/* * Copyright (c) 2005 Jakub Jermar * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * * - Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * - Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * - The name of the author may not be used to endorse or promote products * derived from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ /** @addtogroup ia32 * @{ */ /** @file */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #ifdef CONFIG_SMP static struct smp_config_operations *ops = NULL; void smp_init(void) { uintptr_t l_apic_address, io_apic_address; if (acpi_madt) { acpi_madt_parse(); ops = &madt_config_operations; } if (config.cpu_count == 1) { mps_init(); ops = &mps_config_operations; } l_apic_address = (uintptr_t) frame_alloc(ONE_FRAME, FRAME_ATOMIC | FRAME_KA); if (!l_apic_address) panic("cannot allocate address for l_apic\n"); io_apic_address = (uintptr_t) frame_alloc(ONE_FRAME, FRAME_ATOMIC | FRAME_KA); if (!io_apic_address) panic("cannot allocate address for io_apic\n"); if (config.cpu_count > 1) { page_mapping_insert(AS_KERNEL, l_apic_address, (uintptr_t) l_apic, PAGE_NOT_CACHEABLE); page_mapping_insert(AS_KERNEL, io_apic_address, (uintptr_t) io_apic, PAGE_NOT_CACHEABLE); l_apic = (uint32_t *) l_apic_address; io_apic = (uint32_t *) io_apic_address; } } /* * Kernel thread for bringing up application processors. It becomes clear * that we need an arrangement like this (AP's being initialized by a kernel * thread), for a thread has its dedicated stack. (The stack used during the * BSP initialization (prior the very first call to scheduler()) will be used * as an initialization stack for each AP.) */ void kmp(void *arg) { unsigned int i; ASSERT(ops != NULL); /* * We need to access data in frame 0. * We boldly make use of kernel address space mapping. */ /* * Set the warm-reset vector to the real-mode address of 4K-aligned ap_boot() */ *((uint16_t *) (PA2KA(0x467+0))) = ((uintptr_t) ap_boot) >> 4; /* segment */ *((uint16_t *) (PA2KA(0x467+2))) = 0; /* offset */ /* * Save 0xa to address 0xf of the CMOS RAM. * BIOS will not do the POST after the INIT signal. */ outb(0x70,0xf); outb(0x71,0xa); pic_disable_irqs(0xffff); apic_init(); for (i = 0; i < ops->cpu_count(); i++) { struct descriptor *gdt_new; /* * Skip processors marked unusable. */ if (!ops->cpu_enabled(i)) continue; /* * The bootstrap processor is already up. */ if (ops->cpu_bootstrap(i)) continue; if (ops->cpu_apic_id(i) == l_apic_id()) { printf("%s: bad processor entry #%d, will not send IPI to myself\n", __FUNCTION__, i); continue; } /* * Prepare new GDT for CPU in question. */ if (!(gdt_new = (struct descriptor *) malloc(GDT_ITEMS*sizeof(struct descriptor), FRAME_ATOMIC))) panic("couldn't allocate memory for GDT\n"); memcpy(gdt_new, gdt, GDT_ITEMS * sizeof(struct descriptor)); memsetb((uintptr_t)(&gdt_new[TSS_DES]), sizeof(struct descriptor), 0); protected_ap_gdtr.limit = GDT_ITEMS * sizeof(struct descriptor); protected_ap_gdtr.base = KA2PA((uintptr_t) gdt_new); gdtr.base = (uintptr_t) gdt_new; if (l_apic_send_init_ipi(ops->cpu_apic_id(i))) { /* * There may be just one AP being initialized at * the time. After it comes completely up, it is * supposed to wake us up. */ if (waitq_sleep_timeout(&ap_completion_wq, 1000000, SYNCH_FLAGS_NONE) == ESYNCH_TIMEOUT) printf("%s: waiting for cpu%d (APIC ID = %d) timed out\n", __FUNCTION__, config.cpu_active > i ? config.cpu_active : i, ops->cpu_apic_id(i)); } else printf("INIT IPI for l_apic%d failed\n", ops->cpu_apic_id(i)); } } int smp_irq_to_pin(int irq) { ASSERT(ops != NULL); return ops->irq_to_pin(irq); } #endif /* CONFIG_SMP */ /** @} */