2 2005.12 yhlu add coreboot_ram cross the vga font buffer handling
3 2005.12 yhlu add CONFIG_RAMBASE above 1M support for SMP
4 2008.05 stepan add support for going back to sipi wait state
7 #include <cpu/x86/lapic.h>
10 #include <console/console.h>
12 #include <device/device.h>
13 #include <device/path.h>
14 #include <smp/atomic.h>
15 #include <smp/spinlock.h>
20 #if CONFIG_RAMBASE >= 0x100000
21 /* This is a lot more paranoid now, since Linux can NOT handle
22 * being told there is a CPU when none exists. So any errors
23 * will return 0, meaning no CPU.
25 * We actually handling that case by noting which cpus startup
26 * and not telling anyone about the ones that dont.
28 static unsigned long get_valid_start_eip(unsigned long orig_start_eip)
30 return (unsigned long)orig_start_eip & 0xffff; // 16 bit to avoid 0xa0000
34 #if CONFIG_HAVE_ACPI_RESUME == 1
36 char *lowmem_backup_ptr;
37 int lowmem_backup_size;
40 extern char _secondary_start[];
42 static void copy_secondary_start_to_1m_below(void)
44 #if CONFIG_RAMBASE >= 0x100000
45 extern char _secondary_start_end[];
46 unsigned long code_size;
47 unsigned long start_eip;
49 /* _secondary_start need to be masked 20 above bit, because 16 bit code in secondary.S
50 Also We need to copy the _secondary_start to the below 1M region
52 start_eip = get_valid_start_eip((unsigned long)_secondary_start);
53 code_size = (unsigned long)_secondary_start_end - (unsigned long)_secondary_start;
55 #if CONFIG_HAVE_ACPI_RESUME == 1
56 /* need to save it for RAM resume */
57 lowmem_backup_size = code_size;
58 lowmem_backup = malloc(code_size);
59 lowmem_backup_ptr = (char *)start_eip;
61 if (lowmem_backup == NULL)
62 die("Out of backup memory\n");
64 memcpy(lowmem_backup, lowmem_backup_ptr, lowmem_backup_size);
66 /* copy the _secondary_start to the ram below 1M*/
67 memcpy((unsigned char *)start_eip, (unsigned char *)_secondary_start, code_size);
69 printk(BIOS_DEBUG, "start_eip=0x%08lx, offset=0x%08lx, code_size=0x%08lx\n", start_eip, ((unsigned long)_secondary_start - start_eip), code_size);
73 static int lapic_start_cpu(unsigned long apicid)
76 unsigned long send_status, accept_status, start_eip;
77 int j, num_starts, maxlvt;
80 * Starting actual IPI sequence...
83 printk(BIOS_SPEW, "Asserting INIT.\n");
86 * Turn INIT on target chip
88 lapic_write_around(LAPIC_ICR2, SET_LAPIC_DEST_FIELD(apicid));
94 lapic_write_around(LAPIC_ICR, LAPIC_INT_LEVELTRIG | LAPIC_INT_ASSERT
97 printk(BIOS_SPEW, "Waiting for send to finish...\n");
100 printk(BIOS_SPEW, "+");
102 send_status = lapic_read(LAPIC_ICR) & LAPIC_ICR_BUSY;
103 } while (send_status && (timeout++ < 1000));
104 if (timeout >= 1000) {
105 printk(BIOS_ERR, "CPU %ld: First apic write timed out. Disabling\n",
108 printk(BIOS_ERR, "ESR is 0x%lx\n", lapic_read(LAPIC_ESR));
109 if (lapic_read(LAPIC_ESR)) {
110 printk(BIOS_ERR, "Try to reset ESR\n");
111 lapic_write_around(LAPIC_ESR, 0);
112 printk(BIOS_ERR, "ESR is 0x%lx\n", lapic_read(LAPIC_ESR));
116 #if !defined (CONFIG_CPU_AMD_MODEL_10XXX) && !defined (CONFIG_CPU_AMD_MODEL_14XXX)
120 printk(BIOS_SPEW, "Deasserting INIT.\n");
123 lapic_write_around(LAPIC_ICR2, SET_LAPIC_DEST_FIELD(apicid));
126 lapic_write_around(LAPIC_ICR, LAPIC_INT_LEVELTRIG | LAPIC_DM_INIT);
128 printk(BIOS_SPEW, "Waiting for send to finish...\n");
131 printk(BIOS_SPEW, "+");
133 send_status = lapic_read(LAPIC_ICR) & LAPIC_ICR_BUSY;
134 } while (send_status && (timeout++ < 1000));
135 if (timeout >= 1000) {
136 printk(BIOS_ERR, "CPU %ld: Second apic write timed out. Disabling\n",
142 #if CONFIG_RAMBASE >= 0x100000
143 start_eip = get_valid_start_eip((unsigned long)_secondary_start);
145 start_eip = (unsigned long)_secondary_start;
148 #if !defined (CONFIG_CPU_AMD_MODEL_10XXX) && !defined (CONFIG_CPU_AMD_MODEL_14XXX)
155 * Run STARTUP IPI loop.
157 printk(BIOS_SPEW, "#startup loops: %d.\n", num_starts);
161 for (j = 1; j <= num_starts; j++) {
162 printk(BIOS_SPEW, "Sending STARTUP #%d to %lu.\n", j, apicid);
163 lapic_read_around(LAPIC_SPIV);
164 lapic_write(LAPIC_ESR, 0);
165 lapic_read(LAPIC_ESR);
166 printk(BIOS_SPEW, "After apic_write.\n");
173 lapic_write_around(LAPIC_ICR2, SET_LAPIC_DEST_FIELD(apicid));
175 /* Boot on the stack */
176 /* Kick the second */
177 lapic_write_around(LAPIC_ICR, LAPIC_DM_STARTUP
178 | (start_eip >> 12));
181 * Give the other CPU some time to accept the IPI.
185 printk(BIOS_SPEW, "Startup point 1.\n");
187 printk(BIOS_SPEW, "Waiting for send to finish...\n");
190 printk(BIOS_SPEW, "+");
192 send_status = lapic_read(LAPIC_ICR) & LAPIC_ICR_BUSY;
193 } while (send_status && (timeout++ < 1000));
196 * Give the other CPU some time to accept the IPI.
200 * Due to the Pentium erratum 3AP.
203 lapic_read_around(LAPIC_SPIV);
204 lapic_write(LAPIC_ESR, 0);
206 accept_status = (lapic_read(LAPIC_ESR) & 0xEF);
207 if (send_status || accept_status)
210 printk(BIOS_SPEW, "After Startup.\n");
212 printk(BIOS_WARNING, "APIC never delivered???\n");
214 printk(BIOS_WARNING, "APIC delivery error (%lx).\n", accept_status);
215 if (send_status || accept_status)
220 /* Number of cpus that are currently running in coreboot */
221 static atomic_t active_cpus = ATOMIC_INIT(1);
223 /* start_cpu_lock covers last_cpu_index and secondary_stack.
224 * Only starting one cpu at a time let's me remove the logic
225 * for select the stack from assembly language.
227 * In addition communicating by variables to the cpu I
228 * am starting allows me to veryify it has started before
232 static spinlock_t start_cpu_lock = SPIN_LOCK_UNLOCKED;
233 static unsigned last_cpu_index = 0;
234 volatile unsigned long secondary_stack;
236 int start_cpu(device_t cpu)
238 extern unsigned char _estack[];
239 struct cpu_info *info;
240 unsigned long stack_end;
241 unsigned long apicid;
246 spin_lock(&start_cpu_lock);
248 /* Get the cpu's apicid */
249 apicid = cpu->path.apic.apic_id;
251 /* Get an index for the new processor */
252 index = ++last_cpu_index;
254 /* Find end of the new processors stack */
255 stack_end = ((unsigned long)_estack) - (CONFIG_STACK_SIZE*index) - sizeof(struct cpu_info);
257 /* Record the index and which cpu structure we are using */
258 info = (struct cpu_info *)stack_end;
262 /* Advertise the new stack to start_cpu */
263 secondary_stack = stack_end;
265 /* Until the cpu starts up report the cpu is not enabled */
267 cpu->initialized = 0;
270 result = lapic_start_cpu(apicid);
274 /* Wait 1s or until the new the new cpu calls in */
275 for(count = 0; count < 100000 ; count++) {
276 if (secondary_stack == 0) {
284 spin_unlock(&start_cpu_lock);
288 #if CONFIG_AP_IN_SIPI_WAIT == 1
290 * Normally this function is defined in lapic.h as an always inline function
291 * that just keeps the CPU in a hlt() loop. This does not work on all CPUs.
292 * I think all hyperthreading CPUs might need this version, but I could only
293 * verify this on the Intel Core Duo
295 void stop_this_cpu(void)
298 unsigned long send_status;
301 id = lapic_read(LAPIC_ID) >> 24;
303 printk(BIOS_DEBUG, "CPU %ld going down...\n", id);
305 /* send an LAPIC INIT to myself */
306 lapic_write_around(LAPIC_ICR2, SET_LAPIC_DEST_FIELD(id));
307 lapic_write_around(LAPIC_ICR, LAPIC_INT_LEVELTRIG | LAPIC_INT_ASSERT | LAPIC_DM_INIT);
309 /* wait for the ipi send to finish */
311 // When these two printk(BIOS_SPEW, ...) calls are not removed, the
312 // machine will hang when log level is SPEW. Why?
313 printk(BIOS_SPEW, "Waiting for send to finish...\n");
318 printk(BIOS_SPEW, "+");
321 send_status = lapic_read(LAPIC_ICR) & LAPIC_ICR_BUSY;
322 } while (send_status && (timeout++ < 1000));
323 if (timeout >= 1000) {
324 printk(BIOS_ERR, "timed out\n");
328 printk(BIOS_SPEW, "Deasserting INIT.\n");
329 /* Deassert the LAPIC INIT */
330 lapic_write_around(LAPIC_ICR2, SET_LAPIC_DEST_FIELD(id));
331 lapic_write_around(LAPIC_ICR, LAPIC_INT_LEVELTRIG | LAPIC_DM_INIT);
333 printk(BIOS_SPEW, "Waiting for send to finish...\n");
336 printk(BIOS_SPEW, "+");
338 send_status = lapic_read(LAPIC_ICR) & LAPIC_ICR_BUSY;
339 } while (send_status && (timeout++ < 1000));
340 if (timeout >= 1000) {
341 printk(BIOS_ERR, "timed out\n");
350 /* C entry point of secondary cpus */
351 void secondary_cpu_init(void)
353 atomic_inc(&active_cpus);
354 #if CONFIG_SERIAL_CPU_INIT == 1
355 #if CONFIG_MAX_CPUS>2
356 spin_lock(&start_cpu_lock);
360 #if CONFIG_SERIAL_CPU_INIT == 1
361 #if CONFIG_MAX_CPUS>2
362 spin_unlock(&start_cpu_lock);
366 atomic_dec(&active_cpus);
371 static void start_other_cpus(struct bus *cpu_bus, device_t bsp_cpu)
374 /* Loop through the cpus once getting them started */
376 for(cpu = cpu_bus->children; cpu ; cpu = cpu->sibling) {
377 if (cpu->path.type != DEVICE_PATH_APIC) {
380 #if CONFIG_SERIAL_CPU_INIT == 0
390 if (cpu->initialized) {
394 if (!start_cpu(cpu)) {
395 /* Record the error in cpu? */
396 printk(BIOS_ERR, "CPU 0x%02x would not start!\n",
397 cpu->path.apic.apic_id);
399 #if CONFIG_SERIAL_CPU_INIT == 1
400 #if CONFIG_MAX_CPUS>2
408 static void wait_other_cpus_stop(struct bus *cpu_bus)
411 int old_active_count, active_count;
412 /* Now loop until the other cpus have finished initializing */
413 old_active_count = 1;
414 active_count = atomic_read(&active_cpus);
415 while(active_count > 1) {
416 if (active_count != old_active_count) {
417 printk(BIOS_INFO, "Waiting for %d CPUS to stop\n", active_count - 1);
418 old_active_count = active_count;
421 active_count = atomic_read(&active_cpus);
423 for(cpu = cpu_bus->children; cpu; cpu = cpu->sibling) {
424 if (cpu->path.type != DEVICE_PATH_APIC) {
427 if (!cpu->initialized) {
428 printk(BIOS_ERR, "CPU 0x%02x did not initialize!\n",
429 cpu->path.apic.apic_id);
432 printk(BIOS_DEBUG, "All AP CPUs stopped\n");
435 #else /* CONFIG_SMP */
436 #define initialize_other_cpus(root) do {} while(0)
437 #endif /* CONFIG_SMP */
439 void initialize_cpus(struct bus *cpu_bus)
441 struct device_path cpu_path;
442 struct cpu_info *info;
444 /* Find the info struct for this cpu */
448 /* Ensure the local apic is enabled */
451 /* Get the device path of the boot cpu */
452 cpu_path.type = DEVICE_PATH_APIC;
453 cpu_path.apic.apic_id = lapicid();
455 /* Get the device path of the boot cpu */
456 cpu_path.type = DEVICE_PATH_CPU;
460 /* Find the device structure for the boot cpu */
461 info->cpu = alloc_find_dev(cpu_bus, &cpu_path);
464 copy_secondary_start_to_1m_below(); // why here? In case some day we can start core1 in amd_sibling_init
467 #if CONFIG_HAVE_SMI_HANDLER
471 cpus_ready_for_init();
474 #if CONFIG_SERIAL_CPU_INIT == 0
475 /* start all aps at first, so we can init ECC all together */
476 start_other_cpus(cpu_bus, info->cpu);
480 /* Initialize the bootstrap processor */
484 #if CONFIG_SERIAL_CPU_INIT == 1
485 start_other_cpus(cpu_bus, info->cpu);
488 /* Now wait the rest of the cpus stop*/
489 wait_other_cpus_stop(cpu_bus);