]> git.proxmox.com Git - mirror_ubuntu-zesty-kernel.git/blob - arch/i386/kernel/acpi/boot.c
ACPICA: minimal patch to integrate new tables into Linux
[mirror_ubuntu-zesty-kernel.git] / arch / i386 / kernel / acpi / boot.c
1 /*
2 * boot.c - Architecture-Specific Low-Level ACPI Boot Support
3 *
4 * Copyright (C) 2001, 2002 Paul Diefenbaugh <paul.s.diefenbaugh@intel.com>
5 * Copyright (C) 2001 Jun Nakajima <jun.nakajima@intel.com>
6 *
7 * ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
8 *
9 * This program is free software; you can redistribute it and/or modify
10 * it under the terms of the GNU General Public License as published by
11 * the Free Software Foundation; either version 2 of the License, or
12 * (at your option) any later version.
13 *
14 * This program is distributed in the hope that it will be useful,
15 * but WITHOUT ANY WARRANTY; without even the implied warranty of
16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 * GNU General Public License for more details.
18 *
19 * You should have received a copy of the GNU General Public License
20 * along with this program; if not, write to the Free Software
21 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
22 *
23 * ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
24 */
25
26 #include <linux/init.h>
27 #include <linux/acpi.h>
28 #include <linux/efi.h>
29 #include <linux/cpumask.h>
30 #include <linux/module.h>
31 #include <linux/dmi.h>
32 #include <linux/irq.h>
33 #include <linux/bootmem.h>
34 #include <linux/ioport.h>
35
36 #include <asm/pgtable.h>
37 #include <asm/io_apic.h>
38 #include <asm/apic.h>
39 #include <asm/io.h>
40 #include <asm/mpspec.h>
41
42 static int __initdata acpi_force = 0;
43
44 #ifdef CONFIG_ACPI
45 int acpi_disabled = 0;
46 #else
47 int acpi_disabled = 1;
48 #endif
49 EXPORT_SYMBOL(acpi_disabled);
50
51 #ifdef CONFIG_X86_64
52
53 #include <asm/proto.h>
54
55 static inline int acpi_madt_oem_check(char *oem_id, char *oem_table_id) { return 0; }
56
57
58 #else /* X86 */
59
60 #ifdef CONFIG_X86_LOCAL_APIC
61 #include <mach_apic.h>
62 #include <mach_mpparse.h>
63 #endif /* CONFIG_X86_LOCAL_APIC */
64
65 #endif /* X86 */
66
67 #define BAD_MADT_ENTRY(entry, end) ( \
68 (!entry) || (unsigned long)entry + sizeof(*entry) > end || \
69 ((acpi_table_entry_header *)entry)->length < sizeof(*entry))
70
71 #define PREFIX "ACPI: "
72
73 int acpi_noirq; /* skip ACPI IRQ initialization */
74 int acpi_pci_disabled __initdata; /* skip ACPI PCI scan and IRQ initialization */
75 int acpi_ht __initdata = 1; /* enable HT */
76
77 int acpi_lapic;
78 int acpi_ioapic;
79 int acpi_strict;
80 EXPORT_SYMBOL(acpi_strict);
81
82 acpi_interrupt_flags acpi_sci_flags __initdata;
83 int acpi_sci_override_gsi __initdata;
84 int acpi_skip_timer_override __initdata;
85 int acpi_use_timer_override __initdata;
86
87 #ifdef CONFIG_X86_LOCAL_APIC
88 static u64 acpi_lapic_addr __initdata = APIC_DEFAULT_PHYS_BASE;
89 #endif
90
91 #ifndef __HAVE_ARCH_CMPXCHG
92 #warning ACPI uses CMPXCHG, i486 and later hardware
93 #endif
94
95 #define MAX_MADT_ENTRIES 256
96 u8 x86_acpiid_to_apicid[MAX_MADT_ENTRIES] =
97 {[0 ... MAX_MADT_ENTRIES - 1] = 0xff };
98 EXPORT_SYMBOL(x86_acpiid_to_apicid);
99
100 /* --------------------------------------------------------------------------
101 Boot-time Configuration
102 -------------------------------------------------------------------------- */
103
104 /*
105 * The default interrupt routing model is PIC (8259). This gets
106 * overriden if IOAPICs are enumerated (below).
107 */
108 enum acpi_irq_model_id acpi_irq_model = ACPI_IRQ_MODEL_PIC;
109
110 #ifdef CONFIG_X86_64
111
112 /* rely on all ACPI tables being in the direct mapping */
113 char *__acpi_map_table(unsigned long phys_addr, unsigned long size)
114 {
115 if (!phys_addr || !size)
116 return NULL;
117
118 if (phys_addr+size <= (end_pfn_map << PAGE_SHIFT) + PAGE_SIZE)
119 return __va(phys_addr);
120
121 return NULL;
122 }
123
124 #else
125
126 /*
127 * Temporarily use the virtual area starting from FIX_IO_APIC_BASE_END,
128 * to map the target physical address. The problem is that set_fixmap()
129 * provides a single page, and it is possible that the page is not
130 * sufficient.
131 * By using this area, we can map up to MAX_IO_APICS pages temporarily,
132 * i.e. until the next __va_range() call.
133 *
134 * Important Safety Note: The fixed I/O APIC page numbers are *subtracted*
135 * from the fixed base. That's why we start at FIX_IO_APIC_BASE_END and
136 * count idx down while incrementing the phys address.
137 */
138 char *__acpi_map_table(unsigned long phys, unsigned long size)
139 {
140 unsigned long base, offset, mapped_size;
141 int idx;
142
143 if (phys + size < 8 * 1024 * 1024)
144 return __va(phys);
145
146 offset = phys & (PAGE_SIZE - 1);
147 mapped_size = PAGE_SIZE - offset;
148 set_fixmap(FIX_ACPI_END, phys);
149 base = fix_to_virt(FIX_ACPI_END);
150
151 /*
152 * Most cases can be covered by the below.
153 */
154 idx = FIX_ACPI_END;
155 while (mapped_size < size) {
156 if (--idx < FIX_ACPI_BEGIN)
157 return NULL; /* cannot handle this */
158 phys += PAGE_SIZE;
159 set_fixmap(idx, phys);
160 mapped_size += PAGE_SIZE;
161 }
162
163 return ((unsigned char *)base + offset);
164 }
165 #endif
166
167 #ifdef CONFIG_PCI_MMCONFIG
168 /* The physical address of the MMCONFIG aperture. Set from ACPI tables. */
169 struct acpi_table_mcfg_config *pci_mmcfg_config;
170 int pci_mmcfg_config_num;
171
172 int __init acpi_parse_mcfg(unsigned long phys_addr, unsigned long size)
173 {
174 struct acpi_table_mcfg *mcfg;
175 unsigned long i;
176 int config_size;
177
178 if (!phys_addr || !size)
179 return -EINVAL;
180
181 mcfg = (struct acpi_table_mcfg *)__acpi_map_table(phys_addr, size);
182 if (!mcfg) {
183 printk(KERN_WARNING PREFIX "Unable to map MCFG\n");
184 return -ENODEV;
185 }
186
187 /* how many config structures do we have */
188 pci_mmcfg_config_num = 0;
189 i = size - sizeof(struct acpi_table_mcfg);
190 while (i >= sizeof(struct acpi_table_mcfg_config)) {
191 ++pci_mmcfg_config_num;
192 i -= sizeof(struct acpi_table_mcfg_config);
193 };
194 if (pci_mmcfg_config_num == 0) {
195 printk(KERN_ERR PREFIX "MMCONFIG has no entries\n");
196 return -ENODEV;
197 }
198
199 config_size = pci_mmcfg_config_num * sizeof(*pci_mmcfg_config);
200 pci_mmcfg_config = kmalloc(config_size, GFP_KERNEL);
201 if (!pci_mmcfg_config) {
202 printk(KERN_WARNING PREFIX
203 "No memory for MCFG config tables\n");
204 return -ENOMEM;
205 }
206
207 memcpy(pci_mmcfg_config, &mcfg->config, config_size);
208 for (i = 0; i < pci_mmcfg_config_num; ++i) {
209 if (mcfg->config[i].base_reserved) {
210 printk(KERN_ERR PREFIX
211 "MMCONFIG not in low 4GB of memory\n");
212 kfree(pci_mmcfg_config);
213 pci_mmcfg_config_num = 0;
214 return -ENODEV;
215 }
216 }
217
218 return 0;
219 }
220 #endif /* CONFIG_PCI_MMCONFIG */
221
222 #ifdef CONFIG_X86_LOCAL_APIC
223 static int __init acpi_parse_madt(unsigned long phys_addr, unsigned long size)
224 {
225 struct acpi_table_madt *madt = NULL;
226
227 if (!phys_addr || !size || !cpu_has_apic)
228 return -EINVAL;
229
230 madt = (struct acpi_table_madt *)__acpi_map_table(phys_addr, size);
231 if (!madt) {
232 printk(KERN_WARNING PREFIX "Unable to map MADT\n");
233 return -ENODEV;
234 }
235
236 if (madt->lapic_address) {
237 acpi_lapic_addr = (u64) madt->lapic_address;
238
239 printk(KERN_DEBUG PREFIX "Local APIC address 0x%08x\n",
240 madt->lapic_address);
241 }
242
243 acpi_madt_oem_check(madt->header.oem_id, madt->header.oem_table_id);
244
245 return 0;
246 }
247
248 static int __init
249 acpi_parse_lapic(acpi_table_entry_header * header, const unsigned long end)
250 {
251 struct acpi_table_lapic *processor = NULL;
252
253 processor = (struct acpi_table_lapic *)header;
254
255 if (BAD_MADT_ENTRY(processor, end))
256 return -EINVAL;
257
258 acpi_table_print_madt_entry(header);
259
260 /* Record local apic id only when enabled */
261 if (processor->flags.enabled)
262 x86_acpiid_to_apicid[processor->acpi_id] = processor->id;
263
264 /*
265 * We need to register disabled CPU as well to permit
266 * counting disabled CPUs. This allows us to size
267 * cpus_possible_map more accurately, to permit
268 * to not preallocating memory for all NR_CPUS
269 * when we use CPU hotplug.
270 */
271 mp_register_lapic(processor->id, /* APIC ID */
272 processor->flags.enabled); /* Enabled? */
273
274 return 0;
275 }
276
277 static int __init
278 acpi_parse_lapic_addr_ovr(acpi_table_entry_header * header,
279 const unsigned long end)
280 {
281 struct acpi_table_lapic_addr_ovr *lapic_addr_ovr = NULL;
282
283 lapic_addr_ovr = (struct acpi_table_lapic_addr_ovr *)header;
284
285 if (BAD_MADT_ENTRY(lapic_addr_ovr, end))
286 return -EINVAL;
287
288 acpi_lapic_addr = lapic_addr_ovr->address;
289
290 return 0;
291 }
292
293 static int __init
294 acpi_parse_lapic_nmi(acpi_table_entry_header * header, const unsigned long end)
295 {
296 struct acpi_table_lapic_nmi *lapic_nmi = NULL;
297
298 lapic_nmi = (struct acpi_table_lapic_nmi *)header;
299
300 if (BAD_MADT_ENTRY(lapic_nmi, end))
301 return -EINVAL;
302
303 acpi_table_print_madt_entry(header);
304
305 if (lapic_nmi->lint != 1)
306 printk(KERN_WARNING PREFIX "NMI not connected to LINT 1!\n");
307
308 return 0;
309 }
310
311 #endif /*CONFIG_X86_LOCAL_APIC */
312
313 #ifdef CONFIG_X86_IO_APIC
314
315 static int __init
316 acpi_parse_ioapic(acpi_table_entry_header * header, const unsigned long end)
317 {
318 struct acpi_table_ioapic *ioapic = NULL;
319
320 ioapic = (struct acpi_table_ioapic *)header;
321
322 if (BAD_MADT_ENTRY(ioapic, end))
323 return -EINVAL;
324
325 acpi_table_print_madt_entry(header);
326
327 mp_register_ioapic(ioapic->id,
328 ioapic->address, ioapic->global_irq_base);
329
330 return 0;
331 }
332
333 /*
334 * Parse Interrupt Source Override for the ACPI SCI
335 */
336 static void __init acpi_sci_ioapic_setup(u32 gsi, u16 polarity, u16 trigger)
337 {
338 if (trigger == 0) /* compatible SCI trigger is level */
339 trigger = 3;
340
341 if (polarity == 0) /* compatible SCI polarity is low */
342 polarity = 3;
343
344 /* Command-line over-ride via acpi_sci= */
345 if (acpi_sci_flags.trigger)
346 trigger = acpi_sci_flags.trigger;
347
348 if (acpi_sci_flags.polarity)
349 polarity = acpi_sci_flags.polarity;
350
351 /*
352 * mp_config_acpi_legacy_irqs() already setup IRQs < 16
353 * If GSI is < 16, this will update its flags,
354 * else it will create a new mp_irqs[] entry.
355 */
356 mp_override_legacy_irq(gsi, polarity, trigger, gsi);
357
358 /*
359 * stash over-ride to indicate we've been here
360 * and for later update of acpi_fadt
361 */
362 acpi_sci_override_gsi = gsi;
363 return;
364 }
365
366 static int __init
367 acpi_parse_int_src_ovr(acpi_table_entry_header * header,
368 const unsigned long end)
369 {
370 struct acpi_table_int_src_ovr *intsrc = NULL;
371
372 intsrc = (struct acpi_table_int_src_ovr *)header;
373
374 if (BAD_MADT_ENTRY(intsrc, end))
375 return -EINVAL;
376
377 acpi_table_print_madt_entry(header);
378
379 if (intsrc->bus_irq == acpi_fadt.sci_int) {
380 acpi_sci_ioapic_setup(intsrc->global_irq,
381 intsrc->flags.polarity,
382 intsrc->flags.trigger);
383 return 0;
384 }
385
386 if (acpi_skip_timer_override &&
387 intsrc->bus_irq == 0 && intsrc->global_irq == 2) {
388 printk(PREFIX "BIOS IRQ0 pin2 override ignored.\n");
389 return 0;
390 }
391
392 mp_override_legacy_irq(intsrc->bus_irq,
393 intsrc->flags.polarity,
394 intsrc->flags.trigger, intsrc->global_irq);
395
396 return 0;
397 }
398
399 static int __init
400 acpi_parse_nmi_src(acpi_table_entry_header * header, const unsigned long end)
401 {
402 struct acpi_table_nmi_src *nmi_src = NULL;
403
404 nmi_src = (struct acpi_table_nmi_src *)header;
405
406 if (BAD_MADT_ENTRY(nmi_src, end))
407 return -EINVAL;
408
409 acpi_table_print_madt_entry(header);
410
411 /* TBD: Support nimsrc entries? */
412
413 return 0;
414 }
415
416 #endif /* CONFIG_X86_IO_APIC */
417
418 /*
419 * acpi_pic_sci_set_trigger()
420 *
421 * use ELCR to set PIC-mode trigger type for SCI
422 *
423 * If a PIC-mode SCI is not recognized or gives spurious IRQ7's
424 * it may require Edge Trigger -- use "acpi_sci=edge"
425 *
426 * Port 0x4d0-4d1 are ECLR1 and ECLR2, the Edge/Level Control Registers
427 * for the 8259 PIC. bit[n] = 1 means irq[n] is Level, otherwise Edge.
428 * ECLR1 is IRQ's 0-7 (IRQ 0, 1, 2 must be 0)
429 * ECLR2 is IRQ's 8-15 (IRQ 8, 13 must be 0)
430 */
431
432 void __init acpi_pic_sci_set_trigger(unsigned int irq, u16 trigger)
433 {
434 unsigned int mask = 1 << irq;
435 unsigned int old, new;
436
437 /* Real old ELCR mask */
438 old = inb(0x4d0) | (inb(0x4d1) << 8);
439
440 /*
441 * If we use ACPI to set PCI irq's, then we should clear ELCR
442 * since we will set it correctly as we enable the PCI irq
443 * routing.
444 */
445 new = acpi_noirq ? old : 0;
446
447 /*
448 * Update SCI information in the ELCR, it isn't in the PCI
449 * routing tables..
450 */
451 switch (trigger) {
452 case 1: /* Edge - clear */
453 new &= ~mask;
454 break;
455 case 3: /* Level - set */
456 new |= mask;
457 break;
458 }
459
460 if (old == new)
461 return;
462
463 printk(PREFIX "setting ELCR to %04x (from %04x)\n", new, old);
464 outb(new, 0x4d0);
465 outb(new >> 8, 0x4d1);
466 }
467
468 int acpi_gsi_to_irq(u32 gsi, unsigned int *irq)
469 {
470 *irq = gsi;
471 return 0;
472 }
473
474 /*
475 * success: return IRQ number (>=0)
476 * failure: return < 0
477 */
478 int acpi_register_gsi(u32 gsi, int triggering, int polarity)
479 {
480 unsigned int irq;
481 unsigned int plat_gsi = gsi;
482
483 #ifdef CONFIG_PCI
484 /*
485 * Make sure all (legacy) PCI IRQs are set as level-triggered.
486 */
487 if (acpi_irq_model == ACPI_IRQ_MODEL_PIC) {
488 extern void eisa_set_level_irq(unsigned int irq);
489
490 if (triggering == ACPI_LEVEL_SENSITIVE)
491 eisa_set_level_irq(gsi);
492 }
493 #endif
494
495 #ifdef CONFIG_X86_IO_APIC
496 if (acpi_irq_model == ACPI_IRQ_MODEL_IOAPIC) {
497 plat_gsi = mp_register_gsi(gsi, triggering, polarity);
498 }
499 #endif
500 acpi_gsi_to_irq(plat_gsi, &irq);
501 return irq;
502 }
503
504 EXPORT_SYMBOL(acpi_register_gsi);
505
506 /*
507 * ACPI based hotplug support for CPU
508 */
509 #ifdef CONFIG_ACPI_HOTPLUG_CPU
510 int acpi_map_lsapic(acpi_handle handle, int *pcpu)
511 {
512 struct acpi_buffer buffer = { ACPI_ALLOCATE_BUFFER, NULL };
513 union acpi_object *obj;
514 struct acpi_table_lapic *lapic;
515 cpumask_t tmp_map, new_map;
516 u8 physid;
517 int cpu;
518
519 if (ACPI_FAILURE(acpi_evaluate_object(handle, "_MAT", NULL, &buffer)))
520 return -EINVAL;
521
522 if (!buffer.length || !buffer.pointer)
523 return -EINVAL;
524
525 obj = buffer.pointer;
526 if (obj->type != ACPI_TYPE_BUFFER ||
527 obj->buffer.length < sizeof(*lapic)) {
528 kfree(buffer.pointer);
529 return -EINVAL;
530 }
531
532 lapic = (struct acpi_table_lapic *)obj->buffer.pointer;
533
534 if ((lapic->header.type != ACPI_MADT_LAPIC) ||
535 (!lapic->flags.enabled)) {
536 kfree(buffer.pointer);
537 return -EINVAL;
538 }
539
540 physid = lapic->id;
541
542 kfree(buffer.pointer);
543 buffer.length = ACPI_ALLOCATE_BUFFER;
544 buffer.pointer = NULL;
545
546 tmp_map = cpu_present_map;
547 mp_register_lapic(physid, lapic->flags.enabled);
548
549 /*
550 * If mp_register_lapic successfully generates a new logical cpu
551 * number, then the following will get us exactly what was mapped
552 */
553 cpus_andnot(new_map, cpu_present_map, tmp_map);
554 if (cpus_empty(new_map)) {
555 printk ("Unable to map lapic to logical cpu number\n");
556 return -EINVAL;
557 }
558
559 cpu = first_cpu(new_map);
560
561 *pcpu = cpu;
562 return 0;
563 }
564
565 EXPORT_SYMBOL(acpi_map_lsapic);
566
567 int acpi_unmap_lsapic(int cpu)
568 {
569 int i;
570
571 for_each_possible_cpu(i) {
572 if (x86_acpiid_to_apicid[i] == x86_cpu_to_apicid[cpu]) {
573 x86_acpiid_to_apicid[i] = -1;
574 break;
575 }
576 }
577 x86_cpu_to_apicid[cpu] = -1;
578 cpu_clear(cpu, cpu_present_map);
579 num_processors--;
580
581 return (0);
582 }
583
584 EXPORT_SYMBOL(acpi_unmap_lsapic);
585 #endif /* CONFIG_ACPI_HOTPLUG_CPU */
586
587 int acpi_register_ioapic(acpi_handle handle, u64 phys_addr, u32 gsi_base)
588 {
589 /* TBD */
590 return -EINVAL;
591 }
592
593 EXPORT_SYMBOL(acpi_register_ioapic);
594
595 int acpi_unregister_ioapic(acpi_handle handle, u32 gsi_base)
596 {
597 /* TBD */
598 return -EINVAL;
599 }
600
601 EXPORT_SYMBOL(acpi_unregister_ioapic);
602
603 static unsigned long __init
604 acpi_scan_rsdp(unsigned long start, unsigned long length)
605 {
606 unsigned long offset = 0;
607 unsigned long sig_len = sizeof("RSD PTR ") - 1;
608
609 /*
610 * Scan all 16-byte boundaries of the physical memory region for the
611 * RSDP signature.
612 */
613 for (offset = 0; offset < length; offset += 16) {
614 if (strncmp((char *)(phys_to_virt(start) + offset), "RSD PTR ", sig_len))
615 continue;
616 return (start + offset);
617 }
618
619 return 0;
620 }
621
622 static int __init acpi_parse_sbf(unsigned long phys_addr, unsigned long size)
623 {
624 struct acpi_table_sbf *sb;
625
626 if (!phys_addr || !size)
627 return -EINVAL;
628
629 sb = (struct acpi_table_sbf *)__acpi_map_table(phys_addr, size);
630 if (!sb) {
631 printk(KERN_WARNING PREFIX "Unable to map SBF\n");
632 return -ENODEV;
633 }
634
635 sbf_port = sb->sbf_cmos; /* Save CMOS port */
636
637 return 0;
638 }
639
640 #ifdef CONFIG_HPET_TIMER
641
642 static int __init acpi_parse_hpet(unsigned long phys, unsigned long size)
643 {
644 struct acpi_table_hpet *hpet_tbl;
645 struct resource *hpet_res;
646 resource_size_t res_start;
647
648 if (!phys || !size)
649 return -EINVAL;
650
651 hpet_tbl = (struct acpi_table_hpet *)__acpi_map_table(phys, size);
652 if (!hpet_tbl) {
653 printk(KERN_WARNING PREFIX "Unable to map HPET\n");
654 return -ENODEV;
655 }
656
657 if (hpet_tbl->addr.space_id != ACPI_SPACE_MEM) {
658 printk(KERN_WARNING PREFIX "HPET timers must be located in "
659 "memory.\n");
660 return -1;
661 }
662
663 #define HPET_RESOURCE_NAME_SIZE 9
664 hpet_res = alloc_bootmem(sizeof(*hpet_res) + HPET_RESOURCE_NAME_SIZE);
665 if (hpet_res) {
666 memset(hpet_res, 0, sizeof(*hpet_res));
667 hpet_res->name = (void *)&hpet_res[1];
668 hpet_res->flags = IORESOURCE_MEM | IORESOURCE_BUSY;
669 snprintf((char *)hpet_res->name, HPET_RESOURCE_NAME_SIZE,
670 "HPET %u", hpet_tbl->number);
671 hpet_res->end = (1 * 1024) - 1;
672 }
673
674 #ifdef CONFIG_X86_64
675 vxtime.hpet_address = hpet_tbl->addr.addrl |
676 ((long)hpet_tbl->addr.addrh << 32);
677
678 printk(KERN_INFO PREFIX "HPET id: %#x base: %#lx\n",
679 hpet_tbl->id, vxtime.hpet_address);
680
681 res_start = vxtime.hpet_address;
682 #else /* X86 */
683 {
684 extern unsigned long hpet_address;
685
686 hpet_address = hpet_tbl->addr.addrl;
687 printk(KERN_INFO PREFIX "HPET id: %#x base: %#lx\n",
688 hpet_tbl->id, hpet_address);
689
690 res_start = hpet_address;
691 }
692 #endif /* X86 */
693
694 if (hpet_res) {
695 hpet_res->start = res_start;
696 hpet_res->end += res_start;
697 insert_resource(&iomem_resource, hpet_res);
698 }
699
700 return 0;
701 }
702 #else
703 #define acpi_parse_hpet NULL
704 #endif
705
706 #ifdef CONFIG_X86_PM_TIMER
707 extern u32 pmtmr_ioport;
708 #endif
709
710 static int __init acpi_parse_fadt(unsigned long phys, unsigned long size)
711 {
712 struct fadt_descriptor *fadt = NULL;
713
714 fadt = (struct fadt_descriptor *)__acpi_map_table(phys, size);
715 if (!fadt) {
716 printk(KERN_WARNING PREFIX "Unable to map FADT\n");
717 return 0;
718 }
719
720 #ifdef CONFIG_X86_PM_TIMER
721 /* detect the location of the ACPI PM Timer */
722 if (fadt->header.revision >= FADT2_REVISION_ID) {
723 /* FADT rev. 2 */
724 if (fadt->xpm_timer_block.space_id !=
725 ACPI_ADR_SPACE_SYSTEM_IO)
726 return 0;
727
728 pmtmr_ioport = fadt->xpm_timer_block.address;
729 /*
730 * "X" fields are optional extensions to the original V1.0
731 * fields, so we must selectively expand V1.0 fields if the
732 * corresponding X field is zero.
733 */
734 if (!pmtmr_ioport)
735 pmtmr_ioport = fadt->pm_timer_block;
736 } else {
737 /* FADT rev. 1 */
738 pmtmr_ioport = fadt->pm_timer_block;
739 }
740 if (pmtmr_ioport)
741 printk(KERN_INFO PREFIX "PM-Timer IO Port: %#x\n",
742 pmtmr_ioport);
743 #endif
744 return 0;
745 }
746
747 unsigned long __init acpi_find_rsdp(void)
748 {
749 unsigned long rsdp_phys = 0;
750
751 if (efi_enabled) {
752 if (efi.acpi20 != EFI_INVALID_TABLE_ADDR)
753 return efi.acpi20;
754 else if (efi.acpi != EFI_INVALID_TABLE_ADDR)
755 return efi.acpi;
756 }
757 /*
758 * Scan memory looking for the RSDP signature. First search EBDA (low
759 * memory) paragraphs and then search upper memory (E0000-FFFFF).
760 */
761 rsdp_phys = acpi_scan_rsdp(0, 0x400);
762 if (!rsdp_phys)
763 rsdp_phys = acpi_scan_rsdp(0xE0000, 0x20000);
764
765 return rsdp_phys;
766 }
767
768 #ifdef CONFIG_X86_LOCAL_APIC
769 /*
770 * Parse LAPIC entries in MADT
771 * returns 0 on success, < 0 on error
772 */
773 static int __init acpi_parse_madt_lapic_entries(void)
774 {
775 int count;
776
777 if (!cpu_has_apic)
778 return -ENODEV;
779
780 /*
781 * Note that the LAPIC address is obtained from the MADT (32-bit value)
782 * and (optionally) overriden by a LAPIC_ADDR_OVR entry (64-bit value).
783 */
784
785 count =
786 acpi_table_parse_madt(ACPI_MADT_LAPIC_ADDR_OVR,
787 acpi_parse_lapic_addr_ovr, 0);
788 if (count < 0) {
789 printk(KERN_ERR PREFIX
790 "Error parsing LAPIC address override entry\n");
791 return count;
792 }
793
794 mp_register_lapic_address(acpi_lapic_addr);
795
796 count = acpi_table_parse_madt(ACPI_MADT_LAPIC, acpi_parse_lapic,
797 MAX_APICS);
798 if (!count) {
799 printk(KERN_ERR PREFIX "No LAPIC entries present\n");
800 /* TBD: Cleanup to allow fallback to MPS */
801 return -ENODEV;
802 } else if (count < 0) {
803 printk(KERN_ERR PREFIX "Error parsing LAPIC entry\n");
804 /* TBD: Cleanup to allow fallback to MPS */
805 return count;
806 }
807
808 count =
809 acpi_table_parse_madt(ACPI_MADT_LAPIC_NMI, acpi_parse_lapic_nmi, 0);
810 if (count < 0) {
811 printk(KERN_ERR PREFIX "Error parsing LAPIC NMI entry\n");
812 /* TBD: Cleanup to allow fallback to MPS */
813 return count;
814 }
815 return 0;
816 }
817 #endif /* CONFIG_X86_LOCAL_APIC */
818
819 #ifdef CONFIG_X86_IO_APIC
820 /*
821 * Parse IOAPIC related entries in MADT
822 * returns 0 on success, < 0 on error
823 */
824 static int __init acpi_parse_madt_ioapic_entries(void)
825 {
826 int count;
827
828 /*
829 * ACPI interpreter is required to complete interrupt setup,
830 * so if it is off, don't enumerate the io-apics with ACPI.
831 * If MPS is present, it will handle them,
832 * otherwise the system will stay in PIC mode
833 */
834 if (acpi_disabled || acpi_noirq) {
835 return -ENODEV;
836 }
837
838 if (!cpu_has_apic)
839 return -ENODEV;
840
841 /*
842 * if "noapic" boot option, don't look for IO-APICs
843 */
844 if (skip_ioapic_setup) {
845 printk(KERN_INFO PREFIX "Skipping IOAPIC probe "
846 "due to 'noapic' option.\n");
847 return -ENODEV;
848 }
849
850 count =
851 acpi_table_parse_madt(ACPI_MADT_IOAPIC, acpi_parse_ioapic,
852 MAX_IO_APICS);
853 if (!count) {
854 printk(KERN_ERR PREFIX "No IOAPIC entries present\n");
855 return -ENODEV;
856 } else if (count < 0) {
857 printk(KERN_ERR PREFIX "Error parsing IOAPIC entry\n");
858 return count;
859 }
860
861 count =
862 acpi_table_parse_madt(ACPI_MADT_INT_SRC_OVR, acpi_parse_int_src_ovr,
863 NR_IRQ_VECTORS);
864 if (count < 0) {
865 printk(KERN_ERR PREFIX
866 "Error parsing interrupt source overrides entry\n");
867 /* TBD: Cleanup to allow fallback to MPS */
868 return count;
869 }
870
871 /*
872 * If BIOS did not supply an INT_SRC_OVR for the SCI
873 * pretend we got one so we can set the SCI flags.
874 */
875 if (!acpi_sci_override_gsi)
876 acpi_sci_ioapic_setup(acpi_fadt.sci_int, 0, 0);
877
878 /* Fill in identity legacy mapings where no override */
879 mp_config_acpi_legacy_irqs();
880
881 count =
882 acpi_table_parse_madt(ACPI_MADT_NMI_SRC, acpi_parse_nmi_src,
883 NR_IRQ_VECTORS);
884 if (count < 0) {
885 printk(KERN_ERR PREFIX "Error parsing NMI SRC entry\n");
886 /* TBD: Cleanup to allow fallback to MPS */
887 return count;
888 }
889
890 return 0;
891 }
892 #else
893 static inline int acpi_parse_madt_ioapic_entries(void)
894 {
895 return -1;
896 }
897 #endif /* !CONFIG_X86_IO_APIC */
898
899 static void __init acpi_process_madt(void)
900 {
901 #ifdef CONFIG_X86_LOCAL_APIC
902 int count, error;
903
904 count = acpi_table_parse(ACPI_APIC, acpi_parse_madt);
905 if (count >= 1) {
906
907 /*
908 * Parse MADT LAPIC entries
909 */
910 error = acpi_parse_madt_lapic_entries();
911 if (!error) {
912 acpi_lapic = 1;
913
914 #ifdef CONFIG_X86_GENERICARCH
915 generic_bigsmp_probe();
916 #endif
917 /*
918 * Parse MADT IO-APIC entries
919 */
920 error = acpi_parse_madt_ioapic_entries();
921 if (!error) {
922 acpi_irq_model = ACPI_IRQ_MODEL_IOAPIC;
923 acpi_irq_balance_set(NULL);
924 acpi_ioapic = 1;
925
926 smp_found_config = 1;
927 clustered_apic_check();
928 }
929 }
930 if (error == -EINVAL) {
931 /*
932 * Dell Precision Workstation 410, 610 come here.
933 */
934 printk(KERN_ERR PREFIX
935 "Invalid BIOS MADT, disabling ACPI\n");
936 disable_acpi();
937 }
938 }
939 #endif
940 return;
941 }
942
943 #ifdef __i386__
944
945 static int __init disable_acpi_irq(struct dmi_system_id *d)
946 {
947 if (!acpi_force) {
948 printk(KERN_NOTICE "%s detected: force use of acpi=noirq\n",
949 d->ident);
950 acpi_noirq_set();
951 }
952 return 0;
953 }
954
955 static int __init disable_acpi_pci(struct dmi_system_id *d)
956 {
957 if (!acpi_force) {
958 printk(KERN_NOTICE "%s detected: force use of pci=noacpi\n",
959 d->ident);
960 acpi_disable_pci();
961 }
962 return 0;
963 }
964
965 static int __init dmi_disable_acpi(struct dmi_system_id *d)
966 {
967 if (!acpi_force) {
968 printk(KERN_NOTICE "%s detected: acpi off\n", d->ident);
969 disable_acpi();
970 } else {
971 printk(KERN_NOTICE
972 "Warning: DMI blacklist says broken, but acpi forced\n");
973 }
974 return 0;
975 }
976
977 /*
978 * Limit ACPI to CPU enumeration for HT
979 */
980 static int __init force_acpi_ht(struct dmi_system_id *d)
981 {
982 if (!acpi_force) {
983 printk(KERN_NOTICE "%s detected: force use of acpi=ht\n",
984 d->ident);
985 disable_acpi();
986 acpi_ht = 1;
987 } else {
988 printk(KERN_NOTICE
989 "Warning: acpi=force overrules DMI blacklist: acpi=ht\n");
990 }
991 return 0;
992 }
993
994 /*
995 * If your system is blacklisted here, but you find that acpi=force
996 * works for you, please contact acpi-devel@sourceforge.net
997 */
998 static struct dmi_system_id __initdata acpi_dmi_table[] = {
999 /*
1000 * Boxes that need ACPI disabled
1001 */
1002 {
1003 .callback = dmi_disable_acpi,
1004 .ident = "IBM Thinkpad",
1005 .matches = {
1006 DMI_MATCH(DMI_BOARD_VENDOR, "IBM"),
1007 DMI_MATCH(DMI_BOARD_NAME, "2629H1G"),
1008 },
1009 },
1010
1011 /*
1012 * Boxes that need acpi=ht
1013 */
1014 {
1015 .callback = force_acpi_ht,
1016 .ident = "FSC Primergy T850",
1017 .matches = {
1018 DMI_MATCH(DMI_SYS_VENDOR, "FUJITSU SIEMENS"),
1019 DMI_MATCH(DMI_PRODUCT_NAME, "PRIMERGY T850"),
1020 },
1021 },
1022 {
1023 .callback = force_acpi_ht,
1024 .ident = "DELL GX240",
1025 .matches = {
1026 DMI_MATCH(DMI_BOARD_VENDOR, "Dell Computer Corporation"),
1027 DMI_MATCH(DMI_BOARD_NAME, "OptiPlex GX240"),
1028 },
1029 },
1030 {
1031 .callback = force_acpi_ht,
1032 .ident = "HP VISUALIZE NT Workstation",
1033 .matches = {
1034 DMI_MATCH(DMI_BOARD_VENDOR, "Hewlett-Packard"),
1035 DMI_MATCH(DMI_PRODUCT_NAME, "HP VISUALIZE NT Workstation"),
1036 },
1037 },
1038 {
1039 .callback = force_acpi_ht,
1040 .ident = "Compaq Workstation W8000",
1041 .matches = {
1042 DMI_MATCH(DMI_SYS_VENDOR, "Compaq"),
1043 DMI_MATCH(DMI_PRODUCT_NAME, "Workstation W8000"),
1044 },
1045 },
1046 {
1047 .callback = force_acpi_ht,
1048 .ident = "ASUS P4B266",
1049 .matches = {
1050 DMI_MATCH(DMI_BOARD_VENDOR, "ASUSTeK Computer INC."),
1051 DMI_MATCH(DMI_BOARD_NAME, "P4B266"),
1052 },
1053 },
1054 {
1055 .callback = force_acpi_ht,
1056 .ident = "ASUS P2B-DS",
1057 .matches = {
1058 DMI_MATCH(DMI_BOARD_VENDOR, "ASUSTeK Computer INC."),
1059 DMI_MATCH(DMI_BOARD_NAME, "P2B-DS"),
1060 },
1061 },
1062 {
1063 .callback = force_acpi_ht,
1064 .ident = "ASUS CUR-DLS",
1065 .matches = {
1066 DMI_MATCH(DMI_BOARD_VENDOR, "ASUSTeK Computer INC."),
1067 DMI_MATCH(DMI_BOARD_NAME, "CUR-DLS"),
1068 },
1069 },
1070 {
1071 .callback = force_acpi_ht,
1072 .ident = "ABIT i440BX-W83977",
1073 .matches = {
1074 DMI_MATCH(DMI_BOARD_VENDOR, "ABIT <http://www.abit.com>"),
1075 DMI_MATCH(DMI_BOARD_NAME, "i440BX-W83977 (BP6)"),
1076 },
1077 },
1078 {
1079 .callback = force_acpi_ht,
1080 .ident = "IBM Bladecenter",
1081 .matches = {
1082 DMI_MATCH(DMI_BOARD_VENDOR, "IBM"),
1083 DMI_MATCH(DMI_BOARD_NAME, "IBM eServer BladeCenter HS20"),
1084 },
1085 },
1086 {
1087 .callback = force_acpi_ht,
1088 .ident = "IBM eServer xSeries 360",
1089 .matches = {
1090 DMI_MATCH(DMI_BOARD_VENDOR, "IBM"),
1091 DMI_MATCH(DMI_BOARD_NAME, "eServer xSeries 360"),
1092 },
1093 },
1094 {
1095 .callback = force_acpi_ht,
1096 .ident = "IBM eserver xSeries 330",
1097 .matches = {
1098 DMI_MATCH(DMI_BOARD_VENDOR, "IBM"),
1099 DMI_MATCH(DMI_BOARD_NAME, "eserver xSeries 330"),
1100 },
1101 },
1102 {
1103 .callback = force_acpi_ht,
1104 .ident = "IBM eserver xSeries 440",
1105 .matches = {
1106 DMI_MATCH(DMI_BOARD_VENDOR, "IBM"),
1107 DMI_MATCH(DMI_PRODUCT_NAME, "eserver xSeries 440"),
1108 },
1109 },
1110
1111 /*
1112 * Boxes that need ACPI PCI IRQ routing disabled
1113 */
1114 {
1115 .callback = disable_acpi_irq,
1116 .ident = "ASUS A7V",
1117 .matches = {
1118 DMI_MATCH(DMI_BOARD_VENDOR, "ASUSTeK Computer INC"),
1119 DMI_MATCH(DMI_BOARD_NAME, "<A7V>"),
1120 /* newer BIOS, Revision 1011, does work */
1121 DMI_MATCH(DMI_BIOS_VERSION,
1122 "ASUS A7V ACPI BIOS Revision 1007"),
1123 },
1124 },
1125
1126 /*
1127 * Boxes that need ACPI PCI IRQ routing and PCI scan disabled
1128 */
1129 { /* _BBN 0 bug */
1130 .callback = disable_acpi_pci,
1131 .ident = "ASUS PR-DLS",
1132 .matches = {
1133 DMI_MATCH(DMI_BOARD_VENDOR, "ASUSTeK Computer INC."),
1134 DMI_MATCH(DMI_BOARD_NAME, "PR-DLS"),
1135 DMI_MATCH(DMI_BIOS_VERSION,
1136 "ASUS PR-DLS ACPI BIOS Revision 1010"),
1137 DMI_MATCH(DMI_BIOS_DATE, "03/21/2003")
1138 },
1139 },
1140 {
1141 .callback = disable_acpi_pci,
1142 .ident = "Acer TravelMate 36x Laptop",
1143 .matches = {
1144 DMI_MATCH(DMI_SYS_VENDOR, "Acer"),
1145 DMI_MATCH(DMI_PRODUCT_NAME, "TravelMate 360"),
1146 },
1147 },
1148 {}
1149 };
1150
1151 #endif /* __i386__ */
1152
1153 /*
1154 * acpi_boot_table_init() and acpi_boot_init()
1155 * called from setup_arch(), always.
1156 * 1. checksums all tables
1157 * 2. enumerates lapics
1158 * 3. enumerates io-apics
1159 *
1160 * acpi_table_init() is separate to allow reading SRAT without
1161 * other side effects.
1162 *
1163 * side effects of acpi_boot_init:
1164 * acpi_lapic = 1 if LAPIC found
1165 * acpi_ioapic = 1 if IOAPIC found
1166 * if (acpi_lapic && acpi_ioapic) smp_found_config = 1;
1167 * if acpi_blacklisted() acpi_disabled = 1;
1168 * acpi_irq_model=...
1169 * ...
1170 *
1171 * return value: (currently ignored)
1172 * 0: success
1173 * !0: failure
1174 */
1175
1176 int __init acpi_boot_table_init(void)
1177 {
1178 int error;
1179
1180 #ifdef __i386__
1181 dmi_check_system(acpi_dmi_table);
1182 #endif
1183
1184 /*
1185 * If acpi_disabled, bail out
1186 * One exception: acpi=ht continues far enough to enumerate LAPICs
1187 */
1188 if (acpi_disabled && !acpi_ht)
1189 return 1;
1190
1191 /*
1192 * Initialize the ACPI boot-time table parser.
1193 */
1194 error = acpi_table_init();
1195 if (error) {
1196 disable_acpi();
1197 return error;
1198 }
1199
1200 acpi_table_parse(ACPI_BOOT, acpi_parse_sbf);
1201
1202 /*
1203 * blacklist may disable ACPI entirely
1204 */
1205 error = acpi_blacklisted();
1206 if (error) {
1207 if (acpi_force) {
1208 printk(KERN_WARNING PREFIX "acpi=force override\n");
1209 } else {
1210 printk(KERN_WARNING PREFIX "Disabling ACPI support\n");
1211 disable_acpi();
1212 return error;
1213 }
1214 }
1215
1216 return 0;
1217 }
1218
1219 int __init acpi_boot_init(void)
1220 {
1221 /*
1222 * If acpi_disabled, bail out
1223 * One exception: acpi=ht continues far enough to enumerate LAPICs
1224 */
1225 if (acpi_disabled && !acpi_ht)
1226 return 1;
1227
1228 acpi_table_parse(ACPI_BOOT, acpi_parse_sbf);
1229
1230 /*
1231 * set sci_int and PM timer address
1232 */
1233 acpi_table_parse(ACPI_FADT, acpi_parse_fadt);
1234
1235 /*
1236 * Process the Multiple APIC Description Table (MADT), if present
1237 */
1238 acpi_process_madt();
1239
1240 acpi_table_parse(ACPI_HPET, acpi_parse_hpet);
1241
1242 return 0;
1243 }
1244
1245 static int __init parse_acpi(char *arg)
1246 {
1247 if (!arg)
1248 return -EINVAL;
1249
1250 /* "acpi=off" disables both ACPI table parsing and interpreter */
1251 if (strcmp(arg, "off") == 0) {
1252 disable_acpi();
1253 }
1254 /* acpi=force to over-ride black-list */
1255 else if (strcmp(arg, "force") == 0) {
1256 acpi_force = 1;
1257 acpi_ht = 1;
1258 acpi_disabled = 0;
1259 }
1260 /* acpi=strict disables out-of-spec workarounds */
1261 else if (strcmp(arg, "strict") == 0) {
1262 acpi_strict = 1;
1263 }
1264 /* Limit ACPI just to boot-time to enable HT */
1265 else if (strcmp(arg, "ht") == 0) {
1266 if (!acpi_force)
1267 disable_acpi();
1268 acpi_ht = 1;
1269 }
1270 /* "acpi=noirq" disables ACPI interrupt routing */
1271 else if (strcmp(arg, "noirq") == 0) {
1272 acpi_noirq_set();
1273 } else {
1274 /* Core will printk when we return error. */
1275 return -EINVAL;
1276 }
1277 return 0;
1278 }
1279 early_param("acpi", parse_acpi);
1280
1281 /* FIXME: Using pci= for an ACPI parameter is a travesty. */
1282 static int __init parse_pci(char *arg)
1283 {
1284 if (arg && strcmp(arg, "noacpi") == 0)
1285 acpi_disable_pci();
1286 return 0;
1287 }
1288 early_param("pci", parse_pci);
1289
1290 #ifdef CONFIG_X86_IO_APIC
1291 static int __init parse_acpi_skip_timer_override(char *arg)
1292 {
1293 acpi_skip_timer_override = 1;
1294 return 0;
1295 }
1296 early_param("acpi_skip_timer_override", parse_acpi_skip_timer_override);
1297
1298 static int __init parse_acpi_use_timer_override(char *arg)
1299 {
1300 acpi_use_timer_override = 1;
1301 return 0;
1302 }
1303 early_param("acpi_use_timer_override", parse_acpi_use_timer_override);
1304 #endif /* CONFIG_X86_IO_APIC */
1305
1306 static int __init setup_acpi_sci(char *s)
1307 {
1308 if (!s)
1309 return -EINVAL;
1310 if (!strcmp(s, "edge"))
1311 acpi_sci_flags.trigger = 1;
1312 else if (!strcmp(s, "level"))
1313 acpi_sci_flags.trigger = 3;
1314 else if (!strcmp(s, "high"))
1315 acpi_sci_flags.polarity = 1;
1316 else if (!strcmp(s, "low"))
1317 acpi_sci_flags.polarity = 3;
1318 else
1319 return -EINVAL;
1320 return 0;
1321 }
1322 early_param("acpi_sci", setup_acpi_sci);
1323
1324 int __acpi_acquire_global_lock(unsigned int *lock)
1325 {
1326 unsigned int old, new, val;
1327 do {
1328 old = *lock;
1329 new = (((old & ~0x3) + 2) + ((old >> 1) & 0x1));
1330 val = cmpxchg(lock, old, new);
1331 } while (unlikely (val != old));
1332 return (new < 3) ? -1 : 0;
1333 }
1334
1335 int __acpi_release_global_lock(unsigned int *lock)
1336 {
1337 unsigned int old, new, val;
1338 do {
1339 old = *lock;
1340 new = old & ~0x3;
1341 val = cmpxchg(lock, old, new);
1342 } while (unlikely (val != old));
1343 return old & 0x1;
1344 }