]>
Commit | Line | Data |
---|---|---|
1 | /* | |
2 | * acpi_osl.c - OS-dependent functions ($Revision: 83 $) | |
3 | * | |
4 | * Copyright (C) 2000 Andrew Henroid | |
5 | * Copyright (C) 2001, 2002 Andy Grover <andrew.grover@intel.com> | |
6 | * Copyright (C) 2001, 2002 Paul Diefenbaugh <paul.s.diefenbaugh@intel.com> | |
7 | * | |
8 | * ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ | |
9 | * | |
10 | * This program is free software; you can redistribute it and/or modify | |
11 | * it under the terms of the GNU General Public License as published by | |
12 | * the Free Software Foundation; either version 2 of the License, or | |
13 | * (at your option) any later version. | |
14 | * | |
15 | * This program is distributed in the hope that it will be useful, | |
16 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
17 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
18 | * GNU General Public License for more details. | |
19 | * | |
20 | * You should have received a copy of the GNU General Public License | |
21 | * along with this program; if not, write to the Free Software | |
22 | * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA | |
23 | * | |
24 | * ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ | |
25 | * | |
26 | */ | |
27 | ||
28 | #include <linux/module.h> | |
29 | #include <linux/kernel.h> | |
30 | #include <linux/slab.h> | |
31 | #include <linux/mm.h> | |
32 | #include <linux/pci.h> | |
33 | #include <linux/interrupt.h> | |
34 | #include <linux/kmod.h> | |
35 | #include <linux/delay.h> | |
36 | #include <linux/dmi.h> | |
37 | #include <linux/workqueue.h> | |
38 | #include <linux/nmi.h> | |
39 | #include <linux/acpi.h> | |
40 | #include <acpi/acpi.h> | |
41 | #include <asm/io.h> | |
42 | #include <acpi/acpi_bus.h> | |
43 | #include <acpi/processor.h> | |
44 | #include <asm/uaccess.h> | |
45 | ||
46 | #include <linux/efi.h> | |
47 | ||
48 | #define _COMPONENT ACPI_OS_SERVICES | |
49 | ACPI_MODULE_NAME("osl"); | |
50 | #define PREFIX "ACPI: " | |
51 | struct acpi_os_dpc { | |
52 | acpi_osd_exec_callback function; | |
53 | void *context; | |
54 | struct work_struct work; | |
55 | }; | |
56 | ||
57 | #ifdef CONFIG_ACPI_CUSTOM_DSDT | |
58 | #include CONFIG_ACPI_CUSTOM_DSDT_FILE | |
59 | #endif | |
60 | ||
61 | #ifdef ENABLE_DEBUGGER | |
62 | #include <linux/kdb.h> | |
63 | ||
64 | /* stuff for debugger support */ | |
65 | int acpi_in_debugger; | |
66 | EXPORT_SYMBOL(acpi_in_debugger); | |
67 | ||
68 | extern char line_buf[80]; | |
69 | #endif /*ENABLE_DEBUGGER */ | |
70 | ||
71 | static unsigned int acpi_irq_irq; | |
72 | static acpi_osd_handler acpi_irq_handler; | |
73 | static void *acpi_irq_context; | |
74 | static struct workqueue_struct *kacpid_wq; | |
75 | static struct workqueue_struct *kacpi_notify_wq; | |
76 | ||
77 | #define OSI_STRING_LENGTH_MAX 64 /* arbitrary */ | |
78 | static char osi_additional_string[OSI_STRING_LENGTH_MAX]; | |
79 | ||
80 | static int osi_linux; /* disable _OSI(Linux) by default */ | |
81 | ||
82 | #ifdef CONFIG_DMI | |
83 | static struct __initdata dmi_system_id acpi_osl_dmi_table[]; | |
84 | #endif | |
85 | ||
86 | static void __init acpi_request_region (struct acpi_generic_address *addr, | |
87 | unsigned int length, char *desc) | |
88 | { | |
89 | struct resource *res; | |
90 | ||
91 | if (!addr->address || !length) | |
92 | return; | |
93 | ||
94 | if (addr->space_id == ACPI_ADR_SPACE_SYSTEM_IO) | |
95 | res = request_region(addr->address, length, desc); | |
96 | else if (addr->space_id == ACPI_ADR_SPACE_SYSTEM_MEMORY) | |
97 | res = request_mem_region(addr->address, length, desc); | |
98 | } | |
99 | ||
100 | static int __init acpi_reserve_resources(void) | |
101 | { | |
102 | acpi_request_region(&acpi_gbl_FADT.xpm1a_event_block, acpi_gbl_FADT.pm1_event_length, | |
103 | "ACPI PM1a_EVT_BLK"); | |
104 | ||
105 | acpi_request_region(&acpi_gbl_FADT.xpm1b_event_block, acpi_gbl_FADT.pm1_event_length, | |
106 | "ACPI PM1b_EVT_BLK"); | |
107 | ||
108 | acpi_request_region(&acpi_gbl_FADT.xpm1a_control_block, acpi_gbl_FADT.pm1_control_length, | |
109 | "ACPI PM1a_CNT_BLK"); | |
110 | ||
111 | acpi_request_region(&acpi_gbl_FADT.xpm1b_control_block, acpi_gbl_FADT.pm1_control_length, | |
112 | "ACPI PM1b_CNT_BLK"); | |
113 | ||
114 | if (acpi_gbl_FADT.pm_timer_length == 4) | |
115 | acpi_request_region(&acpi_gbl_FADT.xpm_timer_block, 4, "ACPI PM_TMR"); | |
116 | ||
117 | acpi_request_region(&acpi_gbl_FADT.xpm2_control_block, acpi_gbl_FADT.pm2_control_length, | |
118 | "ACPI PM2_CNT_BLK"); | |
119 | ||
120 | /* Length of GPE blocks must be a non-negative multiple of 2 */ | |
121 | ||
122 | if (!(acpi_gbl_FADT.gpe0_block_length & 0x1)) | |
123 | acpi_request_region(&acpi_gbl_FADT.xgpe0_block, | |
124 | acpi_gbl_FADT.gpe0_block_length, "ACPI GPE0_BLK"); | |
125 | ||
126 | if (!(acpi_gbl_FADT.gpe1_block_length & 0x1)) | |
127 | acpi_request_region(&acpi_gbl_FADT.xgpe1_block, | |
128 | acpi_gbl_FADT.gpe1_block_length, "ACPI GPE1_BLK"); | |
129 | ||
130 | return 0; | |
131 | } | |
132 | device_initcall(acpi_reserve_resources); | |
133 | ||
134 | acpi_status __init acpi_os_initialize(void) | |
135 | { | |
136 | dmi_check_system(acpi_osl_dmi_table); | |
137 | return AE_OK; | |
138 | } | |
139 | ||
140 | acpi_status acpi_os_initialize1(void) | |
141 | { | |
142 | /* | |
143 | * Initialize PCI configuration space access, as we'll need to access | |
144 | * it while walking the namespace (bus 0 and root bridges w/ _BBNs). | |
145 | */ | |
146 | if (!raw_pci_ops) { | |
147 | printk(KERN_ERR PREFIX | |
148 | "Access to PCI configuration space unavailable\n"); | |
149 | return AE_NULL_ENTRY; | |
150 | } | |
151 | kacpid_wq = create_singlethread_workqueue("kacpid"); | |
152 | kacpi_notify_wq = create_singlethread_workqueue("kacpi_notify"); | |
153 | BUG_ON(!kacpid_wq); | |
154 | BUG_ON(!kacpi_notify_wq); | |
155 | return AE_OK; | |
156 | } | |
157 | ||
158 | acpi_status acpi_os_terminate(void) | |
159 | { | |
160 | if (acpi_irq_handler) { | |
161 | acpi_os_remove_interrupt_handler(acpi_irq_irq, | |
162 | acpi_irq_handler); | |
163 | } | |
164 | ||
165 | destroy_workqueue(kacpid_wq); | |
166 | destroy_workqueue(kacpi_notify_wq); | |
167 | ||
168 | return AE_OK; | |
169 | } | |
170 | ||
171 | void acpi_os_printf(const char *fmt, ...) | |
172 | { | |
173 | va_list args; | |
174 | va_start(args, fmt); | |
175 | acpi_os_vprintf(fmt, args); | |
176 | va_end(args); | |
177 | } | |
178 | ||
179 | EXPORT_SYMBOL(acpi_os_printf); | |
180 | ||
181 | void acpi_os_vprintf(const char *fmt, va_list args) | |
182 | { | |
183 | static char buffer[512]; | |
184 | ||
185 | vsprintf(buffer, fmt, args); | |
186 | ||
187 | #ifdef ENABLE_DEBUGGER | |
188 | if (acpi_in_debugger) { | |
189 | kdb_printf("%s", buffer); | |
190 | } else { | |
191 | printk("%s", buffer); | |
192 | } | |
193 | #else | |
194 | printk("%s", buffer); | |
195 | #endif | |
196 | } | |
197 | ||
198 | acpi_physical_address __init acpi_os_get_root_pointer(void) | |
199 | { | |
200 | if (efi_enabled) { | |
201 | if (efi.acpi20 != EFI_INVALID_TABLE_ADDR) | |
202 | return efi.acpi20; | |
203 | else if (efi.acpi != EFI_INVALID_TABLE_ADDR) | |
204 | return efi.acpi; | |
205 | else { | |
206 | printk(KERN_ERR PREFIX | |
207 | "System description tables not found\n"); | |
208 | return 0; | |
209 | } | |
210 | } else | |
211 | return acpi_find_rsdp(); | |
212 | } | |
213 | ||
214 | void __iomem *acpi_os_map_memory(acpi_physical_address phys, acpi_size size) | |
215 | { | |
216 | if (phys > ULONG_MAX) { | |
217 | printk(KERN_ERR PREFIX "Cannot map memory that high\n"); | |
218 | return NULL; | |
219 | } | |
220 | if (acpi_gbl_permanent_mmap) | |
221 | /* | |
222 | * ioremap checks to ensure this is in reserved space | |
223 | */ | |
224 | return ioremap((unsigned long)phys, size); | |
225 | else | |
226 | return __acpi_map_table((unsigned long)phys, size); | |
227 | } | |
228 | EXPORT_SYMBOL_GPL(acpi_os_map_memory); | |
229 | ||
230 | void acpi_os_unmap_memory(void __iomem * virt, acpi_size size) | |
231 | { | |
232 | if (acpi_gbl_permanent_mmap) { | |
233 | iounmap(virt); | |
234 | } | |
235 | } | |
236 | EXPORT_SYMBOL_GPL(acpi_os_unmap_memory); | |
237 | ||
238 | #ifdef ACPI_FUTURE_USAGE | |
239 | acpi_status | |
240 | acpi_os_get_physical_address(void *virt, acpi_physical_address * phys) | |
241 | { | |
242 | if (!phys || !virt) | |
243 | return AE_BAD_PARAMETER; | |
244 | ||
245 | *phys = virt_to_phys(virt); | |
246 | ||
247 | return AE_OK; | |
248 | } | |
249 | #endif | |
250 | ||
251 | #define ACPI_MAX_OVERRIDE_LEN 100 | |
252 | ||
253 | static char acpi_os_name[ACPI_MAX_OVERRIDE_LEN]; | |
254 | ||
255 | acpi_status | |
256 | acpi_os_predefined_override(const struct acpi_predefined_names *init_val, | |
257 | acpi_string * new_val) | |
258 | { | |
259 | if (!init_val || !new_val) | |
260 | return AE_BAD_PARAMETER; | |
261 | ||
262 | *new_val = NULL; | |
263 | if (!memcmp(init_val->name, "_OS_", 4) && strlen(acpi_os_name)) { | |
264 | printk(KERN_INFO PREFIX "Overriding _OS definition to '%s'\n", | |
265 | acpi_os_name); | |
266 | *new_val = acpi_os_name; | |
267 | } | |
268 | ||
269 | return AE_OK; | |
270 | } | |
271 | ||
272 | acpi_status | |
273 | acpi_os_table_override(struct acpi_table_header * existing_table, | |
274 | struct acpi_table_header ** new_table) | |
275 | { | |
276 | if (!existing_table || !new_table) | |
277 | return AE_BAD_PARAMETER; | |
278 | ||
279 | #ifdef CONFIG_ACPI_CUSTOM_DSDT | |
280 | if (strncmp(existing_table->signature, "DSDT", 4) == 0) | |
281 | *new_table = (struct acpi_table_header *)AmlCode; | |
282 | else | |
283 | *new_table = NULL; | |
284 | #else | |
285 | *new_table = NULL; | |
286 | #endif | |
287 | return AE_OK; | |
288 | } | |
289 | ||
290 | static irqreturn_t acpi_irq(int irq, void *dev_id) | |
291 | { | |
292 | return (*acpi_irq_handler) (acpi_irq_context) ? IRQ_HANDLED : IRQ_NONE; | |
293 | } | |
294 | ||
295 | acpi_status | |
296 | acpi_os_install_interrupt_handler(u32 gsi, acpi_osd_handler handler, | |
297 | void *context) | |
298 | { | |
299 | unsigned int irq; | |
300 | ||
301 | /* | |
302 | * Ignore the GSI from the core, and use the value in our copy of the | |
303 | * FADT. It may not be the same if an interrupt source override exists | |
304 | * for the SCI. | |
305 | */ | |
306 | gsi = acpi_gbl_FADT.sci_interrupt; | |
307 | if (acpi_gsi_to_irq(gsi, &irq) < 0) { | |
308 | printk(KERN_ERR PREFIX "SCI (ACPI GSI %d) not registered\n", | |
309 | gsi); | |
310 | return AE_OK; | |
311 | } | |
312 | ||
313 | acpi_irq_handler = handler; | |
314 | acpi_irq_context = context; | |
315 | if (request_irq(irq, acpi_irq, IRQF_SHARED, "acpi", acpi_irq)) { | |
316 | printk(KERN_ERR PREFIX "SCI (IRQ%d) allocation failed\n", irq); | |
317 | return AE_NOT_ACQUIRED; | |
318 | } | |
319 | acpi_irq_irq = irq; | |
320 | ||
321 | return AE_OK; | |
322 | } | |
323 | ||
324 | acpi_status acpi_os_remove_interrupt_handler(u32 irq, acpi_osd_handler handler) | |
325 | { | |
326 | if (irq) { | |
327 | free_irq(irq, acpi_irq); | |
328 | acpi_irq_handler = NULL; | |
329 | acpi_irq_irq = 0; | |
330 | } | |
331 | ||
332 | return AE_OK; | |
333 | } | |
334 | ||
335 | /* | |
336 | * Running in interpreter thread context, safe to sleep | |
337 | */ | |
338 | ||
339 | void acpi_os_sleep(acpi_integer ms) | |
340 | { | |
341 | schedule_timeout_interruptible(msecs_to_jiffies(ms)); | |
342 | } | |
343 | ||
344 | EXPORT_SYMBOL(acpi_os_sleep); | |
345 | ||
346 | void acpi_os_stall(u32 us) | |
347 | { | |
348 | while (us) { | |
349 | u32 delay = 1000; | |
350 | ||
351 | if (delay > us) | |
352 | delay = us; | |
353 | udelay(delay); | |
354 | touch_nmi_watchdog(); | |
355 | us -= delay; | |
356 | } | |
357 | } | |
358 | ||
359 | EXPORT_SYMBOL(acpi_os_stall); | |
360 | ||
361 | /* | |
362 | * Support ACPI 3.0 AML Timer operand | |
363 | * Returns 64-bit free-running, monotonically increasing timer | |
364 | * with 100ns granularity | |
365 | */ | |
366 | u64 acpi_os_get_timer(void) | |
367 | { | |
368 | static u64 t; | |
369 | ||
370 | #ifdef CONFIG_HPET | |
371 | /* TBD: use HPET if available */ | |
372 | #endif | |
373 | ||
374 | #ifdef CONFIG_X86_PM_TIMER | |
375 | /* TBD: default to PM timer if HPET was not available */ | |
376 | #endif | |
377 | if (!t) | |
378 | printk(KERN_ERR PREFIX "acpi_os_get_timer() TBD\n"); | |
379 | ||
380 | return ++t; | |
381 | } | |
382 | ||
383 | acpi_status acpi_os_read_port(acpi_io_address port, u32 * value, u32 width) | |
384 | { | |
385 | u32 dummy; | |
386 | ||
387 | if (!value) | |
388 | value = &dummy; | |
389 | ||
390 | *value = 0; | |
391 | if (width <= 8) { | |
392 | *(u8 *) value = inb(port); | |
393 | } else if (width <= 16) { | |
394 | *(u16 *) value = inw(port); | |
395 | } else if (width <= 32) { | |
396 | *(u32 *) value = inl(port); | |
397 | } else { | |
398 | BUG(); | |
399 | } | |
400 | ||
401 | return AE_OK; | |
402 | } | |
403 | ||
404 | EXPORT_SYMBOL(acpi_os_read_port); | |
405 | ||
406 | acpi_status acpi_os_write_port(acpi_io_address port, u32 value, u32 width) | |
407 | { | |
408 | if (width <= 8) { | |
409 | outb(value, port); | |
410 | } else if (width <= 16) { | |
411 | outw(value, port); | |
412 | } else if (width <= 32) { | |
413 | outl(value, port); | |
414 | } else { | |
415 | BUG(); | |
416 | } | |
417 | ||
418 | return AE_OK; | |
419 | } | |
420 | ||
421 | EXPORT_SYMBOL(acpi_os_write_port); | |
422 | ||
423 | acpi_status | |
424 | acpi_os_read_memory(acpi_physical_address phys_addr, u32 * value, u32 width) | |
425 | { | |
426 | u32 dummy; | |
427 | void __iomem *virt_addr; | |
428 | ||
429 | virt_addr = ioremap(phys_addr, width); | |
430 | if (!value) | |
431 | value = &dummy; | |
432 | ||
433 | switch (width) { | |
434 | case 8: | |
435 | *(u8 *) value = readb(virt_addr); | |
436 | break; | |
437 | case 16: | |
438 | *(u16 *) value = readw(virt_addr); | |
439 | break; | |
440 | case 32: | |
441 | *(u32 *) value = readl(virt_addr); | |
442 | break; | |
443 | default: | |
444 | BUG(); | |
445 | } | |
446 | ||
447 | iounmap(virt_addr); | |
448 | ||
449 | return AE_OK; | |
450 | } | |
451 | ||
452 | acpi_status | |
453 | acpi_os_write_memory(acpi_physical_address phys_addr, u32 value, u32 width) | |
454 | { | |
455 | void __iomem *virt_addr; | |
456 | ||
457 | virt_addr = ioremap(phys_addr, width); | |
458 | ||
459 | switch (width) { | |
460 | case 8: | |
461 | writeb(value, virt_addr); | |
462 | break; | |
463 | case 16: | |
464 | writew(value, virt_addr); | |
465 | break; | |
466 | case 32: | |
467 | writel(value, virt_addr); | |
468 | break; | |
469 | default: | |
470 | BUG(); | |
471 | } | |
472 | ||
473 | iounmap(virt_addr); | |
474 | ||
475 | return AE_OK; | |
476 | } | |
477 | ||
478 | acpi_status | |
479 | acpi_os_read_pci_configuration(struct acpi_pci_id * pci_id, u32 reg, | |
480 | void *value, u32 width) | |
481 | { | |
482 | int result, size; | |
483 | ||
484 | if (!value) | |
485 | return AE_BAD_PARAMETER; | |
486 | ||
487 | switch (width) { | |
488 | case 8: | |
489 | size = 1; | |
490 | break; | |
491 | case 16: | |
492 | size = 2; | |
493 | break; | |
494 | case 32: | |
495 | size = 4; | |
496 | break; | |
497 | default: | |
498 | return AE_ERROR; | |
499 | } | |
500 | ||
501 | BUG_ON(!raw_pci_ops); | |
502 | ||
503 | result = raw_pci_ops->read(pci_id->segment, pci_id->bus, | |
504 | PCI_DEVFN(pci_id->device, pci_id->function), | |
505 | reg, size, value); | |
506 | ||
507 | return (result ? AE_ERROR : AE_OK); | |
508 | } | |
509 | ||
510 | EXPORT_SYMBOL(acpi_os_read_pci_configuration); | |
511 | ||
512 | acpi_status | |
513 | acpi_os_write_pci_configuration(struct acpi_pci_id * pci_id, u32 reg, | |
514 | acpi_integer value, u32 width) | |
515 | { | |
516 | int result, size; | |
517 | ||
518 | switch (width) { | |
519 | case 8: | |
520 | size = 1; | |
521 | break; | |
522 | case 16: | |
523 | size = 2; | |
524 | break; | |
525 | case 32: | |
526 | size = 4; | |
527 | break; | |
528 | default: | |
529 | return AE_ERROR; | |
530 | } | |
531 | ||
532 | BUG_ON(!raw_pci_ops); | |
533 | ||
534 | result = raw_pci_ops->write(pci_id->segment, pci_id->bus, | |
535 | PCI_DEVFN(pci_id->device, pci_id->function), | |
536 | reg, size, value); | |
537 | ||
538 | return (result ? AE_ERROR : AE_OK); | |
539 | } | |
540 | ||
541 | /* TODO: Change code to take advantage of driver model more */ | |
542 | static void acpi_os_derive_pci_id_2(acpi_handle rhandle, /* upper bound */ | |
543 | acpi_handle chandle, /* current node */ | |
544 | struct acpi_pci_id **id, | |
545 | int *is_bridge, u8 * bus_number) | |
546 | { | |
547 | acpi_handle handle; | |
548 | struct acpi_pci_id *pci_id = *id; | |
549 | acpi_status status; | |
550 | unsigned long temp; | |
551 | acpi_object_type type; | |
552 | u8 tu8; | |
553 | ||
554 | acpi_get_parent(chandle, &handle); | |
555 | if (handle != rhandle) { | |
556 | acpi_os_derive_pci_id_2(rhandle, handle, &pci_id, is_bridge, | |
557 | bus_number); | |
558 | ||
559 | status = acpi_get_type(handle, &type); | |
560 | if ((ACPI_FAILURE(status)) || (type != ACPI_TYPE_DEVICE)) | |
561 | return; | |
562 | ||
563 | status = | |
564 | acpi_evaluate_integer(handle, METHOD_NAME__ADR, NULL, | |
565 | &temp); | |
566 | if (ACPI_SUCCESS(status)) { | |
567 | pci_id->device = ACPI_HIWORD(ACPI_LODWORD(temp)); | |
568 | pci_id->function = ACPI_LOWORD(ACPI_LODWORD(temp)); | |
569 | ||
570 | if (*is_bridge) | |
571 | pci_id->bus = *bus_number; | |
572 | ||
573 | /* any nicer way to get bus number of bridge ? */ | |
574 | status = | |
575 | acpi_os_read_pci_configuration(pci_id, 0x0e, &tu8, | |
576 | 8); | |
577 | if (ACPI_SUCCESS(status) | |
578 | && ((tu8 & 0x7f) == 1 || (tu8 & 0x7f) == 2)) { | |
579 | status = | |
580 | acpi_os_read_pci_configuration(pci_id, 0x18, | |
581 | &tu8, 8); | |
582 | if (!ACPI_SUCCESS(status)) { | |
583 | /* Certainly broken... FIX ME */ | |
584 | return; | |
585 | } | |
586 | *is_bridge = 1; | |
587 | pci_id->bus = tu8; | |
588 | status = | |
589 | acpi_os_read_pci_configuration(pci_id, 0x19, | |
590 | &tu8, 8); | |
591 | if (ACPI_SUCCESS(status)) { | |
592 | *bus_number = tu8; | |
593 | } | |
594 | } else | |
595 | *is_bridge = 0; | |
596 | } | |
597 | } | |
598 | } | |
599 | ||
600 | void acpi_os_derive_pci_id(acpi_handle rhandle, /* upper bound */ | |
601 | acpi_handle chandle, /* current node */ | |
602 | struct acpi_pci_id **id) | |
603 | { | |
604 | int is_bridge = 1; | |
605 | u8 bus_number = (*id)->bus; | |
606 | ||
607 | acpi_os_derive_pci_id_2(rhandle, chandle, id, &is_bridge, &bus_number); | |
608 | } | |
609 | ||
610 | static void acpi_os_execute_deferred(struct work_struct *work) | |
611 | { | |
612 | struct acpi_os_dpc *dpc = container_of(work, struct acpi_os_dpc, work); | |
613 | if (!dpc) { | |
614 | printk(KERN_ERR PREFIX "Invalid (NULL) context\n"); | |
615 | return; | |
616 | } | |
617 | ||
618 | dpc->function(dpc->context); | |
619 | kfree(dpc); | |
620 | ||
621 | /* Yield cpu to notify thread */ | |
622 | cond_resched(); | |
623 | ||
624 | return; | |
625 | } | |
626 | ||
627 | static void acpi_os_execute_notify(struct work_struct *work) | |
628 | { | |
629 | struct acpi_os_dpc *dpc = container_of(work, struct acpi_os_dpc, work); | |
630 | ||
631 | if (!dpc) { | |
632 | printk(KERN_ERR PREFIX "Invalid (NULL) context\n"); | |
633 | return; | |
634 | } | |
635 | ||
636 | dpc->function(dpc->context); | |
637 | ||
638 | kfree(dpc); | |
639 | ||
640 | return; | |
641 | } | |
642 | ||
643 | /******************************************************************************* | |
644 | * | |
645 | * FUNCTION: acpi_os_execute | |
646 | * | |
647 | * PARAMETERS: Type - Type of the callback | |
648 | * Function - Function to be executed | |
649 | * Context - Function parameters | |
650 | * | |
651 | * RETURN: Status | |
652 | * | |
653 | * DESCRIPTION: Depending on type, either queues function for deferred execution or | |
654 | * immediately executes function on a separate thread. | |
655 | * | |
656 | ******************************************************************************/ | |
657 | ||
658 | acpi_status acpi_os_execute(acpi_execute_type type, | |
659 | acpi_osd_exec_callback function, void *context) | |
660 | { | |
661 | acpi_status status = AE_OK; | |
662 | struct acpi_os_dpc *dpc; | |
663 | ||
664 | ACPI_DEBUG_PRINT((ACPI_DB_EXEC, | |
665 | "Scheduling function [%p(%p)] for deferred execution.\n", | |
666 | function, context)); | |
667 | ||
668 | if (!function) | |
669 | return AE_BAD_PARAMETER; | |
670 | ||
671 | /* | |
672 | * Allocate/initialize DPC structure. Note that this memory will be | |
673 | * freed by the callee. The kernel handles the work_struct list in a | |
674 | * way that allows us to also free its memory inside the callee. | |
675 | * Because we may want to schedule several tasks with different | |
676 | * parameters we can't use the approach some kernel code uses of | |
677 | * having a static work_struct. | |
678 | */ | |
679 | ||
680 | dpc = kmalloc(sizeof(struct acpi_os_dpc), GFP_ATOMIC); | |
681 | if (!dpc) | |
682 | return_ACPI_STATUS(AE_NO_MEMORY); | |
683 | ||
684 | dpc->function = function; | |
685 | dpc->context = context; | |
686 | ||
687 | if (type == OSL_NOTIFY_HANDLER) { | |
688 | INIT_WORK(&dpc->work, acpi_os_execute_notify); | |
689 | if (!queue_work(kacpi_notify_wq, &dpc->work)) { | |
690 | status = AE_ERROR; | |
691 | kfree(dpc); | |
692 | } | |
693 | } else { | |
694 | INIT_WORK(&dpc->work, acpi_os_execute_deferred); | |
695 | if (!queue_work(kacpid_wq, &dpc->work)) { | |
696 | ACPI_DEBUG_PRINT((ACPI_DB_ERROR, | |
697 | "Call to queue_work() failed.\n")); | |
698 | status = AE_ERROR; | |
699 | kfree(dpc); | |
700 | } | |
701 | } | |
702 | return_ACPI_STATUS(status); | |
703 | } | |
704 | ||
705 | EXPORT_SYMBOL(acpi_os_execute); | |
706 | ||
707 | void acpi_os_wait_events_complete(void *context) | |
708 | { | |
709 | flush_workqueue(kacpid_wq); | |
710 | } | |
711 | ||
712 | EXPORT_SYMBOL(acpi_os_wait_events_complete); | |
713 | ||
714 | /* | |
715 | * Allocate the memory for a spinlock and initialize it. | |
716 | */ | |
717 | acpi_status acpi_os_create_lock(acpi_spinlock * handle) | |
718 | { | |
719 | spin_lock_init(*handle); | |
720 | ||
721 | return AE_OK; | |
722 | } | |
723 | ||
724 | /* | |
725 | * Deallocate the memory for a spinlock. | |
726 | */ | |
727 | void acpi_os_delete_lock(acpi_spinlock handle) | |
728 | { | |
729 | return; | |
730 | } | |
731 | ||
732 | acpi_status | |
733 | acpi_os_create_semaphore(u32 max_units, u32 initial_units, acpi_handle * handle) | |
734 | { | |
735 | struct semaphore *sem = NULL; | |
736 | ||
737 | ||
738 | sem = acpi_os_allocate(sizeof(struct semaphore)); | |
739 | if (!sem) | |
740 | return AE_NO_MEMORY; | |
741 | memset(sem, 0, sizeof(struct semaphore)); | |
742 | ||
743 | sema_init(sem, initial_units); | |
744 | ||
745 | *handle = (acpi_handle *) sem; | |
746 | ||
747 | ACPI_DEBUG_PRINT((ACPI_DB_MUTEX, "Creating semaphore[%p|%d].\n", | |
748 | *handle, initial_units)); | |
749 | ||
750 | return AE_OK; | |
751 | } | |
752 | ||
753 | EXPORT_SYMBOL(acpi_os_create_semaphore); | |
754 | ||
755 | /* | |
756 | * TODO: A better way to delete semaphores? Linux doesn't have a | |
757 | * 'delete_semaphore()' function -- may result in an invalid | |
758 | * pointer dereference for non-synchronized consumers. Should | |
759 | * we at least check for blocked threads and signal/cancel them? | |
760 | */ | |
761 | ||
762 | acpi_status acpi_os_delete_semaphore(acpi_handle handle) | |
763 | { | |
764 | struct semaphore *sem = (struct semaphore *)handle; | |
765 | ||
766 | ||
767 | if (!sem) | |
768 | return AE_BAD_PARAMETER; | |
769 | ||
770 | ACPI_DEBUG_PRINT((ACPI_DB_MUTEX, "Deleting semaphore[%p].\n", handle)); | |
771 | ||
772 | kfree(sem); | |
773 | sem = NULL; | |
774 | ||
775 | return AE_OK; | |
776 | } | |
777 | ||
778 | EXPORT_SYMBOL(acpi_os_delete_semaphore); | |
779 | ||
780 | /* | |
781 | * TODO: The kernel doesn't have a 'down_timeout' function -- had to | |
782 | * improvise. The process is to sleep for one scheduler quantum | |
783 | * until the semaphore becomes available. Downside is that this | |
784 | * may result in starvation for timeout-based waits when there's | |
785 | * lots of semaphore activity. | |
786 | * | |
787 | * TODO: Support for units > 1? | |
788 | */ | |
789 | acpi_status acpi_os_wait_semaphore(acpi_handle handle, u32 units, u16 timeout) | |
790 | { | |
791 | acpi_status status = AE_OK; | |
792 | struct semaphore *sem = (struct semaphore *)handle; | |
793 | int ret = 0; | |
794 | ||
795 | ||
796 | if (!sem || (units < 1)) | |
797 | return AE_BAD_PARAMETER; | |
798 | ||
799 | if (units > 1) | |
800 | return AE_SUPPORT; | |
801 | ||
802 | ACPI_DEBUG_PRINT((ACPI_DB_MUTEX, "Waiting for semaphore[%p|%d|%d]\n", | |
803 | handle, units, timeout)); | |
804 | ||
805 | /* | |
806 | * This can be called during resume with interrupts off. | |
807 | * Like boot-time, we should be single threaded and will | |
808 | * always get the lock if we try -- timeout or not. | |
809 | * If this doesn't succeed, then we will oops courtesy of | |
810 | * might_sleep() in down(). | |
811 | */ | |
812 | if (!down_trylock(sem)) | |
813 | return AE_OK; | |
814 | ||
815 | switch (timeout) { | |
816 | /* | |
817 | * No Wait: | |
818 | * -------- | |
819 | * A zero timeout value indicates that we shouldn't wait - just | |
820 | * acquire the semaphore if available otherwise return AE_TIME | |
821 | * (a.k.a. 'would block'). | |
822 | */ | |
823 | case 0: | |
824 | if (down_trylock(sem)) | |
825 | status = AE_TIME; | |
826 | break; | |
827 | ||
828 | /* | |
829 | * Wait Indefinitely: | |
830 | * ------------------ | |
831 | */ | |
832 | case ACPI_WAIT_FOREVER: | |
833 | down(sem); | |
834 | break; | |
835 | ||
836 | /* | |
837 | * Wait w/ Timeout: | |
838 | * ---------------- | |
839 | */ | |
840 | default: | |
841 | // TODO: A better timeout algorithm? | |
842 | { | |
843 | int i = 0; | |
844 | static const int quantum_ms = 1000 / HZ; | |
845 | ||
846 | ret = down_trylock(sem); | |
847 | for (i = timeout; (i > 0 && ret != 0); i -= quantum_ms) { | |
848 | schedule_timeout_interruptible(1); | |
849 | ret = down_trylock(sem); | |
850 | } | |
851 | ||
852 | if (ret != 0) | |
853 | status = AE_TIME; | |
854 | } | |
855 | break; | |
856 | } | |
857 | ||
858 | if (ACPI_FAILURE(status)) { | |
859 | ACPI_DEBUG_PRINT((ACPI_DB_MUTEX, | |
860 | "Failed to acquire semaphore[%p|%d|%d], %s", | |
861 | handle, units, timeout, | |
862 | acpi_format_exception(status))); | |
863 | } else { | |
864 | ACPI_DEBUG_PRINT((ACPI_DB_MUTEX, | |
865 | "Acquired semaphore[%p|%d|%d]", handle, | |
866 | units, timeout)); | |
867 | } | |
868 | ||
869 | return status; | |
870 | } | |
871 | ||
872 | EXPORT_SYMBOL(acpi_os_wait_semaphore); | |
873 | ||
874 | /* | |
875 | * TODO: Support for units > 1? | |
876 | */ | |
877 | acpi_status acpi_os_signal_semaphore(acpi_handle handle, u32 units) | |
878 | { | |
879 | struct semaphore *sem = (struct semaphore *)handle; | |
880 | ||
881 | ||
882 | if (!sem || (units < 1)) | |
883 | return AE_BAD_PARAMETER; | |
884 | ||
885 | if (units > 1) | |
886 | return AE_SUPPORT; | |
887 | ||
888 | ACPI_DEBUG_PRINT((ACPI_DB_MUTEX, "Signaling semaphore[%p|%d]\n", handle, | |
889 | units)); | |
890 | ||
891 | up(sem); | |
892 | ||
893 | return AE_OK; | |
894 | } | |
895 | ||
896 | EXPORT_SYMBOL(acpi_os_signal_semaphore); | |
897 | ||
898 | #ifdef ACPI_FUTURE_USAGE | |
899 | u32 acpi_os_get_line(char *buffer) | |
900 | { | |
901 | ||
902 | #ifdef ENABLE_DEBUGGER | |
903 | if (acpi_in_debugger) { | |
904 | u32 chars; | |
905 | ||
906 | kdb_read(buffer, sizeof(line_buf)); | |
907 | ||
908 | /* remove the CR kdb includes */ | |
909 | chars = strlen(buffer) - 1; | |
910 | buffer[chars] = '\0'; | |
911 | } | |
912 | #endif | |
913 | ||
914 | return 0; | |
915 | } | |
916 | #endif /* ACPI_FUTURE_USAGE */ | |
917 | ||
918 | acpi_status acpi_os_signal(u32 function, void *info) | |
919 | { | |
920 | switch (function) { | |
921 | case ACPI_SIGNAL_FATAL: | |
922 | printk(KERN_ERR PREFIX "Fatal opcode executed\n"); | |
923 | break; | |
924 | case ACPI_SIGNAL_BREAKPOINT: | |
925 | /* | |
926 | * AML Breakpoint | |
927 | * ACPI spec. says to treat it as a NOP unless | |
928 | * you are debugging. So if/when we integrate | |
929 | * AML debugger into the kernel debugger its | |
930 | * hook will go here. But until then it is | |
931 | * not useful to print anything on breakpoints. | |
932 | */ | |
933 | break; | |
934 | default: | |
935 | break; | |
936 | } | |
937 | ||
938 | return AE_OK; | |
939 | } | |
940 | ||
941 | EXPORT_SYMBOL(acpi_os_signal); | |
942 | ||
943 | static int __init acpi_os_name_setup(char *str) | |
944 | { | |
945 | char *p = acpi_os_name; | |
946 | int count = ACPI_MAX_OVERRIDE_LEN - 1; | |
947 | ||
948 | if (!str || !*str) | |
949 | return 0; | |
950 | ||
951 | for (; count-- && str && *str; str++) { | |
952 | if (isalnum(*str) || *str == ' ' || *str == ':') | |
953 | *p++ = *str; | |
954 | else if (*str == '\'' || *str == '"') | |
955 | continue; | |
956 | else | |
957 | break; | |
958 | } | |
959 | *p = 0; | |
960 | ||
961 | return 1; | |
962 | ||
963 | } | |
964 | ||
965 | __setup("acpi_os_name=", acpi_os_name_setup); | |
966 | ||
967 | static void enable_osi_linux(int enable) { | |
968 | ||
969 | if (osi_linux != enable) | |
970 | printk(KERN_INFO PREFIX "%sabled _OSI(Linux)\n", | |
971 | enable ? "En": "Dis"); | |
972 | ||
973 | osi_linux = enable; | |
974 | return; | |
975 | } | |
976 | ||
977 | /* | |
978 | * Modify the list of "OS Interfaces" reported to BIOS via _OSI | |
979 | * | |
980 | * empty string disables _OSI | |
981 | * string starting with '!' disables that string | |
982 | * otherwise string is added to list, augmenting built-in strings | |
983 | */ | |
984 | static int __init acpi_osi_setup(char *str) | |
985 | { | |
986 | if (str == NULL || *str == '\0') { | |
987 | printk(KERN_INFO PREFIX "_OSI method disabled\n"); | |
988 | acpi_gbl_create_osi_method = FALSE; | |
989 | } else if (!strcmp("!Linux", str)) { | |
990 | enable_osi_linux(0); | |
991 | } else if (*str == '!') { | |
992 | if (acpi_osi_invalidate(++str) == AE_OK) | |
993 | printk(KERN_INFO PREFIX "Deleted _OSI(%s)\n", str); | |
994 | } else if (!strcmp("Linux", str)) { | |
995 | enable_osi_linux(1); | |
996 | } else if (*osi_additional_string == '\0') { | |
997 | strncpy(osi_additional_string, str, OSI_STRING_LENGTH_MAX); | |
998 | printk(KERN_INFO PREFIX "Added _OSI(%s)\n", str); | |
999 | } | |
1000 | ||
1001 | return 1; | |
1002 | } | |
1003 | ||
1004 | __setup("acpi_osi=", acpi_osi_setup); | |
1005 | ||
1006 | /* enable serialization to combat AE_ALREADY_EXISTS errors */ | |
1007 | static int __init acpi_serialize_setup(char *str) | |
1008 | { | |
1009 | printk(KERN_INFO PREFIX "serialize enabled\n"); | |
1010 | ||
1011 | acpi_gbl_all_methods_serialized = TRUE; | |
1012 | ||
1013 | return 1; | |
1014 | } | |
1015 | ||
1016 | __setup("acpi_serialize", acpi_serialize_setup); | |
1017 | ||
1018 | /* | |
1019 | * Wake and Run-Time GPES are expected to be separate. | |
1020 | * We disable wake-GPEs at run-time to prevent spurious | |
1021 | * interrupts. | |
1022 | * | |
1023 | * However, if a system exists that shares Wake and | |
1024 | * Run-time events on the same GPE this flag is available | |
1025 | * to tell Linux to keep the wake-time GPEs enabled at run-time. | |
1026 | */ | |
1027 | static int __init acpi_wake_gpes_always_on_setup(char *str) | |
1028 | { | |
1029 | printk(KERN_INFO PREFIX "wake GPEs not disabled\n"); | |
1030 | ||
1031 | acpi_gbl_leave_wake_gpes_disabled = FALSE; | |
1032 | ||
1033 | return 1; | |
1034 | } | |
1035 | ||
1036 | __setup("acpi_wake_gpes_always_on", acpi_wake_gpes_always_on_setup); | |
1037 | ||
1038 | /* | |
1039 | * Acquire a spinlock. | |
1040 | * | |
1041 | * handle is a pointer to the spinlock_t. | |
1042 | */ | |
1043 | ||
1044 | acpi_cpu_flags acpi_os_acquire_lock(acpi_spinlock lockp) | |
1045 | { | |
1046 | acpi_cpu_flags flags; | |
1047 | spin_lock_irqsave(lockp, flags); | |
1048 | return flags; | |
1049 | } | |
1050 | ||
1051 | /* | |
1052 | * Release a spinlock. See above. | |
1053 | */ | |
1054 | ||
1055 | void acpi_os_release_lock(acpi_spinlock lockp, acpi_cpu_flags flags) | |
1056 | { | |
1057 | spin_unlock_irqrestore(lockp, flags); | |
1058 | } | |
1059 | ||
1060 | #ifndef ACPI_USE_LOCAL_CACHE | |
1061 | ||
1062 | /******************************************************************************* | |
1063 | * | |
1064 | * FUNCTION: acpi_os_create_cache | |
1065 | * | |
1066 | * PARAMETERS: name - Ascii name for the cache | |
1067 | * size - Size of each cached object | |
1068 | * depth - Maximum depth of the cache (in objects) <ignored> | |
1069 | * cache - Where the new cache object is returned | |
1070 | * | |
1071 | * RETURN: status | |
1072 | * | |
1073 | * DESCRIPTION: Create a cache object | |
1074 | * | |
1075 | ******************************************************************************/ | |
1076 | ||
1077 | acpi_status | |
1078 | acpi_os_create_cache(char *name, u16 size, u16 depth, acpi_cache_t ** cache) | |
1079 | { | |
1080 | *cache = kmem_cache_create(name, size, 0, 0, NULL); | |
1081 | if (*cache == NULL) | |
1082 | return AE_ERROR; | |
1083 | else | |
1084 | return AE_OK; | |
1085 | } | |
1086 | ||
1087 | /******************************************************************************* | |
1088 | * | |
1089 | * FUNCTION: acpi_os_purge_cache | |
1090 | * | |
1091 | * PARAMETERS: Cache - Handle to cache object | |
1092 | * | |
1093 | * RETURN: Status | |
1094 | * | |
1095 | * DESCRIPTION: Free all objects within the requested cache. | |
1096 | * | |
1097 | ******************************************************************************/ | |
1098 | ||
1099 | acpi_status acpi_os_purge_cache(acpi_cache_t * cache) | |
1100 | { | |
1101 | kmem_cache_shrink(cache); | |
1102 | return (AE_OK); | |
1103 | } | |
1104 | ||
1105 | /******************************************************************************* | |
1106 | * | |
1107 | * FUNCTION: acpi_os_delete_cache | |
1108 | * | |
1109 | * PARAMETERS: Cache - Handle to cache object | |
1110 | * | |
1111 | * RETURN: Status | |
1112 | * | |
1113 | * DESCRIPTION: Free all objects within the requested cache and delete the | |
1114 | * cache object. | |
1115 | * | |
1116 | ******************************************************************************/ | |
1117 | ||
1118 | acpi_status acpi_os_delete_cache(acpi_cache_t * cache) | |
1119 | { | |
1120 | kmem_cache_destroy(cache); | |
1121 | return (AE_OK); | |
1122 | } | |
1123 | ||
1124 | /******************************************************************************* | |
1125 | * | |
1126 | * FUNCTION: acpi_os_release_object | |
1127 | * | |
1128 | * PARAMETERS: Cache - Handle to cache object | |
1129 | * Object - The object to be released | |
1130 | * | |
1131 | * RETURN: None | |
1132 | * | |
1133 | * DESCRIPTION: Release an object to the specified cache. If cache is full, | |
1134 | * the object is deleted. | |
1135 | * | |
1136 | ******************************************************************************/ | |
1137 | ||
1138 | acpi_status acpi_os_release_object(acpi_cache_t * cache, void *object) | |
1139 | { | |
1140 | kmem_cache_free(cache, object); | |
1141 | return (AE_OK); | |
1142 | } | |
1143 | ||
1144 | /****************************************************************************** | |
1145 | * | |
1146 | * FUNCTION: acpi_os_validate_interface | |
1147 | * | |
1148 | * PARAMETERS: interface - Requested interface to be validated | |
1149 | * | |
1150 | * RETURN: AE_OK if interface is supported, AE_SUPPORT otherwise | |
1151 | * | |
1152 | * DESCRIPTION: Match an interface string to the interfaces supported by the | |
1153 | * host. Strings originate from an AML call to the _OSI method. | |
1154 | * | |
1155 | *****************************************************************************/ | |
1156 | ||
1157 | acpi_status | |
1158 | acpi_os_validate_interface (char *interface) | |
1159 | { | |
1160 | if (!strncmp(osi_additional_string, interface, OSI_STRING_LENGTH_MAX)) | |
1161 | return AE_OK; | |
1162 | if (!strcmp("Linux", interface)) { | |
1163 | printk(KERN_WARNING PREFIX | |
1164 | "System BIOS is requesting _OSI(Linux)\n"); | |
1165 | printk(KERN_WARNING PREFIX | |
1166 | "If \"acpi_osi=Linux\" works better,\n" | |
1167 | "Please send dmidecode " | |
1168 | "to linux-acpi@vger.kernel.org\n"); | |
1169 | if(osi_linux) | |
1170 | return AE_OK; | |
1171 | } | |
1172 | return AE_SUPPORT; | |
1173 | } | |
1174 | ||
1175 | /****************************************************************************** | |
1176 | * | |
1177 | * FUNCTION: acpi_os_validate_address | |
1178 | * | |
1179 | * PARAMETERS: space_id - ACPI space ID | |
1180 | * address - Physical address | |
1181 | * length - Address length | |
1182 | * | |
1183 | * RETURN: AE_OK if address/length is valid for the space_id. Otherwise, | |
1184 | * should return AE_AML_ILLEGAL_ADDRESS. | |
1185 | * | |
1186 | * DESCRIPTION: Validate a system address via the host OS. Used to validate | |
1187 | * the addresses accessed by AML operation regions. | |
1188 | * | |
1189 | *****************************************************************************/ | |
1190 | ||
1191 | acpi_status | |
1192 | acpi_os_validate_address ( | |
1193 | u8 space_id, | |
1194 | acpi_physical_address address, | |
1195 | acpi_size length) | |
1196 | { | |
1197 | ||
1198 | return AE_OK; | |
1199 | } | |
1200 | ||
1201 | #ifdef CONFIG_DMI | |
1202 | static int dmi_osi_linux(const struct dmi_system_id *d) | |
1203 | { | |
1204 | printk(KERN_NOTICE "%s detected: enabling _OSI(Linux)\n", d->ident); | |
1205 | enable_osi_linux(1); | |
1206 | return 0; | |
1207 | } | |
1208 | ||
1209 | static struct dmi_system_id acpi_osl_dmi_table[] __initdata = { | |
1210 | /* | |
1211 | * Boxes that need _OSI(Linux) | |
1212 | */ | |
1213 | { | |
1214 | .callback = dmi_osi_linux, | |
1215 | .ident = "Intel Napa CRB", | |
1216 | .matches = { | |
1217 | DMI_MATCH(DMI_BOARD_VENDOR, "Intel Corporation"), | |
1218 | DMI_MATCH(DMI_BOARD_NAME, "MPAD-MSAE Customer Reference Boards"), | |
1219 | }, | |
1220 | }, | |
1221 | {} | |
1222 | }; | |
1223 | #endif /* CONFIG_DMI */ | |
1224 | ||
1225 | #endif |