2 * PowerNV OPAL high level interfaces
4 * Copyright 2011 IBM Corp.
6 * This program is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License
8 * as published by the Free Software Foundation; either version
9 * 2 of the License, or (at your option) any later version.
12 #define pr_fmt(fmt) "opal: " fmt
14 #include <linux/printk.h>
15 #include <linux/types.h>
17 #include <linux/of_fdt.h>
18 #include <linux/of_platform.h>
19 #include <linux/interrupt.h>
20 #include <linux/notifier.h>
21 #include <linux/slab.h>
22 #include <linux/sched.h>
23 #include <linux/kobject.h>
24 #include <linux/delay.h>
25 #include <linux/memblock.h>
26 #include <linux/kthread.h>
27 #include <linux/freezer.h>
29 #include <asm/machdep.h>
31 #include <asm/firmware.h>
36 /* /sys/firmware/opal */
37 struct kobject
*opal_kobj
;
45 struct mcheck_recoverable_range
{
51 static struct mcheck_recoverable_range
*mc_recoverable_range
;
52 static int mc_recoverable_range_len
;
54 struct device_node
*opal_node
;
55 static DEFINE_SPINLOCK(opal_write_lock
);
56 static struct atomic_notifier_head opal_msg_notifier_head
[OPAL_MSG_TYPE_MAX
];
57 static uint32_t opal_heartbeat
;
58 static struct task_struct
*kopald_tsk
;
60 void opal_configure_cores(void)
62 /* Do the actual re-init, This will clobber all FPRs, VRs, etc...
64 * It will preserve non volatile GPRs and HSPRG0/1. It will
65 * also restore HIDs and other SPRs to their original value
66 * but it might clobber a bunch.
69 opal_reinit_cpus(OPAL_REINIT_CPUS_HILE_BE
);
71 opal_reinit_cpus(OPAL_REINIT_CPUS_HILE_LE
);
74 /* Restore some bits */
75 if (cur_cpu_spec
->cpu_restore
)
76 cur_cpu_spec
->cpu_restore();
79 int __init
early_init_dt_scan_opal(unsigned long node
,
80 const char *uname
, int depth
, void *data
)
82 const void *basep
, *entryp
, *sizep
;
83 int basesz
, entrysz
, runtimesz
;
85 if (depth
!= 1 || strcmp(uname
, "ibm,opal") != 0)
88 basep
= of_get_flat_dt_prop(node
, "opal-base-address", &basesz
);
89 entryp
= of_get_flat_dt_prop(node
, "opal-entry-address", &entrysz
);
90 sizep
= of_get_flat_dt_prop(node
, "opal-runtime-size", &runtimesz
);
92 if (!basep
|| !entryp
|| !sizep
)
95 opal
.base
= of_read_number(basep
, basesz
/4);
96 opal
.entry
= of_read_number(entryp
, entrysz
/4);
97 opal
.size
= of_read_number(sizep
, runtimesz
/4);
99 pr_debug("OPAL Base = 0x%llx (basep=%p basesz=%d)\n",
100 opal
.base
, basep
, basesz
);
101 pr_debug("OPAL Entry = 0x%llx (entryp=%p basesz=%d)\n",
102 opal
.entry
, entryp
, entrysz
);
103 pr_debug("OPAL Entry = 0x%llx (sizep=%p runtimesz=%d)\n",
104 opal
.size
, sizep
, runtimesz
);
106 if (of_flat_dt_is_compatible(node
, "ibm,opal-v3")) {
107 powerpc_firmware_features
|= FW_FEATURE_OPAL
;
108 pr_info("OPAL detected !\n");
110 panic("OPAL != V3 detected, no longer supported.\n");
116 int __init
early_init_dt_scan_recoverable_ranges(unsigned long node
,
117 const char *uname
, int depth
, void *data
)
122 if (depth
!= 1 || strcmp(uname
, "ibm,opal") != 0)
125 prop
= of_get_flat_dt_prop(node
, "mcheck-recoverable-ranges", &psize
);
130 pr_debug("Found machine check recoverable ranges.\n");
133 * Calculate number of available entries.
135 * Each recoverable address range entry is (start address, len,
136 * recovery address), 2 cells each for start and recovery address,
137 * 1 cell for len, totalling 5 cells per entry.
139 mc_recoverable_range_len
= psize
/ (sizeof(*prop
) * 5);
142 if (!mc_recoverable_range_len
)
145 /* Size required to hold all the entries. */
146 size
= mc_recoverable_range_len
*
147 sizeof(struct mcheck_recoverable_range
);
150 * Allocate a buffer to hold the MC recoverable ranges. We would be
151 * accessing them in real mode, hence it needs to be within
154 mc_recoverable_range
=__va(memblock_alloc_base(size
, __alignof__(u64
),
156 memset(mc_recoverable_range
, 0, size
);
158 for (i
= 0; i
< mc_recoverable_range_len
; i
++) {
159 mc_recoverable_range
[i
].start_addr
=
160 of_read_number(prop
+ (i
* 5) + 0, 2);
161 mc_recoverable_range
[i
].end_addr
=
162 mc_recoverable_range
[i
].start_addr
+
163 of_read_number(prop
+ (i
* 5) + 2, 1);
164 mc_recoverable_range
[i
].recover_addr
=
165 of_read_number(prop
+ (i
* 5) + 3, 2);
167 pr_debug("Machine check recoverable range: %llx..%llx: %llx\n",
168 mc_recoverable_range
[i
].start_addr
,
169 mc_recoverable_range
[i
].end_addr
,
170 mc_recoverable_range
[i
].recover_addr
);
175 static int __init
opal_register_exception_handlers(void)
177 #ifdef __BIG_ENDIAN__
180 if (!(powerpc_firmware_features
& FW_FEATURE_OPAL
))
183 /* Hookup some exception handlers except machine check. We use the
184 * fwnmi area at 0x7000 to provide the glue space to OPAL
189 * Check if we are running on newer firmware that exports
190 * OPAL_HANDLE_HMI token. If yes, then don't ask OPAL to patch
191 * the HMI interrupt and we catch it directly in Linux.
193 * For older firmware (i.e currently released POWER8 System Firmware
194 * as of today <= SV810_087), we fallback to old behavior and let OPAL
195 * patch the HMI vector and handle it inside OPAL firmware.
197 * For newer firmware (in development/yet to be released) we will
198 * start catching/handling HMI directly in Linux.
200 if (!opal_check_token(OPAL_HANDLE_HMI
)) {
201 pr_info("Old firmware detected, OPAL handles HMIs.\n");
202 opal_register_exception_handler(
203 OPAL_HYPERVISOR_MAINTENANCE_HANDLER
,
208 opal_register_exception_handler(OPAL_SOFTPATCH_HANDLER
, 0, glue
);
213 machine_early_initcall(powernv
, opal_register_exception_handlers
);
216 * Opal message notifier based on message type. Allow subscribers to get
217 * notified for specific messgae type.
219 int opal_message_notifier_register(enum opal_msg_type msg_type
,
220 struct notifier_block
*nb
)
222 if (!nb
|| msg_type
>= OPAL_MSG_TYPE_MAX
) {
223 pr_warning("%s: Invalid arguments, msg_type:%d\n",
228 return atomic_notifier_chain_register(
229 &opal_msg_notifier_head
[msg_type
], nb
);
231 EXPORT_SYMBOL_GPL(opal_message_notifier_register
);
233 int opal_message_notifier_unregister(enum opal_msg_type msg_type
,
234 struct notifier_block
*nb
)
236 return atomic_notifier_chain_unregister(
237 &opal_msg_notifier_head
[msg_type
], nb
);
239 EXPORT_SYMBOL_GPL(opal_message_notifier_unregister
);
241 static void opal_message_do_notify(uint32_t msg_type
, void *msg
)
243 /* notify subscribers */
244 atomic_notifier_call_chain(&opal_msg_notifier_head
[msg_type
],
248 static void opal_handle_message(void)
252 * TODO: pre-allocate a message buffer depending on opal-msg-size
253 * value in /proc/device-tree.
255 static struct opal_msg msg
;
258 ret
= opal_get_msg(__pa(&msg
), sizeof(msg
));
259 /* No opal message pending. */
260 if (ret
== OPAL_RESOURCE
)
263 /* check for errors. */
265 pr_warning("%s: Failed to retrieve opal message, err=%lld\n",
270 type
= be32_to_cpu(msg
.msg_type
);
273 if (type
>= OPAL_MSG_TYPE_MAX
) {
274 pr_warn_once("%s: Unknown message type: %u\n", __func__
, type
);
277 opal_message_do_notify(type
, (void *)&msg
);
280 static irqreturn_t
opal_message_notify(int irq
, void *data
)
282 opal_handle_message();
286 static int __init
opal_message_init(void)
290 for (i
= 0; i
< OPAL_MSG_TYPE_MAX
; i
++)
291 ATOMIC_INIT_NOTIFIER_HEAD(&opal_msg_notifier_head
[i
]);
293 irq
= opal_event_request(ilog2(OPAL_EVENT_MSG_PENDING
));
295 pr_err("%s: Can't register OPAL event irq (%d)\n",
300 ret
= request_irq(irq
, opal_message_notify
,
301 IRQ_TYPE_LEVEL_HIGH
, "opal-msg", NULL
);
303 pr_err("%s: Can't request OPAL event irq (%d)\n",
311 int opal_get_chars(uint32_t vtermno
, char *buf
, int count
)
318 opal_poll_events(&evt
);
319 if ((be64_to_cpu(evt
) & OPAL_EVENT_CONSOLE_INPUT
) == 0)
321 len
= cpu_to_be64(count
);
322 rc
= opal_console_read(vtermno
, &len
, buf
);
323 if (rc
== OPAL_SUCCESS
)
324 return be64_to_cpu(len
);
328 int opal_put_chars(uint32_t vtermno
, const char *data
, int total_len
)
339 /* We want put_chars to be atomic to avoid mangling of hvsi
340 * packets. To do that, we first test for room and return
341 * -EAGAIN if there isn't enough.
343 * Unfortunately, opal_console_write_buffer_space() doesn't
344 * appear to work on opal v1, so we just assume there is
345 * enough room and be done with it
347 spin_lock_irqsave(&opal_write_lock
, flags
);
348 rc
= opal_console_write_buffer_space(vtermno
, &olen
);
349 len
= be64_to_cpu(olen
);
350 if (rc
|| len
< total_len
) {
351 spin_unlock_irqrestore(&opal_write_lock
, flags
);
352 /* Closed -> drop characters */
355 opal_poll_events(NULL
);
359 /* We still try to handle partial completions, though they
360 * should no longer happen.
363 while(total_len
> 0 && (rc
== OPAL_BUSY
||
364 rc
== OPAL_BUSY_EVENT
|| rc
== OPAL_SUCCESS
)) {
365 olen
= cpu_to_be64(total_len
);
366 rc
= opal_console_write(vtermno
, &olen
, data
);
367 len
= be64_to_cpu(olen
);
369 /* Closed or other error drop */
370 if (rc
!= OPAL_SUCCESS
&& rc
!= OPAL_BUSY
&&
371 rc
!= OPAL_BUSY_EVENT
) {
375 if (rc
== OPAL_SUCCESS
) {
380 /* This is a bit nasty but we need that for the console to
381 * flush when there aren't any interrupts. We will clean
382 * things a bit later to limit that to synchronous path
383 * such as the kernel console and xmon/udbg
386 opal_poll_events(&evt
);
387 while(rc
== OPAL_SUCCESS
&&
388 (be64_to_cpu(evt
) & OPAL_EVENT_CONSOLE_OUTPUT
));
390 spin_unlock_irqrestore(&opal_write_lock
, flags
);
394 static int opal_recover_mce(struct pt_regs
*regs
,
395 struct machine_check_event
*evt
)
398 uint64_t ea
= get_mce_fault_addr(evt
);
400 if (!(regs
->msr
& MSR_RI
)) {
401 /* If MSR_RI isn't set, we cannot recover */
402 pr_err("Machine check interrupt unrecoverable: MSR(RI=0)\n");
404 } else if (evt
->disposition
== MCE_DISPOSITION_RECOVERED
) {
405 /* Platform corrected itself */
407 } else if (ea
&& !is_kernel_addr(ea
)) {
409 * Faulting address is not in kernel text. We should be fine.
410 * We need to find which process uses this address.
411 * For now, kill the task if we have received exception when
414 * TODO: Queue up this address for hwpoisioning later.
416 if (user_mode(regs
) && !is_global_init(current
)) {
417 _exception(SIGBUS
, regs
, BUS_MCEERR_AR
, regs
->nip
);
421 } else if (user_mode(regs
) && !is_global_init(current
) &&
422 evt
->severity
== MCE_SEV_ERROR_SYNC
) {
424 * If we have received a synchronous error when in userspace
427 _exception(SIGBUS
, regs
, BUS_MCEERR_AR
, regs
->nip
);
433 int opal_machine_check(struct pt_regs
*regs
)
435 struct machine_check_event evt
;
438 if (!get_mce_event(&evt
, MCE_EVENT_RELEASE
))
441 /* Print things out */
442 if (evt
.version
!= MCE_V1
) {
443 pr_err("Machine Check Exception, Unknown event version %d !\n",
447 machine_check_print_event_info(&evt
);
449 if (opal_recover_mce(regs
, &evt
))
453 * Unrecovered machine check, we are heading to panic path.
455 * We may have hit this MCE in very early stage of kernel
456 * initialization even before opal-prd has started running. If
457 * this is the case then this MCE error may go un-noticed or
458 * un-analyzed if we go down panic path. We need to inform
459 * BMC/OCC about this error so that they can collect relevant
460 * data for error analysis before rebooting.
461 * Use opal_cec_reboot2(OPAL_REBOOT_PLATFORM_ERROR) to do so.
462 * This function may not return on BMC based system.
464 ret
= opal_cec_reboot2(OPAL_REBOOT_PLATFORM_ERROR
,
465 "Unrecoverable Machine Check exception");
466 if (ret
== OPAL_UNSUPPORTED
) {
467 pr_emerg("Reboot type %d not supported\n",
468 OPAL_REBOOT_PLATFORM_ERROR
);
472 * We reached here. There can be three possibilities:
473 * 1. We are running on a firmware level that do not support
475 * 2. We are running on a firmware level that do not support
476 * OPAL_REBOOT_PLATFORM_ERROR reboot type.
477 * 3. We are running on FSP based system that does not need opal
478 * to trigger checkstop explicitly for error analysis. The FSP
479 * PRD component would have already got notified about this
480 * error through other channels.
482 * If hardware marked this as an unrecoverable MCE, we are
483 * going to panic anyway. Even if it didn't, it's not safe to
484 * continue at this point, so we should explicitly panic.
487 panic("PowerNV Unrecovered Machine Check");
491 /* Early hmi handler called in real mode. */
492 int opal_hmi_exception_early(struct pt_regs
*regs
)
497 * call opal hmi handler. Pass paca address as token.
498 * The return value OPAL_SUCCESS is an indication that there is
499 * an HMI event generated waiting to pull by Linux.
501 rc
= opal_handle_hmi();
502 if (rc
== OPAL_SUCCESS
) {
503 local_paca
->hmi_event_available
= 1;
509 /* HMI exception handler called in virtual mode during check_irq_replay. */
510 int opal_handle_hmi_exception(struct pt_regs
*regs
)
516 * Check if HMI event is available.
517 * if Yes, then call opal_poll_events to pull opal messages and
520 if (!local_paca
->hmi_event_available
)
523 local_paca
->hmi_event_available
= 0;
524 rc
= opal_poll_events(&evt
);
525 if (rc
== OPAL_SUCCESS
&& evt
)
526 opal_handle_events(be64_to_cpu(evt
));
531 static uint64_t find_recovery_address(uint64_t nip
)
535 for (i
= 0; i
< mc_recoverable_range_len
; i
++)
536 if ((nip
>= mc_recoverable_range
[i
].start_addr
) &&
537 (nip
< mc_recoverable_range
[i
].end_addr
))
538 return mc_recoverable_range
[i
].recover_addr
;
542 bool opal_mce_check_early_recovery(struct pt_regs
*regs
)
544 uint64_t recover_addr
= 0;
546 if (!opal
.base
|| !opal
.size
)
549 if ((regs
->nip
>= opal
.base
) &&
550 (regs
->nip
< (opal
.base
+ opal
.size
)))
551 recover_addr
= find_recovery_address(regs
->nip
);
554 * Setup regs->nip to rfi into fixup address.
557 regs
->nip
= recover_addr
;
560 return !!recover_addr
;
563 static int opal_sysfs_init(void)
565 opal_kobj
= kobject_create_and_add("opal", firmware_kobj
);
567 pr_warn("kobject_create_and_add opal failed\n");
574 static ssize_t
symbol_map_read(struct file
*fp
, struct kobject
*kobj
,
575 struct bin_attribute
*bin_attr
,
576 char *buf
, loff_t off
, size_t count
)
578 return memory_read_from_buffer(buf
, count
, &off
, bin_attr
->private,
582 static BIN_ATTR_RO(symbol_map
, 0);
584 static void opal_export_symmap(void)
588 struct device_node
*fw
;
591 fw
= of_find_node_by_path("/ibm,opal/firmware");
594 syms
= of_get_property(fw
, "symbol-map", &size
);
595 if (!syms
|| size
!= 2 * sizeof(__be64
))
598 /* Setup attributes */
599 bin_attr_symbol_map
.private = __va(be64_to_cpu(syms
[0]));
600 bin_attr_symbol_map
.size
= be64_to_cpu(syms
[1]);
602 rc
= sysfs_create_bin_file(opal_kobj
, &bin_attr_symbol_map
);
604 pr_warn("Error %d creating OPAL symbols file\n", rc
);
607 static void __init
opal_dump_region_init(void)
613 if (!opal_check_token(OPAL_REGISTER_DUMP_REGION
))
616 /* Register kernel log buffer */
617 addr
= log_buf_addr_get();
621 size
= log_buf_len_get();
625 rc
= opal_register_dump_region(OPAL_DUMP_REGION_LOG_BUF
,
627 /* Don't warn if this is just an older OPAL that doesn't
628 * know about that call
630 if (rc
&& rc
!= OPAL_UNSUPPORTED
)
631 pr_warn("DUMP: Failed to register kernel log buffer. "
635 static void opal_pdev_init(struct device_node
*opal_node
,
636 const char *compatible
)
638 struct device_node
*np
;
640 for_each_child_of_node(opal_node
, np
)
641 if (of_device_is_compatible(np
, compatible
))
642 of_platform_device_create(np
, NULL
, NULL
);
645 static void opal_i2c_create_devs(void)
647 struct device_node
*np
;
649 for_each_compatible_node(np
, NULL
, "ibm,opal-i2c")
650 of_platform_device_create(np
, NULL
, NULL
);
653 static int kopald(void *unused
)
655 unsigned long timeout
= msecs_to_jiffies(opal_heartbeat
) + 1;
661 opal_poll_events(&events
);
662 opal_handle_events(be64_to_cpu(events
));
663 schedule_timeout_interruptible(timeout
);
664 } while (!kthread_should_stop());
669 void opal_wake_poller(void)
672 wake_up_process(kopald_tsk
);
675 static void opal_init_heartbeat(void)
677 /* Old firwmware, we assume the HVC heartbeat is sufficient */
678 if (of_property_read_u32(opal_node
, "ibm,heartbeat-ms",
679 &opal_heartbeat
) != 0)
683 kopald_tsk
= kthread_run(kopald
, NULL
, "kopald");
686 static int __init
opal_init(void)
688 struct device_node
*np
, *consoles
, *leds
;
691 opal_node
= of_find_node_by_path("/ibm,opal");
693 pr_warn("Device node not found\n");
697 /* Register OPAL consoles if any ports */
698 consoles
= of_find_node_by_path("/ibm,opal/consoles");
700 for_each_child_of_node(consoles
, np
) {
701 if (strcmp(np
->name
, "serial"))
703 of_platform_device_create(np
, NULL
, NULL
);
705 of_node_put(consoles
);
708 /* Initialise OPAL messaging system */
711 /* Initialise OPAL asynchronous completion interface */
712 opal_async_comp_init();
714 /* Initialise OPAL sensor interface */
717 /* Initialise OPAL hypervisor maintainence interrupt handling */
718 opal_hmi_handler_init();
720 /* Create i2c platform devices */
721 opal_i2c_create_devs();
723 /* Setup a heatbeat thread if requested by OPAL */
724 opal_init_heartbeat();
726 /* Create leds platform devices */
727 leds
= of_find_node_by_path("/ibm,opal/leds");
729 of_platform_device_create(leds
, "opal_leds", NULL
);
733 /* Initialise OPAL message log interface */
736 /* Create "opal" kobject under /sys/firmware */
737 rc
= opal_sysfs_init();
739 /* Export symbol map to userspace */
740 opal_export_symmap();
741 /* Setup dump region interface */
742 opal_dump_region_init();
743 /* Setup error log interface */
744 rc
= opal_elog_init();
745 /* Setup code update interface */
746 opal_flash_update_init();
747 /* Setup platform dump extract interface */
748 opal_platform_dump_init();
749 /* Setup system parameters interface */
750 opal_sys_param_init();
751 /* Setup message log sysfs interface. */
752 opal_msglog_sysfs_init();
755 /* Initialize platform devices: IPMI backend, PRD & flash interface */
756 opal_pdev_init(opal_node
, "ibm,opal-ipmi");
757 opal_pdev_init(opal_node
, "ibm,opal-flash");
758 opal_pdev_init(opal_node
, "ibm,opal-prd");
760 /* Initialise platform device: oppanel interface */
761 opal_pdev_init(opal_node
, "ibm,opal-oppanel");
763 /* Initialise OPAL kmsg dumper for flushing console on panic */
768 machine_subsys_initcall(powernv
, opal_init
);
770 void opal_shutdown(void)
774 opal_event_shutdown();
777 * Then sync with OPAL which ensure anything that can
778 * potentially write to our memory has completed such
779 * as an ongoing dump retrieval
781 while (rc
== OPAL_BUSY
|| rc
== OPAL_BUSY_EVENT
) {
782 rc
= opal_sync_host_reboot();
784 opal_poll_events(NULL
);
789 /* Unregister memory dump region */
790 if (opal_check_token(OPAL_UNREGISTER_DUMP_REGION
))
791 opal_unregister_dump_region(OPAL_DUMP_REGION_LOG_BUF
);
794 /* Export this so that test modules can use it */
795 EXPORT_SYMBOL_GPL(opal_invalid_call
);
796 EXPORT_SYMBOL_GPL(opal_xscom_read
);
797 EXPORT_SYMBOL_GPL(opal_xscom_write
);
798 EXPORT_SYMBOL_GPL(opal_ipmi_send
);
799 EXPORT_SYMBOL_GPL(opal_ipmi_recv
);
800 EXPORT_SYMBOL_GPL(opal_flash_read
);
801 EXPORT_SYMBOL_GPL(opal_flash_write
);
802 EXPORT_SYMBOL_GPL(opal_flash_erase
);
803 EXPORT_SYMBOL_GPL(opal_prd_msg
);
805 /* Convert a region of vmalloc memory to an opal sg list */
806 struct opal_sg_list
*opal_vmalloc_to_sg_list(void *vmalloc_addr
,
807 unsigned long vmalloc_size
)
809 struct opal_sg_list
*sg
, *first
= NULL
;
812 sg
= kzalloc(PAGE_SIZE
, GFP_KERNEL
);
818 while (vmalloc_size
> 0) {
819 uint64_t data
= vmalloc_to_pfn(vmalloc_addr
) << PAGE_SHIFT
;
820 uint64_t length
= min(vmalloc_size
, PAGE_SIZE
);
822 sg
->entry
[i
].data
= cpu_to_be64(data
);
823 sg
->entry
[i
].length
= cpu_to_be64(length
);
826 if (i
>= SG_ENTRIES_PER_NODE
) {
827 struct opal_sg_list
*next
;
829 next
= kzalloc(PAGE_SIZE
, GFP_KERNEL
);
833 sg
->length
= cpu_to_be64(
834 i
* sizeof(struct opal_sg_entry
) + 16);
836 sg
->next
= cpu_to_be64(__pa(next
));
840 vmalloc_addr
+= length
;
841 vmalloc_size
-= length
;
844 sg
->length
= cpu_to_be64(i
* sizeof(struct opal_sg_entry
) + 16);
849 pr_err("%s : Failed to allocate memory\n", __func__
);
850 opal_free_sg_list(first
);
854 void opal_free_sg_list(struct opal_sg_list
*sg
)
857 uint64_t next
= be64_to_cpu(sg
->next
);
868 int opal_error_code(int rc
)
871 case OPAL_SUCCESS
: return 0;
873 case OPAL_PARAMETER
: return -EINVAL
;
874 case OPAL_ASYNC_COMPLETION
: return -EINPROGRESS
;
875 case OPAL_BUSY_EVENT
: return -EBUSY
;
876 case OPAL_NO_MEM
: return -ENOMEM
;
877 case OPAL_PERMISSION
: return -EPERM
;
879 case OPAL_UNSUPPORTED
: return -EIO
;
880 case OPAL_HARDWARE
: return -EIO
;
881 case OPAL_INTERNAL_ERROR
: return -EIO
;
883 pr_err("%s: unexpected OPAL error %d\n", __func__
, rc
);
888 EXPORT_SYMBOL_GPL(opal_poll_events
);
889 EXPORT_SYMBOL_GPL(opal_rtc_read
);
890 EXPORT_SYMBOL_GPL(opal_rtc_write
);
891 EXPORT_SYMBOL_GPL(opal_tpo_read
);
892 EXPORT_SYMBOL_GPL(opal_tpo_write
);
893 EXPORT_SYMBOL_GPL(opal_i2c_request
);
894 /* Export these symbols for PowerNV LED class driver */
895 EXPORT_SYMBOL_GPL(opal_leds_get_ind
);
896 EXPORT_SYMBOL_GPL(opal_leds_set_ind
);
897 /* Export this symbol for PowerNV Operator Panel class driver */
898 EXPORT_SYMBOL_GPL(opal_write_oppanel_async
);