2 * PowerNV OPAL high level interfaces
4 * Copyright 2011 IBM Corp.
6 * This program is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License
8 * as published by the Free Software Foundation; either version
9 * 2 of the License, or (at your option) any later version.
12 #define pr_fmt(fmt) "opal: " fmt
14 #include <linux/printk.h>
15 #include <linux/types.h>
17 #include <linux/of_fdt.h>
18 #include <linux/of_platform.h>
19 #include <linux/interrupt.h>
20 #include <linux/notifier.h>
21 #include <linux/slab.h>
22 #include <linux/sched.h>
23 #include <linux/kobject.h>
24 #include <linux/delay.h>
25 #include <linux/memblock.h>
26 #include <linux/kthread.h>
27 #include <linux/freezer.h>
29 #include <asm/machdep.h>
31 #include <asm/firmware.h>
36 /* /sys/firmware/opal */
37 struct kobject
*opal_kobj
;
45 struct mcheck_recoverable_range
{
51 static struct mcheck_recoverable_range
*mc_recoverable_range
;
52 static int mc_recoverable_range_len
;
54 struct device_node
*opal_node
;
55 static DEFINE_SPINLOCK(opal_write_lock
);
56 static struct atomic_notifier_head opal_msg_notifier_head
[OPAL_MSG_TYPE_MAX
];
57 static uint32_t opal_heartbeat
;
58 static struct task_struct
*kopald_tsk
;
60 void opal_configure_cores(void)
64 /* Do the actual re-init, This will clobber all FPRs, VRs, etc...
66 * It will preserve non volatile GPRs and HSPRG0/1. It will
67 * also restore HIDs and other SPRs to their original value
68 * but it might clobber a bunch.
71 reinit_flags
|= OPAL_REINIT_CPUS_HILE_BE
;
73 reinit_flags
|= OPAL_REINIT_CPUS_HILE_LE
;
77 * POWER9 always support running hash:
78 * ie. Host hash supports hash guests
79 * Host radix supports hash/radix guests
81 if (early_cpu_has_feature(CPU_FTR_ARCH_300
)) {
82 reinit_flags
|= OPAL_REINIT_CPUS_MMU_HASH
;
83 if (early_radix_enabled())
84 reinit_flags
|= OPAL_REINIT_CPUS_MMU_RADIX
;
87 opal_reinit_cpus(reinit_flags
);
89 /* Restore some bits */
90 if (cur_cpu_spec
->cpu_restore
)
91 cur_cpu_spec
->cpu_restore();
94 int __init
early_init_dt_scan_opal(unsigned long node
,
95 const char *uname
, int depth
, void *data
)
97 const void *basep
, *entryp
, *sizep
;
98 int basesz
, entrysz
, runtimesz
;
100 if (depth
!= 1 || strcmp(uname
, "ibm,opal") != 0)
103 basep
= of_get_flat_dt_prop(node
, "opal-base-address", &basesz
);
104 entryp
= of_get_flat_dt_prop(node
, "opal-entry-address", &entrysz
);
105 sizep
= of_get_flat_dt_prop(node
, "opal-runtime-size", &runtimesz
);
107 if (!basep
|| !entryp
|| !sizep
)
110 opal
.base
= of_read_number(basep
, basesz
/4);
111 opal
.entry
= of_read_number(entryp
, entrysz
/4);
112 opal
.size
= of_read_number(sizep
, runtimesz
/4);
114 pr_debug("OPAL Base = 0x%llx (basep=%p basesz=%d)\n",
115 opal
.base
, basep
, basesz
);
116 pr_debug("OPAL Entry = 0x%llx (entryp=%p basesz=%d)\n",
117 opal
.entry
, entryp
, entrysz
);
118 pr_debug("OPAL Entry = 0x%llx (sizep=%p runtimesz=%d)\n",
119 opal
.size
, sizep
, runtimesz
);
121 if (of_flat_dt_is_compatible(node
, "ibm,opal-v3")) {
122 powerpc_firmware_features
|= FW_FEATURE_OPAL
;
123 pr_info("OPAL detected !\n");
125 panic("OPAL != V3 detected, no longer supported.\n");
131 int __init
early_init_dt_scan_recoverable_ranges(unsigned long node
,
132 const char *uname
, int depth
, void *data
)
137 if (depth
!= 1 || strcmp(uname
, "ibm,opal") != 0)
140 prop
= of_get_flat_dt_prop(node
, "mcheck-recoverable-ranges", &psize
);
145 pr_debug("Found machine check recoverable ranges.\n");
148 * Calculate number of available entries.
150 * Each recoverable address range entry is (start address, len,
151 * recovery address), 2 cells each for start and recovery address,
152 * 1 cell for len, totalling 5 cells per entry.
154 mc_recoverable_range_len
= psize
/ (sizeof(*prop
) * 5);
157 if (!mc_recoverable_range_len
)
160 /* Size required to hold all the entries. */
161 size
= mc_recoverable_range_len
*
162 sizeof(struct mcheck_recoverable_range
);
165 * Allocate a buffer to hold the MC recoverable ranges. We would be
166 * accessing them in real mode, hence it needs to be within
169 mc_recoverable_range
=__va(memblock_alloc_base(size
, __alignof__(u64
),
171 memset(mc_recoverable_range
, 0, size
);
173 for (i
= 0; i
< mc_recoverable_range_len
; i
++) {
174 mc_recoverable_range
[i
].start_addr
=
175 of_read_number(prop
+ (i
* 5) + 0, 2);
176 mc_recoverable_range
[i
].end_addr
=
177 mc_recoverable_range
[i
].start_addr
+
178 of_read_number(prop
+ (i
* 5) + 2, 1);
179 mc_recoverable_range
[i
].recover_addr
=
180 of_read_number(prop
+ (i
* 5) + 3, 2);
182 pr_debug("Machine check recoverable range: %llx..%llx: %llx\n",
183 mc_recoverable_range
[i
].start_addr
,
184 mc_recoverable_range
[i
].end_addr
,
185 mc_recoverable_range
[i
].recover_addr
);
190 static int __init
opal_register_exception_handlers(void)
192 #ifdef __BIG_ENDIAN__
195 if (!(powerpc_firmware_features
& FW_FEATURE_OPAL
))
198 /* Hookup some exception handlers except machine check. We use the
199 * fwnmi area at 0x7000 to provide the glue space to OPAL
204 * Check if we are running on newer firmware that exports
205 * OPAL_HANDLE_HMI token. If yes, then don't ask OPAL to patch
206 * the HMI interrupt and we catch it directly in Linux.
208 * For older firmware (i.e currently released POWER8 System Firmware
209 * as of today <= SV810_087), we fallback to old behavior and let OPAL
210 * patch the HMI vector and handle it inside OPAL firmware.
212 * For newer firmware (in development/yet to be released) we will
213 * start catching/handling HMI directly in Linux.
215 if (!opal_check_token(OPAL_HANDLE_HMI
)) {
216 pr_info("Old firmware detected, OPAL handles HMIs.\n");
217 opal_register_exception_handler(
218 OPAL_HYPERVISOR_MAINTENANCE_HANDLER
,
223 opal_register_exception_handler(OPAL_SOFTPATCH_HANDLER
, 0, glue
);
228 machine_early_initcall(powernv
, opal_register_exception_handlers
);
231 * Opal message notifier based on message type. Allow subscribers to get
232 * notified for specific messgae type.
234 int opal_message_notifier_register(enum opal_msg_type msg_type
,
235 struct notifier_block
*nb
)
237 if (!nb
|| msg_type
>= OPAL_MSG_TYPE_MAX
) {
238 pr_warning("%s: Invalid arguments, msg_type:%d\n",
243 return atomic_notifier_chain_register(
244 &opal_msg_notifier_head
[msg_type
], nb
);
246 EXPORT_SYMBOL_GPL(opal_message_notifier_register
);
248 int opal_message_notifier_unregister(enum opal_msg_type msg_type
,
249 struct notifier_block
*nb
)
251 return atomic_notifier_chain_unregister(
252 &opal_msg_notifier_head
[msg_type
], nb
);
254 EXPORT_SYMBOL_GPL(opal_message_notifier_unregister
);
256 static void opal_message_do_notify(uint32_t msg_type
, void *msg
)
258 /* notify subscribers */
259 atomic_notifier_call_chain(&opal_msg_notifier_head
[msg_type
],
263 static void opal_handle_message(void)
267 * TODO: pre-allocate a message buffer depending on opal-msg-size
268 * value in /proc/device-tree.
270 static struct opal_msg msg
;
273 ret
= opal_get_msg(__pa(&msg
), sizeof(msg
));
274 /* No opal message pending. */
275 if (ret
== OPAL_RESOURCE
)
278 /* check for errors. */
280 pr_warning("%s: Failed to retrieve opal message, err=%lld\n",
285 type
= be32_to_cpu(msg
.msg_type
);
288 if (type
>= OPAL_MSG_TYPE_MAX
) {
289 pr_warn_once("%s: Unknown message type: %u\n", __func__
, type
);
292 opal_message_do_notify(type
, (void *)&msg
);
295 static irqreturn_t
opal_message_notify(int irq
, void *data
)
297 opal_handle_message();
301 static int __init
opal_message_init(void)
305 for (i
= 0; i
< OPAL_MSG_TYPE_MAX
; i
++)
306 ATOMIC_INIT_NOTIFIER_HEAD(&opal_msg_notifier_head
[i
]);
308 irq
= opal_event_request(ilog2(OPAL_EVENT_MSG_PENDING
));
310 pr_err("%s: Can't register OPAL event irq (%d)\n",
315 ret
= request_irq(irq
, opal_message_notify
,
316 IRQ_TYPE_LEVEL_HIGH
, "opal-msg", NULL
);
318 pr_err("%s: Can't request OPAL event irq (%d)\n",
326 int opal_get_chars(uint32_t vtermno
, char *buf
, int count
)
333 opal_poll_events(&evt
);
334 if ((be64_to_cpu(evt
) & OPAL_EVENT_CONSOLE_INPUT
) == 0)
336 len
= cpu_to_be64(count
);
337 rc
= opal_console_read(vtermno
, &len
, buf
);
338 if (rc
== OPAL_SUCCESS
)
339 return be64_to_cpu(len
);
343 int opal_put_chars(uint32_t vtermno
, const char *data
, int total_len
)
354 /* We want put_chars to be atomic to avoid mangling of hvsi
355 * packets. To do that, we first test for room and return
356 * -EAGAIN if there isn't enough.
358 * Unfortunately, opal_console_write_buffer_space() doesn't
359 * appear to work on opal v1, so we just assume there is
360 * enough room and be done with it
362 spin_lock_irqsave(&opal_write_lock
, flags
);
363 rc
= opal_console_write_buffer_space(vtermno
, &olen
);
364 len
= be64_to_cpu(olen
);
365 if (rc
|| len
< total_len
) {
366 spin_unlock_irqrestore(&opal_write_lock
, flags
);
367 /* Closed -> drop characters */
370 opal_poll_events(NULL
);
374 /* We still try to handle partial completions, though they
375 * should no longer happen.
378 while(total_len
> 0 && (rc
== OPAL_BUSY
||
379 rc
== OPAL_BUSY_EVENT
|| rc
== OPAL_SUCCESS
)) {
380 olen
= cpu_to_be64(total_len
);
381 rc
= opal_console_write(vtermno
, &olen
, data
);
382 len
= be64_to_cpu(olen
);
384 /* Closed or other error drop */
385 if (rc
!= OPAL_SUCCESS
&& rc
!= OPAL_BUSY
&&
386 rc
!= OPAL_BUSY_EVENT
) {
390 if (rc
== OPAL_SUCCESS
) {
395 /* This is a bit nasty but we need that for the console to
396 * flush when there aren't any interrupts. We will clean
397 * things a bit later to limit that to synchronous path
398 * such as the kernel console and xmon/udbg
401 opal_poll_events(&evt
);
402 while(rc
== OPAL_SUCCESS
&&
403 (be64_to_cpu(evt
) & OPAL_EVENT_CONSOLE_OUTPUT
));
405 spin_unlock_irqrestore(&opal_write_lock
, flags
);
409 static int opal_recover_mce(struct pt_regs
*regs
,
410 struct machine_check_event
*evt
)
414 if (!(regs
->msr
& MSR_RI
)) {
415 /* If MSR_RI isn't set, we cannot recover */
416 pr_err("Machine check interrupt unrecoverable: MSR(RI=0)\n");
418 } else if (evt
->disposition
== MCE_DISPOSITION_RECOVERED
) {
419 /* Platform corrected itself */
421 } else if (evt
->severity
== MCE_SEV_FATAL
) {
422 /* Fatal machine check */
423 pr_err("Machine check interrupt is fatal\n");
425 } else if ((evt
->severity
== MCE_SEV_ERROR_SYNC
) &&
426 (user_mode(regs
) && !is_global_init(current
))) {
428 * For now, kill the task if we have received exception when
431 * TODO: Queue up this address for hwpoisioning later.
433 _exception(SIGBUS
, regs
, BUS_MCEERR_AR
, regs
->nip
);
439 int opal_machine_check(struct pt_regs
*regs
)
441 struct machine_check_event evt
;
444 if (!get_mce_event(&evt
, MCE_EVENT_RELEASE
))
447 /* Print things out */
448 if (evt
.version
!= MCE_V1
) {
449 pr_err("Machine Check Exception, Unknown event version %d !\n",
453 machine_check_print_event_info(&evt
);
455 if (opal_recover_mce(regs
, &evt
))
459 * Unrecovered machine check, we are heading to panic path.
461 * We may have hit this MCE in very early stage of kernel
462 * initialization even before opal-prd has started running. If
463 * this is the case then this MCE error may go un-noticed or
464 * un-analyzed if we go down panic path. We need to inform
465 * BMC/OCC about this error so that they can collect relevant
466 * data for error analysis before rebooting.
467 * Use opal_cec_reboot2(OPAL_REBOOT_PLATFORM_ERROR) to do so.
468 * This function may not return on BMC based system.
470 ret
= opal_cec_reboot2(OPAL_REBOOT_PLATFORM_ERROR
,
471 "Unrecoverable Machine Check exception");
472 if (ret
== OPAL_UNSUPPORTED
) {
473 pr_emerg("Reboot type %d not supported\n",
474 OPAL_REBOOT_PLATFORM_ERROR
);
478 * We reached here. There can be three possibilities:
479 * 1. We are running on a firmware level that do not support
481 * 2. We are running on a firmware level that do not support
482 * OPAL_REBOOT_PLATFORM_ERROR reboot type.
483 * 3. We are running on FSP based system that does not need opal
484 * to trigger checkstop explicitly for error analysis. The FSP
485 * PRD component would have already got notified about this
486 * error through other channels.
488 * If hardware marked this as an unrecoverable MCE, we are
489 * going to panic anyway. Even if it didn't, it's not safe to
490 * continue at this point, so we should explicitly panic.
493 panic("PowerNV Unrecovered Machine Check");
497 /* Early hmi handler called in real mode. */
498 int opal_hmi_exception_early(struct pt_regs
*regs
)
503 * call opal hmi handler. Pass paca address as token.
504 * The return value OPAL_SUCCESS is an indication that there is
505 * an HMI event generated waiting to pull by Linux.
507 rc
= opal_handle_hmi();
508 if (rc
== OPAL_SUCCESS
) {
509 local_paca
->hmi_event_available
= 1;
515 /* HMI exception handler called in virtual mode during check_irq_replay. */
516 int opal_handle_hmi_exception(struct pt_regs
*regs
)
522 * Check if HMI event is available.
523 * if Yes, then call opal_poll_events to pull opal messages and
526 if (!local_paca
->hmi_event_available
)
529 local_paca
->hmi_event_available
= 0;
530 rc
= opal_poll_events(&evt
);
531 if (rc
== OPAL_SUCCESS
&& evt
)
532 opal_handle_events(be64_to_cpu(evt
));
537 static uint64_t find_recovery_address(uint64_t nip
)
541 for (i
= 0; i
< mc_recoverable_range_len
; i
++)
542 if ((nip
>= mc_recoverable_range
[i
].start_addr
) &&
543 (nip
< mc_recoverable_range
[i
].end_addr
))
544 return mc_recoverable_range
[i
].recover_addr
;
548 bool opal_mce_check_early_recovery(struct pt_regs
*regs
)
550 uint64_t recover_addr
= 0;
552 if (!opal
.base
|| !opal
.size
)
555 if ((regs
->nip
>= opal
.base
) &&
556 (regs
->nip
< (opal
.base
+ opal
.size
)))
557 recover_addr
= find_recovery_address(regs
->nip
);
560 * Setup regs->nip to rfi into fixup address.
563 regs
->nip
= recover_addr
;
566 return !!recover_addr
;
569 static int opal_sysfs_init(void)
571 opal_kobj
= kobject_create_and_add("opal", firmware_kobj
);
573 pr_warn("kobject_create_and_add opal failed\n");
580 static ssize_t
symbol_map_read(struct file
*fp
, struct kobject
*kobj
,
581 struct bin_attribute
*bin_attr
,
582 char *buf
, loff_t off
, size_t count
)
584 return memory_read_from_buffer(buf
, count
, &off
, bin_attr
->private,
588 static BIN_ATTR_RO(symbol_map
, 0);
590 static void opal_export_symmap(void)
594 struct device_node
*fw
;
597 fw
= of_find_node_by_path("/ibm,opal/firmware");
600 syms
= of_get_property(fw
, "symbol-map", &size
);
601 if (!syms
|| size
!= 2 * sizeof(__be64
))
604 /* Setup attributes */
605 bin_attr_symbol_map
.private = __va(be64_to_cpu(syms
[0]));
606 bin_attr_symbol_map
.size
= be64_to_cpu(syms
[1]);
608 rc
= sysfs_create_bin_file(opal_kobj
, &bin_attr_symbol_map
);
610 pr_warn("Error %d creating OPAL symbols file\n", rc
);
613 static void __init
opal_dump_region_init(void)
619 if (!opal_check_token(OPAL_REGISTER_DUMP_REGION
))
622 /* Register kernel log buffer */
623 addr
= log_buf_addr_get();
627 size
= log_buf_len_get();
631 rc
= opal_register_dump_region(OPAL_DUMP_REGION_LOG_BUF
,
633 /* Don't warn if this is just an older OPAL that doesn't
634 * know about that call
636 if (rc
&& rc
!= OPAL_UNSUPPORTED
)
637 pr_warn("DUMP: Failed to register kernel log buffer. "
641 static void opal_pdev_init(const char *compatible
)
643 struct device_node
*np
;
645 for_each_compatible_node(np
, NULL
, compatible
)
646 of_platform_device_create(np
, NULL
, NULL
);
649 static int kopald(void *unused
)
651 unsigned long timeout
= msecs_to_jiffies(opal_heartbeat
) + 1;
657 opal_poll_events(&events
);
658 opal_handle_events(be64_to_cpu(events
));
659 schedule_timeout_interruptible(timeout
);
660 } while (!kthread_should_stop());
665 void opal_wake_poller(void)
668 wake_up_process(kopald_tsk
);
671 static void opal_init_heartbeat(void)
673 /* Old firwmware, we assume the HVC heartbeat is sufficient */
674 if (of_property_read_u32(opal_node
, "ibm,heartbeat-ms",
675 &opal_heartbeat
) != 0)
679 kopald_tsk
= kthread_run(kopald
, NULL
, "kopald");
682 static int __init
opal_init(void)
684 struct device_node
*np
, *consoles
, *leds
;
687 opal_node
= of_find_node_by_path("/ibm,opal");
689 pr_warn("Device node not found\n");
693 /* Register OPAL consoles if any ports */
694 consoles
= of_find_node_by_path("/ibm,opal/consoles");
696 for_each_child_of_node(consoles
, np
) {
697 if (strcmp(np
->name
, "serial"))
699 of_platform_device_create(np
, NULL
, NULL
);
701 of_node_put(consoles
);
704 /* Initialise OPAL messaging system */
707 /* Initialise OPAL asynchronous completion interface */
708 opal_async_comp_init();
710 /* Initialise OPAL sensor interface */
713 /* Initialise OPAL hypervisor maintainence interrupt handling */
714 opal_hmi_handler_init();
716 /* Create i2c platform devices */
717 opal_pdev_init("ibm,opal-i2c");
719 /* Setup a heatbeat thread if requested by OPAL */
720 opal_init_heartbeat();
722 /* Create leds platform devices */
723 leds
= of_find_node_by_path("/ibm,opal/leds");
725 of_platform_device_create(leds
, "opal_leds", NULL
);
729 /* Initialise OPAL message log interface */
732 /* Create "opal" kobject under /sys/firmware */
733 rc
= opal_sysfs_init();
735 /* Export symbol map to userspace */
736 opal_export_symmap();
737 /* Setup dump region interface */
738 opal_dump_region_init();
739 /* Setup error log interface */
740 rc
= opal_elog_init();
741 /* Setup code update interface */
742 opal_flash_update_init();
743 /* Setup platform dump extract interface */
744 opal_platform_dump_init();
745 /* Setup system parameters interface */
746 opal_sys_param_init();
747 /* Setup message log sysfs interface. */
748 opal_msglog_sysfs_init();
751 /* Initialize platform devices: IPMI backend, PRD & flash interface */
752 opal_pdev_init("ibm,opal-ipmi");
753 opal_pdev_init("ibm,opal-flash");
754 opal_pdev_init("ibm,opal-prd");
756 /* Initialise platform device: oppanel interface */
757 opal_pdev_init("ibm,opal-oppanel");
759 /* Initialise OPAL kmsg dumper for flushing console on panic */
764 machine_subsys_initcall(powernv
, opal_init
);
766 void opal_shutdown(void)
770 opal_event_shutdown();
773 * Then sync with OPAL which ensure anything that can
774 * potentially write to our memory has completed such
775 * as an ongoing dump retrieval
777 while (rc
== OPAL_BUSY
|| rc
== OPAL_BUSY_EVENT
) {
778 rc
= opal_sync_host_reboot();
780 opal_poll_events(NULL
);
785 /* Unregister memory dump region */
786 if (opal_check_token(OPAL_UNREGISTER_DUMP_REGION
))
787 opal_unregister_dump_region(OPAL_DUMP_REGION_LOG_BUF
);
790 /* Export this so that test modules can use it */
791 EXPORT_SYMBOL_GPL(opal_invalid_call
);
792 EXPORT_SYMBOL_GPL(opal_xscom_read
);
793 EXPORT_SYMBOL_GPL(opal_xscom_write
);
794 EXPORT_SYMBOL_GPL(opal_ipmi_send
);
795 EXPORT_SYMBOL_GPL(opal_ipmi_recv
);
796 EXPORT_SYMBOL_GPL(opal_flash_read
);
797 EXPORT_SYMBOL_GPL(opal_flash_write
);
798 EXPORT_SYMBOL_GPL(opal_flash_erase
);
799 EXPORT_SYMBOL_GPL(opal_prd_msg
);
801 /* Convert a region of vmalloc memory to an opal sg list */
802 struct opal_sg_list
*opal_vmalloc_to_sg_list(void *vmalloc_addr
,
803 unsigned long vmalloc_size
)
805 struct opal_sg_list
*sg
, *first
= NULL
;
808 sg
= kzalloc(PAGE_SIZE
, GFP_KERNEL
);
814 while (vmalloc_size
> 0) {
815 uint64_t data
= vmalloc_to_pfn(vmalloc_addr
) << PAGE_SHIFT
;
816 uint64_t length
= min(vmalloc_size
, PAGE_SIZE
);
818 sg
->entry
[i
].data
= cpu_to_be64(data
);
819 sg
->entry
[i
].length
= cpu_to_be64(length
);
822 if (i
>= SG_ENTRIES_PER_NODE
) {
823 struct opal_sg_list
*next
;
825 next
= kzalloc(PAGE_SIZE
, GFP_KERNEL
);
829 sg
->length
= cpu_to_be64(
830 i
* sizeof(struct opal_sg_entry
) + 16);
832 sg
->next
= cpu_to_be64(__pa(next
));
836 vmalloc_addr
+= length
;
837 vmalloc_size
-= length
;
840 sg
->length
= cpu_to_be64(i
* sizeof(struct opal_sg_entry
) + 16);
845 pr_err("%s : Failed to allocate memory\n", __func__
);
846 opal_free_sg_list(first
);
850 void opal_free_sg_list(struct opal_sg_list
*sg
)
853 uint64_t next
= be64_to_cpu(sg
->next
);
864 int opal_error_code(int rc
)
867 case OPAL_SUCCESS
: return 0;
869 case OPAL_PARAMETER
: return -EINVAL
;
870 case OPAL_ASYNC_COMPLETION
: return -EINPROGRESS
;
871 case OPAL_BUSY_EVENT
: return -EBUSY
;
872 case OPAL_NO_MEM
: return -ENOMEM
;
873 case OPAL_PERMISSION
: return -EPERM
;
875 case OPAL_UNSUPPORTED
: return -EIO
;
876 case OPAL_HARDWARE
: return -EIO
;
877 case OPAL_INTERNAL_ERROR
: return -EIO
;
879 pr_err("%s: unexpected OPAL error %d\n", __func__
, rc
);
884 void powernv_set_nmmu_ptcr(unsigned long ptcr
)
888 if (firmware_has_feature(FW_FEATURE_OPAL
)) {
889 rc
= opal_nmmu_set_ptcr(-1UL, ptcr
);
890 if (rc
!= OPAL_SUCCESS
&& rc
!= OPAL_UNSUPPORTED
)
891 pr_warn("%s: Unable to set nest mmu ptcr\n", __func__
);
895 EXPORT_SYMBOL_GPL(opal_poll_events
);
896 EXPORT_SYMBOL_GPL(opal_rtc_read
);
897 EXPORT_SYMBOL_GPL(opal_rtc_write
);
898 EXPORT_SYMBOL_GPL(opal_tpo_read
);
899 EXPORT_SYMBOL_GPL(opal_tpo_write
);
900 EXPORT_SYMBOL_GPL(opal_i2c_request
);
901 /* Export these symbols for PowerNV LED class driver */
902 EXPORT_SYMBOL_GPL(opal_leds_get_ind
);
903 EXPORT_SYMBOL_GPL(opal_leds_set_ind
);
904 /* Export this symbol for PowerNV Operator Panel class driver */
905 EXPORT_SYMBOL_GPL(opal_write_oppanel_async
);
906 /* Export this for KVM */
907 EXPORT_SYMBOL_GPL(opal_int_set_mfrr
);