]>
Commit | Line | Data |
---|---|---|
1 | /* SPDX-License-Identifier: GPL-2.0 */ | |
2 | #ifndef INCLUDE_XEN_OPS_H | |
3 | #define INCLUDE_XEN_OPS_H | |
4 | ||
5 | #include <linux/percpu.h> | |
6 | #include <linux/notifier.h> | |
7 | #include <linux/efi.h> | |
8 | #include <asm/xen/interface.h> | |
9 | #include <xen/interface/vcpu.h> | |
10 | ||
11 | DECLARE_PER_CPU(struct vcpu_info *, xen_vcpu); | |
12 | ||
13 | DECLARE_PER_CPU(uint32_t, xen_vcpu_id); | |
14 | static inline uint32_t xen_vcpu_nr(int cpu) | |
15 | { | |
16 | return per_cpu(xen_vcpu_id, cpu); | |
17 | } | |
18 | ||
19 | #define XEN_VCPU_ID_INVALID U32_MAX | |
20 | ||
21 | void xen_arch_pre_suspend(void); | |
22 | void xen_arch_post_suspend(int suspend_cancelled); | |
23 | ||
24 | void xen_timer_resume(void); | |
25 | void xen_arch_resume(void); | |
26 | void xen_arch_suspend(void); | |
27 | ||
28 | void xen_reboot(int reason); | |
29 | ||
30 | void xen_resume_notifier_register(struct notifier_block *nb); | |
31 | void xen_resume_notifier_unregister(struct notifier_block *nb); | |
32 | ||
33 | bool xen_vcpu_stolen(int vcpu); | |
34 | void xen_setup_runstate_info(int cpu); | |
35 | void xen_time_setup_guest(void); | |
36 | void xen_manage_runstate_time(int action); | |
37 | void xen_get_runstate_snapshot(struct vcpu_runstate_info *res); | |
38 | u64 xen_steal_clock(int cpu); | |
39 | ||
40 | int xen_setup_shutdown_event(void); | |
41 | ||
42 | extern unsigned long *xen_contiguous_bitmap; | |
43 | ||
44 | #if defined(CONFIG_XEN_PV) || defined(CONFIG_ARM) || defined(CONFIG_ARM64) | |
45 | int xen_create_contiguous_region(phys_addr_t pstart, unsigned int order, | |
46 | unsigned int address_bits, | |
47 | dma_addr_t *dma_handle); | |
48 | ||
49 | void xen_destroy_contiguous_region(phys_addr_t pstart, unsigned int order); | |
50 | #else | |
51 | static inline int xen_create_contiguous_region(phys_addr_t pstart, | |
52 | unsigned int order, | |
53 | unsigned int address_bits, | |
54 | dma_addr_t *dma_handle) | |
55 | { | |
56 | return 0; | |
57 | } | |
58 | ||
59 | static inline void xen_destroy_contiguous_region(phys_addr_t pstart, | |
60 | unsigned int order) { } | |
61 | #endif | |
62 | ||
63 | struct vm_area_struct; | |
64 | ||
65 | /* | |
66 | * xen_remap_domain_gfn_array() - map an array of foreign frames | |
67 | * @vma: VMA to map the pages into | |
68 | * @addr: Address at which to map the pages | |
69 | * @gfn: Array of GFNs to map | |
70 | * @nr: Number entries in the GFN array | |
71 | * @err_ptr: Returns per-GFN error status. | |
72 | * @prot: page protection mask | |
73 | * @domid: Domain owning the pages | |
74 | * @pages: Array of pages if this domain has an auto-translated physmap | |
75 | * | |
76 | * @gfn and @err_ptr may point to the same buffer, the GFNs will be | |
77 | * overwritten by the error codes after they are mapped. | |
78 | * | |
79 | * Returns the number of successfully mapped frames, or a -ve error | |
80 | * code. | |
81 | */ | |
82 | int xen_remap_domain_gfn_array(struct vm_area_struct *vma, | |
83 | unsigned long addr, | |
84 | xen_pfn_t *gfn, int nr, | |
85 | int *err_ptr, pgprot_t prot, | |
86 | unsigned domid, | |
87 | struct page **pages); | |
88 | ||
89 | /* xen_remap_domain_gfn_range() - map a range of foreign frames | |
90 | * @vma: VMA to map the pages into | |
91 | * @addr: Address at which to map the pages | |
92 | * @gfn: First GFN to map. | |
93 | * @nr: Number frames to map | |
94 | * @prot: page protection mask | |
95 | * @domid: Domain owning the pages | |
96 | * @pages: Array of pages if this domain has an auto-translated physmap | |
97 | * | |
98 | * Returns the number of successfully mapped frames, or a -ve error | |
99 | * code. | |
100 | */ | |
101 | int xen_remap_domain_gfn_range(struct vm_area_struct *vma, | |
102 | unsigned long addr, | |
103 | xen_pfn_t gfn, int nr, | |
104 | pgprot_t prot, unsigned domid, | |
105 | struct page **pages); | |
106 | int xen_unmap_domain_gfn_range(struct vm_area_struct *vma, | |
107 | int numpgs, struct page **pages); | |
108 | ||
109 | #ifdef CONFIG_XEN_AUTO_XLATE | |
110 | int xen_xlate_remap_gfn_array(struct vm_area_struct *vma, | |
111 | unsigned long addr, | |
112 | xen_pfn_t *gfn, int nr, | |
113 | int *err_ptr, pgprot_t prot, | |
114 | unsigned domid, | |
115 | struct page **pages); | |
116 | int xen_xlate_unmap_gfn_range(struct vm_area_struct *vma, | |
117 | int nr, struct page **pages); | |
118 | #else | |
119 | /* | |
120 | * These two functions are called from arch/x86/xen/mmu.c and so stubs | |
121 | * are needed for a configuration not specifying CONFIG_XEN_AUTO_XLATE. | |
122 | */ | |
123 | static inline int xen_xlate_remap_gfn_array(struct vm_area_struct *vma, | |
124 | unsigned long addr, | |
125 | xen_pfn_t *gfn, int nr, | |
126 | int *err_ptr, pgprot_t prot, | |
127 | unsigned int domid, | |
128 | struct page **pages) | |
129 | { | |
130 | return -EOPNOTSUPP; | |
131 | } | |
132 | ||
133 | static inline int xen_xlate_unmap_gfn_range(struct vm_area_struct *vma, | |
134 | int nr, struct page **pages) | |
135 | { | |
136 | return -EOPNOTSUPP; | |
137 | } | |
138 | #endif | |
139 | ||
140 | int xen_xlate_map_ballooned_pages(xen_pfn_t **pfns, void **vaddr, | |
141 | unsigned long nr_grant_frames); | |
142 | ||
143 | bool xen_running_on_version_or_later(unsigned int major, unsigned int minor); | |
144 | ||
145 | efi_status_t xen_efi_get_time(efi_time_t *tm, efi_time_cap_t *tc); | |
146 | efi_status_t xen_efi_set_time(efi_time_t *tm); | |
147 | efi_status_t xen_efi_get_wakeup_time(efi_bool_t *enabled, efi_bool_t *pending, | |
148 | efi_time_t *tm); | |
149 | efi_status_t xen_efi_set_wakeup_time(efi_bool_t enabled, efi_time_t *tm); | |
150 | efi_status_t xen_efi_get_variable(efi_char16_t *name, efi_guid_t *vendor, | |
151 | u32 *attr, unsigned long *data_size, | |
152 | void *data); | |
153 | efi_status_t xen_efi_get_next_variable(unsigned long *name_size, | |
154 | efi_char16_t *name, efi_guid_t *vendor); | |
155 | efi_status_t xen_efi_set_variable(efi_char16_t *name, efi_guid_t *vendor, | |
156 | u32 attr, unsigned long data_size, | |
157 | void *data); | |
158 | efi_status_t xen_efi_query_variable_info(u32 attr, u64 *storage_space, | |
159 | u64 *remaining_space, | |
160 | u64 *max_variable_size); | |
161 | efi_status_t xen_efi_get_next_high_mono_count(u32 *count); | |
162 | efi_status_t xen_efi_update_capsule(efi_capsule_header_t **capsules, | |
163 | unsigned long count, unsigned long sg_list); | |
164 | efi_status_t xen_efi_query_capsule_caps(efi_capsule_header_t **capsules, | |
165 | unsigned long count, u64 *max_size, | |
166 | int *reset_type); | |
167 | void xen_efi_reset_system(int reset_type, efi_status_t status, | |
168 | unsigned long data_size, efi_char16_t *data); | |
169 | ||
170 | ||
171 | #ifdef CONFIG_PREEMPT | |
172 | ||
173 | static inline void xen_preemptible_hcall_begin(void) | |
174 | { | |
175 | } | |
176 | ||
177 | static inline void xen_preemptible_hcall_end(void) | |
178 | { | |
179 | } | |
180 | ||
181 | #else | |
182 | ||
183 | DECLARE_PER_CPU(bool, xen_in_preemptible_hcall); | |
184 | ||
185 | static inline void xen_preemptible_hcall_begin(void) | |
186 | { | |
187 | __this_cpu_write(xen_in_preemptible_hcall, true); | |
188 | } | |
189 | ||
190 | static inline void xen_preemptible_hcall_end(void) | |
191 | { | |
192 | __this_cpu_write(xen_in_preemptible_hcall, false); | |
193 | } | |
194 | ||
195 | #endif /* CONFIG_PREEMPT */ | |
196 | ||
197 | #endif /* INCLUDE_XEN_OPS_H */ |