]>
Commit | Line | Data |
---|---|---|
f8de50eb KA |
1 | /* |
2 | * Copyright (c) 2006, Intel Corporation. | |
3 | * | |
4 | * This file is released under the GPLv2. | |
5 | * | |
98bcef56 | 6 | * Copyright (C) 2006-2008 Intel Corporation |
7 | * Author: Anil S Keshavamurthy <anil.s.keshavamurthy@intel.com> | |
f8de50eb KA |
8 | * |
9 | */ | |
10 | ||
11 | #ifndef _IOVA_H_ | |
12 | #define _IOVA_H_ | |
13 | ||
14 | #include <linux/types.h> | |
15 | #include <linux/kernel.h> | |
16 | #include <linux/rbtree.h> | |
fb418dab | 17 | #include <linux/atomic.h> |
f8de50eb KA |
18 | #include <linux/dma-mapping.h> |
19 | ||
f8de50eb KA |
20 | /* iova structure */ |
21 | struct iova { | |
22 | struct rb_node node; | |
9257b4a2 OP |
23 | unsigned long pfn_hi; /* Highest allocated pfn */ |
24 | unsigned long pfn_lo; /* Lowest allocated pfn */ | |
25 | }; | |
26 | ||
27 | struct iova_magazine; | |
28 | struct iova_cpu_rcache; | |
29 | ||
30 | #define IOVA_RANGE_CACHE_MAX_SIZE 6 /* log of max cached IOVA range size (in pages) */ | |
31 | #define MAX_GLOBAL_MAGS 32 /* magazines per bin */ | |
32 | ||
33 | struct iova_rcache { | |
34 | spinlock_t lock; | |
35 | unsigned long depot_size; | |
36 | struct iova_magazine *depot[MAX_GLOBAL_MAGS]; | |
37 | struct iova_cpu_rcache __percpu *cpu_rcaches; | |
f8de50eb KA |
38 | }; |
39 | ||
42f87e71 JR |
40 | struct iova_domain; |
41 | ||
42 | /* Call-Back from IOVA code into IOMMU drivers */ | |
43 | typedef void (* iova_flush_cb)(struct iova_domain *domain); | |
44 | ||
45 | /* Destructor for per-entry data */ | |
46 | typedef void (* iova_entry_dtor)(unsigned long data); | |
47 | ||
48 | /* Number of entries per Flush Queue */ | |
49 | #define IOVA_FQ_SIZE 256 | |
50 | ||
9a005a80 JR |
51 | /* Timeout (in ms) after which entries are flushed from the Flush-Queue */ |
52 | #define IOVA_FQ_TIMEOUT 10 | |
53 | ||
42f87e71 JR |
54 | /* Flush Queue entry for defered flushing */ |
55 | struct iova_fq_entry { | |
56 | unsigned long iova_pfn; | |
57 | unsigned long pages; | |
58 | unsigned long data; | |
fb418dab | 59 | u64 counter; /* Flush counter when this entrie was added */ |
42f87e71 JR |
60 | }; |
61 | ||
62 | /* Per-CPU Flush Queue structure */ | |
63 | struct iova_fq { | |
64 | struct iova_fq_entry entries[IOVA_FQ_SIZE]; | |
65 | unsigned head, tail; | |
8109c2a2 | 66 | spinlock_t lock; |
42f87e71 JR |
67 | }; |
68 | ||
f8de50eb KA |
69 | /* holds all the iova translations for a domain */ |
70 | struct iova_domain { | |
f8de50eb KA |
71 | spinlock_t iova_rbtree_lock; /* Lock to protect update of rbtree */ |
72 | struct rb_root rbroot; /* iova domain rbtree root */ | |
e60aa7b5 RM |
73 | struct rb_node *cached_node; /* Save last alloced node */ |
74 | struct rb_node *cached32_node; /* Save last 32-bit alloced node */ | |
0fb5fe87 | 75 | unsigned long granule; /* pfn granularity for this domain */ |
1b722500 | 76 | unsigned long start_pfn; /* Lower limit for this domain */ |
f661197e | 77 | unsigned long dma_32bit_pfn; |
bb68b2fb | 78 | struct iova anchor; /* rbtree lookup anchor */ |
9257b4a2 | 79 | struct iova_rcache rcaches[IOVA_RANGE_CACHE_MAX_SIZE]; /* IOVA range caches */ |
42f87e71 JR |
80 | |
81 | iova_flush_cb flush_cb; /* Call-Back function to flush IOMMU | |
82 | TLBs */ | |
83 | ||
84 | iova_entry_dtor entry_dtor; /* IOMMU driver specific destructor for | |
85 | iova entry */ | |
86 | ||
87 | struct iova_fq __percpu *fq; /* Flush Queue */ | |
fb418dab JR |
88 | |
89 | atomic64_t fq_flush_start_cnt; /* Number of TLB flushes that | |
90 | have been started */ | |
91 | ||
92 | atomic64_t fq_flush_finish_cnt; /* Number of TLB flushes that | |
93 | have been finished */ | |
9a005a80 JR |
94 | |
95 | struct timer_list fq_timer; /* Timer to regularily empty the | |
96 | flush-queues */ | |
97 | atomic_t fq_timer_on; /* 1 when timer is active, 0 | |
98 | when not */ | |
f8de50eb KA |
99 | }; |
100 | ||
a156ef99 JL |
101 | static inline unsigned long iova_size(struct iova *iova) |
102 | { | |
103 | return iova->pfn_hi - iova->pfn_lo + 1; | |
104 | } | |
105 | ||
0fb5fe87 RM |
106 | static inline unsigned long iova_shift(struct iova_domain *iovad) |
107 | { | |
108 | return __ffs(iovad->granule); | |
109 | } | |
110 | ||
111 | static inline unsigned long iova_mask(struct iova_domain *iovad) | |
112 | { | |
113 | return iovad->granule - 1; | |
114 | } | |
115 | ||
116 | static inline size_t iova_offset(struct iova_domain *iovad, dma_addr_t iova) | |
117 | { | |
118 | return iova & iova_mask(iovad); | |
119 | } | |
120 | ||
121 | static inline size_t iova_align(struct iova_domain *iovad, size_t size) | |
122 | { | |
123 | return ALIGN(size, iovad->granule); | |
124 | } | |
125 | ||
126 | static inline dma_addr_t iova_dma_addr(struct iova_domain *iovad, struct iova *iova) | |
127 | { | |
128 | return (dma_addr_t)iova->pfn_lo << iova_shift(iovad); | |
129 | } | |
130 | ||
131 | static inline unsigned long iova_pfn(struct iova_domain *iovad, dma_addr_t iova) | |
132 | { | |
133 | return iova >> iova_shift(iovad); | |
134 | } | |
135 | ||
b4d8c7ae | 136 | #if IS_ENABLED(CONFIG_IOMMU_IOVA) |
ae1ff3d6 SA |
137 | int iova_cache_get(void); |
138 | void iova_cache_put(void); | |
85b45456 | 139 | |
f8de50eb KA |
140 | struct iova *alloc_iova_mem(void); |
141 | void free_iova_mem(struct iova *iova); | |
142 | void free_iova(struct iova_domain *iovad, unsigned long pfn); | |
143 | void __free_iova(struct iova_domain *iovad, struct iova *iova); | |
144 | struct iova *alloc_iova(struct iova_domain *iovad, unsigned long size, | |
f76aec76 KA |
145 | unsigned long limit_pfn, |
146 | bool size_aligned); | |
9257b4a2 OP |
147 | void free_iova_fast(struct iova_domain *iovad, unsigned long pfn, |
148 | unsigned long size); | |
19282101 JR |
149 | void queue_iova(struct iova_domain *iovad, |
150 | unsigned long pfn, unsigned long pages, | |
151 | unsigned long data); | |
9257b4a2 | 152 | unsigned long alloc_iova_fast(struct iova_domain *iovad, unsigned long size, |
538d5b33 | 153 | unsigned long limit_pfn, bool flush_rcache); |
f8de50eb KA |
154 | struct iova *reserve_iova(struct iova_domain *iovad, unsigned long pfn_lo, |
155 | unsigned long pfn_hi); | |
156 | void copy_reserved_iova(struct iova_domain *from, struct iova_domain *to); | |
0fb5fe87 | 157 | void init_iova_domain(struct iova_domain *iovad, unsigned long granule, |
aa3ac946 | 158 | unsigned long start_pfn); |
054b54a7 | 159 | bool has_iova_flush_queue(struct iova_domain *iovad); |
42f87e71 JR |
160 | int init_iova_flush_queue(struct iova_domain *iovad, |
161 | iova_flush_cb flush_cb, iova_entry_dtor entry_dtor); | |
f8de50eb KA |
162 | struct iova *find_iova(struct iova_domain *iovad, unsigned long pfn); |
163 | void put_iova_domain(struct iova_domain *iovad); | |
75f05569 JL |
164 | struct iova *split_and_remove_iova(struct iova_domain *iovad, |
165 | struct iova *iova, unsigned long pfn_lo, unsigned long pfn_hi); | |
9257b4a2 | 166 | void free_cpu_cached_iovas(unsigned int cpu, struct iova_domain *iovad); |
21aff52a TR |
167 | #else |
168 | static inline int iova_cache_get(void) | |
169 | { | |
170 | return -ENOTSUPP; | |
171 | } | |
172 | ||
173 | static inline void iova_cache_put(void) | |
174 | { | |
175 | } | |
176 | ||
177 | static inline struct iova *alloc_iova_mem(void) | |
178 | { | |
179 | return NULL; | |
180 | } | |
181 | ||
182 | static inline void free_iova_mem(struct iova *iova) | |
183 | { | |
184 | } | |
185 | ||
186 | static inline void free_iova(struct iova_domain *iovad, unsigned long pfn) | |
187 | { | |
188 | } | |
189 | ||
190 | static inline void __free_iova(struct iova_domain *iovad, struct iova *iova) | |
191 | { | |
192 | } | |
193 | ||
194 | static inline struct iova *alloc_iova(struct iova_domain *iovad, | |
195 | unsigned long size, | |
196 | unsigned long limit_pfn, | |
197 | bool size_aligned) | |
198 | { | |
199 | return NULL; | |
200 | } | |
201 | ||
202 | static inline void free_iova_fast(struct iova_domain *iovad, | |
203 | unsigned long pfn, | |
204 | unsigned long size) | |
205 | { | |
206 | } | |
207 | ||
19282101 JR |
208 | static inline void queue_iova(struct iova_domain *iovad, |
209 | unsigned long pfn, unsigned long pages, | |
210 | unsigned long data) | |
211 | { | |
212 | } | |
213 | ||
21aff52a TR |
214 | static inline unsigned long alloc_iova_fast(struct iova_domain *iovad, |
215 | unsigned long size, | |
538d5b33 TN |
216 | unsigned long limit_pfn, |
217 | bool flush_rcache) | |
21aff52a TR |
218 | { |
219 | return 0; | |
220 | } | |
221 | ||
222 | static inline struct iova *reserve_iova(struct iova_domain *iovad, | |
223 | unsigned long pfn_lo, | |
224 | unsigned long pfn_hi) | |
225 | { | |
226 | return NULL; | |
227 | } | |
228 | ||
229 | static inline void copy_reserved_iova(struct iova_domain *from, | |
230 | struct iova_domain *to) | |
231 | { | |
232 | } | |
233 | ||
234 | static inline void init_iova_domain(struct iova_domain *iovad, | |
235 | unsigned long granule, | |
aa3ac946 | 236 | unsigned long start_pfn) |
21aff52a TR |
237 | { |
238 | } | |
239 | ||
5ba32698 | 240 | static inline bool has_iova_flush_queue(struct iova_domain *iovad) |
054b54a7 DS |
241 | { |
242 | return false; | |
243 | } | |
244 | ||
42f87e71 JR |
245 | static inline int init_iova_flush_queue(struct iova_domain *iovad, |
246 | iova_flush_cb flush_cb, | |
247 | iova_entry_dtor entry_dtor) | |
248 | { | |
249 | return -ENODEV; | |
250 | } | |
251 | ||
21aff52a TR |
252 | static inline struct iova *find_iova(struct iova_domain *iovad, |
253 | unsigned long pfn) | |
254 | { | |
255 | return NULL; | |
256 | } | |
257 | ||
258 | static inline void put_iova_domain(struct iova_domain *iovad) | |
259 | { | |
260 | } | |
261 | ||
262 | static inline struct iova *split_and_remove_iova(struct iova_domain *iovad, | |
263 | struct iova *iova, | |
264 | unsigned long pfn_lo, | |
265 | unsigned long pfn_hi) | |
266 | { | |
267 | return NULL; | |
268 | } | |
269 | ||
270 | static inline void free_cpu_cached_iovas(unsigned int cpu, | |
271 | struct iova_domain *iovad) | |
272 | { | |
273 | } | |
274 | #endif | |
f8de50eb KA |
275 | |
276 | #endif |