]>
Commit | Line | Data |
---|---|---|
1da177e4 LT |
1 | /* |
2 | * eeh.h | |
3 | * Copyright (C) 2001 Dave Engebretsen & Todd Inglett IBM Corporation. | |
4 | * | |
5 | * This program is free software; you can redistribute it and/or modify | |
6 | * it under the terms of the GNU General Public License as published by | |
7 | * the Free Software Foundation; either version 2 of the License, or | |
8 | * (at your option) any later version. | |
9 | * | |
10 | * This program is distributed in the hope that it will be useful, | |
11 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
12 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
13 | * GNU General Public License for more details. | |
14 | * | |
15 | * You should have received a copy of the GNU General Public License | |
16 | * along with this program; if not, write to the Free Software | |
17 | * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA | |
18 | */ | |
19 | ||
20 | #ifndef _PPC64_EEH_H | |
21 | #define _PPC64_EEH_H | |
22 | ||
23 | #include <linux/config.h> | |
24 | #include <linux/init.h> | |
25 | #include <linux/list.h> | |
26 | #include <linux/string.h> | |
27 | ||
28 | struct pci_dev; | |
29 | struct device_node; | |
30 | struct device_node; | |
31 | struct notifier_block; | |
32 | ||
33 | #ifdef CONFIG_EEH | |
34 | ||
35 | /* Values for eeh_mode bits in device_node */ | |
36 | #define EEH_MODE_SUPPORTED (1<<0) | |
37 | #define EEH_MODE_NOCHECK (1<<1) | |
38 | #define EEH_MODE_ISOLATED (1<<2) | |
39 | ||
40 | void __init eeh_init(void); | |
41 | unsigned long eeh_check_failure(const volatile void __iomem *token, | |
42 | unsigned long val); | |
43 | int eeh_dn_check_failure(struct device_node *dn, struct pci_dev *dev); | |
44 | void __init pci_addr_cache_build(void); | |
45 | ||
46 | /** | |
47 | * eeh_add_device_early | |
48 | * eeh_add_device_late | |
49 | * | |
50 | * Perform eeh initialization for devices added after boot. | |
51 | * Call eeh_add_device_early before doing any i/o to the | |
52 | * device (including config space i/o). Call eeh_add_device_late | |
53 | * to finish the eeh setup for this device. | |
54 | */ | |
55 | void eeh_add_device_early(struct device_node *); | |
56 | void eeh_add_device_late(struct pci_dev *); | |
57 | ||
58 | /** | |
59 | * eeh_remove_device - undo EEH setup for the indicated pci device | |
60 | * @dev: pci device to be removed | |
61 | * | |
62 | * This routine should be when a device is removed from a running | |
63 | * system (e.g. by hotplug or dlpar). | |
64 | */ | |
65 | void eeh_remove_device(struct pci_dev *); | |
66 | ||
67 | #define EEH_DISABLE 0 | |
68 | #define EEH_ENABLE 1 | |
69 | #define EEH_RELEASE_LOADSTORE 2 | |
70 | #define EEH_RELEASE_DMA 3 | |
71 | ||
72 | /** | |
73 | * Notifier event flags. | |
74 | */ | |
75 | #define EEH_NOTIFY_FREEZE 1 | |
76 | ||
77 | /** EEH event -- structure holding pci slot data that describes | |
78 | * a change in the isolation status of a PCI slot. A pointer | |
79 | * to this struct is passed as the data pointer in a notify callback. | |
80 | */ | |
81 | struct eeh_event { | |
82 | struct list_head list; | |
83 | struct pci_dev *dev; | |
84 | struct device_node *dn; | |
85 | int reset_state; | |
86 | }; | |
87 | ||
88 | /** Register to find out about EEH events. */ | |
89 | int eeh_register_notifier(struct notifier_block *nb); | |
90 | int eeh_unregister_notifier(struct notifier_block *nb); | |
91 | ||
92 | /** | |
93 | * EEH_POSSIBLE_ERROR() -- test for possible MMIO failure. | |
94 | * | |
95 | * If this macro yields TRUE, the caller relays to eeh_check_failure() | |
96 | * which does further tests out of line. | |
97 | */ | |
98 | #define EEH_POSSIBLE_ERROR(val, type) ((val) == (type)~0) | |
99 | ||
100 | /* | |
101 | * Reads from a device which has been isolated by EEH will return | |
102 | * all 1s. This macro gives an all-1s value of the given size (in | |
103 | * bytes: 1, 2, or 4) for comparing with the result of a read. | |
104 | */ | |
105 | #define EEH_IO_ERROR_VALUE(size) (~0U >> ((4 - (size)) * 8)) | |
106 | ||
107 | #else /* !CONFIG_EEH */ | |
108 | static inline void eeh_init(void) { } | |
109 | ||
110 | static inline unsigned long eeh_check_failure(const volatile void __iomem *token, unsigned long val) | |
111 | { | |
112 | return val; | |
113 | } | |
114 | ||
115 | static inline int eeh_dn_check_failure(struct device_node *dn, struct pci_dev *dev) | |
116 | { | |
117 | return 0; | |
118 | } | |
119 | ||
120 | static inline void pci_addr_cache_build(void) { } | |
121 | ||
122 | static inline void eeh_add_device_early(struct device_node *dn) { } | |
123 | ||
124 | static inline void eeh_add_device_late(struct pci_dev *dev) { } | |
125 | ||
126 | static inline void eeh_remove_device(struct pci_dev *dev) { } | |
127 | ||
128 | #define EEH_POSSIBLE_ERROR(val, type) (0) | |
129 | #define EEH_IO_ERROR_VALUE(size) (-1UL) | |
130 | #endif /* CONFIG_EEH */ | |
131 | ||
132 | /* | |
133 | * MMIO read/write operations with EEH support. | |
134 | */ | |
135 | static inline u8 eeh_readb(const volatile void __iomem *addr) | |
136 | { | |
137 | u8 val = in_8(addr); | |
138 | if (EEH_POSSIBLE_ERROR(val, u8)) | |
139 | return eeh_check_failure(addr, val); | |
140 | return val; | |
141 | } | |
142 | static inline void eeh_writeb(u8 val, volatile void __iomem *addr) | |
143 | { | |
144 | out_8(addr, val); | |
145 | } | |
146 | ||
147 | static inline u16 eeh_readw(const volatile void __iomem *addr) | |
148 | { | |
149 | u16 val = in_le16(addr); | |
150 | if (EEH_POSSIBLE_ERROR(val, u16)) | |
151 | return eeh_check_failure(addr, val); | |
152 | return val; | |
153 | } | |
154 | static inline void eeh_writew(u16 val, volatile void __iomem *addr) | |
155 | { | |
156 | out_le16(addr, val); | |
157 | } | |
158 | static inline u16 eeh_raw_readw(const volatile void __iomem *addr) | |
159 | { | |
160 | u16 val = in_be16(addr); | |
161 | if (EEH_POSSIBLE_ERROR(val, u16)) | |
162 | return eeh_check_failure(addr, val); | |
163 | return val; | |
164 | } | |
165 | static inline void eeh_raw_writew(u16 val, volatile void __iomem *addr) { | |
166 | volatile u16 __iomem *vaddr = (volatile u16 __iomem *) addr; | |
167 | out_be16(vaddr, val); | |
168 | } | |
169 | ||
170 | static inline u32 eeh_readl(const volatile void __iomem *addr) | |
171 | { | |
172 | u32 val = in_le32(addr); | |
173 | if (EEH_POSSIBLE_ERROR(val, u32)) | |
174 | return eeh_check_failure(addr, val); | |
175 | return val; | |
176 | } | |
177 | static inline void eeh_writel(u32 val, volatile void __iomem *addr) | |
178 | { | |
179 | out_le32(addr, val); | |
180 | } | |
181 | static inline u32 eeh_raw_readl(const volatile void __iomem *addr) | |
182 | { | |
183 | u32 val = in_be32(addr); | |
184 | if (EEH_POSSIBLE_ERROR(val, u32)) | |
185 | return eeh_check_failure(addr, val); | |
186 | return val; | |
187 | } | |
188 | static inline void eeh_raw_writel(u32 val, volatile void __iomem *addr) | |
189 | { | |
190 | out_be32(addr, val); | |
191 | } | |
192 | ||
193 | static inline u64 eeh_readq(const volatile void __iomem *addr) | |
194 | { | |
195 | u64 val = in_le64(addr); | |
196 | if (EEH_POSSIBLE_ERROR(val, u64)) | |
197 | return eeh_check_failure(addr, val); | |
198 | return val; | |
199 | } | |
200 | static inline void eeh_writeq(u64 val, volatile void __iomem *addr) | |
201 | { | |
202 | out_le64(addr, val); | |
203 | } | |
204 | static inline u64 eeh_raw_readq(const volatile void __iomem *addr) | |
205 | { | |
206 | u64 val = in_be64(addr); | |
207 | if (EEH_POSSIBLE_ERROR(val, u64)) | |
208 | return eeh_check_failure(addr, val); | |
209 | return val; | |
210 | } | |
211 | static inline void eeh_raw_writeq(u64 val, volatile void __iomem *addr) | |
212 | { | |
213 | out_be64(addr, val); | |
214 | } | |
215 | ||
216 | #define EEH_CHECK_ALIGN(v,a) \ | |
217 | ((((unsigned long)(v)) & ((a) - 1)) == 0) | |
218 | ||
219 | static inline void eeh_memset_io(volatile void __iomem *addr, int c, | |
220 | unsigned long n) | |
221 | { | |
222 | u32 lc = c; | |
223 | lc |= lc << 8; | |
224 | lc |= lc << 16; | |
225 | ||
226 | while(n && !EEH_CHECK_ALIGN(addr, 4)) { | |
227 | *((volatile u8 *)addr) = c; | |
228 | addr = (void *)((unsigned long)addr + 1); | |
229 | n--; | |
230 | } | |
231 | while(n >= 4) { | |
232 | *((volatile u32 *)addr) = lc; | |
233 | addr = (void *)((unsigned long)addr + 4); | |
234 | n -= 4; | |
235 | } | |
236 | while(n) { | |
237 | *((volatile u8 *)addr) = c; | |
238 | addr = (void *)((unsigned long)addr + 1); | |
239 | n--; | |
240 | } | |
241 | __asm__ __volatile__ ("sync" : : : "memory"); | |
242 | } | |
243 | static inline void eeh_memcpy_fromio(void *dest, const volatile void __iomem *src, | |
244 | unsigned long n) | |
245 | { | |
246 | void *vsrc = (void __force *) src; | |
247 | void *destsave = dest; | |
248 | unsigned long nsave = n; | |
249 | ||
250 | while(n && (!EEH_CHECK_ALIGN(vsrc, 4) || !EEH_CHECK_ALIGN(dest, 4))) { | |
251 | *((u8 *)dest) = *((volatile u8 *)vsrc); | |
252 | __asm__ __volatile__ ("eieio" : : : "memory"); | |
253 | vsrc = (void *)((unsigned long)vsrc + 1); | |
254 | dest = (void *)((unsigned long)dest + 1); | |
255 | n--; | |
256 | } | |
257 | while(n > 4) { | |
258 | *((u32 *)dest) = *((volatile u32 *)vsrc); | |
259 | __asm__ __volatile__ ("eieio" : : : "memory"); | |
260 | vsrc = (void *)((unsigned long)vsrc + 4); | |
261 | dest = (void *)((unsigned long)dest + 4); | |
262 | n -= 4; | |
263 | } | |
264 | while(n) { | |
265 | *((u8 *)dest) = *((volatile u8 *)vsrc); | |
266 | __asm__ __volatile__ ("eieio" : : : "memory"); | |
267 | vsrc = (void *)((unsigned long)vsrc + 1); | |
268 | dest = (void *)((unsigned long)dest + 1); | |
269 | n--; | |
270 | } | |
271 | __asm__ __volatile__ ("sync" : : : "memory"); | |
272 | ||
273 | /* Look for ffff's here at dest[n]. Assume that at least 4 bytes | |
274 | * were copied. Check all four bytes. | |
275 | */ | |
276 | if ((nsave >= 4) && | |
277 | (EEH_POSSIBLE_ERROR((*((u32 *) destsave+nsave-4)), u32))) { | |
278 | eeh_check_failure(src, (*((u32 *) destsave+nsave-4))); | |
279 | } | |
280 | } | |
281 | ||
282 | static inline void eeh_memcpy_toio(volatile void __iomem *dest, const void *src, | |
283 | unsigned long n) | |
284 | { | |
285 | void *vdest = (void __force *) dest; | |
286 | ||
287 | while(n && (!EEH_CHECK_ALIGN(vdest, 4) || !EEH_CHECK_ALIGN(src, 4))) { | |
288 | *((volatile u8 *)vdest) = *((u8 *)src); | |
289 | src = (void *)((unsigned long)src + 1); | |
290 | vdest = (void *)((unsigned long)vdest + 1); | |
291 | n--; | |
292 | } | |
293 | while(n > 4) { | |
294 | *((volatile u32 *)vdest) = *((volatile u32 *)src); | |
295 | src = (void *)((unsigned long)src + 4); | |
296 | vdest = (void *)((unsigned long)vdest + 4); | |
297 | n-=4; | |
298 | } | |
299 | while(n) { | |
300 | *((volatile u8 *)vdest) = *((u8 *)src); | |
301 | src = (void *)((unsigned long)src + 1); | |
302 | vdest = (void *)((unsigned long)vdest + 1); | |
303 | n--; | |
304 | } | |
305 | __asm__ __volatile__ ("sync" : : : "memory"); | |
306 | } | |
307 | ||
308 | #undef EEH_CHECK_ALIGN | |
309 | ||
310 | static inline u8 eeh_inb(unsigned long port) | |
311 | { | |
312 | u8 val; | |
313 | if (!_IO_IS_VALID(port)) | |
314 | return ~0; | |
315 | val = in_8((u8 __iomem *)(port+pci_io_base)); | |
316 | if (EEH_POSSIBLE_ERROR(val, u8)) | |
317 | return eeh_check_failure((void __iomem *)(port), val); | |
318 | return val; | |
319 | } | |
320 | ||
321 | static inline void eeh_outb(u8 val, unsigned long port) | |
322 | { | |
323 | if (_IO_IS_VALID(port)) | |
324 | out_8((u8 __iomem *)(port+pci_io_base), val); | |
325 | } | |
326 | ||
327 | static inline u16 eeh_inw(unsigned long port) | |
328 | { | |
329 | u16 val; | |
330 | if (!_IO_IS_VALID(port)) | |
331 | return ~0; | |
332 | val = in_le16((u16 __iomem *)(port+pci_io_base)); | |
333 | if (EEH_POSSIBLE_ERROR(val, u16)) | |
334 | return eeh_check_failure((void __iomem *)(port), val); | |
335 | return val; | |
336 | } | |
337 | ||
338 | static inline void eeh_outw(u16 val, unsigned long port) | |
339 | { | |
340 | if (_IO_IS_VALID(port)) | |
341 | out_le16((u16 __iomem *)(port+pci_io_base), val); | |
342 | } | |
343 | ||
344 | static inline u32 eeh_inl(unsigned long port) | |
345 | { | |
346 | u32 val; | |
347 | if (!_IO_IS_VALID(port)) | |
348 | return ~0; | |
349 | val = in_le32((u32 __iomem *)(port+pci_io_base)); | |
350 | if (EEH_POSSIBLE_ERROR(val, u32)) | |
351 | return eeh_check_failure((void __iomem *)(port), val); | |
352 | return val; | |
353 | } | |
354 | ||
355 | static inline void eeh_outl(u32 val, unsigned long port) | |
356 | { | |
357 | if (_IO_IS_VALID(port)) | |
358 | out_le32((u32 __iomem *)(port+pci_io_base), val); | |
359 | } | |
360 | ||
361 | /* in-string eeh macros */ | |
362 | static inline void eeh_insb(unsigned long port, void * buf, int ns) | |
363 | { | |
364 | _insb((u8 __iomem *)(port+pci_io_base), buf, ns); | |
365 | if (EEH_POSSIBLE_ERROR((*(((u8*)buf)+ns-1)), u8)) | |
366 | eeh_check_failure((void __iomem *)(port), *(u8*)buf); | |
367 | } | |
368 | ||
369 | static inline void eeh_insw_ns(unsigned long port, void * buf, int ns) | |
370 | { | |
371 | _insw_ns((u16 __iomem *)(port+pci_io_base), buf, ns); | |
372 | if (EEH_POSSIBLE_ERROR((*(((u16*)buf)+ns-1)), u16)) | |
373 | eeh_check_failure((void __iomem *)(port), *(u16*)buf); | |
374 | } | |
375 | ||
376 | static inline void eeh_insl_ns(unsigned long port, void * buf, int nl) | |
377 | { | |
378 | _insl_ns((u32 __iomem *)(port+pci_io_base), buf, nl); | |
379 | if (EEH_POSSIBLE_ERROR((*(((u32*)buf)+nl-1)), u32)) | |
380 | eeh_check_failure((void __iomem *)(port), *(u32*)buf); | |
381 | } | |
382 | ||
383 | #endif /* _PPC64_EEH_H */ |