]>
Commit | Line | Data |
---|---|---|
8cfab3cf | 1 | // SPDX-License-Identifier: GPL-2.0 |
1da177e4 | 2 | /* |
2b937303 GKH |
3 | * (C) Copyright 2002-2004, 2007 Greg Kroah-Hartman <greg@kroah.com> |
4 | * (C) Copyright 2007 Novell Inc. | |
1da177e4 LT |
5 | */ |
6 | ||
7 | #include <linux/pci.h> | |
8 | #include <linux/module.h> | |
9 | #include <linux/init.h> | |
10 | #include <linux/device.h> | |
d42c6997 | 11 | #include <linux/mempolicy.h> |
4e57b681 TS |
12 | #include <linux/string.h> |
13 | #include <linux/slab.h> | |
8c65b4a6 | 14 | #include <linux/sched.h> |
69a18b18 | 15 | #include <linux/sched/isolation.h> |
873392ca | 16 | #include <linux/cpu.h> |
6cbf8214 | 17 | #include <linux/pm_runtime.h> |
eea3fc03 | 18 | #include <linux/suspend.h> |
4fc9bbf9 | 19 | #include <linux/kexec.h> |
07397df2 NG |
20 | #include <linux/of_device.h> |
21 | #include <linux/acpi.h> | |
a1fd09e8 | 22 | #include <linux/dma-map-ops.h> |
1da177e4 | 23 | #include "pci.h" |
c6c889d9 | 24 | #include "pcie/portdrv.h" |
1da177e4 | 25 | |
75865858 GKH |
26 | struct pci_dynid { |
27 | struct list_head node; | |
28 | struct pci_device_id id; | |
29 | }; | |
1da177e4 | 30 | |
9dba910e TH |
31 | /** |
32 | * pci_add_dynid - add a new PCI device ID to this driver and re-probe devices | |
33 | * @drv: target pci driver | |
34 | * @vendor: PCI vendor ID | |
35 | * @device: PCI device ID | |
36 | * @subvendor: PCI subvendor ID | |
37 | * @subdevice: PCI subdevice ID | |
38 | * @class: PCI class | |
39 | * @class_mask: PCI class mask | |
40 | * @driver_data: private driver data | |
41 | * | |
42 | * Adds a new dynamic pci device ID to this driver and causes the | |
43 | * driver to probe for all devices again. @drv must have been | |
44 | * registered prior to calling this function. | |
45 | * | |
46 | * CONTEXT: | |
47 | * Does GFP_KERNEL allocation. | |
48 | * | |
49 | * RETURNS: | |
50 | * 0 on success, -errno on failure. | |
51 | */ | |
52 | int pci_add_dynid(struct pci_driver *drv, | |
53 | unsigned int vendor, unsigned int device, | |
54 | unsigned int subvendor, unsigned int subdevice, | |
55 | unsigned int class, unsigned int class_mask, | |
56 | unsigned long driver_data) | |
57 | { | |
58 | struct pci_dynid *dynid; | |
9dba910e TH |
59 | |
60 | dynid = kzalloc(sizeof(*dynid), GFP_KERNEL); | |
61 | if (!dynid) | |
62 | return -ENOMEM; | |
63 | ||
64 | dynid->id.vendor = vendor; | |
65 | dynid->id.device = device; | |
66 | dynid->id.subvendor = subvendor; | |
67 | dynid->id.subdevice = subdevice; | |
68 | dynid->id.class = class; | |
69 | dynid->id.class_mask = class_mask; | |
70 | dynid->id.driver_data = driver_data; | |
3d3c2ae1 | 71 | |
9dba910e TH |
72 | spin_lock(&drv->dynids.lock); |
73 | list_add_tail(&dynid->node, &drv->dynids.list); | |
74 | spin_unlock(&drv->dynids.lock); | |
75 | ||
3b7f1016 | 76 | return driver_attach(&drv->driver); |
9dba910e | 77 | } |
b7fe9434 | 78 | EXPORT_SYMBOL_GPL(pci_add_dynid); |
9dba910e TH |
79 | |
80 | static void pci_free_dynids(struct pci_driver *drv) | |
81 | { | |
82 | struct pci_dynid *dynid, *n; | |
3d3c2ae1 | 83 | |
9dba910e TH |
84 | spin_lock(&drv->dynids.lock); |
85 | list_for_each_entry_safe(dynid, n, &drv->dynids.list, node) { | |
86 | list_del(&dynid->node); | |
87 | kfree(dynid); | |
88 | } | |
89 | spin_unlock(&drv->dynids.lock); | |
90 | } | |
91 | ||
1f40704b ZD |
92 | /** |
93 | * pci_match_id - See if a PCI device matches a given pci_id table | |
94 | * @ids: array of PCI device ID structures to search in | |
95 | * @dev: the PCI device structure to match against. | |
96 | * | |
97 | * Used by a driver to check whether a PCI device is in its list of | |
98 | * supported devices. Returns the matching pci_device_id structure or | |
99 | * %NULL if there is no match. | |
100 | * | |
101 | * Deprecated; don't use this as it will not catch any dynamic IDs | |
102 | * that a driver might want to check for. | |
103 | */ | |
104 | const struct pci_device_id *pci_match_id(const struct pci_device_id *ids, | |
105 | struct pci_dev *dev) | |
106 | { | |
107 | if (ids) { | |
108 | while (ids->vendor || ids->subvendor || ids->class_mask) { | |
109 | if (pci_match_one_device(ids, dev)) | |
110 | return ids; | |
111 | ids++; | |
112 | } | |
113 | } | |
114 | return NULL; | |
115 | } | |
116 | EXPORT_SYMBOL(pci_match_id); | |
117 | ||
118 | static const struct pci_device_id pci_device_id_any = { | |
119 | .vendor = PCI_ANY_ID, | |
120 | .device = PCI_ANY_ID, | |
121 | .subvendor = PCI_ANY_ID, | |
122 | .subdevice = PCI_ANY_ID, | |
123 | }; | |
124 | ||
125 | /** | |
126 | * pci_match_device - See if a device matches a driver's list of IDs | |
127 | * @drv: the PCI driver to match against | |
128 | * @dev: the PCI device structure to match against | |
129 | * | |
130 | * Used by a driver to check whether a PCI device is in its list of | |
131 | * supported devices or in the dynids list, which may have been augmented | |
132 | * via the sysfs "new_id" file. Returns the matching pci_device_id | |
133 | * structure or %NULL if there is no match. | |
134 | */ | |
135 | static const struct pci_device_id *pci_match_device(struct pci_driver *drv, | |
136 | struct pci_dev *dev) | |
137 | { | |
138 | struct pci_dynid *dynid; | |
139 | const struct pci_device_id *found_id = NULL; | |
140 | ||
141 | /* When driver_override is set, only bind to the matching driver */ | |
142 | if (dev->driver_override && strcmp(dev->driver_override, drv->name)) | |
143 | return NULL; | |
144 | ||
145 | /* Look at the dynamic ids first, before the static ones */ | |
146 | spin_lock(&drv->dynids.lock); | |
147 | list_for_each_entry(dynid, &drv->dynids.list, node) { | |
148 | if (pci_match_one_device(&dynid->id, dev)) { | |
149 | found_id = &dynid->id; | |
150 | break; | |
151 | } | |
152 | } | |
153 | spin_unlock(&drv->dynids.lock); | |
154 | ||
155 | if (!found_id) | |
156 | found_id = pci_match_id(drv->id_table, dev); | |
157 | ||
158 | /* driver_override will always match, send a dummy id */ | |
159 | if (!found_id && dev->driver_override) | |
160 | found_id = &pci_device_id_any; | |
161 | ||
162 | return found_id; | |
163 | } | |
164 | ||
1da177e4 | 165 | /** |
2f0cd59c | 166 | * new_id_store - sysfs frontend to pci_add_dynid() |
8f7020d3 RD |
167 | * @driver: target device driver |
168 | * @buf: buffer for scanning device ID data | |
169 | * @count: input size | |
1da177e4 | 170 | * |
9dba910e | 171 | * Allow PCI IDs to be added to an existing driver via sysfs. |
1da177e4 | 172 | */ |
a9427741 | 173 | static ssize_t new_id_store(struct device_driver *driver, const char *buf, |
3c78bc61 | 174 | size_t count) |
1da177e4 | 175 | { |
1da177e4 | 176 | struct pci_driver *pdrv = to_pci_driver(driver); |
b41d6cf3 | 177 | const struct pci_device_id *ids = pdrv->id_table; |
20a796a9 | 178 | u32 vendor, device, subvendor = PCI_ANY_ID, |
3c78bc61 RD |
179 | subdevice = PCI_ANY_ID, class = 0, class_mask = 0; |
180 | unsigned long driver_data = 0; | |
181 | int fields = 0; | |
8895d3bc | 182 | int retval = 0; |
1da177e4 | 183 | |
b41d6cf3 | 184 | fields = sscanf(buf, "%x %x %x %x %x %x %lx", |
1da177e4 LT |
185 | &vendor, &device, &subvendor, &subdevice, |
186 | &class, &class_mask, &driver_data); | |
6ba18636 | 187 | if (fields < 2) |
1da177e4 LT |
188 | return -EINVAL; |
189 | ||
8895d3bc BD |
190 | if (fields != 7) { |
191 | struct pci_dev *pdev = kzalloc(sizeof(*pdev), GFP_KERNEL); | |
192 | if (!pdev) | |
193 | return -ENOMEM; | |
194 | ||
195 | pdev->vendor = vendor; | |
196 | pdev->device = device; | |
197 | pdev->subsystem_vendor = subvendor; | |
198 | pdev->subsystem_device = subdevice; | |
199 | pdev->class = class; | |
200 | ||
3853f912 | 201 | if (pci_match_device(pdrv, pdev)) |
8895d3bc BD |
202 | retval = -EEXIST; |
203 | ||
204 | kfree(pdev); | |
205 | ||
206 | if (retval) | |
207 | return retval; | |
208 | } | |
209 | ||
b41d6cf3 JD |
210 | /* Only accept driver_data values that match an existing id_table |
211 | entry */ | |
2debb4d2 CW |
212 | if (ids) { |
213 | retval = -EINVAL; | |
214 | while (ids->vendor || ids->subvendor || ids->class_mask) { | |
215 | if (driver_data == ids->driver_data) { | |
216 | retval = 0; | |
217 | break; | |
218 | } | |
219 | ids++; | |
b41d6cf3 | 220 | } |
2debb4d2 CW |
221 | if (retval) /* No match */ |
222 | return retval; | |
b41d6cf3 | 223 | } |
b41d6cf3 | 224 | |
9dba910e TH |
225 | retval = pci_add_dynid(pdrv, vendor, device, subvendor, subdevice, |
226 | class, class_mask, driver_data); | |
b19441af GKH |
227 | if (retval) |
228 | return retval; | |
1da177e4 LT |
229 | return count; |
230 | } | |
a9427741 | 231 | static DRIVER_ATTR_WO(new_id); |
1da177e4 | 232 | |
0994375e | 233 | /** |
2f0cd59c | 234 | * remove_id_store - remove a PCI device ID from this driver |
0994375e CW |
235 | * @driver: target device driver |
236 | * @buf: buffer for scanning device ID data | |
237 | * @count: input size | |
238 | * | |
239 | * Removes a dynamic pci device ID to this driver. | |
240 | */ | |
a9427741 | 241 | static ssize_t remove_id_store(struct device_driver *driver, const char *buf, |
3c78bc61 | 242 | size_t count) |
0994375e CW |
243 | { |
244 | struct pci_dynid *dynid, *n; | |
245 | struct pci_driver *pdrv = to_pci_driver(driver); | |
20a796a9 | 246 | u32 vendor, device, subvendor = PCI_ANY_ID, |
0994375e CW |
247 | subdevice = PCI_ANY_ID, class = 0, class_mask = 0; |
248 | int fields = 0; | |
9222097f | 249 | size_t retval = -ENODEV; |
0994375e CW |
250 | |
251 | fields = sscanf(buf, "%x %x %x %x %x %x", | |
252 | &vendor, &device, &subvendor, &subdevice, | |
253 | &class, &class_mask); | |
254 | if (fields < 2) | |
255 | return -EINVAL; | |
256 | ||
257 | spin_lock(&pdrv->dynids.lock); | |
258 | list_for_each_entry_safe(dynid, n, &pdrv->dynids.list, node) { | |
259 | struct pci_device_id *id = &dynid->id; | |
260 | if ((id->vendor == vendor) && | |
261 | (id->device == device) && | |
262 | (subvendor == PCI_ANY_ID || id->subvendor == subvendor) && | |
263 | (subdevice == PCI_ANY_ID || id->subdevice == subdevice) && | |
264 | !((id->class ^ class) & class_mask)) { | |
265 | list_del(&dynid->node); | |
266 | kfree(dynid); | |
9222097f | 267 | retval = count; |
0994375e CW |
268 | break; |
269 | } | |
270 | } | |
271 | spin_unlock(&pdrv->dynids.lock); | |
272 | ||
9222097f | 273 | return retval; |
0994375e | 274 | } |
a9427741 | 275 | static DRIVER_ATTR_WO(remove_id); |
0994375e | 276 | |
2229c1fb GKH |
277 | static struct attribute *pci_drv_attrs[] = { |
278 | &driver_attr_new_id.attr, | |
279 | &driver_attr_remove_id.attr, | |
280 | NULL, | |
bfb09a86 | 281 | }; |
2229c1fb | 282 | ATTRIBUTE_GROUPS(pci_drv); |
0994375e | 283 | |
873392ca RR |
284 | struct drv_dev_and_id { |
285 | struct pci_driver *drv; | |
286 | struct pci_dev *dev; | |
287 | const struct pci_device_id *id; | |
288 | }; | |
289 | ||
290 | static long local_pci_probe(void *_ddi) | |
291 | { | |
292 | struct drv_dev_and_id *ddi = _ddi; | |
967577b0 HY |
293 | struct pci_dev *pci_dev = ddi->dev; |
294 | struct pci_driver *pci_drv = ddi->drv; | |
295 | struct device *dev = &pci_dev->dev; | |
f3ec4f87 AS |
296 | int rc; |
297 | ||
967577b0 HY |
298 | /* |
299 | * Unbound PCI devices are always put in D0, regardless of | |
300 | * runtime PM status. During probe, the device is set to | |
301 | * active and the usage count is incremented. If the driver | |
a8360062 RW |
302 | * supports runtime PM, it should call pm_runtime_put_noidle(), |
303 | * or any other runtime PM helper function decrementing the usage | |
304 | * count, in its probe routine and pm_runtime_get_noresume() in | |
305 | * its remove routine. | |
f3ec4f87 | 306 | */ |
967577b0 HY |
307 | pm_runtime_get_sync(dev); |
308 | pci_dev->driver = pci_drv; | |
309 | rc = pci_drv->probe(pci_dev, ddi->id); | |
f92d74c1 SC |
310 | if (!rc) |
311 | return rc; | |
312 | if (rc < 0) { | |
967577b0 HY |
313 | pci_dev->driver = NULL; |
314 | pm_runtime_put_sync(dev); | |
f92d74c1 | 315 | return rc; |
f3ec4f87 | 316 | } |
f92d74c1 SC |
317 | /* |
318 | * Probe function should return < 0 for failure, 0 for success | |
319 | * Treat values > 0 as success, but warn. | |
320 | */ | |
6941a0c2 BH |
321 | pci_warn(pci_dev, "Driver probe function unexpectedly returned %d\n", |
322 | rc); | |
f92d74c1 | 323 | return 0; |
873392ca RR |
324 | } |
325 | ||
0b2c2a71 TG |
326 | static bool pci_physfn_is_probed(struct pci_dev *dev) |
327 | { | |
328 | #ifdef CONFIG_PCI_IOV | |
329 | return dev->is_virtfn && dev->physfn->is_probed; | |
330 | #else | |
331 | return false; | |
332 | #endif | |
333 | } | |
334 | ||
d42c6997 AK |
335 | static int pci_call_probe(struct pci_driver *drv, struct pci_dev *dev, |
336 | const struct pci_device_id *id) | |
337 | { | |
0b2c2a71 | 338 | int error, node, cpu; |
69a18b18 | 339 | int hk_flags = HK_FLAG_DOMAIN | HK_FLAG_WQ; |
873392ca RR |
340 | struct drv_dev_and_id ddi = { drv, dev, id }; |
341 | ||
12c3156f AD |
342 | /* |
343 | * Execute driver initialization on node where the device is | |
344 | * attached. This way the driver likely allocates its local memory | |
345 | * on the right node. | |
346 | */ | |
873392ca | 347 | node = dev_to_node(&dev->dev); |
0b2c2a71 TG |
348 | dev->is_probed = 1; |
349 | ||
350 | cpu_hotplug_disable(); | |
12c3156f AD |
351 | |
352 | /* | |
0b2c2a71 TG |
353 | * Prevent nesting work_on_cpu() for the case where a Virtual Function |
354 | * device is probed from work_on_cpu() of the Physical device. | |
12c3156f | 355 | */ |
0b2c2a71 TG |
356 | if (node < 0 || node >= MAX_NUMNODES || !node_online(node) || |
357 | pci_physfn_is_probed(dev)) | |
358 | cpu = nr_cpu_ids; | |
359 | else | |
69a18b18 AB |
360 | cpu = cpumask_any_and(cpumask_of_node(node), |
361 | housekeeping_cpumask(hk_flags)); | |
0b2c2a71 TG |
362 | |
363 | if (cpu < nr_cpu_ids) | |
364 | error = work_on_cpu(cpu, local_pci_probe, &ddi); | |
365 | else | |
873392ca | 366 | error = local_pci_probe(&ddi); |
12c3156f | 367 | |
0b2c2a71 TG |
368 | dev->is_probed = 0; |
369 | cpu_hotplug_enable(); | |
d42c6997 AK |
370 | return error; |
371 | } | |
372 | ||
1da177e4 | 373 | /** |
23ea3793 | 374 | * __pci_device_probe - check if a driver wants to claim a specific PCI device |
8f7020d3 RD |
375 | * @drv: driver to call to check if it wants the PCI device |
376 | * @pci_dev: PCI device being probed | |
f7625980 | 377 | * |
8f7020d3 | 378 | * returns 0 on success, else error. |
1da177e4 LT |
379 | * side-effect: pci_dev->driver is set to drv when drv claims pci_dev. |
380 | */ | |
3c78bc61 | 381 | static int __pci_device_probe(struct pci_driver *drv, struct pci_dev *pci_dev) |
75865858 GKH |
382 | { |
383 | const struct pci_device_id *id; | |
1da177e4 LT |
384 | int error = 0; |
385 | ||
386 | if (!pci_dev->driver && drv->probe) { | |
75865858 GKH |
387 | error = -ENODEV; |
388 | ||
389 | id = pci_match_device(drv, pci_dev); | |
390 | if (id) | |
d42c6997 | 391 | error = pci_call_probe(drv, pci_dev, id); |
1da177e4 LT |
392 | } |
393 | return error; | |
394 | } | |
395 | ||
890e4847 JL |
396 | int __weak pcibios_alloc_irq(struct pci_dev *dev) |
397 | { | |
398 | return 0; | |
399 | } | |
400 | ||
401 | void __weak pcibios_free_irq(struct pci_dev *dev) | |
402 | { | |
403 | } | |
404 | ||
0e7df224 BW |
405 | #ifdef CONFIG_PCI_IOV |
406 | static inline bool pci_device_can_probe(struct pci_dev *pdev) | |
407 | { | |
2d2f4273 AW |
408 | return (!pdev->is_virtfn || pdev->physfn->sriov->drivers_autoprobe || |
409 | pdev->driver_override); | |
0e7df224 BW |
410 | } |
411 | #else | |
412 | static inline bool pci_device_can_probe(struct pci_dev *pdev) | |
413 | { | |
414 | return true; | |
415 | } | |
416 | #endif | |
417 | ||
3c78bc61 | 418 | static int pci_device_probe(struct device *dev) |
1da177e4 | 419 | { |
890e4847 JL |
420 | int error; |
421 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
422 | struct pci_driver *drv = to_pci_driver(dev->driver); | |
423 | ||
76002d8b AW |
424 | if (!pci_device_can_probe(pci_dev)) |
425 | return -ENODEV; | |
426 | ||
30fdfb92 MM |
427 | pci_assign_irq(pci_dev); |
428 | ||
890e4847 JL |
429 | error = pcibios_alloc_irq(pci_dev); |
430 | if (error < 0) | |
431 | return error; | |
1da177e4 | 432 | |
1da177e4 | 433 | pci_dev_get(pci_dev); |
76002d8b AW |
434 | error = __pci_device_probe(drv, pci_dev); |
435 | if (error) { | |
436 | pcibios_free_irq(pci_dev); | |
437 | pci_dev_put(pci_dev); | |
890e4847 | 438 | } |
1da177e4 LT |
439 | |
440 | return error; | |
441 | } | |
442 | ||
3c78bc61 | 443 | static int pci_device_remove(struct device *dev) |
1da177e4 | 444 | { |
3c78bc61 RD |
445 | struct pci_dev *pci_dev = to_pci_dev(dev); |
446 | struct pci_driver *drv = pci_dev->driver; | |
1da177e4 LT |
447 | |
448 | if (drv) { | |
f3ec4f87 AS |
449 | if (drv->remove) { |
450 | pm_runtime_get_sync(dev); | |
1da177e4 | 451 | drv->remove(pci_dev); |
f3ec4f87 AS |
452 | pm_runtime_put_noidle(dev); |
453 | } | |
890e4847 | 454 | pcibios_free_irq(pci_dev); |
1da177e4 | 455 | pci_dev->driver = NULL; |
38972375 | 456 | pci_iov_remove(pci_dev); |
1da177e4 LT |
457 | } |
458 | ||
f3ec4f87 | 459 | /* Undo the runtime PM settings in local_pci_probe() */ |
967577b0 | 460 | pm_runtime_put_sync(dev); |
f3ec4f87 | 461 | |
2449e06a SL |
462 | /* |
463 | * If the device is still on, set the power state as "unknown", | |
464 | * since it might change by the next time we load the driver. | |
465 | */ | |
466 | if (pci_dev->current_state == PCI_D0) | |
467 | pci_dev->current_state = PCI_UNKNOWN; | |
468 | ||
1da177e4 LT |
469 | /* |
470 | * We would love to complain here if pci_dev->is_enabled is set, that | |
471 | * the driver should have called pci_disable_device(), but the | |
472 | * unfortunate fact is there are too many odd BIOS and bridge setups | |
f7625980 | 473 | * that don't like drivers doing that all of the time. |
1da177e4 LT |
474 | * Oh well, we can dream of sane hardware when we sleep, no matter how |
475 | * horrible the crap we have to deal with is when we are awake... | |
476 | */ | |
477 | ||
478 | pci_dev_put(pci_dev); | |
479 | return 0; | |
480 | } | |
481 | ||
bbb44d9f RW |
482 | static void pci_device_shutdown(struct device *dev) |
483 | { | |
484 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
485 | struct pci_driver *drv = pci_dev->driver; | |
486 | ||
3ff2de9b HY |
487 | pm_runtime_resume(dev); |
488 | ||
bbb44d9f RW |
489 | if (drv && drv->shutdown) |
490 | drv->shutdown(pci_dev); | |
5b415f1e | 491 | |
b566a22c | 492 | /* |
4fc9bbf9 KA |
493 | * If this is a kexec reboot, turn off Bus Master bit on the |
494 | * device to tell it to not continue to do DMA. Don't touch | |
495 | * devices in D3cold or unknown states. | |
496 | * If it is not a kexec reboot, firmware will hit the PCI | |
497 | * devices with big hammer and stop their DMA any way. | |
b566a22c | 498 | */ |
4fc9bbf9 | 499 | if (kexec_in_progress && (pci_dev->current_state <= PCI_D3hot)) |
6e0eda3c | 500 | pci_clear_master(pci_dev); |
bbb44d9f RW |
501 | } |
502 | ||
aa338601 | 503 | #ifdef CONFIG_PM |
6cbf8214 RW |
504 | |
505 | /* Auxiliary functions used for system resume and run-time resume. */ | |
506 | ||
507 | /** | |
508 | * pci_restore_standard_config - restore standard config registers of PCI device | |
509 | * @pci_dev: PCI device to handle | |
510 | */ | |
511 | static int pci_restore_standard_config(struct pci_dev *pci_dev) | |
512 | { | |
513 | pci_update_current_state(pci_dev, PCI_UNKNOWN); | |
514 | ||
515 | if (pci_dev->current_state != PCI_D0) { | |
516 | int error = pci_set_power_state(pci_dev, PCI_D0); | |
517 | if (error) | |
518 | return error; | |
519 | } | |
520 | ||
1d3c16a8 | 521 | pci_restore_state(pci_dev); |
0ce3fcaf | 522 | pci_pme_restore(pci_dev); |
1d3c16a8 | 523 | return 0; |
6cbf8214 RW |
524 | } |
525 | ||
f7b32a86 BH |
526 | static void pci_pm_default_resume(struct pci_dev *pci_dev) |
527 | { | |
528 | pci_fixup_device(pci_fixup_resume, pci_dev); | |
529 | pci_enable_wake(pci_dev, PCI_D0, false); | |
530 | } | |
531 | ||
db288c9c RW |
532 | #endif |
533 | ||
534 | #ifdef CONFIG_PM_SLEEP | |
535 | ||
6cbf8214 RW |
536 | static void pci_pm_default_resume_early(struct pci_dev *pci_dev) |
537 | { | |
db288c9c | 538 | pci_power_up(pci_dev); |
81cfa590 | 539 | pci_update_current_state(pci_dev, PCI_D0); |
db288c9c | 540 | pci_restore_state(pci_dev); |
0ce3fcaf | 541 | pci_pme_restore(pci_dev); |
6cbf8214 RW |
542 | } |
543 | ||
fa58d305 RW |
544 | /* |
545 | * Default "suspend" method for devices that have no driver provided suspend, | |
546 | * or not even a driver at all (second part). | |
bbb44d9f | 547 | */ |
bb808945 | 548 | static void pci_pm_set_unknown_state(struct pci_dev *pci_dev) |
bbb44d9f | 549 | { |
bbb44d9f RW |
550 | /* |
551 | * mark its power state as "unknown", since we don't know if | |
552 | * e.g. the BIOS will change its device state when we suspend. | |
553 | */ | |
554 | if (pci_dev->current_state == PCI_D0) | |
555 | pci_dev->current_state = PCI_UNKNOWN; | |
556 | } | |
557 | ||
355a72d7 RW |
558 | /* |
559 | * Default "resume" method for devices that have no driver provided resume, | |
560 | * or not even a driver at all (second part). | |
561 | */ | |
bb808945 | 562 | static int pci_pm_reenable_device(struct pci_dev *pci_dev) |
355a72d7 RW |
563 | { |
564 | int retval; | |
565 | ||
bbb44d9f RW |
566 | /* if the device was enabled before suspend, reenable */ |
567 | retval = pci_reenable_device(pci_dev); | |
568 | /* | |
569 | * if the device was busmaster before the suspend, make it busmaster | |
570 | * again | |
571 | */ | |
572 | if (pci_dev->is_busmaster) | |
573 | pci_set_master(pci_dev); | |
574 | ||
575 | return retval; | |
576 | } | |
577 | ||
578 | static int pci_legacy_suspend(struct device *dev, pm_message_t state) | |
1da177e4 | 579 | { |
3c78bc61 RD |
580 | struct pci_dev *pci_dev = to_pci_dev(dev); |
581 | struct pci_driver *drv = pci_dev->driver; | |
46939f8b | 582 | |
02669492 | 583 | if (drv && drv->suspend) { |
99dadce8 | 584 | pci_power_t prev = pci_dev->current_state; |
46939f8b | 585 | int error; |
aa8c6c93 | 586 | |
57ef8026 FP |
587 | error = drv->suspend(pci_dev, state); |
588 | suspend_report_result(drv->suspend, error); | |
589 | if (error) | |
590 | return error; | |
aa8c6c93 | 591 | |
46939f8b | 592 | if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 |
99dadce8 | 593 | && pci_dev->current_state != PCI_UNKNOWN) { |
12bcae44 BH |
594 | pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, |
595 | "PCI PM: Device state not saved by %pS\n", | |
596 | drv->suspend); | |
99dadce8 | 597 | } |
02669492 | 598 | } |
ad8cfa1d RW |
599 | |
600 | pci_fixup_device(pci_fixup_suspend, pci_dev); | |
601 | ||
46939f8b | 602 | return 0; |
1da177e4 LT |
603 | } |
604 | ||
bbb44d9f | 605 | static int pci_legacy_suspend_late(struct device *dev, pm_message_t state) |
cbd69dbb | 606 | { |
3c78bc61 | 607 | struct pci_dev *pci_dev = to_pci_dev(dev); |
46939f8b RW |
608 | |
609 | if (!pci_dev->state_saved) | |
610 | pci_save_state(pci_dev); | |
611 | ||
612 | pci_pm_set_unknown_state(pci_dev); | |
613 | ||
7d2a01b8 AN |
614 | pci_fixup_device(pci_fixup_suspend_late, pci_dev); |
615 | ||
46939f8b | 616 | return 0; |
cbd69dbb | 617 | } |
1da177e4 | 618 | |
bbb44d9f | 619 | static int pci_legacy_resume(struct device *dev) |
1da177e4 | 620 | { |
3c78bc61 RD |
621 | struct pci_dev *pci_dev = to_pci_dev(dev); |
622 | struct pci_driver *drv = pci_dev->driver; | |
1da177e4 | 623 | |
ad8cfa1d RW |
624 | pci_fixup_device(pci_fixup_resume, pci_dev); |
625 | ||
aa8c6c93 RW |
626 | return drv && drv->resume ? |
627 | drv->resume(pci_dev) : pci_pm_reenable_device(pci_dev); | |
1da177e4 LT |
628 | } |
629 | ||
571ff758 RW |
630 | /* Auxiliary functions used by the new power management framework */ |
631 | ||
5294e256 | 632 | static void pci_pm_default_suspend(struct pci_dev *pci_dev) |
73410429 | 633 | { |
5294e256 | 634 | /* Disable non-bridge devices without PM support */ |
326c1cda | 635 | if (!pci_has_subordinate(pci_dev)) |
cbbc2f6b | 636 | pci_disable_enabled_device(pci_dev); |
73410429 RW |
637 | } |
638 | ||
07e836e8 RW |
639 | static bool pci_has_legacy_pm_support(struct pci_dev *pci_dev) |
640 | { | |
641 | struct pci_driver *drv = pci_dev->driver; | |
1a1daf09 | 642 | bool ret = drv && (drv->suspend || drv->resume); |
bb808945 RW |
643 | |
644 | /* | |
645 | * Legacy PM support is used by default, so warn if the new framework is | |
646 | * supported as well. Drivers are supposed to support either the | |
647 | * former, or the latter, but not both at the same time. | |
648 | */ | |
12bcae44 BH |
649 | pci_WARN(pci_dev, ret && drv->driver.pm, "device %04x:%04x\n", |
650 | pci_dev->vendor, pci_dev->device); | |
bb808945 RW |
651 | |
652 | return ret; | |
07e836e8 RW |
653 | } |
654 | ||
571ff758 RW |
655 | /* New power management framework */ |
656 | ||
bbb44d9f RW |
657 | static int pci_pm_prepare(struct device *dev) |
658 | { | |
0c7376ad | 659 | struct pci_dev *pci_dev = to_pci_dev(dev); |
6da2f2cc | 660 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 661 | |
6da2f2cc BH |
662 | if (pm && pm->prepare) { |
663 | int error = pm->prepare(dev); | |
08810a41 | 664 | if (error < 0) |
bac2a909 | 665 | return error; |
08810a41 RW |
666 | |
667 | if (!error && dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_PREPARE)) | |
668 | return 0; | |
bac2a909 | 669 | } |
0c7376ad RW |
670 | if (pci_dev_need_resume(pci_dev)) |
671 | return 0; | |
672 | ||
673 | /* | |
674 | * The PME setting needs to be adjusted here in case the direct-complete | |
675 | * optimization is used with respect to this device. | |
676 | */ | |
677 | pci_dev_adjust_pme(pci_dev); | |
678 | return 1; | |
bbb44d9f RW |
679 | } |
680 | ||
2cef548a RW |
681 | static void pci_pm_complete(struct device *dev) |
682 | { | |
a0d2a959 LW |
683 | struct pci_dev *pci_dev = to_pci_dev(dev); |
684 | ||
685 | pci_dev_complete_resume(pci_dev); | |
686 | pm_generic_complete(dev); | |
687 | ||
688 | /* Resume device if platform firmware has put it in reset-power-on */ | |
bd755d77 | 689 | if (pm_runtime_suspended(dev) && pm_resume_via_firmware()) { |
a0d2a959 LW |
690 | pci_power_t pre_sleep_state = pci_dev->current_state; |
691 | ||
b51033e0 RW |
692 | pci_refresh_power_state(pci_dev); |
693 | /* | |
694 | * On platforms with ACPI this check may also trigger for | |
695 | * devices sharing power resources if one of those power | |
696 | * resources has been activated as a result of a change of the | |
697 | * power state of another device sharing it. However, in that | |
698 | * case it is also better to resume the device, in general. | |
699 | */ | |
a0d2a959 LW |
700 | if (pci_dev->current_state < pre_sleep_state) |
701 | pm_request_resume(dev); | |
702 | } | |
2cef548a | 703 | } |
bbb44d9f | 704 | |
6cbf8214 RW |
705 | #else /* !CONFIG_PM_SLEEP */ |
706 | ||
707 | #define pci_pm_prepare NULL | |
2cef548a | 708 | #define pci_pm_complete NULL |
6cbf8214 RW |
709 | |
710 | #endif /* !CONFIG_PM_SLEEP */ | |
711 | ||
bbb44d9f | 712 | #ifdef CONFIG_SUSPEND |
a39bd851 BH |
713 | static void pcie_pme_root_status_cleanup(struct pci_dev *pci_dev) |
714 | { | |
715 | /* | |
716 | * Some BIOSes forget to clear Root PME Status bits after system | |
717 | * wakeup, which breaks ACPI-based runtime wakeup on PCI Express. | |
718 | * Clear those bits now just in case (shouldn't hurt). | |
719 | */ | |
720 | if (pci_is_pcie(pci_dev) && | |
3620c714 BH |
721 | (pci_pcie_type(pci_dev) == PCI_EXP_TYPE_ROOT_PORT || |
722 | pci_pcie_type(pci_dev) == PCI_EXP_TYPE_RC_EC)) | |
a39bd851 BH |
723 | pcie_clear_root_pme_status(pci_dev); |
724 | } | |
bbb44d9f RW |
725 | |
726 | static int pci_pm_suspend(struct device *dev) | |
727 | { | |
728 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
8150f32b | 729 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 730 | |
d491f2b7 RW |
731 | pci_dev->skip_bus_pm = false; |
732 | ||
ad8cfa1d RW |
733 | if (pci_has_legacy_pm_support(pci_dev)) |
734 | return pci_legacy_suspend(dev, PMSG_SUSPEND); | |
bb808945 | 735 | |
5294e256 RW |
736 | if (!pm) { |
737 | pci_pm_default_suspend(pci_dev); | |
c4b65157 | 738 | return 0; |
5294e256 RW |
739 | } |
740 | ||
7cd0602d | 741 | /* |
c4b65157 RW |
742 | * PCI devices suspended at run time may need to be resumed at this |
743 | * point, because in general it may be necessary to reconfigure them for | |
744 | * system suspend. Namely, if the device is expected to wake up the | |
745 | * system from the sleep state, it may have to be reconfigured for this | |
746 | * purpose, or if the device is not expected to wake up the system from | |
747 | * the sleep state, it should be prevented from signaling wakeup events | |
748 | * going forward. | |
749 | * | |
750 | * Also if the driver of the device does not indicate that its system | |
751 | * suspend callbacks can cope with runtime-suspended devices, it is | |
752 | * better to resume the device from runtime suspend here. | |
7cd0602d | 753 | */ |
c4b65157 | 754 | if (!dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_SUSPEND) || |
0c7376ad | 755 | pci_dev_need_resume(pci_dev)) { |
c4b65157 | 756 | pm_runtime_resume(dev); |
656088aa | 757 | pci_dev->state_saved = false; |
0c7376ad RW |
758 | } else { |
759 | pci_dev_adjust_pme(pci_dev); | |
656088aa | 760 | } |
7cd0602d | 761 | |
5294e256 RW |
762 | if (pm->suspend) { |
763 | pci_power_t prev = pci_dev->current_state; | |
764 | int error; | |
765 | ||
ddb7c9d2 RW |
766 | error = pm->suspend(dev); |
767 | suspend_report_result(pm->suspend, error); | |
5294e256 RW |
768 | if (error) |
769 | return error; | |
770 | ||
46939f8b | 771 | if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 |
5294e256 | 772 | && pci_dev->current_state != PCI_UNKNOWN) { |
12bcae44 BH |
773 | pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, |
774 | "PCI PM: State of device not saved by %pS\n", | |
775 | pm->suspend); | |
5294e256 | 776 | } |
bbb44d9f | 777 | } |
fa58d305 | 778 | |
5294e256 | 779 | return 0; |
bbb44d9f RW |
780 | } |
781 | ||
c4b65157 RW |
782 | static int pci_pm_suspend_late(struct device *dev) |
783 | { | |
fa2bfead | 784 | if (dev_pm_skip_suspend(dev)) |
c4b65157 RW |
785 | return 0; |
786 | ||
787 | pci_fixup_device(pci_fixup_suspend, to_pci_dev(dev)); | |
788 | ||
789 | return pm_generic_suspend_late(dev); | |
790 | } | |
791 | ||
bbb44d9f | 792 | static int pci_pm_suspend_noirq(struct device *dev) |
c8958177 | 793 | { |
355a72d7 | 794 | struct pci_dev *pci_dev = to_pci_dev(dev); |
8150f32b | 795 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
c8958177 | 796 | |
fa2bfead | 797 | if (dev_pm_skip_suspend(dev)) |
c4b65157 RW |
798 | return 0; |
799 | ||
bb808945 RW |
800 | if (pci_has_legacy_pm_support(pci_dev)) |
801 | return pci_legacy_suspend_late(dev, PMSG_SUSPEND); | |
802 | ||
931ff68a RW |
803 | if (!pm) { |
804 | pci_save_state(pci_dev); | |
7d2a01b8 | 805 | goto Fixup; |
931ff68a | 806 | } |
46939f8b RW |
807 | |
808 | if (pm->suspend_noirq) { | |
809 | pci_power_t prev = pci_dev->current_state; | |
810 | int error; | |
811 | ||
812 | error = pm->suspend_noirq(dev); | |
813 | suspend_report_result(pm->suspend_noirq, error); | |
814 | if (error) | |
815 | return error; | |
816 | ||
817 | if (!pci_dev->state_saved && pci_dev->current_state != PCI_D0 | |
818 | && pci_dev->current_state != PCI_UNKNOWN) { | |
12bcae44 BH |
819 | pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, |
820 | "PCI PM: State of device not saved by %pS\n", | |
821 | pm->suspend_noirq); | |
7d2a01b8 | 822 | goto Fixup; |
46939f8b | 823 | } |
bbb44d9f RW |
824 | } |
825 | ||
d491f2b7 RW |
826 | if (pci_dev->skip_bus_pm) { |
827 | /* | |
3e26c5fe RW |
828 | * Either the device is a bridge with a child in D0 below it, or |
829 | * the function is running for the second time in a row without | |
d491f2b7 | 830 | * going through full resume, which is possible only during |
3e26c5fe RW |
831 | * suspend-to-idle in a spurious wakeup case. The device should |
832 | * be in D0 at this point, but if it is a bridge, it may be | |
833 | * necessary to save its state. | |
d491f2b7 | 834 | */ |
3e26c5fe RW |
835 | if (!pci_dev->state_saved) |
836 | pci_save_state(pci_dev); | |
837 | } else if (!pci_dev->state_saved) { | |
46939f8b | 838 | pci_save_state(pci_dev); |
9d26d3a8 | 839 | if (pci_power_manageable(pci_dev)) |
46939f8b RW |
840 | pci_prepare_to_sleep(pci_dev); |
841 | } | |
d67e37d7 | 842 | |
6941a0c2 | 843 | pci_dbg(pci_dev, "PCI PM: Suspend power state: %s\n", |
ca67ab5c RW |
844 | pci_power_name(pci_dev->current_state)); |
845 | ||
3e26c5fe RW |
846 | if (pci_dev->current_state == PCI_D0) { |
847 | pci_dev->skip_bus_pm = true; | |
848 | /* | |
849 | * Per PCI PM r1.2, table 6-1, a bridge must be in D0 if any | |
850 | * downstream device is in D0, so avoid changing the power state | |
851 | * of the parent bridge by setting the skip_bus_pm flag for it. | |
852 | */ | |
853 | if (pci_dev->bus->self) | |
854 | pci_dev->bus->self->skip_bus_pm = true; | |
855 | } | |
856 | ||
471a739a | 857 | if (pci_dev->skip_bus_pm && pm_suspend_no_platform()) { |
6941a0c2 | 858 | pci_dbg(pci_dev, "PCI PM: Skipped\n"); |
3e26c5fe RW |
859 | goto Fixup; |
860 | } | |
861 | ||
46939f8b RW |
862 | pci_pm_set_unknown_state(pci_dev); |
863 | ||
dbf0e4c7 AS |
864 | /* |
865 | * Some BIOSes from ASUS have a bug: If a USB EHCI host controller's | |
866 | * PCI COMMAND register isn't 0, the BIOS assumes that the controller | |
867 | * hasn't been quiesced and tries to turn it off. If the controller | |
868 | * is already in D3, this can hang or cause memory corruption. | |
869 | * | |
870 | * Since the value of the COMMAND register doesn't matter once the | |
871 | * device has been suspended, we can safely set it to 0 here. | |
872 | */ | |
873 | if (pci_dev->class == PCI_CLASS_SERIAL_USB_EHCI) | |
874 | pci_write_config_word(pci_dev, PCI_COMMAND, 0); | |
875 | ||
7d2a01b8 AN |
876 | Fixup: |
877 | pci_fixup_device(pci_fixup_suspend_late, pci_dev); | |
878 | ||
bd755d77 RW |
879 | /* |
880 | * If the target system sleep state is suspend-to-idle, it is sufficient | |
881 | * to check whether or not the device's wakeup settings are good for | |
882 | * runtime PM. Otherwise, the pm_resume_via_firmware() check will cause | |
883 | * pci_pm_complete() to take care of fixing up the device's state | |
884 | * anyway, if need be. | |
885 | */ | |
0fe8a1be RW |
886 | if (device_can_wakeup(dev) && !device_may_wakeup(dev)) |
887 | dev->power.may_skip_resume = false; | |
bd755d77 | 888 | |
46939f8b | 889 | return 0; |
c8958177 | 890 | } |
1da177e4 | 891 | |
f6dc1e5e | 892 | static int pci_pm_resume_noirq(struct device *dev) |
bbb44d9f RW |
893 | { |
894 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
6da2f2cc | 895 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
ad9001f2 MW |
896 | pci_power_t prev_state = pci_dev->current_state; |
897 | bool skip_bus_pm = pci_dev->skip_bus_pm; | |
bbb44d9f | 898 | |
76c70cb5 | 899 | if (dev_pm_skip_resume(dev)) |
bd755d77 RW |
900 | return 0; |
901 | ||
3e26c5fe RW |
902 | /* |
903 | * In the suspend-to-idle case, devices left in D0 during suspend will | |
904 | * stay in D0, so it is not necessary to restore or update their | |
471a739a RW |
905 | * configuration here and attempting to put them into D0 again is |
906 | * pointless, so avoid doing that. | |
3e26c5fe | 907 | */ |
ad9001f2 | 908 | if (!(skip_bus_pm && pm_suspend_no_platform())) |
3e26c5fe RW |
909 | pci_pm_default_resume_early(pci_dev); |
910 | ||
911 | pci_fixup_device(pci_fixup_resume_early, pci_dev); | |
ec6a75ef | 912 | pcie_pme_root_status_cleanup(pci_dev); |
aa8c6c93 | 913 | |
ad9001f2 MW |
914 | if (!skip_bus_pm && prev_state == PCI_D3cold) |
915 | pci_bridge_wait_for_secondary_bus(pci_dev); | |
916 | ||
ad8cfa1d | 917 | if (pci_has_legacy_pm_support(pci_dev)) |
89cdbc35 | 918 | return 0; |
bb808945 | 919 | |
6da2f2cc BH |
920 | if (pm && pm->resume_noirq) |
921 | return pm->resume_noirq(dev); | |
bbb44d9f | 922 | |
6da2f2cc | 923 | return 0; |
bbb44d9f RW |
924 | } |
925 | ||
6e176bf8 RW |
926 | static int pci_pm_resume_early(struct device *dev) |
927 | { | |
76c70cb5 | 928 | if (dev_pm_skip_resume(dev)) |
6e176bf8 RW |
929 | return 0; |
930 | ||
931 | return pm_generic_resume_early(dev); | |
932 | } | |
933 | ||
f6dc1e5e | 934 | static int pci_pm_resume(struct device *dev) |
bbb44d9f | 935 | { |
355a72d7 | 936 | struct pci_dev *pci_dev = to_pci_dev(dev); |
8150f32b | 937 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 938 | |
418e4da3 RW |
939 | /* |
940 | * This is necessary for the suspend error path in which resume is | |
941 | * called without restoring the standard config registers of the device. | |
942 | */ | |
943 | if (pci_dev->state_saved) | |
944 | pci_restore_standard_config(pci_dev); | |
945 | ||
ad8cfa1d | 946 | if (pci_has_legacy_pm_support(pci_dev)) |
f6dc1e5e | 947 | return pci_legacy_resume(dev); |
bb808945 | 948 | |
5294e256 | 949 | pci_pm_default_resume(pci_dev); |
73410429 | 950 | |
5294e256 RW |
951 | if (pm) { |
952 | if (pm->resume) | |
6da2f2cc | 953 | return pm->resume(dev); |
5294e256 RW |
954 | } else { |
955 | pci_pm_reenable_device(pci_dev); | |
956 | } | |
bbb44d9f | 957 | |
6da2f2cc | 958 | return 0; |
bbb44d9f RW |
959 | } |
960 | ||
961 | #else /* !CONFIG_SUSPEND */ | |
962 | ||
963 | #define pci_pm_suspend NULL | |
c4b65157 | 964 | #define pci_pm_suspend_late NULL |
bbb44d9f RW |
965 | #define pci_pm_suspend_noirq NULL |
966 | #define pci_pm_resume NULL | |
6e176bf8 | 967 | #define pci_pm_resume_early NULL |
bbb44d9f RW |
968 | #define pci_pm_resume_noirq NULL |
969 | ||
970 | #endif /* !CONFIG_SUSPEND */ | |
971 | ||
1f112cee | 972 | #ifdef CONFIG_HIBERNATE_CALLBACKS |
bbb44d9f RW |
973 | |
974 | static int pci_pm_freeze(struct device *dev) | |
975 | { | |
976 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
8150f32b | 977 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 978 | |
ad8cfa1d RW |
979 | if (pci_has_legacy_pm_support(pci_dev)) |
980 | return pci_legacy_suspend(dev, PMSG_FREEZE); | |
bb808945 | 981 | |
5294e256 RW |
982 | if (!pm) { |
983 | pci_pm_default_suspend(pci_dev); | |
984 | return 0; | |
bbb44d9f RW |
985 | } |
986 | ||
7cd0602d | 987 | /* |
501debd4 RW |
988 | * Resume all runtime-suspended devices before creating a snapshot |
989 | * image of system memory, because the restore kernel generally cannot | |
990 | * be expected to always handle them consistently and they need to be | |
991 | * put into the runtime-active metastate during system resume anyway, | |
992 | * so it is better to ensure that the state saved in the image will be | |
993 | * always consistent with that. | |
7cd0602d | 994 | */ |
501debd4 RW |
995 | pm_runtime_resume(dev); |
996 | pci_dev->state_saved = false; | |
7cd0602d | 997 | |
5294e256 RW |
998 | if (pm->freeze) { |
999 | int error; | |
1000 | ||
1001 | error = pm->freeze(dev); | |
1002 | suspend_report_result(pm->freeze, error); | |
1003 | if (error) | |
1004 | return error; | |
1005 | } | |
1006 | ||
5294e256 | 1007 | return 0; |
bbb44d9f RW |
1008 | } |
1009 | ||
1010 | static int pci_pm_freeze_noirq(struct device *dev) | |
1011 | { | |
355a72d7 | 1012 | struct pci_dev *pci_dev = to_pci_dev(dev); |
6da2f2cc | 1013 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 1014 | |
bb808945 RW |
1015 | if (pci_has_legacy_pm_support(pci_dev)) |
1016 | return pci_legacy_suspend_late(dev, PMSG_FREEZE); | |
1017 | ||
6da2f2cc | 1018 | if (pm && pm->freeze_noirq) { |
46939f8b RW |
1019 | int error; |
1020 | ||
6da2f2cc BH |
1021 | error = pm->freeze_noirq(dev); |
1022 | suspend_report_result(pm->freeze_noirq, error); | |
46939f8b RW |
1023 | if (error) |
1024 | return error; | |
bbb44d9f RW |
1025 | } |
1026 | ||
46939f8b RW |
1027 | if (!pci_dev->state_saved) |
1028 | pci_save_state(pci_dev); | |
d67e37d7 | 1029 | |
46939f8b RW |
1030 | pci_pm_set_unknown_state(pci_dev); |
1031 | ||
1032 | return 0; | |
bbb44d9f RW |
1033 | } |
1034 | ||
f6dc1e5e | 1035 | static int pci_pm_thaw_noirq(struct device *dev) |
bbb44d9f | 1036 | { |
355a72d7 | 1037 | struct pci_dev *pci_dev = to_pci_dev(dev); |
6da2f2cc | 1038 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
699c1985 | 1039 | |
5839ee73 | 1040 | /* |
89cdbc35 BH |
1041 | * The pm->thaw_noirq() callback assumes the device has been |
1042 | * returned to D0 and its config state has been restored. | |
f2c33cca DC |
1043 | * |
1044 | * In addition, pci_restore_state() restores MSI-X state in MMIO | |
1045 | * space, which requires the device to be in D0, so return it to D0 | |
1046 | * in case the driver's "freeze" callbacks put it into a low-power | |
1047 | * state. | |
5839ee73 RW |
1048 | */ |
1049 | pci_set_power_state(pci_dev, PCI_D0); | |
e60514bd | 1050 | pci_restore_state(pci_dev); |
d67e37d7 | 1051 | |
f2c33cca | 1052 | if (pci_has_legacy_pm_support(pci_dev)) |
89cdbc35 | 1053 | return 0; |
f2c33cca | 1054 | |
6da2f2cc BH |
1055 | if (pm && pm->thaw_noirq) |
1056 | return pm->thaw_noirq(dev); | |
bbb44d9f | 1057 | |
6da2f2cc | 1058 | return 0; |
bbb44d9f RW |
1059 | } |
1060 | ||
f6dc1e5e | 1061 | static int pci_pm_thaw(struct device *dev) |
bbb44d9f | 1062 | { |
355a72d7 | 1063 | struct pci_dev *pci_dev = to_pci_dev(dev); |
8150f32b | 1064 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f RW |
1065 | int error = 0; |
1066 | ||
ad8cfa1d | 1067 | if (pci_has_legacy_pm_support(pci_dev)) |
f6dc1e5e | 1068 | return pci_legacy_resume(dev); |
bb808945 | 1069 | |
5294e256 RW |
1070 | if (pm) { |
1071 | if (pm->thaw) | |
1072 | error = pm->thaw(dev); | |
1073 | } else { | |
1074 | pci_pm_reenable_device(pci_dev); | |
1075 | } | |
bbb44d9f | 1076 | |
4b77b0a2 RW |
1077 | pci_dev->state_saved = false; |
1078 | ||
bbb44d9f RW |
1079 | return error; |
1080 | } | |
1081 | ||
1082 | static int pci_pm_poweroff(struct device *dev) | |
1083 | { | |
355a72d7 | 1084 | struct pci_dev *pci_dev = to_pci_dev(dev); |
8150f32b | 1085 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 1086 | |
ad8cfa1d RW |
1087 | if (pci_has_legacy_pm_support(pci_dev)) |
1088 | return pci_legacy_suspend(dev, PMSG_HIBERNATE); | |
bb808945 | 1089 | |
5294e256 RW |
1090 | if (!pm) { |
1091 | pci_pm_default_suspend(pci_dev); | |
c4b65157 | 1092 | return 0; |
5294e256 RW |
1093 | } |
1094 | ||
7cd0602d | 1095 | /* The reason to do that is the same as in pci_pm_suspend(). */ |
c4b65157 | 1096 | if (!dev_pm_test_driver_flags(dev, DPM_FLAG_SMART_SUSPEND) || |
0c7376ad | 1097 | pci_dev_need_resume(pci_dev)) { |
c4b65157 | 1098 | pm_runtime_resume(dev); |
0c7376ad RW |
1099 | pci_dev->state_saved = false; |
1100 | } else { | |
1101 | pci_dev_adjust_pme(pci_dev); | |
1102 | } | |
7cd0602d | 1103 | |
5294e256 | 1104 | if (pm->poweroff) { |
46939f8b RW |
1105 | int error; |
1106 | ||
ddb7c9d2 RW |
1107 | error = pm->poweroff(dev); |
1108 | suspend_report_result(pm->poweroff, error); | |
46939f8b RW |
1109 | if (error) |
1110 | return error; | |
bbb44d9f RW |
1111 | } |
1112 | ||
46939f8b | 1113 | return 0; |
bbb44d9f | 1114 | } |
c9b9972b | 1115 | |
c4b65157 RW |
1116 | static int pci_pm_poweroff_late(struct device *dev) |
1117 | { | |
fa2bfead | 1118 | if (dev_pm_skip_suspend(dev)) |
c4b65157 | 1119 | return 0; |
699c1985 | 1120 | |
c4b65157 RW |
1121 | pci_fixup_device(pci_fixup_suspend, to_pci_dev(dev)); |
1122 | ||
1123 | return pm_generic_poweroff_late(dev); | |
bbb44d9f RW |
1124 | } |
1125 | ||
1126 | static int pci_pm_poweroff_noirq(struct device *dev) | |
1127 | { | |
46939f8b | 1128 | struct pci_dev *pci_dev = to_pci_dev(dev); |
6da2f2cc | 1129 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 1130 | |
fa2bfead | 1131 | if (dev_pm_skip_suspend(dev)) |
c4b65157 RW |
1132 | return 0; |
1133 | ||
6da2f2cc | 1134 | if (pci_has_legacy_pm_support(pci_dev)) |
bb808945 RW |
1135 | return pci_legacy_suspend_late(dev, PMSG_HIBERNATE); |
1136 | ||
6da2f2cc | 1137 | if (!pm) { |
7d2a01b8 | 1138 | pci_fixup_device(pci_fixup_suspend_late, pci_dev); |
46939f8b | 1139 | return 0; |
7d2a01b8 | 1140 | } |
46939f8b | 1141 | |
6da2f2cc | 1142 | if (pm->poweroff_noirq) { |
46939f8b RW |
1143 | int error; |
1144 | ||
6da2f2cc BH |
1145 | error = pm->poweroff_noirq(dev); |
1146 | suspend_report_result(pm->poweroff_noirq, error); | |
46939f8b RW |
1147 | if (error) |
1148 | return error; | |
bbb44d9f RW |
1149 | } |
1150 | ||
326c1cda | 1151 | if (!pci_dev->state_saved && !pci_has_subordinate(pci_dev)) |
46939f8b RW |
1152 | pci_prepare_to_sleep(pci_dev); |
1153 | ||
0b68c8e2 RW |
1154 | /* |
1155 | * The reason for doing this here is the same as for the analogous code | |
1156 | * in pci_pm_suspend_noirq(). | |
1157 | */ | |
1158 | if (pci_dev->class == PCI_CLASS_SERIAL_USB_EHCI) | |
1159 | pci_write_config_word(pci_dev, PCI_COMMAND, 0); | |
1160 | ||
7d2a01b8 AN |
1161 | pci_fixup_device(pci_fixup_suspend_late, pci_dev); |
1162 | ||
46939f8b | 1163 | return 0; |
bbb44d9f RW |
1164 | } |
1165 | ||
f6dc1e5e | 1166 | static int pci_pm_restore_noirq(struct device *dev) |
bbb44d9f RW |
1167 | { |
1168 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
6da2f2cc | 1169 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
699c1985 | 1170 | |
6cbf8214 | 1171 | pci_pm_default_resume_early(pci_dev); |
3e26c5fe | 1172 | pci_fixup_device(pci_fixup_resume_early, pci_dev); |
aa8c6c93 | 1173 | |
ad8cfa1d | 1174 | if (pci_has_legacy_pm_support(pci_dev)) |
89cdbc35 | 1175 | return 0; |
bb808945 | 1176 | |
6da2f2cc BH |
1177 | if (pm && pm->restore_noirq) |
1178 | return pm->restore_noirq(dev); | |
bbb44d9f | 1179 | |
6da2f2cc | 1180 | return 0; |
bbb44d9f RW |
1181 | } |
1182 | ||
f6dc1e5e | 1183 | static int pci_pm_restore(struct device *dev) |
bbb44d9f RW |
1184 | { |
1185 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
8150f32b | 1186 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
bbb44d9f | 1187 | |
418e4da3 RW |
1188 | /* |
1189 | * This is necessary for the hibernation error path in which restore is | |
1190 | * called without restoring the standard config registers of the device. | |
1191 | */ | |
1192 | if (pci_dev->state_saved) | |
1193 | pci_restore_standard_config(pci_dev); | |
1194 | ||
ad8cfa1d | 1195 | if (pci_has_legacy_pm_support(pci_dev)) |
f6dc1e5e | 1196 | return pci_legacy_resume(dev); |
bb808945 | 1197 | |
5294e256 | 1198 | pci_pm_default_resume(pci_dev); |
73410429 | 1199 | |
5294e256 RW |
1200 | if (pm) { |
1201 | if (pm->restore) | |
6da2f2cc | 1202 | return pm->restore(dev); |
5294e256 RW |
1203 | } else { |
1204 | pci_pm_reenable_device(pci_dev); | |
1205 | } | |
bbb44d9f | 1206 | |
6da2f2cc | 1207 | return 0; |
c8958177 | 1208 | } |
1da177e4 | 1209 | |
1f112cee | 1210 | #else /* !CONFIG_HIBERNATE_CALLBACKS */ |
bbb44d9f RW |
1211 | |
1212 | #define pci_pm_freeze NULL | |
1213 | #define pci_pm_freeze_noirq NULL | |
1214 | #define pci_pm_thaw NULL | |
1215 | #define pci_pm_thaw_noirq NULL | |
1216 | #define pci_pm_poweroff NULL | |
c4b65157 | 1217 | #define pci_pm_poweroff_late NULL |
bbb44d9f RW |
1218 | #define pci_pm_poweroff_noirq NULL |
1219 | #define pci_pm_restore NULL | |
1220 | #define pci_pm_restore_noirq NULL | |
1221 | ||
1f112cee | 1222 | #endif /* !CONFIG_HIBERNATE_CALLBACKS */ |
bbb44d9f | 1223 | |
fbb988be | 1224 | #ifdef CONFIG_PM |
6cbf8214 RW |
1225 | |
1226 | static int pci_pm_runtime_suspend(struct device *dev) | |
1227 | { | |
1228 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
1229 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; | |
1230 | pci_power_t prev = pci_dev->current_state; | |
1231 | int error; | |
1232 | ||
967577b0 | 1233 | /* |
5775b843 RW |
1234 | * If pci_dev->driver is not set (unbound), we leave the device in D0, |
1235 | * but it may go to D3cold when the bridge above it runtime suspends. | |
1236 | * Save its config space in case that happens. | |
967577b0 | 1237 | */ |
5775b843 RW |
1238 | if (!pci_dev->driver) { |
1239 | pci_save_state(pci_dev); | |
967577b0 | 1240 | return 0; |
5775b843 | 1241 | } |
967577b0 | 1242 | |
82fee4d6 | 1243 | pci_dev->state_saved = false; |
c5eb1190 JN |
1244 | if (pm && pm->runtime_suspend) { |
1245 | error = pm->runtime_suspend(dev); | |
06bf403d ID |
1246 | /* |
1247 | * -EBUSY and -EAGAIN is used to request the runtime PM core | |
1248 | * to schedule a new suspend, so log the event only with debug | |
1249 | * log level. | |
1250 | */ | |
c5eb1190 | 1251 | if (error == -EBUSY || error == -EAGAIN) { |
6941a0c2 | 1252 | pci_dbg(pci_dev, "can't suspend now (%ps returned %d)\n", |
06bf403d | 1253 | pm->runtime_suspend, error); |
c5eb1190 JN |
1254 | return error; |
1255 | } else if (error) { | |
6941a0c2 | 1256 | pci_err(pci_dev, "can't suspend (%ps returned %d)\n", |
06bf403d | 1257 | pm->runtime_suspend, error); |
c5eb1190 JN |
1258 | return error; |
1259 | } | |
06bf403d | 1260 | } |
6cbf8214 RW |
1261 | |
1262 | pci_fixup_device(pci_fixup_suspend, pci_dev); | |
1263 | ||
c5eb1190 JN |
1264 | if (pm && pm->runtime_suspend |
1265 | && !pci_dev->state_saved && pci_dev->current_state != PCI_D0 | |
6cbf8214 | 1266 | && pci_dev->current_state != PCI_UNKNOWN) { |
12bcae44 BH |
1267 | pci_WARN_ONCE(pci_dev, pci_dev->current_state != prev, |
1268 | "PCI PM: State of device not saved by %pS\n", | |
1269 | pm->runtime_suspend); | |
6cbf8214 RW |
1270 | return 0; |
1271 | } | |
1272 | ||
42eca230 | 1273 | if (!pci_dev->state_saved) { |
6cbf8214 | 1274 | pci_save_state(pci_dev); |
42eca230 DA |
1275 | pci_finish_runtime_suspend(pci_dev); |
1276 | } | |
6cbf8214 RW |
1277 | |
1278 | return 0; | |
1279 | } | |
1280 | ||
1281 | static int pci_pm_runtime_resume(struct device *dev) | |
1282 | { | |
1283 | struct pci_dev *pci_dev = to_pci_dev(dev); | |
1284 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; | |
ad9001f2 | 1285 | pci_power_t prev_state = pci_dev->current_state; |
6da2f2cc | 1286 | int error = 0; |
6cbf8214 | 1287 | |
967577b0 | 1288 | /* |
5775b843 RW |
1289 | * Restoring config space is necessary even if the device is not bound |
1290 | * to a driver because although we left it in D0, it may have gone to | |
1291 | * D3cold when the bridge above it runtime suspended. | |
967577b0 | 1292 | */ |
5775b843 RW |
1293 | pci_restore_standard_config(pci_dev); |
1294 | ||
967577b0 HY |
1295 | if (!pci_dev->driver) |
1296 | return 0; | |
1297 | ||
db288c9c | 1298 | pci_fixup_device(pci_fixup_resume_early, pci_dev); |
f7b32a86 | 1299 | pci_pm_default_resume(pci_dev); |
6cbf8214 | 1300 | |
ad9001f2 MW |
1301 | if (prev_state == PCI_D3cold) |
1302 | pci_bridge_wait_for_secondary_bus(pci_dev); | |
1303 | ||
c5eb1190 | 1304 | if (pm && pm->runtime_resume) |
6da2f2cc | 1305 | error = pm->runtime_resume(dev); |
448bd857 HY |
1306 | |
1307 | pci_dev->runtime_d3cold = false; | |
1308 | ||
6da2f2cc | 1309 | return error; |
6cbf8214 RW |
1310 | } |
1311 | ||
1312 | static int pci_pm_runtime_idle(struct device *dev) | |
1313 | { | |
967577b0 | 1314 | struct pci_dev *pci_dev = to_pci_dev(dev); |
6cbf8214 RW |
1315 | const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL; |
1316 | ||
967577b0 HY |
1317 | /* |
1318 | * If pci_dev->driver is not set (unbound), the device should | |
1319 | * always remain in D0 regardless of the runtime PM status | |
1320 | */ | |
1321 | if (!pci_dev->driver) | |
45f0a85c | 1322 | return 0; |
967577b0 | 1323 | |
6cbf8214 RW |
1324 | if (!pm) |
1325 | return -ENOSYS; | |
1326 | ||
45f0a85c | 1327 | if (pm->runtime_idle) |
6da2f2cc | 1328 | return pm->runtime_idle(dev); |
6cbf8214 | 1329 | |
6da2f2cc | 1330 | return 0; |
6cbf8214 RW |
1331 | } |
1332 | ||
f91da04d | 1333 | static const struct dev_pm_ops pci_dev_pm_ops = { |
adf09493 | 1334 | .prepare = pci_pm_prepare, |
2cef548a | 1335 | .complete = pci_pm_complete, |
adf09493 | 1336 | .suspend = pci_pm_suspend, |
c4b65157 | 1337 | .suspend_late = pci_pm_suspend_late, |
adf09493 | 1338 | .resume = pci_pm_resume, |
6e176bf8 | 1339 | .resume_early = pci_pm_resume_early, |
adf09493 RW |
1340 | .freeze = pci_pm_freeze, |
1341 | .thaw = pci_pm_thaw, | |
1342 | .poweroff = pci_pm_poweroff, | |
c4b65157 | 1343 | .poweroff_late = pci_pm_poweroff_late, |
adf09493 | 1344 | .restore = pci_pm_restore, |
bbb44d9f RW |
1345 | .suspend_noirq = pci_pm_suspend_noirq, |
1346 | .resume_noirq = pci_pm_resume_noirq, | |
1347 | .freeze_noirq = pci_pm_freeze_noirq, | |
1348 | .thaw_noirq = pci_pm_thaw_noirq, | |
1349 | .poweroff_noirq = pci_pm_poweroff_noirq, | |
1350 | .restore_noirq = pci_pm_restore_noirq, | |
6cbf8214 RW |
1351 | .runtime_suspend = pci_pm_runtime_suspend, |
1352 | .runtime_resume = pci_pm_runtime_resume, | |
1353 | .runtime_idle = pci_pm_runtime_idle, | |
bbb44d9f RW |
1354 | }; |
1355 | ||
adf09493 | 1356 | #define PCI_PM_OPS_PTR (&pci_dev_pm_ops) |
bbb44d9f | 1357 | |
fbb988be RW |
1358 | #else /* !CONFIG_PM */ |
1359 | ||
1360 | #define pci_pm_runtime_suspend NULL | |
1361 | #define pci_pm_runtime_resume NULL | |
1362 | #define pci_pm_runtime_idle NULL | |
bbb44d9f RW |
1363 | |
1364 | #define PCI_PM_OPS_PTR NULL | |
1365 | ||
fbb988be | 1366 | #endif /* !CONFIG_PM */ |
bbb44d9f | 1367 | |
1da177e4 | 1368 | /** |
863b18f4 | 1369 | * __pci_register_driver - register a new pci driver |
1da177e4 | 1370 | * @drv: the driver structure to register |
863b18f4 | 1371 | * @owner: owner module of drv |
f95d882d | 1372 | * @mod_name: module name string |
f7625980 | 1373 | * |
1da177e4 | 1374 | * Adds the driver structure to the list of registered drivers. |
f7625980 BH |
1375 | * Returns a negative value on error, otherwise 0. |
1376 | * If no error occurred, the driver remains registered even if | |
1da177e4 LT |
1377 | * no device was claimed during registration. |
1378 | */ | |
725522b5 GKH |
1379 | int __pci_register_driver(struct pci_driver *drv, struct module *owner, |
1380 | const char *mod_name) | |
1da177e4 | 1381 | { |
1da177e4 LT |
1382 | /* initialize common driver fields */ |
1383 | drv->driver.name = drv->name; | |
1384 | drv->driver.bus = &pci_bus_type; | |
863b18f4 | 1385 | drv->driver.owner = owner; |
725522b5 | 1386 | drv->driver.mod_name = mod_name; |
92d50fc1 | 1387 | drv->driver.groups = drv->groups; |
50b00755 | 1388 | |
75865858 GKH |
1389 | spin_lock_init(&drv->dynids.lock); |
1390 | INIT_LIST_HEAD(&drv->dynids.list); | |
1da177e4 LT |
1391 | |
1392 | /* register with core */ | |
bfb09a86 | 1393 | return driver_register(&drv->driver); |
1da177e4 | 1394 | } |
b7fe9434 | 1395 | EXPORT_SYMBOL(__pci_register_driver); |
1da177e4 LT |
1396 | |
1397 | /** | |
1398 | * pci_unregister_driver - unregister a pci driver | |
1399 | * @drv: the driver structure to unregister | |
f7625980 | 1400 | * |
1da177e4 LT |
1401 | * Deletes the driver structure from the list of registered PCI drivers, |
1402 | * gives it a chance to clean up by calling its remove() function for | |
1403 | * each device it was responsible for, and marks those devices as | |
1404 | * driverless. | |
1405 | */ | |
1406 | ||
3c78bc61 | 1407 | void pci_unregister_driver(struct pci_driver *drv) |
1da177e4 LT |
1408 | { |
1409 | driver_unregister(&drv->driver); | |
1410 | pci_free_dynids(drv); | |
1411 | } | |
b7fe9434 | 1412 | EXPORT_SYMBOL(pci_unregister_driver); |
1da177e4 LT |
1413 | |
1414 | static struct pci_driver pci_compat_driver = { | |
1415 | .name = "compat" | |
1416 | }; | |
1417 | ||
1418 | /** | |
1419 | * pci_dev_driver - get the pci_driver of a device | |
1420 | * @dev: the device to query | |
1421 | * | |
f7625980 | 1422 | * Returns the appropriate pci_driver structure or %NULL if there is no |
1da177e4 LT |
1423 | * registered driver for the device. |
1424 | */ | |
3c78bc61 | 1425 | struct pci_driver *pci_dev_driver(const struct pci_dev *dev) |
1da177e4 LT |
1426 | { |
1427 | if (dev->driver) | |
1428 | return dev->driver; | |
1429 | else { | |
1430 | int i; | |
3c78bc61 | 1431 | for (i = 0; i <= PCI_ROM_RESOURCE; i++) |
1da177e4 LT |
1432 | if (dev->resource[i].flags & IORESOURCE_BUSY) |
1433 | return &pci_compat_driver; | |
1434 | } | |
1435 | return NULL; | |
1436 | } | |
b7fe9434 | 1437 | EXPORT_SYMBOL(pci_dev_driver); |
1da177e4 LT |
1438 | |
1439 | /** | |
1440 | * pci_bus_match - Tell if a PCI device structure has a matching PCI device id structure | |
1da177e4 | 1441 | * @dev: the PCI device structure to match against |
8f7020d3 | 1442 | * @drv: the device driver to search for matching PCI device id structures |
f7625980 | 1443 | * |
1da177e4 | 1444 | * Used by a driver to check whether a PCI device present in the |
8f7020d3 | 1445 | * system is in its list of supported devices. Returns the matching |
1da177e4 LT |
1446 | * pci_device_id structure or %NULL if there is no match. |
1447 | */ | |
75865858 | 1448 | static int pci_bus_match(struct device *dev, struct device_driver *drv) |
1da177e4 | 1449 | { |
75865858 | 1450 | struct pci_dev *pci_dev = to_pci_dev(dev); |
58d9a38f | 1451 | struct pci_driver *pci_drv; |
1da177e4 LT |
1452 | const struct pci_device_id *found_id; |
1453 | ||
58d9a38f YL |
1454 | if (!pci_dev->match_driver) |
1455 | return 0; | |
1456 | ||
1457 | pci_drv = to_pci_driver(drv); | |
75865858 | 1458 | found_id = pci_match_device(pci_drv, pci_dev); |
1da177e4 LT |
1459 | if (found_id) |
1460 | return 1; | |
1461 | ||
75865858 | 1462 | return 0; |
1da177e4 LT |
1463 | } |
1464 | ||
1465 | /** | |
1466 | * pci_dev_get - increments the reference count of the pci device structure | |
1467 | * @dev: the device being referenced | |
1468 | * | |
1469 | * Each live reference to a device should be refcounted. | |
1470 | * | |
1471 | * Drivers for PCI devices should normally record such references in | |
1472 | * their probe() methods, when they bind to a device, and release | |
1473 | * them by calling pci_dev_put(), in their disconnect() methods. | |
1474 | * | |
1475 | * A pointer to the device with the incremented reference counter is returned. | |
1476 | */ | |
1477 | struct pci_dev *pci_dev_get(struct pci_dev *dev) | |
1478 | { | |
1479 | if (dev) | |
1480 | get_device(&dev->dev); | |
1481 | return dev; | |
1482 | } | |
b7fe9434 | 1483 | EXPORT_SYMBOL(pci_dev_get); |
1da177e4 LT |
1484 | |
1485 | /** | |
1486 | * pci_dev_put - release a use of the pci device structure | |
1487 | * @dev: device that's been disconnected | |
1488 | * | |
1489 | * Must be called when a user of a device is finished with it. When the last | |
1490 | * user of the device calls this function, the memory of the device is freed. | |
1491 | */ | |
1492 | void pci_dev_put(struct pci_dev *dev) | |
1493 | { | |
1494 | if (dev) | |
1495 | put_device(&dev->dev); | |
1496 | } | |
b7fe9434 | 1497 | EXPORT_SYMBOL(pci_dev_put); |
1da177e4 | 1498 | |
8ccc9aa1 BP |
1499 | static int pci_uevent(struct device *dev, struct kobj_uevent_env *env) |
1500 | { | |
1501 | struct pci_dev *pdev; | |
1502 | ||
1503 | if (!dev) | |
1504 | return -ENODEV; | |
1505 | ||
1506 | pdev = to_pci_dev(dev); | |
8ccc9aa1 BP |
1507 | |
1508 | if (add_uevent_var(env, "PCI_CLASS=%04X", pdev->class)) | |
1509 | return -ENOMEM; | |
1510 | ||
1511 | if (add_uevent_var(env, "PCI_ID=%04X:%04X", pdev->vendor, pdev->device)) | |
1512 | return -ENOMEM; | |
1513 | ||
1514 | if (add_uevent_var(env, "PCI_SUBSYS_ID=%04X:%04X", pdev->subsystem_vendor, | |
1515 | pdev->subsystem_device)) | |
1516 | return -ENOMEM; | |
1517 | ||
1518 | if (add_uevent_var(env, "PCI_SLOT_NAME=%s", pci_name(pdev))) | |
1519 | return -ENOMEM; | |
1520 | ||
145b3fe5 | 1521 | if (add_uevent_var(env, "MODALIAS=pci:v%08Xd%08Xsv%08Xsd%08Xbc%02Xsc%02Xi%02X", |
8ccc9aa1 BP |
1522 | pdev->vendor, pdev->device, |
1523 | pdev->subsystem_vendor, pdev->subsystem_device, | |
1524 | (u8)(pdev->class >> 16), (u8)(pdev->class >> 8), | |
1525 | (u8)(pdev->class))) | |
1526 | return -ENOMEM; | |
efdd4070 | 1527 | |
8ccc9aa1 BP |
1528 | return 0; |
1529 | } | |
1530 | ||
2e28bc84 | 1531 | #if defined(CONFIG_PCIEPORTBUS) || defined(CONFIG_EEH) |
3ecac020 ME |
1532 | /** |
1533 | * pci_uevent_ers - emit a uevent during recovery path of PCI device | |
1534 | * @pdev: PCI device undergoing error recovery | |
1535 | * @err_type: type of error event | |
1536 | */ | |
1537 | void pci_uevent_ers(struct pci_dev *pdev, enum pci_ers_result err_type) | |
1538 | { | |
1539 | int idx = 0; | |
1540 | char *envp[3]; | |
1541 | ||
1542 | switch (err_type) { | |
1543 | case PCI_ERS_RESULT_NONE: | |
1544 | case PCI_ERS_RESULT_CAN_RECOVER: | |
1545 | envp[idx++] = "ERROR_EVENT=BEGIN_RECOVERY"; | |
1546 | envp[idx++] = "DEVICE_ONLINE=0"; | |
1547 | break; | |
1548 | case PCI_ERS_RESULT_RECOVERED: | |
1549 | envp[idx++] = "ERROR_EVENT=SUCCESSFUL_RECOVERY"; | |
1550 | envp[idx++] = "DEVICE_ONLINE=1"; | |
1551 | break; | |
1552 | case PCI_ERS_RESULT_DISCONNECT: | |
1553 | envp[idx++] = "ERROR_EVENT=FAILED_RECOVERY"; | |
1554 | envp[idx++] = "DEVICE_ONLINE=0"; | |
1555 | break; | |
1556 | default: | |
1557 | break; | |
1558 | } | |
1559 | ||
1560 | if (idx > 0) { | |
1561 | envp[idx++] = NULL; | |
1562 | kobject_uevent_env(&pdev->dev.kobj, KOBJ_CHANGE, envp); | |
1563 | } | |
1564 | } | |
1565 | #endif | |
1566 | ||
02e0bea6 PS |
1567 | static int pci_bus_num_vf(struct device *dev) |
1568 | { | |
1569 | return pci_num_vf(to_pci_dev(dev)); | |
1570 | } | |
1571 | ||
07397df2 NG |
1572 | /** |
1573 | * pci_dma_configure - Setup DMA configuration | |
1574 | * @dev: ptr to dev structure | |
1575 | * | |
1576 | * Function to update PCI devices's DMA configuration using the same | |
1577 | * info from the OF node or ACPI node of host bridge's parent (if any). | |
1578 | */ | |
1579 | static int pci_dma_configure(struct device *dev) | |
1580 | { | |
1581 | struct device *bridge; | |
1582 | int ret = 0; | |
1583 | ||
1584 | bridge = pci_get_host_bridge_device(to_pci_dev(dev)); | |
1585 | ||
1586 | if (IS_ENABLED(CONFIG_OF) && bridge->parent && | |
1587 | bridge->parent->of_node) { | |
3d6ce86e | 1588 | ret = of_dma_configure(dev, bridge->parent->of_node, true); |
07397df2 NG |
1589 | } else if (has_acpi_companion(bridge)) { |
1590 | struct acpi_device *adev = to_acpi_device_node(bridge->fwnode); | |
07397df2 | 1591 | |
e5361ca2 | 1592 | ret = acpi_dma_configure(dev, acpi_get_dma_attr(adev)); |
07397df2 NG |
1593 | } |
1594 | ||
1595 | pci_put_host_bridge_device(bridge); | |
1596 | return ret; | |
1597 | } | |
1598 | ||
1da177e4 LT |
1599 | struct bus_type pci_bus_type = { |
1600 | .name = "pci", | |
1601 | .match = pci_bus_match, | |
312c004d | 1602 | .uevent = pci_uevent, |
b15d686a RK |
1603 | .probe = pci_device_probe, |
1604 | .remove = pci_device_remove, | |
cbd69dbb | 1605 | .shutdown = pci_device_shutdown, |
5136b2da | 1606 | .dev_groups = pci_dev_groups, |
0f49ba55 | 1607 | .bus_groups = pci_bus_groups, |
2229c1fb | 1608 | .drv_groups = pci_drv_groups, |
bbb44d9f | 1609 | .pm = PCI_PM_OPS_PTR, |
02e0bea6 | 1610 | .num_vf = pci_bus_num_vf, |
07397df2 | 1611 | .dma_configure = pci_dma_configure, |
1da177e4 | 1612 | }; |
b7fe9434 | 1613 | EXPORT_SYMBOL(pci_bus_type); |
1da177e4 | 1614 | |
c6c889d9 BH |
1615 | #ifdef CONFIG_PCIEPORTBUS |
1616 | static int pcie_port_bus_match(struct device *dev, struct device_driver *drv) | |
1617 | { | |
1618 | struct pcie_device *pciedev; | |
1619 | struct pcie_port_service_driver *driver; | |
1620 | ||
1621 | if (drv->bus != &pcie_port_bus_type || dev->bus != &pcie_port_bus_type) | |
1622 | return 0; | |
1623 | ||
1624 | pciedev = to_pcie_device(dev); | |
1625 | driver = to_service_driver(drv); | |
1626 | ||
1627 | if (driver->service != pciedev->service) | |
1628 | return 0; | |
1629 | ||
1630 | if (driver->port_type != PCIE_ANY_PORT && | |
1631 | driver->port_type != pci_pcie_type(pciedev->port)) | |
1632 | return 0; | |
1633 | ||
1634 | return 1; | |
1635 | } | |
1636 | ||
1637 | struct bus_type pcie_port_bus_type = { | |
1638 | .name = "pci_express", | |
1639 | .match = pcie_port_bus_match, | |
1640 | }; | |
1641 | EXPORT_SYMBOL_GPL(pcie_port_bus_type); | |
1642 | #endif | |
1643 | ||
1da177e4 LT |
1644 | static int __init pci_driver_init(void) |
1645 | { | |
c6c889d9 BH |
1646 | int ret; |
1647 | ||
1648 | ret = bus_register(&pci_bus_type); | |
1649 | if (ret) | |
1650 | return ret; | |
1651 | ||
1652 | #ifdef CONFIG_PCIEPORTBUS | |
1653 | ret = bus_register(&pcie_port_bus_type); | |
1654 | if (ret) | |
1655 | return ret; | |
1656 | #endif | |
a8651194 | 1657 | dma_debug_add_bus(&pci_bus_type); |
c6c889d9 | 1658 | return 0; |
1da177e4 | 1659 | } |
1da177e4 | 1660 | postcore_initcall(pci_driver_init); |