2 * Support PCI/PCIe on PowerNV platforms
4 * Currently supports only P5IOC2
6 * Copyright 2011 Benjamin Herrenschmidt, IBM Corp.
8 * This program is free software; you can redistribute it and/or
9 * modify it under the terms of the GNU General Public License
10 * as published by the Free Software Foundation; either version
11 * 2 of the License, or (at your option) any later version.
14 #include <linux/kernel.h>
15 #include <linux/pci.h>
16 #include <linux/delay.h>
17 #include <linux/string.h>
18 #include <linux/init.h>
19 #include <linux/bootmem.h>
20 #include <linux/irq.h>
22 #include <linux/msi.h>
23 #include <linux/iommu.h>
25 #include <asm/sections.h>
28 #include <asm/pci-bridge.h>
29 #include <asm/machdep.h>
30 #include <asm/msi_bitmap.h>
31 #include <asm/ppc-pci.h>
33 #include <asm/iommu.h>
35 #include <asm/firmware.h>
36 #include <asm/eeh_event.h>
43 #define PCI_RESET_DELAY_US 3000000
45 #define cfg_dbg(fmt...) do { } while(0)
46 //#define cfg_dbg(fmt...) printk(fmt)
49 static int pnv_msi_check_device(struct pci_dev
* pdev
, int nvec
, int type
)
51 struct pci_controller
*hose
= pci_bus_to_host(pdev
->bus
);
52 struct pnv_phb
*phb
= hose
->private_data
;
53 struct pci_dn
*pdn
= pci_get_pdn(pdev
);
55 if (pdn
&& pdn
->force_32bit_msi
&& !phb
->msi32_support
)
58 return (phb
&& phb
->msi_bmp
.bitmap
) ? 0 : -ENODEV
;
61 static int pnv_setup_msi_irqs(struct pci_dev
*pdev
, int nvec
, int type
)
63 struct pci_controller
*hose
= pci_bus_to_host(pdev
->bus
);
64 struct pnv_phb
*phb
= hose
->private_data
;
65 struct msi_desc
*entry
;
74 list_for_each_entry(entry
, &pdev
->msi_list
, list
) {
75 if (!entry
->msi_attrib
.is_64
&& !phb
->msi32_support
) {
76 pr_warn("%s: Supports only 64-bit MSIs\n",
80 hwirq
= msi_bitmap_alloc_hwirqs(&phb
->msi_bmp
, 1);
82 pr_warn("%s: Failed to find a free MSI\n",
86 virq
= irq_create_mapping(NULL
, phb
->msi_base
+ hwirq
);
88 pr_warn("%s: Failed to map MSI to linux irq\n",
90 msi_bitmap_free_hwirqs(&phb
->msi_bmp
, hwirq
, 1);
93 rc
= phb
->msi_setup(phb
, pdev
, phb
->msi_base
+ hwirq
,
94 virq
, entry
->msi_attrib
.is_64
, &msg
);
96 pr_warn("%s: Failed to setup MSI\n", pci_name(pdev
));
97 irq_dispose_mapping(virq
);
98 msi_bitmap_free_hwirqs(&phb
->msi_bmp
, hwirq
, 1);
101 irq_set_msi_desc(virq
, entry
);
102 write_msi_msg(virq
, &msg
);
107 static void pnv_teardown_msi_irqs(struct pci_dev
*pdev
)
109 struct pci_controller
*hose
= pci_bus_to_host(pdev
->bus
);
110 struct pnv_phb
*phb
= hose
->private_data
;
111 struct msi_desc
*entry
;
116 list_for_each_entry(entry
, &pdev
->msi_list
, list
) {
117 if (entry
->irq
== NO_IRQ
)
119 irq_set_msi_desc(entry
->irq
, NULL
);
120 msi_bitmap_free_hwirqs(&phb
->msi_bmp
,
121 virq_to_hw(entry
->irq
) - phb
->msi_base
, 1);
122 irq_dispose_mapping(entry
->irq
);
125 #endif /* CONFIG_PCI_MSI */
127 static void pnv_pci_dump_p7ioc_diag_data(struct pci_controller
*hose
,
128 struct OpalIoPhbErrorCommon
*common
)
130 struct OpalIoP7IOCPhbErrorData
*data
;
133 data
= (struct OpalIoP7IOCPhbErrorData
*)common
;
134 pr_info("P7IOC PHB#%d Diag-data (Version: %d)\n",
135 hose
->global_number
, common
->version
);
138 pr_info("brdgCtl: %08x\n",
140 if (data
->portStatusReg
|| data
->rootCmplxStatus
||
141 data
->busAgentStatus
)
142 pr_info("UtlSts: %08x %08x %08x\n",
143 data
->portStatusReg
, data
->rootCmplxStatus
,
144 data
->busAgentStatus
);
145 if (data
->deviceStatus
|| data
->slotStatus
||
146 data
->linkStatus
|| data
->devCmdStatus
||
148 pr_info("RootSts: %08x %08x %08x %08x %08x\n",
149 data
->deviceStatus
, data
->slotStatus
,
150 data
->linkStatus
, data
->devCmdStatus
,
152 if (data
->rootErrorStatus
|| data
->uncorrErrorStatus
||
153 data
->corrErrorStatus
)
154 pr_info("RootErrSts: %08x %08x %08x\n",
155 data
->rootErrorStatus
, data
->uncorrErrorStatus
,
156 data
->corrErrorStatus
);
157 if (data
->tlpHdr1
|| data
->tlpHdr2
||
158 data
->tlpHdr3
|| data
->tlpHdr4
)
159 pr_info("RootErrLog: %08x %08x %08x %08x\n",
160 data
->tlpHdr1
, data
->tlpHdr2
,
161 data
->tlpHdr3
, data
->tlpHdr4
);
162 if (data
->sourceId
|| data
->errorClass
||
164 pr_info("RootErrLog1: %08x %016llx %016llx\n",
165 data
->sourceId
, data
->errorClass
,
167 if (data
->p7iocPlssr
|| data
->p7iocCsr
)
168 pr_info("PhbSts: %016llx %016llx\n",
169 data
->p7iocPlssr
, data
->p7iocCsr
);
171 pr_info("Lem: %016llx %016llx %016llx\n",
172 data
->lemFir
, data
->lemErrorMask
,
174 if (data
->phbErrorStatus
)
175 pr_info("PhbErr: %016llx %016llx %016llx %016llx\n",
176 data
->phbErrorStatus
, data
->phbFirstErrorStatus
,
177 data
->phbErrorLog0
, data
->phbErrorLog1
);
178 if (data
->mmioErrorStatus
)
179 pr_info("OutErr: %016llx %016llx %016llx %016llx\n",
180 data
->mmioErrorStatus
, data
->mmioFirstErrorStatus
,
181 data
->mmioErrorLog0
, data
->mmioErrorLog1
);
182 if (data
->dma0ErrorStatus
)
183 pr_info("InAErr: %016llx %016llx %016llx %016llx\n",
184 data
->dma0ErrorStatus
, data
->dma0FirstErrorStatus
,
185 data
->dma0ErrorLog0
, data
->dma0ErrorLog1
);
186 if (data
->dma1ErrorStatus
)
187 pr_info("InBErr: %016llx %016llx %016llx %016llx\n",
188 data
->dma1ErrorStatus
, data
->dma1FirstErrorStatus
,
189 data
->dma1ErrorLog0
, data
->dma1ErrorLog1
);
191 for (i
= 0; i
< OPAL_P7IOC_NUM_PEST_REGS
; i
++) {
192 if ((data
->pestA
[i
] >> 63) == 0 &&
193 (data
->pestB
[i
] >> 63) == 0)
196 pr_info("PE[%3d] A/B: %016llx %016llx\n",
197 i
, data
->pestA
[i
], data
->pestB
[i
]);
201 static void pnv_pci_dump_phb3_diag_data(struct pci_controller
*hose
,
202 struct OpalIoPhbErrorCommon
*common
)
204 struct OpalIoPhb3ErrorData
*data
;
207 data
= (struct OpalIoPhb3ErrorData
*)common
;
208 pr_info("PHB3 PHB#%d Diag-data (Version: %d)\n",
209 hose
->global_number
, be32_to_cpu(common
->version
));
211 pr_info("brdgCtl: %08x\n",
212 be32_to_cpu(data
->brdgCtl
));
213 if (data
->portStatusReg
|| data
->rootCmplxStatus
||
214 data
->busAgentStatus
)
215 pr_info("UtlSts: %08x %08x %08x\n",
216 be32_to_cpu(data
->portStatusReg
),
217 be32_to_cpu(data
->rootCmplxStatus
),
218 be32_to_cpu(data
->busAgentStatus
));
219 if (data
->deviceStatus
|| data
->slotStatus
||
220 data
->linkStatus
|| data
->devCmdStatus
||
222 pr_info("RootSts: %08x %08x %08x %08x %08x\n",
223 be32_to_cpu(data
->deviceStatus
),
224 be32_to_cpu(data
->slotStatus
),
225 be32_to_cpu(data
->linkStatus
),
226 be32_to_cpu(data
->devCmdStatus
),
227 be32_to_cpu(data
->devSecStatus
));
228 if (data
->rootErrorStatus
|| data
->uncorrErrorStatus
||
229 data
->corrErrorStatus
)
230 pr_info("RootErrSts: %08x %08x %08x\n",
231 be32_to_cpu(data
->rootErrorStatus
),
232 be32_to_cpu(data
->uncorrErrorStatus
),
233 be32_to_cpu(data
->corrErrorStatus
));
234 if (data
->tlpHdr1
|| data
->tlpHdr2
||
235 data
->tlpHdr3
|| data
->tlpHdr4
)
236 pr_info("RootErrLog: %08x %08x %08x %08x\n",
237 be32_to_cpu(data
->tlpHdr1
),
238 be32_to_cpu(data
->tlpHdr2
),
239 be32_to_cpu(data
->tlpHdr3
),
240 be32_to_cpu(data
->tlpHdr4
));
241 if (data
->sourceId
|| data
->errorClass
||
243 pr_info("RootErrLog1: %08x %016llx %016llx\n",
244 be32_to_cpu(data
->sourceId
),
245 be64_to_cpu(data
->errorClass
),
246 be64_to_cpu(data
->correlator
));
248 pr_info("nFir: %016llx %016llx %016llx\n",
249 be64_to_cpu(data
->nFir
),
250 be64_to_cpu(data
->nFirMask
),
251 be64_to_cpu(data
->nFirWOF
));
252 if (data
->phbPlssr
|| data
->phbCsr
)
253 pr_info("PhbSts: %016llx %016llx\n",
254 be64_to_cpu(data
->phbPlssr
),
255 be64_to_cpu(data
->phbCsr
));
257 pr_info("Lem: %016llx %016llx %016llx\n",
258 be64_to_cpu(data
->lemFir
),
259 be64_to_cpu(data
->lemErrorMask
),
260 be64_to_cpu(data
->lemWOF
));
261 if (data
->phbErrorStatus
)
262 pr_info("PhbErr: %016llx %016llx %016llx %016llx\n",
263 be64_to_cpu(data
->phbErrorStatus
),
264 be64_to_cpu(data
->phbFirstErrorStatus
),
265 be64_to_cpu(data
->phbErrorLog0
),
266 be64_to_cpu(data
->phbErrorLog1
));
267 if (data
->mmioErrorStatus
)
268 pr_info("OutErr: %016llx %016llx %016llx %016llx\n",
269 be64_to_cpu(data
->mmioErrorStatus
),
270 be64_to_cpu(data
->mmioFirstErrorStatus
),
271 be64_to_cpu(data
->mmioErrorLog0
),
272 be64_to_cpu(data
->mmioErrorLog1
));
273 if (data
->dma0ErrorStatus
)
274 pr_info("InAErr: %016llx %016llx %016llx %016llx\n",
275 be64_to_cpu(data
->dma0ErrorStatus
),
276 be64_to_cpu(data
->dma0FirstErrorStatus
),
277 be64_to_cpu(data
->dma0ErrorLog0
),
278 be64_to_cpu(data
->dma0ErrorLog1
));
279 if (data
->dma1ErrorStatus
)
280 pr_info("InBErr: %016llx %016llx %016llx %016llx\n",
281 be64_to_cpu(data
->dma1ErrorStatus
),
282 be64_to_cpu(data
->dma1FirstErrorStatus
),
283 be64_to_cpu(data
->dma1ErrorLog0
),
284 be64_to_cpu(data
->dma1ErrorLog1
));
286 for (i
= 0; i
< OPAL_PHB3_NUM_PEST_REGS
; i
++) {
287 if ((be64_to_cpu(data
->pestA
[i
]) >> 63) == 0 &&
288 (be64_to_cpu(data
->pestB
[i
]) >> 63) == 0)
291 pr_info("PE[%3d] A/B: %016llx %016llx\n",
292 i
, be64_to_cpu(data
->pestA
[i
]),
293 be64_to_cpu(data
->pestB
[i
]));
297 void pnv_pci_dump_phb_diag_data(struct pci_controller
*hose
,
298 unsigned char *log_buff
)
300 struct OpalIoPhbErrorCommon
*common
;
302 if (!hose
|| !log_buff
)
305 common
= (struct OpalIoPhbErrorCommon
*)log_buff
;
306 switch (be32_to_cpu(common
->ioType
)) {
307 case OPAL_PHB_ERROR_DATA_TYPE_P7IOC
:
308 pnv_pci_dump_p7ioc_diag_data(hose
, common
);
310 case OPAL_PHB_ERROR_DATA_TYPE_PHB3
:
311 pnv_pci_dump_phb3_diag_data(hose
, common
);
314 pr_warn("%s: Unrecognized ioType %d\n",
315 __func__
, be32_to_cpu(common
->ioType
));
319 static void pnv_pci_handle_eeh_config(struct pnv_phb
*phb
, u32 pe_no
)
321 unsigned long flags
, rc
;
324 spin_lock_irqsave(&phb
->lock
, flags
);
326 rc
= opal_pci_get_phb_diag_data2(phb
->opal_id
, phb
->diag
.blob
,
327 PNV_PCI_DIAG_BUF_SIZE
);
328 has_diag
= (rc
== OPAL_SUCCESS
);
330 rc
= opal_pci_eeh_freeze_clear(phb
->opal_id
, pe_no
,
331 OPAL_EEH_ACTION_CLEAR_FREEZE_ALL
);
333 pr_warning("PCI %d: Failed to clear EEH freeze state"
334 " for PE#%d, err %ld\n",
335 phb
->hose
->global_number
, pe_no
, rc
);
337 /* For now, let's only display the diag buffer when we fail to clear
338 * the EEH status. We'll do more sensible things later when we have
339 * proper EEH support. We need to make sure we don't pollute ourselves
340 * with the normal errors generated when probing empty slots
343 pnv_pci_dump_phb_diag_data(phb
->hose
, phb
->diag
.blob
);
345 pr_warning("PCI %d: No diag data available\n",
346 phb
->hose
->global_number
);
349 spin_unlock_irqrestore(&phb
->lock
, flags
);
352 static void pnv_pci_config_check_eeh(struct pnv_phb
*phb
,
353 struct device_node
*dn
)
361 * Get the PE#. During the PCI probe stage, we might not
362 * setup that yet. So all ER errors should be mapped to
365 pe_no
= PCI_DN(dn
)->pe_number
;
366 if (pe_no
== IODA_INVALID_PE
) {
367 if (phb
->type
== PNV_PHB_P5IOC2
)
370 pe_no
= phb
->ioda
.reserved_pe
;
373 /* Read freeze status */
374 rc
= opal_pci_eeh_freeze_status(phb
->opal_id
, pe_no
, &fstate
, &pcierr
,
377 pr_warning("%s: Can't read EEH status (PE#%d) for "
379 __func__
, pe_no
, dn
->full_name
, rc
);
382 cfg_dbg(" -> EEH check, bdfn=%04x PE#%d fstate=%x\n",
383 (PCI_DN(dn
)->busno
<< 8) | (PCI_DN(dn
)->devfn
),
386 pnv_pci_handle_eeh_config(phb
, pe_no
);
389 int pnv_pci_cfg_read(struct device_node
*dn
,
390 int where
, int size
, u32
*val
)
392 struct pci_dn
*pdn
= PCI_DN(dn
);
393 struct pnv_phb
*phb
= pdn
->phb
->private_data
;
394 u32 bdfn
= (pdn
->busno
<< 8) | pdn
->devfn
;
400 rc
= opal_pci_config_read_byte(phb
->opal_id
, bdfn
, where
, &v8
);
401 *val
= (rc
== OPAL_SUCCESS
) ? v8
: 0xff;
406 rc
= opal_pci_config_read_half_word(phb
->opal_id
, bdfn
, where
,
408 *val
= (rc
== OPAL_SUCCESS
) ? be16_to_cpu(v16
) : 0xffff;
413 rc
= opal_pci_config_read_word(phb
->opal_id
, bdfn
, where
, &v32
);
414 *val
= (rc
== OPAL_SUCCESS
) ? be32_to_cpu(v32
) : 0xffffffff;
418 return PCIBIOS_FUNC_NOT_SUPPORTED
;
421 cfg_dbg("%s: bus: %x devfn: %x +%x/%x -> %08x\n",
422 __func__
, pdn
->busno
, pdn
->devfn
, where
, size
, *val
);
423 return PCIBIOS_SUCCESSFUL
;
426 int pnv_pci_cfg_write(struct device_node
*dn
,
427 int where
, int size
, u32 val
)
429 struct pci_dn
*pdn
= PCI_DN(dn
);
430 struct pnv_phb
*phb
= pdn
->phb
->private_data
;
431 u32 bdfn
= (pdn
->busno
<< 8) | pdn
->devfn
;
433 cfg_dbg("%s: bus: %x devfn: %x +%x/%x -> %08x\n",
434 pdn
->busno
, pdn
->devfn
, where
, size
, val
);
437 opal_pci_config_write_byte(phb
->opal_id
, bdfn
, where
, val
);
440 opal_pci_config_write_half_word(phb
->opal_id
, bdfn
, where
, val
);
443 opal_pci_config_write_word(phb
->opal_id
, bdfn
, where
, val
);
446 return PCIBIOS_FUNC_NOT_SUPPORTED
;
449 return PCIBIOS_SUCCESSFUL
;
453 static bool pnv_pci_cfg_check(struct pci_controller
*hose
,
454 struct device_node
*dn
)
456 struct eeh_dev
*edev
= NULL
;
457 struct pnv_phb
*phb
= hose
->private_data
;
459 /* EEH not enabled ? */
460 if (!(phb
->flags
& PNV_PHB_FLAG_EEH
))
463 /* PE reset or device removed ? */
464 edev
= of_node_to_eeh_dev(dn
);
467 (edev
->pe
->state
& EEH_PE_RESET
))
470 if (edev
->mode
& EEH_DEV_REMOVED
)
477 static inline pnv_pci_cfg_check(struct pci_controller
*hose
,
478 struct device_node
*dn
)
482 #endif /* CONFIG_EEH */
484 static int pnv_pci_read_config(struct pci_bus
*bus
,
486 int where
, int size
, u32
*val
)
488 struct device_node
*dn
, *busdn
= pci_bus_to_OF_node(bus
);
495 for (dn
= busdn
->child
; dn
; dn
= dn
->sibling
) {
497 if (pdn
&& pdn
->devfn
== devfn
) {
498 phb
= pdn
->phb
->private_data
;
504 if (!found
|| !pnv_pci_cfg_check(pdn
->phb
, dn
))
505 return PCIBIOS_DEVICE_NOT_FOUND
;
507 ret
= pnv_pci_cfg_read(dn
, where
, size
, val
);
508 if (phb
->flags
& PNV_PHB_FLAG_EEH
) {
509 if (*val
== EEH_IO_ERROR_VALUE(size
) &&
510 eeh_dev_check_failure(of_node_to_eeh_dev(dn
)))
511 return PCIBIOS_DEVICE_NOT_FOUND
;
513 pnv_pci_config_check_eeh(phb
, dn
);
519 static int pnv_pci_write_config(struct pci_bus
*bus
,
521 int where
, int size
, u32 val
)
523 struct device_node
*dn
, *busdn
= pci_bus_to_OF_node(bus
);
529 for (dn
= busdn
->child
; dn
; dn
= dn
->sibling
) {
531 if (pdn
&& pdn
->devfn
== devfn
) {
532 phb
= pdn
->phb
->private_data
;
538 if (!found
|| !pnv_pci_cfg_check(pdn
->phb
, dn
))
539 return PCIBIOS_DEVICE_NOT_FOUND
;
541 ret
= pnv_pci_cfg_write(dn
, where
, size
, val
);
542 if (!(phb
->flags
& PNV_PHB_FLAG_EEH
))
543 pnv_pci_config_check_eeh(phb
, dn
);
548 struct pci_ops pnv_pci_ops
= {
549 .read
= pnv_pci_read_config
,
550 .write
= pnv_pci_write_config
,
553 static int pnv_tce_build(struct iommu_table
*tbl
, long index
, long npages
,
554 unsigned long uaddr
, enum dma_data_direction direction
,
555 struct dma_attrs
*attrs
, bool rm
)
561 proto_tce
= TCE_PCI_READ
; // Read allowed
563 if (direction
!= DMA_TO_DEVICE
)
564 proto_tce
|= TCE_PCI_WRITE
;
566 tces
= tcep
= ((__be64
*)tbl
->it_base
) + index
- tbl
->it_offset
;
567 rpn
= __pa(uaddr
) >> tbl
->it_page_shift
;
570 *(tcep
++) = cpu_to_be64(proto_tce
|
571 (rpn
++ << tbl
->it_page_shift
));
573 /* Some implementations won't cache invalid TCEs and thus may not
574 * need that flush. We'll probably turn it_type into a bit mask
575 * of flags if that becomes the case
577 if (tbl
->it_type
& TCE_PCI_SWINV_CREATE
)
578 pnv_pci_ioda_tce_invalidate(tbl
, tces
, tcep
- 1, rm
);
583 static int pnv_tce_build_vm(struct iommu_table
*tbl
, long index
, long npages
,
585 enum dma_data_direction direction
,
586 struct dma_attrs
*attrs
)
588 return pnv_tce_build(tbl
, index
, npages
, uaddr
, direction
, attrs
,
592 static void pnv_tce_free(struct iommu_table
*tbl
, long index
, long npages
,
597 tces
= tcep
= ((__be64
*)tbl
->it_base
) + index
- tbl
->it_offset
;
600 *(tcep
++) = cpu_to_be64(0);
602 if (tbl
->it_type
& TCE_PCI_SWINV_FREE
)
603 pnv_pci_ioda_tce_invalidate(tbl
, tces
, tcep
- 1, rm
);
606 static void pnv_tce_free_vm(struct iommu_table
*tbl
, long index
, long npages
)
608 pnv_tce_free(tbl
, index
, npages
, false);
611 static unsigned long pnv_tce_get(struct iommu_table
*tbl
, long index
)
613 return ((u64
*)tbl
->it_base
)[index
- tbl
->it_offset
];
616 static int pnv_tce_build_rm(struct iommu_table
*tbl
, long index
, long npages
,
618 enum dma_data_direction direction
,
619 struct dma_attrs
*attrs
)
621 return pnv_tce_build(tbl
, index
, npages
, uaddr
, direction
, attrs
, true);
624 static void pnv_tce_free_rm(struct iommu_table
*tbl
, long index
, long npages
)
626 pnv_tce_free(tbl
, index
, npages
, true);
629 void pnv_pci_setup_iommu_table(struct iommu_table
*tbl
,
630 void *tce_mem
, u64 tce_size
,
631 u64 dma_offset
, unsigned page_shift
)
633 tbl
->it_blocksize
= 16;
634 tbl
->it_base
= (unsigned long)tce_mem
;
635 tbl
->it_page_shift
= page_shift
;
636 tbl
->it_offset
= dma_offset
>> tbl
->it_page_shift
;
638 tbl
->it_size
= tce_size
>> 3;
640 tbl
->it_type
= TCE_PCI
;
643 static struct iommu_table
*pnv_pci_setup_bml_iommu(struct pci_controller
*hose
)
645 struct iommu_table
*tbl
;
646 const __be64
*basep
, *swinvp
;
649 basep
= of_get_property(hose
->dn
, "linux,tce-base", NULL
);
650 sizep
= of_get_property(hose
->dn
, "linux,tce-size", NULL
);
651 if (basep
== NULL
|| sizep
== NULL
) {
652 pr_err("PCI: %s has missing tce entries !\n",
653 hose
->dn
->full_name
);
656 tbl
= kzalloc_node(sizeof(struct iommu_table
), GFP_KERNEL
, hose
->node
);
659 pnv_pci_setup_iommu_table(tbl
, __va(be64_to_cpup(basep
)),
660 be32_to_cpup(sizep
), 0, IOMMU_PAGE_SHIFT_4K
);
661 iommu_init_table(tbl
, hose
->node
);
662 iommu_register_group(tbl
, pci_domain_nr(hose
->bus
), 0);
664 /* Deal with SW invalidated TCEs when needed (BML way) */
665 swinvp
= of_get_property(hose
->dn
, "linux,tce-sw-invalidate-info",
668 tbl
->it_busno
= be64_to_cpu(swinvp
[1]);
669 tbl
->it_index
= (unsigned long)ioremap(be64_to_cpup(swinvp
), 8);
670 tbl
->it_type
= TCE_PCI_SWINV_CREATE
| TCE_PCI_SWINV_FREE
;
675 static void pnv_pci_dma_fallback_setup(struct pci_controller
*hose
,
676 struct pci_dev
*pdev
)
678 struct device_node
*np
= pci_bus_to_OF_node(hose
->bus
);
684 if (!pdn
->iommu_table
)
685 pdn
->iommu_table
= pnv_pci_setup_bml_iommu(hose
);
686 if (!pdn
->iommu_table
)
688 set_iommu_table_base_and_group(&pdev
->dev
, pdn
->iommu_table
);
691 static void pnv_pci_dma_dev_setup(struct pci_dev
*pdev
)
693 struct pci_controller
*hose
= pci_bus_to_host(pdev
->bus
);
694 struct pnv_phb
*phb
= hose
->private_data
;
696 /* If we have no phb structure, try to setup a fallback based on
697 * the device-tree (RTAS PCI for example)
699 if (phb
&& phb
->dma_dev_setup
)
700 phb
->dma_dev_setup(phb
, pdev
);
702 pnv_pci_dma_fallback_setup(hose
, pdev
);
705 int pnv_pci_dma_set_mask(struct pci_dev
*pdev
, u64 dma_mask
)
707 struct pci_controller
*hose
= pci_bus_to_host(pdev
->bus
);
708 struct pnv_phb
*phb
= hose
->private_data
;
710 if (phb
&& phb
->dma_set_mask
)
711 return phb
->dma_set_mask(phb
, pdev
, dma_mask
);
712 return __dma_set_mask(&pdev
->dev
, dma_mask
);
715 void pnv_pci_shutdown(void)
717 struct pci_controller
*hose
;
719 list_for_each_entry(hose
, &hose_list
, list_node
) {
720 struct pnv_phb
*phb
= hose
->private_data
;
722 if (phb
&& phb
->shutdown
)
727 /* Fixup wrong class code in p7ioc and p8 root complex */
728 static void pnv_p7ioc_rc_quirk(struct pci_dev
*dev
)
730 dev
->class = PCI_CLASS_BRIDGE_PCI
<< 8;
732 DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_IBM
, 0x3b9, pnv_p7ioc_rc_quirk
);
734 static int pnv_pci_probe_mode(struct pci_bus
*bus
)
736 struct pci_controller
*hose
= pci_bus_to_host(bus
);
737 const __be64
*tstamp
;
741 /* We hijack this as a way to ensure we have waited long
742 * enough since the reset was lifted on the PCI bus
744 if (bus
!= hose
->bus
)
745 return PCI_PROBE_NORMAL
;
746 tstamp
= of_get_property(hose
->dn
, "reset-clear-timestamp", NULL
);
747 if (!tstamp
|| !*tstamp
)
748 return PCI_PROBE_NORMAL
;
750 now
= mftb() / tb_ticks_per_usec
;
751 target
= (be64_to_cpup(tstamp
) / tb_ticks_per_usec
)
752 + PCI_RESET_DELAY_US
;
754 pr_devel("pci %04d: Reset target: 0x%llx now: 0x%llx\n",
755 hose
->global_number
, target
, now
);
758 msleep((target
- now
+ 999) / 1000);
760 return PCI_PROBE_NORMAL
;
763 void __init
pnv_pci_init(void)
765 struct device_node
*np
;
767 pci_add_flags(PCI_CAN_SKIP_ISA_ALIGN
);
769 /* OPAL absent, try POPAL first then RTAS detection of PHBs */
770 if (!firmware_has_feature(FW_FEATURE_OPAL
)) {
771 #ifdef CONFIG_PPC_POWERNV_RTAS
772 init_pci_config_tokens();
773 find_and_init_phbs();
774 #endif /* CONFIG_PPC_POWERNV_RTAS */
776 /* OPAL is here, do our normal stuff */
780 /* Look for IODA IO-Hubs. We don't support mixing IODA
781 * and p5ioc2 due to the need to change some global
784 for_each_compatible_node(np
, NULL
, "ibm,ioda-hub") {
785 pnv_pci_init_ioda_hub(np
);
789 /* Look for p5ioc2 IO-Hubs */
791 for_each_compatible_node(np
, NULL
, "ibm,p5ioc2")
792 pnv_pci_init_p5ioc2_hub(np
);
794 /* Look for ioda2 built-in PHB3's */
795 for_each_compatible_node(np
, NULL
, "ibm,ioda2-phb")
796 pnv_pci_init_ioda2_phb(np
);
799 /* Setup the linkage between OF nodes and PHBs */
802 /* Configure IOMMU DMA hooks */
803 ppc_md
.pci_dma_dev_setup
= pnv_pci_dma_dev_setup
;
804 ppc_md
.tce_build
= pnv_tce_build_vm
;
805 ppc_md
.tce_free
= pnv_tce_free_vm
;
806 ppc_md
.tce_build_rm
= pnv_tce_build_rm
;
807 ppc_md
.tce_free_rm
= pnv_tce_free_rm
;
808 ppc_md
.tce_get
= pnv_tce_get
;
809 ppc_md
.pci_probe_mode
= pnv_pci_probe_mode
;
810 set_pci_dma_ops(&dma_iommu_ops
);
813 #ifdef CONFIG_PCI_MSI
814 ppc_md
.msi_check_device
= pnv_msi_check_device
;
815 ppc_md
.setup_msi_irqs
= pnv_setup_msi_irqs
;
816 ppc_md
.teardown_msi_irqs
= pnv_teardown_msi_irqs
;
820 static int tce_iommu_bus_notifier(struct notifier_block
*nb
,
821 unsigned long action
, void *data
)
823 struct device
*dev
= data
;
826 case BUS_NOTIFY_ADD_DEVICE
:
827 return iommu_add_device(dev
);
828 case BUS_NOTIFY_DEL_DEVICE
:
829 if (dev
->iommu_group
)
830 iommu_del_device(dev
);
837 static struct notifier_block tce_iommu_bus_nb
= {
838 .notifier_call
= tce_iommu_bus_notifier
,
841 static int __init
tce_iommu_bus_notifier_init(void)
843 bus_register_notifier(&pci_bus_type
, &tce_iommu_bus_nb
);
847 subsys_initcall_sync(tce_iommu_bus_notifier_init
);