1 #ifndef __POWERNV_PCI_H
2 #define __POWERNV_PCI_H
13 /* Precise PHB model for error management */
15 PNV_PHB_MODEL_UNKNOWN
,
22 #define PNV_PCI_DIAG_BUF_SIZE 8192
23 #define PNV_IODA_PE_DEV (1 << 0) /* PE has single PCI device */
24 #define PNV_IODA_PE_BUS (1 << 1) /* PE has primary PCI bus */
25 #define PNV_IODA_PE_BUS_ALL (1 << 2) /* PE has subordinate buses */
26 #define PNV_IODA_PE_MASTER (1 << 3) /* Master PE in compound case */
27 #define PNV_IODA_PE_SLAVE (1 << 4) /* Slave PE in compound case */
28 #define PNV_IODA_PE_VF (1 << 5) /* PE for one VF */
29 #define PNV_IODA_PE_PEER (1 << 6) /* PE has peers */
31 /* Data associated with a PE, including IOMMU tracking etc.. */
37 #define PNV_IODA_MAX_PEER_PES 8
38 struct pnv_ioda_pe
*peers
[PNV_IODA_MAX_PEER_PES
];
40 /* A PE can be associated with a single device or an
41 * entire bus (& children). In the former case, pdev
42 * is populated, in the later case, pbus is.
45 struct pci_dev
*parent_dev
;
50 /* Effective RID (device RID for a device PE and base bus
51 * RID with devfn 0 for a bus PE)
56 unsigned int pe_number
;
58 /* "Weight" assigned to the PE for the sake of DMA resource
61 unsigned int dma_weight
;
63 /* "Base" iommu table, ie, 4K TCEs, 32-bit DMA */
66 struct iommu_table_group table_group
;
68 /* 64-bit TCE bypass region */
69 bool tce_bypass_enabled
;
70 uint64_t tce_bypass_base
;
72 /* MSIs. MVE index is identical for for 32 and 64 bit MSI
73 * and -1 if not supported. (It's actually identical to the
78 /* PEs in compound case */
79 struct pnv_ioda_pe
*master
;
80 struct list_head slaves
;
82 /* Link in list of PE#s */
83 struct list_head dma_link
;
84 struct list_head list
;
87 #define PNV_PHB_FLAG_EEH (1 << 0)
90 struct pci_controller
*hose
;
91 enum pnv_phb_type type
;
92 enum pnv_phb_model model
;
100 #ifdef CONFIG_DEBUG_FS
102 struct dentry
*dbgfs
;
105 #ifdef CONFIG_PCI_MSI
106 unsigned int msi_base
;
107 unsigned int msi32_support
;
108 struct msi_bitmap msi_bmp
;
110 int (*msi_setup
)(struct pnv_phb
*phb
, struct pci_dev
*dev
,
111 unsigned int hwirq
, unsigned int virq
,
112 unsigned int is_64
, struct msi_msg
*msg
);
113 void (*dma_dev_setup
)(struct pnv_phb
*phb
, struct pci_dev
*pdev
);
114 void (*fixup_phb
)(struct pci_controller
*hose
);
115 u32 (*bdfn_to_pe
)(struct pnv_phb
*phb
, struct pci_bus
*bus
, u32 devfn
);
116 int (*init_m64
)(struct pnv_phb
*phb
);
117 void (*reserve_m64_pe
)(struct pci_bus
*bus
,
118 unsigned long *pe_bitmap
, bool all
);
119 int (*pick_m64_pe
)(struct pci_bus
*bus
, bool all
);
120 int (*get_pe_state
)(struct pnv_phb
*phb
, int pe_no
);
121 void (*freeze_pe
)(struct pnv_phb
*phb
, int pe_no
);
122 int (*unfreeze_pe
)(struct pnv_phb
*phb
, int pe_no
, int opt
);
126 struct iommu_table iommu_table
;
127 struct iommu_table_group table_group
;
131 /* Global bridge info */
132 unsigned int total_pe
;
133 unsigned int reserved_pe
;
135 /* 32-bit MMIO window */
136 unsigned int m32_size
;
137 unsigned int m32_segsize
;
138 unsigned int m32_pci_base
;
140 /* 64-bit MMIO window */
141 unsigned int m64_bar_idx
;
142 unsigned long m64_size
;
143 unsigned long m64_segsize
;
144 unsigned long m64_base
;
145 unsigned long m64_bar_alloc
;
148 unsigned int io_size
;
149 unsigned int io_segsize
;
150 unsigned int io_pci_base
;
152 /* PE allocation bitmap */
153 unsigned long *pe_alloc
;
154 /* PE allocation mutex */
155 struct mutex pe_alloc_mutex
;
157 /* M32 & IO segment maps */
158 unsigned int *m32_segmap
;
159 unsigned int *io_segmap
;
160 struct pnv_ioda_pe
*pe_array
;
164 struct irq_chip irq_chip
;
166 /* Sorted list of used PE's based
167 * on the sequence of creation
169 struct list_head pe_list
;
170 struct mutex pe_list_mutex
;
172 /* Reverse map of PEs, will have to extend if
173 * we are to support more than 256 PEs, indexed
176 unsigned char pe_rmap
[0x10000];
178 /* 32-bit TCE tables allocation */
179 unsigned long tce32_count
;
181 /* Total "weight" for the sake of DMA resources
184 unsigned int dma_weight
;
185 unsigned int dma_pe_count
;
187 /* Sorted list of used PE's, sorted at
188 * boot for resource allocation purposes
190 struct list_head pe_dma_list
;
192 /* TCE cache invalidate registers (physical and
195 phys_addr_t tce_inval_reg_phys
;
196 __be64 __iomem
*tce_inval_reg
;
200 /* PHB and hub status structure */
202 unsigned char blob
[PNV_PCI_DIAG_BUF_SIZE
];
203 struct OpalIoP7IOCPhbErrorData p7ioc
;
204 struct OpalIoPhb3ErrorData phb3
;
205 struct OpalIoP7IOCErrorData hub_diag
;
210 extern struct pci_ops pnv_pci_ops
;
211 extern int pnv_tce_build(struct iommu_table
*tbl
, long index
, long npages
,
212 unsigned long uaddr
, enum dma_data_direction direction
,
213 struct dma_attrs
*attrs
);
214 extern void pnv_tce_free(struct iommu_table
*tbl
, long index
, long npages
);
215 extern int pnv_tce_xchg(struct iommu_table
*tbl
, long index
,
216 unsigned long *hpa
, enum dma_data_direction
*direction
);
217 extern unsigned long pnv_tce_get(struct iommu_table
*tbl
, long index
);
219 void pnv_pci_dump_phb_diag_data(struct pci_controller
*hose
,
220 unsigned char *log_buff
);
221 int pnv_pci_cfg_read(struct pci_dn
*pdn
,
222 int where
, int size
, u32
*val
);
223 int pnv_pci_cfg_write(struct pci_dn
*pdn
,
224 int where
, int size
, u32 val
);
225 extern struct iommu_table
*pnv_pci_table_alloc(int nid
);
227 extern long pnv_pci_link_table_and_group(int node
, int num
,
228 struct iommu_table
*tbl
,
229 struct iommu_table_group
*table_group
);
230 extern void pnv_pci_unlink_table_and_group(struct iommu_table
*tbl
,
231 struct iommu_table_group
*table_group
);
232 extern void pnv_pci_setup_iommu_table(struct iommu_table
*tbl
,
233 void *tce_mem
, u64 tce_size
,
234 u64 dma_offset
, unsigned page_shift
);
235 extern void pnv_pci_init_p5ioc2_hub(struct device_node
*np
);
236 extern void pnv_pci_init_ioda_hub(struct device_node
*np
);
237 extern void pnv_pci_init_ioda2_phb(struct device_node
*np
);
238 extern void pnv_pci_init_npu_phb(struct device_node
*np
);
239 extern void pnv_pci_ioda_tce_invalidate(struct iommu_table
*tbl
,
240 __be64
*startp
, __be64
*endp
, bool rm
);
241 extern void pnv_pci_reset_secondary_bus(struct pci_dev
*dev
);
242 extern int pnv_eeh_phb_reset(struct pci_controller
*hose
, int option
);
244 extern void pnv_pci_dma_dev_setup(struct pci_dev
*pdev
);
245 extern void pnv_pci_dma_bus_setup(struct pci_bus
*bus
);
246 extern int pnv_setup_msi_irqs(struct pci_dev
*pdev
, int nvec
, int type
);
247 extern void pnv_teardown_msi_irqs(struct pci_dev
*pdev
);
249 /* Nvlink functions */
250 extern void pnv_npu_tce_invalidate_entire(struct pnv_ioda_pe
*npe
);
251 extern void pnv_npu_tce_invalidate(struct pnv_ioda_pe
*npe
,
252 struct iommu_table
*tbl
,
254 unsigned long npages
,
256 extern void pnv_npu_init_dma_pe(struct pnv_ioda_pe
*npe
);
257 extern void pnv_npu_setup_dma_pe(struct pnv_ioda_pe
*npe
);
258 extern int pnv_npu_dma_set_bypass(struct pnv_ioda_pe
*npe
, bool enabled
);
259 extern int pnv_npu_dma_set_mask(struct pci_dev
*npdev
, u64 dma_mask
);
261 #endif /* __POWERNV_PCI_H */