1 // SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause)
4 #include <linux/module.h>
13 void (*desc_changed
)(void *ctx
);
15 u32 __iomem
*peer_mws
[];
19 * ntb_msi_init() - Initialize the MSI context
20 * @ntb: NTB device context
22 * This function must be called before any other ntb_msi function.
23 * It initializes the context for MSI operations and maps
24 * the peer memory windows.
26 * This function reserves the last N outbound memory windows (where N
27 * is the number of peers).
29 * Return: Zero on success, otherwise a negative error number.
31 int ntb_msi_init(struct ntb_dev
*ntb
,
32 void (*desc_changed
)(void *ctx
))
34 phys_addr_t mw_phys_addr
;
35 resource_size_t mw_size
;
42 peers
= ntb_peer_port_count(ntb
);
46 struct_size
= sizeof(*ntb
->msi
) + sizeof(*ntb
->msi
->peer_mws
) * peers
;
48 ntb
->msi
= devm_kzalloc(&ntb
->dev
, struct_size
, GFP_KERNEL
);
52 ntb
->msi
->desc_changed
= desc_changed
;
54 for (i
= 0; i
< peers
; i
++) {
55 peer_widx
= ntb_peer_mw_count(ntb
) - 1 - i
;
57 ret
= ntb_peer_mw_get_addr(ntb
, peer_widx
, &mw_phys_addr
,
62 ntb
->msi
->peer_mws
[i
] = devm_ioremap(&ntb
->dev
, mw_phys_addr
,
64 if (!ntb
->msi
->peer_mws
[i
]) {
73 for (i
= 0; i
< peers
; i
++)
74 if (ntb
->msi
->peer_mws
[i
])
75 devm_iounmap(&ntb
->dev
, ntb
->msi
->peer_mws
[i
]);
77 devm_kfree(&ntb
->dev
, ntb
->msi
);
81 EXPORT_SYMBOL(ntb_msi_init
);
84 * ntb_msi_setup_mws() - Initialize the MSI inbound memory windows
85 * @ntb: NTB device context
87 * This function sets up the required inbound memory windows. It should be
88 * called from a work function after a link up event.
90 * Over the entire network, this function will reserves the last N
91 * inbound memory windows for each peer (where N is the number of peers).
93 * ntb_msi_init() must be called before this function.
95 * Return: Zero on success, otherwise a negative error number.
97 int ntb_msi_setup_mws(struct ntb_dev
*ntb
)
99 struct msi_desc
*desc
;
102 resource_size_t addr_align
, size_align
, size_max
;
103 resource_size_t mw_size
= SZ_32K
;
104 resource_size_t mw_min_size
= mw_size
;
111 msi_lock_descs(&ntb
->pdev
->dev
);
112 desc
= msi_first_desc(&ntb
->pdev
->dev
, MSI_DESC_ASSOCIATED
);
113 addr
= desc
->msg
.address_lo
+ ((uint64_t)desc
->msg
.address_hi
<< 32);
114 msi_unlock_descs(&ntb
->pdev
->dev
);
116 for (peer
= 0; peer
< ntb_peer_port_count(ntb
); peer
++) {
117 peer_widx
= ntb_peer_highest_mw_idx(ntb
, peer
);
121 ret
= ntb_mw_get_align(ntb
, peer
, peer_widx
, &addr_align
,
126 addr
&= ~(addr_align
- 1);
129 for (peer
= 0; peer
< ntb_peer_port_count(ntb
); peer
++) {
130 peer_widx
= ntb_peer_highest_mw_idx(ntb
, peer
);
136 ret
= ntb_mw_get_align(ntb
, peer
, peer_widx
, NULL
,
137 &size_align
, &size_max
);
141 mw_size
= round_up(mw_size
, size_align
);
142 mw_size
= max(mw_size
, size_max
);
143 if (mw_size
< mw_min_size
)
144 mw_min_size
= mw_size
;
146 ret
= ntb_mw_set_trans(ntb
, peer
, peer_widx
,
152 ntb
->msi
->base_addr
= addr
;
153 ntb
->msi
->end_addr
= addr
+ mw_min_size
;
158 for (i
= 0; i
< peer
; i
++) {
159 peer_widx
= ntb_peer_highest_mw_idx(ntb
, peer
);
163 ntb_mw_clear_trans(ntb
, i
, peer_widx
);
168 EXPORT_SYMBOL(ntb_msi_setup_mws
);
171 * ntb_msi_clear_mws() - Clear all inbound memory windows
172 * @ntb: NTB device context
174 * This function tears down the resources used by ntb_msi_setup_mws().
176 void ntb_msi_clear_mws(struct ntb_dev
*ntb
)
181 for (peer
= 0; peer
< ntb_peer_port_count(ntb
); peer
++) {
182 peer_widx
= ntb_peer_highest_mw_idx(ntb
, peer
);
186 ntb_mw_clear_trans(ntb
, peer
, peer_widx
);
189 EXPORT_SYMBOL(ntb_msi_clear_mws
);
191 struct ntb_msi_devres
{
193 struct msi_desc
*entry
;
194 struct ntb_msi_desc
*msi_desc
;
197 static int ntb_msi_set_desc(struct ntb_dev
*ntb
, struct msi_desc
*entry
,
198 struct ntb_msi_desc
*msi_desc
)
202 addr
= entry
->msg
.address_lo
+
203 ((uint64_t)entry
->msg
.address_hi
<< 32);
205 if (addr
< ntb
->msi
->base_addr
|| addr
>= ntb
->msi
->end_addr
) {
206 dev_warn_once(&ntb
->dev
,
207 "IRQ %d: MSI Address not within the memory window (%llx, [%llx %llx])\n",
208 entry
->irq
, addr
, ntb
->msi
->base_addr
,
213 msi_desc
->addr_offset
= addr
- ntb
->msi
->base_addr
;
214 msi_desc
->data
= entry
->msg
.data
;
219 static void ntb_msi_write_msg(struct msi_desc
*entry
, void *data
)
221 struct ntb_msi_devres
*dr
= data
;
223 WARN_ON(ntb_msi_set_desc(dr
->ntb
, entry
, dr
->msi_desc
));
225 if (dr
->ntb
->msi
->desc_changed
)
226 dr
->ntb
->msi
->desc_changed(dr
->ntb
->ctx
);
229 static void ntbm_msi_callback_release(struct device
*dev
, void *res
)
231 struct ntb_msi_devres
*dr
= res
;
233 dr
->entry
->write_msi_msg
= NULL
;
234 dr
->entry
->write_msi_msg_data
= NULL
;
237 static int ntbm_msi_setup_callback(struct ntb_dev
*ntb
, struct msi_desc
*entry
,
238 struct ntb_msi_desc
*msi_desc
)
240 struct ntb_msi_devres
*dr
;
242 dr
= devres_alloc(ntbm_msi_callback_release
,
243 sizeof(struct ntb_msi_devres
), GFP_KERNEL
);
249 dr
->msi_desc
= msi_desc
;
251 devres_add(&ntb
->dev
, dr
);
253 dr
->entry
->write_msi_msg
= ntb_msi_write_msg
;
254 dr
->entry
->write_msi_msg_data
= dr
;
260 * ntbm_msi_request_threaded_irq() - allocate an MSI interrupt
261 * @ntb: NTB device context
262 * @handler: Function to be called when the IRQ occurs
263 * @thread_fn: Function to be called in a threaded interrupt context. NULL
264 * for clients which handle everything in @handler
265 * @name: An ascii name for the claiming device, dev_name(dev) if NULL
266 * @dev_id: A cookie passed back to the handler function
267 * @msi_desc: MSI descriptor data which triggers the interrupt
269 * This function assigns an interrupt handler to an unused
270 * MSI interrupt and returns the descriptor used to trigger
271 * it. The descriptor can then be sent to a peer to trigger
274 * The interrupt resource is managed with devres so it will
275 * be automatically freed when the NTB device is torn down.
277 * If an IRQ allocated with this function needs to be freed
278 * separately, ntbm_free_irq() must be used.
280 * Return: IRQ number assigned on success, otherwise a negative error number.
282 int ntbm_msi_request_threaded_irq(struct ntb_dev
*ntb
, irq_handler_t handler
,
283 irq_handler_t thread_fn
,
284 const char *name
, void *dev_id
,
285 struct ntb_msi_desc
*msi_desc
)
287 struct device
*dev
= &ntb
->pdev
->dev
;
288 struct msi_desc
*entry
;
295 msi_for_each_desc(entry
, dev
, MSI_DESC_ASSOCIATED
) {
296 if (irq_has_action(entry
->irq
))
299 ret
= devm_request_threaded_irq(&ntb
->dev
, entry
->irq
, handler
,
300 thread_fn
, 0, name
, dev_id
);
304 if (ntb_msi_set_desc(ntb
, entry
, msi_desc
)) {
305 devm_free_irq(&ntb
->dev
, entry
->irq
, dev_id
);
309 ret
= ntbm_msi_setup_callback(ntb
, entry
, msi_desc
);
311 devm_free_irq(&ntb
->dev
, entry
->irq
, dev_id
);
321 msi_unlock_descs(dev
);
324 EXPORT_SYMBOL(ntbm_msi_request_threaded_irq
);
326 static int ntbm_msi_callback_match(struct device
*dev
, void *res
, void *data
)
328 struct ntb_dev
*ntb
= dev_ntb(dev
);
329 struct ntb_msi_devres
*dr
= res
;
331 return dr
->ntb
== ntb
&& dr
->entry
== data
;
335 * ntbm_msi_free_irq() - free an interrupt
336 * @ntb: NTB device context
337 * @irq: Interrupt line to free
338 * @dev_id: Device identity to free
340 * This function should be used to manually free IRQs allocated with
341 * ntbm_request_[threaded_]irq().
343 void ntbm_msi_free_irq(struct ntb_dev
*ntb
, unsigned int irq
, void *dev_id
)
345 struct msi_desc
*entry
= irq_get_msi_desc(irq
);
347 entry
->write_msi_msg
= NULL
;
348 entry
->write_msi_msg_data
= NULL
;
350 WARN_ON(devres_destroy(&ntb
->dev
, ntbm_msi_callback_release
,
351 ntbm_msi_callback_match
, entry
));
353 devm_free_irq(&ntb
->dev
, irq
, dev_id
);
355 EXPORT_SYMBOL(ntbm_msi_free_irq
);
358 * ntb_msi_peer_trigger() - Trigger an interrupt handler on a peer
359 * @ntb: NTB device context
361 * @desc: MSI descriptor data which triggers the interrupt
363 * This function triggers an interrupt on a peer. It requires
364 * the descriptor structure to have been passed from that peer
365 * by some other means.
367 * Return: Zero on success, otherwise a negative error number.
369 int ntb_msi_peer_trigger(struct ntb_dev
*ntb
, int peer
,
370 struct ntb_msi_desc
*desc
)
377 idx
= desc
->addr_offset
/ sizeof(*ntb
->msi
->peer_mws
[peer
]);
379 iowrite32(desc
->data
, &ntb
->msi
->peer_mws
[peer
][idx
]);
383 EXPORT_SYMBOL(ntb_msi_peer_trigger
);
386 * ntb_msi_peer_addr() - Get the DMA address to trigger a peer's MSI interrupt
387 * @ntb: NTB device context
389 * @desc: MSI descriptor data which triggers the interrupt
390 * @msi_addr: Physical address to trigger the interrupt
392 * This function allows using DMA engines to trigger an interrupt
393 * (for example, trigger an interrupt to process the data after
394 * sending it). To trigger the interrupt, write @desc.data to the address
395 * returned in @msi_addr
397 * Return: Zero on success, otherwise a negative error number.
399 int ntb_msi_peer_addr(struct ntb_dev
*ntb
, int peer
,
400 struct ntb_msi_desc
*desc
,
401 phys_addr_t
*msi_addr
)
403 int peer_widx
= ntb_peer_mw_count(ntb
) - 1 - peer
;
404 phys_addr_t mw_phys_addr
;
407 ret
= ntb_peer_mw_get_addr(ntb
, peer_widx
, &mw_phys_addr
, NULL
);
412 *msi_addr
= mw_phys_addr
+ desc
->addr_offset
;
416 EXPORT_SYMBOL(ntb_msi_peer_addr
);