]>
Commit | Line | Data |
---|---|---|
1 | /* | |
2 | * Copyright (c) 2004, 2005 Topspin Communications. All rights reserved. | |
3 | * Copyright (c) 2005 Sun Microsystems, Inc. All rights reserved. | |
4 | * Copyright (c) 2004 Voltaire, Inc. All rights reserved. | |
5 | * | |
6 | * This software is available to you under a choice of one of two | |
7 | * licenses. You may choose to be licensed under the terms of the GNU | |
8 | * General Public License (GPL) Version 2, available from the file | |
9 | * COPYING in the main directory of this source tree, or the | |
10 | * OpenIB.org BSD license below: | |
11 | * | |
12 | * Redistribution and use in source and binary forms, with or | |
13 | * without modification, are permitted provided that the following | |
14 | * conditions are met: | |
15 | * | |
16 | * - Redistributions of source code must retain the above | |
17 | * copyright notice, this list of conditions and the following | |
18 | * disclaimer. | |
19 | * | |
20 | * - Redistributions in binary form must reproduce the above | |
21 | * copyright notice, this list of conditions and the following | |
22 | * disclaimer in the documentation and/or other materials | |
23 | * provided with the distribution. | |
24 | * | |
25 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, | |
26 | * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF | |
27 | * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND | |
28 | * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS | |
29 | * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN | |
30 | * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN | |
31 | * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | |
32 | * SOFTWARE. | |
33 | */ | |
34 | ||
35 | #ifndef _IPOIB_H | |
36 | #define _IPOIB_H | |
37 | ||
38 | #include <linux/list.h> | |
39 | #include <linux/skbuff.h> | |
40 | #include <linux/netdevice.h> | |
41 | #include <linux/workqueue.h> | |
42 | #include <linux/kref.h> | |
43 | #include <linux/if_infiniband.h> | |
44 | #include <linux/mutex.h> | |
45 | ||
46 | #include <net/neighbour.h> | |
47 | #include <net/sch_generic.h> | |
48 | ||
49 | #include <linux/atomic.h> | |
50 | ||
51 | #include <rdma/ib_verbs.h> | |
52 | #include <rdma/ib_pack.h> | |
53 | #include <rdma/ib_sa.h> | |
54 | #include <linux/sched.h> | |
55 | /* constants */ | |
56 | ||
57 | enum ipoib_flush_level { | |
58 | IPOIB_FLUSH_LIGHT, | |
59 | IPOIB_FLUSH_NORMAL, | |
60 | IPOIB_FLUSH_HEAVY | |
61 | }; | |
62 | ||
63 | enum { | |
64 | IPOIB_ENCAP_LEN = 4, | |
65 | IPOIB_PSEUDO_LEN = 20, | |
66 | IPOIB_HARD_LEN = IPOIB_ENCAP_LEN + IPOIB_PSEUDO_LEN, | |
67 | ||
68 | IPOIB_UD_HEAD_SIZE = IB_GRH_BYTES + IPOIB_ENCAP_LEN, | |
69 | IPOIB_UD_RX_SG = 2, /* max buffer needed for 4K mtu */ | |
70 | ||
71 | IPOIB_CM_MTU = 0x10000 - 0x10, /* padding to align header to 16 */ | |
72 | IPOIB_CM_BUF_SIZE = IPOIB_CM_MTU + IPOIB_ENCAP_LEN, | |
73 | IPOIB_CM_HEAD_SIZE = IPOIB_CM_BUF_SIZE % PAGE_SIZE, | |
74 | IPOIB_CM_RX_SG = ALIGN(IPOIB_CM_BUF_SIZE, PAGE_SIZE) / PAGE_SIZE, | |
75 | IPOIB_RX_RING_SIZE = 256, | |
76 | IPOIB_TX_RING_SIZE = 128, | |
77 | IPOIB_MAX_QUEUE_SIZE = 8192, | |
78 | IPOIB_MIN_QUEUE_SIZE = 2, | |
79 | IPOIB_CM_MAX_CONN_QP = 4096, | |
80 | ||
81 | IPOIB_NUM_WC = 4, | |
82 | ||
83 | IPOIB_MAX_PATH_REC_QUEUE = 3, | |
84 | IPOIB_MAX_MCAST_QUEUE = 64, | |
85 | ||
86 | IPOIB_FLAG_OPER_UP = 0, | |
87 | IPOIB_FLAG_INITIALIZED = 1, | |
88 | IPOIB_FLAG_ADMIN_UP = 2, | |
89 | IPOIB_PKEY_ASSIGNED = 3, | |
90 | IPOIB_FLAG_SUBINTERFACE = 5, | |
91 | IPOIB_STOP_REAPER = 7, | |
92 | IPOIB_FLAG_ADMIN_CM = 9, | |
93 | IPOIB_FLAG_UMCAST = 10, | |
94 | IPOIB_STOP_NEIGH_GC = 11, | |
95 | IPOIB_NEIGH_TBL_FLUSH = 12, | |
96 | IPOIB_FLAG_DEV_ADDR_SET = 13, | |
97 | IPOIB_FLAG_DEV_ADDR_CTRL = 14, | |
98 | IPOIB_FLAG_GOING_DOWN = 15, | |
99 | ||
100 | IPOIB_MAX_BACKOFF_SECONDS = 16, | |
101 | ||
102 | IPOIB_MCAST_FLAG_FOUND = 0, /* used in set_multicast_list */ | |
103 | IPOIB_MCAST_FLAG_SENDONLY = 1, | |
104 | /* | |
105 | * For IPOIB_MCAST_FLAG_BUSY | |
106 | * When set, in flight join and mcast->mc is unreliable | |
107 | * When clear and mcast->mc IS_ERR_OR_NULL, need to restart or | |
108 | * haven't started yet | |
109 | * When clear and mcast->mc is valid pointer, join was successful | |
110 | */ | |
111 | IPOIB_MCAST_FLAG_BUSY = 2, | |
112 | IPOIB_MCAST_FLAG_ATTACHED = 3, | |
113 | ||
114 | MAX_SEND_CQE = 16, | |
115 | IPOIB_CM_COPYBREAK = 256, | |
116 | ||
117 | IPOIB_NON_CHILD = 0, | |
118 | IPOIB_LEGACY_CHILD = 1, | |
119 | IPOIB_RTNL_CHILD = 2, | |
120 | }; | |
121 | ||
122 | #define IPOIB_OP_RECV (1ul << 31) | |
123 | #ifdef CONFIG_INFINIBAND_IPOIB_CM | |
124 | #define IPOIB_OP_CM (1ul << 30) | |
125 | #else | |
126 | #define IPOIB_OP_CM (0) | |
127 | #endif | |
128 | ||
129 | #define IPOIB_QPN_MASK ((__force u32) cpu_to_be32(0xFFFFFF)) | |
130 | ||
131 | /* structs */ | |
132 | ||
133 | struct ipoib_header { | |
134 | __be16 proto; | |
135 | u16 reserved; | |
136 | }; | |
137 | ||
138 | struct ipoib_pseudo_header { | |
139 | u8 hwaddr[INFINIBAND_ALEN]; | |
140 | }; | |
141 | ||
142 | static inline void skb_add_pseudo_hdr(struct sk_buff *skb) | |
143 | { | |
144 | char *data = skb_push(skb, IPOIB_PSEUDO_LEN); | |
145 | ||
146 | /* | |
147 | * only the ipoib header is present now, make room for a dummy | |
148 | * pseudo header and set skb field accordingly | |
149 | */ | |
150 | memset(data, 0, IPOIB_PSEUDO_LEN); | |
151 | skb_reset_mac_header(skb); | |
152 | skb_pull(skb, IPOIB_HARD_LEN); | |
153 | } | |
154 | ||
155 | static inline struct ipoib_dev_priv *ipoib_priv(const struct net_device *dev) | |
156 | { | |
157 | struct rdma_netdev *rn = netdev_priv(dev); | |
158 | ||
159 | return rn->clnt_priv; | |
160 | } | |
161 | ||
162 | /* Used for all multicast joins (broadcast, IPv4 mcast and IPv6 mcast) */ | |
163 | struct ipoib_mcast { | |
164 | struct ib_sa_mcmember_rec mcmember; | |
165 | struct ib_sa_multicast *mc; | |
166 | struct ipoib_ah *ah; | |
167 | ||
168 | struct rb_node rb_node; | |
169 | struct list_head list; | |
170 | ||
171 | unsigned long created; | |
172 | unsigned long backoff; | |
173 | unsigned long delay_until; | |
174 | ||
175 | unsigned long flags; | |
176 | unsigned char logcount; | |
177 | ||
178 | struct list_head neigh_list; | |
179 | ||
180 | struct sk_buff_head pkt_queue; | |
181 | ||
182 | struct net_device *dev; | |
183 | struct completion done; | |
184 | }; | |
185 | ||
186 | struct ipoib_rx_buf { | |
187 | struct sk_buff *skb; | |
188 | u64 mapping[IPOIB_UD_RX_SG]; | |
189 | }; | |
190 | ||
191 | struct ipoib_tx_buf { | |
192 | struct sk_buff *skb; | |
193 | u64 mapping[MAX_SKB_FRAGS + 1]; | |
194 | }; | |
195 | ||
196 | struct ipoib_cm_tx_buf { | |
197 | struct sk_buff *skb; | |
198 | u64 mapping; | |
199 | }; | |
200 | ||
201 | struct ib_cm_id; | |
202 | ||
203 | struct ipoib_cm_data { | |
204 | __be32 qpn; /* High byte MUST be ignored on receive */ | |
205 | __be32 mtu; | |
206 | }; | |
207 | ||
208 | /* | |
209 | * Quoting 10.3.1 Queue Pair and EE Context States: | |
210 | * | |
211 | * Note, for QPs that are associated with an SRQ, the Consumer should take the | |
212 | * QP through the Error State before invoking a Destroy QP or a Modify QP to the | |
213 | * Reset State. The Consumer may invoke the Destroy QP without first performing | |
214 | * a Modify QP to the Error State and waiting for the Affiliated Asynchronous | |
215 | * Last WQE Reached Event. However, if the Consumer does not wait for the | |
216 | * Affiliated Asynchronous Last WQE Reached Event, then WQE and Data Segment | |
217 | * leakage may occur. Therefore, it is good programming practice to tear down a | |
218 | * QP that is associated with an SRQ by using the following process: | |
219 | * | |
220 | * - Put the QP in the Error State | |
221 | * - Wait for the Affiliated Asynchronous Last WQE Reached Event; | |
222 | * - either: | |
223 | * drain the CQ by invoking the Poll CQ verb and either wait for CQ | |
224 | * to be empty or the number of Poll CQ operations has exceeded | |
225 | * CQ capacity size; | |
226 | * - or | |
227 | * post another WR that completes on the same CQ and wait for this | |
228 | * WR to return as a WC; | |
229 | * - and then invoke a Destroy QP or Reset QP. | |
230 | * | |
231 | * We use the second option and wait for a completion on the | |
232 | * same CQ before destroying QPs attached to our SRQ. | |
233 | */ | |
234 | ||
235 | enum ipoib_cm_state { | |
236 | IPOIB_CM_RX_LIVE, | |
237 | IPOIB_CM_RX_ERROR, /* Ignored by stale task */ | |
238 | IPOIB_CM_RX_FLUSH /* Last WQE Reached event observed */ | |
239 | }; | |
240 | ||
241 | struct ipoib_cm_rx { | |
242 | struct ib_cm_id *id; | |
243 | struct ib_qp *qp; | |
244 | struct ipoib_cm_rx_buf *rx_ring; | |
245 | struct list_head list; | |
246 | struct net_device *dev; | |
247 | unsigned long jiffies; | |
248 | enum ipoib_cm_state state; | |
249 | int recv_count; | |
250 | }; | |
251 | ||
252 | struct ipoib_cm_tx { | |
253 | struct ib_cm_id *id; | |
254 | struct ib_qp *qp; | |
255 | struct list_head list; | |
256 | struct net_device *dev; | |
257 | struct ipoib_neigh *neigh; | |
258 | struct ipoib_path *path; | |
259 | struct ipoib_tx_buf *tx_ring; | |
260 | unsigned tx_head; | |
261 | unsigned tx_tail; | |
262 | unsigned long flags; | |
263 | u32 mtu; | |
264 | unsigned max_send_sge; | |
265 | }; | |
266 | ||
267 | struct ipoib_cm_rx_buf { | |
268 | struct sk_buff *skb; | |
269 | u64 mapping[IPOIB_CM_RX_SG]; | |
270 | }; | |
271 | ||
272 | struct ipoib_cm_dev_priv { | |
273 | struct ib_srq *srq; | |
274 | struct ipoib_cm_rx_buf *srq_ring; | |
275 | struct ib_cm_id *id; | |
276 | struct list_head passive_ids; /* state: LIVE */ | |
277 | struct list_head rx_error_list; /* state: ERROR */ | |
278 | struct list_head rx_flush_list; /* state: FLUSH, drain not started */ | |
279 | struct list_head rx_drain_list; /* state: FLUSH, drain started */ | |
280 | struct list_head rx_reap_list; /* state: FLUSH, drain done */ | |
281 | struct work_struct start_task; | |
282 | struct work_struct reap_task; | |
283 | struct work_struct skb_task; | |
284 | struct work_struct rx_reap_task; | |
285 | struct delayed_work stale_task; | |
286 | struct sk_buff_head skb_queue; | |
287 | struct list_head start_list; | |
288 | struct list_head reap_list; | |
289 | struct ib_wc ibwc[IPOIB_NUM_WC]; | |
290 | struct ib_sge rx_sge[IPOIB_CM_RX_SG]; | |
291 | struct ib_recv_wr rx_wr; | |
292 | int nonsrq_conn_qp; | |
293 | int max_cm_mtu; | |
294 | int num_frags; | |
295 | }; | |
296 | ||
297 | struct ipoib_ethtool_st { | |
298 | u16 coalesce_usecs; | |
299 | u16 max_coalesced_frames; | |
300 | }; | |
301 | ||
302 | struct ipoib_neigh_table; | |
303 | ||
304 | struct ipoib_neigh_hash { | |
305 | struct ipoib_neigh_table *ntbl; | |
306 | struct ipoib_neigh __rcu **buckets; | |
307 | struct rcu_head rcu; | |
308 | u32 mask; | |
309 | u32 size; | |
310 | }; | |
311 | ||
312 | struct ipoib_neigh_table { | |
313 | struct ipoib_neigh_hash __rcu *htbl; | |
314 | atomic_t entries; | |
315 | struct completion flushed; | |
316 | struct completion deleted; | |
317 | }; | |
318 | ||
319 | struct ipoib_qp_state_validate { | |
320 | struct work_struct work; | |
321 | struct ipoib_dev_priv *priv; | |
322 | }; | |
323 | ||
324 | /* | |
325 | * Device private locking: network stack tx_lock protects members used | |
326 | * in TX fast path, lock protects everything else. lock nests inside | |
327 | * of tx_lock (ie tx_lock must be acquired first if needed). | |
328 | */ | |
329 | struct ipoib_dev_priv { | |
330 | spinlock_t lock; | |
331 | ||
332 | struct net_device *dev; | |
333 | ||
334 | struct napi_struct napi; | |
335 | ||
336 | unsigned long flags; | |
337 | ||
338 | struct rw_semaphore vlan_rwsem; | |
339 | struct mutex mcast_mutex; | |
340 | ||
341 | struct rb_root path_tree; | |
342 | struct list_head path_list; | |
343 | ||
344 | struct ipoib_neigh_table ntbl; | |
345 | ||
346 | struct ipoib_mcast *broadcast; | |
347 | struct list_head multicast_list; | |
348 | struct rb_root multicast_tree; | |
349 | ||
350 | struct workqueue_struct *wq; | |
351 | struct delayed_work mcast_task; | |
352 | struct work_struct carrier_on_task; | |
353 | struct work_struct flush_light; | |
354 | struct work_struct flush_normal; | |
355 | struct work_struct flush_heavy; | |
356 | struct work_struct restart_task; | |
357 | struct delayed_work ah_reap_task; | |
358 | struct delayed_work neigh_reap_task; | |
359 | struct ib_device *ca; | |
360 | u8 port; | |
361 | u16 pkey; | |
362 | u16 pkey_index; | |
363 | struct ib_pd *pd; | |
364 | struct ib_cq *recv_cq; | |
365 | struct ib_cq *send_cq; | |
366 | struct ib_qp *qp; | |
367 | u32 qkey; | |
368 | ||
369 | union ib_gid local_gid; | |
370 | u16 local_lid; | |
371 | ||
372 | unsigned int admin_mtu; | |
373 | unsigned int mcast_mtu; | |
374 | unsigned int max_ib_mtu; | |
375 | ||
376 | struct ipoib_rx_buf *rx_ring; | |
377 | ||
378 | struct ipoib_tx_buf *tx_ring; | |
379 | unsigned tx_head; | |
380 | unsigned tx_tail; | |
381 | struct ib_sge tx_sge[MAX_SKB_FRAGS + 1]; | |
382 | struct ib_ud_wr tx_wr; | |
383 | unsigned tx_outstanding; | |
384 | struct ib_wc send_wc[MAX_SEND_CQE]; | |
385 | ||
386 | struct ib_recv_wr rx_wr; | |
387 | struct ib_sge rx_sge[IPOIB_UD_RX_SG]; | |
388 | ||
389 | struct ib_wc ibwc[IPOIB_NUM_WC]; | |
390 | ||
391 | struct list_head dead_ahs; | |
392 | ||
393 | struct ib_event_handler event_handler; | |
394 | ||
395 | struct net_device *parent; | |
396 | struct list_head child_intfs; | |
397 | struct list_head list; | |
398 | int child_type; | |
399 | ||
400 | #ifdef CONFIG_INFINIBAND_IPOIB_CM | |
401 | struct ipoib_cm_dev_priv cm; | |
402 | #endif | |
403 | ||
404 | #ifdef CONFIG_INFINIBAND_IPOIB_DEBUG | |
405 | struct list_head fs_list; | |
406 | struct dentry *mcg_dentry; | |
407 | struct dentry *path_dentry; | |
408 | #endif | |
409 | u64 hca_caps; | |
410 | struct ipoib_ethtool_st ethtool; | |
411 | struct timer_list poll_timer; | |
412 | unsigned max_send_sge; | |
413 | bool sm_fullmember_sendonly_support; | |
414 | const struct net_device_ops *rn_ops; | |
415 | }; | |
416 | ||
417 | struct ipoib_ah { | |
418 | struct net_device *dev; | |
419 | struct ib_ah *ah; | |
420 | struct list_head list; | |
421 | struct kref ref; | |
422 | unsigned last_send; | |
423 | }; | |
424 | ||
425 | struct ipoib_path { | |
426 | struct net_device *dev; | |
427 | struct sa_path_rec pathrec; | |
428 | struct ipoib_ah *ah; | |
429 | struct sk_buff_head queue; | |
430 | ||
431 | struct list_head neigh_list; | |
432 | ||
433 | int query_id; | |
434 | struct ib_sa_query *query; | |
435 | struct completion done; | |
436 | ||
437 | struct rb_node rb_node; | |
438 | struct list_head list; | |
439 | int valid; | |
440 | }; | |
441 | ||
442 | struct ipoib_neigh { | |
443 | struct ipoib_ah *ah; | |
444 | #ifdef CONFIG_INFINIBAND_IPOIB_CM | |
445 | struct ipoib_cm_tx *cm; | |
446 | #endif | |
447 | u8 daddr[INFINIBAND_ALEN]; | |
448 | struct sk_buff_head queue; | |
449 | ||
450 | struct net_device *dev; | |
451 | ||
452 | struct list_head list; | |
453 | struct ipoib_neigh __rcu *hnext; | |
454 | struct rcu_head rcu; | |
455 | atomic_t refcnt; | |
456 | unsigned long alive; | |
457 | }; | |
458 | ||
459 | #define IPOIB_UD_MTU(ib_mtu) (ib_mtu - IPOIB_ENCAP_LEN) | |
460 | #define IPOIB_UD_BUF_SIZE(ib_mtu) (ib_mtu + IB_GRH_BYTES) | |
461 | ||
462 | void ipoib_neigh_dtor(struct ipoib_neigh *neigh); | |
463 | static inline void ipoib_neigh_put(struct ipoib_neigh *neigh) | |
464 | { | |
465 | if (atomic_dec_and_test(&neigh->refcnt)) | |
466 | ipoib_neigh_dtor(neigh); | |
467 | } | |
468 | struct ipoib_neigh *ipoib_neigh_get(struct net_device *dev, u8 *daddr); | |
469 | struct ipoib_neigh *ipoib_neigh_alloc(u8 *daddr, | |
470 | struct net_device *dev); | |
471 | void ipoib_neigh_free(struct ipoib_neigh *neigh); | |
472 | void ipoib_del_neighs_by_gid(struct net_device *dev, u8 *gid); | |
473 | ||
474 | extern struct workqueue_struct *ipoib_workqueue; | |
475 | ||
476 | /* functions */ | |
477 | ||
478 | int ipoib_poll(struct napi_struct *napi, int budget); | |
479 | void ipoib_ib_completion(struct ib_cq *cq, void *dev_ptr); | |
480 | void ipoib_send_comp_handler(struct ib_cq *cq, void *dev_ptr); | |
481 | ||
482 | struct ipoib_ah *ipoib_create_ah(struct net_device *dev, | |
483 | struct ib_pd *pd, struct rdma_ah_attr *attr); | |
484 | void ipoib_free_ah(struct kref *kref); | |
485 | static inline void ipoib_put_ah(struct ipoib_ah *ah) | |
486 | { | |
487 | kref_put(&ah->ref, ipoib_free_ah); | |
488 | } | |
489 | int ipoib_open(struct net_device *dev); | |
490 | int ipoib_add_pkey_attr(struct net_device *dev); | |
491 | int ipoib_add_umcast_attr(struct net_device *dev); | |
492 | ||
493 | int ipoib_send(struct net_device *dev, struct sk_buff *skb, | |
494 | struct ib_ah *address, u32 dqpn); | |
495 | void ipoib_reap_ah(struct work_struct *work); | |
496 | ||
497 | struct ipoib_path *__path_find(struct net_device *dev, void *gid); | |
498 | void ipoib_mark_paths_invalid(struct net_device *dev); | |
499 | void ipoib_flush_paths(struct net_device *dev); | |
500 | struct ipoib_dev_priv *ipoib_intf_alloc(struct ib_device *hca, u8 port, | |
501 | const char *format); | |
502 | void ipoib_ib_tx_timer_func(unsigned long ctx); | |
503 | void ipoib_ib_dev_flush_light(struct work_struct *work); | |
504 | void ipoib_ib_dev_flush_normal(struct work_struct *work); | |
505 | void ipoib_ib_dev_flush_heavy(struct work_struct *work); | |
506 | void ipoib_pkey_event(struct work_struct *work); | |
507 | void ipoib_ib_dev_cleanup(struct net_device *dev); | |
508 | ||
509 | int ipoib_ib_dev_open_default(struct net_device *dev); | |
510 | int ipoib_ib_dev_open(struct net_device *dev); | |
511 | int ipoib_ib_dev_stop(struct net_device *dev); | |
512 | void ipoib_ib_dev_up(struct net_device *dev); | |
513 | void ipoib_ib_dev_down(struct net_device *dev); | |
514 | int ipoib_ib_dev_stop_default(struct net_device *dev); | |
515 | void ipoib_pkey_dev_check_presence(struct net_device *dev); | |
516 | ||
517 | int ipoib_dev_init(struct net_device *dev, struct ib_device *ca, int port); | |
518 | void ipoib_dev_cleanup(struct net_device *dev); | |
519 | ||
520 | void ipoib_mcast_join_task(struct work_struct *work); | |
521 | void ipoib_mcast_carrier_on_task(struct work_struct *work); | |
522 | void ipoib_mcast_send(struct net_device *dev, u8 *daddr, struct sk_buff *skb); | |
523 | ||
524 | void ipoib_mcast_restart_task(struct work_struct *work); | |
525 | void ipoib_mcast_start_thread(struct net_device *dev); | |
526 | int ipoib_mcast_stop_thread(struct net_device *dev); | |
527 | ||
528 | void ipoib_mcast_dev_down(struct net_device *dev); | |
529 | void ipoib_mcast_dev_flush(struct net_device *dev); | |
530 | ||
531 | int ipoib_dma_map_tx(struct ib_device *ca, struct ipoib_tx_buf *tx_req); | |
532 | void ipoib_dma_unmap_tx(struct ipoib_dev_priv *priv, | |
533 | struct ipoib_tx_buf *tx_req); | |
534 | ||
535 | static inline void ipoib_build_sge(struct ipoib_dev_priv *priv, | |
536 | struct ipoib_tx_buf *tx_req) | |
537 | { | |
538 | int i, off; | |
539 | struct sk_buff *skb = tx_req->skb; | |
540 | skb_frag_t *frags = skb_shinfo(skb)->frags; | |
541 | int nr_frags = skb_shinfo(skb)->nr_frags; | |
542 | u64 *mapping = tx_req->mapping; | |
543 | ||
544 | if (skb_headlen(skb)) { | |
545 | priv->tx_sge[0].addr = mapping[0]; | |
546 | priv->tx_sge[0].length = skb_headlen(skb); | |
547 | off = 1; | |
548 | } else | |
549 | off = 0; | |
550 | ||
551 | for (i = 0; i < nr_frags; ++i) { | |
552 | priv->tx_sge[i + off].addr = mapping[i + off]; | |
553 | priv->tx_sge[i + off].length = skb_frag_size(&frags[i]); | |
554 | } | |
555 | priv->tx_wr.wr.num_sge = nr_frags + off; | |
556 | } | |
557 | ||
558 | #ifdef CONFIG_INFINIBAND_IPOIB_DEBUG | |
559 | struct ipoib_mcast_iter *ipoib_mcast_iter_init(struct net_device *dev); | |
560 | int ipoib_mcast_iter_next(struct ipoib_mcast_iter *iter); | |
561 | void ipoib_mcast_iter_read(struct ipoib_mcast_iter *iter, | |
562 | union ib_gid *gid, | |
563 | unsigned long *created, | |
564 | unsigned int *queuelen, | |
565 | unsigned int *complete, | |
566 | unsigned int *send_only); | |
567 | ||
568 | struct ipoib_path_iter *ipoib_path_iter_init(struct net_device *dev); | |
569 | int ipoib_path_iter_next(struct ipoib_path_iter *iter); | |
570 | void ipoib_path_iter_read(struct ipoib_path_iter *iter, | |
571 | struct ipoib_path *path); | |
572 | #endif | |
573 | ||
574 | int ipoib_mcast_attach(struct net_device *dev, struct ib_device *hca, | |
575 | union ib_gid *mgid, u16 mlid, int set_qkey, u32 qkey); | |
576 | int ipoib_mcast_detach(struct net_device *dev, struct ib_device *hca, | |
577 | union ib_gid *mgid, u16 mlid); | |
578 | void ipoib_mcast_remove_list(struct list_head *remove_list); | |
579 | void ipoib_check_and_add_mcast_sendonly(struct ipoib_dev_priv *priv, u8 *mgid, | |
580 | struct list_head *remove_list); | |
581 | ||
582 | int ipoib_init_qp(struct net_device *dev); | |
583 | int ipoib_transport_dev_init(struct net_device *dev, struct ib_device *ca); | |
584 | void ipoib_transport_dev_cleanup(struct net_device *dev); | |
585 | ||
586 | void ipoib_event(struct ib_event_handler *handler, | |
587 | struct ib_event *record); | |
588 | ||
589 | int ipoib_vlan_add(struct net_device *pdev, unsigned short pkey); | |
590 | int ipoib_vlan_delete(struct net_device *pdev, unsigned short pkey); | |
591 | ||
592 | int __ipoib_vlan_add(struct ipoib_dev_priv *ppriv, struct ipoib_dev_priv *priv, | |
593 | u16 pkey, int child_type); | |
594 | ||
595 | int __init ipoib_netlink_init(void); | |
596 | void __exit ipoib_netlink_fini(void); | |
597 | ||
598 | void ipoib_set_umcast(struct net_device *ndev, int umcast_val); | |
599 | int ipoib_set_mode(struct net_device *dev, const char *buf); | |
600 | ||
601 | void ipoib_setup_common(struct net_device *dev); | |
602 | ||
603 | void ipoib_pkey_open(struct ipoib_dev_priv *priv); | |
604 | void ipoib_drain_cq(struct net_device *dev); | |
605 | ||
606 | void ipoib_set_ethtool_ops(struct net_device *dev); | |
607 | void ipoib_set_dev_features(struct ipoib_dev_priv *priv, struct ib_device *hca); | |
608 | ||
609 | #define IPOIB_FLAGS_RC 0x80 | |
610 | #define IPOIB_FLAGS_UC 0x40 | |
611 | ||
612 | /* We don't support UC connections at the moment */ | |
613 | #define IPOIB_CM_SUPPORTED(ha) (ha[0] & (IPOIB_FLAGS_RC)) | |
614 | ||
615 | #ifdef CONFIG_INFINIBAND_IPOIB_CM | |
616 | ||
617 | extern int ipoib_max_conn_qp; | |
618 | ||
619 | static inline int ipoib_cm_admin_enabled(struct net_device *dev) | |
620 | { | |
621 | struct ipoib_dev_priv *priv = ipoib_priv(dev); | |
622 | return IPOIB_CM_SUPPORTED(dev->dev_addr) && | |
623 | test_bit(IPOIB_FLAG_ADMIN_CM, &priv->flags); | |
624 | } | |
625 | ||
626 | static inline int ipoib_cm_enabled(struct net_device *dev, u8 *hwaddr) | |
627 | { | |
628 | struct ipoib_dev_priv *priv = ipoib_priv(dev); | |
629 | return IPOIB_CM_SUPPORTED(hwaddr) && | |
630 | test_bit(IPOIB_FLAG_ADMIN_CM, &priv->flags); | |
631 | } | |
632 | ||
633 | static inline int ipoib_cm_up(struct ipoib_neigh *neigh) | |
634 | ||
635 | { | |
636 | return test_bit(IPOIB_FLAG_OPER_UP, &neigh->cm->flags); | |
637 | } | |
638 | ||
639 | static inline struct ipoib_cm_tx *ipoib_cm_get(struct ipoib_neigh *neigh) | |
640 | { | |
641 | return neigh->cm; | |
642 | } | |
643 | ||
644 | static inline void ipoib_cm_set(struct ipoib_neigh *neigh, struct ipoib_cm_tx *tx) | |
645 | { | |
646 | neigh->cm = tx; | |
647 | } | |
648 | ||
649 | static inline int ipoib_cm_has_srq(struct net_device *dev) | |
650 | { | |
651 | struct ipoib_dev_priv *priv = ipoib_priv(dev); | |
652 | return !!priv->cm.srq; | |
653 | } | |
654 | ||
655 | static inline unsigned int ipoib_cm_max_mtu(struct net_device *dev) | |
656 | { | |
657 | struct ipoib_dev_priv *priv = ipoib_priv(dev); | |
658 | return priv->cm.max_cm_mtu; | |
659 | } | |
660 | ||
661 | void ipoib_cm_send(struct net_device *dev, struct sk_buff *skb, struct ipoib_cm_tx *tx); | |
662 | int ipoib_cm_dev_open(struct net_device *dev); | |
663 | void ipoib_cm_dev_stop(struct net_device *dev); | |
664 | int ipoib_cm_dev_init(struct net_device *dev); | |
665 | int ipoib_cm_add_mode_attr(struct net_device *dev); | |
666 | void ipoib_cm_dev_cleanup(struct net_device *dev); | |
667 | struct ipoib_cm_tx *ipoib_cm_create_tx(struct net_device *dev, struct ipoib_path *path, | |
668 | struct ipoib_neigh *neigh); | |
669 | void ipoib_cm_destroy_tx(struct ipoib_cm_tx *tx); | |
670 | void ipoib_cm_skb_too_long(struct net_device *dev, struct sk_buff *skb, | |
671 | unsigned int mtu); | |
672 | void ipoib_cm_handle_rx_wc(struct net_device *dev, struct ib_wc *wc); | |
673 | void ipoib_cm_handle_tx_wc(struct net_device *dev, struct ib_wc *wc); | |
674 | #else | |
675 | ||
676 | struct ipoib_cm_tx; | |
677 | ||
678 | #define ipoib_max_conn_qp 0 | |
679 | ||
680 | static inline int ipoib_cm_admin_enabled(struct net_device *dev) | |
681 | { | |
682 | return 0; | |
683 | } | |
684 | static inline int ipoib_cm_enabled(struct net_device *dev, u8 *hwaddr) | |
685 | ||
686 | { | |
687 | return 0; | |
688 | } | |
689 | ||
690 | static inline int ipoib_cm_up(struct ipoib_neigh *neigh) | |
691 | ||
692 | { | |
693 | return 0; | |
694 | } | |
695 | ||
696 | static inline struct ipoib_cm_tx *ipoib_cm_get(struct ipoib_neigh *neigh) | |
697 | { | |
698 | return NULL; | |
699 | } | |
700 | ||
701 | static inline void ipoib_cm_set(struct ipoib_neigh *neigh, struct ipoib_cm_tx *tx) | |
702 | { | |
703 | } | |
704 | ||
705 | static inline int ipoib_cm_has_srq(struct net_device *dev) | |
706 | { | |
707 | return 0; | |
708 | } | |
709 | ||
710 | static inline unsigned int ipoib_cm_max_mtu(struct net_device *dev) | |
711 | { | |
712 | return 0; | |
713 | } | |
714 | ||
715 | static inline | |
716 | void ipoib_cm_send(struct net_device *dev, struct sk_buff *skb, struct ipoib_cm_tx *tx) | |
717 | { | |
718 | return; | |
719 | } | |
720 | ||
721 | static inline | |
722 | int ipoib_cm_dev_open(struct net_device *dev) | |
723 | { | |
724 | return 0; | |
725 | } | |
726 | ||
727 | static inline | |
728 | void ipoib_cm_dev_stop(struct net_device *dev) | |
729 | { | |
730 | return; | |
731 | } | |
732 | ||
733 | static inline | |
734 | int ipoib_cm_dev_init(struct net_device *dev) | |
735 | { | |
736 | return -ENOSYS; | |
737 | } | |
738 | ||
739 | static inline | |
740 | void ipoib_cm_dev_cleanup(struct net_device *dev) | |
741 | { | |
742 | return; | |
743 | } | |
744 | ||
745 | static inline | |
746 | struct ipoib_cm_tx *ipoib_cm_create_tx(struct net_device *dev, struct ipoib_path *path, | |
747 | struct ipoib_neigh *neigh) | |
748 | { | |
749 | return NULL; | |
750 | } | |
751 | ||
752 | static inline | |
753 | void ipoib_cm_destroy_tx(struct ipoib_cm_tx *tx) | |
754 | { | |
755 | return; | |
756 | } | |
757 | ||
758 | static inline | |
759 | int ipoib_cm_add_mode_attr(struct net_device *dev) | |
760 | { | |
761 | return 0; | |
762 | } | |
763 | ||
764 | static inline void ipoib_cm_skb_too_long(struct net_device *dev, struct sk_buff *skb, | |
765 | unsigned int mtu) | |
766 | { | |
767 | dev_kfree_skb_any(skb); | |
768 | } | |
769 | ||
770 | static inline void ipoib_cm_handle_rx_wc(struct net_device *dev, struct ib_wc *wc) | |
771 | { | |
772 | } | |
773 | ||
774 | static inline void ipoib_cm_handle_tx_wc(struct net_device *dev, struct ib_wc *wc) | |
775 | { | |
776 | } | |
777 | #endif | |
778 | ||
779 | #ifdef CONFIG_INFINIBAND_IPOIB_DEBUG | |
780 | void ipoib_create_debug_files(struct net_device *dev); | |
781 | void ipoib_delete_debug_files(struct net_device *dev); | |
782 | int ipoib_register_debugfs(void); | |
783 | void ipoib_unregister_debugfs(void); | |
784 | #else | |
785 | static inline void ipoib_create_debug_files(struct net_device *dev) { } | |
786 | static inline void ipoib_delete_debug_files(struct net_device *dev) { } | |
787 | static inline int ipoib_register_debugfs(void) { return 0; } | |
788 | static inline void ipoib_unregister_debugfs(void) { } | |
789 | #endif | |
790 | ||
791 | #define ipoib_printk(level, priv, format, arg...) \ | |
792 | printk(level "%s: " format, ((struct ipoib_dev_priv *) priv)->dev->name , ## arg) | |
793 | #define ipoib_warn(priv, format, arg...) \ | |
794 | do { \ | |
795 | static DEFINE_RATELIMIT_STATE(_rs, \ | |
796 | 10 * HZ /*10 seconds */, \ | |
797 | 100); \ | |
798 | if (__ratelimit(&_rs)) \ | |
799 | ipoib_printk(KERN_WARNING, priv, format , ## arg);\ | |
800 | } while (0) | |
801 | ||
802 | extern int ipoib_sendq_size; | |
803 | extern int ipoib_recvq_size; | |
804 | ||
805 | extern struct ib_sa_client ipoib_sa_client; | |
806 | ||
807 | #ifdef CONFIG_INFINIBAND_IPOIB_DEBUG | |
808 | extern int ipoib_debug_level; | |
809 | ||
810 | #define ipoib_dbg(priv, format, arg...) \ | |
811 | do { \ | |
812 | if (ipoib_debug_level > 0) \ | |
813 | ipoib_printk(KERN_DEBUG, priv, format , ## arg); \ | |
814 | } while (0) | |
815 | #define ipoib_dbg_mcast(priv, format, arg...) \ | |
816 | do { \ | |
817 | if (mcast_debug_level > 0) \ | |
818 | ipoib_printk(KERN_DEBUG, priv, format , ## arg); \ | |
819 | } while (0) | |
820 | #else /* CONFIG_INFINIBAND_IPOIB_DEBUG */ | |
821 | #define ipoib_dbg(priv, format, arg...) \ | |
822 | do { (void) (priv); } while (0) | |
823 | #define ipoib_dbg_mcast(priv, format, arg...) \ | |
824 | do { (void) (priv); } while (0) | |
825 | #endif /* CONFIG_INFINIBAND_IPOIB_DEBUG */ | |
826 | ||
827 | #ifdef CONFIG_INFINIBAND_IPOIB_DEBUG_DATA | |
828 | #define ipoib_dbg_data(priv, format, arg...) \ | |
829 | do { \ | |
830 | if (data_debug_level > 0) \ | |
831 | ipoib_printk(KERN_DEBUG, priv, format , ## arg); \ | |
832 | } while (0) | |
833 | #else /* CONFIG_INFINIBAND_IPOIB_DEBUG_DATA */ | |
834 | #define ipoib_dbg_data(priv, format, arg...) \ | |
835 | do { (void) (priv); } while (0) | |
836 | #endif /* CONFIG_INFINIBAND_IPOIB_DEBUG_DATA */ | |
837 | ||
838 | #define IPOIB_QPN(ha) (be32_to_cpup((__be32 *) ha) & 0xffffff) | |
839 | ||
840 | extern const char ipoib_driver_version[]; | |
841 | ||
842 | #endif /* _IPOIB_H */ |