]> git.proxmox.com Git - mirror_ubuntu-kernels.git/blob - include/rdma/ib_mad.h
Merge branches 'for-5.1/upstream-fixes', 'for-5.2/core', 'for-5.2/ish', 'for-5.2...
[mirror_ubuntu-kernels.git] / include / rdma / ib_mad.h
1 /*
2 * Copyright (c) 2004 Mellanox Technologies Ltd. All rights reserved.
3 * Copyright (c) 2004 Infinicon Corporation. All rights reserved.
4 * Copyright (c) 2004 Intel Corporation. All rights reserved.
5 * Copyright (c) 2004 Topspin Corporation. All rights reserved.
6 * Copyright (c) 2004-2006 Voltaire Corporation. All rights reserved.
7 *
8 * This software is available to you under a choice of one of two
9 * licenses. You may choose to be licensed under the terms of the GNU
10 * General Public License (GPL) Version 2, available from the file
11 * COPYING in the main directory of this source tree, or the
12 * OpenIB.org BSD license below:
13 *
14 * Redistribution and use in source and binary forms, with or
15 * without modification, are permitted provided that the following
16 * conditions are met:
17 *
18 * - Redistributions of source code must retain the above
19 * copyright notice, this list of conditions and the following
20 * disclaimer.
21 *
22 * - Redistributions in binary form must reproduce the above
23 * copyright notice, this list of conditions and the following
24 * disclaimer in the documentation and/or other materials
25 * provided with the distribution.
26 *
27 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
28 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
29 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
30 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
31 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
32 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
33 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
34 * SOFTWARE.
35 */
36
37 #if !defined(IB_MAD_H)
38 #define IB_MAD_H
39
40 #include <linux/list.h>
41
42 #include <rdma/ib_verbs.h>
43 #include <uapi/rdma/ib_user_mad.h>
44
45 /* Management base versions */
46 #define IB_MGMT_BASE_VERSION 1
47 #define OPA_MGMT_BASE_VERSION 0x80
48
49 #define OPA_SM_CLASS_VERSION 0x80
50
51 /* Management classes */
52 #define IB_MGMT_CLASS_SUBN_LID_ROUTED 0x01
53 #define IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE 0x81
54 #define IB_MGMT_CLASS_SUBN_ADM 0x03
55 #define IB_MGMT_CLASS_PERF_MGMT 0x04
56 #define IB_MGMT_CLASS_BM 0x05
57 #define IB_MGMT_CLASS_DEVICE_MGMT 0x06
58 #define IB_MGMT_CLASS_CM 0x07
59 #define IB_MGMT_CLASS_SNMP 0x08
60 #define IB_MGMT_CLASS_DEVICE_ADM 0x10
61 #define IB_MGMT_CLASS_BOOT_MGMT 0x11
62 #define IB_MGMT_CLASS_BIS 0x12
63 #define IB_MGMT_CLASS_CONG_MGMT 0x21
64 #define IB_MGMT_CLASS_VENDOR_RANGE2_START 0x30
65 #define IB_MGMT_CLASS_VENDOR_RANGE2_END 0x4F
66
67 #define IB_OPENIB_OUI (0x001405)
68
69 /* Management methods */
70 #define IB_MGMT_METHOD_GET 0x01
71 #define IB_MGMT_METHOD_SET 0x02
72 #define IB_MGMT_METHOD_GET_RESP 0x81
73 #define IB_MGMT_METHOD_SEND 0x03
74 #define IB_MGMT_METHOD_TRAP 0x05
75 #define IB_MGMT_METHOD_REPORT 0x06
76 #define IB_MGMT_METHOD_REPORT_RESP 0x86
77 #define IB_MGMT_METHOD_TRAP_REPRESS 0x07
78
79 #define IB_MGMT_METHOD_RESP 0x80
80 #define IB_BM_ATTR_MOD_RESP cpu_to_be32(1)
81
82 #define IB_MGMT_MAX_METHODS 128
83
84 /* MAD Status field bit masks */
85 #define IB_MGMT_MAD_STATUS_SUCCESS 0x0000
86 #define IB_MGMT_MAD_STATUS_BUSY 0x0001
87 #define IB_MGMT_MAD_STATUS_REDIRECT_REQD 0x0002
88 #define IB_MGMT_MAD_STATUS_BAD_VERSION 0x0004
89 #define IB_MGMT_MAD_STATUS_UNSUPPORTED_METHOD 0x0008
90 #define IB_MGMT_MAD_STATUS_UNSUPPORTED_METHOD_ATTRIB 0x000c
91 #define IB_MGMT_MAD_STATUS_INVALID_ATTRIB_VALUE 0x001c
92
93 /* RMPP information */
94 #define IB_MGMT_RMPP_VERSION 1
95
96 #define IB_MGMT_RMPP_TYPE_DATA 1
97 #define IB_MGMT_RMPP_TYPE_ACK 2
98 #define IB_MGMT_RMPP_TYPE_STOP 3
99 #define IB_MGMT_RMPP_TYPE_ABORT 4
100
101 #define IB_MGMT_RMPP_FLAG_ACTIVE 1
102 #define IB_MGMT_RMPP_FLAG_FIRST (1<<1)
103 #define IB_MGMT_RMPP_FLAG_LAST (1<<2)
104
105 #define IB_MGMT_RMPP_NO_RESPTIME 0x1F
106
107 #define IB_MGMT_RMPP_STATUS_SUCCESS 0
108 #define IB_MGMT_RMPP_STATUS_RESX 1
109 #define IB_MGMT_RMPP_STATUS_ABORT_MIN 118
110 #define IB_MGMT_RMPP_STATUS_T2L 118
111 #define IB_MGMT_RMPP_STATUS_BAD_LEN 119
112 #define IB_MGMT_RMPP_STATUS_BAD_SEG 120
113 #define IB_MGMT_RMPP_STATUS_BADT 121
114 #define IB_MGMT_RMPP_STATUS_W2S 122
115 #define IB_MGMT_RMPP_STATUS_S2B 123
116 #define IB_MGMT_RMPP_STATUS_BAD_STATUS 124
117 #define IB_MGMT_RMPP_STATUS_UNV 125
118 #define IB_MGMT_RMPP_STATUS_TMR 126
119 #define IB_MGMT_RMPP_STATUS_UNSPEC 127
120 #define IB_MGMT_RMPP_STATUS_ABORT_MAX 127
121
122 #define IB_QP0 0
123 #define IB_QP1 cpu_to_be32(1)
124 #define IB_QP1_QKEY 0x80010000
125 #define IB_QP_SET_QKEY 0x80000000
126
127 #define IB_DEFAULT_PKEY_PARTIAL 0x7FFF
128 #define IB_DEFAULT_PKEY_FULL 0xFFFF
129
130 /*
131 * Generic trap/notice types
132 */
133 #define IB_NOTICE_TYPE_FATAL 0x80
134 #define IB_NOTICE_TYPE_URGENT 0x81
135 #define IB_NOTICE_TYPE_SECURITY 0x82
136 #define IB_NOTICE_TYPE_SM 0x83
137 #define IB_NOTICE_TYPE_INFO 0x84
138
139 /*
140 * Generic trap/notice producers
141 */
142 #define IB_NOTICE_PROD_CA cpu_to_be16(1)
143 #define IB_NOTICE_PROD_SWITCH cpu_to_be16(2)
144 #define IB_NOTICE_PROD_ROUTER cpu_to_be16(3)
145 #define IB_NOTICE_PROD_CLASS_MGR cpu_to_be16(4)
146
147 enum {
148 IB_MGMT_MAD_HDR = 24,
149 IB_MGMT_MAD_DATA = 232,
150 IB_MGMT_RMPP_HDR = 36,
151 IB_MGMT_RMPP_DATA = 220,
152 IB_MGMT_VENDOR_HDR = 40,
153 IB_MGMT_VENDOR_DATA = 216,
154 IB_MGMT_SA_HDR = 56,
155 IB_MGMT_SA_DATA = 200,
156 IB_MGMT_DEVICE_HDR = 64,
157 IB_MGMT_DEVICE_DATA = 192,
158 IB_MGMT_MAD_SIZE = IB_MGMT_MAD_HDR + IB_MGMT_MAD_DATA,
159 OPA_MGMT_MAD_DATA = 2024,
160 OPA_MGMT_RMPP_DATA = 2012,
161 OPA_MGMT_MAD_SIZE = IB_MGMT_MAD_HDR + OPA_MGMT_MAD_DATA,
162 };
163
164 struct ib_mad_hdr {
165 u8 base_version;
166 u8 mgmt_class;
167 u8 class_version;
168 u8 method;
169 __be16 status;
170 __be16 class_specific;
171 __be64 tid;
172 __be16 attr_id;
173 __be16 resv;
174 __be32 attr_mod;
175 };
176
177 struct ib_rmpp_hdr {
178 u8 rmpp_version;
179 u8 rmpp_type;
180 u8 rmpp_rtime_flags;
181 u8 rmpp_status;
182 __be32 seg_num;
183 __be32 paylen_newwin;
184 };
185
186 typedef u64 __bitwise ib_sa_comp_mask;
187
188 #define IB_SA_COMP_MASK(n) ((__force ib_sa_comp_mask) cpu_to_be64(1ull << (n)))
189
190 /*
191 * ib_sa_hdr and ib_sa_mad structures must be packed because they have
192 * 64-bit fields that are only 32-bit aligned. 64-bit architectures will
193 * lay them out wrong otherwise. (And unfortunately they are sent on
194 * the wire so we can't change the layout)
195 */
196 struct ib_sa_hdr {
197 __be64 sm_key;
198 __be16 attr_offset;
199 __be16 reserved;
200 ib_sa_comp_mask comp_mask;
201 } __attribute__ ((packed));
202
203 struct ib_mad {
204 struct ib_mad_hdr mad_hdr;
205 u8 data[IB_MGMT_MAD_DATA];
206 };
207
208 struct opa_mad {
209 struct ib_mad_hdr mad_hdr;
210 u8 data[OPA_MGMT_MAD_DATA];
211 };
212
213 struct ib_rmpp_mad {
214 struct ib_mad_hdr mad_hdr;
215 struct ib_rmpp_hdr rmpp_hdr;
216 u8 data[IB_MGMT_RMPP_DATA];
217 };
218
219 struct opa_rmpp_mad {
220 struct ib_mad_hdr mad_hdr;
221 struct ib_rmpp_hdr rmpp_hdr;
222 u8 data[OPA_MGMT_RMPP_DATA];
223 };
224
225 struct ib_sa_mad {
226 struct ib_mad_hdr mad_hdr;
227 struct ib_rmpp_hdr rmpp_hdr;
228 struct ib_sa_hdr sa_hdr;
229 u8 data[IB_MGMT_SA_DATA];
230 } __attribute__ ((packed));
231
232 struct ib_vendor_mad {
233 struct ib_mad_hdr mad_hdr;
234 struct ib_rmpp_hdr rmpp_hdr;
235 u8 reserved;
236 u8 oui[3];
237 u8 data[IB_MGMT_VENDOR_DATA];
238 };
239
240 #define IB_MGMT_CLASSPORTINFO_ATTR_ID cpu_to_be16(0x0001)
241
242 #define IB_CLASS_PORT_INFO_RESP_TIME_MASK 0x1F
243 #define IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE 5
244
245 struct ib_class_port_info {
246 u8 base_version;
247 u8 class_version;
248 __be16 capability_mask;
249 /* 27 bits for cap_mask2, 5 bits for resp_time */
250 __be32 cap_mask2_resp_time;
251 u8 redirect_gid[16];
252 __be32 redirect_tcslfl;
253 __be16 redirect_lid;
254 __be16 redirect_pkey;
255 __be32 redirect_qp;
256 __be32 redirect_qkey;
257 u8 trap_gid[16];
258 __be32 trap_tcslfl;
259 __be16 trap_lid;
260 __be16 trap_pkey;
261 __be32 trap_hlqp;
262 __be32 trap_qkey;
263 };
264
265 /* PortInfo CapabilityMask */
266 enum ib_port_capability_mask_bits {
267 IB_PORT_SM = 1 << 1,
268 IB_PORT_NOTICE_SUP = 1 << 2,
269 IB_PORT_TRAP_SUP = 1 << 3,
270 IB_PORT_OPT_IPD_SUP = 1 << 4,
271 IB_PORT_AUTO_MIGR_SUP = 1 << 5,
272 IB_PORT_SL_MAP_SUP = 1 << 6,
273 IB_PORT_MKEY_NVRAM = 1 << 7,
274 IB_PORT_PKEY_NVRAM = 1 << 8,
275 IB_PORT_LED_INFO_SUP = 1 << 9,
276 IB_PORT_SM_DISABLED = 1 << 10,
277 IB_PORT_SYS_IMAGE_GUID_SUP = 1 << 11,
278 IB_PORT_PKEY_SW_EXT_PORT_TRAP_SUP = 1 << 12,
279 IB_PORT_EXTENDED_SPEEDS_SUP = 1 << 14,
280 IB_PORT_CAP_MASK2_SUP = 1 << 15,
281 IB_PORT_CM_SUP = 1 << 16,
282 IB_PORT_SNMP_TUNNEL_SUP = 1 << 17,
283 IB_PORT_REINIT_SUP = 1 << 18,
284 IB_PORT_DEVICE_MGMT_SUP = 1 << 19,
285 IB_PORT_VENDOR_CLASS_SUP = 1 << 20,
286 IB_PORT_DR_NOTICE_SUP = 1 << 21,
287 IB_PORT_CAP_MASK_NOTICE_SUP = 1 << 22,
288 IB_PORT_BOOT_MGMT_SUP = 1 << 23,
289 IB_PORT_LINK_LATENCY_SUP = 1 << 24,
290 IB_PORT_CLIENT_REG_SUP = 1 << 25,
291 IB_PORT_OTHER_LOCAL_CHANGES_SUP = 1 << 26,
292 IB_PORT_LINK_SPEED_WIDTH_TABLE_SUP = 1 << 27,
293 IB_PORT_VENDOR_SPECIFIC_MADS_TABLE_SUP = 1 << 28,
294 IB_PORT_MCAST_PKEY_TRAP_SUPPRESSION_SUP = 1 << 29,
295 IB_PORT_MCAST_FDB_TOP_SUP = 1 << 30,
296 IB_PORT_HIERARCHY_INFO_SUP = 1ULL << 31,
297 };
298
299 enum ib_port_capability_mask2_bits {
300 IB_PORT_SET_NODE_DESC_SUP = 1 << 0,
301 IB_PORT_EX_PORT_INFO_EX_SUP = 1 << 1,
302 IB_PORT_VIRT_SUP = 1 << 2,
303 IB_PORT_SWITCH_PORT_STATE_TABLE_SUP = 1 << 3,
304 IB_PORT_LINK_WIDTH_2X_SUP = 1 << 4,
305 IB_PORT_LINK_SPEED_HDR_SUP = 1 << 5,
306 };
307
308 #define OPA_CLASS_PORT_INFO_PR_SUPPORT BIT(26)
309
310 struct opa_class_port_info {
311 u8 base_version;
312 u8 class_version;
313 __be16 cap_mask;
314 __be32 cap_mask2_resp_time;
315
316 u8 redirect_gid[16];
317 __be32 redirect_tc_fl;
318 __be32 redirect_lid;
319 __be32 redirect_sl_qp;
320 __be32 redirect_qkey;
321
322 u8 trap_gid[16];
323 __be32 trap_tc_fl;
324 __be32 trap_lid;
325 __be32 trap_hl_qp;
326 __be32 trap_qkey;
327
328 __be16 trap_pkey;
329 __be16 redirect_pkey;
330
331 u8 trap_sl_rsvd;
332 u8 reserved[3];
333 } __packed;
334
335 /**
336 * ib_get_cpi_resp_time - Returns the resp_time value from
337 * cap_mask2_resp_time in ib_class_port_info.
338 * @cpi: A struct ib_class_port_info mad.
339 */
340 static inline u8 ib_get_cpi_resp_time(struct ib_class_port_info *cpi)
341 {
342 return (u8)(be32_to_cpu(cpi->cap_mask2_resp_time) &
343 IB_CLASS_PORT_INFO_RESP_TIME_MASK);
344 }
345
346 /**
347 * ib_set_cpi_resptime - Sets the response time in an
348 * ib_class_port_info mad.
349 * @cpi: A struct ib_class_port_info.
350 * @rtime: The response time to set.
351 */
352 static inline void ib_set_cpi_resp_time(struct ib_class_port_info *cpi,
353 u8 rtime)
354 {
355 cpi->cap_mask2_resp_time =
356 (cpi->cap_mask2_resp_time &
357 cpu_to_be32(~IB_CLASS_PORT_INFO_RESP_TIME_MASK)) |
358 cpu_to_be32(rtime & IB_CLASS_PORT_INFO_RESP_TIME_MASK);
359 }
360
361 /**
362 * ib_get_cpi_capmask2 - Returns the capmask2 value from
363 * cap_mask2_resp_time in ib_class_port_info.
364 * @cpi: A struct ib_class_port_info mad.
365 */
366 static inline u32 ib_get_cpi_capmask2(struct ib_class_port_info *cpi)
367 {
368 return (be32_to_cpu(cpi->cap_mask2_resp_time) >>
369 IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE);
370 }
371
372 /**
373 * ib_set_cpi_capmask2 - Sets the capmask2 in an
374 * ib_class_port_info mad.
375 * @cpi: A struct ib_class_port_info.
376 * @capmask2: The capmask2 to set.
377 */
378 static inline void ib_set_cpi_capmask2(struct ib_class_port_info *cpi,
379 u32 capmask2)
380 {
381 cpi->cap_mask2_resp_time =
382 (cpi->cap_mask2_resp_time &
383 cpu_to_be32(IB_CLASS_PORT_INFO_RESP_TIME_MASK)) |
384 cpu_to_be32(capmask2 <<
385 IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE);
386 }
387
388 /**
389 * opa_get_cpi_capmask2 - Returns the capmask2 value from
390 * cap_mask2_resp_time in ib_class_port_info.
391 * @cpi: A struct opa_class_port_info mad.
392 */
393 static inline u32 opa_get_cpi_capmask2(struct opa_class_port_info *cpi)
394 {
395 return (be32_to_cpu(cpi->cap_mask2_resp_time) >>
396 IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE);
397 }
398
399 struct ib_mad_notice_attr {
400 u8 generic_type;
401 u8 prod_type_msb;
402 __be16 prod_type_lsb;
403 __be16 trap_num;
404 __be16 issuer_lid;
405 __be16 toggle_count;
406
407 union {
408 struct {
409 u8 details[54];
410 } raw_data;
411
412 struct {
413 __be16 reserved;
414 __be16 lid; /* where violation happened */
415 u8 port_num; /* where violation happened */
416 } __packed ntc_129_131;
417
418 struct {
419 __be16 reserved;
420 __be16 lid; /* LID where change occurred */
421 u8 reserved2;
422 u8 local_changes; /* low bit - local changes */
423 __be32 new_cap_mask; /* new capability mask */
424 u8 reserved3;
425 u8 change_flags; /* low 3 bits only */
426 } __packed ntc_144;
427
428 struct {
429 __be16 reserved;
430 __be16 lid; /* lid where sys guid changed */
431 __be16 reserved2;
432 __be64 new_sys_guid;
433 } __packed ntc_145;
434
435 struct {
436 __be16 reserved;
437 __be16 lid;
438 __be16 dr_slid;
439 u8 method;
440 u8 reserved2;
441 __be16 attr_id;
442 __be32 attr_mod;
443 __be64 mkey;
444 u8 reserved3;
445 u8 dr_trunc_hop;
446 u8 dr_rtn_path[30];
447 } __packed ntc_256;
448
449 struct {
450 __be16 reserved;
451 __be16 lid1;
452 __be16 lid2;
453 __be32 key;
454 __be32 sl_qp1; /* SL: high 4 bits */
455 __be32 qp2; /* high 8 bits reserved */
456 union ib_gid gid1;
457 union ib_gid gid2;
458 } __packed ntc_257_258;
459
460 } details;
461 };
462
463 /**
464 * ib_mad_send_buf - MAD data buffer and work request for sends.
465 * @next: A pointer used to chain together MADs for posting.
466 * @mad: References an allocated MAD data buffer for MADs that do not have
467 * RMPP active. For MADs using RMPP, references the common and management
468 * class specific headers.
469 * @mad_agent: MAD agent that allocated the buffer.
470 * @ah: The address handle to use when sending the MAD.
471 * @context: User-controlled context fields.
472 * @hdr_len: Indicates the size of the data header of the MAD. This length
473 * includes the common MAD, RMPP, and class specific headers.
474 * @data_len: Indicates the total size of user-transferred data.
475 * @seg_count: The number of RMPP segments allocated for this send.
476 * @seg_size: Size of the data in each RMPP segment. This does not include
477 * class specific headers.
478 * @seg_rmpp_size: Size of each RMPP segment including the class specific
479 * headers.
480 * @timeout_ms: Time to wait for a response.
481 * @retries: Number of times to retry a request for a response. For MADs
482 * using RMPP, this applies per window. On completion, returns the number
483 * of retries needed to complete the transfer.
484 *
485 * Users are responsible for initializing the MAD buffer itself, with the
486 * exception of any RMPP header. Additional segment buffer space allocated
487 * beyond data_len is padding.
488 */
489 struct ib_mad_send_buf {
490 struct ib_mad_send_buf *next;
491 void *mad;
492 struct ib_mad_agent *mad_agent;
493 struct ib_ah *ah;
494 void *context[2];
495 int hdr_len;
496 int data_len;
497 int seg_count;
498 int seg_size;
499 int seg_rmpp_size;
500 int timeout_ms;
501 int retries;
502 };
503
504 /**
505 * ib_response_mad - Returns if the specified MAD has been generated in
506 * response to a sent request or trap.
507 */
508 int ib_response_mad(const struct ib_mad_hdr *hdr);
509
510 /**
511 * ib_get_rmpp_resptime - Returns the RMPP response time.
512 * @rmpp_hdr: An RMPP header.
513 */
514 static inline u8 ib_get_rmpp_resptime(struct ib_rmpp_hdr *rmpp_hdr)
515 {
516 return rmpp_hdr->rmpp_rtime_flags >> 3;
517 }
518
519 /**
520 * ib_get_rmpp_flags - Returns the RMPP flags.
521 * @rmpp_hdr: An RMPP header.
522 */
523 static inline u8 ib_get_rmpp_flags(struct ib_rmpp_hdr *rmpp_hdr)
524 {
525 return rmpp_hdr->rmpp_rtime_flags & 0x7;
526 }
527
528 /**
529 * ib_set_rmpp_resptime - Sets the response time in an RMPP header.
530 * @rmpp_hdr: An RMPP header.
531 * @rtime: The response time to set.
532 */
533 static inline void ib_set_rmpp_resptime(struct ib_rmpp_hdr *rmpp_hdr, u8 rtime)
534 {
535 rmpp_hdr->rmpp_rtime_flags = ib_get_rmpp_flags(rmpp_hdr) | (rtime << 3);
536 }
537
538 /**
539 * ib_set_rmpp_flags - Sets the flags in an RMPP header.
540 * @rmpp_hdr: An RMPP header.
541 * @flags: The flags to set.
542 */
543 static inline void ib_set_rmpp_flags(struct ib_rmpp_hdr *rmpp_hdr, u8 flags)
544 {
545 rmpp_hdr->rmpp_rtime_flags = (rmpp_hdr->rmpp_rtime_flags & 0xF8) |
546 (flags & 0x7);
547 }
548
549 struct ib_mad_agent;
550 struct ib_mad_send_wc;
551 struct ib_mad_recv_wc;
552
553 /**
554 * ib_mad_send_handler - callback handler for a sent MAD.
555 * @mad_agent: MAD agent that sent the MAD.
556 * @mad_send_wc: Send work completion information on the sent MAD.
557 */
558 typedef void (*ib_mad_send_handler)(struct ib_mad_agent *mad_agent,
559 struct ib_mad_send_wc *mad_send_wc);
560
561 /**
562 * ib_mad_snoop_handler - Callback handler for snooping sent MADs.
563 * @mad_agent: MAD agent that snooped the MAD.
564 * @send_buf: send MAD data buffer.
565 * @mad_send_wc: Work completion information on the sent MAD. Valid
566 * only for snooping that occurs on a send completion.
567 *
568 * Clients snooping MADs should not modify data referenced by the @send_buf
569 * or @mad_send_wc.
570 */
571 typedef void (*ib_mad_snoop_handler)(struct ib_mad_agent *mad_agent,
572 struct ib_mad_send_buf *send_buf,
573 struct ib_mad_send_wc *mad_send_wc);
574
575 /**
576 * ib_mad_recv_handler - callback handler for a received MAD.
577 * @mad_agent: MAD agent requesting the received MAD.
578 * @send_buf: Send buffer if found, else NULL
579 * @mad_recv_wc: Received work completion information on the received MAD.
580 *
581 * MADs received in response to a send request operation will be handed to
582 * the user before the send operation completes. All data buffers given
583 * to registered agents through this routine are owned by the receiving
584 * client, except for snooping agents. Clients snooping MADs should not
585 * modify the data referenced by @mad_recv_wc.
586 */
587 typedef void (*ib_mad_recv_handler)(struct ib_mad_agent *mad_agent,
588 struct ib_mad_send_buf *send_buf,
589 struct ib_mad_recv_wc *mad_recv_wc);
590
591 /**
592 * ib_mad_agent - Used to track MAD registration with the access layer.
593 * @device: Reference to device registration is on.
594 * @qp: Reference to QP used for sending and receiving MADs.
595 * @mr: Memory region for system memory usable for DMA.
596 * @recv_handler: Callback handler for a received MAD.
597 * @send_handler: Callback handler for a sent MAD.
598 * @snoop_handler: Callback handler for snooped sent MADs.
599 * @context: User-specified context associated with this registration.
600 * @hi_tid: Access layer assigned transaction ID for this client.
601 * Unsolicited MADs sent by this client will have the upper 32-bits
602 * of their TID set to this value.
603 * @flags: registration flags
604 * @port_num: Port number on which QP is registered
605 * @rmpp_version: If set, indicates the RMPP version used by this agent.
606 */
607 enum {
608 IB_MAD_USER_RMPP = IB_USER_MAD_USER_RMPP,
609 };
610 struct ib_mad_agent {
611 struct ib_device *device;
612 struct ib_qp *qp;
613 ib_mad_recv_handler recv_handler;
614 ib_mad_send_handler send_handler;
615 ib_mad_snoop_handler snoop_handler;
616 void *context;
617 u32 hi_tid;
618 u32 flags;
619 void *security;
620 struct list_head mad_agent_sec_list;
621 u8 port_num;
622 u8 rmpp_version;
623 bool smp_allowed;
624 };
625
626 /**
627 * ib_mad_send_wc - MAD send completion information.
628 * @send_buf: Send MAD data buffer associated with the send MAD request.
629 * @status: Completion status.
630 * @vendor_err: Optional vendor error information returned with a failed
631 * request.
632 */
633 struct ib_mad_send_wc {
634 struct ib_mad_send_buf *send_buf;
635 enum ib_wc_status status;
636 u32 vendor_err;
637 };
638
639 /**
640 * ib_mad_recv_buf - received MAD buffer information.
641 * @list: Reference to next data buffer for a received RMPP MAD.
642 * @grh: References a data buffer containing the global route header.
643 * The data refereced by this buffer is only valid if the GRH is
644 * valid.
645 * @mad: References the start of the received MAD.
646 */
647 struct ib_mad_recv_buf {
648 struct list_head list;
649 struct ib_grh *grh;
650 union {
651 struct ib_mad *mad;
652 struct opa_mad *opa_mad;
653 };
654 };
655
656 /**
657 * ib_mad_recv_wc - received MAD information.
658 * @wc: Completion information for the received data.
659 * @recv_buf: Specifies the location of the received data buffer(s).
660 * @rmpp_list: Specifies a list of RMPP reassembled received MAD buffers.
661 * @mad_len: The length of the received MAD, without duplicated headers.
662 * @mad_seg_size: The size of individual MAD segments
663 *
664 * For received response, the wr_id contains a pointer to the ib_mad_send_buf
665 * for the corresponding send request.
666 */
667 struct ib_mad_recv_wc {
668 struct ib_wc *wc;
669 struct ib_mad_recv_buf recv_buf;
670 struct list_head rmpp_list;
671 int mad_len;
672 size_t mad_seg_size;
673 };
674
675 /**
676 * ib_mad_reg_req - MAD registration request
677 * @mgmt_class: Indicates which management class of MADs should be receive
678 * by the caller. This field is only required if the user wishes to
679 * receive unsolicited MADs, otherwise it should be 0.
680 * @mgmt_class_version: Indicates which version of MADs for the given
681 * management class to receive.
682 * @oui: Indicates IEEE OUI when mgmt_class is a vendor class
683 * in the range from 0x30 to 0x4f. Otherwise not used.
684 * @method_mask: The caller will receive unsolicited MADs for any method
685 * where @method_mask = 1.
686 *
687 */
688 struct ib_mad_reg_req {
689 u8 mgmt_class;
690 u8 mgmt_class_version;
691 u8 oui[3];
692 DECLARE_BITMAP(method_mask, IB_MGMT_MAX_METHODS);
693 };
694
695 /**
696 * ib_register_mad_agent - Register to send/receive MADs.
697 * @device: The device to register with.
698 * @port_num: The port on the specified device to use.
699 * @qp_type: Specifies which QP to access. Must be either
700 * IB_QPT_SMI or IB_QPT_GSI.
701 * @mad_reg_req: Specifies which unsolicited MADs should be received
702 * by the caller. This parameter may be NULL if the caller only
703 * wishes to receive solicited responses.
704 * @rmpp_version: If set, indicates that the client will send
705 * and receive MADs that contain the RMPP header for the given version.
706 * If set to 0, indicates that RMPP is not used by this client.
707 * @send_handler: The completion callback routine invoked after a send
708 * request has completed.
709 * @recv_handler: The completion callback routine invoked for a received
710 * MAD.
711 * @context: User specified context associated with the registration.
712 * @registration_flags: Registration flags to set for this agent
713 */
714 struct ib_mad_agent *ib_register_mad_agent(struct ib_device *device,
715 u8 port_num,
716 enum ib_qp_type qp_type,
717 struct ib_mad_reg_req *mad_reg_req,
718 u8 rmpp_version,
719 ib_mad_send_handler send_handler,
720 ib_mad_recv_handler recv_handler,
721 void *context,
722 u32 registration_flags);
723
724 enum ib_mad_snoop_flags {
725 /*IB_MAD_SNOOP_POSTED_SENDS = 1,*/
726 /*IB_MAD_SNOOP_RMPP_SENDS = (1<<1),*/
727 IB_MAD_SNOOP_SEND_COMPLETIONS = (1<<2),
728 /*IB_MAD_SNOOP_RMPP_SEND_COMPLETIONS = (1<<3),*/
729 IB_MAD_SNOOP_RECVS = (1<<4)
730 /*IB_MAD_SNOOP_RMPP_RECVS = (1<<5),*/
731 /*IB_MAD_SNOOP_REDIRECTED_QPS = (1<<6)*/
732 };
733
734 /**
735 * ib_register_mad_snoop - Register to snoop sent and received MADs.
736 * @device: The device to register with.
737 * @port_num: The port on the specified device to use.
738 * @qp_type: Specifies which QP traffic to snoop. Must be either
739 * IB_QPT_SMI or IB_QPT_GSI.
740 * @mad_snoop_flags: Specifies information where snooping occurs.
741 * @send_handler: The callback routine invoked for a snooped send.
742 * @recv_handler: The callback routine invoked for a snooped receive.
743 * @context: User specified context associated with the registration.
744 */
745 struct ib_mad_agent *ib_register_mad_snoop(struct ib_device *device,
746 u8 port_num,
747 enum ib_qp_type qp_type,
748 int mad_snoop_flags,
749 ib_mad_snoop_handler snoop_handler,
750 ib_mad_recv_handler recv_handler,
751 void *context);
752
753 /**
754 * ib_unregister_mad_agent - Unregisters a client from using MAD services.
755 * @mad_agent: Corresponding MAD registration request to deregister.
756 *
757 * After invoking this routine, MAD services are no longer usable by the
758 * client on the associated QP.
759 */
760 void ib_unregister_mad_agent(struct ib_mad_agent *mad_agent);
761
762 /**
763 * ib_post_send_mad - Posts MAD(s) to the send queue of the QP associated
764 * with the registered client.
765 * @send_buf: Specifies the information needed to send the MAD(s).
766 * @bad_send_buf: Specifies the MAD on which an error was encountered. This
767 * parameter is optional if only a single MAD is posted.
768 *
769 * Sent MADs are not guaranteed to complete in the order that they were posted.
770 *
771 * If the MAD requires RMPP, the data buffer should contain a single copy
772 * of the common MAD, RMPP, and class specific headers, followed by the class
773 * defined data. If the class defined data would not divide evenly into
774 * RMPP segments, then space must be allocated at the end of the referenced
775 * buffer for any required padding. To indicate the amount of class defined
776 * data being transferred, the paylen_newwin field in the RMPP header should
777 * be set to the size of the class specific header plus the amount of class
778 * defined data being transferred. The paylen_newwin field should be
779 * specified in network-byte order.
780 */
781 int ib_post_send_mad(struct ib_mad_send_buf *send_buf,
782 struct ib_mad_send_buf **bad_send_buf);
783
784
785 /**
786 * ib_free_recv_mad - Returns data buffers used to receive a MAD.
787 * @mad_recv_wc: Work completion information for a received MAD.
788 *
789 * Clients receiving MADs through their ib_mad_recv_handler must call this
790 * routine to return the work completion buffers to the access layer.
791 */
792 void ib_free_recv_mad(struct ib_mad_recv_wc *mad_recv_wc);
793
794 /**
795 * ib_cancel_mad - Cancels an outstanding send MAD operation.
796 * @mad_agent: Specifies the registration associated with sent MAD.
797 * @send_buf: Indicates the MAD to cancel.
798 *
799 * MADs will be returned to the user through the corresponding
800 * ib_mad_send_handler.
801 */
802 void ib_cancel_mad(struct ib_mad_agent *mad_agent,
803 struct ib_mad_send_buf *send_buf);
804
805 /**
806 * ib_modify_mad - Modifies an outstanding send MAD operation.
807 * @mad_agent: Specifies the registration associated with sent MAD.
808 * @send_buf: Indicates the MAD to modify.
809 * @timeout_ms: New timeout value for sent MAD.
810 *
811 * This call will reset the timeout value for a sent MAD to the specified
812 * value.
813 */
814 int ib_modify_mad(struct ib_mad_agent *mad_agent,
815 struct ib_mad_send_buf *send_buf, u32 timeout_ms);
816
817 /**
818 * ib_redirect_mad_qp - Registers a QP for MAD services.
819 * @qp: Reference to a QP that requires MAD services.
820 * @rmpp_version: If set, indicates that the client will send
821 * and receive MADs that contain the RMPP header for the given version.
822 * If set to 0, indicates that RMPP is not used by this client.
823 * @send_handler: The completion callback routine invoked after a send
824 * request has completed.
825 * @recv_handler: The completion callback routine invoked for a received
826 * MAD.
827 * @context: User specified context associated with the registration.
828 *
829 * Use of this call allows clients to use MAD services, such as RMPP,
830 * on user-owned QPs. After calling this routine, users may send
831 * MADs on the specified QP by calling ib_mad_post_send.
832 */
833 struct ib_mad_agent *ib_redirect_mad_qp(struct ib_qp *qp,
834 u8 rmpp_version,
835 ib_mad_send_handler send_handler,
836 ib_mad_recv_handler recv_handler,
837 void *context);
838
839 /**
840 * ib_process_mad_wc - Processes a work completion associated with a
841 * MAD sent or received on a redirected QP.
842 * @mad_agent: Specifies the registered MAD service using the redirected QP.
843 * @wc: References a work completion associated with a sent or received
844 * MAD segment.
845 *
846 * This routine is used to complete or continue processing on a MAD request.
847 * If the work completion is associated with a send operation, calling
848 * this routine is required to continue an RMPP transfer or to wait for a
849 * corresponding response, if it is a request. If the work completion is
850 * associated with a receive operation, calling this routine is required to
851 * process an inbound or outbound RMPP transfer, or to match a response MAD
852 * with its corresponding request.
853 */
854 int ib_process_mad_wc(struct ib_mad_agent *mad_agent,
855 struct ib_wc *wc);
856
857 /**
858 * ib_create_send_mad - Allocate and initialize a data buffer and work request
859 * for sending a MAD.
860 * @mad_agent: Specifies the registered MAD service to associate with the MAD.
861 * @remote_qpn: Specifies the QPN of the receiving node.
862 * @pkey_index: Specifies which PKey the MAD will be sent using. This field
863 * is valid only if the remote_qpn is QP 1.
864 * @rmpp_active: Indicates if the send will enable RMPP.
865 * @hdr_len: Indicates the size of the data header of the MAD. This length
866 * should include the common MAD header, RMPP header, plus any class
867 * specific header.
868 * @data_len: Indicates the size of any user-transferred data. The call will
869 * automatically adjust the allocated buffer size to account for any
870 * additional padding that may be necessary.
871 * @gfp_mask: GFP mask used for the memory allocation.
872 * @base_version: Base Version of this MAD
873 *
874 * This routine allocates a MAD for sending. The returned MAD send buffer
875 * will reference a data buffer usable for sending a MAD, along
876 * with an initialized work request structure. Users may modify the returned
877 * MAD data buffer before posting the send.
878 *
879 * The returned MAD header, class specific headers, and any padding will be
880 * cleared. Users are responsible for initializing the common MAD header,
881 * any class specific header, and MAD data area.
882 * If @rmpp_active is set, the RMPP header will be initialized for sending.
883 */
884 struct ib_mad_send_buf *ib_create_send_mad(struct ib_mad_agent *mad_agent,
885 u32 remote_qpn, u16 pkey_index,
886 int rmpp_active,
887 int hdr_len, int data_len,
888 gfp_t gfp_mask,
889 u8 base_version);
890
891 /**
892 * ib_is_mad_class_rmpp - returns whether given management class
893 * supports RMPP.
894 * @mgmt_class: management class
895 *
896 * This routine returns whether the management class supports RMPP.
897 */
898 int ib_is_mad_class_rmpp(u8 mgmt_class);
899
900 /**
901 * ib_get_mad_data_offset - returns the data offset for a given
902 * management class.
903 * @mgmt_class: management class
904 *
905 * This routine returns the data offset in the MAD for the management
906 * class requested.
907 */
908 int ib_get_mad_data_offset(u8 mgmt_class);
909
910 /**
911 * ib_get_rmpp_segment - returns the data buffer for a given RMPP segment.
912 * @send_buf: Previously allocated send data buffer.
913 * @seg_num: number of segment to return
914 *
915 * This routine returns a pointer to the data buffer of an RMPP MAD.
916 * Users must provide synchronization to @send_buf around this call.
917 */
918 void *ib_get_rmpp_segment(struct ib_mad_send_buf *send_buf, int seg_num);
919
920 /**
921 * ib_free_send_mad - Returns data buffers used to send a MAD.
922 * @send_buf: Previously allocated send data buffer.
923 */
924 void ib_free_send_mad(struct ib_mad_send_buf *send_buf);
925
926 /**
927 * ib_mad_kernel_rmpp_agent - Returns if the agent is performing RMPP.
928 * @agent: the agent in question
929 * @return: true if agent is performing rmpp, false otherwise.
930 */
931 int ib_mad_kernel_rmpp_agent(const struct ib_mad_agent *agent);
932
933 #endif /* IB_MAD_H */