Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
torvalds
GitHub Repository: torvalds/linux
Path: blob/master/include/rdma/ib_mad.h
29509 views
1
/* SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB */
2
/*
3
* Copyright (c) 2004 Mellanox Technologies Ltd. All rights reserved.
4
* Copyright (c) 2004 Infinicon Corporation. All rights reserved.
5
* Copyright (c) 2004 Intel Corporation. All rights reserved.
6
* Copyright (c) 2004 Topspin Corporation. All rights reserved.
7
* Copyright (c) 2004-2006 Voltaire Corporation. All rights reserved.
8
*/
9
10
#ifndef IB_MAD_H
11
#define IB_MAD_H
12
13
#include <linux/list.h>
14
15
#include <rdma/ib_verbs.h>
16
#include <uapi/rdma/ib_user_mad.h>
17
18
/* Management base versions */
19
#define IB_MGMT_BASE_VERSION 1
20
#define OPA_MGMT_BASE_VERSION 0x80
21
22
#define OPA_SM_CLASS_VERSION 0x80
23
24
/* Management classes */
25
#define IB_MGMT_CLASS_SUBN_LID_ROUTED 0x01
26
#define IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE 0x81
27
#define IB_MGMT_CLASS_SUBN_ADM 0x03
28
#define IB_MGMT_CLASS_PERF_MGMT 0x04
29
#define IB_MGMT_CLASS_BM 0x05
30
#define IB_MGMT_CLASS_DEVICE_MGMT 0x06
31
#define IB_MGMT_CLASS_CM 0x07
32
#define IB_MGMT_CLASS_SNMP 0x08
33
#define IB_MGMT_CLASS_DEVICE_ADM 0x10
34
#define IB_MGMT_CLASS_BOOT_MGMT 0x11
35
#define IB_MGMT_CLASS_BIS 0x12
36
#define IB_MGMT_CLASS_CONG_MGMT 0x21
37
#define IB_MGMT_CLASS_VENDOR_RANGE2_START 0x30
38
#define IB_MGMT_CLASS_VENDOR_RANGE2_END 0x4F
39
40
#define IB_OPENIB_OUI (0x001405)
41
42
/* Management methods */
43
#define IB_MGMT_METHOD_GET 0x01
44
#define IB_MGMT_METHOD_SET 0x02
45
#define IB_MGMT_METHOD_GET_RESP 0x81
46
#define IB_MGMT_METHOD_SEND 0x03
47
#define IB_MGMT_METHOD_TRAP 0x05
48
#define IB_MGMT_METHOD_REPORT 0x06
49
#define IB_MGMT_METHOD_REPORT_RESP 0x86
50
#define IB_MGMT_METHOD_TRAP_REPRESS 0x07
51
#define IB_MGMT_METHOD_GET_TABLE 0x12
52
53
#define IB_MGMT_METHOD_RESP 0x80
54
#define IB_BM_ATTR_MOD_RESP cpu_to_be32(1)
55
56
#define IB_MGMT_MAX_METHODS 128
57
58
/* MAD Status field bit masks */
59
#define IB_MGMT_MAD_STATUS_SUCCESS 0x0000
60
#define IB_MGMT_MAD_STATUS_BUSY 0x0001
61
#define IB_MGMT_MAD_STATUS_REDIRECT_REQD 0x0002
62
#define IB_MGMT_MAD_STATUS_BAD_VERSION 0x0004
63
#define IB_MGMT_MAD_STATUS_UNSUPPORTED_METHOD 0x0008
64
#define IB_MGMT_MAD_STATUS_UNSUPPORTED_METHOD_ATTRIB 0x000c
65
#define IB_MGMT_MAD_STATUS_INVALID_ATTRIB_VALUE 0x001c
66
67
/* RMPP information */
68
#define IB_MGMT_RMPP_VERSION 1
69
70
#define IB_MGMT_RMPP_TYPE_DATA 1
71
#define IB_MGMT_RMPP_TYPE_ACK 2
72
#define IB_MGMT_RMPP_TYPE_STOP 3
73
#define IB_MGMT_RMPP_TYPE_ABORT 4
74
75
#define IB_MGMT_RMPP_FLAG_ACTIVE 1
76
#define IB_MGMT_RMPP_FLAG_FIRST (1<<1)
77
#define IB_MGMT_RMPP_FLAG_LAST (1<<2)
78
79
#define IB_MGMT_RMPP_NO_RESPTIME 0x1F
80
81
#define IB_MGMT_RMPP_STATUS_SUCCESS 0
82
#define IB_MGMT_RMPP_STATUS_RESX 1
83
#define IB_MGMT_RMPP_STATUS_ABORT_MIN 118
84
#define IB_MGMT_RMPP_STATUS_T2L 118
85
#define IB_MGMT_RMPP_STATUS_BAD_LEN 119
86
#define IB_MGMT_RMPP_STATUS_BAD_SEG 120
87
#define IB_MGMT_RMPP_STATUS_BADT 121
88
#define IB_MGMT_RMPP_STATUS_W2S 122
89
#define IB_MGMT_RMPP_STATUS_S2B 123
90
#define IB_MGMT_RMPP_STATUS_BAD_STATUS 124
91
#define IB_MGMT_RMPP_STATUS_UNV 125
92
#define IB_MGMT_RMPP_STATUS_TMR 126
93
#define IB_MGMT_RMPP_STATUS_UNSPEC 127
94
#define IB_MGMT_RMPP_STATUS_ABORT_MAX 127
95
96
#define IB_QP0 0
97
#define IB_QP1 cpu_to_be32(1)
98
#define IB_QP1_QKEY 0x80010000
99
#define IB_QP_SET_QKEY 0x80000000
100
101
#define IB_DEFAULT_PKEY_PARTIAL 0x7FFF
102
#define IB_DEFAULT_PKEY_FULL 0xFFFF
103
104
/*
105
* Generic trap/notice types
106
*/
107
#define IB_NOTICE_TYPE_FATAL 0x80
108
#define IB_NOTICE_TYPE_URGENT 0x81
109
#define IB_NOTICE_TYPE_SECURITY 0x82
110
#define IB_NOTICE_TYPE_SM 0x83
111
#define IB_NOTICE_TYPE_INFO 0x84
112
113
/*
114
* Generic trap/notice producers
115
*/
116
#define IB_NOTICE_PROD_CA cpu_to_be16(1)
117
#define IB_NOTICE_PROD_SWITCH cpu_to_be16(2)
118
#define IB_NOTICE_PROD_ROUTER cpu_to_be16(3)
119
#define IB_NOTICE_PROD_CLASS_MGR cpu_to_be16(4)
120
121
enum {
122
IB_MGMT_MAD_HDR = 24,
123
IB_MGMT_MAD_DATA = 232,
124
IB_MGMT_RMPP_HDR = 36,
125
IB_MGMT_RMPP_DATA = 220,
126
IB_MGMT_VENDOR_HDR = 40,
127
IB_MGMT_VENDOR_DATA = 216,
128
IB_MGMT_SA_HDR = 56,
129
IB_MGMT_SA_DATA = 200,
130
IB_MGMT_DEVICE_HDR = 64,
131
IB_MGMT_DEVICE_DATA = 192,
132
IB_MGMT_MAD_SIZE = IB_MGMT_MAD_HDR + IB_MGMT_MAD_DATA,
133
OPA_MGMT_MAD_DATA = 2024,
134
OPA_MGMT_RMPP_DATA = 2012,
135
OPA_MGMT_MAD_SIZE = IB_MGMT_MAD_HDR + OPA_MGMT_MAD_DATA,
136
};
137
138
struct ib_mad_hdr {
139
u8 base_version;
140
u8 mgmt_class;
141
u8 class_version;
142
u8 method;
143
__be16 status;
144
__be16 class_specific;
145
__be64 tid;
146
__be16 attr_id;
147
__be16 resv;
148
__be32 attr_mod;
149
};
150
151
struct ib_rmpp_hdr {
152
u8 rmpp_version;
153
u8 rmpp_type;
154
u8 rmpp_rtime_flags;
155
u8 rmpp_status;
156
__be32 seg_num;
157
__be32 paylen_newwin;
158
};
159
160
typedef u64 __bitwise ib_sa_comp_mask;
161
162
#define IB_SA_COMP_MASK(n) ((__force ib_sa_comp_mask) cpu_to_be64(1ull << (n)))
163
164
/*
165
* ib_sa_hdr and ib_sa_mad structures must be packed because they have
166
* 64-bit fields that are only 32-bit aligned. 64-bit architectures will
167
* lay them out wrong otherwise. (And unfortunately they are sent on
168
* the wire so we can't change the layout)
169
*/
170
struct ib_sa_hdr {
171
__be64 sm_key;
172
__be16 attr_offset;
173
__be16 reserved;
174
ib_sa_comp_mask comp_mask;
175
} __packed;
176
177
struct ib_mad {
178
struct ib_mad_hdr mad_hdr;
179
u8 data[IB_MGMT_MAD_DATA];
180
};
181
182
struct opa_mad {
183
struct ib_mad_hdr mad_hdr;
184
u8 data[OPA_MGMT_MAD_DATA];
185
};
186
187
struct ib_rmpp_mad {
188
struct ib_mad_hdr mad_hdr;
189
struct ib_rmpp_hdr rmpp_hdr;
190
u8 data[IB_MGMT_RMPP_DATA];
191
};
192
193
struct opa_rmpp_mad {
194
struct ib_mad_hdr mad_hdr;
195
struct ib_rmpp_hdr rmpp_hdr;
196
u8 data[OPA_MGMT_RMPP_DATA];
197
};
198
199
struct ib_sa_mad {
200
struct ib_mad_hdr mad_hdr;
201
struct ib_rmpp_hdr rmpp_hdr;
202
struct ib_sa_hdr sa_hdr;
203
u8 data[IB_MGMT_SA_DATA];
204
} __packed;
205
206
struct ib_vendor_mad {
207
struct ib_mad_hdr mad_hdr;
208
struct ib_rmpp_hdr rmpp_hdr;
209
u8 reserved;
210
u8 oui[3];
211
u8 data[IB_MGMT_VENDOR_DATA];
212
};
213
214
#define IB_MGMT_CLASSPORTINFO_ATTR_ID cpu_to_be16(0x0001)
215
216
#define IB_CLASS_PORT_INFO_RESP_TIME_MASK 0x1F
217
#define IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE 5
218
219
struct ib_class_port_info {
220
u8 base_version;
221
u8 class_version;
222
__be16 capability_mask;
223
/* 27 bits for cap_mask2, 5 bits for resp_time */
224
__be32 cap_mask2_resp_time;
225
u8 redirect_gid[16];
226
__be32 redirect_tcslfl;
227
__be16 redirect_lid;
228
__be16 redirect_pkey;
229
__be32 redirect_qp;
230
__be32 redirect_qkey;
231
u8 trap_gid[16];
232
__be32 trap_tcslfl;
233
__be16 trap_lid;
234
__be16 trap_pkey;
235
__be32 trap_hlqp;
236
__be32 trap_qkey;
237
};
238
239
/* PortInfo CapabilityMask */
240
enum ib_port_capability_mask_bits {
241
IB_PORT_SM = 1 << 1,
242
IB_PORT_NOTICE_SUP = 1 << 2,
243
IB_PORT_TRAP_SUP = 1 << 3,
244
IB_PORT_OPT_IPD_SUP = 1 << 4,
245
IB_PORT_AUTO_MIGR_SUP = 1 << 5,
246
IB_PORT_SL_MAP_SUP = 1 << 6,
247
IB_PORT_MKEY_NVRAM = 1 << 7,
248
IB_PORT_PKEY_NVRAM = 1 << 8,
249
IB_PORT_LED_INFO_SUP = 1 << 9,
250
IB_PORT_SM_DISABLED = 1 << 10,
251
IB_PORT_SYS_IMAGE_GUID_SUP = 1 << 11,
252
IB_PORT_PKEY_SW_EXT_PORT_TRAP_SUP = 1 << 12,
253
IB_PORT_EXTENDED_SPEEDS_SUP = 1 << 14,
254
IB_PORT_CAP_MASK2_SUP = 1 << 15,
255
IB_PORT_CM_SUP = 1 << 16,
256
IB_PORT_SNMP_TUNNEL_SUP = 1 << 17,
257
IB_PORT_REINIT_SUP = 1 << 18,
258
IB_PORT_DEVICE_MGMT_SUP = 1 << 19,
259
IB_PORT_VENDOR_CLASS_SUP = 1 << 20,
260
IB_PORT_DR_NOTICE_SUP = 1 << 21,
261
IB_PORT_CAP_MASK_NOTICE_SUP = 1 << 22,
262
IB_PORT_BOOT_MGMT_SUP = 1 << 23,
263
IB_PORT_LINK_LATENCY_SUP = 1 << 24,
264
IB_PORT_CLIENT_REG_SUP = 1 << 25,
265
IB_PORT_OTHER_LOCAL_CHANGES_SUP = 1 << 26,
266
IB_PORT_LINK_SPEED_WIDTH_TABLE_SUP = 1 << 27,
267
IB_PORT_VENDOR_SPECIFIC_MADS_TABLE_SUP = 1 << 28,
268
IB_PORT_MCAST_PKEY_TRAP_SUPPRESSION_SUP = 1 << 29,
269
IB_PORT_MCAST_FDB_TOP_SUP = 1 << 30,
270
IB_PORT_HIERARCHY_INFO_SUP = 1ULL << 31,
271
};
272
273
enum ib_port_capability_mask2_bits {
274
IB_PORT_SET_NODE_DESC_SUP = 1 << 0,
275
IB_PORT_EX_PORT_INFO_EX_SUP = 1 << 1,
276
IB_PORT_VIRT_SUP = 1 << 2,
277
IB_PORT_SWITCH_PORT_STATE_TABLE_SUP = 1 << 3,
278
IB_PORT_LINK_WIDTH_2X_SUP = 1 << 4,
279
IB_PORT_LINK_SPEED_HDR_SUP = 1 << 5,
280
IB_PORT_LINK_SPEED_NDR_SUP = 1 << 10,
281
IB_PORT_EXTENDED_SPEEDS2_SUP = 1 << 11,
282
IB_PORT_LINK_SPEED_XDR_SUP = 1 << 12,
283
};
284
285
#define OPA_CLASS_PORT_INFO_PR_SUPPORT BIT(26)
286
287
struct opa_class_port_info {
288
u8 base_version;
289
u8 class_version;
290
__be16 cap_mask;
291
__be32 cap_mask2_resp_time;
292
293
u8 redirect_gid[16];
294
__be32 redirect_tc_fl;
295
__be32 redirect_lid;
296
__be32 redirect_sl_qp;
297
__be32 redirect_qkey;
298
299
u8 trap_gid[16];
300
__be32 trap_tc_fl;
301
__be32 trap_lid;
302
__be32 trap_hl_qp;
303
__be32 trap_qkey;
304
305
__be16 trap_pkey;
306
__be16 redirect_pkey;
307
308
u8 trap_sl_rsvd;
309
u8 reserved[3];
310
} __packed;
311
312
/**
313
* ib_get_cpi_resp_time - Returns the resp_time value from
314
* cap_mask2_resp_time in ib_class_port_info.
315
* @cpi: A struct ib_class_port_info mad.
316
*/
317
static inline u8 ib_get_cpi_resp_time(struct ib_class_port_info *cpi)
318
{
319
return (u8)(be32_to_cpu(cpi->cap_mask2_resp_time) &
320
IB_CLASS_PORT_INFO_RESP_TIME_MASK);
321
}
322
323
/**
324
* ib_set_cpi_resptime - Sets the response time in an
325
* ib_class_port_info mad.
326
* @cpi: A struct ib_class_port_info.
327
* @rtime: The response time to set.
328
*/
329
static inline void ib_set_cpi_resp_time(struct ib_class_port_info *cpi,
330
u8 rtime)
331
{
332
cpi->cap_mask2_resp_time =
333
(cpi->cap_mask2_resp_time &
334
cpu_to_be32(~IB_CLASS_PORT_INFO_RESP_TIME_MASK)) |
335
cpu_to_be32(rtime & IB_CLASS_PORT_INFO_RESP_TIME_MASK);
336
}
337
338
/**
339
* ib_get_cpi_capmask2 - Returns the capmask2 value from
340
* cap_mask2_resp_time in ib_class_port_info.
341
* @cpi: A struct ib_class_port_info mad.
342
*/
343
static inline u32 ib_get_cpi_capmask2(struct ib_class_port_info *cpi)
344
{
345
return (be32_to_cpu(cpi->cap_mask2_resp_time) >>
346
IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE);
347
}
348
349
/**
350
* ib_set_cpi_capmask2 - Sets the capmask2 in an
351
* ib_class_port_info mad.
352
* @cpi: A struct ib_class_port_info.
353
* @capmask2: The capmask2 to set.
354
*/
355
static inline void ib_set_cpi_capmask2(struct ib_class_port_info *cpi,
356
u32 capmask2)
357
{
358
cpi->cap_mask2_resp_time =
359
(cpi->cap_mask2_resp_time &
360
cpu_to_be32(IB_CLASS_PORT_INFO_RESP_TIME_MASK)) |
361
cpu_to_be32(capmask2 <<
362
IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE);
363
}
364
365
/**
366
* opa_get_cpi_capmask2 - Returns the capmask2 value from
367
* cap_mask2_resp_time in ib_class_port_info.
368
* @cpi: A struct opa_class_port_info mad.
369
*/
370
static inline u32 opa_get_cpi_capmask2(struct opa_class_port_info *cpi)
371
{
372
return (be32_to_cpu(cpi->cap_mask2_resp_time) >>
373
IB_CLASS_PORT_INFO_RESP_TIME_FIELD_SIZE);
374
}
375
376
struct ib_mad_notice_attr {
377
u8 generic_type;
378
u8 prod_type_msb;
379
__be16 prod_type_lsb;
380
__be16 trap_num;
381
__be16 issuer_lid;
382
__be16 toggle_count;
383
384
union {
385
struct {
386
u8 details[54];
387
} raw_data;
388
389
struct {
390
__be16 reserved;
391
__be16 lid; /* where violation happened */
392
u8 port_num; /* where violation happened */
393
} __packed ntc_129_131;
394
395
struct {
396
__be16 reserved;
397
__be16 lid; /* LID where change occurred */
398
u8 reserved2;
399
u8 local_changes; /* low bit - local changes */
400
__be32 new_cap_mask; /* new capability mask */
401
u8 reserved3;
402
u8 change_flags; /* low 3 bits only */
403
} __packed ntc_144;
404
405
struct {
406
__be16 reserved;
407
__be16 lid; /* lid where sys guid changed */
408
__be16 reserved2;
409
__be64 new_sys_guid;
410
} __packed ntc_145;
411
412
struct {
413
__be16 reserved;
414
__be16 lid;
415
__be16 dr_slid;
416
u8 method;
417
u8 reserved2;
418
__be16 attr_id;
419
__be32 attr_mod;
420
__be64 mkey;
421
u8 reserved3;
422
u8 dr_trunc_hop;
423
u8 dr_rtn_path[30];
424
} __packed ntc_256;
425
426
struct {
427
__be16 reserved;
428
__be16 lid1;
429
__be16 lid2;
430
__be32 key;
431
__be32 sl_qp1; /* SL: high 4 bits */
432
__be32 qp2; /* high 8 bits reserved */
433
union ib_gid gid1;
434
union ib_gid gid2;
435
} __packed ntc_257_258;
436
437
} details;
438
};
439
440
/**
441
* ib_mad_send_buf - MAD data buffer and work request for sends.
442
* @next: A pointer used to chain together MADs for posting.
443
* @mad: References an allocated MAD data buffer for MADs that do not have
444
* RMPP active. For MADs using RMPP, references the common and management
445
* class specific headers.
446
* @mad_agent: MAD agent that allocated the buffer.
447
* @ah: The address handle to use when sending the MAD.
448
* @context: User-controlled context fields.
449
* @hdr_len: Indicates the size of the data header of the MAD. This length
450
* includes the common MAD, RMPP, and class specific headers.
451
* @data_len: Indicates the total size of user-transferred data.
452
* @seg_count: The number of RMPP segments allocated for this send.
453
* @seg_size: Size of the data in each RMPP segment. This does not include
454
* class specific headers.
455
* @seg_rmpp_size: Size of each RMPP segment including the class specific
456
* headers.
457
* @timeout_ms: Time to wait for a response.
458
* @retries: Number of times to retry a request for a response. For MADs
459
* using RMPP, this applies per window. On completion, returns the number
460
* of retries needed to complete the transfer.
461
*
462
* Users are responsible for initializing the MAD buffer itself, with the
463
* exception of any RMPP header. Additional segment buffer space allocated
464
* beyond data_len is padding.
465
*/
466
struct ib_mad_send_buf {
467
struct ib_mad_send_buf *next;
468
void *mad;
469
struct ib_mad_agent *mad_agent;
470
struct ib_ah *ah;
471
void *context[2];
472
int hdr_len;
473
int data_len;
474
int seg_count;
475
int seg_size;
476
int seg_rmpp_size;
477
int timeout_ms;
478
int retries;
479
};
480
481
/**
482
* ib_response_mad - Returns if the specified MAD has been generated in
483
* response to a sent request or trap.
484
*/
485
int ib_response_mad(const struct ib_mad_hdr *hdr);
486
487
/**
488
* ib_get_rmpp_resptime - Returns the RMPP response time.
489
* @rmpp_hdr: An RMPP header.
490
*/
491
static inline u8 ib_get_rmpp_resptime(struct ib_rmpp_hdr *rmpp_hdr)
492
{
493
return rmpp_hdr->rmpp_rtime_flags >> 3;
494
}
495
496
/**
497
* ib_get_rmpp_flags - Returns the RMPP flags.
498
* @rmpp_hdr: An RMPP header.
499
*/
500
static inline u8 ib_get_rmpp_flags(struct ib_rmpp_hdr *rmpp_hdr)
501
{
502
return rmpp_hdr->rmpp_rtime_flags & 0x7;
503
}
504
505
/**
506
* ib_set_rmpp_resptime - Sets the response time in an RMPP header.
507
* @rmpp_hdr: An RMPP header.
508
* @rtime: The response time to set.
509
*/
510
static inline void ib_set_rmpp_resptime(struct ib_rmpp_hdr *rmpp_hdr, u8 rtime)
511
{
512
rmpp_hdr->rmpp_rtime_flags = ib_get_rmpp_flags(rmpp_hdr) | (rtime << 3);
513
}
514
515
/**
516
* ib_set_rmpp_flags - Sets the flags in an RMPP header.
517
* @rmpp_hdr: An RMPP header.
518
* @flags: The flags to set.
519
*/
520
static inline void ib_set_rmpp_flags(struct ib_rmpp_hdr *rmpp_hdr, u8 flags)
521
{
522
rmpp_hdr->rmpp_rtime_flags = (rmpp_hdr->rmpp_rtime_flags & 0xF8) |
523
(flags & 0x7);
524
}
525
526
struct ib_mad_agent;
527
struct ib_mad_send_wc;
528
struct ib_mad_recv_wc;
529
530
/**
531
* ib_mad_send_handler - callback handler for a sent MAD.
532
* @mad_agent: MAD agent that sent the MAD.
533
* @mad_send_wc: Send work completion information on the sent MAD.
534
*/
535
typedef void (*ib_mad_send_handler)(struct ib_mad_agent *mad_agent,
536
struct ib_mad_send_wc *mad_send_wc);
537
538
/**
539
* ib_mad_recv_handler - callback handler for a received MAD.
540
* @mad_agent: MAD agent requesting the received MAD.
541
* @send_buf: Send buffer if found, else NULL
542
* @mad_recv_wc: Received work completion information on the received MAD.
543
*
544
* MADs received in response to a send request operation will be handed to
545
* the user before the send operation completes. All data buffers given
546
* to registered agents through this routine are owned by the receiving
547
* client.
548
*/
549
typedef void (*ib_mad_recv_handler)(struct ib_mad_agent *mad_agent,
550
struct ib_mad_send_buf *send_buf,
551
struct ib_mad_recv_wc *mad_recv_wc);
552
553
/**
554
* ib_mad_agent - Used to track MAD registration with the access layer.
555
* @device: Reference to device registration is on.
556
* @qp: Reference to QP used for sending and receiving MADs.
557
* @mr: Memory region for system memory usable for DMA.
558
* @recv_handler: Callback handler for a received MAD.
559
* @send_handler: Callback handler for a sent MAD.
560
* @context: User-specified context associated with this registration.
561
* @hi_tid: Access layer assigned transaction ID for this client.
562
* Unsolicited MADs sent by this client will have the upper 32-bits
563
* of their TID set to this value.
564
* @flags: registration flags
565
* @port_num: Port number on which QP is registered
566
* @rmpp_version: If set, indicates the RMPP version used by this agent.
567
*/
568
enum {
569
IB_MAD_USER_RMPP = IB_USER_MAD_USER_RMPP,
570
};
571
struct ib_mad_agent {
572
struct ib_device *device;
573
struct ib_qp *qp;
574
ib_mad_recv_handler recv_handler;
575
ib_mad_send_handler send_handler;
576
void *context;
577
u32 hi_tid;
578
u32 flags;
579
void *security;
580
struct list_head mad_agent_sec_list;
581
u8 port_num;
582
u8 rmpp_version;
583
bool smp_allowed;
584
};
585
586
/**
587
* ib_mad_send_wc - MAD send completion information.
588
* @send_buf: Send MAD data buffer associated with the send MAD request.
589
* @status: Completion status.
590
* @vendor_err: Optional vendor error information returned with a failed
591
* request.
592
*/
593
struct ib_mad_send_wc {
594
struct ib_mad_send_buf *send_buf;
595
enum ib_wc_status status;
596
u32 vendor_err;
597
};
598
599
/**
600
* ib_mad_recv_buf - received MAD buffer information.
601
* @list: Reference to next data buffer for a received RMPP MAD.
602
* @grh: References a data buffer containing the global route header.
603
* The data refereced by this buffer is only valid if the GRH is
604
* valid.
605
* @mad: References the start of the received MAD.
606
*/
607
struct ib_mad_recv_buf {
608
struct list_head list;
609
struct ib_grh *grh;
610
union {
611
struct ib_mad *mad;
612
struct opa_mad *opa_mad;
613
};
614
};
615
616
/**
617
* ib_mad_recv_wc - received MAD information.
618
* @wc: Completion information for the received data.
619
* @recv_buf: Specifies the location of the received data buffer(s).
620
* @rmpp_list: Specifies a list of RMPP reassembled received MAD buffers.
621
* @mad_len: The length of the received MAD, without duplicated headers.
622
* @mad_seg_size: The size of individual MAD segments
623
*
624
* For received response, the wr_id contains a pointer to the ib_mad_send_buf
625
* for the corresponding send request.
626
*/
627
struct ib_mad_recv_wc {
628
struct ib_wc *wc;
629
struct ib_mad_recv_buf recv_buf;
630
struct list_head rmpp_list;
631
int mad_len;
632
size_t mad_seg_size;
633
};
634
635
/**
636
* ib_mad_reg_req - MAD registration request
637
* @mgmt_class: Indicates which management class of MADs should be receive
638
* by the caller. This field is only required if the user wishes to
639
* receive unsolicited MADs, otherwise it should be 0.
640
* @mgmt_class_version: Indicates which version of MADs for the given
641
* management class to receive.
642
* @oui: Indicates IEEE OUI when mgmt_class is a vendor class
643
* in the range from 0x30 to 0x4f. Otherwise not used.
644
* @method_mask: The caller will receive unsolicited MADs for any method
645
* where @method_mask = 1.
646
*
647
*/
648
struct ib_mad_reg_req {
649
u8 mgmt_class;
650
u8 mgmt_class_version;
651
u8 oui[3];
652
DECLARE_BITMAP(method_mask, IB_MGMT_MAX_METHODS);
653
};
654
655
/**
656
* ib_register_mad_agent - Register to send/receive MADs.
657
* @device: The device to register with.
658
* @port_num: The port on the specified device to use.
659
* @qp_type: Specifies which QP to access. Must be either
660
* IB_QPT_SMI or IB_QPT_GSI.
661
* @mad_reg_req: Specifies which unsolicited MADs should be received
662
* by the caller. This parameter may be NULL if the caller only
663
* wishes to receive solicited responses.
664
* @rmpp_version: If set, indicates that the client will send
665
* and receive MADs that contain the RMPP header for the given version.
666
* If set to 0, indicates that RMPP is not used by this client.
667
* @send_handler: The completion callback routine invoked after a send
668
* request has completed.
669
* @recv_handler: The completion callback routine invoked for a received
670
* MAD.
671
* @context: User specified context associated with the registration.
672
* @registration_flags: Registration flags to set for this agent
673
*/
674
struct ib_mad_agent *ib_register_mad_agent(struct ib_device *device,
675
u32 port_num,
676
enum ib_qp_type qp_type,
677
struct ib_mad_reg_req *mad_reg_req,
678
u8 rmpp_version,
679
ib_mad_send_handler send_handler,
680
ib_mad_recv_handler recv_handler,
681
void *context,
682
u32 registration_flags);
683
/**
684
* ib_unregister_mad_agent - Unregisters a client from using MAD services.
685
* @mad_agent: Corresponding MAD registration request to deregister.
686
*
687
* After invoking this routine, MAD services are no longer usable by the
688
* client on the associated QP.
689
*/
690
void ib_unregister_mad_agent(struct ib_mad_agent *mad_agent);
691
692
/**
693
* ib_post_send_mad - Posts MAD(s) to the send queue of the QP associated
694
* with the registered client.
695
* @send_buf: Specifies the information needed to send the MAD(s).
696
* @bad_send_buf: Specifies the MAD on which an error was encountered. This
697
* parameter is optional if only a single MAD is posted.
698
*
699
* Sent MADs are not guaranteed to complete in the order that they were posted.
700
*
701
* If the MAD requires RMPP, the data buffer should contain a single copy
702
* of the common MAD, RMPP, and class specific headers, followed by the class
703
* defined data. If the class defined data would not divide evenly into
704
* RMPP segments, then space must be allocated at the end of the referenced
705
* buffer for any required padding. To indicate the amount of class defined
706
* data being transferred, the paylen_newwin field in the RMPP header should
707
* be set to the size of the class specific header plus the amount of class
708
* defined data being transferred. The paylen_newwin field should be
709
* specified in network-byte order.
710
*/
711
int ib_post_send_mad(struct ib_mad_send_buf *send_buf,
712
struct ib_mad_send_buf **bad_send_buf);
713
714
715
/**
716
* ib_free_recv_mad - Returns data buffers used to receive a MAD.
717
* @mad_recv_wc: Work completion information for a received MAD.
718
*
719
* Clients receiving MADs through their ib_mad_recv_handler must call this
720
* routine to return the work completion buffers to the access layer.
721
*/
722
void ib_free_recv_mad(struct ib_mad_recv_wc *mad_recv_wc);
723
724
/**
725
* ib_modify_mad - Modifies an outstanding send MAD operation.
726
* @send_buf: Indicates the MAD to modify.
727
* @timeout_ms: New timeout value for sent MAD.
728
*
729
* This call will reset the timeout value for a sent MAD to the specified
730
* value.
731
*/
732
int ib_modify_mad(struct ib_mad_send_buf *send_buf, u32 timeout_ms);
733
734
/**
735
* ib_cancel_mad - Cancels an outstanding send MAD operation.
736
* @send_buf: Indicates the MAD to cancel.
737
*
738
* MADs will be returned to the user through the corresponding
739
* ib_mad_send_handler.
740
*/
741
static inline void ib_cancel_mad(struct ib_mad_send_buf *send_buf)
742
{
743
ib_modify_mad(send_buf, 0);
744
}
745
746
/**
747
* ib_create_send_mad - Allocate and initialize a data buffer and work request
748
* for sending a MAD.
749
* @mad_agent: Specifies the registered MAD service to associate with the MAD.
750
* @remote_qpn: Specifies the QPN of the receiving node.
751
* @pkey_index: Specifies which PKey the MAD will be sent using. This field
752
* is valid only if the remote_qpn is QP 1.
753
* @rmpp_active: Indicates if the send will enable RMPP.
754
* @hdr_len: Indicates the size of the data header of the MAD. This length
755
* should include the common MAD header, RMPP header, plus any class
756
* specific header.
757
* @data_len: Indicates the size of any user-transferred data. The call will
758
* automatically adjust the allocated buffer size to account for any
759
* additional padding that may be necessary.
760
* @gfp_mask: GFP mask used for the memory allocation.
761
* @base_version: Base Version of this MAD
762
*
763
* This routine allocates a MAD for sending. The returned MAD send buffer
764
* will reference a data buffer usable for sending a MAD, along
765
* with an initialized work request structure. Users may modify the returned
766
* MAD data buffer before posting the send.
767
*
768
* The returned MAD header, class specific headers, and any padding will be
769
* cleared. Users are responsible for initializing the common MAD header,
770
* any class specific header, and MAD data area.
771
* If @rmpp_active is set, the RMPP header will be initialized for sending.
772
*/
773
struct ib_mad_send_buf *ib_create_send_mad(struct ib_mad_agent *mad_agent,
774
u32 remote_qpn, u16 pkey_index,
775
int rmpp_active,
776
int hdr_len, int data_len,
777
gfp_t gfp_mask,
778
u8 base_version);
779
780
/**
781
* ib_is_mad_class_rmpp - returns whether given management class
782
* supports RMPP.
783
* @mgmt_class: management class
784
*
785
* This routine returns whether the management class supports RMPP.
786
*/
787
int ib_is_mad_class_rmpp(u8 mgmt_class);
788
789
/**
790
* ib_get_mad_data_offset - returns the data offset for a given
791
* management class.
792
* @mgmt_class: management class
793
*
794
* This routine returns the data offset in the MAD for the management
795
* class requested.
796
*/
797
int ib_get_mad_data_offset(u8 mgmt_class);
798
799
/**
800
* ib_get_rmpp_segment - returns the data buffer for a given RMPP segment.
801
* @send_buf: Previously allocated send data buffer.
802
* @seg_num: number of segment to return
803
*
804
* This routine returns a pointer to the data buffer of an RMPP MAD.
805
* Users must provide synchronization to @send_buf around this call.
806
*/
807
void *ib_get_rmpp_segment(struct ib_mad_send_buf *send_buf, int seg_num);
808
809
/**
810
* ib_free_send_mad - Returns data buffers used to send a MAD.
811
* @send_buf: Previously allocated send data buffer.
812
*/
813
void ib_free_send_mad(struct ib_mad_send_buf *send_buf);
814
815
/**
816
* ib_mad_kernel_rmpp_agent - Returns if the agent is performing RMPP.
817
* @agent: the agent in question
818
* @return: true if agent is performing rmpp, false otherwise.
819
*/
820
int ib_mad_kernel_rmpp_agent(const struct ib_mad_agent *agent);
821
822
#endif /* IB_MAD_H */
823
824