2 * Copyright (c) 2004 Intel Corporation. All rights reserved.
3 * Copyright (c) 2004 Topspin Corporation. All rights reserved.
4 * Copyright (c) 2004 Voltaire Corporation. All rights reserved.
6 * This software is available to you under a choice of one of two
7 * licenses. You may choose to be licensed under the terms of the GNU
8 * General Public License (GPL) Version 2, available from the file
9 * COPYING the madirectory of this source tree, or the
10 * OpenIB.org BSD license below:
12 * Redistribution and use source and binary forms, with or
13 * withmodification, are permitted provided that the following
16 * - Redistributions of source code must retathe above
17 * copyright notice, this list of conditions and the following
20 * - Redistributions binary form must reproduce the above
21 * copyright notice, this list of conditions and the following
22 * disclaimer the documentation and/or other materials
23 * provided with the distribution.
25 * THE SOFTWARE IS PROVIDED "AS IS", WITHWARRANTY OF ANY KIND,
26 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
27 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
28 * NONINFRINGEMENT. NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
29 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER AN
30 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OF OR IN
31 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS THE
34 #if !defined(CM_MSGS_H)
37 #include <rdma/ib_mad.h>
40 * Parameters to routines below should be in network-byte order, and values
41 * are returned in network-byte order.
44 #define IB_CM_CLASS_VERSION 2 /* IB specification 1.2 */
46 #define CM_REQ_ATTR_ID __constant_htons(0x0010)
47 #define CM_MRA_ATTR_ID __constant_htons(0x0011)
48 #define CM_REJ_ATTR_ID __constant_htons(0x0012)
49 #define CM_REP_ATTR_ID __constant_htons(0x0013)
50 #define CM_RTU_ATTR_ID __constant_htons(0x0014)
51 #define CM_DREQ_ATTR_ID __constant_htons(0x0015)
52 #define CM_DREP_ATTR_ID __constant_htons(0x0016)
53 #define CM_SIDR_REQ_ATTR_ID __constant_htons(0x0017)
54 #define CM_SIDR_REP_ATTR_ID __constant_htons(0x0018)
55 #define CM_LAP_ATTR_ID __constant_htons(0x0019)
56 #define CM_APR_ATTR_ID __constant_htons(0x001A)
58 enum cm_msg_sequence {
66 struct ib_mad_hdr hdr;
74 /* local QPN:24, responder resources:8 */
76 /* local EECN:24, initiator depth:8 */
79 * remote EECN:24, remote CM response timeout:5,
80 * transport service type:2, end-to-end flow control:1
83 /* starting PSN:24, local CM response timeout:5, retry count:3 */
86 /* path MTU:4, RDC exists:1, RNR retry count:3. */
88 /* max CM Retries:4, SRQ:1, rsvd:3 */
91 __be16 primary_local_lid;
92 __be16 primary_remote_lid;
93 union ib_gid primary_local_gid;
94 union ib_gid primary_remote_gid;
95 /* flow label:20, rsvd:6, packet rate:6 */
96 __be32 primary_offset88;
97 u8 primary_traffic_class;
99 /* SL:4, subnet local:1, rsvd:3 */
101 /* local ACK timeout:5, rsvd:3 */
104 __be16 alt_local_lid;
105 __be16 alt_remote_lid;
106 union ib_gid alt_local_gid;
107 union ib_gid alt_remote_gid;
108 /* flow label:20, rsvd:6, packet rate:6 */
109 __be32 alt_offset132;
110 u8 alt_traffic_class;
112 /* SL:4, subnet local:1, rsvd:3 */
114 /* local ACK timeout:5, rsvd:3 */
117 u8 private_data[IB_CM_REQ_PRIVATE_DATA_SIZE];
119 } __attribute__ ((packed));
121 static inline __be32 cm_req_get_local_qpn(struct cm_req_msg *req_msg)
123 return cpu_to_be32(be32_to_cpu(req_msg->offset32) >> 8);
126 static inline void cm_req_set_local_qpn(struct cm_req_msg *req_msg, __be32 qpn)
128 req_msg->offset32 = cpu_to_be32((be32_to_cpu(qpn) << 8) |
129 (be32_to_cpu(req_msg->offset32) &
133 static inline u8 cm_req_get_resp_res(struct cm_req_msg *req_msg)
135 return (u8) be32_to_cpu(req_msg->offset32);
138 static inline void cm_req_set_resp_res(struct cm_req_msg *req_msg, u8 resp_res)
140 req_msg->offset32 = cpu_to_be32(resp_res |
141 (be32_to_cpu(req_msg->offset32) &
145 static inline u8 cm_req_get_init_depth(struct cm_req_msg *req_msg)
147 return (u8) be32_to_cpu(req_msg->offset36);
150 static inline void cm_req_set_init_depth(struct cm_req_msg *req_msg,
153 req_msg->offset36 = cpu_to_be32(init_depth |
154 (be32_to_cpu(req_msg->offset36) &
158 static inline u8 cm_req_get_remote_resp_timeout(struct cm_req_msg *req_msg)
160 return (u8) ((be32_to_cpu(req_msg->offset40) & 0xF8) >> 3);
163 static inline void cm_req_set_remote_resp_timeout(struct cm_req_msg *req_msg,
166 req_msg->offset40 = cpu_to_be32((resp_timeout << 3) |
167 (be32_to_cpu(req_msg->offset40) &
171 static inline enum ib_qp_type cm_req_get_qp_type(struct cm_req_msg *req_msg)
173 u8 transport_type = (u8) (be32_to_cpu(req_msg->offset40) & 0x06) >> 1;
174 switch(transport_type) {
175 case 0: return IB_QPT_RC;
176 case 1: return IB_QPT_UC;
181 static inline void cm_req_set_qp_type(struct cm_req_msg *req_msg,
182 enum ib_qp_type qp_type)
186 req_msg->offset40 = cpu_to_be32((be32_to_cpu(
191 req_msg->offset40 = cpu_to_be32(be32_to_cpu(
197 static inline u8 cm_req_get_flow_ctrl(struct cm_req_msg *req_msg)
199 return be32_to_cpu(req_msg->offset40) & 0x1;
202 static inline void cm_req_set_flow_ctrl(struct cm_req_msg *req_msg,
205 req_msg->offset40 = cpu_to_be32((flow_ctrl & 0x1) |
206 (be32_to_cpu(req_msg->offset40) &
210 static inline __be32 cm_req_get_starting_psn(struct cm_req_msg *req_msg)
212 return cpu_to_be32(be32_to_cpu(req_msg->offset44) >> 8);
215 static inline void cm_req_set_starting_psn(struct cm_req_msg *req_msg,
218 req_msg->offset44 = cpu_to_be32((be32_to_cpu(starting_psn) << 8) |
219 (be32_to_cpu(req_msg->offset44) & 0x000000FF));
222 static inline u8 cm_req_get_local_resp_timeout(struct cm_req_msg *req_msg)
224 return (u8) ((be32_to_cpu(req_msg->offset44) & 0xF8) >> 3);
227 static inline void cm_req_set_local_resp_timeout(struct cm_req_msg *req_msg,
230 req_msg->offset44 = cpu_to_be32((resp_timeout << 3) |
231 (be32_to_cpu(req_msg->offset44) & 0xFFFFFF07));
234 static inline u8 cm_req_get_retry_count(struct cm_req_msg *req_msg)
236 return (u8) (be32_to_cpu(req_msg->offset44) & 0x7);
239 static inline void cm_req_set_retry_count(struct cm_req_msg *req_msg,
242 req_msg->offset44 = cpu_to_be32((retry_count & 0x7) |
243 (be32_to_cpu(req_msg->offset44) & 0xFFFFFFF8));
246 static inline u8 cm_req_get_path_mtu(struct cm_req_msg *req_msg)
248 return req_msg->offset50 >> 4;
251 static inline void cm_req_set_path_mtu(struct cm_req_msg *req_msg, u8 path_mtu)
253 req_msg->offset50 = (u8) ((req_msg->offset50 & 0xF) | (path_mtu << 4));
256 static inline u8 cm_req_get_rnr_retry_count(struct cm_req_msg *req_msg)
258 return req_msg->offset50 & 0x7;
261 static inline void cm_req_set_rnr_retry_count(struct cm_req_msg *req_msg,
264 req_msg->offset50 = (u8) ((req_msg->offset50 & 0xF8) |
265 (rnr_retry_count & 0x7));
268 static inline u8 cm_req_get_max_cm_retries(struct cm_req_msg *req_msg)
270 return req_msg->offset51 >> 4;
273 static inline void cm_req_set_max_cm_retries(struct cm_req_msg *req_msg,
276 req_msg->offset51 = (u8) ((req_msg->offset51 & 0xF) | (retries << 4));
279 static inline u8 cm_req_get_srq(struct cm_req_msg *req_msg)
281 return (req_msg->offset51 & 0x8) >> 3;
284 static inline void cm_req_set_srq(struct cm_req_msg *req_msg, u8 srq)
286 req_msg->offset51 = (u8) ((req_msg->offset51 & 0xF7) |
290 static inline __be32 cm_req_get_primary_flow_label(struct cm_req_msg *req_msg)
292 return cpu_to_be32(be32_to_cpu(req_msg->primary_offset88) >> 12);
295 static inline void cm_req_set_primary_flow_label(struct cm_req_msg *req_msg,
298 req_msg->primary_offset88 = cpu_to_be32(
299 (be32_to_cpu(req_msg->primary_offset88) &
301 (be32_to_cpu(flow_label) << 12));
304 static inline u8 cm_req_get_primary_packet_rate(struct cm_req_msg *req_msg)
306 return (u8) (be32_to_cpu(req_msg->primary_offset88) & 0x3F);
309 static inline void cm_req_set_primary_packet_rate(struct cm_req_msg *req_msg,
312 req_msg->primary_offset88 = cpu_to_be32(
313 (be32_to_cpu(req_msg->primary_offset88) &
314 0xFFFFFFC0) | (rate & 0x3F));
317 static inline u8 cm_req_get_primary_sl(struct cm_req_msg *req_msg)
319 return (u8) (req_msg->primary_offset94 >> 4);
322 static inline void cm_req_set_primary_sl(struct cm_req_msg *req_msg, u8 sl)
324 req_msg->primary_offset94 = (u8) ((req_msg->primary_offset94 & 0x0F) |
328 static inline u8 cm_req_get_primary_subnet_local(struct cm_req_msg *req_msg)
330 return (u8) ((req_msg->primary_offset94 & 0x08) >> 3);
333 static inline void cm_req_set_primary_subnet_local(struct cm_req_msg *req_msg,
336 req_msg->primary_offset94 = (u8) ((req_msg->primary_offset94 & 0xF7) |
337 ((subnet_local & 0x1) << 3));
340 static inline u8 cm_req_get_primary_local_ack_timeout(struct cm_req_msg *req_msg)
342 return (u8) (req_msg->primary_offset95 >> 3);
345 static inline void cm_req_set_primary_local_ack_timeout(struct cm_req_msg *req_msg,
346 u8 local_ack_timeout)
348 req_msg->primary_offset95 = (u8) ((req_msg->primary_offset95 & 0x07) |
349 (local_ack_timeout << 3));
352 static inline __be32 cm_req_get_alt_flow_label(struct cm_req_msg *req_msg)
354 return cpu_to_be32(be32_to_cpu(req_msg->alt_offset132) >> 12);
357 static inline void cm_req_set_alt_flow_label(struct cm_req_msg *req_msg,
360 req_msg->alt_offset132 = cpu_to_be32(
361 (be32_to_cpu(req_msg->alt_offset132) &
363 (be32_to_cpu(flow_label) << 12));
366 static inline u8 cm_req_get_alt_packet_rate(struct cm_req_msg *req_msg)
368 return (u8) (be32_to_cpu(req_msg->alt_offset132) & 0x3F);
371 static inline void cm_req_set_alt_packet_rate(struct cm_req_msg *req_msg,
374 req_msg->alt_offset132 = cpu_to_be32(
375 (be32_to_cpu(req_msg->alt_offset132) &
376 0xFFFFFFC0) | (rate & 0x3F));
379 static inline u8 cm_req_get_alt_sl(struct cm_req_msg *req_msg)
381 return (u8) (req_msg->alt_offset138 >> 4);
384 static inline void cm_req_set_alt_sl(struct cm_req_msg *req_msg, u8 sl)
386 req_msg->alt_offset138 = (u8) ((req_msg->alt_offset138 & 0x0F) |
390 static inline u8 cm_req_get_alt_subnet_local(struct cm_req_msg *req_msg)
392 return (u8) ((req_msg->alt_offset138 & 0x08) >> 3);
395 static inline void cm_req_set_alt_subnet_local(struct cm_req_msg *req_msg,
398 req_msg->alt_offset138 = (u8) ((req_msg->alt_offset138 & 0xF7) |
399 ((subnet_local & 0x1) << 3));
402 static inline u8 cm_req_get_alt_local_ack_timeout(struct cm_req_msg *req_msg)
404 return (u8) (req_msg->alt_offset139 >> 3);
407 static inline void cm_req_set_alt_local_ack_timeout(struct cm_req_msg *req_msg,
408 u8 local_ack_timeout)
410 req_msg->alt_offset139 = (u8) ((req_msg->alt_offset139 & 0x07) |
411 (local_ack_timeout << 3));
414 /* Message REJected or MRAed */
415 enum cm_msg_response {
416 CM_MSG_RESPONSE_REQ = 0x0,
417 CM_MSG_RESPONSE_REP = 0x1,
418 CM_MSG_RESPONSE_OTHER = 0x2
422 struct ib_mad_hdr hdr;
424 __be32 local_comm_id;
425 __be32 remote_comm_id;
426 /* message MRAed:2, rsvd:6 */
428 /* service timeout:5, rsvd:3 */
431 u8 private_data[IB_CM_MRA_PRIVATE_DATA_SIZE];
433 } __attribute__ ((packed));
435 static inline u8 cm_mra_get_msg_mraed(struct cm_mra_msg *mra_msg)
437 return (u8) (mra_msg->offset8 >> 6);
440 static inline void cm_mra_set_msg_mraed(struct cm_mra_msg *mra_msg, u8 msg)
442 mra_msg->offset8 = (u8) ((mra_msg->offset8 & 0x3F) | (msg << 6));
445 static inline u8 cm_mra_get_service_timeout(struct cm_mra_msg *mra_msg)
447 return (u8) (mra_msg->offset9 >> 3);
450 static inline void cm_mra_set_service_timeout(struct cm_mra_msg *mra_msg,
453 mra_msg->offset9 = (u8) ((mra_msg->offset9 & 0x07) |
454 (service_timeout << 3));
458 struct ib_mad_hdr hdr;
460 __be32 local_comm_id;
461 __be32 remote_comm_id;
462 /* message REJected:2, rsvd:6 */
464 /* reject info length:7, rsvd:1. */
467 u8 ari[IB_CM_REJ_ARI_LENGTH];
469 u8 private_data[IB_CM_REJ_PRIVATE_DATA_SIZE];
471 } __attribute__ ((packed));
473 static inline u8 cm_rej_get_msg_rejected(struct cm_rej_msg *rej_msg)
475 return (u8) (rej_msg->offset8 >> 6);
478 static inline void cm_rej_set_msg_rejected(struct cm_rej_msg *rej_msg, u8 msg)
480 rej_msg->offset8 = (u8) ((rej_msg->offset8 & 0x3F) | (msg << 6));
483 static inline u8 cm_rej_get_reject_info_len(struct cm_rej_msg *rej_msg)
485 return (u8) (rej_msg->offset9 >> 1);
488 static inline void cm_rej_set_reject_info_len(struct cm_rej_msg *rej_msg,
491 rej_msg->offset9 = (u8) ((rej_msg->offset9 & 0x1) | (len << 1));
495 struct ib_mad_hdr hdr;
497 __be32 local_comm_id;
498 __be32 remote_comm_id;
500 /* local QPN:24, rsvd:8 */
502 /* local EECN:24, rsvd:8 */
504 /* starting PSN:24 rsvd:8 */
508 /* target ACK delay:5, failover accepted:2, end-to-end flow control:1 */
510 /* RNR retry count:3, SRQ:1, rsvd:5 */
512 __be64 local_ca_guid;
514 u8 private_data[IB_CM_REP_PRIVATE_DATA_SIZE];
516 } __attribute__ ((packed));
518 static inline __be32 cm_rep_get_local_qpn(struct cm_rep_msg *rep_msg)
520 return cpu_to_be32(be32_to_cpu(rep_msg->offset12) >> 8);
523 static inline void cm_rep_set_local_qpn(struct cm_rep_msg *rep_msg, __be32 qpn)
525 rep_msg->offset12 = cpu_to_be32((be32_to_cpu(qpn) << 8) |
526 (be32_to_cpu(rep_msg->offset12) & 0x000000FF));
529 static inline __be32 cm_rep_get_starting_psn(struct cm_rep_msg *rep_msg)
531 return cpu_to_be32(be32_to_cpu(rep_msg->offset20) >> 8);
534 static inline void cm_rep_set_starting_psn(struct cm_rep_msg *rep_msg,
537 rep_msg->offset20 = cpu_to_be32((be32_to_cpu(starting_psn) << 8) |
538 (be32_to_cpu(rep_msg->offset20) & 0x000000FF));
541 static inline u8 cm_rep_get_target_ack_delay(struct cm_rep_msg *rep_msg)
543 return (u8) (rep_msg->offset26 >> 3);
546 static inline void cm_rep_set_target_ack_delay(struct cm_rep_msg *rep_msg,
549 rep_msg->offset26 = (u8) ((rep_msg->offset26 & 0x07) |
550 (target_ack_delay << 3));
553 static inline u8 cm_rep_get_failover(struct cm_rep_msg *rep_msg)
555 return (u8) ((rep_msg->offset26 & 0x06) >> 1);
558 static inline void cm_rep_set_failover(struct cm_rep_msg *rep_msg, u8 failover)
560 rep_msg->offset26 = (u8) ((rep_msg->offset26 & 0xF9) |
561 ((failover & 0x3) << 1));
564 static inline u8 cm_rep_get_flow_ctrl(struct cm_rep_msg *rep_msg)
566 return (u8) (rep_msg->offset26 & 0x01);
569 static inline void cm_rep_set_flow_ctrl(struct cm_rep_msg *rep_msg,
572 rep_msg->offset26 = (u8) ((rep_msg->offset26 & 0xFE) |
576 static inline u8 cm_rep_get_rnr_retry_count(struct cm_rep_msg *rep_msg)
578 return (u8) (rep_msg->offset27 >> 5);
581 static inline void cm_rep_set_rnr_retry_count(struct cm_rep_msg *rep_msg,
584 rep_msg->offset27 = (u8) ((rep_msg->offset27 & 0x1F) |
585 (rnr_retry_count << 5));
588 static inline u8 cm_rep_get_srq(struct cm_rep_msg *rep_msg)
590 return (u8) ((rep_msg->offset27 >> 4) & 0x1);
593 static inline void cm_rep_set_srq(struct cm_rep_msg *rep_msg, u8 srq)
595 rep_msg->offset27 = (u8) ((rep_msg->offset27 & 0xEF) |
600 struct ib_mad_hdr hdr;
602 __be32 local_comm_id;
603 __be32 remote_comm_id;
605 u8 private_data[IB_CM_RTU_PRIVATE_DATA_SIZE];
607 } __attribute__ ((packed));
610 struct ib_mad_hdr hdr;
612 __be32 local_comm_id;
613 __be32 remote_comm_id;
614 /* remote QPN/EECN:24, rsvd:8 */
617 u8 private_data[IB_CM_DREQ_PRIVATE_DATA_SIZE];
619 } __attribute__ ((packed));
621 static inline __be32 cm_dreq_get_remote_qpn(struct cm_dreq_msg *dreq_msg)
623 return cpu_to_be32(be32_to_cpu(dreq_msg->offset8) >> 8);
626 static inline void cm_dreq_set_remote_qpn(struct cm_dreq_msg *dreq_msg, __be32 qpn)
628 dreq_msg->offset8 = cpu_to_be32((be32_to_cpu(qpn) << 8) |
629 (be32_to_cpu(dreq_msg->offset8) & 0x000000FF));
633 struct ib_mad_hdr hdr;
635 __be32 local_comm_id;
636 __be32 remote_comm_id;
638 u8 private_data[IB_CM_DREP_PRIVATE_DATA_SIZE];
640 } __attribute__ ((packed));
643 struct ib_mad_hdr hdr;
645 __be32 local_comm_id;
646 __be32 remote_comm_id;
649 /* remote QPN/EECN:24, remote CM response timeout:5, rsvd:3 */
653 __be16 alt_local_lid;
654 __be16 alt_remote_lid;
655 union ib_gid alt_local_gid;
656 union ib_gid alt_remote_gid;
657 /* flow label:20, rsvd:4, traffic class:8 */
660 /* rsvd:2, packet rate:6 */
662 /* SL:4, subnet local:1, rsvd:3 */
664 /* local ACK timeout:5, rsvd:3 */
667 u8 private_data[IB_CM_LAP_PRIVATE_DATA_SIZE];
668 } __attribute__ ((packed));
670 static inline __be32 cm_lap_get_remote_qpn(struct cm_lap_msg *lap_msg)
672 return cpu_to_be32(be32_to_cpu(lap_msg->offset12) >> 8);
675 static inline void cm_lap_set_remote_qpn(struct cm_lap_msg *lap_msg, __be32 qpn)
677 lap_msg->offset12 = cpu_to_be32((be32_to_cpu(qpn) << 8) |
678 (be32_to_cpu(lap_msg->offset12) &
682 static inline u8 cm_lap_get_remote_resp_timeout(struct cm_lap_msg *lap_msg)
684 return (u8) ((be32_to_cpu(lap_msg->offset12) & 0xF8) >> 3);
687 static inline void cm_lap_set_remote_resp_timeout(struct cm_lap_msg *lap_msg,
690 lap_msg->offset12 = cpu_to_be32((resp_timeout << 3) |
691 (be32_to_cpu(lap_msg->offset12) &
695 static inline __be32 cm_lap_get_flow_label(struct cm_lap_msg *lap_msg)
697 return cpu_to_be32(be32_to_cpu(lap_msg->offset56) >> 12);
700 static inline void cm_lap_set_flow_label(struct cm_lap_msg *lap_msg,
703 lap_msg->offset56 = cpu_to_be32(
704 (be32_to_cpu(lap_msg->offset56) & 0x00000FFF) |
705 (be32_to_cpu(flow_label) << 12));
708 static inline u8 cm_lap_get_traffic_class(struct cm_lap_msg *lap_msg)
710 return (u8) be32_to_cpu(lap_msg->offset56);
713 static inline void cm_lap_set_traffic_class(struct cm_lap_msg *lap_msg,
716 lap_msg->offset56 = cpu_to_be32(traffic_class |
717 (be32_to_cpu(lap_msg->offset56) &
721 static inline u8 cm_lap_get_packet_rate(struct cm_lap_msg *lap_msg)
723 return lap_msg->offset61 & 0x3F;
726 static inline void cm_lap_set_packet_rate(struct cm_lap_msg *lap_msg,
729 lap_msg->offset61 = (packet_rate & 0x3F) | (lap_msg->offset61 & 0xC0);
732 static inline u8 cm_lap_get_sl(struct cm_lap_msg *lap_msg)
734 return lap_msg->offset62 >> 4;
737 static inline void cm_lap_set_sl(struct cm_lap_msg *lap_msg, u8 sl)
739 lap_msg->offset62 = (sl << 4) | (lap_msg->offset62 & 0x0F);
742 static inline u8 cm_lap_get_subnet_local(struct cm_lap_msg *lap_msg)
744 return (lap_msg->offset62 >> 3) & 0x1;
747 static inline void cm_lap_set_subnet_local(struct cm_lap_msg *lap_msg,
750 lap_msg->offset62 = ((subnet_local & 0x1) << 3) |
751 (lap_msg->offset61 & 0xF7);
753 static inline u8 cm_lap_get_local_ack_timeout(struct cm_lap_msg *lap_msg)
755 return lap_msg->offset63 >> 3;
758 static inline void cm_lap_set_local_ack_timeout(struct cm_lap_msg *lap_msg,
759 u8 local_ack_timeout)
761 lap_msg->offset63 = (local_ack_timeout << 3) |
762 (lap_msg->offset63 & 0x07);
766 struct ib_mad_hdr hdr;
768 __be32 local_comm_id;
769 __be32 remote_comm_id;
773 u8 info[IB_CM_APR_INFO_LENGTH];
775 u8 private_data[IB_CM_APR_PRIVATE_DATA_SIZE];
776 } __attribute__ ((packed));
778 struct cm_sidr_req_msg {
779 struct ib_mad_hdr hdr;
786 u8 private_data[IB_CM_SIDR_REQ_PRIVATE_DATA_SIZE];
787 } __attribute__ ((packed));
789 struct cm_sidr_rep_msg {
790 struct ib_mad_hdr hdr;
800 u8 info[IB_CM_SIDR_REP_INFO_LENGTH];
802 u8 private_data[IB_CM_SIDR_REP_PRIVATE_DATA_SIZE];
803 } __attribute__ ((packed));
805 static inline __be32 cm_sidr_rep_get_qpn(struct cm_sidr_rep_msg *sidr_rep_msg)
807 return cpu_to_be32(be32_to_cpu(sidr_rep_msg->offset8) >> 8);
810 static inline void cm_sidr_rep_set_qpn(struct cm_sidr_rep_msg *sidr_rep_msg,
813 sidr_rep_msg->offset8 = cpu_to_be32((be32_to_cpu(qpn) << 8) |
814 (be32_to_cpu(sidr_rep_msg->offset8) &
818 #endif /* CM_MSGS_H */