1/* SPDX-License-Identifier: GPL-2.0 or BSD-3-Clause */ 2 3/* Authors: Bernard Metzler <bmt@zurich.ibm.com> */ 4/* Copyright (c) 2008-2019, IBM Corporation */ 5 6#ifndef _IWARP_H 7#define _IWARP_H 8 9#include <rdma/rdma_user_cm.h> /* RDMA_MAX_PRIVATE_DATA */ 10#include <linux/types.h> 11#include <asm/byteorder.h> 12 13#define RDMAP_VERSION 1 14#define DDP_VERSION 1 15#define MPA_REVISION_1 1 16#define MPA_REVISION_2 2 17#define MPA_MAX_PRIVDATA RDMA_MAX_PRIVATE_DATA 18#define MPA_KEY_REQ "MPA ID Req Frame" 19#define MPA_KEY_REP "MPA ID Rep Frame" 20#define MPA_IRD_ORD_MASK 0x3fff 21 22struct mpa_rr_params { 23 __be16 bits; 24 __be16 pd_len; 25}; 26 27/* 28 * MPA request/response header bits & fields 29 */ 30enum { 31 MPA_RR_FLAG_MARKERS = cpu_to_be16(0x8000), 32 MPA_RR_FLAG_CRC = cpu_to_be16(0x4000), 33 MPA_RR_FLAG_REJECT = cpu_to_be16(0x2000), 34 MPA_RR_FLAG_ENHANCED = cpu_to_be16(0x1000), 35 MPA_RR_FLAG_GSO_EXP = cpu_to_be16(0x0800), 36 MPA_RR_MASK_REVISION = cpu_to_be16(0x00ff) 37}; 38 39/* 40 * MPA request/reply header 41 */ 42struct mpa_rr { 43 __u8 key[16]; 44 struct mpa_rr_params params; 45}; 46 47static inline void __mpa_rr_set_revision(__be16 *bits, u8 rev) 48{ 49 *bits = (*bits & ~MPA_RR_MASK_REVISION) | 50 (cpu_to_be16(rev) & MPA_RR_MASK_REVISION); 51} 52 53static inline u8 __mpa_rr_revision(__be16 mpa_rr_bits) 54{ 55 __be16 rev = mpa_rr_bits & MPA_RR_MASK_REVISION; 56 57 return be16_to_cpu(rev); 58} 59 60enum mpa_v2_ctrl { 61 MPA_V2_PEER_TO_PEER = cpu_to_be16(0x8000), 62 MPA_V2_ZERO_LENGTH_RTR = cpu_to_be16(0x4000), 63 MPA_V2_RDMA_WRITE_RTR = cpu_to_be16(0x8000), 64 MPA_V2_RDMA_READ_RTR = cpu_to_be16(0x4000), 65 MPA_V2_RDMA_NO_RTR = cpu_to_be16(0x0000), 66 MPA_V2_MASK_IRD_ORD = cpu_to_be16(0x3fff) 67}; 68 69struct mpa_v2_data { 70 __be16 ird; 71 __be16 ord; 72}; 73 74struct mpa_marker { 75 __be16 rsvd; 76 __be16 fpdu_hmd; /* FPDU header-marker distance (= MPA's FPDUPTR) */ 77}; 78 79/* 80 * maximum MPA trailer 81 */ 82struct mpa_trailer { 83 __u8 pad[4]; 84 __be32 crc; 85}; 86 87#define MPA_HDR_SIZE 2 88#define MPA_CRC_SIZE 4 89 90/* 91 * Common portion of iWARP headers (MPA, DDP, RDMAP) 92 * for any FPDU 93 */ 94struct iwarp_ctrl { 95 __be16 mpa_len; 96 __be16 ddp_rdmap_ctrl; 97}; 98 99/* 100 * DDP/RDMAP Hdr bits & fields 101 */ 102enum { 103 DDP_FLAG_TAGGED = cpu_to_be16(0x8000), 104 DDP_FLAG_LAST = cpu_to_be16(0x4000), 105 DDP_MASK_RESERVED = cpu_to_be16(0x3C00), 106 DDP_MASK_VERSION = cpu_to_be16(0x0300), 107 RDMAP_MASK_VERSION = cpu_to_be16(0x00C0), 108 RDMAP_MASK_RESERVED = cpu_to_be16(0x0030), 109 RDMAP_MASK_OPCODE = cpu_to_be16(0x000f) 110}; 111 112static inline u8 __ddp_get_version(struct iwarp_ctrl *ctrl) 113{ 114 return be16_to_cpu(ctrl->ddp_rdmap_ctrl & DDP_MASK_VERSION) >> 8; 115} 116 117static inline void __ddp_set_version(struct iwarp_ctrl *ctrl, u8 version) 118{ 119 ctrl->ddp_rdmap_ctrl = 120 (ctrl->ddp_rdmap_ctrl & ~DDP_MASK_VERSION) | 121 (cpu_to_be16((u16)version << 8) & DDP_MASK_VERSION); 122} 123 124static inline u8 __rdmap_get_version(struct iwarp_ctrl *ctrl) 125{ 126 __be16 ver = ctrl->ddp_rdmap_ctrl & RDMAP_MASK_VERSION; 127 128 return be16_to_cpu(ver) >> 6; 129} 130 131static inline void __rdmap_set_version(struct iwarp_ctrl *ctrl, u8 version) 132{ 133 ctrl->ddp_rdmap_ctrl = (ctrl->ddp_rdmap_ctrl & ~RDMAP_MASK_VERSION) | 134 (cpu_to_be16(version << 6) & RDMAP_MASK_VERSION); 135} 136 137static inline u8 __rdmap_get_opcode(struct iwarp_ctrl *ctrl) 138{ 139 return be16_to_cpu(ctrl->ddp_rdmap_ctrl & RDMAP_MASK_OPCODE); 140} 141 142static inline void __rdmap_set_opcode(struct iwarp_ctrl *ctrl, u8 opcode) 143{ 144 ctrl->ddp_rdmap_ctrl = (ctrl->ddp_rdmap_ctrl & ~RDMAP_MASK_OPCODE) | 145 (cpu_to_be16(opcode) & RDMAP_MASK_OPCODE); 146} 147 148struct iwarp_rdma_write { 149 struct iwarp_ctrl ctrl; 150 __be32 sink_stag; 151 __be64 sink_to; 152}; 153 154struct iwarp_rdma_rreq { 155 struct iwarp_ctrl ctrl; 156 __be32 rsvd; 157 __be32 ddp_qn; 158 __be32 ddp_msn; 159 __be32 ddp_mo; 160 __be32 sink_stag; 161 __be64 sink_to; 162 __be32 read_size; 163 __be32 source_stag; 164 __be64 source_to; 165}; 166 167struct iwarp_rdma_rresp { 168 struct iwarp_ctrl ctrl; 169 __be32 sink_stag; 170 __be64 sink_to; 171}; 172 173struct iwarp_send { 174 struct iwarp_ctrl ctrl; 175 __be32 rsvd; 176 __be32 ddp_qn; 177 __be32 ddp_msn; 178 __be32 ddp_mo; 179}; 180 181struct iwarp_send_inv { 182 struct iwarp_ctrl ctrl; 183 __be32 inval_stag; 184 __be32 ddp_qn; 185 __be32 ddp_msn; 186 __be32 ddp_mo; 187}; 188 189struct iwarp_terminate { 190 struct iwarp_ctrl ctrl; 191 __be32 rsvd; 192 __be32 ddp_qn; 193 __be32 ddp_msn; 194 __be32 ddp_mo; 195#if defined(__LITTLE_ENDIAN_BITFIELD) 196 __be32 layer : 4; 197 __be32 etype : 4; 198 __be32 ecode : 8; 199 __be32 flag_m : 1; 200 __be32 flag_d : 1; 201 __be32 flag_r : 1; 202 __be32 reserved : 13; 203#elif defined(__BIG_ENDIAN_BITFIELD) 204 __be32 reserved : 13; 205 __be32 flag_r : 1; 206 __be32 flag_d : 1; 207 __be32 flag_m : 1; 208 __be32 ecode : 8; 209 __be32 etype : 4; 210 __be32 layer : 4; 211#else 212#error "undefined byte order" 213#endif 214}; 215 216/* 217 * Terminate Hdr bits & fields 218 */ 219enum { 220 TERM_MASK_LAYER = cpu_to_be32(0xf0000000), 221 TERM_MASK_ETYPE = cpu_to_be32(0x0f000000), 222 TERM_MASK_ECODE = cpu_to_be32(0x00ff0000), 223 TERM_FLAG_M = cpu_to_be32(0x00008000), 224 TERM_FLAG_D = cpu_to_be32(0x00004000), 225 TERM_FLAG_R = cpu_to_be32(0x00002000), 226 TERM_MASK_RESVD = cpu_to_be32(0x00001fff) 227}; 228 229static inline u8 __rdmap_term_layer(struct iwarp_terminate *term) 230{ 231 return term->layer; 232} 233 234static inline void __rdmap_term_set_layer(struct iwarp_terminate *term, 235 u8 layer) 236{ 237 term->layer = layer & 0xf; 238} 239 240static inline u8 __rdmap_term_etype(struct iwarp_terminate *term) 241{ 242 return term->etype; 243} 244 245static inline void __rdmap_term_set_etype(struct iwarp_terminate *term, 246 u8 etype) 247{ 248 term->etype = etype & 0xf; 249} 250 251static inline u8 __rdmap_term_ecode(struct iwarp_terminate *term) 252{ 253 return term->ecode; 254} 255 256static inline void __rdmap_term_set_ecode(struct iwarp_terminate *term, 257 u8 ecode) 258{ 259 term->ecode = ecode; 260} 261 262/* 263 * Common portion of iWARP headers (MPA, DDP, RDMAP) 264 * for an FPDU carrying an untagged DDP segment 265 */ 266struct iwarp_ctrl_untagged { 267 struct iwarp_ctrl ctrl; 268 __be32 rsvd; 269 __be32 ddp_qn; 270 __be32 ddp_msn; 271 __be32 ddp_mo; 272}; 273 274/* 275 * Common portion of iWARP headers (MPA, DDP, RDMAP) 276 * for an FPDU carrying a tagged DDP segment 277 */ 278struct iwarp_ctrl_tagged { 279 struct iwarp_ctrl ctrl; 280 __be32 ddp_stag; 281 __be64 ddp_to; 282}; 283 284union iwarp_hdr { 285 struct iwarp_ctrl ctrl; 286 struct iwarp_ctrl_untagged c_untagged; 287 struct iwarp_ctrl_tagged c_tagged; 288 struct iwarp_rdma_write rwrite; 289 struct iwarp_rdma_rreq rreq; 290 struct iwarp_rdma_rresp rresp; 291 struct iwarp_terminate terminate; 292 struct iwarp_send send; 293 struct iwarp_send_inv send_inv; 294}; 295 296enum term_elayer { 297 TERM_ERROR_LAYER_RDMAP = 0x00, 298 TERM_ERROR_LAYER_DDP = 0x01, 299 TERM_ERROR_LAYER_LLP = 0x02 /* eg., MPA */ 300}; 301 302enum ddp_etype { 303 DDP_ETYPE_CATASTROPHIC = 0x0, 304 DDP_ETYPE_TAGGED_BUF = 0x1, 305 DDP_ETYPE_UNTAGGED_BUF = 0x2, 306 DDP_ETYPE_RSVD = 0x3 307}; 308 309enum ddp_ecode { 310 /* unspecified, set to zero */ 311 DDP_ECODE_CATASTROPHIC = 0x00, 312 /* Tagged Buffer Errors */ 313 DDP_ECODE_T_INVALID_STAG = 0x00, 314 DDP_ECODE_T_BASE_BOUNDS = 0x01, 315 DDP_ECODE_T_STAG_NOT_ASSOC = 0x02, 316 DDP_ECODE_T_TO_WRAP = 0x03, 317 DDP_ECODE_T_VERSION = 0x04, 318 /* Untagged Buffer Errors */ 319 DDP_ECODE_UT_INVALID_QN = 0x01, 320 DDP_ECODE_UT_INVALID_MSN_NOBUF = 0x02, 321 DDP_ECODE_UT_INVALID_MSN_RANGE = 0x03, 322 DDP_ECODE_UT_INVALID_MO = 0x04, 323 DDP_ECODE_UT_MSG_TOOLONG = 0x05, 324 DDP_ECODE_UT_VERSION = 0x06 325}; 326 327enum rdmap_untagged_qn { 328 RDMAP_UNTAGGED_QN_SEND = 0, 329 RDMAP_UNTAGGED_QN_RDMA_READ = 1, 330 RDMAP_UNTAGGED_QN_TERMINATE = 2, 331 RDMAP_UNTAGGED_QN_COUNT = 3 332}; 333 334enum rdmap_etype { 335 RDMAP_ETYPE_CATASTROPHIC = 0x0, 336 RDMAP_ETYPE_REMOTE_PROTECTION = 0x1, 337 RDMAP_ETYPE_REMOTE_OPERATION = 0x2 338}; 339 340enum rdmap_ecode { 341 RDMAP_ECODE_INVALID_STAG = 0x00, 342 RDMAP_ECODE_BASE_BOUNDS = 0x01, 343 RDMAP_ECODE_ACCESS_RIGHTS = 0x02, 344 RDMAP_ECODE_STAG_NOT_ASSOC = 0x03, 345 RDMAP_ECODE_TO_WRAP = 0x04, 346 RDMAP_ECODE_VERSION = 0x05, 347 RDMAP_ECODE_OPCODE = 0x06, 348 RDMAP_ECODE_CATASTROPHIC_STREAM = 0x07, 349 RDMAP_ECODE_CATASTROPHIC_GLOBAL = 0x08, 350 RDMAP_ECODE_CANNOT_INVALIDATE = 0x09, 351 RDMAP_ECODE_UNSPECIFIED = 0xff 352}; 353 354enum llp_ecode { 355 LLP_ECODE_TCP_STREAM_LOST = 0x01, /* How to transfer this ?? */ 356 LLP_ECODE_RECEIVED_CRC = 0x02, 357 LLP_ECODE_FPDU_START = 0x03, 358 LLP_ECODE_INVALID_REQ_RESP = 0x04, 359 360 /* Errors for Enhanced Connection Establishment only */ 361 LLP_ECODE_LOCAL_CATASTROPHIC = 0x05, 362 LLP_ECODE_INSUFFICIENT_IRD = 0x06, 363 LLP_ECODE_NO_MATCHING_RTR = 0x07 364}; 365 366enum llp_etype { LLP_ETYPE_MPA = 0x00 }; 367 368enum rdma_opcode { 369 RDMAP_RDMA_WRITE = 0x0, 370 RDMAP_RDMA_READ_REQ = 0x1, 371 RDMAP_RDMA_READ_RESP = 0x2, 372 RDMAP_SEND = 0x3, 373 RDMAP_SEND_INVAL = 0x4, 374 RDMAP_SEND_SE = 0x5, 375 RDMAP_SEND_SE_INVAL = 0x6, 376 RDMAP_TERMINATE = 0x7, 377 RDMAP_NOT_SUPPORTED = RDMAP_TERMINATE + 1 378}; 379 380#endif 381