1 /* SPDX-License-Identifier: GPL-2.0-or-later */
2 /*
3 * RDMA Transport Layer
4 *
5 * Copyright (c) 2014 - 2018 ProfitBricks GmbH. All rights reserved.
6 * Copyright (c) 2018 - 2019 1&1 IONOS Cloud GmbH. All rights reserved.
7 * Copyright (c) 2019 - 2020 1&1 IONOS SE. All rights reserved.
8 */
9
10 #ifndef RTRS_PRI_H
11 #define RTRS_PRI_H
12
13 #include <linux/uuid.h>
14 #include <rdma/rdma_cm.h>
15 #include <rdma/ib_verbs.h>
16 #include <rdma/ib.h>
17
18 #include "rtrs.h"
19
20 #define RTRS_PROTO_VER_MAJOR 2
21 #define RTRS_PROTO_VER_MINOR 0
22
23 #define RTRS_PROTO_VER_STRING __stringify(RTRS_PROTO_VER_MAJOR) "." \
24 __stringify(RTRS_PROTO_VER_MINOR)
25
26 /*
27 * Max IB immediate data size is 2^28 (MAX_IMM_PAYL_BITS)
28 * and the minimum chunk size is 4096 (2^12).
29 * So the maximum sess_queue_depth is 65535 (2^16 - 1) in theory
30 * since queue_depth in rtrs_msg_conn_rsp is defined as le16.
31 * Therefore the pratical max value of sess_queue_depth is
32 * somewhere between 1 and 65535 and it depends on the system.
33 */
34 #define MAX_SESS_QUEUE_DEPTH 65535
35
36 enum rtrs_imm_const {
37 MAX_IMM_TYPE_BITS = 4,
38 MAX_IMM_TYPE_MASK = ((1 << MAX_IMM_TYPE_BITS) - 1),
39 MAX_IMM_PAYL_BITS = 28,
40 MAX_IMM_PAYL_MASK = ((1 << MAX_IMM_PAYL_BITS) - 1),
41 };
42
43 enum rtrs_imm_type {
44 RTRS_IO_REQ_IMM = 0, /* client to server */
45 RTRS_IO_RSP_IMM = 1, /* server to client */
46 RTRS_IO_RSP_W_INV_IMM = 2, /* server to client */
47
48 RTRS_HB_MSG_IMM = 8, /* HB: HeartBeat */
49 RTRS_HB_ACK_IMM = 9,
50
51 RTRS_LAST_IMM,
52 };
53
54 enum {
55 SERVICE_CON_QUEUE_DEPTH = 512,
56
57 MAX_PATHS_NUM = 128,
58
59 MIN_CHUNK_SIZE = 8192,
60
61 RTRS_HB_INTERVAL_MS = 5000,
62 RTRS_HB_MISSED_MAX = 5,
63
64 RTRS_MAGIC = 0x1BBD,
65 RTRS_PROTO_VER = (RTRS_PROTO_VER_MAJOR << 8) | RTRS_PROTO_VER_MINOR,
66 };
67
68 struct rtrs_ib_dev;
69
70 struct rtrs_rdma_dev_pd_ops {
71 struct rtrs_ib_dev *(*alloc)(void);
72 void (*free)(struct rtrs_ib_dev *dev);
73 int (*init)(struct rtrs_ib_dev *dev);
74 void (*deinit)(struct rtrs_ib_dev *dev);
75 };
76
77 struct rtrs_rdma_dev_pd {
78 struct mutex mutex;
79 struct list_head list;
80 enum ib_pd_flags pd_flags;
81 const struct rtrs_rdma_dev_pd_ops *ops;
82 };
83
84 struct rtrs_ib_dev {
85 struct ib_device *ib_dev;
86 struct ib_pd *ib_pd;
87 struct kref ref;
88 struct list_head entry;
89 struct rtrs_rdma_dev_pd *pool;
90 };
91
92 struct rtrs_con {
93 struct rtrs_path *path;
94 struct ib_qp *qp;
95 struct ib_cq *cq;
96 struct rdma_cm_id *cm_id;
97 unsigned int cid;
98 int nr_cqe;
99 atomic_t wr_cnt;
100 atomic_t sq_wr_avail;
101 };
102
103 struct rtrs_path {
104 struct list_head entry;
105 struct sockaddr_storage dst_addr;
106 struct sockaddr_storage src_addr;
107 char sessname[NAME_MAX];
108 uuid_t uuid;
109 struct rtrs_con **con;
110 unsigned int con_num;
111 unsigned int irq_con_num;
112 unsigned int recon_cnt;
113 unsigned int signal_interval;
114 struct rtrs_ib_dev *dev;
115 int dev_ref;
116 struct ib_cqe *hb_cqe;
117 void (*hb_err_handler)(struct rtrs_con *con);
118 struct workqueue_struct *hb_wq;
119 struct delayed_work hb_dwork;
120 unsigned int hb_interval_ms;
121 unsigned int hb_missed_cnt;
122 unsigned int hb_missed_max;
123 ktime_t hb_last_sent;
124 ktime_t hb_cur_latency;
125 };
126
127 /* rtrs information unit */
128 struct rtrs_iu {
129 struct ib_cqe cqe;
130 dma_addr_t dma_addr;
131 void *buf;
132 size_t size;
133 enum dma_data_direction direction;
134 };
135
136 /**
137 * enum rtrs_msg_types - RTRS message types, see also rtrs/README
138 * @RTRS_MSG_INFO_REQ: Client additional info request to the server
139 * @RTRS_MSG_INFO_RSP: Server additional info response to the client
140 * @RTRS_MSG_WRITE: Client writes data per RDMA to server
141 * @RTRS_MSG_READ: Client requests data transfer from server
142 * @RTRS_MSG_RKEY_RSP: Server refreshed rkey for rbuf
143 */
144 enum rtrs_msg_types {
145 RTRS_MSG_INFO_REQ,
146 RTRS_MSG_INFO_RSP,
147 RTRS_MSG_WRITE,
148 RTRS_MSG_READ,
149 RTRS_MSG_RKEY_RSP,
150 };
151
152 /**
153 * enum rtrs_msg_flags - RTRS message flags.
154 * @RTRS_NEED_INVAL: Send invalidation in response.
155 * @RTRS_MSG_NEW_RKEY_F: Send refreshed rkey in response.
156 */
157 enum rtrs_msg_flags {
158 RTRS_MSG_NEED_INVAL_F = 1 << 0,
159 RTRS_MSG_NEW_RKEY_F = 1 << 1,
160 };
161
162 /**
163 * struct rtrs_sg_desc - RDMA-Buffer entry description
164 * @addr: Address of RDMA destination buffer
165 * @key: Authorization rkey to write to the buffer
166 * @len: Size of the buffer
167 */
168 struct rtrs_sg_desc {
169 __le64 addr;
170 __le32 key;
171 __le32 len;
172 };
173
174 /**
175 * struct rtrs_msg_conn_req - Client connection request to the server
176 * @magic: RTRS magic
177 * @version: RTRS protocol version
178 * @cid: Current connection id
179 * @cid_num: Number of connections per session
180 * @recon_cnt: Reconnections counter
181 * @sess_uuid: UUID of a session (path)
182 * @paths_uuid: UUID of a group of sessions (paths)
183 *
184 * NOTE: max size 56 bytes, see man rdma_connect().
185 */
186 struct rtrs_msg_conn_req {
187 /* Is set to 0 by cma.c in case of AF_IB, do not touch that.
188 * see https://www.spinics.net/lists/linux-rdma/msg22397.html
189 */
190 u8 __cma_version;
191 /* On sender side that should be set to 0, or cma_save_ip_info()
192 * extract garbage and will fail.
193 */
194 u8 __ip_version;
195 __le16 magic;
196 __le16 version;
197 __le16 cid;
198 __le16 cid_num;
199 __le16 recon_cnt;
200 uuid_t sess_uuid;
201 uuid_t paths_uuid;
202 u8 first_conn : 1;
203 u8 reserved_bits : 7;
204 u8 reserved[11];
205 };
206
207 /**
208 * struct rtrs_msg_conn_rsp - Server connection response to the client
209 * @magic: RTRS magic
210 * @version: RTRS protocol version
211 * @errno: If rdma_accept() then 0, if rdma_reject() indicates error
212 * @queue_depth: max inflight messages (queue-depth) in this session
213 * @max_io_size: max io size server supports
214 * @max_hdr_size: max msg header size server supports
215 *
216 * NOTE: size is 56 bytes, max possible is 136 bytes, see man rdma_accept().
217 */
218 struct rtrs_msg_conn_rsp {
219 __le16 magic;
220 __le16 version;
221 __le16 errno;
222 __le16 queue_depth;
223 __le32 max_io_size;
224 __le32 max_hdr_size;
225 __le32 flags;
226 u8 reserved[36];
227 };
228
229 /**
230 * struct rtrs_msg_info_req
231 * @type: @RTRS_MSG_INFO_REQ
232 * @pathname: Path name chosen by client
233 */
234 struct rtrs_msg_info_req {
235 __le16 type;
236 u8 pathname[NAME_MAX];
237 u8 reserved[15];
238 };
239
240 /**
241 * struct rtrs_msg_info_rsp
242 * @type: @RTRS_MSG_INFO_RSP
243 * @sg_cnt: Number of @desc entries
244 * @desc: RDMA buffers where the client can write to server
245 */
246 struct rtrs_msg_info_rsp {
247 __le16 type;
248 __le16 sg_cnt;
249 u8 reserved[4];
250 struct rtrs_sg_desc desc[];
251 };
252
253 /**
254 * struct rtrs_msg_rkey_rsp
255 * @type: @RTRS_MSG_RKEY_RSP
256 * @buf_id: RDMA buf_id of the new rkey
257 * @rkey: new remote key for RDMA buffers id from server
258 */
259 struct rtrs_msg_rkey_rsp {
260 __le16 type;
261 __le16 buf_id;
262 __le32 rkey;
263 };
264
265 /**
266 * struct rtrs_msg_rdma_read - RDMA data transfer request from client
267 * @type: always @RTRS_MSG_READ
268 * @usr_len: length of user payload
269 * @sg_cnt: number of @desc entries
270 * @desc: RDMA buffers where the server can write the result to
271 */
272 struct rtrs_msg_rdma_read {
273 __le16 type;
274 __le16 usr_len;
275 __le16 flags;
276 __le16 sg_cnt;
277 struct rtrs_sg_desc desc[];
278 };
279
280 /**
281 * struct_msg_rdma_write - Message transferred to server with RDMA-Write
282 * @type: always @RTRS_MSG_WRITE
283 * @usr_len: length of user payload
284 */
285 struct rtrs_msg_rdma_write {
286 __le16 type;
287 __le16 usr_len;
288 };
289
290 /**
291 * struct_msg_rdma_hdr - header for read or write request
292 * @type: @RTRS_MSG_WRITE | @RTRS_MSG_READ
293 */
294 struct rtrs_msg_rdma_hdr {
295 __le16 type;
296 };
297
298 /* rtrs.c */
299
300 struct rtrs_iu *rtrs_iu_alloc(u32 queue_num, size_t size, gfp_t t,
301 struct ib_device *dev, enum dma_data_direction,
302 void (*done)(struct ib_cq *cq, struct ib_wc *wc));
303 void rtrs_iu_free(struct rtrs_iu *iu, struct ib_device *dev, u32 queue_num);
304 int rtrs_iu_post_recv(struct rtrs_con *con, struct rtrs_iu *iu);
305 int rtrs_iu_post_send(struct rtrs_con *con, struct rtrs_iu *iu, size_t size,
306 struct ib_send_wr *head);
307 int rtrs_iu_post_rdma_write_imm(struct rtrs_con *con, struct rtrs_iu *iu,
308 struct ib_sge *sge, unsigned int num_sge,
309 u32 rkey, u64 rdma_addr, u32 imm_data,
310 enum ib_send_flags flags,
311 struct ib_send_wr *head,
312 struct ib_send_wr *tail);
313
314 int rtrs_post_recv_empty(struct rtrs_con *con, struct ib_cqe *cqe);
315
316 int rtrs_cq_qp_create(struct rtrs_path *path, struct rtrs_con *con,
317 u32 max_send_sge, int cq_vector, int nr_cqe,
318 u32 max_send_wr, u32 max_recv_wr,
319 enum ib_poll_context poll_ctx);
320 void rtrs_cq_qp_destroy(struct rtrs_con *con);
321
322 void rtrs_init_hb(struct rtrs_path *path, struct ib_cqe *cqe,
323 unsigned int interval_ms, unsigned int missed_max,
324 void (*err_handler)(struct rtrs_con *con),
325 struct workqueue_struct *wq);
326 void rtrs_start_hb(struct rtrs_path *path);
327 void rtrs_stop_hb(struct rtrs_path *path);
328 void rtrs_send_hb_ack(struct rtrs_path *path);
329
330 void rtrs_rdma_dev_pd_init(enum ib_pd_flags pd_flags,
331 struct rtrs_rdma_dev_pd *pool);
332 void rtrs_rdma_dev_pd_deinit(struct rtrs_rdma_dev_pd *pool);
333
334 struct rtrs_ib_dev *rtrs_ib_dev_find_or_add(struct ib_device *ib_dev,
335 struct rtrs_rdma_dev_pd *pool);
336 int rtrs_ib_dev_put(struct rtrs_ib_dev *dev);
337
rtrs_to_imm(u32 type,u32 payload)338 static inline u32 rtrs_to_imm(u32 type, u32 payload)
339 {
340 BUILD_BUG_ON(MAX_IMM_PAYL_BITS + MAX_IMM_TYPE_BITS != 32);
341 BUILD_BUG_ON(RTRS_LAST_IMM > (1<<MAX_IMM_TYPE_BITS));
342 return ((type & MAX_IMM_TYPE_MASK) << MAX_IMM_PAYL_BITS) |
343 (payload & MAX_IMM_PAYL_MASK);
344 }
345
rtrs_from_imm(u32 imm,u32 * type,u32 * payload)346 static inline void rtrs_from_imm(u32 imm, u32 *type, u32 *payload)
347 {
348 *payload = imm & MAX_IMM_PAYL_MASK;
349 *type = imm >> MAX_IMM_PAYL_BITS;
350 }
351
rtrs_to_io_req_imm(u32 addr)352 static inline u32 rtrs_to_io_req_imm(u32 addr)
353 {
354 return rtrs_to_imm(RTRS_IO_REQ_IMM, addr);
355 }
356
rtrs_to_io_rsp_imm(u32 msg_id,int errno,bool w_inval)357 static inline u32 rtrs_to_io_rsp_imm(u32 msg_id, int errno, bool w_inval)
358 {
359 enum rtrs_imm_type type;
360 u32 payload;
361
362 /* 9 bits for errno, 19 bits for msg_id */
363 payload = (abs(errno) & 0x1ff) << 19 | (msg_id & 0x7ffff);
364 type = w_inval ? RTRS_IO_RSP_W_INV_IMM : RTRS_IO_RSP_IMM;
365
366 return rtrs_to_imm(type, payload);
367 }
368
rtrs_from_io_rsp_imm(u32 payload,u32 * msg_id,int * errno)369 static inline void rtrs_from_io_rsp_imm(u32 payload, u32 *msg_id, int *errno)
370 {
371 /* 9 bits for errno, 19 bits for msg_id */
372 *msg_id = payload & 0x7ffff;
373 *errno = -(int)((payload >> 19) & 0x1ff);
374 }
375
376 #define STAT_STORE_FUNC(type, set_value, reset) \
377 static ssize_t set_value##_store(struct kobject *kobj, \
378 struct kobj_attribute *attr, \
379 const char *buf, size_t count) \
380 { \
381 int ret = -EINVAL; \
382 type *stats = container_of(kobj, type, kobj_stats); \
383 \
384 if (sysfs_streq(buf, "1")) \
385 ret = reset(stats, true); \
386 else if (sysfs_streq(buf, "0")) \
387 ret = reset(stats, false); \
388 if (ret) \
389 return ret; \
390 \
391 return count; \
392 }
393
394 #define STAT_SHOW_FUNC(type, get_value, print) \
395 static ssize_t get_value##_show(struct kobject *kobj, \
396 struct kobj_attribute *attr, \
397 char *page) \
398 { \
399 type *stats = container_of(kobj, type, kobj_stats); \
400 \
401 return print(stats, page); \
402 }
403
404 #define STAT_ATTR(type, stat, print, reset) \
405 STAT_STORE_FUNC(type, stat, reset) \
406 STAT_SHOW_FUNC(type, stat, print) \
407 static struct kobj_attribute stat##_attr = __ATTR_RW(stat)
408
409 #endif /* RTRS_PRI_H */
410