Lines Matching refs:msg

39 static int fill_sq(struct sk_buff *msg, struct t4_wq *wq)  in fill_sq()  argument
42 if (rdma_nl_put_driver_u32(msg, "sqid", wq->sq.qid)) in fill_sq()
44 if (rdma_nl_put_driver_u32(msg, "flushed", wq->flushed)) in fill_sq()
46 if (rdma_nl_put_driver_u32(msg, "memsize", wq->sq.memsize)) in fill_sq()
48 if (rdma_nl_put_driver_u32(msg, "cidx", wq->sq.cidx)) in fill_sq()
50 if (rdma_nl_put_driver_u32(msg, "pidx", wq->sq.pidx)) in fill_sq()
52 if (rdma_nl_put_driver_u32(msg, "wq_pidx", wq->sq.wq_pidx)) in fill_sq()
54 if (rdma_nl_put_driver_u32(msg, "flush_cidx", wq->sq.flush_cidx)) in fill_sq()
56 if (rdma_nl_put_driver_u32(msg, "in_use", wq->sq.in_use)) in fill_sq()
58 if (rdma_nl_put_driver_u32(msg, "size", wq->sq.size)) in fill_sq()
60 if (rdma_nl_put_driver_u32_hex(msg, "flags", wq->sq.flags)) in fill_sq()
67 static int fill_rq(struct sk_buff *msg, struct t4_wq *wq) in fill_rq() argument
70 if (rdma_nl_put_driver_u32(msg, "rqid", wq->rq.qid)) in fill_rq()
72 if (rdma_nl_put_driver_u32(msg, "memsize", wq->rq.memsize)) in fill_rq()
74 if (rdma_nl_put_driver_u32(msg, "cidx", wq->rq.cidx)) in fill_rq()
76 if (rdma_nl_put_driver_u32(msg, "pidx", wq->rq.pidx)) in fill_rq()
78 if (rdma_nl_put_driver_u32(msg, "wq_pidx", wq->rq.wq_pidx)) in fill_rq()
80 if (rdma_nl_put_driver_u32(msg, "msn", wq->rq.msn)) in fill_rq()
82 if (rdma_nl_put_driver_u32_hex(msg, "rqt_hwaddr", wq->rq.rqt_hwaddr)) in fill_rq()
84 if (rdma_nl_put_driver_u32(msg, "rqt_size", wq->rq.rqt_size)) in fill_rq()
86 if (rdma_nl_put_driver_u32(msg, "in_use", wq->rq.in_use)) in fill_rq()
88 if (rdma_nl_put_driver_u32(msg, "size", wq->rq.size)) in fill_rq()
95 static int fill_swsqe(struct sk_buff *msg, struct t4_sq *sq, u16 idx, in fill_swsqe() argument
98 if (rdma_nl_put_driver_u32(msg, "idx", idx)) in fill_swsqe()
100 if (rdma_nl_put_driver_u32(msg, "opcode", sqe->opcode)) in fill_swsqe()
102 if (rdma_nl_put_driver_u32(msg, "complete", sqe->complete)) in fill_swsqe()
105 rdma_nl_put_driver_u32(msg, "cqe_status", CQE_STATUS(&sqe->cqe))) in fill_swsqe()
107 if (rdma_nl_put_driver_u32(msg, "signaled", sqe->signaled)) in fill_swsqe()
109 if (rdma_nl_put_driver_u32(msg, "flushed", sqe->flushed)) in fill_swsqe()
119 static int fill_swsqes(struct sk_buff *msg, struct t4_sq *sq, in fill_swsqes() argument
125 if (fill_swsqe(msg, sq, first_idx, first_sqe)) in fill_swsqes()
129 if (fill_swsqe(msg, sq, last_idx, last_sqe)) in fill_swsqes()
137 int c4iw_fill_res_qp_entry(struct sk_buff *msg, struct ib_qp *ibqp) in c4iw_fill_res_qp_entry() argument
150 table_attr = nla_nest_start_noflag(msg, RDMA_NLDEV_ATTR_DRIVER); in c4iw_fill_res_qp_entry()
173 if (fill_sq(msg, &wq)) in c4iw_fill_res_qp_entry()
176 if (fill_swsqes(msg, &wq.sq, first_sq_idx, fsp, last_sq_idx, lsp)) in c4iw_fill_res_qp_entry()
179 if (fill_rq(msg, &wq)) in c4iw_fill_res_qp_entry()
182 nla_nest_end(msg, table_attr); in c4iw_fill_res_qp_entry()
186 nla_nest_cancel(msg, table_attr); in c4iw_fill_res_qp_entry()
196 int c4iw_fill_res_cm_id_entry(struct sk_buff *msg, in c4iw_fill_res_cm_id_entry() argument
216 table_attr = nla_nest_start_noflag(msg, RDMA_NLDEV_ATTR_DRIVER); in c4iw_fill_res_cm_id_entry()
234 if (rdma_nl_put_driver_u32(msg, "state", epcp->state)) in c4iw_fill_res_cm_id_entry()
236 if (rdma_nl_put_driver_u64_hex(msg, "flags", epcp->flags)) in c4iw_fill_res_cm_id_entry()
238 if (rdma_nl_put_driver_u64_hex(msg, "history", epcp->history)) in c4iw_fill_res_cm_id_entry()
242 if (rdma_nl_put_driver_u32(msg, "stid", listen_ep->stid)) in c4iw_fill_res_cm_id_entry()
244 if (rdma_nl_put_driver_u32(msg, "backlog", listen_ep->backlog)) in c4iw_fill_res_cm_id_entry()
247 if (rdma_nl_put_driver_u32(msg, "hwtid", ep->hwtid)) in c4iw_fill_res_cm_id_entry()
249 if (rdma_nl_put_driver_u32(msg, "ord", ep->ord)) in c4iw_fill_res_cm_id_entry()
251 if (rdma_nl_put_driver_u32(msg, "ird", ep->ird)) in c4iw_fill_res_cm_id_entry()
253 if (rdma_nl_put_driver_u32(msg, "emss", ep->emss)) in c4iw_fill_res_cm_id_entry()
256 if (!ep->parent_ep && rdma_nl_put_driver_u32(msg, "atid", in c4iw_fill_res_cm_id_entry()
260 nla_nest_end(msg, table_attr); in c4iw_fill_res_cm_id_entry()
265 nla_nest_cancel(msg, table_attr); in c4iw_fill_res_cm_id_entry()
271 static int fill_cq(struct sk_buff *msg, struct t4_cq *cq) in fill_cq() argument
273 if (rdma_nl_put_driver_u32(msg, "cqid", cq->cqid)) in fill_cq()
275 if (rdma_nl_put_driver_u32(msg, "memsize", cq->memsize)) in fill_cq()
277 if (rdma_nl_put_driver_u32(msg, "size", cq->size)) in fill_cq()
279 if (rdma_nl_put_driver_u32(msg, "cidx", cq->cidx)) in fill_cq()
281 if (rdma_nl_put_driver_u32(msg, "cidx_inc", cq->cidx_inc)) in fill_cq()
283 if (rdma_nl_put_driver_u32(msg, "sw_cidx", cq->sw_cidx)) in fill_cq()
285 if (rdma_nl_put_driver_u32(msg, "sw_pidx", cq->sw_pidx)) in fill_cq()
287 if (rdma_nl_put_driver_u32(msg, "sw_in_use", cq->sw_in_use)) in fill_cq()
289 if (rdma_nl_put_driver_u32(msg, "vector", cq->vector)) in fill_cq()
291 if (rdma_nl_put_driver_u32(msg, "gen", cq->gen)) in fill_cq()
293 if (rdma_nl_put_driver_u32(msg, "error", cq->error)) in fill_cq()
295 if (rdma_nl_put_driver_u64_hex(msg, "bits_type_ts", in fill_cq()
298 if (rdma_nl_put_driver_u64_hex(msg, "flags", cq->flags)) in fill_cq()
307 static int fill_cqe(struct sk_buff *msg, struct t4_cqe *cqe, u16 idx, in fill_cqe() argument
310 if (rdma_nl_put_driver_u32(msg, qstr, idx)) in fill_cqe()
312 if (rdma_nl_put_driver_u32_hex(msg, "header", in fill_cqe()
315 if (rdma_nl_put_driver_u32(msg, "len", be32_to_cpu(cqe->len))) in fill_cqe()
317 if (rdma_nl_put_driver_u32_hex(msg, "wrid_hi", in fill_cqe()
320 if (rdma_nl_put_driver_u32_hex(msg, "wrid_low", in fill_cqe()
323 if (rdma_nl_put_driver_u64_hex(msg, "bits_type_ts", in fill_cqe()
333 static int fill_hwcqes(struct sk_buff *msg, struct t4_cq *cq, in fill_hwcqes() argument
339 if (fill_cqe(msg, cqes, idx, "hwcq_idx")) in fill_hwcqes()
342 if (fill_cqe(msg, cqes + 1, idx, "hwcq_idx")) in fill_hwcqes()
350 static int fill_swcqes(struct sk_buff *msg, struct t4_cq *cq, in fill_swcqes() argument
359 if (fill_cqe(msg, cqes, idx, "swcq_idx")) in fill_swcqes()
364 if (fill_cqe(msg, cqes + 1, idx, "swcq_idx")) in fill_swcqes()
372 int c4iw_fill_res_cq_entry(struct sk_buff *msg, struct ib_cq *ibcq) in c4iw_fill_res_cq_entry() argument
385 table_attr = nla_nest_start_noflag(msg, RDMA_NLDEV_ATTR_DRIVER); in c4iw_fill_res_cq_entry()
413 if (fill_cq(msg, &cq)) in c4iw_fill_res_cq_entry()
416 if (fill_swcqes(msg, &cq, swcqes)) in c4iw_fill_res_cq_entry()
419 if (fill_hwcqes(msg, &cq, hwcqes)) in c4iw_fill_res_cq_entry()
422 nla_nest_end(msg, table_attr); in c4iw_fill_res_cq_entry()
426 nla_nest_cancel(msg, table_attr); in c4iw_fill_res_cq_entry()
431 int c4iw_fill_res_mr_entry(struct sk_buff *msg, struct ib_mr *ibmr) in c4iw_fill_res_mr_entry() argument
443 table_attr = nla_nest_start_noflag(msg, RDMA_NLDEV_ATTR_DRIVER); in c4iw_fill_res_mr_entry()
454 if (rdma_nl_put_driver_u32_hex(msg, "idx", stag >> 8)) in c4iw_fill_res_mr_entry()
456 if (rdma_nl_put_driver_u32(msg, "valid", in c4iw_fill_res_mr_entry()
459 if (rdma_nl_put_driver_u32_hex(msg, "key", stag & 0xff)) in c4iw_fill_res_mr_entry()
461 if (rdma_nl_put_driver_u32(msg, "state", in c4iw_fill_res_mr_entry()
464 if (rdma_nl_put_driver_u32(msg, "pdid", in c4iw_fill_res_mr_entry()
467 if (rdma_nl_put_driver_u32_hex(msg, "perm", in c4iw_fill_res_mr_entry()
470 if (rdma_nl_put_driver_u32(msg, "ps", in c4iw_fill_res_mr_entry()
473 if (rdma_nl_put_driver_u64(msg, "len", in c4iw_fill_res_mr_entry()
476 if (rdma_nl_put_driver_u32_hex(msg, "pbl_addr", in c4iw_fill_res_mr_entry()
480 nla_nest_end(msg, table_attr); in c4iw_fill_res_mr_entry()
484 nla_nest_cancel(msg, table_attr); in c4iw_fill_res_mr_entry()