/linux-6.1.9/include/rdma/ |
D | rdmavt_mr.h | 80 u8 num_sge; member 95 while (ss->num_sge) { in rvt_put_ss() 97 if (--ss->num_sge) in rvt_put_ss() 125 if (--ss->num_sge) in rvt_update_sge()
|
/linux-6.1.9/drivers/infiniband/sw/siw/ |
D | siw_verbs.c | 648 int num_sge = core_wr->num_sge, bytes = 0; in siw_copy_inline_sgl() local 653 while (num_sge--) { in siw_copy_inline_sgl() 670 sqe->num_sge = bytes > 0 ? 1 : 0; in siw_copy_inline_sgl() 826 if (wr->num_sge > qp->attrs.sq_max_sges) { in siw_post_send() 827 siw_dbg_qp(qp, "too many sge's: %d\n", wr->num_sge); in siw_post_send() 848 wr->num_sge); in siw_post_send() 849 sqe->num_sge = wr->num_sge; in siw_post_send() 857 sqe->num_sge = 1; in siw_post_send() 876 if (unlikely(wr->num_sge != 1)) { in siw_post_send() 886 sqe->num_sge = 1; in siw_post_send() [all …]
|
D | siw_mem.h | 32 static inline void siw_unref_mem_sgl(struct siw_mem **mem, unsigned int num_sge) in siw_unref_mem_sgl() argument 34 while (num_sge) { in siw_unref_mem_sgl() 41 num_sge--; in siw_unref_mem_sgl()
|
D | siw_verbs.h | 25 int num_sge) in siw_copy_sgl() argument 27 while (num_sge--) { in siw_copy_sgl()
|
D | siw_qp_rx.c | 352 int num_sge = rqe->num_sge; in siw_rqe_get() local 354 if (likely(num_sge <= SIW_MAX_SGE)) { in siw_rqe_get() 364 wqe->rqe.num_sge = num_sge; in siw_rqe_get() 366 while (i < num_sge) { in siw_rqe_get() 377 siw_dbg_qp(qp, "too many sge's: %d\n", rqe->num_sge); in siw_rqe_get() 573 wqe->rqe.num_sge = 1; in siw_proc_write() 714 resp->num_sge = length ? 1 : 0; in siw_init_rresp() 764 wqe->sqe.num_sge = 1; in siw_orqe_start_rx()
|
/linux-6.1.9/drivers/infiniband/sw/rxe/ |
D | rxe_verbs.c | 245 int num_sge = ibwr->num_sge; in post_one_recv() local 254 if (unlikely(num_sge > rq->max_sge)) { in post_one_recv() 260 for (i = 0; i < num_sge; i++) in post_one_recv() 267 num_sge * sizeof(struct ib_sge)); in post_one_recv() 271 recv_wqe->dma.num_sge = num_sge; in post_one_recv() 500 int num_sge = ibwr->num_sge; in validate_send_wr() local 503 if (unlikely(num_sge > sq->max_sge)) in validate_send_wr() 588 for (i = 0; i < ibwr->num_sge; i++, sge++) { in copy_inline_data_to_wqe() 598 int num_sge = ibwr->num_sge; in init_send_wqe() local 613 num_sge * sizeof(struct ib_sge)); in init_send_wqe() [all …]
|
/linux-6.1.9/drivers/infiniband/hw/erdma/ |
D | erdma_qp.c | 217 while (i < send_wr->num_sge) { in fill_inline_data() 258 if (send_wr->num_sge > qp->dev->attrs.max_send_sge) in fill_sgl() 264 while (i < send_wr->num_sge) { in fill_sgl() 343 if (unlikely(send_wr->num_sge != 1)) in erdma_push_one_sqe() 367 send_wr->num_sge * sizeof(struct ib_sge); in erdma_push_one_sqe() 442 wqe_size += send_wr->num_sge * sizeof(struct ib_sge); in erdma_push_one_sqe() 444 send_wr->num_sge); in erdma_push_one_sqe() 514 if (recv_wr->num_sge == 0) { in erdma_post_recv_one() 516 } else if (recv_wr->num_sge == 1) { in erdma_post_recv_one()
|
/linux-6.1.9/drivers/infiniband/sw/rdmavt/ |
D | trace_tx.h | 68 __field(int, num_sge) 88 __entry->num_sge = wqe->wr.num_sge; 109 __entry->num_sge,
|
D | rc.c | 166 ss->num_sge = wqe->wr.num_sge; in rvt_restart_sge()
|
D | qp.c | 627 for (i = 0; i < wqe->wr.num_sge; i++) { in rvt_swqe_has_lkey() 865 qp->r_sge.num_sge = 0; in rvt_init_qp() 1803 if ((unsigned)wr->num_sge > qp->r_rq.max_sge) { in rvt_post_recv() 1829 wqe->num_sge = wr->num_sge; in rvt_post_recv() 1830 for (i = 0; i < wr->num_sge; i++) { in rvt_post_recv() 1879 (wr->num_sge == 0 || in rvt_qp_valid_operation() 1983 if (unlikely(wr->num_sge > qp->s_max_sge)) in rvt_post_one_wr() 2047 if (wr->num_sge) { in rvt_post_one_wr() 2052 for (i = 0; i < wr->num_sge; i++) { in rvt_post_one_wr() 2066 wqe->wr.num_sge = j; in rvt_post_one_wr() [all …]
|
D | srq.c | 202 p->num_sge = wqe->num_sge; in rvt_modify_srq() 203 for (i = 0; i < wqe->num_sge; i++) in rvt_modify_srq()
|
/linux-6.1.9/net/rds/ |
D | ib_send.c | 630 send->s_wr.num_sge = 1; in rds_ib_xmit() 653 send->s_wr.num_sge = 2; in rds_ib_xmit() 683 &send->s_wr, send->s_wr.num_sge, send->s_wr.next); in rds_ib_xmit() 802 send->s_atomic_wr.wr.num_sge = 1; in rds_ib_xmit_atomic() 869 int num_sge; in rds_ib_xmit_rdma() local 915 num_sge = op->op_count; in rds_ib_xmit_rdma() 930 if (num_sge > max_sge) { in rds_ib_xmit_rdma() 931 send->s_rdma_wr.wr.num_sge = max_sge; in rds_ib_xmit_rdma() 932 num_sge -= max_sge; in rds_ib_xmit_rdma() 934 send->s_rdma_wr.wr.num_sge = num_sge; in rds_ib_xmit_rdma() [all …]
|
/linux-6.1.9/drivers/infiniband/core/ |
D | uverbs_std_types_mr.c | 55 int num_sge; in UVERBS_HANDLER() local 72 num_sge = uverbs_attr_ptr_get_array_size( in UVERBS_HANDLER() 74 if (num_sge <= 0) in UVERBS_HANDLER() 75 return num_sge; in UVERBS_HANDLER() 79 return ib_dev->ops.advise_mr(pd, advice, flags, sg_list, num_sge, in UVERBS_HANDLER()
|
/linux-6.1.9/drivers/infiniband/hw/mlx5/ |
D | wr.c | 272 for (i = 0; i < wr->num_sge; i++) { in set_data_inl_seg() 569 if (unlikely(send_wr->num_sge != 0) || in set_pi_umr_wr() 935 u8 next_fence, int *num_sge) in handle_qpt_rc() argument 955 *num_sge = 0; in handle_qpt_rc() 962 *num_sge = 0; in handle_qpt_rc() 971 *num_sge = 0; in handle_qpt_rc() 1064 int num_sge; in mlx5_ib_post_send() local 1090 num_sge = wr->num_sge; in mlx5_ib_post_send() 1091 if (unlikely(num_sge > qp->sq.max_gs)) { in mlx5_ib_post_send() 1131 next_fence, &num_sge); in mlx5_ib_post_send() [all …]
|
D | odp.c | 1659 u32 num_sge; member 1671 for (i = 0; i < work->num_sge; ++i) in destroy_prefetch_work() 1725 WARN_ON(!work->num_sge); in mlx5_ib_prefetch_mr_work() 1726 for (i = 0; i < work->num_sge; ++i) { in mlx5_ib_prefetch_mr_work() 1741 struct ib_sge *sg_list, u32 num_sge) in init_prefetch_work() argument 1748 for (i = 0; i < num_sge; ++i) { in init_prefetch_work() 1753 work->num_sge = i; in init_prefetch_work() 1760 work->num_sge = num_sge; in init_prefetch_work() 1767 u32 num_sge) in mlx5_ib_prefetch_sg_list() argument 1773 for (i = 0; i < num_sge; ++i) { in mlx5_ib_prefetch_sg_list() [all …]
|
/linux-6.1.9/include/uapi/rdma/ |
D | siw-abi.h | 110 __u8 num_sge; member 128 __u8 num_sge; member
|
D | vmw_pvrdma-abi.h | 231 __u32 num_sge; /* size of s/g array */ member 239 __u32 num_sge; /* size of s/g array */ member
|
D | rvt-abi.h | 45 __u8 num_sge; member
|
/linux-6.1.9/drivers/infiniband/hw/qib/ |
D | qib_uc.c | 101 qp->s_sge.num_sge = wqe->wr.num_sge; in qib_make_uc_req() 278 qp->r_sge.num_sge = 0; in qib_uc_rcv() 430 qp->r_sge.num_sge = 1; in qib_uc_rcv() 432 qp->r_sge.num_sge = 0; in qib_uc_rcv() 512 qp->r_sge.num_sge = 0; in qib_uc_rcv()
|
D | qib_ud.c | 173 ssge.num_sge = swqe->wr.num_sge; in qib_ud_loopback() 183 if (--ssge.num_sge) in qib_ud_loopback() 320 qp->s_sge.num_sge = wqe->wr.num_sge; in qib_make_ud_req()
|
/linux-6.1.9/drivers/infiniband/hw/hfi1/ |
D | uc.c | 114 qp->s_sge.num_sge = wqe->wr.num_sge; in hfi1_make_uc_req() 293 qp->r_sge.num_sge = 0; in hfi1_uc_rcv() 458 qp->r_sge.num_sge = 1; in hfi1_uc_rcv() 460 qp->r_sge.num_sge = 0; in hfi1_uc_rcv() 535 qp->r_sge.num_sge = 0; in hfi1_uc_rcv()
|
/linux-6.1.9/drivers/infiniband/ulp/rtrs/ |
D | rtrs.c | 88 .num_sge = 1, in rtrs_iu_post_recv() 143 .num_sge = 1, in rtrs_iu_post_send() 153 struct ib_sge *sge, unsigned int num_sge, in rtrs_iu_post_rdma_write_imm() argument 165 .wr.num_sge = num_sge, in rtrs_iu_post_rdma_write_imm() 177 for (i = 0; i < num_sge; i++) in rtrs_iu_post_rdma_write_imm()
|
/linux-6.1.9/drivers/infiniband/hw/vmw_pvrdma/ |
D | pvrdma_qp.c | 704 if (unlikely(wr->num_sge > qp->sq.max_sg || wr->num_sge < 0)) { in pvrdma_post_send() 761 wqe_hdr->num_sge = wr->num_sge; in pvrdma_post_send() 844 for (i = 0; i < wr->num_sge; i++) { in pvrdma_post_send() 912 if (unlikely(wr->num_sge > qp->rq.max_sg || in pvrdma_post_recv() 913 wr->num_sge < 0)) { in pvrdma_post_recv() 932 wqe_hdr->num_sge = wr->num_sge; in pvrdma_post_recv() 936 for (i = 0; i < wr->num_sge; i++) { in pvrdma_post_recv()
|
/linux-6.1.9/drivers/infiniband/hw/cxgb4/ |
D | qp.c | 423 for (i = 0; i < wr->num_sge; i++) { in build_immd() 455 int num_sge, u32 *plenp) in build_isgl() argument 467 for (i = 0; i < num_sge; i++) { in build_isgl() 482 isglp->nsge = cpu_to_be16(num_sge); in build_isgl() 496 if (wr->num_sge > T4_MAX_SEND_SGE) in build_rdma_send() 525 if (wr->num_sge) { in build_rdma_send() 537 wr->sg_list, wr->num_sge, &plen); in build_rdma_send() 541 wr->num_sge * sizeof(struct fw_ri_sge); in build_rdma_send() 563 if (wr->num_sge > T4_MAX_SEND_SGE) in build_rdma_write() 576 if (wr->num_sge) { in build_rdma_write() [all …]
|
/linux-6.1.9/drivers/infiniband/hw/qedr/ |
D | qedr_roce_cm.c | 407 for (i = 0; i < swr->num_sge; ++i) in qedr_gsi_build_header() 531 packet->n_seg = swr->num_sge; in qedr_gsi_build_packet() 559 if (wr->num_sge > RDMA_MAX_SGE_PER_SQ_WQE) { in qedr_gsi_post_send() 561 wr->num_sge, RDMA_MAX_SGE_PER_SQ_WQE); in qedr_gsi_post_send() 632 if (wr->num_sge > QEDR_GSI_MAX_RECV_SGE) { in qedr_gsi_post_recv() 635 wr->num_sge, QEDR_GSI_MAX_RECV_SGE); in qedr_gsi_post_recv()
|