Lines Matching refs:wqe
228 struct mlx5r_umr_wqe *wqe, bool with_data) in mlx5r_umr_post_send() argument
259 mlx5r_memcpy_send_wqe(&qp->sq, &cur_edge, &seg, &size, wqe, wqe_size); in mlx5r_umr_post_send()
289 struct mlx5r_umr_wqe *wqe, bool with_data) in mlx5r_umr_post_send_wait() argument
295 err = umr_check_mkey_mask(dev, be64_to_cpu(wqe->ctrl_seg.mkey_mask)); in mlx5r_umr_post_send_wait()
316 err = mlx5r_umr_post_send(umrc->qp, mkey, &umr_context.cqe, wqe, in mlx5r_umr_post_send_wait()
361 struct mlx5r_umr_wqe wqe = {}; in mlx5r_umr_revoke_mr() local
366 wqe.ctrl_seg.mkey_mask |= get_umr_update_pd_mask(); in mlx5r_umr_revoke_mr()
367 wqe.ctrl_seg.mkey_mask |= get_umr_disable_mr_mask(); in mlx5r_umr_revoke_mr()
368 wqe.ctrl_seg.flags |= MLX5_UMR_INLINE; in mlx5r_umr_revoke_mr()
370 MLX5_SET(mkc, &wqe.mkey_seg, free, 1); in mlx5r_umr_revoke_mr()
371 MLX5_SET(mkc, &wqe.mkey_seg, pd, to_mpd(dev->umrc.pd)->pdn); in mlx5r_umr_revoke_mr()
372 MLX5_SET(mkc, &wqe.mkey_seg, qpn, 0xffffff); in mlx5r_umr_revoke_mr()
373 MLX5_SET(mkc, &wqe.mkey_seg, mkey_7_0, in mlx5r_umr_revoke_mr()
376 return mlx5r_umr_post_send_wait(dev, mr->mmkey.key, &wqe, false); in mlx5r_umr_revoke_mr()
398 struct mlx5r_umr_wqe wqe = {}; in mlx5r_umr_rereg_pd_access() local
401 wqe.ctrl_seg.mkey_mask = get_umr_update_access_mask(dev); in mlx5r_umr_rereg_pd_access()
402 wqe.ctrl_seg.mkey_mask |= get_umr_update_pd_mask(); in mlx5r_umr_rereg_pd_access()
403 wqe.ctrl_seg.flags = MLX5_UMR_CHECK_FREE; in mlx5r_umr_rereg_pd_access()
404 wqe.ctrl_seg.flags |= MLX5_UMR_INLINE; in mlx5r_umr_rereg_pd_access()
406 mlx5r_umr_set_access_flags(dev, &wqe.mkey_seg, access_flags); in mlx5r_umr_rereg_pd_access()
407 MLX5_SET(mkc, &wqe.mkey_seg, pd, to_mpd(pd)->pdn); in mlx5r_umr_rereg_pd_access()
408 MLX5_SET(mkc, &wqe.mkey_seg, qpn, 0xffffff); in mlx5r_umr_rereg_pd_access()
409 MLX5_SET(mkc, &wqe.mkey_seg, mkey_7_0, in mlx5r_umr_rereg_pd_access()
412 err = mlx5r_umr_post_send_wait(dev, mr->mmkey.key, &wqe, false); in mlx5r_umr_rereg_pd_access()
571 struct mlx5r_umr_wqe *wqe, in mlx5r_umr_final_update_xlt() argument
578 wqe->ctrl_seg.mkey_mask |= get_umr_enable_mr_mask(); in mlx5r_umr_final_update_xlt()
585 wqe->ctrl_seg.mkey_mask |= get_umr_update_access_mask(dev); in mlx5r_umr_final_update_xlt()
586 wqe->ctrl_seg.mkey_mask |= get_umr_update_pd_mask(); in mlx5r_umr_final_update_xlt()
593 wqe->ctrl_seg.mkey_mask |= get_umr_update_translation_mask(); in mlx5r_umr_final_update_xlt()
595 MLX5_SET(mkc, &wqe->mkey_seg, length64, 1); in mlx5r_umr_final_update_xlt()
598 wqe->ctrl_seg.xlt_octowords = in mlx5r_umr_final_update_xlt()
600 wqe->data_seg.byte_count = cpu_to_be32(sg->length); in mlx5r_umr_final_update_xlt()
612 struct mlx5r_umr_wqe wqe = {}; in mlx5r_umr_update_mr_pas() local
633 mlx5r_umr_set_update_xlt_ctrl_seg(&wqe.ctrl_seg, flags, &sg); in mlx5r_umr_update_mr_pas()
634 mlx5r_umr_set_update_xlt_mkey_seg(dev, &wqe.mkey_seg, mr, in mlx5r_umr_update_mr_pas()
636 mlx5r_umr_set_update_xlt_data_seg(&wqe.data_seg, &sg); in mlx5r_umr_update_mr_pas()
646 err = mlx5r_umr_post_send_wait(dev, mr->mmkey.key, &wqe, in mlx5r_umr_update_mr_pas()
653 mlx5r_umr_update_offset(&wqe.ctrl_seg, offset); in mlx5r_umr_update_mr_pas()
671 mlx5r_umr_final_update_xlt(dev, &wqe, mr, &sg, flags); in mlx5r_umr_update_mr_pas()
674 err = mlx5r_umr_post_send_wait(dev, mr->mmkey.key, &wqe, true); in mlx5r_umr_update_mr_pas()
697 struct mlx5r_umr_wqe wqe = {}; in mlx5r_umr_update_xlt() local
737 mlx5r_umr_set_update_xlt_ctrl_seg(&wqe.ctrl_seg, flags, &sg); in mlx5r_umr_update_xlt()
738 mlx5r_umr_set_update_xlt_mkey_seg(dev, &wqe.mkey_seg, mr, page_shift); in mlx5r_umr_update_xlt()
739 mlx5r_umr_set_update_xlt_data_seg(&wqe.data_seg, &sg); in mlx5r_umr_update_xlt()
754 mlx5r_umr_final_update_xlt(dev, &wqe, mr, &sg, flags); in mlx5r_umr_update_xlt()
755 mlx5r_umr_update_offset(&wqe.ctrl_seg, idx * desc_size); in mlx5r_umr_update_xlt()
756 err = mlx5r_umr_post_send_wait(dev, mr->mmkey.key, &wqe, true); in mlx5r_umr_update_xlt()