/linux-6.1.9/drivers/media/dvb-frontends/ |
D | dib3000mb.c | 148 wr(DIB3000MB_REG_LOCK1_MASK, DIB3000MB_LOCK1_SEARCH_4); in dib3000mb_set_frontend() 153 wr(DIB3000MB_REG_FFT, DIB3000_TRANSMISSION_MODE_2K); in dib3000mb_set_frontend() 157 wr(DIB3000MB_REG_FFT, DIB3000_TRANSMISSION_MODE_8K); in dib3000mb_set_frontend() 169 wr(DIB3000MB_REG_GUARD_TIME, DIB3000_GUARD_TIME_1_32); in dib3000mb_set_frontend() 173 wr(DIB3000MB_REG_GUARD_TIME, DIB3000_GUARD_TIME_1_16); in dib3000mb_set_frontend() 177 wr(DIB3000MB_REG_GUARD_TIME, DIB3000_GUARD_TIME_1_8); in dib3000mb_set_frontend() 181 wr(DIB3000MB_REG_GUARD_TIME, DIB3000_GUARD_TIME_1_4); in dib3000mb_set_frontend() 193 wr(DIB3000MB_REG_DDS_INV, DIB3000_DDS_INVERSION_OFF); in dib3000mb_set_frontend() 200 wr(DIB3000MB_REG_DDS_INV, DIB3000_DDS_INVERSION_ON); in dib3000mb_set_frontend() 209 wr(DIB3000MB_REG_QAM, DIB3000_CONSTELLATION_QPSK); in dib3000mb_set_frontend() [all …]
|
/linux-6.1.9/lib/ |
D | decompress_unlzma.c | 292 static inline size_t INIT get_pos(struct writer *wr) in get_pos() argument 295 wr->global_pos + wr->buffer_pos; in get_pos() 298 static inline uint8_t INIT peek_old_byte(struct writer *wr, in peek_old_byte() argument 301 if (!wr->flush) { in peek_old_byte() 303 while (offs > wr->header->dict_size) in peek_old_byte() 304 offs -= wr->header->dict_size; in peek_old_byte() 305 pos = wr->buffer_pos - offs; in peek_old_byte() 306 return wr->buffer[pos]; in peek_old_byte() 308 uint32_t pos = wr->buffer_pos - offs; in peek_old_byte() 309 while (pos >= wr->header->dict_size) in peek_old_byte() [all …]
|
/linux-6.1.9/tools/bpf/bpftool/ |
D | json_writer.c | 314 json_writer_t *wr = jsonw_new(stdout); in main() local 316 jsonw_start_object(wr); in main() 317 jsonw_pretty(wr, true); in main() 318 jsonw_name(wr, "Vyatta"); in main() 319 jsonw_start_object(wr); in main() 320 jsonw_string_field(wr, "url", "http://vyatta.com"); in main() 321 jsonw_uint_field(wr, "downloads", 2000000ul); in main() 322 jsonw_float_field(wr, "stock", 8.16); in main() 324 jsonw_name(wr, "ARGV"); in main() 325 jsonw_start_array(wr); in main() [all …]
|
/linux-6.1.9/include/trace/events/ |
D | ib_mad.h | 24 TP_PROTO(struct ib_mad_send_wr_private *wr, 26 TP_ARGS(wr, qp_info), 55 __entry->dev_index = wr->mad_agent_priv->agent.device->index; 56 __entry->port_num = wr->mad_agent_priv->agent.port_num; 57 __entry->qp_num = wr->mad_agent_priv->qp_info->qp->qp_num; 58 __entry->agent_priv = wr->mad_agent_priv; 59 __entry->wrtid = wr->tid; 60 __entry->max_retries = wr->max_retries; 61 __entry->retries_left = wr->retries_left; 62 __entry->retry = wr->retry; [all …]
|
/linux-6.1.9/drivers/infiniband/hw/mlx5/ |
D | wr.c | 54 static void set_eth_seg(const struct ib_send_wr *wr, struct mlx5_ib_qp *qp, in set_eth_seg() argument 61 if (wr->send_flags & IB_SEND_IP_CSUM) in set_eth_seg() 65 if (wr->opcode == IB_WR_LSO) { in set_eth_seg() 66 struct ib_ud_wr *ud_wr = container_of(wr, struct ib_ud_wr, wr); in set_eth_seg() 103 const struct ib_send_wr *wr) in set_datagram_seg() argument 105 memcpy(&dseg->av, &to_mah(ud_wr(wr)->ah)->av, sizeof(struct mlx5_av)); in set_datagram_seg() 107 cpu_to_be32(ud_wr(wr)->remote_qpn | MLX5_EXTENDED_UD_AV); in set_datagram_seg() 108 dseg->av.key.qkey.qkey = cpu_to_be32(ud_wr(wr)->remote_qkey); in set_datagram_seg() 228 static __be32 send_ieth(const struct ib_send_wr *wr) in send_ieth() argument 230 switch (wr->opcode) { in send_ieth() [all …]
|
D | gsi.c | 51 struct mlx5_ib_gsi_wr *wr; in generate_completions() local 56 wr = &gsi->outstanding_wrs[index % gsi->cap.max_send_wr]; in generate_completions() 58 if (!wr->completed) in generate_completions() 61 WARN_ON_ONCE(mlx5_ib_generate_wc(gsi_cq, &wr->wc)); in generate_completions() 62 wr->completed = false; in generate_completions() 71 struct mlx5_ib_gsi_wr *wr = in handle_single_completion() local 78 wr->completed = true; in handle_single_completion() 79 wr_id = wr->wc.wr_id; in handle_single_completion() 80 wr->wc = *wc; in handle_single_completion() 81 wr->wc.wr_id = wr_id; in handle_single_completion() [all …]
|
D | wr.h | 104 int mlx5_ib_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, 106 int mlx5_ib_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, 110 const struct ib_send_wr *wr, in mlx5_ib_post_send_nodrain() argument 113 return mlx5_ib_post_send(ibqp, wr, bad_wr, false); in mlx5_ib_post_send_nodrain() 117 const struct ib_send_wr *wr, in mlx5_ib_post_send_drain() argument 120 return mlx5_ib_post_send(ibqp, wr, bad_wr, true); in mlx5_ib_post_send_drain() 124 const struct ib_recv_wr *wr, in mlx5_ib_post_recv_nodrain() argument 127 return mlx5_ib_post_recv(ibqp, wr, bad_wr, false); in mlx5_ib_post_recv_nodrain() 131 const struct ib_recv_wr *wr, in mlx5_ib_post_recv_drain() argument 134 return mlx5_ib_post_recv(ibqp, wr, bad_wr, true); in mlx5_ib_post_recv_drain()
|
/linux-6.1.9/fs/orangefs/ |
D | inode.c | 23 struct orangefs_write_range *wr = NULL; in orangefs_writepage_locked() local 34 wr = (struct orangefs_write_range *)page_private(page); in orangefs_writepage_locked() 35 WARN_ON(wr->pos >= len); in orangefs_writepage_locked() 36 off = wr->pos; in orangefs_writepage_locked() 37 if (off + wr->len > len) in orangefs_writepage_locked() 40 wlen = wr->len; in orangefs_writepage_locked() 59 len, wr, NULL, NULL); in orangefs_writepage_locked() 94 struct orangefs_write_range *wrp, wr; in orangefs_writepages_work() local 122 wr.uid = ow->uid; in orangefs_writepages_work() 123 wr.gid = ow->gid; in orangefs_writepages_work() [all …]
|
/linux-6.1.9/drivers/infiniband/core/ |
D | rw.c | 81 reg->inv_wr.next = ®->reg_wr.wr; in rdma_rw_inv_key() 112 reg->reg_wr.wr.opcode = IB_WR_REG_MR; in rdma_rw_init_one_mr() 152 prev->wr.wr.next = ®->inv_wr; in rdma_rw_init_mr_wrs() 154 prev->wr.wr.next = ®->reg_wr.wr; in rdma_rw_init_mr_wrs() 157 reg->reg_wr.wr.next = ®->wr.wr; in rdma_rw_init_mr_wrs() 159 reg->wr.wr.sg_list = ®->sge; in rdma_rw_init_mr_wrs() 160 reg->wr.wr.num_sge = 1; in rdma_rw_init_mr_wrs() 161 reg->wr.remote_addr = remote_addr; in rdma_rw_init_mr_wrs() 162 reg->wr.rkey = rkey; in rdma_rw_init_mr_wrs() 164 reg->wr.wr.opcode = IB_WR_RDMA_WRITE; in rdma_rw_init_mr_wrs() [all …]
|
/linux-6.1.9/drivers/infiniband/ulp/iser/ |
D | iser_memory.c | 257 struct ib_reg_wr *wr = &tx_desc->reg_wr; in iser_reg_sig_mr() local 268 iser_inv_rkey(&tx_desc->inv_wr, mr, cqe, &wr->wr); in iser_reg_sig_mr() 280 memset(wr, 0, sizeof(*wr)); in iser_reg_sig_mr() 281 wr->wr.next = &tx_desc->send_wr; in iser_reg_sig_mr() 282 wr->wr.opcode = IB_WR_REG_MR_INTEGRITY; in iser_reg_sig_mr() 283 wr->wr.wr_cqe = cqe; in iser_reg_sig_mr() 284 wr->wr.num_sge = 0; in iser_reg_sig_mr() 285 wr->wr.send_flags = 0; in iser_reg_sig_mr() 286 wr->mr = mr; in iser_reg_sig_mr() 287 wr->key = mr->rkey; in iser_reg_sig_mr() [all …]
|
/linux-6.1.9/drivers/infiniband/hw/vmw_pvrdma/ |
D | pvrdma_qp.c | 647 const struct ib_reg_wr *wr) in set_reg_seg() argument 649 struct pvrdma_user_mr *mr = to_vmr(wr->mr); in set_reg_seg() 651 wqe_hdr->wr.fast_reg.iova_start = mr->ibmr.iova; in set_reg_seg() 652 wqe_hdr->wr.fast_reg.pl_pdir_dma = mr->pdir.dir_dma; in set_reg_seg() 653 wqe_hdr->wr.fast_reg.page_shift = mr->page_shift; in set_reg_seg() 654 wqe_hdr->wr.fast_reg.page_list_len = mr->npages; in set_reg_seg() 655 wqe_hdr->wr.fast_reg.length = mr->ibmr.length; in set_reg_seg() 656 wqe_hdr->wr.fast_reg.access_flags = wr->access; in set_reg_seg() 657 wqe_hdr->wr.fast_reg.rkey = wr->key; in set_reg_seg() 671 int pvrdma_post_send(struct ib_qp *ibqp, const struct ib_send_wr *wr, in pvrdma_post_send() argument [all …]
|
/linux-6.1.9/tools/testing/selftests/breakpoints/ |
D | breakpoint_test_arm64.c | 33 static void child(int size, int wr) in child() argument 35 volatile uint8_t *addr = &var[32 + wr]; in child() 112 static bool run_test(int wr_size, int wp_size, int wr, int wp) in run_test() argument 125 child(wr_size, wr); in run_test() 204 int wr, wp, size; in main() local 215 for (wr = 0; wr <= 32; wr = wr + size) { in main() 216 for (wp = wr - size; wp <= wr + size; wp = wp + size) { in main() 217 result = run_test(size, MIN(size, 8), wr, wp); in main() 218 if ((result && wr == wp) || in main() 219 (!result && wr != wp)) in main() [all …]
|
/linux-6.1.9/include/linux/ |
D | hdlcdrv.h | 28 unsigned rd, wr; member 35 unsigned int wr; member 49 buf->buffer[buf->wr] = buf->shreg; in hdlcdrv_add_bitbuffer() 50 buf->wr = (buf->wr+1) % sizeof(buf->buffer); in hdlcdrv_add_bitbuffer() 58 buf->buffer[buf->wr] = bits & 0xff; in hdlcdrv_add_bitbuffer_word() 59 buf->wr = (buf->wr+1) % sizeof(buf->buffer); in hdlcdrv_add_bitbuffer_word() 60 buf->buffer[buf->wr] = (bits >> 8) & 0xff; in hdlcdrv_add_bitbuffer_word() 61 buf->wr = (buf->wr+1) % sizeof(buf->buffer); in hdlcdrv_add_bitbuffer_word() 164 ret = !((HDLCDRV_HDLCBUFFER - 1 + hb->rd - hb->wr) % HDLCDRV_HDLCBUFFER); in hdlcdrv_hbuf_full() 177 ret = (hb->rd == hb->wr); in hdlcdrv_hbuf_empty() [all …]
|
/linux-6.1.9/drivers/infiniband/sw/rxe/ |
D | rxe_mw.c | 83 if (unlikely(!mr || wqe->wr.wr.mw.length == 0)) { in rxe_check_bind_mw() 117 if (unlikely(wqe->wr.wr.mw.length > mr->ibmr.length)) { in rxe_check_bind_mw() 123 if (unlikely((wqe->wr.wr.mw.addr < mr->ibmr.iova) || in rxe_check_bind_mw() 124 ((wqe->wr.wr.mw.addr + wqe->wr.wr.mw.length) > in rxe_check_bind_mw() 138 u32 key = wqe->wr.wr.mw.rkey & 0xff; in rxe_do_bind_mw() 141 mw->access = wqe->wr.wr.mw.access; in rxe_do_bind_mw() 143 mw->addr = wqe->wr.wr.mw.addr; in rxe_do_bind_mw() 144 mw->length = wqe->wr.wr.mw.length; in rxe_do_bind_mw() 170 u32 mw_rkey = wqe->wr.wr.mw.mw_rkey; in rxe_bind_mw() 171 u32 mr_lkey = wqe->wr.wr.mw.mr_lkey; in rxe_bind_mw() [all …]
|
D | rxe_verbs.c | 373 static int rxe_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr, in rxe_post_srq_recv() argument 382 while (wr) { in rxe_post_srq_recv() 383 err = post_one_recv(&srq->rq, wr); in rxe_post_srq_recv() 386 wr = wr->next; in rxe_post_srq_recv() 392 *bad_wr = wr; in rxe_post_srq_recv() 524 static void init_send_wr(struct rxe_qp *qp, struct rxe_send_wr *wr, in init_send_wr() argument 527 wr->wr_id = ibwr->wr_id; in init_send_wr() 528 wr->opcode = ibwr->opcode; in init_send_wr() 529 wr->send_flags = ibwr->send_flags; in init_send_wr() 535 wr->wr.ud.remote_qpn = ud_wr(ibwr)->remote_qpn; in init_send_wr() [all …]
|
/linux-6.1.9/drivers/infiniband/sw/siw/ |
D | siw_verbs.c | 676 static int siw_sq_flush_wr(struct siw_qp *qp, const struct ib_send_wr *wr, in siw_sq_flush_wr() argument 681 while (wr) { in siw_sq_flush_wr() 684 switch (wr->opcode) { in siw_sq_flush_wr() 714 sqe.id = wr->wr_id; in siw_sq_flush_wr() 720 *bad_wr = wr; in siw_sq_flush_wr() 723 wr = wr->next; in siw_sq_flush_wr() 729 static int siw_rq_flush_wr(struct siw_qp *qp, const struct ib_recv_wr *wr, in siw_rq_flush_wr() argument 735 while (wr) { in siw_rq_flush_wr() 736 rqe.id = wr->wr_id; in siw_rq_flush_wr() 740 *bad_wr = wr; in siw_rq_flush_wr() [all …]
|
/linux-6.1.9/drivers/infiniband/hw/mthca/ |
D | mthca_qp.c | 1501 const struct ib_ud_wr *wr, in build_mlx_header() argument 1511 mthca_ah_grh_present(to_mah(wr->ah)), 0, 0, 0, in build_mlx_header() 1514 err = mthca_read_ah(dev, to_mah(wr->ah), &sqp->ud_header); in build_mlx_header() 1525 switch (wr->wr.opcode) { in build_mlx_header() 1533 sqp->ud_header.immediate_data = wr->wr.ex.imm_data; in build_mlx_header() 1542 sqp->ud_header.bth.solicited_event = !!(wr->wr.send_flags & IB_SEND_SOLICITED); in build_mlx_header() 1547 ib_get_cached_pkey(&dev->ib_dev, qp->port, wr->pkey_index, in build_mlx_header() 1550 sqp->ud_header.bth.destination_qpn = cpu_to_be32(wr->remote_qpn); in build_mlx_header() 1552 sqp->ud_header.deth.qkey = cpu_to_be32(wr->remote_qkey & 0x80000000 ? in build_mlx_header() 1553 sqp->qkey : wr->remote_qkey); in build_mlx_header() [all …]
|
/linux-6.1.9/arch/mips/mm/ |
D | tlbex.c | 2014 struct work_registers wr = build_get_work_registers(p); in build_r4000_tlbchange_handler_head() local 2017 build_get_pmde64(p, l, r, wr.r1, wr.r2); /* get pmd in ptr */ in build_r4000_tlbchange_handler_head() 2019 build_get_pgde32(p, wr.r1, wr.r2); /* get pgd in ptr */ in build_r4000_tlbchange_handler_head() 2028 build_is_huge_pte(p, r, wr.r1, wr.r2, label_tlb_huge_update); in build_r4000_tlbchange_handler_head() 2031 UASM_i_MFC0(p, wr.r1, C0_BADVADDR); in build_r4000_tlbchange_handler_head() 2032 UASM_i_LW(p, wr.r2, 0, wr.r2); in build_r4000_tlbchange_handler_head() 2033 UASM_i_SRL(p, wr.r1, wr.r1, PAGE_SHIFT - PTE_T_LOG2); in build_r4000_tlbchange_handler_head() 2034 uasm_i_andi(p, wr.r1, wr.r1, (PTRS_PER_PTE - 1) << PTE_T_LOG2); in build_r4000_tlbchange_handler_head() 2035 UASM_i_ADDU(p, wr.r2, wr.r2, wr.r1); in build_r4000_tlbchange_handler_head() 2040 iPTE_LW(p, wr.r1, wr.r2); /* get even pte */ in build_r4000_tlbchange_handler_head() [all …]
|
/linux-6.1.9/drivers/infiniband/hw/cxgb4/ |
D | qp.c | 415 const struct ib_send_wr *wr, int max, u32 *plenp) in build_immd() argument 423 for (i = 0; i < wr->num_sge; i++) { in build_immd() 424 if ((plen + wr->sg_list[i].length) > max) in build_immd() 426 srcp = (u8 *)(unsigned long)wr->sg_list[i].addr; in build_immd() 427 plen += wr->sg_list[i].length; in build_immd() 428 rem = wr->sg_list[i].length; in build_immd() 490 const struct ib_send_wr *wr, u8 *len16) in build_rdma_send() argument 496 if (wr->num_sge > T4_MAX_SEND_SGE) in build_rdma_send() 498 switch (wr->opcode) { in build_rdma_send() 500 if (wr->send_flags & IB_SEND_SOLICITED) in build_rdma_send() [all …]
|
/linux-6.1.9/drivers/infiniband/ulp/rtrs/ |
D | rtrs.c | 73 struct ib_recv_wr wr; in rtrs_iu_post_recv() local 85 wr = (struct ib_recv_wr) { in rtrs_iu_post_recv() 91 return ib_post_recv(con->qp, &wr, NULL); in rtrs_iu_post_recv() 97 struct ib_recv_wr wr; in rtrs_post_recv_empty() local 99 wr = (struct ib_recv_wr) { in rtrs_post_recv_empty() 103 return ib_post_recv(con->qp, &wr, NULL); in rtrs_post_recv_empty() 108 struct ib_send_wr *wr, struct ib_send_wr *tail) in rtrs_post_send() argument 115 next->next = wr; in rtrs_post_send() 117 head = wr; in rtrs_post_send() 121 wr->next = tail; in rtrs_post_send() [all …]
|
D | rtrs-srv.c | 200 struct ib_rdma_wr *wr = NULL; in rdma_write_sg() local 217 wr = &id->tx_wr; in rdma_write_sg() 233 wr->wr.sg_list = plist; in rdma_write_sg() 234 wr->wr.num_sge = 1; in rdma_write_sg() 235 wr->remote_addr = le64_to_cpu(id->rd_msg->desc[0].addr); in rdma_write_sg() 236 wr->rkey = le32_to_cpu(id->rd_msg->desc[0].key); in rdma_write_sg() 238 rkey = wr->rkey; in rdma_write_sg() 241 WARN_ON_ONCE(rkey != wr->rkey); in rdma_write_sg() 243 wr->wr.opcode = IB_WR_RDMA_WRITE; in rdma_write_sg() 244 wr->wr.wr_cqe = &io_comp_cqe; in rdma_write_sg() [all …]
|
/linux-6.1.9/arch/sparc/kernel/ |
D | trampoline_32.S | 47 wr %g1, 0x0, %psr ! traps off though 52 wr %g1, 0x0, %wim 56 wr %g3, 0x0, %tbr 71 wr %g1, PSR_ET, %psr ! traps on 101 wr %g1, 0x0, %psr ! traps off though 106 wr %g1, 0x0, %wim 111 wr %g1, 0x0, %tbr 132 wr %g1, PSR_ET, %psr ! traps on 160 wr %g1, 0x0, %psr ! traps off though 165 wr %g1, 0x0, %wim [all …]
|
D | entry.S | 159 wr %l0, 0x0, %psr 181 wr %l4, 0x0, %psr 183 wr %l4, PSR_ET, %psr 200 wr %l0, PSR_ET, %psr 230 wr %g2, 0x0, %psr 232 wr %g2, PSR_ET, %psr 239 wr %g2, PSR_ET, %psr ! keep ET up 249 wr %g2, 0x0, %psr 251 wr %g2, PSR_ET, %psr 255 wr %l0, PSR_ET, %psr [all …]
|
/linux-6.1.9/arch/arm/boot/dts/ |
D | omap3430-sdp.dts | 65 gpmc,cs-wr-off-ns = <186>; 68 gpmc,adv-wr-off-ns = <48>; 74 gpmc,wr-cycle-ns = <186>; 79 gpmc,wr-data-mux-bus-ns = <90>; 80 gpmc,wr-access-ns = <186>; 115 gpmc,cs-wr-off-ns = <36>; 118 gpmc,adv-wr-off-ns = <36>; 124 gpmc,wr-cycle-ns = <72>; 126 gpmc,wr-access-ns = <30>; 161 gpmc,cs-wr-off-ns = <72>; [all …]
|
/linux-6.1.9/drivers/infiniband/sw/rdmavt/ |
D | qp.c | 627 for (i = 0; i < wqe->wr.num_sge; i++) { in rvt_swqe_has_lkey() 1783 int rvt_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *wr, in rvt_post_recv() argument 1794 *bad_wr = wr; in rvt_post_recv() 1798 for (; wr; wr = wr->next) { in rvt_post_recv() 1803 if ((unsigned)wr->num_sge > qp->r_rq.max_sge) { in rvt_post_recv() 1804 *bad_wr = wr; in rvt_post_recv() 1814 *bad_wr = wr; in rvt_post_recv() 1823 wc.wr_id = wr->wr_id; in rvt_post_recv() 1828 wqe->wr_id = wr->wr_id; in rvt_post_recv() 1829 wqe->num_sge = wr->num_sge; in rvt_post_recv() [all …]
|