/linux-6.6.21/net/sunrpc/xprtrdma/ |
D | rpc_rdma.c | 399 int nsegs, nchunks; in rpcrdma_encode_write_list() local 419 nchunks = 0; in rpcrdma_encode_write_list() 431 nchunks++; in rpcrdma_encode_write_list() 443 nchunks++; in rpcrdma_encode_write_list() 448 *segcount = cpu_to_be32(nchunks); in rpcrdma_encode_write_list() 476 int nsegs, nchunks; in rpcrdma_encode_reply_chunk() local 497 nchunks = 0; in rpcrdma_encode_reply_chunk() 509 nchunks++; in rpcrdma_encode_reply_chunk() 514 *segcount = cpu_to_be32(nchunks); in rpcrdma_encode_reply_chunk()
|
/linux-6.6.21/drivers/dma/sh/ |
D | rcar-dmac.c | 81 unsigned int nchunks; member 394 chan->index, desc, desc->nchunks, &desc->hwdescs.dma); in rcar_dmac_chan_start_xfer() 408 RCAR_DMACHCRB_DCNT(desc->nchunks - 1) | in rcar_dmac_chan_start_xfer() 765 rcar_dmac_realloc_hwdesc(chan, desc, desc->nchunks * sizeof(*hwdesc)); in rcar_dmac_fill_hwdesc() 936 unsigned int nchunks = 0; in rcar_dmac_chan_prep_sg() local 1026 nchunks++; in rcar_dmac_chan_prep_sg() 1030 desc->nchunks = nchunks; in rcar_dmac_chan_prep_sg() 1041 desc->hwdescs.use = !cross_boundary && nchunks > 1; in rcar_dmac_chan_prep_sg() 1401 dptr = desc->nchunks; in rcar_dmac_chan_get_residue() 1403 WARN_ON(dptr >= desc->nchunks); in rcar_dmac_chan_get_residue() [all …]
|
/linux-6.6.21/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_cs.h | 55 unsigned nchunks; member
|
D | amdgpu_cs.c | 201 p->nchunks = cs->in.num_chunks; in amdgpu_cs_pass1() 202 p->chunks = kvmalloc_array(p->nchunks, sizeof(struct amdgpu_cs_chunk), in amdgpu_cs_pass1() 209 for (i = 0; i < p->nchunks; i++) { in amdgpu_cs_pass1() 313 i = p->nchunks - 1; in amdgpu_cs_pass1() 319 p->nchunks = 0; in amdgpu_cs_pass1() 593 for (i = 0; i < p->nchunks; ++i) { in amdgpu_cs_pass2() 1372 for (i = 0; i < parser->nchunks; i++) in amdgpu_cs_parser_fini()
|
/linux-6.6.21/drivers/infiniband/hw/vmw_pvrdma/ |
D | pvrdma_dev_api.h | 462 u32 nchunks; member 484 u32 nchunks; member 516 u32 nchunks; member
|
D | pvrdma_mr.c | 171 cmd->nchunks = npages; in pvrdma_reg_user_mr() 243 cmd->nchunks = max_num_sg; in pvrdma_alloc_mr()
|
D | pvrdma_srq.c | 176 cmd->nchunks = srq->npages; in pvrdma_create_srq()
|
D | pvrdma_cq.c | 184 cmd->nchunks = npages; in pvrdma_create_cq()
|
/linux-6.6.21/drivers/crypto/cavium/cpt/ |
D | cptvf.h | 54 u32 nchunks; /* Number of command chunks */ member
|
D | cptvf_main.c | 192 queue->nchunks = 0; in free_command_queues() 239 i, queue->nchunks); in alloc_command_queues() 245 if (queue->nchunks == 0) { in alloc_command_queues() 254 queue->nchunks++; in alloc_command_queues()
|
/linux-6.6.21/net/sctp/ |
D | auth.c | 759 __u16 nchunks; in sctp_auth_ep_add_chunkid() local 768 nchunks = param_len - sizeof(struct sctp_paramhdr); in sctp_auth_ep_add_chunkid() 769 if (nchunks == SCTP_NUM_CHUNK_TYPES) in sctp_auth_ep_add_chunkid() 772 p->chunks[nchunks] = chunk_id; in sctp_auth_ep_add_chunkid()
|
/linux-6.6.21/drivers/mtd/nand/raw/ |
D | sunxi_nand.c | 350 int chunksize, int nchunks, in sunxi_nfc_dma_op_prepare() argument 364 sg_init_one(sg, buf, nchunks * chunksize); in sunxi_nfc_dma_op_prepare() 379 writel(nchunks, nfc->regs + NFC_REG_SECTOR_NUM); in sunxi_nfc_dma_op_prepare() 385 writel(chunksize * nchunks, nfc->regs + NFC_REG_MDMA_CNT); in sunxi_nfc_dma_op_prepare() 904 int nchunks) in sunxi_nfc_hw_ecc_read_chunks_dma() argument 919 ret = sunxi_nfc_dma_op_prepare(nfc, buf, ecc->size, nchunks, in sunxi_nfc_hw_ecc_read_chunks_dma() 955 for (i = 0; i < nchunks; i++) { in sunxi_nfc_hw_ecc_read_chunks_dma() 987 for (i = 0; i < nchunks; i++) { in sunxi_nfc_hw_ecc_read_chunks_dma() 1189 int nchunks = DIV_ROUND_UP(data_offs + readlen, nand->ecc.size); in sunxi_nfc_hw_ecc_read_subpage_dma() local 1196 ret = sunxi_nfc_hw_ecc_read_chunks_dma(nand, buf, false, page, nchunks); in sunxi_nfc_hw_ecc_read_subpage_dma()
|
D | marvell_nand.c | 261 int nchunks; member 276 .nchunks = nc, \ 1277 for (chunk = 0; chunk < lt->nchunks; chunk++) { in marvell_nfc_hw_ecc_bch_read_page_raw() 1336 else if (chunk < lt->nchunks - 1) in marvell_nfc_hw_ecc_bch_read_chunk() 1394 for (chunk = 0; chunk < lt->nchunks; chunk++) { in marvell_nfc_hw_ecc_bch_read_page() 1436 for (chunk = 0; chunk < lt->nchunks; chunk++) { in marvell_nfc_hw_ecc_bch_read_page() 1529 for (chunk = 0; chunk < lt->nchunks; chunk++) { in marvell_nfc_hw_ecc_bch_write_page_raw() 1589 if (lt->nchunks == 1) in marvell_nfc_hw_ecc_bch_write_chunk() 1599 } else if (chunk < lt->nchunks - 1) { in marvell_nfc_hw_ecc_bch_write_chunk() 1606 if (chunk == lt->nchunks - 1) in marvell_nfc_hw_ecc_bch_write_chunk() [all …]
|
D | renesas-nand-controller.c | 498 unsigned int nchunks = real_len / chip->ecc.size; in rnandc_read_subpage_hw_ecc() local 549 for (i = start_chunk; i < nchunks; i++) { in rnandc_read_subpage_hw_ecc()
|
/linux-6.6.21/drivers/gpu/drm/radeon/ |
D | radeon_cs.c | 302 p->nchunks = cs->num_chunks; in radeon_cs_parser_init() 303 p->chunks = kvcalloc(p->nchunks, sizeof(struct radeon_cs_chunk), GFP_KERNEL); in radeon_cs_parser_init() 307 for (i = 0; i < p->nchunks; i++) { in radeon_cs_parser_init() 457 for (i = 0; i < parser->nchunks; i++) in radeon_cs_parser_fini()
|
D | radeon.h | 1023 unsigned nchunks; member
|
/linux-6.6.21/drivers/vhost/ |
D | vdpa.c | 991 unsigned long lock_limit, sz2pin, nchunks, i; in vhost_vdpa_pa_map() local 1020 nchunks = 0; in vhost_vdpa_pa_map() 1035 nchunks++; in vhost_vdpa_pa_map() 1066 nchunks = 0; in vhost_vdpa_pa_map() 1081 if (nchunks) { in vhost_vdpa_pa_map()
|
/linux-6.6.21/sound/drivers/vx/ |
D | vx_pcm.c | 677 struct vx_pipe *pipe, int nchunks) in vx_pcm_playback_transfer() argument 684 for (i = 0; i < nchunks; i++) { in vx_pcm_playback_transfer()
|