Home
last modified time | relevance | path

Searched refs:nchunks (Results 1 – 18 of 18) sorted by relevance

/linux-6.6.21/net/sunrpc/xprtrdma/
Drpc_rdma.c399 int nsegs, nchunks; in rpcrdma_encode_write_list() local
419 nchunks = 0; in rpcrdma_encode_write_list()
431 nchunks++; in rpcrdma_encode_write_list()
443 nchunks++; in rpcrdma_encode_write_list()
448 *segcount = cpu_to_be32(nchunks); in rpcrdma_encode_write_list()
476 int nsegs, nchunks; in rpcrdma_encode_reply_chunk() local
497 nchunks = 0; in rpcrdma_encode_reply_chunk()
509 nchunks++; in rpcrdma_encode_reply_chunk()
514 *segcount = cpu_to_be32(nchunks); in rpcrdma_encode_reply_chunk()
/linux-6.6.21/drivers/dma/sh/
Drcar-dmac.c81 unsigned int nchunks; member
394 chan->index, desc, desc->nchunks, &desc->hwdescs.dma); in rcar_dmac_chan_start_xfer()
408 RCAR_DMACHCRB_DCNT(desc->nchunks - 1) | in rcar_dmac_chan_start_xfer()
765 rcar_dmac_realloc_hwdesc(chan, desc, desc->nchunks * sizeof(*hwdesc)); in rcar_dmac_fill_hwdesc()
936 unsigned int nchunks = 0; in rcar_dmac_chan_prep_sg() local
1026 nchunks++; in rcar_dmac_chan_prep_sg()
1030 desc->nchunks = nchunks; in rcar_dmac_chan_prep_sg()
1041 desc->hwdescs.use = !cross_boundary && nchunks > 1; in rcar_dmac_chan_prep_sg()
1401 dptr = desc->nchunks; in rcar_dmac_chan_get_residue()
1403 WARN_ON(dptr >= desc->nchunks); in rcar_dmac_chan_get_residue()
[all …]
/linux-6.6.21/drivers/gpu/drm/amd/amdgpu/
Damdgpu_cs.h55 unsigned nchunks; member
Damdgpu_cs.c201 p->nchunks = cs->in.num_chunks; in amdgpu_cs_pass1()
202 p->chunks = kvmalloc_array(p->nchunks, sizeof(struct amdgpu_cs_chunk), in amdgpu_cs_pass1()
209 for (i = 0; i < p->nchunks; i++) { in amdgpu_cs_pass1()
313 i = p->nchunks - 1; in amdgpu_cs_pass1()
319 p->nchunks = 0; in amdgpu_cs_pass1()
593 for (i = 0; i < p->nchunks; ++i) { in amdgpu_cs_pass2()
1372 for (i = 0; i < parser->nchunks; i++) in amdgpu_cs_parser_fini()
/linux-6.6.21/drivers/infiniband/hw/vmw_pvrdma/
Dpvrdma_dev_api.h462 u32 nchunks; member
484 u32 nchunks; member
516 u32 nchunks; member
Dpvrdma_mr.c171 cmd->nchunks = npages; in pvrdma_reg_user_mr()
243 cmd->nchunks = max_num_sg; in pvrdma_alloc_mr()
Dpvrdma_srq.c176 cmd->nchunks = srq->npages; in pvrdma_create_srq()
Dpvrdma_cq.c184 cmd->nchunks = npages; in pvrdma_create_cq()
/linux-6.6.21/drivers/crypto/cavium/cpt/
Dcptvf.h54 u32 nchunks; /* Number of command chunks */ member
Dcptvf_main.c192 queue->nchunks = 0; in free_command_queues()
239 i, queue->nchunks); in alloc_command_queues()
245 if (queue->nchunks == 0) { in alloc_command_queues()
254 queue->nchunks++; in alloc_command_queues()
/linux-6.6.21/net/sctp/
Dauth.c759 __u16 nchunks; in sctp_auth_ep_add_chunkid() local
768 nchunks = param_len - sizeof(struct sctp_paramhdr); in sctp_auth_ep_add_chunkid()
769 if (nchunks == SCTP_NUM_CHUNK_TYPES) in sctp_auth_ep_add_chunkid()
772 p->chunks[nchunks] = chunk_id; in sctp_auth_ep_add_chunkid()
/linux-6.6.21/drivers/mtd/nand/raw/
Dsunxi_nand.c350 int chunksize, int nchunks, in sunxi_nfc_dma_op_prepare() argument
364 sg_init_one(sg, buf, nchunks * chunksize); in sunxi_nfc_dma_op_prepare()
379 writel(nchunks, nfc->regs + NFC_REG_SECTOR_NUM); in sunxi_nfc_dma_op_prepare()
385 writel(chunksize * nchunks, nfc->regs + NFC_REG_MDMA_CNT); in sunxi_nfc_dma_op_prepare()
904 int nchunks) in sunxi_nfc_hw_ecc_read_chunks_dma() argument
919 ret = sunxi_nfc_dma_op_prepare(nfc, buf, ecc->size, nchunks, in sunxi_nfc_hw_ecc_read_chunks_dma()
955 for (i = 0; i < nchunks; i++) { in sunxi_nfc_hw_ecc_read_chunks_dma()
987 for (i = 0; i < nchunks; i++) { in sunxi_nfc_hw_ecc_read_chunks_dma()
1189 int nchunks = DIV_ROUND_UP(data_offs + readlen, nand->ecc.size); in sunxi_nfc_hw_ecc_read_subpage_dma() local
1196 ret = sunxi_nfc_hw_ecc_read_chunks_dma(nand, buf, false, page, nchunks); in sunxi_nfc_hw_ecc_read_subpage_dma()
Dmarvell_nand.c261 int nchunks; member
276 .nchunks = nc, \
1277 for (chunk = 0; chunk < lt->nchunks; chunk++) { in marvell_nfc_hw_ecc_bch_read_page_raw()
1336 else if (chunk < lt->nchunks - 1) in marvell_nfc_hw_ecc_bch_read_chunk()
1394 for (chunk = 0; chunk < lt->nchunks; chunk++) { in marvell_nfc_hw_ecc_bch_read_page()
1436 for (chunk = 0; chunk < lt->nchunks; chunk++) { in marvell_nfc_hw_ecc_bch_read_page()
1529 for (chunk = 0; chunk < lt->nchunks; chunk++) { in marvell_nfc_hw_ecc_bch_write_page_raw()
1589 if (lt->nchunks == 1) in marvell_nfc_hw_ecc_bch_write_chunk()
1599 } else if (chunk < lt->nchunks - 1) { in marvell_nfc_hw_ecc_bch_write_chunk()
1606 if (chunk == lt->nchunks - 1) in marvell_nfc_hw_ecc_bch_write_chunk()
[all …]
Drenesas-nand-controller.c498 unsigned int nchunks = real_len / chip->ecc.size; in rnandc_read_subpage_hw_ecc() local
549 for (i = start_chunk; i < nchunks; i++) { in rnandc_read_subpage_hw_ecc()
/linux-6.6.21/drivers/gpu/drm/radeon/
Dradeon_cs.c302 p->nchunks = cs->num_chunks; in radeon_cs_parser_init()
303 p->chunks = kvcalloc(p->nchunks, sizeof(struct radeon_cs_chunk), GFP_KERNEL); in radeon_cs_parser_init()
307 for (i = 0; i < p->nchunks; i++) { in radeon_cs_parser_init()
457 for (i = 0; i < parser->nchunks; i++) in radeon_cs_parser_fini()
Dradeon.h1023 unsigned nchunks; member
/linux-6.6.21/drivers/vhost/
Dvdpa.c991 unsigned long lock_limit, sz2pin, nchunks, i; in vhost_vdpa_pa_map() local
1020 nchunks = 0; in vhost_vdpa_pa_map()
1035 nchunks++; in vhost_vdpa_pa_map()
1066 nchunks = 0; in vhost_vdpa_pa_map()
1081 if (nchunks) { in vhost_vdpa_pa_map()
/linux-6.6.21/sound/drivers/vx/
Dvx_pcm.c677 struct vx_pipe *pipe, int nchunks) in vx_pcm_playback_transfer() argument
684 for (i = 0; i < nchunks; i++) { in vx_pcm_playback_transfer()