/linux-6.1.9/drivers/infiniband/hw/bnxt_re/ |
D | qplib_res.c | 63 static void __free_pbl(struct bnxt_qplib_res *res, struct bnxt_qplib_pbl *pbl, in __free_pbl() argument 70 for (i = 0; i < pbl->pg_count; i++) { in __free_pbl() 71 if (pbl->pg_arr[i]) in __free_pbl() 72 dma_free_coherent(&pdev->dev, pbl->pg_size, in __free_pbl() 74 pbl->pg_arr[i] & in __free_pbl() 76 pbl->pg_map_arr[i]); in __free_pbl() 80 pbl->pg_arr[i] = NULL; in __free_pbl() 83 vfree(pbl->pg_arr); in __free_pbl() 84 pbl->pg_arr = NULL; in __free_pbl() 85 vfree(pbl->pg_map_arr); in __free_pbl() [all …]
|
D | qplib_res.h | 145 struct bnxt_qplib_pbl pbl[PBL_LVL_MAX + 1]; member 283 struct bnxt_qplib_pbl *pbl; in bnxt_qplib_base_pg_size() local 285 pbl = &hwq->pbl[PBL_LVL_0]; in bnxt_qplib_base_pg_size() 286 switch (pbl->pg_size) { in bnxt_qplib_base_pg_size()
|
D | qplib_fp.c | 609 struct bnxt_qplib_pbl *pbl; in bnxt_qplib_create_srq() local 637 pbl = &srq->hwq.pbl[PBL_LVL_0]; in bnxt_qplib_create_srq() 643 req.pbl = cpu_to_le64(pbl->pg_map_arr[0]); in bnxt_qplib_create_srq() 818 struct bnxt_qplib_pbl *pbl; in bnxt_qplib_create_qp1() local 847 pbl = &sq->hwq.pbl[PBL_LVL_0]; in bnxt_qplib_create_qp1() 848 req.sq_pbl = cpu_to_le64(pbl->pg_map_arr[0]); in bnxt_qplib_create_qp1() 872 pbl = &rq->hwq.pbl[PBL_LVL_0]; in bnxt_qplib_create_qp1() 873 req.rq_pbl = cpu_to_le64(pbl->pg_map_arr[0]); in bnxt_qplib_create_qp1() 960 struct bnxt_qplib_pbl *pbl; in bnxt_qplib_create_qp() local 1001 pbl = &sq->hwq.pbl[PBL_LVL_0]; in bnxt_qplib_create_qp() [all …]
|
D | qplib_rcfw.c | 522 cpu_to_le64(ctx->qpc_tbl.pbl[PBL_LVL_0].pg_map_arr[0]); in bnxt_qplib_init_rcfw() 524 cpu_to_le64(ctx->mrw_tbl.pbl[PBL_LVL_0].pg_map_arr[0]); in bnxt_qplib_init_rcfw() 526 cpu_to_le64(ctx->srqc_tbl.pbl[PBL_LVL_0].pg_map_arr[0]); in bnxt_qplib_init_rcfw() 528 cpu_to_le64(ctx->cq_tbl.pbl[PBL_LVL_0].pg_map_arr[0]); in bnxt_qplib_init_rcfw() 530 cpu_to_le64(ctx->tim_tbl.pbl[PBL_LVL_0].pg_map_arr[0]); in bnxt_qplib_init_rcfw() 532 cpu_to_le64(ctx->tqm_ctx.pde.pbl[PBL_LVL_0].pg_map_arr[0]); in bnxt_qplib_init_rcfw() 785 init.cmdq_pbl = cpu_to_le64(cmdq->hwq.pbl[PBL_LVL_0].pg_map_arr[0]); in bnxt_qplib_start_rcfw()
|
/linux-6.1.9/drivers/media/platform/amphion/ |
D | vpu_helpers.c | 121 u32 stride, u32 interlaced, u32 *pbl) in get_nv12_plane_size() argument 127 if (pbl) in get_nv12_plane_size() 128 bytesperline = max(bytesperline, *pbl); in get_nv12_plane_size() 134 if (pbl) in get_nv12_plane_size() 135 *pbl = bytesperline; in get_nv12_plane_size() 141 u32 stride, u32 interlaced, u32 *pbl) in get_tiled_8l128_plane_size() argument 156 if (pbl) in get_tiled_8l128_plane_size() 157 bytesperline = max(bytesperline, *pbl); in get_tiled_8l128_plane_size() 163 if (pbl) in get_tiled_8l128_plane_size() 164 *pbl = bytesperline; in get_tiled_8l128_plane_size() [all …]
|
/linux-6.1.9/drivers/infiniband/hw/efa/ |
D | efa_verbs.c | 1258 static int pbl_chunk_list_create(struct efa_dev *dev, struct pbl_context *pbl) in pbl_chunk_list_create() argument 1260 struct pbl_chunk_list *chunk_list = &pbl->phys.indirect.chunk_list; in pbl_chunk_list_create() 1261 int page_cnt = pbl->phys.indirect.pbl_buf_size_in_pages; in pbl_chunk_list_create() 1262 struct scatterlist *pages_sgl = pbl->phys.indirect.sgl; in pbl_chunk_list_create() 1264 int sg_dma_cnt = pbl->phys.indirect.sg_dma_cnt; in pbl_chunk_list_create() 1358 static void pbl_chunk_list_destroy(struct efa_dev *dev, struct pbl_context *pbl) in pbl_chunk_list_destroy() argument 1360 struct pbl_chunk_list *chunk_list = &pbl->phys.indirect.chunk_list; in pbl_chunk_list_destroy() 1374 struct pbl_context *pbl) in pbl_continuous_initialize() argument 1378 dma_addr = dma_map_single(&dev->pdev->dev, pbl->pbl_buf, in pbl_continuous_initialize() 1379 pbl->pbl_buf_size_in_bytes, DMA_TO_DEVICE); in pbl_continuous_initialize() [all …]
|
D | efa_com_cmd.c | 244 memcpy(mr_cmd.pbl.inline_pbl_array, in efa_com_register_mr() 245 params->pbl.inline_pbl_array, in efa_com_register_mr() 246 sizeof(mr_cmd.pbl.inline_pbl_array)); in efa_com_register_mr() 248 mr_cmd.pbl.pbl.length = params->pbl.pbl.length; in efa_com_register_mr() 249 mr_cmd.pbl.pbl.address.mem_addr_low = in efa_com_register_mr() 250 params->pbl.pbl.address.mem_addr_low; in efa_com_register_mr() 251 mr_cmd.pbl.pbl.address.mem_addr_high = in efa_com_register_mr() 252 params->pbl.pbl.address.mem_addr_high; in efa_com_register_mr()
|
D | efa_com_cmd.h | 184 struct efa_com_ctrl_buff_info pbl; member 185 } pbl; member
|
/linux-6.1.9/include/linux/qed/ |
D | qed_chain.h | 97 } pbl; member 311 *p_next_elem = p_chain->pbl.pp_addr_tbl[page_index].virt_addr; in qed_chain_advance_page() 377 p_prod_page_idx = &p_chain->pbl.c.u16.prod_page_idx; in qed_chain_produce() 386 p_prod_page_idx = &p_chain->pbl.c.u32.prod_page_idx; in qed_chain_produce() 448 p_cons_page_idx = &p_chain->pbl.c.u16.cons_page_idx; in qed_chain_consume() 457 p_cons_page_idx = &p_chain->pbl.c.u32.cons_page_idx; in qed_chain_consume() 501 p_chain->pbl.c.u16.prod_page_idx = (u16)reset_val; in qed_chain_reset() 502 p_chain->pbl.c.u16.cons_page_idx = (u16)reset_val; in qed_chain_reset() 504 p_chain->pbl.c.u32.prod_page_idx = reset_val; in qed_chain_reset() 505 p_chain->pbl.c.u32.cons_page_idx = reset_val; in qed_chain_reset() [all …]
|
/linux-6.1.9/drivers/infiniband/sw/siw/ |
D | siw_mem.c | 150 kfree(mem->pbl); in siw_free_mem() 325 dma_addr_t siw_pbl_get_buffer(struct siw_pbl *pbl, u64 off, int *len, int *idx) in siw_pbl_get_buffer() argument 329 while (i < pbl->num_buf) { in siw_pbl_get_buffer() 330 struct siw_pble *pble = &pbl->pbe[i]; in siw_pbl_get_buffer() 351 struct siw_pbl *pbl; in siw_pbl_alloc() local 356 pbl = kzalloc(struct_size(pbl, pbe, num_buf), GFP_KERNEL); in siw_pbl_alloc() 357 if (!pbl) in siw_pbl_alloc() 360 pbl->max_buf = num_buf; in siw_pbl_alloc() 362 return pbl; in siw_pbl_alloc()
|
D | siw_verbs.c | 1417 struct siw_pbl *pbl = NULL; in siw_alloc_mr() local 1435 pbl = siw_pbl_alloc(max_sge); in siw_alloc_mr() 1436 if (IS_ERR(pbl)) { in siw_alloc_mr() 1437 rv = PTR_ERR(pbl); in siw_alloc_mr() 1439 pbl = NULL; in siw_alloc_mr() 1447 rv = siw_mr_add_mem(mr, pd, pbl, 0, max_sge * PAGE_SIZE, 0); in siw_alloc_mr() 1461 kfree(pbl); in siw_alloc_mr() 1484 struct siw_pbl *pbl = mem->pbl; in siw_map_mr_sg() local 1489 if (!pbl) { in siw_map_mr_sg() 1493 pble = pbl->pbe; in siw_map_mr_sg() [all …]
|
/linux-6.1.9/drivers/infiniband/hw/qedr/ |
D | verbs.c | 507 struct qedr_pbl_info *pbl_info, struct qedr_pbl *pbl) in qedr_free_pbl() argument 513 if (!pbl[i].va) in qedr_free_pbl() 516 pbl[i].va, pbl[i].pa); in qedr_free_pbl() 519 kfree(pbl); in qedr_free_pbl() 622 struct qedr_pbl *pbl, in qedr_populate_pbls() argument 637 pbl_tbl = &pbl[1]; in qedr_populate_pbls() 639 pbl_tbl = pbl; in qedr_populate_pbls() 729 cq->latest_cqe = qed_chain_consume(&cq->pbl); in consume_cqe() 980 rc = dev->ops->common->chain_alloc(dev->cdev, &cq->pbl, in qedr_create_cq() 985 page_cnt = qed_chain_get_page_cnt(&cq->pbl); in qedr_create_cq() [all …]
|
D | qedr.h | 75 struct qed_chain pbl; member 291 struct qed_chain pbl; member 320 struct qed_chain pbl; member 342 qed_chain_get_capacity(p_info->pbl) \ 348 struct qed_chain pbl; member
|
D | main.c | 316 dev->ops->common->chain_free(dev->cdev, &dev->cnq_array[i].pbl); in qedr_free_resources() 382 rc = dev->ops->common->chain_alloc(dev->cdev, &cnq->pbl, in qedr_alloc_resources() 395 i, qed_chain_get_cons_idx(&cnq->pbl)); in qedr_alloc_resources() 403 dev->ops->common->chain_free(dev->cdev, &dev->cnq_array[i].pbl); in qedr_alloc_resources() 447 sw_comp_cons = qed_chain_get_cons_idx(&cnq->pbl); in qedr_irq_handler() 453 cq_handle = (struct regpair *)qed_chain_consume(&cnq->pbl); in qedr_irq_handler() 487 sw_comp_cons = qed_chain_get_cons_idx(&cnq->pbl); in qedr_irq_handler() 802 page_cnt = qed_chain_get_page_cnt(&dev->cnq_array[i].pbl); in qedr_init_hw() 805 p_phys_table = qed_chain_get_pbl_phys(&dev->cnq_array[i].pbl); in qedr_init_hw()
|
/linux-6.1.9/drivers/infiniband/hw/cxgb4/ |
D | resource.c | 271 rdev->stats.pbl.cur += roundup(size, 1 << MIN_PBL_SHIFT); in c4iw_pblpool_alloc() 272 if (rdev->stats.pbl.cur > rdev->stats.pbl.max) in c4iw_pblpool_alloc() 273 rdev->stats.pbl.max = rdev->stats.pbl.cur; in c4iw_pblpool_alloc() 276 rdev->stats.pbl.fail++; in c4iw_pblpool_alloc() 294 rdev->stats.pbl.cur -= roundup(size, 1 << MIN_PBL_SHIFT); in c4iw_pblpool_free() 308 pbl_start = rdev->lldi.vr->pbl.start; in c4iw_pblpool_create() 309 pbl_chunk = rdev->lldi.vr->pbl.size; in c4iw_pblpool_create()
|
/linux-6.1.9/drivers/scsi/bnx2fc/ |
D | bnx2fc_tgt.c | 667 u32 *pbl; in bnx2fc_alloc_session_resc() local 722 pbl = (u32 *)tgt->rq_pbl; in bnx2fc_alloc_session_resc() 725 *pbl = (u32)page; in bnx2fc_alloc_session_resc() 726 pbl++; in bnx2fc_alloc_session_resc() 727 *pbl = (u32)((u64)page >> 32); in bnx2fc_alloc_session_resc() 728 pbl++; in bnx2fc_alloc_session_resc() 776 pbl = (u32 *)tgt->confq_pbl; in bnx2fc_alloc_session_resc() 779 *pbl = (u32)page; in bnx2fc_alloc_session_resc() 780 pbl++; in bnx2fc_alloc_session_resc() 781 *pbl = (u32)((u64)page >> 32); in bnx2fc_alloc_session_resc() [all …]
|
D | bnx2fc_hwi.c | 1987 u32 *pbl; in bnx2fc_free_hash_table() local 1991 pbl = hba->hash_tbl_pbl; in bnx2fc_free_hash_table() 1992 if (pbl) { in bnx2fc_free_hash_table() 1997 dma_address = le32_to_cpu(*pbl); in bnx2fc_free_hash_table() 1998 ++pbl; in bnx2fc_free_hash_table() 1999 dma_address += ((u64)le32_to_cpu(*pbl)) << 32; in bnx2fc_free_hash_table() 2000 ++pbl; in bnx2fc_free_hash_table() 2028 u32 *pbl; in bnx2fc_allocate_hash_table() local 2069 pbl = hba->hash_tbl_pbl; in bnx2fc_allocate_hash_table() 2072 *pbl = cpu_to_le32((u32) paddr); in bnx2fc_allocate_hash_table() [all …]
|
/linux-6.1.9/drivers/net/ethernet/qlogic/qed/ |
D | qed_chain.c | 111 if (!chain->pbl.pp_addr_tbl) in qed_chain_free_pbl() 115 entry = chain->pbl.pp_addr_tbl + i; in qed_chain_free_pbl() 128 vfree(chain->pbl.pp_addr_tbl); in qed_chain_free_pbl() 129 chain->pbl.pp_addr_tbl = NULL; in qed_chain_free_pbl() 269 chain->pbl.pp_addr_tbl = addr_tbl; in qed_chain_alloc_pbl()
|
/linux-6.1.9/drivers/scsi/qedi/ |
D | qedi_main.c | 1539 struct scsi_bd *pbl; in qedi_alloc_bdq() local 1576 pbl = (struct scsi_bd *)qedi->bdq_pbl; in qedi_alloc_bdq() 1578 pbl->address.hi = in qedi_alloc_bdq() 1580 pbl->address.lo = in qedi_alloc_bdq() 1584 pbl, pbl->address.hi, pbl->address.lo, i); in qedi_alloc_bdq() 1585 pbl->opaque.iscsi_opaque.reserved_zero[0] = 0; in qedi_alloc_bdq() 1586 pbl->opaque.iscsi_opaque.reserved_zero[1] = 0; in qedi_alloc_bdq() 1587 pbl->opaque.iscsi_opaque.reserved_zero[2] = 0; in qedi_alloc_bdq() 1588 pbl->opaque.iscsi_opaque.opaque = cpu_to_le16(i); in qedi_alloc_bdq() 1589 pbl++; in qedi_alloc_bdq() [all …]
|
/linux-6.1.9/Documentation/devicetree/bindings/net/ |
D | samsung-sxgbe.txt | 15 - samsung,pbl: Integer, Programmable Burst Length. 47 samsung,pbl = <0x08>
|
D | stm32-dwmac.yaml | 128 snps,pbl = <2>; 145 snps,pbl = <8>; 161 snps,pbl = <8>;
|
/linux-6.1.9/drivers/net/ethernet/samsung/sxgbe/ |
D | sxgbe_dma.c | 44 int fix_burst, int pbl, dma_addr_t dma_tx, in sxgbe_dma_channel_init() argument 57 reg_val |= (pbl << SXGBE_DMA_TXPBL_LSHIFT); in sxgbe_dma_channel_init() 61 reg_val |= (pbl << SXGBE_DMA_RXPBL_LSHIFT); in sxgbe_dma_channel_init()
|
/linux-6.1.9/include/linux/ |
D | sxgbe_platform.h | 33 int pbl; member
|
/linux-6.1.9/drivers/net/ethernet/tehuti/ |
D | tehuti.h | 315 struct pbl { struct 333 struct pbl pbl[]; /* Fragments */ member
|
D | tehuti.c | 1472 struct pbl *pbl = &txdd->pbl[0]; in bdx_tx_map_skb() local 1479 pbl->len = CPU_CHIP_SWAP32(db->wptr->len); in bdx_tx_map_skb() 1480 pbl->pa_lo = CPU_CHIP_SWAP32(L32_64(db->wptr->addr.dma)); in bdx_tx_map_skb() 1481 pbl->pa_hi = CPU_CHIP_SWAP32(H32_64(db->wptr->addr.dma)); in bdx_tx_map_skb() 1482 DBG("=== pbl len: 0x%x ================\n", pbl->len); in bdx_tx_map_skb() 1483 DBG("=== pbl pa_lo: 0x%x ================\n", pbl->pa_lo); in bdx_tx_map_skb() 1484 DBG("=== pbl pa_hi: 0x%x ================\n", pbl->pa_hi); in bdx_tx_map_skb() 1496 pbl++; in bdx_tx_map_skb() 1497 pbl->len = CPU_CHIP_SWAP32(db->wptr->len); in bdx_tx_map_skb() 1498 pbl->pa_lo = CPU_CHIP_SWAP32(L32_64(db->wptr->addr.dma)); in bdx_tx_map_skb() [all …]
|