/linux-5.19.10/drivers/infiniband/hw/bnxt_re/ |
D | qplib_res.h | 118 u32 pg_size; member 282 u8 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_4K; in bnxt_qplib_base_pg_size() local 286 switch (pbl->pg_size) { in bnxt_qplib_base_pg_size() 288 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_4K; in bnxt_qplib_base_pg_size() 291 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_8K; in bnxt_qplib_base_pg_size() 294 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_64K; in bnxt_qplib_base_pg_size() 297 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_2M; in bnxt_qplib_base_pg_size() 300 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_8M; in bnxt_qplib_base_pg_size() 303 pg_size = BNXT_QPLIB_HWRM_PG_SIZE_1G; in bnxt_qplib_base_pg_size() 309 return pg_size; in bnxt_qplib_base_pg_size()
|
D | qplib_res.c | 72 dma_free_coherent(&pdev->dev, pbl->pg_size, in __free_pbl() 88 pbl->pg_size = 0; in __free_pbl() 132 pbl->pg_size = sginfo->pgsize; in __alloc_pbl() 137 pbl->pg_size, in __alloc_pbl() 186 u32 npages, aux_slots, pg_size, aux_pages = 0, aux_size = 0; in bnxt_qplib_alloc_init_hwq() local 196 pg_size = hwq_attr->sginfo->pgsize; in bnxt_qplib_alloc_init_hwq() 204 aux_pages = (aux_slots * aux_size) / pg_size; in bnxt_qplib_alloc_init_hwq() 205 if ((aux_slots * aux_size) % pg_size) in bnxt_qplib_alloc_init_hwq() 211 npages = (depth * stride) / pg_size + aux_pages; in bnxt_qplib_alloc_init_hwq() 212 if ((depth * stride) % pg_size) in bnxt_qplib_alloc_init_hwq() [all …]
|
D | qplib_sp.c | 577 u32 pg_size; in bnxt_qplib_reg_mr() local 612 pg_size = PAGE_SIZE; in bnxt_qplib_reg_mr() 617 pg_size = buf_pg_size ? buf_pg_size : PAGE_SIZE; in bnxt_qplib_reg_mr() 619 ((ilog2(pg_size) << in bnxt_qplib_reg_mr()
|
D | ib_verbs.c | 3881 resp.pg_size = PAGE_SIZE; in bnxt_re_alloc_ucontext()
|
/linux-5.19.10/drivers/iommu/amd/ |
D | io_pgtable.c | 53 unsigned long pte_mask, pg_size, cnt; in first_pte_l7() local 56 pg_size = PTE_PAGE_SIZE(*pte); in first_pte_l7() 57 cnt = PAGE_SIZE_PTE_COUNT(pg_size); in first_pte_l7() 62 *page_size = pg_size; in first_pte_l7()
|
/linux-5.19.10/include/uapi/rdma/ |
D | bnxt_re-abi.h | 65 __u32 pg_size; member
|
/linux-5.19.10/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
D | chtls_io.c | 1106 int pg_size = PAGE_SIZE; in chtls_sendmsg() local 1111 pg_size = page_size(page); in chtls_sendmsg() 1112 if (off < pg_size && in chtls_sendmsg() 1122 if (page && off == pg_size) { in chtls_sendmsg() 1125 pg_size = PAGE_SIZE; in chtls_sendmsg() 1138 pg_size <<= order; in chtls_sendmsg() 1142 pg_size = PAGE_SIZE; in chtls_sendmsg() 1149 if (copy > pg_size - off) in chtls_sendmsg() 1150 copy = pg_size - off; in chtls_sendmsg() 1171 if (off + copy < pg_size) { in chtls_sendmsg()
|
/linux-5.19.10/drivers/net/ethernet/huawei/hinic/ |
D | hinic_hw_eqs.c | 27 #define GET_EQ_NUM_PAGES(eq, pg_size) \ argument 28 (ALIGN((eq)->q_len * (eq)->elem_size, pg_size) / (pg_size)) 30 #define GET_EQ_NUM_ELEMS_IN_PG(eq, pg_size) ((pg_size) / (eq)->elem_size) argument
|
D | hinic_hw_io.c | 38 #define HINIC_PAGE_SIZE_HW(pg_size) ((u8)ilog2((u32)((pg_size) >> 12))) argument
|
/linux-5.19.10/drivers/crypto/caam/ |
D | ctrl.c | 632 int pg_size; in caam_probe() local 702 pg_size = (comp_params & CTPR_MS_PG_SZ_MASK) >> CTPR_MS_PG_SZ_SHIFT; in caam_probe() 703 if (pg_size == 0) in caam_probe()
|
/linux-5.19.10/tools/testing/selftests/kvm/lib/x86_64/ |
D | processor.c | 199 const uint64_t pg_size = PG_LEVEL_SIZE(level); in __virt_pg_map() local 206 TEST_ASSERT((vaddr % pg_size) == 0, in __virt_pg_map() 208 "vaddr: 0x%lx page size: 0x%lx", vaddr, pg_size); in __virt_pg_map() 211 TEST_ASSERT((paddr % pg_size) == 0, in __virt_pg_map() 213 " paddr: 0x%lx page size: 0x%lx", paddr, pg_size); in __virt_pg_map()
|
/linux-5.19.10/drivers/iommu/ |
D | iommu.c | 789 unsigned long pg_size; in iommu_create_device_direct_mappings() local 797 pg_size = 1UL << __ffs(domain->pgsize_bitmap); in iommu_create_device_direct_mappings() 807 start = ALIGN(entry->start, pg_size); in iommu_create_device_direct_mappings() 808 end = ALIGN(entry->start + entry->length, pg_size); in iommu_create_device_direct_mappings() 814 for (addr = start; addr <= end; addr += pg_size) { in iommu_create_device_direct_mappings() 822 map_size += pg_size; in iommu_create_device_direct_mappings()
|
/linux-5.19.10/drivers/infiniband/hw/irdma/ |
D | verbs.c | 2263 static bool irdma_check_mem_contiguous(u64 *arr, u32 npages, u32 pg_size) in irdma_check_mem_contiguous() argument 2268 if ((*arr + (pg_size * pg_idx)) != arr[pg_idx]) in irdma_check_mem_contiguous() 2281 u32 pg_size) in irdma_check_mr_contiguous() argument 2293 pg_size); in irdma_check_mr_contiguous() 2301 if ((*start_addr + (i * pg_size * PBLE_PER_PAGE)) != *arr) in irdma_check_mr_contiguous() 2303 ret = irdma_check_mem_contiguous(arr, leaf->cnt, pg_size); in irdma_check_mr_contiguous() 2367 u32 pg_size, total; in irdma_handle_q_mem() local 2371 pg_size = iwmr->page_size; in irdma_handle_q_mem() 2393 pg_size); in irdma_handle_q_mem() 2397 pg_size); in irdma_handle_q_mem() [all …]
|
/linux-5.19.10/drivers/misc/bcm-vk/ |
D | bcm_vk_dev.c | 1208 unsigned long pg_size; in bcm_vk_mmap() local 1213 pg_size = ((pci_resource_len(vk->pdev, VK_MMAPABLE_BAR) - 1) in bcm_vk_mmap() 1215 if (vma->vm_pgoff + vma_pages(vma) > pg_size) in bcm_vk_mmap()
|
/linux-5.19.10/drivers/net/wireless/realtek/rtw88/ |
D | fw.c | 1354 u32 pg_size; in rtw_download_drv_rsvd_page() local 1358 pg_size = rtwdev->chip->page_size; in rtw_download_drv_rsvd_page() 1359 pg_num = size / pg_size + ((size & (pg_size - 1)) ? 1 : 0); in rtw_download_drv_rsvd_page()
|
/linux-5.19.10/drivers/target/sbp/ |
D | sbp_target.c | 1242 int tcode, sg_miter_flags, max_payload, pg_size, speed, node_id, in sbp_rw_data() local 1261 pg_size = CMDBLK_ORB_PG_SIZE(be32_to_cpu(req->orb.misc)); in sbp_rw_data() 1262 if (pg_size) { in sbp_rw_data()
|
/linux-5.19.10/drivers/net/ethernet/chelsio/cxgb3/ |
D | t3_hw.c | 2522 unsigned int pg_size) in pm_num_pages() argument 2524 unsigned int n = mem_size / pg_size; in pm_num_pages()
|