Lines Matching refs:pent

49 #define lv2ent_fault(pent) ((*(pent) & 3) == 0)  argument
50 #define lv2ent_small(pent) ((*(pent) & 2) == 2) argument
51 #define lv2ent_large(pent) ((*(pent) & 3) == 1) argument
98 #define lpage_phys(pent) (sect_to_phys(*(pent)) & LPAGE_MASK) argument
100 #define spage_phys(pent) (sect_to_phys(*(pent)) & SPAGE_MASK) argument
1006 sysmmu_pte_t *pent; in alloc_lv2entry() local
1009 pent = kmem_cache_zalloc(lv2table_kmem_cache, GFP_ATOMIC); in alloc_lv2entry()
1010 BUG_ON((uintptr_t)pent & (LV2TABLE_SIZE - 1)); in alloc_lv2entry()
1011 if (!pent) in alloc_lv2entry()
1014 exynos_iommu_set_pte(sent, mk_lv1ent_page(virt_to_phys(pent))); in alloc_lv2entry()
1015 kmemleak_ignore(pent); in alloc_lv2entry()
1017 handle = dma_map_single(dma_dev, pent, LV2TABLE_SIZE, in alloc_lv2entry()
1020 kmem_cache_free(lv2table_kmem_cache, pent); in alloc_lv2entry()
1092 static int lv2set_page(sysmmu_pte_t *pent, phys_addr_t paddr, size_t size, in lv2set_page() argument
1096 if (WARN_ON(!lv2ent_fault(pent))) in lv2set_page()
1099 exynos_iommu_set_pte(pent, mk_lv2ent_spage(paddr, prot)); in lv2set_page()
1103 dma_addr_t pent_base = virt_to_phys(pent); in lv2set_page()
1106 sizeof(*pent) * SPAGES_PER_LPAGE, in lv2set_page()
1108 for (i = 0; i < SPAGES_PER_LPAGE; i++, pent++) { in lv2set_page()
1109 if (WARN_ON(!lv2ent_fault(pent))) { in lv2set_page()
1111 memset(pent - i, 0, sizeof(*pent) * i); in lv2set_page()
1115 *pent = mk_lv2ent_lpage(paddr, prot); in lv2set_page()
1118 sizeof(*pent) * SPAGES_PER_LPAGE, in lv2set_page()
1173 sysmmu_pte_t *pent; in exynos_iommu_map() local
1175 pent = alloc_lv2entry(domain, entry, iova, in exynos_iommu_map()
1178 if (IS_ERR(pent)) in exynos_iommu_map()
1179 ret = PTR_ERR(pent); in exynos_iommu_map()
1181 ret = lv2set_page(pent, paddr, size, prot, in exynos_iommu_map()