/linux-6.1.9/arch/xtensa/kernel/ |
D | setup.c | 298 !mem_reserve(__pa(initrd_start), __pa(initrd_end))) in setup_arch() 304 mem_reserve(__pa(_stext), __pa(_end)); in setup_arch() 306 mem_reserve(__pa(_xip_start), __pa(_xip_end)); in setup_arch() 311 mem_reserve(__pa(_WindowVectors_text_start), in setup_arch() 312 __pa(_WindowVectors_text_end)); in setup_arch() 315 mem_reserve(__pa(_DebugInterruptVector_text_start), in setup_arch() 316 __pa(_DebugInterruptVector_text_end)); in setup_arch() 318 mem_reserve(__pa(_KernelExceptionVector_text_start), in setup_arch() 319 __pa(_KernelExceptionVector_text_end)); in setup_arch() 321 mem_reserve(__pa(_UserExceptionVector_text_start), in setup_arch() [all …]
|
/linux-6.1.9/arch/parisc/kernel/ |
D | firmware.c | 162 __pa(pdc_result), 0); in set_firmware_width_unlocked() 248 retval = mem_pdc_call(PDC_INSTR, 0UL, __pa(pdc_result)); in pdc_instr() 273 __pa(pdc_result), __pa(pdc_result2), len); in pdc_chassis_info() 297 retval = mem_pdc_call(PDC_PAT_CHASSIS_LOG, PDC_PAT_CHASSIS_WRITE_LOG, __pa(&state), __pa(&data)); in pdc_pat_chassis_send_log() 364 __pa(pdc_result)); in pdc_pat_get_PDC_entrypoint() 380 retval = mem_pdc_call(PDC_CHASSIS, PDC_CHASSIS_WARN, __pa(pdc_result)); in pdc_chassis_warn() 391 ret = mem_pdc_call(PDC_COPROC, PDC_COPROC_CFG, __pa(pdc_result)); in pdc_coproc_cfg_unlocked() 438 retval = mem_pdc_call(PDC_IODC, PDC_IODC_READ, __pa(pdc_result), hpa, in pdc_iodc_read() 439 index, __pa(pdc_result2), iodc_data_size); in pdc_iodc_read() 465 retval = mem_pdc_call(PDC_SYSTEM_MAP, PDC_FIND_MODULE, __pa(pdc_result), in pdc_system_map_find_mods() [all …]
|
/linux-6.1.9/arch/x86/include/asm/ |
D | pgalloc.h | 67 paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); in pmd_populate_kernel() 68 set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); in pmd_populate_kernel() 74 paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); in pmd_populate_kernel_safe() 75 set_pmd_safe(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); in pmd_populate_kernel_safe() 101 paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); in pud_populate() 102 set_pud(pud, __pud(_PAGE_TABLE | __pa(pmd))); in pud_populate() 107 paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); in pud_populate_safe() 108 set_pud_safe(pud, __pud(_PAGE_TABLE | __pa(pmd))); in pud_populate_safe() 115 paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT); in p4d_populate() 116 set_p4d(p4d, __p4d(_PAGE_TABLE | __pa(pud))); in p4d_populate() [all …]
|
D | page.h | 41 #ifndef __pa 42 #define __pa(x) __phys_addr((unsigned long)(x)) macro 63 #define __boot_pa(x) __pa(x) 69 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
|
/linux-6.1.9/arch/x86/kernel/ |
D | head32.c | 74 #ifdef __pa in mk_early_pgtbl_32() 75 #undef __pa in mk_early_pgtbl_32() 77 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET) in mk_early_pgtbl_32() macro 82 const unsigned long limit = __pa(_end) + in mk_early_pgtbl_32() 85 pmd_t pl2, *pl2p = (pmd_t *)__pa(initial_pg_pmd); in mk_early_pgtbl_32() 88 pgd_t pl2, *pl2p = (pgd_t *)__pa(initial_page_table); in mk_early_pgtbl_32() 92 ptep = (pte_t *)__pa(__brk_base); in mk_early_pgtbl_32() 112 ptr = (unsigned long *)__pa(&max_pfn_mapped); in mk_early_pgtbl_32() 116 ptr = (unsigned long *)__pa(&_brk_end); in mk_early_pgtbl_32()
|
D | machine_kexec_32.c | 89 set_pgd(pgd, __pgd(__pa(pmd) | _PAGE_PRESENT)); in machine_kexec_page_table_set_one() 95 set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); in machine_kexec_page_table_set_one() 111 (unsigned long)control_page, __pa(control_page)); in machine_kexec_prepare_page_tables() 117 __pa(control_page), __pa(control_page)); in machine_kexec_prepare_page_tables() 201 page_list[PA_CONTROL_PAGE] = __pa(control_page); in machine_kexec() 203 page_list[PA_PGD] = __pa(image->arch.pgd); in machine_kexec()
|
D | espfix_64.c | 170 pud = __pud(__pa(pmd_p) | (PGTABLE_PROT & ptemask)); in init_espfix_ap() 171 paravirt_alloc_pmd(&init_mm, __pa(pmd_p) >> PAGE_SHIFT); in init_espfix_ap() 182 pmd = __pmd(__pa(pte_p) | (PGTABLE_PROT & ptemask)); in init_espfix_ap() 183 paravirt_alloc_pte(&init_mm, __pa(pte_p) >> PAGE_SHIFT); in init_espfix_ap() 194 pte = __pte(__pa(stack_page) | ((__PAGE_KERNEL_RO | _PAGE_ENC) & ptemask)); in init_espfix_ap()
|
/linux-6.1.9/arch/arm/mach-omap2/ |
D | omap-secure.c | 81 outer_clean_range(__pa(param), __pa(param + 5)); in omap_secure_dispatcher() 82 ret = omap_smc2(idx, flag, __pa(param)); in omap_secure_dispatcher() 136 param[1] = __pa(addr); /* Physical address for saving */ in omap3_save_secure_ram() 141 ret = save_secure_ram_context(__pa(param)); in omap3_save_secure_ram() 179 outer_clean_range(__pa(param), __pa(param + 5)); in rx51_secure_dispatcher() 180 ret = omap_smc3(idx, process, flag, __pa(param)); in rx51_secure_dispatcher()
|
/linux-6.1.9/arch/csky/mm/ |
D | init.c | 55 if (__pa(initrd_end) > PFN_PHYS(max_low_pfn)) { in setup_initrd() 62 if (memblock_is_region_reserved(__pa(initrd_start), size)) { in setup_initrd() 64 __pa(initrd_start), size); in setup_initrd() 68 memblock_reserve(__pa(initrd_start), size); in setup_initrd() 122 p[i] = __pa(invalid_pte_table); in pgd_init() 133 swapper_pg_dir[i].pgd = __pa(invalid_pte_table); in mmu_init() 137 __pa(kernel_pte_tables + (PTRS_PER_PTE * (i - USER_PTRS_PER_PGD))); in mmu_init() 182 set_pmd(pmd, __pmd(__pa(pte))); in fixrange_init()
|
/linux-6.1.9/arch/powerpc/mm/nohash/ |
D | 8xx.c | 30 return __pa(va); in v_block_mapped() 140 unsigned long etext8 = ALIGN(__pa(_etext), SZ_8M); in mmu_mapin_ram() 141 unsigned long sinittext = __pa(_sinittext); in mmu_mapin_ram() 144 unsigned long einittext8 = ALIGN(__pa(_einittext), SZ_8M); in mmu_mapin_ram() 168 unsigned long etext8 = ALIGN(__pa(_etext), SZ_8M); in mmu_mark_initmem_nx() 169 unsigned long sinittext = __pa(_sinittext); in mmu_mark_initmem_nx() 171 unsigned long einittext8 = ALIGN(__pa(_einittext), SZ_8M); in mmu_mark_initmem_nx() 182 unsigned long sinittext = __pa(_sinittext); in mmu_mark_rodata_ro()
|
/linux-6.1.9/arch/um/kernel/ |
D | physmem.c | 107 os_seek_file(physmem_fd, __pa(__syscall_stub_start)); in setup_physmem() 111 memblock_add(__pa(start), len + highmem); in setup_physmem() 112 memblock_reserve(__pa(start), reserve); in setup_physmem() 114 min_low_pfn = PFN_UP(__pa(reserve_end)); in setup_physmem() 126 else if (phys < __pa(end_iomem)) { in phys_mapping() 139 else if (phys < __pa(end_iomem) + highmem) { in phys_mapping() 212 region->phys = __pa(region->virt); in setup_iomem()
|
/linux-6.1.9/arch/arm/mm/ |
D | pmsa-v8.c | 255 subtract_range(mem, ARRAY_SIZE(mem), __pa(KERNEL_START), __pa(KERNEL_END)); in pmsav8_setup() 256 subtract_range(io, ARRAY_SIZE(io), __pa(KERNEL_START), __pa(KERNEL_END)); in pmsav8_setup() 260 subtract_range(mem, ARRAY_SIZE(mem), CONFIG_XIP_PHYS_ADDR, __pa(_exiprom)); in pmsav8_setup() 261 subtract_range(io, ARRAY_SIZE(io), CONFIG_XIP_PHYS_ADDR, __pa(_exiprom)); in pmsav8_setup() 277 err |= pmsav8_setup_fixed(PMSAv8_XIP_REGION, CONFIG_XIP_PHYS_ADDR, __pa(_exiprom)); in pmsav8_setup() 280 err |= pmsav8_setup_fixed(region++, __pa(KERNEL_START), __pa(KERNEL_END)); in pmsav8_setup()
|
/linux-6.1.9/arch/parisc/mm/ |
D | init.c | 284 memblock_reserve(__pa(KERNEL_BINARY_TEXT_START), in setup_bootmem() 300 if (__pa(initrd_start) < mem_max) { in setup_bootmem() 303 if (__pa(initrd_end) > mem_max) { in setup_bootmem() 304 initrd_reserve = mem_max - __pa(initrd_start); in setup_bootmem() 309 …printk(KERN_INFO "initrd: reserving %08lx-%08lx (mem_max %08lx)\n", __pa(initrd_start), __pa(initr… in setup_bootmem() 311 memblock_reserve(__pa(initrd_start), initrd_reserve); in setup_bootmem() 357 ro_start = __pa((unsigned long)_text); in map_pages() 358 ro_end = __pa((unsigned long)&data_start); in map_pages() 359 kernel_start = __pa((unsigned long)&__init_begin); in map_pages() 360 kernel_end = __pa((unsigned long)&_end); in map_pages() [all …]
|
/linux-6.1.9/arch/x86/xen/ |
D | p2m.c | 346 paravirt_alloc_pte(&init_mm, __pa(p2m_missing_pte) >> PAGE_SHIFT); in xen_rebuild_p2m_list() 348 paravirt_alloc_pte(&init_mm, __pa(p2m_identity_pte) >> PAGE_SHIFT); in xen_rebuild_p2m_list() 351 pfn_pte(PFN_DOWN(__pa(p2m_missing)), PAGE_KERNEL_RO)); in xen_rebuild_p2m_list() 353 pfn_pte(PFN_DOWN(__pa(p2m_identity)), PAGE_KERNEL_RO)); in xen_rebuild_p2m_list() 386 pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL)); in xen_rebuild_p2m_list() 396 pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL_RO)); in xen_rebuild_p2m_list() 406 set_pmd(pmdp, __pmd(__pa(ptep) | _KERNPG_TABLE)); in xen_rebuild_p2m_list() 455 if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity))) in get_phys_to_machine() 493 paravirt_alloc_pte(&init_mm, __pa(pte_newpg[i]) >> PAGE_SHIFT); in alloc_p2m_pmd() 505 __pmd(__pa(pte_newpg[i]) | _KERNPG_TABLE)); in alloc_p2m_pmd() [all …]
|
/linux-6.1.9/arch/x86/power/ |
D | hibernate_64.c | 71 __pud(__pa(pmd) | pgprot_val(pgtable_prot))); in set_up_temporary_text_mapping() 73 p4d_t new_p4d = __p4d(__pa(pud) | pgprot_val(pgtable_prot)); in set_up_temporary_text_mapping() 74 pgd_t new_pgd = __pgd(__pa(p4d) | pgprot_val(pgtable_prot)); in set_up_temporary_text_mapping() 80 pgd_t new_pgd = __pgd(__pa(pud) | pgprot_val(pgtable_prot)); in set_up_temporary_text_mapping() 123 temp_pgt = __pa(pgd); in set_up_temporary_mappings()
|
D | hibernate_32.c | 41 set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT)); in resume_one_md_table_init() 66 set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE)); in resume_one_page_table_init() 141 __pgd(__pa(empty_zero_page) | _PAGE_PRESENT)); in resume_init_first_level_page_table() 189 temp_pgt = __pa(resume_pg_dir); in swsusp_arch_resume()
|
/linux-6.1.9/arch/riscv/mm/ |
D | kasan_init.c | 55 set_pmd(pmd, pfn_pmd(PFN_DOWN(__pa(base_pte)), PAGE_TABLE)); in kasan_populate_pte() 94 set_pud(pud, pfn_pud(PFN_DOWN(__pa(base_pmd)), PAGE_TABLE)); in kasan_populate_pmd() 134 phys_addr = __pa(((uintptr_t)kasan_early_shadow_pmd)); in kasan_populate_pud() 156 set_pgd(pgd, pfn_pgd(PFN_DOWN(__pa(base_pud)), PAGE_TABLE)); in kasan_populate_pud() 192 phys_addr = __pa(((uintptr_t)kasan_early_shadow_pud)); in kasan_populate_p4d() 214 set_pgd(pgd, pfn_pgd(PFN_DOWN(__pa(base_p4d)), PAGE_TABLE)); in kasan_populate_p4d() 241 phys_addr = __pa((uintptr_t)kasan_early_shadow_pgd_next); in kasan_populate_pgd() 279 (__pa((uintptr_t)kasan_early_shadow_pte)), in kasan_early_init() 286 (__pa(((uintptr_t)kasan_early_shadow_pmd))), in kasan_early_init() 294 (__pa(((uintptr_t)kasan_early_shadow_pud))), in kasan_early_init() [all …]
|
D | init.c | 226 max_mapped_addr = __pa(~(ulong)0); in setup_bootmem() 367 return __pa(vaddr); in alloc_pte_late() 450 return __pa(vaddr); in alloc_pmd_late() 515 return __pa(vaddr); in alloc_pud_late() 553 return __pa(vaddr); in alloc_p4d_late() 614 #define get_pgd_next_virt(__pa) (pgtable_l5_enabled ? \ argument 615 pt_ops.get_p4d_virt(__pa) : (pgd_next_t *)(pgtable_l4_enabled ? \ 616 pt_ops.get_pud_virt(__pa) : (pud_t *)pt_ops.get_pmd_virt(__pa))) 617 #define create_pgd_next_mapping(__nextp, __va, __pa, __sz, __prot) \ argument 619 create_p4d_mapping(__nextp, __va, __pa, __sz, __prot) : \ [all …]
|
/linux-6.1.9/arch/s390/mm/ |
D | kasan_init.c | 178 set_pmd(pm_dir, __pmd(__pa(page) | sgt_prot)); in kasan_early_pgtable_populate() 197 set_pte(pt_dir, __pte(__pa(page) | pgt_prot)); in kasan_early_pgtable_populate() 202 set_pte(pt_dir, __pte(__pa(page) | pgt_prot)); in kasan_early_pgtable_populate() 206 set_pte(pt_dir, __pte(__pa(page) | pgt_prot_zero)); in kasan_early_pgtable_populate() 222 S390_lowcore.kernel_asce = (__pa(pgd) & PAGE_MASK) | asce_bits; in kasan_set_pgd() 259 pmd_t pmd_z = __pmd(__pa(kasan_early_shadow_pte) | _SEGMENT_ENTRY); in kasan_early_init() 260 pud_t pud_z = __pud(__pa(kasan_early_shadow_pmd) | _REGION3_ENTRY); in kasan_early_init() 261 p4d_t p4d_z = __p4d(__pa(kasan_early_shadow_pud) | _REGION2_ENTRY); in kasan_early_init() 266 pte_z = __pte(__pa(kasan_early_shadow_page) | pgt_prot); in kasan_early_init()
|
/linux-6.1.9/arch/hexagon/include/asm/ |
D | page.h | 86 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + PHYS_OFFSET) macro 93 #define virt_to_page(kaddr) pfn_to_page(PFN_DOWN(__pa(kaddr))) 99 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) 129 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT)
|
D | pgalloc.h | 39 mm->context.ptbase = __pa(pgd); in pgd_alloc() 77 set_pmd(pmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE)); in pmd_populate_kernel() 85 set_pmd(ppmd, __pmd(((unsigned long)__pa(pte)) | HEXAGON_L1_PTE_SIZE)); in pmd_populate_kernel()
|
/linux-6.1.9/arch/ia64/include/asm/ |
D | pgalloc.h | 36 p4d_val(*p4d_entry) = __pa(pud); in p4d_populate() 45 pud_val(*pud_entry) = __pa(pmd); in pud_populate() 59 pmd_val(*pmd_entry) = __pa(pte); in pmd_populate_kernel()
|
/linux-6.1.9/arch/powerpc/platforms/pseries/ |
D | svm.c | 48 uv_unshare_page(PHYS_PFN(__pa(addr)), numpages); in set_memory_encrypted() 61 uv_share_page(PHYS_PFN(__pa(addr)), numpages); in set_memory_decrypted() 85 unsigned long pfn = PHYS_PFN(__pa(addr)); in dtl_cache_ctor()
|
/linux-6.1.9/arch/x86/mm/ |
D | init_32.c | 77 paravirt_alloc_pmd(&init_mm, __pa(pmd_table) >> PAGE_SHIFT); in one_md_table_init() 78 set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT)); in one_md_table_init() 102 paravirt_alloc_pte(&init_mm, __pa(page_table) >> PAGE_SHIFT); in one_page_table_init() 103 set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE)); in one_page_table_init() 184 paravirt_alloc_pte(&init_mm, __pa(newpte) >> PAGE_SHIFT); in page_table_kmap_check() 185 set_pmd(pmd, __pmd(__pa(newpte)|_PAGE_TABLE)); in page_table_kmap_check() 189 paravirt_release_pte(__pa(pte) >> PAGE_SHIFT); in page_table_kmap_check() 473 pfn, pmd, __pa(pmd)); in native_pagetable_init() 482 pfn, pmd, __pa(pmd), pte, __pa(pte)); in native_pagetable_init() 485 paravirt_alloc_pmd(&init_mm, __pa(base) >> PAGE_SHIFT); in native_pagetable_init()
|
/linux-6.1.9/drivers/pci/controller/ |
D | pci-versatile.c | 94 writel(__pa(PAGE_OFFSET) >> 28, PCI_SMAP(mem)); in versatile_pci_probe() 131 writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_0); in versatile_pci_probe() 132 writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_1); in versatile_pci_probe() 133 writel(__pa(PAGE_OFFSET), local_pci_cfg_base + PCI_BASE_ADDRESS_2); in versatile_pci_probe()
|