Searched refs:PAGE_SHIFT (Results 1 – 15 of 15) sorted by relevance
/DragonOS/kernel/src/mm/allocator/ |
H A D | bump.rs | 185 let area_base = (area.base.data() + MMA::PAGE_SHIFT) & !(MMA::PAGE_SHIFT); in usage() 187 let area_end = (area.base.data() + area.size) & !(MMA::PAGE_SHIFT); in usage() 189 total += (area_end - area_base) >> MMA::PAGE_SHIFT; in usage() 192 used += (area_end - area_base) >> MMA::PAGE_SHIFT; in usage() 197 used += (self.offset - area_base) >> MMA::PAGE_SHIFT; in usage()
|
H A D | buddy.rs | 245 self.buddy_free(page_list_addr, MMArch::PAGE_SHIFT as u8); in pop_front() 288 unsafe { self.buddy_free(page_list_addr, MMArch::PAGE_SHIFT as u8) }; in pop_front()
|
H A D | page_frame.rs | 24 number: paddr.data() >> MMArch::PAGE_SHIFT, in new()
|
/DragonOS/kernel/src/virt/kvm/ |
H A D | host_mem.rs | 37 pub const PAGE_SHIFT: u32 = 12; constant 38 pub const PAGE_SIZE: u32 = 1 << PAGE_SHIFT; 152 return Ok(hpa.data() as u64 >> PAGE_SHIFT); in hva_to_pfn() 158 return Ok(hpa.data() as u64 >> PAGE_SHIFT); in hva_to_pfn()
|
H A D | vm.rs | 15 KVM_MEM_SLOTS_NUM, KVM_USER_MEM_SLOTS, PAGE_SHIFT, 77 let base_gfn = mem.guest_phys_addr >> PAGE_SHIFT; in set_user_memory_region() 78 let npages = mem.memory_size >> PAGE_SHIFT; in set_user_memory_region()
|
/DragonOS/kernel/src/driver/firmware/efi/ |
H A D | init.rs | 110 let size = (md.page_count << (MMArch::PAGE_SHIFT as u64)) as usize; in efi_find_mirror() 240 && ((efi_vaddr - md.virt_start) < (md.page_count << (MMArch::PAGE_SHIFT as u64))) in efi_vaddr_2_paddr() 260 (md.phys_start + (md.page_count << (MMArch::PAGE_SHIFT as u64))) as usize, in reserve_memory_regions() 266 let size = (page_count << (MMArch::PAGE_SHIFT as u64)) as usize; in reserve_memory_regions()
|
/DragonOS/kernel/src/arch/x86_64/kvm/vmx/ |
H A D | mmu.rs | 5 virt::kvm::host_mem::{__gfn_to_pfn, kvm_vcpu_gfn_to_memslot, PAGE_MASK, PAGE_SHIFT}, 109 let gfn = gpa >> PAGE_SHIFT; // 物理地址右移12位得到物理页框号(相对于虚拟机而言) in tdp_page_fault() 223 assert!(ept_mapper.walk(gpa, pfn << PAGE_SHIFT, page_flags).is_ok()); in __direct_map()
|
/DragonOS/kernel/src/mm/ |
H A D | fault.rs | 75 ((address - guard.region().start()) >> MMArch::PAGE_SHIFT) + file_page_offset in new() 536 let vm_pgoff = (address - vma_region.start()) >> MMArch::PAGE_SHIFT; in do_fault_around() 539 let pte_pgoff = (address.data() >> MMArch::PAGE_SHIFT) & (1 << MMArch::PAGE_ENTRY_SHIFT); in do_fault_around() 544 let vma_pages_count = (vma_region.end() - vma_region.start()) >> MMArch::PAGE_SHIFT; in do_fault_around() 561 1 << MMArch::PAGE_SHIFT, in do_fault_around() 608 << MMArch::PAGE_SHIFT); in filemap_map_pages() 617 VirtAddr::new(addr.data() + ((pgoff - start_pgoff) << MMArch::PAGE_SHIFT)); in filemap_map_pages()
|
H A D | mod.rs | 445 const PAGE_SHIFT: usize; constant 493 const PAGE_SIZE: usize = 1 << Self::PAGE_SHIFT; 500 const PAGE_ADDRESS_SHIFT: usize = Self::PAGE_LEVELS * Self::PAGE_ENTRY_SHIFT + Self::PAGE_SHIFT; 506 const PAGE_ENTRY_SIZE: usize = 1 << (Self::PAGE_SHIFT - Self::PAGE_ENTRY_SHIFT);
|
H A D | syscall.rs | 612 + (guard.file_page_offset().unwrap_or(0) << MMArch::PAGE_SHIFT); in msync()
|
H A D | page.rs | 528 let shift = self.level * Arch::PAGE_ENTRY_SHIFT + Arch::PAGE_SHIFT; in entry_base() 583 let shift = self.level * Arch::PAGE_ENTRY_SHIFT + Arch::PAGE_SHIFT; in index_of()
|
H A D | ucontext.rs | 373 let pgoff = offset >> MMArch::PAGE_SHIFT; in file_mapping() 1667 self.region.start + ((index - self.file_pgoff.unwrap()) << MMArch::PAGE_SHIFT); in page_address()
|
/DragonOS/kernel/src/arch/x86_64/mm/ |
H A D | mod.rs | 79 const PAGE_SHIFT: usize = 12; constant 257 const PAGE_SIZE: usize = 1 << Self::PAGE_SHIFT; 263 const PAGE_ADDRESS_SHIFT: usize = Self::PAGE_LEVELS * Self::PAGE_ENTRY_SHIFT + Self::PAGE_SHIFT; 269 const PAGE_ENTRY_SIZE: usize = 1 << (Self::PAGE_SHIFT - Self::PAGE_ENTRY_SHIFT);
|
/DragonOS/kernel/src/arch/riscv64/mm/ |
H A D | mod.rs | 91 const PAGE_SHIFT: usize = 12; constant
|
/DragonOS/kernel/src/filesystem/vfs/ |
H A D | file.rs | 141 .range(0..((MMArch::PAGE_ADDRESS_SIZE >> MMArch::PAGE_SHIFT) as u64)) in fmt()
|