Lines Matching refs:new
42 base: PhysAddr::new(0),
47 static mut INITIAL_CR3_VALUE: PhysAddr = PhysAddr::new(0);
53 static INNER_ALLOCATOR: SpinLock<Option<BuddyAllocator<MMArch>>> = SpinLock::new(None);
80 static XD_RESERVED: AtomicBool = AtomicBool::new(false);
121 const USER_END_VADDR: VirtAddr = VirtAddr::new(0x0000_7eff_ffff_ffff);
122 const USER_BRK_START: VirtAddr = VirtAddr::new(0x700000000000);
123 const USER_STACK_START: VirtAddr = VirtAddr::new(0x6ffff0a00000);
177 return PhysAddr::new(paddr); in table()
224 return Ok(crate::mm::ucontext::UserMapper::new(new_umapper)); in setup_new_usermapper()
253 PHYS_MEMORY_AREAS[areas_count].base = PhysAddr::new(mb2_mem_info[i].addr as usize); in init_memory_area_from_multiboot2()
299 static _CALL_ONCE: AtomicBool = AtomicBool::new(false); in mm_init()
311 MMArch::phys_2_virt(PhysAddr::new(0)).unwrap().data() in mm_init()
325 unsafe { MMArch::virt_2_phys(VirtAddr::new(page_align_up(virt_offset))) }.unwrap(); in allocator_init()
329 BumpAllocator::<X86_64MMArch>::new(&PHYS_MEMORY_AREAS, phy_offset.data()); in allocator_init()
356 let empty_entry = PageEntry::<MMArch>::new(0); in allocator_init()
393 let buddy_allocator = unsafe { BuddyAllocator::<X86_64MMArch>::new(bump_allocator).unwrap() }; in allocator_init()
408 let mapper = crate::mm::page::PageMapper::<MMArch, _>::new( in allocator_init()
445 let mut addr_set: HashSet<PhysAddr> = HashSet::new(); in test_buddy()
562 return PageFlags::new().set_execute(true).set_write(true); in kernel_page_flags()
565 return PageFlags::new().set_execute(true); in kernel_page_flags()
567 return PageFlags::new().set_write(true).set_execute(true); in kernel_page_flags()
572 static FLAG: AtomicBool = AtomicBool::new(false); in set_inner_allocator()
595 let paddr = PhysAddr::new(i * MMArch::PAGE_SIZE); in remap_at_low_address()
596 let vaddr = VirtAddr::new(i * MMArch::PAGE_SIZE); in remap_at_low_address()
612 let vaddr = VirtAddr::new(i * MMArch::PAGE_SIZE); in unmap_at_low_address()