/linux-6.1.9/mm/ |
D | folio-compat.c | 14 return folio_mapping(page_folio(page)); in page_mapping() 20 return folio_unlock(page_folio(page)); in unlock_page() 26 return folio_end_writeback(page_folio(page)); in end_page_writeback() 32 return folio_wait_writeback(page_folio(page)); in wait_on_page_writeback() 38 return folio_wait_stable(page_folio(page)); in wait_for_stable_page() 44 return folio_mapped(page_folio(page)); in page_mapped() 50 folio_mark_accessed(page_folio(page)); in mark_page_accessed() 56 return folio_start_writeback(page_folio(page)); in set_page_writeback() 62 return folio_mark_dirty(page_folio(page)); in set_page_dirty() 68 return filemap_dirty_folio(page_mapping(page), page_folio(page)); in __set_page_dirty_nobuffers() [all …]
|
D | migrate_device.c | 224 folio_mark_dirty(page_folio(page)); in migrate_vma_collect_pmd() 401 folio = page_folio(page); in migrate_device_unmap() 427 folio = page_folio(page); in migrate_device_unmap() 621 if (mem_cgroup_charge(page_folio(page), vma->vm_mm, GFP_KERNEL)) in migrate_vma_insert_page() 767 r = migrate_folio_extra(mapping, page_folio(newpage), in __migrate_device_pages() 768 page_folio(page), in __migrate_device_pages() 771 r = migrate_folio(mapping, page_folio(newpage), in __migrate_device_pages() 772 page_folio(page), MIGRATE_SYNC_NO_COPY); in __migrate_device_pages() 852 src = page_folio(page); in migrate_device_finalize() 853 dst = page_folio(newpage); in migrate_device_finalize()
|
D | mlock.c | 67 lruvec = folio_lruvec_relock_irq(page_folio(page), lruvec); in __mlock_page() 107 lruvec = folio_lruvec_relock_irq(page_folio(page), lruvec); in __mlock_new_page() 131 lruvec = folio_lruvec_relock_irq(page_folio(page), lruvec); in __munlock_page() 325 mlock_folio(page_folio(page)); in mlock_pte_range() 341 mlock_folio(page_folio(page)); in mlock_pte_range()
|
D | gup.c | 52 struct folio *folio = page_folio(page); in sanity_check_pinned_pages() 74 folio = page_folio(page); in try_get_folio() 89 if (unlikely(page_folio(page) != folio)) { in try_get_folio() 211 struct folio *folio = page_folio(page); in try_grab_page() 250 gup_put_folio(page_folio(page), 1, FOLL_PIN); in unpin_user_page() 258 struct folio *folio = page_folio(next); in gup_folio_range_next() 272 struct folio *folio = page_folio(list[i]); in gup_folio_next() 276 if (page_folio(list[nr]) != folio) in gup_folio_next() 1957 struct folio *folio = page_folio(pages[i]); in collect_longterm_unpinnable_pages() 2007 struct folio *folio = page_folio(pages[i]); in migrate_longterm_unpinnable_pages()
|
D | swap.c | 732 struct folio *folio = page_folio(page); in deactivate_page() 755 struct folio *folio = page_folio(page); in mark_page_lazyfree() 986 struct folio *folio = page_folio(pages[i]); in release_pages()
|
D | page_idle.c | 88 struct folio *folio = page_folio(page); in page_idle_clear_pte_refs()
|
D | debug.c | 51 struct folio *folio = page_folio(page); in __dump_page()
|
D | migrate.c | 1163 struct folio *dst, *src = page_folio(page); in unmap_and_move() 1181 dst = page_folio(newpage); in unmap_and_move() 1255 struct folio *dst, *src = page_folio(hpage); in unmap_and_move_huge_page() 1281 dst = page_folio(new_hpage); in unmap_and_move_huge_page() 1608 struct folio *folio = page_folio(page); in alloc_migration_target()
|
D | truncate.c | 266 return truncate_inode_folio(mapping, page_folio(page)); in generic_error_remove_page() 297 struct folio *folio = page_folio(page); in invalidate_inode_page()
|
D | huge_memory.c | 662 if (mem_cgroup_charge(page_folio(page), vma->vm_mm, gfp)) { in __do_huge_pmd_anonymous_page() 1330 folio = page_folio(page); in do_huge_pmd_wp_page() 2292 if (folio && folio != page_folio(pmd_page(*pmd))) in __split_huge_pmd() 2467 if (!folio_test_swapcache(page_folio(head))) { in __split_huge_page_tail() 2505 struct folio *folio = page_folio(page); in __split_huge_page() 2533 struct folio *tail = page_folio(head + i); in __split_huge_page() 2635 struct folio *folio = page_folio(page); in split_huge_page_to_list() 3024 if (!can_split_folio(page_folio(page), NULL)) in split_huge_pages_pid()
|
D | memory-failure.c | 503 struct folio *folio = page_folio(page); in collect_procs_anon() 813 mem_cgroup_uncharge(page_folio(p)); in delete_from_lru_cache() 1058 struct folio *folio = page_folio(p); in me_swapcache_clean() 1403 struct folio *folio = page_folio(hpage); in hwpoison_user_mappings()
|
D | rmap.c | 749 struct folio *folio = page_folio(page); in page_address_in_vma() 1101 struct folio *folio = page_folio(page); in page_move_anon_rmap() 1164 struct folio *folio = page_folio(page); in __page_check_anon_rmap() 2352 struct folio *folio = page_folio(pages[i]); in make_device_exclusive_range()
|
/linux-6.1.9/include/linux/ |
D | page_idle.h | 125 return folio_test_young(page_folio(page)); in page_is_young() 130 folio_set_young(page_folio(page)); in set_page_young() 135 return folio_test_clear_young(page_folio(page)); in test_and_clear_page_young() 140 return folio_test_idle(page_folio(page)); in page_is_idle() 145 folio_set_idle(page_folio(page)); in set_page_idle() 150 folio_clear_idle(page_folio(page)); in clear_page_idle()
|
D | netfs.h | 93 folio_start_fscache(page_folio(page)); in set_page_fscache() 98 folio_end_private_2(page_folio(page)); in end_page_fscache() 103 folio_wait_private_2(page_folio(page)); in wait_on_page_fscache() 108 return folio_wait_private_2_killable(page_folio(page)); in wait_on_page_fscache_killable()
|
D | pagemap.h | 374 return folio_file_mapping(page_folio(page)); in page_file_mapping() 382 struct folio *folio = page_folio(page); in page_mapping_file() 461 folio_attach_private(page_folio(page), data); in attach_page_private() 466 return folio_detach_private(page_folio(page)); in detach_page_private() 910 return folio_trylock(page_folio(page)); in trylock_page() 958 folio = page_folio(page); in lock_page() 988 return folio_lock_killable(page_folio(page)); in lock_page_killable() 1034 folio_wait_locked(page_folio(page)); in wait_on_page_locked() 1039 return folio_wait_locked_killable(page_folio(page)); in wait_on_page_locked_killable() 1053 __folio_mark_dirty(page_folio(page), mapping, warn); in __set_page_dirty() [all …]
|
D | page-flags.h | 275 #define page_folio(p) (_Generic((p), \ macro 549 return folio_test_swapcache(page_folio(page)); in PageSwapCache() 663 return folio_test_anon(page_folio(page)); in PageAnon() 693 return folio_test_ksm(page_folio(page)); in PageKsm() 730 return folio_test_uptodate(page_folio(page)); in PageUptodate() 772 folio_start_writeback_keepwrite(page_folio(page)); in set_page_writeback_keepwrite()
|
D | mm_inline.h | 34 return folio_is_file_lru(page_folio(page)); in page_is_file_lru() 329 lruvec_add_folio(lruvec, page_folio(page)); in add_page_to_lru_list() 363 lruvec_del_folio(lruvec, page_folio(page)); in del_page_from_lru_list()
|
D | mm.h | 836 return folio_entire_mapcount(page_folio(page)); in compound_mapcount() 871 return folio_mapcount(page_folio(page)); in total_mapcount() 892 return page_folio(page); in virt_to_folio() 1129 folio_get(page_folio(page)); in get_page() 1203 struct folio *folio = page_folio(page); in put_page() 1529 return page_folio(pfn_to_page(pfn)); in pfn_folio() 1581 return folio_maybe_dma_pinned(page_folio(page)); in page_maybe_dma_pinned()
|
D | writeback.h | 388 folio_account_redirty(page_folio(page)); in account_page_redirty()
|
D | page_ref.h | 94 return folio_ref_count(page_folio(page)); in page_count()
|
D | rmap.h | 493 return folio_mkclean(page_folio(page)); in page_mkclean()
|
/linux-6.1.9/mm/damon/ |
D | paddr.c | 46 folio = page_folio(page); in damon_pa_mkold() 140 folio = page_folio(page); in damon_pa_young()
|
/linux-6.1.9/fs/9p/ |
D | vfs_addr.c | 201 struct folio *folio = page_folio(page); in v9fs_vfs_writepage() 302 struct folio *folio = page_folio(subpage); in v9fs_write_end()
|
/linux-6.1.9/fs/afs/ |
D | write.c | 125 struct folio *folio = page_folio(subpage); in afs_write_end() 523 folio = page_folio(pvec.pages[i]); in afs_extend_writeback() 672 struct folio *folio = page_folio(subpage); in afs_writepage() 716 folio = page_folio(head_page); in afs_writepages_region() 894 struct folio *folio = page_folio(vmf->page); in afs_page_mkwrite()
|
/linux-6.1.9/fs/ext4/ |
D | readpage.c | 399 block_read_full_folio(page_folio(page), ext4_get_block); in ext4_mpage_readpages()
|