Home
last modified time | relevance | path

Searched refs:page_folio (Results 1 – 25 of 69) sorted by relevance

123

/linux-6.1.9/mm/
Dfolio-compat.c14 return folio_mapping(page_folio(page)); in page_mapping()
20 return folio_unlock(page_folio(page)); in unlock_page()
26 return folio_end_writeback(page_folio(page)); in end_page_writeback()
32 return folio_wait_writeback(page_folio(page)); in wait_on_page_writeback()
38 return folio_wait_stable(page_folio(page)); in wait_for_stable_page()
44 return folio_mapped(page_folio(page)); in page_mapped()
50 folio_mark_accessed(page_folio(page)); in mark_page_accessed()
56 return folio_start_writeback(page_folio(page)); in set_page_writeback()
62 return folio_mark_dirty(page_folio(page)); in set_page_dirty()
68 return filemap_dirty_folio(page_mapping(page), page_folio(page)); in __set_page_dirty_nobuffers()
[all …]
Dmigrate_device.c224 folio_mark_dirty(page_folio(page)); in migrate_vma_collect_pmd()
401 folio = page_folio(page); in migrate_device_unmap()
427 folio = page_folio(page); in migrate_device_unmap()
621 if (mem_cgroup_charge(page_folio(page), vma->vm_mm, GFP_KERNEL)) in migrate_vma_insert_page()
767 r = migrate_folio_extra(mapping, page_folio(newpage), in __migrate_device_pages()
768 page_folio(page), in __migrate_device_pages()
771 r = migrate_folio(mapping, page_folio(newpage), in __migrate_device_pages()
772 page_folio(page), MIGRATE_SYNC_NO_COPY); in __migrate_device_pages()
852 src = page_folio(page); in migrate_device_finalize()
853 dst = page_folio(newpage); in migrate_device_finalize()
Dmlock.c67 lruvec = folio_lruvec_relock_irq(page_folio(page), lruvec); in __mlock_page()
107 lruvec = folio_lruvec_relock_irq(page_folio(page), lruvec); in __mlock_new_page()
131 lruvec = folio_lruvec_relock_irq(page_folio(page), lruvec); in __munlock_page()
325 mlock_folio(page_folio(page)); in mlock_pte_range()
341 mlock_folio(page_folio(page)); in mlock_pte_range()
Dgup.c52 struct folio *folio = page_folio(page); in sanity_check_pinned_pages()
74 folio = page_folio(page); in try_get_folio()
89 if (unlikely(page_folio(page) != folio)) { in try_get_folio()
211 struct folio *folio = page_folio(page); in try_grab_page()
250 gup_put_folio(page_folio(page), 1, FOLL_PIN); in unpin_user_page()
258 struct folio *folio = page_folio(next); in gup_folio_range_next()
272 struct folio *folio = page_folio(list[i]); in gup_folio_next()
276 if (page_folio(list[nr]) != folio) in gup_folio_next()
1957 struct folio *folio = page_folio(pages[i]); in collect_longterm_unpinnable_pages()
2007 struct folio *folio = page_folio(pages[i]); in migrate_longterm_unpinnable_pages()
Dswap.c732 struct folio *folio = page_folio(page); in deactivate_page()
755 struct folio *folio = page_folio(page); in mark_page_lazyfree()
986 struct folio *folio = page_folio(pages[i]); in release_pages()
Dpage_idle.c88 struct folio *folio = page_folio(page); in page_idle_clear_pte_refs()
Ddebug.c51 struct folio *folio = page_folio(page); in __dump_page()
Dmigrate.c1163 struct folio *dst, *src = page_folio(page); in unmap_and_move()
1181 dst = page_folio(newpage); in unmap_and_move()
1255 struct folio *dst, *src = page_folio(hpage); in unmap_and_move_huge_page()
1281 dst = page_folio(new_hpage); in unmap_and_move_huge_page()
1608 struct folio *folio = page_folio(page); in alloc_migration_target()
Dtruncate.c266 return truncate_inode_folio(mapping, page_folio(page)); in generic_error_remove_page()
297 struct folio *folio = page_folio(page); in invalidate_inode_page()
Dhuge_memory.c662 if (mem_cgroup_charge(page_folio(page), vma->vm_mm, gfp)) { in __do_huge_pmd_anonymous_page()
1330 folio = page_folio(page); in do_huge_pmd_wp_page()
2292 if (folio && folio != page_folio(pmd_page(*pmd))) in __split_huge_pmd()
2467 if (!folio_test_swapcache(page_folio(head))) { in __split_huge_page_tail()
2505 struct folio *folio = page_folio(page); in __split_huge_page()
2533 struct folio *tail = page_folio(head + i); in __split_huge_page()
2635 struct folio *folio = page_folio(page); in split_huge_page_to_list()
3024 if (!can_split_folio(page_folio(page), NULL)) in split_huge_pages_pid()
Dmemory-failure.c503 struct folio *folio = page_folio(page); in collect_procs_anon()
813 mem_cgroup_uncharge(page_folio(p)); in delete_from_lru_cache()
1058 struct folio *folio = page_folio(p); in me_swapcache_clean()
1403 struct folio *folio = page_folio(hpage); in hwpoison_user_mappings()
Drmap.c749 struct folio *folio = page_folio(page); in page_address_in_vma()
1101 struct folio *folio = page_folio(page); in page_move_anon_rmap()
1164 struct folio *folio = page_folio(page); in __page_check_anon_rmap()
2352 struct folio *folio = page_folio(pages[i]); in make_device_exclusive_range()
/linux-6.1.9/include/linux/
Dpage_idle.h125 return folio_test_young(page_folio(page)); in page_is_young()
130 folio_set_young(page_folio(page)); in set_page_young()
135 return folio_test_clear_young(page_folio(page)); in test_and_clear_page_young()
140 return folio_test_idle(page_folio(page)); in page_is_idle()
145 folio_set_idle(page_folio(page)); in set_page_idle()
150 folio_clear_idle(page_folio(page)); in clear_page_idle()
Dnetfs.h93 folio_start_fscache(page_folio(page)); in set_page_fscache()
98 folio_end_private_2(page_folio(page)); in end_page_fscache()
103 folio_wait_private_2(page_folio(page)); in wait_on_page_fscache()
108 return folio_wait_private_2_killable(page_folio(page)); in wait_on_page_fscache_killable()
Dpagemap.h374 return folio_file_mapping(page_folio(page)); in page_file_mapping()
382 struct folio *folio = page_folio(page); in page_mapping_file()
461 folio_attach_private(page_folio(page), data); in attach_page_private()
466 return folio_detach_private(page_folio(page)); in detach_page_private()
910 return folio_trylock(page_folio(page)); in trylock_page()
958 folio = page_folio(page); in lock_page()
988 return folio_lock_killable(page_folio(page)); in lock_page_killable()
1034 folio_wait_locked(page_folio(page)); in wait_on_page_locked()
1039 return folio_wait_locked_killable(page_folio(page)); in wait_on_page_locked_killable()
1053 __folio_mark_dirty(page_folio(page), mapping, warn); in __set_page_dirty()
[all …]
Dpage-flags.h275 #define page_folio(p) (_Generic((p), \ macro
549 return folio_test_swapcache(page_folio(page)); in PageSwapCache()
663 return folio_test_anon(page_folio(page)); in PageAnon()
693 return folio_test_ksm(page_folio(page)); in PageKsm()
730 return folio_test_uptodate(page_folio(page)); in PageUptodate()
772 folio_start_writeback_keepwrite(page_folio(page)); in set_page_writeback_keepwrite()
Dmm_inline.h34 return folio_is_file_lru(page_folio(page)); in page_is_file_lru()
329 lruvec_add_folio(lruvec, page_folio(page)); in add_page_to_lru_list()
363 lruvec_del_folio(lruvec, page_folio(page)); in del_page_from_lru_list()
Dmm.h836 return folio_entire_mapcount(page_folio(page)); in compound_mapcount()
871 return folio_mapcount(page_folio(page)); in total_mapcount()
892 return page_folio(page); in virt_to_folio()
1129 folio_get(page_folio(page)); in get_page()
1203 struct folio *folio = page_folio(page); in put_page()
1529 return page_folio(pfn_to_page(pfn)); in pfn_folio()
1581 return folio_maybe_dma_pinned(page_folio(page)); in page_maybe_dma_pinned()
Dwriteback.h388 folio_account_redirty(page_folio(page)); in account_page_redirty()
Dpage_ref.h94 return folio_ref_count(page_folio(page)); in page_count()
Drmap.h493 return folio_mkclean(page_folio(page)); in page_mkclean()
/linux-6.1.9/mm/damon/
Dpaddr.c46 folio = page_folio(page); in damon_pa_mkold()
140 folio = page_folio(page); in damon_pa_young()
/linux-6.1.9/fs/9p/
Dvfs_addr.c201 struct folio *folio = page_folio(page); in v9fs_vfs_writepage()
302 struct folio *folio = page_folio(subpage); in v9fs_write_end()
/linux-6.1.9/fs/afs/
Dwrite.c125 struct folio *folio = page_folio(subpage); in afs_write_end()
523 folio = page_folio(pvec.pages[i]); in afs_extend_writeback()
672 struct folio *folio = page_folio(subpage); in afs_writepage()
716 folio = page_folio(head_page); in afs_writepages_region()
894 struct folio *folio = page_folio(vmf->page); in afs_page_mkwrite()
/linux-6.1.9/fs/ext4/
Dreadpage.c399 block_read_full_folio(page_folio(page), ext4_get_block); in ext4_mpage_readpages()

123