/linux-6.6.21/fs/freevxfs/ |
D | vxfs_immed.c | 38 memcpy_to_page(folio_page(folio, i), 0, src, PAGE_SIZE); in vxfs_immed_read_folio()
|
/linux-6.6.21/arch/openrisc/mm/ |
D | cache.c | 58 sync_icache_dcache(folio_page(folio, nr)); in update_cache()
|
/linux-6.6.21/arch/arm/mm/ |
D | flush.c | 220 void *addr = kmap_high_get(folio_page(folio, i)); in __flush_dcache_folio() 223 kunmap_high(folio_page(folio, i)); in __flush_dcache_folio()
|
/linux-6.6.21/mm/ |
D | util.c | 806 copy_highpage(folio_page(dst, i), folio_page(src, i)); in folio_copy() 1143 flush_dcache_page(folio_page(folio, i)); in flush_dcache_folio()
|
D | rmap.c | 1089 mapcount += atomic_read(&folio_page(folio, i)->_mapcount); in folio_total_mapcount() 1538 subpage = folio_page(folio, pfn - folio_pfn(folio)); in try_to_unmap_one() 1874 subpage = folio_page(folio, in try_to_migrate_one() 1907 subpage = folio_page(folio, pfn - folio_pfn(folio)); in try_to_migrate_one() 2213 subpage = folio_page(folio, in page_make_device_exclusive_one()
|
D | slab_common.c | 1001 mod_lruvec_page_state(folio_page(folio, 0), NR_SLAB_UNRECLAIMABLE_B, in free_large_kmalloc() 1003 __free_pages(folio_page(folio, 0), order); in free_large_kmalloc()
|
D | slab.h | 170 #define slab_page(s) folio_page(slab_folio(s), 0)
|
D | hugetlb.c | 1530 p = folio_page(folio, i); in __destroy_compound_gigantic_folio() 1997 p = folio_page(folio, i); in __prep_compound_gigantic_folio() 2052 p = folio_page(folio, j); in __prep_compound_gigantic_folio() 2059 p = folio_page(folio, j); in __prep_compound_gigantic_folio() 3648 subpage = folio_page(folio, i); in demote_free_hugetlb_folio()
|
D | shmem.c | 2112 clear_highpage(folio_page(folio, i)); in shmem_get_folio_gfp() 3538 if (PageHWPoison(folio_page(folio, 0)) || in shmem_get_link() 3549 if (PageHWPoison(folio_page(folio, 0))) { in shmem_get_link()
|
D | filemap.c | 2853 page = folio_page(folio, offset / PAGE_SIZE); in splice_folio_into_pipe() 3491 struct page *page = folio_page(folio, start); in filemap_map_folio_range()
|
D | memory.c | 6033 dst_page = folio_page(dst, i); in copy_user_gigantic_page() 6034 src_page = folio_page(src, i); in copy_user_gigantic_page() 6094 subpage = folio_page(dst_folio, i); in copy_folio_from_user()
|
D | migrate.c | 199 new = folio_page(folio, idx); in remove_migration_pte()
|
D | memcontrol.c | 3424 folio_page(folio, i)->memcg_data = folio->memcg_data; in split_page_memcg()
|
/linux-6.6.21/include/linux/ |
D | highmem-internal.h | 78 struct page *page = folio_page(folio, offset / PAGE_SIZE); in kmap_local_folio()
|
D | nfs_page.h | 199 return folio_page(folio, pgbase >> PAGE_SHIFT); in nfs_page_to_page()
|
D | mm.h | 2091 return (struct folio *)folio_page(folio, folio_nr_pages(folio)); in folio_next() 2139 return page_mapcount(folio_page(folio, 0)); in folio_estimated_sharers() 2156 ret = arch_make_page_accessible(folio_page(folio, i)); in arch_make_folio_accessible()
|
D | page-flags.h | 282 #define folio_page(folio, n) nth_page(&(folio)->page, n) macro
|
D | pagemap.h | 812 return folio_page(folio, index & (folio_nr_pages(folio) - 1)); in folio_file_page()
|
/linux-6.6.21/fs/crypto/ |
D | crypto.c | 272 struct page *page = folio_page(folio, i >> PAGE_SHIFT); in fscrypt_decrypt_pagecache_blocks()
|
/linux-6.6.21/fs/smb/client/ |
D | cifsencrypt.c | 131 p = kmap_local_page(folio_page(folio, j)); in cifs_shash_xarray()
|
D | smbdirect.c | 2554 if (!smb_set_sge(rdma, folio_page(folio, 0), off, len)) { in smb_extract_xarray_to_rdma()
|
/linux-6.6.21/lib/ |
D | scatterlist.c | 1296 sg_set_page(sg, folio_page(folio, 0), len, offset); in extract_xarray_to_sg()
|
/linux-6.6.21/fs/ceph/ |
D | addr.c | 1475 snapc = ceph_find_incompatible(folio_page(*foliop, 0)); in ceph_netfs_check_write_begin() 1895 pages[0] = folio_page(folio, 0); in ceph_uninline_data()
|
/linux-6.6.21/block/ |
D | bio.c | 1159 page = folio_page(fi.folio, fi.offset / PAGE_SIZE); in __bio_release_pages()
|
/linux-6.6.21/fs/f2fs/ |
D | data.c | 3116 pages[nr_pages] = folio_page(folio, idx); in f2fs_write_cache_pages()
|