/linux-6.6.21/include/linux/ |
D | pagemap.h | 22 pgoff_t start, pgoff_t end); 32 pgoff_t start, pgoff_t end); 553 pgoff_t page_cache_next_miss(struct address_space *mapping, 554 pgoff_t index, unsigned long max_scan); 555 pgoff_t page_cache_prev_miss(struct address_space *mapping, 556 pgoff_t index, unsigned long max_scan); 616 void *filemap_get_entry(struct address_space *mapping, pgoff_t index); 617 struct folio *__filemap_get_folio(struct address_space *mapping, pgoff_t index, 619 struct page *pagecache_get_page(struct address_space *mapping, pgoff_t index, 634 pgoff_t index) in filemap_get_folio() [all …]
|
D | shmem_fs.h | 26 pgoff_t fallocend; /* highest fallocate endindex */ 109 pgoff_t index, gfp_t gfp_mask); 113 extern bool shmem_is_huge(struct inode *inode, pgoff_t index, bool shmem_huge_force, 124 pgoff_t start, pgoff_t end); 135 int shmem_get_folio(struct inode *inode, pgoff_t index, struct folio **foliop, 138 pgoff_t index, gfp_t gfp); 141 pgoff_t index) in shmem_read_folio() 147 struct address_space *mapping, pgoff_t index) in shmem_read_mapping_page() 169 static inline pgoff_t shmem_fallocend(struct inode *inode, pgoff_t eof) in shmem_fallocend()
|
D | swapops.h | 86 static inline swp_entry_t swp_entry(unsigned long type, pgoff_t offset) in swp_entry() 107 static inline pgoff_t swp_offset(swp_entry_t entry) in swp_offset() 168 static inline swp_entry_t make_readable_device_private_entry(pgoff_t offset) in make_readable_device_private_entry() 173 static inline swp_entry_t make_writable_device_private_entry(pgoff_t offset) in make_writable_device_private_entry() 189 static inline swp_entry_t make_readable_device_exclusive_entry(pgoff_t offset) in make_readable_device_exclusive_entry() 194 static inline swp_entry_t make_writable_device_exclusive_entry(pgoff_t offset) in make_writable_device_exclusive_entry() 210 static inline swp_entry_t make_readable_device_private_entry(pgoff_t offset) in make_readable_device_private_entry() 215 static inline swp_entry_t make_writable_device_private_entry(pgoff_t offset) in make_writable_device_private_entry() 230 static inline swp_entry_t make_readable_device_exclusive_entry(pgoff_t offset) in make_readable_device_exclusive_entry() 235 static inline swp_entry_t make_writable_device_exclusive_entry(pgoff_t offset) in make_writable_device_exclusive_entry() [all …]
|
D | dax.h | 28 long (*direct_access)(struct dax_device *, pgoff_t, long, 37 int (*zero_page_range)(struct dax_device *, pgoff_t, size_t); 42 size_t (*recovery_write)(struct dax_device *dax_dev, pgoff_t pgoff, 67 size_t dax_recovery_write(struct dax_device *dax_dev, pgoff_t pgoff, 121 pgoff_t pgoff, void *addr, size_t bytes, struct iov_iter *i) in dax_recovery_write() 174 …ct page *dax_layout_busy_page_range(struct address_space *mapping, pgoff_t start, pgoff_t nr_pages) in dax_layout_busy_page_range() 230 long dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, long nr_pages, 232 size_t dax_copy_from_iter(struct dax_device *dax_dev, pgoff_t pgoff, void *addr, 234 size_t dax_copy_to_iter(struct dax_device *dax_dev, pgoff_t pgoff, void *addr, 236 int dax_zero_page_range(struct dax_device *dax_dev, pgoff_t pgoff, [all …]
|
D | zswap.h | 15 void zswap_invalidate(int type, pgoff_t offset); 31 static inline void zswap_invalidate(int type, pgoff_t offset) {} in zswap_invalidate()
|
D | writeback.h | 368 pgoff_t start, pgoff_t end); 375 pgoff_t start, pgoff_t end);
|
D | pagewalk.h | 129 int walk_page_mapping(struct address_space *mapping, pgoff_t first_index, 130 pgoff_t nr, const struct mm_walk_ops *ops,
|
/linux-6.6.21/mm/ |
D | truncate.c | 32 pgoff_t index, void *entry) in __clear_shadow_entry() 42 static void clear_shadow_entry(struct address_space *mapping, pgoff_t index, in clear_shadow_entry() 60 struct folio_batch *fbatch, pgoff_t *indices) in truncate_folio_batch_exceptionals() 84 pgoff_t index = indices[i]; in truncate_folio_batch_exceptionals() 113 pgoff_t index, void *entry) in invalidate_exceptional_entry() 127 pgoff_t index, void *entry) in invalidate_exceptional_entry2() 332 pgoff_t start; /* inclusive */ in truncate_inode_pages_range() 333 pgoff_t end; /* exclusive */ in truncate_inode_pages_range() 335 pgoff_t indices[PAGEVEC_SIZE]; in truncate_inode_pages_range() 336 pgoff_t index; in truncate_inode_pages_range() [all …]
|
D | mapping_dirty_helpers.c | 67 pgoff_t bitmap_pgoff; 69 pgoff_t start; 70 pgoff_t end; 97 pgoff_t pgoff = ((addr - walk->vma->vm_start) >> PAGE_SHIFT) + in clean_record_pte() 264 pgoff_t first_index, pgoff_t nr) in wp_shared_mapping_range() 314 pgoff_t first_index, pgoff_t nr, in clean_record_shared_mapping_range() 315 pgoff_t bitmap_pgoff, in clean_record_shared_mapping_range() 317 pgoff_t *start, in clean_record_shared_mapping_range() 318 pgoff_t *end) in clean_record_shared_mapping_range()
|
D | readahead.c | 287 pgoff_t end_index; /* The last page we want to read */ in do_page_cache_ra() 418 static pgoff_t count_history_pages(struct address_space *mapping, in count_history_pages() 419 pgoff_t index, unsigned long max) in count_history_pages() 421 pgoff_t head; in count_history_pages() 435 pgoff_t index, in try_context_readahead() 439 pgoff_t size; in try_context_readahead() 464 static inline int ra_alloc_folio(struct readahead_control *ractl, pgoff_t index, in ra_alloc_folio() 465 pgoff_t mark, unsigned int order, gfp_t gfp) in ra_alloc_folio() 490 pgoff_t index = readahead_index(ractl); in page_cache_ra_order() 491 pgoff_t limit = (i_size_read(mapping->host) - 1) >> PAGE_SHIFT; in page_cache_ra_order() [all …]
|
D | filemap.c | 477 pgoff_t max = end_byte >> PAGE_SHIFT; in filemap_range_has_page() 506 pgoff_t index = start_byte >> PAGE_SHIFT; in __filemap_fdatawait_range() 507 pgoff_t end = end_byte >> PAGE_SHIFT; in __filemap_fdatawait_range() 635 pgoff_t max = end_byte >> PAGE_SHIFT; in filemap_range_has_writeback() 811 pgoff_t offset = old->index; in replace_page_cache_folio() 845 struct folio *folio, pgoff_t index, gfp_t gfp, void **shadowp) in __filemap_add_folio() 933 pgoff_t index, gfp_t gfp) in filemap_add_folio() 1731 pgoff_t page_cache_next_miss(struct address_space *mapping, in page_cache_next_miss() 1732 pgoff_t index, unsigned long max_scan) in page_cache_next_miss() 1767 pgoff_t page_cache_prev_miss(struct address_space *mapping, in page_cache_prev_miss() [all …]
|
D | swap_cgroup.c | 66 pgoff_t offset) in __lookup_swap_cgroup() 79 pgoff_t offset = swp_offset(ent); in lookup_swap_cgroup() 133 pgoff_t offset = swp_offset(ent); in swap_cgroup_record() 134 pgoff_t end = offset + nr_ents; in swap_cgroup_record()
|
D | folio-compat.c | 87 pgoff_t index, gfp_t gfp) in add_to_page_cache_lru() 94 struct page *pagecache_get_page(struct address_space *mapping, pgoff_t index, in pagecache_get_page() 107 pgoff_t index) in grab_cache_page_write_begin()
|
D | internal.h | 127 struct file *file, pgoff_t index, unsigned long nr_to_read) in force_page_cache_readahead() 133 unsigned find_lock_entries(struct address_space *mapping, pgoff_t *start, 134 pgoff_t end, struct folio_batch *fbatch, pgoff_t *indices); 135 unsigned find_get_entries(struct address_space *mapping, pgoff_t *start, 136 pgoff_t end, struct folio_batch *fbatch, pgoff_t *indices); 143 pgoff_t start, pgoff_t end, unsigned long *nr_failed); 640 vma_pgoff_address(pgoff_t pgoff, unsigned long nr_pages, in vma_pgoff_address() 679 pgoff_t pgoff; in vma_address_end()
|
D | shmem.c | 106 pgoff_t start; /* start of range currently being fallocated */ 107 pgoff_t next; /* the next page offset to be fallocated */ 108 pgoff_t nr_falloced; /* how many new pages have been fallocated */ 109 pgoff_t nr_unswapped; /* how often writepage refused to swap out */ 148 static int shmem_swapin_folio(struct inode *inode, pgoff_t index, 474 pgoff_t index, void *expected, void *replacement) in shmem_replace_entry() 496 pgoff_t index, swp_entry_t swap) in shmem_confirm_swap() 538 bool shmem_is_huge(struct inode *inode, pgoff_t index, bool shmem_huge_force, in shmem_is_huge() 663 pgoff_t index; in shmem_unused_huge_shrink() 745 bool shmem_is_huge(struct inode *inode, pgoff_t index, bool shmem_huge_force, in shmem_is_huge() [all …]
|
/linux-6.6.21/include/trace/events/ |
D | fs_dax.h | 12 pgoff_t max_pgoff, int result), 20 __field(pgoff_t, pgoff) 21 __field(pgoff_t, max_pgoff) 57 pgoff_t max_pgoff, int result), \ 160 __field(pgoff_t, pgoff) 228 TP_PROTO(struct inode *inode, pgoff_t start_index, pgoff_t end_index), 232 __field(pgoff_t, start_index) 233 __field(pgoff_t, end_index) 253 TP_PROTO(struct inode *inode, pgoff_t start_index, pgoff_t end_index),\ 260 TP_PROTO(struct inode *inode, pgoff_t pgoff, pgoff_t pglen), [all …]
|
/linux-6.6.21/drivers/gpu/drm/vmwgfx/ |
D | vmwgfx_page_dirty.c | 85 pgoff_t offset = drm_vma_node_start(&vbo->tbo.base.vma_node); in vmw_bo_dirty_scan_pagetable() 87 pgoff_t num_marked; in vmw_bo_dirty_scan_pagetable() 125 pgoff_t num_marked; in vmw_bo_dirty_scan_mkwrite() 141 pgoff_t start = 0; in vmw_bo_dirty_scan_mkwrite() 142 pgoff_t end = dirty->bitmap_size; in vmw_bo_dirty_scan_mkwrite() 185 pgoff_t start, pgoff_t end) in vmw_bo_dirty_pre_unmap() 210 pgoff_t start, pgoff_t end) in vmw_bo_dirty_unmap() 233 pgoff_t num_pages = PFN_UP(vbo->tbo.resource->size); in vmw_bo_dirty_add() 257 pgoff_t offset = drm_vma_node_start(&vbo->tbo.base.vma_node); in vmw_bo_dirty_add() 311 pgoff_t start, cur, end; in vmw_bo_dirty_transfer_to_res() [all …]
|
/linux-6.6.21/fs/smb/client/ |
D | fscache.h | 78 pgoff_t first, unsigned int nr_pages, 79 pgoff_t *_data_first, 83 pgoff_t first, unsigned int nr_pages, in cifs_fscache_query_occupancy() 84 pgoff_t *_data_first, in cifs_fscache_query_occupancy() 129 pgoff_t first, unsigned int nr_pages, in cifs_fscache_query_occupancy() 130 pgoff_t *_data_first, in cifs_fscache_query_occupancy()
|
/linux-6.6.21/drivers/gpu/drm/i915/gem/ |
D | i915_gem_object.h | 378 pgoff_t n, 402 static_assert(castable_to_type(n, pgoff_t)); \ 425 __i915_gem_object_get_sg(struct drm_i915_gem_object *obj, pgoff_t n, in __i915_gem_object_get_sg() 447 static_assert(castable_to_type(n, pgoff_t)); \ 470 __i915_gem_object_get_sg_dma(struct drm_i915_gem_object *obj, pgoff_t n, in __i915_gem_object_get_sg_dma() 492 static_assert(castable_to_type(n, pgoff_t)); \ 512 __i915_gem_object_get_page(struct drm_i915_gem_object *obj, pgoff_t n); 528 static_assert(castable_to_type(n, pgoff_t)); \ 547 __i915_gem_object_get_dirty_page(struct drm_i915_gem_object *obj, pgoff_t n); 563 static_assert(castable_to_type(n, pgoff_t)); \ [all …]
|
/linux-6.6.21/fs/nilfs2/ |
D | page.c | 29 __nilfs_get_page_block(struct page *page, unsigned long block, pgoff_t index, in __nilfs_get_page_block() 53 pgoff_t index = blkoff >> (PAGE_SHIFT - blkbits); in nilfs_grab_buffer() 245 pgoff_t index = 0; in nilfs_copy_dirty_pages() 250 if (!filemap_get_folios_tag(smap, &index, (pgoff_t)-1, in nilfs_copy_dirty_pages() 300 pgoff_t start = 0; in nilfs_copy_back_pages() 310 pgoff_t index = folio->index; in nilfs_copy_back_pages() 363 pgoff_t index = 0; in nilfs_clear_dirty_pages() 367 while (filemap_get_folios_tag(mapping, &index, (pgoff_t)-1, in nilfs_clear_dirty_pages() 493 pgoff_t index; in nilfs_find_uncommitted_extent()
|
/linux-6.6.21/drivers/md/ |
D | dm-linear.c | 159 static struct dax_device *linear_dax_pgoff(struct dm_target *ti, pgoff_t *pgoff) in linear_dax_pgoff() 168 static long linear_dax_direct_access(struct dm_target *ti, pgoff_t pgoff, in linear_dax_direct_access() 177 static int linear_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff, in linear_dax_zero_page_range() 185 static size_t linear_dax_recovery_write(struct dm_target *ti, pgoff_t pgoff, in linear_dax_recovery_write()
|
/linux-6.6.21/fs/f2fs/ |
D | node.h | 196 static inline pgoff_t current_nat_addr(struct f2fs_sb_info *sbi, nid_t start) in current_nat_addr() 199 pgoff_t block_off; in current_nat_addr() 200 pgoff_t block_addr; in current_nat_addr() 209 block_addr = (pgoff_t)(nm_i->nat_blkaddr + in current_nat_addr() 219 static inline pgoff_t next_nat_addr(struct f2fs_sb_info *sbi, in next_nat_addr() 220 pgoff_t block_addr) in next_nat_addr()
|
D | file.c | 393 pgoff_t index, int whence) in __found_offset() 416 pgoff_t pgofs, end_offset; in f2fs_seek_block() 438 pgofs = (pgoff_t)(offset >> PAGE_SHIFT); in f2fs_seek_block() 615 pgoff_t fofs; in f2fs_truncate_data_blocks_range() 637 pgoff_t index = from >> PAGE_SHIFT; in truncate_partial_data_page() 671 pgoff_t free_from; in f2fs_do_truncate_blocks() 678 free_from = (pgoff_t)F2FS_BLK_ALIGN(from); in f2fs_do_truncate_blocks() 1052 static int fill_zero(struct inode *inode, pgoff_t index, in fill_zero() 1077 int f2fs_truncate_hole(struct inode *inode, pgoff_t pg_start, pgoff_t pg_end) in f2fs_truncate_hole() 1083 pgoff_t end_offset, count; in f2fs_truncate_hole() [all …]
|
/linux-6.6.21/fs/ |
D | dax.c | 125 pgoff_t entry_start; 490 dax_entry_t dax_lock_mapping_entry(struct address_space *mapping, pgoff_t index, in dax_lock_mapping_entry() 533 void dax_unlock_mapping_entry(struct address_space *mapping, pgoff_t index, in dax_unlock_mapping_entry() 683 pgoff_t start_idx = start >> PAGE_SHIFT; in dax_layout_busy_page_range() 684 pgoff_t end_idx; in dax_layout_busy_page_range() 746 pgoff_t index, bool trunc) in __dax_invalidate_entry() 771 pgoff_t start, pgoff_t end) in __dax_clear_dirty_range() 801 int dax_delete_mapping_entry(struct address_space *mapping, pgoff_t index) in dax_delete_mapping_entry() 820 pgoff_t index) in dax_invalidate_mapping_entry_sync() 825 static pgoff_t dax_iomap_pgoff(const struct iomap *iomap, loff_t pos) in dax_iomap_pgoff() [all …]
|
/linux-6.6.21/include/drm/ttm/ |
D | ttm_resource.h | 280 pgoff_t i; 281 pgoff_t end; 282 pgoff_t offs;
|