/linux-6.1.9/fs/nilfs2/ |
D | btnode.c | 198 xa_lock_irq(&btnc->i_pages); in nilfs_btnode_prepare_change_key() 199 err = __xa_insert(&btnc->i_pages, newkey, opage, GFP_NOFS); in nilfs_btnode_prepare_change_key() 200 xa_unlock_irq(&btnc->i_pages); in nilfs_btnode_prepare_change_key() 255 xa_lock_irq(&btnc->i_pages); in nilfs_btnode_commit_change_key() 256 __xa_erase(&btnc->i_pages, oldkey); in nilfs_btnode_commit_change_key() 257 __xa_set_mark(&btnc->i_pages, newkey, PAGECACHE_TAG_DIRTY); in nilfs_btnode_commit_change_key() 258 xa_unlock_irq(&btnc->i_pages); in nilfs_btnode_commit_change_key() 286 xa_erase_irq(&btnc->i_pages, newkey); in nilfs_btnode_abort_change_key()
|
D | page.c | 324 xa_lock_irq(&smap->i_pages); in nilfs_copy_back_pages() 325 f = __xa_erase(&smap->i_pages, index); in nilfs_copy_back_pages() 328 xa_unlock_irq(&smap->i_pages); in nilfs_copy_back_pages() 330 xa_lock_irq(&dmap->i_pages); in nilfs_copy_back_pages() 331 f = __xa_store(&dmap->i_pages, index, folio, GFP_NOFS); in nilfs_copy_back_pages() 340 __xa_set_mark(&dmap->i_pages, index, in nilfs_copy_back_pages() 343 xa_unlock_irq(&dmap->i_pages); in nilfs_copy_back_pages() 454 xa_lock_irq(&mapping->i_pages); in __nilfs_clear_page_dirty() 456 __xa_clear_mark(&mapping->i_pages, page_index(page), in __nilfs_clear_page_dirty() 458 xa_unlock_irq(&mapping->i_pages); in __nilfs_clear_page_dirty() [all …]
|
/linux-6.1.9/mm/ |
D | truncate.c | 35 XA_STATE(xas, &mapping->i_pages, index); in __clear_shadow_entry() 47 xa_lock_irq(&mapping->i_pages); in clear_shadow_entry() 49 xa_unlock_irq(&mapping->i_pages); in clear_shadow_entry() 80 xa_lock_irq(&mapping->i_pages); in truncate_folio_batch_exceptionals() 101 xa_unlock_irq(&mapping->i_pages); in truncate_folio_batch_exceptionals() 483 xa_lock_irq(&mapping->i_pages); in truncate_inode_pages_final() 484 xa_unlock_irq(&mapping->i_pages); in truncate_inode_pages_final() 589 xa_lock_irq(&mapping->i_pages); in invalidate_complete_folio2() 595 xa_unlock_irq(&mapping->i_pages); in invalidate_complete_folio2() 603 xa_unlock_irq(&mapping->i_pages); in invalidate_complete_folio2()
|
D | workingset.c | 553 mapping = container_of(node->array, struct address_space, i_pages); in workingset_update_node() 554 lockdep_assert_held(&mapping->i_pages.xa_lock); in workingset_update_node() 647 mapping = container_of(node->array, struct address_space, i_pages); in shadow_lru_isolate() 650 if (!xa_trylock(&mapping->i_pages)) { in shadow_lru_isolate() 657 xa_unlock(&mapping->i_pages); in shadow_lru_isolate() 681 xa_unlock_irq(&mapping->i_pages); in shadow_lru_isolate()
|
D | swap_state.c | 78 page = xa_load(&address_space->i_pages, idx); in get_shadow_from_swap_cache() 93 XA_STATE_ORDER(xas, &address_space->i_pages, idx, folio_order(folio)); in add_to_swap_cache() 146 XA_STATE(xas, &address_space->i_pages, idx); in __delete_from_swap_cache() 236 xa_lock_irq(&address_space->i_pages); in delete_from_swap_cache() 238 xa_unlock_irq(&address_space->i_pages); in delete_from_swap_cache() 253 XA_STATE(xas, &address_space->i_pages, curr); in clear_shadow_from_swap_cache() 255 xa_lock_irq(&address_space->i_pages); in clear_shadow_from_swap_cache() 261 xa_unlock_irq(&address_space->i_pages); in clear_shadow_from_swap_cache() 675 xa_init_flags(&space->i_pages, XA_FLAGS_LOCK_IRQ); in init_swap_address_space()
|
D | filemap.c | 127 XA_STATE(xas, &mapping->i_pages, folio->index); in page_cache_delete() 254 xa_lock_irq(&mapping->i_pages); in filemap_remove_folio() 256 xa_unlock_irq(&mapping->i_pages); in filemap_remove_folio() 280 XA_STATE(xas, &mapping->i_pages, fbatch->folios[0]->index); in page_cache_delete_batch() 327 xa_lock_irq(&mapping->i_pages); in delete_from_page_cache_batch() 335 xa_unlock_irq(&mapping->i_pages); in delete_from_page_cache_batch() 474 XA_STATE(xas, &mapping->i_pages, start_byte >> PAGE_SHIFT); in filemap_range_has_page() 633 XA_STATE(xas, &mapping->i_pages, start_byte >> PAGE_SHIFT); in filemap_range_has_writeback() 807 XA_STATE(xas, &mapping->i_pages, offset); in replace_page_cache_page() 842 XA_STATE(xas, &mapping->i_pages, index); in __filemap_add_folio() [all …]
|
D | readahead.c | 231 struct folio *folio = xa_load(&mapping->i_pages, index + i); in page_cache_ra_unbounded() 804 struct page *page = xa_load(&mapping->i_pages, index); in readahead_expand() 827 struct page *page = xa_load(&mapping->i_pages, index); in readahead_expand()
|
D | page-writeback.c | 2232 XA_STATE(xas, &mapping->i_pages, start); in tag_pages_for_writeback() 2613 xa_lock_irqsave(&mapping->i_pages, flags); in __folio_mark_dirty() 2617 __xa_set_mark(&mapping->i_pages, folio_index(folio), in __folio_mark_dirty() 2620 xa_unlock_irqrestore(&mapping->i_pages, flags); in __folio_mark_dirty() 2923 xa_lock_irqsave(&mapping->i_pages, flags); in __folio_end_writeback() 2926 __xa_clear_mark(&mapping->i_pages, folio_index(folio), in __folio_end_writeback() 2943 xa_unlock_irqrestore(&mapping->i_pages, flags); in __folio_end_writeback() 2965 XA_STATE(xas, &mapping->i_pages, folio_index(folio)); in __folio_start_writeback()
|
D | shmem.c | 403 XA_STATE(xas, &mapping->i_pages, index); in shmem_replace_entry() 425 return xa_load(&mapping->i_pages, index) == swp_to_radix_entry(swap); in shmem_confirm_swap() 694 XA_STATE_ORDER(xas, &mapping->i_pages, index, folio_order(folio)); in shmem_add_to_page_cache() 764 xa_lock_irq(&mapping->i_pages); in shmem_delete_from_page_cache() 770 xa_unlock_irq(&mapping->i_pages); in shmem_delete_from_page_cache() 783 old = xa_cmpxchg_irq(&mapping->i_pages, index, radswap, NULL, 0); in shmem_free_swap() 800 XA_STATE(xas, &mapping->i_pages, start); in shmem_partial_swap_usage() 1183 XA_STATE(xas, &mapping->i_pages, start); in shmem_find_swap_entries() 1551 if (xa_find(&mapping->i_pages, &hindex, hindex + HPAGE_PMD_NR - 1, in shmem_alloc_hugefolio() 1662 xa_lock_irq(&swap_mapping->i_pages); in shmem_replace_folio() [all …]
|
D | memfd.c | 76 XA_STATE(xas, &mapping->i_pages, 0); in memfd_wait_for_pins()
|
/linux-6.1.9/arch/nios2/include/asm/ |
D | cacheflush.h | 49 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages) 50 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages)
|
/linux-6.1.9/arch/csky/abiv1/inc/abi/ |
D | cacheflush.h | 17 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages) 18 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages)
|
/linux-6.1.9/arch/parisc/include/asm/ |
D | cacheflush.h | 49 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages) 50 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages)
|
/linux-6.1.9/include/linux/ |
D | backing-dev.h | 245 !lockdep_is_held(&inode->i_mapping->i_pages.xa_lock) && in inode_to_wb() 289 xa_lock_irqsave(&inode->i_mapping->i_pages, cookie->flags); in unlocked_inode_to_wb_begin() 307 xa_unlock_irqrestore(&inode->i_mapping->i_pages, cookie->flags); in unlocked_inode_to_wb_end()
|
D | pagemap.h | 137 return xa_empty(&mapping->i_pages); in mapping_empty() 174 head = rcu_access_pointer(mapping->i_pages.xa_head); in mapping_shrinkable() 1252 folio = xa_load(&ractl->mapping->i_pages, ractl->_index); in __readahead_folio() 1296 XA_STATE(xas, &rac->mapping->i_pages, 0); in __readahead_batch()
|
/linux-6.1.9/Documentation/translations/zh_CN/mm/ |
D | page_migration.rst | 84 5. i_pages的锁被持有。这将导致所有试图通过映射访问该页的进程在自旋锁上阻塞。 99 11. i_pages锁被放弃。这样一来,在映射中的查找又变得可能了。进程将从在锁上自旋到在
|
/linux-6.1.9/arch/arm/include/asm/ |
D | cacheflush.h | 316 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages) 317 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages)
|
/linux-6.1.9/fs/netfs/ |
D | io.c | 26 iov_iter_xarray(&iter, READ, &subreq->rreq->mapping->i_pages, in netfs_clear_unread() 52 iov_iter_xarray(&iter, READ, &rreq->mapping->i_pages, in netfs_read_from_cache() 121 XA_STATE(xas, &rreq->mapping->i_pages, subreq->start / PAGE_SIZE); in netfs_rreq_unmark_after_write() 211 iov_iter_xarray(&iter, WRITE, &rreq->mapping->i_pages, in netfs_rreq_do_write_to_cache()
|
/linux-6.1.9/fs/ |
D | dax.c | 456 xas.xa = &mapping->i_pages; in dax_lock_page() 481 XA_STATE(xas, &mapping->i_pages, page->index); in dax_unlock_page() 510 xas.xa = &mapping->i_pages; in dax_lock_mapping_entry() 544 XA_STATE(xas, &mapping->i_pages, index); in dax_unlock_mapping_entry() 693 XA_STATE(xas, &mapping->i_pages, start_idx); in dax_layout_busy_page_range() 756 XA_STATE(xas, &mapping->i_pages, index); in __dax_invalidate_entry() 1014 XA_STATE(xas, &mapping->i_pages, wbc->range_start >> PAGE_SHIFT); in dax_writeback_mapping_range() 1590 XA_STATE(xas, &mapping->i_pages, vmf->pgoff); in dax_iomap_pte_fault() 1700 XA_STATE_ORDER(xas, &mapping->i_pages, vmf->pgoff, PMD_ORDER); in dax_iomap_pmd_fault() 1824 XA_STATE_ORDER(xas, &mapping->i_pages, vmf->pgoff, order); in dax_insert_pfn_mkwrite()
|
/linux-6.1.9/fs/fscache/ |
D | io.c | 219 XA_STATE(xas, &mapping->i_pages, first); in __fscache_clear_page_bits() 289 iov_iter_xarray(&iter, WRITE, &mapping->i_pages, start, len); in __fscache_write_to_cache()
|
/linux-6.1.9/fs/erofs/ |
D | fscache.c | 70 XA_STATE(xas, &rreq->mapping->i_pages, start_page); in erofs_fscache_rreq_unlock_folios() 197 iov_iter_xarray(&iter, READ, &rreq->mapping->i_pages, in erofs_fscache_read_folios_async() 293 iov_iter_xarray(&iter, READ, &mapping->i_pages, pos, PAGE_SIZE); in erofs_fscache_data_read() 305 iov_iter_xarray(&iter, READ, &mapping->i_pages, pos, count); in erofs_fscache_data_read()
|
/linux-6.1.9/fs/9p/ |
D | vfs_addr.c | 43 iov_iter_xarray(&to, READ, &rreq->mapping->i_pages, pos, len); in v9fs_issue_read() 175 iov_iter_xarray(&from, WRITE, &folio_mapping(folio)->i_pages, start, len); in v9fs_vfs_write_folio_locked()
|
/linux-6.1.9/fs/orangefs/ |
D | inode.c | 253 struct xarray *i_pages; in orangefs_readahead() local 271 i_pages = &rac->mapping->i_pages; in orangefs_readahead() 273 iov_iter_xarray(&iter, READ, i_pages, offset, readahead_length(rac)); in orangefs_readahead()
|
/linux-6.1.9/fs/afs/ |
D | file.c | 328 &fsreq->vnode->netfs.inode.i_mapping->i_pages, in afs_issue_read() 349 iov_iter_xarray(&fsreq->def_iter, READ, &folio->mapping->i_pages, in afs_symlink_read_folio()
|
D | write.c | 264 XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE); in afs_pages_written_back() 446 XA_STATE(xas, &mapping->i_pages, index); in afs_extend_writeback() 612 iov_iter_xarray(&iter, WRITE, &mapping->i_pages, start, len); in afs_write_back_from_locked_folio()
|