/linux-2.4.37.9/include/linux/ |
D | swap.h | 97 extern void __remove_inode_page(struct page *); 107 extern void FASTCALL(lru_cache_add(struct page *)); 108 extern void FASTCALL(__lru_cache_del(struct page *)); 109 extern void FASTCALL(lru_cache_del(struct page *)); 111 extern void FASTCALL(activate_page(struct page *)); 122 extern void rw_swap_page(int, struct page *); 132 extern int add_to_swap_cache(struct page *, swp_entry_t); 133 extern void __delete_from_swap_cache(struct page *page); 134 extern void delete_from_swap_cache(struct page *page); 135 extern void free_page_and_swap_cache(struct page *page); [all …]
|
D | mm.h | 137 struct page * (*nopage)(struct vm_area_struct * area, unsigned long address, int unused); 155 typedef struct page { struct 159 struct page *next_hash; /* Next page sharing our hash bucket in argument 166 struct page **pprev_hash; /* Complement to *next_hash. */ argument 203 static inline struct page *nth_page(struct page *page, int n) in nth_page() argument 205 return page + n; in nth_page() 311 #define arch_set_page_uptodate(page) argument 315 #define UnlockPage(page) unlock_page(page) argument 316 #define Page_Uptodate(page) test_bit(PG_uptodate, &(page)->flags) argument 318 #define SetPageUptodate(page) set_bit(PG_uptodate, &(page)->flags) argument [all …]
|
D | pagemap.h | 34 static inline struct page *page_cache_alloc(struct address_space *x) in page_cache_alloc() 49 extern struct page **page_hash_table; 73 extern struct page * __find_get_page(struct address_space *mapping, 74 unsigned long index, struct page **hash); 77 extern struct page * __find_lock_page (struct address_space * mapping, 78 unsigned long index, struct page **hash); 79 extern struct page * find_or_create_page(struct address_space *mapping, 82 extern void FASTCALL(lock_page(struct page *page)); 83 extern void FASTCALL(unlock_page(struct page *page)); 86 extern struct page *find_trylock_page(struct address_space *, unsigned long); [all …]
|
D | highmem.h | 9 extern struct page *highmem_start_page; 68 static inline void *kmap(struct page *page) { return page_address(page); } in kmap() argument 70 #define kunmap(page) do { } while (0) argument 72 #define kmap_atomic(page,idx) kmap(page) argument 73 #define kunmap_atomic(page,idx) kunmap(page) argument 77 #define kmap_nonblock(page) kmap(page) argument 84 static inline void clear_user_highpage(struct page *page, unsigned long vaddr) in clear_user_highpage() argument 86 void *addr = kmap_atomic(page, KM_USER0); in clear_user_highpage() 91 static inline void clear_highpage(struct page *page) in clear_highpage() argument 93 clear_page(kmap(page)); in clear_highpage() [all …]
|
/linux-2.4.37.9/arch/arm/mm/ |
D | small_page.c | 53 struct page *queue; 68 #error unsupported page size 78 static void add_page_to_queue(struct page *page, struct page **p) in add_page_to_queue() argument 81 if (page->pprev_hash) in add_page_to_queue() 82 PAGE_BUG(page); in add_page_to_queue() 84 page->next_hash = *p; in add_page_to_queue() 86 (*p)->pprev_hash = &page->next_hash; in add_page_to_queue() 87 *p = page; in add_page_to_queue() 88 page->pprev_hash = p; in add_page_to_queue() 91 static void remove_page_from_queue(struct page *page) in remove_page_from_queue() argument [all …]
|
D | consistent.c | 42 struct page *page, *end, *free; in consistent_alloc() local 53 page = alloc_pages(gfp, order); in consistent_alloc() 54 if (!page) in consistent_alloc() 57 *dma_handle = page_to_bus(page); in consistent_alloc() 58 ret = __ioremap(page_to_pfn(page) << PAGE_SHIFT, size, 0); in consistent_alloc() 64 void *virt = page_address(page); in consistent_alloc() 81 free = page + (size >> PAGE_SHIFT); in consistent_alloc() 82 end = page + (1 << order); in consistent_alloc() 84 for (; page < end; page++) { in consistent_alloc() 85 set_page_count(page, 1); in consistent_alloc() [all …]
|
/linux-2.4.37.9/fs/sysv/ |
D | dir.c | 28 static inline void dir_put_page(struct page *page) in dir_put_page() argument 30 kunmap(page); in dir_put_page() 31 page_cache_release(page); in dir_put_page() 39 static int dir_commit_chunk(struct page *page, unsigned from, unsigned to) in dir_commit_chunk() argument 41 struct inode *dir = (struct inode *)page->mapping->host; in dir_commit_chunk() 45 page->mapping->a_ops->commit_write(NULL, page, from, to); in dir_commit_chunk() 48 err = writeout_one_page(page); in dir_commit_chunk() 49 err2 = waitfor_one_page(page); in dir_commit_chunk() 56 static struct page * dir_get_page(struct inode *dir, unsigned long n) in dir_get_page() 59 struct page *page = read_cache_page(mapping, n, in dir_get_page() local [all …]
|
/linux-2.4.37.9/mm/ |
D | filemap.c | 47 struct page **page_hash_table; 70 static void FASTCALL(add_page_to_hash_queue(struct page * page, struct page **p)); 71 static void fastcall add_page_to_hash_queue(struct page * page, struct page **p) in add_page_to_hash_queue() argument 73 struct page *next = *p; in add_page_to_hash_queue() 75 *p = page; in add_page_to_hash_queue() 76 page->next_hash = next; in add_page_to_hash_queue() 77 page->pprev_hash = p; in add_page_to_hash_queue() 79 next->pprev_hash = &page->next_hash; in add_page_to_hash_queue() 80 if (page->buffers) in add_page_to_hash_queue() 81 PAGE_BUG(page); in add_page_to_hash_queue() [all …]
|
D | swap_state.c | 24 static int swap_writepage(struct page *page) in swap_writepage() argument 26 if (remove_exclusive_swap_page(page)) { in swap_writepage() 27 UnlockPage(page); in swap_writepage() 30 rw_swap_page(WRITE, page); in swap_writepage() 70 int add_to_swap_cache(struct page *page, swp_entry_t entry) in add_to_swap_cache() argument 72 if (page->mapping) in add_to_swap_cache() 78 if (add_to_page_cache_unique(page, &swapper_space, entry.val, in add_to_swap_cache() 84 if (!PageLocked(page)) in add_to_swap_cache() 86 if (!PageSwapCache(page)) in add_to_swap_cache() 96 void __delete_from_swap_cache(struct page *page) in __delete_from_swap_cache() argument [all …]
|
D | swap.c | 39 static inline void activate_page_nolock(struct page * page) in activate_page_nolock() argument 41 if (PageLRU(page) && !PageActive(page)) { in activate_page_nolock() 42 del_page_from_inactive_list(page); in activate_page_nolock() 43 add_page_to_active_list(page); in activate_page_nolock() 47 void fastcall activate_page(struct page * page) in activate_page() argument 50 activate_page_nolock(page); in activate_page() 58 void fastcall lru_cache_add(struct page * page) in lru_cache_add() argument 60 if (!PageLRU(page)) { in lru_cache_add() 62 if (!TestSetPageLRU(page)) in lru_cache_add() 63 add_page_to_inactive_list(page); in lru_cache_add() [all …]
|
D | page_alloc.c | 49 static void FASTCALL(__free_pages_ok (struct page *page, unsigned int order)); 52 struct page * free_pages_ok_no_irq_head; 56 struct page * page, * __page; in do_free_pages_ok_no_irq() local 60 page = free_pages_ok_no_irq_head; in do_free_pages_ok_no_irq() 65 while (page) { in do_free_pages_ok_no_irq() 66 __page = page; in do_free_pages_ok_no_irq() 67 page = page->next_hash; in do_free_pages_ok_no_irq() 80 #define BAD_RANGE(zone, page) \ argument 82 (((page) - mem_map) >= ((zone)->zone_start_mapnr+(zone)->size)) \ 83 || (((page) - mem_map) < (zone)->zone_start_mapnr) \ [all …]
|
D | page_io.c | 36 static int rw_swap_page_base(int rw, swp_entry_t entry, struct page *page) in rw_swap_page_base() argument 46 ClearPageUptodate(page); in rw_swap_page_base() 74 brw_page(rw, page, dev, zones, block_size); in rw_swap_page_base() 85 void rw_swap_page(int rw, struct page *page) in rw_swap_page() argument 89 entry.val = page->index; in rw_swap_page() 91 if (!PageLocked(page)) in rw_swap_page() 92 PAGE_BUG(page); in rw_swap_page() 93 if (!PageSwapCache(page)) in rw_swap_page() 94 PAGE_BUG(page); in rw_swap_page() 95 if (!rw_swap_page_base(rw, entry, page)) in rw_swap_page() [all …]
|
D | vmscan.c | 98 …vm_area_struct* vma, unsigned long address, pte_t * page_table, struct page *page, zone_t * classz… in try_to_swap_out() argument 105 mark_page_accessed(page); in try_to_swap_out() 110 if (PageActive(page)) in try_to_swap_out() 114 if (!memclass(page_zone(page), classzone)) in try_to_swap_out() 117 if (TryLockPage(page)) in try_to_swap_out() 130 set_page_dirty(page); in try_to_swap_out() 137 if (PageSwapCache(page)) { in try_to_swap_out() 138 entry.val = page->index; in try_to_swap_out() 144 UnlockPage(page); in try_to_swap_out() 146 int freeable = page_count(page) - !!page->buffers <= 2; in try_to_swap_out() [all …]
|
D | highmem.c | 49 struct page *page; in flush_all_zero_pkmaps() local 72 page = pte_page(pkmap_page_table[i]); in flush_all_zero_pkmaps() 75 page->virtual = NULL; in flush_all_zero_pkmaps() 80 static inline unsigned long map_new_virtual(struct page *page, int nonblocking) in map_new_virtual() argument 116 if (page->virtual) in map_new_virtual() 117 return (unsigned long) page->virtual; in map_new_virtual() 124 set_pte(&(pkmap_page_table[last_pkmap_nr]), mk_pte(page, kmap_prot)); in map_new_virtual() 127 page->virtual = (void *) vaddr; in map_new_virtual() 132 void fastcall *kmap_high(struct page *page, int nonblocking) in kmap_high() argument 143 vaddr = (unsigned long) page->virtual; in kmap_high() [all …]
|
/linux-2.4.37.9/fs/minix/ |
D | dir.c | 23 static inline void dir_put_page(struct page *page) in dir_put_page() argument 25 kunmap(page); in dir_put_page() 26 page_cache_release(page); in dir_put_page() 34 static int dir_commit_chunk(struct page *page, unsigned from, unsigned to) in dir_commit_chunk() argument 36 struct inode *dir = (struct inode *)page->mapping->host; in dir_commit_chunk() 38 page->mapping->a_ops->commit_write(NULL, page, from, to); in dir_commit_chunk() 41 err = writeout_one_page(page); in dir_commit_chunk() 42 err2 = waitfor_one_page(page); in dir_commit_chunk() 49 static struct page * dir_get_page(struct inode *dir, unsigned long n) in dir_get_page() 52 struct page *page = read_cache_page(mapping, n, in dir_get_page() local [all …]
|
/linux-2.4.37.9/fs/nfs/ |
D | symlink.c | 30 static int nfs_symlink_filler(struct inode *inode, struct page *page) in nfs_symlink_filler() argument 39 error = NFS_PROTO(inode)->readlink(inode, page); in nfs_symlink_filler() 43 SetPageUptodate(page); in nfs_symlink_filler() 44 UnlockPage(page); in nfs_symlink_filler() 48 SetPageError(page); in nfs_symlink_filler() 49 UnlockPage(page); in nfs_symlink_filler() 53 static char *nfs_getlink(struct inode *inode, struct page **ppage) in nfs_getlink() 55 struct page *page; in nfs_getlink() local 59 page = read_cache_page(&inode->i_data, 0, in nfs_getlink() 61 if (IS_ERR(page)) in nfs_getlink() [all …]
|
/linux-2.4.37.9/fs/ext2/ |
D | dir.c | 39 static inline void ext2_put_page(struct page *page) in ext2_put_page() argument 41 kunmap(page); in ext2_put_page() 42 page_cache_release(page); in ext2_put_page() 50 static int ext2_commit_chunk(struct page *page, unsigned from, unsigned to) in ext2_commit_chunk() argument 52 struct inode *dir = page->mapping->host; in ext2_commit_chunk() 55 page->mapping->a_ops->commit_write(NULL, page, from, to); in ext2_commit_chunk() 58 err = writeout_one_page(page); in ext2_commit_chunk() 59 err2 = waitfor_one_page(page); in ext2_commit_chunk() 66 static void ext2_check_page(struct page *page, int quiet) in ext2_check_page() argument 68 struct inode *dir = page->mapping->host; in ext2_check_page() [all …]
|
/linux-2.4.37.9/arch/ia64/mm/ |
D | hugetlbpage.c | 31 static struct page *alloc_hugetlb_page(void) in alloc_hugetlb_page() 34 struct page *page; in alloc_hugetlb_page() local 42 page = list_entry(htlbpage_freelist.next, struct page, list); in alloc_hugetlb_page() 43 list_del(&page->list); in alloc_hugetlb_page() 46 set_page_count(page, 1); in alloc_hugetlb_page() 48 clear_highpage(&page[i]); in alloc_hugetlb_page() 49 return page; in alloc_hugetlb_page() 89 struct page *page, pte_t * page_table, int write_access) in set_huge_pte() argument 96 pte_mkwrite(pte_mkdirty(mk_pte(page, vma->vm_page_prot))); in set_huge_pte() 98 entry = pte_wrprotect(mk_pte(page, vma->vm_page_prot)); in set_huge_pte() [all …]
|
/linux-2.4.37.9/include/asm-mips/ |
D | highmem.h | 49 extern void * kmap_high(struct page *page, int nonblocking); 50 extern void kunmap_high(struct page *page); 52 static inline void *__kmap(struct page *page, int nonblocking) in __kmap() argument 56 if (page < highmem_start_page) in __kmap() 57 return page_address(page); in __kmap() 58 return kmap_high(page, nonblocking); in __kmap() 61 #define kmap(page) __kmap(page, 0) argument 62 #define kmap_nonblock(page) __kmap(page, 1) argument 64 static inline void kunmap(struct page *page) in kunmap() argument 68 if (page < highmem_start_page) in kunmap() [all …]
|
/linux-2.4.37.9/fs/xfs/linux-2.4/ |
D | xfs_aops.c | 57 STATIC void xfs_count_page_state(struct page *, int *, int *, int *); 58 STATIC void xfs_convert_page(struct inode *, struct page *, 66 struct page *page, in xfs_page_trace() argument 73 loff_t offset = page->index << PAGE_CACHE_SHIFT; in xfs_page_trace() 76 if (page_has_buffers(page)) in xfs_page_trace() 77 xfs_count_page_state(page, &delalloc, &unmapped, &unwritten); in xfs_page_trace() 88 (void *)page, in xfs_page_trace() 103 #define xfs_page_trace(tag, inode, page, mask) argument 168 struct page *page, in xfs_offset_to_map() argument 176 full_offset = page->index; /* NB: using 64bit number */ in xfs_offset_to_map() [all …]
|
/linux-2.4.37.9/arch/arm/common/ |
D | pcipool.c | 140 struct pci_page *page; in pool_alloc_page() local 147 page = (struct pci_page *) kmalloc (mapsize + sizeof *page, mem_flags); in pool_alloc_page() 148 if (!page) in pool_alloc_page() 150 page->vaddr = pci_alloc_consistent (pool->dev, in pool_alloc_page() 152 &page->dma); in pool_alloc_page() 153 if (page->vaddr) { in pool_alloc_page() 154 memset (page->bitmap, 0xff, mapsize); // bit set == free in pool_alloc_page() 156 memset (page->vaddr, POOL_POISON_BYTE, pool->allocation); in pool_alloc_page() 158 list_add (&page->page_list, &pool->page_list); in pool_alloc_page() 160 kfree (page); in pool_alloc_page() [all …]
|
/linux-2.4.37.9/include/asm-sparc/ |
D | highmem.h | 30 extern void *kmap_high(struct page *page, int nonblocking); 31 extern void kunmap_high(struct page *page); 60 void *kmap_atomic(struct page *page, enum km_type type); 63 #define kmap(page) __kmap(page, 0) argument 64 #define kmap_nonblock(page) __kmap(page, 1) argument 66 static inline void *__kmap(struct page *page, int nonblocking) in __kmap() argument 70 if (page < highmem_start_page) in __kmap() 71 return page_address(page); in __kmap() 72 return kmap_high(page, nonblocking); in __kmap() 75 static inline void kunmap(struct page *page) in kunmap() argument [all …]
|
/linux-2.4.37.9/include/asm-ppc/ |
D | highmem.h | 52 extern void *kmap_high(struct page *page, int nonblock); 53 extern void kunmap_high(struct page *page); 55 #define kmap(page) __kmap(page, 0) argument 56 #define kmap_nonblock(page) __kmap(page, 1) argument 58 static inline void *__kmap(struct page *page, int nonblock) in __kmap() argument 62 if (page < highmem_start_page) in __kmap() 63 return page_address(page); in __kmap() 64 return kmap_high(page, nonblock); in __kmap() 67 static inline void kunmap(struct page *page) in kunmap() argument 71 if (page < highmem_start_page) in kunmap() [all …]
|
/linux-2.4.37.9/include/asm-i386/ |
D | highmem.h | 59 extern void * FASTCALL(kmap_high(struct page *page, int nonblocking)); 60 extern void FASTCALL(kunmap_high(struct page *page)); 62 #define kmap(page) __kmap(page, 0) argument 63 #define kmap_nonblock(page) __kmap(page, 1) argument 65 static inline void *__kmap(struct page *page, int nonblocking) in __kmap() argument 69 if (page < highmem_start_page) in __kmap() 70 return page_address(page); in __kmap() 71 return kmap_high(page, nonblocking); in __kmap() 74 static inline void kunmap(struct page *page) in kunmap() argument 78 if (page < highmem_start_page) in kunmap() [all …]
|
/linux-2.4.37.9/include/asm-sparc64/ |
D | pgalloc.h | 20 #define flush_cache_page(vma, page) \ argument 24 #define flush_page_to_ram(page) do { } while (0) argument 35 extern void flush_dcache_page_impl(struct page *page); 37 extern void smp_flush_dcache_page_impl(struct page *page, int cpu); 38 extern void flush_dcache_page_all(struct mm_struct *mm, struct page *page); 40 #define smp_flush_dcache_page_impl(page,cpu) flush_dcache_page_impl(page) argument 41 #define flush_dcache_page_all(mm,page) flush_dcache_page_impl(page) argument 44 extern void flush_dcache_page(struct page *page); 55 extern void __flush_tlb_page(unsigned long context, unsigned long page, unsigned long r); 77 #define flush_tlb_page(vma, page) \ argument [all …]
|