Home
last modified time | relevance | path

Searched refs:page (Results 1 – 25 of 889) sorted by relevance

12345678910>>...36

/linux-2.4.37.9/include/linux/
Dswap.h97 extern void __remove_inode_page(struct page *);
107 extern void FASTCALL(lru_cache_add(struct page *));
108 extern void FASTCALL(__lru_cache_del(struct page *));
109 extern void FASTCALL(lru_cache_del(struct page *));
111 extern void FASTCALL(activate_page(struct page *));
122 extern void rw_swap_page(int, struct page *);
132 extern int add_to_swap_cache(struct page *, swp_entry_t);
133 extern void __delete_from_swap_cache(struct page *page);
134 extern void delete_from_swap_cache(struct page *page);
135 extern void free_page_and_swap_cache(struct page *page);
[all …]
Dmm.h137 struct page * (*nopage)(struct vm_area_struct * area, unsigned long address, int unused);
155 typedef struct page { struct
159 struct page *next_hash; /* Next page sharing our hash bucket in argument
166 struct page **pprev_hash; /* Complement to *next_hash. */ argument
203 static inline struct page *nth_page(struct page *page, int n) in nth_page() argument
205 return page + n; in nth_page()
311 #define arch_set_page_uptodate(page) argument
315 #define UnlockPage(page) unlock_page(page) argument
316 #define Page_Uptodate(page) test_bit(PG_uptodate, &(page)->flags) argument
318 #define SetPageUptodate(page) set_bit(PG_uptodate, &(page)->flags) argument
[all …]
Dpagemap.h34 static inline struct page *page_cache_alloc(struct address_space *x) in page_cache_alloc()
49 extern struct page **page_hash_table;
73 extern struct page * __find_get_page(struct address_space *mapping,
74 unsigned long index, struct page **hash);
77 extern struct page * __find_lock_page (struct address_space * mapping,
78 unsigned long index, struct page **hash);
79 extern struct page * find_or_create_page(struct address_space *mapping,
82 extern void FASTCALL(lock_page(struct page *page));
83 extern void FASTCALL(unlock_page(struct page *page));
86 extern struct page *find_trylock_page(struct address_space *, unsigned long);
[all …]
Dhighmem.h9 extern struct page *highmem_start_page;
68 static inline void *kmap(struct page *page) { return page_address(page); } in kmap() argument
70 #define kunmap(page) do { } while (0) argument
72 #define kmap_atomic(page,idx) kmap(page) argument
73 #define kunmap_atomic(page,idx) kunmap(page) argument
77 #define kmap_nonblock(page) kmap(page) argument
84 static inline void clear_user_highpage(struct page *page, unsigned long vaddr) in clear_user_highpage() argument
86 void *addr = kmap_atomic(page, KM_USER0); in clear_user_highpage()
91 static inline void clear_highpage(struct page *page) in clear_highpage() argument
93 clear_page(kmap(page)); in clear_highpage()
[all …]
/linux-2.4.37.9/arch/arm/mm/
Dsmall_page.c53 struct page *queue;
68 #error unsupported page size
78 static void add_page_to_queue(struct page *page, struct page **p) in add_page_to_queue() argument
81 if (page->pprev_hash) in add_page_to_queue()
82 PAGE_BUG(page); in add_page_to_queue()
84 page->next_hash = *p; in add_page_to_queue()
86 (*p)->pprev_hash = &page->next_hash; in add_page_to_queue()
87 *p = page; in add_page_to_queue()
88 page->pprev_hash = p; in add_page_to_queue()
91 static void remove_page_from_queue(struct page *page) in remove_page_from_queue() argument
[all …]
Dconsistent.c42 struct page *page, *end, *free; in consistent_alloc() local
53 page = alloc_pages(gfp, order); in consistent_alloc()
54 if (!page) in consistent_alloc()
57 *dma_handle = page_to_bus(page); in consistent_alloc()
58 ret = __ioremap(page_to_pfn(page) << PAGE_SHIFT, size, 0); in consistent_alloc()
64 void *virt = page_address(page); in consistent_alloc()
81 free = page + (size >> PAGE_SHIFT); in consistent_alloc()
82 end = page + (1 << order); in consistent_alloc()
84 for (; page < end; page++) { in consistent_alloc()
85 set_page_count(page, 1); in consistent_alloc()
[all …]
/linux-2.4.37.9/fs/sysv/
Ddir.c28 static inline void dir_put_page(struct page *page) in dir_put_page() argument
30 kunmap(page); in dir_put_page()
31 page_cache_release(page); in dir_put_page()
39 static int dir_commit_chunk(struct page *page, unsigned from, unsigned to) in dir_commit_chunk() argument
41 struct inode *dir = (struct inode *)page->mapping->host; in dir_commit_chunk()
45 page->mapping->a_ops->commit_write(NULL, page, from, to); in dir_commit_chunk()
48 err = writeout_one_page(page); in dir_commit_chunk()
49 err2 = waitfor_one_page(page); in dir_commit_chunk()
56 static struct page * dir_get_page(struct inode *dir, unsigned long n) in dir_get_page()
59 struct page *page = read_cache_page(mapping, n, in dir_get_page() local
[all …]
/linux-2.4.37.9/mm/
Dfilemap.c47 struct page **page_hash_table;
70 static void FASTCALL(add_page_to_hash_queue(struct page * page, struct page **p));
71 static void fastcall add_page_to_hash_queue(struct page * page, struct page **p) in add_page_to_hash_queue() argument
73 struct page *next = *p; in add_page_to_hash_queue()
75 *p = page; in add_page_to_hash_queue()
76 page->next_hash = next; in add_page_to_hash_queue()
77 page->pprev_hash = p; in add_page_to_hash_queue()
79 next->pprev_hash = &page->next_hash; in add_page_to_hash_queue()
80 if (page->buffers) in add_page_to_hash_queue()
81 PAGE_BUG(page); in add_page_to_hash_queue()
[all …]
Dswap_state.c24 static int swap_writepage(struct page *page) in swap_writepage() argument
26 if (remove_exclusive_swap_page(page)) { in swap_writepage()
27 UnlockPage(page); in swap_writepage()
30 rw_swap_page(WRITE, page); in swap_writepage()
70 int add_to_swap_cache(struct page *page, swp_entry_t entry) in add_to_swap_cache() argument
72 if (page->mapping) in add_to_swap_cache()
78 if (add_to_page_cache_unique(page, &swapper_space, entry.val, in add_to_swap_cache()
84 if (!PageLocked(page)) in add_to_swap_cache()
86 if (!PageSwapCache(page)) in add_to_swap_cache()
96 void __delete_from_swap_cache(struct page *page) in __delete_from_swap_cache() argument
[all …]
Dswap.c39 static inline void activate_page_nolock(struct page * page) in activate_page_nolock() argument
41 if (PageLRU(page) && !PageActive(page)) { in activate_page_nolock()
42 del_page_from_inactive_list(page); in activate_page_nolock()
43 add_page_to_active_list(page); in activate_page_nolock()
47 void fastcall activate_page(struct page * page) in activate_page() argument
50 activate_page_nolock(page); in activate_page()
58 void fastcall lru_cache_add(struct page * page) in lru_cache_add() argument
60 if (!PageLRU(page)) { in lru_cache_add()
62 if (!TestSetPageLRU(page)) in lru_cache_add()
63 add_page_to_inactive_list(page); in lru_cache_add()
[all …]
Dpage_alloc.c49 static void FASTCALL(__free_pages_ok (struct page *page, unsigned int order));
52 struct page * free_pages_ok_no_irq_head;
56 struct page * page, * __page; in do_free_pages_ok_no_irq() local
60 page = free_pages_ok_no_irq_head; in do_free_pages_ok_no_irq()
65 while (page) { in do_free_pages_ok_no_irq()
66 __page = page; in do_free_pages_ok_no_irq()
67 page = page->next_hash; in do_free_pages_ok_no_irq()
80 #define BAD_RANGE(zone, page) \ argument
82 (((page) - mem_map) >= ((zone)->zone_start_mapnr+(zone)->size)) \
83 || (((page) - mem_map) < (zone)->zone_start_mapnr) \
[all …]
Dpage_io.c36 static int rw_swap_page_base(int rw, swp_entry_t entry, struct page *page) in rw_swap_page_base() argument
46 ClearPageUptodate(page); in rw_swap_page_base()
74 brw_page(rw, page, dev, zones, block_size); in rw_swap_page_base()
85 void rw_swap_page(int rw, struct page *page) in rw_swap_page() argument
89 entry.val = page->index; in rw_swap_page()
91 if (!PageLocked(page)) in rw_swap_page()
92 PAGE_BUG(page); in rw_swap_page()
93 if (!PageSwapCache(page)) in rw_swap_page()
94 PAGE_BUG(page); in rw_swap_page()
95 if (!rw_swap_page_base(rw, entry, page)) in rw_swap_page()
[all …]
Dvmscan.c98 …vm_area_struct* vma, unsigned long address, pte_t * page_table, struct page *page, zone_t * classz… in try_to_swap_out() argument
105 mark_page_accessed(page); in try_to_swap_out()
110 if (PageActive(page)) in try_to_swap_out()
114 if (!memclass(page_zone(page), classzone)) in try_to_swap_out()
117 if (TryLockPage(page)) in try_to_swap_out()
130 set_page_dirty(page); in try_to_swap_out()
137 if (PageSwapCache(page)) { in try_to_swap_out()
138 entry.val = page->index; in try_to_swap_out()
144 UnlockPage(page); in try_to_swap_out()
146 int freeable = page_count(page) - !!page->buffers <= 2; in try_to_swap_out()
[all …]
Dhighmem.c49 struct page *page; in flush_all_zero_pkmaps() local
72 page = pte_page(pkmap_page_table[i]); in flush_all_zero_pkmaps()
75 page->virtual = NULL; in flush_all_zero_pkmaps()
80 static inline unsigned long map_new_virtual(struct page *page, int nonblocking) in map_new_virtual() argument
116 if (page->virtual) in map_new_virtual()
117 return (unsigned long) page->virtual; in map_new_virtual()
124 set_pte(&(pkmap_page_table[last_pkmap_nr]), mk_pte(page, kmap_prot)); in map_new_virtual()
127 page->virtual = (void *) vaddr; in map_new_virtual()
132 void fastcall *kmap_high(struct page *page, int nonblocking) in kmap_high() argument
143 vaddr = (unsigned long) page->virtual; in kmap_high()
[all …]
/linux-2.4.37.9/fs/minix/
Ddir.c23 static inline void dir_put_page(struct page *page) in dir_put_page() argument
25 kunmap(page); in dir_put_page()
26 page_cache_release(page); in dir_put_page()
34 static int dir_commit_chunk(struct page *page, unsigned from, unsigned to) in dir_commit_chunk() argument
36 struct inode *dir = (struct inode *)page->mapping->host; in dir_commit_chunk()
38 page->mapping->a_ops->commit_write(NULL, page, from, to); in dir_commit_chunk()
41 err = writeout_one_page(page); in dir_commit_chunk()
42 err2 = waitfor_one_page(page); in dir_commit_chunk()
49 static struct page * dir_get_page(struct inode *dir, unsigned long n) in dir_get_page()
52 struct page *page = read_cache_page(mapping, n, in dir_get_page() local
[all …]
/linux-2.4.37.9/fs/nfs/
Dsymlink.c30 static int nfs_symlink_filler(struct inode *inode, struct page *page) in nfs_symlink_filler() argument
39 error = NFS_PROTO(inode)->readlink(inode, page); in nfs_symlink_filler()
43 SetPageUptodate(page); in nfs_symlink_filler()
44 UnlockPage(page); in nfs_symlink_filler()
48 SetPageError(page); in nfs_symlink_filler()
49 UnlockPage(page); in nfs_symlink_filler()
53 static char *nfs_getlink(struct inode *inode, struct page **ppage) in nfs_getlink()
55 struct page *page; in nfs_getlink() local
59 page = read_cache_page(&inode->i_data, 0, in nfs_getlink()
61 if (IS_ERR(page)) in nfs_getlink()
[all …]
/linux-2.4.37.9/fs/ext2/
Ddir.c39 static inline void ext2_put_page(struct page *page) in ext2_put_page() argument
41 kunmap(page); in ext2_put_page()
42 page_cache_release(page); in ext2_put_page()
50 static int ext2_commit_chunk(struct page *page, unsigned from, unsigned to) in ext2_commit_chunk() argument
52 struct inode *dir = page->mapping->host; in ext2_commit_chunk()
55 page->mapping->a_ops->commit_write(NULL, page, from, to); in ext2_commit_chunk()
58 err = writeout_one_page(page); in ext2_commit_chunk()
59 err2 = waitfor_one_page(page); in ext2_commit_chunk()
66 static void ext2_check_page(struct page *page, int quiet) in ext2_check_page() argument
68 struct inode *dir = page->mapping->host; in ext2_check_page()
[all …]
/linux-2.4.37.9/arch/ia64/mm/
Dhugetlbpage.c31 static struct page *alloc_hugetlb_page(void) in alloc_hugetlb_page()
34 struct page *page; in alloc_hugetlb_page() local
42 page = list_entry(htlbpage_freelist.next, struct page, list); in alloc_hugetlb_page()
43 list_del(&page->list); in alloc_hugetlb_page()
46 set_page_count(page, 1); in alloc_hugetlb_page()
48 clear_highpage(&page[i]); in alloc_hugetlb_page()
49 return page; in alloc_hugetlb_page()
89 struct page *page, pte_t * page_table, int write_access) in set_huge_pte() argument
96 pte_mkwrite(pte_mkdirty(mk_pte(page, vma->vm_page_prot))); in set_huge_pte()
98 entry = pte_wrprotect(mk_pte(page, vma->vm_page_prot)); in set_huge_pte()
[all …]
/linux-2.4.37.9/include/asm-mips/
Dhighmem.h49 extern void * kmap_high(struct page *page, int nonblocking);
50 extern void kunmap_high(struct page *page);
52 static inline void *__kmap(struct page *page, int nonblocking) in __kmap() argument
56 if (page < highmem_start_page) in __kmap()
57 return page_address(page); in __kmap()
58 return kmap_high(page, nonblocking); in __kmap()
61 #define kmap(page) __kmap(page, 0) argument
62 #define kmap_nonblock(page) __kmap(page, 1) argument
64 static inline void kunmap(struct page *page) in kunmap() argument
68 if (page < highmem_start_page) in kunmap()
[all …]
/linux-2.4.37.9/fs/xfs/linux-2.4/
Dxfs_aops.c57 STATIC void xfs_count_page_state(struct page *, int *, int *, int *);
58 STATIC void xfs_convert_page(struct inode *, struct page *,
66 struct page *page, in xfs_page_trace() argument
73 loff_t offset = page->index << PAGE_CACHE_SHIFT; in xfs_page_trace()
76 if (page_has_buffers(page)) in xfs_page_trace()
77 xfs_count_page_state(page, &delalloc, &unmapped, &unwritten); in xfs_page_trace()
88 (void *)page, in xfs_page_trace()
103 #define xfs_page_trace(tag, inode, page, mask) argument
168 struct page *page, in xfs_offset_to_map() argument
176 full_offset = page->index; /* NB: using 64bit number */ in xfs_offset_to_map()
[all …]
/linux-2.4.37.9/arch/arm/common/
Dpcipool.c140 struct pci_page *page; in pool_alloc_page() local
147 page = (struct pci_page *) kmalloc (mapsize + sizeof *page, mem_flags); in pool_alloc_page()
148 if (!page) in pool_alloc_page()
150 page->vaddr = pci_alloc_consistent (pool->dev, in pool_alloc_page()
152 &page->dma); in pool_alloc_page()
153 if (page->vaddr) { in pool_alloc_page()
154 memset (page->bitmap, 0xff, mapsize); // bit set == free in pool_alloc_page()
156 memset (page->vaddr, POOL_POISON_BYTE, pool->allocation); in pool_alloc_page()
158 list_add (&page->page_list, &pool->page_list); in pool_alloc_page()
160 kfree (page); in pool_alloc_page()
[all …]
/linux-2.4.37.9/include/asm-sparc/
Dhighmem.h30 extern void *kmap_high(struct page *page, int nonblocking);
31 extern void kunmap_high(struct page *page);
60 void *kmap_atomic(struct page *page, enum km_type type);
63 #define kmap(page) __kmap(page, 0) argument
64 #define kmap_nonblock(page) __kmap(page, 1) argument
66 static inline void *__kmap(struct page *page, int nonblocking) in __kmap() argument
70 if (page < highmem_start_page) in __kmap()
71 return page_address(page); in __kmap()
72 return kmap_high(page, nonblocking); in __kmap()
75 static inline void kunmap(struct page *page) in kunmap() argument
[all …]
/linux-2.4.37.9/include/asm-ppc/
Dhighmem.h52 extern void *kmap_high(struct page *page, int nonblock);
53 extern void kunmap_high(struct page *page);
55 #define kmap(page) __kmap(page, 0) argument
56 #define kmap_nonblock(page) __kmap(page, 1) argument
58 static inline void *__kmap(struct page *page, int nonblock) in __kmap() argument
62 if (page < highmem_start_page) in __kmap()
63 return page_address(page); in __kmap()
64 return kmap_high(page, nonblock); in __kmap()
67 static inline void kunmap(struct page *page) in kunmap() argument
71 if (page < highmem_start_page) in kunmap()
[all …]
/linux-2.4.37.9/include/asm-i386/
Dhighmem.h59 extern void * FASTCALL(kmap_high(struct page *page, int nonblocking));
60 extern void FASTCALL(kunmap_high(struct page *page));
62 #define kmap(page) __kmap(page, 0) argument
63 #define kmap_nonblock(page) __kmap(page, 1) argument
65 static inline void *__kmap(struct page *page, int nonblocking) in __kmap() argument
69 if (page < highmem_start_page) in __kmap()
70 return page_address(page); in __kmap()
71 return kmap_high(page, nonblocking); in __kmap()
74 static inline void kunmap(struct page *page) in kunmap() argument
78 if (page < highmem_start_page) in kunmap()
[all …]
/linux-2.4.37.9/include/asm-sparc64/
Dpgalloc.h20 #define flush_cache_page(vma, page) \ argument
24 #define flush_page_to_ram(page) do { } while (0) argument
35 extern void flush_dcache_page_impl(struct page *page);
37 extern void smp_flush_dcache_page_impl(struct page *page, int cpu);
38 extern void flush_dcache_page_all(struct mm_struct *mm, struct page *page);
40 #define smp_flush_dcache_page_impl(page,cpu) flush_dcache_page_impl(page) argument
41 #define flush_dcache_page_all(mm,page) flush_dcache_page_impl(page) argument
44 extern void flush_dcache_page(struct page *page);
55 extern void __flush_tlb_page(unsigned long context, unsigned long page, unsigned long r);
77 #define flush_tlb_page(vma, page) \ argument
[all …]

12345678910>>...36