Home
last modified time | relevance | path

Searched refs:kaddr (Results 1 – 25 of 203) sorted by relevance

123456789

/linux-6.1.9/fs/nilfs2/
Dcpfile.c71 void *kaddr, in nilfs_cpfile_block_add_valid_checkpoints() argument
74 struct nilfs_checkpoint *cp = kaddr + bh_offset(bh); in nilfs_cpfile_block_add_valid_checkpoints()
85 void *kaddr, in nilfs_cpfile_block_sub_valid_checkpoints() argument
88 struct nilfs_checkpoint *cp = kaddr + bh_offset(bh); in nilfs_cpfile_block_sub_valid_checkpoints()
100 void *kaddr) in nilfs_cpfile_block_get_header() argument
102 return kaddr + bh_offset(bh); in nilfs_cpfile_block_get_header()
108 void *kaddr) in nilfs_cpfile_block_get_checkpoint() argument
110 return kaddr + bh_offset(bh) + nilfs_cpfile_get_offset(cpfile, cno) * in nilfs_cpfile_block_get_checkpoint()
116 void *kaddr) in nilfs_cpfile_block_init() argument
118 struct nilfs_checkpoint *cp = kaddr + bh_offset(bh); in nilfs_cpfile_block_init()
[all …]
Ddat.c79 void *kaddr; in nilfs_dat_commit_alloc() local
81 kaddr = kmap_atomic(req->pr_entry_bh->b_page); in nilfs_dat_commit_alloc()
83 req->pr_entry_bh, kaddr); in nilfs_dat_commit_alloc()
87 kunmap_atomic(kaddr); in nilfs_dat_commit_alloc()
103 void *kaddr; in nilfs_dat_commit_free() local
105 kaddr = kmap_atomic(req->pr_entry_bh->b_page); in nilfs_dat_commit_free()
107 req->pr_entry_bh, kaddr); in nilfs_dat_commit_free()
111 kunmap_atomic(kaddr); in nilfs_dat_commit_free()
137 void *kaddr; in nilfs_dat_commit_start() local
139 kaddr = kmap_atomic(req->pr_entry_bh->b_page); in nilfs_dat_commit_start()
[all …]
Dsufile.c75 struct buffer_head *bh, void *kaddr) in nilfs_sufile_block_get_segment_usage() argument
77 return kaddr + bh_offset(bh) + in nilfs_sufile_block_get_segment_usage()
108 void *kaddr; in nilfs_sufile_mod_counter() local
110 kaddr = kmap_atomic(header_bh->b_page); in nilfs_sufile_mod_counter()
111 header = kaddr + bh_offset(header_bh); in nilfs_sufile_mod_counter()
114 kunmap_atomic(kaddr); in nilfs_sufile_mod_counter()
309 void *kaddr; in nilfs_sufile_alloc() local
318 kaddr = kmap_atomic(header_bh->b_page); in nilfs_sufile_alloc()
319 header = kaddr + bh_offset(header_bh); in nilfs_sufile_alloc()
321 kunmap_atomic(kaddr); in nilfs_sufile_alloc()
[all …]
Ddir.c121 char *kaddr = page_address(page); in nilfs_check_page() local
135 p = (struct nilfs_dir_entry *)(kaddr + offs); in nilfs_check_page()
179 p = (struct nilfs_dir_entry *)(kaddr + offs); in nilfs_check_page()
276 char *kaddr, *limit; in nilfs_readdir() local
285 kaddr = page_address(page); in nilfs_readdir()
286 de = (struct nilfs_dir_entry *)(kaddr + offset); in nilfs_readdir()
287 limit = kaddr + nilfs_last_byte(inode, n) - in nilfs_readdir()
348 char *kaddr; in nilfs_find_entry() local
352 kaddr = page_address(page); in nilfs_find_entry()
353 de = (struct nilfs_dir_entry *)kaddr; in nilfs_find_entry()
[all …]
/linux-6.1.9/arch/arm64/kernel/
Dparavirt.c38 struct pvclock_vcpu_stolen_time __rcu *kaddr; member
55 struct pvclock_vcpu_stolen_time *kaddr = NULL; in para_steal_clock() local
67 kaddr = rcu_dereference(reg->kaddr); in para_steal_clock()
68 if (!kaddr) { in para_steal_clock()
73 ret = le64_to_cpu(READ_ONCE(kaddr->stolen_time)); in para_steal_clock()
80 struct pvclock_vcpu_stolen_time *kaddr = NULL; in stolen_time_cpu_down_prepare() local
84 if (!reg->kaddr) in stolen_time_cpu_down_prepare()
87 kaddr = rcu_replace_pointer(reg->kaddr, NULL, true); in stolen_time_cpu_down_prepare()
89 memunmap(kaddr); in stolen_time_cpu_down_prepare()
96 struct pvclock_vcpu_stolen_time *kaddr = NULL; in stolen_time_cpu_online() local
[all …]
/linux-6.1.9/fs/btrfs/
Dstruct-funcs.c63 ASSERT(token->kaddr); \
67 return get_unaligned_le##bits(token->kaddr + oip); \
69 token->kaddr = page_address(token->eb->pages[idx]); \
72 return get_unaligned_le##bits(token->kaddr + oip); \
74 memcpy(lebytes, token->kaddr + oip, part); \
75 token->kaddr = page_address(token->eb->pages[idx + 1]); \
77 memcpy(lebytes + part, token->kaddr, size - part); \
86 char *kaddr = page_address(eb->pages[idx]); \
93 return get_unaligned_le##bits(kaddr + oip); \
95 memcpy(lebytes, kaddr + oip, part); \
[all …]
Dlzo.c138 char *kaddr; in copy_compressed_data_to_page() local
158 kaddr = kmap_local_page(cur_page); in copy_compressed_data_to_page()
159 write_compress_length(kaddr + offset_in_page(*cur_out), in copy_compressed_data_to_page()
170 kunmap_local(kaddr); in copy_compressed_data_to_page()
183 kaddr = kmap_local_page(cur_page); in copy_compressed_data_to_page()
185 memcpy(kaddr + offset_in_page(*cur_out), in copy_compressed_data_to_page()
200 memset(kaddr + offset_in_page(*cur_out), 0, in copy_compressed_data_to_page()
205 kunmap_local(kaddr); in copy_compressed_data_to_page()
332 char *kaddr; in lzo_decompress_bio() local
341 kaddr = kmap_local_page(cb->compressed_pages[0]); in lzo_decompress_bio()
[all …]
/linux-6.1.9/arch/m68k/sun3/
Ddvma.c23 static unsigned long dvma_page(unsigned long kaddr, unsigned long vaddr) in dvma_page() argument
29 j = *(volatile unsigned long *)kaddr; in dvma_page()
30 *(volatile unsigned long *)kaddr = j; in dvma_page()
32 ptep = pfn_pte(virt_to_pfn(kaddr), PAGE_KERNEL); in dvma_page()
40 return (vaddr + (kaddr & ~PAGE_MASK)); in dvma_page()
44 int dvma_map_iommu(unsigned long kaddr, unsigned long baddr, in dvma_map_iommu() argument
56 dvma_page(kaddr, vaddr); in dvma_map_iommu()
57 kaddr += PAGE_SIZE; in dvma_map_iommu()
Dsun3dvma.c283 unsigned long dvma_map_align(unsigned long kaddr, int len, int align) in dvma_map_align() argument
292 if(!kaddr || !len) { in dvma_map_align()
298 pr_debug("dvma_map request %08x bytes from %08lx\n", len, kaddr); in dvma_map_align()
299 off = kaddr & ~DVMA_PAGE_MASK; in dvma_map_align()
300 kaddr &= PAGE_MASK; in dvma_map_align()
312 if(!dvma_map_iommu(kaddr, baddr, len)) in dvma_map_align()
315 pr_crit("dvma_map failed kaddr %lx baddr %lx len %x\n", kaddr, baddr, in dvma_map_align()
340 unsigned long kaddr; in dvma_malloc_align() local
350 if((kaddr = __get_free_pages(GFP_ATOMIC, get_order(len))) == 0) in dvma_malloc_align()
353 if((baddr = (unsigned long)dvma_map_align(kaddr, len, align)) == 0) { in dvma_malloc_align()
[all …]
/linux-6.1.9/fs/ext2/
Ddir.c109 static bool ext2_check_page(struct page *page, int quiet, char *kaddr) in ext2_check_page() argument
128 p = (ext2_dirent *)(kaddr + offs); in ext2_check_page()
180 p = (ext2_dirent *)(kaddr + offs); in ext2_check_page()
284 char *kaddr, *limit; in ext2_readdir() local
286 struct page *page = ext2_get_page(inode, n, 0, (void **)&kaddr); in ext2_readdir()
297 offset = ext2_validate_entry(kaddr, offset, chunk_mask); in ext2_readdir()
303 de = (ext2_dirent *)(kaddr+offset); in ext2_readdir()
304 limit = kaddr + ext2_last_byte(inode, n) - EXT2_DIR_REC_LEN(1); in ext2_readdir()
309 ext2_put_page(page, kaddr); in ext2_readdir()
321 ext2_put_page(page, kaddr); in ext2_readdir()
[all …]
/linux-6.1.9/drivers/acpi/
Dnvs.c78 void *kaddr; member
138 if (entry->kaddr) { in suspend_nvs_free()
140 iounmap(entry->kaddr); in suspend_nvs_free()
143 acpi_os_unmap_iomem(entry->kaddr, in suspend_nvs_free()
146 entry->kaddr = NULL; in suspend_nvs_free()
182 entry->kaddr = acpi_os_get_iomem(phys, size); in suspend_nvs_save()
183 if (!entry->kaddr) { in suspend_nvs_save()
184 entry->kaddr = acpi_os_ioremap(phys, size); in suspend_nvs_save()
185 entry->unmap = !!entry->kaddr; in suspend_nvs_save()
187 if (!entry->kaddr) { in suspend_nvs_save()
[all …]
/linux-6.1.9/fs/sysv/
Ddir.c81 char *kaddr, *limit; in sysv_readdir() local
87 kaddr = (char *)page_address(page); in sysv_readdir()
88 de = (struct sysv_dir_entry *)(kaddr+offset); in sysv_readdir()
89 limit = kaddr + PAGE_SIZE - SYSV_DIRSIZE; in sysv_readdir()
145 char *kaddr; in sysv_find_entry() local
148 kaddr = (char*)page_address(page); in sysv_find_entry()
149 de = (struct sysv_dir_entry *) kaddr; in sysv_find_entry()
150 kaddr += PAGE_SIZE - SYSV_DIRSIZE; in sysv_find_entry()
151 for ( ; (char *) de <= kaddr ; de++) { in sysv_find_entry()
182 char *kaddr; in sysv_add_link() local
[all …]
/linux-6.1.9/fs/minix/
Ddir.c100 char *p, *kaddr, *limit; in minix_readdir() local
105 kaddr = (char *)page_address(page); in minix_readdir()
106 p = kaddr+offset; in minix_readdir()
107 limit = kaddr + minix_last_byte(inode, n) - chunk_size; in minix_readdir()
168 char *kaddr, *limit; in minix_find_entry() local
174 kaddr = (char*)page_address(page); in minix_find_entry()
175 limit = kaddr + minix_last_byte(dir, n) - sbi->s_dirsize; in minix_find_entry()
176 for (p = kaddr; p <= limit; p = minix_next_entry(p, sbi)) { in minix_find_entry()
210 char *kaddr, *p; in minix_add_link() local
231 kaddr = (char*)page_address(page); in minix_add_link()
[all …]
/linux-6.1.9/fs/ufs/
Ddir.c114 char *kaddr = page_address(page); in ufs_check_page() local
129 p = (struct ufs_dir_entry *)(kaddr + offs); in ufs_check_page()
179 p = (struct ufs_dir_entry *)(kaddr + offs); in ufs_check_page()
277 char *kaddr; in ufs_find_entry() local
280 kaddr = page_address(page); in ufs_find_entry()
281 de = (struct ufs_dir_entry *) kaddr; in ufs_find_entry()
282 kaddr += ufs_last_byte(dir, n) - reclen; in ufs_find_entry()
283 while ((char *) de <= kaddr) { in ufs_find_entry()
318 char *kaddr; in ufs_add_link() local
337 kaddr = page_address(page); in ufs_add_link()
[all …]
/linux-6.1.9/arch/m68k/sun3x/
Ddvma.c78 inline int dvma_map_cpu(unsigned long kaddr, in dvma_map_cpu() argument
87 kaddr &= PAGE_MASK; in dvma_map_cpu()
92 pr_debug("dvma: mapping kern %08lx to virt %08lx\n", kaddr, vaddr); in dvma_map_cpu()
127 __pa(kaddr), vaddr); in dvma_map_cpu()
128 set_pte(pte, pfn_pte(virt_to_pfn(kaddr), in dvma_map_cpu()
131 kaddr += PAGE_SIZE; in dvma_map_cpu()
146 inline int dvma_map_iommu(unsigned long kaddr, unsigned long baddr, in dvma_map_iommu() argument
163 dvma_entry_set(index, __pa(kaddr)); in dvma_map_iommu()
168 kaddr += DVMA_PAGE_SIZE; in dvma_map_iommu()
/linux-6.1.9/fs/erofs/
Dinode.c20 void *kaddr; in erofs_read_inode() local
32 kaddr = erofs_read_metabuf(buf, sb, blkaddr, EROFS_KMAP); in erofs_read_inode()
33 if (IS_ERR(kaddr)) { in erofs_read_inode()
35 vi->nid, PTR_ERR(kaddr)); in erofs_read_inode()
36 return kaddr; in erofs_read_inode()
39 dic = kaddr + *ofs; in erofs_read_inode()
73 kaddr = erofs_read_metabuf(buf, sb, blkaddr + 1, in erofs_read_inode()
75 if (IS_ERR(kaddr)) { in erofs_read_inode()
77 vi->nid, PTR_ERR(kaddr)); in erofs_read_inode()
79 return kaddr; in erofs_read_inode()
[all …]
Dxattr.c13 void *kaddr; member
79 it.kaddr = erofs_read_metabuf(&it.buf, sb, it.blkaddr, EROFS_KMAP); in init_inode_xattrs()
80 if (IS_ERR(it.kaddr)) { in init_inode_xattrs()
81 ret = PTR_ERR(it.kaddr); in init_inode_xattrs()
85 ih = (struct erofs_xattr_ibody_header *)(it.kaddr + it.ofs); in init_inode_xattrs()
103 it.kaddr = erofs_read_metabuf(&it.buf, sb, ++it.blkaddr, in init_inode_xattrs()
105 if (IS_ERR(it.kaddr)) { in init_inode_xattrs()
108 ret = PTR_ERR(it.kaddr); in init_inode_xattrs()
114 le32_to_cpu(*(__le32 *)(it.kaddr + it.ofs)); in init_inode_xattrs()
150 it->kaddr = erofs_read_metabuf(&it->buf, it->sb, it->blkaddr, in xattr_iter_fixup()
[all …]
/linux-6.1.9/arch/arm/include/asm/
Dtlbflush.h472 static inline void __local_flush_tlb_kernel_page(unsigned long kaddr) in __local_flush_tlb_kernel_page() argument
477 tlb_op(TLB_V4_U_PAGE, "c8, c7, 1", kaddr); in __local_flush_tlb_kernel_page()
478 tlb_op(TLB_V4_D_PAGE, "c8, c6, 1", kaddr); in __local_flush_tlb_kernel_page()
479 tlb_op(TLB_V4_I_PAGE, "c8, c5, 1", kaddr); in __local_flush_tlb_kernel_page()
483 tlb_op(TLB_V6_U_PAGE, "c8, c7, 1", kaddr); in __local_flush_tlb_kernel_page()
484 tlb_op(TLB_V6_D_PAGE, "c8, c6, 1", kaddr); in __local_flush_tlb_kernel_page()
485 tlb_op(TLB_V6_I_PAGE, "c8, c5, 1", kaddr); in __local_flush_tlb_kernel_page()
488 static inline void local_flush_tlb_kernel_page(unsigned long kaddr) in local_flush_tlb_kernel_page() argument
492 kaddr &= PAGE_MASK; in local_flush_tlb_kernel_page()
497 __local_flush_tlb_kernel_page(kaddr); in local_flush_tlb_kernel_page()
[all …]
/linux-6.1.9/arch/loongarch/include/asm/
Dpage.h98 #define virt_to_pfn(kaddr) PFN_DOWN(PHYSADDR(kaddr)) argument
99 #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) argument
101 extern int __virt_addr_valid(volatile void *kaddr);
102 #define virt_addr_valid(kaddr) __virt_addr_valid((volatile void *)(kaddr)) argument
/linux-6.1.9/fs/ntfs/
Dbitmap.c40 u8 *kaddr; in __ntfs_bitmap_set_bits_in_run() local
68 kaddr = page_address(page); in __ntfs_bitmap_set_bits_in_run()
78 u8 *byte = kaddr + pos; in __ntfs_bitmap_set_bits_in_run()
98 memset(kaddr + pos, value ? 0xff : 0, len); in __ntfs_bitmap_set_bits_in_run()
116 kaddr = page_address(page); in __ntfs_bitmap_set_bits_in_run()
122 memset(kaddr, value ? 0xff : 0, len); in __ntfs_bitmap_set_bits_in_run()
136 byte = kaddr + len; in __ntfs_bitmap_set_bits_in_run()
/linux-6.1.9/tools/testing/nvdimm/
Dpmem-dax.c12 long nr_pages, enum dax_access_mode mode, void **kaddr, in __pmem_direct_access() argument
28 if (kaddr) in __pmem_direct_access()
29 *kaddr = pmem->virt_addr + offset; in __pmem_direct_access()
39 if (kaddr) in __pmem_direct_access()
40 *kaddr = pmem->virt_addr + offset; in __pmem_direct_access()
/linux-6.1.9/arch/arc/include/asm/
Dpage.h94 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument
126 #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) argument
127 #define virt_addr_valid(kaddr) pfn_valid(virt_to_pfn(kaddr)) argument
/linux-6.1.9/arch/hexagon/include/asm/
Dpage.h93 #define virt_to_page(kaddr) pfn_to_page(PFN_DOWN(__pa(kaddr))) argument
99 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) argument
129 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument
/linux-6.1.9/arch/m68k/include/asm/
Dpage_no.h23 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument
33 #define virt_addr_valid(kaddr) (((unsigned long)(kaddr) >= PAGE_OFFSET) && \ argument
34 ((unsigned long)(kaddr) < memory_end))
/linux-6.1.9/arch/x86/include/asm/
Dpage.h69 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) argument
71 extern bool __virt_addr_valid(unsigned long kaddr);
72 #define virt_addr_valid(kaddr) __virt_addr_valid((unsigned long) (kaddr)) argument

123456789