Home
last modified time | relevance | path

Searched refs:page_offset (Results 1 – 25 of 216) sorted by relevance

123456789

/linux-6.6.21/drivers/infiniband/sw/rxe/
Drxe_mr.c240 mr->page_offset = mr->ibmr.iova & (page_size - 1); in rxe_map_mr_sg()
248 unsigned int page_offset = rxe_mr_iova_to_page_offset(mr, iova); in rxe_mr_copy_xarray() local
260 mr_page_size(mr) - page_offset); in rxe_mr_copy_xarray()
263 memcpy(addr, va + page_offset, bytes); in rxe_mr_copy_xarray()
265 memcpy(va + page_offset, addr, bytes); in rxe_mr_copy_xarray()
268 page_offset = 0; in rxe_mr_copy_xarray()
280 unsigned int page_offset = dma_addr & (PAGE_SIZE - 1); in rxe_mr_copy_dma() local
288 PAGE_SIZE - page_offset); in rxe_mr_copy_dma()
292 memcpy(va + page_offset, addr, bytes); in rxe_mr_copy_dma()
294 memcpy(addr, va + page_offset, bytes); in rxe_mr_copy_dma()
[all …]
/linux-6.6.21/drivers/net/ethernet/cavium/liquidio/
Docteon_network.h298 pg_info->page_offset = 0; in recv_buffer_alloc()
300 skb_pg_info->page_offset = 0; in recv_buffer_alloc()
324 skb_pg_info->page_offset = 0; in recv_buffer_fast_alloc()
348 pg_info->page_offset = 0; in recv_buffer_recycle()
353 if (pg_info->page_offset == 0) in recv_buffer_recycle()
354 pg_info->page_offset = LIO_RXBUFFER_SZ; in recv_buffer_recycle()
356 pg_info->page_offset = 0; in recv_buffer_recycle()
384 skb_pg_info->page_offset = pg_info->page_offset; in recv_buffer_reuse()
398 pg_info->page_offset = 0; in recv_buffer_destroy()
415 pg_info->page_offset = 0; in recv_buffer_free()
[all …]
/linux-6.6.21/drivers/scsi/fnic/
Dfnic_trace.c57 fnic_trace_entries.page_offset[fnic_trace_entries.wr_idx]; in fnic_trace_get_buf()
110 fnic_trace_entries.page_offset[rd_idx]; in fnic_get_trace_data()
152 fnic_trace_entries.page_offset[rd_idx]; in fnic_get_trace_data()
476 fnic_trace_entries.page_offset = in fnic_trace_buf_init()
479 if (!fnic_trace_entries.page_offset) { in fnic_trace_buf_init()
489 memset((void *)fnic_trace_entries.page_offset, 0, in fnic_trace_buf_init()
500 fnic_trace_entries.page_offset[i] = fnic_buf_head; in fnic_trace_buf_init()
518 if (fnic_trace_entries.page_offset) { in fnic_trace_free()
519 vfree((void *)fnic_trace_entries.page_offset); in fnic_trace_free()
520 fnic_trace_entries.page_offset = NULL; in fnic_trace_free()
[all …]
/linux-6.6.21/drivers/gpu/drm/vmwgfx/
Dvmwgfx_page_dirty.c381 unsigned long page_offset; in vmw_bo_vm_mkwrite() local
396 page_offset = vmf->pgoff - drm_vma_node_start(&bo->base.vma_node); in vmw_bo_vm_mkwrite()
397 if (unlikely(page_offset >= PFN_UP(bo->resource->size))) { in vmw_bo_vm_mkwrite()
403 !test_bit(page_offset, &vbo->dirty->bitmap[0])) { in vmw_bo_vm_mkwrite()
406 __set_bit(page_offset, &dirty->bitmap[0]); in vmw_bo_vm_mkwrite()
407 dirty->start = min(dirty->start, page_offset); in vmw_bo_vm_mkwrite()
408 dirty->end = max(dirty->end, page_offset + 1); in vmw_bo_vm_mkwrite()
435 unsigned long page_offset; in vmw_bo_vm_fault() local
437 page_offset = vmf->pgoff - in vmw_bo_vm_fault()
439 if (page_offset >= PFN_UP(bo->resource->size) || in vmw_bo_vm_fault()
[all …]
/linux-6.6.21/drivers/vfio/pci/pds/
Dlm.c193 size_t page_offset; in pds_vfio_save_read() local
199 page_offset = (*pos) % PAGE_SIZE; in pds_vfio_save_read()
200 page = pds_vfio_get_file_page(lm_file, *pos - page_offset); in pds_vfio_save_read()
207 page_len = min_t(size_t, len, PAGE_SIZE - page_offset); in pds_vfio_save_read()
209 err = copy_to_user(buf, from_buff + page_offset, page_len); in pds_vfio_save_read()
287 size_t page_offset; in pds_vfio_restore_write() local
293 page_offset = (*pos) % PAGE_SIZE; in pds_vfio_restore_write()
294 page = pds_vfio_get_file_page(lm_file, *pos - page_offset); in pds_vfio_restore_write()
301 page_len = min_t(size_t, len, PAGE_SIZE - page_offset); in pds_vfio_restore_write()
303 err = copy_from_user(to_buff + page_offset, buf, page_len); in pds_vfio_restore_write()
/linux-6.6.21/drivers/gpu/drm/ttm/
Dttm_bo_vm.c85 unsigned long page_offset) in ttm_bo_io_mem_pfn() argument
90 return bdev->funcs->io_mem_pfn(bo, page_offset); in ttm_bo_io_mem_pfn()
92 return (bo->resource->bus.offset >> PAGE_SHIFT) + page_offset; in ttm_bo_io_mem_pfn()
188 unsigned long page_offset; in ttm_bo_vm_fault_reserved() local
210 page_offset = ((address - vma->vm_start) >> PAGE_SHIFT) + in ttm_bo_vm_fault_reserved()
215 if (unlikely(page_offset >= PFN_UP(bo->base.size))) in ttm_bo_vm_fault_reserved()
247 pfn = ttm_bo_io_mem_pfn(bo, page_offset); in ttm_bo_vm_fault_reserved()
249 page = ttm->pages[page_offset]; in ttm_bo_vm_fault_reserved()
277 if (unlikely(++page_offset >= page_last)) in ttm_bo_vm_fault_reserved()
/linux-6.6.21/scripts/
Dleaking_addresses.pl305 state $page_offset = get_page_offset();
311 if (hex($match) < $page_offset) {
321 my $page_offset;
329 $page_offset = get_kernel_config_option('CONFIG_PAGE_OFFSET');
330 if (!$page_offset) {
333 return $page_offset;
/linux-6.6.21/fs/hfs/
Dbnode.c25 off += node->page_offset; in hfs_bnode_read()
77 off += node->page_offset; in hfs_bnode_write()
101 off += node->page_offset; in hfs_bnode_clear()
116 src += src_node->page_offset; in hfs_bnode_copy()
117 dst += dst_node->page_offset; in hfs_bnode_copy()
133 src += node->page_offset; in hfs_bnode_move()
134 dst += node->page_offset; in hfs_bnode_move()
288 node->page_offset = off & ~PAGE_MASK; in __hfs_bnode_create()
344 node->page_offset); in hfs_bnode_find()
434 memzero_page(*pagep, node->page_offset, in hfs_bnode_create()
/linux-6.6.21/net/ceph/
Dmessenger.c739 size_t *page_offset, in ceph_msg_data_bio_next() argument
745 *page_offset = bv.bv_offset; in ceph_msg_data_bio_next()
794 size_t *page_offset, in ceph_msg_data_bvecs_next() argument
800 *page_offset = bv.bv_offset; in ceph_msg_data_bvecs_next()
844 cursor->page_offset = data->alignment & ~PAGE_MASK; in ceph_msg_data_pages_cursor_init()
848 BUG_ON(length > SIZE_MAX - cursor->page_offset); in ceph_msg_data_pages_cursor_init()
853 size_t *page_offset, size_t *length) in ceph_msg_data_pages_next() argument
860 BUG_ON(cursor->page_offset >= PAGE_SIZE); in ceph_msg_data_pages_next()
862 *page_offset = cursor->page_offset; in ceph_msg_data_pages_next()
863 *length = min_t(size_t, cursor->resid, PAGE_SIZE - *page_offset); in ceph_msg_data_pages_next()
[all …]
/linux-6.6.21/drivers/infiniband/hw/mlx5/
Dmem.c65 u64 page_offset; in __mlx5_umem_find_best_quantized_pgoff() local
78 page_offset = ib_umem_dma_offset(umem, page_size); in __mlx5_umem_find_best_quantized_pgoff()
79 while (page_offset & ~(u64)(page_offset_mask * (page_size / scale))) { in __mlx5_umem_find_best_quantized_pgoff()
81 page_offset = ib_umem_dma_offset(umem, page_size); in __mlx5_umem_find_best_quantized_pgoff()
92 (unsigned long)page_offset / (page_size / scale); in __mlx5_umem_find_best_quantized_pgoff()
Dsrq_cmd.c17 u32 page_offset = in->page_offset; in get_pas_size() local
21 u32 rq_sz_po = rq_sz + (page_offset * po_quanta); in get_pas_size()
34 MLX5_SET(wq, wq, page_offset, in->page_offset); in set_wq()
47 MLX5_SET(srqc, srqc, page_offset, in->page_offset); in set_srqc()
62 in->page_offset = MLX5_GET(wq, wq, page_offset); in get_wq()
75 in->page_offset = MLX5_GET(srqc, srqc, page_offset); in get_srqc()
111 MLX5_ADAPTER_PAGE_SHIFT, page_offset, \
112 64, &(in)->page_offset))
/linux-6.6.21/drivers/mtd/tests/
Dnandbiterrs.c45 static unsigned page_offset; variable
46 module_param(page_offset, uint, S_IRUGO);
47 MODULE_PARM_DESC(page_offset, "Page number relative to dev start");
358 offset = (loff_t)page_offset * mtd->writesize; in mtd_nandbiterrs_init()
362 page_offset, offset, eraseblock); in mtd_nandbiterrs_init()
/linux-6.6.21/drivers/nvmem/
Drave-sp-eeprom.c161 const unsigned int page_offset = offset % RAVE_SP_EEPROM_PAGE_SIZE; in rave_sp_eeprom_page_access() local
172 if (WARN_ON(data_len > sizeof(page.data) - page_offset)) in rave_sp_eeprom_page_access()
187 memcpy(&page.data[page_offset], data, data_len); in rave_sp_eeprom_page_access()
199 memcpy(data, &page.data[page_offset], data_len); in rave_sp_eeprom_page_access()
/linux-6.6.21/drivers/iommu/iommufd/
Dpages.c206 return iopt_area_iova(area) - area->page_offset + index * PAGE_SIZE; in iopt_area_index_to_iova()
217 return iopt_area_iova(area) - area->page_offset + in iopt_area_index_to_iova_last()
380 unsigned int page_offset = 0; in batch_from_domain() local
386 page_offset = area->page_offset; in batch_from_domain()
393 phys = iommu_iova_to_phys(domain, iova) - page_offset; in batch_from_domain()
396 iova += PAGE_SIZE - page_offset; in batch_from_domain()
397 page_offset = 0; in batch_from_domain()
408 unsigned int page_offset = 0; in raw_pages_from_domain() local
414 page_offset = area->page_offset; in raw_pages_from_domain()
416 phys = iommu_iova_to_phys(domain, iova) - page_offset; in raw_pages_from_domain()
[all …]
/linux-6.6.21/drivers/gpu/drm/qxl/
Dqxl_image.c166 unsigned int page_base, page_offset, out_offset; in qxl_image_init_helper() local
175 page_offset = offset_in_page(out_offset); in qxl_image_init_helper()
176 size = min((int)(PAGE_SIZE - page_offset), remain); in qxl_image_init_helper()
179 k_data = ptr + page_offset; in qxl_image_init_helper()
/linux-6.6.21/fs/hfsplus/
Dbnode.c27 off += node->page_offset; in hfs_bnode_read()
78 off += node->page_offset; in hfs_bnode_write()
106 off += node->page_offset; in hfs_bnode_clear()
130 src += src_node->page_offset; in hfs_bnode_copy()
131 dst += dst_node->page_offset; in hfs_bnode_copy()
184 src += node->page_offset; in hfs_bnode_move()
185 dst += node->page_offset; in hfs_bnode_move()
438 node->page_offset = off & ~PAGE_MASK; in __hfs_bnode_create()
495 node->page_offset); in hfs_bnode_find()
585 memzero_page(*pagep, node->page_offset, in hfs_bnode_create()
Dwrapper.c74 unsigned int page_offset = offset_in_page(buf); in hfsplus_submit_bio() local
75 unsigned int len = min_t(unsigned int, PAGE_SIZE - page_offset, in hfsplus_submit_bio()
78 ret = bio_add_page(bio, virt_to_page(buf), len, page_offset); in hfsplus_submit_bio()
/linux-6.6.21/drivers/net/ethernet/intel/iavf/
Diavf_txrx.c715 rx_bi->page_offset, in iavf_clean_rx_ring()
728 rx_bi->page_offset = 0; in iavf_clean_rx_ring()
880 bi->page_offset = iavf_rx_offset(rx_ring); in iavf_alloc_mapped_page()
935 bi->page_offset, in iavf_alloc_rx_buffers()
942 rx_desc->read.pkt_addr = cpu_to_le64(bi->dma + bi->page_offset); in iavf_alloc_rx_buffers()
1177 new_buff->page_offset = old_buff->page_offset; in iavf_reuse_rx_page()
1224 if (rx_buffer->page_offset > IAVF_LAST_OFFSET) in iavf_can_reuse_rx_page()
1267 rx_buffer->page_offset, size, truesize); in iavf_add_rx_frag()
1271 rx_buffer->page_offset ^= truesize; in iavf_add_rx_frag()
1273 rx_buffer->page_offset += truesize; in iavf_add_rx_frag()
[all …]
/linux-6.6.21/drivers/net/ethernet/sfc/falcon/
Drx.c59 return page_address(buf->page) + buf->page_offset; in ef4_rx_buf_va()
157 unsigned int page_offset; in ef4_init_rx_buffers() local
188 page_offset = sizeof(struct ef4_rx_page_state); in ef4_init_rx_buffers()
195 rx_buf->page_offset = page_offset + efx->rx_ip_align; in ef4_init_rx_buffers()
201 page_offset += efx->rx_page_buf_step; in ef4_init_rx_buffers()
202 } while (page_offset + efx->rx_page_buf_step <= PAGE_SIZE); in ef4_init_rx_buffers()
454 rx_buf->page, rx_buf->page_offset, in ef4_rx_packet_gro()
499 rx_buf->page_offset += hdr_len; in ef4_rx_mk_skb()
504 rx_buf->page, rx_buf->page_offset, in ef4_rx_mk_skb()
586 rx_buf->page_offset += efx->rx_prefix_size; in ef4_rx_packet()
/linux-6.6.21/drivers/infiniband/hw/hfi1/
Dpin_system.c289 unsigned int page_offset; in add_mapping_to_sdma_packet() local
310 page_offset = start - ALIGN_DOWN(start, PAGE_SIZE); in add_mapping_to_sdma_packet()
311 from_this_page = PAGE_SIZE - page_offset; in add_mapping_to_sdma_packet()
327 page_offset, from_this_page, in add_mapping_to_sdma_packet()
338 ret, page_index, page_offset, from_this_page); in add_mapping_to_sdma_packet()
/linux-6.6.21/drivers/net/ethernet/google/gve/
Dgve_rx_dqo.c190 buf_state->page_info.page_offset = 0; in gve_alloc_page_dqo()
406 buf_state->page_info.page_offset); in gve_rx_post_buffers_dqo()
438 buf_state->page_info.page_offset; in gve_try_recycle_buf()
442 buf_state->page_info.page_offset += data_buffer_size; in gve_try_recycle_buf()
443 buf_state->page_info.page_offset &= (PAGE_SIZE - 1); in gve_try_recycle_buf()
448 if (buf_state->page_info.page_offset == in gve_try_recycle_buf()
542 buf_state->page_info.page_offset, in gve_rx_copy_ondemand()
592 buf_state->page_info.page_offset, in gve_rx_append_frags()
644 buf_state->page_info.page_offset, in gve_rx_dqo()
685 buf_state->page_info.page_offset, buf_len, in gve_rx_dqo()
Dgve_rx.c86 page_info->page_offset = 0; in gve_setup_rx_buffer()
162 rx->qpl_copy_pool[j].page_offset = 0; in gve_prefill_rx_pages()
368 u32 offset = page_info->page_offset + page_info->pad; in gve_rx_add_frags()
411 page_info->page_offset ^= PAGE_SIZE / 2; in gve_rx_flip_buff()
456 void *src = page_info->page_address + page_info->page_offset; in gve_rx_copy_to_pool()
489 alloc_page_info.page_offset = 0; in gve_rx_copy_to_pool()
506 dst = copy_page_info->page_address + copy_page_info->page_offset; in gve_rx_copy_to_pool()
516 copy_page_info->page_offset += rx->packet_buffer_size; in gve_rx_copy_to_pool()
517 copy_page_info->page_offset &= (PAGE_SIZE - 1); in gve_rx_copy_to_pool()
756 va = page_info->page_address + page_info->page_offset; in gve_rx()
[all …]
/linux-6.6.21/drivers/net/ethernet/mellanox/mlx4/
Den_rx.c70 frag->page_offset = priv->rx_headroom; in mlx4_alloc_page()
89 frags->page_offset); in mlx4_en_alloc_frags()
148 frags->page_offset = XDP_PACKET_HEADROOM; in mlx4_en_prepare_rx_desc()
488 dma_sync_single_range_for_cpu(priv->ddev, dma, frags->page_offset, in mlx4_en_complete_rx_desc()
491 __skb_fill_page_desc(skb, nr, page, frags->page_offset, in mlx4_en_complete_rx_desc()
496 frags->page_offset ^= PAGE_SIZE / 2; in mlx4_en_complete_rx_desc()
507 frags->page_offset += sz_align; in mlx4_en_complete_rx_desc()
508 release = frags->page_offset + frag_info->frag_size > PAGE_SIZE; in mlx4_en_complete_rx_desc()
755 va = page_address(frags[0].page) + frags[0].page_offset; in mlx4_en_process_rx_cq()
784 dma = frags[0].dma + frags[0].page_offset; in mlx4_en_process_rx_cq()
[all …]
/linux-6.6.21/tools/testing/selftests/powerpc/primitives/
Dload_unaligned_zeropad.c102 static int do_one_test(char *p, int page_offset) in do_one_test() argument
114 …printf("offset %u load_unaligned_zeropad returned 0x%lx, should be 0x%lx\n", page_offset, got, sho… in do_one_test()
/linux-6.6.21/drivers/gpu/drm/panfrost/
Dpanfrost_mmu.c447 pgoff_t page_offset; in panfrost_mmu_map_fault_addr() local
466 page_offset = addr >> PAGE_SHIFT; in panfrost_mmu_map_fault_addr()
467 page_offset -= bomapping->mmnode.start; in panfrost_mmu_map_fault_addr()
493 if (pages[page_offset]) { in panfrost_mmu_map_fault_addr()
502 for (i = page_offset; i < page_offset + NUM_FAULT_PAGES; i++) { in panfrost_mmu_map_fault_addr()
511 sgt = &bo->sgts[page_offset / (SZ_2M / PAGE_SIZE)]; in panfrost_mmu_map_fault_addr()
512 ret = sg_alloc_table_from_pages(sgt, pages + page_offset, in panfrost_mmu_map_fault_addr()

123456789