/linux-6.6.21/drivers/gpu/drm/i915/ |
D | i915_query.c | 25 if (copy_from_user(query_hdr, u64_to_user_ptr(query_item->data_ptr), in copy_query_item() 68 if (copy_to_user(u64_to_user_ptr(query_item->data_ptr), in fill_topology_info() 72 if (copy_to_user(u64_to_user_ptr(query_item->data_ptr + sizeof(topo)), in fill_topology_info() 76 if (intel_sseu_copy_ssmask_to_user(u64_to_user_ptr(query_item->data_ptr + in fill_topology_info() 81 if (intel_sseu_copy_eumask_to_user(u64_to_user_ptr(query_item->data_ptr + in fill_topology_info() 132 u64_to_user_ptr(query_item->data_ptr); in query_engine_info() 200 u32 __user *p = u64_to_user_ptr(user_regs_ptr); in copy_perf_config_registers_or_number() 230 u64_to_user_ptr(query_item->data_ptr); in query_perf_config_data() 232 u64_to_user_ptr(query_item->data_ptr + in query_perf_config_data() 376 u64_to_user_ptr(query_item->data_ptr); in query_perf_config_list() [all …]
|
D | i915_user_extensions.c | 57 ext = u64_to_user_ptr(next); in i915_user_extensions()
|
D | i915_gem.c | 254 user_data = u64_to_user_ptr(args->data_ptr); in i915_gem_shmem_pread() 406 user_data = u64_to_user_ptr(args->data_ptr); in i915_gem_gtt_pread() 475 if (!access_ok(u64_to_user_ptr(args->data_ptr), in i915_gem_pread_ioctl() 586 user_data = u64_to_user_ptr(args->data_ptr); in i915_gem_gtt_pwrite_fast() 701 user_data = u64_to_user_ptr(args->data_ptr); in i915_gem_shmem_pwrite() 757 if (!access_ok(u64_to_user_ptr(args->data_ptr), args->size)) in i915_gem_pwrite_ioctl()
|
/linux-6.6.21/io_uring/ |
D | fs.c | 61 oldf = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_renameat_prep() 62 newf = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in io_renameat_prep() 120 fname = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_unlinkat_prep() 167 fname = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_mkdirat_prep() 209 oldpath = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_symlinkat_prep() 210 newpath = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in io_symlinkat_prep() 253 oldf = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_linkat_prep() 254 newf = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in io_linkat_prep()
|
D | xattr.c | 56 name = u64_to_user_ptr(READ_ONCE(sqe->addr)); in __io_getxattr_prep() 57 ix->ctx.cvalue = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in __io_getxattr_prep() 97 path = u64_to_user_ptr(READ_ONCE(sqe->addr3)); in io_getxattr_prep() 159 name = u64_to_user_ptr(READ_ONCE(sqe->addr)); in __io_setxattr_prep() 160 ix->ctx.cvalue = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in __io_setxattr_prep() 190 path = u64_to_user_ptr(READ_ONCE(sqe->addr3)); in io_setxattr_prep()
|
D | statx.c | 35 path = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_statx_prep() 36 sx->buffer = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in io_statx_prep()
|
D | rsrc.c | 101 dst->iov_base = u64_to_user_ptr((u64)ciov.iov_base); in io_copy_iov() 359 u64 __user *tags = u64_to_user_ptr(up->tags); in __io_sqe_files_update() 360 __s32 __user *fds = u64_to_user_ptr(up->data); in __io_sqe_files_update() 434 u64 __user *tags = u64_to_user_ptr(up->tags); in __io_sqe_buffers_update() 435 struct iovec iov, __user *iovs = u64_to_user_ptr(up->data); in __io_sqe_buffers_update() 554 return io_sqe_files_register(ctx, u64_to_user_ptr(rr.data), in io_register_rsrc() 555 rr.nr, u64_to_user_ptr(rr.tags)); in io_register_rsrc() 559 return io_sqe_buffers_register(ctx, u64_to_user_ptr(rr.data), in io_register_rsrc() 560 rr.nr, u64_to_user_ptr(rr.tags)); in io_register_rsrc() 586 __s32 __user *fds = u64_to_user_ptr(up->arg); in io_files_update_with_index_alloc()
|
D | epoll.c | 38 ev = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_epoll_ctl_prep()
|
D | openclose.c | 61 fname = u64_to_user_ptr(READ_ONCE(sqe->addr)); in __io_openat_prep() 97 how = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in io_openat2_prep()
|
/linux-6.6.21/net/bpf/ |
D | bpf_dummy_struct_ops.c | 36 ctx_in = u64_to_user_ptr(kattr->test.ctx_in); in dummy_ops_init_args() 41 u_state = u64_to_user_ptr(args->args[0]); in dummy_ops_init_args() 56 u_state = u64_to_user_ptr(args->args[0]); in dummy_ops_copy_args()
|
/linux-6.6.21/drivers/gpu/drm/ |
D | drm_syncobj.c | 1229 u64_to_user_ptr(timeline_wait->points), in drm_syncobj_array_wait() 1316 u64_to_user_ptr(args->handles), in drm_syncobj_wait_ioctl() 1350 u64_to_user_ptr(args->handles), in drm_syncobj_timeline_wait_ioctl() 1477 u64_to_user_ptr(args->handles), in drm_syncobj_reset_ioctl() 1510 u64_to_user_ptr(args->handles), in drm_syncobj_signal_ioctl() 1548 u64_to_user_ptr(args->handles), in drm_syncobj_timeline_signal_ioctl() 1560 if (!u64_to_user_ptr(args->points)) { in drm_syncobj_timeline_signal_ioctl() 1562 } else if (copy_from_user(points, u64_to_user_ptr(args->points), in drm_syncobj_timeline_signal_ioctl() 1605 uint64_t __user *points = u64_to_user_ptr(args->points); in drm_syncobj_query_ioctl() 1619 u64_to_user_ptr(args->handles), in drm_syncobj_query_ioctl()
|
D | drm_property.c | 480 values_ptr = u64_to_user_ptr(out_resp->values_ptr); in drm_mode_getproperty_ioctl() 491 enum_ptr = u64_to_user_ptr(out_resp->enum_blob_ptr); in drm_mode_getproperty_ioctl() 769 if (copy_to_user(u64_to_user_ptr(out_resp->data), in drm_mode_getblob_ioctl() 798 u64_to_user_ptr(out_resp->data), in drm_mode_createblob_ioctl()
|
D | drm_mode_config.c | 113 fb_id = u64_to_user_ptr(card_res->fb_id_ptr); in drm_mode_getresources() 131 crtc_id = u64_to_user_ptr(card_res->crtc_id_ptr); in drm_mode_getresources() 143 encoder_id = u64_to_user_ptr(card_res->encoder_id_ptr); in drm_mode_getresources() 154 connector_id = u64_to_user_ptr(card_res->connector_id_ptr); in drm_mode_getresources()
|
/linux-6.6.21/drivers/gpu/drm/qxl/ |
D | qxl_ioctl.c | 165 if (!access_ok(u64_to_user_ptr(cmd->command), in qxl_process_single_command() 187 u64_to_user_ptr(cmd->command), cmd->command_size); in qxl_process_single_command() 206 struct drm_qxl_reloc __user *u = u64_to_user_ptr(cmd->relocs); in qxl_process_single_command() 283 u64_to_user_ptr(execbuffer->commands); in qxl_execbuffer_ioctl()
|
/linux-6.6.21/drivers/infiniband/core/ |
D | uverbs_ioctl.c | 141 return ib_is_buffer_cleared(u64_to_user_ptr(uattr->data) + len, in uverbs_is_attr_cleared() 201 ret = copy_from_user(idr_vals, u64_to_user_ptr(uattr->data), in uverbs_process_idrs_array() 297 if (copy_from_user(p, u64_to_user_ptr(uattr->data), in uverbs_process_attr() 716 udata->inbuf = u64_to_user_ptr(in->ptr_attr.data); in uverbs_fill_udata() 723 udata->outbuf = u64_to_user_ptr(out->ptr_attr.data); in uverbs_fill_udata() 741 if (copy_to_user(u64_to_user_ptr(attr->ptr_attr.data), from, min_size)) in uverbs_copy_to() 819 if (clear_user(u64_to_user_ptr(attr->ptr_attr.data) + size, in uverbs_copy_to_struct_or_zero()
|
D | ucma.c | 402 if (copy_to_user(u64_to_user_ptr(cmd.response), in ucma_get_event() 472 if (copy_to_user(u64_to_user_ptr(cmd.response), in ucma_create_id() 618 if (copy_to_user(u64_to_user_ptr(cmd.response), in ucma_destroy_id() 862 if (copy_to_user(u64_to_user_ptr(cmd.response), &resp, in ucma_query_route() 1008 response = u64_to_user_ptr(cmd.response); in ucma_query() 1245 if (copy_to_user(u64_to_user_ptr(cmd.response), in ucma_init_qp_attr() 1397 optval = memdup_user(u64_to_user_ptr(cmd.optval), in ucma_set_option() 1493 if (copy_to_user(u64_to_user_ptr(cmd->response), in ucma_process_join() 1603 if (copy_to_user(u64_to_user_ptr(cmd.response), in ucma_leave_multicast() 1675 if (copy_to_user(u64_to_user_ptr(cmd.response), in ucma_migrate_id()
|
/linux-6.6.21/kernel/bpf/ |
D | syscall.c | 1433 void __user *ukey = u64_to_user_ptr(attr->key); in map_lookup_elem() 1434 void __user *uvalue = u64_to_user_ptr(attr->value); in map_lookup_elem() 1619 void __user *ukey = u64_to_user_ptr(attr->key); in map_get_next_key() 1620 void __user *unext_key = u64_to_user_ptr(attr->next_key); in map_get_next_key() 1685 void __user *keys = u64_to_user_ptr(attr->batch.keys); in generic_map_delete_batch() 1742 void __user *values = u64_to_user_ptr(attr->batch.values); in generic_map_update_batch() 1743 void __user *keys = u64_to_user_ptr(attr->batch.keys); in generic_map_update_batch() 1804 void __user *uobatch = u64_to_user_ptr(attr->batch.out_batch); in generic_map_lookup_batch() 1805 void __user *ubatch = u64_to_user_ptr(attr->batch.in_batch); in generic_map_lookup_batch() 1806 void __user *values = u64_to_user_ptr(attr->batch.values); in generic_map_lookup_batch() [all …]
|
D | mprog.c | 420 uprog_id = u64_to_user_ptr(attr->query.prog_ids); in bpf_mprog_query() 421 uprog_flags = u64_to_user_ptr(attr->query.prog_attach_flags); in bpf_mprog_query() 422 ulink_id = u64_to_user_ptr(attr->query.link_ids); in bpf_mprog_query() 423 ulink_flags = u64_to_user_ptr(attr->query.link_attach_flags); in bpf_mprog_query()
|
/linux-6.6.21/drivers/gpu/drm/virtio/ |
D | virtgpu_submit.c | 118 u64_to_user_ptr(address), in virtio_gpu_parse_deps() 208 u64_to_user_ptr(address), in virtio_gpu_parse_post_deps() 311 if (copy_from_user(bo_handles, u64_to_user_ptr(exbuf->bo_handles), in virtio_gpu_init_submit_buflist() 416 submit->buf = vmemdup_user(u64_to_user_ptr(exbuf->command), exbuf->size); in virtio_gpu_init_submit()
|
D | virtgpu_ioctl.c | 113 if (copy_to_user(u64_to_user_ptr(param->value), &value, sizeof(int))) in virtio_gpu_getparam_ioctl() 423 if (copy_to_user(u64_to_user_ptr(args->addr), ptr, size)) in virtio_gpu_get_caps_ioctl() 510 buf = memdup_user(u64_to_user_ptr(rc_blob->cmd), in virtio_gpu_resource_create_blob_ioctl() 586 ctx_set_params = memdup_user(u64_to_user_ptr(args->ctx_set_params), in virtio_gpu_context_init_ioctl()
|
/linux-6.6.21/drivers/iommu/iommufd/ |
D | ioas.c | 80 ranges = u64_to_user_ptr(cmd->allowed_iovas); in iommufd_ioas_iova_ranges() 160 u64_to_user_ptr(cmd->allowed_iovas), in iommufd_ioas_allow_iovas() 223 u64_to_user_ptr(cmd->user_va), cmd->length, in iommufd_ioas_map()
|
/linux-6.6.21/drivers/gpu/drm/i915/gem/ |
D | i915_gem_phys.c | 148 char __user *user_data = u64_to_user_ptr(args->data_ptr); in i915_gem_object_pwrite_phys() 179 char __user *user_data = u64_to_user_ptr(args->data_ptr); in i915_gem_object_pread_phys()
|
/linux-6.6.21/drivers/gpu/drm/etnaviv/ |
D | etnaviv_gem_submit.c | 480 ret = copy_from_user(bos, u64_to_user_ptr(args->bos), in etnaviv_ioctl_gem_submit() 487 ret = copy_from_user(relocs, u64_to_user_ptr(args->relocs), in etnaviv_ioctl_gem_submit() 494 ret = copy_from_user(pmrs, u64_to_user_ptr(args->pmrs), in etnaviv_ioctl_gem_submit() 501 ret = copy_from_user(stream, u64_to_user_ptr(args->stream), in etnaviv_ioctl_gem_submit()
|
/linux-6.6.21/drivers/gpu/drm/msm/ |
D | msm_gem_submit.c | 121 u64_to_user_ptr(args->bos + (i * sizeof(submit_bo))); in submit_lookup_objects() 190 u64_to_user_ptr(args->cmds + (i * sizeof(submit_cmd))); in submit_lookup_cmds() 222 userptr = u64_to_user_ptr(submit_cmd.relocs); in submit_lookup_cmds() 598 u64_to_user_ptr(address), in msm_parse_deps() 672 u64_to_user_ptr(address), in msm_parse_post_deps()
|
/linux-6.6.21/include/linux/ |
D | bpfptr.h | 31 return USER_BPFPTR(u64_to_user_ptr(addr)); in make_bpfptr()
|