/linux-6.6.21/io_uring/ |
D | tctx.c | 49 struct io_uring_task *tctx = tsk->io_uring; in __io_uring_free() 57 tsk->io_uring = NULL; in __io_uring_free() 88 task->io_uring = tctx; in io_uring_alloc_task_context() 96 struct io_uring_task *tctx = current->io_uring; in __io_uring_add_tctx_node() 105 tctx = current->io_uring; in __io_uring_add_tctx_node() 148 current->io_uring->last = ctx; in __io_uring_add_tctx_node_from_submit() 157 struct io_uring_task *tctx = current->io_uring; in io_uring_del_tctx_node() 200 struct io_uring_task *tctx = current->io_uring; in io_uring_unreg_ringfd() 270 tctx = current->io_uring; in io_ringfd_register() 316 struct io_uring_task *tctx = current->io_uring; in io_ringfd_unregister()
|
D | cancel.c | 108 WARN_ON_ONCE(!io_wq_current_is_worker() && tctx != current->io_uring); in io_try_cancel() 178 struct io_uring_task *tctx = node->task->io_uring; in __io_async_cancel() 201 struct io_uring_task *tctx = req->task->io_uring; in io_async_cancel() 294 ret = __io_sync_cancel(current->io_uring, &cd, sc.fd); in io_sync_cancel() 318 ret = __io_sync_cancel(current->io_uring, &cd, sc.fd); in io_sync_cancel()
|
D | tctx.h | 27 struct io_uring_task *tctx = current->io_uring; in io_uring_add_tctx_node()
|
D | Makefile | 5 obj-$(CONFIG_IO_URING) += io_uring.o xattr.o nop.o fs.o splice.o \
|
D | io_uring.c | 405 struct io_uring_task *tctx = req->task->io_uring; in io_clean_op() 422 atomic_inc(&req->task->io_uring->inflight_tracked); in io_req_track_inflight() 512 struct io_uring_task *tctx = req->task->io_uring; in io_queue_iowq() 746 struct io_uring_task *tctx = task->io_uring; in io_put_task_remote() 757 task->io_uring->cached_refs++; in io_put_task_local() 780 struct io_uring_task *tctx = task->io_uring; in io_uring_drop_tctx_refs() 1370 struct io_uring_task *tctx = req->task->io_uring; in io_req_normal_work_add() 2457 current->io_uring->cached_refs += left; in io_submit_sqes() 2523 struct io_uring_task *tctx = current->io_uring; in current_pending_io() 3073 struct io_uring_task *tctx = current->io_uring; in io_tctx_exit_cb() [all …]
|
D | io_uring.h | 343 struct io_uring_task *tctx = current->io_uring; in io_get_task_refs()
|
D | sqpoll.c | 443 ret = io_wq_cpu_affinity(sqd->thread->io_uring, mask); in io_sqpoll_wq_cpu_affinity()
|
D | timeout.c | 316 ret = io_try_cancel(req->task->io_uring, &cd, issue_flags); in io_req_task_link_timeout()
|
/linux-6.6.21/include/linux/ |
D | io_uring.h | 69 if (current->io_uring) { in io_uring_files_cancel() 76 if (current->io_uring) in io_uring_task_cancel() 81 if (tsk->io_uring) in io_uring_free()
|
D | io_uring_types.h | 71 struct io_uring { struct 92 struct io_uring sq, cq; argument
|
D | sched.h | 1097 struct io_uring_task *io_uring; member
|
/linux-6.6.21/tools/testing/selftests/net/ |
D | io_uring_zerocopy_tx.c | 107 struct io_uring { struct 157 static int io_uring_register_buffers(struct io_uring *ring, in io_uring_register_buffers() 215 static int io_uring_queue_init(unsigned entries, struct io_uring *ring, in io_uring_queue_init() 236 static int io_uring_submit(struct io_uring *ring) in io_uring_submit() 291 static struct io_uring_sqe *io_uring_get_sqe(struct io_uring *ring) in io_uring_get_sqe() 300 static int io_uring_wait_cqe(struct io_uring *ring, struct io_uring_cqe **cqe_ptr) in io_uring_wait_cqe() 323 static inline void io_uring_cqe_seen(struct io_uring *ring) in io_uring_cqe_seen() 363 struct io_uring ring; in do_tx()
|
/linux-6.6.21/Documentation/block/ |
D | ublk.rst | 47 ``io_uring`` passthrough command; that is why ublk is also one io_uring based 48 block driver. It has been observed that using io_uring passthrough command can 51 done by io_uring, but also the preferred IO handling in ublk server is io_uring 76 # do anything. all IOs are handled by io_uring 119 pthread & io_uring for handling ublk IO), this command is sent to the 127 io_uring). 232 ublk server needs to create per-queue IO pthread & io_uring for handling IO 233 commands via io_uring passthrough. The per-queue IO pthread 246 The following IO commands are communicated via io_uring passthrough command, 262 the IO notification via io_uring.
|
/linux-6.6.21/ |
D | Kbuild | 91 obj-$(CONFIG_IO_URING) += io_uring/
|
D | MAINTAINERS | 11014 F: include/linux/io_uring.h 11016 F: include/trace/events/io_uring.h 11017 F: include/uapi/linux/io_uring.h 11018 F: io_uring/
|
/linux-6.6.21/init/ |
D | init_task.c | 119 .io_uring = NULL,
|
D | Kconfig | 1662 This option enables support for the io_uring interface, enabling
|
/linux-6.6.21/tools/testing/selftests/mm/ |
D | gup_longterm.c | 184 struct io_uring ring; in do_test()
|
D | cow.c | 352 struct io_uring ring; in do_test_iouring()
|
/linux-6.6.21/drivers/block/ |
D | Kconfig | 377 io_uring based userspace block driver. Together with ublk server, ublk 383 can be used in IO path for replacing io_uring cmd, which will become 397 may help security subsystem to audit io_uring command.
|
/linux-6.6.21/scripts/ |
D | Makefile.package | 10 include init io_uring ipc kernel lib mm net rust \
|
/linux-6.6.21/Documentation/admin-guide/sysctl/ |
D | kernel.rst | 456 Prevents all processes from creating new io_uring instances. Enabling this 460 0 All processes can create io_uring instances as normal. This is the 462 1 io_uring creation is disabled (io_uring_setup() will fail with 464 Existing io_uring instances can still be used. See the 466 2 io_uring creation is disabled for all processes. io_uring_setup() 467 always fails with -EPERM. Existing io_uring instances can still be 477 to create an io_uring instance. If io_uring_group is set to -1 (the 479 io_uring instances.
|
/linux-6.6.21/include/trace/events/ |
D | io_uring.h | 3 #define TRACE_SYSTEM io_uring
|
/linux-6.6.21/Documentation/networking/ |
D | napi.rst | 208 ``net.core.busy_read`` sysctls. An io_uring API for NAPI busy polling
|
/linux-6.6.21/kernel/ |
D | fork.c | 2406 p->io_uring = NULL; in copy_process()
|