/linux-6.1.9/io_uring/ |
D | net.h | 32 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags); 37 int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags); 39 int io_send(struct io_kiocb *req, unsigned int issue_flags); 44 int io_recvmsg(struct io_kiocb *req, unsigned int issue_flags); 45 int io_recv(struct io_kiocb *req, unsigned int issue_flags); 50 int io_accept(struct io_kiocb *req, unsigned int issue_flags); 53 int io_socket(struct io_kiocb *req, unsigned int issue_flags); 57 int io_connect(struct io_kiocb *req, unsigned int issue_flags); 59 int io_send_zc(struct io_kiocb *req, unsigned int issue_flags); 60 int io_sendmsg_zc(struct io_kiocb *req, unsigned int issue_flags);
|
D | net.c | 83 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags) in io_shutdown() argument 89 if (issue_flags & IO_URING_F_NONBLOCK) in io_shutdown() 108 static void io_netmsg_recycle(struct io_kiocb *req, unsigned int issue_flags) in io_netmsg_recycle() argument 112 if (!req_has_async_data(req) || issue_flags & IO_URING_F_UNLOCKED) in io_netmsg_recycle() 123 unsigned int issue_flags) in io_msg_alloc_async() argument 129 if (!(issue_flags & IO_URING_F_UNLOCKED) && in io_msg_alloc_async() 154 unsigned int issue_flags) in io_setup_async_msg() argument 160 async_msg = io_msg_alloc_async(req, issue_flags); in io_setup_async_msg() 206 unsigned int issue_flags) in io_setup_async_addr() argument 213 io = io_msg_alloc_async(req, issue_flags); in io_setup_async_addr() [all …]
|
D | openclose.c | 96 int io_openat2(struct io_kiocb *req, unsigned int issue_flags) in io_openat2() argument 110 if (issue_flags & IO_URING_F_NONBLOCK) { in io_openat2() 140 (!resolve_nonblock && (issue_flags & IO_URING_F_NONBLOCK))) in io_openat2() 145 if ((issue_flags & IO_URING_F_NONBLOCK) && !nonblock_set) in io_openat2() 152 ret = io_fixed_fd_install(req, issue_flags, file, in io_openat2() 163 int io_openat(struct io_kiocb *req, unsigned int issue_flags) in io_openat() argument 165 return io_openat2(req, issue_flags); in io_openat() 176 int __io_close_fixed(struct io_ring_ctx *ctx, unsigned int issue_flags, in __io_close_fixed() argument 181 io_ring_submit_lock(ctx, issue_flags); in __io_close_fixed() 183 io_ring_submit_unlock(ctx, issue_flags); in __io_close_fixed() [all …]
|
D | kbuf.h | 37 unsigned int issue_flags); 41 int io_remove_buffers(struct io_kiocb *req, unsigned int issue_flags); 44 int io_provide_buffers(struct io_kiocb *req, unsigned int issue_flags); 49 unsigned int __io_put_kbuf(struct io_kiocb *req, unsigned issue_flags); 51 void io_kbuf_recycle_legacy(struct io_kiocb *req, unsigned issue_flags); 87 static inline void io_kbuf_recycle(struct io_kiocb *req, unsigned issue_flags) in io_kbuf_recycle() argument 90 io_kbuf_recycle_legacy(req, issue_flags); in io_kbuf_recycle() 125 unsigned issue_flags) in io_put_kbuf() argument 130 return __io_put_kbuf(req, issue_flags); in io_put_kbuf()
|
D | msg_ring.c | 44 unsigned int issue_flags) in io_double_unlock_ctx() argument 46 if (issue_flags & IO_URING_F_UNLOCKED) in io_double_unlock_ctx() 53 unsigned int issue_flags) in io_double_lock_ctx() argument 60 if (!(issue_flags & IO_URING_F_UNLOCKED)) { in io_double_lock_ctx() 78 static int io_msg_send_fd(struct io_kiocb *req, unsigned int issue_flags) in io_msg_send_fd() argument 92 ret = io_double_lock_ctx(ctx, target_ctx, issue_flags); in io_msg_send_fd() 126 io_double_unlock_ctx(ctx, target_ctx, issue_flags); in io_msg_send_fd() 149 int io_msg_ring(struct io_kiocb *req, unsigned int issue_flags) in io_msg_ring() argument 163 ret = io_msg_send_fd(req, issue_flags); in io_msg_ring()
|
D | fs.h | 4 int io_renameat(struct io_kiocb *req, unsigned int issue_flags); 8 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags); 12 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags); 16 int io_symlinkat(struct io_kiocb *req, unsigned int issue_flags); 19 int io_linkat(struct io_kiocb *req, unsigned int issue_flags);
|
D | kbuf.c | 50 void io_kbuf_recycle_legacy(struct io_kiocb *req, unsigned issue_flags) in io_kbuf_recycle_legacy() argument 65 io_ring_submit_lock(ctx, issue_flags); in io_kbuf_recycle_legacy() 73 io_ring_submit_unlock(ctx, issue_flags); in io_kbuf_recycle_legacy() 77 unsigned int __io_put_kbuf(struct io_kiocb *req, unsigned issue_flags) in __io_put_kbuf() argument 96 } else if (issue_flags & IO_URING_F_UNLOCKED) { in __io_put_kbuf() 130 unsigned int issue_flags) in io_ring_buffer_select() argument 154 if (issue_flags & IO_URING_F_UNLOCKED || !file_can_poll(req->file)) { in io_ring_buffer_select() 172 unsigned int issue_flags) in io_buffer_select() argument 178 io_ring_submit_lock(req->ctx, issue_flags); in io_buffer_select() 183 ret = io_ring_buffer_select(req, len, bl, issue_flags); in io_buffer_select() [all …]
|
D | xattr.c | 107 int io_fgetxattr(struct io_kiocb *req, unsigned int issue_flags) in io_fgetxattr() argument 112 if (issue_flags & IO_URING_F_NONBLOCK) in io_fgetxattr() 123 int io_getxattr(struct io_kiocb *req, unsigned int issue_flags) in io_getxattr() argument 130 if (issue_flags & IO_URING_F_NONBLOCK) in io_getxattr() 208 static int __io_setxattr(struct io_kiocb *req, unsigned int issue_flags, in __io_setxattr() argument 223 int io_fsetxattr(struct io_kiocb *req, unsigned int issue_flags) in io_fsetxattr() argument 227 if (issue_flags & IO_URING_F_NONBLOCK) in io_fsetxattr() 230 ret = __io_setxattr(req, issue_flags, &req->file->f_path); in io_fsetxattr() 235 int io_setxattr(struct io_kiocb *req, unsigned int issue_flags) in io_setxattr() argument 242 if (issue_flags & IO_URING_F_NONBLOCK) in io_setxattr() [all …]
|
D | openclose.h | 3 int __io_close_fixed(struct io_ring_ctx *ctx, unsigned int issue_flags, 7 int io_openat(struct io_kiocb *req, unsigned int issue_flags); 11 int io_openat2(struct io_kiocb *req, unsigned int issue_flags); 14 int io_close(struct io_kiocb *req, unsigned int issue_flags);
|
D | xattr.h | 6 int io_fsetxattr(struct io_kiocb *req, unsigned int issue_flags); 9 int io_setxattr(struct io_kiocb *req, unsigned int issue_flags); 12 int io_fgetxattr(struct io_kiocb *req, unsigned int issue_flags); 15 int io_getxattr(struct io_kiocb *req, unsigned int issue_flags);
|
D | poll.h | 27 int io_poll_add(struct io_kiocb *req, unsigned int issue_flags); 30 int io_poll_remove(struct io_kiocb *req, unsigned int issue_flags); 34 unsigned issue_flags); 35 int io_arm_poll_handler(struct io_kiocb *req, unsigned issue_flags);
|
D | sync.c | 38 int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags) in io_sync_file_range() argument 44 if (issue_flags & IO_URING_F_NONBLOCK) in io_sync_file_range() 68 int io_fsync(struct io_kiocb *req, unsigned int issue_flags) in io_fsync() argument 75 if (issue_flags & IO_URING_F_NONBLOCK) in io_fsync() 97 int io_fallocate(struct io_kiocb *req, unsigned int issue_flags) in io_fallocate() argument 103 if (issue_flags & IO_URING_F_NONBLOCK) in io_fallocate()
|
D | splice.c | 47 int io_tee(struct io_kiocb *req, unsigned int issue_flags) in io_tee() argument 55 if (issue_flags & IO_URING_F_NONBLOCK) in io_tee() 59 in = io_file_get_fixed(req, sp->splice_fd_in, issue_flags); in io_tee() 88 int io_splice(struct io_kiocb *req, unsigned int issue_flags) in io_splice() argument 97 if (issue_flags & IO_URING_F_NONBLOCK) in io_splice() 101 in = io_file_get_fixed(req, sp->splice_fd_in, issue_flags); in io_splice()
|
D | fs.c | 80 int io_renameat(struct io_kiocb *req, unsigned int issue_flags) in io_renameat() argument 85 if (issue_flags & IO_URING_F_NONBLOCK) in io_renameat() 129 int io_unlinkat(struct io_kiocb *req, unsigned int issue_flags) in io_unlinkat() argument 134 if (issue_flags & IO_URING_F_NONBLOCK) in io_unlinkat() 176 int io_mkdirat(struct io_kiocb *req, unsigned int issue_flags) in io_mkdirat() argument 181 if (issue_flags & IO_URING_F_NONBLOCK) in io_mkdirat() 226 int io_symlinkat(struct io_kiocb *req, unsigned int issue_flags) in io_symlinkat() argument 231 if (issue_flags & IO_URING_F_NONBLOCK) in io_symlinkat() 271 int io_linkat(struct io_kiocb *req, unsigned int issue_flags) in io_linkat() argument 276 if (issue_flags & IO_URING_F_NONBLOCK) in io_linkat()
|
D | sync.h | 4 int io_sync_file_range(struct io_kiocb *req, unsigned int issue_flags); 7 int io_fsync(struct io_kiocb *req, unsigned int issue_flags); 9 int io_fallocate(struct io_kiocb *req, unsigned int issue_flags);
|
D | cancel.c | 82 unsigned issue_flags) in io_try_cancel() argument 97 ret = io_poll_cancel(ctx, cd, issue_flags); in io_try_cancel() 132 unsigned int issue_flags) in __io_async_cancel() argument 140 ret = io_try_cancel(tctx, cd, issue_flags); in __io_async_cancel() 149 io_ring_submit_lock(ctx, issue_flags); in __io_async_cancel() 161 io_ring_submit_unlock(ctx, issue_flags); in __io_async_cancel() 165 int io_async_cancel(struct io_kiocb *req, unsigned int issue_flags) in io_async_cancel() argument 182 issue_flags); in io_async_cancel() 193 ret = __io_async_cancel(&cd, tctx, issue_flags); in io_async_cancel()
|
D | advise.c | 48 int io_madvise(struct io_kiocb *req, unsigned int issue_flags) in io_madvise() argument 54 if (issue_flags & IO_URING_F_NONBLOCK) in io_madvise() 78 int io_fadvise(struct io_kiocb *req, unsigned int issue_flags) in io_fadvise() argument 83 if (issue_flags & IO_URING_F_NONBLOCK) { in io_fadvise()
|
D | poll.c | 584 unsigned issue_flags) in __io_arm_poll_handler() argument 609 ipt->owning = issue_flags & IO_URING_F_UNLOCKED; in __io_arm_poll_handler() 613 if (issue_flags & IO_URING_F_UNLOCKED) in __io_arm_poll_handler() 672 unsigned issue_flags) in io_req_alloc_apoll() argument 681 } else if (!(issue_flags & IO_URING_F_UNLOCKED) && in io_req_alloc_apoll() 694 int io_arm_poll_handler(struct io_kiocb *req, unsigned issue_flags) in io_arm_poll_handler() argument 729 apoll = io_req_alloc_apoll(req, issue_flags); in io_arm_poll_handler() 735 io_kbuf_recycle(req, issue_flags); in io_arm_poll_handler() 737 ret = __io_arm_poll_handler(req, &apoll->poll, &ipt, mask, issue_flags); in io_arm_poll_handler() 873 unsigned issue_flags) in io_poll_cancel() argument [all …]
|
D | uring_cmd.c | 104 int io_uring_cmd(struct io_kiocb *req, unsigned int issue_flags) in io_uring_cmd() argument 119 issue_flags |= IO_URING_F_SQE128; in io_uring_cmd() 121 issue_flags |= IO_URING_F_CQE32; in io_uring_cmd() 123 issue_flags |= IO_URING_F_IOPOLL; in io_uring_cmd() 131 ret = file->f_op->uring_cmd(ioucmd, issue_flags); in io_uring_cmd()
|
D | splice.h | 4 int io_tee(struct io_kiocb *req, unsigned int issue_flags); 7 int io_splice(struct io_kiocb *req, unsigned int issue_flags);
|
D | advise.h | 4 int io_madvise(struct io_kiocb *req, unsigned int issue_flags); 7 int io_fadvise(struct io_kiocb *req, unsigned int issue_flags);
|
D | io_uring.h | 34 void __io_req_complete(struct io_kiocb *req, unsigned issue_flags); 46 unsigned issue_flags); 183 unsigned issue_flags) in io_ring_submit_unlock() argument 186 if (issue_flags & IO_URING_F_UNLOCKED) in io_ring_submit_unlock() 191 unsigned issue_flags) in io_ring_submit_lock() argument 199 if (issue_flags & IO_URING_F_UNLOCKED) in io_ring_submit_lock()
|
D | rw.c | 324 unsigned int issue_flags) in kiocb_done() argument 339 io_put_kbuf(req, issue_flags)); in kiocb_done() 358 unsigned int issue_flags) in __io_import_iovec() argument 381 buf = io_buffer_select(req, &sqe_len, issue_flags); in __io_import_iovec() 404 unsigned int issue_flags) in io_import_iovec() argument 406 *iovec = __io_import_iovec(rw, req, s, issue_flags); in io_import_iovec() 695 int io_read(struct io_kiocb *req, unsigned int issue_flags) in io_read() argument 701 bool force_nonblock = issue_flags & IO_URING_F_NONBLOCK; in io_read() 707 ret = io_import_iovec(READ, req, &iovec, s, issue_flags); in io_read() 719 ret = io_import_iovec(READ, req, &iovec, s, issue_flags); in io_read() [all …]
|
D | cancel.h | 17 int io_async_cancel(struct io_kiocb *req, unsigned int issue_flags); 20 unsigned int issue_flags);
|
/linux-6.1.9/drivers/nvme/host/ |
D | ioctl.c | 473 struct io_uring_cmd *ioucmd, unsigned int issue_flags, bool vec) in nvme_uring_cmd_io() argument 516 if (issue_flags & IO_URING_F_NONBLOCK) { in nvme_uring_cmd_io() 520 if (issue_flags & IO_URING_F_IOPOLL) in nvme_uring_cmd_io() 537 if (issue_flags & IO_URING_F_IOPOLL && rq_flags & REQ_POLLED) { in nvme_uring_cmd_io() 654 static int nvme_uring_cmd_checks(unsigned int issue_flags) in nvme_uring_cmd_checks() argument 658 if ((issue_flags & (IO_URING_F_SQE128|IO_URING_F_CQE32)) != in nvme_uring_cmd_checks() 665 unsigned int issue_flags) in nvme_ns_uring_cmd() argument 672 ret = nvme_uring_cmd_checks(issue_flags); in nvme_ns_uring_cmd() 678 ret = nvme_uring_cmd_io(ctrl, ns, ioucmd, issue_flags, false); in nvme_ns_uring_cmd() 681 ret = nvme_uring_cmd_io(ctrl, ns, ioucmd, issue_flags, true); in nvme_ns_uring_cmd() [all …]
|