Lines Matching refs:issue_flags

83 int io_shutdown(struct io_kiocb *req, unsigned int issue_flags)  in io_shutdown()  argument
89 if (issue_flags & IO_URING_F_NONBLOCK) in io_shutdown()
108 static void io_netmsg_recycle(struct io_kiocb *req, unsigned int issue_flags) in io_netmsg_recycle() argument
112 if (!req_has_async_data(req) || issue_flags & IO_URING_F_UNLOCKED) in io_netmsg_recycle()
123 unsigned int issue_flags) in io_msg_alloc_async() argument
129 if (!(issue_flags & IO_URING_F_UNLOCKED) && in io_msg_alloc_async()
154 unsigned int issue_flags) in io_setup_async_msg() argument
160 async_msg = io_msg_alloc_async(req, issue_flags); in io_setup_async_msg()
206 unsigned int issue_flags) in io_setup_async_addr() argument
213 io = io_msg_alloc_async(req, issue_flags); in io_setup_async_addr()
269 int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags) in io_sendmsg() argument
293 return io_setup_async_msg(req, kmsg, issue_flags); in io_sendmsg()
296 if (issue_flags & IO_URING_F_NONBLOCK) in io_sendmsg()
304 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_sendmsg()
305 return io_setup_async_msg(req, kmsg, issue_flags); in io_sendmsg()
309 return io_setup_async_msg(req, kmsg, issue_flags); in io_sendmsg()
319 io_netmsg_recycle(req, issue_flags); in io_sendmsg()
328 int io_send(struct io_kiocb *req, unsigned int issue_flags) in io_send() argument
361 return io_setup_async_addr(req, &__address, issue_flags); in io_send()
372 if (issue_flags & IO_URING_F_NONBLOCK) in io_send()
380 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_send()
381 return io_setup_async_addr(req, &__address, issue_flags); in io_send()
388 return io_setup_async_addr(req, &__address, issue_flags); in io_send()
605 unsigned issue_flags) in io_recv_finish() argument
628 if (issue_flags & IO_URING_F_MULTISHOT) in io_recv_finish()
719 int io_recvmsg(struct io_kiocb *req, unsigned int issue_flags) in io_recvmsg() argument
727 bool force_nonblock = issue_flags & IO_URING_F_NONBLOCK; in io_recvmsg()
745 return io_setup_async_msg(req, kmsg, issue_flags); in io_recvmsg()
752 buf = io_buffer_select(req, &len, issue_flags); in io_recvmsg()
759 io_kbuf_recycle(req, issue_flags); in io_recvmsg()
786 ret = io_setup_async_msg(req, kmsg, issue_flags); in io_recvmsg()
787 if (ret == -EAGAIN && (issue_flags & IO_URING_F_MULTISHOT)) { in io_recvmsg()
788 io_kbuf_recycle(req, issue_flags); in io_recvmsg()
796 return io_setup_async_msg(req, kmsg, issue_flags); in io_recvmsg()
810 io_kbuf_recycle(req, issue_flags); in io_recvmsg()
812 cflags = io_put_kbuf(req, issue_flags); in io_recvmsg()
816 if (!io_recv_finish(req, &ret, cflags, mshot_finished, issue_flags)) in io_recvmsg()
823 io_netmsg_recycle(req, issue_flags); in io_recvmsg()
830 int io_recv(struct io_kiocb *req, unsigned int issue_flags) in io_recv() argument
839 bool force_nonblock = issue_flags & IO_URING_F_NONBLOCK; in io_recv()
854 buf = io_buffer_select(req, &len, issue_flags); in io_recv()
882 if (issue_flags & IO_URING_F_MULTISHOT) { in io_recv()
883 io_kbuf_recycle(req, issue_flags); in io_recv()
909 io_kbuf_recycle(req, issue_flags); in io_recv()
911 cflags = io_put_kbuf(req, issue_flags); in io_recv()
915 if (!io_recv_finish(req, &ret, cflags, ret <= 0, issue_flags)) in io_recv()
1060 int io_send_zc(struct io_kiocb *req, unsigned int issue_flags) in io_send_zc() argument
1097 return io_setup_async_addr(req, &__address, issue_flags); in io_send_zc()
1117 if (issue_flags & IO_URING_F_NONBLOCK) in io_send_zc()
1127 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_send_zc()
1128 return io_setup_async_addr(req, &__address, issue_flags); in io_send_zc()
1135 return io_setup_async_addr(req, &__address, issue_flags); in io_send_zc()
1151 if (!(issue_flags & IO_URING_F_UNLOCKED)) { in io_send_zc()
1159 int io_sendmsg_zc(struct io_kiocb *req, unsigned int issue_flags) in io_sendmsg_zc() argument
1184 return io_setup_async_msg(req, kmsg, issue_flags); in io_sendmsg_zc()
1187 if (issue_flags & IO_URING_F_NONBLOCK) in io_sendmsg_zc()
1197 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_sendmsg_zc()
1198 return io_setup_async_msg(req, kmsg, issue_flags); in io_sendmsg_zc()
1203 return io_setup_async_msg(req, kmsg, issue_flags); in io_sendmsg_zc()
1215 io_netmsg_recycle(req, issue_flags); in io_sendmsg_zc()
1225 if (!(issue_flags & IO_URING_F_UNLOCKED)) { in io_sendmsg_zc()
1278 int io_accept(struct io_kiocb *req, unsigned int issue_flags) in io_accept() argument
1282 bool force_nonblock = issue_flags & IO_URING_F_NONBLOCK; in io_accept()
1306 if (issue_flags & IO_URING_F_MULTISHOT) in io_accept()
1317 ret = io_fixed_fd_install(req, issue_flags, file, in io_accept()
1331 return (issue_flags & IO_URING_F_MULTISHOT) ? IOU_STOP_MULTISHOT : IOU_OK; in io_accept()
1355 int io_socket(struct io_kiocb *req, unsigned int issue_flags) in io_socket() argument
1372 if (ret == -EAGAIN && (issue_flags & IO_URING_F_NONBLOCK)) in io_socket()
1381 ret = io_fixed_fd_install(req, issue_flags, file, in io_socket()
1409 int io_connect(struct io_kiocb *req, unsigned int issue_flags) in io_connect() argument
1415 bool force_nonblock = issue_flags & IO_URING_F_NONBLOCK; in io_connect()