Searched refs:sq_entries (Results 1 – 11 of 11) sorted by relevance
/linux-6.1.9/io_uring/ |
D | fdinfo.c | 57 unsigned int sq_mask = ctx->sq_entries - 1, cq_mask = ctx->cq_entries - 1; in __io_uring_show_fdinfo() 64 unsigned int sq_entries, cq_entries; in __io_uring_show_fdinfo() local 88 sq_entries = min(sq_tail - sq_head, ctx->sq_entries); in __io_uring_show_fdinfo() 89 for (i = 0; i < sq_entries; i++) { in __io_uring_show_fdinfo()
|
D | io_uring.c | 2208 unsigned head, mask = ctx->sq_entries - 1; in io_get_sqe() 2220 if (likely(head < ctx->sq_entries)) { in io_get_sqe() 2244 ret = left = min3(nr, ctx->sq_entries, entries); in io_submit_sqes() 2458 static unsigned long rings_size(struct io_ring_ctx *ctx, unsigned int sq_entries, in rings_size() argument 2481 sq_array_size = array_size(sizeof(u32), sq_entries); in rings_size() 3295 ctx->sq_entries = p->sq_entries; in io_allocate_scq_urings() 3298 size = rings_size(ctx, p->sq_entries, p->cq_entries, &sq_array_offset); in io_allocate_scq_urings() 3308 rings->sq_ring_mask = p->sq_entries - 1; in io_allocate_scq_urings() 3310 rings->sq_ring_entries = p->sq_entries; in io_allocate_scq_urings() 3314 size = array_size(2 * sizeof(struct io_uring_sqe), p->sq_entries); in io_allocate_scq_urings() [all …]
|
D | tctx.c | 42 concurrency = min(ctx->sq_entries, 4 * num_online_cpus()); in io_init_wq_offload()
|
D | io_uring.h | 238 return READ_ONCE(r->sq.tail) - ctx->cached_sq_head == ctx->sq_entries; in io_sqring_full()
|
/linux-6.1.9/tools/io_uring/ |
D | setup.c | 17 sq->ring_sz = p->sq_off.array + p->sq_entries * sizeof(unsigned); in io_uring_mmap() 30 size = p->sq_entries * sizeof(struct io_uring_sqe); in io_uring_mmap() 46 munmap(sq->sqes, p->sq_entries * sizeof(struct io_uring_sqe)); in io_uring_mmap()
|
D | io_uring-bench.c | 432 ptr = mmap(0, p.sq_off.array + p.sq_entries * sizeof(__u32), in setup_ring() 444 s->sqes = mmap(0, p.sq_entries * sizeof(struct io_uring_sqe), in setup_ring()
|
/linux-6.1.9/include/trace/events/ |
D | io_uring.h | 29 TP_PROTO(int fd, void *ctx, u32 sq_entries, u32 cq_entries, u32 flags), 31 TP_ARGS(fd, ctx, sq_entries, cq_entries, flags), 36 __field( u32, sq_entries ) 44 __entry->sq_entries = sq_entries; 50 __entry->ctx, __entry->fd, __entry->sq_entries,
|
/linux-6.1.9/tools/testing/selftests/net/ |
D | io_uring_zerocopy_tx.c | 175 sq->ring_sz = p->sq_off.array + p->sq_entries * sizeof(unsigned); in io_uring_mmap() 188 size = p->sq_entries * sizeof(struct io_uring_sqe); in io_uring_mmap() 203 munmap(sq->sqes, p->sq_entries * sizeof(struct io_uring_sqe)); in io_uring_mmap()
|
/linux-6.1.9/drivers/net/ethernet/ibm/ehea/ |
D | ehea_main.c | 52 static int sq_entries = EHEA_DEF_ENTRIES_SQ; variable 60 module_param(sq_entries, int, 0); 76 MODULE_PARM_DESC(sq_entries, " Number of entries for the Send Queue " 2254 pr_cfg.max_entries_scq = sq_entries * 2; in ehea_port_res_setup() 2255 pr_cfg.max_entries_sq = sq_entries; in ehea_port_res_setup() 2261 pr_cfg_small_rx.max_entries_scq = sq_entries; in ehea_port_res_setup() 2262 pr_cfg_small_rx.max_entries_sq = sq_entries; in ehea_port_res_setup() 2962 port->sig_comp_iv = sq_entries / 10; in ehea_setup_single_port() 3517 if ((sq_entries < EHEA_MIN_ENTRIES_QP) || in check_module_parm() 3518 (sq_entries > EHEA_MAX_ENTRIES_SQ)) { in check_module_parm()
|
/linux-6.1.9/include/uapi/linux/ |
D | io_uring.h | 437 __u32 sq_entries; member
|
/linux-6.1.9/include/linux/ |
D | io_uring_types.h | 229 unsigned sq_entries; member
|