Home
last modified time | relevance | path

Searched refs:sqes (Results 1 – 10 of 10) sorted by relevance

/linux-6.1.9/tools/io_uring/
Dsetup.c31 sq->sqes = mmap(0, size, PROT_READ | PROT_WRITE, in io_uring_mmap()
34 if (sq->sqes == MAP_FAILED) { in io_uring_mmap()
46 munmap(sq->sqes, p->sq_entries * sizeof(struct io_uring_sqe)); in io_uring_mmap()
103 munmap(sq->sqes, *sq->kring_entries * sizeof(struct io_uring_sqe)); in io_uring_queue_exit()
Dio_uring-bench.c77 struct io_uring_sqe *sqes; member
145 struct io_uring_sqe *sqe = &s->sqes[index]; in init_io()
444 s->sqes = mmap(0, p.sq_entries * sizeof(struct io_uring_sqe), in setup_ring()
447 printf("sqes ptr = 0x%p\n", s->sqes); in setup_ring()
Dqueue.c153 sqe = &sq->sqes[sq->sqe_tail & *sq->kring_mask]; in io_uring_get_sqe()
Dliburing.h27 struct io_uring_sqe *sqes; member
/linux-6.1.9/tools/testing/selftests/net/
Dio_uring_zerocopy_tx.c88 struct io_uring_sqe *sqes; member
189 sq->sqes = mmap(0, size, PROT_READ | PROT_WRITE, in io_uring_mmap()
191 if (sq->sqes == MAP_FAILED) { in io_uring_mmap()
203 munmap(sq->sqes, p->sq_entries * sizeof(struct io_uring_sqe)); in io_uring_mmap()
297 return &sq->sqes[sq->sqe_tail++ & *sq->kring_mask]; in io_uring_get_sqe()
/linux-6.1.9/drivers/nvme/target/
Dpassthru.c133 id->sqes = min_t(__u8, ((0x6 << 4) | 0x6), id->sqes); in nvmet_passthru_override_id_ctrl()
Dadmin-cmd.c428 id->sqes = (0x6 << 4) | 0x6; in nvmet_execute_identify_ctrl()
/linux-6.1.9/drivers/nvme/host/
Dapple.c134 struct nvme_command *sqes; member
301 memcpy(&q->sqes[tag], cmd, sizeof(*cmd)); in apple_nvme_submit_cmd()
1291 q->sqes = dmam_alloc_coherent(anv->dev, in apple_nvme_queue_alloc()
1294 if (!q->sqes) in apple_nvme_queue_alloc()
Dpci.c35 #define SQ_SIZE(q) ((q)->q_depth << (q)->sqes)
205 u8 sqes; member
499 memcpy(nvmeq->sq_cmds + (nvmeq->sq_tail << nvmeq->sqes), in nvme_sq_copy_cmd()
1591 nvmeq->sqes = qid ? dev->io_sqes : NVME_ADM_SQES; in nvme_alloc_queue()
/linux-6.1.9/include/linux/
Dnvme.h327 __u8 sqes; member