Searched refs:nvme_req (Results 1 – 15 of 15) sorted by relevance
/linux-6.1.9/drivers/nvme/host/ |
D | trace.h | 63 __entry->ctrl_id = nvme_req(req)->ctrl->instance; 99 __entry->ctrl_id = nvme_req(req)->ctrl->instance; 101 __entry->cid = nvme_req(req)->cmd->common.command_id; 102 __entry->result = le64_to_cpu(nvme_req(req)->result.u64); 103 __entry->retries = nvme_req(req)->retries; 104 __entry->flags = nvme_req(req)->flags; 105 __entry->status = nvme_req(req)->status; 155 __entry->ctrl_id = nvme_req(req)->ctrl->instance;
|
D | fault_inject.c | 71 fault_inject = &nvme_req(req)->ctrl->fault_inject; in nvme_should_fail() 79 nvme_req(req)->status = status; in nvme_should_fail()
|
D | ioctl.c | 82 nvme_req(req)->flags |= NVME_REQ_USERCMD; in nvme_alloc_user_request() 166 ctrl = nvme_req(req)->ctrl; in nvme_submit_user_cmd() 171 *result = le64_to_cpu(nvme_req(req)->result.u64); in nvme_submit_user_cmd() 397 if (nvme_req(req)->flags & NVME_REQ_CANCELLED) in nvme_uring_task_meta_cb() 400 status = nvme_req(req)->status; in nvme_uring_task_meta_cb() 402 result = le64_to_cpu(nvme_req(req)->result.u64); in nvme_uring_task_meta_cb() 432 if (nvme_req(req)->flags & NVME_REQ_CANCELLED) in nvme_uring_cmd_end_io() 435 pdu->nvme_status = nvme_req(req)->status; in nvme_uring_cmd_end_io() 436 pdu->u.result = le64_to_cpu(nvme_req(req)->result.u64); in nvme_uring_cmd_end_io()
|
D | nvme.h | 178 static inline struct nvme_request *nvme_req(struct request *req) in nvme_req() function 538 return nvme_cid_install_genctr(nvme_req(rq)->genctr) | rq->tag; in nvme_cid() 554 if (unlikely(nvme_genctr_mask(nvme_req(rq)->genctr) != genctr)) { in nvme_find_rq() 555 dev_err(nvme_req(rq)->ctrl->device, in nvme_find_rq() 557 tag, genctr, nvme_genctr_mask(nvme_req(rq)->genctr)); in nvme_find_rq() 682 struct nvme_request *rq = nvme_req(req); in nvme_try_complete_req()
|
D | core.c | 299 crd = (nvme_req(req)->status & NVME_SC_CRD) >> 11; in nvme_retry_req() 301 delay = nvme_req(req)->ctrl->crdt[crd - 1] * 100; in nvme_retry_req() 303 nvme_req(req)->retries++; in nvme_retry_req() 311 struct nvme_request *nr = nvme_req(req); in nvme_log_error() 348 if (likely(nvme_req(req)->status == 0)) in nvme_decide_disposition() 351 if ((nvme_req(req)->status & 0x7ff) == NVME_SC_AUTH_REQUIRED) in nvme_decide_disposition() 355 (nvme_req(req)->status & NVME_SC_DNR) || in nvme_decide_disposition() 356 nvme_req(req)->retries >= nvme_max_retries) in nvme_decide_disposition() 360 if (nvme_is_path_error(nvme_req(req)->status) || in nvme_decide_disposition() 376 le64_to_cpu(nvme_req(req)->result.u64)); in nvme_end_req_zoned() [all …]
|
D | fabrics.h | 200 nvme_req(rq)->status = NVME_SC_HOST_ABORTED_CMD; in nvmf_complete_timed_out_request()
|
D | rdma.c | 303 nvme_req(rq)->ctrl = &ctrl->ctrl; in nvme_rdma_init_request() 310 req->metadata_sgl = (void *)nvme_req(rq) + in nvme_rdma_init_request() 315 nvme_req(rq)->cmd = req->sqe.data; in nvme_rdma_init_request() 2017 struct nvme_command *c = nvme_req(rq)->cmd; in nvme_rdma_queue_rq() 2106 nvme_req(rq)->status = NVME_SC_INVALID_PI; in nvme_rdma_check_pi_status() 2113 nvme_req(rq)->status = NVME_SC_GUARD_CHECK; in nvme_rdma_check_pi_status() 2116 nvme_req(rq)->status = NVME_SC_REFTAG_CHECK; in nvme_rdma_check_pi_status() 2119 nvme_req(rq)->status = NVME_SC_APPTAG_CHECK; in nvme_rdma_check_pi_status()
|
D | apple.c | 602 !blk_mq_add_to_batch(req, iob, nvme_req(req)->status, in apple_nvme_handle_cqe() 789 struct nvme_request *nreq = nvme_req(req); in apple_nvme_init_request() 890 nvme_req(req)->status = NVME_SC_HOST_ABORTED_CMD; in apple_nvme_timeout() 891 nvme_req(req)->flags |= NVME_REQ_CANCELLED; in apple_nvme_timeout() 916 nvme_req(req)->flags |= NVME_REQ_CANCELLED; in apple_nvme_timeout()
|
D | pci.c | 433 nvme_req(req)->ctrl = &dev->ctrl; in nvme_pci_init_request() 434 nvme_req(req)->cmd = &iod->cmd; in nvme_pci_init_request() 1095 !blk_mq_add_to_batch(req, iob, nvme_req(req)->status, in nvme_handle_cqe() 1277 "Abort status: 0x%x", nvme_req(req)->status); in abort_endio() 1389 nvme_req(req)->flags |= NVME_REQ_CANCELLED; in nvme_timeout() 1407 nvme_req(req)->flags |= NVME_REQ_CANCELLED; in nvme_timeout() 1427 nvme_get_opcode_str(nvme_req(req)->cmd->common.opcode), in nvme_timeout()
|
D | multipath.c | 86 u16 status = nvme_req(req)->status & 0x7ff; in nvme_failover_req()
|
D | tcp.c | 480 nvme_req(rq)->ctrl = &ctrl->ctrl; in nvme_tcp_init_request() 481 nvme_req(rq)->cmd = &pdu->cmd; in nvme_tcp_init_request()
|
D | fc.c | 2149 nvme_req(rq)->ctrl = &ctrl->ctrl; in nvme_fc_init_request() 2150 nvme_req(rq)->cmd = &op->op.cmd_iu.sqe; in nvme_fc_init_request()
|
/linux-6.1.9/drivers/nvme/target/ |
D | passthru.c | 218 struct nvme_ctrl *ctrl = nvme_req(rq)->ctrl; in nvmet_passthru_execute_cmd_work() 240 req->cqe->result = nvme_req(rq)->result; in nvmet_passthru_execute_cmd_work() 253 req->cqe->result = nvme_req(rq)->result; in nvmet_passthru_req_done() 254 nvmet_req_complete(req, nvme_req(rq)->status); in nvmet_passthru_req_done()
|
D | loop.c | 19 struct nvme_request nvme_req; member 64 static void nvme_loop_queue_response(struct nvmet_req *nvme_req); 210 nvme_req(req)->ctrl = &ctrl->ctrl; in nvme_loop_init_request() 211 nvme_req(req)->cmd = &iod->cmd; in nvme_loop_init_request()
|
D | fc.c | 228 nvmet_req_to_fod(struct nvmet_req *nvme_req) in nvmet_req_to_fod() argument 230 return container_of(nvme_req, struct nvmet_fc_fcp_iod, req); in nvmet_req_to_fod() 1911 static void nvmet_fc_fcp_nvme_cmd_done(struct nvmet_req *nvme_req); 2493 nvmet_fc_fcp_nvme_cmd_done(struct nvmet_req *nvme_req) in nvmet_fc_fcp_nvme_cmd_done() argument 2495 struct nvmet_fc_fcp_iod *fod = nvmet_req_to_fod(nvme_req); in nvmet_fc_fcp_nvme_cmd_done()
|