Lines Matching refs:ep_ctx
552 struct xhci_ep_ctx *ep_ctx; in xhci_get_hw_deq() local
562 ep_ctx = xhci_get_ep_ctx(xhci, vdev->out_ctx, ep_index); in xhci_get_hw_deq()
563 return le64_to_cpu(ep_ctx->deq); in xhci_get_hw_deq()
1064 struct xhci_ep_ctx *ep_ctx; in xhci_handle_cmd_stop_ep() local
1082 ep_ctx = xhci_get_ep_ctx(xhci, ep->vdev->out_ctx, ep_index); in xhci_handle_cmd_stop_ep()
1084 trace_xhci_handle_cmd_stop_ep(ep_ctx); in xhci_handle_cmd_stop_ep()
1101 switch (GET_EP_CTX_STATE(ep_ctx)) { in xhci_handle_cmd_stop_ep()
1306 struct xhci_ep_ctx *ep_ctx; in xhci_handle_cmd_set_deq() local
1324 ep_ctx = xhci_get_ep_ctx(xhci, ep->vdev->out_ctx, ep_index); in xhci_handle_cmd_set_deq()
1327 trace_xhci_handle_cmd_set_deq_ep(ep_ctx); in xhci_handle_cmd_set_deq()
1339 ep_state = GET_EP_CTX_STATE(ep_ctx); in xhci_handle_cmd_set_deq()
1369 deq = le64_to_cpu(ep_ctx->deq) & ~EP_CTX_CYCLE_MASK; in xhci_handle_cmd_set_deq()
1412 struct xhci_ep_ctx *ep_ctx; in xhci_handle_cmd_reset_ep() local
1420 ep_ctx = xhci_get_ep_ctx(xhci, ep->vdev->out_ctx, ep_index); in xhci_handle_cmd_reset_ep()
1421 trace_xhci_handle_cmd_reset_ep(ep_ctx); in xhci_handle_cmd_reset_ep()
1473 struct xhci_ep_ctx *ep_ctx; in xhci_handle_cmd_config_ep() local
1496 ep_ctx = xhci_get_ep_ctx(xhci, virt_dev->out_ctx, ep_index); in xhci_handle_cmd_config_ep()
1497 trace_xhci_handle_cmd_config_ep(ep_ctx); in xhci_handle_cmd_config_ep()
2106 struct xhci_ep_ctx *ep_ctx, in xhci_requires_manual_halt_cleanup() argument
2119 if (GET_EP_CTX_STATE(ep_ctx) == EP_STATE_HALTED) in xhci_requires_manual_halt_cleanup()
2143 struct xhci_ep_ctx *ep_ctx; in finish_td() local
2145 ep_ctx = xhci_get_ep_ctx(xhci, ep->vdev->out_ctx, ep->ep_index); in finish_td()
2171 if (GET_EP_CTX_STATE(ep_ctx) != EP_STATE_HALTED) { in finish_td()
2245 struct xhci_ep_ctx *ep_ctx; in process_ctrl_td() local
2251 ep_ctx = xhci_get_ep_ctx(xhci, ep->vdev->out_ctx, ep->ep_index); in process_ctrl_td()
2296 ep_ctx, trb_comp_code)) in process_ctrl_td()
2537 struct xhci_ep_ctx *ep_ctx; in handle_tx_event() local
2555 ep_ctx = xhci_get_ep_ctx(xhci, ep->vdev->out_ctx, ep_index); in handle_tx_event()
2557 if (GET_EP_CTX_STATE(ep_ctx) == EP_STATE_DISABLED) { in handle_tx_event()
2759 xhci_requires_manual_halt_cleanup(xhci, ep_ctx, in handle_tx_event()
2854 xhci_requires_manual_halt_cleanup(xhci, ep_ctx, in handle_tx_event()
3247 struct xhci_ep_ctx *ep_ctx = xhci_get_ep_ctx(xhci, xdev->out_ctx, ep_index); in prepare_transfer() local
3257 ret = prepare_ring(xhci, ep_ring, GET_EP_CTX_STATE(ep_ctx), in prepare_transfer()
3358 struct xhci_ep_ctx *ep_ctx) in check_interval() argument
3363 xhci_interval = EP_INTERVAL_TO_UFRAMES(le32_to_cpu(ep_ctx->ep_info)); in check_interval()
3396 struct xhci_ep_ctx *ep_ctx; in xhci_queue_intr_tx() local
3398 ep_ctx = xhci_get_ep_ctx(xhci, xhci->devs[slot_id]->out_ctx, ep_index); in xhci_queue_intr_tx()
3399 check_interval(xhci, urb, ep_ctx); in xhci_queue_intr_tx()
4177 struct xhci_ep_ctx *ep_ctx; in xhci_queue_isoc_tx_prepare() local
4187 ep_ctx = xhci_get_ep_ctx(xhci, xdev->out_ctx, ep_index); in xhci_queue_isoc_tx_prepare()
4197 ret = prepare_ring(xhci, ep_ring, GET_EP_CTX_STATE(ep_ctx), in xhci_queue_isoc_tx_prepare()
4206 check_interval(xhci, urb, ep_ctx); in xhci_queue_isoc_tx_prepare()
4210 if (GET_EP_CTX_STATE(ep_ctx) == EP_STATE_RUNNING) { in xhci_queue_isoc_tx_prepare()