Home
last modified time | relevance | path

Searched refs:nreq (Results 1 – 25 of 27) sorted by relevance

12

/linux-6.1.9/net/ipv4/
Dinet_connection_sock.c868 struct request_sock *nreq; in inet_reqsk_clone() local
870 nreq = kmem_cache_alloc(req->rsk_ops->slab, GFP_ATOMIC | __GFP_NOWARN); in inet_reqsk_clone()
871 if (!nreq) { in inet_reqsk_clone()
880 nreq_sk = req_to_sk(nreq); in inet_reqsk_clone()
894 nreq->rsk_listener = sk; in inet_reqsk_clone()
899 if (sk->sk_protocol == IPPROTO_TCP && tcp_rsk(nreq)->tfo_listener) in inet_reqsk_clone()
900 rcu_assign_pointer(tcp_sk(nreq->sk)->fastopen_rsk, nreq); in inet_reqsk_clone()
902 return nreq; in inet_reqsk_clone()
965 struct request_sock *nreq = NULL, *oreq = req; in reqsk_timer_handler() local
979 nreq = inet_reqsk_clone(req, nsk); in reqsk_timer_handler()
[all …]
/linux-6.1.9/drivers/infiniband/hw/mlx5/
Dwr.c29 int mlx5r_wq_overflow(struct mlx5_ib_wq *wq, int nreq, struct ib_cq *ib_cq) in mlx5r_wq_overflow() argument
35 if (likely(cur + nreq < wq->max_post)) in mlx5r_wq_overflow()
43 return cur + nreq >= wq->max_post; in mlx5r_wq_overflow()
726 int *size, void **cur_edge, int nreq, __be32 general_id, in mlx5r_begin_wqe() argument
729 if (unlikely(mlx5r_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq))) in mlx5r_begin_wqe()
751 void **cur_edge, int nreq) in begin_wqe() argument
753 return mlx5r_begin_wqe(qp, seg, ctrl, idx, size, cur_edge, nreq, in begin_wqe()
760 u64 wr_id, int nreq, u8 fence, u32 mlx5_opcode) in mlx5r_finish_wqe() argument
773 qp->sq.wqe_head[idx] = qp->sq.head + nreq; in mlx5r_finish_wqe()
815 void **cur_edge, unsigned int *idx, int nreq, in handle_psv() argument
[all …]
Dwr.h94 int mlx5r_wq_overflow(struct mlx5_ib_wq *wq, int nreq, struct ib_cq *ib_cq);
97 int *size, void **cur_edge, int nreq, __be32 general_id,
101 u64 wr_id, int nreq, u8 fence, u32 mlx5_opcode);
102 void mlx5r_ring_db(struct mlx5_ib_qp *qp, unsigned int nreq,
Dsrq.c412 int nreq; in mlx5_ib_post_srq_recv() local
423 for (nreq = 0; wr; nreq++, wr = wr->next) { in mlx5_ib_post_srq_recv()
455 if (likely(nreq)) { in mlx5_ib_post_srq_recv()
456 srq->wqe_ctr += nreq; in mlx5_ib_post_srq_recv()
/linux-6.1.9/crypto/
Dechainiv.c45 SYNC_SKCIPHER_REQUEST_ON_STACK(nreq, ctx->sknull); in echainiv_encrypt()
47 skcipher_request_set_sync_tfm(nreq, ctx->sknull); in echainiv_encrypt()
48 skcipher_request_set_callback(nreq, req->base.flags, in echainiv_encrypt()
50 skcipher_request_set_crypt(nreq, req->src, req->dst, in echainiv_encrypt()
54 err = crypto_skcipher_encrypt(nreq); in echainiv_encrypt()
Dseqiv.c69 SYNC_SKCIPHER_REQUEST_ON_STACK(nreq, ctx->sknull); in seqiv_aead_encrypt()
71 skcipher_request_set_sync_tfm(nreq, ctx->sknull); in seqiv_aead_encrypt()
72 skcipher_request_set_callback(nreq, req->base.flags, in seqiv_aead_encrypt()
74 skcipher_request_set_crypt(nreq, req->src, req->dst, in seqiv_aead_encrypt()
78 err = crypto_skcipher_encrypt(nreq); in seqiv_aead_encrypt()
Dgcm.c966 SYNC_SKCIPHER_REQUEST_ON_STACK(nreq, ctx->null); in crypto_rfc4543_copy_src_to_dst()
968 skcipher_request_set_sync_tfm(nreq, ctx->null); in crypto_rfc4543_copy_src_to_dst()
969 skcipher_request_set_callback(nreq, req->base.flags, NULL, NULL); in crypto_rfc4543_copy_src_to_dst()
970 skcipher_request_set_crypt(nreq, req->src, req->dst, nbytes, NULL); in crypto_rfc4543_copy_src_to_dst()
972 return crypto_skcipher_encrypt(nreq); in crypto_rfc4543_copy_src_to_dst()
/linux-6.1.9/drivers/infiniband/hw/mthca/
Dmthca_qp.c1568 static inline int mthca_wq_overflow(struct mthca_wq *wq, int nreq, in mthca_wq_overflow() argument
1575 if (likely(cur + nreq < wq->max)) in mthca_wq_overflow()
1583 return cur + nreq >= wq->max; in mthca_wq_overflow()
1634 int nreq; in mthca_tavor_post_send() local
1655 for (nreq = 0; wr; ++nreq, wr = wr->next) { in mthca_tavor_post_send()
1656 if (mthca_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) { in mthca_tavor_post_send()
1660 qp->sq.max, nreq); in mthca_tavor_post_send()
1789 cpu_to_be32((nreq ? 0 : MTHCA_NEXT_DBD) | size | in mthca_tavor_post_send()
1793 if (!nreq) { in mthca_tavor_post_send()
1806 if (likely(nreq)) { in mthca_tavor_post_send()
[all …]
Dmthca_srq.c493 int nreq; in mthca_tavor_post_srq_recv() local
502 for (nreq = 0; wr; wr = wr->next) { in mthca_tavor_post_srq_recv()
543 ++nreq; in mthca_tavor_post_srq_recv()
544 if (unlikely(nreq == MTHCA_TAVOR_MAX_WQES_PER_RECV_DB)) { in mthca_tavor_post_srq_recv()
545 nreq = 0; in mthca_tavor_post_srq_recv()
561 if (likely(nreq)) { in mthca_tavor_post_srq_recv()
568 mthca_write64(first_ind << srq->wqe_shift, (srq->srqn << 8) | nreq, in mthca_tavor_post_srq_recv()
586 int nreq; in mthca_arbel_post_srq_recv() local
592 for (nreq = 0; wr; ++nreq, wr = wr->next) { in mthca_arbel_post_srq_recv()
627 if (likely(nreq)) { in mthca_arbel_post_srq_recv()
[all …]
/linux-6.1.9/drivers/usb/gadget/function/
Duvc_queue.c48 unsigned int nreq; in uvc_queue_setup() local
64 nreq = DIV_ROUND_UP(DIV_ROUND_UP(sizes[0], 2), req_size); in uvc_queue_setup()
65 nreq = clamp(nreq, 4U, 64U); in uvc_queue_setup()
66 video->uvc_num_requests = nreq; in uvc_queue_setup()
/linux-6.1.9/drivers/infiniband/hw/mlx4/
Dsrq.c315 int nreq; in mlx4_ib_post_srq_recv() local
326 for (nreq = 0; wr; ++nreq, wr = wr->next) { in mlx4_ib_post_srq_recv()
358 if (likely(nreq)) { in mlx4_ib_post_srq_recv()
359 srq->wqe_ctr += nreq; in mlx4_ib_post_srq_recv()
Dqp.c3239 static int mlx4_wq_overflow(struct mlx4_ib_wq *wq, int nreq, struct ib_cq *ib_cq) in mlx4_wq_overflow() argument
3245 if (likely(cur + nreq < wq->max_post)) in mlx4_wq_overflow()
3253 return cur + nreq >= wq->max_post; in mlx4_wq_overflow()
3491 int nreq; in _mlx4_ib_post_send() local
3527 nreq = 0; in _mlx4_ib_post_send()
3533 for (nreq = 0; wr; ++nreq, wr = wr->next) { in _mlx4_ib_post_send()
3537 if (mlx4_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) { in _mlx4_ib_post_send()
3550 qp->sq.wrid[(qp->sq.head + nreq) & (qp->sq.wqe_cnt - 1)] = wr->wr_id; in _mlx4_ib_post_send()
3776 if (likely(nreq)) { in _mlx4_ib_post_send()
3777 qp->sq.head += nreq; in _mlx4_ib_post_send()
[all …]
/linux-6.1.9/fs/nfs/
Dpnfs_nfs.c459 unsigned int nreq = 0; in pnfs_bucket_alloc_ds_commits() local
471 nreq++; in pnfs_bucket_alloc_ds_commits()
475 return nreq; in pnfs_bucket_alloc_ds_commits()
480 return nreq; in pnfs_bucket_alloc_ds_commits()
515 unsigned int nreq = 0; in pnfs_generic_commit_pagelist() local
526 nreq++; in pnfs_generic_commit_pagelist()
529 nreq += pnfs_alloc_ds_commits_list(&list, fl_cinfo, cinfo); in pnfs_generic_commit_pagelist()
530 if (nreq == 0) in pnfs_generic_commit_pagelist()
/linux-6.1.9/drivers/crypto/inside-secure/
Dsafexcel.c816 int ret, nreq = 0, cdesc = 0, rdesc = 0, commands, results; in safexcel_dequeue() local
856 nreq++; in safexcel_dequeue()
867 if (!nreq) in safexcel_dequeue()
872 priv->ring[ring].requests += nreq; in safexcel_dequeue()
1019 int ret, i, nreq, ndesc, tot_descs, handled = 0; in safexcel_handle_result_descriptor() local
1025 nreq = readl(EIP197_HIA_RDR(priv, ring) + EIP197_HIA_xDR_PROC_COUNT); in safexcel_handle_result_descriptor()
1026 nreq >>= EIP197_xDR_PROC_xD_PKT_OFFSET; in safexcel_handle_result_descriptor()
1027 nreq &= EIP197_xDR_PROC_xD_PKT_MASK; in safexcel_handle_result_descriptor()
1028 if (!nreq) in safexcel_handle_result_descriptor()
1031 for (i = 0; i < nreq; i++) { in safexcel_handle_result_descriptor()
[all …]
/linux-6.1.9/drivers/dma/
Dbcm-sba-raid.c287 struct sba_request *nreq; in sba_free_chained_requests() local
293 list_for_each_entry(nreq, &req->next, next) in sba_free_chained_requests()
294 _sba_free_request(sba, nreq); in sba_free_chained_requests()
410 struct sba_request *nreq, *first = req->first; in sba_process_received_request() local
432 list_for_each_entry(nreq, &first->next, next) in sba_process_received_request()
433 _sba_free_request(sba, nreq); in sba_process_received_request()
518 struct sba_request *req, *nreq; in sba_tx_submit() local
530 list_for_each_entry(nreq, &req->next, next) in sba_tx_submit()
531 _sba_pending_request(sba, nreq); in sba_tx_submit()
/linux-6.1.9/fs/nilfs2/
Dbtree.c1736 union nilfs_bmap_ptr_req *nreq, in nilfs_btree_prepare_convert_and_insert() argument
1763 if (nreq != NULL) { in nilfs_btree_prepare_convert_and_insert()
1764 nreq->bpr_ptr = dreq->bpr_ptr + 1; in nilfs_btree_prepare_convert_and_insert()
1765 ret = nilfs_bmap_prepare_alloc_ptr(btree, nreq, dat); in nilfs_btree_prepare_convert_and_insert()
1769 ret = nilfs_btree_get_new_block(btree, nreq->bpr_ptr, &bh); in nilfs_btree_prepare_convert_and_insert()
1782 nilfs_bmap_abort_alloc_ptr(btree, nreq, dat); in nilfs_btree_prepare_convert_and_insert()
1796 union nilfs_bmap_ptr_req *nreq, in nilfs_btree_commit_convert_and_insert() argument
1814 if (nreq != NULL) { in nilfs_btree_commit_convert_and_insert()
1816 nilfs_bmap_commit_alloc_ptr(btree, nreq, dat); in nilfs_btree_commit_convert_and_insert()
1832 tmpptr = nreq->bpr_ptr; in nilfs_btree_commit_convert_and_insert()
[all …]
/linux-6.1.9/drivers/crypto/qat/qat_common/
Dqat_algs.c1250 struct skcipher_request *nreq = skcipher_request_ctx(req); in qat_alg_skcipher_xts_encrypt() local
1256 memcpy(nreq, req, sizeof(*req)); in qat_alg_skcipher_xts_encrypt()
1257 skcipher_request_set_tfm(nreq, ctx->ftfm); in qat_alg_skcipher_xts_encrypt()
1258 return crypto_skcipher_encrypt(nreq); in qat_alg_skcipher_xts_encrypt()
1317 struct skcipher_request *nreq = skcipher_request_ctx(req); in qat_alg_skcipher_xts_decrypt() local
1323 memcpy(nreq, req, sizeof(*req)); in qat_alg_skcipher_xts_decrypt()
1324 skcipher_request_set_tfm(nreq, ctx->ftfm); in qat_alg_skcipher_xts_decrypt()
1325 return crypto_skcipher_decrypt(nreq); in qat_alg_skcipher_xts_decrypt()
/linux-6.1.9/drivers/infiniband/hw/hns/
Dhns_roce_hw_v2.c707 u32 nreq; in hns_roce_v2_post_send() local
715 nreq = 0; in hns_roce_v2_post_send()
721 for (nreq = 0; wr; ++nreq, wr = wr->next) { in hns_roce_v2_post_send()
722 if (hns_roce_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) { in hns_roce_v2_post_send()
728 wqe_idx = (qp->sq.head + nreq) & (qp->sq.wqe_cnt - 1); in hns_roce_v2_post_send()
741 ~(((qp->sq.head + nreq) >> ilog2(qp->sq.wqe_cnt)) & 0x1); in hns_roce_v2_post_send()
756 if (likely(nreq)) { in hns_roce_v2_post_send()
757 qp->sq.head += nreq; in hns_roce_v2_post_send()
760 if (nreq == 1 && (qp->en_flags & HNS_ROCE_QP_CAP_DIRECT_WQE)) in hns_roce_v2_post_send()
849 u32 wqe_idx, nreq, max_sge; in hns_roce_v2_post_recv() local
[all …]
Dhns_roce_qp.c1490 bool hns_roce_wq_overflow(struct hns_roce_wq *hr_wq, u32 nreq, in hns_roce_wq_overflow() argument
1497 if (likely(cur + nreq < hr_wq->wqe_cnt)) in hns_roce_wq_overflow()
1505 return cur + nreq >= hr_wq->wqe_cnt; in hns_roce_wq_overflow()
Dhns_roce_device.h1195 bool hns_roce_wq_overflow(struct hns_roce_wq *hr_wq, u32 nreq,
/linux-6.1.9/drivers/net/ethernet/mellanox/mlx4/
Dmain.c2943 int nreq = min3(dev->caps.num_ports * in mlx4_enable_msi_x() local
2949 nreq = min_t(int, nreq, msi_x); in mlx4_enable_msi_x()
2951 entries = kcalloc(nreq, sizeof(*entries), GFP_KERNEL); in mlx4_enable_msi_x()
2955 for (i = 0; i < nreq; ++i) in mlx4_enable_msi_x()
2958 nreq = pci_enable_msix_range(dev->persist->pdev, entries, 2, in mlx4_enable_msi_x()
2959 nreq); in mlx4_enable_msi_x()
2961 if (nreq < 0 || nreq < MLX4_EQ_ASYNC) { in mlx4_enable_msi_x()
2966 dev->caps.num_comp_vectors = nreq - 1; in mlx4_enable_msi_x()
/linux-6.1.9/drivers/nvme/host/
Dapple.c789 struct nvme_request *nreq = nvme_req(req); in apple_nvme_init_request() local
792 nreq->ctrl = &anv->ctrl; in apple_nvme_init_request()
793 nreq->cmd = &iod->cmd; in apple_nvme_init_request()
/linux-6.1.9/drivers/usb/isp1760/
Disp1760-udc.c841 struct isp1760_request *req, *nreq; in isp1760_ep_disable() local
868 list_for_each_entry_safe(req, nreq, &req_list, queue) { in isp1760_ep_disable()
/linux-6.1.9/drivers/infiniband/sw/rdmavt/
Dqp.c2152 unsigned nreq = 0; in rvt_post_send() local
2179 nreq++; in rvt_post_send()
2183 if (nreq) { in rvt_post_send()
2188 if (nreq == 1 && call_send) in rvt_post_send()
/linux-6.1.9/drivers/infiniband/hw/hfi1/
Dtid_rdma.c5198 struct tid_rdma_request *req, *nreq; in make_tid_rdma_ack() local
5272 nreq = ack_to_tid_req(&qp->s_ack_queue[next]); in make_tid_rdma_ack()
5273 if (!nreq->comp_seg || nreq->ack_seg == nreq->comp_seg) in make_tid_rdma_ack()

12