Home
last modified time | relevance | path

Searched refs:queue_depth (Results 1 – 25 of 111) sorted by relevance

12345

/linux-6.6.21/drivers/target/
Dtarget_core_tpg.c158 struct se_node_acl *acl, u32 queue_depth) in target_set_nacl_queue_depth() argument
160 acl->queue_depth = queue_depth; in target_set_nacl_queue_depth()
162 if (!acl->queue_depth) { in target_set_nacl_queue_depth()
166 acl->queue_depth = 1; in target_set_nacl_queue_depth()
174 u32 queue_depth; in target_alloc_node_acl() local
191 queue_depth = tpg->se_tpg_tfo->tpg_get_default_depth(tpg); in target_alloc_node_acl()
193 queue_depth = 1; in target_alloc_node_acl()
194 target_set_nacl_queue_depth(tpg, acl, queue_depth); in target_alloc_node_acl()
218 acl->queue_depth, in target_add_node_acl()
369 tpg->se_tpg_tfo->tpg_get_tag(tpg), acl->queue_depth, in core_tpg_del_initiator_node_acl()
[all …]
/linux-6.6.21/arch/um/drivers/
Dvector_kern.c250 int queue_depth; in vector_advancehead() local
258 qi->queue_depth -= advance; in vector_advancehead()
264 if (qi->queue_depth == 0) { in vector_advancehead()
268 queue_depth = qi->queue_depth; in vector_advancehead()
270 return queue_depth; in vector_advancehead()
280 int queue_depth; in vector_advancetail() local
286 qi->queue_depth += advance; in vector_advancetail()
287 queue_depth = qi->queue_depth; in vector_advancetail()
289 return queue_depth; in vector_advancetail()
336 int queue_depth; in vector_enqueue() local
[all …]
Dvector_kern.h47 int queue_depth, head, tail, max_depth, max_iov_frags; member
/linux-6.6.21/drivers/s390/block/
Ddasd_genhd.c28 static unsigned int queue_depth = 32; variable
31 module_param(queue_depth, uint, 0444);
32 MODULE_PARM_DESC(queue_depth, "Default queue depth for new DASD devices");
54 block->tag_set.queue_depth = queue_depth; in dasd_gendisk_alloc()
/linux-6.6.21/drivers/target/iscsi/
Discsi_target_device.c41 sess->cmdsn_window = se_nacl->queue_depth; in iscsit_determine_maxcmdsn()
42 atomic_add(se_nacl->queue_depth - 1, &sess->max_cmd_sn); in iscsit_determine_maxcmdsn()
/linux-6.6.21/drivers/infiniband/ulp/rtrs/
Drtrs-clt.c71 size_t max_depth = clt->queue_depth; in __rtrs_get_permit()
499 if (WARN_ON(msg_id >= clt_path->queue_depth)) in process_io_rsp()
553 if (WARN_ON(buf_id >= clt_path->queue_depth)) in rtrs_clt_rkey_rsp_done()
725 q_size = clt_path->queue_depth; in post_recv_path()
1323 for (i = 0; i < clt_path->queue_depth; ++i) { in fail_all_outstanding_reqs()
1349 for (i = 0; i < clt_path->queue_depth; ++i) { in free_path_reqs()
1365 clt_path->reqs = kcalloc(clt_path->queue_depth, in alloc_path_reqs()
1371 for (i = 0; i < clt_path->queue_depth; ++i) { in alloc_path_reqs()
1411 clt->permits_map = bitmap_zalloc(clt->queue_depth, GFP_KERNEL); in alloc_permits()
1416 clt->permits = kcalloc(clt->queue_depth, permit_size(clt), GFP_KERNEL); in alloc_permits()
[all …]
Drtrs-srv.c109 for (i = 0; i < srv->queue_depth; i++) in rtrs_srv_free_ops_ids()
138 srv_path->ops_ids = kcalloc(srv->queue_depth, in rtrs_srv_alloc_ops_ids()
144 for (i = 0; i < srv->queue_depth; ++i) { in rtrs_srv_alloc_ops_ids()
585 mrs_num = srv->queue_depth; in map_cont_bufs()
589 mrs_num = DIV_ROUND_UP(srv->queue_depth, chunks_per_mr); in map_cont_bufs()
590 chunks_per_mr = DIV_ROUND_UP(srv->queue_depth, mrs_num); in map_cont_bufs()
607 srv->queue_depth - chunks); in map_cont_bufs()
655 chunk_bits = ilog2(srv->queue_depth - 1) + 1; in map_cont_bufs()
989 q_size = srv->queue_depth; in post_recv_path()
1246 if (msg_id >= srv->queue_depth || off >= max_chunk_size) { in rtrs_srv_rdma_done()
[all …]
Drtrs-clt.h144 size_t queue_depth; member
173 size_t queue_depth; member
Drtrs.h111 u32 queue_depth; member
Drtrs-srv.h110 size_t queue_depth; member
/linux-6.6.21/drivers/ata/
Dlibata-sata.c1056 int queue_depth) in ata_change_queue_depth() argument
1065 if (!dev || queue_depth < 1 || queue_depth == sdev->queue_depth) { in ata_change_queue_depth()
1067 return sdev->queue_depth; in ata_change_queue_depth()
1076 if (queue_depth > max_queue_depth) { in ata_change_queue_depth()
1085 if (queue_depth == 1 || !ata_ncq_supported(dev)) { in ata_change_queue_depth()
1087 queue_depth = 1; in ata_change_queue_depth()
1094 if (queue_depth == sdev->queue_depth) in ata_change_queue_depth()
1095 return sdev->queue_depth; in ata_change_queue_depth()
1097 return scsi_change_queue_depth(sdev, queue_depth); in ata_change_queue_depth()
1116 int ata_scsi_change_queue_depth(struct scsi_device *sdev, int queue_depth) in ata_scsi_change_queue_depth() argument
[all …]
/linux-6.6.21/drivers/s390/crypto/
Dap_card.c47 return sysfs_emit(buf, "%d\n", ac->queue_depth); in depth_show()
232 struct ap_card *ap_card_create(int id, int queue_depth, int raw_type, in ap_card_create() argument
244 ac->queue_depth = queue_depth; in ap_card_create()
Dap_bus.h187 int queue_depth; /* AP queue depth.*/ member
292 struct ap_card *ap_card_create(int id, int queue_depth, int raw_type,
/linux-6.6.21/include/net/mana/
Dhw_channel.h139 u16 queue_depth; member
153 u16 queue_depth; member
/linux-6.6.21/block/
Dblk-rq-qos.c122 if (rqd->queue_depth == 1) { in rq_depth_calc_max_depth()
138 rqd->queue_depth); in rq_depth_calc_max_depth()
142 unsigned int maxd = 3 * rqd->queue_depth / 4; in rq_depth_calc_max_depth()
Dblk-mq.c3621 unsigned int queue_depth, struct request *flush_rq) in blk_mq_clear_flush_rq_mapping() argument
3632 for (i = 0; i < queue_depth; i++) in blk_mq_clear_flush_rq_mapping()
3657 set->queue_depth, flush_rq); in blk_mq_exit_hctx()
3848 set->queue_depth); in __blk_mq_alloc_map_and_rqs()
4310 q->nr_requests = set->queue_depth; in blk_mq_init_allocated_queue()
4343 set->queue_depth); in __blk_mq_alloc_rq_maps()
4378 depth = set->queue_depth; in blk_mq_alloc_set_map_and_rqs()
4384 set->queue_depth >>= 1; in blk_mq_alloc_set_map_and_rqs()
4385 if (set->queue_depth < set->reserved_tags + BLK_MQ_TAG_MIN) { in blk_mq_alloc_set_map_and_rqs()
4389 } while (set->queue_depth); in blk_mq_alloc_set_map_and_rqs()
[all …]
/linux-6.6.21/drivers/block/null_blk/
Dnull_blk.h33 unsigned int queue_depth; member
133 unsigned int queue_depth; member
Dmain.c768 tag = find_first_zero_bit(nq->tag_map, nq->queue_depth); in get_tag()
769 if (tag >= nq->queue_depth) in get_tag()
1780 nq->queue_depth = nullb->queue_depth; in null_init_queue()
1882 nq->cmds = kcalloc(nq->queue_depth, sizeof(*cmd), GFP_KERNEL); in setup_commands()
1886 nq->tag_map = bitmap_zalloc(nq->queue_depth, GFP_KERNEL); in setup_commands()
1892 for (i = 0; i < nq->queue_depth; i++) { in setup_commands()
1912 nullb->queue_depth = nullb->dev->hw_queue_depth; in setup_queues()
1965 unsigned int queue_depth; in null_init_tag_set() local
1971 queue_depth = nullb->dev->hw_queue_depth; in null_init_tag_set()
1982 queue_depth = g_hw_queue_depth; in null_init_tag_set()
[all …]
/linux-6.6.21/drivers/block/rnbd/
Drnbd-srv.h25 int queue_depth; member
Drnbd-clt.h86 size_t queue_depth; member
/linux-6.6.21/drivers/scsi/snic/
Dsnic_main.c91 if (qsz < sdev->queue_depth) in snic_change_queue_depth()
93 else if (qsz > sdev->queue_depth) in snic_change_queue_depth()
96 atomic64_set(&snic->s_stats.misc.last_qsz, sdev->queue_depth); in snic_change_queue_depth()
100 return sdev->queue_depth; in snic_change_queue_depth()
/linux-6.6.21/drivers/block/
Dvirtio_blk.c1302 module_param_named(queue_depth, virtblk_queue_depth, uint, 0444);
1315 unsigned int queue_depth; in virtblk_probe() local
1360 queue_depth = vblk->vqs[0].vq->num_free; in virtblk_probe()
1363 queue_depth /= 2; in virtblk_probe()
1365 queue_depth = virtblk_queue_depth; in virtblk_probe()
1370 vblk->tag_set.queue_depth = queue_depth; in virtblk_probe()
/linux-6.6.21/include/target/
Dtarget_core_base.h570 u32 queue_depth; member
714 u32 queue_depth; member
805 u32 queue_depth; member
/linux-6.6.21/drivers/scsi/
Dscsi.c222 sdev->queue_depth = depth; in scsi_change_queue_depth()
229 sbitmap_resize(&sdev->budget_map, sdev->queue_depth); in scsi_change_queue_depth()
231 return sdev->queue_depth; in scsi_change_queue_depth()
/linux-6.6.21/drivers/net/ethernet/microsoft/mana/
Dhw_channel.c317 comp_read = mana_gd_poll_cq(q_self, completions, hwc_cq->queue_depth); in mana_hwc_comp_event()
318 WARN_ON_ONCE(comp_read <= 0 || comp_read > hwc_cq->queue_depth); in mana_hwc_comp_event()
397 hwc_cq->queue_depth = q_depth; in mana_hwc_create_cq()
511 hwc_wq->queue_depth = q_depth; in mana_hwc_create_wq()

12345