/linux-6.1.9/drivers/target/ |
D | target_core_tpg.c | 157 struct se_node_acl *acl, u32 queue_depth) in target_set_nacl_queue_depth() argument 159 acl->queue_depth = queue_depth; in target_set_nacl_queue_depth() 161 if (!acl->queue_depth) { in target_set_nacl_queue_depth() 165 acl->queue_depth = 1; in target_set_nacl_queue_depth() 173 u32 queue_depth; in target_alloc_node_acl() local 190 queue_depth = tpg->se_tpg_tfo->tpg_get_default_depth(tpg); in target_alloc_node_acl() 192 queue_depth = 1; in target_alloc_node_acl() 193 target_set_nacl_queue_depth(tpg, acl, queue_depth); in target_alloc_node_acl() 217 acl->queue_depth, in target_add_node_acl() 368 tpg->se_tpg_tfo->tpg_get_tag(tpg), acl->queue_depth, in core_tpg_del_initiator_node_acl() [all …]
|
/linux-6.1.9/arch/um/drivers/ |
D | vector_kern.c | 250 int queue_depth; in vector_advancehead() local 258 qi->queue_depth -= advance; in vector_advancehead() 264 if (qi->queue_depth == 0) { in vector_advancehead() 268 queue_depth = qi->queue_depth; in vector_advancehead() 270 return queue_depth; in vector_advancehead() 280 int queue_depth; in vector_advancetail() local 286 qi->queue_depth += advance; in vector_advancetail() 287 queue_depth = qi->queue_depth; in vector_advancetail() 289 return queue_depth; in vector_advancetail() 336 int queue_depth; in vector_enqueue() local [all …]
|
D | vector_kern.h | 47 int queue_depth, head, tail, max_depth, max_iov_frags; member
|
/linux-6.1.9/drivers/ata/ |
D | libata-sata.c | 1032 struct scsi_device *sdev, int queue_depth) in ata_change_queue_depth() argument 1037 return sdev->queue_depth; in ata_change_queue_depth() 1039 if (queue_depth < 1 || queue_depth == sdev->queue_depth) in ata_change_queue_depth() 1040 return sdev->queue_depth; in ata_change_queue_depth() 1045 if (queue_depth == 1 || !ata_ncq_enabled(dev)) { in ata_change_queue_depth() 1047 queue_depth = 1; in ata_change_queue_depth() 1052 queue_depth = min(queue_depth, sdev->host->can_queue); in ata_change_queue_depth() 1053 queue_depth = min(queue_depth, ata_id_queue_depth(dev->id)); in ata_change_queue_depth() 1054 queue_depth = min(queue_depth, ATA_MAX_QUEUE); in ata_change_queue_depth() 1056 if (sdev->queue_depth == queue_depth) in ata_change_queue_depth() [all …]
|
/linux-6.1.9/drivers/s390/block/ |
D | dasd_genhd.c | 28 static unsigned int queue_depth = 32; variable 31 module_param(queue_depth, uint, 0444); 32 MODULE_PARM_DESC(queue_depth, "Default queue depth for new DASD devices"); 54 block->tag_set.queue_depth = queue_depth; in dasd_gendisk_alloc()
|
/linux-6.1.9/drivers/target/iscsi/ |
D | iscsi_target_device.c | 41 sess->cmdsn_window = se_nacl->queue_depth; in iscsit_determine_maxcmdsn() 42 atomic_add(se_nacl->queue_depth - 1, &sess->max_cmd_sn); in iscsit_determine_maxcmdsn()
|
/linux-6.1.9/drivers/infiniband/ulp/rtrs/ |
D | rtrs-clt.c | 69 size_t max_depth = clt->queue_depth; in __rtrs_get_permit() 497 if (WARN_ON(msg_id >= clt_path->queue_depth)) in process_io_rsp() 551 if (WARN_ON(buf_id >= clt_path->queue_depth)) in rtrs_clt_rkey_rsp_done() 723 q_size = clt_path->queue_depth; in post_recv_path() 1323 for (i = 0; i < clt_path->queue_depth; ++i) { in fail_all_outstanding_reqs() 1349 for (i = 0; i < clt_path->queue_depth; ++i) { in free_path_reqs() 1365 clt_path->reqs = kcalloc(clt_path->queue_depth, in alloc_path_reqs() 1371 for (i = 0; i < clt_path->queue_depth; ++i) { in alloc_path_reqs() 1411 clt->permits_map = bitmap_zalloc(clt->queue_depth, GFP_KERNEL); in alloc_permits() 1416 clt->permits = kcalloc(clt->queue_depth, permit_size(clt), GFP_KERNEL); in alloc_permits() [all …]
|
D | rtrs-srv.c | 106 for (i = 0; i < srv->queue_depth; i++) in rtrs_srv_free_ops_ids() 135 srv_path->ops_ids = kcalloc(srv->queue_depth, in rtrs_srv_alloc_ops_ids() 141 for (i = 0; i < srv->queue_depth; ++i) { in rtrs_srv_alloc_ops_ids() 577 mrs_num = srv->queue_depth; in map_cont_bufs() 581 mrs_num = DIV_ROUND_UP(srv->queue_depth, chunks_per_mr); in map_cont_bufs() 582 chunks_per_mr = DIV_ROUND_UP(srv->queue_depth, mrs_num); in map_cont_bufs() 601 srv->queue_depth - chunks); in map_cont_bufs() 668 chunk_bits = ilog2(srv->queue_depth - 1) + 1; in map_cont_bufs() 978 q_size = srv->queue_depth; in post_recv_path() 1235 if (msg_id >= srv->queue_depth || off >= max_chunk_size) { in rtrs_srv_rdma_done() [all …]
|
D | rtrs-clt.h | 144 size_t queue_depth; member 173 size_t queue_depth; member
|
D | rtrs.h | 111 u32 queue_depth; member
|
D | rtrs-srv.h | 110 size_t queue_depth; member
|
/linux-6.1.9/drivers/s390/crypto/ |
D | ap_card.c | 47 return scnprintf(buf, PAGE_SIZE, "%d\n", ac->queue_depth); in depth_show() 233 struct ap_card *ap_card_create(int id, int queue_depth, int raw_type, in ap_card_create() argument 245 ac->queue_depth = queue_depth; in ap_card_create()
|
D | ap_bus.h | 182 int queue_depth; /* AP queue depth.*/ member 312 struct ap_card *ap_card_create(int id, int queue_depth, int raw_type,
|
/linux-6.1.9/block/ |
D | blk-rq-qos.c | 122 if (rqd->queue_depth == 1) { in rq_depth_calc_max_depth() 138 rqd->queue_depth); in rq_depth_calc_max_depth() 142 unsigned int maxd = 3 * rqd->queue_depth / 4; in rq_depth_calc_max_depth()
|
D | blk-mq.c | 3526 unsigned int queue_depth, struct request *flush_rq) in blk_mq_clear_flush_rq_mapping() argument 3537 for (i = 0; i < queue_depth; i++) in blk_mq_clear_flush_rq_mapping() 3562 set->queue_depth, flush_rq); in blk_mq_exit_hctx() 3753 set->queue_depth); in __blk_mq_alloc_map_and_rqs() 4224 q->nr_requests = set->queue_depth; in blk_mq_init_allocated_queue() 4265 set->queue_depth); in __blk_mq_alloc_rq_maps() 4300 depth = set->queue_depth; in blk_mq_alloc_set_map_and_rqs() 4306 set->queue_depth >>= 1; in blk_mq_alloc_set_map_and_rqs() 4307 if (set->queue_depth < set->reserved_tags + BLK_MQ_TAG_MIN) { in blk_mq_alloc_set_map_and_rqs() 4311 } while (set->queue_depth); in blk_mq_alloc_set_map_and_rqs() [all …]
|
D | blk-mq-tag.h | 13 unsigned int queue_depth,
|
/linux-6.1.9/drivers/net/ethernet/microsoft/mana/ |
D | hw_channel.h | 135 u16 queue_depth; member 149 u16 queue_depth; member
|
/linux-6.1.9/drivers/block/null_blk/ |
D | null_blk.h | 33 unsigned int queue_depth; member 128 unsigned int queue_depth; member
|
D | main.c | 719 tag = find_first_zero_bit(nq->tag_map, nq->queue_depth); in get_tag() 720 if (tag >= nq->queue_depth) in get_tag() 1715 nq->queue_depth = nullb->queue_depth; in null_init_queue() 1819 nq->cmds = kcalloc(nq->queue_depth, sizeof(*cmd), GFP_KERNEL); in setup_commands() 1823 nq->tag_map = bitmap_zalloc(nq->queue_depth, GFP_KERNEL); in setup_commands() 1829 for (i = 0; i < nq->queue_depth; i++) { in setup_commands() 1849 nullb->queue_depth = nullb->dev->hw_queue_depth; in setup_queues() 1902 unsigned int queue_depth; in null_init_tag_set() local 1908 queue_depth = nullb->dev->hw_queue_depth; in null_init_tag_set() 1919 queue_depth = g_hw_queue_depth; in null_init_tag_set() [all …]
|
/linux-6.1.9/drivers/block/rnbd/ |
D | rnbd-srv.h | 25 int queue_depth; member
|
D | rnbd-clt.h | 86 size_t queue_depth; member
|
/linux-6.1.9/drivers/block/ |
D | virtio_blk.c | 889 module_param_named(queue_depth, virtblk_queue_depth, uint, 0444); 902 unsigned int queue_depth; in virtblk_probe() local 946 queue_depth = vblk->vqs[0].vq->num_free; in virtblk_probe() 949 queue_depth /= 2; in virtblk_probe() 951 queue_depth = virtblk_queue_depth; in virtblk_probe() 956 vblk->tag_set.queue_depth = queue_depth; in virtblk_probe()
|
/linux-6.1.9/drivers/scsi/snic/ |
D | snic_main.c | 91 if (qsz < sdev->queue_depth) in snic_change_queue_depth() 93 else if (qsz > sdev->queue_depth) in snic_change_queue_depth() 96 atomic64_set(&snic->s_stats.misc.last_qsz, sdev->queue_depth); in snic_change_queue_depth() 100 return sdev->queue_depth; in snic_change_queue_depth()
|
/linux-6.1.9/drivers/scsi/ |
D | scsi.c | 222 sdev->queue_depth = depth; in scsi_change_queue_depth() 229 sbitmap_resize(&sdev->budget_map, sdev->queue_depth); in scsi_change_queue_depth() 231 return sdev->queue_depth; in scsi_change_queue_depth()
|
/linux-6.1.9/include/uapi/linux/ |
D | ublk_cmd.h | 106 __u16 queue_depth; member
|