Home
last modified time | relevance | path

Searched refs:max_hw_sectors (Results 1 – 19 of 19) sorted by relevance

/linux-6.1.9/block/
Dblk-settings.c42 lim->max_sectors = lim->max_hw_sectors = BLK_SAFE_MAX_SECTORS; in blk_set_default_limits()
78 lim->max_hw_sectors = UINT_MAX; in blk_set_stacking_limits()
122 void blk_queue_max_hw_sectors(struct request_queue *q, unsigned int max_hw_sectors) in blk_queue_max_hw_sectors() argument
127 if ((max_hw_sectors << 9) < PAGE_SIZE) { in blk_queue_max_hw_sectors()
128 max_hw_sectors = 1 << (PAGE_SHIFT - 9); in blk_queue_max_hw_sectors()
130 __func__, max_hw_sectors); in blk_queue_max_hw_sectors()
133 max_hw_sectors = round_down(max_hw_sectors, in blk_queue_max_hw_sectors()
135 limits->max_hw_sectors = max_hw_sectors; in blk_queue_max_hw_sectors()
137 max_sectors = min_not_zero(max_hw_sectors, limits->max_dev_sectors); in blk_queue_max_hw_sectors()
218 max_sectors = min(q->limits.max_hw_sectors, max_zone_append_sectors); in blk_queue_max_zone_append_sectors()
[all …]
Dblk-merge.c593 return q->limits.max_hw_sectors; in blk_rq_get_max_sectors()
/linux-6.1.9/drivers/nvme/target/
Dpassthru.c86 unsigned int max_hw_sectors; in nvmet_passthru_override_id_ctrl() local
105 max_hw_sectors = min_not_zero(pctrl->max_segments << (PAGE_SHIFT - 9), in nvmet_passthru_override_id_ctrl()
106 pctrl->max_hw_sectors); in nvmet_passthru_override_id_ctrl()
112 max_hw_sectors = min_not_zero(BIO_MAX_VECS << (PAGE_SHIFT - 9), in nvmet_passthru_override_id_ctrl()
113 max_hw_sectors); in nvmet_passthru_override_id_ctrl()
117 id->mdts = ilog2(max_hw_sectors) + 9 - page_shift; in nvmet_passthru_override_id_ctrl()
Dloop.c375 ctrl->ctrl.max_hw_sectors = in nvme_loop_configure_admin_queue()
/linux-6.1.9/drivers/block/rnbd/
Drnbd-proto.h138 __le32 max_hw_sectors; member
DREADME73 information: side, max_hw_sectors, etc.
Drnbd-srv.c547 rsp->max_hw_sectors = in rnbd_srv_fill_msg_open_rsp()
/linux-6.1.9/drivers/nvme/host/
Dcore.c1848 if (ctrl->max_hw_sectors) { in nvme_set_queue_limits()
1850 (ctrl->max_hw_sectors / (NVME_CTRL_PAGE_SIZE >> 9)) + 1; in nvme_set_queue_limits()
1853 blk_queue_max_hw_sectors(q, ctrl->max_hw_sectors); in nvme_set_queue_limits()
1948 is_power_of_2(ctrl->max_hw_sectors)) in nvme_set_chunk_sectors()
1949 iob = ctrl->max_hw_sectors; in nvme_set_chunk_sectors()
3063 ctrl->max_zeroes_sectors = ctrl->max_hw_sectors; in nvme_init_non_mdts_limits()
3096 u32 max_hw_sectors; in nvme_init_identify() local
3157 max_hw_sectors = nvme_mps_to_sectors(ctrl, id->mdts); in nvme_init_identify()
3159 max_hw_sectors = UINT_MAX; in nvme_init_identify()
3160 ctrl->max_hw_sectors = in nvme_init_identify()
[all …]
Dzns.c44 ctrl->max_zone_append = ctrl->max_hw_sectors; in nvme_set_max_append()
Dnvme.h282 u32 max_hw_sectors; member
Dmultipath.c877 size_t max_transfer_size = ctrl->max_hw_sectors << SECTOR_SHIFT; in nvme_mpath_init_identify()
Dapple.c1038 anv->ctrl.max_hw_sectors = min_t(u32, NVME_MAX_KB_SZ << 1, in apple_nvme_reset_work()
Drdma.c867 ctrl->ctrl.max_hw_sectors = ctrl->max_fr_pages << (ilog2(SZ_4K) - 9); in nvme_rdma_configure_admin_queue()
Dpci.c2851 dev->ctrl.max_hw_sectors = min_t(u32, in nvme_reset_work()
Dfc.c3104 ctrl->ctrl.max_hw_sectors = ctrl->ctrl.max_segments << in nvme_fc_create_association()
/linux-6.1.9/include/linux/
Dblkdev.h287 unsigned int max_hw_sectors; member
1138 return q->limits.max_hw_sectors; in queue_max_hw_sectors()
/linux-6.1.9/drivers/block/drbd/
Ddrbd_nl.c1263 unsigned int max_hw_sectors = max_bio_size >> 9; in drbd_setup_queue_param() local
1271 max_hw_sectors = min(queue_max_hw_sectors(b), max_bio_size >> 9); in drbd_setup_queue_param()
1280 blk_queue_max_hw_sectors(q, max_hw_sectors); in drbd_setup_queue_param()
/linux-6.1.9/drivers/md/bcache/
Dsuper.c950 q->limits.max_hw_sectors = UINT_MAX; in bcache_device_init()
/linux-6.1.9/drivers/scsi/
Dsd.c3315 q->limits.max_sectors > q->limits.max_hw_sectors) in sd_revalidate_disk()