/linux-6.6.21/block/ |
D | elevator.c | 63 struct elevator_queue *e = q->elevator; in elv_iosched_allow_bio_merge() 161 struct elevator_queue *e = q->elevator; in elevator_exit() 188 struct elevator_queue *e = q->elevator; in elv_rqhash_add() 204 struct elevator_queue *e = q->elevator; in elv_rqhash_find() 279 struct elevator_queue *e = q->elevator; in elv_merge() 373 struct elevator_queue *e = q->elevator; in elv_merged_request() 387 struct elevator_queue *e = q->elevator; in elv_merge_requests() 398 struct elevator_queue *e = q->elevator; in elv_latter_request() 408 struct elevator_queue *e = q->elevator; in elv_former_request() 465 struct elevator_queue *e = q->elevator; in elv_register_queue() [all …]
|
D | mq-deadline.c | 230 struct deadline_data *dd = q->elevator->elevator_data; in dd_request_merged() 250 struct deadline_data *dd = q->elevator->elevator_data; in dd_merged_requests() 598 struct deadline_data *dd = hctx->queue->elevator->elevator_data; in dd_dispatch_request() 630 struct deadline_data *dd = data->q->elevator->elevator_data; in dd_limit_depth() 647 struct deadline_data *dd = q->elevator->elevator_data; in dd_depth_updated() 731 q->elevator = eq; in dd_init_sched() 746 struct deadline_data *dd = q->elevator->elevator_data; in dd_request_merge() 778 struct deadline_data *dd = q->elevator->elevator_data; in dd_bio_merge() 799 struct deadline_data *dd = q->elevator->elevator_data; in dd_insert_request() 869 struct deadline_data *dd = q->elevator->elevator_data; in dd_insert_requests() [all …]
|
D | blk-mq-sched.h | 41 struct elevator_queue *e = q->elevator; in blk_mq_sched_allow_merge() 52 struct elevator_queue *e = rq->q->elevator; in blk_mq_sched_completed_request() 63 struct elevator_queue *e = q->elevator; in blk_mq_sched_requeue_request() 72 struct elevator_queue *e = hctx->queue->elevator; in blk_mq_sched_has_work()
|
D | blk-mq-sched.c | 90 struct elevator_queue *e = q->elevator; in __blk_mq_do_dispatch_sched() 308 if (hctx->queue->elevator) in __blk_mq_sched_dispatch_requests() 342 struct elevator_queue *e = q->elevator; in blk_mq_sched_bio_merge() 483 eq = q->elevator; in blk_mq_init_sched() 501 q->elevator = NULL; in blk_mq_init_sched() 551 q->elevator = NULL; in blk_mq_exit_sched()
|
D | kyber-iosched.c | 425 q->elevator = eq; in kyber_init_sched() 455 struct kyber_queue_data *kqd = hctx->queue->elevator->elevator_data; in kyber_depth_updated() 561 struct kyber_queue_data *kqd = data->q->elevator->elevator_data; in kyber_limit_depth() 616 struct kyber_queue_data *kqd = rq->q->elevator->elevator_data; in kyber_finish_request() 641 struct kyber_queue_data *kqd = rq->q->elevator->elevator_data; in kyber_completed_request() 805 struct kyber_queue_data *kqd = hctx->queue->elevator->elevator_data; in kyber_dispatch_request() 904 struct kyber_queue_data *kqd = q->elevator->elevator_data; \ 963 struct kyber_queue_data *kqd = q->elevator->elevator_data; in KYBER_DEBUGFS_DOMAIN_ATTRS()
|
D | blk-flush.c | 246 if (!q->elevator) { in flush_end_io() 326 if (!q->elevator) { in blk_kick_flush() 367 if (q->elevator) { in mq_flush_data_end_io()
|
D | blk-ioc.c | 48 struct elevator_type *et = icq->q->elevator->type; in ioc_exit_icq() 77 struct elevator_type *et = q->elevator->type; in ioc_destroy_icq() 366 struct elevator_type *et = q->elevator->type; in ioc_create_icq()
|
D | blk-mq-debugfs.c | 667 if (q->elevator && !q->sched_debugfs_dir) in blk_mq_debugfs_register() 674 if (q->elevator && !hctx->sched_debugfs_dir) in blk_mq_debugfs_register() 748 struct elevator_type *e = q->elevator->type; in blk_mq_debugfs_register_sched() 819 struct elevator_type *e = q->elevator->type; in blk_mq_debugfs_register_sched_hctx()
|
D | Makefile | 6 obj-y := bdev.o fops.o bio.o elevator.o blk-core.o blk-sysfs.o \
|
D | blk-mq.c | 393 struct elevator_queue *e = data->q->elevator; in blk_mq_rq_ctx_init() 450 if (q->elevator) { in __blk_mq_alloc_requests() 463 struct elevator_mq_ops *ops = &q->elevator->type->ops; in __blk_mq_alloc_requests() 660 if (q->elevator) in blk_mq_alloc_request_hctx() 690 q->elevator->type->ops.finish_request(rq); in blk_mq_finish_request() 2569 } else if (q->elevator) { in blk_mq_insert_request() 2575 q->elevator->type->ops.insert_requests(hctx, &list, flags); in blk_mq_insert_request() 2794 } else if (this_hctx->queue->elevator) { in blk_mq_dispatch_plug_list() 2795 this_hctx->queue->elevator->type->ops.insert_requests(this_hctx, in blk_mq_dispatch_plug_list() 4655 if (q->elevator && q->elevator->type->ops.depth_updated) in blk_mq_update_nr_requests() [all …]
|
D | blk-sysfs.c | 815 if (q->elevator) { in blk_register_queue() 831 if (q->elevator) in blk_register_queue() 832 kobject_uevent(&q->elevator->kobj, KOBJ_ADD); in blk_register_queue()
|
D | bfq-iosched.c | 438 return bic->icq.q->elevator->elevator_data; in bic_to_bfqd() 692 struct bfq_data *bfqd = data->q->elevator->elevator_data; in bfq_limit_depth() 2448 struct bfq_data *bfqd = q->elevator->elevator_data; in bfq_bio_merge() 2488 struct bfq_data *bfqd = q->elevator->elevator_data; in bfq_request_merge() 3230 struct bfq_data *bfqd = q->elevator->elevator_data; in bfq_allow_bio_merge() 3670 bfq_update_peak_rate(q->elevator->elevator_data, rq); in bfq_dispatch_remove() 5149 struct bfq_data *bfqd = hctx->queue->elevator->elevator_data; in bfq_has_work() 5161 struct bfq_data *bfqd = hctx->queue->elevator->elevator_data; in __bfq_dispatch_request() 5302 struct bfq_data *bfqd = hctx->queue->elevator->elevator_data; in bfq_dispatch_request() 6243 struct bfq_data *bfqd = q->elevator->elevator_data; in bfq_insert_request() [all …]
|
D | blk-wbt.c | 718 if (q->elevator && in wbt_enable_default() 719 test_bit(ELEVATOR_FLAG_DISABLE_WBT, &q->elevator->flags)) in wbt_enable_default()
|
D | blk-mq-tag.c | 108 if (!data->q->elevator && !(data->flags & BLK_MQ_REQ_RESERVED) && in __blk_mq_get_tag()
|
D | genhd.c | 554 if (disk->queue->elevator) in device_add_disk() 704 if (q->elevator) { in del_gendisk()
|
D | bfq-cgroup.c | 533 struct bfq_data *bfqd = blkg->q->elevator->elevator_data; in bfq_pd_init()
|
/linux-6.6.21/tools/cgroup/ |
D | iocost_coef_gen.py | 107 global elevator_path, nomerges_path, elevator, nomerges 111 f.write(elevator) 143 elevator = re.sub(r'.*\[(.*)\].*', r'\1', f.read().strip()) variable
|
/linux-6.6.21/Documentation/filesystems/ |
D | zonefs.rst | 148 implemented by the block layer elevator. An elevator implementing the sequential 149 write feature for zoned block device (ELEVATOR_F_ZBD_SEQ_WRITE elevator feature) 150 must be used. This type of elevator (e.g. mq-deadline) is set by default
|
/linux-6.6.21/include/linux/ |
D | blkdev.h | 380 struct elevator_queue *elevator; member
|
/linux-6.6.21/Documentation/ABI/stable/ |
D | sysfs-block | 465 [RW] Standard I/O elevator operations include attempts to merge
|
/linux-6.6.21/Documentation/admin-guide/LSM/ |
D | Smack.rst | 6 "Good for you, you've decided to clean the elevator!"
|
/linux-6.6.21/ |
D | CREDITS | 158 D: elevator + block layer rewrites
|