/linux-5.19.10/block/ |
D | blk-rq-qos.c | 32 void __rq_qos_cleanup(struct rq_qos *rqos, struct bio *bio) in __rq_qos_cleanup() argument 35 if (rqos->ops->cleanup) in __rq_qos_cleanup() 36 rqos->ops->cleanup(rqos, bio); in __rq_qos_cleanup() 37 rqos = rqos->next; in __rq_qos_cleanup() 38 } while (rqos); in __rq_qos_cleanup() 41 void __rq_qos_done(struct rq_qos *rqos, struct request *rq) in __rq_qos_done() argument 44 if (rqos->ops->done) in __rq_qos_done() 45 rqos->ops->done(rqos, rq); in __rq_qos_done() 46 rqos = rqos->next; in __rq_qos_done() 47 } while (rqos); in __rq_qos_done() [all …]
|
D | blk-wbt.c | 100 struct bdi_writeback *wb = &rwb->rqos.q->disk->bdi->wb; in wb_recent_wait() 170 static void __wbt_done(struct rq_qos *rqos, enum wbt_flags wb_acct) in __wbt_done() argument 172 struct rq_wb *rwb = RQWB(rqos); in __wbt_done() 186 static void wbt_done(struct rq_qos *rqos, struct request *rq) in wbt_done() argument 188 struct rq_wb *rwb = RQWB(rqos); in wbt_done() 200 __wbt_done(rqos, wbt_flags(rq)); in wbt_done() 237 struct backing_dev_info *bdi = rwb->rqos.q->disk->bdi; in latency_exceeded() 290 struct backing_dev_info *bdi = rwb->rqos.q->disk->bdi; in rwb_trace_step() 360 if (!rwb->rqos.q->disk) in wb_timer_fn() 365 trace_wbt_timer(rwb->rqos.q->disk->bdi, status, rqd->scale_step, in wb_timer_fn() [all …]
|
D | blk-rq-qos.h | 65 struct rq_qos *rqos; in rq_qos_id() local 66 for (rqos = q->rq_qos; rqos; rqos = rqos->next) { in rq_qos_id() 67 if (rqos->id == id) in rq_qos_id() 70 return rqos; in rq_qos_id() 89 static inline int rq_qos_add(struct request_queue *q, struct rq_qos *rqos) in rq_qos_add() argument 101 if (rq_qos_id(q, rqos->id)) in rq_qos_add() 103 rqos->next = q->rq_qos; in rq_qos_add() 104 q->rq_qos = rqos; in rq_qos_add() 109 if (rqos->ops->debugfs_attrs) { in rq_qos_add() 111 blk_mq_debugfs_register_rqos(rqos); in rq_qos_add() [all …]
|
D | blk-ioprio.c | 187 struct rq_qos rqos; member 190 static void blkcg_ioprio_track(struct rq_qos *rqos, struct request *rq, in blkcg_ioprio_track() argument 213 static void blkcg_ioprio_exit(struct rq_qos *rqos) in blkcg_ioprio_exit() argument 216 container_of(rqos, typeof(*blkioprio_blkg), rqos); in blkcg_ioprio_exit() 218 blkcg_deactivate_policy(rqos->q, &ioprio_policy); in blkcg_ioprio_exit() 230 struct rq_qos *rqos; in blk_ioprio_init() local 243 rqos = &blkioprio_blkg->rqos; in blk_ioprio_init() 244 rqos->id = RQ_QOS_IOPRIO; in blk_ioprio_init() 245 rqos->ops = &blkcg_ioprio_ops; in blk_ioprio_init() 246 rqos->q = q; in blk_ioprio_init() [all …]
|
D | blk-iolatency.c | 88 struct rq_qos rqos; member 103 static inline struct blk_iolatency *BLKIOLATENCY(struct rq_qos *rqos) in BLKIOLATENCY() argument 105 return container_of(rqos, struct blk_iolatency, rqos); in BLKIOLATENCY() 286 static void __blkcg_iolatency_throttle(struct rq_qos *rqos, in __blkcg_iolatency_throttle() argument 295 blkcg_schedule_throttle(rqos->q, use_memdelay); in __blkcg_iolatency_throttle() 333 unsigned long qd = blkiolat->rqos.q->nr_requests; in scale_cookie_change() 373 unsigned long qd = iolat->blkiolat->rqos.q->nr_requests; in scale_change() 466 static void blkcg_iolatency_throttle(struct rq_qos *rqos, struct bio *bio) in blkcg_iolatency_throttle() argument 468 struct blk_iolatency *blkiolat = BLKIOLATENCY(rqos); in blkcg_iolatency_throttle() 483 __blkcg_iolatency_throttle(rqos, iolat, issue_as_root, in blkcg_iolatency_throttle() [all …]
|
D | blk-mq-debugfs.h | 36 void blk_mq_debugfs_register_rqos(struct rq_qos *rqos); 37 void blk_mq_debugfs_unregister_rqos(struct rq_qos *rqos); 77 static inline void blk_mq_debugfs_register_rqos(struct rq_qos *rqos) in blk_mq_debugfs_register_rqos() argument 81 static inline void blk_mq_debugfs_unregister_rqos(struct rq_qos *rqos) in blk_mq_debugfs_unregister_rqos() argument
|
D | blk-mq-debugfs.c | 705 struct rq_qos *rqos = q->rq_qos; in blk_mq_debugfs_register() local 707 while (rqos) { in blk_mq_debugfs_register() 708 blk_mq_debugfs_register_rqos(rqos); in blk_mq_debugfs_register() 709 rqos = rqos->next; in blk_mq_debugfs_register() 816 void blk_mq_debugfs_unregister_rqos(struct rq_qos *rqos) in blk_mq_debugfs_unregister_rqos() argument 818 lockdep_assert_held(&rqos->q->debugfs_mutex); in blk_mq_debugfs_unregister_rqos() 820 if (!rqos->q->debugfs_dir) in blk_mq_debugfs_unregister_rqos() 822 debugfs_remove_recursive(rqos->debugfs_dir); in blk_mq_debugfs_unregister_rqos() 823 rqos->debugfs_dir = NULL; in blk_mq_debugfs_unregister_rqos() 826 void blk_mq_debugfs_register_rqos(struct rq_qos *rqos) in blk_mq_debugfs_register_rqos() argument [all …]
|
D | blk-wbt.h | 68 struct rq_qos rqos; member 73 static inline struct rq_wb *RQWB(struct rq_qos *rqos) in RQWB() argument 75 return container_of(rqos, struct rq_wb, rqos); in RQWB()
|
D | blk-iocost.c | 403 struct rq_qos rqos; member 657 static struct ioc *rqos_to_ioc(struct rq_qos *rqos) in rqos_to_ioc() argument 659 return container_of(rqos, struct ioc, rqos); in rqos_to_ioc() 677 return q_name(ioc->rqos.q); in ioc_name() 812 if (!blk_queue_nonrot(ioc->rqos.q)) in ioc_autop_idx() 816 if (blk_queue_depth(ioc->rqos.q) == 1) in ioc_autop_idx() 2560 static void ioc_rqos_throttle(struct rq_qos *rqos, struct bio *bio) in ioc_rqos_throttle() argument 2563 struct ioc *ioc = rqos_to_ioc(rqos); in ioc_rqos_throttle() 2643 blkcg_schedule_throttle(rqos->q, in ioc_rqos_throttle() 2695 static void ioc_rqos_merge(struct rq_qos *rqos, struct request *rq, in ioc_rqos_merge() argument [all …]
|
D | blk-sysfs.c | 474 struct rq_qos *rqos; in queue_wb_lat_store() local 484 rqos = wbt_rq_qos(q); in queue_wb_lat_store() 485 if (!rqos) { in queue_wb_lat_store()
|
/linux-5.19.10/drivers/net/ethernet/freescale/ |
D | ucc_geth.h | 1071 enum ucc_geth_qos_mode rqos; member
|