/linux-6.1.9/drivers/block/rnbd/ |
D | rnbd-proto.h | 247 bio_opf = REQ_OP_DISCARD; in rnbd_to_bio_flags() 278 case REQ_OP_DISCARD: in rq_to_rnbd_flags()
|
/linux-6.1.9/drivers/md/ |
D | dm-io.c | 314 if (op == REQ_OP_DISCARD) in do_region() 318 if ((op == REQ_OP_DISCARD || op == REQ_OP_WRITE_ZEROES) && in do_region() 334 case REQ_OP_DISCARD: in do_region() 349 if (op == REQ_OP_DISCARD || op == REQ_OP_WRITE_ZEROES) { in do_region()
|
D | dm-raid1.c | 620 if (bio_op(bio) == REQ_OP_DISCARD) { in write_callback() 659 if (bio_op(bio) == REQ_OP_DISCARD) { in do_write() 660 io_req.bi_opf = REQ_OP_DISCARD | op_flags; in do_write() 699 (bio_op(bio) == REQ_OP_DISCARD)) { in do_writes() 1246 bio_op(bio) != REQ_OP_DISCARD) in mirror_end_io()
|
D | dm-region-hash.c | 408 if (bio_op(bio) == REQ_OP_DISCARD) in dm_rh_mark_nosync() 531 if (bio->bi_opf & REQ_PREFLUSH || bio_op(bio) == REQ_OP_DISCARD) in dm_rh_inc_pending()
|
D | md-linear.c | 244 if (unlikely((bio_op(bio) == REQ_OP_DISCARD) && in linear_make_request()
|
D | dm-integrity.c | 1617 if (dio->op == REQ_OP_WRITE || unlikely(dio->op == REQ_OP_DISCARD)) in dec_in_flight() 1723 if (likely(dio->op != REQ_OP_DISCARD)) in integrity_metadata() 1737 if (unlikely(dio->op == REQ_OP_DISCARD)) { in integrity_metadata() 1867 if (unlikely(dio->op == REQ_OP_DISCARD)) { in dm_integrity_map() 1907 if (ic->sectors_per_block > 1 && likely(dio->op != REQ_OP_DISCARD)) { in dm_integrity_map() 2113 if (unlikely(dio->op == REQ_OP_DISCARD) && ic->mode != 'D') in dm_integrity_map_continue() 2132 if (ic->mode == 'J' && likely(dio->op != REQ_OP_DISCARD)) { in dm_integrity_map_continue() 2228 if (ic->mode == 'J' && likely(dio->op == REQ_OP_DISCARD) && !discard_retried) { in dm_integrity_map_continue() 2251 if (ic->mode == 'B' && (dio->op == REQ_OP_WRITE || unlikely(dio->op == REQ_OP_DISCARD))) { in dm_integrity_map_continue() 2280 if (unlikely(dio->op == REQ_OP_DISCARD) && likely(ic->mode != 'D')) { in dm_integrity_map_continue()
|
D | dm-thin.c | 413 bio_set_op_attrs(op->bio, REQ_OP_DISCARD, 0); in end_discard() 746 if (bio_op(bio) == REQ_OP_DISCARD) in inc_all_io_entry() 911 if (op_is_flush(bio->bi_opf) || bio_op(bio) == REQ_OP_DISCARD) in __inc_remap_and_issue_cell() 1812 bio_op(bio) == REQ_OP_DISCARD) in __remap_and_issue_shared_cell() 2201 if (bio_op(bio) == REQ_OP_DISCARD) in process_thin_deferred_bios() 2287 if (bio_op(cell->holder) == REQ_OP_DISCARD) in process_thin_deferred_cells() 2719 if (op_is_flush(bio->bi_opf) || bio_op(bio) == REQ_OP_DISCARD) { in thin_bio_map()
|
D | dm-ebs-target.c | 213 } else if (bio_op(bio) == REQ_OP_DISCARD) { in __ebs_process_bios()
|
D | dm-rq.c | 216 if (req_op(clone) == REQ_OP_DISCARD && in dm_done()
|
D | dm-stripe.c | 284 if (unlikely(bio_op(bio) == REQ_OP_DISCARD) || in stripe_map()
|
D | dm-snap.c | 1986 if (unlikely(bio_op(bio) == REQ_OP_DISCARD)) { in snapshot_map() 2005 if (unlikely(bio_op(bio) == REQ_OP_DISCARD) && in snapshot_map() 2016 if (unlikely(bio_op(bio) == REQ_OP_DISCARD)) { in snapshot_map() 2133 if (unlikely(bio_op(bio) == REQ_OP_DISCARD)) { in snapshot_merge_map()
|
/linux-6.1.9/block/ |
D | blk.h | 158 if (req_op(req) == REQ_OP_DISCARD && in blk_discard_mergable() 167 if (unlikely(op == REQ_OP_DISCARD || op == REQ_OP_SECURE_ERASE)) in blk_queue_get_max_sectors() 303 case REQ_OP_DISCARD: in bio_may_exceed_limits()
|
D | blk-merge.c | 351 case REQ_OP_DISCARD: in __bio_split_to_limits() 412 case REQ_OP_DISCARD: in blk_recalc_rq_segments() 581 if (req_op(rq) == REQ_OP_DISCARD) in blk_rq_get_max_segments() 597 req_op(rq) == REQ_OP_DISCARD || in blk_rq_get_max_sectors() 614 if (req_op(req) == REQ_OP_DISCARD) in ll_new_hw_segment()
|
D | blk-wbt.c | 465 if ((opf & REQ_OP_MASK) == REQ_OP_DISCARD) in get_limit() 536 case REQ_OP_DISCARD: in wbt_should_throttle() 555 if (bio_op(bio) == REQ_OP_DISCARD) in bio_to_wbt_flags()
|
D | bounce.c | 176 case REQ_OP_DISCARD: in bounce_clone_bio()
|
D | blk-lib.c | 67 bio = blk_next_bio(bio, bdev, 0, REQ_OP_DISCARD, gfp_mask); in __blkdev_issue_discard()
|
/linux-6.1.9/include/linux/ |
D | blk_types.h | 366 REQ_OP_DISCARD = (__force blk_opf_t)3, enumerator 509 return (op & REQ_OP_MASK) == REQ_OP_DISCARD; in op_is_discard()
|
D | bio.h | 56 bio_op(bio) != REQ_OP_DISCARD && in bio_has_data() 66 return bio_op(bio) == REQ_OP_DISCARD || in bio_no_advance_iter() 184 case REQ_OP_DISCARD: in bio_segments()
|
/linux-6.1.9/drivers/md/bcache/ |
D | writeback.h | 112 if (bio_op(bio) == REQ_OP_DISCARD) in should_writeback()
|
D | request.c | 373 (bio_op(bio) == REQ_OP_DISCARD)) in check_should_bypass() 993 if (bio_op(bio) == REQ_OP_DISCARD) in cached_dev_write() 1007 if (bio_op(bio) == REQ_OP_DISCARD && in cached_dev_write() 1123 if ((bio_op(bio) == REQ_OP_DISCARD) && in detached_dev_do_request() 1309 s->iop.bypass = (bio_op(bio) == REQ_OP_DISCARD) != 0; in flash_dev_submit_bio()
|
/linux-6.1.9/arch/um/drivers/ |
D | ubd_kern.c | 483 if ((io_req->error == BLK_STS_NOTSUPP) && (req_op(io_req->req) == REQ_OP_DISCARD)) { in ubd_handler() 1267 if (op == REQ_OP_WRITE_ZEROES || op == REQ_OP_DISCARD) { in ubd_map_req() 1332 else if (op == REQ_OP_WRITE_ZEROES || op == REQ_OP_DISCARD) in ubd_submit_request() 1369 case REQ_OP_DISCARD: in ubd_queue_rq() 1525 case REQ_OP_DISCARD: in do_io()
|
/linux-6.1.9/drivers/block/drbd/ |
D | drbd_req.c | 35 | (bio_op(bio_src) == REQ_OP_DISCARD ? RQ_UNMAP : 0); in drbd_req_new() 1170 else if (bio_op(bio) == REQ_OP_DISCARD) in drbd_submit_req_private_bio() 1227 bio_op(bio) == REQ_OP_DISCARD) in drbd_request_prepare()
|
/linux-6.1.9/drivers/mmc/core/ |
D | queue.c | 49 case REQ_OP_DISCARD: in mmc_cqe_issue_type()
|
/linux-6.1.9/Documentation/admin-guide/device-mapper/ |
D | log-writes.rst | 42 Any REQ_OP_DISCARD requests are treated like WRITE requests. Otherwise we would
|
/linux-6.1.9/fs/btrfs/ |
D | volumes.h | 593 case REQ_OP_DISCARD: in btrfs_op()
|