/linux-6.1.9/arch/sparc/kernel/ |
D | visemul.c | 294 static void edge(struct pt_regs *regs, unsigned int insn, unsigned int opf) in edge() argument 307 switch (opf) { in edge() 352 switch (opf) { in edge() 372 static void array(struct pt_regs *regs, unsigned int insn, unsigned int opf) in array() argument 394 switch (opf) { in array() 477 static void pformat(struct pt_regs *regs, unsigned int insn, unsigned int opf) in pformat() argument 483 scale = (gsr >> 3) & (opf == FPACK16_OPF ? 0xf : 0x1f); in pformat() 484 switch (opf) { in pformat() 589 static void pmul(struct pt_regs *regs, unsigned int insn, unsigned int opf) in pmul() argument 594 switch (opf) { in pmul() [all …]
|
/linux-6.1.9/drivers/md/ |
D | dm-io.c | 296 static void do_region(const blk_opf_t opf, unsigned region, in do_region() argument 309 const enum req_op op = opf & REQ_OP_MASK; in do_region() 343 bio = bio_alloc_bioset(where->bdev, num_bvecs, opf, GFP_NOIO, in do_region() 372 static void dispatch_io(blk_opf_t opf, unsigned int num_regions, in dispatch_io() argument 382 opf |= REQ_SYNC; in dispatch_io() 390 if (where[i].count || (opf & REQ_PREFLUSH)) in dispatch_io() 391 do_region(opf, i, where + i, dp, io); in dispatch_io() 415 struct dm_io_region *where, blk_opf_t opf, struct dpages *dp, in sync_io() argument 421 if (num_regions > 1 && !op_is_write(opf)) { in sync_io() 438 dispatch_io(opf, num_regions, where, dp, io, 1); in sync_io() [all …]
|
D | dm-snap-persistent.c | 229 static int chunk_io(struct pstore *ps, void *area, chunk_t chunk, blk_opf_t opf, in chunk_io() argument 238 .bi_opf = opf, in chunk_io() 284 static int area_io(struct pstore *ps, blk_opf_t opf) in area_io() argument 288 return chunk_io(ps, ps->area, chunk, opf, 0); in area_io()
|
/linux-6.1.9/block/ |
D | blk-cgroup-rwstat.h | 62 blk_opf_t opf, uint64_t val) in blkg_rwstat_add() argument 66 if (op_is_discard(opf)) in blkg_rwstat_add() 68 else if (op_is_write(opf)) in blkg_rwstat_add() 75 if (op_is_sync(opf)) in blkg_rwstat_add()
|
D | fops.c | 37 blk_opf_t opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; in dio_bio_write_op() local 41 opf |= REQ_FUA; in dio_bio_write_op() 42 return opf; in dio_bio_write_op() 178 blk_opf_t opf = is_read ? REQ_OP_READ : dio_bio_write_op(iocb); in __blkdev_direct_IO() local 186 opf |= REQ_ALLOC_CACHE; in __blkdev_direct_IO() 187 bio = bio_alloc_bioset(bdev, nr_pages, opf, GFP_KERNEL, in __blkdev_direct_IO() 244 bio = bio_alloc(bdev, nr_pages, opf, GFP_KERNEL); in __blkdev_direct_IO() 300 blk_opf_t opf = is_read ? REQ_OP_READ : dio_bio_write_op(iocb); in __blkdev_direct_IO_async() local 310 opf |= REQ_ALLOC_CACHE; in __blkdev_direct_IO_async() 311 bio = bio_alloc_bioset(bdev, nr_pages, opf, GFP_KERNEL, in __blkdev_direct_IO_async()
|
D | blk-wbt.c | 454 static inline unsigned int get_limit(struct rq_wb *rwb, blk_opf_t opf) in get_limit() argument 465 if ((opf & REQ_OP_MASK) == REQ_OP_DISCARD) in get_limit() 476 if ((opf & REQ_HIPRIO) || wb_recent_wait(rwb) || current_is_kswapd()) in get_limit() 478 else if ((opf & REQ_BACKGROUND) || close_io(rwb)) { in get_limit() 493 blk_opf_t opf; member 499 return rq_wait_inc_below(rqw, get_limit(data->rwb, data->opf)); in wbt_inflight_cb() 513 blk_opf_t opf) in __wbt_wait() argument 519 .opf = opf, in __wbt_wait()
|
D | blk-mq.h | 89 static inline enum hctx_type blk_mq_get_hctx_type(blk_opf_t opf) in blk_mq_get_hctx_type() argument 96 if (opf & REQ_POLLED) in blk_mq_get_hctx_type() 98 else if ((opf & REQ_OP_MASK) == REQ_OP_READ) in blk_mq_get_hctx_type() 110 blk_opf_t opf, in blk_mq_map_queue() argument 113 return ctx->hctxs[blk_mq_get_hctx_type(opf)]; in blk_mq_map_queue()
|
D | bio.c | 242 unsigned short max_vecs, blk_opf_t opf) in bio_init() argument 246 bio->bi_opf = opf; in bio_init() 295 void bio_reset(struct bio *bio, struct block_device *bdev, blk_opf_t opf) in bio_reset() argument 303 bio->bi_opf = opf; in bio_reset() 344 unsigned int nr_pages, blk_opf_t opf, gfp_t gfp) in blk_next_bio() argument 346 struct bio *new = bio_alloc(bdev, nr_pages, opf, gfp); in blk_next_bio() 412 unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp, in bio_alloc_percpu_cache() argument 428 bio_init(bio, bdev, nr_vecs ? bio->bi_inline_vecs : NULL, nr_vecs, opf); in bio_alloc_percpu_cache() 471 blk_opf_t opf, gfp_t gfp_mask, in bio_alloc_bioset() argument 482 if (opf & REQ_ALLOC_CACHE) { in bio_alloc_bioset() [all …]
|
D | bfq-cgroup.c | 223 blk_opf_t opf) in bfqg_stats_update_io_add() argument 225 blkg_rwstat_add(&bfqg->stats.queued, opf, 1); in bfqg_stats_update_io_add() 231 void bfqg_stats_update_io_remove(struct bfq_group *bfqg, blk_opf_t opf) in bfqg_stats_update_io_remove() argument 233 blkg_rwstat_add(&bfqg->stats.queued, opf, -1); in bfqg_stats_update_io_remove() 236 void bfqg_stats_update_io_merged(struct bfq_group *bfqg, blk_opf_t opf) in bfqg_stats_update_io_merged() argument 238 blkg_rwstat_add(&bfqg->stats.merged, opf, 1); in bfqg_stats_update_io_merged() 242 u64 io_start_time_ns, blk_opf_t opf) in bfqg_stats_update_completion() argument 248 blkg_rwstat_add(&stats->service_time, opf, in bfqg_stats_update_completion() 251 blkg_rwstat_add(&stats->wait_time, opf, in bfqg_stats_update_completion() 257 void bfqg_stats_update_io_remove(struct bfq_group *bfqg, blk_opf_t opf) { } in bfqg_stats_update_io_remove() argument [all …]
|
D | bfq-iosched.h | 992 void bfqg_stats_update_io_remove(struct bfq_group *bfqg, blk_opf_t opf); 993 void bfqg_stats_update_io_merged(struct bfq_group *bfqg, blk_opf_t opf); 995 u64 io_start_time_ns, blk_opf_t opf); 1003 blk_opf_t opf);
|
D | kyber-iosched.c | 198 static unsigned int kyber_sched_domain(blk_opf_t opf) in kyber_sched_domain() argument 200 switch (opf & REQ_OP_MASK) { in kyber_sched_domain() 556 static void kyber_limit_depth(blk_opf_t opf, struct blk_mq_alloc_data *data) in kyber_limit_depth() argument 562 if (!op_is_sync(opf)) { in kyber_limit_depth()
|
/linux-6.1.9/drivers/nvme/target/ |
D | io-cmd-bdev.c | 248 blk_opf_t opf; in nvmet_bdev_execute_rw() local 263 opf = REQ_OP_WRITE | REQ_SYNC | REQ_IDLE; in nvmet_bdev_execute_rw() 265 opf |= REQ_FUA; in nvmet_bdev_execute_rw() 268 opf = REQ_OP_READ; in nvmet_bdev_execute_rw() 273 opf |= REQ_NOMERGE; in nvmet_bdev_execute_rw() 280 ARRAY_SIZE(req->inline_bvec), opf); in nvmet_bdev_execute_rw() 282 bio = bio_alloc(req->ns->bdev, bio_max_segs(sg_cnt), opf, in nvmet_bdev_execute_rw() 309 opf, GFP_KERNEL); in nvmet_bdev_execute_rw()
|
D | zns.c | 540 const blk_opf_t opf = REQ_OP_ZONE_APPEND | REQ_SYNC | REQ_IDLE; in nvmet_bdev_execute_zone_append() local 571 ARRAY_SIZE(req->inline_bvec), opf); in nvmet_bdev_execute_zone_append() 573 bio = bio_alloc(req->ns->bdev, req->sg_cnt, opf, GFP_KERNEL); in nvmet_bdev_execute_zone_append()
|
/linux-6.1.9/fs/ |
D | direct-io.c | 120 blk_opf_t opf; /* request operation type and flags */ member 169 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in dio_refill_pages() 236 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in dio_complete() 332 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in dio_bio_end_aio() 400 bio = bio_alloc(bdev, nr_vecs, dio->opf, GFP_KERNEL); in dio_bio_alloc() 419 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in dio_bio_submit() 495 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in dio_bio_complete() 623 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in get_more_blocks() 806 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in submit_page_section() 923 const enum req_op dio_op = dio->opf & REQ_OP_MASK; in do_direct_IO() [all …]
|
D | mpage.c | 150 blk_opf_t opf = REQ_OP_READ; in do_mpage_readpage() local 159 opf |= REQ_RAHEAD; in do_mpage_readpage() 277 args->bio = bio_alloc(bdev, bio_max_segs(args->nr_pages), opf, in do_mpage_readpage()
|
D | buffer.c | 55 static void submit_bh_wbc(blk_opf_t opf, struct buffer_head *bh, 2659 static void submit_bh_wbc(blk_opf_t opf, struct buffer_head *bh, in submit_bh_wbc() argument 2662 const enum req_op op = opf & REQ_OP_MASK; in submit_bh_wbc() 2678 opf |= REQ_META; in submit_bh_wbc() 2680 opf |= REQ_PRIO; in submit_bh_wbc() 2682 bio = bio_alloc(bh->b_bdev, 1, opf, GFP_NOIO); in submit_bh_wbc() 2705 void submit_bh(blk_opf_t opf, struct buffer_head *bh) in submit_bh() argument 2707 submit_bh_wbc(opf, bh, NULL); in submit_bh()
|
/linux-6.1.9/drivers/target/ |
D | target_core_iblock.c | 348 blk_opf_t opf) in iblock_get_bio() argument 357 bio = bio_alloc_bioset(ib_dev->ibd_bd, bio_max_segs(sg_num), opf, in iblock_get_bio() 728 blk_opf_t opf; in iblock_execute_rw() local 740 opf = REQ_OP_WRITE; in iblock_execute_rw() 744 opf |= REQ_FUA; in iblock_execute_rw() 746 opf |= REQ_FUA; in iblock_execute_rw() 749 opf = REQ_OP_READ; in iblock_execute_rw() 764 bio = iblock_get_bio(cmd, block_lba, sgl_nents, opf); in iblock_execute_rw() 797 bio = iblock_get_bio(cmd, block_lba, sg_num, opf); in iblock_execute_rw()
|
/linux-6.1.9/fs/hfsplus/ |
D | wrapper.c | 48 void *buf, void **data, blk_opf_t opf) in hfsplus_submit_bio() argument 50 const enum req_op op = opf & REQ_OP_MASK; in hfsplus_submit_bio() 67 bio = bio_alloc(sb->s_bdev, 1, opf, GFP_NOIO); in hfsplus_submit_bio()
|
/linux-6.1.9/fs/nilfs2/ |
D | btnode.c | 73 sector_t pblocknr, blk_opf_t opf, in nilfs_btnode_submit_block() argument 106 if (opf & REQ_RAHEAD) { in nilfs_btnode_submit_block() 125 submit_bh(opf, bh); in nilfs_btnode_submit_block()
|
D | mdt.c | 114 nilfs_mdt_submit_block(struct inode *inode, unsigned long blkoff, blk_opf_t opf, in nilfs_mdt_submit_block() argument 129 if (opf & REQ_RAHEAD) { in nilfs_mdt_submit_block() 151 submit_bh(opf, bh); in nilfs_mdt_submit_block() 155 opf & REQ_OP_MASK); in nilfs_mdt_submit_block()
|
/linux-6.1.9/kernel/trace/ |
D | blktrace.c | 216 const blk_opf_t opf, u32 what, int error, in __blk_add_trace() argument 230 const enum req_op op = opf & REQ_OP_MASK; in __blk_add_trace() 236 what |= MASK_TC_BIT(opf, SYNC); in __blk_add_trace() 237 what |= MASK_TC_BIT(opf, RAHEAD); in __blk_add_trace() 238 what |= MASK_TC_BIT(opf, META); in __blk_add_trace() 239 what |= MASK_TC_BIT(opf, PREFLUSH); in __blk_add_trace() 240 what |= MASK_TC_BIT(opf, FUA); in __blk_add_trace() 1880 void blk_fill_rwbs(char *rwbs, blk_opf_t opf) in blk_fill_rwbs() argument 1884 if (opf & REQ_PREFLUSH) in blk_fill_rwbs() 1887 switch (opf & REQ_OP_MASK) { in blk_fill_rwbs() [all …]
|
/linux-6.1.9/include/linux/ |
D | bio.h | 408 blk_opf_t opf, gfp_t gfp_mask, 421 unsigned short nr_vecs, blk_opf_t opf, gfp_t gfp_mask) in bio_alloc() argument 423 return bio_alloc_bioset(bdev, nr_vecs, opf, gfp_mask, &fs_bio_set); in bio_alloc() 459 unsigned short max_vecs, blk_opf_t opf); 461 void bio_reset(struct bio *bio, struct block_device *bdev, blk_opf_t opf); 792 unsigned int nr_pages, blk_opf_t opf, gfp_t gfp);
|
D | blktrace_api.h | 109 void blk_fill_rwbs(char *rwbs, blk_opf_t opf);
|
/linux-6.1.9/fs/gfs2/ |
D | lops.h | 19 extern void gfs2_log_submit_bio(struct bio **biop, blk_opf_t opf);
|
/linux-6.1.9/fs/iomap/ |
D | direct-io.c | 55 struct iomap_dio *dio, unsigned short nr_vecs, blk_opf_t opf) in iomap_dio_alloc_bio() argument 58 return bio_alloc_bioset(iter->iomap.bdev, nr_vecs, opf, in iomap_dio_alloc_bio() 60 return bio_alloc(iter->iomap.bdev, nr_vecs, opf, GFP_KERNEL); in iomap_dio_alloc_bio()
|