/linux/block/ |
A D | blk-crypto-internal.h | 34 return bio_crypt_ctx_mergeable(req->crypt_ctx, blk_rq_bytes(req), in bio_crypt_ctx_back_mergeable() 48 return bio_crypt_ctx_mergeable(req->crypt_ctx, blk_rq_bytes(req), in bio_crypt_ctx_merge_rq()
|
A D | blk-core.c | 247 rq->bio, rq->biotail, blk_rq_bytes(rq)); in blk_dump_rq_flags() 1162 should_fail_request(rq->rq_disk->part0, blk_rq_bytes(rq))) in blk_insert_cloned_request() 1199 return blk_rq_bytes(rq); in blk_rq_err_bytes() 1215 BUG_ON(blk_rq_bytes(rq) && !bytes); in blk_rq_err_bytes() 1456 rq->__data_len = blk_rq_bytes(rq_src); in blk_rq_prep_clone()
|
A D | bsg-lib.c | 223 buf->payload_len = blk_rq_bytes(req); in bsg_map_buffer()
|
A D | blk-mq.c | 781 if (blk_rq_bytes(req) < blk_rq_cur_bytes(req)) { in blk_update_request() 821 if (blk_update_request(rq, error, blk_rq_bytes(rq))) in blk_mq_end_request() 859 blk_update_request(rq, BLK_STS_OK, blk_rq_bytes(rq)); in blk_mq_end_request_batch() 2673 if (blk_rq_bytes(last) < BLK_PLUG_FLUSH_SIZE) in blk_mq_submit_bio()
|
A D | blk-merge.c | 841 req->__data_len += blk_rq_bytes(next); in attempt_merge()
|
A D | bfq-cgroup.c | 357 blkg_rwstat_add(&bfqg->stats.bytes, rq->cmd_flags, blk_rq_bytes(rq)); in bfqg_stats_update_legacy_io()
|
/linux/include/linux/ |
A D | blk-mq.h | 960 static inline unsigned int blk_rq_bytes(const struct request *rq) in blk_rq_bytes() function 978 return blk_rq_bytes(rq) >> SECTOR_SHIFT; in blk_rq_sectors() 1001 return blk_rq_bytes(rq); in blk_rq_payload_bytes()
|
/linux/drivers/scsi/ |
A D | scsi_lib.c | 842 } else if (blk_rq_bytes(req) == 0 && sense_current) { in scsi_io_completion_nz_result() 938 if (likely(blk_rq_bytes(req) > 0 || blk_stat == BLK_STS_OK)) { in scsi_io_completion() 945 if (scsi_end_request(req, blk_stat, blk_rq_bytes(req))) in scsi_io_completion() 1014 if (blk_rq_bytes(rq) & rq->q->dma_pad_mask) { in scsi_alloc_sgtables() 1016 (rq->q->dma_pad_mask & ~blk_rq_bytes(rq)) + 1; in scsi_alloc_sgtables() 1182 BUG_ON(blk_rq_bytes(req)); in scsi_setup_scsi_cmnd() 1189 cmd->transfersize = blk_rq_bytes(req); in scsi_setup_scsi_cmnd() 1553 if (blk_rq_bytes(req)) in scsi_prepare_cmd()
|
A D | sd_zbc.c | 463 scsi_set_resid(cmd, blk_rq_bytes(rq)); in sd_zbc_zone_wp_update()
|
A D | sd.c | 1111 rq->__data_len = blk_rq_bytes(rq); in sd_setup_write_same_cmnd() 2051 good_bytes = blk_rq_bytes(req); in sd_done() 2055 scsi_set_resid(SCpnt, blk_rq_bytes(req)); in sd_done()
|
/linux/kernel/trace/ |
A D | blktrace.c | 844 blk_add_trace_rq(rq, 0, blk_rq_bytes(rq), BLK_TA_INSERT, in blk_add_trace_rq_insert() 850 blk_add_trace_rq(rq, 0, blk_rq_bytes(rq), BLK_TA_ISSUE, in blk_add_trace_rq_issue() 856 blk_add_trace_rq(rq, 0, blk_rq_bytes(rq), BLK_TA_BACKMERGE, in blk_add_trace_rq_merge() 862 blk_add_trace_rq(rq, 0, blk_rq_bytes(rq), BLK_TA_REQUEUE, in blk_add_trace_rq_requeue() 1051 __blk_add_trace(bt, blk_rq_pos(rq), blk_rq_bytes(rq), in blk_add_trace_rq_remap() 1078 __blk_add_trace(bt, blk_rq_trace_sector(rq), blk_rq_bytes(rq), 0, 0, in blk_add_driver_data()
|
/linux/include/scsi/ |
A D | scsi_cmnd.h | 247 return blk_rq_bytes(scsi_cmd_to_rq(scmd)) >> shift; in scsi_logical_block_count()
|
/linux/include/trace/events/ |
A D | block.h | 167 __entry->bytes = blk_rq_bytes(rq);
|
/linux/drivers/mtd/ubi/ |
A D | block.c | 191 to_read = blk_rq_bytes(req); in ubiblock_read()
|
/linux/drivers/block/ |
A D | loop.c | 321 ret = file->f_op->fallocate(file, mode, pos, blk_rq_bytes(rq)); in lo_fallocate() 343 if (!cmd->use_aio || cmd->ret < 0 || cmd->ret == blk_rq_bytes(rq) || in lo_complete_rq() 441 iov_iter_bvec(&iter, rw, bvec, nr_bvec, blk_rq_bytes(rq)); in lo_rw_aio()
|
A D | nbd.c | 463 blk_rq_bytes(req), (req->timeout / HZ) * cmd->retries); in nbd_xmit_timeout() 559 unsigned long size = blk_rq_bytes(req); in nbd_send_cmd() 614 (unsigned long long)blk_rq_pos(req) << 9, blk_rq_bytes(req)); in nbd_send_cmd()
|
A D | ataflop.c | 463 blk_rq_bytes(fd_request))); in fd_end_request_cur()
|
/linux/drivers/s390/block/ |
A D | scm_blk.c | 190 aidaw = scm_aidaw_fetch(scmrq, blk_rq_bytes(req)); in scm_request_prepare()
|
/linux/drivers/memstick/core/ |
A D | mspro_block.c | 723 count = blk_rq_bytes(msb->block_req); in mspro_block_issue_req() 767 t_len = blk_rq_bytes(msb->block_req); in mspro_block_complete_req()
|
A D | ms_block.c | 1916 blk_rq_bytes(req), &len); in msb_io_work() 1919 blk_rq_bytes(req), &len); in msb_io_work()
|
/linux/arch/um/drivers/ |
A D | ubd_kern.c | 1272 io_req->io_desc[0].length = blk_rq_bytes(req); in ubd_map_req()
|
/linux/drivers/nvme/host/ |
A D | core.c | 813 req->q->write_hints[streamid] += blk_rq_bytes(req) >> 9; in nvme_assign_write_stream() 898 cpu_to_le16((blk_rq_bytes(req) >> ns->lba_shift) - 1); in nvme_setup_write_zeroes() 937 cmnd->rw.length = cpu_to_le16((blk_rq_bytes(req) >> ns->lba_shift) - 1); in nvme_setup_rw()
|
/linux/drivers/block/rnbd/ |
A D | rnbd-clt.c | 1025 msg.bi_size = cpu_to_le32(blk_rq_bytes(rq)); in rnbd_client_xfer_request()
|
/linux/drivers/md/ |
A D | dm-mpath.c | 505 size_t nr_bytes = blk_rq_bytes(rq); in multipath_clone_and_map()
|
/linux/drivers/block/null_blk/ |
A D | main.c | 1288 if (atomic_long_sub_return(blk_rq_bytes(rq), &nullb->cur_bytes) < 0) { in null_handle_throttled()
|