Lines Matching refs:lba

876 			    unsigned long long lba)  in lba2fake_store()  argument
880 lba = do_div(lba, sdebug_store_sectors); in lba2fake_store()
885 return lsip->storep + lba * sdebug_sector_size; in lba2fake_store()
2671 unsigned long long lba) in zbc_zone() argument
2673 return &devip->zstate[lba >> devip->zsize_shift]; in zbc_zone()
2751 unsigned long long lba, unsigned int num) in zbc_inc_wp() argument
2753 struct sdeb_zone_state *zsp = zbc_zone(devip, lba); in zbc_inc_wp()
2767 if (lba != zsp->z_wp) in zbc_inc_wp()
2770 end = lba + num; in zbc_inc_wp()
2772 n = zend - lba; in zbc_inc_wp()
2784 lba += n; in zbc_inc_wp()
2793 unsigned long long lba, unsigned int num, bool write) in check_zbc_access_params() argument
2797 struct sdeb_zone_state *zsp = zbc_zone(devip, lba); in check_zbc_access_params()
2798 struct sdeb_zone_state *zsp_end = zbc_zone(devip, lba + num - 1); in check_zbc_access_params()
2841 if (lba != zsp->z_wp) { in check_zbc_access_params()
2865 (struct scsi_cmnd *scp, unsigned long long lba, in check_device_access_params() argument
2871 if (lba + num > sdebug_capacity) { in check_device_access_params()
2886 return check_zbc_access_params(scp, lba, num, write); in check_device_access_params()
2909 u32 sg_skip, u64 lba, u32 num, bool do_write) in do_device_access() argument
2930 block = do_div(lba, sdebug_store_sectors); in do_device_access()
2966 static bool comp_write_worker(struct sdeb_store_info *sip, u64 lba, u32 num, in comp_write_worker() argument
2975 block = do_div(lba, store_blks); in comp_write_worker()
3127 u64 lba; in resp_read_dt0() local
3135 lba = get_unaligned_be64(cmd + 2); in resp_read_dt0()
3141 lba = get_unaligned_be32(cmd + 2); in resp_read_dt0()
3147 lba = (u32)cmd[3] | (u32)cmd[2] << 8 | in resp_read_dt0()
3154 lba = get_unaligned_be32(cmd + 2); in resp_read_dt0()
3160 lba = get_unaligned_be32(cmd + 2); in resp_read_dt0()
3165 lba = get_unaligned_be64(cmd + 12); in resp_read_dt0()
3189 ret = check_device_access_params(scp, lba, num, false); in resp_read_dt0()
3193 (lba <= (sdebug_medium_error_start + sdebug_medium_error_count - 1)) && in resp_read_dt0()
3194 ((lba + num) > sdebug_medium_error_start))) { in resp_read_dt0()
3200 ret = (lba < OPT_MEDIUM_ERR_ADDR) in resp_read_dt0()
3201 ? OPT_MEDIUM_ERR_ADDR : (int)lba; in resp_read_dt0()
3212 switch (prot_verify_read(scp, lba, num, ei_lba)) { in resp_read_dt0()
3238 ret = do_device_access(sip, scp, 0, lba, num, false); in resp_read_dt0()
3337 static unsigned long lba_to_map_index(sector_t lba) in lba_to_map_index() argument
3340 lba += sdebug_unmap_granularity - sdebug_unmap_alignment; in lba_to_map_index()
3341 sector_div(lba, sdebug_unmap_granularity); in lba_to_map_index()
3342 return lba; in lba_to_map_index()
3347 sector_t lba = index * sdebug_unmap_granularity; in map_index_to_lba() local
3350 lba -= sdebug_unmap_granularity - sdebug_unmap_alignment; in map_index_to_lba()
3351 return lba; in map_index_to_lba()
3354 static unsigned int map_state(struct sdeb_store_info *sip, sector_t lba, in map_state() argument
3362 index = lba_to_map_index(lba); in map_state()
3371 *num = end - lba; in map_state()
3375 static void map_region(struct sdeb_store_info *sip, sector_t lba, in map_region() argument
3378 sector_t end = lba + len; in map_region()
3380 while (lba < end) { in map_region()
3381 unsigned long index = lba_to_map_index(lba); in map_region()
3386 lba = map_index_to_lba(index + 1); in map_region()
3390 static void unmap_region(struct sdeb_store_info *sip, sector_t lba, in unmap_region() argument
3393 sector_t end = lba + len; in unmap_region()
3396 while (lba < end) { in unmap_region()
3397 unsigned long index = lba_to_map_index(lba); in unmap_region()
3399 if (lba == map_index_to_lba(index) && in unmap_region()
3400 lba + sdebug_unmap_granularity <= end && in unmap_region()
3404 memset(fsp + lba * sdebug_sector_size, in unmap_region()
3410 memset(sip->dif_storep + lba, 0xff, in unmap_region()
3415 lba = map_index_to_lba(index + 1); in unmap_region()
3425 u64 lba; in resp_write_dt0() local
3433 lba = get_unaligned_be64(cmd + 2); in resp_write_dt0()
3439 lba = get_unaligned_be32(cmd + 2); in resp_write_dt0()
3445 lba = (u32)cmd[3] | (u32)cmd[2] << 8 | in resp_write_dt0()
3452 lba = get_unaligned_be32(cmd + 2); in resp_write_dt0()
3458 lba = get_unaligned_be32(cmd + 2); in resp_write_dt0()
3463 lba = get_unaligned_be64(cmd + 12); in resp_write_dt0()
3483 ret = check_device_access_params(scp, lba, num, true); in resp_write_dt0()
3491 switch (prot_verify_write(scp, lba, num, ei_lba)) { in resp_write_dt0()
3517 ret = do_device_access(sip, scp, 0, lba, num, true); in resp_write_dt0()
3519 map_region(sip, lba, num); in resp_write_dt0()
3522 zbc_inc_wp(devip, lba, num); in resp_write_dt0()
3569 u64 lba; in resp_write_scat() local
3636 lba = get_unaligned_be64(up + 0); in resp_write_scat()
3641 my_name, __func__, k, lba, num, sg_off); in resp_write_scat()
3644 ret = check_device_access_params(scp, lba, num, true); in resp_write_scat()
3663 int prot_ret = prot_verify_write(scp, lba, num, in resp_write_scat()
3674 ret = do_device_access(sip, scp, sg_off, lba, num, true); in resp_write_scat()
3677 zbc_inc_wp(devip, lba, num); in resp_write_scat()
3679 map_region(sip, lba, num); in resp_write_scat()
3719 static int resp_write_same(struct scsi_cmnd *scp, u64 lba, u32 num, in resp_write_same() argument
3736 ret = check_device_access_params(scp, lba, num, true); in resp_write_same()
3743 unmap_region(sip, lba, num); in resp_write_same()
3746 lbaa = lba; in resp_write_same()
3767 lbaa = lba + i; in resp_write_same()
3772 map_region(sip, lba, num); in resp_write_same()
3775 zbc_inc_wp(devip, lba, num); in resp_write_same()
3786 u32 lba; in resp_write_same_10() local
3798 lba = get_unaligned_be32(cmd + 2); in resp_write_same_10()
3804 return resp_write_same(scp, lba, num, ei_lba, unmap, false); in resp_write_same_10()
3811 u64 lba; in resp_write_same_16() local
3826 lba = get_unaligned_be64(cmd + 2); in resp_write_same_16()
3832 return resp_write_same(scp, lba, num, ei_lba, unmap, ndob); in resp_write_same_16()
3891 u64 lba; in resp_comp_write() local
3898 lba = get_unaligned_be64(cmd + 2); in resp_comp_write()
3912 ret = check_device_access_params(scp, lba, num, false); in resp_comp_write()
3933 if (!comp_write_worker(sip, lba, num, arr, false)) { in resp_comp_write()
3939 map_region(sip, lba, num); in resp_comp_write()
3947 __be64 lba; member
3989 unsigned long long lba = get_unaligned_be64(&desc[i].lba); in resp_unmap() local
3992 ret = check_device_access_params(scp, lba, num, true); in resp_unmap()
3996 unmap_region(sip, lba, num); in resp_unmap()
4014 u64 lba; in resp_get_lba_status() local
4019 lba = get_unaligned_be64(cmd + 2); in resp_get_lba_status()
4025 ret = check_device_access_params(scp, lba, 1, false); in resp_get_lba_status()
4032 mapped = map_state(sip, lba, &num); in resp_get_lba_status()
4037 if (sdebug_capacity - lba <= 0xffffffff) in resp_get_lba_status()
4038 num = sdebug_capacity - lba; in resp_get_lba_status()
4045 put_unaligned_be64(lba, arr + 8); /* LBA */ in resp_get_lba_status()
4056 u64 lba; in resp_sync_cache() local
4061 lba = get_unaligned_be32(cmd + 2); in resp_sync_cache()
4064 lba = get_unaligned_be64(cmd + 2); in resp_sync_cache()
4067 if (lba + num_blocks > sdebug_capacity) { in resp_sync_cache()
4089 u64 lba; in resp_pre_fetch() local
4098 lba = get_unaligned_be32(cmd + 2); in resp_pre_fetch()
4101 lba = get_unaligned_be64(cmd + 2); in resp_pre_fetch()
4104 if (lba + nblks > sdebug_capacity) { in resp_pre_fetch()
4111 block = do_div(lba, sdebug_store_sectors); in resp_pre_fetch()
4238 u64 lba; in resp_verify() local
4255 lba = get_unaligned_be64(cmd + 2); in resp_verify()
4259 lba = get_unaligned_be32(cmd + 2); in resp_verify()
4270 ret = check_device_access_params(scp, lba, a_num, false); in resp_verify()
4297 if (!comp_write_worker(sip, lba, vnum, arr, true)) { in resp_verify()
4318 u64 lba, zs_lba; in resp_report_zones() local
4356 lba = zs_lba + devip->zsize * i; in resp_report_zones()
4357 if (lba > sdebug_capacity) in resp_report_zones()
4359 zsp = zbc_zone(devip, lba); in resp_report_zones()