| /linux/fs/jfs/ |
| A D | jfs_dmap.c | 65 int nblocks); 102 int nblocks); 986 (ip, blkno + nblocks - 1, addnblocks + nblocks, results)); in dbReAlloc() 1617 u64 nblocks; in dbDiscardAG() member 1646 tt->nblocks = nblocks; in dbDiscardAG() 2165 int nblocks) in dbAllocBits() argument 3587 if (nblocks) in dbExtendFS() 3611 if (nblocks) in dbExtendFS() 3746 dp->nblocks = dp->nfree = cpu_to_le32(nblocks); in dbInitDmap() 3755 le32_add_cpu(&dp->nblocks, nblocks); in dbInitDmap() [all …]
|
| A D | jfs_dmap.h | 148 __le32 nblocks; /* 4: num blks covered by this dmap */ member 284 extern int dbFree(struct inode *ipbmap, s64 blkno, s64 nblocks); 287 int free, s64 blkno, s64 nblocks, struct tblock * tblk); 291 extern int dbAlloc(struct inode *ipbmap, s64 hint, s64 nblocks, s64 * results); 294 s64 blkno, s64 nblocks, s64 addnblocks, s64 * results); 297 extern int dbAllocBottomUp(struct inode *ip, s64 blkno, s64 nblocks); 298 extern int dbExtendFS(struct inode *ipbmap, s64 blkno, s64 nblocks);
|
| A D | jfs_discard.c | 32 void jfs_issue_discard(struct inode *ip, u64 blkno, u64 nblocks) in jfs_issue_discard() argument 37 r = sb_issue_discard(sb, blkno, nblocks, GFP_NOFS, 0); in jfs_issue_discard() 41 (unsigned long long)nblocks, r); in jfs_issue_discard() 46 (unsigned long long)nblocks, r); in jfs_issue_discard()
|
| A D | resize.c | 62 s64 XAddress, XSize, nblocks, xoff, xaddr, t64; in jfs_extendfs() local 313 nblocks = min(t64 - mapSize, XSize); in jfs_extendfs() 322 if ((rc = dbExtendFS(ipbmap, XAddress, nblocks))) in jfs_extendfs() 332 XSize -= nblocks; in jfs_extendfs() 382 xlen = min(xlen, (int) nblocks) & ~(sbi->nbperpage - 1); in jfs_extendfs() 387 if ((rc = xtAppend(tid, ipbmap, 0, xoff, nblocks, &xlen, &xaddr, 0))) { in jfs_extendfs()
|
| A D | xattr.c | 207 int nblocks; in ea_write() local 228 rc = dquot_alloc_block(ip, nblocks); in ea_write() 232 rc = dbAlloc(ip, INOHINT(ip), nblocks, &blkno); in ea_write() 235 dquot_free_block(ip, nblocks); in ea_write() 246 for (i = 0; i < nblocks; i += sbi->nbperpage) { in ea_write() 289 DXDlength(ea, nblocks); in ea_write() 300 dquot_free_block(ip, nblocks); in ea_write() 302 dbFree(ip, blkno, nblocks); in ea_write() 356 int nblocks; in ea_read() local 378 nblocks = lengthDXD(&ji->ea) << sbi->l2nbperpage; in ea_read() [all …]
|
| A D | jfs_extent.c | 496 extBalloc(struct inode *ip, s64 hint, s64 * nblocks, s64 * blkno) in extBalloc() argument 512 if (*nblocks >= max && *nblocks > nbperpage) in extBalloc() 515 nb = nblks = *nblocks; in extBalloc() 533 *nblocks = nb; in extBalloc()
|
| /linux/init/ |
| A D | do_mounts_rd.c | 66 int nblocks = -1; in identify_ramdisk_image() local 96 nblocks = 0; in identify_ramdisk_image() 164 nblocks = n; in identify_ramdisk_image() 174 return nblocks; in identify_ramdisk_image() 190 int nblocks, i; in rd_load_image() local 208 if (nblocks < 0) in rd_load_image() 211 if (nblocks == 0) { in rd_load_image() 222 if (nblocks > rd_blocks) { in rd_load_image() 224 nblocks, rd_blocks); in rd_load_image() 232 devblocks = nblocks; in rd_load_image() [all …]
|
| /linux/arch/x86/crypto/ |
| A D | sm4_aesni_avx_glue.c | 23 const u8 *src, int nblocks); 25 const u8 *src, int nblocks); 63 dst += nblocks * SM4_BLOCK_SIZE; in ecb_do_crypt() 64 src += nblocks * SM4_BLOCK_SIZE; in ecb_do_crypt() 157 src, nblocks); in sm4_avx_cbc_decrypt() 172 dst += nblocks * SM4_BLOCK_SIZE; in sm4_avx_cbc_decrypt() 261 if (nblocks > 1) in sm4_avx_cfb_decrypt() 268 keystream, nblocks); in sm4_avx_cfb_decrypt() 271 nblocks * SM4_BLOCK_SIZE); in sm4_avx_cfb_decrypt() 336 keystream, nblocks); in sm4_avx_ctr_crypt() [all …]
|
| A D | blake2s-glue.c | 22 const u8 *block, const size_t nblocks, 25 const u8 *block, const size_t nblocks, 32 const u8 *block, size_t nblocks, in blake2s_compress_arch() argument 39 blake2s_compress_generic(state, block, nblocks, inc); in blake2s_compress_arch() 44 const size_t blocks = min_t(size_t, nblocks, in blake2s_compress_arch() 55 nblocks -= blocks; in blake2s_compress_arch() 57 } while (nblocks); in blake2s_compress_arch()
|
| /linux/arch/arm/crypto/ |
| A D | blake2b-neon-glue.c | 19 const u8 *block, size_t nblocks, u32 inc); 22 const u8 *block, size_t nblocks, u32 inc) in blake2b_compress_arch() argument 25 blake2b_compress_generic(state, block, nblocks, inc); in blake2b_compress_arch() 30 const size_t blocks = min_t(size_t, nblocks, in blake2b_compress_arch() 37 nblocks -= blocks; in blake2b_compress_arch() 39 } while (nblocks); in blake2b_compress_arch()
|
| /linux/include/crypto/internal/ |
| A D | blake2s.h | 15 size_t nblocks, const u32 inc); 18 size_t nblocks, const u32 inc); 28 const u8 *block, size_t nblocks, u32 inc); 48 const size_t nblocks = DIV_ROUND_UP(inlen, BLAKE2S_BLOCK_SIZE); in __blake2s_update() local 50 (*compress)(state, in, nblocks - 1, BLAKE2S_BLOCK_SIZE); in __blake2s_update() 51 in += BLAKE2S_BLOCK_SIZE * (nblocks - 1); in __blake2s_update() 52 inlen -= BLAKE2S_BLOCK_SIZE * (nblocks - 1); in __blake2s_update()
|
| A D | blake2b.h | 15 const u8 *block, size_t nblocks, u32 inc); 23 const u8 *block, size_t nblocks, u32 inc); 41 const size_t nblocks = DIV_ROUND_UP(inlen, BLAKE2B_BLOCK_SIZE); in __blake2b_update() local 43 (*compress)(state, in, nblocks - 1, BLAKE2B_BLOCK_SIZE); in __blake2b_update() 44 in += BLAKE2B_BLOCK_SIZE * (nblocks - 1); in __blake2b_update() 45 inlen -= BLAKE2B_BLOCK_SIZE * (nblocks - 1); in __blake2b_update()
|
| /linux/fs/nilfs2/ |
| A D | sufile.c | 1022 sector_t start = 0, nblocks = 0; in nilfs_sufile_trim_fs() local 1078 if (!nblocks) { in nilfs_sufile_trim_fs() 1093 nblocks -= start_block - start; in nilfs_sufile_trim_fs() 1097 if (nblocks >= minlen) { in nilfs_sufile_trim_fs() 1102 nblocks * sects_per_block, in nilfs_sufile_trim_fs() 1109 ndiscarded += nblocks; in nilfs_sufile_trim_fs() 1124 if (nblocks) { in nilfs_sufile_trim_fs() 1127 nblocks -= start_block - start; in nilfs_sufile_trim_fs() 1133 if (nblocks >= minlen) { in nilfs_sufile_trim_fs() 1136 nblocks * sects_per_block, in nilfs_sufile_trim_fs() [all …]
|
| A D | segbuf.h | 34 unsigned long nblocks; member 131 return segbuf->sb_sum.nblocks == segbuf->sb_sum.nsumblk; in nilfs_segbuf_empty() 139 segbuf->sb_sum.nblocks++; in nilfs_segbuf_add_segsum_buffer() 148 segbuf->sb_sum.nblocks++; in nilfs_segbuf_add_payload_buffer()
|
| A D | the_nilfs.c | 656 sector_t start = 0, nblocks = 0; in nilfs_discard_segments() local 666 if (!nblocks) { in nilfs_discard_segments() 668 nblocks = seg_end - seg_start + 1; in nilfs_discard_segments() 669 } else if (start + nblocks == seg_start) { in nilfs_discard_segments() 670 nblocks += seg_end - seg_start + 1; in nilfs_discard_segments() 674 nblocks * sects_per_block, in nilfs_discard_segments() 678 nblocks = 0; in nilfs_discard_segments() 681 if (nblocks) in nilfs_discard_segments() 684 nblocks * sects_per_block, in nilfs_discard_segments() 689 int nilfs_count_free_blocks(struct the_nilfs *nilfs, sector_t *nblocks) in nilfs_count_free_blocks() argument [all …]
|
| /linux/fs/ext4/ |
| A D | ext4_jbd2.h | 305 #define ext4_journal_start_sb(sb, type, nblocks) \ argument 306 __ext4_journal_start_sb((sb), __LINE__, (type), (nblocks), 0, \ 309 #define ext4_journal_start(inode, type, nblocks) \ argument 310 __ext4_journal_start((inode), __LINE__, (type), (nblocks), 0, \ 344 static inline int ext4_journal_extend(handle_t *handle, int nblocks, int revoke) in ext4_journal_extend() argument 347 return jbd2_journal_extend(handle, nblocks, revoke); in ext4_journal_extend() 351 static inline int ext4_journal_restart(handle_t *handle, int nblocks, in ext4_journal_restart() argument 355 return jbd2__journal_restart(handle, nblocks, revoke, GFP_NOFS); in ext4_journal_restart()
|
| /linux/fs/reiserfs/ |
| A D | xattr.h | 75 size_t nblocks = JOURNAL_BLOCKS_PER_OBJECT(inode->i_sb); in reiserfs_xattr_jcreate_nblocks() local 78 nblocks += JOURNAL_BLOCKS_PER_OBJECT(inode->i_sb); in reiserfs_xattr_jcreate_nblocks() 80 nblocks += JOURNAL_BLOCKS_PER_OBJECT(inode->i_sb); in reiserfs_xattr_jcreate_nblocks() 83 return nblocks; in reiserfs_xattr_jcreate_nblocks()
|
| A D | xattr_acl.c | 369 int nblocks = 0; in reiserfs_cache_default_acl() local 383 nblocks = reiserfs_xattr_jcreate_nblocks(inode); in reiserfs_cache_default_acl() 384 nblocks += JOURNAL_BLOCKS_PER_OBJECT(inode->i_sb); in reiserfs_cache_default_acl() 389 nblocks += reiserfs_xattr_nblocks(inode, size) * 4; in reiserfs_cache_default_acl() 393 return nblocks; in reiserfs_cache_default_acl()
|
| /linux/lib/crypto/ |
| A D | blake2s-generic.c | 41 size_t nblocks, const u32 inc) in blake2s_compress_generic() argument 48 (nblocks > 1 && inc != BLAKE2S_BLOCK_SIZE)); in blake2s_compress_generic() 50 while (nblocks > 0) { in blake2s_compress_generic() 103 --nblocks; in blake2s_compress_generic()
|
| /linux/block/partitions/ |
| A D | sysv68.c | 44 __be32 nblocks; /* slice size (in blocks) */ member 83 if (be32_to_cpu(slice->nblocks)) { in sysv68_partition() 86 be32_to_cpu(slice->nblocks)); in sysv68_partition()
|
| /linux/fs/ |
| A D | mpage.c | 174 unsigned nblocks; in do_mpage_readpage() local 199 nblocks = map_bh->b_size >> blkbits; in do_mpage_readpage() 202 block_in_file < (args->first_logical_block + nblocks)) { in do_mpage_readpage() 204 unsigned last = nblocks - map_offset; in do_mpage_readpage() 262 nblocks = map_bh->b_size >> blkbits; in do_mpage_readpage() 264 if (relative_block == nblocks) { in do_mpage_readpage() 319 nblocks = map_bh->b_size >> blkbits; in do_mpage_readpage() 320 if ((buffer_boundary(map_bh) && relative_block == nblocks) || in do_mpage_readpage()
|
| /linux/fs/jbd2/ |
| A D | transaction.c | 479 static handle_t *new_handle(int nblocks) in new_handle() argument 484 handle->h_total_credits = nblocks; in new_handle() 506 nblocks += DIV_ROUND_UP(revoke_records, in jbd2__journal_start() 508 handle = new_handle(nblocks); in jbd2__journal_start() 536 line_no, nblocks); in jbd2__journal_start() 687 nblocks += DIV_ROUND_UP( in jbd2_journal_extend() 694 wanted = atomic_add_return(nblocks, in jbd2_journal_extend() 708 nblocks); in jbd2_journal_extend() 710 handle->h_total_credits += nblocks; in jbd2_journal_extend() 711 handle->h_requested_credits += nblocks; in jbd2_journal_extend() [all …]
|
| /linux/drivers/media/usb/pwc/ |
| A D | pwc-dec23.c | 594 int compression_index, nblocks; in DecompressBand23() local 605 nblocks = compressed_image_width / 4; in DecompressBand23() 611 while (nblocks) { in DecompressBand23() 615 nblocks--; in DecompressBand23() 619 nblocks = compressed_image_width / 8; in DecompressBand23() 625 while (nblocks) { in DecompressBand23() 634 nblocks -= 2; in DecompressBand23()
|
| /linux/drivers/s390/char/ |
| A D | tape_char.c | 193 int nblocks; in tapechar_write() local 206 nblocks = count / block_size; in tapechar_write() 209 nblocks = 1; in tapechar_write() 217 DBF_EVENT(6, "TCHAR:nblocks: %x\n", nblocks); in tapechar_write() 224 for (i = 0; i < nblocks; i++) { in tapechar_write()
|
| /linux/drivers/misc/ |
| A D | sram.c | 179 unsigned int nblocks, exports = 0; in sram_reserve_regions() local 191 nblocks = (np) ? of_get_available_child_count(np) + 1 : 1; in sram_reserve_regions() 192 rblocks = kcalloc(nblocks, sizeof(*rblocks), GFP_KERNEL); in sram_reserve_regions() 265 rblocks[nblocks - 1].start = size; in sram_reserve_regions() 266 rblocks[nblocks - 1].size = 0; in sram_reserve_regions() 267 list_add_tail(&rblocks[nblocks - 1].list, &reserve_list); in sram_reserve_regions()
|