/linux-4.1.27/init/ |
H A D | do_mounts_rd.c | 74 int nblocks = -1; identify_ramdisk_image() local 103 nblocks = 0; identify_ramdisk_image() 113 nblocks = (ntohl(romfsb->size)+BLOCK_SIZE-1)>>BLOCK_SIZE_BITS; identify_ramdisk_image() 121 nblocks = (cramfsb->size + BLOCK_SIZE - 1) >> BLOCK_SIZE_BITS; identify_ramdisk_image() 130 nblocks = (le64_to_cpu(squashfsb->bytes_used) + BLOCK_SIZE - 1) identify_ramdisk_image() 145 nblocks = (cramfsb->size + BLOCK_SIZE - 1) >> BLOCK_SIZE_BITS; identify_ramdisk_image() 161 nblocks = minixsb->s_nzones << minixsb->s_log_zone_size; identify_ramdisk_image() 171 nblocks = n; identify_ramdisk_image() 182 return nblocks; identify_ramdisk_image() 190 int nblocks, i, disk; rd_load_image() local 206 nblocks = identify_ramdisk_image(in_fd, rd_image_start, &decompressor); rd_load_image() 207 if (nblocks < 0) rd_load_image() 210 if (nblocks == 0) { rd_load_image() 217 * NOTE NOTE: nblocks is not actually blocks but rd_load_image() 232 if (nblocks > rd_blocks) { rd_load_image() 234 nblocks, rd_blocks); rd_load_image() 247 devblocks = nblocks; rd_load_image() 261 nblocks, ((nblocks-1)/devblocks)+1, nblocks>devblocks ? "s" : ""); rd_load_image() 262 for (i = 0, disk = 1; i < nblocks; i++) { rd_load_image()
|
/linux-4.1.27/kernel/ |
H A D | groups.c | 15 int nblocks; groups_alloc() local 18 nblocks = (gidsetsize + NGROUPS_PER_BLOCK - 1) / NGROUPS_PER_BLOCK; groups_alloc() 20 nblocks = nblocks ? : 1; groups_alloc() 21 group_info = kmalloc(sizeof(*group_info) + nblocks*sizeof(gid_t *), GFP_USER); groups_alloc() 25 group_info->nblocks = nblocks; groups_alloc() 31 for (i = 0; i < nblocks; i++) { groups_alloc() 55 for (i = 0; i < group_info->nblocks; i++) groups_free()
|
/linux-4.1.27/block/partitions/ |
H A D | sysv68.c | 44 __be32 nblocks; /* slice size (in blocks) */ member in struct:slice 83 if (be32_to_cpu(slice->nblocks)) { sysv68_partition() 86 be32_to_cpu(slice->nblocks)); sysv68_partition()
|
/linux-4.1.27/fs/jfs/ |
H A D | jfs_discard.c | 38 * nblocks - number of blocks to be trimmed 45 void jfs_issue_discard(struct inode *ip, u64 blkno, u64 nblocks) jfs_issue_discard() argument 50 r = sb_issue_discard(sb, blkno, nblocks, GFP_NOFS, 0); jfs_issue_discard() 55 (unsigned long long)nblocks, r); jfs_issue_discard() 61 (unsigned long long)nblocks, r); jfs_issue_discard()
|
H A D | jfs_dmap.c | 78 int nblocks); 85 static int dbAllocAny(struct bmap * bmp, s64 nblocks, int l2nb, s64 * results); 87 int nblocks); 89 int nblocks, 92 int nblocks); 93 static int dbAllocDmapLev(struct bmap * bmp, struct dmap * dp, int nblocks, 96 static int dbAllocAG(struct bmap * bmp, int agno, s64 nblocks, int l2nb, 98 static int dbAllocCtl(struct bmap * bmp, s64 nblocks, int l2nb, s64 blkno, 100 static int dbExtend(struct inode *ip, s64 blkno, s64 nblocks, s64 addnblocks); 105 int nblocks); 107 int nblocks); 115 int nblocks); 116 static int dbInitDmap(struct dmap * dp, s64 blkno, int nblocks); 120 static int dbGetL2AGSize(s64 nblocks); 325 * nblocks - number of blocks to be freed. 331 int dbFree(struct inode *ip, s64 blkno, s64 nblocks) dbFree() argument 344 if (unlikely((blkno == 0) || (blkno + nblocks > bmp->db_mapsize))) { dbFree() 346 printk(KERN_ERR "blkno = %Lx, nblocks = %Lx\n", dbFree() 348 (unsigned long long) nblocks); dbFree() 357 if (JFS_SBI(sb)->minblks_trim <= nblocks) dbFree() 358 jfs_issue_discard(ipbmap, blkno, nblocks); dbFree() 364 for (rem = nblocks; rem > 0; rem -= nb, blkno += nb) { dbFree() 416 * nblocks - number of contiguous blocks in the range. 425 int free, s64 blkno, s64 nblocks, struct tblock * tblk) dbUpdatePMap() 439 if (blkno + nblocks > bmp->db_mapsize) { dbUpdatePMap() 440 printk(KERN_ERR "blkno = %Lx, nblocks = %Lx\n", dbUpdatePMap() 442 (unsigned long long) nblocks); dbUpdatePMap() 457 for (rem = nblocks; rem > 0; rem -= nblks, blkno += nblks) { dbUpdatePMap() 691 * nblocks - number of contiguous blocks in the range. 700 int dbAlloc(struct inode *ip, s64 hint, s64 nblocks, s64 * results) dbAlloc() argument 712 /* assert that nblocks is valid */ dbAlloc() 713 assert(nblocks > 0); dbAlloc() 719 l2nb = BLKSTOL2(nblocks); dbAlloc() 737 rc = dbAllocAny(bmp, nblocks, l2nb, results); dbAlloc() 775 if (nblocks <= BPERDMAP) { dbAlloc() 791 if ((rc = dbAllocNext(bmp, dp, blkno, (int) nblocks)) dbAlloc() 818 dbAllocNear(bmp, dp, blkno, (int) nblocks, l2nb, results)) dbAlloc() 830 if ((rc = dbAllocDmapLev(bmp, dp, (int) nblocks, l2nb, results)) dbAlloc() 847 if ((rc = dbAllocAG(bmp, agno, nblocks, l2nb, results)) != -ENOSPC) dbAlloc() 863 if ((rc = dbAllocAG(bmp, agno, nblocks, l2nb, results)) == -ENOSPC) dbAlloc() 864 rc = dbAllocAny(bmp, nblocks, l2nb, results); dbAlloc() 886 * nblocks - extent length; 893 int dbAllocExact(struct inode *ip, s64 blkno, int nblocks) dbAllocExact() argument 911 if (nblocks <= 0 || nblocks > BPERDMAP || blkno >= bmp->db_mapsize) { dbAllocExact() 916 if (nblocks > ((s64) 1 << bmp->db_maxfreebud)) { dbAllocExact() 932 rc = dbAllocNext(bmp, dp, blkno, nblocks); dbAllocExact() 963 * nblocks - number of contiguous blocks within the current 979 s64 blkno, s64 nblocks, s64 addnblocks, s64 * results) dbReAlloc() 985 if ((rc = dbExtend(ip, blkno, nblocks, addnblocks)) == 0) { dbReAlloc() 999 (ip, blkno + nblocks - 1, addnblocks + nblocks, results)); dbReAlloc() 1017 * nblocks - number of contiguous blocks within the current 1026 static int dbExtend(struct inode *ip, s64 blkno, s64 nblocks, s64 addnblocks) dbExtend() argument 1041 (rel_block + nblocks + addnblocks > sbi->nbperpage)) dbExtend() 1045 lastblkno = blkno + nblocks - 1; dbExtend() 1116 * nblocks - number of contiguous free blocks of the range. 1126 int nblocks) dbAllocNext() 1151 if (dbitno + nblocks > BPERDMAP) dbAllocNext() 1173 for (rembits = nblocks; rembits > 0; rembits -= nb, dbitno += nb) { dbAllocNext() 1225 return (dbAllocDmap(bmp, dp, blkno, nblocks)); dbAllocNext() 1244 * nblocks - actual number of contiguous free blocks desired. 1258 struct dmap * dp, s64 blkno, int nblocks, int l2nb, s64 * results) dbAllocNear() 1301 if ((rc = dbAllocDmap(bmp, dp, blkno, nblocks)) == 0) dbAllocNear() 1353 * nblocks - actual number of contiguous free blocks desired. 1366 dbAllocAG(struct bmap * bmp, int agno, s64 nblocks, int l2nb, s64 * results) dbAllocAG() argument 1407 rc = dbAllocCtl(bmp, nblocks, l2nb, blkno, results); dbAllocAG() 1412 (unsigned long long) nblocks); dbAllocAG() 1520 rc = dbAllocCtl(bmp, nblocks, l2nb, blkno, results); dbAllocAG() 1552 * nblocks - actual number of contiguous free blocks desired. 1564 static int dbAllocAny(struct bmap * bmp, s64 nblocks, int l2nb, s64 * results) dbAllocAny() argument 1580 rc = dbAllocCtl(bmp, nblocks, l2nb, blkno, results); dbAllocAny() 1623 s64 nblocks, blkno; dbDiscardAG() local 1630 u64 nblocks; dbDiscardAG() member in struct:range2trim 1633 /* max blkno / nblocks pairs to trim */ dbDiscardAG() 1640 nblocks = bmp->db_agfree[agno]; dbDiscardAG() 1641 max_ranges = nblocks; dbDiscardAG() 1652 while (nblocks >= minlen) { dbDiscardAG() 1653 l2nb = BLKSTOL2(nblocks); dbDiscardAG() 1656 rc = dbAllocAG(bmp, agno, nblocks, l2nb, &blkno); dbDiscardAG() 1659 tt->nblocks = nblocks; dbDiscardAG() 1667 nblocks = bmp->db_agfree[agno]; dbDiscardAG() 1671 l2nb = BLKSTOL2(nblocks) - 1; dbDiscardAG() 1672 nblocks = 1 << l2nb; dbDiscardAG() 1685 tt->nblocks = 0; /* mark the current end */ dbDiscardAG() 1686 for (tt = totrim; tt->nblocks != 0; tt++) { dbDiscardAG() 1690 jfs_issue_discard(ip, tt->blkno, tt->nblocks); dbDiscardAG() 1691 dbFree(ip, tt->blkno, tt->nblocks); dbDiscardAG() 1692 trimmed += tt->nblocks; dbDiscardAG() 1831 * nblocks - actual number of contiguous free blocks to allocate. 1846 dbAllocCtl(struct bmap * bmp, s64 nblocks, int l2nb, s64 blkno, s64 * results) dbAllocCtl() argument 1866 rc = dbAllocDmapLev(bmp, dp, (int) nblocks, l2nb, results); dbAllocCtl() 1882 for (n = nblocks, b = blkno; n > 0; n -= nb, b += nb) { dbAllocCtl() 1934 for (n = nblocks - n, b = blkno; n > 0; dbAllocCtl() 1977 * if so, nblocks of blocks are allocated; otherwise, ENOSPC is 1984 * nblocks - actual number of contiguous block desired. 1998 struct dmap * dp, int nblocks, int l2nb, s64 * results) dbAllocDmapLev() 2026 if ((rc = dbAllocDmap(bmp, dp, blkno, nblocks)) == 0) dbAllocDmapLev() 2052 * nblocks - number of blocks to be allocated. 2061 int nblocks) dbAllocDmap() 2072 dbAllocBits(bmp, dp, blkno, nblocks); dbAllocDmap() 2083 dbFreeBits(bmp, dp, blkno, nblocks); dbAllocDmap() 2107 * nblocks - number of blocks to be freed. 2116 int nblocks) dbFreeDmap() 2127 rc = dbFreeBits(bmp, dp, blkno, nblocks); dbFreeDmap() 2148 dbAllocBits(bmp, dp, blkno, nblocks); dbFreeDmap() 2171 * nblocks - number of bits to be allocated. 2178 int nblocks) dbAllocBits() 2195 assert(dbitno + nblocks <= BPERDMAP); dbAllocBits() 2210 for (rembits = nblocks; rembits > 0; rembits -= nb, dbitno += nb) { dbAllocBits() 2282 le32_add_cpu(&dp->nfree, -nblocks); dbAllocBits() 2295 bmp->db_agfree[agno] -= nblocks; dbAllocBits() 2296 bmp->db_nfree -= nblocks; dbAllocBits() 2318 * nblocks - number of bits to be freed. 2325 int nblocks) dbFreeBits() 2340 assert(dbitno + nblocks <= BPERDMAP); dbFreeBits() 2359 for (rembits = nblocks; rembits > 0; rembits -= nb, dbitno += nb) { dbFreeBits() 2426 le32_add_cpu(&dp->nfree, nblocks); dbFreeBits() 2434 bmp->db_nfree += nblocks; dbFreeBits() 2435 bmp->db_agfree[agno] += nblocks; dbFreeBits() 3202 * nblocks - number of blocks to be freed. 3208 int dbAllocBottomUp(struct inode *ip, s64 blkno, s64 nblocks) dbAllocBottomUp() argument 3220 ASSERT(nblocks <= bmp->db_mapsize - blkno); dbAllocBottomUp() 3226 for (rem = nblocks; rem > 0; rem -= nb, blkno += nb) { dbAllocBottomUp() 3264 int nblocks) dbAllocDmapBU() 3283 assert(dbitno + nblocks <= BPERDMAP); dbAllocDmapBU() 3298 for (rembits = nblocks; rembits > 0; rembits -= nb, dbitno += nb) { dbAllocDmapBU() 3331 le32_add_cpu(&dp->nfree, -nblocks); dbAllocDmapBU() 3347 bmp->db_agfree[agno] -= nblocks; dbAllocDmapBU() 3348 bmp->db_nfree -= nblocks; dbAllocDmapBU() 3361 dbFreeBits(bmp, dp, blkno, nblocks); dbAllocDmapBU() 3370 * FUNCTION: extend bmap from blkno for nblocks; 3384 int dbExtendFS(struct inode *ipbmap, s64 blkno, s64 nblocks) dbExtendFS() argument 3399 newsize = blkno + nblocks; dbExtendFS() 3401 jfs_info("dbExtendFS: blkno:%Ld nblocks:%Ld newsize:%Ld", dbExtendFS() 3402 (long long) blkno, (long long) nblocks, (long long) newsize); dbExtendFS() 3559 n = min(nblocks, (s64)BPERDMAP - n); dbExtendFS() 3567 n = min_t(s64, nblocks, BPERDMAP); dbExtendFS() 3583 nblocks -= n; dbExtendFS() 3584 if (nblocks == 0) dbExtendFS() 3596 if (nblocks) dbExtendFS() 3620 if (nblocks) dbExtendFS() 3743 * nblocks - number of blocks this page 3747 static int dbInitDmap(struct dmap * dp, s64 Blkno, int nblocks) dbInitDmap() argument 3755 dp->nblocks = dp->nfree = cpu_to_le32(nblocks); dbInitDmap() 3758 if (nblocks == BPERDMAP) { dbInitDmap() 3764 le32_add_cpu(&dp->nblocks, nblocks); dbInitDmap() 3765 le32_add_cpu(&dp->nfree, nblocks); dbInitDmap() 3776 for (r = nblocks; r > 0; r -= nb, blkno += nb) { dbInitDmap() 3815 /* does nblocks fall on a 32-bit boundary ? */ dbInitDmap() 4000 * nblocks - Number of blocks in aggregate 4004 static int dbGetL2AGSize(s64 nblocks) dbGetL2AGSize() argument 4010 if (nblocks < BPERDMAP * MAXAG) dbGetL2AGSize() 4016 if (m & nblocks) dbGetL2AGSize() 4021 if (sz < nblocks) dbGetL2AGSize() 4055 s64 nblocks; dbMapFileSizeToMapSize() local 4060 nblocks = ipbmap->i_size >> JFS_SBI(sb)->l2bsize; dbMapFileSizeToMapSize() 4061 npages = nblocks >> JFS_SBI(sb)->l2nbperpage; dbMapFileSizeToMapSize() 4089 nblocks = ndmaps << L2BPERDMAP; dbMapFileSizeToMapSize() 4091 return (nblocks); dbMapFileSizeToMapSize() 424 dbUpdatePMap(struct inode *ipbmap, int free, s64 blkno, s64 nblocks, struct tblock * tblk) dbUpdatePMap() argument 978 dbReAlloc(struct inode *ip, s64 blkno, s64 nblocks, s64 addnblocks, s64 * results) dbReAlloc() argument 1125 dbAllocNext(struct bmap * bmp, struct dmap * dp, s64 blkno, int nblocks) dbAllocNext() argument 1257 dbAllocNear(struct bmap * bmp, struct dmap * dp, s64 blkno, int nblocks, int l2nb, s64 * results) dbAllocNear() argument 1997 dbAllocDmapLev(struct bmap * bmp, struct dmap * dp, int nblocks, int l2nb, s64 * results) dbAllocDmapLev() argument 2060 dbAllocDmap(struct bmap * bmp, struct dmap * dp, s64 blkno, int nblocks) dbAllocDmap() argument 2115 dbFreeDmap(struct bmap * bmp, struct dmap * dp, s64 blkno, int nblocks) dbFreeDmap() argument 2177 dbAllocBits(struct bmap * bmp, struct dmap * dp, s64 blkno, int nblocks) dbAllocBits() argument 2324 dbFreeBits(struct bmap * bmp, struct dmap * dp, s64 blkno, int nblocks) dbFreeBits() argument 3263 dbAllocDmapBU(struct bmap * bmp, struct dmap * dp, s64 blkno, int nblocks) dbAllocDmapBU() argument
|
H A D | jfs_discard.h | 23 extern void jfs_issue_discard(struct inode *ip, u64 blkno, u64 nblocks);
|
H A D | resize.c | 74 s64 XAddress, XSize, nblocks, xoff, xaddr, t64; jfs_extendfs() local 326 nblocks = min(t64 - mapSize, XSize); jfs_extendfs() 335 if ((rc = dbExtendFS(ipbmap, XAddress, nblocks))) jfs_extendfs() 341 * the map now has extended to cover additional nblocks: jfs_extendfs() 342 * dn_mapsize = oldMapsize + nblocks; jfs_extendfs() 344 /* ipbmap->i_mapsize += nblocks; */ jfs_extendfs() 345 XSize -= nblocks; jfs_extendfs() 389 xlen = min(xlen, (int) nblocks) & ~(sbi->nbperpage - 1); jfs_extendfs() 394 if ((rc = xtAppend(tid, ipbmap, 0, xoff, nblocks, &xlen, &xaddr, 0))) { jfs_extendfs()
|
H A D | jfs_dmap.h | 161 __le32 nblocks; /* 4: num blks covered by this dmap */ member in struct:dmap 297 extern int dbFree(struct inode *ipbmap, s64 blkno, s64 nblocks); 300 int free, s64 blkno, s64 nblocks, struct tblock * tblk); 304 extern int dbAlloc(struct inode *ipbmap, s64 hint, s64 nblocks, s64 * results); 307 s64 blkno, s64 nblocks, s64 addnblocks, s64 * results); 310 extern int dbAllocBottomUp(struct inode *ip, s64 blkno, s64 nblocks); 311 extern int dbExtendFS(struct inode *ipbmap, s64 blkno, s64 nblocks);
|
H A D | xattr.c | 222 int nblocks; ea_write() local 240 nblocks = (size + (sb->s_blocksize - 1)) >> sb->s_blocksize_bits; ea_write() 243 rc = dquot_alloc_block(ip, nblocks); ea_write() 247 rc = dbAlloc(ip, INOHINT(ip), nblocks, &blkno); ea_write() 250 dquot_free_block(ip, nblocks); ea_write() 255 * Now have nblocks worth of storage to stuff into the FEALIST. ea_write() 261 for (i = 0; i < nblocks; i += sbi->nbperpage) { ea_write() 304 DXDlength(ea, nblocks); ea_write() 315 dquot_free_block(ip, nblocks); ea_write() 317 dbFree(ip, blkno, nblocks); ea_write() 371 int nblocks; ea_read() local 393 nblocks = lengthDXD(&ji->ea) << sbi->l2nbperpage; ea_read() 401 for (i = 0; i < nblocks; i += sbi->nbperpage) { ea_read()
|
H A D | jfs_extent.c | 483 * requested size (nblocks). if this fails (nblocks 495 * *nblocks - pointer to an s64 value. on entry, this value specifies 509 extBalloc(struct inode *ip, s64 hint, s64 * nblocks, s64 * blkno) extBalloc() argument 525 if (*nblocks >= max && *nblocks > nbperpage) extBalloc() 528 nb = nblks = *nblocks; extBalloc() 546 *nblocks = nb; extBalloc()
|
H A D | jfs_xtree.c | 2186 int nsplit, nblocks, xlen; xtAppend() local 2240 nblocks = JFS_SBI(ip->i_sb)->nbperpage; xtAppend() 2241 for (; nsplit > 0; nsplit--, pxd++, xaddr += nblocks, maxblocks -= nblocks) { xtAppend() 2242 if ((rc = dbAllocBottomUp(ip, xaddr, (s64) nblocks)) == 0) { xtAppend() 2244 PXDlength(pxd, nblocks); xtAppend()
|
/linux-4.1.27/crypto/ |
H A D | lrw.c | 225 unsigned int nblocks; lrw_crypt() local 239 nblocks = min(walk.nbytes / bsize, max_blks); lrw_crypt() 255 for (i = 0; i < nblocks; i++) { lrw_crypt() 270 nblocks * bsize); lrw_crypt() 273 for (i = 0; i < nblocks; i++) lrw_crypt() 276 src += nblocks; lrw_crypt() 277 dst += nblocks; lrw_crypt() 278 nbytes -= nblocks * bsize; lrw_crypt() 279 nblocks = min(nbytes / bsize, max_blks); lrw_crypt() 280 } while (nblocks > 0); lrw_crypt() 287 nblocks = min(nbytes / bsize, max_blks); lrw_crypt()
|
H A D | xts.c | 176 unsigned int nblocks; xts_crypt() local 190 nblocks = min(nbytes / bsize, max_blks); xts_crypt() 202 for (i = 0; i < nblocks; i++) { xts_crypt() 213 nblocks * bsize); xts_crypt() 216 for (i = 0; i < nblocks; i++) xts_crypt() 219 src += nblocks; xts_crypt() 220 dst += nblocks; xts_crypt() 221 nbytes -= nblocks * bsize; xts_crypt() 222 nblocks = min(nbytes / bsize, max_blks); xts_crypt() 223 } while (nblocks > 0); xts_crypt() 232 nblocks = min(nbytes / bsize, max_blks); xts_crypt()
|
H A D | tgr192.c | 41 u32 nblocks; member in struct:tgr192_ctx 505 tctx->nblocks = 0; tgr192_init() 522 tctx->nblocks++; tgr192_update() 541 tctx->nblocks++; tgr192_update() 566 t = tctx->nblocks; tgr192_final()
|
/linux-4.1.27/fs/reiserfs/ |
H A D | xattr.h | 77 size_t nblocks = JOURNAL_BLOCKS_PER_OBJECT(inode->i_sb); reiserfs_xattr_jcreate_nblocks() local 80 nblocks += JOURNAL_BLOCKS_PER_OBJECT(inode->i_sb); reiserfs_xattr_jcreate_nblocks() 82 nblocks += JOURNAL_BLOCKS_PER_OBJECT(inode->i_sb); reiserfs_xattr_jcreate_nblocks() 85 return nblocks; reiserfs_xattr_jcreate_nblocks()
|
H A D | xattr_acl.c | 368 int nblocks = 0; reiserfs_cache_default_acl() local 382 nblocks = reiserfs_xattr_jcreate_nblocks(inode); reiserfs_cache_default_acl() 383 nblocks += JOURNAL_BLOCKS_PER_OBJECT(inode->i_sb); reiserfs_cache_default_acl() 388 nblocks += reiserfs_xattr_nblocks(inode, size) * 4; reiserfs_cache_default_acl() 392 return nblocks; reiserfs_cache_default_acl()
|
H A D | journal.c | 112 unsigned long nblocks, int join); 3018 * blocks you expect to use in nblocks. 3021 struct super_block *sb, unsigned long nblocks, do_journal_begin_r() 3033 BUG_ON(nblocks > journal->j_trans_max); do_journal_begin_r() 3068 && (journal->j_len_alloc + nblocks + 2) >= journal->j_max_batch) do_journal_begin_r() 3080 if (!join && (journal->j_len_alloc + nblocks + 2) >= do_journal_begin_r() 3082 ((journal->j_len + nblocks + 2) * 100) < do_journal_begin_r() 3123 journal->j_len_alloc += nblocks; do_journal_begin_r() 3125 th->t_blocks_allocated = nblocks; do_journal_begin_r() 3145 int nblocks) reiserfs_persistent_transaction() 3164 ret = journal_begin(th, s, nblocks); reiserfs_persistent_transaction() 3218 struct super_block *sb, unsigned long nblocks) journal_begin() 3249 ret = do_journal_begin_r(th, sb, nblocks, JBEGIN_REG); journal_begin() 3020 do_journal_begin_r(struct reiserfs_transaction_handle *th, struct super_block *sb, unsigned long nblocks, int join) do_journal_begin_r() argument 3142 reiserfs_persistent_transaction(struct super_block *s, int nblocks) reiserfs_persistent_transaction() argument 3217 journal_begin(struct reiserfs_transaction_handle *th, struct super_block *sb, unsigned long nblocks) journal_begin() argument
|
/linux-4.1.27/fs/nilfs2/ |
H A D | segbuf.h | 35 * @nblocks: Number of blocks included in the partial segment 47 unsigned long nblocks; member in struct:nilfs_segsum_info 143 return segbuf->sb_sum.nblocks == segbuf->sb_sum.nsumblk; nilfs_segbuf_empty() 151 segbuf->sb_sum.nblocks++; nilfs_segbuf_add_segsum_buffer() 160 segbuf->sb_sum.nblocks++; nilfs_segbuf_add_payload_buffer()
|
H A D | sufile.c | 518 * @nblocks: number of live blocks in the segment 522 unsigned long nblocks, time_t modtime) nilfs_sufile_set_segment_usage() 539 su->su_nblocks = cpu_to_le32(nblocks); nilfs_sufile_set_segment_usage() 890 * %-EINVAL - Invalid values in input (segment number, flags or nblocks) 1027 sector_t start = 0, nblocks = 0; nilfs_sufile_trim_fs() local 1083 if (!nblocks) { nilfs_sufile_trim_fs() 1086 nblocks = seg_end - seg_start + 1; nilfs_sufile_trim_fs() 1090 if (start + nblocks == seg_start) { nilfs_sufile_trim_fs() 1092 nblocks += seg_end - seg_start + 1; nilfs_sufile_trim_fs() 1098 nblocks -= start_block - start; nilfs_sufile_trim_fs() 1102 if (nblocks >= minlen) { nilfs_sufile_trim_fs() 1107 nblocks * sects_per_block, nilfs_sufile_trim_fs() 1114 ndiscarded += nblocks; nilfs_sufile_trim_fs() 1122 nblocks = seg_end - seg_start + 1; nilfs_sufile_trim_fs() 1129 if (nblocks) { nilfs_sufile_trim_fs() 1132 nblocks -= start_block - start; nilfs_sufile_trim_fs() 1135 if (start + nblocks > end_block + 1) nilfs_sufile_trim_fs() 1136 nblocks = end_block - start + 1; nilfs_sufile_trim_fs() 1138 if (nblocks >= minlen) { nilfs_sufile_trim_fs() 1141 nblocks * sects_per_block, nilfs_sufile_trim_fs() 1144 ndiscarded += nblocks; nilfs_sufile_trim_fs() 521 nilfs_sufile_set_segment_usage(struct inode *sufile, __u64 segnum, unsigned long nblocks, time_t modtime) nilfs_sufile_set_segment_usage() argument
|
H A D | the_nilfs.c | 664 sector_t start = 0, nblocks = 0; nilfs_discard_segments() local 674 if (!nblocks) { nilfs_discard_segments() 676 nblocks = seg_end - seg_start + 1; nilfs_discard_segments() 677 } else if (start + nblocks == seg_start) { nilfs_discard_segments() 678 nblocks += seg_end - seg_start + 1; nilfs_discard_segments() 682 nblocks * sects_per_block, nilfs_discard_segments() 686 nblocks = 0; nilfs_discard_segments() 689 if (nblocks) nilfs_discard_segments() 692 nblocks * sects_per_block, nilfs_discard_segments() 697 int nilfs_count_free_blocks(struct the_nilfs *nilfs, sector_t *nblocks) nilfs_count_free_blocks() argument 704 *nblocks = (sector_t)ncleansegs * nilfs->ns_blocks_per_segment; nilfs_count_free_blocks()
|
H A D | recovery.c | 339 unsigned long nblocks, ndatablk, nnodeblk; nilfs_scan_dsync_log() local 348 nblocks = le32_to_cpu(finfo->fi_nblocks); nilfs_scan_dsync_log() 350 nnodeblk = nblocks - ndatablk; nilfs_scan_dsync_log() 821 unsigned long nblocks; nilfs_search_super_root() local 857 nblocks = le32_to_cpu(sum->ss_nblocks); nilfs_search_super_root() 858 pseg_end = pseg_start + nblocks - 1; nilfs_search_super_root() 905 nilfs->ns_pseg_offset = pseg_start + nblocks - seg_start; nilfs_search_super_root() 922 pseg_start += nblocks; nilfs_search_super_root()
|
H A D | segbuf.c | 96 segbuf->sb_pseg_start = prev->sb_pseg_start + prev->sb_sum.nblocks; nilfs_segbuf_map_cont() 127 segbuf->sb_pseg_start + segbuf->sb_sum.nblocks); nilfs_segbuf_extend_payload() 141 segbuf->sb_sum.nblocks = segbuf->sb_sum.nsumblk = 0; nilfs_segbuf_reset() 172 raw_sum->ss_nblocks = cpu_to_le32(segbuf->sb_sum.nblocks); nilfs_segbuf_fill_in_segsum() 429 wi->rest_blocks = segbuf->sb_sum.nblocks; nilfs_segbuf_prepare_write()
|
H A D | sufile.h | 43 unsigned long nblocks, time_t modtime);
|
H A D | segment.c | 384 sci->sc_nblk_this_inc += sci->sc_curseg->sb_sum.nblocks; nilfs_segctor_feed_segment() 397 if (segbuf->sb_sum.nblocks >= segbuf->sb_rest_blocks) { nilfs_segctor_add_super_root() 484 if (segbuf->sb_sum.nblocks + required + 1 > segbuf->sb_rest_blocks) { nilfs_segctor_add_file_block() 980 (sci->sc_nblk_this_inc + sci->sc_curseg->sb_sum.nblocks); nilfs_segctor_buffer_rest() 1378 live_blocks = segbuf->sb_sum.nblocks + nilfs_segctor_update_segusage() 1437 sci->sc_nblk_this_inc += sci->sc_curseg->sb_sum.nblocks; nilfs_segctor_collect() 1490 unsigned long nblocks = 0, ndatablk = 0; nilfs_segctor_update_payload_blocknr() local 1513 nblocks = le32_to_cpu(finfo->fi_nblocks); nilfs_segctor_update_payload_blocknr() 1541 if (--nblocks == 0) { nilfs_segctor_update_payload_blocknr() 1751 + segbuf->sb_sum.nblocks; nilfs_set_next_segment()
|
H A D | ioctl.c | 1228 * %-EINVAL - Invalid values in input (segment number, flags or nblocks)
|
/linux-4.1.27/fs/minix/ |
H A D | itree_v1.c | 66 return nblocks(size, sb); V1_minix_blocks()
|
H A D | itree_v2.c | 75 return nblocks(size, sb); V2_minix_blocks()
|
H A D | itree_common.c | 350 static inline unsigned nblocks(loff_t size, struct super_block *sb) nblocks() function
|
/linux-4.1.27/drivers/misc/ |
H A D | sram.c | 66 unsigned int nblocks; sram_probe() local 107 nblocks = (np) ? of_get_available_child_count(np) + 1 : 1; sram_probe() 108 rblocks = kmalloc((nblocks) * sizeof(*rblocks), GFP_KERNEL); sram_probe() 146 rblocks[nblocks - 1].start = size; 147 rblocks[nblocks - 1].size = 0; 148 list_add_tail(&rblocks[nblocks - 1].list, &reserve_list);
|
/linux-4.1.27/fs/freevxfs/ |
H A D | vxfs_lookup.c | 116 u_long npages, page, nblocks, pblocks, block; vxfs_find_entry() local 122 nblocks = dir_blocks(ip); vxfs_find_entry() 134 for (block = 0; block <= nblocks && block <= pblocks; block++) { vxfs_find_entry() 243 u_long page, npages, block, pblocks, nblocks, offset; vxfs_readdir() local 262 nblocks = dir_blocks(ip); vxfs_readdir() 278 for (; block <= nblocks && block <= pblocks; block++) { vxfs_readdir()
|
/linux-4.1.27/drivers/s390/char/ |
H A D | tape_char.c | 192 int nblocks; tapechar_write() local 205 nblocks = count / block_size; tapechar_write() 208 nblocks = 1; tapechar_write() 216 DBF_EVENT(6, "TCHAR:nblocks: %x\n", nblocks); tapechar_write() 223 for (i = 0; i < nblocks; i++) { tapechar_write()
|
/linux-4.1.27/fs/ext4/ |
H A D | ext4_jbd2.h | 299 #define ext4_journal_start_sb(sb, type, nblocks) \ 300 __ext4_journal_start_sb((sb), __LINE__, (type), (nblocks), 0) 302 #define ext4_journal_start(inode, type, nblocks) \ 303 __ext4_journal_start((inode), __LINE__, (type), (nblocks), 0) 336 static inline int ext4_journal_extend(handle_t *handle, int nblocks) ext4_journal_extend() argument 339 return jbd2_journal_extend(handle, nblocks); ext4_journal_extend() 343 static inline int ext4_journal_restart(handle_t *handle, int nblocks) ext4_journal_restart() argument 346 return jbd2_journal_restart(handle, nblocks); ext4_journal_restart()
|
H A D | ialloc.c | 713 unsigned int line_no, int nblocks) __ext4_new_inode() 836 BUG_ON(nblocks <= 0); __ext4_new_inode() 838 handle_type, nblocks, __ext4_new_inode() 710 __ext4_new_inode(handle_t *handle, struct inode *dir, umode_t mode, const struct qstr *qstr, __u32 goal, uid_t *owner, int handle_type, unsigned int line_no, int nblocks) __ext4_new_inode() argument
|
H A D | namei.c | 1372 ext4_lblk_t nblocks; ext4_find_entry() local 1399 nblocks = 1; ext4_find_entry() 1414 nblocks = dir->i_size >> EXT4_BLOCK_SIZE_BITS(sb); ext4_find_entry() 1416 if (start >= nblocks) ext4_find_entry() 1434 if (b >= nblocks || (num && block == start)) { ext4_find_entry() 1484 if (++block >= nblocks) ext4_find_entry() 1492 block = nblocks; ext4_find_entry() 1493 nblocks = dir->i_size >> EXT4_BLOCK_SIZE_BITS(sb); ext4_find_entry() 1494 if (block < nblocks) { ext4_find_entry()
|
H A D | ext4.h | 2231 unsigned int line_no, int nblocks); 2237 type, nblocks) \ 2239 (type), __LINE__, (nblocks)) 2312 extern int ext4_truncate_restart_trans(handle_t *, struct inode *, int nblocks);
|
H A D | inode.c | 158 int nblocks) ext4_truncate_restart_trans() 171 ret = ext4_journal_restart(handle, nblocks); ext4_truncate_restart_trans() 157 ext4_truncate_restart_trans(handle_t *handle, struct inode *inode, int nblocks) ext4_truncate_restart_trans() argument
|
/linux-4.1.27/fs/ |
H A D | mpage.c | 157 unsigned nblocks; do_mpage_readpage() local 173 nblocks = map_bh->b_size >> blkbits; do_mpage_readpage() 175 block_in_file < (*first_logical_block + nblocks)) { do_mpage_readpage() 177 unsigned last = nblocks - map_offset; do_mpage_readpage() 235 nblocks = map_bh->b_size >> blkbits; do_mpage_readpage() 237 if (relative_block == nblocks) { do_mpage_readpage() 293 nblocks = map_bh->b_size >> blkbits; do_mpage_readpage() 294 if ((buffer_boundary(map_bh) && relative_block == nblocks) || do_mpage_readpage()
|
H A D | direct-io.c | 831 unsigned nblocks; clean_blockdev_aliases() local 833 nblocks = map_bh->b_size >> dio->inode->i_blkbits; clean_blockdev_aliases() 835 for (i = 0; i < nblocks; i++) { clean_blockdev_aliases()
|
/linux-4.1.27/sound/synth/ |
H A D | util_mem.c | 128 hdr->nblocks++; __snd_util_memblk_new() 156 hdr->nblocks--; __snd_util_mem_free()
|
/linux-4.1.27/include/sound/ |
H A D | util_mem.h | 42 int nblocks; /* # of allocated blocks */ member in struct:snd_util_memhdr
|
/linux-4.1.27/drivers/media/usb/pwc/ |
H A D | pwc-dec23.c | 606 int compression_index, nblocks; DecompressBand23() local 617 nblocks = compressed_image_width / 4; DecompressBand23() 623 while (nblocks) { DecompressBand23() 627 nblocks--; DecompressBand23() 631 nblocks = compressed_image_width / 8; DecompressBand23() 637 while (nblocks) { DecompressBand23() 646 nblocks -= 2; DecompressBand23()
|
/linux-4.1.27/fs/jbd2/ |
H A D | transaction.c | 381 static handle_t *new_handle(int nblocks) new_handle() argument 386 handle->h_buffer_credits = nblocks; new_handle() 398 * @nblocks: number of block buffer we might modify 400 * We make sure that the transaction can guarantee at least nblocks of 414 handle_t *jbd2__journal_start(journal_t *journal, int nblocks, int rsv_blocks, jbd2__journal_start() argument 430 handle = new_handle(nblocks); jbd2__journal_start() 457 line_no, nblocks); jbd2__journal_start() 463 handle_t *jbd2_journal_start(journal_t *journal, int nblocks) jbd2_journal_start() argument 465 return jbd2__journal_start(journal, nblocks, 0, GFP_NOFS, 0, 0); jbd2_journal_start() 530 * @nblocks: nr blocks to try to extend by. 547 int jbd2_journal_extend(handle_t *handle, int nblocks) jbd2_journal_extend() argument 565 "transaction not running\n", handle, nblocks); jbd2_journal_extend() 570 wanted = atomic_add_return(nblocks, jbd2_journal_extend() 575 "transaction too large\n", handle, nblocks); jbd2_journal_extend() 576 atomic_sub(nblocks, &transaction->t_outstanding_credits); jbd2_journal_extend() 583 "insufficient log space\n", handle, nblocks); jbd2_journal_extend() 584 atomic_sub(nblocks, &transaction->t_outstanding_credits); jbd2_journal_extend() 592 nblocks); jbd2_journal_extend() 594 handle->h_buffer_credits += nblocks; jbd2_journal_extend() 595 handle->h_requested_credits += nblocks; jbd2_journal_extend() 598 jbd_debug(3, "extended handle %p by %d\n", handle, nblocks); jbd2_journal_extend() 610 * @nblocks: nr credits requested 622 int jbd2__journal_restart(handle_t *handle, int nblocks, gfp_t gfp_mask) jbd2__journal_restart() argument 664 handle->h_buffer_credits = nblocks; jbd2__journal_restart() 671 int jbd2_journal_restart(handle_t *handle, int nblocks) jbd2_journal_restart() argument 673 return jbd2__journal_restart(handle, nblocks, GFP_NOFS); jbd2_journal_restart()
|
H A D | checkpoint.c | 113 int nblocks, space_left; __jbd2_log_wait_for_space() local 116 nblocks = jbd2_space_needed(journal); __jbd2_log_wait_for_space() 117 while (jbd2_log_space_left(journal) < nblocks) { __jbd2_log_wait_for_space() 138 nblocks = jbd2_space_needed(journal); __jbd2_log_wait_for_space() 140 if (space_left < nblocks) { __jbd2_log_wait_for_space() 166 __func__, nblocks, space_left); __jbd2_log_wait_for_space()
|
/linux-4.1.27/fs/jbd/ |
H A D | checkpoint.c | 124 int nblocks, space_left; __log_wait_for_space() local 127 nblocks = jbd_space_needed(journal); __log_wait_for_space() 128 while (__log_space_left(journal) < nblocks) { __log_wait_for_space() 147 nblocks = jbd_space_needed(journal); __log_wait_for_space() 149 if (space_left < nblocks) { __log_wait_for_space() 167 __func__, nblocks, space_left); __log_wait_for_space()
|
H A D | transaction.c | 88 int nblocks = handle->h_buffer_credits; start_this_handle() local 92 if (nblocks > journal->j_max_transaction_buffers) { start_this_handle() 94 current->comm, nblocks, start_this_handle() 167 needed = transaction->t_outstanding_credits + nblocks; start_this_handle() 224 transaction->t_outstanding_credits += nblocks; start_this_handle() 228 handle, nblocks, transaction->t_outstanding_credits, start_this_handle() 243 static handle_t *new_handle(int nblocks) new_handle() argument 248 handle->h_buffer_credits = nblocks; new_handle() 259 * @nblocks: number of block buffer we might modify 261 * We make sure that the transaction can guarantee at least nblocks of 271 handle_t *journal_start(journal_t *journal, int nblocks) journal_start() argument 285 handle = new_handle(nblocks); journal_start() 303 * @nblocks: nr blocks to try to extend by. 320 int journal_extend(handle_t *handle, int nblocks) journal_extend() argument 338 "transaction not running\n", handle, nblocks); journal_extend() 343 wanted = transaction->t_outstanding_credits + nblocks; journal_extend() 347 "transaction too large\n", handle, nblocks); journal_extend() 353 "insufficient log space\n", handle, nblocks); journal_extend() 357 handle->h_buffer_credits += nblocks; journal_extend() 358 transaction->t_outstanding_credits += nblocks; journal_extend() 361 jbd_debug(3, "extended handle %p by %d\n", handle, nblocks); journal_extend() 374 * @nblocks: nr credits requested 386 int journal_restart(handle_t *handle, int nblocks) journal_restart() argument 418 handle->h_buffer_credits = nblocks; journal_restart()
|
/linux-4.1.27/drivers/ide/ |
H A D | ide-floppy_ioctl.c | 26 * int nblocks; 159 * int nblocks; ide_floppy_format_unit()
|
/linux-4.1.27/net/sunrpc/auth_gss/ |
H A D | gss_krb5_crypto.c | 653 int nblocks, nbytes; gss_krb5_aes_encrypt() local 714 nblocks = (nbytes + blocksize - 1) / blocksize; gss_krb5_aes_encrypt() 716 if (nblocks > 2) gss_krb5_aes_encrypt() 717 cbcbytes = (nblocks - 2) * blocksize; gss_krb5_aes_encrypt() 770 int nblocks, blocksize, cbcbytes; gss_krb5_aes_decrypt() local 793 nblocks = (subbuf.len + blocksize - 1) / blocksize; gss_krb5_aes_decrypt() 796 if (nblocks > 2) gss_krb5_aes_decrypt() 797 cbcbytes = (nblocks - 2) * blocksize; gss_krb5_aes_decrypt()
|
/linux-4.1.27/drivers/mtd/ |
H A D | ftl.c | 788 u_long sector, u_long nblocks) ftl_read() 796 part, sector, nblocks); ftl_read() 803 for (i = 0; i < nblocks; i++) { ftl_read() 902 u_long sector, u_long nblocks) ftl_write() 910 part, sector, nblocks); ftl_write() 916 while (part->FreeTotal < nblocks) { ftl_write() 925 for (i = 0; i < nblocks; i++) { ftl_write() 787 ftl_read(partition_t *part, caddr_t buffer, u_long sector, u_long nblocks) ftl_read() argument 901 ftl_write(partition_t *part, caddr_t buffer, u_long sector, u_long nblocks) ftl_write() argument
|
/linux-4.1.27/fs/btrfs/ |
H A D | file-item.c | 176 int nblocks; __btrfs_lookup_bio_sums() local 185 nblocks = bio->bi_iter.bi_size >> inode->i_sb->s_blocksize_bits; __btrfs_lookup_bio_sums() 187 if (nblocks * csum_size > BTRFS_BIO_INLINE_CSUM_SIZE) { __btrfs_lookup_bio_sums() 188 btrfs_bio->csum_allocated = kmalloc_array(nblocks, __btrfs_lookup_bio_sums() 227 (u32 *)csum, nblocks); __btrfs_lookup_bio_sums() 276 count = min_t(int, nblocks, (item_last_offset - disk_bytenr) >> __btrfs_lookup_bio_sums() 283 nblocks -= count; __btrfs_lookup_bio_sums()
|
/linux-4.1.27/fs/gfs2/ |
H A D | rgrp.c | 2307 * @nblocks: requested number of blocks/extent length (value/result) 2314 int gfs2_alloc_blocks(struct gfs2_inode *ip, u64 *bn, unsigned int *nblocks, gfs2_alloc_blocks() argument 2335 fs_warn(sdp, "inum=%llu error=%d, nblocks=%u, full=%d fail_pt=%d\n", gfs2_alloc_blocks() 2336 (unsigned long long)ip->i_no_addr, error, *nblocks, gfs2_alloc_blocks() 2342 gfs2_alloc_extent(&rbm, dinode, nblocks); gfs2_alloc_blocks() 2346 gfs2_adjust_reservation(ip, &rbm, *nblocks); gfs2_alloc_blocks() 2347 ndata = *nblocks; gfs2_alloc_blocks() 2363 if (rbm.rgd->rd_free < *nblocks) { gfs2_alloc_blocks() 2364 pr_warn("nblocks=%u\n", *nblocks); gfs2_alloc_blocks() 2368 rbm.rgd->rd_free -= *nblocks; gfs2_alloc_blocks() 2380 gfs2_statfs_change(sdp, 0, -(s64)*nblocks, dinode ? 1 : 0); gfs2_alloc_blocks() 2382 gfs2_trans_add_unrevoke(sdp, block, *nblocks); gfs2_alloc_blocks() 2384 gfs2_quota_change(ip, *nblocks, ip->i_inode.i_uid, ip->i_inode.i_gid); gfs2_alloc_blocks() 2386 rbm.rgd->rd_free_clone -= *nblocks; gfs2_alloc_blocks() 2387 trace_gfs2_block_alloc(ip, rbm.rgd, block, *nblocks, gfs2_alloc_blocks()
|
/linux-4.1.27/sound/synth/emux/ |
H A D | emux_proc.c | 51 snd_iprintf(buf, "Allocated Blocks: %d\n", emu->memhdr->nblocks); snd_emux_proc_info_read()
|
/linux-4.1.27/include/linux/ |
H A D | jbd.h | 852 extern handle_t *journal_start(journal_t *, int nblocks); 853 extern int journal_restart (handle_t *, int nblocks); 854 extern int journal_extend (handle_t *, int nblocks); 1011 int nblocks = journal->j_max_transaction_buffers; jbd_space_needed() local 1013 nblocks += journal->j_committing_transaction-> jbd_space_needed() 1015 return nblocks; jbd_space_needed()
|
H A D | jbd2.h | 1117 extern handle_t *jbd2_journal_start(journal_t *, int nblocks); 1121 extern int jbd2_journal_restart(handle_t *, int nblocks); 1122 extern int jbd2__journal_restart(handle_t *, int nblocks, gfp_t gfp_mask); 1126 extern int jbd2_journal_extend (handle_t *, int nblocks); 1324 int nblocks = journal->j_max_transaction_buffers; jbd2_space_needed() local 1325 return nblocks + (nblocks >> JBD2_CONTROL_BLOCKS_SHIFT); jbd2_space_needed()
|
H A D | cred.h | 35 int nblocks; member in struct:group_info
|
H A D | nilfs2_fs.h | 760 NILFS_SUINFO_UPDATE_FNS(NBLOCKS, nblocks)
|
/linux-4.1.27/fs/ext3/ |
H A D | ext3.h | 1260 handle_t *ext3_journal_start_sb(struct super_block *sb, int nblocks); 1263 static inline handle_t *ext3_journal_start(struct inode *inode, int nblocks) ext3_journal_start() argument 1265 return ext3_journal_start_sb(inode->i_sb, nblocks); ext3_journal_start() 1276 static inline int ext3_journal_extend(handle_t *handle, int nblocks) ext3_journal_extend() argument 1278 return journal_extend(handle, nblocks); ext3_journal_extend() 1281 static inline int ext3_journal_restart(handle_t *handle, int nblocks) ext3_journal_restart() argument 1283 return journal_restart(handle, nblocks); ext3_journal_restart()
|
H A D | namei.c | 856 int nblocks, i, err; ext3_find_entry() local 871 nblocks = 1; ext3_find_entry() 885 nblocks = dir->i_size >> EXT3_BLOCK_SIZE_BITS(sb); ext3_find_entry() 887 if (start >= nblocks) ext3_find_entry() 905 if (b >= nblocks || (num && block == start)) { ext3_find_entry() 942 if (++block >= nblocks) ext3_find_entry() 950 block = nblocks; ext3_find_entry() 951 nblocks = dir->i_size >> EXT3_BLOCK_SIZE_BITS(sb); ext3_find_entry() 952 if (block < nblocks) { ext3_find_entry()
|
H A D | super.c | 69 handle_t *ext3_journal_start_sb(struct super_block *sb, int nblocks) ext3_journal_start_sb() argument 86 return journal_start(journal, nblocks); ext3_journal_start_sb()
|
/linux-4.1.27/fs/ocfs2/ |
H A D | journal.c | 410 * 'nblocks' is what you want to add to the current transaction. 426 int ocfs2_extend_trans(handle_t *handle, int nblocks) ocfs2_extend_trans() argument 431 BUG_ON(nblocks < 0); ocfs2_extend_trans() 433 if (!nblocks) ocfs2_extend_trans() 438 trace_ocfs2_extend_trans(old_nblocks, nblocks); ocfs2_extend_trans() 443 status = jbd2_journal_extend(handle, nblocks); ocfs2_extend_trans() 451 trace_ocfs2_extend_trans_restart(old_nblocks + nblocks); ocfs2_extend_trans() 453 old_nblocks + nblocks); ocfs2_extend_trans()
|
H A D | journal.h | 239 * ocfs2_extend_trans - Extend a handle by nblocks credits. This may 260 int ocfs2_extend_trans(handle_t *handle, int nblocks);
|
H A D | dir.c | 705 int nblocks, i, err; ocfs2_find_entry_el() local 709 nblocks = i_size_read(dir) >> sb->s_blocksize_bits; ocfs2_find_entry_el() 711 if (start >= nblocks) ocfs2_find_entry_el() 730 if (b >= nblocks || (num && block == start)) { ocfs2_find_entry_el() 767 if (++block >= nblocks) ocfs2_find_entry_el() 775 block = nblocks; ocfs2_find_entry_el() 776 nblocks = i_size_read(dir) >> sb->s_blocksize_bits; ocfs2_find_entry_el() 777 if (block < nblocks) { ocfs2_find_entry_el()
|
/linux-4.1.27/fs/xfs/ |
H A D | xfs_rtalloc.c | 764 xfs_extlen_t nblocks, /* new count of blocks */ xfs_growfs_rt_alloc() 782 while (oblocks < nblocks) { xfs_growfs_rt_alloc() 787 resblks = XFS_GROWFSRT_SPACE_RES(mp, nblocks - oblocks); xfs_growfs_rt_alloc() 808 error = xfs_bmapi_write(tp, ip, oblocks, nblocks - oblocks, xfs_growfs_rt_alloc() 761 xfs_growfs_rt_alloc( xfs_mount_t *mp, xfs_extlen_t oblocks, xfs_extlen_t nblocks, xfs_inode_t *ip) xfs_growfs_rt_alloc() argument
|
H A D | xfs_mount.c | 158 __uint64_t nblocks) xfs_sb_validate_fsb_count() 164 if (nblocks >> (PAGE_CACHE_SHIFT - sbp->sb_blocklog) > ULONG_MAX) xfs_sb_validate_fsb_count() 156 xfs_sb_validate_fsb_count( xfs_sb_t *sbp, __uint64_t nblocks) xfs_sb_validate_fsb_count() argument
|
H A D | xfs_inode.c | 3520 "total extents = %d, nblocks = %Ld, ptr 0x%p", xfs_iflush_int()
|
H A D | xfs_log_recover.c | 2655 "dino bp 0x%p, ino %Ld, total extents = %d, nblocks = %Ld", xlog_recover_inode_pass2()
|
/linux-4.1.27/include/uapi/linux/ |
H A D | nvme.h | 542 __u16 nblocks; member in struct:nvme_user_io
|
/linux-4.1.27/drivers/cdrom/ |
H A D | cdrom.c | 2063 int blocksize, int nblocks) cdrom_read_cd() 2073 cgc->cmd[6] = (nblocks >> 16) & 0xff; cdrom_read_cd() 2074 cgc->cmd[7] = (nblocks >> 8) & 0xff; cdrom_read_cd() 2075 cgc->cmd[8] = nblocks & 0xff; cdrom_read_cd() 2076 cgc->buflen = blocksize * nblocks; cdrom_read_cd() 2083 int lba, int nblocks, int format, int blksize) cdrom_read_block() 2097 cgc->cmd[6] = (nblocks >> 16) & 0xff; cdrom_read_block() 2098 cgc->cmd[7] = (nblocks >> 8) & 0xff; cdrom_read_block() 2099 cgc->cmd[8] = nblocks & 0xff; cdrom_read_block() 2100 cgc->buflen = blksize * nblocks; cdrom_read_block() 2061 cdrom_read_cd(struct cdrom_device_info *cdi, struct packet_command *cgc, int lba, int blocksize, int nblocks) cdrom_read_cd() argument 2081 cdrom_read_block(struct cdrom_device_info *cdi, struct packet_command *cgc, int lba, int nblocks, int format, int blksize) cdrom_read_block() argument
|
/linux-4.1.27/fs/isofs/ |
H A D | inode.c | 992 struct buffer_head **bh, unsigned long nblocks) isofs_get_blocks() 1016 while (nblocks) { isofs_get_blocks() 1070 nblocks--; isofs_get_blocks() 991 isofs_get_blocks(struct inode *inode, sector_t iblock, struct buffer_head **bh, unsigned long nblocks) isofs_get_blocks() argument
|
/linux-4.1.27/drivers/net/ethernet/neterion/vxge/ |
H A D | vxge-config.c | 1319 u32 nblocks = 0; vxge_hw_device_initialize() local 1357 nblocks++; vxge_hw_device_initialize() 1365 nblocks += device_config->vp_config[i].ring.ring_blocks; vxge_hw_device_initialize() 1369 nblocks += device_config->vp_config[i].fifo.fifo_blocks; vxge_hw_device_initialize() 1370 nblocks++; vxge_hw_device_initialize() 1375 device_config->dma_blockpool_initial + nblocks, vxge_hw_device_initialize() 1376 device_config->dma_blockpool_max + nblocks) != VXGE_HW_OK) { vxge_hw_device_initialize()
|
/linux-4.1.27/drivers/block/ |
H A D | nvme-core.c | 1760 length = (io.nblocks + 1) << ns->lba_shift; nvme_submit_io() 1761 meta_len = (io.nblocks + 1) * ns->ms; nvme_submit_io() 1813 c.rw.length = cpu_to_le16(io.nblocks); nvme_submit_io()
|
/linux-4.1.27/net/ipv4/ |
H A D | ping.c | 272 for (i = 0; i < group_info->nblocks; i++) { ping_init_sock()
|
/linux-4.1.27/fs/nfs/ |
H A D | nfs4xdr.c | 4905 uint32_t limit_type, nblocks, blocksize; decode_space_limit() local 4916 nblocks = be32_to_cpup(p++); decode_space_limit() 4918 *maxsize = (uint64_t)nblocks * (uint64_t)blocksize; decode_space_limit()
|
/linux-4.1.27/fs/xfs/libxfs/ |
H A D | xfs_inode_fork.c | 93 "corrupt dinode %Lu, extent total = %d, nblocks = %Lu.", xfs_iformat_fork()
|