nbblks 22 fs/xfs/libxfs/xfs_dquot_buf.c unsigned int nbblks) /* basic block units */ nbblks 24 fs/xfs/libxfs/xfs_dquot_buf.c ASSERT(nbblks > 0); nbblks 25 fs/xfs/libxfs/xfs_dquot_buf.c return BBTOB(nbblks) / sizeof(xfs_dqblk_t); nbblks 144 fs/xfs/libxfs/xfs_quota_defs.h extern int xfs_calc_dquots_per_chunk(unsigned int nbblks); nbblks 98 fs/xfs/xfs_log_recover.c int nbblks) nbblks 106 fs/xfs/xfs_log_recover.c if (!xlog_verify_bno(log, 0, nbblks)) { nbblks 108 fs/xfs/xfs_log_recover.c nbblks); nbblks 127 fs/xfs/xfs_log_recover.c if (nbblks > 1 && log->l_sectBBsize > 1) nbblks 128 fs/xfs/xfs_log_recover.c nbblks += log->l_sectBBsize; nbblks 129 fs/xfs/xfs_log_recover.c nbblks = round_up(nbblks, log->l_sectBBsize); nbblks 130 fs/xfs/xfs_log_recover.c return kmem_alloc_io(BBTOB(nbblks), align_mask, KM_MAYFAIL | KM_ZERO); nbblks 149 fs/xfs/xfs_log_recover.c unsigned int nbblks, nbblks 155 fs/xfs/xfs_log_recover.c if (!xlog_verify_bno(log, blk_no, nbblks)) { nbblks 158 fs/xfs/xfs_log_recover.c blk_no, nbblks); nbblks 164 fs/xfs/xfs_log_recover.c nbblks = round_up(nbblks, log->l_sectBBsize); nbblks 165 fs/xfs/xfs_log_recover.c ASSERT(nbblks > 0); nbblks 168 fs/xfs/xfs_log_recover.c BBTOB(nbblks), data, op); nbblks 173 fs/xfs/xfs_log_recover.c blk_no, nbblks, error); nbblks 182 fs/xfs/xfs_log_recover.c int nbblks, nbblks 185 fs/xfs/xfs_log_recover.c return xlog_do_io(log, blk_no, nbblks, data, REQ_OP_READ); nbblks 192 fs/xfs/xfs_log_recover.c int nbblks, nbblks 198 fs/xfs/xfs_log_recover.c error = xlog_do_io(log, blk_no, nbblks, data, REQ_OP_READ); nbblks 208 fs/xfs/xfs_log_recover.c int nbblks, nbblks 211 fs/xfs/xfs_log_recover.c return xlog_do_io(log, blk_no, nbblks, data, REQ_OP_WRITE); nbblks 372 fs/xfs/xfs_log_recover.c int nbblks, nbblks 389 fs/xfs/xfs_log_recover.c bufblks = 1 << ffs(nbblks); nbblks 398 fs/xfs/xfs_log_recover.c for (i = start_blk; i < start_blk + nbblks; i += bufblks) { nbblks 401 fs/xfs/xfs_log_recover.c bcount = min(bufblks, (start_blk + nbblks - i));