blks 439 arch/mips/include/asm/octeon/cvmx-dpi-defs.h uint64_t blks:4; blks 441 arch/mips/include/asm/octeon/cvmx-dpi-defs.h uint64_t blks:4; blks 452 arch/mips/include/asm/octeon/cvmx-dpi-defs.h uint64_t blks:4; blks 454 arch/mips/include/asm/octeon/cvmx-dpi-defs.h uint64_t blks:4; blks 50 arch/powerpc/lib/rheap.c unsigned long blks, blke; blks 70 arch/powerpc/lib/rheap.c blks = (unsigned long)info->block; blks 74 arch/powerpc/lib/rheap.c fixup(blks, blke, delta, &blk->list); blks 76 arch/powerpc/lib/rheap.c fixup(blks, blke, delta, &info->empty_list); blks 77 arch/powerpc/lib/rheap.c fixup(blks, blke, delta, &info->free_list); blks 78 arch/powerpc/lib/rheap.c fixup(blks, blke, delta, &info->taken_list); blks 963 drivers/lightnvm/core.c u8 *blks, int nr_blks, struct nvm_chk_meta *meta) blks 970 drivers/lightnvm/core.c blktype = blks[offset]; blks 973 drivers/lightnvm/core.c if (blks[offset + pl] & blks 975 drivers/lightnvm/core.c blktype = blks[offset + pl]; blks 1007 drivers/lightnvm/core.c u8 *blks; blks 1022 drivers/lightnvm/core.c blks = kmalloc(nr_blks, GFP_KERNEL); blks 1023 drivers/lightnvm/core.c if (!blks) blks 1038 drivers/lightnvm/core.c ret = dev->ops->get_bb_tbl(dev, ppa_dev, blks); blks 1042 drivers/lightnvm/core.c ret = nvm_bb_to_chunk(dev, ppa_gen, blks, nr_blks, blks 1052 drivers/lightnvm/core.c kfree(blks); blks 1195 drivers/net/ethernet/broadcom/cnic.c int total_mem, blks, i; blks 1198 drivers/net/ethernet/broadcom/cnic.c blks = total_mem / ctx_blk_size; blks 1200 drivers/net/ethernet/broadcom/cnic.c blks++; blks 1202 drivers/net/ethernet/broadcom/cnic.c if (blks > cp->ethdev->ctx_tbl_len) blks 1205 drivers/net/ethernet/broadcom/cnic.c cp->ctx_arr = kcalloc(blks, sizeof(struct cnic_ctx), GFP_KERNEL); blks 1209 drivers/net/ethernet/broadcom/cnic.c cp->ctx_blks = blks; blks 1218 drivers/net/ethernet/broadcom/cnic.c for (i = 0; i < blks; i++) { blks 35 drivers/net/ethernet/cisco/enic/vnic_rq.c unsigned int blks = VNIC_RQ_BUF_BLKS_NEEDED(count); blks 37 drivers/net/ethernet/cisco/enic/vnic_rq.c for (i = 0; i < blks; i++) { blks 43 drivers/net/ethernet/cisco/enic/vnic_rq.c for (i = 0; i < blks; i++) { blks 35 drivers/net/ethernet/cisco/enic/vnic_wq.c unsigned int blks = VNIC_WQ_BUF_BLKS_NEEDED(count); blks 37 drivers/net/ethernet/cisco/enic/vnic_wq.c for (i = 0; i < blks; i++) { blks 43 drivers/net/ethernet/cisco/enic/vnic_wq.c for (i = 0; i < blks; i++) { blks 1299 drivers/net/wireless/ti/wl12xx/main.c u32 blks, u32 spare_blks) blks 1302 drivers/net/wireless/ti/wl12xx/main.c desc->wl128x_mem.total_mem_blocks = blks; blks 1305 drivers/net/wireless/ti/wl12xx/main.c desc->wl127x_mem.total_mem_blocks = blks; blks 1074 drivers/net/wireless/ti/wl18xx/main.c u32 blks, u32 spare_blks) blks 1076 drivers/net/wireless/ti/wl18xx/main.c desc->wl18xx_mem.total_mem_blocks = blks; blks 25 drivers/net/wireless/ti/wlcore/hw_ops.h u32 blks, u32 spare_blks) blks 30 drivers/net/wireless/ti/wlcore/hw_ops.h return wl->ops->set_tx_desc_blocks(wl, desc, blks, spare_blks); blks 52 drivers/net/wireless/ti/wlcore/wlcore.h u32 blks, u32 spare_blks); blks 476 drivers/nvme/host/lightnvm.c u8 *blks) blks 525 drivers/nvme/host/lightnvm.c memcpy(blks, bb_tbl->blk, geo->num_chk * geo->num_pln); blks 31 drivers/scsi/fnic/vnic_rq.c unsigned int blks = VNIC_RQ_BUF_BLKS_NEEDED(count); blks 33 drivers/scsi/fnic/vnic_rq.c for (i = 0; i < blks; i++) { blks 41 drivers/scsi/fnic/vnic_rq.c for (i = 0; i < blks; i++) { blks 51 drivers/scsi/fnic/vnic_wq.c unsigned int blks = VNIC_WQ_BUF_BLKS_NEEDED(count); blks 53 drivers/scsi/fnic/vnic_wq.c for (i = 0; i < blks; i++) { blks 61 drivers/scsi/fnic/vnic_wq.c for (i = 0; i < blks; i++) { blks 47 drivers/scsi/snic/vnic_wq.c unsigned int blks = VNIC_WQ_BUF_BLKS_NEEDED(count); blks 49 drivers/scsi/snic/vnic_wq.c for (i = 0; i < blks; i++) { blks 58 drivers/scsi/snic/vnic_wq.c for (i = 0; i < blks; i++) { blks 765 drivers/scsi/st.c int transfer, blks; blks 784 drivers/scsi/st.c blks = transfer / STp->block_size; blks 785 drivers/scsi/st.c cmd[2] = blks >> 16; blks 786 drivers/scsi/st.c cmd[3] = blks >> 8; blks 787 drivers/scsi/st.c cmd[4] = blks; blks 808 drivers/scsi/st.c STps->drv_block += blks; blks 817 drivers/scsi/st.c STps->drv_block += blks; blks 1623 drivers/scsi/st.c ssize_t i, do_count, blks, transfer; blks 1761 drivers/scsi/st.c blks = transfer = do_count; blks 1764 drivers/scsi/st.c blks = STbp->buffer_bytes; blks 1766 drivers/scsi/st.c blks = do_count; blks 1767 drivers/scsi/st.c blks /= STp->block_size; blks 1768 drivers/scsi/st.c transfer = blks * STp->block_size; blks 1770 drivers/scsi/st.c cmd[2] = blks >> 16; blks 1771 drivers/scsi/st.c cmd[3] = blks >> 8; blks 1772 drivers/scsi/st.c cmd[4] = blks; blks 1811 drivers/scsi/st.c blks = (transfer - undone) / STp->block_size; blks 1845 drivers/scsi/st.c blks = do_count = 0; blks 1865 drivers/scsi/st.c STps->drv_block += blks; blks 1902 drivers/scsi/st.c int transfer, blks, bytes; blks 1920 drivers/scsi/st.c blks = bytes = count; blks 1923 drivers/scsi/st.c blks = (STp->buffer)->buffer_blocks; blks 1924 drivers/scsi/st.c bytes = blks * STp->block_size; blks 1929 drivers/scsi/st.c blks = bytes / STp->block_size; blks 1930 drivers/scsi/st.c bytes = blks * STp->block_size; blks 1939 drivers/scsi/st.c cmd[2] = blks >> 16; blks 1940 drivers/scsi/st.c cmd[3] = blks >> 8; blks 1941 drivers/scsi/st.c cmd[4] = blks; blks 2002 drivers/scsi/st.c if (transfer == blks) { /* We did not get anything, error */ blks 2007 drivers/scsi/st.c STps->drv_block += blks - transfer + 1; blks 2012 drivers/scsi/st.c STbp->buffer_bytes = (blks - transfer) * blks 349 fs/erofs/data.c erofs_blk_t blks = i_size_read(inode) >> LOG_BLOCK_SIZE; blks 351 fs/erofs/data.c if (block >> LOG_SECTORS_PER_BLOCK >= blks) blks 362 fs/ext2/inode.c ext2_blks_to_allocate(Indirect * branch, int k, unsigned long blks, blks 373 fs/ext2/inode.c if (blks < blocks_to_boundary + 1) blks 374 fs/ext2/inode.c count += blks; blks 381 fs/ext2/inode.c while (count < blks && count <= blocks_to_boundary blks 399 fs/ext2/inode.c ext2_fsblk_t goal, int indirect_blks, int blks, blks 416 fs/ext2/inode.c target = blks + indirect_blks; blks 479 fs/ext2/inode.c int indirect_blks, int *blks, ext2_fsblk_t goal, blks 491 fs/ext2/inode.c *blks, new_blocks, &err); blks 536 fs/ext2/inode.c *blks = num; blks 561 fs/ext2/inode.c long block, Indirect *where, int num, int blks) blks 578 fs/ext2/inode.c if (num == 0 && blks > 1) { blks 580 fs/ext2/inode.c for (i = 1; i < blks; i++) blks 590 fs/ext2/inode.c block_i->last_alloc_logical_block = block + blks - 1; blks 592 fs/ext2/inode.c le32_to_cpu(where[num].key) + blks - 1; blks 289 fs/ext4/ext4.h #define EXT4_NUM_B2C(sbi, blks) (((blks) + (sbi)->s_cluster_ratio - 1) >> \ blks 270 fs/ext4/indirect.c static int ext4_blks_to_allocate(Indirect *branch, int k, unsigned int blks, blks 281 fs/ext4/indirect.c if (blks < blocks_to_boundary + 1) blks 282 fs/ext4/indirect.c count += blks; blks 289 fs/ext4/indirect.c while (count < blks && count <= blocks_to_boundary && blks 3550 fs/ext4/super.c int blks; blks 3552 fs/ext4/super.c blks = count_overhead(sb, i, buf); blks 3553 fs/ext4/super.c overhead += blks; blks 3554 fs/ext4/super.c if (blks) blks 3410 fs/f2fs/f2fs.h #define stat_inc_tot_blk_count(si, blks) \ blks 3411 fs/f2fs/f2fs.h ((si)->tot_blks += (blks)) blks 3413 fs/f2fs/f2fs.h #define stat_inc_data_blk_count(sbi, blks, gc_type) \ blks 3416 fs/f2fs/f2fs.h stat_inc_tot_blk_count(si, blks); \ blks 3417 fs/f2fs/f2fs.h si->data_blks += (blks); \ blks 3418 fs/f2fs/f2fs.h si->bg_data_blks += ((gc_type) == BG_GC) ? (blks) : 0; \ blks 3421 fs/f2fs/f2fs.h #define stat_inc_node_blk_count(sbi, blks, gc_type) \ blks 3424 fs/f2fs/f2fs.h stat_inc_tot_blk_count(si, blks); \ blks 3425 fs/f2fs/f2fs.h si->node_blks += (blks); \ blks 3426 fs/f2fs/f2fs.h si->bg_node_blks += ((gc_type) == BG_GC) ? (blks) : 0; \ blks 3463 fs/f2fs/f2fs.h #define stat_inc_tot_blk_count(si, blks) do { } while (0) blks 3464 fs/f2fs/f2fs.h #define stat_inc_data_blk_count(sbi, blks, gc_type) do { } while (0) blks 3465 fs/f2fs/f2fs.h #define stat_inc_node_blk_count(sbi, blks, gc_type) do { } while (0) blks 670 fs/gfs2/bmap.c unsigned n, i, blks, alloced = 0, iblks = 0, branch_start = 0; blks 706 fs/gfs2/bmap.c blks = dblks + iblks; blks 709 fs/gfs2/bmap.c n = blks - alloced; blks 400 fs/gfs2/file.c size_t blks = (size + sdp->sd_sb.sb_bsize - 1) >> sdp->sd_sb.sb_bsize_shift; blks 401 fs/gfs2/file.c int hint = min_t(size_t, INT_MAX, blks); blks 49 fs/gfs2/log.c unsigned int blks; blks 52 fs/gfs2/log.c blks = 1; blks 58 fs/gfs2/log.c blks += DIV_ROUND_UP(nstruct - first, second); blks 61 fs/gfs2/log.c return blks; blks 323 fs/gfs2/log.c void gfs2_log_release(struct gfs2_sbd *sdp, unsigned int blks) blks 326 fs/gfs2/log.c atomic_add(blks, &sdp->sd_log_blks_free); blks 327 fs/gfs2/log.c trace_gfs2_log_blocks(sdp, blks); blks 353 fs/gfs2/log.c int gfs2_log_reserve(struct gfs2_sbd *sdp, unsigned int blks) blks 357 fs/gfs2/log.c unsigned wanted = blks + reserved_blks; blks 362 fs/gfs2/log.c if (gfs2_assert_warn(sdp, blks) || blks 363 fs/gfs2/log.c gfs2_assert_warn(sdp, blks <= sdp->sd_jdesc->jd_blocks)) blks 365 fs/gfs2/log.c atomic_add(blks, &sdp->sd_log_blks_needed); blks 382 fs/gfs2/log.c free_blocks - blks) != free_blocks) { blks 387 fs/gfs2/log.c atomic_sub(blks, &sdp->sd_log_blks_needed); blks 388 fs/gfs2/log.c trace_gfs2_log_blocks(sdp, -blks); blks 399 fs/gfs2/log.c gfs2_log_release(sdp, blks); blks 67 fs/gfs2/log.h extern void gfs2_log_release(struct gfs2_sbd *sdp, unsigned int blks); blks 68 fs/gfs2/log.h extern int gfs2_log_reserve(struct gfs2_sbd *sdp, unsigned int blks); blks 755 fs/gfs2/lops.c unsigned int blks = be32_to_cpu(ld->ld_data1); blks 765 fs/gfs2/lops.c for (; blks; gfs2_replay_incr_blk(jd, &start), blks--) { blks 925 fs/gfs2/lops.c unsigned int blks = be32_to_cpu(ld->ld_length); blks 938 fs/gfs2/lops.c for (; blks; gfs2_replay_incr_blk(jd, &start), blks--) { blks 1008 fs/gfs2/lops.c unsigned int blks = be32_to_cpu(ld->ld_data1); blks 1018 fs/gfs2/lops.c for (; blks; gfs2_replay_incr_blk(jd, &start), blks--) { blks 235 fs/gfs2/xattr.c unsigned int blks = 0; blks 249 fs/gfs2/xattr.c blks++; blks 253 fs/gfs2/xattr.c if (!blks) blks 267 fs/gfs2/xattr.c RES_EATTR + RES_STATFS + RES_QUOTA, blks); blks 724 fs/gfs2/xattr.c unsigned int blks, blks 727 fs/gfs2/xattr.c struct gfs2_alloc_parms ap = { .target = blks }; blks 743 fs/gfs2/xattr.c blks + gfs2_rg_blocks(ip, blks) + blks 795 fs/gfs2/xattr.c unsigned int blks = 1; blks 804 fs/gfs2/xattr.c blks += DIV_ROUND_UP(er.er_data_len, jbsize); blks 806 fs/gfs2/xattr.c return ea_alloc_skeleton(ip, &er, blks, ea_init_i, NULL); blks 940 fs/gfs2/xattr.c unsigned int blks; blks 944 fs/gfs2/xattr.c blks = 2 + DIV_ROUND_UP(es->es_er->er_data_len, blks 947 fs/gfs2/xattr.c error = ea_alloc_skeleton(ip, es->es_er, blks, blks 1035 fs/gfs2/xattr.c unsigned int blks = 2; blks 1055 fs/gfs2/xattr.c blks++; blks 1057 fs/gfs2/xattr.c blks += DIV_ROUND_UP(er.er_data_len, GFS2_SB(&ip->i_inode)->sd_jbsize); blks 1059 fs/gfs2/xattr.c return ea_alloc_skeleton(ip, &er, blks, ea_set_block, el); blks 1256 fs/gfs2/xattr.c unsigned int blks = 0; blks 1293 fs/gfs2/xattr.c blks++; blks 1312 fs/gfs2/xattr.c RES_STATFS + RES_QUOTA, blks); blks 325 fs/ocfs2/xattr.c int blks = ocfs2_blocks_per_xattr_bucket(inode->i_sb); blks 327 fs/ocfs2/xattr.c BUG_ON(blks > OCFS2_XATTR_MAX_BLOCKS_PER_BUCKET); blks 332 fs/ocfs2/xattr.c bucket->bu_blocks = blks; blks 4189 fs/ocfs2/xattr.c int blks = ocfs2_blocks_per_xattr_bucket(inode->i_sb); blks 4198 fs/ocfs2/xattr.c char *target = bucket_block(bucket, blks - 1); blks 4204 fs/ocfs2/xattr.c for (i = 0; i < blks; i++) blks 329 fs/squashfs/file.c long long blks; blks 348 fs/squashfs/file.c blks = read_indexes(inode->i_sb, index - res, &start, &offset); blks 349 fs/squashfs/file.c if (blks < 0) blks 350 fs/squashfs/file.c return (int) blks; blks 351 fs/squashfs/file.c *block += blks; blks 510 fs/ufs/super.c unsigned size, blks, i; blks 519 fs/ufs/super.c blks = (size + uspi->s_fsize - 1) >> uspi->s_fshift; blks 524 fs/ufs/super.c for (i = 0; i < blks; i += uspi->s_fpb) { blks 526 fs/ufs/super.c if (i + uspi->s_fpb > blks) blks 527 fs/ufs/super.c size = (blks - i) * uspi->s_fsize; blks 650 fs/ufs/super.c unsigned blks, size, i; blks 657 fs/ufs/super.c blks = (size + uspi->s_fsize - 1) >> uspi->s_fshift; blks 659 fs/ufs/super.c for (i = 0; i < blks; i += uspi->s_fpb) { blks 661 fs/ufs/super.c if (i + uspi->s_fpb > blks) blks 662 fs/ufs/super.c size = (blks - i) * uspi->s_fsize; blks 263 fs/ufs/ufs_fs.h #define ufs_blkstofrags(blks) ((blks) << uspi->s_fpbshift) blks 1061 fs/xfs/libxfs/xfs_bmap.c int blks; /* space reservation */ blks 1071 fs/xfs/libxfs/xfs_bmap.c blks = XFS_ADDAFORK_SPACE_RES(mp); blks 1073 fs/xfs/libxfs/xfs_bmap.c error = xfs_trans_alloc(mp, &M_RES(mp)->tr_addafork, blks, 0, blks 1079 fs/xfs/libxfs/xfs_bmap.c error = xfs_trans_reserve_quota_nblks(tp, ip, blks, 0, rsvd ? blks 123 fs/xfs/scrub/dabtree.c struct xfs_da_state_blk *blks; blks 138 fs/xfs/scrub/dabtree.c blks = ds->state->path.blk; blks 139 fs/xfs/scrub/dabtree.c entry = xchk_da_btree_entry(ds, level - 1, blks[level - 1].index); blks 480 fs/xfs/scrub/dabtree.c struct xfs_da_state_blk *blks; blks 515 fs/xfs/scrub/dabtree.c blks = ds.state->path.blk; blks 524 fs/xfs/scrub/dabtree.c if (blks[level].bp == NULL) blks 527 fs/xfs/scrub/dabtree.c blks[level].index = 0; blks 530 fs/xfs/scrub/dabtree.c if (blks[level].magic != XFS_DA_NODE_MAGIC) { blks 532 fs/xfs/scrub/dabtree.c if (blks[level].index >= ds.maxrecs[level]) { blks 534 fs/xfs/scrub/dabtree.c blks[level - 1].index++; blks 542 fs/xfs/scrub/dabtree.c blks[level].index); blks 550 fs/xfs/scrub/dabtree.c blks[level].index++; blks 556 fs/xfs/scrub/dabtree.c if (blks[level].index >= ds.maxrecs[level]) { blks 558 fs/xfs/scrub/dabtree.c blks[level - 1].index++; blks 565 fs/xfs/scrub/dabtree.c key = xchk_da_btree_entry(&ds, level, blks[level].index); blks 582 fs/xfs/scrub/dabtree.c if (blks[level].bp == NULL) blks 585 fs/xfs/scrub/dabtree.c blks[level].index = 0; blks 591 fs/xfs/scrub/dabtree.c if (blks[level].bp == NULL) blks 593 fs/xfs/scrub/dabtree.c xfs_trans_brelse(sc->tp, blks[level].bp); blks 594 fs/xfs/scrub/dabtree.c blks[level].bp = NULL; blks 498 net/tipc/group.c int mtyp, blks; blks 524 net/tipc/group.c blks = msg_blocks(hdr); blks 574 net/tipc/group.c tipc_group_update_rcv_win(grp, blks, node, port, xmitq); blks 581 net/tipc/group.c void tipc_group_update_rcv_win(struct tipc_group *grp, int blks, u32 node, blks 594 net/tipc/group.c m->advertised -= blks; blks 71 net/tipc/group.h void tipc_group_update_rcv_win(struct tipc_group *grp, int blks, u32 node, blks 843 net/tipc/socket.c int blks = tsk_blocks(GROUP_H_SIZE + dlen); blks 870 net/tipc/socket.c tipc_group_update_member(mb, blks); blks 893 net/tipc/socket.c int blks = tsk_blocks(GROUP_H_SIZE + dlen); blks 909 net/tipc/socket.c !tipc_group_cong(tsk->group, node, port, blks, blks 939 net/tipc/socket.c int blks = tsk_blocks(GROUP_H_SIZE + dlen); blks 968 net/tipc/socket.c cong = tipc_group_cong(tsk->group, node, port, blks, blks 990 net/tipc/socket.c blks, &mbr)); blks 1027 net/tipc/socket.c int blks = tsk_blocks(MCAST_H_SIZE + dlen); blks 1036 net/tipc/socket.c !tipc_group_bc_cong(tsk->group, blks)); blks 1072 net/tipc/socket.c tipc_group_update_bc_members(tsk->group, blks, ack);