blocks 297 arch/alpha/kernel/osf_sys.c tmp.st_blocks = lstat->blocks; blocks 30 arch/arm/crypto/aes-ce-glue.c int rounds, int blocks); blocks 32 arch/arm/crypto/aes-ce-glue.c int rounds, int blocks); blocks 35 arch/arm/crypto/aes-ce-glue.c int rounds, int blocks, u8 iv[]); blocks 37 arch/arm/crypto/aes-ce-glue.c int rounds, int blocks, u8 iv[]); blocks 44 arch/arm/crypto/aes-ce-glue.c int rounds, int blocks, u8 ctr[]); blocks 182 arch/arm/crypto/aes-ce-glue.c unsigned int blocks; blocks 187 arch/arm/crypto/aes-ce-glue.c while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { blocks 190 arch/arm/crypto/aes-ce-glue.c ctx->key_enc, num_rounds(ctx), blocks); blocks 202 arch/arm/crypto/aes-ce-glue.c unsigned int blocks; blocks 207 arch/arm/crypto/aes-ce-glue.c while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { blocks 210 arch/arm/crypto/aes-ce-glue.c ctx->key_dec, num_rounds(ctx), blocks); blocks 222 arch/arm/crypto/aes-ce-glue.c unsigned int blocks; blocks 225 arch/arm/crypto/aes-ce-glue.c while ((blocks = (walk->nbytes / AES_BLOCK_SIZE))) { blocks 228 arch/arm/crypto/aes-ce-glue.c ctx->key_enc, num_rounds(ctx), blocks, blocks 252 arch/arm/crypto/aes-ce-glue.c unsigned int blocks; blocks 255 arch/arm/crypto/aes-ce-glue.c while ((blocks = (walk->nbytes / AES_BLOCK_SIZE))) { blocks 258 arch/arm/crypto/aes-ce-glue.c ctx->key_dec, num_rounds(ctx), blocks, blocks 398 arch/arm/crypto/aes-ce-glue.c int err, blocks; blocks 402 arch/arm/crypto/aes-ce-glue.c while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { blocks 405 arch/arm/crypto/aes-ce-glue.c ctx->key_enc, num_rounds(ctx), blocks, blocks 419 arch/arm/crypto/aes-ce-glue.c blocks = -1; blocks 423 arch/arm/crypto/aes-ce-glue.c blocks, walk.iv); blocks 30 arch/arm/crypto/aes-neonbs-glue.c int rounds, int blocks); blocks 32 arch/arm/crypto/aes-neonbs-glue.c int rounds, int blocks); blocks 35 arch/arm/crypto/aes-neonbs-glue.c int rounds, int blocks, u8 iv[]); blocks 38 arch/arm/crypto/aes-neonbs-glue.c int rounds, int blocks, u8 ctr[], u8 final[]); blocks 41 arch/arm/crypto/aes-neonbs-glue.c int rounds, int blocks, u8 iv[], int); blocks 43 arch/arm/crypto/aes-neonbs-glue.c int rounds, int blocks, u8 iv[], int); blocks 88 arch/arm/crypto/aes-neonbs-glue.c int rounds, int blocks)) blocks 98 arch/arm/crypto/aes-neonbs-glue.c unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; blocks 101 arch/arm/crypto/aes-neonbs-glue.c blocks = round_down(blocks, blocks 106 arch/arm/crypto/aes-neonbs-glue.c ctx->rounds, blocks); blocks 109 arch/arm/crypto/aes-neonbs-glue.c walk.nbytes - blocks * AES_BLOCK_SIZE); blocks 167 arch/arm/crypto/aes-neonbs-glue.c unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; blocks 170 arch/arm/crypto/aes-neonbs-glue.c blocks = round_down(blocks, blocks 175 arch/arm/crypto/aes-neonbs-glue.c ctx->key.rk, ctx->key.rounds, blocks, blocks 179 arch/arm/crypto/aes-neonbs-glue.c walk.nbytes - blocks * AES_BLOCK_SIZE); blocks 231 arch/arm/crypto/aes-neonbs-glue.c unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; blocks 235 arch/arm/crypto/aes-neonbs-glue.c blocks = round_down(blocks, blocks 242 arch/arm/crypto/aes-neonbs-glue.c ctx->rk, ctx->rounds, blocks, walk.iv, final); blocks 246 arch/arm/crypto/aes-neonbs-glue.c u8 *dst = walk.dst.virt.addr + blocks * AES_BLOCK_SIZE; blocks 247 arch/arm/crypto/aes-neonbs-glue.c u8 *src = walk.src.virt.addr + blocks * AES_BLOCK_SIZE; blocks 256 arch/arm/crypto/aes-neonbs-glue.c walk.nbytes - blocks * AES_BLOCK_SIZE); blocks 331 arch/arm/crypto/aes-neonbs-glue.c int rounds, int blocks, u8 iv[], int)) blocks 361 arch/arm/crypto/aes-neonbs-glue.c unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; blocks 365 arch/arm/crypto/aes-neonbs-glue.c blocks = round_down(blocks, blocks 372 arch/arm/crypto/aes-neonbs-glue.c ctx->key.rounds, blocks, walk.iv, reorder_last_tweak); blocks 375 arch/arm/crypto/aes-neonbs-glue.c walk.nbytes - blocks * AES_BLOCK_SIZE); blocks 48 arch/arm/crypto/ghash-ce-glue.c asmlinkage void pmull_ghash_update_p64(int blocks, u64 dg[], const char *src, blocks 52 arch/arm/crypto/ghash-ce-glue.c asmlinkage void pmull_ghash_update_p8(int blocks, u64 dg[], const char *src, blocks 56 arch/arm/crypto/ghash-ce-glue.c static void (*pmull_ghash_update)(int blocks, u64 dg[], const char *src, blocks 68 arch/arm/crypto/ghash-ce-glue.c static void ghash_do_update(int blocks, u64 dg[], const char *src, blocks 73 arch/arm/crypto/ghash-ce-glue.c pmull_ghash_update(blocks, dg, src, key, head); blocks 83 arch/arm/crypto/ghash-ce-glue.c blocks++; blocks 91 arch/arm/crypto/ghash-ce-glue.c } while (--blocks); blocks 108 arch/arm/crypto/ghash-ce-glue.c int blocks; blocks 118 arch/arm/crypto/ghash-ce-glue.c blocks = len / GHASH_BLOCK_SIZE; blocks 121 arch/arm/crypto/ghash-ce-glue.c ghash_do_update(blocks, ctx->digest, src, key, blocks 123 arch/arm/crypto/ghash-ce-glue.c src += blocks * GHASH_BLOCK_SIZE; blocks 27 arch/arm/crypto/sha1-ce-glue.c int blocks); blocks 28 arch/arm/crypto/sha2-ce-glue.c int blocks); blocks 28 arch/arm/crypto/sha512-glue.c asmlinkage void sha512_block_data_order(u64 *state, u8 const *src, int blocks); blocks 24 arch/arm/crypto/sha512-neon-glue.c int blocks); blocks 129 arch/arm/kernel/sys_oabi-compat.c tmp.st_blocks = stat->blocks; blocks 187 arch/arm64/crypto/aes-ce-ccm-glue.c int blocks = walk->nbytes / AES_BLOCK_SIZE; blocks 194 arch/arm64/crypto/aes-ce-ccm-glue.c blocks++; blocks 215 arch/arm64/crypto/aes-ce-ccm-glue.c } while (--blocks); blocks 75 arch/arm64/crypto/aes-glue.c int rounds, int blocks); blocks 77 arch/arm64/crypto/aes-glue.c int rounds, int blocks); blocks 80 arch/arm64/crypto/aes-glue.c int rounds, int blocks, u8 iv[]); blocks 82 arch/arm64/crypto/aes-glue.c int rounds, int blocks, u8 iv[]); blocks 90 arch/arm64/crypto/aes-glue.c int rounds, int blocks, u8 ctr[]); blocks 100 arch/arm64/crypto/aes-glue.c int rounds, int blocks, u8 iv[], blocks 103 arch/arm64/crypto/aes-glue.c int rounds, int blocks, u8 iv[], blocks 107 arch/arm64/crypto/aes-glue.c int blocks, u8 dg[], int enc_before, blocks 197 arch/arm64/crypto/aes-glue.c unsigned int blocks; blocks 201 arch/arm64/crypto/aes-glue.c while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { blocks 204 arch/arm64/crypto/aes-glue.c ctx->key_enc, rounds, blocks); blocks 217 arch/arm64/crypto/aes-glue.c unsigned int blocks; blocks 221 arch/arm64/crypto/aes-glue.c while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { blocks 224 arch/arm64/crypto/aes-glue.c ctx->key_dec, rounds, blocks); blocks 237 arch/arm64/crypto/aes-glue.c unsigned int blocks; blocks 239 arch/arm64/crypto/aes-glue.c while ((blocks = (walk->nbytes / AES_BLOCK_SIZE))) { blocks 242 arch/arm64/crypto/aes-glue.c ctx->key_enc, rounds, blocks, walk->iv); blocks 266 arch/arm64/crypto/aes-glue.c unsigned int blocks; blocks 268 arch/arm64/crypto/aes-glue.c while ((blocks = (walk->nbytes / AES_BLOCK_SIZE))) { blocks 271 arch/arm64/crypto/aes-glue.c ctx->key_dec, rounds, blocks, walk->iv); blocks 425 arch/arm64/crypto/aes-glue.c unsigned int blocks; blocks 429 arch/arm64/crypto/aes-glue.c blocks = walk.nbytes / AES_BLOCK_SIZE; blocks 430 arch/arm64/crypto/aes-glue.c if (blocks) { blocks 433 arch/arm64/crypto/aes-glue.c ctx->key1.key_enc, rounds, blocks, blocks 447 arch/arm64/crypto/aes-glue.c unsigned int blocks; blocks 451 arch/arm64/crypto/aes-glue.c blocks = walk.nbytes / AES_BLOCK_SIZE; blocks 452 arch/arm64/crypto/aes-glue.c if (blocks) { blocks 455 arch/arm64/crypto/aes-glue.c ctx->key1.key_dec, rounds, blocks, blocks 469 arch/arm64/crypto/aes-glue.c int blocks; blocks 473 arch/arm64/crypto/aes-glue.c while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { blocks 476 arch/arm64/crypto/aes-glue.c ctx->key_enc, rounds, blocks, walk.iv); blocks 489 arch/arm64/crypto/aes-glue.c blocks = -1; blocks 493 arch/arm64/crypto/aes-glue.c blocks, walk.iv); blocks 872 arch/arm64/crypto/aes-glue.c static void mac_do_update(struct crypto_aes_ctx *ctx, u8 const in[], int blocks, blocks 879 arch/arm64/crypto/aes-glue.c aes_mac_update(in, ctx->key_enc, rounds, blocks, dg, enc_before, blocks 886 arch/arm64/crypto/aes-glue.c while (blocks--) { blocks 890 arch/arm64/crypto/aes-glue.c if (blocks || enc_after) blocks 907 arch/arm64/crypto/aes-glue.c int blocks = len / AES_BLOCK_SIZE; blocks 911 arch/arm64/crypto/aes-glue.c mac_do_update(&tctx->key, p, blocks, ctx->dg, blocks 914 arch/arm64/crypto/aes-glue.c p += blocks * AES_BLOCK_SIZE; blocks 29 arch/arm64/crypto/aes-neonbs-glue.c int rounds, int blocks); blocks 31 arch/arm64/crypto/aes-neonbs-glue.c int rounds, int blocks); blocks 34 arch/arm64/crypto/aes-neonbs-glue.c int rounds, int blocks, u8 iv[]); blocks 37 arch/arm64/crypto/aes-neonbs-glue.c int rounds, int blocks, u8 iv[], u8 final[]); blocks 40 arch/arm64/crypto/aes-neonbs-glue.c int rounds, int blocks, u8 iv[]); blocks 42 arch/arm64/crypto/aes-neonbs-glue.c int rounds, int blocks, u8 iv[]); blocks 46 arch/arm64/crypto/aes-neonbs-glue.c int rounds, int blocks); blocks 48 arch/arm64/crypto/aes-neonbs-glue.c int rounds, int blocks, u8 iv[]); blocks 99 arch/arm64/crypto/aes-neonbs-glue.c int rounds, int blocks)) blocks 109 arch/arm64/crypto/aes-neonbs-glue.c unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; blocks 112 arch/arm64/crypto/aes-neonbs-glue.c blocks = round_down(blocks, blocks 117 arch/arm64/crypto/aes-neonbs-glue.c ctx->rounds, blocks); blocks 120 arch/arm64/crypto/aes-neonbs-glue.c walk.nbytes - blocks * AES_BLOCK_SIZE); blocks 168 arch/arm64/crypto/aes-neonbs-glue.c unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; blocks 173 arch/arm64/crypto/aes-neonbs-glue.c ctx->enc, ctx->key.rounds, blocks, blocks 191 arch/arm64/crypto/aes-neonbs-glue.c unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; blocks 194 arch/arm64/crypto/aes-neonbs-glue.c blocks = round_down(blocks, blocks 199 arch/arm64/crypto/aes-neonbs-glue.c ctx->key.rk, ctx->key.rounds, blocks, blocks 203 arch/arm64/crypto/aes-neonbs-glue.c walk.nbytes - blocks * AES_BLOCK_SIZE); blocks 239 arch/arm64/crypto/aes-neonbs-glue.c unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; blocks 243 arch/arm64/crypto/aes-neonbs-glue.c blocks = round_down(blocks, blocks 250 arch/arm64/crypto/aes-neonbs-glue.c ctx->rk, ctx->rounds, blocks, walk.iv, final); blocks 254 arch/arm64/crypto/aes-neonbs-glue.c u8 *dst = walk.dst.virt.addr + blocks * AES_BLOCK_SIZE; blocks 255 arch/arm64/crypto/aes-neonbs-glue.c u8 *src = walk.src.virt.addr + blocks * AES_BLOCK_SIZE; blocks 264 arch/arm64/crypto/aes-neonbs-glue.c walk.nbytes - blocks * AES_BLOCK_SIZE); blocks 319 arch/arm64/crypto/aes-neonbs-glue.c int rounds, int blocks, u8 iv[])) blocks 357 arch/arm64/crypto/aes-neonbs-glue.c unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; blocks 360 arch/arm64/crypto/aes-neonbs-glue.c blocks = round_down(blocks, blocks 368 arch/arm64/crypto/aes-neonbs-glue.c if (likely(blocks > 6)) { /* plain NEON is faster otherwise */ blocks 375 arch/arm64/crypto/aes-neonbs-glue.c fn(out, in, ctx->key.rk, ctx->key.rounds, blocks, blocks 378 arch/arm64/crypto/aes-neonbs-glue.c out += blocks * AES_BLOCK_SIZE; blocks 379 arch/arm64/crypto/aes-neonbs-glue.c in += blocks * AES_BLOCK_SIZE; blocks 380 arch/arm64/crypto/aes-neonbs-glue.c nbytes -= blocks * AES_BLOCK_SIZE; blocks 53 arch/arm64/crypto/ghash-ce-glue.c asmlinkage void pmull_ghash_update_p64(int blocks, u64 dg[], const char *src, blocks 57 arch/arm64/crypto/ghash-ce-glue.c asmlinkage void pmull_ghash_update_p8(int blocks, u64 dg[], const char *src, blocks 61 arch/arm64/crypto/ghash-ce-glue.c asmlinkage void pmull_gcm_encrypt(int blocks, u64 dg[], u8 dst[], blocks 66 arch/arm64/crypto/ghash-ce-glue.c asmlinkage void pmull_gcm_decrypt(int blocks, u64 dg[], u8 dst[], blocks 81 arch/arm64/crypto/ghash-ce-glue.c static void ghash_do_update(int blocks, u64 dg[], const char *src, blocks 83 arch/arm64/crypto/ghash-ce-glue.c void (*simd_update)(int blocks, u64 dg[], blocks 90 arch/arm64/crypto/ghash-ce-glue.c simd_update(blocks, dg, src, key, head); blocks 100 arch/arm64/crypto/ghash-ce-glue.c blocks++; blocks 108 arch/arm64/crypto/ghash-ce-glue.c } while (--blocks); blocks 120 arch/arm64/crypto/ghash-ce-glue.c void (*simd_update)(int blocks, u64 dg[], blocks 132 arch/arm64/crypto/ghash-ce-glue.c int blocks; blocks 142 arch/arm64/crypto/ghash-ce-glue.c blocks = len / GHASH_BLOCK_SIZE; blocks 146 arch/arm64/crypto/ghash-ce-glue.c int chunk = min(blocks, MAX_BLOCKS); blocks 152 arch/arm64/crypto/ghash-ce-glue.c blocks -= chunk; blocks 155 arch/arm64/crypto/ghash-ce-glue.c } while (unlikely(blocks > 0)); blocks 348 arch/arm64/crypto/ghash-ce-glue.c int blocks = count / GHASH_BLOCK_SIZE; blocks 350 arch/arm64/crypto/ghash-ce-glue.c ghash_do_update(blocks, dg, src, &ctx->ghash_key, blocks 354 arch/arm64/crypto/ghash-ce-glue.c src += blocks * GHASH_BLOCK_SIZE; blocks 451 arch/arm64/crypto/ghash-ce-glue.c int blocks = walk.nbytes / (2 * AES_BLOCK_SIZE) * 2; blocks 456 arch/arm64/crypto/ghash-ce-glue.c pmull_gcm_encrypt(blocks, dg, walk.dst.virt.addr, blocks 471 arch/arm64/crypto/ghash-ce-glue.c const int blocks = blocks 475 arch/arm64/crypto/ghash-ce-glue.c int remaining = blocks; blocks 486 arch/arm64/crypto/ghash-ce-glue.c ghash_do_update(blocks, dg, blocks 567 arch/arm64/crypto/ghash-ce-glue.c int blocks = walk.nbytes / (2 * AES_BLOCK_SIZE) * 2; blocks 568 arch/arm64/crypto/ghash-ce-glue.c int rem = walk.total - blocks * AES_BLOCK_SIZE; blocks 573 arch/arm64/crypto/ghash-ce-glue.c pmull_gcm_decrypt(blocks, dg, walk.dst.virt.addr, blocks 605 arch/arm64/crypto/ghash-ce-glue.c int blocks = walk.nbytes / (2 * AES_BLOCK_SIZE) * 2; blocks 609 arch/arm64/crypto/ghash-ce-glue.c ghash_do_update(blocks, dg, walk.src.virt.addr, blocks 620 arch/arm64/crypto/ghash-ce-glue.c } while (--blocks > 0); blocks 29 arch/arm64/crypto/sha1-ce-glue.c int blocks); blocks 29 arch/arm64/crypto/sha2-ce-glue.c int blocks); blocks 36 arch/arm64/crypto/sha2-ce-glue.c asmlinkage void sha256_block_data_order(u32 *digest, u8 const *src, int blocks); blocks 27 arch/arm64/crypto/sha3-ce-glue.c asmlinkage void sha3_ce_transform(u64 *st, const u8 *data, int blocks, blocks 40 arch/arm64/crypto/sha3-ce-glue.c int blocks; blocks 55 arch/arm64/crypto/sha3-ce-glue.c blocks = len / sctx->rsiz; blocks 58 arch/arm64/crypto/sha3-ce-glue.c if (blocks) { blocks 60 arch/arm64/crypto/sha3-ce-glue.c sha3_ce_transform(sctx->st, data, blocks, digest_size); blocks 62 arch/arm64/crypto/sha3-ce-glue.c data += blocks * sctx->rsiz; blocks 28 arch/arm64/crypto/sha512-ce-glue.c int blocks); blocks 30 arch/arm64/crypto/sha512-ce-glue.c asmlinkage void sha512_block_data_order(u64 *digest, u8 const *src, int blocks); blocks 24 arch/arm64/crypto/sm3-ce-glue.c int blocks); blocks 41 arch/m68k/emu/nfblock.c static inline s32 nfhd_get_capacity(u32 major, u32 minor, u32 *blocks, blocks 45 arch/m68k/emu/nfblock.c virt_to_phys(blocks), virt_to_phys(blocksize)); blocks 56 arch/m68k/emu/nfblock.c u32 blocks, bsize; blocks 87 arch/m68k/emu/nfblock.c geo->cylinders = dev->blocks >> (6 - dev->bshift); blocks 99 arch/m68k/emu/nfblock.c static int __init nfhd_init_one(int id, u32 blocks, u32 bsize) blocks 105 arch/m68k/emu/nfblock.c blocks, bsize); blocks 117 arch/m68k/emu/nfblock.c dev->blocks = blocks; blocks 138 arch/m68k/emu/nfblock.c set_capacity(dev->disk, (sector_t)blocks * (bsize / 512)); blocks 157 arch/m68k/emu/nfblock.c u32 blocks, bsize; blocks 175 arch/m68k/emu/nfblock.c if (nfhd_get_capacity(i, 0, &blocks, &bsize)) blocks 177 arch/m68k/emu/nfblock.c nfhd_init_one(i, blocks, bsize); blocks 19 arch/powerpc/crypto/md5-glue.c extern void ppc_md5_transform(u32 *state, const u8 *src, u32 blocks); blocks 33 arch/powerpc/crypto/sha1-spe-glue.c extern void ppc_spe_sha1_transform(u32 *state, const u8 *src, u32 blocks); blocks 34 arch/powerpc/crypto/sha256-spe-glue.c extern void ppc_spe_sha256_transform(u32 *state, const u8 *src, u32 blocks); blocks 110 arch/powerpc/kernel/rtas_flash.c struct flash_block blocks[FLASH_BLOCKS_PER_NODE]; blocks 173 arch/powerpc/kernel/rtas_flash.c if (f->blocks[i].data == NULL) { blocks 176 arch/powerpc/kernel/rtas_flash.c block_size = f->blocks[i].length; blocks 201 arch/powerpc/kernel/rtas_flash.c kmem_cache_free(flash_block_cache, f->blocks[i].data); blocks 357 arch/powerpc/kernel/rtas_flash.c fl->blocks[next_free].data = p; blocks 358 arch/powerpc/kernel/rtas_flash.c fl->blocks[next_free].length = count; blocks 610 arch/powerpc/kernel/rtas_flash.c f->blocks[i].data = (char *)cpu_to_be64(__pa(f->blocks[i].data)); blocks 611 arch/powerpc/kernel/rtas_flash.c image_size += f->blocks[i].length; blocks 612 arch/powerpc/kernel/rtas_flash.c f->blocks[i].length = cpu_to_be64(f->blocks[i].length); blocks 29 arch/powerpc/platforms/pseries/papr_scm.c uint64_t blocks; blocks 61 arch/powerpc/platforms/pseries/papr_scm.c p->blocks, BIND_ANY_ADDR, token); blocks 127 arch/powerpc/platforms/pseries/papr_scm.c p->drc_index, p->blocks - 1); blocks 132 arch/powerpc/platforms/pseries/papr_scm.c if ((end_addr - start_addr) != ((p->blocks - 1) * p->block_size)) blocks 367 arch/powerpc/platforms/pseries/papr_scm.c mapping.size = p->blocks * p->block_size; // XXX: potential overflow? blocks 407 arch/powerpc/platforms/pseries/papr_scm.c u64 blocks, block_size; blocks 424 arch/powerpc/platforms/pseries/papr_scm.c if (of_property_read_u64(dn, "ibm,number-of-blocks", &blocks)) { blocks 445 arch/powerpc/platforms/pseries/papr_scm.c p->blocks = blocks; blocks 479 arch/powerpc/platforms/pseries/papr_scm.c p->res.end = p->bound_addr + p->blocks * p->block_size - 1; blocks 142 arch/s390/kernel/compat_linux.c tmp.st_blocks = (u32)stat->blocks; blocks 82 arch/sparc/kernel/sys_sparc32.c err |= put_user(stat->blocks, &statbuf->st_blocks); blocks 321 arch/um/os-Linux/file.c long blocks; blocks 330 arch/um/os-Linux/file.c if (ioctl(fd, BLKGETSIZE, &blocks) < 0) { blocks 337 arch/um/os-Linux/file.c *size_out = ((long long) blocks) * 512; blocks 52 arch/x86/crypto/aegis128-aesni-glue.c struct aegis_block blocks[AEGIS128_STATE_BLOCKS]; blocks 31 arch/x86/crypto/poly1305_glue.c const u32 *r, unsigned int blocks); blocks 33 arch/x86/crypto/poly1305_glue.c unsigned int blocks, const u32 *u); blocks 36 arch/x86/crypto/poly1305_glue.c unsigned int blocks, const u32 *u); blocks 67 arch/x86/crypto/poly1305_glue.c unsigned int blocks, datalen; blocks 92 arch/x86/crypto/poly1305_glue.c blocks = srclen / (POLY1305_BLOCK_SIZE * 4); blocks 93 arch/x86/crypto/poly1305_glue.c poly1305_4block_avx2(dctx->h.h, src, dctx->r.r, blocks, blocks 95 arch/x86/crypto/poly1305_glue.c src += POLY1305_BLOCK_SIZE * 4 * blocks; blocks 96 arch/x86/crypto/poly1305_glue.c srclen -= POLY1305_BLOCK_SIZE * 4 * blocks; blocks 105 arch/x86/crypto/poly1305_glue.c blocks = srclen / (POLY1305_BLOCK_SIZE * 2); blocks 106 arch/x86/crypto/poly1305_glue.c poly1305_2block_sse2(dctx->h.h, src, dctx->r.r, blocks, blocks 108 arch/x86/crypto/poly1305_glue.c src += POLY1305_BLOCK_SIZE * 2 * blocks; blocks 109 arch/x86/crypto/poly1305_glue.c srclen -= POLY1305_BLOCK_SIZE * 2 * blocks; blocks 95 arch/x86/ia32/sys_ia32.c __put_user(stat->blocks, &ubuf->st_blocks)) blocks 57 arch/x86/include/asm/amd_nb.h struct threshold_block *blocks; blocks 1023 arch/x86/kernel/cpu/mce/amd.c first_block = per_cpu(threshold_banks, cpu)[bank]->blocks; blocks 1248 arch/x86/kernel/cpu/mce/amd.c if (tb->blocks) blocks 1249 arch/x86/kernel/cpu/mce/amd.c list_add(&b->miscj, &tb->blocks->miscj); blocks 1251 arch/x86/kernel/cpu/mce/amd.c tb->blocks = b; blocks 1281 arch/x86/kernel/cpu/mce/amd.c struct list_head *head = &b->blocks->miscj; blocks 1286 arch/x86/kernel/cpu/mce/amd.c err = kobject_add(&b->blocks->kobj, b->kobj, b->blocks->kobj.name); blocks 1385 arch/x86/kernel/cpu/mce/amd.c list_for_each_entry_safe(pos, tmp, &head->blocks->miscj, miscj) { blocks 1390 arch/x86/kernel/cpu/mce/amd.c kobject_put(&head->blocks->kobj); blocks 1400 arch/x86/kernel/cpu/mce/amd.c list_for_each_entry_safe(pos, tmp, &b->blocks->miscj, miscj) blocks 1413 arch/x86/kernel/cpu/mce/amd.c if (!b->blocks) blocks 36 block/partitions/sgi.c unsigned int start, blocks; blocks 71 block/partitions/sgi.c blocks = be32_to_cpu(p->num_blocks); blocks 73 block/partitions/sgi.c if (blocks) { blocks 74 block/partitions/sgi.c put_partition(state, slot, start, blocks); blocks 31 crypto/aegis128-core.c union aegis_block blocks[AEGIS128_STATE_BLOCKS]; blocks 80 crypto/aegis128-core.c tmp = state->blocks[AEGIS128_STATE_BLOCKS - 1]; blocks 82 crypto/aegis128-core.c crypto_aegis_aesenc(&state->blocks[i], &state->blocks[i - 1], blocks 83 crypto/aegis128-core.c &state->blocks[i]); blocks 84 crypto/aegis128-core.c crypto_aegis_aesenc(&state->blocks[0], &tmp, &state->blocks[0]); blocks 96 crypto/aegis128-core.c crypto_aegis_block_xor(&state->blocks[0], msg); blocks 107 crypto/aegis128-core.c crypto_xor(state->blocks[0].bytes, msg, AEGIS_BLOCK_SIZE); blocks 120 crypto/aegis128-core.c state->blocks[0] = key_iv; blocks 121 crypto/aegis128-core.c state->blocks[1] = crypto_aegis_const[1]; blocks 122 crypto/aegis128-core.c state->blocks[2] = crypto_aegis_const[0]; blocks 123 crypto/aegis128-core.c state->blocks[3] = *key; blocks 124 crypto/aegis128-core.c state->blocks[4] = *key; blocks 126 crypto/aegis128-core.c crypto_aegis_block_xor(&state->blocks[3], &crypto_aegis_const[0]); blocks 127 crypto/aegis128-core.c crypto_aegis_block_xor(&state->blocks[4], &crypto_aegis_const[1]); blocks 170 crypto/aegis128-core.c tmp = state->blocks[2]; blocks 171 crypto/aegis128-core.c crypto_aegis_block_and(&tmp, &state->blocks[3]); blocks 172 crypto/aegis128-core.c crypto_aegis_block_xor(&tmp, &state->blocks[4]); blocks 173 crypto/aegis128-core.c crypto_aegis_block_xor(&tmp, &state->blocks[1]); blocks 186 crypto/aegis128-core.c tmp = state->blocks[2]; blocks 187 crypto/aegis128-core.c crypto_aegis_block_and(&tmp, &state->blocks[3]); blocks 188 crypto/aegis128-core.c crypto_aegis_block_xor(&tmp, &state->blocks[4]); blocks 189 crypto/aegis128-core.c crypto_aegis_block_xor(&tmp, &state->blocks[1]); blocks 206 crypto/aegis128-core.c tmp = state->blocks[2]; blocks 207 crypto/aegis128-core.c crypto_aegis_block_and(&tmp, &state->blocks[3]); blocks 208 crypto/aegis128-core.c crypto_aegis_block_xor(&tmp, &state->blocks[4]); blocks 209 crypto/aegis128-core.c crypto_aegis_block_xor(&tmp, &state->blocks[1]); blocks 231 crypto/aegis128-core.c tmp = state->blocks[2]; blocks 232 crypto/aegis128-core.c crypto_aegis_block_and(&tmp, &state->blocks[3]); blocks 233 crypto/aegis128-core.c crypto_aegis_block_xor(&tmp, &state->blocks[4]); blocks 234 crypto/aegis128-core.c crypto_aegis_block_xor(&tmp, &state->blocks[1]); blocks 247 crypto/aegis128-core.c tmp = state->blocks[2]; blocks 248 crypto/aegis128-core.c crypto_aegis_block_and(&tmp, &state->blocks[3]); blocks 249 crypto/aegis128-core.c crypto_aegis_block_xor(&tmp, &state->blocks[4]); blocks 250 crypto/aegis128-core.c crypto_aegis_block_xor(&tmp, &state->blocks[1]); blocks 267 crypto/aegis128-core.c tmp = state->blocks[2]; blocks 268 crypto/aegis128-core.c crypto_aegis_block_and(&tmp, &state->blocks[3]); blocks 269 crypto/aegis128-core.c crypto_aegis_block_xor(&tmp, &state->blocks[4]); blocks 270 crypto/aegis128-core.c crypto_aegis_block_xor(&tmp, &state->blocks[1]); blocks 360 crypto/aegis128-core.c crypto_aegis_block_xor(&tmp, &state->blocks[3]); blocks 366 crypto/aegis128-core.c crypto_aegis_block_xor(tag_xor, &state->blocks[i]); blocks 107 crypto/async_tx/async_pq.c do_sync_gen_syndrome(struct page **blocks, unsigned int offset, int disks, blocks 117 crypto/async_tx/async_pq.c srcs = (void **) blocks; blocks 120 crypto/async_tx/async_pq.c if (blocks[i] == NULL) { blocks 124 crypto/async_tx/async_pq.c srcs[i] = page_address(blocks[i]) + offset; blocks 163 crypto/async_tx/async_pq.c async_gen_syndrome(struct page **blocks, unsigned int offset, int disks, blocks 168 crypto/async_tx/async_pq.c &P(blocks, disks), 2, blocks 169 crypto/async_tx/async_pq.c blocks, src_cnt, len); blocks 173 crypto/async_tx/async_pq.c BUG_ON(disks > MAX_DISKS || !(P(blocks, disks) || Q(blocks, disks))); blocks 197 crypto/async_tx/async_pq.c if (blocks[i] == NULL) blocks 199 crypto/async_tx/async_pq.c unmap->addr[j] = dma_map_page(device->dev, blocks[i], offset, blocks 211 crypto/async_tx/async_pq.c if (P(blocks, disks)) blocks 212 crypto/async_tx/async_pq.c unmap->addr[j++] = dma_map_page(device->dev, P(blocks, disks), blocks 220 crypto/async_tx/async_pq.c if (Q(blocks, disks)) blocks 221 crypto/async_tx/async_pq.c unmap->addr[j++] = dma_map_page(device->dev, Q(blocks, disks), blocks 241 crypto/async_tx/async_pq.c if (!P(blocks, disks)) { blocks 242 crypto/async_tx/async_pq.c P(blocks, disks) = pq_scribble_page; blocks 245 crypto/async_tx/async_pq.c if (!Q(blocks, disks)) { blocks 246 crypto/async_tx/async_pq.c Q(blocks, disks) = pq_scribble_page; blocks 249 crypto/async_tx/async_pq.c do_sync_gen_syndrome(blocks, offset, disks, len, submit); blocks 256 crypto/async_tx/async_pq.c pq_val_chan(struct async_submit_ctl *submit, struct page **blocks, int disks, size_t len) blocks 261 crypto/async_tx/async_pq.c return async_tx_find_channel(submit, DMA_PQ_VAL, NULL, 0, blocks, blocks 281 crypto/async_tx/async_pq.c async_syndrome_val(struct page **blocks, unsigned int offset, int disks, blocks 285 crypto/async_tx/async_pq.c struct dma_chan *chan = pq_val_chan(submit, blocks, disks, len); blocks 308 crypto/async_tx/async_pq.c if (likely(blocks[i])) { blocks 309 crypto/async_tx/async_pq.c unmap->addr[j] = dma_map_page(dev, blocks[i], blocks 318 crypto/async_tx/async_pq.c if (!P(blocks, disks)) { blocks 322 crypto/async_tx/async_pq.c pq[0] = dma_map_page(dev, P(blocks, disks), blocks 328 crypto/async_tx/async_pq.c if (!Q(blocks, disks)) { blocks 332 crypto/async_tx/async_pq.c pq[1] = dma_map_page(dev, Q(blocks, disks), blocks 357 crypto/async_tx/async_pq.c struct page *p_src = P(blocks, disks); blocks 358 crypto/async_tx/async_pq.c struct page *q_src = Q(blocks, disks); blocks 384 crypto/async_tx/async_pq.c tx = async_xor(spare, blocks, offset, disks-2, len, submit); blocks 392 crypto/async_tx/async_pq.c P(blocks, disks) = NULL; blocks 393 crypto/async_tx/async_pq.c Q(blocks, disks) = spare; blocks 395 crypto/async_tx/async_pq.c tx = async_gen_syndrome(blocks, offset, disks, len, submit); blocks 403 crypto/async_tx/async_pq.c P(blocks, disks) = p_src; blocks 404 crypto/async_tx/async_pq.c Q(blocks, disks) = q_src; blocks 147 crypto/async_tx/async_raid6_recov.c struct page **blocks, struct async_submit_ctl *submit) blocks 158 crypto/async_tx/async_raid6_recov.c p = blocks[disks-2]; blocks 159 crypto/async_tx/async_raid6_recov.c q = blocks[disks-1]; blocks 161 crypto/async_tx/async_raid6_recov.c a = blocks[faila]; blocks 162 crypto/async_tx/async_raid6_recov.c b = blocks[failb]; blocks 186 crypto/async_tx/async_raid6_recov.c struct page **blocks, struct async_submit_ctl *submit) blocks 201 crypto/async_tx/async_raid6_recov.c if (blocks[i] == NULL) blocks 210 crypto/async_tx/async_raid6_recov.c p = blocks[disks-2]; blocks 211 crypto/async_tx/async_raid6_recov.c q = blocks[disks-1]; blocks 212 crypto/async_tx/async_raid6_recov.c g = blocks[good]; blocks 218 crypto/async_tx/async_raid6_recov.c dp = blocks[faila]; blocks 219 crypto/async_tx/async_raid6_recov.c dq = blocks[failb]; blocks 260 crypto/async_tx/async_raid6_recov.c struct page **blocks, struct async_submit_ctl *submit) blocks 271 crypto/async_tx/async_raid6_recov.c p = blocks[disks-2]; blocks 272 crypto/async_tx/async_raid6_recov.c q = blocks[disks-1]; blocks 278 crypto/async_tx/async_raid6_recov.c dp = blocks[faila]; blocks 279 crypto/async_tx/async_raid6_recov.c blocks[faila] = NULL; blocks 280 crypto/async_tx/async_raid6_recov.c blocks[disks-2] = dp; blocks 281 crypto/async_tx/async_raid6_recov.c dq = blocks[failb]; blocks 282 crypto/async_tx/async_raid6_recov.c blocks[failb] = NULL; blocks 283 crypto/async_tx/async_raid6_recov.c blocks[disks-1] = dq; blocks 286 crypto/async_tx/async_raid6_recov.c tx = async_gen_syndrome(blocks, 0, disks, bytes, submit); blocks 289 crypto/async_tx/async_raid6_recov.c blocks[faila] = dp; blocks 290 crypto/async_tx/async_raid6_recov.c blocks[failb] = dq; blocks 291 crypto/async_tx/async_raid6_recov.c blocks[disks-2] = p; blocks 292 crypto/async_tx/async_raid6_recov.c blocks[disks-1] = q; blocks 337 crypto/async_tx/async_raid6_recov.c struct page **blocks, struct async_submit_ctl *submit) blocks 354 crypto/async_tx/async_raid6_recov.c void **ptrs = scribble ? scribble : (void **) blocks; blocks 358 crypto/async_tx/async_raid6_recov.c if (blocks[i] == NULL) blocks 361 crypto/async_tx/async_raid6_recov.c ptrs[i] = page_address(blocks[i]); blocks 372 crypto/async_tx/async_raid6_recov.c if (blocks[i]) blocks 386 crypto/async_tx/async_raid6_recov.c return __2data_recov_4(disks, bytes, faila, failb, blocks, submit); blocks 393 crypto/async_tx/async_raid6_recov.c return __2data_recov_5(disks, bytes, faila, failb, blocks, submit); blocks 395 crypto/async_tx/async_raid6_recov.c return __2data_recov_n(disks, bytes, faila, failb, blocks, submit); blocks 410 crypto/async_tx/async_raid6_recov.c struct page **blocks, struct async_submit_ctl *submit) blocks 430 crypto/async_tx/async_raid6_recov.c void **ptrs = scribble ? scribble : (void **) blocks; blocks 434 crypto/async_tx/async_raid6_recov.c if (blocks[i] == NULL) blocks 437 crypto/async_tx/async_raid6_recov.c ptrs[i] = page_address(blocks[i]); blocks 451 crypto/async_tx/async_raid6_recov.c if (blocks[i]) { blocks 460 crypto/async_tx/async_raid6_recov.c p = blocks[disks-2]; blocks 461 crypto/async_tx/async_raid6_recov.c q = blocks[disks-1]; blocks 466 crypto/async_tx/async_raid6_recov.c dq = blocks[faila]; blocks 467 crypto/async_tx/async_raid6_recov.c blocks[faila] = NULL; blocks 468 crypto/async_tx/async_raid6_recov.c blocks[disks-1] = dq; blocks 474 crypto/async_tx/async_raid6_recov.c struct page *g = blocks[good]; blocks 486 crypto/async_tx/async_raid6_recov.c tx = async_gen_syndrome(blocks, 0, disks, bytes, submit); blocks 490 crypto/async_tx/async_raid6_recov.c blocks[faila] = dq; blocks 491 crypto/async_tx/async_raid6_recov.c blocks[disks-1] = q; blocks 71 crypto/async_tx/raid6test.c struct page *blocks[NDISKS]; blocks 84 crypto/async_tx/raid6test.c blocks[count++] = ptrs[i]; blocks 89 crypto/async_tx/raid6test.c tx = async_xor(dest, blocks, 0, count, bytes, &submit); blocks 32 crypto/sha1_generic.c int blocks) blocks 36 crypto/sha1_generic.c while (blocks--) { blocks 152 crypto/sha512_generic.c int blocks) blocks 154 crypto/sha512_generic.c while (blocks--) { blocks 137 crypto/sm3_generic.c int blocks) blocks 139 crypto/sm3_generic.c while (blocks--) { blocks 400 crypto/vmac.c const __le64 *mptr, unsigned int blocks) blocks 415 crypto/vmac.c blocks--; blocks 418 crypto/vmac.c while (blocks--) { blocks 196 drivers/block/amiflop.c int blocks; /* total # blocks on disk */ blocks 1586 drivers/block/amiflop.c getprm.size=p->blocks; blocks 1696 drivers/block/amiflop.c unit[drive].blocks=unit[drive].type->heads*unit[drive].type->tracks* blocks 1698 drivers/block/amiflop.c set_capacity(unit[drive].gendisk, unit[drive].blocks); blocks 171 drivers/block/ataflop.c unsigned blocks; /* total number of blocks */ blocks 750 drivers/block/ataflop.c if (!UDT || desc->track >= UDT->blocks/UDT->spt/2 || desc->head >= 2) { blocks 1154 drivers/block/ataflop.c if (SUDT[-1].blocks > ReqBlock) { blocks 1158 drivers/block/ataflop.c SUDT->blocks); blocks 1173 drivers/block/ataflop.c SUDT->blocks); blocks 1445 drivers/block/ataflop.c if (default_params[drive].blocks == 0) blocks 1518 drivers/block/ataflop.c set_capacity(floppy->disk, UDT->blocks); blocks 1536 drivers/block/ataflop.c set_capacity(floppy->disk, UDT->blocks); blocks 1593 drivers/block/ataflop.c getprm.size = dtp->blocks; blocks 1596 drivers/block/ataflop.c getprm.track = dtp->blocks/dtp->spt/2; blocks 1654 drivers/block/ataflop.c if ( dtp->blocks == setprm.size blocks 1661 drivers/block/ataflop.c set_capacity(floppy->disk, UDT->blocks); blocks 1667 drivers/block/ataflop.c default_params[drive].blocks = dtp->blocks; blocks 1687 drivers/block/ataflop.c dtp->blocks = setprm.size; blocks 1697 drivers/block/ataflop.c drive, dtp->blocks, dtp->spt, dtp->stretch); blocks 1700 drivers/block/ataflop.c if (setprm.track != dtp->blocks/dtp->spt/2 || blocks 1705 drivers/block/ataflop.c set_capacity(floppy->disk, UDT->blocks); blocks 1727 drivers/block/ataflop.c default_params[drive].blocks = 0; blocks 74 drivers/block/rsxx/dev.c u64 blocks = card->size8 >> 9; blocks 83 drivers/block/rsxx/dev.c do_div(blocks, (geo->heads * geo->sectors)); blocks 84 drivers/block/rsxx/dev.c geo->cylinders = blocks; blocks 1017 drivers/crypto/atmel-aes.c u16 blocks, start, end; blocks 1027 drivers/crypto/atmel-aes.c blocks = DIV_ROUND_UP(datalen, AES_BLOCK_SIZE); blocks 1032 drivers/crypto/atmel-aes.c end = start + blocks - 1; blocks 1034 drivers/crypto/atmel-aes.c if (blocks >> 16 || end < start) { blocks 576 drivers/crypto/mediatek/mtk-aes.c u32 start, end, ctr, blocks; blocks 587 drivers/crypto/mediatek/mtk-aes.c blocks = DIV_ROUND_UP(datalen, AES_BLOCK_SIZE); blocks 592 drivers/crypto/mediatek/mtk-aes.c end = start + blocks - 1; blocks 150 drivers/edac/edac_device.c inst->blocks = blk_p; blocks 584 drivers/edac/edac_device.c block = instance->blocks + block_nr; blocks 626 drivers/edac/edac_device.c block = instance->blocks + block_nr; blocks 140 drivers/edac/edac_device.h struct edac_device_block *blocks; /* block array */ blocks 648 drivers/edac/edac_device_sysfs.c &instance->blocks[i]); blocks 653 drivers/edac/edac_device_sysfs.c &instance->blocks[j]); blocks 686 drivers/edac/edac_device_sysfs.c edac_device_delete_block(edac_dev, &instance->blocks[i]); blocks 90 drivers/fpga/altera-cvp.c int (*wait_credit)(struct fpga_manager *mgr, u32 blocks); blocks 230 drivers/fpga/altera-cvp.c u32 blocks) blocks 249 drivers/fpga/altera-cvp.c ret = altera_cvp_chk_error(mgr, blocks * ALTERA_CVP_V2_SIZE); blocks 48 drivers/gpio/gpio-sch311x.c struct sch311x_gpio_block blocks[6]; blocks 283 drivers/gpio/gpio-sch311x.c for (i = 0; i < ARRAY_SIZE(priv->blocks); i++) { blocks 284 drivers/gpio/gpio-sch311x.c block = &priv->blocks[i]; blocks 320 drivers/gpio/gpio-sch311x.c gpiochip_remove(&priv->blocks[i].chip); blocks 329 drivers/gpio/gpio-sch311x.c for (i = 0; i < ARRAY_SIZE(priv->blocks); i++) { blocks 330 drivers/gpio/gpio-sch311x.c gpiochip_remove(&priv->blocks[i].chip); blocks 337 drivers/gpu/drm/i915/i915_buddy.c struct list_head *blocks, blocks 406 drivers/gpu/drm/i915/i915_buddy.c list_splice_tail(&allocated, blocks); blocks 121 drivers/gpu/drm/i915/i915_buddy.h struct list_head *blocks, blocks 4350 drivers/gpu/drm/i915/intel_pm.c u32 blocks; blocks 4397 drivers/gpu/drm/i915/intel_pm.c blocks = 0; blocks 4405 drivers/gpu/drm/i915/intel_pm.c blocks = U32_MAX; blocks 4411 drivers/gpu/drm/i915/intel_pm.c blocks += wm->wm[level].min_ddb_alloc; blocks 4412 drivers/gpu/drm/i915/intel_pm.c blocks += wm->uv_wm[level].min_ddb_alloc; blocks 4415 drivers/gpu/drm/i915/intel_pm.c if (blocks <= alloc_size) { blocks 4416 drivers/gpu/drm/i915/intel_pm.c alloc_size -= blocks; blocks 4423 drivers/gpu/drm/i915/intel_pm.c DRM_DEBUG_KMS("minimum required %d/%d\n", blocks, blocks 114 drivers/gpu/drm/i915/selftests/i915_buddy.c struct list_head *blocks, blocks 127 drivers/gpu/drm/i915/selftests/i915_buddy.c list_for_each_entry(block, blocks, link) { blocks 319 drivers/gpu/drm/i915/selftests/i915_buddy.c LIST_HEAD(blocks); blocks 354 drivers/gpu/drm/i915/selftests/i915_buddy.c list_add_tail(&block->link, &blocks); blocks 366 drivers/gpu/drm/i915/selftests/i915_buddy.c err = igt_check_blocks(&mm, &blocks, total, false); blocks 368 drivers/gpu/drm/i915/selftests/i915_buddy.c i915_buddy_free_list(&mm, &blocks); blocks 394 drivers/gpu/drm/i915/selftests/i915_buddy.c LIST_HEAD(blocks); blocks 419 drivers/gpu/drm/i915/selftests/i915_buddy.c list_add_tail(&block->link, &blocks); blocks 429 drivers/gpu/drm/i915/selftests/i915_buddy.c list_add_tail(&block->link, &blocks); blocks 437 drivers/gpu/drm/i915/selftests/i915_buddy.c list_add_tail(&block->link, &blocks); blocks 443 drivers/gpu/drm/i915/selftests/i915_buddy.c block = list_last_entry(&blocks, typeof(*block), link); blocks 449 drivers/gpu/drm/i915/selftests/i915_buddy.c list_for_each_entry_safe(block, bn, &blocks, link) { blocks 475 drivers/gpu/drm/i915/selftests/i915_buddy.c i915_buddy_free_list(&mm, &blocks); blocks 485 drivers/gpu/drm/i915/selftests/i915_buddy.c LIST_HEAD(blocks); blocks 512 drivers/gpu/drm/i915/selftests/i915_buddy.c list_add_tail(&block->link, &blocks); blocks 519 drivers/gpu/drm/i915/selftests/i915_buddy.c list_add_tail(&block->link, &blocks); blocks 525 drivers/gpu/drm/i915/selftests/i915_buddy.c i915_buddy_free_list(&mm, &blocks); blocks 535 drivers/gpu/drm/i915/selftests/i915_buddy.c LIST_HEAD(blocks); blocks 556 drivers/gpu/drm/i915/selftests/i915_buddy.c block = list_first_entry_or_null(&blocks, typeof(*block), link); blocks 570 drivers/gpu/drm/i915/selftests/i915_buddy.c list_add_tail(&block->link, &blocks); blocks 586 drivers/gpu/drm/i915/selftests/i915_buddy.c list_add_tail(&block->link, &blocks); blocks 600 drivers/gpu/drm/i915/selftests/i915_buddy.c list_add_tail(&block->link, &blocks); blocks 607 drivers/gpu/drm/i915/selftests/i915_buddy.c list_splice_tail(&holes, &blocks); blocks 608 drivers/gpu/drm/i915/selftests/i915_buddy.c i915_buddy_free_list(&mm, &blocks); blocks 617 drivers/gpu/drm/i915/selftests/i915_buddy.c LIST_HEAD(blocks); blocks 680 drivers/gpu/drm/i915/selftests/i915_buddy.c list_splice_tail(&tmp, &blocks); blocks 695 drivers/gpu/drm/i915/selftests/i915_buddy.c i915_buddy_free_list(&mm, &blocks); blocks 71 drivers/gpu/drm/nouveau/dispnv50/atom.h u16 blocks:12; blocks 202 drivers/gpu/drm/nouveau/dispnv50/atom.h u16 blocks[3]; blocks 82 drivers/gpu/drm/nouveau/dispnv50/base507c.c asyw->image.blocks[0] << 8 | blocks 49 drivers/gpu/drm/nouveau/dispnv50/base827c.c asyw->image.blocks[0] << 8 | blocks 40 drivers/gpu/drm/nouveau/dispnv50/base907c.c asyw->image.blocks[0] << 8 | blocks 182 drivers/gpu/drm/nouveau/dispnv50/head507d.c asyh->core.blocks << 8 | blocks 228 drivers/gpu/drm/nouveau/dispnv50/head507d.c asyh->core.blocks = 0; blocks 67 drivers/gpu/drm/nouveau/dispnv50/head827d.c asyh->core.blocks << 8 | blocks 177 drivers/gpu/drm/nouveau/dispnv50/head907d.c asyh->core.blocks << 8 | blocks 86 drivers/gpu/drm/nouveau/dispnv50/ovly507e.c asyw->image.blocks[0] << 8 | blocks 44 drivers/gpu/drm/nouveau/dispnv50/ovly827e.c asyw->image.blocks[0] << 8 | blocks 42 drivers/gpu/drm/nouveau/dispnv50/ovly907e.c asyw->image.blocks[0] << 8 | blocks 261 drivers/gpu/drm/nouveau/dispnv50/wndw.c asyw->image.blocks[0] = fb->base.pitches[0] / 64; blocks 266 drivers/gpu/drm/nouveau/dispnv50/wndw.c asyw->image.blocks[0] = 0; blocks 133 drivers/gpu/drm/nouveau/dispnv50/wndwc37e.c evo_data(push, asyw->image.blocks[0] | (asyw->image.pitch[0] >> 6)); blocks 46 drivers/gpu/drm/nouveau/dispnv50/wndwc57e.c evo_data(push, asyw->image.blocks[0] | (asyw->image.pitch[0] >> 6)); blocks 650 drivers/gpu/drm/radeon/radeon_combios.c u8 id, blocks, clk, data; blocks 657 drivers/gpu/drm/radeon/radeon_combios.c blocks = RBIOS8(offset + 2); blocks 658 drivers/gpu/drm/radeon/radeon_combios.c for (i = 0; i < blocks; i++) { blocks 2641 drivers/gpu/drm/radeon/radeon_combios.c u8 rev, blocks, tmp; blocks 2734 drivers/gpu/drm/radeon/radeon_combios.c blocks = RBIOS8(offset + 0x2); blocks 2890 drivers/gpu/drm/radeon/radeon_combios.c uint8_t blocks, slave_addr, rev; blocks 2904 drivers/gpu/drm/radeon/radeon_combios.c blocks = RBIOS8(offset + 3); blocks 2906 drivers/gpu/drm/radeon/radeon_combios.c while (blocks > 0) { blocks 2952 drivers/gpu/drm/radeon/radeon_combios.c blocks--; blocks 1168 drivers/gpu/drm/vmwgfx/device_include/svga3d_surfacedefs.h surf_size_struct blocks; blocks 1170 drivers/gpu/drm/vmwgfx/device_include/svga3d_surfacedefs.h svga3dsurface_get_size_in_blocks(desc, size, &blocks); blocks 1172 drivers/gpu/drm/vmwgfx/device_include/svga3d_surfacedefs.h pitch = blocks.width * desc->pitch_bytes_per_block; blocks 656 drivers/i2c/busses/i2c-qup.c u32 len = 0, blocks, rem; blocks 663 drivers/i2c/busses/i2c-qup.c blocks = qup->blk.count; blocks 664 drivers/i2c/busses/i2c-qup.c rem = msg->len - (blocks - 1) * limit; blocks 667 drivers/i2c/busses/i2c-qup.c while (qup->blk.pos < blocks) { blocks 668 drivers/i2c/busses/i2c-qup.c tlen = (i == (blocks - 1)) ? rem : limit; blocks 699 drivers/i2c/busses/i2c-qup.c while (qup->blk.pos < blocks) { blocks 700 drivers/i2c/busses/i2c-qup.c tlen = (i == (blocks - 1)) ? rem : limit; blocks 1673 drivers/i2c/busses/i2c-qup.c int blocks; blocks 1714 drivers/i2c/busses/i2c-qup.c blocks = (MX_DMA_BLOCKS << 1) + 1; blocks 1716 drivers/i2c/busses/i2c-qup.c blocks, sizeof(*qup->btx.sg), blocks 1722 drivers/i2c/busses/i2c-qup.c sg_init_table(qup->btx.sg, blocks); blocks 1725 drivers/i2c/busses/i2c-qup.c blocks, sizeof(*qup->brx.sg), blocks 1731 drivers/i2c/busses/i2c-qup.c sg_init_table(qup->brx.sg, blocks); blocks 1734 drivers/i2c/busses/i2c-qup.c size = blocks * 2 + 5; blocks 535 drivers/ide/ide-cd.c unsigned long blocks = blk_rq_sectors(rq) / (hard_sect >> 9); blocks 554 drivers/ide/ide-cd.c req->cmd[7] = (blocks >> 8) & 0xff; blocks 555 drivers/ide/ide-cd.c req->cmd[8] = blocks & 0xff; blocks 197 drivers/ide/ide-floppy.c int blocks = blk_rq_sectors(rq) / floppy->bs_factor; blocks 200 drivers/ide/ide-floppy.c ide_debug_log(IDE_DBG_FUNC, "block: %d, blocks: %d", block, blocks); blocks 204 drivers/ide/ide-floppy.c put_unaligned(cpu_to_be16(blocks), (unsigned short *)&pc->c[7]); blocks 360 drivers/ide/ide-floppy.c lba_capacity = floppy->blocks * floppy->block_size; blocks 366 drivers/ide/ide-floppy.c floppy->blocks = floppy->block_size ? blocks 368 drivers/ide/ide-floppy.c drive->capacity64 = floppy->blocks * floppy->bs_factor; blocks 385 drivers/ide/ide-floppy.c int i, rc = 1, blocks, length; blocks 391 drivers/ide/ide-floppy.c floppy->blocks = 0; blocks 407 drivers/ide/ide-floppy.c blocks = be32_to_cpup((__be32 *)&pc_buf[desc_start]); blocks 412 drivers/ide/ide-floppy.c i, blocks * length / 1024, blocks 413 drivers/ide/ide-floppy.c blocks, length); blocks 436 drivers/ide/ide-floppy.c drive->name, blocks * length / 1024, blocks 437 drivers/ide/ide-floppy.c blocks, length); blocks 444 drivers/ide/ide-floppy.c floppy->blocks = blocks; blocks 453 drivers/ide/ide-floppy.c floppy->blocks * floppy->bs_factor; blocks 42 drivers/ide/ide-floppy_ioctl.c int i, blocks, length, u_array_size, u_index; blocks 76 drivers/ide/ide-floppy_ioctl.c blocks = be32_to_cpup((__be32 *)&pc_buf[desc_start]); blocks 79 drivers/ide/ide-floppy_ioctl.c if (put_user(blocks, argp)) blocks 144 drivers/ide/ide-floppy_ioctl.c int blocks, length, flags, err = 0; blocks 169 drivers/ide/ide-floppy_ioctl.c if (get_user(blocks, arg) || blocks 177 drivers/ide/ide-floppy_ioctl.c ide_floppy_create_format_unit_cmd(pc, buf, blocks, length, flags); blocks 34 drivers/ide/ide-gd.h int blocks, block_size, bs_factor; blocks 350 drivers/ide/ide-tape.c unsigned int blocks = blocks 353 drivers/ide/ide-tape.c tape->avg_size += blocks * tape->blk_size; blocks 362 drivers/ide/ide-tape.c tape->first_frame += blocks; blocks 308 drivers/iio/buffer/industrialio-buffer-dma.c for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { blocks 309 drivers/iio/buffer/industrialio-buffer-dma.c block = queue->fileio.blocks[i]; blocks 326 drivers/iio/buffer/industrialio-buffer-dma.c for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { blocks 327 drivers/iio/buffer/industrialio-buffer-dma.c if (queue->fileio.blocks[i]) { blocks 328 drivers/iio/buffer/industrialio-buffer-dma.c block = queue->fileio.blocks[i]; blocks 346 drivers/iio/buffer/industrialio-buffer-dma.c queue->fileio.blocks[i] = block; blocks 644 drivers/iio/buffer/industrialio-buffer-dma.c for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { blocks 645 drivers/iio/buffer/industrialio-buffer-dma.c if (!queue->fileio.blocks[i]) blocks 647 drivers/iio/buffer/industrialio-buffer-dma.c queue->fileio.blocks[i]->state = IIO_BLOCK_STATE_DEAD; blocks 654 drivers/iio/buffer/industrialio-buffer-dma.c for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { blocks 655 drivers/iio/buffer/industrialio-buffer-dma.c if (!queue->fileio.blocks[i]) blocks 657 drivers/iio/buffer/industrialio-buffer-dma.c iio_buffer_block_put(queue->fileio.blocks[i]); blocks 658 drivers/iio/buffer/industrialio-buffer-dma.c queue->fileio.blocks[i] = NULL; blocks 184 drivers/infiniband/hw/hfi1/pio.c int blocks; /* block size of the pool */ blocks 282 drivers/infiniband/hw/hfi1/pio.c mem_pool_info[i].blocks = ab; blocks 395 drivers/infiniband/hw/hfi1/pio.c pi->blocks = (pool_blocks * pi->centipercent) / 10000; blocks 397 drivers/infiniband/hw/hfi1/pio.c if (pi->blocks == 0 && pi->count != 0) { blocks 406 drivers/infiniband/hw/hfi1/pio.c if (pi->blocks != 0) blocks 410 drivers/infiniband/hw/hfi1/pio.c i, pi->blocks); blocks 413 drivers/infiniband/hw/hfi1/pio.c pi->size = pi->blocks / pi->count; blocks 1455 drivers/infiniband/hw/hfi1/pio.c unsigned long blocks = dwords_to_blocks(dw_len); blocks 1468 drivers/infiniband/hw/hfi1/pio.c if (blocks > avail) { blocks 1479 drivers/infiniband/hw/hfi1/pio.c if (blocks > avail) { blocks 1497 drivers/infiniband/hw/hfi1/pio.c sc->fill += blocks; blocks 1499 drivers/infiniband/hw/hfi1/pio.c sc->fill_wrap += blocks; blocks 1372 drivers/md/bcache/btree.c unsigned int i, nodes = 0, keys = 0, blocks; blocks 1389 drivers/md/bcache/btree.c blocks = btree_default_blocks(b->c) * 2 / 3; blocks 1393 drivers/md/bcache/btree.c block_bytes(b->c)) > blocks * (nodes - 1)) blocks 1427 drivers/md/bcache/btree.c block_bytes(b->c)) > blocks) blocks 78 drivers/md/bcache/journal.c size_t blocks, bytes = set_bytes(j); blocks 101 drivers/md/bcache/journal.c blocks = set_blocks(j, block_bytes(ca->set)); blocks 162 drivers/md/bcache/journal.c offset += blocks * ca->sb.block_size; blocks 163 drivers/md/bcache/journal.c len -= blocks * ca->sb.block_size; blocks 164 drivers/md/bcache/journal.c j = ((void *) j) + blocks * block_bytes(ca); blocks 4471 drivers/md/dm-thin.c sector_t blocks; blocks 4482 drivers/md/dm-thin.c blocks = pool->ti->len; blocks 4483 drivers/md/dm-thin.c (void) sector_div(blocks, pool->sectors_per_block); blocks 4484 drivers/md/dm-thin.c if (blocks) blocks 4485 drivers/md/dm-thin.c return fn(ti, tc->pool_dev, 0, pool->sectors_per_block * blocks, data); blocks 537 drivers/md/dm-verity-fec.c (unsigned long long)v->fec->blocks, blocks 624 drivers/md/dm-verity-fec.c v->fec->blocks = num_ll; blocks 717 drivers/md/dm-verity-fec.c if (!f->blocks) { blocks 722 drivers/md/dm-verity-fec.c f->rounds = f->blocks; blocks 730 drivers/md/dm-verity-fec.c if (f->blocks < v->data_blocks + hash_blocks || !f->rounds) { blocks 739 drivers/md/dm-verity-fec.c f->hash_blocks = f->blocks - v->data_blocks; blocks 40 drivers/md/dm-verity-fec.h sector_t blocks; /* number of blocks covered */ blocks 1222 drivers/md/md-bitmap.c sector_t offset, sector_t *blocks, blocks 1235 drivers/md/md-bitmap.c sector_t blocks; blocks 1306 drivers/md/md-bitmap.c bmc = md_bitmap_get_counter(counts, block, &blocks, 0); blocks 1355 drivers/md/md-bitmap.c sector_t offset, sector_t *blocks, blocks 1378 drivers/md/md-bitmap.c *blocks = csize - (offset & (csize - 1)); blocks 1413 drivers/md/md-bitmap.c sector_t blocks; blocks 1417 drivers/md/md-bitmap.c bmc = md_bitmap_get_counter(&bitmap->counts, offset, &blocks, 1); blocks 1450 drivers/md/md-bitmap.c offset += blocks; blocks 1451 drivers/md/md-bitmap.c if (sectors > blocks) blocks 1452 drivers/md/md-bitmap.c sectors -= blocks; blocks 1474 drivers/md/md-bitmap.c sector_t blocks; blocks 1479 drivers/md/md-bitmap.c bmc = md_bitmap_get_counter(&bitmap->counts, offset, &blocks, 0); blocks 1504 drivers/md/md-bitmap.c offset += blocks; blocks 1505 drivers/md/md-bitmap.c if (sectors > blocks) blocks 1506 drivers/md/md-bitmap.c sectors -= blocks; blocks 1513 drivers/md/md-bitmap.c static int __bitmap_start_sync(struct bitmap *bitmap, sector_t offset, sector_t *blocks, blocks 1519 drivers/md/md-bitmap.c *blocks = 1024; blocks 1523 drivers/md/md-bitmap.c bmc = md_bitmap_get_counter(&bitmap->counts, offset, blocks, 0); blocks 1541 drivers/md/md-bitmap.c int md_bitmap_start_sync(struct bitmap *bitmap, sector_t offset, sector_t *blocks, blocks 1554 drivers/md/md-bitmap.c *blocks = 0; blocks 1555 drivers/md/md-bitmap.c while (*blocks < (PAGE_SIZE>>9)) { blocks 1559 drivers/md/md-bitmap.c *blocks += blocks1; blocks 1565 drivers/md/md-bitmap.c void md_bitmap_end_sync(struct bitmap *bitmap, sector_t offset, sector_t *blocks, int aborted) blocks 1571 drivers/md/md-bitmap.c *blocks = 1024; blocks 1575 drivers/md/md-bitmap.c bmc = md_bitmap_get_counter(&bitmap->counts, offset, blocks, 0); blocks 1603 drivers/md/md-bitmap.c sector_t blocks; blocks 1607 drivers/md/md-bitmap.c md_bitmap_end_sync(bitmap, sector, &blocks, 0); blocks 1608 drivers/md/md-bitmap.c sector += blocks; blocks 1616 drivers/md/md-bitmap.c sector_t blocks; blocks 1635 drivers/md/md-bitmap.c md_bitmap_end_sync(bitmap, s, &blocks, 0); blocks 1636 drivers/md/md-bitmap.c s += blocks; blocks 1648 drivers/md/md-bitmap.c sector_t sector, blocks = 0; blocks 1651 drivers/md/md-bitmap.c md_bitmap_end_sync(bitmap, sector, &blocks, 0); blocks 1652 drivers/md/md-bitmap.c sector += blocks; blocks 1654 drivers/md/md-bitmap.c WARN((blocks > new_lo) && old_lo, "alignment is not correct for lo\n"); blocks 1657 drivers/md/md-bitmap.c md_bitmap_start_sync(bitmap, sector, &blocks, 0); blocks 1658 drivers/md/md-bitmap.c sector += blocks; blocks 1660 drivers/md/md-bitmap.c WARN((blocks > new_hi) && old_hi, "alignment is not correct for hi\n"); blocks 1815 drivers/md/md-bitmap.c sector_t blocks = mddev->resync_max_sectors; blocks 1882 drivers/md/md-bitmap.c err = md_bitmap_resize(bitmap, blocks, mddev->bitmap_info.chunksize, 1); blocks 1922 drivers/md/md-bitmap.c sector_t blocks; blocks 1923 drivers/md/md-bitmap.c md_bitmap_start_sync(bitmap, sector, &blocks, 0); blocks 1924 drivers/md/md-bitmap.c sector += blocks; blocks 2053 drivers/md/md-bitmap.c int md_bitmap_resize(struct bitmap *bitmap, sector_t blocks, blocks 2103 drivers/md/md-bitmap.c chunks = DIV_ROUND_UP_SECTOR_T(blocks, 1 << chunkshift); blocks 2111 drivers/md/md-bitmap.c chunks = DIV_ROUND_UP_SECTOR_T(blocks, 1 << chunkshift); blocks 2155 drivers/md/md-bitmap.c blocks = min(old_counts.chunks << old_counts.chunkshift, blocks 2180 drivers/md/md-bitmap.c blocks = old_counts.chunks << old_counts.chunkshift; blocks 2188 drivers/md/md-bitmap.c for (block = 0; block < blocks; ) { blocks 258 drivers/md/md-bitmap.h int md_bitmap_start_sync(struct bitmap *bitmap, sector_t offset, sector_t *blocks, int degraded); blocks 259 drivers/md/md-bitmap.h void md_bitmap_end_sync(struct bitmap *bitmap, sector_t offset, sector_t *blocks, int aborted); blocks 269 drivers/md/md-bitmap.h int md_bitmap_resize(struct bitmap *bitmap, sector_t blocks, blocks 3230 drivers/md/md.c unsigned long long blocks; blocks 3233 drivers/md/md.c if (kstrtoull(buf, 10, &blocks) < 0) blocks 3236 drivers/md/md.c if (blocks & 1ULL << (8 * sizeof(blocks) - 1)) blocks 3239 drivers/md/md.c new = blocks * 2; blocks 3240 drivers/md/md.c if (new != blocks * 2) blocks 8260 drivers/md/md.c void md_done_sync(struct mddev *mddev, int blocks, int ok) blocks 8263 drivers/md/md.c atomic_sub(blocks, &mddev->recovery_active); blocks 701 drivers/md/md.h extern void md_done_sync(struct mddev *mddev, int blocks, int ok); blocks 238 drivers/md/persistent-data/dm-space-map-common.c unsigned old_blocks, blocks; blocks 242 drivers/md/persistent-data/dm-space-map-common.c blocks = dm_sector_div_up(nr_blocks, ll->entries_per_block); blocks 254 drivers/md/persistent-data/dm-space-map-common.c for (i = old_blocks; i < blocks; i++) { blocks 28 drivers/md/persistent-data/dm-transaction-manager.c dm_block_t blocks[PREFETCH_SIZE]; blocks 40 drivers/md/persistent-data/dm-transaction-manager.c p->blocks[i] = PREFETCH_SENTINEL; blocks 54 drivers/md/persistent-data/dm-transaction-manager.c if (p->blocks[h] == PREFETCH_SENTINEL) blocks 55 drivers/md/persistent-data/dm-transaction-manager.c p->blocks[h] = b; blocks 67 drivers/md/persistent-data/dm-transaction-manager.c if (p->blocks[i] != PREFETCH_SENTINEL) { blocks 68 drivers/md/persistent-data/dm-transaction-manager.c dm_bm_prefetch(bm, p->blocks[i]); blocks 69 drivers/md/persistent-data/dm-transaction-manager.c p->blocks[i] = PREFETCH_SENTINEL; blocks 1490 drivers/md/raid5.c struct page **blocks = to_addr_page(percpu, 0); blocks 1519 drivers/md/raid5.c count = set_syndrome_sources(blocks, sh, SYNDROME_SRC_ALL); blocks 1520 drivers/md/raid5.c blocks[count] = NULL; /* regenerating p is not necessary */ blocks 1521 drivers/md/raid5.c BUG_ON(blocks[count+1] != dest); /* q should already be set */ blocks 1525 drivers/md/raid5.c tx = async_gen_syndrome(blocks, 0, count+2, STRIPE_SIZE, &submit); blocks 1532 drivers/md/raid5.c blocks[count++] = sh->dev[i].page; blocks 1538 drivers/md/raid5.c tx = async_xor(dest, blocks, 0, count, STRIPE_SIZE, &submit); blocks 1556 drivers/md/raid5.c struct page **blocks = to_addr_page(percpu, 0); blocks 1570 drivers/md/raid5.c blocks[i] = NULL; blocks 1576 drivers/md/raid5.c blocks[slot] = sh->dev[i].page; blocks 1600 drivers/md/raid5.c return async_gen_syndrome(blocks, 0, syndrome_disks+2, blocks 1617 drivers/md/raid5.c blocks[count++] = sh->dev[i].page; blocks 1624 drivers/md/raid5.c tx = async_xor(dest, blocks, 0, count, STRIPE_SIZE, blocks 1627 drivers/md/raid5.c count = set_syndrome_sources(blocks, sh, SYNDROME_SRC_ALL); blocks 1631 drivers/md/raid5.c return async_gen_syndrome(blocks, 0, count+2, blocks 1642 drivers/md/raid5.c blocks, &submit); blocks 1647 drivers/md/raid5.c blocks, &submit); blocks 1703 drivers/md/raid5.c struct page **blocks = to_addr_page(percpu, 0); blocks 1710 drivers/md/raid5.c count = set_syndrome_sources(blocks, sh, SYNDROME_SRC_WANT_DRAIN); blocks 1714 drivers/md/raid5.c tx = async_gen_syndrome(blocks, 0, count+2, STRIPE_SIZE, &submit); blocks 1929 drivers/md/raid5.c struct page **blocks; blocks 1953 drivers/md/raid5.c blocks = to_addr_page(percpu, j); blocks 1963 drivers/md/raid5.c count = set_syndrome_sources(blocks, sh, synflags); blocks 1975 drivers/md/raid5.c tx = async_gen_syndrome(blocks, 0, count+2, STRIPE_SIZE, &submit); blocks 1902 drivers/media/dvb-frontends/dib7000p.c u64 blocks, ucb; blocks 1956 drivers/media/dvb-frontends/dib7000p.c blocks = 1250000ULL * 1000000ULL; blocks 1957 drivers/media/dvb-frontends/dib7000p.c do_div(blocks, time_us * 8 * 204); blocks 1959 drivers/media/dvb-frontends/dib7000p.c c->block_count.stat[0].uvalue += blocks; blocks 1989 drivers/media/dvb-frontends/dib7000p.c blocks = 1250000ULL * 1000000ULL; blocks 1990 drivers/media/dvb-frontends/dib7000p.c do_div(blocks, time_us * 8 * 204); blocks 1992 drivers/media/dvb-frontends/dib7000p.c c->block_count.stat[0].uvalue += blocks; blocks 4104 drivers/media/dvb-frontends/dib8000.c u64 blocks; blocks 4165 drivers/media/dvb-frontends/dib8000.c blocks = 1250000ULL * 1000000ULL; blocks 4166 drivers/media/dvb-frontends/dib8000.c do_div(blocks, time_us * 8 * 204); blocks 4168 drivers/media/dvb-frontends/dib8000.c c->block_count.stat[0].uvalue += blocks; blocks 4228 drivers/media/dvb-frontends/dib8000.c blocks = 1250000ULL * 1000000ULL; blocks 4229 drivers/media/dvb-frontends/dib8000.c do_div(blocks, time_us * 8 * 204); blocks 4231 drivers/media/dvb-frontends/dib8000.c c->block_count.stat[0].uvalue += blocks; blocks 59 drivers/media/i2c/ad9389b.c u32 blocks; blocks 376 drivers/media/i2c/ad9389b.c edid->segments ? "found" : "no", edid->blocks); blocks 670 drivers/media/i2c/ad9389b.c if (edid->blocks == 0 || edid->blocks > 256) blocks 678 drivers/media/i2c/ad9389b.c if (edid->blocks + edid->start_block >= state->edid.segments * 2) blocks 679 drivers/media/i2c/ad9389b.c edid->blocks = state->edid.segments * 2 - edid->start_block; blocks 681 drivers/media/i2c/ad9389b.c 128 * edid->blocks); blocks 979 drivers/media/i2c/ad9389b.c u32 blocks = state->edid.blocks; blocks 983 drivers/media/i2c/ad9389b.c if ((segment + 1) * 2 <= blocks) blocks 1032 drivers/media/i2c/ad9389b.c state->edid.blocks = state->edid.data[0x7e] + 1; blocks 1034 drivers/media/i2c/ad9389b.c __func__, state->edid.blocks); blocks 471 drivers/media/i2c/adv748x/adv748x-hdmi.c if (edid->start_block == 0 && edid->blocks == 0) { blocks 472 drivers/media/i2c/adv748x/adv748x-hdmi.c edid->blocks = hdmi->edid.blocks; blocks 476 drivers/media/i2c/adv748x/adv748x-hdmi.c if (edid->start_block >= hdmi->edid.blocks) blocks 479 drivers/media/i2c/adv748x/adv748x-hdmi.c if (edid->start_block + edid->blocks > hdmi->edid.blocks) blocks 480 drivers/media/i2c/adv748x/adv748x-hdmi.c edid->blocks = hdmi->edid.blocks - edid->start_block; blocks 483 drivers/media/i2c/adv748x/adv748x-hdmi.c edid->blocks * 128); blocks 523 drivers/media/i2c/adv748x/adv748x-hdmi.c if (edid->blocks == 0) { blocks 524 drivers/media/i2c/adv748x/adv748x-hdmi.c hdmi->edid.blocks = 0; blocks 533 drivers/media/i2c/adv748x/adv748x-hdmi.c edid->blocks << ADV748X_REPEATER_EDID_SZ_SHIFT); blocks 540 drivers/media/i2c/adv748x/adv748x-hdmi.c if (edid->blocks > 4) { blocks 541 drivers/media/i2c/adv748x/adv748x-hdmi.c edid->blocks = 4; blocks 545 drivers/media/i2c/adv748x/adv748x-hdmi.c memcpy(hdmi->edid.edid, edid->edid, 128 * edid->blocks); blocks 546 drivers/media/i2c/adv748x/adv748x-hdmi.c hdmi->edid.blocks = edid->blocks; blocks 552 drivers/media/i2c/adv748x/adv748x-hdmi.c err = adv748x_hdmi_edid_write_block(hdmi, 128 * edid->blocks, blocks 560 drivers/media/i2c/adv748x/adv748x-hdmi.c edid->blocks << ADV748X_REPEATER_EDID_SZ_SHIFT); blocks 129 drivers/media/i2c/adv748x/adv748x.h unsigned int blocks; blocks 73 drivers/media/i2c/adv7511-v4l2.c u32 blocks; blocks 612 drivers/media/i2c/adv7511-v4l2.c edid->blocks); blocks 1209 drivers/media/i2c/adv7511-v4l2.c if (edid->start_block == 0 && edid->blocks == 0) { blocks 1210 drivers/media/i2c/adv7511-v4l2.c edid->blocks = state->edid.segments * 2; blocks 1220 drivers/media/i2c/adv7511-v4l2.c if (edid->start_block + edid->blocks > state->edid.segments * 2) blocks 1221 drivers/media/i2c/adv7511-v4l2.c edid->blocks = state->edid.segments * 2 - edid->start_block; blocks 1224 drivers/media/i2c/adv7511-v4l2.c 128 * edid->blocks); blocks 1634 drivers/media/i2c/adv7511-v4l2.c u32 blocks = state->edid.blocks; blocks 1639 drivers/media/i2c/adv7511-v4l2.c if ((segment + 1) * 2 <= blocks) blocks 1680 drivers/media/i2c/adv7511-v4l2.c state->edid.blocks = state->edid.data[0x7e] + 1; blocks 1681 drivers/media/i2c/adv7511-v4l2.c v4l2_dbg(1, debug, sd, "%s: %d blocks in total\n", __func__, state->edid.blocks); blocks 181 drivers/media/i2c/adv7604.c unsigned blocks; blocks 2267 drivers/media/i2c/adv7604.c if (edid->start_block == 0 && edid->blocks == 0) { blocks 2268 drivers/media/i2c/adv7604.c edid->blocks = data ? state->edid.blocks : 0; blocks 2275 drivers/media/i2c/adv7604.c if (edid->start_block >= state->edid.blocks) blocks 2278 drivers/media/i2c/adv7604.c if (edid->start_block + edid->blocks > state->edid.blocks) blocks 2279 drivers/media/i2c/adv7604.c edid->blocks = state->edid.blocks - edid->start_block; blocks 2281 drivers/media/i2c/adv7604.c memcpy(edid->edid, data + edid->start_block * 128, edid->blocks * 128); blocks 2301 drivers/media/i2c/adv7604.c if (edid->blocks == 0) { blocks 2312 drivers/media/i2c/adv7604.c state->edid.blocks = 0; blocks 2320 drivers/media/i2c/adv7604.c if (edid->blocks > 2) { blocks 2321 drivers/media/i2c/adv7604.c edid->blocks = 2; blocks 2324 drivers/media/i2c/adv7604.c pa = v4l2_get_edid_phys_addr(edid->edid, edid->blocks * 128, &spa_loc); blocks 2377 drivers/media/i2c/adv7604.c memcpy(state->edid.edid, edid->edid, 128 * edid->blocks); blocks 2378 drivers/media/i2c/adv7604.c state->edid.blocks = edid->blocks; blocks 2383 drivers/media/i2c/adv7604.c err = edid_write_block(sd, 128 * edid->blocks, state->edid.edid); blocks 2477 drivers/media/i2c/adv7842.c if (edid->start_block == 0 && edid->blocks == 0) { blocks 2478 drivers/media/i2c/adv7842.c edid->blocks = data ? 2 : 0; blocks 2488 drivers/media/i2c/adv7842.c if (edid->start_block + edid->blocks > 2) blocks 2489 drivers/media/i2c/adv7842.c edid->blocks = 2 - edid->start_block; blocks 2491 drivers/media/i2c/adv7842.c memcpy(edid->edid, data + edid->start_block * 128, edid->blocks * 128); blocks 2507 drivers/media/i2c/adv7842.c if (e->blocks > 2) { blocks 2508 drivers/media/i2c/adv7842.c e->blocks = 2; blocks 2519 drivers/media/i2c/adv7842.c state->vga_edid.present = e->blocks ? 0x1 : 0x0; blocks 2520 drivers/media/i2c/adv7842.c memcpy(&state->vga_edid.edid, e->edid, 128 * e->blocks); blocks 2526 drivers/media/i2c/adv7842.c if (e->blocks) { blocks 2532 drivers/media/i2c/adv7842.c memcpy(&state->hdmi_edid.edid, e->edid, 128 * e->blocks); blocks 1746 drivers/media/i2c/tc358743.c if (edid->start_block == 0 && edid->blocks == 0) { blocks 1747 drivers/media/i2c/tc358743.c edid->blocks = state->edid_blocks_written; blocks 1755 drivers/media/i2c/tc358743.c edid->blocks == 0) blocks 1758 drivers/media/i2c/tc358743.c if (edid->start_block + edid->blocks > state->edid_blocks_written) blocks 1759 drivers/media/i2c/tc358743.c edid->blocks = state->edid_blocks_written - edid->start_block; blocks 1762 drivers/media/i2c/tc358743.c edid->blocks * EDID_BLOCK_SIZE); blocks 1771 drivers/media/i2c/tc358743.c u16 edid_len = edid->blocks * EDID_BLOCK_SIZE; blocks 1777 drivers/media/i2c/tc358743.c __func__, edid->pad, edid->start_block, edid->blocks); blocks 1787 drivers/media/i2c/tc358743.c if (edid->blocks > EDID_NUM_BLOCKS_MAX) { blocks 1788 drivers/media/i2c/tc358743.c edid->blocks = EDID_NUM_BLOCKS_MAX; blocks 1791 drivers/media/i2c/tc358743.c pa = cec_get_edid_phys_addr(edid->edid, edid->blocks * 128, NULL); blocks 1803 drivers/media/i2c/tc358743.c if (edid->blocks == 0) { blocks 1811 drivers/media/i2c/tc358743.c state->edid_blocks_written = edid->blocks; blocks 275 drivers/media/i2c/tda1997x.c unsigned int blocks; blocks 1834 drivers/media/i2c/tda1997x.c if (edid->start_block == 0 && edid->blocks == 0) { blocks 1835 drivers/media/i2c/tda1997x.c edid->blocks = state->edid.blocks; blocks 1842 drivers/media/i2c/tda1997x.c if (edid->start_block >= state->edid.blocks) blocks 1845 drivers/media/i2c/tda1997x.c if (edid->start_block + edid->blocks > state->edid.blocks) blocks 1846 drivers/media/i2c/tda1997x.c edid->blocks = state->edid.blocks - edid->start_block; blocks 1849 drivers/media/i2c/tda1997x.c edid->blocks * 128); blocks 1865 drivers/media/i2c/tda1997x.c if (edid->blocks == 0) { blocks 1866 drivers/media/i2c/tda1997x.c state->edid.blocks = 0; blocks 1872 drivers/media/i2c/tda1997x.c if (edid->blocks > 2) { blocks 1873 drivers/media/i2c/tda1997x.c edid->blocks = 2; blocks 1889 drivers/media/i2c/tda1997x.c state->edid.blocks = edid->blocks; blocks 501 drivers/media/pci/cobalt/cobalt-driver.c .blocks = 2, blocks 594 drivers/media/pci/cobalt/cobalt-driver.c .blocks = 2, blocks 150 drivers/media/pci/saa7134/saa7134-alsa.c if (dev->dmasound.read_count >= dev->dmasound.blksize * (dev->dmasound.blocks-2)) { blocks 153 drivers/media/pci/saa7134/saa7134-alsa.c dev->dmasound.bufsize, dev->dmasound.blocks); blocks 160 drivers/media/pci/saa7134/saa7134-alsa.c next_blk = (dev->dmasound.dma_blk + 2) % dev->dmasound.blocks; blocks 164 drivers/media/pci/saa7134/saa7134-alsa.c next_blk * dev->dmasound.blksize, dev->dmasound.blocks, blocks 168 drivers/media/pci/saa7134/saa7134-alsa.c dev->dmasound.dma_blk = (dev->dmasound.dma_blk + 1) % dev->dmasound.blocks; blocks 367 drivers/media/pci/saa7134/saa7134-alsa.c dev->dmasound.blocks = 0; blocks 680 drivers/media/pci/saa7134/saa7134-alsa.c if (dev->dmasound.blocks == periods && blocks 691 drivers/media/pci/saa7134/saa7134-alsa.c dev->dmasound.blocks = periods; blocks 697 drivers/media/pci/saa7134/saa7134-alsa.c dev->dmasound.blocks = 0; blocks 505 drivers/media/pci/saa7134/saa7134.h unsigned int blocks; blocks 138 drivers/media/pci/ttpci/av7110_hw.c int blocks, rest; blocks 143 drivers/media/pci/ttpci/av7110_hw.c blocks = len / AV7110_BOOT_MAX_SIZE; blocks 147 drivers/media/pci/ttpci/av7110_hw.c for (i = 0; i < blocks; i++) { blocks 1757 drivers/media/platform/vivid/vivid-vid-cap.c if (edid->blocks == 0) { blocks 1764 drivers/media/platform/vivid/vivid-vid-cap.c if (edid->blocks > dev->edid_max_blocks) { blocks 1765 drivers/media/platform/vivid/vivid-vid-cap.c edid->blocks = dev->edid_max_blocks; blocks 1768 drivers/media/platform/vivid/vivid-vid-cap.c phys_addr = cec_get_edid_phys_addr(edid->edid, edid->blocks * 128, NULL); blocks 1776 drivers/media/platform/vivid/vivid-vid-cap.c dev->edid_blocks = edid->blocks; blocks 1777 drivers/media/platform/vivid/vivid-vid-cap.c memcpy(dev->edid, edid->edid, edid->blocks * 128); blocks 1021 drivers/media/platform/vivid/vivid-vid-common.c if (edid->start_block == 0 && edid->blocks == 0) { blocks 1022 drivers/media/platform/vivid/vivid-vid-common.c edid->blocks = dev->edid_blocks; blocks 1029 drivers/media/platform/vivid/vivid-vid-common.c if (edid->blocks > dev->edid_blocks - edid->start_block) blocks 1030 drivers/media/platform/vivid/vivid-vid-common.c edid->blocks = dev->edid_blocks - edid->start_block; blocks 1033 drivers/media/platform/vivid/vivid-vid-common.c memcpy(edid->edid, dev->edid + edid->start_block * 128, edid->blocks * 128); blocks 1063 drivers/media/v4l2-core/v4l2-compat-ioctl32.c __u32 blocks; blocks 1076 drivers/media/v4l2-core/v4l2-compat-ioctl32.c assign_in_user_cast(&p64->blocks, &p32->blocks) || blocks 1092 drivers/media/v4l2-core/v4l2-compat-ioctl32.c assign_in_user(&p32->blocks, &p64->blocks) || blocks 887 drivers/media/v4l2-core/v4l2-ioctl.c p->pad, p->start_block, p->blocks); blocks 2978 drivers/media/v4l2-core/v4l2-ioctl.c if (edid->blocks) { blocks 2979 drivers/media/v4l2-core/v4l2-ioctl.c if (edid->blocks > 256) { blocks 2985 drivers/media/v4l2-core/v4l2-ioctl.c *array_size = edid->blocks * 128; blocks 266 drivers/media/v4l2-core/v4l2-subdev.c if (edid->blocks && edid->edid == NULL) blocks 39 drivers/mfd/stmpe.c unsigned int blocks; blocks 47 drivers/mfd/stmpe.c static int __stmpe_enable(struct stmpe *stmpe, unsigned int blocks) blocks 49 drivers/mfd/stmpe.c return stmpe->variant->enable(stmpe, blocks, true); blocks 52 drivers/mfd/stmpe.c static int __stmpe_disable(struct stmpe *stmpe, unsigned int blocks) blocks 54 drivers/mfd/stmpe.c return stmpe->variant->enable(stmpe, blocks, false); blocks 132 drivers/mfd/stmpe.c int stmpe_enable(struct stmpe *stmpe, unsigned int blocks) blocks 137 drivers/mfd/stmpe.c ret = __stmpe_enable(stmpe, blocks); blocks 149 drivers/mfd/stmpe.c int stmpe_disable(struct stmpe *stmpe, unsigned int blocks) blocks 154 drivers/mfd/stmpe.c ret = __stmpe_disable(stmpe, blocks); blocks 412 drivers/mfd/stmpe.c static int stmpe801_enable(struct stmpe *stmpe, unsigned int blocks, blocks 415 drivers/mfd/stmpe.c if (blocks & STMPE_BLOCK_GPIO) blocks 427 drivers/mfd/stmpe.c .blocks = stmpe801_blocks, blocks 439 drivers/mfd/stmpe.c .blocks = stmpe801_blocks_noirq, blocks 529 drivers/mfd/stmpe.c static int stmpe811_enable(struct stmpe *stmpe, unsigned int blocks, blocks 534 drivers/mfd/stmpe.c if (blocks & STMPE_BLOCK_GPIO) blocks 537 drivers/mfd/stmpe.c if (blocks & STMPE_BLOCK_ADC) blocks 540 drivers/mfd/stmpe.c if (blocks & STMPE_BLOCK_TOUCHSCREEN) blocks 589 drivers/mfd/stmpe.c .blocks = stmpe811_blocks, blocks 604 drivers/mfd/stmpe.c .blocks = stmpe811_blocks, blocks 644 drivers/mfd/stmpe.c static int stmpe1600_enable(struct stmpe *stmpe, unsigned int blocks, blocks 647 drivers/mfd/stmpe.c if (blocks & STMPE_BLOCK_GPIO) blocks 660 drivers/mfd/stmpe.c .blocks = stmpe1600_blocks, blocks 778 drivers/mfd/stmpe.c static int stmpe1601_enable(struct stmpe *stmpe, unsigned int blocks, blocks 783 drivers/mfd/stmpe.c if (blocks & STMPE_BLOCK_GPIO) blocks 788 drivers/mfd/stmpe.c if (blocks & STMPE_BLOCK_KEYPAD) blocks 793 drivers/mfd/stmpe.c if (blocks & STMPE_BLOCK_PWM) blocks 824 drivers/mfd/stmpe.c .blocks = stmpe1601_blocks, blocks 879 drivers/mfd/stmpe.c static int stmpe1801_enable(struct stmpe *stmpe, unsigned int blocks, blocks 883 drivers/mfd/stmpe.c if (blocks & STMPE_BLOCK_GPIO) blocks 886 drivers/mfd/stmpe.c if (blocks & STMPE_BLOCK_KEYPAD) blocks 933 drivers/mfd/stmpe.c .blocks = stmpe1801_blocks, blocks 1001 drivers/mfd/stmpe.c static int stmpe24xx_enable(struct stmpe *stmpe, unsigned int blocks, blocks 1006 drivers/mfd/stmpe.c if (blocks & STMPE_BLOCK_GPIO) blocks 1009 drivers/mfd/stmpe.c if (blocks & STMPE_BLOCK_KEYPAD) blocks 1039 drivers/mfd/stmpe.c .blocks = stmpe24xx_blocks, blocks 1053 drivers/mfd/stmpe.c .blocks = stmpe24xx_blocks, blocks 1306 drivers/mfd/stmpe.c unsigned int platform_blocks = stmpe->pdata->blocks; blocks 1311 drivers/mfd/stmpe.c struct stmpe_variant_block *block = &variant->blocks[i]; blocks 1362 drivers/mfd/stmpe.c pdata->blocks |= STMPE_BLOCK_GPIO; blocks 1364 drivers/mfd/stmpe.c pdata->blocks |= STMPE_BLOCK_KEYPAD; blocks 1366 drivers/mfd/stmpe.c pdata->blocks |= STMPE_BLOCK_TOUCHSCREEN; blocks 1368 drivers/mfd/stmpe.c pdata->blocks |= STMPE_BLOCK_ADC; blocks 1370 drivers/mfd/stmpe.c pdata->blocks |= STMPE_BLOCK_PWM; blocks 1372 drivers/mfd/stmpe.c pdata->blocks |= STMPE_BLOCK_ROTATOR; blocks 70 drivers/mfd/stmpe.h struct stmpe_variant_block *blocks; blocks 73 drivers/mfd/stmpe.h int (*enable)(struct stmpe *stmpe, unsigned int blocks, bool enable); blocks 287 drivers/mfd/tc3589x.c unsigned int blocks = tc3589x->pdata->block; blocks 289 drivers/mfd/tc3589x.c if (blocks & TC3589x_BLOCK_GPIO) { blocks 300 drivers/mfd/tc3589x.c if (blocks & TC3589x_BLOCK_KEYPAD) { blocks 367 drivers/mmc/core/block.c idata->buf_bytes = (u64) idata->ic.blksz * idata->ic.blocks; blocks 526 drivers/mmc/core/block.c data.blocks = idata->ic.blocks; blocks 577 drivers/mmc/core/block.c sbc.arg = data.blocks | (idata->ic.write_flag & BIT(31)); blocks 907 drivers/mmc/core/block.c __be32 *blocks; blocks 932 drivers/mmc/core/block.c data.blocks = 1; blocks 941 drivers/mmc/core/block.c blocks = kmalloc(4, GFP_KERNEL); blocks 942 drivers/mmc/core/block.c if (!blocks) blocks 945 drivers/mmc/core/block.c sg_init_one(&sg, blocks, 4); blocks 949 drivers/mmc/core/block.c result = ntohl(*blocks); blocks 950 drivers/mmc/core/block.c kfree(blocks); blocks 1217 drivers/mmc/core/block.c brq->data.blocks = 1; blocks 1219 drivers/mmc/core/block.c if (brq->data.blocks > card->ext_csd.rel_sectors) blocks 1220 drivers/mmc/core/block.c brq->data.blocks = card->ext_csd.rel_sectors; blocks 1221 drivers/mmc/core/block.c else if (brq->data.blocks < card->ext_csd.rel_sectors) blocks 1222 drivers/mmc/core/block.c brq->data.blocks = 1; blocks 1315 drivers/mmc/core/block.c brq->data.blocks = blk_rq_sectors(req); blocks 1330 drivers/mmc/core/block.c if (brq->data.blocks > card->host->max_blk_count) blocks 1331 drivers/mmc/core/block.c brq->data.blocks = card->host->max_blk_count; blocks 1333 drivers/mmc/core/block.c if (brq->data.blocks > 1) { blocks 1342 drivers/mmc/core/block.c brq->data.blocks--; blocks 1350 drivers/mmc/core/block.c brq->data.blocks = 1; blocks 1357 drivers/mmc/core/block.c brq->data.blocks = card->host->ops->multi_io_quirk(card, blocks 1360 drivers/mmc/core/block.c brq->data.blocks); blocks 1375 drivers/mmc/core/block.c ((brq->data.blocks * brq->data.blksz) >= blocks 1390 drivers/mmc/core/block.c if (brq->data.blocks != blk_rq_sectors(req)) { blocks 1391 drivers/mmc/core/block.c int i, data_size = brq->data.blocks << 9; blocks 1565 drivers/mmc/core/block.c if (brq->data.blocks > 1 || do_rel_wr) { blocks 1603 drivers/mmc/core/block.c brq->sbc.arg = brq->data.blocks | blocks 1763 drivers/mmc/core/block.c u32 blocks; blocks 1801 drivers/mmc/core/block.c if (mmc_sd_num_wr_blocks(card, &blocks)) blocks 1804 drivers/mmc/core/block.c brq->data.bytes_xfered = blocks << 9; blocks 1838 drivers/mmc/core/block.c brq->data.blocks > 1) { blocks 96 drivers/mmc/core/core.c !should_fail(&host->fail_mmc_request, data->blksz * data->blocks)) blocks 286 drivers/mmc/core/core.c mrq->data->blocks, mrq->data->flags, blocks 314 drivers/mmc/core/core.c mrq->data->blocks > host->max_blk_count || blocks 315 drivers/mmc/core/core.c mrq->data->blocks * mrq->data->blksz > host->max_req_size) blocks 320 drivers/mmc/core/core.c if (sz != mrq->data->blocks * mrq->data->blksz) blocks 270 drivers/mmc/core/mmc_ops.c data.blocks = 1; blocks 642 drivers/mmc/core/mmc_ops.c data.blocks = 1; blocks 753 drivers/mmc/core/mmc_ops.c data.blocks = 1; blocks 81 drivers/mmc/core/mmc_test.c unsigned int blocks; blocks 189 drivers/mmc/core/mmc_test.c struct mmc_request *mrq, unsigned int blocks) blocks 201 drivers/mmc/core/mmc_test.c mrq->sbc->arg = blocks; blocks 210 drivers/mmc/core/mmc_test.c unsigned dev_addr, unsigned blocks, unsigned blksz, int write) blocks 215 drivers/mmc/core/mmc_test.c if (blocks > 1) { blocks 229 drivers/mmc/core/mmc_test.c if (blocks == 1) blocks 238 drivers/mmc/core/mmc_test.c mrq->data->blocks = blocks; blocks 243 drivers/mmc/core/mmc_test.c mmc_test_prepare_sbc(test, mrq, blocks); blocks 691 drivers/mmc/core/mmc_test.c if (mrq->data->blocks > 1) { blocks 723 drivers/mmc/core/mmc_test.c mrq->data->blocks * mrq->data->blksz) blocks 753 drivers/mmc/core/mmc_test.c if (mrq->data->blocks > 1) { blocks 840 drivers/mmc/core/mmc_test.c unsigned dev_addr, unsigned blocks, blocks 860 drivers/mmc/core/mmc_test.c mmc_test_prepare_mrq(test, mrq, sg, sg_len, dev_addr, blocks, blocks 870 drivers/mmc/core/mmc_test.c dev_addr += blocks; blocks 885 drivers/mmc/core/mmc_test.c unsigned blocks, unsigned blksz, int write) blocks 897 drivers/mmc/core/mmc_test.c blocks, blksz, write); blocks 910 drivers/mmc/core/mmc_test.c unsigned blocks, unsigned blksz, int write) blocks 923 drivers/mmc/core/mmc_test.c sg_init_one(&sg, test->buffer, blocks * blksz); blocks 925 drivers/mmc/core/mmc_test.c mmc_test_prepare_mrq(test, &mrq, &sg, 1, 0, blocks, blksz, write); blocks 942 drivers/mmc/core/mmc_test.c unsigned blocks, unsigned blksz, int write) blocks 948 drivers/mmc/core/mmc_test.c for (i = 0; i < blocks * blksz; i++) blocks 962 drivers/mmc/core/mmc_test.c blocks, blksz, write); blocks 973 drivers/mmc/core/mmc_test.c sectors = (blocks * blksz + 511) / 512; blocks 974 drivers/mmc/core/mmc_test.c if ((sectors * 512) == (blocks * blksz)) blocks 990 drivers/mmc/core/mmc_test.c for (i = 0; i < blocks * blksz; i++) { blocks 1003 drivers/mmc/core/mmc_test.c for (i = 0; i < blocks * blksz; i++) { blocks 1404 drivers/mmc/core/mmc_test.c t->blocks = sz >> 9; blocks 1429 drivers/mmc/core/mmc_test.c t->blocks, 512, write); blocks 1469 drivers/mmc/core/mmc_test.c dev_addr, t->blocks, 512, write, count); blocks 2369 drivers/mmc/core/mmc_test.c mmc_test_prepare_mrq(test, mrq, t->sg, t->sg_len, dev_addr, t->blocks, blocks 2372 drivers/mmc/core/mmc_test.c if (use_sbc && t->blocks > 1 && !mrq->sbc) { blocks 2449 drivers/mmc/core/mmc_test.c if (repeat_cmd && (t->blocks + 1) << 9 > t->max_tfr) blocks 2451 drivers/mmc/core/mmc_test.c mmc_hostname(test->card->host), count, t->blocks); blocks 240 drivers/mmc/core/sd_ops.c data.blocks = 1; blocks 287 drivers/mmc/core/sd_ops.c data.blocks = 1; blocks 328 drivers/mmc/core/sd_ops.c data.blocks = 1; blocks 332 drivers/mmc/core/sdio_io.c unsigned blocks; blocks 334 drivers/mmc/core/sdio_io.c blocks = remainder / func->cur_blksize; blocks 335 drivers/mmc/core/sdio_io.c if (blocks > max_blocks) blocks 336 drivers/mmc/core/sdio_io.c blocks = max_blocks; blocks 337 drivers/mmc/core/sdio_io.c size = blocks * func->cur_blksize; blocks 341 drivers/mmc/core/sdio_io.c blocks, func->cur_blksize); blocks 115 drivers/mmc/core/sdio_ops.c unsigned addr, int incr_addr, u8 *buf, unsigned blocks, unsigned blksz) blocks 139 drivers/mmc/core/sdio_ops.c if (blocks == 0) blocks 142 drivers/mmc/core/sdio_ops.c cmd.arg |= 0x08000000 | blocks; /* block mode */ blocks 147 drivers/mmc/core/sdio_ops.c data.blocks = blocks ? blocks : 1; blocks 150 drivers/mmc/core/sdio_ops.c left_size = data.blksz * data.blocks; blocks 22 drivers/mmc/core/sdio_ops.h unsigned addr, int incr_addr, u8 *buf, unsigned blocks, unsigned blksz); blocks 57 drivers/mmc/host/alcor.c unsigned int blocks; /* remaining PIO blocks */ blocks 191 drivers/mmc/host/alcor.c if (!host->blocks) blocks 214 drivers/mmc/host/alcor.c host->blocks--; blocks 250 drivers/mmc/host/alcor.c host->blocks = data->blocks; blocks 254 drivers/mmc/host/alcor.c host->sg_count, host->blocks); blocks 357 drivers/mmc/host/alcor.c data->bytes_xfered = data->blksz * data->blocks; blocks 514 drivers/mmc/host/alcor.c if (!host->dma_on && host->blocks) { blocks 542 drivers/mmc/host/alcor.c if ((intmask & AU6601_INT_DATA_END) || !host->blocks || blocks 802 drivers/mmc/host/alcor.c if (data->blocks * data->blksz < AU6601_MAX_DMA_BLOCK_SIZE) blocks 372 drivers/mmc/host/android-goldfish.c GOLDFISH_MMC_WRITE(host, MMC_BLOCK_COUNT, data->blocks - 1); blocks 379 drivers/mmc/host/android-goldfish.c sg_len = (data->blocks == 1) ? 1 : data->sg_len; blocks 431 drivers/mmc/host/atmel-mci.c data->bytes_xfered, data->blocks, blocks 795 drivers/mmc/host/atmel-mci.c if (data->blocks > 1) blocks 911 drivers/mmc/host/atmel-mci.c int transfer_size = host->data->blocks * host->data->blksz; blocks 1016 drivers/mmc/host/atmel-mci.c if (data->blocks * data->blksz < 12 blocks 1017 drivers/mmc/host/atmel-mci.c || (data->blocks * data->blksz) & 3) blocks 1062 drivers/mmc/host/atmel-mci.c host->data_size = data->blocks * data->blksz; blocks 1106 drivers/mmc/host/atmel-mci.c if (data->blocks * data->blksz < ATMCI_DMA_THRESHOLD) blocks 1275 drivers/mmc/host/atmel-mci.c atmci_writel(host, ATMCI_BLKR, ATMCI_BCNT(data->blocks) blocks 1278 drivers/mmc/host/atmel-mci.c ATMCI_BCNT(data->blocks) | ATMCI_BLKLEN(data->blksz)); blocks 1364 drivers/mmc/host/atmel-mci.c if (data && data->blocks > 1 && data->blksz & 3) { blocks 1820 drivers/mmc/host/atmel-mci.c data->bytes_xfered = data->blocks * data->blksz; blocks 1854 drivers/mmc/host/atmel-mci.c data->bytes_xfered = data->blocks blocks 1877 drivers/mmc/host/atmel-mci.c data->bytes_xfered = data->blocks * data->blksz; blocks 290 drivers/mmc/host/au1xmmc.c if (data->blocks > 1) blocks 295 drivers/mmc/host/au1xmmc.c if (data->blocks > 1) blocks 376 drivers/mmc/host/au1xmmc.c (data->blocks * data->blksz) - host->pio.len; blocks 620 drivers/mmc/host/au1xmmc.c int datalen = data->blocks * data->blksz; blocks 158 drivers/mmc/host/bcm2835.c unsigned int blocks; /* remaining PIO blocks */ blocks 217 drivers/mmc/host/bcm2835.c mrq->data->blocks, blocks 476 drivers/mmc/host/bcm2835.c if ((data->blocks > 1) && (dir_data == DMA_FROM_DEVICE)) { blocks 482 drivers/mmc/host/bcm2835.c (u32)data->blocks * data->blksz); blocks 567 drivers/mmc/host/bcm2835.c host->blocks = data->blocks; blocks 573 drivers/mmc/host/bcm2835.c writel(data->blocks, host->ioaddr + SDHBLC); blocks 726 drivers/mmc/host/bcm2835.c data->bytes_xfered = data->error ? 0 : (data->blksz * data->blocks); blocks 937 drivers/mmc/host/bcm2835.c host->blocks--; blocks 938 drivers/mmc/host/bcm2835.c if ((host->blocks == 0) || host->data->error) blocks 952 drivers/mmc/host/bcm2835.c if ((host->blocks == 0) || host->data->error) blocks 964 drivers/mmc/host/bcm2835.c WARN_ON(!host->blocks); blocks 965 drivers/mmc/host/bcm2835.c if (host->data->error || (--host->blocks == 0)) blocks 1210 drivers/mmc/host/bcm2835.c if (host->use_dma && mrq->data && (mrq->data->blocks > PIO_THRESHOLD)) blocks 132 drivers/mmc/host/cavium-octeon.c if (data->blksz * data->blocks <= 1024) blocks 135 drivers/mmc/host/cavium-octeon.c host->n_minus_one = addr + (data->blksz * data->blocks) - 1024; blocks 302 drivers/mmc/host/cavium.c int data_len = req->data->blocks * req->data->blksz; blocks 336 drivers/mmc/host/cavium.c req->data->bytes_xfered = req->data->blocks * req->data->blksz; blocks 375 drivers/mmc/host/cavium.c data->bytes_xfered = data->blocks * data->blksz; blocks 392 drivers/mmc/host/cavium.c data->bytes_xfered = data->blocks * data->blksz; blocks 636 drivers/mmc/host/cavium.c FIELD_PREP(MIO_EMM_DMA_BLOCK_CNT, mrq->data->blocks) | blocks 646 drivers/mmc/host/cavium.c mrq->data->blocks, (emm_dma & MIO_EMM_DMA_MULTI) ? 1 : 0); blocks 669 drivers/mmc/host/cavium.c data->blocks, data->blksz, data->blocks * data->blksz); blocks 717 drivers/mmc/host/cavium.c unsigned int data_len = mrq->data->blocks * mrq->data->blksz; blocks 805 drivers/mmc/host/cavium.c 64 - ((cmd->data->blocks * cmd->data->blksz) / 8)); blocks 260 drivers/mmc/host/cb710-mmc.c return !(data->blksz & 15 && (data->blocks != 1 || data->blksz != 8)); blocks 266 drivers/mmc/host/cb710-mmc.c size_t len, blocks = data->blocks; blocks 271 drivers/mmc/host/cb710-mmc.c if (unlikely(data->blksz & 15 && (data->blocks != 1 || data->blksz != 8))) blocks 281 drivers/mmc/host/cb710-mmc.c while (blocks-- > 0) { blocks 310 drivers/mmc/host/cb710-mmc.c size_t len, blocks = data->blocks; blocks 315 drivers/mmc/host/cb710-mmc.c if (unlikely(data->blocks > 1 && data->blksz & 15)) blocks 323 drivers/mmc/host/cb710-mmc.c while (blocks-- > 0) { blocks 437 drivers/mmc/host/cb710-mmc.c data->bytes_xfered = data->blksz * data->blocks; blocks 455 drivers/mmc/host/cb710-mmc.c cb710_mmc_set_transfer_size(slot, data->blocks, data->blksz); blocks 416 drivers/mmc/host/cqhci.c CQHCI_BLK_COUNT(mrq->data->blocks) | blocks 747 drivers/mmc/host/cqhci.c data->bytes_xfered = data->blksz * data->blocks; blocks 535 drivers/mmc/host/davinci_mmc.c data->blocks, data->blksz); blocks 544 drivers/mmc/host/davinci_mmc.c writel(data->blocks, host->base + DAVINCI_MMCNBLK); blocks 563 drivers/mmc/host/davinci_mmc.c host->bytes_left = data->blocks * data->blksz; blocks 930 drivers/mmc/host/davinci_mmc.c data->bytes_xfered = data->blocks * data->blksz; blocks 132 drivers/mmc/host/dw_mmc.c data->bytes_xfered, data->blocks, blocks 883 drivers/mmc/host/dw_mmc.c if (data->blocks * data->blksz < DW_MCI_DMA_THRESHOLD) blocks 1299 drivers/mmc/host/dw_mmc.c mci_writel(host, BYTCNT, data->blksz*data->blocks); blocks 1892 drivers/mmc/host/dw_mmc.c data->bytes_xfered = data->blocks * data->blksz; blocks 2262 drivers/mmc/host/dw_mmc.c (data->blksz * data->blocks)) blocks 2345 drivers/mmc/host/dw_mmc.c (data->blksz * data->blocks)) blocks 2429 drivers/mmc/host/dw_mmc.c (data->blksz * data->blocks)) blocks 684 drivers/mmc/host/jz4740_mmc.c writew(cmd->data->blocks, host->base + JZ_REG_MMC_NOB); blocks 740 drivers/mmc/host/jz4740_mmc.c data->bytes_xfered = data->blocks * data->blksz; blocks 714 drivers/mmc/host/meson-gx-mmc.c if (data->blocks > 1) { blocks 722 drivers/mmc/host/meson-gx-mmc.c if (data->blocks > 1) blocks 770 drivers/mmc/host/meson-gx-mmc.c if (data->blocks > 1) { blocks 773 drivers/mmc/host/meson-gx-mmc.c data->blocks); blocks 779 drivers/mmc/host/meson-gx-mmc.c xfer_bytes = data->blksz * data->blocks; blocks 892 drivers/mmc/host/meson-gx-mmc.c data->bytes_xfered = data->blksz * data->blocks; blocks 951 drivers/mmc/host/meson-gx-mmc.c xfer_bytes = data->blksz * data->blocks; blocks 194 drivers/mmc/host/meson-mx-sdio.c (cmd->data->blocks - 1)); blocks 454 drivers/mmc/host/meson-mx-sdio.c cmd->data->bytes_xfered = cmd->data->blksz * cmd->data->blocks; blocks 874 drivers/mmc/host/mmc_spi.c int multiple = (data->blocks > 1); blocks 516 drivers/mmc/host/mmci.c data->sg_len, data->blksz, data->blocks, data->flags); blocks 822 drivers/mmc/host/mmci.c if (data->blksz * data->blocks <= variant->fifosize) blocks 998 drivers/mmc/host/mmci.c data->blksz, data->blocks, data->flags); blocks 1001 drivers/mmc/host/mmci.c host->size = data->blksz * data->blocks; blocks 1148 drivers/mmc/host/mmci.c success = data->blksz * data->blocks - remain; blocks 1185 drivers/mmc/host/mmci.c data->bytes_xfered = data->blksz * data->blocks; blocks 70 drivers/mmc/host/mmci_qcom_dml.c writel_relaxed(data->blocks * data->blksz, blocks 278 drivers/mmc/host/mmci_stm32_sdmmc.c host->data->blocks == 1) blocks 374 drivers/mmc/host/moxart-mmc.c host->data_len = data->blocks * data->blksz; blocks 950 drivers/mmc/host/mtk-sd.c if (data->blocks > 1) blocks 962 drivers/mmc/host/mtk-sd.c writel(data->blocks, host->base + SDC_BLK_NUM); blocks 982 drivers/mmc/host/mtk-sd.c __func__, cmd->opcode, data->blocks, read); blocks 1283 drivers/mmc/host/mtk-sd.c data->bytes_xfered = data->blocks * data->blksz; blocks 1296 drivers/mmc/host/mtk-sd.c __func__, mrq->cmd->opcode, data->blocks); blocks 1384 drivers/mmc/host/mtk-sd.c host->data->blocks); blocks 98 drivers/mmc/host/mvsdio.c (u32)sg_virt(data->sg), data->blocks, data->blksz, blocks 104 drivers/mmc/host/mvsdio.c mvsd_write(MVSD_BLK_COUNT, data->blocks); blocks 117 drivers/mmc/host/mvsdio.c host->pio_size = data->blocks * data->blksz; blocks 311 drivers/mmc/host/mvsdio.c (data->blocks - mvsd_read(MVSD_CURR_BLK_LEFT)) * data->blksz; blocks 298 drivers/mmc/host/mxcmmc.c unsigned int nob = data->blocks; blocks 154 drivers/mmc/host/mxs-mmc.c data->bytes_xfered = data->blocks * data->blksz; blocks 356 drivers/mmc/host/mxs-mmc.c unsigned int blocks = data->blocks; blocks 398 drivers/mmc/host/mxs-mmc.c if (data_size != data->blocks * data->blksz) blocks 399 drivers/mmc/host/mxs-mmc.c blocks = 1; blocks 405 drivers/mmc/host/mxs-mmc.c BF_SSP(blocks - 1, CMD0_BLOCK_COUNT); blocks 409 drivers/mmc/host/mxs-mmc.c BF_SSP(blocks - 1, BLOCK_SIZE_BLOCK_COUNT), blocks 907 drivers/mmc/host/omap.c data->bytes_xfered += data->blocks * data->blksz; blocks 963 drivers/mmc/host/omap.c OMAP_MMC_WRITE(host, NBLK, data->blocks - 1); blocks 970 drivers/mmc/host/omap.c sg_len = (data->blocks == 1) ? 1 : data->sg_len; blocks 1059 drivers/mmc/host/omap.c host->total_bytes_left = data->blocks * block_size; blocks 862 drivers/mmc/host/omap_hsmmc.c data->bytes_xfered += data->blocks * (data->blksz); blocks 1347 drivers/mmc/host/omap_hsmmc.c | (req->data->blocks << 16)); blocks 165 drivers/mmc/host/pxamci.c unsigned int nob = data->blocks; blocks 357 drivers/mmc/host/pxamci.c data->bytes_xfered = data->blocks * data->blksz; blocks 102 drivers/mmc/host/rtsx_pci_sdmmc.c static void sd_cmd_set_data_len(struct rtsx_pcr *pcr, u16 blocks, u16 blksz) blocks 104 drivers/mmc/host/rtsx_pci_sdmmc.c rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, SD_BLOCK_CNT_L, 0xFF, blocks); blocks 105 drivers/mmc/host/rtsx_pci_sdmmc.c rtsx_pci_add_cmd(pcr, WRITE_REG_CMD, SD_BLOCK_CNT_H, 0xFF, blocks >> 8); blocks 439 drivers/mmc/host/rtsx_pci_sdmmc.c size_t data_len = data->blksz * data->blocks; blocks 453 drivers/mmc/host/rtsx_pci_sdmmc.c sd_cmd_set_data_len(pcr, data->blocks, data->blksz); blocks 496 drivers/mmc/host/rtsx_pci_sdmmc.c size_t data_len = data->blksz * data->blocks; blocks 512 drivers/mmc/host/rtsx_pci_sdmmc.c sd_cmd_set_data_len(pcr, data->blocks, data->blksz); blocks 826 drivers/mmc/host/rtsx_pci_sdmmc.c data_size = data->blocks * data->blksz; blocks 845 drivers/mmc/host/rtsx_pci_sdmmc.c data->bytes_xfered = data->blocks * data->blksz; blocks 457 drivers/mmc/host/rtsx_usb_sdmmc.c size_t data_len = data->blksz * data->blocks; blocks 479 drivers/mmc/host/rtsx_usb_sdmmc.c 0xFF, (u8)data->blocks); blocks 481 drivers/mmc/host/rtsx_usb_sdmmc.c 0xFF, (u8)(data->blocks >> 8)); blocks 838 drivers/mmc/host/rtsx_usb_sdmmc.c data_size = data->blocks * data->blksz; blocks 863 drivers/mmc/host/rtsx_usb_sdmmc.c data->bytes_xfered = data->blocks * data->blksz; blocks 200 drivers/mmc/host/s3cmci.c cmd->data->blocks, blocks 201 drivers/mmc/host/s3cmci.c cmd->data->blocks * cmd->data->blksz); blocks 890 drivers/mmc/host/s3cmci.c (mrq->data->blocks * mrq->data->blksz); blocks 972 drivers/mmc/host/s3cmci.c if (data->blocks > 1) { blocks 994 drivers/mmc/host/s3cmci.c dcon = data->blocks & S3C2410_SDIDCON_BLKNUM_MASK; blocks 562 drivers/mmc/host/sdhci-esdhc-imx.c && (host->cmd->data->blocks > 1) blocks 566 drivers/mmc/host/sdhci.c if (host->blocks == 0) blocks 580 drivers/mmc/host/sdhci.c (host->data->blocks == 1)) blocks 592 drivers/mmc/host/sdhci.c host->blocks--; blocks 593 drivers/mmc/host/sdhci.c if (host->blocks == 0) blocks 614 drivers/mmc/host/sdhci.c unsigned int length = data->blksz * data->blocks; blocks 894 drivers/mmc/host/sdhci.c host->data_timeout = data->blocks * target_timeout + blocks 1035 drivers/mmc/host/sdhci.c BUG_ON(data->blksz * data->blocks > 524288); blocks 1037 drivers/mmc/host/sdhci.c BUG_ON(data->blocks > 65535); blocks 1123 drivers/mmc/host/sdhci.c host->blocks = data->blocks; blocks 1140 drivers/mmc/host/sdhci.c sdhci_writew(host, data->blocks, SDHCI_32BIT_BLK_CNT); blocks 1142 drivers/mmc/host/sdhci.c sdhci_writew(host, data->blocks, SDHCI_BLOCK_COUNT); blocks 1216 drivers/mmc/host/sdhci.c if (mmc_op_multi(cmd->opcode) || data->blocks > 1) { blocks 1317 drivers/mmc/host/sdhci.c data->bytes_xfered = data->blksz * data->blocks; blocks 546 drivers/mmc/host/sdhci.h unsigned int blocks; /* remaining PIO blocks */ blocks 297 drivers/mmc/host/sdricoh_cs.c "sg length %i\n", data->blksz, data->blocks, blocks 302 drivers/mmc/host/sdricoh_cs.c for (i = 0; i < data->blocks; i++) { blocks 862 drivers/mmc/host/sh_mmcif.c data->blocks << 16); blocks 1256 drivers/mmc/host/sh_mmcif.c data->blocks * data->blksz; blocks 537 drivers/mmc/host/sunxi-mmc.c data->bytes_xfered = data->blocks * data->blksz; blocks 1064 drivers/mmc/host/sunxi-mmc.c mrq->data ? mrq->data->blksz * mrq->data->blocks : 0); blocks 1083 drivers/mmc/host/sunxi-mmc.c mmc_writel(host, REG_BCNTR, data->blksz * data->blocks); blocks 713 drivers/mmc/host/tifm_sd.c writel(r_data->blocks - 1, blocks 769 drivers/mmc/host/tifm_sd.c r_data->bytes_xfered = r_data->blocks blocks 275 drivers/mmc/host/tmio_mmc_core.c if (data->blocks > 1) { blocks 432 drivers/mmc/host/tmio_mmc_core.c data->bytes_xfered = data->blocks * data->blksz; blocks 687 drivers/mmc/host/tmio_mmc_core.c data->blksz, data->blocks); blocks 708 drivers/mmc/host/tmio_mmc_core.c sd_ctrl_write32(host, CTL_XFER_BLK_COUNT, data->blocks); blocks 710 drivers/mmc/host/tmio_mmc_core.c sd_ctrl_write16(host, CTL_XFER_BLK_COUNT, data->blocks); blocks 277 drivers/mmc/host/toshsd.c data->bytes_xfered = data->blocks * data->blksz; blocks 457 drivers/mmc/host/toshsd.c if (data->blocks > 1) { blocks 479 drivers/mmc/host/toshsd.c data->blksz, data->blocks, data->sg->offset); blocks 491 drivers/mmc/host/toshsd.c iowrite16(data->blocks, host->ioaddr + SD_BLOCKCOUNT); blocks 321 drivers/mmc/host/usdhi6rol0.c data->blksz, data->blocks, sg->offset); blocks 346 drivers/mmc/host/usdhi6rol0.c usdhi6_write(host, USDHI6_SD_SECCNT, data->blocks); blocks 630 drivers/mmc/host/usdhi6rol0.c __func__, data->sg_len, data->blocks, data->blksz); blocks 905 drivers/mmc/host/usdhi6rol0.c data ? host->offset : 0, data ? data->blocks : 0, blocks 913 drivers/mmc/host/usdhi6rol0.c __func__, mrq->cmd->opcode, data ? data->blocks : 0, blocks 947 drivers/mmc/host/usdhi6rol0.c mrq->data->blocks > 1)) { blocks 1008 drivers/mmc/host/usdhi6rol0.c if (cmd->opcode == SD_IO_RW_EXTENDED && data->blocks > 1) { blocks 1030 drivers/mmc/host/usdhi6rol0.c __func__, data->blocks, data->blksz); blocks 1037 drivers/mmc/host/usdhi6rol0.c data->blocks > 1)) blocks 1043 drivers/mmc/host/usdhi6rol0.c data->blocks > 1) && blocks 1048 drivers/mmc/host/usdhi6rol0.c data->blksz, data->blocks, data->sg->offset); blocks 1060 drivers/mmc/host/usdhi6rol0.c __func__, cmd->opcode, data->blocks, data->blksz, blocks 1077 drivers/mmc/host/usdhi6rol0.c data && data->blocks > 1 ? USDHI6_SD_STOP_SEC : 0); blocks 1377 drivers/mmc/host/usdhi6rol0.c mrq->data->blocks > 1)) blocks 1384 drivers/mmc/host/usdhi6rol0.c mrq->data->blocks > 1)) blocks 1550 drivers/mmc/host/usdhi6rol0.c data->bytes_xfered = data->blocks * data->blksz; blocks 1709 drivers/mmc/host/usdhi6rol0.c host->offset, data->blocks, data->blksz, data->sg_len, blocks 239 drivers/mmc/host/ushc.c req->data->bytes_xfered = req->data->blksz * req->data->blocks; blocks 486 drivers/mmc/host/via-sdmmc.c BUG_ON(data->blocks > host->mmc->max_blk_count); blocks 502 drivers/mmc/host/via-sdmmc.c blk_reg |= (data->blocks) << 16; blocks 593 drivers/mmc/host/via-sdmmc.c if (data->blocks > 1) { blocks 632 drivers/mmc/host/via-sdmmc.c data->bytes_xfered = data->blocks * data->blksz; blocks 1101 drivers/mmc/host/vub300.c (data->blocks >> 8) & 0xFF; blocks 1103 drivers/mmc/host/vub300.c (data->blocks >> 0) & 0xFF; blocks 1131 drivers/mmc/host/vub300.c vub300->cmnd.head.block_count[0] = (data->blocks >> 8) & 0xFF; blocks 1132 drivers/mmc/host/vub300.c vub300->cmnd.head.block_count[1] = (data->blocks >> 0) & 0xFF; blocks 1947 drivers/mmc/host/vub300.c vub300->datasize = data->blksz * data->blocks; blocks 459 drivers/mmc/host/wbsd.c if ((data->blocks * data->blksz - data->bytes_xfered) < 16) blocks 538 drivers/mmc/host/wbsd.c size = data->blocks * data->blksz; blocks 703 drivers/mmc/host/wbsd.c data->bytes_xfered = host->mrq->data->blocks * blocks 299 drivers/mmc/host/wmt-sdmmc.c req->data->bytes_xfered = req->data->blksz * req->data->blocks; blocks 615 drivers/mmc/host/wmt-sdmmc.c writew(req->data->blocks, priv->sdmmc_base + SDMMC_BLKCNT); blocks 623 drivers/mmc/host/wmt-sdmmc.c if (req->data->blocks > 1) blocks 629 drivers/mmc/host/wmt-sdmmc.c if (req->data->blocks > 1) blocks 646 drivers/mmc/host/wmt-sdmmc.c if (desc_cnt == req->data->blocks) blocks 279 drivers/mtd/chips/jedec_probe.c #define ERASEINFO(size,blocks) (size<<8)|(blocks-1) blocks 196 drivers/mtd/ftl.c unsigned blocks, j; blocks 265 drivers/mtd/ftl.c blocks = le32_to_cpu(header.FormattedSize) >> header.BlockSize; blocks 266 drivers/mtd/ftl.c part->VirtualBlockMap = vmalloc(array_size(blocks, sizeof(uint32_t))); blocks 270 drivers/mtd/ftl.c memset(part->VirtualBlockMap, 0xff, blocks * sizeof(uint32_t)); blocks 298 drivers/mtd/ftl.c (BLOCK_NUMBER(le32_to_cpu(part->bam_cache[j])) < blocks)) blocks 1282 drivers/mtd/mtdswap.c unsigned int i, eblk_bytes, pages, blocks; blocks 1291 drivers/mtd/mtdswap.c blocks = eblocks * d->pages_per_eblk; blocks 1300 drivers/mtd/mtdswap.c d->revmap = vmalloc(array_size(blocks, sizeof(int))); blocks 1312 drivers/mtd/mtdswap.c for (i = 0; i < blocks; i++) blocks 1038 drivers/mtd/nand/raw/diskonchip.c unsigned blocks, maxblocks; blocks 1064 drivers/mtd/nand/raw/diskonchip.c blocks = mtd->size >> this->phys_erase_shift; blocks 1074 drivers/mtd/nand/raw/diskonchip.c while (blocks > maxblocks) { blocks 1075 drivers/mtd/nand/raw/diskonchip.c blocks >>= 1; blocks 1092 drivers/mtd/nand/raw/diskonchip.c blocks = mtd->size >> this->bbt_erase_shift; blocks 1096 drivers/mtd/nand/raw/diskonchip.c if (blocks > maxblocks) { blocks 1143 drivers/mtd/nand/raw/diskonchip.c int blocks; blocks 1188 drivers/mtd/nand/raw/diskonchip.c blocks = mtd->size >> vshift; blocks 1189 drivers/mtd/nand/raw/diskonchip.c if (blocks > 32768) { blocks 1194 drivers/mtd/nand/raw/diskonchip.c blocks = doc->chips_per_floor << (this->chip_shift - this->phys_erase_shift); blocks 1195 drivers/mtd/nand/raw/diskonchip.c if (inftl_bbt_write && (blocks > mtd->erasesize)) { blocks 88 drivers/mtd/rfd_ftl.c struct block *blocks; blocks 95 drivers/mtd/rfd_ftl.c struct block *block = &part->blocks[block_no]; blocks 188 drivers/mtd/rfd_ftl.c part->blocks = kcalloc(part->total_blocks, sizeof(struct block), blocks 190 drivers/mtd/rfd_ftl.c if (!part->blocks) blocks 238 drivers/mtd/rfd_ftl.c kfree(part->blocks); blocks 280 drivers/mtd/rfd_ftl.c erase->addr = part->blocks[block].offset; blocks 283 drivers/mtd/rfd_ftl.c part->blocks[block].state = BLOCK_ERASING; blocks 284 drivers/mtd/rfd_ftl.c part->blocks[block].free_sectors = 0; blocks 291 drivers/mtd/rfd_ftl.c part->blocks[block].state = BLOCK_FAILED; blocks 292 drivers/mtd/rfd_ftl.c part->blocks[block].free_sectors = 0; blocks 293 drivers/mtd/rfd_ftl.c part->blocks[block].used_sectors = 0; blocks 298 drivers/mtd/rfd_ftl.c part->blocks[block].state = BLOCK_ERASED; blocks 299 drivers/mtd/rfd_ftl.c part->blocks[block].free_sectors = part->data_sectors_per_block; blocks 300 drivers/mtd/rfd_ftl.c part->blocks[block].used_sectors = 0; blocks 301 drivers/mtd/rfd_ftl.c part->blocks[block].erases++; blocks 303 drivers/mtd/rfd_ftl.c rc = mtd_write(part->mbd.mtd, part->blocks[block].offset, blocks 310 drivers/mtd/rfd_ftl.c part->mbd.mtd->name, part->blocks[block].offset); blocks 311 drivers/mtd/rfd_ftl.c part->blocks[block].state = BLOCK_FAILED; blocks 313 drivers/mtd/rfd_ftl.c part->blocks[block].state = BLOCK_OK; blocks 339 drivers/mtd/rfd_ftl.c rc = mtd_read(part->mbd.mtd, part->blocks[block_no].offset, blocks 348 drivers/mtd/rfd_ftl.c part->blocks[block_no].offset); blocks 368 drivers/mtd/rfd_ftl.c addr = part->blocks[block_no].offset + blocks 373 drivers/mtd/rfd_ftl.c if (!part->blocks[block_no].used_sectors--) { blocks 436 drivers/mtd/rfd_ftl.c if (part->blocks[block].free_sectors) blocks 439 drivers/mtd/rfd_ftl.c this_score = part->blocks[block].used_sectors; blocks 445 drivers/mtd/rfd_ftl.c if (part->blocks[block].used_sectors == blocks 450 drivers/mtd/rfd_ftl.c this_score += part->blocks[block].erases; blocks 466 drivers/mtd/rfd_ftl.c part->blocks[best_block].used_sectors, blocks 467 drivers/mtd/rfd_ftl.c part->blocks[best_block].free_sectors); blocks 469 drivers/mtd/rfd_ftl.c if (part->blocks[best_block].used_sectors) blocks 491 drivers/mtd/rfd_ftl.c if (part->blocks[block].free_sectors && blocks 495 drivers/mtd/rfd_ftl.c if (part->blocks[block].state == BLOCK_UNUSED) blocks 528 drivers/mtd/rfd_ftl.c rc = mtd_read(part->mbd.mtd, part->blocks[block].offset, blocks 538 drivers/mtd/rfd_ftl.c part->blocks[block].offset); blocks 559 drivers/mtd/rfd_ftl.c addr = part->blocks[block].offset + blocks 575 drivers/mtd/rfd_ftl.c part->blocks[block].used_sectors--; blocks 577 drivers/mtd/rfd_ftl.c if (!part->blocks[block].used_sectors && blocks 578 drivers/mtd/rfd_ftl.c !part->blocks[block].free_sectors) blocks 615 drivers/mtd/rfd_ftl.c !part->blocks[part->current_block].free_sectors) { blocks 622 drivers/mtd/rfd_ftl.c block = &part->blocks[part->current_block]; blocks 774 drivers/mtd/rfd_ftl.c part->mbd.mtd->name, i, part->blocks[i].erases); blocks 780 drivers/mtd/rfd_ftl.c kfree(part->blocks); blocks 43 drivers/mtd/tests/speedtest.c static int multiblock_erase(int ebnum, int blocks) blocks 51 drivers/mtd/tests/speedtest.c ei.len = mtd->erasesize * blocks; blocks 56 drivers/mtd/tests/speedtest.c err, ebnum, blocks); blocks 176 drivers/mtd/tests/speedtest.c int err, i, blocks, j, k; blocks 372 drivers/mtd/tests/speedtest.c blocks = 1 << k; blocks 374 drivers/mtd/tests/speedtest.c blocks); blocks 377 drivers/mtd/tests/speedtest.c for (j = 0; j < blocks && (i + j) < ebcnt; j++) blocks 397 drivers/mtd/tests/speedtest.c blocks, speed); blocks 78 drivers/net/ethernet/cavium/liquidio/octeon_main.h static inline void octeon_swap_8B_data(u64 *data, u32 blocks) blocks 80 drivers/net/ethernet/cavium/liquidio/octeon_main.h while (blocks) { blocks 82 drivers/net/ethernet/cavium/liquidio/octeon_main.h blocks--; blocks 16 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c const struct mlxsw_afk_block *blocks; blocks 26 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c const struct mlxsw_afk_block *block = &mlxsw_afk->blocks[i]; blocks 53 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c mlxsw_afk->blocks = ops->blocks; blocks 75 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c const struct mlxsw_afk_block *blocks[0]; blocks 113 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c const struct mlxsw_afk_block *block = &mlxsw_afk->blocks[i]; blocks 178 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c key_info->blocks[key_info->blocks_count] = blocks 179 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c &mlxsw_afk->blocks[block_index]; blocks 242 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c key_info = kzalloc(struct_size(key_info, blocks, mlxsw_afk->max_blocks), blocks 322 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c block = key_info->blocks[block_index]; blocks 336 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c return key_info->blocks[block_index]->encoding; blocks 207 drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.h const struct mlxsw_afk_block *blocks; blocks 119 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_flex_keys.c .blocks = mlxsw_sp1_afk_blocks, blocks 309 drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_flex_keys.c .blocks = mlxsw_sp2_afk_blocks, blocks 2237 drivers/net/ethernet/qlogic/qed/qed_cxt.c p_info->blocks[i] = p_mngr->ilt_shadow[shadow_line].p_virt; blocks 53 drivers/net/ethernet/qlogic/qed/qed_cxt.h u8 *blocks[MAX_TID_BLOCKS]; /* 4K */ blocks 863 drivers/net/ethernet/qlogic/qed/qed_fcoe.c memcpy(tasks->blocks, tid_info->blocks, blocks 2330 drivers/net/ethernet/qlogic/qed/qed_hsi.h struct dbg_bus_block_data blocks[88]; blocks 1195 drivers/net/ethernet/qlogic/qed/qed_iscsi.c memcpy(tasks->blocks, tid_info->blocks, blocks 230 drivers/net/wireless/ath/ath6kl/sdio.c data->blocks = scat_req->len / HIF_MBOX_BLOCK_SIZE; blocks 235 drivers/net/wireless/ath/ath6kl/sdio.c data->blksz, data->blocks, scat_req->len, blocks 306 drivers/net/wireless/ath/ath6kl/sdio.c data.blocks); blocks 343 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c md->blocks = req_sz / func_blk_sz; blocks 345 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c mc->arg |= md->blocks & 0x1FF; /* block count */ blocks 7519 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c uint *blocks) blocks 7524 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c *blocks = wlc_hw->xmtfifo_sz[fifo]; blocks 622 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.h uint *blocks); blocks 47 drivers/net/wireless/intersil/orinoco/fw.c __le32 blocks; /* Number of blocks to program */ blocks 365 drivers/s390/block/dasd.c block->blocks << block->s2b_shift); blocks 409 drivers/s390/block/dasd.c block->blocks = 0; blocks 437 drivers/s390/block/dasd_diag.c block->blocks = (unsigned long) label->block_count; blocks 439 drivers/s390/block/dasd_diag.c block->blocks = end_block; blocks 455 drivers/s390/block/dasd_diag.c (unsigned long) (block->blocks << blocks 479 drivers/s390/block/dasd_diag.c geo->cylinders = (block->blocks << block->s2b_shift) >> 10; blocks 2339 drivers/s390/block/dasd_eckd.c block->blocks = ((unsigned long) private->real_cyl * blocks 206 drivers/s390/block/dasd_fba.c block->blocks = private->rdc_data.blk_bdsa; blocks 219 drivers/s390/block/dasd_fba.c geo->cylinders = (block->blocks << block->s2b_shift) >> 10; blocks 561 drivers/s390/block/dasd_int.h unsigned long blocks; /* size of volume in blocks */ blocks 95 drivers/s390/block/dasd_proc.c block->bp_block, block->blocks, blocks 97 drivers/s390/block/dasd_proc.c block->blocks) >> 11); blocks 147 drivers/scsi/myrs.c unsigned char *msg, unsigned long blocks, blocks 153 drivers/scsi/myrs.c (100 * (int)(blocks >> 7)) / (int)(size >> 7)); blocks 43 drivers/scsi/ps3rom.c u32 blocks; blocks 99 drivers/scsi/ps3rom.c atapi_cmnd.blocks = atapi_cmnd.arglen = scsi_bufflen(cmd); blocks 447 drivers/scsi/qedf/qedf.h return (void *)(info->blocks[tid / info->num_tids_per_block] + blocks 376 drivers/scsi/qedi/qedi.h return (info->blocks[tid / info->num_tids_per_block] + blocks 3467 drivers/scsi/scsi_debug.c __be32 blocks; blocks 3509 drivers/scsi/scsi_debug.c unsigned int num = get_unaligned_be32(&desc[i].blocks); blocks 171 drivers/scsi/sd.h static inline sector_t logical_to_sectors(struct scsi_device *sdev, sector_t blocks) blocks 173 drivers/scsi/sd.h return blocks << (ilog2(sdev->sector_size) - 9); blocks 176 drivers/scsi/sd.h static inline unsigned int logical_to_bytes(struct scsi_device *sdev, sector_t blocks) blocks 178 drivers/scsi/sd.h return blocks * sdev->sector_size; blocks 516 drivers/soc/ixp4xx/ixp4xx-npe.c struct dl_block blocks[0]; blocks 526 drivers/soc/ixp4xx/ixp4xx-npe.c int i, j, err, data_size, instr_size, blocks, table_end; blocks 597 drivers/soc/ixp4xx/ixp4xx-npe.c for (blocks = 0; blocks * sizeof(struct dl_block) / 4 < image->size; blocks 598 drivers/soc/ixp4xx/ixp4xx-npe.c blocks++) blocks 599 drivers/soc/ixp4xx/ixp4xx-npe.c if (image->blocks[blocks].type == FW_BLOCK_TYPE_EOF) blocks 601 drivers/soc/ixp4xx/ixp4xx-npe.c if (blocks * sizeof(struct dl_block) / 4 >= image->size) { blocks 608 drivers/soc/ixp4xx/ixp4xx-npe.c print_npe(KERN_DEBUG, npe, "%i firmware blocks found\n", blocks); blocks 611 drivers/soc/ixp4xx/ixp4xx-npe.c table_end = blocks * sizeof(struct dl_block) / 4 + 1 /* EOF marker */; blocks 612 drivers/soc/ixp4xx/ixp4xx-npe.c for (i = 0, blk = image->blocks; i < blocks; i++, blk++) { blocks 369 drivers/staging/greybus/sdio.c if (single_op(data->mrq->cmd) && data->blocks > 1) { blocks 374 drivers/staging/greybus/sdio.c left = data->blksz * data->blocks; blocks 468 drivers/staging/greybus/sdio.c request.data_blocks = cpu_to_le16(data->blocks); blocks 116 drivers/staging/most/dim2/hal.c u32 const blocks = DIV_ROUND_UP(size, DBR_BLOCK_SIZE); blocks 117 drivers/staging/most/dim2/hal.c u32 mask = ~((~(u32)0) << blocks); blocks 136 drivers/staging/most/dim2/hal.c u32 const blocks = DIV_ROUND_UP(size, DBR_BLOCK_SIZE); blocks 137 drivers/staging/most/dim2/hal.c u32 mask = ~((~(u32)0) << blocks); blocks 39 drivers/target/target_core_sbc.c u32 blocks; blocks 57 drivers/target/target_core_sbc.c blocks = 0xffffffff; blocks 59 drivers/target/target_core_sbc.c blocks = (u32)blocks_long; blocks 61 drivers/target/target_core_sbc.c put_unaligned_be32(blocks, &buf[0]); blocks 83 drivers/target/target_core_sbc.c unsigned long long blocks = dev->transport->get_blocks(dev); blocks 86 drivers/target/target_core_sbc.c put_unaligned_be64(blocks, &buf[0]); blocks 957 drivers/target/target_core_spc.c static int spc_modesense_blockdesc(unsigned char *buf, u64 blocks, u32 block_size) blocks 960 drivers/target/target_core_spc.c put_unaligned_be32(min(blocks, 0xffffffffull), buf); blocks 966 drivers/target/target_core_spc.c static int spc_modesense_long_blockdesc(unsigned char *buf, u64 blocks, u32 block_size) blocks 968 drivers/target/target_core_spc.c if (blocks <= 0xffffffff) blocks 969 drivers/target/target_core_spc.c return spc_modesense_blockdesc(buf + 3, blocks, block_size) + 3; blocks 974 drivers/target/target_core_spc.c put_unaligned_be64(blocks, buf); blocks 1026 drivers/target/target_core_spc.c u64 blocks = dev->transport->get_blocks(dev); blocks 1032 drivers/target/target_core_spc.c blocks, block_size); blocks 1036 drivers/target/target_core_spc.c blocks, block_size); blocks 1039 drivers/target/target_core_spc.c length += spc_modesense_blockdesc(&buf[length], blocks, blocks 1618 drivers/target/target_core_user.c static void tcmu_blocks_release(struct radix_tree_root *blocks, blocks 1625 drivers/target/target_core_user.c page = radix_tree_delete(blocks, i); blocks 552 drivers/usb/storage/datafab.c unsigned long block, blocks; blocks 605 drivers/usb/storage/datafab.c blocks = ((u32)(srb->cmnd[7]) << 8) | ((u32)(srb->cmnd[8])); blocks 608 drivers/usb/storage/datafab.c block, blocks); blocks 609 drivers/usb/storage/datafab.c return datafab_read_data(us, info, block, blocks); blocks 618 drivers/usb/storage/datafab.c blocks = ((u32)(srb->cmnd[6]) << 24) | ((u32)(srb->cmnd[7]) << 16) | blocks 622 drivers/usb/storage/datafab.c block, blocks); blocks 623 drivers/usb/storage/datafab.c return datafab_read_data(us, info, block, blocks); blocks 630 drivers/usb/storage/datafab.c blocks = ((u32)(srb->cmnd[7]) << 8) | ((u32)(srb->cmnd[8])); blocks 633 drivers/usb/storage/datafab.c block, blocks); blocks 634 drivers/usb/storage/datafab.c return datafab_write_data(us, info, block, blocks); blocks 643 drivers/usb/storage/datafab.c blocks = ((u32)(srb->cmnd[6]) << 24) | ((u32)(srb->cmnd[7]) << 16) | blocks 647 drivers/usb/storage/datafab.c block, blocks); blocks 648 drivers/usb/storage/datafab.c return datafab_write_data(us, info, block, blocks); blocks 478 drivers/usb/storage/jumpshot.c unsigned long block, blocks; blocks 533 drivers/usb/storage/jumpshot.c blocks = ((u32)(srb->cmnd[7]) << 8) | ((u32)(srb->cmnd[8])); blocks 536 drivers/usb/storage/jumpshot.c block, blocks); blocks 537 drivers/usb/storage/jumpshot.c return jumpshot_read_data(us, info, block, blocks); blocks 546 drivers/usb/storage/jumpshot.c blocks = ((u32)(srb->cmnd[6]) << 24) | ((u32)(srb->cmnd[7]) << 16) | blocks 550 drivers/usb/storage/jumpshot.c block, blocks); blocks 551 drivers/usb/storage/jumpshot.c return jumpshot_read_data(us, info, block, blocks); blocks 558 drivers/usb/storage/jumpshot.c blocks = ((u32)(srb->cmnd[7]) << 8) | ((u32)(srb->cmnd[8])); blocks 561 drivers/usb/storage/jumpshot.c block, blocks); blocks 562 drivers/usb/storage/jumpshot.c return jumpshot_write_data(us, info, block, blocks); blocks 571 drivers/usb/storage/jumpshot.c blocks = ((u32)(srb->cmnd[6]) << 24) | ((u32)(srb->cmnd[7]) << 16) | blocks 575 drivers/usb/storage/jumpshot.c block, blocks); blocks 576 drivers/usb/storage/jumpshot.c return jumpshot_write_data(us, info, block, blocks); blocks 1049 drivers/usb/storage/sddr09.c unsigned int blocks, blocks 1054 drivers/usb/storage/sddr09.c address, blocks); blocks 1056 drivers/usb/storage/sddr09.c return sddr09_read21(us, address, blocks, blocks 1684 drivers/usb/storage/shuttle_usbat.c unsigned long block, blocks; blocks 1732 drivers/usb/storage/shuttle_usbat.c blocks = ((u32)(srb->cmnd[7]) << 8) | ((u32)(srb->cmnd[8])); blocks 1735 drivers/usb/storage/shuttle_usbat.c block, blocks); blocks 1736 drivers/usb/storage/shuttle_usbat.c return usbat_flash_read_data(us, info, block, blocks); blocks 1746 drivers/usb/storage/shuttle_usbat.c blocks = ((u32)(srb->cmnd[6]) << 24) | ((u32)(srb->cmnd[7]) << 16) | blocks 1750 drivers/usb/storage/shuttle_usbat.c block, blocks); blocks 1751 drivers/usb/storage/shuttle_usbat.c return usbat_flash_read_data(us, info, block, blocks); blocks 1758 drivers/usb/storage/shuttle_usbat.c blocks = ((u32)(srb->cmnd[7]) << 8) | ((u32)(srb->cmnd[8])); blocks 1761 drivers/usb/storage/shuttle_usbat.c block, blocks); blocks 1762 drivers/usb/storage/shuttle_usbat.c return usbat_flash_write_data(us, info, block, blocks); blocks 1772 drivers/usb/storage/shuttle_usbat.c blocks = ((u32)(srb->cmnd[6]) << 24) | ((u32)(srb->cmnd[7]) << 16) | blocks 1776 drivers/usb/storage/shuttle_usbat.c block, blocks); blocks 1777 drivers/usb/storage/shuttle_usbat.c return usbat_flash_write_data(us, info, block, blocks); blocks 256 fs/9p/vfs_super.c buf->f_blocks = rs.blocks; blocks 134 fs/afs/dir.c if (dbuf->blocks[tmp].hdr.magic != AFS_DIR_MAGIC) { blocks 137 fs/afs/dir.c ntohs(dbuf->blocks[tmp].hdr.magic)); blocks 148 fs/afs/dir.c ((u8 *)&dbuf->blocks[tmp])[AFS_DIR_BLOCK_SIZE - 1] = 0; blocks 184 fs/afs/dir.c union afs_xdr_dir_block *block = &dbuf->blocks[j]; blocks 490 fs/afs/dir.c dblock = &dbuf->blocks[(blkoff % PAGE_SIZE) / blocks 222 fs/afs/dir_edit.c meta = &meta_page->blocks[0]; blocks 257 fs/afs/dir_edit.c block = &dir_page->blocks[b % AFS_DIR_BLOCKS_PER_PAGE]; blocks 402 fs/afs/dir_edit.c meta = &meta_page->blocks[0]; blocks 423 fs/afs/dir_edit.c block = &dir_page->blocks[b % AFS_DIR_BLOCKS_PER_PAGE]; blocks 96 fs/afs/xdr_fs.h union afs_xdr_dir_block blocks[AFS_DIR_BLOCKS_PER_PAGE]; blocks 174 fs/befs/datastream.c befs_blocknr_t blocks; blocks 215 fs/befs/datastream.c blocks = datablocks + metablocks; blocks 216 fs/befs/datastream.c befs_debug(sb, "<--- %s %u blocks", __func__, (unsigned int)blocks); blocks 218 fs/befs/datastream.c return blocks; blocks 1339 fs/btrfs/backref.c static void free_leaf_list(struct ulist *blocks) blocks 1346 fs/btrfs/backref.c while ((node = ulist_next(blocks, &uiter))) { blocks 1354 fs/btrfs/backref.c ulist_free(blocks); blocks 953 fs/btrfs/ctree.h struct rb_root blocks[BTRFS_MAX_LEVEL]; blocks 9598 fs/btrfs/inode.c stat->blocks = (ALIGN(inode_get_bytes(inode), blocksize) + blocks 3792 fs/btrfs/qgroup.c swapped_blocks->blocks[i] = RB_ROOT; blocks 3813 fs/btrfs/qgroup.c struct rb_root *cur_root = &swapped_blocks->blocks[i]; blocks 3820 fs/btrfs/qgroup.c swapped_blocks->blocks[i] = RB_ROOT; blocks 3845 fs/btrfs/qgroup.c struct btrfs_qgroup_swapped_blocks *blocks = &subvol_root->swapped_blocks; blocks 3896 fs/btrfs/qgroup.c spin_lock(&blocks->lock); blocks 3897 fs/btrfs/qgroup.c cur = &blocks->blocks[level].rb_node; blocks 3930 fs/btrfs/qgroup.c rb_insert_color(&block->node, &blocks->blocks[level]); blocks 3931 fs/btrfs/qgroup.c blocks->swapped = true; blocks 3933 fs/btrfs/qgroup.c spin_unlock(&blocks->lock); blocks 3952 fs/btrfs/qgroup.c struct btrfs_qgroup_swapped_blocks *blocks = &root->swapped_blocks; blocks 3967 fs/btrfs/qgroup.c spin_lock(&blocks->lock); blocks 3968 fs/btrfs/qgroup.c if (!blocks->swapped) { blocks 3969 fs/btrfs/qgroup.c spin_unlock(&blocks->lock); blocks 3972 fs/btrfs/qgroup.c node = blocks->blocks[level].rb_node; blocks 3986 fs/btrfs/qgroup.c spin_unlock(&blocks->lock); blocks 3990 fs/btrfs/qgroup.c rb_erase(&block->node, &blocks->blocks[level]); blocks 3992 fs/btrfs/qgroup.c if (RB_EMPTY_ROOT(&blocks->blocks[i])) { blocks 3997 fs/btrfs/qgroup.c blocks->swapped = swapped; blocks 3998 fs/btrfs/qgroup.c spin_unlock(&blocks->lock); blocks 2584 fs/btrfs/relocation.c static void free_block_list(struct rb_root *blocks) blocks 2588 fs/btrfs/relocation.c while ((rb_node = rb_first(blocks))) { blocks 2590 fs/btrfs/relocation.c rb_erase(rb_node, blocks); blocks 3138 fs/btrfs/relocation.c struct reloc_control *rc, struct rb_root *blocks) blocks 3155 fs/btrfs/relocation.c rbtree_postorder_for_each_entry_safe(block, next, blocks, rb_node) { blocks 3161 fs/btrfs/relocation.c rbtree_postorder_for_each_entry_safe(block, next, blocks, rb_node) { blocks 3170 fs/btrfs/relocation.c rbtree_postorder_for_each_entry_safe(block, next, blocks, rb_node) { blocks 3191 fs/btrfs/relocation.c free_block_list(blocks); blocks 3450 fs/btrfs/relocation.c struct rb_root *blocks) blocks 3497 fs/btrfs/relocation.c rb_node = tree_insert(blocks, block->bytenr, &block->rb_node); blocks 3509 fs/btrfs/relocation.c struct rb_root *blocks) blocks 3520 fs/btrfs/relocation.c if (tree_search(blocks, bytenr)) blocks 3571 fs/btrfs/relocation.c ret = add_tree_block(rc, &key, path, blocks); blocks 3652 fs/btrfs/relocation.c struct rb_root *blocks) blocks 3724 fs/btrfs/relocation.c rb_node = tree_search(blocks, leaf->start); blocks 3750 fs/btrfs/relocation.c rb_node = tree_search(blocks, leaf->start); blocks 3794 fs/btrfs/relocation.c rb_node = tree_insert(blocks, block->bytenr, blocks 3820 fs/btrfs/relocation.c struct rb_root *blocks) blocks 3844 fs/btrfs/relocation.c blocks); blocks 3848 fs/btrfs/relocation.c eb, dref, blocks); blocks 3883 fs/btrfs/relocation.c blocks); blocks 3888 fs/btrfs/relocation.c eb, dref, blocks); blocks 3905 fs/btrfs/relocation.c free_block_list(blocks); blocks 4072 fs/btrfs/relocation.c struct rb_root blocks = RB_ROOT; blocks 4142 fs/btrfs/relocation.c ret = add_tree_block(rc, &key, path, &blocks); blocks 4145 fs/btrfs/relocation.c ret = add_data_references(rc, &key, path, &blocks); blocks 4155 fs/btrfs/relocation.c if (!RB_EMPTY_ROOT(&blocks)) { blocks 4156 fs/btrfs/relocation.c ret = relocate_tree_blocks(trans, rc, &blocks); blocks 2368 fs/ceph/inode.c stat->blocks = 0; blocks 42 fs/cramfs/inode.c unsigned long blocks; blocks 573 fs/cramfs/inode.c sbi->blocks = super->fsid.blocks; blocks 577 fs/cramfs/inode.c sbi->blocks = 0; blocks 688 fs/cramfs/inode.c buf->f_blocks = CRAMFS_SB(sb)->blocks; blocks 1007 fs/ecryptfs/inode.c stat->blocks = lower_stat.blocks; blocks 34 fs/erofs/erofs_fs.h __le32 blocks; /* used for statfs */ blocks 69 fs/erofs/internal.h u32 blocks; blocks 135 fs/erofs/super.c sbi->blocks = le32_to_cpu(dsb->blocks); blocks 533 fs/erofs/super.c buf->f_blocks = sbi->blocks; blocks 162 fs/ext4/balloc.c unsigned int blocks; blocks 171 fs/ext4/balloc.c blocks = ext4_blocks_count(EXT4_SB(sb)->s_es) - blocks 174 fs/ext4/balloc.c blocks = EXT4_BLOCKS_PER_GROUP(sb); blocks 175 fs/ext4/balloc.c return EXT4_NUM_B2C(EXT4_SB(sb), blocks); blocks 68 fs/ext4/ext4_jbd2.c int type, int blocks, int rsv_blocks) blocks 73 fs/ext4/ext4_jbd2.c trace_ext4_journal_start(sb, blocks, rsv_blocks, _RET_IP_); blocks 81 fs/ext4/ext4_jbd2.c return jbd2__journal_start(journal, blocks, rsv_blocks, GFP_NOFS, blocks 264 fs/ext4/ext4_jbd2.h int type, int blocks, int rsv_blocks); blocks 304 fs/ext4/ext4_jbd2.h #define ext4_journal_start_with_reserve(inode, type, blocks, rsv_blocks) \ blocks 305 fs/ext4/ext4_jbd2.h __ext4_journal_start((inode), __LINE__, (type), (blocks), (rsv_blocks)) blocks 309 fs/ext4/ext4_jbd2.h int blocks, int rsv_blocks) blocks 311 fs/ext4/ext4_jbd2.h return __ext4_journal_start_sb(inode->i_sb, line, type, blocks, blocks 2323 fs/ext4/inode.c ext4_lblk_t blocks = (i_size_read(inode) + i_blocksize(inode) - 1) blocks 2327 fs/ext4/inode.c blocks = EXT_MAX_BLOCKS; blocks 2332 fs/ext4/inode.c if (lblk >= blocks || !mpage_add_bh_to_extent(mpd, lblk, bh)) { blocks 2349 fs/ext4/inode.c return lblk < blocks; blocks 5783 fs/ext4/inode.c stat->blocks += (stat->size + 511) >> 9; blocks 5797 fs/ext4/inode.c stat->blocks += delalloc_blocks << (inode->i_sb->s_blocksize_bits - 9); blocks 121 fs/ext4/ioctl.c blkcnt_t blocks; blocks 219 fs/ext4/ioctl.c blocks = inode_bl->i_blocks; blocks 241 fs/ext4/ioctl.c inode_bl->i_blocks = blocks; blocks 2173 fs/ext4/namei.c ext4_lblk_t block, blocks; blocks 2220 fs/ext4/namei.c blocks = dir->i_size >> sb->s_blocksize_bits; blocks 2221 fs/ext4/namei.c for (block = 0; block < blocks; block++) { blocks 2238 fs/ext4/namei.c if (blocks == 1 && !dx_fallback && blocks 239 fs/ext4/readpage.c sector_t blocks[MAX_BUF_PER_PAGE]; blocks 293 fs/ext4/readpage.c blocks[page_block] = map.m_pblk + map_offset + blocks 330 fs/ext4/readpage.c if (page_block && blocks[page_block-1] != map.m_pblk-1) blocks 339 fs/ext4/readpage.c blocks[page_block] = map.m_pblk+relative_block; blocks 368 fs/ext4/readpage.c if (bio && (last_block_in_bio != blocks[0] - 1)) { blocks 387 fs/ext4/readpage.c bio->bi_iter.bi_sector = blocks[0] << (blkbits - 9); blocks 404 fs/ext4/readpage.c last_block_in_bio = blocks[blocks_per_page - 1]; blocks 892 fs/ext4/xattr.c int blocks; blocks 927 fs/ext4/xattr.c blocks = (value_len + sb->s_blocksize - 1) >> sb->s_blocksize_bits; blocks 930 fs/ext4/xattr.c blocks += 1; blocks 933 fs/ext4/xattr.c credits += blocks * 2; blocks 936 fs/ext4/xattr.c credits += blocks; blocks 945 fs/ext4/xattr.c blocks = XATTR_SIZE_MAX >> sb->s_blocksize_bits; blocks 950 fs/ext4/xattr.c blocks += 1; blocks 953 fs/ext4/xattr.c credits += blocks * 2; blocks 741 fs/f2fs/file.c stat->blocks += (stat->size + 511) >> 9; blocks 913 fs/fuse/dir.c stat->blocks = attr->blocks; blocks 162 fs/fuse/inode.c inode->i_blocks = attr->blocks; blocks 402 fs/fuse/inode.c stbuf->f_blocks = attr->blocks; blocks 1035 fs/gfs2/bmap.c unsigned int blocks; blocks 1037 fs/gfs2/bmap.c blocks = ((pos & blockmask) + len + blockmask) >> inode->i_blkbits; blocks 1038 fs/gfs2/bmap.c return gfs2_trans_begin(sdp, RES_DINODE + blocks, 0); blocks 2243 fs/gfs2/bmap.c static int gfs2_add_jextent(struct gfs2_jdesc *jd, u64 lblock, u64 dblock, u64 blocks) blocks 2249 fs/gfs2/bmap.c if ((jext->dblock + jext->blocks) == dblock) { blocks 2250 fs/gfs2/bmap.c jext->blocks += blocks; blocks 2260 fs/gfs2/bmap.c jext->blocks = blocks; blocks 523 fs/gfs2/incore.h u64 blocks; blocks 591 fs/gfs2/inode.c unsigned blocks = 1; blocks 695 fs/gfs2/inode.c blocks++; blocks 697 fs/gfs2/inode.c error = alloc_dinode(ip, aflags, &blocks); blocks 701 fs/gfs2/inode.c gfs2_set_inode_blocks(inode, blocks); blocks 713 fs/gfs2/inode.c error = gfs2_trans_begin(sdp, blocks, 0); blocks 717 fs/gfs2/inode.c if (blocks > 1) { blocks 45 fs/gfs2/inode.h static inline void gfs2_set_inode_blocks(struct inode *inode, u64 blocks) blocks 47 fs/gfs2/inode.h inode->i_blocks = blocks << blocks 148 fs/gfs2/lops.c if ((lbn >= je->lblock) && (lbn < (je->lblock + je->blocks))) { blocks 523 fs/gfs2/lops.c for (; block < je->lblock + je->blocks; block++, dblock++) { blocks 544 fs/gfs2/lops.c unsigned int blocks = blocks 547 fs/gfs2/lops.c bio = gfs2_chain_bio(bio, blocks); blocks 876 fs/gfs2/quota.c unsigned int nalloc = 0, blocks; blocks 919 fs/gfs2/quota.c blocks = num_qd * data_blocks + RES_DINODE + num_qd + 3; blocks 928 fs/gfs2/quota.c blocks += gfs2_rg_blocks(ip, reserved) + nalloc * ind_blocks + RES_STATFS; blocks 930 fs/gfs2/quota.c error = gfs2_trans_begin(sdp, blocks, 0); blocks 1339 fs/gfs2/quota.c unsigned int blocks = size >> sdp->sd_sb.sb_bsize_shift; blocks 1351 fs/gfs2/quota.c sdp->sd_quota_slots = blocks * sdp->sd_qc_per_block; blocks 1362 fs/gfs2/quota.c for (x = 0; x < blocks; x++) { blocks 1602 fs/gfs2/quota.c state->s_state[USRQUOTA].blocks = sdp->sd_quota_inode->i_blocks; blocks 1657 fs/gfs2/quota.c unsigned int blocks = 0; blocks 1717 fs/gfs2/quota.c blocks = 1 + data_blocks + ind_blocks; blocks 1718 fs/gfs2/quota.c ap.target = blocks; blocks 1722 fs/gfs2/quota.c blocks += gfs2_rg_blocks(ip, blocks); blocks 1727 fs/gfs2/quota.c error = gfs2_trans_begin(sdp, blocks + RES_DINODE + 2, 0); blocks 384 fs/gfs2/trace_gfs2.h TP_PROTO(const struct gfs2_sbd *sdp, int blocks), blocks 386 fs/gfs2/trace_gfs2.h TP_ARGS(sdp, blocks), blocks 390 fs/gfs2/trace_gfs2.h __field( int, blocks ) blocks 395 fs/gfs2/trace_gfs2.h __entry->blocks = blocks; blocks 399 fs/gfs2/trace_gfs2.h MINOR(__entry->dev), __entry->blocks) blocks 28 fs/gfs2/trans.c int gfs2_trans_begin(struct gfs2_sbd *sdp, unsigned int blocks, blocks 35 fs/gfs2/trans.c BUG_ON(blocks == 0 && revokes == 0); blocks 45 fs/gfs2/trans.c tr->tr_blocks = blocks; blocks 49 fs/gfs2/trans.c if (blocks) blocks 50 fs/gfs2/trans.c tr->tr_reserved += 6 + blocks; blocks 37 fs/gfs2/trans.h extern int gfs2_trans_begin(struct gfs2_sbd *sdp, unsigned int blocks, blocks 290 fs/hfs/extent.c u32 total_blocks, blocks, start; blocks 306 fs/hfs/extent.c blocks = 0; blocks 308 fs/hfs/extent.c blocks += be16_to_cpu(extent[i].count); blocks 310 fs/hfs/extent.c res = hfs_free_extents(sb, extent, blocks, blocks); blocks 313 fs/hfs/extent.c if (total_blocks == blocks) blocks 327 fs/hfs/extent.c } while (total_blocks > blocks); blocks 401 fs/hfsplus/extents.c u32 total_blocks, blocks, start; blocks 408 fs/hfsplus/extents.c blocks = 0; blocks 410 fs/hfsplus/extents.c blocks += be32_to_cpu(fork->extents[i].block_count); blocks 412 fs/hfsplus/extents.c res = hfsplus_free_extents(sb, fork->extents, blocks, blocks); blocks 415 fs/hfsplus/extents.c if (total_blocks == blocks) blocks 434 fs/hfsplus/extents.c } while (total_blocks > blocks); blocks 61 fs/hostfs/hostfs.h unsigned long long blocks; blocks 556 fs/hostfs/hostfs_kern.c ino->i_blocks = st.blocks; blocks 36 fs/hostfs/hostfs_user.c p->blocks = buf->st_blocks; blocks 189 fs/jbd2/transaction.c static void sub_reserved_credits(journal_t *journal, int blocks) blocks 191 fs/jbd2/transaction.c atomic_sub(blocks, &journal->j_reserved_credits); blocks 201 fs/jbd2/transaction.c static int add_transaction_credits(journal_t *journal, int blocks, blocks 206 fs/jbd2/transaction.c int total = blocks + rsv_blocks; blocks 302 fs/jbd2/transaction.c int blocks = handle->h_buffer_credits; blocks 315 fs/jbd2/transaction.c (rsv_blocks + blocks > journal->j_max_transaction_buffers)) { blocks 318 fs/jbd2/transaction.c current->comm, blocks, rsv_blocks, blocks 384 fs/jbd2/transaction.c if (add_transaction_credits(journal, blocks, rsv_blocks)) blocks 398 fs/jbd2/transaction.c sub_reserved_credits(journal, blocks); blocks 407 fs/jbd2/transaction.c handle->h_requested_credits = blocks; blocks 412 fs/jbd2/transaction.c handle, blocks, blocks 381 fs/jffs2/build.c c->blocks = vzalloc(size); blocks 384 fs/jffs2/build.c c->blocks = kzalloc(size, GFP_KERNEL); blocks 385 fs/jffs2/build.c if (!c->blocks) blocks 389 fs/jffs2/build.c INIT_LIST_HEAD(&c->blocks[i].list); blocks 390 fs/jffs2/build.c c->blocks[i].offset = i * c->sector_size; blocks 391 fs/jffs2/build.c c->blocks[i].free_size = c->sector_size; blocks 426 fs/jffs2/build.c kvfree(c->blocks); blocks 518 fs/jffs2/fs.c size_t blocks; blocks 539 fs/jffs2/fs.c blocks = c->flash_size / c->sector_size; blocks 544 fs/jffs2/fs.c if ((c->sector_size * blocks) != c->flash_size) { blocks 545 fs/jffs2/fs.c c->flash_size = c->sector_size * blocks; blocks 604 fs/jffs2/fs.c kvfree(c->blocks); blocks 714 fs/jffs2/gc.c struct jffs2_eraseblock *jeb = &c->blocks[phys_ofs / c->sector_size]; blocks 1230 fs/jffs2/gc.c jeb = &c->blocks[raw->flash_offset / c->sector_size]; blocks 1286 fs/jffs2/gc.c jeb = &c->blocks[raw->flash_offset / c->sector_size]; blocks 91 fs/jffs2/jffs2_fs_sb.h struct jffs2_eraseblock *blocks; /* The whole array of blocks. Used for getting blocks blocks 510 fs/jffs2/nodelist.c this = c->blocks[i].first_node; blocks 520 fs/jffs2/nodelist.c c->blocks[i].first_node = c->blocks[i].last_node = NULL; blocks 703 fs/jffs2/nodelist.c jeb = &c->blocks[ref->flash_offset / c->sector_size]; blocks 728 fs/jffs2/nodelist.c jeb = &c->blocks[ref->flash_offset / c->sector_size]; blocks 507 fs/jffs2/nodemgmt.c jeb = &c->blocks[ofs / c->sector_size]; blocks 603 fs/jffs2/nodemgmt.c jeb = &c->blocks[blocknr]; blocks 114 fs/jffs2/readinode.c jeb = &c->blocks[ref->flash_offset / c->sector_size]; blocks 617 fs/jffs2/readinode.c jeb = &c->blocks[ref->flash_offset / c->sector_size]; blocks 803 fs/jffs2/readinode.c jeb = &c->blocks[ref->flash_offset / c->sector_size]; blocks 144 fs/jffs2/scan.c struct jffs2_eraseblock *jeb = &c->blocks[i]; blocks 260 fs/jffs2/summary.c jeb = &c->blocks[ofs / c->sector_size]; blocks 321 fs/jffs2/super.c kvfree(c->blocks); blocks 285 fs/jffs2/wbuf.c jeb = &c->blocks[c->wbuf_ofs / c->sector_size]; blocks 460 fs/jffs2/wbuf.c new_jeb = &c->blocks[ofs / c->sector_size]; blocks 599 fs/jffs2/wbuf.c wbuf_jeb = &c->blocks[c->wbuf_ofs / c->sector_size]; blocks 926 fs/jffs2/wbuf.c jeb = &c->blocks[outvec_to / c->sector_size]; blocks 131 fs/jffs2/write.c struct jffs2_eraseblock *jeb = &c->blocks[flash_ofs / c->sector_size]; blocks 283 fs/jffs2/write.c struct jffs2_eraseblock *jeb = &c->blocks[flash_ofs / c->sector_size]; blocks 185 fs/jffs2/xattr.c jeb = &c->blocks[ref_offset(raw) / c->sector_size]; blocks 491 fs/jffs2/xattr.c jeb = &c->blocks[ref_offset(raw) / c->sector_size]; blocks 1307 fs/jffs2/xattr.c jeb = &c->blocks[ref_offset(raw) / c->sector_size]; blocks 32 fs/libfs.c stat->blocks = inode->i_mapping->nrpages << (PAGE_SHIFT - 9); blocks 30 fs/minix/bitmap.c unsigned blocks = DIV_ROUND_UP(numbits, blocksize * 8); blocks 32 fs/minix/bitmap.c while (blocks--) { blocks 630 fs/minix/inode.c stat->blocks = (BLOCK_SIZE / 512) * V1_minix_blocks(stat->size, sb); blocks 632 fs/minix/inode.c stat->blocks = (sb->s_blocksize / 512) * V2_minix_blocks(stat->size, sb); blocks 354 fs/minix/itree_common.c unsigned blocks, res, direct = DIRECT, i = DEPTH; blocks 355 fs/minix/itree_common.c blocks = (size + sb->s_blocksize - 1) >> (BLOCK_SIZE_BITS + k); blocks 356 fs/minix/itree_common.c res = blocks; blocks 357 fs/minix/itree_common.c while (--i && blocks > direct) { blocks 358 fs/minix/itree_common.c blocks -= direct; blocks 359 fs/minix/itree_common.c blocks += sb->s_blocksize/sizeof(block_t) - 1; blocks 360 fs/minix/itree_common.c blocks /= sb->s_blocksize/sizeof(block_t); blocks 361 fs/minix/itree_common.c res += blocks; blocks 167 fs/mpage.c sector_t blocks[MAX_BUF_PER_PAGE]; blocks 213 fs/mpage.c blocks[page_block] = map_bh->b_blocknr + map_offset + blocks 260 fs/mpage.c if (page_block && blocks[page_block-1] != map_bh->b_blocknr-1) blocks 269 fs/mpage.c blocks[page_block] = map_bh->b_blocknr+relative_block; blocks 296 fs/mpage.c if (args->bio && (args->last_block_in_bio != blocks[0] - 1)) blocks 302 fs/mpage.c if (!bdev_read_page(bdev, blocks[0] << (blkbits - 9), blocks 306 fs/mpage.c args->bio = mpage_alloc(bdev, blocks[0] << (blkbits - 9), blocks 326 fs/mpage.c args->last_block_in_bio = blocks[blocks_per_page - 1]; blocks 508 fs/mpage.c sector_t blocks[MAX_BUF_PER_PAGE]; blocks 547 fs/mpage.c if (bh->b_blocknr != blocks[page_block-1] + 1) blocks 550 fs/mpage.c blocks[page_block++] = bh->b_blocknr; blocks 591 fs/mpage.c if (map_bh.b_blocknr != blocks[page_block-1] + 1) blocks 594 fs/mpage.c blocks[page_block++] = map_bh.b_blocknr; blocks 626 fs/mpage.c if (bio && mpd->last_block_in_bio != blocks[0] - 1) blocks 632 fs/mpage.c if (!bdev_write_page(bdev, blocks[0] << (blkbits - 9), blocks 636 fs/mpage.c bio = mpage_alloc(bdev, blocks[0] << (blkbits - 9), blocks 669 fs/mpage.c mpd->last_block_in_bio = blocks[blocks_per_page - 1]; blocks 539 fs/nfs/inode.c inode->i_blocks = fattr->du.nfs2.blocks; blocks 2021 fs/nfs/inode.c inode->i_blocks = fattr->du.nfs2.blocks; blocks 298 fs/nfs/nfs2xdr.c fattr->du.nfs2.blocks = be32_to_cpup(p++); blocks 1038 fs/nfs/nfs2xdr.c result->blocks = be32_to_cpup(p++); blocks 537 fs/nfs/proc.c stat->tbytes = (u64)fsinfo.blocks * fsinfo.bsize; blocks 181 fs/nfsd/nfs3xdr.c p = xdr_encode_hyper(p, ((u64)stat->blocks) << 9); blocks 2818 fs/nfsd/nfs4xdr.c dummy64 = (u64)stat.blocks << 9; blocks 156 fs/nfsd/nfsxdr.c *p++ = htonl((u32) stat->blocks); blocks 593 fs/nilfs2/super.c unsigned long long blocks; blocks 606 fs/nilfs2/super.c blocks = nilfs->ns_blocks_per_segment * nilfs->ns_nsegments blocks 643 fs/nilfs2/super.c buf->f_blocks = blocks - overhead; blocks 1945 fs/ocfs2/cluster/heartbeat.c CONFIGFS_ATTR(o2hb_region_, blocks); blocks 1324 fs/ocfs2/file.c stat->blocks += (stat->size + 511)>>9; blocks 545 fs/ocfs2/journal.h int blocks = ocfs2_mknod_credits(sb, 0, 0); blocks 549 fs/ocfs2/journal.h blocks += ocfs2_clusters_to_blocks(sb, 1); blocks 551 fs/ocfs2/journal.h return blocks + ocfs2_quota_trans_credits(sb); blocks 557 fs/ocfs2/journal.h int blocks; blocks 561 fs/ocfs2/journal.h blocks = 1 + 1 + 1 + bitmap_blocks; blocks 562 fs/ocfs2/journal.h return blocks; blocks 1692 fs/ocfs2/namei.c int virtual, blocks, status, i, bytes_left; blocks 1697 fs/ocfs2/namei.c blocks = (bytes_left + sb->s_blocksize - 1) >> sb->s_blocksize_bits; blocks 1700 fs/ocfs2/namei.c i_size_read(inode), blocks); blocks 1710 fs/ocfs2/namei.c bhs = kcalloc(blocks, sizeof(struct buffer_head *), GFP_KERNEL); blocks 1771 fs/ocfs2/namei.c for(i = 0; i < blocks; i++) blocks 727 fs/ocfs2/ocfs2.h u64 blocks) blocks 732 fs/ocfs2/ocfs2.h blocks += (1 << b_to_c_bits) - 1; blocks 733 fs/ocfs2/ocfs2.h return (u32)(blocks >> b_to_c_bits); blocks 737 fs/ocfs2/ocfs2.h u64 blocks) blocks 742 fs/ocfs2/ocfs2.h return (u32)(blocks >> b_to_c_bits); blocks 782 fs/ocfs2/ocfs2.h u64 blocks) blocks 787 fs/ocfs2/ocfs2.h clusters = ocfs2_blocks_to_clusters(sb, blocks); blocks 804 fs/ocfs2/ocfs2.h u64 blocks; blocks 806 fs/ocfs2/ocfs2.h blocks = ocfs2_blocks_for_bytes(sb, bytes); blocks 807 fs/ocfs2/ocfs2.h return blocks << sb->s_blocksize_bits; blocks 3006 fs/ocfs2/refcounttree.c int i, blocks = ocfs2_clusters_to_blocks(sb, new_len); blocks 3016 fs/ocfs2/refcounttree.c for (i = 0; i < blocks; i++, old_block++, new_block++) { blocks 342 fs/ocfs2/slot_map.c unsigned long long blocks, bytes = 0; blocks 350 fs/ocfs2/slot_map.c blocks = ocfs2_blocks_for_bytes(si->si_inode->i_sb, bytes); blocks 351 fs/ocfs2/slot_map.c BUG_ON(blocks > UINT_MAX); blocks 352 fs/ocfs2/slot_map.c si->si_blocks = blocks; blocks 364 fs/ocfs2/slot_map.c BUG_ON((osb->max_slots / si->si_slots_per_block) > blocks); blocks 958 fs/ocfs2/suballoc.c int blocks, blocks 971 fs/ocfs2/suballoc.c (*ac)->ac_bits_wanted = blocks; blocks 67 fs/ocfs2/suballoc.h int blocks, blocks 222 fs/overlayfs/inode.c stat->blocks = lowerstat.blocks; blocks 240 fs/overlayfs/inode.c stat->blocks = lowerdatastat.blocks; blocks 2806 fs/quota/dquot.c tstate->blocks = dqopt->files[type]->i_blocks; blocks 175 fs/quota/quota.c static inline qsize_t qbtos(qsize_t blocks) blocks 177 fs/quota/quota.c return blocks << QIF_DQBLKSIZE_BITS; blocks 362 fs/quota/quota.c fqs->qs_uquota.qfs_nblks = state.s_state[USRQUOTA].blocks; blocks 367 fs/quota/quota.c fqs->qs_gquota.qfs_nblks = state.s_state[GRPQUOTA].blocks; blocks 379 fs/quota/quota.c state.s_state[PRJQUOTA].blocks; blocks 428 fs/quota/quota.c fqs->qs_uquota.qfs_nblks = state.s_state[USRQUOTA].blocks; blocks 433 fs/quota/quota.c fqs->qs_gquota.qfs_nblks = state.s_state[GRPQUOTA].blocks; blocks 438 fs/quota/quota.c fqs->qs_pquota.qfs_nblks = state.s_state[PRJQUOTA].blocks; blocks 476 fs/quota/quota.c static inline u64 quota_bbtob(u64 blocks) blocks 478 fs/quota/quota.c return blocks << XFS_BB_SHIFT; blocks 27 fs/quota/quota_v1.c static inline qsize_t v1_qbtos(qsize_t blocks) blocks 29 fs/quota/quota_v1.c return blocks << QUOTABLOCK_BITS; blocks 130 fs/quota/quota_v1.c ulong blocks; blocks 140 fs/quota/quota_v1.c blocks = isize >> BLOCK_SIZE_BITS; blocks 142 fs/quota/quota_v1.c if ((blocks % sizeof(struct v1_disk_dqblk) * BLOCK_SIZE + off) % blocks 54 fs/quota/quota_v2.c static inline qsize_t v2_qbtos(qsize_t blocks) blocks 56 fs/quota/quota_v2.c return blocks << QUOTABLOCK_BITS; blocks 1199 fs/reiserfs/inode.c static inline loff_t to_real_used_space(struct inode *inode, ulong blocks, blocks 1207 fs/reiserfs/inode.c (((loff_t) blocks) << 9); blocks 1261 fs/reiserfs/inode.c unsigned long blocks; blocks 1279 fs/reiserfs/inode.c blocks = (inode->i_size + 511) >> 9; blocks 1280 fs/reiserfs/inode.c blocks = _ROUND_UP(blocks, inode->i_sb->s_blocksize >> 9); blocks 1290 fs/reiserfs/inode.c if (inode->i_blocks > blocks) { blocks 1291 fs/reiserfs/inode.c inode->i_blocks = blocks; blocks 2325 fs/reiserfs/journal.c unsigned int blocks = BUFNR; blocks 2334 fs/reiserfs/journal.c blocks = max_block - block; blocks 2338 fs/reiserfs/journal.c for (i = 1; i < blocks; i++) { blocks 1124 fs/reiserfs/super.c unsigned long *blocks, blocks 1180 fs/reiserfs/super.c *blocks = 0; blocks 1202 fs/reiserfs/super.c *blocks = blocks 1206 fs/reiserfs/super.c *blocks = simple_strtoul(arg, &p, 0); blocks 1435 fs/reiserfs/super.c unsigned long blocks; blocks 1462 fs/reiserfs/super.c (s, arg, &mount_options, &blocks, NULL, &commit_max_age, blocks 1513 fs/reiserfs/super.c if (blocks) { blocks 1514 fs/reiserfs/super.c err = reiserfs_resize(s, blocks); blocks 1901 fs/reiserfs/super.c unsigned long blocks; blocks 1942 fs/reiserfs/super.c (s, (char *)data, &sbi->s_mount_opt, &blocks, &jdev_name, blocks 1958 fs/reiserfs/super.c if (blocks) { blocks 299 fs/reiserfs/xattr.c int blocks = JOURNAL_PER_BALANCE_CNT * 2 + 2 + blocks 304 fs/reiserfs/xattr.c err = journal_begin(&th, inode->i_sb, blocks); blocks 48 fs/reiserfs/xattr_security.c int blocks = 0; blocks 70 fs/reiserfs/xattr_security.c blocks = reiserfs_xattr_jcreate_nblocks(inode) + blocks 76 fs/reiserfs/xattr_security.c return blocks; blocks 173 fs/squashfs/file.c int blocks = min_t(int, n, PAGE_SIZE >> 2); blocks 176 fs/squashfs/file.c offset, blocks << 2); blocks 183 fs/squashfs/file.c for (i = 0; i < blocks; i++) { blocks 191 fs/squashfs/file.c n -= blocks; blocks 214 fs/squashfs/file.c static inline int calculate_skip(int blocks) blocks 216 fs/squashfs/file.c int skip = blocks / ((SQUASHFS_META_ENTRIES + 1) blocks 275 fs/squashfs/file.c int blocks = skip * SQUASHFS_META_INDEXES; blocks 276 fs/squashfs/file.c long long res = read_indexes(inode->i_sb, blocks, blocks 47 fs/stat.c stat->blocks = inode->i_blocks; blocks 332 fs/stat.c tmp.st_blocks = stat->blocks; blocks 473 fs/stat.c tmp.st_blocks = stat->blocks; blocks 542 fs/stat.c tmp.stx_blocks = stat->blocks; blocks 620 fs/stat.c tmp.st_blocks = stat->blocks; blocks 47 fs/sysv/balloc.c sysv_zone_t *blocks = sbi->s_bcache; blocks 86 fs/sysv/balloc.c memcpy(get_chunk(sb,bh), blocks, count * sizeof(sysv_zone_t)); blocks 165 fs/sysv/balloc.c sysv_zone_t *blocks; blocks 186 fs/sysv/balloc.c blocks = sbi->s_bcache; blocks 192 fs/sysv/balloc.c while (n && (zone = blocks[--n]) != 0) blocks 208 fs/sysv/balloc.c blocks = get_chunk(sb, bh); blocks 433 fs/sysv/itree.c unsigned blocks, res, direct = DIRECT, i = DEPTH; blocks 434 fs/sysv/itree.c blocks = (size + s->s_blocksize - 1) >> s->s_blocksize_bits; blocks 435 fs/sysv/itree.c res = blocks; blocks 436 fs/sysv/itree.c while (--i && blocks > direct) { blocks 437 fs/sysv/itree.c blocks = ((blocks - direct - 1) >> ptrs_bits) + 1; blocks 438 fs/sysv/itree.c res += blocks; blocks 441 fs/sysv/itree.c return blocks; blocks 449 fs/sysv/itree.c stat->blocks = (s->s_blocksize / 512) * sysv_nblocks(s, stat->size); blocks 1614 fs/ubifs/dir.c stat->blocks = size >> 9; blocks 1616 fs/ubifs/dir.c stat->blocks = 0; blocks 1184 fs/udf/super.c sector_t blocks = i_size_read(sb->s_bdev->bd_inode) >> blocks 1189 fs/udf/super.c sbi->s_last_block != blocks - 1) { blocks 1192 fs/udf/super.c (unsigned long)blocks - 1); blocks 1193 fs/udf/super.c udf_find_vat_block(sb, p_index, type1_index, blocks - 1); blocks 119 fs/verity/enable.c u64 blocks; blocks 139 fs/verity/enable.c blocks = (inode->i_size + params->block_size - 1) >> blocks 142 fs/verity/enable.c err = build_merkle_tree_level(inode, level, blocks, params, blocks 146 fs/verity/enable.c blocks = (blocks + params->hashes_per_block - 1) >> blocks 36 fs/verity/open.c u64 blocks; blocks 92 fs/verity/open.c blocks = (inode->i_size + params->block_size - 1) >> log_blocksize; blocks 93 fs/verity/open.c pr_debug("Data is %lld bytes (%llu blocks)\n", inode->i_size, blocks); blocks 94 fs/verity/open.c while (blocks > 1) { blocks 100 fs/verity/open.c blocks = (blocks + params->hashes_per_block - 1) >> blocks 103 fs/verity/open.c params->level_start[params->num_levels++] = blocks; blocks 109 fs/verity/open.c blocks = params->level_start[level]; blocks 112 fs/verity/open.c level, blocks, offset); blocks 113 fs/verity/open.c offset += blocks; blocks 124 fs/xfs/libxfs/xfs_alloc.c unsigned int blocks; blocks 126 fs/xfs/libxfs/xfs_alloc.c blocks = XFS_BB_TO_FSB(mp, XFS_FSS_TO_BB(mp, 4)); /* ag headers */ blocks 127 fs/xfs/libxfs/xfs_alloc.c blocks += XFS_ALLOC_AGFL_RESERVE; blocks 128 fs/xfs/libxfs/xfs_alloc.c blocks += 3; /* AGF, AGI btree root blocks */ blocks 130 fs/xfs/libxfs/xfs_alloc.c blocks++; /* finobt root block */ blocks 132 fs/xfs/libxfs/xfs_alloc.c blocks++; /* rmap root block */ blocks 134 fs/xfs/libxfs/xfs_alloc.c blocks++; /* refcount root block */ blocks 136 fs/xfs/libxfs/xfs_alloc.c return mp->m_sb.sb_agblocks - blocks; blocks 4854 fs/xfs/libxfs/xfs_btree.c xfs_extlen_t *blocks = data; blocks 4855 fs/xfs/libxfs/xfs_btree.c (*blocks)++; blocks 4864 fs/xfs/libxfs/xfs_btree.c xfs_extlen_t *blocks) blocks 4866 fs/xfs/libxfs/xfs_btree.c *blocks = 0; blocks 4868 fs/xfs/libxfs/xfs_btree.c blocks); blocks 488 fs/xfs/libxfs/xfs_btree.h int xfs_btree_count_blocks(struct xfs_btree_cur *cur, xfs_extlen_t *blocks); blocks 71 fs/xfs/libxfs/xfs_trans_resv.c uint blocks; blocks 73 fs/xfs/libxfs/xfs_trans_resv.c blocks = num_ops * 2 * (2 * mp->m_ag_maxlevels - 1); blocks 75 fs/xfs/libxfs/xfs_trans_resv.c blocks += num_ops * (2 * mp->m_rmap_maxlevels - 1); blocks 77 fs/xfs/libxfs/xfs_trans_resv.c blocks += num_ops * (2 * mp->m_refc_maxlevels - 1); blocks 79 fs/xfs/libxfs/xfs_trans_resv.c return blocks; blocks 350 fs/xfs/scrub/agheader.c xfs_extlen_t *blocks = priv; blocks 352 fs/xfs/scrub/agheader.c (*blocks) += rec->ar_blockcount; blocks 362 fs/xfs/scrub/agheader.c xfs_extlen_t blocks = 0; blocks 369 fs/xfs/scrub/agheader.c xchk_agf_record_bno_lengths, &blocks); blocks 372 fs/xfs/scrub/agheader.c if (blocks != be32_to_cpu(agf->agf_freeblks)) blocks 383 fs/xfs/scrub/agheader.c xfs_extlen_t blocks; blocks 401 fs/xfs/scrub/agheader.c error = xfs_alloc_get_rec(sc->sa.cnt_cur, &agbno, &blocks, &have); blocks 404 fs/xfs/scrub/agheader.c if (!have || blocks != be32_to_cpu(agf->agf_longest)) blocks 415 fs/xfs/scrub/agheader.c xfs_agblock_t blocks; blocks 421 fs/xfs/scrub/agheader.c error = xfs_btree_count_blocks(sc->sa.rmap_cur, &blocks); blocks 424 fs/xfs/scrub/agheader.c btreeblks = blocks - 1; blocks 425 fs/xfs/scrub/agheader.c if (blocks != be32_to_cpu(agf->agf_rmap_blocks)) blocks 440 fs/xfs/scrub/agheader.c error = xfs_btree_count_blocks(sc->sa.bno_cur, &blocks); blocks 443 fs/xfs/scrub/agheader.c btreeblks += blocks - 1; blocks 445 fs/xfs/scrub/agheader.c error = xfs_btree_count_blocks(sc->sa.cnt_cur, &blocks); blocks 448 fs/xfs/scrub/agheader.c btreeblks += blocks - 1; blocks 460 fs/xfs/scrub/agheader.c xfs_agblock_t blocks; blocks 466 fs/xfs/scrub/agheader.c error = xfs_btree_count_blocks(sc->sa.refc_cur, &blocks); blocks 469 fs/xfs/scrub/agheader.c if (blocks != be32_to_cpu(agf->agf_refcount_blocks)) blocks 244 fs/xfs/scrub/agheader_repair.c xfs_agblock_t blocks; blocks 253 fs/xfs/scrub/agheader_repair.c error = xfs_btree_count_blocks(cur, &blocks); blocks 257 fs/xfs/scrub/agheader_repair.c btreeblks = blocks - 1; blocks 264 fs/xfs/scrub/agheader_repair.c error = xfs_btree_count_blocks(cur, &blocks); blocks 268 fs/xfs/scrub/agheader_repair.c btreeblks += blocks - 1; blocks 272 fs/xfs/scrub/agheader_repair.c error = xfs_btree_count_blocks(cur, &blocks); blocks 276 fs/xfs/scrub/agheader_repair.c agf->agf_rmap_blocks = cpu_to_be32(blocks); blocks 277 fs/xfs/scrub/agheader_repair.c btreeblks += blocks - 1; blocks 285 fs/xfs/scrub/agheader_repair.c error = xfs_btree_count_blocks(cur, &blocks); blocks 289 fs/xfs/scrub/agheader_repair.c agf->agf_refcount_blocks = cpu_to_be32(blocks); blocks 319 fs/xfs/scrub/common.c xfs_filblks_t *blocks; blocks 339 fs/xfs/scrub/common.c (*sroi->blocks) += rec->rm_blockcount; blocks 353 fs/xfs/scrub/common.c xfs_filblks_t *blocks) blocks 357 fs/xfs/scrub/common.c .blocks = blocks, blocks 360 fs/xfs/scrub/common.c *blocks = 0; blocks 128 fs/xfs/scrub/common.h const struct xfs_owner_info *oinfo, xfs_filblks_t *blocks); blocks 516 fs/xfs/scrub/ialloc.c xfs_filblks_t blocks; blocks 538 fs/xfs/scrub/ialloc.c &XFS_RMAP_OINFO_INOBT, &blocks); blocks 541 fs/xfs/scrub/ialloc.c if (blocks != inobt_blocks + finobt_blocks) blocks 555 fs/xfs/scrub/ialloc.c xfs_filblks_t blocks; blocks 564 fs/xfs/scrub/ialloc.c &XFS_RMAP_OINFO_INODES, &blocks); blocks 568 fs/xfs/scrub/ialloc.c if (blocks != inode_blocks) blocks 378 fs/xfs/scrub/refcount.c xfs_filblks_t blocks; blocks 389 fs/xfs/scrub/refcount.c &XFS_RMAP_OINFO_REFC, &blocks); blocks 392 fs/xfs/scrub/refcount.c if (blocks != refcbt_blocks) blocks 397 fs/xfs/scrub/refcount.c &XFS_RMAP_OINFO_COW, &blocks); blocks 400 fs/xfs/scrub/refcount.c if (blocks != cow_blocks) blocks 513 fs/xfs/xfs_iops.c stat->blocks = blocks 3767 fs/xfs/xfs_log.c int blocks; blocks 3770 fs/xfs/xfs_log.c blocks = blocks 3772 fs/xfs/xfs_log.c if (blocks < BTOBB(iclog->ic_offset)+BTOBB(log->l_iclog_hsize)) blocks 3780 fs/xfs/xfs_log.c blocks = BLOCK_LSN(tail_lsn) - log->l_prev_block; blocks 3781 fs/xfs/xfs_log.c if (blocks < BTOBB(iclog->ic_offset) + 1) blocks 1578 fs/xfs/xfs_log_recover.c int blocks, blocks 1586 fs/xfs/xfs_log_recover.c int end_block = start_block + blocks; blocks 1597 fs/xfs/xfs_log_recover.c bufblks = 1 << ffs(blocks); blocks 38 fs/xfs/xfs_quotaops.c tstate->blocks = ip->i_d.di_nblocks; blocks 710 fs/xfs/xfs_trace.h TP_PROTO(struct xfs_inode *ip, xfs_fsblock_t blocks, int shift, blocks 712 fs/xfs/xfs_trace.h TP_ARGS(ip, blocks, shift, writeio_blocks), blocks 716 fs/xfs/xfs_trace.h __field(xfs_fsblock_t, blocks) blocks 723 fs/xfs/xfs_trace.h __entry->blocks = blocks; blocks 730 fs/xfs/xfs_trace.h __entry->blocks, __entry->shift, __entry->writeio_blocks) blocks 150 fs/xfs/xfs_trans.c uint blocks, blocks 164 fs/xfs/xfs_trans.c if (blocks > 0) { blocks 165 fs/xfs/xfs_trans.c error = xfs_mod_fdblocks(tp->t_mountp, -((int64_t)blocks), rsvd); blocks 170 fs/xfs/xfs_trans.c tp->t_blk_res += blocks; blocks 239 fs/xfs/xfs_trans.c if (blocks > 0) { blocks 240 fs/xfs/xfs_trans.c xfs_mod_fdblocks(tp->t_mountp, (int64_t)blocks, rsvd); blocks 253 fs/xfs/xfs_trans.c uint blocks, blocks 286 fs/xfs/xfs_trans.c error = xfs_trans_reserve(tp, resp, blocks, rtextents); blocks 166 fs/xfs/xfs_trans.h uint blocks, uint rtextents, uint flags, blocks 18 include/crypto/sha1_base.h typedef void (sha1_block_fn)(struct sha1_state *sst, u8 const *src, int blocks); blocks 45 include/crypto/sha1_base.h int blocks; blocks 57 include/crypto/sha1_base.h blocks = len / SHA1_BLOCK_SIZE; blocks 60 include/crypto/sha1_base.h if (blocks) { blocks 61 include/crypto/sha1_base.h block_fn(sctx, data, blocks); blocks 62 include/crypto/sha1_base.h data += blocks * SHA1_BLOCK_SIZE; blocks 19 include/crypto/sha256_base.h int blocks); blocks 46 include/crypto/sha256_base.h int blocks; blocks 58 include/crypto/sha256_base.h blocks = len / SHA256_BLOCK_SIZE; blocks 61 include/crypto/sha256_base.h if (blocks) { blocks 62 include/crypto/sha256_base.h block_fn(sctx, data, blocks); blocks 63 include/crypto/sha256_base.h data += blocks * SHA256_BLOCK_SIZE; blocks 19 include/crypto/sha512_base.h int blocks); blocks 68 include/crypto/sha512_base.h int blocks; blocks 80 include/crypto/sha512_base.h blocks = len / SHA512_BLOCK_SIZE; blocks 83 include/crypto/sha512_base.h if (blocks) { blocks 84 include/crypto/sha512_base.h block_fn(sctx, data, blocks); blocks 85 include/crypto/sha512_base.h data += blocks * SHA512_BLOCK_SIZE; blocks 18 include/crypto/sm3_base.h typedef void (sm3_block_fn)(struct sm3_state *sst, u8 const *src, int blocks); blocks 48 include/crypto/sm3_base.h int blocks; blocks 60 include/crypto/sm3_base.h blocks = len / SM3_BLOCK_SIZE; blocks 63 include/crypto/sm3_base.h if (blocks) { blocks 64 include/crypto/sm3_base.h block_fn(sctx, data, blocks); blocks 65 include/crypto/sm3_base.h data += blocks * SM3_BLOCK_SIZE; blocks 178 include/linux/async_tx.h async_gen_syndrome(struct page **blocks, unsigned int offset, int src_cnt, blocks 182 include/linux/async_tx.h async_syndrome_val(struct page **blocks, unsigned int offset, int src_cnt, blocks 83 include/linux/iio/buffer-dma.h struct iio_dma_buffer_block *blocks[2]; blocks 1360 include/linux/jbd2.h extern handle_t *jbd2__journal_start(journal_t *, int blocks, int rsv_blocks, blocks 157 include/linux/mfd/stmpe.h extern int stmpe_enable(struct stmpe *stmpe, unsigned int blocks); blocks 158 include/linux/mfd/stmpe.h extern int stmpe_disable(struct stmpe *stmpe, unsigned int blocks); blocks 121 include/linux/mmc/core.h unsigned int blocks; /* number of blocks */ blocks 56 include/linux/nfs_xdr.h __u32 blocks; blocks 167 include/linux/nfs_xdr.h __u32 blocks; /* No. of "bsize" blocks on filesystem */ blocks 141 include/linux/pe.h uint16_t blocks; /* pages in file, 0x3 */ blocks 60 include/linux/qed/qed_fcoe_if.h u8 *blocks[MAX_TID_BLOCKS_FCOE]; blocks 148 include/linux/qed/qed_iscsi_if.h u8 *blocks[MAX_TID_BLOCKS_ISCSI]; blocks 405 include/linux/quota.h blkcnt_t blocks; /* Number of 512-byte blocks in the file */ blocks 48 include/linux/stat.h u64 blocks; blocks 498 include/media/cec.h unsigned int blocks = size / 128; blocks 503 include/media/cec.h if (blocks < 2 || size % 128) blocks 513 include/media/cec.h if (edid[0x7e] + 1 < blocks) blocks 514 include/media/cec.h blocks = edid[0x7e] + 1; blocks 516 include/media/cec.h for (block = 1; block < blocks; block++) { blocks 516 include/net/9p/9p.h u64 blocks; blocks 2073 include/net/bluetooth/hci.h __le16 blocks; blocks 173 include/trace/events/btrfs.h __field( blkcnt_t, blocks ) blocks 183 include/trace/events/btrfs.h __entry->blocks = inode->i_blocks; blocks 197 include/trace/events/btrfs.h (unsigned long long)__entry->blocks, blocks 119 include/trace/events/ext4.h __field( __u64, blocks ) blocks 128 include/trace/events/ext4.h __entry->blocks = inode->i_blocks; blocks 135 include/trace/events/ext4.h __entry->uid, __entry->gid, __entry->blocks) blocks 1426 include/trace/events/ext4.h __field( unsigned int, blocks ) blocks 1434 include/trace/events/ext4.h __entry->blocks = max_blocks; blocks 1441 include/trace/events/ext4.h __entry->pos, __entry->blocks, blocks 1501 include/trace/events/ext4.h __field( __u64, blocks ) blocks 1507 include/trace/events/ext4.h __entry->blocks = inode->i_blocks; blocks 1512 include/trace/events/ext4.h (unsigned long) __entry->ino, __entry->blocks) blocks 1748 include/trace/events/ext4.h TP_PROTO(struct super_block *sb, int blocks, int rsv_blocks, blocks 1751 include/trace/events/ext4.h TP_ARGS(sb, blocks, rsv_blocks, IP), blocks 1756 include/trace/events/ext4.h __field( int, blocks ) blocks 1763 include/trace/events/ext4.h __entry->blocks = blocks; blocks 1769 include/trace/events/ext4.h __entry->blocks, __entry->rsv_blocks, (void *)__entry->ip) blocks 1773 include/trace/events/ext4.h TP_PROTO(struct super_block *sb, int blocks, unsigned long IP), blocks 1775 include/trace/events/ext4.h TP_ARGS(sb, blocks, IP), blocks 1780 include/trace/events/ext4.h __field( int, blocks ) blocks 1786 include/trace/events/ext4.h __entry->blocks = blocks; blocks 1791 include/trace/events/ext4.h __entry->blocks, (void *)__entry->ip) blocks 170 include/trace/events/f2fs.h __field(blkcnt_t, blocks) blocks 181 include/trace/events/f2fs.h __entry->blocks = inode->i_blocks; blocks 192 include/trace/events/f2fs.h (unsigned long long)__entry->blocks, blocks 318 include/trace/events/f2fs.h __field(blkcnt_t, blocks) blocks 326 include/trace/events/f2fs.h __entry->blocks = dir->i_blocks; blocks 334 include/trace/events/f2fs.h (unsigned long long)__entry->blocks, blocks 398 include/trace/events/f2fs.h __field(blkcnt_t, blocks) blocks 406 include/trace/events/f2fs.h __entry->blocks = inode->i_blocks; blocks 414 include/trace/events/f2fs.h (unsigned long long)__entry->blocks, blocks 886 include/trace/events/f2fs.h __field(blkcnt_t, blocks) blocks 897 include/trace/events/f2fs.h __entry->blocks = inode->i_blocks; blocks 908 include/trace/events/f2fs.h (unsigned long long)__entry->blocks, blocks 250 include/trace/events/jbd2.h __field( __u32, blocks ) blocks 264 include/trace/events/jbd2.h __entry->blocks = stats->rs_blocks; blocks 278 include/trace/events/jbd2.h __entry->handle_count, __entry->blocks, blocks 32 include/trace/events/mmc.h __field(unsigned int, blocks) blocks 61 include/trace/events/mmc.h __entry->blocks = mrq->data ? mrq->data->blocks : 0; blocks 89 include/trace/events/mmc.h __entry->blocks, __entry->blksz, blocks 48 include/uapi/linux/cramfs_fs.h __u32 blocks; blocks 221 include/uapi/linux/fuse.h uint64_t blocks; blocks 238 include/uapi/linux/fuse.h uint64_t blocks; blocks 23 include/uapi/linux/mmc/ioctl.h unsigned int blocks; blocks 608 include/uapi/linux/soundcard.h int blocks; /* # of fragment transitions since last time */ blocks 90 include/uapi/linux/v4l2-common.h __u32 blocks; blocks 269 kernel/bpf/core.c u32 i, bsize, psize, blocks; blocks 312 kernel/bpf/core.c blocks = bsize / SHA_MESSAGE_BYTES; blocks 318 kernel/bpf/core.c blocks++; blocks 322 kernel/bpf/core.c while (blocks--) { blocks 365 kernel/power/snapshot.c unsigned int blocks; /* Number of Bitmap Blocks */ blocks 435 kernel/power/snapshot.c block_nr = zone->blocks; blocks 464 kernel/power/snapshot.c block_nr = zone->blocks; blocks 482 kernel/power/snapshot.c zone->blocks += 1; blocks 83 mm/dmapool.c unsigned blocks = 0; blocks 88 mm/dmapool.c blocks += page->in_use; blocks 94 mm/dmapool.c pool->name, blocks, blocks 112 mm/shmem.c unsigned long long blocks; blocks 3435 mm/shmem.c ctx->blocks = DIV_ROUND_UP(size, PAGE_SIZE); blocks 3439 mm/shmem.c ctx->blocks = memparse(param->string, &rest); blocks 3549 mm/shmem.c if ((ctx->seen & SHMEM_SEEN_BLOCKS) && ctx->blocks) { blocks 3555 mm/shmem.c ctx->blocks) > 0) { blocks 3574 mm/shmem.c sbinfo->max_blocks = ctx->blocks; blocks 3656 mm/shmem.c ctx->blocks = shmem_default_max_blocks(); blocks 3667 mm/shmem.c sbinfo->max_blocks = ctx->blocks; blocks 1031 mm/vmstat.c unsigned long blocks; blocks 1034 mm/vmstat.c blocks = zone->free_area[order].nr_free; blocks 1035 mm/vmstat.c info->free_blocks_total += blocks; blocks 1038 mm/vmstat.c info->free_pages += blocks << order; blocks 1042 mm/vmstat.c info->free_blocks_suitable += blocks << blocks 1914 net/9p/client.c &sb->bsize, &sb->blocks, &sb->bfree, &sb->bavail, blocks 1926 net/9p/client.c sb->blocks, sb->bfree, sb->bavail, sb->files, sb->ffree, blocks 4106 net/bluetooth/hci_core.c int blocks; blocks 4117 net/bluetooth/hci_core.c blocks = __get_blocks(hdev, skb); blocks 4118 net/bluetooth/hci_core.c if (blocks > hdev->block_cnt) blocks 4127 net/bluetooth/hci_core.c hdev->block_cnt -= blocks; blocks 4128 net/bluetooth/hci_core.c quote -= blocks; blocks 4130 net/bluetooth/hci_core.c chan->sent += blocks; blocks 4131 net/bluetooth/hci_core.c chan->conn->sent += blocks; blocks 3746 net/bluetooth/hci_event.c block_count = __le16_to_cpu(info->blocks); blocks 57 net/mac80211/michael.c size_t block, blocks, left; blocks 63 net/mac80211/michael.c blocks = data_len / 4; blocks 66 net/mac80211/michael.c for (block = 0; block < blocks; block++) blocks 75 net/mac80211/michael.c val |= data[blocks * 4 + left]; blocks 1741 net/sctp/outqueue.c __u16 tsn_offset, blocks; blocks 1760 net/sctp/outqueue.c blocks = ntohs(sack->num_gap_ack_blocks); blocks 1762 net/sctp/outqueue.c for (i = 0; i < blocks; ++i) { blocks 669 security/apparmor/apparmorfs.c u32 bytes, blocks; blocks 681 security/apparmor/apparmorfs.c if (buf_len < sizeof(bytes) + sizeof(blocks)) blocks 698 security/apparmor/apparmorfs.c memset(buf, 0, sizeof(bytes) + sizeof(blocks)); blocks 699 security/apparmor/apparmorfs.c out = buf + sizeof(bytes) + sizeof(blocks); blocks 701 security/apparmor/apparmorfs.c blocks = 0; blocks 720 security/apparmor/apparmorfs.c blocks++; blocks 727 security/apparmor/apparmorfs.c outle32 = __cpu_to_le32(blocks); blocks 2191 sound/core/oss/pcm_oss.c info.blocks = n / runtime->period_size; blocks 2200 sound/core/oss/pcm_oss.c info.blocks = (runtime->oss.buffer_bytes - delay - fixup) / runtime->oss.period_bytes; blocks 2202 sound/core/oss/pcm_oss.c info.blocks = (delay + fixup) / runtime->oss.period_bytes; blocks 2206 sound/core/oss/pcm_oss.c info.blocks = delay / runtime->oss.period_bytes; blocks 41 sound/isa/gus/gus_pcm.c unsigned int blocks; blocks 117 sound/isa/gus/gus_pcm.c if (pcmp->blocks == 1) { blocks 208 sound/isa/gus/gus_pcm.c pcmp->bpos %= pcmp->blocks; blocks 209 sound/isa/gus/gus_pcm.c if (pcmp->bpos + 1 >= pcmp->blocks) { /* last block? */ blocks 498 sound/isa/gus/gus_pcm.c pcmp->blocks = pcmp->dma_size / pcmp->block_size; blocks 2465 sound/soc/codecs/wm_adsp.c int ret, pos, blocks, type, offset, reg; blocks 2514 sound/soc/codecs/wm_adsp.c blocks = 0; blocks 2523 sound/soc/codecs/wm_adsp.c file, blocks, le32_to_cpu(blk->id), blocks 2528 sound/soc/codecs/wm_adsp.c file, blocks, le32_to_cpu(blk->len), offset, type); blocks 2565 sound/soc/codecs/wm_adsp.c file, blocks, le32_to_cpu(blk->len), blocks 2588 sound/soc/codecs/wm_adsp.c file, blocks, type, pos); blocks 2597 sound/soc/codecs/wm_adsp.c file, blocks, region_name, blocks 2614 sound/soc/codecs/wm_adsp.c file, blocks, le32_to_cpu(blk->len), blocks 2621 sound/soc/codecs/wm_adsp.c file, blocks, reg, region_name, ret); blocks 2626 sound/soc/codecs/wm_adsp.c blocks++; blocks 2635 sound/soc/codecs/wm_adsp.c file, blocks, pos - firmware->size); blocks 240 sound/soc/intel/atom/sst/sst.h u32 blocks; blocks 182 sound/soc/intel/atom/sst/sst_loader.c module->blocks, module->type); blocks 187 sound/soc/intel/atom/sst/sst_loader.c for (count = 0; count < module->blocks; count++) { blocks 52 sound/soc/intel/baytrail/sst-baytrail-dsp.c u32 blocks; /* # of blocks */ blocks 75 sound/soc/intel/baytrail/sst-baytrail-dsp.c for (count = 0; count < module->blocks; count++) { blocks 240 sound/soc/intel/baytrail/sst-baytrail-dsp.c int blocks; blocks 324 sound/soc/intel/baytrail/sst-baytrail-dsp.c size = (region[i].end - region[i].start) / region[i].blocks; blocks 327 sound/soc/intel/baytrail/sst-baytrail-dsp.c for (j = 0; j < region[i].blocks; j++) { blocks 70 sound/soc/intel/haswell/sst-haswell-dsp.c __le32 blocks; /* # of blocks */ blocks 102 sound/soc/intel/haswell/sst-haswell-dsp.c module->blocks, type); blocks 119 sound/soc/intel/haswell/sst-haswell-dsp.c for (count = 0; count < le32_to_cpu(module->blocks); count++) { blocks 445 sound/soc/intel/haswell/sst-haswell-dsp.c int blocks; blocks 663 sound/soc/intel/haswell/sst-haswell-dsp.c size = (region[i].end - region[i].start) / region[i].blocks; blocks 666 sound/soc/intel/haswell/sst-haswell-dsp.c for (j = 0; j < region[i].blocks; j++) { blocks 9 tools/perf/util/block-range.c u64 blocks; blocks 298 tools/perf/util/block-range.c block_ranges.blocks++; blocks 320 tools/perf/util/block-range.c if (block_ranges.blocks) blocks 34 tools/perf/util/call-path.c INIT_LIST_HEAD(&cpr->blocks); blocks 42 tools/perf/util/call-path.c list_for_each_entry_safe(pos, n, &cpr->blocks, node) { blocks 59 tools/perf/util/call-path.c cpb = list_last_entry(&cpr->blocks, struct call_path_block, blocks 65 tools/perf/util/call-path.c list_add_tail(&cpb->node, &cpr->blocks); blocks 56 tools/perf/util/call-path.h struct list_head blocks; blocks 630 tools/testing/selftests/net/psock_tpacket.c static void __v1_v2_fill(struct ring *ring, unsigned int blocks) blocks 634 tools/testing/selftests/net/psock_tpacket.c ring->req.tp_block_nr = blocks; blocks 646 tools/testing/selftests/net/psock_tpacket.c static void __v3_fill(struct ring *ring, unsigned int blocks, int type) blocks 655 tools/testing/selftests/net/psock_tpacket.c ring->req3.tp_block_nr = blocks; blocks 670 tools/testing/selftests/net/psock_tpacket.c unsigned int blocks = 256; blocks 680 tools/testing/selftests/net/psock_tpacket.c __v1_v2_fill(ring, blocks); blocks 686 tools/testing/selftests/net/psock_tpacket.c __v3_fill(ring, blocks, type);