blk 18 arch/mips/alchemy/common/vss.c #define VSS_ADDR(blk) (KSEG1ADDR(AU1300_VSS_PHYS_ADDR) + (blk * 0x0c)) blk 87 arch/powerpc/include/asm/rheap.h void rh_dump_blk(rh_info_t *info, rh_block_t *blk); blk 47 arch/powerpc/lib/rheap.c rh_block_t *block, *blk; blk 73 arch/powerpc/lib/rheap.c for (i = 0, blk = block; i < info->max_blocks; i++, blk++) blk 74 arch/powerpc/lib/rheap.c fixup(blks, blke, delta, &blk->list); blk 91 arch/powerpc/lib/rheap.c blk = block + info->max_blocks - new_blocks; blk 92 arch/powerpc/lib/rheap.c for (i = 0; i < new_blocks; i++, blk++) blk 93 arch/powerpc/lib/rheap.c list_add(&blk->list, &info->empty_list); blk 123 arch/powerpc/lib/rheap.c rh_block_t *blk; blk 133 arch/powerpc/lib/rheap.c blk = list_entry(info->empty_list.next, rh_block_t, list); blk 134 arch/powerpc/lib/rheap.c list_del_init(&blk->list); blk 138 arch/powerpc/lib/rheap.c blk->start = 0; blk 139 arch/powerpc/lib/rheap.c blk->size = 0; blk 140 arch/powerpc/lib/rheap.c blk->owner = NULL; blk 142 arch/powerpc/lib/rheap.c return blk; blk 145 arch/powerpc/lib/rheap.c static inline void release_slot(rh_info_t * info, rh_block_t * blk) blk 147 arch/powerpc/lib/rheap.c list_add(&blk->list, &info->empty_list); blk 153 arch/powerpc/lib/rheap.c rh_block_t *blk; blk 173 arch/powerpc/lib/rheap.c blk = list_entry(l, rh_block_t, list); blk 175 arch/powerpc/lib/rheap.c bs = blk->start; blk 176 arch/powerpc/lib/rheap.c be = bs + blk->size; blk 179 arch/powerpc/lib/rheap.c next = blk; blk 182 arch/powerpc/lib/rheap.c before = blk; blk 185 arch/powerpc/lib/rheap.c after = blk; blk 234 arch/powerpc/lib/rheap.c rh_block_t *blk; blk 239 arch/powerpc/lib/rheap.c blk = list_entry(l, rh_block_t, list); blk 240 arch/powerpc/lib/rheap.c if (blk->start > blkn->start) { blk 241 arch/powerpc/lib/rheap.c list_add_tail(&blkn->list, &blk->list); blk 304 arch/powerpc/lib/rheap.c rh_block_t *blk; blk 323 arch/powerpc/lib/rheap.c for (i = 0, blk = block; i < max_blocks; i++, blk++) blk 324 arch/powerpc/lib/rheap.c list_add(&blk->list, &info->empty_list); blk 331 arch/powerpc/lib/rheap.c rh_block_t *blk; blk 358 arch/powerpc/lib/rheap.c blk = get_slot(info); blk 359 arch/powerpc/lib/rheap.c blk->start = start; blk 360 arch/powerpc/lib/rheap.c blk->size = size; blk 361 arch/powerpc/lib/rheap.c blk->owner = NULL; blk 363 arch/powerpc/lib/rheap.c attach_free_block(info, blk); blk 373 arch/powerpc/lib/rheap.c rh_block_t *blk, *newblk; blk 394 arch/powerpc/lib/rheap.c blk = NULL; blk 396 arch/powerpc/lib/rheap.c blk = list_entry(l, rh_block_t, list); blk 398 arch/powerpc/lib/rheap.c bs = blk->start; blk 399 arch/powerpc/lib/rheap.c be = blk->start + blk->size; blk 402 arch/powerpc/lib/rheap.c blk = NULL; blk 405 arch/powerpc/lib/rheap.c if (blk == NULL) blk 411 arch/powerpc/lib/rheap.c list_del(&blk->list); blk 412 arch/powerpc/lib/rheap.c release_slot(info, blk); blk 419 arch/powerpc/lib/rheap.c blk->start += size; blk 420 arch/powerpc/lib/rheap.c blk->size -= size; blk 424 arch/powerpc/lib/rheap.c blk->size = s - bs; blk 431 arch/powerpc/lib/rheap.c list_add(&newblk->list, &blk->list); blk 445 arch/powerpc/lib/rheap.c rh_block_t *blk; blk 459 arch/powerpc/lib/rheap.c blk = NULL; blk 461 arch/powerpc/lib/rheap.c blk = list_entry(l, rh_block_t, list); blk 462 arch/powerpc/lib/rheap.c if (size <= blk->size) { blk 463 arch/powerpc/lib/rheap.c start = (blk->start + alignment - 1) & ~(alignment - 1); blk 464 arch/powerpc/lib/rheap.c if (start + size <= blk->start + blk->size) blk 467 arch/powerpc/lib/rheap.c blk = NULL; blk 470 arch/powerpc/lib/rheap.c if (blk == NULL) blk 474 arch/powerpc/lib/rheap.c if (blk->size == size) { blk 476 arch/powerpc/lib/rheap.c list_del(&blk->list); blk 477 arch/powerpc/lib/rheap.c newblk = blk; blk 481 arch/powerpc/lib/rheap.c sp_size = start - blk->start; blk 486 arch/powerpc/lib/rheap.c spblk->start = blk->start; blk 489 arch/powerpc/lib/rheap.c list_add(&spblk->list, blk->list.prev); blk 497 arch/powerpc/lib/rheap.c blk->start = start + size; blk 498 arch/powerpc/lib/rheap.c blk->size -= sp_size + size; blk 500 arch/powerpc/lib/rheap.c if (blk->size == 0) { blk 501 arch/powerpc/lib/rheap.c list_del(&blk->list); blk 502 arch/powerpc/lib/rheap.c release_slot(info, blk); blk 530 arch/powerpc/lib/rheap.c rh_block_t *blk, *newblk1, *newblk2; blk 551 arch/powerpc/lib/rheap.c blk = NULL; blk 553 arch/powerpc/lib/rheap.c blk = list_entry(l, rh_block_t, list); blk 555 arch/powerpc/lib/rheap.c bs = blk->start; blk 556 arch/powerpc/lib/rheap.c be = blk->start + blk->size; blk 559 arch/powerpc/lib/rheap.c blk = NULL; blk 562 arch/powerpc/lib/rheap.c if (blk == NULL) blk 568 arch/powerpc/lib/rheap.c list_del(&blk->list); blk 569 arch/powerpc/lib/rheap.c blk->owner = owner; blk 571 arch/powerpc/lib/rheap.c start = blk->start; blk 572 arch/powerpc/lib/rheap.c attach_taken_block(info, blk); blk 581 arch/powerpc/lib/rheap.c blk->start += size; blk 582 arch/powerpc/lib/rheap.c blk->size -= size; blk 586 arch/powerpc/lib/rheap.c blk->size = s - bs; blk 593 arch/powerpc/lib/rheap.c list_add(&newblk2->list, &blk->list); blk 614 arch/powerpc/lib/rheap.c rh_block_t *blk, *blk2; blk 619 arch/powerpc/lib/rheap.c blk = NULL; blk 624 arch/powerpc/lib/rheap.c blk = blk2; blk 627 arch/powerpc/lib/rheap.c if (blk == NULL || start > (blk->start + blk->size)) blk 631 arch/powerpc/lib/rheap.c list_del(&blk->list); blk 634 arch/powerpc/lib/rheap.c size = blk->size; blk 635 arch/powerpc/lib/rheap.c attach_free_block(info, blk); blk 643 arch/powerpc/lib/rheap.c rh_block_t *blk; blk 665 arch/powerpc/lib/rheap.c blk = list_entry(l, rh_block_t, list); blk 667 arch/powerpc/lib/rheap.c stats->start = blk->start; blk 668 arch/powerpc/lib/rheap.c stats->size = blk->size; blk 669 arch/powerpc/lib/rheap.c stats->owner = blk->owner; blk 681 arch/powerpc/lib/rheap.c rh_block_t *blk, *blk2; blk 686 arch/powerpc/lib/rheap.c blk = NULL; blk 691 arch/powerpc/lib/rheap.c blk = blk2; blk 694 arch/powerpc/lib/rheap.c if (blk == NULL || start > (blk->start + blk->size)) blk 697 arch/powerpc/lib/rheap.c blk->owner = owner; blk 698 arch/powerpc/lib/rheap.c size = blk->size; blk 740 arch/powerpc/lib/rheap.c void rh_dump_blk(rh_info_t * info, rh_block_t * blk) blk 744 arch/powerpc/lib/rheap.c blk, blk->start, blk->start + blk->size, blk->size); blk 47 arch/s390/crypto/aes_s390.c struct crypto_sync_skcipher *blk; blk 187 arch/s390/crypto/aes_s390.c crypto_sync_skcipher_clear_flags(sctx->fallback.blk, blk 189 arch/s390/crypto/aes_s390.c crypto_sync_skcipher_set_flags(sctx->fallback.blk, tfm->crt_flags & blk 192 arch/s390/crypto/aes_s390.c ret = crypto_sync_skcipher_setkey(sctx->fallback.blk, key, len); blk 195 arch/s390/crypto/aes_s390.c tfm->crt_flags |= crypto_sync_skcipher_get_flags(sctx->fallback.blk) & blk 208 arch/s390/crypto/aes_s390.c SYNC_SKCIPHER_REQUEST_ON_STACK(req, sctx->fallback.blk); blk 210 arch/s390/crypto/aes_s390.c skcipher_request_set_sync_tfm(req, sctx->fallback.blk); blk 227 arch/s390/crypto/aes_s390.c SYNC_SKCIPHER_REQUEST_ON_STACK(req, sctx->fallback.blk); blk 229 arch/s390/crypto/aes_s390.c skcipher_request_set_sync_tfm(req, sctx->fallback.blk); blk 310 arch/s390/crypto/aes_s390.c sctx->fallback.blk = crypto_alloc_sync_skcipher(name, 0, blk 313 arch/s390/crypto/aes_s390.c if (IS_ERR(sctx->fallback.blk)) { blk 316 arch/s390/crypto/aes_s390.c return PTR_ERR(sctx->fallback.blk); blk 326 arch/s390/crypto/aes_s390.c crypto_free_sync_skcipher(sctx->fallback.blk); blk 149 arch/x86/mm/numa.c mi->blk[mi->nr_blks].start = start; blk 150 arch/x86/mm/numa.c mi->blk[mi->nr_blks].end = end; blk 151 arch/x86/mm/numa.c mi->blk[mi->nr_blks].nid = nid; blk 167 arch/x86/mm/numa.c memmove(&mi->blk[idx], &mi->blk[idx + 1], blk 168 arch/x86/mm/numa.c (mi->nr_blks - idx) * sizeof(mi->blk[0])); blk 238 arch/x86/mm/numa.c struct numa_memblk *bi = &mi->blk[i]; blk 253 arch/x86/mm/numa.c struct numa_memblk *bi = &mi->blk[i]; blk 256 arch/x86/mm/numa.c struct numa_memblk *bj = &mi->blk[j]; blk 286 arch/x86/mm/numa.c struct numa_memblk *bk = &mi->blk[k]; blk 305 arch/x86/mm/numa.c for (i = mi->nr_blks; i < ARRAY_SIZE(mi->blk); i++) { blk 306 arch/x86/mm/numa.c mi->blk[i].start = mi->blk[i].end = 0; blk 307 arch/x86/mm/numa.c mi->blk[i].nid = NUMA_NO_NODE; blk 321 arch/x86/mm/numa.c for (i = 0; i < ARRAY_SIZE(mi->blk); i++) blk 322 arch/x86/mm/numa.c if (mi->blk[i].start != mi->blk[i].end && blk 323 arch/x86/mm/numa.c mi->blk[i].nid != NUMA_NO_NODE) blk 324 arch/x86/mm/numa.c node_set(mi->blk[i].nid, *nodemask); blk 443 arch/x86/mm/numa.c u64 s = mi->blk[i].start >> PAGE_SHIFT; blk 444 arch/x86/mm/numa.c u64 e = mi->blk[i].end >> PAGE_SHIFT; blk 446 arch/x86/mm/numa.c numaram -= __absent_pages_in_range(mi->blk[i].nid, s, e); blk 488 arch/x86/mm/numa.c struct numa_memblk *mb = numa_meminfo.blk + i; blk 518 arch/x86/mm/numa.c struct numa_memblk *mb = numa_meminfo.blk + i; blk 539 arch/x86/mm/numa.c struct numa_memblk *mb = &mi->blk[i]; blk 575 arch/x86/mm/numa.c if (nid != mi->blk[i].nid) blk 577 arch/x86/mm/numa.c start = min(mi->blk[i].start, start); blk 578 arch/x86/mm/numa.c end = max(mi->blk[i].end, end); blk 888 arch/x86/mm/numa.c int nid = mi->blk[0].nid; blk 892 arch/x86/mm/numa.c if (mi->blk[i].start <= start && mi->blk[i].end > start) blk 893 arch/x86/mm/numa.c nid = mi->blk[i].nid; blk 26 arch/x86/mm/numa_emulation.c if (mi->blk[i].nid == nid) blk 49 arch/x86/mm/numa_emulation.c struct numa_memblk *eb = &ei->blk[ei->nr_blks]; blk 50 arch/x86/mm/numa_emulation.c struct numa_memblk *pb = &pi->blk[phys_blk]; blk 136 arch/x86/mm/numa_emulation.c start = pi->blk[phys_blk].start; blk 137 arch/x86/mm/numa_emulation.c limit = pi->blk[phys_blk].end; blk 284 arch/x86/mm/numa_emulation.c start = pi->blk[phys_blk].start; blk 285 arch/x86/mm/numa_emulation.c limit = pi->blk[phys_blk].end; blk 411 arch/x86/mm/numa_emulation.c pi.blk[0].start, pi.blk[0].end, 0, blk 412 arch/x86/mm/numa_emulation.c n, &pi.blk[0], nid); blk 475 arch/x86/mm/numa_emulation.c for (i = 0; i < ARRAY_SIZE(ei.blk); i++) blk 476 arch/x86/mm/numa_emulation.c if (ei.blk[i].start != ei.blk[i].end && blk 477 arch/x86/mm/numa_emulation.c ei.blk[i].nid != NUMA_NO_NODE) blk 478 arch/x86/mm/numa_emulation.c node_set(ei.blk[i].nid, numa_nodes_parsed); blk 16 arch/x86/mm/numa_internal.h struct numa_memblk blk[NR_NODE_MEMBLKS]; blk 35 block/partitions/amiga.c int start_sect, nr_sects, blk, part, res = 0; blk 40 block/partitions/amiga.c for (blk = 0; ; blk++, put_dev_sector(sect)) { blk 41 block/partitions/amiga.c if (blk == RDB_ALLOCATION_LIMIT) blk 43 block/partitions/amiga.c data = read_part_sector(state, blk, §); blk 47 block/partitions/amiga.c bdevname(state->bdev, b), blk); blk 64 block/partitions/amiga.c blk); blk 69 block/partitions/amiga.c bdevname(state->bdev, b), blk); blk 82 block/partitions/amiga.c blk = be32_to_cpu(rdb->rdb_PartitionList); blk 84 block/partitions/amiga.c for (part = 1; blk>0 && part<=16; part++, put_dev_sector(sect)) { blk 85 block/partitions/amiga.c blk *= blksize; /* Read in terms partition table understands */ blk 86 block/partitions/amiga.c data = read_part_sector(state, blk, §); blk 90 block/partitions/amiga.c bdevname(state->bdev, b), blk); blk 95 block/partitions/amiga.c blk = be32_to_cpu(pb->pb_Next); blk 139 block/partitions/ibm.c sector_t blk; blk 155 block/partitions/ibm.c blk = cchhb2blk(&label->vol.vtoc, geo) + 1; blk 157 block/partitions/ibm.c data = read_part_sector(state, blk * secperblk, §); blk 166 block/partitions/ibm.c blk++; blk 167 block/partitions/ibm.c data = read_part_sector(state, blk * secperblk, §); blk 184 block/partitions/ibm.c blk++; blk 185 block/partitions/ibm.c data = read_part_sector(state, blk * secperblk, §); blk 143 drivers/accessibility/braille/braille_console.c static int keyboard_notifier_call(struct notifier_block *blk, blk 262 drivers/accessibility/braille/braille_console.c static int vt_notifier_call(struct notifier_block *blk, blk 272 drivers/base/regmap/regcache-rbtree.c u8 *blk; blk 278 drivers/base/regmap/regcache-rbtree.c blk = krealloc(rbnode->block, blk 281 drivers/base/regmap/regcache-rbtree.c if (!blk) blk 289 drivers/base/regmap/regcache-rbtree.c kfree(blk); blk 302 drivers/base/regmap/regcache-rbtree.c memmove(blk + offset * map->cache_word_size, blk 303 drivers/base/regmap/regcache-rbtree.c blk, rbnode->blklen * map->cache_word_size); blk 308 drivers/base/regmap/regcache-rbtree.c rbnode->block = blk; blk 3090 drivers/cdrom/cdrom.c struct cdrom_blk blk; blk 3092 drivers/cdrom/cdrom.c if (copy_from_user(&blk, (struct cdrom_blk __user *)arg, sizeof(blk))) blk 3095 drivers/cdrom/cdrom.c cgc->cmd[2] = (blk.from >> 24) & 0xff; blk 3096 drivers/cdrom/cdrom.c cgc->cmd[3] = (blk.from >> 16) & 0xff; blk 3097 drivers/cdrom/cdrom.c cgc->cmd[4] = (blk.from >> 8) & 0xff; blk 3098 drivers/cdrom/cdrom.c cgc->cmd[5] = blk.from & 0xff; blk 3099 drivers/cdrom/cdrom.c cgc->cmd[7] = (blk.len >> 8) & 0xff; blk 3100 drivers/cdrom/cdrom.c cgc->cmd[8] = blk.len & 0xff; blk 533 drivers/crypto/mxs-dcp.c struct crypto_sync_skcipher *blk; blk 535 drivers/crypto/mxs-dcp.c blk = crypto_alloc_sync_skcipher(name, 0, CRYPTO_ALG_NEED_FALLBACK); blk 536 drivers/crypto/mxs-dcp.c if (IS_ERR(blk)) blk 537 drivers/crypto/mxs-dcp.c return PTR_ERR(blk); blk 539 drivers/crypto/mxs-dcp.c actx->fallback = blk; blk 613 drivers/crypto/omap-aes.c struct crypto_sync_skcipher *blk; blk 615 drivers/crypto/omap-aes.c blk = crypto_alloc_sync_skcipher(name, 0, CRYPTO_ALG_NEED_FALLBACK); blk 616 drivers/crypto/omap-aes.c if (IS_ERR(blk)) blk 617 drivers/crypto/omap-aes.c return PTR_ERR(blk); blk 619 drivers/crypto/omap-aes.c ctx->fallback = blk; blk 60 drivers/edac/edac_device.c struct edac_device_block *dev_blk, *blk_p, *blk; blk 158 drivers/edac/edac_device.c blk = &blk_p[block]; blk 159 drivers/edac/edac_device.c blk->instance = inst; blk 160 drivers/edac/edac_device.c snprintf(blk->name, sizeof(blk->name), blk 164 drivers/edac/edac_device.c instance, inst, block, blk, blk->name); blk 173 drivers/edac/edac_device.c blk->nr_attribs = nr_attrib; blk 175 drivers/edac/edac_device.c blk->block_attributes = attrib_p; blk 178 drivers/edac/edac_device.c blk->block_attributes); blk 195 drivers/edac/edac_device.c attrib->block = blk; /* up link */ blk 65 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h #define HSWEQ_DCN_PIXEL_RATE_REG_LIST(blk, inst) \ blk 66 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h SRII(PIXEL_RATE_CNTL, blk, inst), \ blk 67 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h SRII(PHYPLL_PIXEL_RATE_CNTL, blk, inst) blk 69 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h #define HWSEQ_PIXEL_RATE_REG_LIST(blk) \ blk 70 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h SRII(PIXEL_RATE_CNTL, blk, 0), \ blk 71 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h SRII(PIXEL_RATE_CNTL, blk, 1), \ blk 72 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h SRII(PIXEL_RATE_CNTL, blk, 2), \ blk 73 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h SRII(PIXEL_RATE_CNTL, blk, 3), \ blk 74 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h SRII(PIXEL_RATE_CNTL, blk, 4), \ blk 75 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h SRII(PIXEL_RATE_CNTL, blk, 5) blk 77 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h #define HWSEQ_PHYPLL_REG_LIST(blk) \ blk 78 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h SRII(PHYPLL_PIXEL_RATE_CNTL, blk, 0), \ blk 79 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h SRII(PHYPLL_PIXEL_RATE_CNTL, blk, 1), \ blk 80 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h SRII(PHYPLL_PIXEL_RATE_CNTL, blk, 2), \ blk 81 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h SRII(PHYPLL_PIXEL_RATE_CNTL, blk, 3), \ blk 82 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h SRII(PHYPLL_PIXEL_RATE_CNTL, blk, 4), \ blk 83 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h SRII(PHYPLL_PIXEL_RATE_CNTL, blk, 5) blk 449 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h #define HWSEQ_DCEF_MASK_SH_LIST(mask_sh, blk)\ blk 450 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h HWS_SF(blk, CLOCK_CONTROL, DCFE_CLOCK_ENABLE, mask_sh),\ blk 453 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h #define HWSEQ_BLND_MASK_SH_LIST(mask_sh, blk)\ blk 454 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h HWS_SF(blk, V_UPDATE_LOCK, BLND_DCP_GRPH_V_UPDATE_LOCK, mask_sh),\ blk 455 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h HWS_SF(blk, V_UPDATE_LOCK, BLND_SCL_V_UPDATE_LOCK, mask_sh),\ blk 456 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h HWS_SF(blk, V_UPDATE_LOCK, BLND_DCP_GRPH_SURF_V_UPDATE_LOCK, mask_sh),\ blk 457 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h HWS_SF(blk, V_UPDATE_LOCK, BLND_BLND_V_UPDATE_LOCK, mask_sh),\ blk 458 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h HWS_SF(blk, V_UPDATE_LOCK, BLND_V_UPDATE_LOCK_MODE, mask_sh),\ blk 459 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h HWS_SF(blk, CONTROL, BLND_FEEDTHROUGH_EN, mask_sh),\ blk 460 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h HWS_SF(blk, CONTROL, BLND_ALPHA_MODE, mask_sh),\ blk 461 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h HWS_SF(blk, CONTROL, BLND_MODE, mask_sh),\ blk 462 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h HWS_SF(blk, CONTROL, BLND_MULTIPLIED_MODE, mask_sh) blk 464 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h #define HWSEQ_PIXEL_RATE_MASK_SH_LIST(mask_sh, blk)\ blk 465 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h HWS_SF1(blk, PIXEL_RATE_CNTL, PIXEL_RATE_SOURCE, mask_sh),\ blk 466 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h HWS_SF(blk, PIXEL_RATE_CNTL, DP_DTO0_ENABLE, mask_sh) blk 468 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h #define HWSEQ_PHYPLL_MASK_SH_LIST(mask_sh, blk)\ blk 469 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h HWS_SF1(blk, PHYPLL_PIXEL_RATE_CNTL, PHYPLL_PIXEL_RATE_SOURCE, mask_sh),\ blk 470 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.h HWS_SF1(blk, PHYPLL_PIXEL_RATE_CNTL, PIXEL_RATE_PLL_SOURCE, mask_sh) blk 129 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h #define MI_GFX8_TILE_MASK_SH_LIST(mask_sh, blk)\ blk 130 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_CONTROL, GRPH_NUM_BANKS, mask_sh),\ blk 131 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_CONTROL, GRPH_BANK_WIDTH, mask_sh),\ blk 132 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_CONTROL, GRPH_BANK_HEIGHT, mask_sh),\ blk 133 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_CONTROL, GRPH_MACRO_TILE_ASPECT, mask_sh),\ blk 134 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_CONTROL, GRPH_TILE_SPLIT, mask_sh),\ blk 135 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_CONTROL, GRPH_MICRO_TILE_MODE, mask_sh),\ blk 136 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_CONTROL, GRPH_PIPE_CONFIG, mask_sh),\ blk 137 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_CONTROL, GRPH_ARRAY_MODE, mask_sh),\ blk 138 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_CONTROL, GRPH_COLOR_EXPANSION_MODE, mask_sh) blk 140 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h #define MI_DCP_MASK_SH_LIST(mask_sh, blk)\ blk 141 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_ENABLE, GRPH_ENABLE, mask_sh),\ blk 142 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_CONTROL, GRPH_DEPTH, mask_sh),\ blk 143 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_CONTROL, GRPH_FORMAT, mask_sh),\ blk 144 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_CONTROL, GRPH_NUM_BANKS, mask_sh),\ blk 145 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_X_START, GRPH_X_START, mask_sh),\ blk 146 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_Y_START, GRPH_Y_START, mask_sh),\ blk 147 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_X_END, GRPH_X_END, mask_sh),\ blk 148 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_Y_END, GRPH_Y_END, mask_sh),\ blk 149 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_PITCH, GRPH_PITCH, mask_sh),\ blk 150 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, HW_ROTATION, GRPH_ROTATION_ANGLE, mask_sh),\ blk 151 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_SWAP_CNTL, GRPH_RED_CROSSBAR, mask_sh),\ blk 152 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_SWAP_CNTL, GRPH_BLUE_CROSSBAR, mask_sh),\ blk 153 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, PRESCALE_GRPH_CONTROL, GRPH_PRESCALE_SELECT, mask_sh),\ blk 154 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, PRESCALE_GRPH_CONTROL, GRPH_PRESCALE_R_SIGN, mask_sh),\ blk 155 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, PRESCALE_GRPH_CONTROL, GRPH_PRESCALE_G_SIGN, mask_sh),\ blk 156 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, PRESCALE_GRPH_CONTROL, GRPH_PRESCALE_B_SIGN, mask_sh),\ blk 157 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_SECONDARY_SURFACE_ADDRESS_HIGH, GRPH_SECONDARY_SURFACE_ADDRESS_HIGH, mask_sh),\ blk 158 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_SECONDARY_SURFACE_ADDRESS, GRPH_SECONDARY_SURFACE_ADDRESS, mask_sh),\ blk 159 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_PRIMARY_SURFACE_ADDRESS_HIGH, GRPH_PRIMARY_SURFACE_ADDRESS_HIGH, mask_sh),\ blk 160 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_PRIMARY_SURFACE_ADDRESS, GRPH_PRIMARY_SURFACE_ADDRESS, mask_sh),\ blk 161 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_UPDATE, GRPH_SURFACE_UPDATE_PENDING, mask_sh),\ blk 162 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_UPDATE, GRPH_UPDATE_LOCK, mask_sh),\ blk 163 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_FLIP_CONTROL, GRPH_SURFACE_UPDATE_H_RETRACE_EN, mask_sh) blk 165 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h #define MI_DCP_DCE11_MASK_SH_LIST(mask_sh, blk)\ blk 166 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_PIPE_OUTSTANDING_REQUEST_LIMIT, GRPH_PIPE_OUTSTANDING_REQUEST_LIMIT, mask_sh) blk 168 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h #define MI_DCP_PTE_MASK_SH_LIST(mask_sh, blk)\ blk 169 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DVMM_PTE_CONTROL, DVMM_PAGE_WIDTH, mask_sh),\ blk 170 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DVMM_PTE_CONTROL, DVMM_PAGE_HEIGHT, mask_sh),\ blk 171 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DVMM_PTE_CONTROL, DVMM_MIN_PTE_BEFORE_FLIP, mask_sh),\ blk 172 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DVMM_PTE_ARB_CONTROL, DVMM_PTE_REQ_PER_CHUNK, mask_sh),\ blk 173 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DVMM_PTE_ARB_CONTROL, DVMM_MAX_PTE_REQ_OUTSTANDING, mask_sh) blk 175 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h #define MI_DMIF_PG_MASK_SH_LIST(mask_sh, blk)\ blk 176 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_ARBITRATION_CONTROL1, PIXEL_DURATION, mask_sh),\ blk 177 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_WATERMARK_MASK_CONTROL, URGENCY_WATERMARK_MASK, mask_sh),\ blk 178 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_WATERMARK_MASK_CONTROL, STUTTER_EXIT_SELF_REFRESH_WATERMARK_MASK, mask_sh),\ blk 179 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_URGENCY_CONTROL, URGENCY_LOW_WATERMARK, mask_sh),\ blk 180 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_URGENCY_CONTROL, URGENCY_HIGH_WATERMARK, mask_sh),\ blk 181 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_STUTTER_CONTROL, STUTTER_ENABLE, mask_sh),\ blk 182 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_STUTTER_CONTROL, STUTTER_IGNORE_FBC, mask_sh),\ blk 186 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h #define MI_DMIF_PG_MASK_SH_DCE(mask_sh, blk)\ blk 187 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_STUTTER_CONTROL, STUTTER_EXIT_SELF_REFRESH_WATERMARK, mask_sh),\ blk 188 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_WATERMARK_MASK_CONTROL, NB_PSTATE_CHANGE_WATERMARK_MASK, mask_sh),\ blk 189 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_NB_PSTATE_CHANGE_CONTROL, NB_PSTATE_CHANGE_ENABLE, mask_sh),\ blk 190 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_NB_PSTATE_CHANGE_CONTROL, NB_PSTATE_CHANGE_URGENT_DURING_REQUEST, mask_sh),\ blk 191 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_NB_PSTATE_CHANGE_CONTROL, NB_PSTATE_CHANGE_NOT_SELF_REFRESH_DURING_REQUEST, mask_sh),\ blk 192 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_NB_PSTATE_CHANGE_CONTROL, NB_PSTATE_CHANGE_WATERMARK, mask_sh) blk 208 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h #define MI_GFX9_TILE_MASK_SH_LIST(mask_sh, blk)\ blk 209 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_CONTROL, GRPH_SW_MODE, mask_sh),\ blk 210 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_CONTROL, GRPH_SE_ENABLE, mask_sh),\ blk 211 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_CONTROL, GRPH_NUM_SHADER_ENGINES, mask_sh),\ blk 212 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_CONTROL, GRPH_NUM_PIPES, mask_sh),\ blk 213 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, GRPH_CONTROL, GRPH_COLOR_EXPANSION_MODE, mask_sh) blk 215 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h #define MI_DCE12_DMIF_PG_MASK_SH_LIST(mask_sh, blk)\ blk 216 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_STUTTER_CONTROL2, STUTTER_EXIT_SELF_REFRESH_WATERMARK, mask_sh),\ blk 217 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_STUTTER_CONTROL2, STUTTER_ENTER_SELF_REFRESH_WATERMARK, mask_sh),\ blk 218 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_URGENT_LEVEL_CONTROL, URGENT_LEVEL_LOW_WATERMARK, mask_sh),\ blk 219 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_URGENT_LEVEL_CONTROL, URGENT_LEVEL_HIGH_WATERMARK, mask_sh),\ blk 220 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_URGENCY_CONTROL, URGENCY_LOW_WATERMARK, mask_sh),\ blk 221 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_URGENCY_CONTROL, URGENCY_HIGH_WATERMARK, mask_sh),\ blk 222 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_WATERMARK_MASK_CONTROL, PSTATE_CHANGE_WATERMARK_MASK, mask_sh),\ blk 223 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_LOW_POWER_CONTROL, PSTATE_CHANGE_ENABLE, mask_sh),\ blk 224 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_LOW_POWER_CONTROL, PSTATE_CHANGE_URGENT_DURING_REQUEST, mask_sh),\ blk 225 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_LOW_POWER_CONTROL, PSTATE_CHANGE_NOT_SELF_REFRESH_DURING_REQUEST, mask_sh),\ blk 226 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h SFB(blk, DPG_PIPE_LOW_POWER_CONTROL, PSTATE_CHANGE_WATERMARK, mask_sh) blk 58 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c static u32 get_valid_inputs(struct block_header *blk) blk 63 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c for (i = 0; i < PIPELINE_INFO_N_VALID_INPUTS(blk->pipeline_info); i++) { blk 64 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c get_resources_id(blk->input_ids[i], NULL, &comp_id); blk 342 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c struct block_header *blk, u32 __iomem *reg) blk 348 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c get_resources_id(blk->block_info, &pipe_id, &layer_id); blk 351 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c BLOCK_INFO_INPUT_ID(blk->block_info), blk 353 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c get_valid_inputs(blk), blk 439 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c struct block_header *blk, u32 __iomem *reg) blk 445 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c get_resources_id(blk->block_info, &pipe_id, &layer_id); blk 448 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c layer_id, BLOCK_INFO_INPUT_ID(blk->block_info), blk 450 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c 1, get_valid_inputs(blk), 0, reg, blk 578 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c struct block_header *blk, u32 __iomem *reg) blk 584 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c get_resources_id(blk->block_info, &pipe_id, &comp_id); blk 588 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c BLOCK_INFO_INPUT_ID(blk->block_info), blk 590 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c CU_NUM_INPUT_IDS, get_valid_inputs(blk), blk 735 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c struct block_header *blk, u32 __iomem *reg) blk 741 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c get_resources_id(blk->block_info, &pipe_id, &comp_id); blk 744 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c comp_id, BLOCK_INFO_INPUT_ID(blk->block_info), blk 746 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c 1, get_valid_inputs(blk), 1, reg, blk 748 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c pipe_id, BLOCK_INFO_BLK_ID(blk->block_info)); blk 843 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c struct block_header *blk, u32 __iomem *reg) blk 849 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c get_resources_id(blk->block_info, &pipe_id, &comp_id); blk 853 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c BLOCK_INFO_INPUT_ID(blk->block_info), blk 855 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c 1, get_valid_inputs(blk), 2, reg, blk 913 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c struct block_header *blk, u32 __iomem *reg) blk 919 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c get_resources_id(blk->block_info, &pipe_id, &comp_id); blk 923 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c BLOCK_INFO_INPUT_ID(blk->block_info), blk 925 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c MG_NUM_INPUTS_IDS, get_valid_inputs(blk), blk 990 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c struct block_header *blk, u32 __iomem *reg) blk 996 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c get_resources_id(blk->block_info, &pipe_id, &comp_id); blk 1000 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c BLOCK_INFO_INPUT_ID(blk->block_info), blk 1002 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c get_valid_inputs(blk), blk 1117 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c struct block_header *blk, u32 __iomem *reg) blk 1123 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c get_resources_id(blk->block_info, &pipe_id, &comp_id); blk 1127 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c BLOCK_INFO_INPUT_ID(blk->block_info), blk 1144 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c struct block_header *blk, u32 __iomem *reg) blk 1147 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c int blk_id = BLOCK_INFO_BLK_ID(blk->block_info); blk 1151 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c switch (BLOCK_INFO_BLK_TYPE(blk->block_info)) { blk 1161 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c err = d71_layer_init(d71, blk, reg); blk 1165 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c err = d71_wb_layer_init(d71, blk, reg); blk 1171 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c err = d71_compiz_init(d71, blk, reg); blk 1175 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c err = d71_scaler_init(d71, blk, reg); blk 1179 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c err = d71_splitter_init(d71, blk, reg); blk 1183 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c err = d71_merger_init(d71, blk, reg); blk 1192 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c err = d71_improc_init(d71, blk, reg); blk 1201 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c err = d71_timing_ctrlr_init(d71, blk, reg); blk 1213 drivers/gpu/drm/arm/display/komeda/d71/d71_component.c blk->block_info); blk 310 drivers/gpu/drm/arm/display/komeda/d71/d71_dev.c void d71_read_block_header(u32 __iomem *reg, struct block_header *blk) blk 314 drivers/gpu/drm/arm/display/komeda/d71/d71_dev.c blk->block_info = malidp_read32(reg, BLK_BLOCK_INFO); blk 315 drivers/gpu/drm/arm/display/komeda/d71/d71_dev.c if (BLOCK_INFO_BLK_TYPE(blk->block_info) == D71_BLK_TYPE_RESERVED) blk 318 drivers/gpu/drm/arm/display/komeda/d71/d71_dev.c blk->pipeline_info = malidp_read32(reg, BLK_PIPELINE_INFO); blk 321 drivers/gpu/drm/arm/display/komeda/d71/d71_dev.c for (i = 0; i < PIPELINE_INFO_N_VALID_INPUTS(blk->pipeline_info); i++) blk 322 drivers/gpu/drm/arm/display/komeda/d71/d71_dev.c blk->input_ids[i] = malidp_read32(reg + i, BLK_VALID_INPUT_ID0); blk 323 drivers/gpu/drm/arm/display/komeda/d71/d71_dev.c for (i = 0; i < PIPELINE_INFO_N_OUTPUTS(blk->pipeline_info); i++) blk 324 drivers/gpu/drm/arm/display/komeda/d71/d71_dev.c blk->output_ids[i] = malidp_read32(reg + i, BLK_OUTPUT_ID0); blk 342 drivers/gpu/drm/arm/display/komeda/d71/d71_dev.c struct block_header blk; blk 407 drivers/gpu/drm/arm/display/komeda/d71/d71_dev.c d71_read_block_header(blk_base, &blk); blk 408 drivers/gpu/drm/arm/display/komeda/d71/d71_dev.c if (BLOCK_INFO_BLK_TYPE(blk.block_info) != D71_BLK_TYPE_RESERVED) { blk 409 drivers/gpu/drm/arm/display/komeda/d71/d71_dev.c err = d71_probe_block(d71, &blk, blk_base); blk 49 drivers/gpu/drm/arm/display/komeda/d71/d71_dev.h struct block_header *blk, u32 __iomem *reg); blk 50 drivers/gpu/drm/arm/display/komeda/d71/d71_dev.h void d71_read_block_header(u32 __iomem *reg, struct block_header *blk); blk 525 drivers/gpu/drm/arm/display/komeda/d71/d71_regs.h static inline u32 get_block_type(struct block_header *blk) blk 527 drivers/gpu/drm/arm/display/komeda/d71/d71_regs.h return BLOCK_INFO_BLK_TYPE(blk->block_info); blk 1212 drivers/gpu/drm/i2c/tda998x_drv.c static int read_edid_block(void *data, u8 *buf, unsigned int blk, size_t length) blk 1218 drivers/gpu/drm/i2c/tda998x_drv.c offset = (blk & 1) ? 128 : 0; blk 1219 drivers/gpu/drm/i2c/tda998x_drv.c segptr = blk / 2; blk 1265 drivers/gpu/drm/i2c/tda998x_drv.c blk, ret); blk 128 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h enum dpu_sspp blk); blk 131 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h enum dpu_lm blk); blk 135 drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h enum dpu_intf blk); blk 65 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c if (i->blk && (&i->blk->list == blk_list)) { blk 70 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c i->blk = list_prepare_entry(i->blk, blk_list, list); blk 72 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c list_for_each_entry_continue(i->blk, blk_list, list) { blk 73 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c if (i->enc_id == i->blk->enc_id) { blk 74 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c i->hw = i->blk->hw; blk 76 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c i->type, i->blk->id, i->enc_id); blk 150 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c struct dpu_rm_hw_blk *blk; blk 182 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c blk = kzalloc(sizeof(*blk), GFP_KERNEL); blk 183 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c if (!blk) { blk 188 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c blk->id = id; blk 189 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c blk->hw = hw; blk 190 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c blk->enc_id = 0; blk 191 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c list_add_tail(&blk->list, &rm->hw_blks[type]); blk 341 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c if (iter.blk->id == lm_cfg->pingpong) { blk 342 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c *pp = iter.blk; blk 384 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c lm[lm_count] = iter_i.blk; blk 398 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c if (iter_i.blk == iter_j.blk) blk 402 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c rm, enc_id, reqs, iter_j.blk, blk 403 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c &pp[lm_count], iter_i.blk)) blk 406 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c lm[lm_count] = iter_j.blk; blk 448 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c const struct dpu_hw_ctl *ctl = to_dpu_hw_ctl(iter.blk->hw); blk 452 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c if (RESERVED_BY_OTHER(iter.blk, enc_id)) blk 457 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c DPU_DEBUG("ctl %d caps 0x%lX\n", iter.blk->id, features); blk 462 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c ctls[i] = iter.blk; blk 463 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c DPU_DEBUG("ctl %d match\n", iter.blk->id); blk 492 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c if (iter.blk->id != id) blk 495 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c if (RESERVED_BY_OTHER(iter.blk, enc_id)) { blk 500 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c iter.blk->enc_id = enc_id; blk 501 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c trace_dpu_rm_reserve_intf(iter.blk->id, enc_id); blk 582 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c struct dpu_rm_hw_blk *blk; blk 586 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c list_for_each_entry(blk, &rm->hw_blks[type], list) { blk 587 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c if (blk->enc_id == enc_id) { blk 588 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c blk->enc_id = 0; blk 590 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c type, blk->id); blk 42 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h struct dpu_rm_hw_blk *blk; blk 88 drivers/gpu/drm/msm/disp/mdp5/mdp5_smp.c int blk = find_first_zero_bit(state->state, cnt); blk 89 drivers/gpu/drm/msm/disp/mdp5/mdp5_smp.c set_bit(blk, cs); blk 90 drivers/gpu/drm/msm/disp/mdp5/mdp5_smp.c set_bit(blk, state->state); blk 225 drivers/gpu/drm/msm/disp/mdp5/mdp5_smp.c u32 blk, val; blk 227 drivers/gpu/drm/msm/disp/mdp5/mdp5_smp.c for_each_set_bit(blk, *assigned, cnt) { blk 228 drivers/gpu/drm/msm/disp/mdp5/mdp5_smp.c int idx = blk / 3; blk 229 drivers/gpu/drm/msm/disp/mdp5/mdp5_smp.c int fld = blk % 3; blk 323 drivers/hwtracing/intel_th/msu.c unsigned int blk; blk 326 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { blk 409 drivers/hwtracing/intel_th/msu.c unsigned int blk; blk 421 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { blk 656 drivers/hwtracing/intel_th/msu.c unsigned int blk; blk 660 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { blk 1175 drivers/hwtracing/intel_th/msu.c unsigned int blk; blk 1190 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { blk 1201 drivers/hwtracing/intel_th/msu.c if (blk == win->nr_segs - 1) { blk 1371 drivers/hwtracing/intel_th/msu.c unsigned int blk; blk 1385 drivers/hwtracing/intel_th/msu.c for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { blk 246 drivers/i2c/busses/i2c-qup.c struct qup_i2c_block blk; blk 288 drivers/i2c/busses/i2c-qup.c struct qup_i2c_block *blk = &qup->blk; blk 339 drivers/i2c/busses/i2c-qup.c blk->tx_fifo_free += qup->out_blk_sz; blk 350 drivers/i2c/busses/i2c-qup.c if (!blk->is_rx_blk_mode) { blk 351 drivers/i2c/busses/i2c-qup.c blk->fifo_available += qup->in_fifo_sz; blk 354 drivers/i2c/busses/i2c-qup.c blk->fifo_available += qup->in_blk_sz; blk 360 drivers/i2c/busses/i2c-qup.c if (!blk->rx_bytes_read) blk 370 drivers/i2c/busses/i2c-qup.c if (blk->is_tx_blk_mode && !(opflags & QUP_MX_OUTPUT_DONE)) blk 463 drivers/i2c/busses/i2c-qup.c struct qup_i2c_block *blk = &qup->blk; blk 473 drivers/i2c/busses/i2c-qup.c blk->tx_fifo_free--; blk 479 drivers/i2c/busses/i2c-qup.c while (blk->tx_fifo_free && qup->pos < msg->len) { blk 496 drivers/i2c/busses/i2c-qup.c blk->tx_fifo_free--; blk 503 drivers/i2c/busses/i2c-qup.c qup->blk.pos = 0; blk 504 drivers/i2c/busses/i2c-qup.c qup->blk.data_len = msg->len; blk 505 drivers/i2c/busses/i2c-qup.c qup->blk.count = DIV_ROUND_UP(msg->len, qup->blk_xfer_limit); blk 512 drivers/i2c/busses/i2c-qup.c if (qup->blk.data_len > qup->blk_xfer_limit) blk 515 drivers/i2c/busses/i2c-qup.c data_len = qup->blk.data_len; blk 554 drivers/i2c/busses/i2c-qup.c int last = (qup->blk.pos == (qup->blk.count - 1)) && (qup->is_last); blk 560 drivers/i2c/busses/i2c-qup.c if (qup->blk.pos == 0) { blk 576 drivers/i2c/busses/i2c-qup.c tags[len++] = qup->blk.pos == (qup->blk.count - 1) ? blk 663 drivers/i2c/busses/i2c-qup.c blocks = qup->blk.count; blk 667 drivers/i2c/busses/i2c-qup.c while (qup->blk.pos < blocks) { blk 671 drivers/i2c/busses/i2c-qup.c qup->blk.data_len -= tlen; blk 689 drivers/i2c/busses/i2c-qup.c qup->blk.pos = i; blk 699 drivers/i2c/busses/i2c-qup.c while (qup->blk.pos < blocks) { blk 703 drivers/i2c/busses/i2c-qup.c qup->blk.data_len -= tlen; blk 718 drivers/i2c/busses/i2c-qup.c qup->blk.pos = i; blk 923 drivers/i2c/busses/i2c-qup.c struct qup_i2c_block *blk = &qup->blk; blk 928 drivers/i2c/busses/i2c-qup.c while (blk->fifo_available && qup->pos < msg->len) { blk 937 drivers/i2c/busses/i2c-qup.c blk->fifo_available--; blk 941 drivers/i2c/busses/i2c-qup.c blk->rx_bytes_read = true; blk 960 drivers/i2c/busses/i2c-qup.c struct qup_i2c_block *blk = &qup->blk; blk 964 drivers/i2c/busses/i2c-qup.c blk->is_tx_blk_mode = blk 965 drivers/i2c/busses/i2c-qup.c blk->total_tx_len > qup->out_fifo_sz ? true : false; blk 966 drivers/i2c/busses/i2c-qup.c blk->is_rx_blk_mode = blk 967 drivers/i2c/busses/i2c-qup.c blk->total_rx_len > qup->in_fifo_sz ? true : false; blk 969 drivers/i2c/busses/i2c-qup.c if (blk->is_tx_blk_mode) { blk 972 drivers/i2c/busses/i2c-qup.c writel(blk->total_tx_len, qup->base + QUP_MX_OUTPUT_CNT); blk 975 drivers/i2c/busses/i2c-qup.c writel(blk->total_tx_len, qup->base + QUP_MX_WRITE_CNT); blk 978 drivers/i2c/busses/i2c-qup.c if (blk->total_rx_len) { blk 979 drivers/i2c/busses/i2c-qup.c if (blk->is_rx_blk_mode) { blk 982 drivers/i2c/busses/i2c-qup.c writel(blk->total_rx_len, qup->base + QUP_MX_INPUT_CNT); blk 985 drivers/i2c/busses/i2c-qup.c writel(blk->total_rx_len, qup->base + QUP_MX_READ_CNT); blk 995 drivers/i2c/busses/i2c-qup.c static void qup_i2c_clear_blk_v1(struct qup_i2c_block *blk) blk 997 drivers/i2c/busses/i2c-qup.c blk->tx_fifo_free = 0; blk 998 drivers/i2c/busses/i2c-qup.c blk->fifo_available = 0; blk 999 drivers/i2c/busses/i2c-qup.c blk->rx_bytes_read = false; blk 1004 drivers/i2c/busses/i2c-qup.c struct qup_i2c_block *blk = &qup->blk; blk 1007 drivers/i2c/busses/i2c-qup.c qup_i2c_clear_blk_v1(blk); blk 1021 drivers/i2c/busses/i2c-qup.c if (!blk->is_tx_blk_mode) { blk 1022 drivers/i2c/busses/i2c-qup.c blk->tx_fifo_free = qup->out_fifo_sz; blk 1048 drivers/i2c/busses/i2c-qup.c struct qup_i2c_block *blk = &qup->blk; blk 1051 drivers/i2c/busses/i2c-qup.c blk->total_tx_len = msg->len + 1; blk 1052 drivers/i2c/busses/i2c-qup.c blk->total_rx_len = 0; blk 1059 drivers/i2c/busses/i2c-qup.c struct qup_i2c_block *blk = &qup->blk; blk 1062 drivers/i2c/busses/i2c-qup.c blk->total_tx_len = 2; blk 1063 drivers/i2c/busses/i2c-qup.c blk->total_rx_len = qup->msg->len; blk 1131 drivers/i2c/busses/i2c-qup.c struct qup_i2c_block *blk = &qup->blk; blk 1134 drivers/i2c/busses/i2c-qup.c if (blk->is_tx_blk_mode) blk 1135 drivers/i2c/busses/i2c-qup.c writel(qup->config_run | blk->total_tx_len, blk 1138 drivers/i2c/busses/i2c-qup.c writel(qup->config_run | blk->total_tx_len, blk 1141 drivers/i2c/busses/i2c-qup.c if (blk->total_rx_len) { blk 1142 drivers/i2c/busses/i2c-qup.c if (blk->is_rx_blk_mode) blk 1143 drivers/i2c/busses/i2c-qup.c writel(qup->config_run | blk->total_rx_len, blk 1146 drivers/i2c/busses/i2c-qup.c writel(qup->config_run | blk->total_rx_len, blk 1162 drivers/i2c/busses/i2c-qup.c struct qup_i2c_block *blk = &qup->blk; blk 1165 drivers/i2c/busses/i2c-qup.c if (blk->is_tx_blk_mode) { blk 1172 drivers/i2c/busses/i2c-qup.c if (blk->is_rx_blk_mode) { blk 1183 drivers/i2c/busses/i2c-qup.c static void qup_i2c_clear_blk_v2(struct qup_i2c_block *blk) blk 1185 drivers/i2c/busses/i2c-qup.c blk->send_last_word = false; blk 1186 drivers/i2c/busses/i2c-qup.c blk->tx_tags_sent = false; blk 1187 drivers/i2c/busses/i2c-qup.c blk->tx_fifo_data = 0; blk 1188 drivers/i2c/busses/i2c-qup.c blk->tx_fifo_data_pos = 0; blk 1189 drivers/i2c/busses/i2c-qup.c blk->tx_fifo_free = 0; blk 1191 drivers/i2c/busses/i2c-qup.c blk->rx_tags_fetched = false; blk 1192 drivers/i2c/busses/i2c-qup.c blk->rx_bytes_read = false; blk 1193 drivers/i2c/busses/i2c-qup.c blk->rx_fifo_data = 0; blk 1194 drivers/i2c/busses/i2c-qup.c blk->rx_fifo_data_pos = 0; blk 1195 drivers/i2c/busses/i2c-qup.c blk->fifo_available = 0; blk 1201 drivers/i2c/busses/i2c-qup.c struct qup_i2c_block *blk = &qup->blk; blk 1204 drivers/i2c/busses/i2c-qup.c for (j = blk->rx_fifo_data_pos; blk 1205 drivers/i2c/busses/i2c-qup.c blk->cur_blk_len && blk->fifo_available; blk 1206 drivers/i2c/busses/i2c-qup.c blk->cur_blk_len--, blk->fifo_available--) { blk 1208 drivers/i2c/busses/i2c-qup.c blk->rx_fifo_data = readl(qup->base + QUP_IN_FIFO_BASE); blk 1210 drivers/i2c/busses/i2c-qup.c *(blk->cur_data++) = blk->rx_fifo_data; blk 1211 drivers/i2c/busses/i2c-qup.c blk->rx_fifo_data >>= 8; blk 1219 drivers/i2c/busses/i2c-qup.c blk->rx_fifo_data_pos = j; blk 1225 drivers/i2c/busses/i2c-qup.c struct qup_i2c_block *blk = &qup->blk; blk 1227 drivers/i2c/busses/i2c-qup.c blk->rx_fifo_data = readl(qup->base + QUP_IN_FIFO_BASE); blk 1228 drivers/i2c/busses/i2c-qup.c blk->rx_fifo_data >>= blk->rx_tag_len * 8; blk 1229 drivers/i2c/busses/i2c-qup.c blk->rx_fifo_data_pos = blk->rx_tag_len; blk 1230 drivers/i2c/busses/i2c-qup.c blk->fifo_available -= blk->rx_tag_len; blk 1243 drivers/i2c/busses/i2c-qup.c struct qup_i2c_block *blk = &qup->blk; blk 1245 drivers/i2c/busses/i2c-qup.c if (!blk->rx_tags_fetched) { blk 1247 drivers/i2c/busses/i2c-qup.c blk->rx_tags_fetched = true; blk 1251 drivers/i2c/busses/i2c-qup.c if (!blk->cur_blk_len) blk 1252 drivers/i2c/busses/i2c-qup.c blk->rx_bytes_read = true; blk 1263 drivers/i2c/busses/i2c-qup.c struct qup_i2c_block *blk = &qup->blk; blk 1266 drivers/i2c/busses/i2c-qup.c for (j = blk->tx_fifo_data_pos; *len && blk->tx_fifo_free; blk 1267 drivers/i2c/busses/i2c-qup.c (*len)--, blk->tx_fifo_free--) { blk 1268 drivers/i2c/busses/i2c-qup.c blk->tx_fifo_data |= *(*data)++ << (j * 8); blk 1270 drivers/i2c/busses/i2c-qup.c writel(blk->tx_fifo_data, blk 1272 drivers/i2c/busses/i2c-qup.c blk->tx_fifo_data = 0x0; blk 1279 drivers/i2c/busses/i2c-qup.c blk->tx_fifo_data_pos = j; blk 1285 drivers/i2c/busses/i2c-qup.c struct qup_i2c_block *blk = &qup->blk; blk 1287 drivers/i2c/busses/i2c-qup.c qup_i2c_write_blk_data(qup, &blk->cur_tx_tags, &blk->tx_tag_len); blk 1288 drivers/i2c/busses/i2c-qup.c if (blk->tx_fifo_data_pos) blk 1289 drivers/i2c/busses/i2c-qup.c writel(blk->tx_fifo_data, qup->base + QUP_OUT_FIFO_BASE); blk 1316 drivers/i2c/busses/i2c-qup.c struct qup_i2c_block *blk = &qup->blk; blk 1318 drivers/i2c/busses/i2c-qup.c if (!blk->tx_tags_sent) { blk 1319 drivers/i2c/busses/i2c-qup.c qup_i2c_write_blk_data(qup, &blk->cur_tx_tags, blk 1320 drivers/i2c/busses/i2c-qup.c &blk->tx_tag_len); blk 1321 drivers/i2c/busses/i2c-qup.c blk->tx_tags_sent = true; blk 1324 drivers/i2c/busses/i2c-qup.c if (blk->send_last_word) blk 1327 drivers/i2c/busses/i2c-qup.c qup_i2c_write_blk_data(qup, &blk->cur_data, &blk->cur_blk_len); blk 1328 drivers/i2c/busses/i2c-qup.c if (!blk->cur_blk_len) { blk 1329 drivers/i2c/busses/i2c-qup.c if (!blk->tx_fifo_data_pos) blk 1332 drivers/i2c/busses/i2c-qup.c if (blk->tx_fifo_free) blk 1335 drivers/i2c/busses/i2c-qup.c blk->send_last_word = true; blk 1341 drivers/i2c/busses/i2c-qup.c writel(blk->tx_fifo_data, qup->base + QUP_OUT_FIFO_BASE); blk 1353 drivers/i2c/busses/i2c-qup.c struct qup_i2c_block *blk = &qup->blk; blk 1369 drivers/i2c/busses/i2c-qup.c blk->cur_data += 1; blk 1378 drivers/i2c/busses/i2c-qup.c qup_i2c_clear_blk_v2(blk); blk 1400 drivers/i2c/busses/i2c-qup.c if (!blk->is_tx_blk_mode) { blk 1401 drivers/i2c/busses/i2c-qup.c blk->tx_fifo_free = qup->out_fifo_sz; blk 1439 drivers/i2c/busses/i2c-qup.c struct qup_i2c_block *blk = &qup->blk; blk 1445 drivers/i2c/busses/i2c-qup.c for (i = 0; i < blk->count; i++) { blk 1447 drivers/i2c/busses/i2c-qup.c blk->pos = i; blk 1448 drivers/i2c/busses/i2c-qup.c blk->cur_tx_tags = blk->tags; blk 1449 drivers/i2c/busses/i2c-qup.c blk->cur_blk_len = data_len; blk 1450 drivers/i2c/busses/i2c-qup.c blk->tx_tag_len = blk 1451 drivers/i2c/busses/i2c-qup.c qup_i2c_set_tags(blk->cur_tx_tags, qup, qup->msg); blk 1453 drivers/i2c/busses/i2c-qup.c blk->cur_data = msg_buf; blk 1456 drivers/i2c/busses/i2c-qup.c blk->total_tx_len = blk->tx_tag_len; blk 1457 drivers/i2c/busses/i2c-qup.c blk->rx_tag_len = 2; blk 1458 drivers/i2c/busses/i2c-qup.c blk->total_rx_len = blk->rx_tag_len + data_len; blk 1460 drivers/i2c/busses/i2c-qup.c blk->total_tx_len = blk->tx_tag_len + data_len; blk 1461 drivers/i2c/busses/i2c-qup.c blk->total_rx_len = 0; blk 1465 drivers/i2c/busses/i2c-qup.c !qup->is_last || i < blk->count - 1); blk 1486 drivers/i2c/busses/i2c-qup.c blk->data_len -= qup->blk_xfer_limit; blk 1535 drivers/i2c/busses/i2c-qup.c qup->blk.is_tx_blk_mode = max_tx_len > qup->out_fifo_sz - blk 1537 drivers/i2c/busses/i2c-qup.c qup->blk.is_rx_blk_mode = max_rx_len > qup->in_fifo_sz - blk 966 drivers/lightnvm/core.c int ret, blk, pl, offset, blktype; blk 968 drivers/lightnvm/core.c for (blk = 0; blk < geo->num_chk; blk++) { blk 969 drivers/lightnvm/core.c offset = blk * geo->pln_mode; blk 980 drivers/lightnvm/core.c ppa.g.blk = blk; blk 1014 drivers/lightnvm/core.c if (ppa.g.blk != 0) blk 902 drivers/lightnvm/pblk-core.c ppa.a.blk = line->id; blk 164 drivers/lightnvm/pblk-map.c erase_ppa->a.blk = e_line->id; blk 206 drivers/lightnvm/pblk-map.c erase_ppa->a.blk = e_line->id; blk 962 drivers/lightnvm/pblk.h return p.a.blk; blk 987 drivers/lightnvm/pblk.h ppa.g.blk = line_id; blk 1203 drivers/lightnvm/pblk.h p->g.ch, p->g.lun, p->g.blk, blk 1245 drivers/lightnvm/pblk.h ppa->g.blk < geo->num_chk && blk 35 drivers/md/dm-dust.c static struct badblock *dust_rb_search(struct rb_root *root, sector_t blk) blk 42 drivers/md/dm-dust.c if (bblk->bb > blk) blk 44 drivers/md/dm-dust.c else if (bblk->bb < blk) blk 282 drivers/md/persistent-data/dm-space-map-common.c struct dm_block *blk; blk 290 drivers/md/persistent-data/dm-space-map-common.c &dm_sm_bitmap_validator, &blk); blk 294 drivers/md/persistent-data/dm-space-map-common.c *result = sm_lookup_bitmap(dm_bitmap_data(blk), b); blk 296 drivers/md/persistent-data/dm-space-map-common.c dm_tm_unlock(ll->tm, blk); blk 344 drivers/md/persistent-data/dm-space-map-common.c struct dm_block *blk; blk 356 drivers/md/persistent-data/dm-space-map-common.c &dm_sm_bitmap_validator, &blk); blk 362 drivers/md/persistent-data/dm-space-map-common.c r = sm_find_free(dm_bitmap_data(blk), blk 370 drivers/md/persistent-data/dm-space-map-common.c dm_tm_unlock(ll->tm, blk); blk 374 drivers/md/persistent-data/dm-space-map-common.c dm_tm_unlock(ll->tm, blk); blk 330 drivers/md/persistent-data/dm-transaction-manager.c struct dm_block **blk) blk 333 drivers/md/persistent-data/dm-transaction-manager.c int r = dm_bm_read_try_lock(tm->real->bm, b, v, blk); blk 341 drivers/md/persistent-data/dm-transaction-manager.c return dm_bm_read_lock(tm->bm, b, v, blk); blk 77 drivers/media/platform/mtk-jpeg/mtk_jpeg_hw.c u32 i, comp, blk; blk 96 drivers/media/platform/mtk-jpeg/mtk_jpeg_hw.c for (i = 0, blk = 0, comp = 0; i < MTK_JPEG_BLOCK_MAX; i++) { blk 102 drivers/media/platform/mtk-jpeg/mtk_jpeg_hw.c if (++blk == param->blk_comp[comp]) { blk 104 drivers/media/platform/mtk-jpeg/mtk_jpeg_hw.c blk = 0; blk 31 drivers/media/platform/omap3isp/isph3a_aewb.c u32 blk; blk 55 drivers/media/platform/omap3isp/isph3a_aewb.c blk = conf->blk_ver_win_start << ISPH3A_AEWINBLK_WINSV_SHIFT; blk 56 drivers/media/platform/omap3isp/isph3a_aewb.c blk |= ((conf->blk_win_height >> 1) - 1) << ISPH3A_AEWINBLK_WINH_SHIFT; blk 66 drivers/media/platform/omap3isp/isph3a_aewb.c isp_reg_writel(aewb->isp, blk, OMAP3_ISP_IOMEM_H3A, ISPH3A_AEWINBLK); blk 629 drivers/media/platform/ti-vpe/vpdma.c struct vpdma_buf *blk, u32 dest_offset) blk 632 drivers/media/platform/ti-vpe/vpdma.c int len = blk->size; blk 634 drivers/media/platform/ti-vpe/vpdma.c WARN_ON(blk->dma_addr & VPDMA_DESC_ALIGN); blk 641 drivers/media/platform/ti-vpe/vpdma.c cfd->payload_addr = (u32) blk->dma_addr; blk 235 drivers/media/platform/ti-vpe/vpdma.h struct vpdma_buf *blk, u32 dest_offset); blk 32 drivers/media/platform/vivid/vivid-radio-rx.c unsigned blk; blk 56 drivers/media/platform/vivid/vivid-radio-rx.c blk = ktime_divns(timestamp, VIVID_RDS_NSEC_PER_BLK); blk 57 drivers/media/platform/vivid/vivid-radio-rx.c use_alternates = (blk % VIVID_RDS_GEN_BLOCKS) & 1; blk 65 drivers/media/platform/vivid/vivid-radio-rx.c if (blk >= dev->radio_rx_rds_last_block + VIVID_RDS_GEN_BLOCKS) blk 66 drivers/media/platform/vivid/vivid-radio-rx.c dev->radio_rx_rds_last_block = blk - VIVID_RDS_GEN_BLOCKS + 1; blk 74 drivers/media/platform/vivid/vivid-radio-rx.c if (blk == dev->radio_rx_rds_last_block || !dev->radio_rx_rds_enabled || blk 90 drivers/media/platform/vivid/vivid-radio-rx.c for (i = 0; i < size && blk > dev->radio_rx_rds_last_block; blk 29 drivers/media/platform/vivid/vivid-radio-tx.c unsigned blk; blk 50 drivers/media/platform/vivid/vivid-radio-tx.c blk = ktime_divns(timestamp, VIVID_RDS_NSEC_PER_BLK); blk 51 drivers/media/platform/vivid/vivid-radio-tx.c if (blk - VIVID_RDS_GEN_BLOCKS >= dev->radio_tx_rds_last_block) blk 52 drivers/media/platform/vivid/vivid-radio-tx.c dev->radio_tx_rds_last_block = blk - VIVID_RDS_GEN_BLOCKS + 1; blk 60 drivers/media/platform/vivid/vivid-radio-tx.c if (blk == dev->radio_tx_rds_last_block || blk 72 drivers/media/platform/vivid/vivid-radio-tx.c for (i = 0; i < size && blk > dev->radio_tx_rds_last_block; blk 690 drivers/mtd/ftl.c uint32_t blk; blk 725 drivers/mtd/ftl.c for (blk = 0; blk < part->BlocksPerUnit; blk++) blk 726 drivers/mtd/ftl.c if (BLOCK_FREE(le32_to_cpu(part->bam_cache[blk]))) break; blk 727 drivers/mtd/ftl.c if (blk == part->BlocksPerUnit) { blk 736 drivers/mtd/ftl.c pr_debug("ftl_cs: found free block at %d in %d\n", blk, eun); blk 737 drivers/mtd/ftl.c return blk; blk 797 drivers/mtd/ftl.c uint32_t bsize, blk, le_virt_addr; blk 809 drivers/mtd/ftl.c blk = (log_addr % bsize) / SECTOR_SIZE; blk 810 drivers/mtd/ftl.c offset = (part->EUNInfo[eun].Offset + blk * sizeof(uint32_t) + blk 837 drivers/mtd/ftl.c if (le32_to_cpu(part->bam_cache[blk]) != old_addr) { blk 844 drivers/mtd/ftl.c le32_to_cpu(part->bam_cache[blk]), old_addr); blk 849 drivers/mtd/ftl.c part->bam_cache[blk] = le_virt_addr; blk 865 drivers/mtd/ftl.c uint32_t bsize, log_addr, virt_addr, old_addr, blk; blk 893 drivers/mtd/ftl.c blk = find_free(part); blk 894 drivers/mtd/ftl.c if (blk == 0) { blk 903 drivers/mtd/ftl.c log_addr = part->bam_index * bsize + blk * SECTOR_SIZE; blk 910 drivers/mtd/ftl.c blk * SECTOR_SIZE); blk 303 drivers/mtd/nand/onenand/onenand_base.c unsigned boundary, blk, die = 0; blk 312 drivers/mtd/nand/onenand/onenand_base.c blk = addr >> (this->erase_shift - 1); blk 313 drivers/mtd/nand/onenand/onenand_base.c if (blk > boundary) blk 314 drivers/mtd/nand/onenand/onenand_base.c blk = (blk + boundary + 1) >> 1; blk 316 drivers/mtd/nand/onenand/onenand_base.c blk += die ? this->density_mask : 0; blk 317 drivers/mtd/nand/onenand/onenand_base.c return blk; blk 4576 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c static void _print_next_block(int idx, const char *blk) blk 4578 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c pr_cont("%s%s", idx ? ", " : "", blk); blk 1660 drivers/net/ethernet/broadcom/cnic.c int blk = (cid - ethdev->starting_cid) / cp->cids_per_blk; blk 1669 drivers/net/ethernet/broadcom/cnic.c if (cp->ctx_arr[blk].mapping & mask) blk 1671 drivers/net/ethernet/broadcom/cnic.c (cp->ctx_arr[blk].mapping & mask); blk 1673 drivers/net/ethernet/broadcom/cnic.c ctx_map = cp->ctx_arr[blk].mapping + align_off + blk 1675 drivers/net/ethernet/broadcom/cnic.c ctx = cp->ctx_arr[blk].ctx + align_off + blk 14 drivers/net/ethernet/chelsio/cxgb/vsc7326_reg.h #define CRA(blk,sub,adr) ((((blk) & 0x7) << 13) | (((sub) & 0xf) << 9) | (((adr) & 0xff) << 1)) blk 784 drivers/net/ethernet/intel/ice/ice_flex_pipe.c ice_ptg_find_ptype(struct ice_hw *hw, enum ice_block blk, u16 ptype, u8 *ptg) blk 789 drivers/net/ethernet/intel/ice/ice_flex_pipe.c *ptg = hw->blk[blk].xlt1.ptypes[ptype].ptg; blk 802 drivers/net/ethernet/intel/ice/ice_flex_pipe.c static void ice_ptg_alloc_val(struct ice_hw *hw, enum ice_block blk, u8 ptg) blk 804 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[blk].xlt1.ptg_tbl[ptg].in_use = true; blk 818 drivers/net/ethernet/intel/ice/ice_flex_pipe.c ice_ptg_remove_ptype(struct ice_hw *hw, enum ice_block blk, u16 ptype, u8 ptg) blk 826 drivers/net/ethernet/intel/ice/ice_flex_pipe.c if (!hw->blk[blk].xlt1.ptg_tbl[ptg].in_use) blk 830 drivers/net/ethernet/intel/ice/ice_flex_pipe.c if (!hw->blk[blk].xlt1.ptg_tbl[ptg].first_ptype) blk 834 drivers/net/ethernet/intel/ice/ice_flex_pipe.c p = hw->blk[blk].xlt1.ptg_tbl[ptg].first_ptype; blk 835 drivers/net/ethernet/intel/ice/ice_flex_pipe.c ch = &hw->blk[blk].xlt1.ptg_tbl[ptg].first_ptype; blk 837 drivers/net/ethernet/intel/ice/ice_flex_pipe.c if (ptype == (p - hw->blk[blk].xlt1.ptypes)) { blk 846 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[blk].xlt1.ptypes[ptype].ptg = ICE_DEFAULT_PTG; blk 847 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[blk].xlt1.ptypes[ptype].next_ptype = NULL; blk 865 drivers/net/ethernet/intel/ice/ice_flex_pipe.c ice_ptg_add_mv_ptype(struct ice_hw *hw, enum ice_block blk, u16 ptype, u8 ptg) blk 873 drivers/net/ethernet/intel/ice/ice_flex_pipe.c if (!hw->blk[blk].xlt1.ptg_tbl[ptg].in_use && ptg != ICE_DEFAULT_PTG) blk 876 drivers/net/ethernet/intel/ice/ice_flex_pipe.c status = ice_ptg_find_ptype(hw, blk, ptype, &original_ptg); blk 886 drivers/net/ethernet/intel/ice/ice_flex_pipe.c ice_ptg_remove_ptype(hw, blk, ptype, original_ptg); blk 893 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[blk].xlt1.ptypes[ptype].next_ptype = blk 894 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[blk].xlt1.ptg_tbl[ptg].first_ptype; blk 895 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[blk].xlt1.ptg_tbl[ptg].first_ptype = blk 896 drivers/net/ethernet/intel/ice/ice_flex_pipe.c &hw->blk[blk].xlt1.ptypes[ptype]; blk 898 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[blk].xlt1.ptypes[ptype].ptg = ptg; blk 899 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[blk].xlt1.t[ptype] = ptg; blk 967 drivers/net/ethernet/intel/ice/ice_flex_pipe.c ice_vsig_find_vsi(struct ice_hw *hw, enum ice_block blk, u16 vsi, u16 *vsig) blk 976 drivers/net/ethernet/intel/ice/ice_flex_pipe.c *vsig = hw->blk[blk].xlt2.vsis[vsi].vsig; blk 989 drivers/net/ethernet/intel/ice/ice_flex_pipe.c static u16 ice_vsig_alloc_val(struct ice_hw *hw, enum ice_block blk, u16 vsig) blk 993 drivers/net/ethernet/intel/ice/ice_flex_pipe.c if (!hw->blk[blk].xlt2.vsig_tbl[idx].in_use) { blk 994 drivers/net/ethernet/intel/ice/ice_flex_pipe.c INIT_LIST_HEAD(&hw->blk[blk].xlt2.vsig_tbl[idx].prop_lst); blk 995 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[blk].xlt2.vsig_tbl[idx].in_use = true; blk 1012 drivers/net/ethernet/intel/ice/ice_flex_pipe.c ice_vsig_remove_vsi(struct ice_hw *hw, enum ice_block blk, u16 vsi, u16 vsig) blk 1022 drivers/net/ethernet/intel/ice/ice_flex_pipe.c if (!hw->blk[blk].xlt2.vsig_tbl[idx].in_use) blk 1029 drivers/net/ethernet/intel/ice/ice_flex_pipe.c vsi_head = &hw->blk[blk].xlt2.vsig_tbl[idx].first_vsi; blk 1033 drivers/net/ethernet/intel/ice/ice_flex_pipe.c vsi_tgt = &hw->blk[blk].xlt2.vsis[vsi]; blk 1070 drivers/net/ethernet/intel/ice/ice_flex_pipe.c ice_vsig_add_mv_vsi(struct ice_hw *hw, enum ice_block blk, u16 vsi, u16 vsig) blk 1084 drivers/net/ethernet/intel/ice/ice_flex_pipe.c if (!hw->blk[blk].xlt2.vsig_tbl[idx].in_use && blk 1088 drivers/net/ethernet/intel/ice/ice_flex_pipe.c status = ice_vsig_find_vsi(hw, blk, vsi, &orig_vsig); blk 1098 drivers/net/ethernet/intel/ice/ice_flex_pipe.c status = ice_vsig_remove_vsi(hw, blk, vsi, orig_vsig); blk 1107 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[blk].xlt2.vsis[vsi].vsig = vsig; blk 1108 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[blk].xlt2.vsis[vsi].changed = 1; blk 1111 drivers/net/ethernet/intel/ice/ice_flex_pipe.c tmp = hw->blk[blk].xlt2.vsig_tbl[idx].first_vsi; blk 1112 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[blk].xlt2.vsig_tbl[idx].first_vsi = blk 1113 drivers/net/ethernet/intel/ice/ice_flex_pipe.c &hw->blk[blk].xlt2.vsis[vsi]; blk 1114 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[blk].xlt2.vsis[vsi].next_vsi = tmp; blk 1115 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[blk].xlt2.t[vsi] = vsig; blk 1168 drivers/net/ethernet/intel/ice/ice_flex_pipe.c static void ice_init_sw_xlt1_db(struct ice_hw *hw, enum ice_block blk) blk 1172 drivers/net/ethernet/intel/ice/ice_flex_pipe.c for (pt = 0; pt < hw->blk[blk].xlt1.count; pt++) { blk 1175 drivers/net/ethernet/intel/ice/ice_flex_pipe.c ptg = hw->blk[blk].xlt1.t[pt]; blk 1177 drivers/net/ethernet/intel/ice/ice_flex_pipe.c ice_ptg_alloc_val(hw, blk, ptg); blk 1178 drivers/net/ethernet/intel/ice/ice_flex_pipe.c ice_ptg_add_mv_ptype(hw, blk, pt, ptg); blk 1188 drivers/net/ethernet/intel/ice/ice_flex_pipe.c static void ice_init_sw_xlt2_db(struct ice_hw *hw, enum ice_block blk) blk 1192 drivers/net/ethernet/intel/ice/ice_flex_pipe.c for (vsi = 0; vsi < hw->blk[blk].xlt2.count; vsi++) { blk 1195 drivers/net/ethernet/intel/ice/ice_flex_pipe.c vsig = hw->blk[blk].xlt2.t[vsi]; blk 1197 drivers/net/ethernet/intel/ice/ice_flex_pipe.c ice_vsig_alloc_val(hw, blk, vsig); blk 1198 drivers/net/ethernet/intel/ice/ice_flex_pipe.c ice_vsig_add_mv_vsi(hw, blk, vsi, vsig); blk 1202 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[blk].xlt2.vsis[vsi].changed = 0; blk 1268 drivers/net/ethernet/intel/ice/ice_flex_pipe.c sizeof(*hw->blk[block_id].xlt1.t); blk 1269 drivers/net/ethernet/intel/ice/ice_flex_pipe.c dst = hw->blk[block_id].xlt1.t; blk 1270 drivers/net/ethernet/intel/ice/ice_flex_pipe.c dst_len = hw->blk[block_id].xlt1.count * blk 1271 drivers/net/ethernet/intel/ice/ice_flex_pipe.c sizeof(*hw->blk[block_id].xlt1.t); blk 1281 drivers/net/ethernet/intel/ice/ice_flex_pipe.c sizeof(*hw->blk[block_id].xlt2.t); blk 1282 drivers/net/ethernet/intel/ice/ice_flex_pipe.c dst = (u8 *)hw->blk[block_id].xlt2.t; blk 1283 drivers/net/ethernet/intel/ice/ice_flex_pipe.c dst_len = hw->blk[block_id].xlt2.count * blk 1284 drivers/net/ethernet/intel/ice/ice_flex_pipe.c sizeof(*hw->blk[block_id].xlt2.t); blk 1294 drivers/net/ethernet/intel/ice/ice_flex_pipe.c sizeof(*hw->blk[block_id].prof.t); blk 1295 drivers/net/ethernet/intel/ice/ice_flex_pipe.c dst = (u8 *)hw->blk[block_id].prof.t; blk 1296 drivers/net/ethernet/intel/ice/ice_flex_pipe.c dst_len = hw->blk[block_id].prof.count * blk 1297 drivers/net/ethernet/intel/ice/ice_flex_pipe.c sizeof(*hw->blk[block_id].prof.t); blk 1307 drivers/net/ethernet/intel/ice/ice_flex_pipe.c sizeof(*hw->blk[block_id].prof_redir.t); blk 1308 drivers/net/ethernet/intel/ice/ice_flex_pipe.c dst = hw->blk[block_id].prof_redir.t; blk 1309 drivers/net/ethernet/intel/ice/ice_flex_pipe.c dst_len = hw->blk[block_id].prof_redir.count * blk 1310 drivers/net/ethernet/intel/ice/ice_flex_pipe.c sizeof(*hw->blk[block_id].prof_redir.t); blk 1320 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[block_id].es.fvw) * blk 1321 drivers/net/ethernet/intel/ice/ice_flex_pipe.c sizeof(*hw->blk[block_id].es.t); blk 1322 drivers/net/ethernet/intel/ice/ice_flex_pipe.c dst = (u8 *)hw->blk[block_id].es.t; blk 1323 drivers/net/ethernet/intel/ice/ice_flex_pipe.c dst_len = (u32)(hw->blk[block_id].es.count * blk 1324 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[block_id].es.fvw) * blk 1325 drivers/net/ethernet/intel/ice/ice_flex_pipe.c sizeof(*hw->blk[block_id].es.t); blk 1366 drivers/net/ethernet/intel/ice/ice_flex_pipe.c ice_fill_tbl(hw, blk_id, hw->blk[blk_id].xlt1.sid); blk 1367 drivers/net/ethernet/intel/ice/ice_flex_pipe.c ice_fill_tbl(hw, blk_id, hw->blk[blk_id].xlt2.sid); blk 1368 drivers/net/ethernet/intel/ice/ice_flex_pipe.c ice_fill_tbl(hw, blk_id, hw->blk[blk_id].prof.sid); blk 1369 drivers/net/ethernet/intel/ice/ice_flex_pipe.c ice_fill_tbl(hw, blk_id, hw->blk[blk_id].prof_redir.sid); blk 1370 drivers/net/ethernet/intel/ice/ice_flex_pipe.c ice_fill_tbl(hw, blk_id, hw->blk[blk_id].es.sid); blk 1385 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[i].is_list_init = false; blk 1387 drivers/net/ethernet/intel/ice/ice_flex_pipe.c devm_kfree(ice_hw_to_dev(hw), hw->blk[i].xlt1.ptypes); blk 1388 drivers/net/ethernet/intel/ice/ice_flex_pipe.c devm_kfree(ice_hw_to_dev(hw), hw->blk[i].xlt1.ptg_tbl); blk 1389 drivers/net/ethernet/intel/ice/ice_flex_pipe.c devm_kfree(ice_hw_to_dev(hw), hw->blk[i].xlt1.t); blk 1390 drivers/net/ethernet/intel/ice/ice_flex_pipe.c devm_kfree(ice_hw_to_dev(hw), hw->blk[i].xlt2.t); blk 1391 drivers/net/ethernet/intel/ice/ice_flex_pipe.c devm_kfree(ice_hw_to_dev(hw), hw->blk[i].xlt2.vsig_tbl); blk 1392 drivers/net/ethernet/intel/ice/ice_flex_pipe.c devm_kfree(ice_hw_to_dev(hw), hw->blk[i].xlt2.vsis); blk 1393 drivers/net/ethernet/intel/ice/ice_flex_pipe.c devm_kfree(ice_hw_to_dev(hw), hw->blk[i].prof.t); blk 1394 drivers/net/ethernet/intel/ice/ice_flex_pipe.c devm_kfree(ice_hw_to_dev(hw), hw->blk[i].prof_redir.t); blk 1395 drivers/net/ethernet/intel/ice/ice_flex_pipe.c devm_kfree(ice_hw_to_dev(hw), hw->blk[i].es.t); blk 1396 drivers/net/ethernet/intel/ice/ice_flex_pipe.c devm_kfree(ice_hw_to_dev(hw), hw->blk[i].es.ref_count); blk 1397 drivers/net/ethernet/intel/ice/ice_flex_pipe.c devm_kfree(ice_hw_to_dev(hw), hw->blk[i].es.written); blk 1400 drivers/net/ethernet/intel/ice/ice_flex_pipe.c memset(hw->blk, 0, sizeof(hw->blk)); blk 1412 drivers/net/ethernet/intel/ice/ice_flex_pipe.c struct ice_prof_redir *prof_redir = &hw->blk[i].prof_redir; blk 1413 drivers/net/ethernet/intel/ice/ice_flex_pipe.c struct ice_prof_tcam *prof = &hw->blk[i].prof; blk 1414 drivers/net/ethernet/intel/ice/ice_flex_pipe.c struct ice_xlt1 *xlt1 = &hw->blk[i].xlt1; blk 1415 drivers/net/ethernet/intel/ice/ice_flex_pipe.c struct ice_xlt2 *xlt2 = &hw->blk[i].xlt2; blk 1416 drivers/net/ethernet/intel/ice/ice_flex_pipe.c struct ice_es *es = &hw->blk[i].es; blk 1447 drivers/net/ethernet/intel/ice/ice_flex_pipe.c struct ice_prof_redir *prof_redir = &hw->blk[i].prof_redir; blk 1448 drivers/net/ethernet/intel/ice/ice_flex_pipe.c struct ice_prof_tcam *prof = &hw->blk[i].prof; blk 1449 drivers/net/ethernet/intel/ice/ice_flex_pipe.c struct ice_xlt1 *xlt1 = &hw->blk[i].xlt1; blk 1450 drivers/net/ethernet/intel/ice/ice_flex_pipe.c struct ice_xlt2 *xlt2 = &hw->blk[i].xlt2; blk 1451 drivers/net/ethernet/intel/ice/ice_flex_pipe.c struct ice_es *es = &hw->blk[i].es; blk 1454 drivers/net/ethernet/intel/ice/ice_flex_pipe.c if (hw->blk[i].is_list_init) blk 1457 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[i].is_list_init = true; blk 1459 drivers/net/ethernet/intel/ice/ice_flex_pipe.c hw->blk[i].overwrite = blk_sizes[i].overwrite; blk 497 drivers/net/ethernet/intel/ice/ice_type.h struct ice_blk_info blk[ICE_BLK_COUNT]; blk 2651 drivers/net/ethernet/neterion/s2io.c static void free_rxd_blk(struct s2io_nic *sp, int ring_no, int blk) blk 2665 drivers/net/ethernet/neterion/s2io.c rx_blocks[blk].rxds[j].virt_addr; blk 2712 drivers/net/ethernet/neterion/s2io.c int i, blk = 0, buf_cnt = 0; blk 2719 drivers/net/ethernet/neterion/s2io.c for (blk = 0; blk < rx_ring_sz[i]; blk++) blk 2720 drivers/net/ethernet/neterion/s2io.c free_rxd_blk(sp, i, blk); blk 96 drivers/net/wireless/intersil/orinoco/hermes_dld.c dblock_addr(const struct dblock *blk) blk 98 drivers/net/wireless/intersil/orinoco/hermes_dld.c return le32_to_cpu(blk->addr); blk 102 drivers/net/wireless/intersil/orinoco/hermes_dld.c dblock_len(const struct dblock *blk) blk 104 drivers/net/wireless/intersil/orinoco/hermes_dld.c return le16_to_cpu(blk->len); blk 256 drivers/net/wireless/intersil/orinoco/hermes_dld.c const struct dblock *blk = (const struct dblock *) first_block; blk 260 drivers/net/wireless/intersil/orinoco/hermes_dld.c end -= sizeof(*blk); blk 264 drivers/net/wireless/intersil/orinoco/hermes_dld.c while (((void *) blk <= end) && blk 265 drivers/net/wireless/intersil/orinoco/hermes_dld.c (dblock_addr(blk) != BLOCK_END)) { blk 266 drivers/net/wireless/intersil/orinoco/hermes_dld.c len = dblock_len(blk); blk 267 drivers/net/wireless/intersil/orinoco/hermes_dld.c total_len += sizeof(*blk) + len; blk 268 drivers/net/wireless/intersil/orinoco/hermes_dld.c blk = (struct dblock *) &blk->data[len]; blk 279 drivers/net/wireless/intersil/orinoco/hermes_dld.c const struct dblock *blk; blk 284 drivers/net/wireless/intersil/orinoco/hermes_dld.c blk = (const struct dblock *) first_block; blk 286 drivers/net/wireless/intersil/orinoco/hermes_dld.c if ((void *) blk > (end - sizeof(*blk))) blk 289 drivers/net/wireless/intersil/orinoco/hermes_dld.c blkaddr = dblock_addr(blk); blk 290 drivers/net/wireless/intersil/orinoco/hermes_dld.c blklen = dblock_len(blk); blk 293 drivers/net/wireless/intersil/orinoco/hermes_dld.c (((void *) blk + blklen) <= end)) { blk 297 drivers/net/wireless/intersil/orinoco/hermes_dld.c err = hw->ops->program(hw, blk->data, blkaddr, blklen); blk 301 drivers/net/wireless/intersil/orinoco/hermes_dld.c blk = (const struct dblock *) &blk->data[blklen]; blk 303 drivers/net/wireless/intersil/orinoco/hermes_dld.c if ((void *) blk > (end - sizeof(*blk))) blk 306 drivers/net/wireless/intersil/orinoco/hermes_dld.c blkaddr = dblock_addr(blk); blk 307 drivers/net/wireless/intersil/orinoco/hermes_dld.c blklen = dblock_len(blk); blk 866 drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c int blk, const u8 *pblk, const u8 *fw) blk 888 drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c hdr->seqnum = cpu_to_le32(blk); blk 890 drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c if (blk) blk 919 drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c int blk = 0; blk 924 drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c while (blk < blk_count) { blk 930 drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c len = qtnf_ep_fw_send(ps->base.pdev, fw_size, blk, pblk, fw); blk 934 drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c if (!((blk + 1) & QTN_PCIE_FW_DLMASK) || blk 935 drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c (blk == (blk_count - 1))) { blk 950 drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c if (blk == (blk_count - 1)) { blk 953 drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c blk -= last_round; blk 957 drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c blk -= QTN_PCIE_FW_DLMASK; blk 964 drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c pr_warn("FW upload retry: block #%d\n", blk); blk 972 drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c blk++; blk 975 drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c pr_debug("FW upload completed: totally sent %d blocks\n", blk); blk 174 drivers/nvme/host/lightnvm.c __u8 blk[0]; blk 525 drivers/nvme/host/lightnvm.c memcpy(blks, bb_tbl->blk, geo->num_chk * geo->num_pln); blk 440 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pa-cp-blk-req, 0x05), blk 389 drivers/scsi/megaraid/megaraid_sas_fp.c u64 blk, debugBlk; blk 390 drivers/scsi/megaraid/megaraid_sas_fp.c blk = mega_div64_32((row-le64_to_cpu(quad->logStart)), le32_to_cpu(quad->diff)); blk 391 drivers/scsi/megaraid/megaraid_sas_fp.c debugBlk = blk; blk 393 drivers/scsi/megaraid/megaraid_sas_fp.c blk = (blk + le64_to_cpu(quad->offsetInSpan)) << raid->stripeShift; blk 394 drivers/scsi/megaraid/megaraid_sas_fp.c *span_blk = blk; blk 453 drivers/scsi/megaraid/megaraid_sas_fp.c u64 blk; blk 454 drivers/scsi/megaraid/megaraid_sas_fp.c blk = mega_div64_32 blk 457 drivers/scsi/megaraid/megaraid_sas_fp.c blk = (blk + le64_to_cpu(quad->offsetInSpan)) blk 459 drivers/scsi/megaraid/megaraid_sas_fp.c *span_blk = blk; blk 3169 drivers/scsi/st.c unsigned int blk; blk 3189 drivers/scsi/st.c if (get_location(STp, &blk, &p, 1)) blk 3193 drivers/scsi/st.c STps->last_block_visited = blk; blk 3196 drivers/scsi/st.c blk, STp->partition); blk 3507 drivers/scsi/st.c unsigned int blk; blk 3819 drivers/scsi/st.c if ((i = get_location(STp, &blk, &bt, 0)) < 0) { blk 3823 drivers/scsi/st.c mt_pos.mt_blkno = blk; blk 508 drivers/soc/ixp4xx/ixp4xx-npe.c } *blk; blk 612 drivers/soc/ixp4xx/ixp4xx-npe.c for (i = 0, blk = image->blocks; i < blocks; i++, blk++) { blk 613 drivers/soc/ixp4xx/ixp4xx-npe.c if (blk->offset > image->size - sizeof(struct dl_codeblock) / 4 blk 614 drivers/soc/ixp4xx/ixp4xx-npe.c || blk->offset < table_end) { blk 616 drivers/soc/ixp4xx/ixp4xx-npe.c "firmware block #%i\n", blk->offset, i); blk 620 drivers/soc/ixp4xx/ixp4xx-npe.c cb = (struct dl_codeblock*)&image->data[blk->offset]; blk 621 drivers/soc/ixp4xx/ixp4xx-npe.c if (blk->type == FW_BLOCK_TYPE_INSTR) { blk 625 drivers/soc/ixp4xx/ixp4xx-npe.c } else if (blk->type == FW_BLOCK_TYPE_DATA) { blk 631 drivers/soc/ixp4xx/ixp4xx-npe.c "type 0x%X\n", i, blk->type); blk 634 drivers/soc/ixp4xx/ixp4xx-npe.c if (blk->offset + sizeof(*cb) / 4 + cb->size > image->size) { blk 638 drivers/soc/ixp4xx/ixp4xx-npe.c blk->type == FW_BLOCK_TYPE_INSTR ? 'I' : 'D', blk 656 drivers/soc/ixp4xx/ixp4xx-npe.c blk->type == FW_BLOCK_TYPE_INSTR ? 'I' : 'D', blk 127 drivers/staging/wusbcore/devconnect.c cack_ie->blk[cnt].CDID = dev_itr->cdid; blk 128 drivers/staging/wusbcore/devconnect.c cack_ie->blk[cnt].bDeviceAddress = dev_itr->addr; blk 133 drivers/staging/wusbcore/devconnect.c + cnt * sizeof(cack_ie->blk[0]); blk 112 drivers/staging/wusbcore/include/wusb.h } blk[WUIE_ELT_MAX]; blk 1219 drivers/usb/storage/ene_ub6250.c u16 blk, index = 0; blk 1231 drivers/usb/storage/ene_ub6250.c blk = be16_to_cpu(PageBuf[index]); blk 1232 drivers/usb/storage/ene_ub6250.c if (blk == MS_LB_NOT_USED) blk 1234 drivers/usb/storage/ene_ub6250.c if (blk == info->MS_Lib.Log2PhyMap[0]) { blk 1386 drivers/usb/storage/ene_ub6250.c u16 blk; blk 1394 drivers/usb/storage/ene_ub6250.c for (blk = phyblk + 1; blk != phyblk; blk++) { blk 1395 drivers/usb/storage/ene_ub6250.c if ((blk & MS_PHYSICAL_BLOCKS_PER_SEGMENT_MASK) == 0) blk 1396 drivers/usb/storage/ene_ub6250.c blk -= MS_PHYSICAL_BLOCKS_PER_SEGMENT; blk 1398 drivers/usb/storage/ene_ub6250.c if (info->MS_Lib.Phy2LogMap[blk] == MS_LB_NOT_USED_ERASED) { blk 1399 drivers/usb/storage/ene_ub6250.c return blk; blk 1400 drivers/usb/storage/ene_ub6250.c } else if (info->MS_Lib.Phy2LogMap[blk] == MS_LB_NOT_USED) { blk 1401 drivers/usb/storage/ene_ub6250.c switch (ms_lib_read_extra(us, blk, 0, &extdat)) { blk 1411 drivers/usb/storage/ene_ub6250.c ms_lib_setacquired_errorblock(us, blk); blk 1416 drivers/usb/storage/ene_ub6250.c ms_lib_setacquired_errorblock(us, blk); blk 1420 drivers/usb/storage/ene_ub6250.c switch (ms_lib_erase_phyblock(us, blk)) { blk 1422 drivers/usb/storage/ene_ub6250.c return blk; blk 1427 drivers/usb/storage/ene_ub6250.c ms_lib_error_phyblock(us, blk); blk 131 fs/adfs/dir_f.c int blk = 0; blk 145 fs/adfs/dir_f.c for (blk = 0; blk < size; blk++) { blk 148 fs/adfs/dir_f.c phys = __adfs_block_map(sb, indaddr, blk); blk 151 fs/adfs/dir_f.c indaddr, blk); blk 155 fs/adfs/dir_f.c dir->bh[blk] = sb_bread(sb, phys); blk 156 fs/adfs/dir_f.c if (!dir->bh[blk]) blk 174 fs/adfs/dir_f.c dir->nr_buffers = blk; blk 181 fs/adfs/dir_f.c for (blk -= 1; blk >= 0; blk -= 1) blk 182 fs/adfs/dir_f.c brelse(dir->bh[blk]); blk 17 fs/adfs/dir_fplus.c unsigned int blk, size; blk 69 fs/adfs/dir_fplus.c for (blk = 1; blk < size; blk++) { blk 70 fs/adfs/dir_fplus.c block = __adfs_block_map(sb, id, blk); blk 72 fs/adfs/dir_fplus.c adfs_error(sb, "dir object %X has a hole at offset %d", id, blk); blk 76 fs/adfs/dir_fplus.c dir->bh_fplus[blk] = sb_bread(sb, block); blk 77 fs/adfs/dir_fplus.c if (!dir->bh_fplus[blk]) { blk 79 fs/adfs/dir_fplus.c id, blk, block); blk 19 fs/affs/affs.h #define AFFS_BLOCK(sb, bh, blk) (AFFS_HEAD(bh)->table[AFFS_SB(sb)->s_hashsize-1-(blk)]) blk 44 fs/affs/bitmap.c u32 blk, bmap, bit, mask, tmp; blk 52 fs/affs/bitmap.c blk = block - sbi->s_reserved; blk 53 fs/affs/bitmap.c bmap = blk / sbi->s_bmap_bits; blk 54 fs/affs/bitmap.c bit = blk % sbi->s_bmap_bits; blk 122 fs/affs/bitmap.c u32 blk, bmap, bit, mask, mask2, tmp; blk 144 fs/affs/bitmap.c blk = goal - sbi->s_reserved; blk 145 fs/affs/bitmap.c bmap = blk / sbi->s_bmap_bits; blk 167 fs/affs/bitmap.c blk = bmap * sbi->s_bmap_bits; blk 182 fs/affs/bitmap.c bit = blk % sbi->s_bmap_bits; blk 186 fs/affs/bitmap.c blk &= ~31UL; blk 194 fs/affs/bitmap.c blk += 32; blk 207 fs/affs/bitmap.c blk += bit + sbi->s_reserved; blk 209 fs/affs/bitmap.c AFFS_I(inode)->i_lastalloc = blk; blk 231 fs/affs/bitmap.c pr_debug("%d\n", blk); blk 232 fs/affs/bitmap.c return blk; blk 249 fs/affs/bitmap.c u32 size, blk, end, offset, mask; blk 275 fs/affs/bitmap.c blk = sb->s_blocksize / 4 - 49; blk 276 fs/affs/bitmap.c end = blk + 25; blk 281 fs/affs/bitmap.c bm->bm_key = be32_to_cpu(bmap_blk[blk]); blk 294 fs/affs/bitmap.c pr_debug("read bitmap block %d: %d\n", blk, bm->bm_key); blk 300 fs/affs/bitmap.c if (++blk < end || i == 1) blk 304 fs/affs/bitmap.c bmap_bh = affs_bread(sb, be32_to_cpu(bmap_blk[blk])); blk 311 fs/affs/bitmap.c blk = 0; blk 839 fs/affs/file.c u32 last_blk, blkcnt, blk; blk 896 fs/affs/file.c blk = last_blk; blk 899 fs/affs/file.c blk++; blk 904 fs/affs/file.c if (size > blkcnt - blk + i) blk 905 fs/affs/file.c size = blkcnt - blk + i; blk 906 fs/affs/file.c for (; i < size; i++, blk++) { blk 942 fs/affs/file.c if (size > blkcnt - blk) blk 943 fs/affs/file.c size = blkcnt - blk; blk 944 fs/affs/file.c for (i = 0; i < size; i++, blk++) blk 654 fs/ext2/inode.c ext2_fsblk_t blk; blk 668 fs/ext2/inode.c blk = le32_to_cpu(*(chain[depth-1].p + count)); blk 669 fs/ext2/inode.c if (blk == first_block + count) blk 1496 fs/ext2/super.c sector_t blk = off >> EXT2_BLOCK_SIZE_BITS(sb); blk 1516 fs/ext2/super.c err = ext2_get_block(inode, blk, &tmp_bh, 0); blk 1531 fs/ext2/super.c blk++; blk 1541 fs/ext2/super.c sector_t blk = off >> EXT2_BLOCK_SIZE_BITS(sb); blk 1555 fs/ext2/super.c err = ext2_get_block(inode, blk, &tmp_bh, 1); blk 1576 fs/ext2/super.c blk++; blk 319 fs/ext4/balloc.c ext4_fsblk_t blk; blk 334 fs/ext4/balloc.c blk = ext4_block_bitmap(sb, desc); blk 335 fs/ext4/balloc.c offset = blk - group_first_block; blk 339 fs/ext4/balloc.c return blk; blk 342 fs/ext4/balloc.c blk = ext4_inode_bitmap(sb, desc); blk 343 fs/ext4/balloc.c offset = blk - group_first_block; blk 347 fs/ext4/balloc.c return blk; blk 350 fs/ext4/balloc.c blk = ext4_inode_table(sb, desc); blk 351 fs/ext4/balloc.c offset = blk - group_first_block; blk 354 fs/ext4/balloc.c return blk; blk 361 fs/ext4/balloc.c return blk; blk 370 fs/ext4/balloc.c ext4_fsblk_t blk; blk 389 fs/ext4/balloc.c blk = ext4_valid_block_bitmap(sb, desc, block_group, bh); blk 390 fs/ext4/balloc.c if (unlikely(blk != 0)) { blk 393 fs/ext4/balloc.c block_group, blk); blk 366 fs/ext4/block_validity.c unsigned int blk; blk 374 fs/ext4/block_validity.c blk = le32_to_cpu(*bref++); blk 375 fs/ext4/block_validity.c if (blk && blk 377 fs/ext4/block_validity.c blk, 1))) { blk 378 fs/ext4/block_validity.c es->s_last_error_block = cpu_to_le64(blk); blk 379 fs/ext4/block_validity.c ext4_error_inode(inode, function, line, blk, blk 285 fs/ext4/ext4.h #define EXT4_B2C(sbi, blk) ((blk) >> (sbi)->s_cluster_bits) blk 2842 fs/ext4/ext4.h struct ext4_group_desc *bg, ext4_fsblk_t blk); blk 2844 fs/ext4/ext4.h struct ext4_group_desc *bg, ext4_fsblk_t blk); blk 2846 fs/ext4/ext4.h struct ext4_group_desc *bg, ext4_fsblk_t blk); blk 2896 fs/ext4/ext4.h ext4_fsblk_t blk) blk 2898 fs/ext4/ext4.h es->s_blocks_count_lo = cpu_to_le32((u32)blk); blk 2899 fs/ext4/ext4.h es->s_blocks_count_hi = cpu_to_le32(blk >> 32); blk 2903 fs/ext4/ext4.h ext4_fsblk_t blk) blk 2905 fs/ext4/ext4.h es->s_free_blocks_count_lo = cpu_to_le32((u32)blk); blk 2906 fs/ext4/ext4.h es->s_free_blocks_count_hi = cpu_to_le32(blk >> 32); blk 2910 fs/ext4/ext4.h ext4_fsblk_t blk) blk 2912 fs/ext4/ext4.h es->s_r_blocks_count_lo = cpu_to_le32((u32)blk); blk 2913 fs/ext4/ext4.h es->s_r_blocks_count_hi = cpu_to_le32(blk >> 32); blk 1838 fs/ext4/extents.c ext4_fsblk_t blk; blk 1856 fs/ext4/extents.c blk = ext4_idx_pblock(path[0].p_idx); blk 1869 fs/ext4/extents.c ext4_free_blocks(handle, inode, NULL, blk, 1, blk 84 fs/ext4/ialloc.c ext4_fsblk_t blk; blk 95 fs/ext4/ialloc.c blk = ext4_inode_bitmap(sb, desc); blk 100 fs/ext4/ialloc.c "inode_bitmap = %llu", block_group, blk); blk 1354 fs/ext4/ialloc.c ext4_fsblk_t blk; blk 1404 fs/ext4/ialloc.c blk = ext4_inode_table(sb, gdp) + used_blks; blk 1423 fs/ext4/ialloc.c ret = sb_issue_zeroout(sb, blk, num, GFP_NOFS); blk 541 fs/ext4/indirect.c ext4_fsblk_t blk; blk 543 fs/ext4/indirect.c blk = le32_to_cpu(*(chain[depth-1].p + count)); blk 545 fs/ext4/indirect.c if (blk == first_block + count) blk 621 fs/ext4/migrate.c ext4_fsblk_t blk; blk 661 fs/ext4/migrate.c blk = len = start = end = 0; blk 664 fs/ext4/migrate.c blk = ext4_ext_pblock(ex); blk 676 fs/ext4/migrate.c ei->i_data[i] = cpu_to_le32(blk++); blk 398 fs/ext4/resize.c ext4_fsblk_t blk) blk 403 fs/ext4/resize.c bh = sb_getblk(sb, blk); blk 767 fs/ext4/resize.c const ext4_fsblk_t blk = primary->b_blocknr; blk 777 fs/ext4/resize.c grp * EXT4_BLOCKS_PER_GROUP(sb) + blk){ blk 780 fs/ext4/resize.c blk, grp, blk 783 fs/ext4/resize.c blk); blk 1004 fs/ext4/resize.c ext4_fsblk_t blk; blk 1022 fs/ext4/resize.c blk = EXT4_SB(sb)->s_sbh->b_blocknr + 1 + EXT4_SB(sb)->s_gdb_count; blk 1028 fs/ext4/resize.c for (res = 0; res < reserved_gdb; res++, blk++) { blk 1029 fs/ext4/resize.c if (le32_to_cpu(*data) != blk) { blk 1032 fs/ext4/resize.c blk, blk 1037 fs/ext4/resize.c primary[res] = ext4_sb_bread(sb, blk, 0); blk 1066 fs/ext4/resize.c blk = group * EXT4_BLOCKS_PER_GROUP(sb); blk 1073 fs/ext4/resize.c data[gdbackups] = cpu_to_le32(blk + primary[i]->b_blocknr); blk 284 fs/ext4/super.c struct ext4_group_desc *bg, ext4_fsblk_t blk) blk 286 fs/ext4/super.c bg->bg_block_bitmap_lo = cpu_to_le32((u32)blk); blk 288 fs/ext4/super.c bg->bg_block_bitmap_hi = cpu_to_le32(blk >> 32); blk 292 fs/ext4/super.c struct ext4_group_desc *bg, ext4_fsblk_t blk) blk 294 fs/ext4/super.c bg->bg_inode_bitmap_lo = cpu_to_le32((u32)blk); blk 296 fs/ext4/super.c bg->bg_inode_bitmap_hi = cpu_to_le32(blk >> 32); blk 300 fs/ext4/super.c struct ext4_group_desc *bg, ext4_fsblk_t blk) blk 302 fs/ext4/super.c bg->bg_inode_table_lo = cpu_to_le32((u32)blk); blk 304 fs/ext4/super.c bg->bg_inode_table_hi = cpu_to_le32(blk >> 32); blk 5980 fs/ext4/super.c ext4_lblk_t blk = off >> EXT4_BLOCK_SIZE_BITS(sb); blk 5995 fs/ext4/super.c bh = ext4_bread(NULL, inode, blk, 0); blk 6006 fs/ext4/super.c blk++; blk 6017 fs/ext4/super.c ext4_lblk_t blk = off >> EXT4_BLOCK_SIZE_BITS(sb); blk 6041 fs/ext4/super.c bh = ext4_bread(handle, inode, blk, blk 1454 fs/f2fs/checkpoint.c block_t blk; blk 1459 fs/f2fs/checkpoint.c blk = start_blk + sbi->blocks_per_seg - nm_i->nat_bits_blocks; blk 1462 fs/f2fs/checkpoint.c (i << F2FS_BLKSIZE_BITS), blk + i); blk 838 fs/f2fs/data.c dn->data_blkaddr = ei.blk + index - ei.fofs; blk 859 fs/f2fs/data.c dn.data_blkaddr = ei.blk + index - ei.fofs; blk 1170 fs/f2fs/data.c map->m_pblk = ei.blk + pgofs - ei.fofs; blk 1461 fs/f2fs/data.c static inline loff_t blk_to_logical(struct inode *inode, sector_t blk) blk 1463 fs/f2fs/data.c return (blk << inode->i_blkbits); blk 1979 fs/f2fs/data.c fio->old_blkaddr = ei.blk + page->index - ei.fofs; blk 2575 fs/f2fs/data.c dn.data_blkaddr = ei.blk + index - ei.fofs; blk 439 fs/f2fs/extent_cache.c next_ex->ei.blk = ei->blk; blk 561 fs/f2fs/extent_cache.c end - dei.fofs + dei.blk, blk 568 fs/f2fs/extent_cache.c en->ei.blk += end - dei.fofs; blk 576 fs/f2fs/f2fs.h u32 blk; /* start block address of the extent */ blk 722 fs/f2fs/f2fs.h ext->blk = le32_to_cpu(i_ext->blk); blk 730 fs/f2fs/f2fs.h i_ext->blk = cpu_to_le32(ext->blk); blk 735 fs/f2fs/f2fs.h u32 blk, unsigned int len) blk 738 fs/f2fs/f2fs.h ei->blk = blk; blk 765 fs/f2fs/f2fs.h back->blk + back->len == front->blk); blk 683 fs/f2fs/gc.c dn.data_blkaddr = ei.blk + index - ei.fofs; blk 262 fs/f2fs/inode.c (!f2fs_is_valid_blkaddr(sbi, ei->blk, blk 264 fs/f2fs/inode.c !f2fs_is_valid_blkaddr(sbi, ei->blk + ei->len - 1, blk 269 fs/f2fs/inode.c ei->blk, ei->fofs, ei->len); blk 1048 fs/f2fs/segment.c block_t blk = start; blk 1052 fs/f2fs/segment.c while (blk < end) { blk 1053 fs/f2fs/segment.c segno = GET_SEGNO(sbi, blk); blk 1055 fs/f2fs/segment.c offset = GET_BLKOFF_FROM_SEG0(sbi, blk); blk 1064 fs/f2fs/segment.c blk = START_BLOCK(sbi, segno + 1); blk 149 fs/freevxfs/vxfs_super.c unsigned blk, __fs32 magic) blk 156 fs/freevxfs/vxfs_super.c bp = sb_bread(sbp, blk); blk 162 fs/freevxfs/vxfs_super.c blk); blk 173 fs/freevxfs/vxfs_super.c rsbp->vs_magic, blk); blk 1979 fs/gfs2/dir.c u64 blk, nblk; blk 2005 fs/gfs2/dir.c for (blk = leaf_no; blk; blk = nblk) { blk 2006 fs/gfs2/dir.c if (blk != leaf_no) { blk 2007 fs/gfs2/dir.c error = get_leaf(dip, blk, &bh); blk 2013 fs/gfs2/dir.c if (blk != leaf_no) blk 2016 fs/gfs2/dir.c gfs2_rlist_add(dip, &rlist, blk); blk 2040 fs/gfs2/dir.c for (blk = leaf_no; blk; blk = nblk) { blk 2043 fs/gfs2/dir.c if (blk != leaf_no) { blk 2044 fs/gfs2/dir.c error = get_leaf(dip, blk, &bh); blk 2050 fs/gfs2/dir.c if (blk != leaf_no) blk 2053 fs/gfs2/dir.c rgd = gfs2_blk2rgrpd(sdp, blk, true); blk 2054 fs/gfs2/dir.c gfs2_free_meta(dip, rgd, blk, 1); blk 696 fs/gfs2/quota.c u64 blk; blk 701 fs/gfs2/quota.c blk = index << (PAGE_SHIFT - sdp->sd_sb.sb_bsize_shift); blk 716 fs/gfs2/quota.c blk++; blk 720 fs/gfs2/quota.c gfs2_block_map(inode, blk, bh, 1); blk 32 fs/gfs2/recovery.c int gfs2_replay_read_block(struct gfs2_jdesc *jd, unsigned int blk, blk 42 fs/gfs2/recovery.c error = gfs2_extent_map(&ip->i_inode, blk, &new, &dblock, &extlen); blk 163 fs/gfs2/recovery.c static int get_log_header(struct gfs2_jdesc *jd, unsigned int blk, blk 170 fs/gfs2/recovery.c error = gfs2_replay_read_block(jd, blk, &bh); blk 175 fs/gfs2/recovery.c blk, head); blk 14 fs/gfs2/recovery.h static inline void gfs2_replay_incr_blk(struct gfs2_jdesc *jd, u32 *blk) blk 16 fs/gfs2/recovery.h if (++*blk == jd->jd_blocks) blk 17 fs/gfs2/recovery.h *blk = 0; blk 20 fs/gfs2/recovery.h extern int gfs2_replay_read_block(struct gfs2_jdesc *jd, unsigned int blk, blk 196 fs/gfs2/rgrp.c static inline int rs_cmp(u64 blk, u32 len, struct gfs2_blkreserv *rs) blk 200 fs/gfs2/rgrp.c if (blk >= startblk + rs->rs_free) blk 202 fs/gfs2/rgrp.c if (blk + len - 1 < startblk) blk 508 fs/gfs2/rgrp.c struct gfs2_rgrpd *gfs2_blk2rgrpd(struct gfs2_sbd *sdp, u64 blk, bool exact) blk 518 fs/gfs2/rgrp.c if (blk < cur->rd_addr) blk 520 fs/gfs2/rgrp.c else if (blk >= cur->rd_data0 + cur->rd_data) blk 525 fs/gfs2/rgrp.c if (blk < cur->rd_addr) blk 527 fs/gfs2/rgrp.c if (blk >= cur->rd_data0 + cur->rd_data) blk 1323 fs/gfs2/rgrp.c u64 blk; blk 1344 fs/gfs2/rgrp.c blk = offset + ((bi->bi_start + x) * GFS2_NBBY); blk 1349 fs/gfs2/rgrp.c if ((start + nr_blks) != blk) { blk 1360 fs/gfs2/rgrp.c start = blk; blk 1365 fs/gfs2/rgrp.c blk++; blk 27 fs/gfs2/rgrp.h extern struct gfs2_rgrpd *gfs2_blk2rgrpd(struct gfs2_sbd *sdp, u64 blk, bool exact); blk 992 fs/gfs2/xattr.c u64 blk; blk 994 fs/gfs2/xattr.c error = gfs2_alloc_blocks(ip, &blk, &n, 0, NULL); blk 997 fs/gfs2/xattr.c gfs2_trans_remove_revoke(sdp, blk, 1); blk 998 fs/gfs2/xattr.c indbh = gfs2_meta_new(ip->i_gl, blk); blk 1005 fs/gfs2/xattr.c ip->i_eattr = blk; blk 265 fs/ioctl.c static inline loff_t blk_to_logical(struct inode *inode, sector_t blk) blk 267 fs/ioctl.c return (blk << inode->i_blkbits); blk 739 fs/jfs/super.c sector_t blk = off >> sb->s_blocksize_bits; blk 759 fs/jfs/super.c err = jfs_get_block(inode, blk, &tmp_bh, 0); blk 774 fs/jfs/super.c blk++; blk 784 fs/jfs/super.c sector_t blk = off >> sb->s_blocksize_bits; blk 799 fs/jfs/super.c err = jfs_get_block(inode, blk, &tmp_bh, 1); blk 820 fs/jfs/super.c blk++; blk 1809 fs/ocfs2/dir.c unsigned long offset, blk, last_ra_blk = 0; blk 1822 fs/ocfs2/dir.c blk = ctx->pos >> sb->s_blocksize_bits; blk 1823 fs/ocfs2/dir.c if (ocfs2_read_dir_block(inode, blk, &bh, 0)) { blk 1835 fs/ocfs2/dir.c || (((last_ra_blk - blk) << 9) <= (ra_sectors / 2))) { blk 1839 fs/ocfs2/dir.c if (!ocfs2_read_dir_block(inode, ++blk, &tmp, blk 1843 fs/ocfs2/dir.c last_ra_blk = blk; blk 4323 fs/ocfs2/dir.c u64 blk; blk 4368 fs/ocfs2/dir.c blk = le64_to_cpu(dx_root->dr_blkno); blk 4373 fs/ocfs2/dir.c bg_blkno = ocfs2_which_suballoc_group(blk, bit); blk 693 fs/ocfs2/ocfs2_trace.h TP_PROTO(struct super_block *sb, unsigned long long blk, blk 695 fs/ocfs2/ocfs2_trace.h TP_ARGS(sb, blk, count), blk 699 fs/ocfs2/ocfs2_trace.h __field(unsigned long long, blk) blk 705 fs/ocfs2/ocfs2_trace.h __entry->blk = blk; blk 710 fs/ocfs2/ocfs2_trace.h __entry->blk, __entry->count) blk 174 fs/ocfs2/quota_global.c sector_t blk = off >> sb->s_blocksize_bits; blk 188 fs/ocfs2/quota_global.c err = ocfs2_extent_map_get_blocks(gqinode, blk, &pblock, blk 209 fs/ocfs2/quota_global.c blk++; blk 223 fs/ocfs2/quota_global.c sector_t blk = off >> sb->s_blocksize_bits; blk 252 fs/ocfs2/quota_global.c err = ocfs2_extent_map_get_blocks(gqinode, blk, &pblock, &pcount, NULL); blk 767 fs/ocfs2/refcounttree.c u64 blk = 0, bg_blkno = 0, ref_blkno = le64_to_cpu(di->i_refcount_loc); blk 787 fs/ocfs2/refcounttree.c blk = le64_to_cpu(rb->rf_blkno); blk 792 fs/ocfs2/refcounttree.c bg_blkno = ocfs2_which_suballoc_group(blk, bit); blk 2531 fs/ocfs2/suballoc.c u64 blk = le64_to_cpu(di->i_blkno); blk 2533 fs/ocfs2/suballoc.c u64 bg_blkno = ocfs2_which_suballoc_group(blk, bit); blk 2492 fs/ocfs2/xattr.c u64 blk, bg_blkno; blk 2508 fs/ocfs2/xattr.c blk = le64_to_cpu(xb->xb_blkno); blk 2513 fs/ocfs2/xattr.c bg_blkno = ocfs2_which_suballoc_group(blk, bit); blk 4627 fs/ocfs2/xattr.c u64 blk, blk 4639 fs/ocfs2/xattr.c trace_ocfs2_divide_xattr_bucket_begin((unsigned long long)blk, blk 4650 fs/ocfs2/xattr.c ret = ocfs2_read_xattr_bucket(s_bucket, blk); blk 59 fs/quota/quota_tree.c static ssize_t read_blk(struct qtree_mem_dqinfo *info, uint blk, char *buf) blk 65 fs/quota/quota_tree.c info->dqi_usable_bs, blk << info->dqi_blocksize_bits); blk 68 fs/quota/quota_tree.c static ssize_t write_blk(struct qtree_mem_dqinfo *info, uint blk, char *buf) blk 74 fs/quota/quota_tree.c info->dqi_usable_bs, blk << info->dqi_blocksize_bits); blk 88 fs/quota/quota_tree.c int ret, blk; blk 93 fs/quota/quota_tree.c blk = info->dqi_free_blk; blk 94 fs/quota/quota_tree.c ret = read_blk(info, blk, buf); blk 105 fs/quota/quota_tree.c blk = info->dqi_blocks++; blk 108 fs/quota/quota_tree.c ret = blk; blk 115 fs/quota/quota_tree.c static int put_free_dqblk(struct qtree_mem_dqinfo *info, char *buf, uint blk) blk 123 fs/quota/quota_tree.c err = write_blk(info, blk, buf); blk 126 fs/quota/quota_tree.c info->dqi_free_blk = blk; blk 133 fs/quota/quota_tree.c uint blk) blk 169 fs/quota/quota_tree.c if (write_blk(info, blk, buf) < 0) blk 171 fs/quota/quota_tree.c "with free entries", blk); blk 180 fs/quota/quota_tree.c uint blk) blk 190 fs/quota/quota_tree.c err = write_blk(info, blk, buf); blk 198 fs/quota/quota_tree.c cpu_to_le32(blk); blk 204 fs/quota/quota_tree.c info->dqi_free_entry = blk; blk 228 fs/quota/quota_tree.c uint blk, i; blk 240 fs/quota/quota_tree.c blk = info->dqi_free_entry; blk 241 fs/quota/quota_tree.c *err = read_blk(info, blk, buf); blk 245 fs/quota/quota_tree.c blk = get_free_dqblk(info); blk 246 fs/quota/quota_tree.c if ((int)blk < 0) { blk 247 fs/quota/quota_tree.c *err = blk; blk 254 fs/quota/quota_tree.c info->dqi_free_entry = blk; blk 259 fs/quota/quota_tree.c *err = remove_free_dqentry(info, buf, blk); blk 262 fs/quota/quota_tree.c "from entry free list", blk); blk 281 fs/quota/quota_tree.c *err = write_blk(info, blk, buf); blk 284 fs/quota/quota_tree.c blk); blk 287 fs/quota/quota_tree.c dquot->dq_off = (blk << info->dqi_blocksize_bits) + blk 291 fs/quota/quota_tree.c return blk; blk 414 fs/quota/quota_tree.c uint blk) blk 422 fs/quota/quota_tree.c if (dquot->dq_off >> info->dqi_blocksize_bits != blk) { blk 424 fs/quota/quota_tree.c "other block (%u) than it should (%u)", blk, blk 428 fs/quota/quota_tree.c ret = read_blk(info, blk, buf); blk 431 fs/quota/quota_tree.c blk); blk 437 fs/quota/quota_tree.c ret = remove_free_dqentry(info, buf, blk); blk 439 fs/quota/quota_tree.c ret = put_free_dqblk(info, buf, blk); blk 442 fs/quota/quota_tree.c "(%u) to free list", blk); blk 452 fs/quota/quota_tree.c ret = insert_free_dqentry(info, buf, blk); blk 455 fs/quota/quota_tree.c "data block (%u) to free entry list", blk); blk 459 fs/quota/quota_tree.c ret = write_blk(info, blk, buf); blk 462 fs/quota/quota_tree.c "data block %u", blk); blk 475 fs/quota/quota_tree.c uint *blk, int depth) blk 484 fs/quota/quota_tree.c ret = read_blk(info, *blk, buf); blk 487 fs/quota/quota_tree.c *blk); blk 505 fs/quota/quota_tree.c && *blk != QT_TREEOFF) { blk 506 fs/quota/quota_tree.c put_free_dqblk(info, buf, *blk); blk 507 fs/quota/quota_tree.c *blk = 0; blk 509 fs/quota/quota_tree.c ret = write_blk(info, *blk, buf); blk 513 fs/quota/quota_tree.c *blk); blk 534 fs/quota/quota_tree.c struct dquot *dquot, uint blk) blk 543 fs/quota/quota_tree.c ret = read_blk(info, blk, buf); blk 546 fs/quota/quota_tree.c "block %u", blk); blk 562 fs/quota/quota_tree.c ret = (blk << info->dqi_blocksize_bits) + sizeof(struct blk 572 fs/quota/quota_tree.c struct dquot *dquot, uint blk, int depth) blk 580 fs/quota/quota_tree.c ret = read_blk(info, blk, buf); blk 583 fs/quota/quota_tree.c blk); blk 587 fs/quota/quota_tree.c blk = le32_to_cpu(ref[get_index(info, dquot->dq_id, depth)]); blk 588 fs/quota/quota_tree.c if (!blk) /* No reference? */ blk 591 fs/quota/quota_tree.c ret = find_tree_dqentry(info, dquot, blk, depth+1); blk 593 fs/quota/quota_tree.c ret = find_block_dqentry(info, dquot, blk); blk 680 fs/quota/quota_tree.c unsigned int blk, int depth) blk 695 fs/quota/quota_tree.c ret = read_blk(info, blk, buf); blk 698 fs/quota/quota_tree.c "Can't read quota tree block %u", blk); blk 2495 fs/reiserfs/super.c unsigned long blk = off >> sb->s_blocksize_bits; blk 2516 fs/reiserfs/super.c err = reiserfs_get_block(inode, blk, &tmp_bh, 0); blk 2532 fs/reiserfs/super.c blk++; blk 2545 fs/reiserfs/super.c unsigned long blk = off >> sb->s_blocksize_bits; blk 2561 fs/reiserfs/super.c err = reiserfs_get_block(inode, blk, &tmp_bh, GET_BLOCK_CREATE); blk 2588 fs/reiserfs/super.c blk++; blk 42 fs/squashfs/export.c int blk = SQUASHFS_LOOKUP_BLOCK(ino_num - 1); blk 44 fs/squashfs/export.c u64 start = le64_to_cpu(msblk->inode_lookup_table[blk]); blk 200 fs/squashfs/namei.c unsigned int blk, off, ino_num; blk 202 fs/squashfs/namei.c blk = le32_to_cpu(dirh.start_block); blk 206 fs/squashfs/namei.c ino = SQUASHFS_MKINODE(blk, off); blk 210 fs/squashfs/namei.c blk, off, ino_num); blk 1513 fs/ubifs/journal.c unsigned int blk; blk 1543 fs/ubifs/journal.c blk = new_size >> UBIFS_BLOCK_SHIFT; blk 1544 fs/ubifs/journal.c data_key_init(c, &key, inum, blk); blk 1556 fs/ubifs/journal.c blk, inode->i_ino); blk 1564 fs/ubifs/journal.c err = truncate_data_node(c, inode, blk, dn, &dlen); blk 1622 fs/ubifs/journal.c blk = (new_size >> UBIFS_BLOCK_SHIFT) + (bit ? 1 : 0); blk 1623 fs/ubifs/journal.c data_key_init(c, &key, inum, blk); blk 1626 fs/ubifs/journal.c blk = (old_size >> UBIFS_BLOCK_SHIFT) - (bit ? 0 : 1); blk 1627 fs/ubifs/journal.c data_key_init(c, &to_key, inum, blk); blk 257 fs/ufs/ufs_fs.h #define ufs_lblktosize(blk) ((blk) << uspi->s_bshift) blk 343 fs/ufs/util.h u64 blk) blk 346 fs/ufs/util.h return ubh_get_addr64(ubh, blk); blk 348 fs/ufs/util.h return ubh_get_addr32(ubh, blk); blk 553 fs/ufs/util.h unsigned blk) blk 555 fs/ufs/util.h BUG_ON(blk > UFS_TIND_BLOCK); blk 557 fs/ufs/util.h (void *)&ufsi->i_u1.u2_i_data[blk] : blk 558 fs/ufs/util.h (void *)&ufsi->i_u1.i_data[blk]; blk 849 fs/xfs/libxfs/xfs_attr.c struct xfs_da_state_blk *blk; blk 873 fs/xfs/libxfs/xfs_attr.c blk = &state->path.blk[ state->path.active-1 ]; blk 874 fs/xfs/libxfs/xfs_attr.c ASSERT(blk->magic == XFS_ATTR_LEAF_MAGIC); blk 901 fs/xfs/libxfs/xfs_attr.c retval = xfs_attr3_leaf_add(blk->bp, state->args); blk 1022 fs/xfs/libxfs/xfs_attr.c blk = &state->path.blk[ state->path.active-1 ]; blk 1023 fs/xfs/libxfs/xfs_attr.c ASSERT(blk->magic == XFS_ATTR_LEAF_MAGIC); blk 1024 fs/xfs/libxfs/xfs_attr.c error = xfs_attr3_leaf_remove(blk->bp, args); blk 1076 fs/xfs/libxfs/xfs_attr.c struct xfs_da_state_blk *blk; blk 1106 fs/xfs/libxfs/xfs_attr.c blk = &state->path.blk[ state->path.active-1 ]; blk 1107 fs/xfs/libxfs/xfs_attr.c ASSERT(blk->bp != NULL); blk 1108 fs/xfs/libxfs/xfs_attr.c ASSERT(blk->magic == XFS_ATTR_LEAF_MAGIC); blk 1142 fs/xfs/libxfs/xfs_attr.c blk = &state->path.blk[ state->path.active-1 ]; blk 1143 fs/xfs/libxfs/xfs_attr.c ASSERT(blk->magic == XFS_ATTR_LEAF_MAGIC); blk 1144 fs/xfs/libxfs/xfs_attr.c retval = xfs_attr3_leaf_remove(blk->bp, args); blk 1173 fs/xfs/libxfs/xfs_attr.c ASSERT(state->path.blk[0].bp); blk 1174 fs/xfs/libxfs/xfs_attr.c state->path.blk[0].bp = NULL; blk 1208 fs/xfs/libxfs/xfs_attr.c xfs_da_state_blk_t *blk; blk 1219 fs/xfs/libxfs/xfs_attr.c for (blk = path->blk, level = 0; level < path->active; blk++, level++) { blk 1220 fs/xfs/libxfs/xfs_attr.c if (blk->bp) { blk 1221 fs/xfs/libxfs/xfs_attr.c blk->disk_blkno = XFS_BUF_ADDR(blk->bp); blk 1222 fs/xfs/libxfs/xfs_attr.c blk->bp = NULL; blk 1224 fs/xfs/libxfs/xfs_attr.c blk->disk_blkno = 0; blk 1234 fs/xfs/libxfs/xfs_attr.c for (blk = path->blk, level = 0; level < path->active; blk++, level++) { blk 1235 fs/xfs/libxfs/xfs_attr.c if (blk->bp) { blk 1236 fs/xfs/libxfs/xfs_attr.c blk->disk_blkno = XFS_BUF_ADDR(blk->bp); blk 1237 fs/xfs/libxfs/xfs_attr.c blk->bp = NULL; blk 1239 fs/xfs/libxfs/xfs_attr.c blk->disk_blkno = 0; blk 1256 fs/xfs/libxfs/xfs_attr.c xfs_da_state_blk_t *blk; blk 1267 fs/xfs/libxfs/xfs_attr.c for (blk = path->blk, level = 0; level < path->active; blk++, level++) { blk 1268 fs/xfs/libxfs/xfs_attr.c if (blk->disk_blkno) { blk 1271 fs/xfs/libxfs/xfs_attr.c blk->blkno, blk->disk_blkno, blk 1272 fs/xfs/libxfs/xfs_attr.c &blk->bp, XFS_ATTR_FORK); blk 1276 fs/xfs/libxfs/xfs_attr.c blk->bp = NULL; blk 1286 fs/xfs/libxfs/xfs_attr.c for (blk = path->blk, level = 0; level < path->active; blk++, level++) { blk 1287 fs/xfs/libxfs/xfs_attr.c if (blk->disk_blkno) { blk 1290 fs/xfs/libxfs/xfs_attr.c blk->blkno, blk->disk_blkno, blk 1291 fs/xfs/libxfs/xfs_attr.c &blk->bp, XFS_ATTR_FORK); blk 1295 fs/xfs/libxfs/xfs_attr.c blk->bp = NULL; blk 1315 fs/xfs/libxfs/xfs_attr.c xfs_da_state_blk_t *blk; blk 1339 fs/xfs/libxfs/xfs_attr.c blk = &state->path.blk[state->path.active - 1]; blk 1340 fs/xfs/libxfs/xfs_attr.c retval = xfs_attr3_leaf_getvalue(blk->bp, args); blk 1347 fs/xfs/libxfs/xfs_attr.c xfs_trans_brelse(args->trans, state->path.blk[i].bp); blk 1348 fs/xfs/libxfs/xfs_attr.c state->path.blk[i].bp = NULL; blk 1863 fs/xfs/libxfs/xfs_attr_leaf.c struct xfs_da_state_blk *blk; blk 1880 fs/xfs/libxfs/xfs_attr_leaf.c blk = &state->path.blk[ state->path.active-1 ]; blk 1881 fs/xfs/libxfs/xfs_attr_leaf.c leaf = blk->bp->b_addr; blk 1961 fs/xfs/libxfs/xfs_attr_leaf.c if (blkno < blk->blkno) { blk 96 fs/xfs/libxfs/xfs_da_btree.c state->altpath.blk[i].bp = NULL; blk 401 fs/xfs/libxfs/xfs_da_btree.c ASSERT(state->path.blk[max].magic == XFS_ATTR_LEAF_MAGIC || blk 402 fs/xfs/libxfs/xfs_da_btree.c state->path.blk[max].magic == XFS_DIR2_LEAFN_MAGIC); blk 404 fs/xfs/libxfs/xfs_da_btree.c addblk = &state->path.blk[max]; /* initial dummy value */ blk 406 fs/xfs/libxfs/xfs_da_btree.c oldblk = &state->path.blk[i]; blk 407 fs/xfs/libxfs/xfs_da_btree.c newblk = &state->altpath.blk[i]; blk 482 fs/xfs/libxfs/xfs_da_btree.c state->path.blk[max].magic == XFS_DIR2_LEAFN_MAGIC); blk 488 fs/xfs/libxfs/xfs_da_btree.c oldblk = &state->path.blk[0]; blk 977 fs/xfs/libxfs/xfs_da_btree.c drop_blk = &state->path.blk[ state->path.active-1 ]; blk 978 fs/xfs/libxfs/xfs_da_btree.c save_blk = &state->altpath.blk[ state->path.active-1 ]; blk 979 fs/xfs/libxfs/xfs_da_btree.c ASSERT(state->path.blk[0].magic == XFS_DA_NODE_MAGIC); blk 1045 fs/xfs/libxfs/xfs_da_btree.c error = xfs_da3_root_join(state, &state->path.blk[0]); blk 1153 fs/xfs/libxfs/xfs_da_btree.c struct xfs_da_state_blk *blk; blk 1172 fs/xfs/libxfs/xfs_da_btree.c blk = &state->path.blk[ state->path.active-1 ]; blk 1173 fs/xfs/libxfs/xfs_da_btree.c info = blk->bp->b_addr; blk 1249 fs/xfs/libxfs/xfs_da_btree.c if (blkno < blk->blkno) { blk 1298 fs/xfs/libxfs/xfs_da_btree.c struct xfs_da_state_blk *blk; blk 1309 fs/xfs/libxfs/xfs_da_btree.c blk = &path->blk[ level ]; blk 1310 fs/xfs/libxfs/xfs_da_btree.c switch (blk->magic) { blk 1312 fs/xfs/libxfs/xfs_da_btree.c lasthash = xfs_attr_leaf_lasthash(blk->bp, &count); blk 1317 fs/xfs/libxfs/xfs_da_btree.c lasthash = xfs_dir2_leaf_lasthash(dp, blk->bp, &count); blk 1322 fs/xfs/libxfs/xfs_da_btree.c lasthash = xfs_da3_node_lasthash(dp, blk->bp, &count); blk 1327 fs/xfs/libxfs/xfs_da_btree.c for (blk--, level--; level >= 0; blk--, level--) { blk 1330 fs/xfs/libxfs/xfs_da_btree.c node = blk->bp->b_addr; blk 1333 fs/xfs/libxfs/xfs_da_btree.c if (be32_to_cpu(btree[blk->index].hashval) == lasthash) blk 1335 fs/xfs/libxfs/xfs_da_btree.c blk->hashval = lasthash; blk 1336 fs/xfs/libxfs/xfs_da_btree.c btree[blk->index].hashval = cpu_to_be32(lasthash); blk 1337 fs/xfs/libxfs/xfs_da_btree.c xfs_trans_log_buf(state->args->trans, blk->bp, blk 1338 fs/xfs/libxfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &btree[blk->index], blk 1487 fs/xfs/libxfs/xfs_da_btree.c struct xfs_da_state_blk *blk; blk 1512 fs/xfs/libxfs/xfs_da_btree.c for (blk = &state->path.blk[0], state->path.active = 1; blk 1514 fs/xfs/libxfs/xfs_da_btree.c blk++, state->path.active++) { blk 1518 fs/xfs/libxfs/xfs_da_btree.c blk->blkno = blkno; blk 1520 fs/xfs/libxfs/xfs_da_btree.c -1, &blk->bp, args->whichfork); blk 1522 fs/xfs/libxfs/xfs_da_btree.c blk->blkno = 0; blk 1526 fs/xfs/libxfs/xfs_da_btree.c curr = blk->bp->b_addr; blk 1531 fs/xfs/libxfs/xfs_da_btree.c blk->magic = XFS_ATTR_LEAF_MAGIC; blk 1532 fs/xfs/libxfs/xfs_da_btree.c blk->hashval = xfs_attr_leaf_lasthash(blk->bp, NULL); blk 1538 fs/xfs/libxfs/xfs_da_btree.c blk->magic = XFS_DIR2_LEAFN_MAGIC; blk 1539 fs/xfs/libxfs/xfs_da_btree.c blk->hashval = xfs_dir2_leaf_lasthash(args->dp, blk 1540 fs/xfs/libxfs/xfs_da_btree.c blk->bp, NULL); blk 1547 fs/xfs/libxfs/xfs_da_btree.c blk->magic = XFS_DA_NODE_MAGIC; blk 1552 fs/xfs/libxfs/xfs_da_btree.c node = blk->bp->b_addr; blk 1569 fs/xfs/libxfs/xfs_da_btree.c blk->hashval = be32_to_cpu(btree[max - 1].hashval); blk 1607 fs/xfs/libxfs/xfs_da_btree.c blk->index = max - 1; blk 1610 fs/xfs/libxfs/xfs_da_btree.c blk->index = probe; blk 1629 fs/xfs/libxfs/xfs_da_btree.c if (blk->magic == XFS_DIR2_LEAFN_MAGIC) { blk 1630 fs/xfs/libxfs/xfs_da_btree.c retval = xfs_dir2_leafn_lookup_int(blk->bp, args, blk 1631 fs/xfs/libxfs/xfs_da_btree.c &blk->index, state); blk 1632 fs/xfs/libxfs/xfs_da_btree.c } else if (blk->magic == XFS_ATTR_LEAF_MAGIC) { blk 1633 fs/xfs/libxfs/xfs_da_btree.c retval = xfs_attr3_leaf_lookup_int(blk->bp, args); blk 1634 fs/xfs/libxfs/xfs_da_btree.c blk->index = args->index; blk 1635 fs/xfs/libxfs/xfs_da_btree.c args->blkno = blk->blkno; blk 1641 fs/xfs/libxfs/xfs_da_btree.c (blk->hashval == args->hashval)) { blk 1648 fs/xfs/libxfs/xfs_da_btree.c } else if (blk->magic == XFS_ATTR_LEAF_MAGIC) { blk 1879 fs/xfs/libxfs/xfs_da_btree.c struct xfs_da_state_blk *blk; blk 1903 fs/xfs/libxfs/xfs_da_btree.c for (blk = &path->blk[level]; level >= 0; blk--, level--) { blk 1904 fs/xfs/libxfs/xfs_da_btree.c node = blk->bp->b_addr; blk 1908 fs/xfs/libxfs/xfs_da_btree.c if (forward && (blk->index < nodehdr.count - 1)) { blk 1909 fs/xfs/libxfs/xfs_da_btree.c blk->index++; blk 1910 fs/xfs/libxfs/xfs_da_btree.c blkno = be32_to_cpu(btree[blk->index].before); blk 1912 fs/xfs/libxfs/xfs_da_btree.c } else if (!forward && (blk->index > 0)) { blk 1913 fs/xfs/libxfs/xfs_da_btree.c blk->index--; blk 1914 fs/xfs/libxfs/xfs_da_btree.c blkno = be32_to_cpu(btree[blk->index].before); blk 1928 fs/xfs/libxfs/xfs_da_btree.c for (blk++, level++; level < path->active; blk++, level++) { blk 1944 fs/xfs/libxfs/xfs_da_btree.c xfs_trans_brelse(args->trans, blk->bp); blk 1945 fs/xfs/libxfs/xfs_da_btree.c blk->blkno = blkno; blk 1946 fs/xfs/libxfs/xfs_da_btree.c blk->bp = bp; blk 1948 fs/xfs/libxfs/xfs_da_btree.c info = blk->bp->b_addr; blk 1964 fs/xfs/libxfs/xfs_da_btree.c blk->magic = XFS_DA_NODE_MAGIC; blk 1968 fs/xfs/libxfs/xfs_da_btree.c blk->hashval = be32_to_cpu(btree[nodehdr.count - 1].hashval); blk 1970 fs/xfs/libxfs/xfs_da_btree.c blk->index = 0; blk 1972 fs/xfs/libxfs/xfs_da_btree.c blk->index = nodehdr.count - 1; blk 1973 fs/xfs/libxfs/xfs_da_btree.c blkno = be32_to_cpu(btree[blk->index].before); blk 1977 fs/xfs/libxfs/xfs_da_btree.c blk->magic = XFS_ATTR_LEAF_MAGIC; blk 1979 fs/xfs/libxfs/xfs_da_btree.c blk->index = 0; blk 1980 fs/xfs/libxfs/xfs_da_btree.c blk->hashval = xfs_attr_leaf_lasthash(blk->bp, NULL); blk 1984 fs/xfs/libxfs/xfs_da_btree.c blk->magic = XFS_DIR2_LEAFN_MAGIC; blk 1986 fs/xfs/libxfs/xfs_da_btree.c blk->index = 0; blk 1987 fs/xfs/libxfs/xfs_da_btree.c blk->hashval = xfs_dir2_leaf_lasthash(args->dp, blk 1988 fs/xfs/libxfs/xfs_da_btree.c blk->bp, NULL); blk 112 fs/xfs/libxfs/xfs_da_btree.h xfs_da_state_blk_t blk[XFS_DA_NODE_MAXDEPTH]; blk 1698 fs/xfs/libxfs/xfs_dir2_leaf.c lbp = state->path.blk[0].bp; blk 1778 fs/xfs/libxfs/xfs_dir2_leaf.c state->path.blk[0].bp = NULL; blk 1426 fs/xfs/libxfs/xfs_dir2_node.c xfs_da_state_blk_t *blk; /* leaf block */ blk 1445 fs/xfs/libxfs/xfs_dir2_node.c blk = &state->path.blk[state->path.active - 1]; blk 1446 fs/xfs/libxfs/xfs_dir2_node.c leaf = blk->bp->b_addr; blk 1449 fs/xfs/libxfs/xfs_dir2_node.c xfs_dir3_leaf_check(dp, blk->bp); blk 1535 fs/xfs/libxfs/xfs_dir2_node.c if (blkno < blk->blkno) blk 1965 fs/xfs/libxfs/xfs_dir2_node.c xfs_da_state_blk_t *blk; /* leaf block for insert */ blk 1997 fs/xfs/libxfs/xfs_dir2_node.c blk = &state->path.blk[state->path.active - 1]; blk 1998 fs/xfs/libxfs/xfs_dir2_node.c ASSERT(blk->magic == XFS_DIR2_LEAFN_MAGIC); blk 2002 fs/xfs/libxfs/xfs_dir2_node.c rval = xfs_dir2_leafn_add(blk->bp, args, blk->index); blk 2068 fs/xfs/libxfs/xfs_dir2_node.c xfs_trans_brelse(args->trans, state->path.blk[i].bp); blk 2069 fs/xfs/libxfs/xfs_dir2_node.c state->path.blk[i].bp = NULL; blk 2089 fs/xfs/libxfs/xfs_dir2_node.c struct xfs_da_state_blk *blk; /* leaf block */ blk 2114 fs/xfs/libxfs/xfs_dir2_node.c blk = &state->path.blk[state->path.active - 1]; blk 2115 fs/xfs/libxfs/xfs_dir2_node.c ASSERT(blk->magic == XFS_DIR2_LEAFN_MAGIC); blk 2121 fs/xfs/libxfs/xfs_dir2_node.c error = xfs_dir2_leafn_remove(args, blk->bp, blk->index, blk 2151 fs/xfs/libxfs/xfs_dir2_node.c xfs_da_state_blk_t *blk; /* leaf block */ blk 2195 fs/xfs/libxfs/xfs_dir2_node.c blk = &state->path.blk[state->path.active - 1]; blk 2196 fs/xfs/libxfs/xfs_dir2_node.c ASSERT(blk->magic == XFS_DIR2_LEAFN_MAGIC); blk 2197 fs/xfs/libxfs/xfs_dir2_node.c leaf = blk->bp->b_addr; blk 2199 fs/xfs/libxfs/xfs_dir2_node.c lep = &ents[blk->index]; blk 2231 fs/xfs/libxfs/xfs_dir2_node.c xfs_trans_brelse(args->trans, state->path.blk[i].bp); blk 2232 fs/xfs/libxfs/xfs_dir2_node.c state->path.blk[i].bp = NULL; blk 308 fs/xfs/scrub/attr.c struct xfs_da_state_blk *blk = &ds->state->path.blk[level]; blk 309 fs/xfs/scrub/attr.c struct xfs_buf *bp = blk->bp; blk 323 fs/xfs/scrub/attr.c if (*last_checked == blk->blkno) blk 334 fs/xfs/scrub/attr.c *last_checked = blk->blkno; blk 406 fs/xfs/scrub/attr.c struct xfs_da_state_blk *blk; blk 417 fs/xfs/scrub/attr.c blk = &ds->state->path.blk[level]; blk 432 fs/xfs/scrub/attr.c bp = blk->bp; blk 54 fs/xfs/scrub/dabtree.c ds->state->path.blk[level].blkno), blk 76 fs/xfs/scrub/dabtree.c ds->state->path.blk[level].blkno), blk 88 fs/xfs/scrub/dabtree.c struct xfs_da_state_blk *blk; blk 92 fs/xfs/scrub/dabtree.c blk = &ds->state->path.blk[level]; blk 93 fs/xfs/scrub/dabtree.c baddr = blk->bp->b_addr; blk 94 fs/xfs/scrub/dabtree.c switch (blk->magic) { blk 138 fs/xfs/scrub/dabtree.c blks = ds->state->path.blk; blk 274 fs/xfs/scrub/dabtree.c if (ds->state->altpath.blk[level].bp) blk 276 fs/xfs/scrub/dabtree.c ds->state->altpath.blk[level].bp); blk 279 fs/xfs/scrub/dabtree.c if (ds->state->altpath.blk[level].blkno != sibling) blk 281 fs/xfs/scrub/dabtree.c if (ds->state->altpath.blk[level].bp) { blk 283 fs/xfs/scrub/dabtree.c ds->state->altpath.blk[level].bp); blk 284 fs/xfs/scrub/dabtree.c ds->state->altpath.blk[level].bp = NULL; blk 332 fs/xfs/scrub/dabtree.c struct xfs_da_state_blk *blk; blk 343 fs/xfs/scrub/dabtree.c blk = &ds->state->path.blk[level]; blk 347 fs/xfs/scrub/dabtree.c if (blk->bp) { blk 348 fs/xfs/scrub/dabtree.c xfs_trans_brelse(dargs->trans, blk->bp); blk 349 fs/xfs/scrub/dabtree.c blk->bp = NULL; blk 353 fs/xfs/scrub/dabtree.c blk->blkno = blkno; blk 358 fs/xfs/scrub/dabtree.c error = xfs_da_read_buf(dargs->trans, dargs->dp, blk->blkno, -2, blk 359 fs/xfs/scrub/dabtree.c &blk->bp, dargs->whichfork, blk 363 fs/xfs/scrub/dabtree.c if (blk->bp) blk 364 fs/xfs/scrub/dabtree.c xchk_buffer_recheck(ds->sc, blk->bp); blk 372 fs/xfs/scrub/dabtree.c blk->bp == NULL) blk 376 fs/xfs/scrub/dabtree.c if (blk->bp == NULL) { blk 381 fs/xfs/scrub/dabtree.c hdr3 = blk->bp->b_addr; blk 382 fs/xfs/scrub/dabtree.c blk->magic = be16_to_cpu(hdr3->hdr.magic); blk 402 fs/xfs/scrub/dabtree.c switch (blk->magic) { blk 405 fs/xfs/scrub/dabtree.c xfs_trans_buf_set_type(dargs->trans, blk->bp, blk 407 fs/xfs/scrub/dabtree.c blk->magic = XFS_ATTR_LEAF_MAGIC; blk 408 fs/xfs/scrub/dabtree.c blk->hashval = xfs_attr_leaf_lasthash(blk->bp, pmaxrecs); blk 414 fs/xfs/scrub/dabtree.c xfs_trans_buf_set_type(dargs->trans, blk->bp, blk 416 fs/xfs/scrub/dabtree.c blk->magic = XFS_DIR2_LEAFN_MAGIC; blk 417 fs/xfs/scrub/dabtree.c blk->hashval = xfs_dir2_leaf_lasthash(ip, blk->bp, pmaxrecs); blk 423 fs/xfs/scrub/dabtree.c xfs_trans_buf_set_type(dargs->trans, blk->bp, blk 425 fs/xfs/scrub/dabtree.c blk->magic = XFS_DIR2_LEAF1_MAGIC; blk 426 fs/xfs/scrub/dabtree.c blk->hashval = xfs_dir2_leaf_lasthash(ip, blk->bp, pmaxrecs); blk 432 fs/xfs/scrub/dabtree.c xfs_trans_buf_set_type(dargs->trans, blk->bp, blk 434 fs/xfs/scrub/dabtree.c blk->magic = XFS_DA_NODE_MAGIC; blk 435 fs/xfs/scrub/dabtree.c node = blk->bp->b_addr; blk 439 fs/xfs/scrub/dabtree.c blk->hashval = be32_to_cpu(btree[*pmaxrecs - 1].hashval); blk 463 fs/xfs/scrub/dabtree.c xfs_trans_brelse(dargs->trans, blk->bp); blk 464 fs/xfs/scrub/dabtree.c blk->bp = NULL; blk 466 fs/xfs/scrub/dabtree.c blk->blkno = 0; blk 515 fs/xfs/scrub/dabtree.c blks = ds.state->path.blk; blk 2192 fs/xfs/xfs_log_recover.c void *blk = bp->b_addr; blk 2200 fs/xfs/xfs_log_recover.c magic32 = be32_to_cpu(*(__be32 *)blk); blk 2210 fs/xfs/xfs_log_recover.c struct xfs_btree_block *btb = blk; blk 2218 fs/xfs/xfs_log_recover.c struct xfs_btree_block *btb = blk; blk 2225 fs/xfs/xfs_log_recover.c lsn = be64_to_cpu(((struct xfs_agf *)blk)->agf_lsn); blk 2226 fs/xfs/xfs_log_recover.c uuid = &((struct xfs_agf *)blk)->agf_uuid; blk 2229 fs/xfs/xfs_log_recover.c lsn = be64_to_cpu(((struct xfs_agfl *)blk)->agfl_lsn); blk 2230 fs/xfs/xfs_log_recover.c uuid = &((struct xfs_agfl *)blk)->agfl_uuid; blk 2233 fs/xfs/xfs_log_recover.c lsn = be64_to_cpu(((struct xfs_agi *)blk)->agi_lsn); blk 2234 fs/xfs/xfs_log_recover.c uuid = &((struct xfs_agi *)blk)->agi_uuid; blk 2237 fs/xfs/xfs_log_recover.c lsn = be64_to_cpu(((struct xfs_dsymlink_hdr *)blk)->sl_lsn); blk 2238 fs/xfs/xfs_log_recover.c uuid = &((struct xfs_dsymlink_hdr *)blk)->sl_uuid; blk 2243 fs/xfs/xfs_log_recover.c lsn = be64_to_cpu(((struct xfs_dir3_blk_hdr *)blk)->lsn); blk 2244 fs/xfs/xfs_log_recover.c uuid = &((struct xfs_dir3_blk_hdr *)blk)->uuid; blk 2263 fs/xfs/xfs_log_recover.c lsn = be64_to_cpu(((struct xfs_dsb *)blk)->sb_lsn); blk 2265 fs/xfs/xfs_log_recover.c uuid = &((struct xfs_dsb *)blk)->sb_meta_uuid; blk 2267 fs/xfs/xfs_log_recover.c uuid = &((struct xfs_dsb *)blk)->sb_uuid; blk 2279 fs/xfs/xfs_log_recover.c magicda = be16_to_cpu(((struct xfs_da_blkinfo *)blk)->magic); blk 2284 fs/xfs/xfs_log_recover.c lsn = be64_to_cpu(((struct xfs_da3_blkinfo *)blk)->lsn); blk 2285 fs/xfs/xfs_log_recover.c uuid = &((struct xfs_da3_blkinfo *)blk)->uuid; blk 2308 fs/xfs/xfs_log_recover.c magic16 = be16_to_cpu(*(__be16 *)blk); blk 28 include/linux/f2fs_fs.h #define F2FS_BLK_TO_BYTES(blk) ((blk) << F2FS_BLKSIZE_BITS) blk 202 include/linux/f2fs_fs.h __le32 blk; /* start block address of the extent */ blk 47 include/linux/lightnvm.h u64 blk : NVM_GEN_BLK_BITS; blk 55 include/linux/lightnvm.h u64 blk : NVM_GEN_BLK_BITS; blk 451 include/linux/lightnvm.h l.ppa |= ((u64)r.g.blk) << ppaf->blk_offset; blk 480 include/linux/lightnvm.h l.g.blk = (r.ppa & ppaf->blk_mask) >> ppaf->blk_offset; blk 537 include/linux/lightnvm.h ppa64.g.blk = (ppa32 & ppaf->blk_mask) >> blk 580 include/linux/lightnvm.h ppa32 |= ppa64.g.blk << ppaf->blk_offset; blk 1556 include/sound/emu10k1.h #define snd_emu10k1_memblk_offset(blk) (((blk)->mapped_page << PAGE_SHIFT) | ((blk)->mem.offset & (PAGE_SIZE - 1))) blk 1865 include/sound/emu10k1.h int snd_emu10k1_free_pages(struct snd_emu10k1 *emu, struct snd_util_memblk *blk); blk 1869 include/sound/emu10k1.h int snd_emu10k1_synth_free(struct snd_emu10k1 *emu, struct snd_util_memblk *blk); blk 1870 include/sound/emu10k1.h int snd_emu10k1_synth_bzero(struct snd_emu10k1 *emu, struct snd_util_memblk *blk, int offset, int size); blk 1871 include/sound/emu10k1.h int snd_emu10k1_synth_copy_from_user(struct snd_emu10k1 *emu, struct snd_util_memblk *blk, int offset, const char __user *data, int size); blk 1872 include/sound/emu10k1.h int snd_emu10k1_memblk_map(struct snd_emu10k1 *emu, struct snd_emu10k1_memblk *blk); blk 21 include/sound/util_mem.h #define snd_util_memblk_argptr(blk) (void*)((char*)(blk) + sizeof(struct snd_util_memblk)) blk 41 include/sound/util_mem.h int snd_util_mem_free(struct snd_util_memhdr *hdr, struct snd_util_memblk *blk); blk 46 include/sound/util_mem.h void __snd_util_mem_free(struct snd_util_memhdr *hdr, struct snd_util_memblk *blk); blk 610 include/trace/events/ext4.h TP_PROTO(struct super_block *sb, unsigned long long blk, blk 613 include/trace/events/ext4.h TP_ARGS(sb, blk, count), blk 617 include/trace/events/ext4.h __field( __u64, blk ) blk 624 include/trace/events/ext4.h __entry->blk = blk; blk 630 include/trace/events/ext4.h __entry->blk, __entry->count) blk 1550 include/trace/events/f2fs.h __field(u32, blk) blk 1559 include/trace/events/f2fs.h __entry->blk = ei->blk; blk 1568 include/trace/events/f2fs.h __entry->blk, blk 1583 include/trace/events/f2fs.h __field(u32, blk) blk 1591 include/trace/events/f2fs.h __entry->blk = blkaddr; blk 1599 include/trace/events/f2fs.h __entry->blk, blk 86 sound/pci/ctxfi/cthardware.h int (*src_rsc_put_ctrl_blk)(void *blk); blk 87 sound/pci/ctxfi/cthardware.h int (*src_set_state)(void *blk, unsigned int state); blk 88 sound/pci/ctxfi/cthardware.h int (*src_set_bm)(void *blk, unsigned int bm); blk 89 sound/pci/ctxfi/cthardware.h int (*src_set_rsr)(void *blk, unsigned int rsr); blk 90 sound/pci/ctxfi/cthardware.h int (*src_set_sf)(void *blk, unsigned int sf); blk 91 sound/pci/ctxfi/cthardware.h int (*src_set_wr)(void *blk, unsigned int wr); blk 92 sound/pci/ctxfi/cthardware.h int (*src_set_pm)(void *blk, unsigned int pm); blk 93 sound/pci/ctxfi/cthardware.h int (*src_set_rom)(void *blk, unsigned int rom); blk 94 sound/pci/ctxfi/cthardware.h int (*src_set_vo)(void *blk, unsigned int vo); blk 95 sound/pci/ctxfi/cthardware.h int (*src_set_st)(void *blk, unsigned int st); blk 96 sound/pci/ctxfi/cthardware.h int (*src_set_ie)(void *blk, unsigned int ie); blk 97 sound/pci/ctxfi/cthardware.h int (*src_set_ilsz)(void *blk, unsigned int ilsz); blk 98 sound/pci/ctxfi/cthardware.h int (*src_set_bp)(void *blk, unsigned int bp); blk 99 sound/pci/ctxfi/cthardware.h int (*src_set_cisz)(void *blk, unsigned int cisz); blk 100 sound/pci/ctxfi/cthardware.h int (*src_set_ca)(void *blk, unsigned int ca); blk 101 sound/pci/ctxfi/cthardware.h int (*src_set_sa)(void *blk, unsigned int sa); blk 102 sound/pci/ctxfi/cthardware.h int (*src_set_la)(void *blk, unsigned int la); blk 103 sound/pci/ctxfi/cthardware.h int (*src_set_pitch)(void *blk, unsigned int pitch); blk 104 sound/pci/ctxfi/cthardware.h int (*src_set_clear_zbufs)(void *blk, unsigned int clear); blk 105 sound/pci/ctxfi/cthardware.h int (*src_set_dirty)(void *blk, unsigned int flags); blk 106 sound/pci/ctxfi/cthardware.h int (*src_set_dirty_all)(void *blk); blk 107 sound/pci/ctxfi/cthardware.h int (*src_commit_write)(struct hw *hw, unsigned int idx, void *blk); blk 108 sound/pci/ctxfi/cthardware.h int (*src_get_ca)(struct hw *hw, unsigned int idx, void *blk); blk 109 sound/pci/ctxfi/cthardware.h unsigned int (*src_get_dirty)(void *blk); blk 112 sound/pci/ctxfi/cthardware.h int (*src_mgr_put_ctrl_blk)(void *blk); blk 114 sound/pci/ctxfi/cthardware.h int (*src_mgr_enbs_src)(void *blk, unsigned int idx); blk 116 sound/pci/ctxfi/cthardware.h int (*src_mgr_enb_src)(void *blk, unsigned int idx); blk 118 sound/pci/ctxfi/cthardware.h int (*src_mgr_dsb_src)(void *blk, unsigned int idx); blk 119 sound/pci/ctxfi/cthardware.h int (*src_mgr_commit_write)(struct hw *hw, void *blk); blk 123 sound/pci/ctxfi/cthardware.h int (*srcimp_mgr_put_ctrl_blk)(void *blk); blk 124 sound/pci/ctxfi/cthardware.h int (*srcimp_mgr_set_imaparc)(void *blk, unsigned int slot); blk 125 sound/pci/ctxfi/cthardware.h int (*srcimp_mgr_set_imapuser)(void *blk, unsigned int user); blk 126 sound/pci/ctxfi/cthardware.h int (*srcimp_mgr_set_imapnxt)(void *blk, unsigned int next); blk 127 sound/pci/ctxfi/cthardware.h int (*srcimp_mgr_set_imapaddr)(void *blk, unsigned int addr); blk 128 sound/pci/ctxfi/cthardware.h int (*srcimp_mgr_commit_write)(struct hw *hw, void *blk); blk 132 sound/pci/ctxfi/cthardware.h int (*amixer_rsc_put_ctrl_blk)(void *blk); blk 134 sound/pci/ctxfi/cthardware.h int (*amixer_mgr_put_ctrl_blk)(void *blk); blk 135 sound/pci/ctxfi/cthardware.h int (*amixer_set_mode)(void *blk, unsigned int mode); blk 136 sound/pci/ctxfi/cthardware.h int (*amixer_set_iv)(void *blk, unsigned int iv); blk 137 sound/pci/ctxfi/cthardware.h int (*amixer_set_x)(void *blk, unsigned int x); blk 138 sound/pci/ctxfi/cthardware.h int (*amixer_set_y)(void *blk, unsigned int y); blk 139 sound/pci/ctxfi/cthardware.h int (*amixer_set_sadr)(void *blk, unsigned int sadr); blk 140 sound/pci/ctxfi/cthardware.h int (*amixer_set_se)(void *blk, unsigned int se); blk 141 sound/pci/ctxfi/cthardware.h int (*amixer_set_dirty)(void *blk, unsigned int flags); blk 142 sound/pci/ctxfi/cthardware.h int (*amixer_set_dirty_all)(void *blk); blk 143 sound/pci/ctxfi/cthardware.h int (*amixer_commit_write)(struct hw *hw, unsigned int idx, void *blk); blk 144 sound/pci/ctxfi/cthardware.h int (*amixer_get_y)(void *blk); blk 145 sound/pci/ctxfi/cthardware.h unsigned int (*amixer_get_dirty)(void *blk); blk 149 sound/pci/ctxfi/cthardware.h int (*dai_put_ctrl_blk)(void *blk); blk 150 sound/pci/ctxfi/cthardware.h int (*dai_srt_set_srco)(void *blk, unsigned int src); blk 151 sound/pci/ctxfi/cthardware.h int (*dai_srt_set_srcm)(void *blk, unsigned int src); blk 152 sound/pci/ctxfi/cthardware.h int (*dai_srt_set_rsr)(void *blk, unsigned int rsr); blk 153 sound/pci/ctxfi/cthardware.h int (*dai_srt_set_drat)(void *blk, unsigned int drat); blk 154 sound/pci/ctxfi/cthardware.h int (*dai_srt_set_ec)(void *blk, unsigned int ec); blk 155 sound/pci/ctxfi/cthardware.h int (*dai_srt_set_et)(void *blk, unsigned int et); blk 156 sound/pci/ctxfi/cthardware.h int (*dai_commit_write)(struct hw *hw, unsigned int idx, void *blk); blk 158 sound/pci/ctxfi/cthardware.h int (*dao_put_ctrl_blk)(void *blk); blk 159 sound/pci/ctxfi/cthardware.h int (*dao_set_spos)(void *blk, unsigned int spos); blk 160 sound/pci/ctxfi/cthardware.h int (*dao_commit_write)(struct hw *hw, unsigned int idx, void *blk); blk 161 sound/pci/ctxfi/cthardware.h int (*dao_get_spos)(void *blk, unsigned int *spos); blk 164 sound/pci/ctxfi/cthardware.h int (*daio_mgr_put_ctrl_blk)(void *blk); blk 165 sound/pci/ctxfi/cthardware.h int (*daio_mgr_enb_dai)(void *blk, unsigned int idx); blk 166 sound/pci/ctxfi/cthardware.h int (*daio_mgr_dsb_dai)(void *blk, unsigned int idx); blk 167 sound/pci/ctxfi/cthardware.h int (*daio_mgr_enb_dao)(void *blk, unsigned int idx); blk 168 sound/pci/ctxfi/cthardware.h int (*daio_mgr_dsb_dao)(void *blk, unsigned int idx); blk 169 sound/pci/ctxfi/cthardware.h int (*daio_mgr_dao_init)(void *blk, unsigned int idx, blk 171 sound/pci/ctxfi/cthardware.h int (*daio_mgr_set_imaparc)(void *blk, unsigned int slot); blk 172 sound/pci/ctxfi/cthardware.h int (*daio_mgr_set_imapnxt)(void *blk, unsigned int next); blk 173 sound/pci/ctxfi/cthardware.h int (*daio_mgr_set_imapaddr)(void *blk, unsigned int addr); blk 174 sound/pci/ctxfi/cthardware.h int (*daio_mgr_commit_write)(struct hw *hw, void *blk); blk 157 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *blk; blk 160 sound/pci/ctxfi/cthw20k1.c blk = kzalloc(sizeof(*blk), GFP_KERNEL); blk 161 sound/pci/ctxfi/cthw20k1.c if (!blk) blk 164 sound/pci/ctxfi/cthw20k1.c *rblk = blk; blk 169 sound/pci/ctxfi/cthw20k1.c static int src_put_rsc_ctrl_blk(void *blk) blk 171 sound/pci/ctxfi/cthw20k1.c kfree((struct src_rsc_ctrl_blk *)blk); blk 176 sound/pci/ctxfi/cthw20k1.c static int src_set_state(void *blk, unsigned int state) blk 178 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 185 sound/pci/ctxfi/cthw20k1.c static int src_set_bm(void *blk, unsigned int bm) blk 187 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 194 sound/pci/ctxfi/cthw20k1.c static int src_set_rsr(void *blk, unsigned int rsr) blk 196 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 203 sound/pci/ctxfi/cthw20k1.c static int src_set_sf(void *blk, unsigned int sf) blk 205 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 212 sound/pci/ctxfi/cthw20k1.c static int src_set_wr(void *blk, unsigned int wr) blk 214 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 221 sound/pci/ctxfi/cthw20k1.c static int src_set_pm(void *blk, unsigned int pm) blk 223 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 230 sound/pci/ctxfi/cthw20k1.c static int src_set_rom(void *blk, unsigned int rom) blk 232 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 239 sound/pci/ctxfi/cthw20k1.c static int src_set_vo(void *blk, unsigned int vo) blk 241 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 248 sound/pci/ctxfi/cthw20k1.c static int src_set_st(void *blk, unsigned int st) blk 250 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 257 sound/pci/ctxfi/cthw20k1.c static int src_set_ie(void *blk, unsigned int ie) blk 259 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 266 sound/pci/ctxfi/cthw20k1.c static int src_set_ilsz(void *blk, unsigned int ilsz) blk 268 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 275 sound/pci/ctxfi/cthw20k1.c static int src_set_bp(void *blk, unsigned int bp) blk 277 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 284 sound/pci/ctxfi/cthw20k1.c static int src_set_cisz(void *blk, unsigned int cisz) blk 286 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 293 sound/pci/ctxfi/cthw20k1.c static int src_set_ca(void *blk, unsigned int ca) blk 295 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 302 sound/pci/ctxfi/cthw20k1.c static int src_set_sa(void *blk, unsigned int sa) blk 304 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 311 sound/pci/ctxfi/cthw20k1.c static int src_set_la(void *blk, unsigned int la) blk 313 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 320 sound/pci/ctxfi/cthw20k1.c static int src_set_pitch(void *blk, unsigned int pitch) blk 322 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 329 sound/pci/ctxfi/cthw20k1.c static int src_set_clear_zbufs(void *blk, unsigned int clear) blk 331 sound/pci/ctxfi/cthw20k1.c ((struct src_rsc_ctrl_blk *)blk)->dirty.bf.czbfs = (clear ? 1 : 0); blk 335 sound/pci/ctxfi/cthw20k1.c static int src_set_dirty(void *blk, unsigned int flags) blk 337 sound/pci/ctxfi/cthw20k1.c ((struct src_rsc_ctrl_blk *)blk)->dirty.data = (flags & 0xffff); blk 341 sound/pci/ctxfi/cthw20k1.c static int src_set_dirty_all(void *blk) blk 343 sound/pci/ctxfi/cthw20k1.c ((struct src_rsc_ctrl_blk *)blk)->dirty.data = ~(0x0); blk 359 sound/pci/ctxfi/cthw20k1.c static int src_commit_write(struct hw *hw, unsigned int idx, void *blk) blk 361 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 416 sound/pci/ctxfi/cthw20k1.c static int src_get_ca(struct hw *hw, unsigned int idx, void *blk) blk 418 sound/pci/ctxfi/cthw20k1.c struct src_rsc_ctrl_blk *ctl = blk; blk 426 sound/pci/ctxfi/cthw20k1.c static unsigned int src_get_dirty(void *blk) blk 428 sound/pci/ctxfi/cthw20k1.c return ((struct src_rsc_ctrl_blk *)blk)->dirty.data; blk 436 sound/pci/ctxfi/cthw20k1.c static int src_mgr_enbs_src(void *blk, unsigned int idx) blk 438 sound/pci/ctxfi/cthw20k1.c ((struct src_mgr_ctrl_blk *)blk)->enbsa = ~(0x0); blk 439 sound/pci/ctxfi/cthw20k1.c ((struct src_mgr_ctrl_blk *)blk)->dirty.bf.enbsa = 1; blk 440 sound/pci/ctxfi/cthw20k1.c ((struct src_mgr_ctrl_blk *)blk)->enb[idx/32] |= (0x1 << (idx%32)); blk 444 sound/pci/ctxfi/cthw20k1.c static int src_mgr_enb_src(void *blk, unsigned int idx) blk 446 sound/pci/ctxfi/cthw20k1.c ((struct src_mgr_ctrl_blk *)blk)->enb[idx/32] |= (0x1 << (idx%32)); blk 447 sound/pci/ctxfi/cthw20k1.c ((struct src_mgr_ctrl_blk *)blk)->dirty.data |= (0x1 << (idx/32)); blk 451 sound/pci/ctxfi/cthw20k1.c static int src_mgr_dsb_src(void *blk, unsigned int idx) blk 453 sound/pci/ctxfi/cthw20k1.c ((struct src_mgr_ctrl_blk *)blk)->enb[idx/32] &= ~(0x1 << (idx%32)); blk 454 sound/pci/ctxfi/cthw20k1.c ((struct src_mgr_ctrl_blk *)blk)->dirty.data |= (0x1 << (idx/32)); blk 458 sound/pci/ctxfi/cthw20k1.c static int src_mgr_commit_write(struct hw *hw, void *blk) blk 460 sound/pci/ctxfi/cthw20k1.c struct src_mgr_ctrl_blk *ctl = blk; blk 483 sound/pci/ctxfi/cthw20k1.c struct src_mgr_ctrl_blk *blk; blk 486 sound/pci/ctxfi/cthw20k1.c blk = kzalloc(sizeof(*blk), GFP_KERNEL); blk 487 sound/pci/ctxfi/cthw20k1.c if (!blk) blk 490 sound/pci/ctxfi/cthw20k1.c *rblk = blk; blk 495 sound/pci/ctxfi/cthw20k1.c static int src_mgr_put_ctrl_blk(void *blk) blk 497 sound/pci/ctxfi/cthw20k1.c kfree((struct src_mgr_ctrl_blk *)blk); blk 504 sound/pci/ctxfi/cthw20k1.c struct srcimp_mgr_ctrl_blk *blk; blk 507 sound/pci/ctxfi/cthw20k1.c blk = kzalloc(sizeof(*blk), GFP_KERNEL); blk 508 sound/pci/ctxfi/cthw20k1.c if (!blk) blk 511 sound/pci/ctxfi/cthw20k1.c *rblk = blk; blk 516 sound/pci/ctxfi/cthw20k1.c static int srcimp_mgr_put_ctrl_blk(void *blk) blk 518 sound/pci/ctxfi/cthw20k1.c kfree((struct srcimp_mgr_ctrl_blk *)blk); blk 523 sound/pci/ctxfi/cthw20k1.c static int srcimp_mgr_set_imaparc(void *blk, unsigned int slot) blk 525 sound/pci/ctxfi/cthw20k1.c struct srcimp_mgr_ctrl_blk *ctl = blk; blk 532 sound/pci/ctxfi/cthw20k1.c static int srcimp_mgr_set_imapuser(void *blk, unsigned int user) blk 534 sound/pci/ctxfi/cthw20k1.c struct srcimp_mgr_ctrl_blk *ctl = blk; blk 541 sound/pci/ctxfi/cthw20k1.c static int srcimp_mgr_set_imapnxt(void *blk, unsigned int next) blk 543 sound/pci/ctxfi/cthw20k1.c struct srcimp_mgr_ctrl_blk *ctl = blk; blk 550 sound/pci/ctxfi/cthw20k1.c static int srcimp_mgr_set_imapaddr(void *blk, unsigned int addr) blk 552 sound/pci/ctxfi/cthw20k1.c struct srcimp_mgr_ctrl_blk *ctl = blk; blk 559 sound/pci/ctxfi/cthw20k1.c static int srcimp_mgr_commit_write(struct hw *hw, void *blk) blk 561 sound/pci/ctxfi/cthw20k1.c struct srcimp_mgr_ctrl_blk *ctl = blk; blk 600 sound/pci/ctxfi/cthw20k1.c static int amixer_set_mode(void *blk, unsigned int mode) blk 602 sound/pci/ctxfi/cthw20k1.c struct amixer_rsc_ctrl_blk *ctl = blk; blk 609 sound/pci/ctxfi/cthw20k1.c static int amixer_set_iv(void *blk, unsigned int iv) blk 615 sound/pci/ctxfi/cthw20k1.c static int amixer_set_x(void *blk, unsigned int x) blk 617 sound/pci/ctxfi/cthw20k1.c struct amixer_rsc_ctrl_blk *ctl = blk; blk 624 sound/pci/ctxfi/cthw20k1.c static int amixer_set_y(void *blk, unsigned int y) blk 626 sound/pci/ctxfi/cthw20k1.c struct amixer_rsc_ctrl_blk *ctl = blk; blk 633 sound/pci/ctxfi/cthw20k1.c static int amixer_set_sadr(void *blk, unsigned int sadr) blk 635 sound/pci/ctxfi/cthw20k1.c struct amixer_rsc_ctrl_blk *ctl = blk; blk 642 sound/pci/ctxfi/cthw20k1.c static int amixer_set_se(void *blk, unsigned int se) blk 644 sound/pci/ctxfi/cthw20k1.c struct amixer_rsc_ctrl_blk *ctl = blk; blk 651 sound/pci/ctxfi/cthw20k1.c static int amixer_set_dirty(void *blk, unsigned int flags) blk 653 sound/pci/ctxfi/cthw20k1.c ((struct amixer_rsc_ctrl_blk *)blk)->dirty.data = (flags & 0xffff); blk 657 sound/pci/ctxfi/cthw20k1.c static int amixer_set_dirty_all(void *blk) blk 659 sound/pci/ctxfi/cthw20k1.c ((struct amixer_rsc_ctrl_blk *)blk)->dirty.data = ~(0x0); blk 663 sound/pci/ctxfi/cthw20k1.c static int amixer_commit_write(struct hw *hw, unsigned int idx, void *blk) blk 665 sound/pci/ctxfi/cthw20k1.c struct amixer_rsc_ctrl_blk *ctl = blk; blk 677 sound/pci/ctxfi/cthw20k1.c static int amixer_get_y(void *blk) blk 679 sound/pci/ctxfi/cthw20k1.c struct amixer_rsc_ctrl_blk *ctl = blk; blk 684 sound/pci/ctxfi/cthw20k1.c static unsigned int amixer_get_dirty(void *blk) blk 686 sound/pci/ctxfi/cthw20k1.c return ((struct amixer_rsc_ctrl_blk *)blk)->dirty.data; blk 691 sound/pci/ctxfi/cthw20k1.c struct amixer_rsc_ctrl_blk *blk; blk 694 sound/pci/ctxfi/cthw20k1.c blk = kzalloc(sizeof(*blk), GFP_KERNEL); blk 695 sound/pci/ctxfi/cthw20k1.c if (!blk) blk 698 sound/pci/ctxfi/cthw20k1.c *rblk = blk; blk 703 sound/pci/ctxfi/cthw20k1.c static int amixer_rsc_put_ctrl_blk(void *blk) blk 705 sound/pci/ctxfi/cthw20k1.c kfree((struct amixer_rsc_ctrl_blk *)blk); blk 724 sound/pci/ctxfi/cthw20k1.c static int amixer_mgr_put_ctrl_blk(void *blk) blk 824 sound/pci/ctxfi/cthw20k1.c static int dai_srt_set_srcr(void *blk, unsigned int src) blk 826 sound/pci/ctxfi/cthw20k1.c struct dai_ctrl_blk *ctl = blk; blk 833 sound/pci/ctxfi/cthw20k1.c static int dai_srt_set_srcl(void *blk, unsigned int src) blk 835 sound/pci/ctxfi/cthw20k1.c struct dai_ctrl_blk *ctl = blk; blk 842 sound/pci/ctxfi/cthw20k1.c static int dai_srt_set_rsr(void *blk, unsigned int rsr) blk 844 sound/pci/ctxfi/cthw20k1.c struct dai_ctrl_blk *ctl = blk; blk 851 sound/pci/ctxfi/cthw20k1.c static int dai_srt_set_drat(void *blk, unsigned int drat) blk 853 sound/pci/ctxfi/cthw20k1.c struct dai_ctrl_blk *ctl = blk; blk 860 sound/pci/ctxfi/cthw20k1.c static int dai_srt_set_ec(void *blk, unsigned int ec) blk 862 sound/pci/ctxfi/cthw20k1.c struct dai_ctrl_blk *ctl = blk; blk 869 sound/pci/ctxfi/cthw20k1.c static int dai_srt_set_et(void *blk, unsigned int et) blk 871 sound/pci/ctxfi/cthw20k1.c struct dai_ctrl_blk *ctl = blk; blk 878 sound/pci/ctxfi/cthw20k1.c static int dai_commit_write(struct hw *hw, unsigned int idx, void *blk) blk 880 sound/pci/ctxfi/cthw20k1.c struct dai_ctrl_blk *ctl = blk; blk 898 sound/pci/ctxfi/cthw20k1.c struct dai_ctrl_blk *blk; blk 901 sound/pci/ctxfi/cthw20k1.c blk = kzalloc(sizeof(*blk), GFP_KERNEL); blk 902 sound/pci/ctxfi/cthw20k1.c if (!blk) blk 905 sound/pci/ctxfi/cthw20k1.c *rblk = blk; blk 910 sound/pci/ctxfi/cthw20k1.c static int dai_put_ctrl_blk(void *blk) blk 912 sound/pci/ctxfi/cthw20k1.c kfree((struct dai_ctrl_blk *)blk); blk 917 sound/pci/ctxfi/cthw20k1.c static int dao_set_spos(void *blk, unsigned int spos) blk 919 sound/pci/ctxfi/cthw20k1.c ((struct dao_ctrl_blk *)blk)->spos = spos; blk 920 sound/pci/ctxfi/cthw20k1.c ((struct dao_ctrl_blk *)blk)->dirty.bf.spos = 1; blk 924 sound/pci/ctxfi/cthw20k1.c static int dao_commit_write(struct hw *hw, unsigned int idx, void *blk) blk 926 sound/pci/ctxfi/cthw20k1.c struct dao_ctrl_blk *ctl = blk; blk 939 sound/pci/ctxfi/cthw20k1.c static int dao_get_spos(void *blk, unsigned int *spos) blk 941 sound/pci/ctxfi/cthw20k1.c *spos = ((struct dao_ctrl_blk *)blk)->spos; blk 947 sound/pci/ctxfi/cthw20k1.c struct dao_ctrl_blk *blk; blk 950 sound/pci/ctxfi/cthw20k1.c blk = kzalloc(sizeof(*blk), GFP_KERNEL); blk 951 sound/pci/ctxfi/cthw20k1.c if (!blk) blk 954 sound/pci/ctxfi/cthw20k1.c *rblk = blk; blk 959 sound/pci/ctxfi/cthw20k1.c static int dao_put_ctrl_blk(void *blk) blk 961 sound/pci/ctxfi/cthw20k1.c kfree((struct dao_ctrl_blk *)blk); blk 966 sound/pci/ctxfi/cthw20k1.c static int daio_mgr_enb_dai(void *blk, unsigned int idx) blk 968 sound/pci/ctxfi/cthw20k1.c struct daio_mgr_ctrl_blk *ctl = blk; blk 983 sound/pci/ctxfi/cthw20k1.c static int daio_mgr_dsb_dai(void *blk, unsigned int idx) blk 985 sound/pci/ctxfi/cthw20k1.c struct daio_mgr_ctrl_blk *ctl = blk; blk 1000 sound/pci/ctxfi/cthw20k1.c static int daio_mgr_enb_dao(void *blk, unsigned int idx) blk 1002 sound/pci/ctxfi/cthw20k1.c struct daio_mgr_ctrl_blk *ctl = blk; blk 1017 sound/pci/ctxfi/cthw20k1.c static int daio_mgr_dsb_dao(void *blk, unsigned int idx) blk 1019 sound/pci/ctxfi/cthw20k1.c struct daio_mgr_ctrl_blk *ctl = blk; blk 1034 sound/pci/ctxfi/cthw20k1.c static int daio_mgr_dao_init(void *blk, unsigned int idx, unsigned int conf) blk 1036 sound/pci/ctxfi/cthw20k1.c struct daio_mgr_ctrl_blk *ctl = blk; blk 1071 sound/pci/ctxfi/cthw20k1.c static int daio_mgr_set_imaparc(void *blk, unsigned int slot) blk 1073 sound/pci/ctxfi/cthw20k1.c struct daio_mgr_ctrl_blk *ctl = blk; blk 1080 sound/pci/ctxfi/cthw20k1.c static int daio_mgr_set_imapnxt(void *blk, unsigned int next) blk 1082 sound/pci/ctxfi/cthw20k1.c struct daio_mgr_ctrl_blk *ctl = blk; blk 1089 sound/pci/ctxfi/cthw20k1.c static int daio_mgr_set_imapaddr(void *blk, unsigned int addr) blk 1091 sound/pci/ctxfi/cthw20k1.c struct daio_mgr_ctrl_blk *ctl = blk; blk 1098 sound/pci/ctxfi/cthw20k1.c static int daio_mgr_commit_write(struct hw *hw, void *blk) blk 1100 sound/pci/ctxfi/cthw20k1.c struct daio_mgr_ctrl_blk *ctl = blk; blk 1141 sound/pci/ctxfi/cthw20k1.c struct daio_mgr_ctrl_blk *blk; blk 1144 sound/pci/ctxfi/cthw20k1.c blk = kzalloc(sizeof(*blk), GFP_KERNEL); blk 1145 sound/pci/ctxfi/cthw20k1.c if (!blk) blk 1148 sound/pci/ctxfi/cthw20k1.c blk->i2sctl = hw_read_20kx(hw, I2SCTL); blk 1149 sound/pci/ctxfi/cthw20k1.c blk->spoctl = hw_read_20kx(hw, SPOCTL); blk 1150 sound/pci/ctxfi/cthw20k1.c blk->spictl = hw_read_20kx(hw, SPICTL); blk 1152 sound/pci/ctxfi/cthw20k1.c *rblk = blk; blk 1157 sound/pci/ctxfi/cthw20k1.c static int daio_mgr_put_ctrl_blk(void *blk) blk 1159 sound/pci/ctxfi/cthw20k1.c kfree((struct daio_mgr_ctrl_blk *)blk); blk 157 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *blk; blk 160 sound/pci/ctxfi/cthw20k2.c blk = kzalloc(sizeof(*blk), GFP_KERNEL); blk 161 sound/pci/ctxfi/cthw20k2.c if (!blk) blk 164 sound/pci/ctxfi/cthw20k2.c *rblk = blk; blk 169 sound/pci/ctxfi/cthw20k2.c static int src_put_rsc_ctrl_blk(void *blk) blk 171 sound/pci/ctxfi/cthw20k2.c kfree(blk); blk 176 sound/pci/ctxfi/cthw20k2.c static int src_set_state(void *blk, unsigned int state) blk 178 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 185 sound/pci/ctxfi/cthw20k2.c static int src_set_bm(void *blk, unsigned int bm) blk 187 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 194 sound/pci/ctxfi/cthw20k2.c static int src_set_rsr(void *blk, unsigned int rsr) blk 196 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 203 sound/pci/ctxfi/cthw20k2.c static int src_set_sf(void *blk, unsigned int sf) blk 205 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 212 sound/pci/ctxfi/cthw20k2.c static int src_set_wr(void *blk, unsigned int wr) blk 214 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 221 sound/pci/ctxfi/cthw20k2.c static int src_set_pm(void *blk, unsigned int pm) blk 223 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 230 sound/pci/ctxfi/cthw20k2.c static int src_set_rom(void *blk, unsigned int rom) blk 232 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 239 sound/pci/ctxfi/cthw20k2.c static int src_set_vo(void *blk, unsigned int vo) blk 241 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 248 sound/pci/ctxfi/cthw20k2.c static int src_set_st(void *blk, unsigned int st) blk 250 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 257 sound/pci/ctxfi/cthw20k2.c static int src_set_ie(void *blk, unsigned int ie) blk 259 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 266 sound/pci/ctxfi/cthw20k2.c static int src_set_ilsz(void *blk, unsigned int ilsz) blk 268 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 275 sound/pci/ctxfi/cthw20k2.c static int src_set_bp(void *blk, unsigned int bp) blk 277 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 284 sound/pci/ctxfi/cthw20k2.c static int src_set_cisz(void *blk, unsigned int cisz) blk 286 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 293 sound/pci/ctxfi/cthw20k2.c static int src_set_ca(void *blk, unsigned int ca) blk 295 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 302 sound/pci/ctxfi/cthw20k2.c static int src_set_sa(void *blk, unsigned int sa) blk 304 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 311 sound/pci/ctxfi/cthw20k2.c static int src_set_la(void *blk, unsigned int la) blk 313 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 320 sound/pci/ctxfi/cthw20k2.c static int src_set_pitch(void *blk, unsigned int pitch) blk 322 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 329 sound/pci/ctxfi/cthw20k2.c static int src_set_clear_zbufs(void *blk, unsigned int clear) blk 331 sound/pci/ctxfi/cthw20k2.c ((struct src_rsc_ctrl_blk *)blk)->dirty.bf.czbfs = (clear ? 1 : 0); blk 335 sound/pci/ctxfi/cthw20k2.c static int src_set_dirty(void *blk, unsigned int flags) blk 337 sound/pci/ctxfi/cthw20k2.c ((struct src_rsc_ctrl_blk *)blk)->dirty.data = (flags & 0xffff); blk 341 sound/pci/ctxfi/cthw20k2.c static int src_set_dirty_all(void *blk) blk 343 sound/pci/ctxfi/cthw20k2.c ((struct src_rsc_ctrl_blk *)blk)->dirty.data = ~(0x0); blk 359 sound/pci/ctxfi/cthw20k2.c static int src_commit_write(struct hw *hw, unsigned int idx, void *blk) blk 361 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 416 sound/pci/ctxfi/cthw20k2.c static int src_get_ca(struct hw *hw, unsigned int idx, void *blk) blk 418 sound/pci/ctxfi/cthw20k2.c struct src_rsc_ctrl_blk *ctl = blk; blk 426 sound/pci/ctxfi/cthw20k2.c static unsigned int src_get_dirty(void *blk) blk 428 sound/pci/ctxfi/cthw20k2.c return ((struct src_rsc_ctrl_blk *)blk)->dirty.data; blk 436 sound/pci/ctxfi/cthw20k2.c static int src_mgr_enbs_src(void *blk, unsigned int idx) blk 438 sound/pci/ctxfi/cthw20k2.c ((struct src_mgr_ctrl_blk *)blk)->enbsa |= (0x1 << ((idx%128)/4)); blk 439 sound/pci/ctxfi/cthw20k2.c ((struct src_mgr_ctrl_blk *)blk)->dirty.bf.enbsa = 1; blk 440 sound/pci/ctxfi/cthw20k2.c ((struct src_mgr_ctrl_blk *)blk)->enb[idx/32] |= (0x1 << (idx%32)); blk 444 sound/pci/ctxfi/cthw20k2.c static int src_mgr_enb_src(void *blk, unsigned int idx) blk 446 sound/pci/ctxfi/cthw20k2.c ((struct src_mgr_ctrl_blk *)blk)->enb[idx/32] |= (0x1 << (idx%32)); blk 447 sound/pci/ctxfi/cthw20k2.c ((struct src_mgr_ctrl_blk *)blk)->dirty.data |= (0x1 << (idx/32)); blk 451 sound/pci/ctxfi/cthw20k2.c static int src_mgr_dsb_src(void *blk, unsigned int idx) blk 453 sound/pci/ctxfi/cthw20k2.c ((struct src_mgr_ctrl_blk *)blk)->enb[idx/32] &= ~(0x1 << (idx%32)); blk 454 sound/pci/ctxfi/cthw20k2.c ((struct src_mgr_ctrl_blk *)blk)->dirty.data |= (0x1 << (idx/32)); blk 458 sound/pci/ctxfi/cthw20k2.c static int src_mgr_commit_write(struct hw *hw, void *blk) blk 460 sound/pci/ctxfi/cthw20k2.c struct src_mgr_ctrl_blk *ctl = blk; blk 483 sound/pci/ctxfi/cthw20k2.c struct src_mgr_ctrl_blk *blk; blk 486 sound/pci/ctxfi/cthw20k2.c blk = kzalloc(sizeof(*blk), GFP_KERNEL); blk 487 sound/pci/ctxfi/cthw20k2.c if (!blk) blk 490 sound/pci/ctxfi/cthw20k2.c *rblk = blk; blk 495 sound/pci/ctxfi/cthw20k2.c static int src_mgr_put_ctrl_blk(void *blk) blk 497 sound/pci/ctxfi/cthw20k2.c kfree(blk); blk 504 sound/pci/ctxfi/cthw20k2.c struct srcimp_mgr_ctrl_blk *blk; blk 507 sound/pci/ctxfi/cthw20k2.c blk = kzalloc(sizeof(*blk), GFP_KERNEL); blk 508 sound/pci/ctxfi/cthw20k2.c if (!blk) blk 511 sound/pci/ctxfi/cthw20k2.c *rblk = blk; blk 516 sound/pci/ctxfi/cthw20k2.c static int srcimp_mgr_put_ctrl_blk(void *blk) blk 518 sound/pci/ctxfi/cthw20k2.c kfree(blk); blk 523 sound/pci/ctxfi/cthw20k2.c static int srcimp_mgr_set_imaparc(void *blk, unsigned int slot) blk 525 sound/pci/ctxfi/cthw20k2.c struct srcimp_mgr_ctrl_blk *ctl = blk; blk 532 sound/pci/ctxfi/cthw20k2.c static int srcimp_mgr_set_imapuser(void *blk, unsigned int user) blk 534 sound/pci/ctxfi/cthw20k2.c struct srcimp_mgr_ctrl_blk *ctl = blk; blk 541 sound/pci/ctxfi/cthw20k2.c static int srcimp_mgr_set_imapnxt(void *blk, unsigned int next) blk 543 sound/pci/ctxfi/cthw20k2.c struct srcimp_mgr_ctrl_blk *ctl = blk; blk 550 sound/pci/ctxfi/cthw20k2.c static int srcimp_mgr_set_imapaddr(void *blk, unsigned int addr) blk 552 sound/pci/ctxfi/cthw20k2.c ((struct srcimp_mgr_ctrl_blk *)blk)->srcimap.idx = addr; blk 553 sound/pci/ctxfi/cthw20k2.c ((struct srcimp_mgr_ctrl_blk *)blk)->dirty.bf.srcimap = 1; blk 557 sound/pci/ctxfi/cthw20k2.c static int srcimp_mgr_commit_write(struct hw *hw, void *blk) blk 559 sound/pci/ctxfi/cthw20k2.c struct srcimp_mgr_ctrl_blk *ctl = blk; blk 599 sound/pci/ctxfi/cthw20k2.c static int amixer_set_mode(void *blk, unsigned int mode) blk 601 sound/pci/ctxfi/cthw20k2.c struct amixer_rsc_ctrl_blk *ctl = blk; blk 608 sound/pci/ctxfi/cthw20k2.c static int amixer_set_iv(void *blk, unsigned int iv) blk 610 sound/pci/ctxfi/cthw20k2.c struct amixer_rsc_ctrl_blk *ctl = blk; blk 617 sound/pci/ctxfi/cthw20k2.c static int amixer_set_x(void *blk, unsigned int x) blk 619 sound/pci/ctxfi/cthw20k2.c struct amixer_rsc_ctrl_blk *ctl = blk; blk 626 sound/pci/ctxfi/cthw20k2.c static int amixer_set_y(void *blk, unsigned int y) blk 628 sound/pci/ctxfi/cthw20k2.c struct amixer_rsc_ctrl_blk *ctl = blk; blk 635 sound/pci/ctxfi/cthw20k2.c static int amixer_set_sadr(void *blk, unsigned int sadr) blk 637 sound/pci/ctxfi/cthw20k2.c struct amixer_rsc_ctrl_blk *ctl = blk; blk 644 sound/pci/ctxfi/cthw20k2.c static int amixer_set_se(void *blk, unsigned int se) blk 646 sound/pci/ctxfi/cthw20k2.c struct amixer_rsc_ctrl_blk *ctl = blk; blk 653 sound/pci/ctxfi/cthw20k2.c static int amixer_set_dirty(void *blk, unsigned int flags) blk 655 sound/pci/ctxfi/cthw20k2.c ((struct amixer_rsc_ctrl_blk *)blk)->dirty.data = (flags & 0xffff); blk 659 sound/pci/ctxfi/cthw20k2.c static int amixer_set_dirty_all(void *blk) blk 661 sound/pci/ctxfi/cthw20k2.c ((struct amixer_rsc_ctrl_blk *)blk)->dirty.data = ~(0x0); blk 665 sound/pci/ctxfi/cthw20k2.c static int amixer_commit_write(struct hw *hw, unsigned int idx, void *blk) blk 667 sound/pci/ctxfi/cthw20k2.c struct amixer_rsc_ctrl_blk *ctl = blk; blk 679 sound/pci/ctxfi/cthw20k2.c static int amixer_get_y(void *blk) blk 681 sound/pci/ctxfi/cthw20k2.c struct amixer_rsc_ctrl_blk *ctl = blk; blk 686 sound/pci/ctxfi/cthw20k2.c static unsigned int amixer_get_dirty(void *blk) blk 688 sound/pci/ctxfi/cthw20k2.c return ((struct amixer_rsc_ctrl_blk *)blk)->dirty.data; blk 693 sound/pci/ctxfi/cthw20k2.c struct amixer_rsc_ctrl_blk *blk; blk 696 sound/pci/ctxfi/cthw20k2.c blk = kzalloc(sizeof(*blk), GFP_KERNEL); blk 697 sound/pci/ctxfi/cthw20k2.c if (!blk) blk 700 sound/pci/ctxfi/cthw20k2.c *rblk = blk; blk 705 sound/pci/ctxfi/cthw20k2.c static int amixer_rsc_put_ctrl_blk(void *blk) blk 707 sound/pci/ctxfi/cthw20k2.c kfree(blk); blk 719 sound/pci/ctxfi/cthw20k2.c static int amixer_mgr_put_ctrl_blk(void *blk) blk 814 sound/pci/ctxfi/cthw20k2.c static int dai_srt_set_srco(void *blk, unsigned int src) blk 816 sound/pci/ctxfi/cthw20k2.c struct dai_ctrl_blk *ctl = blk; blk 823 sound/pci/ctxfi/cthw20k2.c static int dai_srt_set_srcm(void *blk, unsigned int src) blk 825 sound/pci/ctxfi/cthw20k2.c struct dai_ctrl_blk *ctl = blk; blk 832 sound/pci/ctxfi/cthw20k2.c static int dai_srt_set_rsr(void *blk, unsigned int rsr) blk 834 sound/pci/ctxfi/cthw20k2.c struct dai_ctrl_blk *ctl = blk; blk 841 sound/pci/ctxfi/cthw20k2.c static int dai_srt_set_drat(void *blk, unsigned int drat) blk 843 sound/pci/ctxfi/cthw20k2.c struct dai_ctrl_blk *ctl = blk; blk 850 sound/pci/ctxfi/cthw20k2.c static int dai_srt_set_ec(void *blk, unsigned int ec) blk 852 sound/pci/ctxfi/cthw20k2.c struct dai_ctrl_blk *ctl = blk; blk 859 sound/pci/ctxfi/cthw20k2.c static int dai_srt_set_et(void *blk, unsigned int et) blk 861 sound/pci/ctxfi/cthw20k2.c struct dai_ctrl_blk *ctl = blk; blk 868 sound/pci/ctxfi/cthw20k2.c static int dai_commit_write(struct hw *hw, unsigned int idx, void *blk) blk 870 sound/pci/ctxfi/cthw20k2.c struct dai_ctrl_blk *ctl = blk; blk 882 sound/pci/ctxfi/cthw20k2.c struct dai_ctrl_blk *blk; blk 885 sound/pci/ctxfi/cthw20k2.c blk = kzalloc(sizeof(*blk), GFP_KERNEL); blk 886 sound/pci/ctxfi/cthw20k2.c if (!blk) blk 889 sound/pci/ctxfi/cthw20k2.c *rblk = blk; blk 894 sound/pci/ctxfi/cthw20k2.c static int dai_put_ctrl_blk(void *blk) blk 896 sound/pci/ctxfi/cthw20k2.c kfree(blk); blk 901 sound/pci/ctxfi/cthw20k2.c static int dao_set_spos(void *blk, unsigned int spos) blk 903 sound/pci/ctxfi/cthw20k2.c ((struct dao_ctrl_blk *)blk)->atxcsl = spos; blk 904 sound/pci/ctxfi/cthw20k2.c ((struct dao_ctrl_blk *)blk)->dirty.bf.atxcsl = 1; blk 908 sound/pci/ctxfi/cthw20k2.c static int dao_commit_write(struct hw *hw, unsigned int idx, void *blk) blk 910 sound/pci/ctxfi/cthw20k2.c struct dao_ctrl_blk *ctl = blk; blk 924 sound/pci/ctxfi/cthw20k2.c static int dao_get_spos(void *blk, unsigned int *spos) blk 926 sound/pci/ctxfi/cthw20k2.c *spos = ((struct dao_ctrl_blk *)blk)->atxcsl; blk 932 sound/pci/ctxfi/cthw20k2.c struct dao_ctrl_blk *blk; blk 935 sound/pci/ctxfi/cthw20k2.c blk = kzalloc(sizeof(*blk), GFP_KERNEL); blk 936 sound/pci/ctxfi/cthw20k2.c if (!blk) blk 939 sound/pci/ctxfi/cthw20k2.c *rblk = blk; blk 944 sound/pci/ctxfi/cthw20k2.c static int dao_put_ctrl_blk(void *blk) blk 946 sound/pci/ctxfi/cthw20k2.c kfree(blk); blk 951 sound/pci/ctxfi/cthw20k2.c static int daio_mgr_enb_dai(void *blk, unsigned int idx) blk 953 sound/pci/ctxfi/cthw20k2.c struct daio_mgr_ctrl_blk *ctl = blk; blk 960 sound/pci/ctxfi/cthw20k2.c static int daio_mgr_dsb_dai(void *blk, unsigned int idx) blk 962 sound/pci/ctxfi/cthw20k2.c struct daio_mgr_ctrl_blk *ctl = blk; blk 970 sound/pci/ctxfi/cthw20k2.c static int daio_mgr_enb_dao(void *blk, unsigned int idx) blk 972 sound/pci/ctxfi/cthw20k2.c struct daio_mgr_ctrl_blk *ctl = blk; blk 979 sound/pci/ctxfi/cthw20k2.c static int daio_mgr_dsb_dao(void *blk, unsigned int idx) blk 981 sound/pci/ctxfi/cthw20k2.c struct daio_mgr_ctrl_blk *ctl = blk; blk 988 sound/pci/ctxfi/cthw20k2.c static int daio_mgr_dao_init(void *blk, unsigned int idx, unsigned int conf) blk 990 sound/pci/ctxfi/cthw20k2.c struct daio_mgr_ctrl_blk *ctl = blk; blk 1026 sound/pci/ctxfi/cthw20k2.c static int daio_mgr_set_imaparc(void *blk, unsigned int slot) blk 1028 sound/pci/ctxfi/cthw20k2.c struct daio_mgr_ctrl_blk *ctl = blk; blk 1035 sound/pci/ctxfi/cthw20k2.c static int daio_mgr_set_imapnxt(void *blk, unsigned int next) blk 1037 sound/pci/ctxfi/cthw20k2.c struct daio_mgr_ctrl_blk *ctl = blk; blk 1044 sound/pci/ctxfi/cthw20k2.c static int daio_mgr_set_imapaddr(void *blk, unsigned int addr) blk 1046 sound/pci/ctxfi/cthw20k2.c ((struct daio_mgr_ctrl_blk *)blk)->daoimap.idx = addr; blk 1047 sound/pci/ctxfi/cthw20k2.c ((struct daio_mgr_ctrl_blk *)blk)->dirty.bf.daoimap = 1; blk 1051 sound/pci/ctxfi/cthw20k2.c static int daio_mgr_commit_write(struct hw *hw, void *blk) blk 1053 sound/pci/ctxfi/cthw20k2.c struct daio_mgr_ctrl_blk *ctl = blk; blk 1082 sound/pci/ctxfi/cthw20k2.c struct daio_mgr_ctrl_blk *blk; blk 1086 sound/pci/ctxfi/cthw20k2.c blk = kzalloc(sizeof(*blk), GFP_KERNEL); blk 1087 sound/pci/ctxfi/cthw20k2.c if (!blk) blk 1091 sound/pci/ctxfi/cthw20k2.c blk->txctl[i] = hw_read_20kx(hw, AUDIO_IO_TX_CTL+(0x40*i)); blk 1092 sound/pci/ctxfi/cthw20k2.c blk->rxctl[i] = hw_read_20kx(hw, AUDIO_IO_RX_CTL+(0x40*i)); blk 1095 sound/pci/ctxfi/cthw20k2.c *rblk = blk; blk 1100 sound/pci/ctxfi/cthw20k2.c static int daio_mgr_put_ctrl_blk(void *blk) blk 1102 sound/pci/ctxfi/cthw20k2.c kfree(blk); blk 69 sound/pci/emu10k1/memory.c static int synth_alloc_pages(struct snd_emu10k1 *hw, struct snd_emu10k1_memblk *blk); blk 70 sound/pci/emu10k1/memory.c static int synth_free_pages(struct snd_emu10k1 *hw, struct snd_emu10k1_memblk *blk); blk 76 sound/pci/emu10k1/memory.c static void emu10k1_memblk_init(struct snd_emu10k1_memblk *blk) blk 78 sound/pci/emu10k1/memory.c blk->mapped_page = -1; blk 79 sound/pci/emu10k1/memory.c INIT_LIST_HEAD(&blk->mapped_link); blk 80 sound/pci/emu10k1/memory.c INIT_LIST_HEAD(&blk->mapped_order_link); blk 81 sound/pci/emu10k1/memory.c blk->map_locked = 0; blk 83 sound/pci/emu10k1/memory.c blk->first_page = get_aligned_page(blk->mem.offset); blk 84 sound/pci/emu10k1/memory.c blk->last_page = get_aligned_page(blk->mem.offset + blk->mem.size - 1); blk 85 sound/pci/emu10k1/memory.c blk->pages = blk->last_page - blk->first_page + 1; blk 104 sound/pci/emu10k1/memory.c struct snd_emu10k1_memblk *blk = get_emu10k1_memblk(pos, mapped_link); blk 105 sound/pci/emu10k1/memory.c if (blk->mapped_page < 0) blk 107 sound/pci/emu10k1/memory.c size = blk->mapped_page - page; blk 118 sound/pci/emu10k1/memory.c page = blk->mapped_page + blk->pages; blk 134 sound/pci/emu10k1/memory.c static int map_memblk(struct snd_emu10k1 *emu, struct snd_emu10k1_memblk *blk) blk 139 sound/pci/emu10k1/memory.c page = search_empty_map_area(emu, blk->pages, &next); blk 147 sound/pci/emu10k1/memory.c list_add_tail(&blk->mapped_link, next); blk 149 sound/pci/emu10k1/memory.c list_add_tail(&blk->mapped_order_link, &emu->mapped_order_link_head); blk 150 sound/pci/emu10k1/memory.c blk->mapped_page = page; blk 152 sound/pci/emu10k1/memory.c for (pg = blk->first_page; pg <= blk->last_page; pg++) { blk 165 sound/pci/emu10k1/memory.c static int unmap_memblk(struct snd_emu10k1 *emu, struct snd_emu10k1_memblk *blk) blk 172 sound/pci/emu10k1/memory.c if ((p = blk->mapped_link.prev) != &emu->mapped_link_head) { blk 177 sound/pci/emu10k1/memory.c if ((p = blk->mapped_link.next) != &emu->mapped_link_head) { blk 184 sound/pci/emu10k1/memory.c list_del(&blk->mapped_link); blk 185 sound/pci/emu10k1/memory.c list_del(&blk->mapped_order_link); blk 187 sound/pci/emu10k1/memory.c mpage = blk->mapped_page; blk 188 sound/pci/emu10k1/memory.c for (pg = blk->first_page; pg <= blk->last_page; pg++) { blk 192 sound/pci/emu10k1/memory.c blk->mapped_page = -1; blk 205 sound/pci/emu10k1/memory.c struct snd_emu10k1_memblk *blk; blk 211 sound/pci/emu10k1/memory.c blk = get_emu10k1_memblk(p, mem.list); blk 212 sound/pci/emu10k1/memory.c if (page + psize <= blk->first_page) blk 214 sound/pci/emu10k1/memory.c page = blk->last_page + 1; blk 221 sound/pci/emu10k1/memory.c blk = (struct snd_emu10k1_memblk *)__snd_util_memblk_new(emu->memhdr, psize << PAGE_SHIFT, p->prev); blk 222 sound/pci/emu10k1/memory.c if (blk == NULL) blk 224 sound/pci/emu10k1/memory.c blk->mem.offset = aligned_page_offset(page); /* set aligned offset */ blk 225 sound/pci/emu10k1/memory.c emu10k1_memblk_init(blk); blk 226 sound/pci/emu10k1/memory.c return blk; blk 254 sound/pci/emu10k1/memory.c int snd_emu10k1_memblk_map(struct snd_emu10k1 *emu, struct snd_emu10k1_memblk *blk) blk 263 sound/pci/emu10k1/memory.c if (blk->mapped_page >= 0) { blk 265 sound/pci/emu10k1/memory.c list_move_tail(&blk->mapped_order_link, blk 270 sound/pci/emu10k1/memory.c if ((err = map_memblk(emu, blk)) < 0) { blk 280 sound/pci/emu10k1/memory.c if (size >= blk->pages) { blk 282 sound/pci/emu10k1/memory.c err = map_memblk(emu, blk); blk 301 sound/pci/emu10k1/memory.c struct snd_emu10k1_memblk *blk; blk 316 sound/pci/emu10k1/memory.c blk = search_empty(emu, runtime->dma_bytes + idx); blk 317 sound/pci/emu10k1/memory.c if (blk == NULL) { blk 325 sound/pci/emu10k1/memory.c for (page = blk->first_page; page <= blk->last_page; page++, idx++) { blk 343 sound/pci/emu10k1/memory.c blk->map_locked = 1; /* do not unmap this block! */ blk 344 sound/pci/emu10k1/memory.c err = snd_emu10k1_memblk_map(emu, blk); blk 346 sound/pci/emu10k1/memory.c __snd_util_mem_free(hdr, (struct snd_util_memblk *)blk); blk 351 sound/pci/emu10k1/memory.c return (struct snd_util_memblk *)blk; blk 358 sound/pci/emu10k1/memory.c int snd_emu10k1_free_pages(struct snd_emu10k1 *emu, struct snd_util_memblk *blk) blk 360 sound/pci/emu10k1/memory.c if (snd_BUG_ON(!emu || !blk)) blk 362 sound/pci/emu10k1/memory.c return snd_emu10k1_synth_free(emu, blk); blk 404 sound/pci/emu10k1/memory.c struct snd_emu10k1_memblk *blk; blk 408 sound/pci/emu10k1/memory.c blk = (struct snd_emu10k1_memblk *)__snd_util_mem_alloc(hdr, size); blk 409 sound/pci/emu10k1/memory.c if (blk == NULL) { blk 413 sound/pci/emu10k1/memory.c if (synth_alloc_pages(hw, blk)) { blk 414 sound/pci/emu10k1/memory.c __snd_util_mem_free(hdr, (struct snd_util_memblk *)blk); blk 418 sound/pci/emu10k1/memory.c snd_emu10k1_memblk_map(hw, blk); blk 420 sound/pci/emu10k1/memory.c return (struct snd_util_memblk *)blk; blk 432 sound/pci/emu10k1/memory.c struct snd_emu10k1_memblk *blk = (struct snd_emu10k1_memblk *)memblk; blk 437 sound/pci/emu10k1/memory.c if (blk->mapped_page >= 0) blk 438 sound/pci/emu10k1/memory.c unmap_memblk(emu, blk); blk 440 sound/pci/emu10k1/memory.c synth_free_pages(emu, blk); blk 450 sound/pci/emu10k1/memory.c struct snd_emu10k1_memblk *blk, blk 456 sound/pci/emu10k1/memory.c first_page = blk->first_page; blk 457 sound/pci/emu10k1/memory.c if ((p = blk->mem.list.prev) != &hdr->block) { blk 462 sound/pci/emu10k1/memory.c last_page = blk->last_page; blk 463 sound/pci/emu10k1/memory.c if ((p = blk->mem.list.next) != &hdr->block) { blk 505 sound/pci/emu10k1/memory.c static int synth_alloc_pages(struct snd_emu10k1 *emu, struct snd_emu10k1_memblk *blk) blk 510 sound/pci/emu10k1/memory.c emu10k1_memblk_init(blk); blk 511 sound/pci/emu10k1/memory.c get_single_page_range(emu->memhdr, blk, &first_page, &last_page); blk 537 sound/pci/emu10k1/memory.c static int synth_free_pages(struct snd_emu10k1 *emu, struct snd_emu10k1_memblk *blk) blk 541 sound/pci/emu10k1/memory.c get_single_page_range(emu->memhdr, blk, &first_page, &last_page); blk 565 sound/pci/emu10k1/memory.c int snd_emu10k1_synth_bzero(struct snd_emu10k1 *emu, struct snd_util_memblk *blk, blk 570 sound/pci/emu10k1/memory.c struct snd_emu10k1_memblk *p = (struct snd_emu10k1_memblk *)blk; blk 572 sound/pci/emu10k1/memory.c offset += blk->offset & (PAGE_SIZE - 1); blk 595 sound/pci/emu10k1/memory.c int snd_emu10k1_synth_copy_from_user(struct snd_emu10k1 *emu, struct snd_util_memblk *blk, blk 600 sound/pci/emu10k1/memory.c struct snd_emu10k1_memblk *p = (struct snd_emu10k1_memblk *)blk; blk 602 sound/pci/emu10k1/memory.c offset += blk->offset & (PAGE_SIZE - 1); blk 423 sound/pci/trident/trident.h int snd_trident_free_pages(struct snd_trident *trident, struct snd_util_memblk *blk); blk 425 sound/pci/trident/trident.h int snd_trident_synth_free(struct snd_trident *trident, struct snd_util_memblk *blk); blk 426 sound/pci/trident/trident.h int snd_trident_synth_copy_from_user(struct snd_trident *trident, struct snd_util_memblk *blk, blk 118 sound/pci/trident/trident_memory.c #define firstpg(blk) (((struct snd_trident_memblk_arg *)snd_util_memblk_argptr(blk))->first_page) blk 119 sound/pci/trident/trident_memory.c #define lastpg(blk) (((struct snd_trident_memblk_arg *)snd_util_memblk_argptr(blk))->last_page) blk 127 sound/pci/trident/trident_memory.c struct snd_util_memblk *blk; blk 134 sound/pci/trident/trident_memory.c blk = list_entry(p, struct snd_util_memblk, list); blk 135 sound/pci/trident/trident_memory.c if (page + psize <= firstpg(blk)) blk 137 sound/pci/trident/trident_memory.c page = lastpg(blk) + 1; blk 144 sound/pci/trident/trident_memory.c blk = __snd_util_memblk_new(hdr, psize * ALIGN_PAGE_SIZE, p->prev); blk 145 sound/pci/trident/trident_memory.c if (blk == NULL) blk 147 sound/pci/trident/trident_memory.c blk->offset = aligned_page_offset(page); /* set aligned offset */ blk 148 sound/pci/trident/trident_memory.c firstpg(blk) = page; blk 149 sound/pci/trident/trident_memory.c lastpg(blk) = page + psize - 1; blk 150 sound/pci/trident/trident_memory.c return blk; blk 178 sound/pci/trident/trident_memory.c struct snd_util_memblk *blk; blk 193 sound/pci/trident/trident_memory.c blk = search_empty(hdr, runtime->dma_bytes); blk 194 sound/pci/trident/trident_memory.c if (blk == NULL) { blk 201 sound/pci/trident/trident_memory.c for (page = firstpg(blk); page <= lastpg(blk); page++, idx++) { blk 207 sound/pci/trident/trident_memory.c __snd_util_mem_free(hdr, blk); blk 214 sound/pci/trident/trident_memory.c return blk; blk 225 sound/pci/trident/trident_memory.c struct snd_util_memblk *blk; blk 240 sound/pci/trident/trident_memory.c blk = search_empty(hdr, runtime->dma_bytes); blk 241 sound/pci/trident/trident_memory.c if (blk == NULL) { blk 249 sound/pci/trident/trident_memory.c for (page = firstpg(blk); page <= lastpg(blk); page++, blk 252 sound/pci/trident/trident_memory.c __snd_util_mem_free(hdr, blk); blk 259 sound/pci/trident/trident_memory.c return blk; blk 282 sound/pci/trident/trident_memory.c struct snd_util_memblk *blk) blk 287 sound/pci/trident/trident_memory.c if (snd_BUG_ON(!trident || !blk)) blk 293 sound/pci/trident/trident_memory.c for (page = firstpg(blk); page <= lastpg(blk); page++) blk 296 sound/pci/trident/trident_memory.c __snd_util_mem_free(hdr, blk); blk 1580 sound/soc/codecs/wm_adsp.c struct wm_coeff_parsed_alg *blk) blk 1590 sound/soc/codecs/wm_adsp.c blk->id = le32_to_cpu(raw->id); blk 1591 sound/soc/codecs/wm_adsp.c blk->name = raw->name; blk 1592 sound/soc/codecs/wm_adsp.c blk->name_len = strlen(raw->name); blk 1593 sound/soc/codecs/wm_adsp.c blk->ncoeff = le32_to_cpu(raw->ncoeff); blk 1596 sound/soc/codecs/wm_adsp.c blk->id = wm_coeff_parse_int(sizeof(raw->id), data); blk 1597 sound/soc/codecs/wm_adsp.c blk->name_len = wm_coeff_parse_string(sizeof(u8), data, blk 1598 sound/soc/codecs/wm_adsp.c &blk->name); blk 1600 sound/soc/codecs/wm_adsp.c blk->ncoeff = wm_coeff_parse_int(sizeof(raw->ncoeff), data); blk 1604 sound/soc/codecs/wm_adsp.c adsp_dbg(dsp, "Algorithm ID: %#x\n", blk->id); blk 1605 sound/soc/codecs/wm_adsp.c adsp_dbg(dsp, "Algorithm name: %.*s\n", blk->name_len, blk->name); blk 1606 sound/soc/codecs/wm_adsp.c adsp_dbg(dsp, "# of coefficient descriptors: %#x\n", blk->ncoeff); blk 1610 sound/soc/codecs/wm_adsp.c struct wm_coeff_parsed_coeff *blk) blk 1622 sound/soc/codecs/wm_adsp.c blk->offset = le16_to_cpu(raw->hdr.offset); blk 1623 sound/soc/codecs/wm_adsp.c blk->mem_type = le16_to_cpu(raw->hdr.type); blk 1624 sound/soc/codecs/wm_adsp.c blk->name = raw->name; blk 1625 sound/soc/codecs/wm_adsp.c blk->name_len = strlen(raw->name); blk 1626 sound/soc/codecs/wm_adsp.c blk->ctl_type = le16_to_cpu(raw->ctl_type); blk 1627 sound/soc/codecs/wm_adsp.c blk->flags = le16_to_cpu(raw->flags); blk 1628 sound/soc/codecs/wm_adsp.c blk->len = le32_to_cpu(raw->len); blk 1632 sound/soc/codecs/wm_adsp.c blk->offset = wm_coeff_parse_int(sizeof(raw->hdr.offset), &tmp); blk 1633 sound/soc/codecs/wm_adsp.c blk->mem_type = wm_coeff_parse_int(sizeof(raw->hdr.type), &tmp); blk 1635 sound/soc/codecs/wm_adsp.c blk->name_len = wm_coeff_parse_string(sizeof(u8), &tmp, blk 1636 sound/soc/codecs/wm_adsp.c &blk->name); blk 1639 sound/soc/codecs/wm_adsp.c blk->ctl_type = wm_coeff_parse_int(sizeof(raw->ctl_type), &tmp); blk 1640 sound/soc/codecs/wm_adsp.c blk->flags = wm_coeff_parse_int(sizeof(raw->flags), &tmp); blk 1641 sound/soc/codecs/wm_adsp.c blk->len = wm_coeff_parse_int(sizeof(raw->len), &tmp); blk 1647 sound/soc/codecs/wm_adsp.c adsp_dbg(dsp, "\tCoefficient type: %#x\n", blk->mem_type); blk 1648 sound/soc/codecs/wm_adsp.c adsp_dbg(dsp, "\tCoefficient offset: %#x\n", blk->offset); blk 1649 sound/soc/codecs/wm_adsp.c adsp_dbg(dsp, "\tCoefficient name: %.*s\n", blk->name_len, blk->name); blk 1650 sound/soc/codecs/wm_adsp.c adsp_dbg(dsp, "\tCoefficient flags: %#x\n", blk->flags); blk 1651 sound/soc/codecs/wm_adsp.c adsp_dbg(dsp, "\tALSA control type: %#x\n", blk->ctl_type); blk 1652 sound/soc/codecs/wm_adsp.c adsp_dbg(dsp, "\tALSA control len: %#x\n", blk->len); blk 2460 sound/soc/codecs/wm_adsp.c struct wmfw_coeff_item *blk; blk 2516 sound/soc/codecs/wm_adsp.c sizeof(*blk) < firmware->size - pos) { blk 2517 sound/soc/codecs/wm_adsp.c blk = (void *)(&firmware->data[pos]); blk 2519 sound/soc/codecs/wm_adsp.c type = le16_to_cpu(blk->type); blk 2520 sound/soc/codecs/wm_adsp.c offset = le16_to_cpu(blk->offset); blk 2523 sound/soc/codecs/wm_adsp.c file, blocks, le32_to_cpu(blk->id), blk 2524 sound/soc/codecs/wm_adsp.c (le32_to_cpu(blk->ver) >> 16) & 0xff, blk 2525 sound/soc/codecs/wm_adsp.c (le32_to_cpu(blk->ver) >> 8) & 0xff, blk 2526 sound/soc/codecs/wm_adsp.c le32_to_cpu(blk->ver) & 0xff); blk 2528 sound/soc/codecs/wm_adsp.c file, blocks, le32_to_cpu(blk->len), offset, type); blk 2541 sound/soc/codecs/wm_adsp.c if (le32_to_cpu(blk->id) == dsp->fw_id && blk 2565 sound/soc/codecs/wm_adsp.c file, blocks, le32_to_cpu(blk->len), blk 2566 sound/soc/codecs/wm_adsp.c type, le32_to_cpu(blk->id)); blk 2575 sound/soc/codecs/wm_adsp.c le32_to_cpu(blk->id)); blk 2582 sound/soc/codecs/wm_adsp.c type, le32_to_cpu(blk->id)); blk 2593 sound/soc/codecs/wm_adsp.c if (le32_to_cpu(blk->len) > blk 2594 sound/soc/codecs/wm_adsp.c firmware->size - pos - sizeof(*blk)) { blk 2598 sound/soc/codecs/wm_adsp.c le32_to_cpu(blk->len), blk 2604 sound/soc/codecs/wm_adsp.c buf = wm_adsp_buf_alloc(blk->data, blk 2605 sound/soc/codecs/wm_adsp.c le32_to_cpu(blk->len), blk 2614 sound/soc/codecs/wm_adsp.c file, blocks, le32_to_cpu(blk->len), blk 2617 sound/soc/codecs/wm_adsp.c le32_to_cpu(blk->len)); blk 2625 sound/soc/codecs/wm_adsp.c pos += (le32_to_cpu(blk->len) + sizeof(*blk) + 3) & ~0x03; blk 62 sound/synth/util_mem.c struct snd_util_memblk *blk; blk 79 sound/synth/util_mem.c blk = get_memblk(p); blk 80 sound/synth/util_mem.c if (blk->offset - prev_offset >= units) blk 82 sound/synth/util_mem.c prev_offset = blk->offset + blk->size; blk 100 sound/synth/util_mem.c struct snd_util_memblk *blk; blk 102 sound/synth/util_mem.c blk = kmalloc(sizeof(struct snd_util_memblk) + hdr->block_extra_size, blk 104 sound/synth/util_mem.c if (blk == NULL) blk 108 sound/synth/util_mem.c blk->offset = 0; blk 111 sound/synth/util_mem.c blk->offset = p->offset + p->size; blk 113 sound/synth/util_mem.c blk->size = units; blk 114 sound/synth/util_mem.c list_add(&blk->list, prev); blk 117 sound/synth/util_mem.c return blk; blk 127 sound/synth/util_mem.c struct snd_util_memblk *blk; blk 129 sound/synth/util_mem.c blk = __snd_util_mem_alloc(hdr, size); blk 131 sound/synth/util_mem.c return blk; blk 140 sound/synth/util_mem.c __snd_util_mem_free(struct snd_util_memhdr *hdr, struct snd_util_memblk *blk) blk 142 sound/synth/util_mem.c list_del(&blk->list); blk 144 sound/synth/util_mem.c hdr->used -= blk->size; blk 145 sound/synth/util_mem.c kfree(blk); blk 151 sound/synth/util_mem.c int snd_util_mem_free(struct snd_util_memhdr *hdr, struct snd_util_memblk *blk) blk 153 sound/synth/util_mem.c if (snd_BUG_ON(!hdr || !blk)) blk 157 sound/synth/util_mem.c __snd_util_mem_free(hdr, blk); blk 42 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c struct intel_pt_blk *blk; blk 283 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c struct intel_pt_blk *blk = stack->blk; blk 285 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c stack->blk = blk->prev; blk 287 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c stack->spare = blk; blk 289 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c free(blk); blk 295 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c if (!stack->blk) blk 298 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c if (!stack->blk) blk 302 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c return stack->blk->ip[--stack->pos]; blk 307 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c struct intel_pt_blk *blk; blk 310 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c blk = stack->spare; blk 313 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c blk = malloc(sizeof(struct intel_pt_blk)); blk 314 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c if (!blk) blk 318 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c blk->prev = stack->blk; blk 319 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c stack->blk = blk; blk 328 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c if (!stack->blk || stack->pos == INTEL_PT_BLK_SIZE) { blk 334 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c stack->blk->ip[stack->pos++] = ip; blk 340 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c while (stack->blk) blk 348 tools/perf/util/intel-pt-decoder/intel-pt-decoder.c zfree(&stack->blk);