btp 571 fs/xfs/libxfs/xfs_da_format.h xfs_dir2_block_leaf_p(struct xfs_dir2_block_tail *btp) btp 573 fs/xfs/libxfs/xfs_da_format.h return ((struct xfs_dir2_leaf_entry *)btp) - be32_to_cpu(btp->count); btp 161 fs/xfs/libxfs/xfs_dir2_block.c struct xfs_dir2_block_tail *btp, btp 180 fs/xfs/libxfs/xfs_dir2_block.c if (btp->stale) { btp 205 fs/xfs/libxfs/xfs_dir2_block.c if (be16_to_cpu(dup->length) + (be32_to_cpu(btp->stale) - 1) * btp 208 fs/xfs/libxfs/xfs_dir2_block.c } else if ((be32_to_cpu(btp->stale) - 1) * (uint)sizeof(*blp) < len) btp 274 fs/xfs/libxfs/xfs_dir2_block.c struct xfs_dir2_block_tail *btp, btp 285 fs/xfs/libxfs/xfs_dir2_block.c fromidx = toidx = be32_to_cpu(btp->count) - 1; btp 301 fs/xfs/libxfs/xfs_dir2_block.c *lfloglow = toidx + 1 - (be32_to_cpu(btp->stale) - 1); btp 302 fs/xfs/libxfs/xfs_dir2_block.c *lfloghigh -= be32_to_cpu(btp->stale) - 1; btp 303 fs/xfs/libxfs/xfs_dir2_block.c be32_add_cpu(&btp->count, -(be32_to_cpu(btp->stale) - 1)); btp 306 fs/xfs/libxfs/xfs_dir2_block.c (xfs_dir2_data_aoff_t)((be32_to_cpu(btp->stale) - 1) * sizeof(*blp)), btp 308 fs/xfs/libxfs/xfs_dir2_block.c btp->stale = cpu_to_be32(1); btp 327 fs/xfs/libxfs/xfs_dir2_block.c xfs_dir2_block_tail_t *btp; /* block tail */ btp 364 fs/xfs/libxfs/xfs_dir2_block.c btp = xfs_dir2_block_tail_p(args->geo, hdr); btp 365 fs/xfs/libxfs/xfs_dir2_block.c blp = xfs_dir2_block_leaf_p(btp); btp 371 fs/xfs/libxfs/xfs_dir2_block.c xfs_dir2_block_need_space(dp, hdr, btp, blp, &tagp, &dup, btp 407 fs/xfs/libxfs/xfs_dir2_block.c xfs_dir2_block_compact(args, bp, hdr, btp, blp, &needlog, btp 410 fs/xfs/libxfs/xfs_dir2_block.c blp = xfs_dir2_block_leaf_p(btp); btp 411 fs/xfs/libxfs/xfs_dir2_block.c } else if (btp->stale) { btp 416 fs/xfs/libxfs/xfs_dir2_block.c lfloglow = be32_to_cpu(btp->count); btp 423 fs/xfs/libxfs/xfs_dir2_block.c for (low = 0, high = be32_to_cpu(btp->count) - 1; low <= high; ) { btp 438 fs/xfs/libxfs/xfs_dir2_block.c if (!btp->stale) { btp 455 fs/xfs/libxfs/xfs_dir2_block.c be32_add_cpu(&btp->count, 1); btp 487 fs/xfs/libxfs/xfs_dir2_block.c highstale < be32_to_cpu(btp->count) && btp 497 fs/xfs/libxfs/xfs_dir2_block.c (highstale == be32_to_cpu(btp->count) || btp 509 fs/xfs/libxfs/xfs_dir2_block.c ASSERT(highstale < be32_to_cpu(btp->count)); btp 517 fs/xfs/libxfs/xfs_dir2_block.c be32_add_cpu(&btp->stale, -1); btp 572 fs/xfs/libxfs/xfs_dir2_block.c xfs_dir2_block_tail_t *btp; btp 574 fs/xfs/libxfs/xfs_dir2_block.c btp = xfs_dir2_block_tail_p(tp->t_mountp->m_dir_geo, hdr); btp 575 fs/xfs/libxfs/xfs_dir2_block.c blp = xfs_dir2_block_leaf_p(btp); btp 589 fs/xfs/libxfs/xfs_dir2_block.c xfs_dir2_block_tail_t *btp; btp 591 fs/xfs/libxfs/xfs_dir2_block.c btp = xfs_dir2_block_tail_p(tp->t_mountp->m_dir_geo, hdr); btp 592 fs/xfs/libxfs/xfs_dir2_block.c xfs_trans_log_buf(tp, bp, (uint)((char *)btp - (char *)hdr), btp 593 fs/xfs/libxfs/xfs_dir2_block.c (uint)((char *)(btp + 1) - (char *)hdr - 1)); btp 607 fs/xfs/libxfs/xfs_dir2_block.c xfs_dir2_block_tail_t *btp; /* block tail */ btp 624 fs/xfs/libxfs/xfs_dir2_block.c btp = xfs_dir2_block_tail_p(args->geo, hdr); btp 625 fs/xfs/libxfs/xfs_dir2_block.c blp = xfs_dir2_block_leaf_p(btp); btp 655 fs/xfs/libxfs/xfs_dir2_block.c xfs_dir2_block_tail_t *btp; /* block tail */ btp 677 fs/xfs/libxfs/xfs_dir2_block.c btp = xfs_dir2_block_tail_p(args->geo, hdr); btp 678 fs/xfs/libxfs/xfs_dir2_block.c blp = xfs_dir2_block_leaf_p(btp); btp 683 fs/xfs/libxfs/xfs_dir2_block.c for (low = 0, high = be32_to_cpu(btp->count) - 1; ; ) { btp 729 fs/xfs/libxfs/xfs_dir2_block.c } while (++mid < be32_to_cpu(btp->count) && btp 757 fs/xfs/libxfs/xfs_dir2_block.c xfs_dir2_block_tail_t *btp; /* block tail */ btp 780 fs/xfs/libxfs/xfs_dir2_block.c btp = xfs_dir2_block_tail_p(args->geo, hdr); btp 781 fs/xfs/libxfs/xfs_dir2_block.c blp = xfs_dir2_block_leaf_p(btp); btp 798 fs/xfs/libxfs/xfs_dir2_block.c be32_add_cpu(&btp->stale, 1); btp 837 fs/xfs/libxfs/xfs_dir2_block.c xfs_dir2_block_tail_t *btp; /* block tail */ btp 854 fs/xfs/libxfs/xfs_dir2_block.c btp = xfs_dir2_block_tail_p(args->geo, hdr); btp 855 fs/xfs/libxfs/xfs_dir2_block.c blp = xfs_dir2_block_leaf_p(btp); btp 901 fs/xfs/libxfs/xfs_dir2_block.c xfs_dir2_block_tail_t *btp; /* block tail */ btp 998 fs/xfs/libxfs/xfs_dir2_block.c btp = xfs_dir2_block_tail_p(args->geo, hdr); btp 999 fs/xfs/libxfs/xfs_dir2_block.c btp->count = cpu_to_be32(leafhdr.count - leafhdr.stale); btp 1000 fs/xfs/libxfs/xfs_dir2_block.c btp->stale = 0; btp 1005 fs/xfs/libxfs/xfs_dir2_block.c lep = xfs_dir2_block_leaf_p(btp); btp 1011 fs/xfs/libxfs/xfs_dir2_block.c ASSERT(to == be32_to_cpu(btp->count)); btp 1012 fs/xfs/libxfs/xfs_dir2_block.c xfs_dir2_block_log_leaf(tp, dbp, 0, be32_to_cpu(btp->count) - 1); btp 1048 fs/xfs/libxfs/xfs_dir2_block.c xfs_dir2_block_tail_t *btp; /* block tail pointer */ btp 1120 fs/xfs/libxfs/xfs_dir2_block.c i = (uint)sizeof(*btp) + btp 1136 fs/xfs/libxfs/xfs_dir2_block.c btp = xfs_dir2_block_tail_p(args->geo, hdr); btp 1137 fs/xfs/libxfs/xfs_dir2_block.c btp->count = cpu_to_be32(sfp->count + 2); /* ., .. */ btp 1138 fs/xfs/libxfs/xfs_dir2_block.c btp->stale = 0; btp 1139 fs/xfs/libxfs/xfs_dir2_block.c blp = xfs_dir2_block_leaf_p(btp); btp 1242 fs/xfs/libxfs/xfs_dir2_block.c xfs_sort(blp, be32_to_cpu(btp->count), sizeof(*blp), xfs_dir2_block_sort); btp 1248 fs/xfs/libxfs/xfs_dir2_block.c xfs_dir2_block_log_leaf(tp, bp, 0, be32_to_cpu(btp->count) - 1); btp 38 fs/xfs/libxfs/xfs_dir2_data.c xfs_dir2_block_tail_t *btp=NULL; /* block tail */ btp 79 fs/xfs/libxfs/xfs_dir2_data.c btp = xfs_dir2_block_tail_p(geo, hdr); btp 80 fs/xfs/libxfs/xfs_dir2_data.c lep = xfs_dir2_block_leaf_p(btp); btp 89 fs/xfs/libxfs/xfs_dir2_data.c if (be32_to_cpu(btp->count) >= btp 90 fs/xfs/libxfs/xfs_dir2_data.c ((char *)btp - p) / sizeof(struct xfs_dir2_leaf_entry)) btp 193 fs/xfs/libxfs/xfs_dir2_data.c for (i = 0; i < be32_to_cpu(btp->count); i++) { btp 198 fs/xfs/libxfs/xfs_dir2_data.c if (i >= be32_to_cpu(btp->count)) btp 210 fs/xfs/libxfs/xfs_dir2_data.c for (i = stale = 0; i < be32_to_cpu(btp->count); i++) { btp 218 fs/xfs/libxfs/xfs_dir2_data.c if (count != be32_to_cpu(btp->count) - be32_to_cpu(btp->stale)) btp 220 fs/xfs/libxfs/xfs_dir2_data.c if (stale != be32_to_cpu(btp->stale)) btp 338 fs/xfs/libxfs/xfs_dir2_leaf.c xfs_dir2_block_tail_t *btp; /* block's tail */ btp 376 fs/xfs/libxfs/xfs_dir2_leaf.c btp = xfs_dir2_block_tail_p(args->geo, hdr); btp 377 fs/xfs/libxfs/xfs_dir2_leaf.c blp = xfs_dir2_block_leaf_p(btp); btp 385 fs/xfs/libxfs/xfs_dir2_leaf.c leafhdr.count = be32_to_cpu(btp->count); btp 386 fs/xfs/libxfs/xfs_dir2_leaf.c leafhdr.stale = be32_to_cpu(btp->stale); btp 394 fs/xfs/libxfs/xfs_dir2_leaf.c memcpy(ents, blp, be32_to_cpu(btp->count) * sizeof(xfs_dir2_leaf_entry_t)); btp 54 fs/xfs/libxfs/xfs_dir2_sf.c xfs_dir2_block_tail_t *btp; /* tail area of the block */ btp 78 fs/xfs/libxfs/xfs_dir2_sf.c btp = xfs_dir2_block_tail_p(geo, hdr); btp 79 fs/xfs/libxfs/xfs_dir2_sf.c blp = xfs_dir2_block_leaf_p(btp); btp 84 fs/xfs/libxfs/xfs_dir2_sf.c for (i = 0; i < be32_to_cpu(btp->count); i++) { btp 569 fs/xfs/xfs_buf.c struct xfs_buftarg *btp, btp 588 fs/xfs/xfs_buf.c ASSERT(!(BBTOB(cmap.bm_len) < btp->bt_meta_sectorsize)); btp 589 fs/xfs/xfs_buf.c ASSERT(!(BBTOB(cmap.bm_bn) & (xfs_off_t)btp->bt_meta_sectormask)); btp 595 fs/xfs/xfs_buf.c eofs = XFS_FSB_TO_BB(btp->bt_mount, btp->bt_mount->m_sb.sb_dblocks); btp 597 fs/xfs/xfs_buf.c xfs_alert(btp->bt_mount, btp 604 fs/xfs/xfs_buf.c pag = xfs_perag_get(btp->bt_mount, btp 605 fs/xfs/xfs_buf.c xfs_daddr_to_agno(btp->bt_mount, cmap.bm_bn)); btp 617 fs/xfs/xfs_buf.c XFS_STATS_INC(btp->bt_mount, xb_miss_locked); btp 638 fs/xfs/xfs_buf.c XFS_STATS_INC(btp->bt_mount, xb_busy_locked); btp 642 fs/xfs/xfs_buf.c XFS_STATS_INC(btp->bt_mount, xb_get_locked_waited); btp 658 fs/xfs/xfs_buf.c XFS_STATS_INC(btp->bt_mount, xb_get_locked); btp 1588 fs/xfs/xfs_buf.c struct xfs_buftarg *btp) btp 1605 fs/xfs/xfs_buf.c while (percpu_counter_sum(&btp->bt_io_count)) btp 1607 fs/xfs/xfs_buf.c flush_workqueue(btp->bt_mount->m_buf_workqueue); btp 1610 fs/xfs/xfs_buf.c while (list_lru_count(&btp->bt_lru)) { btp 1611 fs/xfs/xfs_buf.c list_lru_walk(&btp->bt_lru, xfs_buftarg_wait_rele, btp 1619 fs/xfs/xfs_buf.c xfs_alert(btp->bt_mount, btp 1622 fs/xfs/xfs_buf.c xfs_alert(btp->bt_mount, btp 1669 fs/xfs/xfs_buf.c struct xfs_buftarg *btp = container_of(shrink, btp 1674 fs/xfs/xfs_buf.c freed = list_lru_shrink_walk(&btp->bt_lru, sc, btp 1692 fs/xfs/xfs_buf.c struct xfs_buftarg *btp = container_of(shrink, btp 1694 fs/xfs/xfs_buf.c return list_lru_shrink_count(&btp->bt_lru, sc); btp 1699 fs/xfs/xfs_buf.c struct xfs_buftarg *btp) btp 1701 fs/xfs/xfs_buf.c unregister_shrinker(&btp->bt_shrinker); btp 1702 fs/xfs/xfs_buf.c ASSERT(percpu_counter_sum(&btp->bt_io_count) == 0); btp 1703 fs/xfs/xfs_buf.c percpu_counter_destroy(&btp->bt_io_count); btp 1704 fs/xfs/xfs_buf.c list_lru_destroy(&btp->bt_lru); btp 1706 fs/xfs/xfs_buf.c xfs_blkdev_issue_flush(btp); btp 1708 fs/xfs/xfs_buf.c kmem_free(btp); btp 1713 fs/xfs/xfs_buf.c xfs_buftarg_t *btp, btp 1717 fs/xfs/xfs_buf.c btp->bt_meta_sectorsize = sectorsize; btp 1718 fs/xfs/xfs_buf.c btp->bt_meta_sectormask = sectorsize - 1; btp 1720 fs/xfs/xfs_buf.c if (set_blocksize(btp->bt_bdev, sectorsize)) { btp 1721 fs/xfs/xfs_buf.c xfs_warn(btp->bt_mount, btp 1723 fs/xfs/xfs_buf.c sectorsize, btp->bt_bdev); btp 1728 fs/xfs/xfs_buf.c btp->bt_logical_sectorsize = bdev_logical_block_size(btp->bt_bdev); btp 1729 fs/xfs/xfs_buf.c btp->bt_logical_sectormask = bdev_logical_block_size(btp->bt_bdev) - 1; btp 1741 fs/xfs/xfs_buf.c xfs_buftarg_t *btp, btp 1744 fs/xfs/xfs_buf.c return xfs_setsize_buftarg(btp, bdev_logical_block_size(bdev)); btp 1753 fs/xfs/xfs_buf.c xfs_buftarg_t *btp; btp 1755 fs/xfs/xfs_buf.c btp = kmem_zalloc(sizeof(*btp), KM_NOFS); btp 1757 fs/xfs/xfs_buf.c btp->bt_mount = mp; btp 1758 fs/xfs/xfs_buf.c btp->bt_dev = bdev->bd_dev; btp 1759 fs/xfs/xfs_buf.c btp->bt_bdev = bdev; btp 1760 fs/xfs/xfs_buf.c btp->bt_daxdev = dax_dev; btp 1762 fs/xfs/xfs_buf.c if (xfs_setsize_buftarg_early(btp, bdev)) btp 1765 fs/xfs/xfs_buf.c if (list_lru_init(&btp->bt_lru)) btp 1768 fs/xfs/xfs_buf.c if (percpu_counter_init(&btp->bt_io_count, 0, GFP_KERNEL)) btp 1771 fs/xfs/xfs_buf.c btp->bt_shrinker.count_objects = xfs_buftarg_shrink_count; btp 1772 fs/xfs/xfs_buf.c btp->bt_shrinker.scan_objects = xfs_buftarg_shrink_scan; btp 1773 fs/xfs/xfs_buf.c btp->bt_shrinker.seeks = DEFAULT_SEEKS; btp 1774 fs/xfs/xfs_buf.c btp->bt_shrinker.flags = SHRINKER_NUMA_AWARE; btp 1775 fs/xfs/xfs_buf.c if (register_shrinker(&btp->bt_shrinker)) btp 1777 fs/xfs/xfs_buf.c return btp; btp 1780 fs/xfs/xfs_buf.c percpu_counter_destroy(&btp->bt_io_count); btp 1782 fs/xfs/xfs_buf.c list_lru_destroy(&btp->bt_lru); btp 1784 fs/xfs/xfs_buf.c kmem_free(btp); btp 129 fs/xfs/xfs_pwork.c struct xfs_buftarg *btp = mp->m_ddev_targp; btp 135 fs/xfs/xfs_pwork.c return blk_queue_nonrot(btp->bt_bdev->bd_queue) ? 2 : 1; btp 481 include/linux/trace_events.h int bpf_probe_register(struct bpf_raw_event_map *btp, struct bpf_prog *prog); btp 482 include/linux/trace_events.h int bpf_probe_unregister(struct bpf_raw_event_map *btp, struct bpf_prog *prog); btp 484 include/linux/trace_events.h void bpf_put_raw_tracepoint(struct bpf_raw_event_map *btp); btp 507 include/linux/trace_events.h static inline int bpf_probe_register(struct bpf_raw_event_map *btp, struct bpf_prog *p) btp 511 include/linux/trace_events.h static inline int bpf_probe_unregister(struct bpf_raw_event_map *btp, struct bpf_prog *p) btp 519 include/linux/trace_events.h static inline void bpf_put_raw_tracepoint(struct bpf_raw_event_map *btp) btp 1793 kernel/bpf/syscall.c struct bpf_raw_event_map *btp; btp 1802 kernel/bpf/syscall.c bpf_probe_unregister(raw_tp->btp, raw_tp->prog); btp 1805 kernel/bpf/syscall.c bpf_put_raw_tracepoint(raw_tp->btp); btp 1821 kernel/bpf/syscall.c struct bpf_raw_event_map *btp; btp 1831 kernel/bpf/syscall.c btp = bpf_get_raw_tracepoint(tp_name); btp 1832 kernel/bpf/syscall.c if (!btp) btp 1840 kernel/bpf/syscall.c raw_tp->btp = btp; btp 1853 kernel/bpf/syscall.c err = bpf_probe_register(raw_tp->btp, prog); btp 1861 kernel/bpf/syscall.c bpf_probe_unregister(raw_tp->btp, prog); btp 1872 kernel/bpf/syscall.c bpf_put_raw_tracepoint(btp); btp 2808 kernel/bpf/syscall.c struct bpf_raw_event_map *btp = raw_tp->btp; btp 2813 kernel/bpf/syscall.c btp->tp->name, 0, 0); btp 36 kernel/trace/bpf_trace.c struct bpf_raw_event_map *btp, *ret = NULL; btp 43 kernel/trace/bpf_trace.c btp = &btm->module->bpf_raw_events[i]; btp 44 kernel/trace/bpf_trace.c if (!strcmp(btp->tp->name, name)) { btp 46 kernel/trace/bpf_trace.c ret = btp; btp 1311 kernel/trace/bpf_trace.c struct bpf_raw_event_map *btp = __start__bpf_raw_tp; btp 1313 kernel/trace/bpf_trace.c for (; btp < __stop__bpf_raw_tp; btp++) { btp 1314 kernel/trace/bpf_trace.c if (!strcmp(btp->tp->name, name)) btp 1315 kernel/trace/bpf_trace.c return btp; btp 1321 kernel/trace/bpf_trace.c void bpf_put_raw_tracepoint(struct bpf_raw_event_map *btp) btp 1323 kernel/trace/bpf_trace.c struct module *mod = __module_address((unsigned long)btp); btp 1384 kernel/trace/bpf_trace.c static int __bpf_probe_register(struct bpf_raw_event_map *btp, struct bpf_prog *prog) btp 1386 kernel/trace/bpf_trace.c struct tracepoint *tp = btp->tp; btp 1392 kernel/trace/bpf_trace.c if (prog->aux->max_ctx_offset > btp->num_args * sizeof(u64)) btp 1395 kernel/trace/bpf_trace.c if (prog->aux->max_tp_access > btp->writable_size) btp 1398 kernel/trace/bpf_trace.c return tracepoint_probe_register(tp, (void *)btp->bpf_func, prog); btp 1401 kernel/trace/bpf_trace.c int bpf_probe_register(struct bpf_raw_event_map *btp, struct bpf_prog *prog) btp 1403 kernel/trace/bpf_trace.c return __bpf_probe_register(btp, prog); btp 1406 kernel/trace/bpf_trace.c int bpf_probe_unregister(struct bpf_raw_event_map *btp, struct bpf_prog *prog) btp 1408 kernel/trace/bpf_trace.c return tracepoint_probe_unregister(btp->tp, (void *)btp->bpf_func, prog);