htab 46 arch/powerpc/kvm/book3s_32_mmu_host.c static ulong htab; htab 110 arch/powerpc/kvm/book3s_32_mmu_host.c ulong pteg = htab; htab 123 arch/powerpc/kvm/book3s_32_mmu_host.c htab, hash, htabmask, pteg); htab 383 arch/powerpc/kvm/book3s_32_mmu_host.c htab = (ulong)__va(sdr1 & 0xffff0000); htab 48 drivers/s390/char/sclp_rw.c sclp_make_buffer(void *page, unsigned short columns, unsigned short htab) htab 66 drivers/s390/char/sclp_rw.c buffer->htab = htab; htab 242 drivers/s390/char/sclp_rw.c } while (buffer->current_length % buffer->htab); htab 352 drivers/s390/char/sclp_rw.c sclp_set_htab(struct sclp_buffer *buffer, unsigned short htab) htab 354 drivers/s390/char/sclp_rw.c buffer->htab = htab; htab 75 drivers/s390/char/sclp_rw.h unsigned short htab; htab 47 kernel/bpf/hashtab.c struct bpf_htab *htab; htab 62 kernel/bpf/hashtab.c static bool htab_is_lru(const struct bpf_htab *htab) htab 64 kernel/bpf/hashtab.c return htab->map.map_type == BPF_MAP_TYPE_LRU_HASH || htab 65 kernel/bpf/hashtab.c htab->map.map_type == BPF_MAP_TYPE_LRU_PERCPU_HASH; htab 68 kernel/bpf/hashtab.c static bool htab_is_percpu(const struct bpf_htab *htab) htab 70 kernel/bpf/hashtab.c return htab->map.map_type == BPF_MAP_TYPE_PERCPU_HASH || htab 71 kernel/bpf/hashtab.c htab->map.map_type == BPF_MAP_TYPE_LRU_PERCPU_HASH; htab 74 kernel/bpf/hashtab.c static bool htab_is_prealloc(const struct bpf_htab *htab) htab 76 kernel/bpf/hashtab.c return !(htab->map.map_flags & BPF_F_NO_PREALLOC); htab 95 kernel/bpf/hashtab.c static struct htab_elem *get_htab_elem(struct bpf_htab *htab, int i) htab 97 kernel/bpf/hashtab.c return (struct htab_elem *) (htab->elems + i * htab->elem_size); htab 100 kernel/bpf/hashtab.c static void htab_free_elems(struct bpf_htab *htab) htab 104 kernel/bpf/hashtab.c if (!htab_is_percpu(htab)) htab 107 kernel/bpf/hashtab.c for (i = 0; i < htab->map.max_entries; i++) { htab 110 kernel/bpf/hashtab.c pptr = htab_elem_get_ptr(get_htab_elem(htab, i), htab 111 kernel/bpf/hashtab.c htab->map.key_size); htab 116 kernel/bpf/hashtab.c bpf_map_area_free(htab->elems); htab 119 kernel/bpf/hashtab.c static struct htab_elem *prealloc_lru_pop(struct bpf_htab *htab, void *key, htab 122 kernel/bpf/hashtab.c struct bpf_lru_node *node = bpf_lru_pop_free(&htab->lru, hash); htab 127 kernel/bpf/hashtab.c memcpy(l->key, key, htab->map.key_size); htab 134 kernel/bpf/hashtab.c static int prealloc_init(struct bpf_htab *htab) htab 136 kernel/bpf/hashtab.c u32 num_entries = htab->map.max_entries; htab 139 kernel/bpf/hashtab.c if (!htab_is_percpu(htab) && !htab_is_lru(htab)) htab 142 kernel/bpf/hashtab.c htab->elems = bpf_map_area_alloc(htab->elem_size * num_entries, htab 143 kernel/bpf/hashtab.c htab->map.numa_node); htab 144 kernel/bpf/hashtab.c if (!htab->elems) htab 147 kernel/bpf/hashtab.c if (!htab_is_percpu(htab)) htab 151 kernel/bpf/hashtab.c u32 size = round_up(htab->map.value_size, 8); htab 157 kernel/bpf/hashtab.c htab_elem_set_ptr(get_htab_elem(htab, i), htab->map.key_size, htab 163 kernel/bpf/hashtab.c if (htab_is_lru(htab)) htab 164 kernel/bpf/hashtab.c err = bpf_lru_init(&htab->lru, htab 165 kernel/bpf/hashtab.c htab->map.map_flags & BPF_F_NO_COMMON_LRU, htab 169 kernel/bpf/hashtab.c htab); htab 171 kernel/bpf/hashtab.c err = pcpu_freelist_init(&htab->freelist); htab 176 kernel/bpf/hashtab.c if (htab_is_lru(htab)) htab 177 kernel/bpf/hashtab.c bpf_lru_populate(&htab->lru, htab->elems, htab 179 kernel/bpf/hashtab.c htab->elem_size, num_entries); htab 181 kernel/bpf/hashtab.c pcpu_freelist_populate(&htab->freelist, htab 182 kernel/bpf/hashtab.c htab->elems + offsetof(struct htab_elem, fnode), htab 183 kernel/bpf/hashtab.c htab->elem_size, num_entries); htab 188 kernel/bpf/hashtab.c htab_free_elems(htab); htab 192 kernel/bpf/hashtab.c static void prealloc_destroy(struct bpf_htab *htab) htab 194 kernel/bpf/hashtab.c htab_free_elems(htab); htab 196 kernel/bpf/hashtab.c if (htab_is_lru(htab)) htab 197 kernel/bpf/hashtab.c bpf_lru_destroy(&htab->lru); htab 199 kernel/bpf/hashtab.c pcpu_freelist_destroy(&htab->freelist); htab 202 kernel/bpf/hashtab.c static int alloc_extra_elems(struct bpf_htab *htab) htab 214 kernel/bpf/hashtab.c l = pcpu_freelist_pop(&htab->freelist); htab 221 kernel/bpf/hashtab.c htab->extra_elems = pptr; htab 242 kernel/bpf/hashtab.c BUILD_BUG_ON(offsetof(struct htab_elem, htab) != htab 308 kernel/bpf/hashtab.c struct bpf_htab *htab; htab 312 kernel/bpf/hashtab.c htab = kzalloc(sizeof(*htab), GFP_USER); htab 313 kernel/bpf/hashtab.c if (!htab) htab 316 kernel/bpf/hashtab.c bpf_map_init_from_attr(&htab->map, attr); htab 323 kernel/bpf/hashtab.c htab->map.max_entries = roundup(attr->max_entries, htab 325 kernel/bpf/hashtab.c if (htab->map.max_entries < attr->max_entries) htab 326 kernel/bpf/hashtab.c htab->map.max_entries = rounddown(attr->max_entries, htab 331 kernel/bpf/hashtab.c htab->n_buckets = roundup_pow_of_two(htab->map.max_entries); htab 333 kernel/bpf/hashtab.c htab->elem_size = sizeof(struct htab_elem) + htab 334 kernel/bpf/hashtab.c round_up(htab->map.key_size, 8); htab 336 kernel/bpf/hashtab.c htab->elem_size += sizeof(void *); htab 338 kernel/bpf/hashtab.c htab->elem_size += round_up(htab->map.value_size, 8); htab 342 kernel/bpf/hashtab.c if (htab->n_buckets == 0 || htab 343 kernel/bpf/hashtab.c htab->n_buckets > U32_MAX / sizeof(struct bucket)) htab 346 kernel/bpf/hashtab.c cost = (u64) htab->n_buckets * sizeof(struct bucket) + htab 347 kernel/bpf/hashtab.c (u64) htab->elem_size * htab->map.max_entries; htab 350 kernel/bpf/hashtab.c cost += (u64) round_up(htab->map.value_size, 8) * htab 351 kernel/bpf/hashtab.c num_possible_cpus() * htab->map.max_entries; htab 353 kernel/bpf/hashtab.c cost += (u64) htab->elem_size * num_possible_cpus(); htab 356 kernel/bpf/hashtab.c err = bpf_map_charge_init(&htab->map.memory, cost); htab 361 kernel/bpf/hashtab.c htab->buckets = bpf_map_area_alloc(htab->n_buckets * htab 363 kernel/bpf/hashtab.c htab->map.numa_node); htab 364 kernel/bpf/hashtab.c if (!htab->buckets) htab 367 kernel/bpf/hashtab.c if (htab->map.map_flags & BPF_F_ZERO_SEED) htab 368 kernel/bpf/hashtab.c htab->hashrnd = 0; htab 370 kernel/bpf/hashtab.c htab->hashrnd = get_random_int(); htab 372 kernel/bpf/hashtab.c for (i = 0; i < htab->n_buckets; i++) { htab 373 kernel/bpf/hashtab.c INIT_HLIST_NULLS_HEAD(&htab->buckets[i].head, i); htab 374 kernel/bpf/hashtab.c raw_spin_lock_init(&htab->buckets[i].lock); htab 378 kernel/bpf/hashtab.c err = prealloc_init(htab); htab 386 kernel/bpf/hashtab.c err = alloc_extra_elems(htab); htab 392 kernel/bpf/hashtab.c return &htab->map; htab 395 kernel/bpf/hashtab.c prealloc_destroy(htab); htab 397 kernel/bpf/hashtab.c bpf_map_area_free(htab->buckets); htab 399 kernel/bpf/hashtab.c bpf_map_charge_finish(&htab->map.memory); htab 401 kernel/bpf/hashtab.c kfree(htab); htab 410 kernel/bpf/hashtab.c static inline struct bucket *__select_bucket(struct bpf_htab *htab, u32 hash) htab 412 kernel/bpf/hashtab.c return &htab->buckets[hash & (htab->n_buckets - 1)]; htab 415 kernel/bpf/hashtab.c static inline struct hlist_nulls_head *select_bucket(struct bpf_htab *htab, u32 hash) htab 417 kernel/bpf/hashtab.c return &__select_bucket(htab, hash)->head; htab 463 kernel/bpf/hashtab.c struct bpf_htab *htab = container_of(map, struct bpf_htab, map); htab 473 kernel/bpf/hashtab.c hash = htab_map_hash(key, key_size, htab->hashrnd); htab 475 kernel/bpf/hashtab.c head = select_bucket(htab, hash); htab 477 kernel/bpf/hashtab.c l = lookup_nulls_elem_raw(head, hash, key, key_size, htab->n_buckets); htab 572 kernel/bpf/hashtab.c struct bpf_htab *htab = (struct bpf_htab *)arg; htab 580 kernel/bpf/hashtab.c b = __select_bucket(htab, tgt_l->hash); htab 599 kernel/bpf/hashtab.c struct bpf_htab *htab = container_of(map, struct bpf_htab, map); htab 612 kernel/bpf/hashtab.c hash = htab_map_hash(key, key_size, htab->hashrnd); htab 614 kernel/bpf/hashtab.c head = select_bucket(htab, hash); htab 617 kernel/bpf/hashtab.c l = lookup_nulls_elem_raw(head, hash, key, key_size, htab->n_buckets); htab 633 kernel/bpf/hashtab.c i = hash & (htab->n_buckets - 1); htab 638 kernel/bpf/hashtab.c for (; i < htab->n_buckets; i++) { htab 639 kernel/bpf/hashtab.c head = select_bucket(htab, i); htab 655 kernel/bpf/hashtab.c static void htab_elem_free(struct bpf_htab *htab, struct htab_elem *l) htab 657 kernel/bpf/hashtab.c if (htab->map.map_type == BPF_MAP_TYPE_PERCPU_HASH) htab 658 kernel/bpf/hashtab.c free_percpu(htab_elem_get_ptr(l, htab->map.key_size)); htab 665 kernel/bpf/hashtab.c struct bpf_htab *htab = l->htab; htab 673 kernel/bpf/hashtab.c htab_elem_free(htab, l); htab 678 kernel/bpf/hashtab.c static void free_htab_elem(struct bpf_htab *htab, struct htab_elem *l) htab 680 kernel/bpf/hashtab.c struct bpf_map *map = &htab->map; htab 688 kernel/bpf/hashtab.c if (htab_is_prealloc(htab)) { htab 689 kernel/bpf/hashtab.c __pcpu_freelist_push(&htab->freelist, &l->fnode); htab 691 kernel/bpf/hashtab.c atomic_dec(&htab->count); htab 692 kernel/bpf/hashtab.c l->htab = htab; htab 697 kernel/bpf/hashtab.c static void pcpu_copy_value(struct bpf_htab *htab, void __percpu *pptr, htab 702 kernel/bpf/hashtab.c memcpy(this_cpu_ptr(pptr), value, htab->map.value_size); htab 704 kernel/bpf/hashtab.c u32 size = round_up(htab->map.value_size, 8); htab 715 kernel/bpf/hashtab.c static bool fd_htab_map_needs_adjust(const struct bpf_htab *htab) htab 717 kernel/bpf/hashtab.c return htab->map.map_type == BPF_MAP_TYPE_HASH_OF_MAPS && htab 721 kernel/bpf/hashtab.c static struct htab_elem *alloc_htab_elem(struct bpf_htab *htab, void *key, htab 726 kernel/bpf/hashtab.c u32 size = htab->map.value_size; htab 727 kernel/bpf/hashtab.c bool prealloc = htab_is_prealloc(htab); htab 736 kernel/bpf/hashtab.c pl_new = this_cpu_ptr(htab->extra_elems); htab 742 kernel/bpf/hashtab.c l = __pcpu_freelist_pop(&htab->freelist); htab 748 kernel/bpf/hashtab.c if (atomic_inc_return(&htab->count) > htab->map.max_entries) htab 758 kernel/bpf/hashtab.c l_new = kmalloc_node(htab->elem_size, GFP_ATOMIC | __GFP_NOWARN, htab 759 kernel/bpf/hashtab.c htab->map.numa_node); htab 764 kernel/bpf/hashtab.c check_and_init_map_lock(&htab->map, htab 784 kernel/bpf/hashtab.c pcpu_copy_value(htab, pptr, value, onallcpus); htab 788 kernel/bpf/hashtab.c } else if (fd_htab_map_needs_adjust(htab)) { htab 792 kernel/bpf/hashtab.c copy_map_value(&htab->map, htab 800 kernel/bpf/hashtab.c atomic_dec(&htab->count); htab 804 kernel/bpf/hashtab.c static int check_flags(struct bpf_htab *htab, struct htab_elem *l_old, htab 822 kernel/bpf/hashtab.c struct bpf_htab *htab = container_of(map, struct bpf_htab, map); htab 838 kernel/bpf/hashtab.c hash = htab_map_hash(key, key_size, htab->hashrnd); htab 840 kernel/bpf/hashtab.c b = __select_bucket(htab, hash); htab 848 kernel/bpf/hashtab.c htab->n_buckets); htab 849 kernel/bpf/hashtab.c ret = check_flags(htab, l_old, map_flags); htab 870 kernel/bpf/hashtab.c ret = check_flags(htab, l_old, map_flags); htab 888 kernel/bpf/hashtab.c l_new = alloc_htab_elem(htab, key, value, key_size, hash, false, false, htab 902 kernel/bpf/hashtab.c if (!htab_is_prealloc(htab)) htab 903 kernel/bpf/hashtab.c free_htab_elem(htab, l_old); htab 914 kernel/bpf/hashtab.c struct bpf_htab *htab = container_of(map, struct bpf_htab, map); htab 930 kernel/bpf/hashtab.c hash = htab_map_hash(key, key_size, htab->hashrnd); htab 932 kernel/bpf/hashtab.c b = __select_bucket(htab, hash); htab 940 kernel/bpf/hashtab.c l_new = prealloc_lru_pop(htab, key, hash); htab 950 kernel/bpf/hashtab.c ret = check_flags(htab, l_old, map_flags); htab 968 kernel/bpf/hashtab.c bpf_lru_push_free(&htab->lru, &l_new->lru_node); htab 970 kernel/bpf/hashtab.c bpf_lru_push_free(&htab->lru, &l_old->lru_node); htab 979 kernel/bpf/hashtab.c struct bpf_htab *htab = container_of(map, struct bpf_htab, map); htab 995 kernel/bpf/hashtab.c hash = htab_map_hash(key, key_size, htab->hashrnd); htab 997 kernel/bpf/hashtab.c b = __select_bucket(htab, hash); htab 1005 kernel/bpf/hashtab.c ret = check_flags(htab, l_old, map_flags); htab 1011 kernel/bpf/hashtab.c pcpu_copy_value(htab, htab_elem_get_ptr(l_old, key_size), htab 1014 kernel/bpf/hashtab.c l_new = alloc_htab_elem(htab, key, value, key_size, htab 1032 kernel/bpf/hashtab.c struct bpf_htab *htab = container_of(map, struct bpf_htab, map); htab 1048 kernel/bpf/hashtab.c hash = htab_map_hash(key, key_size, htab->hashrnd); htab 1050 kernel/bpf/hashtab.c b = __select_bucket(htab, hash); htab 1059 kernel/bpf/hashtab.c l_new = prealloc_lru_pop(htab, key, hash); htab 1069 kernel/bpf/hashtab.c ret = check_flags(htab, l_old, map_flags); htab 1077 kernel/bpf/hashtab.c pcpu_copy_value(htab, htab_elem_get_ptr(l_old, key_size), htab 1080 kernel/bpf/hashtab.c pcpu_copy_value(htab, htab_elem_get_ptr(l_new, key_size), htab 1089 kernel/bpf/hashtab.c bpf_lru_push_free(&htab->lru, &l_new->lru_node); htab 1109 kernel/bpf/hashtab.c struct bpf_htab *htab = container_of(map, struct bpf_htab, map); htab 1121 kernel/bpf/hashtab.c hash = htab_map_hash(key, key_size, htab->hashrnd); htab 1122 kernel/bpf/hashtab.c b = __select_bucket(htab, hash); htab 1131 kernel/bpf/hashtab.c free_htab_elem(htab, l); htab 1141 kernel/bpf/hashtab.c struct bpf_htab *htab = container_of(map, struct bpf_htab, map); htab 1153 kernel/bpf/hashtab.c hash = htab_map_hash(key, key_size, htab->hashrnd); htab 1154 kernel/bpf/hashtab.c b = __select_bucket(htab, hash); htab 1168 kernel/bpf/hashtab.c bpf_lru_push_free(&htab->lru, &l->lru_node); htab 1172 kernel/bpf/hashtab.c static void delete_all_elements(struct bpf_htab *htab) htab 1176 kernel/bpf/hashtab.c for (i = 0; i < htab->n_buckets; i++) { htab 1177 kernel/bpf/hashtab.c struct hlist_nulls_head *head = select_bucket(htab, i); htab 1183 kernel/bpf/hashtab.c htab_elem_free(htab, l); htab 1191 kernel/bpf/hashtab.c struct bpf_htab *htab = container_of(map, struct bpf_htab, map); htab 1204 kernel/bpf/hashtab.c if (!htab_is_prealloc(htab)) htab 1205 kernel/bpf/hashtab.c delete_all_elements(htab); htab 1207 kernel/bpf/hashtab.c prealloc_destroy(htab); htab 1209 kernel/bpf/hashtab.c free_percpu(htab->extra_elems); htab 1210 kernel/bpf/hashtab.c bpf_map_area_free(htab->buckets); htab 1211 kernel/bpf/hashtab.c kfree(htab); htab 1318 kernel/bpf/hashtab.c struct bpf_htab *htab = container_of(map, struct bpf_htab, map); htab 1322 kernel/bpf/hashtab.c if (htab_is_lru(htab)) htab 1393 kernel/bpf/hashtab.c struct bpf_htab *htab = container_of(map, struct bpf_htab, map); htab 1399 kernel/bpf/hashtab.c for (i = 0; i < htab->n_buckets; i++) { htab 1400 kernel/bpf/hashtab.c head = select_bucket(htab, i); htab 545 net/core/sock_map.c static struct bpf_htab_bucket *sock_hash_select_bucket(struct bpf_htab *htab, htab 548 net/core/sock_map.c return &htab->buckets[hash & (htab->buckets_num - 1)]; htab 568 net/core/sock_map.c struct bpf_htab *htab = container_of(map, struct bpf_htab, map); htab 576 net/core/sock_map.c bucket = sock_hash_select_bucket(htab, hash); htab 582 net/core/sock_map.c static void sock_hash_free_elem(struct bpf_htab *htab, htab 585 net/core/sock_map.c atomic_dec(&htab->count); htab 592 net/core/sock_map.c struct bpf_htab *htab = container_of(map, struct bpf_htab, map); htab 597 net/core/sock_map.c bucket = sock_hash_select_bucket(htab, elem->hash); htab 609 net/core/sock_map.c sock_hash_free_elem(htab, elem); htab 616 net/core/sock_map.c struct bpf_htab *htab = container_of(map, struct bpf_htab, map); htab 623 net/core/sock_map.c bucket = sock_hash_select_bucket(htab, hash); htab 630 net/core/sock_map.c sock_hash_free_elem(htab, elem); htab 637 net/core/sock_map.c static struct bpf_htab_elem *sock_hash_alloc_elem(struct bpf_htab *htab, htab 644 net/core/sock_map.c if (atomic_inc_return(&htab->count) > htab->map.max_entries) { htab 646 net/core/sock_map.c atomic_dec(&htab->count); htab 651 net/core/sock_map.c new = kmalloc_node(htab->elem_size, GFP_ATOMIC | __GFP_NOWARN, htab 652 net/core/sock_map.c htab->map.numa_node); htab 654 net/core/sock_map.c atomic_dec(&htab->count); htab 666 net/core/sock_map.c struct bpf_htab *htab = container_of(map, struct bpf_htab, map); htab 685 net/core/sock_map.c ret = sock_map_link(map, &htab->progs, sk); htab 693 net/core/sock_map.c bucket = sock_hash_select_bucket(htab, hash); htab 705 net/core/sock_map.c elem_new = sock_hash_alloc_elem(htab, key, key_size, hash, sk, elem); htab 719 net/core/sock_map.c sock_hash_free_elem(htab, elem); htab 766 net/core/sock_map.c struct bpf_htab *htab = container_of(map, struct bpf_htab, map); htab 775 net/core/sock_map.c head = &sock_hash_select_bucket(htab, hash)->head; htab 787 net/core/sock_map.c i = hash & (htab->buckets_num - 1); htab 790 net/core/sock_map.c for (; i < htab->buckets_num; i++) { htab 791 net/core/sock_map.c head = &sock_hash_select_bucket(htab, i)->head; htab 805 net/core/sock_map.c struct bpf_htab *htab; htab 819 net/core/sock_map.c htab = kzalloc(sizeof(*htab), GFP_USER); htab 820 net/core/sock_map.c if (!htab) htab 823 net/core/sock_map.c bpf_map_init_from_attr(&htab->map, attr); htab 825 net/core/sock_map.c htab->buckets_num = roundup_pow_of_two(htab->map.max_entries); htab 826 net/core/sock_map.c htab->elem_size = sizeof(struct bpf_htab_elem) + htab 827 net/core/sock_map.c round_up(htab->map.key_size, 8); htab 828 net/core/sock_map.c if (htab->buckets_num == 0 || htab 829 net/core/sock_map.c htab->buckets_num > U32_MAX / sizeof(struct bpf_htab_bucket)) { htab 834 net/core/sock_map.c cost = (u64) htab->buckets_num * sizeof(struct bpf_htab_bucket) + htab 835 net/core/sock_map.c (u64) htab->elem_size * htab->map.max_entries; htab 841 net/core/sock_map.c htab->buckets = bpf_map_area_alloc(htab->buckets_num * htab 843 net/core/sock_map.c htab->map.numa_node); htab 844 net/core/sock_map.c if (!htab->buckets) { htab 849 net/core/sock_map.c for (i = 0; i < htab->buckets_num; i++) { htab 850 net/core/sock_map.c INIT_HLIST_HEAD(&htab->buckets[i].head); htab 851 net/core/sock_map.c raw_spin_lock_init(&htab->buckets[i].lock); htab 854 net/core/sock_map.c return &htab->map; htab 856 net/core/sock_map.c kfree(htab); htab 862 net/core/sock_map.c struct bpf_htab *htab = container_of(map, struct bpf_htab, map); htab 873 net/core/sock_map.c for (i = 0; i < htab->buckets_num; i++) { htab 874 net/core/sock_map.c bucket = sock_hash_select_bucket(htab, i); htab 891 net/core/sock_map.c bpf_map_area_free(htab->buckets); htab 892 net/core/sock_map.c kfree(htab); htab 4027 net/xfrm/xfrm_policy.c struct xfrm_policy_hash *htab; htab 4033 net/xfrm/xfrm_policy.c htab = &net->xfrm.policy_bydst[dir]; htab 4034 net/xfrm/xfrm_policy.c htab->table = xfrm_hash_alloc(sz); htab 4035 net/xfrm/xfrm_policy.c if (!htab->table) htab 4037 net/xfrm/xfrm_policy.c htab->hmask = hmask; htab 4038 net/xfrm/xfrm_policy.c htab->dbits4 = 32; htab 4039 net/xfrm/xfrm_policy.c htab->sbits4 = 32; htab 4040 net/xfrm/xfrm_policy.c htab->dbits6 = 128; htab 4041 net/xfrm/xfrm_policy.c htab->sbits6 = 128; htab 4058 net/xfrm/xfrm_policy.c struct xfrm_policy_hash *htab; htab 4060 net/xfrm/xfrm_policy.c htab = &net->xfrm.policy_bydst[dir]; htab 4061 net/xfrm/xfrm_policy.c xfrm_hash_free(htab->table, sz); htab 4083 net/xfrm/xfrm_policy.c struct xfrm_policy_hash *htab; htab 4087 net/xfrm/xfrm_policy.c htab = &net->xfrm.policy_bydst[dir]; htab 4088 net/xfrm/xfrm_policy.c sz = (htab->hmask + 1) * sizeof(struct hlist_head); htab 4089 net/xfrm/xfrm_policy.c WARN_ON(!hlist_empty(htab->table)); htab 4090 net/xfrm/xfrm_policy.c xfrm_hash_free(htab->table, sz); htab 5 tools/testing/selftests/bpf/progs/sample_map_ret0.c struct bpf_map_def SEC("maps") htab = { htab 26 tools/testing/selftests/bpf/progs/sample_map_ret0.c value = bpf_map_lookup_elem(&htab, &key);