htab               46 arch/powerpc/kvm/book3s_32_mmu_host.c static ulong htab;
htab              110 arch/powerpc/kvm/book3s_32_mmu_host.c 	ulong pteg = htab;
htab              123 arch/powerpc/kvm/book3s_32_mmu_host.c 		htab, hash, htabmask, pteg);
htab              383 arch/powerpc/kvm/book3s_32_mmu_host.c 	htab = (ulong)__va(sdr1 & 0xffff0000);
htab               48 drivers/s390/char/sclp_rw.c sclp_make_buffer(void *page, unsigned short columns, unsigned short htab)
htab               66 drivers/s390/char/sclp_rw.c 	buffer->htab = htab;
htab              242 drivers/s390/char/sclp_rw.c 			} while (buffer->current_length % buffer->htab);
htab              352 drivers/s390/char/sclp_rw.c sclp_set_htab(struct sclp_buffer *buffer, unsigned short htab)
htab              354 drivers/s390/char/sclp_rw.c 	buffer->htab = htab;
htab               75 drivers/s390/char/sclp_rw.h 	unsigned short htab;
htab               47 kernel/bpf/hashtab.c 				struct bpf_htab *htab;
htab               62 kernel/bpf/hashtab.c static bool htab_is_lru(const struct bpf_htab *htab)
htab               64 kernel/bpf/hashtab.c 	return htab->map.map_type == BPF_MAP_TYPE_LRU_HASH ||
htab               65 kernel/bpf/hashtab.c 		htab->map.map_type == BPF_MAP_TYPE_LRU_PERCPU_HASH;
htab               68 kernel/bpf/hashtab.c static bool htab_is_percpu(const struct bpf_htab *htab)
htab               70 kernel/bpf/hashtab.c 	return htab->map.map_type == BPF_MAP_TYPE_PERCPU_HASH ||
htab               71 kernel/bpf/hashtab.c 		htab->map.map_type == BPF_MAP_TYPE_LRU_PERCPU_HASH;
htab               74 kernel/bpf/hashtab.c static bool htab_is_prealloc(const struct bpf_htab *htab)
htab               76 kernel/bpf/hashtab.c 	return !(htab->map.map_flags & BPF_F_NO_PREALLOC);
htab               95 kernel/bpf/hashtab.c static struct htab_elem *get_htab_elem(struct bpf_htab *htab, int i)
htab               97 kernel/bpf/hashtab.c 	return (struct htab_elem *) (htab->elems + i * htab->elem_size);
htab              100 kernel/bpf/hashtab.c static void htab_free_elems(struct bpf_htab *htab)
htab              104 kernel/bpf/hashtab.c 	if (!htab_is_percpu(htab))
htab              107 kernel/bpf/hashtab.c 	for (i = 0; i < htab->map.max_entries; i++) {
htab              110 kernel/bpf/hashtab.c 		pptr = htab_elem_get_ptr(get_htab_elem(htab, i),
htab              111 kernel/bpf/hashtab.c 					 htab->map.key_size);
htab              116 kernel/bpf/hashtab.c 	bpf_map_area_free(htab->elems);
htab              119 kernel/bpf/hashtab.c static struct htab_elem *prealloc_lru_pop(struct bpf_htab *htab, void *key,
htab              122 kernel/bpf/hashtab.c 	struct bpf_lru_node *node = bpf_lru_pop_free(&htab->lru, hash);
htab              127 kernel/bpf/hashtab.c 		memcpy(l->key, key, htab->map.key_size);
htab              134 kernel/bpf/hashtab.c static int prealloc_init(struct bpf_htab *htab)
htab              136 kernel/bpf/hashtab.c 	u32 num_entries = htab->map.max_entries;
htab              139 kernel/bpf/hashtab.c 	if (!htab_is_percpu(htab) && !htab_is_lru(htab))
htab              142 kernel/bpf/hashtab.c 	htab->elems = bpf_map_area_alloc(htab->elem_size * num_entries,
htab              143 kernel/bpf/hashtab.c 					 htab->map.numa_node);
htab              144 kernel/bpf/hashtab.c 	if (!htab->elems)
htab              147 kernel/bpf/hashtab.c 	if (!htab_is_percpu(htab))
htab              151 kernel/bpf/hashtab.c 		u32 size = round_up(htab->map.value_size, 8);
htab              157 kernel/bpf/hashtab.c 		htab_elem_set_ptr(get_htab_elem(htab, i), htab->map.key_size,
htab              163 kernel/bpf/hashtab.c 	if (htab_is_lru(htab))
htab              164 kernel/bpf/hashtab.c 		err = bpf_lru_init(&htab->lru,
htab              165 kernel/bpf/hashtab.c 				   htab->map.map_flags & BPF_F_NO_COMMON_LRU,
htab              169 kernel/bpf/hashtab.c 				   htab);
htab              171 kernel/bpf/hashtab.c 		err = pcpu_freelist_init(&htab->freelist);
htab              176 kernel/bpf/hashtab.c 	if (htab_is_lru(htab))
htab              177 kernel/bpf/hashtab.c 		bpf_lru_populate(&htab->lru, htab->elems,
htab              179 kernel/bpf/hashtab.c 				 htab->elem_size, num_entries);
htab              181 kernel/bpf/hashtab.c 		pcpu_freelist_populate(&htab->freelist,
htab              182 kernel/bpf/hashtab.c 				       htab->elems + offsetof(struct htab_elem, fnode),
htab              183 kernel/bpf/hashtab.c 				       htab->elem_size, num_entries);
htab              188 kernel/bpf/hashtab.c 	htab_free_elems(htab);
htab              192 kernel/bpf/hashtab.c static void prealloc_destroy(struct bpf_htab *htab)
htab              194 kernel/bpf/hashtab.c 	htab_free_elems(htab);
htab              196 kernel/bpf/hashtab.c 	if (htab_is_lru(htab))
htab              197 kernel/bpf/hashtab.c 		bpf_lru_destroy(&htab->lru);
htab              199 kernel/bpf/hashtab.c 		pcpu_freelist_destroy(&htab->freelist);
htab              202 kernel/bpf/hashtab.c static int alloc_extra_elems(struct bpf_htab *htab)
htab              214 kernel/bpf/hashtab.c 		l = pcpu_freelist_pop(&htab->freelist);
htab              221 kernel/bpf/hashtab.c 	htab->extra_elems = pptr;
htab              242 kernel/bpf/hashtab.c 	BUILD_BUG_ON(offsetof(struct htab_elem, htab) !=
htab              308 kernel/bpf/hashtab.c 	struct bpf_htab *htab;
htab              312 kernel/bpf/hashtab.c 	htab = kzalloc(sizeof(*htab), GFP_USER);
htab              313 kernel/bpf/hashtab.c 	if (!htab)
htab              316 kernel/bpf/hashtab.c 	bpf_map_init_from_attr(&htab->map, attr);
htab              323 kernel/bpf/hashtab.c 		htab->map.max_entries = roundup(attr->max_entries,
htab              325 kernel/bpf/hashtab.c 		if (htab->map.max_entries < attr->max_entries)
htab              326 kernel/bpf/hashtab.c 			htab->map.max_entries = rounddown(attr->max_entries,
htab              331 kernel/bpf/hashtab.c 	htab->n_buckets = roundup_pow_of_two(htab->map.max_entries);
htab              333 kernel/bpf/hashtab.c 	htab->elem_size = sizeof(struct htab_elem) +
htab              334 kernel/bpf/hashtab.c 			  round_up(htab->map.key_size, 8);
htab              336 kernel/bpf/hashtab.c 		htab->elem_size += sizeof(void *);
htab              338 kernel/bpf/hashtab.c 		htab->elem_size += round_up(htab->map.value_size, 8);
htab              342 kernel/bpf/hashtab.c 	if (htab->n_buckets == 0 ||
htab              343 kernel/bpf/hashtab.c 	    htab->n_buckets > U32_MAX / sizeof(struct bucket))
htab              346 kernel/bpf/hashtab.c 	cost = (u64) htab->n_buckets * sizeof(struct bucket) +
htab              347 kernel/bpf/hashtab.c 	       (u64) htab->elem_size * htab->map.max_entries;
htab              350 kernel/bpf/hashtab.c 		cost += (u64) round_up(htab->map.value_size, 8) *
htab              351 kernel/bpf/hashtab.c 			num_possible_cpus() * htab->map.max_entries;
htab              353 kernel/bpf/hashtab.c 	       cost += (u64) htab->elem_size * num_possible_cpus();
htab              356 kernel/bpf/hashtab.c 	err = bpf_map_charge_init(&htab->map.memory, cost);
htab              361 kernel/bpf/hashtab.c 	htab->buckets = bpf_map_area_alloc(htab->n_buckets *
htab              363 kernel/bpf/hashtab.c 					   htab->map.numa_node);
htab              364 kernel/bpf/hashtab.c 	if (!htab->buckets)
htab              367 kernel/bpf/hashtab.c 	if (htab->map.map_flags & BPF_F_ZERO_SEED)
htab              368 kernel/bpf/hashtab.c 		htab->hashrnd = 0;
htab              370 kernel/bpf/hashtab.c 		htab->hashrnd = get_random_int();
htab              372 kernel/bpf/hashtab.c 	for (i = 0; i < htab->n_buckets; i++) {
htab              373 kernel/bpf/hashtab.c 		INIT_HLIST_NULLS_HEAD(&htab->buckets[i].head, i);
htab              374 kernel/bpf/hashtab.c 		raw_spin_lock_init(&htab->buckets[i].lock);
htab              378 kernel/bpf/hashtab.c 		err = prealloc_init(htab);
htab              386 kernel/bpf/hashtab.c 			err = alloc_extra_elems(htab);
htab              392 kernel/bpf/hashtab.c 	return &htab->map;
htab              395 kernel/bpf/hashtab.c 	prealloc_destroy(htab);
htab              397 kernel/bpf/hashtab.c 	bpf_map_area_free(htab->buckets);
htab              399 kernel/bpf/hashtab.c 	bpf_map_charge_finish(&htab->map.memory);
htab              401 kernel/bpf/hashtab.c 	kfree(htab);
htab              410 kernel/bpf/hashtab.c static inline struct bucket *__select_bucket(struct bpf_htab *htab, u32 hash)
htab              412 kernel/bpf/hashtab.c 	return &htab->buckets[hash & (htab->n_buckets - 1)];
htab              415 kernel/bpf/hashtab.c static inline struct hlist_nulls_head *select_bucket(struct bpf_htab *htab, u32 hash)
htab              417 kernel/bpf/hashtab.c 	return &__select_bucket(htab, hash)->head;
htab              463 kernel/bpf/hashtab.c 	struct bpf_htab *htab = container_of(map, struct bpf_htab, map);
htab              473 kernel/bpf/hashtab.c 	hash = htab_map_hash(key, key_size, htab->hashrnd);
htab              475 kernel/bpf/hashtab.c 	head = select_bucket(htab, hash);
htab              477 kernel/bpf/hashtab.c 	l = lookup_nulls_elem_raw(head, hash, key, key_size, htab->n_buckets);
htab              572 kernel/bpf/hashtab.c 	struct bpf_htab *htab = (struct bpf_htab *)arg;
htab              580 kernel/bpf/hashtab.c 	b = __select_bucket(htab, tgt_l->hash);
htab              599 kernel/bpf/hashtab.c 	struct bpf_htab *htab = container_of(map, struct bpf_htab, map);
htab              612 kernel/bpf/hashtab.c 	hash = htab_map_hash(key, key_size, htab->hashrnd);
htab              614 kernel/bpf/hashtab.c 	head = select_bucket(htab, hash);
htab              617 kernel/bpf/hashtab.c 	l = lookup_nulls_elem_raw(head, hash, key, key_size, htab->n_buckets);
htab              633 kernel/bpf/hashtab.c 	i = hash & (htab->n_buckets - 1);
htab              638 kernel/bpf/hashtab.c 	for (; i < htab->n_buckets; i++) {
htab              639 kernel/bpf/hashtab.c 		head = select_bucket(htab, i);
htab              655 kernel/bpf/hashtab.c static void htab_elem_free(struct bpf_htab *htab, struct htab_elem *l)
htab              657 kernel/bpf/hashtab.c 	if (htab->map.map_type == BPF_MAP_TYPE_PERCPU_HASH)
htab              658 kernel/bpf/hashtab.c 		free_percpu(htab_elem_get_ptr(l, htab->map.key_size));
htab              665 kernel/bpf/hashtab.c 	struct bpf_htab *htab = l->htab;
htab              673 kernel/bpf/hashtab.c 	htab_elem_free(htab, l);
htab              678 kernel/bpf/hashtab.c static void free_htab_elem(struct bpf_htab *htab, struct htab_elem *l)
htab              680 kernel/bpf/hashtab.c 	struct bpf_map *map = &htab->map;
htab              688 kernel/bpf/hashtab.c 	if (htab_is_prealloc(htab)) {
htab              689 kernel/bpf/hashtab.c 		__pcpu_freelist_push(&htab->freelist, &l->fnode);
htab              691 kernel/bpf/hashtab.c 		atomic_dec(&htab->count);
htab              692 kernel/bpf/hashtab.c 		l->htab = htab;
htab              697 kernel/bpf/hashtab.c static void pcpu_copy_value(struct bpf_htab *htab, void __percpu *pptr,
htab              702 kernel/bpf/hashtab.c 		memcpy(this_cpu_ptr(pptr), value, htab->map.value_size);
htab              704 kernel/bpf/hashtab.c 		u32 size = round_up(htab->map.value_size, 8);
htab              715 kernel/bpf/hashtab.c static bool fd_htab_map_needs_adjust(const struct bpf_htab *htab)
htab              717 kernel/bpf/hashtab.c 	return htab->map.map_type == BPF_MAP_TYPE_HASH_OF_MAPS &&
htab              721 kernel/bpf/hashtab.c static struct htab_elem *alloc_htab_elem(struct bpf_htab *htab, void *key,
htab              726 kernel/bpf/hashtab.c 	u32 size = htab->map.value_size;
htab              727 kernel/bpf/hashtab.c 	bool prealloc = htab_is_prealloc(htab);
htab              736 kernel/bpf/hashtab.c 			pl_new = this_cpu_ptr(htab->extra_elems);
htab              742 kernel/bpf/hashtab.c 			l = __pcpu_freelist_pop(&htab->freelist);
htab              748 kernel/bpf/hashtab.c 		if (atomic_inc_return(&htab->count) > htab->map.max_entries)
htab              758 kernel/bpf/hashtab.c 		l_new = kmalloc_node(htab->elem_size, GFP_ATOMIC | __GFP_NOWARN,
htab              759 kernel/bpf/hashtab.c 				     htab->map.numa_node);
htab              764 kernel/bpf/hashtab.c 		check_and_init_map_lock(&htab->map,
htab              784 kernel/bpf/hashtab.c 		pcpu_copy_value(htab, pptr, value, onallcpus);
htab              788 kernel/bpf/hashtab.c 	} else if (fd_htab_map_needs_adjust(htab)) {
htab              792 kernel/bpf/hashtab.c 		copy_map_value(&htab->map,
htab              800 kernel/bpf/hashtab.c 	atomic_dec(&htab->count);
htab              804 kernel/bpf/hashtab.c static int check_flags(struct bpf_htab *htab, struct htab_elem *l_old,
htab              822 kernel/bpf/hashtab.c 	struct bpf_htab *htab = container_of(map, struct bpf_htab, map);
htab              838 kernel/bpf/hashtab.c 	hash = htab_map_hash(key, key_size, htab->hashrnd);
htab              840 kernel/bpf/hashtab.c 	b = __select_bucket(htab, hash);
htab              848 kernel/bpf/hashtab.c 					      htab->n_buckets);
htab              849 kernel/bpf/hashtab.c 		ret = check_flags(htab, l_old, map_flags);
htab              870 kernel/bpf/hashtab.c 	ret = check_flags(htab, l_old, map_flags);
htab              888 kernel/bpf/hashtab.c 	l_new = alloc_htab_elem(htab, key, value, key_size, hash, false, false,
htab              902 kernel/bpf/hashtab.c 		if (!htab_is_prealloc(htab))
htab              903 kernel/bpf/hashtab.c 			free_htab_elem(htab, l_old);
htab              914 kernel/bpf/hashtab.c 	struct bpf_htab *htab = container_of(map, struct bpf_htab, map);
htab              930 kernel/bpf/hashtab.c 	hash = htab_map_hash(key, key_size, htab->hashrnd);
htab              932 kernel/bpf/hashtab.c 	b = __select_bucket(htab, hash);
htab              940 kernel/bpf/hashtab.c 	l_new = prealloc_lru_pop(htab, key, hash);
htab              950 kernel/bpf/hashtab.c 	ret = check_flags(htab, l_old, map_flags);
htab              968 kernel/bpf/hashtab.c 		bpf_lru_push_free(&htab->lru, &l_new->lru_node);
htab              970 kernel/bpf/hashtab.c 		bpf_lru_push_free(&htab->lru, &l_old->lru_node);
htab              979 kernel/bpf/hashtab.c 	struct bpf_htab *htab = container_of(map, struct bpf_htab, map);
htab              995 kernel/bpf/hashtab.c 	hash = htab_map_hash(key, key_size, htab->hashrnd);
htab              997 kernel/bpf/hashtab.c 	b = __select_bucket(htab, hash);
htab             1005 kernel/bpf/hashtab.c 	ret = check_flags(htab, l_old, map_flags);
htab             1011 kernel/bpf/hashtab.c 		pcpu_copy_value(htab, htab_elem_get_ptr(l_old, key_size),
htab             1014 kernel/bpf/hashtab.c 		l_new = alloc_htab_elem(htab, key, value, key_size,
htab             1032 kernel/bpf/hashtab.c 	struct bpf_htab *htab = container_of(map, struct bpf_htab, map);
htab             1048 kernel/bpf/hashtab.c 	hash = htab_map_hash(key, key_size, htab->hashrnd);
htab             1050 kernel/bpf/hashtab.c 	b = __select_bucket(htab, hash);
htab             1059 kernel/bpf/hashtab.c 		l_new = prealloc_lru_pop(htab, key, hash);
htab             1069 kernel/bpf/hashtab.c 	ret = check_flags(htab, l_old, map_flags);
htab             1077 kernel/bpf/hashtab.c 		pcpu_copy_value(htab, htab_elem_get_ptr(l_old, key_size),
htab             1080 kernel/bpf/hashtab.c 		pcpu_copy_value(htab, htab_elem_get_ptr(l_new, key_size),
htab             1089 kernel/bpf/hashtab.c 		bpf_lru_push_free(&htab->lru, &l_new->lru_node);
htab             1109 kernel/bpf/hashtab.c 	struct bpf_htab *htab = container_of(map, struct bpf_htab, map);
htab             1121 kernel/bpf/hashtab.c 	hash = htab_map_hash(key, key_size, htab->hashrnd);
htab             1122 kernel/bpf/hashtab.c 	b = __select_bucket(htab, hash);
htab             1131 kernel/bpf/hashtab.c 		free_htab_elem(htab, l);
htab             1141 kernel/bpf/hashtab.c 	struct bpf_htab *htab = container_of(map, struct bpf_htab, map);
htab             1153 kernel/bpf/hashtab.c 	hash = htab_map_hash(key, key_size, htab->hashrnd);
htab             1154 kernel/bpf/hashtab.c 	b = __select_bucket(htab, hash);
htab             1168 kernel/bpf/hashtab.c 		bpf_lru_push_free(&htab->lru, &l->lru_node);
htab             1172 kernel/bpf/hashtab.c static void delete_all_elements(struct bpf_htab *htab)
htab             1176 kernel/bpf/hashtab.c 	for (i = 0; i < htab->n_buckets; i++) {
htab             1177 kernel/bpf/hashtab.c 		struct hlist_nulls_head *head = select_bucket(htab, i);
htab             1183 kernel/bpf/hashtab.c 			htab_elem_free(htab, l);
htab             1191 kernel/bpf/hashtab.c 	struct bpf_htab *htab = container_of(map, struct bpf_htab, map);
htab             1204 kernel/bpf/hashtab.c 	if (!htab_is_prealloc(htab))
htab             1205 kernel/bpf/hashtab.c 		delete_all_elements(htab);
htab             1207 kernel/bpf/hashtab.c 		prealloc_destroy(htab);
htab             1209 kernel/bpf/hashtab.c 	free_percpu(htab->extra_elems);
htab             1210 kernel/bpf/hashtab.c 	bpf_map_area_free(htab->buckets);
htab             1211 kernel/bpf/hashtab.c 	kfree(htab);
htab             1318 kernel/bpf/hashtab.c 	struct bpf_htab *htab = container_of(map, struct bpf_htab, map);
htab             1322 kernel/bpf/hashtab.c 	if (htab_is_lru(htab))
htab             1393 kernel/bpf/hashtab.c 	struct bpf_htab *htab = container_of(map, struct bpf_htab, map);
htab             1399 kernel/bpf/hashtab.c 	for (i = 0; i < htab->n_buckets; i++) {
htab             1400 kernel/bpf/hashtab.c 		head = select_bucket(htab, i);
htab              545 net/core/sock_map.c static struct bpf_htab_bucket *sock_hash_select_bucket(struct bpf_htab *htab,
htab              548 net/core/sock_map.c 	return &htab->buckets[hash & (htab->buckets_num - 1)];
htab              568 net/core/sock_map.c 	struct bpf_htab *htab = container_of(map, struct bpf_htab, map);
htab              576 net/core/sock_map.c 	bucket = sock_hash_select_bucket(htab, hash);
htab              582 net/core/sock_map.c static void sock_hash_free_elem(struct bpf_htab *htab,
htab              585 net/core/sock_map.c 	atomic_dec(&htab->count);
htab              592 net/core/sock_map.c 	struct bpf_htab *htab = container_of(map, struct bpf_htab, map);
htab              597 net/core/sock_map.c 	bucket = sock_hash_select_bucket(htab, elem->hash);
htab              609 net/core/sock_map.c 		sock_hash_free_elem(htab, elem);
htab              616 net/core/sock_map.c 	struct bpf_htab *htab = container_of(map, struct bpf_htab, map);
htab              623 net/core/sock_map.c 	bucket = sock_hash_select_bucket(htab, hash);
htab              630 net/core/sock_map.c 		sock_hash_free_elem(htab, elem);
htab              637 net/core/sock_map.c static struct bpf_htab_elem *sock_hash_alloc_elem(struct bpf_htab *htab,
htab              644 net/core/sock_map.c 	if (atomic_inc_return(&htab->count) > htab->map.max_entries) {
htab              646 net/core/sock_map.c 			atomic_dec(&htab->count);
htab              651 net/core/sock_map.c 	new = kmalloc_node(htab->elem_size, GFP_ATOMIC | __GFP_NOWARN,
htab              652 net/core/sock_map.c 			   htab->map.numa_node);
htab              654 net/core/sock_map.c 		atomic_dec(&htab->count);
htab              666 net/core/sock_map.c 	struct bpf_htab *htab = container_of(map, struct bpf_htab, map);
htab              685 net/core/sock_map.c 	ret = sock_map_link(map, &htab->progs, sk);
htab              693 net/core/sock_map.c 	bucket = sock_hash_select_bucket(htab, hash);
htab              705 net/core/sock_map.c 	elem_new = sock_hash_alloc_elem(htab, key, key_size, hash, sk, elem);
htab              719 net/core/sock_map.c 		sock_hash_free_elem(htab, elem);
htab              766 net/core/sock_map.c 	struct bpf_htab *htab = container_of(map, struct bpf_htab, map);
htab              775 net/core/sock_map.c 	head = &sock_hash_select_bucket(htab, hash)->head;
htab              787 net/core/sock_map.c 	i = hash & (htab->buckets_num - 1);
htab              790 net/core/sock_map.c 	for (; i < htab->buckets_num; i++) {
htab              791 net/core/sock_map.c 		head = &sock_hash_select_bucket(htab, i)->head;
htab              805 net/core/sock_map.c 	struct bpf_htab *htab;
htab              819 net/core/sock_map.c 	htab = kzalloc(sizeof(*htab), GFP_USER);
htab              820 net/core/sock_map.c 	if (!htab)
htab              823 net/core/sock_map.c 	bpf_map_init_from_attr(&htab->map, attr);
htab              825 net/core/sock_map.c 	htab->buckets_num = roundup_pow_of_two(htab->map.max_entries);
htab              826 net/core/sock_map.c 	htab->elem_size = sizeof(struct bpf_htab_elem) +
htab              827 net/core/sock_map.c 			  round_up(htab->map.key_size, 8);
htab              828 net/core/sock_map.c 	if (htab->buckets_num == 0 ||
htab              829 net/core/sock_map.c 	    htab->buckets_num > U32_MAX / sizeof(struct bpf_htab_bucket)) {
htab              834 net/core/sock_map.c 	cost = (u64) htab->buckets_num * sizeof(struct bpf_htab_bucket) +
htab              835 net/core/sock_map.c 	       (u64) htab->elem_size * htab->map.max_entries;
htab              841 net/core/sock_map.c 	htab->buckets = bpf_map_area_alloc(htab->buckets_num *
htab              843 net/core/sock_map.c 					   htab->map.numa_node);
htab              844 net/core/sock_map.c 	if (!htab->buckets) {
htab              849 net/core/sock_map.c 	for (i = 0; i < htab->buckets_num; i++) {
htab              850 net/core/sock_map.c 		INIT_HLIST_HEAD(&htab->buckets[i].head);
htab              851 net/core/sock_map.c 		raw_spin_lock_init(&htab->buckets[i].lock);
htab              854 net/core/sock_map.c 	return &htab->map;
htab              856 net/core/sock_map.c 	kfree(htab);
htab              862 net/core/sock_map.c 	struct bpf_htab *htab = container_of(map, struct bpf_htab, map);
htab              873 net/core/sock_map.c 	for (i = 0; i < htab->buckets_num; i++) {
htab              874 net/core/sock_map.c 		bucket = sock_hash_select_bucket(htab, i);
htab              891 net/core/sock_map.c 	bpf_map_area_free(htab->buckets);
htab              892 net/core/sock_map.c 	kfree(htab);
htab             4027 net/xfrm/xfrm_policy.c 		struct xfrm_policy_hash *htab;
htab             4033 net/xfrm/xfrm_policy.c 		htab = &net->xfrm.policy_bydst[dir];
htab             4034 net/xfrm/xfrm_policy.c 		htab->table = xfrm_hash_alloc(sz);
htab             4035 net/xfrm/xfrm_policy.c 		if (!htab->table)
htab             4037 net/xfrm/xfrm_policy.c 		htab->hmask = hmask;
htab             4038 net/xfrm/xfrm_policy.c 		htab->dbits4 = 32;
htab             4039 net/xfrm/xfrm_policy.c 		htab->sbits4 = 32;
htab             4040 net/xfrm/xfrm_policy.c 		htab->dbits6 = 128;
htab             4041 net/xfrm/xfrm_policy.c 		htab->sbits6 = 128;
htab             4058 net/xfrm/xfrm_policy.c 		struct xfrm_policy_hash *htab;
htab             4060 net/xfrm/xfrm_policy.c 		htab = &net->xfrm.policy_bydst[dir];
htab             4061 net/xfrm/xfrm_policy.c 		xfrm_hash_free(htab->table, sz);
htab             4083 net/xfrm/xfrm_policy.c 		struct xfrm_policy_hash *htab;
htab             4087 net/xfrm/xfrm_policy.c 		htab = &net->xfrm.policy_bydst[dir];
htab             4088 net/xfrm/xfrm_policy.c 		sz = (htab->hmask + 1) * sizeof(struct hlist_head);
htab             4089 net/xfrm/xfrm_policy.c 		WARN_ON(!hlist_empty(htab->table));
htab             4090 net/xfrm/xfrm_policy.c 		xfrm_hash_free(htab->table, sz);
htab                5 tools/testing/selftests/bpf/progs/sample_map_ret0.c struct bpf_map_def SEC("maps") htab = {
htab               26 tools/testing/selftests/bpf/progs/sample_map_ret0.c 	value = bpf_map_lookup_elem(&htab, &key);