/linux-4.1.27/drivers/s390/scsi/ |
D | zfcp_reqlist.h | 23 struct list_head buckets[ZFCP_REQ_LIST_BUCKETS]; member 49 INIT_LIST_HEAD(&rl->buckets[i]); in zfcp_reqlist_alloc() 65 if (!list_empty(&rl->buckets[i])) in zfcp_reqlist_isempty() 89 list_for_each_entry(req, &rl->buckets[i], list) in _zfcp_reqlist_find() 162 list_add_tail(&req->list, &rl->buckets[i]); in zfcp_reqlist_add() 179 list_splice_init(&rl->buckets[i], list); in zfcp_reqlist_move()
|
/linux-4.1.27/net/openvswitch/ |
D | flow_table.c | 121 struct flex_array *buckets; in alloc_buckets() local 124 buckets = flex_array_alloc(sizeof(struct hlist_head), in alloc_buckets() 126 if (!buckets) in alloc_buckets() 129 err = flex_array_prealloc(buckets, 0, n_buckets, GFP_KERNEL); in alloc_buckets() 131 flex_array_free(buckets); in alloc_buckets() 137 flex_array_get(buckets, i)); in alloc_buckets() 139 return buckets; in alloc_buckets() 174 static void free_buckets(struct flex_array *buckets) in free_buckets() argument 176 flex_array_free(buckets); in free_buckets() 182 free_buckets(ti->buckets); in __table_instance_destroy() [all …]
|
D | flow_table.h | 40 struct flex_array *buckets; member
|
/linux-4.1.27/fs/nfs/filelayout/ |
D | filelayout.c | 754 kfree(flo->commit_info.buckets); in filelayout_free_lseg() 755 flo->commit_info.buckets = NULL; in filelayout_free_lseg() 766 struct pnfs_commit_bucket *buckets; in filelayout_alloc_commit_info() local 785 buckets = kcalloc(size, sizeof(struct pnfs_commit_bucket), in filelayout_alloc_commit_info() 787 if (!buckets) in filelayout_alloc_commit_info() 790 INIT_LIST_HEAD(&buckets[i].written); in filelayout_alloc_commit_info() 791 INIT_LIST_HEAD(&buckets[i].committing); in filelayout_alloc_commit_info() 793 buckets[i].direct_verf.committed = NFS_INVALID_STABLE_HOW; in filelayout_alloc_commit_info() 800 list_splice(&cinfo->ds->buckets[i].written, in filelayout_alloc_commit_info() 801 &buckets[i].written); in filelayout_alloc_commit_info() [all …]
|
/linux-4.1.27/kernel/bpf/ |
D | hashtab.c | 19 struct hlist_head *buckets; member 73 htab->buckets = kmalloc_array(htab->n_buckets, sizeof(struct hlist_head), in htab_map_alloc() 76 if (!htab->buckets) { in htab_map_alloc() 77 htab->buckets = vmalloc(htab->n_buckets * sizeof(struct hlist_head)); in htab_map_alloc() 78 if (!htab->buckets) in htab_map_alloc() 83 INIT_HLIST_HEAD(&htab->buckets[i]); in htab_map_alloc() 105 return &htab->buckets[hash & (htab->n_buckets - 1)]; in select_bucket() 344 kvfree(htab->buckets); in htab_map_free()
|
/linux-4.1.27/net/ceph/crush/ |
D | crush.c | 119 if (map->buckets) { in crush_destroy() 122 if (map->buckets[b] == NULL) in crush_destroy() 124 crush_destroy_bucket(map->buckets[b]); in crush_destroy() 126 kfree(map->buckets); in crush_destroy()
|
D | mapper.c | 471 itemtype = map->buckets[-1-item]->type; in crush_choose_firstn() 484 in = map->buckets[-1-item]; in crush_choose_firstn() 506 map->buckets[-1-item], in crush_choose_firstn() 658 itemtype = map->buckets[-1-item]->type; in crush_choose_indep() 675 in = map->buckets[-1-item]; in crush_choose_indep() 693 map->buckets[-1-item], in crush_choose_indep() 864 map->buckets[-1-w[i]], in crush_do_rule() 883 map->buckets[-1-w[i]], in crush_do_rule()
|
/linux-4.1.27/fs/nfs/ |
D | pnfs_nfs.c | 146 cnt = pnfs_generic_scan_ds_commit_list(&cinfo->ds->buckets[i], in pnfs_generic_scan_commit_lists() 165 for (i = 0, b = cinfo->ds->buckets; i < cinfo->ds->nbuckets; i++, b++) { in pnfs_generic_recover_commit_reqs() 188 bucket = &fl_cinfo->buckets[i]; in pnfs_generic_retry_commit() 211 bucket = fl_cinfo->buckets; in pnfs_generic_alloc_ds_commits() 274 struct pnfs_commit_bucket *buckets; in pnfs_generic_commit_pagelist() local 276 buckets = cinfo->ds->buckets; in pnfs_generic_commit_pagelist() 278 &buckets[data->ds_commit_index].committing, in pnfs_generic_commit_pagelist() 854 struct pnfs_commit_bucket *buckets; in pnfs_layout_mark_request_commit() local 857 buckets = cinfo->ds->buckets; in pnfs_layout_mark_request_commit() 858 list = &buckets[ds_commit_idx].written; in pnfs_layout_mark_request_commit() [all …]
|
D | direct.c | 172 verfp = &dreq->ds_cinfo.buckets[commit_idx].direct_verf; in nfs_direct_select_verf()
|
/linux-4.1.27/net/netfilter/ipvs/ |
D | ip_vs_sh.c | 75 struct ip_vs_sh_bucket buckets[IP_VS_SH_TAB_SIZE]; member 112 struct ip_vs_dest *dest = rcu_dereference(s->buckets[hash].dest); in ip_vs_sh_get() 134 dest = rcu_dereference(s->buckets[ihash].dest); in ip_vs_sh_get_fallback() 149 dest = rcu_dereference(s->buckets[hash].dest); in ip_vs_sh_get_fallback() 176 b = &s->buckets[0]; in ip_vs_sh_reassign() 220 b = &s->buckets[0]; in ip_vs_sh_flush()
|
D | ip_vs_dh.c | 68 struct ip_vs_dh_bucket buckets[IP_VS_DH_TAB_SIZE]; member 94 return rcu_dereference(s->buckets[ip_vs_dh_hashkey(af, addr)].dest); in ip_vs_dh_get() 110 b = &s->buckets[0]; in ip_vs_dh_reassign() 144 b = &s->buckets[0]; in ip_vs_dh_flush()
|
/linux-4.1.27/drivers/md/ |
D | dm-region-hash.c | 71 struct list_head *buckets; member 204 rh->buckets = vmalloc(nr_buckets * sizeof(*rh->buckets)); in dm_region_hash_create() 205 if (!rh->buckets) { in dm_region_hash_create() 212 INIT_LIST_HEAD(rh->buckets + i); in dm_region_hash_create() 226 vfree(rh->buckets); in dm_region_hash_create() 242 list_for_each_entry_safe(reg, nreg, rh->buckets + h, in dm_region_hash_destroy() 255 vfree(rh->buckets); in dm_region_hash_destroy() 274 struct list_head *bucket = rh->buckets + rh_hash(rh, region); in __rh_lookup() 285 list_add(®->hash_list, rh->buckets + rh_hash(rh, reg->key)); in __rh_insert()
|
D | dm-ioctl.c | 72 static void init_buckets(struct list_head *buckets) in init_buckets() argument 77 INIT_LIST_HEAD(buckets + i); in init_buckets()
|
/linux-4.1.27/include/linux/ |
D | rhashtable.h | 78 struct rhash_head __rcu *buckets[] ____cacheline_aligned_in_smp; member 392 rht_for_each_continue(pos, (tbl)->buckets[hash], tbl, hash) 417 rht_for_each_entry_continue(tpos, pos, (tbl)->buckets[hash], \ 433 for (pos = rht_dereference_bucket((tbl)->buckets[hash], tbl, hash), \ 469 rht_for_each_rcu_continue(pos, (tbl)->buckets[hash], tbl, hash) 503 rht_for_each_entry_rcu_continue(tpos, pos, (tbl)->buckets[hash],\ 636 head = rht_dereference_bucket(tbl->buckets[hash], tbl, hash); in __rhashtable_insert_fast() 640 rcu_assign_pointer(tbl->buckets[hash], obj); in __rhashtable_insert_fast() 756 pprev = &tbl->buckets[hash]; in __rhashtable_remove_fast()
|
D | nfs_xdr.h | 1125 struct pnfs_commit_bucket *buckets; member 1253 kfree(cinfo->buckets); in nfs_free_pnfs_ds_cinfo()
|
D | netdevice.h | 2433 u8 buckets[]; member
|
/linux-4.1.27/lib/ |
D | rhashtable.c | 119 size = sizeof(*tbl) + nbuckets * sizeof(tbl->buckets[0]); in bucket_table_alloc() 140 INIT_RHT_NULLS_HEAD(tbl->buckets[i], ht, i); in bucket_table_alloc() 163 struct rhash_head __rcu **pprev = &old_tbl->buckets[old_hash]; in rhashtable_rehash_one() 187 head = rht_dereference_bucket(new_tbl->buckets[new_hash], in rhashtable_rehash_one() 195 rcu_assign_pointer(new_tbl->buckets[new_hash], entry); in rhashtable_rehash_one() 462 head = rht_dereference_bucket(tbl->buckets[hash], tbl, hash); in rhashtable_insert_slow() 466 rcu_assign_pointer(tbl->buckets[hash], obj); in rhashtable_insert_slow() 828 for (pos = rht_dereference(tbl->buckets[i], ht), in rhashtable_free_and_destroy()
|
D | test_rhashtable.c | 110 i, tbl->buckets[i], cnt); in test_bucket_stats()
|
/linux-4.1.27/drivers/md/persistent-data/ |
D | dm-transaction-manager.c | 96 struct hlist_head buckets[DM_HASH_SIZE]; member 110 hlist_for_each_entry(si, tm->buckets + bucket, hlist) in is_shadow() 134 hlist_add_head(&si->hlist, tm->buckets + bucket); in insert_shadow() 148 bucket = tm->buckets + i; in wipe_shadow_table() 177 INIT_HLIST_HEAD(tm->buckets + i); in dm_tm_create()
|
/linux-4.1.27/net/sched/ |
D | sch_hhf.c | 128 struct wdrr_bucket buckets[WDRR_BUCKET_CNT]; member 371 bucket = &q->buckets[WDRR_BUCKET_FOR_HH]; in hhf_drop() 373 bucket = &q->buckets[WDRR_BUCKET_FOR_NON_HH]; in hhf_drop() 385 return bucket - q->buckets; in hhf_drop() 396 bucket = &q->buckets[idx]; in hhf_enqueue() 449 int weight = (bucket - q->buckets == WDRR_BUCKET_FOR_HH) ? in hhf_dequeue() 661 struct wdrr_bucket *bucket = q->buckets + i; in hhf_init()
|
/linux-4.1.27/drivers/infiniband/ulp/ipoib/ |
D | ipoib_main.c | 896 for (neigh = rcu_dereference_bh(htbl->buckets[hash_val]); in ipoib_neigh_get() 945 struct ipoib_neigh __rcu **np = &htbl->buckets[i]; in __ipoib_reap_neigh() 1021 for (neigh = rcu_dereference_protected(htbl->buckets[hash_val], in ipoib_neigh_alloc() 1047 rcu_dereference_protected(htbl->buckets[hash_val], in ipoib_neigh_alloc() 1049 rcu_assign_pointer(htbl->buckets[hash_val], neigh); in ipoib_neigh_alloc() 1106 np = &htbl->buckets[hash_val]; in ipoib_neigh_free() 1131 struct ipoib_neigh **buckets; in ipoib_neigh_hash_init() local 1141 buckets = kzalloc(size * sizeof(*buckets), GFP_KERNEL); in ipoib_neigh_hash_init() 1142 if (!buckets) { in ipoib_neigh_hash_init() 1148 htbl->buckets = buckets; in ipoib_neigh_hash_init() [all …]
|
D | ipoib.h | 288 struct ipoib_neigh __rcu **buckets; member
|
/linux-4.1.27/drivers/md/bcache/ |
D | alloc.c | 145 trace_bcache_invalidate(ca, b - ca->buckets); in __bch_invalidate_one_bucket() 156 fifo_push(&ca->free_inc, b - ca->buckets); in bch_invalidate_one_bucket() 225 b = ca->buckets + ca->fifo_last_bucket++; in invalidate_buckets_fifo() 250 b = ca->buckets + n; in invalidate_buckets_random() 428 b = ca->buckets + r; in bch_bucket_alloc() 481 k->ptr[i] = PTR(ca->buckets[b].gen, in __bch_bucket_alloc_set()
|
D | bcache.h | 434 struct bucket *buckets; member 750 return PTR_CACHE(c, k, ptr)->buckets + PTR_BUCKET_NR(c, k, ptr); in PTR_BUCKET() 820 for (b = (ca)->buckets + (ca)->sb.first_bucket; \ 821 b < (ca)->buckets + (ca)->sb.nbuckets; b++)
|
D | super.c | 568 for (b = ca->buckets + i * prios_per_bucket(ca); in bch_prio_write() 569 b < ca->buckets + ca->sb.nbuckets && d < end; in bch_prio_write() 587 atomic_dec_bug(&ca->buckets[bucket].pin); in bch_prio_write() 604 &ca->buckets[ca->prio_last_buckets[i]]); in bch_prio_write() 617 for (b = ca->buckets; in prio_read() 618 b < ca->buckets + ca->sb.nbuckets; in prio_read() 1821 vfree(ca->buckets); in bch_cache_release() 1859 !(ca->buckets = vzalloc(sizeof(struct bucket) * in cache_alloc()
|
D | btree.c | 1701 SET_GC_MARK(ca->buckets + *i, GC_MARK_METADATA); in bch_btree_gc_finish() 1705 SET_GC_MARK(ca->buckets + *i, GC_MARK_METADATA); in bch_btree_gc_finish() 1876 b - ca->buckets); in bch_initial_gc_finish()
|
D | sysfs.c | 786 p[i] = ca->buckets[i].prio; in SHOW()
|
/linux-4.1.27/Documentation/networking/ |
D | nf_conntrack-sysctl.txt | 13 by 16384 to determine the number of buckets but the hash table will 14 never have fewer than 32 and limited to 16384 buckets. For systems 15 with more than 4GB of memory it will be 65536 buckets.
|
D | scaling.txt | 202 the same that selects a CPU in RPS, but as the number of buckets can 205 table has 4096 buckets. This value can be modified through sysctl
|
/linux-4.1.27/fs/nfs/flexfilelayout/ |
D | flexfilelayout.c | 403 kfree(ffl->commit_info.buckets); in ff_layout_free_lseg() 404 ffl->commit_info.buckets = NULL; in ff_layout_free_lseg() 424 struct pnfs_commit_bucket *buckets; in ff_layout_alloc_commit_info() local 438 buckets = kcalloc(size, sizeof(struct pnfs_commit_bucket), in ff_layout_alloc_commit_info() 440 if (!buckets) in ff_layout_alloc_commit_info() 447 kfree(buckets); in ff_layout_alloc_commit_info() 449 cinfo->ds->buckets = buckets; in ff_layout_alloc_commit_info() 452 INIT_LIST_HEAD(&buckets[i].written); in ff_layout_alloc_commit_info() 453 INIT_LIST_HEAD(&buckets[i].committing); in ff_layout_alloc_commit_info() 455 buckets[i].direct_verf.committed = in ff_layout_alloc_commit_info()
|
/linux-4.1.27/drivers/message/fusion/ |
D | mptlan.c | 1158 u32 curr, buckets, count, max; in mpt_lan_post_receive_buckets() local 1164 buckets = (priv->max_buckets_out - curr); in mpt_lan_post_receive_buckets() 1168 __func__, buckets, curr)); in mpt_lan_post_receive_buckets() 1173 while (buckets) { in mpt_lan_post_receive_buckets() 1179 __func__, buckets)); in mpt_lan_post_receive_buckets() 1186 count = buckets; in mpt_lan_post_receive_buckets() 1283 buckets -= i; in mpt_lan_post_receive_buckets() 1289 __func__, buckets, atomic_read(&priv->buckets_out))); in mpt_lan_post_receive_buckets()
|
/linux-4.1.27/include/linux/crush/ |
D | crush.h | 165 struct crush_bucket **buckets; member
|
/linux-4.1.27/Documentation/ |
D | bcache.txt | 10 in erase block sized buckets, and it uses a hybrid btree/log to track cached 325 Size of buckets 411 Size of buckets 423 increase the number of buckets kept on the freelist, which lets you 426 since buckets are discarded when they move on to the freelist will also make 437 Total buckets in this cache 443 metadata overhead. Average is the average priority of cache buckets.
|
D | kernel-parameters.txt | 865 Set number of hash buckets for dentry cache. 1383 Set number of hash buckets for inode cache. 3288 Set number of hash buckets for route cache 3566 Set number of hash buckets for TCP connection 3729 Set number of hash buckets for UDP/UDP-Lite connections
|
/linux-4.1.27/drivers/gpu/drm/radeon/ |
D | radeon_cs.c | 79 struct radeon_cs_buckets buckets; in radeon_cs_parser_relocs() local 96 radeon_cs_buckets_init(&buckets); in radeon_cs_parser_relocs() 170 radeon_cs_buckets_add(&buckets, &p->relocs[i].tv.head, in radeon_cs_parser_relocs() 174 radeon_cs_buckets_get_list(&buckets, &p->validated); in radeon_cs_parser_relocs()
|
/linux-4.1.27/drivers/media/v4l2-core/ |
D | v4l2-ctrls.c | 1725 hdl->buckets = kcalloc(hdl->nr_of_buckets, sizeof(hdl->buckets[0]), in v4l2_ctrl_handler_init_class() 1727 hdl->error = hdl->buckets ? 0 : -ENOMEM; in v4l2_ctrl_handler_init_class() 1739 if (hdl == NULL || hdl->buckets == NULL) in v4l2_ctrl_handler_free() 1755 kfree(hdl->buckets); in v4l2_ctrl_handler_free() 1756 hdl->buckets = NULL; in v4l2_ctrl_handler_free() 1808 ref = hdl->buckets ? hdl->buckets[bucket] : NULL; in find_ref() 1903 new_ref->next = hdl->buckets[bucket]; in handler_new_ref() 1904 hdl->buckets[bucket] = new_ref; in handler_new_ref()
|
/linux-4.1.27/Documentation/ABI/testing/ |
D | sysfs-block-bcache | 128 For a cache, the number of usable buckets. 141 Number of btree buckets/nodes that are currently cached in
|
/linux-4.1.27/drivers/staging/lustre/lustre/libcfs/ |
D | hash.c | 885 cfs_hash_buckets_free(struct cfs_hash_bucket **buckets, in cfs_hash_buckets_free() argument 891 if (buckets[i] != NULL) in cfs_hash_buckets_free() 892 LIBCFS_FREE(buckets[i], bkt_size); in cfs_hash_buckets_free() 895 LIBCFS_FREE(buckets, sizeof(buckets[0]) * size); in cfs_hash_buckets_free()
|
/linux-4.1.27/net/ceph/ |
D | osdmap.c | 189 c->buckets = kcalloc(c->max_buckets, sizeof(*c->buckets), GFP_NOFS); in crush_decode() 190 if (c->buckets == NULL) in crush_decode() 204 c->buckets[i] = NULL; in crush_decode() 231 b = c->buckets[i] = kzalloc(size, GFP_NOFS); in crush_decode()
|
/linux-4.1.27/kernel/trace/ |
D | ftrace.c | 1130 struct hlist_head *buckets; member 1143 .buckets = (struct hlist_head *)empty_buckets, 1230 hhd = &hash->buckets[key]; in ftrace_lookup_ip() 1250 hhd = &hash->buckets[key]; in __add_hash_entry() 1298 hhd = &hash->buckets[i]; in ftrace_hash_clear() 1310 kfree(hash->buckets); in free_ftrace_hash() 1346 hash->buckets = kcalloc(size, sizeof(*hash->buckets), GFP_KERNEL); in alloc_ftrace_hash() 1348 if (!hash->buckets) { in alloc_ftrace_hash() 1377 hlist_for_each_entry(entry, &hash->buckets[i], hlist) { in alloc_and_copy_ftrace_hash() 1443 hhd = &src->buckets[i]; in ftrace_hash_move()
|
/linux-4.1.27/drivers/gpu/drm/i915/ |
D | i915_gem_execbuffer.c | 49 struct hlist_head buckets[0]; member 89 memset(eb->buckets, 0, (eb->and+1)*sizeof(struct hlist_head)); in eb_reset() 164 &eb->buckets[handle & eb->and]); in eb_lookup_vmas() 198 head = &eb->buckets[handle & eb->and]; in eb_get_vma()
|
/linux-4.1.27/Documentation/filesystems/ |
D | omfs.txt | 81 until a match is found on i_name. Empty buckets are represented by block
|
D | logfs.txt | 188 is limited to 2^31. Secondly the table contains hash buckets with 16 191 Using single-entry buckets would result in birthday "attacks". At
|
D | f2fs.txt | 470 a hash table with dedicated number of hash buckets as shown below. Note that 489 The number of blocks and buckets are determined by, 497 # of buckets in level #n = | 509 bucket number to scan in level #n = (hash value) % (# of buckets in level #n)
|
/linux-4.1.27/net/core/ |
D | neighbour.c | 317 struct neighbour __rcu **buckets; in neigh_hash_alloc() local 324 buckets = kzalloc(size, GFP_ATOMIC); in neigh_hash_alloc() 326 buckets = (struct neighbour __rcu **) in neigh_hash_alloc() 329 if (!buckets) { in neigh_hash_alloc() 333 ret->hash_buckets = buckets; in neigh_hash_alloc() 346 struct neighbour __rcu **buckets = nht->hash_buckets; in neigh_hash_free_rcu() local 349 kfree(buckets); in neigh_hash_free_rcu() 351 free_pages((unsigned long)buckets, get_order(size)); in neigh_hash_free_rcu()
|
D | dev.c | 3315 if (likely(fl->buckets[old_flow])) in skb_flow_limit() 3316 fl->buckets[old_flow]--; in skb_flow_limit() 3318 if (++fl->buckets[new_flow] > (FLOW_LIMIT_HISTORY >> 1)) { in skb_flow_limit()
|
/linux-4.1.27/include/media/ |
D | v4l2-ctrls.h | 255 struct v4l2_ctrl_ref **buckets; member
|
/linux-4.1.27/Documentation/filesystems/nfs/ |
D | pnfs.txt | 36 data structure. The hlist size of 32 buckets needs better
|
/linux-4.1.27/Documentation/cris/ |
D | README | 127 IP: routing cache hash table of 1024 buckets, 8Kbytes
|
/linux-4.1.27/include/trace/events/ |
D | bcache.h | 418 __entry->sectors = GC_SECTORS_USED(&ca->buckets[bucket]);
|
/linux-4.1.27/fs/jffs2/ |
D | README.Locking | 105 has to be locked while walking the c->inocache_list hash buckets.
|
/linux-4.1.27/Documentation/locking/ |
D | lockdep-design.txt | 248 a hash table with 8192 buckets where each bucket has its own
|
/linux-4.1.27/net/netfilter/ |
D | Kconfig | 1123 of limit buckets, based on your selection of source/destination
|