bkt 50 arch/mips/netlogic/xlr/fmn-config.c int bkt; bkt 54 arch/mips/netlogic/xlr/fmn-config.c for (bkt = 0; bkt < 16; bkt++) bkt 56 arch/mips/netlogic/xlr/fmn-config.c xlr_board_fmn_config.bucket_size[(bkt * 8) + 0], bkt 57 arch/mips/netlogic/xlr/fmn-config.c xlr_board_fmn_config.bucket_size[(bkt * 8) + 1], bkt 58 arch/mips/netlogic/xlr/fmn-config.c xlr_board_fmn_config.bucket_size[(bkt * 8) + 2], bkt 59 arch/mips/netlogic/xlr/fmn-config.c xlr_board_fmn_config.bucket_size[(bkt * 8) + 3], bkt 60 arch/mips/netlogic/xlr/fmn-config.c xlr_board_fmn_config.bucket_size[(bkt * 8) + 4], bkt 61 arch/mips/netlogic/xlr/fmn-config.c xlr_board_fmn_config.bucket_size[(bkt * 8) + 5], bkt 62 arch/mips/netlogic/xlr/fmn-config.c xlr_board_fmn_config.bucket_size[(bkt * 8) + 6], bkt 63 arch/mips/netlogic/xlr/fmn-config.c xlr_board_fmn_config.bucket_size[(bkt * 8) + 7]); bkt 68 arch/mips/netlogic/xlr/fmn-config.c for (bkt = 0; bkt < 16; bkt++) bkt 70 arch/mips/netlogic/xlr/fmn-config.c fmn_info->credit_config[(bkt * 8) + 0], bkt 71 arch/mips/netlogic/xlr/fmn-config.c fmn_info->credit_config[(bkt * 8) + 1], bkt 72 arch/mips/netlogic/xlr/fmn-config.c fmn_info->credit_config[(bkt * 8) + 2], bkt 73 arch/mips/netlogic/xlr/fmn-config.c fmn_info->credit_config[(bkt * 8) + 3], bkt 74 arch/mips/netlogic/xlr/fmn-config.c fmn_info->credit_config[(bkt * 8) + 4], bkt 75 arch/mips/netlogic/xlr/fmn-config.c fmn_info->credit_config[(bkt * 8) + 5], bkt 76 arch/mips/netlogic/xlr/fmn-config.c fmn_info->credit_config[(bkt * 8) + 6], bkt 77 arch/mips/netlogic/xlr/fmn-config.c fmn_info->credit_config[(bkt * 8) + 7]); bkt 84 arch/mips/netlogic/xlr/fmn-config.c int bkt, n, total_credits, ncores; bkt 87 arch/mips/netlogic/xlr/fmn-config.c for (bkt = 0; bkt < 128; bkt++) { bkt 90 arch/mips/netlogic/xlr/fmn-config.c total_credits += cfg->cpu[n].credit_config[bkt]; bkt 91 arch/mips/netlogic/xlr/fmn-config.c total_credits += cfg->gmac[0].credit_config[bkt]; bkt 92 arch/mips/netlogic/xlr/fmn-config.c total_credits += cfg->gmac[1].credit_config[bkt]; bkt 93 arch/mips/netlogic/xlr/fmn-config.c total_credits += cfg->dma.credit_config[bkt]; bkt 94 arch/mips/netlogic/xlr/fmn-config.c total_credits += cfg->cmp.credit_config[bkt]; bkt 95 arch/mips/netlogic/xlr/fmn-config.c total_credits += cfg->sae.credit_config[bkt]; bkt 96 arch/mips/netlogic/xlr/fmn-config.c total_credits += cfg->xgmac[0].credit_config[bkt]; bkt 97 arch/mips/netlogic/xlr/fmn-config.c total_credits += cfg->xgmac[1].credit_config[bkt]; bkt 98 arch/mips/netlogic/xlr/fmn-config.c if (total_credits > cfg->bucket_size[bkt]) bkt 100 arch/mips/netlogic/xlr/fmn-config.c bkt, total_credits, cfg->bucket_size[bkt]); bkt 107 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c int bkt; bkt 112 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c vnic_hash_for_each_safe(mactbl, bkt, tmp, node, hlist) { bkt 157 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c int bkt; bkt 168 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c vnic_hash_for_each(mactbl, bkt, node, hlist) { bkt 209 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c int i, bkt, rc = 0; bkt 263 drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c vnic_hash_for_each(old_mactbl, bkt, node, hlist) { bkt 280 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h #define vnic_hash_for_each_safe(name, bkt, tmp, obj, member) \ bkt 281 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h for ((bkt) = 0, obj = NULL; \ bkt 282 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h !obj && (bkt) < OPA_VNIC_MAC_TBL_SIZE; (bkt)++) \ bkt 283 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h hlist_for_each_entry_safe(obj, tmp, &name[bkt], member) bkt 289 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h #define vnic_hash_for_each(name, bkt, obj, member) \ bkt 290 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h for ((bkt) = 0, obj = NULL; \ bkt 291 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h !obj && (bkt) < OPA_VNIC_MAC_TBL_SIZE; (bkt)++) \ bkt 292 drivers/infiniband/ulp/opa_vnic/opa_vnic_internal.h hlist_for_each_entry(obj, &name[bkt], member) bkt 114 drivers/net/ethernet/intel/i40e/i40e_debugfs.c int i, bkt; bkt 147 drivers/net/ethernet/intel/i40e/i40e_debugfs.c hash_for_each(vsi->mac_filter_hash, bkt, f, hlist) { bkt 1235 drivers/net/ethernet/intel/i40e/i40e_main.c int bkt, new_vlan; bkt 1262 drivers/net/ethernet/intel/i40e/i40e_main.c hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, hlist) { bkt 1474 drivers/net/ethernet/intel/i40e/i40e_main.c int bkt; bkt 1483 drivers/net/ethernet/intel/i40e/i40e_main.c hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, hlist) { bkt 1509 drivers/net/ethernet/intel/i40e/i40e_main.c int bkt; bkt 1512 drivers/net/ethernet/intel/i40e/i40e_main.c hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, hlist) { bkt 2279 drivers/net/ethernet/intel/i40e/i40e_main.c int bkt; bkt 2309 drivers/net/ethernet/intel/i40e/i40e_main.c hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, hlist) { bkt 2486 drivers/net/ethernet/intel/i40e/i40e_main.c hash_for_each(vsi->mac_filter_hash, bkt, f, hlist) { bkt 2765 drivers/net/ethernet/intel/i40e/i40e_main.c int bkt; bkt 2767 drivers/net/ethernet/intel/i40e/i40e_main.c hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, hlist) { bkt 2836 drivers/net/ethernet/intel/i40e/i40e_main.c int bkt; bkt 2838 drivers/net/ethernet/intel/i40e/i40e_main.c hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, hlist) { bkt 13086 drivers/net/ethernet/intel/i40e/i40e_main.c int bkt; bkt 13292 drivers/net/ethernet/intel/i40e/i40e_main.c hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, hlist) { bkt 13331 drivers/net/ethernet/intel/i40e/i40e_main.c int i, n, bkt; bkt 13370 drivers/net/ethernet/intel/i40e/i40e_main.c hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, hlist) bkt 1132 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c int bkt; bkt 1169 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c hash_for_each(vsi->mac_filter_hash, bkt, f, hlist) { bkt 1985 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c int num_vlans = 0, bkt; bkt 1987 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c hash_for_each(vsi->mac_filter_hash, bkt, f, hlist) { bkt 3135 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c int bkt; bkt 3173 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, bkt 3965 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c int bkt; bkt 4021 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c hash_for_each_safe(vsi->mac_filter_hash, bkt, h, f, hlist) bkt 4057 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c int bkt; bkt 4065 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c hash_for_each(priv->fs.tc.hairpin_tbl, bkt, hpe, hairpin_hlist) bkt 219 drivers/net/ethernet/mellanox/mlx5/core/lib/vxlan.c int bkt; bkt 225 drivers/net/ethernet/mellanox/mlx5/core/lib/vxlan.c hash_for_each_safe(vxlan->htable, bkt, tmp, vxlanp, hlist) { bkt 1952 drivers/net/ethernet/rocker/rocker_ofdpa.c int bkt; bkt 1963 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_for_each_safe(ofdpa->fdb_tbl, bkt, tmp, found, entry) { bkt 1993 drivers/net/ethernet/rocker/rocker_ofdpa.c int bkt; bkt 1997 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_for_each_safe(ofdpa->fdb_tbl, bkt, tmp, entry, entry) { bkt 2386 drivers/net/ethernet/rocker/rocker_ofdpa.c int bkt; bkt 2392 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_for_each_safe(ofdpa->flow_tbl, bkt, tmp, flow_entry, entry) bkt 2397 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_for_each_safe(ofdpa->group_tbl, bkt, tmp, group_entry, entry) bkt 2402 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_for_each_safe(ofdpa->fdb_tbl, bkt, tmp, fdb_entry, entry) bkt 2407 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_for_each_safe(ofdpa->internal_vlan_tbl, bkt, bkt 2413 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_for_each_safe(ofdpa->neigh_tbl, bkt, tmp, neigh_entry, entry) bkt 2781 drivers/net/ethernet/rocker/rocker_ofdpa.c int bkt; bkt 2787 drivers/net/ethernet/rocker/rocker_ofdpa.c hash_for_each_safe(ofdpa->flow_tbl, bkt, tmp, flow_entry, entry) { bkt 96 drivers/staging/netlogic/xlr_net.c static void xlr_net_fmn_handler(int bkt, int src_stnid, int size, int code, bkt 126 include/linux/hashtable.h #define hash_for_each(name, bkt, obj, member) \ bkt 127 include/linux/hashtable.h for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\ bkt 128 include/linux/hashtable.h (bkt)++)\ bkt 129 include/linux/hashtable.h hlist_for_each_entry(obj, &name[bkt], member) bkt 138 include/linux/hashtable.h #define hash_for_each_rcu(name, bkt, obj, member) \ bkt 139 include/linux/hashtable.h for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\ bkt 140 include/linux/hashtable.h (bkt)++)\ bkt 141 include/linux/hashtable.h hlist_for_each_entry_rcu(obj, &name[bkt], member) bkt 152 include/linux/hashtable.h #define hash_for_each_safe(name, bkt, tmp, obj, member) \ bkt 153 include/linux/hashtable.h for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\ bkt 154 include/linux/hashtable.h (bkt)++)\ bkt 155 include/linux/hashtable.h hlist_for_each_entry_safe(obj, tmp, &name[bkt], member) bkt 328 include/linux/rhashtable.h struct rhash_lock_head **bkt) bkt 331 include/linux/rhashtable.h bit_spin_lock(0, (unsigned long *)bkt); bkt 345 include/linux/rhashtable.h struct rhash_lock_head **bkt) bkt 348 include/linux/rhashtable.h bit_spin_unlock(0, (unsigned long *)bkt); bkt 353 include/linux/rhashtable.h struct rhash_lock_head *const *bkt) bkt 356 include/linux/rhashtable.h ((unsigned long)*bkt & ~BIT(0) ?: bkt 357 include/linux/rhashtable.h (unsigned long)RHT_NULLS_MARKER(bkt)); bkt 368 include/linux/rhashtable.h struct rhash_lock_head *const *bkt) bkt 370 include/linux/rhashtable.h struct rhash_head __rcu *p = __rht_ptr(bkt); bkt 376 include/linux/rhashtable.h struct rhash_lock_head *const *bkt, bkt 380 include/linux/rhashtable.h return rht_dereference_bucket(__rht_ptr(bkt), tbl, hash); bkt 384 include/linux/rhashtable.h struct rhash_lock_head *const *bkt) bkt 386 include/linux/rhashtable.h return rcu_dereference_protected(__rht_ptr(bkt), 1); bkt 389 include/linux/rhashtable.h static inline void rht_assign_locked(struct rhash_lock_head **bkt, bkt 392 include/linux/rhashtable.h struct rhash_head __rcu **p = (struct rhash_head __rcu **)bkt; bkt 400 include/linux/rhashtable.h struct rhash_lock_head **bkt, bkt 403 include/linux/rhashtable.h struct rhash_head __rcu **p = (struct rhash_head __rcu **)bkt; bkt 596 include/linux/rhashtable.h struct rhash_lock_head *const *bkt; bkt 604 include/linux/rhashtable.h bkt = rht_bucket(tbl, hash); bkt 606 include/linux/rhashtable.h rht_for_each_rcu_from(he, rht_ptr_rcu(bkt), tbl, hash) { bkt 616 include/linux/rhashtable.h } while (he != RHT_NULLS_MARKER(bkt)); bkt 712 include/linux/rhashtable.h struct rhash_lock_head **bkt; bkt 725 include/linux/rhashtable.h bkt = rht_bucket_insert(ht, tbl, hash); bkt 727 include/linux/rhashtable.h if (!bkt) bkt 730 include/linux/rhashtable.h rht_lock(tbl, bkt); bkt 734 include/linux/rhashtable.h rht_unlock(tbl, bkt); bkt 739 include/linux/rhashtable.h rht_for_each_from(head, rht_ptr(bkt, tbl, hash), tbl, hash) { bkt 766 include/linux/rhashtable.h rht_unlock(tbl, bkt); bkt 768 include/linux/rhashtable.h rht_assign_unlock(tbl, bkt, obj); bkt 784 include/linux/rhashtable.h head = rht_ptr(bkt, tbl, hash); bkt 795 include/linux/rhashtable.h rht_assign_unlock(tbl, bkt, obj); bkt 807 include/linux/rhashtable.h rht_unlock(tbl, bkt); bkt 998 include/linux/rhashtable.h struct rhash_lock_head **bkt; bkt 1005 include/linux/rhashtable.h bkt = rht_bucket_var(tbl, hash); bkt 1006 include/linux/rhashtable.h if (!bkt) bkt 1009 include/linux/rhashtable.h rht_lock(tbl, bkt); bkt 1011 include/linux/rhashtable.h rht_for_each_from(he, rht_ptr(bkt, tbl, hash), tbl, hash) { bkt 1053 include/linux/rhashtable.h rht_unlock(tbl, bkt); bkt 1055 include/linux/rhashtable.h rht_assign_unlock(tbl, bkt, obj); bkt 1060 include/linux/rhashtable.h rht_unlock(tbl, bkt); bkt 1150 include/linux/rhashtable.h struct rhash_lock_head **bkt; bkt 1163 include/linux/rhashtable.h bkt = rht_bucket_var(tbl, hash); bkt 1164 include/linux/rhashtable.h if (!bkt) bkt 1168 include/linux/rhashtable.h rht_lock(tbl, bkt); bkt 1170 include/linux/rhashtable.h rht_for_each_from(he, rht_ptr(bkt, tbl, hash), tbl, hash) { bkt 1179 include/linux/rhashtable.h rht_unlock(tbl, bkt); bkt 1181 include/linux/rhashtable.h rht_assign_unlock(tbl, bkt, obj_new); bkt 1187 include/linux/rhashtable.h rht_unlock(tbl, bkt); bkt 4660 kernel/workqueue.c int bkt; bkt 4669 kernel/workqueue.c hash_for_each(pool->busy_hash, bkt, worker, hentry) { bkt 4679 kernel/workqueue.c hash_for_each(pool->busy_hash, bkt, worker, hentry) { bkt 5975 kernel/workqueue.c int cpu, bkt; bkt 6013 kernel/workqueue.c hash_for_each(unbound_pool_hash, bkt, pool, hash_node) bkt 216 lib/rhashtable.c struct rhash_lock_head **bkt, bkt 231 lib/rhashtable.c rht_for_each_from(entry, rht_ptr(bkt, old_tbl, old_hash), bkt 259 lib/rhashtable.c rht_assign_locked(bkt, next); bkt 269 lib/rhashtable.c struct rhash_lock_head **bkt = rht_bucket_var(old_tbl, old_hash); bkt 272 lib/rhashtable.c if (!bkt) bkt 274 lib/rhashtable.c rht_lock(old_tbl, bkt); bkt 276 lib/rhashtable.c while (!(err = rhashtable_rehash_one(ht, bkt, old_hash))) bkt 281 lib/rhashtable.c rht_unlock(old_tbl, bkt); bkt 479 lib/rhashtable.c struct rhash_lock_head **bkt, bkt 492 lib/rhashtable.c rht_for_each_from(head, rht_ptr(bkt, tbl, hash), tbl, hash) { bkt 518 lib/rhashtable.c rht_assign_locked(bkt, obj); bkt 530 lib/rhashtable.c struct rhash_lock_head **bkt, bkt 558 lib/rhashtable.c head = rht_ptr(bkt, tbl, hash); bkt 571 lib/rhashtable.c rht_assign_locked(bkt, obj); bkt 585 lib/rhashtable.c struct rhash_lock_head **bkt; bkt 596 lib/rhashtable.c bkt = rht_bucket_var(tbl, hash); bkt 598 lib/rhashtable.c bkt = rht_bucket_insert(ht, tbl, hash); bkt 599 lib/rhashtable.c if (bkt == NULL) { bkt 603 lib/rhashtable.c rht_lock(tbl, bkt); bkt 604 lib/rhashtable.c data = rhashtable_lookup_one(ht, bkt, tbl, bkt 606 lib/rhashtable.c new_tbl = rhashtable_insert_one(ht, bkt, tbl, bkt 611 lib/rhashtable.c rht_unlock(tbl, bkt); bkt 238 net/ipv4/cipso_ipv4.c u32 bkt; bkt 247 net/ipv4/cipso_ipv4.c bkt = hash & (CIPSO_V4_CACHE_BUCKETS - 1); bkt 248 net/ipv4/cipso_ipv4.c spin_lock_bh(&cipso_v4_cache[bkt].lock); bkt 249 net/ipv4/cipso_ipv4.c list_for_each_entry(entry, &cipso_v4_cache[bkt].list, list) { bkt 259 net/ipv4/cipso_ipv4.c spin_unlock_bh(&cipso_v4_cache[bkt].lock); bkt 274 net/ipv4/cipso_ipv4.c spin_unlock_bh(&cipso_v4_cache[bkt].lock); bkt 279 net/ipv4/cipso_ipv4.c spin_unlock_bh(&cipso_v4_cache[bkt].lock); bkt 301 net/ipv4/cipso_ipv4.c u32 bkt; bkt 324 net/ipv4/cipso_ipv4.c bkt = entry->hash & (CIPSO_V4_CACHE_BUCKETS - 1); bkt 325 net/ipv4/cipso_ipv4.c spin_lock_bh(&cipso_v4_cache[bkt].lock); bkt 326 net/ipv4/cipso_ipv4.c if (cipso_v4_cache[bkt].size < cipso_v4_cache_bucketsize) { bkt 327 net/ipv4/cipso_ipv4.c list_add(&entry->list, &cipso_v4_cache[bkt].list); bkt 328 net/ipv4/cipso_ipv4.c cipso_v4_cache[bkt].size += 1; bkt 330 net/ipv4/cipso_ipv4.c old_entry = list_entry(cipso_v4_cache[bkt].list.prev, bkt 333 net/ipv4/cipso_ipv4.c list_add(&entry->list, &cipso_v4_cache[bkt].list); bkt 336 net/ipv4/cipso_ipv4.c spin_unlock_bh(&cipso_v4_cache[bkt].lock); bkt 200 net/ipv6/calipso.c u32 bkt; bkt 209 net/ipv6/calipso.c bkt = hash & (CALIPSO_CACHE_BUCKETS - 1); bkt 210 net/ipv6/calipso.c spin_lock_bh(&calipso_cache[bkt].lock); bkt 211 net/ipv6/calipso.c list_for_each_entry(entry, &calipso_cache[bkt].list, list) { bkt 221 net/ipv6/calipso.c spin_unlock_bh(&calipso_cache[bkt].lock); bkt 236 net/ipv6/calipso.c spin_unlock_bh(&calipso_cache[bkt].lock); bkt 241 net/ipv6/calipso.c spin_unlock_bh(&calipso_cache[bkt].lock); bkt 265 net/ipv6/calipso.c u32 bkt; bkt 288 net/ipv6/calipso.c bkt = entry->hash & (CALIPSO_CACHE_BUCKETS - 1); bkt 289 net/ipv6/calipso.c spin_lock_bh(&calipso_cache[bkt].lock); bkt 290 net/ipv6/calipso.c if (calipso_cache[bkt].size < calipso_cache_bucketsize) { bkt 291 net/ipv6/calipso.c list_add(&entry->list, &calipso_cache[bkt].list); bkt 292 net/ipv6/calipso.c calipso_cache[bkt].size += 1; bkt 294 net/ipv6/calipso.c old_entry = list_entry(calipso_cache[bkt].list.prev, bkt 297 net/ipv6/calipso.c list_add(&entry->list, &calipso_cache[bkt].list); bkt 300 net/ipv6/calipso.c spin_unlock_bh(&calipso_cache[bkt].lock); bkt 139 net/netlabel/netlabel_domainhash.c u32 bkt; bkt 144 net/netlabel/netlabel_domainhash.c bkt = netlbl_domhsh_hash(domain); bkt 145 net/netlabel/netlabel_domainhash.c bkt_list = &netlbl_domhsh_rcu_deref(netlbl_domhsh)->tbl[bkt]; bkt 434 net/netlabel/netlabel_domainhash.c u32 bkt = netlbl_domhsh_hash(entry->domain); bkt 436 net/netlabel/netlabel_domainhash.c &rcu_dereference(netlbl_domhsh)->tbl[bkt]); bkt 204 net/netlabel/netlabel_unlabeled.c u32 bkt; bkt 208 net/netlabel/netlabel_unlabeled.c bkt = netlbl_unlhsh_hash(ifindex); bkt 209 net/netlabel/netlabel_unlabeled.c bkt_list = &netlbl_unlhsh_rcu_deref(netlbl_unlhsh)->tbl[bkt]; bkt 313 net/netlabel/netlabel_unlabeled.c u32 bkt; bkt 327 net/netlabel/netlabel_unlabeled.c bkt = netlbl_unlhsh_hash(ifindex); bkt 331 net/netlabel/netlabel_unlabeled.c &netlbl_unlhsh_rcu_deref(netlbl_unlhsh)->tbl[bkt]); bkt 54 net/sched/cls_route.c struct route4_bucket *bkt; bkt 326 net/sched/cls_route.c b = f->bkt; bkt 454 net/sched/cls_route.c f->bkt = b; bkt 508 net/sched/cls_route.c f->bkt = fold->bkt; bkt 518 net/sched/cls_route.c fp = &f->bkt->ht[h]; bkt 568 tools/bpf/bpftool/btf.c unsigned int bkt; bkt 570 tools/bpf/bpftool/btf.c hash_for_each_safe(tab->table, bkt, tmp, obj, hash) { bkt 427 tools/bpf/bpftool/common.c unsigned int bkt; bkt 429 tools/bpf/bpftool/common.c hash_for_each_safe(tab->table, bkt, tmp, obj, hash) { bkt 105 tools/include/linux/hashtable.h #define hash_for_each(name, bkt, obj, member) \ bkt 106 tools/include/linux/hashtable.h for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\ bkt 107 tools/include/linux/hashtable.h (bkt)++)\ bkt 108 tools/include/linux/hashtable.h hlist_for_each_entry(obj, &name[bkt], member) bkt 119 tools/include/linux/hashtable.h #define hash_for_each_safe(name, bkt, tmp, obj, member) \ bkt 120 tools/include/linux/hashtable.h for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\ bkt 121 tools/include/linux/hashtable.h (bkt)++)\ bkt 122 tools/include/linux/hashtable.h hlist_for_each_entry_safe(obj, tmp, &name[bkt], member) bkt 94 tools/lib/bpf/hashmap.c int bkt; bkt 105 tools/lib/bpf/hashmap.c hashmap__for_each_entry_safe(map, cur, tmp, bkt) { bkt 138 tools/lib/bpf/hashmap.h #define hashmap__for_each_entry(map, cur, bkt) \ bkt 139 tools/lib/bpf/hashmap.h for (bkt = 0; bkt < map->cap; bkt++) \ bkt 140 tools/lib/bpf/hashmap.h for (cur = map->buckets[bkt]; cur; cur = cur->next) bkt 150 tools/lib/bpf/hashmap.h #define hashmap__for_each_entry_safe(map, cur, tmp, bkt) \ bkt 151 tools/lib/bpf/hashmap.h for (bkt = 0; bkt < map->cap; bkt++) \ bkt 152 tools/lib/bpf/hashmap.h for (cur = map->buckets[bkt]; \ bkt 163 tools/lib/bpf/hashmap.h for (cur = ({ size_t bkt = hash_bits(map->hash_fn((_key), map->ctx),\ bkt 165 tools/lib/bpf/hashmap.h map->buckets ? map->buckets[bkt] : NULL; }); \ bkt 171 tools/lib/bpf/hashmap.h for (cur = ({ size_t bkt = hash_bits(map->hash_fn((_key), map->ctx),\ bkt 173 tools/lib/bpf/hashmap.h cur = map->buckets ? map->buckets[bkt] : NULL; }); \ bkt 55 tools/testing/selftests/bpf/test_hashmap.c int err, bkt, found_cnt, i; bkt 101 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_entry(map, entry, bkt) { bkt 144 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_entry_safe(map, entry, tmp, bkt) { bkt 201 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_entry_safe(map, entry, tmp, bkt) { bkt 235 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_entry(map, entry, bkt) { bkt 242 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_entry(map, entry, bkt) { bkt 263 tools/testing/selftests/bpf/test_hashmap.c int err, bkt; bkt 303 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_entry(map, entry, bkt) { bkt 335 tools/testing/selftests/bpf/test_hashmap.c int bkt; bkt 357 tools/testing/selftests/bpf/test_hashmap.c hashmap__for_each_entry(map, entry, bkt) {