ht 258 arch/mips/include/asm/octeon/cvmx-pci-defs.h uint32_t ht:8; ht 264 arch/mips/include/asm/octeon/cvmx-pci-defs.h uint32_t ht:8; ht 79 arch/mips/include/asm/txx9/tx3927.h endian_def_b4(unused0, ht, mlt, cls); ht 225 arch/mips/kernel/kgdb.c struct hard_trap_info *ht; ht 227 arch/mips/kernel/kgdb.c for (ht = hard_trap_info; ht->tt && ht->signo; ht++) ht 228 arch/mips/kernel/kgdb.c if (ht->tt == tt) ht 229 arch/mips/kernel/kgdb.c return ht->signo; ht 166 arch/mips/kernel/ptrace.c u16 ht[NUM_WATCH_REGS]; ht 187 arch/mips/kernel/ptrace.c __get_user(ht[i], &addr->WATCH_STYLE.watchhi[i]); ht 188 arch/mips/kernel/ptrace.c if (ht[i] & ~MIPS_WATCHHI_MASK) ht 197 arch/mips/kernel/ptrace.c child->thread.watch.mips3264.watchhi[i] = ht[i]; ht 96 arch/powerpc/kernel/kgdb.c struct hard_trap_info *ht; ht 98 arch/powerpc/kernel/kgdb.c for (ht = hard_trap_info; ht->tt && ht->signo; ht++) ht 99 arch/powerpc/kernel/kgdb.c if (ht->tt == tt) ht 100 arch/powerpc/kernel/kgdb.c return ht->signo; ht 591 arch/powerpc/platforms/maple/pci.c struct device_node *ht = NULL; ht 613 arch/powerpc/platforms/maple/pci.c ht = np; ht 620 arch/powerpc/platforms/maple/pci.c if (ht && maple_add_bridge(ht) != 0) ht 621 arch/powerpc/platforms/maple/pci.c of_node_put(ht); ht 901 arch/powerpc/platforms/powermac/pci.c struct device_node *ht __maybe_unused = NULL; ht 920 arch/powerpc/platforms/powermac/pci.c ht = np; ht 929 arch/powerpc/platforms/powermac/pci.c if (ht && pmac_add_bridge(ht) != 0) ht 930 arch/powerpc/platforms/powermac/pci.c of_node_put(ht); ht 1292 arch/x86/events/intel/p4.c PMU_FORMAT_ATTR(ht, "config:63" ); ht 758 arch/x86/events/perf_event.h #define EVENT_ATTR_STR_HT(_name, v, noht, ht) \ ht 763 arch/x86/events/perf_event.h .event_str_ht = ht, \ ht 227 drivers/clk/pxa/clk-pxa27x.c unsigned int t, ht; ht 233 drivers/clk/pxa/clk-pxa27x.c ht = clkcfg & (1 << 2); ht 324 drivers/clk/pxa/clk-pxa27x.c unsigned int t, ht, osc_forced; ht 333 drivers/clk/pxa/clk-pxa27x.c ht = clkcfg & (1 << 2); ht 335 drivers/clk/pxa/clk-pxa27x.c if (ht || t) ht 252 drivers/gpu/drm/ast/ast_drv.h u32 ht; ht 188 drivers/gpu/drm/ast/ast_mode.c adjusted_mode->crtc_htotal = vbios_mode->enh_table->ht; ht 190 drivers/gpu/drm/ast/ast_mode.c adjusted_mode->crtc_hblank_end = vbios_mode->enh_table->ht - hborder; ht 45 drivers/gpu/drm/drm_hashtab.c int drm_ht_create(struct drm_open_hash *ht, unsigned int order) ht 49 drivers/gpu/drm/drm_hashtab.c ht->order = order; ht 50 drivers/gpu/drm/drm_hashtab.c ht->table = NULL; ht 51 drivers/gpu/drm/drm_hashtab.c if (size <= PAGE_SIZE / sizeof(*ht->table)) ht 52 drivers/gpu/drm/drm_hashtab.c ht->table = kcalloc(size, sizeof(*ht->table), GFP_KERNEL); ht 54 drivers/gpu/drm/drm_hashtab.c ht->table = vzalloc(array_size(size, sizeof(*ht->table))); ht 55 drivers/gpu/drm/drm_hashtab.c if (!ht->table) { ht 63 drivers/gpu/drm/drm_hashtab.c void drm_ht_verbose_list(struct drm_open_hash *ht, unsigned long key) ht 70 drivers/gpu/drm/drm_hashtab.c hashed_key = hash_long(key, ht->order); ht 72 drivers/gpu/drm/drm_hashtab.c h_list = &ht->table[hashed_key]; ht 77 drivers/gpu/drm/drm_hashtab.c static struct hlist_node *drm_ht_find_key(struct drm_open_hash *ht, ht 84 drivers/gpu/drm/drm_hashtab.c hashed_key = hash_long(key, ht->order); ht 85 drivers/gpu/drm/drm_hashtab.c h_list = &ht->table[hashed_key]; ht 95 drivers/gpu/drm/drm_hashtab.c static struct hlist_node *drm_ht_find_key_rcu(struct drm_open_hash *ht, ht 102 drivers/gpu/drm/drm_hashtab.c hashed_key = hash_long(key, ht->order); ht 103 drivers/gpu/drm/drm_hashtab.c h_list = &ht->table[hashed_key]; ht 113 drivers/gpu/drm/drm_hashtab.c int drm_ht_insert_item(struct drm_open_hash *ht, struct drm_hash_item *item) ht 121 drivers/gpu/drm/drm_hashtab.c hashed_key = hash_long(key, ht->order); ht 122 drivers/gpu/drm/drm_hashtab.c h_list = &ht->table[hashed_key]; ht 144 drivers/gpu/drm/drm_hashtab.c int drm_ht_just_insert_please(struct drm_open_hash *ht, struct drm_hash_item *item, ht 156 drivers/gpu/drm/drm_hashtab.c ret = drm_ht_insert_item(ht, item); ht 169 drivers/gpu/drm/drm_hashtab.c int drm_ht_find_item(struct drm_open_hash *ht, unsigned long key, ht 174 drivers/gpu/drm/drm_hashtab.c list = drm_ht_find_key_rcu(ht, key); ht 183 drivers/gpu/drm/drm_hashtab.c int drm_ht_remove_key(struct drm_open_hash *ht, unsigned long key) ht 187 drivers/gpu/drm/drm_hashtab.c list = drm_ht_find_key(ht, key); ht 195 drivers/gpu/drm/drm_hashtab.c int drm_ht_remove_item(struct drm_open_hash *ht, struct drm_hash_item *item) ht 202 drivers/gpu/drm/drm_hashtab.c void drm_ht_remove(struct drm_open_hash *ht) ht 204 drivers/gpu/drm/drm_hashtab.c if (ht->table) { ht 205 drivers/gpu/drm/drm_hashtab.c kvfree(ht->table); ht 206 drivers/gpu/drm/drm_hashtab.c ht->table = NULL; ht 105 drivers/gpu/drm/i2c/ch7006_mode.c #define __MODE(f, hd, vd, ht, vt, hsynp, vsynp, \ ht 115 drivers/gpu/drm/i2c/ch7006_mode.c .htotal = ht, \ ht 135 drivers/gpu/drm/i2c/ch7006_mode.c #define MODE(f, hd, vd, ht, vt, hsynp, vsynp, \ ht 137 drivers/gpu/drm/i2c/ch7006_mode.c __MODE(f, hd, vd, ht, vt, hsynp, vsynp, subc, scale, \ ht 258 drivers/gpu/drm/ingenic/ingenic-drm.c unsigned int vpe, vds, vde, vt, hpe, hds, hde, ht; ht 268 drivers/gpu/drm/ingenic/ingenic-drm.c ht = hde + mode->hsync_start - mode->hdisplay; ht 279 drivers/gpu/drm/ingenic/ingenic-drm.c ht << JZ_LCD_VAT_HT_OFFSET | ht 3215 drivers/gpu/drm/omapdrm/dss/dispc.c unsigned long ht, vt; ht 3231 drivers/gpu/drm/omapdrm/dss/dispc.c ht = vm->pixelclock / xtot; ht 3245 drivers/gpu/drm/omapdrm/dss/dispc.c DSSDBG("hsync %luHz, vsync %luHz\n", ht, vt); ht 248 drivers/gpu/drm/vmwgfx/ttm_object.c struct drm_open_hash *ht = &tfile->ref_hash[TTM_REF_USAGE]; ht 252 drivers/gpu/drm/vmwgfx/ttm_object.c ret = drm_ht_find_item_rcu(ht, key, &hash); ht 268 drivers/gpu/drm/vmwgfx/ttm_object.c struct drm_open_hash *ht = &tfile->ref_hash[TTM_REF_USAGE]; ht 272 drivers/gpu/drm/vmwgfx/ttm_object.c ret = drm_ht_find_item_rcu(ht, key, &hash); ht 312 drivers/gpu/drm/vmwgfx/ttm_object.c struct drm_open_hash *ht = &tfile->ref_hash[TTM_REF_USAGE]; ht 317 drivers/gpu/drm/vmwgfx/ttm_object.c if (unlikely(drm_ht_find_item_rcu(ht, base->handle, &hash) != 0)) ht 349 drivers/gpu/drm/vmwgfx/ttm_object.c struct drm_open_hash *ht = &tfile->ref_hash[ref_type]; ht 367 drivers/gpu/drm/vmwgfx/ttm_object.c ret = drm_ht_find_item_rcu(ht, base->handle, &hash); ht 398 drivers/gpu/drm/vmwgfx/ttm_object.c ret = drm_ht_insert_item_rcu(ht, &ref->hash); ht 426 drivers/gpu/drm/vmwgfx/ttm_object.c struct drm_open_hash *ht; ht 429 drivers/gpu/drm/vmwgfx/ttm_object.c ht = &tfile->ref_hash[ref->ref_type]; ht 430 drivers/gpu/drm/vmwgfx/ttm_object.c (void)drm_ht_remove_item_rcu(ht, &ref->hash); ht 446 drivers/gpu/drm/vmwgfx/ttm_object.c struct drm_open_hash *ht = &tfile->ref_hash[ref_type]; ht 452 drivers/gpu/drm/vmwgfx/ttm_object.c ret = drm_ht_find_item(ht, key, &hash); ht 182 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c if (ctx->ht) { ht 185 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c if (!drm_ht_find_item(ctx->ht, (unsigned long) vbo, &hash)) ht 219 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c if (ctx->ht) { ht 222 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c if (!drm_ht_find_item(ctx->ht, (unsigned long) res, &hash)) ht 277 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c if (ctx->ht) { ht 279 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c ret = drm_ht_insert_item(ctx->ht, &bo_node->hash); ht 332 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c if (ctx->ht) { ht 334 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c ret = drm_ht_insert_item(ctx->ht, &node->hash); ht 642 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c if (!ctx->ht) ht 646 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c (void) drm_ht_remove_item(ctx->ht, &entry->hash); ht 649 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c (void) drm_ht_remove_item(ctx->ht, &val->hash); ht 652 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c (void) drm_ht_remove_item(ctx->ht, &val->hash); ht 654 drivers/gpu/drm/vmwgfx/vmwgfx_validation.c ctx->ht = NULL; ht 76 drivers/gpu/drm/vmwgfx/vmwgfx_validation.h struct drm_open_hash *ht; ht 108 drivers/gpu/drm/vmwgfx/vmwgfx_validation.h { .ht = _ht, \ ht 154 drivers/gpu/drm/vmwgfx/vmwgfx_validation.h struct drm_open_hash *ht) ht 156 drivers/gpu/drm/vmwgfx/vmwgfx_validation.h ctx->ht = ht; ht 136 drivers/infiniband/hw/hfi1/affinity.c int possible, curr_cpu, i, ht; ht 147 drivers/infiniband/hw/hfi1/affinity.c ht = cpumask_weight(topology_sibling_cpumask( ht 155 drivers/infiniband/hw/hfi1/affinity.c for (i = 0; i < possible / ht; i++) ht 583 drivers/md/dm-cache-policy-smq.c static int h_init(struct smq_hash_table *ht, struct entry_space *es, unsigned nr_entries) ht 587 drivers/md/dm-cache-policy-smq.c ht->es = es; ht 589 drivers/md/dm-cache-policy-smq.c ht->hash_bits = __ffs(nr_buckets); ht 591 drivers/md/dm-cache-policy-smq.c ht->buckets = vmalloc(array_size(nr_buckets, sizeof(*ht->buckets))); ht 592 drivers/md/dm-cache-policy-smq.c if (!ht->buckets) ht 596 drivers/md/dm-cache-policy-smq.c ht->buckets[i] = INDEXER_NULL; ht 601 drivers/md/dm-cache-policy-smq.c static void h_exit(struct smq_hash_table *ht) ht 603 drivers/md/dm-cache-policy-smq.c vfree(ht->buckets); ht 606 drivers/md/dm-cache-policy-smq.c static struct entry *h_head(struct smq_hash_table *ht, unsigned bucket) ht 608 drivers/md/dm-cache-policy-smq.c return to_entry(ht->es, ht->buckets[bucket]); ht 611 drivers/md/dm-cache-policy-smq.c static struct entry *h_next(struct smq_hash_table *ht, struct entry *e) ht 613 drivers/md/dm-cache-policy-smq.c return to_entry(ht->es, e->hash_next); ht 616 drivers/md/dm-cache-policy-smq.c static void __h_insert(struct smq_hash_table *ht, unsigned bucket, struct entry *e) ht 618 drivers/md/dm-cache-policy-smq.c e->hash_next = ht->buckets[bucket]; ht 619 drivers/md/dm-cache-policy-smq.c ht->buckets[bucket] = to_index(ht->es, e); ht 622 drivers/md/dm-cache-policy-smq.c static void h_insert(struct smq_hash_table *ht, struct entry *e) ht 624 drivers/md/dm-cache-policy-smq.c unsigned h = hash_64(from_oblock(e->oblock), ht->hash_bits); ht 625 drivers/md/dm-cache-policy-smq.c __h_insert(ht, h, e); ht 628 drivers/md/dm-cache-policy-smq.c static struct entry *__h_lookup(struct smq_hash_table *ht, unsigned h, dm_oblock_t oblock, ht 634 drivers/md/dm-cache-policy-smq.c for (e = h_head(ht, h); e; e = h_next(ht, e)) { ht 644 drivers/md/dm-cache-policy-smq.c static void __h_unlink(struct smq_hash_table *ht, unsigned h, ht 650 drivers/md/dm-cache-policy-smq.c ht->buckets[h] = e->hash_next; ht 656 drivers/md/dm-cache-policy-smq.c static struct entry *h_lookup(struct smq_hash_table *ht, dm_oblock_t oblock) ht 659 drivers/md/dm-cache-policy-smq.c unsigned h = hash_64(from_oblock(oblock), ht->hash_bits); ht 661 drivers/md/dm-cache-policy-smq.c e = __h_lookup(ht, h, oblock, &prev); ht 667 drivers/md/dm-cache-policy-smq.c __h_unlink(ht, h, e, prev); ht 668 drivers/md/dm-cache-policy-smq.c __h_insert(ht, h, e); ht 674 drivers/md/dm-cache-policy-smq.c static void h_remove(struct smq_hash_table *ht, struct entry *e) ht 676 drivers/md/dm-cache-policy-smq.c unsigned h = hash_64(from_oblock(e->oblock), ht->hash_bits); ht 683 drivers/md/dm-cache-policy-smq.c e = __h_lookup(ht, h, e->oblock, &prev); ht 685 drivers/md/dm-cache-policy-smq.c __h_unlink(ht, h, e, prev); ht 96 drivers/md/dm-clone-target.c struct hash_table_bucket *ht; ht 584 drivers/md/dm-clone-target.c clone->ht = kvmalloc(sz * sizeof(struct hash_table_bucket), GFP_KERNEL); ht 585 drivers/md/dm-clone-target.c if (!clone->ht) ht 589 drivers/md/dm-clone-target.c bucket = clone->ht + i; ht 600 drivers/md/dm-clone-target.c kvfree(clone->ht); ht 606 drivers/md/dm-clone-target.c return &clone->ht[hash_long(region_nr, HASH_TABLE_BITS)]; ht 17 drivers/net/ethernet/mellanox/mlx5/core/en/fs.h struct rhashtable ht; ht 785 drivers/net/ethernet/mellanox/mlx5/core/en_rx.c int ht = (cht & CQE_RSS_HTYPE_L4) ? PKT_HASH_TYPE_L4 : ht 788 drivers/net/ethernet/mellanox/mlx5/core/en_rx.c skb_set_hash(skb, be32_to_cpu(cqe->rss_hash_result), ht); ht 3500 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c return &priv->fs.tc.ht; ht 4117 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c err = rhashtable_init(&tc->ht, &tc_ht_params); ht 4149 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c rhashtable_destroy(&tc->ht); ht 88 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c struct mlx5i_pkey_qpn_ht *ht = ipriv->qpn_htbl; ht 98 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c spin_lock_bh(&ht->ht_lock); ht 99 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c hlist_add_head(&new_node->hlist, &ht->buckets[key]); ht 100 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c spin_unlock_bh(&ht->ht_lock); ht 109 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c struct mlx5i_pkey_qpn_ht *ht = ipriv->qpn_htbl; ht 112 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c node = mlx5i_find_qpn_to_netdev_node(ht->buckets, qpn); ht 118 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c spin_lock_bh(&ht->ht_lock); ht 120 drivers/net/ethernet/mellanox/mlx5/core/ipoib/ipoib_vlan.c spin_unlock_bh(&ht->ht_lock); ht 432 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c struct rhashtable ht; ht 464 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c err = rhashtable_init(&fib->ht, &mlxsw_sp_fib_ht_params); ht 490 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c rhashtable_destroy(&fib->ht); ht 4476 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c return rhashtable_insert_fast(&fib->ht, &fib_node->ht_node, ht 4483 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c rhashtable_remove_fast(&fib->ht, &fib_node->ht_node, ht 4496 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c return rhashtable_lookup_fast(&fib->ht, &key, mlxsw_sp_fib_ht_params); ht 474 drivers/net/wireless/ath/ath10k/core.h u64 ht[ATH10K_COUNTER_TYPE_MAX][ATH10K_HT_MCS_NUM]; ht 692 drivers/net/wireless/ath/ath10k/debugfs_sta.c " %llu ", stats->ht[j][i]); ht 3362 drivers/net/wireless/ath/ath10k/htt_rx.c STATS_OP_FMT(SUCC).ht[0][ht_idx] += pstats->succ_bytes; ht 3363 drivers/net/wireless/ath/ath10k/htt_rx.c STATS_OP_FMT(SUCC).ht[1][ht_idx] += pstats->succ_pkts; ht 3364 drivers/net/wireless/ath/ath10k/htt_rx.c STATS_OP_FMT(FAIL).ht[0][ht_idx] += pstats->failed_bytes; ht 3365 drivers/net/wireless/ath/ath10k/htt_rx.c STATS_OP_FMT(FAIL).ht[1][ht_idx] += pstats->failed_pkts; ht 3366 drivers/net/wireless/ath/ath10k/htt_rx.c STATS_OP_FMT(RETRY).ht[0][ht_idx] += pstats->retry_bytes; ht 3367 drivers/net/wireless/ath/ath10k/htt_rx.c STATS_OP_FMT(RETRY).ht[1][ht_idx] += pstats->retry_pkts; ht 3383 drivers/net/wireless/ath/ath10k/htt_rx.c STATS_OP_FMT(AMPDU).ht[0][ht_idx] += ht 3385 drivers/net/wireless/ath/ath10k/htt_rx.c STATS_OP_FMT(AMPDU).ht[1][ht_idx] += ht 2312 drivers/net/wireless/ath/ath10k/mac.c arg->peer_flags |= ar->wmi.peer_flags->ht; ht 4378 drivers/net/wireless/ath/ath10k/wmi-tlv.c .ht = WMI_TLV_PEER_HT, ht 1632 drivers/net/wireless/ath/ath10k/wmi.c .ht = WMI_PEER_HT, ht 1652 drivers/net/wireless/ath/ath10k/wmi.c .ht = WMI_10X_PEER_HT, ht 1670 drivers/net/wireless/ath/ath10k/wmi.c .ht = WMI_10_2_PEER_HT, ht 6344 drivers/net/wireless/ath/ath10k/wmi.h u32 ht; ht 3866 drivers/net/wireless/ath/ath6kl/cfg80211.c bool band_2gig = false, band_5gig = false, ht = false; ht 3901 drivers/net/wireless/ath/ath6kl/cfg80211.c ht = true; ht 3907 drivers/net/wireless/ath/ath6kl/cfg80211.c ht = true; ht 3913 drivers/net/wireless/ath/ath6kl/cfg80211.c ht = true; ht 3929 drivers/net/wireless/ath/ath6kl/cfg80211.c if (!(ht && ht 3937 drivers/net/wireless/ath/ath6kl/cfg80211.c if (ht) ht 1730 drivers/net/wireless/ath/wcn36xx/hal.h u8 ht; ht 1867 drivers/net/wireless/ath/wcn36xx/hal.h u8 ht; ht 130 drivers/net/wireless/ath/wcn36xx/smd.c bss_params->ht = sta->ht_cap.ht_supported; ht 1318 drivers/net/wireless/ath/wcn36xx/smd.c msg_body->bss_params.ht = orig->bss_params.ht; ht 213 drivers/net/wireless/broadcom/b43/phy_common.h struct b43_phy_ht *ht; ht 198 drivers/net/wireless/broadcom/b43/phy_ht.c struct b43_phy_ht *htphy = dev->phy.ht; ht 337 drivers/net/wireless/broadcom/b43/phy_ht.c struct b43_phy_ht *phy_ht = dev->phy.ht; ht 377 drivers/net/wireless/broadcom/b43/phy_ht.c struct b43_phy_ht *phy_ht = dev->phy.ht; ht 535 drivers/net/wireless/broadcom/b43/phy_ht.c struct b43_phy_ht *phy_ht = dev->phy.ht; ht 575 drivers/net/wireless/broadcom/b43/phy_ht.c struct b43_phy_ht *phy_ht = dev->phy.ht; ht 623 drivers/net/wireless/broadcom/b43/phy_ht.c struct b43_phy_ht *phy_ht = dev->phy.ht; ht 848 drivers/net/wireless/broadcom/b43/phy_ht.c dev->phy.ht = phy_ht; ht 856 drivers/net/wireless/broadcom/b43/phy_ht.c struct b43_phy_ht *phy_ht = phy->ht; ht 871 drivers/net/wireless/broadcom/b43/phy_ht.c struct b43_phy_ht *phy_ht = dev->phy.ht; ht 1015 drivers/net/wireless/broadcom/b43/phy_ht.c struct b43_phy_ht *phy_ht = phy->ht; ht 1018 drivers/net/wireless/broadcom/b43/phy_ht.c phy->ht = NULL; ht 56 drivers/net/wireless/intel/iwlegacy/4965-debug.c struct stats_rx_ht_phy *ht, *accum_ht, *delta_ht, *max_ht; ht 75 drivers/net/wireless/intel/iwlegacy/4965-debug.c ht = &il->_4965.stats.rx.ofdm_ht; ht 407 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(ht->plcp_err), accum_ht->plcp_err, ht 411 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(ht->overrun_err), accum_ht->overrun_err, ht 415 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(ht->early_overrun_err), ht 420 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(ht->crc32_good), accum_ht->crc32_good, ht 424 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(ht->crc32_err), accum_ht->crc32_err, ht 428 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(ht->mh_format_err), accum_ht->mh_format_err, ht 432 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(ht->agg_crc32_good), accum_ht->agg_crc32_good, ht 436 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(ht->agg_mpdu_cnt), accum_ht->agg_mpdu_cnt, ht 440 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(ht->agg_cnt), accum_ht->agg_cnt, ht 444 drivers/net/wireless/intel/iwlegacy/4965-debug.c le32_to_cpu(ht->unsupport_mcs), accum_ht->unsupport_mcs, ht 6099 drivers/net/wireless/intel/iwlegacy/4965-mac.c il->ht.is_40mhz = false; ht 6100 drivers/net/wireless/intel/iwlegacy/4965-mac.c il->ht.extension_chan_offset = IEEE80211_HT_PARAM_CHA_SEC_NONE; ht 6103 drivers/net/wireless/intel/iwlegacy/4965-mac.c il->ht.extension_chan_offset = IEEE80211_HT_PARAM_CHA_SEC_BELOW; ht 6104 drivers/net/wireless/intel/iwlegacy/4965-mac.c il->ht.is_40mhz = true; ht 6107 drivers/net/wireless/intel/iwlegacy/4965-mac.c il->ht.extension_chan_offset = IEEE80211_HT_PARAM_CHA_SEC_ABOVE; ht 6108 drivers/net/wireless/intel/iwlegacy/4965-mac.c il->ht.is_40mhz = true; ht 631 drivers/net/wireless/intel/iwlegacy/4965-rs.c !il->ht.non_gf_sta_present; ht 2375 drivers/net/wireless/intel/iwlegacy/common.c if (il->ht.enabled) ht 3576 drivers/net/wireless/intel/iwlegacy/common.c if (!il->ht.enabled || !il->ht.is_40mhz) ht 3593 drivers/net/wireless/intel/iwlegacy/common.c il->ht.extension_chan_offset); ht 3847 drivers/net/wireless/intel/iwlegacy/common.c if (!il->ht.enabled) { ht 3856 drivers/net/wireless/intel/iwlegacy/common.c cpu_to_le32(il->ht.protection << RXON_FLG_HT_OPERATING_MODE_POS); ht 3865 drivers/net/wireless/intel/iwlegacy/common.c if (il->ht.protection == IEEE80211_HT_OP_MODE_PROTECTION_20MHZ) { ht 3868 drivers/net/wireless/intel/iwlegacy/common.c switch (il->ht.extension_chan_offset) { ht 3879 drivers/net/wireless/intel/iwlegacy/common.c switch (il->ht.extension_chan_offset) { ht 3905 drivers/net/wireless/intel/iwlegacy/common.c il->ht.protection, il->ht.extension_chan_offset); ht 5004 drivers/net/wireless/intel/iwlegacy/common.c if (il->ht.enabled) ht 5082 drivers/net/wireless/intel/iwlegacy/common.c if (il->ht.enabled != conf_is_ht(conf)) { ht 5083 drivers/net/wireless/intel/iwlegacy/common.c il->ht.enabled = conf_is_ht(conf); ht 5086 drivers/net/wireless/intel/iwlegacy/common.c if (il->ht.enabled) { ht 5088 drivers/net/wireless/intel/iwlegacy/common.c il->ht.extension_chan_offset = ht 5090 drivers/net/wireless/intel/iwlegacy/common.c il->ht.is_40mhz = true; ht 5092 drivers/net/wireless/intel/iwlegacy/common.c il->ht.extension_chan_offset = ht 5094 drivers/net/wireless/intel/iwlegacy/common.c il->ht.is_40mhz = true; ht 5096 drivers/net/wireless/intel/iwlegacy/common.c il->ht.extension_chan_offset = ht 5098 drivers/net/wireless/intel/iwlegacy/common.c il->ht.is_40mhz = false; ht 5101 drivers/net/wireless/intel/iwlegacy/common.c il->ht.is_40mhz = false; ht 5107 drivers/net/wireless/intel/iwlegacy/common.c il->ht.protection = IEEE80211_HT_OP_MODE_PROTECTION_NONE; ht 5219 drivers/net/wireless/intel/iwlegacy/common.c if (!il->ht.enabled) ht 5222 drivers/net/wireless/intel/iwlegacy/common.c il->ht.protection = ht 5224 drivers/net/wireless/intel/iwlegacy/common.c il->ht.non_gf_sta_present = ht 1219 drivers/net/wireless/intel/iwlegacy/common.h } ht; ht 685 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c struct statistics_rx_ht_phy *ht, *accum_ht, *delta_ht, *max_ht; ht 703 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c ht = &priv->statistics.rx_ofdm_ht; ht 1051 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(ht->plcp_err), accum_ht->plcp_err, ht 1055 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(ht->overrun_err), accum_ht->overrun_err, ht 1059 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(ht->early_overrun_err), ht 1065 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(ht->crc32_good), accum_ht->crc32_good, ht 1069 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(ht->crc32_err), accum_ht->crc32_err, ht 1073 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(ht->mh_format_err), ht 1078 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(ht->agg_crc32_good), ht 1083 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(ht->agg_mpdu_cnt), ht 1088 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(ht->agg_cnt), accum_ht->agg_cnt, ht 1092 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c le32_to_cpu(ht->unsupport_mcs), ht 522 drivers/net/wireless/intel/iwlwifi/dvm/dev.h } ht; ht 979 drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c ctx->ht.is_40mhz = false; ht 980 drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c ctx->ht.extension_chan_offset = IEEE80211_HT_PARAM_CHA_SEC_NONE; ht 983 drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c ctx->ht.extension_chan_offset = IEEE80211_HT_PARAM_CHA_SEC_BELOW; ht 984 drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c ctx->ht.is_40mhz = true; ht 987 drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c ctx->ht.extension_chan_offset = IEEE80211_HT_PARAM_CHA_SEC_ABOVE; ht 988 drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c ctx->ht.is_40mhz = true; ht 171 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c if (ctx->ht.enabled) ht 608 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c if (!ctx->ht.enabled) { ht 619 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c rxon->flags |= cpu_to_le32(ctx->ht.protection << ht 629 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c if (ctx->ht.protection == ht 636 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c switch (ctx->ht.extension_chan_offset) { ht 651 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c switch (ctx->ht.extension_chan_offset) { ht 680 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c le32_to_cpu(rxon->flags), ctx->ht.protection, ht 681 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c ctx->ht.extension_chan_offset); ht 1143 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c ctx->ht.extension_chan_offset = ht 1145 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c ctx->ht.is_40mhz = true; ht 1147 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c ctx->ht.extension_chan_offset = ht 1149 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c ctx->ht.is_40mhz = true; ht 1151 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c ctx->ht.extension_chan_offset = ht 1153 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c ctx->ht.is_40mhz = false; ht 1198 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c if (ctx->ht.enabled != conf_is_ht(conf)) ht 1199 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c ctx->ht.enabled = conf_is_ht(conf); ht 1201 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c if (ctx->ht.enabled) { ht 1204 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c if (!ctx->ht.is_40mhz || ht 1208 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c ctx->ht.is_40mhz = false; ht 1214 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c ctx->ht.protection = IEEE80211_HT_OP_MODE_PROTECTION_NONE; ht 1448 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c if (ctx->ht.enabled) { ht 1449 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c ctx->ht.protection = bss_conf->ht_operation_mode & ht 1451 drivers/net/wireless/intel/iwlwifi/dvm/rxon.c ctx->ht.non_gf_sta_present = !!(bss_conf->ht_operation_mode & ht 135 drivers/net/wireless/intel/iwlwifi/dvm/sta.c if (!ctx->ht.enabled || !ctx->ht.is_40mhz) ht 793 drivers/net/wireless/intel/iwlwifi/dvm/sta.c if (ctx->ht.enabled) ht 927 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c struct mvm_statistics_rx_ht_phy_v1 *ht = ht 930 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c PRINT_STATS_LE32(ht, plcp_err); ht 931 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c PRINT_STATS_LE32(ht, overrun_err); ht 932 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c PRINT_STATS_LE32(ht, early_overrun_err); ht 933 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c PRINT_STATS_LE32(ht, crc32_good); ht 934 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c PRINT_STATS_LE32(ht, crc32_err); ht 935 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c PRINT_STATS_LE32(ht, mh_format_err); ht 936 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c PRINT_STATS_LE32(ht, agg_crc32_good); ht 937 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c PRINT_STATS_LE32(ht, agg_mpdu_cnt); ht 938 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c PRINT_STATS_LE32(ht, agg_cnt); ht 939 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c PRINT_STATS_LE32(ht, unsupport_mcs); ht 941 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c struct mvm_statistics_rx_ht_phy *ht = ht 944 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c PRINT_STATS_LE32(ht, mh_format_err); ht 945 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c PRINT_STATS_LE32(ht, agg_mpdu_cnt); ht 946 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c PRINT_STATS_LE32(ht, agg_cnt); ht 947 drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c PRINT_STATS_LE32(ht, unsupport_mcs); ht 71 drivers/net/wireless/mediatek/mt76/debugfs.c mt76_seq_puts_array(s, "HT", dev->rate_power.ht, ht 72 drivers/net/wireless/mediatek/mt76/debugfs.c ARRAY_SIZE(dev->rate_power.ht)); ht 338 drivers/net/wireless/mediatek/mt76/mt76.h s8 ht[16]; ht 1369 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c wtbl_ht->ht = 1; ht 259 drivers/net/wireless/mediatek/mt76/mt7615/mcu.h u8 ht; ht 182 drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c t->ht[0] = t->ht[1] = t->vht[0] = t->vht[1] = s6_to_s8(val); ht 183 drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c t->ht[2] = t->ht[3] = t->vht[2] = t->vht[3] = s6_to_s8(val >> 8); ht 188 drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c t->ht[4] = t->ht[5] = t->vht[4] = t->vht[5] = s6_to_s8(val); ht 189 drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c t->ht[6] = t->ht[7] = t->vht[6] = t->vht[7] = s6_to_s8(val >> 8); ht 657 drivers/net/wireless/mediatek/mt76/mt76x0/phy.c *target_power = cur_power + dev->mt76.rate_power.ht[tx_rate]; ht 84 drivers/net/wireless/mediatek/mt76/mt76x0/phy.h struct mt76x0_rate_pwr_item ht[8]; ht 104 drivers/net/wireless/mediatek/mt76/mt76x02_phy.c mt76x02_tx_power_mask(t->ofdm[4], t->ofdm[6], t->ht[0], ht 105 drivers/net/wireless/mediatek/mt76/mt76x02_phy.c t->ht[2])); ht 107 drivers/net/wireless/mediatek/mt76/mt76x02_phy.c mt76x02_tx_power_mask(t->ht[4], t->ht[6], t->ht[8], ht 108 drivers/net/wireless/mediatek/mt76/mt76x02_phy.c t->ht[10])); ht 110 drivers/net/wireless/mediatek/mt76/mt76x02_phy.c mt76x02_tx_power_mask(t->ht[12], t->ht[14], t->stbc[0], ht 115 drivers/net/wireless/mediatek/mt76/mt76x02_phy.c mt76x02_tx_power_mask(t->ofdm[7], t->vht[8], t->ht[7], ht 118 drivers/net/wireless/mediatek/mt76/mt76x02_phy.c mt76x02_tx_power_mask(t->ht[14], 0, t->vht[8], t->vht[9])); ht 120 drivers/net/wireless/mediatek/mt76/mt76x02_phy.c mt76x02_tx_power_mask(t->ht[7], 0, t->stbc[8], t->stbc[9])); ht 72 drivers/net/wireless/mediatek/mt76/mt76x02_txrx.c max_txpwr = dev->mt76.rate_power.ht[idx & 0xf]; ht 75 drivers/net/wireless/mediatek/mt76/mt76x02_txrx.c max_txpwr = dev->mt76.rate_power.ht[rate->idx & 0xf]; ht 307 drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c t->ht[0] = t->ht[1] = mt76x02_rate_power_val(val); ht 308 drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c t->ht[2] = t->ht[3] = mt76x02_rate_power_val(val >> 8); ht 311 drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c t->ht[4] = t->ht[5] = mt76x02_rate_power_val(val); ht 312 drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c t->ht[6] = t->ht[7] = mt76x02_rate_power_val(val >> 8); ht 315 drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c t->ht[8] = t->ht[9] = mt76x02_rate_power_val(val); ht 316 drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c t->ht[10] = t->ht[11] = mt76x02_rate_power_val(val >> 8); ht 319 drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c t->ht[12] = t->ht[13] = mt76x02_rate_power_val(val); ht 320 drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c t->ht[14] = t->ht[15] = mt76x02_rate_power_val(val >> 8); ht 335 drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c memcpy(t->stbc, t->ht, sizeof(t->stbc[0]) * 8); ht 114 drivers/net/wireless/mediatek/mt7601u/debugfs.c rp->ht[i].raw, rp->ht[i].bw20, rp->ht[i].bw40); ht 281 drivers/net/wireless/mediatek/mt7601u/eeprom.c mt7601u_set_power_rate(&t->ht[0], delta, (val >> 16) & 0xff); ht 282 drivers/net/wireless/mediatek/mt7601u/eeprom.c mt7601u_set_power_rate(&t->ht[1], delta, (val >> 24) & 0xff); ht 285 drivers/net/wireless/mediatek/mt7601u/eeprom.c mt7601u_set_power_rate(&t->ht[2], delta, (val >> 0) & 0xff); ht 286 drivers/net/wireless/mediatek/mt7601u/eeprom.c mt7601u_set_power_rate(&t->ht[3], delta, (val >> 8) & 0xff); ht 85 drivers/net/wireless/mediatek/mt7601u/eeprom.h struct power_per_rate ht[4]; ht 833 drivers/net/wireless/mediatek/mt7601u/phy.c rate_table = dev->ee->power_rate_table.ht; ht 794 drivers/net/wireless/ralink/rt2x00/rt2800lib.c txdesc->u.ht.mpdu_density); ht 795 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2x00_set_field32(&word, TXWI_W0_TX_OP, txdesc->u.ht.txop); ht 796 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2x00_set_field32(&word, TXWI_W0_MCS, txdesc->u.ht.mcs); ht 801 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2x00_set_field32(&word, TXWI_W0_STBC, txdesc->u.ht.stbc); ht 810 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2x00_set_field32(&word, TXWI_W1_BW_WIN_SIZE, txdesc->u.ht.ba_size); ht 813 drivers/net/wireless/ralink/rt2x00/rt2800lib.c txdesc->key_idx : txdesc->u.ht.wcid); ht 10078 drivers/net/wireless/ralink/rt2x00/rt2800lib.c spec->ht.ht_supported = true; ht 10080 drivers/net/wireless/ralink/rt2x00/rt2800lib.c spec->ht.ht_supported = false; ht 10082 drivers/net/wireless/ralink/rt2x00/rt2800lib.c spec->ht.cap = ht 10092 drivers/net/wireless/ralink/rt2x00/rt2800lib.c spec->ht.cap |= IEEE80211_HT_CAP_TX_STBC; ht 10094 drivers/net/wireless/ralink/rt2x00/rt2800lib.c spec->ht.cap |= rx_chains << IEEE80211_HT_CAP_RX_STBC_SHIFT; ht 10096 drivers/net/wireless/ralink/rt2x00/rt2800lib.c spec->ht.ampdu_factor = (rx_chains > 1) ? 3 : 2; ht 10097 drivers/net/wireless/ralink/rt2x00/rt2800lib.c spec->ht.ampdu_density = 4; ht 10098 drivers/net/wireless/ralink/rt2x00/rt2800lib.c spec->ht.mcs.tx_params = IEEE80211_HT_MCS_TX_DEFINED; ht 10100 drivers/net/wireless/ralink/rt2x00/rt2800lib.c spec->ht.mcs.tx_params |= IEEE80211_HT_MCS_TX_RX_DIFF; ht 10101 drivers/net/wireless/ralink/rt2x00/rt2800lib.c spec->ht.mcs.tx_params |= ht 10107 drivers/net/wireless/ralink/rt2x00/rt2800lib.c spec->ht.mcs.rx_mask[2] = 0xff; ht 10110 drivers/net/wireless/ralink/rt2x00/rt2800lib.c spec->ht.mcs.rx_mask[1] = 0xff; ht 10113 drivers/net/wireless/ralink/rt2x00/rt2800lib.c spec->ht.mcs.rx_mask[0] = 0xff; ht 10114 drivers/net/wireless/ralink/rt2x00/rt2800lib.c spec->ht.mcs.rx_mask[4] = 0x1; /* MCS32 */ ht 410 drivers/net/wireless/ralink/rt2x00/rt2x00.h struct ieee80211_sta_ht_cap ht; ht 1057 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c &spec->ht, sizeof(spec->ht)); ht 1075 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c &spec->ht, sizeof(spec->ht)); ht 305 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c txdesc->u.ht.wcid = sta_priv->wcid; ht 314 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c txdesc->u.ht.mcs = txrate->idx; ht 320 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c if (sta && txdesc->u.ht.mcs > 7 && ht 324 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c txdesc->u.ht.mcs = rt2x00_get_rate_mcs(hwrate->mcs); ht 326 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c txdesc->u.ht.mcs |= 0x08; ht 331 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c txdesc->u.ht.txop = TXOP_SIFS; ht 333 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c txdesc->u.ht.txop = TXOP_BACKOFF; ht 343 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c txdesc->u.ht.stbc = 1; ht 352 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c txdesc->u.ht.mpdu_density = density; ht 353 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c txdesc->u.ht.ba_size = 7; /* FIXME: What value is needed? */ ht 377 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c txdesc->u.ht.txop = TXOP_BACKOFF; ht 379 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c txdesc->u.ht.txop = TXOP_SIFS; ht 381 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c txdesc->u.ht.txop = TXOP_HTTXOP; ht 315 drivers/net/wireless/ralink/rt2x00/rt2x00queue.h } ht; ht 1137 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c bool ht = true; ht 1147 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c ht = false; ht 1234 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c if (ht) ht 1261 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c bool ht = true; ht 1275 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c ht = false; ht 1357 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c if (ht) ht 283 drivers/net/wireless/ti/wl12xx/main.c .ht = { ht 120 drivers/net/wireless/ti/wl18xx/conf.h struct wl18xx_ht_settings ht; ht 426 drivers/net/wireless/ti/wl18xx/main.c .ht = { ht 509 drivers/net/wireless/ti/wl18xx/main.c .ht = { ht 1263 drivers/net/wireless/ti/wl18xx/main.c (priv->conf.ht.mode != HT_MODE_WIDE) && ht 1264 drivers/net/wireless/ti/wl18xx/main.c (priv->conf.ht.mode != HT_MODE_SISO20); ht 1949 drivers/net/wireless/ti/wl18xx/main.c priv->conf.ht.mode = HT_MODE_DEFAULT; ht 1951 drivers/net/wireless/ti/wl18xx/main.c priv->conf.ht.mode = HT_MODE_WIDE; ht 1953 drivers/net/wireless/ti/wl18xx/main.c priv->conf.ht.mode = HT_MODE_SISO20; ht 1960 drivers/net/wireless/ti/wl18xx/main.c if (priv->conf.ht.mode == HT_MODE_DEFAULT) { ht 1975 drivers/net/wireless/ti/wl18xx/main.c } else if (priv->conf.ht.mode == HT_MODE_WIDE) { ht 1980 drivers/net/wireless/ti/wl18xx/main.c } else if (priv->conf.ht.mode == HT_MODE_SISO20) { ht 1387 drivers/net/wireless/ti/wlcore/acx.c acx->tid_bitmap = wl->conf.ht.tx_ba_tid_bitmap; ht 1388 drivers/net/wireless/ti/wlcore/acx.c acx->win_size = wl->conf.ht.tx_ba_win_size; ht 1389 drivers/net/wireless/ti/wlcore/acx.c acx->inactivity_timeout = wl->conf.ht.inactivity_timeout; ht 1140 drivers/net/wireless/ti/wlcore/conf.h struct conf_ht_setting ht; ht 6346 drivers/net/wireless/ti/wlcore/main.c wl->hw->max_rx_aggregation_subframes = wl->conf.ht.rx_ba_win_size; ht 2020 drivers/staging/rtl8192e/rtllib_rx.c struct bss_ht *ht = &network->bssht; ht 2024 drivers/staging/rtl8192e/rtllib_rx.c ht->bdHTSpecVer = HT_SPEC_VER_EWC; ht 2025 drivers/staging/rtl8192e/rtllib_rx.c ht->bdHTCapLen = min_t(u16, *tmp_htcap_len, ht 2026 drivers/staging/rtl8192e/rtllib_rx.c sizeof(ht->bdHTCapBuf)); ht 2027 drivers/staging/rtl8192e/rtllib_rx.c memcpy(ht->bdHTCapBuf, info_element->data, ht->bdHTCapLen); ht 2029 drivers/staging/rtl8192e/rtllib_rx.c ht->bdSupportHT = true; ht 2030 drivers/staging/rtl8192e/rtllib_rx.c ht->bdHT1R = ((((struct ht_capab_ele *) ht 2031 drivers/staging/rtl8192e/rtllib_rx.c ht->bdHTCapBuf))->MCS[1]) == 0; ht 2033 drivers/staging/rtl8192e/rtllib_rx.c ht->bdBandWidth = (enum ht_channel_width) ht 2035 drivers/staging/rtl8192e/rtllib_rx.c (ht->bdHTCapBuf))->ChlWidth); ht 2037 drivers/staging/rtl8192e/rtllib_rx.c ht->bdSupportHT = false; ht 2038 drivers/staging/rtl8192e/rtllib_rx.c ht->bdHT1R = false; ht 2039 drivers/staging/rtl8192e/rtllib_rx.c ht->bdBandWidth = HT_CHANNEL_WIDTH_20; ht 1556 drivers/staging/speakup/main.c speakup_console[vc->vc_num]->ht.cy = vc->vc_y; ht 1572 drivers/staging/speakup/main.c hi = speakup_console[vc_num]->ht.highsize[bi]; ht 1575 drivers/staging/speakup/main.c if (speakup_console[vc_num]->ht.highsize[bi] == 0) { ht 1576 drivers/staging/speakup/main.c speakup_console[vc_num]->ht.rpos[bi] = vc->vc_pos; ht 1577 drivers/staging/speakup/main.c speakup_console[vc_num]->ht.rx[bi] = vc->vc_x; ht 1578 drivers/staging/speakup/main.c speakup_console[vc_num]->ht.ry[bi] = vc->vc_y; ht 1582 drivers/staging/speakup/main.c speakup_console[vc_num]->ht.highbuf[bi][hi] = ic[i]; ht 1585 drivers/staging/speakup/main.c if (speakup_console[vc_num]->ht.highbuf[bi][hi - 1] != ht 1587 drivers/staging/speakup/main.c speakup_console[vc_num]->ht.highbuf[bi][hi] = ht 1594 drivers/staging/speakup/main.c speakup_console[vc_num]->ht.highsize[bi] = hi; ht 1603 drivers/staging/speakup/main.c speakup_console[vc_num]->ht.highsize[i] = 0; ht 1615 drivers/staging/speakup/main.c speakup_console[vc_num]->ht.bgcount[i] = 0; ht 1624 drivers/staging/speakup/main.c speakup_console[vc_num]->ht.bgcount[bg]++; ht 1631 drivers/staging/speakup/main.c if (speakup_console[vc_num]->ht.bgcount[i] > 0) ht 1647 drivers/staging/speakup/main.c if (speakup_console[vc_num]->ht.bgcount[cptr[i]] > ht 1648 drivers/staging/speakup/main.c speakup_console[vc_num]->ht.bgcount[cptr[j]]) ht 1652 drivers/staging/speakup/main.c if (speakup_console[vc_num]->ht.bgcount[cptr[i]] != 0) ht 1653 drivers/staging/speakup/main.c if (speakup_console[vc_num]->ht.highsize[cptr[i]] > 0) ht 1667 drivers/staging/speakup/main.c d = vc->vc_y - speakup_console[vc_num]->ht.cy; ht 1669 drivers/staging/speakup/main.c if (speakup_console[vc_num]->ht.ry[hc] != vc->vc_y) ht 1673 drivers/staging/speakup/main.c spkup_write(speakup_console[vc_num]->ht.highbuf[hc], ht 1674 drivers/staging/speakup/main.c speakup_console[vc_num]->ht.highsize[hc]); ht 1675 drivers/staging/speakup/main.c spk_pos = spk_cp = speakup_console[vc_num]->ht.rpos[hc]; ht 1676 drivers/staging/speakup/main.c spk_x = spk_cx = speakup_console[vc_num]->ht.rx[hc]; ht 1677 drivers/staging/speakup/main.c spk_y = spk_cy = speakup_console[vc_num]->ht.ry[hc]; ht 77 drivers/staging/speakup/spk_types.h struct spk_highlight_color_track ht; ht 179 drivers/video/fbdev/asiliantfb.c unsigned ht = (p->var.left_margin + p->var.xres + p->var.right_margin + p->var.hsync_len) / 8; ht 193 drivers/video/fbdev/asiliantfb.c write_cr(0x00, (ht - 5) & 0xff); ht 196 drivers/video/fbdev/asiliantfb.c write_cr(0x03, ((ht - 1) & 0x1f) | 0x80); ht 198 drivers/video/fbdev/asiliantfb.c write_cr(0x05, (((ht - 1) & 0x20) <<2) | (he & 0x1f)); ht 199 drivers/video/fbdev/asiliantfb.c write_cr(0x3c, (ht - 1) & 0xc0); ht 214 drivers/video/fbdev/asiliantfb.c write_cr(0x38, ((ht - 5) & 0x100) >> 8); ht 308 drivers/video/fbdev/imsttfb.c __u16 hes, heb, hsb, ht, ves, veb, vsb, vt, vil; ht 506 drivers/video/fbdev/imsttfb.c init->ht = init->hsb + htp; ht 704 drivers/video/fbdev/imsttfb.c write_reg_le32(par->dc_regs, HT, init->ht); ht 1050 drivers/video/fbdev/intelfb/intelfbhw.c u32 *vs, *vb, *vt, *hs, *hb, *ht, *ss, *pipe_conf; ht 1065 drivers/video/fbdev/intelfb/intelfbhw.c ht = &hw->htotal_b; ht 1077 drivers/video/fbdev/intelfb/intelfbhw.c ht = &hw->htotal_a; ht 1236 drivers/video/fbdev/intelfb/intelfbhw.c *ht = (htotal << HTOTAL_SHIFT) | (hactive << HACTIVE_SHIFT); ht 1282 drivers/video/fbdev/intelfb/intelfbhw.c const u32 *hs, *ht, *hb, *vs, *vt, *vb, *ss; ht 1309 drivers/video/fbdev/intelfb/intelfbhw.c ht = &hw->htotal_b; ht 1333 drivers/video/fbdev/intelfb/intelfbhw.c ht = &hw->htotal_a; ht 1423 drivers/video/fbdev/intelfb/intelfbhw.c OUTREG(htotal_reg, *ht); ht 723 drivers/video/fbdev/matrox/matroxfb_maven.c static int maven_find_exact_clocks(unsigned int ht, unsigned int vt, ht 737 drivers/video/fbdev/matrox/matroxfb_maven.c unsigned int h = ht + 2 + x; ht 192 drivers/video/fbdev/matrox/matroxfb_misc.c unsigned int hd, hs, he, hbe, ht; ht 243 drivers/video/fbdev/matrox/matroxfb_misc.c ht = m->HTotal >> 3; ht 252 drivers/video/fbdev/matrox/matroxfb_misc.c ht >>= 1; ht 261 drivers/video/fbdev/matrox/matroxfb_misc.c ht <<= 1; ht 267 drivers/video/fbdev/matrox/matroxfb_misc.c ht = ht - 1; ht 274 drivers/video/fbdev/matrox/matroxfb_misc.c if (((ht & 0x07) == 0x06) || ((ht & 0x0F) == 0x04)) ht 275 drivers/video/fbdev/matrox/matroxfb_misc.c ht++; ht 276 drivers/video/fbdev/matrox/matroxfb_misc.c hbe = ht; ht 283 drivers/video/fbdev/matrox/matroxfb_misc.c hw->CRTCEXT[5] = (hs + he - ht) >> 1; ht 289 drivers/video/fbdev/matrox/matroxfb_misc.c hw->CRTCEXT[1] = (((ht - 4) & 0x100) >> 8) | ht 304 drivers/video/fbdev/matrox/matroxfb_misc.c hw->CRTC[0] = ht-4; ht 445 drivers/video/fbdev/mb862xx/mb862xxfbdrv.c unsigned long hsp, vsp, ht, vt; ht 469 drivers/video/fbdev/mb862xx/mb862xxfbdrv.c ht = ((inreg(disp, GC_HTP) & 0xfff0000) >> 16) + 1; ht 471 drivers/video/fbdev/mb862xx/mb862xxfbdrv.c fbi->var.left_margin = ht - hsp - fbi->var.hsync_len; ht 237 drivers/video/fbdev/mbx/mbxfb.c ushort hbps, ht, hfps, has; ht 290 drivers/video/fbdev/mbx/mbxfb.c ht = hfps + var->right_margin; ht 297 drivers/video/fbdev/mbx/mbxfb.c write_reg_dly((Dht01_Hbps(hbps) | Dht01_Ht(ht)), DHT01); ht 3249 drivers/video/fbdev/omap2/omapfb/dss/dispc.c unsigned long ht, vt; ht 3267 drivers/video/fbdev/omap2/omapfb/dss/dispc.c ht = timings->pixelclock / xtot; ht 3277 drivers/video/fbdev/omap2/omapfb/dss/dispc.c DSSDBG("hsync %luHz, vsync %luHz\n", ht, vt); ht 551 drivers/video/fbdev/tdfxfb.c u32 hd, hs, he, ht, hbs, hbe; ht 584 drivers/video/fbdev/tdfxfb.c ht = (htotal >> 3) - 1; ht 586 drivers/video/fbdev/tdfxfb.c hbe = ht; ht 643 drivers/video/fbdev/tdfxfb.c reg.crt[0x00] = ht - 4; ht 668 drivers/video/fbdev/tdfxfb.c reg.ext[0x00] = (((ht & 0x100) >> 8) | ht 1981 fs/gfs2/dir.c char *ht; ht 1991 fs/gfs2/dir.c ht = kzalloc(size, GFP_NOFS | __GFP_NOWARN); ht 1992 fs/gfs2/dir.c if (ht == NULL) ht 1993 fs/gfs2/dir.c ht = __vmalloc(size, GFP_NOFS | __GFP_NOWARN | __GFP_ZERO, ht 1995 fs/gfs2/dir.c if (!ht) ht 2058 fs/gfs2/dir.c error = gfs2_dir_write_data(dip, ht, index * sizeof(u64), size); ht 2085 fs/gfs2/dir.c kvfree(ht); ht 52 include/drm/drm_hashtab.h int drm_ht_create(struct drm_open_hash *ht, unsigned int order); ht 53 include/drm/drm_hashtab.h int drm_ht_insert_item(struct drm_open_hash *ht, struct drm_hash_item *item); ht 54 include/drm/drm_hashtab.h int drm_ht_just_insert_please(struct drm_open_hash *ht, struct drm_hash_item *item, ht 57 include/drm/drm_hashtab.h int drm_ht_find_item(struct drm_open_hash *ht, unsigned long key, struct drm_hash_item **item); ht 59 include/drm/drm_hashtab.h void drm_ht_verbose_list(struct drm_open_hash *ht, unsigned long key); ht 60 include/drm/drm_hashtab.h int drm_ht_remove_key(struct drm_open_hash *ht, unsigned long key); ht 61 include/drm/drm_hashtab.h int drm_ht_remove_item(struct drm_open_hash *ht, struct drm_hash_item *item); ht 62 include/drm/drm_hashtab.h void drm_ht_remove(struct drm_open_hash *ht); ht 134 include/drm/drm_modes.h #define DRM_MODE(nm, t, c, hd, hss, hse, ht, hsk, vd, vss, vse, vt, vs, f) \ ht 137 include/drm/drm_modes.h .htotal = (ht), .hskew = (hsk), .vdisplay = (vd), \ ht 34 include/linux/hashtable.h static inline void __hash_init(struct hlist_head *ht, unsigned int sz) ht 39 include/linux/hashtable.h INIT_HLIST_HEAD(&ht[i]); ht 81 include/linux/hashtable.h static inline bool __hash_empty(struct hlist_head *ht, unsigned int sz) ht 86 include/linux/hashtable.h if (!hlist_empty(&ht[i])) ht 34 include/linux/rhashtable-types.h struct rhashtable *ht; ht 98 include/linux/rhashtable-types.h struct rhashtable ht; ht 121 include/linux/rhashtable-types.h struct rhashtable *ht; ht 130 include/linux/rhashtable-types.h int rhashtable_init(struct rhashtable *ht, ht 113 include/linux/rhashtable.h static inline void *rht_obj(const struct rhashtable *ht, ht 116 include/linux/rhashtable.h return (char *)he - ht->p.head_offset; ht 125 include/linux/rhashtable.h static inline unsigned int rht_key_get_hash(struct rhashtable *ht, ht 133 include/linux/rhashtable.h hash = ht->p.hashfn(key, ht->key_len, hash_rnd); ht 144 include/linux/rhashtable.h unsigned int key_len = ht->p.key_len; ht 156 include/linux/rhashtable.h struct rhashtable *ht, const struct bucket_table *tbl, ht 159 include/linux/rhashtable.h unsigned int hash = rht_key_get_hash(ht, key, params, tbl->hash_rnd); ht 165 include/linux/rhashtable.h struct rhashtable *ht, const struct bucket_table *tbl, ht 168 include/linux/rhashtable.h const char *ptr = rht_obj(ht, he); ht 172 include/linux/rhashtable.h ht->p.key_len, ht 174 include/linux/rhashtable.h rht_key_hashfn(ht, tbl, ptr + params.key_offset, params); ht 182 include/linux/rhashtable.h static inline bool rht_grow_above_75(const struct rhashtable *ht, ht 186 include/linux/rhashtable.h return atomic_read(&ht->nelems) > (tbl->size / 4 * 3) && ht 187 include/linux/rhashtable.h (!ht->p.max_size || tbl->size < ht->p.max_size); ht 195 include/linux/rhashtable.h static inline bool rht_shrink_below_30(const struct rhashtable *ht, ht 199 include/linux/rhashtable.h return atomic_read(&ht->nelems) < (tbl->size * 3 / 10) && ht 200 include/linux/rhashtable.h tbl->size > ht->p.min_size; ht 208 include/linux/rhashtable.h static inline bool rht_grow_above_100(const struct rhashtable *ht, ht 211 include/linux/rhashtable.h return atomic_read(&ht->nelems) > tbl->size && ht 212 include/linux/rhashtable.h (!ht->p.max_size || tbl->size < ht->p.max_size); ht 220 include/linux/rhashtable.h static inline bool rht_grow_above_max(const struct rhashtable *ht, ht 223 include/linux/rhashtable.h return atomic_read(&ht->nelems) >= ht->max_elems; ht 227 include/linux/rhashtable.h int lockdep_rht_mutex_is_held(struct rhashtable *ht); ht 230 include/linux/rhashtable.h static inline int lockdep_rht_mutex_is_held(struct rhashtable *ht) ht 242 include/linux/rhashtable.h void *rhashtable_insert_slow(struct rhashtable *ht, const void *key, ht 245 include/linux/rhashtable.h void rhashtable_walk_enter(struct rhashtable *ht, ht 259 include/linux/rhashtable.h void rhashtable_free_and_destroy(struct rhashtable *ht, ht 262 include/linux/rhashtable.h void rhashtable_destroy(struct rhashtable *ht); ht 268 include/linux/rhashtable.h struct rhash_lock_head **rht_bucket_nested_insert(struct rhashtable *ht, ht 272 include/linux/rhashtable.h #define rht_dereference(p, ht) \ ht 273 include/linux/rhashtable.h rcu_dereference_protected(p, lockdep_rht_mutex_is_held(ht)) ht 275 include/linux/rhashtable.h #define rht_dereference_rcu(p, ht) \ ht 276 include/linux/rhashtable.h rcu_dereference_check(p, lockdep_rht_mutex_is_held(ht)) ht 302 include/linux/rhashtable.h struct rhashtable *ht, struct bucket_table *tbl, unsigned int hash) ht 304 include/linux/rhashtable.h return unlikely(tbl->nest) ? rht_bucket_nested_insert(ht, tbl, hash) : ht 581 include/linux/rhashtable.h struct rhashtable *ht = arg->ht; ht 584 include/linux/rhashtable.h return memcmp(ptr + ht->p.key_offset, arg->key, ht->p.key_len); ht 589 include/linux/rhashtable.h struct rhashtable *ht, const void *key, ht 593 include/linux/rhashtable.h .ht = ht, ht 601 include/linux/rhashtable.h tbl = rht_dereference_rcu(ht->tbl, ht); ht 603 include/linux/rhashtable.h hash = rht_key_hashfn(ht, tbl, key, params); ht 608 include/linux/rhashtable.h params.obj_cmpfn(&arg, rht_obj(ht, he)) : ht 609 include/linux/rhashtable.h rhashtable_compare(&arg, rht_obj(ht, he))) ht 621 include/linux/rhashtable.h tbl = rht_dereference_rcu(tbl->future_tbl, ht); ht 642 include/linux/rhashtable.h struct rhashtable *ht, const void *key, ht 645 include/linux/rhashtable.h struct rhash_head *he = __rhashtable_lookup(ht, key, params); ht 647 include/linux/rhashtable.h return he ? rht_obj(ht, he) : NULL; ht 665 include/linux/rhashtable.h struct rhashtable *ht, const void *key, ht 671 include/linux/rhashtable.h obj = rhashtable_lookup(ht, key, params); ht 695 include/linux/rhashtable.h struct rhash_head *he = __rhashtable_lookup(&hlt->ht, key, params); ht 705 include/linux/rhashtable.h struct rhashtable *ht, const void *key, struct rhash_head *obj, ht 709 include/linux/rhashtable.h .ht = ht, ht 722 include/linux/rhashtable.h tbl = rht_dereference_rcu(ht->tbl, ht); ht 723 include/linux/rhashtable.h hash = rht_head_hashfn(ht, tbl, obj, params); ht 725 include/linux/rhashtable.h bkt = rht_bucket_insert(ht, tbl, hash); ht 736 include/linux/rhashtable.h return rhashtable_insert_slow(ht, key, obj); ht 746 include/linux/rhashtable.h params.obj_cmpfn(&arg, rht_obj(ht, head)) : ht 747 include/linux/rhashtable.h rhashtable_compare(&arg, rht_obj(ht, head)))) { ht 752 include/linux/rhashtable.h data = rht_obj(ht, head); ht 777 include/linux/rhashtable.h if (unlikely(rht_grow_above_max(ht, tbl))) ht 780 include/linux/rhashtable.h if (unlikely(rht_grow_above_100(ht, tbl))) ht 794 include/linux/rhashtable.h atomic_inc(&ht->nelems); ht 797 include/linux/rhashtable.h if (rht_grow_above_75(ht, tbl)) ht 798 include/linux/rhashtable.h schedule_work(&ht->run_work); ht 827 include/linux/rhashtable.h struct rhashtable *ht, struct rhash_head *obj, ht 832 include/linux/rhashtable.h ret = __rhashtable_insert_fast(ht, NULL, obj, params, false); ht 859 include/linux/rhashtable.h return PTR_ERR(__rhashtable_insert_fast(&hlt->ht, key, &list->rhead, ht 882 include/linux/rhashtable.h const char *key = rht_obj(&hlt->ht, &list->rhead); ht 904 include/linux/rhashtable.h struct rhashtable *ht, struct rhash_head *obj, ht 907 include/linux/rhashtable.h const char *key = rht_obj(ht, obj); ht 910 include/linux/rhashtable.h BUG_ON(ht->p.obj_hashfn); ht 912 include/linux/rhashtable.h ret = __rhashtable_insert_fast(ht, key + ht->p.key_offset, obj, params, ht 931 include/linux/rhashtable.h struct rhashtable *ht, struct rhash_head *obj, ht 934 include/linux/rhashtable.h const char *key = rht_obj(ht, obj); ht 936 include/linux/rhashtable.h BUG_ON(ht->p.obj_hashfn); ht 938 include/linux/rhashtable.h return __rhashtable_insert_fast(ht, key + ht->p.key_offset, obj, params, ht 958 include/linux/rhashtable.h struct rhashtable *ht, const void *key, struct rhash_head *obj, ht 963 include/linux/rhashtable.h BUG_ON(!ht->p.obj_hashfn || !key); ht 965 include/linux/rhashtable.h ret = __rhashtable_insert_fast(ht, key, obj, params, false); ht 984 include/linux/rhashtable.h struct rhashtable *ht, const void *key, struct rhash_head *obj, ht 987 include/linux/rhashtable.h BUG_ON(!ht->p.obj_hashfn || !key); ht 989 include/linux/rhashtable.h return __rhashtable_insert_fast(ht, key, obj, params, false); ht 994 include/linux/rhashtable.h struct rhashtable *ht, struct bucket_table *tbl, ht 1004 include/linux/rhashtable.h hash = rht_head_hashfn(ht, tbl, obj, params); ht 1063 include/linux/rhashtable.h atomic_dec(&ht->nelems); ht 1064 include/linux/rhashtable.h if (unlikely(ht->p.automatic_shrinking && ht 1065 include/linux/rhashtable.h rht_shrink_below_30(ht, tbl))) ht 1066 include/linux/rhashtable.h schedule_work(&ht->run_work); ht 1075 include/linux/rhashtable.h struct rhashtable *ht, struct rhash_head *obj, ht 1083 include/linux/rhashtable.h tbl = rht_dereference_rcu(ht->tbl, ht); ht 1090 include/linux/rhashtable.h while ((err = __rhashtable_remove_fast_one(ht, tbl, obj, params, ht 1092 include/linux/rhashtable.h (tbl = rht_dereference_rcu(tbl->future_tbl, ht))) ht 1116 include/linux/rhashtable.h struct rhashtable *ht, struct rhash_head *obj, ht 1119 include/linux/rhashtable.h return __rhashtable_remove_fast(ht, obj, params, false); ht 1141 include/linux/rhashtable.h return __rhashtable_remove_fast(&hlt->ht, &list->rhead, params, true); ht 1146 include/linux/rhashtable.h struct rhashtable *ht, struct bucket_table *tbl, ht 1159 include/linux/rhashtable.h hash = rht_head_hashfn(ht, tbl, obj_old, params); ht 1160 include/linux/rhashtable.h if (hash != rht_head_hashfn(ht, tbl, obj_new, params)) ht 1208 include/linux/rhashtable.h struct rhashtable *ht, struct rhash_head *obj_old, ht 1217 include/linux/rhashtable.h tbl = rht_dereference_rcu(ht->tbl, ht); ht 1224 include/linux/rhashtable.h while ((err = __rhashtable_replace_fast(ht, tbl, obj_old, ht 1226 include/linux/rhashtable.h (tbl = rht_dereference_rcu(tbl->future_tbl, ht))) ht 1258 include/linux/rhashtable.h return rhashtable_walk_enter(&hlt->ht, iter); ht 1274 include/linux/rhashtable.h return rhashtable_free_and_destroy(&hlt->ht, free_fn, arg); ht 37 include/net/raw.h struct hlist_head ht[RAW_HTABLE_SIZE]; ht 18 include/net/smc.h struct hlist_head ht; ht 88 kernel/smpboot.c struct smp_hotplug_thread *ht; ht 110 kernel/smpboot.c struct smp_hotplug_thread *ht = td->ht; ht 119 kernel/smpboot.c if (ht->cleanup && td->status != HP_THREAD_NONE) ht 120 kernel/smpboot.c ht->cleanup(td->cpu, cpu_online(td->cpu)); ht 128 kernel/smpboot.c if (ht->park && td->status == HP_THREAD_ACTIVE) { ht 130 kernel/smpboot.c ht->park(td->cpu); ht 145 kernel/smpboot.c if (ht->setup) ht 146 kernel/smpboot.c ht->setup(td->cpu); ht 153 kernel/smpboot.c if (ht->unpark) ht 154 kernel/smpboot.c ht->unpark(td->cpu); ht 159 kernel/smpboot.c if (!ht->thread_should_run(td->cpu)) { ht 165 kernel/smpboot.c ht->thread_fn(td->cpu); ht 171 kernel/smpboot.c __smpboot_create_thread(struct smp_hotplug_thread *ht, unsigned int cpu) ht 173 kernel/smpboot.c struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); ht 183 kernel/smpboot.c td->ht = ht; ht 186 kernel/smpboot.c ht->thread_comm); ht 197 kernel/smpboot.c *per_cpu_ptr(ht->store, cpu) = tsk; ht 198 kernel/smpboot.c if (ht->create) { ht 208 kernel/smpboot.c ht->create(cpu); ht 228 kernel/smpboot.c static void smpboot_unpark_thread(struct smp_hotplug_thread *ht, unsigned int cpu) ht 230 kernel/smpboot.c struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); ht 232 kernel/smpboot.c if (!ht->selfparking) ht 247 kernel/smpboot.c static void smpboot_park_thread(struct smp_hotplug_thread *ht, unsigned int cpu) ht 249 kernel/smpboot.c struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); ht 251 kernel/smpboot.c if (tsk && !ht->selfparking) ht 266 kernel/smpboot.c static void smpboot_destroy_threads(struct smp_hotplug_thread *ht) ht 272 kernel/smpboot.c struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu); ht 277 kernel/smpboot.c *per_cpu_ptr(ht->store, cpu) = NULL; ht 912 lib/objagg.c struct rhashtable *ht = arg->ht; ht 914 lib/objagg.c container_of(ht, struct objagg_hints, node_ht); ht 918 lib/objagg.c ptr += ht->p.key_offset; ht 920 lib/objagg.c memcmp(ptr, arg->key, ht->p.key_len); ht 37 lib/rhashtable.c static u32 head_hashfn(struct rhashtable *ht, ht 41 lib/rhashtable.c return rht_head_hashfn(ht, tbl, he, ht->p); ht 47 lib/rhashtable.c int lockdep_rht_mutex_is_held(struct rhashtable *ht) ht 49 lib/rhashtable.c return (debug_locks) ? lockdep_is_held(&ht->mutex) : 1; ht 113 lib/rhashtable.c static union nested_table *nested_table_alloc(struct rhashtable *ht, ht 138 lib/rhashtable.c static struct bucket_table *nested_bucket_table_alloc(struct rhashtable *ht, ht 155 lib/rhashtable.c if (!nested_table_alloc(ht, (union nested_table __rcu **)tbl->buckets, ht 166 lib/rhashtable.c static struct bucket_table *bucket_table_alloc(struct rhashtable *ht, ht 180 lib/rhashtable.c tbl = nested_bucket_table_alloc(ht, nbuckets, gfp); ht 202 lib/rhashtable.c static struct bucket_table *rhashtable_last_table(struct rhashtable *ht, ht 209 lib/rhashtable.c tbl = rht_dereference_rcu(tbl->future_tbl, ht); ht 215 lib/rhashtable.c static int rhashtable_rehash_one(struct rhashtable *ht, ht 219 lib/rhashtable.c struct bucket_table *old_tbl = rht_dereference(ht->tbl, ht); ht 220 lib/rhashtable.c struct bucket_table *new_tbl = rhashtable_last_table(ht, old_tbl); ht 245 lib/rhashtable.c new_hash = head_hashfn(ht, new_tbl, entry); ht 265 lib/rhashtable.c static int rhashtable_rehash_chain(struct rhashtable *ht, ht 268 lib/rhashtable.c struct bucket_table *old_tbl = rht_dereference(ht->tbl, ht); ht 276 lib/rhashtable.c while (!(err = rhashtable_rehash_one(ht, bkt, old_hash))) ht 286 lib/rhashtable.c static int rhashtable_rehash_attach(struct rhashtable *ht, ht 303 lib/rhashtable.c static int rhashtable_rehash_table(struct rhashtable *ht) ht 305 lib/rhashtable.c struct bucket_table *old_tbl = rht_dereference(ht->tbl, ht); ht 311 lib/rhashtable.c new_tbl = rht_dereference(old_tbl->future_tbl, ht); ht 316 lib/rhashtable.c err = rhashtable_rehash_chain(ht, old_hash); ht 323 lib/rhashtable.c rcu_assign_pointer(ht->tbl, new_tbl); ht 325 lib/rhashtable.c spin_lock(&ht->lock); ht 337 lib/rhashtable.c spin_unlock(&ht->lock); ht 339 lib/rhashtable.c return rht_dereference(new_tbl->future_tbl, ht) ? -EAGAIN : 0; ht 342 lib/rhashtable.c static int rhashtable_rehash_alloc(struct rhashtable *ht, ht 349 lib/rhashtable.c ASSERT_RHT_MUTEX(ht); ht 351 lib/rhashtable.c new_tbl = bucket_table_alloc(ht, size, GFP_KERNEL); ht 355 lib/rhashtable.c err = rhashtable_rehash_attach(ht, old_tbl, new_tbl); ht 378 lib/rhashtable.c static int rhashtable_shrink(struct rhashtable *ht) ht 380 lib/rhashtable.c struct bucket_table *old_tbl = rht_dereference(ht->tbl, ht); ht 381 lib/rhashtable.c unsigned int nelems = atomic_read(&ht->nelems); ht 386 lib/rhashtable.c if (size < ht->p.min_size) ht 387 lib/rhashtable.c size = ht->p.min_size; ht 392 lib/rhashtable.c if (rht_dereference(old_tbl->future_tbl, ht)) ht 395 lib/rhashtable.c return rhashtable_rehash_alloc(ht, old_tbl, size); ht 400 lib/rhashtable.c struct rhashtable *ht; ht 404 lib/rhashtable.c ht = container_of(work, struct rhashtable, run_work); ht 405 lib/rhashtable.c mutex_lock(&ht->mutex); ht 407 lib/rhashtable.c tbl = rht_dereference(ht->tbl, ht); ht 408 lib/rhashtable.c tbl = rhashtable_last_table(ht, tbl); ht 410 lib/rhashtable.c if (rht_grow_above_75(ht, tbl)) ht 411 lib/rhashtable.c err = rhashtable_rehash_alloc(ht, tbl, tbl->size * 2); ht 412 lib/rhashtable.c else if (ht->p.automatic_shrinking && rht_shrink_below_30(ht, tbl)) ht 413 lib/rhashtable.c err = rhashtable_shrink(ht); ht 415 lib/rhashtable.c err = rhashtable_rehash_alloc(ht, tbl, tbl->size); ht 420 lib/rhashtable.c nerr = rhashtable_rehash_table(ht); ht 424 lib/rhashtable.c mutex_unlock(&ht->mutex); ht 427 lib/rhashtable.c schedule_work(&ht->run_work); ht 430 lib/rhashtable.c static int rhashtable_insert_rehash(struct rhashtable *ht, ht 438 lib/rhashtable.c old_tbl = rht_dereference_rcu(ht->tbl, ht); ht 444 lib/rhashtable.c if (rht_grow_above_75(ht, tbl)) ht 452 lib/rhashtable.c new_tbl = bucket_table_alloc(ht, size, GFP_ATOMIC | __GFP_NOWARN); ht 456 lib/rhashtable.c err = rhashtable_rehash_attach(ht, tbl, new_tbl); ht 462 lib/rhashtable.c schedule_work(&ht->run_work); ht 473 lib/rhashtable.c schedule_work(&ht->run_work); ht 478 lib/rhashtable.c static void *rhashtable_lookup_one(struct rhashtable *ht, ht 484 lib/rhashtable.c .ht = ht, ht 498 lib/rhashtable.c (ht->p.obj_cmpfn ? ht 499 lib/rhashtable.c ht->p.obj_cmpfn(&arg, rht_obj(ht, head)) : ht 500 lib/rhashtable.c rhashtable_compare(&arg, rht_obj(ht, head)))) { ht 505 lib/rhashtable.c if (!ht->rhlist) ht 506 lib/rhashtable.c return rht_obj(ht, head); ht 529 lib/rhashtable.c static struct bucket_table *rhashtable_insert_one(struct rhashtable *ht, ht 545 lib/rhashtable.c new_tbl = rht_dereference_rcu(tbl->future_tbl, ht); ht 552 lib/rhashtable.c if (unlikely(rht_grow_above_max(ht, tbl))) ht 555 lib/rhashtable.c if (unlikely(rht_grow_above_100(ht, tbl))) ht 561 lib/rhashtable.c if (ht->rhlist) { ht 573 lib/rhashtable.c atomic_inc(&ht->nelems); ht 574 lib/rhashtable.c if (rht_grow_above_75(ht, tbl)) ht 575 lib/rhashtable.c schedule_work(&ht->run_work); ht 580 lib/rhashtable.c static void *rhashtable_try_insert(struct rhashtable *ht, const void *key, ht 589 lib/rhashtable.c new_tbl = rcu_dereference(ht->tbl); ht 593 lib/rhashtable.c hash = rht_head_hashfn(ht, tbl, obj, ht->p); ht 598 lib/rhashtable.c bkt = rht_bucket_insert(ht, tbl, hash); ht 600 lib/rhashtable.c new_tbl = rht_dereference_rcu(tbl->future_tbl, ht); ht 604 lib/rhashtable.c data = rhashtable_lookup_one(ht, bkt, tbl, ht 606 lib/rhashtable.c new_tbl = rhashtable_insert_one(ht, bkt, tbl, ht 616 lib/rhashtable.c data = ERR_PTR(rhashtable_insert_rehash(ht, tbl) ?: ht 622 lib/rhashtable.c void *rhashtable_insert_slow(struct rhashtable *ht, const void *key, ht 629 lib/rhashtable.c data = rhashtable_try_insert(ht, key, obj); ht 658 lib/rhashtable.c void rhashtable_walk_enter(struct rhashtable *ht, struct rhashtable_iter *iter) ht 660 lib/rhashtable.c iter->ht = ht; ht 666 lib/rhashtable.c spin_lock(&ht->lock); ht 668 lib/rhashtable.c rcu_dereference_protected(ht->tbl, lockdep_is_held(&ht->lock)); ht 670 lib/rhashtable.c spin_unlock(&ht->lock); ht 682 lib/rhashtable.c spin_lock(&iter->ht->lock); ht 685 lib/rhashtable.c spin_unlock(&iter->ht->lock); ht 710 lib/rhashtable.c struct rhashtable *ht = iter->ht; ht 711 lib/rhashtable.c bool rhlist = ht->rhlist; ht 715 lib/rhashtable.c spin_lock(&ht->lock); ht 718 lib/rhashtable.c spin_unlock(&ht->lock); ht 723 lib/rhashtable.c iter->walker.tbl = rht_dereference_rcu(ht->tbl, ht); ht 784 lib/rhashtable.c struct rhashtable *ht = iter->ht; ht 786 lib/rhashtable.c bool rhlist = ht->rhlist; ht 817 lib/rhashtable.c return rht_obj(ht, rhlist ? &list->rhead : p); ht 828 lib/rhashtable.c iter->walker.tbl = rht_dereference_rcu(tbl->future_tbl, ht); ht 855 lib/rhashtable.c struct rhashtable *ht = iter->ht; ht 857 lib/rhashtable.c bool rhlist = ht->rhlist; ht 868 lib/rhashtable.c return rht_obj(ht, rhlist ? &list->rhead : p); ht 894 lib/rhashtable.c struct rhashtable *ht = iter->ht; ht 898 lib/rhashtable.c return rht_obj(ht, ht->rhlist ? &list->rhead : p); ht 926 lib/rhashtable.c struct rhashtable *ht; ht 932 lib/rhashtable.c ht = iter->ht; ht 934 lib/rhashtable.c spin_lock(&ht->lock); ht 940 lib/rhashtable.c spin_unlock(&ht->lock); ht 1008 lib/rhashtable.c int rhashtable_init(struct rhashtable *ht, ht 1018 lib/rhashtable.c memset(ht, 0, sizeof(*ht)); ht 1019 lib/rhashtable.c mutex_init(&ht->mutex); ht 1020 lib/rhashtable.c spin_lock_init(&ht->lock); ht 1021 lib/rhashtable.c memcpy(&ht->p, params, sizeof(*params)); ht 1024 lib/rhashtable.c ht->p.min_size = roundup_pow_of_two(params->min_size); ht 1027 lib/rhashtable.c ht->max_elems = 1u << 31; ht 1030 lib/rhashtable.c ht->p.max_size = rounddown_pow_of_two(params->max_size); ht 1031 lib/rhashtable.c if (ht->p.max_size < ht->max_elems / 2) ht 1032 lib/rhashtable.c ht->max_elems = ht->p.max_size * 2; ht 1035 lib/rhashtable.c ht->p.min_size = max_t(u16, ht->p.min_size, HASH_MIN_SIZE); ht 1037 lib/rhashtable.c size = rounded_hashtable_size(&ht->p); ht 1039 lib/rhashtable.c ht->key_len = ht->p.key_len; ht 1041 lib/rhashtable.c ht->p.hashfn = jhash; ht 1043 lib/rhashtable.c if (!(ht->key_len & (sizeof(u32) - 1))) { ht 1044 lib/rhashtable.c ht->key_len /= sizeof(u32); ht 1045 lib/rhashtable.c ht->p.hashfn = rhashtable_jhash2; ht 1054 lib/rhashtable.c tbl = bucket_table_alloc(ht, size, GFP_KERNEL); ht 1056 lib/rhashtable.c size = max_t(u16, ht->p.min_size, HASH_MIN_SIZE); ht 1057 lib/rhashtable.c tbl = bucket_table_alloc(ht, size, GFP_KERNEL | __GFP_NOFAIL); ht 1060 lib/rhashtable.c atomic_set(&ht->nelems, 0); ht 1062 lib/rhashtable.c RCU_INIT_POINTER(ht->tbl, tbl); ht 1064 lib/rhashtable.c INIT_WORK(&ht->run_work, rht_deferred_worker); ht 1083 lib/rhashtable.c err = rhashtable_init(&hlt->ht, params); ht 1084 lib/rhashtable.c hlt->ht.rhlist = true; ht 1089 lib/rhashtable.c static void rhashtable_free_one(struct rhashtable *ht, struct rhash_head *obj, ht 1095 lib/rhashtable.c if (!ht->rhlist) { ht 1096 lib/rhashtable.c free_fn(rht_obj(ht, obj), arg); ht 1103 lib/rhashtable.c list = rht_dereference(list->next, ht); ht 1104 lib/rhashtable.c free_fn(rht_obj(ht, obj), arg); ht 1123 lib/rhashtable.c void rhashtable_free_and_destroy(struct rhashtable *ht, ht 1130 lib/rhashtable.c cancel_work_sync(&ht->run_work); ht 1132 lib/rhashtable.c mutex_lock(&ht->mutex); ht 1133 lib/rhashtable.c tbl = rht_dereference(ht->tbl, ht); ht 1142 lib/rhashtable.c rht_dereference(pos->next, ht) : NULL; ht 1146 lib/rhashtable.c rht_dereference(pos->next, ht) : NULL) ht 1147 lib/rhashtable.c rhashtable_free_one(ht, pos, free_fn, arg); ht 1151 lib/rhashtable.c next_tbl = rht_dereference(tbl->future_tbl, ht); ht 1157 lib/rhashtable.c mutex_unlock(&ht->mutex); ht 1161 lib/rhashtable.c void rhashtable_destroy(struct rhashtable *ht) ht 1163 lib/rhashtable.c return rhashtable_free_and_destroy(ht, NULL, NULL); ht 1207 lib/rhashtable.c struct rhash_lock_head **rht_bucket_nested_insert(struct rhashtable *ht, ht 1218 lib/rhashtable.c ntbl = nested_table_alloc(ht, &ntbl[index].table, ht 1225 lib/rhashtable.c ntbl = nested_table_alloc(ht, &ntbl[index].table, ht 115 lib/test_rhashtable.c static int insert_retry(struct rhashtable *ht, struct test_obj *obj, ht 123 lib/test_rhashtable.c err = rhashtable_insert_fast(ht, &obj->node, params); ht 137 lib/test_rhashtable.c static int __init test_rht_lookup(struct rhashtable *ht, struct test_obj *array, ht 152 lib/test_rhashtable.c obj = rhashtable_lookup_fast(ht, &key, test_rht_params); ht 175 lib/test_rhashtable.c static void test_bucket_stats(struct rhashtable *ht, unsigned int entries) ht 181 lib/test_rhashtable.c rhashtable_walk_enter(ht, &hti); ht 202 lib/test_rhashtable.c total, atomic_read(&ht->nelems), entries, chain_len); ht 204 lib/test_rhashtable.c if (total != atomic_read(&ht->nelems) || total != entries) ht 208 lib/test_rhashtable.c static s64 __init test_rhashtable(struct rhashtable *ht, struct test_obj *array, ht 226 lib/test_rhashtable.c err = insert_retry(ht, obj, test_rht_params); ht 237 lib/test_rhashtable.c test_bucket_stats(ht, entries); ht 239 lib/test_rhashtable.c test_rht_lookup(ht, array, entries); ht 242 lib/test_rhashtable.c test_bucket_stats(ht, entries); ht 251 lib/test_rhashtable.c obj = rhashtable_lookup_fast(ht, &key, test_rht_params); ht 254 lib/test_rhashtable.c rhashtable_remove_fast(ht, &obj->node, test_rht_params); ht 266 lib/test_rhashtable.c static struct rhashtable ht; ht 455 lib/test_rhashtable.c err = rhashtable_init(&ht, &test_rht_params); ht 459 lib/test_rhashtable.c for (i = 0; i < ht.max_elems; i++) { ht 463 lib/test_rhashtable.c err = insert_retry(&ht, obj, test_rht_params); ht 470 lib/test_rhashtable.c err = insert_retry(&ht, &array[ht.max_elems], test_rht_params); ht 475 lib/test_rhashtable.c ht.max_elems, -E2BIG, err); ht 480 lib/test_rhashtable.c rhashtable_destroy(&ht); ht 487 lib/test_rhashtable.c struct rhashtable *ht; ht 492 lib/test_rhashtable.c ht = &rhlt->ht; ht 494 lib/test_rhashtable.c mutex_lock(&ht->mutex); ht 495 lib/test_rhashtable.c tbl = rht_dereference(ht->tbl, ht); ht 501 lib/test_rhashtable.c next = !rht_is_a_nulls(pos) ? rht_dereference(pos->next, ht) : NULL; ht 512 lib/test_rhashtable.c list = rht_dereference(list->next, ht); ht 513 lib/test_rhashtable.c p = rht_obj(ht, pos); ht 522 lib/test_rhashtable.c rht_dereference(pos->next, ht) : NULL; ht 528 lib/test_rhashtable.c mutex_unlock(&ht->mutex); ht 553 lib/test_rhashtable.c key = rht_obj(&rhlt->ht, &rhl_test_objects[i].list_node.rhead); ht 557 lib/test_rhashtable.c err = PTR_ERR(rhashtable_insert_slow(&rhlt->ht, key, ht 613 lib/test_rhashtable.c obj = rhashtable_lookup_fast(&ht, &key, test_rht_params); ht 646 lib/test_rhashtable.c err = insert_retry(&ht, &tdata->objs[i], test_rht_params); ht 670 lib/test_rhashtable.c err = rhashtable_remove_fast(&ht, &tdata->objs[i].node, ht 727 lib/test_rhashtable.c err = rhashtable_init(&ht, &test_rht_params); ht 734 lib/test_rhashtable.c time = test_rhashtable(&ht, objs, entries); ht 735 lib/test_rhashtable.c rhashtable_destroy(&ht); ht 772 lib/test_rhashtable.c err = rhashtable_init(&ht, &test_rht_params); ht 807 lib/test_rhashtable.c rhashtable_destroy(&ht); ht 145 net/decnet/dn_table.c struct dn_fib_node **ht, **old_ht; ht 166 net/decnet/dn_table.c ht = kcalloc(new_divisor, sizeof(struct dn_fib_node*), GFP_KERNEL); ht 167 net/decnet/dn_table.c if (ht == NULL) ht 172 net/decnet/dn_table.c dz->dz_hash = ht; ht 65 net/hsr/hsr_main.h static inline u16 get_hsr_tag_path(struct hsr_tag *ht) ht 67 net/hsr/hsr_main.h return ntohs(ht->path_and_LSDU_size) >> 12; ht 70 net/hsr/hsr_main.h static inline u16 get_hsr_tag_LSDU_size(struct hsr_tag *ht) ht 72 net/hsr/hsr_main.h return ntohs(ht->path_and_LSDU_size) & 0x0FFF; ht 75 net/hsr/hsr_main.h static inline void set_hsr_tag_path(struct hsr_tag *ht, u16 path) ht 77 net/hsr/hsr_main.h ht->path_and_LSDU_size = ht 78 net/hsr/hsr_main.h htons((ntohs(ht->path_and_LSDU_size) & 0x0FFF) | (path << 12)); ht 81 net/hsr/hsr_main.h static inline void set_hsr_tag_LSDU_size(struct hsr_tag *ht, u16 LSDU_size) ht 83 net/hsr/hsr_main.h ht->path_and_LSDU_size = htons((ntohs(ht->path_and_LSDU_size) & ht 98 net/ipv4/raw.c head = &h->ht[inet_sk(sk)->inet_num & (RAW_HTABLE_SIZE - 1)]; ht 179 net/ipv4/raw.c head = &raw_v4_hashinfo.ht[hash]; ht 213 net/ipv4/raw.c raw_sk = sk_head(&raw_v4_hashinfo.ht[hash]); ht 296 net/ipv4/raw.c raw_sk = sk_head(&raw_v4_hashinfo.ht[hash]); ht 999 net/ipv4/raw.c sk_for_each(sk, &h->ht[state->bucket]) ht 1020 net/ipv4/raw.c sk = sk_head(&h->ht[state->bucket]); ht 69 net/ipv4/raw_diag.c sk_for_each(s, &hashinfo->ht[slot]) { ht 166 net/ipv4/raw_diag.c sk_for_each(sk, &hashinfo->ht[slot]) { ht 172 net/ipv6/raw.c sk = sk_head(&raw_v6_hashinfo.ht[hash]); ht 234 net/ipv6/raw.c raw_sk = sk_head(&raw_v6_hashinfo.ht[nexthdr & (RAW_HTABLE_SIZE - 1)]); ht 372 net/ipv6/raw.c sk = sk_head(&raw_v6_hashinfo.ht[hash]); ht 888 net/mac80211/rc80211_minstrel_ht.c struct minstrel_ht_sta *mi = &msp->ht; ht 1295 net/mac80211/rc80211_minstrel_ht.c struct minstrel_ht_sta *mi = &msp->ht; ht 1388 net/mac80211/rc80211_minstrel_ht.c struct minstrel_ht_sta *mi = &msp->ht; ht 1699 net/mac80211/rc80211_minstrel_ht.c struct minstrel_ht_sta *mi = &msp->ht; ht 112 net/mac80211/rc80211_minstrel_ht.h struct minstrel_ht_sta ht; ht 124 net/mac80211/rc80211_minstrel_ht_debugfs.c struct minstrel_ht_sta *mi = &msp->ht; ht 274 net/mac80211/rc80211_minstrel_ht_debugfs.c struct minstrel_ht_sta *mi = &msp->ht; ht 742 net/netfilter/ipset/ip_set_hash_gen.h struct hbucket *ht; ht 747 net/netfilter/ipset/ip_set_hash_gen.h ht = kzalloc(sizeof(*ht) + ht 751 net/netfilter/ipset/ip_set_hash_gen.h if (!ht) ht 756 net/netfilter/ipset/ip_set_hash_gen.h memcpy(ht, m, sizeof(struct hbucket) + ht 758 net/netfilter/ipset/ip_set_hash_gen.h ht->size = m->size + AHASH_INIT_SIZE; ht 763 net/netfilter/ipset/ip_set_hash_gen.h m = ht; ht 764 net/netfilter/ipset/ip_set_hash_gen.h RCU_INIT_POINTER(hbucket(t, key), ht); ht 25 net/netfilter/nft_set_hash.c struct rhashtable ht; ht 88 net/netfilter/nft_set_hash.c he = rhashtable_lookup(&priv->ht, &arg, nft_rhash_params); ht 106 net/netfilter/nft_set_hash.c he = rhashtable_lookup(&priv->ht, &arg, nft_rhash_params); ht 129 net/netfilter/nft_set_hash.c he = rhashtable_lookup(&priv->ht, &arg, nft_rhash_params); ht 137 net/netfilter/nft_set_hash.c prev = rhashtable_lookup_get_insert_key(&priv->ht, &arg, &he->node, ht 171 net/netfilter/nft_set_hash.c prev = rhashtable_lookup_get_insert_key(&priv->ht, &arg, &he->node, ht 217 net/netfilter/nft_set_hash.c he = rhashtable_lookup(&priv->ht, &arg, nft_rhash_params); ht 234 net/netfilter/nft_set_hash.c rhashtable_remove_fast(&priv->ht, &he->node, nft_rhash_params); ht 248 net/netfilter/nft_set_hash.c he = rhashtable_lookup(&priv->ht, &arg, nft_rhash_params); ht 252 net/netfilter/nft_set_hash.c return rhashtable_remove_fast(&priv->ht, &he->node, nft_rhash_params) == 0; ht 263 net/netfilter/nft_set_hash.c rhashtable_walk_enter(&priv->ht, &hti); ht 307 net/netfilter/nft_set_hash.c rhashtable_walk_enter(&priv->ht, &hti); ht 333 net/netfilter/nft_set_hash.c rhashtable_remove_fast(&priv->ht, &he->node, nft_rhash_params); ht 370 net/netfilter/nft_set_hash.c err = rhashtable_init(&priv->ht, ¶ms); ht 392 net/netfilter/nft_set_hash.c rhashtable_free_and_destroy(&priv->ht, nft_rhash_elem_destroy, ht 184 net/netfilter/xt_hashlimit.c hash_dst(const struct xt_hashlimit_htable *ht, const struct dsthash_dst *dst) ht 188 net/netfilter/xt_hashlimit.c ht->rnd); ht 195 net/netfilter/xt_hashlimit.c return reciprocal_scale(hash, ht->cfg.size); ht 199 net/netfilter/xt_hashlimit.c dsthash_find(const struct xt_hashlimit_htable *ht, ht 203 net/netfilter/xt_hashlimit.c u_int32_t hash = hash_dst(ht, dst); ht 205 net/netfilter/xt_hashlimit.c if (!hlist_empty(&ht->hash[hash])) { ht 206 net/netfilter/xt_hashlimit.c hlist_for_each_entry_rcu(ent, &ht->hash[hash], node) ht 217 net/netfilter/xt_hashlimit.c dsthash_alloc_init(struct xt_hashlimit_htable *ht, ht 222 net/netfilter/xt_hashlimit.c spin_lock(&ht->lock); ht 227 net/netfilter/xt_hashlimit.c ent = dsthash_find(ht, dst); ht 229 net/netfilter/xt_hashlimit.c spin_unlock(&ht->lock); ht 236 net/netfilter/xt_hashlimit.c if (unlikely(!ht->rnd_initialized)) { ht 237 net/netfilter/xt_hashlimit.c get_random_bytes(&ht->rnd, sizeof(ht->rnd)); ht 238 net/netfilter/xt_hashlimit.c ht->rnd_initialized = true; ht 241 net/netfilter/xt_hashlimit.c if (ht->cfg.max && ht->count >= ht->cfg.max) { ht 243 net/netfilter/xt_hashlimit.c net_err_ratelimited("max count of %u reached\n", ht->cfg.max); ht 252 net/netfilter/xt_hashlimit.c hlist_add_head_rcu(&ent->node, &ht->hash[hash_dst(ht, dst)]); ht 253 net/netfilter/xt_hashlimit.c ht->count++; ht 255 net/netfilter/xt_hashlimit.c spin_unlock(&ht->lock); ht 267 net/netfilter/xt_hashlimit.c dsthash_free(struct xt_hashlimit_htable *ht, struct dsthash_ent *ent) ht 271 net/netfilter/xt_hashlimit.c ht->count--; ht 361 net/netfilter/xt_hashlimit.c static void htable_selective_cleanup(struct xt_hashlimit_htable *ht, bool select_all) ht 365 net/netfilter/xt_hashlimit.c for (i = 0; i < ht->cfg.size; i++) { ht 369 net/netfilter/xt_hashlimit.c spin_lock_bh(&ht->lock); ht 370 net/netfilter/xt_hashlimit.c hlist_for_each_entry_safe(dh, n, &ht->hash[i], node) { ht 372 net/netfilter/xt_hashlimit.c dsthash_free(ht, dh); ht 374 net/netfilter/xt_hashlimit.c spin_unlock_bh(&ht->lock); ht 381 net/netfilter/xt_hashlimit.c struct xt_hashlimit_htable *ht; ht 383 net/netfilter/xt_hashlimit.c ht = container_of(work, struct xt_hashlimit_htable, gc_work.work); ht 385 net/netfilter/xt_hashlimit.c htable_selective_cleanup(ht, false); ht 388 net/netfilter/xt_hashlimit.c &ht->gc_work, msecs_to_jiffies(ht->cfg.gc_interval)); ht 1128 net/netfilter/xt_hashlimit.c struct xt_hashlimit_htable *ht = PDE_DATA(file_inode(s->file)); ht 1132 net/netfilter/xt_hashlimit.c rateinfo_recalc(ent, jiffies, ht->cfg.mode, 2); ht 1143 net/netfilter/xt_hashlimit.c struct xt_hashlimit_htable *ht = PDE_DATA(file_inode(s->file)); ht 1147 net/netfilter/xt_hashlimit.c rateinfo_recalc(ent, jiffies, ht->cfg.mode, 1); ht 1158 net/netfilter/xt_hashlimit.c struct xt_hashlimit_htable *ht = PDE_DATA(file_inode(s->file)); ht 1162 net/netfilter/xt_hashlimit.c rateinfo_recalc(ent, jiffies, ht->cfg.mode, 3); ht 77 net/sched/cls_flower.c struct rhashtable ht; ht 94 net/sched/cls_flower.c struct rhashtable ht; ht 249 net/sched/cls_flower.c return rhashtable_lookup_fast(&mask->ht, fl_key_get_start(mkey, mask), ht 345 net/sched/cls_flower.c return rhashtable_init(&head->ht, &mask_ht_params); ht 353 net/sched/cls_flower.c rhashtable_destroy(&mask->ht); ht 379 net/sched/cls_flower.c rhashtable_remove_fast(&head->ht, &mask->ht_node, mask_ht_params); ht 536 net/sched/cls_flower.c rhashtable_remove_fast(&f->mask->ht, &f->ht_node, ht 557 net/sched/cls_flower.c rhashtable_destroy(&head->ht); ht 1293 net/sched/cls_flower.c return rhashtable_init(&mask->ht, &mask->filter_ht_params); ht 1399 net/sched/cls_flower.c err = rhashtable_replace_fast(&head->ht, &mask->ht_node, ht 1411 net/sched/cls_flower.c rhashtable_destroy(&newmask->ht); ht 1432 net/sched/cls_flower.c fnew->mask = rhashtable_lookup_get_insert_fast(&head->ht, ht 1463 net/sched/cls_flower.c rhashtable_remove_fast(&head->ht, &mask->ht_node, ht 1513 net/sched/cls_flower.c err = rhashtable_lookup_insert_fast(&mask->ht, ht 1635 net/sched/cls_flower.c err = rhashtable_insert_fast(&fnew->mask->ht, ht 1644 net/sched/cls_flower.c rhashtable_remove_fast(&fold->mask->ht, ht 1704 net/sched/cls_flower.c rhashtable_remove_fast(&fnew->mask->ht, &fnew->ht_node, ht 29 net/sched/cls_fw.c struct fw_filter __rcu *ht[HTSIZE]; ht 61 net/sched/cls_fw.c for (f = rcu_dereference_bh(head->ht[fw_hash(id)]); f; ht 97 net/sched/cls_fw.c f = rtnl_dereference(head->ht[fw_hash(handle)]); ht 141 net/sched/cls_fw.c while ((f = rtnl_dereference(head->ht[h])) != NULL) { ht 142 net/sched/cls_fw.c RCU_INIT_POINTER(head->ht[h], ht 167 net/sched/cls_fw.c fp = &head->ht[fw_hash(f->id)]; ht 183 net/sched/cls_fw.c if (rcu_access_pointer(head->ht[h])) { ht 287 net/sched/cls_fw.c fp = &head->ht[fw_hash(fnew->id)]; ht 333 net/sched/cls_fw.c RCU_INIT_POINTER(f->next, head->ht[fw_hash(handle)]); ht 334 net/sched/cls_fw.c rcu_assign_pointer(head->ht[fw_hash(handle)], f); ht 360 net/sched/cls_fw.c for (f = rtnl_dereference(head->ht[h]); f; ht 42 net/sched/cls_route.c struct route4_filter __rcu *ht[16 + 16 + 1]; ht 164 net/sched/cls_route.c for (f = rcu_dereference_bh(b->ht[route4_hash_from(id)]); ht 170 net/sched/cls_route.c for (f = rcu_dereference_bh(b->ht[route4_hash_iif(iif)]); ht 176 net/sched/cls_route.c for (f = rcu_dereference_bh(b->ht[route4_hash_wild()]); ht 232 net/sched/cls_route.c for (f = rtnl_dereference(b->ht[h2]); ht 292 net/sched/cls_route.c while ((f = rtnl_dereference(b->ht[h2])) != NULL) { ht 296 net/sched/cls_route.c RCU_INIT_POINTER(b->ht[h2], next); ht 328 net/sched/cls_route.c fp = &b->ht[from_hash(h >> 16)]; ht 350 net/sched/cls_route.c rt = rtnl_dereference(b->ht[i]); ht 438 net/sched/cls_route.c for (fp = rtnl_dereference(b->ht[h2]); ht 518 net/sched/cls_route.c fp = &f->bkt->ht[h]; ht 534 net/sched/cls_route.c fp = &b->ht[h]; ht 577 net/sched/cls_route.c for (f = rtnl_dereference(b->ht[h1]); ht 69 net/sched/cls_rsvp.h struct rsvp_session __rcu *ht[256]; ht 80 net/sched/cls_rsvp.h struct rsvp_filter __rcu *ht[16 + 1]; ht 170 net/sched/cls_rsvp.h for (s = rcu_dereference_bh(head->ht[h1]); s; ht 183 net/sched/cls_rsvp.h for (f = rcu_dereference_bh(s->ht[h2]); f; ht 208 net/sched/cls_rsvp.h for (f = rcu_dereference_bh(s->ht[16]); f; ht 229 net/sched/cls_rsvp.h for (s = rtnl_dereference(head->ht[h1]); s; ht 231 net/sched/cls_rsvp.h for (ins = &s->ht[h2], pins = rtnl_dereference(*ins); ; ht 258 net/sched/cls_rsvp.h for (s = rtnl_dereference(head->ht[h1]); s; ht 260 net/sched/cls_rsvp.h for (f = rtnl_dereference(s->ht[h2]); f; ht 323 net/sched/cls_rsvp.h while ((s = rtnl_dereference(data->ht[h1])) != NULL) { ht 324 net/sched/cls_rsvp.h RCU_INIT_POINTER(data->ht[h1], s->next); ht 329 net/sched/cls_rsvp.h while ((f = rtnl_dereference(s->ht[h2])) != NULL) { ht 330 net/sched/cls_rsvp.h rcu_assign_pointer(s->ht[h2], f->next); ht 351 net/sched/cls_rsvp.h fp = &s->ht[(h >> 8) & 0xFF]; ht 361 net/sched/cls_rsvp.h if (s->ht[i]) ht 365 net/sched/cls_rsvp.h sp = &head->ht[h & 0xFF]; ht 382 net/sched/cls_rsvp.h if (rcu_access_pointer(head->ht[h1])) { ht 421 net/sched/cls_rsvp.h struct rsvp_session __rcu **sht = data->ht; ht 434 net/sched/cls_rsvp.h for (f = rtnl_dereference(s->ht[h2]); f; ht 584 net/sched/cls_rsvp.h for (sp = &data->ht[h1]; ht 600 net/sched/cls_rsvp.h fp = &s->ht[h2]; ht 608 net/sched/cls_rsvp.h fp = &s->ht[h2]; ht 637 net/sched/cls_rsvp.h sp = &data->ht[h1]; ht 668 net/sched/cls_rsvp.h for (s = rtnl_dereference(head->ht[h]); s; ht 673 net/sched/cls_rsvp.h for (f = rtnl_dereference(s->ht[h1]); f; ht 82 net/sched/cls_u32.c struct tc_u_knode __rcu *ht[1]; ht 111 net/sched/cls_u32.c struct tc_u_hnode *ht = rcu_dereference_bh(tp->root); ht 123 net/sched/cls_u32.c n = rcu_dereference_bh(ht->ht[sel]); ht 168 net/sched/cls_u32.c ht = rcu_dereference_bh(n->ht_down); ht 169 net/sched/cls_u32.c if (!ht) { ht 200 net/sched/cls_u32.c ht = rcu_dereference_bh(n->ht_down); ht 202 net/sched/cls_u32.c if (ht->divisor) { ht 209 net/sched/cls_u32.c sel = ht->divisor & u32_hash_fold(*data, &n->sel, ht 242 net/sched/cls_u32.c ht = rcu_dereference_bh(n->ht_up); ht 256 net/sched/cls_u32.c struct tc_u_hnode *ht; ht 258 net/sched/cls_u32.c for (ht = rtnl_dereference(tp_c->hlist); ht 259 net/sched/cls_u32.c ht; ht 260 net/sched/cls_u32.c ht = rtnl_dereference(ht->next)) ht 261 net/sched/cls_u32.c if (ht->handle == handle) ht 264 net/sched/cls_u32.c return ht; ht 267 net/sched/cls_u32.c static struct tc_u_knode *u32_lookup_key(struct tc_u_hnode *ht, u32 handle) ht 273 net/sched/cls_u32.c if (sel > ht->divisor) ht 276 net/sched/cls_u32.c for (n = rtnl_dereference(ht->ht[sel]); ht 288 net/sched/cls_u32.c struct tc_u_hnode *ht; ht 292 net/sched/cls_u32.c ht = rtnl_dereference(tp->root); ht 294 net/sched/cls_u32.c ht = u32_lookup_ht(tp_c, TC_U32_HTID(handle)); ht 296 net/sched/cls_u32.c if (!ht) ht 300 net/sched/cls_u32.c return ht; ht 302 net/sched/cls_u32.c return u32_lookup_key(ht, handle); ht 391 net/sched/cls_u32.c struct tc_u_hnode *ht = rtnl_dereference(n->ht_down); ht 395 net/sched/cls_u32.c if (ht && --ht->refcnt == 0) ht 396 net/sched/cls_u32.c kfree(ht); ht 449 net/sched/cls_u32.c struct tc_u_hnode *ht = rtnl_dereference(key->ht_up); ht 451 net/sched/cls_u32.c if (ht) { ht 452 net/sched/cls_u32.c kp = &ht->ht[TC_U32_HASH(key->handle)]; ht 460 net/sched/cls_u32.c idr_remove(&ht->handle_idr, key->handle); ht 532 net/sched/cls_u32.c struct tc_u_hnode *ht = rtnl_dereference(n->ht_down); ht 553 net/sched/cls_u32.c cls_u32.knode.link_handle = ht->handle; ht 568 net/sched/cls_u32.c static void u32_clear_hnode(struct tcf_proto *tp, struct tc_u_hnode *ht, ht 575 net/sched/cls_u32.c for (h = 0; h <= ht->divisor; h++) { ht 576 net/sched/cls_u32.c while ((n = rtnl_dereference(ht->ht[h])) != NULL) { ht 577 net/sched/cls_u32.c RCU_INIT_POINTER(ht->ht[h], ht 582 net/sched/cls_u32.c idr_remove(&ht->handle_idr, n->handle); ht 591 net/sched/cls_u32.c static int u32_destroy_hnode(struct tcf_proto *tp, struct tc_u_hnode *ht, ht 598 net/sched/cls_u32.c WARN_ON(--ht->refcnt); ht 600 net/sched/cls_u32.c u32_clear_hnode(tp, ht, extack); ht 606 net/sched/cls_u32.c if (phn == ht) { ht 607 net/sched/cls_u32.c u32_clear_hw_hnode(tp, ht, extack); ht 608 net/sched/cls_u32.c idr_destroy(&ht->handle_idr); ht 609 net/sched/cls_u32.c idr_remove(&tp_c->handle_idr, ht->handle); ht 610 net/sched/cls_u32.c RCU_INIT_POINTER(*hn, ht->next); ht 611 net/sched/cls_u32.c kfree_rcu(ht, rcu); ht 631 net/sched/cls_u32.c struct tc_u_hnode *ht; ht 635 net/sched/cls_u32.c while ((ht = rtnl_dereference(tp_c->hlist)) != NULL) { ht 636 net/sched/cls_u32.c u32_clear_hnode(tp, ht, extack); ht 637 net/sched/cls_u32.c RCU_INIT_POINTER(tp_c->hlist, ht->next); ht 642 net/sched/cls_u32.c if (--ht->refcnt == 0) ht 643 net/sched/cls_u32.c kfree_rcu(ht, rcu); ht 656 net/sched/cls_u32.c struct tc_u_hnode *ht = arg; ht 660 net/sched/cls_u32.c if (TC_U32_KEY(ht->handle)) { ht 661 net/sched/cls_u32.c u32_remove_hw_knode(tp, (struct tc_u_knode *)ht, extack); ht 662 net/sched/cls_u32.c ret = u32_delete_key(tp, (struct tc_u_knode *)ht); ht 666 net/sched/cls_u32.c if (ht->is_root) { ht 671 net/sched/cls_u32.c if (ht->refcnt == 1) { ht 672 net/sched/cls_u32.c u32_destroy_hnode(tp, ht, extack); ht 683 net/sched/cls_u32.c static u32 gen_new_kid(struct tc_u_hnode *ht, u32 htid) ht 688 net/sched/cls_u32.c if (idr_alloc_u32(&ht->handle_idr, NULL, &index, max, GFP_KERNEL)) { ht 690 net/sched/cls_u32.c if (idr_alloc_u32(&ht->handle_idr, NULL, &index, max, ht 770 net/sched/cls_u32.c struct tc_u_hnode *ht; ht 773 net/sched/cls_u32.c ht = rtnl_dereference(tp->root); ht 775 net/sched/cls_u32.c ht = u32_lookup_ht(tp_c, TC_U32_HTID(n->handle)); ht 777 net/sched/cls_u32.c ins = &ht->ht[TC_U32_HASH(n->handle)]; ht 787 net/sched/cls_u32.c idr_replace(&ht->handle_idr, n, n->handle); ht 795 net/sched/cls_u32.c struct tc_u_hnode *ht = rtnl_dereference(n->ht_down); ht 813 net/sched/cls_u32.c RCU_INIT_POINTER(new->ht_down, ht); ht 816 net/sched/cls_u32.c if (ht) ht 817 net/sched/cls_u32.c ht->refcnt++; ht 849 net/sched/cls_u32.c struct tc_u_hnode *ht; ht 941 net/sched/cls_u32.c ht = kzalloc(sizeof(*ht) + divisor*sizeof(void *), GFP_KERNEL); ht 942 net/sched/cls_u32.c if (ht == NULL) ht 945 net/sched/cls_u32.c handle = gen_new_htid(tp->data, ht); ht 947 net/sched/cls_u32.c kfree(ht); ht 951 net/sched/cls_u32.c err = idr_alloc_u32(&tp_c->handle_idr, ht, &handle, ht 954 net/sched/cls_u32.c kfree(ht); ht 958 net/sched/cls_u32.c ht->refcnt = 1; ht 959 net/sched/cls_u32.c ht->divisor = divisor; ht 960 net/sched/cls_u32.c ht->handle = handle; ht 961 net/sched/cls_u32.c ht->prio = tp->prio; ht 962 net/sched/cls_u32.c idr_init(&ht->handle_idr); ht 963 net/sched/cls_u32.c ht->flags = flags; ht 965 net/sched/cls_u32.c err = u32_replace_hw_hnode(tp, ht, flags, extack); ht 968 net/sched/cls_u32.c kfree(ht); ht 972 net/sched/cls_u32.c RCU_INIT_POINTER(ht->next, tp_c->hlist); ht 973 net/sched/cls_u32.c rcu_assign_pointer(tp_c->hlist, ht); ht 974 net/sched/cls_u32.c *arg = ht; ht 982 net/sched/cls_u32.c ht = rtnl_dereference(tp->root); ht 983 net/sched/cls_u32.c htid = ht->handle; ht 985 net/sched/cls_u32.c ht = u32_lookup_ht(tp->data, TC_U32_HTID(htid)); ht 986 net/sched/cls_u32.c if (!ht) { ht 992 net/sched/cls_u32.c ht = rtnl_dereference(tp->root); ht 993 net/sched/cls_u32.c htid = ht->handle; ht 996 net/sched/cls_u32.c if (ht->divisor < TC_U32_HASH(htid)) { ht 1007 net/sched/cls_u32.c err = idr_alloc_u32(&ht->handle_idr, NULL, &handle, handle, ht 1012 net/sched/cls_u32.c handle = gen_new_kid(ht, htid); ht 1043 net/sched/cls_u32.c RCU_INIT_POINTER(n->ht_up, ht); ht 1081 net/sched/cls_u32.c ins = &ht->ht[TC_U32_HASH(handle)]; ht 1107 net/sched/cls_u32.c idr_remove(&ht->handle_idr, handle); ht 1115 net/sched/cls_u32.c struct tc_u_hnode *ht; ht 1122 net/sched/cls_u32.c for (ht = rtnl_dereference(tp_c->hlist); ht 1123 net/sched/cls_u32.c ht; ht 1124 net/sched/cls_u32.c ht = rtnl_dereference(ht->next)) { ht 1125 net/sched/cls_u32.c if (ht->prio != tp->prio) ht 1128 net/sched/cls_u32.c if (arg->fn(tp, ht, arg) < 0) { ht 1134 net/sched/cls_u32.c for (h = 0; h <= ht->divisor; h++) { ht 1135 net/sched/cls_u32.c for (n = rtnl_dereference(ht->ht[h]); ht 1152 net/sched/cls_u32.c static int u32_reoffload_hnode(struct tcf_proto *tp, struct tc_u_hnode *ht, ht 1159 net/sched/cls_u32.c tc_cls_common_offload_init(&cls_u32.common, tp, ht->flags, extack); ht 1161 net/sched/cls_u32.c cls_u32.hnode.divisor = ht->divisor; ht 1162 net/sched/cls_u32.c cls_u32.hnode.handle = ht->handle; ht 1163 net/sched/cls_u32.c cls_u32.hnode.prio = ht->prio; ht 1166 net/sched/cls_u32.c if (err && add && tc_skip_sw(ht->flags)) ht 1176 net/sched/cls_u32.c struct tc_u_hnode *ht = rtnl_dereference(n->ht_down); ht 1199 net/sched/cls_u32.c cls_u32.knode.link_handle = ht->handle; ht 1215 net/sched/cls_u32.c struct tc_u_hnode *ht; ht 1220 net/sched/cls_u32.c for (ht = rtnl_dereference(tp_c->hlist); ht 1221 net/sched/cls_u32.c ht; ht 1222 net/sched/cls_u32.c ht = rtnl_dereference(ht->next)) { ht 1223 net/sched/cls_u32.c if (ht->prio != tp->prio) ht 1230 net/sched/cls_u32.c if (add && !tc_skip_hw(ht->flags)) { ht 1231 net/sched/cls_u32.c err = u32_reoffload_hnode(tp, ht, add, cb, cb_priv, ht 1237 net/sched/cls_u32.c for (h = 0; h <= ht->divisor; h++) { ht 1238 net/sched/cls_u32.c for (n = rtnl_dereference(ht->ht[h]); ht 1251 net/sched/cls_u32.c if (!add && !tc_skip_hw(ht->flags)) ht 1252 net/sched/cls_u32.c u32_reoffload_hnode(tp, ht, add, cb, cb_priv, extack); ht 1288 net/sched/cls_u32.c struct tc_u_hnode *ht = fh; ht 1289 net/sched/cls_u32.c u32 divisor = ht->divisor + 1; ht 126 net/sched/sch_sfq.c sfq_index *ht; /* Hash table ('divisor' slots) */ ht 321 net/sched/sch_sfq.c q->ht[slot->hash] = SFQ_EMPTY_SLOT; ht 365 net/sched/sch_sfq.c x = q->ht[hash]; ht 371 net/sched/sch_sfq.c q->ht[hash] = x; ht 507 net/sched/sch_sfq.c q->ht[slot->hash] = SFQ_EMPTY_SLOT; ht 558 net/sched/sch_sfq.c q->ht[slot->hash] = SFQ_EMPTY_SLOT; ht 564 net/sched/sch_sfq.c sfq_index x = q->ht[hash]; ht 577 net/sched/sch_sfq.c q->ht[hash] = x; ht 726 net/sched/sch_sfq.c sfq_free(q->ht); ht 767 net/sched/sch_sfq.c q->ht = sfq_alloc(sizeof(q->ht[0]) * q->divisor); ht 769 net/sched/sch_sfq.c if (!q->ht || !q->slots) { ht 775 net/sched/sch_sfq.c q->ht[i] = SFQ_EMPTY_SLOT; ht 866 net/sched/sch_sfq.c sfq_index idx = q->ht[cl - 1]; ht 891 net/sched/sch_sfq.c if (q->ht[i] == SFQ_EMPTY_SLOT || ht 80 net/smc/af_smc.c head = &h->ht; ht 2071 net/smc/af_smc.c INIT_HLIST_HEAD(&smc_v4_hashinfo.ht); ht 2072 net/smc/af_smc.c INIT_HLIST_HEAD(&smc_v6_hashinfo.ht); ht 203 net/smc/smc_diag.c head = &prot->h.smc_hash->ht; ht 31 tools/include/linux/hashtable.h static inline void __hash_init(struct hlist_head *ht, unsigned int sz) ht 36 tools/include/linux/hashtable.h INIT_HLIST_HEAD(&ht[i]); ht 69 tools/include/linux/hashtable.h static inline bool __hash_empty(struct hlist_head *ht, unsigned int sz) ht 74 tools/include/linux/hashtable.h if (!hlist_empty(&ht[i])) ht 1379 tools/perf/util/auxtrace.c struct hlist_head *ht; ht 1388 tools/perf/util/auxtrace.c ht = calloc(sz, sizeof(struct hlist_head)); ht 1389 tools/perf/util/auxtrace.c if (!ht) ht 1393 tools/perf/util/auxtrace.c INIT_HLIST_HEAD(&ht[i]); ht 1395 tools/perf/util/auxtrace.c c->hashtable = ht;