stab              361 arch/mips/include/asm/sgiarcs.h 	unsigned long	stab;		   /* Symbol table. */
stab              115 arch/powerpc/platforms/cell/iommu.c 	unsigned long *stab;
stab              312 arch/powerpc/platforms/cell/iommu.c 	iommu->stab = page_address(page);
stab              313 arch/powerpc/platforms/cell/iommu.c 	memset(iommu->stab, 0, stab_size);
stab              345 arch/powerpc/platforms/cell/iommu.c 			__func__, iommu->nid, iommu->stab, ptab,
stab              368 arch/powerpc/platforms/cell/iommu.c 		iommu->stab[i] = reg | (__pa(ptab) + (n_pte_pages << 12) *
stab              370 arch/powerpc/platforms/cell/iommu.c 		pr_debug("\t[%d] 0x%016lx\n", i, iommu->stab[i]);
stab              407 arch/powerpc/platforms/cell/iommu.c 	reg = IOC_IOST_Origin_E | __pa(iommu->stab) | IOC_IOST_Origin_HW;
stab              649 arch/powerpc/platforms/cell/iommu.c 	iommu->stab = NULL;
stab              747 include/asm-generic/vmlinux.lds.h 		.stab 0 : { *(.stab) }					\
stab              749 include/asm-generic/vmlinux.lds.h 		.stab.excl 0 : { *(.stab.excl) }			\
stab              750 include/asm-generic/vmlinux.lds.h 		.stab.exclstr 0 : { *(.stab.exclstr) }			\
stab              751 include/asm-generic/vmlinux.lds.h 		.stab.index 0 : { *(.stab.index) }			\
stab              752 include/asm-generic/vmlinux.lds.h 		.stab.indexstr 0 : { *(.stab.indexstr) }		\
stab              184 include/net/red.h 				 u8 Scell_log, u8 *stab, u32 max_P)
stab              216 include/net/red.h 	if (stab)
stab              217 include/net/red.h 		memcpy(p->Stab, stab, sizeof(p->Stab));
stab               84 include/net/sch_generic.h 	struct qdisc_size_table	__rcu *stab;
stab              676 include/net/sch_generic.h 			       const struct qdisc_size_table *stab);
stab              785 include/net/sch_generic.h 	struct qdisc_size_table *stab = rcu_dereference_bh(sch->stab);
stab              787 include/net/sch_generic.h 	if (stab)
stab              788 include/net/sch_generic.h 		__qdisc_calculate_pkt_len(skb, stab);
stab               26 net/core/sock_map.c 	struct bpf_stab *stab;
stab               38 net/core/sock_map.c 	stab = kzalloc(sizeof(*stab), GFP_USER);
stab               39 net/core/sock_map.c 	if (!stab)
stab               42 net/core/sock_map.c 	bpf_map_init_from_attr(&stab->map, attr);
stab               43 net/core/sock_map.c 	raw_spin_lock_init(&stab->lock);
stab               46 net/core/sock_map.c 	cost = (u64) stab->map.max_entries * sizeof(struct sock *);
stab               47 net/core/sock_map.c 	err = bpf_map_charge_init(&stab->map.memory, cost);
stab               51 net/core/sock_map.c 	stab->sks = bpf_map_area_alloc(stab->map.max_entries *
stab               53 net/core/sock_map.c 				       stab->map.numa_node);
stab               54 net/core/sock_map.c 	if (stab->sks)
stab               55 net/core/sock_map.c 		return &stab->map;
stab               57 net/core/sock_map.c 	bpf_map_charge_finish(&stab->map.memory);
stab               59 net/core/sock_map.c 	kfree(stab);
stab              116 net/core/sock_map.c 			struct bpf_stab *stab = container_of(map, struct bpf_stab,
stab              118 net/core/sock_map.c 			if (psock->parser.enabled && stab->progs.skb_parser)
stab              233 net/core/sock_map.c 	struct bpf_stab *stab = container_of(map, struct bpf_stab, map);
stab              241 net/core/sock_map.c 	for (i = 0; i < stab->map.max_entries; i++) {
stab              242 net/core/sock_map.c 		struct sock **psk = &stab->sks[i];
stab              258 net/core/sock_map.c 	bpf_map_area_free(stab->sks);
stab              259 net/core/sock_map.c 	kfree(stab);
stab              269 net/core/sock_map.c 	struct bpf_stab *stab = container_of(map, struct bpf_stab, map);
stab              275 net/core/sock_map.c 	return READ_ONCE(stab->sks[key]);
stab              283 net/core/sock_map.c static int __sock_map_delete(struct bpf_stab *stab, struct sock *sk_test,
stab              289 net/core/sock_map.c 	raw_spin_lock_bh(&stab->lock);
stab              299 net/core/sock_map.c 	raw_spin_unlock_bh(&stab->lock);
stab              306 net/core/sock_map.c 	struct bpf_stab *stab = container_of(map, struct bpf_stab, map);
stab              308 net/core/sock_map.c 	__sock_map_delete(stab, sk, link_raw);
stab              313 net/core/sock_map.c 	struct bpf_stab *stab = container_of(map, struct bpf_stab, map);
stab              320 net/core/sock_map.c 	psk = &stab->sks[i];
stab              321 net/core/sock_map.c 	return __sock_map_delete(stab, NULL, psk);
stab              326 net/core/sock_map.c 	struct bpf_stab *stab = container_of(map, struct bpf_stab, map);
stab              330 net/core/sock_map.c 	if (i == stab->map.max_entries - 1)
stab              332 net/core/sock_map.c 	if (i >= stab->map.max_entries)
stab              342 net/core/sock_map.c 	struct bpf_stab *stab = container_of(map, struct bpf_stab, map);
stab              361 net/core/sock_map.c 	ret = sock_map_link(map, &stab->progs, sk);
stab              368 net/core/sock_map.c 	raw_spin_lock_bh(&stab->lock);
stab              369 net/core/sock_map.c 	osk = stab->sks[idx];
stab              378 net/core/sock_map.c 	sock_map_add_link(psock, link, map, &stab->sks[idx]);
stab              379 net/core/sock_map.c 	stab->sks[idx] = sk;
stab              381 net/core/sock_map.c 		sock_map_unref(osk, &stab->sks[idx]);
stab              382 net/core/sock_map.c 	raw_spin_unlock_bh(&stab->lock);
stab              385 net/core/sock_map.c 	raw_spin_unlock_bh(&stab->lock);
stab              472 net/sched/sch_api.c 	struct qdisc_size_table *stab;
stab              503 net/sched/sch_api.c 	list_for_each_entry(stab, &qdisc_stab_list, list) {
stab              504 net/sched/sch_api.c 		if (memcmp(&stab->szopts, s, sizeof(*s)))
stab              506 net/sched/sch_api.c 		if (tsize > 0 && memcmp(stab->data, tab, tsize * sizeof(u16)))
stab              508 net/sched/sch_api.c 		stab->refcnt++;
stab              509 net/sched/sch_api.c 		return stab;
stab              512 net/sched/sch_api.c 	stab = kmalloc(sizeof(*stab) + tsize * sizeof(u16), GFP_KERNEL);
stab              513 net/sched/sch_api.c 	if (!stab)
stab              516 net/sched/sch_api.c 	stab->refcnt = 1;
stab              517 net/sched/sch_api.c 	stab->szopts = *s;
stab              519 net/sched/sch_api.c 		memcpy(stab->data, tab, tsize * sizeof(u16));
stab              521 net/sched/sch_api.c 	list_add_tail(&stab->list, &qdisc_stab_list);
stab              523 net/sched/sch_api.c 	return stab;
stab              538 net/sched/sch_api.c static int qdisc_dump_stab(struct sk_buff *skb, struct qdisc_size_table *stab)
stab              545 net/sched/sch_api.c 	if (nla_put(skb, TCA_STAB_BASE, sizeof(stab->szopts), &stab->szopts))
stab              556 net/sched/sch_api.c 			       const struct qdisc_size_table *stab)
stab              560 net/sched/sch_api.c 	pkt_len = skb->len + stab->szopts.overhead;
stab              561 net/sched/sch_api.c 	if (unlikely(!stab->szopts.tsize))
stab              564 net/sched/sch_api.c 	slot = pkt_len + stab->szopts.cell_align;
stab              568 net/sched/sch_api.c 	slot >>= stab->szopts.cell_log;
stab              569 net/sched/sch_api.c 	if (likely(slot < stab->szopts.tsize))
stab              570 net/sched/sch_api.c 		pkt_len = stab->data[slot];
stab              572 net/sched/sch_api.c 		pkt_len = stab->data[stab->szopts.tsize - 1] *
stab              573 net/sched/sch_api.c 				(slot / stab->szopts.tsize) +
stab              574 net/sched/sch_api.c 				stab->data[slot % stab->szopts.tsize];
stab              576 net/sched/sch_api.c 	pkt_len <<= stab->szopts.size_log;
stab              876 net/sched/sch_api.c 	struct qdisc_size_table *stab;
stab              912 net/sched/sch_api.c 	stab = rtnl_dereference(q->stab);
stab              913 net/sched/sch_api.c 	if (stab && qdisc_dump_stab(skb, stab) < 0)
stab             1158 net/sched/sch_api.c 	struct qdisc_size_table *stab;
stab             1243 net/sched/sch_api.c 		stab = qdisc_get_stab(tca[TCA_STAB], extack);
stab             1244 net/sched/sch_api.c 		if (IS_ERR(stab)) {
stab             1245 net/sched/sch_api.c 			err = PTR_ERR(stab);
stab             1248 net/sched/sch_api.c 		rcu_assign_pointer(sch->stab, stab);
stab             1300 net/sched/sch_api.c 	qdisc_put_stab(rtnl_dereference(sch->stab));
stab             1309 net/sched/sch_api.c 	struct qdisc_size_table *ostab, *stab = NULL;
stab             1327 net/sched/sch_api.c 		stab = qdisc_get_stab(tca[TCA_STAB], extack);
stab             1328 net/sched/sch_api.c 		if (IS_ERR(stab))
stab             1329 net/sched/sch_api.c 			return PTR_ERR(stab);
stab             1332 net/sched/sch_api.c 	ostab = rtnl_dereference(sch->stab);
stab             1333 net/sched/sch_api.c 	rcu_assign_pointer(sch->stab, stab);
stab              957 net/sched/sch_generic.c 	qdisc_put_stab(rtnl_dereference(qdisc->stab));
stab              476 net/sched/sch_gred.c 				 u8 *stab, u32 max_P,
stab              508 net/sched/sch_gred.c 		      ctl->Scell_log, stab, max_P);
stab              642 net/sched/sch_gred.c 	u8 *stab;
stab              670 net/sched/sch_gred.c 	stab = nla_data(tb[TCA_GRED_STAB]);
stab              702 net/sched/sch_gred.c 	err = gred_change_vq(sch, ctl->DP, ctl, prio, stab, max_P, &prealloc,