stab 361 arch/mips/include/asm/sgiarcs.h unsigned long stab; /* Symbol table. */ stab 115 arch/powerpc/platforms/cell/iommu.c unsigned long *stab; stab 312 arch/powerpc/platforms/cell/iommu.c iommu->stab = page_address(page); stab 313 arch/powerpc/platforms/cell/iommu.c memset(iommu->stab, 0, stab_size); stab 345 arch/powerpc/platforms/cell/iommu.c __func__, iommu->nid, iommu->stab, ptab, stab 368 arch/powerpc/platforms/cell/iommu.c iommu->stab[i] = reg | (__pa(ptab) + (n_pte_pages << 12) * stab 370 arch/powerpc/platforms/cell/iommu.c pr_debug("\t[%d] 0x%016lx\n", i, iommu->stab[i]); stab 407 arch/powerpc/platforms/cell/iommu.c reg = IOC_IOST_Origin_E | __pa(iommu->stab) | IOC_IOST_Origin_HW; stab 649 arch/powerpc/platforms/cell/iommu.c iommu->stab = NULL; stab 747 include/asm-generic/vmlinux.lds.h .stab 0 : { *(.stab) } \ stab 749 include/asm-generic/vmlinux.lds.h .stab.excl 0 : { *(.stab.excl) } \ stab 750 include/asm-generic/vmlinux.lds.h .stab.exclstr 0 : { *(.stab.exclstr) } \ stab 751 include/asm-generic/vmlinux.lds.h .stab.index 0 : { *(.stab.index) } \ stab 752 include/asm-generic/vmlinux.lds.h .stab.indexstr 0 : { *(.stab.indexstr) } \ stab 184 include/net/red.h u8 Scell_log, u8 *stab, u32 max_P) stab 216 include/net/red.h if (stab) stab 217 include/net/red.h memcpy(p->Stab, stab, sizeof(p->Stab)); stab 84 include/net/sch_generic.h struct qdisc_size_table __rcu *stab; stab 676 include/net/sch_generic.h const struct qdisc_size_table *stab); stab 785 include/net/sch_generic.h struct qdisc_size_table *stab = rcu_dereference_bh(sch->stab); stab 787 include/net/sch_generic.h if (stab) stab 788 include/net/sch_generic.h __qdisc_calculate_pkt_len(skb, stab); stab 26 net/core/sock_map.c struct bpf_stab *stab; stab 38 net/core/sock_map.c stab = kzalloc(sizeof(*stab), GFP_USER); stab 39 net/core/sock_map.c if (!stab) stab 42 net/core/sock_map.c bpf_map_init_from_attr(&stab->map, attr); stab 43 net/core/sock_map.c raw_spin_lock_init(&stab->lock); stab 46 net/core/sock_map.c cost = (u64) stab->map.max_entries * sizeof(struct sock *); stab 47 net/core/sock_map.c err = bpf_map_charge_init(&stab->map.memory, cost); stab 51 net/core/sock_map.c stab->sks = bpf_map_area_alloc(stab->map.max_entries * stab 53 net/core/sock_map.c stab->map.numa_node); stab 54 net/core/sock_map.c if (stab->sks) stab 55 net/core/sock_map.c return &stab->map; stab 57 net/core/sock_map.c bpf_map_charge_finish(&stab->map.memory); stab 59 net/core/sock_map.c kfree(stab); stab 116 net/core/sock_map.c struct bpf_stab *stab = container_of(map, struct bpf_stab, stab 118 net/core/sock_map.c if (psock->parser.enabled && stab->progs.skb_parser) stab 233 net/core/sock_map.c struct bpf_stab *stab = container_of(map, struct bpf_stab, map); stab 241 net/core/sock_map.c for (i = 0; i < stab->map.max_entries; i++) { stab 242 net/core/sock_map.c struct sock **psk = &stab->sks[i]; stab 258 net/core/sock_map.c bpf_map_area_free(stab->sks); stab 259 net/core/sock_map.c kfree(stab); stab 269 net/core/sock_map.c struct bpf_stab *stab = container_of(map, struct bpf_stab, map); stab 275 net/core/sock_map.c return READ_ONCE(stab->sks[key]); stab 283 net/core/sock_map.c static int __sock_map_delete(struct bpf_stab *stab, struct sock *sk_test, stab 289 net/core/sock_map.c raw_spin_lock_bh(&stab->lock); stab 299 net/core/sock_map.c raw_spin_unlock_bh(&stab->lock); stab 306 net/core/sock_map.c struct bpf_stab *stab = container_of(map, struct bpf_stab, map); stab 308 net/core/sock_map.c __sock_map_delete(stab, sk, link_raw); stab 313 net/core/sock_map.c struct bpf_stab *stab = container_of(map, struct bpf_stab, map); stab 320 net/core/sock_map.c psk = &stab->sks[i]; stab 321 net/core/sock_map.c return __sock_map_delete(stab, NULL, psk); stab 326 net/core/sock_map.c struct bpf_stab *stab = container_of(map, struct bpf_stab, map); stab 330 net/core/sock_map.c if (i == stab->map.max_entries - 1) stab 332 net/core/sock_map.c if (i >= stab->map.max_entries) stab 342 net/core/sock_map.c struct bpf_stab *stab = container_of(map, struct bpf_stab, map); stab 361 net/core/sock_map.c ret = sock_map_link(map, &stab->progs, sk); stab 368 net/core/sock_map.c raw_spin_lock_bh(&stab->lock); stab 369 net/core/sock_map.c osk = stab->sks[idx]; stab 378 net/core/sock_map.c sock_map_add_link(psock, link, map, &stab->sks[idx]); stab 379 net/core/sock_map.c stab->sks[idx] = sk; stab 381 net/core/sock_map.c sock_map_unref(osk, &stab->sks[idx]); stab 382 net/core/sock_map.c raw_spin_unlock_bh(&stab->lock); stab 385 net/core/sock_map.c raw_spin_unlock_bh(&stab->lock); stab 472 net/sched/sch_api.c struct qdisc_size_table *stab; stab 503 net/sched/sch_api.c list_for_each_entry(stab, &qdisc_stab_list, list) { stab 504 net/sched/sch_api.c if (memcmp(&stab->szopts, s, sizeof(*s))) stab 506 net/sched/sch_api.c if (tsize > 0 && memcmp(stab->data, tab, tsize * sizeof(u16))) stab 508 net/sched/sch_api.c stab->refcnt++; stab 509 net/sched/sch_api.c return stab; stab 512 net/sched/sch_api.c stab = kmalloc(sizeof(*stab) + tsize * sizeof(u16), GFP_KERNEL); stab 513 net/sched/sch_api.c if (!stab) stab 516 net/sched/sch_api.c stab->refcnt = 1; stab 517 net/sched/sch_api.c stab->szopts = *s; stab 519 net/sched/sch_api.c memcpy(stab->data, tab, tsize * sizeof(u16)); stab 521 net/sched/sch_api.c list_add_tail(&stab->list, &qdisc_stab_list); stab 523 net/sched/sch_api.c return stab; stab 538 net/sched/sch_api.c static int qdisc_dump_stab(struct sk_buff *skb, struct qdisc_size_table *stab) stab 545 net/sched/sch_api.c if (nla_put(skb, TCA_STAB_BASE, sizeof(stab->szopts), &stab->szopts)) stab 556 net/sched/sch_api.c const struct qdisc_size_table *stab) stab 560 net/sched/sch_api.c pkt_len = skb->len + stab->szopts.overhead; stab 561 net/sched/sch_api.c if (unlikely(!stab->szopts.tsize)) stab 564 net/sched/sch_api.c slot = pkt_len + stab->szopts.cell_align; stab 568 net/sched/sch_api.c slot >>= stab->szopts.cell_log; stab 569 net/sched/sch_api.c if (likely(slot < stab->szopts.tsize)) stab 570 net/sched/sch_api.c pkt_len = stab->data[slot]; stab 572 net/sched/sch_api.c pkt_len = stab->data[stab->szopts.tsize - 1] * stab 573 net/sched/sch_api.c (slot / stab->szopts.tsize) + stab 574 net/sched/sch_api.c stab->data[slot % stab->szopts.tsize]; stab 576 net/sched/sch_api.c pkt_len <<= stab->szopts.size_log; stab 876 net/sched/sch_api.c struct qdisc_size_table *stab; stab 912 net/sched/sch_api.c stab = rtnl_dereference(q->stab); stab 913 net/sched/sch_api.c if (stab && qdisc_dump_stab(skb, stab) < 0) stab 1158 net/sched/sch_api.c struct qdisc_size_table *stab; stab 1243 net/sched/sch_api.c stab = qdisc_get_stab(tca[TCA_STAB], extack); stab 1244 net/sched/sch_api.c if (IS_ERR(stab)) { stab 1245 net/sched/sch_api.c err = PTR_ERR(stab); stab 1248 net/sched/sch_api.c rcu_assign_pointer(sch->stab, stab); stab 1300 net/sched/sch_api.c qdisc_put_stab(rtnl_dereference(sch->stab)); stab 1309 net/sched/sch_api.c struct qdisc_size_table *ostab, *stab = NULL; stab 1327 net/sched/sch_api.c stab = qdisc_get_stab(tca[TCA_STAB], extack); stab 1328 net/sched/sch_api.c if (IS_ERR(stab)) stab 1329 net/sched/sch_api.c return PTR_ERR(stab); stab 1332 net/sched/sch_api.c ostab = rtnl_dereference(sch->stab); stab 1333 net/sched/sch_api.c rcu_assign_pointer(sch->stab, stab); stab 957 net/sched/sch_generic.c qdisc_put_stab(rtnl_dereference(qdisc->stab)); stab 476 net/sched/sch_gred.c u8 *stab, u32 max_P, stab 508 net/sched/sch_gred.c ctl->Scell_log, stab, max_P); stab 642 net/sched/sch_gred.c u8 *stab; stab 670 net/sched/sch_gred.c stab = nla_data(tb[TCA_GRED_STAB]); stab 702 net/sched/sch_gred.c err = gred_change_vq(sch, ctl->DP, ctl, prio, stab, max_P, &prealloc,