/linux-4.4.14/net/sched/ |
D | sch_fq_codel.c | 55 struct fq_codel_flow *flows; /* Flows table [flows_cnt] */ member 154 flow = &q->flows[idx]; in fq_codel_drop() 192 flow = &q->flows[idx]; in fq_codel_enqueue() 232 q->backlogs[flow - q->flows] -= qdisc_pkt_len(skb); in dequeue() 302 struct fq_codel_flow *flow = q->flows + i; in fq_codel_reset() 341 if (q->flows) in fq_codel_change() 412 fq_codel_free(q->flows); in fq_codel_destroy() 436 if (!q->flows) { in fq_codel_init() 437 q->flows = fq_codel_zalloc(q->flows_cnt * in fq_codel_init() 439 if (!q->flows) in fq_codel_init() [all …]
|
D | sch_atm.c | 63 struct list_head flows; /* NB: "link" is also on this member 75 list_for_each_entry(flow, &p->flows, list) { in lookup_flow() 338 list_for_each_entry(flow, &p->flows, list) { in atm_tc_walk() 375 list_for_each_entry(flow, &p->flows, list) { in atm_tc_enqueue() 458 list_for_each_entry(flow, &p->flows, list) { in sch_atm_dequeue() 529 list_for_each_entry(flow, &p->flows, list) { in atm_tc_drop() 541 INIT_LIST_HEAD(&p->flows); in atm_tc_init() 543 list_add(&p->link.list, &p->flows); in atm_tc_init() 564 list_for_each_entry(flow, &p->flows, list) in atm_tc_reset() 575 list_for_each_entry(flow, &p->flows, list) in atm_tc_destroy() [all …]
|
D | sch_fq.c | 101 u32 flows; member 205 q->flows -= fcnt; in fq_gc() 248 if (q->flows >= (2U << q->fq_trees_log) && in fq_classify() 249 q->inactive_flows > q->flows/2) in fq_classify() 292 q->flows++; in fq_classify() 548 q->flows = 0; in fq_reset() 593 q->flows -= fcnt; in fq_rehash() 828 .flows = q->flows, in fq_dump_stats()
|
D | sch_sfq.c | 644 if (ctl->flows) in sfq_change() 645 q->maxflows = min_t(u32, ctl->flows, SFQ_MAX_FLOWS); in sfq_change() 774 opt.v0.flows = q->maxflows; in sfq_dump()
|
D | Kconfig | 236 and Keep for responsive flows, CHOose and Kill for unresponsive 237 flows). This is a variation of RED which trys to penalize flows
|
/linux-4.4.14/Documentation/networking/ |
D | scaling.txt | 27 of logical flows. Packets for each flow are steered to a separate receive 170 to the same CPU is CPU load imbalance if flows vary in packet rate. 176 Flow Limit is an optional RPS feature that prioritizes small flows 177 during CPU contention by dropping packets from large flows slightly 178 ahead of those from small flows. It is active only when an RPS or RFS 184 new packet is dropped. Packets from other flows are still only 188 even large flows maintain connectivity. 204 identification of large flows and fewer false positives. The default 239 flows to the CPUs where those flows are being processed. The flow hash 244 same CPU. Indeed, with many flows and few CPUs, it is very likely that [all …]
|
D | pktgen.txt | 92 flows: 0 flowlen: 0 107 flows: 0 250 pktgen.conf-1-1-flows # 1 CPU 1 dev multiple flows. 268 pgset "flows 1" 362 flows
|
D | openvswitch.txt | 13 table" that userspace populates with "flows" that map from keys based 101 A wildcarded flow can represent a group of exact match flows. Each '1' bit 105 by reduce the number of new flows need to be processed by the user space program. 117 two possible approaches: reactively install flows as they miss the kernel 127 The behavior when using overlapping wildcarded flows is undefined. It is the 130 performs best-effort detection of overlapping wildcarded flows and may reject 143 future operations. The kernel is not required to index flows by the original
|
D | mac80211-auth-assoc-deauth.txt | 3 # deauthentication/disassociation flows.
|
D | ixgbe.txt | 124 Supports advanced filters that direct receive packets by their flows to 126 Matches flows and CPU cores for flow affinity. Supports multiple parameters
|
D | bonding.txt | 784 Specifies if dynamic shuffling of flows is enabled in tlb 787 The default behavior of tlb mode is to shuffle active flows across 924 flows. 934 flows. 1675 even successfully terminating some portion of flows). Though this is not 2288 destination. In a local configuration, traffic flows directly to and
|
D | ip-sysctl.txt | 708 on the local machine, hurting latency of other flows, for 1264 identify packet flows for mechanisms like Equal Cost Multipath
|
/linux-4.4.14/net/core/ |
D | pktgen.c | 385 struct flow_state *flows; member 2289 return !!(pkt_dev->flows[flow].flags & F_INIT); in f_seen() 2297 if (pkt_dev->flows[flow].count >= pkt_dev->lflow) { in f_pick() 2299 pkt_dev->flows[flow].count = 0; in f_pick() 2300 pkt_dev->flows[flow].flags = 0; in f_pick() 2309 if (pkt_dev->flows[flow].count > pkt_dev->lflow) { in f_pick() 2310 pkt_dev->flows[flow].count = 0; in f_pick() 2311 pkt_dev->flows[flow].flags = 0; in f_pick() 2326 struct xfrm_state *x = pkt_dev->flows[flow].x; in get_ipsec_sa() 2345 pkt_dev->flows[flow].x = x; in get_ipsec_sa() [all …]
|
D | net-sysfs.c | 809 table->flows[count].cpu = RPS_NO_CPU; in store_rps_dev_flow_table_cnt()
|
D | dev.c | 3257 rflow = &flow_table->flows[flow_id]; in set_rps_cpu() 3328 rflow = &flow_table->flows[hash & flow_table->mask]; in get_rps_cpu() 3396 rflow = &flow_table->flows[flow_id]; in rps_may_expire_flow()
|
/linux-4.4.14/net/dccp/ccids/ |
D | Kconfig | 19 be reasonably fair when competing for bandwidth with TCP-like flows,
|
/linux-4.4.14/include/uapi/linux/ |
D | pkt_sched.h | 195 unsigned flows; /* Maximal number of flows */ member 798 __u32 flows; member
|
/linux-4.4.14/drivers/net/ |
D | tun.c | 200 struct hlist_head flows[TUN_NUM_FLOW_ENTRIES]; member 327 hlist_for_each_entry_safe(e, n, &tun->flows[i], hash_link) in tun_flow_flush() 342 hlist_for_each_entry_safe(e, n, &tun->flows[i], hash_link) { in tun_flow_delete_by_queue() 365 hlist_for_each_entry_safe(e, n, &tun->flows[i], hash_link) { in tun_flow_cleanup() 392 head = &tun->flows[tun_hashfn(rxhash)]; in tun_flow_update() 453 e = tun_flow_find(&tun->flows[tun_hashfn(txq)], txq); in tun_select_queue() 834 e = tun_flow_find(&tun->flows[tun_hashfn(rxhash)], in tun_net_xmit() 976 INIT_HLIST_HEAD(&tun->flows[i]); in tun_flow_init()
|
/linux-4.4.14/Documentation/ABI/testing/ |
D | sysfs-platform-kim | 8 The device name flows down to architecture specific board
|
D | sysfs-class-net-queues | 16 Number of Receive Packet Steering flows being currently
|
/linux-4.4.14/Documentation/devicetree/bindings/soc/ti/ |
D | keystone-navigator-dma.txt | 5 channels and flows for the QMSS(Queue Manager SubSystem) who triggers
|
/linux-4.4.14/Documentation/devicetree/bindings/arm/ |
D | coresight.txt | 6 sinks, links and sources. Trace data produced by one or more sources flows
|
/linux-4.4.14/Documentation/devicetree/bindings/soc/fsl/ |
D | qman.txt | 15 DPAA logic modules, maintains packet ordering within flows. Besides providing
|
/linux-4.4.14/net/ |
D | Kconfig | 294 backlog reaches netdev_max_backlog. If a few out of many active flows 296 maintain capacity for the other flows. This feature provides servers
|
/linux-4.4.14/Documentation/ |
D | media-framework.txt | 32 flows from the entity's output to one or more entity inputs. Pads should 36 on the same entity or on different entities. Data flows from a source
|
/linux-4.4.14/net/ipv4/ |
D | Kconfig | 497 other Reno and H-TCP flows. 591 - Low latency (short flows, queries), 613 o Coexist with flows that use loss-based congestion control.
|
/linux-4.4.14/arch/blackfin/ |
D | Kconfig.debug | 166 flows) and should normally be turned off, except in those nasty
|
/linux-4.4.14/Documentation/misc-devices/mei/ |
D | mei.txt | 36 - Selective network isolation of Ethernet and IP protocol flows based
|
/linux-4.4.14/Documentation/trace/ |
D | coresight.txt | 23 flows through the coresight system (via ATB bus) using links that are connecting
|
/linux-4.4.14/Documentation/nvdimm/ |
D | btt.txt | 211 g. Summarizing - Read and Write flows
|
/linux-4.4.14/Documentation/block/ |
D | biodoc.txt | 969 4.2 Request flows seen by I/O schedulers 971 flows.
|
/linux-4.4.14/include/linux/ |
D | netdevice.h | 641 struct rps_dev_flow flows[0]; member
|
/linux-4.4.14/net/netfilter/ |
D | Kconfig | 348 policies to flows, instead of using the global timeout policy.
|