flows 3254 drivers/infiniband/core/uverbs_cmd.c flow_attr = kzalloc(struct_size(flow_attr, flows, flows 342 drivers/infiniband/hw/hfi1/hfi.h struct tid_flow_state flows[RXE_NUM_TID_FLOWS]; flows 824 drivers/infiniband/hw/hfi1/rc.c &req->flows[req->setup_head]; flows 1114 drivers/infiniband/hw/hfi1/rc.c flow = &req->flows[req->flow_idx]; flows 752 drivers/infiniband/hw/hfi1/tid_rdma.c u32 generation = rcd->flows[flow_idx].generation; flows 770 drivers/infiniband/hw/hfi1/tid_rdma.c rcd->flows[flow_idx].generation = flows 771 drivers/infiniband/hw/hfi1/tid_rdma.c kern_flow_generation_next(rcd->flows[flow_idx].generation); flows 799 drivers/infiniband/hw/hfi1/tid_rdma.c rcd->flows[fs->index].generation = fs->generation; flows 848 drivers/infiniband/hw/hfi1/tid_rdma.c rcd->flows[i].generation = mask_generation(prandom_u32()); flows 1458 drivers/infiniband/hw/hfi1/tid_rdma.c struct tid_rdma_flow *flow = &req->flows[req->setup_head]; flows 1549 drivers/infiniband/hw/hfi1/tid_rdma.c struct tid_rdma_flow *flow = &req->flows[req->clear_tail]; flows 1605 drivers/infiniband/hw/hfi1/tid_rdma.c kfree(req->flows); flows 1606 drivers/infiniband/hw/hfi1/tid_rdma.c req->flows = NULL; flows 1627 drivers/infiniband/hw/hfi1/tid_rdma.c struct tid_rdma_flow *flows; flows 1630 drivers/infiniband/hw/hfi1/tid_rdma.c if (likely(req->flows)) flows 1632 drivers/infiniband/hw/hfi1/tid_rdma.c flows = kmalloc_node(MAX_FLOWS * sizeof(*flows), gfp, flows 1634 drivers/infiniband/hw/hfi1/tid_rdma.c if (!flows) flows 1638 drivers/infiniband/hw/hfi1/tid_rdma.c flows[i].req = req; flows 1639 drivers/infiniband/hw/hfi1/tid_rdma.c flows[i].npagesets = 0; flows 1640 drivers/infiniband/hw/hfi1/tid_rdma.c flows[i].pagesets[0].mapped = 0; flows 1641 drivers/infiniband/hw/hfi1/tid_rdma.c flows[i].resync_npkts = 0; flows 1643 drivers/infiniband/hw/hfi1/tid_rdma.c req->flows = flows; flows 1684 drivers/infiniband/hw/hfi1/tid_rdma.c flow = &req->flows[tail]; flows 1701 drivers/infiniband/hw/hfi1/tid_rdma.c struct tid_rdma_flow *flow = &req->flows[req->flow_idx]; flows 1857 drivers/infiniband/hw/hfi1/tid_rdma.c flow = &req->flows[req->flow_idx]; flows 1896 drivers/infiniband/hw/hfi1/tid_rdma.c flow = &req->flows[req->setup_head]; flows 2120 drivers/infiniband/hw/hfi1/tid_rdma.c fstate = &req->flows[req->clear_tail].flow_state; flows 2344 drivers/infiniband/hw/hfi1/tid_rdma.c struct tid_rdma_flow *flow = &req->flows[req->clear_tail]; flows 2467 drivers/infiniband/hw/hfi1/tid_rdma.c flow = &req->flows[req->clear_tail]; flows 2631 drivers/infiniband/hw/hfi1/tid_rdma.c flow = &req->flows[req->clear_tail]; flows 2749 drivers/infiniband/hw/hfi1/tid_rdma.c flow = &req->flows[req->clear_tail]; flows 2937 drivers/infiniband/hw/hfi1/tid_rdma.c flow = &req->flows[req->clear_tail]; flows 3061 drivers/infiniband/hw/hfi1/tid_rdma.c flow = &req->flows[fidx]; flows 3133 drivers/infiniband/hw/hfi1/tid_rdma.c req->flows[fidx].sent = 0; flows 3134 drivers/infiniband/hw/hfi1/tid_rdma.c req->flows[fidx].pkt = 0; flows 3135 drivers/infiniband/hw/hfi1/tid_rdma.c req->flows[fidx].tid_idx = 0; flows 3136 drivers/infiniband/hw/hfi1/tid_rdma.c req->flows[fidx].tid_offset = 0; flows 3137 drivers/infiniband/hw/hfi1/tid_rdma.c req->flows[fidx].resync_npkts = 0; flows 3839 drivers/infiniband/hw/hfi1/tid_rdma.c flow = &req->flows[req->flow_idx]; flows 4093 drivers/infiniband/hw/hfi1/tid_rdma.c flow = &req->flows[req->setup_head]; flows 4197 drivers/infiniband/hw/hfi1/tid_rdma.c struct tid_rdma_flow *flow = &req->flows[req->clear_tail]; flows 4282 drivers/infiniband/hw/hfi1/tid_rdma.c flow = &req->flows[req->clear_tail]; flows 4336 drivers/infiniband/hw/hfi1/tid_rdma.c rcd->flows[flow->idx].psn = psn & HFI1_KDETH_BTH_SEQ_MASK; flows 4420 drivers/infiniband/hw/hfi1/tid_rdma.c struct tid_rdma_flow *flow = &req->flows[iflow]; flows 4532 drivers/infiniband/hw/hfi1/tid_rdma.c flow = &req->flows[req->acked_tail]; flows 4562 drivers/infiniband/hw/hfi1/tid_rdma.c flow = &req->flows[req->acked_tail]; flows 4624 drivers/infiniband/hw/hfi1/tid_rdma.c flow = &req->flows[req->acked_tail]; flows 4662 drivers/infiniband/hw/hfi1/tid_rdma.c flow = &rptr->flows[fidx]; flows 4708 drivers/infiniband/hw/hfi1/tid_rdma.c if (!req->flows) flows 4710 drivers/infiniband/hw/hfi1/tid_rdma.c flow = &req->flows[req->acked_tail]; flows 4837 drivers/infiniband/hw/hfi1/tid_rdma.c struct tid_rdma_flow *flow = &req->flows[fidx]; flows 4899 drivers/infiniband/hw/hfi1/tid_rdma.c rcd->flows[fs->index].generation = generation; flows 4934 drivers/infiniband/hw/hfi1/tid_rdma.c flow = &req->flows[flow_idx]; flows 5276 drivers/infiniband/hw/hfi1/tid_rdma.c full_flow_psn(&req->flows[flow], flows 5277 drivers/infiniband/hw/hfi1/tid_rdma.c req->flows[flow].flow_state.lpsn)) > 0))) { flows 105 drivers/infiniband/hw/hfi1/tid_rdma.h struct tid_rdma_flow *flows; /* array of tid flows */ flows 163 drivers/net/ethernet/mellanox/mlx5/core/en_rep.h struct list_head flows; flows 175 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c struct list_head flows; flows 200 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c struct list_head flows; flows 335 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c WARN_ON(!list_empty(&mh->flows)); flows 391 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c INIT_LIST_HEAD(&mh->flows); flows 412 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c list_add(&flow->mod_hdr, &mh->flows); flows 729 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c WARN_ON(!list_empty(&hpe->flows)); flows 819 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c INIT_LIST_HEAD(&hpe->flows); flows 872 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c list_add(&flow->hairpin, &hpe->flows); flows 1414 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c list_for_each_entry(efi, &e->flows, list) { flows 1510 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c list_for_each_entry_safe(efi, tmp, &e->flows, list) { flows 1555 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c WARN_ON(!list_empty(&e->flows)); flows 3054 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c INIT_LIST_HEAD(&e->flows); flows 3074 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c list_add(&flow->encaps[out_index].list, &e->flows); flows 685 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c struct mlx5_flow_handle **flows; flows 699 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c flows = kvzalloc(nvports * sizeof(*flows), GFP_KERNEL); flows 700 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c if (!flows) { flows 719 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c flows[MLX5_VPORT_PF] = flow; flows 730 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c flows[mlx5_eswitch_ecpf_idx(esw)] = flow; flows 744 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c flows[i] = flow; flows 747 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c esw->fdb_table.offloads.peer_miss_rules = flows; flows 755 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c mlx5_del_flow_rules(flows[i]); flows 758 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c mlx5_del_flow_rules(flows[mlx5_eswitch_ecpf_idx(esw)]); flows 761 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c mlx5_del_flow_rules(flows[MLX5_VPORT_PF]); flows 764 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c kvfree(flows); flows 772 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c struct mlx5_flow_handle **flows; flows 775 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c flows = esw->fdb_table.offloads.peer_miss_rules; flows 779 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c mlx5_del_flow_rules(flows[i]); flows 782 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c mlx5_del_flow_rules(flows[mlx5_eswitch_ecpf_idx(esw)]); flows 785 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c mlx5_del_flow_rules(flows[MLX5_VPORT_PF]); flows 787 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c kvfree(flows); flows 232 drivers/net/tun.c struct hlist_head flows[TUN_NUM_FLOW_ENTRIES]; flows 457 drivers/net/tun.c hlist_for_each_entry_safe(e, n, &tun->flows[i], hash_link) flows 472 drivers/net/tun.c hlist_for_each_entry_safe(e, n, &tun->flows[i], hash_link) { flows 495 drivers/net/tun.c hlist_for_each_entry_safe(e, n, &tun->flows[i], hash_link) { flows 522 drivers/net/tun.c head = &tun->flows[tun_hashfn(rxhash)]; flows 574 drivers/net/tun.c e = tun_flow_find(&tun->flows[tun_hashfn(txq)], txq); flows 1042 drivers/net/tun.c e = tun_flow_find(&tun->flows[tun_hashfn(rxhash)], rxhash); flows 1364 drivers/net/tun.c INIT_HLIST_HEAD(&tun->flows[i]); flows 689 include/linux/netdevice.h struct rps_dev_flow flows[0]; flows 1373 include/net/cfg80211.h u32 flows; flows 54 include/net/fq.h u32 flows; flows 68 include/net/fq.h struct fq_flow *flows; flows 125 include/net/fq_impl.h flow = &fq->flows[idx]; flows 133 include/net/fq_impl.h tin->flows++; flows 316 include/net/fq_impl.h fq->flows = kvcalloc(fq->flows_cnt, sizeof(fq->flows[0]), GFP_KERNEL); flows 317 include/net/fq_impl.h if (!fq->flows) flows 321 include/net/fq_impl.h fq_flow_init(&fq->flows[i]); flows 332 include/net/fq_impl.h fq_flow_reset(fq, &fq->flows[i], free_func); flows 334 include/net/fq_impl.h kvfree(fq->flows); flows 335 include/net/fq_impl.h fq->flows = NULL; flows 2056 include/rdma/ib_verbs.h union ib_flow_spec flows[]; flows 218 include/uapi/linux/pkt_sched.h unsigned flows; /* Maximal number of flows */ flows 911 include/uapi/linux/pkt_sched.h __u32 flows; flows 3928 net/core/dev.c rflow = &flow_table->flows[flow_id]; flows 3999 net/core/dev.c rflow = &flow_table->flows[hash & flow_table->mask]; flows 4067 net/core/dev.c rflow = &flow_table->flows[flow_id]; flows 830 net/core/net-sysfs.c table->flows[count].cpu = RPS_NO_CPU; flows 399 net/core/pktgen.c struct flow_state *flows; flows 2202 net/core/pktgen.c return !!(pkt_dev->flows[flow].flags & F_INIT); flows 2210 net/core/pktgen.c if (pkt_dev->flows[flow].count >= pkt_dev->lflow) { flows 2212 net/core/pktgen.c pkt_dev->flows[flow].count = 0; flows 2213 net/core/pktgen.c pkt_dev->flows[flow].flags = 0; flows 2222 net/core/pktgen.c if (pkt_dev->flows[flow].count > pkt_dev->lflow) { flows 2223 net/core/pktgen.c pkt_dev->flows[flow].count = 0; flows 2224 net/core/pktgen.c pkt_dev->flows[flow].flags = 0; flows 2239 net/core/pktgen.c struct xfrm_state *x = pkt_dev->flows[flow].x; flows 2258 net/core/pktgen.c pkt_dev->flows[flow].x = x; flows 2416 net/core/pktgen.c pkt_dev->cur_daddr = pkt_dev->flows[flow].cur_daddr; flows 2445 net/core/pktgen.c pkt_dev->flows[flow].flags |= F_INIT; flows 2446 net/core/pktgen.c pkt_dev->flows[flow].cur_daddr = flows 2487 net/core/pktgen.c pkt_dev->flows[flow].count++; flows 2499 net/core/pktgen.c struct xfrm_state *x = pkt_dev->flows[pkt_dev->curfl].x; flows 2542 net/core/pktgen.c struct xfrm_state *x = pkt_dev->flows[i].x; flows 2545 net/core/pktgen.c pkt_dev->flows[i].x = NULL; flows 2555 net/core/pktgen.c struct xfrm_state *x = pkt_dev->flows[pkt_dev->curfl].x; flows 3609 net/core/pktgen.c pkt_dev->flows = vzalloc_node(array_size(MAX_CFLOWS, flows 3612 net/core/pktgen.c if (pkt_dev->flows == NULL) { flows 3672 net/core/pktgen.c vfree(pkt_dev->flows); flows 3776 net/core/pktgen.c vfree(pkt_dev->flows); flows 3858 net/mac80211/cfg.c txqstats->flows = txqi->tin.flows; flows 508 net/mac80211/debugfs_netdev.c txqi->tin.flows, flows 174 net/mac80211/debugfs_sta.c txqi->tin.flows, flows 1320 net/mac80211/tx.c flow = &fq->flows[cvars - local->cvars]; flows 1364 net/mac80211/tx.c cvars = &local->cvars[flow - fq->flows]; flows 66 net/sched/sch_atm.c struct list_head flows; /* NB: "link" is also on this flows 78 net/sched/sch_atm.c list_for_each_entry(flow, &p->flows, list) { flows 355 net/sched/sch_atm.c list_for_each_entry(flow, &p->flows, list) { flows 393 net/sched/sch_atm.c list_for_each_entry(flow, &p->flows, list) { flows 477 net/sched/sch_atm.c list_for_each_entry(flow, &p->flows, list) { flows 548 net/sched/sch_atm.c INIT_LIST_HEAD(&p->flows); flows 550 net/sched/sch_atm.c list_add(&p->link.list, &p->flows); flows 576 net/sched/sch_atm.c list_for_each_entry(flow, &p->flows, list) flows 587 net/sched/sch_atm.c list_for_each_entry(flow, &p->flows, list) { flows 592 net/sched/sch_atm.c list_for_each_entry_safe(flow, tmp, &p->flows, list) { flows 150 net/sched/sch_cake.c struct cake_flow flows[CAKE_QUEUES]; flows 707 net/sched/sch_cake.c q->flows[reduced_hash].set)) { flows 725 net/sched/sch_cake.c if (!q->flows[outer_hash + k].set) { flows 740 net/sched/sch_cake.c if (!q->flows[outer_hash + k].set) { flows 752 net/sched/sch_cake.c if (q->flows[outer_hash + k].set == CAKE_SET_BULK) { flows 753 net/sched/sch_cake.c q->hosts[q->flows[reduced_hash].srchost].srchost_bulk_flow_count--; flows 754 net/sched/sch_cake.c q->hosts[q->flows[reduced_hash].dsthost].dsthost_bulk_flow_count--; flows 781 net/sched/sch_cake.c if (q->flows[reduced_hash].set == CAKE_SET_BULK) flows 783 net/sched/sch_cake.c q->flows[reduced_hash].srchost = srchost_idx; flows 804 net/sched/sch_cake.c if (q->flows[reduced_hash].set == CAKE_SET_BULK) flows 806 net/sched/sch_cake.c q->flows[reduced_hash].dsthost = dsthost_idx; flows 1485 net/sched/sch_cake.c flow = &b->flows[idx]; flows 1651 net/sched/sch_cake.c flow = &b->flows[idx]; flows 1852 net/sched/sch_cake.c struct cake_flow *flow = &b->flows[q->cur_flow]; flows 1990 net/sched/sch_cake.c q->cur_flow = flow - b->flows; flows 2712 net/sched/sch_cake.c struct cake_flow *flow = b->flows + j; flows 2959 net/sched/sch_cake.c flow = &b->flows[idx % CAKE_QUEUES]; flows 3031 net/sched/sch_cake.c if (list_empty(&b->flows[j].flowchain) || flows 113 net/sched/sch_fq.c u32 flows; flows 230 net/sched/sch_fq.c q->flows -= fcnt; flows 284 net/sched/sch_fq.c if (q->flows >= (2U << q->fq_trees_log) && flows 285 net/sched/sch_fq.c q->inactive_flows > q->flows/2) flows 339 net/sched/sch_fq.c q->flows++; flows 648 net/sched/sch_fq.c q->flows = 0; flows 693 net/sched/sch_fq.c q->flows -= fcnt; flows 952 net/sched/sch_fq.c st.flows = q->flows; flows 54 net/sched/sch_fq_codel.c struct fq_codel_flow *flows; /* Flows table [flows_cnt] */ flows 165 net/sched/sch_fq_codel.c flow = &q->flows[idx]; flows 205 net/sched/sch_fq_codel.c flow = &q->flows[idx]; flows 266 net/sched/sch_fq_codel.c q->backlogs[flow - q->flows] -= qdisc_pkt_len(skb); flows 344 net/sched/sch_fq_codel.c struct fq_codel_flow *flow = q->flows + i; flows 383 net/sched/sch_fq_codel.c if (q->flows) flows 447 net/sched/sch_fq_codel.c kvfree(q->flows); flows 479 net/sched/sch_fq_codel.c if (!q->flows) { flows 480 net/sched/sch_fq_codel.c q->flows = kvcalloc(q->flows_cnt, flows 483 net/sched/sch_fq_codel.c if (!q->flows) { flows 493 net/sched/sch_fq_codel.c struct fq_codel_flow *flow = q->flows + i; flows 506 net/sched/sch_fq_codel.c kvfree(q->flows); flows 507 net/sched/sch_fq_codel.c q->flows = NULL; flows 624 net/sched/sch_fq_codel.c const struct fq_codel_flow *flow = &q->flows[idx]; flows 671 net/sched/sch_fq_codel.c if (list_empty(&q->flows[i].flowchain) || flows 663 net/sched/sch_sfq.c if (ctl->flows) flows 664 net/sched/sch_sfq.c q->maxflows = min_t(u32, ctl->flows, SFQ_MAX_FLOWS); flows 800 net/sched/sch_sfq.c opt.v0.flows = q->maxflows; flows 1010 net/wireless/nl80211.c PUT_TXQVAL_U32(FLOWS, flows); flows 217 tools/include/uapi/linux/pkt_sched.h unsigned flows; /* Maximal number of flows */ flows 910 tools/include/uapi/linux/pkt_sched.h __u32 flows;