flows            3254 drivers/infiniband/core/uverbs_cmd.c 	flow_attr = kzalloc(struct_size(flow_attr, flows,
flows             342 drivers/infiniband/hw/hfi1/hfi.h 	struct tid_flow_state flows[RXE_NUM_TID_FLOWS];
flows             824 drivers/infiniband/hw/hfi1/rc.c 					&req->flows[req->setup_head];
flows            1114 drivers/infiniband/hw/hfi1/rc.c 		flow = &req->flows[req->flow_idx];
flows             752 drivers/infiniband/hw/hfi1/tid_rdma.c 	u32 generation = rcd->flows[flow_idx].generation;
flows             770 drivers/infiniband/hw/hfi1/tid_rdma.c 	rcd->flows[flow_idx].generation =
flows             771 drivers/infiniband/hw/hfi1/tid_rdma.c 		kern_flow_generation_next(rcd->flows[flow_idx].generation);
flows             799 drivers/infiniband/hw/hfi1/tid_rdma.c 		rcd->flows[fs->index].generation = fs->generation;
flows             848 drivers/infiniband/hw/hfi1/tid_rdma.c 		rcd->flows[i].generation = mask_generation(prandom_u32());
flows            1458 drivers/infiniband/hw/hfi1/tid_rdma.c 	struct tid_rdma_flow *flow = &req->flows[req->setup_head];
flows            1549 drivers/infiniband/hw/hfi1/tid_rdma.c 	struct tid_rdma_flow *flow = &req->flows[req->clear_tail];
flows            1605 drivers/infiniband/hw/hfi1/tid_rdma.c 	kfree(req->flows);
flows            1606 drivers/infiniband/hw/hfi1/tid_rdma.c 	req->flows = NULL;
flows            1627 drivers/infiniband/hw/hfi1/tid_rdma.c 	struct tid_rdma_flow *flows;
flows            1630 drivers/infiniband/hw/hfi1/tid_rdma.c 	if (likely(req->flows))
flows            1632 drivers/infiniband/hw/hfi1/tid_rdma.c 	flows = kmalloc_node(MAX_FLOWS * sizeof(*flows), gfp,
flows            1634 drivers/infiniband/hw/hfi1/tid_rdma.c 	if (!flows)
flows            1638 drivers/infiniband/hw/hfi1/tid_rdma.c 		flows[i].req = req;
flows            1639 drivers/infiniband/hw/hfi1/tid_rdma.c 		flows[i].npagesets = 0;
flows            1640 drivers/infiniband/hw/hfi1/tid_rdma.c 		flows[i].pagesets[0].mapped =  0;
flows            1641 drivers/infiniband/hw/hfi1/tid_rdma.c 		flows[i].resync_npkts = 0;
flows            1643 drivers/infiniband/hw/hfi1/tid_rdma.c 	req->flows = flows;
flows            1684 drivers/infiniband/hw/hfi1/tid_rdma.c 		flow = &req->flows[tail];
flows            1701 drivers/infiniband/hw/hfi1/tid_rdma.c 	struct tid_rdma_flow *flow = &req->flows[req->flow_idx];
flows            1857 drivers/infiniband/hw/hfi1/tid_rdma.c 	flow = &req->flows[req->flow_idx];
flows            1896 drivers/infiniband/hw/hfi1/tid_rdma.c 	flow = &req->flows[req->setup_head];
flows            2120 drivers/infiniband/hw/hfi1/tid_rdma.c 			fstate = &req->flows[req->clear_tail].flow_state;
flows            2344 drivers/infiniband/hw/hfi1/tid_rdma.c 	struct tid_rdma_flow *flow = &req->flows[req->clear_tail];
flows            2467 drivers/infiniband/hw/hfi1/tid_rdma.c 	flow = &req->flows[req->clear_tail];
flows            2631 drivers/infiniband/hw/hfi1/tid_rdma.c 	flow = &req->flows[req->clear_tail];
flows            2749 drivers/infiniband/hw/hfi1/tid_rdma.c 			flow = &req->flows[req->clear_tail];
flows            2937 drivers/infiniband/hw/hfi1/tid_rdma.c 	flow = &req->flows[req->clear_tail];
flows            3061 drivers/infiniband/hw/hfi1/tid_rdma.c 		flow = &req->flows[fidx];
flows            3133 drivers/infiniband/hw/hfi1/tid_rdma.c 				req->flows[fidx].sent = 0;
flows            3134 drivers/infiniband/hw/hfi1/tid_rdma.c 				req->flows[fidx].pkt = 0;
flows            3135 drivers/infiniband/hw/hfi1/tid_rdma.c 				req->flows[fidx].tid_idx = 0;
flows            3136 drivers/infiniband/hw/hfi1/tid_rdma.c 				req->flows[fidx].tid_offset = 0;
flows            3137 drivers/infiniband/hw/hfi1/tid_rdma.c 				req->flows[fidx].resync_npkts = 0;
flows            3839 drivers/infiniband/hw/hfi1/tid_rdma.c 	flow = &req->flows[req->flow_idx];
flows            4093 drivers/infiniband/hw/hfi1/tid_rdma.c 	flow = &req->flows[req->setup_head];
flows            4197 drivers/infiniband/hw/hfi1/tid_rdma.c 	struct tid_rdma_flow *flow = &req->flows[req->clear_tail];
flows            4282 drivers/infiniband/hw/hfi1/tid_rdma.c 	flow = &req->flows[req->clear_tail];
flows            4336 drivers/infiniband/hw/hfi1/tid_rdma.c 	rcd->flows[flow->idx].psn = psn & HFI1_KDETH_BTH_SEQ_MASK;
flows            4420 drivers/infiniband/hw/hfi1/tid_rdma.c 	struct tid_rdma_flow *flow = &req->flows[iflow];
flows            4532 drivers/infiniband/hw/hfi1/tid_rdma.c 	flow = &req->flows[req->acked_tail];
flows            4562 drivers/infiniband/hw/hfi1/tid_rdma.c 		flow = &req->flows[req->acked_tail];
flows            4624 drivers/infiniband/hw/hfi1/tid_rdma.c 			flow = &req->flows[req->acked_tail];
flows            4662 drivers/infiniband/hw/hfi1/tid_rdma.c 					flow = &rptr->flows[fidx];
flows            4708 drivers/infiniband/hw/hfi1/tid_rdma.c 			if (!req->flows)
flows            4710 drivers/infiniband/hw/hfi1/tid_rdma.c 			flow = &req->flows[req->acked_tail];
flows            4837 drivers/infiniband/hw/hfi1/tid_rdma.c 	struct tid_rdma_flow *flow = &req->flows[fidx];
flows            4899 drivers/infiniband/hw/hfi1/tid_rdma.c 		rcd->flows[fs->index].generation = generation;
flows            4934 drivers/infiniband/hw/hfi1/tid_rdma.c 				flow = &req->flows[flow_idx];
flows            5276 drivers/infiniband/hw/hfi1/tid_rdma.c 		      full_flow_psn(&req->flows[flow],
flows            5277 drivers/infiniband/hw/hfi1/tid_rdma.c 				    req->flows[flow].flow_state.lpsn)) > 0))) {
flows             105 drivers/infiniband/hw/hfi1/tid_rdma.h 	struct tid_rdma_flow *flows;	/* array of tid flows */
flows             163 drivers/net/ethernet/mellanox/mlx5/core/en_rep.h 	struct list_head flows;
flows             175 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	struct list_head flows;
flows             200 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	struct list_head flows;
flows             335 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	WARN_ON(!list_empty(&mh->flows));
flows             391 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	INIT_LIST_HEAD(&mh->flows);
flows             412 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	list_add(&flow->mod_hdr, &mh->flows);
flows             729 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	WARN_ON(!list_empty(&hpe->flows));
flows             819 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	INIT_LIST_HEAD(&hpe->flows);
flows             872 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	list_add(&flow->hairpin, &hpe->flows);
flows            1414 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	list_for_each_entry(efi, &e->flows, list) {
flows            1510 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 		list_for_each_entry_safe(efi, tmp, &e->flows, list) {
flows            1555 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	WARN_ON(!list_empty(&e->flows));
flows            3054 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	INIT_LIST_HEAD(&e->flows);
flows            3074 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c 	list_add(&flow->encaps[out_index].list, &e->flows);
flows             685 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 	struct mlx5_flow_handle **flows;
flows             699 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 	flows = kvzalloc(nvports * sizeof(*flows), GFP_KERNEL);
flows             700 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 	if (!flows) {
flows             719 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 		flows[MLX5_VPORT_PF] = flow;
flows             730 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 		flows[mlx5_eswitch_ecpf_idx(esw)] = flow;
flows             744 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 		flows[i] = flow;
flows             747 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 	esw->fdb_table.offloads.peer_miss_rules = flows;
flows             755 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 		mlx5_del_flow_rules(flows[i]);
flows             758 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 		mlx5_del_flow_rules(flows[mlx5_eswitch_ecpf_idx(esw)]);
flows             761 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 		mlx5_del_flow_rules(flows[MLX5_VPORT_PF]);
flows             764 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 	kvfree(flows);
flows             772 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 	struct mlx5_flow_handle **flows;
flows             775 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 	flows = esw->fdb_table.offloads.peer_miss_rules;
flows             779 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 		mlx5_del_flow_rules(flows[i]);
flows             782 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 		mlx5_del_flow_rules(flows[mlx5_eswitch_ecpf_idx(esw)]);
flows             785 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 		mlx5_del_flow_rules(flows[MLX5_VPORT_PF]);
flows             787 drivers/net/ethernet/mellanox/mlx5/core/eswitch_offloads.c 	kvfree(flows);
flows             232 drivers/net/tun.c 	struct hlist_head flows[TUN_NUM_FLOW_ENTRIES];
flows             457 drivers/net/tun.c 		hlist_for_each_entry_safe(e, n, &tun->flows[i], hash_link)
flows             472 drivers/net/tun.c 		hlist_for_each_entry_safe(e, n, &tun->flows[i], hash_link) {
flows             495 drivers/net/tun.c 		hlist_for_each_entry_safe(e, n, &tun->flows[i], hash_link) {
flows             522 drivers/net/tun.c 	head = &tun->flows[tun_hashfn(rxhash)];
flows             574 drivers/net/tun.c 	e = tun_flow_find(&tun->flows[tun_hashfn(txq)], txq);
flows            1042 drivers/net/tun.c 		e = tun_flow_find(&tun->flows[tun_hashfn(rxhash)], rxhash);
flows            1364 drivers/net/tun.c 		INIT_HLIST_HEAD(&tun->flows[i]);
flows             689 include/linux/netdevice.h 	struct rps_dev_flow flows[0];
flows            1373 include/net/cfg80211.h 	u32 flows;
flows              54 include/net/fq.h 	u32 flows;
flows              68 include/net/fq.h 	struct fq_flow *flows;
flows             125 include/net/fq_impl.h 	flow = &fq->flows[idx];
flows             133 include/net/fq_impl.h 		tin->flows++;
flows             316 include/net/fq_impl.h 	fq->flows = kvcalloc(fq->flows_cnt, sizeof(fq->flows[0]), GFP_KERNEL);
flows             317 include/net/fq_impl.h 	if (!fq->flows)
flows             321 include/net/fq_impl.h 		fq_flow_init(&fq->flows[i]);
flows             332 include/net/fq_impl.h 		fq_flow_reset(fq, &fq->flows[i], free_func);
flows             334 include/net/fq_impl.h 	kvfree(fq->flows);
flows             335 include/net/fq_impl.h 	fq->flows = NULL;
flows            2056 include/rdma/ib_verbs.h 	union ib_flow_spec flows[];
flows             218 include/uapi/linux/pkt_sched.h 	unsigned	flows;		/* Maximal number of flows  */
flows             911 include/uapi/linux/pkt_sched.h 	__u32	flows;
flows            3928 net/core/dev.c 		rflow = &flow_table->flows[flow_id];
flows            3999 net/core/dev.c 		rflow = &flow_table->flows[hash & flow_table->mask];
flows            4067 net/core/dev.c 		rflow = &flow_table->flows[flow_id];
flows             830 net/core/net-sysfs.c 			table->flows[count].cpu = RPS_NO_CPU;
flows             399 net/core/pktgen.c 	struct flow_state *flows;
flows            2202 net/core/pktgen.c 	return !!(pkt_dev->flows[flow].flags & F_INIT);
flows            2210 net/core/pktgen.c 		if (pkt_dev->flows[flow].count >= pkt_dev->lflow) {
flows            2212 net/core/pktgen.c 			pkt_dev->flows[flow].count = 0;
flows            2213 net/core/pktgen.c 			pkt_dev->flows[flow].flags = 0;
flows            2222 net/core/pktgen.c 		if (pkt_dev->flows[flow].count > pkt_dev->lflow) {
flows            2223 net/core/pktgen.c 			pkt_dev->flows[flow].count = 0;
flows            2224 net/core/pktgen.c 			pkt_dev->flows[flow].flags = 0;
flows            2239 net/core/pktgen.c 	struct xfrm_state *x = pkt_dev->flows[flow].x;
flows            2258 net/core/pktgen.c 			pkt_dev->flows[flow].x = x;
flows            2416 net/core/pktgen.c 			pkt_dev->cur_daddr = pkt_dev->flows[flow].cur_daddr;
flows            2445 net/core/pktgen.c 				pkt_dev->flows[flow].flags |= F_INIT;
flows            2446 net/core/pktgen.c 				pkt_dev->flows[flow].cur_daddr =
flows            2487 net/core/pktgen.c 	pkt_dev->flows[flow].count++;
flows            2499 net/core/pktgen.c 	struct xfrm_state *x = pkt_dev->flows[pkt_dev->curfl].x;
flows            2542 net/core/pktgen.c 			struct xfrm_state *x = pkt_dev->flows[i].x;
flows            2545 net/core/pktgen.c 				pkt_dev->flows[i].x = NULL;
flows            2555 net/core/pktgen.c 		struct xfrm_state *x = pkt_dev->flows[pkt_dev->curfl].x;
flows            3609 net/core/pktgen.c 	pkt_dev->flows = vzalloc_node(array_size(MAX_CFLOWS,
flows            3612 net/core/pktgen.c 	if (pkt_dev->flows == NULL) {
flows            3672 net/core/pktgen.c 	vfree(pkt_dev->flows);
flows            3776 net/core/pktgen.c 	vfree(pkt_dev->flows);
flows            3858 net/mac80211/cfg.c 		txqstats->flows = txqi->tin.flows;
flows             508 net/mac80211/debugfs_netdev.c 			txqi->tin.flows,
flows             174 net/mac80211/debugfs_sta.c 			       txqi->tin.flows,
flows            1320 net/mac80211/tx.c 		flow = &fq->flows[cvars - local->cvars];
flows            1364 net/mac80211/tx.c 		cvars = &local->cvars[flow - fq->flows];
flows              66 net/sched/sch_atm.c 	struct list_head	flows;		/* NB: "link" is also on this
flows              78 net/sched/sch_atm.c 	list_for_each_entry(flow, &p->flows, list) {
flows             355 net/sched/sch_atm.c 	list_for_each_entry(flow, &p->flows, list) {
flows             393 net/sched/sch_atm.c 		list_for_each_entry(flow, &p->flows, list) {
flows             477 net/sched/sch_atm.c 	list_for_each_entry(flow, &p->flows, list) {
flows             548 net/sched/sch_atm.c 	INIT_LIST_HEAD(&p->flows);
flows             550 net/sched/sch_atm.c 	list_add(&p->link.list, &p->flows);
flows             576 net/sched/sch_atm.c 	list_for_each_entry(flow, &p->flows, list)
flows             587 net/sched/sch_atm.c 	list_for_each_entry(flow, &p->flows, list) {
flows             592 net/sched/sch_atm.c 	list_for_each_entry_safe(flow, tmp, &p->flows, list) {
flows             150 net/sched/sch_cake.c 	struct cake_flow flows[CAKE_QUEUES];
flows             707 net/sched/sch_cake.c 		   q->flows[reduced_hash].set)) {
flows             725 net/sched/sch_cake.c 				if (!q->flows[outer_hash + k].set) {
flows             740 net/sched/sch_cake.c 			if (!q->flows[outer_hash + k].set) {
flows             752 net/sched/sch_cake.c 		if (q->flows[outer_hash + k].set == CAKE_SET_BULK) {
flows             753 net/sched/sch_cake.c 			q->hosts[q->flows[reduced_hash].srchost].srchost_bulk_flow_count--;
flows             754 net/sched/sch_cake.c 			q->hosts[q->flows[reduced_hash].dsthost].dsthost_bulk_flow_count--;
flows             781 net/sched/sch_cake.c 			if (q->flows[reduced_hash].set == CAKE_SET_BULK)
flows             783 net/sched/sch_cake.c 			q->flows[reduced_hash].srchost = srchost_idx;
flows             804 net/sched/sch_cake.c 			if (q->flows[reduced_hash].set == CAKE_SET_BULK)
flows             806 net/sched/sch_cake.c 			q->flows[reduced_hash].dsthost = dsthost_idx;
flows            1485 net/sched/sch_cake.c 	flow = &b->flows[idx];
flows            1651 net/sched/sch_cake.c 	flow = &b->flows[idx];
flows            1852 net/sched/sch_cake.c 	struct cake_flow *flow = &b->flows[q->cur_flow];
flows            1990 net/sched/sch_cake.c 	q->cur_flow = flow - b->flows;
flows            2712 net/sched/sch_cake.c 			struct cake_flow *flow = b->flows + j;
flows            2959 net/sched/sch_cake.c 		flow = &b->flows[idx % CAKE_QUEUES];
flows            3031 net/sched/sch_cake.c 			if (list_empty(&b->flows[j].flowchain) ||
flows             113 net/sched/sch_fq.c 	u32		flows;
flows             230 net/sched/sch_fq.c 	q->flows -= fcnt;
flows             284 net/sched/sch_fq.c 	if (q->flows >= (2U << q->fq_trees_log) &&
flows             285 net/sched/sch_fq.c 	    q->inactive_flows > q->flows/2)
flows             339 net/sched/sch_fq.c 	q->flows++;
flows             648 net/sched/sch_fq.c 	q->flows		= 0;
flows             693 net/sched/sch_fq.c 	q->flows -= fcnt;
flows             952 net/sched/sch_fq.c 	st.flows		  = q->flows;
flows              54 net/sched/sch_fq_codel.c 	struct fq_codel_flow *flows;	/* Flows table [flows_cnt] */
flows             165 net/sched/sch_fq_codel.c 	flow = &q->flows[idx];
flows             205 net/sched/sch_fq_codel.c 	flow = &q->flows[idx];
flows             266 net/sched/sch_fq_codel.c 		q->backlogs[flow - q->flows] -= qdisc_pkt_len(skb);
flows             344 net/sched/sch_fq_codel.c 		struct fq_codel_flow *flow = q->flows + i;
flows             383 net/sched/sch_fq_codel.c 		if (q->flows)
flows             447 net/sched/sch_fq_codel.c 	kvfree(q->flows);
flows             479 net/sched/sch_fq_codel.c 	if (!q->flows) {
flows             480 net/sched/sch_fq_codel.c 		q->flows = kvcalloc(q->flows_cnt,
flows             483 net/sched/sch_fq_codel.c 		if (!q->flows) {
flows             493 net/sched/sch_fq_codel.c 			struct fq_codel_flow *flow = q->flows + i;
flows             506 net/sched/sch_fq_codel.c 	kvfree(q->flows);
flows             507 net/sched/sch_fq_codel.c 	q->flows = NULL;
flows             624 net/sched/sch_fq_codel.c 		const struct fq_codel_flow *flow = &q->flows[idx];
flows             671 net/sched/sch_fq_codel.c 		if (list_empty(&q->flows[i].flowchain) ||
flows             663 net/sched/sch_sfq.c 	if (ctl->flows)
flows             664 net/sched/sch_sfq.c 		q->maxflows = min_t(u32, ctl->flows, SFQ_MAX_FLOWS);
flows             800 net/sched/sch_sfq.c 	opt.v0.flows	= q->maxflows;
flows            1010 net/wireless/nl80211.c 	PUT_TXQVAL_U32(FLOWS, flows);
flows             217 tools/include/uapi/linux/pkt_sched.h 	unsigned	flows;		/* Maximal number of flows  */
flows             910 tools/include/uapi/linux/pkt_sched.h 	__u32	flows;