tci              3563 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 	struct virtchnl_tc_info *tci =
tci              3593 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 	if (!tci->num_tc || tci->num_tc > I40E_MAX_VF_VSI) {
tci              3596 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 			vf->vf_id, tci->num_tc, I40E_MAX_VF_VSI);
tci              3602 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 	for (i = 0; i < tci->num_tc; i++)
tci              3603 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 		if (!tci->list[i].count ||
tci              3604 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 		    tci->list[i].count > I40E_DEFAULT_QUEUES_PER_VF) {
tci              3607 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 				vf->vf_id, i, tci->list[i].count,
tci              3658 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 	vf->num_tc = tci->num_tc;
tci              3660 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 		if (tci->list[i].max_tx_rate) {
tci              3661 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 			if (tci->list[i].max_tx_rate > speed) {
tci              3664 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 					tci->list[i].max_tx_rate,
tci              3670 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 					tci->list[i].max_tx_rate;
tci              3673 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c 		vf->ch[i].num_qps = tci->list[i].count;
tci               271 drivers/net/ethernet/netronome/nfp/flower/cmsg.h 	__be16 tci;
tci                38 drivers/net/ethernet/netronome/nfp/flower/match.c 		ext->tci = cpu_to_be16(tmp_tci);
tci                45 drivers/net/ethernet/netronome/nfp/flower/match.c 		msk->tci = cpu_to_be16(tmp_tci);
tci                72 drivers/net/ethernet/netronome/nfp/flower/offload.c 			__be16 tci;
tci               540 drivers/net/ethernet/netronome/nfp/flower/offload.c 				merge->tci = cpu_to_be16(0xffff);
tci               543 drivers/net/ethernet/netronome/nfp/flower/offload.c 			merge->tci = cpu_to_be16(0);
tci               633 drivers/net/ethernet/netronome/nfp/flower/offload.c 	merge->tci = meta_tci->tci;
tci              1029 drivers/net/ethernet/netronome/nfp/flower/offload.c 	if (meta_tci->tci & cpu_to_be16(NFP_FLOWER_MASK_VLAN_PRESENT)) {
tci              1030 drivers/net/ethernet/netronome/nfp/flower/offload.c 		u16 vlan_tci = be16_to_cpu(meta_tci->tci);
tci              1081 drivers/net/ethernet/qlogic/qed/qed_hsi.h 	__le16 tci;
tci               365 drivers/net/ethernet/qlogic/qed/qed_sp_commands.c 	p_ramrod->outer_tag_config.outer_tag.tci =
tci               387 drivers/net/ethernet/qlogic/qed/qed_sp_commands.c 		p_ramrod->outer_tag_config.outer_tag.tci |=
tci               435 drivers/net/ethernet/qlogic/qed/qed_sp_commands.c 		   sb, sb_index, p_ramrod->outer_tag_config.outer_tag.tci);
tci               324 drivers/net/ethernet/sfc/siena_sriov.c 	vlan = ntohs(vf->addr.tci) & VLAN_VID_MASK;
tci               362 drivers/net/ethernet/sfc/siena_sriov.c 	vlan = ntohs(vf->addr.tci) & VLAN_VID_MASK;
tci              1119 drivers/net/ethernet/sfc/siena_sriov.c 		peer->tci = 0;
tci              1615 drivers/net/ethernet/sfc/siena_sriov.c 	u16 tci;
tci              1622 drivers/net/ethernet/sfc/siena_sriov.c 	tci = (vlan & VLAN_VID_MASK) | ((qos & 0x7) << VLAN_PRIO_SHIFT);
tci              1623 drivers/net/ethernet/sfc/siena_sriov.c 	vf->addr.tci = htons(tci);
tci              1659 drivers/net/ethernet/sfc/siena_sriov.c 	u16 tci;
tci              1669 drivers/net/ethernet/sfc/siena_sriov.c 	tci = ntohs(vf->addr.tci);
tci              1670 drivers/net/ethernet/sfc/siena_sriov.c 	ivi->vlan = tci & VLAN_VID_MASK;
tci              1671 drivers/net/ethernet/sfc/siena_sriov.c 	ivi->qos = (tci >> VLAN_PRIO_SHIFT) & 0x7;
tci                87 drivers/net/ethernet/sfc/vfdi.h 	__be16 tci;
tci              1187 drivers/net/usb/aqc111.c 	u16 tci = 0;
tci              1206 drivers/net/usb/aqc111.c 	if (vlan_get_tag(skb, &tci) >= 0) {
tci              1208 drivers/net/usb/aqc111.c 		tx_desc |= ((u64)tci & AQ_TX_DESC_VLAN_MASK) <<
tci               225 drivers/net/usb/cdc_mbim.c 	u16 tci = 0;
tci               243 drivers/net/usb/cdc_mbim.c 		if (vlan_get_tag(skb, &tci) < 0 && skb->len > VLAN_ETH_HLEN &&
tci               244 drivers/net/usb/cdc_mbim.c 		    __vlan_get_tag(skb, &tci) == 0) {
tci               255 drivers/net/usb/cdc_mbim.c 			if (!tci)
tci               258 drivers/net/usb/cdc_mbim.c 			if (tci == MBIM_IPS0_VID)
tci               259 drivers/net/usb/cdc_mbim.c 				tci = 0;
tci               270 drivers/net/usb/cdc_mbim.c 		switch (tci & 0x0f00) {
tci               275 drivers/net/usb/cdc_mbim.c 			c[3] = tci;
tci               282 drivers/net/usb/cdc_mbim.c 			c[3] = tci;
tci               286 drivers/net/usb/cdc_mbim.c 				  "unsupported tci=0x%04x\n", tci);
tci               307 drivers/net/usb/cdc_mbim.c static void do_neigh_solicit(struct usbnet *dev, u8 *buf, u16 tci)
tci               324 drivers/net/usb/cdc_mbim.c 	if (tci) {
tci               326 drivers/net/usb/cdc_mbim.c 						  tci);
tci               365 drivers/net/usb/cdc_mbim.c static struct sk_buff *cdc_mbim_process_dgram(struct usbnet *dev, u8 *buf, size_t len, u16 tci)
tci               370 drivers/net/usb/cdc_mbim.c 	if (tci < 256 || tci == MBIM_IPS0_VID) { /* IPS session? */
tci               380 drivers/net/usb/cdc_mbim.c 				do_neigh_solicit(dev, buf, tci);
tci               403 drivers/net/usb/cdc_mbim.c 	if (tci)
tci               404 drivers/net/usb/cdc_mbim.c 		__vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q), tci);
tci               424 drivers/net/usb/cdc_mbim.c 	u16 tci;
tci               440 drivers/net/usb/cdc_mbim.c 		tci = c[3];
tci               442 drivers/net/usb/cdc_mbim.c 		if (!tci && info->flags & FLAG_IPS0_VLAN)
tci               443 drivers/net/usb/cdc_mbim.c 			tci = MBIM_IPS0_VID;
tci               447 drivers/net/usb/cdc_mbim.c 		tci = c[3] + 256;
tci               481 drivers/net/usb/cdc_mbim.c 			skb = cdc_mbim_process_dgram(dev, skb_in->data + offset, len, tci);
tci               139 drivers/net/vmxnet3/vmxnet3_defs.h 	u32 tci:16;     /* Tag to Insert */
tci               153 drivers/net/vmxnet3/vmxnet3_defs.h 	u32 tci:16;     /* Tag to Insert */
tci               251 drivers/net/vmxnet3/vmxnet3_defs.h 	u32		tci:16;       /* Tag stripped */
tci               259 drivers/net/vmxnet3/vmxnet3_defs.h 	u32		tci:16;       /* Tag stripped */
tci              1107 drivers/net/vmxnet3/vmxnet3_drv.c 		gdesc->txd.tci = skb_vlan_tag_get(skb);
tci              1530 drivers/net/vmxnet3/vmxnet3_drv.c 				__vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q), rcd->tci);
tci              2551 drivers/scsi/isci/host.c void isci_tci_free(struct isci_host *ihost, u16 tci)
tci              2555 drivers/scsi/isci/host.c 	ihost->tci_pool[tail] = tci;
tci              2562 drivers/scsi/isci/host.c 	u16 tci = ihost->tci_pool[head];
tci              2565 drivers/scsi/isci/host.c 	return tci;
tci              2576 drivers/scsi/isci/host.c 		u16 tci = isci_tci_alloc(ihost);
tci              2577 drivers/scsi/isci/host.c 		u8 seq = ihost->io_request_sequence[tci];
tci              2579 drivers/scsi/isci/host.c 		return ISCI_TAG(seq, tci);
tci              2587 drivers/scsi/isci/host.c 	u16 tci = ISCI_TAG_TCI(io_tag);
tci              2594 drivers/scsi/isci/host.c 	if (seq == ihost->io_request_sequence[tci]) {
tci              2595 drivers/scsi/isci/host.c 		ihost->io_request_sequence[tci] = (seq+1) & (SCI_MAX_SEQ-1);
tci              2597 drivers/scsi/isci/host.c 		isci_tci_free(ihost, tci);
tci               350 drivers/scsi/isci/host.h #define ISCI_TAG(seq, tci) (((u16) (seq)) << 12 | tci)
tci               477 drivers/scsi/isci/host.h void isci_tci_free(struct isci_host *ihost, u16 tci);
tci               239 drivers/soc/fsl/qe/ucc_fast.c 	if (uf_info->tci)
tci               277 drivers/soc/fsl/qe/ucc_slow.c 	if (us_info->tci)
tci                21 include/linux/dsa/8021q.h 			       u16 tpid, u16 tci);
tci                40 include/linux/dsa/8021q.h 			       u16 tpid, u16 tci)
tci               164 include/net/erspan.h 		__be16 tci;
tci               181 include/net/erspan.h 		vlan_tci = ntohs(qp->tci);
tci               258 include/net/erspan.h 		__be16 tci;
tci               275 include/net/erspan.h 		vlan_tci = ntohs(qp->tci);
tci               157 include/soc/fsl/qe/ucc_fast.h 	int tci;
tci               170 include/soc/fsl/qe/ucc_slow.h 	int tci;
tci              2071 net/core/dev.c 			     int tci, u16 index)
tci              2077 net/core/dev.c 		map = xmap_dereference(dev_maps->attr_map[tci]);
tci              2090 net/core/dev.c 		RCU_INIT_POINTER(dev_maps->attr_map[tci], NULL);
tci              2104 net/core/dev.c 	int tci;
tci              2106 net/core/dev.c 	for (tci = cpu * num_tc; num_tc--; tci++) {
tci              2110 net/core/dev.c 			if (!remove_xps_queue(dev_maps, tci, j))
tci              2245 net/core/dev.c 	int i, j, tci, numa_node_id = -2;
tci              2293 net/core/dev.c 		tci = j * num_tc + tc;
tci              2294 net/core/dev.c 		map = dev_maps ? xmap_dereference(dev_maps->attr_map[tci]) :
tci              2301 net/core/dev.c 		RCU_INIT_POINTER(new_dev_maps->attr_map[tci], map);
tci              2317 net/core/dev.c 		for (i = tc, tci = j * num_tc; dev_maps && i--; tci++) {
tci              2319 net/core/dev.c 			map = xmap_dereference(dev_maps->attr_map[tci]);
tci              2320 net/core/dev.c 			RCU_INIT_POINTER(new_dev_maps->attr_map[tci], map);
tci              2326 net/core/dev.c 		tci = j * num_tc + tc;
tci              2333 net/core/dev.c 			map = xmap_dereference(new_dev_maps->attr_map[tci]);
tci              2349 net/core/dev.c 			map = xmap_dereference(dev_maps->attr_map[tci]);
tci              2350 net/core/dev.c 			RCU_INIT_POINTER(new_dev_maps->attr_map[tci], map);
tci              2354 net/core/dev.c 		for (i = num_tc - tc, tci++; dev_maps && --i; tci++) {
tci              2356 net/core/dev.c 			map = xmap_dereference(dev_maps->attr_map[tci]);
tci              2357 net/core/dev.c 			RCU_INIT_POINTER(new_dev_maps->attr_map[tci], map);
tci              2372 net/core/dev.c 		for (i = num_tc, tci = j * num_tc; i--; tci++) {
tci              2373 net/core/dev.c 			new_map = xmap_dereference(new_dev_maps->attr_map[tci]);
tci              2374 net/core/dev.c 			map = xmap_dereference(dev_maps->attr_map[tci]);
tci              2400 net/core/dev.c 		for (i = tc, tci = j * num_tc; i--; tci++)
tci              2401 net/core/dev.c 			active |= remove_xps_queue(dev_maps, tci, index);
tci              2404 net/core/dev.c 			active |= remove_xps_queue(dev_maps, tci, index);
tci              2405 net/core/dev.c 		for (i = num_tc - tc, tci++; --i; tci++)
tci              2406 net/core/dev.c 			active |= remove_xps_queue(dev_maps, tci, index);
tci              2421 net/core/dev.c 		for (i = num_tc, tci = j * num_tc; i--; tci++) {
tci              2422 net/core/dev.c 			new_map = xmap_dereference(new_dev_maps->attr_map[tci]);
tci              2424 net/core/dev.c 			      xmap_dereference(dev_maps->attr_map[tci]) :
tci              3539 net/core/dev.c 			       struct xps_dev_maps *dev_maps, unsigned int tci)
tci              3545 net/core/dev.c 		tci *= dev->num_tc;
tci              3546 net/core/dev.c 		tci += netdev_get_prio_tc_map(dev, skb->priority);
tci              3549 net/core/dev.c 	map = rcu_dereference(dev_maps->attr_map[tci]);
tci              3580 net/core/dev.c 		int tci = sk_rx_queue_get(sk);
tci              3582 net/core/dev.c 		if (tci >= 0 && tci < dev->num_rx_queues)
tci              3584 net/core/dev.c 							  tci);
tci              3591 net/core/dev.c 			unsigned int tci = skb->sender_cpu - 1;
tci              3594 net/core/dev.c 							  tci);
tci              1270 net/core/net-sysfs.c 			int i, tci = cpu * num_tc + tc;
tci              1273 net/core/net-sysfs.c 			map = rcu_dereference(dev_maps->attr_map[tci]);
tci              1353 net/core/net-sysfs.c 		int i, tci = j * num_tc + tc;
tci              1356 net/core/net-sysfs.c 		map = rcu_dereference(dev_maps->attr_map[tci]);
tci               293 net/dsa/tag_8021q.c 			       u16 tpid, u16 tci)
tci               298 net/dsa/tag_8021q.c 	return vlan_insert_tag(skb, htons(tpid), tci);
tci               242 net/dsa/tag_sja1105.c 	u16 tpid, vid, tci;
tci               258 net/dsa/tag_sja1105.c 		__skb_vlan_pop(skb, &tci);
tci               263 net/dsa/tag_sja1105.c 		vid = tci & VLAN_VID_MASK;
tci               266 net/dsa/tag_sja1105.c 		skb->priority = (tci & VLAN_PRIO_MASK) >> VLAN_PRIO_SHIFT;
tci               217 net/openvswitch/actions.c 		key->eth.vlan.tci = 0;
tci               229 net/openvswitch/actions.c 		key->eth.vlan.tci = vlan->vlan_tci;
tci               317 net/openvswitch/flow.c 	key_vh->tci = vh->tci | htons(VLAN_CFI_MASK);
tci               322 net/openvswitch/flow.c 		u16 tci;
tci               326 net/openvswitch/flow.c 		err = __skb_vlan_pop(skb, &tci);
tci               330 net/openvswitch/flow.c 		__vlan_hwaccel_put_tag(skb, key_vh->tpid, tci);
tci               339 net/openvswitch/flow.c 	key->eth.vlan.tci = 0;
tci               341 net/openvswitch/flow.c 	key->eth.cvlan.tci = 0;
tci               350 net/openvswitch/flow.c 		key->eth.vlan.tci = htons(skb->vlan_tci) | htons(VLAN_CFI_MASK);
tci               794 net/openvswitch/flow.c 		if (key->eth.cvlan.tci & htons(VLAN_CFI_MASK))
tci                49 net/openvswitch/flow.h 	__be16 tci;  /* 0 if no VLAN, VLAN_CFI_MASK set otherwise. */
tci               964 net/openvswitch/flow_netlink.c 	__be16 tci = 0;
tci               968 net/openvswitch/flow_netlink.c 		tci = nla_get_be16(a[OVS_KEY_ATTR_VLAN]);
tci               975 net/openvswitch/flow_netlink.c 		SW_FLOW_KEY_PUT(match, eth.vlan.tci, tci, is_mask);
tci               978 net/openvswitch/flow_netlink.c 		SW_FLOW_KEY_PUT(match, eth.cvlan.tci, tci, is_mask);
tci               987 net/openvswitch/flow_netlink.c 	__be16 tci = 0;
tci              1003 net/openvswitch/flow_netlink.c 		tci = nla_get_be16(a[OVS_KEY_ATTR_VLAN]);
tci              1005 net/openvswitch/flow_netlink.c 	if (!(tci & htons(VLAN_CFI_MASK))) {
tci              1006 net/openvswitch/flow_netlink.c 		if (tci) {
tci              1025 net/openvswitch/flow_netlink.c 	__be16 tci = 0;
tci              1027 net/openvswitch/flow_netlink.c 	bool encap_valid = !!(match->key->eth.vlan.tci &
tci              1029 net/openvswitch/flow_netlink.c 	bool i_encap_valid = !!(match->key->eth.cvlan.tci &
tci              1044 net/openvswitch/flow_netlink.c 		tci = nla_get_be16(a[OVS_KEY_ATTR_VLAN]);
tci              1054 net/openvswitch/flow_netlink.c 	if (!(tci & htons(VLAN_CFI_MASK))) {
tci              1110 net/openvswitch/flow_netlink.c 	encap_valid = !!(match->key->eth.vlan.tci & htons(VLAN_CFI_MASK));
tci              1816 net/openvswitch/flow_netlink.c 		SW_FLOW_KEY_PUT(match, eth.vlan.tci, htons(0xffff), true);
tci              1817 net/openvswitch/flow_netlink.c 		SW_FLOW_KEY_PUT(match, eth.cvlan.tci, htons(0xffff), true);
tci              1937 net/openvswitch/flow_netlink.c 	    nla_put_be16(skb, OVS_KEY_ATTR_VLAN, vh->tci))
tci              2027 net/openvswitch/flow_netlink.c 		if (swkey->eth.vlan.tci || eth_type_vlan(swkey->eth.type)) {
tci              2031 net/openvswitch/flow_netlink.c 			if (!swkey->eth.vlan.tci)
tci              2034 net/openvswitch/flow_netlink.c 			if (swkey->eth.cvlan.tci || eth_type_vlan(swkey->eth.type)) {
tci              2039 net/openvswitch/flow_netlink.c 				if (!swkey->eth.cvlan.tci)
tci              3229 net/openvswitch/flow_netlink.c 				     key->eth.vlan.tci, log);
tci                29 net/sched/act_vlan.c 	u16 tci;
tci                62 net/sched/act_vlan.c 			tci = skb_vlan_tag_get(skb);
tci                66 net/sched/act_vlan.c 			err = __skb_vlan_pop(skb, &tci);
tci                71 net/sched/act_vlan.c 		tci = (tci & ~VLAN_VID_MASK) | p->tcfv_push_vid;
tci                74 net/sched/act_vlan.c 			tci &= ~VLAN_PRIO_MASK;
tci                75 net/sched/act_vlan.c 			tci |= p->tcfv_push_prio << VLAN_PRIO_SHIFT;
tci                78 net/sched/act_vlan.c 		__vlan_hwaccel_put_tag(skb, p->tcfv_push_proto, tci);