q_vector 104 drivers/net/ethernet/intel/fm10k/fm10k.h struct fm10k_q_vector *q_vector;/* backpointer to host q_vector */ q_vector 343 drivers/net/ethernet/intel/fm10k/fm10k.h struct fm10k_q_vector *q_vector[MAX_Q_VECTORS]; q_vector 548 drivers/net/ethernet/intel/fm10k/fm10k.h void fm10k_dbg_q_vector_init(struct fm10k_q_vector *q_vector); q_vector 549 drivers/net/ethernet/intel/fm10k/fm10k.h void fm10k_dbg_q_vector_exit(struct fm10k_q_vector *q_vector); q_vector 555 drivers/net/ethernet/intel/fm10k/fm10k.h static inline void fm10k_dbg_q_vector_init(struct fm10k_q_vector *q_vector) {} q_vector 556 drivers/net/ethernet/intel/fm10k/fm10k.h static inline void fm10k_dbg_q_vector_exit(struct fm10k_q_vector *q_vector) {} q_vector 116 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c struct fm10k_q_vector *q_vector = ring->q_vector; q_vector 120 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c if (ring < q_vector->rx.ring) q_vector 150 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c void fm10k_dbg_q_vector_init(struct fm10k_q_vector *q_vector) q_vector 152 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c struct fm10k_intfc *interface = q_vector->interface; q_vector 160 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c snprintf(name, sizeof(name), "q_vector.%03d", q_vector->v_idx); q_vector 162 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c q_vector->dbg_q_vector = debugfs_create_dir(name, interface->dbg_intfc); q_vector 165 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c for (i = 0; i < q_vector->tx.count; i++) { q_vector 166 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c struct fm10k_ring *ring = &q_vector->tx.ring[i]; q_vector 171 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c q_vector->dbg_q_vector, ring, q_vector 176 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c for (i = 0; i < q_vector->rx.count; i++) { q_vector 177 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c struct fm10k_ring *ring = &q_vector->rx.ring[i]; q_vector 182 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c q_vector->dbg_q_vector, ring, q_vector 191 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c void fm10k_dbg_q_vector_exit(struct fm10k_q_vector *q_vector) q_vector 193 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c struct fm10k_intfc *interface = q_vector->interface; q_vector 196 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c debugfs_remove_recursive(q_vector->dbg_q_vector); q_vector 197 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c q_vector->dbg_q_vector = NULL; q_vector 678 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c struct fm10k_q_vector *qv = interface->q_vector[i]; q_vector 322 drivers/net/ethernet/intel/fm10k/fm10k_main.c skb = napi_alloc_skb(&rx_ring->q_vector->napi, q_vector 559 drivers/net/ethernet/intel/fm10k/fm10k_main.c static void fm10k_receive_skb(struct fm10k_q_vector *q_vector, q_vector 562 drivers/net/ethernet/intel/fm10k/fm10k_main.c napi_gro_receive(&q_vector->napi, skb); q_vector 565 drivers/net/ethernet/intel/fm10k/fm10k_main.c static int fm10k_clean_rx_irq(struct fm10k_q_vector *q_vector, q_vector 615 drivers/net/ethernet/intel/fm10k/fm10k_main.c fm10k_receive_skb(q_vector, skb); q_vector 631 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->rx.total_packets += total_packets; q_vector 632 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->rx.total_bytes += total_bytes; q_vector 1125 drivers/net/ethernet/intel/fm10k/fm10k_main.c struct fm10k_intfc *interface = ring->q_vector->interface; q_vector 1189 drivers/net/ethernet/intel/fm10k/fm10k_main.c static bool fm10k_clean_tx_irq(struct fm10k_q_vector *q_vector, q_vector 1192 drivers/net/ethernet/intel/fm10k/fm10k_main.c struct fm10k_intfc *interface = q_vector->interface; q_vector 1196 drivers/net/ethernet/intel/fm10k/fm10k_main.c unsigned int budget = q_vector->tx.work_limit; q_vector 1284 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->tx.total_bytes += total_bytes; q_vector 1285 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->tx.total_packets += total_packets; q_vector 1410 drivers/net/ethernet/intel/fm10k/fm10k_main.c static void fm10k_qv_enable(struct fm10k_q_vector *q_vector) q_vector 1416 drivers/net/ethernet/intel/fm10k/fm10k_main.c fm10k_update_itr(&q_vector->tx); q_vector 1419 drivers/net/ethernet/intel/fm10k/fm10k_main.c fm10k_update_itr(&q_vector->rx); q_vector 1422 drivers/net/ethernet/intel/fm10k/fm10k_main.c itr |= (q_vector->tx.itr & FM10K_ITR_MAX); q_vector 1425 drivers/net/ethernet/intel/fm10k/fm10k_main.c itr |= (q_vector->rx.itr & FM10K_ITR_MAX) << FM10K_ITR_INTERVAL1_SHIFT; q_vector 1428 drivers/net/ethernet/intel/fm10k/fm10k_main.c writel(itr, q_vector->itr); q_vector 1433 drivers/net/ethernet/intel/fm10k/fm10k_main.c struct fm10k_q_vector *q_vector = q_vector 1439 drivers/net/ethernet/intel/fm10k/fm10k_main.c fm10k_for_each_ring(ring, q_vector->tx) { q_vector 1440 drivers/net/ethernet/intel/fm10k/fm10k_main.c if (!fm10k_clean_tx_irq(q_vector, ring, budget)) q_vector 1451 drivers/net/ethernet/intel/fm10k/fm10k_main.c if (q_vector->rx.count > 1) q_vector 1452 drivers/net/ethernet/intel/fm10k/fm10k_main.c per_ring_budget = max(budget / q_vector->rx.count, 1); q_vector 1456 drivers/net/ethernet/intel/fm10k/fm10k_main.c fm10k_for_each_ring(ring, q_vector->rx) { q_vector 1457 drivers/net/ethernet/intel/fm10k/fm10k_main.c int work = fm10k_clean_rx_irq(q_vector, ring, per_ring_budget); q_vector 1472 drivers/net/ethernet/intel/fm10k/fm10k_main.c fm10k_qv_enable(q_vector); q_vector 1604 drivers/net/ethernet/intel/fm10k/fm10k_main.c struct fm10k_q_vector *q_vector; q_vector 1611 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector = kzalloc(struct_size(q_vector, ring, ring_count), GFP_KERNEL); q_vector 1612 drivers/net/ethernet/intel/fm10k/fm10k_main.c if (!q_vector) q_vector 1616 drivers/net/ethernet/intel/fm10k/fm10k_main.c netif_napi_add(interface->netdev, &q_vector->napi, q_vector 1620 drivers/net/ethernet/intel/fm10k/fm10k_main.c interface->q_vector[v_idx] = q_vector; q_vector 1621 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->interface = interface; q_vector 1622 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->v_idx = v_idx; q_vector 1625 drivers/net/ethernet/intel/fm10k/fm10k_main.c ring = q_vector->ring; q_vector 1628 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->tx.ring = ring; q_vector 1629 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->tx.work_limit = FM10K_DEFAULT_TX_WORK; q_vector 1630 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->tx.itr = interface->tx_itr; q_vector 1631 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->tx.itr_scale = interface->hw.mac.itr_scale; q_vector 1632 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->tx.count = txr_count; q_vector 1640 drivers/net/ethernet/intel/fm10k/fm10k_main.c ring->q_vector = q_vector; q_vector 1658 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->rx.ring = ring; q_vector 1659 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->rx.itr = interface->rx_itr; q_vector 1660 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->rx.itr_scale = interface->hw.mac.itr_scale; q_vector 1661 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->rx.count = rxr_count; q_vector 1670 drivers/net/ethernet/intel/fm10k/fm10k_main.c ring->q_vector = q_vector; q_vector 1687 drivers/net/ethernet/intel/fm10k/fm10k_main.c fm10k_dbg_q_vector_init(q_vector); q_vector 1703 drivers/net/ethernet/intel/fm10k/fm10k_main.c struct fm10k_q_vector *q_vector = interface->q_vector[v_idx]; q_vector 1706 drivers/net/ethernet/intel/fm10k/fm10k_main.c fm10k_dbg_q_vector_exit(q_vector); q_vector 1708 drivers/net/ethernet/intel/fm10k/fm10k_main.c fm10k_for_each_ring(ring, q_vector->tx) q_vector 1711 drivers/net/ethernet/intel/fm10k/fm10k_main.c fm10k_for_each_ring(ring, q_vector->rx) q_vector 1714 drivers/net/ethernet/intel/fm10k/fm10k_main.c interface->q_vector[v_idx] = NULL; q_vector 1715 drivers/net/ethernet/intel/fm10k/fm10k_main.c netif_napi_del(&q_vector->napi); q_vector 1716 drivers/net/ethernet/intel/fm10k/fm10k_main.c kfree_rcu(q_vector, rcu); q_vector 720 drivers/net/ethernet/intel/fm10k/fm10k_pci.c struct fm10k_q_vector *qv = interface->q_vector[i]; q_vector 900 drivers/net/ethernet/intel/fm10k/fm10k_pci.c if (ring->q_vector) { q_vector 901 drivers/net/ethernet/intel/fm10k/fm10k_pci.c txint = ring->q_vector->v_idx + NON_Q_VECTORS; q_vector 913 drivers/net/ethernet/intel/fm10k/fm10k_pci.c ring->q_vector) q_vector 915 drivers/net/ethernet/intel/fm10k/fm10k_pci.c &ring->q_vector->affinity_mask, q_vector 1039 drivers/net/ethernet/intel/fm10k/fm10k_pci.c if (ring->q_vector) { q_vector 1040 drivers/net/ethernet/intel/fm10k/fm10k_pci.c rxint = ring->q_vector->v_idx + NON_Q_VECTORS; q_vector 1174 drivers/net/ethernet/intel/fm10k/fm10k_pci.c struct fm10k_q_vector *q_vector; q_vector 1178 drivers/net/ethernet/intel/fm10k/fm10k_pci.c q_vector = interface->q_vector[q_idx]; q_vector 1179 drivers/net/ethernet/intel/fm10k/fm10k_pci.c napi_enable(&q_vector->napi); q_vector 1185 drivers/net/ethernet/intel/fm10k/fm10k_pci.c struct fm10k_q_vector *q_vector = data; q_vector 1187 drivers/net/ethernet/intel/fm10k/fm10k_pci.c if (q_vector->rx.count || q_vector->tx.count) q_vector 1188 drivers/net/ethernet/intel/fm10k/fm10k_pci.c napi_schedule_irqoff(&q_vector->napi); q_vector 1728 drivers/net/ethernet/intel/fm10k/fm10k_pci.c struct fm10k_q_vector *q_vector; q_vector 1732 drivers/net/ethernet/intel/fm10k/fm10k_pci.c q_vector = interface->q_vector[vector]; q_vector 1734 drivers/net/ethernet/intel/fm10k/fm10k_pci.c if (!q_vector->tx.count && !q_vector->rx.count) q_vector 1741 drivers/net/ethernet/intel/fm10k/fm10k_pci.c writel(FM10K_ITR_MASK_SET, q_vector->itr); q_vector 1743 drivers/net/ethernet/intel/fm10k/fm10k_pci.c free_irq(entry->vector, q_vector); q_vector 1765 drivers/net/ethernet/intel/fm10k/fm10k_pci.c struct fm10k_q_vector *q_vector = interface->q_vector[vector]; q_vector 1768 drivers/net/ethernet/intel/fm10k/fm10k_pci.c if (q_vector->tx.count && q_vector->rx.count) { q_vector 1769 drivers/net/ethernet/intel/fm10k/fm10k_pci.c snprintf(q_vector->name, sizeof(q_vector->name), q_vector 1772 drivers/net/ethernet/intel/fm10k/fm10k_pci.c } else if (q_vector->rx.count) { q_vector 1773 drivers/net/ethernet/intel/fm10k/fm10k_pci.c snprintf(q_vector->name, sizeof(q_vector->name), q_vector 1775 drivers/net/ethernet/intel/fm10k/fm10k_pci.c } else if (q_vector->tx.count) { q_vector 1776 drivers/net/ethernet/intel/fm10k/fm10k_pci.c snprintf(q_vector->name, sizeof(q_vector->name), q_vector 1784 drivers/net/ethernet/intel/fm10k/fm10k_pci.c q_vector->itr = (hw->mac.type == fm10k_mac_pf) ? q_vector 1790 drivers/net/ethernet/intel/fm10k/fm10k_pci.c q_vector->name, q_vector); q_vector 1799 drivers/net/ethernet/intel/fm10k/fm10k_pci.c irq_set_affinity_hint(entry->vector, &q_vector->affinity_mask); q_vector 1802 drivers/net/ethernet/intel/fm10k/fm10k_pci.c writel(FM10K_ITR_ENABLE, q_vector->itr); q_vector 1812 drivers/net/ethernet/intel/fm10k/fm10k_pci.c struct fm10k_q_vector *q_vector; q_vector 1816 drivers/net/ethernet/intel/fm10k/fm10k_pci.c q_vector = interface->q_vector[vector]; q_vector 1818 drivers/net/ethernet/intel/fm10k/fm10k_pci.c if (!q_vector->tx.count && !q_vector->rx.count) q_vector 1825 drivers/net/ethernet/intel/fm10k/fm10k_pci.c writel(FM10K_ITR_MASK_SET, q_vector->itr); q_vector 1827 drivers/net/ethernet/intel/fm10k/fm10k_pci.c free_irq(entry->vector, q_vector); q_vector 1871 drivers/net/ethernet/intel/fm10k/fm10k_pci.c struct fm10k_q_vector *q_vector; q_vector 1875 drivers/net/ethernet/intel/fm10k/fm10k_pci.c q_vector = interface->q_vector[q_idx]; q_vector 1876 drivers/net/ethernet/intel/fm10k/fm10k_pci.c napi_disable(&q_vector->napi); q_vector 2817 drivers/net/ethernet/intel/i40e/i40e_ethtool.c struct i40e_q_vector *q_vector; q_vector 2835 drivers/net/ethernet/intel/i40e/i40e_ethtool.c q_vector = rx_ring->q_vector; q_vector 2836 drivers/net/ethernet/intel/i40e/i40e_ethtool.c q_vector->rx.target_itr = ITR_TO_REG(rx_ring->itr_setting); q_vector 2838 drivers/net/ethernet/intel/i40e/i40e_ethtool.c q_vector = tx_ring->q_vector; q_vector 2839 drivers/net/ethernet/intel/i40e/i40e_ethtool.c q_vector->tx.target_itr = ITR_TO_REG(tx_ring->itr_setting); q_vector 2846 drivers/net/ethernet/intel/i40e/i40e_ethtool.c wr32(hw, I40E_PFINT_RATEN(q_vector->reg_idx), intrl); q_vector 360 drivers/net/ethernet/intel/i40e/i40e_main.c I40E_PFINT_DYN_CTLN(tx_ring->q_vector->v_idx + q_vector 3093 drivers/net/ethernet/intel/i40e/i40e_main.c if (!ring->q_vector || !ring->netdev || ring->ch) q_vector 3100 drivers/net/ethernet/intel/i40e/i40e_main.c cpu = cpumask_local_spread(ring->q_vector->v_idx, -1); q_vector 3531 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_q_vector *q_vector = vsi->q_vectors[i]; q_vector 3533 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->rx.next_update = jiffies + 1; q_vector 3534 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->rx.target_itr = q_vector 3537 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->rx.target_itr >> 1); q_vector 3538 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->rx.current_itr = q_vector->rx.target_itr; q_vector 3540 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.next_update = jiffies + 1; q_vector 3541 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.target_itr = q_vector 3544 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.target_itr >> 1); q_vector 3545 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.current_itr = q_vector->tx.target_itr; q_vector 3552 drivers/net/ethernet/intel/i40e/i40e_main.c for (q = 0; q < q_vector->num_ringpairs; q++) { q_vector 3584 drivers/net/ethernet/intel/i40e/i40e_main.c if (q == (q_vector->num_ringpairs - 1)) q_vector 3641 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_q_vector *q_vector = vsi->q_vectors[0]; q_vector 3647 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->rx.next_update = jiffies + 1; q_vector 3648 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->rx.target_itr = ITR_TO_REG(vsi->rx_rings[0]->itr_setting); q_vector 3649 drivers/net/ethernet/intel/i40e/i40e_main.c wr32(hw, I40E_PFINT_ITR0(I40E_RX_ITR), q_vector->rx.target_itr >> 1); q_vector 3650 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->rx.current_itr = q_vector->rx.target_itr; q_vector 3651 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.next_update = jiffies + 1; q_vector 3652 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.target_itr = ITR_TO_REG(vsi->tx_rings[0]->itr_setting); q_vector 3653 drivers/net/ethernet/intel/i40e/i40e_main.c wr32(hw, I40E_PFINT_ITR0(I40E_TX_ITR), q_vector->tx.target_itr >> 1); q_vector 3654 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.current_itr = q_vector->tx.target_itr; q_vector 3723 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_q_vector *q_vector = data; q_vector 3725 drivers/net/ethernet/intel/i40e/i40e_main.c if (!q_vector->tx.ring && !q_vector->rx.ring) q_vector 3728 drivers/net/ethernet/intel/i40e/i40e_main.c napi_schedule_irqoff(&q_vector->napi); q_vector 3744 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_q_vector *q_vector = q_vector 3747 drivers/net/ethernet/intel/i40e/i40e_main.c cpumask_copy(&q_vector->affinity_mask, mask); q_vector 3779 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_q_vector *q_vector = vsi->q_vectors[vector]; q_vector 3783 drivers/net/ethernet/intel/i40e/i40e_main.c if (q_vector->tx.ring && q_vector->rx.ring) { q_vector 3784 drivers/net/ethernet/intel/i40e/i40e_main.c snprintf(q_vector->name, sizeof(q_vector->name) - 1, q_vector 3787 drivers/net/ethernet/intel/i40e/i40e_main.c } else if (q_vector->rx.ring) { q_vector 3788 drivers/net/ethernet/intel/i40e/i40e_main.c snprintf(q_vector->name, sizeof(q_vector->name) - 1, q_vector 3790 drivers/net/ethernet/intel/i40e/i40e_main.c } else if (q_vector->tx.ring) { q_vector 3791 drivers/net/ethernet/intel/i40e/i40e_main.c snprintf(q_vector->name, sizeof(q_vector->name) - 1, q_vector 3800 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->name, q_vector 3801 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector); q_vector 3809 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->affinity_notify.notify = i40e_irq_affinity_notify; q_vector 3810 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->affinity_notify.release = i40e_irq_affinity_release; q_vector 3811 drivers/net/ethernet/intel/i40e/i40e_main.c irq_set_affinity_notifier(irq_num, &q_vector->affinity_notify); q_vector 3818 drivers/net/ethernet/intel/i40e/i40e_main.c cpu = cpumask_local_spread(q_vector->v_idx, -1); q_vector 3958 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_q_vector *q_vector = vsi->q_vectors[0]; q_vector 3967 drivers/net/ethernet/intel/i40e/i40e_main.c napi_schedule_irqoff(&q_vector->napi); q_vector 4131 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_irq_dynamic_enable(vsi, tx_ring->q_vector->v_idx); q_vector 4143 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_q_vector *q_vector = data; q_vector 4146 drivers/net/ethernet/intel/i40e/i40e_main.c if (!q_vector->tx.ring) q_vector 4149 drivers/net/ethernet/intel/i40e/i40e_main.c vsi = q_vector->tx.ring->vsi; q_vector 4150 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_clean_fdir_tx_irq(q_vector->tx.ring, vsi->work_limit); q_vector 4163 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_q_vector *q_vector = vsi->q_vectors[v_idx]; q_vector 4167 drivers/net/ethernet/intel/i40e/i40e_main.c tx_ring->q_vector = q_vector; q_vector 4168 drivers/net/ethernet/intel/i40e/i40e_main.c tx_ring->next = q_vector->tx.ring; q_vector 4169 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.ring = tx_ring; q_vector 4170 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.count++; q_vector 4176 drivers/net/ethernet/intel/i40e/i40e_main.c xdp_ring->q_vector = q_vector; q_vector 4177 drivers/net/ethernet/intel/i40e/i40e_main.c xdp_ring->next = q_vector->tx.ring; q_vector 4178 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.ring = xdp_ring; q_vector 4179 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.count++; q_vector 4182 drivers/net/ethernet/intel/i40e/i40e_main.c rx_ring->q_vector = q_vector; q_vector 4183 drivers/net/ethernet/intel/i40e/i40e_main.c rx_ring->next = q_vector->rx.ring; q_vector 4184 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->rx.ring = rx_ring; q_vector 4185 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->rx.count++; q_vector 4213 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_q_vector *q_vector = vsi->q_vectors[v_start]; q_vector 4217 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->num_ringpairs = num_ringpairs; q_vector 4218 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->reg_idx = q_vector->v_idx + vsi->base_vector - 1; q_vector 4220 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->rx.count = 0; q_vector 4221 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.count = 0; q_vector 4222 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->rx.ring = NULL; q_vector 4223 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->tx.ring = NULL; q_vector 4726 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_q_vector *q_vector = vsi->q_vectors[v_idx]; q_vector 4729 drivers/net/ethernet/intel/i40e/i40e_main.c if (!q_vector) q_vector 4733 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_for_each_ring(ring, q_vector->tx) q_vector 4734 drivers/net/ethernet/intel/i40e/i40e_main.c ring->q_vector = NULL; q_vector 4736 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_for_each_ring(ring, q_vector->rx) q_vector 4737 drivers/net/ethernet/intel/i40e/i40e_main.c ring->q_vector = NULL; q_vector 4741 drivers/net/ethernet/intel/i40e/i40e_main.c netif_napi_del(&q_vector->napi); q_vector 4745 drivers/net/ethernet/intel/i40e/i40e_main.c kfree_rcu(q_vector, rcu); q_vector 4817 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_q_vector *q_vector = vsi->q_vectors[q_idx]; q_vector 4819 drivers/net/ethernet/intel/i40e/i40e_main.c if (q_vector->rx.ring || q_vector->tx.ring) q_vector 4820 drivers/net/ethernet/intel/i40e/i40e_main.c napi_enable(&q_vector->napi); q_vector 4836 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_q_vector *q_vector = vsi->q_vectors[q_idx]; q_vector 4838 drivers/net/ethernet/intel/i40e/i40e_main.c if (q_vector->rx.ring || q_vector->tx.ring) q_vector 4839 drivers/net/ethernet/intel/i40e/i40e_main.c napi_disable(&q_vector->napi); q_vector 11157 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_q_vector *q_vector; q_vector 11160 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector = kzalloc(sizeof(struct i40e_q_vector), GFP_KERNEL); q_vector 11161 drivers/net/ethernet/intel/i40e/i40e_main.c if (!q_vector) q_vector 11164 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->vsi = vsi; q_vector 11165 drivers/net/ethernet/intel/i40e/i40e_main.c q_vector->v_idx = v_idx; q_vector 11166 drivers/net/ethernet/intel/i40e/i40e_main.c cpumask_copy(&q_vector->affinity_mask, cpu_possible_mask); q_vector 11169 drivers/net/ethernet/intel/i40e/i40e_main.c netif_napi_add(vsi->netdev, &q_vector->napi, q_vector 11173 drivers/net/ethernet/intel/i40e/i40e_main.c vsi->q_vectors[v_idx] = q_vector; q_vector 12632 drivers/net/ethernet/intel/i40e/i40e_main.c struct i40e_q_vector *q_vector = rxr->q_vector; q_vector 12638 drivers/net/ethernet/intel/i40e/i40e_main.c if (q_vector->rx.ring || q_vector->tx.ring) { q_vector 12640 drivers/net/ethernet/intel/i40e/i40e_main.c napi_enable(&q_vector->napi); q_vector 12642 drivers/net/ethernet/intel/i40e/i40e_main.c napi_disable(&q_vector->napi); q_vector 12713 drivers/net/ethernet/intel/i40e/i40e_main.c i40e_irq_dynamic_enable(vsi, rxr->q_vector->v_idx); q_vector 12738 drivers/net/ethernet/intel/i40e/i40e_main.c u32 intpf = vsi->base_vector + rxr->q_vector->v_idx; q_vector 752 drivers/net/ethernet/intel/i40e/i40e_txrx.c i40e_force_wb(vsi, tx_ring->q_vector); q_vector 906 drivers/net/ethernet/intel/i40e/i40e_txrx.c struct i40e_q_vector *q_vector) q_vector 908 drivers/net/ethernet/intel/i40e/i40e_txrx.c u16 flags = q_vector->tx.ring[0].flags; q_vector 914 drivers/net/ethernet/intel/i40e/i40e_txrx.c if (q_vector->arm_wb_state) q_vector 922 drivers/net/ethernet/intel/i40e/i40e_txrx.c I40E_PFINT_DYN_CTLN(q_vector->reg_idx), q_vector 930 drivers/net/ethernet/intel/i40e/i40e_txrx.c q_vector->arm_wb_state = true; q_vector 939 drivers/net/ethernet/intel/i40e/i40e_txrx.c void i40e_force_wb(struct i40e_vsi *vsi, struct i40e_q_vector *q_vector) q_vector 949 drivers/net/ethernet/intel/i40e/i40e_txrx.c I40E_PFINT_DYN_CTLN(q_vector->reg_idx), val); q_vector 961 drivers/net/ethernet/intel/i40e/i40e_txrx.c static inline bool i40e_container_is_rx(struct i40e_q_vector *q_vector, q_vector 964 drivers/net/ethernet/intel/i40e/i40e_txrx.c return &q_vector->rx == rc; q_vector 967 drivers/net/ethernet/intel/i40e/i40e_txrx.c static inline unsigned int i40e_itr_divisor(struct i40e_q_vector *q_vector) q_vector 971 drivers/net/ethernet/intel/i40e/i40e_txrx.c switch (q_vector->vsi->back->hw.phy.link_info.link_speed) { q_vector 1005 drivers/net/ethernet/intel/i40e/i40e_txrx.c static void i40e_update_itr(struct i40e_q_vector *q_vector, q_vector 1020 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr = i40e_container_is_rx(q_vector, rc) ? q_vector 1038 drivers/net/ethernet/intel/i40e/i40e_txrx.c if (q_vector->itr_countdown) { q_vector 1046 drivers/net/ethernet/intel/i40e/i40e_txrx.c if (i40e_container_is_rx(q_vector, rc)) { q_vector 1053 drivers/net/ethernet/intel/i40e/i40e_txrx.c (q_vector->tx.target_itr & I40E_ITR_ADAPTIVE_LATENCY)) { q_vector 1064 drivers/net/ethernet/intel/i40e/i40e_txrx.c (q_vector->rx.target_itr & I40E_ITR_MASK) == q_vector 1092 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr = min(q_vector->tx.current_itr, q_vector->rx.current_itr); q_vector 1179 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr += DIV_ROUND_UP(avg_wire_size, i40e_itr_divisor(q_vector)) * q_vector 2029 drivers/net/ethernet/intel/i40e/i40e_txrx.c skb = __napi_alloc_skb(&rx_ring->q_vector->napi, q_vector 2291 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_ring->q_vector->rx.total_packets += total_rx_packets; q_vector 2292 drivers/net/ethernet/intel/i40e/i40e_txrx.c rx_ring->q_vector->rx.total_bytes += total_rx_bytes; q_vector 2440 drivers/net/ethernet/intel/i40e/i40e_txrx.c napi_gro_receive(&rx_ring->q_vector->napi, skb); q_vector 2503 drivers/net/ethernet/intel/i40e/i40e_txrx.c struct i40e_q_vector *q_vector) q_vector 2515 drivers/net/ethernet/intel/i40e/i40e_txrx.c i40e_update_itr(q_vector, &q_vector->tx); q_vector 2516 drivers/net/ethernet/intel/i40e/i40e_txrx.c i40e_update_itr(q_vector, &q_vector->rx); q_vector 2526 drivers/net/ethernet/intel/i40e/i40e_txrx.c if (q_vector->rx.target_itr < q_vector->rx.current_itr) { q_vector 2529 drivers/net/ethernet/intel/i40e/i40e_txrx.c q_vector->rx.target_itr); q_vector 2530 drivers/net/ethernet/intel/i40e/i40e_txrx.c q_vector->rx.current_itr = q_vector->rx.target_itr; q_vector 2531 drivers/net/ethernet/intel/i40e/i40e_txrx.c q_vector->itr_countdown = ITR_COUNTDOWN_START; q_vector 2532 drivers/net/ethernet/intel/i40e/i40e_txrx.c } else if ((q_vector->tx.target_itr < q_vector->tx.current_itr) || q_vector 2533 drivers/net/ethernet/intel/i40e/i40e_txrx.c ((q_vector->rx.target_itr - q_vector->rx.current_itr) < q_vector 2534 drivers/net/ethernet/intel/i40e/i40e_txrx.c (q_vector->tx.target_itr - q_vector->tx.current_itr))) { q_vector 2539 drivers/net/ethernet/intel/i40e/i40e_txrx.c q_vector->tx.target_itr); q_vector 2540 drivers/net/ethernet/intel/i40e/i40e_txrx.c q_vector->tx.current_itr = q_vector->tx.target_itr; q_vector 2541 drivers/net/ethernet/intel/i40e/i40e_txrx.c q_vector->itr_countdown = ITR_COUNTDOWN_START; q_vector 2542 drivers/net/ethernet/intel/i40e/i40e_txrx.c } else if (q_vector->rx.current_itr != q_vector->rx.target_itr) { q_vector 2545 drivers/net/ethernet/intel/i40e/i40e_txrx.c q_vector->rx.target_itr); q_vector 2546 drivers/net/ethernet/intel/i40e/i40e_txrx.c q_vector->rx.current_itr = q_vector->rx.target_itr; q_vector 2547 drivers/net/ethernet/intel/i40e/i40e_txrx.c q_vector->itr_countdown = ITR_COUNTDOWN_START; q_vector 2551 drivers/net/ethernet/intel/i40e/i40e_txrx.c if (q_vector->itr_countdown) q_vector 2552 drivers/net/ethernet/intel/i40e/i40e_txrx.c q_vector->itr_countdown--; q_vector 2556 drivers/net/ethernet/intel/i40e/i40e_txrx.c wr32(hw, INTREG(q_vector->reg_idx), intval); q_vector 2570 drivers/net/ethernet/intel/i40e/i40e_txrx.c struct i40e_q_vector *q_vector = q_vector 2572 drivers/net/ethernet/intel/i40e/i40e_txrx.c struct i40e_vsi *vsi = q_vector->vsi; q_vector 2587 drivers/net/ethernet/intel/i40e/i40e_txrx.c i40e_for_each_ring(ring, q_vector->tx) { q_vector 2607 drivers/net/ethernet/intel/i40e/i40e_txrx.c budget_per_ring = max(budget/q_vector->num_ringpairs, 1); q_vector 2609 drivers/net/ethernet/intel/i40e/i40e_txrx.c i40e_for_each_ring(ring, q_vector->rx) { q_vector 2631 drivers/net/ethernet/intel/i40e/i40e_txrx.c if (!cpumask_test_cpu(cpu_id, &q_vector->affinity_mask)) { q_vector 2636 drivers/net/ethernet/intel/i40e/i40e_txrx.c i40e_force_wb(vsi, q_vector); q_vector 2643 drivers/net/ethernet/intel/i40e/i40e_txrx.c q_vector->tx.ring[0].tx_stats.tx_force_wb++; q_vector 2644 drivers/net/ethernet/intel/i40e/i40e_txrx.c i40e_enable_wb_on_itr(vsi, q_vector); q_vector 2650 drivers/net/ethernet/intel/i40e/i40e_txrx.c q_vector->arm_wb_state = false; q_vector 2656 drivers/net/ethernet/intel/i40e/i40e_txrx.c i40e_update_enable_itr(vsi, q_vector); q_vector 406 drivers/net/ethernet/intel/i40e/i40e_txrx.h struct i40e_q_vector *q_vector; /* Backreference to associated vector */ q_vector 491 drivers/net/ethernet/intel/i40e/i40e_txrx.h void i40e_force_wb(struct i40e_vsi *vsi, struct i40e_q_vector *q_vector); q_vector 55 drivers/net/ethernet/intel/i40e/i40e_txrx_common.h tx_ring->q_vector->tx.total_bytes += total_bytes; q_vector 56 drivers/net/ethernet/intel/i40e/i40e_txrx_common.h tx_ring->q_vector->tx.total_packets += total_packets; q_vector 495 drivers/net/ethernet/intel/i40e/i40e_xsk.c skb = __napi_alloc_skb(&rx_ring->q_vector->napi, q_vector 623 drivers/net/ethernet/intel/i40e/i40e_xsk.c napi_gro_receive(&rx_ring->q_vector->napi, skb); q_vector 816 drivers/net/ethernet/intel/i40e/i40e_xsk.c if (!napi_if_scheduled_mark_missed(&ring->q_vector->napi)) q_vector 817 drivers/net/ethernet/intel/i40e/i40e_xsk.c i40e_force_wb(vsi, ring->q_vector); q_vector 717 drivers/net/ethernet/intel/iavf/iavf_ethtool.c struct iavf_q_vector *q_vector; q_vector 730 drivers/net/ethernet/intel/iavf/iavf_ethtool.c q_vector = rx_ring->q_vector; q_vector 731 drivers/net/ethernet/intel/iavf/iavf_ethtool.c q_vector->rx.target_itr = ITR_TO_REG(rx_ring->itr_setting); q_vector 733 drivers/net/ethernet/intel/iavf/iavf_ethtool.c q_vector = tx_ring->q_vector; q_vector 734 drivers/net/ethernet/intel/iavf/iavf_ethtool.c q_vector->tx.target_itr = ITR_TO_REG(tx_ring->itr_setting); q_vector 285 drivers/net/ethernet/intel/iavf/iavf_main.c struct iavf_q_vector *q_vector = data; q_vector 287 drivers/net/ethernet/intel/iavf/iavf_main.c if (!q_vector->tx.ring && !q_vector->rx.ring) q_vector 290 drivers/net/ethernet/intel/iavf/iavf_main.c napi_schedule_irqoff(&q_vector->napi); q_vector 304 drivers/net/ethernet/intel/iavf/iavf_main.c struct iavf_q_vector *q_vector = &adapter->q_vectors[v_idx]; q_vector 308 drivers/net/ethernet/intel/iavf/iavf_main.c rx_ring->q_vector = q_vector; q_vector 309 drivers/net/ethernet/intel/iavf/iavf_main.c rx_ring->next = q_vector->rx.ring; q_vector 311 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->rx.ring = rx_ring; q_vector 312 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->rx.count++; q_vector 313 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->rx.next_update = jiffies + 1; q_vector 314 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->rx.target_itr = ITR_TO_REG(rx_ring->itr_setting); q_vector 315 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->ring_mask |= BIT(r_idx); q_vector 316 drivers/net/ethernet/intel/iavf/iavf_main.c wr32(hw, IAVF_VFINT_ITRN1(IAVF_RX_ITR, q_vector->reg_idx), q_vector 317 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->rx.current_itr >> 1); q_vector 318 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->rx.current_itr = q_vector->rx.target_itr; q_vector 330 drivers/net/ethernet/intel/iavf/iavf_main.c struct iavf_q_vector *q_vector = &adapter->q_vectors[v_idx]; q_vector 334 drivers/net/ethernet/intel/iavf/iavf_main.c tx_ring->q_vector = q_vector; q_vector 335 drivers/net/ethernet/intel/iavf/iavf_main.c tx_ring->next = q_vector->tx.ring; q_vector 337 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->tx.ring = tx_ring; q_vector 338 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->tx.count++; q_vector 339 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->tx.next_update = jiffies + 1; q_vector 340 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->tx.target_itr = ITR_TO_REG(tx_ring->itr_setting); q_vector 341 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->num_ringpairs++; q_vector 342 drivers/net/ethernet/intel/iavf/iavf_main.c wr32(hw, IAVF_VFINT_ITRN1(IAVF_TX_ITR, q_vector->reg_idx), q_vector 343 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->tx.target_itr >> 1); q_vector 344 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->tx.current_itr = q_vector->tx.target_itr; q_vector 390 drivers/net/ethernet/intel/iavf/iavf_main.c struct iavf_q_vector *q_vector = q_vector 393 drivers/net/ethernet/intel/iavf/iavf_main.c cpumask_copy(&q_vector->affinity_mask, mask); q_vector 427 drivers/net/ethernet/intel/iavf/iavf_main.c struct iavf_q_vector *q_vector = &adapter->q_vectors[vector]; q_vector 431 drivers/net/ethernet/intel/iavf/iavf_main.c if (q_vector->tx.ring && q_vector->rx.ring) { q_vector 432 drivers/net/ethernet/intel/iavf/iavf_main.c snprintf(q_vector->name, sizeof(q_vector->name), q_vector 435 drivers/net/ethernet/intel/iavf/iavf_main.c } else if (q_vector->rx.ring) { q_vector 436 drivers/net/ethernet/intel/iavf/iavf_main.c snprintf(q_vector->name, sizeof(q_vector->name), q_vector 438 drivers/net/ethernet/intel/iavf/iavf_main.c } else if (q_vector->tx.ring) { q_vector 439 drivers/net/ethernet/intel/iavf/iavf_main.c snprintf(q_vector->name, sizeof(q_vector->name), q_vector 448 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->name, q_vector 449 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector); q_vector 456 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->affinity_notify.notify = iavf_irq_affinity_notify; q_vector 457 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->affinity_notify.release = q_vector 459 drivers/net/ethernet/intel/iavf/iavf_main.c irq_set_affinity_notifier(irq_num, &q_vector->affinity_notify); q_vector 464 drivers/net/ethernet/intel/iavf/iavf_main.c cpu = cpumask_local_spread(q_vector->v_idx, -1); q_vector 893 drivers/net/ethernet/intel/iavf/iavf_main.c struct iavf_q_vector *q_vector; q_vector 899 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector = &adapter->q_vectors[q_idx]; q_vector 900 drivers/net/ethernet/intel/iavf/iavf_main.c napi = &q_vector->napi; q_vector 912 drivers/net/ethernet/intel/iavf/iavf_main.c struct iavf_q_vector *q_vector; q_vector 916 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector = &adapter->q_vectors[q_idx]; q_vector 917 drivers/net/ethernet/intel/iavf/iavf_main.c napi_disable(&q_vector->napi); q_vector 1337 drivers/net/ethernet/intel/iavf/iavf_main.c struct iavf_q_vector *q_vector; q_vector 1340 drivers/net/ethernet/intel/iavf/iavf_main.c adapter->q_vectors = kcalloc(num_q_vectors, sizeof(*q_vector), q_vector 1346 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector = &adapter->q_vectors[q_idx]; q_vector 1347 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->adapter = adapter; q_vector 1348 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->vsi = &adapter->vsi; q_vector 1349 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->v_idx = q_idx; q_vector 1350 drivers/net/ethernet/intel/iavf/iavf_main.c q_vector->reg_idx = q_idx; q_vector 1351 drivers/net/ethernet/intel/iavf/iavf_main.c cpumask_copy(&q_vector->affinity_mask, cpu_possible_mask); q_vector 1352 drivers/net/ethernet/intel/iavf/iavf_main.c netif_napi_add(adapter->netdev, &q_vector->napi, q_vector 1379 drivers/net/ethernet/intel/iavf/iavf_main.c struct iavf_q_vector *q_vector = &adapter->q_vectors[q_idx]; q_vector 1382 drivers/net/ethernet/intel/iavf/iavf_main.c netif_napi_del(&q_vector->napi); q_vector 166 drivers/net/ethernet/intel/iavf/iavf_txrx.c iavf_force_wb(vsi, tx_ring->q_vector); q_vector 285 drivers/net/ethernet/intel/iavf/iavf_txrx.c tx_ring->q_vector->tx.total_bytes += total_bytes; q_vector 286 drivers/net/ethernet/intel/iavf/iavf_txrx.c tx_ring->q_vector->tx.total_packets += total_packets; q_vector 333 drivers/net/ethernet/intel/iavf/iavf_txrx.c struct iavf_q_vector *q_vector) q_vector 335 drivers/net/ethernet/intel/iavf/iavf_txrx.c u16 flags = q_vector->tx.ring[0].flags; q_vector 341 drivers/net/ethernet/intel/iavf/iavf_txrx.c if (q_vector->arm_wb_state) q_vector 348 drivers/net/ethernet/intel/iavf/iavf_txrx.c IAVF_VFINT_DYN_CTLN1(q_vector->reg_idx), val); q_vector 349 drivers/net/ethernet/intel/iavf/iavf_txrx.c q_vector->arm_wb_state = true; q_vector 358 drivers/net/ethernet/intel/iavf/iavf_txrx.c void iavf_force_wb(struct iavf_vsi *vsi, struct iavf_q_vector *q_vector) q_vector 367 drivers/net/ethernet/intel/iavf/iavf_txrx.c IAVF_VFINT_DYN_CTLN1(q_vector->reg_idx), q_vector 371 drivers/net/ethernet/intel/iavf/iavf_txrx.c static inline bool iavf_container_is_rx(struct iavf_q_vector *q_vector, q_vector 374 drivers/net/ethernet/intel/iavf/iavf_txrx.c return &q_vector->rx == rc; q_vector 377 drivers/net/ethernet/intel/iavf/iavf_txrx.c static inline unsigned int iavf_itr_divisor(struct iavf_q_vector *q_vector) q_vector 381 drivers/net/ethernet/intel/iavf/iavf_txrx.c switch (q_vector->adapter->link_speed) { q_vector 415 drivers/net/ethernet/intel/iavf/iavf_txrx.c static void iavf_update_itr(struct iavf_q_vector *q_vector, q_vector 430 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr = iavf_container_is_rx(q_vector, rc) ? q_vector 448 drivers/net/ethernet/intel/iavf/iavf_txrx.c if (q_vector->itr_countdown) { q_vector 456 drivers/net/ethernet/intel/iavf/iavf_txrx.c if (iavf_container_is_rx(q_vector, rc)) { q_vector 463 drivers/net/ethernet/intel/iavf/iavf_txrx.c (q_vector->tx.target_itr & IAVF_ITR_ADAPTIVE_LATENCY)) { q_vector 474 drivers/net/ethernet/intel/iavf/iavf_txrx.c (q_vector->rx.target_itr & IAVF_ITR_MASK) == q_vector 502 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr = min(q_vector->tx.current_itr, q_vector->rx.current_itr); q_vector 589 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr += DIV_ROUND_UP(avg_wire_size, iavf_itr_divisor(q_vector)) * q_vector 863 drivers/net/ethernet/intel/iavf/iavf_txrx.c struct iavf_q_vector *q_vector = rx_ring->q_vector; q_vector 869 drivers/net/ethernet/intel/iavf/iavf_txrx.c napi_gro_receive(&q_vector->napi, skb); q_vector 1318 drivers/net/ethernet/intel/iavf/iavf_txrx.c skb = __napi_alloc_skb(&rx_ring->q_vector->napi, q_vector 1590 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_ring->q_vector->rx.total_packets += total_rx_packets; q_vector 1591 drivers/net/ethernet/intel/iavf/iavf_txrx.c rx_ring->q_vector->rx.total_bytes += total_rx_bytes; q_vector 1644 drivers/net/ethernet/intel/iavf/iavf_txrx.c struct iavf_q_vector *q_vector) q_vector 1650 drivers/net/ethernet/intel/iavf/iavf_txrx.c iavf_update_itr(q_vector, &q_vector->tx); q_vector 1651 drivers/net/ethernet/intel/iavf/iavf_txrx.c iavf_update_itr(q_vector, &q_vector->rx); q_vector 1661 drivers/net/ethernet/intel/iavf/iavf_txrx.c if (q_vector->rx.target_itr < q_vector->rx.current_itr) { q_vector 1664 drivers/net/ethernet/intel/iavf/iavf_txrx.c q_vector->rx.target_itr); q_vector 1665 drivers/net/ethernet/intel/iavf/iavf_txrx.c q_vector->rx.current_itr = q_vector->rx.target_itr; q_vector 1666 drivers/net/ethernet/intel/iavf/iavf_txrx.c q_vector->itr_countdown = ITR_COUNTDOWN_START; q_vector 1667 drivers/net/ethernet/intel/iavf/iavf_txrx.c } else if ((q_vector->tx.target_itr < q_vector->tx.current_itr) || q_vector 1668 drivers/net/ethernet/intel/iavf/iavf_txrx.c ((q_vector->rx.target_itr - q_vector->rx.current_itr) < q_vector 1669 drivers/net/ethernet/intel/iavf/iavf_txrx.c (q_vector->tx.target_itr - q_vector->tx.current_itr))) { q_vector 1674 drivers/net/ethernet/intel/iavf/iavf_txrx.c q_vector->tx.target_itr); q_vector 1675 drivers/net/ethernet/intel/iavf/iavf_txrx.c q_vector->tx.current_itr = q_vector->tx.target_itr; q_vector 1676 drivers/net/ethernet/intel/iavf/iavf_txrx.c q_vector->itr_countdown = ITR_COUNTDOWN_START; q_vector 1677 drivers/net/ethernet/intel/iavf/iavf_txrx.c } else if (q_vector->rx.current_itr != q_vector->rx.target_itr) { q_vector 1680 drivers/net/ethernet/intel/iavf/iavf_txrx.c q_vector->rx.target_itr); q_vector 1681 drivers/net/ethernet/intel/iavf/iavf_txrx.c q_vector->rx.current_itr = q_vector->rx.target_itr; q_vector 1682 drivers/net/ethernet/intel/iavf/iavf_txrx.c q_vector->itr_countdown = ITR_COUNTDOWN_START; q_vector 1686 drivers/net/ethernet/intel/iavf/iavf_txrx.c if (q_vector->itr_countdown) q_vector 1687 drivers/net/ethernet/intel/iavf/iavf_txrx.c q_vector->itr_countdown--; q_vector 1691 drivers/net/ethernet/intel/iavf/iavf_txrx.c wr32(hw, INTREG(q_vector->reg_idx), intval); q_vector 1705 drivers/net/ethernet/intel/iavf/iavf_txrx.c struct iavf_q_vector *q_vector = q_vector 1707 drivers/net/ethernet/intel/iavf/iavf_txrx.c struct iavf_vsi *vsi = q_vector->vsi; q_vector 1722 drivers/net/ethernet/intel/iavf/iavf_txrx.c iavf_for_each_ring(ring, q_vector->tx) { q_vector 1738 drivers/net/ethernet/intel/iavf/iavf_txrx.c budget_per_ring = max(budget/q_vector->num_ringpairs, 1); q_vector 1740 drivers/net/ethernet/intel/iavf/iavf_txrx.c iavf_for_each_ring(ring, q_vector->rx) { q_vector 1760 drivers/net/ethernet/intel/iavf/iavf_txrx.c if (!cpumask_test_cpu(cpu_id, &q_vector->affinity_mask)) { q_vector 1765 drivers/net/ethernet/intel/iavf/iavf_txrx.c iavf_force_wb(vsi, q_vector); q_vector 1772 drivers/net/ethernet/intel/iavf/iavf_txrx.c q_vector->tx.ring[0].tx_stats.tx_force_wb++; q_vector 1773 drivers/net/ethernet/intel/iavf/iavf_txrx.c iavf_enable_wb_on_itr(vsi, q_vector); q_vector 1779 drivers/net/ethernet/intel/iavf/iavf_txrx.c q_vector->arm_wb_state = false; q_vector 1785 drivers/net/ethernet/intel/iavf/iavf_txrx.c iavf_update_enable_itr(vsi, q_vector); q_vector 378 drivers/net/ethernet/intel/iavf/iavf_txrx.h struct iavf_q_vector *q_vector; /* Backreference to associated vector */ q_vector 448 drivers/net/ethernet/intel/iavf/iavf_txrx.h void iavf_force_wb(struct iavf_vsi *vsi, struct iavf_q_vector *q_vector); q_vector 354 drivers/net/ethernet/intel/iavf/iavf_virtchnl.c struct iavf_q_vector *q_vector; q_vector 376 drivers/net/ethernet/intel/iavf/iavf_virtchnl.c q_vector = &adapter->q_vectors[v_idx]; q_vector 381 drivers/net/ethernet/intel/iavf/iavf_virtchnl.c vecmap->txq_map = q_vector->ring_mask; q_vector 382 drivers/net/ethernet/intel/iavf/iavf_virtchnl.c vecmap->rxq_map = q_vector->ring_mask; q_vector 391 drivers/net/ethernet/intel/ice/ice.h struct ice_q_vector *q_vector) q_vector 393 drivers/net/ethernet/intel/ice/ice.h u32 vector = (vsi && q_vector) ? q_vector->reg_idx : q_vector 3120 drivers/net/ethernet/intel/ice/ice_ethtool.c ec->rx_coalesce_usecs_high = rc->ring->q_vector->intrl; q_vector 3149 drivers/net/ethernet/intel/ice/ice_ethtool.c &vsi->rx_rings[q_num]->q_vector->rx)) q_vector 3152 drivers/net/ethernet/intel/ice/ice_ethtool.c &vsi->tx_rings[q_num]->q_vector->tx)) q_vector 3156 drivers/net/ethernet/intel/ice/ice_ethtool.c &vsi->rx_rings[q_num]->q_vector->rx)) q_vector 3160 drivers/net/ethernet/intel/ice/ice_ethtool.c &vsi->tx_rings[q_num]->q_vector->tx)) q_vector 3243 drivers/net/ethernet/intel/ice/ice_ethtool.c if (ec->rx_coalesce_usecs_high != rc->ring->q_vector->intrl) { q_vector 3244 drivers/net/ethernet/intel/ice/ice_ethtool.c rc->ring->q_vector->intrl = ec->rx_coalesce_usecs_high; q_vector 3245 drivers/net/ethernet/intel/ice/ice_ethtool.c wr32(&pf->hw, GLINT_RATE(rc->ring->q_vector->reg_idx), q_vector 3322 drivers/net/ethernet/intel/ice/ice_ethtool.c &vsi->rx_rings[q_num]->q_vector->rx, q_vector 3327 drivers/net/ethernet/intel/ice/ice_ethtool.c &vsi->tx_rings[q_num]->q_vector->tx, q_vector 3332 drivers/net/ethernet/intel/ice/ice_ethtool.c &vsi->rx_rings[q_num]->q_vector->rx, q_vector 3337 drivers/net/ethernet/intel/ice/ice_ethtool.c &vsi->tx_rings[q_num]->q_vector->tx, q_vector 521 drivers/net/ethernet/intel/ice/ice_lib.c struct ice_q_vector *q_vector = (struct ice_q_vector *)data; q_vector 523 drivers/net/ethernet/intel/ice/ice_lib.c if (!q_vector->tx.ring && !q_vector->rx.ring) q_vector 526 drivers/net/ethernet/intel/ice/ice_lib.c napi_schedule(&q_vector->napi); q_vector 1107 drivers/net/ethernet/intel/ice/ice_lib.c struct ice_q_vector *q_vector; q_vector 1116 drivers/net/ethernet/intel/ice/ice_lib.c q_vector = vsi->q_vectors[v_idx]; q_vector 1118 drivers/net/ethernet/intel/ice/ice_lib.c ice_for_each_ring(ring, q_vector->tx) q_vector 1119 drivers/net/ethernet/intel/ice/ice_lib.c ring->q_vector = NULL; q_vector 1120 drivers/net/ethernet/intel/ice/ice_lib.c ice_for_each_ring(ring, q_vector->rx) q_vector 1121 drivers/net/ethernet/intel/ice/ice_lib.c ring->q_vector = NULL; q_vector 1125 drivers/net/ethernet/intel/ice/ice_lib.c netif_napi_del(&q_vector->napi); q_vector 1127 drivers/net/ethernet/intel/ice/ice_lib.c devm_kfree(&pf->pdev->dev, q_vector); q_vector 1153 drivers/net/ethernet/intel/ice/ice_lib.c struct ice_q_vector *q_vector; q_vector 1156 drivers/net/ethernet/intel/ice/ice_lib.c q_vector = devm_kzalloc(&pf->pdev->dev, sizeof(*q_vector), GFP_KERNEL); q_vector 1157 drivers/net/ethernet/intel/ice/ice_lib.c if (!q_vector) q_vector 1160 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->vsi = vsi; q_vector 1161 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->v_idx = v_idx; q_vector 1166 drivers/net/ethernet/intel/ice/ice_lib.c cpumask_set_cpu(v_idx, &q_vector->affinity_mask); q_vector 1173 drivers/net/ethernet/intel/ice/ice_lib.c netif_napi_add(vsi->netdev, &q_vector->napi, ice_napi_poll, q_vector 1178 drivers/net/ethernet/intel/ice/ice_lib.c vsi->q_vectors[v_idx] = q_vector; q_vector 1366 drivers/net/ethernet/intel/ice/ice_lib.c struct ice_q_vector *q_vector = vsi->q_vectors[v_id]; q_vector 1371 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->num_ring_tx = tx_rings_per_v; q_vector 1372 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->tx.ring = NULL; q_vector 1373 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->tx.itr_idx = ICE_TX_ITR; q_vector 1379 drivers/net/ethernet/intel/ice/ice_lib.c tx_ring->q_vector = q_vector; q_vector 1380 drivers/net/ethernet/intel/ice/ice_lib.c tx_ring->next = q_vector->tx.ring; q_vector 1381 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->tx.ring = tx_ring; q_vector 1387 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->num_ring_rx = rx_rings_per_v; q_vector 1388 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->rx.ring = NULL; q_vector 1389 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->rx.itr_idx = ICE_RX_ITR; q_vector 1395 drivers/net/ethernet/intel/ice/ice_lib.c rx_ring->q_vector = q_vector; q_vector 1396 drivers/net/ethernet/intel/ice/ice_lib.c rx_ring->next = q_vector->rx.ring; q_vector 1397 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->rx.ring = rx_ring; q_vector 1882 drivers/net/ethernet/intel/ice/ice_lib.c ice_cfg_itr(struct ice_hw *hw, struct ice_q_vector *q_vector) q_vector 1886 drivers/net/ethernet/intel/ice/ice_lib.c if (q_vector->num_ring_rx) { q_vector 1887 drivers/net/ethernet/intel/ice/ice_lib.c struct ice_ring_container *rc = &q_vector->rx; q_vector 1896 drivers/net/ethernet/intel/ice/ice_lib.c wr32(hw, GLINT_ITR(rc->itr_idx, q_vector->reg_idx), q_vector 1900 drivers/net/ethernet/intel/ice/ice_lib.c if (q_vector->num_ring_tx) { q_vector 1901 drivers/net/ethernet/intel/ice/ice_lib.c struct ice_ring_container *rc = &q_vector->tx; q_vector 1910 drivers/net/ethernet/intel/ice/ice_lib.c wr32(hw, GLINT_ITR(rc->itr_idx, q_vector->reg_idx), q_vector 1992 drivers/net/ethernet/intel/ice/ice_lib.c struct ice_q_vector *q_vector = vsi->q_vectors[i]; q_vector 1993 drivers/net/ethernet/intel/ice/ice_lib.c u16 reg_idx = q_vector->reg_idx; q_vector 1995 drivers/net/ethernet/intel/ice/ice_lib.c ice_cfg_itr(hw, q_vector); q_vector 1998 drivers/net/ethernet/intel/ice/ice_lib.c ice_intrl_usec_to_reg(q_vector->intrl, hw->intrl_gran)); q_vector 2011 drivers/net/ethernet/intel/ice/ice_lib.c for (q = 0; q < q_vector->num_ring_tx; q++) { q_vector 2013 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->tx.itr_idx); q_vector 2017 drivers/net/ethernet/intel/ice/ice_lib.c for (q = 0; q < q_vector->num_ring_rx; q++) { q_vector 2019 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->rx.itr_idx); q_vector 2141 drivers/net/ethernet/intel/ice/ice_lib.c void ice_trigger_sw_intr(struct ice_hw *hw, struct ice_q_vector *q_vector) q_vector 2143 drivers/net/ethernet/intel/ice/ice_lib.c wr32(hw, GLINT_DYN_CTL(q_vector->reg_idx), q_vector 2166 drivers/net/ethernet/intel/ice/ice_lib.c struct ice_q_vector *q_vector; q_vector 2182 drivers/net/ethernet/intel/ice/ice_lib.c q_vector = ring->q_vector; q_vector 2183 drivers/net/ethernet/intel/ice/ice_lib.c if (q_vector) q_vector 2184 drivers/net/ethernet/intel/ice/ice_lib.c ice_trigger_sw_intr(hw, q_vector); q_vector 2379 drivers/net/ethernet/intel/ice/ice_lib.c struct ice_q_vector *q_vector = vsi->q_vectors[i]; q_vector 2381 drivers/net/ethernet/intel/ice/ice_lib.c if (!q_vector) { q_vector 2391 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->reg_idx = ice_calc_vf_reg_idx(vf, q_vector); q_vector 2393 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->reg_idx = q_vector 2394 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->v_idx + vsi->base_vector; q_vector 2402 drivers/net/ethernet/intel/ice/ice_lib.c struct ice_q_vector *q_vector = vsi->q_vectors[i]; q_vector 2404 drivers/net/ethernet/intel/ice/ice_lib.c if (q_vector) q_vector 2405 drivers/net/ethernet/intel/ice/ice_lib.c q_vector->reg_idx = 0; q_vector 2686 drivers/net/ethernet/intel/ice/ice_lib.c struct ice_q_vector *q_vector = vsi->q_vectors[i]; q_vector 2687 drivers/net/ethernet/intel/ice/ice_lib.c u16 reg_idx = q_vector->reg_idx; q_vector 2691 drivers/net/ethernet/intel/ice/ice_lib.c for (q = 0; q < q_vector->num_ring_tx; q++) { q_vector 2696 drivers/net/ethernet/intel/ice/ice_lib.c for (q = 0; q < q_vector->num_ring_rx; q++) { q_vector 103 drivers/net/ethernet/intel/ice/ice_lib.h void ice_trigger_sw_intr(struct ice_hw *hw, struct ice_q_vector *q_vector); q_vector 107 drivers/net/ethernet/intel/ice/ice_main.c ice_trigger_sw_intr(hw, tx_ring->q_vector); q_vector 1565 drivers/net/ethernet/intel/ice/ice_main.c struct ice_q_vector *q_vector = q_vector 1568 drivers/net/ethernet/intel/ice/ice_main.c cpumask_copy(&q_vector->affinity_mask, mask); q_vector 1613 drivers/net/ethernet/intel/ice/ice_main.c struct ice_q_vector *q_vector = vsi->q_vectors[vector]; q_vector 1617 drivers/net/ethernet/intel/ice/ice_main.c if (q_vector->tx.ring && q_vector->rx.ring) { q_vector 1618 drivers/net/ethernet/intel/ice/ice_main.c snprintf(q_vector->name, sizeof(q_vector->name) - 1, q_vector 1621 drivers/net/ethernet/intel/ice/ice_main.c } else if (q_vector->rx.ring) { q_vector 1622 drivers/net/ethernet/intel/ice/ice_main.c snprintf(q_vector->name, sizeof(q_vector->name) - 1, q_vector 1624 drivers/net/ethernet/intel/ice/ice_main.c } else if (q_vector->tx.ring) { q_vector 1625 drivers/net/ethernet/intel/ice/ice_main.c snprintf(q_vector->name, sizeof(q_vector->name) - 1, q_vector 1633 drivers/net/ethernet/intel/ice/ice_main.c q_vector->name, q_vector); q_vector 1641 drivers/net/ethernet/intel/ice/ice_main.c q_vector->affinity_notify.notify = ice_irq_affinity_notify; q_vector 1642 drivers/net/ethernet/intel/ice/ice_main.c q_vector->affinity_notify.release = ice_irq_affinity_release; q_vector 1643 drivers/net/ethernet/intel/ice/ice_main.c irq_set_affinity_notifier(irq_num, &q_vector->affinity_notify); q_vector 1646 drivers/net/ethernet/intel/ice/ice_main.c irq_set_affinity_hint(irq_num, &q_vector->affinity_mask); q_vector 3526 drivers/net/ethernet/intel/ice/ice_main.c struct ice_q_vector *q_vector = vsi->q_vectors[q_idx]; q_vector 3528 drivers/net/ethernet/intel/ice/ice_main.c if (q_vector->rx.ring || q_vector->tx.ring) q_vector 3529 drivers/net/ethernet/intel/ice/ice_main.c napi_enable(&q_vector->napi); q_vector 3893 drivers/net/ethernet/intel/ice/ice_main.c struct ice_q_vector *q_vector = vsi->q_vectors[q_idx]; q_vector 3895 drivers/net/ethernet/intel/ice/ice_main.c if (q_vector->rx.ring || q_vector->tx.ring) q_vector 3896 drivers/net/ethernet/intel/ice/ice_main.c napi_disable(&q_vector->napi); q_vector 4693 drivers/net/ethernet/intel/ice/ice_main.c val = rd32(hw, GLINT_DYN_CTL(tx_ring->q_vector->reg_idx)); q_vector 195 drivers/net/ethernet/intel/ice/ice_txrx.c tx_ring->q_vector->tx.total_bytes += total_bytes; q_vector 196 drivers/net/ethernet/intel/ice/ice_txrx.c tx_ring->q_vector->tx.total_pkts += total_pkts; q_vector 706 drivers/net/ethernet/intel/ice/ice_txrx.c skb = __napi_alloc_skb(&rx_ring->q_vector->napi, ICE_RX_HDR_SIZE, q_vector 990 drivers/net/ethernet/intel/ice/ice_txrx.c napi_gro_receive(&rx_ring->q_vector->napi, skb); q_vector 1107 drivers/net/ethernet/intel/ice/ice_txrx.c rx_ring->q_vector->rx.total_pkts += total_rx_pkts; q_vector 1108 drivers/net/ethernet/intel/ice/ice_txrx.c rx_ring->q_vector->rx.total_bytes += total_rx_bytes; q_vector 1193 drivers/net/ethernet/intel/ice/ice_txrx.c ice_update_itr(struct ice_q_vector *q_vector, struct ice_ring_container *rc) q_vector 1208 drivers/net/ethernet/intel/ice/ice_txrx.c if (q_vector->itr_countdown) { q_vector 1213 drivers/net/ethernet/intel/ice/ice_txrx.c container_is_rx = (&q_vector->rx == rc); q_vector 1229 drivers/net/ethernet/intel/ice/ice_txrx.c prefetch(q_vector->vsi->port_info); q_vector 1241 drivers/net/ethernet/intel/ice/ice_txrx.c (q_vector->tx.target_itr & ICE_ITR_ADAPTIVE_LATENCY)) { q_vector 1252 drivers/net/ethernet/intel/ice/ice_txrx.c (q_vector->rx.target_itr & ICE_ITR_MASK) == q_vector 1280 drivers/net/ethernet/intel/ice/ice_txrx.c itr = min(q_vector->tx.current_itr, q_vector->rx.current_itr); q_vector 1314 drivers/net/ethernet/intel/ice/ice_txrx.c itr = ice_adjust_itr_by_size_and_speed(q_vector->vsi->port_info, q_vector 1362 drivers/net/ethernet/intel/ice/ice_txrx.c static void ice_update_ena_itr(struct ice_q_vector *q_vector) q_vector 1364 drivers/net/ethernet/intel/ice/ice_txrx.c struct ice_ring_container *tx = &q_vector->tx; q_vector 1365 drivers/net/ethernet/intel/ice/ice_txrx.c struct ice_ring_container *rx = &q_vector->rx; q_vector 1366 drivers/net/ethernet/intel/ice/ice_txrx.c struct ice_vsi *vsi = q_vector->vsi; q_vector 1373 drivers/net/ethernet/intel/ice/ice_txrx.c if (q_vector->itr_countdown == ICE_IN_WB_ON_ITR_MODE) { q_vector 1375 drivers/net/ethernet/intel/ice/ice_txrx.c wr32(&vsi->back->hw, GLINT_DYN_CTL(q_vector->reg_idx), itr_val); q_vector 1382 drivers/net/ethernet/intel/ice/ice_txrx.c q_vector->itr_countdown = 0; q_vector 1387 drivers/net/ethernet/intel/ice/ice_txrx.c ice_update_itr(q_vector, tx); q_vector 1388 drivers/net/ethernet/intel/ice/ice_txrx.c ice_update_itr(q_vector, rx); q_vector 1402 drivers/net/ethernet/intel/ice/ice_txrx.c q_vector->itr_countdown = ITR_COUNTDOWN_START; q_vector 1411 drivers/net/ethernet/intel/ice/ice_txrx.c q_vector->itr_countdown = ITR_COUNTDOWN_START; q_vector 1416 drivers/net/ethernet/intel/ice/ice_txrx.c q_vector->itr_countdown = ITR_COUNTDOWN_START; q_vector 1420 drivers/net/ethernet/intel/ice/ice_txrx.c if (q_vector->itr_countdown) q_vector 1421 drivers/net/ethernet/intel/ice/ice_txrx.c q_vector->itr_countdown--; q_vector 1424 drivers/net/ethernet/intel/ice/ice_txrx.c if (!test_bit(__ICE_DOWN, q_vector->vsi->state)) q_vector 1425 drivers/net/ethernet/intel/ice/ice_txrx.c wr32(&q_vector->vsi->back->hw, q_vector 1426 drivers/net/ethernet/intel/ice/ice_txrx.c GLINT_DYN_CTL(q_vector->reg_idx), q_vector 1444 drivers/net/ethernet/intel/ice/ice_txrx.c static void ice_set_wb_on_itr(struct ice_q_vector *q_vector) q_vector 1446 drivers/net/ethernet/intel/ice/ice_txrx.c struct ice_vsi *vsi = q_vector->vsi; q_vector 1449 drivers/net/ethernet/intel/ice/ice_txrx.c if (q_vector->itr_countdown == ICE_IN_WB_ON_ITR_MODE) q_vector 1452 drivers/net/ethernet/intel/ice/ice_txrx.c if (q_vector->num_ring_rx) q_vector 1453 drivers/net/ethernet/intel/ice/ice_txrx.c wr32(&vsi->back->hw, GLINT_DYN_CTL(q_vector->reg_idx), q_vector 1457 drivers/net/ethernet/intel/ice/ice_txrx.c if (q_vector->num_ring_tx) q_vector 1458 drivers/net/ethernet/intel/ice/ice_txrx.c wr32(&vsi->back->hw, GLINT_DYN_CTL(q_vector->reg_idx), q_vector 1462 drivers/net/ethernet/intel/ice/ice_txrx.c q_vector->itr_countdown = ICE_IN_WB_ON_ITR_MODE; q_vector 1476 drivers/net/ethernet/intel/ice/ice_txrx.c struct ice_q_vector *q_vector = q_vector 1486 drivers/net/ethernet/intel/ice/ice_txrx.c ice_for_each_ring(ring, q_vector->tx) q_vector 1495 drivers/net/ethernet/intel/ice/ice_txrx.c if (unlikely(q_vector->num_ring_rx > 1)) q_vector 1500 drivers/net/ethernet/intel/ice/ice_txrx.c budget_per_ring = max(budget / q_vector->num_ring_rx, 1); q_vector 1505 drivers/net/ethernet/intel/ice/ice_txrx.c ice_for_each_ring(ring, q_vector->rx) { q_vector 1523 drivers/net/ethernet/intel/ice/ice_txrx.c ice_update_ena_itr(q_vector); q_vector 1525 drivers/net/ethernet/intel/ice/ice_txrx.c ice_set_wb_on_itr(q_vector); q_vector 172 drivers/net/ethernet/intel/ice/ice_txrx.h struct ice_q_vector *q_vector; /* Backreference to associated vector */ q_vector 757 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c int ice_calc_vf_reg_idx(struct ice_vf *vf, struct ice_q_vector *q_vector) q_vector 761 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c if (!vf || !q_vector) q_vector 768 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c q_vector->v_idx + 1; q_vector 2159 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c struct ice_q_vector *q_vector; q_vector 2180 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c q_vector = vsi->q_vectors[vector_id - ICE_NONQ_VECS_VF]; q_vector 2181 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c if (!q_vector) { q_vector 2188 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c q_vector->num_ring_rx = 0; q_vector 2194 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c q_vector->num_ring_rx++; q_vector 2195 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c q_vector->rx.itr_idx = map->rxitr_idx; q_vector 2196 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c vsi->rx_rings[vsi_q_id]->q_vector = q_vector; q_vector 2198 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c q_vector->rx.itr_idx); q_vector 2202 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c q_vector->num_ring_tx = 0; q_vector 2208 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c q_vector->num_ring_tx++; q_vector 2209 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c q_vector->tx.itr_idx = map->txitr_idx; q_vector 2210 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c vsi->tx_rings[vsi_q_id]->q_vector = q_vector; q_vector 2212 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c q_vector->tx.itr_idx); q_vector 121 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.h int ice_calc_vf_reg_idx(struct ice_vf *vf, struct ice_q_vector *q_vector); q_vector 192 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.h struct ice_q_vector __always_unused *q_vector) q_vector 249 drivers/net/ethernet/intel/igb/igb.h struct igb_q_vector *q_vector; /* backlink to q_vector */ q_vector 532 drivers/net/ethernet/intel/igb/igb.h struct igb_q_vector *q_vector[MAX_Q_VECTORS]; q_vector 677 drivers/net/ethernet/intel/igb/igb.h void igb_ptp_rx_rgtstamp(struct igb_q_vector *q_vector, struct sk_buff *skb); q_vector 678 drivers/net/ethernet/intel/igb/igb.h void igb_ptp_rx_pktstamp(struct igb_q_vector *q_vector, void *va, q_vector 2235 drivers/net/ethernet/intel/igb/igb_ethtool.c struct igb_q_vector *q_vector = adapter->q_vector[i]; q_vector 2236 drivers/net/ethernet/intel/igb/igb_ethtool.c q_vector->tx.work_limit = adapter->tx_work_limit; q_vector 2237 drivers/net/ethernet/intel/igb/igb_ethtool.c if (q_vector->rx.ring) q_vector 2238 drivers/net/ethernet/intel/igb/igb_ethtool.c q_vector->itr_val = adapter->rx_itr_setting; q_vector 2240 drivers/net/ethernet/intel/igb/igb_ethtool.c q_vector->itr_val = adapter->tx_itr_setting; q_vector 2241 drivers/net/ethernet/intel/igb/igb_ethtool.c if (q_vector->itr_val && q_vector->itr_val <= 3) q_vector 2242 drivers/net/ethernet/intel/igb/igb_ethtool.c q_vector->itr_val = IGB_START_ITR; q_vector 2243 drivers/net/ethernet/intel/igb/igb_ethtool.c q_vector->set_itr = 1; q_vector 790 drivers/net/ethernet/intel/igb/igb_main.c static void igb_assign_vector(struct igb_q_vector *q_vector, int msix_vector) q_vector 792 drivers/net/ethernet/intel/igb/igb_main.c struct igb_adapter *adapter = q_vector->adapter; q_vector 798 drivers/net/ethernet/intel/igb/igb_main.c if (q_vector->rx.ring) q_vector 799 drivers/net/ethernet/intel/igb/igb_main.c rx_queue = q_vector->rx.ring->reg_idx; q_vector 800 drivers/net/ethernet/intel/igb/igb_main.c if (q_vector->tx.ring) q_vector 801 drivers/net/ethernet/intel/igb/igb_main.c tx_queue = q_vector->tx.ring->reg_idx; q_vector 817 drivers/net/ethernet/intel/igb/igb_main.c q_vector->eims_value = msixbm; q_vector 833 drivers/net/ethernet/intel/igb/igb_main.c q_vector->eims_value = BIT(msix_vector); q_vector 854 drivers/net/ethernet/intel/igb/igb_main.c q_vector->eims_value = BIT(msix_vector); q_vector 862 drivers/net/ethernet/intel/igb/igb_main.c adapter->eims_enable_mask |= q_vector->eims_value; q_vector 865 drivers/net/ethernet/intel/igb/igb_main.c q_vector->set_itr = 1; q_vector 929 drivers/net/ethernet/intel/igb/igb_main.c igb_assign_vector(adapter->q_vector[i], vector++); q_vector 952 drivers/net/ethernet/intel/igb/igb_main.c struct igb_q_vector *q_vector = adapter->q_vector[i]; q_vector 956 drivers/net/ethernet/intel/igb/igb_main.c q_vector->itr_register = adapter->io_addr + E1000_EITR(vector); q_vector 958 drivers/net/ethernet/intel/igb/igb_main.c if (q_vector->rx.ring && q_vector->tx.ring) q_vector 959 drivers/net/ethernet/intel/igb/igb_main.c sprintf(q_vector->name, "%s-TxRx-%u", netdev->name, q_vector 960 drivers/net/ethernet/intel/igb/igb_main.c q_vector->rx.ring->queue_index); q_vector 961 drivers/net/ethernet/intel/igb/igb_main.c else if (q_vector->tx.ring) q_vector 962 drivers/net/ethernet/intel/igb/igb_main.c sprintf(q_vector->name, "%s-tx-%u", netdev->name, q_vector 963 drivers/net/ethernet/intel/igb/igb_main.c q_vector->tx.ring->queue_index); q_vector 964 drivers/net/ethernet/intel/igb/igb_main.c else if (q_vector->rx.ring) q_vector 965 drivers/net/ethernet/intel/igb/igb_main.c sprintf(q_vector->name, "%s-rx-%u", netdev->name, q_vector 966 drivers/net/ethernet/intel/igb/igb_main.c q_vector->rx.ring->queue_index); q_vector 968 drivers/net/ethernet/intel/igb/igb_main.c sprintf(q_vector->name, "%s-unused", netdev->name); q_vector 971 drivers/net/ethernet/intel/igb/igb_main.c igb_msix_ring, 0, q_vector->name, q_vector 972 drivers/net/ethernet/intel/igb/igb_main.c q_vector); q_vector 987 drivers/net/ethernet/intel/igb/igb_main.c adapter->q_vector[i]); q_vector 1002 drivers/net/ethernet/intel/igb/igb_main.c struct igb_q_vector *q_vector = adapter->q_vector[v_idx]; q_vector 1004 drivers/net/ethernet/intel/igb/igb_main.c adapter->q_vector[v_idx] = NULL; q_vector 1009 drivers/net/ethernet/intel/igb/igb_main.c if (q_vector) q_vector 1010 drivers/net/ethernet/intel/igb/igb_main.c kfree_rcu(q_vector, rcu); q_vector 1023 drivers/net/ethernet/intel/igb/igb_main.c struct igb_q_vector *q_vector = adapter->q_vector[v_idx]; q_vector 1028 drivers/net/ethernet/intel/igb/igb_main.c if (!q_vector) q_vector 1031 drivers/net/ethernet/intel/igb/igb_main.c if (q_vector->tx.ring) q_vector 1032 drivers/net/ethernet/intel/igb/igb_main.c adapter->tx_ring[q_vector->tx.ring->queue_index] = NULL; q_vector 1034 drivers/net/ethernet/intel/igb/igb_main.c if (q_vector->rx.ring) q_vector 1035 drivers/net/ethernet/intel/igb/igb_main.c adapter->rx_ring[q_vector->rx.ring->queue_index] = NULL; q_vector 1037 drivers/net/ethernet/intel/igb/igb_main.c netif_napi_del(&q_vector->napi); q_vector 1192 drivers/net/ethernet/intel/igb/igb_main.c struct igb_q_vector *q_vector; q_vector 1202 drivers/net/ethernet/intel/igb/igb_main.c size = struct_size(q_vector, ring, ring_count); q_vector 1205 drivers/net/ethernet/intel/igb/igb_main.c q_vector = adapter->q_vector[v_idx]; q_vector 1206 drivers/net/ethernet/intel/igb/igb_main.c if (!q_vector) { q_vector 1207 drivers/net/ethernet/intel/igb/igb_main.c q_vector = kzalloc(size, GFP_KERNEL); q_vector 1208 drivers/net/ethernet/intel/igb/igb_main.c } else if (size > ksize(q_vector)) { q_vector 1209 drivers/net/ethernet/intel/igb/igb_main.c kfree_rcu(q_vector, rcu); q_vector 1210 drivers/net/ethernet/intel/igb/igb_main.c q_vector = kzalloc(size, GFP_KERNEL); q_vector 1212 drivers/net/ethernet/intel/igb/igb_main.c memset(q_vector, 0, size); q_vector 1214 drivers/net/ethernet/intel/igb/igb_main.c if (!q_vector) q_vector 1218 drivers/net/ethernet/intel/igb/igb_main.c netif_napi_add(adapter->netdev, &q_vector->napi, q_vector 1222 drivers/net/ethernet/intel/igb/igb_main.c adapter->q_vector[v_idx] = q_vector; q_vector 1223 drivers/net/ethernet/intel/igb/igb_main.c q_vector->adapter = adapter; q_vector 1226 drivers/net/ethernet/intel/igb/igb_main.c q_vector->tx.work_limit = adapter->tx_work_limit; q_vector 1229 drivers/net/ethernet/intel/igb/igb_main.c q_vector->itr_register = adapter->io_addr + E1000_EITR(0); q_vector 1230 drivers/net/ethernet/intel/igb/igb_main.c q_vector->itr_val = IGB_START_ITR; q_vector 1233 drivers/net/ethernet/intel/igb/igb_main.c ring = q_vector->ring; q_vector 1239 drivers/net/ethernet/intel/igb/igb_main.c q_vector->itr_val = adapter->rx_itr_setting; q_vector 1243 drivers/net/ethernet/intel/igb/igb_main.c q_vector->itr_val = adapter->tx_itr_setting; q_vector 1252 drivers/net/ethernet/intel/igb/igb_main.c ring->q_vector = q_vector; q_vector 1255 drivers/net/ethernet/intel/igb/igb_main.c igb_add_ring(ring, &q_vector->tx); q_vector 1287 drivers/net/ethernet/intel/igb/igb_main.c ring->q_vector = q_vector; q_vector 1290 drivers/net/ethernet/intel/igb/igb_main.c igb_add_ring(ring, &q_vector->rx); q_vector 1435 drivers/net/ethernet/intel/igb/igb_main.c igb_assign_vector(adapter->q_vector[0], 0); q_vector 1468 drivers/net/ethernet/intel/igb/igb_main.c adapter->q_vector[i]); q_vector 2131 drivers/net/ethernet/intel/igb/igb_main.c napi_enable(&(adapter->q_vector[i]->napi)); q_vector 2136 drivers/net/ethernet/intel/igb/igb_main.c igb_assign_vector(adapter->q_vector[0], 0); q_vector 2199 drivers/net/ethernet/intel/igb/igb_main.c if (adapter->q_vector[i]) { q_vector 2200 drivers/net/ethernet/intel/igb/igb_main.c napi_synchronize(&adapter->q_vector[i]->napi); q_vector 2201 drivers/net/ethernet/intel/igb/igb_main.c napi_disable(&adapter->q_vector[i]->napi); q_vector 3945 drivers/net/ethernet/intel/igb/igb_main.c napi_enable(&(adapter->q_vector[i]->napi)); q_vector 5415 drivers/net/ethernet/intel/igb/igb_main.c eics |= adapter->q_vector[i]->eims_value; q_vector 5463 drivers/net/ethernet/intel/igb/igb_main.c static void igb_update_ring_itr(struct igb_q_vector *q_vector) q_vector 5465 drivers/net/ethernet/intel/igb/igb_main.c int new_val = q_vector->itr_val; q_vector 5467 drivers/net/ethernet/intel/igb/igb_main.c struct igb_adapter *adapter = q_vector->adapter; q_vector 5478 drivers/net/ethernet/intel/igb/igb_main.c packets = q_vector->rx.total_packets; q_vector 5480 drivers/net/ethernet/intel/igb/igb_main.c avg_wire_size = q_vector->rx.total_bytes / packets; q_vector 5482 drivers/net/ethernet/intel/igb/igb_main.c packets = q_vector->tx.total_packets; q_vector 5485 drivers/net/ethernet/intel/igb/igb_main.c q_vector->tx.total_bytes / packets); q_vector 5505 drivers/net/ethernet/intel/igb/igb_main.c ((q_vector->rx.ring && adapter->rx_itr_setting == 3) || q_vector 5506 drivers/net/ethernet/intel/igb/igb_main.c (!q_vector->rx.ring && adapter->tx_itr_setting == 3))) q_vector 5510 drivers/net/ethernet/intel/igb/igb_main.c if (new_val != q_vector->itr_val) { q_vector 5511 drivers/net/ethernet/intel/igb/igb_main.c q_vector->itr_val = new_val; q_vector 5512 drivers/net/ethernet/intel/igb/igb_main.c q_vector->set_itr = 1; q_vector 5515 drivers/net/ethernet/intel/igb/igb_main.c q_vector->rx.total_bytes = 0; q_vector 5516 drivers/net/ethernet/intel/igb/igb_main.c q_vector->rx.total_packets = 0; q_vector 5517 drivers/net/ethernet/intel/igb/igb_main.c q_vector->tx.total_bytes = 0; q_vector 5518 drivers/net/ethernet/intel/igb/igb_main.c q_vector->tx.total_packets = 0; q_vector 5537 drivers/net/ethernet/intel/igb/igb_main.c static void igb_update_itr(struct igb_q_vector *q_vector, q_vector 5589 drivers/net/ethernet/intel/igb/igb_main.c static void igb_set_itr(struct igb_q_vector *q_vector) q_vector 5591 drivers/net/ethernet/intel/igb/igb_main.c struct igb_adapter *adapter = q_vector->adapter; q_vector 5592 drivers/net/ethernet/intel/igb/igb_main.c u32 new_itr = q_vector->itr_val; q_vector 5602 drivers/net/ethernet/intel/igb/igb_main.c igb_update_itr(q_vector, &q_vector->tx); q_vector 5603 drivers/net/ethernet/intel/igb/igb_main.c igb_update_itr(q_vector, &q_vector->rx); q_vector 5605 drivers/net/ethernet/intel/igb/igb_main.c current_itr = max(q_vector->rx.itr, q_vector->tx.itr); q_vector 5609 drivers/net/ethernet/intel/igb/igb_main.c ((q_vector->rx.ring && adapter->rx_itr_setting == 3) || q_vector 5610 drivers/net/ethernet/intel/igb/igb_main.c (!q_vector->rx.ring && adapter->tx_itr_setting == 3))) q_vector 5629 drivers/net/ethernet/intel/igb/igb_main.c if (new_itr != q_vector->itr_val) { q_vector 5634 drivers/net/ethernet/intel/igb/igb_main.c new_itr = new_itr > q_vector->itr_val ? q_vector 5635 drivers/net/ethernet/intel/igb/igb_main.c max((new_itr * q_vector->itr_val) / q_vector 5636 drivers/net/ethernet/intel/igb/igb_main.c (new_itr + (q_vector->itr_val >> 2)), q_vector 5644 drivers/net/ethernet/intel/igb/igb_main.c q_vector->itr_val = new_itr; q_vector 5645 drivers/net/ethernet/intel/igb/igb_main.c q_vector->set_itr = 1; q_vector 6560 drivers/net/ethernet/intel/igb/igb_main.c static void igb_write_itr(struct igb_q_vector *q_vector) q_vector 6562 drivers/net/ethernet/intel/igb/igb_main.c struct igb_adapter *adapter = q_vector->adapter; q_vector 6563 drivers/net/ethernet/intel/igb/igb_main.c u32 itr_val = q_vector->itr_val & 0x7FFC; q_vector 6565 drivers/net/ethernet/intel/igb/igb_main.c if (!q_vector->set_itr) q_vector 6576 drivers/net/ethernet/intel/igb/igb_main.c writel(itr_val, q_vector->itr_register); q_vector 6577 drivers/net/ethernet/intel/igb/igb_main.c q_vector->set_itr = 0; q_vector 6582 drivers/net/ethernet/intel/igb/igb_main.c struct igb_q_vector *q_vector = data; q_vector 6585 drivers/net/ethernet/intel/igb/igb_main.c igb_write_itr(q_vector); q_vector 6587 drivers/net/ethernet/intel/igb/igb_main.c napi_schedule(&q_vector->napi); q_vector 6634 drivers/net/ethernet/intel/igb/igb_main.c static void igb_update_dca(struct igb_q_vector *q_vector) q_vector 6636 drivers/net/ethernet/intel/igb/igb_main.c struct igb_adapter *adapter = q_vector->adapter; q_vector 6639 drivers/net/ethernet/intel/igb/igb_main.c if (q_vector->cpu == cpu) q_vector 6642 drivers/net/ethernet/intel/igb/igb_main.c if (q_vector->tx.ring) q_vector 6643 drivers/net/ethernet/intel/igb/igb_main.c igb_update_tx_dca(adapter, q_vector->tx.ring, cpu); q_vector 6645 drivers/net/ethernet/intel/igb/igb_main.c if (q_vector->rx.ring) q_vector 6646 drivers/net/ethernet/intel/igb/igb_main.c igb_update_rx_dca(adapter, q_vector->rx.ring, cpu); q_vector 6648 drivers/net/ethernet/intel/igb/igb_main.c q_vector->cpu = cpu; q_vector 6665 drivers/net/ethernet/intel/igb/igb_main.c adapter->q_vector[i]->cpu = -1; q_vector 6666 drivers/net/ethernet/intel/igb/igb_main.c igb_update_dca(adapter->q_vector[i]); q_vector 7601 drivers/net/ethernet/intel/igb/igb_main.c struct igb_q_vector *q_vector = adapter->q_vector[0]; q_vector 7606 drivers/net/ethernet/intel/igb/igb_main.c igb_write_itr(q_vector); q_vector 7625 drivers/net/ethernet/intel/igb/igb_main.c napi_schedule(&q_vector->napi); q_vector 7638 drivers/net/ethernet/intel/igb/igb_main.c struct igb_q_vector *q_vector = adapter->q_vector[0]; q_vector 7651 drivers/net/ethernet/intel/igb/igb_main.c igb_write_itr(q_vector); q_vector 7671 drivers/net/ethernet/intel/igb/igb_main.c napi_schedule(&q_vector->napi); q_vector 7676 drivers/net/ethernet/intel/igb/igb_main.c static void igb_ring_irq_enable(struct igb_q_vector *q_vector) q_vector 7678 drivers/net/ethernet/intel/igb/igb_main.c struct igb_adapter *adapter = q_vector->adapter; q_vector 7681 drivers/net/ethernet/intel/igb/igb_main.c if ((q_vector->rx.ring && (adapter->rx_itr_setting & 3)) || q_vector 7682 drivers/net/ethernet/intel/igb/igb_main.c (!q_vector->rx.ring && (adapter->tx_itr_setting & 3))) { q_vector 7684 drivers/net/ethernet/intel/igb/igb_main.c igb_set_itr(q_vector); q_vector 7686 drivers/net/ethernet/intel/igb/igb_main.c igb_update_ring_itr(q_vector); q_vector 7691 drivers/net/ethernet/intel/igb/igb_main.c wr32(E1000_EIMS, q_vector->eims_value); q_vector 7704 drivers/net/ethernet/intel/igb/igb_main.c struct igb_q_vector *q_vector = container_of(napi, q_vector 7711 drivers/net/ethernet/intel/igb/igb_main.c if (q_vector->adapter->flags & IGB_FLAG_DCA_ENABLED) q_vector 7712 drivers/net/ethernet/intel/igb/igb_main.c igb_update_dca(q_vector); q_vector 7714 drivers/net/ethernet/intel/igb/igb_main.c if (q_vector->tx.ring) q_vector 7715 drivers/net/ethernet/intel/igb/igb_main.c clean_complete = igb_clean_tx_irq(q_vector, budget); q_vector 7717 drivers/net/ethernet/intel/igb/igb_main.c if (q_vector->rx.ring) { q_vector 7718 drivers/net/ethernet/intel/igb/igb_main.c int cleaned = igb_clean_rx_irq(q_vector, budget); q_vector 7733 drivers/net/ethernet/intel/igb/igb_main.c igb_ring_irq_enable(q_vector); q_vector 7745 drivers/net/ethernet/intel/igb/igb_main.c static bool igb_clean_tx_irq(struct igb_q_vector *q_vector, int napi_budget) q_vector 7747 drivers/net/ethernet/intel/igb/igb_main.c struct igb_adapter *adapter = q_vector->adapter; q_vector 7748 drivers/net/ethernet/intel/igb/igb_main.c struct igb_ring *tx_ring = q_vector->tx.ring; q_vector 7752 drivers/net/ethernet/intel/igb/igb_main.c unsigned int budget = q_vector->tx.work_limit; q_vector 7841 drivers/net/ethernet/intel/igb/igb_main.c q_vector->tx.total_bytes += total_bytes; q_vector 7842 drivers/net/ethernet/intel/igb/igb_main.c q_vector->tx.total_packets += total_packets; q_vector 8027 drivers/net/ethernet/intel/igb/igb_main.c skb = napi_alloc_skb(&rx_ring->q_vector->napi, IGB_RX_HDR_LEN); q_vector 8032 drivers/net/ethernet/intel/igb/igb_main.c igb_ptp_rx_pktstamp(rx_ring->q_vector, va, skb); q_vector 8094 drivers/net/ethernet/intel/igb/igb_main.c igb_ptp_rx_pktstamp(rx_ring->q_vector, skb->data, skb); q_vector 8242 drivers/net/ethernet/intel/igb/igb_main.c igb_ptp_rx_rgtstamp(rx_ring->q_vector, skb); q_vector 8303 drivers/net/ethernet/intel/igb/igb_main.c static int igb_clean_rx_irq(struct igb_q_vector *q_vector, const int budget) q_vector 8305 drivers/net/ethernet/intel/igb/igb_main.c struct igb_ring *rx_ring = q_vector->rx.ring; q_vector 8369 drivers/net/ethernet/intel/igb/igb_main.c napi_gro_receive(&q_vector->napi, skb); q_vector 8385 drivers/net/ethernet/intel/igb/igb_main.c q_vector->rx.total_packets += total_packets; q_vector 8386 drivers/net/ethernet/intel/igb/igb_main.c q_vector->rx.total_bytes += total_bytes; q_vector 869 drivers/net/ethernet/intel/igb/igb_ptp.c void igb_ptp_rx_pktstamp(struct igb_q_vector *q_vector, void *va, q_vector 873 drivers/net/ethernet/intel/igb/igb_ptp.c struct igb_adapter *adapter = q_vector->adapter; q_vector 909 drivers/net/ethernet/intel/igb/igb_ptp.c void igb_ptp_rx_rgtstamp(struct igb_q_vector *q_vector, q_vector 912 drivers/net/ethernet/intel/igb/igb_ptp.c struct igb_adapter *adapter = q_vector->adapter; q_vector 244 drivers/net/ethernet/intel/igc/igc.h struct igc_q_vector *q_vector; /* backlink to q_vector */ q_vector 404 drivers/net/ethernet/intel/igc/igc.h struct igc_q_vector *q_vector[MAX_Q_VECTORS]; q_vector 862 drivers/net/ethernet/intel/igc/igc_ethtool.c struct igc_q_vector *q_vector = adapter->q_vector[i]; q_vector 864 drivers/net/ethernet/intel/igc/igc_ethtool.c q_vector->tx.work_limit = adapter->tx_work_limit; q_vector 865 drivers/net/ethernet/intel/igc/igc_ethtool.c if (q_vector->rx.ring) q_vector 866 drivers/net/ethernet/intel/igc/igc_ethtool.c q_vector->itr_val = adapter->rx_itr_setting; q_vector 868 drivers/net/ethernet/intel/igc/igc_ethtool.c q_vector->itr_val = adapter->tx_itr_setting; q_vector 869 drivers/net/ethernet/intel/igc/igc_ethtool.c if (q_vector->itr_val && q_vector->itr_val <= 3) q_vector 870 drivers/net/ethernet/intel/igc/igc_ethtool.c q_vector->itr_val = IGC_START_ITR; q_vector 871 drivers/net/ethernet/intel/igc/igc_ethtool.c q_vector->set_itr = 1; q_vector 60 drivers/net/ethernet/intel/igc/igc_main.c static void igc_write_itr(struct igc_q_vector *q_vector); q_vector 61 drivers/net/ethernet/intel/igc/igc_main.c static void igc_assign_vector(struct igc_q_vector *q_vector, int msix_vector); q_vector 1307 drivers/net/ethernet/intel/igc/igc_main.c skb = napi_alloc_skb(&rx_ring->q_vector->napi, IGC_RX_HDR_LEN); q_vector 1557 drivers/net/ethernet/intel/igc/igc_main.c static int igc_clean_rx_irq(struct igc_q_vector *q_vector, const int budget) q_vector 1560 drivers/net/ethernet/intel/igc/igc_main.c struct igc_ring *rx_ring = q_vector->rx.ring; q_vector 1623 drivers/net/ethernet/intel/igc/igc_main.c napi_gro_receive(&q_vector->napi, skb); q_vector 1639 drivers/net/ethernet/intel/igc/igc_main.c q_vector->rx.total_packets += total_packets; q_vector 1640 drivers/net/ethernet/intel/igc/igc_main.c q_vector->rx.total_bytes += total_bytes; q_vector 1701 drivers/net/ethernet/intel/igc/igc_main.c static bool igc_clean_tx_irq(struct igc_q_vector *q_vector, int napi_budget) q_vector 1703 drivers/net/ethernet/intel/igc/igc_main.c struct igc_adapter *adapter = q_vector->adapter; q_vector 1705 drivers/net/ethernet/intel/igc/igc_main.c unsigned int budget = q_vector->tx.work_limit; q_vector 1706 drivers/net/ethernet/intel/igc/igc_main.c struct igc_ring *tx_ring = q_vector->tx.ring; q_vector 1798 drivers/net/ethernet/intel/igc/igc_main.c q_vector->tx.total_bytes += total_bytes; q_vector 1799 drivers/net/ethernet/intel/igc/igc_main.c q_vector->tx.total_packets += total_packets; q_vector 1880 drivers/net/ethernet/intel/igc/igc_main.c napi_enable(&adapter->q_vector[i]->napi); q_vector 1885 drivers/net/ethernet/intel/igc/igc_main.c igc_assign_vector(adapter->q_vector[0], 0); q_vector 2137 drivers/net/ethernet/intel/igc/igc_main.c if (adapter->q_vector[i]) { q_vector 2138 drivers/net/ethernet/intel/igc/igc_main.c napi_synchronize(&adapter->q_vector[i]->napi); q_vector 2139 drivers/net/ethernet/intel/igc/igc_main.c napi_disable(&adapter->q_vector[i]->napi); q_vector 2591 drivers/net/ethernet/intel/igc/igc_main.c static void igc_assign_vector(struct igc_q_vector *q_vector, int msix_vector) q_vector 2593 drivers/net/ethernet/intel/igc/igc_main.c struct igc_adapter *adapter = q_vector->adapter; q_vector 2598 drivers/net/ethernet/intel/igc/igc_main.c if (q_vector->rx.ring) q_vector 2599 drivers/net/ethernet/intel/igc/igc_main.c rx_queue = q_vector->rx.ring->reg_idx; q_vector 2600 drivers/net/ethernet/intel/igc/igc_main.c if (q_vector->tx.ring) q_vector 2601 drivers/net/ethernet/intel/igc/igc_main.c tx_queue = q_vector->tx.ring->reg_idx; q_vector 2613 drivers/net/ethernet/intel/igc/igc_main.c q_vector->eims_value = BIT(msix_vector); q_vector 2621 drivers/net/ethernet/intel/igc/igc_main.c adapter->eims_enable_mask |= q_vector->eims_value; q_vector 2624 drivers/net/ethernet/intel/igc/igc_main.c q_vector->set_itr = 1; q_vector 2666 drivers/net/ethernet/intel/igc/igc_main.c igc_assign_vector(adapter->q_vector[i], vector++); q_vector 2673 drivers/net/ethernet/intel/igc/igc_main.c struct igc_q_vector *q_vector = data; q_vector 2676 drivers/net/ethernet/intel/igc/igc_main.c igc_write_itr(q_vector); q_vector 2678 drivers/net/ethernet/intel/igc/igc_main.c napi_schedule(&q_vector->napi); q_vector 2701 drivers/net/ethernet/intel/igc/igc_main.c struct igc_q_vector *q_vector = adapter->q_vector[i]; q_vector 2705 drivers/net/ethernet/intel/igc/igc_main.c q_vector->itr_register = adapter->io_addr + IGC_EITR(vector); q_vector 2707 drivers/net/ethernet/intel/igc/igc_main.c if (q_vector->rx.ring && q_vector->tx.ring) q_vector 2708 drivers/net/ethernet/intel/igc/igc_main.c sprintf(q_vector->name, "%s-TxRx-%u", netdev->name, q_vector 2709 drivers/net/ethernet/intel/igc/igc_main.c q_vector->rx.ring->queue_index); q_vector 2710 drivers/net/ethernet/intel/igc/igc_main.c else if (q_vector->tx.ring) q_vector 2711 drivers/net/ethernet/intel/igc/igc_main.c sprintf(q_vector->name, "%s-tx-%u", netdev->name, q_vector 2712 drivers/net/ethernet/intel/igc/igc_main.c q_vector->tx.ring->queue_index); q_vector 2713 drivers/net/ethernet/intel/igc/igc_main.c else if (q_vector->rx.ring) q_vector 2714 drivers/net/ethernet/intel/igc/igc_main.c sprintf(q_vector->name, "%s-rx-%u", netdev->name, q_vector 2715 drivers/net/ethernet/intel/igc/igc_main.c q_vector->rx.ring->queue_index); q_vector 2717 drivers/net/ethernet/intel/igc/igc_main.c sprintf(q_vector->name, "%s-unused", netdev->name); q_vector 2720 drivers/net/ethernet/intel/igc/igc_main.c igc_msix_ring, 0, q_vector->name, q_vector 2721 drivers/net/ethernet/intel/igc/igc_main.c q_vector); q_vector 2736 drivers/net/ethernet/intel/igc/igc_main.c adapter->q_vector[i]); q_vector 2752 drivers/net/ethernet/intel/igc/igc_main.c struct igc_q_vector *q_vector = adapter->q_vector[v_idx]; q_vector 2757 drivers/net/ethernet/intel/igc/igc_main.c if (!q_vector) q_vector 2760 drivers/net/ethernet/intel/igc/igc_main.c if (q_vector->tx.ring) q_vector 2761 drivers/net/ethernet/intel/igc/igc_main.c adapter->tx_ring[q_vector->tx.ring->queue_index] = NULL; q_vector 2763 drivers/net/ethernet/intel/igc/igc_main.c if (q_vector->rx.ring) q_vector 2764 drivers/net/ethernet/intel/igc/igc_main.c adapter->rx_ring[q_vector->rx.ring->queue_index] = NULL; q_vector 2766 drivers/net/ethernet/intel/igc/igc_main.c netif_napi_del(&q_vector->napi); q_vector 2829 drivers/net/ethernet/intel/igc/igc_main.c struct igc_q_vector *q_vector = adapter->q_vector[v_idx]; q_vector 2831 drivers/net/ethernet/intel/igc/igc_main.c adapter->q_vector[v_idx] = NULL; q_vector 2836 drivers/net/ethernet/intel/igc/igc_main.c if (q_vector) q_vector 2837 drivers/net/ethernet/intel/igc/igc_main.c kfree_rcu(q_vector, rcu); q_vector 3058 drivers/net/ethernet/intel/igc/igc_main.c eics |= adapter->q_vector[i]->eims_value; q_vector 3089 drivers/net/ethernet/intel/igc/igc_main.c static void igc_update_ring_itr(struct igc_q_vector *q_vector) q_vector 3091 drivers/net/ethernet/intel/igc/igc_main.c struct igc_adapter *adapter = q_vector->adapter; q_vector 3092 drivers/net/ethernet/intel/igc/igc_main.c int new_val = q_vector->itr_val; q_vector 3108 drivers/net/ethernet/intel/igc/igc_main.c packets = q_vector->rx.total_packets; q_vector 3110 drivers/net/ethernet/intel/igc/igc_main.c avg_wire_size = q_vector->rx.total_bytes / packets; q_vector 3112 drivers/net/ethernet/intel/igc/igc_main.c packets = q_vector->tx.total_packets; q_vector 3115 drivers/net/ethernet/intel/igc/igc_main.c q_vector->tx.total_bytes / packets); q_vector 3135 drivers/net/ethernet/intel/igc/igc_main.c ((q_vector->rx.ring && adapter->rx_itr_setting == 3) || q_vector 3136 drivers/net/ethernet/intel/igc/igc_main.c (!q_vector->rx.ring && adapter->tx_itr_setting == 3))) q_vector 3140 drivers/net/ethernet/intel/igc/igc_main.c if (new_val != q_vector->itr_val) { q_vector 3141 drivers/net/ethernet/intel/igc/igc_main.c q_vector->itr_val = new_val; q_vector 3142 drivers/net/ethernet/intel/igc/igc_main.c q_vector->set_itr = 1; q_vector 3145 drivers/net/ethernet/intel/igc/igc_main.c q_vector->rx.total_bytes = 0; q_vector 3146 drivers/net/ethernet/intel/igc/igc_main.c q_vector->rx.total_packets = 0; q_vector 3147 drivers/net/ethernet/intel/igc/igc_main.c q_vector->tx.total_bytes = 0; q_vector 3148 drivers/net/ethernet/intel/igc/igc_main.c q_vector->tx.total_packets = 0; q_vector 3166 drivers/net/ethernet/intel/igc/igc_main.c static void igc_update_itr(struct igc_q_vector *q_vector, q_vector 3226 drivers/net/ethernet/intel/igc/igc_main.c struct igc_q_vector *q_vector = adapter->q_vector[0]; q_vector 3231 drivers/net/ethernet/intel/igc/igc_main.c igc_write_itr(q_vector); q_vector 3247 drivers/net/ethernet/intel/igc/igc_main.c napi_schedule(&q_vector->napi); q_vector 3260 drivers/net/ethernet/intel/igc/igc_main.c struct igc_q_vector *q_vector = adapter->q_vector[0]; q_vector 3273 drivers/net/ethernet/intel/igc/igc_main.c igc_write_itr(q_vector); q_vector 3290 drivers/net/ethernet/intel/igc/igc_main.c napi_schedule(&q_vector->napi); q_vector 3295 drivers/net/ethernet/intel/igc/igc_main.c static void igc_set_itr(struct igc_q_vector *q_vector) q_vector 3297 drivers/net/ethernet/intel/igc/igc_main.c struct igc_adapter *adapter = q_vector->adapter; q_vector 3298 drivers/net/ethernet/intel/igc/igc_main.c u32 new_itr = q_vector->itr_val; q_vector 3312 drivers/net/ethernet/intel/igc/igc_main.c igc_update_itr(q_vector, &q_vector->tx); q_vector 3313 drivers/net/ethernet/intel/igc/igc_main.c igc_update_itr(q_vector, &q_vector->rx); q_vector 3315 drivers/net/ethernet/intel/igc/igc_main.c current_itr = max(q_vector->rx.itr, q_vector->tx.itr); q_vector 3319 drivers/net/ethernet/intel/igc/igc_main.c ((q_vector->rx.ring && adapter->rx_itr_setting == 3) || q_vector 3320 drivers/net/ethernet/intel/igc/igc_main.c (!q_vector->rx.ring && adapter->tx_itr_setting == 3))) q_vector 3339 drivers/net/ethernet/intel/igc/igc_main.c if (new_itr != q_vector->itr_val) { q_vector 3344 drivers/net/ethernet/intel/igc/igc_main.c new_itr = new_itr > q_vector->itr_val ? q_vector 3345 drivers/net/ethernet/intel/igc/igc_main.c max((new_itr * q_vector->itr_val) / q_vector 3346 drivers/net/ethernet/intel/igc/igc_main.c (new_itr + (q_vector->itr_val >> 2)), q_vector 3354 drivers/net/ethernet/intel/igc/igc_main.c q_vector->itr_val = new_itr; q_vector 3355 drivers/net/ethernet/intel/igc/igc_main.c q_vector->set_itr = 1; q_vector 3359 drivers/net/ethernet/intel/igc/igc_main.c static void igc_ring_irq_enable(struct igc_q_vector *q_vector) q_vector 3361 drivers/net/ethernet/intel/igc/igc_main.c struct igc_adapter *adapter = q_vector->adapter; q_vector 3364 drivers/net/ethernet/intel/igc/igc_main.c if ((q_vector->rx.ring && (adapter->rx_itr_setting & 3)) || q_vector 3365 drivers/net/ethernet/intel/igc/igc_main.c (!q_vector->rx.ring && (adapter->tx_itr_setting & 3))) { q_vector 3367 drivers/net/ethernet/intel/igc/igc_main.c igc_set_itr(q_vector); q_vector 3369 drivers/net/ethernet/intel/igc/igc_main.c igc_update_ring_itr(q_vector); q_vector 3374 drivers/net/ethernet/intel/igc/igc_main.c wr32(IGC_EIMS, q_vector->eims_value); q_vector 3387 drivers/net/ethernet/intel/igc/igc_main.c struct igc_q_vector *q_vector = container_of(napi, q_vector 3393 drivers/net/ethernet/intel/igc/igc_main.c if (q_vector->tx.ring) q_vector 3394 drivers/net/ethernet/intel/igc/igc_main.c clean_complete = igc_clean_tx_irq(q_vector, budget); q_vector 3396 drivers/net/ethernet/intel/igc/igc_main.c if (q_vector->rx.ring) { q_vector 3397 drivers/net/ethernet/intel/igc/igc_main.c int cleaned = igc_clean_rx_irq(q_vector, budget); q_vector 3412 drivers/net/ethernet/intel/igc/igc_main.c igc_ring_irq_enable(q_vector); q_vector 3510 drivers/net/ethernet/intel/igc/igc_main.c struct igc_q_vector *q_vector; q_vector 3521 drivers/net/ethernet/intel/igc/igc_main.c q_vector = adapter->q_vector[v_idx]; q_vector 3522 drivers/net/ethernet/intel/igc/igc_main.c if (!q_vector) q_vector 3523 drivers/net/ethernet/intel/igc/igc_main.c q_vector = kzalloc(struct_size(q_vector, ring, ring_count), q_vector 3526 drivers/net/ethernet/intel/igc/igc_main.c memset(q_vector, 0, struct_size(q_vector, ring, ring_count)); q_vector 3527 drivers/net/ethernet/intel/igc/igc_main.c if (!q_vector) q_vector 3531 drivers/net/ethernet/intel/igc/igc_main.c netif_napi_add(adapter->netdev, &q_vector->napi, q_vector 3535 drivers/net/ethernet/intel/igc/igc_main.c adapter->q_vector[v_idx] = q_vector; q_vector 3536 drivers/net/ethernet/intel/igc/igc_main.c q_vector->adapter = adapter; q_vector 3539 drivers/net/ethernet/intel/igc/igc_main.c q_vector->tx.work_limit = adapter->tx_work_limit; q_vector 3542 drivers/net/ethernet/intel/igc/igc_main.c q_vector->itr_register = adapter->io_addr + IGC_EITR(0); q_vector 3543 drivers/net/ethernet/intel/igc/igc_main.c q_vector->itr_val = IGC_START_ITR; q_vector 3546 drivers/net/ethernet/intel/igc/igc_main.c ring = q_vector->ring; q_vector 3552 drivers/net/ethernet/intel/igc/igc_main.c q_vector->itr_val = adapter->rx_itr_setting; q_vector 3556 drivers/net/ethernet/intel/igc/igc_main.c q_vector->itr_val = adapter->tx_itr_setting; q_vector 3565 drivers/net/ethernet/intel/igc/igc_main.c ring->q_vector = q_vector; q_vector 3568 drivers/net/ethernet/intel/igc/igc_main.c igc_add_ring(ring, &q_vector->tx); q_vector 3587 drivers/net/ethernet/intel/igc/igc_main.c ring->q_vector = q_vector; q_vector 3590 drivers/net/ethernet/intel/igc/igc_main.c igc_add_ring(ring, &q_vector->rx); q_vector 3722 drivers/net/ethernet/intel/igc/igc_main.c adapter->q_vector[i]); q_vector 3814 drivers/net/ethernet/intel/igc/igc_main.c igc_assign_vector(adapter->q_vector[0], 0); q_vector 3838 drivers/net/ethernet/intel/igc/igc_main.c static void igc_write_itr(struct igc_q_vector *q_vector) q_vector 3840 drivers/net/ethernet/intel/igc/igc_main.c u32 itr_val = q_vector->itr_val & IGC_QVECTOR_MASK; q_vector 3842 drivers/net/ethernet/intel/igc/igc_main.c if (!q_vector->set_itr) q_vector 3850 drivers/net/ethernet/intel/igc/igc_main.c writel(itr_val, q_vector->itr_register); q_vector 3851 drivers/net/ethernet/intel/igc/igc_main.c q_vector->set_itr = 0; q_vector 3912 drivers/net/ethernet/intel/igc/igc_main.c napi_enable(&adapter->q_vector[i]->napi); q_vector 310 drivers/net/ethernet/intel/ixgbe/ixgbe.h struct ixgbe_q_vector *q_vector; /* backpointer to host q_vector */ q_vector 658 drivers/net/ethernet/intel/ixgbe/ixgbe.h struct ixgbe_q_vector *q_vector[MAX_Q_VECTORS]; q_vector 975 drivers/net/ethernet/intel/ixgbe/ixgbe.h ixgbe_ptp_rx_pktstamp(rx_ring->q_vector, skb); q_vector 982 drivers/net/ethernet/intel/ixgbe/ixgbe.h ixgbe_ptp_rx_rgtstamp(rx_ring->q_vector, skb); q_vector 2285 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count) q_vector 2331 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c struct ixgbe_q_vector *q_vector; q_vector 2336 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count) { q_vector 2370 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count) q_vector 2389 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c q_vector = adapter->q_vector[i]; q_vector 2390 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c if (q_vector->tx.count && !q_vector->rx.count) q_vector 2392 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c q_vector->itr = tx_itr_param; q_vector 2395 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c q_vector->itr = rx_itr_param; q_vector 2396 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c ixgbe_write_eitr(q_vector); q_vector 835 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c struct ixgbe_q_vector *q_vector; q_vector 856 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector = kzalloc_node(struct_size(q_vector, ring, ring_count), q_vector 858 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c if (!q_vector) q_vector 859 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector = kzalloc(struct_size(q_vector, ring, ring_count), q_vector 861 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c if (!q_vector) q_vector 866 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c cpumask_set_cpu(cpu, &q_vector->affinity_mask); q_vector 867 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->numa_node = node; q_vector 871 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->cpu = -1; q_vector 875 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c netif_napi_add(adapter->netdev, &q_vector->napi, q_vector 879 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c adapter->q_vector[v_idx] = q_vector; q_vector 880 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->adapter = adapter; q_vector 881 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->v_idx = v_idx; q_vector 884 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->tx.work_limit = adapter->tx_work_limit; q_vector 887 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->tx.itr = IXGBE_ITR_ADAPTIVE_MAX_USECS | q_vector 889 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->rx.itr = IXGBE_ITR_ADAPTIVE_MAX_USECS | q_vector 896 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->itr = IXGBE_12K_ITR; q_vector 898 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->itr = adapter->tx_itr_setting; q_vector 902 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->itr = IXGBE_20K_ITR; q_vector 904 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->itr = adapter->rx_itr_setting; q_vector 908 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c ring = q_vector->ring; q_vector 916 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c ring->q_vector = q_vector; q_vector 919 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c ixgbe_add_ring(ring, &q_vector->tx); q_vector 942 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c ring->q_vector = q_vector; q_vector 945 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c ixgbe_add_ring(ring, &q_vector->tx); q_vector 969 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c ring->q_vector = q_vector; q_vector 972 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c ixgbe_add_ring(ring, &q_vector->rx); q_vector 1020 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c struct ixgbe_q_vector *q_vector = adapter->q_vector[v_idx]; q_vector 1023 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c ixgbe_for_each_ring(ring, q_vector->tx) { q_vector 1030 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c ixgbe_for_each_ring(ring, q_vector->rx) q_vector 1033 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c adapter->q_vector[v_idx] = NULL; q_vector 1034 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c napi_hash_del(&q_vector->napi); q_vector 1035 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c netif_napi_del(&q_vector->napi); q_vector 1041 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c kfree_rcu(q_vector, rcu); q_vector 1112 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c static bool ixgbe_clean_tx_irq(struct ixgbe_q_vector *q_vector, q_vector 1115 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_adapter *adapter = q_vector->adapter; q_vector 1119 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c unsigned int budget = q_vector->tx.work_limit; q_vector 1211 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c q_vector->tx.total_bytes += total_bytes; q_vector 1212 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c q_vector->tx.total_packets += total_packets; q_vector 1343 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c static void ixgbe_update_dca(struct ixgbe_q_vector *q_vector) q_vector 1345 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_adapter *adapter = q_vector->adapter; q_vector 1349 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (q_vector->cpu == cpu) q_vector 1352 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_for_each_ring(ring, q_vector->tx) q_vector 1355 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_for_each_ring(ring, q_vector->rx) q_vector 1358 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c q_vector->cpu = cpu; q_vector 1376 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c adapter->q_vector[i]->cpu = -1; q_vector 1377 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_update_dca(adapter->q_vector[i]); q_vector 1686 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c u32 flags = rx_ring->q_vector->adapter->flags; q_vector 1716 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c void ixgbe_rx_skb(struct ixgbe_q_vector *q_vector, q_vector 1719 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c napi_gro_receive(&q_vector->napi, skb); q_vector 2121 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c skb = napi_alloc_skb(&rx_ring->q_vector->napi, IXGBE_RX_HDR_SIZE); q_vector 2277 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c static int ixgbe_clean_rx_irq(struct ixgbe_q_vector *q_vector, q_vector 2282 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_adapter *adapter = q_vector->adapter; q_vector 2400 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_rx_skb(q_vector, skb); q_vector 2423 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c q_vector->rx.total_packets += total_rx_packets; q_vector 2424 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c q_vector->rx.total_bytes += total_rx_bytes; q_vector 2438 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_q_vector *q_vector; q_vector 2454 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c q_vector = adapter->q_vector[v_idx]; q_vector 2456 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_for_each_ring(ring, q_vector->rx) q_vector 2459 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_for_each_ring(ring, q_vector->tx) q_vector 2462 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_write_eitr(q_vector); q_vector 2504 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c static void ixgbe_update_itr(struct ixgbe_q_vector *q_vector, q_vector 2537 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c itr = (q_vector->itr >> 2) + IXGBE_ITR_ADAPTIVE_MIN_INC; q_vector 2560 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c itr = (q_vector->itr >> 2) + IXGBE_ITR_ADAPTIVE_MIN_INC; q_vector 2570 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c itr = q_vector->itr >> 2; q_vector 2579 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c itr = q_vector->itr >> 3; q_vector 2649 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c switch (q_vector->adapter->link_speed) { q_vector 2687 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c void ixgbe_write_eitr(struct ixgbe_q_vector *q_vector) q_vector 2689 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_adapter *adapter = q_vector->adapter; q_vector 2691 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c int v_idx = q_vector->v_idx; q_vector 2692 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c u32 itr_reg = q_vector->itr & IXGBE_MAX_EITR; q_vector 2716 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c static void ixgbe_set_itr(struct ixgbe_q_vector *q_vector) q_vector 2720 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_update_itr(q_vector, &q_vector->tx); q_vector 2721 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_update_itr(q_vector, &q_vector->rx); q_vector 2724 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c new_itr = min(q_vector->rx.itr, q_vector->tx.itr); q_vector 2730 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (new_itr != q_vector->itr) { q_vector 2732 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c q_vector->itr = new_itr; q_vector 2734 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_write_eitr(q_vector); q_vector 3139 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_q_vector *q_vector = data; q_vector 3143 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (q_vector->rx.ring || q_vector->tx.ring) q_vector 3144 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c napi_schedule_irqoff(&q_vector->napi); q_vector 3158 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_q_vector *q_vector = q_vector 3160 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_adapter *adapter = q_vector->adapter; q_vector 3167 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_update_dca(q_vector); q_vector 3170 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_for_each_ring(ring, q_vector->tx) { q_vector 3172 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_clean_xdp_tx_irq(q_vector, ring, budget) : q_vector 3173 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_clean_tx_irq(q_vector, ring, budget); q_vector 3185 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (q_vector->rx.count > 1) q_vector 3186 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c per_ring_budget = max(budget/q_vector->rx.count, 1); q_vector 3190 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_for_each_ring(ring, q_vector->rx) { q_vector 3192 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_clean_rx_irq_zc(q_vector, ring, q_vector 3194 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_clean_rx_irq(q_vector, ring, q_vector 3209 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_set_itr(q_vector); q_vector 3212 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c BIT_ULL(q_vector->v_idx)); q_vector 3232 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_q_vector *q_vector = adapter->q_vector[vector]; q_vector 3235 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (q_vector->tx.ring && q_vector->rx.ring) { q_vector 3236 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c snprintf(q_vector->name, sizeof(q_vector->name), q_vector 3239 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c } else if (q_vector->rx.ring) { q_vector 3240 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c snprintf(q_vector->name, sizeof(q_vector->name), q_vector 3242 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c } else if (q_vector->tx.ring) { q_vector 3243 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c snprintf(q_vector->name, sizeof(q_vector->name), q_vector 3250 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c q_vector->name, q_vector); q_vector 3260 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c &q_vector->affinity_mask); q_vector 3279 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c adapter->q_vector[vector]); q_vector 3297 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_q_vector *q_vector = adapter->q_vector[0]; q_vector 3350 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c napi_schedule_irqoff(&q_vector->napi); q_vector 3402 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_q_vector *q_vector = adapter->q_vector[vector]; q_vector 3406 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (!q_vector->rx.ring && !q_vector->tx.ring) q_vector 3412 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c free_irq(entry->vector, q_vector); q_vector 3460 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_q_vector *q_vector = adapter->q_vector[0]; q_vector 3462 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_write_eitr(q_vector); q_vector 3513 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (!ring->q_vector || (ring->q_vector->itr < IXGBE_100K_ITR)) q_vector 3536 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_q_vector *q_vector = ring->q_vector; q_vector 3538 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (q_vector) q_vector 3540 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c &q_vector->affinity_mask, q_vector 5001 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c napi_enable(&adapter->q_vector[q_idx]->napi); q_vector 5009 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c napi_disable(&adapter->q_vector[q_idx]->napi); q_vector 6452 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (tx_ring->q_vector) q_vector 6453 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ring_node = tx_ring->q_vector->numa_node; q_vector 6546 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (rx_ring->q_vector) q_vector 6547 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ring_node = rx_ring->q_vector->numa_node; q_vector 7346 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_q_vector *qv = adapter->q_vector[i]; q_vector 8357 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_q_vector *q_vector = ring->q_vector; q_vector 8372 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (!q_vector) q_vector 8394 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_adapter *adapter = q_vector->adapter; q_vector 8497 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ixgbe_fdir_add_signature_filter_82599(&q_vector->adapter->hw, q_vector 10178 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c struct ixgbe_q_vector *qv = ring->q_vector; q_vector 10529 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c napi_disable(&rx_ring->q_vector->napi); q_vector 10559 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c napi_enable(&rx_ring->q_vector->napi); q_vector 885 drivers/net/ethernet/intel/ixgbe/ixgbe_ptp.c void ixgbe_ptp_rx_pktstamp(struct ixgbe_q_vector *q_vector, q_vector 901 drivers/net/ethernet/intel/ixgbe/ixgbe_ptp.c ixgbe_ptp_convert_to_hwtstamp(q_vector->adapter, skb_hwtstamps(skb), q_vector 914 drivers/net/ethernet/intel/ixgbe/ixgbe_ptp.c void ixgbe_ptp_rx_rgtstamp(struct ixgbe_q_vector *q_vector, q_vector 923 drivers/net/ethernet/intel/ixgbe/ixgbe_ptp.c if (!q_vector || !q_vector->adapter) q_vector 926 drivers/net/ethernet/intel/ixgbe/ixgbe_ptp.c adapter = q_vector->adapter; q_vector 23 drivers/net/ethernet/intel/ixgbe/ixgbe_txrx_common.h void ixgbe_rx_skb(struct ixgbe_q_vector *q_vector, q_vector 39 drivers/net/ethernet/intel/ixgbe/ixgbe_txrx_common.h int ixgbe_clean_rx_irq_zc(struct ixgbe_q_vector *q_vector, q_vector 43 drivers/net/ethernet/intel/ixgbe/ixgbe_txrx_common.h bool ixgbe_clean_xdp_tx_irq(struct ixgbe_q_vector *q_vector, q_vector 403 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c skb = __napi_alloc_skb(&rx_ring->q_vector->napi, q_vector 427 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c int ixgbe_clean_rx_irq_zc(struct ixgbe_q_vector *q_vector, q_vector 432 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c struct ixgbe_adapter *adapter = q_vector->adapter; q_vector 526 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c ixgbe_rx_skb(q_vector, skb); q_vector 546 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c q_vector->rx.total_packets += total_rx_packets; q_vector 547 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c q_vector->rx.total_bytes += total_rx_bytes; q_vector 640 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c bool ixgbe_clean_xdp_tx_irq(struct ixgbe_q_vector *q_vector, q_vector 686 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c q_vector->tx.total_bytes += total_bytes; q_vector 687 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c q_vector->tx.total_packets += total_packets; q_vector 695 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c return ixgbe_xmit_zc(tx_ring, q_vector->tx.work_limit); q_vector 720 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c if (!napi_if_scheduled_mark_missed(&ring->q_vector->napi)) { q_vector 721 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c u64 eics = BIT_ULL(ring->q_vector->v_idx); q_vector 803 drivers/net/ethernet/intel/ixgbevf/ethtool.c if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count) q_vector 819 drivers/net/ethernet/intel/ixgbevf/ethtool.c struct ixgbevf_q_vector *q_vector; q_vector 824 drivers/net/ethernet/intel/ixgbevf/ethtool.c if (adapter->q_vector[0]->tx.count && q_vector 825 drivers/net/ethernet/intel/ixgbevf/ethtool.c adapter->q_vector[0]->rx.count && ec->tx_coalesce_usecs) q_vector 855 drivers/net/ethernet/intel/ixgbevf/ethtool.c q_vector = adapter->q_vector[i]; q_vector 856 drivers/net/ethernet/intel/ixgbevf/ethtool.c if (q_vector->tx.count && !q_vector->rx.count) q_vector 858 drivers/net/ethernet/intel/ixgbevf/ethtool.c q_vector->itr = tx_itr_param; q_vector 861 drivers/net/ethernet/intel/ixgbevf/ethtool.c q_vector->itr = rx_itr_param; q_vector 862 drivers/net/ethernet/intel/ixgbevf/ethtool.c ixgbevf_write_eitr(q_vector); q_vector 92 drivers/net/ethernet/intel/ixgbevf/ixgbevf.h struct ixgbevf_q_vector *q_vector; /* backpointer to q_vector */ q_vector 324 drivers/net/ethernet/intel/ixgbevf/ixgbevf.h struct ixgbevf_q_vector *q_vector[MAX_MSIX_Q_VECTORS]; q_vector 460 drivers/net/ethernet/intel/ixgbevf/ixgbevf.h extern void ixgbevf_write_eitr(struct ixgbevf_q_vector *q_vector); q_vector 112 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c static void ixgbevf_set_itr(struct ixgbevf_q_vector *q_vector); q_vector 266 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c static bool ixgbevf_clean_tx_irq(struct ixgbevf_q_vector *q_vector, q_vector 269 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_adapter *adapter = q_vector->adapter; q_vector 365 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->tx.total_bytes += total_bytes; q_vector 366 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->tx.total_packets += total_packets; q_vector 431 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c static void ixgbevf_rx_skb(struct ixgbevf_q_vector *q_vector, q_vector 434 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c napi_gro_receive(&q_vector->napi, skb); q_vector 892 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c skb = napi_alloc_skb(&rx_ring->q_vector->napi, IXGBEVF_RX_HDR_SIZE); q_vector 1115 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c static int ixgbevf_clean_rx_irq(struct ixgbevf_q_vector *q_vector, q_vector 1120 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_adapter *adapter = q_vector->adapter; q_vector 1221 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_rx_skb(q_vector, skb); q_vector 1248 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->rx.total_packets += total_rx_packets; q_vector 1249 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->rx.total_bytes += total_rx_bytes; q_vector 1264 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_q_vector *q_vector = q_vector 1266 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_adapter *adapter = q_vector->adapter; q_vector 1271 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_for_each_ring(ring, q_vector->tx) { q_vector 1272 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c if (!ixgbevf_clean_tx_irq(q_vector, ring, budget)) q_vector 1282 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c if (q_vector->rx.count > 1) q_vector 1283 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c per_ring_budget = max(budget/q_vector->rx.count, 1); q_vector 1287 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_for_each_ring(ring, q_vector->rx) { q_vector 1288 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c int cleaned = ixgbevf_clean_rx_irq(q_vector, ring, q_vector 1304 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_set_itr(q_vector); q_vector 1308 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c BIT(q_vector->v_idx)); q_vector 1318 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c void ixgbevf_write_eitr(struct ixgbevf_q_vector *q_vector) q_vector 1320 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_adapter *adapter = q_vector->adapter; q_vector 1322 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c int v_idx = q_vector->v_idx; q_vector 1323 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c u32 itr_reg = q_vector->itr & IXGBE_MAX_EITR; q_vector 1342 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_q_vector *q_vector; q_vector 1354 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector = adapter->q_vector[v_idx]; q_vector 1356 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_for_each_ring(ring, q_vector->rx) q_vector 1359 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_for_each_ring(ring, q_vector->tx) q_vector 1362 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c if (q_vector->tx.ring && !q_vector->rx.ring) { q_vector 1365 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->itr = IXGBE_12K_ITR; q_vector 1367 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->itr = adapter->tx_itr_setting; q_vector 1371 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->itr = IXGBE_20K_ITR; q_vector 1373 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->itr = adapter->rx_itr_setting; q_vector 1379 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_write_eitr(q_vector); q_vector 1408 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c static void ixgbevf_update_itr(struct ixgbevf_q_vector *q_vector, q_vector 1426 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c timepassed_us = q_vector->itr >> 2; q_vector 1457 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c static void ixgbevf_set_itr(struct ixgbevf_q_vector *q_vector) q_vector 1459 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c u32 new_itr = q_vector->itr; q_vector 1462 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_update_itr(q_vector, &q_vector->tx); q_vector 1463 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_update_itr(q_vector, &q_vector->rx); q_vector 1465 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c current_itr = max(q_vector->rx.itr, q_vector->tx.itr); q_vector 1482 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c if (new_itr != q_vector->itr) { q_vector 1484 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c new_itr = (10 * new_itr * q_vector->itr) / q_vector 1485 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ((9 * new_itr) + q_vector->itr); q_vector 1488 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->itr = new_itr; q_vector 1490 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_write_eitr(q_vector); q_vector 1515 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_q_vector *q_vector = data; q_vector 1518 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c if (q_vector->rx.ring || q_vector->tx.ring) q_vector 1519 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c napi_schedule_irqoff(&q_vector->napi); q_vector 1539 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_q_vector *q_vector = adapter->q_vector[vector]; q_vector 1542 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c if (q_vector->tx.ring && q_vector->rx.ring) { q_vector 1543 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c snprintf(q_vector->name, sizeof(q_vector->name), q_vector 1546 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c } else if (q_vector->rx.ring) { q_vector 1547 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c snprintf(q_vector->name, sizeof(q_vector->name), q_vector 1549 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c } else if (q_vector->tx.ring) { q_vector 1550 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c snprintf(q_vector->name, sizeof(q_vector->name), q_vector 1557 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->name, q_vector); q_vector 1580 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c adapter->q_vector[vector]); q_vector 1628 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c if (!adapter->q_vector[i]->rx.ring && q_vector 1629 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c !adapter->q_vector[i]->tx.ring) q_vector 1633 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c adapter->q_vector[i]); q_vector 2142 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_q_vector *q_vector; q_vector 2146 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector = adapter->q_vector[q_idx]; q_vector 2147 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c napi_enable(&q_vector->napi); q_vector 2154 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_q_vector *q_vector; q_vector 2158 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector = adapter->q_vector[q_idx]; q_vector 2159 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c napi_disable(&q_vector->napi); q_vector 2699 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_q_vector *q_vector; q_vector 2705 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c size = sizeof(*q_vector) + (sizeof(*ring) * ring_count); q_vector 2708 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector = kzalloc(size, GFP_KERNEL); q_vector 2709 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c if (!q_vector) q_vector 2713 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c netif_napi_add(adapter->netdev, &q_vector->napi, ixgbevf_poll, 64); q_vector 2716 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c adapter->q_vector[v_idx] = q_vector; q_vector 2717 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->adapter = adapter; q_vector 2718 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->v_idx = v_idx; q_vector 2721 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ring = q_vector->ring; q_vector 2729 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ring->q_vector = q_vector; q_vector 2732 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_add_ring(ring, &q_vector->tx); q_vector 2757 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ring->q_vector = q_vector; q_vector 2760 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_add_ring(ring, &q_vector->tx); q_vector 2786 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ring->q_vector = q_vector; q_vector 2789 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_add_ring(ring, &q_vector->rx); q_vector 2821 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_q_vector *q_vector = adapter->q_vector[v_idx]; q_vector 2824 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_for_each_ring(ring, q_vector->tx) { q_vector 2831 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ixgbevf_for_each_ring(ring, q_vector->rx) q_vector 2834 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c adapter->q_vector[v_idx] = NULL; q_vector 2835 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c netif_napi_del(&q_vector->napi); q_vector 2840 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c kfree_rcu(q_vector, rcu); q_vector 3205 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c struct ixgbevf_q_vector *qv = adapter->q_vector[i];