Lines Matching refs:q_vector
2419 if (!ring->q_vector || !ring->netdev) in i40e_config_xps_tx_ring()
2426 &ring->q_vector->affinity_mask, in i40e_config_xps_tx_ring()
2797 struct i40e_q_vector *q_vector; in i40e_vsi_configure_msix() local
2811 q_vector = vsi->q_vectors[i]; in i40e_vsi_configure_msix()
2812 q_vector->rx.itr = ITR_TO_REG(vsi->rx_itr_setting); in i40e_vsi_configure_msix()
2813 q_vector->rx.latency_range = I40E_LOW_LATENCY; in i40e_vsi_configure_msix()
2815 q_vector->rx.itr); in i40e_vsi_configure_msix()
2816 q_vector->tx.itr = ITR_TO_REG(vsi->tx_itr_setting); in i40e_vsi_configure_msix()
2817 q_vector->tx.latency_range = I40E_LOW_LATENCY; in i40e_vsi_configure_msix()
2819 q_vector->tx.itr); in i40e_vsi_configure_msix()
2823 for (q = 0; q < q_vector->num_ringpairs; q++) { in i40e_vsi_configure_msix()
2841 if (q == (q_vector->num_ringpairs - 1)) in i40e_vsi_configure_msix()
2894 struct i40e_q_vector *q_vector = vsi->q_vectors[0]; in i40e_configure_msi_and_legacy() local
2900 q_vector->rx.itr = ITR_TO_REG(vsi->rx_itr_setting); in i40e_configure_msi_and_legacy()
2901 q_vector->rx.latency_range = I40E_LOW_LATENCY; in i40e_configure_msi_and_legacy()
2902 wr32(hw, I40E_PFINT_ITR0(I40E_RX_ITR), q_vector->rx.itr); in i40e_configure_msi_and_legacy()
2903 q_vector->tx.itr = ITR_TO_REG(vsi->tx_itr_setting); in i40e_configure_msi_and_legacy()
2904 q_vector->tx.latency_range = I40E_LOW_LATENCY; in i40e_configure_msi_and_legacy()
2905 wr32(hw, I40E_PFINT_ITR0(I40E_TX_ITR), q_vector->tx.itr); in i40e_configure_msi_and_legacy()
2998 struct i40e_q_vector *q_vector = data; in i40e_msix_clean_rings() local
3000 if (!q_vector->tx.ring && !q_vector->rx.ring) in i40e_msix_clean_rings()
3003 napi_schedule(&q_vector->napi); in i40e_msix_clean_rings()
3025 struct i40e_q_vector *q_vector = vsi->q_vectors[vector]; in i40e_vsi_request_irq_msix() local
3027 if (q_vector->tx.ring && q_vector->rx.ring) { in i40e_vsi_request_irq_msix()
3028 snprintf(q_vector->name, sizeof(q_vector->name) - 1, in i40e_vsi_request_irq_msix()
3031 } else if (q_vector->rx.ring) { in i40e_vsi_request_irq_msix()
3032 snprintf(q_vector->name, sizeof(q_vector->name) - 1, in i40e_vsi_request_irq_msix()
3034 } else if (q_vector->tx.ring) { in i40e_vsi_request_irq_msix()
3035 snprintf(q_vector->name, sizeof(q_vector->name) - 1, in i40e_vsi_request_irq_msix()
3044 q_vector->name, in i40e_vsi_request_irq_msix()
3045 q_vector); in i40e_vsi_request_irq_msix()
3054 &q_vector->affinity_mask); in i40e_vsi_request_irq_msix()
3341 tx_ring->q_vector->v_idx + vsi->base_vector); in i40e_clean_fdir_tx_irq()
3353 struct i40e_q_vector *q_vector = data; in i40e_fdir_clean_ring() local
3356 if (!q_vector->tx.ring) in i40e_fdir_clean_ring()
3359 vsi = q_vector->tx.ring->vsi; in i40e_fdir_clean_ring()
3360 i40e_clean_fdir_tx_irq(q_vector->tx.ring, vsi->work_limit); in i40e_fdir_clean_ring()
3373 struct i40e_q_vector *q_vector = vsi->q_vectors[v_idx]; in map_vector_to_qp() local
3377 tx_ring->q_vector = q_vector; in map_vector_to_qp()
3378 tx_ring->next = q_vector->tx.ring; in map_vector_to_qp()
3379 q_vector->tx.ring = tx_ring; in map_vector_to_qp()
3380 q_vector->tx.count++; in map_vector_to_qp()
3382 rx_ring->q_vector = q_vector; in map_vector_to_qp()
3383 rx_ring->next = q_vector->rx.ring; in map_vector_to_qp()
3384 q_vector->rx.ring = rx_ring; in map_vector_to_qp()
3385 q_vector->rx.count++; in map_vector_to_qp()
3413 struct i40e_q_vector *q_vector = vsi->q_vectors[v_start]; in i40e_vsi_map_rings_to_vectors() local
3417 q_vector->num_ringpairs = num_ringpairs; in i40e_vsi_map_rings_to_vectors()
3419 q_vector->rx.count = 0; in i40e_vsi_map_rings_to_vectors()
3420 q_vector->tx.count = 0; in i40e_vsi_map_rings_to_vectors()
3421 q_vector->rx.ring = NULL; in i40e_vsi_map_rings_to_vectors()
3422 q_vector->tx.ring = NULL; in i40e_vsi_map_rings_to_vectors()
3806 struct i40e_q_vector *q_vector = vsi->q_vectors[v_idx]; in i40e_free_q_vector() local
3809 if (!q_vector) in i40e_free_q_vector()
3813 i40e_for_each_ring(ring, q_vector->tx) in i40e_free_q_vector()
3814 ring->q_vector = NULL; in i40e_free_q_vector()
3816 i40e_for_each_ring(ring, q_vector->rx) in i40e_free_q_vector()
3817 ring->q_vector = NULL; in i40e_free_q_vector()
3821 netif_napi_del(&q_vector->napi); in i40e_free_q_vector()
3825 kfree_rcu(q_vector, rcu); in i40e_free_q_vector()
7245 struct i40e_q_vector *q_vector; in i40e_vsi_alloc_q_vector() local
7248 q_vector = kzalloc(sizeof(struct i40e_q_vector), GFP_KERNEL); in i40e_vsi_alloc_q_vector()
7249 if (!q_vector) in i40e_vsi_alloc_q_vector()
7252 q_vector->vsi = vsi; in i40e_vsi_alloc_q_vector()
7253 q_vector->v_idx = v_idx; in i40e_vsi_alloc_q_vector()
7254 cpumask_set_cpu(v_idx, &q_vector->affinity_mask); in i40e_vsi_alloc_q_vector()
7256 netif_napi_add(vsi->netdev, &q_vector->napi, in i40e_vsi_alloc_q_vector()
7259 q_vector->rx.latency_range = I40E_LOW_LATENCY; in i40e_vsi_alloc_q_vector()
7260 q_vector->tx.latency_range = I40E_LOW_LATENCY; in i40e_vsi_alloc_q_vector()
7263 vsi->q_vectors[v_idx] = q_vector; in i40e_vsi_alloc_q_vector()