Home
last modified time | relevance | path

Searched refs:tx_ring (Results 1 – 200 of 223) sorted by relevance

12

/linux-4.4.14/drivers/net/ethernet/intel/i40evf/
Di40e_txrx.c80 void i40evf_clean_tx_ring(struct i40e_ring *tx_ring) in i40evf_clean_tx_ring() argument
86 if (!tx_ring->tx_bi) in i40evf_clean_tx_ring()
90 for (i = 0; i < tx_ring->count; i++) in i40evf_clean_tx_ring()
91 i40e_unmap_and_free_tx_resource(tx_ring, &tx_ring->tx_bi[i]); in i40evf_clean_tx_ring()
93 bi_size = sizeof(struct i40e_tx_buffer) * tx_ring->count; in i40evf_clean_tx_ring()
94 memset(tx_ring->tx_bi, 0, bi_size); in i40evf_clean_tx_ring()
97 memset(tx_ring->desc, 0, tx_ring->size); in i40evf_clean_tx_ring()
99 tx_ring->next_to_use = 0; in i40evf_clean_tx_ring()
100 tx_ring->next_to_clean = 0; in i40evf_clean_tx_ring()
102 if (!tx_ring->netdev) in i40evf_clean_tx_ring()
[all …]
Di40e_txrx.h317 void i40evf_clean_tx_ring(struct i40e_ring *tx_ring);
319 int i40evf_setup_tx_descriptors(struct i40e_ring *tx_ring);
321 void i40evf_free_tx_resources(struct i40e_ring *tx_ring);
Di40evf_main.c373 struct i40e_ring *tx_ring = adapter->tx_rings[t_idx]; in i40evf_map_vector_to_txq() local
375 tx_ring->q_vector = q_vector; in i40evf_map_vector_to_txq()
376 tx_ring->next = q_vector->tx.ring; in i40evf_map_vector_to_txq()
377 tx_ring->vsi = &adapter->vsi; in i40evf_map_vector_to_txq()
378 q_vector->tx.ring = tx_ring; in i40evf_map_vector_to_txq()
1140 struct i40e_ring *tx_ring; in i40evf_alloc_queues() local
1143 tx_ring = kzalloc(sizeof(*tx_ring) * 2, GFP_KERNEL); in i40evf_alloc_queues()
1144 if (!tx_ring) in i40evf_alloc_queues()
1147 tx_ring->queue_index = i; in i40evf_alloc_queues()
1148 tx_ring->netdev = adapter->netdev; in i40evf_alloc_queues()
[all …]
/linux-4.4.14/drivers/net/ethernet/intel/i40e/
Di40e_txrx.c57 struct i40e_ring *tx_ring; in i40e_program_fdir_filter() local
74 tx_ring = vsi->tx_rings[0]; in i40e_program_fdir_filter()
75 dev = tx_ring->dev; in i40e_program_fdir_filter()
79 if (I40E_DESC_UNUSED(tx_ring) > 1) in i40e_program_fdir_filter()
85 if (!(I40E_DESC_UNUSED(tx_ring) > 1)) in i40e_program_fdir_filter()
94 i = tx_ring->next_to_use; in i40e_program_fdir_filter()
95 fdir_desc = I40E_TX_FDIRDESC(tx_ring, i); in i40e_program_fdir_filter()
96 first = &tx_ring->tx_bi[i]; in i40e_program_fdir_filter()
99 tx_ring->next_to_use = ((i + 1) < tx_ring->count) ? i + 1 : 0; in i40e_program_fdir_filter()
147 i = tx_ring->next_to_use; in i40e_program_fdir_filter()
[all …]
Di40e_fcoe.c1005 static void i40e_fcoe_program_ddp(struct i40e_ring *tx_ring, in i40e_fcoe_program_ddp() argument
1012 struct i40e_pf *pf = tx_ring->vsi->back; in i40e_fcoe_program_ddp()
1013 u16 i = tx_ring->next_to_use; in i40e_fcoe_program_ddp()
1035 ddp_desc = I40E_DDP_CONTEXT_DESC(tx_ring, i); in i40e_fcoe_program_ddp()
1037 if (i == tx_ring->count) in i40e_fcoe_program_ddp()
1057 queue_desc = I40E_QUEUE_CONTEXT_DESC(tx_ring, i++); in i40e_fcoe_program_ddp()
1058 if (i == tx_ring->count) in i40e_fcoe_program_ddp()
1067 filter_desc = I40E_FILTER_CONTEXT_DESC(tx_ring, i); in i40e_fcoe_program_ddp()
1069 if (i == tx_ring->count) in i40e_fcoe_program_ddp()
1092 tx_ring->next_to_use = i; in i40e_fcoe_program_ddp()
[all …]
Di40e_txrx.h322 void i40e_clean_tx_ring(struct i40e_ring *tx_ring);
324 int i40e_setup_tx_descriptors(struct i40e_ring *tx_ring);
326 void i40e_free_tx_resources(struct i40e_ring *tx_ring);
330 void i40e_tx_map(struct i40e_ring *tx_ring, struct sk_buff *skb,
333 int i40e_maybe_stop_tx(struct i40e_ring *tx_ring, int size);
334 int i40e_xmit_descriptor_count(struct sk_buff *skb, struct i40e_ring *tx_ring);
336 struct i40e_ring *tx_ring, u32 *flags);
348 static inline u32 i40e_get_head(struct i40e_ring *tx_ring) in i40e_get_head() argument
350 void *head = (struct i40e_tx_desc *)tx_ring->desc + tx_ring->count; in i40e_get_head()
Di40e_debugfs.c538 struct i40e_ring *tx_ring = ACCESS_ONCE(vsi->tx_rings[i]); in i40e_dbg_dump_vsi_seid() local
540 if (!tx_ring) in i40e_dbg_dump_vsi_seid()
545 i, tx_ring->desc); in i40e_dbg_dump_vsi_seid()
548 i, tx_ring->dev, in i40e_dbg_dump_vsi_seid()
549 tx_ring->netdev, in i40e_dbg_dump_vsi_seid()
550 tx_ring->tx_bi); in i40e_dbg_dump_vsi_seid()
553 i, tx_ring->state, in i40e_dbg_dump_vsi_seid()
554 tx_ring->queue_index, in i40e_dbg_dump_vsi_seid()
555 tx_ring->reg_idx); in i40e_dbg_dump_vsi_seid()
558 i, tx_ring->dtype); in i40e_dbg_dump_vsi_seid()
[all …]
Di40e_main.c305 struct i40e_ring *tx_ring = NULL; in i40e_tx_timeout() local
334 tx_ring = vsi->tx_rings[i]; in i40e_tx_timeout()
347 if (tx_ring) { in i40e_tx_timeout()
348 head = i40e_get_head(tx_ring); in i40e_tx_timeout()
352 I40E_PFINT_DYN_CTLN(tx_ring->q_vector->v_idx + in i40e_tx_timeout()
353 tx_ring->vsi->base_vector - 1)); in i40e_tx_timeout()
358 vsi->seid, hung_queue, tx_ring->next_to_clean, in i40e_tx_timeout()
359 head, tx_ring->next_to_use, in i40e_tx_timeout()
360 readl(tx_ring->tail), val); in i40e_tx_timeout()
434 struct i40e_ring *tx_ring, *rx_ring; in i40e_get_netdev_stats_struct() local
[all …]
Di40e_ethtool.c1343 struct i40e_ring *tx_ring, *rx_ring; in i40e_get_ethtool_stats() local
1373 tx_ring = ACCESS_ONCE(vsi->tx_rings[j]); in i40e_get_ethtool_stats()
1375 if (!tx_ring) in i40e_get_ethtool_stats()
1380 start = u64_stats_fetch_begin_irq(&tx_ring->syncp); in i40e_get_ethtool_stats()
1381 data[i] = tx_ring->stats.packets; in i40e_get_ethtool_stats()
1382 data[i + 1] = tx_ring->stats.bytes; in i40e_get_ethtool_stats()
1383 } while (u64_stats_fetch_retry_irq(&tx_ring->syncp, start)); in i40e_get_ethtool_stats()
1387 rx_ring = &tx_ring[1]; in i40e_get_ethtool_stats()
/linux-4.4.14/drivers/net/ethernet/intel/fm10k/
Dfm10k_main.c784 static int fm10k_tso(struct fm10k_ring *tx_ring, in fm10k_tso() argument
817 tx_desc = FM10K_TX_DESC(tx_ring, tx_ring->next_to_use); in fm10k_tso()
823 tx_ring->netdev->features &= ~NETIF_F_GSO_UDP_TUNNEL; in fm10k_tso()
825 netdev_err(tx_ring->netdev, in fm10k_tso()
830 static void fm10k_tx_csum(struct fm10k_ring *tx_ring, in fm10k_tx_csum() argument
850 dev_warn(tx_ring->dev, in fm10k_tx_csum()
852 tx_ring->tx_stats.csum_err++; in fm10k_tx_csum()
871 dev_warn(tx_ring->dev, in fm10k_tx_csum()
875 tx_ring->tx_stats.csum_err++; in fm10k_tx_csum()
888 dev_warn(tx_ring->dev, in fm10k_tx_csum()
[all …]
Dfm10k_netdev.c33 int fm10k_setup_tx_resources(struct fm10k_ring *tx_ring) in fm10k_setup_tx_resources() argument
35 struct device *dev = tx_ring->dev; in fm10k_setup_tx_resources()
38 size = sizeof(struct fm10k_tx_buffer) * tx_ring->count; in fm10k_setup_tx_resources()
40 tx_ring->tx_buffer = vzalloc(size); in fm10k_setup_tx_resources()
41 if (!tx_ring->tx_buffer) in fm10k_setup_tx_resources()
44 u64_stats_init(&tx_ring->syncp); in fm10k_setup_tx_resources()
47 tx_ring->size = tx_ring->count * sizeof(struct fm10k_tx_desc); in fm10k_setup_tx_resources()
48 tx_ring->size = ALIGN(tx_ring->size, 4096); in fm10k_setup_tx_resources()
50 tx_ring->desc = dma_alloc_coherent(dev, tx_ring->size, in fm10k_setup_tx_resources()
51 &tx_ring->dma, GFP_KERNEL); in fm10k_setup_tx_resources()
[all …]
Dfm10k_pci.c346 struct fm10k_ring *tx_ring = interface->tx_ring[i]; in fm10k_update_stats() local
348 restart_queue += tx_ring->tx_stats.restart_queue; in fm10k_update_stats()
349 tx_busy += tx_ring->tx_stats.tx_busy; in fm10k_update_stats()
350 tx_csum_errors += tx_ring->tx_stats.csum_err; in fm10k_update_stats()
351 bytes += tx_ring->stats.bytes; in fm10k_update_stats()
352 pkts += tx_ring->stats.packets; in fm10k_update_stats()
353 hw_csum_tx_good += tx_ring->tx_stats.csum_good; in fm10k_update_stats()
428 struct fm10k_ring *tx_ring = interface->tx_ring[i]; in fm10k_watchdog_flush_tx() local
430 if (tx_ring->next_to_use != tx_ring->next_to_clean) { in fm10k_watchdog_flush_tx()
494 set_check_for_tx_hang(interface->tx_ring[i]); in fm10k_check_hang_subtask()
[all …]
Dfm10k.h273 struct fm10k_ring *tx_ring[MAX_QUEUES] ____cacheline_aligned_in_smp; member
465 struct fm10k_ring *tx_ring);
467 bool fm10k_check_tx_hang(struct fm10k_ring *tx_ring);
Dfm10k_ethtool.c343 ring = interface->tx_ring[i]; in fm10k_get_ethtool_stats()
612 interface->tx_ring[i]->count = new_tx_count; in fm10k_set_ringparam()
638 memcpy(&temp_ring[i], interface->tx_ring[i], in fm10k_set_ringparam()
653 fm10k_free_tx_resources(interface->tx_ring[i]); in fm10k_set_ringparam()
655 memcpy(interface->tx_ring[i], &temp_ring[i], in fm10k_set_ringparam()
/linux-4.4.14/drivers/net/ethernet/intel/igbvf/
Dnetdev.c437 struct igbvf_ring *tx_ring) in igbvf_setup_tx_resources() argument
442 size = sizeof(struct igbvf_buffer) * tx_ring->count; in igbvf_setup_tx_resources()
443 tx_ring->buffer_info = vzalloc(size); in igbvf_setup_tx_resources()
444 if (!tx_ring->buffer_info) in igbvf_setup_tx_resources()
448 tx_ring->size = tx_ring->count * sizeof(union e1000_adv_tx_desc); in igbvf_setup_tx_resources()
449 tx_ring->size = ALIGN(tx_ring->size, 4096); in igbvf_setup_tx_resources()
451 tx_ring->desc = dma_alloc_coherent(&pdev->dev, tx_ring->size, in igbvf_setup_tx_resources()
452 &tx_ring->dma, GFP_KERNEL); in igbvf_setup_tx_resources()
453 if (!tx_ring->desc) in igbvf_setup_tx_resources()
456 tx_ring->adapter = adapter; in igbvf_setup_tx_resources()
[all …]
Dethtool.c205 struct igbvf_ring *tx_ring = adapter->tx_ring; in igbvf_get_ringparam() local
211 ring->tx_pending = tx_ring->count; in igbvf_get_ringparam()
233 if ((new_tx_count == adapter->tx_ring->count) && in igbvf_set_ringparam()
243 adapter->tx_ring->count = new_tx_count; in igbvf_set_ringparam()
260 if (new_tx_count != adapter->tx_ring->count) { in igbvf_set_ringparam()
261 memcpy(temp_ring, adapter->tx_ring, sizeof(struct igbvf_ring)); in igbvf_set_ringparam()
268 igbvf_free_tx_resources(adapter->tx_ring); in igbvf_set_ringparam()
270 memcpy(adapter->tx_ring, temp_ring, sizeof(struct igbvf_ring)); in igbvf_set_ringparam()
Digbvf.h203 struct igbvf_ring *tx_ring /* One per active queue */
/linux-4.4.14/drivers/net/ethernet/qlogic/qlcnic/
Dqlcnic_io.c278 struct qlcnic_host_tx_ring *tx_ring = adapter->tx_ring; in qlcnic_82xx_change_filter() local
282 producer = tx_ring->producer; in qlcnic_82xx_change_filter()
283 hwdesc = &tx_ring->desc_head[tx_ring->producer]; in qlcnic_82xx_change_filter()
299 tx_ring->producer = get_next_index(producer, tx_ring->num_desc); in qlcnic_82xx_change_filter()
374 struct qlcnic_host_tx_ring *tx_ring) in qlcnic_tx_encap_pkt() argument
378 u32 producer = tx_ring->producer; in qlcnic_tx_encap_pkt()
408 hwdesc = &tx_ring->desc_head[producer]; in qlcnic_tx_encap_pkt()
409 tx_ring->cmd_buf_arr[producer].skb = NULL; in qlcnic_tx_encap_pkt()
414 producer = get_next_index(producer, tx_ring->num_desc); in qlcnic_tx_encap_pkt()
417 tx_ring->producer = producer; in qlcnic_tx_encap_pkt()
[all …]
Dqlcnic_ctx.c414 struct qlcnic_host_tx_ring *tx_ring, in qlcnic_82xx_fw_cmd_create_tx_ctx() argument
432 tx_ring->producer = 0; in qlcnic_82xx_fw_cmd_create_tx_ctx()
433 tx_ring->sw_consumer = 0; in qlcnic_82xx_fw_cmd_create_tx_ctx()
434 *(tx_ring->hw_consumer) = 0; in qlcnic_82xx_fw_cmd_create_tx_ctx()
475 prq->cmd_cons_dma_addr = cpu_to_le64(tx_ring->hw_cons_phys_addr); in qlcnic_82xx_fw_cmd_create_tx_ctx()
479 prq_cds->host_phys_addr = cpu_to_le64(tx_ring->phys_addr); in qlcnic_82xx_fw_cmd_create_tx_ctx()
480 prq_cds->ring_size = cpu_to_le32(tx_ring->num_desc); in qlcnic_82xx_fw_cmd_create_tx_ctx()
494 tx_ring->state = le32_to_cpu(prsp->host_ctx_state); in qlcnic_82xx_fw_cmd_create_tx_ctx()
496 tx_ring->crb_cmd_producer = adapter->ahw->pci_base0 + temp; in qlcnic_82xx_fw_cmd_create_tx_ctx()
497 tx_ring->ctx_id = le16_to_cpu(prsp->context_id); in qlcnic_82xx_fw_cmd_create_tx_ctx()
[all …]
Dqlcnic_main.c124 inline void qlcnic_update_cmd_producer(struct qlcnic_host_tx_ring *tx_ring) in qlcnic_update_cmd_producer() argument
126 writel(tx_ring->producer, tx_ring->crb_cmd_producer); in qlcnic_update_cmd_producer()
1722 struct qlcnic_host_tx_ring *tx_ring; in qlcnic_request_irq() local
1792 tx_ring = &adapter->tx_ring[ring]; in qlcnic_request_irq()
1793 snprintf(tx_ring->name, sizeof(tx_ring->name), in qlcnic_request_irq()
1795 err = request_irq(tx_ring->irq, handler, flags, in qlcnic_request_irq()
1796 tx_ring->name, tx_ring); in qlcnic_request_irq()
1810 struct qlcnic_host_tx_ring *tx_ring; in qlcnic_free_irq() local
1829 tx_ring = &adapter->tx_ring[ring]; in qlcnic_free_irq()
1830 if (tx_ring->irq) in qlcnic_free_irq()
[all …]
Dqlcnic_ethtool.c526 struct qlcnic_host_tx_ring *tx_ring; in qlcnic_get_regs() local
554 tx_ring = &adapter->tx_ring[ring]; in qlcnic_get_regs()
555 regs_buff[i++] = le32_to_cpu(*(tx_ring->hw_consumer)); in qlcnic_get_regs()
556 regs_buff[i++] = tx_ring->sw_consumer; in qlcnic_get_regs()
557 regs_buff[i++] = readl(tx_ring->crb_cmd_producer); in qlcnic_get_regs()
558 regs_buff[i++] = tx_ring->producer; in qlcnic_get_regs()
559 if (tx_ring->crb_intr_mask) in qlcnic_get_regs()
560 regs_buff[i++] = readl(tx_ring->crb_intr_mask); in qlcnic_get_regs()
687 u8 rx_ring, u8 tx_ring) in qlcnic_validate_ring_count() argument
689 if (rx_ring == 0 || tx_ring == 0) in qlcnic_validate_ring_count()
[all …]
Dqlcnic.h71 #define TX_BUFF_RINGSIZE(tx_ring) \ argument
72 (sizeof(struct qlcnic_cmd_buffer) * tx_ring->num_desc)
73 #define TX_DESC_RINGSIZE(tx_ring) \ argument
74 (sizeof(struct cmd_desc_type0) * tx_ring->num_desc)
595 struct qlcnic_host_tx_ring *tx_ring; member
1110 struct qlcnic_host_tx_ring *tx_ring; member
1728 static inline u32 qlcnic_tx_avail(struct qlcnic_host_tx_ring *tx_ring) in qlcnic_tx_avail() argument
1730 if (likely(tx_ring->producer < tx_ring->sw_consumer)) in qlcnic_tx_avail()
1731 return tx_ring->sw_consumer - tx_ring->producer; in qlcnic_tx_avail()
1733 return tx_ring->sw_consumer + tx_ring->num_desc - in qlcnic_tx_avail()
[all …]
Dqlcnic_hw.c395 struct qlcnic_host_tx_ring *tx_ring; in qlcnic_send_cmd_descs() local
402 tx_ring = &adapter->tx_ring[0]; in qlcnic_send_cmd_descs()
403 __netif_tx_lock_bh(tx_ring->txq); in qlcnic_send_cmd_descs()
405 producer = tx_ring->producer; in qlcnic_send_cmd_descs()
407 if (nr_desc >= qlcnic_tx_avail(tx_ring)) { in qlcnic_send_cmd_descs()
408 netif_tx_stop_queue(tx_ring->txq); in qlcnic_send_cmd_descs()
410 if (qlcnic_tx_avail(tx_ring) > nr_desc) { in qlcnic_send_cmd_descs()
411 if (qlcnic_tx_avail(tx_ring) > TX_STOP_THRESH) in qlcnic_send_cmd_descs()
412 netif_tx_wake_queue(tx_ring->txq); in qlcnic_send_cmd_descs()
415 __netif_tx_unlock_bh(tx_ring->txq); in qlcnic_send_cmd_descs()
[all …]
Dqlcnic_init.c131 struct qlcnic_host_tx_ring *tx_ring) in qlcnic_release_tx_buffers() argument
137 spin_lock(&tx_ring->tx_clean_lock); in qlcnic_release_tx_buffers()
139 cmd_buf = tx_ring->cmd_buf_arr; in qlcnic_release_tx_buffers()
140 for (i = 0; i < tx_ring->num_desc; i++) { in qlcnic_release_tx_buffers()
163 spin_unlock(&tx_ring->tx_clean_lock); in qlcnic_release_tx_buffers()
251 sds_ring->tx_ring = &adapter->tx_ring[ring]; in qlcnic_alloc_sw_resources()
253 sds_ring->tx_ring = &adapter->tx_ring[0]; in qlcnic_alloc_sw_resources()
Dqlcnic_hw.h195 struct qlcnic_host_tx_ring *tx_ring, int);
/linux-4.4.14/drivers/net/ethernet/intel/ixgbevf/
Dixgbevf_main.c196 static void ixgbevf_unmap_and_free_tx_resource(struct ixgbevf_ring *tx_ring, in ixgbevf_unmap_and_free_tx_resource() argument
202 dma_unmap_single(tx_ring->dev, in ixgbevf_unmap_and_free_tx_resource()
207 dma_unmap_page(tx_ring->dev, in ixgbevf_unmap_and_free_tx_resource()
238 static inline bool ixgbevf_check_tx_hang(struct ixgbevf_ring *tx_ring) in ixgbevf_check_tx_hang() argument
240 u32 tx_done = ixgbevf_get_tx_completed(tx_ring); in ixgbevf_check_tx_hang()
241 u32 tx_done_old = tx_ring->tx_stats.tx_done_old; in ixgbevf_check_tx_hang()
242 u32 tx_pending = ixgbevf_get_tx_pending(tx_ring); in ixgbevf_check_tx_hang()
244 clear_check_for_tx_hang(tx_ring); in ixgbevf_check_tx_hang()
254 &tx_ring->state); in ixgbevf_check_tx_hang()
257 clear_bit(__IXGBEVF_HANG_CHECK_ARMED, &tx_ring->state); in ixgbevf_check_tx_hang()
[all …]
Dethtool.c268 struct ixgbevf_ring *tx_ring = NULL, *rx_ring = NULL; in ixgbevf_set_ringparam() local
293 adapter->tx_ring[i]->count = new_tx_count; in ixgbevf_set_ringparam()
302 tx_ring = vmalloc(adapter->num_tx_queues * sizeof(*tx_ring)); in ixgbevf_set_ringparam()
303 if (!tx_ring) { in ixgbevf_set_ringparam()
310 tx_ring[i] = *adapter->tx_ring[i]; in ixgbevf_set_ringparam()
311 tx_ring[i].count = new_tx_count; in ixgbevf_set_ringparam()
312 err = ixgbevf_setup_tx_resources(&tx_ring[i]); in ixgbevf_set_ringparam()
316 ixgbevf_free_tx_resources(&tx_ring[i]); in ixgbevf_set_ringparam()
319 vfree(tx_ring); in ixgbevf_set_ringparam()
320 tx_ring = NULL; in ixgbevf_set_ringparam()
[all …]
Dixgbevf.h394 struct ixgbevf_ring *tx_ring[MAX_TX_QUEUES]; /* One per active queue */ member
/linux-4.4.14/drivers/net/ethernet/agere/
Det131x.c357 struct tx_ring { struct
488 struct tx_ring tx_ring; member
1642 struct tx_ring *tx_ring = &adapter->tx_ring; in et131x_config_tx_dma_regs() local
1645 writel(upper_32_bits(tx_ring->tx_desc_ring_pa), &txdma->pr_base_hi); in et131x_config_tx_dma_regs()
1646 writel(lower_32_bits(tx_ring->tx_desc_ring_pa), &txdma->pr_base_lo); in et131x_config_tx_dma_regs()
1652 writel(upper_32_bits(tx_ring->tx_status_pa), &txdma->dma_wb_base_hi); in et131x_config_tx_dma_regs()
1653 writel(lower_32_bits(tx_ring->tx_status_pa), &txdma->dma_wb_base_lo); in et131x_config_tx_dma_regs()
1655 *tx_ring->tx_status = 0; in et131x_config_tx_dma_regs()
1658 tx_ring->send_idx = 0; in et131x_config_tx_dma_regs()
1755 struct tx_ring *tx_ring = &adapter->tx_ring; in et131x_init_send() local
[all …]
/linux-4.4.14/drivers/net/ethernet/intel/ixgb/
Dixgb_main.c702 struct ixgb_desc_ring *txdr = &adapter->tx_ring; in ixgb_setup_tx_resources()
739 u64 tdba = adapter->tx_ring.dma; in ixgb_configure_tx()
740 u32 tdlen = adapter->tx_ring.count * sizeof(struct ixgb_tx_desc); in ixgb_configure_tx()
916 vfree(adapter->tx_ring.buffer_info); in ixgb_free_tx_resources()
917 adapter->tx_ring.buffer_info = NULL; in ixgb_free_tx_resources()
919 dma_free_coherent(&pdev->dev, adapter->tx_ring.size, in ixgb_free_tx_resources()
920 adapter->tx_ring.desc, adapter->tx_ring.dma); in ixgb_free_tx_resources()
922 adapter->tx_ring.desc = NULL; in ixgb_free_tx_resources()
957 struct ixgb_desc_ring *tx_ring = &adapter->tx_ring; in ixgb_clean_tx_ring() local
964 for (i = 0; i < tx_ring->count; i++) { in ixgb_clean_tx_ring()
[all …]
Dixgb_param.c278 struct ixgb_desc_ring *tx_ring = &adapter->tx_ring; in ixgb_check_options() local
281 tx_ring->count = TxDescriptors[bd]; in ixgb_check_options()
282 ixgb_validate_option(&tx_ring->count, &opt); in ixgb_check_options()
284 tx_ring->count = opt.def; in ixgb_check_options()
286 tx_ring->count = ALIGN(tx_ring->count, IXGB_REQ_TX_DESCRIPTOR_MULTIPLE); in ixgb_check_options()
Dixgb_ethtool.c489 struct ixgb_desc_ring *txdr = &adapter->tx_ring; in ixgb_get_ringparam()
503 struct ixgb_desc_ring *txdr = &adapter->tx_ring; in ixgb_set_ringparam()
508 tx_old = adapter->tx_ring; in ixgb_set_ringparam()
536 tx_new = adapter->tx_ring; in ixgb_set_ringparam()
538 adapter->tx_ring = tx_old; in ixgb_set_ringparam()
542 adapter->tx_ring = tx_new; in ixgb_set_ringparam()
553 adapter->tx_ring = tx_old; in ixgb_set_ringparam()
Dixgb.h148 struct ixgb_desc_ring tx_ring ____cacheline_aligned_in_smp;
/linux-4.4.14/drivers/net/ethernet/oki-semi/pch_gbe/
Dpch_gbe_main.c631 adapter->tx_ring = devm_kzalloc(&adapter->pdev->dev, in pch_gbe_alloc_queues()
632 sizeof(*adapter->tx_ring), GFP_KERNEL); in pch_gbe_alloc_queues()
633 if (!adapter->tx_ring) in pch_gbe_alloc_queues()
858 (unsigned long long)adapter->tx_ring->dma, in pch_gbe_configure_tx()
859 adapter->tx_ring->size); in pch_gbe_configure_tx()
862 tdba = adapter->tx_ring->dma; in pch_gbe_configure_tx()
863 tdlen = adapter->tx_ring->size - 0x10; in pch_gbe_configure_tx()
977 struct pch_gbe_tx_ring *tx_ring) in pch_gbe_clean_tx_ring() argument
985 for (i = 0; i < tx_ring->count; i++) { in pch_gbe_clean_tx_ring()
986 buffer_info = &tx_ring->buffer_info[i]; in pch_gbe_clean_tx_ring()
[all …]
Dpch_gbe_ethtool.c269 struct pch_gbe_tx_ring *txdr = adapter->tx_ring; in pch_gbe_get_ringparam()
302 tx_old = adapter->tx_ring; in pch_gbe_set_ringparam()
315 adapter->tx_ring = txdr; in pch_gbe_set_ringparam()
331 err = pch_gbe_setup_tx_resources(adapter, adapter->tx_ring); in pch_gbe_set_ringparam()
338 adapter->tx_ring = tx_old; in pch_gbe_set_ringparam()
340 pch_gbe_free_tx_resources(adapter, adapter->tx_ring); in pch_gbe_set_ringparam()
344 adapter->tx_ring = txdr; in pch_gbe_set_ringparam()
351 adapter->tx_ring = txdr; in pch_gbe_set_ringparam()
361 adapter->tx_ring = tx_old; in pch_gbe_set_ringparam()
Dpch_gbe_param.c459 struct pch_gbe_tx_ring *tx_ring = adapter->tx_ring; in pch_gbe_check_options() local
460 tx_ring->count = TxDescriptors; in pch_gbe_check_options()
461 pch_gbe_validate_option(&tx_ring->count, &opt, adapter); in pch_gbe_check_options()
462 tx_ring->count = roundup(tx_ring->count, in pch_gbe_check_options()
Dpch_gbe.h638 struct pch_gbe_tx_ring *tx_ring; member
664 struct pch_gbe_tx_ring *tx_ring);
/linux-4.4.14/drivers/staging/rdma/amso1100/
Dc2.c111 static int c2_tx_ring_alloc(struct c2_ring *tx_ring, void *vaddr, in c2_tx_ring_alloc() argument
119 tx_ring->start = kmalloc(sizeof(*elem) * tx_ring->count, GFP_KERNEL); in c2_tx_ring_alloc()
120 if (!tx_ring->start) in c2_tx_ring_alloc()
123 elem = tx_ring->start; in c2_tx_ring_alloc()
126 for (i = 0; i < tx_ring->count; i++, elem++, tx_desc++, txp_desc++) { in c2_tx_ring_alloc()
141 if (i == tx_ring->count - 1) { in c2_tx_ring_alloc()
142 elem->next = tx_ring->start; in c2_tx_ring_alloc()
151 tx_ring->to_use = tx_ring->to_clean = tx_ring->start; in c2_tx_ring_alloc()
326 struct c2_ring *tx_ring = &c2_port->tx_ring; in c2_tx_clean() local
334 elem = tx_ring->start; in c2_tx_clean()
[all …]
Dc2.h363 struct c2_ring tx_ring; member
/linux-4.4.14/drivers/net/ethernet/intel/e1000e/
Dnetdev.c218 struct e1000_ring *tx_ring = adapter->tx_ring; in e1000e_dump() local
262 buffer_info = &tx_ring->buffer_info[tx_ring->next_to_clean]; in e1000e_dump()
264 0, tx_ring->next_to_use, tx_ring->next_to_clean, in e1000e_dump()
306 for (i = 0; tx_ring->desc && (i < tx_ring->count); i++) { in e1000e_dump()
308 tx_desc = E1000_TX_DESC(*tx_ring, i); in e1000e_dump()
309 buffer_info = &tx_ring->buffer_info[i]; in e1000e_dump()
311 if (i == tx_ring->next_to_use && i == tx_ring->next_to_clean) in e1000e_dump()
313 else if (i == tx_ring->next_to_use) in e1000e_dump()
315 else if (i == tx_ring->next_to_clean) in e1000e_dump()
640 static void e1000e_update_tdt_wa(struct e1000_ring *tx_ring, unsigned int i) in e1000e_update_tdt_wa() argument
[all …]
Dethtool.c694 adapter->tx_ring->count = new_tx_count; in e1000_set_ringparam()
729 memcpy(temp_tx, adapter->tx_ring, size); in e1000_set_ringparam()
745 e1000e_free_tx_resources(adapter->tx_ring); in e1000_set_ringparam()
746 memcpy(adapter->tx_ring, temp_tx, size); in e1000_set_ringparam()
1114 struct e1000_ring *tx_ring = &adapter->test_tx_ring; in e1000_free_desc_rings() local
1120 if (tx_ring->desc && tx_ring->buffer_info) { in e1000_free_desc_rings()
1121 for (i = 0; i < tx_ring->count; i++) { in e1000_free_desc_rings()
1122 buffer_info = &tx_ring->buffer_info[i]; in e1000_free_desc_rings()
1147 if (tx_ring->desc) { in e1000_free_desc_rings()
1148 dma_free_coherent(&pdev->dev, tx_ring->size, tx_ring->desc, in e1000_free_desc_rings()
[all …]
De1000.h237 struct e1000_ring *tx_ring ____cacheline_aligned_in_smp;
/linux-4.4.14/drivers/net/ethernet/atheros/atl1e/
Datl1e_main.c636 hw->tpd_thresh = adapter->tx_ring.count / 2; in atl1e_sw_init()
664 struct atl1e_tx_ring *tx_ring = &adapter->tx_ring; in atl1e_clean_tx_ring() local
669 if (tx_ring->desc == NULL || tx_ring->tx_buffer == NULL) in atl1e_clean_tx_ring()
672 ring_count = tx_ring->count; in atl1e_clean_tx_ring()
675 tx_buffer = &tx_ring->tx_buffer[index]; in atl1e_clean_tx_ring()
688 tx_buffer = &tx_ring->tx_buffer[index]; in atl1e_clean_tx_ring()
695 memset(tx_ring->desc, 0, sizeof(struct atl1e_tpd_desc) * in atl1e_clean_tx_ring()
697 memset(tx_ring->tx_buffer, 0, sizeof(struct atl1e_tx_buffer) * in atl1e_clean_tx_ring()
728 *ring_size = ((u32)(adapter->tx_ring.count * in atl1e_cal_ring_size()
754 rwlock_init(&adapter->tx_ring.tx_lock); in atl1e_init_ring_resources()
[all …]
Datl1e_param.c206 adapter->tx_ring.count = (u16) val & 0xFFFC; in atl1e_check_options()
208 adapter->tx_ring.count = (u16)opt.def; in atl1e_check_options()
Datl1e.h458 struct atl1e_tx_ring tx_ring; member
/linux-4.4.14/drivers/net/ethernet/apm/xgene/
Dxgene_enet_main.c288 static int xgene_enet_setup_tx_desc(struct xgene_enet_desc_ring *tx_ring, in xgene_enet_setup_tx_desc() argument
291 struct device *dev = ndev_to_dev(tx_ring->ndev); in xgene_enet_setup_tx_desc()
292 struct xgene_enet_pdata *pdata = netdev_priv(tx_ring->ndev); in xgene_enet_setup_tx_desc()
297 u16 tail = tx_ring->tail; in xgene_enet_setup_tx_desc()
305 raw_desc = &tx_ring->raw_desc[tail]; in xgene_enet_setup_tx_desc()
306 tail = (tail + 1) & (tx_ring->slots - 1); in xgene_enet_setup_tx_desc()
312 raw_desc->m3 = cpu_to_le64(SET_VAL(HENQNUM, tx_ring->dst_ring_num) | in xgene_enet_setup_tx_desc()
320 netdev_err(tx_ring->ndev, "DMA mapping error\n"); in xgene_enet_setup_tx_desc()
334 exp_desc = (void *)&tx_ring->raw_desc[tail]; in xgene_enet_setup_tx_desc()
335 tail = (tail + 1) & (tx_ring->slots - 1); in xgene_enet_setup_tx_desc()
[all …]
Dxgene_enet_main.h156 struct xgene_enet_desc_ring *tx_ring; member
/linux-4.4.14/drivers/net/ethernet/intel/e1000/
De1000_main.c106 struct e1000_tx_ring *tx_ring);
125 struct e1000_tx_ring *tx_ring);
139 struct e1000_tx_ring *tx_ring);
1245 kfree(adapter->tx_ring); in e1000_probe()
1283 kfree(adapter->tx_ring); in e1000_remove()
1336 adapter->tx_ring = kcalloc(adapter->num_tx_queues, in e1000_alloc_queues()
1338 if (!adapter->tx_ring) in e1000_alloc_queues()
1344 kfree(adapter->tx_ring); in e1000_alloc_queues()
1578 err = e1000_setup_tx_resources(adapter, &adapter->tx_ring[i]); in e1000_setup_all_tx_resources()
1583 &adapter->tx_ring[i]); in e1000_setup_all_tx_resources()
[all …]
De1000_param.c281 struct e1000_tx_ring *tx_ring = adapter->tx_ring; in e1000_check_options() local
298 tx_ring->count = TxDescriptors[bd]; in e1000_check_options()
299 e1000_validate_option(&tx_ring->count, &opt, adapter); in e1000_check_options()
300 tx_ring->count = ALIGN(tx_ring->count, in e1000_check_options()
303 tx_ring->count = opt.def; in e1000_check_options()
306 tx_ring[i].count = tx_ring->count; in e1000_check_options()
De1000_ethtool.c570 struct e1000_tx_ring *txdr = adapter->tx_ring; in e1000_get_ringparam()
600 tx_old = adapter->tx_ring; in e1000_set_ringparam()
614 adapter->tx_ring = txdr; in e1000_set_ringparam()
645 adapter->tx_ring = tx_old; in e1000_set_ringparam()
651 adapter->tx_ring = txdr; in e1000_set_ringparam()
663 adapter->tx_ring = tx_old; in e1000_set_ringparam()
De1000.h252 struct e1000_tx_ring *tx_ring; /* One per active queue */ member
/linux-4.4.14/drivers/net/ethernet/intel/igb/
Digb_main.c369 struct igb_ring *tx_ring; in igb_dump() local
404 tx_ring = adapter->tx_ring[n]; in igb_dump()
405 buffer_info = &tx_ring->tx_buffer_info[tx_ring->next_to_clean]; in igb_dump()
407 n, tx_ring->next_to_use, tx_ring->next_to_clean, in igb_dump()
432 tx_ring = adapter->tx_ring[n]; in igb_dump()
434 pr_info("TX QUEUE INDEX = %d\n", tx_ring->queue_index); in igb_dump()
438 for (i = 0; tx_ring->desc && (i < tx_ring->count); i++) { in igb_dump()
441 tx_desc = IGB_TX_DESC(tx_ring, i); in igb_dump()
442 buffer_info = &tx_ring->tx_buffer_info[i]; in igb_dump()
444 if (i == tx_ring->next_to_use && in igb_dump()
[all …]
Digb.h373 struct igb_ring *tx_ring[16]; member
580 static inline struct netdev_queue *txring_txq(const struct igb_ring *tx_ring) in txring_txq() argument
582 return netdev_get_tx_queue(tx_ring->netdev, tx_ring->queue_index); in txring_txq()
Digb_ethtool.c888 adapter->tx_ring[i]->count = new_tx_count; in igb_set_ringparam()
916 memcpy(&temp_ring[i], adapter->tx_ring[i], in igb_set_ringparam()
931 igb_free_tx_resources(adapter->tx_ring[i]); in igb_set_ringparam()
933 memcpy(adapter->tx_ring[i], &temp_ring[i], in igb_set_ringparam()
1536 struct igb_ring *tx_ring = &adapter->test_tx_ring; in igb_setup_desc_rings() local
1542 tx_ring->count = IGB_DEFAULT_TXD; in igb_setup_desc_rings()
1543 tx_ring->dev = &adapter->pdev->dev; in igb_setup_desc_rings()
1544 tx_ring->netdev = adapter->netdev; in igb_setup_desc_rings()
1545 tx_ring->reg_idx = adapter->vfs_allocated_count; in igb_setup_desc_rings()
1547 if (igb_setup_tx_resources(tx_ring)) { in igb_setup_desc_rings()
[all …]
/linux-4.4.14/drivers/net/ethernet/intel/ixgbe/
Dixgbe_main.c568 struct ixgbe_ring *tx_ring; in ixgbe_dump() local
610 tx_ring = adapter->tx_ring[n]; in ixgbe_dump()
611 tx_buffer = &tx_ring->tx_buffer_info[tx_ring->next_to_clean]; in ixgbe_dump()
613 n, tx_ring->next_to_use, tx_ring->next_to_clean, in ixgbe_dump()
662 tx_ring = adapter->tx_ring[n]; in ixgbe_dump()
664 pr_info("TX QUEUE INDEX = %d\n", tx_ring->queue_index); in ixgbe_dump()
671 for (i = 0; tx_ring->desc && (i < tx_ring->count); i++) { in ixgbe_dump()
672 tx_desc = IXGBE_TX_DESC(tx_ring, i); in ixgbe_dump()
673 tx_buffer = &tx_ring->tx_buffer_info[i]; in ixgbe_dump()
685 if (i == tx_ring->next_to_use && in ixgbe_dump()
[all …]
Dixgbe_lib.c74 adapter->tx_ring[i]->reg_idx = reg_idx; in ixgbe_cache_ring_dcb_sriov()
101 adapter->tx_ring[i]->reg_idx = reg_idx; in ixgbe_cache_ring_dcb_sriov()
184 adapter->tx_ring[offset + i]->reg_idx = tx_idx; in ixgbe_cache_ring_dcb()
186 adapter->tx_ring[offset + i]->dcb_tc = tc; in ixgbe_cache_ring_dcb()
247 adapter->tx_ring[i]->reg_idx = reg_idx; in ixgbe_cache_ring_sriov()
253 adapter->tx_ring[i]->reg_idx = reg_idx; in ixgbe_cache_ring_sriov()
274 adapter->tx_ring[i]->reg_idx = i; in ixgbe_cache_ring_rss()
294 adapter->tx_ring[0]->reg_idx = 0; in ixgbe_cache_ring_register()
900 adapter->tx_ring[txr_idx] = ring; in ixgbe_alloc_q_vector()
975 adapter->tx_ring[ring->queue_index] = NULL; in ixgbe_free_q_vector()
[all …]
Dixgbe_ethtool.c952 struct ixgbe_ring *tx_ring = adapter->tx_ring[0]; in ixgbe_get_ringparam() local
958 ring->tx_pending = tx_ring->count; in ixgbe_get_ringparam()
991 adapter->tx_ring[i]->count = new_tx_count; in ixgbe_set_ringparam()
1018 memcpy(&temp_ring[i], adapter->tx_ring[i], in ixgbe_set_ringparam()
1033 ixgbe_free_tx_resources(adapter->tx_ring[i]); in ixgbe_set_ringparam()
1035 memcpy(adapter->tx_ring[i], &temp_ring[i], in ixgbe_set_ringparam()
1122 ring = adapter->tx_ring[j]; in ixgbe_get_ethtool_stats()
1637 struct ixgbe_ring *tx_ring = &adapter->test_tx_ring; in ixgbe_free_desc_rings() local
1649 reg_ctl = IXGBE_READ_REG(hw, IXGBE_TXDCTL(tx_ring->reg_idx)); in ixgbe_free_desc_rings()
1651 IXGBE_WRITE_REG(hw, IXGBE_TXDCTL(tx_ring->reg_idx), reg_ctl); in ixgbe_free_desc_rings()
[all …]
Dixgbe_fcoe.c511 int ixgbe_fso(struct ixgbe_ring *tx_ring, in ixgbe_fso() argument
523 dev_err(tx_ring->dev, "Wrong gso type %d:expecting SKB_GSO_FCOE\n", in ixgbe_fso()
549 dev_warn(tx_ring->dev, "unknown sof = 0x%x\n", sof); in ixgbe_fso()
575 dev_warn(tx_ring->dev, "unknown eof = 0x%x\n", eof); in ixgbe_fso()
612 ixgbe_tx_ctxtdesc(tx_ring, vlan_macip_lens, fcoe_sof_eof, in ixgbe_fso()
Dixgbe.h670 struct ixgbe_ring *tx_ring[MAX_TX_QUEUES] ____cacheline_aligned_in_smp; member
928 int ixgbe_fso(struct ixgbe_ring *tx_ring, struct ixgbe_tx_buffer *first,
983 struct ixgbe_ring *tx_ring);
/linux-4.4.14/drivers/net/ethernet/packetengines/
Dyellowfin.c310 struct yellowfin_desc *tx_ring; member
441 np->tx_ring = ring_space; in yellowfin_init_one()
514 pci_free_consistent(pdev, TX_TOTAL_SIZE, np->tx_ring, np->tx_ring_dma); in yellowfin_init_one()
700 pr_warn(" Tx ring %p: ", yp->tx_ring); in yellowfin_tx_timeout()
704 yp->tx_ring[i].result_status); in yellowfin_tx_timeout()
762 yp->tx_ring[i].dbdma_cmd = cpu_to_le32(CMD_STOP); in yellowfin_init_ring()
763 yp->tx_ring[i].branch_addr = cpu_to_le32(yp->tx_ring_dma + in yellowfin_init_ring()
767 yp->tx_ring[--i].dbdma_cmd = cpu_to_le32(CMD_STOP | BRANCH_ALWAYS); in yellowfin_init_ring()
775 yp->tx_ring[j].dbdma_cmd = cpu_to_le32(CMD_STOP); in yellowfin_init_ring()
776 yp->tx_ring[j].branch_addr = cpu_to_le32(yp->tx_ring_dma + in yellowfin_init_ring()
[all …]
Dhamachi.c485 struct hamachi_desc *tx_ring; member
651 hmp->tx_ring = ring_space; in hamachi_init_one()
780 pci_free_consistent(pdev, TX_TOTAL_SIZE, hmp->tx_ring, in hamachi_init_one()
1002 if (hmp->tx_ring[entry].status_n_length & cpu_to_le32(DescOwn)) in hamachi_tx()
1008 leXX_to_cpu(hmp->tx_ring[entry].addr), in hamachi_tx()
1013 hmp->tx_ring[entry].status_n_length = 0; in hamachi_tx()
1015 hmp->tx_ring[TX_RING_SIZE-1].status_n_length |= in hamachi_tx()
1063 printk(KERN_DEBUG" Tx ring %p: ", hmp->tx_ring); in hamachi_tx_timeout()
1066 le32_to_cpu(hmp->tx_ring[i].status_n_length)); in hamachi_tx_timeout()
1091 hmp->tx_ring[i].status_n_length = in hamachi_tx_timeout()
[all …]
/linux-4.4.14/drivers/net/ethernet/qlogic/netxen/
Dnetxen_nic_ctx.c448 struct nx_host_tx_ring *tx_ring = adapter->tx_ring; in nx_fw_cmd_create_tx_ctx() local
490 prq_cds->host_phys_addr = cpu_to_le64(tx_ring->phys_addr); in nx_fw_cmd_create_tx_ctx()
491 prq_cds->ring_size = cpu_to_le32(tx_ring->num_desc); in nx_fw_cmd_create_tx_ctx()
503 tx_ring->crb_cmd_producer = netxen_get_ioaddr(adapter, in nx_fw_cmd_create_tx_ctx()
706 struct nx_host_tx_ring *tx_ring; in netxen_init_old_ctx() local
712 tx_ring = adapter->tx_ring; in netxen_init_old_ctx()
716 hwctx->cmd_ring_addr = cpu_to_le64(tx_ring->phys_addr); in netxen_init_old_ctx()
717 hwctx->cmd_ring_size = cpu_to_le32(tx_ring->num_desc); in netxen_init_old_ctx()
762 struct nx_host_tx_ring *tx_ring; in netxen_alloc_hw_resources() local
769 tx_ring = adapter->tx_ring; in netxen_alloc_hw_resources()
[all …]
Dnetxen_nic_init.c136 struct nx_host_tx_ring *tx_ring = adapter->tx_ring; in netxen_release_tx_buffers() local
139 cmd_buf = tx_ring->cmd_buf_arr; in netxen_release_tx_buffers()
140 for (i = 0; i < tx_ring->num_desc; i++) { in netxen_release_tx_buffers()
169 struct nx_host_tx_ring *tx_ring; in netxen_free_sw_resources() local
185 if (adapter->tx_ring == NULL) in netxen_free_sw_resources()
188 tx_ring = adapter->tx_ring; in netxen_free_sw_resources()
189 vfree(tx_ring->cmd_buf_arr); in netxen_free_sw_resources()
190 kfree(tx_ring); in netxen_free_sw_resources()
191 adapter->tx_ring = NULL; in netxen_free_sw_resources()
199 struct nx_host_tx_ring *tx_ring; in netxen_alloc_sw_resources() local
[all …]
Dnetxen_nic_main.c123 struct nx_host_tx_ring *tx_ring) in netxen_nic_update_cmd_producer() argument
125 NXWRIO(adapter, tx_ring->crb_cmd_producer, tx_ring->producer); in netxen_nic_update_cmd_producer()
135 struct nx_host_tx_ring *tx_ring) in netxen_nic_update_cmd_consumer() argument
137 NXWRIO(adapter, tx_ring->crb_cmd_consumer, tx_ring->sw_consumer); in netxen_nic_update_cmd_consumer()
1225 struct nx_host_tx_ring *tx_ring; in netxen_nic_attach() local
1261 tx_ring = adapter->tx_ring; in netxen_nic_attach()
1262 tx_ring->crb_cmd_producer = netxen_get_ioaddr(adapter, in netxen_nic_attach()
1264 tx_ring->crb_cmd_consumer = netxen_get_ioaddr(adapter, in netxen_nic_attach()
1267 tx_ring->producer = 0; in netxen_nic_attach()
1268 tx_ring->sw_consumer = 0; in netxen_nic_attach()
[all …]
Dnetxen_nic_hw.c585 struct nx_host_tx_ring *tx_ring; in netxen_send_cmd_descs() local
592 tx_ring = adapter->tx_ring; in netxen_send_cmd_descs()
593 __netif_tx_lock_bh(tx_ring->txq); in netxen_send_cmd_descs()
595 producer = tx_ring->producer; in netxen_send_cmd_descs()
596 consumer = tx_ring->sw_consumer; in netxen_send_cmd_descs()
598 if (nr_desc >= netxen_tx_avail(tx_ring)) { in netxen_send_cmd_descs()
599 netif_tx_stop_queue(tx_ring->txq); in netxen_send_cmd_descs()
601 if (netxen_tx_avail(tx_ring) > nr_desc) { in netxen_send_cmd_descs()
602 if (netxen_tx_avail(tx_ring) > TX_STOP_THRESH) in netxen_send_cmd_descs()
603 netif_tx_wake_queue(tx_ring->txq); in netxen_send_cmd_descs()
[all …]
Dnetxen_nic.h81 #define TX_BUFF_RINGSIZE(tx_ring) \ argument
82 (sizeof(struct netxen_cmd_buffer) * tx_ring->num_desc)
83 #define TX_DESC_RINGSIZE(tx_ring) \ argument
84 (sizeof(struct cmd_desc_type0) * tx_ring->num_desc)
1657 struct nx_host_tx_ring *tx_ring; member
1806 struct nx_host_tx_ring *tx_ring);
1874 static inline u32 netxen_tx_avail(struct nx_host_tx_ring *tx_ring) in netxen_tx_avail() argument
1877 return find_diff_among(tx_ring->producer, in netxen_tx_avail()
1878 tx_ring->sw_consumer, tx_ring->num_desc); in netxen_tx_avail()
Dnetxen_nic_ethtool.c330 regs_buff[i++] = le32_to_cpu(*(adapter->tx_ring->hw_consumer)); in netxen_nic_get_regs()
342 adapter->tx_ring->crb_cmd_consumer); in netxen_nic_get_regs()
345 regs_buff[i++] = NXRDIO(adapter, adapter->tx_ring->crb_cmd_producer); in netxen_nic_get_regs()
/linux-4.4.14/drivers/net/ethernet/qlogic/qlge/
Dqlge_main.c2111 struct tx_ring *tx_ring; in ql_process_mac_tx_intr() local
2115 tx_ring = &qdev->tx_ring[mac_rsp->txq_idx]; in ql_process_mac_tx_intr()
2116 tx_ring_desc = &tx_ring->q[mac_rsp->tid]; in ql_process_mac_tx_intr()
2118 tx_ring->tx_bytes += (tx_ring_desc->skb)->len; in ql_process_mac_tx_intr()
2119 tx_ring->tx_packets++; in ql_process_mac_tx_intr()
2144 atomic_inc(&tx_ring->tx_count); in ql_process_mac_tx_intr()
2213 struct tx_ring *tx_ring; in ql_clean_outbound_rx_ring() local
2241 tx_ring = &qdev->tx_ring[net_rsp->txq_idx]; in ql_clean_outbound_rx_ring()
2242 if (__netif_subqueue_stopped(qdev->ndev, tx_ring->wq_id)) { in ql_clean_outbound_rx_ring()
2243 if ((atomic_read(&tx_ring->tx_count) > (tx_ring->wq_len / 4))) in ql_clean_outbound_rx_ring()
[all …]
Dqlge_dbg.c1644 DUMP_QDEV_FIELD(qdev, "%p", tx_ring); in ql_dump_qdev()
1669 void ql_dump_tx_ring(struct tx_ring *tx_ring) in ql_dump_tx_ring() argument
1671 if (tx_ring == NULL) in ql_dump_tx_ring()
1674 tx_ring->wq_id); in ql_dump_tx_ring()
1675 pr_err("tx_ring->base = %p\n", tx_ring->wq_base); in ql_dump_tx_ring()
1677 (unsigned long long) tx_ring->wq_base_dma); in ql_dump_tx_ring()
1679 tx_ring->cnsmr_idx_sh_reg, in ql_dump_tx_ring()
1680 tx_ring->cnsmr_idx_sh_reg in ql_dump_tx_ring()
1681 ? ql_read_sh_reg(tx_ring->cnsmr_idx_sh_reg) : 0); in ql_dump_tx_ring()
1682 pr_err("tx_ring->size = %d\n", tx_ring->wq_size); in ql_dump_tx_ring()
[all …]
Dqlge.h1382 struct tx_ring { struct
2116 struct tx_ring tx_ring[MAX_TX_RINGS]; member
2292 void ql_dump_tx_ring(struct tx_ring *tx_ring);
2299 #define QL_DUMP_TX_RING(tx_ring) ql_dump_tx_ring(tx_ring) argument
2307 #define QL_DUMP_TX_RING(tx_ring) argument
/linux-4.4.14/drivers/net/vmxnet3/
Dvmxnet3_drv.c336 BUG_ON(tq->buf_info[eop_idx].sop_idx != tq->tx_ring.next2comp); in vmxnet3_unmap_pkt()
337 BUG_ON(VMXNET3_TXDESC_GET_EOP(&(tq->tx_ring.base[eop_idx].txd)) != 1); in vmxnet3_unmap_pkt()
343 VMXNET3_INC_RING_IDX_ONLY(eop_idx, tq->tx_ring.size); in vmxnet3_unmap_pkt()
345 while (tq->tx_ring.next2comp != eop_idx) { in vmxnet3_unmap_pkt()
346 vmxnet3_unmap_tx_buf(tq->buf_info + tq->tx_ring.next2comp, in vmxnet3_unmap_pkt()
354 vmxnet3_cmd_ring_adv_next2comp(&tq->tx_ring); in vmxnet3_unmap_pkt()
383 vmxnet3_cmd_ring_desc_avail(&tq->tx_ring) > in vmxnet3_tq_tx_complete()
400 while (tq->tx_ring.next2comp != tq->tx_ring.next2fill) { in vmxnet3_tq_cleanup()
403 tbi = tq->buf_info + tq->tx_ring.next2comp; in vmxnet3_tq_cleanup()
410 vmxnet3_cmd_ring_adv_next2comp(&tq->tx_ring); in vmxnet3_tq_cleanup()
[all …]
Dvmxnet3_ethtool.c389 buf[j++] = VMXNET3_GET_ADDR_LO(tq->tx_ring.basePA); in vmxnet3_get_regs()
390 buf[j++] = VMXNET3_GET_ADDR_HI(tq->tx_ring.basePA); in vmxnet3_get_regs()
391 buf[j++] = tq->tx_ring.size; in vmxnet3_get_regs()
392 buf[j++] = tq->tx_ring.next2fill; in vmxnet3_get_regs()
393 buf[j++] = tq->tx_ring.next2comp; in vmxnet3_get_regs()
394 buf[j++] = tq->tx_ring.gen; in vmxnet3_get_regs()
Dvmxnet3_int.h229 struct vmxnet3_cmd_ring tx_ring; member
/linux-4.4.14/drivers/net/ethernet/amd/
Dariadne.c88 volatile struct TDRE *tx_ring[TX_RING_SIZE]; member
100 struct TDRE tx_ring[TX_RING_SIZE]; member
130 volatile struct TDRE *t = &lancedata->tx_ring[i]; in ariadne_init_ring()
138 priv->tx_ring[i] = &lancedata->tx_ring[i]; in ariadne_init_ring()
141 i, &lancedata->tx_ring[i], lancedata->tx_buff[i]); in ariadne_init_ring()
310 int status = lowb(priv->tx_ring[entry]->TMD1); in ariadne_interrupt()
315 priv->tx_ring[entry]->TMD1 &= 0xff00; in ariadne_interrupt()
319 int err_status = priv->tx_ring[entry]->TMD3; in ariadne_interrupt()
456 lance->RDP = swloww(ARIADNE_RAM + offsetof(struct lancedata, tx_ring)); in ariadne_open()
458 lance->RDP = swhighw(ARIADNE_RAM + offsetof(struct lancedata, tx_ring)); in ariadne_open()
[all …]
Dlance.c232 u32 tx_ring; member
238 struct lance_tx_head tx_ring[TX_RING_SIZE]; member
576 lp->init_block.tx_ring = ((u32)isa_virt_to_bus(lp->tx_ring) & 0xffffff) | TX_RING_LEN_BITS; in lance_probe1()
789 (u32) isa_virt_to_bus(lp->tx_ring), in lance_open()
890 lp->tx_ring[i].base = 0; in lance_init_ring()
899 lp->init_block.tx_ring = ((u32)isa_virt_to_bus(lp->tx_ring) & 0xffffff) | TX_RING_LEN_BITS; in lance_init_ring()
939 lp->tx_ring[i].base, -lp->tx_ring[i].length, in lance_tx_timeout()
940 lp->tx_ring[i].misc); in lance_tx_timeout()
981 lp->tx_ring[entry].length = -ETH_ZLEN; in lance_start_xmit()
984 lp->tx_ring[entry].length = -skb->len; in lance_start_xmit()
[all …]
Dpcnet32.c241 __le32 tx_ring; member
263 struct pcnet32_tx_head *tx_ring; member
507 lp->tx_ring, lp->tx_ring_dma_addr); in pcnet32_realloc_tx_ring()
512 lp->tx_ring = new_tx_ring; in pcnet32_realloc_tx_ring()
922 lp->tx_ring[x].length = cpu_to_le16(-skb->len); in pcnet32_loopback_test()
923 lp->tx_ring[x].misc = 0; in pcnet32_loopback_test()
948 lp->tx_ring[x].base = cpu_to_le32(lp->tx_dma_addr[x]); in pcnet32_loopback_test()
950 lp->tx_ring[x].status = cpu_to_le16(status); in pcnet32_loopback_test()
1254 int status = (short)le16_to_cpu(lp->tx_ring[entry].status); in pcnet32_tx()
1259 lp->tx_ring[entry].base = 0; in pcnet32_tx()
[all …]
Damd8111e.c305 if((lp->tx_ring = pci_alloc_consistent(lp->pci_dev, in amd8111e_init_ring()
346 lp->tx_ring[i].buff_phy_addr = 0; in amd8111e_init_ring()
347 lp->tx_ring[i].tx_flags = 0; in amd8111e_init_ring()
348 lp->tx_ring[i].buff_count = 0; in amd8111e_init_ring()
362 sizeof(struct amd8111e_tx_dr)*NUM_TX_RING_DR,lp->tx_ring, in amd8111e_init_ring()
635 if(lp->tx_ring){ in amd8111e_free_ring()
638 lp->tx_ring, lp->tx_ring_dma_addr); in amd8111e_free_ring()
640 lp->tx_ring = NULL; in amd8111e_free_ring()
657 status = le16_to_cpu(lp->tx_ring[tx_index].tx_flags); in amd8111e_tx()
662 lp->tx_ring[tx_index].buff_phy_addr = 0; in amd8111e_tx()
[all …]
Datarilance.c151 struct ringdesc tx_ring; member
620 MEM->init.tx_ring.adr_lo = offsetof( struct lance_memory, tx_head ); in lance_probe1()
621 MEM->init.tx_ring.adr_hi = 0; in lance_probe1()
622 MEM->init.tx_ring.len = TX_RING_LEN_BITS; in lance_probe1()
Damd8111e.h747 struct amd8111e_tx_dr* tx_ring; member
/linux-4.4.14/drivers/infiniband/hw/mlx4/
Dmad.c531 tun_mad = (struct mlx4_rcv_tunnel_mad *) (tun_qp->tx_ring[tun_tx_ix].buf.addr); in mlx4_ib_send_to_slave()
532 if (tun_qp->tx_ring[tun_tx_ix].ah) in mlx4_ib_send_to_slave()
533 ib_destroy_ah(tun_qp->tx_ring[tun_tx_ix].ah); in mlx4_ib_send_to_slave()
534 tun_qp->tx_ring[tun_tx_ix].ah = ah; in mlx4_ib_send_to_slave()
536 tun_qp->tx_ring[tun_tx_ix].buf.map, in mlx4_ib_send_to_slave()
578 tun_qp->tx_ring[tun_tx_ix].buf.map, in mlx4_ib_send_to_slave()
582 list.addr = tun_qp->tx_ring[tun_tx_ix].buf.map; in mlx4_ib_send_to_slave()
1242 sqp_mad = (struct mlx4_mad_snd_buf *) (sqp->tx_ring[wire_tx_ix].buf.addr); in mlx4_ib_send_to_wire()
1243 if (sqp->tx_ring[wire_tx_ix].ah) in mlx4_ib_send_to_wire()
1244 ib_destroy_ah(sqp->tx_ring[wire_tx_ix].ah); in mlx4_ib_send_to_wire()
[all …]
/linux-4.4.14/drivers/net/ethernet/dlink/
Dsundance.c372 struct netdev_desc *tx_ring; member
562 np->tx_ring = (struct netdev_desc *)ring_space; in sundance_probe1()
704 np->tx_ring, np->tx_ring_dma); in sundance_probe1()
990 (unsigned long long)(np->tx_ring_dma + i*sizeof(*np->tx_ring)), in tx_timeout()
991 le32_to_cpu(np->tx_ring[i].next_desc), in tx_timeout()
992 le32_to_cpu(np->tx_ring[i].status), in tx_timeout()
993 (le32_to_cpu(np->tx_ring[i].status) >> 2) & 0xff, in tx_timeout()
994 le32_to_cpu(np->tx_ring[i].frag[0].addr), in tx_timeout()
995 le32_to_cpu(np->tx_ring[i].frag[0].length)); in tx_timeout()
1068 np->tx_ring[i].status = 0; in init_ring()
[all …]
Ddl2k.c239 np->tx_ring = ring_space; in rio_probe1()
310 pci_free_consistent (pdev, TX_TOTAL_SIZE, np->tx_ring, np->tx_ring_dma); in rio_probe1()
604 np->tx_ring[i].status = cpu_to_le64 (TFDDone); in alloc_list()
605 np->tx_ring[i].next_desc = cpu_to_le64 (np->tx_ring_dma + in alloc_list()
658 txdesc = &np->tx_ring[entry]; in start_xmit()
773 if (!(np->tx_ring[entry].status & cpu_to_le64(TFDDone))) in rio_free_tx()
777 desc_to_dma(&np->tx_ring[entry]), in rio_free_tx()
1765 pci_unmap_single(pdev, desc_to_dma(&np->tx_ring[i]), in rio_close()
1786 pci_free_consistent (pdev, TX_TOTAL_SIZE, np->tx_ring, in rio_remove1()
Ddl2k.h370 struct netdev_desc *tx_ring; member
/linux-4.4.14/drivers/net/ethernet/
Dfealnx.c378 struct fealnx_desc *tx_ring; member
574 np->tx_ring = ring_space; in fealnx_init_one()
674 pci_free_consistent(pdev, TX_TOTAL_SIZE, np->tx_ring, np->tx_ring_dma); in fealnx_init_one()
694 pci_free_consistent(pdev, TX_TOTAL_SIZE, np->tx_ring, in fealnx_remove_one()
1157 iowrite32(np->tx_ring_dma + ((char*)np->cur_tx - (char*)np->tx_ring), in enable_rxtx()
1216 printk(KERN_DEBUG " Tx ring %p: ", np->tx_ring); in fealnx_tx_timeout()
1218 printk(KERN_CONT " %4.4x", np->tx_ring[i].status); in fealnx_tx_timeout()
1280 np->cur_tx = &np->tx_ring[0]; in init_ring()
1281 np->cur_tx_copy = &np->tx_ring[0]; in init_ring()
1286 np->tx_ring[i].status = 0; in init_ring()
[all …]
/linux-4.4.14/drivers/net/irda/
Dau1k_ir.c154 volatile struct ring_dest *tx_ring[NUM_IR_DESC]; member
257 aup->tx_ring[i] = (volatile struct ring_dest *) in setup_hw_rings()
312 ptxd = aup->tx_ring[i]; in au1k_irda_set_speed()
422 ptxd = aup->tx_ring[aup->tx_tail]; in au1k_tx_ack()
430 ptxd = aup->tx_ring[aup->tx_tail]; in au1k_tx_ack()
680 ptxd = aup->tx_ring[aup->tx_head]; in au1k_irda_hard_xmit()
861 aup->tx_ring[i]->addr_0 = (u8)(pDB->dma_addr & 0xff); in au1k_irda_net_init()
862 aup->tx_ring[i]->addr_1 = (u8)((pDB->dma_addr >> 8) & 0xff); in au1k_irda_net_init()
863 aup->tx_ring[i]->addr_2 = (u8)((pDB->dma_addr >> 16) & 0xff); in au1k_irda_net_init()
864 aup->tx_ring[i]->addr_3 = (u8)((pDB->dma_addr >> 24) & 0xff); in au1k_irda_net_init()
[all …]
Dvlsi_ir.c357 if (netif_running(ndev) && idev->rx_ring && idev->tx_ring) { in vlsi_seq_show()
361 vlsi_proc_ring(seq, idev->tx_ring); in vlsi_seq_show()
488 idev->tx_ring = vlsi_alloc_ring(idev->pdev, hwmap, ringsize[0], in vlsi_create_hwif()
490 if (idev->tx_ring == NULL) in vlsi_create_hwif()
499 idev->rx_ring = idev->tx_ring = NULL; in vlsi_create_hwif()
509 vlsi_free_ring(idev->tx_ring); in vlsi_destroy_hwif()
510 idev->rx_ring = idev->tx_ring = NULL; in vlsi_destroy_hwif()
847 struct vlsi_ring *r = idev->tx_ring; in vlsi_hard_start_xmit()
891 if (ring_first(idev->tx_ring) == NULL) { in vlsi_hard_start_xmit()
1042 struct vlsi_ring *r = idev->tx_ring; in vlsi_tx_interrupt()
[all …]
Dvlsi_ir.h724 struct vlsi_ring *tx_ring, *rx_ring; member
/linux-4.4.14/drivers/net/ethernet/qualcomm/
Dqca_spi.h56 struct tx_ring { struct
83 struct tx_ring txr; argument
/linux-4.4.14/drivers/net/hippi/
Drrunner.c159 rrpriv->tx_ring = tmpptr; in rr_init_one()
210 if (rrpriv->tx_ring) in rr_init_one()
211 pci_free_consistent(pdev, TX_TOTAL_SIZE, rrpriv->tx_ring, in rr_init_one()
239 pci_free_consistent(pdev, TX_TOTAL_SIZE, rr->tx_ring, in rr_remove_one()
601 rrpriv->tx_ring[i].size = 0; in rr_init1()
602 set_rraddr(&rrpriv->tx_ring[i].addr, 0); in rr_init1()
1069 desc = &(rrpriv->tx_ring[txcon]); in rr_interrupt()
1082 set_rraddr(&rrpriv->tx_ring[txcon].addr, 0); in rr_interrupt()
1115 struct tx_desc *desc = &(rrpriv->tx_ring[i]); in rr_raz_tx()
1291 …i is valid - dumping data (0x%x bytes - DMA len 0x%x)\n", index, len, rrpriv->tx_ring[index].size); in rr_dump()
[all …]
Drrunner.h802 struct tx_desc *tx_ring; member
/linux-4.4.14/drivers/net/ethernet/smsc/
Dsmsc9420.c62 struct smsc9420_dma_desc *tx_ring; member
548 BUG_ON(!pd->tx_ring); in smsc9420_free_tx_ring()
563 pd->tx_ring[i].status = 0; in smsc9420_free_tx_ring()
564 pd->tx_ring[i].length = 0; in smsc9420_free_tx_ring()
565 pd->tx_ring[i].buffer1 = 0; in smsc9420_free_tx_ring()
566 pd->tx_ring[i].buffer2 = 0; in smsc9420_free_tx_ring()
951 status = pd->tx_ring[index].status; in smsc9420_complete_tx()
952 length = pd->tx_ring[index].length; in smsc9420_complete_tx()
970 pd->tx_ring[index].buffer1 = 0; in smsc9420_complete_tx()
990 BUG_ON(pd->tx_ring[index].status & TDES0_OWN_); in smsc9420_hard_start_xmit()
[all …]
Depic100.c254 struct epic_tx_desc *tx_ring; member
383 ep->tx_ring = ring_space; in epic_init_one()
502 pci_free_consistent(pdev, TX_TOTAL_SIZE, ep->tx_ring, ep->tx_ring_dma); in epic_init_one()
937 ep->tx_ring[i].txstatus = 0x0000; in epic_init_ring()
938 ep->tx_ring[i].next = ep->tx_ring_dma + in epic_init_ring()
941 ep->tx_ring[i-1].next = ep->tx_ring_dma; in epic_init_ring()
964 ep->tx_ring[entry].bufaddr = pci_map_single(ep->pci_dev, skb->data, in epic_start_xmit()
977 ep->tx_ring[entry].buflength = ctrl_word | skb->len; in epic_start_xmit()
978 ep->tx_ring[entry].txstatus = in epic_start_xmit()
1031 int txstatus = ep->tx_ring[entry].txstatus; in epic_tx()
[all …]
/linux-4.4.14/drivers/net/ethernet/samsung/sxgbe/
Dsxgbe_main.c395 struct sxgbe_tx_queue *tx_ring, int tx_rsize) in init_tx_ring() argument
398 if (!tx_ring) { in init_tx_ring()
404 tx_ring->dma_tx = dma_zalloc_coherent(dev, in init_tx_ring()
406 &tx_ring->dma_tx_phy, GFP_KERNEL); in init_tx_ring()
407 if (!tx_ring->dma_tx) in init_tx_ring()
411 tx_ring->tx_skbuff_dma = devm_kcalloc(dev, tx_rsize, in init_tx_ring()
413 if (!tx_ring->tx_skbuff_dma) in init_tx_ring()
416 tx_ring->tx_skbuff = devm_kcalloc(dev, tx_rsize, in init_tx_ring()
419 if (!tx_ring->tx_skbuff) in init_tx_ring()
423 tx_ring->queue_no = queue_no; in init_tx_ring()
[all …]
/linux-4.4.14/drivers/net/ethernet/3com/
D3c59x.c598 struct boom_tx_desc* tx_ring; member
1226 vp->tx_ring = (struct boom_tx_desc *)(vp->rx_ring + RX_RING_SIZE); in vortex_probe1()
2134 struct boom_tx_desc *prev_entry = &vp->tx_ring[(vp->cur_tx-1) % TX_RING_SIZE]; in boomerang_start_xmit()
2163 vp->tx_ring[entry].next = 0; in boomerang_start_xmit()
2166 vp->tx_ring[entry].status = cpu_to_le32(skb->len | TxIntrUploaded); in boomerang_start_xmit()
2168 vp->tx_ring[entry].status = cpu_to_le32(skb->len | TxIntrUploaded | AddTCPChksum | AddUDPChksum); in boomerang_start_xmit()
2176 vp->tx_ring[entry].frag[0].addr = cpu_to_le32(dma_addr); in boomerang_start_xmit()
2177 vp->tx_ring[entry].frag[0].length = cpu_to_le32(skb->len | LAST_FRAG); in boomerang_start_xmit()
2186 vp->tx_ring[entry].frag[0].addr = cpu_to_le32(dma_addr); in boomerang_start_xmit()
2187 vp->tx_ring[entry].frag[0].length = cpu_to_le32(skb_headlen(skb)); in boomerang_start_xmit()
[all …]
D3c515.c306 struct boom_tx_desc tx_ring[TX_RING_SIZE]; member
982 &vp->tx_ring[0]); in corkscrew_timeout()
985 &vp->tx_ring[i], in corkscrew_timeout()
986 vp->tx_ring[i].length, vp->tx_ring[i].status); in corkscrew_timeout()
1021 prev_entry = &vp->tx_ring[(vp->cur_tx - 1) % TX_RING_SIZE]; in corkscrew_start_xmit()
1029 vp->tx_ring[entry].next = 0; in corkscrew_start_xmit()
1030 vp->tx_ring[entry].addr = isa_virt_to_bus(skb->data); in corkscrew_start_xmit()
1031 vp->tx_ring[entry].length = skb->len | 0x80000000; in corkscrew_start_xmit()
1032 vp->tx_ring[entry].status = skb->len | 0x80000000; in corkscrew_start_xmit()
1041 prev_entry->next = isa_virt_to_bus(&vp->tx_ring[entry]); in corkscrew_start_xmit()
[all …]
/linux-4.4.14/drivers/net/ethernet/freescale/fs_enet/
Dfs_enet-main.c236 if (fep->tx_free == fep->tx_ring) in fs_enet_tx_napi()
399 fep->tx_free = fep->tx_ring; in fs_init_bds()
431 for (i = 0, bdp = fep->tx_bd_base; i < fep->tx_ring; i++, bdp++) { in fs_init_bds()
435 CBDW_SC(bdp, (i < fep->tx_ring - 1) ? 0 : BD_SC_WRAP); in fs_init_bds()
449 for (i = 0, bdp = fep->tx_bd_base; i < fep->tx_ring; i++, bdp++) { in fs_cleanup_bds()
968 fpi->tx_ring = 64; in fs_enet_probe()
1006 (fpi->rx_ring + fpi->tx_ring) + in fs_enet_probe()
1007 sizeof(char) * fpi->tx_ring; in fs_enet_probe()
1031 fpi->tx_ring); in fs_enet_probe()
1047 fep->tx_ring = fpi->tx_ring; in fs_enet_probe()
Dmac-fcc.c155 (fpi->tx_ring + fpi->rx_ring) * in allocate_bd()
171 (fpi->tx_ring + fpi->rx_ring) * sizeof(cbd_t), in free_bd()
555 last_tx_bd = fep->tx_bd_base + ((fpi->tx_ring - 1) * sizeof(cbd_t)); in tx_restart()
Dmac-fec.c144 (fpi->tx_ring + fpi->rx_ring) * in allocate_bd()
159 dma_free_coherent(fep->dev, (fpi->tx_ring + fpi->rx_ring) in free_bd()
Dfs_enet.h132 int rx_ring, tx_ring; member
Dmac-scc.c147 fep->ring_mem_addr = cpm_dpalloc((fpi->tx_ring + fpi->rx_ring) * in allocate_bd()
/linux-4.4.14/drivers/net/ethernet/dec/tulip/
Dwinbond-840.c297 struct w840_tx_desc *tx_ring; member
801 np->tx_ring = (struct w840_tx_desc*)&np->rx_ring[RX_RING_SIZE]; in init_rxtx_rings()
831 np->tx_ring[i].status = 0; in init_rxtx_rings()
942 printk(KERN_DEBUG " Tx ring %p: ", np->tx_ring); in tx_timeout()
944 printk(KERN_CONT " %08x", np->tx_ring[i].status); in tx_timeout()
1014 np->tx_ring[entry].buffer1 = np->tx_addr[entry]; in start_tx()
1016 np->tx_ring[entry].length = DescWholePkt | skb->len; in start_tx()
1020 np->tx_ring[entry].buffer2 = np->tx_addr[entry]+TX_BUFLIMIT; in start_tx()
1021 np->tx_ring[entry].length = DescWholePkt | (len << 11) | TX_BUFLIMIT; in start_tx()
1024 np->tx_ring[entry].length |= DescEndRing; in start_tx()
[all …]
Dtulip_core.c371 tp->tx_ring[tp->cur_tx].length = cpu_to_le32(0x08000000 | 192); in tulip_up()
372 tp->tx_ring[tp->cur_tx].buffer1 = cpu_to_le32(mapping); in tulip_up()
373 tp->tx_ring[tp->cur_tx].status = cpu_to_le32(DescOwned); in tulip_up()
599 printk(KERN_DEBUG " Tx ring %p: ", tp->tx_ring); in tulip_tx_timeout()
601 pr_cont(" %08x", (unsigned int)tp->tx_ring[i].status); in tulip_tx_timeout()
659 tp->tx_ring[i].status = 0x00000000; in tulip_init_ring()
660 tp->tx_ring[i].buffer2 = cpu_to_le32(tp->tx_ring_dma + sizeof(struct tulip_tx_desc) * (i + 1)); in tulip_init_ring()
662 tp->tx_ring[i-1].buffer2 = cpu_to_le32(tp->tx_ring_dma); in tulip_init_ring()
683 tp->tx_ring[entry].buffer1 = cpu_to_le32(mapping); in tulip_start_xmit()
698 tp->tx_ring[entry].length = cpu_to_le32(skb->len | flag); in tulip_start_xmit()
[all …]
Dde2104x.c307 struct de_desc *tx_ring; member
552 status = le32_to_cpu(de->tx_ring[tx_tail].opts1); in de_tx()
625 txd = &de->tx_ring[entry]; in de_start_xmit()
759 dummy_txd = &de->tx_ring[entry]; in __de_set_rx_mode()
775 txd = &de->tx_ring[entry]; in __de_set_rx_mode()
1310 memset(de->tx_ring, 0, sizeof(struct de_desc) * DE_TX_RING_SIZE); in de_init_rings()
1311 de->tx_ring[DE_TX_RING_SIZE - 1].opts2 = cpu_to_le32(RingEnd); in de_init_rings()
1324 de->tx_ring = &de->rx_ring[DE_RX_RING_SIZE]; in de_alloc_rings()
1335 memset(de->tx_ring, 0, sizeof(struct de_desc) * DE_TX_RING_SIZE); in de_clean_rings()
1336 de->tx_ring[DE_TX_RING_SIZE - 1].opts2 = cpu_to_le32(RingEnd); in de_clean_rings()
[all …]
Dde4x5.c791 struct de4x5_desc *tx_ring; /* TX descriptor ring */ member
1181 lp->tx_ring = lp->rx_ring + NUM_RX_DESC; in de4x5_hw_init()
1224 lp->tx_ring[lp->txRingSize - 1].des1 |= cpu_to_le32(TD_TER); in de4x5_hw_init()
1425 lp->tx_ring[i].status = cpu_to_le32(0); in de4x5_sw_reset()
1440 if ((s32)le32_to_cpu(lp->tx_ring[lp->tx_new].status) >= 0) j=1; in de4x5_sw_reset()
1679 dma_unmap_single(lp->gendev, le32_to_cpu(lp->tx_ring[entry].buf), in de4x5_free_tx_buff()
1680 le32_to_cpu(lp->tx_ring[entry].des1) & TD_TBS1, in de4x5_free_tx_buff()
1699 status = (s32)le32_to_cpu(lp->tx_ring[entry].status); in de4x5_tx()
1901 lp->tx_ring[lp->tx_new].buf = cpu_to_le32(buf_dma); in load_packet()
1902 lp->tx_ring[lp->tx_new].des1 &= cpu_to_le32(TD_TER); in load_packet()
[all …]
Dtulip.h405 struct tulip_tx_desc *tx_ring; member
Dinterrupt.c591 int status = le32_to_cpu(tp->tx_ring[entry].status); in tulip_interrupt()
/linux-4.4.14/drivers/net/ethernet/micrel/
Dks8695net.c175 struct tx_ring_desc *tx_ring; member
379 !(ksp->tx_ring[buff_n].owner & cpu_to_le32(TDES_OWN))) { in ks8695_tx_irq()
387 ksp->tx_ring[buff_n].data_ptr = 0; in ks8695_tx_irq()
687 ksp->tx_ring[buff_n].owner = 0; in ks8695_shutdown()
688 ksp->tx_ring[buff_n].status = 0; in ks8695_shutdown()
689 ksp->tx_ring[buff_n].data_ptr = 0; in ks8695_shutdown()
1193 ksp->tx_ring[buff_n].data_ptr = in ks8695_start_xmit()
1195 ksp->tx_ring[buff_n].status = in ks8695_start_xmit()
1202 ksp->tx_ring[buff_n].owner = cpu_to_le32(TDES_OWN); in ks8695_start_xmit()
1473 ksp->tx_ring = ksp->ring_base; in ks8695_probe()
[all …]
/linux-4.4.14/drivers/net/ethernet/sgi/
Dmeth.c78 tx_packet *tx_ring; member
214 priv->tx_ring = dma_zalloc_coherent(NULL, TX_RING_BUFFER_SIZE, in meth_init_tx_ring()
216 if (!priv->tx_ring) in meth_init_tx_ring()
256 dma_free_coherent(NULL, TX_RING_BUFFER_SIZE, priv->tx_ring, in meth_free_tx_ring()
492 status = priv->tx_ring[priv->tx_read].header.raw; in meth_tx_cleanup()
526 priv->tx_ring[priv->tx_read].header.raw = 0; in meth_tx_cleanup()
611 tx_packet *desc = &priv->tx_ring[priv->tx_write]; in meth_tx_short_prepare()
624 tx_packet *desc = &priv->tx_ring[priv->tx_write]; in meth_tx_1page_prepare()
649 tx_packet *desc = &priv->tx_ring[priv->tx_write]; in meth_tx_2page_prepare()
/linux-4.4.14/drivers/dma/
Dxgene-dma.c296 struct xgene_dma_ring tx_ring; member
438 xgene_dma_init_desc(desc1, chan->tx_ring.dst_ring_num); in xgene_dma_prep_cpy_desc()
484 xgene_dma_init_desc(desc1, chan->tx_ring.dst_ring_num); in xgene_dma_prep_xor_desc()
655 struct xgene_dma_ring *ring = &chan->tx_ring; in xgene_chan_xfer_request()
1347 xgene_dma_delete_ring_one(&chan->tx_ring); in xgene_dma_delete_chan_rings()
1385 struct xgene_dma_ring *tx_ring = &chan->tx_ring; in xgene_dma_create_chan_rings() local
1401 tx_ring->owner = XGENE_DMA_RING_OWNER_DMA; in xgene_dma_create_chan_rings()
1402 tx_ring->buf_num = XGENE_DMA_BUFNUM + chan->id; in xgene_dma_create_chan_rings()
1404 ret = xgene_dma_create_ring_one(chan, tx_ring, in xgene_dma_create_chan_rings()
1411 tx_ring->dst_ring_num = XGENE_DMA_RING_DST_ID(rx_ring->num); in xgene_dma_create_chan_rings()
[all …]
/linux-4.4.14/drivers/net/ethernet/amd/xgbe/
Dxgbe-drv.c136 struct xgbe_ring *tx_ring, *rx_ring; in xgbe_alloc_channels() local
146 tx_ring = kcalloc(pdata->tx_ring_count, sizeof(struct xgbe_ring), in xgbe_alloc_channels()
148 if (!tx_ring) in xgbe_alloc_channels()
177 spin_lock_init(&tx_ring->lock); in xgbe_alloc_channels()
178 channel->tx_ring = tx_ring++; in xgbe_alloc_channels()
189 channel->tx_ring, channel->rx_ring); in xgbe_alloc_channels()
201 kfree(tx_ring); in xgbe_alloc_channels()
216 kfree(pdata->channel->tx_ring); in xgbe_free_channels()
283 if (channel->tx_ring && channel->rx_ring) in xgbe_enable_rx_tx_ints()
285 else if (channel->tx_ring) in xgbe_enable_rx_tx_ints()
[all …]
Dxgbe-desc.c181 xgbe_free_ring(pdata, channel->tx_ring); in xgbe_free_ring_resources()
233 ret = xgbe_init_ring(pdata, channel->tx_ring, in xgbe_alloc_ring_resources()
368 ring = channel->tx_ring; in xgbe_wrapper_tx_descriptor_init()
494 struct xgbe_ring *ring = channel->tx_ring; in xgbe_map_tx_skb()
Dxgbe-dev.c197 if (!channel->tx_ring) in xgbe_config_tx_pbl_val()
236 if (!channel->tx_ring) in xgbe_config_osp_mode()
332 if (!channel->tx_ring) in xgbe_config_tso_mode()
622 if (channel->tx_ring) { in xgbe_enable_dma_interrupts()
1078 struct xgbe_ring *ring = channel->tx_ring; in xgbe_tx_desc_init()
1395 struct xgbe_ring *ring = channel->tx_ring; in xgbe_dev_xmit()
2552 if (!channel->tx_ring) in xgbe_enable_tx()
2575 if (!channel->tx_ring) in xgbe_disable_tx()
2591 if (!channel->tx_ring) in xgbe_disable_tx()
2657 if (!channel->tx_ring) in xgbe_powerup_tx()
[all …]
Dxgbe.h456 struct xgbe_ring *tx_ring; member
/linux-4.4.14/drivers/net/ethernet/hisilicon/hns/
Dhns_dsaf_rcb.c188 (ring_type == RX_RING) ? &q->rx_ring : &q->tx_ring; in hns_rcb_ring_init()
399 ring = &q->tx_ring; in hns_rcb_ring_get_cfg()
774 regs_buff[4] = queue->tx_ring.stats.tx_pkts; in hns_rcb_get_stats()
775 regs_buff[5] = queue->tx_ring.stats.tx_bytes; in hns_rcb_get_stats()
776 regs_buff[6] = queue->tx_ring.stats.tx_err_cnt; in hns_rcb_get_stats()
777 regs_buff[7] = queue->tx_ring.stats.io_err_cnt; in hns_rcb_get_stats()
778 regs_buff[8] = queue->tx_ring.stats.sw_err_cnt; in hns_rcb_get_stats()
779 regs_buff[9] = queue->tx_ring.stats.seg_pkt_cnt; in hns_rcb_get_stats()
780 regs_buff[10] = queue->tx_ring.stats.restart_queue; in hns_rcb_get_stats()
781 regs_buff[11] = queue->tx_ring.stats.tx_busy; in hns_rcb_get_stats()
Dhnae.c240 ret = hnae_init_ring(q, &q->tx_ring, q->tx_ring.flags | RINGF_DIR); in hnae_init_queue()
254 hnae_fini_ring(&q->tx_ring); in hnae_init_queue()
264 hnae_fini_ring(&q->tx_ring); in hnae_fini_queue()
Dhnae.h305 struct hnae_ring rx_ring, tx_ring; member
491 (q)->tx_ring.io_base + RCB_REG_TAIL)
Dhns_ae_adapt.c159 ae_handle->qs[i]->tx_ring.q = ae_handle->qs[i]; in hns_ae_get_handle()
507 tx_bytes += queue->tx_ring.stats.tx_bytes; in hns_ae_update_stats()
508 tx_packets += queue->tx_ring.stats.tx_pkts; in hns_ae_update_stats()
Dhns_enet.c1191 tx_bytes += h->qs[idx]->tx_ring.stats.tx_bytes; in hns_nic_get_stats64()
1192 tx_pkts += h->qs[idx]->tx_ring.stats.tx_pkts; in hns_nic_get_stats64()
1292 i, h->qs[i]->tx_ring.next_to_clean); in hns_nic_dump()
1294 i, h->qs[i]->tx_ring.next_to_use); in hns_nic_dump()
1406 rd->ring = &h->qs[i]->tx_ring; in hns_nic_init_ring_data()
/linux-4.4.14/drivers/net/ethernet/mellanox/mlx4/
Den_netdev.c1307 i, priv->tx_ring[i]->qpn, priv->tx_ring[i]->cqn, in mlx4_en_tx_timeout()
1308 priv->tx_ring[i]->cons, priv->tx_ring[i]->prod); in mlx4_en_tx_timeout()
1531 struct mlx4_en_tx_ring *tx_ring; in mlx4_en_start_port() local
1640 tx_ring = priv->tx_ring[i]; in mlx4_en_start_port()
1641 err = mlx4_en_activate_tx_ring(priv, tx_ring, cq->mcq.cqn, in mlx4_en_start_port()
1648 tx_ring->tx_queue = netdev_get_tx_queue(dev, i); in mlx4_en_start_port()
1654 for (j = 0; j < tx_ring->buf_size; j += STAMP_STRIDE) in mlx4_en_start_port()
1655 *((u32 *) (tx_ring->buf + j)) = 0xffffffff; in mlx4_en_start_port()
1726 mlx4_en_deactivate_tx_ring(priv, priv->tx_ring[tx_index]); in mlx4_en_start_port()
1839 mlx4_en_deactivate_tx_ring(priv, priv->tx_ring[i]); in mlx4_en_stop_port()
[all …]
Den_ethtool.c405 data[index++] = priv->tx_ring[i]->packets; in mlx4_en_get_ethtool_stats()
406 data[index++] = priv->tx_ring[i]->bytes; in mlx4_en_get_ethtool_stats()
1058 tx_size == priv->tx_ring[0]->size) in mlx4_en_set_ringparam()
1100 param->tx_pending = priv->tx_ring[0]->size; in mlx4_en_get_ringparam()
1811 bf_supported &= priv->tx_ring[i]->bf_alloced; in mlx4_en_set_priv_flags()
1824 priv->tx_ring[i]->bf_enabled = bf_enabled_new; in mlx4_en_set_priv_flags()
Den_tx.c390 struct mlx4_en_tx_ring *ring = priv->tx_ring[cq->ring]; in mlx4_en_process_tx_cq()
737 ring = priv->tx_ring[tx_ind]; in mlx4_en_xmit()
Den_port.c196 const struct mlx4_en_tx_ring *ring = priv->tx_ring[i]; in mlx4_en_DUMP_ETH_STATS()
Dmlx4_en.h555 struct mlx4_en_tx_ring **tx_ring; member
/linux-4.4.14/drivers/net/ethernet/via/
Dvia-rhine.c443 struct tx_desc *tx_ring; member
1188 rp->tx_ring = ring + RX_RING_SIZE * sizeof(struct rx_desc); in alloc_ring()
1204 rp->tx_ring = NULL; in free_ring()
1330 rp->tx_ring[i].tx_status = 0; in alloc_tbufs()
1331 rp->tx_ring[i].desc_length = cpu_to_le32(TXDESC); in alloc_tbufs()
1333 rp->tx_ring[i].next_desc = cpu_to_le32(next); in alloc_tbufs()
1337 rp->tx_ring[i-1].next_desc = cpu_to_le32(rp->tx_ring_dma); in alloc_tbufs()
1349 rp->tx_ring[i].tx_status = 0; in free_tbufs()
1350 rp->tx_ring[i].desc_length = cpu_to_le32(TXDESC); in free_tbufs()
1351 rp->tx_ring[i].addr = cpu_to_le32(0xBADF00D0); /* An invalid address. */ in free_tbufs()
[all …]
/linux-4.4.14/drivers/net/xen-netback/
Dxenbus.c56 struct xen_netif_tx_back_ring *tx_ring = &queue->tx; in xenvif_read_io_ring() local
60 if (tx_ring->sring) { in xenvif_read_io_ring()
61 struct xen_netif_tx_sring *sring = tx_ring->sring; in xenvif_read_io_ring()
64 tx_ring->nr_ents); in xenvif_read_io_ring()
68 tx_ring->req_cons, in xenvif_read_io_ring()
69 tx_ring->req_cons - sring->rsp_prod, in xenvif_read_io_ring()
74 tx_ring->rsp_prod_pvt, in xenvif_read_io_ring()
75 tx_ring->rsp_prod_pvt - sring->rsp_prod, in xenvif_read_io_ring()
/linux-4.4.14/drivers/net/wireless/realtek/rtlwifi/
Dpci.c532 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in _rtl_pci_tx_chk_waitq()
567 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[prio]; in _rtl_pci_tx_isr()
1125 ring = &rtlpci->tx_ring[BEACON_QUEUE]; in _rtl_pci_prepare_bcn_tasklet()
1264 rtlpci->tx_ring[prio].buffer_desc = buffer_desc; in _rtl_pci_init_tx_ring()
1265 rtlpci->tx_ring[prio].buffer_desc_dma = buffer_desc_dma; in _rtl_pci_init_tx_ring()
1267 rtlpci->tx_ring[prio].cur_tx_rp = 0; in _rtl_pci_init_tx_ring()
1268 rtlpci->tx_ring[prio].cur_tx_wp = 0; in _rtl_pci_init_tx_ring()
1269 rtlpci->tx_ring[prio].avl_desc = entries; in _rtl_pci_init_tx_ring()
1282 rtlpci->tx_ring[prio].desc = desc; in _rtl_pci_init_tx_ring()
1283 rtlpci->tx_ring[prio].dma = desc_dma; in _rtl_pci_init_tx_ring()
[all …]
Dpci.h210 struct rtl8192_tx_ring tx_ring[RTL_PCI_MAX_TX_QUEUE_COUNT]; member
/linux-4.4.14/drivers/net/ethernet/apple/
Dmacmace.c62 unsigned char *tx_ring; member
389 mp->tx_ring = dma_alloc_coherent(mp->device, in mace_open()
392 if (mp->tx_ring == NULL) in mace_open()
421 mp->tx_ring, mp->tx_ring_phys); in mace_open()
469 skb_copy_from_linear_data(skb, mp->tx_ring, skb->len); in mace_xmit_start()
759 mp->tx_ring, mp->tx_ring_phys); in mac_mace_device_remove()
/linux-4.4.14/drivers/net/ethernet/sis/
Dsis900.c185 BufferDesc *tx_ring; member
486 sis_priv->tx_ring = ring_space; in sis900_probe()
575 pci_free_consistent(pci_dev, TX_TOTAL_SIZE, sis_priv->tx_ring, in sis900_probe()
1136 sis_priv->tx_ring[i].link = sis_priv->tx_ring_dma + in sis900_init_tx_ring()
1138 sis_priv->tx_ring[i].cmdsts = 0; in sis900_init_tx_ring()
1139 sis_priv->tx_ring[i].bufptr = 0; in sis900_init_tx_ring()
1564 sis_priv->tx_ring[i].bufptr, skb->len, in sis900_tx_timeout()
1568 sis_priv->tx_ring[i].cmdsts = 0; in sis900_tx_timeout()
1569 sis_priv->tx_ring[i].bufptr = 0; in sis900_tx_timeout()
1614 sis_priv->tx_ring[entry].bufptr = pci_map_single(sis_priv->pci_dev, in sis900_start_xmit()
[all …]
/linux-4.4.14/net/packet/
Daf_packet.c170 int closing, int tx_ring);
1222 po->tx_ring.pending_refcnt = alloc_percpu(unsigned int); in packet_alloc_pending()
1223 if (unlikely(po->tx_ring.pending_refcnt == NULL)) in packet_alloc_pending()
1231 free_percpu(po->tx_ring.pending_refcnt); in packet_free_pending()
2319 if (likely(po->tx_ring.pg_vec)) { in tpacket_destruct_skb()
2324 packet_dec_pending(&po->tx_ring); in tpacket_destruct_skb()
2381 off_max = po->tx_ring.frame_size - tp_len; in tpacket_fill_skb()
2507 size_max = po->tx_ring.frame_size in tpacket_snd()
2514 ph = packet_current_frame(po, &po->tx_ring, in tpacket_snd()
2546 packet_increment_head(&po->tx_ring); in tpacket_snd()
[all …]
Dinternal.h109 struct packet_ring_buffer tx_ring; member
Ddiag.c106 ret = pdiag_put_ring(&po->tx_ring, po->tp_version, in pdiag_put_rings_cfg()
/linux-4.4.14/drivers/net/wireless/realtek/rtl818x/rtl8180/
Ddev.c345 struct rtl8180_tx_ring *ring = &priv->tx_ring[prio]; in rtl8180_handle_tx()
474 ring = &priv->tx_ring[prio]; in rtl8180_tx()
862 priv->tx_ring[1].dma); in rtl8180_init_hw()
864 priv->tx_ring[0].dma); in rtl8180_init_hw()
867 priv->tx_ring[4].dma); in rtl8180_init_hw()
869 priv->tx_ring[0].dma); in rtl8180_init_hw()
871 priv->tx_ring[1].dma); in rtl8180_init_hw()
873 priv->tx_ring[2].dma); in rtl8180_init_hw()
875 priv->tx_ring[3].dma); in rtl8180_init_hw()
1080 priv->tx_ring[prio].desc = ring; in rtl8180_init_tx_ring()
[all …]
Drtl8180.h117 struct rtl8180_tx_ring tx_ring[RTL818X_NR_TX_QUEUES]; member
/linux-4.4.14/drivers/net/ethernet/realtek/
D8139cp.c343 struct cp_desc *tx_ring; member
657 struct cp_desc *txd = cp->tx_ring + tx_tail; in cp_tx()
726 txd = &cp->tx_ring[index]; in unwind_tx_frag_mapping()
781 struct cp_desc *txd = &cp->tx_ring[entry]; in cp_start_xmit()
844 txd = &cp->tx_ring[entry]; in cp_start_xmit()
856 txd = &cp->tx_ring[first_entry]; in cp_start_xmit()
1106 memset(cp->tx_ring, 0, sizeof(struct cp_desc) * CP_TX_RING_SIZE); in cp_init_rings()
1107 cp->tx_ring[CP_TX_RING_SIZE - 1].opts1 = cpu_to_le32(RingEnd); in cp_init_rings()
1126 cp->tx_ring = &cp->rx_ring[CP_RX_RING_SIZE]; in cp_alloc_rings()
1153 desc = cp->tx_ring + i; in cp_clean_rings()
[all …]
/linux-4.4.14/drivers/net/ethernet/rdc/
Dr6040.c194 struct r6040_descriptor *tx_ring; member
330 lp->tx_remove_ptr = lp->tx_insert_ptr = lp->tx_ring; in r6040_init_txbufs()
331 r6040_init_ring_desc(lp->tx_ring, lp->tx_ring_dma, TX_DCNT); in r6040_init_txbufs()
507 if (lp->tx_ring) { in r6040_close()
509 TX_DESC_SIZE, lp->tx_ring, lp->tx_ring_dma); in r6040_close()
510 lp->tx_ring = NULL; in r6040_close()
780 lp->tx_ring = in r6040_open()
782 if (!lp->tx_ring) { in r6040_open()
797 pci_free_consistent(lp->pdev, TX_DESC_SIZE, lp->tx_ring, in r6040_open()
/linux-4.4.14/drivers/net/ethernet/broadcom/
Db44.c1028 bp->tx_ring[entry].ctrl = cpu_to_le32(ctrl); in b44_start_xmit()
1029 bp->tx_ring[entry].addr = cpu_to_le32((u32) mapping+bp->dma_offset); in b44_start_xmit()
1033 entry * sizeof(bp->tx_ring[0]), in b44_start_xmit()
1139 memset(bp->tx_ring, 0, B44_TX_RING_BYTES); in b44_init_rings()
1176 if (bp->tx_ring) { in b44_free_consistent()
1180 kfree(bp->tx_ring); in b44_free_consistent()
1183 bp->tx_ring, bp->tx_ring_dma); in b44_free_consistent()
1184 bp->tx_ring = NULL; in b44_free_consistent()
1236 bp->tx_ring = dma_alloc_coherent(bp->sdev->dma_dev, size, in b44_alloc_consistent()
1238 if (!bp->tx_ring) { in b44_alloc_consistent()
[all …]
Dbgmac.c583 bgmac_dma_tx_ring_free(bgmac, &bgmac->tx_ring[i]); in bgmac_dma_cleanup()
594 bgmac_dma_ring_desc_free(bgmac, &bgmac->tx_ring[i], in bgmac_dma_free()
621 ring = &bgmac->tx_ring[i]; in bgmac_dma_alloc()
682 ring = &bgmac->tx_ring[i]; in bgmac_dma_init()
1024 bgmac_dma_tx_reset(bgmac, &bgmac->tx_ring[i]); in bgmac_chip_reset()
1258 bgmac_dma_tx_free(bgmac, &bgmac->tx_ring[0]); in bgmac_poll()
1331 ring = &bgmac->tx_ring[0]; in bgmac_start_xmit()
Db44.h362 struct dma_desc *rx_ring, *tx_ring; member
Dbgmac.h447 struct bgmac_dma_ring tx_ring[BGMAC_MAX_TX_RINGS]; member
Dbnx2.c700 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_free_tx_mem()
752 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_alloc_tx_mem()
2848 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_tx_int()
3396 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_has_fast_work()
3490 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_poll_work()
5126 txr = &bnapi->tx_ring; in bnx2_clear_ring_states()
5178 txr = &bnapi->tx_ring; in bnx2_init_tx_ring()
5425 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_free_tx_skbs()
5806 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_run_loopback()
5811 txr = &tx_napi->tx_ring; in bnx2_run_loopback()
[all …]
/linux-4.4.14/drivers/net/ethernet/nvidia/
Dforcedeth.c830 union ring_type tx_ring; member
1932 np->get_tx = np->put_tx = np->first_tx = np->tx_ring; in nv_init_tx()
1935 np->last_tx.orig = &np->tx_ring.orig[np->tx_ring_size-1]; in nv_init_tx()
1937 np->last_tx.ex = &np->tx_ring.ex[np->tx_ring_size-1]; in nv_init_tx()
1948 np->tx_ring.orig[i].flaglen = 0; in nv_init_tx()
1949 np->tx_ring.orig[i].buf = 0; in nv_init_tx()
1951 np->tx_ring.ex[i].flaglen = 0; in nv_init_tx()
1952 np->tx_ring.ex[i].txvlan = 0; in nv_init_tx()
1953 np->tx_ring.ex[i].bufhigh = 0; in nv_init_tx()
1954 np->tx_ring.ex[i].buflow = 0; in nv_init_tx()
[all …]
/linux-4.4.14/drivers/staging/rtl8192e/rtl8192e/
Drtl_core.c274 struct rtl8192_tx_ring *ring = &priv->tx_ring[prio]; in _rtl92e_check_nic_enough_desc()
538 ring = &priv->tx_ring[BEACON_QUEUE]; in _rtl92e_prepare_beacon()
1116 if (skb_queue_len(&(&priv->tx_ring[i])->queue) > 0) { in _rtl92e_is_tx_queue_empty()
1118 i, skb_queue_len(&(&priv->tx_ring[i])->queue)); in _rtl92e_is_tx_queue_empty()
1151 ring = &priv->tx_ring[QueueID]; in _rtl92e_tx_check_stuck()
1595 struct rtl8192_tx_ring *ring = &priv->tx_ring[prio]; in _rtl92e_free_tx_ring()
1681 struct rtl8192_tx_ring *ring = &priv->tx_ring[prio]; in _rtl92e_tx_isr()
1713 ring = &priv->tx_ring[TXCMD_QUEUE]; in _rtl92e_tx_cmd()
1772 ring = &priv->tx_ring[tcb_desc->queue_index]; in _rtl92e_tx()
1864 priv->tx_ring[prio].desc = ring; in _rtl92e_alloc_tx_ring()
[all …]
Drtl_core.h269 struct tx_ring { struct
272 struct tx_ring *next; argument
404 struct rtl8192_tx_ring tx_ring[MAX_TX_QUEUE_COUNT]; member
Dr8192E_phy.c1492 ring = &priv->tx_ring[QueueID]; in _rtl92e_set_rf_power_state()
1521 ring = &priv->tx_ring[QueueID]; in _rtl92e_set_rf_power_state()
/linux-4.4.14/drivers/net/ethernet/marvell/
Dskge.c502 p->tx_pending = skge->tx_ring.count; in skge_get_ring_param()
516 skge->tx_ring.count = p->tx_pending; in skge_set_ring_param()
2546 tx_size = skge->tx_ring.count * sizeof(struct skge_tx_desc); in skge_up()
2570 err = skge_ring_alloc(&skge->tx_ring, skge->mem + rx_size, in skge_up()
2601 BUG_ON(skge->tx_ring.to_use != skge->tx_ring.to_clean); in skge_up()
2603 skge_qset(skge, txqaddr[port], skge->tx_ring.to_use); in skge_up()
2623 kfree(skge->tx_ring.start); in skge_up()
2717 kfree(skge->tx_ring.start); in skge_down()
2744 if (unlikely(skge_avail(&skge->tx_ring) < skb_shinfo(skb)->nr_frags + 1)) in skge_xmit_frame()
2747 e = skge->tx_ring.to_use; in skge_xmit_frame()
[all …]
Dsky2.c1617 sky2->tx_ring = kcalloc(sky2->tx_ring_size, sizeof(struct tx_ring_info), in sky2_alloc_buffers()
1619 if (!sky2->tx_ring) in sky2_alloc_buffers()
1654 kfree(sky2->tx_ring); in sky2_free_buffers()
1657 sky2->tx_ring = NULL; in sky2_free_buffers()
1937 re = sky2->tx_ring + slot; in sky2_xmit_frame()
1966 re = sky2->tx_ring + slot; in sky2_xmit_frame()
1993 re = sky2->tx_ring + i; in sky2_xmit_frame()
2025 struct tx_ring_info *re = sky2->tx_ring + idx; in sky2_tx_complete()
/linux-4.4.14/net/netlink/
Daf_netlink.c311 return nlk_sk(sk)->tx_ring.pg_vec != NULL; in netlink_tx_is_mmaped()
380 __netlink_set_ring(struct sock *sk, struct nl_mmap_req *req, bool tx_ring, void **pg_vec, in __netlink_set_ring() argument
387 queue = tx_ring ? &sk->sk_write_queue : &sk->sk_receive_queue; in __netlink_set_ring()
388 ring = tx_ring ? &nlk->tx_ring : &nlk->rx_ring; in __netlink_set_ring()
411 bool tx_ring) in netlink_set_ring() argument
418 ring = tx_ring ? &nlk->tx_ring : &nlk->rx_ring; in netlink_set_ring()
457 __netlink_set_ring(sk, req, tx_ring, pg_vec, order); in netlink_set_ring()
511 for (ring = &nlk->rx_ring; ring <= &nlk->tx_ring; ring++) { in netlink_mmap()
525 for (ring = &nlk->rx_ring; ring <= &nlk->tx_ring; ring++) { in netlink_mmap()
709 if (nlk->tx_ring.pg_vec) { in netlink_poll()
[all …]
Daf_netlink.h50 struct netlink_ring tx_ring; member
Ddiag.c33 ret = sk_diag_put_ring(&nlk->tx_ring, NETLINK_DIAG_TX_RING, in sk_diag_put_rings_cfg()
/linux-4.4.14/drivers/net/ethernet/pasemi/
Dpasemi_mac.c125 static struct pasemi_mac_txring *tx_ring(const struct pasemi_mac *mac) in tx_ring() function
556 struct pasemi_mac_txring *txring = tx_ring(mac); in pasemi_mac_free_tx_resources()
696 pcnt = *tx_ring(mac)->chan.status & PAS_STATUS_PCNT_M; in pasemi_mac_restart_tx_intr()
700 write_iob_reg(PAS_IOB_DMA_TXCH_RESET(tx_ring(mac)->chan.chno), reg); in pasemi_mac_restart_tx_intr()
727 struct pasemi_dmachan *chan = &tx_ring(mac)->chan; in pasemi_mac_tx_error()
1177 pasemi_dma_start_chan(&tx_ring(mac)->chan, PAS_DMA_TXCHAN_TCMDSTA_SZ | in pasemi_mac_open()
1267 int txch = tx_ring(mac)->chan.chno; in pasemi_mac_pause_txchan()
1332 txch = tx_ring(mac)->chan.chno; in pasemi_mac_close()
1363 pasemi_mac_clean_tx(tx_ring(mac)); in pasemi_mac_close()
1478 struct pasemi_mac_txring * const txring = tx_ring(mac); in pasemi_mac_start_tx()
[all …]
/linux-4.4.14/drivers/infiniband/ulp/ipoib/
Dipoib_cm.c733 tx_req = &tx->tx_ring[tx->tx_head & (ipoib_sendq_size - 1)]; in ipoib_cm_send()
786 tx_req = &tx->tx_ring[wr_id]; in ipoib_cm_handle_tx_wc()
1114 p->tx_ring = __vmalloc(ipoib_sendq_size * sizeof *p->tx_ring, in ipoib_cm_tx_init()
1116 if (!p->tx_ring) { in ipoib_cm_tx_init()
1121 memset(p->tx_ring, 0, ipoib_sendq_size * sizeof *p->tx_ring); in ipoib_cm_tx_init()
1162 vfree(p->tx_ring); in ipoib_cm_tx_init()
1179 if (p->tx_ring) { in ipoib_cm_tx_destroy()
1196 tx_req = &p->tx_ring[p->tx_tail & (ipoib_sendq_size - 1)]; in ipoib_cm_tx_destroy()
1211 vfree(p->tx_ring); in ipoib_cm_tx_destroy()
Dipoib_main.c1541 priv->tx_ring = vzalloc(ipoib_sendq_size * sizeof *priv->tx_ring); in ipoib_dev_init()
1542 if (!priv->tx_ring) { in ipoib_dev_init()
1566 vfree(priv->tx_ring); in ipoib_dev_init()
1602 vfree(priv->tx_ring); in ipoib_dev_cleanup()
1605 priv->tx_ring = NULL; in ipoib_dev_cleanup()
Dipoib.h242 struct ipoib_tx_buf *tx_ring; member
359 struct ipoib_tx_buf *tx_ring; member
Dipoib_ib.c390 tx_req = &priv->tx_ring[wr_id]; in ipoib_ib_handle_tx_wc()
577 tx_req = &priv->tx_ring[priv->tx_head & (ipoib_sendq_size - 1)]; in ipoib_send()
855 tx_req = &priv->tx_ring[priv->tx_tail & in ipoib_ib_dev_stop()
/linux-4.4.14/drivers/net/ethernet/cadence/
Dmacb.c74 return &queue->tx_ring[macb_tx_ring_wrap(index)]; in macb_tx_desc()
1247 desc = &queue->tx_ring[entry]; in macb_tx_map()
1254 desc = &queue->tx_ring[entry]; in macb_tx_map()
1423 if (queue->tx_ring) { in macb_free_consistent()
1425 queue->tx_ring, queue->tx_ring_dma); in macb_free_consistent()
1426 queue->tx_ring = NULL; in macb_free_consistent()
1470 queue->tx_ring = dma_alloc_coherent(&bp->pdev->dev, size, in macb_alloc_consistent()
1473 if (!queue->tx_ring) in macb_alloc_consistent()
1478 queue->tx_ring); in macb_alloc_consistent()
1513 queue->tx_ring[i].addr = 0; in gem_init_rings()
[all …]
Dmacb.h779 struct macb_dma_desc *tx_ring; member
/linux-4.4.14/drivers/net/ethernet/octeon/
Docteon_mgmt.c126 u64 *tx_ring; member
273 re.d64 = p->tx_ring[p->tx_next_clean]; in octeon_mgmt_clean_tx_buffers()
983 p->tx_ring = kzalloc(ring_size_to_bytes(OCTEON_MGMT_TX_RING_SIZE), in octeon_mgmt_open()
985 if (!p->tx_ring) in octeon_mgmt_open()
988 dma_map_single(p->dev, p->tx_ring, in octeon_mgmt_open()
1236 kfree(p->tx_ring); in octeon_mgmt_open()
1269 kfree(p->tx_ring); in octeon_mgmt_stop()
1307 p->tx_ring[p->tx_next] = re.d64; in octeon_mgmt_xmit()
/linux-4.4.14/drivers/net/ethernet/natsemi/
Dnatsemi.c542 struct netdev_desc *tx_ring; member
1863 printk(KERN_DEBUG " Tx ring at %p:\n", np->tx_ring); in dump_ring()
1866 i, np->tx_ring[i].next_desc, in dump_ring()
1867 np->tx_ring[i].cmd_status, in dump_ring()
1868 np->tx_ring[i].addr); in dump_ring()
1920 np->tx_ring = &np->rx_ring[RX_RING_SIZE]; in alloc_ring()
1976 np->tx_ring[i].next_desc = cpu_to_le32(np->ring_dma in init_ring()
1979 np->tx_ring[i].cmd_status = 0; in init_ring()
2081 np->tx_ring[i].cmd_status = 0; in reinit_ring()
2109 np->tx_ring[entry].addr = cpu_to_le32(np->tx_dma[entry]); in start_tx()
[all …]
/linux-4.4.14/drivers/net/ethernet/altera/
Daltera_tse_main.c284 priv->tx_ring = kcalloc(tx_descs, sizeof(struct tse_buffer), in alloc_init_skbufs()
286 if (!priv->tx_ring) in alloc_init_skbufs()
307 kfree(priv->tx_ring); in alloc_init_skbufs()
325 tse_free_tx_buffer(priv, &priv->tx_ring[i]); in free_skbufs()
328 kfree(priv->tx_ring); in free_skbufs()
466 tx_buff = &priv->tx_ring[entry]; in tse_tx_complete()
593 buffer = &priv->tx_ring[entry]; in tse_start_xmit()
Daltera_tse.h434 struct tse_buffer *tx_ring; member
/linux-4.4.14/include/linux/
Dfs_enet_pd.h139 int rx_ring, tx_ring; /* number of buffers on rx */ member
/linux-4.4.14/drivers/net/ethernet/alteon/
Dacenic.c728 if (ap->tx_ring != NULL && !ACE_IS_TIGON_I(ap)) { in ace_free_descriptors()
730 pci_free_consistent(ap->pdev, size, ap->tx_ring, in ace_free_descriptors()
733 ap->tx_ring = NULL; in ace_free_descriptors()
788 ap->tx_ring = pci_alloc_consistent(ap->pdev, size, in ace_allocate_descriptors()
791 if (ap->tx_ring == NULL) in ace_allocate_descriptors()
1286 ap->tx_ring = (__force struct tx_desc *) regs->Window; in ace_init()
1289 writel(0, (__force void __iomem *)ap->tx_ring + i * 4); in ace_init()
1293 memset(ap->tx_ring, 0, in ace_init()
2335 tx = (__force struct tx_desc __iomem *) &ap->tx_ring[i]; in ace_close()
2340 memset(ap->tx_ring + i, 0, in ace_close()
[all …]
Dacenic.h646 struct tx_desc *tx_ring; member
/linux-4.4.14/drivers/crypto/qat/qat_common/
Dadf_transport.c404 struct adf_etr_ring_data *tx_ring; in adf_init_bank() local
440 tx_ring = &bank->rings[i - hw_data->tx_rx_gap]; in adf_init_bank()
441 ring->inflights = tx_ring->inflights; in adf_init_bank()
/linux-4.4.14/drivers/staging/rdma/hfi1/
Dsdma.c563 return sde->tx_ring[sde->tx_head & sde->sdma_mask]; in get_txhead()
592 sde->tx_ring[sde->tx_head++ & sde->sdma_mask] = NULL; in sdma_flush_descq()
986 kvfree(sde->tx_ring); in sdma_clean()
987 sde->tx_ring = NULL; in sdma_clean()
1127 sde->tx_ring = in sdma_init()
1130 if (!sde->tx_ring) in sdma_init()
1131 sde->tx_ring = in sdma_init()
1135 if (!sde->tx_ring) in sdma_init()
1497 sde->tx_ring[sde->tx_head++ & sde->sdma_mask] = NULL; in sdma_make_progress()
2018 WARN_ON_ONCE(sde->tx_ring[sde->tx_tail & sde->sdma_mask]); in submit_tx()
[all …]
Dsdma.h422 struct sdma_txreq **tx_ring; member
/linux-4.4.14/drivers/net/ethernet/renesas/
Dravb_main.c209 if (priv->tx_ring[q]) { in ravb_ring_free()
212 dma_free_coherent(ndev->dev.parent, ring_size, priv->tx_ring[q], in ravb_ring_free()
214 priv->tx_ring[q] = NULL; in ravb_ring_free()
258 memset(priv->tx_ring[q], 0, tx_ring_size); in ravb_ring_format()
260 for (i = 0, tx_desc = priv->tx_ring[q]; i < priv->num_tx_ring[q]; in ravb_ring_format()
323 priv->tx_ring[q] = dma_alloc_coherent(ndev->dev.parent, ring_size, in ravb_ring_init()
326 if (!priv->tx_ring[q]) in ravb_ring_init()
436 desc = &priv->tx_ring[q][entry]; in ravb_tx_free()
1338 desc = &priv->tx_ring[q][entry]; in ravb_start_xmit()
Dsh_eth.c1150 if (mdp->tx_ring) { in sh_eth_ring_free()
1152 dma_free_coherent(NULL, ringsize, mdp->tx_ring, in sh_eth_ring_free()
1154 mdp->tx_ring = NULL; in sh_eth_ring_free()
1219 memset(mdp->tx_ring, 0, tx_ringsize); in sh_eth_ring_format()
1224 txdesc = &mdp->tx_ring[i]; in sh_eth_ring_format()
1277 mdp->tx_ring = dma_alloc_coherent(NULL, tx_ringsize, &mdp->tx_desc_dma, in sh_eth_ring_init()
1279 if (!mdp->tx_ring) in sh_eth_ring_init()
1392 mdp->tx_ring[i].status &= ~cpu_to_edmac(mdp, TD_TACT); in sh_eth_dev_exit()
1423 txdesc = &mdp->tx_ring[entry]; in sh_eth_txfree()
2404 txdesc = &mdp->tx_ring[entry]; in sh_eth_start_xmit()
Dsh_eth.h509 struct sh_eth_txdesc *tx_ring; member
Dravb.h787 struct ravb_tx_desc *tx_ring[NUM_TX_QUEUE]; member
/linux-4.4.14/firmware/keyspan_pda/
Dkeyspan_pda.S231 mov dptr, #tx_ring
752 mov dph1, #HIGH(tx_ring) ; load DPTR1 with target
792 ;; tx_ring[0..15]
823 ;; tx_ring[0..15]
825 mov dptr, #tx_ring ; DPTR1: source
876 mov dph, #HIGH(tx_ring)
1117 tx_ring: label
Dxircom_pgs.S233 mov dptr, #tx_ring
809 mov dph1, #HIGH(tx_ring) ; load DPTR1 with target
853 ;; tx_ring[0..15]
884 ;; tx_ring[0..15]
886 mov dptr, #tx_ring ; DPTR1: source
937 mov dph, #HIGH(tx_ring)
1185 tx_ring: label
/linux-4.4.14/drivers/net/wireless/realtek/rtlwifi/rtl8192ee/
Dtrx.c567 rtlpci->tx_ring[q_idx].avl_desc = point_diff; in rtl92ee_get_available_desc()
595 current_bd_desc = rtlpci->tx_ring[queue_index].cur_tx_wp; in rtl92ee_pre_fill_tx_bd_desc()
649 rtlpci->tx_ring[queue_index].dma + in rtl92ee_pre_fill_tx_bd_desc()
929 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[q_idx]; in rtl92ee_set_desc()
1045 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl92ee_is_tx_desc_closed()
Dhw.c186 ring = &rtlpci->tx_ring[queue]; in _rtl92ee_set_fw_clock_off()
843 ((u64)rtlpci->tx_ring[BEACON_QUEUE].buffer_desc_dma) & in _rtl92ee_init_mac()
846 (u64)rtlpci->tx_ring[MGNT_QUEUE].buffer_desc_dma & in _rtl92ee_init_mac()
849 (u64)rtlpci->tx_ring[VO_QUEUE].buffer_desc_dma & in _rtl92ee_init_mac()
852 (u64)rtlpci->tx_ring[VI_QUEUE].buffer_desc_dma & in _rtl92ee_init_mac()
856 (u64)rtlpci->tx_ring[BE_QUEUE].buffer_desc_dma & in _rtl92ee_init_mac()
862 (u64)rtlpci->tx_ring[BK_QUEUE].buffer_desc_dma & in _rtl92ee_init_mac()
865 (u64)rtlpci->tx_ring[HIGH_QUEUE].buffer_desc_dma & in _rtl92ee_init_mac()
/linux-4.4.14/drivers/infiniband/ulp/srp/
Dib_srp.h165 struct srp_iu **tx_ring; member
Dib_srp.c646 if (ch->tx_ring) { in srp_free_ch_ib()
648 srp_free_iu(target->srp_host, ch->tx_ring[i]); in srp_free_ch_ib()
649 kfree(ch->tx_ring); in srp_free_ch_ib()
650 ch->tx_ring = NULL; in srp_free_ch_ib()
1239 list_add(&ch->tx_ring[j]->list, &ch->free_tx); in srp_rport_reconnect()
2165 ch->tx_ring = kcalloc(target->queue_size, sizeof(*ch->tx_ring), in srp_alloc_iu_bufs()
2167 if (!ch->tx_ring) in srp_alloc_iu_bufs()
2179 ch->tx_ring[i] = srp_alloc_iu(target->srp_host, in srp_alloc_iu_bufs()
2182 if (!ch->tx_ring[i]) in srp_alloc_iu_bufs()
2185 list_add(&ch->tx_ring[i]->list, &ch->free_tx); in srp_alloc_iu_bufs()
[all …]
/linux-4.4.14/drivers/net/wireless/realtek/rtlwifi/rtl8192se/
Dhw.c726 rtl_write_dword(rtlpriv, TBKDA, rtlpci->tx_ring[BK_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
727 rtl_write_dword(rtlpriv, TBEDA, rtlpci->tx_ring[BE_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
728 rtl_write_dword(rtlpriv, TVIDA, rtlpci->tx_ring[VI_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
729 rtl_write_dword(rtlpriv, TVODA, rtlpci->tx_ring[VO_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
730 rtl_write_dword(rtlpriv, TBDA, rtlpci->tx_ring[BEACON_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
731 rtl_write_dword(rtlpriv, TCDA, rtlpci->tx_ring[TXCMD_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
732 rtl_write_dword(rtlpriv, TMDA, rtlpci->tx_ring[MGNT_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
733 rtl_write_dword(rtlpriv, THPDA, rtlpci->tx_ring[HIGH_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
734 rtl_write_dword(rtlpriv, HDA, rtlpci->tx_ring[HCCA_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
Dsw.c245 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl92se_is_tx_desc_closed()
Dfw.c146 ring = &rtlpci->tx_ring[TXCMD_QUEUE]; in _rtl92s_cmd_send_packet()
/linux-4.4.14/drivers/net/wireless/realtek/rtlwifi/rtl8821ae/
Dhw.c50 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[BEACON_QUEUE]; in _rtl8821ae_return_beacon_queue_skb()
208 ring = &rtlpci->tx_ring[queue]; in _rtl8821ae_set_fw_clock_off()
1003 rtlpci->tx_ring[BEACON_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac()
1005 rtlpci->tx_ring[MGNT_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac()
1007 rtlpci->tx_ring[VO_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac()
1009 rtlpci->tx_ring[VI_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac()
1011 rtlpci->tx_ring[BE_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac()
1013 rtlpci->tx_ring[BK_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac()
1015 rtlpci->tx_ring[HIGH_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac()
1457 rtlpci->tx_ring[BEACON_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_trx_desc_hw_address()
[all …]
/linux-4.4.14/drivers/net/ethernet/broadcom/genet/
Dbcmgenet.c2461 struct bcmgenet_tx_ring *tx_ring; in bcmgenet_isr1() local
2493 tx_ring = &priv->tx_rings[index]; in bcmgenet_isr1()
2495 if (likely(napi_schedule_prep(&tx_ring->napi))) { in bcmgenet_isr1()
2496 tx_ring->int_disable(tx_ring); in bcmgenet_isr1()
2497 __napi_schedule(&tx_ring->napi); in bcmgenet_isr1()
2509 struct bcmgenet_tx_ring *tx_ring; in bcmgenet_isr0() local
2532 tx_ring = &priv->tx_rings[DESC_INDEX]; in bcmgenet_isr0()
2534 if (likely(napi_schedule_prep(&tx_ring->napi))) { in bcmgenet_isr0()
2535 tx_ring->int_disable(tx_ring); in bcmgenet_isr0()
2536 __napi_schedule(&tx_ring->napi); in bcmgenet_isr0()
/linux-4.4.14/drivers/net/ethernet/adaptec/
Dstarfire.c534 starfire_tx_desc *tx_ring; member
915 np->tx_ring = (void *) np->rx_done_q + rx_done_q_size; in netdev_open()
917 np->rx_ring = (void *) np->tx_ring + tx_ring_size; in netdev_open()
1239 np->tx_ring[entry].addr = cpu_to_dma(np->tx_info[entry].mapping); in start_tx()
1240 np->tx_ring[entry].status = cpu_to_le32(status); in start_tx()
1922 i, le32_to_cpu(np->tx_ring[i].status), in netdev_close()
1923 (long long) dma_to_cpu(np->tx_ring[i].addr), in netdev_close()
/linux-4.4.14/drivers/net/wireless/realtek/rtlwifi/rtl8188ee/
Dhw.c92 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[BEACON_QUEUE]; in _rtl88ee_return_beacon_queue_skb()
207 ring = &rtlpci->tx_ring[queue]; in _rtl88ee_set_fw_clock_off()
911 ((u64) rtlpci->tx_ring[BEACON_QUEUE].dma) & in _rtl88ee_init_mac()
914 (u64) rtlpci->tx_ring[MGNT_QUEUE].dma & in _rtl88ee_init_mac()
917 (u64) rtlpci->tx_ring[VO_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl88ee_init_mac()
919 (u64) rtlpci->tx_ring[VI_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl88ee_init_mac()
921 (u64) rtlpci->tx_ring[BE_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl88ee_init_mac()
923 (u64) rtlpci->tx_ring[BK_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl88ee_init_mac()
925 (u64) rtlpci->tx_ring[HIGH_QUEUE].dma & in _rtl88ee_init_mac()
/linux-4.4.14/drivers/net/wireless/realtek/rtlwifi/rtl8723be/
Dhw.c53 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[BEACON_QUEUE]; in _rtl8723be_return_beacon_queue_skb()
209 ring = &rtlpci->tx_ring[queue]; in _rtl8723be_set_fw_clock_off()
901 ((u64) rtlpci->tx_ring[BEACON_QUEUE].dma) & in _rtl8723be_init_mac()
904 (u64) rtlpci->tx_ring[MGNT_QUEUE].dma & in _rtl8723be_init_mac()
907 (u64) rtlpci->tx_ring[VO_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8723be_init_mac()
909 (u64) rtlpci->tx_ring[VI_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8723be_init_mac()
911 (u64) rtlpci->tx_ring[BE_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8723be_init_mac()
913 (u64) rtlpci->tx_ring[BK_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8723be_init_mac()
915 (u64) rtlpci->tx_ring[HIGH_QUEUE].dma & in _rtl8723be_init_mac()
Dtrx.c736 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl8723be_is_tx_desc_closed()
/linux-4.4.14/Documentation/networking/
Dnetlink_mmap.txt219 void *rx_ring, *tx_ring;
235 tx_ring = rx_ring + ring_size:
314 hdr = tx_ring + frame_offset;
/linux-4.4.14/drivers/net/wireless/
Dadm8211.c316 u32 status = le32_to_cpu(priv->tx_ring[entry].status); in adm8211_interrupt_tci()
1459 desc = &priv->tx_ring[i]; in adm8211_init_rings()
1644 priv->tx_ring[entry].buffer1 = cpu_to_le32(mapping); in adm8211_tx_raw()
1648 priv->tx_ring[entry].length = cpu_to_le32(flag | skb->len); in adm8211_tx_raw()
1652 priv->tx_ring[entry].status = cpu_to_le32(flag); in adm8211_tx_raw()
1742 priv->tx_ring = priv->rx_ring + priv->rx_ring_size; in adm8211_alloc_rings()
Dadm8211.h541 struct adm8211_desc *tx_ring; member
/linux-4.4.14/drivers/net/wireless/realtek/rtlwifi/rtl8192de/
Dhw.c761 rtlpci->tx_ring[BEACON_QUEUE].dma); in _rtl92de_init_mac()
762 rtl_write_dword(rtlpriv, REG_MGQ_DESA, rtlpci->tx_ring[MGNT_QUEUE].dma); in _rtl92de_init_mac()
763 rtl_write_dword(rtlpriv, REG_VOQ_DESA, rtlpci->tx_ring[VO_QUEUE].dma); in _rtl92de_init_mac()
764 rtl_write_dword(rtlpriv, REG_VIQ_DESA, rtlpci->tx_ring[VI_QUEUE].dma); in _rtl92de_init_mac()
765 rtl_write_dword(rtlpriv, REG_BEQ_DESA, rtlpci->tx_ring[BE_QUEUE].dma); in _rtl92de_init_mac()
766 rtl_write_dword(rtlpriv, REG_BKQ_DESA, rtlpci->tx_ring[BK_QUEUE].dma); in _rtl92de_init_mac()
767 rtl_write_dword(rtlpriv, REG_HQ_DESA, rtlpci->tx_ring[HIGH_QUEUE].dma); in _rtl92de_init_mac()
Dfw.c554 ring = &rtlpci->tx_ring[BEACON_QUEUE]; in _rtl92d_cmd_send_packet()
/linux-4.4.14/drivers/net/wireless/realtek/rtlwifi/rtl8192cu/
Dphy.c448 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl92cu_phy_set_rf_power_state()
491 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl92cu_phy_set_rf_power_state()
/linux-4.4.14/drivers/net/wireless/realtek/rtlwifi/rtl8723ae/
Dhw.c758 ((u64) rtlpci->tx_ring[BEACON_QUEUE].dma) & in _rtl8712e_init_mac()
761 (u64) rtlpci->tx_ring[MGNT_QUEUE].dma & in _rtl8712e_init_mac()
764 (u64) rtlpci->tx_ring[VO_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8712e_init_mac()
766 (u64) rtlpci->tx_ring[VI_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8712e_init_mac()
768 (u64) rtlpci->tx_ring[BE_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8712e_init_mac()
770 (u64) rtlpci->tx_ring[BK_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8712e_init_mac()
772 (u64) rtlpci->tx_ring[HIGH_QUEUE].dma & in _rtl8712e_init_mac()
Dtrx.c687 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl8723e_is_tx_desc_closed()
/linux-4.4.14/drivers/net/wireless/realtek/rtlwifi/rtl8192ce/
Dhw.c761 ((u64) rtlpci->tx_ring[BEACON_QUEUE].dma) & in _rtl92ce_init_mac()
764 (u64) rtlpci->tx_ring[MGNT_QUEUE].dma & in _rtl92ce_init_mac()
767 (u64) rtlpci->tx_ring[VO_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl92ce_init_mac()
769 (u64) rtlpci->tx_ring[VI_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl92ce_init_mac()
771 (u64) rtlpci->tx_ring[BE_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl92ce_init_mac()
773 (u64) rtlpci->tx_ring[BK_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl92ce_init_mac()
775 (u64) rtlpci->tx_ring[HIGH_QUEUE].dma & in _rtl92ce_init_mac()
Dphy.c524 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl92ce_phy_set_rf_power_state()
Dtrx.c745 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl92ce_is_tx_desc_closed()
/linux-4.4.14/drivers/net/wireless/realtek/rtlwifi/rtl8723com/
Dfw_common.c318 ring = &rtlpci->tx_ring[BEACON_QUEUE]; in rtl8723_cmd_send_packet()
/linux-4.4.14/drivers/net/ethernet/rocker/
Drocker.c229 struct rocker_dma_ring_info tx_ring; member
1309 &rocker_port->tx_ring); in rocker_port_dma_rings_init()
1315 err = rocker_dma_ring_bufs_alloc(rocker, &rocker_port->tx_ring, in rocker_port_dma_rings_init()
1355 rocker_dma_ring_bufs_free(rocker, &rocker_port->tx_ring, in rocker_port_dma_rings_init()
1358 rocker_dma_ring_destroy(rocker, &rocker_port->tx_ring); in rocker_port_dma_rings_init()
1370 rocker_dma_ring_bufs_free(rocker, &rocker_port->tx_ring, in rocker_port_dma_rings_fini()
1372 rocker_dma_ring_destroy(rocker, &rocker_port->tx_ring); in rocker_port_dma_rings_fini()
4174 desc_info = rocker_desc_head_get(&rocker_port->tx_ring); in rocker_port_xmit()
4208 rocker_desc_head_set(rocker, &rocker_port->tx_ring, desc_info); in rocker_port_xmit()
4210 desc_info = rocker_desc_head_get(&rocker_port->tx_ring); in rocker_port_xmit()
[all …]
/linux-4.4.14/drivers/net/ethernet/broadcom/bnxt/
Dbnxt.c187 txr = &bnapi->tx_ring; in bnxt_start_xmit()
420 struct bnxt_tx_ring_info *txr = &bnapi->tx_ring; in bnxt_tx_int()
1461 txr = &bnapi->tx_ring; in bnxt_free_tx_skbs()
1743 txr = &bnapi->tx_ring; in bnxt_free_tx_rings()
1785 txr = &bnapi->tx_ring; in bnxt_alloc_tx_rings()
1905 txr = &bnapi->tx_ring; in bnxt_init_ring_struct()
2039 struct bnxt_tx_ring_info *txr = &bnapi->tx_ring; in bnxt_init_tx_rings()
2426 txr = &bnapi->tx_ring; in bnxt_clear_ring_indices()
3316 struct bnxt_tx_ring_info *txr = &bnapi->tx_ring; in bnxt_hwrm_ring_alloc()
3421 struct bnxt_tx_ring_info *txr = &bnapi->tx_ring; in bnxt_hwrm_ring_free()
[all …]

12