Home
last modified time | relevance | path

Searched refs:tx_ring (Results 1 – 200 of 210) sorted by relevance

12

/linux-4.1.27/drivers/net/ethernet/intel/i40evf/
Di40e_txrx.c80 void i40evf_clean_tx_ring(struct i40e_ring *tx_ring) in i40evf_clean_tx_ring() argument
86 if (!tx_ring->tx_bi) in i40evf_clean_tx_ring()
90 for (i = 0; i < tx_ring->count; i++) in i40evf_clean_tx_ring()
91 i40e_unmap_and_free_tx_resource(tx_ring, &tx_ring->tx_bi[i]); in i40evf_clean_tx_ring()
93 bi_size = sizeof(struct i40e_tx_buffer) * tx_ring->count; in i40evf_clean_tx_ring()
94 memset(tx_ring->tx_bi, 0, bi_size); in i40evf_clean_tx_ring()
97 memset(tx_ring->desc, 0, tx_ring->size); in i40evf_clean_tx_ring()
99 tx_ring->next_to_use = 0; in i40evf_clean_tx_ring()
100 tx_ring->next_to_clean = 0; in i40evf_clean_tx_ring()
102 if (!tx_ring->netdev) in i40evf_clean_tx_ring()
[all …]
Di40e_txrx.h294 void i40evf_clean_tx_ring(struct i40e_ring *tx_ring);
296 int i40evf_setup_tx_descriptors(struct i40e_ring *tx_ring);
298 void i40evf_free_tx_resources(struct i40e_ring *tx_ring);
Di40evf_main.c370 struct i40e_ring *tx_ring = adapter->tx_rings[t_idx]; in i40evf_map_vector_to_txq() local
372 tx_ring->q_vector = q_vector; in i40evf_map_vector_to_txq()
373 tx_ring->next = q_vector->tx.ring; in i40evf_map_vector_to_txq()
374 tx_ring->vsi = &adapter->vsi; in i40evf_map_vector_to_txq()
375 q_vector->tx.ring = tx_ring; in i40evf_map_vector_to_txq()
1099 struct i40e_ring *tx_ring; in i40evf_alloc_queues() local
1102 tx_ring = kzalloc(sizeof(*tx_ring) * 2, GFP_KERNEL); in i40evf_alloc_queues()
1103 if (!tx_ring) in i40evf_alloc_queues()
1106 tx_ring->queue_index = i; in i40evf_alloc_queues()
1107 tx_ring->netdev = adapter->netdev; in i40evf_alloc_queues()
[all …]
/linux-4.1.27/drivers/net/ethernet/intel/i40e/
Di40e_txrx.c57 struct i40e_ring *tx_ring; in i40e_program_fdir_filter() local
74 tx_ring = vsi->tx_rings[0]; in i40e_program_fdir_filter()
75 dev = tx_ring->dev; in i40e_program_fdir_filter()
79 if (I40E_DESC_UNUSED(tx_ring) > 1) in i40e_program_fdir_filter()
85 if (!(I40E_DESC_UNUSED(tx_ring) > 1)) in i40e_program_fdir_filter()
94 i = tx_ring->next_to_use; in i40e_program_fdir_filter()
95 fdir_desc = I40E_TX_FDIRDESC(tx_ring, i); in i40e_program_fdir_filter()
96 first = &tx_ring->tx_bi[i]; in i40e_program_fdir_filter()
99 tx_ring->next_to_use = ((i + 1) < tx_ring->count) ? i + 1 : 0; in i40e_program_fdir_filter()
147 i = tx_ring->next_to_use; in i40e_program_fdir_filter()
[all …]
Di40e_fcoe.c1005 static void i40e_fcoe_program_ddp(struct i40e_ring *tx_ring, in i40e_fcoe_program_ddp() argument
1012 struct i40e_pf *pf = tx_ring->vsi->back; in i40e_fcoe_program_ddp()
1013 u16 i = tx_ring->next_to_use; in i40e_fcoe_program_ddp()
1035 ddp_desc = I40E_DDP_CONTEXT_DESC(tx_ring, i); in i40e_fcoe_program_ddp()
1037 if (i == tx_ring->count) in i40e_fcoe_program_ddp()
1057 queue_desc = I40E_QUEUE_CONTEXT_DESC(tx_ring, i++); in i40e_fcoe_program_ddp()
1058 if (i == tx_ring->count) in i40e_fcoe_program_ddp()
1067 filter_desc = I40E_FILTER_CONTEXT_DESC(tx_ring, i); in i40e_fcoe_program_ddp()
1069 if (i == tx_ring->count) in i40e_fcoe_program_ddp()
1092 tx_ring->next_to_use = i; in i40e_fcoe_program_ddp()
[all …]
Di40e_txrx.h297 void i40e_clean_tx_ring(struct i40e_ring *tx_ring);
299 int i40e_setup_tx_descriptors(struct i40e_ring *tx_ring);
301 void i40e_free_tx_resources(struct i40e_ring *tx_ring);
305 void i40e_tx_map(struct i40e_ring *tx_ring, struct sk_buff *skb,
308 int i40e_maybe_stop_tx(struct i40e_ring *tx_ring, int size);
309 int i40e_xmit_descriptor_count(struct sk_buff *skb, struct i40e_ring *tx_ring);
311 struct i40e_ring *tx_ring, u32 *flags);
Di40e_debugfs.c537 struct i40e_ring *tx_ring = ACCESS_ONCE(vsi->tx_rings[i]); in i40e_dbg_dump_vsi_seid() local
538 if (!tx_ring) in i40e_dbg_dump_vsi_seid()
543 i, tx_ring->desc); in i40e_dbg_dump_vsi_seid()
546 i, tx_ring->dev, in i40e_dbg_dump_vsi_seid()
547 tx_ring->netdev, in i40e_dbg_dump_vsi_seid()
548 tx_ring->tx_bi); in i40e_dbg_dump_vsi_seid()
551 i, tx_ring->state, in i40e_dbg_dump_vsi_seid()
552 tx_ring->queue_index, in i40e_dbg_dump_vsi_seid()
553 tx_ring->reg_idx); in i40e_dbg_dump_vsi_seid()
556 i, tx_ring->dtype); in i40e_dbg_dump_vsi_seid()
[all …]
Di40e_main.c385 struct i40e_ring *tx_ring, *rx_ring; in i40e_get_netdev_stats_struct() local
401 tx_ring = ACCESS_ONCE(vsi->tx_rings[i]); in i40e_get_netdev_stats_struct()
402 if (!tx_ring) in i40e_get_netdev_stats_struct()
406 start = u64_stats_fetch_begin_irq(&tx_ring->syncp); in i40e_get_netdev_stats_struct()
407 packets = tx_ring->stats.packets; in i40e_get_netdev_stats_struct()
408 bytes = tx_ring->stats.bytes; in i40e_get_netdev_stats_struct()
409 } while (u64_stats_fetch_retry_irq(&tx_ring->syncp, start)); in i40e_get_netdev_stats_struct()
413 rx_ring = &tx_ring[1]; in i40e_get_netdev_stats_struct()
2711 struct i40e_ring *tx_ring, *rx_ring; in i40e_vsi_config_dcb_rings() local
2719 tx_ring = vsi->tx_rings[i]; in i40e_vsi_config_dcb_rings()
[all …]
Di40e_ethtool.c1267 struct i40e_ring *tx_ring, *rx_ring; in i40e_get_ethtool_stats() local
1297 tx_ring = ACCESS_ONCE(vsi->tx_rings[j]); in i40e_get_ethtool_stats()
1299 if (!tx_ring) in i40e_get_ethtool_stats()
1304 start = u64_stats_fetch_begin_irq(&tx_ring->syncp); in i40e_get_ethtool_stats()
1305 data[i] = tx_ring->stats.packets; in i40e_get_ethtool_stats()
1306 data[i + 1] = tx_ring->stats.bytes; in i40e_get_ethtool_stats()
1307 } while (u64_stats_fetch_retry_irq(&tx_ring->syncp, start)); in i40e_get_ethtool_stats()
1311 rx_ring = &tx_ring[1]; in i40e_get_ethtool_stats()
/linux-4.1.27/drivers/net/ethernet/intel/fm10k/
Dfm10k_main.c793 static int fm10k_tso(struct fm10k_ring *tx_ring, in fm10k_tso() argument
826 tx_desc = FM10K_TX_DESC(tx_ring, tx_ring->next_to_use); in fm10k_tso()
832 tx_ring->netdev->features &= ~NETIF_F_GSO_UDP_TUNNEL; in fm10k_tso()
834 netdev_err(tx_ring->netdev, in fm10k_tso()
839 static void fm10k_tx_csum(struct fm10k_ring *tx_ring, in fm10k_tx_csum() argument
859 dev_warn(tx_ring->dev, in fm10k_tx_csum()
861 tx_ring->tx_stats.csum_err++; in fm10k_tx_csum()
880 dev_warn(tx_ring->dev, in fm10k_tx_csum()
884 tx_ring->tx_stats.csum_err++; in fm10k_tx_csum()
897 dev_warn(tx_ring->dev, in fm10k_tx_csum()
[all …]
Dfm10k_netdev.c33 int fm10k_setup_tx_resources(struct fm10k_ring *tx_ring) in fm10k_setup_tx_resources() argument
35 struct device *dev = tx_ring->dev; in fm10k_setup_tx_resources()
38 size = sizeof(struct fm10k_tx_buffer) * tx_ring->count; in fm10k_setup_tx_resources()
40 tx_ring->tx_buffer = vzalloc(size); in fm10k_setup_tx_resources()
41 if (!tx_ring->tx_buffer) in fm10k_setup_tx_resources()
44 u64_stats_init(&tx_ring->syncp); in fm10k_setup_tx_resources()
47 tx_ring->size = tx_ring->count * sizeof(struct fm10k_tx_desc); in fm10k_setup_tx_resources()
48 tx_ring->size = ALIGN(tx_ring->size, 4096); in fm10k_setup_tx_resources()
50 tx_ring->desc = dma_alloc_coherent(dev, tx_ring->size, in fm10k_setup_tx_resources()
51 &tx_ring->dma, GFP_KERNEL); in fm10k_setup_tx_resources()
[all …]
Dfm10k_pci.c330 struct fm10k_ring *tx_ring = interface->tx_ring[i]; in fm10k_update_stats() local
332 restart_queue += tx_ring->tx_stats.restart_queue; in fm10k_update_stats()
333 tx_busy += tx_ring->tx_stats.tx_busy; in fm10k_update_stats()
334 tx_csum_errors += tx_ring->tx_stats.csum_err; in fm10k_update_stats()
335 bytes += tx_ring->stats.bytes; in fm10k_update_stats()
336 pkts += tx_ring->stats.packets; in fm10k_update_stats()
397 struct fm10k_ring *tx_ring = interface->tx_ring[i]; in fm10k_watchdog_flush_tx() local
399 if (tx_ring->next_to_use != tx_ring->next_to_clean) { in fm10k_watchdog_flush_tx()
463 set_check_for_tx_hang(interface->tx_ring[i]); in fm10k_check_hang_subtask()
596 fm10k_configure_tx_ring(interface, interface->tx_ring[i]); in fm10k_configure_tx()
[all …]
Dfm10k.h265 struct fm10k_ring *tx_ring[MAX_QUEUES] ____cacheline_aligned_in_smp; member
446 struct fm10k_ring *tx_ring);
448 bool fm10k_check_tx_hang(struct fm10k_ring *tx_ring);
Dfm10k_ethtool.c225 ring = interface->tx_ring[i]; in fm10k_get_ethtool_stats()
498 interface->tx_ring[i]->count = new_tx_count; in fm10k_set_ringparam()
524 memcpy(&temp_ring[i], interface->tx_ring[i], in fm10k_set_ringparam()
539 fm10k_free_tx_resources(interface->tx_ring[i]); in fm10k_set_ringparam()
541 memcpy(interface->tx_ring[i], &temp_ring[i], in fm10k_set_ringparam()
/linux-4.1.27/drivers/net/ethernet/intel/igbvf/
Dnetdev.c436 struct igbvf_ring *tx_ring) in igbvf_setup_tx_resources() argument
441 size = sizeof(struct igbvf_buffer) * tx_ring->count; in igbvf_setup_tx_resources()
442 tx_ring->buffer_info = vzalloc(size); in igbvf_setup_tx_resources()
443 if (!tx_ring->buffer_info) in igbvf_setup_tx_resources()
447 tx_ring->size = tx_ring->count * sizeof(union e1000_adv_tx_desc); in igbvf_setup_tx_resources()
448 tx_ring->size = ALIGN(tx_ring->size, 4096); in igbvf_setup_tx_resources()
450 tx_ring->desc = dma_alloc_coherent(&pdev->dev, tx_ring->size, in igbvf_setup_tx_resources()
451 &tx_ring->dma, GFP_KERNEL); in igbvf_setup_tx_resources()
452 if (!tx_ring->desc) in igbvf_setup_tx_resources()
455 tx_ring->adapter = adapter; in igbvf_setup_tx_resources()
[all …]
Dethtool.c207 struct igbvf_ring *tx_ring = adapter->tx_ring; in igbvf_get_ringparam() local
213 ring->tx_pending = tx_ring->count; in igbvf_get_ringparam()
235 if ((new_tx_count == adapter->tx_ring->count) && in igbvf_set_ringparam()
245 adapter->tx_ring->count = new_tx_count; in igbvf_set_ringparam()
262 if (new_tx_count != adapter->tx_ring->count) { in igbvf_set_ringparam()
263 memcpy(temp_ring, adapter->tx_ring, sizeof(struct igbvf_ring)); in igbvf_set_ringparam()
270 igbvf_free_tx_resources(adapter->tx_ring); in igbvf_set_ringparam()
272 memcpy(adapter->tx_ring, temp_ring, sizeof(struct igbvf_ring)); in igbvf_set_ringparam()
Digbvf.h203 struct igbvf_ring *tx_ring /* One per active queue */
/linux-4.1.27/drivers/net/ethernet/qlogic/qlcnic/
Dqlcnic_io.c278 struct qlcnic_host_tx_ring *tx_ring = adapter->tx_ring; in qlcnic_82xx_change_filter() local
282 producer = tx_ring->producer; in qlcnic_82xx_change_filter()
283 hwdesc = &tx_ring->desc_head[tx_ring->producer]; in qlcnic_82xx_change_filter()
299 tx_ring->producer = get_next_index(producer, tx_ring->num_desc); in qlcnic_82xx_change_filter()
374 struct qlcnic_host_tx_ring *tx_ring) in qlcnic_tx_encap_pkt() argument
378 u32 producer = tx_ring->producer; in qlcnic_tx_encap_pkt()
408 hwdesc = &tx_ring->desc_head[producer]; in qlcnic_tx_encap_pkt()
409 tx_ring->cmd_buf_arr[producer].skb = NULL; in qlcnic_tx_encap_pkt()
414 producer = get_next_index(producer, tx_ring->num_desc); in qlcnic_tx_encap_pkt()
417 tx_ring->producer = producer; in qlcnic_tx_encap_pkt()
[all …]
Dqlcnic_ctx.c416 struct qlcnic_host_tx_ring *tx_ring, in qlcnic_82xx_fw_cmd_create_tx_ctx() argument
434 tx_ring->producer = 0; in qlcnic_82xx_fw_cmd_create_tx_ctx()
435 tx_ring->sw_consumer = 0; in qlcnic_82xx_fw_cmd_create_tx_ctx()
436 *(tx_ring->hw_consumer) = 0; in qlcnic_82xx_fw_cmd_create_tx_ctx()
477 prq->cmd_cons_dma_addr = cpu_to_le64(tx_ring->hw_cons_phys_addr); in qlcnic_82xx_fw_cmd_create_tx_ctx()
481 prq_cds->host_phys_addr = cpu_to_le64(tx_ring->phys_addr); in qlcnic_82xx_fw_cmd_create_tx_ctx()
482 prq_cds->ring_size = cpu_to_le32(tx_ring->num_desc); in qlcnic_82xx_fw_cmd_create_tx_ctx()
496 tx_ring->state = le32_to_cpu(prsp->host_ctx_state); in qlcnic_82xx_fw_cmd_create_tx_ctx()
498 tx_ring->crb_cmd_producer = adapter->ahw->pci_base0 + temp; in qlcnic_82xx_fw_cmd_create_tx_ctx()
499 tx_ring->ctx_id = le16_to_cpu(prsp->context_id); in qlcnic_82xx_fw_cmd_create_tx_ctx()
[all …]
Dqlcnic_main.c124 inline void qlcnic_update_cmd_producer(struct qlcnic_host_tx_ring *tx_ring) in qlcnic_update_cmd_producer() argument
126 writel(tx_ring->producer, tx_ring->crb_cmd_producer); in qlcnic_update_cmd_producer()
1712 struct qlcnic_host_tx_ring *tx_ring; in qlcnic_request_irq() local
1782 tx_ring = &adapter->tx_ring[ring]; in qlcnic_request_irq()
1783 snprintf(tx_ring->name, sizeof(tx_ring->name), in qlcnic_request_irq()
1785 err = request_irq(tx_ring->irq, handler, flags, in qlcnic_request_irq()
1786 tx_ring->name, tx_ring); in qlcnic_request_irq()
1800 struct qlcnic_host_tx_ring *tx_ring; in qlcnic_free_irq() local
1819 tx_ring = &adapter->tx_ring[ring]; in qlcnic_free_irq()
1820 if (tx_ring->irq) in qlcnic_free_irq()
[all …]
Dqlcnic_ethtool.c525 struct qlcnic_host_tx_ring *tx_ring; in qlcnic_get_regs() local
553 tx_ring = &adapter->tx_ring[ring]; in qlcnic_get_regs()
554 regs_buff[i++] = le32_to_cpu(*(tx_ring->hw_consumer)); in qlcnic_get_regs()
555 regs_buff[i++] = tx_ring->sw_consumer; in qlcnic_get_regs()
556 regs_buff[i++] = readl(tx_ring->crb_cmd_producer); in qlcnic_get_regs()
557 regs_buff[i++] = tx_ring->producer; in qlcnic_get_regs()
558 if (tx_ring->crb_intr_mask) in qlcnic_get_regs()
559 regs_buff[i++] = readl(tx_ring->crb_intr_mask); in qlcnic_get_regs()
686 u8 rx_ring, u8 tx_ring) in qlcnic_validate_ring_count() argument
688 if (rx_ring == 0 || tx_ring == 0) in qlcnic_validate_ring_count()
[all …]
Dqlcnic.h73 #define TX_BUFF_RINGSIZE(tx_ring) \ argument
74 (sizeof(struct qlcnic_cmd_buffer) * tx_ring->num_desc)
75 #define TX_DESC_RINGSIZE(tx_ring) \ argument
76 (sizeof(struct cmd_desc_type0) * tx_ring->num_desc)
595 struct qlcnic_host_tx_ring *tx_ring; member
1109 struct qlcnic_host_tx_ring *tx_ring; member
1730 static inline u32 qlcnic_tx_avail(struct qlcnic_host_tx_ring *tx_ring) in qlcnic_tx_avail() argument
1732 if (likely(tx_ring->producer < tx_ring->sw_consumer)) in qlcnic_tx_avail()
1733 return tx_ring->sw_consumer - tx_ring->producer; in qlcnic_tx_avail()
1735 return tx_ring->sw_consumer + tx_ring->num_desc - in qlcnic_tx_avail()
[all …]
Dqlcnic_hw.c395 struct qlcnic_host_tx_ring *tx_ring; in qlcnic_send_cmd_descs() local
402 tx_ring = &adapter->tx_ring[0]; in qlcnic_send_cmd_descs()
403 __netif_tx_lock_bh(tx_ring->txq); in qlcnic_send_cmd_descs()
405 producer = tx_ring->producer; in qlcnic_send_cmd_descs()
407 if (nr_desc >= qlcnic_tx_avail(tx_ring)) { in qlcnic_send_cmd_descs()
408 netif_tx_stop_queue(tx_ring->txq); in qlcnic_send_cmd_descs()
410 if (qlcnic_tx_avail(tx_ring) > nr_desc) { in qlcnic_send_cmd_descs()
411 if (qlcnic_tx_avail(tx_ring) > TX_STOP_THRESH) in qlcnic_send_cmd_descs()
412 netif_tx_wake_queue(tx_ring->txq); in qlcnic_send_cmd_descs()
415 __netif_tx_unlock_bh(tx_ring->txq); in qlcnic_send_cmd_descs()
[all …]
Dqlcnic_init.c131 struct qlcnic_host_tx_ring *tx_ring) in qlcnic_release_tx_buffers() argument
137 spin_lock(&tx_ring->tx_clean_lock); in qlcnic_release_tx_buffers()
139 cmd_buf = tx_ring->cmd_buf_arr; in qlcnic_release_tx_buffers()
140 for (i = 0; i < tx_ring->num_desc; i++) { in qlcnic_release_tx_buffers()
163 spin_unlock(&tx_ring->tx_clean_lock); in qlcnic_release_tx_buffers()
251 sds_ring->tx_ring = &adapter->tx_ring[ring]; in qlcnic_alloc_sw_resources()
253 sds_ring->tx_ring = &adapter->tx_ring[0]; in qlcnic_alloc_sw_resources()
Dqlcnic_hw.h194 struct qlcnic_host_tx_ring *tx_ring, int);
Dqlcnic_83xx_hw.c1290 struct qlcnic_host_tx_ring *tx_ring) in qlcnic_83xx_del_tx_ctx() argument
1304 cmd.req.arg[1] = tx_ring->ctx_id | temp; in qlcnic_83xx_del_tx_ctx()
2225 temp = adapter->tx_ring->ctx_id; in qlcnic_83xx_set_tx_intr_coal()
3486 cmd.req.arg[1] = BIT_1 | (adapter->tx_ring->ctx_id << 16); in qlcnic_83xx_get_stats()
/linux-4.1.27/drivers/net/ethernet/intel/ixgbevf/
Dixgbevf_main.c196 static void ixgbevf_unmap_and_free_tx_resource(struct ixgbevf_ring *tx_ring, in ixgbevf_unmap_and_free_tx_resource() argument
202 dma_unmap_single(tx_ring->dev, in ixgbevf_unmap_and_free_tx_resource()
207 dma_unmap_page(tx_ring->dev, in ixgbevf_unmap_and_free_tx_resource()
238 static inline bool ixgbevf_check_tx_hang(struct ixgbevf_ring *tx_ring) in ixgbevf_check_tx_hang() argument
240 u32 tx_done = ixgbevf_get_tx_completed(tx_ring); in ixgbevf_check_tx_hang()
241 u32 tx_done_old = tx_ring->tx_stats.tx_done_old; in ixgbevf_check_tx_hang()
242 u32 tx_pending = ixgbevf_get_tx_pending(tx_ring); in ixgbevf_check_tx_hang()
244 clear_check_for_tx_hang(tx_ring); in ixgbevf_check_tx_hang()
254 &tx_ring->state); in ixgbevf_check_tx_hang()
257 clear_bit(__IXGBEVF_HANG_CHECK_ARMED, &tx_ring->state); in ixgbevf_check_tx_hang()
[all …]
Dethtool.c268 struct ixgbevf_ring *tx_ring = NULL, *rx_ring = NULL; in ixgbevf_set_ringparam() local
293 adapter->tx_ring[i]->count = new_tx_count; in ixgbevf_set_ringparam()
302 tx_ring = vmalloc(adapter->num_tx_queues * sizeof(*tx_ring)); in ixgbevf_set_ringparam()
303 if (!tx_ring) { in ixgbevf_set_ringparam()
310 tx_ring[i] = *adapter->tx_ring[i]; in ixgbevf_set_ringparam()
311 tx_ring[i].count = new_tx_count; in ixgbevf_set_ringparam()
312 err = ixgbevf_setup_tx_resources(&tx_ring[i]); in ixgbevf_set_ringparam()
316 ixgbevf_free_tx_resources(&tx_ring[i]); in ixgbevf_set_ringparam()
319 vfree(tx_ring); in ixgbevf_set_ringparam()
320 tx_ring = NULL; in ixgbevf_set_ringparam()
[all …]
Dixgbevf.h392 struct ixgbevf_ring *tx_ring[MAX_TX_QUEUES]; /* One per active queue */ member
/linux-4.1.27/drivers/net/ethernet/agere/
Det131x.c357 struct tx_ring { struct
488 struct tx_ring tx_ring; member
1642 struct tx_ring *tx_ring = &adapter->tx_ring; in et131x_config_tx_dma_regs() local
1645 writel(upper_32_bits(tx_ring->tx_desc_ring_pa), &txdma->pr_base_hi); in et131x_config_tx_dma_regs()
1646 writel(lower_32_bits(tx_ring->tx_desc_ring_pa), &txdma->pr_base_lo); in et131x_config_tx_dma_regs()
1652 writel(upper_32_bits(tx_ring->tx_status_pa), &txdma->dma_wb_base_hi); in et131x_config_tx_dma_regs()
1653 writel(lower_32_bits(tx_ring->tx_status_pa), &txdma->dma_wb_base_lo); in et131x_config_tx_dma_regs()
1655 *tx_ring->tx_status = 0; in et131x_config_tx_dma_regs()
1658 tx_ring->send_idx = 0; in et131x_config_tx_dma_regs()
1755 struct tx_ring *tx_ring = &adapter->tx_ring; in et131x_init_send() local
[all …]
/linux-4.1.27/drivers/net/ethernet/intel/ixgb/
Dixgb_main.c702 struct ixgb_desc_ring *txdr = &adapter->tx_ring; in ixgb_setup_tx_resources()
739 u64 tdba = adapter->tx_ring.dma; in ixgb_configure_tx()
740 u32 tdlen = adapter->tx_ring.count * sizeof(struct ixgb_tx_desc); in ixgb_configure_tx()
916 vfree(adapter->tx_ring.buffer_info); in ixgb_free_tx_resources()
917 adapter->tx_ring.buffer_info = NULL; in ixgb_free_tx_resources()
919 dma_free_coherent(&pdev->dev, adapter->tx_ring.size, in ixgb_free_tx_resources()
920 adapter->tx_ring.desc, adapter->tx_ring.dma); in ixgb_free_tx_resources()
922 adapter->tx_ring.desc = NULL; in ixgb_free_tx_resources()
957 struct ixgb_desc_ring *tx_ring = &adapter->tx_ring; in ixgb_clean_tx_ring() local
964 for (i = 0; i < tx_ring->count; i++) { in ixgb_clean_tx_ring()
[all …]
Dixgb_param.c278 struct ixgb_desc_ring *tx_ring = &adapter->tx_ring; in ixgb_check_options() local
281 tx_ring->count = TxDescriptors[bd]; in ixgb_check_options()
282 ixgb_validate_option(&tx_ring->count, &opt); in ixgb_check_options()
284 tx_ring->count = opt.def; in ixgb_check_options()
286 tx_ring->count = ALIGN(tx_ring->count, IXGB_REQ_TX_DESCRIPTOR_MULTIPLE); in ixgb_check_options()
Dixgb_ethtool.c492 struct ixgb_desc_ring *txdr = &adapter->tx_ring; in ixgb_get_ringparam()
506 struct ixgb_desc_ring *txdr = &adapter->tx_ring; in ixgb_set_ringparam()
511 tx_old = adapter->tx_ring; in ixgb_set_ringparam()
539 tx_new = adapter->tx_ring; in ixgb_set_ringparam()
541 adapter->tx_ring = tx_old; in ixgb_set_ringparam()
545 adapter->tx_ring = tx_new; in ixgb_set_ringparam()
556 adapter->tx_ring = tx_old; in ixgb_set_ringparam()
Dixgb.h148 struct ixgb_desc_ring tx_ring ____cacheline_aligned_in_smp;
/linux-4.1.27/drivers/net/ethernet/oki-semi/pch_gbe/
Dpch_gbe_main.c631 adapter->tx_ring = devm_kzalloc(&adapter->pdev->dev, in pch_gbe_alloc_queues()
632 sizeof(*adapter->tx_ring), GFP_KERNEL); in pch_gbe_alloc_queues()
633 if (!adapter->tx_ring) in pch_gbe_alloc_queues()
858 (unsigned long long)adapter->tx_ring->dma, in pch_gbe_configure_tx()
859 adapter->tx_ring->size); in pch_gbe_configure_tx()
862 tdba = adapter->tx_ring->dma; in pch_gbe_configure_tx()
863 tdlen = adapter->tx_ring->size - 0x10; in pch_gbe_configure_tx()
977 struct pch_gbe_tx_ring *tx_ring) in pch_gbe_clean_tx_ring() argument
985 for (i = 0; i < tx_ring->count; i++) { in pch_gbe_clean_tx_ring()
986 buffer_info = &tx_ring->buffer_info[i]; in pch_gbe_clean_tx_ring()
[all …]
Dpch_gbe_ethtool.c270 struct pch_gbe_tx_ring *txdr = adapter->tx_ring; in pch_gbe_get_ringparam()
303 tx_old = adapter->tx_ring; in pch_gbe_set_ringparam()
316 adapter->tx_ring = txdr; in pch_gbe_set_ringparam()
332 err = pch_gbe_setup_tx_resources(adapter, adapter->tx_ring); in pch_gbe_set_ringparam()
339 adapter->tx_ring = tx_old; in pch_gbe_set_ringparam()
341 pch_gbe_free_tx_resources(adapter, adapter->tx_ring); in pch_gbe_set_ringparam()
345 adapter->tx_ring = txdr; in pch_gbe_set_ringparam()
352 adapter->tx_ring = txdr; in pch_gbe_set_ringparam()
362 adapter->tx_ring = tx_old; in pch_gbe_set_ringparam()
Dpch_gbe_param.c459 struct pch_gbe_tx_ring *tx_ring = adapter->tx_ring; in pch_gbe_check_options() local
460 tx_ring->count = TxDescriptors; in pch_gbe_check_options()
461 pch_gbe_validate_option(&tx_ring->count, &opt, adapter); in pch_gbe_check_options()
462 tx_ring->count = roundup(tx_ring->count, in pch_gbe_check_options()
Dpch_gbe.h638 struct pch_gbe_tx_ring *tx_ring; member
664 struct pch_gbe_tx_ring *tx_ring);
/linux-4.1.27/drivers/infiniband/hw/amso1100/
Dc2.c111 static int c2_tx_ring_alloc(struct c2_ring *tx_ring, void *vaddr, in c2_tx_ring_alloc() argument
119 tx_ring->start = kmalloc(sizeof(*elem) * tx_ring->count, GFP_KERNEL); in c2_tx_ring_alloc()
120 if (!tx_ring->start) in c2_tx_ring_alloc()
123 elem = tx_ring->start; in c2_tx_ring_alloc()
126 for (i = 0; i < tx_ring->count; i++, elem++, tx_desc++, txp_desc++) { in c2_tx_ring_alloc()
141 if (i == tx_ring->count - 1) { in c2_tx_ring_alloc()
142 elem->next = tx_ring->start; in c2_tx_ring_alloc()
151 tx_ring->to_use = tx_ring->to_clean = tx_ring->start; in c2_tx_ring_alloc()
326 struct c2_ring *tx_ring = &c2_port->tx_ring; in c2_tx_clean() local
334 elem = tx_ring->start; in c2_tx_clean()
[all …]
Dc2.h363 struct c2_ring tx_ring; member
/linux-4.1.27/drivers/net/ethernet/intel/e1000e/
Dnetdev.c218 struct e1000_ring *tx_ring = adapter->tx_ring; in e1000e_dump() local
262 buffer_info = &tx_ring->buffer_info[tx_ring->next_to_clean]; in e1000e_dump()
264 0, tx_ring->next_to_use, tx_ring->next_to_clean, in e1000e_dump()
306 for (i = 0; tx_ring->desc && (i < tx_ring->count); i++) { in e1000e_dump()
308 tx_desc = E1000_TX_DESC(*tx_ring, i); in e1000e_dump()
309 buffer_info = &tx_ring->buffer_info[i]; in e1000e_dump()
311 if (i == tx_ring->next_to_use && i == tx_ring->next_to_clean) in e1000e_dump()
313 else if (i == tx_ring->next_to_use) in e1000e_dump()
315 else if (i == tx_ring->next_to_clean) in e1000e_dump()
640 static void e1000e_update_tdt_wa(struct e1000_ring *tx_ring, unsigned int i) in e1000e_update_tdt_wa() argument
[all …]
Dethtool.c696 adapter->tx_ring->count = new_tx_count; in e1000_set_ringparam()
731 memcpy(temp_tx, adapter->tx_ring, size); in e1000_set_ringparam()
747 e1000e_free_tx_resources(adapter->tx_ring); in e1000_set_ringparam()
748 memcpy(adapter->tx_ring, temp_tx, size); in e1000_set_ringparam()
1116 struct e1000_ring *tx_ring = &adapter->test_tx_ring; in e1000_free_desc_rings() local
1122 if (tx_ring->desc && tx_ring->buffer_info) { in e1000_free_desc_rings()
1123 for (i = 0; i < tx_ring->count; i++) { in e1000_free_desc_rings()
1124 buffer_info = &tx_ring->buffer_info[i]; in e1000_free_desc_rings()
1149 if (tx_ring->desc) { in e1000_free_desc_rings()
1150 dma_free_coherent(&pdev->dev, tx_ring->size, tx_ring->desc, in e1000_free_desc_rings()
[all …]
De1000.h235 struct e1000_ring *tx_ring ____cacheline_aligned_in_smp;
/linux-4.1.27/drivers/net/ethernet/atheros/atl1e/
Datl1e_main.c636 hw->tpd_thresh = adapter->tx_ring.count / 2; in atl1e_sw_init()
664 struct atl1e_tx_ring *tx_ring = &adapter->tx_ring; in atl1e_clean_tx_ring() local
669 if (tx_ring->desc == NULL || tx_ring->tx_buffer == NULL) in atl1e_clean_tx_ring()
672 ring_count = tx_ring->count; in atl1e_clean_tx_ring()
675 tx_buffer = &tx_ring->tx_buffer[index]; in atl1e_clean_tx_ring()
688 tx_buffer = &tx_ring->tx_buffer[index]; in atl1e_clean_tx_ring()
695 memset(tx_ring->desc, 0, sizeof(struct atl1e_tpd_desc) * in atl1e_clean_tx_ring()
697 memset(tx_ring->tx_buffer, 0, sizeof(struct atl1e_tx_buffer) * in atl1e_clean_tx_ring()
728 *ring_size = ((u32)(adapter->tx_ring.count * in atl1e_cal_ring_size()
754 rwlock_init(&adapter->tx_ring.tx_lock); in atl1e_init_ring_resources()
[all …]
Datl1e_param.c206 adapter->tx_ring.count = (u16) val & 0xFFFC; in atl1e_check_options()
208 adapter->tx_ring.count = (u16)opt.def; in atl1e_check_options()
Datl1e.h458 struct atl1e_tx_ring tx_ring; member
/linux-4.1.27/drivers/net/ethernet/apm/xgene/
Dxgene_enet_main.c227 static int xgene_enet_setup_tx_desc(struct xgene_enet_desc_ring *tx_ring, in xgene_enet_setup_tx_desc() argument
230 struct device *dev = ndev_to_dev(tx_ring->ndev); in xgene_enet_setup_tx_desc()
233 u16 tail = tx_ring->tail; in xgene_enet_setup_tx_desc()
236 raw_desc = &tx_ring->raw_desc[tail]; in xgene_enet_setup_tx_desc()
241 netdev_err(tx_ring->ndev, "DMA mapping error\n"); in xgene_enet_setup_tx_desc()
251 raw_desc->m3 = cpu_to_le64(SET_VAL(HENQNUM, tx_ring->dst_ring_num) | in xgene_enet_setup_tx_desc()
253 tx_ring->cp_ring->cp_skb[tail] = skb; in xgene_enet_setup_tx_desc()
262 struct xgene_enet_desc_ring *tx_ring = pdata->tx_ring; in xgene_enet_start_xmit() local
263 struct xgene_enet_desc_ring *cp_ring = tx_ring->cp_ring; in xgene_enet_start_xmit()
266 tx_level = xgene_enet_ring_len(tx_ring); in xgene_enet_start_xmit()
[all …]
Dxgene_enet_main.h116 struct xgene_enet_desc_ring *tx_ring; member
/linux-4.1.27/drivers/net/ethernet/intel/e1000/
De1000_main.c106 struct e1000_tx_ring *tx_ring);
125 struct e1000_tx_ring *tx_ring);
139 struct e1000_tx_ring *tx_ring);
1245 kfree(adapter->tx_ring); in e1000_probe()
1283 kfree(adapter->tx_ring); in e1000_remove()
1336 adapter->tx_ring = kcalloc(adapter->num_tx_queues, in e1000_alloc_queues()
1338 if (!adapter->tx_ring) in e1000_alloc_queues()
1344 kfree(adapter->tx_ring); in e1000_alloc_queues()
1578 err = e1000_setup_tx_resources(adapter, &adapter->tx_ring[i]); in e1000_setup_all_tx_resources()
1583 &adapter->tx_ring[i]); in e1000_setup_all_tx_resources()
[all …]
De1000_param.c281 struct e1000_tx_ring *tx_ring = adapter->tx_ring; in e1000_check_options() local
298 tx_ring->count = TxDescriptors[bd]; in e1000_check_options()
299 e1000_validate_option(&tx_ring->count, &opt, adapter); in e1000_check_options()
300 tx_ring->count = ALIGN(tx_ring->count, in e1000_check_options()
303 tx_ring->count = opt.def; in e1000_check_options()
306 tx_ring[i].count = tx_ring->count; in e1000_check_options()
De1000_ethtool.c572 struct e1000_tx_ring *txdr = adapter->tx_ring; in e1000_get_ringparam()
602 tx_old = adapter->tx_ring; in e1000_set_ringparam()
616 adapter->tx_ring = txdr; in e1000_set_ringparam()
647 adapter->tx_ring = tx_old; in e1000_set_ringparam()
653 adapter->tx_ring = txdr; in e1000_set_ringparam()
665 adapter->tx_ring = tx_old; in e1000_set_ringparam()
De1000.h252 struct e1000_tx_ring *tx_ring; /* One per active queue */ member
/linux-4.1.27/drivers/net/ethernet/intel/igb/
Digb_main.c367 struct igb_ring *tx_ring; in igb_dump() local
402 tx_ring = adapter->tx_ring[n]; in igb_dump()
403 buffer_info = &tx_ring->tx_buffer_info[tx_ring->next_to_clean]; in igb_dump()
405 n, tx_ring->next_to_use, tx_ring->next_to_clean, in igb_dump()
430 tx_ring = adapter->tx_ring[n]; in igb_dump()
432 pr_info("TX QUEUE INDEX = %d\n", tx_ring->queue_index); in igb_dump()
436 for (i = 0; tx_ring->desc && (i < tx_ring->count); i++) { in igb_dump()
439 tx_desc = IGB_TX_DESC(tx_ring, i); in igb_dump()
440 buffer_info = &tx_ring->tx_buffer_info[i]; in igb_dump()
442 if (i == tx_ring->next_to_use && in igb_dump()
[all …]
Digb.h373 struct igb_ring *tx_ring[16]; member
580 static inline struct netdev_queue *txring_txq(const struct igb_ring *tx_ring) in txring_txq() argument
582 return netdev_get_tx_queue(tx_ring->netdev, tx_ring->queue_index); in txring_txq()
Digb_ethtool.c892 adapter->tx_ring[i]->count = new_tx_count; in igb_set_ringparam()
920 memcpy(&temp_ring[i], adapter->tx_ring[i], in igb_set_ringparam()
935 igb_free_tx_resources(adapter->tx_ring[i]); in igb_set_ringparam()
937 memcpy(adapter->tx_ring[i], &temp_ring[i], in igb_set_ringparam()
1540 struct igb_ring *tx_ring = &adapter->test_tx_ring; in igb_setup_desc_rings() local
1546 tx_ring->count = IGB_DEFAULT_TXD; in igb_setup_desc_rings()
1547 tx_ring->dev = &adapter->pdev->dev; in igb_setup_desc_rings()
1548 tx_ring->netdev = adapter->netdev; in igb_setup_desc_rings()
1549 tx_ring->reg_idx = adapter->vfs_allocated_count; in igb_setup_desc_rings()
1551 if (igb_setup_tx_resources(tx_ring)) { in igb_setup_desc_rings()
[all …]
/linux-4.1.27/drivers/net/ethernet/intel/ixgbe/
Dixgbe_main.c554 struct ixgbe_ring *tx_ring; in ixgbe_dump() local
596 tx_ring = adapter->tx_ring[n]; in ixgbe_dump()
597 tx_buffer = &tx_ring->tx_buffer_info[tx_ring->next_to_clean]; in ixgbe_dump()
599 n, tx_ring->next_to_use, tx_ring->next_to_clean, in ixgbe_dump()
648 tx_ring = adapter->tx_ring[n]; in ixgbe_dump()
650 pr_info("TX QUEUE INDEX = %d\n", tx_ring->queue_index); in ixgbe_dump()
657 for (i = 0; tx_ring->desc && (i < tx_ring->count); i++) { in ixgbe_dump()
658 tx_desc = IXGBE_TX_DESC(tx_ring, i); in ixgbe_dump()
659 tx_buffer = &tx_ring->tx_buffer_info[i]; in ixgbe_dump()
671 if (i == tx_ring->next_to_use && in ixgbe_dump()
[all …]
Dixgbe_lib.c74 adapter->tx_ring[i]->reg_idx = reg_idx; in ixgbe_cache_ring_dcb_sriov()
101 adapter->tx_ring[i]->reg_idx = reg_idx; in ixgbe_cache_ring_dcb_sriov()
184 adapter->tx_ring[offset + i]->reg_idx = tx_idx; in ixgbe_cache_ring_dcb()
186 adapter->tx_ring[offset + i]->dcb_tc = tc; in ixgbe_cache_ring_dcb()
247 adapter->tx_ring[i]->reg_idx = reg_idx; in ixgbe_cache_ring_sriov()
253 adapter->tx_ring[i]->reg_idx = reg_idx; in ixgbe_cache_ring_sriov()
274 adapter->tx_ring[i]->reg_idx = i; in ixgbe_cache_ring_rss()
294 adapter->tx_ring[0]->reg_idx = 0; in ixgbe_cache_ring_register()
900 adapter->tx_ring[txr_idx] = ring; in ixgbe_alloc_q_vector()
975 adapter->tx_ring[ring->queue_index] = NULL; in ixgbe_free_q_vector()
[all …]
Dixgbe_ethtool.c947 struct ixgbe_ring *tx_ring = adapter->tx_ring[0]; in ixgbe_get_ringparam() local
953 ring->tx_pending = tx_ring->count; in ixgbe_get_ringparam()
986 adapter->tx_ring[i]->count = new_tx_count; in ixgbe_set_ringparam()
1013 memcpy(&temp_ring[i], adapter->tx_ring[i], in ixgbe_set_ringparam()
1028 ixgbe_free_tx_resources(adapter->tx_ring[i]); in ixgbe_set_ringparam()
1030 memcpy(adapter->tx_ring[i], &temp_ring[i], in ixgbe_set_ringparam()
1117 ring = adapter->tx_ring[j]; in ixgbe_get_ethtool_stats()
1632 struct ixgbe_ring *tx_ring = &adapter->test_tx_ring; in ixgbe_free_desc_rings() local
1644 reg_ctl = IXGBE_READ_REG(hw, IXGBE_TXDCTL(tx_ring->reg_idx)); in ixgbe_free_desc_rings()
1646 IXGBE_WRITE_REG(hw, IXGBE_TXDCTL(tx_ring->reg_idx), reg_ctl); in ixgbe_free_desc_rings()
[all …]
Dixgbe_fcoe.c511 int ixgbe_fso(struct ixgbe_ring *tx_ring, in ixgbe_fso() argument
523 dev_err(tx_ring->dev, "Wrong gso type %d:expecting SKB_GSO_FCOE\n", in ixgbe_fso()
549 dev_warn(tx_ring->dev, "unknown sof = 0x%x\n", sof); in ixgbe_fso()
575 dev_warn(tx_ring->dev, "unknown eof = 0x%x\n", eof); in ixgbe_fso()
612 ixgbe_tx_ctxtdesc(tx_ring, vlan_macip_lens, fcoe_sof_eof, in ixgbe_fso()
Dixgbe.h657 struct ixgbe_ring *tx_ring[MAX_TX_QUEUES] ____cacheline_aligned_in_smp; member
912 int ixgbe_fso(struct ixgbe_ring *tx_ring, struct ixgbe_tx_buffer *first,
967 struct ixgbe_ring *tx_ring);
/linux-4.1.27/drivers/net/ethernet/packetengines/
Dyellowfin.c310 struct yellowfin_desc *tx_ring; member
441 np->tx_ring = ring_space; in yellowfin_init_one()
514 pci_free_consistent(pdev, TX_TOTAL_SIZE, np->tx_ring, np->tx_ring_dma); in yellowfin_init_one()
700 pr_warn(" Tx ring %p: ", yp->tx_ring); in yellowfin_tx_timeout()
704 yp->tx_ring[i].result_status); in yellowfin_tx_timeout()
762 yp->tx_ring[i].dbdma_cmd = cpu_to_le32(CMD_STOP); in yellowfin_init_ring()
763 yp->tx_ring[i].branch_addr = cpu_to_le32(yp->tx_ring_dma + in yellowfin_init_ring()
767 yp->tx_ring[--i].dbdma_cmd = cpu_to_le32(CMD_STOP | BRANCH_ALWAYS); in yellowfin_init_ring()
775 yp->tx_ring[j].dbdma_cmd = cpu_to_le32(CMD_STOP); in yellowfin_init_ring()
776 yp->tx_ring[j].branch_addr = cpu_to_le32(yp->tx_ring_dma + in yellowfin_init_ring()
[all …]
Dhamachi.c485 struct hamachi_desc *tx_ring; member
651 hmp->tx_ring = ring_space; in hamachi_init_one()
780 pci_free_consistent(pdev, TX_TOTAL_SIZE, hmp->tx_ring, in hamachi_init_one()
1002 if (hmp->tx_ring[entry].status_n_length & cpu_to_le32(DescOwn)) in hamachi_tx()
1008 leXX_to_cpu(hmp->tx_ring[entry].addr), in hamachi_tx()
1013 hmp->tx_ring[entry].status_n_length = 0; in hamachi_tx()
1015 hmp->tx_ring[TX_RING_SIZE-1].status_n_length |= in hamachi_tx()
1063 printk(KERN_DEBUG" Tx ring %p: ", hmp->tx_ring); in hamachi_tx_timeout()
1066 le32_to_cpu(hmp->tx_ring[i].status_n_length)); in hamachi_tx_timeout()
1091 hmp->tx_ring[i].status_n_length = in hamachi_tx_timeout()
[all …]
/linux-4.1.27/drivers/net/ethernet/qlogic/netxen/
Dnetxen_nic_ctx.c448 struct nx_host_tx_ring *tx_ring = adapter->tx_ring; in nx_fw_cmd_create_tx_ctx() local
490 prq_cds->host_phys_addr = cpu_to_le64(tx_ring->phys_addr); in nx_fw_cmd_create_tx_ctx()
491 prq_cds->ring_size = cpu_to_le32(tx_ring->num_desc); in nx_fw_cmd_create_tx_ctx()
503 tx_ring->crb_cmd_producer = netxen_get_ioaddr(adapter, in nx_fw_cmd_create_tx_ctx()
706 struct nx_host_tx_ring *tx_ring; in netxen_init_old_ctx() local
712 tx_ring = adapter->tx_ring; in netxen_init_old_ctx()
716 hwctx->cmd_ring_addr = cpu_to_le64(tx_ring->phys_addr); in netxen_init_old_ctx()
717 hwctx->cmd_ring_size = cpu_to_le32(tx_ring->num_desc); in netxen_init_old_ctx()
762 struct nx_host_tx_ring *tx_ring; in netxen_alloc_hw_resources() local
769 tx_ring = adapter->tx_ring; in netxen_alloc_hw_resources()
[all …]
Dnetxen_nic_init.c136 struct nx_host_tx_ring *tx_ring = adapter->tx_ring; in netxen_release_tx_buffers() local
139 cmd_buf = tx_ring->cmd_buf_arr; in netxen_release_tx_buffers()
140 for (i = 0; i < tx_ring->num_desc; i++) { in netxen_release_tx_buffers()
169 struct nx_host_tx_ring *tx_ring; in netxen_free_sw_resources() local
185 if (adapter->tx_ring == NULL) in netxen_free_sw_resources()
188 tx_ring = adapter->tx_ring; in netxen_free_sw_resources()
189 vfree(tx_ring->cmd_buf_arr); in netxen_free_sw_resources()
190 kfree(tx_ring); in netxen_free_sw_resources()
191 adapter->tx_ring = NULL; in netxen_free_sw_resources()
199 struct nx_host_tx_ring *tx_ring; in netxen_alloc_sw_resources() local
[all …]
Dnetxen_nic_main.c123 struct nx_host_tx_ring *tx_ring) in netxen_nic_update_cmd_producer() argument
125 NXWRIO(adapter, tx_ring->crb_cmd_producer, tx_ring->producer); in netxen_nic_update_cmd_producer()
135 struct nx_host_tx_ring *tx_ring) in netxen_nic_update_cmd_consumer() argument
137 NXWRIO(adapter, tx_ring->crb_cmd_consumer, tx_ring->sw_consumer); in netxen_nic_update_cmd_consumer()
1225 struct nx_host_tx_ring *tx_ring; in netxen_nic_attach() local
1261 tx_ring = adapter->tx_ring; in netxen_nic_attach()
1262 tx_ring->crb_cmd_producer = netxen_get_ioaddr(adapter, in netxen_nic_attach()
1264 tx_ring->crb_cmd_consumer = netxen_get_ioaddr(adapter, in netxen_nic_attach()
1267 tx_ring->producer = 0; in netxen_nic_attach()
1268 tx_ring->sw_consumer = 0; in netxen_nic_attach()
[all …]
Dnetxen_nic_hw.c585 struct nx_host_tx_ring *tx_ring; in netxen_send_cmd_descs() local
592 tx_ring = adapter->tx_ring; in netxen_send_cmd_descs()
593 __netif_tx_lock_bh(tx_ring->txq); in netxen_send_cmd_descs()
595 producer = tx_ring->producer; in netxen_send_cmd_descs()
596 consumer = tx_ring->sw_consumer; in netxen_send_cmd_descs()
598 if (nr_desc >= netxen_tx_avail(tx_ring)) { in netxen_send_cmd_descs()
599 netif_tx_stop_queue(tx_ring->txq); in netxen_send_cmd_descs()
601 if (netxen_tx_avail(tx_ring) > nr_desc) { in netxen_send_cmd_descs()
602 if (netxen_tx_avail(tx_ring) > TX_STOP_THRESH) in netxen_send_cmd_descs()
603 netif_tx_wake_queue(tx_ring->txq); in netxen_send_cmd_descs()
[all …]
Dnetxen_nic.h81 #define TX_BUFF_RINGSIZE(tx_ring) \ argument
82 (sizeof(struct netxen_cmd_buffer) * tx_ring->num_desc)
83 #define TX_DESC_RINGSIZE(tx_ring) \ argument
84 (sizeof(struct cmd_desc_type0) * tx_ring->num_desc)
1657 struct nx_host_tx_ring *tx_ring; member
1806 struct nx_host_tx_ring *tx_ring);
1874 static inline u32 netxen_tx_avail(struct nx_host_tx_ring *tx_ring) in netxen_tx_avail() argument
1877 return find_diff_among(tx_ring->producer, in netxen_tx_avail()
1878 tx_ring->sw_consumer, tx_ring->num_desc); in netxen_tx_avail()
Dnetxen_nic_ethtool.c332 regs_buff[i++] = le32_to_cpu(*(adapter->tx_ring->hw_consumer)); in netxen_nic_get_regs()
344 adapter->tx_ring->crb_cmd_consumer); in netxen_nic_get_regs()
347 regs_buff[i++] = NXRDIO(adapter, adapter->tx_ring->crb_cmd_producer); in netxen_nic_get_regs()
/linux-4.1.27/drivers/net/ethernet/qlogic/qlge/
Dqlge_main.c2100 struct tx_ring *tx_ring; in ql_process_mac_tx_intr() local
2104 tx_ring = &qdev->tx_ring[mac_rsp->txq_idx]; in ql_process_mac_tx_intr()
2105 tx_ring_desc = &tx_ring->q[mac_rsp->tid]; in ql_process_mac_tx_intr()
2107 tx_ring->tx_bytes += (tx_ring_desc->skb)->len; in ql_process_mac_tx_intr()
2108 tx_ring->tx_packets++; in ql_process_mac_tx_intr()
2133 atomic_inc(&tx_ring->tx_count); in ql_process_mac_tx_intr()
2202 struct tx_ring *tx_ring; in ql_clean_outbound_rx_ring() local
2230 tx_ring = &qdev->tx_ring[net_rsp->txq_idx]; in ql_clean_outbound_rx_ring()
2231 if (__netif_subqueue_stopped(qdev->ndev, tx_ring->wq_id)) { in ql_clean_outbound_rx_ring()
2232 if ((atomic_read(&tx_ring->tx_count) > (tx_ring->wq_len / 4))) in ql_clean_outbound_rx_ring()
[all …]
Dqlge_dbg.c1644 DUMP_QDEV_FIELD(qdev, "%p", tx_ring); in ql_dump_qdev()
1669 void ql_dump_tx_ring(struct tx_ring *tx_ring) in ql_dump_tx_ring() argument
1671 if (tx_ring == NULL) in ql_dump_tx_ring()
1674 tx_ring->wq_id); in ql_dump_tx_ring()
1675 pr_err("tx_ring->base = %p\n", tx_ring->wq_base); in ql_dump_tx_ring()
1677 (unsigned long long) tx_ring->wq_base_dma); in ql_dump_tx_ring()
1679 tx_ring->cnsmr_idx_sh_reg, in ql_dump_tx_ring()
1680 tx_ring->cnsmr_idx_sh_reg in ql_dump_tx_ring()
1681 ? ql_read_sh_reg(tx_ring->cnsmr_idx_sh_reg) : 0); in ql_dump_tx_ring()
1682 pr_err("tx_ring->size = %d\n", tx_ring->wq_size); in ql_dump_tx_ring()
[all …]
Dqlge.h1382 struct tx_ring { struct
2116 struct tx_ring tx_ring[MAX_TX_RINGS]; member
2292 void ql_dump_tx_ring(struct tx_ring *tx_ring);
2299 #define QL_DUMP_TX_RING(tx_ring) ql_dump_tx_ring(tx_ring) argument
2307 #define QL_DUMP_TX_RING(tx_ring) argument
/linux-4.1.27/drivers/net/vmxnet3/
Dvmxnet3_drv.c336 BUG_ON(tq->buf_info[eop_idx].sop_idx != tq->tx_ring.next2comp); in vmxnet3_unmap_pkt()
337 BUG_ON(VMXNET3_TXDESC_GET_EOP(&(tq->tx_ring.base[eop_idx].txd)) != 1); in vmxnet3_unmap_pkt()
343 VMXNET3_INC_RING_IDX_ONLY(eop_idx, tq->tx_ring.size); in vmxnet3_unmap_pkt()
345 while (tq->tx_ring.next2comp != eop_idx) { in vmxnet3_unmap_pkt()
346 vmxnet3_unmap_tx_buf(tq->buf_info + tq->tx_ring.next2comp, in vmxnet3_unmap_pkt()
354 vmxnet3_cmd_ring_adv_next2comp(&tq->tx_ring); in vmxnet3_unmap_pkt()
383 vmxnet3_cmd_ring_desc_avail(&tq->tx_ring) > in vmxnet3_tq_tx_complete()
400 while (tq->tx_ring.next2comp != tq->tx_ring.next2fill) { in vmxnet3_tq_cleanup()
403 tbi = tq->buf_info + tq->tx_ring.next2comp; in vmxnet3_tq_cleanup()
410 vmxnet3_cmd_ring_adv_next2comp(&tq->tx_ring); in vmxnet3_tq_cleanup()
[all …]
Dvmxnet3_ethtool.c360 buf[j++] = adapter->tx_queue[i].tx_ring.next2fill; in vmxnet3_get_regs()
361 buf[j++] = adapter->tx_queue[i].tx_ring.next2comp; in vmxnet3_get_regs()
362 buf[j++] = adapter->tx_queue[i].tx_ring.gen; in vmxnet3_get_regs()
Dvmxnet3_int.h229 struct vmxnet3_cmd_ring tx_ring; member
/linux-4.1.27/drivers/net/ethernet/amd/
Dariadne.c88 volatile struct TDRE *tx_ring[TX_RING_SIZE]; member
100 struct TDRE tx_ring[TX_RING_SIZE]; member
130 volatile struct TDRE *t = &lancedata->tx_ring[i]; in ariadne_init_ring()
138 priv->tx_ring[i] = &lancedata->tx_ring[i]; in ariadne_init_ring()
141 i, &lancedata->tx_ring[i], lancedata->tx_buff[i]); in ariadne_init_ring()
310 int status = lowb(priv->tx_ring[entry]->TMD1); in ariadne_interrupt()
315 priv->tx_ring[entry]->TMD1 &= 0xff00; in ariadne_interrupt()
319 int err_status = priv->tx_ring[entry]->TMD3; in ariadne_interrupt()
456 lance->RDP = swloww(ARIADNE_RAM + offsetof(struct lancedata, tx_ring)); in ariadne_open()
458 lance->RDP = swhighw(ARIADNE_RAM + offsetof(struct lancedata, tx_ring)); in ariadne_open()
[all …]
Dlance.c232 u32 tx_ring; member
238 struct lance_tx_head tx_ring[TX_RING_SIZE]; member
576 lp->init_block.tx_ring = ((u32)isa_virt_to_bus(lp->tx_ring) & 0xffffff) | TX_RING_LEN_BITS; in lance_probe1()
789 (u32) isa_virt_to_bus(lp->tx_ring), in lance_open()
890 lp->tx_ring[i].base = 0; in lance_init_ring()
899 lp->init_block.tx_ring = ((u32)isa_virt_to_bus(lp->tx_ring) & 0xffffff) | TX_RING_LEN_BITS; in lance_init_ring()
939 lp->tx_ring[i].base, -lp->tx_ring[i].length, in lance_tx_timeout()
940 lp->tx_ring[i].misc); in lance_tx_timeout()
981 lp->tx_ring[entry].length = -ETH_ZLEN; in lance_start_xmit()
984 lp->tx_ring[entry].length = -skb->len; in lance_start_xmit()
[all …]
Dpcnet32.c241 __le32 tx_ring; member
263 struct pcnet32_tx_head *tx_ring; member
507 lp->tx_ring, lp->tx_ring_dma_addr); in pcnet32_realloc_tx_ring()
512 lp->tx_ring = new_tx_ring; in pcnet32_realloc_tx_ring()
922 lp->tx_ring[x].length = cpu_to_le16(-skb->len); in pcnet32_loopback_test()
923 lp->tx_ring[x].misc = 0; in pcnet32_loopback_test()
948 lp->tx_ring[x].base = cpu_to_le32(lp->tx_dma_addr[x]); in pcnet32_loopback_test()
950 lp->tx_ring[x].status = cpu_to_le16(status); in pcnet32_loopback_test()
1254 int status = (short)le16_to_cpu(lp->tx_ring[entry].status); in pcnet32_tx()
1259 lp->tx_ring[entry].base = 0; in pcnet32_tx()
[all …]
Damd8111e.c305 if((lp->tx_ring = pci_alloc_consistent(lp->pci_dev, in amd8111e_init_ring()
346 lp->tx_ring[i].buff_phy_addr = 0; in amd8111e_init_ring()
347 lp->tx_ring[i].tx_flags = 0; in amd8111e_init_ring()
348 lp->tx_ring[i].buff_count = 0; in amd8111e_init_ring()
362 sizeof(struct amd8111e_tx_dr)*NUM_TX_RING_DR,lp->tx_ring, in amd8111e_init_ring()
635 if(lp->tx_ring){ in amd8111e_free_ring()
638 lp->tx_ring, lp->tx_ring_dma_addr); in amd8111e_free_ring()
640 lp->tx_ring = NULL; in amd8111e_free_ring()
657 status = le16_to_cpu(lp->tx_ring[tx_index].tx_flags); in amd8111e_tx()
662 lp->tx_ring[tx_index].buff_phy_addr = 0; in amd8111e_tx()
[all …]
Datarilance.c151 struct ringdesc tx_ring; member
620 MEM->init.tx_ring.adr_lo = offsetof( struct lance_memory, tx_head ); in lance_probe1()
621 MEM->init.tx_ring.adr_hi = 0; in lance_probe1()
622 MEM->init.tx_ring.len = TX_RING_LEN_BITS; in lance_probe1()
Damd8111e.h747 struct amd8111e_tx_dr* tx_ring; member
/linux-4.1.27/drivers/infiniband/hw/mlx4/
Dmad.c536 tun_mad = (struct mlx4_rcv_tunnel_mad *) (tun_qp->tx_ring[tun_tx_ix].buf.addr); in mlx4_ib_send_to_slave()
537 if (tun_qp->tx_ring[tun_tx_ix].ah) in mlx4_ib_send_to_slave()
538 ib_destroy_ah(tun_qp->tx_ring[tun_tx_ix].ah); in mlx4_ib_send_to_slave()
539 tun_qp->tx_ring[tun_tx_ix].ah = ah; in mlx4_ib_send_to_slave()
541 tun_qp->tx_ring[tun_tx_ix].buf.map, in mlx4_ib_send_to_slave()
583 tun_qp->tx_ring[tun_tx_ix].buf.map, in mlx4_ib_send_to_slave()
587 list.addr = tun_qp->tx_ring[tun_tx_ix].buf.map; in mlx4_ib_send_to_slave()
1223 sqp_mad = (struct mlx4_mad_snd_buf *) (sqp->tx_ring[wire_tx_ix].buf.addr); in mlx4_ib_send_to_wire()
1224 if (sqp->tx_ring[wire_tx_ix].ah) in mlx4_ib_send_to_wire()
1225 ib_destroy_ah(sqp->tx_ring[wire_tx_ix].ah); in mlx4_ib_send_to_wire()
[all …]
Dmlx4_ib.h396 struct mlx4_ib_tun_tx_buf *tx_ring; member
/linux-4.1.27/drivers/net/ethernet/dlink/
Dsundance.c372 struct netdev_desc *tx_ring; member
562 np->tx_ring = (struct netdev_desc *)ring_space; in sundance_probe1()
704 np->tx_ring, np->tx_ring_dma); in sundance_probe1()
990 (unsigned long long)(np->tx_ring_dma + i*sizeof(*np->tx_ring)), in tx_timeout()
991 le32_to_cpu(np->tx_ring[i].next_desc), in tx_timeout()
992 le32_to_cpu(np->tx_ring[i].status), in tx_timeout()
993 (le32_to_cpu(np->tx_ring[i].status) >> 2) & 0xff, in tx_timeout()
994 le32_to_cpu(np->tx_ring[i].frag[0].addr), in tx_timeout()
995 le32_to_cpu(np->tx_ring[i].frag[0].length)); in tx_timeout()
1068 np->tx_ring[i].status = 0; in init_ring()
[all …]
Ddl2k.c239 np->tx_ring = ring_space; in rio_probe1()
297 pci_free_consistent (pdev, TX_TOTAL_SIZE, np->tx_ring, np->tx_ring_dma); in rio_probe1()
560 np->tx_ring[i].status = cpu_to_le64 (TFDDone); in alloc_list()
561 np->tx_ring[i].next_desc = cpu_to_le64 (np->tx_ring_dma + in alloc_list()
614 txdesc = &np->tx_ring[entry]; in start_xmit()
729 if (!(np->tx_ring[entry].status & cpu_to_le64(TFDDone))) in rio_free_tx()
733 desc_to_dma(&np->tx_ring[entry]), in rio_free_tx()
1718 pci_unmap_single(pdev, desc_to_dma(&np->tx_ring[i]), in rio_close()
1739 pci_free_consistent (pdev, TX_TOTAL_SIZE, np->tx_ring, in rio_remove1()
Ddl2k.h363 struct netdev_desc *tx_ring; member
/linux-4.1.27/drivers/net/ethernet/
Dfealnx.c378 struct fealnx_desc *tx_ring; member
574 np->tx_ring = ring_space; in fealnx_init_one()
674 pci_free_consistent(pdev, TX_TOTAL_SIZE, np->tx_ring, np->tx_ring_dma); in fealnx_init_one()
694 pci_free_consistent(pdev, TX_TOTAL_SIZE, np->tx_ring, in fealnx_remove_one()
1157 iowrite32(np->tx_ring_dma + ((char*)np->cur_tx - (char*)np->tx_ring), in enable_rxtx()
1216 printk(KERN_DEBUG " Tx ring %p: ", np->tx_ring); in fealnx_tx_timeout()
1218 printk(KERN_CONT " %4.4x", np->tx_ring[i].status); in fealnx_tx_timeout()
1280 np->cur_tx = &np->tx_ring[0]; in init_ring()
1281 np->cur_tx_copy = &np->tx_ring[0]; in init_ring()
1286 np->tx_ring[i].status = 0; in init_ring()
[all …]
/linux-4.1.27/drivers/net/irda/
Dau1k_ir.c154 volatile struct ring_dest *tx_ring[NUM_IR_DESC]; member
257 aup->tx_ring[i] = (volatile struct ring_dest *) in setup_hw_rings()
312 ptxd = aup->tx_ring[i]; in au1k_irda_set_speed()
422 ptxd = aup->tx_ring[aup->tx_tail]; in au1k_tx_ack()
430 ptxd = aup->tx_ring[aup->tx_tail]; in au1k_tx_ack()
680 ptxd = aup->tx_ring[aup->tx_head]; in au1k_irda_hard_xmit()
861 aup->tx_ring[i]->addr_0 = (u8)(pDB->dma_addr & 0xff); in au1k_irda_net_init()
862 aup->tx_ring[i]->addr_1 = (u8)((pDB->dma_addr >> 8) & 0xff); in au1k_irda_net_init()
863 aup->tx_ring[i]->addr_2 = (u8)((pDB->dma_addr >> 16) & 0xff); in au1k_irda_net_init()
864 aup->tx_ring[i]->addr_3 = (u8)((pDB->dma_addr >> 24) & 0xff); in au1k_irda_net_init()
[all …]
Dvlsi_ir.c357 if (netif_running(ndev) && idev->rx_ring && idev->tx_ring) { in vlsi_seq_show()
361 vlsi_proc_ring(seq, idev->tx_ring); in vlsi_seq_show()
488 idev->tx_ring = vlsi_alloc_ring(idev->pdev, hwmap, ringsize[0], in vlsi_create_hwif()
490 if (idev->tx_ring == NULL) in vlsi_create_hwif()
499 idev->rx_ring = idev->tx_ring = NULL; in vlsi_create_hwif()
509 vlsi_free_ring(idev->tx_ring); in vlsi_destroy_hwif()
510 idev->rx_ring = idev->tx_ring = NULL; in vlsi_destroy_hwif()
847 struct vlsi_ring *r = idev->tx_ring; in vlsi_hard_start_xmit()
891 if (ring_first(idev->tx_ring) == NULL) { in vlsi_hard_start_xmit()
1042 struct vlsi_ring *r = idev->tx_ring; in vlsi_tx_interrupt()
[all …]
Dvlsi_ir.h724 struct vlsi_ring *tx_ring, *rx_ring; member
/linux-4.1.27/drivers/net/ethernet/qualcomm/
Dqca_spi.h56 struct tx_ring { struct
83 struct tx_ring txr; argument
Dqca_spi.c400 qcaspi_tx_ring_has_space(struct tx_ring *txr) in qcaspi_tx_ring_has_space()
/linux-4.1.27/drivers/net/hippi/
Drrunner.c159 rrpriv->tx_ring = tmpptr; in rr_init_one()
210 if (rrpriv->tx_ring) in rr_init_one()
211 pci_free_consistent(pdev, TX_TOTAL_SIZE, rrpriv->tx_ring, in rr_init_one()
239 pci_free_consistent(pdev, TX_TOTAL_SIZE, rr->tx_ring, in rr_remove_one()
601 rrpriv->tx_ring[i].size = 0; in rr_init1()
602 set_rraddr(&rrpriv->tx_ring[i].addr, 0); in rr_init1()
1069 desc = &(rrpriv->tx_ring[txcon]); in rr_interrupt()
1082 set_rraddr(&rrpriv->tx_ring[txcon].addr, 0); in rr_interrupt()
1115 struct tx_desc *desc = &(rrpriv->tx_ring[i]); in rr_raz_tx()
1291 …i is valid - dumping data (0x%x bytes - DMA len 0x%x)\n", index, len, rrpriv->tx_ring[index].size); in rr_dump()
[all …]
Drrunner.h802 struct tx_desc *tx_ring; member
/linux-4.1.27/drivers/net/ethernet/smsc/
Dsmsc9420.c62 struct smsc9420_dma_desc *tx_ring; member
548 BUG_ON(!pd->tx_ring); in smsc9420_free_tx_ring()
563 pd->tx_ring[i].status = 0; in smsc9420_free_tx_ring()
564 pd->tx_ring[i].length = 0; in smsc9420_free_tx_ring()
565 pd->tx_ring[i].buffer1 = 0; in smsc9420_free_tx_ring()
566 pd->tx_ring[i].buffer2 = 0; in smsc9420_free_tx_ring()
951 status = pd->tx_ring[index].status; in smsc9420_complete_tx()
952 length = pd->tx_ring[index].length; in smsc9420_complete_tx()
970 pd->tx_ring[index].buffer1 = 0; in smsc9420_complete_tx()
990 BUG_ON(pd->tx_ring[index].status & TDES0_OWN_); in smsc9420_hard_start_xmit()
[all …]
Depic100.c254 struct epic_tx_desc *tx_ring; member
383 ep->tx_ring = ring_space; in epic_init_one()
502 pci_free_consistent(pdev, TX_TOTAL_SIZE, ep->tx_ring, ep->tx_ring_dma); in epic_init_one()
937 ep->tx_ring[i].txstatus = 0x0000; in epic_init_ring()
938 ep->tx_ring[i].next = ep->tx_ring_dma + in epic_init_ring()
941 ep->tx_ring[i-1].next = ep->tx_ring_dma; in epic_init_ring()
964 ep->tx_ring[entry].bufaddr = pci_map_single(ep->pci_dev, skb->data, in epic_start_xmit()
977 ep->tx_ring[entry].buflength = ctrl_word | skb->len; in epic_start_xmit()
978 ep->tx_ring[entry].txstatus = in epic_start_xmit()
1031 int txstatus = ep->tx_ring[entry].txstatus; in epic_tx()
[all …]
/linux-4.1.27/drivers/net/ethernet/samsung/sxgbe/
Dsxgbe_main.c395 struct sxgbe_tx_queue *tx_ring, int tx_rsize) in init_tx_ring() argument
398 if (!tx_ring) { in init_tx_ring()
404 tx_ring->dma_tx = dma_zalloc_coherent(dev, in init_tx_ring()
406 &tx_ring->dma_tx_phy, GFP_KERNEL); in init_tx_ring()
407 if (!tx_ring->dma_tx) in init_tx_ring()
411 tx_ring->tx_skbuff_dma = devm_kcalloc(dev, tx_rsize, in init_tx_ring()
413 if (!tx_ring->tx_skbuff_dma) in init_tx_ring()
416 tx_ring->tx_skbuff = devm_kcalloc(dev, tx_rsize, in init_tx_ring()
419 if (!tx_ring->tx_skbuff) in init_tx_ring()
423 tx_ring->queue_no = queue_no; in init_tx_ring()
[all …]
/linux-4.1.27/net/packet/
Daf_packet.c169 int closing, int tx_ring);
521 int tx_ring, in prb_shutdown_retire_blk_timer() argument
526 pkc = tx_ring ? GET_PBDQC_FROM_RB(&po->tx_ring) : in prb_shutdown_retire_blk_timer()
546 static void prb_setup_retire_blk_timer(struct packet_sock *po, int tx_ring) in prb_setup_retire_blk_timer() argument
550 if (tx_ring) in prb_setup_retire_blk_timer()
553 pkc = tx_ring ? GET_PBDQC_FROM_RB(&po->tx_ring) : in prb_setup_retire_blk_timer()
610 union tpacket_req_u *req_u, int tx_ring) in init_prb_bdqc() argument
637 prb_setup_retire_blk_timer(po, tx_ring); in init_prb_bdqc()
1225 po->tx_ring.pending_refcnt = alloc_percpu(unsigned int); in packet_alloc_pending()
1226 if (unlikely(po->tx_ring.pending_refcnt == NULL)) in packet_alloc_pending()
[all …]
Dinternal.h97 struct packet_ring_buffer tx_ring; member
Ddiag.c106 ret = pdiag_put_ring(&po->tx_ring, po->tp_version, in pdiag_put_rings_cfg()
/linux-4.1.27/drivers/net/ethernet/3com/
D3c59x.c598 struct boom_tx_desc* tx_ring; member
1226 vp->tx_ring = (struct boom_tx_desc *)(vp->rx_ring + RX_RING_SIZE); in vortex_probe1()
2129 struct boom_tx_desc *prev_entry = &vp->tx_ring[(vp->cur_tx-1) % TX_RING_SIZE]; in boomerang_start_xmit()
2158 vp->tx_ring[entry].next = 0; in boomerang_start_xmit()
2161 vp->tx_ring[entry].status = cpu_to_le32(skb->len | TxIntrUploaded); in boomerang_start_xmit()
2163 vp->tx_ring[entry].status = cpu_to_le32(skb->len | TxIntrUploaded | AddTCPChksum | AddUDPChksum); in boomerang_start_xmit()
2171 vp->tx_ring[entry].frag[0].addr = cpu_to_le32(dma_addr); in boomerang_start_xmit()
2172 vp->tx_ring[entry].frag[0].length = cpu_to_le32(skb->len | LAST_FRAG); in boomerang_start_xmit()
2181 vp->tx_ring[entry].frag[0].addr = cpu_to_le32(dma_addr); in boomerang_start_xmit()
2182 vp->tx_ring[entry].frag[0].length = cpu_to_le32(skb_headlen(skb)); in boomerang_start_xmit()
[all …]
D3c515.c306 struct boom_tx_desc tx_ring[TX_RING_SIZE]; member
982 &vp->tx_ring[0]); in corkscrew_timeout()
985 &vp->tx_ring[i], in corkscrew_timeout()
986 vp->tx_ring[i].length, vp->tx_ring[i].status); in corkscrew_timeout()
1021 prev_entry = &vp->tx_ring[(vp->cur_tx - 1) % TX_RING_SIZE]; in corkscrew_start_xmit()
1029 vp->tx_ring[entry].next = 0; in corkscrew_start_xmit()
1030 vp->tx_ring[entry].addr = isa_virt_to_bus(skb->data); in corkscrew_start_xmit()
1031 vp->tx_ring[entry].length = skb->len | 0x80000000; in corkscrew_start_xmit()
1032 vp->tx_ring[entry].status = skb->len | 0x80000000; in corkscrew_start_xmit()
1041 prev_entry->next = isa_virt_to_bus(&vp->tx_ring[entry]); in corkscrew_start_xmit()
[all …]
/linux-4.1.27/drivers/net/ethernet/freescale/fs_enet/
Dfs_enet-main.c239 if (fep->tx_free == fep->tx_ring) in fs_enet_tx_napi()
402 fep->tx_free = fep->tx_ring; in fs_init_bds()
434 for (i = 0, bdp = fep->tx_bd_base; i < fep->tx_ring; i++, bdp++) { in fs_init_bds()
438 CBDW_SC(bdp, (i < fep->tx_ring - 1) ? 0 : BD_SC_WRAP); in fs_init_bds()
452 for (i = 0, bdp = fep->tx_bd_base; i < fep->tx_ring; i++, bdp++) { in fs_cleanup_bds()
950 fpi->tx_ring = 64; in fs_enet_probe()
988 (fpi->rx_ring + fpi->tx_ring) + in fs_enet_probe()
989 sizeof(char) * fpi->tx_ring; in fs_enet_probe()
1013 fpi->tx_ring); in fs_enet_probe()
1029 fep->tx_ring = fpi->tx_ring; in fs_enet_probe()
Dmac-fcc.c155 (fpi->tx_ring + fpi->rx_ring) * in allocate_bd()
171 (fpi->tx_ring + fpi->rx_ring) * sizeof(cbd_t), in free_bd()
555 last_tx_bd = fep->tx_bd_base + (fpi->tx_ring * sizeof(cbd_t)); in tx_restart()
Dmac-fec.c144 (fpi->tx_ring + fpi->rx_ring) * in allocate_bd()
159 dma_free_coherent(fep->dev, (fpi->tx_ring + fpi->rx_ring) in free_bd()
Dfs_enet.h132 int rx_ring, tx_ring; member
Dmac-scc.c147 fep->ring_mem_addr = cpm_dpalloc((fpi->tx_ring + fpi->rx_ring) * in allocate_bd()
/linux-4.1.27/drivers/net/ethernet/dec/tulip/
Dwinbond-840.c297 struct w840_tx_desc *tx_ring; member
801 np->tx_ring = (struct w840_tx_desc*)&np->rx_ring[RX_RING_SIZE]; in init_rxtx_rings()
831 np->tx_ring[i].status = 0; in init_rxtx_rings()
942 printk(KERN_DEBUG " Tx ring %p: ", np->tx_ring); in tx_timeout()
944 printk(KERN_CONT " %08x", np->tx_ring[i].status); in tx_timeout()
1014 np->tx_ring[entry].buffer1 = np->tx_addr[entry]; in start_tx()
1016 np->tx_ring[entry].length = DescWholePkt | skb->len; in start_tx()
1020 np->tx_ring[entry].buffer2 = np->tx_addr[entry]+TX_BUFLIMIT; in start_tx()
1021 np->tx_ring[entry].length = DescWholePkt | (len << 11) | TX_BUFLIMIT; in start_tx()
1024 np->tx_ring[entry].length |= DescEndRing; in start_tx()
[all …]
Dtulip_core.c372 tp->tx_ring[tp->cur_tx].length = cpu_to_le32(0x08000000 | 192); in tulip_up()
373 tp->tx_ring[tp->cur_tx].buffer1 = cpu_to_le32(mapping); in tulip_up()
374 tp->tx_ring[tp->cur_tx].status = cpu_to_le32(DescOwned); in tulip_up()
600 printk(KERN_DEBUG " Tx ring %p: ", tp->tx_ring); in tulip_tx_timeout()
602 pr_cont(" %08x", (unsigned int)tp->tx_ring[i].status); in tulip_tx_timeout()
660 tp->tx_ring[i].status = 0x00000000; in tulip_init_ring()
661 tp->tx_ring[i].buffer2 = cpu_to_le32(tp->tx_ring_dma + sizeof(struct tulip_tx_desc) * (i + 1)); in tulip_init_ring()
663 tp->tx_ring[i-1].buffer2 = cpu_to_le32(tp->tx_ring_dma); in tulip_init_ring()
684 tp->tx_ring[entry].buffer1 = cpu_to_le32(mapping); in tulip_start_xmit()
699 tp->tx_ring[entry].length = cpu_to_le32(skb->len | flag); in tulip_start_xmit()
[all …]
Dde2104x.c307 struct de_desc *tx_ring; member
552 status = le32_to_cpu(de->tx_ring[tx_tail].opts1); in de_tx()
625 txd = &de->tx_ring[entry]; in de_start_xmit()
759 dummy_txd = &de->tx_ring[entry]; in __de_set_rx_mode()
775 txd = &de->tx_ring[entry]; in __de_set_rx_mode()
1310 memset(de->tx_ring, 0, sizeof(struct de_desc) * DE_TX_RING_SIZE); in de_init_rings()
1311 de->tx_ring[DE_TX_RING_SIZE - 1].opts2 = cpu_to_le32(RingEnd); in de_init_rings()
1324 de->tx_ring = &de->rx_ring[DE_RX_RING_SIZE]; in de_alloc_rings()
1335 memset(de->tx_ring, 0, sizeof(struct de_desc) * DE_TX_RING_SIZE); in de_clean_rings()
1336 de->tx_ring[DE_TX_RING_SIZE - 1].opts2 = cpu_to_le32(RingEnd); in de_clean_rings()
[all …]
Dde4x5.c791 struct de4x5_desc *tx_ring; /* TX descriptor ring */ member
1181 lp->tx_ring = lp->rx_ring + NUM_RX_DESC; in de4x5_hw_init()
1224 lp->tx_ring[lp->txRingSize - 1].des1 |= cpu_to_le32(TD_TER); in de4x5_hw_init()
1425 lp->tx_ring[i].status = cpu_to_le32(0); in de4x5_sw_reset()
1440 if ((s32)le32_to_cpu(lp->tx_ring[lp->tx_new].status) >= 0) j=1; in de4x5_sw_reset()
1679 dma_unmap_single(lp->gendev, le32_to_cpu(lp->tx_ring[entry].buf), in de4x5_free_tx_buff()
1680 le32_to_cpu(lp->tx_ring[entry].des1) & TD_TBS1, in de4x5_free_tx_buff()
1699 status = (s32)le32_to_cpu(lp->tx_ring[entry].status); in de4x5_tx()
1901 lp->tx_ring[lp->tx_new].buf = cpu_to_le32(buf_dma); in load_packet()
1902 lp->tx_ring[lp->tx_new].des1 &= cpu_to_le32(TD_TER); in load_packet()
[all …]
Dtulip.h405 struct tulip_tx_desc *tx_ring; member
Dinterrupt.c591 int status = le32_to_cpu(tp->tx_ring[entry].status); in tulip_interrupt()
/linux-4.1.27/drivers/net/ethernet/micrel/
Dks8695net.c175 struct tx_ring_desc *tx_ring; member
379 !(ksp->tx_ring[buff_n].owner & cpu_to_le32(TDES_OWN))) { in ks8695_tx_irq()
387 ksp->tx_ring[buff_n].data_ptr = 0; in ks8695_tx_irq()
687 ksp->tx_ring[buff_n].owner = 0; in ks8695_shutdown()
688 ksp->tx_ring[buff_n].status = 0; in ks8695_shutdown()
689 ksp->tx_ring[buff_n].data_ptr = 0; in ks8695_shutdown()
1193 ksp->tx_ring[buff_n].data_ptr = in ks8695_start_xmit()
1195 ksp->tx_ring[buff_n].status = in ks8695_start_xmit()
1202 ksp->tx_ring[buff_n].owner = cpu_to_le32(TDES_OWN); in ks8695_start_xmit()
1473 ksp->tx_ring = ksp->ring_base; in ks8695_probe()
[all …]
/linux-4.1.27/drivers/net/ethernet/sgi/
Dmeth.c78 tx_packet *tx_ring; member
214 priv->tx_ring = dma_zalloc_coherent(NULL, TX_RING_BUFFER_SIZE, in meth_init_tx_ring()
216 if (!priv->tx_ring) in meth_init_tx_ring()
256 dma_free_coherent(NULL, TX_RING_BUFFER_SIZE, priv->tx_ring, in meth_free_tx_ring()
492 status = priv->tx_ring[priv->tx_read].header.raw; in meth_tx_cleanup()
526 priv->tx_ring[priv->tx_read].header.raw = 0; in meth_tx_cleanup()
611 tx_packet *desc = &priv->tx_ring[priv->tx_write]; in meth_tx_short_prepare()
624 tx_packet *desc = &priv->tx_ring[priv->tx_write]; in meth_tx_1page_prepare()
649 tx_packet *desc = &priv->tx_ring[priv->tx_write]; in meth_tx_2page_prepare()
/linux-4.1.27/drivers/net/ethernet/amd/xgbe/
Dxgbe-drv.c136 struct xgbe_ring *tx_ring, *rx_ring; in xgbe_alloc_channels() local
146 tx_ring = kcalloc(pdata->tx_ring_count, sizeof(struct xgbe_ring), in xgbe_alloc_channels()
148 if (!tx_ring) in xgbe_alloc_channels()
177 spin_lock_init(&tx_ring->lock); in xgbe_alloc_channels()
178 channel->tx_ring = tx_ring++; in xgbe_alloc_channels()
188 channel->dma_irq, channel->tx_ring, channel->rx_ring); in xgbe_alloc_channels()
200 kfree(tx_ring); in xgbe_alloc_channels()
215 kfree(pdata->channel->tx_ring); in xgbe_free_channels()
281 if (channel->tx_ring && channel->rx_ring) in xgbe_enable_rx_tx_ints()
283 else if (channel->tx_ring) in xgbe_enable_rx_tx_ints()
[all …]
Dxgbe-desc.c181 xgbe_free_ring(pdata, channel->tx_ring); in xgbe_free_ring_resources()
230 ret = xgbe_init_ring(pdata, channel->tx_ring, in xgbe_alloc_ring_resources()
362 ring = channel->tx_ring; in xgbe_wrapper_tx_descriptor_init()
490 struct xgbe_ring *ring = channel->tx_ring; in xgbe_map_tx_skb()
Dxgbe-dev.c197 if (!channel->tx_ring) in xgbe_config_tx_pbl_val()
236 if (!channel->tx_ring) in xgbe_config_osp_mode()
332 if (!channel->tx_ring) in xgbe_config_tso_mode()
622 if (channel->tx_ring) { in xgbe_enable_dma_interrupts()
1092 struct xgbe_ring *ring = channel->tx_ring; in xgbe_tx_desc_init()
1403 struct xgbe_ring *ring = channel->tx_ring; in xgbe_dev_xmit()
2601 if (!channel->tx_ring) in xgbe_enable_tx()
2624 if (!channel->tx_ring) in xgbe_disable_tx()
2640 if (!channel->tx_ring) in xgbe_disable_tx()
2706 if (!channel->tx_ring) in xgbe_powerup_tx()
[all …]
Dxgbe.h413 struct xgbe_ring *tx_ring; member
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/
Den_netdev.c1299 i, priv->tx_ring[i]->qpn, priv->tx_ring[i]->cqn, in mlx4_en_tx_timeout()
1300 priv->tx_ring[i]->cons, priv->tx_ring[i]->prod); in mlx4_en_tx_timeout()
1523 struct mlx4_en_tx_ring *tx_ring; in mlx4_en_start_port() local
1629 tx_ring = priv->tx_ring[i]; in mlx4_en_start_port()
1630 err = mlx4_en_activate_tx_ring(priv, tx_ring, cq->mcq.cqn, in mlx4_en_start_port()
1637 tx_ring->tx_queue = netdev_get_tx_queue(dev, i); in mlx4_en_start_port()
1643 for (j = 0; j < tx_ring->buf_size; j += STAMP_STRIDE) in mlx4_en_start_port()
1644 *((u32 *) (tx_ring->buf + j)) = 0xffffffff; in mlx4_en_start_port()
1710 mlx4_en_deactivate_tx_ring(priv, priv->tx_ring[tx_index]); in mlx4_en_start_port()
1822 mlx4_en_deactivate_tx_ring(priv, priv->tx_ring[i]); in mlx4_en_stop_port()
[all …]
Den_ethtool.c396 data[index++] = priv->tx_ring[i]->packets; in mlx4_en_get_ethtool_stats()
397 data[index++] = priv->tx_ring[i]->bytes; in mlx4_en_get_ethtool_stats()
1043 tx_size == priv->tx_ring[0]->size) in mlx4_en_set_ringparam()
1085 param->tx_pending = priv->tx_ring[0]->size; in mlx4_en_get_ringparam()
1794 bf_supported &= priv->tx_ring[i]->bf_alloced; in mlx4_en_set_priv_flags()
1807 priv->tx_ring[i]->bf_enabled = bf_enabled_new; in mlx4_en_set_priv_flags()
Den_port.c195 const struct mlx4_en_tx_ring *ring = priv->tx_ring[i]; in mlx4_en_DUMP_ETH_STATS()
Den_tx.c390 struct mlx4_en_tx_ring *ring = priv->tx_ring[cq->ring]; in mlx4_en_process_tx_cq()
734 ring = priv->tx_ring[tx_ind]; in mlx4_en_xmit()
Dmlx4_en.h554 struct mlx4_en_tx_ring **tx_ring; member
/linux-4.1.27/drivers/net/ethernet/via/
Dvia-rhine.c444 struct tx_desc *tx_ring; member
1190 rp->tx_ring = ring + RX_RING_SIZE * sizeof(struct rx_desc); in alloc_ring()
1206 rp->tx_ring = NULL; in free_ring()
1291 rp->tx_ring[i].tx_status = 0; in alloc_tbufs()
1292 rp->tx_ring[i].desc_length = cpu_to_le32(TXDESC); in alloc_tbufs()
1294 rp->tx_ring[i].next_desc = cpu_to_le32(next); in alloc_tbufs()
1298 rp->tx_ring[i-1].next_desc = cpu_to_le32(rp->tx_ring_dma); in alloc_tbufs()
1310 rp->tx_ring[i].tx_status = 0; in free_tbufs()
1311 rp->tx_ring[i].desc_length = cpu_to_le32(TXDESC); in free_tbufs()
1312 rp->tx_ring[i].addr = cpu_to_le32(0xBADF00D0); /* An invalid address. */ in free_tbufs()
[all …]
/linux-4.1.27/drivers/net/xen-netback/
Dxenbus.c56 struct xen_netif_tx_back_ring *tx_ring = &queue->tx; in xenvif_read_io_ring() local
60 if (tx_ring->sring) { in xenvif_read_io_ring()
61 struct xen_netif_tx_sring *sring = tx_ring->sring; in xenvif_read_io_ring()
64 tx_ring->nr_ents); in xenvif_read_io_ring()
68 tx_ring->req_cons, in xenvif_read_io_ring()
69 tx_ring->req_cons - sring->rsp_prod, in xenvif_read_io_ring()
74 tx_ring->rsp_prod_pvt, in xenvif_read_io_ring()
75 tx_ring->rsp_prod_pvt - sring->rsp_prod, in xenvif_read_io_ring()
/linux-4.1.27/drivers/net/wireless/rtlwifi/
Dpci.c532 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in _rtl_pci_tx_chk_waitq()
567 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[prio]; in _rtl_pci_tx_isr()
1125 ring = &rtlpci->tx_ring[BEACON_QUEUE]; in _rtl_pci_prepare_bcn_tasklet()
1264 rtlpci->tx_ring[prio].buffer_desc = buffer_desc; in _rtl_pci_init_tx_ring()
1265 rtlpci->tx_ring[prio].buffer_desc_dma = buffer_desc_dma; in _rtl_pci_init_tx_ring()
1267 rtlpci->tx_ring[prio].cur_tx_rp = 0; in _rtl_pci_init_tx_ring()
1268 rtlpci->tx_ring[prio].cur_tx_wp = 0; in _rtl_pci_init_tx_ring()
1269 rtlpci->tx_ring[prio].avl_desc = entries; in _rtl_pci_init_tx_ring()
1282 rtlpci->tx_ring[prio].desc = desc; in _rtl_pci_init_tx_ring()
1283 rtlpci->tx_ring[prio].dma = desc_dma; in _rtl_pci_init_tx_ring()
[all …]
Dpci.h210 struct rtl8192_tx_ring tx_ring[RTL_PCI_MAX_TX_QUEUE_COUNT]; member
/linux-4.1.27/drivers/net/ethernet/apple/
Dmacmace.c62 unsigned char *tx_ring; member
389 mp->tx_ring = dma_alloc_coherent(mp->device, in mace_open()
392 if (mp->tx_ring == NULL) in mace_open()
421 mp->tx_ring, mp->tx_ring_phys); in mace_open()
469 skb_copy_from_linear_data(skb, mp->tx_ring, skb->len); in mace_xmit_start()
759 mp->tx_ring, mp->tx_ring_phys); in mac_mace_device_remove()
/linux-4.1.27/drivers/net/ethernet/sis/
Dsis900.c185 BufferDesc *tx_ring; member
486 sis_priv->tx_ring = ring_space; in sis900_probe()
575 pci_free_consistent(pci_dev, TX_TOTAL_SIZE, sis_priv->tx_ring, in sis900_probe()
1136 sis_priv->tx_ring[i].link = sis_priv->tx_ring_dma + in sis900_init_tx_ring()
1138 sis_priv->tx_ring[i].cmdsts = 0; in sis900_init_tx_ring()
1139 sis_priv->tx_ring[i].bufptr = 0; in sis900_init_tx_ring()
1564 sis_priv->tx_ring[i].bufptr, skb->len, in sis900_tx_timeout()
1568 sis_priv->tx_ring[i].cmdsts = 0; in sis900_tx_timeout()
1569 sis_priv->tx_ring[i].bufptr = 0; in sis900_tx_timeout()
1614 sis_priv->tx_ring[entry].bufptr = pci_map_single(sis_priv->pci_dev, in sis900_start_xmit()
[all …]
/linux-4.1.27/drivers/dma/
Dxgene-dma.c326 struct xgene_dma_ring tx_ring; member
467 xgene_dma_init_desc(desc1, chan->tx_ring.dst_ring_num); in xgene_dma_prep_cpy_desc()
521 xgene_dma_init_desc(desc1, chan->tx_ring.dst_ring_num); in xgene_dma_prep_xor_desc()
771 ret = xgene_chan_xfer_request(&chan->tx_ring, desc_sw); in xgene_chan_xfer_ld_pending()
1446 xgene_dma_delete_ring_one(&chan->tx_ring); in xgene_dma_delete_chan_rings()
1481 struct xgene_dma_ring *tx_ring = &chan->tx_ring; in xgene_dma_create_chan_rings() local
1497 tx_ring->owner = XGENE_DMA_RING_OWNER_DMA; in xgene_dma_create_chan_rings()
1498 tx_ring->buf_num = XGENE_DMA_BUFNUM + chan->id; in xgene_dma_create_chan_rings()
1500 ret = xgene_dma_create_ring_one(chan, tx_ring, in xgene_dma_create_chan_rings()
1507 tx_ring->dst_ring_num = XGENE_DMA_RING_DST_ID(rx_ring->num); in xgene_dma_create_chan_rings()
[all …]
/linux-4.1.27/drivers/net/wireless/rtl818x/rtl8180/
Ddev.c345 struct rtl8180_tx_ring *ring = &priv->tx_ring[prio]; in rtl8180_handle_tx()
474 ring = &priv->tx_ring[prio]; in rtl8180_tx()
862 priv->tx_ring[1].dma); in rtl8180_init_hw()
864 priv->tx_ring[0].dma); in rtl8180_init_hw()
867 priv->tx_ring[4].dma); in rtl8180_init_hw()
869 priv->tx_ring[0].dma); in rtl8180_init_hw()
871 priv->tx_ring[1].dma); in rtl8180_init_hw()
873 priv->tx_ring[2].dma); in rtl8180_init_hw()
875 priv->tx_ring[3].dma); in rtl8180_init_hw()
1080 priv->tx_ring[prio].desc = ring; in rtl8180_init_tx_ring()
[all …]
Drtl8180.h117 struct rtl8180_tx_ring tx_ring[RTL818X_NR_TX_QUEUES]; member
/linux-4.1.27/drivers/net/ethernet/rdc/
Dr6040.c194 struct r6040_descriptor *tx_ring; member
330 lp->tx_remove_ptr = lp->tx_insert_ptr = lp->tx_ring; in r6040_init_txbufs()
331 r6040_init_ring_desc(lp->tx_ring, lp->tx_ring_dma, TX_DCNT); in r6040_init_txbufs()
507 if (lp->tx_ring) { in r6040_close()
509 TX_DESC_SIZE, lp->tx_ring, lp->tx_ring_dma); in r6040_close()
510 lp->tx_ring = NULL; in r6040_close()
780 lp->tx_ring = in r6040_open()
782 if (!lp->tx_ring) { in r6040_open()
797 pci_free_consistent(lp->pdev, TX_DESC_SIZE, lp->tx_ring, in r6040_open()
/linux-4.1.27/drivers/net/ethernet/broadcom/
Db44.c1028 bp->tx_ring[entry].ctrl = cpu_to_le32(ctrl); in b44_start_xmit()
1029 bp->tx_ring[entry].addr = cpu_to_le32((u32) mapping+bp->dma_offset); in b44_start_xmit()
1033 entry * sizeof(bp->tx_ring[0]), in b44_start_xmit()
1139 memset(bp->tx_ring, 0, B44_TX_RING_BYTES); in b44_init_rings()
1176 if (bp->tx_ring) { in b44_free_consistent()
1180 kfree(bp->tx_ring); in b44_free_consistent()
1183 bp->tx_ring, bp->tx_ring_dma); in b44_free_consistent()
1184 bp->tx_ring = NULL; in b44_free_consistent()
1236 bp->tx_ring = dma_alloc_coherent(bp->sdev->dma_dev, size, in b44_alloc_consistent()
1238 if (!bp->tx_ring) { in b44_alloc_consistent()
[all …]
Dbgmac.c583 bgmac_dma_tx_ring_free(bgmac, &bgmac->tx_ring[i]); in bgmac_dma_cleanup()
594 bgmac_dma_ring_desc_free(bgmac, &bgmac->tx_ring[i], in bgmac_dma_free()
621 ring = &bgmac->tx_ring[i]; in bgmac_dma_alloc()
682 ring = &bgmac->tx_ring[i]; in bgmac_dma_init()
1024 bgmac_dma_tx_reset(bgmac, &bgmac->tx_ring[i]); in bgmac_chip_reset()
1258 bgmac_dma_tx_free(bgmac, &bgmac->tx_ring[0]); in bgmac_poll()
1331 ring = &bgmac->tx_ring[0]; in bgmac_start_xmit()
Db44.h358 struct dma_desc *rx_ring, *tx_ring; member
Dbgmac.h446 struct bgmac_dma_ring tx_ring[BGMAC_MAX_TX_RINGS]; member
Dbnx2.c700 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_free_tx_mem()
752 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_alloc_tx_mem()
2830 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_tx_int()
3378 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_has_fast_work()
3472 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_poll_work()
5108 txr = &bnapi->tx_ring; in bnx2_clear_ring_states()
5160 txr = &bnapi->tx_ring; in bnx2_init_tx_ring()
5407 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_free_tx_skbs()
5788 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_run_loopback()
5793 txr = &tx_napi->tx_ring; in bnx2_run_loopback()
[all …]
Dtg3.c6542 if (tnapi->tx_ring[sw_idx].len_flags & TXD_FLAG_HWTSTAMP) { in tg3_tx()
7714 tg3_tx_set_bd(&tnapi->tx_ring[*entry], map, in tg3_tx_frag_set()
7725 tg3_tx_set_bd(&tnapi->tx_ring[*entry], map, in tg3_tx_frag_set()
7735 tg3_tx_set_bd(&tnapi->tx_ring[*entry], map, in tg3_tx_frag_set()
8576 if (tnapi->tx_ring) in tg3_init_rings()
8577 memset(tnapi->tx_ring, 0, TG3_TX_RING_BYTES); in tg3_init_rings()
8600 if (tnapi->tx_ring) { in tg3_mem_tx_release()
8602 tnapi->tx_ring, tnapi->tx_desc_mapping); in tg3_mem_tx_release()
8603 tnapi->tx_ring = NULL; in tg3_mem_tx_release()
8628 tnapi->tx_ring = dma_alloc_coherent(&tp->pdev->dev, in tg3_mem_tx_acquire()
[all …]
/linux-4.1.27/drivers/net/ethernet/nvidia/
Dforcedeth.c830 union ring_type tx_ring; member
1932 np->get_tx = np->put_tx = np->first_tx = np->tx_ring; in nv_init_tx()
1935 np->last_tx.orig = &np->tx_ring.orig[np->tx_ring_size-1]; in nv_init_tx()
1937 np->last_tx.ex = &np->tx_ring.ex[np->tx_ring_size-1]; in nv_init_tx()
1948 np->tx_ring.orig[i].flaglen = 0; in nv_init_tx()
1949 np->tx_ring.orig[i].buf = 0; in nv_init_tx()
1951 np->tx_ring.ex[i].flaglen = 0; in nv_init_tx()
1952 np->tx_ring.ex[i].txvlan = 0; in nv_init_tx()
1953 np->tx_ring.ex[i].bufhigh = 0; in nv_init_tx()
1954 np->tx_ring.ex[i].buflow = 0; in nv_init_tx()
[all …]
/linux-4.1.27/drivers/net/ethernet/marvell/
Dskge.c502 p->tx_pending = skge->tx_ring.count; in skge_get_ring_param()
516 skge->tx_ring.count = p->tx_pending; in skge_set_ring_param()
2546 tx_size = skge->tx_ring.count * sizeof(struct skge_tx_desc); in skge_up()
2570 err = skge_ring_alloc(&skge->tx_ring, skge->mem + rx_size, in skge_up()
2601 BUG_ON(skge->tx_ring.to_use != skge->tx_ring.to_clean); in skge_up()
2603 skge_qset(skge, txqaddr[port], skge->tx_ring.to_use); in skge_up()
2623 kfree(skge->tx_ring.start); in skge_up()
2717 kfree(skge->tx_ring.start); in skge_down()
2744 if (unlikely(skge_avail(&skge->tx_ring) < skb_shinfo(skb)->nr_frags + 1)) in skge_xmit_frame()
2747 e = skge->tx_ring.to_use; in skge_xmit_frame()
[all …]
Dsky2.c1617 sky2->tx_ring = kcalloc(sky2->tx_ring_size, sizeof(struct tx_ring_info), in sky2_alloc_buffers()
1619 if (!sky2->tx_ring) in sky2_alloc_buffers()
1654 kfree(sky2->tx_ring); in sky2_free_buffers()
1657 sky2->tx_ring = NULL; in sky2_free_buffers()
1937 re = sky2->tx_ring + slot; in sky2_xmit_frame()
1966 re = sky2->tx_ring + slot; in sky2_xmit_frame()
1993 re = sky2->tx_ring + i; in sky2_xmit_frame()
2025 struct tx_ring_info *re = sky2->tx_ring + idx; in sky2_tx_complete()
Dsky2.h2226 struct tx_ring_info *tx_ring; member
Dskge.h2456 struct skge_ring tx_ring; member
/linux-4.1.27/drivers/net/ethernet/realtek/
D8139cp.c342 struct cp_desc *tx_ring; member
655 struct cp_desc *txd = cp->tx_ring + tx_tail; in cp_tx()
724 txd = &cp->tx_ring[index]; in unwind_tx_frag_mapping()
758 struct cp_desc *txd = &cp->tx_ring[entry]; in cp_start_xmit()
844 txd = &cp->tx_ring[entry]; in cp_start_xmit()
856 txd = &cp->tx_ring[first_entry]; in cp_start_xmit()
1116 memset(cp->tx_ring, 0, sizeof(struct cp_desc) * CP_TX_RING_SIZE); in cp_init_rings()
1117 cp->tx_ring[CP_TX_RING_SIZE - 1].opts1 = cpu_to_le32(RingEnd); in cp_init_rings()
1135 cp->tx_ring = &cp->rx_ring[CP_RX_RING_SIZE]; in cp_alloc_rings()
1162 desc = cp->tx_ring + i; in cp_clean_rings()
[all …]
/linux-4.1.27/drivers/staging/rtl8192e/rtl8192e/
Drtl_core.c445 struct rtl8192_tx_ring *ring = &priv->tx_ring[prio]; in rtl8192_get_nic_desc_num()
460 struct rtl8192_tx_ring *ring = &priv->tx_ring[prio]; in rtl8192_check_nic_enough_desc()
724 ring = &priv->tx_ring[BEACON_QUEUE]; in rtl8192_prepare_beacon()
1349 if (skb_queue_len(&(&priv->tx_ring[i])->queue) > 0) { in rtl8192_is_tx_queue_empty()
1351 i, skb_queue_len(&(&priv->tx_ring[i])->queue)); in rtl8192_is_tx_queue_empty()
1388 ring = &priv->tx_ring[QueueID]; in rtl819x_TxCheckStuck()
1846 struct rtl8192_tx_ring *ring = &priv->tx_ring[prio]; in rtl8192_free_tx_ring()
1940 struct rtl8192_tx_ring *ring = &priv->tx_ring[prio]; in rtl8192_tx_isr()
1972 ring = &priv->tx_ring[TXCMD_QUEUE]; in rtl8192_tx_cmd()
2032 ring = &priv->tx_ring[tcb_desc->queue_index]; in rtl8192_tx()
[all …]
Drtl_core.h472 struct tx_ring { struct
475 struct tx_ring *next; argument
631 struct rtl8192_tx_ring tx_ring[MAX_TX_QUEUE_COUNT]; member
Dr8192E_phy.c1494 ring = &priv->tx_ring[QueueID]; in SetRFPowerState8190()
1523 ring = &priv->tx_ring[QueueID]; in SetRFPowerState8190()
/linux-4.1.27/net/netlink/
Daf_netlink.c308 return nlk_sk(sk)->tx_ring.pg_vec != NULL; in netlink_tx_is_mmaped()
377 __netlink_set_ring(struct sock *sk, struct nl_mmap_req *req, bool tx_ring, void **pg_vec, in __netlink_set_ring() argument
384 queue = tx_ring ? &sk->sk_write_queue : &sk->sk_receive_queue; in __netlink_set_ring()
385 ring = tx_ring ? &nlk->tx_ring : &nlk->rx_ring; in __netlink_set_ring()
408 bool tx_ring) in netlink_set_ring() argument
415 ring = tx_ring ? &nlk->tx_ring : &nlk->rx_ring; in netlink_set_ring()
454 __netlink_set_ring(sk, req, tx_ring, pg_vec, order); in netlink_set_ring()
508 for (ring = &nlk->rx_ring; ring <= &nlk->tx_ring; ring++) { in netlink_mmap()
522 for (ring = &nlk->rx_ring; ring <= &nlk->tx_ring; ring++) { in netlink_mmap()
695 if (nlk->tx_ring.pg_vec) { in netlink_poll()
[all …]
Daf_netlink.h50 struct netlink_ring tx_ring; member
Ddiag.c33 ret = sk_diag_put_ring(&nlk->tx_ring, NETLINK_DIAG_TX_RING, in sk_diag_put_rings_cfg()
/linux-4.1.27/drivers/infiniband/ulp/ipoib/
Dipoib_cm.c736 tx_req = &tx->tx_ring[tx->tx_head & (ipoib_sendq_size - 1)]; in ipoib_cm_send()
792 tx_req = &tx->tx_ring[wr_id]; in ipoib_cm_handle_tx_wc()
1117 p->tx_ring = __vmalloc(ipoib_sendq_size * sizeof *p->tx_ring, in ipoib_cm_tx_init()
1119 if (!p->tx_ring) { in ipoib_cm_tx_init()
1124 memset(p->tx_ring, 0, ipoib_sendq_size * sizeof *p->tx_ring); in ipoib_cm_tx_init()
1165 vfree(p->tx_ring); in ipoib_cm_tx_init()
1182 if (p->tx_ring) { in ipoib_cm_tx_destroy()
1199 tx_req = &p->tx_ring[p->tx_tail & (ipoib_sendq_size - 1)]; in ipoib_cm_tx_destroy()
1215 vfree(p->tx_ring); in ipoib_cm_tx_destroy()
Dipoib_main.c1293 priv->tx_ring = vzalloc(ipoib_sendq_size * sizeof *priv->tx_ring); in ipoib_dev_init()
1294 if (!priv->tx_ring) { in ipoib_dev_init()
1318 vfree(priv->tx_ring); in ipoib_dev_init()
1354 vfree(priv->tx_ring); in ipoib_dev_cleanup()
1357 priv->tx_ring = NULL; in ipoib_dev_cleanup()
Dipoib.h242 struct ipoib_cm_tx_buf *tx_ring; member
360 struct ipoib_tx_buf *tx_ring; member
Dipoib_ib.c390 tx_req = &priv->tx_ring[wr_id]; in ipoib_ib_handle_tx_wc()
591 tx_req = &priv->tx_ring[priv->tx_head & (ipoib_sendq_size - 1)]; in ipoib_send()
869 tx_req = &priv->tx_ring[priv->tx_tail & in ipoib_ib_dev_stop()
/linux-4.1.27/drivers/net/ethernet/pasemi/
Dpasemi_mac.c125 static struct pasemi_mac_txring *tx_ring(const struct pasemi_mac *mac) in tx_ring() function
556 struct pasemi_mac_txring *txring = tx_ring(mac); in pasemi_mac_free_tx_resources()
696 pcnt = *tx_ring(mac)->chan.status & PAS_STATUS_PCNT_M; in pasemi_mac_restart_tx_intr()
700 write_iob_reg(PAS_IOB_DMA_TXCH_RESET(tx_ring(mac)->chan.chno), reg); in pasemi_mac_restart_tx_intr()
727 struct pasemi_dmachan *chan = &tx_ring(mac)->chan; in pasemi_mac_tx_error()
1177 pasemi_dma_start_chan(&tx_ring(mac)->chan, PAS_DMA_TXCHAN_TCMDSTA_SZ | in pasemi_mac_open()
1267 int txch = tx_ring(mac)->chan.chno; in pasemi_mac_pause_txchan()
1332 txch = tx_ring(mac)->chan.chno; in pasemi_mac_close()
1363 pasemi_mac_clean_tx(tx_ring(mac)); in pasemi_mac_close()
1478 struct pasemi_mac_txring * const txring = tx_ring(mac); in pasemi_mac_start_tx()
[all …]
/linux-4.1.27/drivers/net/ethernet/cadence/
Dmacb.c72 return &queue->tx_ring[macb_tx_ring_wrap(index)]; in macb_tx_desc()
1201 desc = &queue->tx_ring[entry]; in macb_tx_map()
1208 desc = &queue->tx_ring[entry]; in macb_tx_map()
1377 if (queue->tx_ring) { in macb_free_consistent()
1379 queue->tx_ring, queue->tx_ring_dma); in macb_free_consistent()
1380 queue->tx_ring = NULL; in macb_free_consistent()
1424 queue->tx_ring = dma_alloc_coherent(&bp->pdev->dev, size, in macb_alloc_consistent()
1427 if (!queue->tx_ring) in macb_alloc_consistent()
1432 queue->tx_ring); in macb_alloc_consistent()
1467 queue->tx_ring[i].addr = 0; in gem_init_rings()
[all …]
Dmacb.h774 struct macb_dma_desc *tx_ring; member
/linux-4.1.27/drivers/net/ethernet/octeon/
Docteon_mgmt.c126 u64 *tx_ring; member
273 re.d64 = p->tx_ring[p->tx_next_clean]; in octeon_mgmt_clean_tx_buffers()
983 p->tx_ring = kzalloc(ring_size_to_bytes(OCTEON_MGMT_TX_RING_SIZE), in octeon_mgmt_open()
985 if (!p->tx_ring) in octeon_mgmt_open()
988 dma_map_single(p->dev, p->tx_ring, in octeon_mgmt_open()
1236 kfree(p->tx_ring); in octeon_mgmt_open()
1269 kfree(p->tx_ring); in octeon_mgmt_stop()
1307 p->tx_ring[p->tx_next] = re.d64; in octeon_mgmt_xmit()
/linux-4.1.27/drivers/net/ethernet/natsemi/
Dnatsemi.c542 struct netdev_desc *tx_ring; member
1863 printk(KERN_DEBUG " Tx ring at %p:\n", np->tx_ring); in dump_ring()
1866 i, np->tx_ring[i].next_desc, in dump_ring()
1867 np->tx_ring[i].cmd_status, in dump_ring()
1868 np->tx_ring[i].addr); in dump_ring()
1920 np->tx_ring = &np->rx_ring[RX_RING_SIZE]; in alloc_ring()
1970 np->tx_ring[i].next_desc = cpu_to_le32(np->ring_dma in init_ring()
1973 np->tx_ring[i].cmd_status = 0; in init_ring()
2075 np->tx_ring[i].cmd_status = 0; in reinit_ring()
2097 np->tx_ring[entry].addr = cpu_to_le32(np->tx_dma[entry]); in start_tx()
[all …]
/linux-4.1.27/drivers/net/ethernet/altera/
Daltera_tse_main.c284 priv->tx_ring = kcalloc(tx_descs, sizeof(struct tse_buffer), in alloc_init_skbufs()
286 if (!priv->tx_ring) in alloc_init_skbufs()
307 kfree(priv->tx_ring); in alloc_init_skbufs()
325 tse_free_tx_buffer(priv, &priv->tx_ring[i]); in free_skbufs()
328 kfree(priv->tx_ring); in free_skbufs()
466 tx_buff = &priv->tx_ring[entry]; in tse_tx_complete()
593 buffer = &priv->tx_ring[entry]; in tse_start_xmit()
Daltera_tse.h434 struct tse_buffer *tx_ring; member
/linux-4.1.27/include/linux/
Dfs_enet_pd.h139 int rx_ring, tx_ring; /* number of buffers on rx */ member
/linux-4.1.27/drivers/net/ethernet/alteon/
Dacenic.c728 if (ap->tx_ring != NULL && !ACE_IS_TIGON_I(ap)) { in ace_free_descriptors()
730 pci_free_consistent(ap->pdev, size, ap->tx_ring, in ace_free_descriptors()
733 ap->tx_ring = NULL; in ace_free_descriptors()
788 ap->tx_ring = pci_alloc_consistent(ap->pdev, size, in ace_allocate_descriptors()
791 if (ap->tx_ring == NULL) in ace_allocate_descriptors()
1286 ap->tx_ring = (__force struct tx_desc *) regs->Window; in ace_init()
1289 writel(0, (__force void __iomem *)ap->tx_ring + i * 4); in ace_init()
1293 memset(ap->tx_ring, 0, in ace_init()
2335 tx = (__force struct tx_desc __iomem *) &ap->tx_ring[i]; in ace_close()
2340 memset(ap->tx_ring + i, 0, in ace_close()
[all …]
Dacenic.h646 struct tx_desc *tx_ring; member
/linux-4.1.27/drivers/crypto/qat/qat_common/
Dadf_transport.c401 struct adf_etr_ring_data *tx_ring; in adf_init_bank() local
437 tx_ring = &bank->rings[i - hw_data->tx_rx_gap]; in adf_init_bank()
438 ring->inflights = tx_ring->inflights; in adf_init_bank()
/linux-4.1.27/firmware/keyspan_pda/
Dkeyspan_pda.S231 mov dptr, #tx_ring
752 mov dph1, #HIGH(tx_ring) ; load DPTR1 with target
792 ;; tx_ring[0..15]
823 ;; tx_ring[0..15]
825 mov dptr, #tx_ring ; DPTR1: source
876 mov dph, #HIGH(tx_ring)
1117 tx_ring: label
Dxircom_pgs.S233 mov dptr, #tx_ring
809 mov dph1, #HIGH(tx_ring) ; load DPTR1 with target
853 ;; tx_ring[0..15]
884 ;; tx_ring[0..15]
886 mov dptr, #tx_ring ; DPTR1: source
937 mov dph, #HIGH(tx_ring)
1185 tx_ring: label
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8192ee/
Dtrx.c567 rtlpci->tx_ring[q_idx].avl_desc = point_diff; in rtl92ee_get_available_desc()
595 current_bd_desc = rtlpci->tx_ring[queue_index].cur_tx_wp; in rtl92ee_pre_fill_tx_bd_desc()
649 rtlpci->tx_ring[queue_index].dma + in rtl92ee_pre_fill_tx_bd_desc()
929 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[q_idx]; in rtl92ee_set_desc()
1045 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl92ee_is_tx_desc_closed()
Dhw.c186 ring = &rtlpci->tx_ring[queue]; in _rtl92ee_set_fw_clock_off()
843 ((u64)rtlpci->tx_ring[BEACON_QUEUE].buffer_desc_dma) & in _rtl92ee_init_mac()
846 (u64)rtlpci->tx_ring[MGNT_QUEUE].buffer_desc_dma & in _rtl92ee_init_mac()
849 (u64)rtlpci->tx_ring[VO_QUEUE].buffer_desc_dma & in _rtl92ee_init_mac()
852 (u64)rtlpci->tx_ring[VI_QUEUE].buffer_desc_dma & in _rtl92ee_init_mac()
856 (u64)rtlpci->tx_ring[BE_QUEUE].buffer_desc_dma & in _rtl92ee_init_mac()
862 (u64)rtlpci->tx_ring[BK_QUEUE].buffer_desc_dma & in _rtl92ee_init_mac()
865 (u64)rtlpci->tx_ring[HIGH_QUEUE].buffer_desc_dma & in _rtl92ee_init_mac()
/linux-4.1.27/drivers/infiniband/ulp/srp/
Dib_srp.h165 struct srp_iu **tx_ring; member
Dib_srp.c646 if (ch->tx_ring) { in srp_free_ch_ib()
648 srp_free_iu(target->srp_host, ch->tx_ring[i]); in srp_free_ch_ib()
649 kfree(ch->tx_ring); in srp_free_ch_ib()
650 ch->tx_ring = NULL; in srp_free_ch_ib()
1239 list_add(&ch->tx_ring[j]->list, &ch->free_tx); in srp_rport_reconnect()
2106 ch->tx_ring = kcalloc(target->queue_size, sizeof(*ch->tx_ring), in srp_alloc_iu_bufs()
2108 if (!ch->tx_ring) in srp_alloc_iu_bufs()
2120 ch->tx_ring[i] = srp_alloc_iu(target->srp_host, in srp_alloc_iu_bufs()
2123 if (!ch->tx_ring[i]) in srp_alloc_iu_bufs()
2126 list_add(&ch->tx_ring[i]->list, &ch->free_tx); in srp_alloc_iu_bufs()
[all …]
/linux-4.1.27/drivers/net/ethernet/renesas/
Dsh_eth.c1178 memset(mdp->tx_ring, 0, tx_ringsize); in sh_eth_ring_format()
1183 txdesc = &mdp->tx_ring[i]; in sh_eth_ring_format()
1242 mdp->tx_ring = dma_alloc_coherent(NULL, tx_ringsize, &mdp->tx_desc_dma, in sh_eth_ring_init()
1244 if (!mdp->tx_ring) { in sh_eth_ring_init()
1257 mdp->tx_ring = NULL; in sh_eth_ring_init()
1274 if (mdp->tx_ring) { in sh_eth_free_dma_buffer()
1276 dma_free_coherent(NULL, ringsize, mdp->tx_ring, in sh_eth_free_dma_buffer()
1278 mdp->tx_ring = NULL; in sh_eth_free_dma_buffer()
1384 mdp->tx_ring[i].status &= ~cpu_to_edmac(mdp, TD_TACT); in sh_eth_dev_exit()
1415 txdesc = &mdp->tx_ring[entry]; in sh_eth_txfree()
[all …]
Dsh_eth.h510 struct sh_eth_txdesc *tx_ring; member
/linux-4.1.27/drivers/net/ethernet/broadcom/genet/
Dbcmgenet.c2289 struct bcmgenet_tx_ring *tx_ring; in bcmgenet_isr1() local
2321 tx_ring = &priv->tx_rings[index]; in bcmgenet_isr1()
2323 if (likely(napi_schedule_prep(&tx_ring->napi))) { in bcmgenet_isr1()
2324 tx_ring->int_disable(tx_ring); in bcmgenet_isr1()
2325 __napi_schedule(&tx_ring->napi); in bcmgenet_isr1()
2337 struct bcmgenet_tx_ring *tx_ring; in bcmgenet_isr0() local
2360 tx_ring = &priv->tx_rings[DESC_INDEX]; in bcmgenet_isr0()
2362 if (likely(napi_schedule_prep(&tx_ring->napi))) { in bcmgenet_isr0()
2363 tx_ring->int_disable(tx_ring); in bcmgenet_isr0()
2364 __napi_schedule(&tx_ring->napi); in bcmgenet_isr0()
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8192se/
Dhw.c726 rtl_write_dword(rtlpriv, TBKDA, rtlpci->tx_ring[BK_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
727 rtl_write_dword(rtlpriv, TBEDA, rtlpci->tx_ring[BE_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
728 rtl_write_dword(rtlpriv, TVIDA, rtlpci->tx_ring[VI_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
729 rtl_write_dword(rtlpriv, TVODA, rtlpci->tx_ring[VO_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
730 rtl_write_dword(rtlpriv, TBDA, rtlpci->tx_ring[BEACON_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
731 rtl_write_dword(rtlpriv, TCDA, rtlpci->tx_ring[TXCMD_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
732 rtl_write_dword(rtlpriv, TMDA, rtlpci->tx_ring[MGNT_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
733 rtl_write_dword(rtlpriv, THPDA, rtlpci->tx_ring[HIGH_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
734 rtl_write_dword(rtlpriv, HDA, rtlpci->tx_ring[HCCA_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
Dsw.c245 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl92se_is_tx_desc_closed()
Dfw.c146 ring = &rtlpci->tx_ring[TXCMD_QUEUE]; in _rtl92s_cmd_send_packet()
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8821ae/
Dhw.c50 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[BEACON_QUEUE]; in _rtl8821ae_return_beacon_queue_skb()
208 ring = &rtlpci->tx_ring[queue]; in _rtl8821ae_set_fw_clock_off()
1003 rtlpci->tx_ring[BEACON_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac()
1005 rtlpci->tx_ring[MGNT_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac()
1007 rtlpci->tx_ring[VO_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac()
1009 rtlpci->tx_ring[VI_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac()
1011 rtlpci->tx_ring[BE_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac()
1013 rtlpci->tx_ring[BK_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac()
1015 rtlpci->tx_ring[HIGH_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac()
1457 rtlpci->tx_ring[BEACON_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_trx_desc_hw_address()
[all …]
Dtrx.c974 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl8821ae_is_tx_desc_closed()
/linux-4.1.27/drivers/net/ethernet/adaptec/
Dstarfire.c534 starfire_tx_desc *tx_ring; member
915 np->tx_ring = (void *) np->rx_done_q + rx_done_q_size; in netdev_open()
917 np->rx_ring = (void *) np->tx_ring + tx_ring_size; in netdev_open()
1239 np->tx_ring[entry].addr = cpu_to_dma(np->tx_info[entry].mapping); in start_tx()
1240 np->tx_ring[entry].status = cpu_to_le32(status); in start_tx()
1922 i, le32_to_cpu(np->tx_ring[i].status), in netdev_close()
1923 (long long) dma_to_cpu(np->tx_ring[i].addr), in netdev_close()
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8188ee/
Dhw.c92 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[BEACON_QUEUE]; in _rtl88ee_return_beacon_queue_skb()
207 ring = &rtlpci->tx_ring[queue]; in _rtl88ee_set_fw_clock_off()
911 ((u64) rtlpci->tx_ring[BEACON_QUEUE].dma) & in _rtl88ee_init_mac()
914 (u64) rtlpci->tx_ring[MGNT_QUEUE].dma & in _rtl88ee_init_mac()
917 (u64) rtlpci->tx_ring[VO_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl88ee_init_mac()
919 (u64) rtlpci->tx_ring[VI_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl88ee_init_mac()
921 (u64) rtlpci->tx_ring[BE_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl88ee_init_mac()
923 (u64) rtlpci->tx_ring[BK_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl88ee_init_mac()
925 (u64) rtlpci->tx_ring[HIGH_QUEUE].dma & in _rtl88ee_init_mac()
Dtrx.c829 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl88ee_is_tx_desc_closed()
Dphy.c2247 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl88ee_phy_set_rf_power_state()
2291 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl88ee_phy_set_rf_power_state()
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8723be/
Dhw.c53 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[BEACON_QUEUE]; in _rtl8723be_return_beacon_queue_skb()
209 ring = &rtlpci->tx_ring[queue]; in _rtl8723be_set_fw_clock_off()
901 ((u64) rtlpci->tx_ring[BEACON_QUEUE].dma) & in _rtl8723be_init_mac()
904 (u64) rtlpci->tx_ring[MGNT_QUEUE].dma & in _rtl8723be_init_mac()
907 (u64) rtlpci->tx_ring[VO_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8723be_init_mac()
909 (u64) rtlpci->tx_ring[VI_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8723be_init_mac()
911 (u64) rtlpci->tx_ring[BE_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8723be_init_mac()
913 (u64) rtlpci->tx_ring[BK_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8723be_init_mac()
915 (u64) rtlpci->tx_ring[HIGH_QUEUE].dma & in _rtl8723be_init_mac()
Dtrx.c736 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl8723be_is_tx_desc_closed()
Dphy.c2629 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl8723be_phy_set_rf_power_state()
2678 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl8723be_phy_set_rf_power_state()
/linux-4.1.27/Documentation/networking/
Dnetlink_mmap.txt219 void *rx_ring, *tx_ring;
235 tx_ring = rx_ring + ring_size:
314 hdr = tx_ring + frame_offset;
/linux-4.1.27/drivers/net/wireless/
Dadm8211.c316 u32 status = le32_to_cpu(priv->tx_ring[entry].status); in adm8211_interrupt_tci()
1460 desc = &priv->tx_ring[i]; in adm8211_init_rings()
1645 priv->tx_ring[entry].buffer1 = cpu_to_le32(mapping); in adm8211_tx_raw()
1649 priv->tx_ring[entry].length = cpu_to_le32(flag | skb->len); in adm8211_tx_raw()
1653 priv->tx_ring[entry].status = cpu_to_le32(flag); in adm8211_tx_raw()
1743 priv->tx_ring = priv->rx_ring + priv->rx_ring_size; in adm8211_alloc_rings()
Dadm8211.h541 struct adm8211_desc *tx_ring; member
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8192de/
Dhw.c761 rtlpci->tx_ring[BEACON_QUEUE].dma); in _rtl92de_init_mac()
762 rtl_write_dword(rtlpriv, REG_MGQ_DESA, rtlpci->tx_ring[MGNT_QUEUE].dma); in _rtl92de_init_mac()
763 rtl_write_dword(rtlpriv, REG_VOQ_DESA, rtlpci->tx_ring[VO_QUEUE].dma); in _rtl92de_init_mac()
764 rtl_write_dword(rtlpriv, REG_VIQ_DESA, rtlpci->tx_ring[VI_QUEUE].dma); in _rtl92de_init_mac()
765 rtl_write_dword(rtlpriv, REG_BEQ_DESA, rtlpci->tx_ring[BE_QUEUE].dma); in _rtl92de_init_mac()
766 rtl_write_dword(rtlpriv, REG_BKQ_DESA, rtlpci->tx_ring[BK_QUEUE].dma); in _rtl92de_init_mac()
767 rtl_write_dword(rtlpriv, REG_HQ_DESA, rtlpci->tx_ring[HIGH_QUEUE].dma); in _rtl92de_init_mac()
Dfw.c554 ring = &rtlpci->tx_ring[BEACON_QUEUE]; in _rtl92d_cmd_send_packet()
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8192ce/
Dhw.c761 ((u64) rtlpci->tx_ring[BEACON_QUEUE].dma) & in _rtl92ce_init_mac()
764 (u64) rtlpci->tx_ring[MGNT_QUEUE].dma & in _rtl92ce_init_mac()
767 (u64) rtlpci->tx_ring[VO_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl92ce_init_mac()
769 (u64) rtlpci->tx_ring[VI_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl92ce_init_mac()
771 (u64) rtlpci->tx_ring[BE_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl92ce_init_mac()
773 (u64) rtlpci->tx_ring[BK_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl92ce_init_mac()
775 (u64) rtlpci->tx_ring[HIGH_QUEUE].dma & in _rtl92ce_init_mac()
Dphy.c524 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl92ce_phy_set_rf_power_state()
Dtrx.c745 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl92ce_is_tx_desc_closed()
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8723ae/
Dhw.c758 ((u64) rtlpci->tx_ring[BEACON_QUEUE].dma) & in _rtl8712e_init_mac()
761 (u64) rtlpci->tx_ring[MGNT_QUEUE].dma & in _rtl8712e_init_mac()
764 (u64) rtlpci->tx_ring[VO_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8712e_init_mac()
766 (u64) rtlpci->tx_ring[VI_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8712e_init_mac()
768 (u64) rtlpci->tx_ring[BE_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8712e_init_mac()
770 (u64) rtlpci->tx_ring[BK_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8712e_init_mac()
772 (u64) rtlpci->tx_ring[HIGH_QUEUE].dma & in _rtl8712e_init_mac()
Dtrx.c687 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl8723e_is_tx_desc_closed()
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8192cu/
Dphy.c460 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl92cu_phy_set_rf_power_state()
503 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl92cu_phy_set_rf_power_state()
/linux-4.1.27/drivers/net/ethernet/rocker/
Drocker.c226 struct rocker_dma_ring_info tx_ring; member
1229 &rocker_port->tx_ring); in rocker_port_dma_rings_init()
1235 err = rocker_dma_ring_bufs_alloc(rocker, &rocker_port->tx_ring, in rocker_port_dma_rings_init()
1275 rocker_dma_ring_bufs_free(rocker, &rocker_port->tx_ring, in rocker_port_dma_rings_init()
1278 rocker_dma_ring_destroy(rocker, &rocker_port->tx_ring); in rocker_port_dma_rings_init()
1290 rocker_dma_ring_bufs_free(rocker, &rocker_port->tx_ring, in rocker_port_dma_rings_fini()
1292 rocker_dma_ring_destroy(rocker, &rocker_port->tx_ring); in rocker_port_dma_rings_fini()
3947 desc_info = rocker_desc_head_get(&rocker_port->tx_ring); in rocker_port_xmit()
3978 rocker_desc_head_set(rocker, &rocker_port->tx_ring, desc_info); in rocker_port_xmit()
3980 desc_info = rocker_desc_head_get(&rocker_port->tx_ring); in rocker_port_xmit()
[all …]
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8723com/
Dfw_common.c318 ring = &rtlpci->tx_ring[BEACON_QUEUE]; in rtl8723_cmd_send_packet()

12