/linux-4.1.27/drivers/net/ethernet/intel/i40evf/ |
D | i40e_txrx.c | 80 void i40evf_clean_tx_ring(struct i40e_ring *tx_ring) in i40evf_clean_tx_ring() argument 86 if (!tx_ring->tx_bi) in i40evf_clean_tx_ring() 90 for (i = 0; i < tx_ring->count; i++) in i40evf_clean_tx_ring() 91 i40e_unmap_and_free_tx_resource(tx_ring, &tx_ring->tx_bi[i]); in i40evf_clean_tx_ring() 93 bi_size = sizeof(struct i40e_tx_buffer) * tx_ring->count; in i40evf_clean_tx_ring() 94 memset(tx_ring->tx_bi, 0, bi_size); in i40evf_clean_tx_ring() 97 memset(tx_ring->desc, 0, tx_ring->size); in i40evf_clean_tx_ring() 99 tx_ring->next_to_use = 0; in i40evf_clean_tx_ring() 100 tx_ring->next_to_clean = 0; in i40evf_clean_tx_ring() 102 if (!tx_ring->netdev) in i40evf_clean_tx_ring() [all …]
|
D | i40e_txrx.h | 294 void i40evf_clean_tx_ring(struct i40e_ring *tx_ring); 296 int i40evf_setup_tx_descriptors(struct i40e_ring *tx_ring); 298 void i40evf_free_tx_resources(struct i40e_ring *tx_ring);
|
D | i40evf_main.c | 370 struct i40e_ring *tx_ring = adapter->tx_rings[t_idx]; in i40evf_map_vector_to_txq() local 372 tx_ring->q_vector = q_vector; in i40evf_map_vector_to_txq() 373 tx_ring->next = q_vector->tx.ring; in i40evf_map_vector_to_txq() 374 tx_ring->vsi = &adapter->vsi; in i40evf_map_vector_to_txq() 375 q_vector->tx.ring = tx_ring; in i40evf_map_vector_to_txq() 1099 struct i40e_ring *tx_ring; in i40evf_alloc_queues() local 1102 tx_ring = kzalloc(sizeof(*tx_ring) * 2, GFP_KERNEL); in i40evf_alloc_queues() 1103 if (!tx_ring) in i40evf_alloc_queues() 1106 tx_ring->queue_index = i; in i40evf_alloc_queues() 1107 tx_ring->netdev = adapter->netdev; in i40evf_alloc_queues() [all …]
|
/linux-4.1.27/drivers/net/ethernet/intel/i40e/ |
D | i40e_txrx.c | 57 struct i40e_ring *tx_ring; in i40e_program_fdir_filter() local 74 tx_ring = vsi->tx_rings[0]; in i40e_program_fdir_filter() 75 dev = tx_ring->dev; in i40e_program_fdir_filter() 79 if (I40E_DESC_UNUSED(tx_ring) > 1) in i40e_program_fdir_filter() 85 if (!(I40E_DESC_UNUSED(tx_ring) > 1)) in i40e_program_fdir_filter() 94 i = tx_ring->next_to_use; in i40e_program_fdir_filter() 95 fdir_desc = I40E_TX_FDIRDESC(tx_ring, i); in i40e_program_fdir_filter() 96 first = &tx_ring->tx_bi[i]; in i40e_program_fdir_filter() 99 tx_ring->next_to_use = ((i + 1) < tx_ring->count) ? i + 1 : 0; in i40e_program_fdir_filter() 147 i = tx_ring->next_to_use; in i40e_program_fdir_filter() [all …]
|
D | i40e_fcoe.c | 1005 static void i40e_fcoe_program_ddp(struct i40e_ring *tx_ring, in i40e_fcoe_program_ddp() argument 1012 struct i40e_pf *pf = tx_ring->vsi->back; in i40e_fcoe_program_ddp() 1013 u16 i = tx_ring->next_to_use; in i40e_fcoe_program_ddp() 1035 ddp_desc = I40E_DDP_CONTEXT_DESC(tx_ring, i); in i40e_fcoe_program_ddp() 1037 if (i == tx_ring->count) in i40e_fcoe_program_ddp() 1057 queue_desc = I40E_QUEUE_CONTEXT_DESC(tx_ring, i++); in i40e_fcoe_program_ddp() 1058 if (i == tx_ring->count) in i40e_fcoe_program_ddp() 1067 filter_desc = I40E_FILTER_CONTEXT_DESC(tx_ring, i); in i40e_fcoe_program_ddp() 1069 if (i == tx_ring->count) in i40e_fcoe_program_ddp() 1092 tx_ring->next_to_use = i; in i40e_fcoe_program_ddp() [all …]
|
D | i40e_txrx.h | 297 void i40e_clean_tx_ring(struct i40e_ring *tx_ring); 299 int i40e_setup_tx_descriptors(struct i40e_ring *tx_ring); 301 void i40e_free_tx_resources(struct i40e_ring *tx_ring); 305 void i40e_tx_map(struct i40e_ring *tx_ring, struct sk_buff *skb, 308 int i40e_maybe_stop_tx(struct i40e_ring *tx_ring, int size); 309 int i40e_xmit_descriptor_count(struct sk_buff *skb, struct i40e_ring *tx_ring); 311 struct i40e_ring *tx_ring, u32 *flags);
|
D | i40e_debugfs.c | 537 struct i40e_ring *tx_ring = ACCESS_ONCE(vsi->tx_rings[i]); in i40e_dbg_dump_vsi_seid() local 538 if (!tx_ring) in i40e_dbg_dump_vsi_seid() 543 i, tx_ring->desc); in i40e_dbg_dump_vsi_seid() 546 i, tx_ring->dev, in i40e_dbg_dump_vsi_seid() 547 tx_ring->netdev, in i40e_dbg_dump_vsi_seid() 548 tx_ring->tx_bi); in i40e_dbg_dump_vsi_seid() 551 i, tx_ring->state, in i40e_dbg_dump_vsi_seid() 552 tx_ring->queue_index, in i40e_dbg_dump_vsi_seid() 553 tx_ring->reg_idx); in i40e_dbg_dump_vsi_seid() 556 i, tx_ring->dtype); in i40e_dbg_dump_vsi_seid() [all …]
|
D | i40e_main.c | 385 struct i40e_ring *tx_ring, *rx_ring; in i40e_get_netdev_stats_struct() local 401 tx_ring = ACCESS_ONCE(vsi->tx_rings[i]); in i40e_get_netdev_stats_struct() 402 if (!tx_ring) in i40e_get_netdev_stats_struct() 406 start = u64_stats_fetch_begin_irq(&tx_ring->syncp); in i40e_get_netdev_stats_struct() 407 packets = tx_ring->stats.packets; in i40e_get_netdev_stats_struct() 408 bytes = tx_ring->stats.bytes; in i40e_get_netdev_stats_struct() 409 } while (u64_stats_fetch_retry_irq(&tx_ring->syncp, start)); in i40e_get_netdev_stats_struct() 413 rx_ring = &tx_ring[1]; in i40e_get_netdev_stats_struct() 2711 struct i40e_ring *tx_ring, *rx_ring; in i40e_vsi_config_dcb_rings() local 2719 tx_ring = vsi->tx_rings[i]; in i40e_vsi_config_dcb_rings() [all …]
|
D | i40e_ethtool.c | 1267 struct i40e_ring *tx_ring, *rx_ring; in i40e_get_ethtool_stats() local 1297 tx_ring = ACCESS_ONCE(vsi->tx_rings[j]); in i40e_get_ethtool_stats() 1299 if (!tx_ring) in i40e_get_ethtool_stats() 1304 start = u64_stats_fetch_begin_irq(&tx_ring->syncp); in i40e_get_ethtool_stats() 1305 data[i] = tx_ring->stats.packets; in i40e_get_ethtool_stats() 1306 data[i + 1] = tx_ring->stats.bytes; in i40e_get_ethtool_stats() 1307 } while (u64_stats_fetch_retry_irq(&tx_ring->syncp, start)); in i40e_get_ethtool_stats() 1311 rx_ring = &tx_ring[1]; in i40e_get_ethtool_stats()
|
/linux-4.1.27/drivers/net/ethernet/intel/fm10k/ |
D | fm10k_main.c | 793 static int fm10k_tso(struct fm10k_ring *tx_ring, in fm10k_tso() argument 826 tx_desc = FM10K_TX_DESC(tx_ring, tx_ring->next_to_use); in fm10k_tso() 832 tx_ring->netdev->features &= ~NETIF_F_GSO_UDP_TUNNEL; in fm10k_tso() 834 netdev_err(tx_ring->netdev, in fm10k_tso() 839 static void fm10k_tx_csum(struct fm10k_ring *tx_ring, in fm10k_tx_csum() argument 859 dev_warn(tx_ring->dev, in fm10k_tx_csum() 861 tx_ring->tx_stats.csum_err++; in fm10k_tx_csum() 880 dev_warn(tx_ring->dev, in fm10k_tx_csum() 884 tx_ring->tx_stats.csum_err++; in fm10k_tx_csum() 897 dev_warn(tx_ring->dev, in fm10k_tx_csum() [all …]
|
D | fm10k_netdev.c | 33 int fm10k_setup_tx_resources(struct fm10k_ring *tx_ring) in fm10k_setup_tx_resources() argument 35 struct device *dev = tx_ring->dev; in fm10k_setup_tx_resources() 38 size = sizeof(struct fm10k_tx_buffer) * tx_ring->count; in fm10k_setup_tx_resources() 40 tx_ring->tx_buffer = vzalloc(size); in fm10k_setup_tx_resources() 41 if (!tx_ring->tx_buffer) in fm10k_setup_tx_resources() 44 u64_stats_init(&tx_ring->syncp); in fm10k_setup_tx_resources() 47 tx_ring->size = tx_ring->count * sizeof(struct fm10k_tx_desc); in fm10k_setup_tx_resources() 48 tx_ring->size = ALIGN(tx_ring->size, 4096); in fm10k_setup_tx_resources() 50 tx_ring->desc = dma_alloc_coherent(dev, tx_ring->size, in fm10k_setup_tx_resources() 51 &tx_ring->dma, GFP_KERNEL); in fm10k_setup_tx_resources() [all …]
|
D | fm10k_pci.c | 330 struct fm10k_ring *tx_ring = interface->tx_ring[i]; in fm10k_update_stats() local 332 restart_queue += tx_ring->tx_stats.restart_queue; in fm10k_update_stats() 333 tx_busy += tx_ring->tx_stats.tx_busy; in fm10k_update_stats() 334 tx_csum_errors += tx_ring->tx_stats.csum_err; in fm10k_update_stats() 335 bytes += tx_ring->stats.bytes; in fm10k_update_stats() 336 pkts += tx_ring->stats.packets; in fm10k_update_stats() 397 struct fm10k_ring *tx_ring = interface->tx_ring[i]; in fm10k_watchdog_flush_tx() local 399 if (tx_ring->next_to_use != tx_ring->next_to_clean) { in fm10k_watchdog_flush_tx() 463 set_check_for_tx_hang(interface->tx_ring[i]); in fm10k_check_hang_subtask() 596 fm10k_configure_tx_ring(interface, interface->tx_ring[i]); in fm10k_configure_tx() [all …]
|
D | fm10k.h | 265 struct fm10k_ring *tx_ring[MAX_QUEUES] ____cacheline_aligned_in_smp; member 446 struct fm10k_ring *tx_ring); 448 bool fm10k_check_tx_hang(struct fm10k_ring *tx_ring);
|
D | fm10k_ethtool.c | 225 ring = interface->tx_ring[i]; in fm10k_get_ethtool_stats() 498 interface->tx_ring[i]->count = new_tx_count; in fm10k_set_ringparam() 524 memcpy(&temp_ring[i], interface->tx_ring[i], in fm10k_set_ringparam() 539 fm10k_free_tx_resources(interface->tx_ring[i]); in fm10k_set_ringparam() 541 memcpy(interface->tx_ring[i], &temp_ring[i], in fm10k_set_ringparam()
|
/linux-4.1.27/drivers/net/ethernet/intel/igbvf/ |
D | netdev.c | 436 struct igbvf_ring *tx_ring) in igbvf_setup_tx_resources() argument 441 size = sizeof(struct igbvf_buffer) * tx_ring->count; in igbvf_setup_tx_resources() 442 tx_ring->buffer_info = vzalloc(size); in igbvf_setup_tx_resources() 443 if (!tx_ring->buffer_info) in igbvf_setup_tx_resources() 447 tx_ring->size = tx_ring->count * sizeof(union e1000_adv_tx_desc); in igbvf_setup_tx_resources() 448 tx_ring->size = ALIGN(tx_ring->size, 4096); in igbvf_setup_tx_resources() 450 tx_ring->desc = dma_alloc_coherent(&pdev->dev, tx_ring->size, in igbvf_setup_tx_resources() 451 &tx_ring->dma, GFP_KERNEL); in igbvf_setup_tx_resources() 452 if (!tx_ring->desc) in igbvf_setup_tx_resources() 455 tx_ring->adapter = adapter; in igbvf_setup_tx_resources() [all …]
|
D | ethtool.c | 207 struct igbvf_ring *tx_ring = adapter->tx_ring; in igbvf_get_ringparam() local 213 ring->tx_pending = tx_ring->count; in igbvf_get_ringparam() 235 if ((new_tx_count == adapter->tx_ring->count) && in igbvf_set_ringparam() 245 adapter->tx_ring->count = new_tx_count; in igbvf_set_ringparam() 262 if (new_tx_count != adapter->tx_ring->count) { in igbvf_set_ringparam() 263 memcpy(temp_ring, adapter->tx_ring, sizeof(struct igbvf_ring)); in igbvf_set_ringparam() 270 igbvf_free_tx_resources(adapter->tx_ring); in igbvf_set_ringparam() 272 memcpy(adapter->tx_ring, temp_ring, sizeof(struct igbvf_ring)); in igbvf_set_ringparam()
|
D | igbvf.h | 203 struct igbvf_ring *tx_ring /* One per active queue */
|
/linux-4.1.27/drivers/net/ethernet/qlogic/qlcnic/ |
D | qlcnic_io.c | 278 struct qlcnic_host_tx_ring *tx_ring = adapter->tx_ring; in qlcnic_82xx_change_filter() local 282 producer = tx_ring->producer; in qlcnic_82xx_change_filter() 283 hwdesc = &tx_ring->desc_head[tx_ring->producer]; in qlcnic_82xx_change_filter() 299 tx_ring->producer = get_next_index(producer, tx_ring->num_desc); in qlcnic_82xx_change_filter() 374 struct qlcnic_host_tx_ring *tx_ring) in qlcnic_tx_encap_pkt() argument 378 u32 producer = tx_ring->producer; in qlcnic_tx_encap_pkt() 408 hwdesc = &tx_ring->desc_head[producer]; in qlcnic_tx_encap_pkt() 409 tx_ring->cmd_buf_arr[producer].skb = NULL; in qlcnic_tx_encap_pkt() 414 producer = get_next_index(producer, tx_ring->num_desc); in qlcnic_tx_encap_pkt() 417 tx_ring->producer = producer; in qlcnic_tx_encap_pkt() [all …]
|
D | qlcnic_ctx.c | 416 struct qlcnic_host_tx_ring *tx_ring, in qlcnic_82xx_fw_cmd_create_tx_ctx() argument 434 tx_ring->producer = 0; in qlcnic_82xx_fw_cmd_create_tx_ctx() 435 tx_ring->sw_consumer = 0; in qlcnic_82xx_fw_cmd_create_tx_ctx() 436 *(tx_ring->hw_consumer) = 0; in qlcnic_82xx_fw_cmd_create_tx_ctx() 477 prq->cmd_cons_dma_addr = cpu_to_le64(tx_ring->hw_cons_phys_addr); in qlcnic_82xx_fw_cmd_create_tx_ctx() 481 prq_cds->host_phys_addr = cpu_to_le64(tx_ring->phys_addr); in qlcnic_82xx_fw_cmd_create_tx_ctx() 482 prq_cds->ring_size = cpu_to_le32(tx_ring->num_desc); in qlcnic_82xx_fw_cmd_create_tx_ctx() 496 tx_ring->state = le32_to_cpu(prsp->host_ctx_state); in qlcnic_82xx_fw_cmd_create_tx_ctx() 498 tx_ring->crb_cmd_producer = adapter->ahw->pci_base0 + temp; in qlcnic_82xx_fw_cmd_create_tx_ctx() 499 tx_ring->ctx_id = le16_to_cpu(prsp->context_id); in qlcnic_82xx_fw_cmd_create_tx_ctx() [all …]
|
D | qlcnic_main.c | 124 inline void qlcnic_update_cmd_producer(struct qlcnic_host_tx_ring *tx_ring) in qlcnic_update_cmd_producer() argument 126 writel(tx_ring->producer, tx_ring->crb_cmd_producer); in qlcnic_update_cmd_producer() 1712 struct qlcnic_host_tx_ring *tx_ring; in qlcnic_request_irq() local 1782 tx_ring = &adapter->tx_ring[ring]; in qlcnic_request_irq() 1783 snprintf(tx_ring->name, sizeof(tx_ring->name), in qlcnic_request_irq() 1785 err = request_irq(tx_ring->irq, handler, flags, in qlcnic_request_irq() 1786 tx_ring->name, tx_ring); in qlcnic_request_irq() 1800 struct qlcnic_host_tx_ring *tx_ring; in qlcnic_free_irq() local 1819 tx_ring = &adapter->tx_ring[ring]; in qlcnic_free_irq() 1820 if (tx_ring->irq) in qlcnic_free_irq() [all …]
|
D | qlcnic_ethtool.c | 525 struct qlcnic_host_tx_ring *tx_ring; in qlcnic_get_regs() local 553 tx_ring = &adapter->tx_ring[ring]; in qlcnic_get_regs() 554 regs_buff[i++] = le32_to_cpu(*(tx_ring->hw_consumer)); in qlcnic_get_regs() 555 regs_buff[i++] = tx_ring->sw_consumer; in qlcnic_get_regs() 556 regs_buff[i++] = readl(tx_ring->crb_cmd_producer); in qlcnic_get_regs() 557 regs_buff[i++] = tx_ring->producer; in qlcnic_get_regs() 558 if (tx_ring->crb_intr_mask) in qlcnic_get_regs() 559 regs_buff[i++] = readl(tx_ring->crb_intr_mask); in qlcnic_get_regs() 686 u8 rx_ring, u8 tx_ring) in qlcnic_validate_ring_count() argument 688 if (rx_ring == 0 || tx_ring == 0) in qlcnic_validate_ring_count() [all …]
|
D | qlcnic.h | 73 #define TX_BUFF_RINGSIZE(tx_ring) \ argument 74 (sizeof(struct qlcnic_cmd_buffer) * tx_ring->num_desc) 75 #define TX_DESC_RINGSIZE(tx_ring) \ argument 76 (sizeof(struct cmd_desc_type0) * tx_ring->num_desc) 595 struct qlcnic_host_tx_ring *tx_ring; member 1109 struct qlcnic_host_tx_ring *tx_ring; member 1730 static inline u32 qlcnic_tx_avail(struct qlcnic_host_tx_ring *tx_ring) in qlcnic_tx_avail() argument 1732 if (likely(tx_ring->producer < tx_ring->sw_consumer)) in qlcnic_tx_avail() 1733 return tx_ring->sw_consumer - tx_ring->producer; in qlcnic_tx_avail() 1735 return tx_ring->sw_consumer + tx_ring->num_desc - in qlcnic_tx_avail() [all …]
|
D | qlcnic_hw.c | 395 struct qlcnic_host_tx_ring *tx_ring; in qlcnic_send_cmd_descs() local 402 tx_ring = &adapter->tx_ring[0]; in qlcnic_send_cmd_descs() 403 __netif_tx_lock_bh(tx_ring->txq); in qlcnic_send_cmd_descs() 405 producer = tx_ring->producer; in qlcnic_send_cmd_descs() 407 if (nr_desc >= qlcnic_tx_avail(tx_ring)) { in qlcnic_send_cmd_descs() 408 netif_tx_stop_queue(tx_ring->txq); in qlcnic_send_cmd_descs() 410 if (qlcnic_tx_avail(tx_ring) > nr_desc) { in qlcnic_send_cmd_descs() 411 if (qlcnic_tx_avail(tx_ring) > TX_STOP_THRESH) in qlcnic_send_cmd_descs() 412 netif_tx_wake_queue(tx_ring->txq); in qlcnic_send_cmd_descs() 415 __netif_tx_unlock_bh(tx_ring->txq); in qlcnic_send_cmd_descs() [all …]
|
D | qlcnic_init.c | 131 struct qlcnic_host_tx_ring *tx_ring) in qlcnic_release_tx_buffers() argument 137 spin_lock(&tx_ring->tx_clean_lock); in qlcnic_release_tx_buffers() 139 cmd_buf = tx_ring->cmd_buf_arr; in qlcnic_release_tx_buffers() 140 for (i = 0; i < tx_ring->num_desc; i++) { in qlcnic_release_tx_buffers() 163 spin_unlock(&tx_ring->tx_clean_lock); in qlcnic_release_tx_buffers() 251 sds_ring->tx_ring = &adapter->tx_ring[ring]; in qlcnic_alloc_sw_resources() 253 sds_ring->tx_ring = &adapter->tx_ring[0]; in qlcnic_alloc_sw_resources()
|
D | qlcnic_hw.h | 194 struct qlcnic_host_tx_ring *tx_ring, int);
|
D | qlcnic_83xx_hw.c | 1290 struct qlcnic_host_tx_ring *tx_ring) in qlcnic_83xx_del_tx_ctx() argument 1304 cmd.req.arg[1] = tx_ring->ctx_id | temp; in qlcnic_83xx_del_tx_ctx() 2225 temp = adapter->tx_ring->ctx_id; in qlcnic_83xx_set_tx_intr_coal() 3486 cmd.req.arg[1] = BIT_1 | (adapter->tx_ring->ctx_id << 16); in qlcnic_83xx_get_stats()
|
/linux-4.1.27/drivers/net/ethernet/intel/ixgbevf/ |
D | ixgbevf_main.c | 196 static void ixgbevf_unmap_and_free_tx_resource(struct ixgbevf_ring *tx_ring, in ixgbevf_unmap_and_free_tx_resource() argument 202 dma_unmap_single(tx_ring->dev, in ixgbevf_unmap_and_free_tx_resource() 207 dma_unmap_page(tx_ring->dev, in ixgbevf_unmap_and_free_tx_resource() 238 static inline bool ixgbevf_check_tx_hang(struct ixgbevf_ring *tx_ring) in ixgbevf_check_tx_hang() argument 240 u32 tx_done = ixgbevf_get_tx_completed(tx_ring); in ixgbevf_check_tx_hang() 241 u32 tx_done_old = tx_ring->tx_stats.tx_done_old; in ixgbevf_check_tx_hang() 242 u32 tx_pending = ixgbevf_get_tx_pending(tx_ring); in ixgbevf_check_tx_hang() 244 clear_check_for_tx_hang(tx_ring); in ixgbevf_check_tx_hang() 254 &tx_ring->state); in ixgbevf_check_tx_hang() 257 clear_bit(__IXGBEVF_HANG_CHECK_ARMED, &tx_ring->state); in ixgbevf_check_tx_hang() [all …]
|
D | ethtool.c | 268 struct ixgbevf_ring *tx_ring = NULL, *rx_ring = NULL; in ixgbevf_set_ringparam() local 293 adapter->tx_ring[i]->count = new_tx_count; in ixgbevf_set_ringparam() 302 tx_ring = vmalloc(adapter->num_tx_queues * sizeof(*tx_ring)); in ixgbevf_set_ringparam() 303 if (!tx_ring) { in ixgbevf_set_ringparam() 310 tx_ring[i] = *adapter->tx_ring[i]; in ixgbevf_set_ringparam() 311 tx_ring[i].count = new_tx_count; in ixgbevf_set_ringparam() 312 err = ixgbevf_setup_tx_resources(&tx_ring[i]); in ixgbevf_set_ringparam() 316 ixgbevf_free_tx_resources(&tx_ring[i]); in ixgbevf_set_ringparam() 319 vfree(tx_ring); in ixgbevf_set_ringparam() 320 tx_ring = NULL; in ixgbevf_set_ringparam() [all …]
|
D | ixgbevf.h | 392 struct ixgbevf_ring *tx_ring[MAX_TX_QUEUES]; /* One per active queue */ member
|
/linux-4.1.27/drivers/net/ethernet/agere/ |
D | et131x.c | 357 struct tx_ring { struct 488 struct tx_ring tx_ring; member 1642 struct tx_ring *tx_ring = &adapter->tx_ring; in et131x_config_tx_dma_regs() local 1645 writel(upper_32_bits(tx_ring->tx_desc_ring_pa), &txdma->pr_base_hi); in et131x_config_tx_dma_regs() 1646 writel(lower_32_bits(tx_ring->tx_desc_ring_pa), &txdma->pr_base_lo); in et131x_config_tx_dma_regs() 1652 writel(upper_32_bits(tx_ring->tx_status_pa), &txdma->dma_wb_base_hi); in et131x_config_tx_dma_regs() 1653 writel(lower_32_bits(tx_ring->tx_status_pa), &txdma->dma_wb_base_lo); in et131x_config_tx_dma_regs() 1655 *tx_ring->tx_status = 0; in et131x_config_tx_dma_regs() 1658 tx_ring->send_idx = 0; in et131x_config_tx_dma_regs() 1755 struct tx_ring *tx_ring = &adapter->tx_ring; in et131x_init_send() local [all …]
|
/linux-4.1.27/drivers/net/ethernet/intel/ixgb/ |
D | ixgb_main.c | 702 struct ixgb_desc_ring *txdr = &adapter->tx_ring; in ixgb_setup_tx_resources() 739 u64 tdba = adapter->tx_ring.dma; in ixgb_configure_tx() 740 u32 tdlen = adapter->tx_ring.count * sizeof(struct ixgb_tx_desc); in ixgb_configure_tx() 916 vfree(adapter->tx_ring.buffer_info); in ixgb_free_tx_resources() 917 adapter->tx_ring.buffer_info = NULL; in ixgb_free_tx_resources() 919 dma_free_coherent(&pdev->dev, adapter->tx_ring.size, in ixgb_free_tx_resources() 920 adapter->tx_ring.desc, adapter->tx_ring.dma); in ixgb_free_tx_resources() 922 adapter->tx_ring.desc = NULL; in ixgb_free_tx_resources() 957 struct ixgb_desc_ring *tx_ring = &adapter->tx_ring; in ixgb_clean_tx_ring() local 964 for (i = 0; i < tx_ring->count; i++) { in ixgb_clean_tx_ring() [all …]
|
D | ixgb_param.c | 278 struct ixgb_desc_ring *tx_ring = &adapter->tx_ring; in ixgb_check_options() local 281 tx_ring->count = TxDescriptors[bd]; in ixgb_check_options() 282 ixgb_validate_option(&tx_ring->count, &opt); in ixgb_check_options() 284 tx_ring->count = opt.def; in ixgb_check_options() 286 tx_ring->count = ALIGN(tx_ring->count, IXGB_REQ_TX_DESCRIPTOR_MULTIPLE); in ixgb_check_options()
|
D | ixgb_ethtool.c | 492 struct ixgb_desc_ring *txdr = &adapter->tx_ring; in ixgb_get_ringparam() 506 struct ixgb_desc_ring *txdr = &adapter->tx_ring; in ixgb_set_ringparam() 511 tx_old = adapter->tx_ring; in ixgb_set_ringparam() 539 tx_new = adapter->tx_ring; in ixgb_set_ringparam() 541 adapter->tx_ring = tx_old; in ixgb_set_ringparam() 545 adapter->tx_ring = tx_new; in ixgb_set_ringparam() 556 adapter->tx_ring = tx_old; in ixgb_set_ringparam()
|
D | ixgb.h | 148 struct ixgb_desc_ring tx_ring ____cacheline_aligned_in_smp;
|
/linux-4.1.27/drivers/net/ethernet/oki-semi/pch_gbe/ |
D | pch_gbe_main.c | 631 adapter->tx_ring = devm_kzalloc(&adapter->pdev->dev, in pch_gbe_alloc_queues() 632 sizeof(*adapter->tx_ring), GFP_KERNEL); in pch_gbe_alloc_queues() 633 if (!adapter->tx_ring) in pch_gbe_alloc_queues() 858 (unsigned long long)adapter->tx_ring->dma, in pch_gbe_configure_tx() 859 adapter->tx_ring->size); in pch_gbe_configure_tx() 862 tdba = adapter->tx_ring->dma; in pch_gbe_configure_tx() 863 tdlen = adapter->tx_ring->size - 0x10; in pch_gbe_configure_tx() 977 struct pch_gbe_tx_ring *tx_ring) in pch_gbe_clean_tx_ring() argument 985 for (i = 0; i < tx_ring->count; i++) { in pch_gbe_clean_tx_ring() 986 buffer_info = &tx_ring->buffer_info[i]; in pch_gbe_clean_tx_ring() [all …]
|
D | pch_gbe_ethtool.c | 270 struct pch_gbe_tx_ring *txdr = adapter->tx_ring; in pch_gbe_get_ringparam() 303 tx_old = adapter->tx_ring; in pch_gbe_set_ringparam() 316 adapter->tx_ring = txdr; in pch_gbe_set_ringparam() 332 err = pch_gbe_setup_tx_resources(adapter, adapter->tx_ring); in pch_gbe_set_ringparam() 339 adapter->tx_ring = tx_old; in pch_gbe_set_ringparam() 341 pch_gbe_free_tx_resources(adapter, adapter->tx_ring); in pch_gbe_set_ringparam() 345 adapter->tx_ring = txdr; in pch_gbe_set_ringparam() 352 adapter->tx_ring = txdr; in pch_gbe_set_ringparam() 362 adapter->tx_ring = tx_old; in pch_gbe_set_ringparam()
|
D | pch_gbe_param.c | 459 struct pch_gbe_tx_ring *tx_ring = adapter->tx_ring; in pch_gbe_check_options() local 460 tx_ring->count = TxDescriptors; in pch_gbe_check_options() 461 pch_gbe_validate_option(&tx_ring->count, &opt, adapter); in pch_gbe_check_options() 462 tx_ring->count = roundup(tx_ring->count, in pch_gbe_check_options()
|
D | pch_gbe.h | 638 struct pch_gbe_tx_ring *tx_ring; member 664 struct pch_gbe_tx_ring *tx_ring);
|
/linux-4.1.27/drivers/infiniband/hw/amso1100/ |
D | c2.c | 111 static int c2_tx_ring_alloc(struct c2_ring *tx_ring, void *vaddr, in c2_tx_ring_alloc() argument 119 tx_ring->start = kmalloc(sizeof(*elem) * tx_ring->count, GFP_KERNEL); in c2_tx_ring_alloc() 120 if (!tx_ring->start) in c2_tx_ring_alloc() 123 elem = tx_ring->start; in c2_tx_ring_alloc() 126 for (i = 0; i < tx_ring->count; i++, elem++, tx_desc++, txp_desc++) { in c2_tx_ring_alloc() 141 if (i == tx_ring->count - 1) { in c2_tx_ring_alloc() 142 elem->next = tx_ring->start; in c2_tx_ring_alloc() 151 tx_ring->to_use = tx_ring->to_clean = tx_ring->start; in c2_tx_ring_alloc() 326 struct c2_ring *tx_ring = &c2_port->tx_ring; in c2_tx_clean() local 334 elem = tx_ring->start; in c2_tx_clean() [all …]
|
D | c2.h | 363 struct c2_ring tx_ring; member
|
/linux-4.1.27/drivers/net/ethernet/intel/e1000e/ |
D | netdev.c | 218 struct e1000_ring *tx_ring = adapter->tx_ring; in e1000e_dump() local 262 buffer_info = &tx_ring->buffer_info[tx_ring->next_to_clean]; in e1000e_dump() 264 0, tx_ring->next_to_use, tx_ring->next_to_clean, in e1000e_dump() 306 for (i = 0; tx_ring->desc && (i < tx_ring->count); i++) { in e1000e_dump() 308 tx_desc = E1000_TX_DESC(*tx_ring, i); in e1000e_dump() 309 buffer_info = &tx_ring->buffer_info[i]; in e1000e_dump() 311 if (i == tx_ring->next_to_use && i == tx_ring->next_to_clean) in e1000e_dump() 313 else if (i == tx_ring->next_to_use) in e1000e_dump() 315 else if (i == tx_ring->next_to_clean) in e1000e_dump() 640 static void e1000e_update_tdt_wa(struct e1000_ring *tx_ring, unsigned int i) in e1000e_update_tdt_wa() argument [all …]
|
D | ethtool.c | 696 adapter->tx_ring->count = new_tx_count; in e1000_set_ringparam() 731 memcpy(temp_tx, adapter->tx_ring, size); in e1000_set_ringparam() 747 e1000e_free_tx_resources(adapter->tx_ring); in e1000_set_ringparam() 748 memcpy(adapter->tx_ring, temp_tx, size); in e1000_set_ringparam() 1116 struct e1000_ring *tx_ring = &adapter->test_tx_ring; in e1000_free_desc_rings() local 1122 if (tx_ring->desc && tx_ring->buffer_info) { in e1000_free_desc_rings() 1123 for (i = 0; i < tx_ring->count; i++) { in e1000_free_desc_rings() 1124 buffer_info = &tx_ring->buffer_info[i]; in e1000_free_desc_rings() 1149 if (tx_ring->desc) { in e1000_free_desc_rings() 1150 dma_free_coherent(&pdev->dev, tx_ring->size, tx_ring->desc, in e1000_free_desc_rings() [all …]
|
D | e1000.h | 235 struct e1000_ring *tx_ring ____cacheline_aligned_in_smp;
|
/linux-4.1.27/drivers/net/ethernet/atheros/atl1e/ |
D | atl1e_main.c | 636 hw->tpd_thresh = adapter->tx_ring.count / 2; in atl1e_sw_init() 664 struct atl1e_tx_ring *tx_ring = &adapter->tx_ring; in atl1e_clean_tx_ring() local 669 if (tx_ring->desc == NULL || tx_ring->tx_buffer == NULL) in atl1e_clean_tx_ring() 672 ring_count = tx_ring->count; in atl1e_clean_tx_ring() 675 tx_buffer = &tx_ring->tx_buffer[index]; in atl1e_clean_tx_ring() 688 tx_buffer = &tx_ring->tx_buffer[index]; in atl1e_clean_tx_ring() 695 memset(tx_ring->desc, 0, sizeof(struct atl1e_tpd_desc) * in atl1e_clean_tx_ring() 697 memset(tx_ring->tx_buffer, 0, sizeof(struct atl1e_tx_buffer) * in atl1e_clean_tx_ring() 728 *ring_size = ((u32)(adapter->tx_ring.count * in atl1e_cal_ring_size() 754 rwlock_init(&adapter->tx_ring.tx_lock); in atl1e_init_ring_resources() [all …]
|
D | atl1e_param.c | 206 adapter->tx_ring.count = (u16) val & 0xFFFC; in atl1e_check_options() 208 adapter->tx_ring.count = (u16)opt.def; in atl1e_check_options()
|
D | atl1e.h | 458 struct atl1e_tx_ring tx_ring; member
|
/linux-4.1.27/drivers/net/ethernet/apm/xgene/ |
D | xgene_enet_main.c | 227 static int xgene_enet_setup_tx_desc(struct xgene_enet_desc_ring *tx_ring, in xgene_enet_setup_tx_desc() argument 230 struct device *dev = ndev_to_dev(tx_ring->ndev); in xgene_enet_setup_tx_desc() 233 u16 tail = tx_ring->tail; in xgene_enet_setup_tx_desc() 236 raw_desc = &tx_ring->raw_desc[tail]; in xgene_enet_setup_tx_desc() 241 netdev_err(tx_ring->ndev, "DMA mapping error\n"); in xgene_enet_setup_tx_desc() 251 raw_desc->m3 = cpu_to_le64(SET_VAL(HENQNUM, tx_ring->dst_ring_num) | in xgene_enet_setup_tx_desc() 253 tx_ring->cp_ring->cp_skb[tail] = skb; in xgene_enet_setup_tx_desc() 262 struct xgene_enet_desc_ring *tx_ring = pdata->tx_ring; in xgene_enet_start_xmit() local 263 struct xgene_enet_desc_ring *cp_ring = tx_ring->cp_ring; in xgene_enet_start_xmit() 266 tx_level = xgene_enet_ring_len(tx_ring); in xgene_enet_start_xmit() [all …]
|
D | xgene_enet_main.h | 116 struct xgene_enet_desc_ring *tx_ring; member
|
/linux-4.1.27/drivers/net/ethernet/intel/e1000/ |
D | e1000_main.c | 106 struct e1000_tx_ring *tx_ring); 125 struct e1000_tx_ring *tx_ring); 139 struct e1000_tx_ring *tx_ring); 1245 kfree(adapter->tx_ring); in e1000_probe() 1283 kfree(adapter->tx_ring); in e1000_remove() 1336 adapter->tx_ring = kcalloc(adapter->num_tx_queues, in e1000_alloc_queues() 1338 if (!adapter->tx_ring) in e1000_alloc_queues() 1344 kfree(adapter->tx_ring); in e1000_alloc_queues() 1578 err = e1000_setup_tx_resources(adapter, &adapter->tx_ring[i]); in e1000_setup_all_tx_resources() 1583 &adapter->tx_ring[i]); in e1000_setup_all_tx_resources() [all …]
|
D | e1000_param.c | 281 struct e1000_tx_ring *tx_ring = adapter->tx_ring; in e1000_check_options() local 298 tx_ring->count = TxDescriptors[bd]; in e1000_check_options() 299 e1000_validate_option(&tx_ring->count, &opt, adapter); in e1000_check_options() 300 tx_ring->count = ALIGN(tx_ring->count, in e1000_check_options() 303 tx_ring->count = opt.def; in e1000_check_options() 306 tx_ring[i].count = tx_ring->count; in e1000_check_options()
|
D | e1000_ethtool.c | 572 struct e1000_tx_ring *txdr = adapter->tx_ring; in e1000_get_ringparam() 602 tx_old = adapter->tx_ring; in e1000_set_ringparam() 616 adapter->tx_ring = txdr; in e1000_set_ringparam() 647 adapter->tx_ring = tx_old; in e1000_set_ringparam() 653 adapter->tx_ring = txdr; in e1000_set_ringparam() 665 adapter->tx_ring = tx_old; in e1000_set_ringparam()
|
D | e1000.h | 252 struct e1000_tx_ring *tx_ring; /* One per active queue */ member
|
/linux-4.1.27/drivers/net/ethernet/intel/igb/ |
D | igb_main.c | 367 struct igb_ring *tx_ring; in igb_dump() local 402 tx_ring = adapter->tx_ring[n]; in igb_dump() 403 buffer_info = &tx_ring->tx_buffer_info[tx_ring->next_to_clean]; in igb_dump() 405 n, tx_ring->next_to_use, tx_ring->next_to_clean, in igb_dump() 430 tx_ring = adapter->tx_ring[n]; in igb_dump() 432 pr_info("TX QUEUE INDEX = %d\n", tx_ring->queue_index); in igb_dump() 436 for (i = 0; tx_ring->desc && (i < tx_ring->count); i++) { in igb_dump() 439 tx_desc = IGB_TX_DESC(tx_ring, i); in igb_dump() 440 buffer_info = &tx_ring->tx_buffer_info[i]; in igb_dump() 442 if (i == tx_ring->next_to_use && in igb_dump() [all …]
|
D | igb.h | 373 struct igb_ring *tx_ring[16]; member 580 static inline struct netdev_queue *txring_txq(const struct igb_ring *tx_ring) in txring_txq() argument 582 return netdev_get_tx_queue(tx_ring->netdev, tx_ring->queue_index); in txring_txq()
|
D | igb_ethtool.c | 892 adapter->tx_ring[i]->count = new_tx_count; in igb_set_ringparam() 920 memcpy(&temp_ring[i], adapter->tx_ring[i], in igb_set_ringparam() 935 igb_free_tx_resources(adapter->tx_ring[i]); in igb_set_ringparam() 937 memcpy(adapter->tx_ring[i], &temp_ring[i], in igb_set_ringparam() 1540 struct igb_ring *tx_ring = &adapter->test_tx_ring; in igb_setup_desc_rings() local 1546 tx_ring->count = IGB_DEFAULT_TXD; in igb_setup_desc_rings() 1547 tx_ring->dev = &adapter->pdev->dev; in igb_setup_desc_rings() 1548 tx_ring->netdev = adapter->netdev; in igb_setup_desc_rings() 1549 tx_ring->reg_idx = adapter->vfs_allocated_count; in igb_setup_desc_rings() 1551 if (igb_setup_tx_resources(tx_ring)) { in igb_setup_desc_rings() [all …]
|
/linux-4.1.27/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_main.c | 554 struct ixgbe_ring *tx_ring; in ixgbe_dump() local 596 tx_ring = adapter->tx_ring[n]; in ixgbe_dump() 597 tx_buffer = &tx_ring->tx_buffer_info[tx_ring->next_to_clean]; in ixgbe_dump() 599 n, tx_ring->next_to_use, tx_ring->next_to_clean, in ixgbe_dump() 648 tx_ring = adapter->tx_ring[n]; in ixgbe_dump() 650 pr_info("TX QUEUE INDEX = %d\n", tx_ring->queue_index); in ixgbe_dump() 657 for (i = 0; tx_ring->desc && (i < tx_ring->count); i++) { in ixgbe_dump() 658 tx_desc = IXGBE_TX_DESC(tx_ring, i); in ixgbe_dump() 659 tx_buffer = &tx_ring->tx_buffer_info[i]; in ixgbe_dump() 671 if (i == tx_ring->next_to_use && in ixgbe_dump() [all …]
|
D | ixgbe_lib.c | 74 adapter->tx_ring[i]->reg_idx = reg_idx; in ixgbe_cache_ring_dcb_sriov() 101 adapter->tx_ring[i]->reg_idx = reg_idx; in ixgbe_cache_ring_dcb_sriov() 184 adapter->tx_ring[offset + i]->reg_idx = tx_idx; in ixgbe_cache_ring_dcb() 186 adapter->tx_ring[offset + i]->dcb_tc = tc; in ixgbe_cache_ring_dcb() 247 adapter->tx_ring[i]->reg_idx = reg_idx; in ixgbe_cache_ring_sriov() 253 adapter->tx_ring[i]->reg_idx = reg_idx; in ixgbe_cache_ring_sriov() 274 adapter->tx_ring[i]->reg_idx = i; in ixgbe_cache_ring_rss() 294 adapter->tx_ring[0]->reg_idx = 0; in ixgbe_cache_ring_register() 900 adapter->tx_ring[txr_idx] = ring; in ixgbe_alloc_q_vector() 975 adapter->tx_ring[ring->queue_index] = NULL; in ixgbe_free_q_vector() [all …]
|
D | ixgbe_ethtool.c | 947 struct ixgbe_ring *tx_ring = adapter->tx_ring[0]; in ixgbe_get_ringparam() local 953 ring->tx_pending = tx_ring->count; in ixgbe_get_ringparam() 986 adapter->tx_ring[i]->count = new_tx_count; in ixgbe_set_ringparam() 1013 memcpy(&temp_ring[i], adapter->tx_ring[i], in ixgbe_set_ringparam() 1028 ixgbe_free_tx_resources(adapter->tx_ring[i]); in ixgbe_set_ringparam() 1030 memcpy(adapter->tx_ring[i], &temp_ring[i], in ixgbe_set_ringparam() 1117 ring = adapter->tx_ring[j]; in ixgbe_get_ethtool_stats() 1632 struct ixgbe_ring *tx_ring = &adapter->test_tx_ring; in ixgbe_free_desc_rings() local 1644 reg_ctl = IXGBE_READ_REG(hw, IXGBE_TXDCTL(tx_ring->reg_idx)); in ixgbe_free_desc_rings() 1646 IXGBE_WRITE_REG(hw, IXGBE_TXDCTL(tx_ring->reg_idx), reg_ctl); in ixgbe_free_desc_rings() [all …]
|
D | ixgbe_fcoe.c | 511 int ixgbe_fso(struct ixgbe_ring *tx_ring, in ixgbe_fso() argument 523 dev_err(tx_ring->dev, "Wrong gso type %d:expecting SKB_GSO_FCOE\n", in ixgbe_fso() 549 dev_warn(tx_ring->dev, "unknown sof = 0x%x\n", sof); in ixgbe_fso() 575 dev_warn(tx_ring->dev, "unknown eof = 0x%x\n", eof); in ixgbe_fso() 612 ixgbe_tx_ctxtdesc(tx_ring, vlan_macip_lens, fcoe_sof_eof, in ixgbe_fso()
|
D | ixgbe.h | 657 struct ixgbe_ring *tx_ring[MAX_TX_QUEUES] ____cacheline_aligned_in_smp; member 912 int ixgbe_fso(struct ixgbe_ring *tx_ring, struct ixgbe_tx_buffer *first, 967 struct ixgbe_ring *tx_ring);
|
/linux-4.1.27/drivers/net/ethernet/packetengines/ |
D | yellowfin.c | 310 struct yellowfin_desc *tx_ring; member 441 np->tx_ring = ring_space; in yellowfin_init_one() 514 pci_free_consistent(pdev, TX_TOTAL_SIZE, np->tx_ring, np->tx_ring_dma); in yellowfin_init_one() 700 pr_warn(" Tx ring %p: ", yp->tx_ring); in yellowfin_tx_timeout() 704 yp->tx_ring[i].result_status); in yellowfin_tx_timeout() 762 yp->tx_ring[i].dbdma_cmd = cpu_to_le32(CMD_STOP); in yellowfin_init_ring() 763 yp->tx_ring[i].branch_addr = cpu_to_le32(yp->tx_ring_dma + in yellowfin_init_ring() 767 yp->tx_ring[--i].dbdma_cmd = cpu_to_le32(CMD_STOP | BRANCH_ALWAYS); in yellowfin_init_ring() 775 yp->tx_ring[j].dbdma_cmd = cpu_to_le32(CMD_STOP); in yellowfin_init_ring() 776 yp->tx_ring[j].branch_addr = cpu_to_le32(yp->tx_ring_dma + in yellowfin_init_ring() [all …]
|
D | hamachi.c | 485 struct hamachi_desc *tx_ring; member 651 hmp->tx_ring = ring_space; in hamachi_init_one() 780 pci_free_consistent(pdev, TX_TOTAL_SIZE, hmp->tx_ring, in hamachi_init_one() 1002 if (hmp->tx_ring[entry].status_n_length & cpu_to_le32(DescOwn)) in hamachi_tx() 1008 leXX_to_cpu(hmp->tx_ring[entry].addr), in hamachi_tx() 1013 hmp->tx_ring[entry].status_n_length = 0; in hamachi_tx() 1015 hmp->tx_ring[TX_RING_SIZE-1].status_n_length |= in hamachi_tx() 1063 printk(KERN_DEBUG" Tx ring %p: ", hmp->tx_ring); in hamachi_tx_timeout() 1066 le32_to_cpu(hmp->tx_ring[i].status_n_length)); in hamachi_tx_timeout() 1091 hmp->tx_ring[i].status_n_length = in hamachi_tx_timeout() [all …]
|
/linux-4.1.27/drivers/net/ethernet/qlogic/netxen/ |
D | netxen_nic_ctx.c | 448 struct nx_host_tx_ring *tx_ring = adapter->tx_ring; in nx_fw_cmd_create_tx_ctx() local 490 prq_cds->host_phys_addr = cpu_to_le64(tx_ring->phys_addr); in nx_fw_cmd_create_tx_ctx() 491 prq_cds->ring_size = cpu_to_le32(tx_ring->num_desc); in nx_fw_cmd_create_tx_ctx() 503 tx_ring->crb_cmd_producer = netxen_get_ioaddr(adapter, in nx_fw_cmd_create_tx_ctx() 706 struct nx_host_tx_ring *tx_ring; in netxen_init_old_ctx() local 712 tx_ring = adapter->tx_ring; in netxen_init_old_ctx() 716 hwctx->cmd_ring_addr = cpu_to_le64(tx_ring->phys_addr); in netxen_init_old_ctx() 717 hwctx->cmd_ring_size = cpu_to_le32(tx_ring->num_desc); in netxen_init_old_ctx() 762 struct nx_host_tx_ring *tx_ring; in netxen_alloc_hw_resources() local 769 tx_ring = adapter->tx_ring; in netxen_alloc_hw_resources() [all …]
|
D | netxen_nic_init.c | 136 struct nx_host_tx_ring *tx_ring = adapter->tx_ring; in netxen_release_tx_buffers() local 139 cmd_buf = tx_ring->cmd_buf_arr; in netxen_release_tx_buffers() 140 for (i = 0; i < tx_ring->num_desc; i++) { in netxen_release_tx_buffers() 169 struct nx_host_tx_ring *tx_ring; in netxen_free_sw_resources() local 185 if (adapter->tx_ring == NULL) in netxen_free_sw_resources() 188 tx_ring = adapter->tx_ring; in netxen_free_sw_resources() 189 vfree(tx_ring->cmd_buf_arr); in netxen_free_sw_resources() 190 kfree(tx_ring); in netxen_free_sw_resources() 191 adapter->tx_ring = NULL; in netxen_free_sw_resources() 199 struct nx_host_tx_ring *tx_ring; in netxen_alloc_sw_resources() local [all …]
|
D | netxen_nic_main.c | 123 struct nx_host_tx_ring *tx_ring) in netxen_nic_update_cmd_producer() argument 125 NXWRIO(adapter, tx_ring->crb_cmd_producer, tx_ring->producer); in netxen_nic_update_cmd_producer() 135 struct nx_host_tx_ring *tx_ring) in netxen_nic_update_cmd_consumer() argument 137 NXWRIO(adapter, tx_ring->crb_cmd_consumer, tx_ring->sw_consumer); in netxen_nic_update_cmd_consumer() 1225 struct nx_host_tx_ring *tx_ring; in netxen_nic_attach() local 1261 tx_ring = adapter->tx_ring; in netxen_nic_attach() 1262 tx_ring->crb_cmd_producer = netxen_get_ioaddr(adapter, in netxen_nic_attach() 1264 tx_ring->crb_cmd_consumer = netxen_get_ioaddr(adapter, in netxen_nic_attach() 1267 tx_ring->producer = 0; in netxen_nic_attach() 1268 tx_ring->sw_consumer = 0; in netxen_nic_attach() [all …]
|
D | netxen_nic_hw.c | 585 struct nx_host_tx_ring *tx_ring; in netxen_send_cmd_descs() local 592 tx_ring = adapter->tx_ring; in netxen_send_cmd_descs() 593 __netif_tx_lock_bh(tx_ring->txq); in netxen_send_cmd_descs() 595 producer = tx_ring->producer; in netxen_send_cmd_descs() 596 consumer = tx_ring->sw_consumer; in netxen_send_cmd_descs() 598 if (nr_desc >= netxen_tx_avail(tx_ring)) { in netxen_send_cmd_descs() 599 netif_tx_stop_queue(tx_ring->txq); in netxen_send_cmd_descs() 601 if (netxen_tx_avail(tx_ring) > nr_desc) { in netxen_send_cmd_descs() 602 if (netxen_tx_avail(tx_ring) > TX_STOP_THRESH) in netxen_send_cmd_descs() 603 netif_tx_wake_queue(tx_ring->txq); in netxen_send_cmd_descs() [all …]
|
D | netxen_nic.h | 81 #define TX_BUFF_RINGSIZE(tx_ring) \ argument 82 (sizeof(struct netxen_cmd_buffer) * tx_ring->num_desc) 83 #define TX_DESC_RINGSIZE(tx_ring) \ argument 84 (sizeof(struct cmd_desc_type0) * tx_ring->num_desc) 1657 struct nx_host_tx_ring *tx_ring; member 1806 struct nx_host_tx_ring *tx_ring); 1874 static inline u32 netxen_tx_avail(struct nx_host_tx_ring *tx_ring) in netxen_tx_avail() argument 1877 return find_diff_among(tx_ring->producer, in netxen_tx_avail() 1878 tx_ring->sw_consumer, tx_ring->num_desc); in netxen_tx_avail()
|
D | netxen_nic_ethtool.c | 332 regs_buff[i++] = le32_to_cpu(*(adapter->tx_ring->hw_consumer)); in netxen_nic_get_regs() 344 adapter->tx_ring->crb_cmd_consumer); in netxen_nic_get_regs() 347 regs_buff[i++] = NXRDIO(adapter, adapter->tx_ring->crb_cmd_producer); in netxen_nic_get_regs()
|
/linux-4.1.27/drivers/net/ethernet/qlogic/qlge/ |
D | qlge_main.c | 2100 struct tx_ring *tx_ring; in ql_process_mac_tx_intr() local 2104 tx_ring = &qdev->tx_ring[mac_rsp->txq_idx]; in ql_process_mac_tx_intr() 2105 tx_ring_desc = &tx_ring->q[mac_rsp->tid]; in ql_process_mac_tx_intr() 2107 tx_ring->tx_bytes += (tx_ring_desc->skb)->len; in ql_process_mac_tx_intr() 2108 tx_ring->tx_packets++; in ql_process_mac_tx_intr() 2133 atomic_inc(&tx_ring->tx_count); in ql_process_mac_tx_intr() 2202 struct tx_ring *tx_ring; in ql_clean_outbound_rx_ring() local 2230 tx_ring = &qdev->tx_ring[net_rsp->txq_idx]; in ql_clean_outbound_rx_ring() 2231 if (__netif_subqueue_stopped(qdev->ndev, tx_ring->wq_id)) { in ql_clean_outbound_rx_ring() 2232 if ((atomic_read(&tx_ring->tx_count) > (tx_ring->wq_len / 4))) in ql_clean_outbound_rx_ring() [all …]
|
D | qlge_dbg.c | 1644 DUMP_QDEV_FIELD(qdev, "%p", tx_ring); in ql_dump_qdev() 1669 void ql_dump_tx_ring(struct tx_ring *tx_ring) in ql_dump_tx_ring() argument 1671 if (tx_ring == NULL) in ql_dump_tx_ring() 1674 tx_ring->wq_id); in ql_dump_tx_ring() 1675 pr_err("tx_ring->base = %p\n", tx_ring->wq_base); in ql_dump_tx_ring() 1677 (unsigned long long) tx_ring->wq_base_dma); in ql_dump_tx_ring() 1679 tx_ring->cnsmr_idx_sh_reg, in ql_dump_tx_ring() 1680 tx_ring->cnsmr_idx_sh_reg in ql_dump_tx_ring() 1681 ? ql_read_sh_reg(tx_ring->cnsmr_idx_sh_reg) : 0); in ql_dump_tx_ring() 1682 pr_err("tx_ring->size = %d\n", tx_ring->wq_size); in ql_dump_tx_ring() [all …]
|
D | qlge.h | 1382 struct tx_ring { struct 2116 struct tx_ring tx_ring[MAX_TX_RINGS]; member 2292 void ql_dump_tx_ring(struct tx_ring *tx_ring); 2299 #define QL_DUMP_TX_RING(tx_ring) ql_dump_tx_ring(tx_ring) argument 2307 #define QL_DUMP_TX_RING(tx_ring) argument
|
/linux-4.1.27/drivers/net/vmxnet3/ |
D | vmxnet3_drv.c | 336 BUG_ON(tq->buf_info[eop_idx].sop_idx != tq->tx_ring.next2comp); in vmxnet3_unmap_pkt() 337 BUG_ON(VMXNET3_TXDESC_GET_EOP(&(tq->tx_ring.base[eop_idx].txd)) != 1); in vmxnet3_unmap_pkt() 343 VMXNET3_INC_RING_IDX_ONLY(eop_idx, tq->tx_ring.size); in vmxnet3_unmap_pkt() 345 while (tq->tx_ring.next2comp != eop_idx) { in vmxnet3_unmap_pkt() 346 vmxnet3_unmap_tx_buf(tq->buf_info + tq->tx_ring.next2comp, in vmxnet3_unmap_pkt() 354 vmxnet3_cmd_ring_adv_next2comp(&tq->tx_ring); in vmxnet3_unmap_pkt() 383 vmxnet3_cmd_ring_desc_avail(&tq->tx_ring) > in vmxnet3_tq_tx_complete() 400 while (tq->tx_ring.next2comp != tq->tx_ring.next2fill) { in vmxnet3_tq_cleanup() 403 tbi = tq->buf_info + tq->tx_ring.next2comp; in vmxnet3_tq_cleanup() 410 vmxnet3_cmd_ring_adv_next2comp(&tq->tx_ring); in vmxnet3_tq_cleanup() [all …]
|
D | vmxnet3_ethtool.c | 360 buf[j++] = adapter->tx_queue[i].tx_ring.next2fill; in vmxnet3_get_regs() 361 buf[j++] = adapter->tx_queue[i].tx_ring.next2comp; in vmxnet3_get_regs() 362 buf[j++] = adapter->tx_queue[i].tx_ring.gen; in vmxnet3_get_regs()
|
D | vmxnet3_int.h | 229 struct vmxnet3_cmd_ring tx_ring; member
|
/linux-4.1.27/drivers/net/ethernet/amd/ |
D | ariadne.c | 88 volatile struct TDRE *tx_ring[TX_RING_SIZE]; member 100 struct TDRE tx_ring[TX_RING_SIZE]; member 130 volatile struct TDRE *t = &lancedata->tx_ring[i]; in ariadne_init_ring() 138 priv->tx_ring[i] = &lancedata->tx_ring[i]; in ariadne_init_ring() 141 i, &lancedata->tx_ring[i], lancedata->tx_buff[i]); in ariadne_init_ring() 310 int status = lowb(priv->tx_ring[entry]->TMD1); in ariadne_interrupt() 315 priv->tx_ring[entry]->TMD1 &= 0xff00; in ariadne_interrupt() 319 int err_status = priv->tx_ring[entry]->TMD3; in ariadne_interrupt() 456 lance->RDP = swloww(ARIADNE_RAM + offsetof(struct lancedata, tx_ring)); in ariadne_open() 458 lance->RDP = swhighw(ARIADNE_RAM + offsetof(struct lancedata, tx_ring)); in ariadne_open() [all …]
|
D | lance.c | 232 u32 tx_ring; member 238 struct lance_tx_head tx_ring[TX_RING_SIZE]; member 576 lp->init_block.tx_ring = ((u32)isa_virt_to_bus(lp->tx_ring) & 0xffffff) | TX_RING_LEN_BITS; in lance_probe1() 789 (u32) isa_virt_to_bus(lp->tx_ring), in lance_open() 890 lp->tx_ring[i].base = 0; in lance_init_ring() 899 lp->init_block.tx_ring = ((u32)isa_virt_to_bus(lp->tx_ring) & 0xffffff) | TX_RING_LEN_BITS; in lance_init_ring() 939 lp->tx_ring[i].base, -lp->tx_ring[i].length, in lance_tx_timeout() 940 lp->tx_ring[i].misc); in lance_tx_timeout() 981 lp->tx_ring[entry].length = -ETH_ZLEN; in lance_start_xmit() 984 lp->tx_ring[entry].length = -skb->len; in lance_start_xmit() [all …]
|
D | pcnet32.c | 241 __le32 tx_ring; member 263 struct pcnet32_tx_head *tx_ring; member 507 lp->tx_ring, lp->tx_ring_dma_addr); in pcnet32_realloc_tx_ring() 512 lp->tx_ring = new_tx_ring; in pcnet32_realloc_tx_ring() 922 lp->tx_ring[x].length = cpu_to_le16(-skb->len); in pcnet32_loopback_test() 923 lp->tx_ring[x].misc = 0; in pcnet32_loopback_test() 948 lp->tx_ring[x].base = cpu_to_le32(lp->tx_dma_addr[x]); in pcnet32_loopback_test() 950 lp->tx_ring[x].status = cpu_to_le16(status); in pcnet32_loopback_test() 1254 int status = (short)le16_to_cpu(lp->tx_ring[entry].status); in pcnet32_tx() 1259 lp->tx_ring[entry].base = 0; in pcnet32_tx() [all …]
|
D | amd8111e.c | 305 if((lp->tx_ring = pci_alloc_consistent(lp->pci_dev, in amd8111e_init_ring() 346 lp->tx_ring[i].buff_phy_addr = 0; in amd8111e_init_ring() 347 lp->tx_ring[i].tx_flags = 0; in amd8111e_init_ring() 348 lp->tx_ring[i].buff_count = 0; in amd8111e_init_ring() 362 sizeof(struct amd8111e_tx_dr)*NUM_TX_RING_DR,lp->tx_ring, in amd8111e_init_ring() 635 if(lp->tx_ring){ in amd8111e_free_ring() 638 lp->tx_ring, lp->tx_ring_dma_addr); in amd8111e_free_ring() 640 lp->tx_ring = NULL; in amd8111e_free_ring() 657 status = le16_to_cpu(lp->tx_ring[tx_index].tx_flags); in amd8111e_tx() 662 lp->tx_ring[tx_index].buff_phy_addr = 0; in amd8111e_tx() [all …]
|
D | atarilance.c | 151 struct ringdesc tx_ring; member 620 MEM->init.tx_ring.adr_lo = offsetof( struct lance_memory, tx_head ); in lance_probe1() 621 MEM->init.tx_ring.adr_hi = 0; in lance_probe1() 622 MEM->init.tx_ring.len = TX_RING_LEN_BITS; in lance_probe1()
|
D | amd8111e.h | 747 struct amd8111e_tx_dr* tx_ring; member
|
/linux-4.1.27/drivers/infiniband/hw/mlx4/ |
D | mad.c | 536 tun_mad = (struct mlx4_rcv_tunnel_mad *) (tun_qp->tx_ring[tun_tx_ix].buf.addr); in mlx4_ib_send_to_slave() 537 if (tun_qp->tx_ring[tun_tx_ix].ah) in mlx4_ib_send_to_slave() 538 ib_destroy_ah(tun_qp->tx_ring[tun_tx_ix].ah); in mlx4_ib_send_to_slave() 539 tun_qp->tx_ring[tun_tx_ix].ah = ah; in mlx4_ib_send_to_slave() 541 tun_qp->tx_ring[tun_tx_ix].buf.map, in mlx4_ib_send_to_slave() 583 tun_qp->tx_ring[tun_tx_ix].buf.map, in mlx4_ib_send_to_slave() 587 list.addr = tun_qp->tx_ring[tun_tx_ix].buf.map; in mlx4_ib_send_to_slave() 1223 sqp_mad = (struct mlx4_mad_snd_buf *) (sqp->tx_ring[wire_tx_ix].buf.addr); in mlx4_ib_send_to_wire() 1224 if (sqp->tx_ring[wire_tx_ix].ah) in mlx4_ib_send_to_wire() 1225 ib_destroy_ah(sqp->tx_ring[wire_tx_ix].ah); in mlx4_ib_send_to_wire() [all …]
|
D | mlx4_ib.h | 396 struct mlx4_ib_tun_tx_buf *tx_ring; member
|
/linux-4.1.27/drivers/net/ethernet/dlink/ |
D | sundance.c | 372 struct netdev_desc *tx_ring; member 562 np->tx_ring = (struct netdev_desc *)ring_space; in sundance_probe1() 704 np->tx_ring, np->tx_ring_dma); in sundance_probe1() 990 (unsigned long long)(np->tx_ring_dma + i*sizeof(*np->tx_ring)), in tx_timeout() 991 le32_to_cpu(np->tx_ring[i].next_desc), in tx_timeout() 992 le32_to_cpu(np->tx_ring[i].status), in tx_timeout() 993 (le32_to_cpu(np->tx_ring[i].status) >> 2) & 0xff, in tx_timeout() 994 le32_to_cpu(np->tx_ring[i].frag[0].addr), in tx_timeout() 995 le32_to_cpu(np->tx_ring[i].frag[0].length)); in tx_timeout() 1068 np->tx_ring[i].status = 0; in init_ring() [all …]
|
D | dl2k.c | 239 np->tx_ring = ring_space; in rio_probe1() 297 pci_free_consistent (pdev, TX_TOTAL_SIZE, np->tx_ring, np->tx_ring_dma); in rio_probe1() 560 np->tx_ring[i].status = cpu_to_le64 (TFDDone); in alloc_list() 561 np->tx_ring[i].next_desc = cpu_to_le64 (np->tx_ring_dma + in alloc_list() 614 txdesc = &np->tx_ring[entry]; in start_xmit() 729 if (!(np->tx_ring[entry].status & cpu_to_le64(TFDDone))) in rio_free_tx() 733 desc_to_dma(&np->tx_ring[entry]), in rio_free_tx() 1718 pci_unmap_single(pdev, desc_to_dma(&np->tx_ring[i]), in rio_close() 1739 pci_free_consistent (pdev, TX_TOTAL_SIZE, np->tx_ring, in rio_remove1()
|
D | dl2k.h | 363 struct netdev_desc *tx_ring; member
|
/linux-4.1.27/drivers/net/ethernet/ |
D | fealnx.c | 378 struct fealnx_desc *tx_ring; member 574 np->tx_ring = ring_space; in fealnx_init_one() 674 pci_free_consistent(pdev, TX_TOTAL_SIZE, np->tx_ring, np->tx_ring_dma); in fealnx_init_one() 694 pci_free_consistent(pdev, TX_TOTAL_SIZE, np->tx_ring, in fealnx_remove_one() 1157 iowrite32(np->tx_ring_dma + ((char*)np->cur_tx - (char*)np->tx_ring), in enable_rxtx() 1216 printk(KERN_DEBUG " Tx ring %p: ", np->tx_ring); in fealnx_tx_timeout() 1218 printk(KERN_CONT " %4.4x", np->tx_ring[i].status); in fealnx_tx_timeout() 1280 np->cur_tx = &np->tx_ring[0]; in init_ring() 1281 np->cur_tx_copy = &np->tx_ring[0]; in init_ring() 1286 np->tx_ring[i].status = 0; in init_ring() [all …]
|
/linux-4.1.27/drivers/net/irda/ |
D | au1k_ir.c | 154 volatile struct ring_dest *tx_ring[NUM_IR_DESC]; member 257 aup->tx_ring[i] = (volatile struct ring_dest *) in setup_hw_rings() 312 ptxd = aup->tx_ring[i]; in au1k_irda_set_speed() 422 ptxd = aup->tx_ring[aup->tx_tail]; in au1k_tx_ack() 430 ptxd = aup->tx_ring[aup->tx_tail]; in au1k_tx_ack() 680 ptxd = aup->tx_ring[aup->tx_head]; in au1k_irda_hard_xmit() 861 aup->tx_ring[i]->addr_0 = (u8)(pDB->dma_addr & 0xff); in au1k_irda_net_init() 862 aup->tx_ring[i]->addr_1 = (u8)((pDB->dma_addr >> 8) & 0xff); in au1k_irda_net_init() 863 aup->tx_ring[i]->addr_2 = (u8)((pDB->dma_addr >> 16) & 0xff); in au1k_irda_net_init() 864 aup->tx_ring[i]->addr_3 = (u8)((pDB->dma_addr >> 24) & 0xff); in au1k_irda_net_init() [all …]
|
D | vlsi_ir.c | 357 if (netif_running(ndev) && idev->rx_ring && idev->tx_ring) { in vlsi_seq_show() 361 vlsi_proc_ring(seq, idev->tx_ring); in vlsi_seq_show() 488 idev->tx_ring = vlsi_alloc_ring(idev->pdev, hwmap, ringsize[0], in vlsi_create_hwif() 490 if (idev->tx_ring == NULL) in vlsi_create_hwif() 499 idev->rx_ring = idev->tx_ring = NULL; in vlsi_create_hwif() 509 vlsi_free_ring(idev->tx_ring); in vlsi_destroy_hwif() 510 idev->rx_ring = idev->tx_ring = NULL; in vlsi_destroy_hwif() 847 struct vlsi_ring *r = idev->tx_ring; in vlsi_hard_start_xmit() 891 if (ring_first(idev->tx_ring) == NULL) { in vlsi_hard_start_xmit() 1042 struct vlsi_ring *r = idev->tx_ring; in vlsi_tx_interrupt() [all …]
|
D | vlsi_ir.h | 724 struct vlsi_ring *tx_ring, *rx_ring; member
|
/linux-4.1.27/drivers/net/ethernet/qualcomm/ |
D | qca_spi.h | 56 struct tx_ring { struct 83 struct tx_ring txr; argument
|
D | qca_spi.c | 400 qcaspi_tx_ring_has_space(struct tx_ring *txr) in qcaspi_tx_ring_has_space()
|
/linux-4.1.27/drivers/net/hippi/ |
D | rrunner.c | 159 rrpriv->tx_ring = tmpptr; in rr_init_one() 210 if (rrpriv->tx_ring) in rr_init_one() 211 pci_free_consistent(pdev, TX_TOTAL_SIZE, rrpriv->tx_ring, in rr_init_one() 239 pci_free_consistent(pdev, TX_TOTAL_SIZE, rr->tx_ring, in rr_remove_one() 601 rrpriv->tx_ring[i].size = 0; in rr_init1() 602 set_rraddr(&rrpriv->tx_ring[i].addr, 0); in rr_init1() 1069 desc = &(rrpriv->tx_ring[txcon]); in rr_interrupt() 1082 set_rraddr(&rrpriv->tx_ring[txcon].addr, 0); in rr_interrupt() 1115 struct tx_desc *desc = &(rrpriv->tx_ring[i]); in rr_raz_tx() 1291 …i is valid - dumping data (0x%x bytes - DMA len 0x%x)\n", index, len, rrpriv->tx_ring[index].size); in rr_dump() [all …]
|
D | rrunner.h | 802 struct tx_desc *tx_ring; member
|
/linux-4.1.27/drivers/net/ethernet/smsc/ |
D | smsc9420.c | 62 struct smsc9420_dma_desc *tx_ring; member 548 BUG_ON(!pd->tx_ring); in smsc9420_free_tx_ring() 563 pd->tx_ring[i].status = 0; in smsc9420_free_tx_ring() 564 pd->tx_ring[i].length = 0; in smsc9420_free_tx_ring() 565 pd->tx_ring[i].buffer1 = 0; in smsc9420_free_tx_ring() 566 pd->tx_ring[i].buffer2 = 0; in smsc9420_free_tx_ring() 951 status = pd->tx_ring[index].status; in smsc9420_complete_tx() 952 length = pd->tx_ring[index].length; in smsc9420_complete_tx() 970 pd->tx_ring[index].buffer1 = 0; in smsc9420_complete_tx() 990 BUG_ON(pd->tx_ring[index].status & TDES0_OWN_); in smsc9420_hard_start_xmit() [all …]
|
D | epic100.c | 254 struct epic_tx_desc *tx_ring; member 383 ep->tx_ring = ring_space; in epic_init_one() 502 pci_free_consistent(pdev, TX_TOTAL_SIZE, ep->tx_ring, ep->tx_ring_dma); in epic_init_one() 937 ep->tx_ring[i].txstatus = 0x0000; in epic_init_ring() 938 ep->tx_ring[i].next = ep->tx_ring_dma + in epic_init_ring() 941 ep->tx_ring[i-1].next = ep->tx_ring_dma; in epic_init_ring() 964 ep->tx_ring[entry].bufaddr = pci_map_single(ep->pci_dev, skb->data, in epic_start_xmit() 977 ep->tx_ring[entry].buflength = ctrl_word | skb->len; in epic_start_xmit() 978 ep->tx_ring[entry].txstatus = in epic_start_xmit() 1031 int txstatus = ep->tx_ring[entry].txstatus; in epic_tx() [all …]
|
/linux-4.1.27/drivers/net/ethernet/samsung/sxgbe/ |
D | sxgbe_main.c | 395 struct sxgbe_tx_queue *tx_ring, int tx_rsize) in init_tx_ring() argument 398 if (!tx_ring) { in init_tx_ring() 404 tx_ring->dma_tx = dma_zalloc_coherent(dev, in init_tx_ring() 406 &tx_ring->dma_tx_phy, GFP_KERNEL); in init_tx_ring() 407 if (!tx_ring->dma_tx) in init_tx_ring() 411 tx_ring->tx_skbuff_dma = devm_kcalloc(dev, tx_rsize, in init_tx_ring() 413 if (!tx_ring->tx_skbuff_dma) in init_tx_ring() 416 tx_ring->tx_skbuff = devm_kcalloc(dev, tx_rsize, in init_tx_ring() 419 if (!tx_ring->tx_skbuff) in init_tx_ring() 423 tx_ring->queue_no = queue_no; in init_tx_ring() [all …]
|
/linux-4.1.27/net/packet/ |
D | af_packet.c | 169 int closing, int tx_ring); 521 int tx_ring, in prb_shutdown_retire_blk_timer() argument 526 pkc = tx_ring ? GET_PBDQC_FROM_RB(&po->tx_ring) : in prb_shutdown_retire_blk_timer() 546 static void prb_setup_retire_blk_timer(struct packet_sock *po, int tx_ring) in prb_setup_retire_blk_timer() argument 550 if (tx_ring) in prb_setup_retire_blk_timer() 553 pkc = tx_ring ? GET_PBDQC_FROM_RB(&po->tx_ring) : in prb_setup_retire_blk_timer() 610 union tpacket_req_u *req_u, int tx_ring) in init_prb_bdqc() argument 637 prb_setup_retire_blk_timer(po, tx_ring); in init_prb_bdqc() 1225 po->tx_ring.pending_refcnt = alloc_percpu(unsigned int); in packet_alloc_pending() 1226 if (unlikely(po->tx_ring.pending_refcnt == NULL)) in packet_alloc_pending() [all …]
|
D | internal.h | 97 struct packet_ring_buffer tx_ring; member
|
D | diag.c | 106 ret = pdiag_put_ring(&po->tx_ring, po->tp_version, in pdiag_put_rings_cfg()
|
/linux-4.1.27/drivers/net/ethernet/3com/ |
D | 3c59x.c | 598 struct boom_tx_desc* tx_ring; member 1226 vp->tx_ring = (struct boom_tx_desc *)(vp->rx_ring + RX_RING_SIZE); in vortex_probe1() 2129 struct boom_tx_desc *prev_entry = &vp->tx_ring[(vp->cur_tx-1) % TX_RING_SIZE]; in boomerang_start_xmit() 2158 vp->tx_ring[entry].next = 0; in boomerang_start_xmit() 2161 vp->tx_ring[entry].status = cpu_to_le32(skb->len | TxIntrUploaded); in boomerang_start_xmit() 2163 vp->tx_ring[entry].status = cpu_to_le32(skb->len | TxIntrUploaded | AddTCPChksum | AddUDPChksum); in boomerang_start_xmit() 2171 vp->tx_ring[entry].frag[0].addr = cpu_to_le32(dma_addr); in boomerang_start_xmit() 2172 vp->tx_ring[entry].frag[0].length = cpu_to_le32(skb->len | LAST_FRAG); in boomerang_start_xmit() 2181 vp->tx_ring[entry].frag[0].addr = cpu_to_le32(dma_addr); in boomerang_start_xmit() 2182 vp->tx_ring[entry].frag[0].length = cpu_to_le32(skb_headlen(skb)); in boomerang_start_xmit() [all …]
|
D | 3c515.c | 306 struct boom_tx_desc tx_ring[TX_RING_SIZE]; member 982 &vp->tx_ring[0]); in corkscrew_timeout() 985 &vp->tx_ring[i], in corkscrew_timeout() 986 vp->tx_ring[i].length, vp->tx_ring[i].status); in corkscrew_timeout() 1021 prev_entry = &vp->tx_ring[(vp->cur_tx - 1) % TX_RING_SIZE]; in corkscrew_start_xmit() 1029 vp->tx_ring[entry].next = 0; in corkscrew_start_xmit() 1030 vp->tx_ring[entry].addr = isa_virt_to_bus(skb->data); in corkscrew_start_xmit() 1031 vp->tx_ring[entry].length = skb->len | 0x80000000; in corkscrew_start_xmit() 1032 vp->tx_ring[entry].status = skb->len | 0x80000000; in corkscrew_start_xmit() 1041 prev_entry->next = isa_virt_to_bus(&vp->tx_ring[entry]); in corkscrew_start_xmit() [all …]
|
/linux-4.1.27/drivers/net/ethernet/freescale/fs_enet/ |
D | fs_enet-main.c | 239 if (fep->tx_free == fep->tx_ring) in fs_enet_tx_napi() 402 fep->tx_free = fep->tx_ring; in fs_init_bds() 434 for (i = 0, bdp = fep->tx_bd_base; i < fep->tx_ring; i++, bdp++) { in fs_init_bds() 438 CBDW_SC(bdp, (i < fep->tx_ring - 1) ? 0 : BD_SC_WRAP); in fs_init_bds() 452 for (i = 0, bdp = fep->tx_bd_base; i < fep->tx_ring; i++, bdp++) { in fs_cleanup_bds() 950 fpi->tx_ring = 64; in fs_enet_probe() 988 (fpi->rx_ring + fpi->tx_ring) + in fs_enet_probe() 989 sizeof(char) * fpi->tx_ring; in fs_enet_probe() 1013 fpi->tx_ring); in fs_enet_probe() 1029 fep->tx_ring = fpi->tx_ring; in fs_enet_probe()
|
D | mac-fcc.c | 155 (fpi->tx_ring + fpi->rx_ring) * in allocate_bd() 171 (fpi->tx_ring + fpi->rx_ring) * sizeof(cbd_t), in free_bd() 555 last_tx_bd = fep->tx_bd_base + (fpi->tx_ring * sizeof(cbd_t)); in tx_restart()
|
D | mac-fec.c | 144 (fpi->tx_ring + fpi->rx_ring) * in allocate_bd() 159 dma_free_coherent(fep->dev, (fpi->tx_ring + fpi->rx_ring) in free_bd()
|
D | fs_enet.h | 132 int rx_ring, tx_ring; member
|
D | mac-scc.c | 147 fep->ring_mem_addr = cpm_dpalloc((fpi->tx_ring + fpi->rx_ring) * in allocate_bd()
|
/linux-4.1.27/drivers/net/ethernet/dec/tulip/ |
D | winbond-840.c | 297 struct w840_tx_desc *tx_ring; member 801 np->tx_ring = (struct w840_tx_desc*)&np->rx_ring[RX_RING_SIZE]; in init_rxtx_rings() 831 np->tx_ring[i].status = 0; in init_rxtx_rings() 942 printk(KERN_DEBUG " Tx ring %p: ", np->tx_ring); in tx_timeout() 944 printk(KERN_CONT " %08x", np->tx_ring[i].status); in tx_timeout() 1014 np->tx_ring[entry].buffer1 = np->tx_addr[entry]; in start_tx() 1016 np->tx_ring[entry].length = DescWholePkt | skb->len; in start_tx() 1020 np->tx_ring[entry].buffer2 = np->tx_addr[entry]+TX_BUFLIMIT; in start_tx() 1021 np->tx_ring[entry].length = DescWholePkt | (len << 11) | TX_BUFLIMIT; in start_tx() 1024 np->tx_ring[entry].length |= DescEndRing; in start_tx() [all …]
|
D | tulip_core.c | 372 tp->tx_ring[tp->cur_tx].length = cpu_to_le32(0x08000000 | 192); in tulip_up() 373 tp->tx_ring[tp->cur_tx].buffer1 = cpu_to_le32(mapping); in tulip_up() 374 tp->tx_ring[tp->cur_tx].status = cpu_to_le32(DescOwned); in tulip_up() 600 printk(KERN_DEBUG " Tx ring %p: ", tp->tx_ring); in tulip_tx_timeout() 602 pr_cont(" %08x", (unsigned int)tp->tx_ring[i].status); in tulip_tx_timeout() 660 tp->tx_ring[i].status = 0x00000000; in tulip_init_ring() 661 tp->tx_ring[i].buffer2 = cpu_to_le32(tp->tx_ring_dma + sizeof(struct tulip_tx_desc) * (i + 1)); in tulip_init_ring() 663 tp->tx_ring[i-1].buffer2 = cpu_to_le32(tp->tx_ring_dma); in tulip_init_ring() 684 tp->tx_ring[entry].buffer1 = cpu_to_le32(mapping); in tulip_start_xmit() 699 tp->tx_ring[entry].length = cpu_to_le32(skb->len | flag); in tulip_start_xmit() [all …]
|
D | de2104x.c | 307 struct de_desc *tx_ring; member 552 status = le32_to_cpu(de->tx_ring[tx_tail].opts1); in de_tx() 625 txd = &de->tx_ring[entry]; in de_start_xmit() 759 dummy_txd = &de->tx_ring[entry]; in __de_set_rx_mode() 775 txd = &de->tx_ring[entry]; in __de_set_rx_mode() 1310 memset(de->tx_ring, 0, sizeof(struct de_desc) * DE_TX_RING_SIZE); in de_init_rings() 1311 de->tx_ring[DE_TX_RING_SIZE - 1].opts2 = cpu_to_le32(RingEnd); in de_init_rings() 1324 de->tx_ring = &de->rx_ring[DE_RX_RING_SIZE]; in de_alloc_rings() 1335 memset(de->tx_ring, 0, sizeof(struct de_desc) * DE_TX_RING_SIZE); in de_clean_rings() 1336 de->tx_ring[DE_TX_RING_SIZE - 1].opts2 = cpu_to_le32(RingEnd); in de_clean_rings() [all …]
|
D | de4x5.c | 791 struct de4x5_desc *tx_ring; /* TX descriptor ring */ member 1181 lp->tx_ring = lp->rx_ring + NUM_RX_DESC; in de4x5_hw_init() 1224 lp->tx_ring[lp->txRingSize - 1].des1 |= cpu_to_le32(TD_TER); in de4x5_hw_init() 1425 lp->tx_ring[i].status = cpu_to_le32(0); in de4x5_sw_reset() 1440 if ((s32)le32_to_cpu(lp->tx_ring[lp->tx_new].status) >= 0) j=1; in de4x5_sw_reset() 1679 dma_unmap_single(lp->gendev, le32_to_cpu(lp->tx_ring[entry].buf), in de4x5_free_tx_buff() 1680 le32_to_cpu(lp->tx_ring[entry].des1) & TD_TBS1, in de4x5_free_tx_buff() 1699 status = (s32)le32_to_cpu(lp->tx_ring[entry].status); in de4x5_tx() 1901 lp->tx_ring[lp->tx_new].buf = cpu_to_le32(buf_dma); in load_packet() 1902 lp->tx_ring[lp->tx_new].des1 &= cpu_to_le32(TD_TER); in load_packet() [all …]
|
D | tulip.h | 405 struct tulip_tx_desc *tx_ring; member
|
D | interrupt.c | 591 int status = le32_to_cpu(tp->tx_ring[entry].status); in tulip_interrupt()
|
/linux-4.1.27/drivers/net/ethernet/micrel/ |
D | ks8695net.c | 175 struct tx_ring_desc *tx_ring; member 379 !(ksp->tx_ring[buff_n].owner & cpu_to_le32(TDES_OWN))) { in ks8695_tx_irq() 387 ksp->tx_ring[buff_n].data_ptr = 0; in ks8695_tx_irq() 687 ksp->tx_ring[buff_n].owner = 0; in ks8695_shutdown() 688 ksp->tx_ring[buff_n].status = 0; in ks8695_shutdown() 689 ksp->tx_ring[buff_n].data_ptr = 0; in ks8695_shutdown() 1193 ksp->tx_ring[buff_n].data_ptr = in ks8695_start_xmit() 1195 ksp->tx_ring[buff_n].status = in ks8695_start_xmit() 1202 ksp->tx_ring[buff_n].owner = cpu_to_le32(TDES_OWN); in ks8695_start_xmit() 1473 ksp->tx_ring = ksp->ring_base; in ks8695_probe() [all …]
|
/linux-4.1.27/drivers/net/ethernet/sgi/ |
D | meth.c | 78 tx_packet *tx_ring; member 214 priv->tx_ring = dma_zalloc_coherent(NULL, TX_RING_BUFFER_SIZE, in meth_init_tx_ring() 216 if (!priv->tx_ring) in meth_init_tx_ring() 256 dma_free_coherent(NULL, TX_RING_BUFFER_SIZE, priv->tx_ring, in meth_free_tx_ring() 492 status = priv->tx_ring[priv->tx_read].header.raw; in meth_tx_cleanup() 526 priv->tx_ring[priv->tx_read].header.raw = 0; in meth_tx_cleanup() 611 tx_packet *desc = &priv->tx_ring[priv->tx_write]; in meth_tx_short_prepare() 624 tx_packet *desc = &priv->tx_ring[priv->tx_write]; in meth_tx_1page_prepare() 649 tx_packet *desc = &priv->tx_ring[priv->tx_write]; in meth_tx_2page_prepare()
|
/linux-4.1.27/drivers/net/ethernet/amd/xgbe/ |
D | xgbe-drv.c | 136 struct xgbe_ring *tx_ring, *rx_ring; in xgbe_alloc_channels() local 146 tx_ring = kcalloc(pdata->tx_ring_count, sizeof(struct xgbe_ring), in xgbe_alloc_channels() 148 if (!tx_ring) in xgbe_alloc_channels() 177 spin_lock_init(&tx_ring->lock); in xgbe_alloc_channels() 178 channel->tx_ring = tx_ring++; in xgbe_alloc_channels() 188 channel->dma_irq, channel->tx_ring, channel->rx_ring); in xgbe_alloc_channels() 200 kfree(tx_ring); in xgbe_alloc_channels() 215 kfree(pdata->channel->tx_ring); in xgbe_free_channels() 281 if (channel->tx_ring && channel->rx_ring) in xgbe_enable_rx_tx_ints() 283 else if (channel->tx_ring) in xgbe_enable_rx_tx_ints() [all …]
|
D | xgbe-desc.c | 181 xgbe_free_ring(pdata, channel->tx_ring); in xgbe_free_ring_resources() 230 ret = xgbe_init_ring(pdata, channel->tx_ring, in xgbe_alloc_ring_resources() 362 ring = channel->tx_ring; in xgbe_wrapper_tx_descriptor_init() 490 struct xgbe_ring *ring = channel->tx_ring; in xgbe_map_tx_skb()
|
D | xgbe-dev.c | 197 if (!channel->tx_ring) in xgbe_config_tx_pbl_val() 236 if (!channel->tx_ring) in xgbe_config_osp_mode() 332 if (!channel->tx_ring) in xgbe_config_tso_mode() 622 if (channel->tx_ring) { in xgbe_enable_dma_interrupts() 1092 struct xgbe_ring *ring = channel->tx_ring; in xgbe_tx_desc_init() 1403 struct xgbe_ring *ring = channel->tx_ring; in xgbe_dev_xmit() 2601 if (!channel->tx_ring) in xgbe_enable_tx() 2624 if (!channel->tx_ring) in xgbe_disable_tx() 2640 if (!channel->tx_ring) in xgbe_disable_tx() 2706 if (!channel->tx_ring) in xgbe_powerup_tx() [all …]
|
D | xgbe.h | 413 struct xgbe_ring *tx_ring; member
|
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/ |
D | en_netdev.c | 1299 i, priv->tx_ring[i]->qpn, priv->tx_ring[i]->cqn, in mlx4_en_tx_timeout() 1300 priv->tx_ring[i]->cons, priv->tx_ring[i]->prod); in mlx4_en_tx_timeout() 1523 struct mlx4_en_tx_ring *tx_ring; in mlx4_en_start_port() local 1629 tx_ring = priv->tx_ring[i]; in mlx4_en_start_port() 1630 err = mlx4_en_activate_tx_ring(priv, tx_ring, cq->mcq.cqn, in mlx4_en_start_port() 1637 tx_ring->tx_queue = netdev_get_tx_queue(dev, i); in mlx4_en_start_port() 1643 for (j = 0; j < tx_ring->buf_size; j += STAMP_STRIDE) in mlx4_en_start_port() 1644 *((u32 *) (tx_ring->buf + j)) = 0xffffffff; in mlx4_en_start_port() 1710 mlx4_en_deactivate_tx_ring(priv, priv->tx_ring[tx_index]); in mlx4_en_start_port() 1822 mlx4_en_deactivate_tx_ring(priv, priv->tx_ring[i]); in mlx4_en_stop_port() [all …]
|
D | en_ethtool.c | 396 data[index++] = priv->tx_ring[i]->packets; in mlx4_en_get_ethtool_stats() 397 data[index++] = priv->tx_ring[i]->bytes; in mlx4_en_get_ethtool_stats() 1043 tx_size == priv->tx_ring[0]->size) in mlx4_en_set_ringparam() 1085 param->tx_pending = priv->tx_ring[0]->size; in mlx4_en_get_ringparam() 1794 bf_supported &= priv->tx_ring[i]->bf_alloced; in mlx4_en_set_priv_flags() 1807 priv->tx_ring[i]->bf_enabled = bf_enabled_new; in mlx4_en_set_priv_flags()
|
D | en_port.c | 195 const struct mlx4_en_tx_ring *ring = priv->tx_ring[i]; in mlx4_en_DUMP_ETH_STATS()
|
D | en_tx.c | 390 struct mlx4_en_tx_ring *ring = priv->tx_ring[cq->ring]; in mlx4_en_process_tx_cq() 734 ring = priv->tx_ring[tx_ind]; in mlx4_en_xmit()
|
D | mlx4_en.h | 554 struct mlx4_en_tx_ring **tx_ring; member
|
/linux-4.1.27/drivers/net/ethernet/via/ |
D | via-rhine.c | 444 struct tx_desc *tx_ring; member 1190 rp->tx_ring = ring + RX_RING_SIZE * sizeof(struct rx_desc); in alloc_ring() 1206 rp->tx_ring = NULL; in free_ring() 1291 rp->tx_ring[i].tx_status = 0; in alloc_tbufs() 1292 rp->tx_ring[i].desc_length = cpu_to_le32(TXDESC); in alloc_tbufs() 1294 rp->tx_ring[i].next_desc = cpu_to_le32(next); in alloc_tbufs() 1298 rp->tx_ring[i-1].next_desc = cpu_to_le32(rp->tx_ring_dma); in alloc_tbufs() 1310 rp->tx_ring[i].tx_status = 0; in free_tbufs() 1311 rp->tx_ring[i].desc_length = cpu_to_le32(TXDESC); in free_tbufs() 1312 rp->tx_ring[i].addr = cpu_to_le32(0xBADF00D0); /* An invalid address. */ in free_tbufs() [all …]
|
/linux-4.1.27/drivers/net/xen-netback/ |
D | xenbus.c | 56 struct xen_netif_tx_back_ring *tx_ring = &queue->tx; in xenvif_read_io_ring() local 60 if (tx_ring->sring) { in xenvif_read_io_ring() 61 struct xen_netif_tx_sring *sring = tx_ring->sring; in xenvif_read_io_ring() 64 tx_ring->nr_ents); in xenvif_read_io_ring() 68 tx_ring->req_cons, in xenvif_read_io_ring() 69 tx_ring->req_cons - sring->rsp_prod, in xenvif_read_io_ring() 74 tx_ring->rsp_prod_pvt, in xenvif_read_io_ring() 75 tx_ring->rsp_prod_pvt - sring->rsp_prod, in xenvif_read_io_ring()
|
/linux-4.1.27/drivers/net/wireless/rtlwifi/ |
D | pci.c | 532 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in _rtl_pci_tx_chk_waitq() 567 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[prio]; in _rtl_pci_tx_isr() 1125 ring = &rtlpci->tx_ring[BEACON_QUEUE]; in _rtl_pci_prepare_bcn_tasklet() 1264 rtlpci->tx_ring[prio].buffer_desc = buffer_desc; in _rtl_pci_init_tx_ring() 1265 rtlpci->tx_ring[prio].buffer_desc_dma = buffer_desc_dma; in _rtl_pci_init_tx_ring() 1267 rtlpci->tx_ring[prio].cur_tx_rp = 0; in _rtl_pci_init_tx_ring() 1268 rtlpci->tx_ring[prio].cur_tx_wp = 0; in _rtl_pci_init_tx_ring() 1269 rtlpci->tx_ring[prio].avl_desc = entries; in _rtl_pci_init_tx_ring() 1282 rtlpci->tx_ring[prio].desc = desc; in _rtl_pci_init_tx_ring() 1283 rtlpci->tx_ring[prio].dma = desc_dma; in _rtl_pci_init_tx_ring() [all …]
|
D | pci.h | 210 struct rtl8192_tx_ring tx_ring[RTL_PCI_MAX_TX_QUEUE_COUNT]; member
|
/linux-4.1.27/drivers/net/ethernet/apple/ |
D | macmace.c | 62 unsigned char *tx_ring; member 389 mp->tx_ring = dma_alloc_coherent(mp->device, in mace_open() 392 if (mp->tx_ring == NULL) in mace_open() 421 mp->tx_ring, mp->tx_ring_phys); in mace_open() 469 skb_copy_from_linear_data(skb, mp->tx_ring, skb->len); in mace_xmit_start() 759 mp->tx_ring, mp->tx_ring_phys); in mac_mace_device_remove()
|
/linux-4.1.27/drivers/net/ethernet/sis/ |
D | sis900.c | 185 BufferDesc *tx_ring; member 486 sis_priv->tx_ring = ring_space; in sis900_probe() 575 pci_free_consistent(pci_dev, TX_TOTAL_SIZE, sis_priv->tx_ring, in sis900_probe() 1136 sis_priv->tx_ring[i].link = sis_priv->tx_ring_dma + in sis900_init_tx_ring() 1138 sis_priv->tx_ring[i].cmdsts = 0; in sis900_init_tx_ring() 1139 sis_priv->tx_ring[i].bufptr = 0; in sis900_init_tx_ring() 1564 sis_priv->tx_ring[i].bufptr, skb->len, in sis900_tx_timeout() 1568 sis_priv->tx_ring[i].cmdsts = 0; in sis900_tx_timeout() 1569 sis_priv->tx_ring[i].bufptr = 0; in sis900_tx_timeout() 1614 sis_priv->tx_ring[entry].bufptr = pci_map_single(sis_priv->pci_dev, in sis900_start_xmit() [all …]
|
/linux-4.1.27/drivers/dma/ |
D | xgene-dma.c | 326 struct xgene_dma_ring tx_ring; member 467 xgene_dma_init_desc(desc1, chan->tx_ring.dst_ring_num); in xgene_dma_prep_cpy_desc() 521 xgene_dma_init_desc(desc1, chan->tx_ring.dst_ring_num); in xgene_dma_prep_xor_desc() 771 ret = xgene_chan_xfer_request(&chan->tx_ring, desc_sw); in xgene_chan_xfer_ld_pending() 1446 xgene_dma_delete_ring_one(&chan->tx_ring); in xgene_dma_delete_chan_rings() 1481 struct xgene_dma_ring *tx_ring = &chan->tx_ring; in xgene_dma_create_chan_rings() local 1497 tx_ring->owner = XGENE_DMA_RING_OWNER_DMA; in xgene_dma_create_chan_rings() 1498 tx_ring->buf_num = XGENE_DMA_BUFNUM + chan->id; in xgene_dma_create_chan_rings() 1500 ret = xgene_dma_create_ring_one(chan, tx_ring, in xgene_dma_create_chan_rings() 1507 tx_ring->dst_ring_num = XGENE_DMA_RING_DST_ID(rx_ring->num); in xgene_dma_create_chan_rings() [all …]
|
/linux-4.1.27/drivers/net/wireless/rtl818x/rtl8180/ |
D | dev.c | 345 struct rtl8180_tx_ring *ring = &priv->tx_ring[prio]; in rtl8180_handle_tx() 474 ring = &priv->tx_ring[prio]; in rtl8180_tx() 862 priv->tx_ring[1].dma); in rtl8180_init_hw() 864 priv->tx_ring[0].dma); in rtl8180_init_hw() 867 priv->tx_ring[4].dma); in rtl8180_init_hw() 869 priv->tx_ring[0].dma); in rtl8180_init_hw() 871 priv->tx_ring[1].dma); in rtl8180_init_hw() 873 priv->tx_ring[2].dma); in rtl8180_init_hw() 875 priv->tx_ring[3].dma); in rtl8180_init_hw() 1080 priv->tx_ring[prio].desc = ring; in rtl8180_init_tx_ring() [all …]
|
D | rtl8180.h | 117 struct rtl8180_tx_ring tx_ring[RTL818X_NR_TX_QUEUES]; member
|
/linux-4.1.27/drivers/net/ethernet/rdc/ |
D | r6040.c | 194 struct r6040_descriptor *tx_ring; member 330 lp->tx_remove_ptr = lp->tx_insert_ptr = lp->tx_ring; in r6040_init_txbufs() 331 r6040_init_ring_desc(lp->tx_ring, lp->tx_ring_dma, TX_DCNT); in r6040_init_txbufs() 507 if (lp->tx_ring) { in r6040_close() 509 TX_DESC_SIZE, lp->tx_ring, lp->tx_ring_dma); in r6040_close() 510 lp->tx_ring = NULL; in r6040_close() 780 lp->tx_ring = in r6040_open() 782 if (!lp->tx_ring) { in r6040_open() 797 pci_free_consistent(lp->pdev, TX_DESC_SIZE, lp->tx_ring, in r6040_open()
|
/linux-4.1.27/drivers/net/ethernet/broadcom/ |
D | b44.c | 1028 bp->tx_ring[entry].ctrl = cpu_to_le32(ctrl); in b44_start_xmit() 1029 bp->tx_ring[entry].addr = cpu_to_le32((u32) mapping+bp->dma_offset); in b44_start_xmit() 1033 entry * sizeof(bp->tx_ring[0]), in b44_start_xmit() 1139 memset(bp->tx_ring, 0, B44_TX_RING_BYTES); in b44_init_rings() 1176 if (bp->tx_ring) { in b44_free_consistent() 1180 kfree(bp->tx_ring); in b44_free_consistent() 1183 bp->tx_ring, bp->tx_ring_dma); in b44_free_consistent() 1184 bp->tx_ring = NULL; in b44_free_consistent() 1236 bp->tx_ring = dma_alloc_coherent(bp->sdev->dma_dev, size, in b44_alloc_consistent() 1238 if (!bp->tx_ring) { in b44_alloc_consistent() [all …]
|
D | bgmac.c | 583 bgmac_dma_tx_ring_free(bgmac, &bgmac->tx_ring[i]); in bgmac_dma_cleanup() 594 bgmac_dma_ring_desc_free(bgmac, &bgmac->tx_ring[i], in bgmac_dma_free() 621 ring = &bgmac->tx_ring[i]; in bgmac_dma_alloc() 682 ring = &bgmac->tx_ring[i]; in bgmac_dma_init() 1024 bgmac_dma_tx_reset(bgmac, &bgmac->tx_ring[i]); in bgmac_chip_reset() 1258 bgmac_dma_tx_free(bgmac, &bgmac->tx_ring[0]); in bgmac_poll() 1331 ring = &bgmac->tx_ring[0]; in bgmac_start_xmit()
|
D | b44.h | 358 struct dma_desc *rx_ring, *tx_ring; member
|
D | bgmac.h | 446 struct bgmac_dma_ring tx_ring[BGMAC_MAX_TX_RINGS]; member
|
D | bnx2.c | 700 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_free_tx_mem() 752 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_alloc_tx_mem() 2830 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_tx_int() 3378 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_has_fast_work() 3472 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_poll_work() 5108 txr = &bnapi->tx_ring; in bnx2_clear_ring_states() 5160 txr = &bnapi->tx_ring; in bnx2_init_tx_ring() 5407 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_free_tx_skbs() 5788 struct bnx2_tx_ring_info *txr = &bnapi->tx_ring; in bnx2_run_loopback() 5793 txr = &tx_napi->tx_ring; in bnx2_run_loopback() [all …]
|
D | tg3.c | 6542 if (tnapi->tx_ring[sw_idx].len_flags & TXD_FLAG_HWTSTAMP) { in tg3_tx() 7714 tg3_tx_set_bd(&tnapi->tx_ring[*entry], map, in tg3_tx_frag_set() 7725 tg3_tx_set_bd(&tnapi->tx_ring[*entry], map, in tg3_tx_frag_set() 7735 tg3_tx_set_bd(&tnapi->tx_ring[*entry], map, in tg3_tx_frag_set() 8576 if (tnapi->tx_ring) in tg3_init_rings() 8577 memset(tnapi->tx_ring, 0, TG3_TX_RING_BYTES); in tg3_init_rings() 8600 if (tnapi->tx_ring) { in tg3_mem_tx_release() 8602 tnapi->tx_ring, tnapi->tx_desc_mapping); in tg3_mem_tx_release() 8603 tnapi->tx_ring = NULL; in tg3_mem_tx_release() 8628 tnapi->tx_ring = dma_alloc_coherent(&tp->pdev->dev, in tg3_mem_tx_acquire() [all …]
|
/linux-4.1.27/drivers/net/ethernet/nvidia/ |
D | forcedeth.c | 830 union ring_type tx_ring; member 1932 np->get_tx = np->put_tx = np->first_tx = np->tx_ring; in nv_init_tx() 1935 np->last_tx.orig = &np->tx_ring.orig[np->tx_ring_size-1]; in nv_init_tx() 1937 np->last_tx.ex = &np->tx_ring.ex[np->tx_ring_size-1]; in nv_init_tx() 1948 np->tx_ring.orig[i].flaglen = 0; in nv_init_tx() 1949 np->tx_ring.orig[i].buf = 0; in nv_init_tx() 1951 np->tx_ring.ex[i].flaglen = 0; in nv_init_tx() 1952 np->tx_ring.ex[i].txvlan = 0; in nv_init_tx() 1953 np->tx_ring.ex[i].bufhigh = 0; in nv_init_tx() 1954 np->tx_ring.ex[i].buflow = 0; in nv_init_tx() [all …]
|
/linux-4.1.27/drivers/net/ethernet/marvell/ |
D | skge.c | 502 p->tx_pending = skge->tx_ring.count; in skge_get_ring_param() 516 skge->tx_ring.count = p->tx_pending; in skge_set_ring_param() 2546 tx_size = skge->tx_ring.count * sizeof(struct skge_tx_desc); in skge_up() 2570 err = skge_ring_alloc(&skge->tx_ring, skge->mem + rx_size, in skge_up() 2601 BUG_ON(skge->tx_ring.to_use != skge->tx_ring.to_clean); in skge_up() 2603 skge_qset(skge, txqaddr[port], skge->tx_ring.to_use); in skge_up() 2623 kfree(skge->tx_ring.start); in skge_up() 2717 kfree(skge->tx_ring.start); in skge_down() 2744 if (unlikely(skge_avail(&skge->tx_ring) < skb_shinfo(skb)->nr_frags + 1)) in skge_xmit_frame() 2747 e = skge->tx_ring.to_use; in skge_xmit_frame() [all …]
|
D | sky2.c | 1617 sky2->tx_ring = kcalloc(sky2->tx_ring_size, sizeof(struct tx_ring_info), in sky2_alloc_buffers() 1619 if (!sky2->tx_ring) in sky2_alloc_buffers() 1654 kfree(sky2->tx_ring); in sky2_free_buffers() 1657 sky2->tx_ring = NULL; in sky2_free_buffers() 1937 re = sky2->tx_ring + slot; in sky2_xmit_frame() 1966 re = sky2->tx_ring + slot; in sky2_xmit_frame() 1993 re = sky2->tx_ring + i; in sky2_xmit_frame() 2025 struct tx_ring_info *re = sky2->tx_ring + idx; in sky2_tx_complete()
|
D | sky2.h | 2226 struct tx_ring_info *tx_ring; member
|
D | skge.h | 2456 struct skge_ring tx_ring; member
|
/linux-4.1.27/drivers/net/ethernet/realtek/ |
D | 8139cp.c | 342 struct cp_desc *tx_ring; member 655 struct cp_desc *txd = cp->tx_ring + tx_tail; in cp_tx() 724 txd = &cp->tx_ring[index]; in unwind_tx_frag_mapping() 758 struct cp_desc *txd = &cp->tx_ring[entry]; in cp_start_xmit() 844 txd = &cp->tx_ring[entry]; in cp_start_xmit() 856 txd = &cp->tx_ring[first_entry]; in cp_start_xmit() 1116 memset(cp->tx_ring, 0, sizeof(struct cp_desc) * CP_TX_RING_SIZE); in cp_init_rings() 1117 cp->tx_ring[CP_TX_RING_SIZE - 1].opts1 = cpu_to_le32(RingEnd); in cp_init_rings() 1135 cp->tx_ring = &cp->rx_ring[CP_RX_RING_SIZE]; in cp_alloc_rings() 1162 desc = cp->tx_ring + i; in cp_clean_rings() [all …]
|
/linux-4.1.27/drivers/staging/rtl8192e/rtl8192e/ |
D | rtl_core.c | 445 struct rtl8192_tx_ring *ring = &priv->tx_ring[prio]; in rtl8192_get_nic_desc_num() 460 struct rtl8192_tx_ring *ring = &priv->tx_ring[prio]; in rtl8192_check_nic_enough_desc() 724 ring = &priv->tx_ring[BEACON_QUEUE]; in rtl8192_prepare_beacon() 1349 if (skb_queue_len(&(&priv->tx_ring[i])->queue) > 0) { in rtl8192_is_tx_queue_empty() 1351 i, skb_queue_len(&(&priv->tx_ring[i])->queue)); in rtl8192_is_tx_queue_empty() 1388 ring = &priv->tx_ring[QueueID]; in rtl819x_TxCheckStuck() 1846 struct rtl8192_tx_ring *ring = &priv->tx_ring[prio]; in rtl8192_free_tx_ring() 1940 struct rtl8192_tx_ring *ring = &priv->tx_ring[prio]; in rtl8192_tx_isr() 1972 ring = &priv->tx_ring[TXCMD_QUEUE]; in rtl8192_tx_cmd() 2032 ring = &priv->tx_ring[tcb_desc->queue_index]; in rtl8192_tx() [all …]
|
D | rtl_core.h | 472 struct tx_ring { struct 475 struct tx_ring *next; argument 631 struct rtl8192_tx_ring tx_ring[MAX_TX_QUEUE_COUNT]; member
|
D | r8192E_phy.c | 1494 ring = &priv->tx_ring[QueueID]; in SetRFPowerState8190() 1523 ring = &priv->tx_ring[QueueID]; in SetRFPowerState8190()
|
/linux-4.1.27/net/netlink/ |
D | af_netlink.c | 308 return nlk_sk(sk)->tx_ring.pg_vec != NULL; in netlink_tx_is_mmaped() 377 __netlink_set_ring(struct sock *sk, struct nl_mmap_req *req, bool tx_ring, void **pg_vec, in __netlink_set_ring() argument 384 queue = tx_ring ? &sk->sk_write_queue : &sk->sk_receive_queue; in __netlink_set_ring() 385 ring = tx_ring ? &nlk->tx_ring : &nlk->rx_ring; in __netlink_set_ring() 408 bool tx_ring) in netlink_set_ring() argument 415 ring = tx_ring ? &nlk->tx_ring : &nlk->rx_ring; in netlink_set_ring() 454 __netlink_set_ring(sk, req, tx_ring, pg_vec, order); in netlink_set_ring() 508 for (ring = &nlk->rx_ring; ring <= &nlk->tx_ring; ring++) { in netlink_mmap() 522 for (ring = &nlk->rx_ring; ring <= &nlk->tx_ring; ring++) { in netlink_mmap() 695 if (nlk->tx_ring.pg_vec) { in netlink_poll() [all …]
|
D | af_netlink.h | 50 struct netlink_ring tx_ring; member
|
D | diag.c | 33 ret = sk_diag_put_ring(&nlk->tx_ring, NETLINK_DIAG_TX_RING, in sk_diag_put_rings_cfg()
|
/linux-4.1.27/drivers/infiniband/ulp/ipoib/ |
D | ipoib_cm.c | 736 tx_req = &tx->tx_ring[tx->tx_head & (ipoib_sendq_size - 1)]; in ipoib_cm_send() 792 tx_req = &tx->tx_ring[wr_id]; in ipoib_cm_handle_tx_wc() 1117 p->tx_ring = __vmalloc(ipoib_sendq_size * sizeof *p->tx_ring, in ipoib_cm_tx_init() 1119 if (!p->tx_ring) { in ipoib_cm_tx_init() 1124 memset(p->tx_ring, 0, ipoib_sendq_size * sizeof *p->tx_ring); in ipoib_cm_tx_init() 1165 vfree(p->tx_ring); in ipoib_cm_tx_init() 1182 if (p->tx_ring) { in ipoib_cm_tx_destroy() 1199 tx_req = &p->tx_ring[p->tx_tail & (ipoib_sendq_size - 1)]; in ipoib_cm_tx_destroy() 1215 vfree(p->tx_ring); in ipoib_cm_tx_destroy()
|
D | ipoib_main.c | 1293 priv->tx_ring = vzalloc(ipoib_sendq_size * sizeof *priv->tx_ring); in ipoib_dev_init() 1294 if (!priv->tx_ring) { in ipoib_dev_init() 1318 vfree(priv->tx_ring); in ipoib_dev_init() 1354 vfree(priv->tx_ring); in ipoib_dev_cleanup() 1357 priv->tx_ring = NULL; in ipoib_dev_cleanup()
|
D | ipoib.h | 242 struct ipoib_cm_tx_buf *tx_ring; member 360 struct ipoib_tx_buf *tx_ring; member
|
D | ipoib_ib.c | 390 tx_req = &priv->tx_ring[wr_id]; in ipoib_ib_handle_tx_wc() 591 tx_req = &priv->tx_ring[priv->tx_head & (ipoib_sendq_size - 1)]; in ipoib_send() 869 tx_req = &priv->tx_ring[priv->tx_tail & in ipoib_ib_dev_stop()
|
/linux-4.1.27/drivers/net/ethernet/pasemi/ |
D | pasemi_mac.c | 125 static struct pasemi_mac_txring *tx_ring(const struct pasemi_mac *mac) in tx_ring() function 556 struct pasemi_mac_txring *txring = tx_ring(mac); in pasemi_mac_free_tx_resources() 696 pcnt = *tx_ring(mac)->chan.status & PAS_STATUS_PCNT_M; in pasemi_mac_restart_tx_intr() 700 write_iob_reg(PAS_IOB_DMA_TXCH_RESET(tx_ring(mac)->chan.chno), reg); in pasemi_mac_restart_tx_intr() 727 struct pasemi_dmachan *chan = &tx_ring(mac)->chan; in pasemi_mac_tx_error() 1177 pasemi_dma_start_chan(&tx_ring(mac)->chan, PAS_DMA_TXCHAN_TCMDSTA_SZ | in pasemi_mac_open() 1267 int txch = tx_ring(mac)->chan.chno; in pasemi_mac_pause_txchan() 1332 txch = tx_ring(mac)->chan.chno; in pasemi_mac_close() 1363 pasemi_mac_clean_tx(tx_ring(mac)); in pasemi_mac_close() 1478 struct pasemi_mac_txring * const txring = tx_ring(mac); in pasemi_mac_start_tx() [all …]
|
/linux-4.1.27/drivers/net/ethernet/cadence/ |
D | macb.c | 72 return &queue->tx_ring[macb_tx_ring_wrap(index)]; in macb_tx_desc() 1201 desc = &queue->tx_ring[entry]; in macb_tx_map() 1208 desc = &queue->tx_ring[entry]; in macb_tx_map() 1377 if (queue->tx_ring) { in macb_free_consistent() 1379 queue->tx_ring, queue->tx_ring_dma); in macb_free_consistent() 1380 queue->tx_ring = NULL; in macb_free_consistent() 1424 queue->tx_ring = dma_alloc_coherent(&bp->pdev->dev, size, in macb_alloc_consistent() 1427 if (!queue->tx_ring) in macb_alloc_consistent() 1432 queue->tx_ring); in macb_alloc_consistent() 1467 queue->tx_ring[i].addr = 0; in gem_init_rings() [all …]
|
D | macb.h | 774 struct macb_dma_desc *tx_ring; member
|
/linux-4.1.27/drivers/net/ethernet/octeon/ |
D | octeon_mgmt.c | 126 u64 *tx_ring; member 273 re.d64 = p->tx_ring[p->tx_next_clean]; in octeon_mgmt_clean_tx_buffers() 983 p->tx_ring = kzalloc(ring_size_to_bytes(OCTEON_MGMT_TX_RING_SIZE), in octeon_mgmt_open() 985 if (!p->tx_ring) in octeon_mgmt_open() 988 dma_map_single(p->dev, p->tx_ring, in octeon_mgmt_open() 1236 kfree(p->tx_ring); in octeon_mgmt_open() 1269 kfree(p->tx_ring); in octeon_mgmt_stop() 1307 p->tx_ring[p->tx_next] = re.d64; in octeon_mgmt_xmit()
|
/linux-4.1.27/drivers/net/ethernet/natsemi/ |
D | natsemi.c | 542 struct netdev_desc *tx_ring; member 1863 printk(KERN_DEBUG " Tx ring at %p:\n", np->tx_ring); in dump_ring() 1866 i, np->tx_ring[i].next_desc, in dump_ring() 1867 np->tx_ring[i].cmd_status, in dump_ring() 1868 np->tx_ring[i].addr); in dump_ring() 1920 np->tx_ring = &np->rx_ring[RX_RING_SIZE]; in alloc_ring() 1970 np->tx_ring[i].next_desc = cpu_to_le32(np->ring_dma in init_ring() 1973 np->tx_ring[i].cmd_status = 0; in init_ring() 2075 np->tx_ring[i].cmd_status = 0; in reinit_ring() 2097 np->tx_ring[entry].addr = cpu_to_le32(np->tx_dma[entry]); in start_tx() [all …]
|
/linux-4.1.27/drivers/net/ethernet/altera/ |
D | altera_tse_main.c | 284 priv->tx_ring = kcalloc(tx_descs, sizeof(struct tse_buffer), in alloc_init_skbufs() 286 if (!priv->tx_ring) in alloc_init_skbufs() 307 kfree(priv->tx_ring); in alloc_init_skbufs() 325 tse_free_tx_buffer(priv, &priv->tx_ring[i]); in free_skbufs() 328 kfree(priv->tx_ring); in free_skbufs() 466 tx_buff = &priv->tx_ring[entry]; in tse_tx_complete() 593 buffer = &priv->tx_ring[entry]; in tse_start_xmit()
|
D | altera_tse.h | 434 struct tse_buffer *tx_ring; member
|
/linux-4.1.27/include/linux/ |
D | fs_enet_pd.h | 139 int rx_ring, tx_ring; /* number of buffers on rx */ member
|
/linux-4.1.27/drivers/net/ethernet/alteon/ |
D | acenic.c | 728 if (ap->tx_ring != NULL && !ACE_IS_TIGON_I(ap)) { in ace_free_descriptors() 730 pci_free_consistent(ap->pdev, size, ap->tx_ring, in ace_free_descriptors() 733 ap->tx_ring = NULL; in ace_free_descriptors() 788 ap->tx_ring = pci_alloc_consistent(ap->pdev, size, in ace_allocate_descriptors() 791 if (ap->tx_ring == NULL) in ace_allocate_descriptors() 1286 ap->tx_ring = (__force struct tx_desc *) regs->Window; in ace_init() 1289 writel(0, (__force void __iomem *)ap->tx_ring + i * 4); in ace_init() 1293 memset(ap->tx_ring, 0, in ace_init() 2335 tx = (__force struct tx_desc __iomem *) &ap->tx_ring[i]; in ace_close() 2340 memset(ap->tx_ring + i, 0, in ace_close() [all …]
|
D | acenic.h | 646 struct tx_desc *tx_ring; member
|
/linux-4.1.27/drivers/crypto/qat/qat_common/ |
D | adf_transport.c | 401 struct adf_etr_ring_data *tx_ring; in adf_init_bank() local 437 tx_ring = &bank->rings[i - hw_data->tx_rx_gap]; in adf_init_bank() 438 ring->inflights = tx_ring->inflights; in adf_init_bank()
|
/linux-4.1.27/firmware/keyspan_pda/ |
D | keyspan_pda.S | 231 mov dptr, #tx_ring 752 mov dph1, #HIGH(tx_ring) ; load DPTR1 with target 792 ;; tx_ring[0..15] 823 ;; tx_ring[0..15] 825 mov dptr, #tx_ring ; DPTR1: source 876 mov dph, #HIGH(tx_ring) 1117 tx_ring: label
|
D | xircom_pgs.S | 233 mov dptr, #tx_ring 809 mov dph1, #HIGH(tx_ring) ; load DPTR1 with target 853 ;; tx_ring[0..15] 884 ;; tx_ring[0..15] 886 mov dptr, #tx_ring ; DPTR1: source 937 mov dph, #HIGH(tx_ring) 1185 tx_ring: label
|
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8192ee/ |
D | trx.c | 567 rtlpci->tx_ring[q_idx].avl_desc = point_diff; in rtl92ee_get_available_desc() 595 current_bd_desc = rtlpci->tx_ring[queue_index].cur_tx_wp; in rtl92ee_pre_fill_tx_bd_desc() 649 rtlpci->tx_ring[queue_index].dma + in rtl92ee_pre_fill_tx_bd_desc() 929 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[q_idx]; in rtl92ee_set_desc() 1045 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl92ee_is_tx_desc_closed()
|
D | hw.c | 186 ring = &rtlpci->tx_ring[queue]; in _rtl92ee_set_fw_clock_off() 843 ((u64)rtlpci->tx_ring[BEACON_QUEUE].buffer_desc_dma) & in _rtl92ee_init_mac() 846 (u64)rtlpci->tx_ring[MGNT_QUEUE].buffer_desc_dma & in _rtl92ee_init_mac() 849 (u64)rtlpci->tx_ring[VO_QUEUE].buffer_desc_dma & in _rtl92ee_init_mac() 852 (u64)rtlpci->tx_ring[VI_QUEUE].buffer_desc_dma & in _rtl92ee_init_mac() 856 (u64)rtlpci->tx_ring[BE_QUEUE].buffer_desc_dma & in _rtl92ee_init_mac() 862 (u64)rtlpci->tx_ring[BK_QUEUE].buffer_desc_dma & in _rtl92ee_init_mac() 865 (u64)rtlpci->tx_ring[HIGH_QUEUE].buffer_desc_dma & in _rtl92ee_init_mac()
|
/linux-4.1.27/drivers/infiniband/ulp/srp/ |
D | ib_srp.h | 165 struct srp_iu **tx_ring; member
|
D | ib_srp.c | 646 if (ch->tx_ring) { in srp_free_ch_ib() 648 srp_free_iu(target->srp_host, ch->tx_ring[i]); in srp_free_ch_ib() 649 kfree(ch->tx_ring); in srp_free_ch_ib() 650 ch->tx_ring = NULL; in srp_free_ch_ib() 1239 list_add(&ch->tx_ring[j]->list, &ch->free_tx); in srp_rport_reconnect() 2106 ch->tx_ring = kcalloc(target->queue_size, sizeof(*ch->tx_ring), in srp_alloc_iu_bufs() 2108 if (!ch->tx_ring) in srp_alloc_iu_bufs() 2120 ch->tx_ring[i] = srp_alloc_iu(target->srp_host, in srp_alloc_iu_bufs() 2123 if (!ch->tx_ring[i]) in srp_alloc_iu_bufs() 2126 list_add(&ch->tx_ring[i]->list, &ch->free_tx); in srp_alloc_iu_bufs() [all …]
|
/linux-4.1.27/drivers/net/ethernet/renesas/ |
D | sh_eth.c | 1178 memset(mdp->tx_ring, 0, tx_ringsize); in sh_eth_ring_format() 1183 txdesc = &mdp->tx_ring[i]; in sh_eth_ring_format() 1242 mdp->tx_ring = dma_alloc_coherent(NULL, tx_ringsize, &mdp->tx_desc_dma, in sh_eth_ring_init() 1244 if (!mdp->tx_ring) { in sh_eth_ring_init() 1257 mdp->tx_ring = NULL; in sh_eth_ring_init() 1274 if (mdp->tx_ring) { in sh_eth_free_dma_buffer() 1276 dma_free_coherent(NULL, ringsize, mdp->tx_ring, in sh_eth_free_dma_buffer() 1278 mdp->tx_ring = NULL; in sh_eth_free_dma_buffer() 1384 mdp->tx_ring[i].status &= ~cpu_to_edmac(mdp, TD_TACT); in sh_eth_dev_exit() 1415 txdesc = &mdp->tx_ring[entry]; in sh_eth_txfree() [all …]
|
D | sh_eth.h | 510 struct sh_eth_txdesc *tx_ring; member
|
/linux-4.1.27/drivers/net/ethernet/broadcom/genet/ |
D | bcmgenet.c | 2289 struct bcmgenet_tx_ring *tx_ring; in bcmgenet_isr1() local 2321 tx_ring = &priv->tx_rings[index]; in bcmgenet_isr1() 2323 if (likely(napi_schedule_prep(&tx_ring->napi))) { in bcmgenet_isr1() 2324 tx_ring->int_disable(tx_ring); in bcmgenet_isr1() 2325 __napi_schedule(&tx_ring->napi); in bcmgenet_isr1() 2337 struct bcmgenet_tx_ring *tx_ring; in bcmgenet_isr0() local 2360 tx_ring = &priv->tx_rings[DESC_INDEX]; in bcmgenet_isr0() 2362 if (likely(napi_schedule_prep(&tx_ring->napi))) { in bcmgenet_isr0() 2363 tx_ring->int_disable(tx_ring); in bcmgenet_isr0() 2364 __napi_schedule(&tx_ring->napi); in bcmgenet_isr0()
|
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8192se/ |
D | hw.c | 726 rtl_write_dword(rtlpriv, TBKDA, rtlpci->tx_ring[BK_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload() 727 rtl_write_dword(rtlpriv, TBEDA, rtlpci->tx_ring[BE_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload() 728 rtl_write_dword(rtlpriv, TVIDA, rtlpci->tx_ring[VI_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload() 729 rtl_write_dword(rtlpriv, TVODA, rtlpci->tx_ring[VO_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload() 730 rtl_write_dword(rtlpriv, TBDA, rtlpci->tx_ring[BEACON_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload() 731 rtl_write_dword(rtlpriv, TCDA, rtlpci->tx_ring[TXCMD_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload() 732 rtl_write_dword(rtlpriv, TMDA, rtlpci->tx_ring[MGNT_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload() 733 rtl_write_dword(rtlpriv, THPDA, rtlpci->tx_ring[HIGH_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload() 734 rtl_write_dword(rtlpriv, HDA, rtlpci->tx_ring[HCCA_QUEUE].dma); in _rtl92se_macconfig_before_fwdownload()
|
D | sw.c | 245 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl92se_is_tx_desc_closed()
|
D | fw.c | 146 ring = &rtlpci->tx_ring[TXCMD_QUEUE]; in _rtl92s_cmd_send_packet()
|
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8821ae/ |
D | hw.c | 50 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[BEACON_QUEUE]; in _rtl8821ae_return_beacon_queue_skb() 208 ring = &rtlpci->tx_ring[queue]; in _rtl8821ae_set_fw_clock_off() 1003 rtlpci->tx_ring[BEACON_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac() 1005 rtlpci->tx_ring[MGNT_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac() 1007 rtlpci->tx_ring[VO_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac() 1009 rtlpci->tx_ring[VI_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac() 1011 rtlpci->tx_ring[BE_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac() 1013 rtlpci->tx_ring[BK_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac() 1015 rtlpci->tx_ring[HIGH_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_mac() 1457 rtlpci->tx_ring[BEACON_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8821ae_init_trx_desc_hw_address() [all …]
|
D | trx.c | 974 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl8821ae_is_tx_desc_closed()
|
/linux-4.1.27/drivers/net/ethernet/adaptec/ |
D | starfire.c | 534 starfire_tx_desc *tx_ring; member 915 np->tx_ring = (void *) np->rx_done_q + rx_done_q_size; in netdev_open() 917 np->rx_ring = (void *) np->tx_ring + tx_ring_size; in netdev_open() 1239 np->tx_ring[entry].addr = cpu_to_dma(np->tx_info[entry].mapping); in start_tx() 1240 np->tx_ring[entry].status = cpu_to_le32(status); in start_tx() 1922 i, le32_to_cpu(np->tx_ring[i].status), in netdev_close() 1923 (long long) dma_to_cpu(np->tx_ring[i].addr), in netdev_close()
|
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8188ee/ |
D | hw.c | 92 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[BEACON_QUEUE]; in _rtl88ee_return_beacon_queue_skb() 207 ring = &rtlpci->tx_ring[queue]; in _rtl88ee_set_fw_clock_off() 911 ((u64) rtlpci->tx_ring[BEACON_QUEUE].dma) & in _rtl88ee_init_mac() 914 (u64) rtlpci->tx_ring[MGNT_QUEUE].dma & in _rtl88ee_init_mac() 917 (u64) rtlpci->tx_ring[VO_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl88ee_init_mac() 919 (u64) rtlpci->tx_ring[VI_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl88ee_init_mac() 921 (u64) rtlpci->tx_ring[BE_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl88ee_init_mac() 923 (u64) rtlpci->tx_ring[BK_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl88ee_init_mac() 925 (u64) rtlpci->tx_ring[HIGH_QUEUE].dma & in _rtl88ee_init_mac()
|
D | trx.c | 829 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl88ee_is_tx_desc_closed()
|
D | phy.c | 2247 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl88ee_phy_set_rf_power_state() 2291 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl88ee_phy_set_rf_power_state()
|
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8723be/ |
D | hw.c | 53 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[BEACON_QUEUE]; in _rtl8723be_return_beacon_queue_skb() 209 ring = &rtlpci->tx_ring[queue]; in _rtl8723be_set_fw_clock_off() 901 ((u64) rtlpci->tx_ring[BEACON_QUEUE].dma) & in _rtl8723be_init_mac() 904 (u64) rtlpci->tx_ring[MGNT_QUEUE].dma & in _rtl8723be_init_mac() 907 (u64) rtlpci->tx_ring[VO_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8723be_init_mac() 909 (u64) rtlpci->tx_ring[VI_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8723be_init_mac() 911 (u64) rtlpci->tx_ring[BE_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8723be_init_mac() 913 (u64) rtlpci->tx_ring[BK_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8723be_init_mac() 915 (u64) rtlpci->tx_ring[HIGH_QUEUE].dma & in _rtl8723be_init_mac()
|
D | trx.c | 736 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl8723be_is_tx_desc_closed()
|
D | phy.c | 2629 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl8723be_phy_set_rf_power_state() 2678 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl8723be_phy_set_rf_power_state()
|
/linux-4.1.27/Documentation/networking/ |
D | netlink_mmap.txt | 219 void *rx_ring, *tx_ring; 235 tx_ring = rx_ring + ring_size: 314 hdr = tx_ring + frame_offset;
|
/linux-4.1.27/drivers/net/wireless/ |
D | adm8211.c | 316 u32 status = le32_to_cpu(priv->tx_ring[entry].status); in adm8211_interrupt_tci() 1460 desc = &priv->tx_ring[i]; in adm8211_init_rings() 1645 priv->tx_ring[entry].buffer1 = cpu_to_le32(mapping); in adm8211_tx_raw() 1649 priv->tx_ring[entry].length = cpu_to_le32(flag | skb->len); in adm8211_tx_raw() 1653 priv->tx_ring[entry].status = cpu_to_le32(flag); in adm8211_tx_raw() 1743 priv->tx_ring = priv->rx_ring + priv->rx_ring_size; in adm8211_alloc_rings()
|
D | adm8211.h | 541 struct adm8211_desc *tx_ring; member
|
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8192de/ |
D | hw.c | 761 rtlpci->tx_ring[BEACON_QUEUE].dma); in _rtl92de_init_mac() 762 rtl_write_dword(rtlpriv, REG_MGQ_DESA, rtlpci->tx_ring[MGNT_QUEUE].dma); in _rtl92de_init_mac() 763 rtl_write_dword(rtlpriv, REG_VOQ_DESA, rtlpci->tx_ring[VO_QUEUE].dma); in _rtl92de_init_mac() 764 rtl_write_dword(rtlpriv, REG_VIQ_DESA, rtlpci->tx_ring[VI_QUEUE].dma); in _rtl92de_init_mac() 765 rtl_write_dword(rtlpriv, REG_BEQ_DESA, rtlpci->tx_ring[BE_QUEUE].dma); in _rtl92de_init_mac() 766 rtl_write_dword(rtlpriv, REG_BKQ_DESA, rtlpci->tx_ring[BK_QUEUE].dma); in _rtl92de_init_mac() 767 rtl_write_dword(rtlpriv, REG_HQ_DESA, rtlpci->tx_ring[HIGH_QUEUE].dma); in _rtl92de_init_mac()
|
D | fw.c | 554 ring = &rtlpci->tx_ring[BEACON_QUEUE]; in _rtl92d_cmd_send_packet()
|
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8192ce/ |
D | hw.c | 761 ((u64) rtlpci->tx_ring[BEACON_QUEUE].dma) & in _rtl92ce_init_mac() 764 (u64) rtlpci->tx_ring[MGNT_QUEUE].dma & in _rtl92ce_init_mac() 767 (u64) rtlpci->tx_ring[VO_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl92ce_init_mac() 769 (u64) rtlpci->tx_ring[VI_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl92ce_init_mac() 771 (u64) rtlpci->tx_ring[BE_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl92ce_init_mac() 773 (u64) rtlpci->tx_ring[BK_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl92ce_init_mac() 775 (u64) rtlpci->tx_ring[HIGH_QUEUE].dma & in _rtl92ce_init_mac()
|
D | phy.c | 524 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl92ce_phy_set_rf_power_state()
|
D | trx.c | 745 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl92ce_is_tx_desc_closed()
|
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8723ae/ |
D | hw.c | 758 ((u64) rtlpci->tx_ring[BEACON_QUEUE].dma) & in _rtl8712e_init_mac() 761 (u64) rtlpci->tx_ring[MGNT_QUEUE].dma & in _rtl8712e_init_mac() 764 (u64) rtlpci->tx_ring[VO_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8712e_init_mac() 766 (u64) rtlpci->tx_ring[VI_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8712e_init_mac() 768 (u64) rtlpci->tx_ring[BE_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8712e_init_mac() 770 (u64) rtlpci->tx_ring[BK_QUEUE].dma & DMA_BIT_MASK(32)); in _rtl8712e_init_mac() 772 (u64) rtlpci->tx_ring[HIGH_QUEUE].dma & in _rtl8712e_init_mac()
|
D | trx.c | 687 struct rtl8192_tx_ring *ring = &rtlpci->tx_ring[hw_queue]; in rtl8723e_is_tx_desc_closed()
|
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8192cu/ |
D | phy.c | 460 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl92cu_phy_set_rf_power_state() 503 ring = &pcipriv->dev.tx_ring[queue_id]; in _rtl92cu_phy_set_rf_power_state()
|
/linux-4.1.27/drivers/net/ethernet/rocker/ |
D | rocker.c | 226 struct rocker_dma_ring_info tx_ring; member 1229 &rocker_port->tx_ring); in rocker_port_dma_rings_init() 1235 err = rocker_dma_ring_bufs_alloc(rocker, &rocker_port->tx_ring, in rocker_port_dma_rings_init() 1275 rocker_dma_ring_bufs_free(rocker, &rocker_port->tx_ring, in rocker_port_dma_rings_init() 1278 rocker_dma_ring_destroy(rocker, &rocker_port->tx_ring); in rocker_port_dma_rings_init() 1290 rocker_dma_ring_bufs_free(rocker, &rocker_port->tx_ring, in rocker_port_dma_rings_fini() 1292 rocker_dma_ring_destroy(rocker, &rocker_port->tx_ring); in rocker_port_dma_rings_fini() 3947 desc_info = rocker_desc_head_get(&rocker_port->tx_ring); in rocker_port_xmit() 3978 rocker_desc_head_set(rocker, &rocker_port->tx_ring, desc_info); in rocker_port_xmit() 3980 desc_info = rocker_desc_head_get(&rocker_port->tx_ring); in rocker_port_xmit() [all …]
|
/linux-4.1.27/drivers/net/wireless/rtlwifi/rtl8723com/ |
D | fw_common.c | 318 ring = &rtlpci->tx_ring[BEACON_QUEUE]; in rtl8723_cmd_send_packet()
|