/linux-4.4.14/drivers/net/wireless/ath/ath5k/ |
D | qcu.c | 286 struct ath5k_txq_info *tq = &ah->ah_txq[queue]; in ath5k_hw_set_tx_retry_limits() local 292 (tq->tqi_cw_min << AR5K_NODCU_RETRY_LMT_CW_MIN_S) in ath5k_hw_set_tx_retry_limits() 326 struct ath5k_txq_info *tq = &ah->ah_txq[queue]; in ath5k_hw_reset_tx_queue() local 330 tq = &ah->ah_txq[queue]; in ath5k_hw_reset_tx_queue() 335 (tq->tqi_type == AR5K_TX_QUEUE_INACTIVE)) in ath5k_hw_reset_tx_queue() 343 AR5K_REG_SM(tq->tqi_cw_min, AR5K_DCU_LCL_IFS_CW_MIN) | in ath5k_hw_reset_tx_queue() 344 AR5K_REG_SM(tq->tqi_cw_max, AR5K_DCU_LCL_IFS_CW_MAX) | in ath5k_hw_reset_tx_queue() 345 AR5K_REG_SM(tq->tqi_aifs, AR5K_DCU_LCL_IFS_AIFS), in ath5k_hw_reset_tx_queue() 368 if (tq->tqi_cbr_period) { in ath5k_hw_reset_tx_queue() 369 ath5k_hw_reg_write(ah, AR5K_REG_SM(tq->tqi_cbr_period, in ath5k_hw_reset_tx_queue() [all …]
|
/linux-4.4.14/drivers/net/vmxnet3/ |
D | vmxnet3_drv.c | 103 vmxnet3_tq_stopped(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) in vmxnet3_tq_stopped() argument 105 return tq->stopped; in vmxnet3_tq_stopped() 110 vmxnet3_tq_start(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) in vmxnet3_tq_start() argument 112 tq->stopped = false; in vmxnet3_tq_start() 113 netif_start_subqueue(adapter->netdev, tq - adapter->tx_queue); in vmxnet3_tq_start() 118 vmxnet3_tq_wake(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) in vmxnet3_tq_wake() argument 120 tq->stopped = false; in vmxnet3_tq_wake() 121 netif_wake_subqueue(adapter->netdev, (tq - adapter->tx_queue)); in vmxnet3_tq_wake() 126 vmxnet3_tq_stop(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter) in vmxnet3_tq_stop() argument 128 tq->stopped = true; in vmxnet3_tq_stop() [all …]
|
D | vmxnet3_ethtool.c | 384 struct vmxnet3_tx_queue *tq = &adapter->tx_queue[i]; in vmxnet3_get_regs() local 389 buf[j++] = VMXNET3_GET_ADDR_LO(tq->tx_ring.basePA); in vmxnet3_get_regs() 390 buf[j++] = VMXNET3_GET_ADDR_HI(tq->tx_ring.basePA); in vmxnet3_get_regs() 391 buf[j++] = tq->tx_ring.size; in vmxnet3_get_regs() 392 buf[j++] = tq->tx_ring.next2fill; in vmxnet3_get_regs() 393 buf[j++] = tq->tx_ring.next2comp; in vmxnet3_get_regs() 394 buf[j++] = tq->tx_ring.gen; in vmxnet3_get_regs() 396 buf[j++] = VMXNET3_GET_ADDR_LO(tq->data_ring.basePA); in vmxnet3_get_regs() 397 buf[j++] = VMXNET3_GET_ADDR_HI(tq->data_ring.basePA); in vmxnet3_get_regs() 398 buf[j++] = tq->data_ring.size; in vmxnet3_get_regs() [all …]
|
D | vmxnet3_int.h | 383 #define VMXNET3_WAKE_QUEUE_THRESHOLD(tq) (5) argument
|
/linux-4.4.14/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | sge.c | 233 static inline unsigned int txq_avail(const struct sge_txq *tq) in txq_avail() argument 235 return tq->size - 1 - tq->in_use; in txq_avail() 307 const struct ulptx_sgl *sgl, const struct sge_txq *tq) in unmap_sgl() argument 326 if (likely((u8 *)(p + 1) <= (u8 *)tq->stat)) { in unmap_sgl() 333 } else if ((u8 *)p == (u8 *)tq->stat) { in unmap_sgl() 334 p = (const struct ulptx_sge_pair *)tq->desc; in unmap_sgl() 336 } else if ((u8 *)p + 8 == (u8 *)tq->stat) { in unmap_sgl() 337 const __be64 *addr = (const __be64 *)tq->desc; in unmap_sgl() 345 const __be64 *addr = (const __be64 *)tq->desc; in unmap_sgl() 357 if ((u8 *)p == (u8 *)tq->stat) in unmap_sgl() [all …]
|
D | cxgb4vf_main.c | 476 struct sge_txq *tq; in fwevtq_handler() local 493 tq = s->egr_map[eq_idx]; in fwevtq_handler() 494 if (unlikely(tq == NULL)) { in fwevtq_handler() 499 txq = container_of(tq, struct sge_eth_txq, q); in fwevtq_handler() 500 if (unlikely(tq->abs_id != qid)) { in fwevtq_handler() 503 qid, tq->abs_id); in fwevtq_handler()
|
/linux-4.4.14/drivers/net/ |
D | ifb.c | 54 struct sk_buff_head tq; member 72 skb = skb_peek(&txp->tq); in ifb_ri_tasklet() 76 skb_queue_splice_tail_init(&txp->rq, &txp->tq); in ifb_ri_tasklet() 80 while ((skb = __skb_dequeue(&txp->tq)) != NULL) { in ifb_ri_tasklet() 97 if (skb_queue_len(&txp->tq) != 0) in ifb_ri_tasklet() 178 __skb_queue_head_init(&txp->tq); in ifb_dev_init() 211 __skb_queue_purge(&txp->tq); in ifb_dev_free()
|
/linux-4.4.14/drivers/input/serio/ |
D | hp_sdc.c | 190 curr = hp_sdc.tq[hp_sdc.rcurr]; in hp_sdc_take() 319 curr = hp_sdc.tq[hp_sdc.rcurr]; in hp_sdc_tasklet() 381 if (hp_sdc.tq[curridx] != NULL) in hp_sdc_put() 395 if (hp_sdc.tq[curridx] != NULL) in hp_sdc_put() 415 curr = hp_sdc.tq[curridx]; in hp_sdc_put() 419 hp_sdc.tq[curridx] = NULL; in hp_sdc_put() 433 hp_sdc.tq[curridx] = NULL; in hp_sdc_put() 576 hp_sdc.tq[curridx] = NULL; in hp_sdc_put() 612 if (hp_sdc.tq[i] == this) in __hp_sdc_enqueue_transaction() 620 if (hp_sdc.tq[i] == NULL) { in __hp_sdc_enqueue_transaction() [all …]
|
/linux-4.4.14/drivers/input/keyboard/ |
D | sunkbd.c | 79 struct work_struct tq; member 118 schedule_work(&sunkbd->tq); in sunkbd_interrupt() 225 struct sunkbd *sunkbd = container_of(work, struct sunkbd, tq); in sunkbd_reinit() 268 INIT_WORK(&sunkbd->tq, sunkbd_reinit); in sunkbd_connect()
|
D | lkkbd.c | 282 struct work_struct tq; member 469 schedule_work(&lk->tq); in lkkbd_interrupt() 580 struct lkkbd *lk = container_of(work, struct lkkbd, tq); in lkkbd_reinit() 635 INIT_WORK(&lk->tq, lkkbd_reinit); in lkkbd_connect()
|
/linux-4.4.14/net/batman-adv/ |
D | bat_iv_ogm.c | 334 batadv_ogm_packet->tq = BATADV_TQ_MAX_VALUE; in batadv_iv_ogm_iface_enable() 387 static u8 batadv_hop_penalty(u8 tq, const struct batadv_priv *bat_priv) in batadv_hop_penalty() argument 392 new_tq = tq * (BATADV_TQ_MAX_VALUE - hop_penalty); in batadv_hop_penalty() 453 batadv_ogm_packet->tq, batadv_ogm_packet->ttl, in batadv_iv_ogm_send_to_if() 838 batadv_ogm_packet->tq = batadv_hop_penalty(batadv_ogm_packet->tq, in batadv_iv_ogm_forward() 843 batadv_ogm_packet->tq, batadv_ogm_packet->ttl); in batadv_iv_ogm_forward() 1055 batadv_ogm_packet->tq); in batadv_iv_ogm_orig_update() 1237 combined_tq = batadv_ogm_packet->tq * in batadv_iv_ogm_calc_tq() 1244 batadv_ogm_packet->tq = combined_tq; in batadv_iv_ogm_calc_tq() 1250 batadv_ogm_packet->tq, if_incoming->net_dev->name, in batadv_iv_ogm_calc_tq() [all …]
|
D | packet.h | 224 u8 tq; member
|
D | network-coding.c | 764 if (ogm_packet->tq < bat_priv->nc.min_tq) in batadv_can_nc_with_orig() 995 static u8 batadv_nc_random_weight_tq(u8 tq) in batadv_nc_random_weight_tq() argument 1002 rand_tq = rand_val * (BATADV_TQ_MAX_VALUE - tq); in batadv_nc_random_weight_tq()
|
/linux-4.4.14/fs/ncpfs/ |
D | ncp_fs_sb.h | 116 struct work_struct tq; /* STREAM/DGRAM: data/error ready */ member 135 struct work_struct tq; /* STREAM only: transmitter ready */ member
|
D | sock.c | 105 schedule_work(&server->rcv.tq); in ncp_tcp_data_ready() 113 schedule_work(&server->rcv.tq); in ncp_tcp_error_report() 124 schedule_work(&server->tx.tq); in ncp_tcp_write_space() 385 container_of(work, struct ncp_server, rcv.tq); in ncpdgram_rcv_proc() 683 container_of(work, struct ncp_server, rcv.tq); in ncp_tcp_rcv_proc() 693 container_of(work, struct ncp_server, tx.tq); in ncp_tcp_tx_proc()
|
D | inode.c | 323 flush_work(&server->rcv.tq); in ncp_stop_tasks() 325 flush_work(&server->tx.tq); in ncp_stop_tasks() 648 INIT_WORK(&server->rcv.tq, ncp_tcp_rcv_proc); in ncp_fill_super() 649 INIT_WORK(&server->tx.tq, ncp_tcp_tx_proc); in ncp_fill_super() 652 INIT_WORK(&server->rcv.tq, ncpdgram_rcv_proc); in ncp_fill_super()
|
/linux-4.4.14/arch/arm/boot/dts/ |
D | imx53-mba53.dts | 18 compatible = "tq,mba53", "tq,tqma53", "fsl,imx53"; 71 compatible = "tq,imx53-mba53-sgtl5000",
|
D | imx53-tqma53.dtsi | 17 compatible = "tq,tqma53", "fsl,imx53";
|
/linux-4.4.14/drivers/staging/lustre/lnet/lnet/ |
D | lib-move.c | 789 struct lnet_tx_queue *tq = ni->ni_tx_queues[cpt]; in lnet_post_send_locked() local 844 LASSERT((tq->tq_credits < 0) == in lnet_post_send_locked() 845 !list_empty(&tq->tq_delayed)); in lnet_post_send_locked() 848 tq->tq_credits--; in lnet_post_send_locked() 850 if (tq->tq_credits < tq->tq_credits_min) in lnet_post_send_locked() 851 tq->tq_credits_min = tq->tq_credits; in lnet_post_send_locked() 853 if (tq->tq_credits < 0) { in lnet_post_send_locked() 855 list_add_tail(&msg->msg_list, &tq->tq_delayed); in lnet_post_send_locked() 972 struct lnet_tx_queue *tq = ni->ni_tx_queues[msg->msg_tx_cpt]; in lnet_return_tx_credits_locked() local 977 LASSERT((tq->tq_credits < 0) == in lnet_return_tx_credits_locked() [all …]
|
D | router_proc.c | 693 struct lnet_tx_queue *tq; in proc_lnet_nis() local 715 cfs_percpt_for_each(tq, i, ni->ni_tx_queues) { in proc_lnet_nis() 734 tq->tq_credits_max, in proc_lnet_nis() 735 tq->tq_credits, tq->tq_credits_min); in proc_lnet_nis()
|
D | api-ni.c | 961 struct lnet_tx_queue *tq; in lnet_startup_lndnis() local 1063 cfs_percpt_for_each(tq, i, ni->ni_tx_queues) { in lnet_startup_lndnis() 1064 tq->tq_credits_min = in lnet_startup_lndnis() 1065 tq->tq_credits_max = in lnet_startup_lndnis() 1066 tq->tq_credits = lnet_ni_tq_credits(ni); in lnet_startup_lndnis()
|
D | config.c | 114 struct lnet_tx_queue *tq; in lnet_ni_alloc() local 144 cfs_percpt_for_each(tq, i, ni->ni_tx_queues) in lnet_ni_alloc() 145 INIT_LIST_HEAD(&tq->tq_delayed); in lnet_ni_alloc()
|
/linux-4.4.14/include/uapi/linux/can/ |
D | netlink.h | 33 __u32 tq; /* Time quanta (TQ) in nanoseconds */ member
|
/linux-4.4.14/ipc/ |
D | sem.c | 208 struct sem_queue *q, *tq; in unmerge_queues() local 218 list_for_each_entry_safe(q, tq, &sma->pending_alter, list) { in unmerge_queues() 1090 struct sem_queue *q, *tq; in freeary() local 1108 list_for_each_entry_safe(q, tq, &sma->pending_const, list) { in freeary() 1113 list_for_each_entry_safe(q, tq, &sma->pending_alter, list) { in freeary() 1119 list_for_each_entry_safe(q, tq, &sem->pending_const, list) { in freeary() 1123 list_for_each_entry_safe(q, tq, &sem->pending_alter, list) { in freeary()
|
/linux-4.4.14/drivers/net/can/ |
D | dev.c | 179 bt->tq = (u32)v64; in can_calc_bittiming() 232 brp64 = (u64)priv->clock.freq * (u64)bt->tq; in can_fixup_bittiming() 266 if (!bt->tq && bt->bitrate) in can_get_bittiming() 268 else if (bt->tq && !bt->bitrate) in can_get_bittiming()
|
D | Kconfig | 52 arguments "tq", "prop_seg", "phase_seg1", "phase_seg2" and "sjw".
|
/linux-4.4.14/include/linux/ |
D | hp_sdc.h | 281 hp_sdc_transaction *tq[HP_SDC_QUEUE_LEN]; /* All pending read/writes */ member
|
/linux-4.4.14/Documentation/ABI/testing/ |
D | sysfs-class-net-mesh | 69 originator message's tq-field on every hop.
|
/linux-4.4.14/Documentation/networking/ |
D | can.txt | 1022 [ tq TQ prop-seg PROP_SEG phase-seg1 PHASE-SEG1 1057 tq 125 prop-seg 6 phase-seg1 7 phase-seg2 2 sjw 1 1091 "tq 125 prop-seg 6 phase-seg1 7 phase-seg2 2 sjw 1" 1094 tq. They allow to define the CAN bit-timing in a hardware 1102 and 2, the synchronisation jump width in units of tq, the 1117 specifying the arguments "tq", "prop_seg", "phase_seg1", "phase_seg2" 1120 $ ip link set canX type can tq 125 prop-seg 6 \ 1228 tq 50 prop-seg 14 phase-seg1 15 phase-seg2 10 sjw 1
|
/linux-4.4.14/drivers/tty/ipwireless/ |
D | hardware.c | 1728 struct ipw_tx_packet *tp, *tq; in ipwireless_hardware_free() local 1738 list_for_each_entry_safe(tp, tq, &hw->tx_queue[i], queue) { in ipwireless_hardware_free()
|