tq                270 drivers/input/keyboard/lkkbd.c 	struct work_struct tq;
tq                457 drivers/input/keyboard/lkkbd.c 		schedule_work(&lk->tq);
tq                568 drivers/input/keyboard/lkkbd.c 	struct lkkbd *lk = container_of(work, struct lkkbd, tq);
tq                623 drivers/input/keyboard/lkkbd.c 	INIT_WORK(&lk->tq, lkkbd_reinit);
tq                 63 drivers/input/keyboard/sunkbd.c 	struct work_struct tq;
tq                102 drivers/input/keyboard/sunkbd.c 		schedule_work(&sunkbd->tq);
tq                209 drivers/input/keyboard/sunkbd.c 	struct sunkbd *sunkbd = container_of(work, struct sunkbd, tq);
tq                252 drivers/input/keyboard/sunkbd.c 	INIT_WORK(&sunkbd->tq, sunkbd_reinit);
tq                190 drivers/input/serio/hp_sdc.c 	curr = hp_sdc.tq[hp_sdc.rcurr];
tq                316 drivers/input/serio/hp_sdc.c 			curr = hp_sdc.tq[hp_sdc.rcurr];
tq                378 drivers/input/serio/hp_sdc.c 	if (hp_sdc.tq[curridx] != NULL)
tq                392 drivers/input/serio/hp_sdc.c 		if (hp_sdc.tq[curridx] != NULL)
tq                412 drivers/input/serio/hp_sdc.c 	curr = hp_sdc.tq[curridx];
tq                416 drivers/input/serio/hp_sdc.c 		hp_sdc.tq[curridx] = NULL;
tq                430 drivers/input/serio/hp_sdc.c 		hp_sdc.tq[curridx] = NULL;
tq                573 drivers/input/serio/hp_sdc.c 		hp_sdc.tq[curridx] = NULL;
tq                609 drivers/input/serio/hp_sdc.c 		if (hp_sdc.tq[i] == this)
tq                617 drivers/input/serio/hp_sdc.c 		if (hp_sdc.tq[i] == NULL) {
tq                618 drivers/input/serio/hp_sdc.c 			hp_sdc.tq[i] = this;
tq                652 drivers/input/serio/hp_sdc.c 		if (hp_sdc.tq[i] == this)
tq                653 drivers/input/serio/hp_sdc.c 			hp_sdc.tq[i] = NULL;
tq                853 drivers/input/serio/hp_sdc.c 	memset(&hp_sdc.tq, 0, sizeof(hp_sdc.tq));
tq                205 drivers/net/can/dev.c 	bt->tq = (u32)v64;
tq                259 drivers/net/can/dev.c 	brp64 = (u64)priv->clock.freq * (u64)bt->tq;
tq                310 drivers/net/can/dev.c 	if (!bt->tq && bt->bitrate && btc)
tq                312 drivers/net/can/dev.c 	else if (bt->tq && !bt->bitrate && btc)
tq                314 drivers/net/can/dev.c 	else if (!bt->tq && bt->bitrate && bitrate_const)
tq                140 drivers/net/can/usb/ucan.c 	__le32 tq;           /* Time quanta (TQ) in nanoseconds */
tq               1249 drivers/net/can/usb/ucan.c 	cmd_set_bittiming->tq = cpu_to_le32(up->can.bittiming.tq);
tq               1941 drivers/net/ethernet/chelsio/cxgb4/sge.c 		struct sge_txq *tq = &txq->q;
tq               1950 drivers/net/ethernet/chelsio/cxgb4/sge.c 		if (unlikely((void *)sgl == (void *)tq->stat)) {
tq               1951 drivers/net/ethernet/chelsio/cxgb4/sge.c 			sgl = (void *)tq->desc;
tq               1952 drivers/net/ethernet/chelsio/cxgb4/sge.c 			end = (void *)((void *)tq->desc +
tq               1953 drivers/net/ethernet/chelsio/cxgb4/sge.c 				       ((void *)end - (void *)tq->stat));
tq               1956 drivers/net/ethernet/chelsio/cxgb4/sge.c 		cxgb4_write_sgl(skb, tq, sgl, end, 0, addr);
tq               1959 drivers/net/ethernet/chelsio/cxgb4/sge.c 		last_desc = tq->pidx + ndesc - 1;
tq               1960 drivers/net/ethernet/chelsio/cxgb4/sge.c 		if (last_desc >= tq->size)
tq               1961 drivers/net/ethernet/chelsio/cxgb4/sge.c 			last_desc -= tq->size;
tq               1962 drivers/net/ethernet/chelsio/cxgb4/sge.c 		tq->sdesc[last_desc].skb = skb;
tq               1963 drivers/net/ethernet/chelsio/cxgb4/sge.c 		tq->sdesc[last_desc].sgl = sgl;
tq                537 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		struct sge_txq *tq;
tq                554 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		tq = s->egr_map[eq_idx];
tq                555 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		if (unlikely(tq == NULL)) {
tq                560 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		txq = container_of(tq, struct sge_eth_txq, q);
tq                561 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		if (unlikely(tq->abs_id != qid)) {
tq                564 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 				qid, tq->abs_id);
tq                233 drivers/net/ethernet/chelsio/cxgb4vf/sge.c static inline unsigned int txq_avail(const struct sge_txq *tq)
tq                235 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	return tq->size - 1 - tq->in_use;
tq                307 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		      const struct ulptx_sgl *sgl, const struct sge_txq *tq)
tq                326 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		if (likely((u8 *)(p + 1) <= (u8 *)tq->stat)) {
tq                333 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		} else if ((u8 *)p == (u8 *)tq->stat) {
tq                334 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			p = (const struct ulptx_sge_pair *)tq->desc;
tq                336 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		} else if ((u8 *)p + 8 == (u8 *)tq->stat) {
tq                337 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			const __be64 *addr = (const __be64 *)tq->desc;
tq                345 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			const __be64 *addr = (const __be64 *)tq->desc;
tq                357 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		if ((u8 *)p == (u8 *)tq->stat)
tq                358 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			p = (const struct ulptx_sge_pair *)tq->desc;
tq                359 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		addr = ((u8 *)p + 16 <= (u8 *)tq->stat
tq                361 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			: *(const __be64 *)tq->desc);
tq                377 drivers/net/ethernet/chelsio/cxgb4vf/sge.c static void free_tx_desc(struct adapter *adapter, struct sge_txq *tq,
tq                381 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	unsigned int cidx = tq->cidx;
tq                386 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	sdesc = &tq->sdesc[cidx];
tq                394 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 				unmap_sgl(dev, sdesc->skb, sdesc->sgl, tq);
tq                400 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		if (++cidx == tq->size) {
tq                402 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			sdesc = tq->sdesc;
tq                405 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	tq->cidx = cidx;
tq                411 drivers/net/ethernet/chelsio/cxgb4vf/sge.c static inline int reclaimable(const struct sge_txq *tq)
tq                413 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	int hw_cidx = be16_to_cpu(tq->stat->cidx);
tq                414 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	int reclaimable = hw_cidx - tq->cidx;
tq                416 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		reclaimable += tq->size;
tq                431 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 					struct sge_txq *tq,
tq                434 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	int avail = reclaimable(tq);
tq                444 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		free_tx_desc(adapter, tq, avail, unmap);
tq                445 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		tq->in_use -= avail;
tq                901 drivers/net/ethernet/chelsio/cxgb4vf/sge.c static void write_sgl(const struct sk_buff *skb, struct sge_txq *tq,
tq                930 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	to = (u8 *)end > (u8 *)tq->stat ? buf : sgl->sge;
tq                943 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	if (unlikely((u8 *)end > (u8 *)tq->stat)) {
tq                944 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		unsigned int part0 = (u8 *)tq->stat - (u8 *)sgl->sge, part1;
tq                948 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		part1 = (u8 *)end - (u8 *)tq->stat;
tq                949 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		memcpy(tq->desc, (u8 *)buf + part0, part1);
tq                950 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		end = (void *)tq->desc + part1;
tq                964 drivers/net/ethernet/chelsio/cxgb4vf/sge.c static inline void ring_tx_db(struct adapter *adapter, struct sge_txq *tq,
tq                975 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	if (unlikely(tq->bar2_addr == NULL)) {
tq                979 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			     QID_V(tq->cntxt_id) | val);
tq                995 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		if (n == 1 && tq->bar2_qid == 0) {
tq                996 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			unsigned int index = (tq->pidx
tq                997 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 					      ? (tq->pidx - 1)
tq                998 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 					      : (tq->size - 1));
tq                999 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			__be64 *src = (__be64 *)&tq->desc[index];
tq               1000 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			__be64 __iomem *dst = (__be64 __iomem *)(tq->bar2_addr +
tq               1024 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			writel(val | QID_V(tq->bar2_qid),
tq               1025 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			       tq->bar2_addr + SGE_UDB_KDOORBELL);
tq               1052 drivers/net/ethernet/chelsio/cxgb4vf/sge.c static void inline_tx_skb(const struct sk_buff *skb, const struct sge_txq *tq,
tq               1056 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	int left = (void *)tq->stat - pos;
tq               1066 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		skb_copy_bits(skb, left, tq->desc, skb->len - left);
tq               1067 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		pos = (void *)tq->desc + (skb->len - left);
tq               1142 drivers/net/ethernet/chelsio/cxgb4vf/sge.c static inline void txq_advance(struct sge_txq *tq, unsigned int n)
tq               1144 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	tq->in_use += n;
tq               1145 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	tq->pidx += n;
tq               1146 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	if (tq->pidx >= tq->size)
tq               1147 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		tq->pidx -= tq->size;
tq               1420 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		struct sge_txq *tq = &txq->q;
tq               1430 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		if (unlikely((void *)sgl == (void *)tq->stat)) {
tq               1431 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			sgl = (void *)tq->desc;
tq               1432 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			end = ((void *)tq->desc + ((void *)end - (void *)tq->stat));
tq               1435 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		write_sgl(skb, tq, sgl, end, 0, addr);
tq               1438 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		last_desc = tq->pidx + ndesc - 1;
tq               1439 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		if (last_desc >= tq->size)
tq               1440 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			last_desc -= tq->size;
tq               1441 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		tq->sdesc[last_desc].skb = skb;
tq               1442 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 		tq->sdesc[last_desc].sgl = sgl;
tq               2513 drivers/net/ethernet/chelsio/cxgb4vf/sge.c static void free_txq(struct adapter *adapter, struct sge_txq *tq)
tq               2518 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			  tq->size * sizeof(*tq->desc) + s->stat_len,
tq               2519 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 			  tq->desc, tq->phys_addr);
tq               2520 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	tq->cntxt_id = 0;
tq               2521 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	tq->sdesc = NULL;
tq               2522 drivers/net/ethernet/chelsio/cxgb4vf/sge.c 	tq->desc = NULL;
tq                 51 drivers/net/ifb.c 	struct sk_buff_head     tq;
tq                 69 drivers/net/ifb.c 	skb = skb_peek(&txp->tq);
tq                 73 drivers/net/ifb.c 		skb_queue_splice_tail_init(&txp->rq, &txp->tq);
tq                 77 drivers/net/ifb.c 	while ((skb = __skb_dequeue(&txp->tq)) != NULL) {
tq                 92 drivers/net/ifb.c 			if (skb_queue_len(&txp->tq) != 0)
tq                170 drivers/net/ifb.c 		__skb_queue_head_init(&txp->tq);
tq                204 drivers/net/ifb.c 		__skb_queue_purge(&txp->tq);
tq                103 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_stopped(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter)
tq                105 drivers/net/vmxnet3/vmxnet3_drv.c 	return tq->stopped;
tq                110 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_start(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter)
tq                112 drivers/net/vmxnet3/vmxnet3_drv.c 	tq->stopped = false;
tq                113 drivers/net/vmxnet3/vmxnet3_drv.c 	netif_start_subqueue(adapter->netdev, tq - adapter->tx_queue);
tq                118 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_wake(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter)
tq                120 drivers/net/vmxnet3/vmxnet3_drv.c 	tq->stopped = false;
tq                121 drivers/net/vmxnet3/vmxnet3_drv.c 	netif_wake_subqueue(adapter->netdev, (tq - adapter->tx_queue));
tq                126 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_stop(struct vmxnet3_tx_queue *tq, struct vmxnet3_adapter *adapter)
tq                128 drivers/net/vmxnet3/vmxnet3_drv.c 	tq->stopped = true;
tq                129 drivers/net/vmxnet3/vmxnet3_drv.c 	tq->num_stop++;
tq                130 drivers/net/vmxnet3/vmxnet3_drv.c 	netif_stop_subqueue(adapter->netdev, (tq - adapter->tx_queue));
tq                329 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_unmap_pkt(u32 eop_idx, struct vmxnet3_tx_queue *tq,
tq                336 drivers/net/vmxnet3/vmxnet3_drv.c 	BUG_ON(tq->buf_info[eop_idx].sop_idx != tq->tx_ring.next2comp);
tq                337 drivers/net/vmxnet3/vmxnet3_drv.c 	BUG_ON(VMXNET3_TXDESC_GET_EOP(&(tq->tx_ring.base[eop_idx].txd)) != 1);
tq                339 drivers/net/vmxnet3/vmxnet3_drv.c 	skb = tq->buf_info[eop_idx].skb;
tq                341 drivers/net/vmxnet3/vmxnet3_drv.c 	tq->buf_info[eop_idx].skb = NULL;
tq                343 drivers/net/vmxnet3/vmxnet3_drv.c 	VMXNET3_INC_RING_IDX_ONLY(eop_idx, tq->tx_ring.size);
tq                345 drivers/net/vmxnet3/vmxnet3_drv.c 	while (tq->tx_ring.next2comp != eop_idx) {
tq                346 drivers/net/vmxnet3/vmxnet3_drv.c 		vmxnet3_unmap_tx_buf(tq->buf_info + tq->tx_ring.next2comp,
tq                354 drivers/net/vmxnet3/vmxnet3_drv.c 		vmxnet3_cmd_ring_adv_next2comp(&tq->tx_ring);
tq                364 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_tx_complete(struct vmxnet3_tx_queue *tq,
tq                370 drivers/net/vmxnet3/vmxnet3_drv.c 	gdesc = tq->comp_ring.base + tq->comp_ring.next2proc;
tq                371 drivers/net/vmxnet3/vmxnet3_drv.c 	while (VMXNET3_TCD_GET_GEN(&gdesc->tcd) == tq->comp_ring.gen) {
tq                378 drivers/net/vmxnet3/vmxnet3_drv.c 					       &gdesc->tcd), tq, adapter->pdev,
tq                381 drivers/net/vmxnet3/vmxnet3_drv.c 		vmxnet3_comp_ring_adv_next2proc(&tq->comp_ring);
tq                382 drivers/net/vmxnet3/vmxnet3_drv.c 		gdesc = tq->comp_ring.base + tq->comp_ring.next2proc;
tq                386 drivers/net/vmxnet3/vmxnet3_drv.c 		spin_lock(&tq->tx_lock);
tq                387 drivers/net/vmxnet3/vmxnet3_drv.c 		if (unlikely(vmxnet3_tq_stopped(tq, adapter) &&
tq                388 drivers/net/vmxnet3/vmxnet3_drv.c 			     vmxnet3_cmd_ring_desc_avail(&tq->tx_ring) >
tq                389 drivers/net/vmxnet3/vmxnet3_drv.c 			     VMXNET3_WAKE_QUEUE_THRESHOLD(tq) &&
tq                391 drivers/net/vmxnet3/vmxnet3_drv.c 			vmxnet3_tq_wake(tq, adapter);
tq                393 drivers/net/vmxnet3/vmxnet3_drv.c 		spin_unlock(&tq->tx_lock);
tq                400 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_cleanup(struct vmxnet3_tx_queue *tq,
tq                405 drivers/net/vmxnet3/vmxnet3_drv.c 	while (tq->tx_ring.next2comp != tq->tx_ring.next2fill) {
tq                408 drivers/net/vmxnet3/vmxnet3_drv.c 		tbi = tq->buf_info + tq->tx_ring.next2comp;
tq                415 drivers/net/vmxnet3/vmxnet3_drv.c 		vmxnet3_cmd_ring_adv_next2comp(&tq->tx_ring);
tq                419 drivers/net/vmxnet3/vmxnet3_drv.c 	for (i = 0; i < tq->tx_ring.size; i++) {
tq                420 drivers/net/vmxnet3/vmxnet3_drv.c 		BUG_ON(tq->buf_info[i].skb != NULL ||
tq                421 drivers/net/vmxnet3/vmxnet3_drv.c 		       tq->buf_info[i].map_type != VMXNET3_MAP_NONE);
tq                424 drivers/net/vmxnet3/vmxnet3_drv.c 	tq->tx_ring.gen = VMXNET3_INIT_GEN;
tq                425 drivers/net/vmxnet3/vmxnet3_drv.c 	tq->tx_ring.next2fill = tq->tx_ring.next2comp = 0;
tq                427 drivers/net/vmxnet3/vmxnet3_drv.c 	tq->comp_ring.gen = VMXNET3_INIT_GEN;
tq                428 drivers/net/vmxnet3/vmxnet3_drv.c 	tq->comp_ring.next2proc = 0;
tq                433 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_destroy(struct vmxnet3_tx_queue *tq,
tq                436 drivers/net/vmxnet3/vmxnet3_drv.c 	if (tq->tx_ring.base) {
tq                437 drivers/net/vmxnet3/vmxnet3_drv.c 		dma_free_coherent(&adapter->pdev->dev, tq->tx_ring.size *
tq                439 drivers/net/vmxnet3/vmxnet3_drv.c 				  tq->tx_ring.base, tq->tx_ring.basePA);
tq                440 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->tx_ring.base = NULL;
tq                442 drivers/net/vmxnet3/vmxnet3_drv.c 	if (tq->data_ring.base) {
tq                444 drivers/net/vmxnet3/vmxnet3_drv.c 				  tq->data_ring.size * tq->txdata_desc_size,
tq                445 drivers/net/vmxnet3/vmxnet3_drv.c 				  tq->data_ring.base, tq->data_ring.basePA);
tq                446 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->data_ring.base = NULL;
tq                448 drivers/net/vmxnet3/vmxnet3_drv.c 	if (tq->comp_ring.base) {
tq                449 drivers/net/vmxnet3/vmxnet3_drv.c 		dma_free_coherent(&adapter->pdev->dev, tq->comp_ring.size *
tq                451 drivers/net/vmxnet3/vmxnet3_drv.c 				  tq->comp_ring.base, tq->comp_ring.basePA);
tq                452 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->comp_ring.base = NULL;
tq                454 drivers/net/vmxnet3/vmxnet3_drv.c 	if (tq->buf_info) {
tq                456 drivers/net/vmxnet3/vmxnet3_drv.c 				  tq->tx_ring.size * sizeof(tq->buf_info[0]),
tq                457 drivers/net/vmxnet3/vmxnet3_drv.c 				  tq->buf_info, tq->buf_info_pa);
tq                458 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->buf_info = NULL;
tq                475 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_init(struct vmxnet3_tx_queue *tq,
tq                481 drivers/net/vmxnet3/vmxnet3_drv.c 	memset(tq->tx_ring.base, 0, tq->tx_ring.size *
tq                483 drivers/net/vmxnet3/vmxnet3_drv.c 	tq->tx_ring.next2fill = tq->tx_ring.next2comp = 0;
tq                484 drivers/net/vmxnet3/vmxnet3_drv.c 	tq->tx_ring.gen = VMXNET3_INIT_GEN;
tq                486 drivers/net/vmxnet3/vmxnet3_drv.c 	memset(tq->data_ring.base, 0,
tq                487 drivers/net/vmxnet3/vmxnet3_drv.c 	       tq->data_ring.size * tq->txdata_desc_size);
tq                490 drivers/net/vmxnet3/vmxnet3_drv.c 	memset(tq->comp_ring.base, 0, tq->comp_ring.size *
tq                492 drivers/net/vmxnet3/vmxnet3_drv.c 	tq->comp_ring.next2proc = 0;
tq                493 drivers/net/vmxnet3/vmxnet3_drv.c 	tq->comp_ring.gen = VMXNET3_INIT_GEN;
tq                496 drivers/net/vmxnet3/vmxnet3_drv.c 	memset(tq->buf_info, 0, sizeof(tq->buf_info[0]) * tq->tx_ring.size);
tq                497 drivers/net/vmxnet3/vmxnet3_drv.c 	for (i = 0; i < tq->tx_ring.size; i++)
tq                498 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->buf_info[i].map_type = VMXNET3_MAP_NONE;
tq                505 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_create(struct vmxnet3_tx_queue *tq,
tq                510 drivers/net/vmxnet3/vmxnet3_drv.c 	BUG_ON(tq->tx_ring.base || tq->data_ring.base ||
tq                511 drivers/net/vmxnet3/vmxnet3_drv.c 	       tq->comp_ring.base || tq->buf_info);
tq                513 drivers/net/vmxnet3/vmxnet3_drv.c 	tq->tx_ring.base = dma_alloc_coherent(&adapter->pdev->dev,
tq                514 drivers/net/vmxnet3/vmxnet3_drv.c 			tq->tx_ring.size * sizeof(struct Vmxnet3_TxDesc),
tq                515 drivers/net/vmxnet3/vmxnet3_drv.c 			&tq->tx_ring.basePA, GFP_KERNEL);
tq                516 drivers/net/vmxnet3/vmxnet3_drv.c 	if (!tq->tx_ring.base) {
tq                521 drivers/net/vmxnet3/vmxnet3_drv.c 	tq->data_ring.base = dma_alloc_coherent(&adapter->pdev->dev,
tq                522 drivers/net/vmxnet3/vmxnet3_drv.c 			tq->data_ring.size * tq->txdata_desc_size,
tq                523 drivers/net/vmxnet3/vmxnet3_drv.c 			&tq->data_ring.basePA, GFP_KERNEL);
tq                524 drivers/net/vmxnet3/vmxnet3_drv.c 	if (!tq->data_ring.base) {
tq                529 drivers/net/vmxnet3/vmxnet3_drv.c 	tq->comp_ring.base = dma_alloc_coherent(&adapter->pdev->dev,
tq                530 drivers/net/vmxnet3/vmxnet3_drv.c 			tq->comp_ring.size * sizeof(struct Vmxnet3_TxCompDesc),
tq                531 drivers/net/vmxnet3/vmxnet3_drv.c 			&tq->comp_ring.basePA, GFP_KERNEL);
tq                532 drivers/net/vmxnet3/vmxnet3_drv.c 	if (!tq->comp_ring.base) {
tq                537 drivers/net/vmxnet3/vmxnet3_drv.c 	sz = tq->tx_ring.size * sizeof(tq->buf_info[0]);
tq                538 drivers/net/vmxnet3/vmxnet3_drv.c 	tq->buf_info = dma_alloc_coherent(&adapter->pdev->dev, sz,
tq                539 drivers/net/vmxnet3/vmxnet3_drv.c 					  &tq->buf_info_pa, GFP_KERNEL);
tq                540 drivers/net/vmxnet3/vmxnet3_drv.c 	if (!tq->buf_info)
tq                546 drivers/net/vmxnet3/vmxnet3_drv.c 	vmxnet3_tq_destroy(tq, adapter);
tq                675 drivers/net/vmxnet3/vmxnet3_drv.c 		struct vmxnet3_tx_queue *tq, struct pci_dev *pdev,
tq                687 drivers/net/vmxnet3/vmxnet3_drv.c 	dw2 = (tq->tx_ring.gen ^ 0x1) << VMXNET3_TXD_GEN_SHIFT;
tq                689 drivers/net/vmxnet3/vmxnet3_drv.c 	ctx->sop_txd = tq->tx_ring.base + tq->tx_ring.next2fill;
tq                694 drivers/net/vmxnet3/vmxnet3_drv.c 		ctx->sop_txd->txd.addr = cpu_to_le64(tq->data_ring.basePA +
tq                695 drivers/net/vmxnet3/vmxnet3_drv.c 					tq->tx_ring.next2fill *
tq                696 drivers/net/vmxnet3/vmxnet3_drv.c 					tq->txdata_desc_size);
tq                700 drivers/net/vmxnet3/vmxnet3_drv.c 		tbi = tq->buf_info + tq->tx_ring.next2fill;
tq                705 drivers/net/vmxnet3/vmxnet3_drv.c 			tq->tx_ring.next2fill,
tq                708 drivers/net/vmxnet3/vmxnet3_drv.c 		vmxnet3_cmd_ring_adv_next2fill(&tq->tx_ring);
tq                711 drivers/net/vmxnet3/vmxnet3_drv.c 		dw2 = tq->tx_ring.gen << VMXNET3_TXD_GEN_SHIFT;
tq                728 drivers/net/vmxnet3/vmxnet3_drv.c 		tbi = tq->buf_info + tq->tx_ring.next2fill;
tq                738 drivers/net/vmxnet3/vmxnet3_drv.c 		gdesc = tq->tx_ring.base + tq->tx_ring.next2fill;
tq                739 drivers/net/vmxnet3/vmxnet3_drv.c 		BUG_ON(gdesc->txd.gen == tq->tx_ring.gen);
tq                747 drivers/net/vmxnet3/vmxnet3_drv.c 			tq->tx_ring.next2fill, le64_to_cpu(gdesc->txd.addr),
tq                749 drivers/net/vmxnet3/vmxnet3_drv.c 		vmxnet3_cmd_ring_adv_next2fill(&tq->tx_ring);
tq                750 drivers/net/vmxnet3/vmxnet3_drv.c 		dw2 = tq->tx_ring.gen << VMXNET3_TXD_GEN_SHIFT;
tq                763 drivers/net/vmxnet3/vmxnet3_drv.c 			tbi = tq->buf_info + tq->tx_ring.next2fill;
tq                780 drivers/net/vmxnet3/vmxnet3_drv.c 			gdesc = tq->tx_ring.base + tq->tx_ring.next2fill;
tq                781 drivers/net/vmxnet3/vmxnet3_drv.c 			BUG_ON(gdesc->txd.gen == tq->tx_ring.gen);
tq                789 drivers/net/vmxnet3/vmxnet3_drv.c 				tq->tx_ring.next2fill, le64_to_cpu(gdesc->txd.addr),
tq                791 drivers/net/vmxnet3/vmxnet3_drv.c 			vmxnet3_cmd_ring_adv_next2fill(&tq->tx_ring);
tq                792 drivers/net/vmxnet3/vmxnet3_drv.c 			dw2 = tq->tx_ring.gen << VMXNET3_TXD_GEN_SHIFT;
tq                803 drivers/net/vmxnet3/vmxnet3_drv.c 	tbi->sop_idx = ctx->sop_txd - tq->tx_ring.base;
tq                838 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_parse_hdr(struct sk_buff *skb, struct vmxnet3_tx_queue *tq,
tq                881 drivers/net/vmxnet3/vmxnet3_drv.c 					       tq->txdata_desc_size,
tq                893 drivers/net/vmxnet3/vmxnet3_drv.c 	if (unlikely(ctx->copy_size > tq->txdata_desc_size)) {
tq                894 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->stats.oversized_hdr++;
tq                915 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_copy_hdr(struct sk_buff *skb, struct vmxnet3_tx_queue *tq,
tq                921 drivers/net/vmxnet3/vmxnet3_drv.c 	tdd = (struct Vmxnet3_TxDataDesc *)((u8 *)tq->data_ring.base +
tq                922 drivers/net/vmxnet3/vmxnet3_drv.c 					    tq->tx_ring.next2fill *
tq                923 drivers/net/vmxnet3/vmxnet3_drv.c 					    tq->txdata_desc_size);
tq                928 drivers/net/vmxnet3/vmxnet3_drv.c 		ctx->copy_size, tq->tx_ring.next2fill);
tq                979 drivers/net/vmxnet3/vmxnet3_drv.c vmxnet3_tq_xmit(struct sk_buff *skb, struct vmxnet3_tx_queue *tq,
tq               1004 drivers/net/vmxnet3/vmxnet3_drv.c 				tq->stats.drop_tso++;
tq               1007 drivers/net/vmxnet3/vmxnet3_drv.c 			tq->stats.copy_skb_header++;
tq               1017 drivers/net/vmxnet3/vmxnet3_drv.c 				tq->stats.drop_too_many_frags++;
tq               1020 drivers/net/vmxnet3/vmxnet3_drv.c 			tq->stats.linearized++;
tq               1027 drivers/net/vmxnet3/vmxnet3_drv.c 	ret = vmxnet3_parse_hdr(skb, tq, &ctx, adapter);
tq               1034 drivers/net/vmxnet3/vmxnet3_drv.c 				tq->stats.drop_oversized_hdr++;
tq               1042 drivers/net/vmxnet3/vmxnet3_drv.c 					tq->stats.drop_oversized_hdr++;
tq               1048 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->stats.drop_hdr_inspect_err++;
tq               1052 drivers/net/vmxnet3/vmxnet3_drv.c 	spin_lock_irqsave(&tq->tx_lock, flags);
tq               1054 drivers/net/vmxnet3/vmxnet3_drv.c 	if (count > vmxnet3_cmd_ring_desc_avail(&tq->tx_ring)) {
tq               1055 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->stats.tx_ring_full++;
tq               1059 drivers/net/vmxnet3/vmxnet3_drv.c 			tq->tx_ring.next2comp, tq->tx_ring.next2fill);
tq               1061 drivers/net/vmxnet3/vmxnet3_drv.c 		vmxnet3_tq_stop(tq, adapter);
tq               1062 drivers/net/vmxnet3/vmxnet3_drv.c 		spin_unlock_irqrestore(&tq->tx_lock, flags);
tq               1067 drivers/net/vmxnet3/vmxnet3_drv.c 	vmxnet3_copy_hdr(skb, tq, &ctx, adapter);
tq               1070 drivers/net/vmxnet3/vmxnet3_drv.c 	if (vmxnet3_map_pkt(skb, &ctx, tq, adapter->pdev, adapter))
tq               1084 drivers/net/vmxnet3/vmxnet3_drv.c 	tx_num_deferred = le32_to_cpu(tq->shared->txNumDeferred);
tq               1102 drivers/net/vmxnet3/vmxnet3_drv.c 	le32_add_cpu(&tq->shared->txNumDeferred, num_pkts);
tq               1129 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->tx_ring.base), le64_to_cpu(gdesc->txd.addr),
tq               1132 drivers/net/vmxnet3/vmxnet3_drv.c 	spin_unlock_irqrestore(&tq->tx_lock, flags);
tq               1134 drivers/net/vmxnet3/vmxnet3_drv.c 	if (tx_num_deferred >= le32_to_cpu(tq->shared->txThreshold)) {
tq               1135 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->shared->txNumDeferred = 0;
tq               1137 drivers/net/vmxnet3/vmxnet3_drv.c 				       VMXNET3_REG_TXPROD + tq->qid * 8,
tq               1138 drivers/net/vmxnet3/vmxnet3_drv.c 				       tq->tx_ring.next2fill);
tq               1144 drivers/net/vmxnet3/vmxnet3_drv.c 	spin_unlock_irqrestore(&tq->tx_lock, flags);
tq               1146 drivers/net/vmxnet3/vmxnet3_drv.c 	tq->stats.drop_total++;
tq               1911 drivers/net/vmxnet3/vmxnet3_drv.c 		struct vmxnet3_tx_queue *tq =
tq               1913 drivers/net/vmxnet3/vmxnet3_drv.c 		vmxnet3_tq_tx_complete(tq, adapter);
tq               1936 drivers/net/vmxnet3/vmxnet3_drv.c 	struct vmxnet3_tx_queue *tq = data;
tq               1937 drivers/net/vmxnet3/vmxnet3_drv.c 	struct vmxnet3_adapter *adapter = tq->adapter;
tq               1940 drivers/net/vmxnet3/vmxnet3_drv.c 		vmxnet3_disable_intr(adapter, tq->comp_ring.intr_idx);
tq               1950 drivers/net/vmxnet3/vmxnet3_drv.c 		vmxnet3_tq_tx_complete(tq, adapter);
tq               1952 drivers/net/vmxnet3/vmxnet3_drv.c 	vmxnet3_enable_intr(adapter, tq->comp_ring.intr_idx);
tq               2444 drivers/net/vmxnet3/vmxnet3_drv.c 		struct vmxnet3_tx_queue	*tq = &adapter->tx_queue[i];
tq               2447 drivers/net/vmxnet3/vmxnet3_drv.c 		tqc->txRingBasePA   = cpu_to_le64(tq->tx_ring.basePA);
tq               2448 drivers/net/vmxnet3/vmxnet3_drv.c 		tqc->dataRingBasePA = cpu_to_le64(tq->data_ring.basePA);
tq               2449 drivers/net/vmxnet3/vmxnet3_drv.c 		tqc->compRingBasePA = cpu_to_le64(tq->comp_ring.basePA);
tq               2450 drivers/net/vmxnet3/vmxnet3_drv.c 		tqc->ddPA           = cpu_to_le64(tq->buf_info_pa);
tq               2451 drivers/net/vmxnet3/vmxnet3_drv.c 		tqc->txRingSize     = cpu_to_le32(tq->tx_ring.size);
tq               2452 drivers/net/vmxnet3/vmxnet3_drv.c 		tqc->dataRingSize   = cpu_to_le32(tq->data_ring.size);
tq               2453 drivers/net/vmxnet3/vmxnet3_drv.c 		tqc->txDataRingDescSize = cpu_to_le32(tq->txdata_desc_size);
tq               2454 drivers/net/vmxnet3/vmxnet3_drv.c 		tqc->compRingSize   = cpu_to_le32(tq->comp_ring.size);
tq               2458 drivers/net/vmxnet3/vmxnet3_drv.c 		tqc->intrIdx        = tq->comp_ring.intr_idx;
tq               2827 drivers/net/vmxnet3/vmxnet3_drv.c 		struct vmxnet3_tx_queue	*tq = &adapter->tx_queue[i];
tq               2828 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->tx_ring.size   = tx_ring_size;
tq               2829 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->data_ring.size = tx_ring_size;
tq               2830 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->comp_ring.size = tx_ring_size;
tq               2831 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->txdata_desc_size = txdata_desc_size;
tq               2832 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->shared = &adapter->tqd_start[i].ctrl;
tq               2833 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->stopped = true;
tq               2834 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->adapter = adapter;
tq               2835 drivers/net/vmxnet3/vmxnet3_drv.c 		tq->qid = i;
tq               2836 drivers/net/vmxnet3/vmxnet3_drv.c 		err = vmxnet3_tq_create(tq, adapter);
tq                392 drivers/net/vmxnet3/vmxnet3_ethtool.c 		struct vmxnet3_tx_queue *tq = &adapter->tx_queue[i];
tq                397 drivers/net/vmxnet3/vmxnet3_ethtool.c 		buf[j++] = VMXNET3_GET_ADDR_LO(tq->tx_ring.basePA);
tq                398 drivers/net/vmxnet3/vmxnet3_ethtool.c 		buf[j++] = VMXNET3_GET_ADDR_HI(tq->tx_ring.basePA);
tq                399 drivers/net/vmxnet3/vmxnet3_ethtool.c 		buf[j++] = tq->tx_ring.size;
tq                400 drivers/net/vmxnet3/vmxnet3_ethtool.c 		buf[j++] = tq->tx_ring.next2fill;
tq                401 drivers/net/vmxnet3/vmxnet3_ethtool.c 		buf[j++] = tq->tx_ring.next2comp;
tq                402 drivers/net/vmxnet3/vmxnet3_ethtool.c 		buf[j++] = tq->tx_ring.gen;
tq                404 drivers/net/vmxnet3/vmxnet3_ethtool.c 		buf[j++] = VMXNET3_GET_ADDR_LO(tq->data_ring.basePA);
tq                405 drivers/net/vmxnet3/vmxnet3_ethtool.c 		buf[j++] = VMXNET3_GET_ADDR_HI(tq->data_ring.basePA);
tq                406 drivers/net/vmxnet3/vmxnet3_ethtool.c 		buf[j++] = tq->data_ring.size;
tq                407 drivers/net/vmxnet3/vmxnet3_ethtool.c 		buf[j++] = tq->txdata_desc_size;
tq                409 drivers/net/vmxnet3/vmxnet3_ethtool.c 		buf[j++] = VMXNET3_GET_ADDR_LO(tq->comp_ring.basePA);
tq                410 drivers/net/vmxnet3/vmxnet3_ethtool.c 		buf[j++] = VMXNET3_GET_ADDR_HI(tq->comp_ring.basePA);
tq                411 drivers/net/vmxnet3/vmxnet3_ethtool.c 		buf[j++] = tq->comp_ring.size;
tq                412 drivers/net/vmxnet3/vmxnet3_ethtool.c 		buf[j++] = tq->comp_ring.next2proc;
tq                413 drivers/net/vmxnet3/vmxnet3_ethtool.c 		buf[j++] = tq->comp_ring.gen;
tq                415 drivers/net/vmxnet3/vmxnet3_ethtool.c 		buf[j++] = tq->stopped;
tq                404 drivers/net/vmxnet3/vmxnet3_int.h #define VMXNET3_WAKE_QUEUE_THRESHOLD(tq)  (5)
tq                286 drivers/net/wireless/ath/ath5k/qcu.c 		struct ath5k_txq_info *tq = &ah->ah_txq[queue];
tq                292 drivers/net/wireless/ath/ath5k/qcu.c 			(tq->tqi_cw_min << AR5K_NODCU_RETRY_LMT_CW_MIN_S)
tq                326 drivers/net/wireless/ath/ath5k/qcu.c 	struct ath5k_txq_info *tq = &ah->ah_txq[queue];
tq                333 drivers/net/wireless/ath/ath5k/qcu.c 	(tq->tqi_type == AR5K_TX_QUEUE_INACTIVE))
tq                341 drivers/net/wireless/ath/ath5k/qcu.c 		AR5K_REG_SM(tq->tqi_cw_min, AR5K_DCU_LCL_IFS_CW_MIN) |
tq                342 drivers/net/wireless/ath/ath5k/qcu.c 		AR5K_REG_SM(tq->tqi_cw_max, AR5K_DCU_LCL_IFS_CW_MAX) |
tq                343 drivers/net/wireless/ath/ath5k/qcu.c 		AR5K_REG_SM(tq->tqi_aifs, AR5K_DCU_LCL_IFS_AIFS),
tq                366 drivers/net/wireless/ath/ath5k/qcu.c 	if (tq->tqi_cbr_period) {
tq                367 drivers/net/wireless/ath/ath5k/qcu.c 		ath5k_hw_reg_write(ah, AR5K_REG_SM(tq->tqi_cbr_period,
tq                369 drivers/net/wireless/ath/ath5k/qcu.c 					AR5K_REG_SM(tq->tqi_cbr_overflow_limit,
tq                376 drivers/net/wireless/ath/ath5k/qcu.c 		if (tq->tqi_cbr_overflow_limit)
tq                382 drivers/net/wireless/ath/ath5k/qcu.c 	if (tq->tqi_ready_time && (tq->tqi_type != AR5K_TX_QUEUE_CAB))
tq                383 drivers/net/wireless/ath/ath5k/qcu.c 		ath5k_hw_reg_write(ah, AR5K_REG_SM(tq->tqi_ready_time,
tq                388 drivers/net/wireless/ath/ath5k/qcu.c 	if (tq->tqi_burst_time) {
tq                389 drivers/net/wireless/ath/ath5k/qcu.c 		ath5k_hw_reg_write(ah, AR5K_REG_SM(tq->tqi_burst_time,
tq                394 drivers/net/wireless/ath/ath5k/qcu.c 		if (tq->tqi_flags & AR5K_TXQ_FLAG_RDYTIME_EXP_POLICY_ENABLE)
tq                400 drivers/net/wireless/ath/ath5k/qcu.c 	if (tq->tqi_flags & AR5K_TXQ_FLAG_BACKOFF_DISABLE)
tq                405 drivers/net/wireless/ath/ath5k/qcu.c 	if (tq->tqi_flags & AR5K_TXQ_FLAG_FRAG_BURST_BACKOFF_ENABLE)
tq                412 drivers/net/wireless/ath/ath5k/qcu.c 	switch (tq->tqi_type) {
tq                434 drivers/net/wireless/ath/ath5k/qcu.c 		ath5k_hw_reg_write(ah, ((tq->tqi_ready_time -
tq                462 drivers/net/wireless/ath/ath5k/qcu.c 	if (tq->tqi_flags & AR5K_TXQ_FLAG_TXOKINT_ENABLE)
tq                465 drivers/net/wireless/ath/ath5k/qcu.c 	if (tq->tqi_flags & AR5K_TXQ_FLAG_TXERRINT_ENABLE)
tq                468 drivers/net/wireless/ath/ath5k/qcu.c 	if (tq->tqi_flags & AR5K_TXQ_FLAG_TXURNINT_ENABLE)
tq                471 drivers/net/wireless/ath/ath5k/qcu.c 	if (tq->tqi_flags & AR5K_TXQ_FLAG_TXDESCINT_ENABLE)
tq                474 drivers/net/wireless/ath/ath5k/qcu.c 	if (tq->tqi_flags & AR5K_TXQ_FLAG_TXEOLINT_ENABLE)
tq                477 drivers/net/wireless/ath/ath5k/qcu.c 	if (tq->tqi_flags & AR5K_TXQ_FLAG_CBRORNINT_ENABLE)
tq                480 drivers/net/wireless/ath/ath5k/qcu.c 	if (tq->tqi_flags & AR5K_TXQ_FLAG_CBRURNINT_ENABLE)
tq                483 drivers/net/wireless/ath/ath5k/qcu.c 	if (tq->tqi_flags & AR5K_TXQ_FLAG_QTRIGINT_ENABLE)
tq                486 drivers/net/wireless/ath/ath5k/qcu.c 	if (tq->tqi_flags & AR5K_TXQ_FLAG_TXNOFRMINT_ENABLE)
tq                300 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(tq-full,				0x08),
tq                403 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(tq-bank-conflict-issue-stall,	0x0b),
tq                404 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(tq-full,				0x0c),
tq               1735 drivers/tty/ipwireless/hardware.c 	struct ipw_tx_packet *tp, *tq;
tq               1745 drivers/tty/ipwireless/hardware.c 		list_for_each_entry_safe(tp, tq, &hw->tx_queue[i], queue) {
tq                281 include/linux/hp_sdc.h 	hp_sdc_transaction *tq[HP_SDC_QUEUE_LEN]; /* All pending read/writes */
tq                217 include/uapi/linux/batadv_packet.h 	__u8   tq;
tq                 34 include/uapi/linux/can/netlink.h 	__u32 tq;		/* Time quanta (TQ) in nanoseconds */
tq                260 ipc/sem.c      	struct sem_queue *q, *tq;
tq                270 ipc/sem.c      	list_for_each_entry_safe(q, tq, &sma->pending_alter, list) {
tq               1127 ipc/sem.c      	struct sem_queue *q, *tq;
tq               1144 ipc/sem.c      	list_for_each_entry_safe(q, tq, &sma->pending_const, list) {
tq               1149 ipc/sem.c      	list_for_each_entry_safe(q, tq, &sma->pending_alter, list) {
tq               1155 ipc/sem.c      		list_for_each_entry_safe(q, tq, &sem->pending_const, list) {
tq               1159 ipc/sem.c      		list_for_each_entry_safe(q, tq, &sem->pending_alter, list) {
tq                219 net/batman-adv/bat_iv_ogm.c 	batadv_ogm_packet->tq = BATADV_TQ_MAX_VALUE;
tq                295 net/batman-adv/bat_iv_ogm.c static u8 batadv_hop_penalty(u8 tq, const struct batadv_priv *bat_priv)
tq                300 net/batman-adv/bat_iv_ogm.c 	new_tq = tq * (BATADV_TQ_MAX_VALUE - hop_penalty);
tq                374 net/batman-adv/bat_iv_ogm.c 			   batadv_ogm_packet->tq, batadv_ogm_packet->ttl,
tq                716 net/batman-adv/bat_iv_ogm.c 	batadv_ogm_packet->tq = batadv_hop_penalty(batadv_ogm_packet->tq,
tq                721 net/batman-adv/bat_iv_ogm.c 		   batadv_ogm_packet->tq, batadv_ogm_packet->ttl);
tq                999 net/batman-adv/bat_iv_ogm.c 			       batadv_ogm_packet->tq);
tq               1171 net/batman-adv/bat_iv_ogm.c 	combined_tq = batadv_ogm_packet->tq *
tq               1178 net/batman-adv/bat_iv_ogm.c 	batadv_ogm_packet->tq = combined_tq;
tq               1184 net/batman-adv/bat_iv_ogm.c 		   batadv_ogm_packet->tq, if_incoming->net_dev->name,
tq               1190 net/batman-adv/bat_iv_ogm.c 	if (batadv_ogm_packet->tq >= BATADV_TQ_TOTAL_BIDRECT_LIMIT)
tq               1357 net/batman-adv/bat_iv_ogm.c 	if (ogm_packet->tq == 0) {
tq               1604 net/batman-adv/bat_iv_ogm.c 		   ogm_packet->tq, ogm_packet->ttl,
tq                778 net/batman-adv/network-coding.c 	if (ogm_packet->tq < bat_priv->nc.min_tq)
tq               1010 net/batman-adv/network-coding.c static u8 batadv_nc_random_weight_tq(u8 tq)
tq               1013 net/batman-adv/network-coding.c 	u8 rand_tq = prandom_u32_max(BATADV_TQ_MAX_VALUE + 1 - tq);
tq                284 net/tipc/trace.h 	TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq),
tq                286 net/tipc/trace.h 	TP_ARGS(r, f, t, tq),
tq                301 net/tipc/trace.h 		__entry->len = skb_queue_len(tq);
tq                302 net/tipc/trace.h 		__entry->fseqno = msg_seqno(buf_msg(skb_peek(tq)));
tq                303 net/tipc/trace.h 		__entry->lseqno = msg_seqno(buf_msg(skb_peek_tail(tq)));
tq                312 net/tipc/trace.h 	TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq),
tq                313 net/tipc/trace.h 	TP_ARGS(r, f, t, tq)
tq                317 net/tipc/trace.h 	TP_PROTO(struct tipc_link *r, u16 f, u16 t, struct sk_buff_head *tq),
tq                318 net/tipc/trace.h 	TP_ARGS(r, f, t, tq),