tx                238 arch/mips/include/asm/ip32/mace.h 	volatile unsigned long tx;
tx               1064 arch/mips/include/asm/octeon/cvmx-agl-defs.h 		uint64_t tx:2;
tx               1070 arch/mips/include/asm/octeon/cvmx-agl-defs.h 		uint64_t tx:2;
tx               1077 arch/mips/include/asm/octeon/cvmx-agl-defs.h 		uint64_t tx:1;
tx               1083 arch/mips/include/asm/octeon/cvmx-agl-defs.h 		uint64_t tx:1;
tx                216 arch/mips/include/asm/processor.h 	u64	tx[4];
tx                 55 arch/mips/netlogic/xlp/cop2-ex.c 		: "r"(r->tx), "r"(r->rx));
tx                 85 arch/mips/netlogic/xlp/cop2-ex.c 		: : "m"(*r), "r"(r->tx), "r"(r->rx));
tx                 27 arch/s390/include/uapi/asm/guarded_storage.h 			__u8 tx	: 1;
tx               4226 arch/x86/events/intel/core.c EVENT_ATTR_STR(tx-start,	tx_start,	"event=0xc9,umask=0x1");
tx               4227 arch/x86/events/intel/core.c EVENT_ATTR_STR(tx-commit,	tx_commit,	"event=0xc9,umask=0x2");
tx               4228 arch/x86/events/intel/core.c EVENT_ATTR_STR(tx-abort,	tx_abort,	"event=0xc9,umask=0x4");
tx               4229 arch/x86/events/intel/core.c EVENT_ATTR_STR(tx-capacity,	tx_capacity,	"event=0x54,umask=0x2");
tx               4230 arch/x86/events/intel/core.c EVENT_ATTR_STR(tx-conflict,	tx_conflict,	"event=0x54,umask=0x1");
tx               4272 arch/x86/events/intel/core.c EVENT_ATTR_STR(tx-capacity-read,  tx_capacity_read,  "event=0x54,umask=0x80");
tx               4273 arch/x86/events/intel/core.c EVENT_ATTR_STR(tx-capacity-write, tx_capacity_write, "event=0x54,umask=0x2");
tx                 39 crypto/async_tx/async_memcpy.c 	struct dma_async_tx_descriptor *tx = NULL;
tx                 61 crypto/async_tx/async_memcpy.c 		tx = device->device_prep_dma_memcpy(chan, unmap->addr[1],
tx                 66 crypto/async_tx/async_memcpy.c 	if (tx) {
tx                 69 crypto/async_tx/async_memcpy.c 		dma_set_unmap(tx, unmap);
tx                 70 crypto/async_tx/async_memcpy.c 		async_tx_submit(chan, tx, submit);
tx                 91 crypto/async_tx/async_memcpy.c 	return tx;
tx                 41 crypto/async_tx/async_pq.c 	struct dma_async_tx_descriptor *tx = NULL;
tx                 78 crypto/async_tx/async_pq.c 			tx = dma->device_prep_dma_pq(chan, dma_dest,
tx                 83 crypto/async_tx/async_pq.c 			if (likely(tx))
tx                 89 crypto/async_tx/async_pq.c 		dma_set_unmap(tx, unmap);
tx                 90 crypto/async_tx/async_pq.c 		async_tx_submit(chan, tx, submit);
tx                 91 crypto/async_tx/async_pq.c 		submit->depend_tx = tx;
tx                100 crypto/async_tx/async_pq.c 	return tx;
tx                183 crypto/async_tx/async_pq.c 		struct dma_async_tx_descriptor *tx;
tx                228 crypto/async_tx/async_pq.c 		tx = do_async_gen_syndrome(chan, coefs, j, unmap, dma_flags, submit);
tx                230 crypto/async_tx/async_pq.c 		return tx;
tx                287 crypto/async_tx/async_pq.c 	struct dma_async_tx_descriptor *tx;
tx                342 crypto/async_tx/async_pq.c 			tx = device->device_prep_dma_pq_val(chan, pq,
tx                348 crypto/async_tx/async_pq.c 			if (likely(tx))
tx                354 crypto/async_tx/async_pq.c 		dma_set_unmap(tx, unmap);
tx                355 crypto/async_tx/async_pq.c 		async_tx_submit(chan, tx, submit);
tx                379 crypto/async_tx/async_pq.c 		tx = NULL;
tx                384 crypto/async_tx/async_pq.c 			tx = async_xor(spare, blocks, offset, disks-2, len, submit);
tx                385 crypto/async_tx/async_pq.c 			async_tx_quiesce(&tx);
tx                395 crypto/async_tx/async_pq.c 			tx = async_gen_syndrome(blocks, offset, disks, len, submit);
tx                396 crypto/async_tx/async_pq.c 			async_tx_quiesce(&tx);
tx                410 crypto/async_tx/async_pq.c 		tx = NULL;
tx                414 crypto/async_tx/async_pq.c 	return tx;
tx                 35 crypto/async_tx/async_raid6_recov.c 		struct dma_async_tx_descriptor *tx;
tx                 50 crypto/async_tx/async_raid6_recov.c 		tx = dma->device_prep_dma_pq(chan, pq, unmap->addr, 2, coef,
tx                 52 crypto/async_tx/async_raid6_recov.c 		if (tx) {
tx                 53 crypto/async_tx/async_raid6_recov.c 			dma_set_unmap(tx, unmap);
tx                 54 crypto/async_tx/async_raid6_recov.c 			async_tx_submit(chan, tx, submit);
tx                 56 crypto/async_tx/async_raid6_recov.c 			return tx;
tx                 99 crypto/async_tx/async_raid6_recov.c 		struct dma_async_tx_descriptor *tx;
tx                115 crypto/async_tx/async_raid6_recov.c 		tx = dma->device_prep_dma_pq(chan, dma_dest, unmap->addr,
tx                118 crypto/async_tx/async_raid6_recov.c 		if (tx) {
tx                119 crypto/async_tx/async_raid6_recov.c 			dma_set_unmap(tx, unmap);
tx                121 crypto/async_tx/async_raid6_recov.c 			async_tx_submit(chan, tx, submit);
tx                122 crypto/async_tx/async_raid6_recov.c 			return tx;
tx                149 crypto/async_tx/async_raid6_recov.c 	struct dma_async_tx_descriptor *tx = NULL;
tx                170 crypto/async_tx/async_raid6_recov.c 	init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble);
tx                171 crypto/async_tx/async_raid6_recov.c 	tx = async_sum_product(b, srcs, coef, bytes, submit);
tx                176 crypto/async_tx/async_raid6_recov.c 	init_async_submit(submit, flags | ASYNC_TX_XOR_ZERO_DST, tx, cb_fn,
tx                178 crypto/async_tx/async_raid6_recov.c 	tx = async_xor(a, srcs, 0, 2, bytes, submit);
tx                180 crypto/async_tx/async_raid6_recov.c 	return tx;
tx                188 crypto/async_tx/async_raid6_recov.c 	struct dma_async_tx_descriptor *tx = NULL;
tx                221 crypto/async_tx/async_raid6_recov.c 	init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble);
tx                222 crypto/async_tx/async_raid6_recov.c 	tx = async_memcpy(dp, g, 0, 0, bytes, submit);
tx                223 crypto/async_tx/async_raid6_recov.c 	init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble);
tx                224 crypto/async_tx/async_raid6_recov.c 	tx = async_mult(dq, g, raid6_gfexp[good], bytes, submit);
tx                229 crypto/async_tx/async_raid6_recov.c 	init_async_submit(submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_DROP_DST, tx,
tx                231 crypto/async_tx/async_raid6_recov.c 	tx = async_xor(dp, srcs, 0, 2, bytes, submit);
tx                236 crypto/async_tx/async_raid6_recov.c 	init_async_submit(submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_DROP_DST, tx,
tx                238 crypto/async_tx/async_raid6_recov.c 	tx = async_xor(dq, srcs, 0, 2, bytes, submit);
tx                245 crypto/async_tx/async_raid6_recov.c 	init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble);
tx                246 crypto/async_tx/async_raid6_recov.c 	tx = async_sum_product(dq, srcs, coef, bytes, submit);
tx                251 crypto/async_tx/async_raid6_recov.c 	init_async_submit(submit, flags | ASYNC_TX_XOR_DROP_DST, tx, cb_fn,
tx                253 crypto/async_tx/async_raid6_recov.c 	tx = async_xor(dp, srcs, 0, 2, bytes, submit);
tx                255 crypto/async_tx/async_raid6_recov.c 	return tx;
tx                262 crypto/async_tx/async_raid6_recov.c 	struct dma_async_tx_descriptor *tx = NULL;
tx                285 crypto/async_tx/async_raid6_recov.c 	init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble);
tx                286 crypto/async_tx/async_raid6_recov.c 	tx = async_gen_syndrome(blocks, 0, disks, bytes, submit);
tx                297 crypto/async_tx/async_raid6_recov.c 	init_async_submit(submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_DROP_DST, tx,
tx                299 crypto/async_tx/async_raid6_recov.c 	tx = async_xor(dp, srcs, 0, 2, bytes, submit);
tx                304 crypto/async_tx/async_raid6_recov.c 	init_async_submit(submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_DROP_DST, tx,
tx                306 crypto/async_tx/async_raid6_recov.c 	tx = async_xor(dq, srcs, 0, 2, bytes, submit);
tx                313 crypto/async_tx/async_raid6_recov.c 	init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble);
tx                314 crypto/async_tx/async_raid6_recov.c 	tx = async_sum_product(dq, srcs, coef, bytes, submit);
tx                319 crypto/async_tx/async_raid6_recov.c 	init_async_submit(submit, flags | ASYNC_TX_XOR_DROP_DST, tx, cb_fn,
tx                321 crypto/async_tx/async_raid6_recov.c 	tx = async_xor(dp, srcs, 0, 2, bytes, submit);
tx                323 crypto/async_tx/async_raid6_recov.c 	return tx;
tx                412 crypto/async_tx/async_raid6_recov.c 	struct dma_async_tx_descriptor *tx = NULL;
tx                476 crypto/async_tx/async_raid6_recov.c 		init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL,
tx                478 crypto/async_tx/async_raid6_recov.c 		tx = async_memcpy(p, g, 0, 0, bytes, submit);
tx                480 crypto/async_tx/async_raid6_recov.c 		init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL,
tx                482 crypto/async_tx/async_raid6_recov.c 		tx = async_mult(dq, g, raid6_gfexp[good], bytes, submit);
tx                484 crypto/async_tx/async_raid6_recov.c 		init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL,
tx                486 crypto/async_tx/async_raid6_recov.c 		tx = async_gen_syndrome(blocks, 0, disks, bytes, submit);
tx                498 crypto/async_tx/async_raid6_recov.c 	init_async_submit(submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_DROP_DST, tx,
tx                500 crypto/async_tx/async_raid6_recov.c 	tx = async_xor(dq, srcs, 0, 2, bytes, submit);
tx                502 crypto/async_tx/async_raid6_recov.c 	init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble);
tx                503 crypto/async_tx/async_raid6_recov.c 	tx = async_mult(dq, dq, coef, bytes, submit);
tx                507 crypto/async_tx/async_raid6_recov.c 	init_async_submit(submit, flags | ASYNC_TX_XOR_DROP_DST, tx, cb_fn,
tx                509 crypto/async_tx/async_raid6_recov.c 	tx = async_xor(p, srcs, 0, 2, bytes, submit);
tx                511 crypto/async_tx/async_raid6_recov.c 	return tx;
tx                 66 crypto/async_tx/async_tx.c 			struct dma_async_tx_descriptor *tx)
tx                 74 crypto/async_tx/async_tx.c 	if (txd_parent(depend_tx) && depend_tx->chan == tx->chan) {
tx                 75 crypto/async_tx/async_tx.c 		txd_chain(depend_tx, tx);
tx                100 crypto/async_tx/async_tx.c 		txd_chain(intr_tx, tx);
tx                121 crypto/async_tx/async_tx.c 		tx->tx_submit(tx);
tx                143 crypto/async_tx/async_tx.c async_tx_submit(struct dma_chan *chan, struct dma_async_tx_descriptor *tx,
tx                148 crypto/async_tx/async_tx.c 	tx->callback = submit->cb_fn;
tx                149 crypto/async_tx/async_tx.c 	tx->callback_param = submit->cb_param;
tx                161 crypto/async_tx/async_tx.c 		       txd_parent(tx));
tx                173 crypto/async_tx/async_tx.c 				txd_chain(depend_tx, tx);
tx                192 crypto/async_tx/async_tx.c 			async_tx_channel_switch(depend_tx, tx);
tx                195 crypto/async_tx/async_tx.c 			txd_clear_parent(tx);
tx                196 crypto/async_tx/async_tx.c 			tx->tx_submit(tx);
tx                200 crypto/async_tx/async_tx.c 		txd_clear_parent(tx);
tx                201 crypto/async_tx/async_tx.c 		tx->tx_submit(tx);
tx                205 crypto/async_tx/async_tx.c 		async_tx_ack(tx);
tx                225 crypto/async_tx/async_tx.c 	struct dma_async_tx_descriptor *tx;
tx                238 crypto/async_tx/async_tx.c 		tx = device ? device->device_prep_dma_interrupt(chan, 0) : NULL;
tx                240 crypto/async_tx/async_tx.c 		tx = NULL;
tx                242 crypto/async_tx/async_tx.c 	if (tx) {
tx                245 crypto/async_tx/async_tx.c 		async_tx_submit(chan, tx, submit);
tx                255 crypto/async_tx/async_tx.c 	return tx;
tx                263 crypto/async_tx/async_tx.c void async_tx_quiesce(struct dma_async_tx_descriptor **tx)
tx                265 crypto/async_tx/async_tx.c 	if (*tx) {
tx                269 crypto/async_tx/async_tx.c 		BUG_ON(async_tx_test_ack(*tx));
tx                270 crypto/async_tx/async_tx.c 		if (dma_wait_for_async_tx(*tx) != DMA_COMPLETE)
tx                273 crypto/async_tx/async_tx.c 		async_tx_ack(*tx);
tx                274 crypto/async_tx/async_tx.c 		*tx = NULL;
tx                 27 crypto/async_tx/async_xor.c 	struct dma_async_tx_descriptor *tx = NULL;
tx                 65 crypto/async_tx/async_xor.c 		tx = dma->device_prep_dma_xor(chan, dma_dest, src_list,
tx                 69 crypto/async_tx/async_xor.c 		if (unlikely(!tx))
tx                 73 crypto/async_tx/async_xor.c 		while (unlikely(!tx)) {
tx                 75 crypto/async_tx/async_xor.c 			tx = dma->device_prep_dma_xor(chan, dma_dest,
tx                 82 crypto/async_tx/async_xor.c 		dma_set_unmap(tx, unmap);
tx                 83 crypto/async_tx/async_xor.c 		async_tx_submit(chan, tx, submit);
tx                 84 crypto/async_tx/async_xor.c 		submit->depend_tx = tx;
tx                 96 crypto/async_tx/async_xor.c 	return tx;
tx                175 crypto/async_tx/async_xor.c 		struct dma_async_tx_descriptor *tx;
tx                195 crypto/async_tx/async_xor.c 		tx = do_async_xor(chan, unmap, submit);
tx                197 crypto/async_tx/async_xor.c 		return tx;
tx                262 crypto/async_tx/async_xor.c 	struct dma_async_tx_descriptor *tx = NULL;
tx                289 crypto/async_tx/async_xor.c 		tx = device->device_prep_dma_xor_val(chan, unmap->addr, src_cnt,
tx                292 crypto/async_tx/async_xor.c 		if (unlikely(!tx)) {
tx                295 crypto/async_tx/async_xor.c 			while (!tx) {
tx                297 crypto/async_tx/async_xor.c 				tx = device->device_prep_dma_xor_val(chan,
tx                302 crypto/async_tx/async_xor.c 		dma_set_unmap(tx, unmap);
tx                303 crypto/async_tx/async_xor.c 		async_tx_submit(chan, tx, submit);
tx                315 crypto/async_tx/async_xor.c 		tx = async_xor(dest, src_list, offset, src_cnt, len, submit);
tx                317 crypto/async_tx/async_xor.c 		async_tx_quiesce(&tx);
tx                326 crypto/async_tx/async_xor.c 	return tx;
tx                 59 crypto/async_tx/raid6test.c 	struct dma_async_tx_descriptor *tx = NULL;
tx                 69 crypto/async_tx/raid6test.c 			tx = async_gen_syndrome(ptrs, 0, disks, bytes, &submit);
tx                 89 crypto/async_tx/raid6test.c 			tx = async_xor(dest, blocks, 0, count, bytes, &submit);
tx                 91 crypto/async_tx/raid6test.c 			init_async_submit(&submit, 0, tx, NULL, NULL, addr_conv);
tx                 92 crypto/async_tx/raid6test.c 			tx = async_gen_syndrome(ptrs, 0, disks, bytes, &submit);
tx                 98 crypto/async_tx/raid6test.c 			tx = async_raid6_datap_recov(disks, bytes, faila, ptrs, &submit);
tx                102 crypto/async_tx/raid6test.c 			tx = async_raid6_2data_recov(disks, bytes, faila, failb, ptrs, &submit);
tx                106 crypto/async_tx/raid6test.c 	init_async_submit(&submit, ASYNC_TX_ACK, tx, callback, &cmp, addr_conv);
tx                107 crypto/async_tx/raid6test.c 	tx = async_syndrome_val(ptrs, 0, disks, bytes, &result, spare, &submit);
tx                108 crypto/async_tx/raid6test.c 	async_tx_issue_pending(tx);
tx                146 crypto/async_tx/raid6test.c 	struct dma_async_tx_descriptor *tx;
tx                165 crypto/async_tx/raid6test.c 	tx = async_gen_syndrome(dataptrs, 0, disks, PAGE_SIZE, &submit);
tx                166 crypto/async_tx/raid6test.c 	async_tx_issue_pending(tx);
tx               1267 crypto/ecc.c   			u64 tx[ECC_MAX_DIGITS];
tx               1271 crypto/ecc.c   			vli_set(tx, point->x, ndigits);
tx               1273 crypto/ecc.c   			apply_z(tx, ty, z, curve->p, ndigits);
tx               1274 crypto/ecc.c   			vli_mod_sub(tz, rx, tx, curve->p, ndigits);
tx               1275 crypto/ecc.c   			xycz_add(tx, ty, rx, ry, curve->p, ndigits);
tx                237 drivers/acpi/processor_throttling.c 		if (p_limit->thermal.tx > target_state)
tx                238 drivers/acpi/processor_throttling.c 			target_state = p_limit->thermal.tx;
tx                239 drivers/acpi/processor_throttling.c 		if (p_limit->user.tx > target_state)
tx                240 drivers/acpi/processor_throttling.c 			target_state = p_limit->user.tx;
tx                333 drivers/acpi/processor_throttling.c 		if (limit->thermal.tx > target_state)
tx                334 drivers/acpi/processor_throttling.c 			target_state = limit->thermal.tx;
tx                335 drivers/acpi/processor_throttling.c 		if (limit->user.tx > target_state)
tx                336 drivers/acpi/processor_throttling.c 			target_state = limit->user.tx;
tx                534 drivers/acpi/processor_throttling.c 		struct acpi_processor_tx_tss *tx =
tx                539 drivers/acpi/processor_throttling.c 		state.pointer = tx;
tx                552 drivers/acpi/processor_throttling.c 		if (!tx->freqpercentage) {
tx                836 drivers/acpi/processor_throttling.c 		struct acpi_processor_tx_tss *tx =
tx                839 drivers/acpi/processor_throttling.c 		if (tx->control == value)
tx                851 drivers/acpi/processor_throttling.c 		struct acpi_processor_tx_tss *tx =
tx                854 drivers/acpi/processor_throttling.c 		*value = tx->control;
tx                397 drivers/ata/pata_arasan_cf.c 	struct dma_async_tx_descriptor *tx;
tx                403 drivers/ata/pata_arasan_cf.c 	tx = chan->device->device_prep_dma_memcpy(chan, dest, src, len, flags);
tx                404 drivers/ata/pata_arasan_cf.c 	if (!tx) {
tx                409 drivers/ata/pata_arasan_cf.c 	tx->callback = dma_callback;
tx                410 drivers/ata/pata_arasan_cf.c 	tx->callback_param = acdev;
tx                411 drivers/ata/pata_arasan_cf.c 	cookie = tx->tx_submit(tx);
tx                 50 drivers/ata/pata_pxa.c 	struct dma_async_tx_descriptor *tx;
tx                 57 drivers/ata/pata_pxa.c 	tx = dmaengine_prep_slave_sg(pd->dma_chan, qc->sg, qc->n_elem, dir,
tx                 59 drivers/ata/pata_pxa.c 	if (!tx) {
tx                 63 drivers/ata/pata_pxa.c 	tx->callback = pxa_ata_dma_irq;
tx                 64 drivers/ata/pata_pxa.c 	tx->callback_param = pd;
tx                 65 drivers/ata/pata_pxa.c 	pd->dma_cookie = dmaengine_submit(tx);
tx                118 drivers/atm/adummy.c 	atomic_inc(&vcc->stats->tx);
tx                435 drivers/atm/ambassador.c static void tx_complete (amb_dev * dev, tx_out * tx) {
tx                436 drivers/atm/ambassador.c   tx_simple * tx_descr = bus_to_virt (tx->handle);
tx                439 drivers/atm/ambassador.c   PRINTD (DBG_FLOW|DBG_TX, "tx_complete %p %p", dev, tx);
tx                442 drivers/atm/ambassador.c   atomic_inc(&ATM_SKB(skb)->vcc->stats->tx);
tx                612 drivers/atm/ambassador.c static int tx_give (amb_dev * dev, tx_in * tx) {
tx                626 drivers/atm/ambassador.c     *txq->in.ptr = *tx;
tx               1300 drivers/atm/ambassador.c   tx_in tx;
tx               1357 drivers/atm/ambassador.c   tx.vc = cpu_to_be16 (vcc->tx_frame_bits | vc);
tx               1358 drivers/atm/ambassador.c   tx.tx_descr_length = cpu_to_be16 (sizeof(tx_frag)+sizeof(tx_frag_end));
tx               1359 drivers/atm/ambassador.c   tx.tx_descr_addr = cpu_to_be32 (virt_to_bus (&tx_descr->tx_frag));
tx               1361 drivers/atm/ambassador.c   while (tx_give (dev, &tx))
tx                229 drivers/atm/atmtcp.c 	atomic_inc(&vcc->stats->tx);
tx                316 drivers/atm/atmtcp.c 	atomic_inc(&vcc->stats->tx);
tx                189 drivers/atm/eni.c 		if (eni_dev->tx[i].send)
tx                191 drivers/atm/eni.c 			    eni_dev->tx[i].send,eni_dev->tx[i].words*4);
tx               1038 drivers/atm/eni.c 	struct eni_tx *tx;
tx               1052 drivers/atm/eni.c 	tx = eni_vcc->tx;
tx               1053 drivers/atm/eni.c 	NULLCHECK(tx);
tx               1092 drivers/atm/eni.c 	if (!NEPMOK(tx->tx_pos,size+TX_GAP,
tx               1093 drivers/atm/eni.c 	    eni_in(MID_TX_RDPTR(tx->index)),tx->words)) {
tx               1110 drivers/atm/eni.c 	DPRINTK("dma_wr is %d, tx_pos is %ld\n",dma_wr,tx->tx_pos);
tx               1122 drivers/atm/eni.c 	eni_dev->dma[j++] = (((tx->tx_pos+TX_DESCR_SIZE) & (tx->words-1)) <<
tx               1123 drivers/atm/eni.c 	     MID_DMA_COUNT_SHIFT) | (tx->index << MID_DMA_CHAN_SHIFT) |
tx               1127 drivers/atm/eni.c 		if (aal5) put_dma(tx->index,eni_dev->dma,&j,paddr,skb->len);
tx               1128 drivers/atm/eni.c 		else put_dma(tx->index,eni_dev->dma,&j,paddr+4,skb->len-4);
tx               1133 drivers/atm/eni.c 				put_dma(tx->index,eni_dev->dma,&j,(unsigned long)
tx               1137 drivers/atm/eni.c 				put_dma(tx->index,eni_dev->dma,&j,(unsigned long)
tx               1143 drivers/atm/eni.c 		put_dma(tx->index, eni_dev->dma, &j, eni_dev->zero.dma,
tx               1147 drivers/atm/eni.c 	eni_dev->dma[j++] = (((tx->tx_pos+size) & (tx->words-1)) <<
tx               1148 drivers/atm/eni.c 	     MID_DMA_COUNT_SHIFT) | (tx->index << MID_DMA_CHAN_SHIFT) |
tx               1154 drivers/atm/eni.c 	    (aal5 ? MID_SEG_AAL5 : 0) | (tx->prescaler << MID_SEG_PR_SHIFT) |
tx               1155 drivers/atm/eni.c 	    (tx->resolution << MID_SEG_RATE_SHIFT) |
tx               1156 drivers/atm/eni.c 	    (size/(ATM_CELL_PAYLOAD/4)),tx->send+tx->tx_pos*4);
tx               1161 drivers/atm/eni.c 	    tx->send+((tx->tx_pos+1) & (tx->words-1))*4);
tx               1164 drivers/atm/eni.c 		writel(skb->len,tx->send+
tx               1165 drivers/atm/eni.c                     ((tx->tx_pos+size-AAL5_TRAILER) & (tx->words-1))*4);
tx               1172 drivers/atm/eni.c 	ENI_PRV_POS(skb) = tx->tx_pos;
tx               1175 drivers/atm/eni.c 	tx->tx_pos = (tx->tx_pos+size) & (tx->words-1);
tx               1176 drivers/atm/eni.c 	DPRINTK("dma_wr set to %d, tx_pos is now %ld\n",dma_wr,tx->tx_pos);
tx               1186 drivers/atm/eni.c 	struct eni_tx *tx;
tx               1193 drivers/atm/eni.c 		tx = &ENI_DEV(dev)->tx[i];
tx               1194 drivers/atm/eni.c 		if (tx->send)
tx               1195 drivers/atm/eni.c 			while ((skb = skb_dequeue(&tx->backlog))) {
tx               1199 drivers/atm/eni.c 				skb_queue_head(&tx->backlog,skb);
tx               1213 drivers/atm/eni.c 	struct eni_tx *tx;
tx               1221 drivers/atm/eni.c 		tx = ENI_VCC(vcc)->tx;
tx               1222 drivers/atm/eni.c 		NULLCHECK(ENI_VCC(vcc)->tx);
tx               1224 drivers/atm/eni.c 		    (unsigned) eni_in(MID_TX_DESCRSTART(tx->index)));
tx               1225 drivers/atm/eni.c 		if (ENI_VCC(vcc)->txing < tx->words && ENI_PRV_POS(skb) ==
tx               1226 drivers/atm/eni.c 		    eni_in(MID_TX_DESCRSTART(tx->index))) {
tx               1235 drivers/atm/eni.c 		atomic_inc(&vcc->stats->tx);
tx               1247 drivers/atm/eni.c 		if (!eni_dev->tx[i].send) return eni_dev->tx+i;
tx               1294 drivers/atm/eni.c 	struct eni_tx *tx;
tx               1313 drivers/atm/eni.c 			eni_vcc->tx = eni_dev->ubr;
tx               1319 drivers/atm/eni.c 	new_tx = !eni_vcc->tx;
tx               1321 drivers/atm/eni.c 	if (!new_tx) tx = eni_vcc->tx;
tx               1325 drivers/atm/eni.c 		tx = alloc_tx(eni_dev,unlimited);
tx               1326 drivers/atm/eni.c 		if (!tx) {
tx               1330 drivers/atm/eni.c 		DPRINTK("got chan %d\n",tx->index);
tx               1331 drivers/atm/eni.c 		tx->reserved = tx->shaping = 0;
tx               1332 drivers/atm/eni.c 		tx->send = mem;
tx               1333 drivers/atm/eni.c 		tx->words = size >> 2;
tx               1334 drivers/atm/eni.c 		skb_queue_head_init(&tx->backlog);
tx               1337 drivers/atm/eni.c 		    ((tx->send-eni_dev->ram) >> (MID_LOC_SKIP+2)),
tx               1338 drivers/atm/eni.c 		    MID_TX_PLACE(tx->index));
tx               1339 drivers/atm/eni.c 		tx->tx_pos = eni_in(MID_TX_DESCRSTART(tx->index)) &
tx               1342 drivers/atm/eni.c 	error = comp_tx(eni_dev,&rate,tx->reserved,&pre,&res,unlimited);
tx               1346 drivers/atm/eni.c 	if (!error && !ubr && rate > eni_dev->tx_bw+tx->reserved)
tx               1348 drivers/atm/eni.c 	if (!error && set_rsv && !set_shp && rate < tx->shaping)
tx               1350 drivers/atm/eni.c 	if (!error && !set_rsv && rate > tx->reserved && !ubr)
tx               1354 drivers/atm/eni.c 			tx->send = NULL;
tx               1361 drivers/atm/eni.c 		eni_dev->tx_bw += tx->reserved;
tx               1362 drivers/atm/eni.c 		tx->reserved = rate;
tx               1366 drivers/atm/eni.c 		if (unlimited && new_tx) eni_dev->ubr = tx;
tx               1367 drivers/atm/eni.c 		tx->prescaler = pre;
tx               1368 drivers/atm/eni.c 		tx->resolution = res;
tx               1369 drivers/atm/eni.c 		tx->shaping = rate;
tx               1371 drivers/atm/eni.c 	if (set_shp) eni_vcc->tx = tx;
tx               1372 drivers/atm/eni.c 	DPRINTK("rsv %d shp %d\n",tx->reserved,tx->shaping);
tx               1379 drivers/atm/eni.c 	ENI_VCC(vcc)->tx = NULL;
tx               1399 drivers/atm/eni.c 	if (!eni_vcc->tx) return;
tx               1409 drivers/atm/eni.c 		txing = skb_peek(&eni_vcc->tx->backlog) || eni_vcc->txing;
tx               1418 drivers/atm/eni.c 	if (eni_vcc->tx != eni_dev->ubr) {
tx               1424 drivers/atm/eni.c 		while (eni_in(MID_TX_RDPTR(eni_vcc->tx->index)) !=
tx               1425 drivers/atm/eni.c 		    eni_in(MID_TX_DESCRSTART(eni_vcc->tx->index)))
tx               1427 drivers/atm/eni.c 		eni_free_mem(eni_dev,eni_vcc->tx->send,eni_vcc->tx->words << 2);
tx               1428 drivers/atm/eni.c 		eni_vcc->tx->send = NULL;
tx               1429 drivers/atm/eni.c 		eni_dev->tx_bw += eni_vcc->tx->reserved;
tx               1431 drivers/atm/eni.c 	eni_vcc->tx = NULL;
tx               1449 drivers/atm/eni.c 		eni_dev->tx[i].send = NULL;
tx               1450 drivers/atm/eni.c 		eni_dev->tx[i].index = i;
tx               1933 drivers/atm/eni.c 		eni_vcc->tx = NULL; /* for eni_close after open_rx */
tx               1961 drivers/atm/eni.c 	struct eni_tx *tx = ENI_VCC(vcc)->tx;
tx               1966 drivers/atm/eni.c 	if (tx == eni_dev->ubr) return -EBADFD;
tx               1970 drivers/atm/eni.c 	if ((flgs & ATM_MF_DEC_RSV) && rate && rate < tx->reserved) rsv = 1;
tx               1971 drivers/atm/eni.c 	if ((flgs & ATM_MF_INC_RSV) && (!rate || rate > tx->reserved)) rsv = 1;
tx               1972 drivers/atm/eni.c 	if ((flgs & ATM_MF_DEC_SHP) && rate && rate < tx->shaping) shp = 1;
tx               1973 drivers/atm/eni.c 	if ((flgs & ATM_MF_INC_SHP) && (!rate || rate > tx->shaping)) shp = 1;
tx               1987 drivers/atm/eni.c 		dsc = tx->send+ENI_PRV_POS(skb)*4;
tx               1989 drivers/atm/eni.c 		    (tx->prescaler << MID_SEG_PR_SHIFT) |
tx               1990 drivers/atm/eni.c 		    (tx->resolution << MID_SEG_RATE_SHIFT), dsc);
tx               2015 drivers/atm/eni.c 		if ((mult.tx && mult.tx <= 100) || (mult.rx &&mult.rx <= 100) ||
tx               2016 drivers/atm/eni.c 		    mult.tx > 65536 || mult.rx > 65536)
tx               2018 drivers/atm/eni.c 		if (mult.tx) eni_dev->tx_mult = mult.tx;
tx               2056 drivers/atm/eni.c 	if (!ENI_VCC(vcc)->tx) {
tx               2080 drivers/atm/eni.c 	skb_queue_tail(&ENI_VCC(vcc)->tx->backlog,skb);
tx               2160 drivers/atm/eni.c 		struct eni_tx *tx = eni_dev->tx+i;
tx               2162 drivers/atm/eni.c 		if (!tx->send) continue;
tx               2166 drivers/atm/eni.c 			    (unsigned long) (tx->send - eni_dev->ram),
tx               2167 drivers/atm/eni.c 			    tx->send-eni_dev->ram+tx->words*4-1,tx->words*4,
tx               2168 drivers/atm/eni.c 			    tx->reserved,tx->shaping,
tx               2169 drivers/atm/eni.c 			    tx == eni_dev->ubr ? " (UBR)" : "");
tx               2173 drivers/atm/eni.c 		    skb_queue_len(&tx->backlog));
tx               2195 drivers/atm/eni.c 				if (eni_vcc->tx) length += sprintf(page+length,", ");
tx               2197 drivers/atm/eni.c 			if (eni_vcc->tx)
tx               2199 drivers/atm/eni.c 				    eni_vcc->tx->index,eni_vcc->txing);
tx                 59 drivers/atm/eni.h 	struct eni_tx *tx;		/* TXer, NULL if none */
tx                 85 drivers/atm/eni.h 	struct eni_tx tx[NR_CHAN];	/* TX channels */
tx                741 drivers/atm/firestream.c 			atomic_inc(&ATM_SKB(skb)->vcc->stats->tx);
tx                848 drivers/atm/fore200e.c 		    atomic_inc(&vcc->stats->tx);
tx               2605 drivers/atm/he.c 	atomic_inc(&vcc->stats->tx);
tx               1159 drivers/atm/horizon.c 	atomic_inc(&ATM_SKB(skb)->vcc->stats->tx);
tx                814 drivers/atm/idt77252.c 		atomic_inc(&vcc->stats->tx);
tx               2948 drivers/atm/iphase.c                 atomic_inc(&vcc->stats->tx);
tx               3053 drivers/atm/iphase.c         atomic_inc(&vcc->stats->tx);
tx               3060 drivers/atm/iphase.c         if (atomic_read(&vcc->stats->tx) % 20 == 0) {
tx                241 drivers/atm/lanai.c 	} tx;
tx                719 drivers/atm/lanai.c 	dma_addr_t dmaaddr = lvcc->tx.buf.dmaaddr;
tx                727 drivers/atm/lanai.c 		(lvcc->tx.atmvcc->qos.txtp.traffic_class == ATM_CBR) ?
tx                732 drivers/atm/lanai.c 	    TXADDR1_SET_SIZE(lanai_buf_size_cardorder(&lvcc->tx.buf)),
tx                773 drivers/atm/lanai.c 	while ((skb = skb_dequeue(&lvcc->tx.backlog)) != NULL)
tx                774 drivers/atm/lanai.c 		lanai_free_skb(lvcc->tx.atmvcc, skb);
tx                784 drivers/atm/lanai.c 	    (((lanai_buf_size(&lvcc->tx.buf) / 1024) * HZ) >> 7);
tx                789 drivers/atm/lanai.c 		    (lvcc->tx.atmvcc->qos.txtp.traffic_class != ATM_CBR ||
tx                800 drivers/atm/lanai.c 			    lvcc->tx.atmvcc->dev->number, lvcc->vci);
tx               1138 drivers/atm/lanai.c 	r -= ((unsigned long) lvcc->tx.buf.ptr) -
tx               1139 drivers/atm/lanai.c 	    ((unsigned long) lvcc->tx.buf.start);
tx               1142 drivers/atm/lanai.c 		r += lanai_buf_size(&lvcc->tx.buf);
tx               1149 drivers/atm/lanai.c 	return !skb_queue_empty(&lvcc->tx.backlog);
tx               1163 drivers/atm/lanai.c 	APRINTK((((unsigned long) lvcc->tx.buf.ptr) & 15) == 0,
tx               1164 drivers/atm/lanai.c 	    "vcc_tx_add_aal5_descriptor: bad ptr=%p\n", lvcc->tx.buf.ptr);
tx               1165 drivers/atm/lanai.c 	lvcc->tx.buf.ptr += 4;	/* Hope the values REALLY don't matter */
tx               1166 drivers/atm/lanai.c 	pos = ((unsigned char *) lvcc->tx.buf.ptr) -
tx               1167 drivers/atm/lanai.c 	    (unsigned char *) lvcc->tx.buf.start;
tx               1171 drivers/atm/lanai.c 	    lvcc->tx.buf.start, lvcc->tx.buf.ptr, lvcc->tx.buf.end);
tx               1172 drivers/atm/lanai.c 	pos = (pos + len) & (lanai_buf_size(&lvcc->tx.buf) - 1);
tx               1176 drivers/atm/lanai.c 	    lvcc->tx.buf.start, lvcc->tx.buf.ptr, lvcc->tx.buf.end);
tx               1177 drivers/atm/lanai.c 	lvcc->tx.buf.ptr[-1] =
tx               1179 drivers/atm/lanai.c 	    ((lvcc->tx.atmvcc->atm_options & ATM_ATMOPT_CLP) ?
tx               1181 drivers/atm/lanai.c 	if (lvcc->tx.buf.ptr >= lvcc->tx.buf.end)
tx               1182 drivers/atm/lanai.c 		lvcc->tx.buf.ptr = lvcc->tx.buf.start;
tx               1189 drivers/atm/lanai.c 	APRINTK((((unsigned long) lvcc->tx.buf.ptr) & 15) == 8,
tx               1190 drivers/atm/lanai.c 	    "vcc_tx_add_aal5_trailer: bad ptr=%p\n", lvcc->tx.buf.ptr);
tx               1191 drivers/atm/lanai.c 	lvcc->tx.buf.ptr += 2;
tx               1192 drivers/atm/lanai.c 	lvcc->tx.buf.ptr[-2] = cpu_to_be32((uu << 24) | (cpi << 16) | len);
tx               1193 drivers/atm/lanai.c 	if (lvcc->tx.buf.ptr >= lvcc->tx.buf.end)
tx               1194 drivers/atm/lanai.c 		lvcc->tx.buf.ptr = lvcc->tx.buf.start;
tx               1202 drivers/atm/lanai.c 	e = ((unsigned char *) lvcc->tx.buf.ptr) + n;
tx               1203 drivers/atm/lanai.c 	m = e - (unsigned char *) lvcc->tx.buf.end;
tx               1206 drivers/atm/lanai.c 	memcpy(lvcc->tx.buf.ptr, src, n - m);
tx               1208 drivers/atm/lanai.c 		memcpy(lvcc->tx.buf.start, src + n - m, m);
tx               1209 drivers/atm/lanai.c 		e = ((unsigned char *) lvcc->tx.buf.start) + m;
tx               1211 drivers/atm/lanai.c 	lvcc->tx.buf.ptr = (u32 *) e;
tx               1220 drivers/atm/lanai.c 	e = ((unsigned char *) lvcc->tx.buf.ptr) + n;
tx               1221 drivers/atm/lanai.c 	m = e - (unsigned char *) lvcc->tx.buf.end;
tx               1224 drivers/atm/lanai.c 	memset(lvcc->tx.buf.ptr, 0, n - m);
tx               1226 drivers/atm/lanai.c 		memset(lvcc->tx.buf.start, 0, m);
tx               1227 drivers/atm/lanai.c 		e = ((unsigned char *) lvcc->tx.buf.start) + m;
tx               1229 drivers/atm/lanai.c 	lvcc->tx.buf.ptr = (u32 *) e;
tx               1236 drivers/atm/lanai.c 	int i, ptr = ((unsigned char *) lvcc->tx.buf.ptr) -
tx               1237 drivers/atm/lanai.c 	    (unsigned char *) lvcc->tx.buf.start;
tx               1240 drivers/atm/lanai.c 	    ptr, lvcc->vci, lvcc->tx.buf.start, lvcc->tx.buf.ptr,
tx               1241 drivers/atm/lanai.c 	    lvcc->tx.buf.end);
tx               1293 drivers/atm/lanai.c 	lanai_free_skb(lvcc->tx.atmvcc, skb);
tx               1294 drivers/atm/lanai.c 	atomic_inc(&lvcc->tx.atmvcc->stats->tx);
tx               1308 drivers/atm/lanai.c 		skb = skb_dequeue(&lvcc->tx.backlog);
tx               1314 drivers/atm/lanai.c 			skb_queue_head(&lvcc->tx.backlog, skb);
tx               1340 drivers/atm/lanai.c 		skb_queue_tail(&lvcc->tx.backlog, skb);
tx               1358 drivers/atm/lanai.c 	lanai_free_skb(lvcc->tx.atmvcc, skb);
tx               1470 drivers/atm/lanai.c 		skb_queue_head_init(&lvcc->tx.backlog);
tx               1512 drivers/atm/lanai.c 		lvcc->tx.unqueue = vcc_tx_unqueue_aal0;
tx               1516 drivers/atm/lanai.c 		lvcc->tx.unqueue = vcc_tx_unqueue_aal5;
tx               1520 drivers/atm/lanai.c 	return lanai_get_sized_buffer(lanai, &lvcc->tx.buf, max_sdu,
tx               1634 drivers/atm/lanai.c 		if (unlikely(lvcc->tx.atmvcc == NULL)) {
tx               1642 drivers/atm/lanai.c 		lvcc->tx.endptr = SERVICE_GET_END(s);
tx               1703 drivers/atm/lanai.c 		lvcc->tx.unqueue(lanai, lvcc, lvcc->tx.endptr);
tx               1750 drivers/atm/lanai.c 	if (lvcc == NULL || lvcc->tx.atmvcc == NULL ||
tx               1756 drivers/atm/lanai.c 	lvcc->tx.unqueue(lanai, lvcc, endptr);
tx               1999 drivers/atm/lanai.c 		    lvcc->tx.atmvcc != NULL && lvcc->tx.atmvcc != atmvcc)
tx               2288 drivers/atm/lanai.c 	if (lvcc->tx.atmvcc == atmvcc) {
tx               2295 drivers/atm/lanai.c 		lanai_buf_deallocate(&lvcc->tx.buf, lanai->pci);
tx               2296 drivers/atm/lanai.c 		lvcc->tx.atmvcc = NULL;
tx               2356 drivers/atm/lanai.c 		APRINTK(lvcc->tx.atmvcc == NULL, "tx.atmvcc!=NULL, vci=%d\n",
tx               2361 drivers/atm/lanai.c 		lvcc->tx.atmvcc = atmvcc;
tx               2376 drivers/atm/lanai.c 	if (atmvcc == lvcc->tx.atmvcc) {
tx               2395 drivers/atm/lanai.c 	      lvcc->tx.atmvcc != atmvcc))
tx               2519 drivers/atm/lanai.c 	if (lvcc->tx.atmvcc != NULL)
tx               2522 drivers/atm/lanai.c 		    lvcc->tx.atmvcc->qos.aal == ATM_AAL5 ? 5 : 0,
tx               2523 drivers/atm/lanai.c 		    lanai_buf_size(&lvcc->tx.buf),
tx               2524 drivers/atm/lanai.c 		    lvcc->tx.atmvcc == lanai->cbrvcc ? 'C' : 'U',
tx               1257 drivers/atm/nicstar.c 	if (vcc->qos.txtp.traffic_class != ATM_NONE && vc->tx)
tx               1379 drivers/atm/nicstar.c 			vc->tx = 1;
tx               1461 drivers/atm/nicstar.c 		vc->tx = 0;
tx               1642 drivers/atm/nicstar.c 	if (!vc->tx) {
tx               1712 drivers/atm/nicstar.c 	atomic_inc(&vcc->stats->tx);
tx                698 drivers/atm/nicstar.h 	volatile unsigned int tx:1;	/* TX vc? */
tx               1121 drivers/atm/solos-pci.c 				atomic_inc(&vcc->stats->tx);
tx                739 drivers/atm/zatm.c 	atomic_inc(&vcc->stats->tx);
tx                206 drivers/block/aoe/aoenet.c 	kts.fn = tx;
tx                390 drivers/bus/moxtet.c 		.tx_buf = moxtet->tx,
tx                433 drivers/bus/moxtet.c 	moxtet->tx[moxtet->count - mdev->idx] = val;
tx                435 drivers/bus/moxtet.c 	ret = spi_write(to_spi_device(moxtet->dev), moxtet->tx,
tx                452 drivers/bus/moxtet.c 	return moxtet->tx[moxtet->count - mdev->idx];
tx                502 drivers/bus/moxtet.c 		p = hex_byte_pack(p, moxtet->tx[moxtet->count - i]);
tx                538 drivers/bus/moxtet.c 		moxtet->tx[moxtet->count - i] = bin[i];
tx                540 drivers/bus/moxtet.c 	err = spi_write(to_spi_device(moxtet->dev), moxtet->tx,
tx               2207 drivers/char/pcmcia/synclink_cs.c 	icount->tx = cnow.tx;
tx               2583 drivers/char/pcmcia/synclink_cs.c 			      info->icount.tx, info->icount.rx);
tx                 31 drivers/clk/mediatek/clk-apmixed.c 	struct mtk_ref2usb_tx *tx = to_mtk_ref2usb_tx(hw);
tx                 33 drivers/clk/mediatek/clk-apmixed.c 	return (readl(tx->base_addr) & REF2USB_EN_MASK) == REF2USB_EN_MASK;
tx                 38 drivers/clk/mediatek/clk-apmixed.c 	struct mtk_ref2usb_tx *tx = to_mtk_ref2usb_tx(hw);
tx                 41 drivers/clk/mediatek/clk-apmixed.c 	val = readl(tx->base_addr);
tx                 44 drivers/clk/mediatek/clk-apmixed.c 	writel(val, tx->base_addr);
tx                 48 drivers/clk/mediatek/clk-apmixed.c 	writel(val, tx->base_addr);
tx                 51 drivers/clk/mediatek/clk-apmixed.c 	writel(val, tx->base_addr);
tx                 58 drivers/clk/mediatek/clk-apmixed.c 	struct mtk_ref2usb_tx *tx = to_mtk_ref2usb_tx(hw);
tx                 61 drivers/clk/mediatek/clk-apmixed.c 	val = readl(tx->base_addr);
tx                 63 drivers/clk/mediatek/clk-apmixed.c 	writel(val, tx->base_addr);
tx                 75 drivers/clk/mediatek/clk-apmixed.c 	struct mtk_ref2usb_tx *tx;
tx                 79 drivers/clk/mediatek/clk-apmixed.c 	tx = kzalloc(sizeof(*tx), GFP_KERNEL);
tx                 80 drivers/clk/mediatek/clk-apmixed.c 	if (!tx)
tx                 83 drivers/clk/mediatek/clk-apmixed.c 	tx->base_addr = reg;
tx                 84 drivers/clk/mediatek/clk-apmixed.c 	tx->hw.init = &init;
tx                 91 drivers/clk/mediatek/clk-apmixed.c 	clk = clk_register(NULL, &tx->hw);
tx                 95 drivers/clk/mediatek/clk-apmixed.c 		kfree(tx);
tx                 50 drivers/clk/tegra/clk-bpmp.c 	} tx;
tx                 77 drivers/clk/tegra/clk-bpmp.c 	memcpy(req + 4, clk->tx.data, clk->tx.size);
tx                 81 drivers/clk/tegra/clk-bpmp.c 	msg.tx.data = &request;
tx                 82 drivers/clk/tegra/clk-bpmp.c 	msg.tx.size = sizeof(request);
tx                155 drivers/clk/tegra/clk-bpmp.c 	msg.tx.data = &request;
tx                156 drivers/clk/tegra/clk-bpmp.c 	msg.tx.size = sizeof(request);
tx                182 drivers/clk/tegra/clk-bpmp.c 	msg.tx.data = &request;
tx                183 drivers/clk/tegra/clk-bpmp.c 	msg.tx.size = sizeof(request);
tx                208 drivers/clk/tegra/clk-bpmp.c 	msg.tx.data = &request;
tx                209 drivers/clk/tegra/clk-bpmp.c 	msg.tx.size = sizeof(request);
tx                264 drivers/clk/tegra/clk-bpmp.c 	msg.tx.data = &request;
tx                265 drivers/clk/tegra/clk-bpmp.c 	msg.tx.size = sizeof(request);
tx                128 drivers/cpufreq/tegra186-cpufreq.c 	msg.tx.data = &req;
tx                129 drivers/cpufreq/tegra186-cpufreq.c 	msg.tx.size = sizeof(req);
tx                586 drivers/crypto/omap-sham.c 	struct dma_async_tx_descriptor *tx;
tx                610 drivers/crypto/omap-sham.c 	tx = dmaengine_prep_slave_sg(dd->dma_lch, ctx->sg, ctx->sg_len,
tx                614 drivers/crypto/omap-sham.c 	if (!tx) {
tx                619 drivers/crypto/omap-sham.c 	tx->callback = omap_sham_dma_callback;
tx                620 drivers/crypto/omap-sham.c 	tx->callback_param = dd;
tx                632 drivers/crypto/omap-sham.c 	dmaengine_submit(tx);
tx                200 drivers/dma/altera-msgdma.c #define tx_to_desc(tx)	container_of(tx, struct msgdma_sw_desc, async_tx)
tx                301 drivers/dma/altera-msgdma.c static dma_cookie_t msgdma_tx_submit(struct dma_async_tx_descriptor *tx)
tx                303 drivers/dma/altera-msgdma.c 	struct msgdma_device *mdev = to_mdev(tx->chan);
tx                308 drivers/dma/altera-msgdma.c 	new = tx_to_desc(tx);
tx                310 drivers/dma/altera-msgdma.c 	cookie = dma_cookie_assign(tx);
tx                320 drivers/dma/amba-pl08x.c static inline struct pl08x_txd *to_pl08x_txd(struct dma_async_tx_descriptor *tx)
tx                322 drivers/dma/amba-pl08x.c 	return container_of(tx, struct pl08x_txd, vd.tx);
tx                532 drivers/dma/amba-pl08x.c 	struct pl08x_txd *txd = to_pl08x_txd(&vd->tx);
tx               1510 drivers/dma/amba-pl08x.c 	struct pl08x_txd *txd = to_pl08x_txd(&vd->tx);
tx               1511 drivers/dma/amba-pl08x.c 	struct pl08x_dma_chan *plchan = to_pl08x_chan(vd->tx.chan);
tx               1513 drivers/dma/amba-pl08x.c 	dma_descriptor_unmap(&vd->tx);
tx               1580 drivers/dma/amba-pl08x.c 			struct pl08x_txd *txd = to_pl08x_txd(&vd->tx);
tx               2316 drivers/dma/amba-pl08x.c 			struct pl08x_txd *tx;
tx               2326 drivers/dma/amba-pl08x.c 			tx = plchan->at;
tx               2327 drivers/dma/amba-pl08x.c 			if (tx && tx->cyclic) {
tx               2328 drivers/dma/amba-pl08x.c 				vchan_cyclic_callback(&tx->vd);
tx               2329 drivers/dma/amba-pl08x.c 			} else if (tx) {
tx               2336 drivers/dma/amba-pl08x.c 				tx->done = true;
tx               2337 drivers/dma/amba-pl08x.c 				vchan_cookie_complete(&tx->vd);
tx                 59 drivers/dma/at_hdmac.c static dma_cookie_t atc_tx_submit(struct dma_async_tx_descriptor *tx);
tx                667 drivers/dma/at_hdmac.c static dma_cookie_t atc_tx_submit(struct dma_async_tx_descriptor *tx)
tx                669 drivers/dma/at_hdmac.c 	struct at_desc		*desc = txd_to_at_desc(tx);
tx                670 drivers/dma/at_hdmac.c 	struct at_dma_chan	*atchan = to_at_dma_chan(tx->chan);
tx                675 drivers/dma/at_hdmac.c 	cookie = dma_cookie_assign(tx);
tx                678 drivers/dma/at_hdmac.c 		dev_vdbg(chan2dev(tx->chan), "tx_submit: started %u\n",
tx                683 drivers/dma/at_hdmac.c 		dev_vdbg(chan2dev(tx->chan), "tx_submit: queued %u\n",
tx                420 drivers/dma/at_xdmac.c static dma_cookie_t at_xdmac_tx_submit(struct dma_async_tx_descriptor *tx)
tx                422 drivers/dma/at_xdmac.c 	struct at_xdmac_desc	*desc = txd_to_at_desc(tx);
tx                423 drivers/dma/at_xdmac.c 	struct at_xdmac_chan	*atchan = to_at_xdmac_chan(tx->chan);
tx                428 drivers/dma/at_xdmac.c 	cookie = dma_cookie_assign(tx);
tx                430 drivers/dma/at_xdmac.c 	dev_vdbg(chan2dev(tx->chan), "%s: atchan 0x%p, add desc 0x%p to xfers_list\n",
tx                 93 drivers/dma/bcm-sba-raid.c #define to_sba_request(tx)		\
tx                 94 drivers/dma/bcm-sba-raid.c 	container_of(tx, struct sba_request, tx)
tx                121 drivers/dma/bcm-sba-raid.c 	struct dma_async_tx_descriptor tx;
tx                211 drivers/dma/bcm-sba-raid.c 		if (async_tx_test_ack(&req->tx)) {
tx                235 drivers/dma/bcm-sba-raid.c 	dma_async_tx_descriptor_init(&req->tx, &sba->dma_chan);
tx                236 drivers/dma/bcm-sba-raid.c 	async_tx_ack(&req->tx);
tx                419 drivers/dma/bcm-sba-raid.c 	struct dma_async_tx_descriptor *tx;
tx                424 drivers/dma/bcm-sba-raid.c 		tx = &first->tx;
tx                426 drivers/dma/bcm-sba-raid.c 		WARN_ON(tx->cookie < 0);
tx                427 drivers/dma/bcm-sba-raid.c 		if (tx->cookie > 0) {
tx                429 drivers/dma/bcm-sba-raid.c 			dma_cookie_complete(tx);
tx                431 drivers/dma/bcm-sba-raid.c 			dmaengine_desc_get_callback_invoke(tx, NULL);
tx                432 drivers/dma/bcm-sba-raid.c 			dma_descriptor_unmap(tx);
tx                433 drivers/dma/bcm-sba-raid.c 			tx->callback = NULL;
tx                434 drivers/dma/bcm-sba-raid.c 			tx->callback_result = NULL;
tx                437 drivers/dma/bcm-sba-raid.c 		dma_run_dependencies(tx);
tx                467 drivers/dma/bcm-sba-raid.c 		if (async_tx_test_ack(&req->tx))
tx                523 drivers/dma/bcm-sba-raid.c static dma_cookie_t sba_tx_submit(struct dma_async_tx_descriptor *tx)
tx                530 drivers/dma/bcm-sba-raid.c 	if (unlikely(!tx))
tx                533 drivers/dma/bcm-sba-raid.c 	sba = to_sba_device(tx->chan);
tx                534 drivers/dma/bcm-sba-raid.c 	req = to_sba_request(tx);
tx                538 drivers/dma/bcm-sba-raid.c 	cookie = dma_cookie_assign(tx);
tx                569 drivers/dma/bcm-sba-raid.c 	dma_addr_t resp_dma = req->tx.phys;
tx                643 drivers/dma/bcm-sba-raid.c 	req->tx.flags = flags;
tx                644 drivers/dma/bcm-sba-raid.c 	req->tx.cookie = -EBUSY;
tx                646 drivers/dma/bcm-sba-raid.c 	return &req->tx;
tx                657 drivers/dma/bcm-sba-raid.c 	dma_addr_t resp_dma = req->tx.phys;
tx                729 drivers/dma/bcm-sba-raid.c 	req->tx.flags = flags;
tx                730 drivers/dma/bcm-sba-raid.c 	req->tx.cookie = -EBUSY;
tx                765 drivers/dma/bcm-sba-raid.c 	return (first) ? &first->tx : NULL;
tx                777 drivers/dma/bcm-sba-raid.c 	dma_addr_t resp_dma = req->tx.phys;
tx                868 drivers/dma/bcm-sba-raid.c 	req->tx.flags = flags;
tx                869 drivers/dma/bcm-sba-raid.c 	req->tx.cookie = -EBUSY;
tx                908 drivers/dma/bcm-sba-raid.c 	return (first) ? &first->tx : NULL;
tx                922 drivers/dma/bcm-sba-raid.c 	dma_addr_t resp_dma = req->tx.phys;
tx               1080 drivers/dma/bcm-sba-raid.c 	req->tx.flags = flags;
tx               1081 drivers/dma/bcm-sba-raid.c 	req->tx.cookie = -EBUSY;
tx               1097 drivers/dma/bcm-sba-raid.c 	dma_addr_t resp_dma = req->tx.phys;
tx               1335 drivers/dma/bcm-sba-raid.c 	req->tx.flags = flags;
tx               1336 drivers/dma/bcm-sba-raid.c 	req->tx.cookie = -EBUSY;
tx               1432 drivers/dma/bcm-sba-raid.c 	return (first) ? &first->tx : NULL;
tx               1520 drivers/dma/bcm-sba-raid.c 		dma_async_tx_descriptor_init(&req->tx, &sba->dma_chan);
tx               1521 drivers/dma/bcm-sba-raid.c 		async_tx_ack(&req->tx);
tx               1522 drivers/dma/bcm-sba-raid.c 		req->tx.tx_submit = sba_tx_submit;
tx               1523 drivers/dma/bcm-sba-raid.c 		req->tx.phys = sba->resp_dma_base + i * sba->hw_resp_size;
tx                203 drivers/dma/bcm2835-dma.c 	return container_of(t, struct bcm2835_desc, vd.tx);
tx                452 drivers/dma/bcm2835-dma.c 	c->desc = d = to_bcm2835_dma_desc(&vd->tx);
tx                580 drivers/dma/bcm2835-dma.c 			bcm2835_dma_desc_size(to_bcm2835_dma_desc(&vd->tx));
tx                581 drivers/dma/bcm2835-dma.c 	} else if (c->desc && c->desc->vd.tx.cookie == cookie) {
tx                800 drivers/dma/bcm2835-dma.c 		if (c->desc->vd.tx.flags & DMA_PREP_INTERRUPT)
tx               2195 drivers/dma/coh901318.c coh901318_tx_submit(struct dma_async_tx_descriptor *tx)
tx               2197 drivers/dma/coh901318.c 	struct coh901318_desc *cohd = container_of(tx, struct coh901318_desc,
tx               2199 drivers/dma/coh901318.c 	struct coh901318_chan *cohc = to_coh901318_chan(tx->chan);
tx               2204 drivers/dma/coh901318.c 	cookie = dma_cookie_assign(tx);
tx                253 drivers/dma/dma-axi-dmac.c 	if (chan->hw_cyclic && desc->cyclic && !desc->vdesc.tx.callback &&
tx                257 drivers/dma/dma-jz4780.c 	struct jz4780_dma_chan *jzchan = to_jz4780_dma_chan(vdesc->tx.chan);
tx                495 drivers/dma/dma-jz4780.c 		if (jzchan->desc->type == DMA_CYCLIC && vdesc->tx.callback) {
tx                653 drivers/dma/dma-jz4780.c 	} else if (cookie == jzchan->desc->vdesc.tx.cookie) {
tx               1297 drivers/dma/dmaengine.c void dma_async_tx_descriptor_init(struct dma_async_tx_descriptor *tx,
tx               1300 drivers/dma/dmaengine.c 	tx->chan = chan;
tx               1302 drivers/dma/dmaengine.c 	spin_lock_init(&tx->lock);
tx               1311 drivers/dma/dmaengine.c dma_wait_for_async_tx(struct dma_async_tx_descriptor *tx)
tx               1315 drivers/dma/dmaengine.c 	if (!tx)
tx               1318 drivers/dma/dmaengine.c 	while (tx->cookie == -EBUSY) {
tx               1320 drivers/dma/dmaengine.c 			dev_err(tx->chan->device->dev,
tx               1327 drivers/dma/dmaengine.c 	return dma_sync_wait(tx->chan, tx->cookie);
tx               1335 drivers/dma/dmaengine.c void dma_run_dependencies(struct dma_async_tx_descriptor *tx)
tx               1337 drivers/dma/dmaengine.c 	struct dma_async_tx_descriptor *dep = txd_next(tx);
tx               1345 drivers/dma/dmaengine.c 	txd_clear_next(tx);
tx                 29 drivers/dma/dmaengine.h static inline dma_cookie_t dma_cookie_assign(struct dma_async_tx_descriptor *tx)
tx                 31 drivers/dma/dmaengine.h 	struct dma_chan *chan = tx->chan;
tx                 37 drivers/dma/dmaengine.h 	tx->cookie = chan->cookie = cookie;
tx                 52 drivers/dma/dmaengine.h static inline void dma_cookie_complete(struct dma_async_tx_descriptor *tx)
tx                 54 drivers/dma/dmaengine.h 	BUG_ON(tx->cookie < DMA_MIN_COOKIE);
tx                 55 drivers/dma/dmaengine.h 	tx->chan->completed_cookie = tx->cookie;
tx                 56 drivers/dma/dmaengine.h 	tx->cookie = 0;
tx                106 drivers/dma/dmaengine.h dmaengine_desc_get_callback(struct dma_async_tx_descriptor *tx,
tx                109 drivers/dma/dmaengine.h 	cb->callback = tx->callback;
tx                110 drivers/dma/dmaengine.h 	cb->callback_result = tx->callback_result;
tx                111 drivers/dma/dmaengine.h 	cb->callback_param = tx->callback_param;
tx                152 drivers/dma/dmaengine.h dmaengine_desc_get_callback_invoke(struct dma_async_tx_descriptor *tx,
tx                157 drivers/dma/dmaengine.h 	dmaengine_desc_get_callback(tx, &cb);
tx                667 drivers/dma/dmatest.c 		struct dma_async_tx_descriptor *tx = NULL;
tx                762 drivers/dma/dmatest.c 			tx = dev->device_prep_dma_memcpy(chan,
tx                766 drivers/dma/dmatest.c 			tx = dev->device_prep_dma_memset(chan,
tx                771 drivers/dma/dmatest.c 			tx = dev->device_prep_dma_xor(chan,
tx                778 drivers/dma/dmatest.c 			tx = dev->device_prep_dma_pq(chan, dma_pq, srcs,
tx                783 drivers/dma/dmatest.c 		if (!tx) {
tx                792 drivers/dma/dmatest.c 			tx->callback = dmatest_callback;
tx                793 drivers/dma/dmatest.c 			tx->callback_param = done;
tx                795 drivers/dma/dmatest.c 		cookie = tx->tx_submit(tx);
tx                213 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c 	desc->vd.tx.phys = phys;
tx                228 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c 		dma_pool_free(dw->desc_pool, child, child->vd.tx.phys);
tx                232 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c 	dma_pool_free(dw->desc_pool, desc, desc->vd.tx.phys);
tx                298 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c 	write_chan_llp(chan, first->vd.tx.phys | lms);
tx                321 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c 		vd->tx.cookie);
tx                495 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c 			write_desc_llp(prev, desc->vd.tx.phys | lms);
tx                560 drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c 		axi_chan_name(chan), vd->tx.cookie, status);
tx                 56 drivers/dma/dw/core.c static dma_cookie_t dwc_tx_submit(struct dma_async_tx_descriptor *tx)
tx                 58 drivers/dma/dw/core.c 	struct dw_desc		*desc = txd_to_dw_desc(tx);
tx                 59 drivers/dma/dw/core.c 	struct dw_dma_chan	*dwc = to_dw_dma_chan(tx->chan);
tx                 64 drivers/dma/dw/core.c 	cookie = dma_cookie_assign(tx);
tx                 74 drivers/dma/dw/core.c 	dev_vdbg(chan2dev(tx->chan), "%s: queued %u\n",
tx                829 drivers/dma/ep93xx_dma.c static dma_cookie_t ep93xx_dma_tx_submit(struct dma_async_tx_descriptor *tx)
tx                831 drivers/dma/ep93xx_dma.c 	struct ep93xx_dma_chan *edmac = to_ep93xx_dma_chan(tx->chan);
tx                837 drivers/dma/ep93xx_dma.c 	cookie = dma_cookie_assign(tx);
tx                839 drivers/dma/ep93xx_dma.c 	desc = container_of(tx, struct ep93xx_dma_desc, txd);
tx                325 drivers/dma/fsl-edma-common.c 	if (fsl_chan->edesc && cookie == fsl_chan->edesc->vdesc.tx.cookie)
tx                 83 drivers/dma/fsl_raid.c #define to_fsl_re_dma_desc(tx) container_of(tx, struct fsl_re_desc, async_tx)
tx                 86 drivers/dma/fsl_raid.c static dma_cookie_t fsl_re_tx_submit(struct dma_async_tx_descriptor *tx)
tx                 93 drivers/dma/fsl_raid.c 	desc = to_fsl_re_dma_desc(tx);
tx                 94 drivers/dma/fsl_raid.c 	re_chan = container_of(tx->chan, struct fsl_re_chan, chan);
tx                 97 drivers/dma/fsl_raid.c 	cookie = dma_cookie_assign(tx);
tx                430 drivers/dma/fsl_raid.c 		struct dma_async_tx_descriptor *tx;
tx                438 drivers/dma/fsl_raid.c 		tx = fsl_re_prep_dma_genq(chan, dest[1], dma_src, 2, coef, len,
tx                440 drivers/dma/fsl_raid.c 		if (tx)
tx                441 drivers/dma/fsl_raid.c 			desc = to_fsl_re_dma_desc(tx);
tx                443 drivers/dma/fsl_raid.c 		return tx;
tx                406 drivers/dma/fsldma.c static dma_cookie_t fsl_dma_tx_submit(struct dma_async_tx_descriptor *tx)
tx                408 drivers/dma/fsldma.c 	struct fsldma_chan *chan = to_fsl_chan(tx->chan);
tx                409 drivers/dma/fsldma.c 	struct fsl_desc_sw *desc = tx_to_fsl_desc(tx);
tx                192 drivers/dma/fsldma.h #define tx_to_fsl_desc(tx) container_of(tx, struct fsl_desc_sw, async_tx)
tx                330 drivers/dma/hsu/hsu.c 	if (hsuc->desc && cookie == hsuc->desc->vdesc.tx.cookie) {
tx                224 drivers/dma/idma64.c 	struct idma64_chan *idma64c = to_idma64_chan(vdesc->tx.chan);
tx                380 drivers/dma/idma64.c 	if (idma64c->desc && cookie == idma64c->desc->vdesc.tx.cookie) {
tx                176 drivers/dma/img-mdc-dma.c 	struct virt_dma_desc *vdesc = container_of(t, struct virt_dma_desc, tx);
tx                279 drivers/dma/img-mdc-dma.c 	struct mdc_tx_desc *mdesc = to_mdc_desc(&vd->tx);
tx                537 drivers/dma/img-mdc-dma.c 	mdesc = to_mdc_desc(&vd->tx);
tx                591 drivers/dma/img-mdc-dma.c 		mdesc = to_mdc_desc(&vd->tx);
tx                593 drivers/dma/img-mdc-dma.c 	} else if (mchan->desc && mchan->desc->vd.tx.cookie == cookie) {
tx                744 drivers/dma/imx-dma.c static dma_cookie_t imxdma_tx_submit(struct dma_async_tx_descriptor *tx)
tx                746 drivers/dma/imx-dma.c 	struct imxdma_channel *imxdmac = to_imxdma_chan(tx->chan);
tx                753 drivers/dma/imx-dma.c 	cookie = dma_cookie_assign(tx);
tx                748 drivers/dma/imx-sdma.c 	return container_of(t, struct sdma_desc, vd.tx);
tx                762 drivers/dma/imx-sdma.c 	sdmac->desc = desc = to_sdma_desc(&vd->tx);
tx                817 drivers/dma/imx-sdma.c 		dmaengine_desc_get_callback_invoke(&desc->vd.tx, NULL);
tx               1665 drivers/dma/imx-sdma.c 		desc = to_sdma_desc(&vd->tx);
tx               1671 drivers/dma/imx-sdma.c 	} else if (sdmac->desc && sdmac->desc->vd.tx.cookie == cookie) {
tx                297 drivers/dma/ioat/dma.c static dma_cookie_t ioat_tx_submit_unlock(struct dma_async_tx_descriptor *tx)
tx                300 drivers/dma/ioat/dma.c 	struct dma_chan *c = tx->chan;
tx                304 drivers/dma/ioat/dma.c 	cookie = dma_cookie_assign(tx);
tx                599 drivers/dma/ioat/dma.c 		struct dma_async_tx_descriptor *tx;
tx                609 drivers/dma/ioat/dma.c 		tx = &desc->txd;
tx                610 drivers/dma/ioat/dma.c 		if (tx->cookie) {
tx                611 drivers/dma/ioat/dma.c 			dma_cookie_complete(tx);
tx                612 drivers/dma/ioat/dma.c 			dma_descriptor_unmap(tx);
tx                613 drivers/dma/ioat/dma.c 			dmaengine_desc_get_callback_invoke(tx, NULL);
tx                614 drivers/dma/ioat/dma.c 			tx->callback = NULL;
tx                615 drivers/dma/ioat/dma.c 			tx->callback_result = NULL;
tx                618 drivers/dma/ioat/dma.c 		if (tx->phys == phys_complete)
tx                722 drivers/dma/ioat/dma.c 		struct dma_async_tx_descriptor *tx;
tx                727 drivers/dma/ioat/dma.c 		tx = &desc->txd;
tx                728 drivers/dma/ioat/dma.c 		if (tx->cookie) {
tx                731 drivers/dma/ioat/dma.c 			dma_cookie_complete(tx);
tx                732 drivers/dma/ioat/dma.c 			dma_descriptor_unmap(tx);
tx                734 drivers/dma/ioat/dma.c 			dmaengine_desc_get_callback_invoke(tx, &res);
tx                735 drivers/dma/ioat/dma.c 			tx->callback = NULL;
tx                736 drivers/dma/ioat/dma.c 			tx->callback_result = NULL;
tx                763 drivers/dma/ioat/dma.c 	struct dma_async_tx_descriptor *tx;
tx                832 drivers/dma/ioat/dma.c 	tx = &desc->txd;
tx                833 drivers/dma/ioat/dma.c 	if (tx->cookie) {
tx                834 drivers/dma/ioat/dma.c 		dma_cookie_complete(tx);
tx                835 drivers/dma/ioat/dma.c 		dma_descriptor_unmap(tx);
tx                836 drivers/dma/ioat/dma.c 		dmaengine_desc_get_callback_invoke(tx, &res);
tx                837 drivers/dma/ioat/dma.c 		tx->callback = NULL;
tx                838 drivers/dma/ioat/dma.c 		tx->callback_result = NULL;
tx                218 drivers/dma/ioat/dma.h 		struct dma_async_tx_descriptor *tx, int id)
tx                224 drivers/dma/ioat/dma.h 		(unsigned long long) tx->phys,
tx                225 drivers/dma/ioat/dma.h 		(unsigned long long) hw->next, tx->cookie, tx->flags,
tx                308 drivers/dma/ioat/init.c 	struct dma_async_tx_descriptor *tx;
tx                351 drivers/dma/ioat/init.c 	tx = ioat_dma->dma_dev.device_prep_dma_memcpy(dma_chan, dma_dest,
tx                354 drivers/dma/ioat/init.c 	if (!tx) {
tx                360 drivers/dma/ioat/init.c 	async_tx_ack(tx);
tx                362 drivers/dma/ioat/init.c 	tx->callback = ioat_dma_test_callback;
tx                363 drivers/dma/ioat/init.c 	tx->callback_param = &cmp;
tx                364 drivers/dma/ioat/init.c 	cookie = tx->tx_submit(tx);
tx                793 drivers/dma/ioat/init.c 	struct dma_async_tx_descriptor *tx;
tx                867 drivers/dma/ioat/init.c 	tx = dma->device_prep_dma_xor(dma_chan, dest_dma, dma_srcs,
tx                871 drivers/dma/ioat/init.c 	if (!tx) {
tx                877 drivers/dma/ioat/init.c 	async_tx_ack(tx);
tx                879 drivers/dma/ioat/init.c 	tx->callback = ioat_dma_test_callback;
tx                880 drivers/dma/ioat/init.c 	tx->callback_param = &cmp;
tx                881 drivers/dma/ioat/init.c 	cookie = tx->tx_submit(tx);
tx                936 drivers/dma/ioat/init.c 	tx = dma->device_prep_dma_xor_val(dma_chan, dma_srcs,
tx                939 drivers/dma/ioat/init.c 	if (!tx) {
tx                945 drivers/dma/ioat/init.c 	async_tx_ack(tx);
tx                947 drivers/dma/ioat/init.c 	tx->callback = ioat_dma_test_callback;
tx                948 drivers/dma/ioat/init.c 	tx->callback_param = &cmp;
tx                949 drivers/dma/ioat/init.c 	cookie = tx->tx_submit(tx);
tx                989 drivers/dma/ioat/init.c 	tx = dma->device_prep_dma_xor_val(dma_chan, dma_srcs,
tx                992 drivers/dma/ioat/init.c 	if (!tx) {
tx                998 drivers/dma/ioat/init.c 	async_tx_ack(tx);
tx               1000 drivers/dma/ioat/init.c 	tx->callback = ioat_dma_test_callback;
tx               1001 drivers/dma/ioat/init.c 	tx->callback_param = &cmp;
tx               1002 drivers/dma/ioat/init.c 	cookie = tx->tx_submit(tx);
tx                 31 drivers/dma/iop-adma.c #define tx_to_iop_adma_slot(tx) \
tx                 32 drivers/dma/iop-adma.c 	container_of(tx, struct iop_adma_desc_slot, async_tx)
tx                 55 drivers/dma/iop-adma.c 	struct dma_async_tx_descriptor *tx = &desc->async_tx;
tx                 57 drivers/dma/iop-adma.c 	BUG_ON(tx->cookie < 0);
tx                 58 drivers/dma/iop-adma.c 	if (tx->cookie > 0) {
tx                 59 drivers/dma/iop-adma.c 		cookie = tx->cookie;
tx                 60 drivers/dma/iop-adma.c 		tx->cookie = 0;
tx                 65 drivers/dma/iop-adma.c 		dmaengine_desc_get_callback_invoke(tx, NULL);
tx                 67 drivers/dma/iop-adma.c 		dma_descriptor_unmap(tx);
tx                 73 drivers/dma/iop-adma.c 	dma_run_dependencies(tx);
tx                361 drivers/dma/iop-adma.c iop_adma_tx_submit(struct dma_async_tx_descriptor *tx)
tx                363 drivers/dma/iop-adma.c 	struct iop_adma_desc_slot *sw_desc = tx_to_iop_adma_slot(tx);
tx                364 drivers/dma/iop-adma.c 	struct iop_adma_chan *iop_chan = to_iop_adma_chan(tx->chan);
tx                374 drivers/dma/iop-adma.c 	cookie = dma_cookie_assign(tx);
tx                838 drivers/dma/iop-adma.c 	struct dma_async_tx_descriptor *tx;
tx                870 drivers/dma/iop-adma.c 	tx = iop_adma_prep_dma_memcpy(dma_chan, dest_dma, src_dma,
tx                874 drivers/dma/iop-adma.c 	cookie = iop_adma_tx_submit(tx);
tx                914 drivers/dma/iop-adma.c 	struct dma_async_tx_descriptor *tx;
tx                970 drivers/dma/iop-adma.c 	tx = iop_adma_prep_dma_xor(dma_chan, dest_dma, dma_srcs,
tx                974 drivers/dma/iop-adma.c 	cookie = iop_adma_tx_submit(tx);
tx               1016 drivers/dma/iop-adma.c 	tx = iop_adma_prep_dma_xor_val(dma_chan, dma_srcs,
tx               1021 drivers/dma/iop-adma.c 	cookie = iop_adma_tx_submit(tx);
tx               1045 drivers/dma/iop-adma.c 	tx = iop_adma_prep_dma_xor_val(dma_chan, dma_srcs,
tx               1050 drivers/dma/iop-adma.c 	cookie = iop_adma_tx_submit(tx);
tx               1092 drivers/dma/iop-adma.c 	struct dma_async_tx_descriptor *tx;
tx               1139 drivers/dma/iop-adma.c 	tx = iop_adma_prep_dma_pq(dma_chan, pq_dest, pq_src,
tx               1145 drivers/dma/iop-adma.c 	cookie = iop_adma_tx_submit(tx);
tx               1177 drivers/dma/iop-adma.c 	tx = iop_adma_prep_dma_pq_val(dma_chan, &pq_src[IOP_ADMA_NUM_SRC_TEST],
tx               1182 drivers/dma/iop-adma.c 	cookie = iop_adma_tx_submit(tx);
tx               1209 drivers/dma/iop-adma.c 	tx = iop_adma_prep_dma_pq_val(dma_chan, &pq_src[IOP_ADMA_NUM_SRC_TEST],
tx               1214 drivers/dma/iop-adma.c 	cookie = iop_adma_tx_submit(tx);
tx                819 drivers/dma/ipu/ipu_idmac.c static dma_cookie_t idmac_tx_submit(struct dma_async_tx_descriptor *tx)
tx                821 drivers/dma/ipu/ipu_idmac.c 	struct idmac_tx_desc *desc = to_tx_desc(tx);
tx                822 drivers/dma/ipu/ipu_idmac.c 	struct idmac_channel *ichan = to_idmac_chan(tx->chan);
tx                823 drivers/dma/ipu/ipu_idmac.c 	struct idmac *idmac = to_idmac(tx->chan->device);
tx                833 drivers/dma/ipu/ipu_idmac.c 		dev_err(dev, "Descriptor %p not prepared!\n", tx);
tx                839 drivers/dma/ipu/ipu_idmac.c 	async_tx_clear_ack(tx);
tx                866 drivers/dma/ipu/ipu_idmac.c 	cookie = dma_cookie_assign(tx);
tx                897 drivers/dma/ipu/ipu_idmac.c 		tx->cookie = cookie;
tx                715 drivers/dma/k3dma.c 	struct k3_dma_dev *d = to_k3_dma(vd->tx.chan->device);
tx                372 drivers/dma/mediatek/mtk-cqdma.c 		dma_run_dependencies(&cvd->vd.tx);
tx                428 drivers/dma/mediatek/mtk-cqdma.c 		if (vd->tx.cookie == cookie) {
tx                435 drivers/dma/mediatek/mtk-cqdma.c 		if (vd->tx.cookie == cookie)
tx                492 drivers/dma/mediatek/mtk-cqdma.c 	struct dma_async_tx_descriptor *tx = NULL, *prev_tx = NULL;
tx                529 drivers/dma/mediatek/mtk-cqdma.c 		tx = vchan_tx_prep(to_virt_chan(c), &cvd[i]->vd, flags);
tx                530 drivers/dma/mediatek/mtk-cqdma.c 		tx->next = NULL;
tx                535 drivers/dma/mediatek/mtk-cqdma.c 			prev_tx->next = tx;
tx                545 drivers/dma/mediatek/mtk-cqdma.c 		prev_tx = tx;
tx                548 drivers/dma/mediatek/mtk-cqdma.c 	return &cvd[0]->vd.tx;
tx                584 drivers/dma/mediatek/mtk-hsdma.c 			hvc = to_hsdma_vchan(cb->vd->tx.chan);
tx                667 drivers/dma/mediatek/mtk-hsdma.c 		if (vd->tx.cookie == cookie)
tx                671 drivers/dma/mediatek/mtk-hsdma.c 		if (vd->tx.cookie == cookie)
tx                118 drivers/dma/mediatek/mtk-uart-apdma.c 	return container_of(t, struct mtk_uart_apdma_desc, vd.tx);
tx                134 drivers/dma/mediatek/mtk-uart-apdma.c 	struct dma_chan *chan = vd->tx.chan;
tx                371 drivers/dma/mediatek/mtk-uart-apdma.c 		c->desc = to_mtk_uart_apdma_desc(&vd->tx);
tx                 80 drivers/dma/mic_x100_dma.c 	struct dma_async_tx_descriptor *tx;
tx                 93 drivers/dma/mic_x100_dma.c 		tx = &ch->tx_array[last_tail];
tx                 94 drivers/dma/mic_x100_dma.c 		if (tx->cookie) {
tx                 95 drivers/dma/mic_x100_dma.c 			dma_cookie_complete(tx);
tx                 96 drivers/dma/mic_x100_dma.c 			dmaengine_desc_get_callback_invoke(tx, NULL);
tx                 97 drivers/dma/mic_x100_dma.c 			tx->callback = NULL;
tx                238 drivers/dma/mic_x100_dma.c static dma_cookie_t mic_dma_tx_submit_unlock(struct dma_async_tx_descriptor *tx)
tx                240 drivers/dma/mic_x100_dma.c 	struct mic_dma_chan *mic_ch = to_mic_dma_chan(tx->chan);
tx                243 drivers/dma/mic_x100_dma.c 	dma_cookie_assign(tx);
tx                244 drivers/dma/mic_x100_dma.c 	cookie = tx->cookie;
tx                261 drivers/dma/mic_x100_dma.c 	struct dma_async_tx_descriptor *tx = &ch->tx_array[idx];
tx                263 drivers/dma/mic_x100_dma.c 	dma_async_tx_descriptor_init(tx, &ch->api_ch);
tx                264 drivers/dma/mic_x100_dma.c 	tx->tx_submit = mic_dma_tx_submit_unlock;
tx                265 drivers/dma/mic_x100_dma.c 	return tx;
tx                499 drivers/dma/mic_x100_dma.c 	struct dma_async_tx_descriptor *tx;
tx                503 drivers/dma/mic_x100_dma.c 	tx = mic_dma_prep_memcpy_lock(&ch->api_ch, 0, 0, 0, DMA_PREP_FENCE);
tx                504 drivers/dma/mic_x100_dma.c 	if (!tx) {
tx                509 drivers/dma/mic_x100_dma.c 	cookie = tx->tx_submit(tx);
tx                132 drivers/dma/mmp_pdma.c #define tx_to_mmp_pdma_desc(tx)					\
tx                133 drivers/dma/mmp_pdma.c 	container_of(tx, struct mmp_pdma_desc_sw, async_tx)
tx                341 drivers/dma/mmp_pdma.c static dma_cookie_t mmp_pdma_tx_submit(struct dma_async_tx_descriptor *tx)
tx                343 drivers/dma/mmp_pdma.c 	struct mmp_pdma_chan *chan = to_mmp_pdma_chan(tx->chan);
tx                344 drivers/dma/mmp_pdma.c 	struct mmp_pdma_desc_sw *desc = tx_to_mmp_pdma_desc(tx);
tx                372 drivers/dma/mmp_tdma.c static dma_cookie_t mmp_tdma_tx_submit(struct dma_async_tx_descriptor *tx)
tx                374 drivers/dma/mmp_tdma.c 	struct mmp_tdma_chan *tdmac = to_mmp_tdma_chan(tx->chan);
tx                178 drivers/dma/moxart-dma.c 	return container_of(t, struct moxart_desc, vd.tx);
tx                433 drivers/dma/moxart-dma.c 	ch->desc = to_moxart_dma_desc(&vd->tx);
tx                495 drivers/dma/moxart-dma.c 		d = to_moxart_dma_desc(&vd->tx);
tx                497 drivers/dma/moxart-dma.c 	} else if (ch->desc && ch->desc->vd.tx.cookie == cookie) {
tx                 42 drivers/dma/mv_xor.c #define to_mv_xor_slot(tx)		\
tx                 43 drivers/dma/mv_xor.c 	container_of(tx, struct mv_xor_desc_slot, async_tx)
tx                382 drivers/dma/mv_xor.c mv_xor_tx_submit(struct dma_async_tx_descriptor *tx)
tx                384 drivers/dma/mv_xor.c 	struct mv_xor_desc_slot *sw_desc = to_mv_xor_slot(tx);
tx                385 drivers/dma/mv_xor.c 	struct mv_xor_chan *mv_chan = to_mv_xor_chan(tx->chan);
tx                395 drivers/dma/mv_xor.c 	cookie = dma_cookie_assign(tx);
tx                768 drivers/dma/mv_xor.c 	struct dma_async_tx_descriptor *tx;
tx                823 drivers/dma/mv_xor.c 	tx = mv_xor_prep_dma_memcpy(dma_chan, dest_dma, src_dma,
tx                825 drivers/dma/mv_xor.c 	if (!tx) {
tx                832 drivers/dma/mv_xor.c 	cookie = mv_xor_tx_submit(tx);
tx                841 drivers/dma/mv_xor.c 	async_tx_ack(tx);
tx                879 drivers/dma/mv_xor.c 	struct dma_async_tx_descriptor *tx;
tx                956 drivers/dma/mv_xor.c 	tx = mv_xor_prep_dma_xor(dma_chan, dest_dma, dma_srcs,
tx                958 drivers/dma/mv_xor.c 	if (!tx) {
tx                965 drivers/dma/mv_xor.c 	cookie = mv_xor_tx_submit(tx);
tx                974 drivers/dma/mv_xor.c 	async_tx_ack(tx);
tx                299 drivers/dma/mv_xor_v2.c mv_xor_v2_tx_submit(struct dma_async_tx_descriptor *tx)
tx                304 drivers/dma/mv_xor_v2.c 		container_of(tx, struct mv_xor_v2_sw_desc, async_tx);
tx                306 drivers/dma/mv_xor_v2.c 		container_of(tx->chan, struct mv_xor_v2_device, dmachan);
tx                314 drivers/dma/mv_xor_v2.c 	cookie = dma_cookie_assign(tx);
tx                318 drivers/dma/mxs-dma.c static dma_cookie_t mxs_dma_tx_submit(struct dma_async_tx_descriptor *tx)
tx                320 drivers/dma/mxs-dma.c 	return dma_cookie_assign(tx);
tx                666 drivers/dma/nbpfaxi.c static dma_cookie_t nbpf_tx_submit(struct dma_async_tx_descriptor *tx)
tx                668 drivers/dma/nbpfaxi.c 	struct nbpf_desc *desc = container_of(tx, struct nbpf_desc, async_tx);
tx                674 drivers/dma/nbpfaxi.c 	cookie = dma_cookie_assign(tx);
tx                293 drivers/dma/owl-dma.c static inline struct owl_dma_txd *to_owl_txd(struct dma_async_tx_descriptor *tx)
tx                295 drivers/dma/owl-dma.c 	return container_of(tx, struct owl_dma_txd, vd.tx);
tx                500 drivers/dma/owl-dma.c 	struct owl_dma_txd *txd = to_owl_txd(&vd->tx);
tx                651 drivers/dma/owl-dma.c 	struct owl_dma *od = to_owl_dma(vd->tx.chan->device);
tx                652 drivers/dma/owl-dma.c 	struct owl_dma_txd *txd = to_owl_txd(&vd->tx);
tx                783 drivers/dma/owl-dma.c 		txd = to_owl_txd(&vd->tx);
tx               2025 drivers/dma/pl330.c to_desc(struct dma_async_tx_descriptor *tx)
tx               2027 drivers/dma/pl330.c 	return container_of(tx, struct dma_pl330_desc, txd);
tx               2489 drivers/dma/pl330.c static dma_cookie_t pl330_tx_submit(struct dma_async_tx_descriptor *tx)
tx               2491 drivers/dma/pl330.c 	struct dma_pl330_desc *desc, *last = to_desc(tx);
tx               2492 drivers/dma/pl330.c 	struct dma_pl330_chan *pch = to_pchan(tx->chan);
tx               1227 drivers/dma/ppc4xx/adma.c ppc440spe_adma_tx_submit(struct dma_async_tx_descriptor *tx);
tx               1229 drivers/dma/ppc4xx/adma.c static void ppc440spe_adma_set_dest(struct ppc440spe_adma_desc_slot *tx,
tx               1232 drivers/dma/ppc4xx/adma.c ppc440spe_adma_memcpy_xor_set_src(struct ppc440spe_adma_desc_slot *tx,
tx               1236 drivers/dma/ppc4xx/adma.c ppc440spe_adma_pq_set_dest(struct ppc440spe_adma_desc_slot *tx,
tx               1239 drivers/dma/ppc4xx/adma.c ppc440spe_adma_pq_set_src(struct ppc440spe_adma_desc_slot *tx,
tx               1242 drivers/dma/ppc4xx/adma.c ppc440spe_adma_pq_set_src_mult(struct ppc440spe_adma_desc_slot *tx,
tx               1245 drivers/dma/ppc4xx/adma.c ppc440spe_adma_pqzero_sum_set_dest(struct ppc440spe_adma_desc_slot *tx,
tx               1897 drivers/dma/ppc4xx/adma.c static dma_cookie_t ppc440spe_adma_tx_submit(struct dma_async_tx_descriptor *tx)
tx               1900 drivers/dma/ppc4xx/adma.c 	struct ppc440spe_adma_chan *chan = to_ppc440spe_adma_chan(tx->chan);
tx               1906 drivers/dma/ppc4xx/adma.c 	sw_desc = tx_to_ppc440spe_adma_slot(tx);
tx               1913 drivers/dma/ppc4xx/adma.c 	cookie = dma_cookie_assign(tx);
tx               2768 drivers/dma/ppc4xx/adma.c 	struct dma_async_tx_descriptor *tx;
tx               2776 drivers/dma/ppc4xx/adma.c 	tx = ppc440spe_adma_prep_dma_pqzero_sum(chan, pq, &src[1],
tx               2779 drivers/dma/ppc4xx/adma.c 	return tx;
tx                 22 drivers/dma/ppc4xx/adma.h #define tx_to_ppc440spe_adma_slot(tx) \
tx                 23 drivers/dma/ppc4xx/adma.h 		container_of(tx, struct ppc440spe_adma_desc_slot, async_tx)
tx                136 drivers/dma/pxa_dma.c #define tx_to_pxad_desc(tx)					\
tx                137 drivers/dma/pxa_dma.c 	container_of(tx, struct pxad_desc_sw, async_tx)
tx                624 drivers/dma/pxa_dma.c 			__func__, vd, vd->tx.cookie, vd_completed,
tx                626 drivers/dma/pxa_dma.c 		last_started = vd->tx.cookie;
tx                776 drivers/dma/pxa_dma.c static dma_cookie_t pxad_tx_submit(struct dma_async_tx_descriptor *tx)
tx                778 drivers/dma/pxa_dma.c 	struct virt_dma_chan *vc = to_virt_chan(tx->chan);
tx                781 drivers/dma/pxa_dma.c 		*vd = container_of(tx, struct virt_dma_desc, tx);
tx                785 drivers/dma/pxa_dma.c 	set_updater_desc(to_pxad_sw_desc(vd), tx->flags);
tx                788 drivers/dma/pxa_dma.c 	cookie = dma_cookie_assign(tx);
tx                839 drivers/dma/pxa_dma.c 		"%s(): txd %p[%x]", __func__, vd_first, vd_first->tx.cookie);
tx                852 drivers/dma/pxa_dma.c 	struct dma_async_tx_descriptor *tx;
tx                856 drivers/dma/pxa_dma.c 	tx = vchan_tx_prep(vc, vd, tx_flags);
tx                857 drivers/dma/pxa_dma.c 	tx->tx_submit = pxad_tx_submit;
tx                860 drivers/dma/pxa_dma.c 		vc, vd, vd->tx.cookie,
tx                863 drivers/dma/pxa_dma.c 	return tx;
tx               1103 drivers/dma/pxa_dma.c 			vd, vd->tx.cookie, is_desc_completed(vd));
tx                931 drivers/dma/qcom/bam_dma.c 			if (async_desc->vd.tx.cookie != cookie)
tx               1028 drivers/dma/qcom/bam_dma.c 		dmaengine_desc_get_callback(&async_desc->vd.tx, &cb);
tx                410 drivers/dma/s3c24xx-dma.c struct s3c24xx_txd *to_s3c24xx_txd(struct dma_async_tx_descriptor *tx)
tx                412 drivers/dma/s3c24xx-dma.c 	return container_of(tx, struct s3c24xx_txd, vd.tx);
tx                507 drivers/dma/s3c24xx-dma.c 	struct s3c24xx_txd *txd = to_s3c24xx_txd(&vd->tx);
tx                637 drivers/dma/s3c24xx-dma.c 	struct s3c24xx_txd *txd = to_s3c24xx_txd(&vd->tx);
tx                638 drivers/dma/s3c24xx-dma.c 	struct s3c24xx_dma_chan *s3cchan = to_s3c24xx_dma_chan(vd->tx.chan);
tx                641 drivers/dma/s3c24xx-dma.c 		dma_descriptor_unmap(&vd->tx);
tx                783 drivers/dma/s3c24xx-dma.c 		txd = to_s3c24xx_txd(&vd->tx);
tx                155 drivers/dma/sa11x0-dma.c 		p->num, &txd->vd, txd->vd.tx.cookie, txd->ddar);
tx                450 drivers/dma/sa11x0-dma.c 		if (p->txd_done && p->txd_done->vd.tx.cookie == cookie)
tx                452 drivers/dma/sa11x0-dma.c 		else if (p->txd_load && p->txd_load->vd.tx.cookie == cookie)
tx                460 drivers/dma/sh/rcar-dmac.c static dma_cookie_t rcar_dmac_tx_submit(struct dma_async_tx_descriptor *tx)
tx                462 drivers/dma/sh/rcar-dmac.c 	struct rcar_dmac_chan *chan = to_rcar_dmac_chan(tx->chan);
tx                463 drivers/dma/sh/rcar-dmac.c 	struct rcar_dmac_desc *desc = to_rcar_dmac_desc(tx);
tx                469 drivers/dma/sh/rcar-dmac.c 	cookie = dma_cookie_assign(tx);
tx                472 drivers/dma/sh/rcar-dmac.c 		chan->index, tx->cookie, desc);
tx                 70 drivers/dma/sh/shdma-base.c static dma_cookie_t shdma_tx_submit(struct dma_async_tx_descriptor *tx)
tx                 73 drivers/dma/sh/shdma-base.c 		container_of(tx, struct shdma_desc, async_tx);
tx                 74 drivers/dma/sh/shdma-base.c 	struct shdma_chan *schan = to_shdma_chan(tx->chan);
tx                 75 drivers/dma/sh/shdma-base.c 	dma_async_tx_callback callback = tx->callback;
tx                 83 drivers/dma/sh/shdma-base.c 	cookie = dma_cookie_assign(tx);
tx                 99 drivers/dma/sh/shdma-base.c 			chunk->async_tx.callback_param = tx->callback_param;
tx                108 drivers/dma/sh/shdma-base.c 			tx->cookie, &chunk->async_tx, schan->id);
tx                337 drivers/dma/sh/shdma-base.c 		struct dma_async_tx_descriptor *tx = &desc->async_tx;
tx                339 drivers/dma/sh/shdma-base.c 		BUG_ON(tx->cookie > 0 && tx->cookie != desc->cookie);
tx                353 drivers/dma/sh/shdma-base.c 		if (tx->cookie > 0)
tx                354 drivers/dma/sh/shdma-base.c 			cookie = tx->cookie;
tx                366 drivers/dma/sh/shdma-base.c 		if (desc->mark == DESC_COMPLETED && tx->callback) {
tx                368 drivers/dma/sh/shdma-base.c 			dmaengine_desc_get_callback(tx, &cb);
tx                369 drivers/dma/sh/shdma-base.c 			callback = tx->callback;
tx                371 drivers/dma/sh/shdma-base.c 				tx->cookie, tx, schan->id);
tx                376 drivers/dma/sh/shdma-base.c 		if (tx->cookie > 0 || tx->cookie == -EBUSY) {
tx                378 drivers/dma/sh/shdma-base.c 				BUG_ON(tx->cookie < 0);
tx                381 drivers/dma/sh/shdma-base.c 			head_acked = async_tx_test_ack(tx);
tx                394 drivers/dma/sh/shdma-base.c 			tx, tx->cookie);
tx                884 drivers/dma/sh/shdma-base.c 			struct dma_async_tx_descriptor *tx = &sdesc->async_tx;
tx                887 drivers/dma/sh/shdma-base.c 			dmaengine_desc_get_callback_invoke(tx, NULL);
tx                 57 drivers/dma/sh/shdma.h #define tx_to_sh_desc(tx) container_of(tx, struct sh_desc, async_tx)
tx                568 drivers/dma/sh/usb-dmac.c 	struct usb_dmac_chan *chan = to_usb_dmac_chan(vd->tx.chan);
tx                588 drivers/dma/sh/usb-dmac.c 		desc->done_cookie = desc->vd.tx.cookie;
tx                662 drivers/dma/sprd-dma.c 	} else if (schan->cur_desc && schan->cur_desc->vd.tx.cookie == cookie) {
tx                594 drivers/dma/st_fdma.c 	if (fchan->fdesc && cookie == fchan->fdesc->vdesc.tx.cookie)
tx               1465 drivers/dma/ste_dma40.c static dma_cookie_t d40_tx_submit(struct dma_async_tx_descriptor *tx)
tx               1467 drivers/dma/ste_dma40.c 	struct d40_chan *d40c = container_of(tx->chan,
tx               1470 drivers/dma/ste_dma40.c 	struct d40_desc *d40d = container_of(tx, struct d40_desc, txd);
tx               1475 drivers/dma/ste_dma40.c 	cookie = dma_cookie_assign(tx);
tx               1156 drivers/dma/stm32-dma.c 	if (chan->desc && cookie == chan->desc->vdesc.tx.cookie)
tx                371 drivers/dma/stm32-mdma.c 	struct stm32_mdma_chan *chan = to_stm32_mdma_chan(vdesc->tx.chan);
tx               1319 drivers/dma/stm32-mdma.c 	if (chan->desc && cookie == chan->desc->vdesc.tx.cookie)
tx                221 drivers/dma/sun6i-dma.c to_sun6i_desc(struct dma_async_tx_descriptor *tx)
tx                223 drivers/dma/sun6i-dma.c 	return container_of(tx, struct sun6i_desc, vd.tx);
tx                403 drivers/dma/sun6i-dma.c 	struct sun6i_desc *txd = to_sun6i_desc(&vd->tx);
tx                404 drivers/dma/sun6i-dma.c 	struct sun6i_dma_dev *sdev = to_sun6i_dma_dev(vd->tx.chan->device);
tx                445 drivers/dma/sun6i-dma.c 	pchan->desc = to_sun6i_desc(&desc->tx);
tx                947 drivers/dma/sun6i-dma.c 	txd = to_sun6i_desc(&vd->tx);
tx                269 drivers/dma/tegra20-apb-dma.c static dma_cookie_t tegra_dma_tx_submit(struct dma_async_tx_descriptor *tx);
tx                198 drivers/dma/tegra210-adma.c 	return container_of(td, struct tegra_adma_desc, vd.tx);
tx                362 drivers/dma/tegra210-adma.c 	desc = to_tegra_adma_desc(&vd->tx);
tx                526 drivers/dma/tegra210-adma.c 		desc = to_tegra_adma_desc(&vd->tx);
tx                528 drivers/dma/tegra210-adma.c 	} else if (tdc->desc && tdc->desc->vd.tx.cookie == cookie) {
tx                362 drivers/dma/ti/cppi41.c static dma_cookie_t cppi41_tx_submit(struct dma_async_tx_descriptor *tx)
tx                366 drivers/dma/ti/cppi41.c 	cookie = dma_cookie_assign(tx);
tx                751 drivers/dma/ti/edma.c static inline struct edma_desc *to_edma_desc(struct dma_async_tx_descriptor *tx)
tx                753 drivers/dma/ti/edma.c 	return container_of(tx, struct edma_desc, vdesc.tx);
tx                776 drivers/dma/ti/edma.c 		echan->edesc = to_edma_desc(&vdesc->tx);
tx               1844 drivers/dma/ti/edma.c 	if (echan->edesc && echan->edesc->vdesc.tx.cookie == cookie) {
tx               1851 drivers/dma/ti/edma.c 			txstate->residue = to_edma_desc(&vdesc->tx)->residue;
tx               1862 drivers/dma/ti/edma.c 	    echan->edesc->vdesc.tx.cookie == cookie) {
tx                223 drivers/dma/ti/omap-dma.c 	return container_of(t, struct omap_desc, vd.tx);
tx                228 drivers/dma/ti/omap-dma.c 	struct omap_desc *d = to_omap_dma_desc(&vd->tx);
tx                231 drivers/dma/ti/omap-dma.c 		struct omap_dmadev *od = to_omap_dma_dev(vd->tx.chan->device);
tx                542 drivers/dma/ti/omap-dma.c 	c->desc = d = to_omap_dma_desc(&vd->tx);
tx                826 drivers/dma/ti/omap-dma.c 	if (c->desc && c->desc->vd.tx.cookie == cookie)
tx                848 drivers/dma/ti/omap-dma.c 						to_omap_dma_desc(&vd->tx));
tx                187 drivers/dma/txx9dmac.c static dma_cookie_t txx9dmac_tx_submit(struct dma_async_tx_descriptor *tx);
tx                690 drivers/dma/txx9dmac.c static dma_cookie_t txx9dmac_tx_submit(struct dma_async_tx_descriptor *tx)
tx                692 drivers/dma/txx9dmac.c 	struct txx9dmac_desc *desc = txd_to_txx9dmac_desc(tx);
tx                693 drivers/dma/txx9dmac.c 	struct txx9dmac_chan *dc = to_txx9dmac_chan(tx->chan);
tx                697 drivers/dma/txx9dmac.c 	cookie = dma_cookie_assign(tx);
tx                699 drivers/dma/txx9dmac.c 	dev_vdbg(chan2dev(tx->chan), "tx_submit: queued %u %p\n",
tx                304 drivers/dma/uniphier-mdmac.c 	if (mc->md && mc->md->vd.tx.cookie == cookie) {
tx                 14 drivers/dma/virt-dma.c static struct virt_dma_desc *to_virt_desc(struct dma_async_tx_descriptor *tx)
tx                 16 drivers/dma/virt-dma.c 	return container_of(tx, struct virt_dma_desc, tx);
tx                 19 drivers/dma/virt-dma.c dma_cookie_t vchan_tx_submit(struct dma_async_tx_descriptor *tx)
tx                 21 drivers/dma/virt-dma.c 	struct virt_dma_chan *vc = to_virt_chan(tx->chan);
tx                 22 drivers/dma/virt-dma.c 	struct virt_dma_desc *vd = to_virt_desc(tx);
tx                 27 drivers/dma/virt-dma.c 	cookie = dma_cookie_assign(tx);
tx                 49 drivers/dma/virt-dma.c int vchan_tx_desc_free(struct dma_async_tx_descriptor *tx)
tx                 51 drivers/dma/virt-dma.c 	struct virt_dma_chan *vc = to_virt_chan(tx->chan);
tx                 52 drivers/dma/virt-dma.c 	struct virt_dma_desc *vd = to_virt_desc(tx);
tx                 60 drivers/dma/virt-dma.c 		vc, vd, vd->tx.cookie);
tx                 72 drivers/dma/virt-dma.c 		if (vd->tx.cookie == cookie)
tx                 95 drivers/dma/virt-dma.c 		dmaengine_desc_get_callback(&vd->tx, &cb);
tx                104 drivers/dma/virt-dma.c 		dmaengine_desc_get_callback(&vd->tx, &cb);
tx                117 drivers/dma/virt-dma.c 		if (dmaengine_desc_test_reuse(&vd->tx)) {
tx                 16 drivers/dma/virt-dma.h 	struct dma_async_tx_descriptor tx;
tx                 61 drivers/dma/virt-dma.h 	dma_async_tx_descriptor_init(&vd->tx, &vc->chan);
tx                 62 drivers/dma/virt-dma.h 	vd->tx.flags = tx_flags;
tx                 63 drivers/dma/virt-dma.h 	vd->tx.tx_submit = vchan_tx_submit;
tx                 64 drivers/dma/virt-dma.h 	vd->tx.desc_free = vchan_tx_desc_free;
tx                 73 drivers/dma/virt-dma.h 	return &vd->tx;
tx                 96 drivers/dma/virt-dma.h 	struct virt_dma_chan *vc = to_virt_chan(vd->tx.chan);
tx                 99 drivers/dma/virt-dma.h 	cookie = vd->tx.cookie;
tx                100 drivers/dma/virt-dma.h 	dma_cookie_complete(&vd->tx);
tx                114 drivers/dma/virt-dma.h 	struct virt_dma_chan *vc = to_virt_chan(vd->tx.chan);
tx                116 drivers/dma/virt-dma.h 	if (dmaengine_desc_test_reuse(&vd->tx))
tx                128 drivers/dma/virt-dma.h 	struct virt_dma_chan *vc = to_virt_chan(vd->tx.chan);
tx                142 drivers/dma/virt-dma.h 	struct virt_dma_chan *vc = to_virt_chan(vd->tx.chan);
tx                193 drivers/dma/virt-dma.h 		dmaengine_desc_clear_reuse(&vd->tx);
tx                188 drivers/dma/xgene-dma.c #define to_dma_desc_sw(tx)		\
tx                189 drivers/dma/xgene-dma.c 	container_of(tx, struct xgene_dma_desc_sw, tx)
tx                241 drivers/dma/xgene-dma.c 	struct dma_async_tx_descriptor tx;
tx                454 drivers/dma/xgene-dma.c static dma_cookie_t xgene_dma_tx_submit(struct dma_async_tx_descriptor *tx)
tx                460 drivers/dma/xgene-dma.c 	if (unlikely(!tx))
tx                463 drivers/dma/xgene-dma.c 	chan = to_dma_chan(tx->chan);
tx                464 drivers/dma/xgene-dma.c 	desc = to_dma_desc_sw(tx);
tx                468 drivers/dma/xgene-dma.c 	cookie = dma_cookie_assign(tx);
tx                483 drivers/dma/xgene-dma.c 	dma_pool_free(chan->desc_pool, desc, desc->tx.phys);
tx                499 drivers/dma/xgene-dma.c 	desc->tx.phys = phys;
tx                500 drivers/dma/xgene-dma.c 	desc->tx.tx_submit = xgene_dma_tx_submit;
tx                501 drivers/dma/xgene-dma.c 	dma_async_tx_descriptor_init(&desc->tx, &chan->dma_chan);
tx                521 drivers/dma/xgene-dma.c 		if (async_tx_test_ack(&desc->tx))
tx                537 drivers/dma/xgene-dma.c 	struct dma_async_tx_descriptor *tx = &desc->tx;
tx                546 drivers/dma/xgene-dma.c 	if (tx->cookie == 0)
tx                549 drivers/dma/xgene-dma.c 	dma_cookie_complete(tx);
tx                550 drivers/dma/xgene-dma.c 	dma_descriptor_unmap(tx);
tx                553 drivers/dma/xgene-dma.c 	dmaengine_desc_get_callback_invoke(tx, NULL);
tx                556 drivers/dma/xgene-dma.c 	dma_run_dependencies(tx);
tx                578 drivers/dma/xgene-dma.c 	if (!async_tx_test_ack(&desc->tx)) {
tx                588 drivers/dma/xgene-dma.c 	dma_pool_free(chan->desc_pool, desc, desc->tx.phys);
tx                862 drivers/dma/xgene-dma.c 		new->tx.cookie = 0;
tx                863 drivers/dma/xgene-dma.c 		async_tx_ack(&new->tx);
tx                869 drivers/dma/xgene-dma.c 	new->tx.flags = flags; /* client is in control of this ack */
tx                870 drivers/dma/xgene-dma.c 	new->tx.cookie = -EBUSY;
tx                873 drivers/dma/xgene-dma.c 	return &new->tx;
tx                920 drivers/dma/xgene-dma.c 		new->tx.cookie = 0;
tx                921 drivers/dma/xgene-dma.c 		async_tx_ack(&new->tx);
tx                946 drivers/dma/xgene-dma.c 	new->tx.flags = flags; /* client is in control of this ack */
tx                947 drivers/dma/xgene-dma.c 	new->tx.cookie = -EBUSY;
tx                950 drivers/dma/xgene-dma.c 	return &new->tx;
tx                454 drivers/dma/xilinx/xilinx_dma.c #define to_dma_tx_descriptor(tx) \
tx                455 drivers/dma/xilinx/xilinx_dma.c 	container_of(tx, struct xilinx_dma_tx_descriptor, async_tx)
tx               1588 drivers/dma/xilinx/xilinx_dma.c static dma_cookie_t xilinx_dma_tx_submit(struct dma_async_tx_descriptor *tx)
tx               1590 drivers/dma/xilinx/xilinx_dma.c 	struct xilinx_dma_tx_descriptor *desc = to_dma_tx_descriptor(tx);
tx               1591 drivers/dma/xilinx/xilinx_dma.c 	struct xilinx_dma_chan *chan = to_xilinx_chan(tx->chan);
tx               1613 drivers/dma/xilinx/xilinx_dma.c 	cookie = dma_cookie_assign(tx);
tx                145 drivers/dma/xilinx/zynqmp_dma.c #define tx_to_desc(tx)		container_of(tx, struct zynqmp_dma_desc_sw, \
tx                369 drivers/dma/xilinx/zynqmp_dma.c static dma_cookie_t zynqmp_dma_tx_submit(struct dma_async_tx_descriptor *tx)
tx                371 drivers/dma/xilinx/zynqmp_dma.c 	struct zynqmp_dma_chan *chan = to_chan(tx->chan);
tx                376 drivers/dma/xilinx/zynqmp_dma.c 	new = tx_to_desc(tx);
tx                378 drivers/dma/xilinx/zynqmp_dma.c 	cookie = dma_cookie_assign(tx);
tx                719 drivers/dma/zx_dma.c 	struct zx_dma_dev *d = to_zx_dma(vd->tx.chan->device);
tx                154 drivers/firmware/arm_scmi/base.c 	num_skip = t->tx.buf;
tx                207 drivers/firmware/arm_scmi/base.c 	put_unaligned_le32(id, t->tx.buf);
tx                110 drivers/firmware/arm_scmi/clock.c 	put_unaligned_le32(clk_id, t->tx.buf);
tx                141 drivers/firmware/arm_scmi/clock.c 	clk_desc = t->tx.buf;
tx                207 drivers/firmware/arm_scmi/clock.c 	put_unaligned_le32(clk_id, t->tx.buf);
tx                235 drivers/firmware/arm_scmi/clock.c 	cfg = t->tx.buf;
tx                265 drivers/firmware/arm_scmi/clock.c 	cfg = t->tx.buf;
tx                 94 drivers/firmware/arm_scmi/common.h 	struct scmi_msg tx;
tx                262 drivers/firmware/arm_scmi/driver.c 	iowrite32(sizeof(mem->msg_header) + t->tx.len, &mem->length);
tx                264 drivers/firmware/arm_scmi/driver.c 	if (t->tx.buf)
tx                265 drivers/firmware/arm_scmi/driver.c 		memcpy_toio(mem->msg_payload, t->tx.buf, t->tx.len);
tx                549 drivers/firmware/arm_scmi/driver.c 	xfer->tx.len = tx_size;
tx                707 drivers/firmware/arm_scmi/driver.c 		xfer->tx.buf = xfer->rx.buf;
tx                723 drivers/firmware/arm_scmi/driver.c 				int prot_id, bool tx)
tx                732 drivers/firmware/arm_scmi/driver.c 	const char *desc = tx ? "Tx" : "Rx";
tx                735 drivers/firmware/arm_scmi/driver.c 	idx = tx ? 0 : 1;
tx                736 drivers/firmware/arm_scmi/driver.c 	idr = tx ? &info->tx_idr : &info->rx_idr;
tx                754 drivers/firmware/arm_scmi/driver.c 	cl->tx_prepare = tx ? scmi_tx_prepare : NULL;
tx                756 drivers/firmware/arm_scmi/driver.c 	cl->knows_txdone = tx;
tx                198 drivers/firmware/arm_scmi/perf.c 	put_unaligned_le32(domain, t->tx.buf);
tx                253 drivers/firmware/arm_scmi/perf.c 	dom_info = t->tx.buf;
tx                343 drivers/firmware/arm_scmi/perf.c 	limits = t->tx.buf;
tx                382 drivers/firmware/arm_scmi/perf.c 	put_unaligned_le32(domain, t->tx.buf);
tx                424 drivers/firmware/arm_scmi/perf.c 	lvl = t->tx.buf;
tx                461 drivers/firmware/arm_scmi/perf.c 	put_unaligned_le32(domain, t->tx.buf);
tx                518 drivers/firmware/arm_scmi/perf.c 	info = t->tx.buf;
tx                 99 drivers/firmware/arm_scmi/power.c 	put_unaligned_le32(domain, t->tx.buf);
tx                128 drivers/firmware/arm_scmi/power.c 	st = t->tx.buf;
tx                150 drivers/firmware/arm_scmi/power.c 	put_unaligned_le32(domain, t->tx.buf);
tx                 91 drivers/firmware/arm_scmi/reset.c 	put_unaligned_le32(domain, t->tx.buf);
tx                150 drivers/firmware/arm_scmi/reset.c 	dom = t->tx.buf;
tx                123 drivers/firmware/arm_scmi/sensors.c 		put_unaligned_le32(desc_index, t->tx.buf);
tx                180 drivers/firmware/arm_scmi/sensors.c 	cfg = t->tx.buf;
tx                204 drivers/firmware/arm_scmi/sensors.c 	trip = t->tx.buf;
tx                231 drivers/firmware/arm_scmi/sensors.c 	sensor = t->tx.buf;
tx                116 drivers/firmware/tegra/bpmp-debugfs.c 		.tx = {
tx                151 drivers/firmware/tegra/bpmp-debugfs.c 		.tx = {
tx                173 drivers/firmware/tegra/bpmp-debugfs.c 		.tx = {
tx                 23 drivers/firmware/tegra/bpmp-tegra186.c 	} tx, rx;
tx                126 drivers/firmware/tegra/bpmp-tegra186.c 			     priv->tx.virt + offset, priv->tx.phys + offset,
tx                176 drivers/firmware/tegra/bpmp-tegra186.c 	priv->tx.pool = of_gen_pool_get(bpmp->dev->of_node, "shmem", 0);
tx                177 drivers/firmware/tegra/bpmp-tegra186.c 	if (!priv->tx.pool) {
tx                182 drivers/firmware/tegra/bpmp-tegra186.c 	priv->tx.virt = gen_pool_dma_alloc(priv->tx.pool, 4096, &priv->tx.phys);
tx                183 drivers/firmware/tegra/bpmp-tegra186.c 	if (!priv->tx.virt) {
tx                256 drivers/firmware/tegra/bpmp-tegra186.c 	gen_pool_free(priv->tx.pool, (unsigned long)priv->tx.virt, 4096);
tx                275 drivers/firmware/tegra/bpmp-tegra186.c 	gen_pool_free(priv->tx.pool, (unsigned long)priv->tx.virt, 4096);
tx                 95 drivers/firmware/tegra/bpmp.c 	return (msg->tx.size <= MSG_DATA_MIN_SZ) &&
tx                 97 drivers/firmware/tegra/bpmp.c 	       (msg->tx.size == 0 || msg->tx.data) &&
tx                333 drivers/firmware/tegra/bpmp.c 				       msg->tx.data, msg->tx.size);
tx                367 drivers/firmware/tegra/bpmp.c 	channel = tegra_bpmp_write_threaded(bpmp, msg->mrq, msg->tx.data,
tx                368 drivers/firmware/tegra/bpmp.c 					    msg->tx.size);
tx                508 drivers/firmware/tegra/bpmp.c 		.tx = {
tx                558 drivers/firmware/tegra/bpmp.c 	msg.tx.data = &request;
tx                559 drivers/firmware/tegra/bpmp.c 	msg.tx.size = sizeof(request);
tx                602 drivers/firmware/tegra/bpmp.c 	msg.tx.data = &request;
tx                603 drivers/firmware/tegra/bpmp.c 	msg.tx.size = sizeof(request);
tx                 62 drivers/firmware/tegra/ivc.c 	} tx;
tx                 97 drivers/firmware/tegra/ivc.c 	u32 tx = READ_ONCE(header->tx.count);
tx                110 drivers/firmware/tegra/ivc.c 	if (tx - rx > ivc->num_frames)
tx                113 drivers/firmware/tegra/ivc.c 	return tx == rx;
tx                119 drivers/firmware/tegra/ivc.c 	u32 tx = READ_ONCE(header->tx.count);
tx                126 drivers/firmware/tegra/ivc.c 	return tx - rx >= ivc->num_frames;
tx                132 drivers/firmware/tegra/ivc.c 	u32 tx = READ_ONCE(header->tx.count);
tx                141 drivers/firmware/tegra/ivc.c 	return tx - rx;
tx                146 drivers/firmware/tegra/ivc.c 	WRITE_ONCE(ivc->tx.channel->tx.count,
tx                147 drivers/firmware/tegra/ivc.c 		   READ_ONCE(ivc->tx.channel->tx.count) + 1);
tx                149 drivers/firmware/tegra/ivc.c 	if (ivc->tx.position == ivc->num_frames - 1)
tx                150 drivers/firmware/tegra/ivc.c 		ivc->tx.position = 0;
tx                152 drivers/firmware/tegra/ivc.c 		ivc->tx.position++;
tx                168 drivers/firmware/tegra/ivc.c 	unsigned int offset = offsetof(struct tegra_ivc_header, tx.count);
tx                178 drivers/firmware/tegra/ivc.c 	if (ivc->tx.channel->tx.state != TEGRA_IVC_STATE_ESTABLISHED)
tx                203 drivers/firmware/tegra/ivc.c 	if (ivc->tx.channel->tx.state != TEGRA_IVC_STATE_ESTABLISHED)
tx                206 drivers/firmware/tegra/ivc.c 	if (!tegra_ivc_full(ivc, ivc->tx.channel))
tx                209 drivers/firmware/tegra/ivc.c 	tegra_ivc_invalidate(ivc, ivc->tx.phys + offset);
tx                211 drivers/firmware/tegra/ivc.c 	if (tegra_ivc_full(ivc, ivc->tx.channel))
tx                294 drivers/firmware/tegra/ivc.c 	unsigned int tx = offsetof(struct tegra_ivc_header, tx.count);
tx                321 drivers/firmware/tegra/ivc.c 	tegra_ivc_invalidate(ivc, ivc->rx.phys + tx);
tx                339 drivers/firmware/tegra/ivc.c 	return tegra_ivc_frame_virt(ivc, ivc->tx.channel, ivc->tx.position);
tx                346 drivers/firmware/tegra/ivc.c 	unsigned int tx = offsetof(struct tegra_ivc_header, tx.count);
tx                354 drivers/firmware/tegra/ivc.c 	tegra_ivc_flush_frame(ivc, ivc->tx.phys, ivc->tx.position, 0,
tx                364 drivers/firmware/tegra/ivc.c 	tegra_ivc_flush(ivc, ivc->tx.phys + tx);
tx                377 drivers/firmware/tegra/ivc.c 	tegra_ivc_invalidate(ivc, ivc->tx.phys + rx);
tx                379 drivers/firmware/tegra/ivc.c 	if (tegra_ivc_available(ivc, ivc->tx.channel) == 1)
tx                388 drivers/firmware/tegra/ivc.c 	unsigned int offset = offsetof(struct tegra_ivc_header, tx.count);
tx                390 drivers/firmware/tegra/ivc.c 	ivc->tx.channel->tx.state = TEGRA_IVC_STATE_SYNC;
tx                391 drivers/firmware/tegra/ivc.c 	tegra_ivc_flush(ivc, ivc->tx.phys + offset);
tx                418 drivers/firmware/tegra/ivc.c 	unsigned int offset = offsetof(struct tegra_ivc_header, tx.count);
tx                423 drivers/firmware/tegra/ivc.c 	state = READ_ONCE(ivc->rx.channel->tx.state);
tx                426 drivers/firmware/tegra/ivc.c 		offset = offsetof(struct tegra_ivc_header, tx.count);
tx                439 drivers/firmware/tegra/ivc.c 		ivc->tx.channel->tx.count = 0;
tx                442 drivers/firmware/tegra/ivc.c 		ivc->tx.position = 0;
tx                455 drivers/firmware/tegra/ivc.c 		ivc->tx.channel->tx.state = TEGRA_IVC_STATE_ACK;
tx                456 drivers/firmware/tegra/ivc.c 		tegra_ivc_flush(ivc, ivc->tx.phys + offset);
tx                463 drivers/firmware/tegra/ivc.c 	} else if (ivc->tx.channel->tx.state == TEGRA_IVC_STATE_SYNC &&
tx                465 drivers/firmware/tegra/ivc.c 		offset = offsetof(struct tegra_ivc_header, tx.count);
tx                478 drivers/firmware/tegra/ivc.c 		ivc->tx.channel->tx.count = 0;
tx                481 drivers/firmware/tegra/ivc.c 		ivc->tx.position = 0;
tx                495 drivers/firmware/tegra/ivc.c 		ivc->tx.channel->tx.state = TEGRA_IVC_STATE_ESTABLISHED;
tx                496 drivers/firmware/tegra/ivc.c 		tegra_ivc_flush(ivc, ivc->tx.phys + offset);
tx                503 drivers/firmware/tegra/ivc.c 	} else if (ivc->tx.channel->tx.state == TEGRA_IVC_STATE_ACK) {
tx                504 drivers/firmware/tegra/ivc.c 		offset = offsetof(struct tegra_ivc_header, tx.count);
tx                519 drivers/firmware/tegra/ivc.c 		ivc->tx.channel->tx.state = TEGRA_IVC_STATE_ESTABLISHED;
tx                520 drivers/firmware/tegra/ivc.c 		tegra_ivc_flush(ivc, ivc->tx.phys + offset);
tx                536 drivers/firmware/tegra/ivc.c 	if (ivc->tx.channel->tx.state != TEGRA_IVC_STATE_ESTABLISHED)
tx                561 drivers/firmware/tegra/ivc.c static int tegra_ivc_check_params(unsigned long rx, unsigned long tx,
tx                564 drivers/firmware/tegra/ivc.c 	BUILD_BUG_ON(!IS_ALIGNED(offsetof(struct tegra_ivc_header, tx.count),
tx                590 drivers/firmware/tegra/ivc.c 	if (!IS_ALIGNED(tx, TEGRA_IVC_ALIGN)) {
tx                591 drivers/firmware/tegra/ivc.c 		pr_err("IVC channel start not aligned: %#lx\n", tx);
tx                595 drivers/firmware/tegra/ivc.c 	if (rx < tx) {
tx                596 drivers/firmware/tegra/ivc.c 		if (rx + frame_size * num_frames > tx) {
tx                598 drivers/firmware/tegra/ivc.c 			       rx, frame_size * num_frames, tx);
tx                602 drivers/firmware/tegra/ivc.c 		if (tx + frame_size * num_frames > rx) {
tx                604 drivers/firmware/tegra/ivc.c 			       tx, frame_size * num_frames, rx);
tx                613 drivers/firmware/tegra/ivc.c 		   dma_addr_t rx_phys, void *tx, dma_addr_t tx_phys,
tx                631 drivers/firmware/tegra/ivc.c 	err = tegra_ivc_check_params((unsigned long)rx, (unsigned long)tx,
tx                644 drivers/firmware/tegra/ivc.c 		ivc->tx.phys = dma_map_single(peer, tx, queue_size,
tx                646 drivers/firmware/tegra/ivc.c 		if (dma_mapping_error(peer, ivc->tx.phys)) {
tx                653 drivers/firmware/tegra/ivc.c 		ivc->tx.phys = tx_phys;
tx                657 drivers/firmware/tegra/ivc.c 	ivc->tx.channel = tx;
tx                668 drivers/firmware/tegra/ivc.c 	ivc->tx.position = 0;
tx                683 drivers/firmware/tegra/ivc.c 		dma_unmap_single(ivc->peer, ivc->tx.phys, size,
tx                 69 drivers/fpga/machxo2-spi.c 	struct spi_transfer rx, tx;
tx                 74 drivers/fpga/machxo2-spi.c 	memset(&tx, 0, sizeof(tx));
tx                 75 drivers/fpga/machxo2-spi.c 	tx.tx_buf = cmd;
tx                 76 drivers/fpga/machxo2-spi.c 	tx.len = sizeof(cmd);
tx                 80 drivers/fpga/machxo2-spi.c 	spi_message_add_tail(&tx, &msg);
tx                139 drivers/fpga/machxo2-spi.c 	struct spi_transfer tx[2];
tx                144 drivers/fpga/machxo2-spi.c 	memset(tx, 0, sizeof(tx));
tx                146 drivers/fpga/machxo2-spi.c 	tx[0].tx_buf = &erase;
tx                147 drivers/fpga/machxo2-spi.c 	tx[0].len = sizeof(erase);
tx                148 drivers/fpga/machxo2-spi.c 	spi_message_add_tail(&tx[0], &msg);
tx                158 drivers/fpga/machxo2-spi.c 	tx[1].tx_buf = &refresh;
tx                159 drivers/fpga/machxo2-spi.c 	tx[1].len = sizeof(refresh);
tx                160 drivers/fpga/machxo2-spi.c 	tx[1].delay_usecs = MACHXO2_REFRESH_USEC;
tx                161 drivers/fpga/machxo2-spi.c 	spi_message_add_tail(&tx[1], &msg);
tx                192 drivers/fpga/machxo2-spi.c 	struct spi_transfer tx[3];
tx                207 drivers/fpga/machxo2-spi.c 	memset(tx, 0, sizeof(tx));
tx                209 drivers/fpga/machxo2-spi.c 	tx[0].tx_buf = &enable;
tx                210 drivers/fpga/machxo2-spi.c 	tx[0].len = sizeof(enable);
tx                211 drivers/fpga/machxo2-spi.c 	tx[0].delay_usecs = MACHXO2_LOW_DELAY_USEC;
tx                212 drivers/fpga/machxo2-spi.c 	spi_message_add_tail(&tx[0], &msg);
tx                214 drivers/fpga/machxo2-spi.c 	tx[1].tx_buf = &erase;
tx                215 drivers/fpga/machxo2-spi.c 	tx[1].len = sizeof(erase);
tx                216 drivers/fpga/machxo2-spi.c 	spi_message_add_tail(&tx[1], &msg);
tx                231 drivers/fpga/machxo2-spi.c 	tx[2].tx_buf = &initaddr;
tx                232 drivers/fpga/machxo2-spi.c 	tx[2].len = sizeof(initaddr);
tx                233 drivers/fpga/machxo2-spi.c 	spi_message_add_tail(&tx[2], &msg);
tx                253 drivers/fpga/machxo2-spi.c 	struct spi_transfer tx;
tx                268 drivers/fpga/machxo2-spi.c 		memset(&tx, 0, sizeof(tx));
tx                270 drivers/fpga/machxo2-spi.c 		tx.tx_buf = payload;
tx                271 drivers/fpga/machxo2-spi.c 		tx.len = MACHXO2_BUF_SIZE;
tx                272 drivers/fpga/machxo2-spi.c 		tx.delay_usecs = MACHXO2_HIGH_DELAY_USEC;
tx                273 drivers/fpga/machxo2-spi.c 		spi_message_add_tail(&tx, &msg);
tx                291 drivers/fpga/machxo2-spi.c 	struct spi_transfer tx[2];
tx                297 drivers/fpga/machxo2-spi.c 	memset(tx, 0, sizeof(tx));
tx                299 drivers/fpga/machxo2-spi.c 	tx[0].tx_buf = &progdone;
tx                300 drivers/fpga/machxo2-spi.c 	tx[0].len = sizeof(progdone);
tx                301 drivers/fpga/machxo2-spi.c 	spi_message_add_tail(&tx[0], &msg);
tx                318 drivers/fpga/machxo2-spi.c 		tx[1].tx_buf = &refresh;
tx                319 drivers/fpga/machxo2-spi.c 		tx[1].len = sizeof(refresh);
tx                320 drivers/fpga/machxo2-spi.c 		tx[1].delay_usecs = MACHXO2_REFRESH_USEC;
tx                321 drivers/fpga/machxo2-spi.c 		spi_message_add_tail(&tx[1], &msg);
tx                 78 drivers/gpu/drm/amd/powerplay/hwmgr/smu8_hwmgr.h #define DDI_POWERGATING_ARG(phyID, lanemask, rx, tx, core) \
tx                 82 drivers/gpu/drm/amd/powerplay/hwmgr/smu8_hwmgr.h 		((tx) ? DISPLAYPHY_TX_SELECT : 0) | \
tx                472 drivers/gpu/drm/drm_mipi_dsi.c 		const u8 *tx = msg->tx_buf;
tx                474 drivers/gpu/drm/drm_mipi_dsi.c 		packet->header[1] = (msg->tx_len > 0) ? tx[0] : 0;
tx                475 drivers/gpu/drm/drm_mipi_dsi.c 		packet->header[2] = (msg->tx_len > 1) ? tx[1] : 0;
tx                536 drivers/gpu/drm/drm_mipi_dsi.c 	u8 tx[2] = { value & 0xff, value >> 8 };
tx                540 drivers/gpu/drm/drm_mipi_dsi.c 		.tx_len = sizeof(tx),
tx                541 drivers/gpu/drm/drm_mipi_dsi.c 		.tx_buf = tx,
tx                698 drivers/gpu/drm/drm_mipi_dsi.c 	u8 *tx;
tx                703 drivers/gpu/drm/drm_mipi_dsi.c 		tx = kmalloc(size, GFP_KERNEL);
tx                704 drivers/gpu/drm/drm_mipi_dsi.c 		if (!tx)
tx                708 drivers/gpu/drm/drm_mipi_dsi.c 		tx[0] = cmd;
tx                709 drivers/gpu/drm/drm_mipi_dsi.c 		memcpy(&tx[1], data, len);
tx                711 drivers/gpu/drm/drm_mipi_dsi.c 		tx = &cmd;
tx                715 drivers/gpu/drm/drm_mipi_dsi.c 	err = mipi_dsi_dcs_write_buffer(dsi, tx, size);
tx                718 drivers/gpu/drm/drm_mipi_dsi.c 		kfree(tx);
tx                 57 drivers/gpu/drm/i915/display/intel_lvds.c 	int tx;
tx                168 drivers/gpu/drm/i915/display/intel_lvds.c 	pps->tx = REG_FIELD_GET(PANEL_LIGHT_OFF_DELAY_MASK, val);
tx                184 drivers/gpu/drm/i915/display/intel_lvds.c 	    pps->t1_t2 == 0 && pps->t5 == 0 && pps->t3 == 0 && pps->tx == 0) {
tx                192 drivers/gpu/drm/i915/display/intel_lvds.c 		pps->tx = 200 * 10;
tx                197 drivers/gpu/drm/i915/display/intel_lvds.c 			 pps->t1_t2, pps->t3, pps->t4, pps->t5, pps->tx,
tx                219 drivers/gpu/drm/i915/display/intel_lvds.c 		   REG_FIELD_PREP(PANEL_LIGHT_OFF_DELAY_MASK, pps->tx));
tx                179 drivers/gpu/drm/mcde/mcde_dsi.c 	const u8 *tx = msg->tx_buf;
tx                226 drivers/gpu/drm/mcde/mcde_dsi.c 			val |= tx[i] << (i & 3) * 8;
tx                232 drivers/gpu/drm/mcde/mcde_dsi.c 			val |= tx[i + 4] << (i & 3) * 8;
tx                238 drivers/gpu/drm/mcde/mcde_dsi.c 			val |= tx[i + 8] << (i & 3) * 8;
tx                244 drivers/gpu/drm/mcde/mcde_dsi.c 			val |= tx[i + 12] << (i & 3) * 8;
tx               2071 drivers/gpu/drm/msm/dsi/dsi_host.c 		u8 tx[2] = {pkt_size & 0xff, pkt_size >> 8};
tx               2076 drivers/gpu/drm/msm/dsi/dsi_host.c 			.tx_buf = tx,
tx                133 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c 	u32 hi, lo, rq, tx;
tx                138 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c 	i2c->func->aux_stat(i2c, &hi, &lo, &rq, &tx);
tx                139 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c 	if (!hi && !lo && !rq && !tx)
tx                147 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c 		if (tx & aux->intr) mask |= NVKM_I2C_DONE;
tx                 28 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/g94.c g94_aux_stat(struct nvkm_i2c *i2c, u32 *hi, u32 *lo, u32 *rq, u32 *tx)
tx                 33 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/g94.c 	for (i = 0, *hi = *lo = *rq = *tx = 0; i < 8; i++) {
tx                 37 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/g94.c 		if ((stat & (8 << (i * 4)))) *tx |= 1 << i;
tx                 28 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/gk104.c gk104_aux_stat(struct nvkm_i2c *i2c, u32 *hi, u32 *lo, u32 *rq, u32 *tx)
tx                 33 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/gk104.c 	for (i = 0, *hi = *lo = *rq = *tx = 0; i < 8; i++) {
tx                 37 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/gk104.c 		if ((stat & (8 << (i * 4)))) *tx |= 1 << i;
tx                 86 drivers/gpu/drm/omapdrm/omap_dmm_tiler.c 	struct dma_async_tx_descriptor *tx;
tx                 90 drivers/gpu/drm/omapdrm/omap_dmm_tiler.c 	tx = dma_dev->device_prep_dma_memcpy(dmm->wa_dma_chan, dst, src, 4, 0);
tx                 91 drivers/gpu/drm/omapdrm/omap_dmm_tiler.c 	if (!tx) {
tx                 96 drivers/gpu/drm/omapdrm/omap_dmm_tiler.c 	cookie = tx->tx_submit(tx);
tx                101 drivers/gpu/drm/tiny/repaper.c 				const void *tx, void *rx, size_t len)
tx                117 drivers/gpu/drm/tiny/repaper.c 	if (tx && len <= 32) {
tx                118 drivers/gpu/drm/tiny/repaper.c 		txbuf = kmemdup(tx, len, GFP_KERNEL);
tx                133 drivers/gpu/drm/tiny/repaper.c 	tr[1].tx_buf = txbuf ? txbuf : tx;
tx                551 drivers/gpu/drm/vc4/vc4_dsi.c 	struct dma_async_tx_descriptor *tx;
tx                563 drivers/gpu/drm/vc4/vc4_dsi.c 	tx = chan->device->device_prep_dma_memcpy(chan,
tx                567 drivers/gpu/drm/vc4/vc4_dsi.c 	if (!tx) {
tx                572 drivers/gpu/drm/vc4/vc4_dsi.c 	cookie = tx->tx_submit(tx);
tx                 52 drivers/hid/hid-led.c 	} tx;
tx                291 drivers/hid/hid-led.c 	union delcom_packet dp = { .tx.major_cmd = 101, .tx.minor_cmd = 12 };
tx                293 drivers/hid/hid-led.c 	dp.tx.data_lsb = 1 << delcom_get_lednum(led);
tx                294 drivers/hid/hid-led.c 	dp.tx.data_msb = 0;
tx                301 drivers/hid/hid-led.c 	union delcom_packet dp = { .tx.major_cmd = 101, .tx.minor_cmd = 34 };
tx                303 drivers/hid/hid-led.c 	dp.tx.data_lsb = delcom_get_lednum(led);
tx                304 drivers/hid/hid-led.c 	dp.tx.data_msb = led->cdev.brightness;
tx                162 drivers/hwmon/sht21.c 	u8 tx[2];
tx                170 drivers/hwmon/sht21.c 			.buf = tx,
tx                181 drivers/hwmon/sht21.c 	tx[0] = SHT21_READ_SNB_CMD1;
tx                182 drivers/hwmon/sht21.c 	tx[1] = SHT21_READ_SNB_CMD2;
tx                191 drivers/hwmon/sht21.c 	tx[0] = SHT21_READ_SNAC_CMD1;
tx                192 drivers/hwmon/sht21.c 	tx[1] = SHT21_READ_SNAC_CMD2;
tx                179 drivers/i2c/busses/i2c-cpm.c 	struct i2c_msg *pmsg, int num, int tx, int rx)
tx                188 drivers/i2c/busses/i2c-cpm.c 	tbdf = cpm->tbase + tx;
tx                193 drivers/i2c/busses/i2c-cpm.c 	tb = cpm->txbuf[tx];
tx                207 drivers/i2c/busses/i2c-cpm.c 	if (tx + 1 == num)
tx                238 drivers/i2c/busses/i2c-cpm.c 	struct i2c_msg *pmsg, int tx, int rx)
tx                246 drivers/i2c/busses/i2c-cpm.c 	tbdf = cpm->tbase + tx;
tx                249 drivers/i2c/busses/i2c-cpm.c 	tb = cpm->txbuf[tx];
tx                275 drivers/i2c/busses/i2c-cpm.c 		dev_dbg(&adap->dev, "tx sc %d 0x%04x\n", tx,
tx                 58 drivers/i2c/busses/i2c-dln2.c 	} tx;
tx                 60 drivers/i2c/busses/i2c-dln2.c 	tx.port = dln2->port;
tx                 67 drivers/i2c/busses/i2c-dln2.c 	return dln2_transfer_tx(dln2->pdev, cmd, &tx, sizeof(tx));
tx                 81 drivers/i2c/busses/i2c-dln2.c 	} __packed *tx = dln2->buf;
tx                 84 drivers/i2c/busses/i2c-dln2.c 	BUILD_BUG_ON(sizeof(*tx) > DLN2_I2C_BUF_SIZE);
tx                 86 drivers/i2c/busses/i2c-dln2.c 	tx->port = dln2->port;
tx                 87 drivers/i2c/busses/i2c-dln2.c 	tx->addr = addr;
tx                 88 drivers/i2c/busses/i2c-dln2.c 	tx->mem_addr_len = 0;
tx                 89 drivers/i2c/busses/i2c-dln2.c 	tx->mem_addr = 0;
tx                 90 drivers/i2c/busses/i2c-dln2.c 	tx->buf_len = cpu_to_le16(data_len);
tx                 91 drivers/i2c/busses/i2c-dln2.c 	memcpy(tx->buf, data, data_len);
tx                 93 drivers/i2c/busses/i2c-dln2.c 	len = sizeof(*tx) + data_len - DLN2_I2C_MAX_XFER_SIZE;
tx                 94 drivers/i2c/busses/i2c-dln2.c 	ret = dln2_transfer_tx(dln2->pdev, DLN2_I2C_WRITE, tx, len);
tx                111 drivers/i2c/busses/i2c-dln2.c 	} __packed tx;
tx                120 drivers/i2c/busses/i2c-dln2.c 	tx.port = dln2->port;
tx                121 drivers/i2c/busses/i2c-dln2.c 	tx.addr = addr;
tx                122 drivers/i2c/busses/i2c-dln2.c 	tx.mem_addr_len = 0;
tx                123 drivers/i2c/busses/i2c-dln2.c 	tx.mem_addr = 0;
tx                124 drivers/i2c/busses/i2c-dln2.c 	tx.buf_len = cpu_to_le16(data_len);
tx                126 drivers/i2c/busses/i2c-dln2.c 	ret = dln2_transfer(dln2->pdev, DLN2_I2C_READ, &tx, sizeof(tx),
tx                210 drivers/i2c/busses/i2c-tegra-bpmp.c 	msg.tx.data = request;
tx                211 drivers/i2c/busses/i2c-tegra-bpmp.c 	msg.tx.size = sizeof(*request);
tx                171 drivers/iio/accel/sca3000.c 	u8				tx[6] ____cacheline_aligned;
tx                279 drivers/iio/accel/sca3000.c 	st->tx[0] = SCA3000_WRITE_REG(address);
tx                280 drivers/iio/accel/sca3000.c 	st->tx[1] = val;
tx                281 drivers/iio/accel/sca3000.c 	return spi_write(st->us, st->tx, 2);
tx                291 drivers/iio/accel/sca3000.c 			.tx_buf = st->tx,
tx                297 drivers/iio/accel/sca3000.c 	st->tx[0] = SCA3000_READ_REG(reg_address_high);
tx                333 drivers/iio/accel/sca3000.c 			.tx_buf = st->tx,
tx                337 drivers/iio/accel/sca3000.c 			.tx_buf = st->tx + 2,
tx                340 drivers/iio/accel/sca3000.c 			.tx_buf = st->tx + 4,
tx                343 drivers/iio/accel/sca3000.c 	st->tx[0] = SCA3000_WRITE_REG(SCA3000_REG_UNLOCK_ADDR);
tx                344 drivers/iio/accel/sca3000.c 	st->tx[1] = 0x00;
tx                345 drivers/iio/accel/sca3000.c 	st->tx[2] = SCA3000_WRITE_REG(SCA3000_REG_UNLOCK_ADDR);
tx                346 drivers/iio/accel/sca3000.c 	st->tx[3] = 0x50;
tx                347 drivers/iio/accel/sca3000.c 	st->tx[4] = SCA3000_WRITE_REG(SCA3000_REG_UNLOCK_ADDR);
tx                348 drivers/iio/accel/sca3000.c 	st->tx[5] = 0xA0;
tx                973 drivers/iio/accel/sca3000.c 			.tx_buf = st->tx,
tx                980 drivers/iio/accel/sca3000.c 	st->tx[0] = SCA3000_READ_REG(reg_address_high);
tx                 67 drivers/iio/adc/ad7949.c 	struct spi_transfer tx[] = {
tx                 77 drivers/iio/adc/ad7949.c 	spi_message_init_with_transfers(&msg, tx, 1);
tx                 96 drivers/iio/adc/ad7949.c 	struct spi_transfer tx[] = {
tx                122 drivers/iio/adc/ad7949.c 	spi_message_init_with_transfers(&msg, tx, 1);
tx                793 drivers/iio/adc/max1363.c 	u8 tx[2] = { st->setupbyte,
tx                800 drivers/iio/adc/max1363.c 	st->send(st->client, tx, 2);
tx                 76 drivers/iio/gyro/adxrs450.c 	__be32			tx ____cacheline_aligned;
tx                 93 drivers/iio/gyro/adxrs450.c 	u32 tx;
tx                 97 drivers/iio/gyro/adxrs450.c 			.tx_buf = &st->tx,
tx                 99 drivers/iio/gyro/adxrs450.c 			.len = sizeof(st->tx),
tx                109 drivers/iio/gyro/adxrs450.c 	tx = ADXRS450_READ_DATA | (reg_address << 17);
tx                111 drivers/iio/gyro/adxrs450.c 	if (!(hweight32(tx) & 1))
tx                112 drivers/iio/gyro/adxrs450.c 		tx |= ADXRS450_P;
tx                114 drivers/iio/gyro/adxrs450.c 	st->tx = cpu_to_be32(tx);
tx                141 drivers/iio/gyro/adxrs450.c 	u32 tx;
tx                145 drivers/iio/gyro/adxrs450.c 	tx = ADXRS450_WRITE_DATA | (reg_address << 17) | (val << 1);
tx                147 drivers/iio/gyro/adxrs450.c 	if (!(hweight32(tx) & 1))
tx                148 drivers/iio/gyro/adxrs450.c 		tx |= ADXRS450_P;
tx                150 drivers/iio/gyro/adxrs450.c 	st->tx = cpu_to_be32(tx);
tx                151 drivers/iio/gyro/adxrs450.c 	ret = spi_write(st->us, &st->tx, sizeof(st->tx));
tx                171 drivers/iio/gyro/adxrs450.c 			.tx_buf = &st->tx,
tx                173 drivers/iio/gyro/adxrs450.c 			.len = sizeof(st->tx),
tx                183 drivers/iio/gyro/adxrs450.c 	st->tx = cpu_to_be32(ADXRS450_SENSOR_DATA);
tx                208 drivers/iio/gyro/adxrs450.c 	u32 tx;
tx                210 drivers/iio/gyro/adxrs450.c 		.tx_buf = &st->tx,
tx                213 drivers/iio/gyro/adxrs450.c 		.len = sizeof(st->tx),
tx                217 drivers/iio/gyro/adxrs450.c 	tx = ADXRS450_SENSOR_DATA;
tx                219 drivers/iio/gyro/adxrs450.c 		tx |= (ADXRS450_CHK | ADXRS450_P);
tx                220 drivers/iio/gyro/adxrs450.c 	st->tx = cpu_to_be32(tx);
tx                 25 drivers/iio/gyro/itg3200_buffer.c 	u8 tx = 0x80 | ITG3200_REG_TEMP_OUT_H;
tx                 31 drivers/iio/gyro/itg3200_buffer.c 			.buf = &tx,
tx                214 drivers/iio/health/afe4403.c 	u8 tx[4] = {AFE440X_CONTROL0, 0x0, 0x0, AFE440X_CONTROL0_READ};
tx                219 drivers/iio/health/afe4403.c 	ret = spi_write_then_read(afe->spi, tx, 4, NULL, 0);
tx                232 drivers/iio/health/afe4403.c 	tx[3] = AFE440X_CONTROL0_WRITE;
tx                233 drivers/iio/health/afe4403.c 	ret = spi_write_then_read(afe->spi, tx, 4, NULL, 0);
tx                313 drivers/iio/health/afe4403.c 	u8 tx[4] = {AFE440X_CONTROL0, 0x0, 0x0, AFE440X_CONTROL0_READ};
tx                317 drivers/iio/health/afe4403.c 	ret = spi_write_then_read(afe->spi, tx, 4, NULL, 0);
tx                335 drivers/iio/health/afe4403.c 	tx[3] = AFE440X_CONTROL0_WRITE;
tx                336 drivers/iio/health/afe4403.c 	ret = spi_write_then_read(afe->spi, tx, 4, NULL, 0);
tx                 37 drivers/iio/imu/adis.c 			.tx_buf = adis->tx,
tx                 45 drivers/iio/imu/adis.c 			.tx_buf = adis->tx + 2,
tx                 53 drivers/iio/imu/adis.c 			.tx_buf = adis->tx + 4,
tx                 61 drivers/iio/imu/adis.c 			.tx_buf = adis->tx + 6,
tx                 66 drivers/iio/imu/adis.c 			.tx_buf = adis->tx + 8,
tx                 78 drivers/iio/imu/adis.c 		adis->tx[0] = ADIS_WRITE_REG(ADIS_REG_PAGE_ID);
tx                 79 drivers/iio/imu/adis.c 		adis->tx[1] = page;
tx                 85 drivers/iio/imu/adis.c 		adis->tx[8] = ADIS_WRITE_REG(reg + 3);
tx                 86 drivers/iio/imu/adis.c 		adis->tx[9] = (value >> 24) & 0xff;
tx                 87 drivers/iio/imu/adis.c 		adis->tx[6] = ADIS_WRITE_REG(reg + 2);
tx                 88 drivers/iio/imu/adis.c 		adis->tx[7] = (value >> 16) & 0xff;
tx                 91 drivers/iio/imu/adis.c 		adis->tx[4] = ADIS_WRITE_REG(reg + 1);
tx                 92 drivers/iio/imu/adis.c 		adis->tx[5] = (value >> 8) & 0xff;
tx                 95 drivers/iio/imu/adis.c 		adis->tx[2] = ADIS_WRITE_REG(reg);
tx                 96 drivers/iio/imu/adis.c 		adis->tx[3] = value & 0xff;
tx                137 drivers/iio/imu/adis.c 			.tx_buf = adis->tx,
tx                145 drivers/iio/imu/adis.c 			.tx_buf = adis->tx + 2,
tx                153 drivers/iio/imu/adis.c 			.tx_buf = adis->tx + 4,
tx                173 drivers/iio/imu/adis.c 		adis->tx[0] = ADIS_WRITE_REG(ADIS_REG_PAGE_ID);
tx                174 drivers/iio/imu/adis.c 		adis->tx[1] = page;
tx                180 drivers/iio/imu/adis.c 		adis->tx[2] = ADIS_READ_REG(reg + 2);
tx                181 drivers/iio/imu/adis.c 		adis->tx[3] = 0;
tx                185 drivers/iio/imu/adis.c 		adis->tx[4] = ADIS_READ_REG(reg);
tx                186 drivers/iio/imu/adis.c 		adis->tx[5] = 0;
tx                 27 drivers/iio/imu/adis_buffer.c 	u8 *tx;
tx                 44 drivers/iio/imu/adis_buffer.c 	tx = adis->buffer + burst_length;
tx                 45 drivers/iio/imu/adis_buffer.c 	tx[0] = ADIS_READ_REG(adis->burst->reg_cmd);
tx                 46 drivers/iio/imu/adis_buffer.c 	tx[1] = 0;
tx                 48 drivers/iio/imu/adis_buffer.c 	adis->xfer[0].tx_buf = tx;
tx                 69 drivers/iio/imu/adis_buffer.c 	__be16 *tx, *rx;
tx                 91 drivers/iio/imu/adis_buffer.c 	tx = rx + scan_count;
tx                102 drivers/iio/imu/adis_buffer.c 			adis->xfer[j].tx_buf = &tx[j];
tx                113 drivers/iio/imu/adis_buffer.c 			*tx++ = cpu_to_be16((chan->address + 2) << 8);
tx                114 drivers/iio/imu/adis_buffer.c 		*tx++ = cpu_to_be16(chan->address << 8);
tx                134 drivers/iio/imu/adis_buffer.c 			adis->tx[0] = ADIS_WRITE_REG(ADIS_REG_PAGE_ID);
tx                135 drivers/iio/imu/adis_buffer.c 			adis->tx[1] = 0;
tx                136 drivers/iio/imu/adis_buffer.c 			spi_write(adis->spi, adis->tx, 2);
tx                 19 drivers/iio/pressure/mpl115_spi.c 	u8 tx[4];
tx                 42 drivers/iio/pressure/mpl115_spi.c 		.tx_buf = buf->tx,
tx                 48 drivers/iio/pressure/mpl115_spi.c 	buf->tx[0] = MPL115_SPI_READ(address);
tx                 49 drivers/iio/pressure/mpl115_spi.c 	buf->tx[2] = MPL115_SPI_READ(address + 1);
tx                 63 drivers/iio/pressure/mpl115_spi.c 		.tx_buf = buf->tx,
tx                 67 drivers/iio/pressure/mpl115_spi.c 	buf->tx[0] = MPL115_SPI_WRITE(address);
tx                 68 drivers/iio/pressure/mpl115_spi.c 	buf->tx[1] = value;
tx               7333 drivers/infiniband/hw/hfi1/chip.c 	u16 tx, rx;
tx               7346 drivers/infiniband/hw/hfi1/chip.c 	tx = nibble_to_count(enable_lane_tx);
tx               7374 drivers/infiniband/hw/hfi1/chip.c 		    enable_lane_tx, tx, enable_lane_rx, rx);
tx               7375 drivers/infiniband/hw/hfi1/chip.c 	*tx_width = link_width_to_bits(dd, tx);
tx               7395 drivers/infiniband/hw/hfi1/chip.c 	u16 widths, tx, rx;
tx               7400 drivers/infiniband/hw/hfi1/chip.c 	tx = widths >> 12;
tx               7403 drivers/infiniband/hw/hfi1/chip.c 	*tx_width = link_width_to_bits(dd, tx);
tx               7625 drivers/infiniband/hw/hfi1/chip.c 	u16 tx, rx;
tx               7651 drivers/infiniband/hw/hfi1/chip.c 		get_link_widths(ppd->dd, &tx, &rx);
tx               7652 drivers/infiniband/hw/hfi1/chip.c 		ppd->link_width_downgrade_tx_active = tx;
tx                 46 drivers/infiniband/hw/hfi1/iowait.c 			      struct sdma_txreq *tx,
tx                144 drivers/infiniband/hw/hfi1/iowait.h 		struct sdma_txreq *tx,
tx                176 drivers/infiniband/hw/hfi1/iowait.h 			      struct sdma_txreq *tx,
tx                328 drivers/infiniband/hw/hfi1/iowait.h 	struct sdma_txreq *tx = NULL;
tx                331 drivers/infiniband/hw/hfi1/iowait.h 		tx = list_first_entry(
tx                335 drivers/infiniband/hw/hfi1/iowait.h 		list_del_init(&tx->list);
tx                337 drivers/infiniband/hw/hfi1/iowait.h 	return tx;
tx                343 drivers/infiniband/hw/hfi1/iowait.h 	struct sdma_txreq *tx = NULL;
tx                346 drivers/infiniband/hw/hfi1/iowait.h 		tx = list_first_entry(&w->tx_head, struct sdma_txreq,
tx                348 drivers/infiniband/hw/hfi1/iowait.h 		num_desc = tx->num_desc;
tx                349 drivers/infiniband/hw/hfi1/iowait.h 		if (tx->flags & SDMA_TXREQ_F_VIP)
tx                366 drivers/infiniband/hw/hfi1/iowait.h 	struct sdma_txreq *tx = NULL;
tx                369 drivers/infiniband/hw/hfi1/iowait.h 		tx = list_first_entry(&w->tx_head, struct sdma_txreq,
tx                371 drivers/infiniband/hw/hfi1/iowait.h 		if (tx->flags & SDMA_TXREQ_F_VIP)
tx                152 drivers/infiniband/hw/hfi1/qp.c 		struct sdma_txreq *tx;
tx                154 drivers/infiniband/hw/hfi1/qp.c 		tx = list_first_entry(
tx                158 drivers/infiniband/hw/hfi1/qp.c 		list_del_init(&tx->list);
tx                160 drivers/infiniband/hw/hfi1/qp.c 			container_of(tx, struct verbs_txreq, txreq));
tx                495 drivers/infiniband/hw/hfi1/qp.c 	struct verbs_txreq *tx = container_of(stx, struct verbs_txreq, txreq);
tx                501 drivers/infiniband/hw/hfi1/qp.c 	qp = tx->qp;
tx                535 drivers/infiniband/hw/hfi1/qp.c 		hfi1_put_txreq(tx);
tx                365 drivers/infiniband/hw/hfi1/sdma.c 			       struct sdma_txreq *tx,
tx                369 drivers/infiniband/hw/hfi1/sdma.c 	struct iowait *wait = tx->wait;
tx                370 drivers/infiniband/hw/hfi1/sdma.c 	callback_t complete = tx->complete;
tx                373 drivers/infiniband/hw/hfi1/sdma.c 	trace_hfi1_sdma_out_sn(sde, tx->sn);
tx                374 drivers/infiniband/hw/hfi1/sdma.c 	if (WARN_ON_ONCE(sde->head_sn != tx->sn))
tx                376 drivers/infiniband/hw/hfi1/sdma.c 			   sde->head_sn, tx->sn);
tx                379 drivers/infiniband/hw/hfi1/sdma.c 	__sdma_txclean(sde->dd, tx);
tx                381 drivers/infiniband/hw/hfi1/sdma.c 		(*complete)(tx, res);
tx               1658 drivers/infiniband/hw/hfi1/sdma.c static inline u8 ahg_mode(struct sdma_txreq *tx)
tx               1660 drivers/infiniband/hw/hfi1/sdma.c 	return (tx->descp[0].qw[1] & SDMA_DESC1_HEADER_MODE_SMASK)
tx               1677 drivers/infiniband/hw/hfi1/sdma.c 	struct sdma_txreq *tx)
tx               1681 drivers/infiniband/hw/hfi1/sdma.c 	if (tx->num_desc) {
tx               1682 drivers/infiniband/hw/hfi1/sdma.c 		u8 skip = 0, mode = ahg_mode(tx);
tx               1685 drivers/infiniband/hw/hfi1/sdma.c 		sdma_unmap_desc(dd, &tx->descp[0]);
tx               1689 drivers/infiniband/hw/hfi1/sdma.c 		for (i = 1 + skip; i < tx->num_desc; i++)
tx               1690 drivers/infiniband/hw/hfi1/sdma.c 			sdma_unmap_desc(dd, &tx->descp[i]);
tx               1691 drivers/infiniband/hw/hfi1/sdma.c 		tx->num_desc = 0;
tx               1693 drivers/infiniband/hw/hfi1/sdma.c 	kfree(tx->coalesce_buf);
tx               1694 drivers/infiniband/hw/hfi1/sdma.c 	tx->coalesce_buf = NULL;
tx               1696 drivers/infiniband/hw/hfi1/sdma.c 	if (unlikely(tx->desc_limit > ARRAY_SIZE(tx->descs))) {
tx               1697 drivers/infiniband/hw/hfi1/sdma.c 		tx->desc_limit = ARRAY_SIZE(tx->descs);
tx               1698 drivers/infiniband/hw/hfi1/sdma.c 		kfree(tx->descp);
tx               2307 drivers/infiniband/hw/hfi1/sdma.c static inline u16 submit_tx(struct sdma_engine *sde, struct sdma_txreq *tx)
tx               2311 drivers/infiniband/hw/hfi1/sdma.c 	struct sdma_desc *descp = tx->descp;
tx               2312 drivers/infiniband/hw/hfi1/sdma.c 	u8 skip = 0, mode = ahg_mode(tx);
tx               2323 drivers/infiniband/hw/hfi1/sdma.c 	for (i = 1; i < tx->num_desc; i++, descp++) {
tx               2340 drivers/infiniband/hw/hfi1/sdma.c 	tx->next_descq_idx = tail;
tx               2342 drivers/infiniband/hw/hfi1/sdma.c 	tx->sn = sde->tail_sn++;
tx               2343 drivers/infiniband/hw/hfi1/sdma.c 	trace_hfi1_sdma_in_sn(sde, tx->sn);
tx               2346 drivers/infiniband/hw/hfi1/sdma.c 	sde->tx_ring[sde->tx_tail++ & sde->sdma_mask] = tx;
tx               2347 drivers/infiniband/hw/hfi1/sdma.c 	sde->desc_avail -= tx->num_desc;
tx               2357 drivers/infiniband/hw/hfi1/sdma.c 	struct sdma_txreq *tx,
tx               2363 drivers/infiniband/hw/hfi1/sdma.c 	if (tx->num_desc <= sde->desc_avail)
tx               2371 drivers/infiniband/hw/hfi1/sdma.c 		ret = wait->iow->sleep(sde, wait, tx, seq, pkts_sent);
tx               2397 drivers/infiniband/hw/hfi1/sdma.c 		    struct sdma_txreq *tx,
tx               2405 drivers/infiniband/hw/hfi1/sdma.c 	if (unlikely(tx->tlen))
tx               2407 drivers/infiniband/hw/hfi1/sdma.c 	tx->wait = iowait_ioww_to_iow(wait);
tx               2412 drivers/infiniband/hw/hfi1/sdma.c 	if (unlikely(tx->num_desc > sde->desc_avail))
tx               2414 drivers/infiniband/hw/hfi1/sdma.c 	tail = submit_tx(sde, tx);
tx               2424 drivers/infiniband/hw/hfi1/sdma.c 	tx->next_descq_idx = 0;
tx               2426 drivers/infiniband/hw/hfi1/sdma.c 	tx->sn = sde->tail_sn++;
tx               2427 drivers/infiniband/hw/hfi1/sdma.c 	trace_hfi1_sdma_in_sn(sde, tx->sn);
tx               2430 drivers/infiniband/hw/hfi1/sdma.c 	list_add_tail(&tx->list, &sde->flushlist);
tx               2432 drivers/infiniband/hw/hfi1/sdma.c 	iowait_inc_wait_count(wait, tx->num_desc);
tx               2437 drivers/infiniband/hw/hfi1/sdma.c 	ret = sdma_check_progress(sde, wait, tx, pkts_sent);
tx               2477 drivers/infiniband/hw/hfi1/sdma.c 	struct sdma_txreq *tx, *tx_next;
tx               2485 drivers/infiniband/hw/hfi1/sdma.c 	list_for_each_entry_safe(tx, tx_next, tx_list, list) {
tx               2486 drivers/infiniband/hw/hfi1/sdma.c 		tx->wait = iowait_ioww_to_iow(wait);
tx               2489 drivers/infiniband/hw/hfi1/sdma.c 		if (unlikely(tx->num_desc > sde->desc_avail))
tx               2491 drivers/infiniband/hw/hfi1/sdma.c 		if (unlikely(tx->tlen)) {
tx               2495 drivers/infiniband/hw/hfi1/sdma.c 		list_del_init(&tx->list);
tx               2496 drivers/infiniband/hw/hfi1/sdma.c 		tail = submit_tx(sde, tx);
tx               2518 drivers/infiniband/hw/hfi1/sdma.c 	list_for_each_entry_safe(tx, tx_next, tx_list, list) {
tx               2519 drivers/infiniband/hw/hfi1/sdma.c 		tx->wait = iowait_ioww_to_iow(wait);
tx               2520 drivers/infiniband/hw/hfi1/sdma.c 		list_del_init(&tx->list);
tx               2521 drivers/infiniband/hw/hfi1/sdma.c 		tx->next_descq_idx = 0;
tx               2523 drivers/infiniband/hw/hfi1/sdma.c 		tx->sn = sde->tail_sn++;
tx               2524 drivers/infiniband/hw/hfi1/sdma.c 		trace_hfi1_sdma_in_sn(sde, tx->sn);
tx               2526 drivers/infiniband/hw/hfi1/sdma.c 		list_add_tail(&tx->list, &sde->flushlist);
tx               2528 drivers/infiniband/hw/hfi1/sdma.c 		iowait_inc_wait_count(wait, tx->num_desc);
tx               2535 drivers/infiniband/hw/hfi1/sdma.c 	ret = sdma_check_progress(sde, wait, tx, submit_count > 0);
tx               3056 drivers/infiniband/hw/hfi1/sdma.c static int _extend_sdma_tx_descs(struct hfi1_devdata *dd, struct sdma_txreq *tx)
tx               3061 drivers/infiniband/hw/hfi1/sdma.c 	if (unlikely((tx->num_desc == (MAX_DESC - 1)))) {
tx               3063 drivers/infiniband/hw/hfi1/sdma.c 		if (!tx->tlen) {
tx               3064 drivers/infiniband/hw/hfi1/sdma.c 			tx->desc_limit = MAX_DESC;
tx               3065 drivers/infiniband/hw/hfi1/sdma.c 		} else if (!tx->coalesce_buf) {
tx               3067 drivers/infiniband/hw/hfi1/sdma.c 			tx->coalesce_buf = kmalloc(tx->tlen + sizeof(u32),
tx               3069 drivers/infiniband/hw/hfi1/sdma.c 			if (!tx->coalesce_buf)
tx               3071 drivers/infiniband/hw/hfi1/sdma.c 			tx->coalesce_idx = 0;
tx               3076 drivers/infiniband/hw/hfi1/sdma.c 	if (unlikely(tx->num_desc == MAX_DESC))
tx               3079 drivers/infiniband/hw/hfi1/sdma.c 	tx->descp = kmalloc_array(
tx               3083 drivers/infiniband/hw/hfi1/sdma.c 	if (!tx->descp)
tx               3087 drivers/infiniband/hw/hfi1/sdma.c 	tx->desc_limit = MAX_DESC - 1;
tx               3089 drivers/infiniband/hw/hfi1/sdma.c 	for (i = 0; i < tx->num_desc; i++)
tx               3090 drivers/infiniband/hw/hfi1/sdma.c 		tx->descp[i] = tx->descs[i];
tx               3093 drivers/infiniband/hw/hfi1/sdma.c 	__sdma_txclean(dd, tx);
tx               3113 drivers/infiniband/hw/hfi1/sdma.c int ext_coal_sdma_tx_descs(struct hfi1_devdata *dd, struct sdma_txreq *tx,
tx               3120 drivers/infiniband/hw/hfi1/sdma.c 	rval = _extend_sdma_tx_descs(dd, tx);
tx               3122 drivers/infiniband/hw/hfi1/sdma.c 		__sdma_txclean(dd, tx);
tx               3127 drivers/infiniband/hw/hfi1/sdma.c 	if (tx->coalesce_buf) {
tx               3129 drivers/infiniband/hw/hfi1/sdma.c 			__sdma_txclean(dd, tx);
tx               3137 drivers/infiniband/hw/hfi1/sdma.c 			__sdma_txclean(dd, tx);
tx               3141 drivers/infiniband/hw/hfi1/sdma.c 		memcpy(tx->coalesce_buf + tx->coalesce_idx, kvaddr, len);
tx               3142 drivers/infiniband/hw/hfi1/sdma.c 		tx->coalesce_idx += len;
tx               3147 drivers/infiniband/hw/hfi1/sdma.c 		if (tx->tlen - tx->coalesce_idx)
tx               3151 drivers/infiniband/hw/hfi1/sdma.c 		pad_len = tx->packet_len & (sizeof(u32) - 1);
tx               3154 drivers/infiniband/hw/hfi1/sdma.c 			memset(tx->coalesce_buf + tx->coalesce_idx, 0, pad_len);
tx               3156 drivers/infiniband/hw/hfi1/sdma.c 			tx->packet_len += pad_len;
tx               3157 drivers/infiniband/hw/hfi1/sdma.c 			tx->tlen += pad_len;
tx               3162 drivers/infiniband/hw/hfi1/sdma.c 				      tx->coalesce_buf,
tx               3163 drivers/infiniband/hw/hfi1/sdma.c 				      tx->tlen,
tx               3167 drivers/infiniband/hw/hfi1/sdma.c 			__sdma_txclean(dd, tx);
tx               3172 drivers/infiniband/hw/hfi1/sdma.c 		tx->desc_limit = MAX_DESC;
tx               3173 drivers/infiniband/hw/hfi1/sdma.c 		return _sdma_txadd_daddr(dd, SDMA_MAP_SINGLE, tx,
tx               3174 drivers/infiniband/hw/hfi1/sdma.c 					 addr, tx->tlen);
tx               3201 drivers/infiniband/hw/hfi1/sdma.c int _pad_sdma_tx_descs(struct hfi1_devdata *dd, struct sdma_txreq *tx)
tx               3205 drivers/infiniband/hw/hfi1/sdma.c 	tx->num_desc++;
tx               3206 drivers/infiniband/hw/hfi1/sdma.c 	if ((unlikely(tx->num_desc == tx->desc_limit))) {
tx               3207 drivers/infiniband/hw/hfi1/sdma.c 		rval = _extend_sdma_tx_descs(dd, tx);
tx               3209 drivers/infiniband/hw/hfi1/sdma.c 			__sdma_txclean(dd, tx);
tx               3215 drivers/infiniband/hw/hfi1/sdma.c 		tx,
tx               3218 drivers/infiniband/hw/hfi1/sdma.c 		sizeof(u32) - (tx->packet_len & (sizeof(u32) - 1)));
tx               3219 drivers/infiniband/hw/hfi1/sdma.c 	_sdma_close_tx(dd, tx);
tx               3231 drivers/infiniband/hw/hfi1/sdma.c 	struct sdma_txreq *tx,
tx               3248 drivers/infiniband/hw/hfi1/sdma.c 	tx->num_desc++;
tx               3252 drivers/infiniband/hw/hfi1/sdma.c 		tx->num_desc++;
tx               3253 drivers/infiniband/hw/hfi1/sdma.c 		tx->descs[2].qw[0] = 0;
tx               3254 drivers/infiniband/hw/hfi1/sdma.c 		tx->descs[2].qw[1] = 0;
tx               3257 drivers/infiniband/hw/hfi1/sdma.c 		tx->num_desc++;
tx               3258 drivers/infiniband/hw/hfi1/sdma.c 		tx->descs[1].qw[0] = 0;
tx               3259 drivers/infiniband/hw/hfi1/sdma.c 		tx->descs[1].qw[1] = 0;
tx               3263 drivers/infiniband/hw/hfi1/sdma.c 	tx->descs[0].qw[1] |=
tx               3275 drivers/infiniband/hw/hfi1/sdma.c 		tx->descs[desc].qw[!!(i & 2)] |=
tx                481 drivers/infiniband/hw/hfi1/sdma.h 	struct sdma_txreq *tx,
tx                540 drivers/infiniband/hw/hfi1/sdma.h 	struct sdma_txreq *tx,
tx                553 drivers/infiniband/hw/hfi1/sdma.h 	tx->desc_limit = ARRAY_SIZE(tx->descs);
tx                554 drivers/infiniband/hw/hfi1/sdma.h 	tx->descp = &tx->descs[0];
tx                555 drivers/infiniband/hw/hfi1/sdma.h 	INIT_LIST_HEAD(&tx->list);
tx                556 drivers/infiniband/hw/hfi1/sdma.h 	tx->num_desc = 0;
tx                557 drivers/infiniband/hw/hfi1/sdma.h 	tx->flags = flags;
tx                558 drivers/infiniband/hw/hfi1/sdma.h 	tx->complete = cb;
tx                559 drivers/infiniband/hw/hfi1/sdma.h 	tx->coalesce_buf = NULL;
tx                560 drivers/infiniband/hw/hfi1/sdma.h 	tx->wait = NULL;
tx                561 drivers/infiniband/hw/hfi1/sdma.h 	tx->packet_len = tlen;
tx                562 drivers/infiniband/hw/hfi1/sdma.h 	tx->tlen = tx->packet_len;
tx                563 drivers/infiniband/hw/hfi1/sdma.h 	tx->descs[0].qw[0] = SDMA_DESC0_FIRST_DESC_FLAG;
tx                564 drivers/infiniband/hw/hfi1/sdma.h 	tx->descs[0].qw[1] = 0;
tx                566 drivers/infiniband/hw/hfi1/sdma.h 		tx->descs[0].qw[1] |=
tx                572 drivers/infiniband/hw/hfi1/sdma.h 		_sdma_txreq_ahgadd(tx, num_ahg, ahg_entry, ahg, ahg_hlen);
tx                608 drivers/infiniband/hw/hfi1/sdma.h 	struct sdma_txreq *tx,
tx                613 drivers/infiniband/hw/hfi1/sdma.h 	return sdma_txinit_ahg(tx, flags, tlen, 0, 0, NULL, 0, cb);
tx                636 drivers/infiniband/hw/hfi1/sdma.h 	struct sdma_txreq *tx,
tx                641 drivers/infiniband/hw/hfi1/sdma.h 	struct sdma_desc *desc = &tx->descp[tx->num_desc];
tx                643 drivers/infiniband/hw/hfi1/sdma.h 	if (!tx->num_desc) {
tx                659 drivers/infiniband/hw/hfi1/sdma.h int ext_coal_sdma_tx_descs(struct hfi1_devdata *dd, struct sdma_txreq *tx,
tx                665 drivers/infiniband/hw/hfi1/sdma.h static inline void sdma_txclean(struct hfi1_devdata *dd, struct sdma_txreq *tx)
tx                667 drivers/infiniband/hw/hfi1/sdma.h 	if (tx->num_desc)
tx                668 drivers/infiniband/hw/hfi1/sdma.h 		__sdma_txclean(dd, tx);
tx                673 drivers/infiniband/hw/hfi1/sdma.h 				  struct sdma_txreq *tx)
tx                675 drivers/infiniband/hw/hfi1/sdma.h 	tx->descp[tx->num_desc].qw[0] |=
tx                677 drivers/infiniband/hw/hfi1/sdma.h 	tx->descp[tx->num_desc].qw[1] |=
tx                679 drivers/infiniband/hw/hfi1/sdma.h 	if (tx->flags & SDMA_TXREQ_F_URGENT)
tx                680 drivers/infiniband/hw/hfi1/sdma.h 		tx->descp[tx->num_desc].qw[1] |=
tx                688 drivers/infiniband/hw/hfi1/sdma.h 	struct sdma_txreq *tx,
tx                695 drivers/infiniband/hw/hfi1/sdma.h 		tx,
tx                698 drivers/infiniband/hw/hfi1/sdma.h 	WARN_ON(len > tx->tlen);
tx                699 drivers/infiniband/hw/hfi1/sdma.h 	tx->tlen -= len;
tx                701 drivers/infiniband/hw/hfi1/sdma.h 	if (!tx->tlen) {
tx                702 drivers/infiniband/hw/hfi1/sdma.h 		if (tx->packet_len & (sizeof(u32) - 1)) {
tx                703 drivers/infiniband/hw/hfi1/sdma.h 			rval = _pad_sdma_tx_descs(dd, tx);
tx                707 drivers/infiniband/hw/hfi1/sdma.h 			_sdma_close_tx(dd, tx);
tx                710 drivers/infiniband/hw/hfi1/sdma.h 	tx->num_desc++;
tx                732 drivers/infiniband/hw/hfi1/sdma.h 	struct sdma_txreq *tx,
tx                740 drivers/infiniband/hw/hfi1/sdma.h 	if ((unlikely(tx->num_desc == tx->desc_limit))) {
tx                741 drivers/infiniband/hw/hfi1/sdma.h 		rval = ext_coal_sdma_tx_descs(dd, tx, SDMA_MAP_PAGE,
tx                755 drivers/infiniband/hw/hfi1/sdma.h 		__sdma_txclean(dd, tx);
tx                760 drivers/infiniband/hw/hfi1/sdma.h 			dd, SDMA_MAP_PAGE, tx, addr, len);
tx                781 drivers/infiniband/hw/hfi1/sdma.h 	struct sdma_txreq *tx,
tx                787 drivers/infiniband/hw/hfi1/sdma.h 	if ((unlikely(tx->num_desc == tx->desc_limit))) {
tx                788 drivers/infiniband/hw/hfi1/sdma.h 		rval = ext_coal_sdma_tx_descs(dd, tx, SDMA_MAP_NONE,
tx                794 drivers/infiniband/hw/hfi1/sdma.h 	return _sdma_txadd_daddr(dd, SDMA_MAP_NONE, tx, addr, len);
tx                815 drivers/infiniband/hw/hfi1/sdma.h 	struct sdma_txreq *tx,
tx                822 drivers/infiniband/hw/hfi1/sdma.h 	if ((unlikely(tx->num_desc == tx->desc_limit))) {
tx                823 drivers/infiniband/hw/hfi1/sdma.h 		rval = ext_coal_sdma_tx_descs(dd, tx, SDMA_MAP_SINGLE,
tx                836 drivers/infiniband/hw/hfi1/sdma.h 		__sdma_txclean(dd, tx);
tx                841 drivers/infiniband/hw/hfi1/sdma.h 			dd, SDMA_MAP_SINGLE, tx, addr, len);
tx                848 drivers/infiniband/hw/hfi1/sdma.h 		    struct sdma_txreq *tx,
tx                899 drivers/infiniband/hw/hfi1/sdma.h 				     struct sdma_txreq *tx)
tx                903 drivers/infiniband/hw/hfi1/sdma.h 		if (tx->num_desc > sde->desc_avail)
tx                131 drivers/infiniband/hw/hfi1/sdma_txreq.h static inline int sdma_txreq_built(struct sdma_txreq *tx)
tx                133 drivers/infiniband/hw/hfi1/sdma_txreq.h 	return tx->num_desc;
tx                 91 drivers/infiniband/hw/hfi1/user_sdma.c 			    struct user_sdma_txreq *tx, u32 datalen);
tx                 93 drivers/infiniband/hw/hfi1/user_sdma.c 				struct user_sdma_txreq *tx, u32 len);
tx                632 drivers/infiniband/hw/hfi1/user_sdma.c 				      struct user_sdma_txreq *tx)
tx                653 drivers/infiniband/hw/hfi1/user_sdma.c 			       (sizeof(tx->hdr) - 4));
tx                701 drivers/infiniband/hw/hfi1/user_sdma.c 			       struct user_sdma_txreq *tx,
tx                717 drivers/infiniband/hw/hfi1/user_sdma.c 	memcpy(&tx->hdr, &req->hdr, sizeof(tx->hdr));
tx                720 drivers/infiniband/hw/hfi1/user_sdma.c 		tx->hdr.pbc[0] = cpu_to_le16(pbclen);
tx                722 drivers/infiniband/hw/hfi1/user_sdma.c 	ret = check_header_template(req, &tx->hdr, lrhlen, datalen);
tx                725 drivers/infiniband/hw/hfi1/user_sdma.c 	ret = sdma_txinit_ahg(&tx->txreq, SDMA_TXREQ_F_AHG_COPY,
tx                726 drivers/infiniband/hw/hfi1/user_sdma.c 			      sizeof(tx->hdr) + datalen, req->ahg_idx,
tx                730 drivers/infiniband/hw/hfi1/user_sdma.c 	ret = sdma_txadd_kvaddr(pq->dd, &tx->txreq, &tx->hdr, sizeof(tx->hdr));
tx                732 drivers/infiniband/hw/hfi1/user_sdma.c 		sdma_txclean(pq->dd, &tx->txreq);
tx                737 drivers/infiniband/hw/hfi1/user_sdma.c 			   struct user_sdma_txreq *tx,
tx                756 drivers/infiniband/hw/hfi1/user_sdma.c 	ret = sdma_txadd_page(pq->dd, &tx->txreq, iovec->pages[pageidx],
tx                783 drivers/infiniband/hw/hfi1/user_sdma.c 	struct user_sdma_txreq *tx = NULL;
tx                820 drivers/infiniband/hw/hfi1/user_sdma.c 		tx = kmem_cache_alloc(pq->txreq_cache, GFP_KERNEL);
tx                821 drivers/infiniband/hw/hfi1/user_sdma.c 		if (!tx)
tx                824 drivers/infiniband/hw/hfi1/user_sdma.c 		tx->flags = 0;
tx                825 drivers/infiniband/hw/hfi1/user_sdma.c 		tx->req = req;
tx                826 drivers/infiniband/hw/hfi1/user_sdma.c 		INIT_LIST_HEAD(&tx->list);
tx                833 drivers/infiniband/hw/hfi1/user_sdma.c 			tx->flags |= (TXREQ_FLAGS_REQ_ACK |
tx                852 drivers/infiniband/hw/hfi1/user_sdma.c 			datalen = compute_data_length(req, tx);
tx                868 drivers/infiniband/hw/hfi1/user_sdma.c 				tx->flags |= TXREQ_FLAGS_REQ_DISABLE_SH;
tx                874 drivers/infiniband/hw/hfi1/user_sdma.c 				ret = user_sdma_txadd_ahg(req, tx, datalen);
tx                880 drivers/infiniband/hw/hfi1/user_sdma.c 				changes = set_txreq_header_ahg(req, tx,
tx                888 drivers/infiniband/hw/hfi1/user_sdma.c 			ret = sdma_txinit(&tx->txreq, 0, sizeof(req->hdr) +
tx                898 drivers/infiniband/hw/hfi1/user_sdma.c 			ret = set_txreq_header(req, tx, datalen);
tx                909 drivers/infiniband/hw/hfi1/user_sdma.c 			ret = user_sdma_txadd(req, tx, iovec, datalen,
tx                924 drivers/infiniband/hw/hfi1/user_sdma.c 		list_add_tail(&tx->txreq.list, &req->txps);
tx                930 drivers/infiniband/hw/hfi1/user_sdma.c 		tx->seqnum = req->seqnum++;
tx                951 drivers/infiniband/hw/hfi1/user_sdma.c 	sdma_txclean(pq->dd, &tx->txreq);
tx                953 drivers/infiniband/hw/hfi1/user_sdma.c 	kmem_cache_free(pq->txreq_cache, tx);
tx               1157 drivers/infiniband/hw/hfi1/user_sdma.c 			    struct user_sdma_txreq *tx, u32 datalen)
tx               1160 drivers/infiniband/hw/hfi1/user_sdma.c 	struct hfi1_pkt_header *hdr = &tx->hdr;
tx               1214 drivers/infiniband/hw/hfi1/user_sdma.c 	if (unlikely(tx->flags & TXREQ_FLAGS_REQ_ACK))
tx               1249 drivers/infiniband/hw/hfi1/user_sdma.c 		if (unlikely(tx->flags & TXREQ_FLAGS_REQ_DISABLE_SH))
tx               1267 drivers/infiniband/hw/hfi1/user_sdma.c 	return sdma_txadd_kvaddr(pq->dd, &tx->txreq, hdr, sizeof(*hdr));
tx               1271 drivers/infiniband/hw/hfi1/user_sdma.c 				struct user_sdma_txreq *tx, u32 datalen)
tx               1301 drivers/infiniband/hw/hfi1/user_sdma.c 	if (unlikely(tx->flags & TXREQ_FLAGS_REQ_ACK))
tx               1357 drivers/infiniband/hw/hfi1/user_sdma.c 		if (unlikely(tx->flags & TXREQ_FLAGS_REQ_DISABLE_SH)) {
tx               1378 drivers/infiniband/hw/hfi1/user_sdma.c 	sdma_txinit_ahg(&tx->txreq,
tx               1399 drivers/infiniband/hw/hfi1/user_sdma.c 	struct user_sdma_txreq *tx =
tx               1406 drivers/infiniband/hw/hfi1/user_sdma.c 	if (!tx->req)
tx               1409 drivers/infiniband/hw/hfi1/user_sdma.c 	req = tx->req;
tx               1420 drivers/infiniband/hw/hfi1/user_sdma.c 	req->seqcomp = tx->seqnum;
tx               1421 drivers/infiniband/hw/hfi1/user_sdma.c 	kmem_cache_free(pq->txreq_cache, tx);
tx               1446 drivers/infiniband/hw/hfi1/user_sdma.c 			struct user_sdma_txreq *tx =
tx               1450 drivers/infiniband/hw/hfi1/user_sdma.c 			kmem_cache_free(req->pq->txreq_cache, tx);
tx                630 drivers/infiniband/hw/hfi1/verbs.c 	struct verbs_txreq *tx =
tx                632 drivers/infiniband/hw/hfi1/verbs.c 	struct rvt_qp *qp = tx->qp;
tx                635 drivers/infiniband/hw/hfi1/verbs.c 	if (tx->wqe) {
tx                636 drivers/infiniband/hw/hfi1/verbs.c 		rvt_send_complete(qp, tx->wqe, IB_WC_SUCCESS);
tx                640 drivers/infiniband/hw/hfi1/verbs.c 		hdr = &tx->phdr.hdr;
tx                647 drivers/infiniband/hw/hfi1/verbs.c 	hfi1_put_txreq(tx);
tx                698 drivers/infiniband/hw/hfi1/verbs.c 	struct verbs_txreq *tx)
tx                700 drivers/infiniband/hw/hfi1/verbs.c 	struct rvt_sge_state *ss = tx->ss;
tx                712 drivers/infiniband/hw/hfi1/verbs.c 			&tx->txreq,
tx                762 drivers/infiniband/hw/hfi1/verbs.c 	struct verbs_txreq *tx,
tx                767 drivers/infiniband/hw/hfi1/verbs.c 	struct hfi1_sdma_header *phdr = &tx->phdr;
tx                768 drivers/infiniband/hw/hfi1/verbs.c 	u16 hdrbytes = (tx->hdr_dwords + sizeof(pbc) / 4) << 2;
tx                771 drivers/infiniband/hw/hfi1/verbs.c 	if (tx->phdr.hdr.hdr_type) {
tx                781 drivers/infiniband/hw/hfi1/verbs.c 			&tx->txreq,
tx                795 drivers/infiniband/hw/hfi1/verbs.c 			&tx->txreq,
tx                802 drivers/infiniband/hw/hfi1/verbs.c 			&tx->txreq,
tx                814 drivers/infiniband/hw/hfi1/verbs.c 	if (tx->ss) {
tx                815 drivers/infiniband/hw/hfi1/verbs.c 		ret = build_verbs_ulp_payload(sde, length, tx);
tx                822 drivers/infiniband/hw/hfi1/verbs.c 		ret = sdma_txadd_daddr(sde->dd, &tx->txreq,
tx                848 drivers/infiniband/hw/hfi1/verbs.c 	struct verbs_txreq *tx;
tx                863 drivers/infiniband/hw/hfi1/verbs.c 	tx = ps->s_txreq;
tx                864 drivers/infiniband/hw/hfi1/verbs.c 	if (!sdma_txreq_built(&tx->txreq)) {
tx                888 drivers/infiniband/hw/hfi1/verbs.c 		tx->wqe = qp->s_wqe;
tx                889 drivers/infiniband/hw/hfi1/verbs.c 		ret = build_verbs_tx_desc(tx->sde, len, tx, ahg_info, pbc);
tx                893 drivers/infiniband/hw/hfi1/verbs.c 	ret =  sdma_send_txreq(tx->sde, ps->wait, &tx->txreq, ps->pkts_sent);
tx               1219 drivers/infiniband/hw/hfi1/verbs.c 	struct verbs_txreq *tx = ps->s_txreq;
tx               1232 drivers/infiniband/hw/hfi1/verbs.c 			(tx->s_cur_size + priv->s_running_pkt_size) / 2;
tx               1237 drivers/infiniband/hw/hfi1/verbs.c 		    !sdma_txreq_built(&tx->txreq))
tx                402 drivers/infiniband/hw/hfi1/verbs.h void hfi1_put_txreq(struct verbs_txreq *tx);
tx                 55 drivers/infiniband/hw/hfi1/verbs_txreq.c void hfi1_put_txreq(struct verbs_txreq *tx)
tx                 63 drivers/infiniband/hw/hfi1/verbs_txreq.c 	qp = tx->qp;
tx                 66 drivers/infiniband/hw/hfi1/verbs_txreq.c 	if (tx->mr)
tx                 67 drivers/infiniband/hw/hfi1/verbs_txreq.c 		rvt_put_mr(tx->mr);
tx                 69 drivers/infiniband/hw/hfi1/verbs_txreq.c 	sdma_txclean(dd_from_dev(dev), &tx->txreq);
tx                 72 drivers/infiniband/hw/hfi1/verbs_txreq.c 	kmem_cache_free(dev->verbs_txreq_cache, tx);
tx                 97 drivers/infiniband/hw/hfi1/verbs_txreq.c 	struct verbs_txreq *tx = NULL;
tx                103 drivers/infiniband/hw/hfi1/verbs_txreq.c 		tx = kmem_cache_alloc(dev->verbs_txreq_cache, VERBS_TXREQ_GFP);
tx                104 drivers/infiniband/hw/hfi1/verbs_txreq.c 		if (tx)
tx                119 drivers/infiniband/hw/hfi1/verbs_txreq.c 	return tx;
tx                 80 drivers/infiniband/hw/hfi1/verbs_txreq.h 	struct verbs_txreq *tx;
tx                 83 drivers/infiniband/hw/hfi1/verbs_txreq.h 	tx = kmem_cache_alloc(dev->verbs_txreq_cache, VERBS_TXREQ_GFP);
tx                 84 drivers/infiniband/hw/hfi1/verbs_txreq.h 	if (unlikely(!tx)) {
tx                 86 drivers/infiniband/hw/hfi1/verbs_txreq.h 		tx = __get_txreq(dev, qp);
tx                 87 drivers/infiniband/hw/hfi1/verbs_txreq.h 		if (!tx)
tx                 88 drivers/infiniband/hw/hfi1/verbs_txreq.h 			return tx;
tx                 90 drivers/infiniband/hw/hfi1/verbs_txreq.h 	tx->qp = qp;
tx                 91 drivers/infiniband/hw/hfi1/verbs_txreq.h 	tx->mr = NULL;
tx                 92 drivers/infiniband/hw/hfi1/verbs_txreq.h 	tx->sde = priv->s_sde;
tx                 93 drivers/infiniband/hw/hfi1/verbs_txreq.h 	tx->psc = priv->s_sendcontext;
tx                 95 drivers/infiniband/hw/hfi1/verbs_txreq.h 	tx->txreq.num_desc = 0;
tx                 97 drivers/infiniband/hw/hfi1/verbs_txreq.h 	tx->phdr.hdr.hdr_type = priv->hdr_type;
tx                 98 drivers/infiniband/hw/hfi1/verbs_txreq.h 	tx->txreq.flags = 0;
tx                 99 drivers/infiniband/hw/hfi1/verbs_txreq.h 	return tx;
tx                102 drivers/infiniband/hw/hfi1/verbs_txreq.h static inline struct sdma_txreq *get_sdma_txreq(struct verbs_txreq *tx)
tx                104 drivers/infiniband/hw/hfi1/verbs_txreq.h 	return &tx->txreq;
tx                122 drivers/infiniband/hw/hfi1/verbs_txreq.h void hfi1_put_txreq(struct verbs_txreq *tx);
tx                 83 drivers/infiniband/hw/hfi1/vnic_sdma.c 	struct vnic_txreq *tx = container_of(txreq, struct vnic_txreq, txreq);
tx                 84 drivers/infiniband/hw/hfi1/vnic_sdma.c 	struct hfi1_vnic_sdma *vnic_sdma = tx->sdma;
tx                 87 drivers/infiniband/hw/hfi1/vnic_sdma.c 	dev_kfree_skb_any(tx->skb);
tx                 88 drivers/infiniband/hw/hfi1/vnic_sdma.c 	kmem_cache_free(vnic_sdma->dd->vnic.txreq_cache, tx);
tx                 92 drivers/infiniband/hw/hfi1/vnic_sdma.c 					   struct vnic_txreq *tx)
tx                 98 drivers/infiniband/hw/hfi1/vnic_sdma.c 		&tx->txreq,
tx                 99 drivers/infiniband/hw/hfi1/vnic_sdma.c 		tx->skb->data,
tx                100 drivers/infiniband/hw/hfi1/vnic_sdma.c 		skb_headlen(tx->skb));
tx                104 drivers/infiniband/hw/hfi1/vnic_sdma.c 	for (i = 0; i < skb_shinfo(tx->skb)->nr_frags; i++) {
tx                105 drivers/infiniband/hw/hfi1/vnic_sdma.c 		skb_frag_t *frag = &skb_shinfo(tx->skb)->frags[i];
tx                109 drivers/infiniband/hw/hfi1/vnic_sdma.c 				      &tx->txreq,
tx                117 drivers/infiniband/hw/hfi1/vnic_sdma.c 	if (tx->plen)
tx                118 drivers/infiniband/hw/hfi1/vnic_sdma.c 		ret = sdma_txadd_kvaddr(sde->dd, &tx->txreq,
tx                119 drivers/infiniband/hw/hfi1/vnic_sdma.c 					tx->pad + HFI1_VNIC_MAX_PAD - tx->plen,
tx                120 drivers/infiniband/hw/hfi1/vnic_sdma.c 					tx->plen);
tx                127 drivers/infiniband/hw/hfi1/vnic_sdma.c 			      struct vnic_txreq *tx,
tx                134 drivers/infiniband/hw/hfi1/vnic_sdma.c 		&tx->txreq,
tx                136 drivers/infiniband/hw/hfi1/vnic_sdma.c 		hdrbytes + tx->skb->len + tx->plen,
tx                146 drivers/infiniband/hw/hfi1/vnic_sdma.c 	tx->pbc_val = cpu_to_le64(pbc);
tx                149 drivers/infiniband/hw/hfi1/vnic_sdma.c 		&tx->txreq,
tx                150 drivers/infiniband/hw/hfi1/vnic_sdma.c 		&tx->pbc_val,
tx                156 drivers/infiniband/hw/hfi1/vnic_sdma.c 	ret = build_vnic_ulp_payload(sde, tx);
tx                173 drivers/infiniband/hw/hfi1/vnic_sdma.c 	struct vnic_txreq *tx;
tx                182 drivers/infiniband/hw/hfi1/vnic_sdma.c 	tx = kmem_cache_alloc(dd->vnic.txreq_cache, GFP_ATOMIC);
tx                183 drivers/infiniband/hw/hfi1/vnic_sdma.c 	if (unlikely(!tx)) {
tx                188 drivers/infiniband/hw/hfi1/vnic_sdma.c 	tx->sdma = vnic_sdma;
tx                189 drivers/infiniband/hw/hfi1/vnic_sdma.c 	tx->skb = skb;
tx                190 drivers/infiniband/hw/hfi1/vnic_sdma.c 	hfi1_vnic_update_pad(tx->pad, plen);
tx                191 drivers/infiniband/hw/hfi1/vnic_sdma.c 	tx->plen = plen;
tx                192 drivers/infiniband/hw/hfi1/vnic_sdma.c 	ret = build_vnic_tx_desc(sde, tx, pbc);
tx                197 drivers/infiniband/hw/hfi1/vnic_sdma.c 			      &tx->txreq, vnic_sdma->pkts_sent);
tx                209 drivers/infiniband/hw/hfi1/vnic_sdma.c 	sdma_txclean(dd, &tx->txreq);
tx                210 drivers/infiniband/hw/hfi1/vnic_sdma.c 	kmem_cache_free(dd->vnic.txreq_cache, tx);
tx                490 drivers/infiniband/hw/qib/qib_sdma.c 				  struct qib_verbs_txreq *tx)
tx                492 drivers/infiniband/hw/qib/qib_sdma.c 	struct qib_qp_priv *priv = tx->qp->priv;
tx                496 drivers/infiniband/hw/qib/qib_sdma.c 	tx->txreq.start_idx = 0;
tx                497 drivers/infiniband/hw/qib/qib_sdma.c 	tx->txreq.next_descq_idx = 0;
tx                498 drivers/infiniband/hw/qib/qib_sdma.c 	list_add_tail(&tx->txreq.list, &ppd->sdma_activelist);
tx                513 drivers/infiniband/hw/qib/qib_sdma.c 			struct qib_verbs_txreq *tx)
tx                530 drivers/infiniband/hw/qib/qib_sdma.c 		complete_sdma_err_req(ppd, tx);
tx                534 drivers/infiniband/hw/qib/qib_sdma.c 	if (tx->txreq.sg_count > qib_sdma_descq_freecnt(ppd)) {
tx                543 drivers/infiniband/hw/qib/qib_sdma.c 	dwoffset = tx->hdr_dwords;
tx                544 drivers/infiniband/hw/qib/qib_sdma.c 	make_sdma_desc(ppd, sdmadesc, (u64) tx->txreq.addr, dwoffset, 0);
tx                547 drivers/infiniband/hw/qib/qib_sdma.c 	if (tx->txreq.flags & QIB_SDMA_TXREQ_F_USELARGEBUF)
tx                563 drivers/infiniband/hw/qib/qib_sdma.c 	tx->txreq.start_idx = tail;
tx                580 drivers/infiniband/hw/qib/qib_sdma.c 		if (tx->txreq.flags & QIB_SDMA_TXREQ_F_USELARGEBUF)
tx                601 drivers/infiniband/hw/qib/qib_sdma.c 	if (tx->txreq.flags & QIB_SDMA_TXREQ_F_HEADTOHOST)
tx                603 drivers/infiniband/hw/qib/qib_sdma.c 	if (tx->txreq.flags & QIB_SDMA_TXREQ_F_INTREQ)
tx                605 drivers/infiniband/hw/qib/qib_sdma.c 	priv = tx->qp->priv;
tx                607 drivers/infiniband/hw/qib/qib_sdma.c 	tx->txreq.next_descq_idx = tail;
tx                609 drivers/infiniband/hw/qib/qib_sdma.c 	ppd->sdma_descq_added += tx->txreq.sg_count;
tx                610 drivers/infiniband/hw/qib/qib_sdma.c 	list_add_tail(&tx->txreq.list, &ppd->sdma_activelist);
tx                623 drivers/infiniband/hw/qib/qib_sdma.c 	qp = tx->qp;
tx                625 drivers/infiniband/hw/qib/qib_sdma.c 	qib_put_txreq(tx);
tx                640 drivers/infiniband/hw/qib/qib_sdma.c 	qp = tx->qp;
tx                651 drivers/infiniband/hw/qib/qib_sdma.c 		tx->ss = ss;
tx                652 drivers/infiniband/hw/qib/qib_sdma.c 		tx->dwords = dwords;
tx                653 drivers/infiniband/hw/qib/qib_sdma.c 		priv->s_tx = tx;
tx                670 drivers/infiniband/hw/qib/qib_sdma.c 		qib_put_txreq(tx);
tx                564 drivers/infiniband/hw/qib/qib_verbs.c 	struct qib_verbs_txreq *tx;
tx                576 drivers/infiniband/hw/qib/qib_verbs.c 		tx = list_entry(l, struct qib_verbs_txreq, txreq.list);
tx                587 drivers/infiniband/hw/qib/qib_verbs.c 		tx = ERR_PTR(-EBUSY);
tx                589 drivers/infiniband/hw/qib/qib_verbs.c 	return tx;
tx                595 drivers/infiniband/hw/qib/qib_verbs.c 	struct qib_verbs_txreq *tx;
tx                605 drivers/infiniband/hw/qib/qib_verbs.c 		tx = list_entry(l, struct qib_verbs_txreq, txreq.list);
tx                609 drivers/infiniband/hw/qib/qib_verbs.c 		tx =  __get_txreq(dev, qp);
tx                611 drivers/infiniband/hw/qib/qib_verbs.c 	return tx;
tx                614 drivers/infiniband/hw/qib/qib_verbs.c void qib_put_txreq(struct qib_verbs_txreq *tx)
tx                621 drivers/infiniband/hw/qib/qib_verbs.c 	qp = tx->qp;
tx                624 drivers/infiniband/hw/qib/qib_verbs.c 	if (tx->mr) {
tx                625 drivers/infiniband/hw/qib/qib_verbs.c 		rvt_put_mr(tx->mr);
tx                626 drivers/infiniband/hw/qib/qib_verbs.c 		tx->mr = NULL;
tx                628 drivers/infiniband/hw/qib/qib_verbs.c 	if (tx->txreq.flags & QIB_SDMA_TXREQ_F_FREEBUF) {
tx                629 drivers/infiniband/hw/qib/qib_verbs.c 		tx->txreq.flags &= ~QIB_SDMA_TXREQ_F_FREEBUF;
tx                631 drivers/infiniband/hw/qib/qib_verbs.c 				 tx->txreq.addr, tx->hdr_dwords << 2,
tx                633 drivers/infiniband/hw/qib/qib_verbs.c 		kfree(tx->align_buf);
tx                639 drivers/infiniband/hw/qib/qib_verbs.c 	list_add(&tx->txreq.list, &dev->txreq_free);
tx                714 drivers/infiniband/hw/qib/qib_verbs.c 	struct qib_verbs_txreq *tx =
tx                716 drivers/infiniband/hw/qib/qib_verbs.c 	struct rvt_qp *qp = tx->qp;
tx                720 drivers/infiniband/hw/qib/qib_verbs.c 	if (tx->wqe)
tx                721 drivers/infiniband/hw/qib/qib_verbs.c 		rvt_send_complete(qp, tx->wqe, IB_WC_SUCCESS);
tx                725 drivers/infiniband/hw/qib/qib_verbs.c 		if (tx->txreq.flags & QIB_SDMA_TXREQ_F_FREEBUF)
tx                726 drivers/infiniband/hw/qib/qib_verbs.c 			hdr = &tx->align_buf->hdr;
tx                730 drivers/infiniband/hw/qib/qib_verbs.c 			hdr = &dev->pio_hdrs[tx->hdr_inx].hdr;
tx                744 drivers/infiniband/hw/qib/qib_verbs.c 	qib_put_txreq(tx);
tx                780 drivers/infiniband/hw/qib/qib_verbs.c 	struct qib_verbs_txreq *tx;
tx                786 drivers/infiniband/hw/qib/qib_verbs.c 	tx = priv->s_tx;
tx                787 drivers/infiniband/hw/qib/qib_verbs.c 	if (tx) {
tx                790 drivers/infiniband/hw/qib/qib_verbs.c 		ret = qib_sdma_verbs_send(ppd, tx->ss, tx->dwords, tx);
tx                794 drivers/infiniband/hw/qib/qib_verbs.c 	tx = get_txreq(dev, qp);
tx                795 drivers/infiniband/hw/qib/qib_verbs.c 	if (IS_ERR(tx))
tx                800 drivers/infiniband/hw/qib/qib_verbs.c 	tx->qp = qp;
tx                801 drivers/infiniband/hw/qib/qib_verbs.c 	tx->wqe = qp->s_wqe;
tx                802 drivers/infiniband/hw/qib/qib_verbs.c 	tx->mr = qp->s_rdma_mr;
tx                805 drivers/infiniband/hw/qib/qib_verbs.c 	tx->txreq.callback = sdma_complete;
tx                807 drivers/infiniband/hw/qib/qib_verbs.c 		tx->txreq.flags = QIB_SDMA_TXREQ_F_HEADTOHOST;
tx                809 drivers/infiniband/hw/qib/qib_verbs.c 		tx->txreq.flags = QIB_SDMA_TXREQ_F_INTREQ;
tx                811 drivers/infiniband/hw/qib/qib_verbs.c 		tx->txreq.flags |= QIB_SDMA_TXREQ_F_USELARGEBUF;
tx                824 drivers/infiniband/hw/qib/qib_verbs.c 		phdr = &dev->pio_hdrs[tx->hdr_inx];
tx                828 drivers/infiniband/hw/qib/qib_verbs.c 		tx->txreq.flags |= QIB_SDMA_TXREQ_F_FREEDESC;
tx                829 drivers/infiniband/hw/qib/qib_verbs.c 		tx->txreq.sg_count = ndesc;
tx                830 drivers/infiniband/hw/qib/qib_verbs.c 		tx->txreq.addr = dev->pio_hdrs_phys +
tx                831 drivers/infiniband/hw/qib/qib_verbs.c 			tx->hdr_inx * sizeof(struct qib_pio_header);
tx                832 drivers/infiniband/hw/qib/qib_verbs.c 		tx->hdr_dwords = hdrwords + 2; /* add PBC length */
tx                833 drivers/infiniband/hw/qib/qib_verbs.c 		ret = qib_sdma_verbs_send(ppd, ss, dwords, tx);
tx                838 drivers/infiniband/hw/qib/qib_verbs.c 	tx->hdr_dwords = plen + 1;
tx                839 drivers/infiniband/hw/qib/qib_verbs.c 	phdr = kmalloc(tx->hdr_dwords << 2, GFP_ATOMIC);
tx                847 drivers/infiniband/hw/qib/qib_verbs.c 	tx->txreq.addr = dma_map_single(&dd->pcidev->dev, phdr,
tx                848 drivers/infiniband/hw/qib/qib_verbs.c 					tx->hdr_dwords << 2, DMA_TO_DEVICE);
tx                849 drivers/infiniband/hw/qib/qib_verbs.c 	if (dma_mapping_error(&dd->pcidev->dev, tx->txreq.addr))
tx                851 drivers/infiniband/hw/qib/qib_verbs.c 	tx->align_buf = phdr;
tx                852 drivers/infiniband/hw/qib/qib_verbs.c 	tx->txreq.flags |= QIB_SDMA_TXREQ_F_FREEBUF;
tx                853 drivers/infiniband/hw/qib/qib_verbs.c 	tx->txreq.sg_count = 1;
tx                854 drivers/infiniband/hw/qib/qib_verbs.c 	ret = qib_sdma_verbs_send(ppd, NULL, 0, tx);
tx                860 drivers/infiniband/hw/qib/qib_verbs.c 	qib_put_txreq(tx);
tx                867 drivers/infiniband/hw/qib/qib_verbs.c 	ret = PTR_ERR(tx);
tx               1532 drivers/infiniband/hw/qib/qib_verbs.c 		struct qib_verbs_txreq *tx;
tx               1534 drivers/infiniband/hw/qib/qib_verbs.c 		tx = kzalloc(sizeof(*tx), GFP_KERNEL);
tx               1535 drivers/infiniband/hw/qib/qib_verbs.c 		if (!tx) {
tx               1539 drivers/infiniband/hw/qib/qib_verbs.c 		tx->hdr_inx = i;
tx               1540 drivers/infiniband/hw/qib/qib_verbs.c 		list_add(&tx->txreq.list, &dev->txreq_free);
tx               1630 drivers/infiniband/hw/qib/qib_verbs.c 		struct qib_verbs_txreq *tx;
tx               1633 drivers/infiniband/hw/qib/qib_verbs.c 		tx = list_entry(l, struct qib_verbs_txreq, txreq.list);
tx               1634 drivers/infiniband/hw/qib/qib_verbs.c 		kfree(tx);
tx               1666 drivers/infiniband/hw/qib/qib_verbs.c 		struct qib_verbs_txreq *tx;
tx               1669 drivers/infiniband/hw/qib/qib_verbs.c 		tx = list_entry(l, struct qib_verbs_txreq, txreq.list);
tx               1670 drivers/infiniband/hw/qib/qib_verbs.c 		kfree(tx);
tx                290 drivers/infiniband/hw/qib/qib_verbs.h void qib_put_txreq(struct qib_verbs_txreq *tx);
tx                 59 drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h 	struct pvrdma_ring tx;	/* Tx ring. */
tx                492 drivers/infiniband/sw/siw/siw.h #define tx_qp(tx) container_of(tx, struct siw_qp, tx_ctx)
tx                649 drivers/infiniband/ulp/ipoib/ipoib.h static inline void ipoib_cm_set(struct ipoib_neigh *neigh, struct ipoib_cm_tx *tx)
tx                651 drivers/infiniband/ulp/ipoib/ipoib.h 	neigh->cm = tx;
tx                666 drivers/infiniband/ulp/ipoib/ipoib.h void ipoib_cm_send(struct net_device *dev, struct sk_buff *skb, struct ipoib_cm_tx *tx);
tx                674 drivers/infiniband/ulp/ipoib/ipoib.h void ipoib_cm_destroy_tx(struct ipoib_cm_tx *tx);
tx                706 drivers/infiniband/ulp/ipoib/ipoib.h static inline void ipoib_cm_set(struct ipoib_neigh *neigh, struct ipoib_cm_tx *tx)
tx                721 drivers/infiniband/ulp/ipoib/ipoib.h void ipoib_cm_send(struct net_device *dev, struct sk_buff *skb, struct ipoib_cm_tx *tx)
tx                758 drivers/infiniband/ulp/ipoib/ipoib.h void ipoib_cm_destroy_tx(struct ipoib_cm_tx *tx)
tx                697 drivers/infiniband/ulp/ipoib/ipoib_cm.c 			    struct ipoib_cm_tx *tx,
tx                705 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	return ib_post_send(tx->qp, &priv->tx_wr.wr, NULL);
tx                708 drivers/infiniband/ulp/ipoib/ipoib_cm.c void ipoib_cm_send(struct net_device *dev, struct sk_buff *skb, struct ipoib_cm_tx *tx)
tx                713 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	unsigned int usable_sge = tx->max_send_sge - !!skb_headlen(skb);
tx                715 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	if (unlikely(skb->len > tx->mtu)) {
tx                717 drivers/infiniband/ulp/ipoib/ipoib_cm.c 			   skb->len, tx->mtu);
tx                720 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		ipoib_cm_skb_too_long(dev, skb, tx->mtu - IPOIB_ENCAP_LEN);
tx                741 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		       tx->tx_head, skb->len, tx->qp->qp_num);
tx                750 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	tx_req = &tx->tx_ring[tx->tx_head & (ipoib_sendq_size - 1)];
tx                762 drivers/infiniband/ulp/ipoib/ipoib_cm.c 			  tx->qp->qp_num);
tx                778 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	rc = post_send(priv, tx, tx->tx_head & (ipoib_sendq_size - 1), tx_req);
tx                789 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		++tx->tx_head;
tx                797 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	struct ipoib_cm_tx *tx = wc->qp->qp_context;
tx                811 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	tx_req = &tx->tx_ring[wr_id];
tx                823 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	++tx->tx_tail;
tx                850 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		neigh = tx->neigh;
tx                856 drivers/infiniband/ulp/ipoib/ipoib_cm.c 			tx->neigh = NULL;
tx                859 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		if (test_and_clear_bit(IPOIB_FLAG_INITIALIZED, &tx->flags)) {
tx                860 drivers/infiniband/ulp/ipoib/ipoib_cm.c 			list_move(&tx->list, &priv->cm.reap_list);
tx                864 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		clear_bit(IPOIB_FLAG_OPER_UP, &tx->flags);
tx               1057 drivers/infiniband/ulp/ipoib/ipoib_cm.c static struct ib_qp *ipoib_cm_create_tx_qp(struct net_device *dev, struct ipoib_cm_tx *tx)
tx               1068 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		.qp_context		= tx,
tx               1078 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	tx->max_send_sge = attr.cap.max_send_sge;
tx               1256 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	struct ipoib_cm_tx *tx = cm_id->context;
tx               1257 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	struct ipoib_dev_priv *priv = ipoib_priv(tx->dev);
tx               1281 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		neigh = tx->neigh;
tx               1287 drivers/infiniband/ulp/ipoib/ipoib_cm.c 			tx->neigh = NULL;
tx               1290 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		if (test_and_clear_bit(IPOIB_FLAG_INITIALIZED, &tx->flags)) {
tx               1291 drivers/infiniband/ulp/ipoib/ipoib_cm.c 			list_move(&tx->list, &priv->cm.reap_list);
tx               1309 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	struct ipoib_cm_tx *tx;
tx               1311 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	tx = kzalloc(sizeof(*tx), GFP_ATOMIC);
tx               1312 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	if (!tx)
tx               1315 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	neigh->cm = tx;
tx               1316 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	tx->neigh = neigh;
tx               1317 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	tx->dev = dev;
tx               1318 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	list_add(&tx->list, &priv->cm.start_list);
tx               1319 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	set_bit(IPOIB_FLAG_INITIALIZED, &tx->flags);
tx               1321 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	return tx;
tx               1324 drivers/infiniband/ulp/ipoib/ipoib_cm.c void ipoib_cm_destroy_tx(struct ipoib_cm_tx *tx)
tx               1326 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	struct ipoib_dev_priv *priv = ipoib_priv(tx->dev);
tx               1328 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	if (test_and_clear_bit(IPOIB_FLAG_INITIALIZED, &tx->flags)) {
tx               1330 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		list_move(&tx->list, &priv->cm.reap_list);
tx               1333 drivers/infiniband/ulp/ipoib/ipoib_cm.c 			  tx->neigh->daddr + 4);
tx               1334 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		tx->neigh = NULL;
tx                 80 drivers/input/gameport/gameport.c 	unsigned int i, t, tx;
tx                 87 drivers/input/gameport/gameport.c 	tx = ~0;
tx                 99 drivers/input/gameport/gameport.c 		if (t < tx)
tx                100 drivers/input/gameport/gameport.c 			tx = t;
tx                105 drivers/input/gameport/gameport.c 	if (tx)
tx                106 drivers/input/gameport/gameport.c 		t /= tx;
tx                114 drivers/input/gameport/gameport.c 	unsigned int i, t, t1, t2, t3, tx;
tx                120 drivers/input/gameport/gameport.c 	tx = 1 << 30;
tx                130 drivers/input/gameport/gameport.c 		if ((t = DELTA(t2,t1) - DELTA(t3,t2)) < tx) tx = t;
tx                134 drivers/input/gameport/gameport.c 	return 59659 / (tx < 1 ? 1 : tx);
tx                139 drivers/input/gameport/gameport.c 	unsigned long tx, t1, t2, flags;
tx                144 drivers/input/gameport/gameport.c 	tx = 1 << 30;
tx                153 drivers/input/gameport/gameport.c 		if (t2 - t1 < tx) tx = t2 - t1;
tx                158 drivers/input/gameport/gameport.c 		(unsigned long)HZ / (1000 / 50)) / (tx < 1 ? 1 : tx);
tx                377 drivers/input/joystick/analog.c 	unsigned int i, t, tx;
tx                397 drivers/input/joystick/analog.c 	tx = ~0;
tx                410 drivers/input/joystick/analog.c 		if (t < tx) tx = t;
tx                413 drivers/input/joystick/analog.c         port->loop = tx / 50;
tx                216 drivers/input/rmi4/rmi_f54.c 	u8 tx = drv_data->num_tx_electrodes ? : f54->num_tx_electrodes;
tx                221 drivers/input/rmi4/rmi_f54.c 		size = rx * tx;
tx                228 drivers/input/rmi4/rmi_f54.c 		size = sizeof(u16) * rx * tx;
tx                403 drivers/input/rmi4/rmi_f54.c 	u8 tx = drv_data->num_tx_electrodes ? : f54->num_tx_electrodes;
tx                419 drivers/input/rmi4/rmi_f54.c 	f->height = tx;
tx                651 drivers/input/rmi4/rmi_f54.c 	u8 rx, tx;
tx                668 drivers/input/rmi4/rmi_f54.c 	tx = f54->num_tx_electrodes;
tx                670 drivers/input/rmi4/rmi_f54.c 					array3_size(tx, rx, sizeof(u16)),
tx                 62 drivers/input/serio/maceps2.c 			port->tx = val;
tx                 23 drivers/input/touchscreen/tsc2004.c 	u8 tx = TSC200X_CMD | TSC200X_CMD_12BIT | cmd;
tx                 27 drivers/input/touchscreen/tsc2004.c 	data = i2c_smbus_write_byte(i2c, tx);
tx                 26 drivers/input/touchscreen/tsc2005.c 	u8 tx = TSC200X_CMD | TSC200X_CMD_12BIT | cmd;
tx                 28 drivers/input/touchscreen/tsc2005.c 		.tx_buf         = &tx,
tx                106 drivers/ipack/devices/ipoctal.c 	stats->tx = 0;
tx                140 drivers/ipack/devices/ipoctal.c 	icount->tx = channel->stats.tx;
tx                203 drivers/ipack/devices/ipoctal.c 	channel->stats.tx++;
tx                 30 drivers/ipack/devices/ipoctal.h 	unsigned long tx;
tx                236 drivers/leds/trigger/ledtrig-netdev.c static DEVICE_ATTR_RW(tx);
tx                990 drivers/mailbox/bcm-pdc-mailbox.c 	struct pdc_ring_alloc tx;
tx                994 drivers/mailbox/bcm-pdc-mailbox.c 	tx.vbase = dma_pool_zalloc(pdcs->ring_pool, GFP_KERNEL, &tx.dmabase);
tx                995 drivers/mailbox/bcm-pdc-mailbox.c 	if (unlikely(!tx.vbase)) {
tx               1007 drivers/mailbox/bcm-pdc-mailbox.c 	dev_dbg(dev, " - base DMA addr of tx ring      %pad", &tx.dmabase);
tx               1008 drivers/mailbox/bcm-pdc-mailbox.c 	dev_dbg(dev, " - base virtual addr of tx ring  %p", tx.vbase);
tx               1012 drivers/mailbox/bcm-pdc-mailbox.c 	memcpy(&pdcs->tx_ring_alloc, &tx, sizeof(tx));
tx               1078 drivers/mailbox/bcm-pdc-mailbox.c 	dma_pool_free(pdcs->ring_pool, tx.vbase, tx.dmabase);
tx                 39 drivers/md/raid5-log.h 		       struct dma_async_tx_descriptor *tx);
tx                159 drivers/md/raid5-ppl.c 		       struct dma_async_tx_descriptor *tx)
tx                189 drivers/md/raid5-ppl.c 		return tx;
tx                192 drivers/md/raid5-ppl.c 	init_async_submit(&submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_ZERO_DST, tx,
tx                196 drivers/md/raid5-ppl.c 		tx = async_memcpy(sh->ppl_page, srcs[0], 0, 0, PAGE_SIZE,
tx                199 drivers/md/raid5-ppl.c 		tx = async_xor(sh->ppl_page, srcs, 0, count, PAGE_SIZE,
tx                202 drivers/md/raid5-ppl.c 	return tx;
tx                719 drivers/md/raid5-ppl.c 	struct dma_async_tx_descriptor *tx;
tx                724 drivers/md/raid5-ppl.c 	tx = async_xor(page1, xor_srcs, 0, 2, size, &submit);
tx                726 drivers/md/raid5-ppl.c 	async_tx_quiesce(&tx);
tx               1229 drivers/md/raid5.c 	sector_t sector, struct dma_async_tx_descriptor *tx,
tx               1246 drivers/md/raid5.c 	init_async_submit(&submit, flags, tx, NULL, NULL, NULL);
tx               1274 drivers/md/raid5.c 					tx = async_memcpy(*page, bio_page, page_offset,
tx               1277 drivers/md/raid5.c 				tx = async_memcpy(bio_page, *page, b_offset,
tx               1281 drivers/md/raid5.c 		submit.depend_tx = tx;
tx               1288 drivers/md/raid5.c 	return tx;
tx               1330 drivers/md/raid5.c 	struct dma_async_tx_descriptor *tx = NULL;
tx               1348 drivers/md/raid5.c 				tx = async_copy_data(0, rbi, &dev->page,
tx               1349 drivers/md/raid5.c 						     dev->sector, tx, sh, 0);
tx               1356 drivers/md/raid5.c 	init_async_submit(&submit, ASYNC_TX_ACK, tx, ops_complete_biofill, sh, NULL);
tx               1413 drivers/md/raid5.c 	struct dma_async_tx_descriptor *tx;
tx               1432 drivers/md/raid5.c 		tx = async_memcpy(xor_dest, xor_srcs[0], 0, 0, STRIPE_SIZE, &submit);
tx               1434 drivers/md/raid5.c 		tx = async_xor(xor_dest, xor_srcs, 0, count, STRIPE_SIZE, &submit);
tx               1436 drivers/md/raid5.c 	return tx;
tx               1493 drivers/md/raid5.c 	struct dma_async_tx_descriptor *tx;
tx               1525 drivers/md/raid5.c 		tx = async_gen_syndrome(blocks, 0, count+2, STRIPE_SIZE, &submit);
tx               1538 drivers/md/raid5.c 		tx = async_xor(dest, blocks, 0, count, STRIPE_SIZE, &submit);
tx               1541 drivers/md/raid5.c 	return tx;
tx               1555 drivers/md/raid5.c 	struct dma_async_tx_descriptor *tx;
tx               1624 drivers/md/raid5.c 			tx = async_xor(dest, blocks, 0, count, STRIPE_SIZE,
tx               1628 drivers/md/raid5.c 			init_async_submit(&submit, ASYNC_TX_FENCE, tx,
tx               1669 drivers/md/raid5.c 		struct dma_async_tx_descriptor *tx)
tx               1692 drivers/md/raid5.c 	init_async_submit(&submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_DROP_DST, tx,
tx               1694 drivers/md/raid5.c 	tx = async_xor(xor_dest, xor_srcs, 0, count, STRIPE_SIZE, &submit);
tx               1696 drivers/md/raid5.c 	return tx;
tx               1701 drivers/md/raid5.c 		struct dma_async_tx_descriptor *tx)
tx               1712 drivers/md/raid5.c 	init_async_submit(&submit, ASYNC_TX_FENCE|ASYNC_TX_PQ_XOR_DST, tx,
tx               1714 drivers/md/raid5.c 	tx = async_gen_syndrome(blocks, 0, count+2, STRIPE_SIZE,  &submit);
tx               1716 drivers/md/raid5.c 	return tx;
tx               1720 drivers/md/raid5.c ops_run_biodrain(struct stripe_head *sh, struct dma_async_tx_descriptor *tx)
tx               1763 drivers/md/raid5.c 					tx = async_copy_data(1, wbi, &dev->page,
tx               1764 drivers/md/raid5.c 							     dev->sector, tx, sh,
tx               1787 drivers/md/raid5.c 	return tx;
tx               1839 drivers/md/raid5.c 		     struct dma_async_tx_descriptor *tx)
tx               1904 drivers/md/raid5.c 		init_async_submit(&submit, flags, tx, ops_complete_reconstruct, head_sh,
tx               1908 drivers/md/raid5.c 		init_async_submit(&submit, flags, tx, NULL, NULL,
tx               1913 drivers/md/raid5.c 		tx = async_memcpy(xor_dest, xor_srcs[0], 0, 0, STRIPE_SIZE, &submit);
tx               1915 drivers/md/raid5.c 		tx = async_xor(xor_dest, xor_srcs, 0, count, STRIPE_SIZE, &submit);
tx               1926 drivers/md/raid5.c 		     struct dma_async_tx_descriptor *tx)
tx               1970 drivers/md/raid5.c 		init_async_submit(&submit, txflags, tx, ops_complete_reconstruct,
tx               1973 drivers/md/raid5.c 		init_async_submit(&submit, 0, tx, NULL, NULL,
tx               1975 drivers/md/raid5.c 	tx = async_gen_syndrome(blocks, 0, count+2, STRIPE_SIZE,  &submit);
tx               2003 drivers/md/raid5.c 	struct dma_async_tx_descriptor *tx;
tx               2023 drivers/md/raid5.c 	tx = async_xor_val(xor_dest, xor_srcs, 0, count, STRIPE_SIZE,
tx               2027 drivers/md/raid5.c 	init_async_submit(&submit, ASYNC_TX_ACK, tx, ops_complete_check, sh, NULL);
tx               2028 drivers/md/raid5.c 	tx = async_trigger_callback(&submit);
tx               2055 drivers/md/raid5.c 	struct dma_async_tx_descriptor *tx = NULL;
tx               2070 drivers/md/raid5.c 			tx = ops_run_compute5(sh, percpu);
tx               2073 drivers/md/raid5.c 				tx = ops_run_compute6_1(sh, percpu);
tx               2075 drivers/md/raid5.c 				tx = ops_run_compute6_2(sh, percpu);
tx               2078 drivers/md/raid5.c 		if (tx && !test_bit(STRIPE_OP_RECONSTRUCT, &ops_request))
tx               2079 drivers/md/raid5.c 			async_tx_ack(tx);
tx               2084 drivers/md/raid5.c 			tx = ops_run_prexor5(sh, percpu, tx);
tx               2086 drivers/md/raid5.c 			tx = ops_run_prexor6(sh, percpu, tx);
tx               2090 drivers/md/raid5.c 		tx = ops_run_partial_parity(sh, percpu, tx);
tx               2093 drivers/md/raid5.c 		tx = ops_run_biodrain(sh, tx);
tx               2099 drivers/md/raid5.c 			ops_run_reconstruct5(sh, percpu, tx);
tx               2101 drivers/md/raid5.c 			ops_run_reconstruct6(sh, percpu, tx);
tx               4310 drivers/md/raid5.c 	struct dma_async_tx_descriptor *tx = NULL;
tx               4337 drivers/md/raid5.c 			init_async_submit(&submit, 0, tx, NULL, NULL, NULL);
tx               4338 drivers/md/raid5.c 			tx = async_memcpy(sh2->dev[dd_idx].page,
tx               4357 drivers/md/raid5.c 	async_tx_quiesce(&tx);
tx                 19 drivers/media/dvb-frontends/cxd2880/cxd2880_spi_device.c 	struct spi_transfer tx;
tx                 27 drivers/media/dvb-frontends/cxd2880/cxd2880_spi_device.c 	memset(&tx, 0, sizeof(tx));
tx                 28 drivers/media/dvb-frontends/cxd2880/cxd2880_spi_device.c 	tx.tx_buf = data;
tx                 29 drivers/media/dvb-frontends/cxd2880/cxd2880_spi_device.c 	tx.len = size;
tx                 32 drivers/media/dvb-frontends/cxd2880/cxd2880_spi_device.c 	spi_message_add_tail(&tx, &msg);
tx               2014 drivers/media/dvb-frontends/dib7000p.c 	u8 *tx, *rx;
tx               2021 drivers/media/dvb-frontends/dib7000p.c 	tx = kzalloc(2, GFP_KERNEL);
tx               2022 drivers/media/dvb-frontends/dib7000p.c 	if (!tx)
tx               2030 drivers/media/dvb-frontends/dib7000p.c 	msg[0].buf = tx;
tx               2033 drivers/media/dvb-frontends/dib7000p.c 	tx[0] = 0x03;
tx               2034 drivers/media/dvb-frontends/dib7000p.c 	tx[1] = 0x00;
tx               2057 drivers/media/dvb-frontends/dib7000p.c 	kfree(tx);
tx                208 drivers/media/dvb-frontends/dib9000.c static int dib9000_risc_apb_access_read(struct dib9000_state *state, u32 address, u16 attribute, const u8 * tx, u32 txlen, u8 * b, u32 len);
tx               1012 drivers/media/dvb-frontends/dib9000.c static int dib9000_risc_apb_access_read(struct dib9000_state *state, u32 address, u16 attribute, const u8 * tx, u32 txlen, u8 * b, u32 len)
tx                263 drivers/media/dvb-frontends/dibx000_common.c static int dibx000_i2c_gate_ctrl(struct dibx000_i2c_master *mst, u8 tx[4],
tx                277 drivers/media/dvb-frontends/dibx000_common.c 	tx[0] = (((mst->base_reg + 1) >> 8) & 0xff);
tx                278 drivers/media/dvb-frontends/dibx000_common.c 	tx[1] = ((mst->base_reg + 1) & 0xff);
tx                279 drivers/media/dvb-frontends/dibx000_common.c 	tx[2] = val >> 8;
tx                280 drivers/media/dvb-frontends/dibx000_common.c 	tx[3] = val & 0xff;
tx                413 drivers/media/dvb-frontends/dibx000_common.c 	u8 tx[4];
tx                414 drivers/media/dvb-frontends/dibx000_common.c 	struct i2c_msg m = {.addr = mst->i2c_addr,.buf = tx,.len = 4 };
tx                416 drivers/media/dvb-frontends/dibx000_common.c 	dibx000_i2c_gate_ctrl(mst, tx, 0, 0);
tx                285 drivers/media/i2c/adv748x/adv748x-afe.c 	ret = adv748x_tx_power(afe->tx, enable);
tx                319 drivers/media/i2c/adv748x/adv748x-afe.c 	struct v4l2_subdev *tx;
tx                321 drivers/media/i2c/adv748x/adv748x-afe.c 	tx = adv748x_get_remote_sd(&afe->pads[ADV748X_AFE_SOURCE]);
tx                322 drivers/media/i2c/adv748x/adv748x-afe.c 	if (!tx)
tx                330 drivers/media/i2c/adv748x/adv748x-afe.c 	return adv748x_csi2_set_pixelrate(tx, 14318180);
tx                237 drivers/media/i2c/adv748x/adv748x-core.c static int adv748x_power_up_tx(struct adv748x_csi2 *tx)
tx                239 drivers/media/i2c/adv748x/adv748x-core.c 	struct adv748x_state *state = tx->state;
tx                240 drivers/media/i2c/adv748x/adv748x-core.c 	u8 page = is_txa(tx) ? ADV748X_PAGE_TXA : ADV748X_PAGE_TXB;
tx                244 drivers/media/i2c/adv748x/adv748x-core.c 	adv748x_write_check(state, page, 0x00, 0x80 | tx->num_lanes, &ret);
tx                247 drivers/media/i2c/adv748x/adv748x-core.c 	adv748x_write_check(state, page, 0x00, 0xa0 | tx->num_lanes, &ret);
tx                250 drivers/media/i2c/adv748x/adv748x-core.c 	if (tx->src == &state->hdmi.sd) {
tx                273 drivers/media/i2c/adv748x/adv748x-core.c 	adv748x_write_check(state, page, 0x00, 0x20 | tx->num_lanes, &ret);
tx                284 drivers/media/i2c/adv748x/adv748x-core.c static int adv748x_power_down_tx(struct adv748x_csi2 *tx)
tx                286 drivers/media/i2c/adv748x/adv748x-core.c 	struct adv748x_state *state = tx->state;
tx                287 drivers/media/i2c/adv748x/adv748x-core.c 	u8 page = is_txa(tx) ? ADV748X_PAGE_TXA : ADV748X_PAGE_TXB;
tx                295 drivers/media/i2c/adv748x/adv748x-core.c 	adv748x_write_check(state, page, 0x00, 0x80 | tx->num_lanes, &ret);
tx                306 drivers/media/i2c/adv748x/adv748x-core.c int adv748x_tx_power(struct adv748x_csi2 *tx, bool on)
tx                310 drivers/media/i2c/adv748x/adv748x-core.c 	if (!is_tx_enabled(tx))
tx                313 drivers/media/i2c/adv748x/adv748x-core.c 	val = tx_read(tx, ADV748X_CSI_FS_AS_LS);
tx                325 drivers/media/i2c/adv748x/adv748x-core.c 	return on ? adv748x_power_up_tx(tx) : adv748x_power_down_tx(tx);
tx                338 drivers/media/i2c/adv748x/adv748x-core.c 	struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd);
tx                346 drivers/media/i2c/adv748x/adv748x-core.c 	if (enable && tx->src)
tx                351 drivers/media/i2c/adv748x/adv748x-core.c 		state->afe.tx = enable ? tx : NULL;
tx                353 drivers/media/i2c/adv748x/adv748x-core.c 		state->hdmi.tx = enable ? tx : NULL;
tx                355 drivers/media/i2c/adv748x/adv748x-core.c 	tx->src = enable ? rsd : NULL;
tx                357 drivers/media/i2c/adv748x/adv748x-core.c 	if (state->afe.tx) {
tx                360 drivers/media/i2c/adv748x/adv748x-core.c 		if (is_txa(tx))
tx                366 drivers/media/i2c/adv748x/adv748x-core.c 	if (state->hdmi.tx)
tx                 17 drivers/media/i2c/adv748x/adv748x-csi2.c static int adv748x_csi2_set_virtual_channel(struct adv748x_csi2 *tx,
tx                 20 drivers/media/i2c/adv748x/adv748x-csi2.c 	return tx_write(tx, ADV748X_CSI_VC_REF, vc << ADV748X_CSI_VC_REF_SHIFT);
tx                 35 drivers/media/i2c/adv748x/adv748x-csi2.c static int adv748x_csi2_register_link(struct adv748x_csi2 *tx,
tx                 50 drivers/media/i2c/adv748x/adv748x-csi2.c 				    &tx->sd.entity, ADV748X_CSI2_SINK,
tx                 56 drivers/media/i2c/adv748x/adv748x-csi2.c 		tx->src = src;
tx                 71 drivers/media/i2c/adv748x/adv748x-csi2.c 	struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd);
tx                 72 drivers/media/i2c/adv748x/adv748x-csi2.c 	struct adv748x_state *state = tx->state;
tx                 75 drivers/media/i2c/adv748x/adv748x-csi2.c 	adv_dbg(state, "Registered %s (%s)", is_txa(tx) ? "TXA":"TXB",
tx                 85 drivers/media/i2c/adv748x/adv748x-csi2.c 		ret = adv748x_csi2_register_link(tx, sd->v4l2_dev,
tx                 88 drivers/media/i2c/adv748x/adv748x-csi2.c 						 is_txb(tx));
tx                 93 drivers/media/i2c/adv748x/adv748x-csi2.c 		if (is_txb(tx))
tx                 94 drivers/media/i2c/adv748x/adv748x-csi2.c 			state->afe.tx = tx;
tx                 98 drivers/media/i2c/adv748x/adv748x-csi2.c 	if (is_txb(tx) || !is_hdmi_enabled(state))
tx                101 drivers/media/i2c/adv748x/adv748x-csi2.c 	ret = adv748x_csi2_register_link(tx, sd->v4l2_dev, &state->hdmi.sd,
tx                107 drivers/media/i2c/adv748x/adv748x-csi2.c 	state->hdmi.tx = tx;
tx                122 drivers/media/i2c/adv748x/adv748x-csi2.c 	struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd);
tx                125 drivers/media/i2c/adv748x/adv748x-csi2.c 	src = adv748x_get_remote_sd(&tx->pads[ADV748X_CSI2_SINK]);
tx                148 drivers/media/i2c/adv748x/adv748x-csi2.c 	struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd);
tx                153 drivers/media/i2c/adv748x/adv748x-csi2.c 	return &tx->format;
tx                160 drivers/media/i2c/adv748x/adv748x-csi2.c 	struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd);
tx                161 drivers/media/i2c/adv748x/adv748x-csi2.c 	struct adv748x_state *state = tx->state;
tx                182 drivers/media/i2c/adv748x/adv748x-csi2.c 	struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd);
tx                183 drivers/media/i2c/adv748x/adv748x-csi2.c 	struct adv748x_state *state = tx->state;
tx                237 drivers/media/i2c/adv748x/adv748x-csi2.c 	struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd);
tx                239 drivers/media/i2c/adv748x/adv748x-csi2.c 	if (!tx->pixel_rate)
tx                242 drivers/media/i2c/adv748x/adv748x-csi2.c 	return v4l2_ctrl_s_ctrl_int64(tx->pixel_rate, rate);
tx                259 drivers/media/i2c/adv748x/adv748x-csi2.c static int adv748x_csi2_init_controls(struct adv748x_csi2 *tx)
tx                262 drivers/media/i2c/adv748x/adv748x-csi2.c 	v4l2_ctrl_handler_init(&tx->ctrl_hdl, 1);
tx                264 drivers/media/i2c/adv748x/adv748x-csi2.c 	tx->pixel_rate = v4l2_ctrl_new_std(&tx->ctrl_hdl,
tx                269 drivers/media/i2c/adv748x/adv748x-csi2.c 	tx->sd.ctrl_handler = &tx->ctrl_hdl;
tx                270 drivers/media/i2c/adv748x/adv748x-csi2.c 	if (tx->ctrl_hdl.error) {
tx                271 drivers/media/i2c/adv748x/adv748x-csi2.c 		v4l2_ctrl_handler_free(&tx->ctrl_hdl);
tx                272 drivers/media/i2c/adv748x/adv748x-csi2.c 		return tx->ctrl_hdl.error;
tx                275 drivers/media/i2c/adv748x/adv748x-csi2.c 	return v4l2_ctrl_handler_setup(&tx->ctrl_hdl);
tx                278 drivers/media/i2c/adv748x/adv748x-csi2.c int adv748x_csi2_init(struct adv748x_state *state, struct adv748x_csi2 *tx)
tx                282 drivers/media/i2c/adv748x/adv748x-csi2.c 	if (!is_tx_enabled(tx))
tx                286 drivers/media/i2c/adv748x/adv748x-csi2.c 	adv748x_csi2_set_virtual_channel(tx, 0);
tx                288 drivers/media/i2c/adv748x/adv748x-csi2.c 	adv748x_subdev_init(&tx->sd, state, &adv748x_csi2_ops,
tx                290 drivers/media/i2c/adv748x/adv748x-csi2.c 			    is_txa(tx) ? "txa" : "txb");
tx                293 drivers/media/i2c/adv748x/adv748x-csi2.c 	tx->sd.fwnode = of_fwnode_handle(state->endpoints[tx->port]);
tx                296 drivers/media/i2c/adv748x/adv748x-csi2.c 	tx->sd.internal_ops = &adv748x_csi2_internal_ops;
tx                298 drivers/media/i2c/adv748x/adv748x-csi2.c 	tx->pads[ADV748X_CSI2_SINK].flags = MEDIA_PAD_FL_SINK;
tx                299 drivers/media/i2c/adv748x/adv748x-csi2.c 	tx->pads[ADV748X_CSI2_SOURCE].flags = MEDIA_PAD_FL_SOURCE;
tx                301 drivers/media/i2c/adv748x/adv748x-csi2.c 	ret = media_entity_pads_init(&tx->sd.entity, ADV748X_CSI2_NR_PADS,
tx                302 drivers/media/i2c/adv748x/adv748x-csi2.c 				     tx->pads);
tx                306 drivers/media/i2c/adv748x/adv748x-csi2.c 	ret = adv748x_csi2_init_controls(tx);
tx                310 drivers/media/i2c/adv748x/adv748x-csi2.c 	ret = v4l2_async_register_subdev(&tx->sd);
tx                317 drivers/media/i2c/adv748x/adv748x-csi2.c 	v4l2_ctrl_handler_free(&tx->ctrl_hdl);
tx                319 drivers/media/i2c/adv748x/adv748x-csi2.c 	media_entity_cleanup(&tx->sd.entity);
tx                324 drivers/media/i2c/adv748x/adv748x-csi2.c void adv748x_csi2_cleanup(struct adv748x_csi2 *tx)
tx                326 drivers/media/i2c/adv748x/adv748x-csi2.c 	if (!is_tx_enabled(tx))
tx                329 drivers/media/i2c/adv748x/adv748x-csi2.c 	v4l2_async_unregister_subdev(&tx->sd);
tx                330 drivers/media/i2c/adv748x/adv748x-csi2.c 	media_entity_cleanup(&tx->sd.entity);
tx                331 drivers/media/i2c/adv748x/adv748x-csi2.c 	v4l2_ctrl_handler_free(&tx->ctrl_hdl);
tx                361 drivers/media/i2c/adv748x/adv748x-hdmi.c 	ret = adv748x_tx_power(hdmi->tx, enable);
tx                399 drivers/media/i2c/adv748x/adv748x-hdmi.c 	struct v4l2_subdev *tx;
tx                402 drivers/media/i2c/adv748x/adv748x-hdmi.c 	tx = adv748x_get_remote_sd(&hdmi->pads[ADV748X_HDMI_SOURCE]);
tx                403 drivers/media/i2c/adv748x/adv748x-hdmi.c 	if (!tx)
tx                408 drivers/media/i2c/adv748x/adv748x-hdmi.c 	return adv748x_csi2_set_pixelrate(tx, timings.bt.pixelclock);
tx                124 drivers/media/i2c/adv748x/adv748x.h 	struct adv748x_csi2 *tx;
tx                156 drivers/media/i2c/adv748x/adv748x.h 	struct adv748x_csi2 *tx;
tx                433 drivers/media/i2c/adv748x/adv748x.h int adv748x_tx_power(struct adv748x_csi2 *tx, bool on);
tx                438 drivers/media/i2c/adv748x/adv748x.h int adv748x_csi2_init(struct adv748x_state *state, struct adv748x_csi2 *tx);
tx                439 drivers/media/i2c/adv748x/adv748x.h void adv748x_csi2_cleanup(struct adv748x_csi2 *tx);
tx                 93 drivers/media/i2c/st-mipid02.c 	struct v4l2_fwnode_endpoint tx;
tx                489 drivers/media/i2c/st-mipid02.c 	struct v4l2_fwnode_endpoint *ep = &bridge->tx;
tx                930 drivers/media/i2c/st-mipid02.c 	bridge->tx = ep;
tx                277 drivers/media/platform/exynos-gsc/gsc-core.c 				 u32 tx, u32 ty, u32 *hratio)
tx                295 drivers/media/platform/exynos-gsc/gsc-core.c 		gsc_cal_prescaler_ratio(var, s_frame->crop.width, tx, wratio);
tx                641 drivers/media/platform/exynos-gsc/gsc-core.c 	int tx, ty;
tx                655 drivers/media/platform/exynos-gsc/gsc-core.c 		tx = d_frame->crop.height;
tx                657 drivers/media/platform/exynos-gsc/gsc-core.c 		tx = d_frame->crop.width;
tx                661 drivers/media/platform/exynos-gsc/gsc-core.c 	if (tx <= 0 || ty <= 0) {
tx                662 drivers/media/platform/exynos-gsc/gsc-core.c 		dev_err(dev, "Invalid target size: %dx%d", tx, ty);
tx                667 drivers/media/platform/exynos-gsc/gsc-core.c 				      tx, &sc->pre_hratio);
tx                681 drivers/media/platform/exynos-gsc/gsc-core.c 				 tx, ty, &sc->pre_vratio);
tx                686 drivers/media/platform/exynos-gsc/gsc-core.c 	sc->main_hratio = (s_frame->crop.width << 16) / tx;
tx                690 drivers/media/platform/exynos-gsc/gsc-core.c 			s_frame->crop.width, s_frame->crop.height, tx, ty);
tx                397 drivers/media/platform/exynos-gsc/gsc-core.h 				u32 *wratio, u32 tx, u32 ty, u32 *hratio);
tx                232 drivers/media/platform/exynos4-is/fimc-core.c 	int tx, ty, sx, sy;
tx                237 drivers/media/platform/exynos4-is/fimc-core.c 		tx = d_frame->height;
tx                239 drivers/media/platform/exynos4-is/fimc-core.c 		tx = d_frame->width;
tx                242 drivers/media/platform/exynos4-is/fimc-core.c 	if (tx <= 0 || ty <= 0) {
tx                243 drivers/media/platform/exynos4-is/fimc-core.c 		dev_err(dev, "Invalid target size: %dx%d\n", tx, ty);
tx                256 drivers/media/platform/exynos4-is/fimc-core.c 	ret = fimc_get_scaler_factor(sx, tx, &sc->pre_hratio, &sc->hfactor);
tx                268 drivers/media/platform/exynos4-is/fimc-core.c 		sc->main_hratio = (sx << 14) / (tx << sc->hfactor);
tx                271 drivers/media/platform/exynos4-is/fimc-core.c 		sc->main_hratio = (sx << 8) / (tx << sc->hfactor);
tx                276 drivers/media/platform/exynos4-is/fimc-core.c 	sc->scaleup_h = (tx >= sx) ? 1 : 0;
tx                210 drivers/media/platform/m2m-deinterlace.c 	struct dma_async_tx_descriptor *tx;
tx                328 drivers/media/platform/m2m-deinterlace.c 	tx = dmadev->device_prep_interleaved_dma(chan, ctx->xt, flags);
tx                329 drivers/media/platform/m2m-deinterlace.c 	if (tx == NULL) {
tx                335 drivers/media/platform/m2m-deinterlace.c 		tx->callback = dma_callback;
tx                336 drivers/media/platform/m2m-deinterlace.c 		tx->callback_param = ctx;
tx                339 drivers/media/platform/m2m-deinterlace.c 	ctx->cookie = dmaengine_submit(tx);
tx                233 drivers/media/platform/omap/omap_vout_vrfb.c 	struct dma_async_tx_descriptor *tx;
tx                270 drivers/media/platform/omap/omap_vout_vrfb.c 	tx = dmaengine_prep_interleaved_dma(chan, xt, flags);
tx                271 drivers/media/platform/omap/omap_vout_vrfb.c 	if (tx == NULL) {
tx                276 drivers/media/platform/omap/omap_vout_vrfb.c 	tx->callback = omap_vout_vrfb_dma_tx_callback;
tx                277 drivers/media/platform/omap/omap_vout_vrfb.c 	tx->callback_param = &vout->vrfb_dma_tx;
tx                279 drivers/media/platform/omap/omap_vout_vrfb.c 	cookie = dmaengine_submit(tx);
tx                176 drivers/media/platform/omap3isp/isphist.c 	struct dma_async_tx_descriptor *tx;
tx                203 drivers/media/platform/omap3isp/isphist.c 	tx = dmaengine_prep_slave_single(hist->dma_ch, dma_addr,
tx                206 drivers/media/platform/omap3isp/isphist.c 	if (tx == NULL) {
tx                212 drivers/media/platform/omap3isp/isphist.c 	tx->callback = hist_dma_cb;
tx                213 drivers/media/platform/omap3isp/isphist.c 	tx->callback_param = hist;
tx                214 drivers/media/platform/omap3isp/isphist.c 	cookie = tx->tx_submit(tx);
tx                864 drivers/media/platform/pxa_camera.c 	struct dma_async_tx_descriptor *tx;
tx                866 drivers/media/platform/pxa_camera.c 	tx = dmaengine_prep_slave_sg(dma_chan, sg, sglen, DMA_DEV_TO_MEM,
tx                868 drivers/media/platform/pxa_camera.c 	if (!tx) {
tx                874 drivers/media/platform/pxa_camera.c 	tx->callback_param = pcdev;
tx                877 drivers/media/platform/pxa_camera.c 		tx->callback = pxa_camera_dma_irq_y;
tx                880 drivers/media/platform/pxa_camera.c 		tx->callback = pxa_camera_dma_irq_u;
tx                883 drivers/media/platform/pxa_camera.c 		tx->callback = pxa_camera_dma_irq_v;
tx                887 drivers/media/platform/pxa_camera.c 	buf->descs[channel] = tx;
tx                892 drivers/media/platform/pxa_camera.c 		__func__, buf, tx);
tx                 93 drivers/media/platform/s5p-cec/s5p_cec.c 			cec->tx = STATE_NACK;
tx                 96 drivers/media/platform/s5p-cec/s5p_cec.c 			cec->tx = STATE_ERROR;
tx                 99 drivers/media/platform/s5p-cec/s5p_cec.c 			cec->tx = STATE_DONE;
tx                132 drivers/media/platform/s5p-cec/s5p_cec.c 	switch (cec->tx) {
tx                135 drivers/media/platform/s5p-cec/s5p_cec.c 		cec->tx = STATE_IDLE;
tx                141 drivers/media/platform/s5p-cec/s5p_cec.c 		cec->tx = STATE_IDLE;
tx                147 drivers/media/platform/s5p-cec/s5p_cec.c 		cec->tx = STATE_IDLE;
tx                 72 drivers/media/platform/s5p-cec/s5p_cec.h 	enum cec_state		tx;
tx                 57 drivers/media/radio/radio-keene.c 	u8 tx;
tx                109 drivers/media/radio/radio-keene.c 	radio->buffer[2] = radio->tx;
tx                253 drivers/media/radio/radio-keene.c 		radio->tx = db2tx[(ctrl->val - (s32)ctrl->minimum) / (s32)ctrl->step];
tx                336 drivers/media/radio/radio-keene.c 	radio->tx = 0x32;
tx                225 drivers/media/radio/wl128x/fmdrv_tx.c 	struct fmtx_data *tx = &fmdev->tx_data;
tx                230 drivers/media/radio/wl128x/fmdrv_tx.c 	payload = tx->audio_io;
tx                243 drivers/media/radio/wl128x/fmdrv_tx.c 	struct fmtx_data *tx = &fmdev->tx_data;
tx                273 drivers/media/radio/wl128x/fmdrv_tx.c 	tx->xmit_state = new_xmit_state;
tx                282 drivers/media/radio/wl128x/fmdrv_tx.c 	struct fmtx_data *tx = &fmdev->tx_data;
tx                291 drivers/media/radio/wl128x/fmdrv_tx.c 		tx->pwr_lvl = new_pwr_lvl;
tx                309 drivers/media/radio/wl128x/fmdrv_tx.c 	tx->pwr_lvl = new_pwr_lvl;
tx                320 drivers/media/radio/wl128x/fmdrv_tx.c 	struct fmtx_data *tx = &fmdev->tx_data;
tx                344 drivers/media/radio/wl128x/fmdrv_tx.c 	tx->preemph = payload;
tx                372 drivers/media/radio/wl128x/fmdrv_tx.c 	struct fmtx_data *tx = &fmdev->tx_data;
tx                388 drivers/media/radio/wl128x/fmdrv_tx.c 	tx->tx_frq = (unsigned long)freq_to_set;
tx                389 drivers/media/radio/wl128x/fmdrv_tx.c 	fmdbg("tx: freq_to_set %ld\n", (long int)tx->tx_frq);
tx                400 drivers/media/radio/wl128x/fmdrv_tx.c 	fm_tx_set_pwr_lvl(fmdev, tx->pwr_lvl);
tx                401 drivers/media/radio/wl128x/fmdrv_tx.c 	fm_tx_set_preemph_filter(fmdev, tx->preemph);
tx                403 drivers/media/radio/wl128x/fmdrv_tx.c 	tx->audio_io = 0x01;	/* I2S */
tx                408 drivers/media/radio/wl128x/fmdrv_tx.c 	tx->aud_mode = FM_STEREO_MODE;
tx                409 drivers/media/radio/wl128x/fmdrv_tx.c 	tx->rds.flag = FM_RDS_DISABLE;
tx                 62 drivers/media/rc/fintek-cir.h 	} tx;
tx                121 drivers/media/rc/imon.c 	} tx;
tx                596 drivers/media/rc/imon.c 	reinit_completion(&ictx->tx.finished);
tx                597 drivers/media/rc/imon.c 	ictx->tx.busy = true;
tx                602 drivers/media/rc/imon.c 		ictx->tx.busy = false;
tx                609 drivers/media/rc/imon.c 				&ictx->tx.finished);
tx                616 drivers/media/rc/imon.c 		retval = ictx->tx.status;
tx                936 drivers/media/rc/imon.c 	if (copy_from_user(ictx->tx.data_buf, buf, n_bytes)) {
tx                943 drivers/media/rc/imon.c 		ictx->tx.data_buf[i] = ' ';
tx                946 drivers/media/rc/imon.c 		ictx->tx.data_buf[i] = 0xFF;
tx                952 drivers/media/rc/imon.c 		memcpy(ictx->usb_tx_buf, ictx->tx.data_buf + offset, 7);
tx               1050 drivers/media/rc/imon.c 	ictx->tx.status = urb->status;
tx               1053 drivers/media/rc/imon.c 	ictx->tx.busy = false;
tx               1055 drivers/media/rc/imon.c 	complete(&ictx->tx.finished);
tx               2200 drivers/media/rc/imon.c 	init_completion(&ictx->tx.finished);
tx               2478 drivers/media/rc/imon.c 	if (ictx->tx.busy) {
tx               2480 drivers/media/rc/imon.c 		complete(&ictx->tx.finished);
tx                 65 drivers/media/spi/cxd2880-spi.c 	struct spi_transfer tx = {};
tx                 72 drivers/media/spi/cxd2880-spi.c 	tx.tx_buf = data;
tx                 73 drivers/media/spi/cxd2880-spi.c 	tx.len = size;
tx                 76 drivers/media/spi/cxd2880-spi.c 	spi_message_add_tail(&tx, &msg);
tx                146 drivers/media/spi/gs1662.c 	struct spi_transfer tx[] = {
tx                159 drivers/media/spi/gs1662.c 	spi_message_add_tail(&tx[0], &msg);
tx                160 drivers/media/spi/gs1662.c 	spi_message_add_tail(&tx[1], &msg);
tx                174 drivers/media/spi/gs1662.c 	struct spi_transfer tx[] = {
tx                187 drivers/media/spi/gs1662.c 	spi_message_add_tail(&tx[0], &msg);
tx                188 drivers/media/spi/gs1662.c 	spi_message_add_tail(&tx[1], &msg);
tx                 59 drivers/media/usb/dvb-usb/dib0700.h extern int dib0700_ctrl_rd(struct dvb_usb_device *d, u8 *tx, u8 txlen, u8 *rx, u8 rxlen);
tx                 53 drivers/media/usb/dvb-usb/dib0700_core.c static int dib0700_ctrl_wr(struct dvb_usb_device *d, u8 *tx, u8 txlen)
tx                 58 drivers/media/usb/dvb-usb/dib0700_core.c 	debug_dump(tx, txlen, deb_data);
tx                 61 drivers/media/usb/dvb-usb/dib0700_core.c 		tx[0], USB_TYPE_VENDOR | USB_DIR_OUT, 0, 0, tx, txlen,
tx                 71 drivers/media/usb/dvb-usb/dib0700_core.c int dib0700_ctrl_rd(struct dvb_usb_device *d, u8 *tx, u8 txlen, u8 *rx, u8 rxlen)
tx                 86 drivers/media/usb/dvb-usb/dib0700_core.c 	debug_dump(tx,txlen,deb_data);
tx                 88 drivers/media/usb/dvb-usb/dib0700_core.c 	value = ((txlen - 2) << 8) | tx[1];
tx                 91 drivers/media/usb/dvb-usb/dib0700_core.c 		index |= (tx[2] << 8);
tx                 93 drivers/media/usb/dvb-usb/dib0700_core.c 		index |= tx[3];
tx                 95 drivers/media/usb/dvb-usb/dib0700_core.c 	status = usb_control_msg(d->udev, usb_rcvctrlpipe(d->udev,0), tx[0],
tx                 85 drivers/media/usb/dvb-usb/technisat-usb2.c 		u8 device_addr, u8 *tx, u8 txlen, u8 *rx, u8 rxlen)
tx                 95 drivers/media/usb/dvb-usb/technisat-usb2.c 	debug_dump(tx, txlen, deb_i2c);
tx                117 drivers/media/usb/dvb-usb/technisat-usb2.c 	memcpy(&b[2], tx, txlen);
tx                242 drivers/media/usb/hackrf/hackrf.c 	const bool tx = test_bit(TX_ON, &dev->flags);
tx                264 drivers/media/usb/hackrf/hackrf.c 	if (!rx && !tx) {
tx                275 drivers/media/usb/hackrf/hackrf.c 	} else if (tx && test_and_clear_bit(TX_DAC_FREQUENCY, &dev->flags)) {
tx                314 drivers/media/usb/hackrf/hackrf.c 	} else if (tx && test_and_clear_bit(TX_BANDWIDTH, &dev->flags)) {
tx                351 drivers/media/usb/hackrf/hackrf.c 	} else if (tx && test_and_clear_bit(TX_RF_FREQUENCY, &dev->flags)) {
tx                386 drivers/media/usb/hackrf/hackrf.c 	if (tx && test_and_clear_bit(TX_RF_GAIN, &dev->flags)) {
tx                420 drivers/media/usb/hackrf/hackrf.c 	if (tx && test_and_clear_bit(TX_LNA_GAIN, &dev->flags)) {
tx                 33 drivers/mfd/ipaq-micro.c 	struct ipaq_micro_txdev *tx = &micro->tx;
tx                 40 drivers/mfd/ipaq-micro.c 	tx->buf[bp++] = CHAR_SOF;
tx                 43 drivers/mfd/ipaq-micro.c 	tx->buf[bp++] = checksum;
tx                 46 drivers/mfd/ipaq-micro.c 		tx->buf[bp++] = msg->tx_data[i];
tx                 50 drivers/mfd/ipaq-micro.c 	tx->buf[bp++] = checksum;
tx                 51 drivers/mfd/ipaq-micro.c 	tx->len = bp;
tx                 52 drivers/mfd/ipaq-micro.c 	tx->index = 0;
tx                283 drivers/mfd/ipaq-micro.c 	struct ipaq_micro_txdev *tx = &micro->tx;
tx                286 drivers/mfd/ipaq-micro.c 	while ((tx->index < tx->len) &&
tx                288 drivers/mfd/ipaq-micro.c 		writel(tx->buf[tx->index], micro->base + UTDR);
tx                289 drivers/mfd/ipaq-micro.c 		tx->index++;
tx                335 drivers/mfd/ipaq-micro.c 	struct ipaq_micro_txdev *tx = &micro->tx;
tx                357 drivers/mfd/ipaq-micro.c 	} while (((tx->index < tx->len) && (status & UTSR0_TFS)) ||
tx                262 drivers/misc/echo/echo.c int16_t oslec_update(struct oslec_state *ec, int16_t tx, int16_t rx)
tx                275 drivers/misc/echo/echo.c 	ec->tx = tx;
tx                277 drivers/misc/echo/echo.c 	tx >>= 1;
tx                333 drivers/misc/echo/echo.c 		new = (int)tx * (int)tx;
tx                344 drivers/misc/echo/echo.c 	ec->ltxacc += abs(tx) - ec->ltx;
tx                352 drivers/misc/echo/echo.c 	echo_value = fir16(&ec->fir_state, tx);
tx                359 drivers/misc/echo/echo.c 	echo_value = fir16(&ec->fir_state_bg, tx);
tx                554 drivers/misc/echo/echo.c int16_t oslec_hpf_tx(struct oslec_state *ec, int16_t tx)
tx                560 drivers/misc/echo/echo.c 		tmp = tx << 15;
tx                578 drivers/misc/echo/echo.c 		tx = tmp1;
tx                582 drivers/misc/echo/echo.c 	return tx;
tx                118 drivers/misc/echo/echo.h 	int16_t tx;
tx                 70 drivers/misc/echo/oslec.h int16_t oslec_update(struct oslec_state *ec, int16_t tx, int16_t rx);
tx                 79 drivers/misc/echo/oslec.h int16_t oslec_hpf_tx(struct oslec_state *ec, int16_t tx);
tx                413 drivers/misc/mic/scif/scif_dma.c 	struct dma_async_tx_descriptor *tx = NULL;
tx                426 drivers/misc/mic/scif/scif_dma.c 	tx = ddev->device_prep_dma_memcpy(chan, 0, 0, 0, flags);
tx                427 drivers/misc/mic/scif/scif_dma.c 	if (!tx) {
tx                433 drivers/misc/mic/scif/scif_dma.c 	cookie = tx->tx_submit(tx);
tx                479 drivers/misc/mic/scif/scif_dma.c 	struct dma_async_tx_descriptor *tx = NULL;
tx                493 drivers/misc/mic/scif/scif_dma.c 	tx = ddev->device_prep_dma_memcpy(chan, 0, 0, 0, flags);
tx                494 drivers/misc/mic/scif/scif_dma.c 	if (!tx) {
tx                501 drivers/misc/mic/scif/scif_dma.c 	tx->callback = scif_dma_callback;
tx                502 drivers/misc/mic/scif/scif_dma.c 	tx->callback_param = &done_wait;
tx                503 drivers/misc/mic/scif/scif_dma.c 	cookie = tx->tx_submit(tx);
tx                880 drivers/misc/mic/scif/scif_dma.c 	struct dma_async_tx_descriptor *tx;
tx                954 drivers/misc/mic/scif/scif_dma.c 				tx =
tx                961 drivers/misc/mic/scif/scif_dma.c 				if (!tx) {
tx                965 drivers/misc/mic/scif/scif_dma.c 				cookie = tx->tx_submit(tx);
tx                978 drivers/misc/mic/scif/scif_dma.c 				tx =
tx                983 drivers/misc/mic/scif/scif_dma.c 				if (!tx) {
tx                987 drivers/misc/mic/scif/scif_dma.c 				cookie = tx->tx_submit(tx);
tx                994 drivers/misc/mic/scif/scif_dma.c 				tx =
tx                999 drivers/misc/mic/scif/scif_dma.c 				if (!tx) {
tx               1003 drivers/misc/mic/scif/scif_dma.c 				cookie = tx->tx_submit(tx);
tx               1011 drivers/misc/mic/scif/scif_dma.c 			tx = dev->device_prep_dma_memcpy(chan, temp_dma_addr,
tx               1013 drivers/misc/mic/scif/scif_dma.c 			if (!tx) {
tx               1017 drivers/misc/mic/scif/scif_dma.c 			cookie = tx->tx_submit(tx);
tx               1061 drivers/misc/mic/scif/scif_dma.c 	tx = dev->device_prep_dma_memcpy(chan, 0, 0, 0, DMA_PREP_INTERRUPT);
tx               1062 drivers/misc/mic/scif/scif_dma.c 	if (!tx) {
tx               1066 drivers/misc/mic/scif/scif_dma.c 	tx->callback = &scif_rma_completion_cb;
tx               1067 drivers/misc/mic/scif/scif_dma.c 	tx->callback_param = comp_cb;
tx               1068 drivers/misc/mic/scif/scif_dma.c 	cookie = tx->tx_submit(tx);
tx               1101 drivers/misc/mic/scif/scif_dma.c 	struct dma_async_tx_descriptor *tx;
tx               1143 drivers/misc/mic/scif/scif_dma.c 			tx = dev->device_prep_dma_memcpy(chan, dst_dma_addr,
tx               1147 drivers/misc/mic/scif/scif_dma.c 			if (!tx) {
tx               1151 drivers/misc/mic/scif/scif_dma.c 			cookie = tx->tx_submit(tx);
tx               1164 drivers/misc/mic/scif/scif_dma.c 			tx = dev->device_prep_dma_memcpy(chan, dst_dma_addr,
tx               1166 drivers/misc/mic/scif/scif_dma.c 			if (!tx) {
tx               1170 drivers/misc/mic/scif/scif_dma.c 			cookie = tx->tx_submit(tx);
tx               1177 drivers/misc/mic/scif/scif_dma.c 			tx = dev->device_prep_dma_memcpy(chan, dst_dma_addr,
tx               1179 drivers/misc/mic/scif/scif_dma.c 			if (!tx) {
tx               1183 drivers/misc/mic/scif/scif_dma.c 			cookie = tx->tx_submit(tx);
tx               1221 drivers/misc/mic/scif/scif_dma.c 	struct dma_async_tx_descriptor *tx;
tx               1314 drivers/misc/mic/scif/scif_dma.c 			tx = dev->device_prep_dma_memcpy(chan, dst_dma_addr,
tx               1319 drivers/misc/mic/scif/scif_dma.c 			if (!tx) {
tx               1323 drivers/misc/mic/scif/scif_dma.c 			cookie = tx->tx_submit(tx);
tx               1337 drivers/misc/mic/scif/scif_dma.c 			tx = dev->device_prep_dma_memcpy(chan, dst_dma_addr,
tx               1340 drivers/misc/mic/scif/scif_dma.c 			if (!tx) {
tx               1344 drivers/misc/mic/scif/scif_dma.c 			cookie = tx->tx_submit(tx);
tx               1351 drivers/misc/mic/scif/scif_dma.c 			tx = dev->device_prep_dma_memcpy(chan, dst_dma_addr,
tx               1354 drivers/misc/mic/scif/scif_dma.c 			if (!tx) {
tx               1358 drivers/misc/mic/scif/scif_dma.c 			cookie = tx->tx_submit(tx);
tx                202 drivers/misc/mic/scif/scif_fence.c 	struct dma_async_tx_descriptor *tx;
tx                209 drivers/misc/mic/scif/scif_fence.c 	tx = ddev->device_prep_dma_memcpy(chan, 0, 0, 0, DMA_PREP_FENCE);
tx                210 drivers/misc/mic/scif/scif_fence.c 	if (!tx) {
tx                216 drivers/misc/mic/scif/scif_fence.c 	cookie = tx->tx_submit(tx);
tx                229 drivers/misc/mic/scif/scif_fence.c 		tx = ddev->device_prep_dma_imm_data(chan, dst, val, 0);
tx                243 drivers/misc/mic/scif/scif_fence.c 		tx = ddev->device_prep_dma_memcpy(chan, dst, src, sizeof(val),
tx                246 drivers/misc/mic/scif/scif_fence.c 	if (!tx) {
tx                261 drivers/misc/mic/scif/scif_fence.c 		tx->callback = scif_prog_signal_cb;
tx                262 drivers/misc/mic/scif/scif_fence.c 		tx->callback_param = cb_arg;
tx                264 drivers/misc/mic/scif/scif_fence.c 	cookie = tx->tx_submit(tx);
tx                591 drivers/misc/mic/scif/scif_fence.c 	struct dma_async_tx_descriptor *tx;
tx                595 drivers/misc/mic/scif/scif_fence.c 	tx = ddev->device_prep_dma_memcpy(chan, 0, 0, 0, DMA_PREP_FENCE);
tx                596 drivers/misc/mic/scif/scif_fence.c 	if (!tx) {
tx                602 drivers/misc/mic/scif/scif_fence.c 	cookie = tx->tx_submit(tx);
tx                610 drivers/misc/mic/scif/scif_fence.c 	tx = ddev->device_prep_dma_interrupt(chan, DMA_PREP_INTERRUPT);
tx                611 drivers/misc/mic/scif/scif_fence.c 	if (!tx) {
tx                617 drivers/misc/mic/scif/scif_fence.c 	tx->callback = scif_fence_mark_cb;
tx                618 drivers/misc/mic/scif/scif_fence.c 	tx->callback_param = ep;
tx                619 drivers/misc/mic/scif/scif_fence.c 	*mark = cookie = tx->tx_submit(tx);
tx                474 drivers/misc/mic/vop/vop_vringh.c 	struct dma_async_tx_descriptor *tx;
tx                483 drivers/misc/mic/vop/vop_vringh.c 	tx = ddev->device_prep_dma_memcpy(vop_ch, dst, src, len,
tx                485 drivers/misc/mic/vop/vop_vringh.c 	if (!tx) {
tx                491 drivers/misc/mic/vop/vop_vringh.c 		cookie = tx->tx_submit(tx);
tx                 56 drivers/mmc/core/sdio_uart.c 	__u32	tx;
tx                434 drivers/mmc/core/sdio_uart.c 		port->icount.tx++;
tx                451 drivers/mmc/core/sdio_uart.c 		port->icount.tx++;
tx                974 drivers/mmc/core/sdio_uart.c 					      port->icount.tx, port->icount.rx);
tx                983 drivers/mmc/host/omap.c 		struct dma_async_tx_descriptor *tx;
tx               1039 drivers/mmc/host/omap.c 		tx = dmaengine_prep_slave_sg(c, data->sg, host->sg_len,
tx               1042 drivers/mmc/host/omap.c 		if (!tx)
tx               1047 drivers/mmc/host/omap.c 		tx->callback = mmc_omap_dma_callback;
tx               1048 drivers/mmc/host/omap.c 		tx->callback_param = host;
tx               1049 drivers/mmc/host/omap.c 		dmaengine_submit(tx);
tx               1242 drivers/mmc/host/omap_hsmmc.c 	struct dma_async_tx_descriptor *tx;
tx               1281 drivers/mmc/host/omap_hsmmc.c 	tx = dmaengine_prep_slave_sg(chan, data->sg, data->sg_len,
tx               1284 drivers/mmc/host/omap_hsmmc.c 	if (!tx) {
tx               1290 drivers/mmc/host/omap_hsmmc.c 	tx->callback = omap_hsmmc_dma_callback;
tx               1291 drivers/mmc/host/omap_hsmmc.c 	tx->callback_param = host;
tx               1294 drivers/mmc/host/omap_hsmmc.c 	dmaengine_submit(tx);
tx                161 drivers/mmc/host/pxamci.c 	struct dma_async_tx_descriptor *tx;
tx                209 drivers/mmc/host/pxamci.c 	tx = dmaengine_prep_slave_sg(chan, data->sg, host->dma_len, direction,
tx                211 drivers/mmc/host/pxamci.c 	if (!tx) {
tx                217 drivers/mmc/host/pxamci.c 		tx->callback = pxamci_dma_irq;
tx                218 drivers/mmc/host/pxamci.c 		tx->callback_param = host;
tx                221 drivers/mmc/host/pxamci.c 	host->dma_cookie = dmaengine_submit(tx);
tx                579 drivers/mmc/host/rtsx_usb_sdmmc.c static int sd_change_phase(struct rtsx_usb_sdmmc *host, u8 sample_point, int tx)
tx                585 drivers/mmc/host/rtsx_usb_sdmmc.c 			__func__, tx ? "TX" : "RX", sample_point);
tx                591 drivers/mmc/host/rtsx_usb_sdmmc.c 	if (tx)
tx                328 drivers/mtd/nand/onenand/omap2.c 	struct dma_async_tx_descriptor *tx;
tx                331 drivers/mtd/nand/onenand/omap2.c 	tx = dmaengine_prep_dma_memcpy(c->dma_chan, dst, src, count,
tx                333 drivers/mtd/nand/onenand/omap2.c 	if (!tx) {
tx                340 drivers/mtd/nand/onenand/omap2.c 	tx->callback = omap2_onenand_dma_complete_func;
tx                341 drivers/mtd/nand/onenand/omap2.c 	tx->callback_param = &c->dma_done;
tx                343 drivers/mtd/nand/onenand/omap2.c 	cookie = tx->tx_submit(tx);
tx                369 drivers/mtd/nand/raw/atmel/nand-controller.c 	struct dma_async_tx_descriptor *tx;
tx                387 drivers/mtd/nand/raw/atmel/nand-controller.c 	tx = dmaengine_prep_dma_memcpy(nc->dmac, dst_dma, src_dma, len,
tx                389 drivers/mtd/nand/raw/atmel/nand-controller.c 	if (!tx) {
tx                394 drivers/mtd/nand/raw/atmel/nand-controller.c 	tx->callback = atmel_nand_dma_transfer_finished;
tx                395 drivers/mtd/nand/raw/atmel/nand-controller.c 	tx->callback_param = &finished;
tx                397 drivers/mtd/nand/raw/atmel/nand-controller.c 	cookie = dmaengine_submit(tx);
tx                462 drivers/mtd/nand/raw/fsmc_nand.c 	struct dma_async_tx_descriptor *tx;
tx                487 drivers/mtd/nand/raw/fsmc_nand.c 	tx = dma_dev->device_prep_dma_memcpy(chan, dma_dst, dma_src,
tx                489 drivers/mtd/nand/raw/fsmc_nand.c 	if (!tx) {
tx                495 drivers/mtd/nand/raw/fsmc_nand.c 	tx->callback = dma_complete;
tx                496 drivers/mtd/nand/raw/fsmc_nand.c 	tx->callback_param = host;
tx                497 drivers/mtd/nand/raw/fsmc_nand.c 	cookie = tx->tx_submit(tx);
tx                823 drivers/mtd/nand/raw/marvell_nand.c 	struct dma_async_tx_descriptor *tx;
tx                832 drivers/mtd/nand/raw/marvell_nand.c 	tx = dmaengine_prep_slave_sg(nfc->dma_chan, &sg, 1,
tx                836 drivers/mtd/nand/raw/marvell_nand.c 	if (!tx) {
tx                842 drivers/mtd/nand/raw/marvell_nand.c 	cookie = dmaengine_submit(tx);
tx               1331 drivers/mtd/nand/raw/nandsim.c 	ssize_t tx;
tx               1339 drivers/mtd/nand/raw/nandsim.c 	tx = kernel_read(file, buf, count, &pos);
tx               1342 drivers/mtd/nand/raw/nandsim.c 	return tx;
tx               1347 drivers/mtd/nand/raw/nandsim.c 	ssize_t tx;
tx               1355 drivers/mtd/nand/raw/nandsim.c 	tx = kernel_write(file, buf, count, &pos);
tx               1358 drivers/mtd/nand/raw/nandsim.c 	return tx;
tx               1419 drivers/mtd/nand/raw/nandsim.c 			ssize_t tx;
tx               1426 drivers/mtd/nand/raw/nandsim.c 			tx = read_file(ns, ns->cfile, ns->buf.byte, num, pos);
tx               1427 drivers/mtd/nand/raw/nandsim.c 			if (tx != num) {
tx               1428 drivers/mtd/nand/raw/nandsim.c 				NS_ERR("read_page: read error for page %d ret %ld\n", ns->regs.row, (long)tx);
tx               1489 drivers/mtd/nand/raw/nandsim.c 		ssize_t tx;
tx               1500 drivers/mtd/nand/raw/nandsim.c 			tx = read_file(ns, ns->cfile, pg_off, num, off);
tx               1501 drivers/mtd/nand/raw/nandsim.c 			if (tx != num) {
tx               1502 drivers/mtd/nand/raw/nandsim.c 				NS_ERR("prog_page: read error for page %d ret %ld\n", ns->regs.row, (long)tx);
tx               1510 drivers/mtd/nand/raw/nandsim.c 			tx = write_file(ns, ns->cfile, ns->file_buf, ns->geom.pgszoob, pos);
tx               1511 drivers/mtd/nand/raw/nandsim.c 			if (tx != ns->geom.pgszoob) {
tx               1512 drivers/mtd/nand/raw/nandsim.c 				NS_ERR("prog_page: write error for page %d ret %ld\n", ns->regs.row, (long)tx);
tx               1517 drivers/mtd/nand/raw/nandsim.c 			tx = write_file(ns, ns->cfile, pg_off, num, off);
tx               1518 drivers/mtd/nand/raw/nandsim.c 			if (tx != num) {
tx               1519 drivers/mtd/nand/raw/nandsim.c 				NS_ERR("prog_page: write error for page %d ret %ld\n", ns->regs.row, (long)tx);
tx                458 drivers/mtd/nand/raw/omap2.c 	struct dma_async_tx_descriptor *tx;
tx                478 drivers/mtd/nand/raw/omap2.c 	tx = dmaengine_prep_slave_sg(info->dma, &sg, n,
tx                481 drivers/mtd/nand/raw/omap2.c 	if (!tx)
tx                484 drivers/mtd/nand/raw/omap2.c 	tx->callback = omap_nand_dma_callback;
tx                485 drivers/mtd/nand/raw/omap2.c 	tx->callback_param = &info->comp;
tx                486 drivers/mtd/nand/raw/omap2.c 	dmaengine_submit(tx);
tx                971 drivers/mtd/spi-nor/cadence-quadspi.c 	struct dma_async_tx_descriptor *tx;
tx                985 drivers/mtd/spi-nor/cadence-quadspi.c 	tx = dmaengine_prep_dma_memcpy(cqspi->rx_chan, dma_dst, dma_src,
tx                987 drivers/mtd/spi-nor/cadence-quadspi.c 	if (!tx) {
tx                993 drivers/mtd/spi-nor/cadence-quadspi.c 	tx->callback = cqspi_rx_dma_callback;
tx                994 drivers/mtd/spi-nor/cadence-quadspi.c 	tx->callback_param = cqspi;
tx                995 drivers/mtd/spi-nor/cadence-quadspi.c 	cookie = tx->tx_submit(tx);
tx                154 drivers/mtd/spi-nor/mtk-quadspi.c 			    u8 *tx, int txlen, u8 *rx, int rxlen)
tx                173 drivers/mtd/spi-nor/mtk-quadspi.c 		writeb(tx[i], mtk_nor->base + MTK_NOR_PRG_REG(idx));
tx                223 drivers/net/can/grcan.c 	struct grcan_dma_buffer tx;
tx                527 drivers/net/can/grcan.c 					     dma->tx.size);
tx                558 drivers/net/can/grcan.c 		txrd = grcan_ring_add(txrd, GRCAN_MSG_SIZE, dma->tx.size);
tx                695 drivers/net/can/grcan.c 				if (grcan_txspace(dma->tx.size, txwr,
tx                863 drivers/net/can/grcan.c 		if (grcan_txspace(priv->dma.tx.size, txwr, priv->eskbp) &&
tx                937 drivers/net/can/grcan.c 	struct grcan_dma_buffer *large = rsize > tsize ? &dma->rx : &dma->tx;
tx                938 drivers/net/can/grcan.c 	struct grcan_dma_buffer *small = rsize > tsize ? &dma->tx : &dma->rx;
tx                960 drivers/net/can/grcan.c 	dma->tx.size = tsize;
tx                982 drivers/net/can/grcan.c 	grcan_write_reg(&regs->txaddr, priv->dma.tx.handle);
tx                983 drivers/net/can/grcan.c 	grcan_write_reg(&regs->txsize, priv->dma.tx.size);
tx               1056 drivers/net/can/grcan.c 	priv->echo_skb = kcalloc(dma->tx.size, sizeof(*priv->echo_skb),
tx               1062 drivers/net/can/grcan.c 	priv->can.echo_skb_max = dma->tx.size;
tx               1065 drivers/net/can/grcan.c 	priv->txdlc = kcalloc(dma->tx.size, sizeof(*priv->txdlc), GFP_KERNEL);
tx               1307 drivers/net/can/grcan.c 		if (grcan_txspace(dma->tx.size, txwr, priv->eskbp))
tx               1379 drivers/net/can/grcan.c 	space = grcan_txspace(dma->tx.size, txwr, priv->eskbp);
tx               1382 drivers/net/can/grcan.c 	slot = dma->tx.buf + txwr;
tx               1434 drivers/net/can/grcan.c 		if (unlikely(grcan_ring_sub(txwr, txrd, dma->tx.size) == 1)) {
tx               1460 drivers/net/can/grcan.c 			grcan_ring_add(txwr, GRCAN_MSG_SIZE, dma->tx.size));
tx                335 drivers/net/can/kvaser_pciefd.c static int kvaser_pciefd_spi_cmd(struct kvaser_pciefd *pcie, const u8 *tx,
tx                349 drivers/net/can/kvaser_pciefd.c 		iowrite32(*tx++, pcie->reg_base + KVASER_PCIEFD_SPI_TX_REG);
tx                234 drivers/net/can/mscan/mscan.c 		out_be16(&regs->tx.idr3_2, can_id);
tx                246 drivers/net/can/mscan/mscan.c 	out_be16(&regs->tx.idr1_0, can_id);
tx                249 drivers/net/can/mscan/mscan.c 		void __iomem *data = &regs->tx.dsr1_0;
tx                261 drivers/net/can/mscan/mscan.c 	out_8(&regs->tx.dlr, frame->can_dlc);
tx                262 drivers/net/can/mscan/mscan.c 	out_8(&regs->tx.tbpr, priv->cur_pri);
tx                448 drivers/net/can/mscan/mscan.c 			stats->tx_bytes += in_8(&regs->tx.dlr);
tx                217 drivers/net/can/mscan/mscan.h 	} tx;
tx                 29 drivers/net/can/softing/softing.h 	} tx;
tx                 64 drivers/net/can/softing/softing.h 	} tx;
tx                334 drivers/net/can/softing/softing_fw.c 	card->tx.pending = 0;
tx                455 drivers/net/can/softing/softing_fw.c 			priv->tx.pending = 0;
tx                456 drivers/net/can/softing/softing_fw.c 			priv->tx.echo_put = 0;
tx                457 drivers/net/can/softing/softing_fw.c 			priv->tx.echo_get = 0;
tx                470 drivers/net/can/softing/softing_fw.c 	card->tx.pending = 0;
tx                 69 drivers/net/can/softing/softing_main.c 			(card->tx.pending >= TXMAX) ||
tx                 70 drivers/net/can/softing/softing_main.c 			(priv->tx.pending >= TX_ECHO_SKB_MAX))
tx                104 drivers/net/can/softing/softing_main.c 	card->tx.last_bus = priv->index;
tx                105 drivers/net/can/softing/softing_main.c 	++card->tx.pending;
tx                106 drivers/net/can/softing/softing_main.c 	++priv->tx.pending;
tx                107 drivers/net/can/softing/softing_main.c 	can_put_echo_skb(skb, dev, priv->tx.echo_put);
tx                108 drivers/net/can/softing/softing_main.c 	++priv->tx.echo_put;
tx                109 drivers/net/can/softing/softing_main.c 	if (priv->tx.echo_put >= TX_ECHO_SKB_MAX)
tx                110 drivers/net/can/softing/softing_main.c 		priv->tx.echo_put = 0;
tx                115 drivers/net/can/softing/softing_main.c 	if (card->tx.pending >= TXMAX) {
tx                284 drivers/net/can/softing/softing_main.c 			skb = priv->can.echo_skb[priv->tx.echo_get];
tx                287 drivers/net/can/softing/softing_main.c 			can_get_echo_skb(netdev, priv->tx.echo_get);
tx                288 drivers/net/can/softing/softing_main.c 			++priv->tx.echo_get;
tx                289 drivers/net/can/softing/softing_main.c 			if (priv->tx.echo_get >= TX_ECHO_SKB_MAX)
tx                290 drivers/net/can/softing/softing_main.c 				priv->tx.echo_get = 0;
tx                291 drivers/net/can/softing/softing_main.c 			if (priv->tx.pending)
tx                292 drivers/net/can/softing/softing_main.c 				--priv->tx.pending;
tx                293 drivers/net/can/softing/softing_main.c 			if (card->tx.pending)
tx                294 drivers/net/can/softing/softing_main.c 				--card->tx.pending;
tx                333 drivers/net/can/softing/softing_main.c 	offset = card->tx.last_bus;
tx                335 drivers/net/can/softing/softing_main.c 		if (card->tx.pending >= TXMAX)
tx                344 drivers/net/can/softing/softing_main.c 		if (priv->tx.pending >= TX_ECHO_SKB_MAX)
tx                186 drivers/net/can/usb/ems_usb.c 	u8 tx;
tx                167 drivers/net/can/usb/esd_usb2.c 		struct tx_msg tx;
tx                739 drivers/net/can/usb/esd_usb2.c 	msg->msg.tx.net = priv->index;
tx                740 drivers/net/can/usb/esd_usb2.c 	msg->msg.tx.dlc = cf->can_dlc;
tx                741 drivers/net/can/usb/esd_usb2.c 	msg->msg.tx.id = cpu_to_le32(cf->can_id & CAN_ERR_MASK);
tx                744 drivers/net/can/usb/esd_usb2.c 		msg->msg.tx.dlc |= ESD_RTR;
tx                747 drivers/net/can/usb/esd_usb2.c 		msg->msg.tx.id |= cpu_to_le32(ESD_EXTID);
tx                750 drivers/net/can/usb/esd_usb2.c 		msg->msg.tx.data[i] = cf->data[i];
tx                775 drivers/net/can/usb/esd_usb2.c 	msg->msg.tx.hnd = 0x80000000 | i; /* returned in TX done message */
tx                 18 drivers/net/dsa/sja1105/sja1105_spi.c 				const void *tx, void *rx, int size)
tx                 22 drivers/net/dsa/sja1105/sja1105_spi.c 		.tx_buf = tx,
tx                910 drivers/net/dsa/vitesse-vsc73xx-core.c 		     bool tx)
tx                916 drivers/net/dsa/vitesse-vsc73xx-core.c 	if (tx) {
tx                954 drivers/net/ethernet/3com/3c574_cs.c 	u8 rx, tx, up;
tx                976 drivers/net/ethernet/3com/3c574_cs.c 	tx					 = inw(ioaddr + 12);
tx               1511 drivers/net/ethernet/3com/typhoon.c 	struct tx_desc *tx;
tx               1517 drivers/net/ethernet/3com/typhoon.c 		tx = (struct tx_desc *) (txRing->ringBase + lastRead);
tx               1518 drivers/net/ethernet/3com/typhoon.c 		type = tx->flags & TYPHOON_TYPE_MASK;
tx               1523 drivers/net/ethernet/3com/typhoon.c 			unsigned long ptr = tx->tx_addr;
tx               1529 drivers/net/ethernet/3com/typhoon.c 			skb_dma = (dma_addr_t) le32_to_cpu(tx->frag.addr);
tx               1530 drivers/net/ethernet/3com/typhoon.c 			dma_len = le16_to_cpu(tx->len);
tx               1535 drivers/net/ethernet/3com/typhoon.c 		tx->flags = 0;
tx               2332 drivers/net/ethernet/alteon/acenic.c 				struct tx_desc __iomem *tx;
tx               2333 drivers/net/ethernet/alteon/acenic.c 				tx = (__force struct tx_desc __iomem *) &ap->tx_ring[i];
tx               2334 drivers/net/ethernet/alteon/acenic.c 				writel(0, &tx->addr.addrhi);
tx               2335 drivers/net/ethernet/alteon/acenic.c 				writel(0, &tx->addr.addrlo);
tx               2336 drivers/net/ethernet/alteon/acenic.c 				writel(0, &tx->flagsize);
tx                654 drivers/net/ethernet/amazon/ena/ena_admin_defs.h 	u32 tx;
tx                 56 drivers/net/ethernet/amazon/ena/ena_ethtool.c 	ENA_STAT_ENTRY(stat, tx)
tx               3238 drivers/net/ethernet/amazon/ena/ena_netdev.c 	if (feat->offload.tx &
tx               3242 drivers/net/ethernet/amazon/ena/ena_netdev.c 	if (feat->offload.tx &
tx               3246 drivers/net/ethernet/amazon/ena/ena_netdev.c 	if (feat->offload.tx & ENA_ADMIN_FEATURE_OFFLOAD_DESC_TSO_IPV4_MASK)
tx               3249 drivers/net/ethernet/amazon/ena/ena_netdev.c 	if (feat->offload.tx & ENA_ADMIN_FEATURE_OFFLOAD_DESC_TSO_IPV6_MASK)
tx               3252 drivers/net/ethernet/amazon/ena/ena_netdev.c 	if (feat->offload.tx & ENA_ADMIN_FEATURE_OFFLOAD_DESC_TSO_ECN_MASK)
tx                262 drivers/net/ethernet/amd/sunlance.c 	void (*tx)(struct net_device *);
tx                839 drivers/net/ethernet/amd/sunlance.c 		lp->tx(dev);
tx               1370 drivers/net/ethernet/amd/sunlance.c 		lp->tx = lance_tx_pio;
tx               1382 drivers/net/ethernet/amd/sunlance.c 		lp->tx = lance_tx_dvma;
tx                419 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 		memset(&ring->tx, 0, sizeof(ring->tx));
tx                510 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	memset(&rdata->tx, 0, sizeof(rdata->tx));
tx                551 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	if ((tso && (packet->mss != ring->tx.cur_mss)) ||
tx                552 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	    (vlan && (packet->vlan_ctag != ring->tx.cur_vlan_ctag)))
tx               1663 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 	ring->tx.xmit_more = 0;
tx               1695 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 	if (tso && (packet->mss != ring->tx.cur_mss))
tx               1700 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 	if (vlan && (packet->vlan_ctag != ring->tx.cur_vlan_ctag))
tx               1747 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 			ring->tx.cur_mss = packet->mss;
tx               1767 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 			ring->tx.cur_vlan_ctag = packet->vlan_ctag;
tx               1865 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 	rdata->tx.packets = tx_packets;
tx               1866 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 	rdata->tx.bytes = tx_bytes;
tx               1895 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 		ring->tx.xmit_more = 1;
tx                282 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 		ring->tx.queue_stopped = 1;
tx                287 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 		if (ring->tx.xmit_more)
tx               1851 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 		if (skb_shinfo(skb)->gso_size != ring->tx.cur_mss) {
tx               1873 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 		if (skb_vlan_tag_get(skb) != ring->tx.cur_vlan_ctag)
tx               2655 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 			tx_packets += rdata->tx.packets;
tx               2656 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 			tx_bytes += rdata->tx.bytes;
tx               2672 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	if ((ring->tx.queue_stopped == 1) &&
tx               2674 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 		ring->tx.queue_stopped = 0;
tx                660 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	unsigned int rx, tx;
tx                689 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	tx = __rounddown_pow_of_two(ringparam->tx_pending);
tx                690 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	if (tx != ringparam->tx_pending)
tx                693 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 			      tx);
tx                696 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	    (tx == pdata->tx_desc_count))
tx                700 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	pdata->tx_desc_count = tx;
tx                712 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	unsigned int rx, tx, combined;
tx                722 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	tx = min(pdata->hw_feat.tx_ch_cnt, pdata->tx_max_channel_count);
tx                723 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	tx = min(tx, pdata->channel_irq_count);
tx                724 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	tx = min(tx, pdata->tx_max_q_count);
tx                726 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	combined = min(rx, tx);
tx                730 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	channels->max_tx = tx ? tx - 1 : 0;
tx                734 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	tx = pdata->new_tx_ring_count ? : pdata->tx_ring_count;
tx                736 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	combined = min(rx, tx);
tx                738 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	tx -= combined;
tx                742 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	channels->tx_count = tx;
tx                757 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	unsigned int rx, rx_curr, tx, tx_curr, combined;
tx                767 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	tx = min(pdata->hw_feat.tx_ch_cnt, pdata->tx_max_channel_count);
tx                768 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	tx = min(tx, pdata->tx_max_q_count);
tx                769 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	tx = min(tx, pdata->channel_irq_count);
tx                771 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	combined = min(rx, tx);
tx                814 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	if ((channels->combined_count + channels->tx_count) > tx) {
tx                817 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 			   channels->combined_count + channels->tx_count, tx);
tx                823 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	tx = channels->combined_count + channels->tx_count;
tx                828 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	if ((rx == rx_curr) && (tx == tx_curr))
tx                832 drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c 	pdata->new_tx_ring_count = tx;
tx                430 drivers/net/ethernet/amd/xgbe/xgbe.h 	struct xgbe_tx_ring_data tx;	/* Tx-related data */
tx                488 drivers/net/ethernet/amd/xgbe/xgbe.h 		} tx;
tx                607 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c 				    u32 *rx, u32 *tx)
tx                613 drivers/net/ethernet/apm/xgene/xgene_enet_hw.c 	*tx = ECM_DROP_COUNT(count);
tx                149 drivers/net/ethernet/apm/xgene/xgene_enet_main.h 	void (*get_drop_cnt)(struct xgene_enet_pdata *pdata, u32 *rx, u32 *tx);
tx                 87 drivers/net/ethernet/apm/xgene/xgene_enet_sgmac.c 				     u32 *rx, u32 *tx)
tx                 96 drivers/net/ethernet/apm/xgene/xgene_enet_sgmac.c 	*tx = ECM_DROP_COUNT(count);
tx                172 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 				     u32 *rx, u32 *tx)
tx                178 drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c 	*tx = ECM_DROP_COUNT(count);
tx                208 drivers/net/ethernet/aquantia/atlantic/aq_ring.c 		ring->stats.tx.queue_restarts++;
tx                248 drivers/net/ethernet/aquantia/atlantic/aq_ring.c 			self->stats.tx.bytes += buff->skb->len;
tx                108 drivers/net/ethernet/aquantia/atlantic/aq_ring.h 	struct aq_ring_stats_tx_s tx;
tx                343 drivers/net/ethernet/aquantia/atlantic/aq_vec.c 		struct aq_ring_stats_tx_s *tx = &ring[AQ_VEC_TX_ID].stats.tx;
tx                355 drivers/net/ethernet/aquantia/atlantic/aq_vec.c 		stats_tx->packets += tx->packets;
tx                356 drivers/net/ethernet/aquantia/atlantic/aq_vec.c 		stats_tx->bytes += tx->bytes;
tx                357 drivers/net/ethernet/aquantia/atlantic/aq_vec.c 		stats_tx->errors += tx->errors;
tx                358 drivers/net/ethernet/aquantia/atlantic/aq_vec.c 		stats_tx->queue_restarts += tx->queue_restarts;
tx                129 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c 				  u32 tx)
tx                163 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c 	aq_hw_write_reg_bit(aq_hw, itr_imr_txten_adr[tx],
tx                164 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c 			    itr_imr_txten_msk[tx],
tx                165 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c 			    itr_imr_txten_shift[tx],
tx                209 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c void hw_atl_itr_irq_map_tx_set(struct aq_hw_s *aq_hw, u32 irq_map_tx, u32 tx)
tx                243 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c 	aq_hw_write_reg_bit(aq_hw, itr_imr_txt_adr[tx],
tx                244 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c 			    itr_imr_txt_msk[tx],
tx                245 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c 			    itr_imr_txt_shift[tx],
tx                106 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.h 				  u32 tx);
tx                112 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.h void hw_atl_itr_irq_map_tx_set(struct aq_hw_s *aq_hw, u32 irq_map_tx, u32 tx);
tx                255 drivers/net/ethernet/atheros/ag71xx.c 		} tx;
tx                649 drivers/net/ethernet/atheros/ag71xx.c 		skb = ring->buf[i].tx.skb;
tx                669 drivers/net/ethernet/atheros/ag71xx.c 		ring->buf[i].tx.skb = NULL;
tx                671 drivers/net/ethernet/atheros/ag71xx.c 		bytes_compl += ring->buf[i].tx.len;
tx                706 drivers/net/ethernet/atheros/ag71xx.c 		u32 rx, tx;
tx                711 drivers/net/ethernet/atheros/ag71xx.c 		tx = ag71xx_rr(ag, AG71XX_REG_TX_CTRL) & TX_CTRL_TXE;
tx                712 drivers/net/ethernet/atheros/ag71xx.c 		if (!rx && !tx)
tx                972 drivers/net/ethernet/atheros/ag71xx.c 		if (ring->buf[i].tx.skb) {
tx                973 drivers/net/ethernet/atheros/ag71xx.c 			bytes_compl += ring->buf[i].tx.len;
tx                975 drivers/net/ethernet/atheros/ag71xx.c 			dev_kfree_skb_any(ring->buf[i].tx.skb);
tx                977 drivers/net/ethernet/atheros/ag71xx.c 		ring->buf[i].tx.skb = NULL;
tx               1001 drivers/net/ethernet/atheros/ag71xx.c 		ring->buf[i].tx.skb = NULL;
tx               1133 drivers/net/ethernet/atheros/ag71xx.c 	struct ag71xx_ring *tx = &ag->tx_ring;
tx               1137 drivers/net/ethernet/atheros/ag71xx.c 	ring_size = BIT(tx->order) + BIT(rx->order);
tx               1138 drivers/net/ethernet/atheros/ag71xx.c 	tx_size = BIT(tx->order);
tx               1140 drivers/net/ethernet/atheros/ag71xx.c 	tx->buf = kcalloc(ring_size, sizeof(*tx->buf), GFP_KERNEL);
tx               1141 drivers/net/ethernet/atheros/ag71xx.c 	if (!tx->buf)
tx               1144 drivers/net/ethernet/atheros/ag71xx.c 	tx->descs_cpu = dma_alloc_coherent(&ag->pdev->dev,
tx               1146 drivers/net/ethernet/atheros/ag71xx.c 					   &tx->descs_dma, GFP_KERNEL);
tx               1147 drivers/net/ethernet/atheros/ag71xx.c 	if (!tx->descs_cpu) {
tx               1148 drivers/net/ethernet/atheros/ag71xx.c 		kfree(tx->buf);
tx               1149 drivers/net/ethernet/atheros/ag71xx.c 		tx->buf = NULL;
tx               1153 drivers/net/ethernet/atheros/ag71xx.c 	rx->buf = &tx->buf[tx_size];
tx               1154 drivers/net/ethernet/atheros/ag71xx.c 	rx->descs_cpu = ((void *)tx->descs_cpu) + tx_size * AG71XX_DESC_SIZE;
tx               1155 drivers/net/ethernet/atheros/ag71xx.c 	rx->descs_dma = tx->descs_dma + tx_size * AG71XX_DESC_SIZE;
tx               1163 drivers/net/ethernet/atheros/ag71xx.c 	struct ag71xx_ring *tx = &ag->tx_ring;
tx               1167 drivers/net/ethernet/atheros/ag71xx.c 	ring_size = BIT(tx->order) + BIT(rx->order);
tx               1169 drivers/net/ethernet/atheros/ag71xx.c 	if (tx->descs_cpu)
tx               1171 drivers/net/ethernet/atheros/ag71xx.c 				  tx->descs_cpu, tx->descs_dma);
tx               1173 drivers/net/ethernet/atheros/ag71xx.c 	kfree(tx->buf);
tx               1175 drivers/net/ethernet/atheros/ag71xx.c 	tx->descs_cpu = NULL;
tx               1177 drivers/net/ethernet/atheros/ag71xx.c 	tx->buf = NULL;
tx               1358 drivers/net/ethernet/atheros/ag71xx.c 	ring->buf[i].tx.len = skb->len;
tx               1359 drivers/net/ethernet/atheros/ag71xx.c 	ring->buf[i].tx.skb = skb;
tx               1130 drivers/net/ethernet/aurora/nb8800.c 	u32 rx, tx;
tx               1134 drivers/net/ethernet/aurora/nb8800.c 		tx = nb8800_read_stat(dev, i | 0x80);
tx               1136 drivers/net/ethernet/aurora/nb8800.c 		st[i + NB8800_NUM_STATS / 2] = tx;
tx                235 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_64_octets", mib.tx.pkt_cnt.cnt_64),
tx                236 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_65_127_oct", mib.tx.pkt_cnt.cnt_127),
tx                237 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_128_255_oct", mib.tx.pkt_cnt.cnt_255),
tx                238 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_256_511_oct", mib.tx.pkt_cnt.cnt_511),
tx                239 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_512_1023_oct", mib.tx.pkt_cnt.cnt_1023),
tx                240 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_1024_1518_oct", mib.tx.pkt_cnt.cnt_1518),
tx                241 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_vlan_1519_1522_oct", mib.tx.pkt_cnt.cnt_mgv),
tx                242 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_1522_2047_oct", mib.tx.pkt_cnt.cnt_2047),
tx                243 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_2048_4095_oct", mib.tx.pkt_cnt.cnt_4095),
tx                244 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_4096_9216_oct", mib.tx.pkt_cnt.cnt_9216),
tx                245 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_pkts", mib.tx.pkts),
tx                246 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_multicast", mib.tx.mca),
tx                247 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_broadcast", mib.tx.bca),
tx                248 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_pause", mib.tx.pf),
tx                249 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_control", mib.tx.cf),
tx                250 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_fcs_err", mib.tx.fcs),
tx                251 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_oversize", mib.tx.ovr),
tx                252 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_defer", mib.tx.drf),
tx                253 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_excess_defer", mib.tx.edf),
tx                254 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_single_col", mib.tx.scl),
tx                255 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_multi_col", mib.tx.mcl),
tx                256 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_late_col", mib.tx.lcl),
tx                257 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_excess_col", mib.tx.ecl),
tx                258 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_frags", mib.tx.frg),
tx                259 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_total_col", mib.tx.ncl),
tx                260 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_jabber", mib.tx.jbr),
tx                261 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_bytes", mib.tx.bytes),
tx                262 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_good_pkts", mib.tx.pok),
tx                263 drivers/net/ethernet/broadcom/bcmsysport.c 	STAT_MIB_TX("tx_unicast", mib.tx.uc),
tx                590 drivers/net/ethernet/broadcom/bcmsysport.h 	struct bcm_sysport_tx_counters tx;
tx               7338 drivers/net/ethernet/broadcom/bnx2.c bnx2_change_ring_size(struct bnx2 *bp, u32 rx, u32 tx, bool reset_irq)
tx               7357 drivers/net/ethernet/broadcom/bnx2.c 	bp->tx_ring_size = tx;
tx               1980 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	int rc, tx, rx;
tx               1982 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	tx = BNX2X_NUM_ETH_QUEUES(bp) * bp->max_cos;
tx               1988 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		tx++;
tx               1991 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	rc = netif_set_real_num_tx_queues(bp->dev, tx);
tx               2003 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 			  tx, rx);
tx               3741 drivers/net/ethernet/broadcom/bnx2x/bnx2x_hsi.h 	struct client_init_tx_data tx;
tx               4814 drivers/net/ethernet/broadcom/bnx2x/bnx2x_hsi.h 	struct client_init_tx_data tx;
tx               12159 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c 	u32 rx = 0, tx = 0, i;
tx               12170 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c 			tx = REG_RD(bp, shmem_base +
tx               12178 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c 			tx = REG_RD(bp, shmem_base +
tx               12186 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c 		phy->tx_preemphasis[i << 1] = ((tx>>16) & 0xffff);
tx               12187 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c 		phy->tx_preemphasis[(i << 1) + 1] = (tx & 0xffff);
tx               8727 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		__set_bit(BNX2X_Q_FLG_HC, &init_params->tx.flags);
tx               8733 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		__set_bit(BNX2X_Q_FLG_HC_EN, &init_params->tx.flags);
tx               8738 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		init_params->tx.hc_rate = bp->tx_ticks ?
tx               8742 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		init_params->rx.fw_sb_id = init_params->tx.fw_sb_id =
tx               8750 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c 		init_params->tx.sb_cq_index = HC_INDEX_ETH_FIRST_TX_CQ_CONS;
tx               4956 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 				  &data->tx,
tx               4981 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 				  &data->tx,
tx               4986 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 			 data->tx.tx_bd_page_base.lo,
tx               4987 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 			 data->tx.tx_bd_page_base.hi);
tx               5011 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 	    test_bit(BNX2X_Q_FLG_HC, &init->tx.flags)) {
tx               5012 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 		hc_usec = init->tx.hc_rate ? 1000000 / init->tx.hc_rate : 0;
tx               5014 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 		bnx2x_update_coalesce_sb_index(bp, init->tx.fw_sb_id,
tx               5015 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 			init->tx.sb_cq_index,
tx               5016 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 			!test_bit(BNX2X_Q_FLG_HC_EN, &init->tx.flags),
tx                886 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.h 	} tx;
tx                136 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	   init_params->tx.sb_cq_index,
tx                137 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	   init_params->tx.hc_rate,
tx                181 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	if (test_bit(BNX2X_Q_FLG_HC, &init_p->tx.flags))
tx                182 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 		__set_bit(BNX2X_Q_FLG_HC_EN, &init_p->tx.flags);
tx                186 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c 	init_p->tx.fw_sb_id = vf_igu_sb(vf, q->sb_idx);
tx               1565 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 			init_p->tx.hc_rate = setup_q->txq.hc_rate;
tx               1566 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 			init_p->tx.sb_cq_index = setup_q->txq.sb_index;
tx               1569 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c 						 &init_p->tx.flags);
tx               5650 drivers/net/ethernet/broadcom/bnxt/bnxt.c static int bnxt_trim_rings(struct bnxt *bp, int *rx, int *tx, int max,
tx               5685 drivers/net/ethernet/broadcom/bnxt/bnxt.c 			int tx = hw_resc->resv_tx_rings;
tx               5689 drivers/net/ethernet/broadcom/bnxt/bnxt.c 			if (cp < (rx + tx)) {
tx               5690 drivers/net/ethernet/broadcom/bnxt/bnxt.c 				bnxt_trim_rings(bp, &rx, &tx, cp, false);
tx               5694 drivers/net/ethernet/broadcom/bnxt/bnxt.c 				hw_resc->resv_tx_rings = tx;
tx               5864 drivers/net/ethernet/broadcom/bnxt/bnxt.c static int bnxt_hwrm_reserve_rings(struct bnxt *bp, int tx, int rx, int grp,
tx               5868 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		return bnxt_hwrm_reserve_pf_rings(bp, tx, rx, grp, cp, stat,
tx               5871 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		return bnxt_hwrm_reserve_vf_rings(bp, tx, rx, grp, cp, stat,
tx               5950 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	int tx = bp->tx_nr_rings;
tx               5968 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	rc = bnxt_hwrm_reserve_rings(bp, tx, rx, grp, cp, stat, vnic);
tx               5972 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	tx = hw_resc->resv_tx_rings;
tx               6001 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	rc = bnxt_trim_rings(bp, &rx_rings, &tx, cp, sh);
tx               6004 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	cp = sh ? max_t(int, tx, rx_rings) : tx + rx_rings;
tx               6005 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	bp->tx_nr_rings = tx;
tx               6009 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	if (!tx || !rx || !cp || !grp || !vnic || !stat)
tx               7854 drivers/net/ethernet/broadcom/bnxt/bnxt.c static int bnxt_trim_rings(struct bnxt *bp, int *rx, int *tx, int max,
tx               7857 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	int _rx = *rx, _tx = *tx;
tx               7861 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		*tx = min_t(int, _tx, max);
tx               7873 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		*tx = _tx;
tx               9498 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		struct tx_port_stats *tx = bp->hw_tx_port_stats;
tx               9507 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		stats->collisions = le64_to_cpu(tx->tx_total_collisions);
tx               9508 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		stats->tx_fifo_errors = le64_to_cpu(tx->tx_fifo_underruns);
tx               9509 drivers/net/ethernet/broadcom/bnxt/bnxt.c 		stats->tx_errors = le64_to_cpu(tx->tx_err);
tx               10342 drivers/net/ethernet/broadcom/bnxt/bnxt.c int bnxt_check_rings(struct bnxt *bp, int tx, int rx, bool sh, int tcs,
tx               10360 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	tx_rings_needed = tx * tx_sets + tx_xdp;
tx               11494 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	int rx, tx, cp;
tx               11496 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	_bnxt_get_max_rings(bp, &rx, &tx, &cp);
tx               11498 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	*max_tx = tx;
tx               11499 drivers/net/ethernet/broadcom/bnxt/bnxt.c 	if (!rx || !tx || !cp)
tx               1997 drivers/net/ethernet/broadcom/bnxt/bnxt.h int bnxt_check_rings(struct bnxt *bp, int tx, int rx, bool sh, int tcs,
tx                302 drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c 		int rx, tx;
tx                305 drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c 		bnxt_get_max_rings(bp, &rx, &tx, true);
tx               4961 drivers/net/ethernet/broadcom/cnic.c 	data->tx.tx_bd_page_base.hi = cpu_to_le32(val);
tx               4966 drivers/net/ethernet/broadcom/cnic.c 	data->tx.tx_bd_page_base.lo = cpu_to_le32(val);
tx               4969 drivers/net/ethernet/broadcom/cnic.c 	data->tx.tx_sb_index_number = HC_SP_INDEX_ETH_ISCSI_CQ_CONS;
tx               4970 drivers/net/ethernet/broadcom/cnic.c 	data->tx.tx_status_block_id = BNX2X_DEF_SB_ID;
tx                817 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_64_octets", mib.tx.pkt_cnt.cnt_64),
tx                818 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_65_127_oct", mib.tx.pkt_cnt.cnt_127),
tx                819 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_128_255_oct", mib.tx.pkt_cnt.cnt_255),
tx                820 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_256_511_oct", mib.tx.pkt_cnt.cnt_511),
tx                821 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_512_1023_oct", mib.tx.pkt_cnt.cnt_1023),
tx                822 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_1024_1518_oct", mib.tx.pkt_cnt.cnt_1518),
tx                823 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_vlan_1519_1522_oct", mib.tx.pkt_cnt.cnt_mgv),
tx                824 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_1522_2047_oct", mib.tx.pkt_cnt.cnt_2047),
tx                825 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_2048_4095_oct", mib.tx.pkt_cnt.cnt_4095),
tx                826 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_4096_9216_oct", mib.tx.pkt_cnt.cnt_9216),
tx                827 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_pkts", mib.tx.pkts),
tx                828 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_multicast", mib.tx.mca),
tx                829 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_broadcast", mib.tx.bca),
tx                830 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_pause", mib.tx.pf),
tx                831 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_control", mib.tx.cf),
tx                832 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_fcs_err", mib.tx.fcs),
tx                833 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_oversize", mib.tx.ovr),
tx                834 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_defer", mib.tx.drf),
tx                835 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_excess_defer", mib.tx.edf),
tx                836 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_single_col", mib.tx.scl),
tx                837 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_multi_col", mib.tx.mcl),
tx                838 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_late_col", mib.tx.lcl),
tx                839 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_excess_col", mib.tx.ecl),
tx                840 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_frags", mib.tx.frg),
tx                841 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_total_col", mib.tx.ncl),
tx                842 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_jabber", mib.tx.jbr),
tx                843 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_bytes", mib.tx.bytes),
tx                844 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_good_pkts", mib.tx.pok),
tx                845 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	STAT_GENET_MIB_TX("tx_unicast", mib.tx.uc),
tx                136 drivers/net/ethernet/broadcom/genet/bcmgenet.h 	struct bcmgenet_tx_counters tx;
tx                289 drivers/net/ethernet/brocade/bna/bna.h void bna_bfi_tx_enet_start_rsp(struct bna_tx *tx,
tx                291 drivers/net/ethernet/brocade/bna/bna.h void bna_bfi_tx_enet_stop_rsp(struct bna_tx *tx,
tx                312 drivers/net/ethernet/brocade/bna/bna.h void bna_tx_destroy(struct bna_tx *tx);
tx                313 drivers/net/ethernet/brocade/bna/bna.h void bna_tx_enable(struct bna_tx *tx);
tx                314 drivers/net/ethernet/brocade/bna/bna.h void bna_tx_disable(struct bna_tx *tx, enum bna_cleanup_type type,
tx                316 drivers/net/ethernet/brocade/bna/bna.h void bna_tx_cleanup_complete(struct bna_tx *tx);
tx                317 drivers/net/ethernet/brocade/bna/bna.h void bna_tx_coalescing_timeo_set(struct bna_tx *tx, int coalescing_timeo);
tx                276 drivers/net/ethernet/brocade/bna/bna_enet.c 	struct bna_tx *tx;
tx                322 drivers/net/ethernet/brocade/bna/bna_enet.c 		bna_tx_from_rid(bna, msghdr->enet_id, tx);
tx                323 drivers/net/ethernet/brocade/bna/bna_enet.c 		if (tx)
tx                324 drivers/net/ethernet/brocade/bna/bna_enet.c 			bna_bfi_tx_enet_start_rsp(tx, msghdr);
tx                328 drivers/net/ethernet/brocade/bna/bna_enet.c 		bna_tx_from_rid(bna, msghdr->enet_id, tx);
tx                329 drivers/net/ethernet/brocade/bna/bna_enet.c 		if (tx)
tx                330 drivers/net/ethernet/brocade/bna/bna_enet.c 			bna_bfi_tx_enet_stop_rsp(tx, msghdr);
tx               2758 drivers/net/ethernet/brocade/bna/bna_tx_rx.c #define call_tx_stop_cbfn(tx)						\
tx               2760 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	if ((tx)->stop_cbfn) {						\
tx               2763 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		cbfn = (tx)->stop_cbfn;					\
tx               2764 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		cbarg = (tx)->stop_cbarg;				\
tx               2765 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		(tx)->stop_cbfn = NULL;					\
tx               2766 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		(tx)->stop_cbarg = NULL;				\
tx               2767 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		cbfn(cbarg, (tx));					\
tx               2771 drivers/net/ethernet/brocade/bna/bna_tx_rx.c static void bna_tx_mod_cb_tx_stopped(void *tx_mod, struct bna_tx *tx);
tx               2772 drivers/net/ethernet/brocade/bna/bna_tx_rx.c static void bna_bfi_tx_enet_start(struct bna_tx *tx);
tx               2773 drivers/net/ethernet/brocade/bna/bna_tx_rx.c static void bna_tx_enet_stop(struct bna_tx *tx);
tx               2800 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_stopped_entry(struct bna_tx *tx)
tx               2802 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	call_tx_stop_cbfn(tx);
tx               2806 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_stopped(struct bna_tx *tx, enum bna_tx_event event)
tx               2810 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_start_wait);
tx               2814 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		call_tx_stop_cbfn(tx);
tx               2831 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_start_wait_entry(struct bna_tx *tx)
tx               2833 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	bna_bfi_tx_enet_start(tx);
tx               2837 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_start_wait(struct bna_tx *tx, enum bna_tx_event event)
tx               2841 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		tx->flags &= ~BNA_TX_F_BW_UPDATED;
tx               2842 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_stop_wait);
tx               2846 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		tx->flags &= ~BNA_TX_F_BW_UPDATED;
tx               2847 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_stopped);
tx               2851 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		if (tx->flags & BNA_TX_F_BW_UPDATED) {
tx               2852 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			tx->flags &= ~BNA_TX_F_BW_UPDATED;
tx               2853 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			bfa_fsm_set_state(tx, bna_tx_sm_prio_stop_wait);
tx               2855 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			bfa_fsm_set_state(tx, bna_tx_sm_started);
tx               2859 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		tx->flags |= BNA_TX_F_BW_UPDATED;
tx               2868 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_started_entry(struct bna_tx *tx)
tx               2871 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	int is_regular = (tx->type == BNA_TX_T_REGULAR);
tx               2873 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	list_for_each_entry(txq, &tx->txq_q, qe) {
tx               2876 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bna_ib_start(tx->bna, &txq->ib, is_regular);
tx               2878 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->tx_resume_cbfn(tx->bna->bnad, tx);
tx               2882 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_started(struct bna_tx *tx, enum bna_tx_event event)
tx               2886 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_stop_wait);
tx               2887 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		tx->tx_stall_cbfn(tx->bna->bnad, tx);
tx               2888 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bna_tx_enet_stop(tx);
tx               2892 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_failed);
tx               2893 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		tx->tx_stall_cbfn(tx->bna->bnad, tx);
tx               2894 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		tx->tx_cleanup_cbfn(tx->bna->bnad, tx);
tx               2898 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_prio_stop_wait);
tx               2907 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_stop_wait_entry(struct bna_tx *tx)
tx               2912 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_stop_wait(struct bna_tx *tx, enum bna_tx_event event)
tx               2917 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_cleanup_wait);
tx               2918 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		tx->tx_cleanup_cbfn(tx->bna->bnad, tx);
tx               2926 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bna_tx_enet_stop(tx);
tx               2939 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_cleanup_wait_entry(struct bna_tx *tx)
tx               2944 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_cleanup_wait(struct bna_tx *tx, enum bna_tx_event event)
tx               2953 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_stopped);
tx               2962 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_prio_stop_wait_entry(struct bna_tx *tx)
tx               2964 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->tx_stall_cbfn(tx->bna->bnad, tx);
tx               2965 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	bna_tx_enet_stop(tx);
tx               2969 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_prio_stop_wait(struct bna_tx *tx, enum bna_tx_event event)
tx               2973 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_stop_wait);
tx               2977 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_failed);
tx               2978 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		tx->tx_cleanup_cbfn(tx->bna->bnad, tx);
tx               2982 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_prio_cleanup_wait);
tx               2995 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_prio_cleanup_wait_entry(struct bna_tx *tx)
tx               2997 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->tx_cleanup_cbfn(tx->bna->bnad, tx);
tx               3001 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_prio_cleanup_wait(struct bna_tx *tx, enum bna_tx_event event)
tx               3005 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_cleanup_wait);
tx               3009 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_failed);
tx               3017 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_start_wait);
tx               3026 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_failed_entry(struct bna_tx *tx)
tx               3031 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_failed(struct bna_tx *tx, enum bna_tx_event event)
tx               3035 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_quiesce_wait);
tx               3039 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_cleanup_wait);
tx               3047 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_stopped);
tx               3056 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_quiesce_wait_entry(struct bna_tx *tx)
tx               3061 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_sm_quiesce_wait(struct bna_tx *tx, enum bna_tx_event event)
tx               3065 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_cleanup_wait);
tx               3069 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_failed);
tx               3073 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_set_state(tx, bna_tx_sm_start_wait);
tx               3086 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_bfi_tx_enet_start(struct bna_tx *tx)
tx               3088 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	struct bfi_enet_tx_cfg_req *cfg_req = &tx->bfi_enet_cmd.cfg_req;
tx               3093 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		BFI_ENET_H2I_TX_CFG_SET_REQ, 0, tx->rid);
tx               3097 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	cfg_req->num_queues = tx->num_txq;
tx               3098 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	for (i = 0; i < tx->num_txq; i++) {
tx               3100 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			: list_first_entry(&tx->txq_q, struct bna_txq, qe);
tx               3125 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	cfg_req->tx_cfg.vlan_id = htons((u16)tx->txf_vlan_id);
tx               3129 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	bfa_msgq_cmd_set(&tx->msgq_cmd, NULL, NULL,
tx               3131 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	bfa_msgq_cmd_post(&tx->bna->msgq, &tx->msgq_cmd);
tx               3135 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_bfi_tx_enet_stop(struct bna_tx *tx)
tx               3137 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	struct bfi_enet_req *req = &tx->bfi_enet_cmd.req;
tx               3140 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		BFI_ENET_H2I_TX_CFG_CLR_REQ, 0, tx->rid);
tx               3143 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	bfa_msgq_cmd_set(&tx->msgq_cmd, NULL, NULL, sizeof(struct bfi_enet_req),
tx               3145 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	bfa_msgq_cmd_post(&tx->bna->msgq, &tx->msgq_cmd);
tx               3149 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_enet_stop(struct bna_tx *tx)
tx               3154 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	list_for_each_entry(txq, &tx->txq_q, qe)
tx               3155 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bna_ib_stop(tx->bna, &txq->ib);
tx               3157 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	bna_bfi_tx_enet_stop(tx);
tx               3199 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	struct bna_tx *tx = NULL;
tx               3204 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		tx = list_first_entry(&tx_mod->tx_free_q, struct bna_tx, qe);
tx               3206 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		tx = list_last_entry(&tx_mod->tx_free_q, struct bna_tx, qe);
tx               3207 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	list_del(&tx->qe);
tx               3208 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->type = type;
tx               3210 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	return tx;
tx               3214 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_free(struct bna_tx *tx)
tx               3216 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	struct bna_tx_mod *tx_mod = &tx->bna->tx_mod;
tx               3220 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	while (!list_empty(&tx->txq_q)) {
tx               3221 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq = list_first_entry(&tx->txq_q, struct bna_txq, qe);
tx               3223 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tx = NULL;
tx               3228 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		if (qe == &tx->qe) {
tx               3229 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			list_del(&tx->qe);
tx               3234 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->bna = NULL;
tx               3235 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->priv = NULL;
tx               3238 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		if (((struct bna_tx *)qe)->rid < tx->rid)
tx               3241 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	list_add(&tx->qe, qe);
tx               3245 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_start(struct bna_tx *tx)
tx               3247 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->flags |= BNA_TX_F_ENET_STARTED;
tx               3248 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	if (tx->flags & BNA_TX_F_ENABLED)
tx               3249 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_send_event(tx, TX_E_START);
tx               3253 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_stop(struct bna_tx *tx)
tx               3255 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->stop_cbfn = bna_tx_mod_cb_tx_stopped;
tx               3256 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->stop_cbarg = &tx->bna->tx_mod;
tx               3258 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->flags &= ~BNA_TX_F_ENET_STARTED;
tx               3259 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	bfa_fsm_send_event(tx, TX_E_STOP);
tx               3263 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_fail(struct bna_tx *tx)
tx               3265 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->flags &= ~BNA_TX_F_ENET_STARTED;
tx               3266 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	bfa_fsm_send_event(tx, TX_E_FAIL);
tx               3270 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_bfi_tx_enet_start_rsp(struct bna_tx *tx, struct bfi_msgq_mhdr *msghdr)
tx               3272 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	struct bfi_enet_tx_cfg_rsp *cfg_rsp = &tx->bfi_enet_cmd.cfg_rsp;
tx               3276 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	bfa_msgq_rsp_copy(&tx->bna->msgq, (u8 *)cfg_rsp,
tx               3279 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->hw_id = cfg_rsp->hw_id;
tx               3281 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	for (i = 0, txq = list_first_entry(&tx->txq_q, struct bna_txq, qe);
tx               3282 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	     i < tx->num_txq; i++, txq = list_next_entry(txq, qe)) {
tx               3285 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			tx->bna->pcidev.pci_bar_kva
tx               3288 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			tx->bna->pcidev.pci_bar_kva
tx               3297 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	bfa_fsm_send_event(tx, TX_E_STARTED);
tx               3301 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_bfi_tx_enet_stop_rsp(struct bna_tx *tx, struct bfi_msgq_mhdr *msghdr)
tx               3303 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	bfa_fsm_send_event(tx, TX_E_STOPPED);
tx               3309 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	struct bna_tx *tx;
tx               3311 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	list_for_each_entry(tx, &tx_mod->tx_active_q, qe)
tx               3312 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_send_event(tx, TX_E_BW_UPDATE);
tx               3370 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	struct bna_tx *tx;
tx               3388 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx = bna_tx_get(tx_mod, tx_cfg->tx_type);
tx               3389 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	if (!tx)
tx               3391 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->bna = bna;
tx               3392 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->priv = priv;
tx               3396 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	INIT_LIST_HEAD(&tx->txq_q);
tx               3402 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		list_move_tail(&txq->qe, &tx->txq_q);
tx               3403 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		txq->tx = tx;
tx               3412 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->tcb_setup_cbfn = tx_cbfn->tcb_setup_cbfn;
tx               3413 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->tcb_destroy_cbfn = tx_cbfn->tcb_destroy_cbfn;
tx               3415 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->tx_stall_cbfn = tx_cbfn->tx_stall_cbfn;
tx               3416 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->tx_resume_cbfn = tx_cbfn->tx_resume_cbfn;
tx               3417 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->tx_cleanup_cbfn = tx_cbfn->tx_cleanup_cbfn;
tx               3419 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	list_add_tail(&tx->qe, &tx_mod->tx_active_q);
tx               3421 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->num_txq = tx_cfg->num_txq;
tx               3423 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->flags = 0;
tx               3424 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	if (tx->bna->tx_mod.flags & BNA_TX_MOD_F_ENET_STARTED) {
tx               3425 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		switch (tx->type) {
tx               3427 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			if (!(tx->bna->tx_mod.flags &
tx               3429 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 				tx->flags |= BNA_TX_F_ENET_STARTED;
tx               3432 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			if (tx->bna->tx_mod.flags & BNA_TX_MOD_F_ENET_LOOPBACK)
tx               3433 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 				tx->flags |= BNA_TX_F_ENET_STARTED;
tx               3441 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	list_for_each_entry(txq, &tx->txq_q, qe) {
tx               3486 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		if (tx->tcb_setup_cbfn)
tx               3487 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			(tx->tcb_setup_cbfn)(bna->bnad, txq->tcb);
tx               3497 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->txf_vlan_id = 0;
tx               3499 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	bfa_fsm_set_state(tx, bna_tx_sm_stopped);
tx               3501 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx_mod->rid_mask |= BIT(tx->rid);
tx               3503 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	return tx;
tx               3506 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	bna_tx_free(tx);
tx               3511 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_destroy(struct bna_tx *tx)
tx               3515 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	list_for_each_entry(txq, &tx->txq_q, qe)
tx               3516 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		if (tx->tcb_destroy_cbfn)
tx               3517 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			(tx->tcb_destroy_cbfn)(tx->bna->bnad, txq->tcb);
tx               3519 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->bna->tx_mod.rid_mask &= ~BIT(tx->rid);
tx               3520 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	bna_tx_free(tx);
tx               3524 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_enable(struct bna_tx *tx)
tx               3526 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	if (tx->fsm != (bfa_sm_t)bna_tx_sm_stopped)
tx               3529 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->flags |= BNA_TX_F_ENABLED;
tx               3531 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	if (tx->flags & BNA_TX_F_ENET_STARTED)
tx               3532 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bfa_fsm_send_event(tx, TX_E_START);
tx               3536 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_disable(struct bna_tx *tx, enum bna_cleanup_type type,
tx               3540 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		(*cbfn)(tx->bna->bnad, tx);
tx               3544 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->stop_cbfn = cbfn;
tx               3545 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->stop_cbarg = tx->bna->bnad;
tx               3547 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx->flags &= ~BNA_TX_F_ENABLED;
tx               3549 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	bfa_fsm_send_event(tx, TX_E_STOP);
tx               3553 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_cleanup_complete(struct bna_tx *tx)
tx               3555 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	bfa_fsm_send_event(tx, TX_E_CLEANUP_DONE);
tx               3559 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_mod_cb_tx_stopped(void *arg, struct bna_tx *tx)
tx               3585 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	tx_mod->tx = (struct bna_tx *)
tx               3596 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		tx_mod->tx[i].rid = i;
tx               3597 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		list_add_tail(&tx_mod->tx[i].qe, &tx_mod->tx_free_q);
tx               3616 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	struct bna_tx *tx;
tx               3622 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	list_for_each_entry(tx, &tx_mod->tx_active_q, qe)
tx               3623 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		if (tx->type == type)
tx               3624 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			bna_tx_start(tx);
tx               3630 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	struct bna_tx *tx;
tx               3639 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	list_for_each_entry(tx, &tx_mod->tx_active_q, qe)
tx               3640 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		if (tx->type == type) {
tx               3642 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 			bna_tx_stop(tx);
tx               3651 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	struct bna_tx *tx;
tx               3656 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	list_for_each_entry(tx, &tx_mod->tx_active_q, qe)
tx               3657 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 		bna_tx_fail(tx);
tx               3661 drivers/net/ethernet/brocade/bna/bna_tx_rx.c bna_tx_coalescing_timeo_set(struct bna_tx *tx, int coalescing_timeo)
tx               3665 drivers/net/ethernet/brocade/bna/bna_tx_rx.c 	list_for_each_entry(txq, &tx->txq_q, qe)
tx                448 drivers/net/ethernet/brocade/bna/bna_types.h 	struct bna_tx *tx;
tx                480 drivers/net/ethernet/brocade/bna/bna_types.h 	void (*stop_cbfn)(void *arg, struct bna_tx *tx);
tx                514 drivers/net/ethernet/brocade/bna/bna_types.h 	struct bna_tx *tx;		/* BFI_MAX_TXQ entries */
tx                991 drivers/net/ethernet/brocade/bna/bnad.c bnad_cb_tx_disabled(void *arg, struct bna_tx *tx)
tx               1002 drivers/net/ethernet/brocade/bna/bnad.c 			(struct bnad_tx_info *)tcb->txq->tx->priv;
tx               1012 drivers/net/ethernet/brocade/bna/bnad.c 			(struct bnad_tx_info *)tcb->txq->tx->priv;
tx               1038 drivers/net/ethernet/brocade/bna/bnad.c bnad_cb_tx_stall(struct bnad *bnad, struct bna_tx *tx)
tx               1041 drivers/net/ethernet/brocade/bna/bnad.c 			(struct bnad_tx_info *)tx->priv;
tx               1057 drivers/net/ethernet/brocade/bna/bnad.c bnad_cb_tx_resume(struct bnad *bnad, struct bna_tx *tx)
tx               1059 drivers/net/ethernet/brocade/bna/bnad.c 	struct bnad_tx_info *tx_info = (struct bnad_tx_info *)tx->priv;
tx               1129 drivers/net/ethernet/brocade/bna/bnad.c 	bna_tx_cleanup_complete(tx_info->tx);
tx               1134 drivers/net/ethernet/brocade/bna/bnad.c bnad_cb_tx_cleanup(struct bnad *bnad, struct bna_tx *tx)
tx               1136 drivers/net/ethernet/brocade/bna/bnad.c 	struct bnad_tx_info *tx_info = (struct bnad_tx_info *)tx->priv;
tx               1917 drivers/net/ethernet/brocade/bna/bnad.c 	if (!tx_info->tx)
tx               1922 drivers/net/ethernet/brocade/bna/bnad.c 	bna_tx_disable(tx_info->tx, BNA_HARD_CLEANUP, bnad_cb_tx_disabled);
tx               1931 drivers/net/ethernet/brocade/bna/bnad.c 	bna_tx_destroy(tx_info->tx);
tx               1934 drivers/net/ethernet/brocade/bna/bnad.c 	tx_info->tx = NULL;
tx               1958 drivers/net/ethernet/brocade/bna/bnad.c 	struct bna_tx *tx;
tx               1987 drivers/net/ethernet/brocade/bna/bnad.c 	tx = bna_tx_create(&bnad->bna, bnad, tx_config, &tx_cbfn, res_info,
tx               1990 drivers/net/ethernet/brocade/bna/bnad.c 	if (!tx) {
tx               1994 drivers/net/ethernet/brocade/bna/bnad.c 	tx_info->tx = tx;
tx               2008 drivers/net/ethernet/brocade/bna/bnad.c 	bna_tx_enable(tx);
tx               2015 drivers/net/ethernet/brocade/bna/bnad.c 	bna_tx_destroy(tx_info->tx);
tx               2017 drivers/net/ethernet/brocade/bna/bnad.c 	tx_info->tx = NULL;
tx               2300 drivers/net/ethernet/brocade/bna/bnad.c 	if (!tx_info->tx)
tx               2303 drivers/net/ethernet/brocade/bna/bnad.c 	bna_tx_coalescing_timeo_set(tx_info->tx, bnad->tx_coalescing_timeo);
tx                204 drivers/net/ethernet/brocade/bna/bnad.h 	struct bna_tx *tx; /* 1:1 between tx_info & tx */
tx                480 drivers/net/ethernet/brocade/bna/bnad_ethtool.c 			if (!bnad->tx_info[i].tx)
tx                682 drivers/net/ethernet/brocade/bna/bnad_ethtool.c 			if (!bnad->tx_info[i].tx)
tx                745 drivers/net/ethernet/brocade/bna/bnad_ethtool.c 		if (!bnad->tx_info[i].tx)
tx                823 drivers/net/ethernet/brocade/bna/bnad_ethtool.c 		if (!bnad->tx_info[i].tx)
tx                644 drivers/net/ethernet/calxeda/xgmac.c static int xgmac_set_flow_ctrl(struct xgmac_priv *priv, int rx, int tx)
tx                650 drivers/net/ethernet/calxeda/xgmac.c 	priv->tx_pause = tx;
tx                652 drivers/net/ethernet/calxeda/xgmac.c 	if (rx || tx) {
tx                655 drivers/net/ethernet/calxeda/xgmac.c 		if (tx)
tx               2696 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c #define T3(fmt_spec, s, v) S3(fmt_spec, s, tx[i].v)
tx               2697 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c #define T(s, v) S3("u", s, tx[i].v)
tx               2706 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		const struct sge_eth_txq *tx = &s->ethtxq[base_qset];
tx               2722 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		    4*(7-tx[i].dcb_prio)) & 0xf);
tx               2725 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		    1*(7-tx[i].dcb_prio)) & 0x1);
tx               2764 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		const struct sge_uld_txq *tx;
tx               2767 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		tx = &utxq_info->uldtxq[r * 4];
tx               2913 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		const struct sge_uld_txq *tx;
tx               2917 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		tx = &utxq_info->uldtxq[r * 4];
tx               2946 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		const struct sge_ctrl_txq *tx = &s->ctrlq[r * 4];
tx                244 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	const struct sge_eth_txq *tx = &adap->sge.ethtxq[p->first_qset];
tx                248 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 	for (i = 0; i < p->nqsets; i++, rx++, tx++) {
tx                249 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 		s->tso += tx->tso;
tx                250 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 		s->tx_csum += tx->tx_cso;
tx                253 drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c 		s->vlan_ins += tx->vlan_ins;
tx                282 drivers/net/ethernet/cisco/enic/enic_ethtool.c 		*(data++) = ((u64 *)&vstats->tx)[enic_tx_stats[i].index];
tx                924 drivers/net/ethernet/cisco/enic/enic_main.c 	net_stats->tx_packets = stats->tx.tx_frames_ok;
tx                925 drivers/net/ethernet/cisco/enic/enic_main.c 	net_stats->tx_bytes = stats->tx.tx_bytes_ok;
tx                926 drivers/net/ethernet/cisco/enic/enic_main.c 	net_stats->tx_errors = stats->tx.tx_errors;
tx                927 drivers/net/ethernet/cisco/enic/enic_main.c 	net_stats->tx_dropped = stats->tx.tx_drops;
tx                 71 drivers/net/ethernet/cisco/enic/vnic_stats.h 	struct vnic_tx_stats tx;
tx                272 drivers/net/ethernet/cortina/gemini.c static void gmac_set_flow_control(struct net_device *netdev, bool tx, bool rx)
tx                282 drivers/net/ethernet/cortina/gemini.c 	if (tx)
tx                523 drivers/net/ethernet/dec/tulip/interrupt.c 	int tx = 0;
tx                640 drivers/net/ethernet/dec/tulip/interrupt.c 				tx++;
tx                731 drivers/net/ethernet/dec/tulip/interrupt.c 		if (tx > maxtx || rx > maxrx || oi > maxoi) {
tx                734 drivers/net/ethernet/dec/tulip/interrupt.c 					 csr5, tp->nir, tx, rx, oi);
tx                304 drivers/net/ethernet/freescale/fman/fman_port.c 	struct fman_port_tx_bmi_regs tx;
tx                569 drivers/net/ethernet/freescale/fman/fman_port.c 	struct fman_port_tx_bmi_regs __iomem *regs = &port->bmi_regs->tx;
tx               1029 drivers/net/ethernet/freescale/fman/fman_port.c 			iowrite32be(0xFFFFFF, &port->bmi_regs->tx.fmbm_tcfqid);
tx               1031 drivers/net/ethernet/freescale/fman/fman_port.c 				    &port->bmi_regs->tx.fmbm_tfene);
tx               1405 drivers/net/ethernet/freescale/fman/fman_port.c 		iowrite32be(reg, &port->bmi_regs->tx.fmbm_tfp);
tx               1605 drivers/net/ethernet/freescale/fman/fman_port.c 		bmi_cfg_reg = &port->bmi_regs->tx.fmbm_tcfg;
tx               1606 drivers/net/ethernet/freescale/fman/fman_port.c 		bmi_status_reg = &port->bmi_regs->tx.fmbm_tst;
tx               1680 drivers/net/ethernet/freescale/fman/fman_port.c 		bmi_cfg_reg = &port->bmi_regs->tx.fmbm_tcfg;
tx                334 drivers/net/ethernet/freescale/fman/mac.c int fman_set_mac_active_pause(struct mac_device *mac_dev, bool rx, bool tx)
tx                345 drivers/net/ethernet/freescale/fman/mac.c 	if (tx != mac_dev->tx_pause_active) {
tx                346 drivers/net/ethernet/freescale/fman/mac.c 		u16 pause_time = (tx ? FSL_FM_PAUSE_TIME_ENABLE :
tx                352 drivers/net/ethernet/freescale/fman/mac.c 			mac_dev->tx_pause_active = tx;
tx                 96 drivers/net/ethernet/freescale/fman/mac.h int fman_set_mac_active_pause(struct mac_device *mac_dev, bool rx, bool tx);
tx                148 drivers/net/ethernet/google/gve/gve.h 	struct gve_tx_ring *tx; /* tx rings on this block */
tx                166 drivers/net/ethernet/google/gve/gve.h 	struct gve_tx_ring *tx; /* array of tx_cfg.num_queues */
tx                439 drivers/net/ethernet/google/gve/gve.h 				 struct gve_tx_ring *tx);
tx                195 drivers/net/ethernet/google/gve/gve_adminq.c 	struct gve_tx_ring *tx = &priv->tx[queue_index];
tx                203 drivers/net/ethernet/google/gve/gve_adminq.c 		.queue_resources_addr = cpu_to_be64(tx->q_resources_bus),
tx                204 drivers/net/ethernet/google/gve/gve_adminq.c 		.tx_ring_addr = cpu_to_be64(tx->bus),
tx                205 drivers/net/ethernet/google/gve/gve_adminq.c 		.queue_page_list_id = cpu_to_be32(tx->tx_fifo.qpl->id),
tx                206 drivers/net/ethernet/google/gve/gve_adminq.c 		.ntfy_id = cpu_to_be32(tx->ntfy_id),
tx                285 drivers/net/ethernet/google/gve/gve_adminq.c 	if (priv->tx_desc_cnt * sizeof(priv->tx->desc[0]) < PAGE_SIZE) {
tx                115 drivers/net/ethernet/google/gve/gve_ethtool.c 		if (priv->tx) {
tx                118 drivers/net/ethernet/google/gve/gve_ethtool.c 				  u64_stats_fetch_begin(&priv->tx[ring].statss);
tx                119 drivers/net/ethernet/google/gve/gve_ethtool.c 				tx_pkts += priv->tx[ring].pkt_done;
tx                120 drivers/net/ethernet/google/gve/gve_ethtool.c 				tx_bytes += priv->tx[ring].bytes_done;
tx                121 drivers/net/ethernet/google/gve/gve_ethtool.c 			} while (u64_stats_fetch_retry(&priv->tx[ring].statss,
tx                148 drivers/net/ethernet/google/gve/gve_ethtool.c 	if (priv->tx) {
tx                150 drivers/net/ethernet/google/gve/gve_ethtool.c 			struct gve_tx_ring *tx = &priv->tx[ring];
tx                152 drivers/net/ethernet/google/gve/gve_ethtool.c 			data[i++] = tx->req;
tx                153 drivers/net/ethernet/google/gve/gve_ethtool.c 			data[i++] = tx->done;
tx                154 drivers/net/ethernet/google/gve/gve_ethtool.c 			data[i++] = tx->wake_queue;
tx                155 drivers/net/ethernet/google/gve/gve_ethtool.c 			data[i++] = tx->stop_queue;
tx                157 drivers/net/ethernet/google/gve/gve_ethtool.c 									  tx));
tx                 46 drivers/net/ethernet/google/gve/gve_main.c 	if (priv->tx) {
tx                 50 drivers/net/ethernet/google/gve/gve_main.c 				  u64_stats_fetch_begin(&priv->tx[ring].statss);
tx                 51 drivers/net/ethernet/google/gve/gve_main.c 				s->tx_packets += priv->tx[ring].pkt_done;
tx                 52 drivers/net/ethernet/google/gve/gve_main.c 				s->tx_bytes += priv->tx[ring].bytes_done;
tx                 53 drivers/net/ethernet/google/gve/gve_main.c 			} while (u64_stats_fetch_retry(&priv->tx[ring].statss,
tx                109 drivers/net/ethernet/google/gve/gve_main.c 	if (block->tx)
tx                125 drivers/net/ethernet/google/gve/gve_main.c 	if (block->tx)
tx                415 drivers/net/ethernet/google/gve/gve_main.c 	priv->tx = kvzalloc(priv->tx_cfg.num_queues * sizeof(*priv->tx),
tx                417 drivers/net/ethernet/google/gve/gve_main.c 	if (!priv->tx)
tx                434 drivers/net/ethernet/google/gve/gve_main.c 		u64_stats_init(&priv->tx[i].statss);
tx                453 drivers/net/ethernet/google/gve/gve_main.c 	kvfree(priv->tx);
tx                454 drivers/net/ethernet/google/gve/gve_main.c 	priv->tx = NULL;
tx                497 drivers/net/ethernet/google/gve/gve_main.c 	if (priv->tx) {
tx                503 drivers/net/ethernet/google/gve/gve_main.c 		kvfree(priv->tx);
tx                504 drivers/net/ethernet/google/gve/gve_main.c 		priv->tx = NULL;
tx                139 drivers/net/ethernet/google/gve/gve_tx.c 	block->tx = NULL;
tx                142 drivers/net/ethernet/google/gve/gve_tx.c static int gve_clean_tx_done(struct gve_priv *priv, struct gve_tx_ring *tx,
tx                147 drivers/net/ethernet/google/gve/gve_tx.c 	struct gve_tx_ring *tx = &priv->tx[idx];
tx                153 drivers/net/ethernet/google/gve/gve_tx.c 	slots = tx->mask + 1;
tx                154 drivers/net/ethernet/google/gve/gve_tx.c 	gve_clean_tx_done(priv, tx, tx->req, false);
tx                155 drivers/net/ethernet/google/gve/gve_tx.c 	netdev_tx_reset_queue(tx->netdev_txq);
tx                157 drivers/net/ethernet/google/gve/gve_tx.c 	dma_free_coherent(hdev, sizeof(*tx->q_resources),
tx                158 drivers/net/ethernet/google/gve/gve_tx.c 			  tx->q_resources, tx->q_resources_bus);
tx                159 drivers/net/ethernet/google/gve/gve_tx.c 	tx->q_resources = NULL;
tx                161 drivers/net/ethernet/google/gve/gve_tx.c 	gve_tx_fifo_release(priv, &tx->tx_fifo);
tx                162 drivers/net/ethernet/google/gve/gve_tx.c 	gve_unassign_qpl(priv, tx->tx_fifo.qpl->id);
tx                163 drivers/net/ethernet/google/gve/gve_tx.c 	tx->tx_fifo.qpl = NULL;
tx                165 drivers/net/ethernet/google/gve/gve_tx.c 	bytes = sizeof(*tx->desc) * slots;
tx                166 drivers/net/ethernet/google/gve/gve_tx.c 	dma_free_coherent(hdev, bytes, tx->desc, tx->bus);
tx                167 drivers/net/ethernet/google/gve/gve_tx.c 	tx->desc = NULL;
tx                169 drivers/net/ethernet/google/gve/gve_tx.c 	vfree(tx->info);
tx                170 drivers/net/ethernet/google/gve/gve_tx.c 	tx->info = NULL;
tx                179 drivers/net/ethernet/google/gve/gve_tx.c 	struct gve_tx_ring *tx = &priv->tx[queue_idx];
tx                181 drivers/net/ethernet/google/gve/gve_tx.c 	block->tx = tx;
tx                182 drivers/net/ethernet/google/gve/gve_tx.c 	tx->ntfy_id = ntfy_idx;
tx                187 drivers/net/ethernet/google/gve/gve_tx.c 	struct gve_tx_ring *tx = &priv->tx[idx];
tx                193 drivers/net/ethernet/google/gve/gve_tx.c 	memset(tx, 0, sizeof(*tx));
tx                194 drivers/net/ethernet/google/gve/gve_tx.c 	tx->q_num = idx;
tx                196 drivers/net/ethernet/google/gve/gve_tx.c 	tx->mask = slots - 1;
tx                199 drivers/net/ethernet/google/gve/gve_tx.c 	tx->info = vzalloc(sizeof(*tx->info) * slots);
tx                200 drivers/net/ethernet/google/gve/gve_tx.c 	if (!tx->info)
tx                204 drivers/net/ethernet/google/gve/gve_tx.c 	bytes = sizeof(*tx->desc) * slots;
tx                205 drivers/net/ethernet/google/gve/gve_tx.c 	tx->desc = dma_alloc_coherent(hdev, bytes, &tx->bus, GFP_KERNEL);
tx                206 drivers/net/ethernet/google/gve/gve_tx.c 	if (!tx->desc)
tx                209 drivers/net/ethernet/google/gve/gve_tx.c 	tx->tx_fifo.qpl = gve_assign_tx_qpl(priv);
tx                212 drivers/net/ethernet/google/gve/gve_tx.c 	if (gve_tx_fifo_init(priv, &tx->tx_fifo))
tx                215 drivers/net/ethernet/google/gve/gve_tx.c 	tx->q_resources =
tx                217 drivers/net/ethernet/google/gve/gve_tx.c 				   sizeof(*tx->q_resources),
tx                218 drivers/net/ethernet/google/gve/gve_tx.c 				   &tx->q_resources_bus,
tx                220 drivers/net/ethernet/google/gve/gve_tx.c 	if (!tx->q_resources)
tx                224 drivers/net/ethernet/google/gve/gve_tx.c 		  (unsigned long)tx->bus);
tx                225 drivers/net/ethernet/google/gve/gve_tx.c 	tx->netdev_txq = netdev_get_tx_queue(priv->dev, idx);
tx                231 drivers/net/ethernet/google/gve/gve_tx.c 	gve_tx_fifo_release(priv, &tx->tx_fifo);
tx                233 drivers/net/ethernet/google/gve/gve_tx.c 	dma_free_coherent(hdev, bytes, tx->desc, tx->bus);
tx                234 drivers/net/ethernet/google/gve/gve_tx.c 	tx->desc = NULL;
tx                236 drivers/net/ethernet/google/gve/gve_tx.c 	vfree(tx->info);
tx                237 drivers/net/ethernet/google/gve/gve_tx.c 	tx->info = NULL;
tx                280 drivers/net/ethernet/google/gve/gve_tx.c static inline u32 gve_tx_avail(struct gve_tx_ring *tx)
tx                282 drivers/net/ethernet/google/gve/gve_tx.c 	return tx->mask + 1 - (tx->req - tx->done);
tx                285 drivers/net/ethernet/google/gve/gve_tx.c static inline int gve_skb_fifo_bytes_required(struct gve_tx_ring *tx,
tx                295 drivers/net/ethernet/google/gve/gve_tx.c 	pad_bytes = gve_tx_fifo_pad_alloc_one_frag(&tx->tx_fifo,
tx                313 drivers/net/ethernet/google/gve/gve_tx.c static inline bool gve_can_tx(struct gve_tx_ring *tx, int bytes_required)
tx                315 drivers/net/ethernet/google/gve/gve_tx.c 	return (gve_tx_avail(tx) >= MAX_TX_DESC_NEEDED &&
tx                316 drivers/net/ethernet/google/gve/gve_tx.c 		gve_tx_fifo_can_alloc(&tx->tx_fifo, bytes_required));
tx                320 drivers/net/ethernet/google/gve/gve_tx.c static int gve_maybe_stop_tx(struct gve_tx_ring *tx, struct sk_buff *skb)
tx                324 drivers/net/ethernet/google/gve/gve_tx.c 	bytes_required = gve_skb_fifo_bytes_required(tx, skb);
tx                325 drivers/net/ethernet/google/gve/gve_tx.c 	if (likely(gve_can_tx(tx, bytes_required)))
tx                329 drivers/net/ethernet/google/gve/gve_tx.c 	tx->stop_queue++;
tx                330 drivers/net/ethernet/google/gve/gve_tx.c 	netif_tx_stop_queue(tx->netdev_txq);
tx                345 drivers/net/ethernet/google/gve/gve_tx.c 	if (likely(!gve_can_tx(tx, bytes_required)))
tx                348 drivers/net/ethernet/google/gve/gve_tx.c 	netif_tx_start_queue(tx->netdev_txq);
tx                349 drivers/net/ethernet/google/gve/gve_tx.c 	tx->wake_queue++;
tx                407 drivers/net/ethernet/google/gve/gve_tx.c static int gve_tx_add_skb(struct gve_tx_ring *tx, struct sk_buff *skb,
tx                414 drivers/net/ethernet/google/gve/gve_tx.c 	u32 idx = tx->req & tx->mask;
tx                420 drivers/net/ethernet/google/gve/gve_tx.c 	info = &tx->info[idx];
tx                421 drivers/net/ethernet/google/gve/gve_tx.c 	pkt_desc = &tx->desc[idx];
tx                436 drivers/net/ethernet/google/gve/gve_tx.c 	pad_bytes = gve_tx_fifo_pad_alloc_one_frag(&tx->tx_fifo, hlen);
tx                437 drivers/net/ethernet/google/gve/gve_tx.c 	hdr_nfrags = gve_tx_alloc_fifo(&tx->tx_fifo, hlen + pad_bytes,
tx                440 drivers/net/ethernet/google/gve/gve_tx.c 	payload_nfrags = gve_tx_alloc_fifo(&tx->tx_fifo, skb->len - hlen,
tx                448 drivers/net/ethernet/google/gve/gve_tx.c 		      tx->tx_fifo.base + info->iov[hdr_nfrags - 1].iov_offset,
tx                450 drivers/net/ethernet/google/gve/gve_tx.c 	gve_dma_sync_for_device(dev, tx->tx_fifo.qpl->page_buses,
tx                456 drivers/net/ethernet/google/gve/gve_tx.c 		next_idx = (tx->req + 1 + i - payload_iov) & tx->mask;
tx                457 drivers/net/ethernet/google/gve/gve_tx.c 		seg_desc = &tx->desc[next_idx];
tx                464 drivers/net/ethernet/google/gve/gve_tx.c 			      tx->tx_fifo.base + info->iov[i].iov_offset,
tx                466 drivers/net/ethernet/google/gve/gve_tx.c 		gve_dma_sync_for_device(dev, tx->tx_fifo.qpl->page_buses,
tx                478 drivers/net/ethernet/google/gve/gve_tx.c 	struct gve_tx_ring *tx;
tx                483 drivers/net/ethernet/google/gve/gve_tx.c 	tx = &priv->tx[skb_get_queue_mapping(skb)];
tx                484 drivers/net/ethernet/google/gve/gve_tx.c 	if (unlikely(gve_maybe_stop_tx(tx, skb))) {
tx                490 drivers/net/ethernet/google/gve/gve_tx.c 		gve_tx_put_doorbell(priv, tx->q_resources, tx->req);
tx                493 drivers/net/ethernet/google/gve/gve_tx.c 	nsegs = gve_tx_add_skb(tx, skb, &priv->pdev->dev);
tx                495 drivers/net/ethernet/google/gve/gve_tx.c 	netdev_tx_sent_queue(tx->netdev_txq, skb->len);
tx                499 drivers/net/ethernet/google/gve/gve_tx.c 	tx->req += nsegs;
tx                501 drivers/net/ethernet/google/gve/gve_tx.c 	if (!netif_xmit_stopped(tx->netdev_txq) && netdev_xmit_more())
tx                504 drivers/net/ethernet/google/gve/gve_tx.c 	gve_tx_put_doorbell(priv, tx->q_resources, tx->req);
tx                510 drivers/net/ethernet/google/gve/gve_tx.c static int gve_clean_tx_done(struct gve_priv *priv, struct gve_tx_ring *tx,
tx                521 drivers/net/ethernet/google/gve/gve_tx.c 		idx = tx->done & tx->mask;
tx                524 drivers/net/ethernet/google/gve/gve_tx.c 			   tx->q_num, __func__, idx, tx->req, tx->done);
tx                525 drivers/net/ethernet/google/gve/gve_tx.c 		info = &tx->info[idx];
tx                542 drivers/net/ethernet/google/gve/gve_tx.c 		tx->done++;
tx                545 drivers/net/ethernet/google/gve/gve_tx.c 	gve_tx_free_fifo(&tx->tx_fifo, space_freed);
tx                546 drivers/net/ethernet/google/gve/gve_tx.c 	u64_stats_update_begin(&tx->statss);
tx                547 drivers/net/ethernet/google/gve/gve_tx.c 	tx->bytes_done += bytes;
tx                548 drivers/net/ethernet/google/gve/gve_tx.c 	tx->pkt_done += pkts;
tx                549 drivers/net/ethernet/google/gve/gve_tx.c 	u64_stats_update_end(&tx->statss);
tx                550 drivers/net/ethernet/google/gve/gve_tx.c 	netdev_tx_completed_queue(tx->netdev_txq, pkts, bytes);
tx                557 drivers/net/ethernet/google/gve/gve_tx.c 	if (try_to_wake && netif_tx_queue_stopped(tx->netdev_txq) &&
tx                558 drivers/net/ethernet/google/gve/gve_tx.c 	    likely(gve_can_tx(tx, GVE_TX_START_THRESH))) {
tx                559 drivers/net/ethernet/google/gve/gve_tx.c 		tx->wake_queue++;
tx                560 drivers/net/ethernet/google/gve/gve_tx.c 		netif_tx_wake_queue(tx->netdev_txq);
tx                567 drivers/net/ethernet/google/gve/gve_tx.c 				 struct gve_tx_ring *tx)
tx                569 drivers/net/ethernet/google/gve/gve_tx.c 	u32 counter_index = be32_to_cpu((tx->q_resources->counter_index));
tx                577 drivers/net/ethernet/google/gve/gve_tx.c 	struct gve_tx_ring *tx = block->tx;
tx                587 drivers/net/ethernet/google/gve/gve_tx.c 	tx->last_nic_done = gve_tx_load_event_counter(priv, tx);
tx                588 drivers/net/ethernet/google/gve/gve_tx.c 	nic_done = be32_to_cpu(tx->last_nic_done);
tx                593 drivers/net/ethernet/google/gve/gve_tx.c 		to_do = min_t(u32, (nic_done - tx->done), budget);
tx                594 drivers/net/ethernet/google/gve/gve_tx.c 		gve_clean_tx_done(priv, tx, to_do, true);
tx                597 drivers/net/ethernet/google/gve/gve_tx.c 	repoll |= (nic_done != tx->done);
tx                330 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c static void hix5hd2_set_desc_depth(struct hix5hd2_priv *priv, int rx, int tx)
tx                341 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c 	writel_relaxed(tx << 3, priv->base + TX_BQ_DEPTH);
tx                345 drivers/net/ethernet/hisilicon/hix5hd2_gmac.c 	writel_relaxed(tx << 3, priv->base + TX_RQ_DEPTH);
tx                194 drivers/net/ethernet/hisilicon/hns/hnae.h 		} tx;
tx                 96 drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c static void hns_gmac_get_en(void *mac_drv, u32 *rx, u32 *tx)
tx                102 drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c 	*tx = dsaf_get_bit(porten, GMAC_PORT_TX_EN_B);
tx                511 drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c 	u32 tx;
tx                522 drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c 	hns_gmac_get_en(mac_drv, &rx, &tx);
tx                523 drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c 	mac_info->port_en = rx && tx;
tx                 60 drivers/net/ethernet/hisilicon/hns/hns_enet.c 	desc->tx.send_size = cpu_to_le16((u16)send_sz);
tx                113 drivers/net/ethernet/hisilicon/hns/hns_enet.c 			desc->tx.ip_offset = ip_offset;
tx                114 drivers/net/ethernet/hisilicon/hns/hns_enet.c 			desc->tx.tse_vlan_snap_v6_sctp_nth = tvsvsn;
tx                115 drivers/net/ethernet/hisilicon/hns/hns_enet.c 			desc->tx.mss = cpu_to_le16(mss);
tx                116 drivers/net/ethernet/hisilicon/hns/hns_enet.c 			desc->tx.l4_len = l4_len;
tx                117 drivers/net/ethernet/hisilicon/hns/hns_enet.c 			desc->tx.paylen = cpu_to_le16(paylen);
tx                123 drivers/net/ethernet/hisilicon/hns/hns_enet.c 	desc->tx.bn_pid = bn_pid;
tx                124 drivers/net/ethernet/hisilicon/hns/hns_enet.c 	desc->tx.ra_ri_cs_fe_vld = rrcfv;
tx                162 drivers/net/ethernet/hisilicon/hns/hns_enet.c 	desc->tx.send_size = cpu_to_le16((u16)size);
tx                199 drivers/net/ethernet/hisilicon/hns/hns_enet.c 	desc->tx.asid_bufnum_pid = cpu_to_le16(asid_bufnum_pid);
tx                200 drivers/net/ethernet/hisilicon/hns/hns_enet.c 	desc->tx.flag_ipoffset = cpu_to_le32(flag_ipoffset);
tx                201 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c 	dev_info(dev, "(TX)vlan_tag: %u\n", tx_desc->tx.vlan_tag);
tx                202 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c 	dev_info(dev, "(TX)send_size: %u\n", tx_desc->tx.send_size);
tx                203 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c 	dev_info(dev, "(TX)vlan_tso: %u\n", tx_desc->tx.type_cs_vlan_tso);
tx                204 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c 	dev_info(dev, "(TX)l2_len: %u\n", tx_desc->tx.l2_len);
tx                205 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c 	dev_info(dev, "(TX)l3_len: %u\n", tx_desc->tx.l3_len);
tx                206 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c 	dev_info(dev, "(TX)l4_len: %u\n", tx_desc->tx.l4_len);
tx                207 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c 	dev_info(dev, "(TX)vlan_tag: %u\n", tx_desc->tx.outer_vlan_tag);
tx                208 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c 	dev_info(dev, "(TX)tv: %u\n", tx_desc->tx.tv);
tx                209 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c 	dev_info(dev, "(TX)vlan_msec: %u\n", tx_desc->tx.ol_type_vlan_msec);
tx                210 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c 	dev_info(dev, "(TX)ol2_len: %u\n", tx_desc->tx.ol2_len);
tx                211 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c 	dev_info(dev, "(TX)ol3_len: %u\n", tx_desc->tx.ol3_len);
tx                212 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c 	dev_info(dev, "(TX)ol4_len: %u\n", tx_desc->tx.ol4_len);
tx                213 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c 	dev_info(dev, "(TX)paylen: %u\n", tx_desc->tx.paylen);
tx                214 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c 	dev_info(dev, "(TX)vld_ra_ri: %u\n", tx_desc->tx.bdtp_fe_sc_vld_ra_ri);
tx                215 drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c 	dev_info(dev, "(TX)mss: %u\n", tx_desc->tx.mss);
tx               1095 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	desc->tx.ol_type_vlan_len_msec =
tx               1097 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	desc->tx.type_cs_vlan_tso_len = cpu_to_le32(type_cs_vlan_tso);
tx               1098 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	desc->tx.paylen = cpu_to_le32(paylen);
tx               1099 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	desc->tx.mss = cpu_to_le16(mss);
tx               1100 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	desc->tx.vlan_tag = cpu_to_le16(inner_vtag);
tx               1101 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 	desc->tx.outer_vlan_tag = cpu_to_le16(out_vtag);
tx               1148 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 		desc->tx.send_size = cpu_to_le16(size);
tx               1150 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 		desc->tx.bdtp_fe_sc_vld_ra_ri =
tx               1173 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 		desc->tx.send_size = cpu_to_le16((k == frag_buf_num - 1) ?
tx               1178 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 		desc->tx.bdtp_fe_sc_vld_ra_ri =
tx               4089 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c 		ring->desc[ring->next_to_clean].tx.bdtp_fe_sc_vld_ra_ri = 0;
tx                277 drivers/net/ethernet/hisilicon/hns3/hns3_enet.h 		} tx;
tx                161 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_tm.c int hclge_mac_pause_en_cfg(struct hclge_dev *hdev, bool tx, bool rx)
tx                167 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_tm.c 	desc.data[0] = cpu_to_le32((tx ? HCLGE_TX_MAC_PAUSE_EN_MSK : 0) |
tx                153 drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_tm.h int hclge_mac_pause_en_cfg(struct hclge_dev *hdev, bool tx, bool rx);
tx                673 drivers/net/ethernet/i825xx/ether1.c 	tx_t tx;
tx                701 drivers/net/ethernet/i825xx/ether1.c 	tx.tx_status = 0;
tx                702 drivers/net/ethernet/i825xx/ether1.c 	tx.tx_command = CMD_TX | CMD_INTR;
tx                703 drivers/net/ethernet/i825xx/ether1.c 	tx.tx_link = nopaddr;
tx                704 drivers/net/ethernet/i825xx/ether1.c 	tx.tx_tbdoffset = tbdaddr;
tx                714 drivers/net/ethernet/i825xx/ether1.c 	ether1_writebuffer (dev, &tx, txaddr, TX_SIZE);
tx                189 drivers/net/ethernet/intel/fm10k/fm10k.h 	struct fm10k_ring_container rx, tx;
tx                165 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c 	for (i = 0; i < q_vector->tx.count; i++) {
tx                166 drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c 		struct fm10k_ring *ring = &q_vector->tx.ring[i];
tx                680 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c 		qv->tx.itr = tx_itr;
tx               1196 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	unsigned int budget = q_vector->tx.work_limit;
tx               1284 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	q_vector->tx.total_bytes += total_bytes;
tx               1285 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	q_vector->tx.total_packets += total_packets;
tx               1416 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	fm10k_update_itr(&q_vector->tx);
tx               1422 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	itr |= (q_vector->tx.itr & FM10K_ITR_MAX);
tx               1439 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	fm10k_for_each_ring(ring, q_vector->tx) {
tx               1628 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	q_vector->tx.ring = ring;
tx               1629 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	q_vector->tx.work_limit = FM10K_DEFAULT_TX_WORK;
tx               1630 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	q_vector->tx.itr = interface->tx_itr;
tx               1631 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	q_vector->tx.itr_scale = interface->hw.mac.itr_scale;
tx               1632 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	q_vector->tx.count = txr_count;
tx               1708 drivers/net/ethernet/intel/fm10k/fm10k_main.c 	fm10k_for_each_ring(ring, q_vector->tx)
tx                333 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 	struct fm10k_mbx_fifo *fifo = &mbx->tx;
tx                386 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 	struct fm10k_mbx_fifo *fifo = &mbx->tx;
tx                622 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 	head = fm10k_fifo_head_offset(&mbx->tx, mbx->pulled);
tx                625 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 	mbx->local = fm10k_fifo_crc(&mbx->tx, head, len, mbx->local);
tx                678 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 	u16 fifo_unused = fm10k_fifo_unused(&mbx->tx);
tx                691 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 	return fm10k_fifo_empty(&mbx->tx);
tx                754 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 	err = fm10k_fifo_enqueue(&mbx->tx, msg);
tx                761 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 		err = fm10k_fifo_enqueue(&mbx->tx, msg);
tx                856 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 	struct fm10k_mbx_fifo *fifo = &mbx->tx;
tx               1070 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 	while (fm10k_fifo_head_len(&mbx->tx) && mbx->pulled) {
tx               1071 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 		len = fm10k_fifo_head_drop(&mbx->tx);
tx               1106 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 	for (len = fm10k_fifo_head_len(&mbx->tx);
tx               1108 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 	     len = fm10k_fifo_head_len(&mbx->tx)) {
tx               1109 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 		fm10k_fifo_head_drop(&mbx->tx);
tx               1417 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 	fm10k_fifo_drop_all(&mbx->tx);
tx               1607 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 	fm10k_fifo_init(&mbx->tx, mbx->buffer, FM10K_MBX_TX_BUFFER_SIZE);
tx               1758 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 	fm10k_fifo_drop_all(&mbx->tx);
tx               1920 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 	struct fm10k_mbx_fifo *fifo = &mbx->tx;
tx               2165 drivers/net/ethernet/intel/fm10k/fm10k_mbx.c 	fm10k_fifo_init(&mbx->tx, mbx->buffer, FM10K_MBX_TX_BUFFER_SIZE);
tx                257 drivers/net/ethernet/intel/fm10k/fm10k_mbx.h 	struct fm10k_mbx_fifo tx;
tx                722 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 			if (!qv->tx.count && !qv->rx.count)
tx               1187 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 	if (q_vector->rx.count || q_vector->tx.count)
tx               1734 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 		if (!q_vector->tx.count && !q_vector->rx.count)
tx               1768 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 		if (q_vector->tx.count && q_vector->rx.count) {
tx               1775 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 		} else if (q_vector->tx.count) {
tx               1818 drivers/net/ethernet/intel/fm10k/fm10k_pci.c 		if (!q_vector->tx.count && !q_vector->rx.count)
tx                864 drivers/net/ethernet/intel/i40e/i40e.h 	struct i40e_ring_container tx;
tx               2839 drivers/net/ethernet/intel/i40e/i40e_ethtool.c 	q_vector->tx.target_itr = ITR_TO_REG(tx_ring->itr_setting);
tx               3540 drivers/net/ethernet/intel/i40e/i40e_main.c 		q_vector->tx.next_update = jiffies + 1;
tx               3541 drivers/net/ethernet/intel/i40e/i40e_main.c 		q_vector->tx.target_itr =
tx               3544 drivers/net/ethernet/intel/i40e/i40e_main.c 		     q_vector->tx.target_itr >> 1);
tx               3545 drivers/net/ethernet/intel/i40e/i40e_main.c 		q_vector->tx.current_itr = q_vector->tx.target_itr;
tx               3651 drivers/net/ethernet/intel/i40e/i40e_main.c 	q_vector->tx.next_update = jiffies + 1;
tx               3652 drivers/net/ethernet/intel/i40e/i40e_main.c 	q_vector->tx.target_itr = ITR_TO_REG(vsi->tx_rings[0]->itr_setting);
tx               3653 drivers/net/ethernet/intel/i40e/i40e_main.c 	wr32(hw, I40E_PFINT_ITR0(I40E_TX_ITR), q_vector->tx.target_itr >> 1);
tx               3654 drivers/net/ethernet/intel/i40e/i40e_main.c 	q_vector->tx.current_itr = q_vector->tx.target_itr;
tx               3725 drivers/net/ethernet/intel/i40e/i40e_main.c 	if (!q_vector->tx.ring && !q_vector->rx.ring)
tx               3783 drivers/net/ethernet/intel/i40e/i40e_main.c 		if (q_vector->tx.ring && q_vector->rx.ring) {
tx               3790 drivers/net/ethernet/intel/i40e/i40e_main.c 		} else if (q_vector->tx.ring) {
tx               4146 drivers/net/ethernet/intel/i40e/i40e_main.c 	if (!q_vector->tx.ring)
tx               4149 drivers/net/ethernet/intel/i40e/i40e_main.c 	vsi = q_vector->tx.ring->vsi;
tx               4150 drivers/net/ethernet/intel/i40e/i40e_main.c 	i40e_clean_fdir_tx_irq(q_vector->tx.ring, vsi->work_limit);
tx               4168 drivers/net/ethernet/intel/i40e/i40e_main.c 	tx_ring->next = q_vector->tx.ring;
tx               4169 drivers/net/ethernet/intel/i40e/i40e_main.c 	q_vector->tx.ring = tx_ring;
tx               4170 drivers/net/ethernet/intel/i40e/i40e_main.c 	q_vector->tx.count++;
tx               4177 drivers/net/ethernet/intel/i40e/i40e_main.c 		xdp_ring->next = q_vector->tx.ring;
tx               4178 drivers/net/ethernet/intel/i40e/i40e_main.c 		q_vector->tx.ring = xdp_ring;
tx               4179 drivers/net/ethernet/intel/i40e/i40e_main.c 		q_vector->tx.count++;
tx               4221 drivers/net/ethernet/intel/i40e/i40e_main.c 		q_vector->tx.count = 0;
tx               4223 drivers/net/ethernet/intel/i40e/i40e_main.c 		q_vector->tx.ring = NULL;
tx               4733 drivers/net/ethernet/intel/i40e/i40e_main.c 	i40e_for_each_ring(ring, q_vector->tx)
tx               4819 drivers/net/ethernet/intel/i40e/i40e_main.c 		if (q_vector->rx.ring || q_vector->tx.ring)
tx               4838 drivers/net/ethernet/intel/i40e/i40e_main.c 		if (q_vector->rx.ring || q_vector->tx.ring)
tx               12638 drivers/net/ethernet/intel/i40e/i40e_main.c 	if (q_vector->rx.ring || q_vector->tx.ring) {
tx                908 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	u16 flags = q_vector->tx.ring[0].flags;
tx               1053 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		    (q_vector->tx.target_itr & I40E_ITR_ADAPTIVE_LATENCY)) {
tx               1092 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		itr = min(q_vector->tx.current_itr, q_vector->rx.current_itr);
tx               2515 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	i40e_update_itr(q_vector, &q_vector->tx);
tx               2532 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	} else if ((q_vector->tx.target_itr < q_vector->tx.current_itr) ||
tx               2534 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		    (q_vector->tx.target_itr - q_vector->tx.current_itr))) {
tx               2539 drivers/net/ethernet/intel/i40e/i40e_txrx.c 					   q_vector->tx.target_itr);
tx               2540 drivers/net/ethernet/intel/i40e/i40e_txrx.c 		q_vector->tx.current_itr = q_vector->tx.target_itr;
tx               2587 drivers/net/ethernet/intel/i40e/i40e_txrx.c 	i40e_for_each_ring(ring, q_vector->tx) {
tx               2643 drivers/net/ethernet/intel/i40e/i40e_txrx.c 			q_vector->tx.ring[0].tx_stats.tx_force_wb++;
tx                 55 drivers/net/ethernet/intel/i40e/i40e_txrx_common.h 	tx_ring->q_vector->tx.total_bytes += total_bytes;
tx                 56 drivers/net/ethernet/intel/i40e/i40e_txrx_common.h 	tx_ring->q_vector->tx.total_packets += total_packets;
tx                 98 drivers/net/ethernet/intel/iavf/iavf.h 	struct iavf_ring_container tx;
tx                734 drivers/net/ethernet/intel/iavf/iavf_ethtool.c 	q_vector->tx.target_itr = ITR_TO_REG(tx_ring->itr_setting);
tx                287 drivers/net/ethernet/intel/iavf/iavf_main.c 	if (!q_vector->tx.ring && !q_vector->rx.ring)
tx                335 drivers/net/ethernet/intel/iavf/iavf_main.c 	tx_ring->next = q_vector->tx.ring;
tx                337 drivers/net/ethernet/intel/iavf/iavf_main.c 	q_vector->tx.ring = tx_ring;
tx                338 drivers/net/ethernet/intel/iavf/iavf_main.c 	q_vector->tx.count++;
tx                339 drivers/net/ethernet/intel/iavf/iavf_main.c 	q_vector->tx.next_update = jiffies + 1;
tx                340 drivers/net/ethernet/intel/iavf/iavf_main.c 	q_vector->tx.target_itr = ITR_TO_REG(tx_ring->itr_setting);
tx                343 drivers/net/ethernet/intel/iavf/iavf_main.c 	     q_vector->tx.target_itr >> 1);
tx                344 drivers/net/ethernet/intel/iavf/iavf_main.c 	q_vector->tx.current_itr = q_vector->tx.target_itr;
tx                431 drivers/net/ethernet/intel/iavf/iavf_main.c 		if (q_vector->tx.ring && q_vector->rx.ring) {
tx                438 drivers/net/ethernet/intel/iavf/iavf_main.c 		} else if (q_vector->tx.ring) {
tx                285 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	tx_ring->q_vector->tx.total_bytes += total_bytes;
tx                286 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	tx_ring->q_vector->tx.total_packets += total_packets;
tx                335 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	u16 flags = q_vector->tx.ring[0].flags;
tx                463 drivers/net/ethernet/intel/iavf/iavf_txrx.c 		    (q_vector->tx.target_itr & IAVF_ITR_ADAPTIVE_LATENCY)) {
tx                502 drivers/net/ethernet/intel/iavf/iavf_txrx.c 		itr = min(q_vector->tx.current_itr, q_vector->rx.current_itr);
tx               1650 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	iavf_update_itr(q_vector, &q_vector->tx);
tx               1667 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	} else if ((q_vector->tx.target_itr < q_vector->tx.current_itr) ||
tx               1669 drivers/net/ethernet/intel/iavf/iavf_txrx.c 		    (q_vector->tx.target_itr - q_vector->tx.current_itr))) {
tx               1674 drivers/net/ethernet/intel/iavf/iavf_txrx.c 					   q_vector->tx.target_itr);
tx               1675 drivers/net/ethernet/intel/iavf/iavf_txrx.c 		q_vector->tx.current_itr = q_vector->tx.target_itr;
tx               1722 drivers/net/ethernet/intel/iavf/iavf_txrx.c 	iavf_for_each_ring(ring, q_vector->tx) {
tx               1772 drivers/net/ethernet/intel/iavf/iavf_txrx.c 			q_vector->tx.ring[0].tx_stats.tx_force_wb++;
tx                296 drivers/net/ethernet/intel/ice/ice.h 	struct ice_ring_container tx;
tx               3152 drivers/net/ethernet/intel/ice/ice_ethtool.c 					&vsi->tx_rings[q_num]->q_vector->tx))
tx               3160 drivers/net/ethernet/intel/ice/ice_ethtool.c 					&vsi->tx_rings[q_num]->q_vector->tx))
tx               3327 drivers/net/ethernet/intel/ice/ice_ethtool.c 					&vsi->tx_rings[q_num]->q_vector->tx,
tx               3337 drivers/net/ethernet/intel/ice/ice_ethtool.c 					&vsi->tx_rings[q_num]->q_vector->tx,
tx                523 drivers/net/ethernet/intel/ice/ice_lib.c 	if (!q_vector->tx.ring && !q_vector->rx.ring)
tx               1118 drivers/net/ethernet/intel/ice/ice_lib.c 	ice_for_each_ring(ring, q_vector->tx)
tx               1372 drivers/net/ethernet/intel/ice/ice_lib.c 		q_vector->tx.ring = NULL;
tx               1373 drivers/net/ethernet/intel/ice/ice_lib.c 		q_vector->tx.itr_idx = ICE_TX_ITR;
tx               1380 drivers/net/ethernet/intel/ice/ice_lib.c 			tx_ring->next = q_vector->tx.ring;
tx               1381 drivers/net/ethernet/intel/ice/ice_lib.c 			q_vector->tx.ring = tx_ring;
tx               1901 drivers/net/ethernet/intel/ice/ice_lib.c 		struct ice_ring_container *rc = &q_vector->tx;
tx               2013 drivers/net/ethernet/intel/ice/ice_lib.c 					      q_vector->tx.itr_idx);
tx               2457 drivers/net/ethernet/intel/ice/ice_lib.c void ice_cfg_sw_lldp(struct ice_vsi *vsi, bool tx, bool create)
tx               2472 drivers/net/ethernet/intel/ice/ice_lib.c 	if (tx) {
tx               2493 drivers/net/ethernet/intel/ice/ice_lib.c 			create ? "adding" : "removing", tx ? "TX" : "RX",
tx                 72 drivers/net/ethernet/intel/ice/ice_lib.h void ice_cfg_sw_lldp(struct ice_vsi *vsi, bool tx, bool create);
tx               1617 drivers/net/ethernet/intel/ice/ice_main.c 		if (q_vector->tx.ring && q_vector->rx.ring) {
tx               1624 drivers/net/ethernet/intel/ice/ice_main.c 		} else if (q_vector->tx.ring) {
tx               3528 drivers/net/ethernet/intel/ice/ice_main.c 		if (q_vector->rx.ring || q_vector->tx.ring)
tx               3895 drivers/net/ethernet/intel/ice/ice_main.c 		if (q_vector->rx.ring || q_vector->tx.ring)
tx                195 drivers/net/ethernet/intel/ice/ice_txrx.c 	tx_ring->q_vector->tx.total_bytes += total_bytes;
tx                196 drivers/net/ethernet/intel/ice/ice_txrx.c 	tx_ring->q_vector->tx.total_pkts += total_pkts;
tx               1241 drivers/net/ethernet/intel/ice/ice_txrx.c 		    (q_vector->tx.target_itr & ICE_ITR_ADAPTIVE_LATENCY)) {
tx               1280 drivers/net/ethernet/intel/ice/ice_txrx.c 		itr = min(q_vector->tx.current_itr, q_vector->rx.current_itr);
tx               1364 drivers/net/ethernet/intel/ice/ice_txrx.c 	struct ice_ring_container *tx = &q_vector->tx;
tx               1387 drivers/net/ethernet/intel/ice/ice_txrx.c 	ice_update_itr(q_vector, tx);
tx               1403 drivers/net/ethernet/intel/ice/ice_txrx.c 	} else if ((tx->target_itr < tx->current_itr) ||
tx               1405 drivers/net/ethernet/intel/ice/ice_txrx.c 		    (tx->target_itr - tx->current_itr))) {
tx               1409 drivers/net/ethernet/intel/ice/ice_txrx.c 		itr_val = ice_buildreg_itr(tx->itr_idx, tx->target_itr);
tx               1410 drivers/net/ethernet/intel/ice/ice_txrx.c 		tx->current_itr = tx->target_itr;
tx               1486 drivers/net/ethernet/intel/ice/ice_txrx.c 	ice_for_each_ring(ring, q_vector->tx)
tx               2209 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c 			q_vector->tx.itr_idx = map->txitr_idx;
tx               2212 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c 					      q_vector->tx.itr_idx);
tx                302 drivers/net/ethernet/intel/igb/igb.h 	struct igb_ring_container rx, tx;
tx               2236 drivers/net/ethernet/intel/igb/igb_ethtool.c 		q_vector->tx.work_limit = adapter->tx_work_limit;
tx                800 drivers/net/ethernet/intel/igb/igb_main.c 	if (q_vector->tx.ring)
tx                801 drivers/net/ethernet/intel/igb/igb_main.c 		tx_queue = q_vector->tx.ring->reg_idx;
tx                958 drivers/net/ethernet/intel/igb/igb_main.c 		if (q_vector->rx.ring && q_vector->tx.ring)
tx                961 drivers/net/ethernet/intel/igb/igb_main.c 		else if (q_vector->tx.ring)
tx                963 drivers/net/ethernet/intel/igb/igb_main.c 				q_vector->tx.ring->queue_index);
tx               1031 drivers/net/ethernet/intel/igb/igb_main.c 	if (q_vector->tx.ring)
tx               1032 drivers/net/ethernet/intel/igb/igb_main.c 		adapter->tx_ring[q_vector->tx.ring->queue_index] = NULL;
tx               1226 drivers/net/ethernet/intel/igb/igb_main.c 	q_vector->tx.work_limit = adapter->tx_work_limit;
tx               1255 drivers/net/ethernet/intel/igb/igb_main.c 		igb_add_ring(ring, &q_vector->tx);
tx               5482 drivers/net/ethernet/intel/igb/igb_main.c 	packets = q_vector->tx.total_packets;
tx               5485 drivers/net/ethernet/intel/igb/igb_main.c 				      q_vector->tx.total_bytes / packets);
tx               5517 drivers/net/ethernet/intel/igb/igb_main.c 	q_vector->tx.total_bytes = 0;
tx               5518 drivers/net/ethernet/intel/igb/igb_main.c 	q_vector->tx.total_packets = 0;
tx               5602 drivers/net/ethernet/intel/igb/igb_main.c 	igb_update_itr(q_vector, &q_vector->tx);
tx               5605 drivers/net/ethernet/intel/igb/igb_main.c 	current_itr = max(q_vector->rx.itr, q_vector->tx.itr);
tx               6642 drivers/net/ethernet/intel/igb/igb_main.c 	if (q_vector->tx.ring)
tx               6643 drivers/net/ethernet/intel/igb/igb_main.c 		igb_update_tx_dca(adapter, q_vector->tx.ring, cpu);
tx               7714 drivers/net/ethernet/intel/igb/igb_main.c 	if (q_vector->tx.ring)
tx               7748 drivers/net/ethernet/intel/igb/igb_main.c 	struct igb_ring *tx_ring = q_vector->tx.ring;
tx               7752 drivers/net/ethernet/intel/igb/igb_main.c 	unsigned int budget = q_vector->tx.work_limit;
tx               7841 drivers/net/ethernet/intel/igb/igb_main.c 	q_vector->tx.total_bytes += total_bytes;
tx               7842 drivers/net/ethernet/intel/igb/igb_main.c 	q_vector->tx.total_packets += total_packets;
tx                292 drivers/net/ethernet/intel/igc/igc.h 	struct igc_ring_container rx, tx;
tx                864 drivers/net/ethernet/intel/igc/igc_ethtool.c 		q_vector->tx.work_limit = adapter->tx_work_limit;
tx               1705 drivers/net/ethernet/intel/igc/igc_main.c 	unsigned int budget = q_vector->tx.work_limit;
tx               1706 drivers/net/ethernet/intel/igc/igc_main.c 	struct igc_ring *tx_ring = q_vector->tx.ring;
tx               1798 drivers/net/ethernet/intel/igc/igc_main.c 	q_vector->tx.total_bytes += total_bytes;
tx               1799 drivers/net/ethernet/intel/igc/igc_main.c 	q_vector->tx.total_packets += total_packets;
tx               2600 drivers/net/ethernet/intel/igc/igc_main.c 	if (q_vector->tx.ring)
tx               2601 drivers/net/ethernet/intel/igc/igc_main.c 		tx_queue = q_vector->tx.ring->reg_idx;
tx               2707 drivers/net/ethernet/intel/igc/igc_main.c 		if (q_vector->rx.ring && q_vector->tx.ring)
tx               2710 drivers/net/ethernet/intel/igc/igc_main.c 		else if (q_vector->tx.ring)
tx               2712 drivers/net/ethernet/intel/igc/igc_main.c 				q_vector->tx.ring->queue_index);
tx               2760 drivers/net/ethernet/intel/igc/igc_main.c 	if (q_vector->tx.ring)
tx               2761 drivers/net/ethernet/intel/igc/igc_main.c 		adapter->tx_ring[q_vector->tx.ring->queue_index] = NULL;
tx               3112 drivers/net/ethernet/intel/igc/igc_main.c 	packets = q_vector->tx.total_packets;
tx               3115 drivers/net/ethernet/intel/igc/igc_main.c 				      q_vector->tx.total_bytes / packets);
tx               3147 drivers/net/ethernet/intel/igc/igc_main.c 	q_vector->tx.total_bytes = 0;
tx               3148 drivers/net/ethernet/intel/igc/igc_main.c 	q_vector->tx.total_packets = 0;
tx               3312 drivers/net/ethernet/intel/igc/igc_main.c 	igc_update_itr(q_vector, &q_vector->tx);
tx               3315 drivers/net/ethernet/intel/igc/igc_main.c 	current_itr = max(q_vector->rx.itr, q_vector->tx.itr);
tx               3393 drivers/net/ethernet/intel/igc/igc_main.c 	if (q_vector->tx.ring)
tx               3539 drivers/net/ethernet/intel/igc/igc_main.c 	q_vector->tx.work_limit = adapter->tx_work_limit;
tx               3568 drivers/net/ethernet/intel/igc/igc_main.c 		igc_add_ring(ring, &q_vector->tx);
tx                456 drivers/net/ethernet/intel/ixgbe/ixgbe.h 	struct ixgbe_ring_container rx, tx;
tx                 30 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c 	int tx = DCB_TX_CONFIG;
tx                 48 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c 		if (dst->path[tx].prio_type != src->path[tx].prio_type) {
tx                 49 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c 			dst->path[tx].prio_type = src->path[tx].prio_type;
tx                 53 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c 		if (dst->path[tx].bwg_id != src->path[tx].bwg_id) {
tx                 54 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c 			dst->path[tx].bwg_id = src->path[tx].bwg_id;
tx                 58 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c 		if (dst->path[tx].bwg_percent != src->path[tx].bwg_percent) {
tx                 59 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c 			dst->path[tx].bwg_percent = src->path[tx].bwg_percent;
tx                 63 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c 		if (dst->path[tx].up_to_tc_bitmap !=
tx                 64 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c 				src->path[tx].up_to_tc_bitmap) {
tx                 65 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c 			dst->path[tx].up_to_tc_bitmap =
tx                 66 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c 				src->path[tx].up_to_tc_bitmap;
tx                 95 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c 		if (dcfg->bw_percentage[tx][j] != scfg->bw_percentage[tx][j]) {
tx                 96 drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c 			dcfg->bw_percentage[tx][j] = scfg->bw_percentage[tx][j];
tx               2285 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c 	if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count)
tx               2336 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c 	if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count) {
tx               2370 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c 	if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count)
tx               2390 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c 		if (q_vector->tx.count && !q_vector->rx.count)
tx                 91 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 				    unsigned int *tx, unsigned int *rx)
tx                 96 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 	*tx = 0;
tx                102 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 		*tx = tc << 2; /* 0, 4,  8, 12, 16, 20, 24, 28 */
tx                118 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 				*tx = tc << 5;		/*   0,  32,  64 */
tx                120 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 				*tx = (tc + 2) << 4;	/*  80,  96 */
tx                122 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 				*tx = (tc + 8) << 3;	/* 104, 112, 120 */
tx                131 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 				*tx = tc << 6;		/*  0,  64 */
tx                133 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 				*tx = (tc + 4) << 4;	/* 96, 112 */
tx                884 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 	q_vector->tx.work_limit = adapter->tx_work_limit;
tx                887 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 	q_vector->tx.itr = IXGBE_ITR_ADAPTIVE_MAX_USECS |
tx                919 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 		ixgbe_add_ring(ring, &q_vector->tx);
tx                945 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 		ixgbe_add_ring(ring, &q_vector->tx);
tx               1023 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c 	ixgbe_for_each_ring(ring, q_vector->tx) {
tx               1119 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	unsigned int budget = q_vector->tx.work_limit;
tx               1211 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	q_vector->tx.total_bytes += total_bytes;
tx               1212 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	q_vector->tx.total_packets += total_packets;
tx               1352 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	ixgbe_for_each_ring(ring, q_vector->tx)
tx               2459 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		ixgbe_for_each_ring(ring, q_vector->tx)
tx               2720 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	ixgbe_update_itr(q_vector, &q_vector->tx);
tx               2724 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	new_itr = min(q_vector->rx.itr, q_vector->tx.itr);
tx               3143 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	if (q_vector->rx.ring || q_vector->tx.ring)
tx               3170 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 	ixgbe_for_each_ring(ring, q_vector->tx) {
tx               3235 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		if (q_vector->tx.ring && q_vector->rx.ring) {
tx               3242 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		} else if (q_vector->tx.ring) {
tx               3406 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 		if (!q_vector->rx.ring && !q_vector->tx.ring)
tx               7347 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 			if (qv->rx.ring || qv->tx.ring)
tx                686 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	q_vector->tx.total_bytes += total_bytes;
tx                687 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	q_vector->tx.total_packets += total_packets;
tx                695 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 	return ixgbe_xmit_zc(tx_ring, q_vector->tx.work_limit);
tx                803 drivers/net/ethernet/intel/ixgbevf/ethtool.c 	if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count)
tx                824 drivers/net/ethernet/intel/ixgbevf/ethtool.c 	if (adapter->q_vector[0]->tx.count &&
tx                856 drivers/net/ethernet/intel/ixgbevf/ethtool.c 		if (q_vector->tx.count && !q_vector->rx.count)
tx                239 drivers/net/ethernet/intel/ixgbevf/ixgbevf.h 	struct ixgbevf_ring_container rx, tx;
tx                365 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	q_vector->tx.total_bytes += total_bytes;
tx                366 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	q_vector->tx.total_packets += total_packets;
tx               1271 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	ixgbevf_for_each_ring(ring, q_vector->tx) {
tx               1359 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		ixgbevf_for_each_ring(ring, q_vector->tx)
tx               1362 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		if (q_vector->tx.ring && !q_vector->rx.ring) {
tx               1462 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	ixgbevf_update_itr(q_vector, &q_vector->tx);
tx               1465 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	current_itr = max(q_vector->rx.itr, q_vector->tx.itr);
tx               1518 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	if (q_vector->rx.ring || q_vector->tx.ring)
tx               1542 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		if (q_vector->tx.ring && q_vector->rx.ring) {
tx               1549 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		} else if (q_vector->tx.ring) {
tx               1629 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		    !adapter->q_vector[i]->tx.ring)
tx               2732 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		ixgbevf_add_ring(ring, &q_vector->tx);
tx               2760 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		ixgbevf_add_ring(ring, &q_vector->tx);
tx               2824 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 	ixgbevf_for_each_ring(ring, q_vector->tx) {
tx               3207 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c 		if (qv->rx.ring || qv->tx.ring)
tx                558 drivers/net/ethernet/marvell/octeontx2/af/mbox.h 		} tx;
tx               2226 drivers/net/ethernet/marvell/sky2.c static enum flow_control sky2_flow(int rx, int tx)
tx               2229 drivers/net/ethernet/marvell/sky2.c 		return tx ? FC_BOTH : FC_RX;
tx               2231 drivers/net/ethernet/marvell/sky2.c 		return tx ? FC_TX : FC_NONE;
tx                 72 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c 		struct mlx5e_ipsec_tx_metadata tx;
tx                223 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c 		mdata->content.tx.mss_inv = mlx5e_ipsec_mss_inv(skb);
tx                224 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c 		mdata->content.tx.seq = htons(ntohl(tcph->seq) & 0xFFFF);
tx                228 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c 	mdata->content.tx.esp_next_proto = xo->proto;
tx                231 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c 		   mdata->syndrome, mdata->content.tx.esp_next_proto,
tx                232 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c 		   ntohs(mdata->content.tx.mss_inv),
tx                233 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c 		   ntohs(mdata->content.tx.seq));
tx                421 drivers/net/ethernet/mellanox/mlx5/core/fpga/tls.c 	if (MLX5_GET(tls_extended_cap, buf, tx))
tx                686 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_action.c 					action->dest_tbl.tbl->tx.s_anchor->chunk->icm_addr;
tx                245 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c 		dmn->info.tx.ste_type = MLX5DR_STE_TYPE_TX;
tx                246 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c 		dmn->info.tx.default_icm_addr = dmn->info.caps.nic_tx_allow_address;
tx                247 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c 		dmn->info.tx.drop_icm_addr = dmn->info.caps.nic_tx_drop_address;
tx                257 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c 		dmn->info.tx.ste_type = MLX5DR_STE_TYPE_TX;
tx                265 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c 		dmn->info.tx.default_icm_addr = vport_cap->icm_address_tx;
tx                268 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c 		dmn->info.tx.drop_icm_addr = dmn->info.caps.esw_tx_drop_address;
tx                477 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c 		ret = dr_matcher_connect(dmn, &matcher->tx,
tx                478 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c 					 next_matcher ? &next_matcher->tx : NULL,
tx                479 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c 					 prev_matcher ?	&prev_matcher->tx : NULL);
tx                504 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c 	dr_matcher_uninit_nic(&matcher->tx);
tx                516 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c 		dr_matcher_uninit_nic(&matcher->tx);
tx                581 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c 	ret = dr_matcher_init_nic(matcher, &matcher->tx);
tx                619 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c 		matcher->tx.nic_tbl = &tbl->tx;
tx                620 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c 		ret = dr_matcher_init_nic(matcher, &matcher->tx);
tx                624 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c 		matcher->tx.nic_tbl = &tbl->tx;
tx                738 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c 		ret = dr_matcher_disconnect(dmn, &tbl->tx,
tx                739 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c 					    next_matcher ? &next_matcher->tx : NULL,
tx                740 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_matcher.c 					    prev_matcher ? &prev_matcher->tx : NULL);
tx                948 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c 	dr_rule_destroy_rule_nic(rule, &rule->tx);
tx                961 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c 		dr_rule_destroy_rule_nic(rule, &rule->tx);
tx               1155 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c 	ret = dr_rule_create_rule_nic(rule, &rule->tx, &copy_param,
tx               1199 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c 		rule->tx.nic_matcher = &matcher->tx;
tx               1200 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c 		ret = dr_rule_create_rule_nic(rule, &rule->tx, &param,
tx               1205 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_rule.c 		rule->tx.nic_matcher = &matcher->tx;
tx                 51 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c 			last_htbl = last_matcher->tx.e_anchor;
tx                 53 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c 			last_htbl = tbl->tx.s_anchor;
tx                 55 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c 		tbl->tx.default_icm_addr = action ?
tx                 56 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c 			action->dest_tbl.tbl->tx.s_anchor->chunk->icm_addr :
tx                 57 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c 			tbl->tx.nic_dmn->default_icm_addr;
tx                 60 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c 		info.miss_icm_addr = tbl->tx.default_icm_addr;
tx                 63 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c 							tbl->tx.nic_dmn,
tx                 93 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c 	dr_table_uninit_nic(&tbl->tx);
tx                105 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c 		dr_table_uninit_nic(&tbl->tx);
tx                159 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c 	ret = dr_table_init_nic(tbl->dmn, &tbl->tx);
tx                186 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c 		tbl->tx.nic_dmn = &tbl->dmn->info.tx;
tx                187 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c 		ret = dr_table_init_nic(tbl->dmn, &tbl->tx);
tx                192 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c 		tbl->tx.nic_dmn = &tbl->dmn->info.tx;
tx                221 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c 	if (tbl->tx.s_anchor)
tx                222 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_table.c 		icm_addr_tx = tbl->tx.s_anchor->chunk->icm_addr;
tx                639 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_types.h 	struct mlx5dr_domain_rx_tx tx;
tx                671 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_types.h 	struct mlx5dr_table_rx_tx tx;
tx                696 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_types.h 	struct mlx5dr_matcher_rx_tx tx;
tx                781 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_types.h 	struct mlx5dr_rule_rx_tx tx;
tx                 82 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, version, 0x00, 28, 4);
tx                 89 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, ctl, 0x00, 26, 2);
tx                 94 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, proto, 0x00, 21, 3);
tx                 99 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, rx_is_router, 0x00, 19, 1);
tx                105 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, fid_valid, 0x00, 16, 1);
tx                110 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, swid, 0x00, 12, 3);
tx                116 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, control_tclass, 0x00, 6, 1);
tx                121 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, etclass, 0x00, 0, 4);
tx                131 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, port_mid, 0x04, 16, 16);
tx                138 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, fid, 0x08, 0, 16);
tx                144 drivers/net/ethernet/mellanox/mlxsw/spectrum.c MLXSW_ITEM32(tx, hdr, type, 0x0C, 0, 4);
tx                 58 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, version, 0x00, 28, 4);
tx                 65 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, ctl, 0x00, 26, 2);
tx                 70 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, proto, 0x00, 21, 3);
tx                 77 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, etclass, 0x00, 18, 3);
tx                 82 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, swid, 0x00, 12, 3);
tx                 92 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, port_mid, 0x04, 16, 16);
tx                 97 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, ctclass3, 0x04, 14, 1);
tx                103 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, rdq, 0x04, 9, 5);
tx                108 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, cpu_sig, 0x04, 0, 9);
tx                113 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, sig, 0x0C, 16, 16);
tx                118 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, stclass, 0x0C, 13, 3);
tx                123 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, emad, 0x0C, 5, 1);
tx                129 drivers/net/ethernet/mellanox/mlxsw/switchx2.c MLXSW_ITEM32(tx, hdr, type, 0x0C, 0, 4);
tx                922 drivers/net/ethernet/micrel/ksz884x.c 	struct ksz_desc_tx_stat tx;
tx                928 drivers/net/ethernet/micrel/ksz884x.c 	struct ksz_desc_tx_buf tx;
tx               1598 drivers/net/ethernet/micrel/ksz884x.c 	desc->sw.ctrl.tx.hw_owned = 1;
tx               1642 drivers/net/ethernet/micrel/ksz884x.c 	desc->sw.buf.tx.buf_size = len;
tx               3204 drivers/net/ethernet/micrel/ksz884x.c static void set_flow_ctrl(struct ksz_hw *hw, int rx, int tx)
tx               3215 drivers/net/ethernet/micrel/ksz884x.c 	if (tx)
tx               3231 drivers/net/ethernet/micrel/ksz884x.c 	int tx;
tx               3236 drivers/net/ethernet/micrel/ksz884x.c 	rx = tx = 0;
tx               3238 drivers/net/ethernet/micrel/ksz884x.c 		rx = tx = 1;
tx               3241 drivers/net/ethernet/micrel/ksz884x.c 			rx = tx = 1;
tx               3245 drivers/net/ethernet/micrel/ksz884x.c 			tx = 1;
tx               3252 drivers/net/ethernet/micrel/ksz884x.c 		set_flow_ctrl(hw, rx, tx);
tx               3978 drivers/net/ethernet/micrel/ksz884x.c 	hw->tx_desc_info.cur->sw.buf.tx.first_seg = 1;
tx               4004 drivers/net/ethernet/micrel/ksz884x.c 	cur->sw.buf.tx.last_seg = 1;
tx               4008 drivers/net/ethernet/micrel/ksz884x.c 		cur->sw.buf.tx.intr = 1;
tx               4014 drivers/net/ethernet/micrel/ksz884x.c 	cur->sw.buf.tx.dest_port = hw->dst_ports;
tx               4711 drivers/net/ethernet/micrel/ksz884x.c 		(desc)->sw.buf.tx.csum_gen_tcp = 1;
tx               4712 drivers/net/ethernet/micrel/ksz884x.c 		(desc)->sw.buf.tx.csum_gen_udp = 1;
tx               4751 drivers/net/ethernet/micrel/ksz884x.c 		if (status.tx.hw_owned) {
tx                 90 drivers/net/ethernet/microchip/enc28j60.c 	struct spi_transfer tx = {
tx                104 drivers/net/ethernet/microchip/enc28j60.c 	spi_message_add_tail(&tx, &msg);
tx                155 drivers/net/ethernet/microchip/lan743x_main.c 	struct lan743x_tx *tx = context;
tx                156 drivers/net/ethernet/microchip/lan743x_main.c 	struct lan743x_adapter *adapter = tx->adapter;
tx                163 drivers/net/ethernet/microchip/lan743x_main.c 				  INT_BIT_DMA_TX_(tx->channel_number));
tx                166 drivers/net/ethernet/microchip/lan743x_main.c 	if (int_sts & INT_BIT_DMA_TX_(tx->channel_number)) {
tx                167 drivers/net/ethernet/microchip/lan743x_main.c 		u32 ioc_bit = DMAC_INT_BIT_TX_IOC_(tx->channel_number);
tx                184 drivers/net/ethernet/microchip/lan743x_main.c 			napi_schedule(&tx->napi);
tx                192 drivers/net/ethernet/microchip/lan743x_main.c 				  INT_BIT_DMA_TX_(tx->channel_number));
tx                259 drivers/net/ethernet/microchip/lan743x_main.c 				lan743x_tx_isr(&adapter->tx[channel],
tx                606 drivers/net/ethernet/microchip/lan743x_main.c 							&adapter->tx[index]);
tx               1209 drivers/net/ethernet/microchip/lan743x_main.c static void lan743x_tx_release_desc(struct lan743x_tx *tx,
tx               1217 drivers/net/ethernet/microchip/lan743x_main.c 	descriptor = &tx->ring_cpu_ptr[descriptor_index];
tx               1218 drivers/net/ethernet/microchip/lan743x_main.c 	buffer_info = &tx->buffer_info[descriptor_index];
tx               1233 drivers/net/ethernet/microchip/lan743x_main.c 			dma_unmap_page(&tx->adapter->pdev->dev,
tx               1238 drivers/net/ethernet/microchip/lan743x_main.c 			dma_unmap_single(&tx->adapter->pdev->dev,
tx               1255 drivers/net/ethernet/microchip/lan743x_main.c 		lan743x_ptp_unrequest_tx_timestamp(tx->adapter);
tx               1260 drivers/net/ethernet/microchip/lan743x_main.c 		lan743x_ptp_tx_timestamp_skb(tx->adapter,
tx               1275 drivers/net/ethernet/microchip/lan743x_main.c static int lan743x_tx_next_index(struct lan743x_tx *tx, int index)
tx               1277 drivers/net/ethernet/microchip/lan743x_main.c 	return ((++index) % tx->ring_size);
tx               1280 drivers/net/ethernet/microchip/lan743x_main.c static void lan743x_tx_release_completed_descriptors(struct lan743x_tx *tx)
tx               1282 drivers/net/ethernet/microchip/lan743x_main.c 	while ((*tx->head_cpu_ptr) != (tx->last_head)) {
tx               1283 drivers/net/ethernet/microchip/lan743x_main.c 		lan743x_tx_release_desc(tx, tx->last_head, false);
tx               1284 drivers/net/ethernet/microchip/lan743x_main.c 		tx->last_head = lan743x_tx_next_index(tx, tx->last_head);
tx               1288 drivers/net/ethernet/microchip/lan743x_main.c static void lan743x_tx_release_all_descriptors(struct lan743x_tx *tx)
tx               1292 drivers/net/ethernet/microchip/lan743x_main.c 	original_head = tx->last_head;
tx               1294 drivers/net/ethernet/microchip/lan743x_main.c 		lan743x_tx_release_desc(tx, tx->last_head, true);
tx               1295 drivers/net/ethernet/microchip/lan743x_main.c 		tx->last_head = lan743x_tx_next_index(tx, tx->last_head);
tx               1296 drivers/net/ethernet/microchip/lan743x_main.c 	} while (tx->last_head != original_head);
tx               1297 drivers/net/ethernet/microchip/lan743x_main.c 	memset(tx->ring_cpu_ptr, 0,
tx               1298 drivers/net/ethernet/microchip/lan743x_main.c 	       sizeof(*tx->ring_cpu_ptr) * (tx->ring_size));
tx               1299 drivers/net/ethernet/microchip/lan743x_main.c 	memset(tx->buffer_info, 0,
tx               1300 drivers/net/ethernet/microchip/lan743x_main.c 	       sizeof(*tx->buffer_info) * (tx->ring_size));
tx               1303 drivers/net/ethernet/microchip/lan743x_main.c static int lan743x_tx_get_desc_cnt(struct lan743x_tx *tx,
tx               1316 drivers/net/ethernet/microchip/lan743x_main.c static int lan743x_tx_get_avail_desc(struct lan743x_tx *tx)
tx               1318 drivers/net/ethernet/microchip/lan743x_main.c 	int last_head = tx->last_head;
tx               1319 drivers/net/ethernet/microchip/lan743x_main.c 	int last_tail = tx->last_tail;
tx               1322 drivers/net/ethernet/microchip/lan743x_main.c 		return tx->ring_size - last_tail + last_head - 1;
tx               1327 drivers/net/ethernet/microchip/lan743x_main.c void lan743x_tx_set_timestamping_mode(struct lan743x_tx *tx,
tx               1332 drivers/net/ethernet/microchip/lan743x_main.c 		tx->ts_flags |= TX_TS_FLAG_TIMESTAMPING_ENABLED;
tx               1334 drivers/net/ethernet/microchip/lan743x_main.c 		tx->ts_flags &= ~TX_TS_FLAG_TIMESTAMPING_ENABLED;
tx               1336 drivers/net/ethernet/microchip/lan743x_main.c 		tx->ts_flags |= TX_TS_FLAG_ONE_STEP_SYNC;
tx               1338 drivers/net/ethernet/microchip/lan743x_main.c 		tx->ts_flags &= ~TX_TS_FLAG_ONE_STEP_SYNC;
tx               1341 drivers/net/ethernet/microchip/lan743x_main.c static int lan743x_tx_frame_start(struct lan743x_tx *tx,
tx               1353 drivers/net/ethernet/microchip/lan743x_main.c 	struct lan743x_adapter *adapter = tx->adapter;
tx               1357 drivers/net/ethernet/microchip/lan743x_main.c 	tx->frame_flags |= TX_FRAME_FLAG_IN_PROGRESS;
tx               1358 drivers/net/ethernet/microchip/lan743x_main.c 	tx->frame_first = tx->last_tail;
tx               1359 drivers/net/ethernet/microchip/lan743x_main.c 	tx->frame_tail = tx->frame_first;
tx               1361 drivers/net/ethernet/microchip/lan743x_main.c 	tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail];
tx               1362 drivers/net/ethernet/microchip/lan743x_main.c 	buffer_info = &tx->buffer_info[tx->frame_tail];
tx               1378 drivers/net/ethernet/microchip/lan743x_main.c 	tx->frame_data0 = (first_buffer_length &
tx               1384 drivers/net/ethernet/microchip/lan743x_main.c 		tx->frame_data0 |= TX_DESC_DATA0_TSE_;
tx               1387 drivers/net/ethernet/microchip/lan743x_main.c 		tx->frame_data0 |= TX_DESC_DATA0_ICE_ |
tx               1395 drivers/net/ethernet/microchip/lan743x_main.c static void lan743x_tx_frame_add_lso(struct lan743x_tx *tx,
tx               1406 drivers/net/ethernet/microchip/lan743x_main.c 	tx->frame_data0 |= TX_DESC_DATA0_EXT_;
tx               1408 drivers/net/ethernet/microchip/lan743x_main.c 		tx->frame_data0 |= TX_DESC_DATA0_LS_;
tx               1409 drivers/net/ethernet/microchip/lan743x_main.c 		tx->frame_data0 |= TX_DESC_DATA0_IOC_;
tx               1411 drivers/net/ethernet/microchip/lan743x_main.c 	tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail];
tx               1412 drivers/net/ethernet/microchip/lan743x_main.c 	tx_descriptor->data0 = tx->frame_data0;
tx               1415 drivers/net/ethernet/microchip/lan743x_main.c 	tx->frame_tail = lan743x_tx_next_index(tx, tx->frame_tail);
tx               1416 drivers/net/ethernet/microchip/lan743x_main.c 	tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail];
tx               1417 drivers/net/ethernet/microchip/lan743x_main.c 	buffer_info = &tx->buffer_info[tx->frame_tail];
tx               1429 drivers/net/ethernet/microchip/lan743x_main.c 	tx->frame_data0 = (frame_length & TX_DESC_DATA0_EXT_PAY_LENGTH_MASK_) |
tx               1436 drivers/net/ethernet/microchip/lan743x_main.c static int lan743x_tx_frame_add_fragment(struct lan743x_tx *tx,
tx               1445 drivers/net/ethernet/microchip/lan743x_main.c 	struct lan743x_adapter *adapter = tx->adapter;
tx               1455 drivers/net/ethernet/microchip/lan743x_main.c 	tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail];
tx               1456 drivers/net/ethernet/microchip/lan743x_main.c 	tx_descriptor->data0 = tx->frame_data0;
tx               1459 drivers/net/ethernet/microchip/lan743x_main.c 	tx->frame_tail = lan743x_tx_next_index(tx, tx->frame_tail);
tx               1460 drivers/net/ethernet/microchip/lan743x_main.c 	tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail];
tx               1461 drivers/net/ethernet/microchip/lan743x_main.c 	buffer_info = &tx->buffer_info[tx->frame_tail];
tx               1469 drivers/net/ethernet/microchip/lan743x_main.c 		desc_index = tx->frame_first;
tx               1470 drivers/net/ethernet/microchip/lan743x_main.c 		while (desc_index != tx->frame_tail) {
tx               1471 drivers/net/ethernet/microchip/lan743x_main.c 			lan743x_tx_release_desc(tx, desc_index, true);
tx               1472 drivers/net/ethernet/microchip/lan743x_main.c 			desc_index = lan743x_tx_next_index(tx, desc_index);
tx               1475 drivers/net/ethernet/microchip/lan743x_main.c 		tx->frame_flags &= ~TX_FRAME_FLAG_IN_PROGRESS;
tx               1476 drivers/net/ethernet/microchip/lan743x_main.c 		tx->frame_first = 0;
tx               1477 drivers/net/ethernet/microchip/lan743x_main.c 		tx->frame_data0 = 0;
tx               1478 drivers/net/ethernet/microchip/lan743x_main.c 		tx->frame_tail = 0;
tx               1493 drivers/net/ethernet/microchip/lan743x_main.c 	tx->frame_data0 = (fragment_length & TX_DESC_DATA0_BUF_LENGTH_MASK_) |
tx               1501 drivers/net/ethernet/microchip/lan743x_main.c static void lan743x_tx_frame_end(struct lan743x_tx *tx,
tx               1511 drivers/net/ethernet/microchip/lan743x_main.c 	struct lan743x_adapter *adapter = tx->adapter;
tx               1515 drivers/net/ethernet/microchip/lan743x_main.c 	if ((tx->frame_data0 & TX_DESC_DATA0_DTYPE_MASK_) ==
tx               1517 drivers/net/ethernet/microchip/lan743x_main.c 		tx->frame_data0 |= TX_DESC_DATA0_LS_;
tx               1518 drivers/net/ethernet/microchip/lan743x_main.c 		tx->frame_data0 |= TX_DESC_DATA0_IOC_;
tx               1521 drivers/net/ethernet/microchip/lan743x_main.c 	tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail];
tx               1522 drivers/net/ethernet/microchip/lan743x_main.c 	buffer_info = &tx->buffer_info[tx->frame_tail];
tx               1529 drivers/net/ethernet/microchip/lan743x_main.c 	tx_descriptor->data0 = tx->frame_data0;
tx               1530 drivers/net/ethernet/microchip/lan743x_main.c 	tx->frame_tail = lan743x_tx_next_index(tx, tx->frame_tail);
tx               1531 drivers/net/ethernet/microchip/lan743x_main.c 	tx->last_tail = tx->frame_tail;
tx               1535 drivers/net/ethernet/microchip/lan743x_main.c 	if (tx->vector_flags & LAN743X_VECTOR_FLAG_VECTOR_ENABLE_AUTO_SET)
tx               1537 drivers/net/ethernet/microchip/lan743x_main.c 	if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_SET)
tx               1541 drivers/net/ethernet/microchip/lan743x_main.c 	lan743x_csr_write(adapter, TX_TAIL(tx->channel_number),
tx               1542 drivers/net/ethernet/microchip/lan743x_main.c 			  tx_tail_flags | tx->frame_tail);
tx               1543 drivers/net/ethernet/microchip/lan743x_main.c 	tx->frame_flags &= ~TX_FRAME_FLAG_IN_PROGRESS;
tx               1546 drivers/net/ethernet/microchip/lan743x_main.c static netdev_tx_t lan743x_tx_xmit_frame(struct lan743x_tx *tx,
tx               1560 drivers/net/ethernet/microchip/lan743x_main.c 	required_number_of_descriptors = lan743x_tx_get_desc_cnt(tx, skb);
tx               1562 drivers/net/ethernet/microchip/lan743x_main.c 	spin_lock_irqsave(&tx->ring_lock, irq_flags);
tx               1564 drivers/net/ethernet/microchip/lan743x_main.c 		lan743x_tx_get_avail_desc(tx)) {
tx               1565 drivers/net/ethernet/microchip/lan743x_main.c 		if (required_number_of_descriptors > (tx->ring_size - 1)) {
tx               1569 drivers/net/ethernet/microchip/lan743x_main.c 			tx->overflow_skb = skb;
tx               1570 drivers/net/ethernet/microchip/lan743x_main.c 			netif_stop_queue(tx->adapter->netdev);
tx               1577 drivers/net/ethernet/microchip/lan743x_main.c 	    (tx->ts_flags & TX_TS_FLAG_TIMESTAMPING_ENABLED) &&
tx               1578 drivers/net/ethernet/microchip/lan743x_main.c 	    (lan743x_ptp_request_tx_timestamp(tx->adapter))) {
tx               1581 drivers/net/ethernet/microchip/lan743x_main.c 		if (tx->ts_flags & TX_TS_FLAG_ONE_STEP_SYNC)
tx               1594 drivers/net/ethernet/microchip/lan743x_main.c 	if (lan743x_tx_frame_start(tx,
tx               1604 drivers/net/ethernet/microchip/lan743x_main.c 		lan743x_tx_frame_add_lso(tx, frame_length, nr_frags);
tx               1612 drivers/net/ethernet/microchip/lan743x_main.c 		if (lan743x_tx_frame_add_fragment(tx, frag, frame_length)) {
tx               1624 drivers/net/ethernet/microchip/lan743x_main.c 	lan743x_tx_frame_end(tx, skb, do_timestamp, ignore_sync);
tx               1627 drivers/net/ethernet/microchip/lan743x_main.c 	spin_unlock_irqrestore(&tx->ring_lock, irq_flags);
tx               1633 drivers/net/ethernet/microchip/lan743x_main.c 	struct lan743x_tx *tx = container_of(napi, struct lan743x_tx, napi);
tx               1634 drivers/net/ethernet/microchip/lan743x_main.c 	struct lan743x_adapter *adapter = tx->adapter;
tx               1640 drivers/net/ethernet/microchip/lan743x_main.c 	ioc_bit = DMAC_INT_BIT_TX_IOC_(tx->channel_number);
tx               1642 drivers/net/ethernet/microchip/lan743x_main.c 	if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_W2C)
tx               1644 drivers/net/ethernet/microchip/lan743x_main.c 	spin_lock_irqsave(&tx->ring_lock, irq_flags);
tx               1647 drivers/net/ethernet/microchip/lan743x_main.c 	lan743x_tx_release_completed_descriptors(tx);
tx               1649 drivers/net/ethernet/microchip/lan743x_main.c 		if (tx->overflow_skb) {
tx               1650 drivers/net/ethernet/microchip/lan743x_main.c 			if (lan743x_tx_get_desc_cnt(tx, tx->overflow_skb) <=
tx               1651 drivers/net/ethernet/microchip/lan743x_main.c 				lan743x_tx_get_avail_desc(tx))
tx               1657 drivers/net/ethernet/microchip/lan743x_main.c 	spin_unlock_irqrestore(&tx->ring_lock, irq_flags);
tx               1661 drivers/net/ethernet/microchip/lan743x_main.c 		lan743x_tx_xmit_frame(tx, tx->overflow_skb);
tx               1662 drivers/net/ethernet/microchip/lan743x_main.c 		tx->overflow_skb = NULL;
tx               1671 drivers/net/ethernet/microchip/lan743x_main.c 			  INT_BIT_DMA_TX_(tx->channel_number));
tx               1678 drivers/net/ethernet/microchip/lan743x_main.c static void lan743x_tx_ring_cleanup(struct lan743x_tx *tx)
tx               1680 drivers/net/ethernet/microchip/lan743x_main.c 	if (tx->head_cpu_ptr) {
tx               1681 drivers/net/ethernet/microchip/lan743x_main.c 		pci_free_consistent(tx->adapter->pdev,
tx               1682 drivers/net/ethernet/microchip/lan743x_main.c 				    sizeof(*tx->head_cpu_ptr),
tx               1683 drivers/net/ethernet/microchip/lan743x_main.c 				    (void *)(tx->head_cpu_ptr),
tx               1684 drivers/net/ethernet/microchip/lan743x_main.c 				    tx->head_dma_ptr);
tx               1685 drivers/net/ethernet/microchip/lan743x_main.c 		tx->head_cpu_ptr = NULL;
tx               1686 drivers/net/ethernet/microchip/lan743x_main.c 		tx->head_dma_ptr = 0;
tx               1688 drivers/net/ethernet/microchip/lan743x_main.c 	kfree(tx->buffer_info);
tx               1689 drivers/net/ethernet/microchip/lan743x_main.c 	tx->buffer_info = NULL;
tx               1691 drivers/net/ethernet/microchip/lan743x_main.c 	if (tx->ring_cpu_ptr) {
tx               1692 drivers/net/ethernet/microchip/lan743x_main.c 		pci_free_consistent(tx->adapter->pdev,
tx               1693 drivers/net/ethernet/microchip/lan743x_main.c 				    tx->ring_allocation_size,
tx               1694 drivers/net/ethernet/microchip/lan743x_main.c 				    tx->ring_cpu_ptr,
tx               1695 drivers/net/ethernet/microchip/lan743x_main.c 				    tx->ring_dma_ptr);
tx               1696 drivers/net/ethernet/microchip/lan743x_main.c 		tx->ring_allocation_size = 0;
tx               1697 drivers/net/ethernet/microchip/lan743x_main.c 		tx->ring_cpu_ptr = NULL;
tx               1698 drivers/net/ethernet/microchip/lan743x_main.c 		tx->ring_dma_ptr = 0;
tx               1700 drivers/net/ethernet/microchip/lan743x_main.c 	tx->ring_size = 0;
tx               1703 drivers/net/ethernet/microchip/lan743x_main.c static int lan743x_tx_ring_init(struct lan743x_tx *tx)
tx               1710 drivers/net/ethernet/microchip/lan743x_main.c 	tx->ring_size = LAN743X_TX_RING_SIZE;
tx               1711 drivers/net/ethernet/microchip/lan743x_main.c 	if (tx->ring_size & ~TX_CFG_B_TX_RING_LEN_MASK_) {
tx               1715 drivers/net/ethernet/microchip/lan743x_main.c 	ring_allocation_size = ALIGN(tx->ring_size *
tx               1719 drivers/net/ethernet/microchip/lan743x_main.c 	cpu_ptr = pci_zalloc_consistent(tx->adapter->pdev,
tx               1726 drivers/net/ethernet/microchip/lan743x_main.c 	tx->ring_allocation_size = ring_allocation_size;
tx               1727 drivers/net/ethernet/microchip/lan743x_main.c 	tx->ring_cpu_ptr = (struct lan743x_tx_descriptor *)cpu_ptr;
tx               1728 drivers/net/ethernet/microchip/lan743x_main.c 	tx->ring_dma_ptr = dma_ptr;
tx               1730 drivers/net/ethernet/microchip/lan743x_main.c 	cpu_ptr = kcalloc(tx->ring_size, sizeof(*tx->buffer_info), GFP_KERNEL);
tx               1735 drivers/net/ethernet/microchip/lan743x_main.c 	tx->buffer_info = (struct lan743x_tx_buffer_info *)cpu_ptr;
tx               1737 drivers/net/ethernet/microchip/lan743x_main.c 	cpu_ptr = pci_zalloc_consistent(tx->adapter->pdev,
tx               1738 drivers/net/ethernet/microchip/lan743x_main.c 					sizeof(*tx->head_cpu_ptr), &dma_ptr);
tx               1744 drivers/net/ethernet/microchip/lan743x_main.c 	tx->head_cpu_ptr = cpu_ptr;
tx               1745 drivers/net/ethernet/microchip/lan743x_main.c 	tx->head_dma_ptr = dma_ptr;
tx               1746 drivers/net/ethernet/microchip/lan743x_main.c 	if (tx->head_dma_ptr & 0x3) {
tx               1754 drivers/net/ethernet/microchip/lan743x_main.c 	lan743x_tx_ring_cleanup(tx);
tx               1758 drivers/net/ethernet/microchip/lan743x_main.c static void lan743x_tx_close(struct lan743x_tx *tx)
tx               1760 drivers/net/ethernet/microchip/lan743x_main.c 	struct lan743x_adapter *adapter = tx->adapter;
tx               1764 drivers/net/ethernet/microchip/lan743x_main.c 			  DMAC_CMD_STOP_T_(tx->channel_number));
tx               1765 drivers/net/ethernet/microchip/lan743x_main.c 	lan743x_dmac_tx_wait_till_stopped(adapter, tx->channel_number);
tx               1769 drivers/net/ethernet/microchip/lan743x_main.c 			  DMAC_INT_BIT_TX_IOC_(tx->channel_number));
tx               1771 drivers/net/ethernet/microchip/lan743x_main.c 			  INT_BIT_DMA_TX_(tx->channel_number));
tx               1772 drivers/net/ethernet/microchip/lan743x_main.c 	napi_disable(&tx->napi);
tx               1773 drivers/net/ethernet/microchip/lan743x_main.c 	netif_napi_del(&tx->napi);
tx               1776 drivers/net/ethernet/microchip/lan743x_main.c 			  FCT_TX_CTL_DIS_(tx->channel_number));
tx               1778 drivers/net/ethernet/microchip/lan743x_main.c 				 FCT_TX_CTL_EN_(tx->channel_number),
tx               1781 drivers/net/ethernet/microchip/lan743x_main.c 	lan743x_tx_release_all_descriptors(tx);
tx               1783 drivers/net/ethernet/microchip/lan743x_main.c 	if (tx->overflow_skb) {
tx               1784 drivers/net/ethernet/microchip/lan743x_main.c 		dev_kfree_skb(tx->overflow_skb);
tx               1785 drivers/net/ethernet/microchip/lan743x_main.c 		tx->overflow_skb = NULL;
tx               1788 drivers/net/ethernet/microchip/lan743x_main.c 	lan743x_tx_ring_cleanup(tx);
tx               1791 drivers/net/ethernet/microchip/lan743x_main.c static int lan743x_tx_open(struct lan743x_tx *tx)
tx               1797 drivers/net/ethernet/microchip/lan743x_main.c 	adapter = tx->adapter;
tx               1798 drivers/net/ethernet/microchip/lan743x_main.c 	ret = lan743x_tx_ring_init(tx);
tx               1804 drivers/net/ethernet/microchip/lan743x_main.c 			  FCT_TX_CTL_RESET_(tx->channel_number));
tx               1806 drivers/net/ethernet/microchip/lan743x_main.c 				 FCT_TX_CTL_RESET_(tx->channel_number),
tx               1811 drivers/net/ethernet/microchip/lan743x_main.c 			  FCT_TX_CTL_EN_(tx->channel_number));
tx               1815 drivers/net/ethernet/microchip/lan743x_main.c 			  DMAC_CMD_TX_SWR_(tx->channel_number));
tx               1817 drivers/net/ethernet/microchip/lan743x_main.c 				 DMAC_CMD_TX_SWR_(tx->channel_number),
tx               1822 drivers/net/ethernet/microchip/lan743x_main.c 			  TX_BASE_ADDRH(tx->channel_number),
tx               1823 drivers/net/ethernet/microchip/lan743x_main.c 			  DMA_ADDR_HIGH32(tx->ring_dma_ptr));
tx               1825 drivers/net/ethernet/microchip/lan743x_main.c 			  TX_BASE_ADDRL(tx->channel_number),
tx               1826 drivers/net/ethernet/microchip/lan743x_main.c 			  DMA_ADDR_LOW32(tx->ring_dma_ptr));
tx               1829 drivers/net/ethernet/microchip/lan743x_main.c 	data = lan743x_csr_read(adapter, TX_CFG_B(tx->channel_number));
tx               1831 drivers/net/ethernet/microchip/lan743x_main.c 	data |= ((tx->ring_size) & TX_CFG_B_TX_RING_LEN_MASK_);
tx               1834 drivers/net/ethernet/microchip/lan743x_main.c 	lan743x_csr_write(adapter, TX_CFG_B(tx->channel_number), data);
tx               1844 drivers/net/ethernet/microchip/lan743x_main.c 	lan743x_csr_write(adapter, TX_CFG_A(tx->channel_number), data);
tx               1848 drivers/net/ethernet/microchip/lan743x_main.c 			  TX_HEAD_WRITEBACK_ADDRH(tx->channel_number),
tx               1849 drivers/net/ethernet/microchip/lan743x_main.c 			  DMA_ADDR_HIGH32(tx->head_dma_ptr));
tx               1851 drivers/net/ethernet/microchip/lan743x_main.c 			  TX_HEAD_WRITEBACK_ADDRL(tx->channel_number),
tx               1852 drivers/net/ethernet/microchip/lan743x_main.c 			  DMA_ADDR_LOW32(tx->head_dma_ptr));
tx               1855 drivers/net/ethernet/microchip/lan743x_main.c 	tx->last_head = lan743x_csr_read(adapter, TX_HEAD(tx->channel_number));
tx               1858 drivers/net/ethernet/microchip/lan743x_main.c 	tx->last_tail = 0;
tx               1859 drivers/net/ethernet/microchip/lan743x_main.c 	lan743x_csr_write(adapter, TX_TAIL(tx->channel_number),
tx               1860 drivers/net/ethernet/microchip/lan743x_main.c 			  (u32)(tx->last_tail));
tx               1861 drivers/net/ethernet/microchip/lan743x_main.c 	tx->vector_flags = lan743x_intr_get_vector_flags(adapter,
tx               1863 drivers/net/ethernet/microchip/lan743x_main.c 							 (tx->channel_number));
tx               1865 drivers/net/ethernet/microchip/lan743x_main.c 			  &tx->napi, lan743x_tx_napi_poll,
tx               1866 drivers/net/ethernet/microchip/lan743x_main.c 			  tx->ring_size - 1);
tx               1867 drivers/net/ethernet/microchip/lan743x_main.c 	napi_enable(&tx->napi);
tx               1870 drivers/net/ethernet/microchip/lan743x_main.c 	if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_CLEAR)
tx               1872 drivers/net/ethernet/microchip/lan743x_main.c 	if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_AUTO_CLEAR)
tx               1874 drivers/net/ethernet/microchip/lan743x_main.c 	if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_R2C)
tx               1876 drivers/net/ethernet/microchip/lan743x_main.c 	if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_R2C)
tx               1878 drivers/net/ethernet/microchip/lan743x_main.c 	lan743x_csr_write(adapter, TX_CFG_C(tx->channel_number), data);
tx               1880 drivers/net/ethernet/microchip/lan743x_main.c 	if (!(tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_SET))
tx               1882 drivers/net/ethernet/microchip/lan743x_main.c 				  INT_BIT_DMA_TX_(tx->channel_number));
tx               1884 drivers/net/ethernet/microchip/lan743x_main.c 			  DMAC_INT_BIT_TX_IOC_(tx->channel_number));
tx               1888 drivers/net/ethernet/microchip/lan743x_main.c 			  DMAC_CMD_START_T_(tx->channel_number));
tx               2465 drivers/net/ethernet/microchip/lan743x_main.c 	lan743x_tx_close(&adapter->tx[0]);
tx               2511 drivers/net/ethernet/microchip/lan743x_main.c 	ret = lan743x_tx_open(&adapter->tx[0]);
tx               2544 drivers/net/ethernet/microchip/lan743x_main.c 	return lan743x_tx_xmit_frame(&adapter->tx[0], skb);
tx               2671 drivers/net/ethernet/microchip/lan743x_main.c 	struct lan743x_tx *tx;
tx               2706 drivers/net/ethernet/microchip/lan743x_main.c 	tx = &adapter->tx[0];
tx               2707 drivers/net/ethernet/microchip/lan743x_main.c 	tx->adapter = adapter;
tx               2708 drivers/net/ethernet/microchip/lan743x_main.c 	tx->channel_number = 0;
tx               2709 drivers/net/ethernet/microchip/lan743x_main.c 	spin_lock_init(&tx->ring_lock);
tx                668 drivers/net/ethernet/microchip/lan743x_main.h void lan743x_tx_set_timestamping_mode(struct lan743x_tx *tx,
tx                718 drivers/net/ethernet/microchip/lan743x_main.h 	struct lan743x_tx       tx[LAN743X_MAX_TX_CHANNELS];
tx               1132 drivers/net/ethernet/microchip/lan743x_ptp.c 			lan743x_tx_set_timestamping_mode(&adapter->tx[index],
tx               1139 drivers/net/ethernet/microchip/lan743x_ptp.c 			lan743x_tx_set_timestamping_mode(&adapter->tx[index],
tx               1146 drivers/net/ethernet/microchip/lan743x_ptp.c 			lan743x_tx_set_timestamping_mode(&adapter->tx[index],
tx                173 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	struct myri10ge_tx_buf tx;	/* transmit ring        */
tx               1049 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		ss->tx.req = 0;
tx               1050 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		ss->tx.done = 0;
tx               1051 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		ss->tx.pkt_start = 0;
tx               1052 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		ss->tx.pkt_done = 0;
tx               1057 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		ss->tx.wake_queue = 0;
tx               1058 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		ss->tx.stop_queue = 0;
tx               1390 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	struct myri10ge_tx_buf *tx = &ss->tx;
tx               1395 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	while (tx->pkt_done != mcp_index) {
tx               1396 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		idx = tx->done & tx->mask;
tx               1397 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		skb = tx->info[idx].skb;
tx               1400 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		tx->info[idx].skb = NULL;
tx               1401 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		if (tx->info[idx].last) {
tx               1402 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 			tx->pkt_done++;
tx               1403 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 			tx->info[idx].last = 0;
tx               1405 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		tx->done++;
tx               1406 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		len = dma_unmap_len(&tx->info[idx], len);
tx               1407 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		dma_unmap_len_set(&tx->info[idx], len, 0);
tx               1414 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 						 dma_unmap_addr(&tx->info[idx],
tx               1420 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 					       dma_unmap_addr(&tx->info[idx],
tx               1438 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		if (tx->req == tx->done) {
tx               1439 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 			tx->queue_active = 0;
tx               1440 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 			put_be32(htonl(1), tx->send_stop);
tx               1448 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	    tx->req - tx->done < (tx->mask >> 1) &&
tx               1450 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		tx->wake_queue++;
tx               1555 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	struct myri10ge_tx_buf *tx = &ss->tx;
tx               1589 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		if (send_done_count != tx->pkt_done)
tx               1720 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	ring->tx_max_pending = mgp->ss[0].tx.mask + 1;
tx               1844 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		data[i++] = (unsigned int)ss->tx.pkt_start;
tx               1845 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		data[i++] = (unsigned int)ss->tx.pkt_done;
tx               1846 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		data[i++] = (unsigned int)ss->tx.req;
tx               1847 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		data[i++] = (unsigned int)ss->tx.done;
tx               1850 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		data[i++] = (unsigned int)ss->tx.wake_queue;
tx               1851 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		data[i++] = (unsigned int)ss->tx.stop_queue;
tx               1852 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		data[i++] = (unsigned int)ss->tx.linearized;
tx               1962 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	ss->tx.mask = tx_ring_entries - 1;
tx               1970 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	    * sizeof(*ss->tx.req_list);
tx               1971 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	ss->tx.req_bytes = kzalloc(bytes, GFP_KERNEL);
tx               1972 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	if (ss->tx.req_bytes == NULL)
tx               1976 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	ss->tx.req_list = (struct mcp_kreq_ether_send *)
tx               1977 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	    ALIGN((unsigned long)ss->tx.req_bytes, 8);
tx               1978 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	ss->tx.queue_active = 0;
tx               1992 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	bytes = tx_ring_entries * sizeof(*ss->tx.info);
tx               1993 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	ss->tx.info = kzalloc(bytes, GFP_KERNEL);
tx               1994 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	if (ss->tx.info == NULL)
tx               2062 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	kfree(ss->tx.info);
tx               2071 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	kfree(ss->tx.req_bytes);
tx               2072 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	ss->tx.req_bytes = NULL;
tx               2073 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	ss->tx.req_list = NULL;
tx               2083 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	struct myri10ge_tx_buf *tx;
tx               2087 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	if (ss->tx.req_list == NULL)
tx               2110 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	tx = &ss->tx;
tx               2111 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	while (tx->done != tx->req) {
tx               2112 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		idx = tx->done & tx->mask;
tx               2113 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		skb = tx->info[idx].skb;
tx               2116 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		tx->info[idx].skb = NULL;
tx               2117 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		tx->done++;
tx               2118 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		len = dma_unmap_len(&tx->info[idx], len);
tx               2119 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		dma_unmap_len_set(&tx->info[idx], len, 0);
tx               2125 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 						 dma_unmap_addr(&tx->info[idx],
tx               2131 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 					       dma_unmap_addr(&tx->info[idx],
tx               2140 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	kfree(ss->tx.info);
tx               2146 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	kfree(ss->tx.req_bytes);
tx               2147 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	ss->tx.req_bytes = NULL;
tx               2148 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	ss->tx.req_list = NULL;
tx               2246 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		ss->tx.lanai = (struct mcp_kreq_ether_send __iomem *)
tx               2260 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	ss->tx.send_go = (__iomem __be32 *)
tx               2262 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	ss->tx.send_stop = (__iomem __be32 *)
tx               2485 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	if (mgp->ss[0].tx.req_bytes == NULL)
tx               2522 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_submit_req_backwards(struct myri10ge_tx_buf *tx,
tx               2526 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	starting_slot = tx->req;
tx               2529 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		idx = (starting_slot + cnt) & tx->mask;
tx               2530 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		myri10ge_pio_copy(&tx->lanai[idx], &src[cnt], sizeof(*src));
tx               2543 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_submit_req(struct myri10ge_tx_buf *tx, struct mcp_kreq_ether_send *src,
tx               2551 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	idx = tx->req & tx->mask;
tx               2556 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	dst = dstp = &tx->lanai[idx];
tx               2559 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	if ((idx + cnt) < tx->mask) {
tx               2569 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		myri10ge_submit_req_backwards(tx, src, cnt);
tx               2581 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	tx->req += cnt;
tx               2586 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 				  struct myri10ge_tx_buf *tx, int idx)
tx               2592 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	last_idx = (idx + 1) & tx->mask;
tx               2593 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	idx = tx->req & tx->mask;
tx               2595 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		len = dma_unmap_len(&tx->info[idx], len);
tx               2597 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 			if (tx->info[idx].skb != NULL)
tx               2599 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 						 dma_unmap_addr(&tx->info[idx],
tx               2604 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 					       dma_unmap_addr(&tx->info[idx],
tx               2607 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 			dma_unmap_len_set(&tx->info[idx], len, 0);
tx               2608 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 			tx->info[idx].skb = NULL;
tx               2610 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		idx = (idx + 1) & tx->mask;
tx               2630 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	struct myri10ge_tx_buf *tx;
tx               2645 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	tx = &ss->tx;
tx               2648 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	req = tx->req_list;
tx               2649 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	avail = tx->mask - 1 - (tx->req - tx->done);
tx               2661 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		tx->stop_queue++;
tx               2733 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	idx = tx->req & tx->mask;
tx               2734 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	tx->info[idx].skb = skb;
tx               2735 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	dma_unmap_addr_set(&tx->info[idx], bus, bus);
tx               2736 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	dma_unmap_len_set(&tx->info[idx], len, len);
tx               2839 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 			myri10ge_unmap_tx_dma(mgp, tx, idx);
tx               2842 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		idx = (count + tx->req) & tx->mask;
tx               2843 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		dma_unmap_addr_set(&tx->info[idx], bus, bus);
tx               2844 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		dma_unmap_len_set(&tx->info[idx], len, len);
tx               2854 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	idx = ((count - 1) + tx->req) & tx->mask;
tx               2855 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	tx->info[idx].last = 1;
tx               2856 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	myri10ge_submit_req(tx, tx->req_list, count);
tx               2859 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	if ((mgp->dev->real_num_tx_queues > 1) && tx->queue_active == 0) {
tx               2860 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		tx->queue_active = 1;
tx               2861 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		put_be32(htonl(1), tx->send_go);
tx               2864 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	tx->pkt_start++;
tx               2866 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 		tx->stop_queue++;
tx               2872 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	myri10ge_unmap_tx_dma(mgp, tx, idx);
tx               2882 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	tx->linearized++;
tx               3375 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	if (ss->tx.req != ss->tx.done &&
tx               3376 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	    ss->tx.done == ss->watchdog_tx_done &&
tx               3386 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 				    slice, ss->tx.queue_active, ss->tx.req,
tx               3387 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 				    ss->tx.done, ss->tx.pkt_start,
tx               3388 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 				    ss->tx.pkt_done,
tx               3395 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	if (ss->watchdog_tx_done != ss->tx.done ||
tx               3399 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	ss->watchdog_tx_done = ss->tx.done;
tx               3400 drivers/net/ethernet/myricom/myri10ge/myri10ge.c 	ss->watchdog_tx_req = ss->tx.req;
tx               1903 drivers/net/ethernet/neterion/vxge/vxge-config.c 						 u32 port, u32 *tx, u32 *rx)
tx               1925 drivers/net/ethernet/neterion/vxge/vxge-config.c 		*tx = 1;
tx               1938 drivers/net/ethernet/neterion/vxge/vxge-config.c 						 u32 port, u32 tx, u32 rx)
tx               1959 drivers/net/ethernet/neterion/vxge/vxge-config.c 	if (tx)
tx               1888 drivers/net/ethernet/neterion/vxge/vxge-config.h 	 u32 *tx,
tx               1894 drivers/net/ethernet/neterion/vxge/vxge-config.h 	u32 tx,
tx                150 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c 	struct dentry *queues, *tx, *rx, *xdp;
tx                167 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c 	tx = debugfs_create_dir("tx", queues);
tx                180 drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c 		debugfs_create_file(name, 0400, tx,
tx                111 drivers/net/ethernet/pasemi/pasemi_mac.c 	return mac->tx;
tx               1079 drivers/net/ethernet/pasemi/pasemi_mac.c 	mac->tx = pasemi_mac_setup_tx_resources(dev);
tx               1081 drivers/net/ethernet/pasemi/pasemi_mac.c 	if (!mac->tx)
tx               1104 drivers/net/ethernet/pasemi/pasemi_mac.c 	write_iob_reg(PAS_IOB_DMA_TXCH_CFG(mac->tx->chan.chno),
tx               1169 drivers/net/ethernet/pasemi/pasemi_mac.c 	ret = request_irq(mac->tx->chan.irq, pasemi_mac_tx_intr, 0,
tx               1170 drivers/net/ethernet/pasemi/pasemi_mac.c 			  mac->tx_irq_name, mac->tx);
tx               1173 drivers/net/ethernet/pasemi/pasemi_mac.c 			mac->tx->chan.irq, ret);
tx               1191 drivers/net/ethernet/pasemi/pasemi_mac.c 	timer_setup(&mac->tx->clean_timer, pasemi_mac_tx_timer, 0);
tx               1192 drivers/net/ethernet/pasemi/pasemi_mac.c 	mod_timer(&mac->tx->clean_timer, jiffies + HZ);
tx               1197 drivers/net/ethernet/pasemi/pasemi_mac.c 	free_irq(mac->tx->chan.irq, mac->tx);
tx               1202 drivers/net/ethernet/pasemi/pasemi_mac.c 	if (mac->tx)
tx               1287 drivers/net/ethernet/pasemi/pasemi_mac.c 	del_timer_sync(&mac->tx->clean_timer);
tx               1319 drivers/net/ethernet/pasemi/pasemi_mac.c 	free_irq(mac->tx->chan.irq, mac->tx);
tx               1583 drivers/net/ethernet/pasemi/pasemi_mac.c 	disable_irq(mac->tx->chan.irq);
tx               1584 drivers/net/ethernet/pasemi/pasemi_mac.c 	pasemi_mac_tx_intr(mac->tx->chan.irq, mac->tx);
tx               1585 drivers/net/ethernet/pasemi/pasemi_mac.c 	enable_irq(mac->tx->chan.irq);
tx               1804 drivers/net/ethernet/pasemi/pasemi_mac.c 	pasemi_dma_free_chan(&mac->tx->chan);
tx                 76 drivers/net/ethernet/pasemi/pasemi_mac.h 	struct pasemi_mac_txring *tx;
tx                 94 drivers/net/ethernet/pasemi/pasemi_mac.h #define TX_DESC(tx, num)	((tx)->chan.ring_virt[(num) & (TX_RING_SIZE-1)])
tx                 95 drivers/net/ethernet/pasemi/pasemi_mac.h #define TX_DESC_INFO(tx, num)	((tx)->ring_info[(num) & (TX_RING_SIZE-1)])
tx                 77 drivers/net/ethernet/pasemi/pasemi_mac_ethtool.c 	ering->tx_pending = RING_USED(mac->tx)/2;
tx                 56 drivers/net/ethernet/pensando/ionic/ionic_lif.h 		struct ionic_tx_stats tx;
tx                 81 drivers/net/ethernet/pensando/ionic/ionic_lif.h #define q_to_tx_stats(q)	(&q_to_qcq(q)->stats->tx)
tx                185 drivers/net/ethernet/pensando/ionic/ionic_lif.h #define lif_to_txstats(lif, i)	((lif)->txqcqs[i].stats->tx)
tx                257 drivers/net/ethernet/pensando/ionic/ionic_lif.h 	qcq->stats->tx.sg_cntr[num_sg_elems]++;
tx                 87 drivers/net/ethernet/pensando/ionic/ionic_stats.c 			tstats = &txqcq->stats->tx;
tx               5734 drivers/net/ethernet/qlogic/qed/qed_hsi.h 	u8 tx;
tx               2353 drivers/net/ethernet/qlogic/qed/qed_int.c 			  u8 timer_res, u16 sb_id, bool tx)
tx               2371 drivers/net/ethernet/qlogic/qed/qed_int.c 	if (tx)
tx                430 drivers/net/ethernet/qlogic/qed/qed_int.h 			  u8 timer_res, u16 sb_id, bool tx);
tx                 93 drivers/net/ethernet/qlogic/qed/qed_l2.c 		u8 rx = 0, tx = 0;
tx                 96 drivers/net/ethernet/qlogic/qed/qed_l2.c 		qed_vf_get_num_txqs(p_hwfn, &tx);
tx                 98 drivers/net/ethernet/qlogic/qed/qed_l2.c 		p_l2_info->queues = max_t(u8, rx, tx);
tx               1281 drivers/net/ethernet/qlogic/qed/qed_l2.c 	p_ramrod->filter_cmd_hdr.tx = p_filter_cmd->is_tx_filter ? 1 : 0;
tx               1510 drivers/net/ethernet/qlogic/qlcnic/qlcnic.h 	struct __qlcnic_esw_statistics tx;
tx               1315 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c 			      struct qlcnic_host_tx_ring *tx, int ring)
tx               1327 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c 	tx->producer = 0;
tx               1328 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c 	tx->sw_consumer = 0;
tx               1329 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c 	*(tx->hw_consumer) = 0;
tx               1334 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c 	mbx.phys_addr_low = LSD(tx->phys_addr);
tx               1335 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c 	mbx.phys_addr_high = MSD(tx->phys_addr);
tx               1336 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c 	mbx.cnsmr_index_low = LSD(tx->hw_cons_phys_addr);
tx               1337 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c 	mbx.cnsmr_index_high = MSD(tx->hw_cons_phys_addr);
tx               1338 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c 	mbx.size = tx->num_desc;
tx               1378 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c 	tx->crb_cmd_producer = ahw->pci_base0 + mbx_out->host_prod;
tx               1379 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c 	tx->ctx_id = mbx_out->ctx_id;
tx               1383 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c 		tx->crb_intr_mask = ahw->pci_base0 + intr_mask;
tx               1387 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c 		    tx->ctx_id, mbx_out->state);
tx               1393 drivers/net/ethernet/qlogic/qlcnic/qlcnic_ethtool.c 			QLCNIC_QUERY_TX_COUNTER, &port_stats.tx);
tx               1397 drivers/net/ethernet/qlogic/qlcnic/qlcnic_ethtool.c 	qlcnic_fill_stats(data, &port_stats.tx, QLCNIC_ESW_STATS);
tx                799 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c 				    &port_stats.tx);
tx                834 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c 				       &esw_stats.tx);
tx                221 drivers/net/ethernet/smsc/smc911x.h 	struct dma_async_tx_descriptor *tx;
tx                233 drivers/net/ethernet/smsc/smc911x.h 	tx = dmaengine_prep_slave_single(dma, rx_dmabuf, rx_dmalen,
tx                235 drivers/net/ethernet/smsc/smc911x.h 	if (tx) {
tx                236 drivers/net/ethernet/smsc/smc911x.h 		tx->callback = smc911x_rx_dma_irq;
tx                237 drivers/net/ethernet/smsc/smc911x.h 		tx->callback_param = lp;
tx                238 drivers/net/ethernet/smsc/smc911x.h 		dmaengine_submit(tx);
tx                253 drivers/net/ethernet/smsc/smc911x.h 	struct dma_async_tx_descriptor *tx;
tx                265 drivers/net/ethernet/smsc/smc911x.h 	tx = dmaengine_prep_slave_single(dma, tx_dmabuf, tx_dmalen,
tx                267 drivers/net/ethernet/smsc/smc911x.h 	if (tx) {
tx                268 drivers/net/ethernet/smsc/smc911x.h 		tx->callback = smc911x_tx_dma_irq;
tx                269 drivers/net/ethernet/smsc/smc911x.h 		tx->callback_param = lp;
tx                270 drivers/net/ethernet/smsc/smc911x.h 		dmaengine_submit(tx);
tx                302 drivers/net/ethernet/smsc/smc91x.h 	struct dma_async_tx_descriptor *tx;
tx                308 drivers/net/ethernet/smsc/smc91x.h 	tx = dmaengine_prep_slave_single(lp->dma_chan, dmabuf, len,
tx                310 drivers/net/ethernet/smsc/smc91x.h 	if (tx) {
tx                311 drivers/net/ethernet/smsc/smc91x.h 		cookie = dmaengine_submit(tx);
tx                278 drivers/net/ethernet/socionext/sni_ave.c 	struct ave_desc_info	tx;
tx                302 drivers/net/ethernet/socionext/sni_ave.c 	addr = ((id == AVE_DESCID_TX) ? priv->tx.daddr : priv->rx.daddr)
tx                320 drivers/net/ethernet/socionext/sni_ave.c 	addr = ((id == AVE_DESCID_TX) ? priv->tx.daddr : priv->rx.daddr)
tx                687 drivers/net/ethernet/socionext/sni_ave.c 	proc_idx = priv->tx.proc_idx;
tx                688 drivers/net/ethernet/socionext/sni_ave.c 	done_idx = priv->tx.done_idx;
tx                689 drivers/net/ethernet/socionext/sni_ave.c 	ndesc    = priv->tx.ndesc;
tx                715 drivers/net/ethernet/socionext/sni_ave.c 		if (priv->tx.desc[done_idx].skbs) {
tx                716 drivers/net/ethernet/socionext/sni_ave.c 			ave_dma_unmap(ndev, &priv->tx.desc[done_idx],
tx                718 drivers/net/ethernet/socionext/sni_ave.c 			dev_consume_skb_any(priv->tx.desc[done_idx].skbs);
tx                719 drivers/net/ethernet/socionext/sni_ave.c 			priv->tx.desc[done_idx].skbs = NULL;
tx                725 drivers/net/ethernet/socionext/sni_ave.c 	priv->tx.done_idx = done_idx;
tx               1275 drivers/net/ethernet/socionext/sni_ave.c 	priv->tx.desc = kcalloc(priv->tx.ndesc, sizeof(*priv->tx.desc),
tx               1277 drivers/net/ethernet/socionext/sni_ave.c 	if (!priv->tx.desc) {
tx               1285 drivers/net/ethernet/socionext/sni_ave.c 		kfree(priv->tx.desc);
tx               1291 drivers/net/ethernet/socionext/sni_ave.c 	priv->tx.proc_idx = 0;
tx               1292 drivers/net/ethernet/socionext/sni_ave.c 	priv->tx.done_idx = 0;
tx               1293 drivers/net/ethernet/socionext/sni_ave.c 	for (entry = 0; entry < priv->tx.ndesc; entry++) {
tx               1298 drivers/net/ethernet/socionext/sni_ave.c 	       (((priv->tx.ndesc * priv->desc_size) << 16) & AVE_TXDC_SIZE),
tx               1368 drivers/net/ethernet/socionext/sni_ave.c 	for (entry = 0; entry < priv->tx.ndesc; entry++) {
tx               1369 drivers/net/ethernet/socionext/sni_ave.c 		if (!priv->tx.desc[entry].skbs)
tx               1372 drivers/net/ethernet/socionext/sni_ave.c 		ave_dma_unmap(ndev, &priv->tx.desc[entry], DMA_TO_DEVICE);
tx               1373 drivers/net/ethernet/socionext/sni_ave.c 		dev_kfree_skb_any(priv->tx.desc[entry].skbs);
tx               1374 drivers/net/ethernet/socionext/sni_ave.c 		priv->tx.desc[entry].skbs = NULL;
tx               1376 drivers/net/ethernet/socionext/sni_ave.c 	priv->tx.proc_idx = 0;
tx               1377 drivers/net/ethernet/socionext/sni_ave.c 	priv->tx.done_idx = 0;
tx               1391 drivers/net/ethernet/socionext/sni_ave.c 	kfree(priv->tx.desc);
tx               1404 drivers/net/ethernet/socionext/sni_ave.c 	proc_idx = priv->tx.proc_idx;
tx               1405 drivers/net/ethernet/socionext/sni_ave.c 	done_idx = priv->tx.done_idx;
tx               1406 drivers/net/ethernet/socionext/sni_ave.c 	ndesc = priv->tx.ndesc;
tx               1424 drivers/net/ethernet/socionext/sni_ave.c 	ret = ave_dma_map(ndev, &priv->tx.desc[proc_idx],
tx               1432 drivers/net/ethernet/socionext/sni_ave.c 	priv->tx.desc[proc_idx].skbs = skb;
tx               1450 drivers/net/ethernet/socionext/sni_ave.c 	priv->tx.proc_idx = (proc_idx + 1) % ndesc;
tx               1624 drivers/net/ethernet/socionext/sni_ave.c 		priv->tx.daddr  = AVE_TXDM_64;
tx               1629 drivers/net/ethernet/socionext/sni_ave.c 		priv->tx.daddr  = AVE_TXDM_32;
tx               1637 drivers/net/ethernet/socionext/sni_ave.c 	priv->tx.ndesc = AVE_NR_TXDESC;
tx                 74 drivers/net/ethernet/stmicro/stmmac/dwmac-rk.c #define DELAY_ENABLE(soc, tx, rx) \
tx                 75 drivers/net/ethernet/stmicro/stmmac/dwmac-rk.c 	(((tx) ? soc##_GMAC_TXCLK_DLY_ENABLE : soc##_GMAC_TXCLK_DLY_DISABLE) | \
tx                 18 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 	u32 tx, rx;
tx                 20 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 	tx = readl(ioaddr + XGMAC_TX_CONFIG);
tx                 23 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 	tx |= XGMAC_CORE_INIT_TX;
tx                 27 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		tx |= XGMAC_CONFIG_TE;
tx                 28 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		tx &= ~hw->link.speed_mask;
tx                 32 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 			tx |= hw->link.xgmii.speed10000;
tx                 35 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 			tx |= hw->link.speed2500;
tx                 39 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 			tx |= hw->link.speed1000;
tx                 44 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 	writel(tx, ioaddr + XGMAC_TX_CONFIG);
tx                 51 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 	u32 tx = readl(ioaddr + XGMAC_TX_CONFIG);
tx                 55 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		tx |= XGMAC_CONFIG_TE;
tx                 58 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 		tx &= ~XGMAC_CONFIG_TE;
tx                 62 drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c 	writel(tx, ioaddr + XGMAC_TX_CONFIG);
tx                 61 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 	memset(&desc_data->tx, 0, sizeof(desc_data->tx));
tx                452 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 		memset(&ring->tx, 0, sizeof(ring->tx));
tx                527 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 	if ((tso && (pkt_info->mss != ring->tx.cur_mss)) ||
tx                528 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 	    (vlan && (pkt_info->vlan_ctag != ring->tx.cur_vlan_ctag)))
tx                699 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 	ring->tx.xmit_more = 0;
tx                727 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 	if (tso && (pkt_info->mss != ring->tx.cur_mss))
tx                732 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 	if (vlan && (pkt_info->vlan_ctag != ring->tx.cur_vlan_ctag))
tx                789 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 			ring->tx.cur_mss = pkt_info->mss;
tx                818 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 			ring->tx.cur_vlan_ctag = pkt_info->vlan_ctag;
tx                973 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 	desc_data->tx.packets = pkt_info->tx_packets;
tx                974 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 	desc_data->tx.bytes = pkt_info->tx_bytes;
tx               1003 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 		ring->tx.xmit_more = 1;
tx                 49 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 		ring->tx.queue_stopped = 1;
tx                 54 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 		if (ring->tx.xmit_more)
tx                134 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 		if (skb_shinfo(skb)->gso_size != ring->tx.cur_mss) {
tx                161 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 		if (skb_vlan_tag_get(skb) != ring->tx.cur_vlan_ctag)
tx               1079 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 			tx_packets += desc_data->tx.packets;
tx               1080 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 			tx_bytes += desc_data->tx.bytes;
tx               1096 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 	if ((ring->tx.queue_stopped == 1) &&
tx               1098 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 		ring->tx.queue_stopped = 0;
tx                297 drivers/net/ethernet/synopsys/dwc-xlgmac.h 	struct xlgmac_tx_desc_data tx;
tx                351 drivers/net/ethernet/synopsys/dwc-xlgmac.h 		} tx;
tx               1336 drivers/net/ethernet/ti/netcp_core.c 	config.u.tx.filt_einfo = false;
tx               1337 drivers/net/ethernet/ti/netcp_core.c 	config.u.tx.filt_pswords = false;
tx               1338 drivers/net/ethernet/ti/netcp_core.c 	config.u.tx.priority = DMA_PRIO_MED_L;
tx                776 drivers/net/ethernet/toshiba/ps3_gelic_net.c 					     card->vlan[type].tx);
tx               1572 drivers/net/ethernet/toshiba/ps3_gelic_net.c 		int tx;
tx               1576 drivers/net/ethernet/toshiba/ps3_gelic_net.c 			.tx = GELIC_LV1_VLAN_TX_ETHERNET_0,
tx               1580 drivers/net/ethernet/toshiba/ps3_gelic_net.c 			.tx = GELIC_LV1_VLAN_TX_WIRELESS,
tx               1589 drivers/net/ethernet/toshiba/ps3_gelic_net.c 					 vlan_id_ix[i].tx,
tx               1595 drivers/net/ethernet/toshiba/ps3_gelic_net.c 					vlan_id_ix[i].tx, status);
tx               1596 drivers/net/ethernet/toshiba/ps3_gelic_net.c 			card->vlan[i].tx = 0;
tx               1600 drivers/net/ethernet/toshiba/ps3_gelic_net.c 		card->vlan[i].tx = (u16)v1;
tx               1612 drivers/net/ethernet/toshiba/ps3_gelic_net.c 			card->vlan[i].tx = 0;
tx               1619 drivers/net/ethernet/toshiba/ps3_gelic_net.c 			i, card->vlan[i].tx, card->vlan[i].rx);
tx               1622 drivers/net/ethernet/toshiba/ps3_gelic_net.c 	if (card->vlan[GELIC_PORT_ETHERNET_0].tx) {
tx               1623 drivers/net/ethernet/toshiba/ps3_gelic_net.c 		BUG_ON(!card->vlan[GELIC_PORT_WIRELESS].tx);
tx               1630 drivers/net/ethernet/toshiba/ps3_gelic_net.c 		card->vlan[GELIC_PORT_WIRELESS].tx = 0;
tx                257 drivers/net/ethernet/toshiba/ps3_gelic_net.h 	u16 tx;
tx               2585 drivers/net/ethernet/toshiba/ps3_gelic_wireless.c 	if (!card->vlan[GELIC_PORT_WIRELESS].tx)
tx               2627 drivers/net/ethernet/toshiba/ps3_gelic_wireless.c 	if (!card->vlan[GELIC_PORT_WIRELESS].tx)
tx                614 drivers/net/ethernet/tundra/tsi108_eth.c 	int tx;
tx                619 drivers/net/ethernet/tundra/tsi108_eth.c 		tx = data->txtail;
tx                621 drivers/net/ethernet/tundra/tsi108_eth.c 		if (data->txring[tx].misc & TSI108_TX_OWN)
tx                624 drivers/net/ethernet/tundra/tsi108_eth.c 		skb = data->txskbs[tx];
tx                626 drivers/net/ethernet/tundra/tsi108_eth.c 		if (!(data->txring[tx].misc & TSI108_TX_OK))
tx                628 drivers/net/ethernet/tundra/tsi108_eth.c 			       dev->name, data->txring[tx].misc);
tx                633 drivers/net/ethernet/tundra/tsi108_eth.c 		if (data->txring[tx].misc & TSI108_TX_EOF) {
tx                678 drivers/net/ethernet/tundra/tsi108_eth.c 		int tx = data->txhead;
tx                690 drivers/net/ethernet/tundra/tsi108_eth.c 		if ((tx % TSI108_TX_INT_FREQ == 0) &&
tx                694 drivers/net/ethernet/tundra/tsi108_eth.c 		data->txskbs[tx] = skb;
tx                697 drivers/net/ethernet/tundra/tsi108_eth.c 			data->txring[tx].buf0 = dma_map_single(&data->pdev->dev,
tx                700 drivers/net/ethernet/tundra/tsi108_eth.c 			data->txring[tx].len = skb_headlen(skb);
tx                705 drivers/net/ethernet/tundra/tsi108_eth.c 			data->txring[tx].buf0 =
tx                709 drivers/net/ethernet/tundra/tsi108_eth.c 			data->txring[tx].len = skb_frag_size(frag);
tx                723 drivers/net/ethernet/tundra/tsi108_eth.c 		data->txring[tx].misc = misc | TSI108_TX_OWN;
tx               1401 drivers/net/ethernet/tundra/tsi108_eth.c 		int tx = data->txtail;
tx               1403 drivers/net/ethernet/tundra/tsi108_eth.c 		skb = data->txskbs[tx];
tx               1399 drivers/net/ethernet/via/via-velocity.c 		for (i = 0; i < vptr->tx.numq; i++) {
tx               1400 drivers/net/ethernet/via/via-velocity.c 			writel(vptr->tx.pool_dma[i], &regs->TDBaseLo[i]);
tx               1475 drivers/net/ethernet/via/via-velocity.c 	pool = dma_alloc_coherent(vptr->dev, tx_ring_size * vptr->tx.numq +
tx               1489 drivers/net/ethernet/via/via-velocity.c 	for (i = 0; i < vptr->tx.numq; i++) {
tx               1490 drivers/net/ethernet/via/via-velocity.c 		vptr->tx.rings[i] = pool;
tx               1491 drivers/net/ethernet/via/via-velocity.c 		vptr->tx.pool_dma[i] = pool_dma;
tx               1648 drivers/net/ethernet/via/via-velocity.c 	for (j = 0; j < vptr->tx.numq; j++) {
tx               1650 drivers/net/ethernet/via/via-velocity.c 		vptr->tx.infos[j] = kcalloc(vptr->options.numtx,
tx               1653 drivers/net/ethernet/via/via-velocity.c 		if (!vptr->tx.infos[j])	{
tx               1655 drivers/net/ethernet/via/via-velocity.c 				kfree(vptr->tx.infos[j]);
tx               1659 drivers/net/ethernet/via/via-velocity.c 		vptr->tx.tail[j] = vptr->tx.curr[j] = vptr->tx.used[j] = 0;
tx               1673 drivers/net/ethernet/via/via-velocity.c 		vptr->options.numtx * sizeof(struct tx_desc) * vptr->tx.numq;
tx               1743 drivers/net/ethernet/via/via-velocity.c 	struct velocity_td_info *td_info = &(vptr->tx.infos[q][n]);
tx               1773 drivers/net/ethernet/via/via-velocity.c 	for (j = 0; j < vptr->tx.numq; j++) {
tx               1774 drivers/net/ethernet/via/via-velocity.c 		if (vptr->tx.infos[j] == NULL)
tx               1779 drivers/net/ethernet/via/via-velocity.c 		kfree(vptr->tx.infos[j]);
tx               1780 drivers/net/ethernet/via/via-velocity.c 		vptr->tx.infos[j] = NULL;
tx               1898 drivers/net/ethernet/via/via-velocity.c 	for (qnum = 0; qnum < vptr->tx.numq; qnum++) {
tx               1899 drivers/net/ethernet/via/via-velocity.c 		for (idx = vptr->tx.tail[qnum]; vptr->tx.used[qnum] > 0;
tx               1905 drivers/net/ethernet/via/via-velocity.c 			td = &(vptr->tx.rings[qnum][idx]);
tx               1906 drivers/net/ethernet/via/via-velocity.c 			tdinfo = &(vptr->tx.infos[qnum][idx]);
tx               1930 drivers/net/ethernet/via/via-velocity.c 			vptr->tx.used[qnum]--;
tx               1932 drivers/net/ethernet/via/via-velocity.c 		vptr->tx.tail[qnum] = idx;
tx               2286 drivers/net/ethernet/via/via-velocity.c 		struct tx_info tx;
tx               2298 drivers/net/ethernet/via/via-velocity.c 		tmp_vptr->tx.numq = vptr->tx.numq;
tx               2312 drivers/net/ethernet/via/via-velocity.c 		tx = vptr->tx;
tx               2315 drivers/net/ethernet/via/via-velocity.c 		vptr->tx = tmp_vptr->tx;
tx               2318 drivers/net/ethernet/via/via-velocity.c 		tmp_vptr->tx = tx;
tx               2540 drivers/net/ethernet/via/via-velocity.c 	index = vptr->tx.curr[qnum];
tx               2541 drivers/net/ethernet/via/via-velocity.c 	td_ptr = &(vptr->tx.rings[qnum][index]);
tx               2542 drivers/net/ethernet/via/via-velocity.c 	tdinfo = &(vptr->tx.infos[qnum][index]);
tx               2597 drivers/net/ethernet/via/via-velocity.c 	vptr->tx.used[qnum]++;
tx               2598 drivers/net/ethernet/via/via-velocity.c 	vptr->tx.curr[qnum] = (index + 1) % vptr->options.numtx;
tx               2603 drivers/net/ethernet/via/via-velocity.c 	td_ptr = &(vptr->tx.rings[qnum][prev]);
tx               2642 drivers/net/ethernet/via/via-velocity.c 	vptr->tx.numq = info->txqueue;
tx               3216 drivers/net/ethernet/via/via-velocity.c 	for (i = 0; i < vptr->tx.numq; i++) {
tx               3217 drivers/net/ethernet/via/via-velocity.c 		if (vptr->tx.used[i])
tx               1423 drivers/net/ethernet/via/via-velocity.h #define AVAIL_TD(p,q)   ((p)->options.numtx-((p)->tx.used[(q)]))
tx               1451 drivers/net/ethernet/via/via-velocity.h 	} tx;
tx                309 drivers/net/fddi/skfp/fplustm.c 	smc->hw.fp.tx[QUEUE_S] = queue = &smc->hw.fp.tx_q[QUEUE_S] ;
tx                320 drivers/net/fddi/skfp/fplustm.c 	smc->hw.fp.tx[QUEUE_A0] = queue = &smc->hw.fp.tx_q[QUEUE_A0] ;
tx                185 drivers/net/fddi/skfp/h/fplustm.h 	struct s_smt_tx_queue *tx[USED_QUEUES] ;
tx                362 drivers/net/fddi/skfp/hwmtm.c 	queue = smc->hw.fp.tx[QUEUE_A0] ;
tx                376 drivers/net/fddi/skfp/hwmtm.c 	queue = smc->hw.fp.tx[QUEUE_S] ;
tx                563 drivers/net/fddi/skfp/hwmtm.c 	phys = repair_txd_ring(smc,smc->hw.fp.tx[QUEUE_A0]) ;
tx                568 drivers/net/fddi/skfp/hwmtm.c 	phys = repair_txd_ring(smc,smc->hw.fp.tx[QUEUE_S]) ;
tx               1550 drivers/net/fddi/skfp/hwmtm.c 	smc->os.hwm.tx_p = smc->hw.fp.tx[frame_status & QUEUE_A0] ;
tx               1848 drivers/net/fddi/skfp/hwmtm.c 	queue = smc->hw.fp.tx[QUEUE_A0] ;
tx               1950 drivers/net/fddi/skfp/hwmtm.c 		queue = smc->hw.fp.tx[i] ;
tx               2041 drivers/net/fddi/skfp/hwmtm.c 		queue = smc->hw.fp.tx[i] ;
tx               2065 drivers/net/fddi/skfp/hwmtm.c 		queue = smc->hw.fp.tx[i] ;
tx                317 drivers/net/fddi/skfp/rmt.c 			int	tx ;
tx                323 drivers/net/fddi/skfp/rmt.c 			if ((tx =  sm_mac_get_tx_state(smc)) == 4 || tx == 5) {
tx                342 drivers/net/fddi/skfp/rmt.c 				tx, smc->r.bn_flag);
tx                395 drivers/net/fddi/skfp/rmt.c 			int	tx ;
tx                401 drivers/net/fddi/skfp/rmt.c 			if ((tx =  sm_mac_get_tx_state(smc)) == 4 || tx == 5) {
tx                420 drivers/net/fddi/skfp/rmt.c 				tx, smc->r.bn_flag);
tx                246 drivers/net/fjes/fjes_hw.c 			result = fjes_hw_alloc_epbuf(&buf_pair->tx);
tx                255 drivers/net/fjes/fjes_hw.c 			fjes_hw_setup_epbuf(&buf_pair->tx, mac,
tx                295 drivers/net/fjes/fjes_hw.c 		fjes_hw_free_epbuf(&hw->ep_shm_info[epidx].tx);
tx                493 drivers/net/fjes/fjes_hw.c 						buf_pair->tx.size,
tx                498 drivers/net/fjes/fjes_hw.c 	req_buf->share_buffer.buffer[idx++] = buf_pair->tx.size;
tx                499 drivers/net/fjes/fjes_hw.c 	page_count = buf_pair->tx.size / EP_BUFFER_INFO_SIZE;
tx                501 drivers/net/fjes/fjes_hw.c 		addr = ((u8 *)(buf_pair->tx.buffer)) +
tx                734 drivers/net/fjes/fjes_hw.c 	return (hw->ep_shm_info[src_epid].tx.info->v1i.rx_status &
tx                784 drivers/net/fjes/fjes_hw.c 		hw->ep_shm_info[epidx].tx.info->v1i.rx_status |=
tx               1062 drivers/net/fjes/fjes_hw.c 			fjes_hw_setup_epbuf(&hw->ep_shm_info[epidx].tx,
tx               1117 drivers/net/fjes/fjes_hw.c 					&hw->ep_shm_info[epidx].tx,
tx               1132 drivers/net/fjes/fjes_hw.c 			hw->ep_shm_info[epidx].tx.
tx               1161 drivers/net/fjes/fjes_hw.c 					tx.info->v1i.rx_status |=
tx                252 drivers/net/fjes/fjes_hw.h 	} tx, rx;
tx                331 drivers/net/fjes/fjes_main.c 				   .tx.info->v1i.rx_status &=
tx                404 drivers/net/fjes/fjes_main.c 		fjes_hw_setup_epbuf(&buf_pair->tx, netdev->dev_addr,
tx                459 drivers/net/fjes/fjes_main.c 		fjes_hw_setup_epbuf(&buf_pair->tx,
tx                521 drivers/net/fjes/fjes_main.c 			info = adapter->hw.ep_shm_info[epid].tx.info;
tx                576 drivers/net/fjes/fjes_main.c 				hw->ep_shm_info[epid].tx.info->v1i.tx_status;
tx                580 drivers/net/fjes/fjes_main.c 				hw->ep_shm_info[epid].tx.info->v1i.tx_status =
tx                610 drivers/net/fjes/fjes_main.c 	retval = fjes_hw_epbuf_tx_pkt_send(&adapter->hw.ep_shm_info[dest].tx,
tx                615 drivers/net/fjes/fjes_main.c 	adapter->hw.ep_shm_info[dest].tx.info->v1i.tx_status =
tx                841 drivers/net/fjes/fjes_main.c 			hw->ep_shm_info[epidx].tx.info->v1i.rx_status &=
tx                864 drivers/net/fjes/fjes_main.c 			fjes_hw_setup_epbuf(&hw->ep_shm_info[epidx].tx,
tx                868 drivers/net/fjes/fjes_main.c 			hw->ep_shm_info[epidx].tx.info->v1i.rx_status |=
tx                894 drivers/net/fjes/fjes_main.c 			&adapter->hw.ep_shm_info[epid].tx, vid))
tx                896 drivers/net/fjes/fjes_main.c 				&adapter->hw.ep_shm_info[epid].tx, vid);
tx                912 drivers/net/fjes/fjes_main.c 		fjes_hw_del_vlan_id(&adapter->hw.ep_shm_info[epid].tx, vid);
tx                935 drivers/net/fjes/fjes_main.c 			hw->ep_shm_info[src_epid].tx.info->v1i.rx_status |=
tx                973 drivers/net/fjes/fjes_main.c 		hw->ep_shm_info[src_epid].tx.info->v1i.rx_status |=
tx               1129 drivers/net/fjes/fjes_main.c 				   .tx.info->v1i.rx_status |= FJES_RX_POLL_WORK;
tx               1194 drivers/net/fjes/fjes_main.c 					adapter->hw.ep_shm_info[epidx].tx
tx               1442 drivers/net/fjes/fjes_main.c 			fjes_hw_setup_epbuf(&hw->ep_shm_info[epidx].tx,
tx               1487 drivers/net/fjes/fjes_main.c 					&hw->ep_shm_info[epidx].tx,
tx               1499 drivers/net/fjes/fjes_main.c 				hw->ep_shm_info[epidx].tx.info->v1i.rx_status &=
tx                100 drivers/net/fjes/fjes_trace.h 		__field(u64, tx)
tx                106 drivers/net/fjes/fjes_trace.h 		void *tx, *rx;
tx                108 drivers/net/fjes/fjes_trace.h 		tx = (void *)buf_pair->tx.buffer;
tx                112 drivers/net/fjes/fjes_trace.h 		__entry->tx_size = buf_pair->tx.size;
tx                114 drivers/net/fjes/fjes_trace.h 		__entry->tx = page_to_phys(vmalloc_to_page(tx)) +
tx                115 drivers/net/fjes/fjes_trace.h 				offset_in_page(tx);
tx                120 drivers/net/fjes/fjes_trace.h 		  __entry->length, __entry->epid, __entry->tx, __entry->tx_size,
tx                285 drivers/net/fjes/fjes_trace.h 			hw->ep_shm_info[src_epid].tx.info->v1i.rx_status;
tx                305 drivers/net/fjes/fjes_trace.h 		__entry->rx_status = hw->ep_shm_info[src_epid].tx.info->v1i.rx_status;
tx                329 drivers/net/fjes/fjes_trace.h 			hw->ep_shm_info[src_epid].tx.info->v1i.rx_status;
tx                350 drivers/net/fjes/fjes_trace.h 			hw->ep_shm_info[src_epid].tx.info->v1i.rx_status;
tx                203 drivers/net/hamradio/scc.c static void scc_key_trx (struct scc_channel *scc, char tx);
tx                899 drivers/net/hamradio/scc.c static void scc_key_trx(struct scc_channel *scc, char tx)
tx                904 drivers/net/hamradio/scc.c 		Outb(scc->ctrl + 4, scc->option | (tx? 0x80 : 0));
tx                909 drivers/net/hamradio/scc.c 	time_const = (unsigned) (scc->clock / (scc->modem.speed * (tx? 2:64))) - 2;
tx                913 drivers/net/hamradio/scc.c 	if (tx)
tx                921 drivers/net/hamradio/scc.c 		if (tx)
tx                957 drivers/net/hamradio/scc.c 		if (tx)
tx                105 drivers/net/ieee802154/at86rf230.c 	struct at86rf230_state_change tx;
tx                908 drivers/net/ieee802154/at86rf230.c 	struct at86rf230_state_change *ctx = &lp->tx;
tx               1712 drivers/net/ieee802154/at86rf230.c 	at86rf230_setup_spi_messages(lp, &lp->tx);
tx                 34 drivers/net/netdevsim/ipsec.c 		      ipsec->count, ipsec->tx);
tx                270 drivers/net/netdevsim/ipsec.c 	ipsec->tx++;
tx                 47 drivers/net/netdevsim/netdevsim.h 	u32 tx;
tx               2094 drivers/net/phy/phy_device.c void phy_set_sym_pause(struct phy_device *phydev, bool rx, bool tx,
tx               2099 drivers/net/phy/phy_device.c 	if (rx && tx && autoneg)
tx               2118 drivers/net/phy/phy_device.c void phy_set_asym_pause(struct phy_device *phydev, bool rx, bool tx)
tx               2136 drivers/net/phy/phy_device.c 	if (tx)
tx               1610 drivers/net/usb/hso.c 	icount->tx          = cnow.tx;
tx                338 drivers/net/veth.c 	struct pcpu_lstats tx;
tx                340 drivers/net/veth.c 	tot->tx_dropped = veth_stats_tx(&tx, dev);
tx                341 drivers/net/veth.c 	tot->tx_bytes = tx.bytes;
tx                342 drivers/net/veth.c 	tot->tx_packets = tx.packets;
tx                352 drivers/net/veth.c 		tot->rx_dropped += veth_stats_tx(&tx, peer);
tx                353 drivers/net/veth.c 		tot->rx_bytes += tx.bytes;
tx                354 drivers/net/veth.c 		tot->rx_packets += tx.packets;
tx                328 drivers/net/wan/ixp4xx_hss.c 	int tx, txdone, rx, rxfree;
tx                889 drivers/net/wan/ixp4xx_hss.c 	queue_put_desc(queue_ids[port->id].tx, tx_desc_phys(port, n), desc);
tx                927 drivers/net/wan/ixp4xx_hss.c 	err = qmgr_request_queue(queue_ids[port->id].tx, TX_DESCS, 0, 0,
tx                946 drivers/net/wan/ixp4xx_hss.c 	qmgr_release_queue(queue_ids[port->id].tx);
tx                961 drivers/net/wan/ixp4xx_hss.c 	qmgr_release_queue(queue_ids[port->id].tx);
tx               1132 drivers/net/wan/ixp4xx_hss.c 	while (queue_get_desc(queue_ids[port->id].tx, port, 1) >= 0)
tx                485 drivers/net/wan/z85230.c 	.tx = z8530_tx,
tx                605 drivers/net/wan/z85230.c 	.tx = z8530_dma_tx,
tx                611 drivers/net/wan/z85230.c 	.tx = z8530_dma_tx,
tx                679 drivers/net/wan/z85230.c 	.tx = z8530_tx_clear,
tx                739 drivers/net/wan/z85230.c 				irqs->tx(&dev->chanA);
tx                751 drivers/net/wan/z85230.c 				irqs->tx(&dev->chanB);
tx                256 drivers/net/wan/z85230.h 	void (*tx)(struct z8530_channel *);
tx                204 drivers/net/wimax/i2400m/debugfs.c 	d_level_register_debugfs("dl_", tx, dentry);
tx                982 drivers/net/wimax/i2400m/driver.c 	D_SUBMODULE_DEFINE(tx),
tx                252 drivers/net/wimax/i2400m/tx.c #define D_SUBMODULE tx
tx                 74 drivers/net/wimax/i2400m/usb-tx.c #define D_SUBMODULE tx
tx                365 drivers/net/wimax/i2400m/usb.c 	D_SUBMODULE_DEFINE(tx),
tx                381 drivers/net/wimax/i2400m/usb.c 	d_level_register_debugfs("dl_", tx, dentry);
tx               1766 drivers/net/wireless/admtek/adm8211.c 	.tx			= adm8211_tx,
tx               1353 drivers/net/wireless/ath/ar5523/ar5523.c 	.tx			= ar5523_tx,
tx               8184 drivers/net/wireless/ath/ath10k/mac.c 	.tx				= ath10k_mac_op_tx,
tx               1390 drivers/net/wireless/ath/ath10k/wmi-tlv.c 		ath10k_wmi_pull_pdev_stats_tx(&src->tx, dst);
tx               2952 drivers/net/wireless/ath/ath10k/wmi.c 		ath10k_wmi_pull_pdev_stats_tx(&src->tx, dst);
tx               3006 drivers/net/wireless/ath/ath10k/wmi.c 		ath10k_wmi_pull_pdev_stats_tx(&src->tx, dst);
tx               3067 drivers/net/wireless/ath/ath10k/wmi.c 		ath10k_wmi_pull_pdev_stats_tx(&src->tx, dst);
tx               3144 drivers/net/wireless/ath/ath10k/wmi.c 		ath10k_wmi_pull_pdev_stats_tx(&src->tx, dst);
tx               3236 drivers/net/wireless/ath/ath10k/wmi.c 		ath10k_wmi_10_4_pull_pdev_stats_tx(&src->tx, dst);
tx               5670 drivers/net/wireless/ath/ath10k/wmi.c 	u64 busy, total, tx, rx, rx_bss;
tx               5683 drivers/net/wireless/ath/ath10k/wmi.c 	tx          = __le64_to_cpu(ev->cycle_tx);
tx               5689 drivers/net/wireless/ath/ath10k/wmi.c 		   freq, noise_floor, busy, total, tx, rx, rx_bss);
tx               5705 drivers/net/wireless/ath/ath10k/wmi.c 	survey->time_tx   = div_u64(tx, cc_freq_hz);
tx               4639 drivers/net/wireless/ath/ath10k/wmi.h 	struct wmi_pdev_stats_tx tx;
tx               4655 drivers/net/wireless/ath/ath10k/wmi.h 	struct wmi_pdev_stats_tx tx;
tx               4668 drivers/net/wireless/ath/ath10k/wmi.h 	struct wmi_pdev_stats_tx tx;
tx               4679 drivers/net/wireless/ath/ath10k/wmi.h 	struct wmi_10_4_pdev_stats_tx tx;
tx                740 drivers/net/wireless/ath/ath5k/mac80211-ops.c 				u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max)
tx                744 drivers/net/wireless/ath/ath5k/mac80211-ops.c 	*tx = ah->txqs[AR5K_TX_QUEUE_ID_DATA_MIN].txq_max;
tx                751 drivers/net/wireless/ath/ath5k/mac80211-ops.c static int ath5k_set_ringparam(struct ieee80211_hw *hw, u32 tx, u32 rx)
tx                761 drivers/net/wireless/ath/ath5k/mac80211-ops.c 	if (!tx || tx > ATH5K_TXQ_LEN_MAX)
tx                771 drivers/net/wireless/ath/ath5k/mac80211-ops.c 		ah->txqs[qnum].txq_max = tx;
tx                781 drivers/net/wireless/ath/ath5k/mac80211-ops.c 	.tx			= ath5k_tx,
tx               3408 drivers/net/wireless/ath/ath6kl/cfg80211.c 		.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
tx               3414 drivers/net/wireless/ath/ath6kl/cfg80211.c 		.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
tx               3420 drivers/net/wireless/ath/ath6kl/cfg80211.c 		.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
tx               3426 drivers/net/wireless/ath/ath6kl/cfg80211.c 		.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
tx                 48 drivers/net/wireless/ath/ath6kl/htc-ops.h 	return target->dev->ar->htc_ops->tx(target, packet);
tx                317 drivers/net/wireless/ath/ath6kl/htc.h 		struct htc_tx_packet_info tx;
tx                555 drivers/net/wireless/ath/ath6kl/htc.h 	int  (*tx)(struct htc_target *target, struct htc_packet *packet);
tx                643 drivers/net/wireless/ath/ath6kl/htc.h 	packet->info.tx.tag = tag;
tx                384 drivers/net/wireless/ath/ath6kl/htc_mbox.c 					      bool tx)
tx                389 drivers/net/wireless/ath/ath6kl/htc_mbox.c 	buf_list = tx ? &target->free_ctrl_txbuf : &target->free_ctrl_rxbuf;
tx                402 drivers/net/wireless/ath/ath6kl/htc_mbox.c 	if (tx)
tx                420 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		   packet->info.tx.cred_used);
tx                425 drivers/net/wireless/ath/ath6kl/htc_mbox.c 				packet->info.tx.cred_used;
tx                458 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		   packet->info.tx.seqno);
tx                523 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		   send_len, packet->info.tx.seqno, padded_len,
tx                633 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		packet->info.tx.cred_used = req_cred;
tx                641 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		packet->info.tx.flags = flags;
tx                642 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		packet->info.tx.seqno = endpoint->seqno;
tx                715 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		flags = packet->info.tx.flags | HTC_FLAGS_SEND_BUNDLE;
tx                717 drivers/net/wireless/ath/ath6kl/htc_mbox.c 				       cred_pad, packet->info.tx.seqno);
tx                728 drivers/net/wireless/ath/ath6kl/htc_mbox.c 			   i, packet, packet->info.tx.seqno, len, rem_scat);
tx                925 drivers/net/wireless/ath/ath6kl/htc_mbox.c 			ath6kl_htc_tx_prep_pkt(packet, packet->info.tx.flags,
tx                926 drivers/net/wireless/ath/ath6kl/htc_mbox.c 					       0, packet->info.tx.seqno);
tx               1169 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		    (tag == packet->info.tx.tag))
tx               1181 drivers/net/wireless/ath/ath6kl/htc_mbox.c 			   packet->endpoint, packet->info.tx.tag);
tx               2920 drivers/net/wireless/ath/ath6kl/htc_mbox.c 	.tx = ath6kl_htc_mbox_tx,
tx                 31 drivers/net/wireless/ath/ath6kl/htc_pipe.c 	if (packet->info.tx.flags & HTC_FLAGS_TX_FIXUP_NETBUF) {
tx                 33 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		packet->info.tx.flags &= ~HTC_FLAGS_TX_FIXUP_NETBUF;
tx                163 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		packet->info.tx.cred_used = credits_required;
tx                165 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		packet->info.tx.flags = send_flags;
tx                166 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		packet->info.tx.seqno = ep->seqno;
tx                192 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		packet->info.tx.seqno = ep->seqno;
tx                193 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		packet->info.tx.flags = 0;
tx                194 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		packet->info.tx.cred_used = 0;
tx                238 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		packet->info.tx.flags |= HTC_FLAGS_TX_FIXUP_NETBUF;
tx                242 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		htc_hdr->flags = packet->info.tx.flags;
tx                245 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		htc_hdr->ctrl[1] = (u8) packet->info.tx.seqno;
tx                272 drivers/net/wireless/ath/ath6kl/htc_pipe.c 			ep->cred_dist.credits += packet->info.tx.cred_used;
tx               1709 drivers/net/wireless/ath/ath6kl/htc_pipe.c 	.tx = ath6kl_htc_pipe_tx,
tx                715 drivers/net/wireless/ath/ath6kl/main.c 	stats->tx_pkt += le32_to_cpu(tgt_stats->stats.tx.pkt);
tx                716 drivers/net/wireless/ath/ath6kl/main.c 	stats->tx_byte += le32_to_cpu(tgt_stats->stats.tx.byte);
tx                717 drivers/net/wireless/ath/ath6kl/main.c 	stats->tx_ucast_pkt += le32_to_cpu(tgt_stats->stats.tx.ucast_pkt);
tx                718 drivers/net/wireless/ath/ath6kl/main.c 	stats->tx_ucast_byte += le32_to_cpu(tgt_stats->stats.tx.ucast_byte);
tx                719 drivers/net/wireless/ath/ath6kl/main.c 	stats->tx_mcast_pkt += le32_to_cpu(tgt_stats->stats.tx.mcast_pkt);
tx                720 drivers/net/wireless/ath/ath6kl/main.c 	stats->tx_mcast_byte += le32_to_cpu(tgt_stats->stats.tx.mcast_byte);
tx                721 drivers/net/wireless/ath/ath6kl/main.c 	stats->tx_bcast_pkt  += le32_to_cpu(tgt_stats->stats.tx.bcast_pkt);
tx                722 drivers/net/wireless/ath/ath6kl/main.c 	stats->tx_bcast_byte += le32_to_cpu(tgt_stats->stats.tx.bcast_byte);
tx                724 drivers/net/wireless/ath/ath6kl/main.c 		le32_to_cpu(tgt_stats->stats.tx.rts_success_cnt);
tx                728 drivers/net/wireless/ath/ath6kl/main.c 			le32_to_cpu(tgt_stats->stats.tx.pkt_per_ac[ac]);
tx                730 drivers/net/wireless/ath/ath6kl/main.c 	stats->tx_err += le32_to_cpu(tgt_stats->stats.tx.err);
tx                731 drivers/net/wireless/ath/ath6kl/main.c 	stats->tx_fail_cnt += le32_to_cpu(tgt_stats->stats.tx.fail_cnt);
tx                732 drivers/net/wireless/ath/ath6kl/main.c 	stats->tx_retry_cnt += le32_to_cpu(tgt_stats->stats.tx.retry_cnt);
tx                734 drivers/net/wireless/ath/ath6kl/main.c 		le32_to_cpu(tgt_stats->stats.tx.mult_retry_cnt);
tx                736 drivers/net/wireless/ath/ath6kl/main.c 		le32_to_cpu(tgt_stats->stats.tx.rts_fail_cnt);
tx                738 drivers/net/wireless/ath/ath6kl/main.c 	rate = a_sle32_to_cpu(tgt_stats->stats.tx.ucast_rate);
tx                 92 drivers/net/wireless/ath/ath6kl/trace.h 		__field(unsigned int, tx)
tx                106 drivers/net/wireless/ath/ath6kl/trace.h 			__entry->tx = 1;
tx                108 drivers/net/wireless/ath/ath6kl/trace.h 			__entry->tx = 0;
tx                113 drivers/net/wireless/ath/ath6kl/trace.h 		__entry->tx ? "tx" : "rx",
tx                127 drivers/net/wireless/ath/ath6kl/trace.h 		__field(unsigned int, tx)
tx                147 drivers/net/wireless/ath/ath6kl/trace.h 			__entry->tx = 1;
tx                149 drivers/net/wireless/ath/ath6kl/trace.h 			__entry->tx = 0;
tx                166 drivers/net/wireless/ath/ath6kl/trace.h 		__entry->tx ? "tx" : "rx",
tx                607 drivers/net/wireless/ath/ath6kl/txrx.c 	if (packet->info.tx.tag == ATH6KL_CONTROL_PKT_TAG)
tx               1832 drivers/net/wireless/ath/ath6kl/wmi.h 	struct tx_stats tx;
tx                676 drivers/net/wireless/ath/ath9k/ar9003_phy.c void ar9003_hw_set_chain_masks(struct ath_hw *ah, u8 rx, u8 tx)
tx                685 drivers/net/wireless/ath/ath9k/ar9003_phy.c 	if ((ah->caps.hw_caps & ATH9K_HW_CAP_APM) && (tx == 0x7))
tx                686 drivers/net/wireless/ath/ath9k/ar9003_phy.c 		tx = 3;
tx                688 drivers/net/wireless/ath/ath9k/ar9003_phy.c 	REG_WRITE(ah, AR_SELFGEN_MASK, tx);
tx                 83 drivers/net/wireless/ath/ath9k/ath9k.h #define	ATH_TXQ_SETUP(sc, i) ((sc)->tx.txqsetup & (1<<i))
tx               1026 drivers/net/wireless/ath/ath9k/ath9k.h 	struct ath_tx tx;
tx                 50 drivers/net/wireless/ath/ath9k/beacon.c 		txq = sc->tx.txq_map[IEEE80211_AC_BE];
tx               1016 drivers/net/wireless/ath/ath9k/channel.c 	txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO];
tx               1137 drivers/net/wireless/ath/ath9k/channel.c 	txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO];
tx                647 drivers/net/wireless/ath/ath9k/debug.c 		txq = sc->tx.txq_map[i];
tx                195 drivers/net/wireless/ath/ath9k/debug.h #define PR_QNUM(_n) sc->tx.txq_map[_n]->axq_qnum
tx                429 drivers/net/wireless/ath/ath9k/gpio.c 		txq = sc->tx.txq_map[IEEE80211_AC_BE];
tx                162 drivers/net/wireless/ath/ath9k/hif_usb.c 		spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
tx                163 drivers/net/wireless/ath/ath9k/hif_usb.c 		if (hif_dev->tx.flags & HIF_USB_TX_FLUSH) {
tx                164 drivers/net/wireless/ath/ath9k/hif_usb.c 			spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
tx                169 drivers/net/wireless/ath/ath9k/hif_usb.c 		spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
tx                279 drivers/net/wireless/ath/ath9k/hif_usb.c 		spin_lock(&hif_dev->tx.tx_lock);
tx                280 drivers/net/wireless/ath/ath9k/hif_usb.c 		if (hif_dev->tx.flags & HIF_USB_TX_FLUSH) {
tx                281 drivers/net/wireless/ath/ath9k/hif_usb.c 			spin_unlock(&hif_dev->tx.tx_lock);
tx                285 drivers/net/wireless/ath/ath9k/hif_usb.c 		spin_unlock(&hif_dev->tx.tx_lock);
tx                300 drivers/net/wireless/ath/ath9k/hif_usb.c 	spin_lock(&hif_dev->tx.tx_lock);
tx                301 drivers/net/wireless/ath/ath9k/hif_usb.c 	list_move_tail(&tx_buf->list, &hif_dev->tx.tx_buf);
tx                302 drivers/net/wireless/ath/ath9k/hif_usb.c 	hif_dev->tx.tx_buf_cnt++;
tx                303 drivers/net/wireless/ath/ath9k/hif_usb.c 	if (!(hif_dev->tx.flags & HIF_USB_TX_STOP))
tx                306 drivers/net/wireless/ath/ath9k/hif_usb.c 	spin_unlock(&hif_dev->tx.tx_lock);
tx                319 drivers/net/wireless/ath/ath9k/hif_usb.c 	if (hif_dev->tx.tx_skb_cnt == 0)
tx                323 drivers/net/wireless/ath/ath9k/hif_usb.c 	if (list_empty(&hif_dev->tx.tx_buf))
tx                326 drivers/net/wireless/ath/ath9k/hif_usb.c 	tx_buf = list_first_entry(&hif_dev->tx.tx_buf, struct tx_buf, list);
tx                327 drivers/net/wireless/ath/ath9k/hif_usb.c 	list_move_tail(&tx_buf->list, &hif_dev->tx.tx_pending);
tx                328 drivers/net/wireless/ath/ath9k/hif_usb.c 	hif_dev->tx.tx_buf_cnt--;
tx                330 drivers/net/wireless/ath/ath9k/hif_usb.c 	tx_skb_cnt = min_t(u16, hif_dev->tx.tx_skb_cnt, MAX_TX_AGGR_NUM);
tx                333 drivers/net/wireless/ath/ath9k/hif_usb.c 		nskb = __skb_dequeue(&hif_dev->tx.tx_skb_queue);
tx                338 drivers/net/wireless/ath/ath9k/hif_usb.c 		hif_dev->tx.tx_skb_cnt--;
tx                369 drivers/net/wireless/ath/ath9k/hif_usb.c 		list_move_tail(&tx_buf->list, &hif_dev->tx.tx_buf);
tx                370 drivers/net/wireless/ath/ath9k/hif_usb.c 		hif_dev->tx.tx_buf_cnt++;
tx                385 drivers/net/wireless/ath/ath9k/hif_usb.c 	spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
tx                387 drivers/net/wireless/ath/ath9k/hif_usb.c 	if (hif_dev->tx.flags & HIF_USB_TX_STOP) {
tx                388 drivers/net/wireless/ath/ath9k/hif_usb.c 		spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
tx                393 drivers/net/wireless/ath/ath9k/hif_usb.c 	if (hif_dev->tx.tx_skb_cnt > MAX_TX_BUF_NUM) {
tx                394 drivers/net/wireless/ath/ath9k/hif_usb.c 		spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
tx                398 drivers/net/wireless/ath/ath9k/hif_usb.c 	spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
tx                408 drivers/net/wireless/ath/ath9k/hif_usb.c 	spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
tx                412 drivers/net/wireless/ath/ath9k/hif_usb.c 		__skb_queue_tail(&hif_dev->tx.tx_skb_queue, skb);
tx                413 drivers/net/wireless/ath/ath9k/hif_usb.c 		hif_dev->tx.tx_skb_cnt++;
tx                417 drivers/net/wireless/ath/ath9k/hif_usb.c 	if ((hif_dev->tx.tx_buf_cnt == MAX_TX_URB_NUM) &&
tx                418 drivers/net/wireless/ath/ath9k/hif_usb.c 	    (hif_dev->tx.tx_skb_cnt < 2)) {
tx                422 drivers/net/wireless/ath/ath9k/hif_usb.c 	spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
tx                434 drivers/net/wireless/ath/ath9k/hif_usb.c 	spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
tx                435 drivers/net/wireless/ath/ath9k/hif_usb.c 	hif_dev->tx.flags &= ~HIF_USB_TX_STOP;
tx                436 drivers/net/wireless/ath/ath9k/hif_usb.c 	spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
tx                445 drivers/net/wireless/ath/ath9k/hif_usb.c 	spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
tx                446 drivers/net/wireless/ath/ath9k/hif_usb.c 	ath9k_skb_queue_complete(hif_dev, &hif_dev->tx.tx_skb_queue, false);
tx                447 drivers/net/wireless/ath/ath9k/hif_usb.c 	hif_dev->tx.tx_skb_cnt = 0;
tx                448 drivers/net/wireless/ath/ath9k/hif_usb.c 	hif_dev->tx.flags |= HIF_USB_TX_STOP;
tx                449 drivers/net/wireless/ath/ath9k/hif_usb.c 	spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
tx                453 drivers/net/wireless/ath/ath9k/hif_usb.c 				 &hif_dev->tx.tx_pending, list) {
tx                501 drivers/net/wireless/ath/ath9k/hif_usb.c 	spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
tx                503 drivers/net/wireless/ath/ath9k/hif_usb.c 	skb_queue_walk_safe(&hif_dev->tx.tx_skb_queue, skb, tmp) {
tx                505 drivers/net/wireless/ath/ath9k/hif_usb.c 			__skb_unlink(skb, &hif_dev->tx.tx_skb_queue);
tx                508 drivers/net/wireless/ath/ath9k/hif_usb.c 			hif_dev->tx.tx_skb_cnt--;
tx                513 drivers/net/wireless/ath/ath9k/hif_usb.c 	spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
tx                764 drivers/net/wireless/ath/ath9k/hif_usb.c 				 &hif_dev->tx.tx_buf, list) {
tx                772 drivers/net/wireless/ath/ath9k/hif_usb.c 	spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
tx                773 drivers/net/wireless/ath/ath9k/hif_usb.c 	hif_dev->tx.flags |= HIF_USB_TX_FLUSH;
tx                774 drivers/net/wireless/ath/ath9k/hif_usb.c 	spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
tx                777 drivers/net/wireless/ath/ath9k/hif_usb.c 				 &hif_dev->tx.tx_pending, list) {
tx                793 drivers/net/wireless/ath/ath9k/hif_usb.c 	INIT_LIST_HEAD(&hif_dev->tx.tx_buf);
tx                794 drivers/net/wireless/ath/ath9k/hif_usb.c 	INIT_LIST_HEAD(&hif_dev->tx.tx_pending);
tx                795 drivers/net/wireless/ath/ath9k/hif_usb.c 	spin_lock_init(&hif_dev->tx.tx_lock);
tx                796 drivers/net/wireless/ath/ath9k/hif_usb.c 	__skb_queue_head_init(&hif_dev->tx.tx_skb_queue);
tx                815 drivers/net/wireless/ath/ath9k/hif_usb.c 		list_add_tail(&tx_buf->list, &hif_dev->tx.tx_buf);
tx                818 drivers/net/wireless/ath/ath9k/hif_usb.c 	hif_dev->tx.tx_buf_cnt = MAX_TX_URB_NUM;
tx                123 drivers/net/wireless/ath/ath9k/hif_usb.h 	struct hif_usb_tx tx;
tx                499 drivers/net/wireless/ath/ath9k/htc.h 	struct ath9k_htc_tx tx;
tx                180 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c 			spin_lock_bh(&priv->tx.tx_lock);
tx                181 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c 			priv->tx.queued_cnt++;
tx                182 drivers/net/wireless/ath/ath9k/htc_drv_beacon.c 			spin_unlock_bh(&priv->tx.tx_lock);
tx                296 drivers/net/wireless/ath/ath9k/htc_drv_debug.c 	spin_lock_bh(&priv->tx.tx_lock);
tx                300 drivers/net/wireless/ath/ath9k/htc_drv_debug.c 			MAX_TX_BUF_NUM, priv->tx.tx_slot,
tx                301 drivers/net/wireless/ath/ath9k/htc_drv_debug.c 			bitmap_weight(priv->tx.tx_slot, MAX_TX_BUF_NUM));
tx                302 drivers/net/wireless/ath/ath9k/htc_drv_debug.c 	spin_unlock_bh(&priv->tx.tx_lock);
tx                321 drivers/net/wireless/ath/ath9k/htc_drv_debug.c 			 "Mgmt endpoint", skb_queue_len(&priv->tx.mgmt_ep_queue));
tx                324 drivers/net/wireless/ath/ath9k/htc_drv_debug.c 			 "Cab endpoint", skb_queue_len(&priv->tx.cab_ep_queue));
tx                327 drivers/net/wireless/ath/ath9k/htc_drv_debug.c 			 "Data BE endpoint", skb_queue_len(&priv->tx.data_be_queue));
tx                330 drivers/net/wireless/ath/ath9k/htc_drv_debug.c 			 "Data BK endpoint", skb_queue_len(&priv->tx.data_bk_queue));
tx                333 drivers/net/wireless/ath/ath9k/htc_drv_debug.c 			 "Data VI endpoint", skb_queue_len(&priv->tx.data_vi_queue));
tx                336 drivers/net/wireless/ath/ath9k/htc_drv_debug.c 			 "Data VO endpoint", skb_queue_len(&priv->tx.data_vo_queue));
tx                339 drivers/net/wireless/ath/ath9k/htc_drv_debug.c 			 "Failed queue", skb_queue_len(&priv->tx.tx_failed));
tx                341 drivers/net/wireless/ath/ath9k/htc_drv_debug.c 	spin_lock_bh(&priv->tx.tx_lock);
tx                343 drivers/net/wireless/ath/ath9k/htc_drv_debug.c 			 "Queued count", priv->tx.queued_cnt);
tx                344 drivers/net/wireless/ath/ath9k/htc_drv_debug.c 	spin_unlock_bh(&priv->tx.tx_lock);
tx                122 drivers/net/wireless/ath/ath9k/htc_drv_init.c 					void (*tx) (void *,
tx                135 drivers/net/wireless/ath/ath9k/htc_drv_init.c 	req.ep_callbacks.tx = tx;
tx                645 drivers/net/wireless/ath/ath9k/htc_drv_init.c 	spin_lock_init(&priv->tx.tx_lock);
tx                655 drivers/net/wireless/ath/ath9k/htc_drv_init.c 	timer_setup(&priv->tx.cleanup_timer, ath9k_htc_tx_cleanup_timer, 0);
tx                201 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	del_timer_sync(&priv->tx.cleanup_timer);
tx                233 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	mod_timer(&priv->tx.cleanup_timer,
tx                263 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	del_timer_sync(&priv->tx.cleanup_timer);
tx                311 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	mod_timer(&priv->tx.cleanup_timer,
tx                735 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	spin_lock_bh(&priv->tx.tx_lock);
tx                737 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	spin_unlock_bh(&priv->tx.tx_lock);
tx                960 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	spin_lock_bh(&priv->tx.tx_lock);
tx                961 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	priv->tx.flags &= ~ATH9K_HTC_OP_TX_QUEUES_STOP;
tx                962 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	spin_unlock_bh(&priv->tx.tx_lock);
tx                966 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	mod_timer(&priv->tx.cleanup_timer,
tx               1000 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	del_timer_sync(&priv->tx.cleanup_timer);
tx               1691 drivers/net/wireless/ath/ath9k/htc_drv_main.c 		spin_lock_bh(&priv->tx.tx_lock);
tx               1693 drivers/net/wireless/ath/ath9k/htc_drv_main.c 		spin_unlock_bh(&priv->tx.tx_lock);
tx               1870 drivers/net/wireless/ath/ath9k/htc_drv_main.c 	.tx                 = ath9k_htc_tx,
tx                 58 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	spin_lock_bh(&priv->tx.tx_lock);
tx                 59 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	priv->tx.queued_cnt++;
tx                 60 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	if ((priv->tx.queued_cnt >= ATH9K_HTC_TX_THRESHOLD) &&
tx                 61 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	    !(priv->tx.flags & ATH9K_HTC_OP_TX_QUEUES_STOP)) {
tx                 62 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 		priv->tx.flags |= ATH9K_HTC_OP_TX_QUEUES_STOP;
tx                 65 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	spin_unlock_bh(&priv->tx.tx_lock);
tx                 70 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	spin_lock_bh(&priv->tx.tx_lock);
tx                 71 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	if ((priv->tx.queued_cnt < ATH9K_HTC_TX_THRESHOLD) &&
tx                 72 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	    (priv->tx.flags & ATH9K_HTC_OP_TX_QUEUES_STOP)) {
tx                 73 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 		priv->tx.flags &= ~ATH9K_HTC_OP_TX_QUEUES_STOP;
tx                 76 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	spin_unlock_bh(&priv->tx.tx_lock);
tx                 83 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	spin_lock_bh(&priv->tx.tx_lock);
tx                 84 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	slot = find_first_zero_bit(priv->tx.tx_slot, MAX_TX_BUF_NUM);
tx                 86 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 		spin_unlock_bh(&priv->tx.tx_lock);
tx                 89 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	__set_bit(slot, priv->tx.tx_slot);
tx                 90 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	spin_unlock_bh(&priv->tx.tx_lock);
tx                 97 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	spin_lock_bh(&priv->tx.tx_lock);
tx                 98 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	__clear_bit(slot, priv->tx.tx_slot);
tx                 99 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	spin_unlock_bh(&priv->tx.tx_lock);
tx                137 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 		epid_queue = &priv->tx.mgmt_ep_queue;
tx                139 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 		epid_queue = &priv->tx.cab_ep_queue;
tx                141 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 		epid_queue = &priv->tx.data_be_queue;
tx                143 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 		epid_queue = &priv->tx.data_bk_queue;
tx                145 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 		epid_queue = &priv->tx.data_vi_queue;
tx                147 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 		epid_queue = &priv->tx.data_vo_queue;
tx                392 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	spin_lock_bh(&priv->tx.tx_lock);
tx                395 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	spin_unlock_bh(&priv->tx.tx_lock);
tx                430 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 				spin_lock_bh(&priv->tx.tx_lock);
tx                432 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 				spin_unlock_bh(&priv->tx.tx_lock);
tx                504 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	spin_lock_bh(&priv->tx.tx_lock);
tx                505 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	if (WARN_ON(--priv->tx.queued_cnt < 0))
tx                506 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 		priv->tx.queued_cnt = 0;
tx                507 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	spin_unlock_bh(&priv->tx.tx_lock);
tx                538 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	spin_lock_bh(&priv->tx.tx_lock);
tx                539 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	priv->tx.flags |= ATH9K_HTC_OP_TX_DRAIN;
tx                540 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	spin_unlock_bh(&priv->tx.tx_lock);
tx                550 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	ath9k_htc_tx_drainq(priv, &priv->tx.mgmt_ep_queue);
tx                551 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	ath9k_htc_tx_drainq(priv, &priv->tx.cab_ep_queue);
tx                552 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	ath9k_htc_tx_drainq(priv, &priv->tx.data_be_queue);
tx                553 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	ath9k_htc_tx_drainq(priv, &priv->tx.data_bk_queue);
tx                554 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	ath9k_htc_tx_drainq(priv, &priv->tx.data_vi_queue);
tx                555 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	ath9k_htc_tx_drainq(priv, &priv->tx.data_vo_queue);
tx                556 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	ath9k_htc_tx_drainq(priv, &priv->tx.tx_failed);
tx                568 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	spin_lock_bh(&priv->tx.tx_lock);
tx                569 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	priv->tx.flags &= ~ATH9K_HTC_OP_TX_DRAIN;
tx                570 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	spin_unlock_bh(&priv->tx.tx_lock);
tx                577 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	spin_lock(&priv->tx.tx_lock);
tx                578 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	if (priv->tx.flags & ATH9K_HTC_OP_TX_DRAIN) {
tx                579 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 		spin_unlock(&priv->tx.tx_lock);
tx                582 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	spin_unlock(&priv->tx.tx_lock);
tx                584 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	ath9k_htc_tx_drainq(priv, &priv->tx.tx_failed);
tx                696 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 		skb_queue_tail(&priv->tx.tx_failed, skb);
tx                757 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	struct ath9k_htc_priv *priv = from_timer(priv, t, tx.cleanup_timer);
tx                788 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	ath9k_htc_tx_cleanup_queue(priv, &priv->tx.mgmt_ep_queue);
tx                789 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	ath9k_htc_tx_cleanup_queue(priv, &priv->tx.cab_ep_queue);
tx                790 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	ath9k_htc_tx_cleanup_queue(priv, &priv->tx.data_be_queue);
tx                791 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	ath9k_htc_tx_cleanup_queue(priv, &priv->tx.data_bk_queue);
tx                792 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	ath9k_htc_tx_cleanup_queue(priv, &priv->tx.data_vi_queue);
tx                793 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	ath9k_htc_tx_cleanup_queue(priv, &priv->tx.data_vo_queue);
tx                798 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	mod_timer(&priv->tx.cleanup_timer,
tx                804 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	skb_queue_head_init(&priv->tx.mgmt_ep_queue);
tx                805 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	skb_queue_head_init(&priv->tx.cab_ep_queue);
tx                806 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	skb_queue_head_init(&priv->tx.data_be_queue);
tx                807 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	skb_queue_head_init(&priv->tx.data_bk_queue);
tx                808 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	skb_queue_head_init(&priv->tx.data_vi_queue);
tx                809 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	skb_queue_head_init(&priv->tx.data_vo_queue);
tx                810 drivers/net/wireless/ath/ath9k/htc_drv_txrx.c 	skb_queue_head_init(&priv->tx.tx_failed);
tx                345 drivers/net/wireless/ath/ath9k/htc_hst.c 		if (endpoint->ep_callbacks.tx) {
tx                346 drivers/net/wireless/ath/ath9k/htc_hst.c 			endpoint->ep_callbacks.tx(endpoint->ep_callbacks.priv,
tx                 94 drivers/net/wireless/ath/ath9k/htc_hst.h 	void (*tx) (void *, struct sk_buff *, enum htc_endpoint_id, bool txok);
tx               1127 drivers/net/wireless/ath/ath9k/hw.h void ar9003_hw_set_chain_masks(struct ath_hw *ah, u8 rx, u8 tx);
tx                414 drivers/net/wireless/ath/ath9k/init.c 	sc->tx.uapsdq = ath_txq_setup(sc, ATH9K_TX_QUEUE_UAPSD, 0);
tx                417 drivers/net/wireless/ath/ath9k/init.c 		sc->tx.txq_map[i] = ath_txq_setup(sc, ATH9K_TX_QUEUE_DATA, i);
tx                418 drivers/net/wireless/ath/ath9k/init.c 		sc->tx.txq_map[i]->mac80211_qnum = i;
tx                784 drivers/net/wireless/ath/ath9k/init.c 			ath_tx_cleanupq(sc, &sc->tx.txq[i]);
tx               1109 drivers/net/wireless/ath/ath9k/init.c 			ath_tx_cleanupq(sc, &sc->tx.txq[i]);
tx                 32 drivers/net/wireless/ath/ath9k/link.c 		txq = sc->tx.txq_map[i];
tx                182 drivers/net/wireless/ath/ath9k/link.c 	txctl.txq = sc->tx.txq_map[IEEE80211_AC_BE];
tx                299 drivers/net/wireless/ath/ath9k/mac.h 		} tx;
tx                317 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl2     u.tx.ctl2
tx                318 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl3     u.tx.ctl3
tx                319 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl4     u.tx.ctl4
tx                320 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl5     u.tx.ctl5
tx                321 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl6     u.tx.ctl6
tx                322 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl7     u.tx.ctl7
tx                323 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl8     u.tx.ctl8
tx                324 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl9     u.tx.ctl9
tx                325 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl10    u.tx.ctl10
tx                326 drivers/net/wireless/ath/ath9k/mac.h #define ds_ctl11    u.tx.ctl11
tx                328 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus0    u.tx.status0
tx                329 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus1    u.tx.status1
tx                330 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus2    u.tx.status2
tx                331 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus3    u.tx.status3
tx                332 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus4    u.tx.status4
tx                333 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus5    u.tx.status5
tx                334 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus6    u.tx.status6
tx                335 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus7    u.tx.status7
tx                336 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus8    u.tx.status8
tx                337 drivers/net/wireless/ath/ath9k/mac.h #define ds_txstatus9    u.tx.status9
tx                805 drivers/net/wireless/ath/ath9k/main.c 	txctl.txq = sc->tx.txq_map[skb_get_queue_mapping(skb)];
tx               1643 drivers/net/wireless/ath/ath9k/main.c 	txq = sc->tx.txq_map[queue];
tx               2048 drivers/net/wireless/ath/ath9k/main.c 		npend = ath9k_has_pending_frames(sc, &sc->tx.txq[i],
tx               2664 drivers/net/wireless/ath/ath9k/main.c 	.tx 		    = ath9k_tx,
tx                126 drivers/net/wireless/ath/ath9k/tx99.c 	txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO];
tx                173 drivers/net/wireless/ath/ath9k/wmi.c 			spin_lock_bh(&priv->tx.tx_lock);
tx                174 drivers/net/wireless/ath/ath9k/wmi.c 			if (priv->tx.flags & ATH9K_HTC_OP_TX_DRAIN) {
tx                175 drivers/net/wireless/ath/ath9k/wmi.c 				spin_unlock_bh(&priv->tx.tx_lock);
tx                178 drivers/net/wireless/ath/ath9k/wmi.c 			spin_unlock_bh(&priv->tx.tx_lock);
tx                264 drivers/net/wireless/ath/ath9k/wmi.c 	connect.ep_callbacks.tx = ath9k_wmi_ctrl_tx;
tx                173 drivers/net/wireless/ath/ath9k/xmit.c 	txq = sc->tx.txq_map[q];
tx                211 drivers/net/wireless/ath/ath9k/xmit.c 	if (tid->txq == sc->tx.txq_map[q]) {
tx                368 drivers/net/wireless/ath/ath9k/xmit.c 	spin_lock_bh(&sc->tx.txbuflock);
tx                370 drivers/net/wireless/ath/ath9k/xmit.c 	if (unlikely(list_empty(&sc->tx.txbuf))) {
tx                371 drivers/net/wireless/ath/ath9k/xmit.c 		spin_unlock_bh(&sc->tx.txbuflock);
tx                375 drivers/net/wireless/ath/ath9k/xmit.c 	bf = list_first_entry(&sc->tx.txbuf, struct ath_buf, list);
tx                378 drivers/net/wireless/ath/ath9k/xmit.c 	spin_unlock_bh(&sc->tx.txbuflock);
tx                385 drivers/net/wireless/ath/ath9k/xmit.c 	spin_lock_bh(&sc->tx.txbuflock);
tx                386 drivers/net/wireless/ath/ath9k/xmit.c 	list_add_tail(&bf->list, &sc->tx.txbuf);
tx                387 drivers/net/wireless/ath/ath9k/xmit.c 	spin_unlock_bh(&sc->tx.txbuflock);
tx                792 drivers/net/wireless/ath/ath9k/xmit.c 		frmlen = sc->tx.max_aggr_framelen[q][modeidx][rates[i].idx];
tx               1118 drivers/net/wireless/ath/ath9k/xmit.c 	cur_ht20 = sc->tx.max_aggr_framelen[queue][MCS_HT20];
tx               1119 drivers/net/wireless/ath/ath9k/xmit.c 	cur_ht20_sgi = sc->tx.max_aggr_framelen[queue][MCS_HT20_SGI];
tx               1120 drivers/net/wireless/ath/ath9k/xmit.c 	cur_ht40 = sc->tx.max_aggr_framelen[queue][MCS_HT40];
tx               1121 drivers/net/wireless/ath/ath9k/xmit.c 	cur_ht40_sgi = sc->tx.max_aggr_framelen[queue][MCS_HT40_SGI];
tx               1374 drivers/net/wireless/ath/ath9k/xmit.c 			    txq == sc->tx.uapsdq)
tx               1631 drivers/net/wireless/ath/ath9k/xmit.c 	struct ath_txq *txq = sc->tx.uapsdq;
tx               1649 drivers/net/wireless/ath/ath9k/xmit.c 			ret = ath_tx_get_tid_subframe(sc, sc->tx.uapsdq,
tx               1745 drivers/net/wireless/ath/ath9k/xmit.c 		struct ath_txq *txq = &sc->tx.txq[axq_qnum];
tx               1756 drivers/net/wireless/ath/ath9k/xmit.c 		sc->tx.txqsetup |= 1<<axq_qnum;
tx               1762 drivers/net/wireless/ath/ath9k/xmit.c 	return &sc->tx.txq[axq_qnum];
tx               1772 drivers/net/wireless/ath/ath9k/xmit.c 	BUG_ON(sc->tx.txq[qnum].axq_qnum != qnum);
tx               1882 drivers/net/wireless/ath/ath9k/xmit.c 		if (!sc->tx.txq[i].axq_depth)
tx               1885 drivers/net/wireless/ath/ath9k/xmit.c 		if (ath9k_hw_numtxpending(ah, sc->tx.txq[i].axq_qnum))
tx               1899 drivers/net/wireless/ath/ath9k/xmit.c 		txq = &sc->tx.txq[i];
tx               1909 drivers/net/wireless/ath/ath9k/xmit.c 	sc->tx.txqsetup &= ~(1<<txq->axq_qnum);
tx               1960 drivers/net/wireless/ath/ath9k/xmit.c 		txq = sc->tx.txq_map[i];
tx               2294 drivers/net/wireless/ath/ath9k/xmit.c 		txq = sc->tx.uapsdq;
tx               2302 drivers/net/wireless/ath/ath9k/xmit.c 	if (txq == sc->tx.txq_map[q]) {
tx               2490 drivers/net/wireless/ath/ath9k/xmit.c 	spin_lock_irqsave(&sc->tx.txbuflock, flags);
tx               2491 drivers/net/wireless/ath/ath9k/xmit.c 	list_splice_tail_init(bf_q, &sc->tx.txbuf);
tx               2492 drivers/net/wireless/ath/ath9k/xmit.c 	spin_unlock_irqrestore(&sc->tx.txbuflock, flags);
tx               2637 drivers/net/wireless/ath/ath9k/xmit.c 			ath_tx_processq(sc, &sc->tx.txq[i]);
tx               2680 drivers/net/wireless/ath/ath9k/xmit.c 		txq = &sc->tx.txq[ts.qid];
tx               2763 drivers/net/wireless/ath/ath9k/xmit.c 	spin_lock_init(&sc->tx.txbuflock);
tx               2765 drivers/net/wireless/ath/ath9k/xmit.c 	error = ath_descdma_setup(sc, &sc->tx.txdma, &sc->tx.txbuf,
tx               2804 drivers/net/wireless/ath/ath9k/xmit.c 		tid->txq = sc->tx.txq_map[acno];
tx               1753 drivers/net/wireless/ath/carl9170/main.c 	.tx			= carl9170_op_tx,
tx                144 drivers/net/wireless/ath/hw.c 	u32 cycles, busy, rx, tx;
tx                154 drivers/net/wireless/ath/hw.c 	tx = REG_READ(ah, AR_TFCNT);
tx                169 drivers/net/wireless/ath/hw.c 	common->cc_ani.tx_frame += tx;
tx                174 drivers/net/wireless/ath/hw.c 	common->cc_survey.tx_frame += tx;
tx               1149 drivers/net/wireless/ath/wcn36xx/main.c 	.tx			= wcn36xx_tx,
tx                286 drivers/net/wireless/ath/wil6210/cfg80211.c 		.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
tx                292 drivers/net/wireless/ath/wil6210/cfg80211.c 		.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
tx                307 drivers/net/wireless/ath/wil6210/cfg80211.c 		.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
tx                313 drivers/net/wireless/ath/wil6210/cfg80211.c 		.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
tx                319 drivers/net/wireless/ath/wil6210/cfg80211.c 		.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
tx                 73 drivers/net/wireless/ath/wil6210/debugfs.c 			&ring->va[idx].tx.enhanced;
tx                133 drivers/net/wireless/ath/wil6210/debugfs.c 					&ring->va[i].tx.legacy;
tx                375 drivers/net/wireless/ath/wil6210/debugfs.c 		       offsetof(struct wil6210_mbox_ctl, tx));
tx               1083 drivers/net/wireless/ath/wil6210/debugfs.c 	bool tx;
tx               1096 drivers/net/wireless/ath/wil6210/debugfs.c 		tx = ring_idx > 0; /* desc ring 0 is reserved for RX */
tx               1103 drivers/net/wireless/ath/wil6210/debugfs.c 		tx = (ring_idx < WIL6210_MAX_TX_RINGS);
tx               1106 drivers/net/wireless/ath/wil6210/debugfs.c 	ring = tx ? &wil->ring_tx[ring_idx] : &wil->ring_rx;
tx               1109 drivers/net/wireless/ath/wil6210/debugfs.c 		if (tx)
tx               1117 drivers/net/wireless/ath/wil6210/debugfs.c 		if (tx)
tx               1129 drivers/net/wireless/ath/wil6210/debugfs.c 	d = &ring->va[txdesc_idx].tx.legacy;
tx               1134 drivers/net/wireless/ath/wil6210/debugfs.c 		if (tx) {
tx               1151 drivers/net/wireless/ath/wil6210/debugfs.c 	if (tx)
tx               1179 drivers/net/wireless/ath/wil6210/debugfs.c 	bool tx;
tx               1189 drivers/net/wireless/ath/wil6210/debugfs.c 	tx = !sring->is_rx;
tx               1192 drivers/net/wireless/ath/wil6210/debugfs.c 		seq_printf(s, "No %cX status ring\n", tx ? 'T' : 'R');
tx               1198 drivers/net/wireless/ath/wil6210/debugfs.c 			   tx ? 'T' : 'R', status_msg_idx, sring->size);
tx               1205 drivers/net/wireless/ath/wil6210/debugfs.c 		   tx ? 'T' : 'R', sring_idx, status_msg_idx);
tx               1209 drivers/net/wireless/ath/wil6210/debugfs.c 	if (!tx && !wil->use_compressed_rx_status)
tx               1478 drivers/net/wireless/ath/wil6210/debugfs.c 	int tx = atomic_xchg(&wil->isr_count_tx, 0);
tx               1487 drivers/net/wireless/ath/wil6210/debugfs.c 	seq_printf(s, "Tx irqs:packets : %8d : %8ld\n", tx, txf - txf_old);
tx                522 drivers/net/wireless/ath/wil6210/interrupt.c 	wil_mbox_ring_le2cpus(&wil->mbox_ctl.tx);
tx                535 drivers/net/wireless/ath/wil6210/interrupt.c 	if (wil->mbox_ctl.tx.entry_size < min_size) {
tx                537 drivers/net/wireless/ath/wil6210/interrupt.c 			wil->mbox_ctl.tx.entry_size);
tx                170 drivers/net/wireless/ath/wil6210/txrx.c 			&vring->va[i].tx.legacy;
tx                226 drivers/net/wireless/ath/wil6210/txrx.c 					&vring->va[vring->swtail].tx.legacy;
tx               1825 drivers/net/wireless/ath/wil6210/txrx.c 	_hdr_desc = &vring->va[i].tx.legacy;
tx               1891 drivers/net/wireless/ath/wil6210/txrx.c 			_desc = &vring->va[i].tx.legacy;
tx               2020 drivers/net/wireless/ath/wil6210/txrx.c 		d = (struct vring_tx_desc *)&vring->va[i].tx.legacy;
tx               2021 drivers/net/wireless/ath/wil6210/txrx.c 		_desc = &vring->va[i].tx.legacy;
tx               2063 drivers/net/wireless/ath/wil6210/txrx.c 	_d = &ring->va[i].tx.legacy;
tx               2103 drivers/net/wireless/ath/wil6210/txrx.c 		_d = &ring->va[i].tx.legacy;
tx               2178 drivers/net/wireless/ath/wil6210/txrx.c 		_d = &ring->va[i].tx.legacy;
tx               2469 drivers/net/wireless/ath/wil6210/txrx.c 		_d = &vring->va[lf].tx.legacy;
tx               2481 drivers/net/wireless/ath/wil6210/txrx.c 			_d = &vring->va[vring->swtail].tx.legacy;
tx                497 drivers/net/wireless/ath/wil6210/txrx.h 	union wil_tx_desc tx;
tx                468 drivers/net/wireless/ath/wil6210/txrx_edma.c 			&ring->va[ring->swtail].tx.enhanced;
tx               1235 drivers/net/wireless/ath/wil6210/txrx_edma.c 				&ring->va[ring->swtail].tx.enhanced;
tx               1363 drivers/net/wireless/ath/wil6210/txrx_edma.c 		&ring->va[i].tx.enhanced;
tx               1538 drivers/net/wireless/ath/wil6210/txrx_edma.c 			&ring->va[i].tx.enhanced;
tx                480 drivers/net/wireless/ath/wil6210/wil6210.h 	struct wil6210_mbox_ring tx;
tx                660 drivers/net/wireless/ath/wil6210/wmi.c 	struct wil6210_mbox_ring *r = &wil->mbox_ctl.tx;
tx                721 drivers/net/wireless/ath/wil6210/wmi.c 				offsetof(struct wil6210_mbox_ctl, tx.tail));
tx                751 drivers/net/wireless/ath/wil6210/wmi.c 	wil_w(wil, RGF_MBOX + offsetof(struct wil6210_mbox_ctl, tx.head),
tx               2182 drivers/net/wireless/atmel/at76c50x-usb.c 	.tx = at76_mac80211_tx,
tx                289 drivers/net/wireless/broadcom/b43/dma.c 		       ring->tx ? "TX" : "RX", ring->index);
tx                304 drivers/net/wireless/broadcom/b43/dma.c 	B43_WARN_ON(!ring->tx);
tx                348 drivers/net/wireless/broadcom/b43/dma.c 			      unsigned char *buf, size_t len, int tx)
tx                352 drivers/net/wireless/broadcom/b43/dma.c 	if (tx) {
tx                365 drivers/net/wireless/broadcom/b43/dma.c 			  dma_addr_t addr, size_t len, int tx)
tx                367 drivers/net/wireless/broadcom/b43/dma.c 	if (tx) {
tx                380 drivers/net/wireless/broadcom/b43/dma.c 	B43_WARN_ON(ring->tx);
tx                389 drivers/net/wireless/broadcom/b43/dma.c 	B43_WARN_ON(ring->tx);
tx                399 drivers/net/wireless/broadcom/b43/dma.c 		if (ring->tx)
tx                599 drivers/net/wireless/broadcom/b43/dma.c 	B43_WARN_ON(ring->tx);
tx                681 drivers/net/wireless/broadcom/b43/dma.c 	if (ring->tx) {
tx                755 drivers/net/wireless/broadcom/b43/dma.c 	if (ring->tx) {
tx                786 drivers/net/wireless/broadcom/b43/dma.c 			B43_WARN_ON(!ring->tx);
tx                789 drivers/net/wireless/broadcom/b43/dma.c 		if (ring->tx) {
tx                865 drivers/net/wireless/broadcom/b43/dma.c 		ring->tx = true;
tx               1360 drivers/net/wireless/broadcom/b43/dma.c 	B43_WARN_ON(!ring->tx);
tx               1430 drivers/net/wireless/broadcom/b43/dma.c 	B43_WARN_ON(!ring->tx);
tx               1672 drivers/net/wireless/broadcom/b43/dma.c 	B43_WARN_ON(ring->tx);
tx               1693 drivers/net/wireless/broadcom/b43/dma.c 	B43_WARN_ON(ring->tx);
tx               1709 drivers/net/wireless/broadcom/b43/dma.c 	B43_WARN_ON(!ring->tx);
tx               1715 drivers/net/wireless/broadcom/b43/dma.c 	B43_WARN_ON(!ring->tx);
tx                255 drivers/net/wireless/broadcom/b43/dma.h 	bool tx;
tx               5171 drivers/net/wireless/broadcom/b43/main.c 	.tx			= b43_op_tx,
tx                746 drivers/net/wireless/broadcom/b43/phy_lp.c static void lpphy_set_trsw_over(struct b43_wldev *dev, bool tx, bool rx)
tx                748 drivers/net/wireless/broadcom/b43/phy_lp.c 	u16 trsw = (tx << 1) | rx;
tx               1839 drivers/net/wireless/broadcom/b43/phy_lp.c static int lpphy_rx_iq_cal(struct b43_wldev *dev, bool noise, bool tx,
tx               1875 drivers/net/wireless/broadcom/b43/phy_lp.c 		tx = true;
tx               1880 drivers/net/wireless/broadcom/b43/phy_lp.c 	lpphy_set_trsw_over(dev, tx, rx);
tx                 32 drivers/net/wireless/broadcom/b43/radio_2056.c 	const struct b2056_inittab_entry *tx;
tx               3011 drivers/net/wireless/broadcom/b43/radio_2056.c 		.tx		= prefix##_tx,			\
tx               10229 drivers/net/wireless/broadcom/b43/radio_2056.c 				B2056_TX0, pts->tx, pts->tx_length);
tx               10231 drivers/net/wireless/broadcom/b43/radio_2056.c 				B2056_TX1, pts->tx, pts->tx_length);
tx                149 drivers/net/wireless/broadcom/b43legacy/dma.c 		       ring->tx ? "TX" : "RX",
tx                165 drivers/net/wireless/broadcom/b43legacy/dma.c 	B43legacy_WARN_ON(!ring->tx);
tx                250 drivers/net/wireless/broadcom/b43legacy/dma.c 			  int tx)
tx                254 drivers/net/wireless/broadcom/b43legacy/dma.c 	if (tx)
tx                270 drivers/net/wireless/broadcom/b43legacy/dma.c 		      int tx)
tx                272 drivers/net/wireless/broadcom/b43legacy/dma.c 	if (tx)
tx                287 drivers/net/wireless/broadcom/b43legacy/dma.c 	B43legacy_WARN_ON(ring->tx);
tx                298 drivers/net/wireless/broadcom/b43legacy/dma.c 	B43legacy_WARN_ON(ring->tx);
tx                450 drivers/net/wireless/broadcom/b43legacy/dma.c 	B43legacy_WARN_ON(ring->tx);
tx                535 drivers/net/wireless/broadcom/b43legacy/dma.c 	if (ring->tx) {
tx                571 drivers/net/wireless/broadcom/b43legacy/dma.c 	if (ring->tx) {
tx                593 drivers/net/wireless/broadcom/b43legacy/dma.c 			B43legacy_WARN_ON(!ring->tx);
tx                596 drivers/net/wireless/broadcom/b43legacy/dma.c 		if (ring->tx)
tx                689 drivers/net/wireless/broadcom/b43legacy/dma.c 		ring->tx = true;
tx                735 drivers/net/wireless/broadcom/b43legacy/dma.c 		     (ring->tx) ? "TX" : "RX", ring->max_used_slots,
tx               1076 drivers/net/wireless/broadcom/b43legacy/dma.c 	B43legacy_WARN_ON(!ring->tx);
tx               1134 drivers/net/wireless/broadcom/b43legacy/dma.c 	B43legacy_WARN_ON(!ring->tx);
tx               1345 drivers/net/wireless/broadcom/b43legacy/dma.c 	B43legacy_WARN_ON(ring->tx);
tx               1361 drivers/net/wireless/broadcom/b43legacy/dma.c 	B43legacy_WARN_ON(!ring->tx);
tx               1367 drivers/net/wireless/broadcom/b43legacy/dma.c 	B43legacy_WARN_ON(!ring->tx);
tx                149 drivers/net/wireless/broadcom/b43legacy/dma.h 	bool tx;
tx               3533 drivers/net/wireless/broadcom/b43legacy/main.c 	.tx			= b43legacy_op_tx,
tx               1291 drivers/net/wireless/broadcom/b43legacy/phy.c 					       u16 tx)
tx               1300 drivers/net/wireless/broadcom/b43legacy/phy.c 	if (tx == 3)
tx               6403 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c 		.tx = 0xffff,
tx               6408 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c 		.tx = 0xffff,
tx               6413 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c 		.tx = 0xffff,
tx               6423 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c 		.tx = 0xffff,
tx               6428 drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c 		.tx = 0xffff,
tx                305 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c static void brcmf_p2p_print_actframe(bool tx, void *frame, u32 frame_len)
tx                319 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                323 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                327 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                331 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                335 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                339 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                343 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                347 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                351 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                355 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                363 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                367 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                371 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                375 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                379 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                387 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                391 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                395 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                399 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                403 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c 				  (tx) ? "TX" : "RX");
tx                411 drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c static void brcmf_p2p_print_actframe(bool tx, void *frame, u32 frame_len)
tx                 70 drivers/net/wireless/broadcom/brcm80211/brcmsmac/d11.h 	struct pio2regs tx;
tx                 82 drivers/net/wireless/broadcom/brcm80211/brcmsmac/d11.h 	struct pio4regs tx;
tx                966 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 	.tx = brcms_ops_tx,
tx               1152 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c static void wlc_lcnphy_set_trsw_override(struct brcms_phy *pi, bool tx, bool rx)
tx               1157 drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c 		    (0x1 << 0), (tx ? (0x1 << 1) : 0) | (rx ? (0x1 << 0) : 0));
tx                969 drivers/net/wireless/cisco/airo.c 	u32		 tx;	    // Tx sequence number
tx               1319 drivers/net/wireless/cisco/airo.c 	cur->tx      = 0;  /* Tx sequence numbers */
tx               1418 drivers/net/wireless/cisco/airo.c 	mic->seq = htonl(context->tx);
tx               1419 drivers/net/wireless/cisco/airo.c 	context->tx += 2;
tx                307 drivers/net/wireless/intel/iwlegacy/3945-debug.c 	struct iwl39_stats_tx *tx, *accum_tx, *delta_tx, *max_tx;
tx                323 drivers/net/wireless/intel/iwlegacy/3945-debug.c 	tx = &il->_3945.stats.tx;
tx                324 drivers/net/wireless/intel/iwlegacy/3945-debug.c 	accum_tx = &il->_3945.accum_stats.tx;
tx                325 drivers/net/wireless/intel/iwlegacy/3945-debug.c 	delta_tx = &il->_3945.delta_stats.tx;
tx                326 drivers/net/wireless/intel/iwlegacy/3945-debug.c 	max_tx = &il->_3945.max_delta.tx;
tx                336 drivers/net/wireless/intel/iwlegacy/3945-debug.c 		      le32_to_cpu(tx->preamble_cnt), accum_tx->preamble_cnt,
tx                341 drivers/net/wireless/intel/iwlegacy/3945-debug.c 		      le32_to_cpu(tx->rx_detected_cnt),
tx                347 drivers/net/wireless/intel/iwlegacy/3945-debug.c 		      le32_to_cpu(tx->bt_prio_defer_cnt),
tx                353 drivers/net/wireless/intel/iwlegacy/3945-debug.c 		      le32_to_cpu(tx->bt_prio_kill_cnt),
tx                359 drivers/net/wireless/intel/iwlegacy/3945-debug.c 		      le32_to_cpu(tx->few_bytes_cnt), accum_tx->few_bytes_cnt,
tx                364 drivers/net/wireless/intel/iwlegacy/3945-debug.c 		      le32_to_cpu(tx->cts_timeout), accum_tx->cts_timeout,
tx                369 drivers/net/wireless/intel/iwlegacy/3945-debug.c 		      le32_to_cpu(tx->ack_timeout), accum_tx->ack_timeout,
tx                374 drivers/net/wireless/intel/iwlegacy/3945-debug.c 		      le32_to_cpu(tx->expected_ack_cnt),
tx                380 drivers/net/wireless/intel/iwlegacy/3945-debug.c 		      le32_to_cpu(tx->actual_ack_cnt), accum_tx->actual_ack_cnt,
tx               1493 drivers/net/wireless/intel/iwlegacy/3945-mac.c 		il->isr_stats.tx++;
tx               3445 drivers/net/wireless/intel/iwlegacy/3945-mac.c 	.tx = il3945_mac_tx,
tx               2423 drivers/net/wireless/intel/iwlegacy/3945.c 	tx_beacon_cmd->tx.sta_id = il->hw_params.bcast_id;
tx               2424 drivers/net/wireless/intel/iwlegacy/3945.c 	tx_beacon_cmd->tx.stop_time.life_time = TX_CMD_LIFE_TIME_INFINITE;
tx               2431 drivers/net/wireless/intel/iwlegacy/3945.c 	tx_beacon_cmd->tx.len = cpu_to_le16((u16) frame_size);
tx               2433 drivers/net/wireless/intel/iwlegacy/3945.c 	tx_beacon_cmd->tx.rate = rate;
tx               2434 drivers/net/wireless/intel/iwlegacy/3945.c 	tx_beacon_cmd->tx.tx_flags =
tx               2438 drivers/net/wireless/intel/iwlegacy/3945.c 	tx_beacon_cmd->tx.supp_rates[0] =
tx               2441 drivers/net/wireless/intel/iwlegacy/3945.c 	tx_beacon_cmd->tx.supp_rates[1] = (IL_CCK_BASIC_RATES_MASK & 0xF);
tx                461 drivers/net/wireless/intel/iwlegacy/4965-debug.c 	struct stats_tx *tx, *accum_tx, *delta_tx, *max_tx;
tx                476 drivers/net/wireless/intel/iwlegacy/4965-debug.c 	tx = &il->_4965.stats.tx;
tx                477 drivers/net/wireless/intel/iwlegacy/4965-debug.c 	accum_tx = &il->_4965.accum_stats.tx;
tx                478 drivers/net/wireless/intel/iwlegacy/4965-debug.c 	delta_tx = &il->_4965.delta_stats.tx;
tx                479 drivers/net/wireless/intel/iwlegacy/4965-debug.c 	max_tx = &il->_4965.max_delta.tx;
tx                485 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->preamble_cnt), accum_tx->preamble_cnt,
tx                489 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->rx_detected_cnt),
tx                494 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->bt_prio_defer_cnt),
tx                499 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->bt_prio_kill_cnt),
tx                504 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->few_bytes_cnt), accum_tx->few_bytes_cnt,
tx                508 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->cts_timeout), accum_tx->cts_timeout,
tx                512 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->ack_timeout), accum_tx->ack_timeout,
tx                516 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->expected_ack_cnt),
tx                521 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->actual_ack_cnt), accum_tx->actual_ack_cnt,
tx                525 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->dump_msdu_cnt), accum_tx->dump_msdu_cnt,
tx                530 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->burst_abort_next_frame_mismatch_cnt),
tx                537 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->burst_abort_missing_next_frame_cnt),
tx                544 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->cts_timeout_collision),
tx                551 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->ack_or_ba_timeout_collision),
tx                557 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->agg.ba_timeout), accum_tx->agg.ba_timeout,
tx                562 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->agg.ba_reschedule_frames),
tx                569 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->agg.scd_query_agg_frame_cnt),
tx                576 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->agg.scd_query_no_agg),
tx                582 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->agg.scd_query_agg),
tx                588 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->agg.scd_query_mismatch),
tx                594 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->agg.frame_not_ready),
tx                600 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->agg.underrun), accum_tx->agg.underrun,
tx                604 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->agg.bt_prio_kill),
tx                609 drivers/net/wireless/intel/iwlegacy/4965-debug.c 		      le32_to_cpu(tx->agg.rx_ba_rsp_cnt),
tx               1770 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	tx_cmd = &out_cmd->cmd.tx;
tx               3813 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	tx_beacon_cmd->tx.len = cpu_to_le16((u16) frame_size);
tx               3814 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	tx_beacon_cmd->tx.sta_id = il->hw_params.bcast_id;
tx               3815 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	tx_beacon_cmd->tx.stop_time.life_time = TX_CMD_LIFE_TIME_INFINITE;
tx               3816 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	tx_beacon_cmd->tx.tx_flags =
tx               3830 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	tx_beacon_cmd->tx.rate_n_flags = cpu_to_le32(rate | rate_flags);
tx               4495 drivers/net/wireless/intel/iwlegacy/4965-mac.c 		il->isr_stats.tx++;
tx               6314 drivers/net/wireless/intel/iwlegacy/4965-mac.c 	.tx = il4965_mac_tx,
tx               2663 drivers/net/wireless/intel/iwlegacy/commands.h 	struct il3945_tx_cmd tx;
tx               2671 drivers/net/wireless/intel/iwlegacy/commands.h 	struct il_tx_cmd tx;
tx               2980 drivers/net/wireless/intel/iwlegacy/commands.h 	struct iwl39_stats_tx tx;
tx               2987 drivers/net/wireless/intel/iwlegacy/commands.h 	struct stats_tx tx;
tx                556 drivers/net/wireless/intel/iwlegacy/common.h 		struct il_tx_cmd tx;
tx               1017 drivers/net/wireless/intel/iwlegacy/common.h 	u32 tx;
tx                704 drivers/net/wireless/intel/iwlegacy/debug.c 		      il->isr_stats.tx);
tx               2474 drivers/net/wireless/intel/iwlwifi/dvm/commands.h 	struct iwl_tx_cmd tx;
tx               2760 drivers/net/wireless/intel/iwlwifi/dvm/commands.h 	struct statistics_tx tx;
tx               2767 drivers/net/wireless/intel/iwlwifi/dvm/commands.h 	struct statistics_tx tx;
tx               3835 drivers/net/wireless/intel/iwlwifi/dvm/commands.h 	u8 tx[IWLAGN_MIC_KEY_SIZE];
tx               3848 drivers/net/wireless/intel/iwlwifi/dvm/commands.h 	struct iwlagn_p1k_cache tx;
tx               1112 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 	struct statistics_tx *tx, *accum_tx, *delta_tx, *max_tx;
tx               1127 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 	tx = &priv->statistics.tx;
tx               1128 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 	accum_tx = &priv->accum_stats.tx;
tx               1129 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 	delta_tx = &priv->delta_stats.tx;
tx               1130 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 	max_tx = &priv->max_delta_stats.tx;
tx               1137 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->preamble_cnt),
tx               1142 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->rx_detected_cnt),
tx               1147 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->bt_prio_defer_cnt),
tx               1153 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->bt_prio_kill_cnt),
tx               1159 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->few_bytes_cnt),
tx               1164 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->cts_timeout), accum_tx->cts_timeout,
tx               1168 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->ack_timeout),
tx               1173 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->expected_ack_cnt),
tx               1179 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->actual_ack_cnt),
tx               1185 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->dump_msdu_cnt),
tx               1191 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->burst_abort_next_frame_mismatch_cnt),
tx               1197 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->burst_abort_missing_next_frame_cnt),
tx               1203 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->cts_timeout_collision),
tx               1209 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->ack_or_ba_timeout_collision),
tx               1215 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->agg.ba_timeout),
tx               1221 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->agg.ba_reschedule_frames),
tx               1227 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->agg.scd_query_agg_frame_cnt),
tx               1233 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->agg.scd_query_no_agg),
tx               1239 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->agg.scd_query_agg),
tx               1245 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->agg.scd_query_mismatch),
tx               1251 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->agg.frame_not_ready),
tx               1257 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->agg.underrun),
tx               1262 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->agg.bt_prio_kill),
tx               1268 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 			 le32_to_cpu(tx->agg.rx_ba_rsp_cnt),
tx               1273 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 	if (tx->tx_power.ant_a || tx->tx_power.ant_b || tx->tx_power.ant_c) {
tx               1277 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 		    tx->tx_power.ant_a)
tx               1280 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 					tx->tx_power.ant_a);
tx               1282 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 		    tx->tx_power.ant_b)
tx               1285 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 					tx->tx_power.ant_b);
tx               1287 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 		    tx->tx_power.ant_c)
tx               1290 drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c 					tx->tx_power.ant_c);
tx                771 drivers/net/wireless/intel/iwlwifi/dvm/dev.h 		struct statistics_tx tx;
tx                785 drivers/net/wireless/intel/iwlwifi/dvm/dev.h 		struct statistics_tx tx;
tx                923 drivers/net/wireless/intel/iwlwifi/dvm/lib.c 			iwlagn_convert_p1k(p1k, data->tkip->tx.p1k);
tx                925 drivers/net/wireless/intel/iwlwifi/dvm/lib.c 			memcpy(data->tkip->mic_keys.tx,
tx               1576 drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c 	.tx = iwlagn_mac_tx,
tx                222 drivers/net/wireless/intel/iwlwifi/dvm/main.c 	tx_beacon_cmd->tx.len = cpu_to_le16((u16)frame_size);
tx                223 drivers/net/wireless/intel/iwlwifi/dvm/main.c 	tx_beacon_cmd->tx.sta_id = priv->beacon_ctx->bcast_sta_id;
tx                224 drivers/net/wireless/intel/iwlwifi/dvm/main.c 	tx_beacon_cmd->tx.stop_time.life_time = TX_CMD_LIFE_TIME_INFINITE;
tx                225 drivers/net/wireless/intel/iwlwifi/dvm/main.c 	tx_beacon_cmd->tx.tx_flags = TX_CMD_FLG_SEQ_CTL_MSK |
tx                256 drivers/net/wireless/intel/iwlwifi/dvm/main.c 	tx_beacon_cmd->tx.rate_n_flags =
tx                215 drivers/net/wireless/intel/iwlwifi/dvm/rx.c 				struct statistics_tx *tx,
tx                312 drivers/net/wireless/intel/iwlwifi/dvm/rx.c 			    struct statistics_tx *tx,
tx                328 drivers/net/wireless/intel/iwlwifi/dvm/rx.c 	ACCUM(tx);
tx                341 drivers/net/wireless/intel/iwlwifi/dvm/rx.c 			    struct statistics_tx *tx,
tx                361 drivers/net/wireless/intel/iwlwifi/dvm/rx.c 	struct statistics_tx *tx;
tx                378 drivers/net/wireless/intel/iwlwifi/dvm/rx.c 		tx = &stats->tx;
tx                396 drivers/net/wireless/intel/iwlwifi/dvm/rx.c 		tx = &stats->tx;
tx                411 drivers/net/wireless/intel/iwlwifi/dvm/rx.c 				    rx_ofdm_ht, rx_cck, tx, bt_activity);
tx                413 drivers/net/wireless/intel/iwlwifi/dvm/rx.c 	iwlagn_recover_from_statistics(priv, rx_ofdm, rx_ofdm_ht, tx, stamp);
tx                421 drivers/net/wireless/intel/iwlwifi/dvm/rx.c 	memcpy(&priv->statistics.tx, tx, sizeof(*tx));
tx                452 drivers/net/wireless/intel/iwlwifi/fw/api/d3.h 	u8 tx[IWL_MIC_KEY_SIZE];
tx                466 drivers/net/wireless/intel/iwlwifi/fw/api/d3.h 	struct iwl_p1k_cache tx;
tx                421 drivers/net/wireless/intel/iwlwifi/fw/api/stats.h 	struct mvm_statistics_tx_v4 tx;
tx                428 drivers/net/wireless/intel/iwlwifi/fw/api/stats.h 	struct mvm_statistics_tx_v4 tx;
tx                436 drivers/net/wireless/intel/iwlwifi/fw/api/stats.h 	struct mvm_statistics_tx tx;
tx                783 drivers/net/wireless/intel/iwlwifi/fw/api/tx.h 	struct iwl_tx_cmd tx;
tx                802 drivers/net/wireless/intel/iwlwifi/fw/api/tx.h 	struct iwl_tx_cmd tx;
tx                558 drivers/net/wireless/intel/iwlwifi/iwl-trans.h 	int (*tx)(struct iwl_trans *trans, struct sk_buff *skb,
tx                961 drivers/net/wireless/intel/iwlwifi/iwl-trans.h 	return trans->ops->tx(trans, skb, dev_cmd, queue);
tx                264 drivers/net/wireless/intel/iwlwifi/mvm/d3.c 			iwl_mvm_convert_p1k(p1k, data->tkip->tx.p1k);
tx                266 drivers/net/wireless/intel/iwlwifi/mvm/d3.c 			memcpy(data->tkip->mic_keys.tx,
tx                870 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c 				    struct iwl_tx_cmd *tx)
tx                880 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c 	tx->len = cpu_to_le16((u16)beacon->len);
tx                881 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c 	tx->sta_id = mvmvif->bcast_sta.sta_id;
tx                882 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c 	tx->life_time = cpu_to_le32(TX_CMD_LIFE_TIME_INFINITE);
tx                887 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c 	tx->tx_flags = cpu_to_le32(tx_flags);
tx                893 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c 	tx->rate_n_flags =
tx                899 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c 	tx->rate_n_flags |= cpu_to_le32(iwl_mvm_mac80211_idx_to_hwrate(rate));
tx                901 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c 		tx->rate_n_flags |= cpu_to_le32(RATE_MCS_CCK_MSK);
tx                931 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c 	iwl_mvm_mac_ctxt_set_tx(mvm, vif, beacon, &beacon_cmd.tx);
tx                951 drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c 	iwl_mvm_mac_ctxt_set_tx(mvm, vif, beacon, &beacon_cmd.tx);
tx               5009 drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c 	.tx = iwl_mvm_mac_tx,
tx                615 drivers/net/wireless/intel/iwlwifi/mvm/mvm.h 	} tx;
tx               1042 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	mdata->tx.airtime += airtime;
tx               1057 drivers/net/wireless/intel/iwlwifi/mvm/tx.c 	mdata->tx.pkts[ac]++;
tx               1248 drivers/net/wireless/intel/iwlwifi/mvm/utils.c 		u32 airtime = mdata->rx.airtime + mdata->tx.airtime;
tx               1260 drivers/net/wireless/intel/iwlwifi/mvm/utils.c 				      mdata->tx.pkts[ac];
tx               1271 drivers/net/wireless/intel/iwlwifi/mvm/utils.c 			memset(&mdata->tx.pkts, 0, sizeof(mdata->tx.pkts));
tx               1282 drivers/net/wireless/intel/iwlwifi/mvm/utils.c 		memset(&mdata->tx.airtime, 0, sizeof(mdata->tx.airtime));
tx               1392 drivers/net/wireless/intel/iwlwifi/mvm/utils.c 		memset(&mdata->tx.pkts, 0, sizeof(mdata->tx.pkts));
tx               1394 drivers/net/wireless/intel/iwlwifi/mvm/utils.c 		memset(&mdata->tx.airtime, 0, sizeof(mdata->tx.airtime));
tx                132 drivers/net/wireless/intel/iwlwifi/pcie/internal.h 	u32 tx;
tx               1950 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 		isr_stats->tx++;
tx               2153 drivers/net/wireless/intel/iwlwifi/pcie/rx.c 		isr_stats->tx++;
tx               2684 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 		isr_stats->tx);
tx               3419 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	.tx = iwl_trans_pcie_tx,
tx               3446 drivers/net/wireless/intel/iwlwifi/pcie/trans.c 	.tx = iwl_trans_pcie_gen2_tx,
tx                 33 drivers/net/wireless/intersil/hostap/hostap.h 			   const struct hfa384x_tx_frame *tx);
tx                389 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 	struct hostap_tx_data tx;
tx                398 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 	tx.skb = skb;
tx                399 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 	tx.sta_ptr = NULL;
tx                414 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 		tx.crypt = local->crypt_info.crypt[local->crypt_info.tx_keyidx];
tx                415 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 		tx.host_encrypt = 1;
tx                417 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 		tx.crypt = NULL;
tx                418 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 		tx.host_encrypt = 0;
tx                433 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 	tx_ret = hostap_handle_sta_tx(local, &tx);
tx                434 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 	skb = tx.skb;
tx                483 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 		tx.crypt = NULL;
tx                486 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 	if (local->ieee_802_1x && meta->ethertype == ETH_P_PAE && tx.crypt &&
tx                491 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 		tx.crypt = NULL; /* no encryption for IEEE 802.1X frames */
tx                494 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 	if (tx.crypt && (!tx.crypt->ops || !tx.crypt->ops->encrypt_mpdu))
tx                495 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 		tx.crypt = NULL;
tx                496 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 	else if ((tx.crypt ||
tx                515 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 	if (tx.crypt) {
tx                516 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 		skb = hostap_tx_encrypt(skb, tx.crypt);
tx                535 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 	if (local->func->tx == NULL || local->func->tx(skb, dev)) {
tx                548 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 	if (tx.sta_ptr)
tx                549 drivers/net/wireless/intersil/hostap/hostap_80211_tx.c 		hostap_handle_sta_release(tx.sta_ptr);
tx               2682 drivers/net/wireless/intersil/hostap/hostap_ap.c ap_tx_ret hostap_handle_sta_tx(local_info_t *local, struct hostap_tx_data *tx)
tx               2685 drivers/net/wireless/intersil/hostap/hostap_ap.c 	struct sk_buff *skb = tx->skb;
tx               2802 drivers/net/wireless/intersil/hostap/hostap_ap.c 		    sta->crypt && tx->host_encrypt) {
tx               2803 drivers/net/wireless/intersil/hostap/hostap_ap.c 			tx->crypt = sta->crypt;
tx               2804 drivers/net/wireless/intersil/hostap/hostap_ap.c 			tx->sta_ptr = sta; /* hostap_handle_sta_release() will
tx                236 drivers/net/wireless/intersil/hostap/hostap_ap.h ap_tx_ret hostap_handle_sta_tx(local_info_t *local, struct hostap_tx_data *tx);
tx               3100 drivers/net/wireless/intersil/hostap/hostap_hw.c 	local->func->tx = prism2_tx_80211;
tx                549 drivers/net/wireless/intersil/hostap/hostap_main.c void hostap_dump_tx_header(const char *name, const struct hfa384x_tx_frame *tx)
tx                555 drivers/net/wireless/intersil/hostap/hostap_main.c 	       name, __le16_to_cpu(tx->status), tx->retry_count, tx->tx_rate,
tx                556 drivers/net/wireless/intersil/hostap/hostap_main.c 	       __le16_to_cpu(tx->tx_control), jiffies);
tx                558 drivers/net/wireless/intersil/hostap/hostap_main.c 	fc = __le16_to_cpu(tx->frame_control);
tx                563 drivers/net/wireless/intersil/hostap/hostap_main.c 	       __le16_to_cpu(tx->duration_id), __le16_to_cpu(tx->seq_ctrl),
tx                564 drivers/net/wireless/intersil/hostap/hostap_main.c 	       __le16_to_cpu(tx->data_len),
tx                569 drivers/net/wireless/intersil/hostap/hostap_main.c 	       tx->addr1, tx->addr2, tx->addr3, tx->addr4);
tx                572 drivers/net/wireless/intersil/hostap/hostap_main.c 	       tx->dst_addr, tx->src_addr,
tx                573 drivers/net/wireless/intersil/hostap/hostap_main.c 	       __be16_to_cpu(tx->len));
tx                600 drivers/net/wireless/intersil/hostap/hostap_wlan.h 	int (*tx)(struct sk_buff *skb, struct net_device *dev);
tx                313 drivers/net/wireless/intersil/p54/main.c 	priv->survey_raw.tx = 0;
tx                706 drivers/net/wireless/intersil/p54/main.c 	.tx			= p54_tx_80211,
tx                167 drivers/net/wireless/intersil/p54/p54.h 	void (*tx)(struct ieee80211_hw *dev, struct sk_buff *skb);
tx                212 drivers/net/wireless/intersil/p54/p54.h 		u64 tx;
tx                615 drivers/net/wireless/intersil/p54/p54pci.c 	priv->common.tx = p54p_tx;
tx                645 drivers/net/wireless/intersil/p54/p54spi.c 	priv->common.tx = p54spi_op_tx;
tx               1040 drivers/net/wireless/intersil/p54/p54usb.c 		priv->common.tx = p54u_tx_lm87;
tx               1046 drivers/net/wireless/intersil/p54/p54usb.c 		priv->common.tx = p54u_tx_net2280;
tx                163 drivers/net/wireless/intersil/p54/txrx.c 		priv->tx(priv->hw, skb);
tx                522 drivers/net/wireless/intersil/p54/txrx.c 	unsigned int i, rssi, tx, cca, dtime, dtotal, dcca, dtx, drssi, unit;
tx                552 drivers/net/wireless/intersil/p54/txrx.c 	tx = le32_to_cpu(stats->sample_tx);
tx                559 drivers/net/wireless/intersil/p54/txrx.c 	dtx = tx - priv->survey_raw.cached_tx;
tx                577 drivers/net/wireless/intersil/p54/txrx.c 			priv->survey_raw.tx += dtx * unit;
tx                578 drivers/net/wireless/intersil/p54/txrx.c 			priv->survey_raw.cached_tx = tx;
tx                597 drivers/net/wireless/intersil/p54/txrx.c 		survey->time_tx = priv->survey_raw.tx;
tx                598 drivers/net/wireless/intersil/p54/txrx.c 		survey->time_busy = priv->survey_raw.tx +
tx               2337 drivers/net/wireless/mac80211_hwsim.c 	.tx = mac80211_hwsim_tx,				\
tx                476 drivers/net/wireless/marvell/libertas_tf/main.c 	.tx			= lbtf_op_tx,
tx               1668 drivers/net/wireless/marvell/mwifiex/cfg80211.c 		.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
tx               1674 drivers/net/wireless/marvell/mwifiex/cfg80211.c 		.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
tx               1680 drivers/net/wireless/marvell/mwifiex/cfg80211.c 		.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
tx               1686 drivers/net/wireless/marvell/mwifiex/cfg80211.c 		.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
tx               1675 drivers/net/wireless/marvell/mwl8k.c 		int tx;
tx               1687 drivers/net/wireless/marvell/mwl8k.c 		tx = txq->head;
tx               1688 drivers/net/wireless/marvell/mwl8k.c 		tx_desc = txq->txd + tx;
tx               1699 drivers/net/wireless/marvell/mwl8k.c 		txq->head = (tx + 1) % MWL8K_TX_DESCS;
tx               1706 drivers/net/wireless/marvell/mwl8k.c 		skb = txq->skb[tx];
tx               1707 drivers/net/wireless/marvell/mwl8k.c 		txq->skb[tx] = NULL;
tx               1903 drivers/net/wireless/marvell/mwl8k.c 	struct mwl8k_tx_desc *tx;
tx               2090 drivers/net/wireless/marvell/mwl8k.c 	tx = txq->txd + txq->tail;
tx               2091 drivers/net/wireless/marvell/mwl8k.c 	tx->data_rate = txdatarate;
tx               2092 drivers/net/wireless/marvell/mwl8k.c 	tx->tx_priority = txpriority;
tx               2093 drivers/net/wireless/marvell/mwl8k.c 	tx->qos_control = cpu_to_le16(qos);
tx               2094 drivers/net/wireless/marvell/mwl8k.c 	tx->pkt_phys_addr = cpu_to_le32(dma);
tx               2095 drivers/net/wireless/marvell/mwl8k.c 	tx->pkt_len = cpu_to_le16(skb->len);
tx               2096 drivers/net/wireless/marvell/mwl8k.c 	tx->rate_info = 0;
tx               2098 drivers/net/wireless/marvell/mwl8k.c 		tx->peer_id = MWL8K_STA(sta)->peer_id;
tx               2100 drivers/net/wireless/marvell/mwl8k.c 		tx->peer_id = 0;
tx               2103 drivers/net/wireless/marvell/mwl8k.c 		tx->timestamp = cpu_to_le32(ioread32(priv->regs +
tx               2106 drivers/net/wireless/marvell/mwl8k.c 		tx->timestamp = 0;
tx               2109 drivers/net/wireless/marvell/mwl8k.c 	tx->status = cpu_to_le32(MWL8K_TXD_STATUS_FW_OWNED | txstatus);
tx               3537 drivers/net/wireless/marvell/mwl8k.c static int mwl8k_cmd_mimo_config(struct ieee80211_hw *hw, __u8 rx, __u8 tx)
tx               3550 drivers/net/wireless/marvell/mwl8k.c 	cmd->tx_antenna_map = tx;
tx               5605 drivers/net/wireless/marvell/mwl8k.c 	.tx			= mwl8k_tx,
tx                657 drivers/net/wireless/mediatek/mt76/mt7603/main.c 	.tx = mt7603_tx,
tx                493 drivers/net/wireless/mediatek/mt76/mt7615/main.c 	.tx = mt7615_tx,
tx                 74 drivers/net/wireless/mediatek/mt76/mt76x0/pci.c 	.tx = mt76x02_tx,
tx                120 drivers/net/wireless/mediatek/mt76/mt76x0/usb.c 	.tx = mt76x02_tx,
tx                157 drivers/net/wireless/mediatek/mt76/mt76x2/pci_main.c 	.tx = mt76x02_tx,
tx                104 drivers/net/wireless/mediatek/mt76/mt76x2/usb_main.c 	.tx = mt76x02_tx,
tx                409 drivers/net/wireless/mediatek/mt7601u/main.c 	.tx = mt7601u_tx,
tx                 56 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c 		.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
tx                 63 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c 		.tx = BIT(IEEE80211_STYPE_ACTION >> 4),
tx                401 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	       ant->tx == ANTENNA_SW_DIVERSITY);
tx                409 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	switch (ant->tx) {
tx                766 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	rt2x00_set_field32(&reg, TXCSR2_TXD_SIZE, rt2x00dev->tx[0].desc_size);
tx                767 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	rt2x00_set_field32(&reg, TXCSR2_NUM_TXD, rt2x00dev->tx[1].limit);
tx                769 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	rt2x00_set_field32(&reg, TXCSR2_NUM_PRIO, rt2x00dev->tx[0].limit);
tx                772 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	entry_priv = rt2x00dev->tx[1].entries[0].priv_data;
tx                778 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	entry_priv = rt2x00dev->tx[0].entries[0].priv_data;
tx               1492 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	rt2x00dev->default_ant.tx =
tx               1503 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	if (rt2x00dev->default_ant.tx == ANTENNA_SW_DIVERSITY)
tx               1504 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 		rt2x00dev->default_ant.tx = ANTENNA_HW_DIVERSITY;
tx               1675 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 			    rt2x00dev->tx->cw_min, rt2x00dev->tx->cw_max);
tx               1705 drivers/net/wireless/ralink/rt2x00/rt2400pci.c 	.tx			= rt2x00mac_tx,
tx                408 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	       ant->tx == ANTENNA_SW_DIVERSITY);
tx                417 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	switch (ant->tx) {
tx                851 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	rt2x00_set_field32(&reg, TXCSR2_TXD_SIZE, rt2x00dev->tx[0].desc_size);
tx                852 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	rt2x00_set_field32(&reg, TXCSR2_NUM_TXD, rt2x00dev->tx[1].limit);
tx                854 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	rt2x00_set_field32(&reg, TXCSR2_NUM_PRIO, rt2x00dev->tx[0].limit);
tx                857 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	entry_priv = rt2x00dev->tx[1].entries[0].priv_data;
tx                863 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	entry_priv = rt2x00dev->tx[0].entries[0].priv_data;
tx               1651 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	rt2x00dev->default_ant.tx =
tx               2004 drivers/net/wireless/ralink/rt2x00/rt2500pci.c 	.tx			= rt2x00mac_tx,
tx                513 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	       ant->tx == ANTENNA_SW_DIVERSITY);
tx                523 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	switch (ant->tx) {
tx               1461 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	rt2x00dev->default_ant.tx =
tx               1472 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	if (rt2x00dev->default_ant.tx == ANTENNA_SW_DIVERSITY)
tx               1473 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 		rt2x00dev->default_ant.tx = ANTENNA_HW_DIVERSITY;
tx               1797 drivers/net/wireless/ralink/rt2x00/rt2500usb.c 	.tx			= rt2x00mac_tx,
tx               9479 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 			rt2x00dev->default_ant.tx = ANTENNA_A;
tx               9483 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 			rt2x00dev->default_ant.tx = ANTENNA_A;
tx               9488 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		rt2x00dev->default_ant.tx = ANTENNA_A;
tx               9495 drivers/net/wireless/ralink/rt2x00/rt2800lib.c 		rt2x00dev->default_ant.tx = ANTENNA_HW_DIVERSITY; /* Unused */
tx                665 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	entry_priv = rt2x00dev->tx[0].entries[0].priv_data;
tx                669 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 				  rt2x00dev->tx[0].limit);
tx                673 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	entry_priv = rt2x00dev->tx[1].entries[0].priv_data;
tx                677 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 				  rt2x00dev->tx[1].limit);
tx                681 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	entry_priv = rt2x00dev->tx[2].entries[0].priv_data;
tx                685 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 				  rt2x00dev->tx[2].limit);
tx                689 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 	entry_priv = rt2x00dev->tx[3].entries[0].priv_data;
tx                693 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c 				  rt2x00dev->tx[3].limit);
tx                290 drivers/net/wireless/ralink/rt2x00/rt2800pci.c 	.tx			= rt2x00mac_tx,
tx                135 drivers/net/wireless/ralink/rt2x00/rt2800soc.c 	.tx			= rt2x00mac_tx,
tx                632 drivers/net/wireless/ralink/rt2x00/rt2800usb.c 	.tx			= rt2x00mac_tx,
tx                202 drivers/net/wireless/ralink/rt2x00/rt2x00.h 	enum antenna tx;
tx                962 drivers/net/wireless/ralink/rt2x00/rt2x00.h 	struct data_queue *tx;
tx               1304 drivers/net/wireless/ralink/rt2x00/rt2x00.h 	if (queue < rt2x00dev->ops->tx_queues && rt2x00dev->tx)
tx               1305 drivers/net/wireless/ralink/rt2x00/rt2x00.h 		return &rt2x00dev->tx[queue];
tx               1480 drivers/net/wireless/ralink/rt2x00/rt2x00.h 			     u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max);
tx                131 drivers/net/wireless/ralink/rt2x00/rt2x00config.c 		if (config.tx == ANTENNA_SW_DIVERSITY) {
tx                134 drivers/net/wireless/ralink/rt2x00/rt2x00config.c 			if (def->tx == ANTENNA_SW_DIVERSITY)
tx                135 drivers/net/wireless/ralink/rt2x00/rt2x00config.c 				config.tx = ANTENNA_B;
tx                137 drivers/net/wireless/ralink/rt2x00/rt2x00config.c 				config.tx = def->tx;
tx                139 drivers/net/wireless/ralink/rt2x00/rt2x00config.c 	} else if (config.tx == ANTENNA_SW_DIVERSITY)
tx                140 drivers/net/wireless/ralink/rt2x00/rt2x00config.c 		config.tx = active->tx;
tx               1154 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 					   rt2x00dev->tx->limit *
tx               1351 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	if (WARN_ON(!rt2x00dev->tx))
tx               1355 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 		return rt2x00dev->tx[0].winfo_size + rt2x00dev->tx[0].desc_size;
tx               1357 drivers/net/wireless/ralink/rt2x00/rt2x00dev.c 	return rt2x00dev->tx[0].winfo_size;
tx                103 drivers/net/wireless/ralink/rt2x00/rt2x00link.c 		new_ant.tx = other_antenna;
tx                143 drivers/net/wireless/ralink/rt2x00/rt2x00link.c 		new_ant.tx = (new_ant.tx == ANTENNA_A) ? ANTENNA_B : ANTENNA_A;
tx                735 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c 	if (tx_ant == 3 && def->tx == ANTENNA_SW_DIVERSITY) {
tx                745 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c 	setup.tx = tx_ant;
tx                767 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c 		*tx_ant = active->tx;
tx                779 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c 			     u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max)
tx                785 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c 		*tx += queue->length;
tx               1256 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	rt2x00dev->tx = &queue[1];
tx               1286 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	rt2x00dev->tx = NULL;
tx                505 drivers/net/wireless/ralink/rt2x00/rt2x00queue.h 	&(__dev)->tx[(__dev)->ops->tx_queues]
tx                552 drivers/net/wireless/ralink/rt2x00/rt2x00queue.h 	queue_loop(__entry, (__dev)->tx, tx_queue_end(__dev))
tx                563 drivers/net/wireless/ralink/rt2x00/rt2x00queue.h 	queue_loop(__entry, (__dev)->tx, queue_end(__dev))
tx                601 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	struct data_queue *queue = rt2x00dev->tx;
tx                628 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c 	if (!rt2x00dev->rx->usb_endpoint || !rt2x00dev->tx->usb_endpoint) {
tx                716 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	       ant->tx == ANTENNA_SW_DIVERSITY);
tx               1337 drivers/net/wireless/ralink/rt2x00/rt61pci.c 			   rt2x00dev->tx[0].limit);
tx               1339 drivers/net/wireless/ralink/rt2x00/rt61pci.c 			   rt2x00dev->tx[1].limit);
tx               1341 drivers/net/wireless/ralink/rt2x00/rt61pci.c 			   rt2x00dev->tx[2].limit);
tx               1343 drivers/net/wireless/ralink/rt2x00/rt61pci.c 			   rt2x00dev->tx[3].limit);
tx               1348 drivers/net/wireless/ralink/rt2x00/rt61pci.c 			   rt2x00dev->tx[0].desc_size / 4);
tx               1351 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	entry_priv = rt2x00dev->tx[0].entries[0].priv_data;
tx               1357 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	entry_priv = rt2x00dev->tx[1].entries[0].priv_data;
tx               1363 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	entry_priv = rt2x00dev->tx[2].entries[0].priv_data;
tx               1369 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	entry_priv = rt2x00dev->tx[3].entries[0].priv_data;
tx               2081 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	for (i = 0; i < rt2x00dev->tx->limit; i++) {
tx               2439 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	rt2x00dev->default_ant.tx =
tx               2484 drivers/net/wireless/ralink/rt2x00/rt61pci.c 		rt2x00dev->default_ant.tx =
tx               2488 drivers/net/wireless/ralink/rt2x00/rt61pci.c 			rt2x00dev->default_ant.tx = ANTENNA_SW_DIVERSITY;
tx               2870 drivers/net/wireless/ralink/rt2x00/rt61pci.c 	.tx			= rt2x00mac_tx,
tx                697 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	       ant->tx == ANTENNA_SW_DIVERSITY);
tx               1879 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	rt2x00dev->default_ant.tx =
tx               2293 drivers/net/wireless/ralink/rt2x00/rt73usb.c 	.tx			= rt2x00mac_tx,
tx               1601 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 	.tx			= rtl8180_tx,
tx               1379 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 	.tx			= rtl8187_tx,
tx               5891 drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu_core.c 	.tx = rtl8xxxu_tx,
tx               1879 drivers/net/wireless/realtek/rtlwifi/core.c 	.tx = rtl_op_tx,
tx                 10 drivers/net/wireless/realtek/rtw88/hci.h 	int (*tx)(struct rtw_dev *rtwdev,
tx                 32 drivers/net/wireless/realtek/rtw88/hci.h 	return rtwdev->hci.ops->tx(rtwdev, pkt_info, skb);
tx                516 drivers/net/wireless/realtek/rtw88/mac80211.c 	.tx			= rtw_ops_tx,
tx               1146 drivers/net/wireless/realtek/rtw88/pci.c 	.tx = rtw_pci_tx,
tx               1975 drivers/net/wireless/rsi/rsi_91x_mac80211.c 	.tx = rsi_mac80211_tx,
tx                239 drivers/net/wireless/st/cw1200/bh.c 			       int *tx)
tx                328 drivers/net/wireless/st/cw1200/bh.c 			*tx = 1;
tx                423 drivers/net/wireless/st/cw1200/bh.c 	int rx, tx, term, suspend;
tx                458 drivers/net/wireless/st/cw1200/bh.c 				tx = atomic_xchg(&priv->bh_tx, 0);
tx                462 drivers/net/wireless/st/cw1200/bh.c 				(rx || tx || term || suspend || priv->bh_error);
tx                466 drivers/net/wireless/st/cw1200/bh.c 			 rx, tx, term, suspend, priv->bh_error, status);
tx                480 drivers/net/wireless/st/cw1200/bh.c 			if (priv->hw_bufs_used && (!rx || !tx)) {
tx                545 drivers/net/wireless/st/cw1200/bh.c 		tx += pending_tx;
tx                553 drivers/net/wireless/st/cw1200/bh.c 			ret = cw1200_bh_rx_helper(priv, &ctrl_reg, &tx);
tx                558 drivers/net/wireless/st/cw1200/bh.c 				ret = cw1200_bh_rx_helper(priv, &ctrl_reg, &tx);
tx                564 drivers/net/wireless/st/cw1200/bh.c 	tx:
tx                565 drivers/net/wireless/st/cw1200/bh.c 		if (tx) {
tx                566 drivers/net/wireless/st/cw1200/bh.c 			tx = 0;
tx                576 drivers/net/wireless/st/cw1200/bh.c 				pending_tx = tx;
tx                583 drivers/net/wireless/st/cw1200/bh.c 				tx = ret;
tx                595 drivers/net/wireless/st/cw1200/bh.c 		if (tx)
tx                596 drivers/net/wireless/st/cw1200/bh.c 			goto tx;
tx                266 drivers/net/wireless/st/cw1200/debug.c 		   d->tx);
tx                 14 drivers/net/wireless/st/cw1200/debug.h 	int tx;
tx                 35 drivers/net/wireless/st/cw1200/debug.h 	++priv->debug->tx;
tx                211 drivers/net/wireless/st/cw1200/main.c 	.tx			= cw1200_tx,
tx                331 drivers/net/wireless/st/cw1200/queue.c 		     struct wsm_tx **tx,
tx                349 drivers/net/wireless/st/cw1200/queue.c 		*tx = (struct wsm_tx *)item->skb->data;
tx                352 drivers/net/wireless/st/cw1200/queue.c 		(*tx)->packet_id = item->packet_id;
tx                 84 drivers/net/wireless/st/cw1200/queue.h 		     struct wsm_tx **tx,
tx                151 drivers/net/wireless/st/cw1200/sta.c 	int tx = 0;
tx                174 drivers/net/wireless/st/cw1200/sta.c 			tx = 1;
tx                182 drivers/net/wireless/st/cw1200/sta.c 			tx = 1;
tx                190 drivers/net/wireless/st/cw1200/sta.c 		tx = 0;
tx                193 drivers/net/wireless/st/cw1200/sta.c 	if (tx) {
tx                898 drivers/net/wireless/st/cw1200/txrx.c 		struct ieee80211_tx_info *tx = IEEE80211_SKB_CB(skb);
tx                920 drivers/net/wireless/st/cw1200/txrx.c 			tx->flags |= IEEE80211_TX_STAT_ACK;
tx                936 drivers/net/wireless/st/cw1200/txrx.c 			if (tx->status.rates[i].count >= tx_count) {
tx                937 drivers/net/wireless/st/cw1200/txrx.c 				tx->status.rates[i].count = tx_count;
tx                940 drivers/net/wireless/st/cw1200/txrx.c 			tx_count -= tx->status.rates[i].count;
tx                941 drivers/net/wireless/st/cw1200/txrx.c 			if (tx->status.rates[i].flags & IEEE80211_TX_RC_MCS)
tx                942 drivers/net/wireless/st/cw1200/txrx.c 				tx->status.rates[i].flags |= ht_flags;
tx                946 drivers/net/wireless/st/cw1200/txrx.c 			tx->status.rates[i].count = 0;
tx                947 drivers/net/wireless/st/cw1200/txrx.c 			tx->status.rates[i].idx = -1;
tx                951 drivers/net/wireless/st/cw1200/txrx.c 		if (tx->control.hw_key) {
tx                952 drivers/net/wireless/st/cw1200/txrx.c 			skb_trim(skb, skb->len - tx->control.hw_key->icv_len);
tx                953 drivers/net/wireless/st/cw1200/txrx.c 			if (tx->control.hw_key->cipher == WLAN_CIPHER_SUITE_TKIP)
tx               1063 drivers/net/wireless/ti/wl1251/acx.h 	struct acx_tx_statistics tx;
tx                101 drivers/net/wireless/ti/wl1251/debugfs.c DEBUGFS_FWSTATS_FILE(tx, internal_desc_overflow, 20, "%u");
tx                241 drivers/net/wireless/ti/wl1251/debugfs.c 	DEBUGFS_FWSTATS_DEL(tx, internal_desc_overflow);
tx                340 drivers/net/wireless/ti/wl1251/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, internal_desc_overflow);
tx               1360 drivers/net/wireless/ti/wl1251/main.c 	.tx = wl1251_op_tx,
tx                244 drivers/net/wireless/ti/wl12xx/acx.h 	struct wl12xx_acx_tx_statistics tx;
tx                 19 drivers/net/wireless/ti/wl12xx/debugfs.c WL12XX_DEBUGFS_FWSTATS_FILE(tx, internal_desc_overflow, "%u");
tx                119 drivers/net/wireless/ti/wl12xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, internal_desc_overflow);
tx                115 drivers/net/wireless/ti/wl12xx/main.c 	.tx = {
tx               1659 drivers/net/wireless/ti/wl12xx/main.c 		thold = wl->conf.tx.fast_link_thold;
tx               1661 drivers/net/wireless/ti/wl12xx/main.c 		thold = wl->conf.tx.slow_link_thold;
tx                270 drivers/net/wireless/ti/wl18xx/acx.h 	struct wl18xx_acx_tx_stats		tx;
tx                 45 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_prepared_descs, "%u");
tx                 46 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_cmplt, "%u");
tx                 47 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_template_prepared, "%u");
tx                 48 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_data_prepared, "%u");
tx                 49 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_template_programmed, "%u");
tx                 50 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_data_programmed, "%u");
tx                 51 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_burst_programmed, "%u");
tx                 52 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_starts, "%u");
tx                 53 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_stop, "%u");
tx                 54 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_start_templates, "%u");
tx                 55 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_start_int_templates, "%u");
tx                 56 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_start_fw_gen, "%u");
tx                 57 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_start_data, "%u");
tx                 58 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_start_null_frame, "%u");
tx                 59 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_exch, "%u");
tx                 60 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_retry_template, "%u");
tx                 61 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_retry_data, "%u");
tx                 62 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE_ARRAY(tx, tx_retry_per_rate,
tx                 64 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_exch_pending, "%u");
tx                 65 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_exch_expiry, "%u");
tx                 66 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_done_template, "%u");
tx                 67 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_done_data, "%u");
tx                 68 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_done_int_template, "%u");
tx                 69 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_cfe1, "%u");
tx                 70 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_cfe2, "%u");
tx                 71 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_called, "%u");
tx                 72 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_mpdu_alloc_failed, "%u");
tx                 73 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_init_called, "%u");
tx                 74 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_in_process_called, "%u");
tx                 75 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_tkip_called, "%u");
tx                 76 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_key_not_found, "%u");
tx                 77 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_need_fragmentation, "%u");
tx                 78 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_bad_mblk_num, "%u");
tx                 79 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_failed, "%u");
tx                 80 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_cache_hit, "%u");
tx                 81 drivers/net/wireless/ti/wl18xx/debugfs.c WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_cache_miss, "%u");
tx                436 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_prepared_descs);
tx                437 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_cmplt);
tx                438 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_template_prepared);
tx                439 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_data_prepared);
tx                440 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_template_programmed);
tx                441 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_data_programmed);
tx                442 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_burst_programmed);
tx                443 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_starts);
tx                444 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_stop);
tx                445 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_start_templates);
tx                446 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_start_int_templates);
tx                447 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_start_fw_gen);
tx                448 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_start_data);
tx                449 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_start_null_frame);
tx                450 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_exch);
tx                451 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_retry_template);
tx                452 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_retry_data);
tx                453 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_retry_per_rate);
tx                454 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_exch_pending);
tx                455 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_exch_expiry);
tx                456 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_done_template);
tx                457 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_done_data);
tx                458 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_done_int_template);
tx                459 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_cfe1);
tx                460 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, tx_cfe2);
tx                461 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, frag_called);
tx                462 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, frag_mpdu_alloc_failed);
tx                463 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, frag_init_called);
tx                464 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, frag_in_process_called);
tx                465 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, frag_tkip_called);
tx                466 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, frag_key_not_found);
tx                467 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, frag_need_fragmentation);
tx                468 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, frag_bad_mblk_num);
tx                469 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, frag_failed);
tx                470 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, frag_cache_hit);
tx                471 drivers/net/wireless/ti/wl18xx/debugfs.c 	DEBUGFS_FWSTATS_ADD(tx, frag_cache_miss);
tx                256 drivers/net/wireless/ti/wl18xx/main.c 	.tx = {
tx                551 drivers/net/wireless/ti/wlcore/acx.c 	detection->tx_energy_detection = wl->conf.tx.tx_energy_detection;
tx                725 drivers/net/wireless/ti/wlcore/acx.c 	struct conf_tx_rate_class *c = &wl->conf.tx.sta_rc_conf;
tx                906 drivers/net/wireless/ti/wlcore/acx.c 		frag_threshold = wl->conf.tx.frag_threshold;
tx                943 drivers/net/wireless/ti/wlcore/acx.c 	acx->tx_compl_timeout = cpu_to_le16(wl->conf.tx.tx_compl_timeout);
tx                944 drivers/net/wireless/ti/wlcore/acx.c 	acx->tx_compl_threshold = cpu_to_le16(wl->conf.tx.tx_compl_threshold);
tx               1538 drivers/net/wireless/ti/wlcore/acx.c 	acx->max_tx_retry = cpu_to_le16(wl->conf.tx.max_tx_retries);
tx                657 drivers/net/wireless/ti/wlcore/cmd.c 	cmd->ap.aging_period = cpu_to_le16(wl->conf.tx.ap_aging_period);
tx               1035 drivers/net/wireless/ti/wlcore/cmd.c 	cmd->short_retry_limit = wl->conf.tx.tmpl_short_retry_limit;
tx               1036 drivers/net/wireless/ti/wlcore/cmd.c 	cmd->long_retry_limit = wl->conf.tx.tmpl_long_retry_limit;
tx               1133 drivers/net/wireless/ti/wlcore/conf.h 	struct conf_tx_settings tx;
tx                247 drivers/net/wireless/ti/wlcore/event.c 	u32 num_packets = wl->conf.tx.max_tx_retries;
tx                468 drivers/net/wireless/ti/wlcore/init.c 	for (i = 0; i < wl->conf.tx.ac_conf_count; i++) {
tx                611 drivers/net/wireless/ti/wlcore/init.c 	BUG_ON(wl->conf.tx.tid_conf_count != wl->conf.tx.ac_conf_count);
tx                612 drivers/net/wireless/ti/wlcore/init.c 	for (i = 0; i < wl->conf.tx.tid_conf_count; i++) {
tx                613 drivers/net/wireless/ti/wlcore/init.c 		conf_ac = &wl->conf.tx.ac_conf[i];
tx                620 drivers/net/wireless/ti/wlcore/init.c 		conf_tid = &wl->conf.tx.tid_conf[i];
tx                211 drivers/net/wireless/ti/wlcore/main.c 		msecs_to_jiffies(wl->conf.tx.tx_watchdog_timeout));
tx                272 drivers/net/wireless/ti/wlcore/main.c 			     wl->conf.tx.tx_watchdog_timeout);
tx                283 drivers/net/wireless/ti/wlcore/main.c 			     wl->conf.tx.tx_watchdog_timeout);
tx                297 drivers/net/wireless/ti/wlcore/main.c 			      wl->conf.tx.tx_watchdog_timeout,
tx                304 drivers/net/wireless/ti/wlcore/main.c 		     wl->conf.tx.tx_watchdog_timeout);
tx               2279 drivers/net/wireless/ti/wlcore/main.c 	wlvif->bitrate_masks[NL80211_BAND_2GHZ] = wl->conf.tx.basic_rate;
tx               2280 drivers/net/wireless/ti/wlcore/main.c 	wlvif->bitrate_masks[NL80211_BAND_5GHZ] = wl->conf.tx.basic_rate_5;
tx               6009 drivers/net/wireless/ti/wlcore/main.c 	.tx = wl1271_op_tx,
tx               1346 drivers/net/wireless/zydas/zd1211rw/zd_mac.c 	.tx			= zd_op_tx,
tx                885 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	struct zd_usb_tx *tx = &usb->tx;
tx                888 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	atomic_set(&tx->enabled, 0);
tx                891 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	usb_kill_anchored_urbs(&tx->submitted);
tx                893 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	spin_lock_irqsave(&tx->lock, flags);
tx                894 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	WARN_ON(!skb_queue_empty(&tx->submitted_skbs));
tx                895 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	WARN_ON(tx->submitted_urbs != 0);
tx                896 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	tx->submitted_urbs = 0;
tx                897 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	spin_unlock_irqrestore(&tx->lock, flags);
tx                914 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	struct zd_usb_tx *tx = &usb->tx;
tx                916 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	spin_lock_irqsave(&tx->lock, flags);
tx                917 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	atomic_set(&tx->enabled, 1);
tx                918 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	tx->submitted_urbs = 0;
tx                920 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	tx->stopped = 0;
tx                921 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	spin_unlock_irqrestore(&tx->lock, flags);
tx                926 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	struct zd_usb_tx *tx = &usb->tx;
tx                929 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	spin_lock_irqsave(&tx->lock, flags);
tx                930 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	--tx->submitted_urbs;
tx                931 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	if (tx->stopped && tx->submitted_urbs <= ZD_USB_TX_LOW) {
tx                933 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 		tx->stopped = 0;
tx                935 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	spin_unlock_irqrestore(&tx->lock, flags);
tx                940 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	struct zd_usb_tx *tx = &usb->tx;
tx                943 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	spin_lock_irqsave(&tx->lock, flags);
tx                944 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	++tx->submitted_urbs;
tx                945 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	if (!tx->stopped && tx->submitted_urbs > ZD_USB_TX_HIGH) {
tx                947 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 		tx->stopped = 1;
tx                949 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	spin_unlock_irqrestore(&tx->lock, flags);
tx                965 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	struct zd_usb_tx *tx;
tx                974 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	tx = &usb->tx;
tx                992 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	skb_unlink(skb, &usb->tx.submitted_skbs);
tx                998 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	usb_anchor_urb(urb, &tx->submitted);
tx               1025 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	struct zd_usb_tx *tx = &usb->tx;
tx               1027 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	if (!atomic_read(&tx->enabled)) {
tx               1042 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	skb_queue_tail(&tx->submitted_skbs, skb);
tx               1043 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	usb_anchor_urb(urb, &tx->submitted);
tx               1049 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 		skb_unlink(skb, &tx->submitted_skbs);
tx               1062 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	struct zd_usb_tx *tx = &usb->tx;
tx               1063 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	struct sk_buff_head *q = &tx->submitted_skbs;
tx               1087 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 		container_of(work, struct zd_usb, tx.watchdog_work.work);
tx               1088 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	struct zd_usb_tx *tx = &usb->tx;
tx               1090 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	if (!atomic_read(&tx->enabled) || !tx->watchdog_enabled)
tx               1103 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	queue_delayed_work(zd_workqueue, &tx->watchdog_work,
tx               1109 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	struct zd_usb_tx *tx = &usb->tx;
tx               1111 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	if (!tx->watchdog_enabled) {
tx               1113 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 		queue_delayed_work(zd_workqueue, &tx->watchdog_work,
tx               1115 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 		tx->watchdog_enabled = 1;
tx               1121 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	struct zd_usb_tx *tx = &usb->tx;
tx               1123 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	if (tx->watchdog_enabled) {
tx               1125 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 		tx->watchdog_enabled = 0;
tx               1126 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 		cancel_delayed_work_sync(&tx->watchdog_work);
tx               1189 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	struct zd_usb_tx *tx = &usb->tx;
tx               1191 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	spin_lock_init(&tx->lock);
tx               1192 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	atomic_set(&tx->enabled, 0);
tx               1193 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	tx->stopped = 0;
tx               1194 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	skb_queue_head_init(&tx->submitted_skbs);
tx               1195 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	init_usb_anchor(&tx->submitted);
tx               1196 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	tx->submitted_urbs = 0;
tx               1197 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	tx->watchdog_enabled = 0;
tx               1198 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 	INIT_DELAYED_WORK(&tx->watchdog_work, zd_tx_watchdog_handler);
tx                209 drivers/net/wireless/zydas/zd1211rw/zd_usb.h 	struct zd_usb_tx tx;
tx                149 drivers/net/xen-netback/common.h 	struct xen_netif_tx_back_ring tx;
tx                 84 drivers/net/xen-netback/interface.c 	if (RING_HAS_UNCONSUMED_REQUESTS(&queue->tx))
tx                161 drivers/net/xen-netback/netback.c 	RING_FINAL_CHECK_FOR_REQUESTS(&queue->tx, more_to_do);
tx                197 drivers/net/xen-netback/netback.c 	RING_IDX cons = queue->tx.req_cons;
tx                207 drivers/net/xen-netback/netback.c 		RING_COPY_REQUEST(&queue->tx, cons++, txp);
tx                210 drivers/net/xen-netback/netback.c 	queue->tx.req_cons = cons;
tx                228 drivers/net/xen-netback/netback.c 	RING_IDX cons = queue->tx.req_cons;
tx                276 drivers/net/xen-netback/netback.c 		RING_COPY_REQUEST(&queue->tx, cons + slots, txp);
tx                604 drivers/net/xen-netback/netback.c 	RING_IDX cons = queue->tx.req_cons;
tx                613 drivers/net/xen-netback/netback.c 		RING_COPY_REQUEST(&queue->tx, cons, &extra);
tx                615 drivers/net/xen-netback/netback.c 		queue->tx.req_cons = ++cons;
tx                809 drivers/net/xen-netback/netback.c 		if (queue->tx.sring->req_prod - queue->tx.req_cons >
tx                814 drivers/net/xen-netback/netback.c 				   queue->tx.sring->req_prod, queue->tx.req_cons,
tx                820 drivers/net/xen-netback/netback.c 		work_to_do = RING_HAS_UNCONSUMED_REQUESTS(&queue->tx);
tx                824 drivers/net/xen-netback/netback.c 		idx = queue->tx.req_cons;
tx                826 drivers/net/xen-netback/netback.c 		RING_COPY_REQUEST(&queue->tx, idx, &txreq);
tx                836 drivers/net/xen-netback/netback.c 		queue->tx.req_cons = ++idx;
tx                844 drivers/net/xen-netback/netback.c 			idx = queue->tx.req_cons;
tx               1018 drivers/net/xen-netback/netback.c 		queue->tx.req_cons = idx;
tx               1379 drivers/net/xen-netback/netback.c 	RING_IDX i = queue->tx.rsp_prod_pvt;
tx               1382 drivers/net/xen-netback/netback.c 	resp = RING_GET_RESPONSE(&queue->tx, i);
tx               1387 drivers/net/xen-netback/netback.c 		RING_GET_RESPONSE(&queue->tx, ++i)->status = XEN_NETIF_RSP_NULL;
tx               1389 drivers/net/xen-netback/netback.c 	queue->tx.rsp_prod_pvt = ++i;
tx               1396 drivers/net/xen-netback/netback.c 	RING_PUSH_RESPONSES_AND_CHECK_NOTIFY(&queue->tx, notify);
tx               1428 drivers/net/xen-netback/netback.c 	if (likely(RING_HAS_UNCONSUMED_REQUESTS(&queue->tx)))
tx               1441 drivers/net/xen-netback/netback.c 	if (queue->tx.sring)
tx               1443 drivers/net/xen-netback/netback.c 					queue->tx.sring);
tx               1465 drivers/net/xen-netback/netback.c 	BACK_RING_INIT(&queue->tx, txs, XEN_PAGE_SIZE);
tx                 29 drivers/net/xen-netback/xenbus.c 	struct xen_netif_tx_back_ring *tx_ring = &queue->tx;
tx                117 drivers/net/xen-netfront.c 	struct xen_netif_tx_front_ring tx;
tx                241 drivers/net/xen-netfront.c 	return (queue->tx.req_prod_pvt - queue->tx.rsp_cons) <
tx                384 drivers/net/xen-netfront.c 		prod = queue->tx.sring->rsp_prod;
tx                387 drivers/net/xen-netfront.c 		for (cons = queue->tx.rsp_cons; cons != prod; cons++) {
tx                390 drivers/net/xen-netfront.c 			txrsp = RING_GET_RESPONSE(&queue->tx, cons);
tx                412 drivers/net/xen-netfront.c 		queue->tx.rsp_cons = prod;
tx                414 drivers/net/xen-netfront.c 		RING_FINAL_CHECK_FOR_RESPONSES(&queue->tx, more_to_do);
tx                424 drivers/net/xen-netfront.c 	struct xen_netif_tx_request *tx; /* Last request */
tx                433 drivers/net/xen-netfront.c 	struct xen_netif_tx_request *tx;
tx                441 drivers/net/xen-netfront.c 	tx = RING_GET_REQUEST(&queue->tx, queue->tx.req_prod_pvt++);
tx                452 drivers/net/xen-netfront.c 	tx->id = id;
tx                453 drivers/net/xen-netfront.c 	tx->gref = ref;
tx                454 drivers/net/xen-netfront.c 	tx->offset = offset;
tx                455 drivers/net/xen-netfront.c 	tx->size = len;
tx                456 drivers/net/xen-netfront.c 	tx->flags = 0;
tx                458 drivers/net/xen-netfront.c 	info->tx = tx;
tx                459 drivers/net/xen-netfront.c 	info->size += tx->size;
tx                475 drivers/net/xen-netfront.c 	return info.tx;
tx                483 drivers/net/xen-netfront.c 	info->tx->flags |= XEN_NETTXF_more_data;
tx                489 drivers/net/xen-netfront.c 	struct netfront_queue *queue, struct xen_netif_tx_request *tx,
tx                496 drivers/net/xen-netfront.c 		.tx = tx,
tx                516 drivers/net/xen-netfront.c 	return info.tx;
tx                569 drivers/net/xen-netfront.c 	struct xen_netif_tx_request *tx, *first_tx;
tx                635 drivers/net/xen-netfront.c 	first_tx = tx = xennet_make_first_txreq(queue, skb,
tx                637 drivers/net/xen-netfront.c 	offset += tx->size;
tx                642 drivers/net/xen-netfront.c 	len -= tx->size;
tx                646 drivers/net/xen-netfront.c 		tx->flags |= XEN_NETTXF_csum_blank | XEN_NETTXF_data_validated;
tx                649 drivers/net/xen-netfront.c 		tx->flags |= XEN_NETTXF_data_validated;
tx                656 drivers/net/xen-netfront.c 			RING_GET_REQUEST(&queue->tx, queue->tx.req_prod_pvt++);
tx                658 drivers/net/xen-netfront.c 		tx->flags |= XEN_NETTXF_extra_info;
tx                672 drivers/net/xen-netfront.c 	tx = xennet_make_txreqs(queue, tx, skb, page, offset, len);
tx                677 drivers/net/xen-netfront.c 		tx = xennet_make_txreqs(queue, tx, skb, skb_frag_page(frag),
tx                685 drivers/net/xen-netfront.c 	RING_PUSH_REQUESTS_AND_CHECK_NOTIFY(&queue->tx, notify);
tx               1415 drivers/net/xen-netfront.c 		xennet_end_access(queue->tx_ring_ref, queue->tx.sring);
tx               1420 drivers/net/xen-netfront.c 		queue->tx.sring = NULL;
tx               1543 drivers/net/xen-netfront.c 	queue->tx.sring = NULL;
tx               1552 drivers/net/xen-netfront.c 	FRONT_RING_INIT(&queue->tx, txs, XEN_PAGE_SIZE);
tx                781 drivers/ntb/test/ntb_perf.c 	struct dma_async_tx_descriptor *tx;
tx                819 drivers/ntb/test/ntb_perf.c 		tx = dmaengine_prep_dma_memcpy(pthr->dma_chan, unmap->addr[1],
tx                821 drivers/ntb/test/ntb_perf.c 		if (!tx)
tx                823 drivers/ntb/test/ntb_perf.c 	} while (!tx && (try++ < DMA_TRIES));
tx                825 drivers/ntb/test/ntb_perf.c 	if (!tx) {
tx                830 drivers/ntb/test/ntb_perf.c 	tx->callback = perf_dma_copy_callback;
tx                831 drivers/ntb/test/ntb_perf.c 	tx->callback_param = pthr;
tx                832 drivers/ntb/test/ntb_perf.c 	dma_set_unmap(tx, unmap);
tx                834 drivers/ntb/test/ntb_perf.c 	ret = dma_submit_error(dmaengine_submit(tx));
tx               1014 drivers/pci/controller/dwc/pcie-tegra194.c 	msg.tx.data = &req;
tx               1015 drivers/pci/controller/dwc/pcie-tegra194.c 	msg.tx.size = sizeof(req);
tx                953 drivers/phy/qualcomm/phy-qcom-qmp.c 	void __iomem *tx;
tx               1447 drivers/phy/qualcomm/phy-qcom-qmp.c 	void __iomem *tx = qphy->tx;
tx               1504 drivers/phy/qualcomm/phy-qcom-qmp.c 	qcom_qmp_phy_configure(tx, cfg->regs, cfg->tx_tbl, cfg->tx_tbl_num);
tx               1881 drivers/phy/qualcomm/phy-qcom-qmp.c 	qphy->tx = of_iomap(np, 0);
tx               1882 drivers/phy/qualcomm/phy-qcom-qmp.c 	if (!qphy->tx)
tx               1908 drivers/phy/qualcomm/phy-qcom-qmp.c 			qphy->tx2 = qphy->tx + QMP_PHY_LEGACY_LANE_STRIDE;
tx                353 drivers/pinctrl/pinctrl-mcp23s08.c 	u8 tx[2];
tx                358 drivers/pinctrl/pinctrl-mcp23s08.c 	tx[0] = mcp->addr | 0x01;
tx                359 drivers/pinctrl/pinctrl-mcp23s08.c 	tx[1] = *((u8 *) reg);
tx                361 drivers/pinctrl/pinctrl-mcp23s08.c 	return spi_write_then_read(spi, tx, sizeof(tx), val, val_size);
tx               1295 drivers/pinctrl/sh-pfc/pfc-r8a7778.c #define CAN_PFC_DATA(name, tx, rx)		SH_PFC_MUX2(name, tx, rx)
tx               1338 drivers/pinctrl/sh-pfc/pfc-r8a7778.c #define SCIF_PFC_DAT(name, tx, rx)	SH_PFC_MUX2(name, tx, rx)
tx               1370 drivers/pinctrl/sh-pfc/pfc-r8a7778.c #define HSPI_PFC_DAT(name, clk, cs, rx, tx)	SH_PFC_MUX4(name, clk, cs, rx, tx)
tx                399 drivers/platform/mellanox/mlxbf-tmfifo.c 	int rx, tx;
tx                402 drivers/platform/mellanox/mlxbf-tmfifo.c 	tx = !test_and_set_bit(MLXBF_TM_TX_LWM_IRQ, &fifo->pend_events);
tx                404 drivers/platform/mellanox/mlxbf-tmfifo.c 	if (rx || tx)
tx                148 drivers/ps3/ps3-vuart.c 	u64 tx;
tx                151 drivers/ps3/ps3-vuart.c 		PARAM_TX_TRIGGER, &tx);
tx                152 drivers/ps3/ps3-vuart.c 	trig->tx = tx;
tx                181 drivers/ps3/ps3-vuart.c 		trig->tx, trig->rx);
tx                186 drivers/ps3/ps3-vuart.c int ps3_vuart_set_triggers(struct ps3_system_bus_device *dev, unsigned int tx,
tx                193 drivers/ps3/ps3-vuart.c 		PARAM_TX_TRIGGER, tx);
tx                220 drivers/ps3/ps3-vuart.c 		tx, rx);
tx                 75 drivers/ps3/vuart.h 	unsigned long tx;
tx                 80 drivers/ps3/vuart.h int ps3_vuart_set_triggers(struct ps3_system_bus_device *dev, unsigned int tx,
tx                115 drivers/ptp/ptp_clock.c static int ptp_clock_adjtime(struct posix_clock *pc, struct __kernel_timex *tx)
tx                123 drivers/ptp/ptp_clock.c 	if (tx->modes & ADJ_SETOFFSET) {
tx                128 drivers/ptp/ptp_clock.c 		ts.tv_sec  = tx->time.tv_sec;
tx                129 drivers/ptp/ptp_clock.c 		ts.tv_nsec = tx->time.tv_usec;
tx                131 drivers/ptp/ptp_clock.c 		if (!(tx->modes & ADJ_NANO))
tx                140 drivers/ptp/ptp_clock.c 	} else if (tx->modes & ADJ_FREQUENCY) {
tx                141 drivers/ptp/ptp_clock.c 		s32 ppb = scaled_ppm_to_ppb(tx->freq);
tx                145 drivers/ptp/ptp_clock.c 			err = ops->adjfine(ops, tx->freq);
tx                148 drivers/ptp/ptp_clock.c 		ptp->dialed_frequency = tx->freq;
tx                149 drivers/ptp/ptp_clock.c 	} else if (tx->modes == 0) {
tx                150 drivers/ptp/ptp_clock.c 		tx->freq = ptp->dialed_frequency;
tx                700 drivers/rapidio/devices/rio_mport_cdev.c 	struct dma_async_tx_descriptor *tx;
tx                719 drivers/rapidio/devices/rio_mport_cdev.c 	tx = prep_dma_xfer(chan, xfer, sgt, nents, dir,
tx                722 drivers/rapidio/devices/rio_mport_cdev.c 	if (!tx) {
tx                728 drivers/rapidio/devices/rio_mport_cdev.c 	} else if (IS_ERR(tx)) {
tx                729 drivers/rapidio/devices/rio_mport_cdev.c 		ret = PTR_ERR(tx);
tx                736 drivers/rapidio/devices/rio_mport_cdev.c 	tx->callback = dma_xfer_callback;
tx                737 drivers/rapidio/devices/rio_mport_cdev.c 	tx->callback_param = req;
tx                742 drivers/rapidio/devices/rio_mport_cdev.c 	cookie = dmaengine_submit(tx);
tx                 30 drivers/reset/tegra/reset-bpmp.c 	msg.tx.data = &request;
tx                 31 drivers/reset/tegra/reset-bpmp.c 	msg.tx.size = sizeof(request);
tx               1580 drivers/rpmsg/qcom_glink_native.c 					   struct qcom_glink_pipe *tx,
tx               1592 drivers/rpmsg/qcom_glink_native.c 	glink->tx_pipe = tx;
tx                 32 drivers/rpmsg/qcom_glink_native.h 					   struct qcom_glink_pipe *tx,
tx                186 drivers/rpmsg/qcom_glink_rpm.c 			       struct glink_rpm_pipe *tx)
tx                235 drivers/rpmsg/qcom_glink_rpm.c 			tx->native.length = size;
tx                237 drivers/rpmsg/qcom_glink_rpm.c 			tx->tail = msg_ram + offset;
tx                238 drivers/rpmsg/qcom_glink_rpm.c 			tx->head = msg_ram + offset + sizeof(u32);
tx                239 drivers/rpmsg/qcom_glink_rpm.c 			tx->fifo = msg_ram + offset + 2 * sizeof(u32);
tx                244 drivers/rpmsg/qcom_glink_rpm.c 	if (!rx->fifo || !tx->fifo) {
tx                249 drivers/rpmsg/qcom_smd.c 	struct smd_channel_info tx;
tx                271 drivers/rpmsg/qcom_smd.c 	struct smd_channel_info_word tx;
tx                311 drivers/rpmsg/qcom_smd.c 		BUILD_BUG_ON(sizeof(channel->info->tx.param) != sizeof(u8)); \
tx                313 drivers/rpmsg/qcom_smd.c 			le32_to_cpu(channel->info_word->tx.param) :          \
tx                314 drivers/rpmsg/qcom_smd.c 			channel->info->tx.param;			     \
tx                319 drivers/rpmsg/qcom_smd.c 		BUILD_BUG_ON(sizeof(channel->info->tx.param) != sizeof(u32)); \
tx                321 drivers/rpmsg/qcom_smd.c 			channel->info_word->tx.param :			      \
tx                322 drivers/rpmsg/qcom_smd.c 			channel->info->tx.param);			      \
tx                327 drivers/rpmsg/qcom_smd.c 		BUILD_BUG_ON(sizeof(channel->info->tx.param) != sizeof(u8)); \
tx                329 drivers/rpmsg/qcom_smd.c 			channel->info_word->tx.param = cpu_to_le32(value);   \
tx                331 drivers/rpmsg/qcom_smd.c 			channel->info->tx.param = value;		     \
tx                336 drivers/rpmsg/qcom_smd.c 		BUILD_BUG_ON(sizeof(channel->info->tx.param) != sizeof(u32)); \
tx                338 drivers/rpmsg/qcom_smd.c 			channel->info_word->tx.param = cpu_to_le32(value);   \
tx                340 drivers/rpmsg/qcom_smd.c 			channel->info->tx.param = cpu_to_le32(value);	      \
tx                495 drivers/rtc/rtc-ds1305.c 		u8 *addr, size_t count, char *tx, char *rx)
tx                506 drivers/rtc/rtc-ds1305.c 	x->tx_buf = tx;
tx                 64 drivers/rtc/rtc-mcp795.c 	u8 tx[2];
tx                 66 drivers/rtc/rtc-mcp795.c 	tx[0] = MCP795_READ;
tx                 67 drivers/rtc/rtc-mcp795.c 	tx[1] = addr;
tx                 68 drivers/rtc/rtc-mcp795.c 	ret = spi_write_then_read(spi, tx, sizeof(tx), buf, count);
tx                 81 drivers/rtc/rtc-mcp795.c 	u8 tx[257];
tx                 83 drivers/rtc/rtc-mcp795.c 	tx[0] = MCP795_WRITE;
tx                 84 drivers/rtc/rtc-mcp795.c 	tx[1] = addr;
tx                 85 drivers/rtc/rtc-mcp795.c 	memcpy(&tx[2], data, count);
tx                 87 drivers/rtc/rtc-mcp795.c 	ret = spi_write(spi, tx, 2 + count);
tx                226 drivers/scsi/cxgbi/libcxgbi.h 		struct cxgbi_skb_tx_cb tx;
tx                239 drivers/scsi/cxgbi/libcxgbi.h #define cxgbi_skcb_tx_wr_next(skb)	(CXGBI_SKB_CB(skb)->tx.wr_next)
tx                229 drivers/scsi/fnic/fnic_main.c 	stats->tx_frames = vs->tx.tx_unicast_frames_ok;
tx                230 drivers/scsi/fnic/fnic_main.c 	stats->tx_words  = vs->tx.tx_unicast_bytes_ok / 4;
tx                233 drivers/scsi/fnic/fnic_main.c 	stats->error_frames = vs->tx.tx_errors + vs->rx.rx_errors;
tx                234 drivers/scsi/fnic/fnic_main.c 	stats->dumped_frames = vs->tx.tx_drops + vs->rx.rx_drop;
tx                 64 drivers/scsi/fnic/vnic_stats.h 	struct vnic_tx_stats tx;
tx                205 drivers/scsi/mvsas/mv_chips.h static inline void mvs_start_delivery(struct mvs_info *mvi, u32 tx)
tx                208 drivers/scsi/mvsas/mv_chips.h 	mw32(MVS_TX_PROD_IDX, tx);
tx                113 drivers/scsi/mvsas/mv_init.c 	if (mvi->tx)
tx                115 drivers/scsi/mvsas/mv_init.c 				  sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ,
tx                116 drivers/scsi/mvsas/mv_init.c 				  mvi->tx, mvi->tx_dma);
tx                239 drivers/scsi/mvsas/mv_init.c 	mvi->tx = dma_alloc_coherent(mvi->dev,
tx                240 drivers/scsi/mvsas/mv_init.c 				     sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ,
tx                242 drivers/scsi/mvsas/mv_init.c 	if (!mvi->tx)
tx                244 drivers/scsi/mvsas/mv_init.c 	memset(mvi->tx, 0, sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ);
tx                379 drivers/scsi/mvsas/mv_sas.c 	slot->tx = mvi->tx_prod;
tx                380 drivers/scsi/mvsas/mv_sas.c 	mvi->tx[mvi->tx_prod] = cpu_to_le32((TXQ_CMD_SMP << TXQ_CMD_SHIFT) |
tx                452 drivers/scsi/mvsas/mv_sas.c 	slot->tx = mvi->tx_prod;
tx                457 drivers/scsi/mvsas/mv_sas.c 	mvi->tx[mvi->tx_prod] = cpu_to_le32(del_q);
tx                580 drivers/scsi/mvsas/mv_sas.c 	slot->tx = mvi->tx_prod;
tx                581 drivers/scsi/mvsas/mv_sas.c 	mvi->tx[mvi->tx_prod] = cpu_to_le32(TXQ_MODE_I | tag |
tx                121 drivers/scsi/mvsas/mv_sas.h 	void (*start_delivery)(struct mvs_info *mvi, u32 tx);
tx                313 drivers/scsi/mvsas/mv_sas.h 	u32 tx;
tx                349 drivers/scsi/mvsas/mv_sas.h 	__le32 *tx;
tx                 64 drivers/scsi/snic/vnic_stats.h 	struct vnic_tx_stats tx;
tx                110 drivers/slimbus/qcom-ctrl.c 	struct slim_ctrl_buf	tx;
tx                155 drivers/slimbus/qcom-ctrl.c 	spin_lock_irqsave(&ctrl->tx.lock, flags);
tx                156 drivers/slimbus/qcom-ctrl.c 	idx = ctrl->tx.head;
tx                157 drivers/slimbus/qcom-ctrl.c 	ctrl->tx.head = (ctrl->tx.head + 1) % ctrl->tx.n;
tx                158 drivers/slimbus/qcom-ctrl.c 	spin_unlock_irqrestore(&ctrl->tx.lock, flags);
tx                309 drivers/slimbus/qcom-ctrl.c 	spin_lock_irqsave(&ctrl->tx.lock, flags);
tx                310 drivers/slimbus/qcom-ctrl.c 	if (((ctrl->tx.head + 1) % ctrl->tx.n) == ctrl->tx.tail) {
tx                311 drivers/slimbus/qcom-ctrl.c 		spin_unlock_irqrestore(&ctrl->tx.lock, flags);
tx                315 drivers/slimbus/qcom-ctrl.c 	idx = ctrl->tx.tail;
tx                317 drivers/slimbus/qcom-ctrl.c 	ctrl->tx.tail = (ctrl->tx.tail + 1) % ctrl->tx.n;
tx                319 drivers/slimbus/qcom-ctrl.c 	spin_unlock_irqrestore(&ctrl->tx.lock, flags);
tx                321 drivers/slimbus/qcom-ctrl.c 	return ctrl->tx.base + (idx * ctrl->tx.sl_sz);
tx                537 drivers/slimbus/qcom-ctrl.c 	ctrl->tx.n = QCOM_TX_MSGS;
tx                538 drivers/slimbus/qcom-ctrl.c 	ctrl->tx.sl_sz = SLIM_MSGQ_BUF_LEN;
tx                547 drivers/slimbus/qcom-ctrl.c 	spin_lock_init(&ctrl->tx.lock);
tx                578 drivers/slimbus/qcom-ctrl.c 	ctrl->tx.base = devm_kcalloc(&pdev->dev, ctrl->tx.n, ctrl->tx.sl_sz,
tx                580 drivers/slimbus/qcom-ctrl.c 	if (!ctrl->tx.base) {
tx                415 drivers/slimbus/slimbus.h 					    struct slim_msg_txn *tx);
tx                141 drivers/soc/fsl/qbman/qman_test_stash.c 	struct qman_fq tx;
tx                286 drivers/soc/fsl/qbman/qman_test_stash.c 	if (qman_enqueue(&handler->tx, &dqrr->fd)) {
tx                302 drivers/soc/fsl/qbman/qman_test_stash.c 		if (qman_enqueue(&handler->tx, &dqrr->fd)) {
tx                366 drivers/soc/fsl/qbman/qman_test_stash.c 		qman_destroy_fq(&handler->tx);
tx                420 drivers/soc/fsl/qbman/qman_test_stash.c 	memset(&handler->tx, 0, sizeof(handler->tx));
tx                422 drivers/soc/fsl/qbman/qman_test_stash.c 			     &handler->tx);
tx                541 drivers/soc/fsl/qbman/qman_test_stash.c 	err = qman_enqueue(&special_handler->tx, &fd);
tx                 46 drivers/soc/tegra/powergate-bpmp.c 	msg.tx.data = &request;
tx                 47 drivers/soc/tegra/powergate-bpmp.c 	msg.tx.size = sizeof(request);
tx                 74 drivers/soc/tegra/powergate-bpmp.c 	msg.tx.data = &request;
tx                 75 drivers/soc/tegra/powergate-bpmp.c 	msg.tx.size = sizeof(request);
tx                102 drivers/soc/tegra/powergate-bpmp.c 	msg.tx.data = &request;
tx                103 drivers/soc/tegra/powergate-bpmp.c 	msg.tx.size = sizeof(request);
tx                132 drivers/soc/tegra/powergate-bpmp.c 	msg.tx.data = &request;
tx                133 drivers/soc/tegra/powergate-bpmp.c 	msg.tx.size = sizeof(request);
tx                159 drivers/soc/ti/knav_dma.c 		if (cfg->u.tx.filt_pswords)
tx                161 drivers/soc/ti/knav_dma.c 		if (cfg->u.tx.filt_einfo)
tx                168 drivers/soc/ti/knav_dma.c 		writel_relaxed(cfg->u.tx.priority, &chan->reg_tx_sched->prio);
tx                330 drivers/soc/ti/knav_dma.c 			chan->cfg.u.tx.filt_einfo,
tx                331 drivers/soc/ti/knav_dma.c 			chan->cfg.u.tx.filt_pswords,
tx                332 drivers/soc/ti/knav_dma.c 			chan->cfg.u.tx.priority);
tx                 52 drivers/spi/spi-altera.c 	const unsigned char *tx;
tx                 80 drivers/spi/spi-altera.c 	if (hw->tx) {
tx                 83 drivers/spi/spi-altera.c 			txd = hw->tx[hw->count];
tx                 86 drivers/spi/spi-altera.c 			txd = (hw->tx[hw->count * 2]
tx                 87 drivers/spi/spi-altera.c 				| (hw->tx[hw->count * 2 + 1] << 8));
tx                120 drivers/spi/spi-altera.c 	hw->tx = t->tx_buf;
tx                 49 drivers/spi/spi-au1550.c 	const u8 *tx;
tx                316 drivers/spi/spi-au1550.c 	hw->tx = t->tx_buf;
tx                366 drivers/spi/spi-au1550.c 		hw->tx = hw->rx;
tx                375 drivers/spi/spi-au1550.c 	res = au1xxx_dbdma_put_source(hw->dma_tx_ch, virt_to_phys(hw->tx),
tx                485 drivers/spi/spi-au1550.c 	if (hw->tx) {							\
tx                486 drivers/spi/spi-au1550.c 		fifoword = *(u##size *)hw->tx & (u32)(mask);		\
tx                487 drivers/spi/spi-au1550.c 		hw->tx += (size) / 8;					\
tx                508 drivers/spi/spi-au1550.c 	hw->tx = t->tx_buf;
tx                162 drivers/spi/spi-bcm63xx-hsspi.c 	const u8 *tx = t->tx_buf;
tx                168 drivers/spi/spi-bcm63xx-hsspi.c 	if (tx && rx)
tx                170 drivers/spi/spi-bcm63xx-hsspi.c 	else if (tx)
tx                190 drivers/spi/spi-bcm63xx-hsspi.c 		if (tx) {
tx                191 drivers/spi/spi-bcm63xx-hsspi.c 			memcpy_toio(bs->fifo + HSSPI_OPCODE_LEN, tx, curr_step);
tx                192 drivers/spi/spi-bcm63xx-hsspi.c 			tx += curr_step;
tx                 66 drivers/spi/spi-bitbang.c 	const u8		*tx = t->tx_buf;
tx                 72 drivers/spi/spi-bitbang.c 		if (tx)
tx                 73 drivers/spi/spi-bitbang.c 			word = *tx++;
tx                 94 drivers/spi/spi-bitbang.c 	const u16		*tx = t->tx_buf;
tx                100 drivers/spi/spi-bitbang.c 		if (tx)
tx                101 drivers/spi/spi-bitbang.c 			word = *tx++;
tx                122 drivers/spi/spi-bitbang.c 	const u32		*tx = t->tx_buf;
tx                128 drivers/spi/spi-bitbang.c 		if (tx)
tx                129 drivers/spi/spi-bitbang.c 			word = *tx++;
tx                 43 drivers/spi/spi-cavium-octeon.c 	p->regs.tx = 0x10;
tx                 49 drivers/spi/spi-cavium-thunderx.c 	p->regs.tx = 0x1010;
tx                 13 drivers/spi/spi-cavium.h 	int tx;
tx                 28 drivers/spi/spi-cavium.h #define OCTEON_SPI_TX(x)	(x->regs.tx)
tx                115 drivers/spi/spi-davinci.c 	const void		*tx;
tx                157 drivers/spi/spi-davinci.c 	if (dspi->tx) {
tx                158 drivers/spi/spi-davinci.c 		const u8 *tx = dspi->tx;
tx                160 drivers/spi/spi-davinci.c 		data = *tx++;
tx                161 drivers/spi/spi-davinci.c 		dspi->tx = tx;
tx                170 drivers/spi/spi-davinci.c 	if (dspi->tx) {
tx                171 drivers/spi/spi-davinci.c 		const u16 *tx = dspi->tx;
tx                173 drivers/spi/spi-davinci.c 		data = *tx++;
tx                174 drivers/spi/spi-davinci.c 		dspi->tx = tx;
tx                590 drivers/spi/spi-davinci.c 	dspi->tx = t->tx_buf;
tx                107 drivers/spi/spi-dln2.c 	} tx;
tx                108 drivers/spi/spi-dln2.c 	unsigned len = sizeof(tx);
tx                110 drivers/spi/spi-dln2.c 	tx.port = dln2->port;
tx                114 drivers/spi/spi-dln2.c 		len -= sizeof(tx.wait_for_completion);
tx                116 drivers/spi/spi-dln2.c 		tx.wait_for_completion = DLN2_TRANSFERS_WAIT_COMPLETE;
tx                120 drivers/spi/spi-dln2.c 	return dln2_transfer_tx(dln2->pdev, cmd, &tx, len);
tx                136 drivers/spi/spi-dln2.c 	} tx;
tx                138 drivers/spi/spi-dln2.c 	tx.port = dln2->port;
tx                145 drivers/spi/spi-dln2.c 	tx.cs = ~cs_mask;
tx                147 drivers/spi/spi-dln2.c 	return dln2_transfer_tx(dln2->pdev, DLN2_SPI_SET_SS, &tx, sizeof(tx));
tx                166 drivers/spi/spi-dln2.c 	} tx;
tx                169 drivers/spi/spi-dln2.c 	tx.port = dln2->port;
tx                170 drivers/spi/spi-dln2.c 	tx.cs = cs_mask;
tx                173 drivers/spi/spi-dln2.c 	return dln2_transfer_tx(dln2->pdev, cmd, &tx, sizeof(tx));
tx                188 drivers/spi/spi-dln2.c 	} tx;
tx                194 drivers/spi/spi-dln2.c 	tx.port = dln2->port;
tx                195 drivers/spi/spi-dln2.c 	ret = dln2_transfer(dln2->pdev, DLN2_SPI_GET_SS_COUNT, &tx, sizeof(tx),
tx                214 drivers/spi/spi-dln2.c 	} tx;
tx                220 drivers/spi/spi-dln2.c 	tx.port = dln2->port;
tx                222 drivers/spi/spi-dln2.c 	ret = dln2_transfer(dln2->pdev, cmd, &tx, sizeof(tx), &rx, &rx_len);
tx                264 drivers/spi/spi-dln2.c 	} __packed tx;
tx                270 drivers/spi/spi-dln2.c 	tx.port = dln2->port;
tx                271 drivers/spi/spi-dln2.c 	tx.speed = cpu_to_le32(speed);
tx                273 drivers/spi/spi-dln2.c 	ret = dln2_transfer(dln2->pdev, DLN2_SPI_SET_FREQUENCY, &tx, sizeof(tx),
tx                291 drivers/spi/spi-dln2.c 	} tx;
tx                293 drivers/spi/spi-dln2.c 	tx.port = dln2->port;
tx                294 drivers/spi/spi-dln2.c 	tx.mode = mode;
tx                296 drivers/spi/spi-dln2.c 	return dln2_transfer_tx(dln2->pdev, DLN2_SPI_SET_MODE, &tx, sizeof(tx));
tx                307 drivers/spi/spi-dln2.c 	} tx;
tx                309 drivers/spi/spi-dln2.c 	tx.port = dln2->port;
tx                310 drivers/spi/spi-dln2.c 	tx.bpw = bpw;
tx                313 drivers/spi/spi-dln2.c 				&tx, sizeof(tx));
tx                322 drivers/spi/spi-dln2.c 	} tx;
tx                330 drivers/spi/spi-dln2.c 	tx.port = dln2->port;
tx                333 drivers/spi/spi-dln2.c 			    &tx, sizeof(tx), rx, &rx_len);
tx                426 drivers/spi/spi-dln2.c 	} __packed *tx = dln2->buf;
tx                429 drivers/spi/spi-dln2.c 	BUILD_BUG_ON(sizeof(*tx) > DLN2_SPI_BUF_SIZE);
tx                434 drivers/spi/spi-dln2.c 	tx->port = dln2->port;
tx                435 drivers/spi/spi-dln2.c 	tx->size = cpu_to_le16(data_len);
tx                436 drivers/spi/spi-dln2.c 	tx->attr = attr;
tx                438 drivers/spi/spi-dln2.c 	dln2_spi_copy_to_buf(tx->buf, data, data_len, dln2->bpw);
tx                440 drivers/spi/spi-dln2.c 	tx_len = sizeof(*tx) + data_len - DLN2_SPI_MAX_XFER_SIZE;
tx                441 drivers/spi/spi-dln2.c 	return dln2_transfer_tx(dln2->pdev, DLN2_SPI_WRITE, tx, tx_len);
tx                455 drivers/spi/spi-dln2.c 	} __packed tx;
tx                467 drivers/spi/spi-dln2.c 	tx.port = dln2->port;
tx                468 drivers/spi/spi-dln2.c 	tx.size = cpu_to_le16(data_len);
tx                469 drivers/spi/spi-dln2.c 	tx.attr = attr;
tx                471 drivers/spi/spi-dln2.c 	ret = dln2_transfer(dln2->pdev, DLN2_SPI_READ, &tx, sizeof(tx),
tx                497 drivers/spi/spi-dln2.c 	} __packed *tx;
tx                504 drivers/spi/spi-dln2.c 	BUILD_BUG_ON(sizeof(*tx) > DLN2_SPI_BUF_SIZE ||
tx                515 drivers/spi/spi-dln2.c 	tx = dln2->buf;
tx                518 drivers/spi/spi-dln2.c 	tx->port = dln2->port;
tx                519 drivers/spi/spi-dln2.c 	tx->size = cpu_to_le16(data_len);
tx                520 drivers/spi/spi-dln2.c 	tx->attr = attr;
tx                522 drivers/spi/spi-dln2.c 	dln2_spi_copy_to_buf(tx->buf, tx_data, data_len, dln2->bpw);
tx                524 drivers/spi/spi-dln2.c 	tx_len = sizeof(*tx) + data_len - DLN2_SPI_MAX_XFER_SIZE;
tx                527 drivers/spi/spi-dln2.c 	ret = dln2_transfer(dln2->pdev, DLN2_SPI_READ_WRITE, tx, tx_len,
tx                 41 drivers/spi/spi-dw-mid.c 	struct dw_dma_slave *tx = dws->dma_tx;
tx                 64 drivers/spi/spi-dw-mid.c 	tx->dma_dev = &dma_dev->dev;
tx                 65 drivers/spi/spi-dw-mid.c 	dws->txchan = dma_request_channel(mask, mid_spi_dma_chan_filter, tx);
tx                156 drivers/spi/spi-dw.c 	tx_left = (dws->tx_end - dws->tx) / dws->n_bytes;
tx                167 drivers/spi/spi-dw.c 	rxtx_gap =  ((dws->rx_end - dws->rx) - (dws->tx_end - dws->tx))
tx                192 drivers/spi/spi-dw.c 				txw = *(u8 *)(dws->tx);
tx                194 drivers/spi/spi-dw.c 				txw = *(u16 *)(dws->tx);
tx                197 drivers/spi/spi-dw.c 		dws->tx += dws->n_bytes;
tx                301 drivers/spi/spi-dw.c 	dws->tx = (void *)transfer->tx_buf;
tx                302 drivers/spi/spi-dw.c 	dws->tx_end = dws->tx + transfer->len;
tx                340 drivers/spi/spi-dw.c 		if (dws->rx && dws->tx)
tx                121 drivers/spi/spi-dw.h 	void			*tx;
tx                 89 drivers/spi/spi-ep93xx.c 	size_t				tx;
tx                183 drivers/spi/spi-ep93xx.c 			val = ((u16 *)xfer->tx_buf)[espi->tx];
tx                184 drivers/spi/spi-ep93xx.c 		espi->tx += 2;
tx                187 drivers/spi/spi-ep93xx.c 			val = ((u8 *)xfer->tx_buf)[espi->tx];
tx                188 drivers/spi/spi-ep93xx.c 		espi->tx += 1;
tx                234 drivers/spi/spi-ep93xx.c 	while (espi->fifo_level < SPI_FIFO_SIZE && espi->tx < xfer->len) {
tx                494 drivers/spi/spi-ep93xx.c 	espi->tx = 0;
tx                125 drivers/spi/spi-fsl-cpm.c 		void *nonconst_tx = (void *)mspi->tx; /* shut up gcc */
tx                186 drivers/spi/spi-fsl-dspi.c 	const void				*tx;
tx                204 drivers/spi/spi-fsl-dspi.c 	if (dspi->tx) {
tx                206 drivers/spi/spi-fsl-dspi.c 			txdata = *(u8 *)dspi->tx;
tx                208 drivers/spi/spi-fsl-dspi.c 			txdata = *(u16 *)dspi->tx;
tx                210 drivers/spi/spi-fsl-dspi.c 			txdata = *(u32 *)dspi->tx;
tx                211 drivers/spi/spi-fsl-dspi.c 		dspi->tx += dspi->bytes_per_word;
tx                748 drivers/spi/spi-fsl-dspi.c 		dspi->tx = transfer->tx_buf;
tx                 42 drivers/spi/spi-fsl-lib.c 	const type *tx = mpc8xxx_spi->tx;			\
tx                 43 drivers/spi/spi-fsl-lib.c 	if (!tx)						\
tx                 45 drivers/spi/spi-fsl-lib.c 	data = *tx++ << mpc8xxx_spi->tx_shift;			\
tx                 46 drivers/spi/spi-fsl-lib.c 	mpc8xxx_spi->tx = tx;					\
tx                 25 drivers/spi/spi-fsl-lib.h 	const void *tx;
tx                110 drivers/spi/spi-fsl-lpspi.c 	void (*tx)(struct fsl_lpspi_data *);
tx                254 drivers/spi/spi-fsl-lpspi.c 		fsl_lpspi->tx(fsl_lpspi);
tx                358 drivers/spi/spi-fsl-lpspi.c 	struct dma_slave_config rx = {}, tx = {};
tx                376 drivers/spi/spi-fsl-lpspi.c 	tx.direction = DMA_MEM_TO_DEV;
tx                377 drivers/spi/spi-fsl-lpspi.c 	tx.dst_addr = fsl_lpspi->base_phys + IMX7ULP_TDR;
tx                378 drivers/spi/spi-fsl-lpspi.c 	tx.dst_addr_width = buswidth;
tx                379 drivers/spi/spi-fsl-lpspi.c 	tx.dst_maxburst = 1;
tx                380 drivers/spi/spi-fsl-lpspi.c 	ret = dmaengine_slave_config(controller->dma_tx, &tx);
tx                457 drivers/spi/spi-fsl-lpspi.c 		fsl_lpspi->tx = fsl_lpspi_buf_tx_u8;
tx                460 drivers/spi/spi-fsl-lpspi.c 		fsl_lpspi->tx = fsl_lpspi_buf_tx_u16;
tx                463 drivers/spi/spi-fsl-lpspi.c 		fsl_lpspi->tx = fsl_lpspi_buf_tx_u32;
tx                572 drivers/spi/spi-fsl-lpspi.c 	struct sg_table *tx = &transfer->tx_sg, *rx = &transfer->rx_sg;
tx                592 drivers/spi/spi-fsl-lpspi.c 				tx->sgl, tx->nents, DMA_MEM_TO_DEV,
tx                335 drivers/spi/spi-fsl-spi.c 	mpc8xxx_spi->tx = t->tx_buf;
tx                 98 drivers/spi/spi-imx.c 	void (*tx)(struct spi_imx_data *);
tx               1112 drivers/spi/spi-imx.c 		spi_imx->tx(spi_imx);
tx               1154 drivers/spi/spi-imx.c 	struct dma_slave_config rx = {}, tx = {};
tx               1171 drivers/spi/spi-imx.c 	tx.direction = DMA_MEM_TO_DEV;
tx               1172 drivers/spi/spi-imx.c 	tx.dst_addr = spi_imx->base_phys + MXC_CSPITXDATA;
tx               1173 drivers/spi/spi-imx.c 	tx.dst_addr_width = buswidth;
tx               1174 drivers/spi/spi-imx.c 	tx.dst_maxburst = spi_imx->wml;
tx               1175 drivers/spi/spi-imx.c 	ret = dmaengine_slave_config(master->dma_tx, &tx);
tx               1215 drivers/spi/spi-imx.c 		spi_imx->tx = spi_imx_buf_tx_swap;
tx               1221 drivers/spi/spi-imx.c 			spi_imx->tx = spi_imx_buf_tx_u8;
tx               1224 drivers/spi/spi-imx.c 			spi_imx->tx = spi_imx_buf_tx_u16;
tx               1227 drivers/spi/spi-imx.c 			spi_imx->tx = spi_imx_buf_tx_u32;
tx               1239 drivers/spi/spi-imx.c 		spi_imx->tx = mx53_ecspi_tx_slave;
tx               1340 drivers/spi/spi-imx.c 	struct sg_table *tx = &transfer->tx_sg, *rx = &transfer->rx_sg;
tx               1384 drivers/spi/spi-imx.c 				tx->sgl, tx->nents, DMA_MEM_TO_DEV,
tx                106 drivers/spi/spi-jcore.c 	const unsigned char *tx;
tx                114 drivers/spi/spi-jcore.c 	tx = t->tx_buf;
tx                122 drivers/spi/spi-jcore.c 		writel(tx ? *tx++ : 0, data_reg);
tx                178 drivers/spi/spi-lantiq-ssc.c 	const u8			*tx;
tx                487 drivers/spi/spi-lantiq-ssc.c 			tx8 = spi->tx;
tx                490 drivers/spi/spi-lantiq-ssc.c 			spi->tx++;
tx                493 drivers/spi/spi-lantiq-ssc.c 			tx16 = (u16 *) spi->tx;
tx                496 drivers/spi/spi-lantiq-ssc.c 			spi->tx += 2;
tx                499 drivers/spi/spi-lantiq-ssc.c 			tx32 = (u32 *) spi->tx;
tx                502 drivers/spi/spi-lantiq-ssc.c 			spi->tx += 4;
tx                617 drivers/spi/spi-lantiq-ssc.c 	if (spi->tx) {
tx                685 drivers/spi/spi-lantiq-ssc.c 	spi->tx = t->tx_buf;
tx                699 drivers/spi/spi-lantiq-ssc.c 		if (!spi->tx)
tx                560 drivers/spi/spi-loopback-test.c 					  void *tx, void *rx)
tx                622 drivers/spi/spi-loopback-test.c 			      void *tx, void *rx)
tx                649 drivers/spi/spi-loopback-test.c 		*ptr = tx + off;
tx                749 drivers/spi/spi-loopback-test.c 			      void *tx, void *rx)
tx                767 drivers/spi/spi-loopback-test.c 					 (void *)tx, rx);
tx                773 drivers/spi/spi-loopback-test.c 					 (void *)tx, rx);
tx                788 drivers/spi/spi-loopback-test.c 		ret = test->execute_msg(spi, test, tx, rx);
tx                790 drivers/spi/spi-loopback-test.c 		ret = spi_test_execute_msg(spi, test, tx, rx);
tx                811 drivers/spi/spi-loopback-test.c 			     void *tx, void *rx,
tx                876 drivers/spi/spi-loopback-test.c 	return _spi_test_run_iter(spi, &test, tx, rx);
tx                890 drivers/spi/spi-loopback-test.c 			 void *tx, void *rx)
tx                932 drivers/spi/spi-loopback-test.c 		ret = spi_test_check_loopback_result(spi, msg, tx, rx);
tx                963 drivers/spi/spi-loopback-test.c 		      void *tx, void *rx)
tx               1001 drivers/spi/spi-loopback-test.c 							tx, rx,
tx               1026 drivers/spi/spi-loopback-test.c 	char *rx = NULL, *tx = NULL;
tx               1042 drivers/spi/spi-loopback-test.c 		tx = vmalloc(SPI_TEST_MAX_SIZE_PLUS);
tx               1044 drivers/spi/spi-loopback-test.c 		tx = kzalloc(SPI_TEST_MAX_SIZE_PLUS, GFP_KERNEL);
tx               1045 drivers/spi/spi-loopback-test.c 	if (!tx) {
tx               1058 drivers/spi/spi-loopback-test.c 			ret = test->run_test(spi, test, tx, rx);
tx               1060 drivers/spi/spi-loopback-test.c 			ret = spi_test_run_test(spi, test, tx, rx);
tx               1072 drivers/spi/spi-loopback-test.c 	kvfree(tx);
tx                105 drivers/spi/spi-lp8841-rtc.c 	const u8		*tx = t->tx_buf;
tx                110 drivers/spi/spi-lp8841-rtc.c 	if (tx) {
tx                114 drivers/spi/spi-lp8841-rtc.c 			word = *tx++;
tx                102 drivers/spi/spi-mem.c static int spi_check_buswidth_req(struct spi_mem *mem, u8 buswidth, bool tx)
tx                111 drivers/spi/spi-mem.c 		if ((tx && (mode & (SPI_TX_DUAL | SPI_TX_QUAD))) ||
tx                112 drivers/spi/spi-mem.c 		    (!tx && (mode & (SPI_RX_DUAL | SPI_RX_QUAD))))
tx                118 drivers/spi/spi-mem.c 		if ((tx && (mode & SPI_TX_QUAD)) ||
tx                119 drivers/spi/spi-mem.c 		    (!tx && (mode & SPI_RX_QUAD)))
tx                125 drivers/spi/spi-mem.c 		if ((tx && (mode & SPI_TX_OCTAL)) ||
tx                126 drivers/spi/spi-mem.c 		    (!tx && (mode & SPI_RX_OCTAL)))
tx                191 drivers/spi/spi-omap-100k.c 		const u8        *tx;
tx                194 drivers/spi/spi-omap-100k.c 		tx = xfer->tx_buf;
tx                198 drivers/spi/spi-omap-100k.c 				spi100k_write_data(spi->master, word_len, *tx++);
tx                204 drivers/spi/spi-omap-100k.c 		const u16       *tx;
tx                207 drivers/spi/spi-omap-100k.c 		tx = xfer->tx_buf;
tx                211 drivers/spi/spi-omap-100k.c 				spi100k_write_data(spi->master, word_len, *tx++);
tx                217 drivers/spi/spi-omap-100k.c 		const u32       *tx;
tx                220 drivers/spi/spi-omap-100k.c 		tx = xfer->tx_buf;
tx                224 drivers/spi/spi-omap-100k.c 				spi100k_write_data(spi->master, word_len, *tx);
tx                405 drivers/spi/spi-omap2-mcspi.c 		struct dma_async_tx_descriptor *tx;
tx                409 drivers/spi/spi-omap2-mcspi.c 		tx = dmaengine_prep_slave_sg(mcspi_dma->dma_tx, xfer->tx_sg.sgl,
tx                413 drivers/spi/spi-omap2-mcspi.c 		if (tx) {
tx                414 drivers/spi/spi-omap2-mcspi.c 			tx->callback = omap2_mcspi_tx_callback;
tx                415 drivers/spi/spi-omap2-mcspi.c 			tx->callback_param = spi;
tx                416 drivers/spi/spi-omap2-mcspi.c 			dmaengine_submit(tx);
tx                466 drivers/spi/spi-omap2-mcspi.c 		struct dma_async_tx_descriptor *tx;
tx                502 drivers/spi/spi-omap2-mcspi.c 		tx = dmaengine_prep_slave_sg(mcspi_dma->dma_rx,
tx                507 drivers/spi/spi-omap2-mcspi.c 		if (tx) {
tx                508 drivers/spi/spi-omap2-mcspi.c 			tx->callback = omap2_mcspi_rx_callback;
tx                509 drivers/spi/spi-omap2-mcspi.c 			tx->callback_param = spi;
tx                510 drivers/spi/spi-omap2-mcspi.c 			dmaengine_submit(tx);
tx                588 drivers/spi/spi-omap2-mcspi.c 	const u8		*tx;
tx                621 drivers/spi/spi-omap2-mcspi.c 	tx = xfer->tx_buf;
tx                627 drivers/spi/spi-omap2-mcspi.c 	if (tx) {
tx                639 drivers/spi/spi-omap2-mcspi.c 	if (tx != NULL) {
tx                718 drivers/spi/spi-omap2-mcspi.c 		const u8	*tx;
tx                721 drivers/spi/spi-omap2-mcspi.c 		tx = xfer->tx_buf;
tx                725 drivers/spi/spi-omap2-mcspi.c 			if (tx != NULL) {
tx                732 drivers/spi/spi-omap2-mcspi.c 						word_len, *tx);
tx                733 drivers/spi/spi-omap2-mcspi.c 				writel_relaxed(*tx++, tx_reg);
tx                742 drivers/spi/spi-omap2-mcspi.c 				if (c == 1 && tx == NULL &&
tx                755 drivers/spi/spi-omap2-mcspi.c 				} else if (c == 0 && tx == NULL) {
tx                766 drivers/spi/spi-omap2-mcspi.c 		const u16	*tx;
tx                769 drivers/spi/spi-omap2-mcspi.c 		tx = xfer->tx_buf;
tx                772 drivers/spi/spi-omap2-mcspi.c 			if (tx != NULL) {
tx                779 drivers/spi/spi-omap2-mcspi.c 						word_len, *tx);
tx                780 drivers/spi/spi-omap2-mcspi.c 				writel_relaxed(*tx++, tx_reg);
tx                789 drivers/spi/spi-omap2-mcspi.c 				if (c == 2 && tx == NULL &&
tx                802 drivers/spi/spi-omap2-mcspi.c 				} else if (c == 0 && tx == NULL) {
tx                813 drivers/spi/spi-omap2-mcspi.c 		const u32	*tx;
tx                816 drivers/spi/spi-omap2-mcspi.c 		tx = xfer->tx_buf;
tx                819 drivers/spi/spi-omap2-mcspi.c 			if (tx != NULL) {
tx                826 drivers/spi/spi-omap2-mcspi.c 						word_len, *tx);
tx                827 drivers/spi/spi-omap2-mcspi.c 				writel_relaxed(*tx++, tx_reg);
tx                836 drivers/spi/spi-omap2-mcspi.c 				if (c == 4 && tx == NULL &&
tx                849 drivers/spi/spi-omap2-mcspi.c 				} else if (c == 0 && tx == NULL) {
tx                463 drivers/spi/spi-orion.c 		const u8 *tx = xfer->tx_buf;
tx                467 drivers/spi/spi-orion.c 			if (orion_spi_write_read_8bit(spi, &tx, &rx) < 0)
tx                474 drivers/spi/spi-orion.c 		const u16 *tx = xfer->tx_buf;
tx                478 drivers/spi/spi-orion.c 			if (orion_spi_write_read_16bit(spi, &tx, &rx) < 0)
tx                114 drivers/spi/spi-pic32.c 	const void		*tx;
tx                165 drivers/spi/spi-pic32.c 	tx_left = (pic32s->tx_end - pic32s->tx) / n_bytes;
tx                177 drivers/spi/spi-pic32.c 		    (pic32s->tx_end - pic32s->tx)) / n_bytes;
tx                209 drivers/spi/spi-pic32.c 			v = *(__type *)(pic32s->tx);		\
tx                211 drivers/spi/spi-pic32.c 		pic32s->tx += sizeof(__type);			\
tx                287 drivers/spi/spi-pic32.c 	if (pic32s->tx_end == pic32s->tx)
tx                536 drivers/spi/spi-pic32.c 		pic32s->tx = (const void *)transfer->tx_buf;
tx                538 drivers/spi/spi-pic32.c 		pic32s->tx_end = pic32s->tx + transfer->len;
tx                378 drivers/spi/spi-pl022.c 	void				*tx;
tx                685 drivers/spi/spi-pl022.c 		__func__, pl022->rx, pl022->rx_end, pl022->tx, pl022->tx_end);
tx                714 drivers/spi/spi-pl022.c 	       && (pl022->tx < pl022->tx_end)) {
tx                720 drivers/spi/spi-pl022.c 			writew(*(u8 *) (pl022->tx), SSP_DR(pl022->virtbase));
tx                723 drivers/spi/spi-pl022.c 			writew((*(u16 *) (pl022->tx)), SSP_DR(pl022->virtbase));
tx                726 drivers/spi/spi-pl022.c 			writel(*(u32 *) (pl022->tx), SSP_DR(pl022->virtbase));
tx                729 drivers/spi/spi-pl022.c 		pl022->tx += (pl022->cur_chip->n_bytes);
tx               1046 drivers/spi/spi-pl022.c 	setup_dma_scatter(pl022, pl022->tx,
tx               1303 drivers/spi/spi-pl022.c 	if (pl022->tx == pl022->tx_end) {
tx               1358 drivers/spi/spi-pl022.c 	pl022->tx = (void *)transfer->tx_buf;
tx               1359 drivers/spi/spi-pl022.c 	pl022->tx_end = pl022->tx + pl022->cur_transfer->len;
tx               1363 drivers/spi/spi-pl022.c 	    pl022->tx ? pl022->cur_chip->write : WRITING_NULL;
tx               1548 drivers/spi/spi-pl022.c 		while (pl022->tx < pl022->tx_end || pl022->rx < pl022->rx_end) {
tx                127 drivers/spi/spi-ppc4xx.c 	const unsigned char *tx;
tx                152 drivers/spi/spi-ppc4xx.c 	hw->tx = t->tx_buf;
tx                158 drivers/spi/spi-ppc4xx.c 	data = hw->tx ? hw->tx[0] : 0;
tx                335 drivers/spi/spi-ppc4xx.c 		data = hw->tx ? hw->tx[count] : 0;
tx                106 drivers/spi/spi-pxa2xx-pci.c 	struct dw_dma_slave *tx, *rx;
tx                131 drivers/spi/spi-pxa2xx-pci.c 	tx = c->tx_param;
tx                132 drivers/spi/spi-pxa2xx-pci.c 	tx->dma_dev = &dma_dev->dev;
tx                490 drivers/spi/spi-pxa2xx.c 		|| (drv_data->tx == drv_data->tx_end))
tx                494 drivers/spi/spi-pxa2xx.c 	drv_data->tx += n_bytes;
tx                515 drivers/spi/spi-pxa2xx.c 		|| (drv_data->tx == drv_data->tx_end))
tx                518 drivers/spi/spi-pxa2xx.c 	pxa2xx_spi_write(drv_data, SSDR, *(u8 *)(drv_data->tx));
tx                519 drivers/spi/spi-pxa2xx.c 	++drv_data->tx;
tx                538 drivers/spi/spi-pxa2xx.c 		|| (drv_data->tx == drv_data->tx_end))
tx                541 drivers/spi/spi-pxa2xx.c 	pxa2xx_spi_write(drv_data, SSDR, *(u16 *)(drv_data->tx));
tx                542 drivers/spi/spi-pxa2xx.c 	drv_data->tx += 2;
tx                561 drivers/spi/spi-pxa2xx.c 		|| (drv_data->tx == drv_data->tx_end))
tx                564 drivers/spi/spi-pxa2xx.c 	pxa2xx_spi_write(drv_data, SSDR, *(u32 *)(drv_data->tx));
tx                565 drivers/spi/spi-pxa2xx.c 	drv_data->tx += 4;
tx                669 drivers/spi/spi-pxa2xx.c 	if (drv_data->tx == drv_data->tx_end) {
tx                982 drivers/spi/spi-pxa2xx.c 	drv_data->tx = (void *)transfer->tx_buf;
tx                983 drivers/spi/spi-pxa2xx.c 	drv_data->tx_end = drv_data->tx + transfer->len;
tx                986 drivers/spi/spi-pxa2xx.c 	drv_data->write = drv_data->tx ? chip->write : null_writer;
tx               1109 drivers/spi/spi-pxa2xx.c 			drv_data->tx += tx_level;
tx                 50 drivers/spi/spi-pxa2xx.h 	void *tx;
tx                170 drivers/spi/spi-rockchip.c 	const void *tx;
tx                277 drivers/spi/spi-rockchip.c 			txw = *(u8 *)rs->tx;
tx                279 drivers/spi/spi-rockchip.c 			txw = *(u16 *)rs->tx;
tx                282 drivers/spi/spi-rockchip.c 		rs->tx += rs->n_bytes;
tx                341 drivers/spi/spi-rockchip.c 	rs->tx = xfer->tx_buf;
tx                343 drivers/spi/spi-rockchip.c 	rs->tx_left = rs->tx ? xfer->len / rs->n_bytes : 0;
tx                493 drivers/spi/spi-rspi.c static int rspi_pio_transfer(struct rspi_data *rspi, const u8 *tx, u8 *rx,
tx                497 drivers/spi/spi-rspi.c 		if (tx) {
tx                498 drivers/spi/spi-rspi.c 			int ret = rspi_data_out(rspi, *tx++);
tx                521 drivers/spi/spi-rspi.c static int rspi_dma_transfer(struct rspi_data *rspi, struct sg_table *tx,
tx                551 drivers/spi/spi-rspi.c 	if (tx) {
tx                552 drivers/spi/spi-rspi.c 		desc_tx = dmaengine_prep_slave_sg(rspi->ctlr->dma_tx, tx->sgl,
tx                553 drivers/spi/spi-rspi.c 					tx->nents, DMA_MEM_TO_DEV,
tx                580 drivers/spi/spi-rspi.c 	if (tx)
tx                591 drivers/spi/spi-rspi.c 	if (tx)
tx                603 drivers/spi/spi-rspi.c 		if (tx)
tx                611 drivers/spi/spi-rspi.c 	if (tx)
tx                733 drivers/spi/spi-rspi.c static int qspi_trigger_transfer_out_in(struct rspi_data *rspi, const u8 *tx,
tx                748 drivers/spi/spi-rspi.c 			rspi_write_data(rspi, *tx++);
tx                781 drivers/spi/spi-rspi.c 	const u8 *tx = xfer->tx_buf;
tx                800 drivers/spi/spi-rspi.c 			rspi_write_data(rspi, *tx++);
tx                 70 drivers/spi/spi-s3c24xx.c 	const unsigned char	*tx;
tx                209 drivers/spi/spi-s3c24xx.c 	return hw->tx ? hw->tx[count] : 0;
tx                275 drivers/spi/spi-s3c24xx.c 	if (hw->tx && !hw->rx)
tx                277 drivers/spi/spi-s3c24xx.c 	else if (hw->rx && !hw->tx)
tx                284 drivers/spi/spi-s3c24xx.c 	regs.uregs[fiq_rtx]  = (long)hw->tx + 1;
tx                402 drivers/spi/spi-s3c24xx.c 	hw->tx = t->tx_buf;
tx                148 drivers/spi/spi-sh-hspi.c 	u32 tx;
tx                174 drivers/spi/spi-sh-hspi.c 			tx = 0;
tx                176 drivers/spi/spi-sh-hspi.c 				tx = (u32)((u8 *)t->tx_buf)[i];
tx                178 drivers/spi/spi-sh-hspi.c 			hspi_write(hspi, SPTBR, tx);
tx                743 drivers/spi/spi-sh-msiof.c static int sh_msiof_dma_once(struct sh_msiof_spi_priv *p, const void *tx,
tx                767 drivers/spi/spi-sh-msiof.c 	if (tx) {
tx                792 drivers/spi/spi-sh-msiof.c 	sh_msiof_spi_set_mode_regs(p, tx, rx, 32, len / 4);
tx                797 drivers/spi/spi-sh-msiof.c 	if (tx)
tx                804 drivers/spi/spi-sh-msiof.c 	if (tx)
tx                813 drivers/spi/spi-sh-msiof.c 	if (tx) {
tx                854 drivers/spi/spi-sh-msiof.c 	if (tx)
tx                260 drivers/spi/spi-sirf.c 	const void *tx;
tx                330 drivers/spi/spi-sirf.c 	const u8 *tx = sspi->tx;
tx                332 drivers/spi/spi-sirf.c 	if (tx) {
tx                333 drivers/spi/spi-sirf.c 		data = *tx++;
tx                334 drivers/spi/spi-sirf.c 		sspi->tx = tx;
tx                358 drivers/spi/spi-sirf.c 	const u16 *tx = sspi->tx;
tx                360 drivers/spi/spi-sirf.c 	if (tx) {
tx                361 drivers/spi/spi-sirf.c 		data = *tx++;
tx                362 drivers/spi/spi-sirf.c 		sspi->tx = tx;
tx                388 drivers/spi/spi-sirf.c 	const u32 *tx = sspi->tx;
tx                390 drivers/spi/spi-sirf.c 	if (tx) {
tx                391 drivers/spi/spi-sirf.c 		data = *tx++;
tx                392 drivers/spi/spi-sirf.c 		sspi->tx = tx;
tx                469 drivers/spi/spi-sirf.c 	memcpy(&cmd, sspi->tx, t->len);
tx                556 drivers/spi/spi-sirf.c 	sspi->src_start = dma_map_single(&spi->dev, (void *)sspi->tx, t->len,
tx                709 drivers/spi/spi-sirf.c 	sspi->tx = t->tx_buf;
tx                211 drivers/spi/spi-stm32.c 	const struct stm32_spi_reg tx;
tx                331 drivers/spi/spi-stm32.c 	.tx = { STM32F4_SPI_DR },
tx                350 drivers/spi/spi-stm32.c 	.tx = { STM32H7_SPI_TXDR },
tx               1141 drivers/spi/spi-stm32.c 		dma_conf->dst_addr = spi->phys_addr + spi->cfg->regs->tx.reg;
tx                 78 drivers/spi/spi-test.h 			void *tx, void *rx);
tx                 80 drivers/spi/spi-test.h 			   void *tx, void *rx);
tx                109 drivers/spi/spi-test.h 		      void *tx, void *rx);
tx                114 drivers/spi/spi-test.h 			 void *tx, void *rx);
tx                401 drivers/spi/spi-ti-qspi.c 	struct dma_async_tx_descriptor *tx;
tx                404 drivers/spi/spi-ti-qspi.c 	tx = dmaengine_prep_dma_memcpy(chan, dma_dst, dma_src, len, flags);
tx                405 drivers/spi/spi-ti-qspi.c 	if (!tx) {
tx                410 drivers/spi/spi-ti-qspi.c 	tx->callback = ti_qspi_dma_callback;
tx                411 drivers/spi/spi-ti-qspi.c 	tx->callback_param = qspi;
tx                412 drivers/spi/spi-ti-qspi.c 	cookie = tx->tx_submit(tx);
tx                171 drivers/staging/comedi/drivers/ni_usb6501.c 	u8 *tx = devpriv->usb_tx_buf;
tx                183 drivers/staging/comedi/drivers/ni_usb6501.c 		memcpy(tx, READ_PORT_REQUEST, request_size);
tx                184 drivers/staging/comedi/drivers/ni_usb6501.c 		tx[14] = val & 0xff;
tx                189 drivers/staging/comedi/drivers/ni_usb6501.c 		memcpy(tx, WRITE_PORT_REQUEST, request_size);
tx                190 drivers/staging/comedi/drivers/ni_usb6501.c 		tx[14] = val & 0xff;
tx                191 drivers/staging/comedi/drivers/ni_usb6501.c 		tx[17] = *bitmap;
tx                196 drivers/staging/comedi/drivers/ni_usb6501.c 		memcpy(tx, SET_PORT_DIR_REQUEST, request_size);
tx                197 drivers/staging/comedi/drivers/ni_usb6501.c 		tx[14] = val & 0xff;
tx                198 drivers/staging/comedi/drivers/ni_usb6501.c 		tx[15] = (val >> 8) & 0xff;
tx                199 drivers/staging/comedi/drivers/ni_usb6501.c 		tx[16] = (val >> 16) & 0xff;
tx                253 drivers/staging/comedi/drivers/ni_usb6501.c 	u8 *tx = devpriv->usb_tx_buf;
tx                265 drivers/staging/comedi/drivers/ni_usb6501.c 		memcpy(tx, START_COUNTER_REQUEST, request_size);
tx                270 drivers/staging/comedi/drivers/ni_usb6501.c 		memcpy(tx, STOP_COUNTER_REQUEST, request_size);
tx                275 drivers/staging/comedi/drivers/ni_usb6501.c 		memcpy(tx, READ_COUNTER_REQUEST, request_size);
tx                280 drivers/staging/comedi/drivers/ni_usb6501.c 		memcpy(tx, WRITE_COUNTER_REQUEST, request_size);
tx                283 drivers/staging/comedi/drivers/ni_usb6501.c 		*((__be32 *)&tx[12]) = cpu_to_be32(*val);
tx                683 drivers/staging/fwserial/fwserial.c 		port->icount.tx += txn->dma_pended.len;
tx               1390 drivers/staging/fwserial/fwserial.c 	icount->tx  = port->icount.tx + stats.xchars;
tx               1408 drivers/staging/fwserial/fwserial.c 		   port->icount.tx + stats.xchars, port->icount.rx);
tx                145 drivers/staging/gdm724x/gdm_usb.c static struct usb_tx_sdu *get_tx_sdu_struct(struct tx_cxt *tx, int *no_spc)
tx                149 drivers/staging/gdm724x/gdm_usb.c 	if (list_empty(&tx->free_list))
tx                152 drivers/staging/gdm724x/gdm_usb.c 	t_sdu = list_entry(tx->free_list.next, struct usb_tx_sdu, list);
tx                155 drivers/staging/gdm724x/gdm_usb.c 	tx->avail_count--;
tx                157 drivers/staging/gdm724x/gdm_usb.c 	*no_spc = list_empty(&tx->free_list) ? 1 : 0;
tx                162 drivers/staging/gdm724x/gdm_usb.c static void put_tx_struct(struct tx_cxt *tx, struct usb_tx_sdu *t_sdu)
tx                164 drivers/staging/gdm724x/gdm_usb.c 	list_add_tail(&t_sdu->list, &tx->free_list);
tx                165 drivers/staging/gdm724x/gdm_usb.c 	tx->avail_count++;
tx                247 drivers/staging/gdm724x/gdm_usb.c 	struct tx_cxt	*tx = &udev->tx;
tx                253 drivers/staging/gdm724x/gdm_usb.c 	spin_lock_irqsave(&tx->lock, flags);
tx                254 drivers/staging/gdm724x/gdm_usb.c 	list_for_each_entry_safe(t_sdu, t_sdu_next, &tx->sdu_list, list) {
tx                259 drivers/staging/gdm724x/gdm_usb.c 	list_for_each_entry_safe(t, t_next, &tx->hci_list, list) {
tx                264 drivers/staging/gdm724x/gdm_usb.c 	list_for_each_entry_safe(t_sdu, t_sdu_next, &tx->free_list, list) {
tx                268 drivers/staging/gdm724x/gdm_usb.c 	spin_unlock_irqrestore(&tx->lock, flags);
tx                300 drivers/staging/gdm724x/gdm_usb.c 	struct tx_cxt *tx = &udev->tx;
tx                310 drivers/staging/gdm724x/gdm_usb.c 	INIT_LIST_HEAD(&tx->sdu_list);
tx                311 drivers/staging/gdm724x/gdm_usb.c 	INIT_LIST_HEAD(&tx->hci_list);
tx                312 drivers/staging/gdm724x/gdm_usb.c 	INIT_LIST_HEAD(&tx->free_list);
tx                316 drivers/staging/gdm724x/gdm_usb.c 	spin_lock_init(&tx->lock);
tx                321 drivers/staging/gdm724x/gdm_usb.c 	tx->avail_count = 0;
tx                333 drivers/staging/gdm724x/gdm_usb.c 		list_add(&t_sdu->list, &tx->free_list);
tx                334 drivers/staging/gdm724x/gdm_usb.c 		tx->avail_count++;
tx                542 drivers/staging/gdm724x/gdm_usb.c 	struct tx_cxt *tx = t->tx;
tx                543 drivers/staging/gdm724x/gdm_usb.c 	struct lte_udev *udev = container_of(tx, struct lte_udev, tx);
tx                556 drivers/staging/gdm724x/gdm_usb.c 	spin_lock_irqsave(&tx->lock, flags);
tx                559 drivers/staging/gdm724x/gdm_usb.c 	spin_unlock_irqrestore(&tx->lock, flags);
tx                590 drivers/staging/gdm724x/gdm_usb.c 	struct tx_cxt *tx = &udev->tx;
tx                600 drivers/staging/gdm724x/gdm_usb.c 		spin_lock_irqsave(&tx->lock, flags);
tx                601 drivers/staging/gdm724x/gdm_usb.c 		if (list_empty(&tx->sdu_list)) {
tx                602 drivers/staging/gdm724x/gdm_usb.c 			spin_unlock_irqrestore(&tx->lock, flags);
tx                606 drivers/staging/gdm724x/gdm_usb.c 		t_sdu = list_entry(tx->sdu_list.next, struct usb_tx_sdu, list);
tx                608 drivers/staging/gdm724x/gdm_usb.c 			spin_unlock_irqrestore(&tx->lock, flags);
tx                613 drivers/staging/gdm724x/gdm_usb.c 		spin_unlock_irqrestore(&tx->lock, flags);
tx                620 drivers/staging/gdm724x/gdm_usb.c 		if (tx->avail_count > 10)
tx                623 drivers/staging/gdm724x/gdm_usb.c 		spin_lock_irqsave(&tx->lock, flags);
tx                624 drivers/staging/gdm724x/gdm_usb.c 		put_tx_struct(tx, t_sdu);
tx                625 drivers/staging/gdm724x/gdm_usb.c 		spin_unlock_irqrestore(&tx->lock, flags);
tx                639 drivers/staging/gdm724x/gdm_usb.c 	struct tx_cxt *tx = &udev->tx;
tx                651 drivers/staging/gdm724x/gdm_usb.c 	spin_lock_irqsave(&tx->lock, flags);
tx                653 drivers/staging/gdm724x/gdm_usb.c 		spin_unlock_irqrestore(&tx->lock, flags);
tx                658 drivers/staging/gdm724x/gdm_usb.c 	if (!list_empty(&tx->hci_list)) {
tx                659 drivers/staging/gdm724x/gdm_usb.c 		t = list_entry(tx->hci_list.next, struct usb_tx, list);
tx                664 drivers/staging/gdm724x/gdm_usb.c 	} else if (!list_empty(&tx->sdu_list)) {
tx                667 drivers/staging/gdm724x/gdm_usb.c 			spin_unlock_irqrestore(&tx->lock, flags);
tx                673 drivers/staging/gdm724x/gdm_usb.c 			spin_unlock_irqrestore(&tx->lock, flags);
tx                677 drivers/staging/gdm724x/gdm_usb.c 		t->tx = tx;
tx                684 drivers/staging/gdm724x/gdm_usb.c 		spin_unlock_irqrestore(&tx->lock, flags);
tx                687 drivers/staging/gdm724x/gdm_usb.c 	spin_unlock_irqrestore(&tx->lock, flags);
tx                706 drivers/staging/gdm724x/gdm_usb.c 	struct tx_cxt *tx = &udev->tx;
tx                718 drivers/staging/gdm724x/gdm_usb.c 	spin_lock_irqsave(&tx->lock, flags);
tx                719 drivers/staging/gdm724x/gdm_usb.c 	t_sdu = get_tx_sdu_struct(tx, &no_spc);
tx                720 drivers/staging/gdm724x/gdm_usb.c 	spin_unlock_irqrestore(&tx->lock, flags);
tx                747 drivers/staging/gdm724x/gdm_usb.c 	spin_lock_irqsave(&tx->lock, flags);
tx                748 drivers/staging/gdm724x/gdm_usb.c 	list_add_tail(&t_sdu->list, &tx->sdu_list);
tx                750 drivers/staging/gdm724x/gdm_usb.c 	spin_unlock_irqrestore(&tx->lock, flags);
tx                762 drivers/staging/gdm724x/gdm_usb.c 	struct tx_cxt *tx = &udev->tx;
tx                781 drivers/staging/gdm724x/gdm_usb.c 	t->tx = tx;
tx                784 drivers/staging/gdm724x/gdm_usb.c 	spin_lock_irqsave(&tx->lock, flags);
tx                785 drivers/staging/gdm724x/gdm_usb.c 	list_add_tail(&t->list, &tx->hci_list);
tx                787 drivers/staging/gdm724x/gdm_usb.c 	spin_unlock_irqrestore(&tx->lock, flags);
tx                939 drivers/staging/gdm724x/gdm_usb.c 	struct tx_cxt *tx;
tx                967 drivers/staging/gdm724x/gdm_usb.c 	tx = &udev->tx;
tx                968 drivers/staging/gdm724x/gdm_usb.c 	spin_lock_irqsave(&tx->lock, flags);
tx                970 drivers/staging/gdm724x/gdm_usb.c 	spin_unlock_irqrestore(&tx->lock, flags);
tx                 42 drivers/staging/gdm724x/gdm_usb.h 	struct tx_cxt *tx;
tx                 86 drivers/staging/gdm724x/gdm_usb.h 	struct tx_cxt tx;
tx                 28 drivers/staging/iio/meter/ade7854-i2c.c 	st->tx[0] = (reg_address >> 8) & 0xFF;
tx                 29 drivers/staging/iio/meter/ade7854-i2c.c 	st->tx[1] = reg_address & 0xFF;
tx                 33 drivers/staging/iio/meter/ade7854-i2c.c 		st->tx[2] = val & 0xFF;
tx                 37 drivers/staging/iio/meter/ade7854-i2c.c 		st->tx[2] = (val >> 8) & 0xFF;
tx                 38 drivers/staging/iio/meter/ade7854-i2c.c 		st->tx[3] = val & 0xFF;
tx                 42 drivers/staging/iio/meter/ade7854-i2c.c 		st->tx[2] = (val >> 16) & 0xFF;
tx                 43 drivers/staging/iio/meter/ade7854-i2c.c 		st->tx[3] = (val >> 8) & 0xFF;
tx                 44 drivers/staging/iio/meter/ade7854-i2c.c 		st->tx[4] = val & 0xFF;
tx                 48 drivers/staging/iio/meter/ade7854-i2c.c 		st->tx[2] = (val >> 24) & 0xFF;
tx                 49 drivers/staging/iio/meter/ade7854-i2c.c 		st->tx[3] = (val >> 16) & 0xFF;
tx                 50 drivers/staging/iio/meter/ade7854-i2c.c 		st->tx[4] = (val >> 8) & 0xFF;
tx                 51 drivers/staging/iio/meter/ade7854-i2c.c 		st->tx[5] = val & 0xFF;
tx                 59 drivers/staging/iio/meter/ade7854-i2c.c 	ret = i2c_master_send(st->i2c, st->tx, count);
tx                 77 drivers/staging/iio/meter/ade7854-i2c.c 	st->tx[0] = (reg_address >> 8) & 0xFF;
tx                 78 drivers/staging/iio/meter/ade7854-i2c.c 	st->tx[1] = reg_address & 0xFF;
tx                 80 drivers/staging/iio/meter/ade7854-i2c.c 	ret = i2c_master_send(st->i2c, st->tx, 2);
tx                 26 drivers/staging/iio/meter/ade7854-spi.c 		.tx_buf = st->tx,
tx                 32 drivers/staging/iio/meter/ade7854-spi.c 	st->tx[0] = ADE7854_WRITE_REG;
tx                 33 drivers/staging/iio/meter/ade7854-spi.c 	st->tx[1] = (reg_address >> 8) & 0xFF;
tx                 34 drivers/staging/iio/meter/ade7854-spi.c 	st->tx[2] = reg_address & 0xFF;
tx                 37 drivers/staging/iio/meter/ade7854-spi.c 		st->tx[3] = val & 0xFF;
tx                 41 drivers/staging/iio/meter/ade7854-spi.c 		st->tx[3] = (val >> 8) & 0xFF;
tx                 42 drivers/staging/iio/meter/ade7854-spi.c 		st->tx[4] = val & 0xFF;
tx                 46 drivers/staging/iio/meter/ade7854-spi.c 		st->tx[3] = (val >> 16) & 0xFF;
tx                 47 drivers/staging/iio/meter/ade7854-spi.c 		st->tx[4] = (val >> 8) & 0xFF;
tx                 48 drivers/staging/iio/meter/ade7854-spi.c 		st->tx[5] = val & 0xFF;
tx                 52 drivers/staging/iio/meter/ade7854-spi.c 		st->tx[3] = (val >> 24) & 0xFF;
tx                 53 drivers/staging/iio/meter/ade7854-spi.c 		st->tx[4] = (val >> 16) & 0xFF;
tx                 54 drivers/staging/iio/meter/ade7854-spi.c 		st->tx[5] = (val >> 8) & 0xFF;
tx                 55 drivers/staging/iio/meter/ade7854-spi.c 		st->tx[6] = val & 0xFF;
tx                 79 drivers/staging/iio/meter/ade7854-spi.c 			.tx_buf = st->tx,
tx                 91 drivers/staging/iio/meter/ade7854-spi.c 	st->tx[0] = ADE7854_READ_REG;
tx                 92 drivers/staging/iio/meter/ade7854-spi.c 	st->tx[1] = (reg_address >> 8) & 0xFF;
tx                 93 drivers/staging/iio/meter/ade7854-spi.c 	st->tx[2] = reg_address & 0xFF;
tx                165 drivers/staging/iio/meter/ade7854.h 	u8 tx[ADE7854_MAX_TX] ____cacheline_aligned;
tx                 98 drivers/staging/iio/resolver/ad2s1210.c 	u8 tx[2] ____cacheline_aligned;
tx                121 drivers/staging/iio/resolver/ad2s1210.c 	st->tx[0] = data;
tx                122 drivers/staging/iio/resolver/ad2s1210.c 	ret = spi_write(st->sdev, st->tx, 1);
tx                137 drivers/staging/iio/resolver/ad2s1210.c 			.tx_buf = &st->tx[0],
tx                142 drivers/staging/iio/resolver/ad2s1210.c 			.tx_buf = &st->tx[1],
tx                148 drivers/staging/iio/resolver/ad2s1210.c 	st->tx[0] = address | AD2S1210_MSB_IS_HIGH;
tx                149 drivers/staging/iio/resolver/ad2s1210.c 	st->tx[1] = AD2S1210_REG_FAULT;
tx                138 drivers/staging/kpc2000/kpc2000_spi.c 		unsigned int tx    :  1; /* Tx Status       */
tx                215 drivers/staging/kpc2000/kpc2000_spi.c 	const u8 *tx = transfer->tx_buf;
tx                218 drivers/staging/kpc2000/kpc2000_spi.c 	if (tx) {
tx                220 drivers/staging/kpc2000/kpc2000_spi.c 			char val = *tx++;
tx                105 drivers/staging/media/soc_camera/imx074.c 	unsigned char tx[3];
tx                109 drivers/staging/media/soc_camera/imx074.c 	msg.buf = tx;
tx                113 drivers/staging/media/soc_camera/imx074.c 	tx[0] = addr >> 8;
tx                114 drivers/staging/media/soc_camera/imx074.c 	tx[1] = addr & 0xff;
tx                115 drivers/staging/media/soc_camera/imx074.c 	tx[2] = data;
tx                 66 drivers/staging/most/net/net.c 	struct net_dev_channel tx;
tx                199 drivers/staging/most/net/net.c 	if (most_start_channel(nd->iface, nd->tx.ch_id, &comp)) {
tx                213 drivers/staging/most/net/net.c 		nd->iface->request_netinfo(nd->iface, nd->tx.ch_id, on_netinfo);
tx                226 drivers/staging/most/net/net.c 		nd->iface->request_netinfo(nd->iface, nd->tx.ch_id, NULL);
tx                228 drivers/staging/most/net/net.c 	most_stop_channel(nd->iface, nd->tx.ch_id, &comp);
tx                240 drivers/staging/most/net/net.c 	mbo = most_get_mbo(nd->iface, nd->tx.ch_id, &comp);
tx                297 drivers/staging/most/net/net.c 	if (nd && nd->rx.linked && nd->tx.linked)
tx                339 drivers/staging/most/net/net.c 		ch = ccfg->direction == MOST_CH_TX ? &nd->tx : &nd->rx;
tx                341 drivers/staging/most/net/net.c 		ch = ccfg->direction == MOST_CH_TX ? &nd->tx : &nd->rx;
tx                379 drivers/staging/most/net/net.c 	} else if (nd->tx.linked && channel_idx == nd->tx.ch_id) {
tx                380 drivers/staging/most/net/net.c 		ch = &nd->tx;
tx                386 drivers/staging/most/net/net.c 	if (nd->rx.linked && nd->tx.linked) {
tx                418 drivers/staging/most/net/net.c 	if (nd->tx.ch_id != channel_idx)
tx                473 drivers/staging/nvec/nvec.c 	if (nvec->tx->pos != nvec->tx->size) {
tx                475 drivers/staging/nvec/nvec.c 		nvec->tx->pos = 0;
tx                554 drivers/staging/nvec/nvec.c 		nvec->tx = &nvec->tx_scratch;
tx                555 drivers/staging/nvec/nvec.c 		list_add_tail(&nvec->tx->node, &nvec->tx_data);
tx                557 drivers/staging/nvec/nvec.c 		nvec->tx = list_first_entry(&nvec->tx_data, struct nvec_msg,
tx                559 drivers/staging/nvec/nvec.c 		nvec->tx->pos = 0;
tx                564 drivers/staging/nvec/nvec.c 		(uint)nvec->tx->size, nvec->tx->data[1]);
tx                639 drivers/staging/nvec/nvec.c 			to_send = nvec->tx->data[0];
tx                640 drivers/staging/nvec/nvec.c 			nvec->tx->pos = 1;
tx                654 drivers/staging/nvec/nvec.c 		} else if (nvec->tx && nvec->tx->pos < nvec->tx->size) {
tx                655 drivers/staging/nvec/nvec.c 			to_send = nvec->tx->data[nvec->tx->pos++];
tx                659 drivers/staging/nvec/nvec.c 				nvec->tx,
tx                660 drivers/staging/nvec/nvec.c 				(uint)(nvec->tx ? nvec->tx->pos : 0),
tx                661 drivers/staging/nvec/nvec.c 				(uint)(nvec->tx ? nvec->tx->size : 0));
tx                149 drivers/staging/nvec/nvec.h 	struct nvec_msg *tx;
tx                671 drivers/staging/ralink-gdma/ralink-gdma.c 	if (desc && (cookie == desc->vdesc.tx.cookie)) {
tx               1201 drivers/staging/rtl8188eu/core/rtw_ap.c void update_beacon(struct adapter *padapter, u8 ie_id, u8 *oui, u8 tx)
tx               1235 drivers/staging/rtl8188eu/core/rtw_ap.c 	if (tx)
tx                 23 drivers/staging/rtl8188eu/include/rtw_ap.h 		   u8 *oui, u8 tx);
tx               1748 drivers/staging/rtl8723bs/core/rtw_ap.c void update_beacon(struct adapter *padapter, u8 ie_id, u8 *oui, u8 tx)
tx               1820 drivers/staging/rtl8723bs/core/rtw_ap.c 	if (tx) {
tx                205 drivers/staging/rtl8723bs/core/rtw_pwrctrl.c void traffic_check_for_leave_lps(struct adapter *padapter, u8 tx, u32 tx_packets)
tx                214 drivers/staging/rtl8723bs/core/rtw_pwrctrl.c 	if (tx) { /* from tx */
tx                248 drivers/staging/rtl8723bs/core/rtw_pwrctrl.c 		rtw_lps_ctrl_wk_cmd(padapter, LPS_CTRL_LEAVE, tx?0:1);
tx                 13 drivers/staging/rtl8723bs/include/rtw_ap.h void update_beacon(struct adapter *padapter, u8 ie_id, u8 *oui, u8 tx);
tx                225 drivers/staging/rtl8723bs/include/rtw_mp.h 	struct mp_tx tx;
tx                348 drivers/staging/rtl8723bs/include/rtw_pwrctrl.h void traffic_check_for_leave_lps(struct adapter *padapter, u8 tx, u32 tx_packets);
tx                174 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c 		.tx = 0xffff,
tx                178 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c 		.tx = 0xffff,
tx                183 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c 		.tx = 0xffff,
tx                194 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c 		.tx = 0xffff,
tx                204 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c 		.tx = 0xffff,
tx                209 drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c 		.tx = 0xffff,
tx               1621 drivers/staging/vt6655/device_main.c 	.tx			= vnt_tx_80211,
tx                936 drivers/staging/vt6656/main_usb.c 	.tx			= vnt_tx_80211,
tx                588 drivers/staging/vt6656/rxtx.c 	union vnt_tx_data_head *head = &tx_head->tx_rts.tx.head;
tx                607 drivers/staging/vt6656/rxtx.c 		head = &tx_head->tx_rts.tx.mic.head;
tx                620 drivers/staging/vt6656/rxtx.c 	union vnt_tx_data_head *head = &tx_head->tx_cts.tx.head;
tx                634 drivers/staging/vt6656/rxtx.c 		head = &tx_head->tx_cts.tx.mic.head;
tx                648 drivers/staging/vt6656/rxtx.c 	union vnt_tx_data_head *head = &tx_head->tx_ab.tx.head;
tx                657 drivers/staging/vt6656/rxtx.c 		head = &tx_head->tx_ab.tx.mic.head;
tx                692 drivers/staging/vt6656/rxtx.c 					&tx_buffer->tx_head.tx_rts.tx.mic.hdr;
tx                699 drivers/staging/vt6656/rxtx.c 			*mic_hdr = &tx_buffer->tx_head.tx_cts.tx.mic.hdr;
tx                705 drivers/staging/vt6656/rxtx.c 		*mic_hdr = &tx_buffer->tx_head.tx_ab.tx.mic.hdr;
tx                198 drivers/staging/vt6656/rxtx.h 		union vnt_tx tx;
tx                202 drivers/staging/vt6656/rxtx.h 		union vnt_tx tx;
tx                206 drivers/staging/vt6656/rxtx.h 		union vnt_tx tx;
tx                 35 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c 		.tx = 0xffff,
tx                 40 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c 		.tx = 0xffff,
tx                 50 drivers/staging/wilc1000/wilc_wfi_cfgoperations.c 		.tx = 0xffff,
tx                 85 drivers/target/iscsi/cxgbit/cxgbit.h 			struct cxgbit_skb_tx_cb tx;
tx                 99 drivers/target/iscsi/cxgbit/cxgbit.h #define cxgbit_skcb_submode(skb)	(CXGBIT_SKB_CB(skb)->tx.submode)
tx                101 drivers/target/iscsi/cxgbit/cxgbit.h #define cxgbit_skcb_tx_extralen(skb)	(CXGBIT_SKB_CB(skb)->tx.extra_len)
tx                 47 drivers/thermal/tegra/tegra-bpmp-thermal.c 	msg.tx.data = &req;
tx                 48 drivers/thermal/tegra/tegra-bpmp-thermal.c 	msg.tx.size = sizeof(req);
tx                 76 drivers/thermal/tegra/tegra-bpmp-thermal.c 	msg.tx.data = &req;
tx                 77 drivers/thermal/tegra/tegra-bpmp-thermal.c 	msg.tx.size = sizeof(req);
tx                135 drivers/thermal/tegra/tegra-bpmp-thermal.c 	msg.tx.data = &req;
tx                136 drivers/thermal/tegra/tegra-bpmp-thermal.c 	msg.tx.size = sizeof(req);
tx                 27 drivers/thunderbolt/ctl.c 	struct tb_ring *tx;
tx                364 drivers/thunderbolt/ctl.c 	res = tb_ring_tx(ctl->tx, &pkg->frame);
tx                619 drivers/thunderbolt/ctl.c 	ctl->tx = tb_ring_alloc_tx(nhi, 0, 10, RING_FLAG_NO_SUSPEND);
tx                620 drivers/thunderbolt/ctl.c 	if (!ctl->tx)
tx                658 drivers/thunderbolt/ctl.c 	if (ctl->tx)
tx                659 drivers/thunderbolt/ctl.c 		tb_ring_free(ctl->tx);
tx                677 drivers/thunderbolt/ctl.c 	tb_ring_start(ctl->tx); /* is used to ack hotplug packets, start first */
tx                700 drivers/thunderbolt/ctl.c 	tb_ring_stop(ctl->tx);
tx                345 drivers/tty/amiserial.c 		info->icount.tx++;
tx                361 drivers/tty/amiserial.c 	info->icount.tx++;
tx               1195 drivers/tty/amiserial.c 	icount->tx = cnow.tx;
tx               1516 drivers/tty/amiserial.c 	seq_printf(m, " tx:%d rx:%d", state->icount.tx, state->icount.rx);
tx               1649 drivers/tty/amiserial.c 	state->icount.rx = state->icount.tx = 0;
tx                601 drivers/tty/cyclades.c 		info->icount.tx++;
tx                656 drivers/tty/cyclades.c 			info->icount.tx++;
tx                664 drivers/tty/cyclades.c 				info->icount.tx++;
tx               1043 drivers/tty/cyclades.c 		info->icount.tx++;
tx               1058 drivers/tty/cyclades.c 		info->icount.tx += small_count;
tx               1073 drivers/tty/cyclades.c 		info->icount.tx++;
tx               2752 drivers/tty/cyclades.c 	sic->tx = cnow.tx;
tx               1105 drivers/tty/ipwireless/hardware.c 	int tx = 0;
tx               1183 drivers/tty/ipwireless/hardware.c 		tx = 1;
tx               1185 drivers/tty/ipwireless/hardware.c 	if (tx)
tx               1191 drivers/tty/ipwireless/hardware.c 	if (tx || rx)
tx                176 drivers/tty/isicom.c static DEFINE_TIMER(tx, isicom_tx);
tx                500 drivers/tty/isicom.c 	mod_timer(&tx, jiffies + msecs_to_jiffies(10));
tx               1669 drivers/tty/isicom.c 	mod_timer(&tx, jiffies + 1);
tx               1682 drivers/tty/isicom.c 	del_timer_sync(&tx);
tx               1840 drivers/tty/mxser.c 	icount->tx = cnow.tx;
tx               2169 drivers/tty/mxser.c 		port->icount.tx++;
tx               2198 drivers/tty/mxser.c 	port->icount.tx += (cnt - port->xmit_cnt);
tx               1738 drivers/tty/nozomi.c 	icount->tx = cnow.tx;
tx                127 drivers/tty/serial/21285.c 		port->icount.tx++;
tx                139 drivers/tty/serial/21285.c 		port->icount.tx++;
tx                 31 drivers/tty/serial/8250/8250_dma.c 	p->port.icount.tx += dma->tx_size;
tx                914 drivers/tty/serial/8250/8250_omap.c 	p->port.icount.tx += dma->tx_size;
tx               1731 drivers/tty/serial/8250/8250_port.c 		port->icount.tx++;
tx               1748 drivers/tty/serial/8250/8250_port.c 		port->icount.tx++;
tx                149 drivers/tty/serial/altera_jtaguart.c 		port->icount.tx++;
tx                166 drivers/tty/serial/altera_jtaguart.c 				port->icount.tx++;
tx                260 drivers/tty/serial/altera_uart.c 		port->icount.tx++;
tx                271 drivers/tty/serial/altera_uart.c 		port->icount.tx++;
tx                178 drivers/tty/serial/amba-pl010.c 		uap->port.icount.tx++;
tx                191 drivers/tty/serial/amba-pl010.c 		uap->port.icount.tx++;
tx                684 drivers/tty/serial/amba-pl011.c 	uap->port.icount.tx += count;
tx                792 drivers/tty/serial/amba-pl011.c 	uap->port.icount.tx++;
tx               1386 drivers/tty/serial/amba-pl011.c 	uap->port.icount.tx++;
tx                136 drivers/tty/serial/apbuart.c 		port->icount.tx++;
tx                151 drivers/tty/serial/apbuart.c 		port->icount.tx++;
tx                354 drivers/tty/serial/ar933x_uart.c 			up->port.icount.tx++;
tx                365 drivers/tty/serial/ar933x_uart.c 		up->port.icount.tx++;
tx                168 drivers/tty/serial/arc_uart.c 		port->icount.tx++;
tx                174 drivers/tty/serial/arc_uart.c 		port->icount.tx++;
tx                837 drivers/tty/serial/atmel_serial.c 		port->icount.tx++;
tx                847 drivers/tty/serial/atmel_serial.c 		port->icount.tx++;
tx                876 drivers/tty/serial/atmel_serial.c 	port->icount.tx += atmel_port->tx_len;
tx               1465 drivers/tty/serial/atmel_serial.c 	port->icount.tx += pdc->ofs;
tx                317 drivers/tty/serial/bcm63xx_uart.c 		port->icount.tx++;
tx                341 drivers/tty/serial/bcm63xx_uart.c 		port->icount.tx++;
tx                156 drivers/tty/serial/clps711x.c 		port->icount.tx++;
tx                174 drivers/tty/serial/clps711x.c 		port->icount.tx++;
tx                704 drivers/tty/serial/cpm_uart/cpm_uart_core.c 		port->icount.tx++;
tx                724 drivers/tty/serial/cpm_uart/cpm_uart_core.c 			port->icount.tx++;
tx                193 drivers/tty/serial/digicolor-usart.c 		port->icount.tx++;
tx                206 drivers/tty/serial/digicolor-usart.c 		port->icount.tx++;
tx                269 drivers/tty/serial/dz.c 		dport->port.icount.tx++;
tx                288 drivers/tty/serial/dz.c 	dport->port.icount.tx++;
tx                144 drivers/tty/serial/efm32-uart.c 			port->icount.tx++;
tx                151 drivers/tty/serial/efm32-uart.c 			port->icount.tx++;
tx                182 drivers/tty/serial/fsl_linflexuart.c 		sport->icount.tx++;
tx                474 drivers/tty/serial/fsl_lpuart.c 	sport->port.icount.tx += sport->dma_tx_bytes;
tx                684 drivers/tty/serial/fsl_lpuart.c 		sport->port.icount.tx++;
tx                698 drivers/tty/serial/fsl_lpuart.c 		sport->port.icount.tx++;
tx                715 drivers/tty/serial/fsl_lpuart.c 		sport->port.icount.tx++;
tx                731 drivers/tty/serial/fsl_lpuart.c 		sport->port.icount.tx++;
tx                707 drivers/tty/serial/icom.c 		icom_port->uart_port.icount.tx += count;
tx                503 drivers/tty/serial/imx.c 		sport->port.icount.tx++;
tx                538 drivers/tty/serial/imx.c 		sport->port.icount.tx++;
tx                566 drivers/tty/serial/imx.c 	sport->port.icount.tx += sport->tx_bytes;
tx                398 drivers/tty/serial/ip22zilog.c 		up->port.icount.tx++;
tx                417 drivers/tty/serial/ip22zilog.c 	up->port.icount.tx++;
tx                605 drivers/tty/serial/ip22zilog.c 		port->icount.tx++;
tx                617 drivers/tty/serial/ip22zilog.c 		port->icount.tx++;
tx                237 drivers/tty/serial/lantiq.c 			port->icount.tx++;
tx                248 drivers/tty/serial/lantiq.c 		port->icount.tx++;
tx                290 drivers/tty/serial/lpc32xx_hs.c 		port->icount.tx++;
tx                304 drivers/tty/serial/lpc32xx_hs.c 		port->icount.tx++;
tx                191 drivers/tty/serial/max3100.c static int max3100_sr(struct max3100_port *s, u16 tx, u16 *rx)
tx                202 drivers/tty/serial/max3100.c 	etx = cpu_to_be16(tx);
tx                212 drivers/tty/serial/max3100.c 	dev_dbg(&s->spi->dev, "%04x - %04x\n", tx, *rx);
tx                260 drivers/tty/serial/max3100.c 	u16 tx, rx;
tx                287 drivers/tty/serial/max3100.c 			tx = 0xffff;
tx                289 drivers/tty/serial/max3100.c 				tx = s->port.x_char;
tx                290 drivers/tty/serial/max3100.c 				s->port.icount.tx++;
tx                294 drivers/tty/serial/max3100.c 				tx = xmit->buf[xmit->tail];
tx                297 drivers/tty/serial/max3100.c 				s->port.icount.tx++;
tx                299 drivers/tty/serial/max3100.c 			if (tx != 0xffff) {
tx                300 drivers/tty/serial/max3100.c 				max3100_calc_parity(s, &tx);
tx                301 drivers/tty/serial/max3100.c 				tx |= MAX3100_WD | (s->rts ? MAX3100_RTS : 0);
tx                302 drivers/tty/serial/max3100.c 				max3100_sr(s, tx, &rx);
tx                571 drivers/tty/serial/max3100.c 		u16 tx, rx;
tx                573 drivers/tty/serial/max3100.c 		tx = MAX3100_WC | MAX3100_SHDN;
tx                574 drivers/tty/serial/max3100.c 		max3100_sr(s, tx, &rx);
tx                616 drivers/tty/serial/max3100.c 		u16 tx, rx;
tx                617 drivers/tty/serial/max3100.c 		tx = 0x4001;
tx                618 drivers/tty/serial/max3100.c 		max3100_sr(s, tx, &rx);
tx                741 drivers/tty/serial/max3100.c 	u16 tx, rx;
tx                804 drivers/tty/serial/max3100.c 		tx = MAX3100_WC | MAX3100_SHDN;
tx                805 drivers/tty/serial/max3100.c 		max3100_sr(max3100s[i], tx, &rx);
tx                860 drivers/tty/serial/max3100.c 		u16 tx, rx;
tx                862 drivers/tty/serial/max3100.c 		tx = MAX3100_WC | MAX3100_SHDN;
tx                863 drivers/tty/serial/max3100.c 		max3100_sr(s, tx, &rx);
tx                755 drivers/tty/serial/max310x.c 		port->icount.tx++;
tx                782 drivers/tty/serial/max310x.c 		port->icount.tx += to_send;
tx                338 drivers/tty/serial/mcf.c 		port->icount.tx++;
tx                347 drivers/tty/serial/mcf.c 		port->icount.tx++;
tx                359 drivers/tty/serial/men_z135_uart.c 	port->icount.tx += n;
tx                151 drivers/tty/serial/meson_uart.c 			port->icount.tx++;
tx                162 drivers/tty/serial/meson_uart.c 		port->icount.tx++;
tx                 90 drivers/tty/serial/milbeaut_usio.c 		port->icount.tx++;
tx                106 drivers/tty/serial/milbeaut_usio.c 		port->icount.tx++;
tx               1447 drivers/tty/serial/mpc52xx_uart.c 		port->icount.tx++;
tx               1462 drivers/tty/serial/mpc52xx_uart.c 		port->icount.tx++;
tx                138 drivers/tty/serial/mps2-uart.c 			port->icount.tx++;
tx                147 drivers/tty/serial/mps2-uart.c 		port->icount.tx++;
tx                320 drivers/tty/serial/mps2-uart.c 				  MAKE_NAME(-tx), mps_port);
tx                458 drivers/tty/serial/msm_serial.c 	port->icount.tx += count;
tx                845 drivers/tty/serial/msm_serial.c 			port->icount.tx++;
tx                883 drivers/tty/serial/msm_serial.c 		port->icount.tx++;
tx                192 drivers/tty/serial/mux.c 		port->icount.tx++;
tx                206 drivers/tty/serial/mux.c 		port->icount.tx++;
tx                200 drivers/tty/serial/mvebu-uart.c 		port->icount.tx++;
tx                306 drivers/tty/serial/mvebu-uart.c 		port->icount.tx++;
tx                319 drivers/tty/serial/mvebu-uart.c 		port->icount.tx++;
tx                623 drivers/tty/serial/mxs-auart.c 			s->port.icount.tx++;
tx                629 drivers/tty/serial/mxs-auart.c 			s->port.icount.tx++;
tx                374 drivers/tty/serial/omap-serial.c 		up->port.icount.tx++;
tx                386 drivers/tty/serial/omap-serial.c 		up->port.icount.tx++;
tx                190 drivers/tty/serial/owl-uart.c 		port->icount.tx++;
tx                201 drivers/tty/serial/owl-uart.c 		port->icount.tx++;
tx                787 drivers/tty/serial/pch_uart.c 		port->icount.tx += sg_dma_len(sg);
tx                905 drivers/tty/serial/pch_uart.c 		port->icount.tx++;
tx                915 drivers/tty/serial/pch_uart.c 		port->icount.tx += tx_size;
tx                962 drivers/tty/serial/pch_uart.c 		port->icount.tx++;
tx                285 drivers/tty/serial/pic32_uart.c 		port->icount.tx++;
tx                314 drivers/tty/serial/pic32_uart.c 		port->icount.tx++;
tx                404 drivers/tty/serial/pmac_zilog.c 		uap->port.icount.tx++;
tx                424 drivers/tty/serial/pmac_zilog.c 	uap->port.icount.tx++;
tx                636 drivers/tty/serial/pmac_zilog.c 		port->icount.tx++;
tx                646 drivers/tty/serial/pmac_zilog.c 		port->icount.tx++;
tx                252 drivers/tty/serial/pnx8xxx_uart.c 		sport->port.icount.tx++;
tx                275 drivers/tty/serial/pnx8xxx_uart.c 		sport->port.icount.tx++;
tx                185 drivers/tty/serial/pxa.c 		up->port.icount.tx++;
tx                198 drivers/tty/serial/pxa.c 		up->port.icount.tx++;
tx                740 drivers/tty/serial/qcom_geni_serial.c 		uport->icount.tx += tx_bytes;
tx                344 drivers/tty/serial/rda-uart.c 		port->icount.tx++;
tx                355 drivers/tty/serial/rda-uart.c 		port->icount.tx++;
tx                446 drivers/tty/serial/rp2.c 			up->port.icount.tx++;
tx                455 drivers/tty/serial/rp2.c 		up->port.icount.tx++;
tx                243 drivers/tty/serial/sa1100.c 		sport->port.icount.tx++;
tx                266 drivers/tty/serial/sa1100.c 		sport->port.icount.tx++;
tx                182 drivers/tty/serial/samsung.c 		port->icount.tx += count;
tx                217 drivers/tty/serial/samsung.c 	port->icount.tx += count;
tx                728 drivers/tty/serial/samsung.c 		port->icount.tx++;
tx                755 drivers/tty/serial/samsung.c 		port->icount.tx++;
tx                396 drivers/tty/serial/sb1250-duart.c 		sport->port.icount.tx++;
tx                408 drivers/tty/serial/sb1250-duart.c 		sport->port.icount.tx++;
tx                641 drivers/tty/serial/sc16is7xx.c 		port->icount.tx++;
tx                663 drivers/tty/serial/sc16is7xx.c 		port->icount.tx += to_send;
tx                452 drivers/tty/serial/sccnxp.c 		port->icount.tx++;
tx                476 drivers/tty/serial/sccnxp.c 		port->icount.tx++;
tx                494 drivers/tty/serial/serial-tegra.c 		tup->uport.icount.tx++;
tx               1273 drivers/tty/serial/serial_core.c 	icount->tx          = cnow.tx;
tx               1857 drivers/tty/serial/serial_core.c 				uport->icount.tx, uport->icount.rx);
tx                350 drivers/tty/serial/serial_txx9.c 		up->port.icount.tx++;
tx                363 drivers/tty/serial/serial_txx9.c 		up->port.icount.tx++;
tx                834 drivers/tty/serial/sh-sci.c 		port->icount.tx++;
tx               1195 drivers/tty/serial/sh-sci.c 	port->icount.tx += s->tx_dma_len;
tx                307 drivers/tty/serial/sifive.c 		ssp->port.icount.tx++;
tx                319 drivers/tty/serial/sifive.c 		ssp->port.icount.tx++;
tx                435 drivers/tty/serial/sirfsoc_uart.c 		port->icount.tx++;
tx                453 drivers/tty/serial/sirfsoc_uart.c 	port->icount.tx += sirfport->transfer_size;
tx                223 drivers/tty/serial/sprd_serial.c 		port->icount.tx += trans_len;
tx                261 drivers/tty/serial/sprd_serial.c 	port->icount.tx += sp->tx_dma.trans_len;
tx                333 drivers/tty/serial/sprd_serial.c 		port->icount.tx++;
tx                638 drivers/tty/serial/sprd_serial.c 		port->icount.tx++;
tx                652 drivers/tty/serial/sprd_serial.c 		port->icount.tx++;
tx                255 drivers/tty/serial/st-asc.c 		port->icount.tx++;
tx                280 drivers/tty/serial/st-asc.c 		port->icount.tx++;
tx                342 drivers/tty/serial/stm32-usart.c 		port->icount.tx++;
tx                410 drivers/tty/serial/stm32-usart.c 	port->icount.tx += count;
tx                424 drivers/tty/serial/stm32-usart.c 		port->icount.tx++;
tx                 55 drivers/tty/serial/sunhv.c 		port->icount.tx++;
tx                 71 drivers/tty/serial/sunhv.c 		port->icount.tx += sent;
tx                274 drivers/tty/serial/sunsab.c 		up->port.icount.tx++;
tx                461 drivers/tty/serial/sunsab.c 		up->port.icount.tx++;
tx                407 drivers/tty/serial/sunsu.c 		up->port.icount.tx++;
tx                424 drivers/tty/serial/sunsu.c 		up->port.icount.tx++;
tx                496 drivers/tty/serial/sunzilog.c 		up->port.icount.tx++;
tx                516 drivers/tty/serial/sunzilog.c 	up->port.icount.tx++;
tx                705 drivers/tty/serial/sunzilog.c 		port->icount.tx++;
tx                717 drivers/tty/serial/sunzilog.c 		port->icount.tx++;
tx                 31 drivers/tty/serial/tegra-tcu.c 	struct mbox_chan *tx, *rx;
tx                 59 drivers/tty/serial/tegra-tcu.c 	mbox_send_message(tcu->tx, msg);
tx                 60 drivers/tty/serial/tegra-tcu.c 	mbox_flush(tcu->tx, 1000);
tx                191 drivers/tty/serial/tegra-tcu.c 	tcu->tx = mbox_request_channel_byname(&tcu->tx_client, "tx");
tx                192 drivers/tty/serial/tegra-tcu.c 	if (IS_ERR(tcu->tx)) {
tx                193 drivers/tty/serial/tegra-tcu.c 		err = PTR_ERR(tcu->tx);
tx                261 drivers/tty/serial/tegra-tcu.c 	mbox_free_channel(tcu->tx);
tx                276 drivers/tty/serial/tegra-tcu.c 	mbox_free_channel(tcu->tx);
tx                107 drivers/tty/serial/timbuart.c 		port->icount.tx++;
tx                113 drivers/tty/serial/timbuart.c 		port->icount.tx,
tx                186 drivers/tty/serial/uartlite.c 		port->icount.tx++;
tx                195 drivers/tty/serial/uartlite.c 	port->icount.tx++;
tx                354 drivers/tty/serial/ucc_uart.c 		port->icount.tx++;
tx                374 drivers/tty/serial/ucc_uart.c 			port->icount.tx++;
tx                386 drivers/tty/serial/vr41xx_siu.c 		port->icount.tx++;
tx                399 drivers/tty/serial/vr41xx_siu.c 		port->icount.tx++;
tx                202 drivers/tty/serial/vt8500_serial.c 		port->icount.tx++;
tx                217 drivers/tty/serial/vt8500_serial.c 		port->icount.tx++;
tx                328 drivers/tty/serial/xilinx_uartps.c 			port->icount.tx++;
tx                617 drivers/tty/serial/zs.c 		zport->port.icount.tx++;
tx                631 drivers/tty/serial/zs.c 	zport->port.icount.tx++;
tx               2931 drivers/tty/synclink.c 	icount->tx = cnow.tx;
tx               3483 drivers/tty/synclink.c 			      info->icount.tx, info->icount.rx);
tx               5654 drivers/tty/synclink.c 			info->icount.tx += 2;
tx               5670 drivers/tty/synclink.c 			info->icount.tx++;
tx               1113 drivers/tty/synclink_gt.c 	icount->tx = cnow.tx;
tx               1275 drivers/tty/synclink_gt.c 			       info->icount.tx, info->icount.rx);
tx               1322 drivers/tty/synclinkmp.c 	icount->tx = cnow.tx;
tx               1388 drivers/tty/synclinkmp.c 			      info->icount.tx, info->icount.rx);
tx               4304 drivers/tty/synclinkmp.c 			info->icount.tx += 2;
tx               4318 drivers/tty/synclinkmp.c 			info->icount.tx++;
tx                601 drivers/usb/atm/usbatm.c 			atomic_inc(&vcc->stats->tx);
tx                745 drivers/usb/atm/usbatm.c 			       atomic_read(&atm_dev->stats.aal5.tx),
tx                148 drivers/usb/host/max3421-hcd.c 	struct max3421_dma_buf *tx;
tx                364 drivers/usb/host/max3421-hcd.c 	max3421_hcd->tx->data[0] =
tx                368 drivers/usb/host/max3421-hcd.c 	transfer.tx_buf = max3421_hcd->tx->data;
tx                390 drivers/usb/host/max3421-hcd.c 	max3421_hcd->tx->data[0] =
tx                393 drivers/usb/host/max3421-hcd.c 	max3421_hcd->tx->data[1] = val;
tx                395 drivers/usb/host/max3421-hcd.c 	transfer.tx_buf = max3421_hcd->tx->data;
tx                414 drivers/usb/host/max3421-hcd.c 	max3421_hcd->tx->data[0] =
tx                417 drivers/usb/host/max3421-hcd.c 	transfer[0].tx_buf = max3421_hcd->tx->data;
tx                440 drivers/usb/host/max3421-hcd.c 	max3421_hcd->tx->data[0] =
tx                444 drivers/usb/host/max3421-hcd.c 	transfer[0].tx_buf = max3421_hcd->tx->data;
tx               1904 drivers/usb/host/max3421-hcd.c 	max3421_hcd->tx = kmalloc(sizeof(*max3421_hcd->tx), GFP_KERNEL);
tx               1905 drivers/usb/host/max3421-hcd.c 	if (!max3421_hcd->tx)
tx               1940 drivers/usb/host/max3421-hcd.c 		kfree(max3421_hcd->tx);
tx                 98 drivers/usb/musb/cppi_dma.c static void cppi_reset_tx(struct cppi_tx_stateram __iomem *tx, u32 ptr)
tx                100 drivers/usb/musb/cppi_dma.c 	musb_writel(&tx->tx_head, 0, 0);
tx                101 drivers/usb/musb/cppi_dma.c 	musb_writel(&tx->tx_buf, 0, 0);
tx                102 drivers/usb/musb/cppi_dma.c 	musb_writel(&tx->tx_current, 0, 0);
tx                103 drivers/usb/musb/cppi_dma.c 	musb_writel(&tx->tx_buf_current, 0, 0);
tx                104 drivers/usb/musb/cppi_dma.c 	musb_writel(&tx->tx_info, 0, 0);
tx                105 drivers/usb/musb/cppi_dma.c 	musb_writel(&tx->tx_rem_len, 0, 0);
tx                107 drivers/usb/musb/cppi_dma.c 	musb_writel(&tx->tx_complete, 0, ptr);
tx                161 drivers/usb/musb/cppi_dma.c 	for (i = 0; i < ARRAY_SIZE(controller->tx); i++) {
tx                162 drivers/usb/musb/cppi_dma.c 		controller->tx[i].transmit = true;
tx                163 drivers/usb/musb/cppi_dma.c 		controller->tx[i].index = i;
tx                171 drivers/usb/musb/cppi_dma.c 	for (i = 0; i < ARRAY_SIZE(controller->tx); i++)
tx                172 drivers/usb/musb/cppi_dma.c 		cppi_pool_init(controller, controller->tx + i);
tx                180 drivers/usb/musb/cppi_dma.c 	for (i = 0; i < ARRAY_SIZE(controller->tx); i++) {
tx                181 drivers/usb/musb/cppi_dma.c 		struct cppi_channel	*tx_ch = controller->tx + i;
tx                182 drivers/usb/musb/cppi_dma.c 		struct cppi_tx_stateram __iomem *tx;
tx                186 drivers/usb/musb/cppi_dma.c 		tx = tibase + DAVINCI_TXCPPI_STATERAM_OFFSET(i);
tx                187 drivers/usb/musb/cppi_dma.c 		tx_ch->state_ram = tx;
tx                188 drivers/usb/musb/cppi_dma.c 		cppi_reset_tx(tx, 0);
tx                238 drivers/usb/musb/cppi_dma.c 	for (i = 0; i < ARRAY_SIZE(controller->tx); i++) {
tx                240 drivers/usb/musb/cppi_dma.c 		controller->tx[i].last_processed = NULL;
tx                241 drivers/usb/musb/cppi_dma.c 		cppi_pool_free(controller->tx + i);
tx                301 drivers/usb/musb/cppi_dma.c 		if (index >= ARRAY_SIZE(controller->tx)) {
tx                305 drivers/usb/musb/cppi_dma.c 		cppi_ch = controller->tx + index;
tx                385 drivers/usb/musb/cppi_dma.c 	struct cppi_tx_stateram __iomem	*tx = c->state_ram;
tx                396 drivers/usb/musb/cppi_dma.c 		musb_readl(&tx->tx_head, 0),
tx                397 drivers/usb/musb/cppi_dma.c 		musb_readl(&tx->tx_buf, 0),
tx                398 drivers/usb/musb/cppi_dma.c 		musb_readl(&tx->tx_current, 0),
tx                399 drivers/usb/musb/cppi_dma.c 		musb_readl(&tx->tx_buf_current, 0),
tx                401 drivers/usb/musb/cppi_dma.c 		musb_readl(&tx->tx_info, 0),
tx                402 drivers/usb/musb/cppi_dma.c 		musb_readl(&tx->tx_rem_len, 0),
tx                404 drivers/usb/musb/cppi_dma.c 		musb_readl(&tx->tx_complete, 0)
tx                561 drivers/usb/musb/cppi_dma.c cppi_next_tx_segment(struct musb *musb, struct cppi_channel *tx)
tx                563 drivers/usb/musb/cppi_dma.c 	unsigned		maxpacket = tx->maxpacket;
tx                564 drivers/usb/musb/cppi_dma.c 	dma_addr_t		addr = tx->buf_dma + tx->offset;
tx                565 drivers/usb/musb/cppi_dma.c 	size_t			length = tx->buf_len - tx->offset;
tx                569 drivers/usb/musb/cppi_dma.c 	struct cppi_tx_stateram	__iomem *tx_ram = tx->state_ram;
tx                595 drivers/usb/musb/cppi_dma.c 			tx->index,
tx                601 drivers/usb/musb/cppi_dma.c 	cppi_rndis_update(tx, 0, musb->ctrl_base, rndis);
tx                609 drivers/usb/musb/cppi_dma.c 	bd = tx->freelist;
tx                610 drivers/usb/musb/cppi_dma.c 	tx->head = bd;
tx                611 drivers/usb/musb/cppi_dma.c 	tx->last_processed = NULL;
tx                627 drivers/usb/musb/cppi_dma.c 		bd->hw_bufp = tx->buf_dma + tx->offset;
tx                632 drivers/usb/musb/cppi_dma.c 		if ((tx->offset + maxpacket) <= tx->buf_len) {
tx                633 drivers/usb/musb/cppi_dma.c 			tx->offset += maxpacket;
tx                641 drivers/usb/musb/cppi_dma.c 			partial_len = tx->buf_len - tx->offset;
tx                642 drivers/usb/musb/cppi_dma.c 			tx->offset = tx->buf_len;
tx                656 drivers/usb/musb/cppi_dma.c 		tx->tail = bd;
tx                664 drivers/usb/musb/cppi_dma.c 	musb_writel(&tx_ram->tx_head, 0, (u32)tx->freelist->dma);
tx                666 drivers/usb/musb/cppi_dma.c 	cppi_dump_tx(5, tx, "/S");
tx               1147 drivers/usb/musb/cppi_dma.c 	u32			rx, tx;
tx               1157 drivers/usb/musb/cppi_dma.c 	tx = musb_readl(tibase, DAVINCI_TXCPPI_MASKED_REG);
tx               1160 drivers/usb/musb/cppi_dma.c 	if (!tx && !rx) {
tx               1166 drivers/usb/musb/cppi_dma.c 	musb_dbg(musb, "CPPI IRQ Tx%x Rx%x", tx, rx);
tx               1169 drivers/usb/musb/cppi_dma.c 	for (index = 0; tx; tx = tx >> 1, index++) {
tx               1175 drivers/usb/musb/cppi_dma.c 		if (!(tx & 1))
tx               1178 drivers/usb/musb/cppi_dma.c 		tx_ch = cppi->tx + index;
tx                116 drivers/usb/musb/cppi_dma.h 	struct cppi_channel		tx[4];
tx                 62 drivers/usb/musb/musb_gadget.c 				request->tx
tx                 75 drivers/usb/musb/musb_gadget.c 			request->tx
tx                100 drivers/usb/musb/musb_gadget.c 			request->tx
tx                108 drivers/usb/musb/musb_gadget.c 			request->tx
tx               1173 drivers/usb/musb/musb_gadget.c 	if (req->tx)
tx               1228 drivers/usb/musb/musb_gadget.c 	request->tx = musb_ep->is_in;
tx                 58 drivers/usb/musb/musb_gadget.h 	u8 tx;			/* endpoint direction */
tx                927 drivers/usb/musb/musb_gadget_ep0.c 	req->tx = ep->is_in;
tx                226 drivers/usb/musb/musb_trace.h 		__entry->is_tx = req->tx;
tx                 34 drivers/usb/musb/tusb6010_omap.c 	u8			tx;
tx                 71 drivers/usb/musb/tusb6010_omap.c 	if (chdat->tx)
tx                112 drivers/usb/musb/tusb6010_omap.c 		chdat->epnum, chdat->tx ? "tx" : "rx");
tx                114 drivers/usb/musb/tusb6010_omap.c 	if (chdat->tx)
tx                124 drivers/usb/musb/tusb6010_omap.c 			chdat->tx ? "tx" : "rx", remaining);
tx                139 drivers/usb/musb/tusb6010_omap.c 		if (chdat->tx) {
tx                158 drivers/usb/musb/tusb6010_omap.c 	musb_dma_completion(musb, chdat->epnum, chdat->tx);
tx                168 drivers/usb/musb/tusb6010_omap.c 		if (chdat->tx) {
tx                219 drivers/usb/musb/tusb6010_omap.c 	if (chdat->tx)
tx                227 drivers/usb/musb/tusb6010_omap.c 			chdat->tx ? "tx" : "rx", dma_remaining);
tx                260 drivers/usb/musb/tusb6010_omap.c 	if (chdat->tx) {
tx                287 drivers/usb/musb/tusb6010_omap.c 		chdat->epnum, chdat->tx ? "tx" : "rx", &dma_addr,
tx                317 drivers/usb/musb/tusb6010_omap.c 		chdat->epnum, chdat->tx ? "tx" : "rx",
tx                327 drivers/usb/musb/tusb6010_omap.c 	if (chdat->tx) {
tx                344 drivers/usb/musb/tusb6010_omap.c 	if (chdat->tx) {
tx                396 drivers/usb/musb/tusb6010_omap.c 	if (chdat->tx)
tx                424 drivers/usb/musb/tusb6010_omap.c 		u8 tx)
tx                438 drivers/usb/musb/tusb6010_omap.c 		dev_dbg(musb->controller, "Not allowing DMA for ep0 %s\n", tx ? "tx" : "rx");
tx                461 drivers/usb/musb/tusb6010_omap.c 	if (tx)
tx                462 drivers/usb/musb/tusb6010_omap.c 		chdat->tx = 1;
tx                464 drivers/usb/musb/tusb6010_omap.c 		chdat->tx = 0;
tx                484 drivers/usb/musb/tusb6010_omap.c 		chdat->tx ? "tx" : "rx",
tx               2466 drivers/usb/serial/ftdi_sio.c 			port->icount.tx += c;
tx               2474 drivers/usb/serial/ftdi_sio.c 		port->icount.tx += count;
tx                562 drivers/usb/serial/generic.c 	icount->tx = cnow.tx;
tx               1372 drivers/usb/serial/io_edgeport.c 	edge_port->port->icount.tx += count;
tx               1384 drivers/usb/serial/io_edgeport.c 		edge_port->port->icount.tx -= count;
tx               2078 drivers/usb/serial/io_ti.c 		edge_port->port->icount.tx += count;
tx               1375 drivers/usb/serial/mos7840.c 	port->icount.tx += transfer_size;
tx               1376 drivers/usb/serial/mos7840.c 	dev_dbg(&port->dev, "icount.tx is %d:\n", port->icount.tx);
tx               1334 drivers/usb/serial/ti_usb_3410_5052.c 		port->icount.tx += count;
tx                388 drivers/video/fbdev/mmp/hw/mmp_ctrl.h #define	 CFG_TXBITS(tx)				(((tx) - 1)<<8) /* 0x1F~0x1 */
tx                396 drivers/video/fbdev/mmp/hw/mmp_ctrl.h #define	 CFG_TXBITSTO0(tx)			((tx)<<4)
tx                179 drivers/video/fbdev/pxa168fb.h #define     CFG_TXBITS(tx)			((tx) << 8)    /* 0x1F~0x1 */
tx                187 drivers/video/fbdev/pxa168fb.h #define     CFG_TXBITSTO0(tx)			((tx) << 4)
tx                589 fs/dlm/lowcomms.c 			     bool tx, bool rx)
tx                593 fs/dlm/lowcomms.c 	if (tx && !closing && cancel_work_sync(&con->swork)) {
tx                195 include/acpi/processor.h 	int tx;			/* throttle level */
tx                 88 include/linux/async_tx.h static inline void async_tx_issue_pending(struct dma_async_tx_descriptor *tx)
tx                 90 include/linux/async_tx.h 	if (likely(tx)) {
tx                 91 include/linux/async_tx.h 		struct dma_chan *chan = tx->chan;
tx                112 include/linux/async_tx.h static inline void async_tx_issue_pending(struct dma_async_tx_descriptor *tx)
tx                147 include/linux/async_tx.h 		  struct dma_async_tx_descriptor *tx,
tx                152 include/linux/async_tx.h 	args->depend_tx = tx;
tx                158 include/linux/async_tx.h void async_tx_submit(struct dma_chan *chan, struct dma_async_tx_descriptor *tx,
tx                194 include/linux/async_tx.h void async_tx_quiesce(struct dma_async_tx_descriptor **tx);
tx                105 include/linux/cyclades.h 	__u32	cts, dsr, rng, dcd, tx, rx;
tx                171 include/linux/dma/ipu-dma.h #define to_tx_desc(tx) container_of(tx, struct idmac_tx_desc, txd)
tx                501 include/linux/dmaengine.h 	dma_cookie_t (*tx_submit)(struct dma_async_tx_descriptor *tx);
tx                502 include/linux/dmaengine.h 	int (*desc_free)(struct dma_async_tx_descriptor *tx);
tx                515 include/linux/dmaengine.h static inline void dma_set_unmap(struct dma_async_tx_descriptor *tx,
tx                519 include/linux/dmaengine.h 	tx->unmap = unmap;
tx                526 include/linux/dmaengine.h static inline void dma_set_unmap(struct dma_async_tx_descriptor *tx,
tx                540 include/linux/dmaengine.h static inline void dma_descriptor_unmap(struct dma_async_tx_descriptor *tx)
tx                542 include/linux/dmaengine.h 	if (tx->unmap) {
tx                543 include/linux/dmaengine.h 		dmaengine_unmap_put(tx->unmap);
tx                544 include/linux/dmaengine.h 		tx->unmap = NULL;
tx               1184 include/linux/dmaengine.h void dma_async_tx_descriptor_init(struct dma_async_tx_descriptor *tx,
tx               1187 include/linux/dmaengine.h static inline void async_tx_ack(struct dma_async_tx_descriptor *tx)
tx               1189 include/linux/dmaengine.h 	tx->flags |= DMA_CTRL_ACK;
tx               1192 include/linux/dmaengine.h static inline void async_tx_clear_ack(struct dma_async_tx_descriptor *tx)
tx               1194 include/linux/dmaengine.h 	tx->flags &= ~DMA_CTRL_ACK;
tx               1197 include/linux/dmaengine.h static inline bool async_tx_test_ack(struct dma_async_tx_descriptor *tx)
tx               1199 include/linux/dmaengine.h 	return (tx->flags & DMA_CTRL_ACK) == DMA_CTRL_ACK;
tx               1202 include/linux/dmaengine.h #define dma_cap_set(tx, mask) __dma_cap_set((tx), &(mask))
tx               1209 include/linux/dmaengine.h #define dma_cap_clear(tx, mask) __dma_cap_clear((tx), &(mask))
tx               1222 include/linux/dmaengine.h #define dma_has_cap(tx, mask) __dma_has_cap((tx), &(mask))
tx               1304 include/linux/dmaengine.h enum dma_status dma_wait_for_async_tx(struct dma_async_tx_descriptor *tx);
tx               1325 include/linux/dmaengine.h static inline enum dma_status dma_wait_for_async_tx(struct dma_async_tx_descriptor *tx)
tx               1366 include/linux/dmaengine.h static inline int dmaengine_desc_set_reuse(struct dma_async_tx_descriptor *tx)
tx               1371 include/linux/dmaengine.h 	ret = dma_get_slave_caps(tx->chan, &caps);
tx               1376 include/linux/dmaengine.h 		tx->flags |= DMA_CTRL_REUSE;
tx               1383 include/linux/dmaengine.h static inline void dmaengine_desc_clear_reuse(struct dma_async_tx_descriptor *tx)
tx               1385 include/linux/dmaengine.h 	tx->flags &= ~DMA_CTRL_REUSE;
tx               1388 include/linux/dmaengine.h static inline bool dmaengine_desc_test_reuse(struct dma_async_tx_descriptor *tx)
tx               1390 include/linux/dmaengine.h 	return (tx->flags & DMA_CTRL_REUSE) == DMA_CTRL_REUSE;
tx               1407 include/linux/dmaengine.h void dma_run_dependencies(struct dma_async_tx_descriptor *tx);
tx                 70 include/linux/iio/imu/adis.h 	uint8_t			tx[10] ____cacheline_aligned;
tx                114 include/linux/mfd/ipaq-micro.h 	struct ipaq_micro_txdev tx;	/* transmit ISR state */
tx                403 include/linux/mlx5/mlx5_ifc_fpga.h 	u8         tx[0x1];
tx                 45 include/linux/moxtet.h 	u8				tx[TURRIS_MOX_MAX_MODULES];
tx               1165 include/linux/phy.h void phy_set_sym_pause(struct phy_device *phydev, bool rx, bool tx,
tx               1167 include/linux/phy.h void phy_set_asym_pause(struct phy_device *phydev, bool rx, bool tx);
tx                 41 include/linux/posix-clock.h 	int  (*clock_adjtime)(struct posix_clock *pc, struct __kernel_timex *tx);
tx                 21 include/linux/serial.h 	__u32	cts, dsr, rng, dcd, tx, rx;
tx                 94 include/linux/serial_core.h 	__u32	tx;
tx                133 include/linux/soc/ti/knav_dma.h 		struct knav_dma_tx_cfg	tx;
tx                929 include/linux/syscalls.h 				struct __kernel_timex __user *tx);
tx                931 include/linux/syscalls.h 				struct old_timex32 __user *tx);
tx               4155 include/net/cfg80211.h 	u16 tx, rx;
tx               3757 include/net/mac80211.h 	void (*tx)(struct ieee80211_hw *hw,
tx               3931 include/net/mac80211.h 	int (*set_ringparam)(struct ieee80211_hw *hw, u32 tx, u32 rx);
tx               3933 include/net/mac80211.h 			      u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max);
tx                861 include/net/tcp.h 		} tx;   /* only used for outgoing skbs */
tx                268 include/net/tls.h 	struct cipher_context tx;
tx                358 include/net/tls.h int tls_set_sw_offload(struct sock *sk, struct tls_context *ctx, int tx);
tx                430 include/net/tls.h static inline u16 tls_user_config(struct tls_context *ctx, bool tx)
tx                432 include/net/tls.h 	u16 config = tx ? ctx->tx_conf : ctx->rx_conf;
tx                517 include/net/tls.h 		       ctx->tx.iv + TLS_CIPHER_AES_GCM_128_SALT_SIZE, iv_size);
tx                 95 include/net/xdp_sock.h 	struct xsk_queue *tx ____cacheline_aligned_in_smp;
tx                202 include/soc/fsl/qe/immap_qe.h 	u8 	tx[0x400];
tx                100 include/soc/tegra/bpmp.h 	} tx;
tx                 21 include/soc/tegra/ivc.h 	} rx, tx;
tx                 95 include/soc/tegra/ivc.h 		   dma_addr_t rx_phys, void *tx, dma_addr_t tx_phys,
tx                 15 include/uapi/linux/atm_eni.h 	int tx,rx;	/* values are in percent and must be > 100 */
tx                 35 include/uapi/linux/atmdev.h     __HANDLE_ITEM(tx);			/* TX okay */ \
tx                 52 include/uapi/linux/if_xdp.h 	struct xdp_ring_offset tx;
tx                104 include/uapi/linux/serial.h 	int rx, tx;
tx                214 include/uapi/linux/synclink.h 	__u32	cts, dsr, rng, dcd, tx, rx;
tx                226 kernel/time/posix-clock.c static int pc_clock_adjtime(clockid_t id, struct __kernel_timex *tx)
tx                241 kernel/time/posix-clock.c 		err = cd.clk->ops.clock_adjtime(cd.clk, tx);
tx                 11 kernel/time/posix-timers.h 	int	(*clock_adj)(const clockid_t which_clock, struct __kernel_timex *tx);
tx                 48 net/atm/proc.c 		   atomic_read(&stats->tx), atomic_read(&stats->tx_err),
tx                275 net/can/j1939/j1939-priv.h 		unsigned int tx;
tx                704 net/can/j1939/transport.c 		session->pkt.tx = 0;
tx                751 net/can/j1939/transport.c 	session->pkt.tx = session->pkt.tx_acked;
tx                783 net/can/j1939/transport.c 	while (session->pkt.tx < pkt_end) {
tx                784 net/can/j1939/transport.c 		dat[0] = session->pkt.tx - session->pkt.dpo + 1;
tx                785 net/can/j1939/transport.c 		offset = (session->pkt.tx * 7) - skcb->offset;
tx                803 net/can/j1939/transport.c 		session->pkt.tx++;
tx                807 net/can/j1939/transport.c 		if (session->pkt.tx < session->pkt.total && pdelay) {
tx               1375 net/can/j1939/transport.c 	session->pkt.tx = session->pkt.tx_acked;
tx               1572 net/can/j1939/transport.c 	session->pkt.tx = 0;
tx               8916 net/core/dev.c 	struct netdev_queue *tx;
tx               8917 net/core/dev.c 	size_t sz = count * sizeof(*tx);
tx               8922 net/core/dev.c 	tx = kvzalloc(sz, GFP_KERNEL | __GFP_RETRY_MAYFAIL);
tx               8923 net/core/dev.c 	if (!tx)
tx               8926 net/core/dev.c 	dev->_tx = tx;
tx               1448 net/dccp/feat.c 	} tx, rx;
tx               1474 net/dccp/feat.c 	if (ccid_get_builtin_ccids(&tx.val, &tx.len))
tx               1477 net/dccp/feat.c 		kfree(tx.val);
tx               1481 net/dccp/feat.c 	if (!dccp_feat_prefer(sysctl_dccp_tx_ccid, tx.val, tx.len) ||
tx               1485 net/dccp/feat.c 	rc = __feat_register_sp(fn, DCCPF_CCID, true, false, tx.val, tx.len);
tx               1492 net/dccp/feat.c 	kfree(tx.val);
tx               1354 net/ipv4/tcp_input.c 	if (unlikely(TCP_SKB_CB(prev)->tx.delivered_mstamp))
tx               1355 net/ipv4/tcp_input.c 		TCP_SKB_CB(prev)->tx.delivered_mstamp = 0;
tx               3123 net/ipv4/tcp_input.c 			last_in_flight = TCP_SKB_CB(skb)->tx.in_flight;
tx               1038 net/ipv4/tcp_output.c 		TCP_SKB_CB(skb)->tx.in_flight = TCP_SKB_CB(skb)->end_seq
tx               1376 net/ipv4/tcp_output.c 	TCP_SKB_CB(buff)->tx = TCP_SKB_CB(skb)->tx;
tx                 65 net/ipv4/tcp_rate.c 	TCP_SKB_CB(skb)->tx.first_tx_mstamp	= tp->first_tx_mstamp;
tx                 66 net/ipv4/tcp_rate.c 	TCP_SKB_CB(skb)->tx.delivered_mstamp	= tp->delivered_mstamp;
tx                 67 net/ipv4/tcp_rate.c 	TCP_SKB_CB(skb)->tx.delivered		= tp->delivered;
tx                 68 net/ipv4/tcp_rate.c 	TCP_SKB_CB(skb)->tx.is_app_limited	= tp->app_limited ? 1 : 0;
tx                 84 net/ipv4/tcp_rate.c 	if (!scb->tx.delivered_mstamp)
tx                 88 net/ipv4/tcp_rate.c 	    after(scb->tx.delivered, rs->prior_delivered)) {
tx                 89 net/ipv4/tcp_rate.c 		rs->prior_delivered  = scb->tx.delivered;
tx                 90 net/ipv4/tcp_rate.c 		rs->prior_mstamp     = scb->tx.delivered_mstamp;
tx                 91 net/ipv4/tcp_rate.c 		rs->is_app_limited   = scb->tx.is_app_limited;
tx                 98 net/ipv4/tcp_rate.c 						     scb->tx.first_tx_mstamp);
tx                106 net/ipv4/tcp_rate.c 		scb->tx.delivered_mstamp = 0;
tx                 59 net/mac80211/agg-rx.c 				     u16 initiator, u16 reason, bool tx)
tx                 95 net/mac80211/agg-rx.c 	if (initiator == WLAN_BACK_RECIPIENT && tx)
tx                118 net/mac80211/agg-rx.c 				    u16 initiator, u16 reason, bool tx)
tx                121 net/mac80211/agg-rx.c 	___ieee80211_stop_rx_ba_session(sta, tid, initiator, reason, tx);
tx                271 net/mac80211/agg-rx.c 				      u16 buf_size, bool tx, bool auto_seq,
tx                453 net/mac80211/agg-rx.c 	if (tx)
tx                462 net/mac80211/agg-rx.c 					    u16 tid, u16 buf_size, bool tx,
tx                469 net/mac80211/agg-rx.c 					 buf_size, tx, auto_seq, addbaext);
tx                303 net/mac80211/debugfs_sta.c 	bool start, tx;
tx                321 net/mac80211/debugfs_sta.c 		tx = true;
tx                323 net/mac80211/debugfs_sta.c 		tx = false;
tx                332 net/mac80211/debugfs_sta.c 		if (!tx)
tx                345 net/mac80211/debugfs_sta.c 		if (!buf || !tx || !start)
tx                353 net/mac80211/debugfs_sta.c 	if (tx) {
tx                 36 net/mac80211/driver-ops.h 	local->ops->tx(&local->hw, control, skb);
tx                711 net/mac80211/driver-ops.h 				    u32 tx, u32 rx)
tx                717 net/mac80211/driver-ops.h 	trace_drv_set_ringparam(local, tx, rx);
tx                719 net/mac80211/driver-ops.h 		ret = local->ops->set_ringparam(&local->hw, tx, rx);
tx                726 net/mac80211/driver-ops.h 				     u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max)
tx                730 net/mac80211/driver-ops.h 	trace_drv_get_ringparam(local, tx, tx_max, rx, rx_max);
tx                732 net/mac80211/driver-ops.h 		local->ops->get_ringparam(&local->hw, tx, tx_max, rx, rx_max);
tx               1595 net/mac80211/ieee80211_i.h void ieee80211_tx_set_protected(struct ieee80211_tx_data *tx);
tx               1810 net/mac80211/ieee80211_i.h 				      u16 buf_size, bool tx, bool auto_seq,
tx                 75 net/mac80211/key.h 			struct tkip_ctx tx;
tx                411 net/mac80211/main.c 		.tx = 0xffff,
tx                418 net/mac80211/main.c 		.tx = 0xffff,
tx                423 net/mac80211/main.c 		.tx = 0xffff,
tx                434 net/mac80211/main.c 		.tx = 0xffff,
tx                444 net/mac80211/main.c 		.tx = 0xffff,
tx                449 net/mac80211/main.c 		.tx = 0xffff,
tx                459 net/mac80211/main.c 		.tx = 0xffff,
tx                465 net/mac80211/main.c 		.tx = 0xffff,
tx                516 net/mac80211/main.c 	if (WARN_ON(!ops->tx || !ops->start || !ops->stop || !ops->config ||
tx                252 net/mac80211/mesh.h 				    bool tx, bool acked);
tx                534 net/mac80211/mesh_ps.c 				    bool tx, bool acked)
tx                539 net/mac80211/mesh_ps.c 	if (tx) {
tx               2223 net/mac80211/mlme.c 				   u16 stype, u16 reason, bool tx,
tx               2232 net/mac80211/mlme.c 	if (WARN_ON_ONCE(tx && !frame_buf))
tx               2266 net/mac80211/mlme.c 	if (tx)
tx               2270 net/mac80211/mlme.c 	if (tx || frame_buf) {
tx               2283 net/mac80211/mlme.c 					       tx, frame_buf);
tx               2287 net/mac80211/mlme.c 	if (tx)
tx               2653 net/mac80211/mlme.c 					const u8 *buf, size_t len, bool tx,
tx               2658 net/mac80211/mlme.c 		.u.mlme.data = tx ? DEAUTH_TX_EVENT : DEAUTH_RX_EVENT,
tx               2662 net/mac80211/mlme.c 	if (tx)
tx               2675 net/mac80211/mlme.c 	bool tx;
tx               2683 net/mac80211/mlme.c 	tx = !sdata->csa_block_tx;
tx               2692 net/mac80211/mlme.c 			       tx, frame_buf);
tx               2703 net/mac80211/mlme.c 	ieee80211_report_disconnect(sdata, frame_buf, sizeof(frame_buf), tx,
tx               4158 net/mac80211/mlme.c 					  u8 *bssid, u8 reason, bool tx)
tx               4163 net/mac80211/mlme.c 			       tx, frame_buf);
tx               5541 net/mac80211/mlme.c 	bool tx = !req->local_state_change;
tx               5553 net/mac80211/mlme.c 					       req->reason_code, tx,
tx               5573 net/mac80211/mlme.c 					       req->reason_code, tx,
tx               5590 net/mac80211/mlme.c 				       req->reason_code, tx, frame_buf);
tx                155 net/mac80211/tkip.c 	struct tkip_ctx *ctx = &key->u.tkip.tx;
tx                176 net/mac80211/tkip.c 	struct tkip_ctx *ctx = &key->u.tkip.tx;
tx                202 net/mac80211/tkip.c 	struct tkip_ctx *ctx = &key->u.tkip.tx;
tx               1252 net/mac80211/trace.h 	TP_PROTO(struct ieee80211_local *local, u32 tx, u32 rx),
tx               1254 net/mac80211/trace.h 	TP_ARGS(local, tx, rx),
tx               1258 net/mac80211/trace.h 		__field(u32, tx)
tx               1264 net/mac80211/trace.h 		__entry->tx = tx;
tx               1270 net/mac80211/trace.h 		LOCAL_PR_ARG, __entry->tx, __entry->rx
tx               1275 net/mac80211/trace.h 	TP_PROTO(struct ieee80211_local *local, u32 *tx, u32 *tx_max,
tx               1278 net/mac80211/trace.h 	TP_ARGS(local, tx, tx_max, rx, rx_max),
tx               1282 net/mac80211/trace.h 		__field(u32, tx)
tx               1290 net/mac80211/trace.h 		__entry->tx = *tx;
tx               1299 net/mac80211/trace.h 		__entry->tx, __entry->tx_max, __entry->rx, __entry->rx_max
tx                 51 net/mac80211/tx.c static __le16 ieee80211_duration(struct ieee80211_tx_data *tx,
tx                 57 net/mac80211/tx.c 	struct ieee80211_local *local = tx->local;
tx                 65 net/mac80211/tx.c 	if (tx->rate.flags & (IEEE80211_TX_RC_MCS | IEEE80211_TX_RC_VHT_MCS))
tx                 69 net/mac80211/tx.c 	chanctx_conf = rcu_dereference(tx->sdata->vif.chanctx_conf);
tx                 77 net/mac80211/tx.c 	if (WARN_ON_ONCE(tx->rate.idx < 0))
tx                 81 net/mac80211/tx.c 	txrate = &sband->bitrates[tx->rate.idx];
tx                150 net/mac80211/tx.c 		if (tx->sdata->vif.bss_conf.basic_rates & BIT(i))
tx                156 net/mac80211/tx.c 			if (tx->sdata->flags & IEEE80211_SDATA_OPERATING_GMODE)
tx                191 net/mac80211/tx.c 				tx->sdata->vif.bss_conf.use_short_preamble,
tx                201 net/mac80211/tx.c 				tx->sdata->vif.bss_conf.use_short_preamble,
tx                210 net/mac80211/tx.c ieee80211_tx_h_dynamic_ps(struct ieee80211_tx_data *tx)
tx                212 net/mac80211/tx.c 	struct ieee80211_local *local = tx->local;
tx                214 net/mac80211/tx.c 	struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
tx                240 net/mac80211/tx.c 	if (tx->sdata->vif.type != NL80211_IFTYPE_STATION)
tx                246 net/mac80211/tx.c 	ifmgd = &tx->sdata->u.mgd;
tx                263 net/mac80211/tx.c 	    skb_get_queue_mapping(tx->skb) == IEEE80211_AC_VO)
tx                287 net/mac80211/tx.c ieee80211_tx_h_check_assoc(struct ieee80211_tx_data *tx)
tx                290 net/mac80211/tx.c 	struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)tx->skb->data;
tx                291 net/mac80211/tx.c 	struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
tx                297 net/mac80211/tx.c 	if (unlikely(test_bit(SCAN_SW_SCANNING, &tx->local->scanning)) &&
tx                298 net/mac80211/tx.c 	    test_bit(SDATA_STATE_OFFCHANNEL, &tx->sdata->state) &&
tx                314 net/mac80211/tx.c 	if (tx->sdata->vif.type == NL80211_IFTYPE_OCB)
tx                317 net/mac80211/tx.c 	if (tx->sdata->vif.type == NL80211_IFTYPE_WDS)
tx                320 net/mac80211/tx.c 	if (tx->flags & IEEE80211_TX_PS_BUFFERED)
tx                323 net/mac80211/tx.c 	if (tx->sta)
tx                324 net/mac80211/tx.c 		assoc = test_sta_flag(tx->sta, WLAN_STA_ASSOC);
tx                326 net/mac80211/tx.c 	if (likely(tx->flags & IEEE80211_TX_UNICAST)) {
tx                330 net/mac80211/tx.c 			sdata_info(tx->sdata,
tx                334 net/mac80211/tx.c 			I802_DEBUG_INC(tx->local->tx_handlers_drop_not_assoc);
tx                338 net/mac80211/tx.c 			    ieee80211_vif_get_num_mcast_if(tx->sdata) == 0)) {
tx                401 net/mac80211/tx.c ieee80211_tx_h_multicast_ps_buf(struct ieee80211_tx_data *tx)
tx                403 net/mac80211/tx.c 	struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
tx                404 net/mac80211/tx.c 	struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)tx->skb->data;
tx                416 net/mac80211/tx.c 	if (tx->sdata->vif.type == NL80211_IFTYPE_AP ||
tx                417 net/mac80211/tx.c 	    tx->sdata->vif.type == NL80211_IFTYPE_AP_VLAN) {
tx                418 net/mac80211/tx.c 		if (!tx->sdata->bss)
tx                421 net/mac80211/tx.c 		ps = &tx->sdata->bss->ps;
tx                422 net/mac80211/tx.c 	} else if (ieee80211_vif_is_mesh(&tx->sdata->vif)) {
tx                423 net/mac80211/tx.c 		ps = &tx->sdata->u.mesh.ps;
tx                436 net/mac80211/tx.c 	if (ieee80211_hw_check(&tx->local->hw, QUEUE_CONTROL))
tx                437 net/mac80211/tx.c 		info->hw_queue = tx->sdata->vif.cab_queue;
tx                446 net/mac80211/tx.c 	if (!ieee80211_hw_check(&tx->local->hw, HOST_BROADCAST_PS_BUFFERING))
tx                450 net/mac80211/tx.c 	if (tx->local->total_ps_buffered >= TOTAL_MAX_TX_BUFFER)
tx                451 net/mac80211/tx.c 		purge_old_ps_buffers(tx->local);
tx                454 net/mac80211/tx.c 		ps_dbg(tx->sdata,
tx                456 net/mac80211/tx.c 		ieee80211_free_txskb(&tx->local->hw, skb_dequeue(&ps->bc_buf));
tx                458 net/mac80211/tx.c 		tx->local->total_ps_buffered++;
tx                460 net/mac80211/tx.c 	skb_queue_tail(&ps->bc_buf, tx->skb);
tx                481 net/mac80211/tx.c ieee80211_tx_h_unicast_ps_buf(struct ieee80211_tx_data *tx)
tx                483 net/mac80211/tx.c 	struct sta_info *sta = tx->sta;
tx                484 net/mac80211/tx.c 	struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
tx                485 net/mac80211/tx.c 	struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)tx->skb->data;
tx                486 net/mac80211/tx.c 	struct ieee80211_local *local = tx->local;
tx                495 net/mac80211/tx.c 		int ac = skb_get_queue_mapping(tx->skb);
tx                505 net/mac80211/tx.c 		if (tx->local->total_ps_buffered >= TOTAL_MAX_TX_BUFFER)
tx                506 net/mac80211/tx.c 			purge_old_ps_buffers(tx->local);
tx                524 net/mac80211/tx.c 			ps_dbg(tx->sdata,
tx                529 net/mac80211/tx.c 			tx->local->total_ps_buffered++;
tx                532 net/mac80211/tx.c 		info->control.vif = &tx->sdata->vif;
tx                535 net/mac80211/tx.c 		skb_queue_tail(&sta->ps_tx_buf[ac], tx->skb);
tx                551 net/mac80211/tx.c 		ps_dbg(tx->sdata,
tx                560 net/mac80211/tx.c ieee80211_tx_h_ps_buf(struct ieee80211_tx_data *tx)
tx                562 net/mac80211/tx.c 	if (unlikely(tx->flags & IEEE80211_TX_PS_BUFFERED))
tx                565 net/mac80211/tx.c 	if (tx->flags & IEEE80211_TX_UNICAST)
tx                566 net/mac80211/tx.c 		return ieee80211_tx_h_unicast_ps_buf(tx);
tx                568 net/mac80211/tx.c 		return ieee80211_tx_h_multicast_ps_buf(tx);
tx                572 net/mac80211/tx.c ieee80211_tx_h_check_control_port_protocol(struct ieee80211_tx_data *tx)
tx                574 net/mac80211/tx.c 	struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
tx                576 net/mac80211/tx.c 	if (unlikely(tx->sdata->control_port_protocol == tx->skb->protocol)) {
tx                577 net/mac80211/tx.c 		if (tx->sdata->control_port_no_encrypt)
tx                587 net/mac80211/tx.c ieee80211_tx_h_select_key(struct ieee80211_tx_data *tx)
tx                590 net/mac80211/tx.c 	struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
tx                591 net/mac80211/tx.c 	struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)tx->skb->data;
tx                594 net/mac80211/tx.c 		tx->key = NULL;
tx                598 net/mac80211/tx.c 	if (tx->sta &&
tx                599 net/mac80211/tx.c 	    (key = rcu_dereference(tx->sta->ptk[tx->sta->ptk_idx])))
tx                600 net/mac80211/tx.c 		tx->key = key;
tx                601 net/mac80211/tx.c 	else if (ieee80211_is_group_privacy_action(tx->skb) &&
tx                602 net/mac80211/tx.c 		(key = rcu_dereference(tx->sdata->default_multicast_key)))
tx                603 net/mac80211/tx.c 		tx->key = key;
tx                606 net/mac80211/tx.c 		 ieee80211_is_robust_mgmt_frame(tx->skb) &&
tx                607 net/mac80211/tx.c 		 (key = rcu_dereference(tx->sdata->default_mgmt_key)))
tx                608 net/mac80211/tx.c 		tx->key = key;
tx                610 net/mac80211/tx.c 		 (key = rcu_dereference(tx->sdata->default_multicast_key)))
tx                611 net/mac80211/tx.c 		tx->key = key;
tx                613 net/mac80211/tx.c 		 (key = rcu_dereference(tx->sdata->default_unicast_key)))
tx                614 net/mac80211/tx.c 		tx->key = key;
tx                616 net/mac80211/tx.c 		tx->key = NULL;
tx                618 net/mac80211/tx.c 	if (tx->key) {
tx                623 net/mac80211/tx.c 		switch (tx->key->conf.cipher) {
tx                628 net/mac80211/tx.c 				tx->key = NULL;
tx                635 net/mac80211/tx.c 			    !ieee80211_use_mfp(hdr->frame_control, tx->sta,
tx                636 net/mac80211/tx.c 					       tx->skb) &&
tx                637 net/mac80211/tx.c 			    !ieee80211_is_group_privacy_action(tx->skb))
tx                638 net/mac80211/tx.c 				tx->key = NULL;
tx                640 net/mac80211/tx.c 				skip_hw = (tx->key->conf.flags &
tx                649 net/mac80211/tx.c 				tx->key = NULL;
tx                653 net/mac80211/tx.c 		if (unlikely(tx->key && tx->key->flags & KEY_FLAG_TAINTED &&
tx                657 net/mac80211/tx.c 		if (!skip_hw && tx->key &&
tx                658 net/mac80211/tx.c 		    tx->key->flags & KEY_FLAG_UPLOADED_TO_HARDWARE)
tx                659 net/mac80211/tx.c 			info->control.hw_key = &tx->key->conf;
tx                660 net/mac80211/tx.c 	} else if (!ieee80211_is_mgmt(hdr->frame_control) && tx->sta &&
tx                661 net/mac80211/tx.c 		   test_sta_flag(tx->sta, WLAN_STA_USES_ENCRYPTION)) {
tx                669 net/mac80211/tx.c ieee80211_tx_h_rate_ctrl(struct ieee80211_tx_data *tx)
tx                671 net/mac80211/tx.c 	struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
tx                672 net/mac80211/tx.c 	struct ieee80211_hdr *hdr = (void *)tx->skb->data;
tx                681 net/mac80211/tx.c 	sband = tx->local->hw.wiphy->bands[info->band];
tx                683 net/mac80211/tx.c 	len = min_t(u32, tx->skb->len + FCS_LEN,
tx                684 net/mac80211/tx.c 			 tx->local->hw.wiphy->frag_threshold);
tx                687 net/mac80211/tx.c 	txrc.hw = &tx->local->hw;
tx                689 net/mac80211/tx.c 	txrc.bss_conf = &tx->sdata->vif.bss_conf;
tx                690 net/mac80211/tx.c 	txrc.skb = tx->skb;
tx                692 net/mac80211/tx.c 	txrc.rate_idx_mask = tx->sdata->rc_rateidx_mask[info->band];
tx                694 net/mac80211/tx.c 	if (tx->sdata->rc_has_mcs_mask[info->band])
tx                696 net/mac80211/tx.c 			tx->sdata->rc_rateidx_mcs_mask[info->band];
tx                698 net/mac80211/tx.c 	txrc.bss = (tx->sdata->vif.type == NL80211_IFTYPE_AP ||
tx                699 net/mac80211/tx.c 		    tx->sdata->vif.type == NL80211_IFTYPE_MESH_POINT ||
tx                700 net/mac80211/tx.c 		    tx->sdata->vif.type == NL80211_IFTYPE_ADHOC ||
tx                701 net/mac80211/tx.c 		    tx->sdata->vif.type == NL80211_IFTYPE_OCB);
tx                704 net/mac80211/tx.c 	if (len > tx->local->hw.wiphy->rts_threshold) {
tx                709 net/mac80211/tx.c 	info->control.use_cts_prot = tx->sdata->vif.bss_conf.use_cts_prot;
tx                717 net/mac80211/tx.c 	if (tx->sdata->vif.bss_conf.use_short_preamble &&
tx                719 net/mac80211/tx.c 	     (tx->sta && test_sta_flag(tx->sta, WLAN_STA_SHORT_PREAMBLE))))
tx                728 net/mac80211/tx.c 	if (tx->sta)
tx                729 net/mac80211/tx.c 		assoc = test_sta_flag(tx->sta, WLAN_STA_ASSOC);
tx                735 net/mac80211/tx.c 	if (WARN(test_bit(SCAN_SW_SCANNING, &tx->local->scanning) && assoc &&
tx                736 net/mac80211/tx.c 		 !rate_usable_index_exists(sband, &tx->sta->sta),
tx                740 net/mac80211/tx.c 		 tx->sdata->name, hdr->addr1,
tx                748 net/mac80211/tx.c 	rate_control_get_rate(tx->sdata, tx->sta, &txrc);
tx                750 net/mac80211/tx.c 	if (tx->sta && !info->control.skip_table)
tx                751 net/mac80211/tx.c 		ratetbl = rcu_dereference(tx->sta->sta.rates);
tx                764 net/mac80211/tx.c 			tx->rate = rate;
tx                769 net/mac80211/tx.c 		tx->rate = info->control.rates[0];
tx                773 net/mac80211/tx.c 		txrc.reported_rate = tx->rate;
tx                774 net/mac80211/tx.c 		if (tx->sta && ieee80211_is_data(hdr->frame_control))
tx                775 net/mac80211/tx.c 			tx->sta->tx_stats.last_rate = txrc.reported_rate;
tx                776 net/mac80211/tx.c 	} else if (tx->sta)
tx                777 net/mac80211/tx.c 		tx->sta->tx_stats.last_rate = txrc.reported_rate;
tx                804 net/mac80211/tx.c ieee80211_tx_h_sequence(struct ieee80211_tx_data *tx)
tx                806 net/mac80211/tx.c 	struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
tx                807 net/mac80211/tx.c 	struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)tx->skb->data;
tx                835 net/mac80211/tx.c 		if (tx->flags & IEEE80211_TX_NO_SEQNO)
tx                840 net/mac80211/tx.c 		hdr->seq_ctrl = cpu_to_le16(tx->sdata->sequence_number);
tx                841 net/mac80211/tx.c 		tx->sdata->sequence_number += 0x10;
tx                842 net/mac80211/tx.c 		if (tx->sta)
tx                843 net/mac80211/tx.c 			tx->sta->tx_stats.msdu[IEEE80211_NUM_TIDS]++;
tx                852 net/mac80211/tx.c 	if (!tx->sta)
tx                857 net/mac80211/tx.c 	tx->sta->tx_stats.msdu[tid]++;
tx                859 net/mac80211/tx.c 	hdr->seq_ctrl = ieee80211_tx_next_seq(tx->sta, tid);
tx                864 net/mac80211/tx.c static int ieee80211_fragment(struct ieee80211_tx_data *tx,
tx                868 net/mac80211/tx.c 	struct ieee80211_local *local = tx->local;
tx                888 net/mac80211/tx.c 				    tx->sdata->encrypt_headroom +
tx                893 net/mac80211/tx.c 		__skb_queue_tail(&tx->skbs, tmp);
tx                896 net/mac80211/tx.c 			    local->tx_headroom + tx->sdata->encrypt_headroom);
tx                925 net/mac80211/tx.c ieee80211_tx_h_fragment(struct ieee80211_tx_data *tx)
tx                927 net/mac80211/tx.c 	struct sk_buff *skb = tx->skb;
tx                930 net/mac80211/tx.c 	int frag_threshold = tx->local->hw.wiphy->frag_threshold;
tx                935 net/mac80211/tx.c 	__skb_queue_tail(&tx->skbs, skb);
tx                936 net/mac80211/tx.c 	tx->skb = NULL;
tx                941 net/mac80211/tx.c 	if (ieee80211_hw_check(&tx->local->hw, SUPPORTS_TX_FRAG))
tx                966 net/mac80211/tx.c 	if (ieee80211_fragment(tx, skb, hdrlen, frag_threshold))
tx                972 net/mac80211/tx.c 	skb_queue_walk(&tx->skbs, skb) {
tx                978 net/mac80211/tx.c 		if (!skb_queue_is_last(&tx->skbs, skb)) {
tx               1000 net/mac80211/tx.c ieee80211_tx_h_stats(struct ieee80211_tx_data *tx)
tx               1005 net/mac80211/tx.c 	if (!tx->sta)
tx               1008 net/mac80211/tx.c 	skb_queue_walk(&tx->skbs, skb) {
tx               1010 net/mac80211/tx.c 		tx->sta->tx_stats.bytes[ac] += skb->len;
tx               1013 net/mac80211/tx.c 		tx->sta->tx_stats.packets[ac]++;
tx               1019 net/mac80211/tx.c ieee80211_tx_h_encrypt(struct ieee80211_tx_data *tx)
tx               1021 net/mac80211/tx.c 	if (!tx->key)
tx               1024 net/mac80211/tx.c 	switch (tx->key->conf.cipher) {
tx               1027 net/mac80211/tx.c 		return ieee80211_crypto_wep_encrypt(tx);
tx               1029 net/mac80211/tx.c 		return ieee80211_crypto_tkip_encrypt(tx);
tx               1032 net/mac80211/tx.c 			tx, IEEE80211_CCMP_MIC_LEN);
tx               1035 net/mac80211/tx.c 			tx, IEEE80211_CCMP_256_MIC_LEN);
tx               1037 net/mac80211/tx.c 		return ieee80211_crypto_aes_cmac_encrypt(tx);
tx               1039 net/mac80211/tx.c 		return ieee80211_crypto_aes_cmac_256_encrypt(tx);
tx               1042 net/mac80211/tx.c 		return ieee80211_crypto_aes_gmac_encrypt(tx);
tx               1045 net/mac80211/tx.c 		return ieee80211_crypto_gcmp_encrypt(tx);
tx               1047 net/mac80211/tx.c 		return ieee80211_crypto_hw_encrypt(tx);
tx               1054 net/mac80211/tx.c ieee80211_tx_h_calculate_duration(struct ieee80211_tx_data *tx)
tx               1061 net/mac80211/tx.c 	skb_queue_walk(&tx->skbs, skb) {
tx               1065 net/mac80211/tx.c 		if (!skb_queue_is_last(&tx->skbs, skb)) {
tx               1066 net/mac80211/tx.c 			struct sk_buff *next = skb_queue_next(&tx->skbs, skb);
tx               1073 net/mac80211/tx.c 			ieee80211_duration(tx, skb, group_addr, next_len);
tx               1081 net/mac80211/tx.c static bool ieee80211_tx_prep_agg(struct ieee80211_tx_data *tx,
tx               1099 net/mac80211/tx.c 	} else if (!tx->sta->sta.txq[tid]) {
tx               1100 net/mac80211/tx.c 		spin_lock(&tx->sta->lock);
tx               1119 net/mac80211/tx.c 		tid_tx = rcu_dereference_protected_tid_tx(tx->sta, tid);
tx               1129 net/mac80211/tx.c 				clear_sta_flag(tx->sta, WLAN_STA_SP);
tx               1130 net/mac80211/tx.c 				ps_dbg(tx->sta->sdata,
tx               1132 net/mac80211/tx.c 				       tx->sta->sta.addr, tx->sta->sta.aid);
tx               1134 net/mac80211/tx.c 			info->control.vif = &tx->sdata->vif;
tx               1141 net/mac80211/tx.c 		spin_unlock(&tx->sta->lock);
tx               1144 net/mac80211/tx.c 			ieee80211_free_txskb(&tx->local->hw, purge_skb);
tx               1161 net/mac80211/tx.c 		     struct ieee80211_tx_data *tx,
tx               1169 net/mac80211/tx.c 	memset(tx, 0, sizeof(*tx));
tx               1170 net/mac80211/tx.c 	tx->skb = skb;
tx               1171 net/mac80211/tx.c 	tx->local = local;
tx               1172 net/mac80211/tx.c 	tx->sdata = sdata;
tx               1173 net/mac80211/tx.c 	__skb_queue_head_init(&tx->skbs);
tx               1186 net/mac80211/tx.c 			tx->sta = sta;
tx               1189 net/mac80211/tx.c 			tx->sta = rcu_dereference(sdata->u.vlan.sta);
tx               1190 net/mac80211/tx.c 			if (!tx->sta && sdata->wdev.use_4addr)
tx               1194 net/mac80211/tx.c 			   tx->sdata->control_port_protocol == tx->skb->protocol) {
tx               1195 net/mac80211/tx.c 			tx->sta = sta_info_get_bss(sdata, hdr->addr1);
tx               1197 net/mac80211/tx.c 		if (!tx->sta && !is_multicast_ether_addr(hdr->addr1))
tx               1198 net/mac80211/tx.c 			tx->sta = sta_info_get(sdata, hdr->addr1);
tx               1201 net/mac80211/tx.c 	if (tx->sta && ieee80211_is_data_qos(hdr->frame_control) &&
tx               1209 net/mac80211/tx.c 		tid_tx = rcu_dereference(tx->sta->ampdu_mlme.tid_tx[tid]);
tx               1213 net/mac80211/tx.c 			queued = ieee80211_tx_prep_agg(tx, skb, info,
tx               1222 net/mac80211/tx.c 		tx->flags &= ~IEEE80211_TX_UNICAST;
tx               1225 net/mac80211/tx.c 		tx->flags |= IEEE80211_TX_UNICAST;
tx               1228 net/mac80211/tx.c 		if (!(tx->flags & IEEE80211_TX_UNICAST) ||
tx               1234 net/mac80211/tx.c 	if (!tx->sta)
tx               1236 net/mac80211/tx.c 	else if (test_and_clear_sta_flag(tx->sta, WLAN_STA_CLEAR_PS_FILT)) {
tx               1238 net/mac80211/tx.c 		ieee80211_check_fast_xmit(tx->sta);
tx               1771 net/mac80211/tx.c static int invoke_tx_handlers_early(struct ieee80211_tx_data *tx)
tx               1777 net/mac80211/tx.c 		res = txh(tx);		\
tx               1787 net/mac80211/tx.c 	if (!ieee80211_hw_check(&tx->local->hw, HAS_RATE_CONTROL))
tx               1792 net/mac80211/tx.c 		I802_DEBUG_INC(tx->local->tx_handlers_drop);
tx               1793 net/mac80211/tx.c 		if (tx->skb)
tx               1794 net/mac80211/tx.c 			ieee80211_free_txskb(&tx->local->hw, tx->skb);
tx               1796 net/mac80211/tx.c 			ieee80211_purge_tx_queue(&tx->local->hw, &tx->skbs);
tx               1799 net/mac80211/tx.c 		I802_DEBUG_INC(tx->local->tx_handlers_queued);
tx               1810 net/mac80211/tx.c static int invoke_tx_handlers_late(struct ieee80211_tx_data *tx)
tx               1812 net/mac80211/tx.c 	struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
tx               1816 net/mac80211/tx.c 		__skb_queue_tail(&tx->skbs, tx->skb);
tx               1817 net/mac80211/tx.c 		tx->skb = NULL;
tx               1827 net/mac80211/tx.c 	if (!ieee80211_hw_check(&tx->local->hw, HAS_RATE_CONTROL))
tx               1833 net/mac80211/tx.c 		I802_DEBUG_INC(tx->local->tx_handlers_drop);
tx               1834 net/mac80211/tx.c 		if (tx->skb)
tx               1835 net/mac80211/tx.c 			ieee80211_free_txskb(&tx->local->hw, tx->skb);
tx               1837 net/mac80211/tx.c 			ieee80211_purge_tx_queue(&tx->local->hw, &tx->skbs);
tx               1840 net/mac80211/tx.c 		I802_DEBUG_INC(tx->local->tx_handlers_queued);
tx               1847 net/mac80211/tx.c static int invoke_tx_handlers(struct ieee80211_tx_data *tx)
tx               1849 net/mac80211/tx.c 	int r = invoke_tx_handlers_early(tx);
tx               1853 net/mac80211/tx.c 	return invoke_tx_handlers_late(tx);
tx               1862 net/mac80211/tx.c 	struct ieee80211_tx_data tx;
tx               1865 net/mac80211/tx.c 	if (ieee80211_tx_prepare(sdata, &tx, NULL, skb) == TX_DROP)
tx               1872 net/mac80211/tx.c 	if (invoke_tx_handlers(&tx))
tx               1876 net/mac80211/tx.c 		if (tx.sta)
tx               1877 net/mac80211/tx.c 			*sta = &tx.sta->sta;
tx               1883 net/mac80211/tx.c 	skb2 = __skb_dequeue(&tx.skbs);
tx               1884 net/mac80211/tx.c 	if (WARN_ON(skb2 != skb || !skb_queue_empty(&tx.skbs))) {
tx               1886 net/mac80211/tx.c 		ieee80211_purge_tx_queue(hw, &tx.skbs);
tx               1902 net/mac80211/tx.c 	struct ieee80211_tx_data tx;
tx               1915 net/mac80211/tx.c 	res_prepare = ieee80211_tx_prepare(sdata, &tx, sta, skb);
tx               1917 net/mac80211/tx.c 	tx.flags |= txdata_flags;
tx               1932 net/mac80211/tx.c 	if (invoke_tx_handlers_early(&tx))
tx               1935 net/mac80211/tx.c 	if (ieee80211_queue_skb(local, sdata, tx.sta, tx.skb))
tx               1938 net/mac80211/tx.c 	if (!invoke_tx_handlers_late(&tx))
tx               1939 net/mac80211/tx.c 		result = __ieee80211_tx(local, &tx.skbs, led_len,
tx               1940 net/mac80211/tx.c 					tx.sta, txpending);
tx               3434 net/mac80211/tx.c 	struct ieee80211_tx_data tx;
tx               3515 net/mac80211/tx.c 	__skb_queue_head_init(&tx.skbs);
tx               3517 net/mac80211/tx.c 	tx.flags = IEEE80211_TX_UNICAST;
tx               3518 net/mac80211/tx.c 	tx.local = local;
tx               3519 net/mac80211/tx.c 	tx.sdata = sdata;
tx               3520 net/mac80211/tx.c 	tx.sta = sta;
tx               3521 net/mac80211/tx.c 	tx.key = fast_tx->key;
tx               3524 net/mac80211/tx.c 		tx.skb = skb;
tx               3525 net/mac80211/tx.c 		r = ieee80211_tx_h_rate_ctrl(&tx);
tx               3526 net/mac80211/tx.c 		skb = tx.skb;
tx               3527 net/mac80211/tx.c 		tx.skb = NULL;
tx               3546 net/mac80211/tx.c 	__skb_queue_tail(&tx.skbs, skb);
tx               3547 net/mac80211/tx.c 	ieee80211_tx_frags(local, &sdata->vif, &sta->sta, &tx.skbs, false);
tx               3561 net/mac80211/tx.c 	struct ieee80211_tx_data tx;
tx               3593 net/mac80211/tx.c 	memset(&tx, 0, sizeof(tx));
tx               3594 net/mac80211/tx.c 	__skb_queue_head_init(&tx.skbs);
tx               3595 net/mac80211/tx.c 	tx.local = local;
tx               3596 net/mac80211/tx.c 	tx.skb = skb;
tx               3597 net/mac80211/tx.c 	tx.sdata = vif_to_sdata(info->control.vif);
tx               3600 net/mac80211/tx.c 		tx.sta = container_of(txq->sta, struct sta_info, sta);
tx               3606 net/mac80211/tx.c 			     !ieee80211_vif_is_mesh(&tx.sdata->vif) &&
tx               3607 net/mac80211/tx.c 			     tx.sdata->vif.type != NL80211_IFTYPE_OCB &&
tx               3609 net/mac80211/tx.c 			     !test_sta_flag(tx.sta, WLAN_STA_AUTHORIZED) &&
tx               3612 net/mac80211/tx.c 			      !ether_addr_equal(tx.sdata->vif.addr,
tx               3624 net/mac80211/tx.c 	r = ieee80211_tx_h_select_key(&tx);
tx               3640 net/mac80211/tx.c 		if (tx.key &&
tx               3641 net/mac80211/tx.c 		    (tx.key->conf.flags & IEEE80211_KEY_FLAG_GENERATE_IV))
tx               3645 net/mac80211/tx.c 					   tx.key, skb);
tx               3647 net/mac80211/tx.c 		if (invoke_tx_handlers_late(&tx))
tx               3650 net/mac80211/tx.c 		skb = __skb_dequeue(&tx.skbs);
tx               3652 net/mac80211/tx.c 		if (!skb_queue_empty(&tx.skbs)) {
tx               3654 net/mac80211/tx.c 			skb_queue_splice_tail(&tx.skbs, &txqi->frags);
tx               3667 net/mac80211/tx.c 	switch (tx.sdata->vif.type) {
tx               3669 net/mac80211/tx.c 		if (tx.sdata->u.mntr.flags & MONITOR_FLAG_ACTIVE) {
tx               3670 net/mac80211/tx.c 			vif = &tx.sdata->vif;
tx               3673 net/mac80211/tx.c 		tx.sdata = rcu_dereference(local->monitor_sdata);
tx               3674 net/mac80211/tx.c 		if (tx.sdata) {
tx               3675 net/mac80211/tx.c 			vif = &tx.sdata->vif;
tx               3686 net/mac80211/tx.c 		tx.sdata = container_of(tx.sdata->bss,
tx               3690 net/mac80211/tx.c 		vif = &tx.sdata->vif;
tx               4070 net/mac80211/tx.c 	struct ieee80211_tx_data tx = {
tx               4089 net/mac80211/tx.c 	tx.sta = sta_info_get(sdata, hdr->addr1);
tx               4090 net/mac80211/tx.c 	tx.skb = skb;
tx               4092 net/mac80211/tx.c 	if (ieee80211_tx_h_select_key(&tx) != TX_CONTINUE) {
tx               4898 net/mac80211/tx.c 	struct ieee80211_tx_data tx;
tx               4947 net/mac80211/tx.c 		if (!ieee80211_tx_prepare(sdata, &tx, NULL, skb))
tx               4954 net/mac80211/tx.c 	tx.flags |= IEEE80211_TX_PS_BUFFERED;
tx               4957 net/mac80211/tx.c 	if (invoke_tx_handlers(&tx))
tx                 49 net/mac80211/util.c void ieee80211_tx_set_protected(struct ieee80211_tx_data *tx)
tx                 54 net/mac80211/util.c 	skb_queue_walk(&tx->skbs, skb) {
tx                272 net/mac80211/wep.c static int wep_encrypt_skb(struct ieee80211_tx_data *tx, struct sk_buff *skb)
tx                278 net/mac80211/wep.c 		if (ieee80211_wep_encrypt(tx->local, skb, tx->key->conf.key,
tx                279 net/mac80211/wep.c 					  tx->key->conf.keylen,
tx                280 net/mac80211/wep.c 					  tx->key->conf.keyidx))
tx                284 net/mac80211/wep.c 		if (!ieee80211_wep_add_iv(tx->local, skb,
tx                285 net/mac80211/wep.c 					  tx->key->conf.keylen,
tx                286 net/mac80211/wep.c 					  tx->key->conf.keyidx))
tx                294 net/mac80211/wep.c ieee80211_crypto_wep_encrypt(struct ieee80211_tx_data *tx)
tx                298 net/mac80211/wep.c 	ieee80211_tx_set_protected(tx);
tx                300 net/mac80211/wep.c 	skb_queue_walk(&tx->skbs, skb) {
tx                301 net/mac80211/wep.c 		if (wep_encrypt_skb(tx, skb) < 0) {
tx                302 net/mac80211/wep.c 			I802_DEBUG_INC(tx->local->tx_handlers_drop_wep);
tx                 28 net/mac80211/wep.h ieee80211_crypto_wep_encrypt(struct ieee80211_tx_data *tx);
tx                 29 net/mac80211/wpa.c ieee80211_tx_h_michael_mic_add(struct ieee80211_tx_data *tx)
tx                 35 net/mac80211/wpa.c 	struct sk_buff *skb = tx->skb;
tx                 40 net/mac80211/wpa.c 	if (!tx->key || tx->key->conf.cipher != WLAN_CIPHER_SUITE_TKIP ||
tx                 58 net/mac80211/wpa.c 	     ieee80211_hw_check(&tx->local->hw, SUPPORTS_TX_FRAG)) &&
tx                 59 net/mac80211/wpa.c 	    !(tx->key->conf.flags & (IEEE80211_KEY_FLAG_GENERATE_MMIC |
tx                 78 net/mac80211/wpa.c 	if (tx->key->conf.flags & IEEE80211_KEY_FLAG_PUT_MIC_SPACE) {
tx                 84 net/mac80211/wpa.c 	key = &tx->key->conf.key[NL80211_TKIP_DATA_OFFSET_TX_MIC_KEY];
tx                193 net/mac80211/wpa.c static int tkip_encrypt_skb(struct ieee80211_tx_data *tx, struct sk_buff *skb)
tx                196 net/mac80211/wpa.c 	struct ieee80211_key *key = tx->key;
tx                242 net/mac80211/wpa.c 	return ieee80211_tkip_encrypt_data(&tx->local->wep_tx_ctx,
tx                248 net/mac80211/wpa.c ieee80211_crypto_tkip_encrypt(struct ieee80211_tx_data *tx)
tx                252 net/mac80211/wpa.c 	ieee80211_tx_set_protected(tx);
tx                254 net/mac80211/wpa.c 	skb_queue_walk(&tx->skbs, skb) {
tx                255 net/mac80211/wpa.c 		if (tkip_encrypt_skb(tx, skb) < 0)
tx                405 net/mac80211/wpa.c static int ccmp_encrypt_skb(struct ieee80211_tx_data *tx, struct sk_buff *skb,
tx                409 net/mac80211/wpa.c 	struct ieee80211_key *key = tx->key;
tx                477 net/mac80211/wpa.c ieee80211_crypto_ccmp_encrypt(struct ieee80211_tx_data *tx,
tx                482 net/mac80211/wpa.c 	ieee80211_tx_set_protected(tx);
tx                484 net/mac80211/wpa.c 	skb_queue_walk(&tx->skbs, skb) {
tx                485 net/mac80211/wpa.c 		if (ccmp_encrypt_skb(tx, skb, mic_len) < 0)
tx                637 net/mac80211/wpa.c static int gcmp_encrypt_skb(struct ieee80211_tx_data *tx, struct sk_buff *skb)
tx                640 net/mac80211/wpa.c 	struct ieee80211_key *key = tx->key;
tx                708 net/mac80211/wpa.c ieee80211_crypto_gcmp_encrypt(struct ieee80211_tx_data *tx)
tx                712 net/mac80211/wpa.c 	ieee80211_tx_set_protected(tx);
tx                714 net/mac80211/wpa.c 	skb_queue_walk(&tx->skbs, skb) {
tx                715 net/mac80211/wpa.c 		if (gcmp_encrypt_skb(tx, skb) < 0)
tx                796 net/mac80211/wpa.c ieee80211_crypto_cs_encrypt(struct ieee80211_tx_data *tx,
tx                800 net/mac80211/wpa.c 	struct ieee80211_key *key = tx->key;
tx                933 net/mac80211/wpa.c ieee80211_crypto_aes_cmac_encrypt(struct ieee80211_tx_data *tx)
tx                937 net/mac80211/wpa.c 	struct ieee80211_key *key = tx->key;
tx                942 net/mac80211/wpa.c 	if (WARN_ON(skb_queue_len(&tx->skbs) != 1))
tx                945 net/mac80211/wpa.c 	skb = skb_peek(&tx->skbs);
tx                981 net/mac80211/wpa.c ieee80211_crypto_aes_cmac_256_encrypt(struct ieee80211_tx_data *tx)
tx                985 net/mac80211/wpa.c 	struct ieee80211_key *key = tx->key;
tx                990 net/mac80211/wpa.c 	if (WARN_ON(skb_queue_len(&tx->skbs) != 1))
tx                993 net/mac80211/wpa.c 	skb = skb_peek(&tx->skbs);
tx               1124 net/mac80211/wpa.c ieee80211_crypto_aes_gmac_encrypt(struct ieee80211_tx_data *tx)
tx               1128 net/mac80211/wpa.c 	struct ieee80211_key *key = tx->key;
tx               1135 net/mac80211/wpa.c 	if (WARN_ON(skb_queue_len(&tx->skbs) != 1))
tx               1138 net/mac80211/wpa.c 	skb = skb_peek(&tx->skbs);
tx               1233 net/mac80211/wpa.c ieee80211_crypto_hw_encrypt(struct ieee80211_tx_data *tx)
tx               1239 net/mac80211/wpa.c 	skb_queue_walk(&tx->skbs, skb) {
tx               1246 net/mac80211/wpa.c 		if (tx->key->flags & KEY_FLAG_CIPHER_SCHEME) {
tx               1247 net/mac80211/wpa.c 			res = ieee80211_crypto_cs_encrypt(tx, skb);
tx               1253 net/mac80211/wpa.c 	ieee80211_tx_set_protected(tx);
tx                 14 net/mac80211/wpa.h ieee80211_tx_h_michael_mic_add(struct ieee80211_tx_data *tx);
tx                 19 net/mac80211/wpa.h ieee80211_crypto_tkip_encrypt(struct ieee80211_tx_data *tx);
tx                 24 net/mac80211/wpa.h ieee80211_crypto_ccmp_encrypt(struct ieee80211_tx_data *tx,
tx                 31 net/mac80211/wpa.h ieee80211_crypto_aes_cmac_encrypt(struct ieee80211_tx_data *tx);
tx                 33 net/mac80211/wpa.h ieee80211_crypto_aes_cmac_256_encrypt(struct ieee80211_tx_data *tx);
tx                 39 net/mac80211/wpa.h ieee80211_crypto_aes_gmac_encrypt(struct ieee80211_tx_data *tx);
tx                 43 net/mac80211/wpa.h ieee80211_crypto_hw_encrypt(struct ieee80211_tx_data *tx);
tx                 48 net/mac80211/wpa.h ieee80211_crypto_gcmp_encrypt(struct ieee80211_tx_data *tx);
tx                177 net/nfc/nci/spi.c 	struct spi_transfer tx, rx;
tx                183 net/nfc/nci/spi.c 	memset(&tx, 0, sizeof(struct spi_transfer));
tx                186 net/nfc/nci/spi.c 	tx.tx_buf = req;
tx                187 net/nfc/nci/spi.c 	tx.len = 2;
tx                188 net/nfc/nci/spi.c 	tx.cs_change = 0;
tx                189 net/nfc/nci/spi.c 	tx.speed_hz = nspi->xfer_speed_hz;
tx                190 net/nfc/nci/spi.c 	spi_message_add_tail(&tx, &m);
tx                 57 net/tls/tls_device.c 		kfree(ctx->tx.rec_seq);
tx                 58 net/tls/tls_device.c 		kfree(ctx->tx.iv);
tx                217 net/tls/tls_device.c 	rcd_sn = tls_ctx->tx.rec_seq;
tx                273 net/tls/tls_device.c 	tls_advance_record_sn(sk, prot, &ctx->tx);
tx                967 net/tls/tls_device.c 	ctx->tx.iv = kmalloc(iv_size + TLS_CIPHER_AES_GCM_128_SALT_SIZE,
tx                969 net/tls/tls_device.c 	if (!ctx->tx.iv) {
tx                974 net/tls/tls_device.c 	memcpy(ctx->tx.iv + TLS_CIPHER_AES_GCM_128_SALT_SIZE, iv, iv_size);
tx                977 net/tls/tls_device.c 	ctx->tx.rec_seq = kmemdup(rec_seq, rec_seq_size, GFP_KERNEL);
tx                978 net/tls/tls_device.c 	if (!ctx->tx.rec_seq) {
tx                988 net/tls/tls_device.c 	memcpy(&rcd_sn, ctx->tx.rec_seq, sizeof(rcd_sn));
tx               1065 net/tls/tls_device.c 	kfree(ctx->tx.rec_seq);
tx               1067 net/tls/tls_device.c 	kfree(ctx->tx.iv);
tx                278 net/tls/tls_main.c 		kfree(ctx->tx.rec_seq);
tx                279 net/tls/tls_main.c 		kfree(ctx->tx.iv);
tx                376 net/tls/tls_main.c 		       ctx->tx.iv + TLS_CIPHER_AES_GCM_128_SALT_SIZE,
tx                378 net/tls/tls_main.c 		memcpy(crypto_info_aes_gcm_128->rec_seq, ctx->tx.rec_seq,
tx                400 net/tls/tls_main.c 		       ctx->tx.iv + TLS_CIPHER_AES_GCM_256_SALT_SIZE,
tx                402 net/tls/tls_main.c 		memcpy(crypto_info_aes_gcm_256->rec_seq, ctx->tx.rec_seq,
tx                448 net/tls/tls_main.c 				  unsigned int optlen, int tx)
tx                462 net/tls/tls_main.c 	if (tx) {
tx                526 net/tls/tls_main.c 	if (tx) {
tx                547 net/tls/tls_main.c 	if (tx)
tx                552 net/tls/tls_main.c 	if (tx) {
tx                502 net/tls/tls_sw.c 	memcpy(&rec->iv_data[iv_offset], tls_ctx->tx.iv,
tx                505 net/tls/tls_sw.c 	xor_iv_with_seq(prot->version, rec->iv_data, tls_ctx->tx.rec_seq);
tx                541 net/tls/tls_sw.c 	tls_advance_record_sn(sk, prot, &tls_ctx->tx);
tx                748 net/tls/tls_sw.c 		     tls_ctx->tx.rec_seq, prot->rec_seq_size,
tx               2258 net/tls/tls_sw.c int tls_set_sw_offload(struct sock *sk, struct tls_context *ctx, int tx)
tx               2282 net/tls/tls_sw.c 	if (tx) {
tx               2308 net/tls/tls_sw.c 	if (tx) {
tx               2312 net/tls/tls_sw.c 		cctx = &ctx->tx;
tx               2471 net/tls/tls_sw.c 	if (tx) {
tx                605 net/wireless/mlme.c 	if (!(wdev->wiphy->mgmt_stypes[wdev->iftype].tx & BIT(stype >> 4)))
tx               1678 net/wireless/nl80211.c 		stypes = mgmt_stypes[ift].tx;
tx               12297 net/wireless/nl80211.c 			     bool tx)
tx               12319 net/wireless/nl80211.c 	if (tx) {
tx               1590 net/wireless/trace.h 	TP_PROTO(struct wiphy *wiphy, int ret, u32 tx, u32 rx),
tx               1591 net/wireless/trace.h 	TP_ARGS(wiphy, ret, tx, rx),
tx               1595 net/wireless/trace.h 		__field(u32, tx)
tx               1601 net/wireless/trace.h 		__entry->tx = tx;
tx               1605 net/wireless/trace.h 		  WIPHY_PR_ARG, __entry->ret, __entry->tx, __entry->rx)
tx               1609 net/wireless/trace.h 	TP_PROTO(struct wiphy *wiphy, u32 tx, u32 tx_max,
tx               1611 net/wireless/trace.h 	TP_ARGS(wiphy, tx, tx_max, rx, rx_max),
tx               1614 net/wireless/trace.h 		__field(u32, tx)
tx               1621 net/wireless/trace.h 		__entry->tx = tx;
tx               1627 net/wireless/trace.h 		  WIPHY_PR_ARG, __entry->tx, __entry->tx_max, __entry->rx,
tx               1632 net/wireless/trace.h 	TP_PROTO(struct wiphy *wiphy, u32 tx, u32 rx),
tx               1633 net/wireless/trace.h 	TP_ARGS(wiphy, rx, tx),
tx               1636 net/wireless/trace.h 		__field(u32, tx)
tx               1641 net/wireless/trace.h 		__entry->tx = tx;
tx               1645 net/wireless/trace.h 		  WIPHY_PR_ARG, __entry->tx, __entry->rx)
tx               1649 net/wireless/trace.h 	TP_PROTO(struct wiphy *wiphy, u32 tx, u32 rx),
tx               1650 net/wireless/trace.h 	TP_ARGS(wiphy, rx, tx)
tx                 30 net/xdp/xdp_umem.c 	if (!xs->tx)
tx                 42 net/xdp/xdp_umem.c 	if (!xs->tx)
tx                 77 net/xdp/xsk.c  		xs->tx->ring->flags |= XDP_RING_NEED_WAKEUP;
tx                104 net/xdp/xsk.c  		xs->tx->ring->flags &= ~XDP_RING_NEED_WAKEUP;
tx                292 net/xdp/xsk.c  		if (!xskq_peek_desc(xs->tx, desc, umem))
tx                298 net/xdp/xsk.c  		xskq_discard_desc(xs->tx);
tx                353 net/xdp/xsk.c  	while (xskq_peek_desc(xs->tx, &desc, xs->umem)) {
tx                386 net/xdp/xsk.c  		xskq_discard_desc(xs->tx);
tx                411 net/xdp/xsk.c  	if (unlikely(!xs->tx))
tx                454 net/xdp/xsk.c  	if (xs->tx && !xskq_full_desc(xs->tx))
tx                564 net/xdp/xsk.c  	xskq_destroy(xs->tx);
tx                642 net/xdp/xsk.c  	if (!xs->rx && !xs->tx) {
tx                708 net/xdp/xsk.c  	xskq_set_umem(xs->tx, xs->umem->size, xs->umem->chunk_mask);
tx                761 net/xdp/xsk.c  		q = (optname == XDP_TX_RING) ? &xs->tx : &xs->rx;
tx                765 net/xdp/xsk.c  			xs->tx->ring->flags |= XDP_RING_NEED_WAKEUP;
tx                873 net/xdp/xsk.c  		stats.tx_invalid_descs = xskq_nb_invalid_descs(xs->tx);
tx                902 net/xdp/xsk.c  					       &off.tx);
tx                909 net/xdp/xsk.c  			off.tx.flags = offsetof(struct xdp_rxtx_ring,
tx                920 net/xdp/xsk.c  			xsk_enter_rxtx_offsets(&off_v1.tx);
tx                979 net/xdp/xsk.c  		q = READ_ONCE(xs->tx);
tx                 15 net/xdp/xsk.h  	struct xdp_ring_offset_v1 tx;
tx                 42 net/xdp/xsk_diag.c 	if (!err && xs->tx)
tx                 43 net/xdp/xsk_diag.c 		err = xsk_diag_put_ring(xs->tx, XDP_DIAG_TX_RING, nlskb);
tx                 89 samples/bpf/xdpsock_user.c 	struct xsk_ring_prod tx;
tx                329 samples/bpf/xdpsock_user.c 				 &xsk->rx, &xsk->tx, &cfg);
tx                497 samples/bpf/xdpsock_user.c 	if (!opt_need_wakeup || xsk_ring_prod__needs_wakeup(&xsk->tx))
tx                537 samples/bpf/xdpsock_user.c 	if (!opt_need_wakeup || xsk_ring_prod__needs_wakeup(&xsk->tx))
tx                615 samples/bpf/xdpsock_user.c 	if (xsk_ring_prod__reserve(&xsk->tx, BATCH_SIZE, &idx) == BATCH_SIZE) {
tx                619 samples/bpf/xdpsock_user.c 			xsk_ring_prod__tx_desc(&xsk->tx, idx + i)->addr	=
tx                621 samples/bpf/xdpsock_user.c 			xsk_ring_prod__tx_desc(&xsk->tx, idx + i)->len =
tx                625 samples/bpf/xdpsock_user.c 		xsk_ring_prod__submit(&xsk->tx, BATCH_SIZE);
tx                676 samples/bpf/xdpsock_user.c 	ret = xsk_ring_prod__reserve(&xsk->tx, rcvd, &idx_tx);
tx                680 samples/bpf/xdpsock_user.c 		if (xsk_ring_prod__needs_wakeup(&xsk->tx))
tx                682 samples/bpf/xdpsock_user.c 		ret = xsk_ring_prod__reserve(&xsk->tx, rcvd, &idx_tx);
tx                696 samples/bpf/xdpsock_user.c 		xsk_ring_prod__tx_desc(&xsk->tx, idx_tx)->addr = orig;
tx                697 samples/bpf/xdpsock_user.c 		xsk_ring_prod__tx_desc(&xsk->tx, idx_tx++)->len = len;
tx                700 samples/bpf/xdpsock_user.c 	xsk_ring_prod__submit(&xsk->tx, rcvd);
tx                377 samples/mic/mpssd/mpssd.c static inline void txrx_prepare(int type, bool tx, struct mic_vring *vr,
tx                380 samples/mic/mpssd/mpssd.c 	copy->vr_idx = tx ? 0 : 1;
tx                890 scripts/kconfig/gconf.c 	gint tx = (gint) event->x;
tx                894 scripts/kconfig/gconf.c 	gtk_tree_view_get_path_at_pos(view, tx, ty, &path, &column, &cx,
tx               1006 scripts/kconfig/gconf.c 	gint tx = (gint) event->x;
tx               1010 scripts/kconfig/gconf.c 	gtk_tree_view_get_path_at_pos(view, tx, ty, &path, &column, &cx,
tx                472 sound/firewire/amdtp-stream.c 	params->header_length = s->ctx_data.tx.ctx_header_size;
tx                473 sound/firewire/amdtp-stream.c 	params->payload_length = s->ctx_data.tx.max_ctx_payload_length;
tx                590 sound/firewire/amdtp-stream.c 		if (*data_blocks > 0 && s->ctx_data.tx.dbc_interval > 0)
tx                591 sound/firewire/amdtp-stream.c 			dbc_interval = s->ctx_data.tx.dbc_interval;
tx                623 sound/firewire/amdtp-stream.c 	if (*payload_length > s->ctx_data.tx.ctx_header_size +
tx                624 sound/firewire/amdtp-stream.c 					s->ctx_data.tx.max_ctx_payload_length) {
tx                627 sound/firewire/amdtp-stream.c 			*payload_length, s->ctx_data.tx.max_ctx_payload_length);
tx                715 sound/firewire/amdtp-stream.c 			s->ctx_data.tx.ctx_header_size / sizeof(*ctx_header);
tx                832 sound/firewire/amdtp-stream.c 	packets = header_length / s->ctx_data.tx.ctx_header_size;
tx                975 sound/firewire/amdtp-stream.c 		s->ctx_data.tx.max_ctx_payload_length = max_ctx_payload_size;
tx                976 sound/firewire/amdtp-stream.c 		s->ctx_data.tx.ctx_header_size = ctx_header_size;
tx                134 sound/firewire/amdtp-stream.h 		} tx;
tx                 98 sound/firewire/dice/dice-proc.c 		} tx;
tx                172 sound/firewire/dice/dice-proc.c 	quadlets = min_t(u32, tx_rx_header.size, sizeof(buf.tx) / 4);
tx                174 sound/firewire/dice/dice-proc.c 		if (dice_proc_read_mem(dice, &buf.tx, sections[2] + 2 +
tx                179 sound/firewire/dice/dice-proc.c 		snd_iprintf(buffer, "  iso channel: %d\n", (int)buf.tx.iso);
tx                181 sound/firewire/dice/dice-proc.c 			    buf.tx.number_audio);
tx                182 sound/firewire/dice/dice-proc.c 		snd_iprintf(buffer, "  midi ports: %u\n", buf.tx.number_midi);
tx                183 sound/firewire/dice/dice-proc.c 		snd_iprintf(buffer, "  speed: S%u\n", 100u << buf.tx.speed);
tx                185 sound/firewire/dice/dice-proc.c 			dice_proc_fixup_string(buf.tx.names, TX_NAMES_SIZE);
tx                186 sound/firewire/dice/dice-proc.c 			snd_iprintf(buffer, "  names: %s\n", buf.tx.names);
tx                190 sound/firewire/dice/dice-proc.c 				    buf.tx.ac3_caps);
tx                192 sound/firewire/dice/dice-proc.c 				    buf.tx.ac3_enable);
tx                 58 sound/firewire/fireworks/fireworks_stream.c 			efw->tx_stream.ctx_data.tx.dbc_interval = 8;
tx                154 sound/soc/bcm/bcm2835-i2s.c 				    bool tx, bool rx)
tx                164 sound/soc/bcm/bcm2835-i2s.c 	off =  tx ? BCM2835_I2S_TXON : 0;
tx                167 sound/soc/bcm/bcm2835-i2s.c 	clr =  tx ? BCM2835_I2S_TXCLR : 0;
tx                266 sound/soc/codecs/cs42xx8.c 	bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx                273 sound/soc/codecs/cs42xx8.c 	if (tx)
tx                276 sound/soc/codecs/cs42xx8.c 	rate[tx]  = params_rate(params);
tx                277 sound/soc/codecs/cs42xx8.c 	rate[!tx] = cs42xx8->rate[!tx];
tx                279 sound/soc/codecs/cs42xx8.c 	ratio[tx] = rate[tx] > 0 ? cs42xx8->sysclk / rate[tx] : 0;
tx                280 sound/soc/codecs/cs42xx8.c 	ratio[!tx] = rate[!tx] > 0 ? cs42xx8->sysclk / rate[!tx] : 0;
tx                303 sound/soc/codecs/cs42xx8.c 		condition1 = ((fm[tx] == CS42XX8_FM_AUTO) ?
tx                304 sound/soc/codecs/cs42xx8.c 			(cs42xx8_ratios[i].ratio[0] == ratio[tx] ||
tx                305 sound/soc/codecs/cs42xx8.c 			cs42xx8_ratios[i].ratio[1] == ratio[tx] ||
tx                306 sound/soc/codecs/cs42xx8.c 			cs42xx8_ratios[i].ratio[2] == ratio[tx]) :
tx                307 sound/soc/codecs/cs42xx8.c 			(cs42xx8_ratios[i].ratio[fm[tx]] == ratio[tx])) &&
tx                311 sound/soc/codecs/cs42xx8.c 		if (!ratio[tx])
tx                315 sound/soc/codecs/cs42xx8.c 		condition2 = ((fm[!tx] == CS42XX8_FM_AUTO) ?
tx                316 sound/soc/codecs/cs42xx8.c 			(cs42xx8_ratios[i].ratio[0] == ratio[!tx] ||
tx                317 sound/soc/codecs/cs42xx8.c 			cs42xx8_ratios[i].ratio[1] == ratio[!tx] ||
tx                318 sound/soc/codecs/cs42xx8.c 			cs42xx8_ratios[i].ratio[2] == ratio[!tx]) :
tx                319 sound/soc/codecs/cs42xx8.c 			(cs42xx8_ratios[i].ratio[fm[!tx]] == ratio[!tx]));
tx                321 sound/soc/codecs/cs42xx8.c 		if (!ratio[!tx])
tx                337 sound/soc/codecs/cs42xx8.c 	cs42xx8->rate[tx] = params_rate(params);
tx                343 sound/soc/codecs/cs42xx8.c 			   CS42XX8_FUNCMOD_xC_FM_MASK(tx) | mask,
tx                344 sound/soc/codecs/cs42xx8.c 			   CS42XX8_FUNCMOD_xC_FM(tx, fm[tx]) | val);
tx                354 sound/soc/codecs/cs42xx8.c 	bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx                357 sound/soc/codecs/cs42xx8.c 	cs42xx8->rate[tx] = 0;
tx                360 sound/soc/codecs/cs42xx8.c 			   CS42XX8_FUNCMOD_xC_FM_MASK(tx),
tx                361 sound/soc/codecs/cs42xx8.c 			   CS42XX8_FUNCMOD_xC_FM(tx, CS42XX8_FM_AUTO));
tx               3972 sound/soc/codecs/wcd9335.c 	bool tx;
tx               3982 sound/soc/codecs/wcd9335.c 		tx = (j >= 16 ? true : false);
tx               3983 sound/soc/codecs/wcd9335.c 		port_id = (tx ? j - 16 : j);
tx               3987 sound/soc/codecs/wcd9335.c 			if (!tx)
tx               4006 sound/soc/codecs/wcd9335.c 			   __func__, (tx ? "TX" : "RX"), port_id, val);
tx               4011 sound/soc/codecs/wcd9335.c 			   __func__, (tx ? "TX" : "RX"), port_id, val);
tx               4015 sound/soc/codecs/wcd9335.c 			if (!tx)
tx                816 sound/soc/codecs/wm8960.c 	bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx                848 sound/soc/codecs/wm8960.c 	if (tx) {
tx                861 sound/soc/codecs/wm8960.c 	wm8960->is_stream_in_use[tx] = true;
tx                863 sound/soc/codecs/wm8960.c 	if (!wm8960->is_stream_in_use[!tx])
tx                874 sound/soc/codecs/wm8960.c 	bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx                876 sound/soc/codecs/wm8960.c 	wm8960->is_stream_in_use[tx] = false;
tx                143 sound/soc/fsl/fsl-asoc-card.c 	bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx                162 sound/soc/fsl/fsl-asoc-card.c 	ret = snd_soc_dai_set_sysclk(rtd->cpu_dai, cpu_priv->sysclk_id[tx],
tx                163 sound/soc/fsl/fsl-asoc-card.c 				     cpu_priv->sysclk_freq[tx],
tx                164 sound/soc/fsl/fsl-asoc-card.c 				     cpu_priv->sysclk_dir[tx]);
tx                134 sound/soc/fsl/fsl_asrc_dma.c 	bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx                149 sound/soc/fsl/fsl_asrc_dma.c 	u8 dir = tx ? OUT : IN;
tx                203 sound/soc/fsl/fsl_asrc_dma.c 	tmp_chan = dma_request_slave_channel(dev_be, tx ? "tx" : "rx");
tx                233 sound/soc/fsl/fsl_asrc_dma.c 	if (tx) {
tx                138 sound/soc/fsl/fsl_esai.c static int fsl_esai_divisor_cal(struct snd_soc_dai *dai, bool tx, u32 ratio,
tx                207 sound/soc/fsl/fsl_esai.c 	regmap_update_bits(esai_priv->regmap, REG_ESAI_xCCR(tx),
tx                216 sound/soc/fsl/fsl_esai.c 	regmap_update_bits(esai_priv->regmap, REG_ESAI_xCCR(tx),
tx                238 sound/soc/fsl/fsl_esai.c 	bool tx = (clk_id <= ESAI_HCKT_EXTAL || esai_priv->synchronous);
tx                246 sound/soc/fsl/fsl_esai.c 			in ? "in" : "out", tx ? 'T' : 'R');
tx                251 sound/soc/fsl/fsl_esai.c 	if (freq == esai_priv->hck_rate[tx] && dir == esai_priv->hck_dir[tx])
tx                255 sound/soc/fsl/fsl_esai.c 	esai_priv->sck_div[tx] = true;
tx                258 sound/soc/fsl/fsl_esai.c 	regmap_update_bits(esai_priv->regmap, REG_ESAI_xCCR(tx),
tx                297 sound/soc/fsl/fsl_esai.c 				tx ? 'T' : 'R');
tx                304 sound/soc/fsl/fsl_esai.c 		ecr |= tx ? ESAI_ECR_ETO : ESAI_ECR_ERO;
tx                309 sound/soc/fsl/fsl_esai.c 				tx ? 'T' : 'R');
tx                313 sound/soc/fsl/fsl_esai.c 	ret = fsl_esai_divisor_cal(dai, tx, ratio, false, 0);
tx                317 sound/soc/fsl/fsl_esai.c 	esai_priv->sck_div[tx] = false;
tx                320 sound/soc/fsl/fsl_esai.c 	esai_priv->hck_dir[tx] = dir;
tx                321 sound/soc/fsl/fsl_esai.c 	esai_priv->hck_rate[tx] = freq;
tx                324 sound/soc/fsl/fsl_esai.c 			   tx ? ESAI_ECR_ETI | ESAI_ECR_ETO :
tx                333 sound/soc/fsl/fsl_esai.c static int fsl_esai_set_bclk(struct snd_soc_dai *dai, bool tx, u32 freq)
tx                336 sound/soc/fsl/fsl_esai.c 	u32 hck_rate = esai_priv->hck_rate[tx];
tx                341 sound/soc/fsl/fsl_esai.c 	if (esai_priv->slave_mode || esai_priv->sck_rate[tx] == freq)
tx                354 sound/soc/fsl/fsl_esai.c 				tx ? 'T' : 'R');
tx                359 sound/soc/fsl/fsl_esai.c 	if (!esai_priv->sck_div[tx] && (ratio > 16 || ratio == 0)) {
tx                364 sound/soc/fsl/fsl_esai.c 	ret = fsl_esai_divisor_cal(dai, tx, ratio, true,
tx                365 sound/soc/fsl/fsl_esai.c 			esai_priv->sck_div[tx] ? 0 : ratio);
tx                370 sound/soc/fsl/fsl_esai.c 	esai_priv->sck_rate[tx] = freq;
tx                509 sound/soc/fsl/fsl_esai.c 	bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx                523 sound/soc/fsl/fsl_esai.c 	ret = fsl_esai_set_bclk(dai, esai_priv->synchronous || tx, bclk);
tx                530 sound/soc/fsl/fsl_esai.c 	regmap_update_bits(esai_priv->regmap, REG_ESAI_xCR(tx), mask, val);
tx                532 sound/soc/fsl/fsl_esai.c 	if (!tx && esai_priv->synchronous)
tx                536 sound/soc/fsl/fsl_esai.c 	regmap_update_bits(esai_priv->regmap, REG_ESAI_xCR(tx),
tx                540 sound/soc/fsl/fsl_esai.c 	regmap_update_bits(esai_priv->regmap, REG_ESAI_xFCR(tx),
tx                544 sound/soc/fsl/fsl_esai.c 	      (tx ? ESAI_xFCR_TE_MASK | ESAI_xFCR_TIEN : ESAI_xFCR_RE_MASK);
tx                546 sound/soc/fsl/fsl_esai.c 	     (tx ? ESAI_xFCR_TE(pins) | ESAI_xFCR_TIEN : ESAI_xFCR_RE(pins));
tx                548 sound/soc/fsl/fsl_esai.c 	regmap_update_bits(esai_priv->regmap, REG_ESAI_xFCR(tx), mask, val);
tx                550 sound/soc/fsl/fsl_esai.c 	if (tx)
tx                618 sound/soc/fsl/fsl_esai.c static void fsl_esai_trigger_start(struct fsl_esai *esai_priv, bool tx)
tx                620 sound/soc/fsl/fsl_esai.c 	u8 i, channels = esai_priv->channels[tx];
tx                624 sound/soc/fsl/fsl_esai.c 	regmap_update_bits(esai_priv->regmap, REG_ESAI_xFCR(tx),
tx                628 sound/soc/fsl/fsl_esai.c 	for (i = 0; tx && i < channels; i++)
tx                643 sound/soc/fsl/fsl_esai.c 	regmap_update_bits(esai_priv->regmap, REG_ESAI_xCR(tx),
tx                644 sound/soc/fsl/fsl_esai.c 			   tx ? ESAI_xCR_TE_MASK : ESAI_xCR_RE_MASK,
tx                645 sound/soc/fsl/fsl_esai.c 			   tx ? ESAI_xCR_TE(pins) : ESAI_xCR_RE(pins));
tx                646 sound/soc/fsl/fsl_esai.c 	mask = tx ? esai_priv->tx_mask : esai_priv->rx_mask;
tx                648 sound/soc/fsl/fsl_esai.c 	regmap_update_bits(esai_priv->regmap, REG_ESAI_xSMB(tx),
tx                650 sound/soc/fsl/fsl_esai.c 	regmap_update_bits(esai_priv->regmap, REG_ESAI_xSMA(tx),
tx                654 sound/soc/fsl/fsl_esai.c 	regmap_update_bits(esai_priv->regmap, REG_ESAI_xCR(tx),
tx                658 sound/soc/fsl/fsl_esai.c static void fsl_esai_trigger_stop(struct fsl_esai *esai_priv, bool tx)
tx                660 sound/soc/fsl/fsl_esai.c 	regmap_update_bits(esai_priv->regmap, REG_ESAI_xCR(tx),
tx                663 sound/soc/fsl/fsl_esai.c 	regmap_update_bits(esai_priv->regmap, REG_ESAI_xCR(tx),
tx                664 sound/soc/fsl/fsl_esai.c 			   tx ? ESAI_xCR_TE_MASK : ESAI_xCR_RE_MASK, 0);
tx                665 sound/soc/fsl/fsl_esai.c 	regmap_update_bits(esai_priv->regmap, REG_ESAI_xSMA(tx),
tx                667 sound/soc/fsl/fsl_esai.c 	regmap_update_bits(esai_priv->regmap, REG_ESAI_xSMB(tx),
tx                671 sound/soc/fsl/fsl_esai.c 	regmap_update_bits(esai_priv->regmap, REG_ESAI_xFCR(tx),
tx                673 sound/soc/fsl/fsl_esai.c 	regmap_update_bits(esai_priv->regmap, REG_ESAI_xFCR(tx),
tx                680 sound/soc/fsl/fsl_esai.c 	bool tx = true, rx = false, enabled[2];
tx                688 sound/soc/fsl/fsl_esai.c 	enabled[tx] = tfcr & ESAI_xFCR_xFEN;
tx                692 sound/soc/fsl/fsl_esai.c 	fsl_esai_trigger_stop(esai_priv, tx);
tx                718 sound/soc/fsl/fsl_esai.c 	if (enabled[tx])
tx                719 sound/soc/fsl/fsl_esai.c 		fsl_esai_trigger_start(esai_priv, tx);
tx                730 sound/soc/fsl/fsl_esai.c 	bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx                733 sound/soc/fsl/fsl_esai.c 	esai_priv->channels[tx] = substream->runtime->channels;
tx                740 sound/soc/fsl/fsl_esai.c 		fsl_esai_trigger_start(esai_priv, tx);
tx                747 sound/soc/fsl/fsl_esai.c 		fsl_esai_trigger_stop(esai_priv, tx);
tx                 22 sound/soc/fsl/fsl_esai.h #define REG_ESAI_xFCR(tx)	(tx ? REG_ESAI_TFCR : REG_ESAI_RFCR)
tx                 23 sound/soc/fsl/fsl_esai.h #define REG_ESAI_xFSR(tx)	(tx ? REG_ESAI_TFSR : REG_ESAI_RFSR)
tx                 41 sound/soc/fsl/fsl_esai.h #define REG_ESAI_xCR(tx)	(tx ? REG_ESAI_TCR : REG_ESAI_RCR)
tx                 42 sound/soc/fsl/fsl_esai.h #define REG_ESAI_xCCR(tx)	(tx ? REG_ESAI_TCCR : REG_ESAI_RCCR)
tx                 47 sound/soc/fsl/fsl_esai.h #define REG_ESAI_xSMA(tx)	(tx ? REG_ESAI_TSMA : REG_ESAI_RSMA)
tx                 48 sound/soc/fsl/fsl_esai.h #define REG_ESAI_xSMB(tx)	(tx ? REG_ESAI_TSMB : REG_ESAI_RSMB)
tx                155 sound/soc/fsl/fsl_sai.c 	bool tx = fsl_dir == FSL_FMT_TRANSMITTER;
tx                175 sound/soc/fsl/fsl_sai.c 	regmap_update_bits(sai->regmap, FSL_SAI_xCR2(tx, ofs),
tx                209 sound/soc/fsl/fsl_sai.c 	bool tx = fsl_dir == FSL_FMT_TRANSMITTER;
tx                303 sound/soc/fsl/fsl_sai.c 	regmap_update_bits(sai->regmap, FSL_SAI_xCR2(tx, ofs),
tx                305 sound/soc/fsl/fsl_sai.c 	regmap_update_bits(sai->regmap, FSL_SAI_xCR4(tx, ofs),
tx                329 sound/soc/fsl/fsl_sai.c static int fsl_sai_set_bclk(struct snd_soc_dai *dai, bool tx, u32 freq)
tx                369 sound/soc/fsl/fsl_sai.c 			sai->mclk_id[tx] = id;
tx                379 sound/soc/fsl/fsl_sai.c 				tx ? 'T' : 'R', freq);
tx                394 sound/soc/fsl/fsl_sai.c 	    (!tx && !sai->synchronous[RX])) {
tx                397 sound/soc/fsl/fsl_sai.c 				   FSL_SAI_CR2_MSEL(sai->mclk_id[tx]));
tx                401 sound/soc/fsl/fsl_sai.c 		   (tx && !sai->synchronous[TX])) {
tx                404 sound/soc/fsl/fsl_sai.c 				   FSL_SAI_CR2_MSEL(sai->mclk_id[tx]));
tx                410 sound/soc/fsl/fsl_sai.c 			sai->mclk_id[tx], savediv, savesub);
tx                421 sound/soc/fsl/fsl_sai.c 	bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx                437 sound/soc/fsl/fsl_sai.c 			ret = fsl_sai_set_bclk(cpu_dai, tx,
tx                441 sound/soc/fsl/fsl_sai.c 			ret = fsl_sai_set_bclk(cpu_dai, tx,
tx                449 sound/soc/fsl/fsl_sai.c 			ret = clk_prepare_enable(sai->mclk_clk[sai->mclk_id[tx]]);
tx                478 sound/soc/fsl/fsl_sai.c 		if (!sai->synchronous[TX] && sai->synchronous[RX] && !tx) {
tx                487 sound/soc/fsl/fsl_sai.c 		} else if (!sai->synchronous[RX] && sai->synchronous[TX] && tx) {
tx                499 sound/soc/fsl/fsl_sai.c 	regmap_update_bits(sai->regmap, FSL_SAI_xCR4(tx, ofs),
tx                502 sound/soc/fsl/fsl_sai.c 	regmap_update_bits(sai->regmap, FSL_SAI_xCR5(tx, ofs),
tx                505 sound/soc/fsl/fsl_sai.c 	regmap_write(sai->regmap, FSL_SAI_xMR(tx), ~0UL - ((1 << channels) - 1));
tx                514 sound/soc/fsl/fsl_sai.c 	bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx                518 sound/soc/fsl/fsl_sai.c 		clk_disable_unprepare(sai->mclk_clk[sai->mclk_id[tx]]);
tx                532 sound/soc/fsl/fsl_sai.c 	bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx                553 sound/soc/fsl/fsl_sai.c 		regmap_update_bits(sai->regmap, FSL_SAI_xCSR(tx, ofs),
tx                561 sound/soc/fsl/fsl_sai.c 		regmap_update_bits(sai->regmap, FSL_SAI_xCSR(tx, ofs),
tx                567 sound/soc/fsl/fsl_sai.c 		regmap_update_bits(sai->regmap, FSL_SAI_xCSR(tx, ofs),
tx                569 sound/soc/fsl/fsl_sai.c 		regmap_update_bits(sai->regmap, FSL_SAI_xCSR(tx, ofs),
tx                573 sound/soc/fsl/fsl_sai.c 		regmap_read(sai->regmap, FSL_SAI_xCSR(!tx, ofs), &xcsr);
tx                585 sound/soc/fsl/fsl_sai.c 					    FSL_SAI_xCSR(tx, ofs), &xcsr);
tx                624 sound/soc/fsl/fsl_sai.c 	bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx                627 sound/soc/fsl/fsl_sai.c 	regmap_update_bits(sai->regmap, FSL_SAI_xCR3(tx, ofs),
tx                638 sound/soc/fsl/fsl_sai.c 					   tx ? sai->dma_params_tx.maxburst :
tx                652 sound/soc/fsl/fsl_sai.c 	bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx                654 sound/soc/fsl/fsl_sai.c 	regmap_update_bits(sai->regmap, FSL_SAI_xCR3(tx, ofs),
tx                 64 sound/soc/fsl/fsl_sai.h #define FSL_SAI_xCSR(tx, ofs)	(tx ? FSL_SAI_TCSR(ofs) : FSL_SAI_RCSR(ofs))
tx                 65 sound/soc/fsl/fsl_sai.h #define FSL_SAI_xCR1(tx, ofs)	(tx ? FSL_SAI_TCR1(ofs) : FSL_SAI_RCR1(ofs))
tx                 66 sound/soc/fsl/fsl_sai.h #define FSL_SAI_xCR2(tx, ofs)	(tx ? FSL_SAI_TCR2(ofs) : FSL_SAI_RCR2(ofs))
tx                 67 sound/soc/fsl/fsl_sai.h #define FSL_SAI_xCR3(tx, ofs)	(tx ? FSL_SAI_TCR3(ofs) : FSL_SAI_RCR3(ofs))
tx                 68 sound/soc/fsl/fsl_sai.h #define FSL_SAI_xCR4(tx, ofs)	(tx ? FSL_SAI_TCR4(ofs) : FSL_SAI_RCR4(ofs))
tx                 69 sound/soc/fsl/fsl_sai.h #define FSL_SAI_xCR5(tx, ofs)	(tx ? FSL_SAI_TCR5(ofs) : FSL_SAI_RCR5(ofs))
tx                 70 sound/soc/fsl/fsl_sai.h #define FSL_SAI_xDR(tx, ofs)	(tx ? FSL_SAI_TDR(ofs) : FSL_SAI_RDR(ofs))
tx                 71 sound/soc/fsl/fsl_sai.h #define FSL_SAI_xFR(tx, ofs)	(tx ? FSL_SAI_TFR(ofs) : FSL_SAI_RFR(ofs))
tx                 72 sound/soc/fsl/fsl_sai.h #define FSL_SAI_xMR(tx)		(tx ? FSL_SAI_TMR : FSL_SAI_RMR)
tx                 36 sound/soc/fsl/fsl_spdif.c #define SIE_INTR_FOR(tx)	(tx ? INTR_FOR_PLAYBACK : INTR_FOR_CAPTURE)
tx                602 sound/soc/fsl/fsl_spdif.c 	bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx                603 sound/soc/fsl/fsl_spdif.c 	u32 intr = SIE_INTR_FOR(tx);
tx                604 sound/soc/fsl/fsl_spdif.c 	u32 dmaen = SCR_DMA_xX_EN(tx);
tx                 93 sound/soc/fsl/fsl_spdif.h #define SCR_DMA_xX_EN(tx)		(tx ? SCR_DMA_TX_EN : SCR_DMA_RX_EN)
tx                405 sound/soc/fsl/fsl_ssi.c static void fsl_ssi_config_enable(struct fsl_ssi *ssi, bool tx)
tx                408 sound/soc/fsl/fsl_ssi.c 	int dir = tx ? TX : RX;
tx                413 sound/soc/fsl/fsl_ssi.c 			   SSI_SOR_xX_CLR(tx), SSI_SOR_xX_CLR(tx));
tx                450 sound/soc/fsl/fsl_ssi.c 	if (ssi->use_dma && tx) {
tx                505 sound/soc/fsl/fsl_ssi.c static void fsl_ssi_config_disable(struct fsl_ssi *ssi, bool tx)
tx                509 sound/soc/fsl/fsl_ssi.c 	int adir = tx ? RX : TX;
tx                510 sound/soc/fsl/fsl_ssi.c 	int dir = tx ? TX : RX;
tx                563 sound/soc/fsl/fsl_ssi.c 			   SSI_SOR_xX_CLR(tx), SSI_SOR_xX_CLR(tx));
tx                676 sound/soc/fsl/fsl_ssi.c 	bool tx2, tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx                765 sound/soc/fsl/fsl_ssi.c 	tx2 = tx || ssi->synchronous;
tx                794 sound/soc/fsl/fsl_ssi.c 	bool tx2, tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx                847 sound/soc/fsl/fsl_ssi.c 	tx2 = tx || ssi->synchronous;
tx               1063 sound/soc/fsl/fsl_ssi.c 	bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx               1076 sound/soc/fsl/fsl_ssi.c 		if (tx && fsl_ssi_is_ac97(ssi))
tx               1078 sound/soc/fsl/fsl_ssi.c 		fsl_ssi_config_enable(ssi, tx);
tx               1084 sound/soc/fsl/fsl_ssi.c 		fsl_ssi_config_disable(ssi, tx);
tx                 33 sound/soc/fsl/fsl_ssi.h #define REG_SSI_SxCR(tx)		((tx) ? REG_SSI_STCR : REG_SSI_SRCR)
tx                 38 sound/soc/fsl/fsl_ssi.h #define REG_SSI_SxCCR(tx)		((tx) ? REG_SSI_STCCR : REG_SSI_SRCCR)
tx                 67 sound/soc/fsl/fsl_ssi.h #define REG_SSI_SxMSK(tx)		((tx) ? REG_SSI_STMSK : REG_SSI_SRMSK)
tx                250 sound/soc/fsl/fsl_ssi.h #define SSI_SOR_xX_CLR(tx)		((tx) ? SSI_SOR_TX_CLR : SSI_SOR_RX_CLR)
tx                 78 sound/soc/fsl/imx-audmix.c 	bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx                 84 sound/soc/fsl/imx-audmix.c 	fmt |= tx ? SND_SOC_DAIFMT_CBS_CFS : SND_SOC_DAIFMT_CBM_CFM;
tx                 85 sound/soc/fsl/imx-audmix.c 	dir  = tx ? SND_SOC_CLOCK_OUT : SND_SOC_CLOCK_IN;
tx                117 sound/soc/fsl/imx-audmix.c 	bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
tx                121 sound/soc/fsl/imx-audmix.c 	if (!tx)
tx                158 sound/soc/intel/atom/sst-atom-controls.c 	unsigned int is_tx = e->tx;
tx                217 sound/soc/intel/atom/sst-atom-controls.c 	unsigned int is_tx = e->tx;
tx                838 sound/soc/intel/atom/sst-atom-controls.h 	bool tx;
tx                847 sound/soc/intel/atom/sst-atom-controls.h 	(struct sst_enum){ .reg = s_ch_no, .tx = is_tx, .max = 4+1, .texts = xtexts, }
tx                214 sound/soc/intel/baytrail/sst-baytrail-ipc.c 	u64 header = msg->tx.header;
tx                625 sound/soc/intel/baytrail/sst-baytrail-ipc.c 	if (msg->tx.header & IPC_HEADER_LARGE(true))
tx                626 sound/soc/intel/baytrail/sst-baytrail-ipc.c 		sst_dsp_outbox_write(ipc->dsp, msg->tx.data, msg->tx.size);
tx                628 sound/soc/intel/baytrail/sst-baytrail-ipc.c 	sst_dsp_shim_write64_unlocked(ipc->dsp, SST_IPCX, msg->tx.header);
tx                650 sound/soc/intel/baytrail/sst-baytrail-ipc.c 	*(u32 *)msg->tx.data = (u32)(msg->tx.header & (u32)-1);
tx                651 sound/soc/intel/baytrail/sst-baytrail-ipc.c 	memcpy(msg->tx.data + sizeof(u32), tx_data, tx_size);
tx                652 sound/soc/intel/baytrail/sst-baytrail-ipc.c 	msg->tx.size += sizeof(u32);
tx                 93 sound/soc/intel/common/sst-ipc.c 	msg->tx.header = request.header;
tx                 94 sound/soc/intel/common/sst-ipc.c 	msg->tx.size = request.size;
tx                125 sound/soc/intel/common/sst-ipc.c 		ipc->msg[i].tx.data = kzalloc(ipc->tx_data_max_size, GFP_KERNEL);
tx                126 sound/soc/intel/common/sst-ipc.c 		if (ipc->msg[i].tx.data == NULL)
tx                131 sound/soc/intel/common/sst-ipc.c 			kfree(ipc->msg[i].tx.data);
tx                143 sound/soc/intel/common/sst-ipc.c 		kfree(ipc->msg[i-1].tx.data);
tx                236 sound/soc/intel/common/sst-ipc.c 		if ((msg->tx.header & mask) == header)
tx                310 sound/soc/intel/common/sst-ipc.c 			kfree(ipc->msg[i].tx.data);
tx                 28 sound/soc/intel/common/sst-ipc.h 	struct sst_ipc_message tx;
tx                514 sound/soc/intel/haswell/sst-haswell-ipc.c 	u32 header = msg->tx.header & ~(IPC_STATUS_MASK | IPC_GLB_REPLY_MASK);
tx               2056 sound/soc/intel/haswell/sst-haswell-ipc.c 	sst_dsp_outbox_write(ipc->dsp, msg->tx.data, msg->tx.size);
tx               2057 sound/soc/intel/haswell/sst-haswell-ipc.c 	sst_dsp_ipc_msg_tx(ipc->dsp, msg->tx.header);
tx               2078 sound/soc/intel/haswell/sst-haswell-ipc.c 	memcpy(msg->tx.data, tx_data, tx_size);
tx                369 sound/soc/intel/skylake/cnl-sst.c 	struct skl_ipc_header *header = (struct skl_ipc_header *)(&msg->tx.header);
tx                371 sound/soc/intel/skylake/cnl-sst.c 	if (msg->tx.size)
tx                372 sound/soc/intel/skylake/cnl-sst.c 		sst_dsp_outbox_write(ipc->dsp, msg->tx.data, msg->tx.size);
tx                284 sound/soc/intel/skylake/skl-sst-ipc.c 		memcpy(msg->tx.data, tx_data, tx_size);
tx                298 sound/soc/intel/skylake/skl-sst-ipc.c 	struct skl_ipc_header *header = (struct skl_ipc_header *)(&msg->tx.header);
tx                300 sound/soc/intel/skylake/skl-sst-ipc.c 	if (msg->tx.size)
tx                301 sound/soc/intel/skylake/skl-sst-ipc.c 		sst_dsp_outbox_write(ipc->dsp, msg->tx.data, msg->tx.size);
tx                135 sound/soc/mediatek/common/mtk-btcvsd.c 	struct mtk_btcvsd_snd_stream *tx;
tx                212 sound/soc/mediatek/common/mtk-btcvsd.c 		bt->tx->state, bt->rx->state, bt->irq_disabled);
tx                216 sound/soc/mediatek/common/mtk-btcvsd.c 	if (bt->tx->state == BT_SCO_STATE_IDLE &&
tx                234 sound/soc/mediatek/common/mtk-btcvsd.c 	memset(bt->tx, 0, sizeof(*bt->tx));
tx                237 sound/soc/mediatek/common/mtk-btcvsd.c 	bt->tx->packet_size = BTCVSD_TX_PACKET_SIZE;
tx                238 sound/soc/mediatek/common/mtk-btcvsd.c 	bt->tx->buf_size = BTCVSD_TX_BUF_SIZE;
tx                239 sound/soc/mediatek/common/mtk-btcvsd.c 	bt->tx->timeout = 0;
tx                240 sound/soc/mediatek/common/mtk-btcvsd.c 	bt->tx->rw_cnt = 0;
tx                241 sound/soc/mediatek/common/mtk-btcvsd.c 	bt->tx->stream = SNDRV_PCM_STREAM_PLAYBACK;
tx                261 sound/soc/mediatek/common/mtk-btcvsd.c 	ts->time_stamp_us = bt->tx->time_stamp;
tx                262 sound/soc/mediatek/common/mtk-btcvsd.c 	ts->data_count_equi_time = bt->tx->buf_data_equivalent_time;
tx                327 sound/soc/mediatek/common/mtk-btcvsd.c 		memset(bt->tx->temp_packet_buf, 170, SCO_PACKET_180);
tx                329 sound/soc/mediatek/common/mtk-btcvsd.c 		memcpy(bt->tx->temp_packet_buf,
tx                334 sound/soc/mediatek/common/mtk-btcvsd.c 	num_valid_addr = bt->tx->buffer_info.num_valid_addr;
tx                343 sound/soc/mediatek/common/mtk-btcvsd.c 			 bt->tx->buffer_info.bt_sram_addr[i]);
tx                345 sound/soc/mediatek/common/mtk-btcvsd.c 		dst = (void *)bt->tx->buffer_info.bt_sram_addr[i];
tx                348 sound/soc/mediatek/common/mtk-btcvsd.c 					     bt->tx->temp_packet_buf, dst,
tx                349 sound/soc/mediatek/common/mtk-btcvsd.c 					     bt->tx->buffer_info.packet_length,
tx                350 sound/soc/mediatek/common/mtk-btcvsd.c 					     bt->tx->buffer_info.packet_num);
tx                435 sound/soc/mediatek/common/mtk-btcvsd.c 		memcpy(bt->tx->temp_packet_buf + (bt->tx->packet_size * i),
tx                437 sound/soc/mediatek/common/mtk-btcvsd.c 			(bt->tx->packet_r % SCO_TX_PACKER_BUF_NUM) *
tx                438 sound/soc/mediatek/common/mtk-btcvsd.c 			bt->tx->packet_size),
tx                439 sound/soc/mediatek/common/mtk-btcvsd.c 		       bt->tx->packet_size);
tx                441 sound/soc/mediatek/common/mtk-btcvsd.c 		bt->tx->packet_r++;
tx                447 sound/soc/mediatek/common/mtk-btcvsd.c 	if (!bt->tx->mute) {
tx                449 sound/soc/mediatek/common/mtk-btcvsd.c 					     bt->tx->temp_packet_buf, dst,
tx                454 sound/soc/mediatek/common/mtk-btcvsd.c 	bt->tx->buffer_info.packet_length = packet_length;
tx                455 sound/soc/mediatek/common/mtk-btcvsd.c 	bt->tx->buffer_info.packet_num = packet_num;
tx                456 sound/soc/mediatek/common/mtk-btcvsd.c 	for (i = 0; i < bt->tx->buffer_info.num_valid_addr; i++) {
tx                457 sound/soc/mediatek/common/mtk-btcvsd.c 		if (bt->tx->buffer_info.bt_sram_addr[i] == ap_addr_tx) {
tx                466 sound/soc/mediatek/common/mtk-btcvsd.c 		bt->tx->buffer_info.num_valid_addr++;
tx                467 sound/soc/mediatek/common/mtk-btcvsd.c 		next_idx = bt->tx->buffer_info.num_valid_addr - 1;
tx                468 sound/soc/mediatek/common/mtk-btcvsd.c 		bt->tx->buffer_info.bt_sram_addr[next_idx] = ap_addr_tx;
tx                472 sound/soc/mediatek/common/mtk-btcvsd.c 			 bt->tx->buffer_info.num_valid_addr);
tx                475 sound/soc/mediatek/common/mtk-btcvsd.c 	if (bt->tx->mute)
tx                489 sound/soc/mediatek/common/mtk-btcvsd.c 	    bt->tx->state != BT_SCO_STATE_RUNNING &&
tx                490 sound/soc/mediatek/common/mtk-btcvsd.c 	    bt->tx->state != BT_SCO_STATE_ENDING &&
tx                491 sound/soc/mediatek/common/mtk-btcvsd.c 	    bt->tx->state != BT_SCO_STATE_LOOPBACK) {
tx                493 sound/soc/mediatek/common/mtk-btcvsd.c 			 __func__, bt->rx->state, bt->tx->state);
tx                517 sound/soc/mediatek/common/mtk-btcvsd.c 	if (bt->tx->state == BT_SCO_STATE_LOOPBACK) {
tx                542 sound/soc/mediatek/common/mtk-btcvsd.c 					     bt->tx->temp_packet_buf,
tx                546 sound/soc/mediatek/common/mtk-btcvsd.c 					     bt->tx->temp_packet_buf, dst,
tx                550 sound/soc/mediatek/common/mtk-btcvsd.c 		bt->tx->rw_cnt++;
tx                585 sound/soc/mediatek/common/mtk-btcvsd.c 	bt->tx->timeout = 0;
tx                586 sound/soc/mediatek/common/mtk-btcvsd.c 	if ((bt->tx->state == BT_SCO_STATE_RUNNING ||
tx                587 sound/soc/mediatek/common/mtk-btcvsd.c 	     bt->tx->state == BT_SCO_STATE_ENDING) &&
tx                588 sound/soc/mediatek/common/mtk-btcvsd.c 	    bt->tx->trigger_start) {
tx                589 sound/soc/mediatek/common/mtk-btcvsd.c 		if (bt->tx->xrun) {
tx                593 sound/soc/mediatek/common/mtk-btcvsd.c 			if (bt->tx->packet_w - bt->tx->packet_r >=
tx                595 sound/soc/mediatek/common/mtk-btcvsd.c 				bt->tx->xrun = 0;
tx                601 sound/soc/mediatek/common/mtk-btcvsd.c 		if ((!bt->tx->xrun &&
tx                602 sound/soc/mediatek/common/mtk-btcvsd.c 		     (bt->tx->packet_w - bt->tx->packet_r >= buf_cnt_tx)) ||
tx                603 sound/soc/mediatek/common/mtk-btcvsd.c 		    bt->tx->state == BT_SCO_STATE_ENDING) {
tx                609 sound/soc/mediatek/common/mtk-btcvsd.c 			bt->tx->rw_cnt++;
tx                611 sound/soc/mediatek/common/mtk-btcvsd.c 			bt->tx->xrun = 1;
tx                624 sound/soc/mediatek/common/mtk-btcvsd.c 	if (bt->tx->state == BT_SCO_STATE_RUNNING ||
tx                625 sound/soc/mediatek/common/mtk-btcvsd.c 	    bt->tx->state == BT_SCO_STATE_ENDING) {
tx                626 sound/soc/mediatek/common/mtk-btcvsd.c 		bt->tx->wait_flag = 1;
tx                628 sound/soc/mediatek/common/mtk-btcvsd.c 		snd_pcm_period_elapsed(bt->tx->substream);
tx                786 sound/soc/mediatek/common/mtk-btcvsd.c 	unsigned int packet_size = bt->tx->packet_size;
tx                792 sound/soc/mediatek/common/mtk-btcvsd.c 	bt->tx->time_stamp = sched_clock();
tx                793 sound/soc/mediatek/common/mtk-btcvsd.c 	bt->tx->buf_data_equivalent_time =
tx                794 sound/soc/mediatek/common/mtk-btcvsd.c 		(unsigned long long)(bt->tx->packet_w - bt->tx->packet_r) *
tx                798 sound/soc/mediatek/common/mtk-btcvsd.c 	bt->tx->buf_data_equivalent_time *= 1000;
tx                803 sound/soc/mediatek/common/mtk-btcvsd.c 		avail = bt->tx->buf_size -
tx                804 sound/soc/mediatek/common/mtk-btcvsd.c 			(bt->tx->packet_w - bt->tx->packet_r) * packet_size;
tx                806 sound/soc/mediatek/common/mtk-btcvsd.c 		cur_write_idx = (bt->tx->packet_w % SCO_TX_PACKER_BUF_NUM) *
tx                834 sound/soc/mediatek/common/mtk-btcvsd.c 		cont = bt->tx->buf_size - cur_write_idx;
tx                848 sound/soc/mediatek/common/mtk-btcvsd.c 		bt->tx->packet_w += write_size / packet_size;
tx                861 sound/soc/mediatek/common/mtk-btcvsd.c 		return bt->tx;
tx                893 sound/soc/mediatek/common/mtk-btcvsd.c 		bt->tx->substream = substream;
tx                926 sound/soc/mediatek/common/mtk-btcvsd.c 	    params_buffer_bytes(hw_params) % bt->tx->packet_size != 0) {
tx               1013 sound/soc/mediatek/common/mtk-btcvsd.c 		bt_stream = bt->tx;
tx               1021 sound/soc/mediatek/common/mtk-btcvsd.c 			bt->tx->packet_r : bt->rx->packet_w;
tx               1113 sound/soc/mediatek/common/mtk-btcvsd.c 	bool lpbk_en = bt->tx->state == BT_SCO_STATE_LOOPBACK;
tx               1126 sound/soc/mediatek/common/mtk-btcvsd.c 		mtk_btcvsd_snd_set_state(bt, bt->tx, BT_SCO_STATE_LOOPBACK);
tx               1129 sound/soc/mediatek/common/mtk-btcvsd.c 		mtk_btcvsd_snd_set_state(bt, bt->tx, BT_SCO_STATE_RUNNING);
tx               1141 sound/soc/mediatek/common/mtk-btcvsd.c 	if (!bt->tx) {
tx               1146 sound/soc/mediatek/common/mtk-btcvsd.c 	ucontrol->value.integer.value[0] = bt->tx->mute;
tx               1156 sound/soc/mediatek/common/mtk-btcvsd.c 	if (!bt->tx)
tx               1159 sound/soc/mediatek/common/mtk-btcvsd.c 	bt->tx->mute = ucontrol->value.integer.value[0];
tx               1223 sound/soc/mediatek/common/mtk-btcvsd.c 	if (!bt->tx)
tx               1226 sound/soc/mediatek/common/mtk-btcvsd.c 	ucontrol->value.integer.value[0] = bt->tx->rw_cnt ? 1 : 0;
tx               1236 sound/soc/mediatek/common/mtk-btcvsd.c 	ucontrol->value.integer.value[0] = bt->tx->timeout;
tx               1323 sound/soc/mediatek/common/mtk-btcvsd.c 	btcvsd->tx = devm_kzalloc(btcvsd->dev, sizeof(*btcvsd->tx), GFP_KERNEL);
tx               1324 sound/soc/mediatek/common/mtk-btcvsd.c 	if (!btcvsd->tx)
tx               1394 sound/soc/mediatek/common/mtk-btcvsd.c 	mtk_btcvsd_snd_set_state(btcvsd, btcvsd->tx, BT_SCO_STATE_IDLE);
tx                 19 sound/soc/meson/axg-card.c 	u32 tx;
tx                215 sound/soc/meson/axg-card.c 					       be->codec_masks[i].tx,
tx                339 sound/soc/meson/axg-card.c 	u32 tx, rx;
tx                349 sound/soc/meson/axg-card.c 	for (i = 0, tx = 0; i < AXG_TDM_NUM_LANES; i++) {
tx                352 sound/soc/meson/axg-card.c 		tx = max(tx, be->tx_mask[i]);
tx                356 sound/soc/meson/axg-card.c 	if (!tx)
tx                370 sound/soc/meson/axg-card.c 	if (!tx && !rx) {
tx                381 sound/soc/meson/axg-card.c 		be->slots = fls(max(tx, rx));
tx                382 sound/soc/meson/axg-card.c 	} else if (be->slots < fls(max(tx, rx)) || be->slots > 32) {
tx                415 sound/soc/meson/axg-card.c 					 &codec_mask->tx);
tx                 40 sound/soc/meson/axg-tdm-interface.c 	struct axg_tdm_stream *tx = (struct axg_tdm_stream *)
tx                 83 sound/soc/meson/axg-tdm-interface.c 	if (tx) {
tx                 84 sound/soc/meson/axg-tdm-interface.c 		tx->mask = tx_mask;
tx                 16 sound/soc/samsung/dma.h 				       const char *tx, const char *rx,
tx                 18 sound/soc/samsung/dmaengine.c 				       const char *tx, const char *rx,
tx                 31 sound/soc/samsung/dmaengine.c 	pcm_conf->chan_names[SNDRV_PCM_STREAM_PLAYBACK] = tx;
tx                419 sound/soc/ti/omap-mcbsp.c 	int tx = (stream == SNDRV_PCM_STREAM_PLAYBACK);
tx                420 sound/soc/ti/omap-mcbsp.c 	int rx = !tx;
tx                440 sound/soc/ti/omap-mcbsp.c 	tx &= 1;
tx                442 sound/soc/ti/omap-mcbsp.c 	MCBSP_WRITE(mcbsp, SPCR2, w | tx);
tx                465 sound/soc/ti/omap-mcbsp.c 		w &= ~(tx ? XDISABLE : 0);
tx                478 sound/soc/ti/omap-mcbsp.c 	int tx = (stream == SNDRV_PCM_STREAM_PLAYBACK);
tx                479 sound/soc/ti/omap-mcbsp.c 	int rx = !tx;
tx                484 sound/soc/ti/omap-mcbsp.c 	tx &= 1;
tx                487 sound/soc/ti/omap-mcbsp.c 		w |= (tx ? XDISABLE : 0);
tx                491 sound/soc/ti/omap-mcbsp.c 	MCBSP_WRITE(mcbsp, SPCR2, w & ~tx);
tx                834 sound/soc/ti/omap-mcbsp.c 	int tx = (substream->stream == SNDRV_PCM_STREAM_PLAYBACK);
tx                835 sound/soc/ti/omap-mcbsp.c 	int stream1 = tx ? SNDRV_PCM_STREAM_PLAYBACK : SNDRV_PCM_STREAM_CAPTURE;
tx                836 sound/soc/ti/omap-mcbsp.c 	int stream2 = tx ? SNDRV_PCM_STREAM_CAPTURE : SNDRV_PCM_STREAM_PLAYBACK;
tx                857 sound/soc/ti/omap-mcbsp.c 	int tx = (substream->stream == SNDRV_PCM_STREAM_PLAYBACK);
tx                858 sound/soc/ti/omap-mcbsp.c 	int stream1 = tx ? SNDRV_PCM_STREAM_PLAYBACK : SNDRV_PCM_STREAM_CAPTURE;
tx                859 sound/soc/ti/omap-mcbsp.c 	int stream2 = tx ? SNDRV_PCM_STREAM_CAPTURE : SNDRV_PCM_STREAM_PLAYBACK;
tx                268 sound/soc/ti/omap-mcpdm.c 	int tx = (substream->stream == SNDRV_PCM_STREAM_PLAYBACK);
tx                269 sound/soc/ti/omap-mcpdm.c 	int stream1 = tx ? SNDRV_PCM_STREAM_PLAYBACK : SNDRV_PCM_STREAM_CAPTURE;
tx                270 sound/soc/ti/omap-mcpdm.c 	int stream2 = tx ? SNDRV_PCM_STREAM_CAPTURE : SNDRV_PCM_STREAM_PLAYBACK;
tx                380 sound/soc/ti/omap-mcpdm.c 	int tx = (substream->stream == SNDRV_PCM_STREAM_PLAYBACK);
tx                381 sound/soc/ti/omap-mcpdm.c 	int stream1 = tx ? SNDRV_PCM_STREAM_PLAYBACK : SNDRV_PCM_STREAM_CAPTURE;
tx                382 sound/soc/ti/omap-mcpdm.c 	int stream2 = tx ? SNDRV_PCM_STREAM_CAPTURE : SNDRV_PCM_STREAM_PLAYBACK;
tx                 52 tools/include/uapi/linux/if_xdp.h 	struct xdp_ring_offset tx;
tx                 58 tools/lib/bpf/xsk.c 	struct xsk_ring_prod *tx;
tx                 86 tools/lib/bpf/xsk.c 	struct xdp_ring_offset_v1 tx;
tx                166 tools/lib/bpf/xsk.c 	off->tx.producer = off_v1.tx.producer;
tx                167 tools/lib/bpf/xsk.c 	off->tx.consumer = off_v1.tx.consumer;
tx                168 tools/lib/bpf/xsk.c 	off->tx.desc = off_v1.tx.desc;
tx                169 tools/lib/bpf/xsk.c 	off->tx.flags = off_v1.tx.consumer + sizeof(__u32);
tx                560 tools/lib/bpf/xsk.c 		       struct xsk_ring_cons *rx, struct xsk_ring_prod *tx,
tx                569 tools/lib/bpf/xsk.c 	if (!umem || !xsk_ptr || !rx || !tx)
tx                615 tools/lib/bpf/xsk.c 	if (tx) {
tx                650 tools/lib/bpf/xsk.c 	if (tx) {
tx                651 tools/lib/bpf/xsk.c 		tx_map = mmap(NULL, off.tx.desc +
tx                660 tools/lib/bpf/xsk.c 		tx->mask = xsk->config.tx_size - 1;
tx                661 tools/lib/bpf/xsk.c 		tx->size = xsk->config.tx_size;
tx                662 tools/lib/bpf/xsk.c 		tx->producer = tx_map + off.tx.producer;
tx                663 tools/lib/bpf/xsk.c 		tx->consumer = tx_map + off.tx.consumer;
tx                664 tools/lib/bpf/xsk.c 		tx->flags = tx_map + off.tx.flags;
tx                665 tools/lib/bpf/xsk.c 		tx->ring = tx_map + off.tx.desc;
tx                666 tools/lib/bpf/xsk.c 		tx->cached_cons = xsk->config.tx_size;
tx                668 tools/lib/bpf/xsk.c 	xsk->tx = tx;
tx                693 tools/lib/bpf/xsk.c 	if (tx)
tx                694 tools/lib/bpf/xsk.c 		munmap(tx_map, off.tx.desc +
tx                753 tools/lib/bpf/xsk.c 		if (xsk->tx) {
tx                754 tools/lib/bpf/xsk.c 			munmap(xsk->tx->ring - off.tx.desc,
tx                755 tools/lib/bpf/xsk.c 			       off.tx.desc + xsk->config.tx_size * desc_sz);
tx                 64 tools/lib/bpf/xsk.h static inline struct xdp_desc *xsk_ring_prod__tx_desc(struct xsk_ring_prod *tx,
tx                 67 tools/lib/bpf/xsk.h 	struct xdp_desc *descs = (struct xdp_desc *)tx->ring;
tx                 69 tools/lib/bpf/xsk.h 	return &descs[idx & tx->mask];
tx                235 tools/lib/bpf/xsk.h 				  struct xsk_ring_prod *tx,
tx                 90 tools/perf/util/stat.c 	ID(TRANSACTION_START,	cpu/tx-start/),
tx                113 tools/spi/spidev_test.c static void transfer(int fd, uint8_t const *tx, uint8_t const *rx, size_t len)
tx                118 tools/spi/spidev_test.c 		.tx_buf = (unsigned long)tx,
tx                146 tools/spi/spidev_test.c 		hex_dump(tx, len, 32, "TX");
tx                300 tools/spi/spidev_test.c 	uint8_t *tx;
tx                303 tools/spi/spidev_test.c 	tx = malloc(size);
tx                304 tools/spi/spidev_test.c 	if (!tx)
tx                311 tools/spi/spidev_test.c 	size = unescape((char *)tx, str, size);
tx                312 tools/spi/spidev_test.c 	transfer(fd, tx, rx, size);
tx                314 tools/spi/spidev_test.c 	free(tx);
tx                322 tools/spi/spidev_test.c 	uint8_t *tx;
tx                332 tools/spi/spidev_test.c 	tx = malloc(sb.st_size);
tx                333 tools/spi/spidev_test.c 	if (!tx)
tx                340 tools/spi/spidev_test.c 	bytes = read(tx_fd, tx, sb.st_size);
tx                344 tools/spi/spidev_test.c 	transfer(fd, tx, rx, sb.st_size);
tx                346 tools/spi/spidev_test.c 	free(tx);
tx                369 tools/spi/spidev_test.c 	uint8_t *tx;
tx                373 tools/spi/spidev_test.c 	tx = malloc(len);
tx                374 tools/spi/spidev_test.c 	if (!tx)
tx                377 tools/spi/spidev_test.c 		tx[i] = random();
tx                383 tools/spi/spidev_test.c 	transfer(fd, tx, rx, len);
tx                389 tools/spi/spidev_test.c 		if (memcmp(tx, rx, len)) {
tx                391 tools/spi/spidev_test.c 			hex_dump(tx, len, 32, "TX");
tx                398 tools/spi/spidev_test.c 	free(tx);
tx                514 tools/testing/selftests/bpf/test_flow_dissector.c 	int fdt = -1, fdr = -1, len, tx = 0, rx = 0;
tx                529 tools/testing/selftests/bpf/test_flow_dissector.c 			tx += do_tx(fdt, buf, len);
tx                539 tools/testing/selftests/bpf/test_flow_dissector.c 				fprintf(stderr, "pkts: tx=%u rx=%u\n", tx, rx);
tx                540 tools/testing/selftests/bpf/test_flow_dissector.c 				tx = 0;
tx                545 tools/testing/selftests/bpf/test_flow_dissector.c 			if (tx == cfg_num_pkt)
tx                551 tools/testing/selftests/bpf/test_flow_dissector.c 	if (rx < tx) {
tx                553 tools/testing/selftests/bpf/test_flow_dissector.c 		while (rx < tx) {
tx                563 tools/testing/selftests/bpf/test_flow_dissector.c 	fprintf(stderr, "pkts: tx=%u rx=%u\n", tx, rx);
tx                577 tools/testing/selftests/bpf/test_flow_dissector.c 		return rx != tx;
tx                447 tools/testing/selftests/bpf/test_sockmap.c 		    struct msg_stats *s, bool tx,
tx                455 tools/testing/selftests/bpf/test_sockmap.c 	err = msg_alloc_iov(&msg, iov_count, iov_length, data, tx);
tx                459 tools/testing/selftests/bpf/test_sockmap.c 		err = msg_alloc_iov(&msg_peek, iov_count, iov_length, data, tx);
tx                464 tools/testing/selftests/bpf/test_sockmap.c 	if (tx) {
tx                446 tools/testing/selftests/net/psock_tpacket.c 				struct tpacket3_hdr *tx = next;
tx                448 tools/testing/selftests/net/psock_tpacket.c 				tx->tp_snaplen = packet_len;
tx                449 tools/testing/selftests/net/psock_tpacket.c 				tx->tp_len = packet_len;
tx                450 tools/testing/selftests/net/psock_tpacket.c 				tx->tp_next_offset = 0;
tx                452 tools/testing/selftests/net/psock_tpacket.c 				memcpy((uint8_t *)tx + TPACKET3_HDRLEN -
tx                455 tools/testing/selftests/net/psock_tpacket.c 				total_bytes += tx->tp_snaplen;
tx                 41 tools/testing/selftests/ptp/testptp.c static int clock_adjtime(clockid_t id, struct timex *tx)
tx                 43 tools/testing/selftests/ptp/testptp.c 	return syscall(__NR_clock_adjtime, id, tx);
tx                152 tools/testing/selftests/ptp/testptp.c 	struct timex tx;
tx                284 tools/testing/selftests/ptp/testptp.c 		memset(&tx, 0, sizeof(tx));
tx                285 tools/testing/selftests/ptp/testptp.c 		tx.modes = ADJ_FREQUENCY;
tx                286 tools/testing/selftests/ptp/testptp.c 		tx.freq = ppb_to_scaled_ppm(adjfreq);
tx                287 tools/testing/selftests/ptp/testptp.c 		if (clock_adjtime(clkid, &tx)) {
tx                295 tools/testing/selftests/ptp/testptp.c 		memset(&tx, 0, sizeof(tx));
tx                296 tools/testing/selftests/ptp/testptp.c 		tx.modes = ADJ_SETOFFSET;
tx                297 tools/testing/selftests/ptp/testptp.c 		tx.time.tv_sec = adjtime;
tx                298 tools/testing/selftests/ptp/testptp.c 		tx.time.tv_usec = 0;
tx                299 tools/testing/selftests/ptp/testptp.c 		if (clock_adjtime(clkid, &tx) < 0) {
tx                 38 tools/testing/selftests/timers/change_skew.c 	struct timex tx;
tx                 41 tools/testing/selftests/timers/change_skew.c 	tx.modes = ADJ_FREQUENCY;
tx                 42 tools/testing/selftests/timers/change_skew.c 	tx.freq = ppm << 16;
tx                 44 tools/testing/selftests/timers/change_skew.c 	ret = adjtimex(&tx);
tx                 60 tools/testing/selftests/timers/change_skew.c 	struct timex tx;
tx                 69 tools/testing/selftests/timers/change_skew.c 	tx.modes = ADJ_OFFSET;
tx                 70 tools/testing/selftests/timers/change_skew.c 	tx.offset = 0;
tx                 71 tools/testing/selftests/timers/change_skew.c 	ret = adjtimex(&tx);
tx                 86 tools/testing/selftests/timers/change_skew.c 	tx.modes = ADJ_FREQUENCY;
tx                 87 tools/testing/selftests/timers/change_skew.c 	tx.offset = 0;
tx                 88 tools/testing/selftests/timers/change_skew.c 	adjtimex(&tx);
tx                 97 tools/testing/selftests/timers/leap-a-day.c 	struct timex tx;
tx                106 tools/testing/selftests/timers/leap-a-day.c 	tx.modes = ADJ_STATUS;
tx                107 tools/testing/selftests/timers/leap-a-day.c 	tx.status = STA_PLL;
tx                108 tools/testing/selftests/timers/leap-a-day.c 	ret = adjtimex(&tx);
tx                111 tools/testing/selftests/timers/leap-a-day.c 	tx.modes = ADJ_MAXERROR;
tx                112 tools/testing/selftests/timers/leap-a-day.c 	tx.maxerror = 0;
tx                113 tools/testing/selftests/timers/leap-a-day.c 	ret = adjtimex(&tx);
tx                116 tools/testing/selftests/timers/leap-a-day.c 	tx.modes = ADJ_STATUS;
tx                117 tools/testing/selftests/timers/leap-a-day.c 	tx.status = 0;
tx                118 tools/testing/selftests/timers/leap-a-day.c 	ret = adjtimex(&tx);
tx                132 tools/testing/selftests/timers/leap-a-day.c 	struct timex tx;
tx                135 tools/testing/selftests/timers/leap-a-day.c 	tx.modes = 0;
tx                136 tools/testing/selftests/timers/leap-a-day.c 	ret = adjtimex(&tx);
tx                138 tools/testing/selftests/timers/leap-a-day.c 	if (tx.time.tv_sec < next_leap) {
tx                142 tools/testing/selftests/timers/leap-a-day.c 					tx.time.tv_sec,
tx                143 tools/testing/selftests/timers/leap-a-day.c 					tx.time.tv_usec,
tx                144 tools/testing/selftests/timers/leap-a-day.c 					tx.tai,
tx                151 tools/testing/selftests/timers/leap-a-day.c 					tx.time.tv_sec,
tx                152 tools/testing/selftests/timers/leap-a-day.c 					tx.time.tv_usec,
tx                153 tools/testing/selftests/timers/leap-a-day.c 					tx.tai,
tx                239 tools/testing/selftests/timers/leap-a-day.c 		struct timex tx;
tx                262 tools/testing/selftests/timers/leap-a-day.c 		tx.modes = ADJ_STATUS;
tx                264 tools/testing/selftests/timers/leap-a-day.c 			tx.status = STA_INS;
tx                266 tools/testing/selftests/timers/leap-a-day.c 			tx.status = STA_DEL;
tx                267 tools/testing/selftests/timers/leap-a-day.c 		ret = adjtimex(&tx);
tx                275 tools/testing/selftests/timers/leap-a-day.c 		tx.modes = 0;
tx                276 tools/testing/selftests/timers/leap-a-day.c 		ret = adjtimex(&tx);
tx                277 tools/testing/selftests/timers/leap-a-day.c 		if (tx.status != STA_INS && tx.status != STA_DEL) {
tx                315 tools/testing/selftests/timers/leap-a-day.c 		tx.modes = 0;
tx                316 tools/testing/selftests/timers/leap-a-day.c 		ret = adjtimex(&tx);
tx                317 tools/testing/selftests/timers/leap-a-day.c 		if (tx.status != STA_INS && tx.status != STA_DEL) {
tx                319 tools/testing/selftests/timers/leap-a-day.c 			tx.modes = ADJ_STATUS;
tx                321 tools/testing/selftests/timers/leap-a-day.c 				tx.status = STA_INS;
tx                323 tools/testing/selftests/timers/leap-a-day.c 				tx.status = STA_DEL;
tx                324 tools/testing/selftests/timers/leap-a-day.c 			ret = adjtimex(&tx);
tx                328 tools/testing/selftests/timers/leap-a-day.c 		now = tx.time.tv_sec;
tx                334 tools/testing/selftests/timers/leap-a-day.c 			tx.modes = 0;
tx                335 tools/testing/selftests/timers/leap-a-day.c 			ret = adjtimex(&tx);
tx                344 tools/testing/selftests/timers/leap-a-day.c 				ctime_r(&tx.time.tv_sec, buf);
tx                349 tools/testing/selftests/timers/leap-a-day.c 						tx.time.tv_usec,
tx                350 tools/testing/selftests/timers/leap-a-day.c 						tx.tai,
tx                353 tools/testing/selftests/timers/leap-a-day.c 			now = tx.time.tv_sec;
tx                 30 tools/testing/selftests/timers/leapcrash.c 	struct timex tx;
tx                 39 tools/testing/selftests/timers/leapcrash.c 	tx.modes = ADJ_STATUS;
tx                 40 tools/testing/selftests/timers/leapcrash.c 	tx.status = STA_PLL;
tx                 41 tools/testing/selftests/timers/leapcrash.c 	ret = adjtimex(&tx);
tx                 43 tools/testing/selftests/timers/leapcrash.c 	tx.modes = ADJ_STATUS;
tx                 44 tools/testing/selftests/timers/leapcrash.c 	tx.status = 0;
tx                 45 tools/testing/selftests/timers/leapcrash.c 	ret = adjtimex(&tx);
tx                 60 tools/testing/selftests/timers/leapcrash.c 	struct timex tx;
tx                 92 tools/testing/selftests/timers/leapcrash.c 		tx.modes = 0;
tx                 93 tools/testing/selftests/timers/leapcrash.c 		adjtimex(&tx);
tx                 96 tools/testing/selftests/timers/leapcrash.c 		while (tx.time.tv_sec < next_leap + 1) {
tx                 98 tools/testing/selftests/timers/leapcrash.c 			tx.modes = ADJ_STATUS;
tx                 99 tools/testing/selftests/timers/leapcrash.c 			tx.status = STA_INS;
tx                100 tools/testing/selftests/timers/leapcrash.c 			adjtimex(&tx);
tx                 30 tools/testing/selftests/timers/set-tai.c 	struct timex tx;
tx                 32 tools/testing/selftests/timers/set-tai.c 	memset(&tx, 0, sizeof(tx));
tx                 34 tools/testing/selftests/timers/set-tai.c 	tx.modes = ADJ_TAI;
tx                 35 tools/testing/selftests/timers/set-tai.c 	tx.constant = offset;
tx                 37 tools/testing/selftests/timers/set-tai.c 	return adjtimex(&tx);
tx                 42 tools/testing/selftests/timers/set-tai.c 	struct timex tx;
tx                 44 tools/testing/selftests/timers/set-tai.c 	memset(&tx, 0, sizeof(tx));
tx                 46 tools/testing/selftests/timers/set-tai.c 	adjtimex(&tx);
tx                 47 tools/testing/selftests/timers/set-tai.c 	return tx.tai;
tx                 43 tools/testing/selftests/timers/skew_consistency.c 	struct timex tx;
tx                 59 tools/testing/selftests/timers/skew_consistency.c 		tx.modes = ADJ_FREQUENCY;
tx                 60 tools/testing/selftests/timers/skew_consistency.c 		tx.freq = ppm << 16;
tx                 61 tools/testing/selftests/timers/skew_consistency.c 		adjtimex(&tx);
tx                 66 tools/testing/selftests/timers/skew_consistency.c 	tx.modes = ADJ_FREQUENCY;
tx                 67 tools/testing/selftests/timers/skew_consistency.c 	tx.offset = 0;
tx                 68 tools/testing/selftests/timers/skew_consistency.c 	adjtimex(&tx);
tx                 43 tools/testing/selftests/timers/valid-adjtimex.c static int clock_adjtime(clockid_t id, struct timex *tx)
tx                 45 tools/testing/selftests/timers/valid-adjtimex.c 	return syscall(__NR_clock_adjtime, id, tx);
tx                 52 tools/testing/selftests/timers/valid-adjtimex.c 	struct timex tx;
tx                 55 tools/testing/selftests/timers/valid-adjtimex.c 	tx.modes = ADJ_STATUS;
tx                 56 tools/testing/selftests/timers/valid-adjtimex.c 	tx.status = 0;
tx                 57 tools/testing/selftests/timers/valid-adjtimex.c 	ret = adjtimex(&tx);
tx                116 tools/testing/selftests/timers/valid-adjtimex.c 	struct timex tx;
tx                122 tools/testing/selftests/timers/valid-adjtimex.c 	memset(&tx, 0, sizeof(struct timex));
tx                128 tools/testing/selftests/timers/valid-adjtimex.c 		tx.modes = ADJ_FREQUENCY;
tx                129 tools/testing/selftests/timers/valid-adjtimex.c 		tx.freq = valid_freq[i];
tx                131 tools/testing/selftests/timers/valid-adjtimex.c 		ret = adjtimex(&tx);
tx                139 tools/testing/selftests/timers/valid-adjtimex.c 		tx.modes = 0;
tx                140 tools/testing/selftests/timers/valid-adjtimex.c 		ret = adjtimex(&tx);
tx                141 tools/testing/selftests/timers/valid-adjtimex.c 		if (tx.freq != valid_freq[i]) {
tx                143 tools/testing/selftests/timers/valid-adjtimex.c 					tx.freq, valid_freq[i]);
tx                147 tools/testing/selftests/timers/valid-adjtimex.c 		tx.modes = ADJ_FREQUENCY;
tx                148 tools/testing/selftests/timers/valid-adjtimex.c 		tx.freq = outofrange_freq[i];
tx                150 tools/testing/selftests/timers/valid-adjtimex.c 		ret = adjtimex(&tx);
tx                158 tools/testing/selftests/timers/valid-adjtimex.c 		tx.modes = 0;
tx                159 tools/testing/selftests/timers/valid-adjtimex.c 		ret = adjtimex(&tx);
tx                160 tools/testing/selftests/timers/valid-adjtimex.c 		if (tx.freq == outofrange_freq[i]) {
tx                163 tools/testing/selftests/timers/valid-adjtimex.c 					tx.freq);
tx                172 tools/testing/selftests/timers/valid-adjtimex.c 			tx.modes = ADJ_FREQUENCY;
tx                173 tools/testing/selftests/timers/valid-adjtimex.c 			tx.freq = invalid_freq[i];
tx                174 tools/testing/selftests/timers/valid-adjtimex.c 			ret = adjtimex(&tx);
tx                188 tools/testing/selftests/timers/valid-adjtimex.c 	tx.modes = ADJ_FREQUENCY;
tx                189 tools/testing/selftests/timers/valid-adjtimex.c 	tx.freq = 0;
tx                190 tools/testing/selftests/timers/valid-adjtimex.c 	ret = adjtimex(&tx);