Lines Matching refs:skb_cb

343 	struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(msdu);  in ath10k_htt_mgmt_tx()  local
344 u8 vdev_id = skb_cb->vdev_id; in ath10k_htt_mgmt_tx()
371 skb_cb->paddr = dma_map_single(dev, msdu->data, msdu->len, in ath10k_htt_mgmt_tx()
373 res = dma_mapping_error(dev, skb_cb->paddr); in ath10k_htt_mgmt_tx()
389 skb_cb->htt.txbuf = NULL; in ath10k_htt_mgmt_tx()
398 dma_unmap_single(dev, skb_cb->paddr, msdu->len, DMA_TO_DEVICE); in ath10k_htt_mgmt_tx()
416 struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(msdu); in ath10k_htt_tx() local
419 u8 vdev_id = skb_cb->vdev_id; in ath10k_htt_tx()
420 u8 tid = skb_cb->htt.tid; in ath10k_htt_tx()
451 skb_cb->htt.txbuf = dma_pool_alloc(htt->tx_pool, GFP_ATOMIC, in ath10k_htt_tx()
453 if (!skb_cb->htt.txbuf) { in ath10k_htt_tx()
457 skb_cb->htt.txbuf_paddr = paddr; in ath10k_htt_tx()
465 skb_cb->paddr = dma_map_single(dev, msdu->data, msdu->len, in ath10k_htt_tx()
467 res = dma_mapping_error(dev, skb_cb->paddr); in ath10k_htt_tx()
474 frags = skb_cb->htt.txbuf->frags; in ath10k_htt_tx()
476 frags[0].paddr = __cpu_to_le32(skb_cb->paddr); in ath10k_htt_tx()
484 frags_paddr = skb_cb->htt.txbuf_paddr; in ath10k_htt_tx()
489 frags_paddr = skb_cb->paddr; in ath10k_htt_tx()
508 skb_cb->htt.txbuf->htc_hdr.eid = htt->eid; in ath10k_htt_tx()
509 skb_cb->htt.txbuf->htc_hdr.len = __cpu_to_le16( in ath10k_htt_tx()
510 sizeof(skb_cb->htt.txbuf->cmd_hdr) + in ath10k_htt_tx()
511 sizeof(skb_cb->htt.txbuf->cmd_tx) + in ath10k_htt_tx()
513 skb_cb->htt.txbuf->htc_hdr.flags = 0; in ath10k_htt_tx()
533 skb_cb->htt.txbuf->cmd_hdr.msg_type = HTT_H2T_MSG_TYPE_TX_FRM; in ath10k_htt_tx()
534 skb_cb->htt.txbuf->cmd_tx.flags0 = flags0; in ath10k_htt_tx()
535 skb_cb->htt.txbuf->cmd_tx.flags1 = __cpu_to_le16(flags1); in ath10k_htt_tx()
536 skb_cb->htt.txbuf->cmd_tx.len = __cpu_to_le16(msdu->len); in ath10k_htt_tx()
537 skb_cb->htt.txbuf->cmd_tx.id = __cpu_to_le16(msdu_id); in ath10k_htt_tx()
538 skb_cb->htt.txbuf->cmd_tx.frags_paddr = __cpu_to_le32(frags_paddr); in ath10k_htt_tx()
539 skb_cb->htt.txbuf->cmd_tx.peerid = __cpu_to_le16(HTT_INVALID_PEERID); in ath10k_htt_tx()
540 skb_cb->htt.txbuf->cmd_tx.freq = __cpu_to_le16(skb_cb->htt.freq); in ath10k_htt_tx()
546 (u32)skb_cb->paddr, vdev_id, tid, skb_cb->htt.freq); in ath10k_htt_tx()
554 sg_items[0].vaddr = &skb_cb->htt.txbuf->htc_hdr; in ath10k_htt_tx()
555 sg_items[0].paddr = skb_cb->htt.txbuf_paddr + in ath10k_htt_tx()
556 sizeof(skb_cb->htt.txbuf->frags); in ath10k_htt_tx()
557 sg_items[0].len = sizeof(skb_cb->htt.txbuf->htc_hdr) + in ath10k_htt_tx()
558 sizeof(skb_cb->htt.txbuf->cmd_hdr) + in ath10k_htt_tx()
559 sizeof(skb_cb->htt.txbuf->cmd_tx); in ath10k_htt_tx()
564 sg_items[1].paddr = skb_cb->paddr; in ath10k_htt_tx()
576 dma_unmap_single(dev, skb_cb->paddr, msdu->len, DMA_TO_DEVICE); in ath10k_htt_tx()
579 skb_cb->htt.txbuf, in ath10k_htt_tx()
580 skb_cb->htt.txbuf_paddr); in ath10k_htt_tx()