/linux-4.1.27/drivers/parisc/ |
H A D | iommu-helpers.h | 21 unsigned long dma_offset = 0, dma_len = 0; iommu_fill_pdir() local 46 BUG_ON(pdirp && (dma_len != sg_dma_len(dma_sg))); iommu_fill_pdir() 50 dma_len = sg_dma_len(startsg); iommu_fill_pdir() 105 unsigned long dma_offset, dma_len; /* start/len of DMA stream */ iommu_coalesce_chunks() local 119 dma_len = startsg->length; iommu_coalesce_chunks() 148 if (unlikely(ALIGN(dma_len + dma_offset + startsg->length, IOVP_SIZE) > iommu_coalesce_chunks() 162 dma_len += startsg->length; iommu_coalesce_chunks() 170 sg_dma_len(contig_sg) = dma_len; iommu_coalesce_chunks() 171 dma_len = ALIGN(dma_len + dma_offset, IOVP_SIZE); iommu_coalesce_chunks() 174 | (iommu_alloc_range(ioc, dev, dma_len) << IOVP_SHIFT) iommu_coalesce_chunks()
|
H A D | ccio-dma.c | 949 ** o dma_len will contain the number of bytes to map ccio_map_sg()
|
H A D | sba_iommu.c | 983 ** o dma_len will contain the number of bytes to map sba_map_sg()
|
/linux-4.1.27/drivers/net/wireless/brcm80211/brcmsmac/ |
H A D | brcms_trace_brcmsmac_tx.h | 76 u16 dma_len), 78 dma_len), 85 __field(u16, dma_len) 93 __entry->dma_len = dma_len; 95 TP_printk("[%s] ampdu session max_len=%u max_frames=%u len=%u frames=%u dma_len=%u", 98 __entry->ampdu_frames, __entry->dma_len)
|
H A D | ampdu.h | 29 * dma_len: DMA length of this AMPDU 37 u16 dma_len; member in struct:brcms_ampdu_session
|
H A D | ampdu.c | 509 session->dma_len = 0; brcms_c_ampdu_reset_session() 568 session->dma_len += p->len; brcms_c_ampdu_add_frame() 644 u16 dma_len = session->dma_len; brcms_c_ampdu_finalize() local 771 dma_len = min(dma_len, f->ampdu_pld_size); brcms_c_ampdu_finalize() 772 txh->PreloadSize = cpu_to_le16(dma_len); brcms_c_ampdu_finalize()
|
H A D | dma.c | 1325 session->dma_len); ampdu_finalize()
|
/linux-4.1.27/arch/arm/mach-rpc/ |
H A D | dma.c | 35 unsigned int dma_len; member in struct:iomd_dma 64 end = offset + idma->dma_len; iomd_get_next_sg() 74 idma->dma_len -= end - offset; iomd_get_next_sg() 77 if (idma->dma_len == 0) { iomd_get_next_sg() 81 idma->dma_len = idma->dma.sg->length; iomd_get_next_sg()
|
/linux-4.1.27/net/rds/ |
H A D | iw_rdma.c | 256 sg->dma_len = 0; rds_iw_set_scatterlist() 268 WARN_ON(sg->dma_len); rds_iw_map_scatterlist() 270 sg->dma_len = ib_dma_map_sg(dev, sg->list, sg->len, DMA_BIDIRECTIONAL); rds_iw_map_scatterlist() 271 if (unlikely(!sg->dma_len)) { rds_iw_map_scatterlist() 280 for (i = 0; i < sg->dma_len; ++i) { rds_iw_map_scatterlist() 281 unsigned int dma_len = ib_sg_dma_len(dev, &sg->list[i]); rds_iw_map_scatterlist() local 285 sg->bytes += dma_len; rds_iw_map_scatterlist() 287 end_addr = dma_addr + dma_len; rds_iw_map_scatterlist() 294 if (i < sg->dma_len - 1) rds_iw_map_scatterlist() 312 for (i = j = 0; i < sg->dma_len; ++i) { rds_iw_map_scatterlist() 313 unsigned int dma_len = ib_sg_dma_len(dev, &sg->list[i]); rds_iw_map_scatterlist() local 317 end_addr = dma_addr + dma_len; rds_iw_map_scatterlist() 328 sg->dma_len = 0; rds_iw_map_scatterlist() 467 ibmr->mapping.m_sg.dma_len, DMA_BIDIRECTIONAL); rds_iw_sync_mr() 471 ibmr->mapping.m_sg.dma_len, DMA_BIDIRECTIONAL); rds_iw_sync_mr() 678 /* FIXME - this is overkill, but mapping->m_sg.dma_len/mapping->m_sg.dma_npages rds_iw_init_fastreg() 716 f_wr.wr.fast_reg.page_list_len = mapping->m_sg.dma_len; rds_iw_rdma_build_fastreg() 777 if (mapping->m_sg.dma_len > pool->max_message_size) { rds_iw_map_fastreg() 806 if (!ibmr->mapping.m_sg.dma_len) rds_iw_free_fastreg()
|
H A D | ib_rdma.c | 390 unsigned int dma_len = ib_sg_dma_len(dev, &scat[i]); rds_ib_map_fmr() local 399 if ((dma_addr + dma_len) & ~PAGE_MASK) { rds_ib_map_fmr() 406 len += dma_len; rds_ib_map_fmr() 420 unsigned int dma_len = ib_sg_dma_len(dev, &scat[i]); rds_ib_map_fmr() local 423 for (j = 0; j < dma_len; j += PAGE_SIZE) rds_ib_map_fmr()
|
H A D | iw.h | 57 int dma_len; member in struct:rds_iw_scatterlist
|
/linux-4.1.27/drivers/mmc/host/ |
H A D | bfin_sdh.c | 69 int dma_len; member in struct:sdh_host 158 host->dma_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, host->dma_dir); sdh_setup_data() 169 for_each_sg(data->sg, sg, host->dma_len, i) { sdh_setup_data() 183 host->dma_len * sizeof(struct dma_desc_array)); sdh_setup_data() 185 host->sg_cpu[host->dma_len - 1].cfg &= ~(DMAFLOW | NDSIZE); sdh_setup_data() 186 host->sg_cpu[host->dma_len - 1].cfg |= DI_EN; sdh_setup_data()
|
H A D | tifm_sd.c | 267 unsigned int dma_len, dma_blk_cnt, dma_off; tifm_sd_set_dma_data() local 283 dma_len = sg_dma_len(&r_data->sg[host->sg_pos]) - host->block_pos; tifm_sd_set_dma_data() 284 if (!dma_len) { tifm_sd_set_dma_data() 289 dma_len = sg_dma_len(&r_data->sg[host->sg_pos]); tifm_sd_set_dma_data() 292 if (dma_len < t_size) { tifm_sd_set_dma_data() 293 dma_blk_cnt = dma_len / r_data->blksz; tifm_sd_set_dma_data() 304 else if (dma_len) { tifm_sd_set_dma_data()
|
H A D | omap_hsmmc.c | 171 unsigned int dma_len; member in struct:omap_hsmmc_next 197 unsigned int dma_len; member in struct:omap_hsmmc_host 1290 int dma_len; omap_hsmmc_pre_dma_transfer() local 1302 dma_len = dma_map_sg(chan->device->dev, data->sg, data->sg_len, omap_hsmmc_pre_dma_transfer() 1306 dma_len = host->next_data.dma_len; omap_hsmmc_pre_dma_transfer() 1307 host->next_data.dma_len = 0; omap_hsmmc_pre_dma_transfer() 1311 if (dma_len == 0) omap_hsmmc_pre_dma_transfer() 1315 next->dma_len = dma_len; omap_hsmmc_pre_dma_transfer() 1318 host->dma_len = dma_len; omap_hsmmc_pre_dma_transfer()
|
H A D | pxamci.c | 74 unsigned int dma_len; member in struct:pxamci_host 212 host->dma_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, pxamci_setup_data() 215 for (i = 0; i < host->dma_len; i++) { pxamci_setup_data() 233 host->sg_cpu[host->dma_len - 1].ddadr = DDADR_STOP; pxamci_setup_data()
|
H A D | sunxi-mmc.c | 336 u32 i, dma_len; sunxi_mmc_map_dma() local 339 dma_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, sunxi_mmc_map_dma() 341 if (dma_len == 0) { sunxi_mmc_map_dma()
|
/linux-4.1.27/drivers/scsi/ |
H A D | am53c974.c | 268 static u32 pci_esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len) pci_esp_dma_length_limit() argument 280 if (dma_len > (1U << dma_limit)) pci_esp_dma_length_limit() 281 dma_len = (1U << dma_limit); pci_esp_dma_length_limit() 287 end = base + dma_len; pci_esp_dma_length_limit() 290 dma_len = end - base; pci_esp_dma_length_limit() 292 return dma_len; pci_esp_dma_length_limit()
|
H A D | mac53c94.c | 367 u32 dma_len; set_dma_cmds() local 381 dma_len = sg_dma_len(scl); scsi_for_each_sg() 382 if (dma_len > 0xffff) scsi_for_each_sg() 384 total += dma_len; scsi_for_each_sg() 385 dcmds->req_count = cpu_to_le16(dma_len); scsi_for_each_sg()
|
H A D | esp_scsi.c | 519 static u32 esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len) esp_dma_length_limit() argument 523 if (dma_len > (1U << 24)) esp_dma_length_limit() 524 dma_len = (1U << 24); esp_dma_length_limit() 534 if (dma_len > (1U << 16)) esp_dma_length_limit() 535 dma_len = (1U << 16); esp_dma_length_limit() 541 end = base + dma_len; esp_dma_length_limit() 544 dma_len = end - base; esp_dma_length_limit() 546 return dma_len; esp_dma_length_limit() 1758 unsigned int dma_len = esp_cur_dma_len(ent, cmd); esp_process_event() local 1769 dma_len = esp->ops->dma_length_limit(esp, dma_addr, esp_process_event() 1770 dma_len); esp_process_event() 1772 dma_len = esp_dma_length_limit(esp, dma_addr, dma_len); esp_process_event() 1774 esp->data_dma_len = dma_len; esp_process_event() 1776 if (!dma_len) { esp_process_event() 1788 (unsigned long long)dma_addr, dma_len, write); esp_process_event() 1790 esp->ops->send_dma_cmd(esp, dma_addr, dma_len, dma_len, esp_process_event()
|
H A D | mac_esp.c | 437 static u32 mac_esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len) mac_esp_dma_length_limit() argument 439 return dma_len > 0xFFFF ? 0xFFFF : dma_len; mac_esp_dma_length_limit()
|
H A D | mesh.c | 1281 u32 dma_len = sg_dma_len(scl); scsi_for_each_sg() local 1284 if (off >= dma_len) { scsi_for_each_sg() 1285 off -= dma_len; scsi_for_each_sg() 1288 if (dma_len > 0xffff) scsi_for_each_sg() 1290 dcmds->req_count = cpu_to_le16(dma_len - off); scsi_for_each_sg() 1295 dtot += dma_len - off; scsi_for_each_sg()
|
H A D | atari_NCR5380.c | 783 hostdata->dma_len = 0; NCR5380_init() 1074 && !hostdata->dma_len NCR5380_main() 1160 transferred = hostdata->dma_len - NCR5380_dma_residual(instance); NCR5380_dma_complete() 1161 hostdata->dma_len = 0; NCR5380_dma_complete() 1835 hostdata->dma_len = c; NCR5380_transfer_dma() 1898 hostdata->dma_len = (p & SR_IO) ? NCR5380_transfer_dma() 1916 hostdata->dma_len = (p & SR_IO) ? NCR5380_transfer_dma() 2920 hostdata->dma_len = 0; NCR5380_bus_reset()
|
H A D | NCR5380.h | 266 volatile int dma_len; /* requested length of DMA */ member in struct:NCR5380_hostdata
|
H A D | esp_scsi.h | 390 u32 dma_len);
|
H A D | atari_scsi.c | 165 (atari_scsi_host->hostdata))->dma_len)
|
H A D | NCR5380.c | 1748 hostdata->dma_len = (p & SR_IO) ? NCR5380_dma_read_setup(instance, d, c) : NCR5380_dma_write_setup(instance, d, c); NCR5380_transfer_dma()
|
/linux-4.1.27/drivers/net/ethernet/sfc/ |
H A D | tx.c | 342 unsigned int dma_len; efx_enqueue_skb() local 391 dma_len = efx_max_tx_len(efx, dma_addr); efx_enqueue_skb() 392 if (likely(dma_len >= len)) efx_enqueue_skb() 393 dma_len = len; efx_enqueue_skb() 396 buffer->len = dma_len; efx_enqueue_skb() 399 len -= dma_len; efx_enqueue_skb() 400 dma_addr += dma_len; efx_enqueue_skb() 930 unsigned dma_len; efx_tx_queue_insert() local 944 dma_len = efx_max_tx_len(efx, dma_addr); efx_tx_queue_insert() 947 if (dma_len >= len) efx_tx_queue_insert() 950 buffer->len = dma_len; efx_tx_queue_insert() 952 dma_addr += dma_len; efx_tx_queue_insert() 953 len -= dma_len; efx_tx_queue_insert()
|
H A D | mcdi_port.c | 932 u32 dma_len = action != EFX_STATS_DISABLE ? efx_mcdi_mac_stats() local 946 MCDI_SET_DWORD(inbuf, MAC_STATS_IN_DMA_LEN, dma_len); efx_mcdi_mac_stats()
|
/linux-4.1.27/drivers/usb/host/whci/ |
H A D | qset.c | 464 size_t dma_len; qset_add_urb_sg() local 492 dma_len = dma_remaining; qset_add_urb_sg() 500 if (std->len + dma_len > QTD_MAX_XFER_SIZE) { qset_add_urb_sg() 501 dma_len = (QTD_MAX_XFER_SIZE / qset->max_packet) qset_add_urb_sg() 505 std->len += dma_len; qset_add_urb_sg() 509 ep = dma_addr + dma_len; qset_add_urb_sg() 529 dma_remaining -= dma_len; qset_add_urb_sg() 530 remaining -= dma_len; qset_add_urb_sg()
|
/linux-4.1.27/drivers/net/wireless/p54/ |
H A D | p54pci.h | 62 __le32 dma_len; member in struct:p54p_csr
|
H A D | p54usb.c | 772 P54U_WRITE(NET2280_DEV_U32, &devreg->dma_len, p54u_upload_firmware_net2280()
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb/ |
H A D | sge.c | 165 DEFINE_DMA_UNMAP_LEN(dma_len); 171 DEFINE_DMA_UNMAP_LEN(dma_len); 513 dma_unmap_len(ce, dma_len), free_freelQ_buffers() 627 if (likely(dma_unmap_len(ce, dma_len))) { free_cmdQ_buffers() 629 dma_unmap_len(ce, dma_len), free_cmdQ_buffers() 830 unsigned int dma_len = q->rx_buffer_size - q->dma_offset; refill_free_list() local 841 mapping = pci_map_single(pdev, skb->data, dma_len, refill_free_list() 847 dma_unmap_len_set(ce, dma_len, dma_len); refill_free_list() 850 e->len_gen = V_CMD_LEN(dma_len) | V_CMD_GEN1(q->genbit); refill_free_list() 1055 dma_unmap_len(ce, dma_len), get_packet() 1060 dma_unmap_len(ce, dma_len), get_packet() 1073 dma_unmap_len(ce, dma_len), PCI_DMA_FROMDEVICE); get_packet() 1096 dma_unmap_len(ce, dma_len), PCI_DMA_FROMDEVICE); unexpected_offload() 1177 dma_unmap_len_set(ce1, dma_len, 0); write_large_page_tx_descs() 1228 dma_unmap_len_set(ce, dma_len, 0); write_tx_descs() 1253 dma_unmap_len_set(ce, dma_len, skb_headlen(skb)); write_tx_descs() 1279 dma_unmap_len_set(ce, dma_len, skb_frag_size(frag)); write_tx_descs()
|
/linux-4.1.27/drivers/crypto/ |
H A D | talitos.c | 714 * @dma_len: length of dma mapped link_tbl space 731 int dma_len; member in struct:talitos_edesc 808 if (edesc->dma_len) ipsec_esp_unmap() 809 dma_unmap_single(dev, edesc->dma_link_tbl, edesc->dma_len, ipsec_esp_unmap() 863 if (edesc->dma_len) ipsec_esp_decrypt_swauth_done() 982 edesc->dma_len, DMA_BIDIRECTIONAL); ipsec_esp() 1032 edesc->dma_len, ipsec_esp() 1075 edesc->dma_len, DMA_BIDIRECTIONAL); ipsec_esp() 1127 int assoc_nents = 0, src_nents, dst_nents, alloc_len, dma_len; talitos_edesc_alloc() local 1177 dma_len = (src_nents + dst_nents + 2 + assoc_nents) * talitos_edesc_alloc() 1179 alloc_len += dma_len; talitos_edesc_alloc() 1181 dma_len = 0; talitos_edesc_alloc() 1208 edesc->dma_len = dma_len; talitos_edesc_alloc() 1209 if (dma_len) talitos_edesc_alloc() 1211 edesc->dma_len, talitos_edesc_alloc() 1283 if (edesc->dma_len) aead_decrypt() 1341 if (edesc->dma_len) common_nonsnoop_unmap() 1342 dma_unmap_single(dev, edesc->dma_link_tbl, edesc->dma_len, common_nonsnoop_unmap() 1410 edesc->dma_len, common_nonsnoop() 1441 edesc->dma_len, DMA_BIDIRECTIONAL); common_nonsnoop() 1525 if (edesc->dma_len) common_nonsnoop_hash_unmap() 1526 dma_unmap_single(dev, edesc->dma_link_tbl, edesc->dma_len, common_nonsnoop_hash_unmap() 1608 edesc->dma_len, common_nonsnoop_hash()
|
/linux-4.1.27/drivers/infiniband/ulp/srp/ |
H A D | ib_srp.c | 1260 unsigned int dma_len, u32 rkey) srp_map_desc() 1266 desc->len = cpu_to_be32(dma_len); srp_map_desc() 1268 state->total_len += dma_len; srp_map_desc() 1287 srp_map_desc(state, 0, state->dma_len, fmr->fmr->rkey); srp_map_finish_fmr() 1319 wr.wr.fast_reg.length = state->dma_len; srp_map_finish_fr() 1328 srp_map_desc(state, state->base_dma_addr, state->dma_len, srp_map_finish_fr() 1344 srp_map_desc(state, state->base_dma_addr, state->dma_len, srp_finish_mapping() 1353 state->dma_len = 0; srp_finish_mapping() 1377 unsigned int dma_len = ib_sg_dma_len(ibdev, sg); srp_map_sg_entry() local 1381 if (!dma_len) srp_map_sg_entry() 1390 srp_map_desc(state, dma_addr, dma_len, target->rkey); srp_map_sg_entry() 1401 dma_len > dev->mr_max_size) { srp_map_sg_entry() 1406 srp_map_desc(state, dma_addr, dma_len, target->rkey); srp_map_sg_entry() 1420 while (dma_len) { srp_map_sg_entry() 1430 len = min_t(unsigned int, dma_len, dev->mr_page_size - offset); srp_map_sg_entry() 1435 state->dma_len += len; srp_map_sg_entry() 1437 dma_len -= len; srp_map_sg_entry() 1483 unsigned int dma_len; for_each_sg() local 1490 dma_len = ib_sg_dma_len(ibdev, sg); for_each_sg() 1491 dma_len -= (state->unmapped_addr - dma_addr); for_each_sg() 1494 srp_map_desc(state, dma_addr, dma_len, target->rkey); for_each_sg() 1259 srp_map_desc(struct srp_map_state *state, dma_addr_t dma_addr, unsigned int dma_len, u32 rkey) srp_map_desc() argument
|
H A D | ib_srp.h | 274 * @dma_len: Number of bytes that will be registered with the next 292 u32 dma_len; member in struct:srp_map_state
|
/linux-4.1.27/drivers/i2c/busses/ |
H A D | i2c-imx.c | 193 unsigned int dma_len; member in struct:imx_i2c_dma 351 dma->dma_len, dma->dma_data_dir); i2c_imx_dma_callback() 364 dma->dma_len, dma->dma_data_dir); i2c_imx_dma_xfer() 371 dma->dma_len, dma->dma_transfer_dir, i2c_imx_dma_xfer() 391 dma->dma_len, dma->dma_data_dir); i2c_imx_dma_xfer() 401 dma->dma_len = 0; i2c_imx_dma_free() 613 dma->dma_len = msgs->len - 1; i2c_imx_dma_write() 681 dma->dma_len = msgs->len - 2; i2c_imx_dma_read()
|
/linux-4.1.27/drivers/block/rsxx/ |
H A D | dma.c | 647 unsigned int dma_len, rsxx_queue_dma() 663 dma->sub_page.cnt = (dma_len >> 9); rsxx_queue_dma() 694 unsigned int dma_len; rsxx_dma_queue_bio() local 734 dma_len = min(bv_len, bio_for_each_segment() 739 dma_off, dma_len, bio_for_each_segment() 747 addr8 += dma_len; bio_for_each_segment() 748 bv_off += dma_len; bio_for_each_segment() 749 bv_len -= dma_len; bio_for_each_segment() 643 rsxx_queue_dma(struct rsxx_cardinfo *card, struct list_head *q, int dir, unsigned int dma_off, unsigned int dma_len, unsigned int laddr, struct page *page, unsigned int pg_off, rsxx_dma_cb cb, void *cb_data) rsxx_queue_dma() argument
|
/linux-4.1.27/drivers/infiniband/ulp/iser/ |
H A D | iser_memory.c | 253 unsigned int dma_len; iser_sg_to_page_vec() local 265 dma_len = ib_sg_dma_len(ibdev, sg); iser_sg_to_page_vec() 266 end_addr = start_addr + dma_len; iser_sg_to_page_vec() 267 total_sz += dma_len; iser_sg_to_page_vec() 346 "off:0x%x sz:0x%x dma_len:0x%x\n", iser_data_buf_dump()
|
/linux-4.1.27/drivers/ata/ |
H A D | pata_arasan_cf.c | 434 u32 xfer_cnt, sglen, dma_len, xfer_ctr; sg_xfer() local 474 dma_len = min(xfer_cnt, FIFO_SIZE); sg_xfer() 475 ret = dma_xfer(acdev, src, dest, dma_len); sg_xfer() 482 src += dma_len; sg_xfer() 484 dest += dma_len; sg_xfer() 486 sglen -= dma_len; sg_xfer() 487 xfer_cnt -= dma_len; sg_xfer()
|
/linux-4.1.27/arch/ia64/hp/common/ |
H A D | sba_iommu.c | 1274 ** one DMA stream, *add* cnt to dma_len. sba_fill_pdir() 1332 unsigned long dma_offset, dma_len; /* start/len of DMA stream */ sba_coalesce_chunks() local 1344 dma_len = vcontig_len = vcontig_end = startsg->length; sba_coalesce_chunks() 1371 if (((dma_len + dma_offset + startsg->length + ~iovp_mask) & iovp_mask) sba_coalesce_chunks() 1375 if (dma_len + startsg->length > max_seg_size) sba_coalesce_chunks() 1388 dma_len += startsg->length; sba_coalesce_chunks() 1419 dma_len += vcontig_len; sba_coalesce_chunks() 1432 dma_len = (dma_len + dma_offset + ~iovp_mask) & iovp_mask; sba_coalesce_chunks() 1433 ASSERT(dma_len <= DMA_CHUNK_SIZE); sba_coalesce_chunks() 1434 idx = sba_alloc_range(ioc, dev, dma_len); sba_coalesce_chunks() 1525 ** o dma_len will contain the number of bytes to map
|
/linux-4.1.27/drivers/media/platform/soc_camera/ |
H A D | pxa_camera.c | 293 int i, offset, dma_len, xfer_len; calculate_dma_sglen() local 298 dma_len = sg_dma_len(sg); for_each_sg() 301 xfer_len = roundup(min(dma_len - offset, size), 8); for_each_sg() 339 int dma_len = 0, xfer_len = 0; pxa_init_dma_channel() local 362 dma_len = sg_dma_len(sg); pxa_init_dma_channel() 365 xfer_len = roundup(min(dma_len - offset, size), 8); pxa_init_dma_channel() 395 * to dma_len (end on PAGE boundary). In this case, the sg element pxa_init_dma_channel() 399 if (xfer_len >= dma_len) { pxa_init_dma_channel() 400 *sg_first_ofs = xfer_len - dma_len; pxa_init_dma_channel()
|
/linux-4.1.27/drivers/net/wireless/ath/ath9k/ |
H A D | init.c | 260 u32 dma_len; ath_descdma_setup() local 263 dma_len = ndesc_skipped * desc_len; ath_descdma_setup() 264 dd->dd_desc_len += dma_len; ath_descdma_setup() 266 ndesc_skipped = ATH_DESC_4KB_BOUND_NUM_SKIPPED(dma_len); ath_descdma_setup()
|
/linux-4.1.27/drivers/net/ethernet/brocade/bna/ |
H A D | bnad.h | 226 DEFINE_DMA_UNMAP_LEN(dma_len);
|
H A D | bnad.c | 146 dma_unmap_len(&unmap->vectors[vector], dma_len), bnad_tx_buff_unmap() 3068 dma_unmap_len_set(&unmap->vectors[vect_id], dma_len, size); bnad_start_xmit()
|
/linux-4.1.27/drivers/infiniband/ulp/srpt/ |
H A D | ib_srpt.c | 1093 u32 dma_len; srpt_map_sg_to_ib_sge() local 1128 dma_len = ib_sg_dma_len(dev, &sg[0]); srpt_map_sg_to_ib_sge() 1150 if (rsize >= dma_len) { srpt_map_sg_to_ib_sge() 1151 tsize -= dma_len; srpt_map_sg_to_ib_sge() 1152 rsize -= dma_len; srpt_map_sg_to_ib_sge() 1153 raddr += dma_len; srpt_map_sg_to_ib_sge() 1159 dma_len = ib_sg_dma_len( srpt_map_sg_to_ib_sge() 1165 dma_len -= rsize; srpt_map_sg_to_ib_sge() 1197 dma_len = ib_sg_dma_len(dev, &sg[0]); srpt_map_sg_to_ib_sge() 1211 if (rsize >= dma_len) { srpt_map_sg_to_ib_sge() 1213 (tsize < dma_len) ? tsize : dma_len; srpt_map_sg_to_ib_sge() 1214 tsize -= dma_len; srpt_map_sg_to_ib_sge() 1215 rsize -= dma_len; srpt_map_sg_to_ib_sge() 1221 dma_len = ib_sg_dma_len( srpt_map_sg_to_ib_sge() 1230 dma_len -= rsize; srpt_map_sg_to_ib_sge()
|
/linux-4.1.27/drivers/block/ |
H A D | nvme-core.c | 642 int dma_len = sg_dma_len(sg); nvme_setup_prps() local 655 dma_len -= (page_size - offset); nvme_setup_prps() 656 if (dma_len) { nvme_setup_prps() 661 dma_len = sg_dma_len(sg); nvme_setup_prps() 699 dma_len -= page_size; nvme_setup_prps() 704 if (dma_len > 0) nvme_setup_prps() 706 BUG_ON(dma_len < 0); nvme_setup_prps() 709 dma_len = sg_dma_len(sg); nvme_setup_prps()
|
/linux-4.1.27/drivers/net/ethernet/broadcom/ |
H A D | bcmsysport.c | 649 dma_unmap_len(cb, dma_len), bcm_sysport_tx_reclaim_one() 656 ndev->stats.tx_bytes += dma_unmap_len(cb, dma_len); bcm_sysport_tx_reclaim_one() 658 dma_unmap_len(cb, dma_len), DMA_TO_DEVICE); bcm_sysport_tx_reclaim_one() 995 dma_unmap_len_set(cb, dma_len, skb_len); bcm_sysport_xmit()
|
H A D | bcmsysport.h | 627 DEFINE_DMA_UNMAP_LEN(dma_len);
|
/linux-4.1.27/drivers/scsi/lpfc/ |
H A D | lpfc_scsi.c | 2272 uint32_t dma_len; lpfc_bg_setup_sgl() local 2336 dma_len = sg_dma_len(sgde); scsi_for_each_sg() 2346 sgl->sge_len = cpu_to_le32(dma_len); scsi_for_each_sg() 2347 dma_offset += dma_len; scsi_for_each_sg() 2412 uint32_t dma_len; lpfc_bg_setup_sgl_prot() local 2553 dma_len = remainder; lpfc_bg_setup_sgl_prot() 2560 dma_len = protgrp_bytes - subtotal; lpfc_bg_setup_sgl_prot() 2561 split_offset += dma_len; lpfc_bg_setup_sgl_prot() 2564 subtotal += dma_len; lpfc_bg_setup_sgl_prot() 2572 sgl->sge_len = cpu_to_le32(dma_len); lpfc_bg_setup_sgl_prot() 2573 dma_offset += dma_len; lpfc_bg_setup_sgl_prot() 3240 uint32_t dma_len; lpfc_scsi_prep_dma_buf_s4() local 3292 dma_len = sg_dma_len(sgel); scsi_for_each_sg() 3303 sgl->sge_len = cpu_to_le32(dma_len); scsi_for_each_sg() 3304 dma_offset += dma_len; scsi_for_each_sg()
|
/linux-4.1.27/drivers/net/ethernet/amd/xgbe/ |
H A D | xgbe-drv.c | 1831 skb = napi_alloc_skb(napi, rdata->rx.hdr.dma_len); xgbe_create_skb() 1838 copy_len = min(rdata->rx.hdr.dma_len, copy_len); xgbe_create_skb() 1998 rdata->rx.hdr.dma_len, xgbe_rx_poll() 2011 rdata->rx.buf.dma_len, xgbe_rx_poll() 2017 put_len, rdata->rx.buf.dma_len); xgbe_rx_poll()
|
H A D | xgbe-desc.c | 302 bd->dma_len = len; xgbe_set_buffer_data()
|
H A D | xgbe.h | 295 unsigned int dma_len; member in struct:xgbe_buffer_data
|
/linux-4.1.27/drivers/net/ethernet/nvidia/ |
H A D | forcedeth.c | 722 unsigned int dma_len:31; member in struct:nv_skb_map 1827 np->put_rx_ctx->dma_len = skb_tailroom(skb); nv_alloc_rx() 1868 np->put_rx_ctx->dma_len = skb_tailroom(skb); nv_alloc_rx_optimized() 1958 np->tx_skb[i].dma_len = 0; nv_init_tx() 1983 tx_skb->dma_len, nv_unmap_txskb() 1987 tx_skb->dma_len, nv_unmap_txskb() 2025 np->tx_skb[i].dma_len = 0; nv_drain_tx() 2240 np->put_tx_ctx->dma_len = bcnt; nv_start_xmit() 2288 np->put_tx_ctx->dma_len = bcnt; nv_start_xmit() 2389 np->put_tx_ctx->dma_len = bcnt; nv_start_xmit_optimized() 2437 np->put_tx_ctx->dma_len = bcnt; nv_start_xmit_optimized() 2816 np->get_rx_ctx->dma_len, nv_rx_process() 2922 np->get_rx_ctx->dma_len, nv_rx_process_optimized()
|
/linux-4.1.27/drivers/scsi/qla2xxx/ |
H A D | qla_iocb.c | 861 uint32_t dma_len; /* OUT */ member in struct:qla2_sgx 896 sgx->dma_len = (blk_sz - cumulative_partial); qla24xx_get_one_block_sg() 901 sgx->dma_len = sg_len - sgx->bytes_consumed; qla24xx_get_one_block_sg() 902 sgx->tot_partial += sgx->dma_len; qla24xx_get_one_block_sg() 906 sgx->bytes_consumed += sgx->dma_len; qla24xx_get_one_block_sg() 963 sle_dma_len = sgx.dma_len; qla24xx_walk_and_build_sglist_no_difb()
|
/linux-4.1.27/drivers/net/ethernet/sun/ |
H A D | sunhme.c | 1930 u32 flags, dma_addr, dma_len; happy_meal_tx() local 1953 dma_len = hme_read_desc32(hp, &this->tx_flags); happy_meal_tx() 1955 dma_len &= TXFLAG_SIZE; happy_meal_tx() 1957 dma_unmap_single(hp->dma_dev, dma_addr, dma_len, DMA_TO_DEVICE); happy_meal_tx() 1959 dma_unmap_page(hp->dma_dev, dma_addr, dma_len, DMA_TO_DEVICE); happy_meal_tx()
|
H A D | sungem.c | 654 u32 dma_len; gem_tx() local 684 dma_len = le64_to_cpu(txd->control_word) & TXDCTRL_BUFSZ; gem_tx() 686 pci_unmap_page(gp->pdev, dma_addr, dma_len, PCI_DMA_TODEVICE); gem_tx()
|
/linux-4.1.27/drivers/net/ethernet/broadcom/genet/ |
H A D | bcmgenet.c | 1056 dma_unmap_len(tx_cb_ptr, dma_len); __bcmgenet_tx_reclaim() 1059 dma_unmap_len(tx_cb_ptr, dma_len), __bcmgenet_tx_reclaim() 1162 dma_unmap_len_set(tx_cb_ptr, dma_len, skb->len); bcmgenet_xmit_single() 1204 dma_unmap_len_set(tx_cb_ptr, dma_len, frag->size); bcmgenet_xmit_frag()
|
H A D | bcmgenet.h | 480 DEFINE_DMA_UNMAP_LEN(dma_len);
|
/linux-4.1.27/drivers/mtd/nand/ |
H A D | pxa3xx_nand.c | 560 int dma_len = ALIGN(info->data_size + info->oob_size, 32); start_data_dma() local 563 desc->dcmd = DCMD_ENDIRQEN | DCMD_WIDTH4 | DCMD_BURST32 | dma_len; start_data_dma()
|
/linux-4.1.27/drivers/net/ethernet/3com/ |
H A D | typhoon.c | 1509 int dma_len; typhoon_clean_tx() local 1526 dma_len = le16_to_cpu(tx->len); typhoon_clean_tx() 1527 pci_unmap_single(tp->pdev, skb_dma, dma_len, typhoon_clean_tx()
|
/linux-4.1.27/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_verbs.c | 811 u32 dma_len = mr->pbl_size; ocrdma_build_pbl_tbl() local 823 va = dma_alloc_coherent(&pdev->dev, dma_len, &pa, GFP_KERNEL); ocrdma_build_pbl_tbl() 829 memset(va, 0, dma_len); ocrdma_build_pbl_tbl()
|
/linux-4.1.27/drivers/block/mtip32xx/ |
H A D | mtip32xx.c | 1762 unsigned int dma_len; fill_command_sg() local 1769 dma_len = sg_dma_len(sg); fill_command_sg() 1770 if (dma_len > 0x400000) fill_command_sg() 1774 cpu_to_le32((dma_len-1) & 0x3FFFFF); fill_command_sg()
|
/linux-4.1.27/drivers/infiniband/ulp/isert/ |
H A D | ib_isert.c | 2407 isert_dbg("RDMA from SGL dma_addr: 0x%llx dma_len: %u, " for_each_sg()
|