Searched refs:dma_len (Results 1 - 63 of 63) sorted by relevance

/linux-4.1.27/drivers/parisc/
H A Diommu-helpers.h21 unsigned long dma_offset = 0, dma_len = 0; iommu_fill_pdir() local
46 BUG_ON(pdirp && (dma_len != sg_dma_len(dma_sg))); iommu_fill_pdir()
50 dma_len = sg_dma_len(startsg); iommu_fill_pdir()
105 unsigned long dma_offset, dma_len; /* start/len of DMA stream */ iommu_coalesce_chunks() local
119 dma_len = startsg->length; iommu_coalesce_chunks()
148 if (unlikely(ALIGN(dma_len + dma_offset + startsg->length, IOVP_SIZE) > iommu_coalesce_chunks()
162 dma_len += startsg->length; iommu_coalesce_chunks()
170 sg_dma_len(contig_sg) = dma_len; iommu_coalesce_chunks()
171 dma_len = ALIGN(dma_len + dma_offset, IOVP_SIZE); iommu_coalesce_chunks()
174 | (iommu_alloc_range(ioc, dev, dma_len) << IOVP_SHIFT) iommu_coalesce_chunks()
H A Dccio-dma.c949 ** o dma_len will contain the number of bytes to map ccio_map_sg()
H A Dsba_iommu.c983 ** o dma_len will contain the number of bytes to map sba_map_sg()
/linux-4.1.27/drivers/net/wireless/brcm80211/brcmsmac/
H A Dbrcms_trace_brcmsmac_tx.h76 u16 dma_len),
78 dma_len),
85 __field(u16, dma_len)
93 __entry->dma_len = dma_len;
95 TP_printk("[%s] ampdu session max_len=%u max_frames=%u len=%u frames=%u dma_len=%u",
98 __entry->ampdu_frames, __entry->dma_len)
H A Dampdu.h29 * dma_len: DMA length of this AMPDU
37 u16 dma_len; member in struct:brcms_ampdu_session
H A Dampdu.c509 session->dma_len = 0; brcms_c_ampdu_reset_session()
568 session->dma_len += p->len; brcms_c_ampdu_add_frame()
644 u16 dma_len = session->dma_len; brcms_c_ampdu_finalize() local
771 dma_len = min(dma_len, f->ampdu_pld_size); brcms_c_ampdu_finalize()
772 txh->PreloadSize = cpu_to_le16(dma_len); brcms_c_ampdu_finalize()
H A Ddma.c1325 session->dma_len); ampdu_finalize()
/linux-4.1.27/arch/arm/mach-rpc/
H A Ddma.c35 unsigned int dma_len; member in struct:iomd_dma
64 end = offset + idma->dma_len; iomd_get_next_sg()
74 idma->dma_len -= end - offset; iomd_get_next_sg()
77 if (idma->dma_len == 0) { iomd_get_next_sg()
81 idma->dma_len = idma->dma.sg->length; iomd_get_next_sg()
/linux-4.1.27/net/rds/
H A Diw_rdma.c256 sg->dma_len = 0; rds_iw_set_scatterlist()
268 WARN_ON(sg->dma_len); rds_iw_map_scatterlist()
270 sg->dma_len = ib_dma_map_sg(dev, sg->list, sg->len, DMA_BIDIRECTIONAL); rds_iw_map_scatterlist()
271 if (unlikely(!sg->dma_len)) { rds_iw_map_scatterlist()
280 for (i = 0; i < sg->dma_len; ++i) { rds_iw_map_scatterlist()
281 unsigned int dma_len = ib_sg_dma_len(dev, &sg->list[i]); rds_iw_map_scatterlist() local
285 sg->bytes += dma_len; rds_iw_map_scatterlist()
287 end_addr = dma_addr + dma_len; rds_iw_map_scatterlist()
294 if (i < sg->dma_len - 1) rds_iw_map_scatterlist()
312 for (i = j = 0; i < sg->dma_len; ++i) { rds_iw_map_scatterlist()
313 unsigned int dma_len = ib_sg_dma_len(dev, &sg->list[i]); rds_iw_map_scatterlist() local
317 end_addr = dma_addr + dma_len; rds_iw_map_scatterlist()
328 sg->dma_len = 0; rds_iw_map_scatterlist()
467 ibmr->mapping.m_sg.dma_len, DMA_BIDIRECTIONAL); rds_iw_sync_mr()
471 ibmr->mapping.m_sg.dma_len, DMA_BIDIRECTIONAL); rds_iw_sync_mr()
678 /* FIXME - this is overkill, but mapping->m_sg.dma_len/mapping->m_sg.dma_npages rds_iw_init_fastreg()
716 f_wr.wr.fast_reg.page_list_len = mapping->m_sg.dma_len; rds_iw_rdma_build_fastreg()
777 if (mapping->m_sg.dma_len > pool->max_message_size) { rds_iw_map_fastreg()
806 if (!ibmr->mapping.m_sg.dma_len) rds_iw_free_fastreg()
H A Dib_rdma.c390 unsigned int dma_len = ib_sg_dma_len(dev, &scat[i]); rds_ib_map_fmr() local
399 if ((dma_addr + dma_len) & ~PAGE_MASK) { rds_ib_map_fmr()
406 len += dma_len; rds_ib_map_fmr()
420 unsigned int dma_len = ib_sg_dma_len(dev, &scat[i]); rds_ib_map_fmr() local
423 for (j = 0; j < dma_len; j += PAGE_SIZE) rds_ib_map_fmr()
H A Diw.h57 int dma_len; member in struct:rds_iw_scatterlist
/linux-4.1.27/drivers/mmc/host/
H A Dbfin_sdh.c69 int dma_len; member in struct:sdh_host
158 host->dma_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, host->dma_dir); sdh_setup_data()
169 for_each_sg(data->sg, sg, host->dma_len, i) { sdh_setup_data()
183 host->dma_len * sizeof(struct dma_desc_array)); sdh_setup_data()
185 host->sg_cpu[host->dma_len - 1].cfg &= ~(DMAFLOW | NDSIZE); sdh_setup_data()
186 host->sg_cpu[host->dma_len - 1].cfg |= DI_EN; sdh_setup_data()
H A Dtifm_sd.c267 unsigned int dma_len, dma_blk_cnt, dma_off; tifm_sd_set_dma_data() local
283 dma_len = sg_dma_len(&r_data->sg[host->sg_pos]) - host->block_pos; tifm_sd_set_dma_data()
284 if (!dma_len) { tifm_sd_set_dma_data()
289 dma_len = sg_dma_len(&r_data->sg[host->sg_pos]); tifm_sd_set_dma_data()
292 if (dma_len < t_size) { tifm_sd_set_dma_data()
293 dma_blk_cnt = dma_len / r_data->blksz; tifm_sd_set_dma_data()
304 else if (dma_len) { tifm_sd_set_dma_data()
H A Domap_hsmmc.c171 unsigned int dma_len; member in struct:omap_hsmmc_next
197 unsigned int dma_len; member in struct:omap_hsmmc_host
1290 int dma_len; omap_hsmmc_pre_dma_transfer() local
1302 dma_len = dma_map_sg(chan->device->dev, data->sg, data->sg_len, omap_hsmmc_pre_dma_transfer()
1306 dma_len = host->next_data.dma_len; omap_hsmmc_pre_dma_transfer()
1307 host->next_data.dma_len = 0; omap_hsmmc_pre_dma_transfer()
1311 if (dma_len == 0) omap_hsmmc_pre_dma_transfer()
1315 next->dma_len = dma_len; omap_hsmmc_pre_dma_transfer()
1318 host->dma_len = dma_len; omap_hsmmc_pre_dma_transfer()
H A Dpxamci.c74 unsigned int dma_len; member in struct:pxamci_host
212 host->dma_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, pxamci_setup_data()
215 for (i = 0; i < host->dma_len; i++) { pxamci_setup_data()
233 host->sg_cpu[host->dma_len - 1].ddadr = DDADR_STOP; pxamci_setup_data()
H A Dsunxi-mmc.c336 u32 i, dma_len; sunxi_mmc_map_dma() local
339 dma_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len, sunxi_mmc_map_dma()
341 if (dma_len == 0) { sunxi_mmc_map_dma()
/linux-4.1.27/drivers/scsi/
H A Dam53c974.c268 static u32 pci_esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len) pci_esp_dma_length_limit() argument
280 if (dma_len > (1U << dma_limit)) pci_esp_dma_length_limit()
281 dma_len = (1U << dma_limit); pci_esp_dma_length_limit()
287 end = base + dma_len; pci_esp_dma_length_limit()
290 dma_len = end - base; pci_esp_dma_length_limit()
292 return dma_len; pci_esp_dma_length_limit()
H A Dmac53c94.c367 u32 dma_len; set_dma_cmds() local
381 dma_len = sg_dma_len(scl); scsi_for_each_sg()
382 if (dma_len > 0xffff) scsi_for_each_sg()
384 total += dma_len; scsi_for_each_sg()
385 dcmds->req_count = cpu_to_le16(dma_len); scsi_for_each_sg()
H A Desp_scsi.c519 static u32 esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len) esp_dma_length_limit() argument
523 if (dma_len > (1U << 24)) esp_dma_length_limit()
524 dma_len = (1U << 24); esp_dma_length_limit()
534 if (dma_len > (1U << 16)) esp_dma_length_limit()
535 dma_len = (1U << 16); esp_dma_length_limit()
541 end = base + dma_len; esp_dma_length_limit()
544 dma_len = end - base; esp_dma_length_limit()
546 return dma_len; esp_dma_length_limit()
1758 unsigned int dma_len = esp_cur_dma_len(ent, cmd); esp_process_event() local
1769 dma_len = esp->ops->dma_length_limit(esp, dma_addr, esp_process_event()
1770 dma_len); esp_process_event()
1772 dma_len = esp_dma_length_limit(esp, dma_addr, dma_len); esp_process_event()
1774 esp->data_dma_len = dma_len; esp_process_event()
1776 if (!dma_len) { esp_process_event()
1788 (unsigned long long)dma_addr, dma_len, write); esp_process_event()
1790 esp->ops->send_dma_cmd(esp, dma_addr, dma_len, dma_len, esp_process_event()
H A Dmac_esp.c437 static u32 mac_esp_dma_length_limit(struct esp *esp, u32 dma_addr, u32 dma_len) mac_esp_dma_length_limit() argument
439 return dma_len > 0xFFFF ? 0xFFFF : dma_len; mac_esp_dma_length_limit()
H A Dmesh.c1281 u32 dma_len = sg_dma_len(scl); scsi_for_each_sg() local
1284 if (off >= dma_len) { scsi_for_each_sg()
1285 off -= dma_len; scsi_for_each_sg()
1288 if (dma_len > 0xffff) scsi_for_each_sg()
1290 dcmds->req_count = cpu_to_le16(dma_len - off); scsi_for_each_sg()
1295 dtot += dma_len - off; scsi_for_each_sg()
H A Datari_NCR5380.c783 hostdata->dma_len = 0; NCR5380_init()
1074 && !hostdata->dma_len NCR5380_main()
1160 transferred = hostdata->dma_len - NCR5380_dma_residual(instance); NCR5380_dma_complete()
1161 hostdata->dma_len = 0; NCR5380_dma_complete()
1835 hostdata->dma_len = c; NCR5380_transfer_dma()
1898 hostdata->dma_len = (p & SR_IO) ? NCR5380_transfer_dma()
1916 hostdata->dma_len = (p & SR_IO) ? NCR5380_transfer_dma()
2920 hostdata->dma_len = 0; NCR5380_bus_reset()
H A DNCR5380.h266 volatile int dma_len; /* requested length of DMA */ member in struct:NCR5380_hostdata
H A Desp_scsi.h390 u32 dma_len);
H A Datari_scsi.c165 (atari_scsi_host->hostdata))->dma_len)
H A DNCR5380.c1748 hostdata->dma_len = (p & SR_IO) ? NCR5380_dma_read_setup(instance, d, c) : NCR5380_dma_write_setup(instance, d, c); NCR5380_transfer_dma()
/linux-4.1.27/drivers/net/ethernet/sfc/
H A Dtx.c342 unsigned int dma_len; efx_enqueue_skb() local
391 dma_len = efx_max_tx_len(efx, dma_addr); efx_enqueue_skb()
392 if (likely(dma_len >= len)) efx_enqueue_skb()
393 dma_len = len; efx_enqueue_skb()
396 buffer->len = dma_len; efx_enqueue_skb()
399 len -= dma_len; efx_enqueue_skb()
400 dma_addr += dma_len; efx_enqueue_skb()
930 unsigned dma_len; efx_tx_queue_insert() local
944 dma_len = efx_max_tx_len(efx, dma_addr); efx_tx_queue_insert()
947 if (dma_len >= len) efx_tx_queue_insert()
950 buffer->len = dma_len; efx_tx_queue_insert()
952 dma_addr += dma_len; efx_tx_queue_insert()
953 len -= dma_len; efx_tx_queue_insert()
H A Dmcdi_port.c932 u32 dma_len = action != EFX_STATS_DISABLE ? efx_mcdi_mac_stats() local
946 MCDI_SET_DWORD(inbuf, MAC_STATS_IN_DMA_LEN, dma_len); efx_mcdi_mac_stats()
/linux-4.1.27/drivers/usb/host/whci/
H A Dqset.c464 size_t dma_len; qset_add_urb_sg() local
492 dma_len = dma_remaining; qset_add_urb_sg()
500 if (std->len + dma_len > QTD_MAX_XFER_SIZE) { qset_add_urb_sg()
501 dma_len = (QTD_MAX_XFER_SIZE / qset->max_packet) qset_add_urb_sg()
505 std->len += dma_len; qset_add_urb_sg()
509 ep = dma_addr + dma_len; qset_add_urb_sg()
529 dma_remaining -= dma_len; qset_add_urb_sg()
530 remaining -= dma_len; qset_add_urb_sg()
/linux-4.1.27/drivers/net/wireless/p54/
H A Dp54pci.h62 __le32 dma_len; member in struct:p54p_csr
H A Dp54usb.c772 P54U_WRITE(NET2280_DEV_U32, &devreg->dma_len, p54u_upload_firmware_net2280()
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb/
H A Dsge.c165 DEFINE_DMA_UNMAP_LEN(dma_len);
171 DEFINE_DMA_UNMAP_LEN(dma_len);
513 dma_unmap_len(ce, dma_len), free_freelQ_buffers()
627 if (likely(dma_unmap_len(ce, dma_len))) { free_cmdQ_buffers()
629 dma_unmap_len(ce, dma_len), free_cmdQ_buffers()
830 unsigned int dma_len = q->rx_buffer_size - q->dma_offset; refill_free_list() local
841 mapping = pci_map_single(pdev, skb->data, dma_len, refill_free_list()
847 dma_unmap_len_set(ce, dma_len, dma_len); refill_free_list()
850 e->len_gen = V_CMD_LEN(dma_len) | V_CMD_GEN1(q->genbit); refill_free_list()
1055 dma_unmap_len(ce, dma_len), get_packet()
1060 dma_unmap_len(ce, dma_len), get_packet()
1073 dma_unmap_len(ce, dma_len), PCI_DMA_FROMDEVICE); get_packet()
1096 dma_unmap_len(ce, dma_len), PCI_DMA_FROMDEVICE); unexpected_offload()
1177 dma_unmap_len_set(ce1, dma_len, 0); write_large_page_tx_descs()
1228 dma_unmap_len_set(ce, dma_len, 0); write_tx_descs()
1253 dma_unmap_len_set(ce, dma_len, skb_headlen(skb)); write_tx_descs()
1279 dma_unmap_len_set(ce, dma_len, skb_frag_size(frag)); write_tx_descs()
/linux-4.1.27/drivers/crypto/
H A Dtalitos.c714 * @dma_len: length of dma mapped link_tbl space
731 int dma_len; member in struct:talitos_edesc
808 if (edesc->dma_len) ipsec_esp_unmap()
809 dma_unmap_single(dev, edesc->dma_link_tbl, edesc->dma_len, ipsec_esp_unmap()
863 if (edesc->dma_len) ipsec_esp_decrypt_swauth_done()
982 edesc->dma_len, DMA_BIDIRECTIONAL); ipsec_esp()
1032 edesc->dma_len, ipsec_esp()
1075 edesc->dma_len, DMA_BIDIRECTIONAL); ipsec_esp()
1127 int assoc_nents = 0, src_nents, dst_nents, alloc_len, dma_len; talitos_edesc_alloc() local
1177 dma_len = (src_nents + dst_nents + 2 + assoc_nents) * talitos_edesc_alloc()
1179 alloc_len += dma_len; talitos_edesc_alloc()
1181 dma_len = 0; talitos_edesc_alloc()
1208 edesc->dma_len = dma_len; talitos_edesc_alloc()
1209 if (dma_len) talitos_edesc_alloc()
1211 edesc->dma_len, talitos_edesc_alloc()
1283 if (edesc->dma_len) aead_decrypt()
1341 if (edesc->dma_len) common_nonsnoop_unmap()
1342 dma_unmap_single(dev, edesc->dma_link_tbl, edesc->dma_len, common_nonsnoop_unmap()
1410 edesc->dma_len, common_nonsnoop()
1441 edesc->dma_len, DMA_BIDIRECTIONAL); common_nonsnoop()
1525 if (edesc->dma_len) common_nonsnoop_hash_unmap()
1526 dma_unmap_single(dev, edesc->dma_link_tbl, edesc->dma_len, common_nonsnoop_hash_unmap()
1608 edesc->dma_len, common_nonsnoop_hash()
/linux-4.1.27/drivers/infiniband/ulp/srp/
H A Dib_srp.c1260 unsigned int dma_len, u32 rkey) srp_map_desc()
1266 desc->len = cpu_to_be32(dma_len); srp_map_desc()
1268 state->total_len += dma_len; srp_map_desc()
1287 srp_map_desc(state, 0, state->dma_len, fmr->fmr->rkey); srp_map_finish_fmr()
1319 wr.wr.fast_reg.length = state->dma_len; srp_map_finish_fr()
1328 srp_map_desc(state, state->base_dma_addr, state->dma_len, srp_map_finish_fr()
1344 srp_map_desc(state, state->base_dma_addr, state->dma_len, srp_finish_mapping()
1353 state->dma_len = 0; srp_finish_mapping()
1377 unsigned int dma_len = ib_sg_dma_len(ibdev, sg); srp_map_sg_entry() local
1381 if (!dma_len) srp_map_sg_entry()
1390 srp_map_desc(state, dma_addr, dma_len, target->rkey); srp_map_sg_entry()
1401 dma_len > dev->mr_max_size) { srp_map_sg_entry()
1406 srp_map_desc(state, dma_addr, dma_len, target->rkey); srp_map_sg_entry()
1420 while (dma_len) { srp_map_sg_entry()
1430 len = min_t(unsigned int, dma_len, dev->mr_page_size - offset); srp_map_sg_entry()
1435 state->dma_len += len; srp_map_sg_entry()
1437 dma_len -= len; srp_map_sg_entry()
1483 unsigned int dma_len; for_each_sg() local
1490 dma_len = ib_sg_dma_len(ibdev, sg); for_each_sg()
1491 dma_len -= (state->unmapped_addr - dma_addr); for_each_sg()
1494 srp_map_desc(state, dma_addr, dma_len, target->rkey); for_each_sg()
1259 srp_map_desc(struct srp_map_state *state, dma_addr_t dma_addr, unsigned int dma_len, u32 rkey) srp_map_desc() argument
H A Dib_srp.h274 * @dma_len: Number of bytes that will be registered with the next
292 u32 dma_len; member in struct:srp_map_state
/linux-4.1.27/drivers/i2c/busses/
H A Di2c-imx.c193 unsigned int dma_len; member in struct:imx_i2c_dma
351 dma->dma_len, dma->dma_data_dir); i2c_imx_dma_callback()
364 dma->dma_len, dma->dma_data_dir); i2c_imx_dma_xfer()
371 dma->dma_len, dma->dma_transfer_dir, i2c_imx_dma_xfer()
391 dma->dma_len, dma->dma_data_dir); i2c_imx_dma_xfer()
401 dma->dma_len = 0; i2c_imx_dma_free()
613 dma->dma_len = msgs->len - 1; i2c_imx_dma_write()
681 dma->dma_len = msgs->len - 2; i2c_imx_dma_read()
/linux-4.1.27/drivers/block/rsxx/
H A Ddma.c647 unsigned int dma_len, rsxx_queue_dma()
663 dma->sub_page.cnt = (dma_len >> 9); rsxx_queue_dma()
694 unsigned int dma_len; rsxx_dma_queue_bio() local
734 dma_len = min(bv_len, bio_for_each_segment()
739 dma_off, dma_len, bio_for_each_segment()
747 addr8 += dma_len; bio_for_each_segment()
748 bv_off += dma_len; bio_for_each_segment()
749 bv_len -= dma_len; bio_for_each_segment()
643 rsxx_queue_dma(struct rsxx_cardinfo *card, struct list_head *q, int dir, unsigned int dma_off, unsigned int dma_len, unsigned int laddr, struct page *page, unsigned int pg_off, rsxx_dma_cb cb, void *cb_data) rsxx_queue_dma() argument
/linux-4.1.27/drivers/infiniband/ulp/iser/
H A Diser_memory.c253 unsigned int dma_len; iser_sg_to_page_vec() local
265 dma_len = ib_sg_dma_len(ibdev, sg); iser_sg_to_page_vec()
266 end_addr = start_addr + dma_len; iser_sg_to_page_vec()
267 total_sz += dma_len; iser_sg_to_page_vec()
346 "off:0x%x sz:0x%x dma_len:0x%x\n", iser_data_buf_dump()
/linux-4.1.27/drivers/ata/
H A Dpata_arasan_cf.c434 u32 xfer_cnt, sglen, dma_len, xfer_ctr; sg_xfer() local
474 dma_len = min(xfer_cnt, FIFO_SIZE); sg_xfer()
475 ret = dma_xfer(acdev, src, dest, dma_len); sg_xfer()
482 src += dma_len; sg_xfer()
484 dest += dma_len; sg_xfer()
486 sglen -= dma_len; sg_xfer()
487 xfer_cnt -= dma_len; sg_xfer()
/linux-4.1.27/arch/ia64/hp/common/
H A Dsba_iommu.c1274 ** one DMA stream, *add* cnt to dma_len. sba_fill_pdir()
1332 unsigned long dma_offset, dma_len; /* start/len of DMA stream */ sba_coalesce_chunks() local
1344 dma_len = vcontig_len = vcontig_end = startsg->length; sba_coalesce_chunks()
1371 if (((dma_len + dma_offset + startsg->length + ~iovp_mask) & iovp_mask) sba_coalesce_chunks()
1375 if (dma_len + startsg->length > max_seg_size) sba_coalesce_chunks()
1388 dma_len += startsg->length; sba_coalesce_chunks()
1419 dma_len += vcontig_len; sba_coalesce_chunks()
1432 dma_len = (dma_len + dma_offset + ~iovp_mask) & iovp_mask; sba_coalesce_chunks()
1433 ASSERT(dma_len <= DMA_CHUNK_SIZE); sba_coalesce_chunks()
1434 idx = sba_alloc_range(ioc, dev, dma_len); sba_coalesce_chunks()
1525 ** o dma_len will contain the number of bytes to map
/linux-4.1.27/drivers/media/platform/soc_camera/
H A Dpxa_camera.c293 int i, offset, dma_len, xfer_len; calculate_dma_sglen() local
298 dma_len = sg_dma_len(sg); for_each_sg()
301 xfer_len = roundup(min(dma_len - offset, size), 8); for_each_sg()
339 int dma_len = 0, xfer_len = 0; pxa_init_dma_channel() local
362 dma_len = sg_dma_len(sg); pxa_init_dma_channel()
365 xfer_len = roundup(min(dma_len - offset, size), 8); pxa_init_dma_channel()
395 * to dma_len (end on PAGE boundary). In this case, the sg element pxa_init_dma_channel()
399 if (xfer_len >= dma_len) { pxa_init_dma_channel()
400 *sg_first_ofs = xfer_len - dma_len; pxa_init_dma_channel()
/linux-4.1.27/drivers/net/wireless/ath/ath9k/
H A Dinit.c260 u32 dma_len; ath_descdma_setup() local
263 dma_len = ndesc_skipped * desc_len; ath_descdma_setup()
264 dd->dd_desc_len += dma_len; ath_descdma_setup()
266 ndesc_skipped = ATH_DESC_4KB_BOUND_NUM_SKIPPED(dma_len); ath_descdma_setup()
/linux-4.1.27/drivers/net/ethernet/brocade/bna/
H A Dbnad.h226 DEFINE_DMA_UNMAP_LEN(dma_len);
H A Dbnad.c146 dma_unmap_len(&unmap->vectors[vector], dma_len), bnad_tx_buff_unmap()
3068 dma_unmap_len_set(&unmap->vectors[vect_id], dma_len, size); bnad_start_xmit()
/linux-4.1.27/drivers/infiniband/ulp/srpt/
H A Dib_srpt.c1093 u32 dma_len; srpt_map_sg_to_ib_sge() local
1128 dma_len = ib_sg_dma_len(dev, &sg[0]); srpt_map_sg_to_ib_sge()
1150 if (rsize >= dma_len) { srpt_map_sg_to_ib_sge()
1151 tsize -= dma_len; srpt_map_sg_to_ib_sge()
1152 rsize -= dma_len; srpt_map_sg_to_ib_sge()
1153 raddr += dma_len; srpt_map_sg_to_ib_sge()
1159 dma_len = ib_sg_dma_len( srpt_map_sg_to_ib_sge()
1165 dma_len -= rsize; srpt_map_sg_to_ib_sge()
1197 dma_len = ib_sg_dma_len(dev, &sg[0]); srpt_map_sg_to_ib_sge()
1211 if (rsize >= dma_len) { srpt_map_sg_to_ib_sge()
1213 (tsize < dma_len) ? tsize : dma_len; srpt_map_sg_to_ib_sge()
1214 tsize -= dma_len; srpt_map_sg_to_ib_sge()
1215 rsize -= dma_len; srpt_map_sg_to_ib_sge()
1221 dma_len = ib_sg_dma_len( srpt_map_sg_to_ib_sge()
1230 dma_len -= rsize; srpt_map_sg_to_ib_sge()
/linux-4.1.27/drivers/block/
H A Dnvme-core.c642 int dma_len = sg_dma_len(sg); nvme_setup_prps() local
655 dma_len -= (page_size - offset); nvme_setup_prps()
656 if (dma_len) { nvme_setup_prps()
661 dma_len = sg_dma_len(sg); nvme_setup_prps()
699 dma_len -= page_size; nvme_setup_prps()
704 if (dma_len > 0) nvme_setup_prps()
706 BUG_ON(dma_len < 0); nvme_setup_prps()
709 dma_len = sg_dma_len(sg); nvme_setup_prps()
/linux-4.1.27/drivers/net/ethernet/broadcom/
H A Dbcmsysport.c649 dma_unmap_len(cb, dma_len), bcm_sysport_tx_reclaim_one()
656 ndev->stats.tx_bytes += dma_unmap_len(cb, dma_len); bcm_sysport_tx_reclaim_one()
658 dma_unmap_len(cb, dma_len), DMA_TO_DEVICE); bcm_sysport_tx_reclaim_one()
995 dma_unmap_len_set(cb, dma_len, skb_len); bcm_sysport_xmit()
H A Dbcmsysport.h627 DEFINE_DMA_UNMAP_LEN(dma_len);
/linux-4.1.27/drivers/scsi/lpfc/
H A Dlpfc_scsi.c2272 uint32_t dma_len; lpfc_bg_setup_sgl() local
2336 dma_len = sg_dma_len(sgde); scsi_for_each_sg()
2346 sgl->sge_len = cpu_to_le32(dma_len); scsi_for_each_sg()
2347 dma_offset += dma_len; scsi_for_each_sg()
2412 uint32_t dma_len; lpfc_bg_setup_sgl_prot() local
2553 dma_len = remainder; lpfc_bg_setup_sgl_prot()
2560 dma_len = protgrp_bytes - subtotal; lpfc_bg_setup_sgl_prot()
2561 split_offset += dma_len; lpfc_bg_setup_sgl_prot()
2564 subtotal += dma_len; lpfc_bg_setup_sgl_prot()
2572 sgl->sge_len = cpu_to_le32(dma_len); lpfc_bg_setup_sgl_prot()
2573 dma_offset += dma_len; lpfc_bg_setup_sgl_prot()
3240 uint32_t dma_len; lpfc_scsi_prep_dma_buf_s4() local
3292 dma_len = sg_dma_len(sgel); scsi_for_each_sg()
3303 sgl->sge_len = cpu_to_le32(dma_len); scsi_for_each_sg()
3304 dma_offset += dma_len; scsi_for_each_sg()
/linux-4.1.27/drivers/net/ethernet/amd/xgbe/
H A Dxgbe-drv.c1831 skb = napi_alloc_skb(napi, rdata->rx.hdr.dma_len); xgbe_create_skb()
1838 copy_len = min(rdata->rx.hdr.dma_len, copy_len); xgbe_create_skb()
1998 rdata->rx.hdr.dma_len, xgbe_rx_poll()
2011 rdata->rx.buf.dma_len, xgbe_rx_poll()
2017 put_len, rdata->rx.buf.dma_len); xgbe_rx_poll()
H A Dxgbe-desc.c302 bd->dma_len = len; xgbe_set_buffer_data()
H A Dxgbe.h295 unsigned int dma_len; member in struct:xgbe_buffer_data
/linux-4.1.27/drivers/net/ethernet/nvidia/
H A Dforcedeth.c722 unsigned int dma_len:31; member in struct:nv_skb_map
1827 np->put_rx_ctx->dma_len = skb_tailroom(skb); nv_alloc_rx()
1868 np->put_rx_ctx->dma_len = skb_tailroom(skb); nv_alloc_rx_optimized()
1958 np->tx_skb[i].dma_len = 0; nv_init_tx()
1983 tx_skb->dma_len, nv_unmap_txskb()
1987 tx_skb->dma_len, nv_unmap_txskb()
2025 np->tx_skb[i].dma_len = 0; nv_drain_tx()
2240 np->put_tx_ctx->dma_len = bcnt; nv_start_xmit()
2288 np->put_tx_ctx->dma_len = bcnt; nv_start_xmit()
2389 np->put_tx_ctx->dma_len = bcnt; nv_start_xmit_optimized()
2437 np->put_tx_ctx->dma_len = bcnt; nv_start_xmit_optimized()
2816 np->get_rx_ctx->dma_len, nv_rx_process()
2922 np->get_rx_ctx->dma_len, nv_rx_process_optimized()
/linux-4.1.27/drivers/scsi/qla2xxx/
H A Dqla_iocb.c861 uint32_t dma_len; /* OUT */ member in struct:qla2_sgx
896 sgx->dma_len = (blk_sz - cumulative_partial); qla24xx_get_one_block_sg()
901 sgx->dma_len = sg_len - sgx->bytes_consumed; qla24xx_get_one_block_sg()
902 sgx->tot_partial += sgx->dma_len; qla24xx_get_one_block_sg()
906 sgx->bytes_consumed += sgx->dma_len; qla24xx_get_one_block_sg()
963 sle_dma_len = sgx.dma_len; qla24xx_walk_and_build_sglist_no_difb()
/linux-4.1.27/drivers/net/ethernet/sun/
H A Dsunhme.c1930 u32 flags, dma_addr, dma_len; happy_meal_tx() local
1953 dma_len = hme_read_desc32(hp, &this->tx_flags); happy_meal_tx()
1955 dma_len &= TXFLAG_SIZE; happy_meal_tx()
1957 dma_unmap_single(hp->dma_dev, dma_addr, dma_len, DMA_TO_DEVICE); happy_meal_tx()
1959 dma_unmap_page(hp->dma_dev, dma_addr, dma_len, DMA_TO_DEVICE); happy_meal_tx()
H A Dsungem.c654 u32 dma_len; gem_tx() local
684 dma_len = le64_to_cpu(txd->control_word) & TXDCTRL_BUFSZ; gem_tx()
686 pci_unmap_page(gp->pdev, dma_addr, dma_len, PCI_DMA_TODEVICE); gem_tx()
/linux-4.1.27/drivers/net/ethernet/broadcom/genet/
H A Dbcmgenet.c1056 dma_unmap_len(tx_cb_ptr, dma_len); __bcmgenet_tx_reclaim()
1059 dma_unmap_len(tx_cb_ptr, dma_len), __bcmgenet_tx_reclaim()
1162 dma_unmap_len_set(tx_cb_ptr, dma_len, skb->len); bcmgenet_xmit_single()
1204 dma_unmap_len_set(tx_cb_ptr, dma_len, frag->size); bcmgenet_xmit_frag()
H A Dbcmgenet.h480 DEFINE_DMA_UNMAP_LEN(dma_len);
/linux-4.1.27/drivers/mtd/nand/
H A Dpxa3xx_nand.c560 int dma_len = ALIGN(info->data_size + info->oob_size, 32); start_data_dma() local
563 desc->dcmd = DCMD_ENDIRQEN | DCMD_WIDTH4 | DCMD_BURST32 | dma_len; start_data_dma()
/linux-4.1.27/drivers/net/ethernet/3com/
H A Dtyphoon.c1509 int dma_len; typhoon_clean_tx() local
1526 dma_len = le16_to_cpu(tx->len); typhoon_clean_tx()
1527 pci_unmap_single(tp->pdev, skb_dma, dma_len, typhoon_clean_tx()
/linux-4.1.27/drivers/infiniband/hw/ocrdma/
H A Docrdma_verbs.c811 u32 dma_len = mr->pbl_size; ocrdma_build_pbl_tbl() local
823 va = dma_alloc_coherent(&pdev->dev, dma_len, &pa, GFP_KERNEL); ocrdma_build_pbl_tbl()
829 memset(va, 0, dma_len); ocrdma_build_pbl_tbl()
/linux-4.1.27/drivers/block/mtip32xx/
H A Dmtip32xx.c1762 unsigned int dma_len; fill_command_sg() local
1769 dma_len = sg_dma_len(sg); fill_command_sg()
1770 if (dma_len > 0x400000) fill_command_sg()
1774 cpu_to_le32((dma_len-1) & 0x3FFFFF); fill_command_sg()
/linux-4.1.27/drivers/infiniband/ulp/isert/
H A Dib_isert.c2407 isert_dbg("RDMA from SGL dma_addr: 0x%llx dma_len: %u, " for_each_sg()

Completed in 2109 milliseconds