Searched refs:dma_buff (Results 1 - 6 of 6) sorted by relevance

/linux-4.4.14/drivers/video/fbdev/
H A Dpxafb.h116 struct pxafb_dma_buff *dma_buff; member in struct:pxafb_info
H A Dpxafb.c521 fbi->palette_cpu = (u16 *)&fbi->dma_buff->palette[0]; pxafb_set_par()
1071 dma_desc = &fbi->dma_buff->dma_desc[dma]; setup_frame_dma()
1082 pal_desc = &fbi->dma_buff->pal_desc[pal]; setup_frame_dma()
1132 dma_desc = &fbi->dma_buff->dma_desc[DMA_CMD]; setup_smart_dma()
1300 fbi->smart_cmds = (uint16_t *) fbi->dma_buff->cmd_buff; pxafb_smart_init()
2164 fbi->dma_buff = dma_alloc_coherent(fbi->dev, fbi->dma_buff_size, pxafb_probe()
2166 if (fbi->dma_buff == NULL) { pxafb_probe()
2251 fbi->dma_buff, fbi->dma_buff_phys); pxafb_probe()
2289 fbi->dma_buff, fbi->dma_buff_phys); pxafb_remove()
/linux-4.4.14/drivers/net/ethernet/cirrus/
H A Dcs89x0.c153 unsigned char *dma_buff; /* points to the beginning of the buffer */ member in struct:net_local
452 memcpy(skb_put(skb, length - semi_cnt), lp->dma_buff, dma_rx()
475 if (lp->dma_buff) { release_dma_buff()
476 free_pages((unsigned long)(lp->dma_buff), release_dma_buff()
478 lp->dma_buff = NULL; release_dma_buff()
886 lp->dma_buff = (unsigned char *)__get_dma_pages(GFP_KERNEL, net_open()
888 if (!lp->dma_buff) { net_open()
895 (unsigned long)lp->dma_buff, net_open()
896 (unsigned long)isa_virt_to_bus(lp->dma_buff)); net_open()
897 if ((unsigned long)lp->dma_buff >= MAX_DMA_ADDRESS || net_open()
898 !dma_page_eq(lp->dma_buff, net_open()
899 lp->dma_buff + lp->dmasize * 1024 - 1)) { net_open()
903 memset(lp->dma_buff, 0, lp->dmasize * 1024); /* Why? */ net_open()
910 lp->rx_dma_ptr = lp->dma_buff; net_open()
911 lp->end_dma_buff = lp->dma_buff + lp->dmasize * 1024; net_open()
916 set_dma_addr(dev->dma, isa_virt_to_bus(lp->dma_buff)); net_open()
/linux-4.4.14/drivers/net/ethernet/intel/i40e/
H A Di40e_adminq.c743 struct i40e_dma_mem *dma_buff = NULL; i40e_asq_send_command() local
831 dma_buff = &(hw->aq.asq.r.asq_bi[hw->aq.asq.next_to_use]); i40e_asq_send_command()
833 memcpy(dma_buff->va, buff, buff_size); i40e_asq_send_command()
840 cpu_to_le32(upper_32_bits(dma_buff->pa)); i40e_asq_send_command()
842 cpu_to_le32(lower_32_bits(dma_buff->pa)); i40e_asq_send_command()
876 memcpy(buff, dma_buff->va, buff_size); i40e_asq_send_command()
/linux-4.4.14/drivers/net/ethernet/intel/i40evf/
H A Di40e_adminq.c676 struct i40e_dma_mem *dma_buff = NULL; i40evf_asq_send_command() local
764 dma_buff = &(hw->aq.asq.r.asq_bi[hw->aq.asq.next_to_use]); i40evf_asq_send_command()
766 memcpy(dma_buff->va, buff, buff_size); i40evf_asq_send_command()
773 cpu_to_le32(upper_32_bits(dma_buff->pa)); i40evf_asq_send_command()
775 cpu_to_le32(lower_32_bits(dma_buff->pa)); i40evf_asq_send_command()
809 memcpy(buff, dma_buff->va, buff_size); i40evf_asq_send_command()
/linux-4.4.14/drivers/net/ethernet/ti/
H A Dnetcp_core.c625 dma_addr_t dma_desc, dma_buff; netcp_process_one_rx_packet() local
641 get_pkt_info(&dma_buff, &buf_len, &dma_desc, desc); netcp_process_one_rx_packet()
651 dma_unmap_single(netcp->dev, dma_buff, buf_len, DMA_FROM_DEVICE); netcp_process_one_rx_packet()
674 get_pkt_info(&dma_buff, &buf_len, &dma_desc, ndesc); netcp_process_one_rx_packet()
677 if (likely(dma_buff && buf_len && page)) { netcp_process_one_rx_packet()
678 dma_unmap_page(netcp->dev, dma_buff, PAGE_SIZE, netcp_process_one_rx_packet()
681 dev_err(netcp->ndev_dev, "Bad Rx desc dma_buff(%p), len(%d), page(%p)\n", netcp_process_one_rx_packet()
682 (void *)dma_buff, buf_len, page); netcp_process_one_rx_packet()
687 offset_in_page(dma_buff), buf_len, PAGE_SIZE); netcp_process_one_rx_packet()

Completed in 266 milliseconds