mmio_read 290 drivers/gpu/drm/i915/intel_uncore.h return uncore->funcs.mmio_read##s__(uncore, reg, (trace__)); \ mmio_read 84 drivers/infiniband/hw/efa/efa_com.c struct efa_com_mmio_read *mmio_read = &edev->mmio_read; mmio_read 90 drivers/infiniband/hw/efa/efa_com.c read_resp = mmio_read->read_resp; mmio_read 92 drivers/infiniband/hw/efa/efa_com.c spin_lock(&mmio_read->lock); mmio_read 93 drivers/infiniband/hw/efa/efa_com.c mmio_read->seq_num++; mmio_read 96 drivers/infiniband/hw/efa/efa_com.c read_resp->req_id = mmio_read->seq_num + 0x9aL; mmio_read 99 drivers/infiniband/hw/efa/efa_com.c mmio_read_reg |= mmio_read->seq_num & mmio_read 104 drivers/infiniband/hw/efa/efa_com.c exp_time = jiffies + usecs_to_jiffies(mmio_read->mmio_read_timeout); mmio_read 106 drivers/infiniband/hw/efa/efa_com.c if (READ_ONCE(read_resp->req_id) == mmio_read->seq_num) mmio_read 111 drivers/infiniband/hw/efa/efa_com.c if (read_resp->req_id != mmio_read->seq_num) { mmio_read 115 drivers/infiniband/hw/efa/efa_com.c mmio_read->seq_num, offset, read_resp->req_id, mmio_read 131 drivers/infiniband/hw/efa/efa_com.c spin_unlock(&mmio_read->lock); mmio_read 901 drivers/infiniband/hw/efa/efa_com.c struct efa_com_mmio_read *mmio_read = &edev->mmio_read; mmio_read 906 drivers/infiniband/hw/efa/efa_com.c addr_high = (mmio_read->read_resp_dma_addr >> 32) & GENMASK(31, 0); mmio_read 907 drivers/infiniband/hw/efa/efa_com.c addr_low = mmio_read->read_resp_dma_addr & GENMASK(31, 0); mmio_read 915 drivers/infiniband/hw/efa/efa_com.c struct efa_com_mmio_read *mmio_read = &edev->mmio_read; mmio_read 917 drivers/infiniband/hw/efa/efa_com.c spin_lock_init(&mmio_read->lock); mmio_read 918 drivers/infiniband/hw/efa/efa_com.c mmio_read->read_resp = mmio_read 919 drivers/infiniband/hw/efa/efa_com.c dma_alloc_coherent(edev->dmadev, sizeof(*mmio_read->read_resp), mmio_read 920 drivers/infiniband/hw/efa/efa_com.c &mmio_read->read_resp_dma_addr, GFP_KERNEL); mmio_read 921 drivers/infiniband/hw/efa/efa_com.c if (!mmio_read->read_resp) mmio_read 926 drivers/infiniband/hw/efa/efa_com.c mmio_read->read_resp->req_id = 0; mmio_read 927 drivers/infiniband/hw/efa/efa_com.c mmio_read->seq_num = 0; mmio_read 928 drivers/infiniband/hw/efa/efa_com.c mmio_read->mmio_read_timeout = EFA_REG_READ_TIMEOUT_US; mmio_read 935 drivers/infiniband/hw/efa/efa_com.c struct efa_com_mmio_read *mmio_read = &edev->mmio_read; mmio_read 937 drivers/infiniband/hw/efa/efa_com.c dma_free_coherent(edev->dmadev, sizeof(*mmio_read->read_resp), mmio_read 938 drivers/infiniband/hw/efa/efa_com.c mmio_read->read_resp, mmio_read->read_resp_dma_addr); mmio_read 111 drivers/infiniband/hw/efa/efa_com.h struct efa_com_mmio_read mmio_read; mmio_read 179 drivers/infiniband/hw/efa/efa_main.c edev->mmio_read.mmio_read_timeout = mmio_read 804 drivers/net/ethernet/amazon/ena/ena_com.c struct ena_com_mmio_read *mmio_read = &ena_dev->mmio_read; mmio_read 806 drivers/net/ethernet/amazon/ena/ena_com.c mmio_read->read_resp; mmio_read 809 drivers/net/ethernet/amazon/ena/ena_com.c u32 timeout = mmio_read->reg_read_to; mmio_read 817 drivers/net/ethernet/amazon/ena/ena_com.c if (!mmio_read->readless_supported) mmio_read 820 drivers/net/ethernet/amazon/ena/ena_com.c spin_lock_irqsave(&mmio_read->lock, flags); mmio_read 821 drivers/net/ethernet/amazon/ena/ena_com.c mmio_read->seq_num++; mmio_read 823 drivers/net/ethernet/amazon/ena/ena_com.c read_resp->req_id = mmio_read->seq_num + 0xDEAD; mmio_read 826 drivers/net/ethernet/amazon/ena/ena_com.c mmio_read_reg |= mmio_read->seq_num & mmio_read 832 drivers/net/ethernet/amazon/ena/ena_com.c if (READ_ONCE(read_resp->req_id) == mmio_read->seq_num) mmio_read 840 drivers/net/ethernet/amazon/ena/ena_com.c mmio_read->seq_num, offset, read_resp->req_id, mmio_read 853 drivers/net/ethernet/amazon/ena/ena_com.c spin_unlock_irqrestore(&mmio_read->lock, flags); mmio_read 1680 drivers/net/ethernet/amazon/ena/ena_com.c struct ena_com_mmio_read *mmio_read = &ena_dev->mmio_read; mmio_read 1682 drivers/net/ethernet/amazon/ena/ena_com.c spin_lock_init(&mmio_read->lock); mmio_read 1683 drivers/net/ethernet/amazon/ena/ena_com.c mmio_read->read_resp = mmio_read 1685 drivers/net/ethernet/amazon/ena/ena_com.c sizeof(*mmio_read->read_resp), mmio_read 1686 drivers/net/ethernet/amazon/ena/ena_com.c &mmio_read->read_resp_dma_addr, GFP_KERNEL); mmio_read 1687 drivers/net/ethernet/amazon/ena/ena_com.c if (unlikely(!mmio_read->read_resp)) mmio_read 1692 drivers/net/ethernet/amazon/ena/ena_com.c mmio_read->read_resp->req_id = 0x0; mmio_read 1693 drivers/net/ethernet/amazon/ena/ena_com.c mmio_read->seq_num = 0x0; mmio_read 1694 drivers/net/ethernet/amazon/ena/ena_com.c mmio_read->readless_supported = true; mmio_read 1705 drivers/net/ethernet/amazon/ena/ena_com.c struct ena_com_mmio_read *mmio_read = &ena_dev->mmio_read; mmio_read 1707 drivers/net/ethernet/amazon/ena/ena_com.c mmio_read->readless_supported = readless_supported; mmio_read 1712 drivers/net/ethernet/amazon/ena/ena_com.c struct ena_com_mmio_read *mmio_read = &ena_dev->mmio_read; mmio_read 1717 drivers/net/ethernet/amazon/ena/ena_com.c dma_free_coherent(ena_dev->dmadev, sizeof(*mmio_read->read_resp), mmio_read 1718 drivers/net/ethernet/amazon/ena/ena_com.c mmio_read->read_resp, mmio_read->read_resp_dma_addr); mmio_read 1720 drivers/net/ethernet/amazon/ena/ena_com.c mmio_read->read_resp = NULL; mmio_read 1725 drivers/net/ethernet/amazon/ena/ena_com.c struct ena_com_mmio_read *mmio_read = &ena_dev->mmio_read; mmio_read 1728 drivers/net/ethernet/amazon/ena/ena_com.c addr_low = ENA_DMA_ADDR_TO_UINT32_LOW(mmio_read->read_resp_dma_addr); mmio_read 1729 drivers/net/ethernet/amazon/ena/ena_com.c addr_high = ENA_DMA_ADDR_TO_UINT32_HIGH(mmio_read->read_resp_dma_addr); mmio_read 333 drivers/net/ethernet/amazon/ena/ena_com.h struct ena_com_mmio_read mmio_read; mmio_read 3075 drivers/net/ethernet/amazon/ena/ena_netdev.c adapter->ena_dev->mmio_read.reg_read_to = mmio_read 252 drivers/video/fbdev/broadsheetfb.c if (par->board->mmio_read) mmio_read 253 drivers/video/fbdev/broadsheetfb.c return par->board->mmio_read(par); mmio_read 72 include/video/broadsheetfb.h u16 (*mmio_read)(struct broadsheetfb_par *); mmio_read 284 kernel/trace/trace_mmiotrace.c .read = mmio_read,