mmio_read         290 drivers/gpu/drm/i915/intel_uncore.h 	return uncore->funcs.mmio_read##s__(uncore, reg, (trace__)); \
mmio_read          84 drivers/infiniband/hw/efa/efa_com.c 	struct efa_com_mmio_read *mmio_read = &edev->mmio_read;
mmio_read          90 drivers/infiniband/hw/efa/efa_com.c 	read_resp = mmio_read->read_resp;
mmio_read          92 drivers/infiniband/hw/efa/efa_com.c 	spin_lock(&mmio_read->lock);
mmio_read          93 drivers/infiniband/hw/efa/efa_com.c 	mmio_read->seq_num++;
mmio_read          96 drivers/infiniband/hw/efa/efa_com.c 	read_resp->req_id = mmio_read->seq_num + 0x9aL;
mmio_read          99 drivers/infiniband/hw/efa/efa_com.c 	mmio_read_reg |= mmio_read->seq_num &
mmio_read         104 drivers/infiniband/hw/efa/efa_com.c 	exp_time = jiffies + usecs_to_jiffies(mmio_read->mmio_read_timeout);
mmio_read         106 drivers/infiniband/hw/efa/efa_com.c 		if (READ_ONCE(read_resp->req_id) == mmio_read->seq_num)
mmio_read         111 drivers/infiniband/hw/efa/efa_com.c 	if (read_resp->req_id != mmio_read->seq_num) {
mmio_read         115 drivers/infiniband/hw/efa/efa_com.c 			mmio_read->seq_num, offset, read_resp->req_id,
mmio_read         131 drivers/infiniband/hw/efa/efa_com.c 	spin_unlock(&mmio_read->lock);
mmio_read         901 drivers/infiniband/hw/efa/efa_com.c 	struct efa_com_mmio_read *mmio_read = &edev->mmio_read;
mmio_read         906 drivers/infiniband/hw/efa/efa_com.c 	addr_high = (mmio_read->read_resp_dma_addr >> 32) & GENMASK(31, 0);
mmio_read         907 drivers/infiniband/hw/efa/efa_com.c 	addr_low = mmio_read->read_resp_dma_addr & GENMASK(31, 0);
mmio_read         915 drivers/infiniband/hw/efa/efa_com.c 	struct efa_com_mmio_read *mmio_read = &edev->mmio_read;
mmio_read         917 drivers/infiniband/hw/efa/efa_com.c 	spin_lock_init(&mmio_read->lock);
mmio_read         918 drivers/infiniband/hw/efa/efa_com.c 	mmio_read->read_resp =
mmio_read         919 drivers/infiniband/hw/efa/efa_com.c 		dma_alloc_coherent(edev->dmadev, sizeof(*mmio_read->read_resp),
mmio_read         920 drivers/infiniband/hw/efa/efa_com.c 				   &mmio_read->read_resp_dma_addr, GFP_KERNEL);
mmio_read         921 drivers/infiniband/hw/efa/efa_com.c 	if (!mmio_read->read_resp)
mmio_read         926 drivers/infiniband/hw/efa/efa_com.c 	mmio_read->read_resp->req_id = 0;
mmio_read         927 drivers/infiniband/hw/efa/efa_com.c 	mmio_read->seq_num = 0;
mmio_read         928 drivers/infiniband/hw/efa/efa_com.c 	mmio_read->mmio_read_timeout = EFA_REG_READ_TIMEOUT_US;
mmio_read         935 drivers/infiniband/hw/efa/efa_com.c 	struct efa_com_mmio_read *mmio_read = &edev->mmio_read;
mmio_read         937 drivers/infiniband/hw/efa/efa_com.c 	dma_free_coherent(edev->dmadev, sizeof(*mmio_read->read_resp),
mmio_read         938 drivers/infiniband/hw/efa/efa_com.c 			  mmio_read->read_resp, mmio_read->read_resp_dma_addr);
mmio_read         111 drivers/infiniband/hw/efa/efa_com.h 	struct efa_com_mmio_read mmio_read;
mmio_read         179 drivers/infiniband/hw/efa/efa_main.c 		edev->mmio_read.mmio_read_timeout =
mmio_read         804 drivers/net/ethernet/amazon/ena/ena_com.c 	struct ena_com_mmio_read *mmio_read = &ena_dev->mmio_read;
mmio_read         806 drivers/net/ethernet/amazon/ena/ena_com.c 		mmio_read->read_resp;
mmio_read         809 drivers/net/ethernet/amazon/ena/ena_com.c 	u32 timeout = mmio_read->reg_read_to;
mmio_read         817 drivers/net/ethernet/amazon/ena/ena_com.c 	if (!mmio_read->readless_supported)
mmio_read         820 drivers/net/ethernet/amazon/ena/ena_com.c 	spin_lock_irqsave(&mmio_read->lock, flags);
mmio_read         821 drivers/net/ethernet/amazon/ena/ena_com.c 	mmio_read->seq_num++;
mmio_read         823 drivers/net/ethernet/amazon/ena/ena_com.c 	read_resp->req_id = mmio_read->seq_num + 0xDEAD;
mmio_read         826 drivers/net/ethernet/amazon/ena/ena_com.c 	mmio_read_reg |= mmio_read->seq_num &
mmio_read         832 drivers/net/ethernet/amazon/ena/ena_com.c 		if (READ_ONCE(read_resp->req_id) == mmio_read->seq_num)
mmio_read         840 drivers/net/ethernet/amazon/ena/ena_com.c 		       mmio_read->seq_num, offset, read_resp->req_id,
mmio_read         853 drivers/net/ethernet/amazon/ena/ena_com.c 	spin_unlock_irqrestore(&mmio_read->lock, flags);
mmio_read        1680 drivers/net/ethernet/amazon/ena/ena_com.c 	struct ena_com_mmio_read *mmio_read = &ena_dev->mmio_read;
mmio_read        1682 drivers/net/ethernet/amazon/ena/ena_com.c 	spin_lock_init(&mmio_read->lock);
mmio_read        1683 drivers/net/ethernet/amazon/ena/ena_com.c 	mmio_read->read_resp =
mmio_read        1685 drivers/net/ethernet/amazon/ena/ena_com.c 				   sizeof(*mmio_read->read_resp),
mmio_read        1686 drivers/net/ethernet/amazon/ena/ena_com.c 				   &mmio_read->read_resp_dma_addr, GFP_KERNEL);
mmio_read        1687 drivers/net/ethernet/amazon/ena/ena_com.c 	if (unlikely(!mmio_read->read_resp))
mmio_read        1692 drivers/net/ethernet/amazon/ena/ena_com.c 	mmio_read->read_resp->req_id = 0x0;
mmio_read        1693 drivers/net/ethernet/amazon/ena/ena_com.c 	mmio_read->seq_num = 0x0;
mmio_read        1694 drivers/net/ethernet/amazon/ena/ena_com.c 	mmio_read->readless_supported = true;
mmio_read        1705 drivers/net/ethernet/amazon/ena/ena_com.c 	struct ena_com_mmio_read *mmio_read = &ena_dev->mmio_read;
mmio_read        1707 drivers/net/ethernet/amazon/ena/ena_com.c 	mmio_read->readless_supported = readless_supported;
mmio_read        1712 drivers/net/ethernet/amazon/ena/ena_com.c 	struct ena_com_mmio_read *mmio_read = &ena_dev->mmio_read;
mmio_read        1717 drivers/net/ethernet/amazon/ena/ena_com.c 	dma_free_coherent(ena_dev->dmadev, sizeof(*mmio_read->read_resp),
mmio_read        1718 drivers/net/ethernet/amazon/ena/ena_com.c 			  mmio_read->read_resp, mmio_read->read_resp_dma_addr);
mmio_read        1720 drivers/net/ethernet/amazon/ena/ena_com.c 	mmio_read->read_resp = NULL;
mmio_read        1725 drivers/net/ethernet/amazon/ena/ena_com.c 	struct ena_com_mmio_read *mmio_read = &ena_dev->mmio_read;
mmio_read        1728 drivers/net/ethernet/amazon/ena/ena_com.c 	addr_low = ENA_DMA_ADDR_TO_UINT32_LOW(mmio_read->read_resp_dma_addr);
mmio_read        1729 drivers/net/ethernet/amazon/ena/ena_com.c 	addr_high = ENA_DMA_ADDR_TO_UINT32_HIGH(mmio_read->read_resp_dma_addr);
mmio_read         333 drivers/net/ethernet/amazon/ena/ena_com.h 	struct ena_com_mmio_read mmio_read;
mmio_read        3075 drivers/net/ethernet/amazon/ena/ena_netdev.c 		adapter->ena_dev->mmio_read.reg_read_to =
mmio_read         252 drivers/video/fbdev/broadsheetfb.c 	if (par->board->mmio_read)
mmio_read         253 drivers/video/fbdev/broadsheetfb.c 		return par->board->mmio_read(par);
mmio_read          72 include/video/broadsheetfb.h 	u16 (*mmio_read)(struct broadsheetfb_par *);
mmio_read         284 kernel/trace/trace_mmiotrace.c 	.read		= mmio_read,