uar               150 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h 	u16 uar;
uar               460 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h 	u16 uar;
uar               729 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h 	u16 uar;
uar               739 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h 	u16 uar;
uar                39 drivers/infiniband/hw/efa/efa_com_cmd.c 	create_qp_cmd.uar = params->uarn;
uar               168 drivers/infiniband/hw/efa/efa_com_cmd.c 	create_cmd.uar = params->uarn;
uar               700 drivers/infiniband/hw/efa/efa_com_cmd.c 	result->uarn = resp.uar;
uar               714 drivers/infiniband/hw/efa/efa_com_cmd.c 	cmd.uar = params->uarn;
uar               724 drivers/infiniband/hw/efa/efa_com_cmd.c 				      cmd.uar, err);
uar               350 drivers/infiniband/hw/hns/hns_roce_cq.c 	struct hns_roce_uar *uar;
uar               370 drivers/infiniband/hw/hns/hns_roce_cq.c 	uar = &hr_dev->priv_uar;
uar               372 drivers/infiniband/hw/hns/hns_roce_cq.c 			 DB_REG_OFFSET * uar->index;
uar               284 drivers/infiniband/hw/hns/hns_roce_device.h 	struct hns_roce_uar	uar;
uar               500 drivers/infiniband/hw/hns/hns_roce_device.h 	struct hns_roce_uar		*uar;
uar              1122 drivers/infiniband/hw/hns/hns_roce_device.h int hns_roce_uar_alloc(struct hns_roce_dev *dev, struct hns_roce_uar *uar);
uar              1123 drivers/infiniband/hw/hns/hns_roce_device.h void hns_roce_uar_free(struct hns_roce_dev *dev, struct hns_roce_uar *uar);
uar               323 drivers/infiniband/hw/hns/hns_roce_main.c 	ret = hns_roce_uar_alloc(hr_dev, &context->uar);
uar               339 drivers/infiniband/hw/hns/hns_roce_main.c 	hns_roce_uar_free(hr_dev, &context->uar);
uar               349 drivers/infiniband/hw/hns/hns_roce_main.c 	hns_roce_uar_free(to_hr_dev(ibcontext->device), &context->uar);
uar               360 drivers/infiniband/hw/hns/hns_roce_main.c 					 to_hr_ucontext(context)->uar.pfn,
uar                92 drivers/infiniband/hw/hns/hns_roce_pd.c int hns_roce_uar_alloc(struct hns_roce_dev *hr_dev, struct hns_roce_uar *uar)
uar                98 drivers/infiniband/hw/hns/hns_roce_pd.c 	ret = hns_roce_bitmap_alloc(&hr_dev->uar_table.bitmap, &uar->logic_idx);
uar               102 drivers/infiniband/hw/hns/hns_roce_pd.c 	if (uar->logic_idx > 0 && hr_dev->caps.phy_num_uars > 1)
uar               103 drivers/infiniband/hw/hns/hns_roce_pd.c 		uar->index = (uar->logic_idx - 1) %
uar               106 drivers/infiniband/hw/hns/hns_roce_pd.c 		uar->index = 0;
uar               114 drivers/infiniband/hw/hns/hns_roce_pd.c 		uar->pfn = ((res->start) >> PAGE_SHIFT) + uar->index;
uar               116 drivers/infiniband/hw/hns/hns_roce_pd.c 		uar->pfn = ((pci_resource_start(hr_dev->pci_dev, 2))
uar               123 drivers/infiniband/hw/hns/hns_roce_pd.c void hns_roce_uar_free(struct hns_roce_dev *hr_dev, struct hns_roce_uar *uar)
uar               125 drivers/infiniband/hw/hns/hns_roce_pd.c 	hns_roce_bitmap_free(&hr_dev->uar_table.bitmap, uar->logic_idx,
uar               183 drivers/infiniband/hw/mlx4/cq.c 	struct mlx4_uar *uar;
uar               223 drivers/infiniband/hw/mlx4/cq.c 		uar = &context->uar;
uar               241 drivers/infiniband/hw/mlx4/cq.c 		uar = &dev->priv_uar;
uar               248 drivers/infiniband/hw/mlx4/cq.c 	err = mlx4_cq_alloc(dev->dev, entries, &cq->buf.mtt, uar, cq->db.dma,
uar              1118 drivers/infiniband/hw/mlx4/main.c 	err = mlx4_uar_alloc(to_mdev(ibdev)->dev, &context->uar);
uar              1134 drivers/infiniband/hw/mlx4/main.c 		mlx4_uar_free(to_mdev(ibdev)->dev, &context->uar);
uar              1145 drivers/infiniband/hw/mlx4/main.c 	mlx4_uar_free(to_mdev(ibcontext->device)->dev, &context->uar);
uar              1159 drivers/infiniband/hw/mlx4/main.c 					 to_mucontext(context)->uar.pfn,
uar              1168 drivers/infiniband/hw/mlx4/main.c 			to_mucontext(context)->uar.pfn +
uar                85 drivers/infiniband/hw/mlx4/mlx4_ib.h 	struct mlx4_uar		uar;
uar              2254 drivers/infiniband/hw/mlx4/qp.c 			mlx4_to_hw_uar_index(dev->dev, ucontext->uar.index));
uar               637 drivers/infiniband/hw/mlx5/cq.c 	void __iomem *uar_page = mdev->priv.uar->map;
uar               884 drivers/infiniband/hw/mlx5/cq.c 	*index = dev->mdev->priv.uar->index;
uar              6608 drivers/infiniband/hw/mlx5/main.c 	dev->mdev->priv.uar = mlx5_get_uars_page(dev->mdev);
uar              6609 drivers/infiniband/hw/mlx5/main.c 	return PTR_ERR_OR_ZERO(dev->mdev->priv.uar);
uar              6614 drivers/infiniband/hw/mlx5/main.c 	mlx5_put_uars_page(dev->mdev, dev->mdev->priv.uar);
uar               828 drivers/infiniband/hw/mthca/mthca_cq.c 		cq_context->logsize_usrpage |= cpu_to_be32(ctx->uar.index);
uar               459 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_uar_alloc(struct mthca_dev *dev, struct mthca_uar *uar);
uar               460 drivers/infiniband/hw/mthca/mthca_dev.h void mthca_uar_free(struct mthca_dev *dev, struct mthca_uar *uar);
uar               439 drivers/infiniband/hw/mthca/mthca_memfree.c static u64 mthca_uarc_virt(struct mthca_dev *dev, struct mthca_uar *uar, int page)
uar               442 drivers/infiniband/hw/mthca/mthca_memfree.c 		uar->index * dev->uar_table.uarc_size +
uar               446 drivers/infiniband/hw/mthca/mthca_memfree.c int mthca_map_user_db(struct mthca_dev *dev, struct mthca_uar *uar,
uar               490 drivers/infiniband/hw/mthca/mthca_memfree.c 				 mthca_uarc_virt(dev, uar, i));
uar               505 drivers/infiniband/hw/mthca/mthca_memfree.c void mthca_unmap_user_db(struct mthca_dev *dev, struct mthca_uar *uar,
uar               547 drivers/infiniband/hw/mthca/mthca_memfree.c void mthca_cleanup_user_db_tab(struct mthca_dev *dev, struct mthca_uar *uar,
uar               557 drivers/infiniband/hw/mthca/mthca_memfree.c 			mthca_UNMAP_ICM(dev, mthca_uarc_virt(dev, uar, i), 1);
uar               165 drivers/infiniband/hw/mthca/mthca_memfree.h int mthca_map_user_db(struct mthca_dev *dev, struct mthca_uar *uar,
uar               167 drivers/infiniband/hw/mthca/mthca_memfree.h void mthca_unmap_user_db(struct mthca_dev *dev, struct mthca_uar *uar,
uar               170 drivers/infiniband/hw/mthca/mthca_memfree.h void mthca_cleanup_user_db_tab(struct mthca_dev *dev, struct mthca_uar *uar,
uar               321 drivers/infiniband/hw/mthca/mthca_provider.c 	err = mthca_uar_alloc(to_mdev(ibdev), &context->uar);
uar               328 drivers/infiniband/hw/mthca/mthca_provider.c 		mthca_uar_free(to_mdev(ibdev), &context->uar);
uar               333 drivers/infiniband/hw/mthca/mthca_provider.c 		mthca_cleanup_user_db_tab(to_mdev(ibdev), &context->uar, context->db_tab);
uar               334 drivers/infiniband/hw/mthca/mthca_provider.c 		mthca_uar_free(to_mdev(ibdev), &context->uar);
uar               345 drivers/infiniband/hw/mthca/mthca_provider.c 	mthca_cleanup_user_db_tab(to_mdev(context->device), &to_mucontext(context)->uar,
uar               347 drivers/infiniband/hw/mthca/mthca_provider.c 	mthca_uar_free(to_mdev(context->device), &to_mucontext(context)->uar);
uar               359 drivers/infiniband/hw/mthca/mthca_provider.c 			       to_mucontext(context)->uar.pfn,
uar               423 drivers/infiniband/hw/mthca/mthca_provider.c 		err = mthca_map_user_db(to_mdev(ibsrq->device), &context->uar,
uar               438 drivers/infiniband/hw/mthca/mthca_provider.c 		mthca_unmap_user_db(to_mdev(ibsrq->device), &context->uar,
uar               461 drivers/infiniband/hw/mthca/mthca_provider.c 		mthca_unmap_user_db(to_mdev(srq->device), &context->uar,
uar               496 drivers/infiniband/hw/mthca/mthca_provider.c 			err = mthca_map_user_db(to_mdev(pd->device), &context->uar,
uar               504 drivers/infiniband/hw/mthca/mthca_provider.c 			err = mthca_map_user_db(to_mdev(pd->device), &context->uar,
uar               509 drivers/infiniband/hw/mthca/mthca_provider.c 						    &context->uar,
uar               529 drivers/infiniband/hw/mthca/mthca_provider.c 					    &context->uar,
uar               533 drivers/infiniband/hw/mthca/mthca_provider.c 					    &context->uar,
uar               591 drivers/infiniband/hw/mthca/mthca_provider.c 				    &context->uar,
uar               595 drivers/infiniband/hw/mthca/mthca_provider.c 				    &context->uar,
uar               627 drivers/infiniband/hw/mthca/mthca_provider.c 		err = mthca_map_user_db(to_mdev(ibdev), &context->uar,
uar               633 drivers/infiniband/hw/mthca/mthca_provider.c 		err = mthca_map_user_db(to_mdev(ibdev), &context->uar,
uar               669 drivers/infiniband/hw/mthca/mthca_provider.c 		mthca_unmap_user_db(to_mdev(ibdev), &context->uar,
uar               674 drivers/infiniband/hw/mthca/mthca_provider.c 		mthca_unmap_user_db(to_mdev(ibdev), &context->uar,
uar               811 drivers/infiniband/hw/mthca/mthca_provider.c 				    &context->uar,
uar               815 drivers/infiniband/hw/mthca/mthca_provider.c 				    &context->uar,
uar                66 drivers/infiniband/hw/mthca/mthca_provider.h 	struct mthca_uar            uar;
uar               627 drivers/infiniband/hw/mthca/mthca_qp.c 		qp_context->usr_page = cpu_to_be32(context->uar.index);
uar                54 drivers/infiniband/hw/mthca/mthca_srq.c 	__be32 uar;
uar               113 drivers/infiniband/hw/mthca/mthca_srq.c 		context->uar = cpu_to_be32(ucontext->uar.index);
uar               115 drivers/infiniband/hw/mthca/mthca_srq.c 		context->uar = cpu_to_be32(dev->driver_uar.index);
uar               141 drivers/infiniband/hw/mthca/mthca_srq.c 		context->logstride_usrpage |= cpu_to_be32(ucontext->uar.index);
uar                38 drivers/infiniband/hw/mthca/mthca_uar.c int mthca_uar_alloc(struct mthca_dev *dev, struct mthca_uar *uar)
uar                40 drivers/infiniband/hw/mthca/mthca_uar.c 	uar->index = mthca_alloc(&dev->uar_table.alloc);
uar                41 drivers/infiniband/hw/mthca/mthca_uar.c 	if (uar->index == -1)
uar                44 drivers/infiniband/hw/mthca/mthca_uar.c 	uar->pfn = (pci_resource_start(dev->pdev, 2) >> PAGE_SHIFT) + uar->index;
uar                49 drivers/infiniband/hw/mthca/mthca_uar.c void mthca_uar_free(struct mthca_dev *dev, struct mthca_uar *uar)
uar                51 drivers/infiniband/hw/mthca/mthca_uar.c 	mthca_free(&dev->uar_table.alloc, uar->index);
uar                90 drivers/infiniband/hw/vmw_pvrdma/pvrdma.h 	struct pvrdma_uar_map *uar;
uar               123 drivers/infiniband/hw/vmw_pvrdma/pvrdma.h 	struct pvrdma_uar_map uar;
uar               531 drivers/infiniband/hw/vmw_pvrdma/pvrdma.h int pvrdma_uar_alloc(struct pvrdma_dev *dev, struct pvrdma_uar_map *uar);
uar               532 drivers/infiniband/hw/vmw_pvrdma/pvrdma.h void pvrdma_uar_free(struct pvrdma_dev *dev, struct pvrdma_uar_map *uar);
uar               200 drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c 		cq->uar = &context->uar;
uar                83 drivers/infiniband/hw/vmw_pvrdma/pvrdma_doorbell.c int pvrdma_uar_alloc(struct pvrdma_dev *dev, struct pvrdma_uar_map *uar)
uar               108 drivers/infiniband/hw/vmw_pvrdma/pvrdma_doorbell.c 	uar->index = obj;
uar               109 drivers/infiniband/hw/vmw_pvrdma/pvrdma_doorbell.c 	uar->pfn = (pci_resource_start(dev->pdev, PVRDMA_PCI_RESOURCE_UAR) >>
uar               110 drivers/infiniband/hw/vmw_pvrdma/pvrdma_doorbell.c 		    PAGE_SHIFT) + uar->index;
uar               115 drivers/infiniband/hw/vmw_pvrdma/pvrdma_doorbell.c void pvrdma_uar_free(struct pvrdma_dev *dev, struct pvrdma_uar_map *uar)
uar               121 drivers/infiniband/hw/vmw_pvrdma/pvrdma_doorbell.c 	obj = uar->index & (tbl->max - 1);
uar               329 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c 	ret = pvrdma_uar_alloc(vdev, &context->uar);
uar               335 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c 		cmd->pfn = context->uar.pfn;
uar               337 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c 		cmd->pfn64 = context->uar.pfn;
uar               353 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c 		pvrdma_uar_free(vdev, &context->uar);
uar               361 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c 	pvrdma_uar_free(vdev, &context->uar);
uar               385 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c 	pvrdma_uar_free(to_vdev(ibcontext->device), &context->uar);
uar               413 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c 	if (io_remap_pfn_range(vma, start, context->uar.pfn, size,
uar               342 drivers/net/ethernet/mellanox/mlx4/cq.c 		  struct mlx4_mtt *mtt, struct mlx4_uar *uar, u64 db_rec,
uar               382 drivers/net/ethernet/mellanox/mlx4/cq.c 			    mlx4_to_hw_uar_index(dev, uar->index));
uar               411 drivers/net/ethernet/mellanox/mlx4/cq.c 	cq->uar        = uar;
uar               123 drivers/net/ethernet/mellanox/mlx4/en_tx.c 		ring->bf.uar = &mdev->priv_uar;
uar               124 drivers/net/ethernet/mellanox/mlx4/en_tx.c 		ring->bf.uar->map = mdev->uar_map;
uar               206 drivers/net/ethernet/mellanox/mlx4/en_tx.c 							 ring->bf.uar->index));
uar               731 drivers/net/ethernet/mellanox/mlx4/en_tx.c 		  ring->bf.uar->map + MLX4_SEND_DOORBELL);
uar               145 drivers/net/ethernet/mellanox/mlx4/pd.c int mlx4_uar_alloc(struct mlx4_dev *dev, struct mlx4_uar *uar)
uar               149 drivers/net/ethernet/mellanox/mlx4/pd.c 	uar->index = mlx4_bitmap_alloc(&mlx4_priv(dev)->uar_table.bitmap);
uar               150 drivers/net/ethernet/mellanox/mlx4/pd.c 	if (uar->index == -1)
uar               154 drivers/net/ethernet/mellanox/mlx4/pd.c 		offset = uar->index % ((int)pci_resource_len(dev->persist->pdev,
uar               158 drivers/net/ethernet/mellanox/mlx4/pd.c 		offset = uar->index;
uar               159 drivers/net/ethernet/mellanox/mlx4/pd.c 	uar->pfn = (pci_resource_start(dev->persist->pdev, 2) >> PAGE_SHIFT)
uar               161 drivers/net/ethernet/mellanox/mlx4/pd.c 	uar->map = NULL;
uar               166 drivers/net/ethernet/mellanox/mlx4/pd.c void mlx4_uar_free(struct mlx4_dev *dev, struct mlx4_uar *uar)
uar               168 drivers/net/ethernet/mellanox/mlx4/pd.c 	mlx4_bitmap_free(&mlx4_priv(dev)->uar_table.bitmap, uar->index, MLX4_USE_RR);
uar               175 drivers/net/ethernet/mellanox/mlx4/pd.c 	struct mlx4_uar *uar;
uar               184 drivers/net/ethernet/mellanox/mlx4/pd.c 		uar = list_entry(priv->bf_list.next, struct mlx4_uar, bf_list);
uar               190 drivers/net/ethernet/mellanox/mlx4/pd.c 		uar = kmalloc_node(sizeof(*uar), GFP_KERNEL, node);
uar               191 drivers/net/ethernet/mellanox/mlx4/pd.c 		if (!uar) {
uar               192 drivers/net/ethernet/mellanox/mlx4/pd.c 			uar = kmalloc(sizeof(*uar), GFP_KERNEL);
uar               193 drivers/net/ethernet/mellanox/mlx4/pd.c 			if (!uar) {
uar               198 drivers/net/ethernet/mellanox/mlx4/pd.c 		err = mlx4_uar_alloc(dev, uar);
uar               202 drivers/net/ethernet/mellanox/mlx4/pd.c 		uar->map = ioremap(uar->pfn << PAGE_SHIFT, PAGE_SIZE);
uar               203 drivers/net/ethernet/mellanox/mlx4/pd.c 		if (!uar->map) {
uar               208 drivers/net/ethernet/mellanox/mlx4/pd.c 		uar->bf_map = io_mapping_map_wc(priv->bf_mapping,
uar               209 drivers/net/ethernet/mellanox/mlx4/pd.c 						uar->index << PAGE_SHIFT,
uar               211 drivers/net/ethernet/mellanox/mlx4/pd.c 		if (!uar->bf_map) {
uar               215 drivers/net/ethernet/mellanox/mlx4/pd.c 		uar->free_bf_bmap = 0;
uar               216 drivers/net/ethernet/mellanox/mlx4/pd.c 		list_add(&uar->bf_list, &priv->bf_list);
uar               219 drivers/net/ethernet/mellanox/mlx4/pd.c 	idx = ffz(uar->free_bf_bmap);
uar               220 drivers/net/ethernet/mellanox/mlx4/pd.c 	uar->free_bf_bmap |= 1 << idx;
uar               221 drivers/net/ethernet/mellanox/mlx4/pd.c 	bf->uar = uar;
uar               224 drivers/net/ethernet/mellanox/mlx4/pd.c 	bf->reg = uar->bf_map + idx * dev->caps.bf_reg_size;
uar               225 drivers/net/ethernet/mellanox/mlx4/pd.c 	if (uar->free_bf_bmap == (1 << dev->caps.bf_regs_per_page) - 1)
uar               226 drivers/net/ethernet/mellanox/mlx4/pd.c 		list_del_init(&uar->bf_list);
uar               231 drivers/net/ethernet/mellanox/mlx4/pd.c 	bf->uar = NULL;
uar               232 drivers/net/ethernet/mellanox/mlx4/pd.c 	iounmap(uar->map);
uar               235 drivers/net/ethernet/mellanox/mlx4/pd.c 	mlx4_uar_free(dev, uar);
uar               238 drivers/net/ethernet/mellanox/mlx4/pd.c 	kfree(uar);
uar               251 drivers/net/ethernet/mellanox/mlx4/pd.c 	if (!bf->uar || !bf->uar->bf_map)
uar               255 drivers/net/ethernet/mellanox/mlx4/pd.c 	idx = (bf->reg - bf->uar->bf_map) / dev->caps.bf_reg_size;
uar               256 drivers/net/ethernet/mellanox/mlx4/pd.c 	bf->uar->free_bf_bmap &= ~(1 << idx);
uar               257 drivers/net/ethernet/mellanox/mlx4/pd.c 	if (!bf->uar->free_bf_bmap) {
uar               258 drivers/net/ethernet/mellanox/mlx4/pd.c 		if (!list_empty(&bf->uar->bf_list))
uar               259 drivers/net/ethernet/mellanox/mlx4/pd.c 			list_del(&bf->uar->bf_list);
uar               261 drivers/net/ethernet/mellanox/mlx4/pd.c 		io_mapping_unmap(bf->uar->bf_map);
uar               262 drivers/net/ethernet/mellanox/mlx4/pd.c 		iounmap(bf->uar->map);
uar               263 drivers/net/ethernet/mellanox/mlx4/pd.c 		mlx4_uar_free(dev, bf->uar);
uar               264 drivers/net/ethernet/mellanox/mlx4/pd.c 		kfree(bf->uar);
uar               265 drivers/net/ethernet/mellanox/mlx4/pd.c 	} else if (list_empty(&bf->uar->bf_list))
uar               266 drivers/net/ethernet/mellanox/mlx4/pd.c 		list_add(&bf->uar->bf_list, &priv->bf_list);
uar               138 drivers/net/ethernet/mellanox/mlx5/core/cq.c 	cq->uar = dev->priv.uar;
uar               147 drivers/net/ethernet/mellanox/mlx5/core/en/txrx.h 	mlx5_cq_arm(mcq, MLX5_CQ_DB_REQ_NOT, mcq->uar->map, cq->wq.cc);
uar              1629 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 	MLX5_SET(cqc,   cqc, uar_page,      mdev->priv.uar->index);
uar              2233 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 	MLX5_SET(cqc, cqc, uar_page, priv->mdev->priv.uar->index);
uar               292 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	MLX5_SET(eqc, eqc, uar_page, priv->uar->index);
uar               305 drivers/net/ethernet/mellanox/mlx5/core/eq.c 	eq->doorbell = priv->uar->map + MLX5_EQ_DOORBEL_OFFSET;
uar               138 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c 	mlx5_write64(wqe, conn->fdev->conn_res.uar->map + MLX5_BF_OFFSET);
uar               362 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c 		    conn->fdev->conn_res.uar->map, conn->cq.wq.cc);
uar               475 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c 	MLX5_SET(cqc, cqc, uar_page, fdev->conn_res.uar->index);
uar               498 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c 	conn->cq.mcq.uar        = fdev->conn_res.uar;
uar               582 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c 	MLX5_SET(qpc, qpc, uar_page, fdev->conn_res.uar->index);
uar              1007 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c 	fdev->conn_res.uar = mlx5_get_uars_page(fdev->mdev);
uar              1008 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c 	if (IS_ERR(fdev->conn_res.uar)) {
uar              1009 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c 		err = PTR_ERR(fdev->conn_res.uar);
uar              1014 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c 		      fdev->conn_res.uar->index);
uar              1036 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c 	mlx5_put_uars_page(fdev->mdev, fdev->conn_res.uar);
uar              1047 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c 	mlx5_put_uars_page(fdev->mdev, fdev->conn_res.uar);
uar                58 drivers/net/ethernet/mellanox/mlx5/core/fpga/core.h 		struct mlx5_uars_page *uar;
uar              1061 drivers/net/ethernet/mellanox/mlx5/core/main.c 	dev->priv.uar = mlx5_get_uars_page(dev);
uar              1062 drivers/net/ethernet/mellanox/mlx5/core/main.c 	if (IS_ERR(dev->priv.uar)) {
uar              1064 drivers/net/ethernet/mellanox/mlx5/core/main.c 		err = PTR_ERR(dev->priv.uar);
uar              1155 drivers/net/ethernet/mellanox/mlx5/core/main.c 	mlx5_put_uars_page(dev, dev->priv.uar);
uar              1173 drivers/net/ethernet/mellanox/mlx5/core/main.c 	mlx5_put_uars_page(dev, dev->priv.uar);
uar                66 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c 	dmn->uar = mlx5_get_uars_page(dmn->mdev);
uar                67 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c 	if (!dmn->uar) {
uar               100 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c 	mlx5_put_uars_page(dmn->mdev, dmn->uar);
uar               112 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c 	mlx5_put_uars_page(dmn->mdev, dmn->uar);
uar                47 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c 	struct mlx5_uars_page *uar;
uar               165 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c 	MLX5_SET(qpc, qpc, uar_page, attr->uar->index);
uar               191 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c 	dr_qp->uar = attr->uar;
uar               221 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c 	mlx5_write64(ctrl, dr_qp->uar->map + MLX5_BF_OFFSET);
uar               699 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c 				      struct mlx5_uars_page *uar,
uar               750 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c 	MLX5_SET(cqc, cqc, uar_page, uar->index);
uar               779 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c 	cq->mcq.uar = uar;
uar               875 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c 	dmn->send_ring->cq = dr_create_cq(dmn->mdev, dmn->uar, cq_size);
uar               883 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c 	init_attr.uar = dmn->uar;
uar               651 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_types.h 	struct mlx5_uars_page *uar;
uar               973 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_types.h 	struct mlx5_uars_page *uar;
uar                49 drivers/net/ethernet/mellanox/mlx5/core/uar.c 		*uarn = MLX5_GET(alloc_uar_out, out, uar);
uar                60 drivers/net/ethernet/mellanox/mlx5/core/uar.c 	MLX5_SET(dealloc_uar_in, in, uar, uarn);
uar               733 include/linux/mlx4/device.h 	struct mlx4_uar	       *uar;
uar               741 include/linux/mlx4/device.h 	struct mlx4_uar	       *uar;
uar              1107 include/linux/mlx4/device.h int mlx4_uar_alloc(struct mlx4_dev *dev, struct mlx4_uar *uar);
uar              1108 include/linux/mlx4/device.h void mlx4_uar_free(struct mlx4_dev *dev, struct mlx4_uar *uar);
uar              1139 include/linux/mlx4/device.h 		  struct mlx4_uar *uar, u64 db_rec, struct mlx4_cq *cq,
uar                45 include/linux/mlx5/cq.h 	struct mlx5_uars_page  *uar;
uar               591 include/linux/mlx5/driver.h 	struct mlx5_uars_page	       *uar;
uar              7109 include/linux/mlx5/mlx5_ifc.h 	u8         uar[0x18];
uar              7941 include/linux/mlx5/mlx5_ifc.h 	u8         uar[0x18];