uar 150 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h u16 uar; uar 460 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h u16 uar; uar 729 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h u16 uar; uar 739 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h u16 uar; uar 39 drivers/infiniband/hw/efa/efa_com_cmd.c create_qp_cmd.uar = params->uarn; uar 168 drivers/infiniband/hw/efa/efa_com_cmd.c create_cmd.uar = params->uarn; uar 700 drivers/infiniband/hw/efa/efa_com_cmd.c result->uarn = resp.uar; uar 714 drivers/infiniband/hw/efa/efa_com_cmd.c cmd.uar = params->uarn; uar 724 drivers/infiniband/hw/efa/efa_com_cmd.c cmd.uar, err); uar 350 drivers/infiniband/hw/hns/hns_roce_cq.c struct hns_roce_uar *uar; uar 370 drivers/infiniband/hw/hns/hns_roce_cq.c uar = &hr_dev->priv_uar; uar 372 drivers/infiniband/hw/hns/hns_roce_cq.c DB_REG_OFFSET * uar->index; uar 284 drivers/infiniband/hw/hns/hns_roce_device.h struct hns_roce_uar uar; uar 500 drivers/infiniband/hw/hns/hns_roce_device.h struct hns_roce_uar *uar; uar 1122 drivers/infiniband/hw/hns/hns_roce_device.h int hns_roce_uar_alloc(struct hns_roce_dev *dev, struct hns_roce_uar *uar); uar 1123 drivers/infiniband/hw/hns/hns_roce_device.h void hns_roce_uar_free(struct hns_roce_dev *dev, struct hns_roce_uar *uar); uar 323 drivers/infiniband/hw/hns/hns_roce_main.c ret = hns_roce_uar_alloc(hr_dev, &context->uar); uar 339 drivers/infiniband/hw/hns/hns_roce_main.c hns_roce_uar_free(hr_dev, &context->uar); uar 349 drivers/infiniband/hw/hns/hns_roce_main.c hns_roce_uar_free(to_hr_dev(ibcontext->device), &context->uar); uar 360 drivers/infiniband/hw/hns/hns_roce_main.c to_hr_ucontext(context)->uar.pfn, uar 92 drivers/infiniband/hw/hns/hns_roce_pd.c int hns_roce_uar_alloc(struct hns_roce_dev *hr_dev, struct hns_roce_uar *uar) uar 98 drivers/infiniband/hw/hns/hns_roce_pd.c ret = hns_roce_bitmap_alloc(&hr_dev->uar_table.bitmap, &uar->logic_idx); uar 102 drivers/infiniband/hw/hns/hns_roce_pd.c if (uar->logic_idx > 0 && hr_dev->caps.phy_num_uars > 1) uar 103 drivers/infiniband/hw/hns/hns_roce_pd.c uar->index = (uar->logic_idx - 1) % uar 106 drivers/infiniband/hw/hns/hns_roce_pd.c uar->index = 0; uar 114 drivers/infiniband/hw/hns/hns_roce_pd.c uar->pfn = ((res->start) >> PAGE_SHIFT) + uar->index; uar 116 drivers/infiniband/hw/hns/hns_roce_pd.c uar->pfn = ((pci_resource_start(hr_dev->pci_dev, 2)) uar 123 drivers/infiniband/hw/hns/hns_roce_pd.c void hns_roce_uar_free(struct hns_roce_dev *hr_dev, struct hns_roce_uar *uar) uar 125 drivers/infiniband/hw/hns/hns_roce_pd.c hns_roce_bitmap_free(&hr_dev->uar_table.bitmap, uar->logic_idx, uar 183 drivers/infiniband/hw/mlx4/cq.c struct mlx4_uar *uar; uar 223 drivers/infiniband/hw/mlx4/cq.c uar = &context->uar; uar 241 drivers/infiniband/hw/mlx4/cq.c uar = &dev->priv_uar; uar 248 drivers/infiniband/hw/mlx4/cq.c err = mlx4_cq_alloc(dev->dev, entries, &cq->buf.mtt, uar, cq->db.dma, uar 1118 drivers/infiniband/hw/mlx4/main.c err = mlx4_uar_alloc(to_mdev(ibdev)->dev, &context->uar); uar 1134 drivers/infiniband/hw/mlx4/main.c mlx4_uar_free(to_mdev(ibdev)->dev, &context->uar); uar 1145 drivers/infiniband/hw/mlx4/main.c mlx4_uar_free(to_mdev(ibcontext->device)->dev, &context->uar); uar 1159 drivers/infiniband/hw/mlx4/main.c to_mucontext(context)->uar.pfn, uar 1168 drivers/infiniband/hw/mlx4/main.c to_mucontext(context)->uar.pfn + uar 85 drivers/infiniband/hw/mlx4/mlx4_ib.h struct mlx4_uar uar; uar 2254 drivers/infiniband/hw/mlx4/qp.c mlx4_to_hw_uar_index(dev->dev, ucontext->uar.index)); uar 637 drivers/infiniband/hw/mlx5/cq.c void __iomem *uar_page = mdev->priv.uar->map; uar 884 drivers/infiniband/hw/mlx5/cq.c *index = dev->mdev->priv.uar->index; uar 6608 drivers/infiniband/hw/mlx5/main.c dev->mdev->priv.uar = mlx5_get_uars_page(dev->mdev); uar 6609 drivers/infiniband/hw/mlx5/main.c return PTR_ERR_OR_ZERO(dev->mdev->priv.uar); uar 6614 drivers/infiniband/hw/mlx5/main.c mlx5_put_uars_page(dev->mdev, dev->mdev->priv.uar); uar 828 drivers/infiniband/hw/mthca/mthca_cq.c cq_context->logsize_usrpage |= cpu_to_be32(ctx->uar.index); uar 459 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_uar_alloc(struct mthca_dev *dev, struct mthca_uar *uar); uar 460 drivers/infiniband/hw/mthca/mthca_dev.h void mthca_uar_free(struct mthca_dev *dev, struct mthca_uar *uar); uar 439 drivers/infiniband/hw/mthca/mthca_memfree.c static u64 mthca_uarc_virt(struct mthca_dev *dev, struct mthca_uar *uar, int page) uar 442 drivers/infiniband/hw/mthca/mthca_memfree.c uar->index * dev->uar_table.uarc_size + uar 446 drivers/infiniband/hw/mthca/mthca_memfree.c int mthca_map_user_db(struct mthca_dev *dev, struct mthca_uar *uar, uar 490 drivers/infiniband/hw/mthca/mthca_memfree.c mthca_uarc_virt(dev, uar, i)); uar 505 drivers/infiniband/hw/mthca/mthca_memfree.c void mthca_unmap_user_db(struct mthca_dev *dev, struct mthca_uar *uar, uar 547 drivers/infiniband/hw/mthca/mthca_memfree.c void mthca_cleanup_user_db_tab(struct mthca_dev *dev, struct mthca_uar *uar, uar 557 drivers/infiniband/hw/mthca/mthca_memfree.c mthca_UNMAP_ICM(dev, mthca_uarc_virt(dev, uar, i), 1); uar 165 drivers/infiniband/hw/mthca/mthca_memfree.h int mthca_map_user_db(struct mthca_dev *dev, struct mthca_uar *uar, uar 167 drivers/infiniband/hw/mthca/mthca_memfree.h void mthca_unmap_user_db(struct mthca_dev *dev, struct mthca_uar *uar, uar 170 drivers/infiniband/hw/mthca/mthca_memfree.h void mthca_cleanup_user_db_tab(struct mthca_dev *dev, struct mthca_uar *uar, uar 321 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_uar_alloc(to_mdev(ibdev), &context->uar); uar 328 drivers/infiniband/hw/mthca/mthca_provider.c mthca_uar_free(to_mdev(ibdev), &context->uar); uar 333 drivers/infiniband/hw/mthca/mthca_provider.c mthca_cleanup_user_db_tab(to_mdev(ibdev), &context->uar, context->db_tab); uar 334 drivers/infiniband/hw/mthca/mthca_provider.c mthca_uar_free(to_mdev(ibdev), &context->uar); uar 345 drivers/infiniband/hw/mthca/mthca_provider.c mthca_cleanup_user_db_tab(to_mdev(context->device), &to_mucontext(context)->uar, uar 347 drivers/infiniband/hw/mthca/mthca_provider.c mthca_uar_free(to_mdev(context->device), &to_mucontext(context)->uar); uar 359 drivers/infiniband/hw/mthca/mthca_provider.c to_mucontext(context)->uar.pfn, uar 423 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_map_user_db(to_mdev(ibsrq->device), &context->uar, uar 438 drivers/infiniband/hw/mthca/mthca_provider.c mthca_unmap_user_db(to_mdev(ibsrq->device), &context->uar, uar 461 drivers/infiniband/hw/mthca/mthca_provider.c mthca_unmap_user_db(to_mdev(srq->device), &context->uar, uar 496 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_map_user_db(to_mdev(pd->device), &context->uar, uar 504 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_map_user_db(to_mdev(pd->device), &context->uar, uar 509 drivers/infiniband/hw/mthca/mthca_provider.c &context->uar, uar 529 drivers/infiniband/hw/mthca/mthca_provider.c &context->uar, uar 533 drivers/infiniband/hw/mthca/mthca_provider.c &context->uar, uar 591 drivers/infiniband/hw/mthca/mthca_provider.c &context->uar, uar 595 drivers/infiniband/hw/mthca/mthca_provider.c &context->uar, uar 627 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_map_user_db(to_mdev(ibdev), &context->uar, uar 633 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_map_user_db(to_mdev(ibdev), &context->uar, uar 669 drivers/infiniband/hw/mthca/mthca_provider.c mthca_unmap_user_db(to_mdev(ibdev), &context->uar, uar 674 drivers/infiniband/hw/mthca/mthca_provider.c mthca_unmap_user_db(to_mdev(ibdev), &context->uar, uar 811 drivers/infiniband/hw/mthca/mthca_provider.c &context->uar, uar 815 drivers/infiniband/hw/mthca/mthca_provider.c &context->uar, uar 66 drivers/infiniband/hw/mthca/mthca_provider.h struct mthca_uar uar; uar 627 drivers/infiniband/hw/mthca/mthca_qp.c qp_context->usr_page = cpu_to_be32(context->uar.index); uar 54 drivers/infiniband/hw/mthca/mthca_srq.c __be32 uar; uar 113 drivers/infiniband/hw/mthca/mthca_srq.c context->uar = cpu_to_be32(ucontext->uar.index); uar 115 drivers/infiniband/hw/mthca/mthca_srq.c context->uar = cpu_to_be32(dev->driver_uar.index); uar 141 drivers/infiniband/hw/mthca/mthca_srq.c context->logstride_usrpage |= cpu_to_be32(ucontext->uar.index); uar 38 drivers/infiniband/hw/mthca/mthca_uar.c int mthca_uar_alloc(struct mthca_dev *dev, struct mthca_uar *uar) uar 40 drivers/infiniband/hw/mthca/mthca_uar.c uar->index = mthca_alloc(&dev->uar_table.alloc); uar 41 drivers/infiniband/hw/mthca/mthca_uar.c if (uar->index == -1) uar 44 drivers/infiniband/hw/mthca/mthca_uar.c uar->pfn = (pci_resource_start(dev->pdev, 2) >> PAGE_SHIFT) + uar->index; uar 49 drivers/infiniband/hw/mthca/mthca_uar.c void mthca_uar_free(struct mthca_dev *dev, struct mthca_uar *uar) uar 51 drivers/infiniband/hw/mthca/mthca_uar.c mthca_free(&dev->uar_table.alloc, uar->index); uar 90 drivers/infiniband/hw/vmw_pvrdma/pvrdma.h struct pvrdma_uar_map *uar; uar 123 drivers/infiniband/hw/vmw_pvrdma/pvrdma.h struct pvrdma_uar_map uar; uar 531 drivers/infiniband/hw/vmw_pvrdma/pvrdma.h int pvrdma_uar_alloc(struct pvrdma_dev *dev, struct pvrdma_uar_map *uar); uar 532 drivers/infiniband/hw/vmw_pvrdma/pvrdma.h void pvrdma_uar_free(struct pvrdma_dev *dev, struct pvrdma_uar_map *uar); uar 200 drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c cq->uar = &context->uar; uar 83 drivers/infiniband/hw/vmw_pvrdma/pvrdma_doorbell.c int pvrdma_uar_alloc(struct pvrdma_dev *dev, struct pvrdma_uar_map *uar) uar 108 drivers/infiniband/hw/vmw_pvrdma/pvrdma_doorbell.c uar->index = obj; uar 109 drivers/infiniband/hw/vmw_pvrdma/pvrdma_doorbell.c uar->pfn = (pci_resource_start(dev->pdev, PVRDMA_PCI_RESOURCE_UAR) >> uar 110 drivers/infiniband/hw/vmw_pvrdma/pvrdma_doorbell.c PAGE_SHIFT) + uar->index; uar 115 drivers/infiniband/hw/vmw_pvrdma/pvrdma_doorbell.c void pvrdma_uar_free(struct pvrdma_dev *dev, struct pvrdma_uar_map *uar) uar 121 drivers/infiniband/hw/vmw_pvrdma/pvrdma_doorbell.c obj = uar->index & (tbl->max - 1); uar 329 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c ret = pvrdma_uar_alloc(vdev, &context->uar); uar 335 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c cmd->pfn = context->uar.pfn; uar 337 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c cmd->pfn64 = context->uar.pfn; uar 353 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c pvrdma_uar_free(vdev, &context->uar); uar 361 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c pvrdma_uar_free(vdev, &context->uar); uar 385 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c pvrdma_uar_free(to_vdev(ibcontext->device), &context->uar); uar 413 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c if (io_remap_pfn_range(vma, start, context->uar.pfn, size, uar 342 drivers/net/ethernet/mellanox/mlx4/cq.c struct mlx4_mtt *mtt, struct mlx4_uar *uar, u64 db_rec, uar 382 drivers/net/ethernet/mellanox/mlx4/cq.c mlx4_to_hw_uar_index(dev, uar->index)); uar 411 drivers/net/ethernet/mellanox/mlx4/cq.c cq->uar = uar; uar 123 drivers/net/ethernet/mellanox/mlx4/en_tx.c ring->bf.uar = &mdev->priv_uar; uar 124 drivers/net/ethernet/mellanox/mlx4/en_tx.c ring->bf.uar->map = mdev->uar_map; uar 206 drivers/net/ethernet/mellanox/mlx4/en_tx.c ring->bf.uar->index)); uar 731 drivers/net/ethernet/mellanox/mlx4/en_tx.c ring->bf.uar->map + MLX4_SEND_DOORBELL); uar 145 drivers/net/ethernet/mellanox/mlx4/pd.c int mlx4_uar_alloc(struct mlx4_dev *dev, struct mlx4_uar *uar) uar 149 drivers/net/ethernet/mellanox/mlx4/pd.c uar->index = mlx4_bitmap_alloc(&mlx4_priv(dev)->uar_table.bitmap); uar 150 drivers/net/ethernet/mellanox/mlx4/pd.c if (uar->index == -1) uar 154 drivers/net/ethernet/mellanox/mlx4/pd.c offset = uar->index % ((int)pci_resource_len(dev->persist->pdev, uar 158 drivers/net/ethernet/mellanox/mlx4/pd.c offset = uar->index; uar 159 drivers/net/ethernet/mellanox/mlx4/pd.c uar->pfn = (pci_resource_start(dev->persist->pdev, 2) >> PAGE_SHIFT) uar 161 drivers/net/ethernet/mellanox/mlx4/pd.c uar->map = NULL; uar 166 drivers/net/ethernet/mellanox/mlx4/pd.c void mlx4_uar_free(struct mlx4_dev *dev, struct mlx4_uar *uar) uar 168 drivers/net/ethernet/mellanox/mlx4/pd.c mlx4_bitmap_free(&mlx4_priv(dev)->uar_table.bitmap, uar->index, MLX4_USE_RR); uar 175 drivers/net/ethernet/mellanox/mlx4/pd.c struct mlx4_uar *uar; uar 184 drivers/net/ethernet/mellanox/mlx4/pd.c uar = list_entry(priv->bf_list.next, struct mlx4_uar, bf_list); uar 190 drivers/net/ethernet/mellanox/mlx4/pd.c uar = kmalloc_node(sizeof(*uar), GFP_KERNEL, node); uar 191 drivers/net/ethernet/mellanox/mlx4/pd.c if (!uar) { uar 192 drivers/net/ethernet/mellanox/mlx4/pd.c uar = kmalloc(sizeof(*uar), GFP_KERNEL); uar 193 drivers/net/ethernet/mellanox/mlx4/pd.c if (!uar) { uar 198 drivers/net/ethernet/mellanox/mlx4/pd.c err = mlx4_uar_alloc(dev, uar); uar 202 drivers/net/ethernet/mellanox/mlx4/pd.c uar->map = ioremap(uar->pfn << PAGE_SHIFT, PAGE_SIZE); uar 203 drivers/net/ethernet/mellanox/mlx4/pd.c if (!uar->map) { uar 208 drivers/net/ethernet/mellanox/mlx4/pd.c uar->bf_map = io_mapping_map_wc(priv->bf_mapping, uar 209 drivers/net/ethernet/mellanox/mlx4/pd.c uar->index << PAGE_SHIFT, uar 211 drivers/net/ethernet/mellanox/mlx4/pd.c if (!uar->bf_map) { uar 215 drivers/net/ethernet/mellanox/mlx4/pd.c uar->free_bf_bmap = 0; uar 216 drivers/net/ethernet/mellanox/mlx4/pd.c list_add(&uar->bf_list, &priv->bf_list); uar 219 drivers/net/ethernet/mellanox/mlx4/pd.c idx = ffz(uar->free_bf_bmap); uar 220 drivers/net/ethernet/mellanox/mlx4/pd.c uar->free_bf_bmap |= 1 << idx; uar 221 drivers/net/ethernet/mellanox/mlx4/pd.c bf->uar = uar; uar 224 drivers/net/ethernet/mellanox/mlx4/pd.c bf->reg = uar->bf_map + idx * dev->caps.bf_reg_size; uar 225 drivers/net/ethernet/mellanox/mlx4/pd.c if (uar->free_bf_bmap == (1 << dev->caps.bf_regs_per_page) - 1) uar 226 drivers/net/ethernet/mellanox/mlx4/pd.c list_del_init(&uar->bf_list); uar 231 drivers/net/ethernet/mellanox/mlx4/pd.c bf->uar = NULL; uar 232 drivers/net/ethernet/mellanox/mlx4/pd.c iounmap(uar->map); uar 235 drivers/net/ethernet/mellanox/mlx4/pd.c mlx4_uar_free(dev, uar); uar 238 drivers/net/ethernet/mellanox/mlx4/pd.c kfree(uar); uar 251 drivers/net/ethernet/mellanox/mlx4/pd.c if (!bf->uar || !bf->uar->bf_map) uar 255 drivers/net/ethernet/mellanox/mlx4/pd.c idx = (bf->reg - bf->uar->bf_map) / dev->caps.bf_reg_size; uar 256 drivers/net/ethernet/mellanox/mlx4/pd.c bf->uar->free_bf_bmap &= ~(1 << idx); uar 257 drivers/net/ethernet/mellanox/mlx4/pd.c if (!bf->uar->free_bf_bmap) { uar 258 drivers/net/ethernet/mellanox/mlx4/pd.c if (!list_empty(&bf->uar->bf_list)) uar 259 drivers/net/ethernet/mellanox/mlx4/pd.c list_del(&bf->uar->bf_list); uar 261 drivers/net/ethernet/mellanox/mlx4/pd.c io_mapping_unmap(bf->uar->bf_map); uar 262 drivers/net/ethernet/mellanox/mlx4/pd.c iounmap(bf->uar->map); uar 263 drivers/net/ethernet/mellanox/mlx4/pd.c mlx4_uar_free(dev, bf->uar); uar 264 drivers/net/ethernet/mellanox/mlx4/pd.c kfree(bf->uar); uar 265 drivers/net/ethernet/mellanox/mlx4/pd.c } else if (list_empty(&bf->uar->bf_list)) uar 266 drivers/net/ethernet/mellanox/mlx4/pd.c list_add(&bf->uar->bf_list, &priv->bf_list); uar 138 drivers/net/ethernet/mellanox/mlx5/core/cq.c cq->uar = dev->priv.uar; uar 147 drivers/net/ethernet/mellanox/mlx5/core/en/txrx.h mlx5_cq_arm(mcq, MLX5_CQ_DB_REQ_NOT, mcq->uar->map, cq->wq.cc); uar 1629 drivers/net/ethernet/mellanox/mlx5/core/en_main.c MLX5_SET(cqc, cqc, uar_page, mdev->priv.uar->index); uar 2233 drivers/net/ethernet/mellanox/mlx5/core/en_main.c MLX5_SET(cqc, cqc, uar_page, priv->mdev->priv.uar->index); uar 292 drivers/net/ethernet/mellanox/mlx5/core/eq.c MLX5_SET(eqc, eqc, uar_page, priv->uar->index); uar 305 drivers/net/ethernet/mellanox/mlx5/core/eq.c eq->doorbell = priv->uar->map + MLX5_EQ_DOORBEL_OFFSET; uar 138 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c mlx5_write64(wqe, conn->fdev->conn_res.uar->map + MLX5_BF_OFFSET); uar 362 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c conn->fdev->conn_res.uar->map, conn->cq.wq.cc); uar 475 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c MLX5_SET(cqc, cqc, uar_page, fdev->conn_res.uar->index); uar 498 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c conn->cq.mcq.uar = fdev->conn_res.uar; uar 582 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c MLX5_SET(qpc, qpc, uar_page, fdev->conn_res.uar->index); uar 1007 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c fdev->conn_res.uar = mlx5_get_uars_page(fdev->mdev); uar 1008 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c if (IS_ERR(fdev->conn_res.uar)) { uar 1009 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c err = PTR_ERR(fdev->conn_res.uar); uar 1014 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c fdev->conn_res.uar->index); uar 1036 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c mlx5_put_uars_page(fdev->mdev, fdev->conn_res.uar); uar 1047 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c mlx5_put_uars_page(fdev->mdev, fdev->conn_res.uar); uar 58 drivers/net/ethernet/mellanox/mlx5/core/fpga/core.h struct mlx5_uars_page *uar; uar 1061 drivers/net/ethernet/mellanox/mlx5/core/main.c dev->priv.uar = mlx5_get_uars_page(dev); uar 1062 drivers/net/ethernet/mellanox/mlx5/core/main.c if (IS_ERR(dev->priv.uar)) { uar 1064 drivers/net/ethernet/mellanox/mlx5/core/main.c err = PTR_ERR(dev->priv.uar); uar 1155 drivers/net/ethernet/mellanox/mlx5/core/main.c mlx5_put_uars_page(dev, dev->priv.uar); uar 1173 drivers/net/ethernet/mellanox/mlx5/core/main.c mlx5_put_uars_page(dev, dev->priv.uar); uar 66 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c dmn->uar = mlx5_get_uars_page(dmn->mdev); uar 67 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c if (!dmn->uar) { uar 100 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c mlx5_put_uars_page(dmn->mdev, dmn->uar); uar 112 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_domain.c mlx5_put_uars_page(dmn->mdev, dmn->uar); uar 47 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c struct mlx5_uars_page *uar; uar 165 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c MLX5_SET(qpc, qpc, uar_page, attr->uar->index); uar 191 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c dr_qp->uar = attr->uar; uar 221 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c mlx5_write64(ctrl, dr_qp->uar->map + MLX5_BF_OFFSET); uar 699 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c struct mlx5_uars_page *uar, uar 750 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c MLX5_SET(cqc, cqc, uar_page, uar->index); uar 779 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c cq->mcq.uar = uar; uar 875 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c dmn->send_ring->cq = dr_create_cq(dmn->mdev, dmn->uar, cq_size); uar 883 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c init_attr.uar = dmn->uar; uar 651 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_types.h struct mlx5_uars_page *uar; uar 973 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_types.h struct mlx5_uars_page *uar; uar 49 drivers/net/ethernet/mellanox/mlx5/core/uar.c *uarn = MLX5_GET(alloc_uar_out, out, uar); uar 60 drivers/net/ethernet/mellanox/mlx5/core/uar.c MLX5_SET(dealloc_uar_in, in, uar, uarn); uar 733 include/linux/mlx4/device.h struct mlx4_uar *uar; uar 741 include/linux/mlx4/device.h struct mlx4_uar *uar; uar 1107 include/linux/mlx4/device.h int mlx4_uar_alloc(struct mlx4_dev *dev, struct mlx4_uar *uar); uar 1108 include/linux/mlx4/device.h void mlx4_uar_free(struct mlx4_dev *dev, struct mlx4_uar *uar); uar 1139 include/linux/mlx4/device.h struct mlx4_uar *uar, u64 db_rec, struct mlx4_cq *cq, uar 45 include/linux/mlx5/cq.h struct mlx5_uars_page *uar; uar 591 include/linux/mlx5/driver.h struct mlx5_uars_page *uar; uar 7109 include/linux/mlx5/mlx5_ifc.h u8 uar[0x18]; uar 7941 include/linux/mlx5/mlx5_ifc.h u8 uar[0x18];