xlt 675 drivers/infiniband/hw/mlx5/mlx5_ib.h u32 xlt; xlt 174 drivers/infiniband/hw/mlx5/mr.c MLX5_SET(mkc, mkc, translations_octword_size, ent->xlt); xlt 643 drivers/infiniband/hw/mlx5/mr.c ent->xlt = (1 << ent->order) * sizeof(struct mlx5_mtt) / xlt 886 drivers/infiniband/hw/mlx5/mr.c void *xlt, int page_shift, size_t size, xlt 895 drivers/infiniband/hw/mlx5/mr.c mlx5_odp_populate_klm(xlt, idx, npages, mr, flags); xlt 903 drivers/infiniband/hw/mlx5/mr.c idx, npages, xlt, xlt 908 drivers/infiniband/hw/mlx5/mr.c memset(xlt + (npages * sizeof(struct mlx5_mtt)), 0, xlt 925 drivers/infiniband/hw/mlx5/mr.c void *xlt; xlt 960 drivers/infiniband/hw/mlx5/mr.c xlt = (void *)__get_free_pages(gfp, get_order(size)); xlt 961 drivers/infiniband/hw/mlx5/mr.c if (!xlt && size > MLX5_SPARE_UMR_CHUNK) { xlt 966 drivers/infiniband/hw/mlx5/mr.c xlt = (void *)__get_free_pages(gfp, get_order(size)); xlt 969 drivers/infiniband/hw/mlx5/mr.c if (!xlt) { xlt 971 drivers/infiniband/hw/mlx5/mr.c xlt = (void *)mlx5_ib_get_xlt_emergency_page(); xlt 973 drivers/infiniband/hw/mlx5/mr.c memset(xlt, 0, size); xlt 977 drivers/infiniband/hw/mlx5/mr.c dma = dma_map_single(ddev, xlt, size, DMA_TO_DEVICE); xlt 1007 drivers/infiniband/hw/mlx5/mr.c npages = populate_xlt(mr, idx, npages, xlt, xlt 1041 drivers/infiniband/hw/mlx5/mr.c free_pages((unsigned long)xlt, get_order(size)); xlt 1564 drivers/infiniband/hw/mlx5/odp.c ent->xlt = MLX5_IMR_MTT_ENTRIES * xlt 1573 drivers/infiniband/hw/mlx5/odp.c ent->xlt = mlx5_imr_ksm_entries *