xlt               675 drivers/infiniband/hw/mlx5/mlx5_ib.h 	u32			xlt;
xlt               174 drivers/infiniband/hw/mlx5/mr.c 		MLX5_SET(mkc, mkc, translations_octword_size, ent->xlt);
xlt               643 drivers/infiniband/hw/mlx5/mr.c 		ent->xlt = (1 << ent->order) * sizeof(struct mlx5_mtt) /
xlt               886 drivers/infiniband/hw/mlx5/mr.c 			       void *xlt, int page_shift, size_t size,
xlt               895 drivers/infiniband/hw/mlx5/mr.c 		mlx5_odp_populate_klm(xlt, idx, npages, mr, flags);
xlt               903 drivers/infiniband/hw/mlx5/mr.c 				       idx, npages, xlt,
xlt               908 drivers/infiniband/hw/mlx5/mr.c 		memset(xlt + (npages * sizeof(struct mlx5_mtt)), 0,
xlt               925 drivers/infiniband/hw/mlx5/mr.c 	void *xlt;
xlt               960 drivers/infiniband/hw/mlx5/mr.c 	xlt = (void *)__get_free_pages(gfp, get_order(size));
xlt               961 drivers/infiniband/hw/mlx5/mr.c 	if (!xlt && size > MLX5_SPARE_UMR_CHUNK) {
xlt               966 drivers/infiniband/hw/mlx5/mr.c 		xlt = (void *)__get_free_pages(gfp, get_order(size));
xlt               969 drivers/infiniband/hw/mlx5/mr.c 	if (!xlt) {
xlt               971 drivers/infiniband/hw/mlx5/mr.c 		xlt = (void *)mlx5_ib_get_xlt_emergency_page();
xlt               973 drivers/infiniband/hw/mlx5/mr.c 		memset(xlt, 0, size);
xlt               977 drivers/infiniband/hw/mlx5/mr.c 	dma = dma_map_single(ddev, xlt, size, DMA_TO_DEVICE);
xlt              1007 drivers/infiniband/hw/mlx5/mr.c 		npages = populate_xlt(mr, idx, npages, xlt,
xlt              1041 drivers/infiniband/hw/mlx5/mr.c 		free_pages((unsigned long)xlt, get_order(size));
xlt              1564 drivers/infiniband/hw/mlx5/odp.c 		ent->xlt = MLX5_IMR_MTT_ENTRIES *
xlt              1573 drivers/infiniband/hw/mlx5/odp.c 		ent->xlt = mlx5_imr_ksm_entries *