uvmm 37 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memgf100.c struct gf100_vmm_map_v0 uvmm = {}; uvmm 47 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memgf100.c uvmm.ro = args->v0.ro; uvmm 48 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memgf100.c uvmm.kind = args->v0.kind; uvmm 59 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memgf100.c ret = nvkm_memory_map(memory, 0, bar, *pvma, &uvmm, sizeof(uvmm)); uvmm 37 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv50.c struct nv50_vmm_map_v0 uvmm = {}; uvmm 48 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv50.c uvmm.ro = args->v0.ro; uvmm 49 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv50.c uvmm.kind = args->v0.kind; uvmm 50 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv50.c uvmm.comp = args->v0.comp; uvmm 62 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/memnv50.c return nvkm_memory_map(memory, 0, bar, *pvma, &uvmm, sizeof(uvmm)); uvmm 46 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c nvkm_uvmm_mthd_pfnclr(struct nvkm_uvmm *uvmm, void *argv, u32 argc) uvmm 48 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_client *client = uvmm->object.client; uvmm 52 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_vmm *vmm = uvmm->vmm; uvmm 75 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c nvkm_uvmm_mthd_pfnmap(struct nvkm_uvmm *uvmm, void *argv, u32 argc) uvmm 77 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_client *client = uvmm->object.client; uvmm 81 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_vmm *vmm = uvmm->vmm; uvmm 109 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c nvkm_uvmm_mthd_unmap(struct nvkm_uvmm *uvmm, void *argv, u32 argc) uvmm 111 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_client *client = uvmm->object.client; uvmm 115 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_vmm *vmm = uvmm->vmm; uvmm 152 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c nvkm_uvmm_mthd_map(struct nvkm_uvmm *uvmm, void *argv, u32 argc) uvmm 154 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_client *client = uvmm->object.client; uvmm 159 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_vmm *vmm = uvmm->vmm; uvmm 231 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c nvkm_uvmm_mthd_put(struct nvkm_uvmm *uvmm, void *argv, u32 argc) uvmm 233 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_client *client = uvmm->object.client; uvmm 237 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_vmm *vmm = uvmm->vmm; uvmm 269 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c nvkm_uvmm_mthd_get(struct nvkm_uvmm *uvmm, void *argv, u32 argc) uvmm 271 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_client *client = uvmm->object.client; uvmm 275 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_vmm *vmm = uvmm->vmm; uvmm 305 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c nvkm_uvmm_mthd_page(struct nvkm_uvmm *uvmm, void *argv, u32 argc) uvmm 314 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c page = uvmm->vmm->func->page; uvmm 335 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_uvmm *uvmm = nvkm_uvmm(object); uvmm 337 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c case NVIF_VMM_V0_PAGE : return nvkm_uvmm_mthd_page (uvmm, argv, argc); uvmm 338 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c case NVIF_VMM_V0_GET : return nvkm_uvmm_mthd_get (uvmm, argv, argc); uvmm 339 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c case NVIF_VMM_V0_PUT : return nvkm_uvmm_mthd_put (uvmm, argv, argc); uvmm 340 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c case NVIF_VMM_V0_MAP : return nvkm_uvmm_mthd_map (uvmm, argv, argc); uvmm 341 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c case NVIF_VMM_V0_UNMAP : return nvkm_uvmm_mthd_unmap (uvmm, argv, argc); uvmm 342 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c case NVIF_VMM_V0_PFNMAP: return nvkm_uvmm_mthd_pfnmap(uvmm, argv, argc); uvmm 343 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c case NVIF_VMM_V0_PFNCLR: return nvkm_uvmm_mthd_pfnclr(uvmm, argv, argc); uvmm 345 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c if (uvmm->vmm->func->mthd) { uvmm 346 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c return uvmm->vmm->func->mthd(uvmm->vmm, uvmm 347 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c uvmm->object.client, uvmm 360 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_uvmm *uvmm = nvkm_uvmm(object); uvmm 361 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c nvkm_vmm_unref(&uvmm->vmm); uvmm 362 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c return uvmm; uvmm 381 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_uvmm *uvmm; uvmm 393 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c if (!(uvmm = kzalloc(sizeof(*uvmm), GFP_KERNEL))) uvmm 395 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c nvkm_object_ctor(&nvkm_uvmm, oclass, &uvmm->object); uvmm 396 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c *pobject = &uvmm->object; uvmm 400 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c NULL, "user", &uvmm->vmm); uvmm 404 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c uvmm->vmm->debug = max(uvmm->vmm->debug, oclass->client->debug); uvmm 409 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c uvmm->vmm = nvkm_vmm_ref(mmu->vmm); uvmm 412 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c page = uvmm->vmm->func->page; uvmm 416 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c args->v0.addr = uvmm->vmm->start; uvmm 417 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c args->v0.size = uvmm->vmm->limit;