agp_be 53 drivers/gpu/drm/ttm/ttm_agp_backend.c struct ttm_agp_backend *agp_be = container_of(ttm, struct ttm_agp_backend, ttm); agp_be 60 drivers/gpu/drm/ttm/ttm_agp_backend.c mem = agp_allocate_memory(agp_be->bridge, ttm->num_pages, AGP_USER_MEMORY); agp_be 73 drivers/gpu/drm/ttm/ttm_agp_backend.c agp_be->mem = mem; agp_be 87 drivers/gpu/drm/ttm/ttm_agp_backend.c struct ttm_agp_backend *agp_be = container_of(ttm, struct ttm_agp_backend, ttm); agp_be 89 drivers/gpu/drm/ttm/ttm_agp_backend.c if (agp_be->mem) { agp_be 90 drivers/gpu/drm/ttm/ttm_agp_backend.c if (agp_be->mem->is_bound) agp_be 91 drivers/gpu/drm/ttm/ttm_agp_backend.c return agp_unbind_memory(agp_be->mem); agp_be 92 drivers/gpu/drm/ttm/ttm_agp_backend.c agp_free_memory(agp_be->mem); agp_be 93 drivers/gpu/drm/ttm/ttm_agp_backend.c agp_be->mem = NULL; agp_be 100 drivers/gpu/drm/ttm/ttm_agp_backend.c struct ttm_agp_backend *agp_be = container_of(ttm, struct ttm_agp_backend, ttm); agp_be 102 drivers/gpu/drm/ttm/ttm_agp_backend.c if (agp_be->mem) agp_be 105 drivers/gpu/drm/ttm/ttm_agp_backend.c kfree(agp_be); agp_be 118 drivers/gpu/drm/ttm/ttm_agp_backend.c struct ttm_agp_backend *agp_be; agp_be 120 drivers/gpu/drm/ttm/ttm_agp_backend.c agp_be = kmalloc(sizeof(*agp_be), GFP_KERNEL); agp_be 121 drivers/gpu/drm/ttm/ttm_agp_backend.c if (!agp_be) agp_be 124 drivers/gpu/drm/ttm/ttm_agp_backend.c agp_be->mem = NULL; agp_be 125 drivers/gpu/drm/ttm/ttm_agp_backend.c agp_be->bridge = bridge; agp_be 126 drivers/gpu/drm/ttm/ttm_agp_backend.c agp_be->ttm.func = &ttm_agp_func; agp_be 128 drivers/gpu/drm/ttm/ttm_agp_backend.c if (ttm_tt_init(&agp_be->ttm, bo, page_flags)) { agp_be 129 drivers/gpu/drm/ttm/ttm_agp_backend.c kfree(agp_be); agp_be 133 drivers/gpu/drm/ttm/ttm_agp_backend.c return &agp_be->ttm;