agp_be             53 drivers/gpu/drm/ttm/ttm_agp_backend.c 	struct ttm_agp_backend *agp_be = container_of(ttm, struct ttm_agp_backend, ttm);
agp_be             60 drivers/gpu/drm/ttm/ttm_agp_backend.c 	mem = agp_allocate_memory(agp_be->bridge, ttm->num_pages, AGP_USER_MEMORY);
agp_be             73 drivers/gpu/drm/ttm/ttm_agp_backend.c 	agp_be->mem = mem;
agp_be             87 drivers/gpu/drm/ttm/ttm_agp_backend.c 	struct ttm_agp_backend *agp_be = container_of(ttm, struct ttm_agp_backend, ttm);
agp_be             89 drivers/gpu/drm/ttm/ttm_agp_backend.c 	if (agp_be->mem) {
agp_be             90 drivers/gpu/drm/ttm/ttm_agp_backend.c 		if (agp_be->mem->is_bound)
agp_be             91 drivers/gpu/drm/ttm/ttm_agp_backend.c 			return agp_unbind_memory(agp_be->mem);
agp_be             92 drivers/gpu/drm/ttm/ttm_agp_backend.c 		agp_free_memory(agp_be->mem);
agp_be             93 drivers/gpu/drm/ttm/ttm_agp_backend.c 		agp_be->mem = NULL;
agp_be            100 drivers/gpu/drm/ttm/ttm_agp_backend.c 	struct ttm_agp_backend *agp_be = container_of(ttm, struct ttm_agp_backend, ttm);
agp_be            102 drivers/gpu/drm/ttm/ttm_agp_backend.c 	if (agp_be->mem)
agp_be            105 drivers/gpu/drm/ttm/ttm_agp_backend.c 	kfree(agp_be);
agp_be            118 drivers/gpu/drm/ttm/ttm_agp_backend.c 	struct ttm_agp_backend *agp_be;
agp_be            120 drivers/gpu/drm/ttm/ttm_agp_backend.c 	agp_be = kmalloc(sizeof(*agp_be), GFP_KERNEL);
agp_be            121 drivers/gpu/drm/ttm/ttm_agp_backend.c 	if (!agp_be)
agp_be            124 drivers/gpu/drm/ttm/ttm_agp_backend.c 	agp_be->mem = NULL;
agp_be            125 drivers/gpu/drm/ttm/ttm_agp_backend.c 	agp_be->bridge = bridge;
agp_be            126 drivers/gpu/drm/ttm/ttm_agp_backend.c 	agp_be->ttm.func = &ttm_agp_func;
agp_be            128 drivers/gpu/drm/ttm/ttm_agp_backend.c 	if (ttm_tt_init(&agp_be->ttm, bo, page_flags)) {
agp_be            129 drivers/gpu/drm/ttm/ttm_agp_backend.c 		kfree(agp_be);
agp_be            133 drivers/gpu/drm/ttm/ttm_agp_backend.c 	return &agp_be->ttm;