/linux-4.1.27/drivers/gpu/drm/nouveau/ |
H A D | nouveau_bo.c | 696 struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem) nve0_bo_move_copy() 709 OUT_RING (chan, new_mem->num_pages); nve0_bo_move_copy() 728 struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem) nvc0_bo_move_copy() 733 u32 page_count = new_mem->num_pages; nvc0_bo_move_copy() 736 page_count = new_mem->num_pages; nvc0_bo_move_copy() 766 struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem) nvc0_bo_move_m2mf() 771 u32 page_count = new_mem->num_pages; nvc0_bo_move_m2mf() 774 page_count = new_mem->num_pages; nvc0_bo_move_m2mf() 805 struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem) nva3_bo_move_copy() 810 u32 page_count = new_mem->num_pages; nva3_bo_move_copy() 813 page_count = new_mem->num_pages; nva3_bo_move_copy() 843 struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem) nv98_bo_move_exec() 854 OUT_RING (chan, new_mem->num_pages << PAGE_SHIFT); nv98_bo_move_exec() 861 struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem) nv84_bo_move_exec() 867 OUT_RING (chan, new_mem->num_pages << PAGE_SHIFT); nv84_bo_move_exec() 895 struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem) nv50_bo_move_m2mf() 898 u64 length = (new_mem->num_pages << PAGE_SHIFT); nv50_bo_move_m2mf() 902 int dst_tiled = !!((struct nvkm_mem *)new_mem->mm_node)->memtype; nv50_bo_move_m2mf() 991 struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem) nv04_bo_move_m2mf() 994 u32 dst_offset = new_mem->start << PAGE_SHIFT; nv04_bo_move_m2mf() 995 u32 page_count = new_mem->num_pages; nv04_bo_move_m2mf() 1004 OUT_RING (chan, nouveau_bo_mem_ctxdma(bo, chan, new_mem)); nv04_bo_move_m2mf() 1006 page_count = new_mem->num_pages; nv04_bo_move_m2mf() 1063 bool no_wait_gpu, struct ttm_mem_reg *new_mem) nouveau_bo_move_m2mf() 1076 ret = nouveau_bo_move_prep(drm, bo, new_mem); nouveau_bo_move_m2mf() 1084 ret = drm->ttm.move(chan, bo, &bo->mem, new_mem); nouveau_bo_move_m2mf() 1092 new_mem); nouveau_bo_move_m2mf() 1163 bool no_wait_gpu, struct ttm_mem_reg *new_mem) nouveau_bo_move_flipd() 1177 tmp_mem = *new_mem; nouveau_bo_move_flipd() 1191 ret = ttm_bo_move_ttm(bo, true, no_wait_gpu, new_mem); nouveau_bo_move_flipd() 1199 bool no_wait_gpu, struct ttm_mem_reg *new_mem) nouveau_bo_move_flips() 1213 tmp_mem = *new_mem; nouveau_bo_move_flips() 1223 ret = nouveau_bo_move_m2mf(bo, true, intr, no_wait_gpu, new_mem); nouveau_bo_move_flips() 1233 nouveau_bo_move_ntfy(struct ttm_buffer_object *bo, struct ttm_mem_reg *new_mem) nouveau_bo_move_ntfy() argument 1243 if (new_mem && new_mem->mem_type != TTM_PL_SYSTEM && nouveau_bo_move_ntfy() 1244 (new_mem->mem_type == TTM_PL_VRAM || nouveau_bo_move_ntfy() 1246 nvkm_vm_map(vma, new_mem->mm_node); nouveau_bo_move_ntfy() 1254 nouveau_bo_vm_bind(struct ttm_buffer_object *bo, struct ttm_mem_reg *new_mem, nouveau_bo_vm_bind() argument 1260 u64 offset = new_mem->start << PAGE_SHIFT; nouveau_bo_vm_bind() 1263 if (new_mem->mem_type != TTM_PL_VRAM) nouveau_bo_vm_bind() 1267 *new_tile = nv10_bo_set_tiling(dev, offset, new_mem->size, nouveau_bo_vm_bind() 1290 bool no_wait_gpu, struct ttm_mem_reg *new_mem) nouveau_bo_move() 1302 ret = nouveau_bo_vm_bind(bo, new_mem, &new_tile); nouveau_bo_move() 1310 bo->mem = *new_mem; nouveau_bo_move() 1311 new_mem->mm_node = NULL; nouveau_bo_move() 1317 if (new_mem->mem_type == TTM_PL_SYSTEM) nouveau_bo_move() 1319 no_wait_gpu, new_mem); nouveau_bo_move() 1322 no_wait_gpu, new_mem); nouveau_bo_move() 1325 no_wait_gpu, new_mem); nouveau_bo_move() 1333 ret = ttm_bo_move_memcpy(bo, evict, no_wait_gpu, new_mem); nouveau_bo_move() 695 nve0_bo_move_copy(struct nouveau_channel *chan, struct ttm_buffer_object *bo, struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem) nve0_bo_move_copy() argument 727 nvc0_bo_move_copy(struct nouveau_channel *chan, struct ttm_buffer_object *bo, struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem) nvc0_bo_move_copy() argument 765 nvc0_bo_move_m2mf(struct nouveau_channel *chan, struct ttm_buffer_object *bo, struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem) nvc0_bo_move_m2mf() argument 804 nva3_bo_move_copy(struct nouveau_channel *chan, struct ttm_buffer_object *bo, struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem) nva3_bo_move_copy() argument 842 nv98_bo_move_exec(struct nouveau_channel *chan, struct ttm_buffer_object *bo, struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem) nv98_bo_move_exec() argument 860 nv84_bo_move_exec(struct nouveau_channel *chan, struct ttm_buffer_object *bo, struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem) nv84_bo_move_exec() argument 894 nv50_bo_move_m2mf(struct nouveau_channel *chan, struct ttm_buffer_object *bo, struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem) nv50_bo_move_m2mf() argument 990 nv04_bo_move_m2mf(struct nouveau_channel *chan, struct ttm_buffer_object *bo, struct ttm_mem_reg *old_mem, struct ttm_mem_reg *new_mem) nv04_bo_move_m2mf() argument 1062 nouveau_bo_move_m2mf(struct ttm_buffer_object *bo, int evict, bool intr, bool no_wait_gpu, struct ttm_mem_reg *new_mem) nouveau_bo_move_m2mf() argument 1162 nouveau_bo_move_flipd(struct ttm_buffer_object *bo, bool evict, bool intr, bool no_wait_gpu, struct ttm_mem_reg *new_mem) nouveau_bo_move_flipd() argument 1198 nouveau_bo_move_flips(struct ttm_buffer_object *bo, bool evict, bool intr, bool no_wait_gpu, struct ttm_mem_reg *new_mem) nouveau_bo_move_flips() argument 1289 nouveau_bo_move(struct ttm_buffer_object *bo, bool evict, bool intr, bool no_wait_gpu, struct ttm_mem_reg *new_mem) nouveau_bo_move() argument
|
/linux-4.1.27/drivers/gpu/drm/ttm/ |
H A D | ttm_bo_util.c | 49 bool no_wait_gpu, struct ttm_mem_reg *new_mem) ttm_bo_move_ttm() 63 ret = ttm_tt_set_placement_caching(ttm, new_mem->placement); ttm_bo_move_ttm() 67 if (new_mem->mem_type != TTM_PL_SYSTEM) { ttm_bo_move_ttm() 68 ret = ttm_tt_bind(ttm, new_mem); ttm_bo_move_ttm() 73 *old_mem = *new_mem; ttm_bo_move_ttm() 74 new_mem->mm_node = NULL; ttm_bo_move_ttm() 325 struct ttm_mem_reg *new_mem) ttm_bo_move_memcpy() 328 struct ttm_mem_type_manager *man = &bdev->man[new_mem->mem_type]; ttm_bo_move_memcpy() 343 ret = ttm_mem_reg_ioremap(bdev, new_mem, &new_iomap); ttm_bo_move_memcpy() 359 memset_io(new_iomap, 0, new_mem->num_pages*PAGE_SIZE); ttm_bo_move_memcpy() 375 if ((old_mem->mem_type == new_mem->mem_type) && ttm_bo_move_memcpy() 376 (new_mem->start < old_mem->start + old_mem->size)) { ttm_bo_move_memcpy() 378 add = new_mem->num_pages - 1; ttm_bo_move_memcpy() 381 for (i = 0; i < new_mem->num_pages; ++i) { ttm_bo_move_memcpy() 389 pgprot_t prot = ttm_io_prot(new_mem->placement, ttm_bo_move_memcpy() 401 *old_mem = *new_mem; ttm_bo_move_memcpy() 402 new_mem->mm_node = NULL; ttm_bo_move_memcpy() 637 struct ttm_mem_reg *new_mem) ttm_bo_move_accel_cleanup() 640 struct ttm_mem_type_manager *man = &bdev->man[new_mem->mem_type]; ttm_bo_move_accel_cleanup() 690 *old_mem = *new_mem; ttm_bo_move_accel_cleanup() 691 new_mem->mm_node = NULL; ttm_bo_move_accel_cleanup() 47 ttm_bo_move_ttm(struct ttm_buffer_object *bo, bool evict, bool no_wait_gpu, struct ttm_mem_reg *new_mem) ttm_bo_move_ttm() argument 323 ttm_bo_move_memcpy(struct ttm_buffer_object *bo, bool evict, bool no_wait_gpu, struct ttm_mem_reg *new_mem) ttm_bo_move_memcpy() argument 633 ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo, struct fence *fence, bool evict, bool no_wait_gpu, struct ttm_mem_reg *new_mem) ttm_bo_move_accel_cleanup() argument
|
/linux-4.1.27/drivers/gpu/drm/radeon/ |
H A D | radeon_ttm.c | 244 struct ttm_mem_reg *new_mem) radeon_move_null() 249 *old_mem = *new_mem; radeon_move_null() 250 new_mem->mm_node = NULL; radeon_move_null() 255 struct ttm_mem_reg *new_mem, radeon_move_blit() 267 new_start = new_mem->start << PAGE_SHIFT; radeon_move_blit() 280 switch (new_mem->mem_type) { radeon_move_blit() 298 num_pages = new_mem->num_pages * (PAGE_SIZE / RADEON_GPU_PAGE_SIZE); radeon_move_blit() 304 evict, no_wait_gpu, new_mem); radeon_move_blit() 312 struct ttm_mem_reg *new_mem) radeon_move_vram_ram() 322 tmp_mem = *new_mem; radeon_move_vram_ram() 350 r = ttm_bo_move_ttm(bo, true, no_wait_gpu, new_mem); radeon_move_vram_ram() 359 struct ttm_mem_reg *new_mem) radeon_move_ram_vram() 369 tmp_mem = *new_mem; radeon_move_ram_vram() 387 r = radeon_move_blit(bo, true, no_wait_gpu, new_mem, old_mem); radeon_move_ram_vram() 399 struct ttm_mem_reg *new_mem) radeon_bo_move() 407 radeon_move_null(bo, new_mem); radeon_bo_move() 411 new_mem->mem_type == TTM_PL_SYSTEM) || radeon_bo_move() 413 new_mem->mem_type == TTM_PL_TT)) { radeon_bo_move() 415 radeon_move_null(bo, new_mem); radeon_bo_move() 425 new_mem->mem_type == TTM_PL_SYSTEM) { radeon_bo_move() 427 no_wait_gpu, new_mem); radeon_bo_move() 429 new_mem->mem_type == TTM_PL_VRAM) { radeon_bo_move() 431 no_wait_gpu, new_mem); radeon_bo_move() 433 r = radeon_move_blit(bo, evict, no_wait_gpu, new_mem, old_mem); radeon_bo_move() 438 r = ttm_bo_move_memcpy(bo, evict, no_wait_gpu, new_mem); radeon_bo_move() 243 radeon_move_null(struct ttm_buffer_object *bo, struct ttm_mem_reg *new_mem) radeon_move_null() argument 253 radeon_move_blit(struct ttm_buffer_object *bo, bool evict, bool no_wait_gpu, struct ttm_mem_reg *new_mem, struct ttm_mem_reg *old_mem) radeon_move_blit() argument 309 radeon_move_vram_ram(struct ttm_buffer_object *bo, bool evict, bool interruptible, bool no_wait_gpu, struct ttm_mem_reg *new_mem) radeon_move_vram_ram() argument 356 radeon_move_ram_vram(struct ttm_buffer_object *bo, bool evict, bool interruptible, bool no_wait_gpu, struct ttm_mem_reg *new_mem) radeon_move_ram_vram() argument 396 radeon_bo_move(struct ttm_buffer_object *bo, bool evict, bool interruptible, bool no_wait_gpu, struct ttm_mem_reg *new_mem) radeon_bo_move() argument
|
H A D | radeon_object.c | 758 struct ttm_mem_reg *new_mem) radeon_bo_move_notify() 770 if (!new_mem) radeon_bo_move_notify() 774 radeon_update_memory_usage(rbo, new_mem->mem_type, 1); radeon_bo_move_notify() 757 radeon_bo_move_notify(struct ttm_buffer_object *bo, struct ttm_mem_reg *new_mem) radeon_bo_move_notify() argument
|
H A D | radeon_object.h | 153 struct ttm_mem_reg *new_mem);
|
/linux-4.1.27/drivers/gpu/drm/qxl/ |
H A D | qxl_ttm.c | 338 struct ttm_mem_reg *new_mem) qxl_move_null() 343 *old_mem = *new_mem; qxl_move_null() 344 new_mem->mm_node = NULL; qxl_move_null() 350 struct ttm_mem_reg *new_mem) qxl_bo_move() 354 qxl_move_null(bo, new_mem); qxl_bo_move() 357 return ttm_bo_move_memcpy(bo, evict, no_wait_gpu, new_mem); qxl_bo_move() 361 struct ttm_mem_reg *new_mem) qxl_bo_move_notify() 372 qxl_surface_evict(qdev, qbo, new_mem ? true : false); qxl_bo_move_notify() 337 qxl_move_null(struct ttm_buffer_object *bo, struct ttm_mem_reg *new_mem) qxl_move_null() argument 347 qxl_bo_move(struct ttm_buffer_object *bo, bool evict, bool interruptible, bool no_wait_gpu, struct ttm_mem_reg *new_mem) qxl_bo_move() argument 360 qxl_bo_move_notify(struct ttm_buffer_object *bo, struct ttm_mem_reg *new_mem) qxl_bo_move_notify() argument
|
H A D | qxl_cmd.c | 487 struct ttm_mem_reg *new_mem) qxl_hw_surface_alloc() 513 if (new_mem) { qxl_hw_surface_alloc() 520 cmd->u.surface_create.data |= (new_mem->start << PAGE_SHIFT) + surf->tbo.bdev->man[new_mem->mem_type].gpu_offset; qxl_hw_surface_alloc() 485 qxl_hw_surface_alloc(struct qxl_device *qdev, struct qxl_bo *surf, struct ttm_mem_reg *new_mem) qxl_hw_surface_alloc() argument
|
/linux-4.1.27/include/drm/ttm/ |
H A D | ttm_bo_driver.h | 392 * @new_mem: the new memory region receiving the buffer 399 struct ttm_mem_reg *new_mem); 419 struct ttm_mem_reg *new_mem); 952 * @new_mem: struct ttm_mem_reg indicating where to move. 956 * free any old aperture space, and set (@new_mem)->mm_node to NULL, 959 * memory space indicated by @new_mem. 966 struct ttm_mem_reg *new_mem); 974 * @new_mem: struct ttm_mem_reg indicating where to move. 978 * free any old aperture space, and set (@new_mem)->mm_node to NULL, 981 * memory space indicated by @new_mem. 988 struct ttm_mem_reg *new_mem); 1006 * @new_mem: struct ttm_mem_reg indicating where to move. 1019 struct ttm_mem_reg *new_mem);
|
/linux-4.1.27/drivers/pci/hotplug/ |
H A D | ibmphp_res.c | 208 struct resource_node *new_mem = NULL; ibmphp_rsrc_init() local 307 new_mem = alloc_resources (curr); ibmphp_rsrc_init() 308 if (!new_mem) ibmphp_rsrc_init() 310 new_mem->type = MEM; ibmphp_rsrc_init() 318 if (ibmphp_add_resource (new_mem) < 0) { ibmphp_rsrc_init() 322 newbus->firstMem = new_mem; ibmphp_rsrc_init() 324 new_mem->rangeno = -1; ibmphp_rsrc_init() 326 debug ("Memory resource for device %x, bus %x, [%x - %x]\n", new_mem->devfunc, new_mem->busno, new_mem->start, new_mem->end); ibmphp_rsrc_init()
|
/linux-4.1.27/drivers/gpu/drm/ast/ |
H A D | ast_ttm.c | 192 struct ttm_mem_reg *new_mem) ast_bo_move() 195 r = ttm_bo_move_memcpy(bo, evict, no_wait_gpu, new_mem); ast_bo_move() 189 ast_bo_move(struct ttm_buffer_object *bo, bool evict, bool interruptible, bool no_wait_gpu, struct ttm_mem_reg *new_mem) ast_bo_move() argument
|
/linux-4.1.27/drivers/gpu/drm/bochs/ |
H A D | bochs_mm.c | 171 struct ttm_mem_reg *new_mem) bochs_bo_move() 173 return ttm_bo_move_memcpy(bo, evict, no_wait_gpu, new_mem); bochs_bo_move() 168 bochs_bo_move(struct ttm_buffer_object *bo, bool evict, bool interruptible, bool no_wait_gpu, struct ttm_mem_reg *new_mem) bochs_bo_move() argument
|
/linux-4.1.27/drivers/gpu/drm/cirrus/ |
H A D | cirrus_ttm.c | 192 struct ttm_mem_reg *new_mem) cirrus_bo_move() 195 r = ttm_bo_move_memcpy(bo, evict, no_wait_gpu, new_mem); cirrus_bo_move() 189 cirrus_bo_move(struct ttm_buffer_object *bo, bool evict, bool interruptible, bool no_wait_gpu, struct ttm_mem_reg *new_mem) cirrus_bo_move() argument
|
/linux-4.1.27/drivers/gpu/drm/mgag200/ |
H A D | mgag200_ttm.c | 192 struct ttm_mem_reg *new_mem) mgag200_bo_move() 195 r = ttm_bo_move_memcpy(bo, evict, no_wait_gpu, new_mem); mgag200_bo_move() 189 mgag200_bo_move(struct ttm_buffer_object *bo, bool evict, bool interruptible, bool no_wait_gpu, struct ttm_mem_reg *new_mem) mgag200_bo_move() argument
|
/linux-4.1.27/drivers/vfio/pci/ |
H A D | vfio_pci_config.c | 504 bool phys_mem, virt_mem, new_mem, phys_io, virt_io, new_io; vfio_basic_config_write() local 515 new_mem = !!(new_cmd & PCI_COMMAND_MEMORY); vfio_basic_config_write() 522 * If the user is writing mem/io enable (new_mem/io) and we vfio_basic_config_write() 529 if ((new_mem && virt_mem && !phys_mem) || vfio_basic_config_write()
|