evict             142 arch/powerpc/kvm/book3s_32_mmu_host.c 	bool evict = false;
evict             172 arch/powerpc/kvm/book3s_32_mmu_host.c 		evict = true;
evict             179 arch/powerpc/kvm/book3s_32_mmu_host.c 	if (!evict && (pteg[rr] & PTE_V)) {
evict            1256 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 			   bool evict,
evict            1267 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	amdgpu_vm_bo_invalidate(adev, abo, evict);
evict            1272 drivers/gpu/drm/amd/amdgpu/amdgpu_object.c 	if (evict)
evict             268 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h 			   bool evict,
evict             426 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 			    bool evict, bool no_wait_gpu,
evict             468 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 		r = ttm_bo_pipeline_move(bo, fence, evict, new_mem);
evict             484 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c static int amdgpu_move_vram_ram(struct ttm_buffer_object *bo, bool evict,
evict             526 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	r = amdgpu_move_blit(bo, evict, ctx->no_wait_gpu, &tmp_mem, old_mem);
evict             543 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c static int amdgpu_move_ram_vram(struct ttm_buffer_object *bo, bool evict,
evict             579 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 	r = amdgpu_move_blit(bo, evict, ctx->no_wait_gpu, new_mem, old_mem);
evict             617 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c static int amdgpu_bo_move(struct ttm_buffer_object *bo, bool evict,
evict             663 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 		r = amdgpu_move_vram_ram(bo, evict, ctx, new_mem);
evict             666 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 		r = amdgpu_move_ram_vram(bo, evict, ctx, new_mem);
evict             668 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c 		r = amdgpu_move_blit(bo, evict, ctx->no_wait_gpu,
evict            1131 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 	struct drm_mm_node evict = arg->vma->node;
evict            1137 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 	err = i915_gem_evict_for_node(vm, &evict, 0);
evict              27 drivers/gpu/drm/i915/selftests/i915_live_selftests.h selftest(evict, i915_gem_evict_live_selftests)
evict              24 drivers/gpu/drm/i915/selftests/i915_mock_selftests.h selftest(evict, i915_gem_evict_mock_selftests)
evict             412 drivers/gpu/drm/nouveau/nouveau_bo.c 	bool force = false, evict = false;
evict             424 drivers/gpu/drm/nouveau/nouveau_bo.c 			evict = true;
evict             429 drivers/gpu/drm/nouveau/nouveau_bo.c 		if (!(memtype & (1 << bo->mem.mem_type)) || evict) {
evict             439 drivers/gpu/drm/nouveau/nouveau_bo.c 	if (evict) {
evict            1124 drivers/gpu/drm/nouveau/nouveau_bo.c nouveau_bo_move_m2mf(struct ttm_buffer_object *bo, int evict, bool intr,
evict            1152 drivers/gpu/drm/nouveau/nouveau_bo.c 								evict,
evict            1231 drivers/gpu/drm/nouveau/nouveau_bo.c nouveau_bo_move_flipd(struct ttm_buffer_object *bo, bool evict, bool intr,
evict            1268 drivers/gpu/drm/nouveau/nouveau_bo.c nouveau_bo_move_flips(struct ttm_buffer_object *bo, bool evict, bool intr,
evict            1304 drivers/gpu/drm/nouveau/nouveau_bo.c nouveau_bo_move_ntfy(struct ttm_buffer_object *bo, bool evict,
evict            1363 drivers/gpu/drm/nouveau/nouveau_bo.c nouveau_bo_move(struct ttm_buffer_object *bo, bool evict,
evict            1397 drivers/gpu/drm/nouveau/nouveau_bo.c 			ret = nouveau_bo_move_flipd(bo, evict,
evict            1401 drivers/gpu/drm/nouveau/nouveau_bo.c 			ret = nouveau_bo_move_flips(bo, evict,
evict            1405 drivers/gpu/drm/nouveau/nouveau_bo.c 			ret = nouveau_bo_move_m2mf(bo, evict,
evict             272 drivers/gpu/drm/qxl/qxl_ttm.c static int qxl_bo_move(struct ttm_buffer_object *bo, bool evict,
evict             291 drivers/gpu/drm/qxl/qxl_ttm.c 			       bool evict,
evict             777 drivers/gpu/drm/radeon/radeon_object.c 			   bool evict,
evict             153 drivers/gpu/drm/radeon/radeon_object.h 				  bool evict,
evict             202 drivers/gpu/drm/radeon/radeon_ttm.c 			bool evict, bool no_wait_gpu,
evict             251 drivers/gpu/drm/radeon/radeon_ttm.c 	r = ttm_bo_move_accel_cleanup(bo, &fence->base, evict, new_mem);
evict             257 drivers/gpu/drm/radeon/radeon_ttm.c 				bool evict, bool interruptible,
evict             302 drivers/gpu/drm/radeon/radeon_ttm.c 				bool evict, bool interruptible,
evict             339 drivers/gpu/drm/radeon/radeon_ttm.c static int radeon_bo_move(struct ttm_buffer_object *bo, bool evict,
evict             378 drivers/gpu/drm/radeon/radeon_ttm.c 		r = radeon_move_vram_ram(bo, evict, ctx->interruptible,
evict             382 drivers/gpu/drm/radeon/radeon_ttm.c 		r = radeon_move_ram_vram(bo, evict, ctx->interruptible,
evict             385 drivers/gpu/drm/radeon/radeon_ttm.c 		r = radeon_move_blit(bo, evict, ctx->no_wait_gpu,
evict              19 drivers/gpu/drm/selftests/drm_mm_selftests.h selftest(evict, igt_evict)
evict             321 drivers/gpu/drm/ttm/ttm_bo.c 				  struct ttm_mem_reg *mem, bool evict,
evict             364 drivers/gpu/drm/ttm/ttm_bo.c 				bdev->driver->move_notify(bo, evict, mem);
evict             372 drivers/gpu/drm/ttm/ttm_bo.c 		bdev->driver->move_notify(bo, evict, mem);
evict             378 drivers/gpu/drm/ttm/ttm_bo.c 		ret = bdev->driver->move(bo, evict, ctx, mem);
evict             683 drivers/gpu/drm/ttm/ttm_bo_util.c 			      bool evict,
evict             693 drivers/gpu/drm/ttm/ttm_bo_util.c 	if (evict) {
evict             744 drivers/gpu/drm/ttm/ttm_bo_util.c 			 struct dma_fence *fence, bool evict,
evict             757 drivers/gpu/drm/ttm/ttm_bo_util.c 	if (!evict) {
evict             845 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 			    bool evict,
evict             247 drivers/infiniband/hw/hfi1/mmu_rb.c 		if (handler->ops->evict(handler->ops_arg, rbnode, evict_arg,
evict              70 drivers/infiniband/hw/hfi1/mmu_rb.h 	int (*evict)(void *ops_arg, struct mmu_rb_node *mnode,
evict             119 drivers/infiniband/hw/hfi1/user_sdma.c 	.evict = sdma_rb_evict,
evict             610 fs/inode.c     		evict(inode);
evict            1571 fs/inode.c     	evict(inode);
evict             310 include/drm/ttm/ttm_bo_driver.h 	int (*move)(struct ttm_buffer_object *bo, bool evict,
evict             337 include/drm/ttm/ttm_bo_driver.h 			    bool evict,
evict             866 include/drm/ttm/ttm_bo_driver.h 			      struct dma_fence *fence, bool evict,
evict             881 include/drm/ttm/ttm_bo_driver.h 			 struct dma_fence *fence, bool evict,
evict              10 include/linux/zbud.h 	int (*evict)(struct zbud_pool *pool, unsigned long handle);
evict              18 include/linux/zpool.h 	int (*evict)(struct zpool *pool, unsigned long handle);
evict              75 mm/z3fold.c    	int (*evict)(struct z3fold_pool *pool, unsigned long handle);
evict            1122 mm/z3fold.c    	if (!pool->ops || !pool->ops->evict || retries == 0) {
evict            1199 mm/z3fold.c    			ret = pool->ops->evict(pool, middle_handle);
evict            1204 mm/z3fold.c    			ret = pool->ops->evict(pool, first_handle);
evict            1209 mm/z3fold.c    			ret = pool->ops->evict(pool, last_handle);
evict            1471 mm/z3fold.c    	if (pool->zpool && pool->zpool_ops && pool->zpool_ops->evict)
evict            1472 mm/z3fold.c    		return pool->zpool_ops->evict(pool->zpool, handle);
evict            1478 mm/z3fold.c    	.evict =	z3fold_zpool_evict
evict             131 mm/zbud.c      	if (pool->zpool && pool->zpool_ops && pool->zpool_ops->evict)
evict             132 mm/zbud.c      		return pool->zpool_ops->evict(pool->zpool, handle);
evict             138 mm/zbud.c      	.evict =	zbud_zpool_evict
evict             509 mm/zbud.c      	if (!pool->ops || !pool->ops->evict || list_empty(&pool->lru) ||
evict             534 mm/zbud.c      			ret = pool->ops->evict(pool, first_handle);
evict             539 mm/zbud.c      			ret = pool->ops->evict(pool, last_handle);
evict             185 mm/zpool.c     	zpool->evictable = driver->shrink && ops && ops->evict;
evict             208 mm/zswap.c     	.evict = zswap_writeback_entry