shrink 145 arch/powerpc/kernel/tau_6xx.c int shrink; shrink 158 arch/powerpc/kernel/tau_6xx.c shrink = (2 + size - min_window) / 4; shrink 159 arch/powerpc/kernel/tau_6xx.c if (shrink) { shrink 160 arch/powerpc/kernel/tau_6xx.c tau[cpu].low += shrink; shrink 161 arch/powerpc/kernel/tau_6xx.c tau[cpu].high -= shrink; shrink 6168 arch/x86/kvm/mmu.c mmu_shrink_scan(struct shrinker *shrink, struct shrink_control *sc) shrink 6229 arch/x86/kvm/mmu.c mmu_shrink_count(struct shrinker *shrink, struct shrink_control *sc) shrink 974 drivers/android/binder_alloc.c binder_shrink_count(struct shrinker *shrink, struct shrink_control *sc) shrink 981 drivers/android/binder_alloc.c binder_shrink_scan(struct shrinker *shrink, struct shrink_control *sc) shrink 97 drivers/cpuidle/governors/haltpoll.c unsigned int shrink = guest_halt_poll_shrink; shrink 100 drivers/cpuidle/governors/haltpoll.c if (shrink == 0) shrink 103 drivers/cpuidle/governors/haltpoll.c val /= shrink; shrink 2399 drivers/gpu/drm/i915/gem/i915_gem_context.c .shrink = i915_global_gem_context_shrink, shrink 317 drivers/gpu/drm/i915/gem/i915_gem_object.c .shrink = i915_global_objects_shrink, shrink 83 drivers/gpu/drm/i915/gem/i915_gem_shmem.c const unsigned int shrink[] = { shrink 86 drivers/gpu/drm/i915/gem/i915_gem_shmem.c }, *s = shrink; shrink 92 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c unsigned long shrink) shrink 97 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c if (shrink & I915_SHRINK_ACTIVE) shrink 149 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c unsigned int shrink) shrink 167 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c if (!shrinker_lock(i915, shrink, &unlock)) shrink 179 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c trace_i915_gem_shrink(i915, target, shrink); shrink 186 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c if (shrink & I915_SHRINK_BOUND) { shrink 189 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c shrink &= ~I915_SHRINK_BOUND; shrink 216 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c if ((shrink & phase->bit) == 0) shrink 235 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c if (shrink & I915_SHRINK_VMAPS && shrink 239 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c if (!(shrink & I915_SHRINK_ACTIVE) && shrink 243 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c if (!(shrink & I915_SHRINK_BOUND) && shrink 255 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c if (unsafe_drop_pages(obj, shrink)) { shrink 260 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c try_to_writeback(obj, shrink); shrink 275 drivers/gpu/drm/i915/gem/i915_gem_shrinker.c if (shrink & I915_SHRINK_BOUND) shrink 266 drivers/gpu/drm/i915/gt/intel_context.c .shrink = i915_global_context_shrink, shrink 756 drivers/gpu/drm/i915/i915_active.c .shrink = i915_global_active_shrink, shrink 31 drivers/gpu/drm/i915/i915_buddy.c .shrink = i915_global_buddy_shrink, shrink 37 drivers/gpu/drm/i915/i915_globals.c global->shrink(); shrink 49 drivers/gpu/drm/i915/i915_globals.c GEM_BUG_ON(!global->shrink); shrink 17 drivers/gpu/drm/i915/i915_globals.h i915_global_func_t shrink; shrink 1579 drivers/gpu/drm/i915/i915_request.c .shrink = i915_global_request_shrink, shrink 516 drivers/gpu/drm/i915/i915_scheduler.c .shrink = i915_global_scheduler_shrink, shrink 1044 drivers/gpu/drm/i915/i915_vma.c .shrink = i915_global_vma_shrink, shrink 268 drivers/gpu/drm/i915/selftests/scatterlist.c shrink, shrink 377 drivers/gpu/drm/ttm/ttm_page_alloc.c ttm_pool_shrink_scan(struct shrinker *shrink, struct shrink_control *sc) shrink 415 drivers/gpu/drm/ttm/ttm_page_alloc.c ttm_pool_shrink_count(struct shrinker *shrink, struct shrink_control *sc) shrink 1093 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c ttm_dma_pool_shrink_scan(struct shrinker *shrink, struct shrink_control *sc) shrink 1135 drivers/gpu/drm/ttm/ttm_page_alloc_dma.c ttm_dma_pool_shrink_count(struct shrinker *shrink, struct shrink_control *sc) shrink 505 drivers/infiniband/hw/mlx5/mr.c int shrink = 0; shrink 528 drivers/infiniband/hw/mlx5/mr.c shrink = 1; shrink 531 drivers/infiniband/hw/mlx5/mr.c if (shrink) shrink 540 drivers/md/bcache/bcache.h struct shrinker shrink; shrink 697 drivers/md/bcache/btree.c static unsigned long bch_mca_scan(struct shrinker *shrink, shrink 700 drivers/md/bcache/btree.c struct cache_set *c = container_of(shrink, struct cache_set, shrink); shrink 766 drivers/md/bcache/btree.c static unsigned long bch_mca_count(struct shrinker *shrink, shrink 769 drivers/md/bcache/btree.c struct cache_set *c = container_of(shrink, struct cache_set, shrink); shrink 787 drivers/md/bcache/btree.c if (c->shrink.list.next) shrink 788 drivers/md/bcache/btree.c unregister_shrinker(&c->shrink); shrink 854 drivers/md/bcache/btree.c c->shrink.count_objects = bch_mca_count; shrink 855 drivers/md/bcache/btree.c c->shrink.scan_objects = bch_mca_scan; shrink 856 drivers/md/bcache/btree.c c->shrink.seeks = 4; shrink 857 drivers/md/bcache/btree.c c->shrink.batch = c->btree_pages * 2; shrink 859 drivers/md/bcache/btree.c if (register_shrinker(&c->shrink)) shrink 1924 drivers/md/bcache/super.c c->shrink.scan_objects(&c->shrink, &sc); shrink 1926 drivers/md/bcache/super.c c->shrink.scan_objects(&c->shrink, &sc); shrink 834 drivers/md/bcache/sysfs.c c->shrink.scan_objects(&c->shrink, &sc); shrink 1575 drivers/md/dm-bufio.c dm_bufio_shrink_scan(struct shrinker *shrink, struct shrink_control *sc) shrink 1580 drivers/md/dm-bufio.c c = container_of(shrink, struct dm_bufio_client, shrinker); shrink 1592 drivers/md/dm-bufio.c dm_bufio_shrink_count(struct shrinker *shrink, struct shrink_control *sc) shrink 1594 drivers/md/dm-bufio.c struct dm_bufio_client *c = container_of(shrink, struct dm_bufio_client, shrinker); shrink 482 drivers/md/dm-zoned-metadata.c static unsigned long dmz_mblock_shrinker_count(struct shrinker *shrink, shrink 485 drivers/md/dm-zoned-metadata.c struct dmz_metadata *zmd = container_of(shrink, struct dmz_metadata, mblk_shrinker); shrink 493 drivers/md/dm-zoned-metadata.c static unsigned long dmz_mblock_shrinker_scan(struct shrinker *shrink, shrink 496 drivers/md/dm-zoned-metadata.c struct dmz_metadata *zmd = container_of(shrink, struct dmz_metadata, mblk_shrinker); shrink 679 drivers/md/persistent-data/dm-array.c r = ((new_size > old_size) ? grow : shrink)(&resize); shrink 6845 drivers/md/raid5.c static unsigned long raid5_cache_scan(struct shrinker *shrink, shrink 6848 drivers/md/raid5.c struct r5conf *conf = container_of(shrink, struct r5conf, shrinker); shrink 6866 drivers/md/raid5.c static unsigned long raid5_cache_count(struct shrinker *shrink, shrink 6869 drivers/md/raid5.c struct r5conf *conf = container_of(shrink, struct r5conf, shrinker); shrink 465 drivers/staging/android/ashmem.c ashmem_shrink_scan(struct shrinker *shrink, struct shrink_control *sc) shrink 507 drivers/staging/android/ashmem.c ashmem_shrink_count(struct shrinker *shrink, struct shrink_control *sc) shrink 594 drivers/staging/android/ion/ion.c if ((heap->flags & ION_HEAP_FLAG_DEFER_FREE) || heap->ops->shrink) { shrink 94 drivers/staging/android/ion/ion.h int (*shrink)(struct ion_heap *heap, gfp_t gfp_mask, int nr_to_scan); shrink 272 drivers/staging/android/ion/ion_heap.c if (heap->ops->shrink) shrink 273 drivers/staging/android/ion/ion_heap.c total += heap->ops->shrink(heap, sc->gfp_mask, 0); shrink 301 drivers/staging/android/ion/ion_heap.c if (heap->ops->shrink) shrink 302 drivers/staging/android/ion/ion_heap.c freed += heap->ops->shrink(heap, sc->gfp_mask, to_scan); shrink 212 drivers/staging/android/ion/ion_system_heap.c .shrink = ion_system_heap_shrink, shrink 256 fs/erofs/utils.c static unsigned long erofs_shrink_count(struct shrinker *shrink, shrink 262 fs/erofs/utils.c static unsigned long erofs_shrink_scan(struct shrinker *shrink, shrink 1533 fs/ext4/extents_status.c static unsigned long ext4_es_count(struct shrinker *shrink, shrink 1539 fs/ext4/extents_status.c sbi = container_of(shrink, struct ext4_sb_info, s_es_shrinker); shrink 1545 fs/ext4/extents_status.c static unsigned long ext4_es_scan(struct shrinker *shrink, shrink 1548 fs/ext4/extents_status.c struct ext4_sb_info *sbi = container_of(shrink, shrink 5614 fs/ext4/inode.c int shrink = (attr->ia_size < inode->i_size); shrink 5628 fs/ext4/inode.c if (shrink) { shrink 5656 fs/ext4/inode.c if (ext4_handle_valid(handle) && shrink) { shrink 5664 fs/ext4/inode.c if (!shrink) { shrink 5684 fs/ext4/inode.c if (!shrink) { shrink 3521 fs/f2fs/f2fs.h unsigned long f2fs_shrink_count(struct shrinker *shrink, shrink 3523 fs/f2fs/f2fs.h unsigned long f2fs_shrink_scan(struct shrinker *shrink, shrink 39 fs/f2fs/shrinker.c unsigned long f2fs_shrink_count(struct shrinker *shrink, shrink 75 fs/f2fs/shrinker.c unsigned long f2fs_shrink_scan(struct shrinker *shrink, shrink 1614 fs/gfs2/glock.c static unsigned long gfs2_glock_shrink_scan(struct shrinker *shrink, shrink 1622 fs/gfs2/glock.c static unsigned long gfs2_glock_shrink_count(struct shrinker *shrink, shrink 163 fs/gfs2/quota.c static unsigned long gfs2_qd_shrink_scan(struct shrinker *shrink, shrink 180 fs/gfs2/quota.c static unsigned long gfs2_qd_shrink_count(struct shrinker *shrink, shrink 270 fs/mbcache.c static unsigned long mb_cache_count(struct shrinker *shrink, shrink 273 fs/mbcache.c struct mb_cache *cache = container_of(shrink, struct mb_cache, shrink 320 fs/mbcache.c static unsigned long mb_cache_scan(struct shrinker *shrink, shrink 323 fs/mbcache.c struct mb_cache *cache = container_of(shrink, struct mb_cache, shrink 2232 fs/nfs/dir.c nfs_access_cache_scan(struct shrinker *shrink, struct shrink_control *sc) shrink 2244 fs/nfs/dir.c nfs_access_cache_count(struct shrinker *shrink, struct shrink_control *sc) shrink 349 fs/nfs/internal.h extern unsigned long nfs_access_cache_count(struct shrinker *shrink, shrink 351 fs/nfs/internal.h extern unsigned long nfs_access_cache_scan(struct shrinker *shrink, shrink 40 fs/nfsd/nfscache.c static unsigned long nfsd_reply_cache_count(struct shrinker *shrink, shrink 42 fs/nfsd/nfscache.c static unsigned long nfsd_reply_cache_scan(struct shrinker *shrink, shrink 272 fs/nfsd/nfscache.c nfsd_reply_cache_count(struct shrinker *shrink, struct shrink_control *sc) shrink 274 fs/nfsd/nfscache.c struct nfsd_net *nn = container_of(shrink, shrink 281 fs/nfsd/nfscache.c nfsd_reply_cache_scan(struct shrinker *shrink, struct shrink_control *sc) shrink 283 fs/nfsd/nfscache.c struct nfsd_net *nn = container_of(shrink, shrink 664 fs/ocfs2/reservations.c unsigned int shrink; shrink 666 fs/ocfs2/reservations.c shrink = min_bits; shrink 668 fs/ocfs2/reservations.c shrink = lru_resv->r_len / 2; shrink 670 fs/ocfs2/reservations.c lru_resv->r_len -= shrink; shrink 673 fs/ocfs2/reservations.c resv->r_len = shrink; shrink 719 fs/quota/dquot.c dqcache_shrink_scan(struct shrinker *shrink, struct shrink_control *sc) shrink 739 fs/quota/dquot.c dqcache_shrink_count(struct shrinker *shrink, struct shrink_control *sc) shrink 61 fs/super.c static unsigned long super_cache_scan(struct shrinker *shrink, shrink 71 fs/super.c sb = container_of(shrink, struct super_block, s_shrink); shrink 118 fs/super.c static unsigned long super_cache_count(struct shrinker *shrink, shrink 124 fs/super.c sb = container_of(shrink, struct super_block, s_shrink); shrink 267 fs/ubifs/shrinker.c unsigned long ubifs_shrink_count(struct shrinker *shrink, shrink 279 fs/ubifs/shrinker.c unsigned long ubifs_shrink_scan(struct shrinker *shrink, shrink 1880 fs/ubifs/ubifs.h unsigned long ubifs_shrink_scan(struct shrinker *shrink, shrink 1882 fs/ubifs/ubifs.h unsigned long ubifs_shrink_count(struct shrinker *shrink, shrink 1666 fs/xfs/xfs_buf.c struct shrinker *shrink, shrink 1669 fs/xfs/xfs_buf.c struct xfs_buftarg *btp = container_of(shrink, shrink 1689 fs/xfs/xfs_buf.c struct shrinker *shrink, shrink 1692 fs/xfs/xfs_buf.c struct xfs_buftarg *btp = container_of(shrink, shrink 496 fs/xfs/xfs_qm.c struct shrinker *shrink, shrink 499 fs/xfs/xfs_qm.c struct xfs_quotainfo *qi = container_of(shrink, shrink 531 fs/xfs/xfs_qm.c struct shrinker *shrink, shrink 534 fs/xfs/xfs_qm.c struct xfs_quotainfo *qi = container_of(shrink, shrink 100 include/linux/zpool.h int (*shrink)(void *pool, unsigned int pages, shrink 196 include/trace/events/vmscan.h __field(void *, shrink) shrink 208 include/trace/events/vmscan.h __entry->shrink = shr->scan_objects; shrink 219 include/trace/events/vmscan.h __entry->shrink, shrink 240 include/trace/events/vmscan.h __field(void *, shrink) shrink 250 include/trace/events/vmscan.h __entry->shrink = shr->scan_objects; shrink 258 include/trace/events/vmscan.h __entry->shrink, shrink 136 mm/huge_memory.c static unsigned long shrink_huge_zero_page_count(struct shrinker *shrink, shrink 143 mm/huge_memory.c static unsigned long shrink_huge_zero_page_scan(struct shrinker *shrink, shrink 2888 mm/huge_memory.c static unsigned long deferred_split_count(struct shrinker *shrink, shrink 2901 mm/huge_memory.c static unsigned long deferred_split_scan(struct shrinker *shrink, shrink 5333 mm/slub.c SLAB_ATTR(shrink); shrink 1552 mm/z3fold.c .shrink = z3fold_zpool_shrink, shrink 211 mm/zbud.c .shrink = zbud_zpool_shrink, shrink 185 mm/zpool.c zpool->evictable = driver->shrink && ops && ops->evict; shrink 318 mm/zpool.c return zpool->driver->shrink ? shrink 319 mm/zpool.c zpool->driver->shrink(zpool->pool, pages, reclaimed) : -EINVAL; shrink 3160 net/core/filter.c bool shrink = len_diff < 0; shrink 3184 net/core/filter.c if ((shrink && (len_diff_abs >= len_cur || shrink 3186 net/core/filter.c (!shrink && (skb->len + len_diff_abs > len_max && shrink 3190 net/core/filter.c ret = shrink ? bpf_skb_net_shrink(skb, off, len_diff_abs, flags) : shrink 527 net/sunrpc/auth.c rpcauth_cache_shrink_scan(struct shrinker *shrink, struct shrink_control *sc) shrink 541 net/sunrpc/auth.c rpcauth_cache_shrink_count(struct shrinker *shrink, struct shrink_control *sc) shrink 74 tools/vm/slabinfo.c int shrink; shrink 1303 tools/vm/slabinfo.c else if (shrink) shrink 1447 tools/vm/slabinfo.c shrink = 1; shrink 1482 tools/vm/slabinfo.c && !validate && !shrink && !set_debug && !show_ops) shrink 2433 virt/kvm/kvm_main.c unsigned int old, val, shrink; shrink 2436 virt/kvm/kvm_main.c shrink = READ_ONCE(halt_poll_ns_shrink); shrink 2437 virt/kvm/kvm_main.c if (shrink == 0) shrink 2440 virt/kvm/kvm_main.c val /= shrink;