Home
last modified time | relevance | path

Searched refs:lru_list (Results 1 – 17 of 17) sorted by relevance

/linux-4.4.14/include/linux/
Dmm_inline.h26 struct lruvec *lruvec, enum lru_list lru) in add_page_to_lru_list()
35 struct lruvec *lruvec, enum lru_list lru) in del_page_from_lru_list()
51 static inline enum lru_list page_lru_base_type(struct page *page) in page_lru_base_type()
65 static __always_inline enum lru_list page_off_lru(struct page *page) in page_off_lru()
67 enum lru_list lru; in page_off_lru()
89 static __always_inline enum lru_list page_lru(struct page *page) in page_lru()
91 enum lru_list lru; in page_lru()
Dmemcontrol.h360 void mem_cgroup_update_lru_size(struct lruvec *lruvec, enum lru_list lru,
378 unsigned long mem_cgroup_get_lru_size(struct lruvec *lruvec, enum lru_list lru) in mem_cgroup_get_lru_size()
599 mem_cgroup_get_lru_size(struct lruvec *lruvec, enum lru_list lru) in mem_cgroup_get_lru_size()
605 mem_cgroup_update_lru_size(struct lruvec *lruvec, enum lru_list lru, in mem_cgroup_update_lru_size()
Dmmzone.h175 enum lru_list { enum
188 static inline int is_file_lru(enum lru_list lru) in is_file_lru()
193 static inline int is_active_lru(enum lru_list lru) in is_active_lru()
198 static inline int is_unevictable_lru(enum lru_list lru) in is_unevictable_lru()
/linux-4.4.14/fs/xfs/
Dxfs_mru_cache.c143 struct list_head *lru_list; in _xfs_mru_cache_migrate() local
156 lru_list = mru->lists + mru->lru_grp; in _xfs_mru_cache_migrate()
157 if (!list_empty(lru_list)) in _xfs_mru_cache_migrate()
158 list_splice_init(lru_list, mru->reap_list.prev); in _xfs_mru_cache_migrate()
182 lru_list = mru->lists + ((mru->lru_grp + grp) % mru->grp_count); in _xfs_mru_cache_migrate()
183 if (!list_empty(lru_list)) in _xfs_mru_cache_migrate()
/linux-4.4.14/net/netfilter/
Dxt_recent.c71 struct list_head lru_list; member
87 struct list_head lru_list; member
151 list_del(&e->lru_list); in recent_entry_remove()
166 e = list_entry(t->lru_list.next, struct recent_entry, lru_list); in recent_entry_reap()
183 e = list_entry(t->lru_list.next, struct recent_entry, lru_list); in recent_entry_init()
202 list_add_tail(&e->lru_list, &t->lru_list); in recent_entry_init()
213 list_move_tail(&e->lru_list, &t->lru_list); in recent_entry_update()
407 INIT_LIST_HEAD(&t->lru_list); in recent_mt_check()
/linux-4.4.14/drivers/md/
Ddm-bufio.c138 struct list_head lru_list; member
483 list_add(&b->lru_list, &c->lru[dirty]); in __link_buffer()
499 list_del(&b->lru_list); in __unlink_buffer()
514 list_move(&b->lru_list, &c->lru[dirty]); in __relink_lru()
752 list_for_each_entry_reverse(b, &c->lru[LIST_CLEAN], lru_list) { in __get_unclaimed_buffer()
764 list_for_each_entry_reverse(b, &c->lru[LIST_DIRTY], lru_list) { in __get_unclaimed_buffer()
841 struct dm_buffer, lru_list); in __alloc_buffer_wait_no_callback()
842 list_del(&b->lru_list); in __alloc_buffer_wait_no_callback()
879 list_add(&b->lru_list, &c->reserved_buffers); in __free_buffer_wake()
891 list_for_each_entry_safe_reverse(b, tmp, &c->lru[LIST_DIRTY], lru_list) { in __write_dirty_buffers_async()
[all …]
/linux-4.4.14/drivers/gpu/drm/i915/
Di915_gem_fence.c240 list_move_tail(&fence->lru_list, &dev_priv->mm.fence_list); in i915_gem_object_update_fence()
244 list_del_init(&fence->lru_list); in i915_gem_object_update_fence()
333 list_for_each_entry(reg, &dev_priv->mm.fence_list, lru_list) { in i915_find_fence_reg()
388 list_move_tail(&reg->lru_list, in i915_gem_object_get_fence()
Di915_gem_execbuffer.c1079 list_move_tail(&dev_priv->fence_regs[obj->fence_reg].lru_list, in i915_gem_execbuffer_move_to_active()
Di915_drv.h478 struct list_head lru_list; member
Di915_gem.c5013 INIT_LIST_HEAD(&dev_priv->fence_regs[i].lru_list); in i915_gem_load()
/linux-4.4.14/drivers/gpu/drm/vmwgfx/
Dvmwgfx_resource.c1376 struct list_head *lru_list = &dev_priv->res_lru[res->func->res_type]; in vmw_resource_validate() local
1393 if (list_empty(lru_list) || !res->func->may_evict) { in vmw_resource_validate()
1402 (list_first_entry(lru_list, struct vmw_resource, in vmw_resource_validate()
1411 list_add_tail(&evict_res->lru_head, lru_list); in vmw_resource_validate()
1636 struct list_head *lru_list = &dev_priv->res_lru[type]; in vmw_resource_evict_type() local
1644 if (list_empty(lru_list)) in vmw_resource_evict_type()
1648 list_first_entry(lru_list, struct vmw_resource, in vmw_resource_evict_type()
1656 list_add_tail(&evict_res->lru_head, lru_list); in vmw_resource_evict_type()
/linux-4.4.14/mm/
Dmmzone.c90 enum lru_list lru; in lruvec_init()
Dvmscan.c215 static unsigned long get_lru_size(struct lruvec *lruvec, enum lru_list lru) in get_lru_size()
1359 isolate_mode_t mode, enum lru_list lru) in isolate_lru_pages()
1559 struct scan_control *sc, enum lru_list lru) in shrink_inactive_list()
1723 enum lru_list lru) in move_active_pages_to_lru()
1764 enum lru_list lru) in shrink_active_list()
1925 static bool inactive_list_is_low(struct lruvec *lruvec, enum lru_list lru) in inactive_list_is_low()
1933 static unsigned long shrink_list(enum lru_list lru, unsigned long nr_to_scan, in shrink_list()
1974 enum lru_list lru; in get_scan_count()
2188 enum lru_list lru; in shrink_lruvec()
3873 enum lru_list lru = page_lru_base_type(page); in check_move_unevictable_pages()
Dswap.c453 enum lru_list lru = page_lru_base_type(page); in pagevec_move_tail_fn()
1038 enum lru_list lru = page_lru(page); in __pagevec_lru_add_fn()
Dmemcontrol.c739 enum lru_list lru; in mem_cgroup_node_nr_lru_pages()
1125 void mem_cgroup_update_lru_size(struct lruvec *lruvec, enum lru_list lru, in mem_cgroup_update_lru_size()
/linux-4.4.14/fs/gfs2/
Dglock.c67 static LIST_HEAD(lru_list);
139 list_add_tail(&gl->gl_lru, &lru_list); in gfs2_glock_add_to_lru()
1331 list_add(&gl->gl_lru, &lru_list); in gfs2_dispose_glock_lru()
1368 while ((nr-- >= 0) && !list_empty(&lru_list)) { in gfs2_scan_glock_lru()
1369 gl = list_entry(lru_list.next, struct gfs2_glock, gl_lru); in gfs2_scan_glock_lru()
1381 list_splice(&skipped, &lru_list); in gfs2_scan_glock_lru()
/linux-4.4.14/Documentation/vm/
Dunevictable-lru.txt126 lru_list enum.
130 lru_list enum element). The memory controller tracks the movement of pages to