Home
last modified time | relevance | path

Searched refs:zone (Results 1 – 192 of 192) sorted by relevance

/linux-4.1.27/drivers/gpu/drm/ttm/
Dttm_memory.c76 struct ttm_mem_zone *zone = in ttm_mem_zone_kobj_release() local
80 zone->name, (unsigned long long)zone->used_mem >> 10); in ttm_mem_zone_kobj_release()
81 kfree(zone); in ttm_mem_zone_kobj_release()
88 struct ttm_mem_zone *zone = in ttm_mem_zone_show() local
92 spin_lock(&zone->glob->lock); in ttm_mem_zone_show()
94 val = zone->zone_mem; in ttm_mem_zone_show()
96 val = zone->emer_mem; in ttm_mem_zone_show()
98 val = zone->max_mem; in ttm_mem_zone_show()
100 val = zone->swap_limit; in ttm_mem_zone_show()
102 val = zone->used_mem; in ttm_mem_zone_show()
[all …]
/linux-4.1.27/mm/
Dcompaction.c103 unsigned long end_pfn, struct zone *zone) in pageblock_pfn_to_page() argument
116 if (page_zone(start_page) != zone) in pageblock_pfn_to_page()
138 void defer_compaction(struct zone *zone, int order) in defer_compaction() argument
140 zone->compact_considered = 0; in defer_compaction()
141 zone->compact_defer_shift++; in defer_compaction()
143 if (order < zone->compact_order_failed) in defer_compaction()
144 zone->compact_order_failed = order; in defer_compaction()
146 if (zone->compact_defer_shift > COMPACT_MAX_DEFER_SHIFT) in defer_compaction()
147 zone->compact_defer_shift = COMPACT_MAX_DEFER_SHIFT; in defer_compaction()
149 trace_mm_compaction_defer_compaction(zone, order); in defer_compaction()
[all …]
Dvmstat.c94 int calculate_pressure_threshold(struct zone *zone) in calculate_pressure_threshold() argument
107 watermark_distance = low_wmark_pages(zone) - min_wmark_pages(zone); in calculate_pressure_threshold()
118 int calculate_normal_threshold(struct zone *zone) in calculate_normal_threshold() argument
153 mem = zone->managed_pages >> (27 - PAGE_SHIFT); in calculate_normal_threshold()
170 struct zone *zone; in refresh_zone_stat_thresholds() local
174 for_each_populated_zone(zone) { in refresh_zone_stat_thresholds()
177 threshold = calculate_normal_threshold(zone); in refresh_zone_stat_thresholds()
180 per_cpu_ptr(zone->pageset, cpu)->stat_threshold in refresh_zone_stat_thresholds()
188 tolerate_drift = low_wmark_pages(zone) - min_wmark_pages(zone); in refresh_zone_stat_thresholds()
191 zone->percpu_drift_mark = high_wmark_pages(zone) + in refresh_zone_stat_thresholds()
[all …]
Dmemory_hotplug.c250 struct zone *zone; in register_page_bootmem_info_node() local
258 zone = &pgdat->node_zones[0]; in register_page_bootmem_info_node()
259 for (; zone < pgdat->node_zones + MAX_NR_ZONES - 1; zone++) { in register_page_bootmem_info_node()
260 if (zone_is_initialized(zone)) { in register_page_bootmem_info_node()
261 nr_pages = zone->wait_table_hash_nr_entries in register_page_bootmem_info_node()
264 page = virt_to_page(zone->wait_table); in register_page_bootmem_info_node()
288 static void __meminit grow_zone_span(struct zone *zone, unsigned long start_pfn, in grow_zone_span() argument
293 zone_span_writelock(zone); in grow_zone_span()
295 old_zone_end_pfn = zone_end_pfn(zone); in grow_zone_span()
296 if (zone_is_empty(zone) || start_pfn < zone->zone_start_pfn) in grow_zone_span()
[all …]
Dpage_alloc.c249 static int page_outside_zone_boundaries(struct zone *zone, struct page *page) in page_outside_zone_boundaries() argument
257 seq = zone_span_seqbegin(zone); in page_outside_zone_boundaries()
258 start_pfn = zone->zone_start_pfn; in page_outside_zone_boundaries()
259 sp = zone->spanned_pages; in page_outside_zone_boundaries()
260 if (!zone_spans_pfn(zone, pfn)) in page_outside_zone_boundaries()
262 } while (zone_span_seqretry(zone, seq)); in page_outside_zone_boundaries()
266 pfn, zone_to_nid(zone), zone->name, in page_outside_zone_boundaries()
272 static int page_is_consistent(struct zone *zone, struct page *page) in page_is_consistent() argument
276 if (zone != page_zone(page)) in page_is_consistent()
284 static int bad_range(struct zone *zone, struct page *page) in bad_range() argument
[all …]
Dvmscan.c164 static unsigned long zone_reclaimable_pages(struct zone *zone) in zone_reclaimable_pages() argument
168 nr = zone_page_state(zone, NR_ACTIVE_FILE) + in zone_reclaimable_pages()
169 zone_page_state(zone, NR_INACTIVE_FILE); in zone_reclaimable_pages()
172 nr += zone_page_state(zone, NR_ACTIVE_ANON) + in zone_reclaimable_pages()
173 zone_page_state(zone, NR_INACTIVE_ANON); in zone_reclaimable_pages()
178 bool zone_reclaimable(struct zone *zone) in zone_reclaimable() argument
180 return zone_page_state(zone, NR_PAGES_SCANNED) < in zone_reclaimable()
181 zone_reclaimable_pages(zone) * 6; in zone_reclaimable()
844 struct zone *zone, in shrink_page_list() argument
882 VM_BUG_ON_PAGE(page_zone(page) != zone, page); in shrink_page_list()
[all …]
Dpage_isolation.c14 struct zone *zone; in set_migratetype_isolate() local
20 zone = page_zone(page); in set_migratetype_isolate()
22 spin_lock_irqsave(&zone->lock, flags); in set_migratetype_isolate()
48 if (!has_unmovable_pages(zone, page, arg.pages_found, in set_migratetype_isolate()
63 zone->nr_isolate_pageblock++; in set_migratetype_isolate()
64 nr_pages = move_freepages_block(zone, page, MIGRATE_ISOLATE); in set_migratetype_isolate()
66 __mod_zone_freepage_state(zone, -nr_pages, migratetype); in set_migratetype_isolate()
69 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate()
71 drain_all_pages(zone); in set_migratetype_isolate()
77 struct zone *zone; in unset_migratetype_isolate() local
[all …]
Dmmzone.c29 struct zone *next_zone(struct zone *zone) in next_zone() argument
31 pg_data_t *pgdat = zone->zone_pgdat; in next_zone()
33 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone()
34 zone++; in next_zone()
38 zone = pgdat->node_zones; in next_zone()
40 zone = NULL; in next_zone()
42 return zone; in next_zone()
68 (z->zone && !zref_in_nodemask(z, nodes))) in next_zones_zonelist()
76 struct page *page, struct zone *zone) in memmap_valid_within() argument
81 if (page_zone(page) != zone) in memmap_valid_within()
Dworkingset.c155 static void *pack_shadow(unsigned long eviction, struct zone *zone) in pack_shadow() argument
157 eviction = (eviction << NODES_SHIFT) | zone_to_nid(zone); in pack_shadow()
158 eviction = (eviction << ZONES_SHIFT) | zone_idx(zone); in pack_shadow()
165 struct zone **zone, in unpack_shadow() argument
181 *zone = NODE_DATA(nid)->node_zones + zid; in unpack_shadow()
183 refault = atomic_long_read(&(*zone)->inactive_age); in unpack_shadow()
215 struct zone *zone = page_zone(page); in workingset_eviction() local
218 eviction = atomic_long_inc_return(&zone->inactive_age); in workingset_eviction()
219 return pack_shadow(eviction, zone); in workingset_eviction()
234 struct zone *zone; in workingset_refault() local
[all …]
Dswap.c55 struct zone *zone = page_zone(page); in __page_cache_release() local
59 spin_lock_irqsave(&zone->lru_lock, flags); in __page_cache_release()
60 lruvec = mem_cgroup_page_lruvec(page, zone); in __page_cache_release()
64 spin_unlock_irqrestore(&zone->lru_lock, flags); in __page_cache_release()
423 struct zone *zone = NULL; in pagevec_lru_move_fn() local
429 struct zone *pagezone = page_zone(page); in pagevec_lru_move_fn()
431 if (pagezone != zone) { in pagevec_lru_move_fn()
432 if (zone) in pagevec_lru_move_fn()
433 spin_unlock_irqrestore(&zone->lru_lock, flags); in pagevec_lru_move_fn()
434 zone = pagezone; in pagevec_lru_move_fn()
[all …]
Dpage_owner.c207 static void init_pages_in_zone(pg_data_t *pgdat, struct zone *zone) in init_pages_in_zone() argument
211 unsigned long pfn = zone->zone_start_pfn, block_end_pfn; in init_pages_in_zone()
212 unsigned long end_pfn = pfn + zone->spanned_pages; in init_pages_in_zone()
216 pfn = zone->zone_start_pfn; in init_pages_in_zone()
265 pgdat->node_id, zone->name, count); in init_pages_in_zone()
270 struct zone *zone; in init_zones_in_node() local
271 struct zone *node_zones = pgdat->node_zones; in init_zones_in_node()
274 for (zone = node_zones; zone - node_zones < MAX_NR_ZONES; ++zone) { in init_zones_in_node()
275 if (!populated_zone(zone)) in init_zones_in_node()
278 spin_lock_irqsave(&zone->lock, flags); in init_zones_in_node()
[all …]
Dmm_init.c33 struct zone *zone; in mminit_verify_zonelist() local
45 zone = &pgdat->node_zones[zoneid]; in mminit_verify_zonelist()
46 if (!populated_zone(zone)) in mminit_verify_zonelist()
52 zone->name); in mminit_verify_zonelist()
55 for_each_zone_zonelist(zone, z, zonelist, zoneid) { in mminit_verify_zonelist()
58 zone->node, zone->name); in mminit_verify_zonelist()
60 printk(KERN_CONT "0:%s ", zone->name); in mminit_verify_zonelist()
133 void __meminit mminit_verify_page_links(struct page *page, enum zone_type zone, in mminit_verify_page_links() argument
137 BUG_ON(page_zonenum(page) != zone); in mminit_verify_page_links()
Dmlock.c176 struct zone *zone = page_zone(page); in munlock_vma_page() local
186 spin_lock_irq(&zone->lru_lock); in munlock_vma_page()
192 __mod_zone_page_state(zone, NR_MLOCK, -nr_pages); in munlock_vma_page()
195 spin_unlock_irq(&zone->lru_lock); in munlock_vma_page()
202 spin_unlock_irq(&zone->lru_lock); in munlock_vma_page()
276 static void __munlock_pagevec(struct pagevec *pvec, struct zone *zone) in __munlock_pagevec() argument
287 spin_lock_irq(&zone->lru_lock); in __munlock_pagevec()
312 __mod_zone_page_state(zone, NR_MLOCK, delta_munlocked); in __munlock_pagevec()
313 spin_unlock_irq(&zone->lru_lock); in __munlock_pagevec()
432 struct zone *zone; in munlock_vma_pages_range() local
[all …]
Dinternal.h101 extern bool zone_reclaimable(struct zone *zone);
128 struct zone *preferred_zone;
191 struct zone *zone; member
366 enum zone_type zone, unsigned long nid, unsigned long pfn);
381 enum zone_type zone, unsigned long nid, unsigned long pfn) in mminit_verify_page_links() argument
420 unsigned long reclaim_clean_pages_from_list(struct zone *zone,
Doom_kill.c202 struct zone *zone; in constrained_alloc() local
234 for_each_zone_zonelist_nodemask(zone, z, zonelist, in constrained_alloc()
236 if (!cpuset_zone_allowed(zone, gfp_mask)) in constrained_alloc()
656 struct zone *zone; in oom_zonelist_trylock() local
660 for_each_zone_zonelist(zone, z, zonelist, gfp_zone(gfp_mask)) in oom_zonelist_trylock()
661 if (test_bit(ZONE_OOM_LOCKED, &zone->flags)) { in oom_zonelist_trylock()
670 for_each_zone_zonelist(zone, z, zonelist, gfp_zone(gfp_mask)) in oom_zonelist_trylock()
671 set_bit(ZONE_OOM_LOCKED, &zone->flags); in oom_zonelist_trylock()
686 struct zone *zone; in oom_zonelist_unlock() local
689 for_each_zone_zonelist(zone, z, zonelist, gfp_zone(gfp_mask)) in oom_zonelist_unlock()
[all …]
Dpage-writeback.c183 static unsigned long zone_dirtyable_memory(struct zone *zone) in zone_dirtyable_memory() argument
187 nr_pages = zone_page_state(zone, NR_FREE_PAGES); in zone_dirtyable_memory()
188 nr_pages -= min(nr_pages, zone->dirty_balance_reserve); in zone_dirtyable_memory()
190 nr_pages += zone_page_state(zone, NR_INACTIVE_FILE); in zone_dirtyable_memory()
191 nr_pages += zone_page_state(zone, NR_ACTIVE_FILE); in zone_dirtyable_memory()
203 struct zone *z = &NODE_DATA(node)->node_zones[ZONE_HIGHMEM]; in highmem_dirtyable_memory()
298 static unsigned long zone_dirty_limit(struct zone *zone) in zone_dirty_limit() argument
300 unsigned long zone_memory = zone_dirtyable_memory(zone); in zone_dirty_limit()
323 bool zone_dirty_ok(struct zone *zone) in zone_dirty_ok() argument
325 unsigned long limit = zone_dirty_limit(zone); in zone_dirty_ok()
[all …]
Dmemcontrol.c586 mem_cgroup_zone_zoneinfo(struct mem_cgroup *memcg, struct zone *zone) in mem_cgroup_zone_zoneinfo() argument
588 int nid = zone_to_nid(zone); in mem_cgroup_zone_zoneinfo()
589 int zid = zone_idx(zone); in mem_cgroup_zone_zoneinfo()
1038 mz = mem_cgroup_zone_zoneinfo(root, reclaim->zone); in mem_cgroup_iter()
1188 struct lruvec *mem_cgroup_zone_lruvec(struct zone *zone, in mem_cgroup_zone_lruvec() argument
1195 lruvec = &zone->lruvec; in mem_cgroup_zone_lruvec()
1199 mz = mem_cgroup_zone_zoneinfo(memcg, zone); in mem_cgroup_zone_lruvec()
1207 if (unlikely(lruvec->zone != zone)) in mem_cgroup_zone_lruvec()
1208 lruvec->zone = zone; in mem_cgroup_zone_lruvec()
1221 struct lruvec *mem_cgroup_page_lruvec(struct page *page, struct zone *zone) in mem_cgroup_page_lruvec() argument
[all …]
Dcma.c102 struct zone *zone; in cma_activate_area() local
110 zone = page_zone(pfn_to_page(pfn)); in cma_activate_area()
124 if (page_zone(pfn_to_page(pfn)) != zone) in cma_activate_area()
Dsparse.c693 int __meminit sparse_add_one_section(struct zone *zone, unsigned long start_pfn) in sparse_add_one_section() argument
696 struct pglist_data *pgdat = zone->zone_pgdat; in sparse_add_one_section()
791 void sparse_remove_one_section(struct zone *zone, struct mem_section *ms) in sparse_remove_one_section() argument
795 struct pglist_data *pgdat = zone->zone_pgdat; in sparse_remove_one_section()
Dquicklist.c30 struct zone *zones = NODE_DATA(node)->node_zones; in max_pages()
Dbacking-dev.c546 long wait_iff_congested(struct zone *zone, int sync, long timeout) in wait_iff_congested() argument
559 !test_bit(ZONE_CONGESTED, &zone->flags)) { in wait_iff_congested()
Dmempolicy.c1614 static int apply_policy_zone(struct mempolicy *policy, enum zone_type zone) in apply_policy_zone() argument
1631 return zone >= dynamic_policy_zone; in apply_policy_zone()
1722 struct zone *zone; in mempolicy_slab_node() local
1727 &zone); in mempolicy_slab_node()
1728 return zone ? zone->node : node; in mempolicy_slab_node()
2253 struct zone *zone; in mpol_misplaced() local
2296 &pol->v.nodes, &zone); in mpol_misplaced()
2297 polnid = zone->node; in mpol_misplaced()
Dmigrate.c1536 struct zone *zone = pgdat->node_zones + z; in migrate_balanced_pgdat() local
1538 if (!populated_zone(zone)) in migrate_balanced_pgdat()
1541 if (!zone_reclaimable(zone)) in migrate_balanced_pgdat()
1545 if (!zone_watermark_ok(zone, 0, in migrate_balanced_pgdat()
1546 high_wmark_pages(zone) + in migrate_balanced_pgdat()
Dhuge_memory.c109 struct zone *zone; in set_recommended_min_free_kbytes() local
113 for_each_populated_zone(zone) in set_recommended_min_free_kbytes()
1639 struct zone *zone = page_zone(page); in __split_huge_page_refcount() local
1644 spin_lock_irq(&zone->lru_lock); in __split_huge_page_refcount()
1645 lruvec = mem_cgroup_page_lruvec(page, zone); in __split_huge_page_refcount()
1724 __mod_zone_page_state(zone, NR_ANON_TRANSPARENT_HUGEPAGES, -1); in __split_huge_page_refcount()
1728 spin_unlock_irq(&zone->lru_lock); in __split_huge_page_refcount()
Dhugetlb.c635 struct zone *zone; in dequeue_huge_page_vma() local
657 for_each_zone_zonelist_nodemask(zone, z, zonelist, in dequeue_huge_page_vma()
659 if (cpuset_zone_allowed(zone, htlb_alloc_mask(h))) { in dequeue_huge_page_vma()
660 page = dequeue_huge_page_node(h, zone_to_nid(zone)); in dequeue_huge_page_vma()
811 static bool zone_spans_last_pfn(const struct zone *zone, in zone_spans_last_pfn() argument
815 return zone_spans_pfn(zone, last_pfn); in zone_spans_last_pfn()
822 struct zone *z; in alloc_gigantic_page()
Dnobootmem.c150 struct zone *z; in reset_node_managed_pages()
Dfilemap.c673 const struct zone *zone = page_zone(page); in page_waitqueue() local
675 return &zone->wait_table[hash_ptr(page, zone->wait_table_bits)]; in page_waitqueue()
Dslab.c3016 struct zone *zone; in fallback_alloc() local
3036 for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { in fallback_alloc()
3037 nid = zone_to_nid(zone); in fallback_alloc()
3039 if (cpuset_zone_allowed(zone, flags) && in fallback_alloc()
Dbootmem.c249 struct zone *z; in reset_node_managed_pages()
Dslub.c1671 struct zone *zone; in get_any_partial() local
1701 for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { in get_any_partial()
1704 n = get_node(s, zone_to_nid(zone)); in get_any_partial()
1706 if (n && cpuset_zone_allowed(zone, flags) && in get_any_partial()
/linux-4.1.27/include/linux/
Dvmstat.h104 #define __count_zone_vm_events(item, zone, delta) \ argument
106 zone_idx(zone), delta)
113 static inline void zone_page_state_add(long x, struct zone *zone, in zone_page_state_add() argument
116 atomic_long_add(x, &zone->vm_stat[item]); in zone_page_state_add()
130 static inline unsigned long zone_page_state(struct zone *zone, in zone_page_state() argument
133 long x = atomic_long_read(&zone->vm_stat[item]); in zone_page_state()
147 static inline unsigned long zone_page_state_snapshot(struct zone *zone, in zone_page_state_snapshot() argument
150 long x = atomic_long_read(&zone->vm_stat[item]); in zone_page_state_snapshot()
155 x += per_cpu_ptr(zone->pageset, cpu)->vm_stat_diff[item]; in zone_page_state_snapshot()
172 struct zone *zones = NODE_DATA(node)->node_zones; in node_page_state()
[all …]
Dmemory_hotplug.h10 struct zone;
62 static inline unsigned zone_span_seqbegin(struct zone *zone) in zone_span_seqbegin() argument
64 return read_seqbegin(&zone->span_seqlock); in zone_span_seqbegin()
66 static inline int zone_span_seqretry(struct zone *zone, unsigned iv) in zone_span_seqretry() argument
68 return read_seqretry(&zone->span_seqlock, iv); in zone_span_seqretry()
70 static inline void zone_span_writelock(struct zone *zone) in zone_span_writelock() argument
72 write_seqlock(&zone->span_seqlock); in zone_span_writelock()
74 static inline void zone_span_writeunlock(struct zone *zone) in zone_span_writeunlock() argument
76 write_sequnlock(&zone->span_seqlock); in zone_span_writeunlock()
78 static inline void zone_seqlock_init(struct zone *zone) in zone_seqlock_init() argument
[all …]
Dmmzone.h220 struct zone *zone; member
327 struct zone { struct
548 static inline unsigned long zone_end_pfn(const struct zone *zone) in zone_end_pfn() argument
550 return zone->zone_start_pfn + zone->spanned_pages; in zone_end_pfn()
553 static inline bool zone_spans_pfn(const struct zone *zone, unsigned long pfn) in zone_spans_pfn() argument
555 return zone->zone_start_pfn <= pfn && pfn < zone_end_pfn(zone); in zone_spans_pfn()
558 static inline bool zone_is_initialized(struct zone *zone) in zone_is_initialized() argument
560 return !!zone->wait_table; in zone_is_initialized()
563 static inline bool zone_is_empty(struct zone *zone) in zone_is_empty() argument
565 return zone->spanned_pages == 0; in zone_is_empty()
[all …]
Dcompaction.h39 extern int fragmentation_index(struct zone *zone, unsigned int order);
45 extern unsigned long compaction_suitable(struct zone *zone, int order,
48 extern void defer_compaction(struct zone *zone, int order);
49 extern bool compaction_deferred(struct zone *zone, int order);
50 extern void compaction_defer_reset(struct zone *zone, int order,
52 extern bool compaction_restarting(struct zone *zone, int order);
71 static inline unsigned long compaction_suitable(struct zone *zone, int order, in compaction_suitable() argument
77 static inline void defer_compaction(struct zone *zone, int order) in defer_compaction() argument
81 static inline bool compaction_deferred(struct zone *zone, int order) in compaction_deferred() argument
Dpage-isolation.h5 static inline bool has_isolate_pageblock(struct zone *zone) in has_isolate_pageblock() argument
7 return zone->nr_isolate_pageblock; in has_isolate_pageblock()
18 static inline bool has_isolate_pageblock(struct zone *zone) in has_isolate_pageblock() argument
32 bool has_unmovable_pages(struct zone *zone, struct page *page, int count,
35 int move_freepages_block(struct zone *zone, struct page *page,
37 int move_freepages(struct zone *zone,
Dmemcontrol.h50 struct zone *zone; member
87 struct lruvec *mem_cgroup_zone_lruvec(struct zone *, struct mem_cgroup *);
88 struct lruvec *mem_cgroup_page_lruvec(struct page *, struct zone *);
179 unsigned long mem_cgroup_soft_limit_reclaim(struct zone *zone, int order,
243 static inline struct lruvec *mem_cgroup_zone_lruvec(struct zone *zone, in mem_cgroup_zone_lruvec() argument
246 return &zone->lruvec; in mem_cgroup_zone_lruvec()
250 struct zone *zone) in mem_cgroup_page_lruvec() argument
252 return &zone->lruvec; in mem_cgroup_page_lruvec()
362 unsigned long mem_cgroup_soft_limit_reclaim(struct zone *zone, int order, in mem_cgroup_soft_limit_reclaim() argument
Dgfp.h378 void drain_zone_pages(struct zone *zone, struct per_cpu_pages *pcp);
379 void drain_all_pages(struct zone *zone);
380 void drain_local_pages(struct zone *zone);
Dswap.h118 struct zone;
328 struct zone *zone,
339 extern int zone_reclaim(struct zone *, gfp_t, unsigned int);
342 static inline int zone_reclaim(struct zone *z, gfp_t mask, unsigned int order) in zone_reclaim()
Dcpuset.h58 static inline int cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed()
174 static inline int cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed()
Dwriteback.h122 bool zone_dirty_ok(struct zone *zone);
Dmm.h747 static inline int zone_to_nid(struct zone *zone) in zone_to_nid() argument
750 return zone->node; in zone_to_nid()
878 static inline struct zone *page_zone(const struct page *page) in page_zone()
896 static inline void set_page_zone(struct page *page, enum zone_type zone) in set_page_zone() argument
899 page->flags |= (zone & ZONES_MASK) << ZONES_PGSHIFT; in set_page_zone()
908 static inline void set_page_links(struct page *page, enum zone_type zone, in set_page_links() argument
911 set_page_zone(page, zone); in set_page_links()
1779 extern void zone_pcp_update(struct zone *zone);
1780 extern void zone_pcp_reset(struct zone *zone);
Dsuspend.h262 extern void mark_free_pages(struct zone *zone);
Dbacking-dev.h279 long wait_iff_congested(struct zone *zone, int sync, long timeout);
/linux-4.1.27/fs/btrfs/
Dreada.c253 struct reada_zone *zone; in reada_find_zone() local
259 zone = NULL; in reada_find_zone()
261 ret = radix_tree_gang_lookup(&dev->reada_zones, (void **)&zone, in reada_find_zone()
264 kref_get(&zone->refcnt); in reada_find_zone()
268 if (logical >= zone->start && logical < zone->end) in reada_find_zone()
269 return zone; in reada_find_zone()
271 kref_put(&zone->refcnt, reada_zone_release); in reada_find_zone()
283 zone = kzalloc(sizeof(*zone), GFP_NOFS); in reada_find_zone()
284 if (!zone) in reada_find_zone()
287 zone->start = start; in reada_find_zone()
[all …]
/linux-4.1.27/include/trace/events/
Dcompaction.h196 TP_PROTO(struct zone *zone,
200 TP_ARGS(zone, order, ret),
210 __entry->nid = zone_to_nid(zone);
211 __entry->name = (char *)zone->name;
225 TP_PROTO(struct zone *zone,
229 TP_ARGS(zone, order, ret)
234 TP_PROTO(struct zone *zone,
238 TP_ARGS(zone, order, ret)
244 TP_PROTO(struct zone *zone, int order),
246 TP_ARGS(zone, order),
[all …]
/linux-4.1.27/kernel/power/
Dsnapshot.c299 struct mem_zone_bm_rtree *zone; member
357 static int add_rtree_block(struct mem_zone_bm_rtree *zone, gfp_t gfp_mask, in add_rtree_block() argument
364 block_nr = zone->blocks; in add_rtree_block()
374 for (i = zone->levels; i < levels_needed; i++) { in add_rtree_block()
376 &zone->nodes); in add_rtree_block()
380 node->data[0] = (unsigned long)zone->rtree; in add_rtree_block()
381 zone->rtree = node; in add_rtree_block()
382 zone->levels += 1; in add_rtree_block()
386 block = alloc_rtree_node(gfp_mask, safe_needed, ca, &zone->leaves); in add_rtree_block()
391 node = zone->rtree; in add_rtree_block()
[all …]
Dpower.h134 extern unsigned int snapshot_additional_pages(struct zone *zone);
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/
Dalloc.c251 struct mlx4_zone_entry *zone = kmalloc(sizeof(*zone), GFP_KERNEL); in mlx4_zone_add_one() local
253 if (NULL == zone) in mlx4_zone_add_one()
256 zone->flags = flags; in mlx4_zone_add_one()
257 zone->bitmap = bitmap; in mlx4_zone_add_one()
258 zone->use_rr = (flags & MLX4_ZONE_USE_RR) ? MLX4_USE_RR : 0; in mlx4_zone_add_one()
259 zone->priority = priority; in mlx4_zone_add_one()
260 zone->offset = offset; in mlx4_zone_add_one()
264 zone->uid = zone_alloc->last_uid++; in mlx4_zone_add_one()
265 zone->allocator = zone_alloc; in mlx4_zone_add_one()
275 list_add_tail(&zone->prio_list, &it->prio_list); in mlx4_zone_add_one()
[all …]
/linux-4.1.27/drivers/mtd/
Dsm_ftl.c195 static loff_t sm_mkoffset(struct sm_ftl *ftl, int zone, int block, int boffset) in sm_mkoffset() argument
198 WARN_ON(zone < 0 || zone >= ftl->zone_count); in sm_mkoffset()
205 return (zone * SM_MAX_ZONE_SIZE + block) * ftl->block_size + boffset; in sm_mkoffset()
210 int *zone, int *block, int *boffset) in sm_break_offset() argument
214 *zone = offset >= ftl->zone_count ? -1 : offset; in sm_break_offset()
237 int zone, int block, int boffset, in sm_read_sector() argument
267 if (zone == 0 && block == ftl->cis_block && boffset == in sm_read_sector()
278 ret = mtd_read_oob(mtd, sm_mkoffset(ftl, zone, block, boffset), &ops); in sm_read_sector()
283 block, zone, ret); in sm_read_sector()
301 " as bad" , block, zone); in sm_read_sector()
[all …]
/linux-4.1.27/drivers/thermal/int340x_thermal/
Dint340x_thermal_zone.c22 static int int340x_thermal_get_zone_temp(struct thermal_zone_device *zone, in int340x_thermal_get_zone_temp() argument
25 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_zone_temp()
30 return d->override_ops->get_temp(zone, temp); in int340x_thermal_get_zone_temp()
51 static int int340x_thermal_get_trip_temp(struct thermal_zone_device *zone, in int340x_thermal_get_trip_temp() argument
54 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_trip_temp()
58 return d->override_ops->get_trip_temp(zone, trip, temp); in int340x_thermal_get_trip_temp()
83 static int int340x_thermal_get_trip_type(struct thermal_zone_device *zone, in int340x_thermal_get_trip_type() argument
87 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_trip_type()
91 return d->override_ops->get_trip_type(zone, trip, type); in int340x_thermal_get_trip_type()
116 static int int340x_thermal_set_trip_temp(struct thermal_zone_device *zone, in int340x_thermal_set_trip_temp() argument
[all …]
Dint340x_thermal_zone.h40 struct thermal_zone_device *zone; member
65 thermal_zone_device_update(tzone->zone); in int340x_thermal_zone_device_update()
Dprocessor_thermal_device.c173 static int proc_thermal_get_zone_temp(struct thermal_zone_device *zone, in proc_thermal_get_zone_temp() argument
/linux-4.1.27/fs/adfs/
Dmap.c192 scan_map(struct adfs_sb_info *asb, unsigned int zone, in scan_map() argument
199 dm = asb->s_map + zone; in scan_map()
200 zone = asb->s_map_size; in scan_map()
201 dm_end = asb->s_map + zone; in scan_map()
212 } while (--zone > 0); in scan_map()
235 unsigned int zone; in adfs_map_free() local
238 zone = asb->s_map_size; in adfs_map_free()
242 } while (--zone > 0); in adfs_map_free()
252 unsigned int zone, mapoff; in adfs_map_lookup() local
260 zone = asb->s_map_size >> 1; in adfs_map_lookup()
[all …]
Dsuper.c305 int i, zone; in adfs_read_map() local
322 for (zone = 0; zone < nzones; zone++, map_addr++) { in adfs_read_map()
323 dm[zone].dm_startbit = 0; in adfs_read_map()
324 dm[zone].dm_endbit = zone_size; in adfs_read_map()
325 dm[zone].dm_startblk = zone * zone_size - ADFS_DR_SIZE_BITS; in adfs_read_map()
326 dm[zone].dm_bh = sb_bread(sb, map_addr); in adfs_read_map()
328 if (!dm[zone].dm_bh) { in adfs_read_map()
335 i = zone - 1; in adfs_read_map()
348 while (--zone >= 0) in adfs_read_map()
349 brelse(dm[zone].dm_bh); in adfs_read_map()
/linux-4.1.27/Documentation/thermal/
Dsysfs-api.txt13 The generic thermal sysfs provides a set of interfaces for thermal zone
17 This how-to focuses on enabling new thermal zone and cooling devices to
19 This solution is platform independent and any type of thermal zone devices
22 The main task of the thermal sysfs driver is to expose thermal zone attributes
25 inputs from thermal zone attributes (the current temperature and trip point
33 1.1 thermal zone device interface
40 This interface function adds a new thermal zone device (sensor) to
44 type: the thermal zone type.
45 trips: the total number of trip points this thermal zone supports.
48 ops: thermal zone device call-backs.
[all …]
Dx86_pkg_temperature_thermal19 zone with maximum two user mode configurable trip points. Number of trip points
27 Each package will register as a thermal zone under /sys/class/thermal.
42 Thermal notifications: To get kobject-uevent notifications, set the thermal zone
Dexynos_thermal_emulation18 /sys/devices/virtual/thermal/thermal_zone'zone id'/emul_temp.
/linux-4.1.27/lib/
Dshow_mem.c26 struct zone *zone = &pgdat->node_zones[zoneid]; in show_mem() local
27 if (!populated_zone(zone)) in show_mem()
30 total += zone->present_pages; in show_mem()
31 reserved += zone->present_pages - zone->managed_pages; in show_mem()
34 highmem += zone->present_pages; in show_mem()
/linux-4.1.27/Documentation/ABI/testing/
Dsysfs-class-powercap25 This status affects every power zone using this "control_type.
27 What: /sys/class/powercap/<control type>/<power zone>
32 A power zone is a single or a collection of devices, which can
33 be independently monitored and controlled. A power zone sysfs
37 What: /sys/class/powercap/<control type>/<power zone>/<child power zone>
45 power zone for a whole CPU package, each CPU core in it can
46 be a child power zone.
48 What: /sys/class/powercap/.../<power zone>/name
53 Specifies the name of this power zone.
55 What: /sys/class/powercap/.../<power zone>/energy_uj
[all …]
Dsysfs-class-backlight-driver-adp88701 What: /sys/class/backlight/<backlight>/<ambient light zone>_max
11 Control the maximum brightness for <ambient light zone>
14 <ambient light zone>.
16 What: /sys/class/backlight/<backlight>/<ambient light zone>_dim
25 Control the dim brightness for <ambient light zone>
29 this <ambient light zone>.
46 Get/Set current ambient light zone. Reading returns
49 to enter the corresponding ambient light zone.
Dsysfs-bus-iio-light-lm3533-als7 in each direction (rising|falling) and a zone change occurs.
8 The corresponding light zone can be read from
32 These thresholds correspond to the eight zone-boundary
41 Get the current light zone (0..4) as defined by the
50 out_currentY_currentZ_raw, where Z is the current zone.
57 Set the output current for channel out_currentY when in zone
Dsysfs-devices-memory69 read-only and is designed to show which zone this memory
/linux-4.1.27/drivers/staging/rts5208/
Dxd.c879 xd_card->zone = vmalloc(size); in xd_init_l2p_tbl()
880 if (!xd_card->zone) { in xd_init_l2p_tbl()
886 xd_card->zone[i].build_flag = 0; in xd_init_l2p_tbl()
887 xd_card->zone[i].l2p_table = NULL; in xd_init_l2p_tbl()
888 xd_card->zone[i].free_table = NULL; in xd_init_l2p_tbl()
889 xd_card->zone[i].get_index = 0; in xd_init_l2p_tbl()
890 xd_card->zone[i].set_index = 0; in xd_init_l2p_tbl()
891 xd_card->zone[i].unused_blk_cnt = 0; in xd_init_l2p_tbl()
897 static inline void free_zone(struct zone_entry *zone) in free_zone() argument
899 if (!zone) in free_zone()
[all …]
Drtsx_chip.h529 struct zone_entry *zone; member
/linux-4.1.27/sound/synth/emux/
Dsoundfont.c525 struct snd_sf_zone *zone; in load_info() local
563 for (zone = sf->zones; zone; zone = zone->next) { in load_info()
564 if (!zone->mapped && in load_info()
565 zone->bank == hdr.bank && in load_info()
566 zone->instr == hdr.instr) in load_info()
595 if ((zone = sf_zone_new(sflist, sf)) == NULL) { in load_info()
600 zone->bank = tmpzone.bank; in load_info()
601 zone->instr = tmpzone.instr; in load_info()
602 zone->v = tmpzone.v; in load_info()
605 zone->sample = set_sample(sf, &zone->v); in load_info()
[all …]
Demux_synth.c115 vp->zone = table[i]; in snd_emux_note_on()
116 if (vp->zone->sample) in snd_emux_note_on()
117 vp->block = vp->zone->sample->block; in snd_emux_note_on()
513 vp->zone = NULL; in terminate_voice()
566 vp->reg = vp->zone->v; in setup_voice()
Demux_effect.c215 origp = (unsigned char*)&vp->zone->v.parm + offset; in snd_emux_send_effect()
/linux-4.1.27/virt/kvm/
Dcoalesced_mmio.c35 if (addr < dev->zone.addr) in coalesced_mmio_in_range()
37 if (addr + len > dev->zone.addr + dev->zone.size) in coalesced_mmio_in_range()
137 struct kvm_coalesced_mmio_zone *zone) in kvm_vm_ioctl_register_coalesced_mmio() argument
148 dev->zone = *zone; in kvm_vm_ioctl_register_coalesced_mmio()
151 ret = kvm_io_bus_register_dev(kvm, KVM_MMIO_BUS, zone->addr, in kvm_vm_ioctl_register_coalesced_mmio()
152 zone->size, &dev->dev); in kvm_vm_ioctl_register_coalesced_mmio()
168 struct kvm_coalesced_mmio_zone *zone) in kvm_vm_ioctl_unregister_coalesced_mmio() argument
175 if (coalesced_mmio_in_range(dev, zone->addr, zone->size)) { in kvm_vm_ioctl_unregister_coalesced_mmio()
Dcoalesced_mmio.h21 struct kvm_coalesced_mmio_zone zone; member
27 struct kvm_coalesced_mmio_zone *zone);
29 struct kvm_coalesced_mmio_zone *zone);
Dkvm_main.c2533 struct kvm_coalesced_mmio_zone zone; in kvm_vm_ioctl() local
2536 if (copy_from_user(&zone, argp, sizeof(zone))) in kvm_vm_ioctl()
2538 r = kvm_vm_ioctl_register_coalesced_mmio(kvm, &zone); in kvm_vm_ioctl()
2542 struct kvm_coalesced_mmio_zone zone; in kvm_vm_ioctl() local
2545 if (copy_from_user(&zone, argp, sizeof(zone))) in kvm_vm_ioctl()
2547 r = kvm_vm_ioctl_unregister_coalesced_mmio(kvm, &zone); in kvm_vm_ioctl()
/linux-4.1.27/arch/x86/mm/
Dhighmem_32.c122 struct zone *zone; in set_highmem_pages_init() local
130 for_each_zone(zone) { in set_highmem_pages_init()
133 if (!is_highmem(zone)) in set_highmem_pages_init()
136 zone_start_pfn = zone->zone_start_pfn; in set_highmem_pages_init()
137 zone_end_pfn = zone_start_pfn + zone->spanned_pages; in set_highmem_pages_init()
139 nid = zone_to_nid(zone); in set_highmem_pages_init()
141 zone->name, nid, zone_start_pfn, zone_end_pfn); in set_highmem_pages_init()
Dinit_32.c829 struct zone *zone = pgdata->node_zones + in arch_add_memory() local
834 return __add_pages(nid, zone, start_pfn, nr_pages); in arch_add_memory()
842 struct zone *zone; in arch_remove_memory() local
844 zone = page_zone(pfn_to_page(start_pfn)); in arch_remove_memory()
845 return __remove_pages(zone, start_pfn, nr_pages); in arch_remove_memory()
Dinit_64.c693 struct zone *zone = pgdat->node_zones + in arch_add_memory() local
701 ret = __add_pages(nid, zone, start_pfn, nr_pages); in arch_add_memory()
1021 struct zone *zone; in arch_remove_memory() local
1024 zone = page_zone(pfn_to_page(start_pfn)); in arch_remove_memory()
1026 ret = __remove_pages(zone, start_pfn, nr_pages); in arch_remove_memory()
/linux-4.1.27/fs/xfs/
Dkmem.h105 kmem_zone_free(kmem_zone_t *zone, void *ptr) in kmem_zone_free() argument
107 kmem_cache_free(zone, ptr); in kmem_zone_free()
111 kmem_zone_destroy(kmem_zone_t *zone) in kmem_zone_destroy() argument
113 if (zone) in kmem_zone_destroy()
114 kmem_cache_destroy(zone); in kmem_zone_destroy()
120 kmem_zone_zalloc(kmem_zone_t *zone, xfs_km_flags_t flags) in kmem_zone_zalloc() argument
122 return kmem_zone_alloc(zone, flags | KM_ZERO); in kmem_zone_zalloc()
Dkmem.c111 kmem_zone_alloc(kmem_zone_t *zone, xfs_km_flags_t flags) in kmem_zone_alloc() argument
118 ptr = kmem_cache_alloc(zone, lflags); in kmem_zone_alloc()
/linux-4.1.27/Documentation/vm/
Dbalance15 is, only when needed (aka zone free memory is 0), instead of making it
24 of incurring the overhead of regular zone balancing.
29 would not be done even when the dma zone was completely empty. 2.2 has
34 In 2.3, zone balancing can be done in one of two ways: depending on the
35 zone size (and possibly of the size of lower class zones), we can decide
37 zone. The good part is, while balancing, we do not need to look at sizes
44 of a zone _and_ all its lower class zones falls below 1/64th of the
45 total memory in the zone and its lower class zones. This fixes the 2.2
52 Note that if the size of the regular zone is huge compared to dma zone,
54 deciding whether to balance the regular zone. The first solution
[all …]
Dnuma71 each memory zone [one or more of DMA, DMA32, NORMAL, HIGH_MEMORY, MOVABLE],
73 selected zone/node cannot satisfy the allocation request. This situation,
74 when a zone has no available memory to satisfy a request, is called
79 fall back to the same zone type on a different node, or to a different zone
82 a default zonelist order based on the sizes of the various zone types relative
93 nodes' zones in the selected zonelist looking for the first zone in the list
120 node the "local memory node"--the node of the first zone in CPU's node's
Dunevictable-lru.txt60 main memory will have over 32 million 4k pages in a single zone. When a large
82 The Unevictable LRU infrastructure consists of an additional, per-zone, LRU list
111 The unevictable list benefits from the "arrayification" of the per-zone LRU
115 unevictable pages are placed directly on the page's zone's unevictable list
116 under the zone lru_lock. This allows us to prevent the stranding of pages on
128 The memory controller data structure automatically gets a per-zone unevictable
129 list as a result of the "arrayification" of the per-zone LRU lists (one per
216 unevictable list for the zone being scanned.
349 is now mlocked and divert the page to the zone's unevictable list. If
411 mlocked, munlock_vma_page() updates that zone statistics for the number of
[all …]
/linux-4.1.27/drivers/iio/light/
Dlm3533-als.c61 atomic_t zone; member
90 static int _lm3533_als_get_zone(struct iio_dev *indio_dev, u8 *zone) in _lm3533_als_get_zone() argument
103 *zone = min_t(u8, val, LM3533_ALS_ZONE_MAX); in _lm3533_als_get_zone()
108 static int lm3533_als_get_zone(struct iio_dev *indio_dev, u8 *zone) in lm3533_als_get_zone() argument
114 *zone = atomic_read(&als->zone); in lm3533_als_get_zone()
116 ret = _lm3533_als_get_zone(indio_dev, zone); in lm3533_als_get_zone()
128 static inline u8 lm3533_als_get_target_reg(unsigned channel, unsigned zone) in lm3533_als_get_target_reg() argument
130 return LM3533_REG_ALS_TARGET_BASE + 5 * channel + zone; in lm3533_als_get_target_reg()
134 unsigned zone, u8 *val) in lm3533_als_get_target() argument
143 if (zone > LM3533_ALS_ZONE_MAX) in lm3533_als_get_target()
[all …]
DKconfig161 light zone through sysfs. A threshold event can be generated on zone
162 changes. The ALS-control output values can be set per zone for the
/linux-4.1.27/drivers/thermal/
Drcar_thermal.c61 struct thermal_zone_device *zone; member
73 #define rcar_zone_to_priv(zone) ((zone)->devdata) argument
203 static int rcar_thermal_get_temp(struct thermal_zone_device *zone, in rcar_thermal_get_temp() argument
206 struct rcar_thermal_priv *priv = rcar_zone_to_priv(zone); in rcar_thermal_get_temp()
218 static int rcar_thermal_get_trip_type(struct thermal_zone_device *zone, in rcar_thermal_get_trip_type() argument
221 struct rcar_thermal_priv *priv = rcar_zone_to_priv(zone); in rcar_thermal_get_trip_type()
237 static int rcar_thermal_get_trip_temp(struct thermal_zone_device *zone, in rcar_thermal_get_trip_temp() argument
240 struct rcar_thermal_priv *priv = rcar_zone_to_priv(zone); in rcar_thermal_get_trip_temp()
256 static int rcar_thermal_notify(struct thermal_zone_device *zone, in rcar_thermal_notify() argument
259 struct rcar_thermal_priv *priv = rcar_zone_to_priv(zone); in rcar_thermal_notify()
[all …]
Dtegra_soctherm.c311 struct tegra_thermctl_zone *zone = data; in tegra_thermctl_get_temp() local
314 val = (readl(zone->reg) >> zone->shift) & SENSOR_TEMP_MASK; in tegra_thermctl_get_temp()
412 struct tegra_thermctl_zone *zone = in tegra_soctherm_probe() local
413 devm_kzalloc(&pdev->dev, sizeof(*zone), GFP_KERNEL); in tegra_soctherm_probe()
414 if (!zone) { in tegra_soctherm_probe()
419 zone->reg = tegra->regs + t124_thermctl_temp_zones[i].offset; in tegra_soctherm_probe()
420 zone->shift = t124_thermctl_temp_zones[i].shift; in tegra_soctherm_probe()
422 tz = thermal_zone_of_sensor_register(&pdev->dev, i, zone, in tegra_soctherm_probe()
Dof-thermal.c395 thermal_zone_of_add_sensor(struct device_node *zone, in thermal_zone_of_add_sensor() argument
402 tzd = thermal_zone_get_zone_by_name(zone->name); in thermal_zone_of_add_sensor()
866 struct thermal_zone_device *zone; in of_parse_thermal_zones() local
894 zone = thermal_zone_device_register(child->name, tz->ntrips, in of_parse_thermal_zones()
899 if (IS_ERR(zone)) { in of_parse_thermal_zones()
901 PTR_ERR(zone)); in of_parse_thermal_zones()
941 struct thermal_zone_device *zone; in of_thermal_destroy_zones() local
947 zone = thermal_zone_get_zone_by_name(child->name); in of_thermal_destroy_zones()
948 if (IS_ERR(zone)) in of_thermal_destroy_zones()
951 thermal_zone_device_unregister(zone); in of_thermal_destroy_zones()
[all …]
DKconfig10 zone and cooling device.
11 Each thermal zone contains its own temperature, trip points,
64 devices based on their 'contribution' to a zone. The
130 Enable this option to make a emul_temp sysfs node in thermal zone
198 management framework. A thermal zone with several trip points will be
200 thermal zone if trip points reached.
227 bound to thermal zone trip points. When a trip point reached, the
248 thermal zone. Each package will have its own thermal zone. There are
259 thermal zone. There are two trip points. One of the trip point can
/linux-4.1.27/arch/s390/mm/
Dinit.c176 struct zone *zone; in arch_add_memory() local
182 for_each_zone(zone) { in arch_add_memory()
183 if (zone_idx(zone) != ZONE_MOVABLE) { in arch_add_memory()
185 zone_start_pfn = zone->zone_start_pfn; in arch_add_memory()
186 zone_end_pfn = zone->zone_start_pfn + in arch_add_memory()
187 zone->spanned_pages; in arch_add_memory()
197 rc = __add_pages(nid, zone, start_pfn, nr_pages); in arch_add_memory()
Dpage-states.c95 struct zone *zone; in arch_set_page_states() local
101 for_each_populated_zone(zone) { in arch_set_page_states()
102 spin_lock_irqsave(&zone->lock, flags); in arch_set_page_states()
104 list_for_each(l, &zone->free_area[order].free_list[t]) { in arch_set_page_states()
112 spin_unlock_irqrestore(&zone->lock, flags); in arch_set_page_states()
/linux-4.1.27/drivers/md/
Draid0.c81 struct strip_zone *zone; in create_strip_zones() local
166 zone = &conf->strip_zone[0]; in create_strip_zones()
218 zone->nb_dev = cnt; in create_strip_zones()
219 zone->zone_end = smallest->sectors * cnt; in create_strip_zones()
221 curr_zone_end = zone->zone_end; in create_strip_zones()
228 zone = conf->strip_zone + i; in create_strip_zones()
232 zone->dev_start = smallest->sectors; in create_strip_zones()
238 if (rdev->sectors <= zone->dev_start) { in create_strip_zones()
258 zone->nb_dev = c; in create_strip_zones()
259 sectors = (smallest->sectors - zone->dev_start) * c; in create_strip_zones()
[all …]
/linux-4.1.27/arch/ia64/mm/
Dinit.c479 unsigned long zone; member
508 args->nid, args->zone, page_to_pfn(map_start), in virtual_memmap_init()
514 memmap_init (unsigned long size, int nid, unsigned long zone, in memmap_init() argument
518 memmap_init_zone(size, nid, zone, start_pfn, MEMMAP_EARLY); in memmap_init()
527 args.zone = zone; in memmap_init()
658 struct zone *zone; in arch_add_memory() local
665 zone = pgdat->node_zones + in arch_add_memory()
667 ret = __add_pages(nid, zone, start_pfn, nr_pages); in arch_add_memory()
681 struct zone *zone; in arch_remove_memory() local
684 zone = page_zone(pfn_to_page(start_pfn)); in arch_remove_memory()
[all …]
/linux-4.1.27/Documentation/devicetree/bindings/thermal/
Dthermal.txt38 the IC. See thermal zone binding for more details
105 get assigned to trip points of the zone. The cooling devices are expected
115 Type: phandle of zone.
119 - contribution: The cooling contribution to the thermal zone of the
122 of all cooling contributions within a thermal zone.
130 * Thermal zone nodes
132 The thermal zone node is the node containing all the required info
133 for describing a thermal zone, including its cooling device bindings. The
134 thermal zone node must contain, apart from its own properties, one sub-node
135 containing trip nodes and one sub-node containing all the zone cooling maps.
[all …]
/linux-4.1.27/Documentation/power/powercap/
Dpowercap.txt19 given zone belongs to. They each contain attributes for monitoring power, as
24 parent power zone containing multiple subzones and so on to reflect the power
26 capping to a set of devices together using the parent power zone and if more
159 In addition to that, each power zone contains a name attribute, allowing the
160 part of the system represented by that zone to be identified.
167 with two different time windows to be applied to each power zone. Thus for
168 each zone there are 2 attributes representing the constraint names, 2 power
194 name (ro): Name of this power zone.
218 Common zone and control type attributes
220 enabled (rw): Enable/Disable controls at zone level or for all zones using
[all …]
/linux-4.1.27/net/ipv6/netfilter/
Dnf_defrag_ipv6_hooks.c36 u16 zone = NF_CT_DEFAULT_ZONE; in nf_ct6_defrag_user() local
40 zone = nf_ct_zone((struct nf_conn *)skb->nfct); in nf_ct6_defrag_user()
46 return IP6_DEFRAG_CONNTRACK_BRIDGE_IN + zone; in nf_ct6_defrag_user()
49 return IP6_DEFRAG_CONNTRACK_IN + zone; in nf_ct6_defrag_user()
51 return IP6_DEFRAG_CONNTRACK_OUT + zone; in nf_ct6_defrag_user()
Dnf_conntrack_proto_icmpv6.c153 u16 zone = tmpl ? nf_ct_zone(tmpl) : NF_CT_DEFAULT_ZONE; in icmpv6_error_message() local
180 h = nf_conntrack_find_get(net, zone, &intuple); in icmpv6_error_message()
/linux-4.1.27/net/netfilter/
Dnf_conntrack_core.c129 static u32 hash_conntrack_raw(const struct nf_conntrack_tuple *tuple, u16 zone) in hash_conntrack_raw() argument
138 return jhash2((u32 *)tuple, n, zone ^ nf_conntrack_hash_rnd ^ in hash_conntrack_raw()
154 u16 zone, unsigned int size) in __hash_conntrack() argument
156 return __hash_bucket(hash_conntrack_raw(tuple, zone), size); in __hash_conntrack()
159 static inline u_int32_t hash_conntrack(const struct net *net, u16 zone, in hash_conntrack() argument
162 return __hash_conntrack(tuple, zone, net->ct.htable_size); in hash_conntrack()
332 u16 zone = nf_ct_zone(ct); in nf_ct_delete_from_lists() local
340 hash = hash_conntrack(net, zone, in nf_ct_delete_from_lists()
342 reply_hash = hash_conntrack(net, zone, in nf_ct_delete_from_lists()
391 u16 zone) in nf_ct_key_equal() argument
[all …]
Dxt_connlimit.c137 u16 zone, in check_hlist() argument
151 found = nf_conntrack_find_get(net, zone, &conn->tuple); in check_hlist()
204 u8 family, u16 zone) in count_tree() argument
232 count = check_hlist(net, &rbconn->hhead, tuple, zone, &addit); in count_tree()
248 check_hlist(net, &rbconn->hhead, tuple, zone, &addit); in count_tree()
293 u_int8_t family, u16 zone) in count_them() argument
309 count = count_tree(net, root, tuple, addr, mask, family, zone); in count_them()
327 u16 zone = NF_CT_DEFAULT_ZONE; in connlimit_mt() local
332 zone = nf_ct_zone(ct); in connlimit_mt()
349 &info->mask, par->family, zone); in connlimit_mt()
Dxt_CT.c197 if (info->zone) in xt_ct_tg_check()
206 ct = nf_conntrack_alloc(par->net, info->zone, &t, &t, GFP_KERNEL); in xt_ct_tg_check()
249 .zone = info->zone, in xt_ct_tg_check_v0()
330 .zone = info->zone, in xt_ct_tg_destroy_v0()
Dnf_nat_core.c121 hash_by_src(const struct net *net, u16 zone, in hash_by_src() argument
128 tuple->dst.protonum ^ zone ^ nf_conntrack_hash_rnd); in hash_by_src()
188 find_appropriate_src(struct net *net, u16 zone, in find_appropriate_src() argument
195 unsigned int h = hash_by_src(net, zone, tuple); in find_appropriate_src()
201 if (same_src(ct, tuple) && nf_ct_zone(ct) == zone) { in find_appropriate_src()
221 find_best_ips_proto(u16 zone, struct nf_conntrack_tuple *tuple, in find_best_ips_proto() argument
261 0 : (__force u32)tuple->dst.u3.all[max] ^ zone); in find_best_ips_proto()
303 u16 zone = nf_ct_zone(ct); in get_unique_tuple() local
326 } else if (find_appropriate_src(net, zone, l3proto, l4proto, in get_unique_tuple()
336 find_best_ips_proto(zone, tuple, range, ct, maniptype); in get_unique_tuple()
Dnf_conntrack_netlink.c968 ctnetlink_parse_zone(const struct nlattr *attr, u16 *zone) in ctnetlink_parse_zone() argument
972 *zone = ntohs(nla_get_be16(attr)); in ctnetlink_parse_zone()
977 *zone = 0; in ctnetlink_parse_zone()
1061 u16 zone; in ctnetlink_del_conntrack() local
1064 err = ctnetlink_parse_zone(cda[CTA_ZONE], &zone); in ctnetlink_del_conntrack()
1081 h = nf_conntrack_find_get(net, zone, &tuple); in ctnetlink_del_conntrack()
1115 u16 zone; in ctnetlink_get_conntrack() local
1136 err = ctnetlink_parse_zone(cda[CTA_ZONE], &zone); in ctnetlink_get_conntrack()
1150 h = nf_conntrack_find_get(net, zone, &tuple); in ctnetlink_get_conntrack()
1648 ctnetlink_create_conntrack(struct net *net, u16 zone, in ctnetlink_create_conntrack() argument
[all …]
Dnf_conntrack_expect.c91 __nf_ct_expect_find(struct net *net, u16 zone, in __nf_ct_expect_find() argument
103 nf_ct_zone(i->master) == zone) in __nf_ct_expect_find()
112 nf_ct_expect_find_get(struct net *net, u16 zone, in nf_ct_expect_find_get() argument
118 i = __nf_ct_expect_find(net, zone, tuple); in nf_ct_expect_find_get()
130 nf_ct_find_expectation(struct net *net, u16 zone, in nf_ct_find_expectation() argument
143 nf_ct_zone(i->master) == zone) { in nf_ct_find_expectation()
Dnf_conntrack_pptp.c148 u16 zone = nf_ct_zone(ct); in destroy_sibling_or_exp() local
153 h = nf_conntrack_find_get(net, zone, t); in destroy_sibling_or_exp()
164 exp = nf_ct_expect_find_get(net, zone, t); in destroy_sibling_or_exp()
/linux-4.1.27/net/sched/
Dact_connmark.c73 thash = nf_conntrack_find_get(dev_net(skb->dev), ca->zone, &tuple); in tcf_connmark()
117 ci->zone = parm->zone; in tcf_connmark_init()
130 ci->zone = parm->zone; in tcf_connmark_init()
147 .zone = ci->zone, in tcf_connmark_dump()
/linux-4.1.27/net/ipv4/netfilter/
Dnf_defrag_ipv4.c46 u16 zone = NF_CT_DEFAULT_ZONE; in nf_ct_defrag_user() local
50 zone = nf_ct_zone((struct nf_conn *)skb->nfct); in nf_ct_defrag_user()
56 return IP_DEFRAG_CONNTRACK_BRIDGE_IN + zone; in nf_ct_defrag_user()
59 return IP_DEFRAG_CONNTRACK_IN + zone; in nf_ct_defrag_user()
61 return IP_DEFRAG_CONNTRACK_OUT + zone; in nf_ct_defrag_user()
Dnf_conntrack_proto_icmp.c137 u16 zone = tmpl ? nf_ct_zone(tmpl) : NF_CT_DEFAULT_ZONE; in icmp_error_message() local
163 h = nf_conntrack_find_get(net, zone, &innertuple); in icmp_error_message()
/linux-4.1.27/arch/powerpc/mm/
Dmem.c119 struct zone *zone; in arch_add_memory() local
130 zone = pgdata->node_zones + in arch_add_memory()
133 return __add_pages(nid, zone, start_pfn, nr_pages); in arch_add_memory()
141 struct zone *zone; in arch_remove_memory() local
144 zone = page_zone(pfn_to_page(start_pfn)); in arch_remove_memory()
145 ret = __remove_pages(zone, start_pfn, nr_pages); in arch_remove_memory()
242 void __init limit_zone_pfn(enum zone_type zone, unsigned long pfn_limit) in limit_zone_pfn() argument
249 for (i = zone; i >= 0; i--) { in limit_zone_pfn()
/linux-4.1.27/drivers/uwb/
Ddrp-ie.c269 void uwb_drp_ie_single_zone_to_bm(struct uwb_mas_bm *bm, u8 zone, u16 mas_bm) in uwb_drp_ie_single_zone_to_bm() argument
277 set_bit(zone * UWB_NUM_ZONES + mas, bm->bm); in uwb_drp_ie_single_zone_to_bm()
302 u8 zone; in uwb_drp_ie_to_bm() local
311 for (zone = 0; zone < UWB_NUM_ZONES; zone++) { in uwb_drp_ie_to_bm()
312 zone_mask = 1 << zone; in uwb_drp_ie_to_bm()
314 uwb_drp_ie_single_zone_to_bm(bm, zone, mas_bm); in uwb_drp_ie_to_bm()
/linux-4.1.27/drivers/usb/storage/
Dalauda.c551 unsigned int zone) in alauda_find_unused_pba() argument
553 u16 *pba_to_lba = info->pba_to_lba[zone]; in alauda_find_unused_pba()
558 return (zone << info->zoneshift) + i; in alauda_find_unused_pba()
567 static int alauda_read_map(struct us_data *us, unsigned int zone) in alauda_read_map() argument
575 unsigned int zone_base_lba = zone * uzonesize; in alauda_read_map()
576 unsigned int zone_base_pba = zone * zonesize; in alauda_read_map()
584 usb_stor_dbg(us, "Mapping blocks for zone %d\n", zone); in alauda_read_map()
675 MEDIA_INFO(us).lba_to_pba[zone] = lba_to_pba; in alauda_read_map()
676 MEDIA_INFO(us).pba_to_lba[zone] = pba_to_lba; in alauda_read_map()
691 static void alauda_ensure_map_for_zone(struct us_data *us, unsigned int zone) in alauda_ensure_map_for_zone() argument
[all …]
Dsddr55.c718 int zone = i / 1024; in sddr55_read_map() local
744 if (info->lba_to_pba[lba + zone * 1000] != NOT_ALLOCATED && in sddr55_read_map()
748 lba + zone * 1000); in sddr55_read_map()
755 info->lba_to_pba[lba + zone * 1000] = i; in sddr55_read_map()
/linux-4.1.27/Documentation/cma/
Ddebugfs.txt4 Each CMA zone represents a directory under <debugfs>/cma/, indexed by the
5 kernel's CMA index. So the first CMA zone would be:
11 - [RO] base_pfn: The base PFN (Page Frame Number) of the zone.
14 - [RO] bitmap: The bitmap of page states in the zone.
/linux-4.1.27/Documentation/networking/
Dcops.txt52 dummy -seed -phase 2 -net 2000 -addr 2000.10 -zone "1033"
53 lt0 -seed -phase 1 -net 1000 -addr 1000.50 -zone "1033"
56 eth0 -seed -phase 2 -net 3000 -addr 3000.20 -zone "1033"
57 lt0 -seed -phase 1 -net 1000 -addr 1000.50 -zone "1033"
61 lt0 -seed -phase 1 -net 1000 -addr 1000.10 -zone "LocalTalk1"
62 lt1 -seed -phase 1 -net 2000 -addr 2000.20 -zone "LocalTalk2"
63 eth0 -seed -phase 2 -net 3000 -addr 3000.30 -zone "EtherTalk"
Dltpc.txt47 dummy -seed -phase 2 -net 2000 -addr 2000.26 -zone "1033"
48 lt0 -seed -phase 1 -net 1033 -addr 1033.27 -zone "1033"
Dfib_trie.txt105 as one "zone"/hash per prefix length.
Dnetlink_mmap.txt83 the ring just contains the frames in a continuous memory zone.
/linux-4.1.27/drivers/hwmon/
Dlm85.c234 static int ZONE_TO_REG(int zone) in ZONE_TO_REG() argument
239 if (zone == lm85_zone_map[i]) in ZONE_TO_REG()
323 struct lm85_zone zone[3]; member
516 data->zone[i].range = val >> 4; in lm85_update_device()
519 data->zone[i].limit = in lm85_update_device()
521 data->zone[i].critical = in lm85_update_device()
527 data->zone[i].limit -= 64; in lm85_update_device()
528 data->zone[i].critical -= 64; in lm85_update_device()
539 data->zone[0].hyst = i >> 4; in lm85_update_device()
540 data->zone[1].hyst = i & 0x0f; in lm85_update_device()
[all …]
Ddme1737.c1562 static SENSOR_DEVICE_ATTR_2(zone##ix##_auto_channels_temp, S_IRUGO, \
1564 static SENSOR_DEVICE_ATTR_2(zone##ix##_auto_point1_temp_hyst, S_IRUGO, \
1566 static SENSOR_DEVICE_ATTR_2(zone##ix##_auto_point1_temp, S_IRUGO, \
1568 static SENSOR_DEVICE_ATTR_2(zone##ix##_auto_point2_temp, S_IRUGO, \
1570 static SENSOR_DEVICE_ATTR_2(zone##ix##_auto_point3_temp, S_IRUGO, \
/linux-4.1.27/arch/m68k/mm/
Dmcfmmu.c42 enum zone_type zone; in paging_init() local
78 for (zone = 0; zone < MAX_NR_ZONES; zone++) in paging_init()
79 zones_size[zone] = 0x0; in paging_init()
/linux-4.1.27/drivers/base/
Dmemory.c389 struct zone *zone; in show_valid_zones() local
399 zone = page_zone(first_page); in show_valid_zones()
401 if (zone_idx(zone) == ZONE_MOVABLE - 1) { in show_valid_zones()
403 if (end_pfn == zone_end_pfn(zone)) in show_valid_zones()
405 zone->name, (zone + 1)->name); in show_valid_zones()
408 if (zone_idx(zone) == ZONE_MOVABLE) { in show_valid_zones()
410 if (start_pfn == zone->zone_start_pfn) in show_valid_zones()
412 zone->name, (zone - 1)->name); in show_valid_zones()
415 return sprintf(buf, "%s\n", zone->name); in show_valid_zones()
/linux-4.1.27/arch/tile/mm/
Dpgtable.c45 struct zone *zone; in show_mem() local
64 for_each_zone(zone) { in show_mem()
67 if (!populated_zone(zone)) in show_mem()
70 spin_lock_irqsave(&zone->lock, flags); in show_mem()
72 int nr = zone->free_area[order].nr_free; in show_mem()
77 spin_unlock_irqrestore(&zone->lock, flags); in show_mem()
79 zone_to_nid(zone), zone->name, in show_mem()
Dinit.c696 struct zone *z; in set_non_bootmem_pages_init()
869 struct zone *zone = pgdata->node_zones + MAX_NR_ZONES-1; in arch_add_memory() local
873 return __add_pages(zone, start_pfn, nr_pages); in arch_add_memory()
/linux-4.1.27/Documentation/hwmon/
Ddme1737165 be retrieved via the zone[1-3]_auto_channels_temp attributes. Each PWM output
166 is assigned to one (or hottest of multiple) temperature zone(s) through the
171 programmed via zone[1-3]_auto_point[1-3]_temp and
172 zone[1-3]_auto_point1_temp_hyst:
178 zone[1-3]_auto_point3_temp full-speed temp (all outputs)
179 zone[1-3]_auto_point2_temp full-speed temp
180 zone[1-3]_auto_point1_temp low-speed temp
181 zone[1-3]_auto_point1_temp_hyst min-speed temp
184 to auto_point2_pwm if the temperature of the associated zone is between
193 based on the temperature of the associated thermal zone:
[all …]
Dasc762117 added PECI and a 4th thermal zone. The Andigilog aSC7611 is the
47 temperature may be mapped to any zone, which has a default assignment
50 Since there is a fan to zone assignment that allows for the "hotter" of
52 indication to the user, we have added an indicator that shows which zone
62 address. One could use our flexible zone assignment to have a zone
181 When the corresponding zone temperature reaches this value,
195 if the temperature for the zone is below the minimum.
197 for the zone is below the minimum.
Dlm85170 sensors and therefore three zones (# 1, 2 and 3). Each zone has the following
182 configured and assigned to a zone for its control value. Each PWM can be
195 PWMs controlled by zone '#'.
199 * pwm#_auto_channels - controls zone that is associated with PWM
Df71882fg93 Note that the lowest numbered temperature zone trip point corresponds to
95 vica versa. So the temperature zone trip points 1-4 (or 1-2) go from high temp
/linux-4.1.27/Documentation/sysctl/
Dvm.txt229 the fragmentation index for each order is in each zone in the system. Values
234 The kernel will not compact memory in a zone if the
288 zone. This is because that memory could then be pinned via the mlock()
316 zone's value is not necessary for following calculation.
321 Each zone has an array of protection pages like this.
324 Node 0, zone DMA
338 These protections are added to score to judge whether this zone should be used
341 In this example, if normal pages (index=2) are required to this DMA zone and
342 watermark[WMARK_HIGH] is used for watermark, the kernel judges this zone should
344 (4 + 2004 = 2008). If this protection value is 0, this zone would be used for
[all …]
/linux-4.1.27/include/net/netfilter/
Dnf_conntrack_expect.h79 __nf_ct_expect_find(struct net *net, u16 zone,
83 nf_ct_expect_find_get(struct net *net, u16 zone,
87 nf_ct_find_expectation(struct net *net, u16 zone,
Dnf_conntrack.h187 __nf_conntrack_find(struct net *net, u16 zone,
254 struct nf_conn *nf_conntrack_alloc(struct net *net, u16 zone,
Dnf_conntrack_core.h55 nf_conntrack_find_get(struct net *net, u16 zone,
/linux-4.1.27/fs/sysv/
Dballoc.c187 sysv_zone_t zone; in sysv_count_free_blocks() local
190 zone = 0; in sysv_count_free_blocks()
191 while (n && (zone = blocks[--n]) != 0) in sysv_count_free_blocks()
193 if (zone == 0) in sysv_count_free_blocks()
196 block = fs32_to_cpu(sbi, zone); in sysv_count_free_blocks()
/linux-4.1.27/fs/proc/
Dmeminfo.c36 struct zone *zone; in meminfo_proc_show() local
57 for_each_zone(zone) in meminfo_proc_show()
58 wmark_low += zone->watermark[WMARK_LOW]; in meminfo_proc_show()
/linux-4.1.27/fs/minix/
Dbitmap.c47 unsigned long bit, zone; in minix_free_block() local
53 zone = block - sbi->s_firstdatazone + 1; in minix_free_block()
54 bit = zone & ((1<<k) - 1); in minix_free_block()
55 zone >>= k; in minix_free_block()
56 if (zone >= sbi->s_zmap_blocks) { in minix_free_block()
60 bh = sbi->s_zmap[zone]; in minix_free_block()
/linux-4.1.27/include/uapi/linux/netfilter/
Dxt_CT.h14 __u16 zone; member
25 __u16 zone; member
/linux-4.1.27/arch/parisc/mm/
Dinit.c660 struct zone *zone = &pgdat->node_zones[zoneid]; in show_mem() local
661 if (!populated_zone(zone)) in show_mem()
664 total += zone->present_pages; in show_mem()
665 reserved = zone->present_pages - zone->managed_pages; in show_mem()
682 struct zone *zone; in show_mem() local
685 for_each_zone_zonelist(zone, z, zl, j) in show_mem()
686 printk("[%d/%s] ", zone_to_nid(zone), in show_mem()
687 zone->name); in show_mem()
/linux-4.1.27/fs/ntfs/
Dlcnalloc.c148 const NTFS_CLUSTER_ALLOCATION_ZONES zone, in ntfs_cluster_alloc() argument
167 zone == MFT_ZONE ? "MFT" : "DATA"); in ntfs_cluster_alloc()
174 BUG_ON(zone < FIRST_ZONE); in ntfs_cluster_alloc()
175 BUG_ON(zone > LAST_ZONE); in ntfs_cluster_alloc()
202 if (zone == DATA_ZONE) in ntfs_cluster_alloc()
213 } else if (zone == DATA_ZONE && zone_start >= vol->mft_zone_start && in ntfs_cluster_alloc()
221 } else if (zone == MFT_ZONE && (zone_start < vol->mft_zone_start || in ntfs_cluster_alloc()
232 if (zone == MFT_ZONE) { in ntfs_cluster_alloc()
696 if (zone == MFT_ZONE || mft_zone_size <= 0) { in ntfs_cluster_alloc()
Dlcnalloc.h45 const NTFS_CLUSTER_ALLOCATION_ZONES zone,
/linux-4.1.27/sound/soc/
Dsoc-jack.c159 struct snd_soc_jack_zone *zone; in snd_soc_jack_get_type() local
161 list_for_each_entry(zone, &jack->jack_zones, list) { in snd_soc_jack_get_type()
162 if (micbias_voltage >= zone->min_mv && in snd_soc_jack_get_type()
163 micbias_voltage < zone->max_mv) in snd_soc_jack_get_type()
164 return zone->jack_type; in snd_soc_jack_get_type()
/linux-4.1.27/drivers/platform/x86/
Dalienware-wmi.c174 static int parse_rgb(const char *buf, struct platform_zone *zone) in parse_rgb() argument
194 zone->colors = repackager.cp; in parse_rgb()
214 static int alienware_update_led(struct platform_zone *zone) in alienware_update_led() argument
223 wmax_args.led_mask = 1 << zone->location; in alienware_update_led()
224 wmax_args.colors = zone->colors; in alienware_update_led()
232 legacy_args.colors = zone->colors; in alienware_update_led()
241 method_id = zone->location + 1; in alienware_update_led()
/linux-4.1.27/arch/powerpc/kernel/
Ddma.c58 int zone; in dma_direct_alloc_coherent() local
66 zone = dma_pfn_limit_to_zone(pfn); in dma_direct_alloc_coherent()
67 if (zone < 0) { in dma_direct_alloc_coherent()
73 switch (zone) { in dma_direct_alloc_coherent()
/linux-4.1.27/Documentation/devicetree/bindings/power_supply/
Dcharger-manager.txt24 - cm-thermal-zone : name of external thermometer's thermal zone
53 cm-thermal-zone = "thermal_zone.1"
/linux-4.1.27/include/uapi/linux/
Dtipc.h63 static inline __u32 tipc_addr(unsigned int zone, in tipc_addr() argument
67 return (zone << 24) | (cluster << 12) | node; in tipc_addr()
/linux-4.1.27/Documentation/trace/
Devents-kmem.txt52 amounts of activity imply high activity on the zone->lock. Taking this lock
64 contention on the zone->lru_lock.
72 for order-0 pages, reduces contention on the zone->lock and reduces the
76 the zone->lock will be taken once and the per-CPU list refilled. The event
85 consecutively imply the zone->lock being taken once. Large amounts of per-CPU
/linux-4.1.27/arch/sh/mm/
Dinit.c522 struct zone *zone; in arch_remove_memory() local
525 zone = page_zone(pfn_to_page(start_pfn)); in arch_remove_memory()
526 ret = __remove_pages(zone, start_pfn, nr_pages); in arch_remove_memory()
DKconfig25 int "Maximum zone order"
35 blocks into "zones", where each zone is a power of two number of
/linux-4.1.27/include/uapi/linux/tc_act/
Dtc_connmark.h11 __u16 zone; member
/linux-4.1.27/include/net/tc_act/
Dtc_connmark.h8 u16 zone; member
/linux-4.1.27/drivers/block/
Dpktcdvd.c1092 static struct packet_data *pkt_get_packet_data(struct pktcdvd_device *pd, int zone) in pkt_get_packet_data() argument
1097 if (pkt->sector == zone || pkt->list.next == &pd->cdrw.pkt_free_list) { in pkt_get_packet_data()
1099 if (pkt->sector != zone) in pkt_get_packet_data()
1198 sector_t zone = 0; /* Suppress gcc warning */ in pkt_handle_queue() local
1223 zone = get_zone(bio->bi_iter.bi_sector, pd); in pkt_handle_queue()
1225 if (p->sector == zone) { in pkt_handle_queue()
1247 pkt = pkt_get_packet_data(pd, zone); in pkt_handle_queue()
1249 pd->current_sector = zone + pd->settings.size; in pkt_handle_queue()
1250 pkt->sector = zone; in pkt_handle_queue()
1259 pkt_dbg(2, pd, "looking for zone %llx\n", (unsigned long long)zone); in pkt_handle_queue()
[all …]
/linux-4.1.27/drivers/net/wireless/rsi/
Drsi_91x_main.c43 void rsi_dbg(u32 zone, const char *fmt, ...) in rsi_dbg() argument
53 if (zone & rsi_zone_enabled) in rsi_dbg()
Drsi_main.h43 extern __printf(2, 3) void rsi_dbg(u32 zone, const char *fmt, ...);
/linux-4.1.27/kernel/sched/
Dwait.c489 const struct zone *zone = page_zone(virt_to_page(word)); in bit_waitqueue() local
492 return &zone->wait_table[hash_long(val, zone->wait_table_bits)]; in bit_waitqueue()
/linux-4.1.27/arch/arm/boot/dts/
Dexynos5440-trip-points.dtsi2 * Device tree sources for default Exynos5440 thermal zone definition
Dexynos5420-trip-points.dtsi2 * Device tree sources for default Exynos5420 thermal zone definition
Dexynos4-cpu-thermal.dtsi2 * Device tree sources for Exynos4 thermal zone
/linux-4.1.27/arch/tile/include/asm/
Dhomecache.h26 struct zone;
/linux-4.1.27/drivers/firmware/efi/libstub/
DMakefile10 -fPIC -fno-strict-aliasing -mno-red-zone \
/linux-4.1.27/arch/metag/mm/
DKconfig52 int "Maximum zone order"
57 blocks into "zones", where each zone is a power of two number of
/linux-4.1.27/arch/c6x/
DKconfig88 int "Maximum zone order"
92 blocks into "zones", where each zone is a power of two number of
/linux-4.1.27/drivers/memstick/core/
Dms_block.c174 int zone = msb_get_zone_from_pba(pba); in msb_mark_block_used() local
188 msb->free_block_count[zone]--; in msb_mark_block_used()
194 int zone = msb_get_zone_from_pba(pba); in msb_mark_block_unused() local
207 msb->free_block_count[zone]++; in msb_mark_block_unused()
1078 static u16 msb_get_free_block(struct msb_data *msb, int zone) in msb_get_free_block() argument
1081 int pba = zone * MS_BLOCKS_IN_ZONE; in msb_get_free_block()
1086 if (!msb->free_block_count[zone]) { in msb_get_free_block()
1087 …O free blocks in the zone %d, to use for a write, (media is WORN out) switching to RO mode", zone); in msb_get_free_block()
1092 pos %= msb->free_block_count[zone]; in msb_get_free_block()
1095 msb->free_block_count[zone], pos); in msb_get_free_block()
[all …]
/linux-4.1.27/Documentation/cris/
DREADME84 zone(0): 2048 pages.
85 zone(1): 0 pages.
86 zone(2): 0 pages.
/linux-4.1.27/include/sound/
Demux_synth.h183 struct snd_sf_zone *zone; /* Zone assigned to this note */ member
/linux-4.1.27/arch/nios2/
DKconfig60 int "Maximum zone order"
65 blocks into "zones", where each zone is a power of two number of
/linux-4.1.27/fs/xfs/libxfs/
Dxfs_da_btree.h25 struct zone;
/linux-4.1.27/Documentation/filesystems/
Dsysv-fs.txt36 * Size of a block or zone (data allocation unit on disk)
196 Notation: We often speak of a "block" but mean a zone (the allocation unit)
Dproc.txt743 Node 0, zone DMA 0 4 5 4 4 3 ...
744 Node 0, zone Normal 1 0 0 1 101 8 ...
745 Node 0, zone HighMem 2 0 0 1 1 0 ...
765 Node 0, zone DMA, type Unmovable 0 0 0 1 1 1 1 1…
766 Node 0, zone DMA, type Reclaimable 0 0 0 0 0 0 0 0…
767 Node 0, zone DMA, type Movable 1 1 2 1 2 1 1 0…
768 Node 0, zone DMA, type Reserve 0 0 0 0 0 0 0 0…
769 Node 0, zone DMA, type Isolate 0 0 0 0 0 0 0 0…
770 Node 0, zone DMA32, type Unmovable 103 54 77 1 1 1 11 8…
771 Node 0, zone DMA32, type Reclaimable 0 0 2 1 0 0 0 0…
[all …]
Df2fs.txt276 -z [int] : Set the number of sections per zone.
321 to 2MB in size. A section is composed of consecutive segments, and a zone
322 consists of a set of sections. By default, section and zone sizes are set to one
328 align with the zone size <-|
373 start block address of Main area with the zone size by reserving some segments
Dntfs.txt192 mft_zone_multiplier= Set the MFT zone multiplier for the volume (this
196 default. The MFT zone multiplier determines how much
198 other space is used up, then the MFT zone will be
205 Value MFT zone size (% of volume size)
Dvfat.txt118 Linux. This is useful when time zone set in sys_tz is
119 not the time zone used by the filesystem. Note that this
/linux-4.1.27/arch/x86/boot/compressed/
DMakefile50 $(obj)/eboot.o: KBUILD_CFLAGS += -fshort-wchar -mno-red-zone
/linux-4.1.27/Documentation/devicetree/bindings/arm/bcm/
Dbrcm,brcmstb.txt62 o offset to the base CPU power zone register
/linux-4.1.27/arch/powerpc/include/asm/
Dpgtable.h218 void limit_zone_pfn(enum zone_type zone, unsigned long max_pfn);
/linux-4.1.27/drivers/misc/
Dlkdtm.c207 struct zone *zone, in jp_shrink_inactive_list() argument
/linux-4.1.27/Documentation/cgroups/
Dmemcg_test.txt110 VM's control (means that it's handled under global zone->lru_lock).
112 list management functions under zone->lru_lock().
Dmemory.txt270 zone->lru_lock
273 per-zone-per-cgroup LRU (cgroup's private LRU) is just guarded by
274 zone->lru_lock, it has no lock of its own.
/linux-4.1.27/kernel/
Dkexec.c1965 VMCOREINFO_STRUCT_SIZE(zone); in crash_save_vmcoreinfo_init()
1983 VMCOREINFO_OFFSET(zone, free_area); in crash_save_vmcoreinfo_init()
1984 VMCOREINFO_OFFSET(zone, vm_stat); in crash_save_vmcoreinfo_init()
1985 VMCOREINFO_OFFSET(zone, spanned_pages); in crash_save_vmcoreinfo_init()
1991 VMCOREINFO_LENGTH(zone.free_area, MAX_ORDER); in crash_save_vmcoreinfo_init()
Dfork.c217 struct zone *zone = page_zone(virt_to_page(ti)); in account_kernel_stack() local
219 mod_zone_page_state(zone, NR_KERNEL_STACK, account); in account_kernel_stack()
/linux-4.1.27/drivers/staging/unisys/Documentation/
Dproc-entries.txt49 clientstring and zone.
/linux-4.1.27/arch/x86/
DMakefile98 KBUILD_CFLAGS += -mno-red-zone
/linux-4.1.27/Documentation/locking/
Dlockstat.txt174 …&(&zone->lru_lock)->rlock: 93000 94934 0.16 59.18 188253.…
/linux-4.1.27/Documentation/acpi/
Dnamespace.txt161 +-| _TZ | Scope(_TZ): the thermal zone namespace
177 | +-| TZ00 | ThermalZone(TZ00); the FAN thermal zone
/linux-4.1.27/arch/ia64/include/asm/
Dpgtable.h547 extern void memmap_init (unsigned long size, int nid, unsigned long zone,
/linux-4.1.27/fs/
Dbuffer.c259 struct zone *zone; in free_more_memory() local
268 &zone); in free_more_memory()
269 if (zone) in free_more_memory()
/linux-4.1.27/Documentation/input/
Diforce-protocol.txt202 Idx 00 Set dead zone (0..2048)
Dsentelic.txt759 bit7 0 RW on-pad tap zone enable
/linux-4.1.27/Documentation/
Dmemory-hotplug.txt176 The first column shows it's default zone.
308 create ZONE_MOVABLE...a zone which is just used for movable pages.
Drtc.txt8 the local time zone or daylight savings time -- unless they dual boot
Dkernel-parameters.txt1241 highmem=nn[KMG] [KNL,BOOT] forces the highmem zone to have an exact
1638 of Movable pages. The Movable zone is used for the
1641 HugeTLB pages may not be allocated from this zone.
1643 use the HighMem zone if it exists, and the Normal
1644 zone if it does not.
2540 one of ['zone', 'node', 'default'] can be specified
3577 Set to disable actions on ACPI thermal zone
3589 Specify global default ACPI thermal zone polling rate
/linux-4.1.27/fs/f2fs/
Dsegment.h250 unsigned int zone; /* current zone number */ member
Dsegment.c864 if (CURSEG_I(sbi, i)->zone == zoneno) in get_new_segment()
892 curseg->zone = GET_ZONENO_FROM_SEGNO(sbi, curseg->segno); in reset_curseg()
/linux-4.1.27/
DREPORTING-BUGS162 around the world, and they may not work in your time zone. Unless you
DCREDITS2657 W: http://zone.pspt.fi/~jsn/
/linux-4.1.27/fs/affs/
DChanges192 zone number was zero, the loop didn't terminate,
/linux-4.1.27/arch/powerpc/
DKconfig584 int "Maximum zone order"
599 blocks into "zones", where each zone is a power of two number of
/linux-4.1.27/arch/sh/kernel/cpu/sh5/
Dentry.S684 bge SP, r6, tr0 ! ? below 1k from bottom of stack : danger zone
/linux-4.1.27/arch/mips/
DKconfig2036 int "Maximum zone order"
2047 blocks into "zones", where each zone is a power of two number of
/linux-4.1.27/arch/arm/
DKconfig1688 int "Maximum zone order" if ARCH_SHMOBILE_LEGACY
1695 blocks into "zones", where each zone is a power of two number of
/linux-4.1.27/Documentation/virtual/uml/
DUserModeLinux-HOWTO.txt3696 zone = 0x100c5160}