Home
last modified time | relevance | path

Searched refs:zone (Results 1 – 200 of 207) sorted by relevance

12

/linux-4.4.14/drivers/gpu/drm/ttm/
Dttm_memory.c76 struct ttm_mem_zone *zone = in ttm_mem_zone_kobj_release() local
80 zone->name, (unsigned long long)zone->used_mem >> 10); in ttm_mem_zone_kobj_release()
81 kfree(zone); in ttm_mem_zone_kobj_release()
88 struct ttm_mem_zone *zone = in ttm_mem_zone_show() local
92 spin_lock(&zone->glob->lock); in ttm_mem_zone_show()
94 val = zone->zone_mem; in ttm_mem_zone_show()
96 val = zone->emer_mem; in ttm_mem_zone_show()
98 val = zone->max_mem; in ttm_mem_zone_show()
100 val = zone->swap_limit; in ttm_mem_zone_show()
102 val = zone->used_mem; in ttm_mem_zone_show()
[all …]
/linux-4.4.14/mm/
Dcompaction.c92 unsigned long end_pfn, struct zone *zone) in pageblock_pfn_to_page() argument
105 if (page_zone(start_page) != zone) in pageblock_pfn_to_page()
127 void defer_compaction(struct zone *zone, int order) in defer_compaction() argument
129 zone->compact_considered = 0; in defer_compaction()
130 zone->compact_defer_shift++; in defer_compaction()
132 if (order < zone->compact_order_failed) in defer_compaction()
133 zone->compact_order_failed = order; in defer_compaction()
135 if (zone->compact_defer_shift > COMPACT_MAX_DEFER_SHIFT) in defer_compaction()
136 zone->compact_defer_shift = COMPACT_MAX_DEFER_SHIFT; in defer_compaction()
138 trace_mm_compaction_defer_compaction(zone, order); in defer_compaction()
[all …]
Dmemory_hotplug.c250 struct zone *zone; in register_page_bootmem_info_node() local
258 zone = &pgdat->node_zones[0]; in register_page_bootmem_info_node()
259 for (; zone < pgdat->node_zones + MAX_NR_ZONES - 1; zone++) { in register_page_bootmem_info_node()
260 if (zone_is_initialized(zone)) { in register_page_bootmem_info_node()
261 nr_pages = zone->wait_table_hash_nr_entries in register_page_bootmem_info_node()
264 page = virt_to_page(zone->wait_table); in register_page_bootmem_info_node()
288 static void __meminit grow_zone_span(struct zone *zone, unsigned long start_pfn, in grow_zone_span() argument
293 zone_span_writelock(zone); in grow_zone_span()
295 old_zone_end_pfn = zone_end_pfn(zone); in grow_zone_span()
296 if (zone_is_empty(zone) || start_pfn < zone->zone_start_pfn) in grow_zone_span()
[all …]
Dvmstat.c94 int calculate_pressure_threshold(struct zone *zone) in calculate_pressure_threshold() argument
107 watermark_distance = low_wmark_pages(zone) - min_wmark_pages(zone); in calculate_pressure_threshold()
118 int calculate_normal_threshold(struct zone *zone) in calculate_normal_threshold() argument
153 mem = zone->managed_pages >> (27 - PAGE_SHIFT); in calculate_normal_threshold()
170 struct zone *zone; in refresh_zone_stat_thresholds() local
174 for_each_populated_zone(zone) { in refresh_zone_stat_thresholds()
177 threshold = calculate_normal_threshold(zone); in refresh_zone_stat_thresholds()
180 per_cpu_ptr(zone->pageset, cpu)->stat_threshold in refresh_zone_stat_thresholds()
188 tolerate_drift = low_wmark_pages(zone) - min_wmark_pages(zone); in refresh_zone_stat_thresholds()
191 zone->percpu_drift_mark = high_wmark_pages(zone) + in refresh_zone_stat_thresholds()
[all …]
Dpage_alloc.c349 static int page_outside_zone_boundaries(struct zone *zone, struct page *page) in page_outside_zone_boundaries() argument
357 seq = zone_span_seqbegin(zone); in page_outside_zone_boundaries()
358 start_pfn = zone->zone_start_pfn; in page_outside_zone_boundaries()
359 sp = zone->spanned_pages; in page_outside_zone_boundaries()
360 if (!zone_spans_pfn(zone, pfn)) in page_outside_zone_boundaries()
362 } while (zone_span_seqretry(zone, seq)); in page_outside_zone_boundaries()
366 pfn, zone_to_nid(zone), zone->name, in page_outside_zone_boundaries()
372 static int page_is_consistent(struct zone *zone, struct page *page) in page_is_consistent() argument
376 if (zone != page_zone(page)) in page_is_consistent()
384 static int bad_range(struct zone *zone, struct page *page) in bad_range() argument
[all …]
Dvmscan.c195 static unsigned long zone_reclaimable_pages(struct zone *zone) in zone_reclaimable_pages() argument
199 nr = zone_page_state(zone, NR_ACTIVE_FILE) + in zone_reclaimable_pages()
200 zone_page_state(zone, NR_INACTIVE_FILE); in zone_reclaimable_pages()
203 nr += zone_page_state(zone, NR_ACTIVE_ANON) + in zone_reclaimable_pages()
204 zone_page_state(zone, NR_INACTIVE_ANON); in zone_reclaimable_pages()
209 bool zone_reclaimable(struct zone *zone) in zone_reclaimable() argument
211 return zone_page_state(zone, NR_PAGES_SCANNED) < in zone_reclaimable()
212 zone_reclaimable_pages(zone) * 6; in zone_reclaimable()
881 struct zone *zone, in shrink_page_list() argument
919 VM_BUG_ON_PAGE(page_zone(page) != zone, page); in shrink_page_list()
[all …]
Dpage_isolation.c15 struct zone *zone; in set_migratetype_isolate() local
21 zone = page_zone(page); in set_migratetype_isolate()
23 spin_lock_irqsave(&zone->lock, flags); in set_migratetype_isolate()
49 if (!has_unmovable_pages(zone, page, arg.pages_found, in set_migratetype_isolate()
64 zone->nr_isolate_pageblock++; in set_migratetype_isolate()
65 nr_pages = move_freepages_block(zone, page, MIGRATE_ISOLATE); in set_migratetype_isolate()
67 __mod_zone_freepage_state(zone, -nr_pages, migratetype); in set_migratetype_isolate()
70 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate()
72 drain_all_pages(zone); in set_migratetype_isolate()
78 struct zone *zone; in unset_migratetype_isolate() local
[all …]
Dmmzone.c29 struct zone *next_zone(struct zone *zone) in next_zone() argument
31 pg_data_t *pgdat = zone->zone_pgdat; in next_zone()
33 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone()
34 zone++; in next_zone()
38 zone = pgdat->node_zones; in next_zone()
40 zone = NULL; in next_zone()
42 return zone; in next_zone()
68 (z->zone && !zref_in_nodemask(z, nodes))) in next_zones_zonelist()
76 struct page *page, struct zone *zone) in memmap_valid_within() argument
81 if (page_zone(page) != zone) in memmap_valid_within()
Dworkingset.c155 static void *pack_shadow(unsigned long eviction, struct zone *zone) in pack_shadow() argument
157 eviction = (eviction << NODES_SHIFT) | zone_to_nid(zone); in pack_shadow()
158 eviction = (eviction << ZONES_SHIFT) | zone_idx(zone); in pack_shadow()
165 struct zone **zone, in unpack_shadow() argument
181 *zone = NODE_DATA(nid)->node_zones + zid; in unpack_shadow()
183 refault = atomic_long_read(&(*zone)->inactive_age); in unpack_shadow()
215 struct zone *zone = page_zone(page); in workingset_eviction() local
218 eviction = atomic_long_inc_return(&zone->inactive_age); in workingset_eviction()
219 return pack_shadow(eviction, zone); in workingset_eviction()
234 struct zone *zone; in workingset_refault() local
[all …]
Dswap.c56 struct zone *zone = page_zone(page); in __page_cache_release() local
60 spin_lock_irqsave(&zone->lru_lock, flags); in __page_cache_release()
61 lruvec = mem_cgroup_page_lruvec(page, zone); in __page_cache_release()
65 spin_unlock_irqrestore(&zone->lru_lock, flags); in __page_cache_release()
423 struct zone *zone = NULL; in pagevec_lru_move_fn() local
429 struct zone *pagezone = page_zone(page); in pagevec_lru_move_fn()
431 if (pagezone != zone) { in pagevec_lru_move_fn()
432 if (zone) in pagevec_lru_move_fn()
433 spin_unlock_irqrestore(&zone->lru_lock, flags); in pagevec_lru_move_fn()
434 zone = pagezone; in pagevec_lru_move_fn()
[all …]
Dpage_owner.c214 static void init_pages_in_zone(pg_data_t *pgdat, struct zone *zone) in init_pages_in_zone() argument
218 unsigned long pfn = zone->zone_start_pfn, block_end_pfn; in init_pages_in_zone()
219 unsigned long end_pfn = pfn + zone->spanned_pages; in init_pages_in_zone()
223 pfn = zone->zone_start_pfn; in init_pages_in_zone()
272 pgdat->node_id, zone->name, count); in init_pages_in_zone()
277 struct zone *zone; in init_zones_in_node() local
278 struct zone *node_zones = pgdat->node_zones; in init_zones_in_node()
281 for (zone = node_zones; zone - node_zones < MAX_NR_ZONES; ++zone) { in init_zones_in_node()
282 if (!populated_zone(zone)) in init_zones_in_node()
285 spin_lock_irqsave(&zone->lock, flags); in init_zones_in_node()
[all …]
Dmm_init.c34 struct zone *zone; in mminit_verify_zonelist() local
46 zone = &pgdat->node_zones[zoneid]; in mminit_verify_zonelist()
47 if (!populated_zone(zone)) in mminit_verify_zonelist()
53 zone->name); in mminit_verify_zonelist()
56 for_each_zone_zonelist(zone, z, zonelist, zoneid) { in mminit_verify_zonelist()
59 zone->node, zone->name); in mminit_verify_zonelist()
61 printk(KERN_CONT "0:%s ", zone->name); in mminit_verify_zonelist()
Dmlock.c176 struct zone *zone = page_zone(page); in munlock_vma_page() local
186 spin_lock_irq(&zone->lru_lock); in munlock_vma_page()
192 __mod_zone_page_state(zone, NR_MLOCK, -nr_pages); in munlock_vma_page()
195 spin_unlock_irq(&zone->lru_lock); in munlock_vma_page()
202 spin_unlock_irq(&zone->lru_lock); in munlock_vma_page()
276 static void __munlock_pagevec(struct pagevec *pvec, struct zone *zone) in __munlock_pagevec() argument
287 spin_lock_irq(&zone->lru_lock); in __munlock_pagevec()
312 __mod_zone_page_state(zone, NR_MLOCK, delta_munlocked); in __munlock_pagevec()
313 spin_unlock_irq(&zone->lru_lock); in __munlock_pagevec()
432 struct zone *zone; in munlock_vma_pages_range() local
[all …]
Dinternal.h120 extern bool zone_reclaimable(struct zone *zone);
147 struct zone *preferred_zone;
212 struct zone *zone; member
432 unsigned long reclaim_clean_pages_from_list(struct zone *zone,
Dpage_idle.c33 struct zone *zone; in page_idle_get_page() local
43 zone = page_zone(page); in page_idle_get_page()
44 spin_lock_irq(&zone->lru_lock); in page_idle_get_page()
49 spin_unlock_irq(&zone->lru_lock); in page_idle_get_page()
Dmemcontrol.c405 mem_cgroup_zone_zoneinfo(struct mem_cgroup *memcg, struct zone *zone) in mem_cgroup_zone_zoneinfo() argument
407 int nid = zone_to_nid(zone); in mem_cgroup_zone_zoneinfo()
408 int zid = zone_idx(zone); in mem_cgroup_zone_zoneinfo()
901 mz = mem_cgroup_zone_zoneinfo(root, reclaim->zone); in mem_cgroup_iter()
1051 struct lruvec *mem_cgroup_zone_lruvec(struct zone *zone, in mem_cgroup_zone_lruvec() argument
1058 lruvec = &zone->lruvec; in mem_cgroup_zone_lruvec()
1062 mz = mem_cgroup_zone_zoneinfo(memcg, zone); in mem_cgroup_zone_lruvec()
1070 if (unlikely(lruvec->zone != zone)) in mem_cgroup_zone_lruvec()
1071 lruvec->zone = zone; in mem_cgroup_zone_lruvec()
1084 struct lruvec *mem_cgroup_page_lruvec(struct page *page, struct zone *zone) in mem_cgroup_page_lruvec() argument
[all …]
Dcma.c102 struct zone *zone; in cma_activate_area() local
110 zone = page_zone(pfn_to_page(pfn)); in cma_activate_area()
124 if (page_zone(pfn_to_page(pfn)) != zone) in cma_activate_area()
Dpage-writeback.c276 static unsigned long zone_dirtyable_memory(struct zone *zone) in zone_dirtyable_memory() argument
280 nr_pages = zone_page_state(zone, NR_FREE_PAGES); in zone_dirtyable_memory()
281 nr_pages -= min(nr_pages, zone->dirty_balance_reserve); in zone_dirtyable_memory()
283 nr_pages += zone_page_state(zone, NR_INACTIVE_FILE); in zone_dirtyable_memory()
284 nr_pages += zone_page_state(zone, NR_ACTIVE_FILE); in zone_dirtyable_memory()
296 struct zone *z = &NODE_DATA(node)->node_zones[ZONE_HIGHMEM]; in highmem_dirtyable_memory()
437 static unsigned long zone_dirty_limit(struct zone *zone) in zone_dirty_limit() argument
439 unsigned long zone_memory = zone_dirtyable_memory(zone); in zone_dirty_limit()
462 bool zone_dirty_ok(struct zone *zone) in zone_dirty_ok() argument
464 unsigned long limit = zone_dirty_limit(zone); in zone_dirty_ok()
[all …]
Dsparse.c693 int __meminit sparse_add_one_section(struct zone *zone, unsigned long start_pfn) in sparse_add_one_section() argument
696 struct pglist_data *pgdat = zone->zone_pgdat; in sparse_add_one_section()
791 void sparse_remove_one_section(struct zone *zone, struct mem_section *ms) in sparse_remove_one_section() argument
795 struct pglist_data *pgdat = zone->zone_pgdat; in sparse_remove_one_section()
Doom_kill.c211 struct zone *zone; in constrained_alloc() local
244 for_each_zone_zonelist_nodemask(zone, z, oc->zonelist, in constrained_alloc()
246 if (!cpuset_zone_allowed(zone, oc->gfp_mask)) in constrained_alloc()
Dquicklist.c30 struct zone *zones = NODE_DATA(node)->node_zones; in max_pages()
Dmigrate.c317 struct zone *oldzone, *newzone; in migrate_page_move_mapping()
1557 struct zone *zone = pgdat->node_zones + z; in migrate_balanced_pgdat() local
1559 if (!populated_zone(zone)) in migrate_balanced_pgdat()
1562 if (!zone_reclaimable(zone)) in migrate_balanced_pgdat()
1566 if (!zone_watermark_ok(zone, 0, in migrate_balanced_pgdat()
1567 high_wmark_pages(zone) + in migrate_balanced_pgdat()
Dmempolicy.c1611 static int apply_policy_zone(struct mempolicy *policy, enum zone_type zone) in apply_policy_zone() argument
1628 return zone >= dynamic_policy_zone; in apply_policy_zone()
1719 struct zone *zone; in mempolicy_slab_node() local
1724 &zone); in mempolicy_slab_node()
1725 return zone ? zone->node : node; in mempolicy_slab_node()
2250 struct zone *zone; in mpol_misplaced() local
2293 &pol->v.nodes, &zone); in mpol_misplaced()
2294 polnid = zone->node; in mpol_misplaced()
Dhuge_memory.c112 struct zone *zone; in set_recommended_min_free_kbytes() local
116 for_each_populated_zone(zone) in set_recommended_min_free_kbytes()
1708 struct zone *zone = page_zone(page); in __split_huge_page_refcount() local
1713 spin_lock_irq(&zone->lru_lock); in __split_huge_page_refcount()
1714 lruvec = mem_cgroup_page_lruvec(page, zone); in __split_huge_page_refcount()
1797 __mod_zone_page_state(zone, NR_ANON_TRANSPARENT_HUGEPAGES, -1); in __split_huge_page_refcount()
1801 spin_unlock_irq(&zone->lru_lock); in __split_huge_page_refcount()
Dbacking-dev.c968 long wait_iff_congested(struct zone *zone, int sync, long timeout) in wait_iff_congested() argument
981 !test_bit(ZONE_CONGESTED, &zone->flags)) { in wait_iff_congested()
Dnobootmem.c163 struct zone *z; in reset_node_managed_pages()
Dhugetlb.c884 struct zone *zone; in dequeue_huge_page_vma() local
906 for_each_zone_zonelist_nodemask(zone, z, zonelist, in dequeue_huge_page_vma()
908 if (cpuset_zone_allowed(zone, htlb_alloc_mask(h))) { in dequeue_huge_page_vma()
909 page = dequeue_huge_page_node(h, zone_to_nid(zone)); in dequeue_huge_page_vma()
1059 static bool zone_spans_last_pfn(const struct zone *zone, in zone_spans_last_pfn() argument
1063 return zone_spans_pfn(zone, last_pfn); in zone_spans_last_pfn()
1070 struct zone *z; in alloc_gigantic_page()
Dfilemap.c740 const struct zone *zone = page_zone(page); in page_waitqueue() local
742 return &zone->wait_table[hash_ptr(page, zone->wait_table_bits)]; in page_waitqueue()
Dslab.c3007 struct zone *zone; in fallback_alloc() local
3027 for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { in fallback_alloc()
3028 nid = zone_to_nid(zone); in fallback_alloc()
3030 if (cpuset_zone_allowed(zone, flags) && in fallback_alloc()
Dbootmem.c250 struct zone *z; in reset_node_managed_pages()
Dslub.c1714 struct zone *zone; in get_any_partial() local
1744 for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { in get_any_partial()
1747 n = get_node(s, zone_to_nid(zone)); in get_any_partial()
1749 if (n && cpuset_zone_allowed(zone, flags) && in get_any_partial()
/linux-4.4.14/include/linux/
Dvmstat.h104 #define __count_zone_vm_events(item, zone, delta) \ argument
106 zone_idx(zone), delta)
113 static inline void zone_page_state_add(long x, struct zone *zone, in zone_page_state_add() argument
116 atomic_long_add(x, &zone->vm_stat[item]); in zone_page_state_add()
130 static inline unsigned long zone_page_state(struct zone *zone, in zone_page_state() argument
133 long x = atomic_long_read(&zone->vm_stat[item]); in zone_page_state()
147 static inline unsigned long zone_page_state_snapshot(struct zone *zone, in zone_page_state_snapshot() argument
150 long x = atomic_long_read(&zone->vm_stat[item]); in zone_page_state_snapshot()
155 x += per_cpu_ptr(zone->pageset, cpu)->vm_stat_diff[item]; in zone_page_state_snapshot()
166 extern void zone_statistics(struct zone *, struct zone *, gfp_t gfp);
[all …]
Dmemory_hotplug.h10 struct zone;
63 static inline unsigned zone_span_seqbegin(struct zone *zone) in zone_span_seqbegin() argument
65 return read_seqbegin(&zone->span_seqlock); in zone_span_seqbegin()
67 static inline int zone_span_seqretry(struct zone *zone, unsigned iv) in zone_span_seqretry() argument
69 return read_seqretry(&zone->span_seqlock, iv); in zone_span_seqretry()
71 static inline void zone_span_writelock(struct zone *zone) in zone_span_writelock() argument
73 write_seqlock(&zone->span_seqlock); in zone_span_writelock()
75 static inline void zone_span_writeunlock(struct zone *zone) in zone_span_writeunlock() argument
77 write_sequnlock(&zone->span_seqlock); in zone_span_writeunlock()
79 static inline void zone_seqlock_init(struct zone *zone) in zone_seqlock_init() argument
[all …]
Dmmzone.h220 struct zone *zone; member
331 struct zone { struct
549 static inline unsigned long zone_end_pfn(const struct zone *zone) in zone_end_pfn() argument
551 return zone->zone_start_pfn + zone->spanned_pages; in zone_end_pfn()
554 static inline bool zone_spans_pfn(const struct zone *zone, unsigned long pfn) in zone_spans_pfn() argument
556 return zone->zone_start_pfn <= pfn && pfn < zone_end_pfn(zone); in zone_spans_pfn()
559 static inline bool zone_is_initialized(struct zone *zone) in zone_is_initialized() argument
561 return !!zone->wait_table; in zone_is_initialized()
564 static inline bool zone_is_empty(struct zone *zone) in zone_is_empty() argument
566 return zone->spanned_pages == 0; in zone_is_empty()
[all …]
Dcompaction.h40 extern int fragmentation_index(struct zone *zone, unsigned int order);
46 extern unsigned long compaction_suitable(struct zone *zone, int order,
49 extern void defer_compaction(struct zone *zone, int order);
50 extern bool compaction_deferred(struct zone *zone, int order);
51 extern void compaction_defer_reset(struct zone *zone, int order,
53 extern bool compaction_restarting(struct zone *zone, int order);
72 static inline unsigned long compaction_suitable(struct zone *zone, int order, in compaction_suitable() argument
78 static inline void defer_compaction(struct zone *zone, int order) in defer_compaction() argument
82 static inline bool compaction_deferred(struct zone *zone, int order) in compaction_deferred() argument
Dpage-isolation.h5 static inline bool has_isolate_pageblock(struct zone *zone) in has_isolate_pageblock() argument
7 return zone->nr_isolate_pageblock; in has_isolate_pageblock()
18 static inline bool has_isolate_pageblock(struct zone *zone) in has_isolate_pageblock() argument
32 bool has_unmovable_pages(struct zone *zone, struct page *page, int count,
35 int move_freepages_block(struct zone *zone, struct page *page,
37 int move_freepages(struct zone *zone,
Dmemcontrol.h56 struct zone *zone; member
306 struct lruvec *mem_cgroup_zone_lruvec(struct zone *, struct mem_cgroup *);
307 struct lruvec *mem_cgroup_page_lruvec(struct page *, struct zone *);
465 unsigned long mem_cgroup_soft_limit_reclaim(struct zone *zone, int order,
545 static inline struct lruvec *mem_cgroup_zone_lruvec(struct zone *zone, in mem_cgroup_zone_lruvec() argument
548 return &zone->lruvec; in mem_cgroup_zone_lruvec()
552 struct zone *zone) in mem_cgroup_page_lruvec() argument
554 return &zone->lruvec; in mem_cgroup_page_lruvec()
657 unsigned long mem_cgroup_soft_limit_reclaim(struct zone *zone, int order, in mem_cgroup_soft_limit_reclaim() argument
Dgfp.h506 void drain_zone_pages(struct zone *zone, struct per_cpu_pages *pcp);
507 void drain_all_pages(struct zone *zone);
508 void drain_local_pages(struct zone *zone);
Dswap.h118 struct zone;
328 struct zone *zone,
339 extern int zone_reclaim(struct zone *, gfp_t, unsigned int);
342 static inline int zone_reclaim(struct zone *z, gfp_t mask, unsigned int order) in zone_reclaim()
Dcpuset.h58 static inline int cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed()
180 static inline int cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed()
Dmm.h709 static inline int zone_to_nid(struct zone *zone) in zone_to_nid() argument
712 return zone->node; in zone_to_nid()
840 static inline struct zone *page_zone(const struct page *page) in page_zone()
858 static inline void set_page_zone(struct page *page, enum zone_type zone) in set_page_zone() argument
861 page->flags |= (zone & ZONES_MASK) << ZONES_PGSHIFT; in set_page_zone()
870 static inline void set_page_links(struct page *page, enum zone_type zone, in set_page_links() argument
873 set_page_zone(page, zone); in set_page_links()
1820 extern void zone_pcp_update(struct zone *zone);
1821 extern void zone_pcp_reset(struct zone *zone);
Dsuspend.h298 extern void mark_free_pages(struct zone *zone);
Dwriteback.h323 bool zone_dirty_ok(struct zone *zone);
Dbacking-dev.h200 long wait_iff_congested(struct zone *zone, int sync, long timeout);
/linux-4.4.14/include/net/netfilter/
Dnf_conntrack_zones.h21 nf_ct_zone_init(struct nf_conntrack_zone *zone, u16 id, u8 dir, u8 flags) in nf_ct_zone_init() argument
23 zone->id = id; in nf_ct_zone_init()
24 zone->flags = flags; in nf_ct_zone_init()
25 zone->dir = dir; in nf_ct_zone_init()
27 return zone; in nf_ct_zone_init()
34 const struct nf_conntrack_zone *zone; in nf_ct_zone_tmpl() local
39 zone = nf_ct_zone(tmpl); in nf_ct_zone_tmpl()
40 if (zone->flags & NF_CT_FLAG_MARK) in nf_ct_zone_tmpl()
41 zone = nf_ct_zone_init(tmp, skb->mark, zone->dir, 0); in nf_ct_zone_tmpl()
43 return zone; in nf_ct_zone_tmpl()
[all …]
Dnf_conntrack_expect.h82 const struct nf_conntrack_zone *zone,
87 const struct nf_conntrack_zone *zone,
92 const struct nf_conntrack_zone *zone,
Dnf_conntrack.h255 const struct nf_conntrack_zone *zone,
296 const struct nf_conntrack_zone *zone,
Dnf_conntrack_core.h57 const struct nf_conntrack_zone *zone,
/linux-4.4.14/fs/btrfs/
Dreada.c253 struct reada_zone *zone; in reada_find_zone() local
259 zone = NULL; in reada_find_zone()
261 ret = radix_tree_gang_lookup(&dev->reada_zones, (void **)&zone, in reada_find_zone()
264 kref_get(&zone->refcnt); in reada_find_zone()
268 if (logical >= zone->start && logical < zone->end) in reada_find_zone()
269 return zone; in reada_find_zone()
271 kref_put(&zone->refcnt, reada_zone_release); in reada_find_zone()
283 zone = kzalloc(sizeof(*zone), GFP_NOFS); in reada_find_zone()
284 if (!zone) in reada_find_zone()
287 zone->start = start; in reada_find_zone()
[all …]
/linux-4.4.14/kernel/power/
Dsnapshot.c299 struct mem_zone_bm_rtree *zone; member
357 static int add_rtree_block(struct mem_zone_bm_rtree *zone, gfp_t gfp_mask, in add_rtree_block() argument
364 block_nr = zone->blocks; in add_rtree_block()
374 for (i = zone->levels; i < levels_needed; i++) { in add_rtree_block()
376 &zone->nodes); in add_rtree_block()
380 node->data[0] = (unsigned long)zone->rtree; in add_rtree_block()
381 zone->rtree = node; in add_rtree_block()
382 zone->levels += 1; in add_rtree_block()
386 block = alloc_rtree_node(gfp_mask, safe_needed, ca, &zone->leaves); in add_rtree_block()
391 node = zone->rtree; in add_rtree_block()
[all …]
Dpower.h134 extern unsigned int snapshot_additional_pages(struct zone *zone);
/linux-4.4.14/include/trace/events/
Dcompaction.h252 TP_PROTO(struct zone *zone,
256 TP_ARGS(zone, order, ret),
266 __entry->nid = zone_to_nid(zone);
267 __entry->idx = zone_idx(zone);
281 TP_PROTO(struct zone *zone,
285 TP_ARGS(zone, order, ret)
290 TP_PROTO(struct zone *zone,
294 TP_ARGS(zone, order, ret)
300 TP_PROTO(struct zone *zone, int order),
302 TP_ARGS(zone, order),
[all …]
/linux-4.4.14/drivers/net/ethernet/mellanox/mlx4/
Dalloc.c251 struct mlx4_zone_entry *zone = kmalloc(sizeof(*zone), GFP_KERNEL); in mlx4_zone_add_one() local
253 if (NULL == zone) in mlx4_zone_add_one()
256 zone->flags = flags; in mlx4_zone_add_one()
257 zone->bitmap = bitmap; in mlx4_zone_add_one()
258 zone->use_rr = (flags & MLX4_ZONE_USE_RR) ? MLX4_USE_RR : 0; in mlx4_zone_add_one()
259 zone->priority = priority; in mlx4_zone_add_one()
260 zone->offset = offset; in mlx4_zone_add_one()
264 zone->uid = zone_alloc->last_uid++; in mlx4_zone_add_one()
265 zone->allocator = zone_alloc; in mlx4_zone_add_one()
275 list_add_tail(&zone->prio_list, &it->prio_list); in mlx4_zone_add_one()
[all …]
/linux-4.4.14/drivers/mtd/
Dsm_ftl.c195 static loff_t sm_mkoffset(struct sm_ftl *ftl, int zone, int block, int boffset) in sm_mkoffset() argument
198 WARN_ON(zone < 0 || zone >= ftl->zone_count); in sm_mkoffset()
205 return (zone * SM_MAX_ZONE_SIZE + block) * ftl->block_size + boffset; in sm_mkoffset()
210 int *zone, int *block, int *boffset) in sm_break_offset() argument
214 *zone = offset >= ftl->zone_count ? -1 : offset; in sm_break_offset()
237 int zone, int block, int boffset, in sm_read_sector() argument
267 if (zone == 0 && block == ftl->cis_block && boffset == in sm_read_sector()
278 ret = mtd_read_oob(mtd, sm_mkoffset(ftl, zone, block, boffset), &ops); in sm_read_sector()
283 block, zone, ret); in sm_read_sector()
301 " as bad" , block, zone); in sm_read_sector()
[all …]
/linux-4.4.14/drivers/thermal/int340x_thermal/
Dint340x_thermal_zone.c22 static int int340x_thermal_get_zone_temp(struct thermal_zone_device *zone, in int340x_thermal_get_zone_temp() argument
25 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_zone_temp()
30 return d->override_ops->get_temp(zone, temp); in int340x_thermal_get_zone_temp()
51 static int int340x_thermal_get_trip_temp(struct thermal_zone_device *zone, in int340x_thermal_get_trip_temp() argument
54 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_trip_temp()
58 return d->override_ops->get_trip_temp(zone, trip, temp); in int340x_thermal_get_trip_temp()
83 static int int340x_thermal_get_trip_type(struct thermal_zone_device *zone, in int340x_thermal_get_trip_type() argument
87 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_trip_type()
91 return d->override_ops->get_trip_type(zone, trip, type); in int340x_thermal_get_trip_type()
116 static int int340x_thermal_set_trip_temp(struct thermal_zone_device *zone, in int340x_thermal_set_trip_temp() argument
[all …]
Dint340x_thermal_zone.h40 struct thermal_zone_device *zone; member
65 thermal_zone_device_update(tzone->zone); in int340x_thermal_zone_device_update()
Dprocessor_thermal_device.c179 static int proc_thermal_get_zone_temp(struct thermal_zone_device *zone, in proc_thermal_get_zone_temp() argument
/linux-4.4.14/fs/adfs/
Dmap.c192 scan_map(struct adfs_sb_info *asb, unsigned int zone, in scan_map() argument
199 dm = asb->s_map + zone; in scan_map()
200 zone = asb->s_map_size; in scan_map()
201 dm_end = asb->s_map + zone; in scan_map()
212 } while (--zone > 0); in scan_map()
235 unsigned int zone; in adfs_map_free() local
238 zone = asb->s_map_size; in adfs_map_free()
242 } while (--zone > 0); in adfs_map_free()
252 unsigned int zone, mapoff; in adfs_map_lookup() local
260 zone = asb->s_map_size >> 1; in adfs_map_lookup()
[all …]
Dsuper.c305 int i, zone; in adfs_read_map() local
322 for (zone = 0; zone < nzones; zone++, map_addr++) { in adfs_read_map()
323 dm[zone].dm_startbit = 0; in adfs_read_map()
324 dm[zone].dm_endbit = zone_size; in adfs_read_map()
325 dm[zone].dm_startblk = zone * zone_size - ADFS_DR_SIZE_BITS; in adfs_read_map()
326 dm[zone].dm_bh = sb_bread(sb, map_addr); in adfs_read_map()
328 if (!dm[zone].dm_bh) { in adfs_read_map()
335 i = zone - 1; in adfs_read_map()
348 while (--zone >= 0) in adfs_read_map()
349 brelse(dm[zone].dm_bh); in adfs_read_map()
/linux-4.4.14/lib/
Dshow_mem.c26 struct zone *zone = &pgdat->node_zones[zoneid]; in show_mem() local
27 if (!populated_zone(zone)) in show_mem()
30 total += zone->present_pages; in show_mem()
31 reserved += zone->present_pages - zone->managed_pages; in show_mem()
34 highmem += zone->present_pages; in show_mem()
/linux-4.4.14/Documentation/thermal/
Dsysfs-api.txt13 The generic thermal sysfs provides a set of interfaces for thermal zone
17 This how-to focuses on enabling new thermal zone and cooling devices to
19 This solution is platform independent and any type of thermal zone devices
22 The main task of the thermal sysfs driver is to expose thermal zone attributes
25 inputs from thermal zone attributes (the current temperature and trip point
33 1.1 thermal zone device interface
40 This interface function adds a new thermal zone device (sensor) to
44 type: the thermal zone type.
45 trips: the total number of trip points this thermal zone supports.
48 ops: thermal zone device call-backs.
[all …]
Dx86_pkg_temperature_thermal19 zone with maximum two user mode configurable trip points. Number of trip points
27 Each package will register as a thermal zone under /sys/class/thermal.
42 Thermal notifications: To get kobject-uevent notifications, set the thermal zone
Dpower_allocator.txt11 point of the thermal zone.
16 thermal zone.
63 provided while registering the thermal zone. This estimates the
76 thermal-zone. For example:
85 Instead, if the thermal zone is registered from the platform code, pass a
184 @tz: thermal zone in which we are currently operating
196 @tz: thermal zone in which we are currently operating
231 If the thermal zone is registered using
233 are passed as part of the thermal zone's `thermal_bind_parameters`.
Dexynos_thermal_emulation18 /sys/devices/virtual/thermal/thermal_zone'zone id'/emul_temp.
/linux-4.4.14/Documentation/ABI/testing/
Dsysfs-class-powercap25 This status affects every power zone using this "control_type.
27 What: /sys/class/powercap/<control type>/<power zone>
32 A power zone is a single or a collection of devices, which can
33 be independently monitored and controlled. A power zone sysfs
37 What: /sys/class/powercap/<control type>/<power zone>/<child power zone>
45 power zone for a whole CPU package, each CPU core in it can
46 be a child power zone.
48 What: /sys/class/powercap/.../<power zone>/name
53 Specifies the name of this power zone.
55 What: /sys/class/powercap/.../<power zone>/energy_uj
[all …]
Dsysfs-class-backlight-driver-adp88701 What: /sys/class/backlight/<backlight>/<ambient light zone>_max
11 Control the maximum brightness for <ambient light zone>
14 <ambient light zone>.
16 What: /sys/class/backlight/<backlight>/<ambient light zone>_dim
25 Control the dim brightness for <ambient light zone>
29 this <ambient light zone>.
46 Get/Set current ambient light zone. Reading returns
49 to enter the corresponding ambient light zone.
Dsysfs-bus-iio-light-lm3533-als7 in each direction (rising|falling) and a zone change occurs.
8 The corresponding light zone can be read from
32 These thresholds correspond to the eight zone-boundary
41 Get the current light zone (0..4) as defined by the
50 out_currentY_currentZ_raw, where Z is the current zone.
57 Set the output current for channel out_currentY when in zone
Dsysfs-devices-memory69 read-only and is designed to show which zone this memory
/linux-4.4.14/drivers/hwmon/
Dscpi-hwmon.c51 struct scpi_thermal_zone *zone = dev; in scpi_read_temp() local
52 struct scpi_sensors *scpi_sensors = zone->scpi_sensors; in scpi_read_temp()
54 struct sensor_data *sensor = &scpi_sensors->data[zone->sensor_id]; in scpi_read_temp()
102 struct scpi_thermal_zone *zone; in unregister_thermal_zones() local
104 zone = list_entry(pos, struct scpi_thermal_zone, list); in unregister_thermal_zones()
105 thermal_zone_of_sensor_unregister(&pdev->dev, zone->tzd); in unregister_thermal_zones()
227 struct scpi_thermal_zone *zone; in scpi_hwmon_probe() local
232 zone = devm_kzalloc(dev, sizeof(*zone), GFP_KERNEL); in scpi_hwmon_probe()
233 if (!zone) { in scpi_hwmon_probe()
238 zone->sensor_id = i; in scpi_hwmon_probe()
[all …]
Dlm85.c234 static int ZONE_TO_REG(int zone) in ZONE_TO_REG() argument
239 if (zone == lm85_zone_map[i]) in ZONE_TO_REG()
323 struct lm85_zone zone[3]; member
516 data->zone[i].range = val >> 4; in lm85_update_device()
519 data->zone[i].limit = in lm85_update_device()
521 data->zone[i].critical = in lm85_update_device()
527 data->zone[i].limit -= 64; in lm85_update_device()
528 data->zone[i].critical -= 64; in lm85_update_device()
539 data->zone[0].hyst = i >> 4; in lm85_update_device()
540 data->zone[1].hyst = i & 0x0f; in lm85_update_device()
[all …]
Ddme1737.c1562 static SENSOR_DEVICE_ATTR_2(zone##ix##_auto_channels_temp, S_IRUGO, \
1564 static SENSOR_DEVICE_ATTR_2(zone##ix##_auto_point1_temp_hyst, S_IRUGO, \
1566 static SENSOR_DEVICE_ATTR_2(zone##ix##_auto_point1_temp, S_IRUGO, \
1568 static SENSOR_DEVICE_ATTR_2(zone##ix##_auto_point2_temp, S_IRUGO, \
1570 static SENSOR_DEVICE_ATTR_2(zone##ix##_auto_point3_temp, S_IRUGO, \
/linux-4.4.14/drivers/staging/rts5208/
Dxd.c879 xd_card->zone = vmalloc(size); in xd_init_l2p_tbl()
880 if (!xd_card->zone) { in xd_init_l2p_tbl()
886 xd_card->zone[i].build_flag = 0; in xd_init_l2p_tbl()
887 xd_card->zone[i].l2p_table = NULL; in xd_init_l2p_tbl()
888 xd_card->zone[i].free_table = NULL; in xd_init_l2p_tbl()
889 xd_card->zone[i].get_index = 0; in xd_init_l2p_tbl()
890 xd_card->zone[i].set_index = 0; in xd_init_l2p_tbl()
891 xd_card->zone[i].unused_blk_cnt = 0; in xd_init_l2p_tbl()
897 static inline void free_zone(struct zone_entry *zone) in free_zone() argument
899 if (!zone) in free_zone()
[all …]
Drtsx_chip.h529 struct zone_entry *zone; member
/linux-4.4.14/sound/synth/emux/
Dsoundfont.c525 struct snd_sf_zone *zone; in load_info() local
563 for (zone = sf->zones; zone; zone = zone->next) { in load_info()
564 if (!zone->mapped && in load_info()
565 zone->bank == hdr.bank && in load_info()
566 zone->instr == hdr.instr) in load_info()
595 if ((zone = sf_zone_new(sflist, sf)) == NULL) { in load_info()
600 zone->bank = tmpzone.bank; in load_info()
601 zone->instr = tmpzone.instr; in load_info()
602 zone->v = tmpzone.v; in load_info()
605 zone->sample = set_sample(sf, &zone->v); in load_info()
[all …]
Demux_synth.c115 vp->zone = table[i]; in snd_emux_note_on()
116 if (vp->zone->sample) in snd_emux_note_on()
117 vp->block = vp->zone->sample->block; in snd_emux_note_on()
513 vp->zone = NULL; in terminate_voice()
566 vp->reg = vp->zone->v; in setup_voice()
Demux_effect.c215 origp = (unsigned char*)&vp->zone->v.parm + offset; in snd_emux_send_effect()
/linux-4.4.14/virt/kvm/
Dcoalesced_mmio.c35 if (addr < dev->zone.addr) in coalesced_mmio_in_range()
37 if (addr + len > dev->zone.addr + dev->zone.size) in coalesced_mmio_in_range()
137 struct kvm_coalesced_mmio_zone *zone) in kvm_vm_ioctl_register_coalesced_mmio() argument
148 dev->zone = *zone; in kvm_vm_ioctl_register_coalesced_mmio()
151 ret = kvm_io_bus_register_dev(kvm, KVM_MMIO_BUS, zone->addr, in kvm_vm_ioctl_register_coalesced_mmio()
152 zone->size, &dev->dev); in kvm_vm_ioctl_register_coalesced_mmio()
168 struct kvm_coalesced_mmio_zone *zone) in kvm_vm_ioctl_unregister_coalesced_mmio() argument
175 if (coalesced_mmio_in_range(dev, zone->addr, zone->size)) { in kvm_vm_ioctl_unregister_coalesced_mmio()
Dcoalesced_mmio.h21 struct kvm_coalesced_mmio_zone zone; member
27 struct kvm_coalesced_mmio_zone *zone);
29 struct kvm_coalesced_mmio_zone *zone);
Dkvm_main.c2783 struct kvm_coalesced_mmio_zone zone; in kvm_vm_ioctl() local
2786 if (copy_from_user(&zone, argp, sizeof(zone))) in kvm_vm_ioctl()
2788 r = kvm_vm_ioctl_register_coalesced_mmio(kvm, &zone); in kvm_vm_ioctl()
2792 struct kvm_coalesced_mmio_zone zone; in kvm_vm_ioctl() local
2795 if (copy_from_user(&zone, argp, sizeof(zone))) in kvm_vm_ioctl()
2797 r = kvm_vm_ioctl_unregister_coalesced_mmio(kvm, &zone); in kvm_vm_ioctl()
/linux-4.4.14/arch/x86/mm/
Dhighmem_32.c109 struct zone *zone; in set_highmem_pages_init() local
117 for_each_zone(zone) { in set_highmem_pages_init()
120 if (!is_highmem(zone)) in set_highmem_pages_init()
123 zone_start_pfn = zone->zone_start_pfn; in set_highmem_pages_init()
124 zone_end_pfn = zone_start_pfn + zone->spanned_pages; in set_highmem_pages_init()
126 nid = zone_to_nid(zone); in set_highmem_pages_init()
128 zone->name, nid, zone_start_pfn, zone_end_pfn); in set_highmem_pages_init()
Dinit_32.c829 struct zone *zone = pgdata->node_zones + in arch_add_memory() local
834 return __add_pages(nid, zone, start_pfn, nr_pages); in arch_add_memory()
842 struct zone *zone; in arch_remove_memory() local
844 zone = page_zone(pfn_to_page(start_pfn)); in arch_remove_memory()
845 return __remove_pages(zone, start_pfn, nr_pages); in arch_remove_memory()
Dinit_64.c693 struct zone *zone = pgdat->node_zones + in arch_add_memory() local
701 ret = __add_pages(nid, zone, start_pfn, nr_pages); in arch_add_memory()
1021 struct zone *zone; in arch_remove_memory() local
1024 zone = page_zone(pfn_to_page(start_pfn)); in arch_remove_memory()
1026 ret = __remove_pages(zone, start_pfn, nr_pages); in arch_remove_memory()
/linux-4.4.14/net/sched/
Dact_connmark.c40 struct nf_conntrack_zone zone; in tcf_connmark() local
74 zone.id = ca->zone; in tcf_connmark()
75 zone.dir = NF_CT_DEFAULT_ZONE_DIR; in tcf_connmark()
77 thash = nf_conntrack_find_get(ca->net, &zone, &tuple); in tcf_connmark()
123 ci->zone = parm->zone; in tcf_connmark_init()
136 ci->zone = parm->zone; in tcf_connmark_init()
153 .zone = ci->zone, in tcf_connmark_dump()
/linux-4.4.14/fs/xfs/
Dkmem.h105 kmem_zone_free(kmem_zone_t *zone, void *ptr) in kmem_zone_free() argument
107 kmem_cache_free(zone, ptr); in kmem_zone_free()
111 kmem_zone_destroy(kmem_zone_t *zone) in kmem_zone_destroy() argument
113 if (zone) in kmem_zone_destroy()
114 kmem_cache_destroy(zone); in kmem_zone_destroy()
120 kmem_zone_zalloc(kmem_zone_t *zone, xfs_km_flags_t flags) in kmem_zone_zalloc() argument
122 return kmem_zone_alloc(zone, flags | KM_ZERO); in kmem_zone_zalloc()
Dkmem.c112 kmem_zone_alloc(kmem_zone_t *zone, xfs_km_flags_t flags) in kmem_zone_alloc() argument
119 ptr = kmem_cache_alloc(zone, lflags); in kmem_zone_alloc()
/linux-4.4.14/Documentation/vm/
Dbalance17 is, only when needed (aka zone free memory is 0), instead of making it
26 of incurring the overhead of regular zone balancing.
31 would not be done even when the dma zone was completely empty. 2.2 has
36 In 2.3, zone balancing can be done in one of two ways: depending on the
37 zone size (and possibly of the size of lower class zones), we can decide
39 zone. The good part is, while balancing, we do not need to look at sizes
46 of a zone _and_ all its lower class zones falls below 1/64th of the
47 total memory in the zone and its lower class zones. This fixes the 2.2
54 Note that if the size of the regular zone is huge compared to dma zone,
56 deciding whether to balance the regular zone. The first solution
[all …]
Dnuma71 each memory zone [one or more of DMA, DMA32, NORMAL, HIGH_MEMORY, MOVABLE],
73 selected zone/node cannot satisfy the allocation request. This situation,
74 when a zone has no available memory to satisfy a request, is called
79 fall back to the same zone type on a different node, or to a different zone
82 a default zonelist order based on the sizes of the various zone types relative
93 nodes' zones in the selected zonelist looking for the first zone in the list
120 node the "local memory node"--the node of the first zone in CPU's node's
Dunevictable-lru.txt60 main memory will have over 32 million 4k pages in a single zone. When a large
82 The Unevictable LRU infrastructure consists of an additional, per-zone, LRU list
111 The unevictable list benefits from the "arrayification" of the per-zone LRU
115 unevictable pages are placed directly on the page's zone's unevictable list
116 under the zone lru_lock. This allows us to prevent the stranding of pages on
128 The memory controller data structure automatically gets a per-zone unevictable
129 list as a result of the "arrayification" of the per-zone LRU lists (one per
216 unevictable list for the zone being scanned.
349 is now mlocked and divert the page to the zone's unevictable list. If
411 mlocked, munlock_vma_page() updates that zone statistics for the number of
[all …]
/linux-4.4.14/drivers/iio/light/
Dlm3533-als.c61 atomic_t zone; member
90 static int _lm3533_als_get_zone(struct iio_dev *indio_dev, u8 *zone) in _lm3533_als_get_zone() argument
103 *zone = min_t(u8, val, LM3533_ALS_ZONE_MAX); in _lm3533_als_get_zone()
108 static int lm3533_als_get_zone(struct iio_dev *indio_dev, u8 *zone) in lm3533_als_get_zone() argument
114 *zone = atomic_read(&als->zone); in lm3533_als_get_zone()
116 ret = _lm3533_als_get_zone(indio_dev, zone); in lm3533_als_get_zone()
128 static inline u8 lm3533_als_get_target_reg(unsigned channel, unsigned zone) in lm3533_als_get_target_reg() argument
130 return LM3533_REG_ALS_TARGET_BASE + 5 * channel + zone; in lm3533_als_get_target_reg()
134 unsigned zone, u8 *val) in lm3533_als_get_target() argument
143 if (zone > LM3533_ALS_ZONE_MAX) in lm3533_als_get_target()
[all …]
DKconfig208 light zone through sysfs. A threshold event can be generated on zone
209 changes. The ALS-control output values can be set per zone for the
/linux-4.4.14/drivers/thermal/
Drcar_thermal.c61 struct thermal_zone_device *zone; member
73 #define rcar_zone_to_priv(zone) ((zone)->devdata) argument
203 static int rcar_thermal_get_temp(struct thermal_zone_device *zone, int *temp) in rcar_thermal_get_temp() argument
205 struct rcar_thermal_priv *priv = rcar_zone_to_priv(zone); in rcar_thermal_get_temp()
217 static int rcar_thermal_get_trip_type(struct thermal_zone_device *zone, in rcar_thermal_get_trip_type() argument
220 struct rcar_thermal_priv *priv = rcar_zone_to_priv(zone); in rcar_thermal_get_trip_type()
236 static int rcar_thermal_get_trip_temp(struct thermal_zone_device *zone, in rcar_thermal_get_trip_temp() argument
239 struct rcar_thermal_priv *priv = rcar_zone_to_priv(zone); in rcar_thermal_get_trip_temp()
255 static int rcar_thermal_notify(struct thermal_zone_device *zone, in rcar_thermal_notify() argument
258 struct rcar_thermal_priv *priv = rcar_zone_to_priv(zone); in rcar_thermal_notify()
[all …]
Dtegra_soctherm.c311 struct tegra_thermctl_zone *zone = data; in tegra_thermctl_get_temp() local
314 val = (readl(zone->reg) >> zone->shift) & SENSOR_TEMP_MASK; in tegra_thermctl_get_temp()
412 struct tegra_thermctl_zone *zone = in tegra_soctherm_probe() local
413 devm_kzalloc(&pdev->dev, sizeof(*zone), GFP_KERNEL); in tegra_soctherm_probe()
414 if (!zone) { in tegra_soctherm_probe()
419 zone->reg = tegra->regs + t124_thermctl_temp_zones[i].offset; in tegra_soctherm_probe()
420 zone->shift = t124_thermctl_temp_zones[i].shift; in tegra_soctherm_probe()
422 tz = thermal_zone_of_sensor_register(&pdev->dev, i, zone, in tegra_soctherm_probe()
Dof-thermal.c400 thermal_zone_of_add_sensor(struct device_node *zone, in thermal_zone_of_add_sensor() argument
407 tzd = thermal_zone_get_zone_by_name(zone->name); in thermal_zone_of_add_sensor()
885 struct thermal_zone_device *zone; in of_parse_thermal_zones() local
925 zone = thermal_zone_device_register(child->name, tz->ntrips, in of_parse_thermal_zones()
930 if (IS_ERR(zone)) { in of_parse_thermal_zones()
932 PTR_ERR(zone)); in of_parse_thermal_zones()
972 struct thermal_zone_device *zone; in of_thermal_destroy_zones() local
978 zone = thermal_zone_get_zone_by_name(child->name); in of_thermal_destroy_zones()
979 if (IS_ERR(zone)) in of_thermal_destroy_zones()
982 thermal_zone_device_unregister(zone); in of_thermal_destroy_zones()
[all …]
DKconfig10 zone and cooling device.
11 Each thermal zone contains its own temperature, trip points,
51 thermal zone but the choice here takes precedence.
75 devices based on their 'contribution' to a zone. The
169 Enable this option to make a emul_temp sysfs node in thermal zone
245 management framework. A thermal zone with several trip points will be
247 thermal zone if trip points reached.
274 bound to thermal zone trip points. When a trip point reached, the
296 thermal zone. Each package will have its own thermal zone. There are
319 thermal zone. There are two trip points. One of the trip point can
[all …]
/linux-4.4.14/arch/s390/mm/
Dpage-states.c95 struct zone *zone; in arch_set_page_states() local
101 for_each_populated_zone(zone) { in arch_set_page_states()
102 spin_lock_irqsave(&zone->lock, flags); in arch_set_page_states()
104 list_for_each(l, &zone->free_area[order].free_list[t]) { in arch_set_page_states()
112 spin_unlock_irqrestore(&zone->lock, flags); in arch_set_page_states()
/linux-4.4.14/net/openvswitch/
Dconntrack.c48 struct nf_conntrack_zone zone; member
133 const struct nf_conntrack_zone *zone, in __ovs_ct_update_key() argument
137 key->ct.zone = zone->id; in __ovs_ct_update_key()
149 const struct nf_conntrack_zone *zone = &nf_ct_zone_dflt; in ovs_ct_update_key() local
161 zone = nf_ct_zone(ct); in ovs_ct_update_key()
165 zone = &info->zone; in ovs_ct_update_key()
167 __ovs_ct_update_key(key, state, zone, ct); in ovs_ct_update_key()
181 nla_put_u16(skb, OVS_KEY_ATTR_CT_ZONE, key->ct.zone)) in ovs_ct_put_key()
305 u16 zone, struct sk_buff *skb) in handle_fragments() argument
310 enum ip_defrag_users user = IP_DEFRAG_CONNTRACK_IN + zone; in handle_fragments()
[all …]
Dconntrack.h78 key->ct.zone = 0; in ovs_ct_fill_key()
Dflow.h117 u16 zone; member
/linux-4.4.14/drivers/md/
Draid0.c81 struct strip_zone *zone; in create_strip_zones() local
166 zone = &conf->strip_zone[0]; in create_strip_zones()
215 zone->nb_dev = cnt; in create_strip_zones()
216 zone->zone_end = smallest->sectors * cnt; in create_strip_zones()
218 curr_zone_end = zone->zone_end; in create_strip_zones()
225 zone = conf->strip_zone + i; in create_strip_zones()
229 zone->dev_start = smallest->sectors; in create_strip_zones()
235 if (rdev->sectors <= zone->dev_start) { in create_strip_zones()
255 zone->nb_dev = c; in create_strip_zones()
256 sectors = (smallest->sectors - zone->dev_start) * c; in create_strip_zones()
[all …]
/linux-4.4.14/Documentation/devicetree/bindings/thermal/
Dthermal.txt38 the IC. See thermal zone binding for more details
106 get assigned to trip points of the zone. The cooling devices are expected
116 Type: phandle of zone.
120 - contribution: The cooling contribution to the thermal zone of the
123 of all cooling contributions within a thermal zone.
131 * Thermal zone nodes
133 The thermal zone node is the node containing all the required info
134 for describing a thermal zone, including its cooling device bindings. The
135 thermal zone node must contain, apart from its own properties, one sub-node
136 containing trip nodes and one sub-node containing all the zone cooling maps.
[all …]
/linux-4.4.14/arch/ia64/mm/
Dinit.c472 unsigned long zone; member
501 args->nid, args->zone, page_to_pfn(map_start), in virtual_memmap_init()
507 memmap_init (unsigned long size, int nid, unsigned long zone, in memmap_init() argument
511 memmap_init_zone(size, nid, zone, start_pfn, MEMMAP_EARLY); in memmap_init()
520 args.zone = zone; in memmap_init()
651 struct zone *zone; in arch_add_memory() local
658 zone = pgdat->node_zones + in arch_add_memory()
660 ret = __add_pages(nid, zone, start_pfn, nr_pages); in arch_add_memory()
674 struct zone *zone; in arch_remove_memory() local
677 zone = page_zone(pfn_to_page(start_pfn)); in arch_remove_memory()
[all …]
/linux-4.4.14/Documentation/power/powercap/
Dpowercap.txt19 given zone belongs to. They each contain attributes for monitoring power, as
24 parent power zone containing multiple subzones and so on to reflect the power
26 capping to a set of devices together using the parent power zone and if more
159 In addition to that, each power zone contains a name attribute, allowing the
160 part of the system represented by that zone to be identified.
167 with two different time windows to be applied to each power zone. Thus for
168 each zone there are 2 attributes representing the constraint names, 2 power
194 name (ro): Name of this power zone.
218 Common zone and control type attributes
220 enabled (rw): Enable/Disable controls at zone level or for all zones using
[all …]
/linux-4.4.14/net/netfilter/
Dxt_CT.c203 struct nf_conntrack_zone zone; in xt_ct_tg_check() local
213 if (info->zone || info->flags & (XT_CT_ZONE_DIR_ORIG | in xt_ct_tg_check()
223 memset(&zone, 0, sizeof(zone)); in xt_ct_tg_check()
224 zone.id = info->zone; in xt_ct_tg_check()
225 zone.dir = xt_ct_flags_to_dir(info); in xt_ct_tg_check()
227 zone.flags |= NF_CT_FLAG_MARK; in xt_ct_tg_check()
229 ct = nf_ct_tmpl_alloc(par->net, &zone, GFP_KERNEL); in xt_ct_tg_check()
273 .zone = info->zone, in xt_ct_tg_check_v0()
356 .zone = info->zone, in xt_ct_tg_destroy_v0()
Dnf_conntrack_netlink.c132 const struct nf_conntrack_zone *zone, int dir) in ctnetlink_dump_zone_id() argument
134 if (zone->id == NF_CT_DEFAULT_ZONE_ID || zone->dir != dir) in ctnetlink_dump_zone_id()
136 if (nla_put_be16(skb, attrtype, htons(zone->id))) in ctnetlink_dump_zone_id()
475 const struct nf_conntrack_zone *zone; in ctnetlink_fill_info() local
491 zone = nf_ct_zone(ct); in ctnetlink_fill_info()
498 if (ctnetlink_dump_zone_id(skb, CTA_TUPLE_ZONE, zone, in ctnetlink_fill_info()
508 if (ctnetlink_dump_zone_id(skb, CTA_TUPLE_ZONE, zone, in ctnetlink_fill_info()
513 if (ctnetlink_dump_zone_id(skb, CTA_ZONE, zone, in ctnetlink_fill_info()
635 const struct nf_conntrack_zone *zone; in ctnetlink_conntrack_event() local
682 zone = nf_ct_zone(ct); in ctnetlink_conntrack_event()
[all …]
Dnf_conntrack_core.c294 const struct nf_conntrack_zone *zone, in nf_ct_tmpl_alloc() argument
306 if (nf_ct_zone_add(tmpl, flags, zone) < 0) in nf_ct_tmpl_alloc()
430 const struct nf_conntrack_zone *zone) in nf_ct_key_equal() argument
438 nf_ct_zone_equal(ct, zone, NF_CT_DIRECTION(h)) && in nf_ct_key_equal()
448 ____nf_conntrack_find(struct net *net, const struct nf_conntrack_zone *zone, in ____nf_conntrack_find() argument
461 if (nf_ct_key_equal(h, tuple, zone)) { in ____nf_conntrack_find()
484 __nf_conntrack_find_get(struct net *net, const struct nf_conntrack_zone *zone, in __nf_conntrack_find_get() argument
492 h = ____nf_conntrack_find(net, zone, tuple, hash); in __nf_conntrack_find_get()
499 if (unlikely(!nf_ct_key_equal(h, tuple, zone))) { in __nf_conntrack_find_get()
511 nf_conntrack_find_get(struct net *net, const struct nf_conntrack_zone *zone, in nf_conntrack_find_get() argument
[all …]
Dxt_connlimit.c137 const struct nf_conntrack_zone *zone, in check_hlist() argument
151 found = nf_conntrack_find_get(net, zone, &conn->tuple); in check_hlist()
204 u8 family, const struct nf_conntrack_zone *zone) in count_tree() argument
232 count = check_hlist(net, &rbconn->hhead, tuple, zone, &addit); in count_tree()
248 check_hlist(net, &rbconn->hhead, tuple, zone, &addit); in count_tree()
294 const struct nf_conntrack_zone *zone) in count_them() argument
310 count = count_tree(net, root, tuple, addr, mask, family, zone); in count_them()
325 const struct nf_conntrack_zone *zone = &nf_ct_zone_dflt; in connlimit_mt() local
333 zone = nf_ct_zone(ct); in connlimit_mt()
350 &info->mask, par->family, zone); in connlimit_mt()
Dnf_conntrack_standalone.c147 const struct nf_conntrack_zone *zone = nf_ct_zone(ct); in ct_show_zone() local
149 if (zone->dir != dir) in ct_show_zone()
151 switch (zone->dir) { in ct_show_zone()
153 seq_printf(s, "zone=%u ", zone->id); in ct_show_zone()
156 seq_printf(s, "zone-orig=%u ", zone->id); in ct_show_zone()
159 seq_printf(s, "zone-reply=%u ", zone->id); in ct_show_zone()
Dnf_nat_core.c188 const struct nf_conntrack_zone *zone, in find_appropriate_src() argument
202 nf_ct_zone_equal(ct, zone, IP_CT_DIR_ORIGINAL)) { in find_appropriate_src()
222 find_best_ips_proto(const struct nf_conntrack_zone *zone, in find_best_ips_proto() argument
263 0 : (__force u32)tuple->dst.u3.all[max] ^ zone->id); in find_best_ips_proto()
302 const struct nf_conntrack_zone *zone; in get_unique_tuple() local
307 zone = nf_ct_zone(ct); in get_unique_tuple()
330 } else if (find_appropriate_src(net, zone, l3proto, l4proto, in get_unique_tuple()
340 find_best_ips_proto(zone, tuple, range, ct, maniptype); in get_unique_tuple()
Dnf_conntrack_expect.c92 const struct nf_conntrack_zone *zone, in __nf_ct_expect_find() argument
104 nf_ct_zone_equal_any(i->master, zone)) in __nf_ct_expect_find()
114 const struct nf_conntrack_zone *zone, in nf_ct_expect_find_get() argument
120 i = __nf_ct_expect_find(net, zone, tuple); in nf_ct_expect_find_get()
133 const struct nf_conntrack_zone *zone, in nf_ct_find_expectation() argument
146 nf_ct_zone_equal_any(i->master, zone)) { in nf_ct_find_expectation()
Dnf_conntrack_pptp.c146 const struct nf_conntrack_zone *zone; in destroy_sibling_or_exp() local
153 zone = nf_ct_zone(ct); in destroy_sibling_or_exp()
154 h = nf_conntrack_find_get(net, zone, t); in destroy_sibling_or_exp()
165 exp = nf_ct_expect_find_get(net, zone, t); in destroy_sibling_or_exp()
/linux-4.4.14/drivers/uwb/
Ddrp-ie.c269 void uwb_drp_ie_single_zone_to_bm(struct uwb_mas_bm *bm, u8 zone, u16 mas_bm) in uwb_drp_ie_single_zone_to_bm() argument
277 set_bit(zone * UWB_NUM_ZONES + mas, bm->bm); in uwb_drp_ie_single_zone_to_bm()
302 u8 zone; in uwb_drp_ie_to_bm() local
311 for (zone = 0; zone < UWB_NUM_ZONES; zone++) { in uwb_drp_ie_to_bm()
312 zone_mask = 1 << zone; in uwb_drp_ie_to_bm()
314 uwb_drp_ie_single_zone_to_bm(bm, zone, mas_bm); in uwb_drp_ie_to_bm()
/linux-4.4.14/drivers/usb/storage/
Dalauda.c554 unsigned int zone) in alauda_find_unused_pba() argument
556 u16 *pba_to_lba = info->pba_to_lba[zone]; in alauda_find_unused_pba()
561 return (zone << info->zoneshift) + i; in alauda_find_unused_pba()
570 static int alauda_read_map(struct us_data *us, unsigned int zone) in alauda_read_map() argument
578 unsigned int zone_base_lba = zone * uzonesize; in alauda_read_map()
579 unsigned int zone_base_pba = zone * zonesize; in alauda_read_map()
587 usb_stor_dbg(us, "Mapping blocks for zone %d\n", zone); in alauda_read_map()
678 MEDIA_INFO(us).lba_to_pba[zone] = lba_to_pba; in alauda_read_map()
679 MEDIA_INFO(us).pba_to_lba[zone] = pba_to_lba; in alauda_read_map()
694 static void alauda_ensure_map_for_zone(struct us_data *us, unsigned int zone) in alauda_ensure_map_for_zone() argument
[all …]
Dsddr55.c721 int zone = i / 1024; in sddr55_read_map() local
747 if (info->lba_to_pba[lba + zone * 1000] != NOT_ALLOCATED && in sddr55_read_map()
751 lba + zone * 1000); in sddr55_read_map()
758 info->lba_to_pba[lba + zone * 1000] = i; in sddr55_read_map()
/linux-4.4.14/arch/powerpc/mm/
Dmem.c119 struct zone *zone; in arch_add_memory() local
130 zone = pgdata->node_zones + in arch_add_memory()
133 return __add_pages(nid, zone, start_pfn, nr_pages); in arch_add_memory()
141 struct zone *zone; in arch_remove_memory() local
144 zone = page_zone(pfn_to_page(start_pfn)); in arch_remove_memory()
145 ret = __remove_pages(zone, start_pfn, nr_pages); in arch_remove_memory()
242 void __init limit_zone_pfn(enum zone_type zone, unsigned long pfn_limit) in limit_zone_pfn() argument
249 for (i = zone; i >= 0; i--) { in limit_zone_pfn()
/linux-4.4.14/Documentation/cma/
Ddebugfs.txt4 Each CMA zone represents a directory under <debugfs>/cma/, indexed by the
5 kernel's CMA index. So the first CMA zone would be:
11 - [RO] base_pfn: The base PFN (Page Frame Number) of the zone.
14 - [RO] bitmap: The bitmap of page states in the zone.
/linux-4.4.14/Documentation/networking/
Dcops.txt52 dummy -seed -phase 2 -net 2000 -addr 2000.10 -zone "1033"
53 lt0 -seed -phase 1 -net 1000 -addr 1000.50 -zone "1033"
56 eth0 -seed -phase 2 -net 3000 -addr 3000.20 -zone "1033"
57 lt0 -seed -phase 1 -net 1000 -addr 1000.50 -zone "1033"
61 lt0 -seed -phase 1 -net 1000 -addr 1000.10 -zone "LocalTalk1"
62 lt1 -seed -phase 1 -net 2000 -addr 2000.20 -zone "LocalTalk2"
63 eth0 -seed -phase 2 -net 3000 -addr 3000.30 -zone "EtherTalk"
Dltpc.txt47 dummy -seed -phase 2 -net 2000 -addr 2000.26 -zone "1033"
48 lt0 -seed -phase 1 -net 1033 -addr 1033.27 -zone "1033"
Dfib_trie.txt105 as one "zone"/hash per prefix length.
Dnetlink_mmap.txt83 the ring just contains the frames in a continuous memory zone.
/linux-4.4.14/arch/m68k/mm/
Dmcfmmu.c42 enum zone_type zone; in paging_init() local
78 for (zone = 0; zone < MAX_NR_ZONES; zone++) in paging_init()
79 zones_size[zone] = 0x0; in paging_init()
/linux-4.4.14/drivers/base/
Dmemory.c393 struct zone *zone; in show_valid_zones() local
403 zone = page_zone(first_page); in show_valid_zones()
405 if (zone_idx(zone) == ZONE_MOVABLE - 1) { in show_valid_zones()
407 if (end_pfn == zone_end_pfn(zone)) in show_valid_zones()
409 zone->name, (zone + 1)->name); in show_valid_zones()
412 if (zone_idx(zone) == ZONE_MOVABLE) { in show_valid_zones()
414 if (start_pfn == zone->zone_start_pfn) in show_valid_zones()
416 zone->name, (zone - 1)->name); in show_valid_zones()
419 return sprintf(buf, "%s\n", zone->name); in show_valid_zones()
/linux-4.4.14/arch/tile/mm/
Dpgtable.c45 struct zone *zone; in show_mem() local
64 for_each_zone(zone) { in show_mem()
67 if (!populated_zone(zone)) in show_mem()
70 spin_lock_irqsave(&zone->lock, flags); in show_mem()
72 int nr = zone->free_area[order].nr_free; in show_mem()
77 spin_unlock_irqrestore(&zone->lock, flags); in show_mem()
79 zone_to_nid(zone), zone->name, in show_mem()
Dinit.c696 struct zone *z; in set_non_bootmem_pages_init()
869 struct zone *zone = pgdata->node_zones + MAX_NR_ZONES-1; in arch_add_memory() local
873 return __add_pages(zone, start_pfn, nr_pages); in arch_add_memory()
/linux-4.4.14/Documentation/hwmon/
Ddme1737165 be retrieved via the zone[1-3]_auto_channels_temp attributes. Each PWM output
166 is assigned to one (or hottest of multiple) temperature zone(s) through the
171 programmed via zone[1-3]_auto_point[1-3]_temp and
172 zone[1-3]_auto_point1_temp_hyst:
178 zone[1-3]_auto_point3_temp full-speed temp (all outputs)
179 zone[1-3]_auto_point2_temp full-speed temp
180 zone[1-3]_auto_point1_temp low-speed temp
181 zone[1-3]_auto_point1_temp_hyst min-speed temp
184 to auto_point2_pwm if the temperature of the associated zone is between
193 based on the temperature of the associated thermal zone:
[all …]
Dasc762117 added PECI and a 4th thermal zone. The Andigilog aSC7611 is the
47 temperature may be mapped to any zone, which has a default assignment
50 Since there is a fan to zone assignment that allows for the "hotter" of
52 indication to the user, we have added an indicator that shows which zone
62 address. One could use our flexible zone assignment to have a zone
181 When the corresponding zone temperature reaches this value,
195 if the temperature for the zone is below the minimum.
197 for the zone is below the minimum.
Dlm85170 sensors and therefore three zones (# 1, 2 and 3). Each zone has the following
182 configured and assigned to a zone for its control value. Each PWM can be
195 PWMs controlled by zone '#'.
199 * pwm#_auto_channels - controls zone that is associated with PWM
Df71882fg93 Note that the lowest numbered temperature zone trip point corresponds to
95 vica versa. So the temperature zone trip points 1-4 (or 1-2) go from high temp
/linux-4.4.14/Documentation/sysctl/
Dvm.txt229 debugfs shows what the fragmentation index for each order is in each zone in
234 The kernel will not compact memory in a zone if the
288 zone. This is because that memory could then be pinned via the mlock()
316 zone's value is not necessary for following calculation.
321 Each zone has an array of protection pages like this.
324 Node 0, zone DMA
338 These protections are added to score to judge whether this zone should be used
341 In this example, if normal pages (index=2) are required to this DMA zone and
342 watermark[WMARK_HIGH] is used for watermark, the kernel judges this zone should
344 (4 + 2004 = 2008). If this protection value is 0, this zone would be used for
[all …]
/linux-4.4.14/fs/sysv/
Dballoc.c187 sysv_zone_t zone; in sysv_count_free_blocks() local
190 zone = 0; in sysv_count_free_blocks()
191 while (n && (zone = blocks[--n]) != 0) in sysv_count_free_blocks()
193 if (zone == 0) in sysv_count_free_blocks()
196 block = fs32_to_cpu(sbi, zone); in sysv_count_free_blocks()
/linux-4.4.14/fs/proc/
Dmeminfo.c35 struct zone *zone; in meminfo_proc_show() local
54 for_each_zone(zone) in meminfo_proc_show()
55 wmark_low += zone->watermark[WMARK_LOW]; in meminfo_proc_show()
/linux-4.4.14/fs/minix/
Dbitmap.c47 unsigned long bit, zone; in minix_free_block() local
53 zone = block - sbi->s_firstdatazone + 1; in minix_free_block()
54 bit = zone & ((1<<k) - 1); in minix_free_block()
55 zone >>= k; in minix_free_block()
56 if (zone >= sbi->s_zmap_blocks) { in minix_free_block()
60 bh = sbi->s_zmap[zone]; in minix_free_block()
/linux-4.4.14/drivers/net/fjes/
Dfjes_hw.c683 (hw->ep_shm_info[hw->my_epid].zone == in fjes_hw_epid_is_same_zone()
687 return (hw->ep_shm_info[epid].zone == in fjes_hw_epid_is_same_zone()
688 hw->ep_shm_info[hw->my_epid].zone); in fjes_hw_epid_is_same_zone()
929 struct my_s {u8 es_status; u8 zone; } *info; in fjes_hw_update_zone_task() member
967 hw->ep_shm_info[epidx].zone = in fjes_hw_update_zone_task()
968 info[epidx].zone; in fjes_hw_update_zone_task()
976 if ((info[epidx].zone != in fjes_hw_update_zone_task()
980 (info[epidx].zone == in fjes_hw_update_zone_task()
981 info[hw->my_epid].zone)) in fjes_hw_update_zone_task()
989 if ((info[epidx].zone == in fjes_hw_update_zone_task()
[all …]
Dfjes_hw.h149 u8 zone; member
244 u8 zone; member
Dfjes_main.c343 hw->ep_shm_info[epidx].zone = in fjes_setup_resources()
344 hw->hw_info.res_buf->info.info[epidx].zone; in fjes_setup_resources()
/linux-4.4.14/include/uapi/linux/netfilter/
Dxt_CT.h20 __u16 zone; member
31 __u16 zone; member
/linux-4.4.14/arch/parisc/mm/
Dinit.c670 struct zone *zone = &pgdat->node_zones[zoneid]; in show_mem() local
671 if (!populated_zone(zone)) in show_mem()
674 total += zone->present_pages; in show_mem()
675 reserved = zone->present_pages - zone->managed_pages; in show_mem()
692 struct zone *zone; in show_mem() local
695 for_each_zone_zonelist(zone, z, zl, j) in show_mem()
696 printk("[%d/%s] ", zone_to_nid(zone), in show_mem()
697 zone->name); in show_mem()
/linux-4.4.14/fs/ntfs/
Dlcnalloc.c148 const NTFS_CLUSTER_ALLOCATION_ZONES zone, in ntfs_cluster_alloc() argument
167 zone == MFT_ZONE ? "MFT" : "DATA"); in ntfs_cluster_alloc()
174 BUG_ON(zone < FIRST_ZONE); in ntfs_cluster_alloc()
175 BUG_ON(zone > LAST_ZONE); in ntfs_cluster_alloc()
202 if (zone == DATA_ZONE) in ntfs_cluster_alloc()
213 } else if (zone == DATA_ZONE && zone_start >= vol->mft_zone_start && in ntfs_cluster_alloc()
221 } else if (zone == MFT_ZONE && (zone_start < vol->mft_zone_start || in ntfs_cluster_alloc()
232 if (zone == MFT_ZONE) { in ntfs_cluster_alloc()
696 if (zone == MFT_ZONE || mft_zone_size <= 0) { in ntfs_cluster_alloc()
Dlcnalloc.h45 const NTFS_CLUSTER_ALLOCATION_ZONES zone,
/linux-4.4.14/sound/soc/
Dsoc-jack.c159 struct snd_soc_jack_zone *zone; in snd_soc_jack_get_type() local
161 list_for_each_entry(zone, &jack->jack_zones, list) { in snd_soc_jack_get_type()
162 if (micbias_voltage >= zone->min_mv && in snd_soc_jack_get_type()
163 micbias_voltage < zone->max_mv) in snd_soc_jack_get_type()
164 return zone->jack_type; in snd_soc_jack_get_type()
/linux-4.4.14/drivers/platform/x86/
Dalienware-wmi.c174 static int parse_rgb(const char *buf, struct platform_zone *zone) in parse_rgb() argument
194 zone->colors = repackager.cp; in parse_rgb()
214 static int alienware_update_led(struct platform_zone *zone) in alienware_update_led() argument
223 wmax_args.led_mask = 1 << zone->location; in alienware_update_led()
224 wmax_args.colors = zone->colors; in alienware_update_led()
232 legacy_args.colors = zone->colors; in alienware_update_led()
241 method_id = zone->location + 1; in alienware_update_led()
/linux-4.4.14/Documentation/devicetree/bindings/power_supply/
Dcharger-manager.txt24 - cm-thermal-zone : name of external thermometer's thermal zone
53 cm-thermal-zone = "thermal_zone.1"
/linux-4.4.14/include/uapi/linux/
Dtipc.h63 static inline __u32 tipc_addr(unsigned int zone, in tipc_addr() argument
67 return (zone << 24) | (cluster << 12) | node; in tipc_addr()
/linux-4.4.14/arch/powerpc/kernel/
Ddma.c81 int zone; in __dma_direct_alloc_coherent() local
89 zone = dma_pfn_limit_to_zone(pfn); in __dma_direct_alloc_coherent()
90 if (zone < 0) { in __dma_direct_alloc_coherent()
96 switch (zone) { in __dma_direct_alloc_coherent()
/linux-4.4.14/Documentation/trace/
Devents-kmem.txt52 amounts of activity imply high activity on the zone->lock. Taking this lock
64 contention on the zone->lru_lock.
72 for order-0 pages, reduces contention on the zone->lock and reduces the
76 the zone->lock will be taken once and the per-CPU list refilled. The event
85 consecutively imply the zone->lock being taken once. Large amounts of per-CPU
/linux-4.4.14/arch/sh/mm/
Dinit.c523 struct zone *zone; in arch_remove_memory() local
526 zone = page_zone(pfn_to_page(start_pfn)); in arch_remove_memory()
527 ret = __remove_pages(zone, start_pfn, nr_pages); in arch_remove_memory()
DKconfig25 int "Maximum zone order"
35 blocks into "zones", where each zone is a power of two number of
/linux-4.4.14/include/uapi/linux/tc_act/
Dtc_connmark.h11 __u16 zone; member
/linux-4.4.14/include/net/tc_act/
Dtc_connmark.h9 u16 zone; member
/linux-4.4.14/drivers/net/wireless/rsi/
Drsi_91x_main.c43 void rsi_dbg(u32 zone, const char *fmt, ...) in rsi_dbg() argument
53 if (zone & rsi_zone_enabled) in rsi_dbg()
Drsi_main.h43 extern __printf(2, 3) void rsi_dbg(u32 zone, const char *fmt, ...);
/linux-4.4.14/kernel/sched/
Dwait.c489 const struct zone *zone = page_zone(virt_to_page(word)); in bit_waitqueue() local
492 return &zone->wait_table[hash_long(val, zone->wait_table_bits)]; in bit_waitqueue()
/linux-4.4.14/arch/arm/boot/dts/
Dexynos5440-trip-points.dtsi2 * Device tree sources for default Exynos5440 thermal zone definition
Dexynos5420-trip-points.dtsi2 * Device tree sources for default Exynos5420 thermal zone definition
Dexynos4-cpu-thermal.dtsi2 * Device tree sources for Exynos4 thermal zone
Dexynos5422-cpu-thermal.dtsi2 * Device tree sources for Exynos5422 thermal zone
/linux-4.4.14/drivers/block/
Dpktcdvd.c1093 static struct packet_data *pkt_get_packet_data(struct pktcdvd_device *pd, int zone) in pkt_get_packet_data() argument
1098 if (pkt->sector == zone || pkt->list.next == &pd->cdrw.pkt_free_list) { in pkt_get_packet_data()
1100 if (pkt->sector != zone) in pkt_get_packet_data()
1199 sector_t zone = 0; /* Suppress gcc warning */ in pkt_handle_queue() local
1224 zone = get_zone(bio->bi_iter.bi_sector, pd); in pkt_handle_queue()
1226 if (p->sector == zone) { in pkt_handle_queue()
1248 pkt = pkt_get_packet_data(pd, zone); in pkt_handle_queue()
1250 pd->current_sector = zone + pd->settings.size; in pkt_handle_queue()
1251 pkt->sector = zone; in pkt_handle_queue()
1260 pkt_dbg(2, pd, "looking for zone %llx\n", (unsigned long long)zone); in pkt_handle_queue()
[all …]
/linux-4.4.14/arch/tile/include/asm/
Dhomecache.h26 struct zone;
/linux-4.4.14/net/ipv4/netfilter/
Dnf_conntrack_proto_icmp.c137 const struct nf_conntrack_zone *zone; in icmp_error_message() local
141 zone = nf_ct_zone_tmpl(tmpl, skb, &tmp); in icmp_error_message()
165 h = nf_conntrack_find_get(net, zone, &innertuple); in icmp_error_message()
/linux-4.4.14/arch/metag/mm/
DKconfig52 int "Maximum zone order"
57 blocks into "zones", where each zone is a power of two number of
/linux-4.4.14/arch/c6x/
DKconfig88 int "Maximum zone order"
92 blocks into "zones", where each zone is a power of two number of
/linux-4.4.14/drivers/memstick/core/
Dms_block.c174 int zone = msb_get_zone_from_pba(pba); in msb_mark_block_used() local
188 msb->free_block_count[zone]--; in msb_mark_block_used()
194 int zone = msb_get_zone_from_pba(pba); in msb_mark_block_unused() local
207 msb->free_block_count[zone]++; in msb_mark_block_unused()
1078 static u16 msb_get_free_block(struct msb_data *msb, int zone) in msb_get_free_block() argument
1081 int pba = zone * MS_BLOCKS_IN_ZONE; in msb_get_free_block()
1086 if (!msb->free_block_count[zone]) { in msb_get_free_block()
1087 …O free blocks in the zone %d, to use for a write, (media is WORN out) switching to RO mode", zone); in msb_get_free_block()
1092 pos %= msb->free_block_count[zone]; in msb_get_free_block()
1095 msb->free_block_count[zone], pos); in msb_get_free_block()
[all …]
/linux-4.4.14/drivers/firmware/efi/libstub/
DMakefile10 -fPIC -fno-strict-aliasing -mno-red-zone \
/linux-4.4.14/Documentation/cris/
DREADME84 zone(0): 2048 pages.
85 zone(1): 0 pages.
86 zone(2): 0 pages.
/linux-4.4.14/kernel/
Dkexec_core.c1381 VMCOREINFO_STRUCT_SIZE(zone); in crash_save_vmcoreinfo_init()
1399 VMCOREINFO_OFFSET(zone, free_area); in crash_save_vmcoreinfo_init()
1400 VMCOREINFO_OFFSET(zone, vm_stat); in crash_save_vmcoreinfo_init()
1401 VMCOREINFO_OFFSET(zone, spanned_pages); in crash_save_vmcoreinfo_init()
1407 VMCOREINFO_LENGTH(zone.free_area, MAX_ORDER); in crash_save_vmcoreinfo_init()
Dfork.c217 struct zone *zone = page_zone(virt_to_page(ti)); in account_kernel_stack() local
219 mod_zone_page_state(zone, NR_KERNEL_STACK, account); in account_kernel_stack()
/linux-4.4.14/include/sound/
Demux_synth.h183 struct snd_sf_zone *zone; /* Zone assigned to this note */ member
/linux-4.4.14/arch/nios2/
DKconfig60 int "Maximum zone order"
65 blocks into "zones", where each zone is a power of two number of
/linux-4.4.14/fs/xfs/libxfs/
Dxfs_da_btree.h25 struct zone;
/linux-4.4.14/Documentation/filesystems/
Dsysv-fs.txt36 * Size of a block or zone (data allocation unit on disk)
196 Notation: We often speak of a "block" but mean a zone (the allocation unit)
Dproc.txt761 Node 0, zone DMA 0 4 5 4 4 3 ...
762 Node 0, zone Normal 1 0 0 1 101 8 ...
763 Node 0, zone HighMem 2 0 0 1 1 0 ...
783 Node 0, zone DMA, type Unmovable 0 0 0 1 1 1 1 1…
784 Node 0, zone DMA, type Reclaimable 0 0 0 0 0 0 0 0…
785 Node 0, zone DMA, type Movable 1 1 2 1 2 1 1 0…
786 Node 0, zone DMA, type Reserve 0 0 0 0 0 0 0 0…
787 Node 0, zone DMA, type Isolate 0 0 0 0 0 0 0 0…
788 Node 0, zone DMA32, type Unmovable 103 54 77 1 1 1 11 8…
789 Node 0, zone DMA32, type Reclaimable 0 0 2 1 0 0 0 0…
[all …]
Df2fs.txt279 -z [int] : Set the number of sections per zone.
324 to 2MB in size. A section is composed of consecutive segments, and a zone
325 consists of a set of sections. By default, section and zone sizes are set to one
331 align with the zone size <-|
376 start block address of Main area with the zone size by reserving some segments
Dntfs.txt192 mft_zone_multiplier= Set the MFT zone multiplier for the volume (this
196 default. The MFT zone multiplier determines how much
198 other space is used up, then the MFT zone will be
205 Value MFT zone size (% of volume size)
Dvfat.txt118 Linux. This is useful when time zone set in sys_tz is
119 not the time zone used by the filesystem. Note that this
/linux-4.4.14/arch/x86/boot/compressed/
DMakefile50 $(obj)/eboot.o: KBUILD_CFLAGS += -fshort-wchar -mno-red-zone
/linux-4.4.14/arch/powerpc/include/asm/
Dpgtable.h229 void limit_zone_pfn(enum zone_type zone, unsigned long max_pfn);
/linux-4.4.14/drivers/misc/
Dlkdtm.c207 struct zone *zone, in jp_shrink_inactive_list() argument
/linux-4.4.14/Documentation/cgroups/
Dmemcg_test.txt110 VM's control (means that it's handled under global zone->lru_lock).
112 list management functions under zone->lru_lock().
Dmemory.txt270 zone->lru_lock
273 per-zone-per-cgroup LRU (cgroup's private LRU) is just guarded by
274 zone->lru_lock, it has no lock of its own.
/linux-4.4.14/drivers/staging/unisys/Documentation/
Dproc-entries.txt49 clientstring and zone.
Doverview.txt234 zoneguid a GUID identifying the channel zone, in
/linux-4.4.14/Documentation/locking/
Dlockstat.txt174 …&(&zone->lru_lock)->rlock: 93000 94934 0.16 59.18 188253.…
/linux-4.4.14/arch/x86/
DMakefile117 KBUILD_CFLAGS += -mno-red-zone
/linux-4.4.14/Documentation/acpi/
Dnamespace.txt161 +-| _TZ | Scope(_TZ): the thermal zone namespace
177 | +-| TZ00 | ThermalZone(TZ00); the FAN thermal zone
/linux-4.4.14/Documentation/devicetree/bindings/arm/
Darm,scpi.txt187 zone. The ID "3" is the sensor identifier for the temperature sensor
/linux-4.4.14/Documentation/devicetree/bindings/arm/bcm/
Dbrcm,brcmstb.txt81 o offset to the base CPU power zone register
/linux-4.4.14/arch/ia64/include/asm/
Dpgtable.h547 extern void memmap_init (unsigned long size, int nid, unsigned long zone,
/linux-4.4.14/arch/xtensa/
DKconfig402 int "Maximum zone order"
406 blocks into "zones", where each zone is a power of two number of
/linux-4.4.14/fs/
Dbuffer.c263 struct zone *zone; in free_more_memory() local
272 &zone); in free_more_memory()
273 if (zone) in free_more_memory()
/linux-4.4.14/Documentation/input/
Diforce-protocol.txt202 Idx 00 Set dead zone (0..2048)
/linux-4.4.14/fs/f2fs/
Dsegment.h263 unsigned int zone; /* current zone number */ member
Dsegment.c939 if (CURSEG_I(sbi, i)->zone == zoneno) in get_new_segment()
967 curseg->zone = GET_ZONENO_FROM_SEGNO(sbi, curseg->segno); in reset_curseg()
/linux-4.4.14/Documentation/
Dmemory-hotplug.txt176 The first column shows it's default zone.
308 create ZONE_MOVABLE...a zone which is just used for movable pages.
Drtc.txt8 the local time zone or daylight savings time -- unless they dual boot
Dkernel-parameters.txt1301 highmem=nn[KMG] [KNL,BOOT] forces the highmem zone to have an exact
1703 of Movable pages. The Movable zone is used for the
1706 HugeTLB pages may not be allocated from this zone.
1708 use the HighMem zone if it exists, and the Normal
1709 zone if it does not.
2617 one of ['zone', 'node', 'default'] can be specified
3697 Set to disable actions on ACPI thermal zone
3709 Specify global default ACPI thermal zone polling rate
/linux-4.4.14/
DREPORTING-BUGS162 around the world, and they may not work in your time zone. Unless you
/linux-4.4.14/fs/affs/
DChanges192 zone number was zero, the loop didn't terminate,
/linux-4.4.14/Documentation/DocBook/
Dkernel-api.xml.db235 API-nr-free-zone-pages
246 API-setup-per-zone-wmarks
Dalsa-driver-api.xml.db221 API-struct-snd-soc-jack-zone
/linux-4.4.14/arch/powerpc/
DKconfig587 int "Maximum zone order"
602 blocks into "zones", where each zone is a power of two number of
/linux-4.4.14/arch/arm64/
DKconfig577 blocks into "zones", where each zone is a power of two number of

12