/linux-4.4.14/drivers/gpu/drm/ttm/ |
D | ttm_memory.c | 76 struct ttm_mem_zone *zone = in ttm_mem_zone_kobj_release() local 80 zone->name, (unsigned long long)zone->used_mem >> 10); in ttm_mem_zone_kobj_release() 81 kfree(zone); in ttm_mem_zone_kobj_release() 88 struct ttm_mem_zone *zone = in ttm_mem_zone_show() local 92 spin_lock(&zone->glob->lock); in ttm_mem_zone_show() 94 val = zone->zone_mem; in ttm_mem_zone_show() 96 val = zone->emer_mem; in ttm_mem_zone_show() 98 val = zone->max_mem; in ttm_mem_zone_show() 100 val = zone->swap_limit; in ttm_mem_zone_show() 102 val = zone->used_mem; in ttm_mem_zone_show() [all …]
|
/linux-4.4.14/mm/ |
D | compaction.c | 92 unsigned long end_pfn, struct zone *zone) in pageblock_pfn_to_page() argument 105 if (page_zone(start_page) != zone) in pageblock_pfn_to_page() 127 void defer_compaction(struct zone *zone, int order) in defer_compaction() argument 129 zone->compact_considered = 0; in defer_compaction() 130 zone->compact_defer_shift++; in defer_compaction() 132 if (order < zone->compact_order_failed) in defer_compaction() 133 zone->compact_order_failed = order; in defer_compaction() 135 if (zone->compact_defer_shift > COMPACT_MAX_DEFER_SHIFT) in defer_compaction() 136 zone->compact_defer_shift = COMPACT_MAX_DEFER_SHIFT; in defer_compaction() 138 trace_mm_compaction_defer_compaction(zone, order); in defer_compaction() [all …]
|
D | memory_hotplug.c | 250 struct zone *zone; in register_page_bootmem_info_node() local 258 zone = &pgdat->node_zones[0]; in register_page_bootmem_info_node() 259 for (; zone < pgdat->node_zones + MAX_NR_ZONES - 1; zone++) { in register_page_bootmem_info_node() 260 if (zone_is_initialized(zone)) { in register_page_bootmem_info_node() 261 nr_pages = zone->wait_table_hash_nr_entries in register_page_bootmem_info_node() 264 page = virt_to_page(zone->wait_table); in register_page_bootmem_info_node() 288 static void __meminit grow_zone_span(struct zone *zone, unsigned long start_pfn, in grow_zone_span() argument 293 zone_span_writelock(zone); in grow_zone_span() 295 old_zone_end_pfn = zone_end_pfn(zone); in grow_zone_span() 296 if (zone_is_empty(zone) || start_pfn < zone->zone_start_pfn) in grow_zone_span() [all …]
|
D | vmstat.c | 94 int calculate_pressure_threshold(struct zone *zone) in calculate_pressure_threshold() argument 107 watermark_distance = low_wmark_pages(zone) - min_wmark_pages(zone); in calculate_pressure_threshold() 118 int calculate_normal_threshold(struct zone *zone) in calculate_normal_threshold() argument 153 mem = zone->managed_pages >> (27 - PAGE_SHIFT); in calculate_normal_threshold() 170 struct zone *zone; in refresh_zone_stat_thresholds() local 174 for_each_populated_zone(zone) { in refresh_zone_stat_thresholds() 177 threshold = calculate_normal_threshold(zone); in refresh_zone_stat_thresholds() 180 per_cpu_ptr(zone->pageset, cpu)->stat_threshold in refresh_zone_stat_thresholds() 188 tolerate_drift = low_wmark_pages(zone) - min_wmark_pages(zone); in refresh_zone_stat_thresholds() 191 zone->percpu_drift_mark = high_wmark_pages(zone) + in refresh_zone_stat_thresholds() [all …]
|
D | page_alloc.c | 349 static int page_outside_zone_boundaries(struct zone *zone, struct page *page) in page_outside_zone_boundaries() argument 357 seq = zone_span_seqbegin(zone); in page_outside_zone_boundaries() 358 start_pfn = zone->zone_start_pfn; in page_outside_zone_boundaries() 359 sp = zone->spanned_pages; in page_outside_zone_boundaries() 360 if (!zone_spans_pfn(zone, pfn)) in page_outside_zone_boundaries() 362 } while (zone_span_seqretry(zone, seq)); in page_outside_zone_boundaries() 366 pfn, zone_to_nid(zone), zone->name, in page_outside_zone_boundaries() 372 static int page_is_consistent(struct zone *zone, struct page *page) in page_is_consistent() argument 376 if (zone != page_zone(page)) in page_is_consistent() 384 static int bad_range(struct zone *zone, struct page *page) in bad_range() argument [all …]
|
D | vmscan.c | 195 static unsigned long zone_reclaimable_pages(struct zone *zone) in zone_reclaimable_pages() argument 199 nr = zone_page_state(zone, NR_ACTIVE_FILE) + in zone_reclaimable_pages() 200 zone_page_state(zone, NR_INACTIVE_FILE); in zone_reclaimable_pages() 203 nr += zone_page_state(zone, NR_ACTIVE_ANON) + in zone_reclaimable_pages() 204 zone_page_state(zone, NR_INACTIVE_ANON); in zone_reclaimable_pages() 209 bool zone_reclaimable(struct zone *zone) in zone_reclaimable() argument 211 return zone_page_state(zone, NR_PAGES_SCANNED) < in zone_reclaimable() 212 zone_reclaimable_pages(zone) * 6; in zone_reclaimable() 881 struct zone *zone, in shrink_page_list() argument 919 VM_BUG_ON_PAGE(page_zone(page) != zone, page); in shrink_page_list() [all …]
|
D | page_isolation.c | 15 struct zone *zone; in set_migratetype_isolate() local 21 zone = page_zone(page); in set_migratetype_isolate() 23 spin_lock_irqsave(&zone->lock, flags); in set_migratetype_isolate() 49 if (!has_unmovable_pages(zone, page, arg.pages_found, in set_migratetype_isolate() 64 zone->nr_isolate_pageblock++; in set_migratetype_isolate() 65 nr_pages = move_freepages_block(zone, page, MIGRATE_ISOLATE); in set_migratetype_isolate() 67 __mod_zone_freepage_state(zone, -nr_pages, migratetype); in set_migratetype_isolate() 70 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 72 drain_all_pages(zone); in set_migratetype_isolate() 78 struct zone *zone; in unset_migratetype_isolate() local [all …]
|
D | mmzone.c | 29 struct zone *next_zone(struct zone *zone) in next_zone() argument 31 pg_data_t *pgdat = zone->zone_pgdat; in next_zone() 33 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) in next_zone() 34 zone++; in next_zone() 38 zone = pgdat->node_zones; in next_zone() 40 zone = NULL; in next_zone() 42 return zone; in next_zone() 68 (z->zone && !zref_in_nodemask(z, nodes))) in next_zones_zonelist() 76 struct page *page, struct zone *zone) in memmap_valid_within() argument 81 if (page_zone(page) != zone) in memmap_valid_within()
|
D | workingset.c | 155 static void *pack_shadow(unsigned long eviction, struct zone *zone) in pack_shadow() argument 157 eviction = (eviction << NODES_SHIFT) | zone_to_nid(zone); in pack_shadow() 158 eviction = (eviction << ZONES_SHIFT) | zone_idx(zone); in pack_shadow() 165 struct zone **zone, in unpack_shadow() argument 181 *zone = NODE_DATA(nid)->node_zones + zid; in unpack_shadow() 183 refault = atomic_long_read(&(*zone)->inactive_age); in unpack_shadow() 215 struct zone *zone = page_zone(page); in workingset_eviction() local 218 eviction = atomic_long_inc_return(&zone->inactive_age); in workingset_eviction() 219 return pack_shadow(eviction, zone); in workingset_eviction() 234 struct zone *zone; in workingset_refault() local [all …]
|
D | swap.c | 56 struct zone *zone = page_zone(page); in __page_cache_release() local 60 spin_lock_irqsave(&zone->lru_lock, flags); in __page_cache_release() 61 lruvec = mem_cgroup_page_lruvec(page, zone); in __page_cache_release() 65 spin_unlock_irqrestore(&zone->lru_lock, flags); in __page_cache_release() 423 struct zone *zone = NULL; in pagevec_lru_move_fn() local 429 struct zone *pagezone = page_zone(page); in pagevec_lru_move_fn() 431 if (pagezone != zone) { in pagevec_lru_move_fn() 432 if (zone) in pagevec_lru_move_fn() 433 spin_unlock_irqrestore(&zone->lru_lock, flags); in pagevec_lru_move_fn() 434 zone = pagezone; in pagevec_lru_move_fn() [all …]
|
D | page_owner.c | 214 static void init_pages_in_zone(pg_data_t *pgdat, struct zone *zone) in init_pages_in_zone() argument 218 unsigned long pfn = zone->zone_start_pfn, block_end_pfn; in init_pages_in_zone() 219 unsigned long end_pfn = pfn + zone->spanned_pages; in init_pages_in_zone() 223 pfn = zone->zone_start_pfn; in init_pages_in_zone() 272 pgdat->node_id, zone->name, count); in init_pages_in_zone() 277 struct zone *zone; in init_zones_in_node() local 278 struct zone *node_zones = pgdat->node_zones; in init_zones_in_node() 281 for (zone = node_zones; zone - node_zones < MAX_NR_ZONES; ++zone) { in init_zones_in_node() 282 if (!populated_zone(zone)) in init_zones_in_node() 285 spin_lock_irqsave(&zone->lock, flags); in init_zones_in_node() [all …]
|
D | mm_init.c | 34 struct zone *zone; in mminit_verify_zonelist() local 46 zone = &pgdat->node_zones[zoneid]; in mminit_verify_zonelist() 47 if (!populated_zone(zone)) in mminit_verify_zonelist() 53 zone->name); in mminit_verify_zonelist() 56 for_each_zone_zonelist(zone, z, zonelist, zoneid) { in mminit_verify_zonelist() 59 zone->node, zone->name); in mminit_verify_zonelist() 61 printk(KERN_CONT "0:%s ", zone->name); in mminit_verify_zonelist()
|
D | mlock.c | 176 struct zone *zone = page_zone(page); in munlock_vma_page() local 186 spin_lock_irq(&zone->lru_lock); in munlock_vma_page() 192 __mod_zone_page_state(zone, NR_MLOCK, -nr_pages); in munlock_vma_page() 195 spin_unlock_irq(&zone->lru_lock); in munlock_vma_page() 202 spin_unlock_irq(&zone->lru_lock); in munlock_vma_page() 276 static void __munlock_pagevec(struct pagevec *pvec, struct zone *zone) in __munlock_pagevec() argument 287 spin_lock_irq(&zone->lru_lock); in __munlock_pagevec() 312 __mod_zone_page_state(zone, NR_MLOCK, delta_munlocked); in __munlock_pagevec() 313 spin_unlock_irq(&zone->lru_lock); in __munlock_pagevec() 432 struct zone *zone; in munlock_vma_pages_range() local [all …]
|
D | internal.h | 120 extern bool zone_reclaimable(struct zone *zone); 147 struct zone *preferred_zone; 212 struct zone *zone; member 432 unsigned long reclaim_clean_pages_from_list(struct zone *zone,
|
D | page_idle.c | 33 struct zone *zone; in page_idle_get_page() local 43 zone = page_zone(page); in page_idle_get_page() 44 spin_lock_irq(&zone->lru_lock); in page_idle_get_page() 49 spin_unlock_irq(&zone->lru_lock); in page_idle_get_page()
|
D | memcontrol.c | 405 mem_cgroup_zone_zoneinfo(struct mem_cgroup *memcg, struct zone *zone) in mem_cgroup_zone_zoneinfo() argument 407 int nid = zone_to_nid(zone); in mem_cgroup_zone_zoneinfo() 408 int zid = zone_idx(zone); in mem_cgroup_zone_zoneinfo() 901 mz = mem_cgroup_zone_zoneinfo(root, reclaim->zone); in mem_cgroup_iter() 1051 struct lruvec *mem_cgroup_zone_lruvec(struct zone *zone, in mem_cgroup_zone_lruvec() argument 1058 lruvec = &zone->lruvec; in mem_cgroup_zone_lruvec() 1062 mz = mem_cgroup_zone_zoneinfo(memcg, zone); in mem_cgroup_zone_lruvec() 1070 if (unlikely(lruvec->zone != zone)) in mem_cgroup_zone_lruvec() 1071 lruvec->zone = zone; in mem_cgroup_zone_lruvec() 1084 struct lruvec *mem_cgroup_page_lruvec(struct page *page, struct zone *zone) in mem_cgroup_page_lruvec() argument [all …]
|
D | cma.c | 102 struct zone *zone; in cma_activate_area() local 110 zone = page_zone(pfn_to_page(pfn)); in cma_activate_area() 124 if (page_zone(pfn_to_page(pfn)) != zone) in cma_activate_area()
|
D | page-writeback.c | 276 static unsigned long zone_dirtyable_memory(struct zone *zone) in zone_dirtyable_memory() argument 280 nr_pages = zone_page_state(zone, NR_FREE_PAGES); in zone_dirtyable_memory() 281 nr_pages -= min(nr_pages, zone->dirty_balance_reserve); in zone_dirtyable_memory() 283 nr_pages += zone_page_state(zone, NR_INACTIVE_FILE); in zone_dirtyable_memory() 284 nr_pages += zone_page_state(zone, NR_ACTIVE_FILE); in zone_dirtyable_memory() 296 struct zone *z = &NODE_DATA(node)->node_zones[ZONE_HIGHMEM]; in highmem_dirtyable_memory() 437 static unsigned long zone_dirty_limit(struct zone *zone) in zone_dirty_limit() argument 439 unsigned long zone_memory = zone_dirtyable_memory(zone); in zone_dirty_limit() 462 bool zone_dirty_ok(struct zone *zone) in zone_dirty_ok() argument 464 unsigned long limit = zone_dirty_limit(zone); in zone_dirty_ok() [all …]
|
D | sparse.c | 693 int __meminit sparse_add_one_section(struct zone *zone, unsigned long start_pfn) in sparse_add_one_section() argument 696 struct pglist_data *pgdat = zone->zone_pgdat; in sparse_add_one_section() 791 void sparse_remove_one_section(struct zone *zone, struct mem_section *ms) in sparse_remove_one_section() argument 795 struct pglist_data *pgdat = zone->zone_pgdat; in sparse_remove_one_section()
|
D | oom_kill.c | 211 struct zone *zone; in constrained_alloc() local 244 for_each_zone_zonelist_nodemask(zone, z, oc->zonelist, in constrained_alloc() 246 if (!cpuset_zone_allowed(zone, oc->gfp_mask)) in constrained_alloc()
|
D | quicklist.c | 30 struct zone *zones = NODE_DATA(node)->node_zones; in max_pages()
|
D | migrate.c | 317 struct zone *oldzone, *newzone; in migrate_page_move_mapping() 1557 struct zone *zone = pgdat->node_zones + z; in migrate_balanced_pgdat() local 1559 if (!populated_zone(zone)) in migrate_balanced_pgdat() 1562 if (!zone_reclaimable(zone)) in migrate_balanced_pgdat() 1566 if (!zone_watermark_ok(zone, 0, in migrate_balanced_pgdat() 1567 high_wmark_pages(zone) + in migrate_balanced_pgdat()
|
D | mempolicy.c | 1611 static int apply_policy_zone(struct mempolicy *policy, enum zone_type zone) in apply_policy_zone() argument 1628 return zone >= dynamic_policy_zone; in apply_policy_zone() 1719 struct zone *zone; in mempolicy_slab_node() local 1724 &zone); in mempolicy_slab_node() 1725 return zone ? zone->node : node; in mempolicy_slab_node() 2250 struct zone *zone; in mpol_misplaced() local 2293 &pol->v.nodes, &zone); in mpol_misplaced() 2294 polnid = zone->node; in mpol_misplaced()
|
D | huge_memory.c | 112 struct zone *zone; in set_recommended_min_free_kbytes() local 116 for_each_populated_zone(zone) in set_recommended_min_free_kbytes() 1708 struct zone *zone = page_zone(page); in __split_huge_page_refcount() local 1713 spin_lock_irq(&zone->lru_lock); in __split_huge_page_refcount() 1714 lruvec = mem_cgroup_page_lruvec(page, zone); in __split_huge_page_refcount() 1797 __mod_zone_page_state(zone, NR_ANON_TRANSPARENT_HUGEPAGES, -1); in __split_huge_page_refcount() 1801 spin_unlock_irq(&zone->lru_lock); in __split_huge_page_refcount()
|
D | backing-dev.c | 968 long wait_iff_congested(struct zone *zone, int sync, long timeout) in wait_iff_congested() argument 981 !test_bit(ZONE_CONGESTED, &zone->flags)) { in wait_iff_congested()
|
D | nobootmem.c | 163 struct zone *z; in reset_node_managed_pages()
|
D | hugetlb.c | 884 struct zone *zone; in dequeue_huge_page_vma() local 906 for_each_zone_zonelist_nodemask(zone, z, zonelist, in dequeue_huge_page_vma() 908 if (cpuset_zone_allowed(zone, htlb_alloc_mask(h))) { in dequeue_huge_page_vma() 909 page = dequeue_huge_page_node(h, zone_to_nid(zone)); in dequeue_huge_page_vma() 1059 static bool zone_spans_last_pfn(const struct zone *zone, in zone_spans_last_pfn() argument 1063 return zone_spans_pfn(zone, last_pfn); in zone_spans_last_pfn() 1070 struct zone *z; in alloc_gigantic_page()
|
D | filemap.c | 740 const struct zone *zone = page_zone(page); in page_waitqueue() local 742 return &zone->wait_table[hash_ptr(page, zone->wait_table_bits)]; in page_waitqueue()
|
D | slab.c | 3007 struct zone *zone; in fallback_alloc() local 3027 for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { in fallback_alloc() 3028 nid = zone_to_nid(zone); in fallback_alloc() 3030 if (cpuset_zone_allowed(zone, flags) && in fallback_alloc()
|
D | bootmem.c | 250 struct zone *z; in reset_node_managed_pages()
|
D | slub.c | 1714 struct zone *zone; in get_any_partial() local 1744 for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { in get_any_partial() 1747 n = get_node(s, zone_to_nid(zone)); in get_any_partial() 1749 if (n && cpuset_zone_allowed(zone, flags) && in get_any_partial()
|
/linux-4.4.14/include/linux/ |
D | vmstat.h | 104 #define __count_zone_vm_events(item, zone, delta) \ argument 106 zone_idx(zone), delta) 113 static inline void zone_page_state_add(long x, struct zone *zone, in zone_page_state_add() argument 116 atomic_long_add(x, &zone->vm_stat[item]); in zone_page_state_add() 130 static inline unsigned long zone_page_state(struct zone *zone, in zone_page_state() argument 133 long x = atomic_long_read(&zone->vm_stat[item]); in zone_page_state() 147 static inline unsigned long zone_page_state_snapshot(struct zone *zone, in zone_page_state_snapshot() argument 150 long x = atomic_long_read(&zone->vm_stat[item]); in zone_page_state_snapshot() 155 x += per_cpu_ptr(zone->pageset, cpu)->vm_stat_diff[item]; in zone_page_state_snapshot() 166 extern void zone_statistics(struct zone *, struct zone *, gfp_t gfp); [all …]
|
D | memory_hotplug.h | 10 struct zone; 63 static inline unsigned zone_span_seqbegin(struct zone *zone) in zone_span_seqbegin() argument 65 return read_seqbegin(&zone->span_seqlock); in zone_span_seqbegin() 67 static inline int zone_span_seqretry(struct zone *zone, unsigned iv) in zone_span_seqretry() argument 69 return read_seqretry(&zone->span_seqlock, iv); in zone_span_seqretry() 71 static inline void zone_span_writelock(struct zone *zone) in zone_span_writelock() argument 73 write_seqlock(&zone->span_seqlock); in zone_span_writelock() 75 static inline void zone_span_writeunlock(struct zone *zone) in zone_span_writeunlock() argument 77 write_sequnlock(&zone->span_seqlock); in zone_span_writeunlock() 79 static inline void zone_seqlock_init(struct zone *zone) in zone_seqlock_init() argument [all …]
|
D | mmzone.h | 220 struct zone *zone; member 331 struct zone { struct 549 static inline unsigned long zone_end_pfn(const struct zone *zone) in zone_end_pfn() argument 551 return zone->zone_start_pfn + zone->spanned_pages; in zone_end_pfn() 554 static inline bool zone_spans_pfn(const struct zone *zone, unsigned long pfn) in zone_spans_pfn() argument 556 return zone->zone_start_pfn <= pfn && pfn < zone_end_pfn(zone); in zone_spans_pfn() 559 static inline bool zone_is_initialized(struct zone *zone) in zone_is_initialized() argument 561 return !!zone->wait_table; in zone_is_initialized() 564 static inline bool zone_is_empty(struct zone *zone) in zone_is_empty() argument 566 return zone->spanned_pages == 0; in zone_is_empty() [all …]
|
D | compaction.h | 40 extern int fragmentation_index(struct zone *zone, unsigned int order); 46 extern unsigned long compaction_suitable(struct zone *zone, int order, 49 extern void defer_compaction(struct zone *zone, int order); 50 extern bool compaction_deferred(struct zone *zone, int order); 51 extern void compaction_defer_reset(struct zone *zone, int order, 53 extern bool compaction_restarting(struct zone *zone, int order); 72 static inline unsigned long compaction_suitable(struct zone *zone, int order, in compaction_suitable() argument 78 static inline void defer_compaction(struct zone *zone, int order) in defer_compaction() argument 82 static inline bool compaction_deferred(struct zone *zone, int order) in compaction_deferred() argument
|
D | page-isolation.h | 5 static inline bool has_isolate_pageblock(struct zone *zone) in has_isolate_pageblock() argument 7 return zone->nr_isolate_pageblock; in has_isolate_pageblock() 18 static inline bool has_isolate_pageblock(struct zone *zone) in has_isolate_pageblock() argument 32 bool has_unmovable_pages(struct zone *zone, struct page *page, int count, 35 int move_freepages_block(struct zone *zone, struct page *page, 37 int move_freepages(struct zone *zone,
|
D | memcontrol.h | 56 struct zone *zone; member 306 struct lruvec *mem_cgroup_zone_lruvec(struct zone *, struct mem_cgroup *); 307 struct lruvec *mem_cgroup_page_lruvec(struct page *, struct zone *); 465 unsigned long mem_cgroup_soft_limit_reclaim(struct zone *zone, int order, 545 static inline struct lruvec *mem_cgroup_zone_lruvec(struct zone *zone, in mem_cgroup_zone_lruvec() argument 548 return &zone->lruvec; in mem_cgroup_zone_lruvec() 552 struct zone *zone) in mem_cgroup_page_lruvec() argument 554 return &zone->lruvec; in mem_cgroup_page_lruvec() 657 unsigned long mem_cgroup_soft_limit_reclaim(struct zone *zone, int order, in mem_cgroup_soft_limit_reclaim() argument
|
D | gfp.h | 506 void drain_zone_pages(struct zone *zone, struct per_cpu_pages *pcp); 507 void drain_all_pages(struct zone *zone); 508 void drain_local_pages(struct zone *zone);
|
D | swap.h | 118 struct zone; 328 struct zone *zone, 339 extern int zone_reclaim(struct zone *, gfp_t, unsigned int); 342 static inline int zone_reclaim(struct zone *z, gfp_t mask, unsigned int order) in zone_reclaim()
|
D | cpuset.h | 58 static inline int cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed() 180 static inline int cpuset_zone_allowed(struct zone *z, gfp_t gfp_mask) in cpuset_zone_allowed()
|
D | mm.h | 709 static inline int zone_to_nid(struct zone *zone) in zone_to_nid() argument 712 return zone->node; in zone_to_nid() 840 static inline struct zone *page_zone(const struct page *page) in page_zone() 858 static inline void set_page_zone(struct page *page, enum zone_type zone) in set_page_zone() argument 861 page->flags |= (zone & ZONES_MASK) << ZONES_PGSHIFT; in set_page_zone() 870 static inline void set_page_links(struct page *page, enum zone_type zone, in set_page_links() argument 873 set_page_zone(page, zone); in set_page_links() 1820 extern void zone_pcp_update(struct zone *zone); 1821 extern void zone_pcp_reset(struct zone *zone);
|
D | suspend.h | 298 extern void mark_free_pages(struct zone *zone);
|
D | writeback.h | 323 bool zone_dirty_ok(struct zone *zone);
|
D | backing-dev.h | 200 long wait_iff_congested(struct zone *zone, int sync, long timeout);
|
/linux-4.4.14/include/net/netfilter/ |
D | nf_conntrack_zones.h | 21 nf_ct_zone_init(struct nf_conntrack_zone *zone, u16 id, u8 dir, u8 flags) in nf_ct_zone_init() argument 23 zone->id = id; in nf_ct_zone_init() 24 zone->flags = flags; in nf_ct_zone_init() 25 zone->dir = dir; in nf_ct_zone_init() 27 return zone; in nf_ct_zone_init() 34 const struct nf_conntrack_zone *zone; in nf_ct_zone_tmpl() local 39 zone = nf_ct_zone(tmpl); in nf_ct_zone_tmpl() 40 if (zone->flags & NF_CT_FLAG_MARK) in nf_ct_zone_tmpl() 41 zone = nf_ct_zone_init(tmp, skb->mark, zone->dir, 0); in nf_ct_zone_tmpl() 43 return zone; in nf_ct_zone_tmpl() [all …]
|
D | nf_conntrack_expect.h | 82 const struct nf_conntrack_zone *zone, 87 const struct nf_conntrack_zone *zone, 92 const struct nf_conntrack_zone *zone,
|
D | nf_conntrack.h | 255 const struct nf_conntrack_zone *zone, 296 const struct nf_conntrack_zone *zone,
|
D | nf_conntrack_core.h | 57 const struct nf_conntrack_zone *zone,
|
/linux-4.4.14/fs/btrfs/ |
D | reada.c | 253 struct reada_zone *zone; in reada_find_zone() local 259 zone = NULL; in reada_find_zone() 261 ret = radix_tree_gang_lookup(&dev->reada_zones, (void **)&zone, in reada_find_zone() 264 kref_get(&zone->refcnt); in reada_find_zone() 268 if (logical >= zone->start && logical < zone->end) in reada_find_zone() 269 return zone; in reada_find_zone() 271 kref_put(&zone->refcnt, reada_zone_release); in reada_find_zone() 283 zone = kzalloc(sizeof(*zone), GFP_NOFS); in reada_find_zone() 284 if (!zone) in reada_find_zone() 287 zone->start = start; in reada_find_zone() [all …]
|
/linux-4.4.14/kernel/power/ |
D | snapshot.c | 299 struct mem_zone_bm_rtree *zone; member 357 static int add_rtree_block(struct mem_zone_bm_rtree *zone, gfp_t gfp_mask, in add_rtree_block() argument 364 block_nr = zone->blocks; in add_rtree_block() 374 for (i = zone->levels; i < levels_needed; i++) { in add_rtree_block() 376 &zone->nodes); in add_rtree_block() 380 node->data[0] = (unsigned long)zone->rtree; in add_rtree_block() 381 zone->rtree = node; in add_rtree_block() 382 zone->levels += 1; in add_rtree_block() 386 block = alloc_rtree_node(gfp_mask, safe_needed, ca, &zone->leaves); in add_rtree_block() 391 node = zone->rtree; in add_rtree_block() [all …]
|
D | power.h | 134 extern unsigned int snapshot_additional_pages(struct zone *zone);
|
/linux-4.4.14/include/trace/events/ |
D | compaction.h | 252 TP_PROTO(struct zone *zone, 256 TP_ARGS(zone, order, ret), 266 __entry->nid = zone_to_nid(zone); 267 __entry->idx = zone_idx(zone); 281 TP_PROTO(struct zone *zone, 285 TP_ARGS(zone, order, ret) 290 TP_PROTO(struct zone *zone, 294 TP_ARGS(zone, order, ret) 300 TP_PROTO(struct zone *zone, int order), 302 TP_ARGS(zone, order), [all …]
|
/linux-4.4.14/drivers/net/ethernet/mellanox/mlx4/ |
D | alloc.c | 251 struct mlx4_zone_entry *zone = kmalloc(sizeof(*zone), GFP_KERNEL); in mlx4_zone_add_one() local 253 if (NULL == zone) in mlx4_zone_add_one() 256 zone->flags = flags; in mlx4_zone_add_one() 257 zone->bitmap = bitmap; in mlx4_zone_add_one() 258 zone->use_rr = (flags & MLX4_ZONE_USE_RR) ? MLX4_USE_RR : 0; in mlx4_zone_add_one() 259 zone->priority = priority; in mlx4_zone_add_one() 260 zone->offset = offset; in mlx4_zone_add_one() 264 zone->uid = zone_alloc->last_uid++; in mlx4_zone_add_one() 265 zone->allocator = zone_alloc; in mlx4_zone_add_one() 275 list_add_tail(&zone->prio_list, &it->prio_list); in mlx4_zone_add_one() [all …]
|
/linux-4.4.14/drivers/mtd/ |
D | sm_ftl.c | 195 static loff_t sm_mkoffset(struct sm_ftl *ftl, int zone, int block, int boffset) in sm_mkoffset() argument 198 WARN_ON(zone < 0 || zone >= ftl->zone_count); in sm_mkoffset() 205 return (zone * SM_MAX_ZONE_SIZE + block) * ftl->block_size + boffset; in sm_mkoffset() 210 int *zone, int *block, int *boffset) in sm_break_offset() argument 214 *zone = offset >= ftl->zone_count ? -1 : offset; in sm_break_offset() 237 int zone, int block, int boffset, in sm_read_sector() argument 267 if (zone == 0 && block == ftl->cis_block && boffset == in sm_read_sector() 278 ret = mtd_read_oob(mtd, sm_mkoffset(ftl, zone, block, boffset), &ops); in sm_read_sector() 283 block, zone, ret); in sm_read_sector() 301 " as bad" , block, zone); in sm_read_sector() [all …]
|
/linux-4.4.14/drivers/thermal/int340x_thermal/ |
D | int340x_thermal_zone.c | 22 static int int340x_thermal_get_zone_temp(struct thermal_zone_device *zone, in int340x_thermal_get_zone_temp() argument 25 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_zone_temp() 30 return d->override_ops->get_temp(zone, temp); in int340x_thermal_get_zone_temp() 51 static int int340x_thermal_get_trip_temp(struct thermal_zone_device *zone, in int340x_thermal_get_trip_temp() argument 54 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_trip_temp() 58 return d->override_ops->get_trip_temp(zone, trip, temp); in int340x_thermal_get_trip_temp() 83 static int int340x_thermal_get_trip_type(struct thermal_zone_device *zone, in int340x_thermal_get_trip_type() argument 87 struct int34x_thermal_zone *d = zone->devdata; in int340x_thermal_get_trip_type() 91 return d->override_ops->get_trip_type(zone, trip, type); in int340x_thermal_get_trip_type() 116 static int int340x_thermal_set_trip_temp(struct thermal_zone_device *zone, in int340x_thermal_set_trip_temp() argument [all …]
|
D | int340x_thermal_zone.h | 40 struct thermal_zone_device *zone; member 65 thermal_zone_device_update(tzone->zone); in int340x_thermal_zone_device_update()
|
D | processor_thermal_device.c | 179 static int proc_thermal_get_zone_temp(struct thermal_zone_device *zone, in proc_thermal_get_zone_temp() argument
|
/linux-4.4.14/fs/adfs/ |
D | map.c | 192 scan_map(struct adfs_sb_info *asb, unsigned int zone, in scan_map() argument 199 dm = asb->s_map + zone; in scan_map() 200 zone = asb->s_map_size; in scan_map() 201 dm_end = asb->s_map + zone; in scan_map() 212 } while (--zone > 0); in scan_map() 235 unsigned int zone; in adfs_map_free() local 238 zone = asb->s_map_size; in adfs_map_free() 242 } while (--zone > 0); in adfs_map_free() 252 unsigned int zone, mapoff; in adfs_map_lookup() local 260 zone = asb->s_map_size >> 1; in adfs_map_lookup() [all …]
|
D | super.c | 305 int i, zone; in adfs_read_map() local 322 for (zone = 0; zone < nzones; zone++, map_addr++) { in adfs_read_map() 323 dm[zone].dm_startbit = 0; in adfs_read_map() 324 dm[zone].dm_endbit = zone_size; in adfs_read_map() 325 dm[zone].dm_startblk = zone * zone_size - ADFS_DR_SIZE_BITS; in adfs_read_map() 326 dm[zone].dm_bh = sb_bread(sb, map_addr); in adfs_read_map() 328 if (!dm[zone].dm_bh) { in adfs_read_map() 335 i = zone - 1; in adfs_read_map() 348 while (--zone >= 0) in adfs_read_map() 349 brelse(dm[zone].dm_bh); in adfs_read_map()
|
/linux-4.4.14/lib/ |
D | show_mem.c | 26 struct zone *zone = &pgdat->node_zones[zoneid]; in show_mem() local 27 if (!populated_zone(zone)) in show_mem() 30 total += zone->present_pages; in show_mem() 31 reserved += zone->present_pages - zone->managed_pages; in show_mem() 34 highmem += zone->present_pages; in show_mem()
|
/linux-4.4.14/Documentation/thermal/ |
D | sysfs-api.txt | 13 The generic thermal sysfs provides a set of interfaces for thermal zone 17 This how-to focuses on enabling new thermal zone and cooling devices to 19 This solution is platform independent and any type of thermal zone devices 22 The main task of the thermal sysfs driver is to expose thermal zone attributes 25 inputs from thermal zone attributes (the current temperature and trip point 33 1.1 thermal zone device interface 40 This interface function adds a new thermal zone device (sensor) to 44 type: the thermal zone type. 45 trips: the total number of trip points this thermal zone supports. 48 ops: thermal zone device call-backs. [all …]
|
D | x86_pkg_temperature_thermal | 19 zone with maximum two user mode configurable trip points. Number of trip points 27 Each package will register as a thermal zone under /sys/class/thermal. 42 Thermal notifications: To get kobject-uevent notifications, set the thermal zone
|
D | power_allocator.txt | 11 point of the thermal zone. 16 thermal zone. 63 provided while registering the thermal zone. This estimates the 76 thermal-zone. For example: 85 Instead, if the thermal zone is registered from the platform code, pass a 184 @tz: thermal zone in which we are currently operating 196 @tz: thermal zone in which we are currently operating 231 If the thermal zone is registered using 233 are passed as part of the thermal zone's `thermal_bind_parameters`.
|
D | exynos_thermal_emulation | 18 /sys/devices/virtual/thermal/thermal_zone'zone id'/emul_temp.
|
/linux-4.4.14/Documentation/ABI/testing/ |
D | sysfs-class-powercap | 25 This status affects every power zone using this "control_type. 27 What: /sys/class/powercap/<control type>/<power zone> 32 A power zone is a single or a collection of devices, which can 33 be independently monitored and controlled. A power zone sysfs 37 What: /sys/class/powercap/<control type>/<power zone>/<child power zone> 45 power zone for a whole CPU package, each CPU core in it can 46 be a child power zone. 48 What: /sys/class/powercap/.../<power zone>/name 53 Specifies the name of this power zone. 55 What: /sys/class/powercap/.../<power zone>/energy_uj [all …]
|
D | sysfs-class-backlight-driver-adp8870 | 1 What: /sys/class/backlight/<backlight>/<ambient light zone>_max 11 Control the maximum brightness for <ambient light zone> 14 <ambient light zone>. 16 What: /sys/class/backlight/<backlight>/<ambient light zone>_dim 25 Control the dim brightness for <ambient light zone> 29 this <ambient light zone>. 46 Get/Set current ambient light zone. Reading returns 49 to enter the corresponding ambient light zone.
|
D | sysfs-bus-iio-light-lm3533-als | 7 in each direction (rising|falling) and a zone change occurs. 8 The corresponding light zone can be read from 32 These thresholds correspond to the eight zone-boundary 41 Get the current light zone (0..4) as defined by the 50 out_currentY_currentZ_raw, where Z is the current zone. 57 Set the output current for channel out_currentY when in zone
|
D | sysfs-devices-memory | 69 read-only and is designed to show which zone this memory
|
/linux-4.4.14/drivers/hwmon/ |
D | scpi-hwmon.c | 51 struct scpi_thermal_zone *zone = dev; in scpi_read_temp() local 52 struct scpi_sensors *scpi_sensors = zone->scpi_sensors; in scpi_read_temp() 54 struct sensor_data *sensor = &scpi_sensors->data[zone->sensor_id]; in scpi_read_temp() 102 struct scpi_thermal_zone *zone; in unregister_thermal_zones() local 104 zone = list_entry(pos, struct scpi_thermal_zone, list); in unregister_thermal_zones() 105 thermal_zone_of_sensor_unregister(&pdev->dev, zone->tzd); in unregister_thermal_zones() 227 struct scpi_thermal_zone *zone; in scpi_hwmon_probe() local 232 zone = devm_kzalloc(dev, sizeof(*zone), GFP_KERNEL); in scpi_hwmon_probe() 233 if (!zone) { in scpi_hwmon_probe() 238 zone->sensor_id = i; in scpi_hwmon_probe() [all …]
|
D | lm85.c | 234 static int ZONE_TO_REG(int zone) in ZONE_TO_REG() argument 239 if (zone == lm85_zone_map[i]) in ZONE_TO_REG() 323 struct lm85_zone zone[3]; member 516 data->zone[i].range = val >> 4; in lm85_update_device() 519 data->zone[i].limit = in lm85_update_device() 521 data->zone[i].critical = in lm85_update_device() 527 data->zone[i].limit -= 64; in lm85_update_device() 528 data->zone[i].critical -= 64; in lm85_update_device() 539 data->zone[0].hyst = i >> 4; in lm85_update_device() 540 data->zone[1].hyst = i & 0x0f; in lm85_update_device() [all …]
|
D | dme1737.c | 1562 static SENSOR_DEVICE_ATTR_2(zone##ix##_auto_channels_temp, S_IRUGO, \ 1564 static SENSOR_DEVICE_ATTR_2(zone##ix##_auto_point1_temp_hyst, S_IRUGO, \ 1566 static SENSOR_DEVICE_ATTR_2(zone##ix##_auto_point1_temp, S_IRUGO, \ 1568 static SENSOR_DEVICE_ATTR_2(zone##ix##_auto_point2_temp, S_IRUGO, \ 1570 static SENSOR_DEVICE_ATTR_2(zone##ix##_auto_point3_temp, S_IRUGO, \
|
/linux-4.4.14/drivers/staging/rts5208/ |
D | xd.c | 879 xd_card->zone = vmalloc(size); in xd_init_l2p_tbl() 880 if (!xd_card->zone) { in xd_init_l2p_tbl() 886 xd_card->zone[i].build_flag = 0; in xd_init_l2p_tbl() 887 xd_card->zone[i].l2p_table = NULL; in xd_init_l2p_tbl() 888 xd_card->zone[i].free_table = NULL; in xd_init_l2p_tbl() 889 xd_card->zone[i].get_index = 0; in xd_init_l2p_tbl() 890 xd_card->zone[i].set_index = 0; in xd_init_l2p_tbl() 891 xd_card->zone[i].unused_blk_cnt = 0; in xd_init_l2p_tbl() 897 static inline void free_zone(struct zone_entry *zone) in free_zone() argument 899 if (!zone) in free_zone() [all …]
|
D | rtsx_chip.h | 529 struct zone_entry *zone; member
|
/linux-4.4.14/sound/synth/emux/ |
D | soundfont.c | 525 struct snd_sf_zone *zone; in load_info() local 563 for (zone = sf->zones; zone; zone = zone->next) { in load_info() 564 if (!zone->mapped && in load_info() 565 zone->bank == hdr.bank && in load_info() 566 zone->instr == hdr.instr) in load_info() 595 if ((zone = sf_zone_new(sflist, sf)) == NULL) { in load_info() 600 zone->bank = tmpzone.bank; in load_info() 601 zone->instr = tmpzone.instr; in load_info() 602 zone->v = tmpzone.v; in load_info() 605 zone->sample = set_sample(sf, &zone->v); in load_info() [all …]
|
D | emux_synth.c | 115 vp->zone = table[i]; in snd_emux_note_on() 116 if (vp->zone->sample) in snd_emux_note_on() 117 vp->block = vp->zone->sample->block; in snd_emux_note_on() 513 vp->zone = NULL; in terminate_voice() 566 vp->reg = vp->zone->v; in setup_voice()
|
D | emux_effect.c | 215 origp = (unsigned char*)&vp->zone->v.parm + offset; in snd_emux_send_effect()
|
/linux-4.4.14/virt/kvm/ |
D | coalesced_mmio.c | 35 if (addr < dev->zone.addr) in coalesced_mmio_in_range() 37 if (addr + len > dev->zone.addr + dev->zone.size) in coalesced_mmio_in_range() 137 struct kvm_coalesced_mmio_zone *zone) in kvm_vm_ioctl_register_coalesced_mmio() argument 148 dev->zone = *zone; in kvm_vm_ioctl_register_coalesced_mmio() 151 ret = kvm_io_bus_register_dev(kvm, KVM_MMIO_BUS, zone->addr, in kvm_vm_ioctl_register_coalesced_mmio() 152 zone->size, &dev->dev); in kvm_vm_ioctl_register_coalesced_mmio() 168 struct kvm_coalesced_mmio_zone *zone) in kvm_vm_ioctl_unregister_coalesced_mmio() argument 175 if (coalesced_mmio_in_range(dev, zone->addr, zone->size)) { in kvm_vm_ioctl_unregister_coalesced_mmio()
|
D | coalesced_mmio.h | 21 struct kvm_coalesced_mmio_zone zone; member 27 struct kvm_coalesced_mmio_zone *zone); 29 struct kvm_coalesced_mmio_zone *zone);
|
D | kvm_main.c | 2783 struct kvm_coalesced_mmio_zone zone; in kvm_vm_ioctl() local 2786 if (copy_from_user(&zone, argp, sizeof(zone))) in kvm_vm_ioctl() 2788 r = kvm_vm_ioctl_register_coalesced_mmio(kvm, &zone); in kvm_vm_ioctl() 2792 struct kvm_coalesced_mmio_zone zone; in kvm_vm_ioctl() local 2795 if (copy_from_user(&zone, argp, sizeof(zone))) in kvm_vm_ioctl() 2797 r = kvm_vm_ioctl_unregister_coalesced_mmio(kvm, &zone); in kvm_vm_ioctl()
|
/linux-4.4.14/arch/x86/mm/ |
D | highmem_32.c | 109 struct zone *zone; in set_highmem_pages_init() local 117 for_each_zone(zone) { in set_highmem_pages_init() 120 if (!is_highmem(zone)) in set_highmem_pages_init() 123 zone_start_pfn = zone->zone_start_pfn; in set_highmem_pages_init() 124 zone_end_pfn = zone_start_pfn + zone->spanned_pages; in set_highmem_pages_init() 126 nid = zone_to_nid(zone); in set_highmem_pages_init() 128 zone->name, nid, zone_start_pfn, zone_end_pfn); in set_highmem_pages_init()
|
D | init_32.c | 829 struct zone *zone = pgdata->node_zones + in arch_add_memory() local 834 return __add_pages(nid, zone, start_pfn, nr_pages); in arch_add_memory() 842 struct zone *zone; in arch_remove_memory() local 844 zone = page_zone(pfn_to_page(start_pfn)); in arch_remove_memory() 845 return __remove_pages(zone, start_pfn, nr_pages); in arch_remove_memory()
|
D | init_64.c | 693 struct zone *zone = pgdat->node_zones + in arch_add_memory() local 701 ret = __add_pages(nid, zone, start_pfn, nr_pages); in arch_add_memory() 1021 struct zone *zone; in arch_remove_memory() local 1024 zone = page_zone(pfn_to_page(start_pfn)); in arch_remove_memory() 1026 ret = __remove_pages(zone, start_pfn, nr_pages); in arch_remove_memory()
|
/linux-4.4.14/net/sched/ |
D | act_connmark.c | 40 struct nf_conntrack_zone zone; in tcf_connmark() local 74 zone.id = ca->zone; in tcf_connmark() 75 zone.dir = NF_CT_DEFAULT_ZONE_DIR; in tcf_connmark() 77 thash = nf_conntrack_find_get(ca->net, &zone, &tuple); in tcf_connmark() 123 ci->zone = parm->zone; in tcf_connmark_init() 136 ci->zone = parm->zone; in tcf_connmark_init() 153 .zone = ci->zone, in tcf_connmark_dump()
|
/linux-4.4.14/fs/xfs/ |
D | kmem.h | 105 kmem_zone_free(kmem_zone_t *zone, void *ptr) in kmem_zone_free() argument 107 kmem_cache_free(zone, ptr); in kmem_zone_free() 111 kmem_zone_destroy(kmem_zone_t *zone) in kmem_zone_destroy() argument 113 if (zone) in kmem_zone_destroy() 114 kmem_cache_destroy(zone); in kmem_zone_destroy() 120 kmem_zone_zalloc(kmem_zone_t *zone, xfs_km_flags_t flags) in kmem_zone_zalloc() argument 122 return kmem_zone_alloc(zone, flags | KM_ZERO); in kmem_zone_zalloc()
|
D | kmem.c | 112 kmem_zone_alloc(kmem_zone_t *zone, xfs_km_flags_t flags) in kmem_zone_alloc() argument 119 ptr = kmem_cache_alloc(zone, lflags); in kmem_zone_alloc()
|
/linux-4.4.14/Documentation/vm/ |
D | balance | 17 is, only when needed (aka zone free memory is 0), instead of making it 26 of incurring the overhead of regular zone balancing. 31 would not be done even when the dma zone was completely empty. 2.2 has 36 In 2.3, zone balancing can be done in one of two ways: depending on the 37 zone size (and possibly of the size of lower class zones), we can decide 39 zone. The good part is, while balancing, we do not need to look at sizes 46 of a zone _and_ all its lower class zones falls below 1/64th of the 47 total memory in the zone and its lower class zones. This fixes the 2.2 54 Note that if the size of the regular zone is huge compared to dma zone, 56 deciding whether to balance the regular zone. The first solution [all …]
|
D | numa | 71 each memory zone [one or more of DMA, DMA32, NORMAL, HIGH_MEMORY, MOVABLE], 73 selected zone/node cannot satisfy the allocation request. This situation, 74 when a zone has no available memory to satisfy a request, is called 79 fall back to the same zone type on a different node, or to a different zone 82 a default zonelist order based on the sizes of the various zone types relative 93 nodes' zones in the selected zonelist looking for the first zone in the list 120 node the "local memory node"--the node of the first zone in CPU's node's
|
D | unevictable-lru.txt | 60 main memory will have over 32 million 4k pages in a single zone. When a large 82 The Unevictable LRU infrastructure consists of an additional, per-zone, LRU list 111 The unevictable list benefits from the "arrayification" of the per-zone LRU 115 unevictable pages are placed directly on the page's zone's unevictable list 116 under the zone lru_lock. This allows us to prevent the stranding of pages on 128 The memory controller data structure automatically gets a per-zone unevictable 129 list as a result of the "arrayification" of the per-zone LRU lists (one per 216 unevictable list for the zone being scanned. 349 is now mlocked and divert the page to the zone's unevictable list. If 411 mlocked, munlock_vma_page() updates that zone statistics for the number of [all …]
|
/linux-4.4.14/drivers/iio/light/ |
D | lm3533-als.c | 61 atomic_t zone; member 90 static int _lm3533_als_get_zone(struct iio_dev *indio_dev, u8 *zone) in _lm3533_als_get_zone() argument 103 *zone = min_t(u8, val, LM3533_ALS_ZONE_MAX); in _lm3533_als_get_zone() 108 static int lm3533_als_get_zone(struct iio_dev *indio_dev, u8 *zone) in lm3533_als_get_zone() argument 114 *zone = atomic_read(&als->zone); in lm3533_als_get_zone() 116 ret = _lm3533_als_get_zone(indio_dev, zone); in lm3533_als_get_zone() 128 static inline u8 lm3533_als_get_target_reg(unsigned channel, unsigned zone) in lm3533_als_get_target_reg() argument 130 return LM3533_REG_ALS_TARGET_BASE + 5 * channel + zone; in lm3533_als_get_target_reg() 134 unsigned zone, u8 *val) in lm3533_als_get_target() argument 143 if (zone > LM3533_ALS_ZONE_MAX) in lm3533_als_get_target() [all …]
|
D | Kconfig | 208 light zone through sysfs. A threshold event can be generated on zone 209 changes. The ALS-control output values can be set per zone for the
|
/linux-4.4.14/drivers/thermal/ |
D | rcar_thermal.c | 61 struct thermal_zone_device *zone; member 73 #define rcar_zone_to_priv(zone) ((zone)->devdata) argument 203 static int rcar_thermal_get_temp(struct thermal_zone_device *zone, int *temp) in rcar_thermal_get_temp() argument 205 struct rcar_thermal_priv *priv = rcar_zone_to_priv(zone); in rcar_thermal_get_temp() 217 static int rcar_thermal_get_trip_type(struct thermal_zone_device *zone, in rcar_thermal_get_trip_type() argument 220 struct rcar_thermal_priv *priv = rcar_zone_to_priv(zone); in rcar_thermal_get_trip_type() 236 static int rcar_thermal_get_trip_temp(struct thermal_zone_device *zone, in rcar_thermal_get_trip_temp() argument 239 struct rcar_thermal_priv *priv = rcar_zone_to_priv(zone); in rcar_thermal_get_trip_temp() 255 static int rcar_thermal_notify(struct thermal_zone_device *zone, in rcar_thermal_notify() argument 258 struct rcar_thermal_priv *priv = rcar_zone_to_priv(zone); in rcar_thermal_notify() [all …]
|
D | tegra_soctherm.c | 311 struct tegra_thermctl_zone *zone = data; in tegra_thermctl_get_temp() local 314 val = (readl(zone->reg) >> zone->shift) & SENSOR_TEMP_MASK; in tegra_thermctl_get_temp() 412 struct tegra_thermctl_zone *zone = in tegra_soctherm_probe() local 413 devm_kzalloc(&pdev->dev, sizeof(*zone), GFP_KERNEL); in tegra_soctherm_probe() 414 if (!zone) { in tegra_soctherm_probe() 419 zone->reg = tegra->regs + t124_thermctl_temp_zones[i].offset; in tegra_soctherm_probe() 420 zone->shift = t124_thermctl_temp_zones[i].shift; in tegra_soctherm_probe() 422 tz = thermal_zone_of_sensor_register(&pdev->dev, i, zone, in tegra_soctherm_probe()
|
D | of-thermal.c | 400 thermal_zone_of_add_sensor(struct device_node *zone, in thermal_zone_of_add_sensor() argument 407 tzd = thermal_zone_get_zone_by_name(zone->name); in thermal_zone_of_add_sensor() 885 struct thermal_zone_device *zone; in of_parse_thermal_zones() local 925 zone = thermal_zone_device_register(child->name, tz->ntrips, in of_parse_thermal_zones() 930 if (IS_ERR(zone)) { in of_parse_thermal_zones() 932 PTR_ERR(zone)); in of_parse_thermal_zones() 972 struct thermal_zone_device *zone; in of_thermal_destroy_zones() local 978 zone = thermal_zone_get_zone_by_name(child->name); in of_thermal_destroy_zones() 979 if (IS_ERR(zone)) in of_thermal_destroy_zones() 982 thermal_zone_device_unregister(zone); in of_thermal_destroy_zones() [all …]
|
D | Kconfig | 10 zone and cooling device. 11 Each thermal zone contains its own temperature, trip points, 51 thermal zone but the choice here takes precedence. 75 devices based on their 'contribution' to a zone. The 169 Enable this option to make a emul_temp sysfs node in thermal zone 245 management framework. A thermal zone with several trip points will be 247 thermal zone if trip points reached. 274 bound to thermal zone trip points. When a trip point reached, the 296 thermal zone. Each package will have its own thermal zone. There are 319 thermal zone. There are two trip points. One of the trip point can [all …]
|
/linux-4.4.14/arch/s390/mm/ |
D | page-states.c | 95 struct zone *zone; in arch_set_page_states() local 101 for_each_populated_zone(zone) { in arch_set_page_states() 102 spin_lock_irqsave(&zone->lock, flags); in arch_set_page_states() 104 list_for_each(l, &zone->free_area[order].free_list[t]) { in arch_set_page_states() 112 spin_unlock_irqrestore(&zone->lock, flags); in arch_set_page_states()
|
/linux-4.4.14/net/openvswitch/ |
D | conntrack.c | 48 struct nf_conntrack_zone zone; member 133 const struct nf_conntrack_zone *zone, in __ovs_ct_update_key() argument 137 key->ct.zone = zone->id; in __ovs_ct_update_key() 149 const struct nf_conntrack_zone *zone = &nf_ct_zone_dflt; in ovs_ct_update_key() local 161 zone = nf_ct_zone(ct); in ovs_ct_update_key() 165 zone = &info->zone; in ovs_ct_update_key() 167 __ovs_ct_update_key(key, state, zone, ct); in ovs_ct_update_key() 181 nla_put_u16(skb, OVS_KEY_ATTR_CT_ZONE, key->ct.zone)) in ovs_ct_put_key() 305 u16 zone, struct sk_buff *skb) in handle_fragments() argument 310 enum ip_defrag_users user = IP_DEFRAG_CONNTRACK_IN + zone; in handle_fragments() [all …]
|
D | conntrack.h | 78 key->ct.zone = 0; in ovs_ct_fill_key()
|
D | flow.h | 117 u16 zone; member
|
/linux-4.4.14/drivers/md/ |
D | raid0.c | 81 struct strip_zone *zone; in create_strip_zones() local 166 zone = &conf->strip_zone[0]; in create_strip_zones() 215 zone->nb_dev = cnt; in create_strip_zones() 216 zone->zone_end = smallest->sectors * cnt; in create_strip_zones() 218 curr_zone_end = zone->zone_end; in create_strip_zones() 225 zone = conf->strip_zone + i; in create_strip_zones() 229 zone->dev_start = smallest->sectors; in create_strip_zones() 235 if (rdev->sectors <= zone->dev_start) { in create_strip_zones() 255 zone->nb_dev = c; in create_strip_zones() 256 sectors = (smallest->sectors - zone->dev_start) * c; in create_strip_zones() [all …]
|
/linux-4.4.14/Documentation/devicetree/bindings/thermal/ |
D | thermal.txt | 38 the IC. See thermal zone binding for more details 106 get assigned to trip points of the zone. The cooling devices are expected 116 Type: phandle of zone. 120 - contribution: The cooling contribution to the thermal zone of the 123 of all cooling contributions within a thermal zone. 131 * Thermal zone nodes 133 The thermal zone node is the node containing all the required info 134 for describing a thermal zone, including its cooling device bindings. The 135 thermal zone node must contain, apart from its own properties, one sub-node 136 containing trip nodes and one sub-node containing all the zone cooling maps. [all …]
|
/linux-4.4.14/arch/ia64/mm/ |
D | init.c | 472 unsigned long zone; member 501 args->nid, args->zone, page_to_pfn(map_start), in virtual_memmap_init() 507 memmap_init (unsigned long size, int nid, unsigned long zone, in memmap_init() argument 511 memmap_init_zone(size, nid, zone, start_pfn, MEMMAP_EARLY); in memmap_init() 520 args.zone = zone; in memmap_init() 651 struct zone *zone; in arch_add_memory() local 658 zone = pgdat->node_zones + in arch_add_memory() 660 ret = __add_pages(nid, zone, start_pfn, nr_pages); in arch_add_memory() 674 struct zone *zone; in arch_remove_memory() local 677 zone = page_zone(pfn_to_page(start_pfn)); in arch_remove_memory() [all …]
|
/linux-4.4.14/Documentation/power/powercap/ |
D | powercap.txt | 19 given zone belongs to. They each contain attributes for monitoring power, as 24 parent power zone containing multiple subzones and so on to reflect the power 26 capping to a set of devices together using the parent power zone and if more 159 In addition to that, each power zone contains a name attribute, allowing the 160 part of the system represented by that zone to be identified. 167 with two different time windows to be applied to each power zone. Thus for 168 each zone there are 2 attributes representing the constraint names, 2 power 194 name (ro): Name of this power zone. 218 Common zone and control type attributes 220 enabled (rw): Enable/Disable controls at zone level or for all zones using [all …]
|
/linux-4.4.14/net/netfilter/ |
D | xt_CT.c | 203 struct nf_conntrack_zone zone; in xt_ct_tg_check() local 213 if (info->zone || info->flags & (XT_CT_ZONE_DIR_ORIG | in xt_ct_tg_check() 223 memset(&zone, 0, sizeof(zone)); in xt_ct_tg_check() 224 zone.id = info->zone; in xt_ct_tg_check() 225 zone.dir = xt_ct_flags_to_dir(info); in xt_ct_tg_check() 227 zone.flags |= NF_CT_FLAG_MARK; in xt_ct_tg_check() 229 ct = nf_ct_tmpl_alloc(par->net, &zone, GFP_KERNEL); in xt_ct_tg_check() 273 .zone = info->zone, in xt_ct_tg_check_v0() 356 .zone = info->zone, in xt_ct_tg_destroy_v0()
|
D | nf_conntrack_netlink.c | 132 const struct nf_conntrack_zone *zone, int dir) in ctnetlink_dump_zone_id() argument 134 if (zone->id == NF_CT_DEFAULT_ZONE_ID || zone->dir != dir) in ctnetlink_dump_zone_id() 136 if (nla_put_be16(skb, attrtype, htons(zone->id))) in ctnetlink_dump_zone_id() 475 const struct nf_conntrack_zone *zone; in ctnetlink_fill_info() local 491 zone = nf_ct_zone(ct); in ctnetlink_fill_info() 498 if (ctnetlink_dump_zone_id(skb, CTA_TUPLE_ZONE, zone, in ctnetlink_fill_info() 508 if (ctnetlink_dump_zone_id(skb, CTA_TUPLE_ZONE, zone, in ctnetlink_fill_info() 513 if (ctnetlink_dump_zone_id(skb, CTA_ZONE, zone, in ctnetlink_fill_info() 635 const struct nf_conntrack_zone *zone; in ctnetlink_conntrack_event() local 682 zone = nf_ct_zone(ct); in ctnetlink_conntrack_event() [all …]
|
D | nf_conntrack_core.c | 294 const struct nf_conntrack_zone *zone, in nf_ct_tmpl_alloc() argument 306 if (nf_ct_zone_add(tmpl, flags, zone) < 0) in nf_ct_tmpl_alloc() 430 const struct nf_conntrack_zone *zone) in nf_ct_key_equal() argument 438 nf_ct_zone_equal(ct, zone, NF_CT_DIRECTION(h)) && in nf_ct_key_equal() 448 ____nf_conntrack_find(struct net *net, const struct nf_conntrack_zone *zone, in ____nf_conntrack_find() argument 461 if (nf_ct_key_equal(h, tuple, zone)) { in ____nf_conntrack_find() 484 __nf_conntrack_find_get(struct net *net, const struct nf_conntrack_zone *zone, in __nf_conntrack_find_get() argument 492 h = ____nf_conntrack_find(net, zone, tuple, hash); in __nf_conntrack_find_get() 499 if (unlikely(!nf_ct_key_equal(h, tuple, zone))) { in __nf_conntrack_find_get() 511 nf_conntrack_find_get(struct net *net, const struct nf_conntrack_zone *zone, in nf_conntrack_find_get() argument [all …]
|
D | xt_connlimit.c | 137 const struct nf_conntrack_zone *zone, in check_hlist() argument 151 found = nf_conntrack_find_get(net, zone, &conn->tuple); in check_hlist() 204 u8 family, const struct nf_conntrack_zone *zone) in count_tree() argument 232 count = check_hlist(net, &rbconn->hhead, tuple, zone, &addit); in count_tree() 248 check_hlist(net, &rbconn->hhead, tuple, zone, &addit); in count_tree() 294 const struct nf_conntrack_zone *zone) in count_them() argument 310 count = count_tree(net, root, tuple, addr, mask, family, zone); in count_them() 325 const struct nf_conntrack_zone *zone = &nf_ct_zone_dflt; in connlimit_mt() local 333 zone = nf_ct_zone(ct); in connlimit_mt() 350 &info->mask, par->family, zone); in connlimit_mt()
|
D | nf_conntrack_standalone.c | 147 const struct nf_conntrack_zone *zone = nf_ct_zone(ct); in ct_show_zone() local 149 if (zone->dir != dir) in ct_show_zone() 151 switch (zone->dir) { in ct_show_zone() 153 seq_printf(s, "zone=%u ", zone->id); in ct_show_zone() 156 seq_printf(s, "zone-orig=%u ", zone->id); in ct_show_zone() 159 seq_printf(s, "zone-reply=%u ", zone->id); in ct_show_zone()
|
D | nf_nat_core.c | 188 const struct nf_conntrack_zone *zone, in find_appropriate_src() argument 202 nf_ct_zone_equal(ct, zone, IP_CT_DIR_ORIGINAL)) { in find_appropriate_src() 222 find_best_ips_proto(const struct nf_conntrack_zone *zone, in find_best_ips_proto() argument 263 0 : (__force u32)tuple->dst.u3.all[max] ^ zone->id); in find_best_ips_proto() 302 const struct nf_conntrack_zone *zone; in get_unique_tuple() local 307 zone = nf_ct_zone(ct); in get_unique_tuple() 330 } else if (find_appropriate_src(net, zone, l3proto, l4proto, in get_unique_tuple() 340 find_best_ips_proto(zone, tuple, range, ct, maniptype); in get_unique_tuple()
|
D | nf_conntrack_expect.c | 92 const struct nf_conntrack_zone *zone, in __nf_ct_expect_find() argument 104 nf_ct_zone_equal_any(i->master, zone)) in __nf_ct_expect_find() 114 const struct nf_conntrack_zone *zone, in nf_ct_expect_find_get() argument 120 i = __nf_ct_expect_find(net, zone, tuple); in nf_ct_expect_find_get() 133 const struct nf_conntrack_zone *zone, in nf_ct_find_expectation() argument 146 nf_ct_zone_equal_any(i->master, zone)) { in nf_ct_find_expectation()
|
D | nf_conntrack_pptp.c | 146 const struct nf_conntrack_zone *zone; in destroy_sibling_or_exp() local 153 zone = nf_ct_zone(ct); in destroy_sibling_or_exp() 154 h = nf_conntrack_find_get(net, zone, t); in destroy_sibling_or_exp() 165 exp = nf_ct_expect_find_get(net, zone, t); in destroy_sibling_or_exp()
|
/linux-4.4.14/drivers/uwb/ |
D | drp-ie.c | 269 void uwb_drp_ie_single_zone_to_bm(struct uwb_mas_bm *bm, u8 zone, u16 mas_bm) in uwb_drp_ie_single_zone_to_bm() argument 277 set_bit(zone * UWB_NUM_ZONES + mas, bm->bm); in uwb_drp_ie_single_zone_to_bm() 302 u8 zone; in uwb_drp_ie_to_bm() local 311 for (zone = 0; zone < UWB_NUM_ZONES; zone++) { in uwb_drp_ie_to_bm() 312 zone_mask = 1 << zone; in uwb_drp_ie_to_bm() 314 uwb_drp_ie_single_zone_to_bm(bm, zone, mas_bm); in uwb_drp_ie_to_bm()
|
/linux-4.4.14/drivers/usb/storage/ |
D | alauda.c | 554 unsigned int zone) in alauda_find_unused_pba() argument 556 u16 *pba_to_lba = info->pba_to_lba[zone]; in alauda_find_unused_pba() 561 return (zone << info->zoneshift) + i; in alauda_find_unused_pba() 570 static int alauda_read_map(struct us_data *us, unsigned int zone) in alauda_read_map() argument 578 unsigned int zone_base_lba = zone * uzonesize; in alauda_read_map() 579 unsigned int zone_base_pba = zone * zonesize; in alauda_read_map() 587 usb_stor_dbg(us, "Mapping blocks for zone %d\n", zone); in alauda_read_map() 678 MEDIA_INFO(us).lba_to_pba[zone] = lba_to_pba; in alauda_read_map() 679 MEDIA_INFO(us).pba_to_lba[zone] = pba_to_lba; in alauda_read_map() 694 static void alauda_ensure_map_for_zone(struct us_data *us, unsigned int zone) in alauda_ensure_map_for_zone() argument [all …]
|
D | sddr55.c | 721 int zone = i / 1024; in sddr55_read_map() local 747 if (info->lba_to_pba[lba + zone * 1000] != NOT_ALLOCATED && in sddr55_read_map() 751 lba + zone * 1000); in sddr55_read_map() 758 info->lba_to_pba[lba + zone * 1000] = i; in sddr55_read_map()
|
/linux-4.4.14/arch/powerpc/mm/ |
D | mem.c | 119 struct zone *zone; in arch_add_memory() local 130 zone = pgdata->node_zones + in arch_add_memory() 133 return __add_pages(nid, zone, start_pfn, nr_pages); in arch_add_memory() 141 struct zone *zone; in arch_remove_memory() local 144 zone = page_zone(pfn_to_page(start_pfn)); in arch_remove_memory() 145 ret = __remove_pages(zone, start_pfn, nr_pages); in arch_remove_memory() 242 void __init limit_zone_pfn(enum zone_type zone, unsigned long pfn_limit) in limit_zone_pfn() argument 249 for (i = zone; i >= 0; i--) { in limit_zone_pfn()
|
/linux-4.4.14/Documentation/cma/ |
D | debugfs.txt | 4 Each CMA zone represents a directory under <debugfs>/cma/, indexed by the 5 kernel's CMA index. So the first CMA zone would be: 11 - [RO] base_pfn: The base PFN (Page Frame Number) of the zone. 14 - [RO] bitmap: The bitmap of page states in the zone.
|
/linux-4.4.14/Documentation/networking/ |
D | cops.txt | 52 dummy -seed -phase 2 -net 2000 -addr 2000.10 -zone "1033" 53 lt0 -seed -phase 1 -net 1000 -addr 1000.50 -zone "1033" 56 eth0 -seed -phase 2 -net 3000 -addr 3000.20 -zone "1033" 57 lt0 -seed -phase 1 -net 1000 -addr 1000.50 -zone "1033" 61 lt0 -seed -phase 1 -net 1000 -addr 1000.10 -zone "LocalTalk1" 62 lt1 -seed -phase 1 -net 2000 -addr 2000.20 -zone "LocalTalk2" 63 eth0 -seed -phase 2 -net 3000 -addr 3000.30 -zone "EtherTalk"
|
D | ltpc.txt | 47 dummy -seed -phase 2 -net 2000 -addr 2000.26 -zone "1033" 48 lt0 -seed -phase 1 -net 1033 -addr 1033.27 -zone "1033"
|
D | fib_trie.txt | 105 as one "zone"/hash per prefix length.
|
D | netlink_mmap.txt | 83 the ring just contains the frames in a continuous memory zone.
|
/linux-4.4.14/arch/m68k/mm/ |
D | mcfmmu.c | 42 enum zone_type zone; in paging_init() local 78 for (zone = 0; zone < MAX_NR_ZONES; zone++) in paging_init() 79 zones_size[zone] = 0x0; in paging_init()
|
/linux-4.4.14/drivers/base/ |
D | memory.c | 393 struct zone *zone; in show_valid_zones() local 403 zone = page_zone(first_page); in show_valid_zones() 405 if (zone_idx(zone) == ZONE_MOVABLE - 1) { in show_valid_zones() 407 if (end_pfn == zone_end_pfn(zone)) in show_valid_zones() 409 zone->name, (zone + 1)->name); in show_valid_zones() 412 if (zone_idx(zone) == ZONE_MOVABLE) { in show_valid_zones() 414 if (start_pfn == zone->zone_start_pfn) in show_valid_zones() 416 zone->name, (zone - 1)->name); in show_valid_zones() 419 return sprintf(buf, "%s\n", zone->name); in show_valid_zones()
|
/linux-4.4.14/arch/tile/mm/ |
D | pgtable.c | 45 struct zone *zone; in show_mem() local 64 for_each_zone(zone) { in show_mem() 67 if (!populated_zone(zone)) in show_mem() 70 spin_lock_irqsave(&zone->lock, flags); in show_mem() 72 int nr = zone->free_area[order].nr_free; in show_mem() 77 spin_unlock_irqrestore(&zone->lock, flags); in show_mem() 79 zone_to_nid(zone), zone->name, in show_mem()
|
D | init.c | 696 struct zone *z; in set_non_bootmem_pages_init() 869 struct zone *zone = pgdata->node_zones + MAX_NR_ZONES-1; in arch_add_memory() local 873 return __add_pages(zone, start_pfn, nr_pages); in arch_add_memory()
|
/linux-4.4.14/Documentation/hwmon/ |
D | dme1737 | 165 be retrieved via the zone[1-3]_auto_channels_temp attributes. Each PWM output 166 is assigned to one (or hottest of multiple) temperature zone(s) through the 171 programmed via zone[1-3]_auto_point[1-3]_temp and 172 zone[1-3]_auto_point1_temp_hyst: 178 zone[1-3]_auto_point3_temp full-speed temp (all outputs) 179 zone[1-3]_auto_point2_temp full-speed temp 180 zone[1-3]_auto_point1_temp low-speed temp 181 zone[1-3]_auto_point1_temp_hyst min-speed temp 184 to auto_point2_pwm if the temperature of the associated zone is between 193 based on the temperature of the associated thermal zone: [all …]
|
D | asc7621 | 17 added PECI and a 4th thermal zone. The Andigilog aSC7611 is the 47 temperature may be mapped to any zone, which has a default assignment 50 Since there is a fan to zone assignment that allows for the "hotter" of 52 indication to the user, we have added an indicator that shows which zone 62 address. One could use our flexible zone assignment to have a zone 181 When the corresponding zone temperature reaches this value, 195 if the temperature for the zone is below the minimum. 197 for the zone is below the minimum.
|
D | lm85 | 170 sensors and therefore three zones (# 1, 2 and 3). Each zone has the following 182 configured and assigned to a zone for its control value. Each PWM can be 195 PWMs controlled by zone '#'. 199 * pwm#_auto_channels - controls zone that is associated with PWM
|
D | f71882fg | 93 Note that the lowest numbered temperature zone trip point corresponds to 95 vica versa. So the temperature zone trip points 1-4 (or 1-2) go from high temp
|
/linux-4.4.14/Documentation/sysctl/ |
D | vm.txt | 229 debugfs shows what the fragmentation index for each order is in each zone in 234 The kernel will not compact memory in a zone if the 288 zone. This is because that memory could then be pinned via the mlock() 316 zone's value is not necessary for following calculation. 321 Each zone has an array of protection pages like this. 324 Node 0, zone DMA 338 These protections are added to score to judge whether this zone should be used 341 In this example, if normal pages (index=2) are required to this DMA zone and 342 watermark[WMARK_HIGH] is used for watermark, the kernel judges this zone should 344 (4 + 2004 = 2008). If this protection value is 0, this zone would be used for [all …]
|
/linux-4.4.14/fs/sysv/ |
D | balloc.c | 187 sysv_zone_t zone; in sysv_count_free_blocks() local 190 zone = 0; in sysv_count_free_blocks() 191 while (n && (zone = blocks[--n]) != 0) in sysv_count_free_blocks() 193 if (zone == 0) in sysv_count_free_blocks() 196 block = fs32_to_cpu(sbi, zone); in sysv_count_free_blocks()
|
/linux-4.4.14/fs/proc/ |
D | meminfo.c | 35 struct zone *zone; in meminfo_proc_show() local 54 for_each_zone(zone) in meminfo_proc_show() 55 wmark_low += zone->watermark[WMARK_LOW]; in meminfo_proc_show()
|
/linux-4.4.14/fs/minix/ |
D | bitmap.c | 47 unsigned long bit, zone; in minix_free_block() local 53 zone = block - sbi->s_firstdatazone + 1; in minix_free_block() 54 bit = zone & ((1<<k) - 1); in minix_free_block() 55 zone >>= k; in minix_free_block() 56 if (zone >= sbi->s_zmap_blocks) { in minix_free_block() 60 bh = sbi->s_zmap[zone]; in minix_free_block()
|
/linux-4.4.14/drivers/net/fjes/ |
D | fjes_hw.c | 683 (hw->ep_shm_info[hw->my_epid].zone == in fjes_hw_epid_is_same_zone() 687 return (hw->ep_shm_info[epid].zone == in fjes_hw_epid_is_same_zone() 688 hw->ep_shm_info[hw->my_epid].zone); in fjes_hw_epid_is_same_zone() 929 struct my_s {u8 es_status; u8 zone; } *info; in fjes_hw_update_zone_task() member 967 hw->ep_shm_info[epidx].zone = in fjes_hw_update_zone_task() 968 info[epidx].zone; in fjes_hw_update_zone_task() 976 if ((info[epidx].zone != in fjes_hw_update_zone_task() 980 (info[epidx].zone == in fjes_hw_update_zone_task() 981 info[hw->my_epid].zone)) in fjes_hw_update_zone_task() 989 if ((info[epidx].zone == in fjes_hw_update_zone_task() [all …]
|
D | fjes_hw.h | 149 u8 zone; member 244 u8 zone; member
|
D | fjes_main.c | 343 hw->ep_shm_info[epidx].zone = in fjes_setup_resources() 344 hw->hw_info.res_buf->info.info[epidx].zone; in fjes_setup_resources()
|
/linux-4.4.14/include/uapi/linux/netfilter/ |
D | xt_CT.h | 20 __u16 zone; member 31 __u16 zone; member
|
/linux-4.4.14/arch/parisc/mm/ |
D | init.c | 670 struct zone *zone = &pgdat->node_zones[zoneid]; in show_mem() local 671 if (!populated_zone(zone)) in show_mem() 674 total += zone->present_pages; in show_mem() 675 reserved = zone->present_pages - zone->managed_pages; in show_mem() 692 struct zone *zone; in show_mem() local 695 for_each_zone_zonelist(zone, z, zl, j) in show_mem() 696 printk("[%d/%s] ", zone_to_nid(zone), in show_mem() 697 zone->name); in show_mem()
|
/linux-4.4.14/fs/ntfs/ |
D | lcnalloc.c | 148 const NTFS_CLUSTER_ALLOCATION_ZONES zone, in ntfs_cluster_alloc() argument 167 zone == MFT_ZONE ? "MFT" : "DATA"); in ntfs_cluster_alloc() 174 BUG_ON(zone < FIRST_ZONE); in ntfs_cluster_alloc() 175 BUG_ON(zone > LAST_ZONE); in ntfs_cluster_alloc() 202 if (zone == DATA_ZONE) in ntfs_cluster_alloc() 213 } else if (zone == DATA_ZONE && zone_start >= vol->mft_zone_start && in ntfs_cluster_alloc() 221 } else if (zone == MFT_ZONE && (zone_start < vol->mft_zone_start || in ntfs_cluster_alloc() 232 if (zone == MFT_ZONE) { in ntfs_cluster_alloc() 696 if (zone == MFT_ZONE || mft_zone_size <= 0) { in ntfs_cluster_alloc()
|
D | lcnalloc.h | 45 const NTFS_CLUSTER_ALLOCATION_ZONES zone,
|
/linux-4.4.14/sound/soc/ |
D | soc-jack.c | 159 struct snd_soc_jack_zone *zone; in snd_soc_jack_get_type() local 161 list_for_each_entry(zone, &jack->jack_zones, list) { in snd_soc_jack_get_type() 162 if (micbias_voltage >= zone->min_mv && in snd_soc_jack_get_type() 163 micbias_voltage < zone->max_mv) in snd_soc_jack_get_type() 164 return zone->jack_type; in snd_soc_jack_get_type()
|
/linux-4.4.14/drivers/platform/x86/ |
D | alienware-wmi.c | 174 static int parse_rgb(const char *buf, struct platform_zone *zone) in parse_rgb() argument 194 zone->colors = repackager.cp; in parse_rgb() 214 static int alienware_update_led(struct platform_zone *zone) in alienware_update_led() argument 223 wmax_args.led_mask = 1 << zone->location; in alienware_update_led() 224 wmax_args.colors = zone->colors; in alienware_update_led() 232 legacy_args.colors = zone->colors; in alienware_update_led() 241 method_id = zone->location + 1; in alienware_update_led()
|
/linux-4.4.14/Documentation/devicetree/bindings/power_supply/ |
D | charger-manager.txt | 24 - cm-thermal-zone : name of external thermometer's thermal zone 53 cm-thermal-zone = "thermal_zone.1"
|
/linux-4.4.14/include/uapi/linux/ |
D | tipc.h | 63 static inline __u32 tipc_addr(unsigned int zone, in tipc_addr() argument 67 return (zone << 24) | (cluster << 12) | node; in tipc_addr()
|
/linux-4.4.14/arch/powerpc/kernel/ |
D | dma.c | 81 int zone; in __dma_direct_alloc_coherent() local 89 zone = dma_pfn_limit_to_zone(pfn); in __dma_direct_alloc_coherent() 90 if (zone < 0) { in __dma_direct_alloc_coherent() 96 switch (zone) { in __dma_direct_alloc_coherent()
|
/linux-4.4.14/Documentation/trace/ |
D | events-kmem.txt | 52 amounts of activity imply high activity on the zone->lock. Taking this lock 64 contention on the zone->lru_lock. 72 for order-0 pages, reduces contention on the zone->lock and reduces the 76 the zone->lock will be taken once and the per-CPU list refilled. The event 85 consecutively imply the zone->lock being taken once. Large amounts of per-CPU
|
/linux-4.4.14/arch/sh/mm/ |
D | init.c | 523 struct zone *zone; in arch_remove_memory() local 526 zone = page_zone(pfn_to_page(start_pfn)); in arch_remove_memory() 527 ret = __remove_pages(zone, start_pfn, nr_pages); in arch_remove_memory()
|
D | Kconfig | 25 int "Maximum zone order" 35 blocks into "zones", where each zone is a power of two number of
|
/linux-4.4.14/include/uapi/linux/tc_act/ |
D | tc_connmark.h | 11 __u16 zone; member
|
/linux-4.4.14/include/net/tc_act/ |
D | tc_connmark.h | 9 u16 zone; member
|
/linux-4.4.14/drivers/net/wireless/rsi/ |
D | rsi_91x_main.c | 43 void rsi_dbg(u32 zone, const char *fmt, ...) in rsi_dbg() argument 53 if (zone & rsi_zone_enabled) in rsi_dbg()
|
D | rsi_main.h | 43 extern __printf(2, 3) void rsi_dbg(u32 zone, const char *fmt, ...);
|
/linux-4.4.14/kernel/sched/ |
D | wait.c | 489 const struct zone *zone = page_zone(virt_to_page(word)); in bit_waitqueue() local 492 return &zone->wait_table[hash_long(val, zone->wait_table_bits)]; in bit_waitqueue()
|
/linux-4.4.14/arch/arm/boot/dts/ |
D | exynos5440-trip-points.dtsi | 2 * Device tree sources for default Exynos5440 thermal zone definition
|
D | exynos5420-trip-points.dtsi | 2 * Device tree sources for default Exynos5420 thermal zone definition
|
D | exynos4-cpu-thermal.dtsi | 2 * Device tree sources for Exynos4 thermal zone
|
D | exynos5422-cpu-thermal.dtsi | 2 * Device tree sources for Exynos5422 thermal zone
|
/linux-4.4.14/drivers/block/ |
D | pktcdvd.c | 1093 static struct packet_data *pkt_get_packet_data(struct pktcdvd_device *pd, int zone) in pkt_get_packet_data() argument 1098 if (pkt->sector == zone || pkt->list.next == &pd->cdrw.pkt_free_list) { in pkt_get_packet_data() 1100 if (pkt->sector != zone) in pkt_get_packet_data() 1199 sector_t zone = 0; /* Suppress gcc warning */ in pkt_handle_queue() local 1224 zone = get_zone(bio->bi_iter.bi_sector, pd); in pkt_handle_queue() 1226 if (p->sector == zone) { in pkt_handle_queue() 1248 pkt = pkt_get_packet_data(pd, zone); in pkt_handle_queue() 1250 pd->current_sector = zone + pd->settings.size; in pkt_handle_queue() 1251 pkt->sector = zone; in pkt_handle_queue() 1260 pkt_dbg(2, pd, "looking for zone %llx\n", (unsigned long long)zone); in pkt_handle_queue() [all …]
|
/linux-4.4.14/arch/tile/include/asm/ |
D | homecache.h | 26 struct zone;
|
/linux-4.4.14/net/ipv4/netfilter/ |
D | nf_conntrack_proto_icmp.c | 137 const struct nf_conntrack_zone *zone; in icmp_error_message() local 141 zone = nf_ct_zone_tmpl(tmpl, skb, &tmp); in icmp_error_message() 165 h = nf_conntrack_find_get(net, zone, &innertuple); in icmp_error_message()
|
/linux-4.4.14/arch/metag/mm/ |
D | Kconfig | 52 int "Maximum zone order" 57 blocks into "zones", where each zone is a power of two number of
|
/linux-4.4.14/arch/c6x/ |
D | Kconfig | 88 int "Maximum zone order" 92 blocks into "zones", where each zone is a power of two number of
|
/linux-4.4.14/drivers/memstick/core/ |
D | ms_block.c | 174 int zone = msb_get_zone_from_pba(pba); in msb_mark_block_used() local 188 msb->free_block_count[zone]--; in msb_mark_block_used() 194 int zone = msb_get_zone_from_pba(pba); in msb_mark_block_unused() local 207 msb->free_block_count[zone]++; in msb_mark_block_unused() 1078 static u16 msb_get_free_block(struct msb_data *msb, int zone) in msb_get_free_block() argument 1081 int pba = zone * MS_BLOCKS_IN_ZONE; in msb_get_free_block() 1086 if (!msb->free_block_count[zone]) { in msb_get_free_block() 1087 …O free blocks in the zone %d, to use for a write, (media is WORN out) switching to RO mode", zone); in msb_get_free_block() 1092 pos %= msb->free_block_count[zone]; in msb_get_free_block() 1095 msb->free_block_count[zone], pos); in msb_get_free_block() [all …]
|
/linux-4.4.14/drivers/firmware/efi/libstub/ |
D | Makefile | 10 -fPIC -fno-strict-aliasing -mno-red-zone \
|
/linux-4.4.14/Documentation/cris/ |
D | README | 84 zone(0): 2048 pages. 85 zone(1): 0 pages. 86 zone(2): 0 pages.
|
/linux-4.4.14/kernel/ |
D | kexec_core.c | 1381 VMCOREINFO_STRUCT_SIZE(zone); in crash_save_vmcoreinfo_init() 1399 VMCOREINFO_OFFSET(zone, free_area); in crash_save_vmcoreinfo_init() 1400 VMCOREINFO_OFFSET(zone, vm_stat); in crash_save_vmcoreinfo_init() 1401 VMCOREINFO_OFFSET(zone, spanned_pages); in crash_save_vmcoreinfo_init() 1407 VMCOREINFO_LENGTH(zone.free_area, MAX_ORDER); in crash_save_vmcoreinfo_init()
|
D | fork.c | 217 struct zone *zone = page_zone(virt_to_page(ti)); in account_kernel_stack() local 219 mod_zone_page_state(zone, NR_KERNEL_STACK, account); in account_kernel_stack()
|
/linux-4.4.14/include/sound/ |
D | emux_synth.h | 183 struct snd_sf_zone *zone; /* Zone assigned to this note */ member
|
/linux-4.4.14/arch/nios2/ |
D | Kconfig | 60 int "Maximum zone order" 65 blocks into "zones", where each zone is a power of two number of
|
/linux-4.4.14/fs/xfs/libxfs/ |
D | xfs_da_btree.h | 25 struct zone;
|
/linux-4.4.14/Documentation/filesystems/ |
D | sysv-fs.txt | 36 * Size of a block or zone (data allocation unit on disk) 196 Notation: We often speak of a "block" but mean a zone (the allocation unit)
|
D | proc.txt | 761 Node 0, zone DMA 0 4 5 4 4 3 ... 762 Node 0, zone Normal 1 0 0 1 101 8 ... 763 Node 0, zone HighMem 2 0 0 1 1 0 ... 783 Node 0, zone DMA, type Unmovable 0 0 0 1 1 1 1 1… 784 Node 0, zone DMA, type Reclaimable 0 0 0 0 0 0 0 0… 785 Node 0, zone DMA, type Movable 1 1 2 1 2 1 1 0… 786 Node 0, zone DMA, type Reserve 0 0 0 0 0 0 0 0… 787 Node 0, zone DMA, type Isolate 0 0 0 0 0 0 0 0… 788 Node 0, zone DMA32, type Unmovable 103 54 77 1 1 1 11 8… 789 Node 0, zone DMA32, type Reclaimable 0 0 2 1 0 0 0 0… [all …]
|
D | f2fs.txt | 279 -z [int] : Set the number of sections per zone. 324 to 2MB in size. A section is composed of consecutive segments, and a zone 325 consists of a set of sections. By default, section and zone sizes are set to one 331 align with the zone size <-| 376 start block address of Main area with the zone size by reserving some segments
|
D | ntfs.txt | 192 mft_zone_multiplier= Set the MFT zone multiplier for the volume (this 196 default. The MFT zone multiplier determines how much 198 other space is used up, then the MFT zone will be 205 Value MFT zone size (% of volume size)
|
D | vfat.txt | 118 Linux. This is useful when time zone set in sys_tz is 119 not the time zone used by the filesystem. Note that this
|
/linux-4.4.14/arch/x86/boot/compressed/ |
D | Makefile | 50 $(obj)/eboot.o: KBUILD_CFLAGS += -fshort-wchar -mno-red-zone
|
/linux-4.4.14/arch/powerpc/include/asm/ |
D | pgtable.h | 229 void limit_zone_pfn(enum zone_type zone, unsigned long max_pfn);
|
/linux-4.4.14/drivers/misc/ |
D | lkdtm.c | 207 struct zone *zone, in jp_shrink_inactive_list() argument
|
/linux-4.4.14/Documentation/cgroups/ |
D | memcg_test.txt | 110 VM's control (means that it's handled under global zone->lru_lock). 112 list management functions under zone->lru_lock().
|
D | memory.txt | 270 zone->lru_lock 273 per-zone-per-cgroup LRU (cgroup's private LRU) is just guarded by 274 zone->lru_lock, it has no lock of its own.
|
/linux-4.4.14/drivers/staging/unisys/Documentation/ |
D | proc-entries.txt | 49 clientstring and zone.
|
D | overview.txt | 234 zoneguid a GUID identifying the channel zone, in
|
/linux-4.4.14/Documentation/locking/ |
D | lockstat.txt | 174 …&(&zone->lru_lock)->rlock: 93000 94934 0.16 59.18 188253.…
|
/linux-4.4.14/arch/x86/ |
D | Makefile | 117 KBUILD_CFLAGS += -mno-red-zone
|
/linux-4.4.14/Documentation/acpi/ |
D | namespace.txt | 161 +-| _TZ | Scope(_TZ): the thermal zone namespace 177 | +-| TZ00 | ThermalZone(TZ00); the FAN thermal zone
|
/linux-4.4.14/Documentation/devicetree/bindings/arm/ |
D | arm,scpi.txt | 187 zone. The ID "3" is the sensor identifier for the temperature sensor
|
/linux-4.4.14/Documentation/devicetree/bindings/arm/bcm/ |
D | brcm,brcmstb.txt | 81 o offset to the base CPU power zone register
|
/linux-4.4.14/arch/ia64/include/asm/ |
D | pgtable.h | 547 extern void memmap_init (unsigned long size, int nid, unsigned long zone,
|
/linux-4.4.14/arch/xtensa/ |
D | Kconfig | 402 int "Maximum zone order" 406 blocks into "zones", where each zone is a power of two number of
|
/linux-4.4.14/fs/ |
D | buffer.c | 263 struct zone *zone; in free_more_memory() local 272 &zone); in free_more_memory() 273 if (zone) in free_more_memory()
|
/linux-4.4.14/Documentation/input/ |
D | iforce-protocol.txt | 202 Idx 00 Set dead zone (0..2048)
|
/linux-4.4.14/fs/f2fs/ |
D | segment.h | 263 unsigned int zone; /* current zone number */ member
|
D | segment.c | 939 if (CURSEG_I(sbi, i)->zone == zoneno) in get_new_segment() 967 curseg->zone = GET_ZONENO_FROM_SEGNO(sbi, curseg->segno); in reset_curseg()
|
/linux-4.4.14/Documentation/ |
D | memory-hotplug.txt | 176 The first column shows it's default zone. 308 create ZONE_MOVABLE...a zone which is just used for movable pages.
|
D | rtc.txt | 8 the local time zone or daylight savings time -- unless they dual boot
|
D | kernel-parameters.txt | 1301 highmem=nn[KMG] [KNL,BOOT] forces the highmem zone to have an exact 1703 of Movable pages. The Movable zone is used for the 1706 HugeTLB pages may not be allocated from this zone. 1708 use the HighMem zone if it exists, and the Normal 1709 zone if it does not. 2617 one of ['zone', 'node', 'default'] can be specified 3697 Set to disable actions on ACPI thermal zone 3709 Specify global default ACPI thermal zone polling rate
|
/linux-4.4.14/ |
D | REPORTING-BUGS | 162 around the world, and they may not work in your time zone. Unless you
|
/linux-4.4.14/fs/affs/ |
D | Changes | 192 zone number was zero, the loop didn't terminate,
|
/linux-4.4.14/Documentation/DocBook/ |
D | kernel-api.xml.db | 235 API-nr-free-zone-pages 246 API-setup-per-zone-wmarks
|
D | alsa-driver-api.xml.db | 221 API-struct-snd-soc-jack-zone
|
/linux-4.4.14/arch/powerpc/ |
D | Kconfig | 587 int "Maximum zone order" 602 blocks into "zones", where each zone is a power of two number of
|
/linux-4.4.14/arch/arm64/ |
D | Kconfig | 577 blocks into "zones", where each zone is a power of two number of
|