Searched refs:start (Results 1 - 200 of 9074) sorted by relevance

1234567891011>>

/linux-4.4.14/arch/ia64/scripts/
H A Dcheck-segrel.S2 data4 @segrel(start)
4 start: label
/linux-4.4.14/kernel/
H A Drange.c10 int add_range(struct range *range, int az, int nr_range, u64 start, u64 end) add_range() argument
12 if (start >= end) add_range()
19 range[nr_range].start = start; add_range()
28 u64 start, u64 end) add_range_with_merge()
32 if (start >= end) add_range_with_merge()
35 /* get new start/end: */ add_range_with_merge()
42 common_start = max(range[i].start, start); add_range_with_merge()
47 /* new start/end, will add it back at last */ add_range_with_merge()
48 start = min(range[i].start, start); add_range_with_merge()
53 range[nr_range - 1].start = 0; add_range_with_merge()
60 return add_range(range, az, nr_range, start, end); add_range_with_merge()
63 void subtract_range(struct range *range, int az, u64 start, u64 end) subtract_range() argument
67 if (start >= end) subtract_range()
74 if (start <= range[j].start && end >= range[j].end) { subtract_range()
75 range[j].start = 0; subtract_range()
80 if (start <= range[j].start && end < range[j].end && subtract_range()
81 range[j].start < end) { subtract_range()
82 range[j].start = end; subtract_range()
87 if (start > range[j].start && end >= range[j].end && subtract_range()
88 range[j].end > start) { subtract_range()
89 range[j].end = start; subtract_range()
93 if (start > range[j].start && end < range[j].end) { subtract_range()
101 range[i].start = end; subtract_range()
106 range[j].end = start; subtract_range()
117 if (r1->start < r2->start) cmp_range()
119 if (r1->start > r2->start) cmp_range()
139 range[i].start = range[k].start; clean_sort_range()
141 range[k].start = 0; clean_sort_range()
27 add_range_with_merge(struct range *range, int az, int nr_range, u64 start, u64 end) add_range_with_merge() argument
H A Dresource.c31 .start = 0,
39 .start = 0,
116 width, (unsigned long long) r->start, r_show()
123 .start = r_start,
210 resource_size_t start = new->start; __request_resource() local
214 if (end < start) __request_resource()
216 if (start < root->start) __request_resource()
223 if (!tmp || tmp->start > end) { __request_resource()
230 if (tmp->end < start) __request_resource()
273 tmp->start = 0; __release_child_resources()
336 * Finds the lowest iomem reosurce exists with-in [res->start.res->end)
337 * the caller must specify res->start, res->end, res->flags and "name".
345 resource_size_t start, end; find_next_iomem_res() local
351 start = res->start; find_next_iomem_res()
353 BUG_ON(start >= end); find_next_iomem_res()
365 if (p->start > end) { find_next_iomem_res()
369 if ((p->end >= start) && (p->start < end)) find_next_iomem_res()
377 if (res->start < p->start) find_next_iomem_res()
378 res->start = p->start; find_next_iomem_res()
387 * All the memory ranges which overlap start,end and also match flags and
392 * @start: start addr
395 int walk_iomem_res(char *name, unsigned long flags, u64 start, u64 end, walk_iomem_res() argument
402 res.start = start; walk_iomem_res()
406 while ((res.start < res.end) && walk_iomem_res()
408 ret = (*func)(res.start, res.end, arg); walk_iomem_res()
411 res.start = res.end + 1; walk_iomem_res()
424 int walk_system_ram_res(u64 start, u64 end, void *arg, walk_system_ram_res() argument
431 res.start = start; walk_system_ram_res()
435 while ((res.start < res.end) && walk_system_ram_res()
437 ret = (*func)(res.start, res.end, arg); walk_system_ram_res()
440 res.start = res.end + 1; walk_system_ram_res()
461 res.start = (u64) start_pfn << PAGE_SHIFT; walk_system_ram_range()
465 while ((res.start < res.end) && walk_system_ram_range()
467 pfn = (res.start + PAGE_SIZE - 1) >> PAGE_SHIFT; walk_system_ram_range()
473 res.start = res.end + 1; walk_system_ram_range()
497 * @start: region start address
513 int region_intersects(resource_size_t start, size_t size, const char *name) region_intersects() argument
516 resource_size_t end = start + size - 1; region_intersects()
524 if (start >= p->start && start <= p->end) region_intersects()
526 if (end >= p->start && end <= p->end) region_intersects()
528 if (p->start >= start && p->end <= end) region_intersects()
551 return avail->start; simple_align_resource()
557 if (res->start < min) resource_clip()
558 res->start = min; resource_clip()
575 tmp.start = root->start; __find_resource()
578 * of this->start - 1 to tmp->end below would cause an underflow. __find_resource()
580 if (this && this->start == root->start) { __find_resource()
581 tmp.start = (this == old) ? old->start : this->end + 1; __find_resource()
586 tmp.end = (this == old) ? this->end : this->start - 1; __find_resource()
590 if (tmp.end < tmp.start) __find_resource()
597 avail.start = ALIGN(tmp.start, constraint->align); __find_resource()
600 if (avail.start >= tmp.start) { __find_resource()
602 alloc.start = constraint->alignf(constraint->alignf_data, &avail, __find_resource()
604 alloc.end = alloc.start + size - 1; __find_resource()
606 new->start = alloc.start; __find_resource()
616 tmp.start = this->end + 1; __find_resource()
656 old->start = new.start; reallocate_resource()
667 old->start = new.start; reallocate_resource()
731 * lookup_resource - find an existing resource by a resource start address
733 * @start: resource start address
737 struct resource *lookup_resource(struct resource *root, resource_size_t start) lookup_resource() argument
743 if (res->start == start) lookup_resource()
769 if ((first->start > new->start) || (first->end < new->end)) __insert_resource()
771 if ((first->start == new->start) && (first->end == new->end)) __insert_resource()
777 if (next->start < new->start || next->end > new->end) __insert_resource()
781 if (next->sibling->start > new->end) __insert_resource()
866 if (conflict->start < new->start) insert_resource_expand_to_fit()
867 new->start = conflict->start; insert_resource_expand_to_fit()
876 static int __adjust_resource(struct resource *res, resource_size_t start, __adjust_resource() argument
880 resource_size_t end = start + size - 1; __adjust_resource()
886 if ((start < parent->start) || (end > parent->end)) __adjust_resource()
889 if (res->sibling && (res->sibling->start <= end)) __adjust_resource()
896 if (start <= tmp->end) __adjust_resource()
902 if ((tmp->start < start) || (tmp->end > end)) __adjust_resource()
905 res->start = start; __adjust_resource()
914 * adjust_resource - modify a resource's start and size
916 * @start: new start value
919 * Given an existing resource, change its start and size to match the
923 int adjust_resource(struct resource *res, resource_size_t start, adjust_resource() argument
929 result = __adjust_resource(res, start, size); adjust_resource()
936 resource_size_t start, resource_size_t end, __reserve_region_with_split()
948 res->start = start; __reserve_region_with_split()
964 if (conflict->start <= res->start && __reserve_region_with_split()
972 if (conflict->start > res->start) { __reserve_region_with_split()
974 res->end = conflict->start - 1; __reserve_region_with_split()
982 next_res->start = conflict->end + 1; __reserve_region_with_split()
987 res->start = conflict->end + 1; __reserve_region_with_split()
994 resource_size_t start, resource_size_t end, reserve_region_with_split()
1000 if (root->start > start || root->end < end) { reserve_region_with_split()
1002 (unsigned long long)start, (unsigned long long)end, reserve_region_with_split()
1004 if (start > root->end || end < root->start) reserve_region_with_split()
1009 if (start < root->start) reserve_region_with_split()
1010 start = root->start; reserve_region_with_split()
1012 (unsigned long long)start, reserve_region_with_split()
1018 __reserve_region_with_split(root, start, end, name); reserve_region_with_split()
1034 return res->start; resource_alignment()
1056 * @start: resource start address
1062 resource_size_t start, resource_size_t n, __request_region()
1072 res->start = start; __request_region()
1073 res->end = start + n - 1; __request_region()
1113 * @start: resource start address
1118 void __release_region(struct resource *parent, resource_size_t start, __release_region() argument
1125 end = start + n - 1; __release_region()
1134 if (res->start <= start && res->end >= end) { __release_region()
1139 if (res->start != start || res->end != end) __release_region()
1154 "<%016llx-%016llx>\n", (unsigned long long)start, __release_region()
1163 * @start: resource start address
1181 resource_size_t start, resource_size_t size) release_mem_region_adjustable()
1189 end = start + size - 1; release_mem_region_adjustable()
1190 if ((start < parent->start) || (end > parent->end)) release_mem_region_adjustable()
1200 if (res->start >= end) release_mem_region_adjustable()
1204 if (res->start > start || res->end < end) { release_mem_region_adjustable()
1218 if (res->start == start && res->end == end) { release_mem_region_adjustable()
1223 } else if (res->start == start && res->end != end) { release_mem_region_adjustable()
1224 /* adjust the start */ release_mem_region_adjustable()
1227 } else if (res->start != start && res->end == end) { release_mem_region_adjustable()
1229 ret = __adjust_resource(res, res->start, release_mem_region_adjustable()
1230 start - res->start); release_mem_region_adjustable()
1238 new_res->start = end + 1; release_mem_region_adjustable()
1245 ret = __adjust_resource(res, res->start, release_mem_region_adjustable()
1246 start - res->start); release_mem_region_adjustable()
1337 resource_size_t start; member in struct:region_devres
1345 __release_region(this->parent, this->start, this->n); devm_region_release()
1353 this->start == match->start && this->n == match->n; devm_region_match()
1357 struct resource *parent, resource_size_t start, __devm_request_region()
1369 dr->start = start; __devm_request_region()
1372 res = __request_region(parent, start, n, name, 0); __devm_request_region()
1383 resource_size_t start, resource_size_t n) __devm_release_region()
1385 struct region_devres match_data = { parent, start, n }; __devm_release_region()
1387 __release_region(parent, start, n); __devm_release_region()
1413 res->start = io_start; reserve_setup()
1417 if (request_resource(res->start >= 0x10000 ? &iomem_resource : &ioport_resource, res) == 0) reserve_setup()
1442 if (p->start >= addr + size) iomem_map_sanity_check()
1446 if (PFN_DOWN(p->start) <= PFN_DOWN(addr) && iomem_map_sanity_check()
1498 if (p->start >= addr + size) iomem_is_exclusive()
935 __reserve_region_with_split(struct resource *root, resource_size_t start, resource_size_t end, const char *name) __reserve_region_with_split() argument
993 reserve_region_with_split(struct resource *root, resource_size_t start, resource_size_t end, const char *name) reserve_region_with_split() argument
1061 __request_region(struct resource *parent, resource_size_t start, resource_size_t n, const char *name, int flags) __request_region() argument
1180 release_mem_region_adjustable(struct resource *parent, resource_size_t start, resource_size_t size) release_mem_region_adjustable() argument
1356 __devm_request_region(struct device *dev, struct resource *parent, resource_size_t start, resource_size_t n, const char *name) __devm_request_region() argument
1382 __devm_release_region(struct device *dev, struct resource *parent, resource_size_t start, resource_size_t n) __devm_release_region() argument
/linux-4.4.14/tools/perf/
H A Dperf-read-vdso.c14 void *start, *end; main() local
17 if (find_vdso_map(&start, &end)) main()
20 size = end - start; main()
23 written = fwrite(start, 1, size, stdout); main()
26 start += written; main()
/linux-4.4.14/arch/mn10300/include/asm/
H A Dcacheflush.h24 extern void mn10300_local_icache_inv_page(unsigned long start);
25 extern void mn10300_local_icache_inv_range(unsigned long start, unsigned long end);
26 extern void mn10300_local_icache_inv_range2(unsigned long start, unsigned long size);
28 extern void mn10300_local_dcache_inv_page(unsigned long start);
29 extern void mn10300_local_dcache_inv_range(unsigned long start, unsigned long end);
30 extern void mn10300_local_dcache_inv_range2(unsigned long start, unsigned long size);
32 extern void mn10300_icache_inv_page(unsigned long start);
33 extern void mn10300_icache_inv_range(unsigned long start, unsigned long end);
34 extern void mn10300_icache_inv_range2(unsigned long start, unsigned long size);
36 extern void mn10300_dcache_inv_page(unsigned long start);
37 extern void mn10300_dcache_inv_range(unsigned long start, unsigned long end);
38 extern void mn10300_dcache_inv_range2(unsigned long start, unsigned long size);
41 extern void mn10300_local_dcache_flush_page(unsigned long start);
42 extern void mn10300_local_dcache_flush_range(unsigned long start, unsigned long end);
43 extern void mn10300_local_dcache_flush_range2(unsigned long start, unsigned long size);
45 extern void mn10300_local_dcache_flush_inv_page(unsigned long start);
46 extern void mn10300_local_dcache_flush_inv_range(unsigned long start, unsigned long end);
47 extern void mn10300_local_dcache_flush_inv_range2(unsigned long start, unsigned long size);
49 extern void mn10300_dcache_flush_page(unsigned long start);
50 extern void mn10300_dcache_flush_range(unsigned long start, unsigned long end);
51 extern void mn10300_dcache_flush_range2(unsigned long start, unsigned long size);
53 extern void mn10300_dcache_flush_inv_page(unsigned long start);
54 extern void mn10300_dcache_flush_inv_range(unsigned long start, unsigned long end);
55 extern void mn10300_dcache_flush_inv_range2(unsigned long start, unsigned long size);
58 #define mn10300_local_dcache_flush_page(start) do {} while (0)
59 #define mn10300_local_dcache_flush_range(start, end) do {} while (0)
60 #define mn10300_local_dcache_flush_range2(start, size) do {} while (0)
63 #define mn10300_local_dcache_flush_inv_page(start) \
64 mn10300_local_dcache_inv_page(start)
65 #define mn10300_local_dcache_flush_inv_range(start, end) \
66 mn10300_local_dcache_inv_range(start, end)
67 #define mn10300_local_dcache_flush_inv_range2(start, size) \
68 mn10300_local_dcache_inv_range2(start, size)
70 #define mn10300_dcache_flush_page(start) do {} while (0)
71 #define mn10300_dcache_flush_range(start, end) do {} while (0)
72 #define mn10300_dcache_flush_range2(start, size) do {} while (0)
74 #define mn10300_dcache_flush_inv_page(start) \
75 mn10300_dcache_inv_page((start))
76 #define mn10300_dcache_flush_inv_range(start, end) \
77 mn10300_dcache_inv_range((start), (end))
78 #define mn10300_dcache_flush_inv_range2(start, size) \
79 mn10300_dcache_inv_range2((start), (size))
83 #define mn10300_local_icache_inv_page(start) do {} while (0)
84 #define mn10300_local_icache_inv_range(start, end) do {} while (0)
85 #define mn10300_local_icache_inv_range2(start, size) do {} while (0)
87 #define mn10300_local_dcache_inv_page(start) do {} while (0)
88 #define mn10300_local_dcache_inv_range(start, end) do {} while (0)
89 #define mn10300_local_dcache_inv_range2(start, size) do {} while (0)
91 #define mn10300_local_dcache_flush_inv_page(start) do {} while (0)
93 #define mn10300_local_dcache_flush_inv_range(start, end)do {} while (0)
94 #define mn10300_local_dcache_flush_inv_range2(start, size) do {} while (0)
95 #define mn10300_local_dcache_flush_page(start) do {} while (0)
96 #define mn10300_local_dcache_flush_range(start, end) do {} while (0)
97 #define mn10300_local_dcache_flush_range2(start, size) do {} while (0)
99 #define mn10300_icache_inv_page(start) do {} while (0)
100 #define mn10300_icache_inv_range(start, end) do {} while (0)
101 #define mn10300_icache_inv_range2(start, size) do {} while (0)
103 #define mn10300_dcache_inv_page(start) do {} while (0)
104 #define mn10300_dcache_inv_range(start, end) do {} while (0)
105 #define mn10300_dcache_inv_range2(start, size) do {} while (0)
107 #define mn10300_dcache_flush_inv_page(start) do {} while (0)
109 #define mn10300_dcache_flush_inv_range(start, end) do {} while (0)
110 #define mn10300_dcache_flush_inv_range2(start, size) do {} while (0)
111 #define mn10300_dcache_flush_page(start) do {} while (0)
112 #define mn10300_dcache_flush_range(start, end) do {} while (0)
113 #define mn10300_dcache_flush_range2(start, size) do {} while (0)
122 #define flush_cache_range(mm, start, end) do {} while (0)
124 #define flush_cache_vmap(start, end) do {} while (0)
125 #define flush_cache_vunmap(start, end) do {} while (0)
136 extern void flush_icache_range(unsigned long start, unsigned long end);
143 extern void flush_icache_range(unsigned long start, unsigned long end);
145 #define flush_icache_range(start, end) do {} while (0)
H A Dtermios.h8 start=^Q stop=^S susp=^Z eol=\0
/linux-4.4.14/arch/hexagon/mm/
H A Dcache.c25 #define spanlines(start, end) \
26 (((end - (start & ~(LINESIZE - 1))) >> LINEBITS) + 1)
28 void flush_dcache_range(unsigned long start, unsigned long end) flush_dcache_range() argument
30 unsigned long lines = spanlines(start, end-1); flush_dcache_range()
33 start &= ~(LINESIZE - 1); flush_dcache_range()
41 : "r" (start) flush_dcache_range()
43 start += LINESIZE; flush_dcache_range()
48 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument
50 unsigned long lines = spanlines(start, end-1); flush_icache_range()
53 start &= ~(LINESIZE - 1); flush_icache_range()
62 : "r" (start) flush_icache_range()
64 start += LINESIZE; flush_icache_range()
73 void hexagon_clean_dcache_range(unsigned long start, unsigned long end) hexagon_clean_dcache_range() argument
75 unsigned long lines = spanlines(start, end-1); hexagon_clean_dcache_range()
78 start &= ~(LINESIZE - 1); hexagon_clean_dcache_range()
86 : "r" (start) hexagon_clean_dcache_range()
88 start += LINESIZE; hexagon_clean_dcache_range()
93 void hexagon_inv_dcache_range(unsigned long start, unsigned long end) hexagon_inv_dcache_range() argument
95 unsigned long lines = spanlines(start, end-1); hexagon_inv_dcache_range()
98 start &= ~(LINESIZE - 1); hexagon_inv_dcache_range()
106 : "r" (start) hexagon_inv_dcache_range()
108 start += LINESIZE; hexagon_inv_dcache_range()
H A Dstrnlen_user.S28 #define start r7 define
49 start = isrc; define
60 tmp1 = memb(start++#1);
65 P1 = cmp.gtu(end,start);
66 mod8 = and(start,#7);
80 dbuf = memd(start);
81 obo = add(start,#1);
88 P0 = cmp.gtu(end,start);
98 if (!P0.new) start = add(obo,tmp1);
101 start = add(start,#8); define
110 start = add(obo,tmp1); define
115 R0 = sub(start,isrc);
/linux-4.4.14/mm/
H A Dmsync.c20 * MS_ASYNC does not start I/O (it used to, up to 2.5.67).
26 * Or the application may run fadvise(FADV_DONTNEED) against the fd to start
31 SYSCALL_DEFINE3(msync, unsigned long, start, size_t, len, int, flags) SYSCALL_DEFINE3()
41 if (offset_in_page(start)) SYSCALL_DEFINE3()
47 end = start + len; SYSCALL_DEFINE3()
48 if (end < start) SYSCALL_DEFINE3()
51 if (end == start) SYSCALL_DEFINE3()
54 * If the interval [start,end) covers some unmapped address ranges, SYSCALL_DEFINE3()
58 vma = find_vma(mm, start); SYSCALL_DEFINE3()
63 /* Still start < end. */ SYSCALL_DEFINE3()
67 /* Here start < vma->vm_end. */ SYSCALL_DEFINE3()
68 if (start < vma->vm_start) { SYSCALL_DEFINE3()
69 start = vma->vm_start; SYSCALL_DEFINE3()
70 if (start >= end) SYSCALL_DEFINE3()
74 /* Here vma->vm_start <= start < vma->vm_end. */ SYSCALL_DEFINE3()
81 fstart = (start - vma->vm_start) + SYSCALL_DEFINE3()
83 fend = fstart + (min(end, vma->vm_end) - start) - 1; SYSCALL_DEFINE3()
84 start = vma->vm_end; SYSCALL_DEFINE3()
91 if (error || start >= end) SYSCALL_DEFINE3()
94 vma = find_vma(mm, start); SYSCALL_DEFINE3()
96 if (start >= end) { SYSCALL_DEFINE3()
H A Dmemtest.c36 u64 *p, *start, *end; memtest() local
42 start = __va(start_phys_aligned); memtest()
43 end = start + (size - (start_phys_aligned - start_phys)) / incr; memtest()
47 for (p = start; p < end; p++) memtest()
50 for (p = start; p < end; p++, start_phys_aligned += incr) { memtest()
65 static void __init do_one_pass(u64 pattern, phys_addr_t start, phys_addr_t end) do_one_pass() argument
72 this_start = clamp(this_start, start, end); do_one_pass()
73 this_end = clamp(this_end, start, end); do_one_pass()
99 void __init early_memtest(phys_addr_t start, phys_addr_t end) early_memtest() argument
110 do_one_pass(patterns[idx], start, end); early_memtest()
H A Dmadvise.c48 unsigned long start, unsigned long end, int behavior) madvise_behavior()
87 error = ksm_madvise(vma, start, end, behavior, &new_flags); madvise_behavior()
104 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT); madvise_behavior()
105 *prev = vma_merge(mm, *prev, start, end, new_flags, vma->anon_vma, madvise_behavior()
115 if (start != vma->vm_start) { madvise_behavior()
116 error = split_vma(mm, vma, start, 1); madvise_behavior()
140 static int swapin_walk_pmd_entry(pmd_t *pmd, unsigned long start, swapin_walk_pmd_entry() argument
150 for (index = start; index != end; index += PAGE_SIZE) { swapin_walk_pmd_entry()
156 orig_pte = pte_offset_map_lock(vma->vm_mm, pmd, start, &ptl); swapin_walk_pmd_entry()
157 pte = *(orig_pte + ((index - start) / PAGE_SIZE)); swapin_walk_pmd_entry()
176 unsigned long start, unsigned long end) force_swapin_readahead()
184 walk_page_range(start, end, &walk); force_swapin_readahead()
190 unsigned long start, unsigned long end, force_shm_swapin_readahead()
197 for (; start < end; start += PAGE_SIZE) { force_shm_swapin_readahead()
198 index = ((start - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; force_shm_swapin_readahead()
222 unsigned long start, unsigned long end) madvise_willneed()
229 force_swapin_readahead(vma, start, end); madvise_willneed()
235 force_shm_swapin_readahead(vma, start, end, madvise_willneed()
250 start = ((start - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; madvise_willneed()
255 force_page_cache_readahead(file->f_mapping, file, start, end - start); madvise_willneed()
280 unsigned long start, unsigned long end) madvise_dontneed()
286 zap_page_range(vma, start, end - start, NULL); madvise_dontneed()
296 unsigned long start, unsigned long end) madvise_remove()
316 offset = (loff_t)(start - vma->vm_start) madvise_remove()
329 offset, end - start); madvise_remove()
339 static int madvise_hwpoison(int bhv, unsigned long start, unsigned long end) madvise_hwpoison() argument
344 for (; start < end; start += PAGE_SIZE << madvise_hwpoison()
348 ret = get_user_pages_fast(start, 1, 0, &p); madvise_hwpoison()
358 page_to_pfn(p), start); madvise_hwpoison()
365 page_to_pfn(p), start); madvise_hwpoison()
375 unsigned long start, unsigned long end, int behavior) madvise_vma()
379 return madvise_remove(vma, prev, start, end); madvise_vma()
381 return madvise_willneed(vma, prev, start, end); madvise_vma()
383 return madvise_dontneed(vma, prev, start, end); madvise_vma()
385 return madvise_behavior(vma, prev, start, end, behavior); madvise_vma()
451 * -EINVAL - start + len < 0, start is not page-aligned,
460 SYSCALL_DEFINE3(madvise, unsigned long, start, size_t, len_in, int, behavior) SYSCALL_DEFINE3()
472 return madvise_hwpoison(behavior, start, start+len_in); SYSCALL_DEFINE3()
477 if (start & ~PAGE_MASK) SYSCALL_DEFINE3()
485 end = start + len; SYSCALL_DEFINE3()
486 if (end < start) SYSCALL_DEFINE3()
490 if (end == start) SYSCALL_DEFINE3()
500 * If the interval [start,end) covers some unmapped address SYSCALL_DEFINE3()
504 vma = find_vma_prev(current->mm, start, &prev); SYSCALL_DEFINE3()
505 if (vma && start > vma->vm_start) SYSCALL_DEFINE3()
510 /* Still start < end. */ SYSCALL_DEFINE3()
515 /* Here start < (end|vma->vm_end). */ SYSCALL_DEFINE3()
516 if (start < vma->vm_start) { SYSCALL_DEFINE3()
518 start = vma->vm_start; SYSCALL_DEFINE3()
519 if (start >= end) SYSCALL_DEFINE3()
523 /* Here vma->vm_start <= start < (end|vma->vm_end) */ SYSCALL_DEFINE3()
528 /* Here vma->vm_start <= start < tmp <= (end|vma->vm_end). */ SYSCALL_DEFINE3()
529 error = madvise_vma(vma, &prev, start, tmp, behavior); SYSCALL_DEFINE3()
532 start = tmp; SYSCALL_DEFINE3()
533 if (prev && start < prev->vm_end) SYSCALL_DEFINE3()
534 start = prev->vm_end; SYSCALL_DEFINE3()
536 if (start >= end) SYSCALL_DEFINE3()
541 vma = find_vma(current->mm, start); SYSCALL_DEFINE3()
46 madvise_behavior(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) madvise_behavior() argument
175 force_swapin_readahead(struct vm_area_struct *vma, unsigned long start, unsigned long end) force_swapin_readahead() argument
189 force_shm_swapin_readahead(struct vm_area_struct *vma, unsigned long start, unsigned long end, struct address_space *mapping) force_shm_swapin_readahead() argument
220 madvise_willneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_willneed() argument
278 madvise_dontneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_dontneed() argument
294 madvise_remove(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_remove() argument
374 madvise_vma(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) madvise_vma() argument
H A Ddebug-pagealloc.c83 unsigned char *start; check_poison_mem() local
86 start = memchr_inv(mem, PAGE_POISON, bytes); check_poison_mem()
87 if (!start) check_poison_mem()
90 for (end = mem + bytes - 1; end > start; end--) { check_poison_mem()
97 else if (start == end && single_bit_flip(*start, PAGE_POISON)) check_poison_mem()
102 print_hex_dump(KERN_ERR, "", DUMP_PREFIX_ADDRESS, 16, 1, start, check_poison_mem()
103 end - start + 1, 1); check_poison_mem()
/linux-4.4.14/arch/mips/bcm63xx/
H A Ddev-uart.c16 /* start & end filled at runtime */
20 /* start filled at runtime */
27 /* start & end filled at runtime */
31 /* start filled at runtime */
62 uart0_resources[0].start = bcm63xx_regset_address(RSET_UART0); bcm63xx_uart_register()
63 uart0_resources[0].end = uart0_resources[0].start + bcm63xx_uart_register()
65 uart0_resources[1].start = bcm63xx_get_irq_number(IRQ_UART0); bcm63xx_uart_register()
69 uart1_resources[0].start = bcm63xx_regset_address(RSET_UART1); bcm63xx_uart_register()
70 uart1_resources[0].end = uart1_resources[0].start + bcm63xx_uart_register()
72 uart1_resources[1].start = bcm63xx_get_irq_number(IRQ_UART1); bcm63xx_uart_register()
H A Ddev-enet.c48 .start = -1, /* filled at runtime */
53 .start = -1, /* filled at runtime */
58 .start = -1, /* filled at runtime */
75 .start = -1, /* filled at runtime */
80 .start = -1, /* filled at runtime */
84 .start = -1, /* filled at runtime */
88 .start = -1, /* filled at runtime */
107 .start = -1, /* filled at runtime */
112 .start = -1, /* filled at runtime */
116 .start = -1, /* filled at runtime */
120 .start = -1, /* filled at runtime */
139 /* start & end filled at runtime */
143 /* start filled at runtime */
147 /* start filled at runtime */
172 shared_res[0].start = bcm63xx_regset_address(RSET_ENETDMA); register_shared()
173 shared_res[0].end = shared_res[0].start; register_shared()
186 shared_res[1].start = bcm63xx_regset_address(RSET_ENETDMAC); register_shared()
187 shared_res[1].end = shared_res[1].start; register_shared()
190 shared_res[2].start = bcm63xx_regset_address(RSET_ENETDMAS); register_shared()
191 shared_res[2].end = shared_res[2].start; register_shared()
220 enet0_res[0].start = bcm63xx_regset_address(RSET_ENET0); bcm63xx_enet_register()
221 enet0_res[0].end = enet0_res[0].start; bcm63xx_enet_register()
223 enet0_res[1].start = bcm63xx_get_irq_number(IRQ_ENET0); bcm63xx_enet_register()
224 enet0_res[2].start = bcm63xx_get_irq_number(IRQ_ENET0_RXDMA); bcm63xx_enet_register()
225 enet0_res[3].start = bcm63xx_get_irq_number(IRQ_ENET0_TXDMA); bcm63xx_enet_register()
228 enet1_res[0].start = bcm63xx_regset_address(RSET_ENET1); bcm63xx_enet_register()
229 enet1_res[0].end = enet1_res[0].start; bcm63xx_enet_register()
231 enet1_res[1].start = bcm63xx_get_irq_number(IRQ_ENET1); bcm63xx_enet_register()
232 enet1_res[2].start = bcm63xx_get_irq_number(IRQ_ENET1_RXDMA); bcm63xx_enet_register()
233 enet1_res[3].start = bcm63xx_get_irq_number(IRQ_ENET1_TXDMA); bcm63xx_enet_register()
286 enetsw_res[0].start = bcm63xx_regset_address(RSET_ENETSW); bcm63xx_enetsw_register()
287 enetsw_res[0].end = enetsw_res[0].start; bcm63xx_enetsw_register()
289 enetsw_res[1].start = bcm63xx_get_irq_number(IRQ_ENETSW_RXDMA0); bcm63xx_enetsw_register()
290 enetsw_res[2].start = bcm63xx_get_irq_number(IRQ_ENETSW_TXDMA0); bcm63xx_enetsw_register()
291 if (!enetsw_res[2].start) bcm63xx_enetsw_register()
292 enetsw_res[2].start = -1; bcm63xx_enetsw_register()
H A Ddev-hsspi.c19 .start = -1, /* filled at runtime */
24 .start = -1, /* filled at runtime */
41 spi_resources[0].start = bcm63xx_regset_address(RSET_HSSPI); bcm63xx_hsspi_register()
42 spi_resources[0].end = spi_resources[0].start; bcm63xx_hsspi_register()
44 spi_resources[1].start = bcm63xx_get_irq_number(IRQ_HSSPI); bcm63xx_hsspi_register()
H A Ddev-spi.c23 .start = -1, /* filled at runtime */
28 .start = -1, /* filled at runtime */
44 spi_resources[0].start = bcm63xx_regset_address(RSET_SPI); bcm63xx_spi_register()
45 spi_resources[0].end = spi_resources[0].start; bcm63xx_spi_register()
46 spi_resources[1].start = bcm63xx_get_irq_number(IRQ_SPI); bcm63xx_spi_register()
H A Ddev-usb-usbd.c47 usbd_resources[0].start = bcm63xx_regset_address(RSET_USBD); bcm63xx_usbd_register()
48 usbd_resources[0].end = usbd_resources[0].start + RSET_USBD_SIZE - 1; bcm63xx_usbd_register()
51 usbd_resources[1].start = bcm63xx_regset_address(RSET_USBDMA); bcm63xx_usbd_register()
52 usbd_resources[1].end = usbd_resources[1].start + RSET_USBDMA_SIZE - 1; bcm63xx_usbd_register()
58 r->start = r->end = bcm63xx_get_irq_number(irq_list[i]); bcm63xx_usbd_register()
H A Ddev-pcmcia.c22 /* start & end filled at runtime */
28 .start = BCM_PCMCIA_COMMON_BASE_PA,
33 .start = BCM_PCMCIA_ATTR_BASE_PA,
38 .start = BCM_PCMCIA_IO_BASE_PA,
45 /* start filled at runtime */
51 .start = BCM_PCMCIA_IO_BASE_PA,
125 pcmcia_resources[0].start = bcm63xx_regset_address(RSET_PCMCIA); bcm63xx_pcmcia_register()
126 pcmcia_resources[0].end = pcmcia_resources[0].start + bcm63xx_pcmcia_register()
128 pcmcia_resources[4].start = bcm63xx_get_irq_number(IRQ_PCMCIA); bcm63xx_pcmcia_register()
H A Ddev-rng.c16 .start = -1, /* filled at runtime */
34 rng_resources[0].start = bcm63xx_regset_address(RSET_RNG); bcm63xx_rng_register()
35 rng_resources[0].end = rng_resources[0].start; bcm63xx_rng_register()
H A Ddev-wdt.c16 .start = -1, /* filled at runtime */
31 wdt_resources[0].start = bcm63xx_regset_address(RSET_WDT); bcm63xx_wdt_register()
32 wdt_resources[0].end = wdt_resources[0].start; bcm63xx_wdt_register()
H A Ddev-dsp.c22 .start = -1, /* filled at runtime */
27 .start = -1, /* filled at runtime */
47 voip_dsp_resources[0].start = val; bcm63xx_dsp_register()
49 voip_dsp_resources[1].start = pd->ext_irq; bcm63xx_dsp_register()
/linux-4.4.14/arch/mn10300/mm/
H A Dcache-smp-flush.c32 * @start: The address of the page of memory to be flushed.
37 void mn10300_dcache_flush_page(unsigned long start) mn10300_dcache_flush_page() argument
41 start &= ~(PAGE_SIZE-1); mn10300_dcache_flush_page()
44 mn10300_local_dcache_flush_page(start); mn10300_dcache_flush_page()
45 smp_cache_call(SMP_DCACHE_FLUSH_RANGE, start, start + PAGE_SIZE); mn10300_dcache_flush_page()
51 * @start: The start address of the region to be flushed.
54 * Flush a range of addresses in the data cache on all CPUs, between start and
57 void mn10300_dcache_flush_range(unsigned long start, unsigned long end) mn10300_dcache_flush_range() argument
62 mn10300_local_dcache_flush_range(start, end); mn10300_dcache_flush_range()
63 smp_cache_call(SMP_DCACHE_FLUSH_RANGE, start, end); mn10300_dcache_flush_range()
69 * @start: The start address of the region to be flushed.
72 * Flush a range of addresses in the data cache on all CPUs, between start and
73 * start+size-1 inclusive.
75 void mn10300_dcache_flush_range2(unsigned long start, unsigned long size) mn10300_dcache_flush_range2() argument
80 mn10300_local_dcache_flush_range2(start, size); mn10300_dcache_flush_range2()
81 smp_cache_call(SMP_DCACHE_FLUSH_RANGE, start, start + size); mn10300_dcache_flush_range2()
103 * @start: The address of the page of memory to be flushed and invalidated.
108 void mn10300_dcache_flush_inv_page(unsigned long start) mn10300_dcache_flush_inv_page() argument
112 start &= ~(PAGE_SIZE-1); mn10300_dcache_flush_inv_page()
115 mn10300_local_dcache_flush_inv_page(start); mn10300_dcache_flush_inv_page()
116 smp_cache_call(SMP_DCACHE_FLUSH_INV_RANGE, start, start + PAGE_SIZE); mn10300_dcache_flush_inv_page()
123 * @start: The start address of the region to be flushed and invalidated.
127 * between start and end-1 inclusive.
129 void mn10300_dcache_flush_inv_range(unsigned long start, unsigned long end) mn10300_dcache_flush_inv_range() argument
134 mn10300_local_dcache_flush_inv_range(start, end); mn10300_dcache_flush_inv_range()
135 smp_cache_call(SMP_DCACHE_FLUSH_INV_RANGE, start, end); mn10300_dcache_flush_inv_range()
142 * @start: The start address of the region to be flushed and invalidated.
146 * between start and start+size-1 inclusive.
148 void mn10300_dcache_flush_inv_range2(unsigned long start, unsigned long size) mn10300_dcache_flush_inv_range2() argument
153 mn10300_local_dcache_flush_inv_range2(start, size); mn10300_dcache_flush_inv_range2()
154 smp_cache_call(SMP_DCACHE_FLUSH_INV_RANGE, start, start + size); mn10300_dcache_flush_inv_range2()
H A Dcache-smp-inv.c32 * @start: The address of the page of memory to be invalidated.
37 void mn10300_icache_inv_page(unsigned long start) mn10300_icache_inv_page() argument
41 start &= ~(PAGE_SIZE-1); mn10300_icache_inv_page()
44 mn10300_local_icache_inv_page(start); mn10300_icache_inv_page()
45 smp_cache_call(SMP_ICACHE_INV_RANGE, start, start + PAGE_SIZE); mn10300_icache_inv_page()
51 * @start: The start address of the region to be invalidated.
55 * between start and end-1 inclusive.
57 void mn10300_icache_inv_range(unsigned long start, unsigned long end) mn10300_icache_inv_range() argument
62 mn10300_local_icache_inv_range(start, end); mn10300_icache_inv_range()
63 smp_cache_call(SMP_ICACHE_INV_RANGE, start, end); mn10300_icache_inv_range()
69 * @start: The start address of the region to be invalidated.
73 * between start and start+size-1 inclusive.
75 void mn10300_icache_inv_range2(unsigned long start, unsigned long size) mn10300_icache_inv_range2() argument
80 mn10300_local_icache_inv_range2(start, size); mn10300_icache_inv_range2()
81 smp_cache_call(SMP_ICACHE_INV_RANGE, start, start + size); mn10300_icache_inv_range2()
102 * @start: The address of the page of memory to be invalidated.
107 void mn10300_dcache_inv_page(unsigned long start) mn10300_dcache_inv_page() argument
111 start &= ~(PAGE_SIZE-1); mn10300_dcache_inv_page()
114 mn10300_local_dcache_inv_page(start); mn10300_dcache_inv_page()
115 smp_cache_call(SMP_DCACHE_INV_RANGE, start, start + PAGE_SIZE); mn10300_dcache_inv_page()
121 * @start: The start address of the region to be invalidated.
124 * Invalidate a range of addresses in the data cache on all CPUs, between start
127 void mn10300_dcache_inv_range(unsigned long start, unsigned long end) mn10300_dcache_inv_range() argument
132 mn10300_local_dcache_inv_range(start, end); mn10300_dcache_inv_range()
133 smp_cache_call(SMP_DCACHE_INV_RANGE, start, end); mn10300_dcache_inv_range()
139 * @start: The start address of the region to be invalidated.
142 * Invalidate a range of addresses in the data cache on all CPUs, between start
143 * and start+size-1 inclusive.
145 void mn10300_dcache_inv_range2(unsigned long start, unsigned long size) mn10300_dcache_inv_range2() argument
150 mn10300_local_dcache_inv_range2(start, size); mn10300_dcache_inv_range2()
151 smp_cache_call(SMP_DCACHE_INV_RANGE, start, start + size); mn10300_dcache_inv_range2()
H A Dcache-flush-icache.c27 unsigned long start = page_to_phys(page); flush_icache_page() local
32 mn10300_local_dcache_flush_page(start); flush_icache_page()
33 mn10300_local_icache_inv_page(start); flush_icache_page()
35 smp_cache_call(SMP_IDCACHE_INV_FLUSH_RANGE, start, start + PAGE_SIZE); flush_icache_page()
43 * @start: The starting virtual address of the page part.
50 static void flush_icache_page_range(unsigned long start, unsigned long end) flush_icache_page_range() argument
60 off = start & ~PAGE_MASK; flush_icache_page_range()
61 size = end - start; flush_icache_page_range()
65 pgd = pgd_offset(current->mm, start); flush_icache_page_range()
69 pud = pud_offset(pgd, start); flush_icache_page_range()
73 pmd = pmd_offset(pud, start); flush_icache_page_range()
77 ppte = pte_offset_map(pmd, start); flush_icache_page_range()
96 smp_cache_call(SMP_IDCACHE_INV_FLUSH_RANGE, start, end); flush_icache_page_range()
101 * @start: The starting virtual address of the region.
108 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument
119 if (start >= end) flush_icache_range()
126 start_page = (start >= 0x80000000UL) ? start : 0x80000000UL; flush_icache_range()
130 if (start_page == start) flush_icache_range()
135 start_page = start & PAGE_MASK; flush_icache_range()
140 flush_icache_page_range(start, end); flush_icache_range()
143 flush_icache_page_range(start, end_page); flush_icache_range()
H A Dcache-inv-icache.c21 * @start: The starting virtual address of the page part.
28 static void flush_icache_page_range(unsigned long start, unsigned long end) flush_icache_page_range() argument
38 off = start & ~PAGE_MASK; flush_icache_page_range()
39 size = end - start; flush_icache_page_range()
43 pgd = pgd_offset(current->mm, start); flush_icache_page_range()
47 pud = pud_offset(pgd, start); flush_icache_page_range()
51 pmd = pmd_offset(pud, start); flush_icache_page_range()
55 ppte = pte_offset_map(pmd, start); flush_icache_page_range()
72 smp_cache_call(SMP_ICACHE_INV_RANGE, start, end); flush_icache_page_range()
77 * @start: The starting virtual address of the region.
84 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument
95 if (start >= end) flush_icache_range()
102 start_page = (start >= 0x80000000UL) ? start : 0x80000000UL; flush_icache_range()
104 smp_cache_call(SMP_ICACHE_INV_RANGE, start, end); flush_icache_range()
105 if (start_page == start) flush_icache_range()
110 start_page = start & PAGE_MASK; flush_icache_range()
115 flush_icache_page_range(start, end); flush_icache_range()
118 flush_icache_page_range(start, end_page); flush_icache_range()
/linux-4.4.14/arch/x86/kernel/
H A Dresource.c4 static void resource_clip(struct resource *res, resource_size_t start, resource_clip() argument
9 if (res->end < start || res->start > end) resource_clip()
12 if (res->start < start) resource_clip()
13 low = start - res->start; resource_clip()
20 res->end = start - 1; resource_clip()
22 res->start = end + 1; resource_clip()
H A Dprobe_roms.c26 .start = 0xf0000,
33 .start = 0xe0000,
40 .start = 0xc8000,
45 .start = 0,
50 .start = 0,
55 .start = 0,
60 .start = 0,
65 .start = 0,
72 .start = 0xc0000,
126 rom = isa_bus_to_virt(res->start); find_oprom()
160 return ioremap(oprom->start, resource_size(oprom)); pci_map_biosrom()
200 unsigned long start, length, upper; probe_roms() local
205 upper = adapter_rom_resources[0].start; probe_roms()
206 for (start = video_rom_resource.start; start < upper; start += 2048) { probe_roms()
207 rom = isa_bus_to_virt(start); probe_roms()
211 video_rom_resource.start = start; probe_roms()
221 video_rom_resource.end = start + length - 1; probe_roms()
227 start = (video_rom_resource.end + 1 + 2047) & ~2047UL; probe_roms()
228 if (start < upper) probe_roms()
229 start = upper; probe_roms()
233 upper = system_rom_resource.start; probe_roms()
236 rom = isa_bus_to_virt(extension_rom_resource.start); probe_roms()
241 upper = extension_rom_resource.start; probe_roms()
246 for (i = 0; i < ARRAY_SIZE(adapter_rom_resources) && start < upper; start += 2048) { probe_roms()
247 rom = isa_bus_to_virt(start); probe_roms()
258 if (!length || start + length > upper || !romchecksum(rom, length)) probe_roms()
261 adapter_rom_resources[i].start = start; probe_roms()
262 adapter_rom_resources[i].end = start + length - 1; probe_roms()
265 start = adapter_rom_resources[i++].end & ~2047UL; probe_roms()
H A Dpci-iommu_table.c10 find_dependents_of(struct iommu_table_entry *start, find_dependents_of() argument
19 for (p = start; p < finish; p++) find_dependents_of()
27 void __init sort_iommu_table(struct iommu_table_entry *start, sort_iommu_table() argument
32 for (p = start; p < finish; p++) { sort_iommu_table()
34 q = find_dependents_of(start, finish, p); sort_iommu_table()
49 void __init check_iommu_entries(struct iommu_table_entry *start, check_iommu_entries() argument
55 for (p = start; p < finish; p++) { check_iommu_entries()
56 q = find_dependents_of(start, finish, p); check_iommu_entries()
57 x = find_dependents_of(start, finish, q); check_iommu_entries()
66 for (p = start; p < finish; p++) { check_iommu_entries()
75 inline void check_iommu_entries(struct iommu_table_entry *start, check_iommu_entries() argument
H A Dcheck.c73 phys_addr_t start, end; setup_bios_corruption_check() local
94 for_each_free_mem_range(i, NUMA_NO_NODE, MEMBLOCK_NONE, &start, &end, setup_bios_corruption_check()
96 start = clamp_t(phys_addr_t, round_up(start, PAGE_SIZE), setup_bios_corruption_check()
100 if (start >= end) setup_bios_corruption_check()
103 memblock_reserve(start, end - start); setup_bios_corruption_check()
104 scan_areas[num_scan_areas].addr = start; setup_bios_corruption_check()
105 scan_areas[num_scan_areas].size = end - start; setup_bios_corruption_check()
108 memset(__va(start), 0, end - start); setup_bios_corruption_check()
/linux-4.4.14/arch/xtensa/mm/
H A Dinit.c43 sysmem.bank[i].start, sysmem.bank[i].end, sysmem_dump()
44 (sysmem.bank[i].end - sysmem.bank[i].start) >> 10); sysmem_dump()
48 * Find bank with maximal .start such that bank.start <= start
50 static inline struct meminfo * __init find_bank(unsigned long start) find_bank() argument
56 if (sysmem.bank[i].start <= start) find_bank()
88 int __init add_sysmem_bank(unsigned long start, unsigned long end) add_sysmem_bank() argument
95 if (start == end || add_sysmem_bank()
96 (start < end) != (PAGE_ALIGN(start) < (end & PAGE_MASK))) { add_sysmem_bank()
98 start, end - start); add_sysmem_bank()
102 start = PAGE_ALIGN(start); add_sysmem_bank()
104 sz = end - start; add_sysmem_bank()
106 it = find_bank(start); add_sysmem_bank()
109 bank_sz = it->end - it->start; add_sysmem_bank()
111 if (it && bank_sz >= start - it->start) { add_sysmem_bank()
112 if (end - it->start > bank_sz) add_sysmem_bank()
123 it->start - start <= sz) { add_sysmem_bank()
124 it->start = start; add_sysmem_bank()
125 if (it->end - it->start < sz) add_sysmem_bank()
132 start, end - start); add_sysmem_bank()
135 it->start = start; add_sysmem_bank()
140 sz = it->end - it->start; add_sysmem_bank()
142 if (sysmem.bank[i].start - it->start <= sz) { add_sysmem_bank()
143 if (sz < sysmem.bank[i].end - it->start) add_sysmem_bank()
154 * mem_reserve(start, end, must_exist)
161 * start Start of region,
170 int __init mem_reserve(unsigned long start, unsigned long end, int must_exist) mem_reserve() argument
177 start = start & PAGE_MASK; mem_reserve()
179 sz = end - start; mem_reserve()
183 it = find_bank(start); mem_reserve()
186 bank_sz = it->end - it->start; mem_reserve()
188 if ((!it || end - it->start > bank_sz) && must_exist) { mem_reserve()
190 start, end); mem_reserve()
194 if (it && start - it->start <= bank_sz) { mem_reserve()
195 if (start == it->start) { mem_reserve()
196 if (end - it->start < bank_sz) { mem_reserve()
197 it->start = end; mem_reserve()
203 it->end = start; mem_reserve()
204 if (end - it->start < bank_sz) mem_reserve()
206 it->start + bank_sz); mem_reserve()
215 if (it->end - start <= sz) { mem_reserve()
219 if (it->start - start < sz) mem_reserve()
220 it->start = end; mem_reserve()
258 pfn = PAGE_ALIGN(sysmem.bank[i].start) >> PAGE_SHIFT; bootmem_init()
279 if (sysmem.bank[i].end - sysmem.bank[i].start >= bootmap_size) { bootmem_init()
280 bootmap_start = sysmem.bank[i].start; bootmem_init()
298 if (sysmem.bank[i].start >> PAGE_SHIFT < max_low_pfn) { bootmem_init()
301 free_bootmem(sysmem.bank[i].start, bootmem_init()
302 end - sysmem.bank[i].start); bootmem_init()
370 void free_initrd_mem(unsigned long start, unsigned long end) free_initrd_mem() argument
373 free_reserved_area((void *)start, (void *)end, -1, "initrd"); free_initrd_mem()
/linux-4.4.14/arch/mips/pci/
H A Dpci-malta.c42 .start = 0x00000000UL,
93 resource_size_t start, end, map, start1, end1, map1, map2, map3, mask; mips_pcibios_init() local
115 start = GT_READ(GT_PCI0M0LD_OFS); mips_pcibios_init()
118 end = (end & GT_PCI_HD_MSK) | (start & ~GT_PCI_HD_MSK); mips_pcibios_init()
124 if (end1 - start1 > end - start) { mips_pcibios_init()
125 start = start1; mips_pcibios_init()
129 mask = ~(start ^ end); mips_pcibios_init()
131 BUG_ON((start & GT_PCI_HD_MSK) != (map & GT_PCI_HD_MSK) && mips_pcibios_init()
133 gt64120_mem_resource.start = start; mips_pcibios_init()
135 gt64120_controller.mem_offset = (start & mask) - (map & mask); mips_pcibios_init()
137 gt64120_mem_resource.start <<= GT_PCI_DCRM_SHF; mips_pcibios_init()
142 start = GT_READ(GT_PCI0IOLD_OFS); mips_pcibios_init()
145 end = (end & GT_PCI_HD_MSK) | (start & ~GT_PCI_HD_MSK); mips_pcibios_init()
146 mask = ~(start ^ end); mips_pcibios_init()
148 BUG_ON((start & GT_PCI_HD_MSK) != (map & GT_PCI_HD_MSK) && mips_pcibios_init()
150 gt64120_io_resource.start = map & mask; mips_pcibios_init()
154 gt64120_io_resource.start <<= GT_PCI_DCRM_SHF; mips_pcibios_init()
172 start = BONITO_PCILO0_BASE; mips_pcibios_init()
176 start = BONITO_PCILO1_BASE; mips_pcibios_init()
181 start = BONITO_PCILO0_BASE; mips_pcibios_init()
184 bonito64_mem_resource.start = start; mips_pcibios_init()
185 bonito64_mem_resource.end = start + mips_pcibios_init()
187 bonito64_controller.mem_offset = start - mips_pcibios_init()
198 MSC_READ(MSC01_PCI_SC2PMBASL, start); mips_pcibios_init()
201 msc_mem_resource.start = start & mask; mips_pcibios_init()
202 msc_mem_resource.end = (start & mask) | ~mask; mips_pcibios_init()
203 msc_controller.mem_offset = (start & mask) - (map & mask); mips_pcibios_init()
205 write_gcr_reg0_base(start); mips_pcibios_init()
209 MSC_READ(MSC01_PCI_SC2PIOBASL, start); mips_pcibios_init()
212 msc_io_resource.start = map & mask; mips_pcibios_init()
217 write_gcr_reg1_base(start); mips_pcibios_init()
222 start = start & mask; mips_pcibios_init()
223 end = start | ~mask; mips_pcibios_init()
224 if ((start >= msc_mem_resource.start && mips_pcibios_init()
225 start <= msc_mem_resource.end) || mips_pcibios_init()
226 (end >= msc_mem_resource.start && mips_pcibios_init()
229 start = max(start, msc_mem_resource.start); mips_pcibios_init()
231 if (start - msc_mem_resource.start >= mips_pcibios_init()
233 msc_mem_resource.end = start - 1; mips_pcibios_init()
235 msc_mem_resource.start = end + 1; mips_pcibios_init()
245 if (controller->io_resource->start < 0x00001000UL) mips_pcibios_init()
246 controller->io_resource->start = 0x00001000UL; mips_pcibios_init()
/linux-4.4.14/arch/x86/include/asm/
H A Dpat.h11 extern int reserve_memtype(u64 start, u64 end,
13 extern int free_memtype(u64 start, u64 end);
18 int io_reserve_memtype(resource_size_t start, resource_size_t end,
21 void io_free_memtype(resource_size_t start, resource_size_t end);
H A Dkasan.h8 * Compiler uses shadow offset assuming that addresses start
9 * from 0. Kernel addresses don't start from 0, so shadow
11 * 'kernel address space start' >> KASAN_SHADOW_SCALE_SHIFT
/linux-4.4.14/include/linux/
H A Drange.h5 u64 start; member in struct:range
10 u64 start, u64 end);
14 u64 start, u64 end);
16 void subtract_range(struct range *range, int az, u64 start, u64 end);
H A Dioport.h19 resource_size_t start; member in struct:resource
46 #define IORESOURCE_STARTALIGN 0x00080000 /* start field is alignment */
112 .start = (_start), \
147 resource_size_t start, resource_size_t end,
161 struct resource *lookup_resource(struct resource *root, resource_size_t start);
162 int adjust_resource(struct resource *res, resource_size_t start,
167 return res->end - res->start + 1; resource_size()
180 return r1->start <= r2->start && r1->end >= r2->end; resource_contains()
185 #define request_region(start,n,name) __request_region(&ioport_resource, (start), (n), (name), 0)
186 #define request_muxed_region(start,n,name) __request_region(&ioport_resource, (start), (n), (name), IORESOURCE_MUXED)
187 #define __request_mem_region(start,n,name, excl) __request_region(&iomem_resource, (start), (n), (name), excl)
188 #define request_mem_region(start,n,name) __request_region(&iomem_resource, (start), (n), (name), 0)
189 #define request_mem_region_exclusive(start,n,name) \
190 __request_region(&iomem_resource, (start), (n), (name), IORESOURCE_EXCLUSIVE)
194 resource_size_t start,
199 #define release_region(start,n) __release_region(&ioport_resource, (start), (n))
200 #define release_mem_region(start,n) __release_region(&iomem_resource, (start), (n))
216 #define devm_request_region(dev,start,n,name) \
217 __devm_request_region(dev, &ioport_resource, (start), (n), (name))
218 #define devm_request_mem_region(dev,start,n,name) \
219 __devm_request_region(dev, &iomem_resource, (start), (n), (name))
222 struct resource *parent, resource_size_t start,
225 #define devm_release_region(dev, start, n) \
226 __devm_release_region(dev, &ioport_resource, (start), (n))
227 #define devm_release_mem_region(dev, start, n) \
228 __devm_release_region(dev, &iomem_resource, (start), (n))
231 resource_size_t start, resource_size_t n);
239 walk_system_ram_res(u64 start, u64 end, void *arg,
242 walk_iomem_res(char *name, unsigned long flags, u64 start, u64 end, void *arg,
248 return (r1->start <= r2->end && r1->end >= r2->start); resource_overlaps()
H A Dinterval_tree.h8 unsigned long start; /* Start of interval */ member in struct:interval_tree_node
21 unsigned long start, unsigned long last);
25 unsigned long start, unsigned long last);
H A Dif_tunnel.h13 #define for_each_ip_tunnel_rcu(pos, start) \
14 for (pos = rcu_dereference(start); pos; pos = rcu_dereference(pos->next))
H A Dfirmware-map.h26 int firmware_map_add_early(u64 start, u64 end, const char *type);
27 int firmware_map_add_hotplug(u64 start, u64 end, const char *type);
28 int firmware_map_remove(u64 start, u64 end, const char *type);
32 static inline int firmware_map_add_early(u64 start, u64 end, const char *type) firmware_map_add_early() argument
37 static inline int firmware_map_add_hotplug(u64 start, u64 end, const char *type) firmware_map_add_hotplug() argument
42 static inline int firmware_map_remove(u64 start, u64 end, const char *type) firmware_map_remove() argument
/linux-4.4.14/arch/sparc/include/asm/
H A Dtlbflush_32.h10 #define flush_tlb_range(vma, start, end) \
11 sparc32_cachetlb_ops->tlb_range(vma, start, end)
18 static inline void flush_tlb_kernel_range(unsigned long start, flush_tlb_kernel_range() argument
/linux-4.4.14/arch/mips/lasat/image/
H A Dhead.S4 .section .text..start, "ax"
15 /* image start and size */
19 /* start of kernel and entrypoint in uncompressed image */
/linux-4.4.14/arch/blackfin/include/asm/
H A Dcacheflush.h28 #define flush_cache_range(vma, start, end) do { } while (0)
30 #define flush_cache_vmap(start, end) do { } while (0)
31 #define flush_cache_vunmap(start, end) do { } while (0)
34 #define flush_icache_range_others(start, end) \
35 smp_icache_flush_range_others((start), (end))
37 #define flush_icache_range_others(start, end) do { } while (0)
40 static inline void flush_icache_range(unsigned start, unsigned end) flush_icache_range() argument
44 blackfin_dcache_flush_range(start, end); flush_icache_range()
47 if (start >= L2_START && end <= L2_START + L2_LENGTH) flush_icache_range()
48 blackfin_dcache_flush_range(start, end); flush_icache_range()
62 blackfin_icache_flush_range(start, end); flush_icache_range()
63 flush_icache_range_others(start, end); flush_icache_range()
67 if (start >= L2_START && end <= L2_START + L2_LENGTH) { flush_icache_range()
68 blackfin_icache_flush_range(start, end); flush_icache_range()
69 flush_icache_range_others(start, end); flush_icache_range()
82 # define invalidate_dcache_range(start,end) blackfin_dcache_invalidate_range((start), (end))
84 # define invalidate_dcache_range(start,end) do { } while (0)
87 # define flush_dcache_range(start,end) blackfin_dcache_flush_range((start), (end))
91 # define flush_dcache_range(start,end) do { } while (0)
/linux-4.4.14/arch/arm/mm/
H A Dcache-feroceon-l2.c33 * Cache range operations are initiated by writing the start and
36 * [start:end].
73 static inline void l2_clean_pa_range(unsigned long start, unsigned long end) l2_clean_pa_range() argument
78 * Make sure 'start' and 'end' reference the same page, as l2_clean_pa_range()
80 * the start address. l2_clean_pa_range()
82 BUG_ON((start ^ end) >> PAGE_SHIFT); l2_clean_pa_range()
84 va_start = l2_get_va(start); l2_clean_pa_range()
85 va_end = va_start + (end - start); l2_clean_pa_range()
104 static inline void l2_inv_pa_range(unsigned long start, unsigned long end) l2_inv_pa_range() argument
109 * Make sure 'start' and 'end' reference the same page, as l2_inv_pa_range()
111 * the start address. l2_inv_pa_range()
113 BUG_ON((start ^ end) >> PAGE_SHIFT); l2_inv_pa_range()
115 va_start = l2_get_va(start); l2_inv_pa_range()
116 va_end = va_start + (end - start); l2_inv_pa_range()
135 * inclusive start and end addresses.
142 static unsigned long calc_range_end(unsigned long start, unsigned long end) calc_range_end() argument
146 BUG_ON(start & (CACHE_LINE_SIZE - 1)); calc_range_end()
150 * Try to process all cache lines between 'start' and 'end'. calc_range_end()
159 if (range_end > start + MAX_RANGE_SIZE) calc_range_end()
160 range_end = start + MAX_RANGE_SIZE; calc_range_end()
165 if (range_end > (start | (PAGE_SIZE - 1)) + 1) calc_range_end()
166 range_end = (start | (PAGE_SIZE - 1)) + 1; calc_range_end()
171 static void feroceon_l2_inv_range(unsigned long start, unsigned long end) feroceon_l2_inv_range() argument
176 if (start & (CACHE_LINE_SIZE - 1)) { feroceon_l2_inv_range()
177 l2_clean_inv_pa(start & ~(CACHE_LINE_SIZE - 1)); feroceon_l2_inv_range()
178 start = (start | (CACHE_LINE_SIZE - 1)) + 1; feroceon_l2_inv_range()
184 if (start < end && end & (CACHE_LINE_SIZE - 1)) { feroceon_l2_inv_range()
190 * Invalidate all full cache lines between 'start' and 'end'. feroceon_l2_inv_range()
192 while (start < end) { feroceon_l2_inv_range()
193 unsigned long range_end = calc_range_end(start, end); feroceon_l2_inv_range()
194 l2_inv_pa_range(start, range_end - CACHE_LINE_SIZE); feroceon_l2_inv_range()
195 start = range_end; feroceon_l2_inv_range()
201 static void feroceon_l2_clean_range(unsigned long start, unsigned long end) feroceon_l2_clean_range() argument
208 start &= ~(CACHE_LINE_SIZE - 1); feroceon_l2_clean_range()
210 while (start != end) { feroceon_l2_clean_range()
211 unsigned long range_end = calc_range_end(start, end); feroceon_l2_clean_range()
212 l2_clean_pa_range(start, range_end - CACHE_LINE_SIZE); feroceon_l2_clean_range()
213 start = range_end; feroceon_l2_clean_range()
220 static void feroceon_l2_flush_range(unsigned long start, unsigned long end) feroceon_l2_flush_range() argument
222 start &= ~(CACHE_LINE_SIZE - 1); feroceon_l2_flush_range()
224 while (start != end) { feroceon_l2_flush_range()
225 unsigned long range_end = calc_range_end(start, end); feroceon_l2_flush_range()
227 l2_clean_pa_range(start, range_end - CACHE_LINE_SIZE); feroceon_l2_flush_range()
228 l2_inv_pa_range(start, range_end - CACHE_LINE_SIZE); feroceon_l2_flush_range()
229 start = range_end; feroceon_l2_flush_range()
H A Dcache-v4wt.S78 * flush_user_cache_range(start, end, flags)
83 * - start - start address (inclusive, page aligned)
101 * coherent_kern_range(start, end)
104 * region described by start. If you have non-snooping
107 * - start - virtual start address
114 * coherent_user_range(start, end)
117 * region described by start. If you have non-snooping
120 * - start - virtual start address
148 * dma_inv_range(start, end)
151 * May not write back any entries. If 'start' or 'end'
155 * - start - virtual start address
167 * dma_flush_range(start, end)
171 * - start - virtual start address
178 * dma_unmap_area(start, size, dir)
179 * - start - kernel virtual start address
190 * dma_map_area(start, size, dir)
191 * - start - kernel virtual start address
H A Dcache-xsc3l2.c98 static void xsc3_l2_inv_range(unsigned long start, unsigned long end) xsc3_l2_inv_range() argument
102 if (start == 0 && end == -1ul) { xsc3_l2_inv_range()
112 if (start & (CACHE_LINE_SIZE - 1)) { xsc3_l2_inv_range()
113 vaddr = l2_map_va(start & ~(CACHE_LINE_SIZE - 1), vaddr); xsc3_l2_inv_range()
116 start = (start | (CACHE_LINE_SIZE - 1)) + 1; xsc3_l2_inv_range()
120 * Invalidate all full cache lines between 'start' and 'end'. xsc3_l2_inv_range()
122 while (start < (end & ~(CACHE_LINE_SIZE - 1))) { xsc3_l2_inv_range()
123 vaddr = l2_map_va(start, vaddr); xsc3_l2_inv_range()
125 start += CACHE_LINE_SIZE; xsc3_l2_inv_range()
131 if (start < end) { xsc3_l2_inv_range()
132 vaddr = l2_map_va(start, vaddr); xsc3_l2_inv_range()
142 static void xsc3_l2_clean_range(unsigned long start, unsigned long end) xsc3_l2_clean_range() argument
148 start &= ~(CACHE_LINE_SIZE - 1); xsc3_l2_clean_range()
149 while (start < end) { xsc3_l2_clean_range()
150 vaddr = l2_map_va(start, vaddr); xsc3_l2_clean_range()
152 start += CACHE_LINE_SIZE; xsc3_l2_clean_range()
180 static void xsc3_l2_flush_range(unsigned long start, unsigned long end) xsc3_l2_flush_range() argument
184 if (start == 0 && end == -1ul) { xsc3_l2_flush_range()
191 start &= ~(CACHE_LINE_SIZE - 1); xsc3_l2_flush_range()
192 while (start < end) { xsc3_l2_flush_range()
193 vaddr = l2_map_va(start, vaddr); xsc3_l2_flush_range()
196 start += CACHE_LINE_SIZE; xsc3_l2_flush_range()
H A Dpageattr.c40 unsigned long start = addr; change_memory_common() local
42 unsigned long end = start + size; change_memory_common()
47 start &= PAGE_MASK; change_memory_common()
48 end = start + size; change_memory_common()
52 if (start < MODULES_VADDR || start >= MODULES_END) change_memory_common()
55 if (end < MODULES_VADDR || start >= MODULES_END) change_memory_common()
61 ret = apply_to_page_range(&init_mm, start, size, change_page_range, change_memory_common()
64 flush_tlb_kernel_range(start, end); change_memory_common()
H A Dcache-fa.S78 * flush_user_cache_range(start, end, flags)
83 * - start - start address (inclusive, page aligned)
106 * coherent_kern_range(start, end)
109 * region described by start. If you have non-snooping
112 * - start - virtual start address
119 * coherent_user_range(start, end)
122 * region described by start. If you have non-snooping
125 * - start - virtual start address
162 * dma_inv_range(start, end)
165 * May not write back any entries. If 'start' or 'end'
169 * - start - virtual start address
188 * dma_clean_range(start, end)
192 * - start - virtual start address
206 * dma_flush_range(start,end)
207 * - start - virtual start address of region
221 * dma_map_area(start, size, dir)
222 * - start - kernel virtual start address
235 * dma_unmap_area(start, size, dir)
236 * - start - kernel virtual start address
H A Dcache-v4wb.S101 * flush_user_cache_range(start, end, flags)
106 * - start - start address (inclusive, page aligned)
142 * coherent_kern_range(start, end)
145 * region described by start. If you have non-snooping
148 * - start - virtual start address
155 * coherent_user_range(start, end)
158 * region described by start. If you have non-snooping
161 * - start - virtual start address
178 * dma_inv_range(start, end)
181 * May not write back any entries. If 'start' or 'end'
185 * - start - virtual start address
202 * dma_clean_range(start, end)
206 * - start - virtual start address
219 * dma_flush_range(start, end)
223 * - start - virtual start address
232 * dma_map_area(start, size, dir)
233 * - start - kernel virtual start address
246 * dma_unmap_area(start, size, dir)
247 * - start - kernel virtual start address
H A Dcache-v6.S89 * v6_flush_cache_range(start, end, flags)
93 * - start - start address (may not be aligned)
104 * v6_coherent_kern_range(start,end)
110 * - start - virtual start address of region
120 * v6_coherent_user_range(start,end)
126 * - start - virtual start address of region
195 * v6_dma_inv_range(start,end)
201 * - start - virtual start address of region
245 * v6_dma_clean_range(start,end)
246 * - start - virtual start address of region
268 * v6_dma_flush_range(start,end)
269 * - start - virtual start address of region
296 * dma_map_area(start, size, dir)
297 * - start - kernel virtual start address
315 * dma_unmap_area(start, size, dir)
316 * - start - kernel virtual start address
/linux-4.4.14/tools/perf/util/
H A Dsvghelper.h6 extern void open_svg(const char *filename, int cpus, int rows, u64 start, u64 end);
7 extern void svg_ubox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int merges);
8 extern void svg_lbox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int merges);
9 extern void svg_fbox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int merges);
10 extern void svg_box(int Yslot, u64 start, u64 end, const char *type);
11 extern void svg_blocked(int Yslot, int cpu, u64 start, u64 end, const char *backtrace);
12 extern void svg_running(int Yslot, int cpu, u64 start, u64 end, const char *backtrace);
13 extern void svg_waiting(int Yslot, int cpu, u64 start, u64 end, const char *backtrace);
17 extern void svg_process(int cpu, u64 start, u64 end, int pid, const char *name, const char *backtrace);
18 extern void svg_cstate(int cpu, u64 start, u64 end, int type);
19 extern void svg_pstate(int cpu, u64 start, u64 end, u64 freq);
25 extern void svg_wakeline(u64 start, int row1, int row2, const char *backtrace);
26 extern void svg_partial_wakeline(u64 start, int row1, char *desc1, int row2, char *desc2, const char *backtrace);
27 extern void svg_interrupt(u64 start, int row, const char *backtrace);
28 extern void svg_text(int Yslot, u64 start, const char *text);
H A Dfind-vdso-map.c1 static int find_vdso_map(void **start, void **end) find_vdso_map() argument
18 start, end, &m)) find_vdso_map()
H A Dsvghelper.c86 void open_svg(const char *filename, int cpus, int rows, u64 start, u64 end) open_svg() argument
95 first_time = start; open_svg()
154 void svg_ubox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int merges) svg_ubox() argument
156 double w = time2pixels(end) - time2pixels(start); svg_ubox()
165 time2pixels(start), svg_ubox()
173 void svg_lbox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int merges) svg_lbox() argument
175 double w = time2pixels(end) - time2pixels(start); svg_lbox()
184 time2pixels(start), svg_lbox()
192 void svg_fbox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int merges) svg_fbox() argument
194 double w = time2pixels(end) - time2pixels(start); svg_fbox()
203 time2pixels(start), svg_fbox()
211 void svg_box(int Yslot, u64 start, u64 end, const char *type) svg_box() argument
217 time2pixels(start), time2pixels(end)-time2pixels(start), Yslot * SLOT_MULT, SLOT_HEIGHT, type); svg_box()
221 void svg_blocked(int Yslot, int cpu, u64 start, u64 end, const char *backtrace) svg_blocked() argument
228 time_to_string(end - start)); svg_blocked()
231 svg_box(Yslot, start, end, "blocked"); svg_blocked()
235 void svg_running(int Yslot, int cpu, u64 start, u64 end, const char *backtrace) svg_running() argument
243 if (svg_highlight && end - start > svg_highlight) svg_running()
250 cpu, time_to_string(end - start)); svg_running()
254 time2pixels(start), time2pixels(end)-time2pixels(start), Yslot * SLOT_MULT, SLOT_HEIGHT, svg_running()
257 text_size = (time2pixels(end)-time2pixels(start)); svg_running()
266 time2pixels(start), Yslot * SLOT_MULT + SLOT_HEIGHT - 1, text_size, cpu + 1); svg_running()
289 void svg_waiting(int Yslot, int cpu, u64 start, u64 end, const char *backtrace) svg_waiting() argument
300 if (end-start > 10 * 1000000) /* 10 msec */ svg_waiting()
303 text = time_to_string(end-start); svg_waiting()
305 font_size = 1.0 * (time2pixels(end)-time2pixels(start)); svg_waiting()
312 fprintf(svgfile, "<g transform=\"translate(%.8f,%.8f)\">\n", time2pixels(start), Yslot * SLOT_MULT); svg_waiting()
313 fprintf(svgfile, "<title>#%d waiting %s</title>\n", cpu, time_to_string(end - start)); svg_waiting()
317 time2pixels(end)-time2pixels(start), SLOT_HEIGHT, style); svg_waiting()
383 void svg_process(int cpu, u64 start, u64 end, int pid, const char *name, const char *backtrace) svg_process() argument
391 if (svg_highlight && end - start >= svg_highlight) svg_process()
398 fprintf(svgfile, "<g transform=\"translate(%.8f,%.8f)\">\n", time2pixels(start), cpu2y(cpu)); svg_process()
399 fprintf(svgfile, "<title>%d %s running %s</title>\n", pid, name, time_to_string(end - start)); svg_process()
403 time2pixels(end)-time2pixels(start), SLOT_MULT+SLOT_HEIGHT, type); svg_process()
404 width = time2pixels(end)-time2pixels(start); svg_process()
417 void svg_cstate(int cpu, u64 start, u64 end, int type) svg_cstate() argument
434 time2pixels(start), time2pixels(end)-time2pixels(start), svg_cstate()
437 width = (time2pixels(end)-time2pixels(start))/2.0; svg_cstate()
445 time2pixels(start), cpu2y(cpu)+width, width, type); svg_cstate()
474 void svg_pstate(int cpu, u64 start, u64 end, u64 freq) svg_pstate() argument
487 time2pixels(start), time2pixels(end), height, height); svg_pstate()
489 time2pixels(start), height+0.9, HzToHuman(freq)); svg_pstate()
495 void svg_partial_wakeline(u64 start, int row1, char *desc1, int row2, char *desc2, const char *backtrace) svg_partial_wakeline() argument
515 time2pixels(start), row1 * SLOT_MULT + SLOT_HEIGHT, time2pixels(start), row1 * SLOT_MULT + SLOT_HEIGHT + SLOT_MULT/32); svg_partial_wakeline()
518 time2pixels(start), row1 * SLOT_MULT + SLOT_HEIGHT + SLOT_HEIGHT/48, desc2); svg_partial_wakeline()
522 time2pixels(start), row2 * SLOT_MULT - SLOT_MULT/32, time2pixels(start), row2 * SLOT_MULT); svg_partial_wakeline()
525 time2pixels(start), row2 * SLOT_MULT - SLOT_MULT/32, desc1); svg_partial_wakeline()
530 time2pixels(start), row2 * SLOT_MULT + SLOT_HEIGHT, time2pixels(start), row2 * SLOT_MULT + SLOT_HEIGHT + SLOT_MULT/32); svg_partial_wakeline()
533 time2pixels(start), row2 * SLOT_MULT + SLOT_HEIGHT + SLOT_MULT/48, desc1); svg_partial_wakeline()
537 time2pixels(start), row1 * SLOT_MULT - SLOT_MULT/32, time2pixels(start), row1 * SLOT_MULT); svg_partial_wakeline()
540 time2pixels(start), row1 * SLOT_MULT - SLOT_HEIGHT/32, desc2); svg_partial_wakeline()
548 time2pixels(start), height); svg_partial_wakeline()
553 void svg_wakeline(u64 start, int row1, int row2, const char *backtrace) svg_wakeline() argument
568 time2pixels(start), row1 * SLOT_MULT + SLOT_HEIGHT, time2pixels(start), row2 * SLOT_MULT); svg_wakeline()
571 time2pixels(start), row2 * SLOT_MULT + SLOT_HEIGHT, time2pixels(start), row1 * SLOT_MULT); svg_wakeline()
577 time2pixels(start), height); svg_wakeline()
582 void svg_interrupt(u64 start, int row, const char *backtrace) svg_interrupt() argument
595 time2pixels(start), row * SLOT_MULT); svg_interrupt()
597 time2pixels(start), row * SLOT_MULT + SLOT_HEIGHT); svg_interrupt()
602 void svg_text(int Yslot, u64 start, const char *text) svg_text() argument
608 time2pixels(start), Yslot * SLOT_MULT+SLOT_HEIGHT/2, text); svg_text()
/linux-4.4.14/arch/s390/mm/
H A Dvmem.c24 unsigned long start; member in struct:memory_segment
78 static int vmem_add_mem(unsigned long start, unsigned long size, int ro) vmem_add_mem() argument
80 unsigned long end = start + size; vmem_add_mem()
81 unsigned long address = start; vmem_add_mem()
146 static void vmem_remove_range(unsigned long start, unsigned long size) vmem_remove_range() argument
148 unsigned long end = start + size; vmem_remove_range()
149 unsigned long address = start; vmem_remove_range()
187 flush_tlb_kernel_range(start, end); vmem_remove_range()
193 int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node) vmemmap_populate() argument
195 unsigned long address = start; vmemmap_populate()
202 for (address = start; address < end;) { vmemmap_populate()
264 void vmemmap_free(unsigned long start, unsigned long end) vmemmap_free() argument
276 if (seg->start + seg->size > VMEM_MAX_PHYS || insert_memory_segment()
277 seg->start + seg->size < seg->start) insert_memory_segment()
281 if (seg->start >= tmp->start + tmp->size) insert_memory_segment()
283 if (seg->start + seg->size <= tmp->start) insert_memory_segment()
302 vmem_remove_range(seg->start, seg->size); __remove_shared_memory()
305 int vmem_remove_mapping(unsigned long start, unsigned long size) vmem_remove_mapping() argument
314 if (seg->start == start && seg->size == size) vmem_remove_mapping()
318 if (seg->start != start || seg->size != size) vmem_remove_mapping()
329 int vmem_add_mapping(unsigned long start, unsigned long size) vmem_add_mapping() argument
339 seg->start = start; vmem_add_mapping()
346 ret = vmem_add_mem(start, size, 0); vmem_add_mapping()
369 phys_addr_t start, end; vmem_map_init() local
374 start = reg->base; for_each_memblock()
376 if (start >= ro_end || end <= ro_start) for_each_memblock()
377 vmem_add_mem(start, end - start, 0); for_each_memblock()
378 else if (start >= ro_start && end <= ro_end) for_each_memblock()
379 vmem_add_mem(start, end - start, 1); for_each_memblock()
380 else if (start >= ro_start) { for_each_memblock()
381 vmem_add_mem(start, ro_end - start, 1); for_each_memblock()
384 vmem_add_mem(start, ro_start - start, 0); for_each_memblock()
387 vmem_add_mem(start, ro_start - start, 0); for_each_memblock()
408 seg->start = reg->base; for_each_memblock()
H A Dextable.c48 void sort_extable(struct exception_table_entry *start, sort_extable() argument
54 /* Normalize entries to being relative to the start of the section */ sort_extable()
55 for (p = start, i = 0; p < finish; p++, i += 8) { sort_extable()
59 sort(start, finish - start, sizeof(*start), cmp_ex, NULL); sort_extable()
61 for (p = start, i = 0; p < finish; p++, i += 8) { sort_extable()
/linux-4.4.14/tools/perf/scripts/python/
H A Dnet_dropmonitor.py33 # Invariant: kallsyms[i][0] <= loc for all 0 <= i <= start
35 start, end = -1, len(kallsyms)
36 while end != start + 1:
37 pivot = (start + end) // 2
41 start = pivot
43 # Now (start == -1 or kallsyms[start][0] <= loc)
44 # and (start == len(kallsyms) - 1 or loc < kallsyms[start + 1][0])
45 if start >= 0:
46 symloc, name = kallsyms[start]
/linux-4.4.14/tools/testing/nvdimm/test/
H A Diomap.c64 - nfit_res->res->start; __nfit_test_ioremap()
78 - nfit_res->res->start; __wrap_devm_ioremap_nocache()
92 return nfit_res->buf + offset - nfit_res->res->start; __wrap_devm_memremap()
106 return nfit_res->buf + offset - nfit_res->res->start; __wrap_memremap()
163 struct resource *parent, resource_size_t start, nfit_test_request_region()
170 nfit_res = get_nfit_res(start); nfit_test_request_region()
175 if (start + n > nfit_res->res->start nfit_test_request_region()
177 pr_debug("%s: start: %llx n: %llx overflow: %pr\n", nfit_test_request_region()
178 __func__, start, n, nfit_test_request_region()
183 res->start = start; nfit_test_request_region()
184 res->end = start + n - 1; nfit_test_request_region()
193 return __devm_request_region(dev, parent, start, n, name); nfit_test_request_region()
194 return __request_region(parent, start, n, name, flags); nfit_test_request_region()
198 resource_size_t start, resource_size_t n, const char *name, __wrap___request_region()
201 return nfit_test_request_region(NULL, parent, start, n, name, flags); __wrap___request_region()
206 struct resource *parent, resource_size_t start, __wrap___devm_request_region()
211 return nfit_test_request_region(dev, parent, start, n, name, 0); __wrap___devm_request_region()
215 void __wrap___release_region(struct resource *parent, resource_size_t start, __wrap___release_region() argument
222 nfit_res = get_nfit_res(start); __wrap___release_region()
227 if (start != res->start || resource_size(res) != n) __wrap___release_region()
228 pr_info("%s: start: %llx n: %llx mismatch: %pr\n", __wrap___release_region()
229 __func__, start, n, res); __wrap___release_region()
235 __release_region(parent, start, n); __wrap___release_region()
162 nfit_test_request_region(struct device *dev, struct resource *parent, resource_size_t start, resource_size_t n, const char *name, int flags) nfit_test_request_region() argument
197 __wrap___request_region(struct resource *parent, resource_size_t start, resource_size_t n, const char *name, int flags) __wrap___request_region() argument
205 __wrap___devm_request_region(struct device *dev, struct resource *parent, resource_size_t start, resource_size_t n, const char *name) __wrap___devm_request_region() argument
/linux-4.4.14/arch/arm/mach-davinci/
H A Ddevices-da8xx.c163 .start = DA8XX_TPCC_BASE,
169 .start = DA8XX_TPTC0_BASE,
175 .start = DA8XX_TPTC1_BASE,
181 .start = IRQ_DA8XX_CCINT0,
186 .start = IRQ_DA8XX_CCERRINT,
194 .start = DA850_TPCC1_BASE,
200 .start = DA850_TPTC2_BASE,
206 .start = IRQ_DA850_CCINT1,
211 .start = IRQ_DA850_CCERRINT1,
266 .start = DA8XX_I2C0_BASE,
271 .start = IRQ_DA8XX_I2CINT0,
286 .start = DA8XX_I2C1_BASE,
291 .start = IRQ_DA8XX_I2CINT1,
322 .start = DA8XX_WDOG_BASE,
355 .start = DA8XX_EMAC_CPPI_PORT_BASE,
360 .start = IRQ_DA8XX_C0_RX_THRESH_PULSE,
365 .start = IRQ_DA8XX_C0_RX_PULSE,
370 .start = IRQ_DA8XX_C0_TX_PULSE,
375 .start = IRQ_DA8XX_C0_MISC_PULSE,
401 .start = DA8XX_EMAC_MDIO_BASE,
428 .start = DAVINCI_DA830_MCASP1_REG_BASE,
435 .start = DAVINCI_DA830_DMA_MCASP1_AXEVT,
442 .start = DAVINCI_DA830_DMA_MCASP1_AREVT,
448 .start = IRQ_DA8XX_MCASPINT,
463 .start = DAVINCI_DA830_MCASP2_REG_BASE,
470 .start = DAVINCI_DA830_DMA_MCASP2_AXEVT,
477 .start = DAVINCI_DA830_DMA_MCASP2_AREVT,
483 .start = IRQ_DA8XX_MCASPINT,
498 .start = DAVINCI_DA8XX_MCASP0_REG_BASE,
505 .start = DAVINCI_DA8XX_DMA_MCASP0_AXEVT,
512 .start = DAVINCI_DA8XX_DMA_MCASP0_AREVT,
518 .start = IRQ_DA8XX_MCASPINT,
561 .start = DA8XX_PRUSS_MEM_BASE,
566 .start = IRQ_DA8XX_EVTOUT0,
571 .start = IRQ_DA8XX_EVTOUT1,
576 .start = IRQ_DA8XX_EVTOUT2,
581 .start = IRQ_DA8XX_EVTOUT3,
586 .start = IRQ_DA8XX_EVTOUT4,
591 .start = IRQ_DA8XX_EVTOUT5,
596 .start = IRQ_DA8XX_EVTOUT6,
601 .start = IRQ_DA8XX_EVTOUT7,
647 .start = DA8XX_LCD_CNTRL_BASE,
652 .start = IRQ_DA8XX_LCDINT,
673 .start = DA8XX_GPIO_BASE,
678 .start = IRQ_DA8XX_GPIO0,
699 .start = DA8XX_MMCSD0_BASE,
704 .start = IRQ_DA8XX_MMCSDINT0,
709 .start = DA8XX_DMA_MMCSD0_RX,
714 .start = DA8XX_DMA_MMCSD0_TX,
736 .start = DA850_MMCSD1_BASE,
741 .start = IRQ_DA850_MMCSDINT0_1,
746 .start = DA850_DMA_MMCSD1_RX,
751 .start = DA850_DMA_MMCSD1_TX,
773 .start = DA8XX_SYSCFG0_BASE + DA8XX_HOST1CFG_REG,
778 .start = DA8XX_SYSCFG0_BASE + DA8XX_CHIPSIG_REG,
783 .start = IRQ_DA8XX_CHIPINT0,
859 .start = DA8XX_RTC_BASE,
864 .start = IRQ_DA8XX_RTC,
869 .start = IRQ_DA8XX_RTC,
902 .start = DA8XX_DDR2_CTL_BASE,
932 .start = DA8XX_SPI0_BASE,
937 .start = IRQ_DA8XX_SPINT0,
942 .start = DA8XX_DMA_SPI0_RX,
947 .start = DA8XX_DMA_SPI0_TX,
955 .start = DA830_SPI1_BASE,
960 .start = IRQ_DA8XX_SPINT1,
965 .start = DA8XX_DMA_SPI1_RX,
970 .start = DA8XX_DMA_SPI1_TX,
1020 da8xx_spi1_resources[0].start = DA850_SPI1_BASE; da8xx_register_spi_bus()
1030 .start = DA850_SATA_BASE,
1035 .start = DA8XX_SYSCFG1_BASE + DA8XX_PWRDN_REG,
1040 .start = IRQ_DA850_SATAINT,
H A Dusb.c55 .start = DAVINCI_USB_OTG_BASE,
60 .start = IRQ_USBINT,
92 usb_dev.resource[1].start = IRQ_DM646X_USBINT; davinci_setup_usb()
93 usb_dev.resource[2].start = IRQ_DM646X_USBDMAINT; davinci_setup_usb()
103 .start = DA8XX_USB0_BASE,
108 .start = IRQ_DA8XX_USB_INT,
146 .start = DA8XX_USB1_BASE,
151 .start = IRQ_DA8XX_IRQN,
/linux-4.4.14/drivers/gpu/drm/exynos/
H A Dregs-fimc.h27 /* Y 1st frame start address for output DMA */
29 /* Y 2nd frame start address for output DMA */
31 /* Y 3rd frame start address for output DMA */
33 /* Y 4th frame start address for output DMA */
35 /* Cb 1st frame start address for output DMA */
37 /* Cb 2nd frame start address for output DMA */
39 /* Cb 3rd frame start address for output DMA */
41 /* Cb 4th frame start address for output DMA */
43 /* Cr 1st frame start address for output DMA */
45 /* Cr 2nd frame start address for output DMA */
47 /* Cr 3rd frame start address for output DMA */
49 /* Cr 4th frame start address for output DMA */
73 /* Y frame start address for input DMA */
75 /* Cb frame start address for input DMA */
77 /* Cr frame start address for input DMA */
89 /* Y frame start address for input DMA */
91 /* Cb frame start address for input DMA */
93 /* Cr frame start address for input DMA */
123 /* Y 5th frame start address for output DMA */
125 /* Y 6th frame start address for output DMA */
127 /* Y 7th frame start address for output DMA */
129 /* Y 8th frame start address for output DMA */
131 /* Y 9th frame start address for output DMA */
133 /* Y 10th frame start address for output DMA */
135 /* Y 11th frame start address for output DMA */
137 /* Y 12th frame start address for output DMA */
139 /* Y 13th frame start address for output DMA */
141 /* Y 14th frame start address for output DMA */
143 /* Y 15th frame start address for output DMA */
145 /* Y 16th frame start address for output DMA */
147 /* Y 17th frame start address for output DMA */
149 /* Y 18th frame start address for output DMA */
151 /* Y 19th frame start address for output DMA */
153 /* Y 20th frame start address for output DMA */
155 /* Y 21th frame start address for output DMA */
157 /* Y 22th frame start address for output DMA */
159 /* Y 23th frame start address for output DMA */
161 /* Y 24th frame start address for output DMA */
163 /* Y 25th frame start address for output DMA */
165 /* Y 26th frame start address for output DMA */
167 /* Y 27th frame start address for output DMA */
169 /* Y 28th frame start address for output DMA */
171 /* Y 29th frame start address for output DMA */
173 /* Y 30th frame start address for output DMA */
175 /* Y 31th frame start address for output DMA */
177 /* Y 32th frame start address for output DMA */
180 /* CB 5th frame start address for output DMA */
182 /* CB 6th frame start address for output DMA */
184 /* CB 7th frame start address for output DMA */
186 /* CB 8th frame start address for output DMA */
188 /* CB 9th frame start address for output DMA */
190 /* CB 10th frame start address for output DMA */
192 /* CB 11th frame start address for output DMA */
194 /* CB 12th frame start address for output DMA */
196 /* CB 13th frame start address for output DMA */
198 /* CB 14th frame start address for output DMA */
200 /* CB 15th frame start address for output DMA */
202 /* CB 16th frame start address for output DMA */
204 /* CB 17th frame start address for output DMA */
206 /* CB 18th frame start address for output DMA */
208 /* CB 19th frame start address for output DMA */
210 /* CB 20th frame start address for output DMA */
212 /* CB 21th frame start address for output DMA */
214 /* CB 22th frame start address for output DMA */
216 /* CB 23th frame start address for output DMA */
218 /* CB 24th frame start address for output DMA */
220 /* CB 25th frame start address for output DMA */
222 /* CB 26th frame start address for output DMA */
224 /* CB 27th frame start address for output DMA */
226 /* CB 28th frame start address for output DMA */
228 /* CB 29th frame start address for output DMA */
230 /* CB 30th frame start address for output DMA */
232 /* CB 31th frame start address for output DMA */
234 /* CB 32th frame start address for output DMA */
237 /* CR 5th frame start address for output DMA */
239 /* CR 6th frame start address for output DMA */
241 /* CR 7th frame start address for output DMA */
243 /* CR 8th frame start address for output DMA */
245 /* CR 9th frame start address for output DMA */
247 /* CR 10th frame start address for output DMA */
249 /* CR 11th frame start address for output DMA */
251 /* CR 12th frame start address for output DMA */
253 /* CR 13th frame start address for output DMA */
255 /* CR 14th frame start address for output DMA */
257 /* CR 15th frame start address for output DMA */
259 /* CR 16th frame start address for output DMA */
261 /* CR 17th frame start address for output DMA */
263 /* CR 18th frame start address for output DMA */
265 /* CR 19th frame start address for output DMA */
267 /* CR 20th frame start address for output DMA */
269 /* CR 21th frame start address for output DMA */
271 /* CR 22th frame start address for output DMA */
273 /* CR 23th frame start address for output DMA */
275 /* CR 24th frame start address for output DMA */
277 /* CR 25th frame start address for output DMA */
279 /* CR 26th frame start address for output DMA */
281 /* CR 27th frame start address for output DMA */
283 /* CR 28th frame start address for output DMA */
285 /* CR 29th frame start address for output DMA */
287 /* CR 30th frame start address for output DMA */
289 /* CR 31th frame start address for output DMA */
291 /* CR 32th frame start address for output DMA */
297 /* frame start address 1 ~ 4, 5 ~ 32 */
/linux-4.4.14/arch/sparc/lib/
H A Duser_fixup.c21 static unsigned long compute_size(unsigned long start, unsigned long size, unsigned long *offset) compute_size() argument
24 unsigned long end = start + size; compute_size()
26 if (fault_addr < start || fault_addr >= end) { compute_size()
29 *offset = fault_addr - start; compute_size()
58 unsigned long start = (unsigned long) to; copy_in_user_fixup() local
59 unsigned long end = start + size; copy_in_user_fixup()
61 if (fault_addr >= start && fault_addr < end) copy_in_user_fixup()
64 start = (unsigned long) from; copy_in_user_fixup()
65 end = start + size; copy_in_user_fixup()
66 if (fault_addr >= start && fault_addr < end) copy_in_user_fixup()
/linux-4.4.14/arch/sh/include/asm/
H A Dtlbflush.h10 * - flush_tlb_range(vma, start, end) flushes a range of pages
11 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages
16 unsigned long start,
20 extern void local_flush_tlb_kernel_range(unsigned long start,
30 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
33 extern void flush_tlb_kernel_range(unsigned long start, unsigned long end);
43 #define flush_tlb_range(vma, start, end) \
44 local_flush_tlb_range(vma, start, end)
46 #define flush_tlb_kernel_range(start, end) \
47 local_flush_tlb_kernel_range(start, end)
H A Dtlb.h24 unsigned long start, end; member in struct:mmu_gather
29 tlb->start = TASK_SIZE; init_tlb_gather()
33 tlb->start = 0; init_tlb_gather()
39 tlb_gather_mmu(struct mmu_gather *tlb, struct mm_struct *mm, unsigned long start, unsigned long end) tlb_gather_mmu() argument
42 tlb->start = start; tlb_gather_mmu()
44 tlb->fullmm = !(start | (end+1)); tlb_gather_mmu()
50 tlb_finish_mmu(struct mmu_gather *tlb, unsigned long start, unsigned long end) tlb_finish_mmu() argument
62 if (tlb->start > address) tlb_remove_tlb_entry()
63 tlb->start = address; tlb_remove_tlb_entry()
84 flush_tlb_range(vma, tlb->start, tlb->end); tlb_end_vma()
/linux-4.4.14/arch/arm/mach-omap1/
H A Dmcbsp.c100 .start = OMAP7XX_MCBSP1_BASE,
106 .start = INT_7XX_McBSP1RX,
111 .start = INT_7XX_McBSP1TX,
116 .start = 9,
121 .start = 8,
127 .start = OMAP7XX_MCBSP2_BASE,
133 .start = INT_7XX_McBSP2RX,
138 .start = INT_7XX_McBSP2TX,
143 .start = 11,
148 .start = 10,
177 .start = OMAP1510_MCBSP1_BASE,
183 .start = INT_McBSP1RX,
188 .start = INT_McBSP1TX,
193 .start = 9,
198 .start = 8,
204 .start = OMAP1510_MCBSP2_BASE,
210 .start = INT_1510_SPI_RX,
215 .start = INT_1510_SPI_TX,
220 .start = 17,
225 .start = 16,
231 .start = OMAP1510_MCBSP3_BASE,
237 .start = INT_McBSP3RX,
242 .start = INT_McBSP3TX,
247 .start = 11,
252 .start = 10,
284 .start = OMAP1610_MCBSP1_BASE,
290 .start = INT_McBSP1RX,
295 .start = INT_McBSP1TX,
300 .start = 9,
305 .start = 8,
311 .start = OMAP1610_MCBSP2_BASE,
317 .start = INT_1610_McBSP2_RX,
322 .start = INT_1610_McBSP2_TX,
327 .start = 17,
332 .start = 16,
338 .start = OMAP1610_MCBSP3_BASE,
344 .start = INT_McBSP3RX,
349 .start = INT_McBSP3TX,
354 .start = 11,
359 .start = 10,
H A Dgpio7xx.c37 .start = OMAP1_MPUIO_VBASE,
42 .start = INT_7XX_MPUIO,
78 .start = OMAP7XX_GPIO1_BASE,
83 .start = INT_7XX_GPIO_BANK1,
117 .start = OMAP7XX_GPIO2_BASE,
122 .start = INT_7XX_GPIO_BANK2,
145 .start = OMAP7XX_GPIO3_BASE,
150 .start = INT_7XX_GPIO_BANK3,
173 .start = OMAP7XX_GPIO4_BASE,
178 .start = INT_7XX_GPIO_BANK4,
201 .start = OMAP7XX_GPIO5_BASE,
206 .start = INT_7XX_GPIO_BANK5,
229 .start = OMAP7XX_GPIO6_BASE,
234 .start = INT_7XX_GPIO_BANK6,
/linux-4.4.14/drivers/usb/core/
H A Ddevices.c181 static char *usb_dump_endpoint_descriptor(int speed, char *start, char *end, usb_dump_endpoint_descriptor() argument
187 if (start > end) usb_dump_endpoint_descriptor()
188 return start; usb_dump_endpoint_descriptor()
230 return start; usb_dump_endpoint_descriptor()
241 start += sprintf(start, format_endpt, desc->bEndpointAddress, dir, usb_dump_endpoint_descriptor()
246 return start; usb_dump_endpoint_descriptor()
249 static char *usb_dump_interface_descriptor(char *start, char *end, usb_dump_interface_descriptor() argument
258 if (start > end) usb_dump_interface_descriptor()
259 return start; usb_dump_interface_descriptor()
267 start += sprintf(start, format_iface, usb_dump_interface_descriptor()
277 return start; usb_dump_interface_descriptor()
280 static char *usb_dump_interface(int speed, char *start, char *end, usb_dump_interface() argument
287 start = usb_dump_interface_descriptor(start, end, intfc, iface, setno); usb_dump_interface()
289 if (start > end) usb_dump_interface()
290 return start; usb_dump_interface()
291 start = usb_dump_endpoint_descriptor(speed, usb_dump_interface()
292 start, end, &desc->endpoint[i].desc); usb_dump_interface()
294 return start; usb_dump_interface()
297 static char *usb_dump_iad_descriptor(char *start, char *end, usb_dump_iad_descriptor() argument
300 if (start > end) usb_dump_iad_descriptor()
301 return start; usb_dump_iad_descriptor()
302 start += sprintf(start, format_iad, usb_dump_iad_descriptor()
309 return start; usb_dump_iad_descriptor()
317 static char *usb_dump_config_descriptor(char *start, char *end, usb_dump_config_descriptor() argument
323 if (start > end) usb_dump_config_descriptor()
324 return start; usb_dump_config_descriptor()
329 start += sprintf(start, format_config, usb_dump_config_descriptor()
336 return start; usb_dump_config_descriptor()
339 static char *usb_dump_config(int speed, char *start, char *end, usb_dump_config() argument
346 if (start > end) usb_dump_config()
347 return start; usb_dump_config()
350 return start + sprintf(start, "(null Cfg. desc.)\n"); usb_dump_config()
351 start = usb_dump_config_descriptor(start, end, &config->desc, active, usb_dump_config()
356 start = usb_dump_iad_descriptor(start, end, usb_dump_config()
363 if (start > end) usb_dump_config()
364 return start; usb_dump_config()
365 start = usb_dump_interface(speed, usb_dump_config()
366 start, end, intfc, interface, j); usb_dump_config()
369 return start; usb_dump_config()
375 static char *usb_dump_device_descriptor(char *start, char *end, usb_dump_device_descriptor() argument
381 if (start > end) usb_dump_device_descriptor()
382 return start; usb_dump_device_descriptor()
383 start += sprintf(start, format_device1, usb_dump_device_descriptor()
391 if (start > end) usb_dump_device_descriptor()
392 return start; usb_dump_device_descriptor()
393 start += sprintf(start, format_device2, usb_dump_device_descriptor()
397 return start; usb_dump_device_descriptor()
403 static char *usb_dump_device_strings(char *start, char *end, usb_dump_device_strings() argument
406 if (start > end) usb_dump_device_strings()
407 return start; usb_dump_device_strings()
409 start += sprintf(start, format_string_manufacturer, usb_dump_device_strings()
411 if (start > end) usb_dump_device_strings()
414 start += sprintf(start, format_string_product, dev->product); usb_dump_device_strings()
415 if (start > end) usb_dump_device_strings()
419 start += sprintf(start, format_string_serialnumber, usb_dump_device_strings()
423 return start; usb_dump_device_strings()
426 static char *usb_dump_desc(char *start, char *end, struct usb_device *dev) usb_dump_desc() argument
430 if (start > end) usb_dump_desc()
431 return start; usb_dump_desc()
433 start = usb_dump_device_descriptor(start, end, &dev->descriptor); usb_dump_desc()
435 if (start > end) usb_dump_desc()
436 return start; usb_dump_desc()
438 start = usb_dump_device_strings(start, end, dev); usb_dump_desc()
441 if (start > end) usb_dump_desc()
442 return start; usb_dump_desc()
443 start = usb_dump_config(dev->speed, usb_dump_desc()
444 start, end, dev->config + i, usb_dump_desc()
448 return start; usb_dump_desc()
454 static char *usb_dump_hub_descriptor(char *start, char *end, usb_dump_hub_descriptor() argument
460 if (start > end) usb_dump_hub_descriptor()
461 return start; usb_dump_hub_descriptor()
462 start += sprintf(start, "Interface:"); usb_dump_hub_descriptor()
463 while (leng && start <= end) { usb_dump_hub_descriptor()
464 start += sprintf(start, " %02x", *ptr); usb_dump_hub_descriptor()
467 *start++ = '\n'; usb_dump_hub_descriptor()
468 return start; usb_dump_hub_descriptor()
471 static char *usb_dump_string(char *start, char *end, usb_dump_string() argument
474 if (start > end) usb_dump_string()
475 return start; usb_dump_string()
476 start += sprintf(start, "Interface:"); usb_dump_string()
479 start += sprintf(start, "%s: %.100s ", id, usb_dump_string()
481 return start; usb_dump_string()
580 /* if we can start copying some data to the user */ usb_device_dump()
/linux-4.4.14/arch/arm64/mm/
H A Dcache.S30 * flush_icache_range(start,end)
36 * - start - virtual start address of region
43 * __flush_cache_user_range(start,end)
49 * - start - virtual start address of region
104 * __inval_cache_range(start, end)
105 * - start - start address of region
112 * __dma_inv_range(start, end)
113 * - start - virtual start address of region
123 1: tst x0, x3 // start cache line aligned?
138 * __dma_clean_range(start, end)
139 * - start - virtual start address of region
160 * __dma_flush_range(start, end)
161 * - start - virtual start address of region
177 * __dma_map_area(start, size, dir)
178 * - start - kernel virtual start address
190 * __dma_unmap_area(start, size, dir)
191 * - start - kernel virtual start address
H A Dpageattr.c42 unsigned long start = addr; change_memory_common() local
44 unsigned long end = start + size; change_memory_common()
49 start &= PAGE_MASK; change_memory_common()
50 end = start + size; change_memory_common()
54 if (start < MODULES_VADDR || start >= MODULES_END) change_memory_common()
66 ret = apply_to_page_range(&init_mm, start, size, change_page_range, change_memory_common()
69 flush_tlb_kernel_range(start, end); change_memory_common()
/linux-4.4.14/drivers/pcmcia/
H A Drsrc_iodyn.c34 resource_size_t start; pcmcia_align() local
36 start = (res->start & ~data->mask) + data->offset; pcmcia_align()
37 if (start < res->start) pcmcia_align()
38 start += data->mask + 1; pcmcia_align()
42 if (start & 0x300) pcmcia_align()
43 start = (start + 0x3ff) & ~0x3ff; pcmcia_align()
49 if ((res->start + size - 1) >= 1024) pcmcia_align()
50 start = res->end; pcmcia_align()
54 return start; pcmcia_align()
104 if ((s->io[i].res->start & (align-1)) == *base) iodyn_find_io()
125 *base = res->start; iodyn_find_io()
137 if (adjust_resource(s->io[i].res, res->start, iodyn_find_io()
147 try = res->start - num; iodyn_find_io()
150 res->start - num, iodyn_find_io()
/linux-4.4.14/arch/nios2/mm/
H A Dcacheflush.c18 static void __flush_dcache(unsigned long start, unsigned long end) __flush_dcache() argument
22 start &= ~(cpuinfo.dcache_line_size - 1); __flush_dcache()
26 if (end > start + cpuinfo.dcache_size) __flush_dcache()
27 end = start + cpuinfo.dcache_size; __flush_dcache()
29 for (addr = start; addr < end; addr += cpuinfo.dcache_line_size) { __flush_dcache()
37 static void __invalidate_dcache(unsigned long start, unsigned long end) __invalidate_dcache() argument
41 start &= ~(cpuinfo.dcache_line_size - 1); __invalidate_dcache()
45 for (addr = start; addr < end; addr += cpuinfo.dcache_line_size) { __invalidate_dcache()
53 static void __flush_icache(unsigned long start, unsigned long end) __flush_icache() argument
57 start &= ~(cpuinfo.icache_line_size - 1); __flush_icache()
61 if (end > start + cpuinfo.icache_size) __flush_icache()
62 end = start + cpuinfo.icache_size; __flush_icache()
64 for (addr = start; addr < end; addr += cpuinfo.icache_line_size) { __flush_icache()
113 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument
115 __flush_dcache(start, end); flush_icache_range()
116 __flush_icache(start, end); flush_icache_range()
119 void flush_dcache_range(unsigned long start, unsigned long end) flush_dcache_range() argument
121 __flush_dcache(start, end); flush_dcache_range()
122 __flush_icache(start, end); flush_dcache_range()
126 void invalidate_dcache_range(unsigned long start, unsigned long end) invalidate_dcache_range() argument
128 __invalidate_dcache(start, end); invalidate_dcache_range()
132 void flush_cache_range(struct vm_area_struct *vma, unsigned long start, flush_cache_range() argument
135 __flush_dcache(start, end); flush_cache_range()
137 __flush_icache(start, end); flush_cache_range()
142 unsigned long start = (unsigned long) page_address(page); flush_icache_page() local
143 unsigned long end = start + PAGE_SIZE; flush_icache_page()
145 __flush_dcache(start, end); flush_icache_page()
146 __flush_icache(start, end); flush_icache_page()
152 unsigned long start = vmaddr; flush_cache_page() local
153 unsigned long end = start + PAGE_SIZE; flush_cache_page()
155 __flush_dcache(start, end); flush_cache_page()
157 __flush_icache(start, end); flush_cache_page()
167 unsigned long start = (unsigned long)page_address(page); __flush_dcache_page() local
169 __flush_dcache(start, start + PAGE_SIZE); __flush_dcache_page()
191 unsigned long start = (unsigned long)page_address(page); flush_dcache_page() local
193 flush_icache_range(start, start + PAGE_SIZE); flush_dcache_page()
/linux-4.4.14/drivers/net/wireless/ti/wlcore/
H A Dio.c83 if ((addr >= part->mem.start) && wlcore_translate_addr()
84 (addr < part->mem.start + part->mem.size)) wlcore_translate_addr()
85 return addr - part->mem.start; wlcore_translate_addr()
86 else if ((addr >= part->reg.start) && wlcore_translate_addr()
87 (addr < part->reg.start + part->reg.size)) wlcore_translate_addr()
88 return addr - part->reg.start + part->mem.size; wlcore_translate_addr()
89 else if ((addr >= part->mem2.start) && wlcore_translate_addr()
90 (addr < part->mem2.start + part->mem2.size)) wlcore_translate_addr()
91 return addr - part->mem2.start + part->mem.size + wlcore_translate_addr()
93 else if ((addr >= part->mem3.start) && wlcore_translate_addr()
94 (addr < part->mem3.start + part->mem3.size)) wlcore_translate_addr()
95 return addr - part->mem3.start + part->mem.size + wlcore_translate_addr()
117 * ...+----+--> mem.start
121 * 00000000 <--+----+... ...+----+--> mem.start + mem.size
129 * + <--+----+... ...+----+--> reg.start
133 * ...+----+--> reg.start + reg.size
146 p->mem.start, p->mem.size); wlcore_set_partition()
148 p->reg.start, p->reg.size); wlcore_set_partition()
150 p->mem2.start, p->mem2.size); wlcore_set_partition()
152 p->mem3.start, p->mem3.size); wlcore_set_partition()
154 ret = wlcore_raw_write32(wl, HW_PART0_START_ADDR, p->mem.start); wlcore_set_partition()
162 ret = wlcore_raw_write32(wl, HW_PART1_START_ADDR, p->reg.start); wlcore_set_partition()
170 ret = wlcore_raw_write32(wl, HW_PART2_START_ADDR, p->mem2.start); wlcore_set_partition()
183 ret = wlcore_raw_write32(wl, HW_PART3_START_ADDR, p->mem3.start); wlcore_set_partition()
/linux-4.4.14/arch/powerpc/include/asm/
H A Dsections.h25 static inline int overlaps_interrupt_vector_text(unsigned long start, overlaps_interrupt_vector_text() argument
32 return start < (unsigned long)__va(real_end) && overlaps_interrupt_vector_text()
36 static inline int overlaps_kernel_text(unsigned long start, unsigned long end) overlaps_kernel_text() argument
38 return start < (unsigned long)__init_end && overlaps_kernel_text()
42 static inline int overlaps_kvm_tmp(unsigned long start, unsigned long end) overlaps_kvm_tmp() argument
46 return start < (unsigned long)kvm_tmp && overlaps_kvm_tmp()
H A Dcacheflush.h22 #define flush_cache_range(vma, start, end) do { } while (0)
25 #define flush_cache_vmap(start, end) do { } while (0)
26 #define flush_cache_vunmap(start, end) do { } while (0)
50 extern void flush_dcache_range(unsigned long start, unsigned long stop);
52 extern void clean_dcache_range(unsigned long start, unsigned long stop);
53 extern void invalidate_dcache_range(unsigned long start, unsigned long stop);
56 extern void flush_inval_dcache_range(unsigned long start, unsigned long stop);
57 extern void flush_dcache_phys_range(unsigned long start, unsigned long stop);
H A Drheap.h21 unsigned long start; member in struct:_rh_block
41 unsigned long start; member in struct:_rh_stats
60 extern int rh_attach_region(rh_info_t * info, unsigned long start, int size);
63 extern unsigned long rh_detach_region(rh_info_t * info, unsigned long start, int size);
73 extern unsigned long rh_alloc_fixed(rh_info_t * info, unsigned long start, int size,
77 extern int rh_free(rh_info_t * info, unsigned long start);
87 extern int rh_set_owner(rh_info_t * info, unsigned long start, const char *owner);
/linux-4.4.14/arch/mips/jz4740/
H A Dplatform.c38 .start = JZ4740_UHC_BASE_ADDR,
43 .start = JZ4740_IRQ_UHC,
68 .start = JZ4740_UDC_BASE_ADDR,
73 .start = JZ4740_IRQ_UDC,
94 .start = JZ4740_MSC_BASE_ADDR,
99 .start = JZ4740_IRQ_MSC,
119 .start = JZ4740_RTC_BASE_ADDR,
124 .start = JZ4740_IRQ_RTC,
140 .start = JZ4740_I2C_BASE_ADDR,
145 .start = JZ4740_IRQ_I2C,
162 .start = JZ4740_EMC_BASE_ADDR,
168 .start = 0x18000000,
174 .start = 0x14000000,
180 .start = 0x0C000000,
186 .start = 0x08000000,
201 .start = JZ4740_LCD_BASE_ADDR,
221 .start = JZ4740_AIC_BASE_ADDR,
243 .start = JZ4740_AIC_BASE_ADDR + 0x80,
259 .start = JZ4740_SADC_BASE_ADDR,
264 .start = JZ4740_IRQ_SADC,
269 .start = JZ4740_IRQ_ADC_BASE,
285 .start = JZ4740_WDT_BASE_ADDR,
307 .start = JZ4740_DMAC_BASE_ADDR,
312 .start = JZ4740_IRQ_DMAC,
/linux-4.4.14/arch/arm/mach-pxa/
H A Ddevices.c38 .start = IRQ_PMU,
52 .start = 0x41100000,
57 .start = IRQ_MMC,
62 .start = 21,
67 .start = 22,
103 .start = 0x40600000,
108 .start = IRQ_USB,
141 .start = 0x54100000,
146 .start = IRQ_USB2,
167 .start = 0x44000000,
172 .start = IRQ_LCD,
199 .start = 0x40100000,
203 .start = IRQ_FFUART,
223 .start = 0x40200000,
227 .start = IRQ_BTUART,
247 .start = 0x40700000,
251 .start = IRQ_STUART,
271 .start = 0x41600000,
275 .start = IRQ_HWUART,
298 .start = 0x40301680,
302 .start = IRQ_I2C,
323 .start = 0x40f00180,
327 .start = IRQ_PWRI2C,
343 .start = 0x40400000,
347 .start = IRQ_I2S,
389 .start = IRQ_STUART,
394 .start = IRQ_ICP,
399 .start = 0x40800000,
404 .start = 0x40700000,
409 .start = 17,
414 .start = 18,
438 .start = 0x40900000,
443 .start = IRQ_RTC1Hz,
449 .start = IRQ_RTCAlrm,
472 .start = 0x40500000,
477 .start = IRQ_AC97,
505 .start = 0x40b00000,
520 .start = 0x40c00000,
537 .start = 0x41000000,
542 .start = IRQ_SSP,
548 .start = 13,
554 .start = 14,
575 .start = 0x41400000,
580 .start = IRQ_NSSP,
586 .start = 15,
592 .start = 16,
613 .start = 0x41500000,
618 .start = IRQ_ASSP,
624 .start = 23,
630 .start = 24,
652 .start = 0x50000000,
657 .start = IRQ_CAMERA,
685 .start = 0x4C000000,
690 .start = IRQ_USBH1,
716 .start = 0x41500000,
721 .start = IRQ_KEYPAD,
743 .start = 0x41000000,
748 .start = IRQ_SSP,
754 .start = 13,
760 .start = 14,
781 .start = 0x41700000,
786 .start = IRQ_SSP2,
792 .start = 15,
798 .start = 16,
819 .start = 0x41900000,
824 .start = IRQ_SSP3,
830 .start = 66,
836 .start = 67,
855 .start = 0x40b00000,
870 .start = 0x40c00000,
887 .start = 0x42000000,
892 .start = IRQ_MMC2,
897 .start = 93,
902 .start = 94,
926 .start = 0x42500000,
931 .start = IRQ_MMC3,
936 .start = 100,
941 .start = 101,
965 .start = 0x54000000,
970 .start = IRQ_GCU,
994 .start = 0x40f500c0,
998 .start = IRQ_PWRI2C,
1013 .start = 0x43100000,
1018 .start = IRQ_NAND,
1024 .start = 97,
1030 .start = 99,
1058 .start = 0x41a00000,
1063 .start = IRQ_SSP4,
1069 .start = 2,
1075 .start = 3,
1134 .start = 0x40e00000,
1138 .start = IRQ_GPIO0,
1143 .start = IRQ_GPIO1,
1148 .start = IRQ_GPIO_2_x,
1210 .start = 0x40000000,
1215 .start = IRQ_DMA,
/linux-4.4.14/arch/alpha/mm/
H A Dnuma.c48 printk(" memcluster %2d, usage %1lx, start %8lu, end %8lu\n", for_each_mem_cluster()
62 unsigned long start, end; setup_memory_node() local
87 start = cluster->start_pfn; for_each_mem_cluster()
88 end = start + cluster->numpages; for_each_mem_cluster()
90 if (start >= node_pfn_end || end <= node_pfn_start) for_each_mem_cluster()
97 printk(" memcluster %2d, usage %1lx, start %8lu, end %8lu\n", for_each_mem_cluster()
101 if (start < node_pfn_start) for_each_mem_cluster()
102 start = node_pfn_start; for_each_mem_cluster()
106 if (start < node_min_pfn) for_each_mem_cluster()
107 node_min_pfn = start; for_each_mem_cluster()
144 printk(" Detected node memory: start %8lu, end %8lu\n",
158 /* Zone start phys-addr must be 2^(MAX_ORDER-1) aligned.
172 start = cluster->start_pfn; for_each_mem_cluster()
173 end = start + cluster->numpages; for_each_mem_cluster()
175 if (start >= node_max_pfn || end <= node_min_pfn) for_each_mem_cluster()
180 if (start < node_min_pfn) for_each_mem_cluster()
181 start = node_min_pfn; for_each_mem_cluster()
183 if (start < start_kernel_pfn) { for_each_mem_cluster()
190 } else if (start < end_kernel_pfn) for_each_mem_cluster()
191 start = end_kernel_pfn; for_each_mem_cluster()
192 if (end - start >= bootmap_pages) { for_each_mem_cluster()
193 bootmap_start = start; for_each_mem_cluster()
212 start = cluster->start_pfn; for_each_mem_cluster()
215 if (start >= node_max_pfn || end <= node_min_pfn) for_each_mem_cluster()
220 if (start < node_min_pfn) for_each_mem_cluster()
221 start = node_min_pfn; for_each_mem_cluster()
223 if (start < start_kernel_pfn) { for_each_mem_cluster()
225 free_bootmem_node(NODE_DATA(nid), PFN_PHYS(start), for_each_mem_cluster()
227 - PFN_PHYS(start))); for_each_mem_cluster()
229 start, start_kernel_pfn); for_each_mem_cluster()
230 start = end_kernel_pfn; for_each_mem_cluster()
233 } else if (start < end_kernel_pfn) for_each_mem_cluster()
234 start = end_kernel_pfn; for_each_mem_cluster()
235 if (start >= end) for_each_mem_cluster()
238 free_bootmem_node(NODE_DATA(nid), PFN_PHYS(start), PFN_PHYS(end) - PFN_PHYS(start)); for_each_mem_cluster()
239 printk(" freeing pages %ld:%ld\n", start, end); for_each_mem_cluster()
/linux-4.4.14/fs/btrfs/tests/
H A Dextent-io-tests.c28 static noinline int process_page_range(struct inode *inode, u64 start, u64 end, process_page_range() argument
33 unsigned long index = start >> PAGE_CACHE_SHIFT; process_page_range()
59 printk(KERN_ERR "stuck in a loop, start %Lu, end %Lu, nr_pages %lu, ret %d\n", start, end, nr_pages, ret); process_page_range()
75 u64 start, end, test_start; test_find_delalloc() local
113 start = 0; test_find_delalloc()
115 found = find_lock_delalloc_range(inode, &tmp, locked_page, &start, test_find_delalloc()
121 if (start != 0 || end != 4095) { test_find_delalloc()
122 test_msg("Expected start 0 end 4095, got start %Lu end %Lu\n", test_find_delalloc()
123 start, end); test_find_delalloc()
126 unlock_extent(&tmp, start, end); test_find_delalloc()
144 start = test_start; test_find_delalloc()
146 found = find_lock_delalloc_range(inode, &tmp, locked_page, &start, test_find_delalloc()
152 if (start != test_start || end != max_bytes - 1) { test_find_delalloc()
153 test_msg("Expected start %Lu end %Lu, got start %Lu, end " test_find_delalloc()
154 "%Lu\n", test_start, max_bytes - 1, start, end); test_find_delalloc()
157 if (process_page_range(inode, start, end, test_find_delalloc()
162 unlock_extent(&tmp, start, end); test_find_delalloc()
178 start = test_start; test_find_delalloc()
180 found = find_lock_delalloc_range(inode, &tmp, locked_page, &start, test_find_delalloc()
199 start = test_start; test_find_delalloc()
201 found = find_lock_delalloc_range(inode, &tmp, locked_page, &start, test_find_delalloc()
207 if (start != test_start || end != total_dirty - 1) { test_find_delalloc()
208 test_msg("Expected start %Lu end %Lu, got start %Lu end %Lu\n", test_find_delalloc()
209 test_start, total_dirty - 1, start, end); test_find_delalloc()
212 if (process_page_range(inode, start, end, test_find_delalloc()
217 unlock_extent(&tmp, start, end); test_find_delalloc()
234 start = test_start; test_find_delalloc()
242 found = find_lock_delalloc_range(inode, &tmp, locked_page, &start, test_find_delalloc()
248 if (start != test_start && end != test_start + PAGE_CACHE_SIZE - 1) { test_find_delalloc()
249 test_msg("Expected start %Lu end %Lu, got start %Lu end %Lu\n", test_find_delalloc()
250 test_start, test_start + PAGE_CACHE_SIZE - 1, start, test_find_delalloc()
254 if (process_page_range(inode, start, end, PROCESS_TEST_LOCKED | test_find_delalloc()
H A Dinode-tests.c26 static void insert_extent(struct btrfs_root *root, u64 start, u64 len, insert_extent() argument
45 key.offset = start; insert_extent()
308 if (em->start != 0 || em->len != 5) { test_btrfs_get_extent()
309 test_msg("Unexpected extent wanted start 0 len 5, got start " test_btrfs_get_extent()
310 "%llu len %llu\n", em->start, em->len); test_btrfs_get_extent()
317 offset = em->start + em->len; test_btrfs_get_extent()
329 if (em->start != offset || em->len != 4091) { test_btrfs_get_extent()
330 test_msg("Unexpected extent wanted start %llu len 1, got start " test_btrfs_get_extent()
331 "%llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent()
343 offset = em->start + em->len; test_btrfs_get_extent()
355 if (em->start != offset || em->len != 4) { test_btrfs_get_extent()
356 test_msg("Unexpected extent wanted start %llu len 4, got start " test_btrfs_get_extent()
357 "%llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent()
364 offset = em->start + em->len; test_btrfs_get_extent()
377 if (em->start != offset || em->len != 4095) { test_btrfs_get_extent()
378 test_msg("Unexpected extent wanted start %llu len 4095, got " test_btrfs_get_extent()
379 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent()
386 if (em->orig_start != em->start) { test_btrfs_get_extent()
387 test_msg("Wrong orig offset, want %llu, have %llu\n", em->start, test_btrfs_get_extent()
391 offset = em->start + em->len; test_btrfs_get_extent()
404 if (em->start != offset || em->len != 4096) { test_btrfs_get_extent()
405 test_msg("Unexpected extent wanted start %llu len 4096, got " test_btrfs_get_extent()
406 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent()
413 if (em->orig_start != em->start) { test_btrfs_get_extent()
414 test_msg("Wrong orig offset, want %llu, have %llu\n", em->start, test_btrfs_get_extent()
419 orig_start = em->start; test_btrfs_get_extent()
420 offset = em->start + em->len; test_btrfs_get_extent()
432 if (em->start != offset || em->len != 4096) { test_btrfs_get_extent()
433 test_msg("Unexpected extent wanted start %llu len 4096, got " test_btrfs_get_extent()
434 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent()
441 offset = em->start + em->len; test_btrfs_get_extent()
453 if (em->start != offset || em->len != 8192) { test_btrfs_get_extent()
454 test_msg("Unexpected extent wanted start %llu len 8192, got " test_btrfs_get_extent()
455 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent()
467 disk_bytenr += (em->start - orig_start); test_btrfs_get_extent()
469 test_msg("Wrong block start, want %llu, have %llu\n", test_btrfs_get_extent()
473 offset = em->start + em->len; test_btrfs_get_extent()
486 if (em->start != offset || em->len != 4096) { test_btrfs_get_extent()
487 test_msg("Unexpected extent wanted start %llu len 4096, got " test_btrfs_get_extent()
488 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent()
496 if (em->orig_start != em->start) { test_btrfs_get_extent()
497 test_msg("Wrong orig offset, want %llu, have %llu\n", em->start, test_btrfs_get_extent()
501 offset = em->start + em->len; test_btrfs_get_extent()
514 if (em->start != offset || em->len != 4096) { test_btrfs_get_extent()
515 test_msg("Unexpected extent wanted start %llu len 4096, got " test_btrfs_get_extent()
516 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent()
524 if (em->orig_start != em->start) { test_btrfs_get_extent()
525 test_msg("Wrong orig offset, want %llu, have %llu\n", em->start, test_btrfs_get_extent()
530 orig_start = em->start; test_btrfs_get_extent()
531 offset = em->start + em->len; test_btrfs_get_extent()
543 if (em->start != offset || em->len != 4096) { test_btrfs_get_extent()
544 test_msg("Unexpected extent wanted start %llu len 4096, got " test_btrfs_get_extent()
545 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent()
557 if (em->block_start != (disk_bytenr + (em->start - em->orig_start))) { test_btrfs_get_extent()
558 test_msg("Unexpected block start, wanted %llu, have %llu\n", test_btrfs_get_extent()
559 disk_bytenr + (em->start - em->orig_start), test_btrfs_get_extent()
563 offset = em->start + em->len; test_btrfs_get_extent()
575 if (em->start != offset || em->len != 8192) { test_btrfs_get_extent()
576 test_msg("Unexpected extent wanted start %llu len 8192, got " test_btrfs_get_extent()
577 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent()
590 if (em->block_start != (disk_bytenr + (em->start - em->orig_start))) { test_btrfs_get_extent()
591 test_msg("Unexpected block start, wanted %llu, have %llu\n", test_btrfs_get_extent()
592 disk_bytenr + (em->start - em->orig_start), test_btrfs_get_extent()
596 offset = em->start + em->len; test_btrfs_get_extent()
609 if (em->start != offset || em->len != 8192) { test_btrfs_get_extent()
610 test_msg("Unexpected extent wanted start %llu len 8192, got " test_btrfs_get_extent()
611 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent()
619 if (em->orig_start != em->start) { test_btrfs_get_extent()
621 em->start, em->orig_start); test_btrfs_get_extent()
629 offset = em->start + em->len; test_btrfs_get_extent()
642 if (em->start != offset || em->len != 4096) { test_btrfs_get_extent()
643 test_msg("Unexpected extent wanted start %llu len 4096, got " test_btrfs_get_extent()
644 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent()
652 if (em->orig_start != em->start) { test_btrfs_get_extent()
654 em->start, em->orig_start); test_btrfs_get_extent()
663 orig_start = em->start; test_btrfs_get_extent()
664 offset = em->start + em->len; test_btrfs_get_extent()
676 if (em->start != offset || em->len != 4096) { test_btrfs_get_extent()
677 test_msg("Unexpected extent wanted start %llu len 4096, got " test_btrfs_get_extent()
678 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent()
685 if (em->orig_start != em->start) { test_btrfs_get_extent()
686 test_msg("Wrong orig offset, want %llu, have %llu\n", em->start, test_btrfs_get_extent()
690 offset = em->start + em->len; test_btrfs_get_extent()
699 test_msg("Block start does not match, want %llu got %llu\n", test_btrfs_get_extent()
703 if (em->start != offset || em->len != 8192) { test_btrfs_get_extent()
704 test_msg("Unexpected extent wanted start %llu len 8192, got " test_btrfs_get_extent()
705 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent()
715 em->start, orig_start); test_btrfs_get_extent()
723 offset = em->start + em->len; test_btrfs_get_extent()
736 if (em->start != offset || em->len != 4096) { test_btrfs_get_extent()
737 test_msg("Unexpected extent wanted start %llu len 4096, got " test_btrfs_get_extent()
738 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent()
745 if (em->orig_start != em->start) { test_btrfs_get_extent()
746 test_msg("Wrong orig offset, want %llu, have %llu\n", em->start, test_btrfs_get_extent()
750 offset = em->start + em->len; test_btrfs_get_extent()
767 if (em->start != offset || em->len != 12288) { test_btrfs_get_extent()
768 test_msg("Unexpected extent wanted start %llu len 12288, got " test_btrfs_get_extent()
769 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent()
777 if (em->orig_start != em->start) { test_btrfs_get_extent()
778 test_msg("Wrong orig offset, want %llu, have %llu\n", em->start, test_btrfs_get_extent()
782 offset = em->start + em->len; test_btrfs_get_extent()
794 if (em->start != offset || em->len != 4096) { test_btrfs_get_extent()
795 test_msg("Unexpected extent wanted start %llu len 4096, got " test_btrfs_get_extent()
796 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent()
803 if (em->orig_start != em->start) { test_btrfs_get_extent()
804 test_msg("Wrong orig offset, want %llu, have %llu\n", em->start, test_btrfs_get_extent()
874 if (em->start != 0 || em->len != 4096) { test_hole_first()
875 test_msg("Unexpected extent wanted start 0 len 4096, got start " test_hole_first()
876 "%llu len %llu\n", em->start, em->len); test_hole_first()
895 if (em->start != 4096 || em->len != 4096) { test_hole_first()
896 test_msg("Unexpected extent wanted start 4096 len 4096, got " test_hole_first()
897 "start %llu len %llu\n", em->start, em->len); test_hole_first()
/linux-4.4.14/drivers/pci/
H A Dhotplug-pci.c10 int pass, busnr, start = parent->busn_res.start; pci_hp_add_bridge() local
13 for (busnr = start; busnr <= end; busnr++) { pci_hp_add_bridge()
H A Dhost-bridge.c62 region->start = res->start - offset; pcibios_resource_to_bus()
70 return region1->start <= region2->start && region1->end >= region2->end; region_contains()
86 bus_region.start = window->res->start - window->offset; pcibios_bus_to_resource()
95 res->start = region->start + offset; pcibios_bus_to_resource()
/linux-4.4.14/arch/x86/mm/
H A Dkasan_init_64.c17 unsigned long start; map_range() local
20 start = (unsigned long)kasan_mem_to_shadow(pfn_to_kaddr(range->start)); map_range()
28 return vmemmap_populate(start, end + 1, NUMA_NO_NODE); map_range()
31 static void __init clear_pgds(unsigned long start, clear_pgds() argument
34 for (; start < end; start += PGDIR_SIZE) clear_pgds()
35 pgd_clear(pgd_offset_k(start)); clear_pgds()
41 unsigned long start = KASAN_SHADOW_START; kasan_map_early_shadow() local
44 for (i = pgd_index(start); start < end; i++) { kasan_map_early_shadow()
47 start += PGDIR_SIZE; kasan_map_early_shadow()
H A Dpat_rbtree.c39 static int is_node_overlap(struct memtype *node, u64 start, u64 end) is_node_overlap() argument
41 if (node->start >= end || node->end <= start) is_node_overlap()
75 /* Find the first (lowest start addr) overlapping range from rb tree */ memtype_rb_lowest_match()
77 u64 start, u64 end) memtype_rb_lowest_match()
85 if (get_subtree_max_end(node->rb_left) > start) { memtype_rb_lowest_match()
88 } else if (is_node_overlap(data, start, end)) { memtype_rb_lowest_match()
91 } else if (start >= data->start) { memtype_rb_lowest_match()
102 u64 start, u64 end) memtype_rb_exact_match()
106 match = memtype_rb_lowest_match(root, start, end); memtype_rb_exact_match()
107 while (match != NULL && match->start < end) { memtype_rb_exact_match()
110 if (match->start == start && match->end == end) memtype_rb_exact_match()
124 u64 start, u64 end, memtype_rb_check_conflict()
132 match = memtype_rb_lowest_match(&memtype_rbroot, start, end); memtype_rb_check_conflict()
139 dprintk("Overlap at 0x%Lx-0x%Lx\n", match->start, match->end); memtype_rb_check_conflict()
146 if (match->start >= end) /* Checked all possible matches */ memtype_rb_check_conflict()
149 if (is_node_overlap(match, start, end) && memtype_rb_check_conflict()
164 current->comm, current->pid, start, end, memtype_rb_check_conflict()
180 if (newdata->start <= data->start) memtype_rb_insert()
182 else if (newdata->start > data->start) memtype_rb_insert()
196 err = memtype_rb_check_conflict(&memtype_rbroot, new->start, new->end, rbt_memtype_check_insert()
209 struct memtype *rbt_memtype_erase(u64 start, u64 end) rbt_memtype_erase() argument
213 data = memtype_rb_exact_match(&memtype_rbroot, start, end); rbt_memtype_erase()
76 memtype_rb_lowest_match(struct rb_root *root, u64 start, u64 end) memtype_rb_lowest_match() argument
101 memtype_rb_exact_match(struct rb_root *root, u64 start, u64 end) memtype_rb_exact_match() argument
123 memtype_rb_check_conflict(struct rb_root *root, u64 start, u64 end, enum page_cache_mode reqtype, enum page_cache_mode *newtype) memtype_rb_check_conflict() argument
H A Ddump_pagetables.c254 * such as the start of vmalloc space etc. note_page()
283 pte_t *start; walk_pte_level() local
286 start = (pte_t *) pmd_page_vaddr(addr); walk_pte_level()
288 prot = pte_flags(*start); walk_pte_level()
291 start++; walk_pte_level()
301 pmd_t *start; walk_pmd_level() local
304 start = (pmd_t *) pud_page_vaddr(addr); walk_pmd_level()
307 if (!pmd_none(*start)) { walk_pmd_level()
308 if (pmd_large(*start) || !pmd_present(*start)) { walk_pmd_level()
309 prot = pmd_flags(*start); walk_pmd_level()
312 walk_pte_level(m, st, *start, walk_pmd_level()
317 start++; walk_pmd_level()
333 pud_t *start; walk_pud_level() local
336 start = (pud_t *) pgd_page_vaddr(addr); walk_pud_level()
340 if (!pud_none(*start)) { walk_pud_level()
341 if (pud_large(*start) || !pud_present(*start)) { walk_pud_level()
342 prot = pud_flags(*start); walk_pud_level()
345 walk_pmd_level(m, st, *start, walk_pud_level()
351 start++; walk_pud_level()
380 pgd_t *start = (pgd_t *) &init_level4_pgt; ptdump_walk_pgd_level_core() local
382 pgd_t *start = swapper_pg_dir; ptdump_walk_pgd_level_core()
389 start = pgd; ptdump_walk_pgd_level_core()
399 if (!pgd_none(*start) && !is_hypervisor_range(i)) { ptdump_walk_pgd_level_core()
400 if (pgd_large(*start) || !pgd_present(*start)) { ptdump_walk_pgd_level_core()
401 prot = pgd_flags(*start); ptdump_walk_pgd_level_core()
404 walk_pud_level(m, &st, *start, ptdump_walk_pgd_level_core()
410 start++; ptdump_walk_pgd_level_core()
/linux-4.4.14/arch/blackfin/mach-bf561/boards/
H A Dtepla.c24 .start = 0x2C000300,
28 .start = IRQ_PROG_INTB,
32 .start = IRQ_PF7,
49 .start = BFIN_UART_THR,
54 .start = IRQ_UART_TX,
59 .start = IRQ_UART_RX,
64 .start = IRQ_UART_ERROR,
69 .start = CH_UART_TX,
74 .start = CH_UART_RX,
100 .start = 0xFFC00400,
105 .start = IRQ_UART0_RX,
110 .start = CH_UART0_RX,
H A Dcm_bf561.c103 .start = SPI0_REGBASE,
108 .start = CH_SPI,
113 .start = IRQ_SPI,
157 .start = 0x28000300,
161 .start = IRQ_PF0,
183 .start = 0x24008000,
188 .start = IRQ_PF43,
215 .start = 0x24000000,
219 .start = IRQ_PF45,
236 .start = 0x24008000,
240 .start = 0x24008004,
244 .start = IRQ_PF47,
276 .start = BFIN_UART_THR,
281 .start = IRQ_UART_TX,
286 .start = IRQ_UART_RX,
291 .start = IRQ_UART_ERROR,
296 .start = CH_UART_TX,
301 .start = CH_UART_RX,
327 .start = 0xFFC00400,
332 .start = IRQ_UART0_RX,
337 .start = CH_UART0_RX,
361 .start = 0x2400C000,
366 .start = 0x2400D018,
371 .start = PATA_INT,
412 .start = 0x20000000,
/linux-4.4.14/arch/c6x/include/asm/
H A Dcache.h66 extern void enable_caching(unsigned long start, unsigned long end);
67 extern void disable_caching(unsigned long start, unsigned long end);
80 extern void L1P_cache_block_invalidate(unsigned int start, unsigned int end);
81 extern void L1D_cache_block_invalidate(unsigned int start, unsigned int end);
82 extern void L1D_cache_block_writeback_invalidate(unsigned int start,
84 extern void L1D_cache_block_writeback(unsigned int start, unsigned int end);
85 extern void L2_cache_block_invalidate(unsigned int start, unsigned int end);
86 extern void L2_cache_block_writeback(unsigned int start, unsigned int end);
87 extern void L2_cache_block_writeback_invalidate(unsigned int start,
89 extern void L2_cache_block_invalidate_nowait(unsigned int start,
91 extern void L2_cache_block_writeback_nowait(unsigned int start,
94 extern void L2_cache_block_writeback_invalidate_nowait(unsigned int start,
/linux-4.4.14/arch/mips/sgi-ip27/
H A Dip27-klconfig.c50 lboard_t *find_lboard(lboard_t *start, unsigned char brd_type) find_lboard() argument
53 while (start) { find_lboard()
54 if (start->brd_type == brd_type) find_lboard()
55 return start; find_lboard()
56 start = KLCF_NEXT(start); find_lboard()
62 lboard_t *find_lboard_class(lboard_t *start, unsigned char brd_type) find_lboard_class() argument
65 while (start) { find_lboard_class()
66 if (KLCLASS(start->brd_type) == KLCLASS(brd_type)) find_lboard_class()
67 return start; find_lboard_class()
68 start = KLCF_NEXT(start); find_lboard_class()
/linux-4.4.14/arch/mips/include/asm/mach-rc32434/
H A Dirq.h17 #define GROUP0_IRQ_BASE 8 /* GRP2 IRQ numbers start here */
18 /* GRP3 IRQ numbers start here */
20 /* GRP4 IRQ numbers start here */
22 /* GRP5 IRQ numbers start here */
/linux-4.4.14/arch/sh/mm/
H A Dcache-sh2.c19 static void sh2__flush_wback_region(void *start, int size) sh2__flush_wback_region() argument
24 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); sh2__flush_wback_region()
25 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2__flush_wback_region()
40 static void sh2__flush_purge_region(void *start, int size) sh2__flush_purge_region() argument
45 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); sh2__flush_purge_region()
46 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2__flush_purge_region()
54 static void sh2__flush_invalidate_region(void *start, int size) sh2__flush_invalidate_region() argument
76 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); sh2__flush_invalidate_region()
77 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2__flush_invalidate_region()
H A Dnuma.c27 void __init setup_bootmem_node(int nid, unsigned long start, unsigned long end) setup_bootmem_node() argument
36 start_pfn = PFN_DOWN(start); setup_bootmem_node()
39 pmb_bolt_mapping((unsigned long)__va(start), start, end - start, setup_bootmem_node()
42 memblock_add(start, end - start); setup_bootmem_node()
H A Dtlbflush_32.c39 void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, local_flush_tlb_range() argument
50 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_range()
60 start &= PAGE_MASK; local_flush_tlb_range()
67 while (start < end) { local_flush_tlb_range()
68 local_flush_tlb_one(asid, start); local_flush_tlb_range()
69 start += PAGE_SIZE; local_flush_tlb_range()
78 void local_flush_tlb_kernel_range(unsigned long start, unsigned long end) local_flush_tlb_kernel_range() argument
85 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_kernel_range()
93 start &= PAGE_MASK; local_flush_tlb_kernel_range()
97 while (start < end) { local_flush_tlb_kernel_range()
98 local_flush_tlb_one(asid, start); local_flush_tlb_kernel_range()
99 start += PAGE_SIZE; local_flush_tlb_kernel_range()
H A Dflush-sh4.c13 static void sh4__flush_wback_region(void *start, int size) sh4__flush_wback_region() argument
17 aligned_start = register_align(start); sh4__flush_wback_region()
47 static void sh4__flush_purge_region(void *start, int size) sh4__flush_purge_region() argument
51 aligned_start = register_align(start); sh4__flush_purge_region()
77 static void sh4__flush_invalidate_region(void *start, int size) sh4__flush_invalidate_region() argument
81 aligned_start = register_align(start); sh4__flush_invalidate_region()
H A Dcache-sh2a.c50 static void sh2a__flush_wback_region(void *start, int size) sh2a__flush_wback_region() argument
58 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); sh2a__flush_wback_region()
59 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2a__flush_wback_region()
92 static void sh2a__flush_purge_region(void *start, int size) sh2a__flush_purge_region() argument
98 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); sh2a__flush_purge_region()
99 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2a__flush_purge_region()
122 static void sh2a__flush_invalidate_region(void *start, int size) sh2a__flush_invalidate_region() argument
128 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); sh2a__flush_invalidate_region()
129 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2a__flush_invalidate_region()
154 unsigned long start, end; sh2a_flush_icache_range() local
158 start = data->addr1 & ~(L1_CACHE_BYTES-1); sh2a_flush_icache_range()
162 sh2a__flush_wback_region((void *)start, end-start); sh2a_flush_icache_range()
170 if (((end - start) >> PAGE_SHIFT) >= MAX_ICACHE_PAGES) { sh2a_flush_icache_range()
174 for (v = start; v < end; v += L1_CACHE_BYTES) sh2a_flush_icache_range()
/linux-4.4.14/arch/avr32/mm/
H A Dcache.c24 void invalidate_dcache_region(void *start, size_t size) invalidate_dcache_region() argument
34 begin = (unsigned long)start; invalidate_dcache_region()
38 flush_dcache_line(start); invalidate_dcache_region()
52 void clean_dcache_region(void *start, size_t size) clean_dcache_region() argument
57 begin = (unsigned long)start & ~(linesz - 1); clean_dcache_region()
58 end = ((unsigned long)start + size + linesz - 1) & ~(linesz - 1); clean_dcache_region()
65 void flush_dcache_region(void *start, size_t size) flush_dcache_region() argument
70 begin = (unsigned long)start & ~(linesz - 1); flush_dcache_region()
71 end = ((unsigned long)start + size + linesz - 1) & ~(linesz - 1); flush_dcache_region()
78 void invalidate_icache_region(void *start, size_t size) invalidate_icache_region() argument
83 begin = (unsigned long)start & ~(linesz - 1); invalidate_icache_region()
84 end = ((unsigned long)start + size + linesz - 1) & ~(linesz - 1); invalidate_icache_region()
90 static inline void __flush_icache_range(unsigned long start, unsigned long end) __flush_icache_range() argument
95 for (v = start; v < end; v += linesz) { __flush_icache_range()
106 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument
111 __flush_icache_range(start & ~(linesz - 1), flush_icache_range()
/linux-4.4.14/arch/blackfin/mach-bf537/boards/
H A Dminotaur.c37 .start = 0x20310000, /* IO PORT */
41 .start = 0x20311000, /* Attribute Memory */
45 .start = IRQ_PF4,
49 .start = IRQ_PF6, /* Card Detect PF6 */
106 .start = 0x20300000,
110 .start = IRQ_PF7,
205 .start = SPI0_REGBASE,
210 .start = CH_SPI,
215 .start = IRQ_SPI,
236 .start = UART0_THR,
241 .start = IRQ_UART0_TX,
246 .start = IRQ_UART0_RX,
251 .start = IRQ_UART0_ERROR,
256 .start = CH_UART0_TX,
261 .start = CH_UART0_RX,
284 .start = UART1_THR,
289 .start = IRQ_UART1_TX,
294 .start = IRQ_UART1_RX,
299 .start = IRQ_UART1_ERROR,
304 .start = CH_UART1_TX,
309 .start = CH_UART1_RX,
335 .start = 0xFFC00400,
340 .start = IRQ_UART0_RX,
345 .start = CH_UART0_RX,
361 .start = 0xFFC02000,
366 .start = IRQ_UART1_RX,
371 .start = CH_UART1_RX,
391 .start = TWI0_REGBASE,
396 .start = IRQ_TWI,
417 .start = SPORT0_TCR1,
422 .start = IRQ_SPORT0_RX,
427 .start = IRQ_SPORT0_ERROR,
451 .start = SPORT1_TCR1,
456 .start = IRQ_SPORT1_RX,
461 .start = IRQ_SPORT1_ERROR,
H A Dtcm_bf537.c114 .start = SPI0_REGBASE,
119 .start = CH_SPI,
124 .start = IRQ_SPI,
172 .start = 0x20200300,
176 .start = IRQ_PF14,
196 .start = 0x20308000,
200 .start = 0x20308004,
204 .start = IRQ_PG15,
235 .start = 0x20300000,
239 .start = IRQ_PG13,
281 .start = 0x20000000,
285 .start = (unsigned long)cm_flash_gpios,
306 .start = UART0_THR,
311 .start = IRQ_UART0_TX,
316 .start = IRQ_UART0_RX,
321 .start = IRQ_UART0_ERROR,
326 .start = CH_UART0_TX,
331 .start = CH_UART0_RX,
354 .start = UART1_THR,
359 .start = IRQ_UART1_TX,
364 .start = IRQ_UART1_RX,
369 .start = IRQ_UART1_ERROR,
374 .start = CH_UART1_TX,
379 .start = CH_UART1_RX,
405 .start = 0xFFC00400,
410 .start = IRQ_UART0_RX,
415 .start = CH_UART0_RX,
431 .start = 0xFFC02000,
436 .start = IRQ_UART1_RX,
441 .start = CH_UART1_RX,
461 .start = TWI0_REGBASE,
466 .start = IRQ_TWI,
487 .start = SPORT0_TCR1,
492 .start = IRQ_SPORT0_RX,
497 .start = IRQ_SPORT0_ERROR,
521 .start = SPORT1_TCR1,
526 .start = IRQ_SPORT1_RX,
531 .start = IRQ_SPORT1_ERROR,
596 .start = 0x2030C000,
601 .start = 0x2030D018,
606 .start = PATA_INT,
H A Dcm_bf537e.c114 .start = SPI0_REGBASE,
119 .start = CH_SPI,
124 .start = IRQ_SPI,
160 .start = SPORT0_TCR1,
165 .start = IRQ_SPORT0_ERROR,
190 .start = SPORT1_TCR1,
195 .start = IRQ_SPORT1_ERROR,
237 .start = 0x20200300,
241 .start = IRQ_PF14,
261 .start = 0x20308000,
265 .start = 0x20308004,
269 .start = IRQ_PG15,
300 .start = 0x20300000,
304 .start = IRQ_PG13,
346 .start = 0x20000000,
350 .start = (unsigned long)cm_flash_gpios,
371 .start = UART0_THR,
376 .start = IRQ_UART0_TX,
381 .start = IRQ_UART0_RX,
386 .start = IRQ_UART0_ERROR,
391 .start = CH_UART0_TX,
396 .start = CH_UART0_RX,
405 .start = -1,
413 .start = -1,
437 .start = UART1_THR,
442 .start = IRQ_UART1_TX,
447 .start = IRQ_UART1_RX,
452 .start = IRQ_UART1_ERROR,
457 .start = CH_UART1_TX,
462 .start = CH_UART1_RX,
471 .start = -1,
479 .start = -1,
506 .start = 0xFFC00400,
511 .start = IRQ_UART0_RX,
516 .start = CH_UART0_RX,
531 .start = 0xFFC02000,
536 .start = IRQ_UART1_RX,
541 .start = CH_UART1_RX,
560 .start = TWI0_REGBASE,
565 .start = IRQ_TWI,
593 .start = SPORT0_TCR1,
598 .start = IRQ_SPORT0_RX,
603 .start = IRQ_SPORT0_ERROR,
622 .start = SPORT1_TCR1,
627 .start = IRQ_SPORT1_RX,
632 .start = IRQ_SPORT1_ERROR,
657 .start = SPORT0_TCR1,
662 .start = IRQ_SPORT0_RX,
667 .start = IRQ_SPORT0_TX,
672 .start = IRQ_SPORT0_ERROR,
677 .start = CH_SPORT0_TX,
682 .start = CH_SPORT0_RX,
740 .start = 0x2030C000,
745 .start = 0x2030D018,
750 .start = PATA_INT,
H A Dpnav10.c36 .start = 0x20310000, /* IO PORT */
40 .start = 0x20311000, /* Attribute Memory */
44 .start = IRQ_PF4,
48 .start = 6, /* Card Detect PF6 */
81 .start = 0x20300300,
86 .start = IRQ_PF7,
138 .start = 0x20300000,
142 .start = IRQ_PF7,
260 .start = SPI0_REGBASE,
265 .start = CH_SPI,
270 .start = IRQ_SPI,
304 .start = UART0_THR,
309 .start = IRQ_UART0_TX,
314 .start = IRQ_UART0_RX,
319 .start = IRQ_UART0_ERROR,
324 .start = CH_UART0_TX,
329 .start = CH_UART0_RX,
352 .start = UART1_THR,
357 .start = IRQ_UART1_TX,
362 .start = IRQ_UART1_RX,
367 .start = IRQ_UART1_ERROR,
372 .start = CH_UART1_TX,
377 .start = CH_UART1_RX,
403 .start = 0xFFC00400,
408 .start = IRQ_UART0_RX,
413 .start = CH_UART0_RX,
429 .start = 0xFFC02000,
434 .start = IRQ_UART1_RX,
439 .start = CH_UART1_RX,
H A Dcm_bf537u.c114 .start = SPI0_REGBASE,
119 .start = CH_SPI,
124 .start = IRQ_SPI,
172 .start = 0x20200300,
176 .start = IRQ_PF14,
196 .start = 0x20308000,
200 .start = 0x20308004,
204 .start = IRQ_PG15,
235 .start = 0x20200000,
239 .start = IRQ_PH14,
281 .start = 0x20000000,
285 .start = (unsigned long)cm_flash_gpios,
306 .start = UART0_THR,
311 .start = IRQ_UART0_TX,
316 .start = IRQ_UART0_RX,
321 .start = IRQ_UART0_ERROR,
326 .start = CH_UART0_TX,
331 .start = CH_UART0_RX,
354 .start = UART1_THR,
359 .start = IRQ_UART1_TX,
364 .start = IRQ_UART1_RX,
369 .start = IRQ_UART1_ERROR,
374 .start = CH_UART1_TX,
379 .start = CH_UART1_RX,
405 .start = 0xFFC00400,
410 .start = IRQ_UART0_RX,
415 .start = CH_UART0_RX,
430 .start = 0xFFC02000,
435 .start = IRQ_UART1_RX,
440 .start = CH_UART1_RX,
459 .start = TWI0_REGBASE,
464 .start = IRQ_TWI,
485 .start = SPORT0_TCR1,
490 .start = IRQ_SPORT0_RX,
495 .start = IRQ_SPORT0_ERROR,
519 .start = SPORT1_TCR1,
524 .start = IRQ_SPORT1_RX,
529 .start = IRQ_SPORT1_ERROR,
594 .start = 0x2030C000,
599 .start = 0x2030D018,
604 .start = PATA_INT,
/linux-4.4.14/arch/blackfin/mach-bf538/boards/
H A Dezkit.c47 .start = UART0_THR,
52 .start = IRQ_UART0_TX,
57 .start = IRQ_UART0_RX,
62 .start = IRQ_UART0_ERROR,
67 .start = CH_UART0_TX,
72 .start = CH_UART0_RX,
78 .start = GPIO_PG7,
83 .start = GPIO_PG6,
107 .start = UART1_THR,
112 .start = IRQ_UART1_TX,
117 .start = IRQ_UART1_RX,
122 .start = IRQ_UART1_ERROR,
127 .start = CH_UART1_TX,
132 .start = CH_UART1_RX,
155 .start = UART2_THR,
160 .start = IRQ_UART2_TX,
165 .start = IRQ_UART2_RX,
170 .start = IRQ_UART2_ERROR,
175 .start = CH_UART2_TX,
180 .start = CH_UART2_RX,
206 .start = 0xFFC00400,
211 .start = IRQ_UART0_RX,
216 .start = CH_UART0_RX,
231 .start = 0xFFC02000,
236 .start = IRQ_UART1_RX,
241 .start = CH_UART1_RX,
256 .start = 0xFFC02100,
261 .start = IRQ_UART2_RX,
266 .start = CH_UART2_RX,
284 .start = SPORT0_TCR1,
289 .start = IRQ_SPORT0_RX,
294 .start = IRQ_SPORT0_ERROR,
318 .start = SPORT1_TCR1,
323 .start = IRQ_SPORT1_RX,
328 .start = IRQ_SPORT1_ERROR,
352 .start = SPORT2_TCR1,
357 .start = IRQ_SPORT2_RX,
362 .start = IRQ_SPORT2_ERROR,
386 .start = SPORT3_TCR1,
391 .start = IRQ_SPORT3_RX,
396 .start = IRQ_SPORT3_ERROR,
426 .start = 0xFFC02A00,
431 .start = IRQ_CAN_RX,
436 .start = IRQ_CAN_TX,
441 .start = IRQ_CAN_ERROR,
473 .start = 0x20310300,
477 .start = IRQ_PF0,
552 .start = IRQ_PPI_ERROR,
615 .start = SPI0_REGBASE,
620 .start = CH_SPI0,
625 .start = IRQ_SPI0,
634 .start = SPI1_REGBASE,
639 .start = CH_SPI1,
644 .start = IRQ_SPI1,
653 .start = SPI2_REGBASE,
658 .start = CH_SPI2,
663 .start = IRQ_SPI2,
723 .start = TWI0_REGBASE,
728 .start = IRQ_TWI0,
748 .start = TWI1_REGBASE,
753 .start = IRQ_TWI1,
839 .start = 0x20000000,
/linux-4.4.14/arch/microblaze/kernel/cpu/
H A Dcache.c92 #define CACHE_LOOP_LIMITS(start, end, cache_line_length, cache_size) \
95 end = min(start + cache_size, end); \
96 start &= align; \
119 * start address is cache aligned
124 #define CACHE_RANGE_LOOP_2(start, end, line_length, op) \
130 count = end - start; \
136 : : "r" (start), "r" (count), \
141 #define CACHE_RANGE_LOOP_1(start, end, line_length, op) \
146 WARN_ON(end < start); \
152 : : "r" (temp), "r" (start), "r" (end), \
158 static void __flush_icache_range_msr_irq(unsigned long start, unsigned long end) __flush_icache_range_msr_irq() argument
164 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __flush_icache_range_msr_irq()
165 (unsigned int)start, (unsigned int) end); __flush_icache_range_msr_irq()
167 CACHE_LOOP_LIMITS(start, end, __flush_icache_range_msr_irq()
174 CACHE_RANGE_LOOP_1(start, end, cpuinfo.icache_line_length, wic); __flush_icache_range_msr_irq()
176 for (i = start; i < end; i += cpuinfo.icache_line_length) __flush_icache_range_msr_irq()
184 static void __flush_icache_range_nomsr_irq(unsigned long start, __flush_icache_range_nomsr_irq() argument
191 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __flush_icache_range_nomsr_irq()
192 (unsigned int)start, (unsigned int) end); __flush_icache_range_nomsr_irq()
194 CACHE_LOOP_LIMITS(start, end, __flush_icache_range_nomsr_irq()
201 CACHE_RANGE_LOOP_1(start, end, cpuinfo.icache_line_length, wic); __flush_icache_range_nomsr_irq()
203 for (i = start; i < end; i += cpuinfo.icache_line_length) __flush_icache_range_nomsr_irq()
212 static void __flush_icache_range_noirq(unsigned long start, __flush_icache_range_noirq() argument
218 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __flush_icache_range_noirq()
219 (unsigned int)start, (unsigned int) end); __flush_icache_range_noirq()
221 CACHE_LOOP_LIMITS(start, end, __flush_icache_range_noirq()
224 CACHE_RANGE_LOOP_1(start, end, cpuinfo.icache_line_length, wic); __flush_icache_range_noirq()
226 for (i = start; i < end; i += cpuinfo.icache_line_length) __flush_icache_range_noirq()
376 static void __invalidate_dcache_range_wb(unsigned long start, __invalidate_dcache_range_wb() argument
382 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __invalidate_dcache_range_wb()
383 (unsigned int)start, (unsigned int) end); __invalidate_dcache_range_wb()
385 CACHE_LOOP_LIMITS(start, end, __invalidate_dcache_range_wb()
388 CACHE_RANGE_LOOP_2(start, end, cpuinfo.dcache_line_length, wdc.clear); __invalidate_dcache_range_wb()
390 for (i = start; i < end; i += cpuinfo.dcache_line_length) __invalidate_dcache_range_wb()
396 static void __invalidate_dcache_range_nomsr_wt(unsigned long start, __invalidate_dcache_range_nomsr_wt() argument
402 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __invalidate_dcache_range_nomsr_wt()
403 (unsigned int)start, (unsigned int) end); __invalidate_dcache_range_nomsr_wt()
404 CACHE_LOOP_LIMITS(start, end, __invalidate_dcache_range_nomsr_wt()
408 CACHE_RANGE_LOOP_1(start, end, cpuinfo.dcache_line_length, wdc); __invalidate_dcache_range_nomsr_wt()
410 for (i = start; i < end; i += cpuinfo.dcache_line_length) __invalidate_dcache_range_nomsr_wt()
416 static void __invalidate_dcache_range_msr_irq_wt(unsigned long start, __invalidate_dcache_range_msr_irq_wt() argument
423 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __invalidate_dcache_range_msr_irq_wt()
424 (unsigned int)start, (unsigned int) end); __invalidate_dcache_range_msr_irq_wt()
425 CACHE_LOOP_LIMITS(start, end, __invalidate_dcache_range_msr_irq_wt()
432 CACHE_RANGE_LOOP_1(start, end, cpuinfo.dcache_line_length, wdc); __invalidate_dcache_range_msr_irq_wt()
434 for (i = start; i < end; i += cpuinfo.dcache_line_length) __invalidate_dcache_range_msr_irq_wt()
443 static void __invalidate_dcache_range_nomsr_irq(unsigned long start, __invalidate_dcache_range_nomsr_irq() argument
450 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __invalidate_dcache_range_nomsr_irq()
451 (unsigned int)start, (unsigned int) end); __invalidate_dcache_range_nomsr_irq()
453 CACHE_LOOP_LIMITS(start, end, __invalidate_dcache_range_nomsr_irq()
460 CACHE_RANGE_LOOP_1(start, end, cpuinfo.dcache_line_length, wdc); __invalidate_dcache_range_nomsr_irq()
462 for (i = start; i < end; i += cpuinfo.dcache_line_length) __invalidate_dcache_range_nomsr_irq()
488 static void __flush_dcache_range_wb(unsigned long start, unsigned long end) __flush_dcache_range_wb() argument
493 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __flush_dcache_range_wb()
494 (unsigned int)start, (unsigned int) end); __flush_dcache_range_wb()
496 CACHE_LOOP_LIMITS(start, end, __flush_dcache_range_wb()
499 CACHE_RANGE_LOOP_2(start, end, cpuinfo.dcache_line_length, wdc.flush); __flush_dcache_range_wb()
501 for (i = start; i < end; i += cpuinfo.dcache_line_length) __flush_dcache_range_wb()
/linux-4.4.14/arch/c6x/platforms/
H A Dcache.c132 static void cache_block_operation(unsigned int *start, cache_block_operation() argument
140 - L2_CACHE_ALIGN_LOW((unsigned int) start)) >> 2; cache_block_operation()
143 for (; wcnt; wcnt -= wc, start += wc) { cache_block_operation()
160 imcr_set(bar_reg, L2_CACHE_ALIGN_LOW((unsigned int) start)); cache_block_operation()
177 static void cache_block_operation_nowait(unsigned int *start, cache_block_operation_nowait() argument
185 - L2_CACHE_ALIGN_LOW((unsigned int) start)) >> 2; cache_block_operation_nowait()
188 for (; wcnt; wcnt -= wc, start += wc) { cache_block_operation_nowait()
192 imcr_set(bar_reg, L2_CACHE_ALIGN_LOW((unsigned int) start)); cache_block_operation_nowait()
325 void enable_caching(unsigned long start, unsigned long end) enable_caching() argument
327 unsigned int mar = IMCR_MAR_BASE + ((start >> 24) << 2); enable_caching()
334 void disable_caching(unsigned long start, unsigned long end) disable_caching() argument
336 unsigned int mar = IMCR_MAR_BASE + ((start >> 24) << 2); disable_caching()
347 void L1P_cache_block_invalidate(unsigned int start, unsigned int end) L1P_cache_block_invalidate() argument
349 cache_block_operation((unsigned int *) start, L1P_cache_block_invalidate()
355 void L1D_cache_block_invalidate(unsigned int start, unsigned int end) L1D_cache_block_invalidate() argument
357 cache_block_operation((unsigned int *) start, L1D_cache_block_invalidate()
362 void L1D_cache_block_writeback_invalidate(unsigned int start, unsigned int end) L1D_cache_block_writeback_invalidate() argument
364 cache_block_operation((unsigned int *) start, L1D_cache_block_writeback_invalidate()
369 void L1D_cache_block_writeback(unsigned int start, unsigned int end) L1D_cache_block_writeback() argument
371 cache_block_operation((unsigned int *) start, L1D_cache_block_writeback()
380 void L2_cache_block_invalidate(unsigned int start, unsigned int end) L2_cache_block_invalidate() argument
382 cache_block_operation((unsigned int *) start, L2_cache_block_invalidate()
387 void L2_cache_block_writeback(unsigned int start, unsigned int end) L2_cache_block_writeback() argument
389 cache_block_operation((unsigned int *) start, L2_cache_block_writeback()
394 void L2_cache_block_writeback_invalidate(unsigned int start, unsigned int end) L2_cache_block_writeback_invalidate() argument
396 cache_block_operation((unsigned int *) start, L2_cache_block_writeback_invalidate()
401 void L2_cache_block_invalidate_nowait(unsigned int start, unsigned int end) L2_cache_block_invalidate_nowait() argument
403 cache_block_operation_nowait((unsigned int *) start, L2_cache_block_invalidate_nowait()
408 void L2_cache_block_writeback_nowait(unsigned int start, unsigned int end) L2_cache_block_writeback_nowait() argument
410 cache_block_operation_nowait((unsigned int *) start, L2_cache_block_writeback_nowait()
415 void L2_cache_block_writeback_invalidate_nowait(unsigned int start, L2_cache_block_writeback_invalidate_nowait() argument
418 cache_block_operation_nowait((unsigned int *) start, L2_cache_block_writeback_invalidate_nowait()
/linux-4.4.14/arch/cris/boot/rescue/
H A Dhead_v32.S2 * Just get started by jumping to CONFIG_ETRAX_PTABLE_SECTOR to start
17 start: label
/linux-4.4.14/arch/nios2/include/asm/
H A Dcacheflush.h26 extern void flush_cache_range(struct vm_area_struct *vma, unsigned long start,
33 extern void flush_icache_range(unsigned long start, unsigned long end);
36 #define flush_cache_vmap(start, end) flush_dcache_range(start, end)
37 #define flush_cache_vunmap(start, end) flush_dcache_range(start, end)
46 extern void flush_dcache_range(unsigned long start, unsigned long end);
47 extern void invalidate_dcache_range(unsigned long start, unsigned long end);
/linux-4.4.14/fs/btrfs/
H A Dextent_io.h71 u64 start, u64 end, int *page_started,
73 int (*writepage_start_hook)(struct page *page, u64 start, u64 end);
74 int (*writepage_io_hook)(struct page *page, u64 start, u64 end);
81 struct page *page, u64 start, u64 end,
83 int (*writepage_end_io_hook)(struct page *page, u64 start, u64 end,
106 u64 start; member in struct:extent_state
126 u64 start; member in struct:extent_buffer
193 u64 start, u64 len,
202 int lock_extent(struct extent_io_tree *tree, u64 start, u64 end);
203 int lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
205 int unlock_extent(struct extent_io_tree *tree, u64 start, u64 end);
206 int unlock_extent_cached(struct extent_io_tree *tree, u64 start, u64 end,
208 int try_lock_extent(struct extent_io_tree *tree, u64 start, u64 end);
217 u64 *start, u64 search_end,
221 int test_range_bit(struct extent_io_tree *tree, u64 start, u64 end,
224 int clear_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
226 int clear_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
229 int clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
232 int set_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
234 int set_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
237 int set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
240 int set_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end,
242 int clear_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end,
244 int set_extent_new(struct extent_io_tree *tree, u64 start, u64 end,
246 int set_extent_dirty(struct extent_io_tree *tree, u64 start, u64 end,
248 int clear_extent_dirty(struct extent_io_tree *tree, u64 start, u64 end,
250 int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
253 int set_extent_delalloc(struct extent_io_tree *tree, u64 start, u64 end,
255 int set_extent_defrag(struct extent_io_tree *tree, u64 start, u64 end,
257 int find_first_extent_bit(struct extent_io_tree *tree, u64 start,
266 u64 start, u64 end, get_extent_t *get_extent,
279 __u64 start, __u64 len, get_extent_t *get_extent);
280 int get_state_private(struct extent_io_tree *tree, u64 start, u64 *private);
284 u64 start);
286 u64 start);
289 u64 start);
296 struct extent_buffer *eb, u64 start, int wait,
300 static inline unsigned long num_extent_pages(u64 start, u64 len) num_extent_pages() argument
302 return ((start + len + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT) - num_extent_pages()
303 (start >> PAGE_CACHE_SHIFT); num_extent_pages()
312 unsigned long start,
315 unsigned long start,
318 unsigned long start,
321 unsigned long start, unsigned long len);
330 unsigned long start, unsigned long len);
341 int extent_range_clear_dirty_for_io(struct inode *inode, u64 start, u64 end);
342 int extent_range_redirty_for_io(struct inode *inode, u64 start, u64 end);
343 int extent_clear_unlock_delalloc(struct inode *inode, u64 start, u64 end,
355 int repair_io_failure(struct inode *inode, u64 start, u64 length, u64 logical,
358 int clean_io_failure(struct inode *inode, u64 start, struct page *page,
360 int end_extent_writepage(struct page *page, int err, u64 start, u64 end);
374 u64 start; member in struct:io_failure_record
383 void btrfs_free_io_failure_record(struct inode *inode, u64 start, u64 end);
384 int btrfs_get_io_failure_record(struct inode *inode, u64 start, u64 end,
396 struct page *locked_page, u64 *start,
400 u64 start);
H A Dextent_io.c67 pr_err("BTRFS: state leak: start %llu end %llu state %u in tree %d refs %d\n", btrfs_leak_debug_check()
68 state->start, state->end, state->state, btrfs_leak_debug_check()
77 printk(KERN_ERR "BTRFS: buffer leak start %llu len %lu " btrfs_leak_debug_check()
79 eb->start, eb->len, atomic_read(&eb->refs)); btrfs_leak_debug_check()
85 #define btrfs_debug_check_extent_io_range(tree, start, end) \
86 __btrfs_debug_check_extent_io_range(__func__, (tree), (start), (end)) __btrfs_debug_check_extent_io_range()
88 struct extent_io_tree *tree, u64 start, u64 end) __btrfs_debug_check_extent_io_range()
101 caller, btrfs_ino(inode), isize, start, end); __btrfs_debug_check_extent_io_range()
114 u64 start; member in struct:tree_entry
146 changeset->bytes_changed += state->end - state->start + 1; add_extent_changeset()
147 ret = ulist_add(changeset->range_changed, state->start, state->end, add_extent_changeset()
278 if (offset < entry->start) tree_insert()
310 if (offset < entry->start) __etree_search()
335 while (prev && offset < prev_entry->start) { __etree_search()
394 if (other->end == state->start - 1 && merge_state()
397 state->start = other->start; merge_state()
406 if (other->start == state->end + 1 && merge_state()
446 struct extent_state *state, u64 start, u64 end, insert_state()
453 if (end < start) insert_state()
454 WARN(1, KERN_ERR "BTRFS: end < start %llu %llu\n", insert_state()
455 end, start); insert_state()
456 state->start = start; insert_state()
467 found->start, found->end, start, end); insert_state()
487 * the tree has 'orig' at [orig->start, orig->end]. After calling, there
489 * prealloc: [orig->start, split - 1]
502 prealloc->start = orig->start; split_state()
505 orig->start = split; split_state()
541 u64 range = state->end - state->start + 1; clear_state_bit()
590 * the range [start, end] is inclusive.
594 static int __clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, __clear_extent_bit() argument
607 btrfs_debug_check_extent_io_range(tree, start, end); __clear_extent_bit()
640 cached->start <= start && cached->end > start) { __clear_extent_bit()
653 node = tree_search(tree, start); __clear_extent_bit()
658 if (state->start > end) __clear_extent_bit()
660 WARN_ON(state->end < start); __clear_extent_bit()
685 if (state->start < start) { __clear_extent_bit()
688 err = split_state(tree, state, prealloc, start); __clear_extent_bit()
708 if (state->start <= end && state->end > end) { __clear_extent_bit()
728 start = last_end + 1; __clear_extent_bit()
729 if (start <= end && state && !need_resched()) __clear_extent_bit()
741 if (start > end) __clear_extent_bit()
764 * The range [start, end] is inclusive.
767 static void wait_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, wait_extent_bit() argument
773 btrfs_debug_check_extent_io_range(tree, start, end); wait_extent_bit()
782 node = tree_search(tree, start); wait_extent_bit()
789 if (state->start > end) wait_extent_bit()
793 start = state->start; wait_extent_bit()
799 start = state->end + 1; wait_extent_bit()
801 if (start > end) wait_extent_bit()
821 u64 range = state->end - state->start + 1; set_state_bits()
852 * part of the range already has the desired bits set. The start of the
855 * [start, end] is inclusive This takes the tree lock.
859 __set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, __set_extent_bit() argument
873 btrfs_debug_check_extent_io_range(tree, start, end); __set_extent_bit()
885 if (state->start <= start && state->end > start && __set_extent_bit()
895 node = tree_search_for_insert(tree, start, &p, &parent); __set_extent_bit()
899 err = insert_state(tree, prealloc, start, end, __set_extent_bit()
910 last_start = state->start; __set_extent_bit()
919 if (state->start == start && state->end <= end) { __set_extent_bit()
921 *failed_start = state->start; __set_extent_bit()
931 start = last_end + 1; __set_extent_bit()
933 if (start < end && state && state->start == start && __set_extent_bit()
955 if (state->start < start) { __set_extent_bit()
957 *failed_start = start; __set_extent_bit()
964 err = split_state(tree, state, prealloc, start); __set_extent_bit()
977 start = last_end + 1; __set_extent_bit()
979 if (start < end && state && state->start == start && __set_extent_bit()
992 if (state->start > start) { __set_extent_bit()
1006 err = insert_state(tree, prealloc, start, this_end, __set_extent_bit()
1013 start = this_end + 1; __set_extent_bit()
1022 if (state->start <= end && state->end > end) { __set_extent_bit()
1024 *failed_start = start; __set_extent_bit()
1052 if (start > end) __set_extent_bit()
1060 int set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, set_extent_bit() argument
1064 return __set_extent_bit(tree, start, end, bits, 0, failed_start, set_extent_bit()
1073 * @start: the start offset in bytes
1086 int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, convert_extent_bit() argument
1100 btrfs_debug_check_extent_io_range(tree, start, end); convert_extent_bit()
1119 if (state->start <= start && state->end > start && convert_extent_bit()
1130 node = tree_search_for_insert(tree, start, &p, &parent); convert_extent_bit()
1137 err = insert_state(tree, prealloc, start, end, convert_extent_bit()
1147 last_start = state->start; convert_extent_bit()
1156 if (state->start == start && state->end <= end) { convert_extent_bit()
1162 start = last_end + 1; convert_extent_bit()
1163 if (start < end && state && state->start == start && convert_extent_bit()
1185 if (state->start < start) { convert_extent_bit()
1191 err = split_state(tree, state, prealloc, start); convert_extent_bit()
1204 start = last_end + 1; convert_extent_bit()
1205 if (start < end && state && state->start == start && convert_extent_bit()
1218 if (state->start > start) { convert_extent_bit()
1235 err = insert_state(tree, prealloc, start, this_end, convert_extent_bit()
1241 start = this_end + 1; convert_extent_bit()
1250 if (state->start <= end && state->end > end) { convert_extent_bit()
1278 if (start > end) convert_extent_bit()
1288 int set_extent_dirty(struct extent_io_tree *tree, u64 start, u64 end, set_extent_dirty() argument
1291 return set_extent_bit(tree, start, end, EXTENT_DIRTY, NULL, set_extent_dirty()
1295 int set_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, set_extent_bits() argument
1298 return set_extent_bit(tree, start, end, bits, NULL, set_extent_bits()
1302 int set_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, set_record_extent_bits() argument
1314 return __set_extent_bit(tree, start, end, bits, 0, NULL, NULL, mask, set_record_extent_bits()
1318 int clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, clear_extent_bit() argument
1322 return __clear_extent_bit(tree, start, end, bits, wake, delete, clear_extent_bit()
1326 int clear_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, clear_extent_bits() argument
1334 return clear_extent_bit(tree, start, end, bits, wake, 0, NULL, mask); clear_extent_bits()
1337 int clear_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, clear_record_extent_bits() argument
1347 return __clear_extent_bit(tree, start, end, bits, 0, 0, NULL, mask, clear_record_extent_bits()
1351 int set_extent_delalloc(struct extent_io_tree *tree, u64 start, u64 end, set_extent_delalloc() argument
1354 return set_extent_bit(tree, start, end, set_extent_delalloc()
1359 int set_extent_defrag(struct extent_io_tree *tree, u64 start, u64 end, set_extent_defrag() argument
1362 return set_extent_bit(tree, start, end, set_extent_defrag()
1367 int clear_extent_dirty(struct extent_io_tree *tree, u64 start, u64 end, clear_extent_dirty() argument
1370 return clear_extent_bit(tree, start, end, clear_extent_dirty()
1375 int set_extent_new(struct extent_io_tree *tree, u64 start, u64 end, set_extent_new() argument
1378 return set_extent_bit(tree, start, end, EXTENT_NEW, NULL, set_extent_new()
1382 int set_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end, set_extent_uptodate() argument
1385 return set_extent_bit(tree, start, end, EXTENT_UPTODATE, NULL, set_extent_uptodate()
1389 int clear_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end, clear_extent_uptodate() argument
1392 return clear_extent_bit(tree, start, end, EXTENT_UPTODATE, 0, 0, clear_extent_uptodate()
1397 * either insert or lock state struct between start and end use mask to tell
1400 int lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, lock_extent_bits() argument
1407 err = __set_extent_bit(tree, start, end, EXTENT_LOCKED | bits, lock_extent_bits()
1412 start = failed_start; lock_extent_bits()
1415 WARN_ON(start > end); lock_extent_bits()
1420 int lock_extent(struct extent_io_tree *tree, u64 start, u64 end) lock_extent() argument
1422 return lock_extent_bits(tree, start, end, 0, NULL); lock_extent()
1425 int try_lock_extent(struct extent_io_tree *tree, u64 start, u64 end) try_lock_extent() argument
1430 err = __set_extent_bit(tree, start, end, EXTENT_LOCKED, EXTENT_LOCKED, try_lock_extent()
1433 if (failed_start > start) try_lock_extent()
1434 clear_extent_bit(tree, start, failed_start - 1, try_lock_extent()
1441 int unlock_extent_cached(struct extent_io_tree *tree, u64 start, u64 end, unlock_extent_cached() argument
1444 return clear_extent_bit(tree, start, end, EXTENT_LOCKED, 1, 0, cached, unlock_extent_cached()
1448 int unlock_extent(struct extent_io_tree *tree, u64 start, u64 end) unlock_extent() argument
1450 return clear_extent_bit(tree, start, end, EXTENT_LOCKED, 1, 0, NULL, unlock_extent()
1454 int extent_range_clear_dirty_for_io(struct inode *inode, u64 start, u64 end) extent_range_clear_dirty_for_io() argument
1456 unsigned long index = start >> PAGE_CACHE_SHIFT; extent_range_clear_dirty_for_io()
1470 int extent_range_redirty_for_io(struct inode *inode, u64 start, u64 end) extent_range_redirty_for_io() argument
1472 unsigned long index = start >> PAGE_CACHE_SHIFT; extent_range_redirty_for_io()
1490 static int set_range_writeback(struct extent_io_tree *tree, u64 start, u64 end) set_range_writeback() argument
1492 unsigned long index = start >> PAGE_CACHE_SHIFT; set_range_writeback()
1506 /* find the first state struct with 'bits' set after 'start', and
1508 * nothing was found after 'start'
1512 u64 start, unsigned bits) find_first_extent_bit_state()
1521 node = tree_search(tree, start); find_first_extent_bit_state()
1527 if (state->end >= start && (state->state & bits)) find_first_extent_bit_state()
1545 int find_first_extent_bit(struct extent_io_tree *tree, u64 start, find_first_extent_bit() argument
1556 if (state->end == start - 1 && extent_state_in_tree(state)) { find_first_extent_bit()
1573 state = find_first_extent_bit_state(tree, start, bits); find_first_extent_bit()
1577 *start_ret = state->start; find_first_extent_bit()
1588 * more than 'max_bytes'. start and end are used to return the range,
1593 u64 *start, u64 *end, u64 max_bytes, find_delalloc_range()
1598 u64 cur_start = *start; find_delalloc_range()
1617 if (found && (state->start != cur_start || find_delalloc_range()
1627 *start = state->start; find_delalloc_range()
1635 total_bytes += state->end - state->start + 1; find_delalloc_range()
1648 u64 start, u64 end) __unlock_for_delalloc()
1652 unsigned long index = start >> PAGE_CACHE_SHIFT; __unlock_for_delalloc()
1689 /* the caller is responsible for locking the start index */ lock_delalloc_pages()
1693 /* skip the page at the start index */ lock_delalloc_pages()
1739 * more than 'max_bytes'. start and end are used to return the range,
1745 struct page *locked_page, u64 *start, find_lock_delalloc_range()
1756 /* step one, find a bunch of delalloc bytes starting at start */ find_lock_delalloc_range()
1757 delalloc_start = *start; find_lock_delalloc_range()
1761 if (!found || delalloc_end <= *start) { find_lock_delalloc_range()
1762 *start = delalloc_start; find_lock_delalloc_range()
1769 * start comes from the offset of locked_page. We have to lock find_lock_delalloc_range()
1773 if (delalloc_start < *start) find_lock_delalloc_range()
1774 delalloc_start = *start; find_lock_delalloc_range()
1782 /* step two, lock all the pages after the page that has start */ find_lock_delalloc_range()
1817 *start = delalloc_start; find_lock_delalloc_range()
1823 int extent_clear_unlock_delalloc(struct inode *inode, u64 start, u64 end, extent_clear_unlock_delalloc() argument
1831 unsigned long index = start >> PAGE_CACHE_SHIFT; extent_clear_unlock_delalloc()
1836 clear_extent_bit(tree, start, end, clear_bits, 1, 0, NULL, GFP_NOFS); extent_clear_unlock_delalloc()
1881 u64 *start, u64 search_end, u64 max_bytes, count_range_bits()
1886 u64 cur_start = *start; count_range_bits()
1909 if (state->start > search_end) count_range_bits()
1911 if (contig && found && state->start > last + 1) count_range_bits()
1915 max(cur_start, state->start); count_range_bits()
1919 *start = max(cur_start, state->start); count_range_bits()
1939 static int set_state_private(struct extent_io_tree *tree, u64 start, u64 private) set_state_private() argument
1950 node = tree_search(tree, start); set_state_private()
1956 if (state->start != start) { set_state_private()
1966 int get_state_private(struct extent_io_tree *tree, u64 start, u64 *private) get_state_private() argument
1977 node = tree_search(tree, start); get_state_private()
1983 if (state->start != start) { get_state_private()
1999 int test_range_bit(struct extent_io_tree *tree, u64 start, u64 end, test_range_bit() argument
2007 if (cached && extent_state_in_tree(cached) && cached->start <= start && test_range_bit()
2008 cached->end > start) test_range_bit()
2011 node = tree_search(tree, start); test_range_bit()
2012 while (node && start <= end) { test_range_bit()
2015 if (filled && state->start > start) { test_range_bit()
2020 if (state->start > end) test_range_bit()
2035 start = state->end + 1; test_range_bit()
2036 if (start > end) test_range_bit()
2055 u64 start = page_offset(page); check_page_uptodate() local
2056 u64 end = start + PAGE_CACHE_SIZE - 1; check_page_uptodate()
2057 if (test_range_bit(tree, start, end, EXTENT_UPTODATE, 1, NULL)) check_page_uptodate()
2067 set_state_private(failure_tree, rec->start, 0); free_io_failure()
2068 ret = clear_extent_bits(failure_tree, rec->start, free_io_failure()
2069 rec->start + rec->len - 1, free_io_failure()
2074 ret = clear_extent_bits(&BTRFS_I(inode)->io_tree, rec->start, free_io_failure()
2075 rec->start + rec->len - 1, free_io_failure()
2094 int repair_io_failure(struct inode *inode, u64 start, u64 length, u64 logical, repair_io_failure() argument
2146 btrfs_ino(inode), start, repair_io_failure()
2155 u64 start = eb->start; repair_eb_io_failure() local
2156 unsigned long i, num_pages = num_extent_pages(eb->start, eb->len); repair_eb_io_failure()
2165 ret = repair_io_failure(root->fs_info->btree_inode, start, repair_eb_io_failure()
2166 PAGE_CACHE_SIZE, start, p, repair_eb_io_failure()
2167 start - page_offset(p), mirror_num); repair_eb_io_failure()
2170 start += PAGE_CACHE_SIZE; repair_eb_io_failure()
2180 int clean_io_failure(struct inode *inode, u64 start, struct page *page, clean_io_failure() argument
2197 ret = get_state_private(&BTRFS_I(inode)->io_failure_tree, start, clean_io_failure()
2208 failrec->start); clean_io_failure()
2216 failrec->start, clean_io_failure()
2220 if (state && state->start <= failrec->start && clean_io_failure()
2221 state->end >= failrec->start + failrec->len - 1) { clean_io_failure()
2225 repair_io_failure(inode, start, failrec->len, clean_io_failure()
2243 void btrfs_free_io_failure_record(struct inode *inode, u64 start, u64 end) btrfs_free_io_failure_record() argument
2253 state = find_first_extent_bit_state(failure_tree, start, EXTENT_DIRTY); btrfs_free_io_failure_record()
2255 if (state->start > end) btrfs_free_io_failure_record()
2271 int btrfs_get_io_failure_record(struct inode *inode, u64 start, u64 end, btrfs_get_io_failure_record() argument
2283 ret = get_state_private(failure_tree, start, &private); btrfs_get_io_failure_record()
2289 failrec->start = start; btrfs_get_io_failure_record()
2290 failrec->len = end - start + 1; btrfs_get_io_failure_record()
2296 em = lookup_extent_mapping(em_tree, start, failrec->len); btrfs_get_io_failure_record()
2303 if (em->start > start || em->start + em->len <= start) { btrfs_get_io_failure_record()
2313 logical = start - em->start; btrfs_get_io_failure_record()
2322 pr_debug("Get IO Failure Record: (new) logical=%llu, start=%llu, len=%llu\n", btrfs_get_io_failure_record()
2323 logical, start, failrec->len); btrfs_get_io_failure_record()
2329 ret = set_extent_bits(failure_tree, start, end, btrfs_get_io_failure_record()
2332 ret = set_state_private(failure_tree, start, btrfs_get_io_failure_record()
2336 ret = set_extent_bits(tree, start, end, EXTENT_DAMAGED, btrfs_get_io_failure_record()
2344 pr_debug("Get IO Failure Record: (found) logical=%llu, start=%llu, len=%llu, validation=%d\n", btrfs_get_io_failure_record()
2345 failrec->logical, failrec->start, failrec->len, btrfs_get_io_failure_record()
2466 struct page *page, u64 start, u64 end, bio_readpage_error()
2478 ret = btrfs_get_io_failure_record(inode, start, end, &failrec); bio_readpage_error()
2495 start - page_offset(page), bio_readpage_error()
2519 int end_extent_writepage(struct page *page, int err, u64 start, u64 end) end_extent_writepage() argument
2528 ret = tree->ops->writepage_end_io_hook(page, start, end_extent_writepage()
2555 u64 start; end_bio_extent_writepage() local
2579 start = page_offset(page); bio_for_each_segment_all()
2580 end = start + bvec->bv_offset + bvec->bv_len - 1; bio_for_each_segment_all()
2582 if (end_extent_writepage(page, bio->bi_error, start, end)) bio_for_each_segment_all()
2592 endio_readpage_release_extent(struct extent_io_tree *tree, u64 start, u64 len, endio_readpage_release_extent() argument
2596 u64 end = start + len - 1; endio_readpage_release_extent()
2599 set_extent_uptodate(tree, start, end, &cached, GFP_ATOMIC); endio_readpage_release_extent()
2600 unlock_extent_cached(tree, start, end, &cached, GFP_ATOMIC); endio_readpage_release_extent()
2621 u64 start; end_bio_extent_readpage() local
2656 start = page_offset(page); bio_for_each_segment_all()
2657 end = start + bvec->bv_offset + bvec->bv_len - 1; bio_for_each_segment_all()
2664 page, start, end, bio_for_each_segment_all()
2669 clean_io_failure(inode, start, page, 0); bio_for_each_segment_all()
2690 ret = bio_readpage_error(bio, offset, page, start, end, bio_for_each_segment_all()
2724 endio_readpage_release_extent(tree, start, bio_for_each_segment_all()
2725 end - start + 1, 0); bio_for_each_segment_all()
2727 extent_start = start; bio_for_each_segment_all()
2728 extent_len = end + 1 - start; bio_for_each_segment_all()
2729 } else if (extent_start + extent_len == start) { bio_for_each_segment_all()
2730 extent_len += end + 1 - start; bio_for_each_segment_all()
2734 extent_start = start; bio_for_each_segment_all()
2735 extent_len = end + 1 - start; bio_for_each_segment_all()
2823 u64 start; submit_one_bio() local
2825 start = page_offset(page) + bvec->bv_offset; submit_one_bio()
2833 mirror_num, bio_flags, start); submit_one_bio()
2942 u64 start, u64 len, get_extent_t *get_extent, __get_extent_map()
2949 if (extent_map_in_tree(em) && start >= em->start && __get_extent_map()
2950 start < extent_map_end(em)) { __get_extent_map()
2959 em = get_extent(inode, page, pg_offset, start, len, 0); __get_extent_map()
2982 u64 start = page_offset(page); __do_readpage() local
2983 u64 page_end = start + PAGE_CACHE_SIZE - 1; __do_readpage()
2985 u64 cur = start; __do_readpage()
3008 unlock_extent(tree, start, end); __do_readpage()
3054 extent_offset = cur - em->start; __do_readpage()
3198 u64 start, u64 end, __do_contiguous_readpages()
3211 lock_extent(tree, start, end); __do_contiguous_readpages()
3212 ordered = btrfs_lookup_ordered_range(inode, start, __do_contiguous_readpages()
3213 end - start + 1); __do_contiguous_readpages()
3216 unlock_extent(tree, start, end); __do_contiguous_readpages()
3236 u64 start = 0; __extent_readpages() local
3245 start = page_start; __extent_readpages()
3246 end = start + PAGE_CACHE_SIZE - 1; __extent_readpages()
3252 index - first_index, start, __extent_readpages()
3256 start = page_start; __extent_readpages()
3257 end = start + PAGE_CACHE_SIZE - 1; __extent_readpages()
3264 index - first_index, start, __extent_readpages()
3278 u64 start = page_offset(page); __extent_read_full_page() local
3279 u64 end = start + PAGE_CACHE_SIZE - 1; __extent_read_full_page()
3283 lock_extent(tree, start, end); __extent_read_full_page()
3284 ordered = btrfs_lookup_ordered_extent(inode, start); __extent_read_full_page()
3287 unlock_extent(tree, start, end); __extent_read_full_page()
3407 /* did the fill delalloc function already unlock and start writepage_delalloc()
3427 * helper for __extent_writepage. This calls the writepage start hooks,
3443 u64 start = page_offset(page); __extent_writepage_io() local
3444 u64 page_end = start + PAGE_CACHE_SIZE - 1; __extent_writepage_io()
3446 u64 cur = start; __extent_writepage_io()
3461 ret = tree->ops->writepage_start_hook(page, start, __extent_writepage_io()
3484 if (i_size <= start) { __extent_writepage_io()
3486 tree->ops->writepage_end_io_hook(page, start, __extent_writepage_io()
3509 extent_offset = cur - em->start; __extent_writepage_io()
3601 u64 start = page_offset(page); __extent_writepage() local
3602 u64 page_end = start + PAGE_CACHE_SIZE - 1; __extent_writepage()
3644 ret = writepage_delalloc(inode, page, wbc, epd, start, &nr_written); __extent_writepage()
3663 end_extent_writepage(page, ret, start, page_end); __extent_writepage()
3733 num_pages = num_extent_pages(eb->start, eb->len); lock_extent_buffer_for_io()
3855 u64 offset = eb->start; write_one_eb()
3862 num_pages = num_extent_pages(eb->start, eb->len); write_one_eb()
4013 * back to the start of the file btree_write_cache_pages()
4150 * back to the start of the file extent_write_cache_pages()
4202 u64 start, u64 end, get_extent_t *get_extent, extent_write_locked_range()
4208 unsigned long nr_pages = (end - start + PAGE_CACHE_SIZE) >> extent_write_locked_range()
4222 .range_start = start, extent_write_locked_range()
4226 while (start <= end) { extent_write_locked_range()
4227 page = find_get_page(mapping, start >> PAGE_CACHE_SHIFT); extent_write_locked_range()
4232 tree->ops->writepage_end_io_hook(page, start, extent_write_locked_range()
4233 start + PAGE_CACHE_SIZE - 1, extent_write_locked_range()
4238 start += PAGE_CACHE_SIZE; extent_write_locked_range()
4321 u64 start = page_offset(page); extent_invalidatepage() local
4322 u64 end = start + PAGE_CACHE_SIZE - 1; extent_invalidatepage()
4325 start += ALIGN(offset, blocksize); extent_invalidatepage()
4326 if (start > end) extent_invalidatepage()
4329 lock_extent_bits(tree, start, end, 0, &cached_state); extent_invalidatepage()
4331 clear_extent_bit(tree, start, end, extent_invalidatepage()
4347 u64 start = page_offset(page); try_release_extent_state() local
4348 u64 end = start + PAGE_CACHE_SIZE - 1; try_release_extent_state()
4351 if (test_range_bit(tree, start, end, try_release_extent_state()
4361 ret = clear_extent_bit(tree, start, end, try_release_extent_state()
4386 u64 start = page_offset(page); try_release_extent_mapping() local
4387 u64 end = start + PAGE_CACHE_SIZE - 1; try_release_extent_mapping()
4392 while (start <= end) { try_release_extent_mapping()
4393 len = end - start + 1; try_release_extent_mapping()
4395 em = lookup_extent_mapping(map, start, len); try_release_extent_mapping()
4401 em->start != start) { try_release_extent_mapping()
4406 if (!test_range_bit(tree, em->start, try_release_extent_mapping()
4414 start = extent_map_end(em); try_release_extent_mapping()
4465 __u64 start, __u64 len, get_extent_t *get_extent) extent_fiemap()
4468 u64 off = start; extent_fiemap()
4469 u64 max = start + len; extent_fiemap()
4494 start = round_down(start, BTRFS_I(inode)->root->sectorsize); extent_fiemap()
4495 len = round_up(max, BTRFS_I(inode)->root->sectorsize) - start; extent_fiemap()
4520 * remember the start of the last extent. There are a extent_fiemap()
4531 * extents. so, we trust isize unless the start of the last extent is extent_fiemap()
4539 lock_extent_bits(&BTRFS_I(inode)->io_tree, start, start + len - 1, 0, extent_fiemap()
4542 em = get_extent_skip_holes(inode, start, last_for_get_extent, extent_fiemap()
4555 if (em->start >= max || extent_map_end(em) < off) extent_fiemap()
4564 em_start = max(em->start, off); extent_fiemap()
4567 * record the offset from the start of the extent extent_fiemap()
4573 offset_in_extent = em_start - em->start; extent_fiemap()
4597 (em->start - em->orig_start); extent_fiemap()
4653 unlock_extent_cached(&BTRFS_I(inode)->io_tree, start, start + len - 1, extent_fiemap()
4682 index = num_extent_pages(eb->start, eb->len); btrfs_release_extent_buffer_page()
4733 __alloc_extent_buffer(struct btrfs_fs_info *fs_info, u64 start, __alloc_extent_buffer() argument
4739 eb->start = start; __alloc_extent_buffer()
4775 unsigned long num_pages = num_extent_pages(src->start, src->len); btrfs_clone_extent_buffer()
4777 new = __alloc_extent_buffer(src->fs_info, src->start, src->len); btrfs_clone_extent_buffer()
4801 u64 start) alloc_dummy_extent_buffer()
4819 eb = __alloc_extent_buffer(fs_info, start, len); alloc_dummy_extent_buffer()
4880 num_pages = num_extent_pages(eb->start, eb->len); mark_extent_buffer_accessed()
4890 u64 start) find_extent_buffer()
4896 start >> PAGE_CACHE_SHIFT); find_extent_buffer()
4928 u64 start) alloc_test_extent_buffer()
4933 eb = find_extent_buffer(fs_info, start); alloc_test_extent_buffer()
4936 eb = alloc_dummy_extent_buffer(fs_info, start); alloc_test_extent_buffer()
4946 start >> PAGE_CACHE_SHIFT, eb); alloc_test_extent_buffer()
4950 exists = find_extent_buffer(fs_info, start); alloc_test_extent_buffer()
4974 u64 start) alloc_extent_buffer()
4977 unsigned long num_pages = num_extent_pages(start, len); alloc_extent_buffer()
4979 unsigned long index = start >> PAGE_CACHE_SHIFT; alloc_extent_buffer()
4987 eb = find_extent_buffer(fs_info, start); alloc_extent_buffer()
4991 eb = __alloc_extent_buffer(fs_info, start, len); alloc_extent_buffer()
5048 start >> PAGE_CACHE_SHIFT, eb); alloc_extent_buffer()
5052 exists = find_extent_buffer(fs_info, start); alloc_extent_buffer()
5111 eb->start >> PAGE_CACHE_SHIFT); release_extent_buffer()
5187 num_pages = num_extent_pages(eb->start, eb->len); clear_extent_buffer_dirty()
5221 num_pages = num_extent_pages(eb->start, eb->len); set_extent_buffer_dirty()
5237 num_pages = num_extent_pages(eb->start, eb->len); clear_extent_buffer_uptodate()
5253 num_pages = num_extent_pages(eb->start, eb->len); set_extent_buffer_uptodate()
5267 struct extent_buffer *eb, u64 start, int wait, read_extent_buffer_pages()
5285 if (start) { read_extent_buffer_pages()
5286 WARN_ON(start < eb->start); read_extent_buffer_pages()
5287 start_i = (start >> PAGE_CACHE_SHIFT) - read_extent_buffer_pages()
5288 (eb->start >> PAGE_CACHE_SHIFT); read_extent_buffer_pages()
5293 num_pages = num_extent_pages(eb->start, eb->len); read_extent_buffer_pages()
5363 unsigned long start, read_extent_buffer()
5371 size_t start_offset = eb->start & ((u64)PAGE_CACHE_SIZE - 1); read_extent_buffer()
5372 unsigned long i = (start_offset + start) >> PAGE_CACHE_SHIFT; read_extent_buffer()
5374 WARN_ON(start > eb->len); read_extent_buffer()
5375 WARN_ON(start + len > eb->start + eb->len); read_extent_buffer()
5377 offset = (start_offset + start) & (PAGE_CACHE_SIZE - 1); read_extent_buffer()
5394 unsigned long start, read_extent_buffer_to_user()
5402 size_t start_offset = eb->start & ((u64)PAGE_CACHE_SIZE - 1); read_extent_buffer_to_user()
5403 unsigned long i = (start_offset + start) >> PAGE_CACHE_SHIFT; read_extent_buffer_to_user()
5406 WARN_ON(start > eb->len); read_extent_buffer_to_user()
5407 WARN_ON(start + len > eb->start + eb->len); read_extent_buffer_to_user()
5409 offset = (start_offset + start) & (PAGE_CACHE_SIZE - 1); read_extent_buffer_to_user()
5430 int map_private_extent_buffer(struct extent_buffer *eb, unsigned long start, map_private_extent_buffer() argument
5435 size_t offset = start & (PAGE_CACHE_SIZE - 1); map_private_extent_buffer()
5438 size_t start_offset = eb->start & ((u64)PAGE_CACHE_SIZE - 1); map_private_extent_buffer()
5439 unsigned long i = (start_offset + start) >> PAGE_CACHE_SHIFT; map_private_extent_buffer()
5440 unsigned long end_i = (start_offset + start + min_len - 1) >> map_private_extent_buffer()
5454 if (start + min_len > eb->len) { map_private_extent_buffer()
5455 WARN(1, KERN_ERR "btrfs bad mapping eb start %llu len %lu, " map_private_extent_buffer()
5457 eb->start, eb->len, start, min_len); map_private_extent_buffer()
5469 unsigned long start, memcmp_extent_buffer()
5477 size_t start_offset = eb->start & ((u64)PAGE_CACHE_SIZE - 1); memcmp_extent_buffer()
5478 unsigned long i = (start_offset + start) >> PAGE_CACHE_SHIFT; memcmp_extent_buffer()
5481 WARN_ON(start > eb->len); memcmp_extent_buffer()
5482 WARN_ON(start + len > eb->start + eb->len); memcmp_extent_buffer()
5484 offset = (start_offset + start) & (PAGE_CACHE_SIZE - 1); memcmp_extent_buffer()
5505 unsigned long start, unsigned long len) write_extent_buffer()
5512 size_t start_offset = eb->start & ((u64)PAGE_CACHE_SIZE - 1); write_extent_buffer()
5513 unsigned long i = (start_offset + start) >> PAGE_CACHE_SHIFT; write_extent_buffer()
5515 WARN_ON(start > eb->len); write_extent_buffer()
5516 WARN_ON(start + len > eb->start + eb->len); write_extent_buffer()
5518 offset = (start_offset + start) & (PAGE_CACHE_SIZE - 1); write_extent_buffer()
5536 unsigned long start, unsigned long len) memset_extent_buffer()
5542 size_t start_offset = eb->start & ((u64)PAGE_CACHE_SIZE - 1); memset_extent_buffer()
5543 unsigned long i = (start_offset + start) >> PAGE_CACHE_SHIFT; memset_extent_buffer()
5545 WARN_ON(start > eb->len); memset_extent_buffer()
5546 WARN_ON(start + len > eb->start + eb->len); memset_extent_buffer()
5548 offset = (start_offset + start) & (PAGE_CACHE_SIZE - 1); memset_extent_buffer()
5573 size_t start_offset = dst->start & ((u64)PAGE_CACHE_SIZE - 1); copy_extent_buffer()
5631 size_t start_offset = dst->start & ((u64)PAGE_CACHE_SIZE - 1); memcpy_extent_buffer()
5679 size_t start_offset = dst->start & ((u64)PAGE_CACHE_SIZE - 1); memmove_extent_buffer()
87 __btrfs_debug_check_extent_io_range(const char *caller, struct extent_io_tree *tree, u64 start, u64 end) __btrfs_debug_check_extent_io_range() argument
445 insert_state(struct extent_io_tree *tree, struct extent_state *state, u64 start, u64 end, struct rb_node ***p, struct rb_node **parent, unsigned *bits, struct extent_changeset *changeset) insert_state() argument
1511 find_first_extent_bit_state(struct extent_io_tree *tree, u64 start, unsigned bits) find_first_extent_bit_state() argument
1592 find_delalloc_range(struct extent_io_tree *tree, u64 *start, u64 *end, u64 max_bytes, struct extent_state **cached_state) find_delalloc_range() argument
1646 __unlock_for_delalloc(struct inode *inode, struct page *locked_page, u64 start, u64 end) __unlock_for_delalloc() argument
1743 find_lock_delalloc_range(struct inode *inode, struct extent_io_tree *tree, struct page *locked_page, u64 *start, u64 *end, u64 max_bytes) find_lock_delalloc_range() argument
1880 count_range_bits(struct extent_io_tree *tree, u64 *start, u64 search_end, u64 max_bytes, unsigned bits, int contig) count_range_bits() argument
2465 bio_readpage_error(struct bio *failed_bio, u64 phy_offset, struct page *page, u64 start, u64 end, int failed_mirror) bio_readpage_error() argument
2941 __get_extent_map(struct inode *inode, struct page *page, size_t pg_offset, u64 start, u64 len, get_extent_t *get_extent, struct extent_map **em_cached) __get_extent_map() argument
3196 __do_contiguous_readpages(struct extent_io_tree *tree, struct page *pages[], int nr_pages, u64 start, u64 end, get_extent_t *get_extent, struct extent_map **em_cached, struct bio **bio, int mirror_num, unsigned long *bio_flags, int rw, u64 *prev_em_start) __do_contiguous_readpages() argument
4201 extent_write_locked_range(struct extent_io_tree *tree, struct inode *inode, u64 start, u64 end, get_extent_t *get_extent, int mode) extent_write_locked_range() argument
4464 extent_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, __u64 start, __u64 len, get_extent_t *get_extent) extent_fiemap() argument
4800 alloc_dummy_extent_buffer(struct btrfs_fs_info *fs_info, u64 start) alloc_dummy_extent_buffer() argument
4889 find_extent_buffer(struct btrfs_fs_info *fs_info, u64 start) find_extent_buffer() argument
4927 alloc_test_extent_buffer(struct btrfs_fs_info *fs_info, u64 start) alloc_test_extent_buffer() argument
4973 alloc_extent_buffer(struct btrfs_fs_info *fs_info, u64 start) alloc_extent_buffer() argument
5266 read_extent_buffer_pages(struct extent_io_tree *tree, struct extent_buffer *eb, u64 start, int wait, get_extent_t *get_extent, int mirror_num) read_extent_buffer_pages() argument
5362 read_extent_buffer(struct extent_buffer *eb, void *dstv, unsigned long start, unsigned long len) read_extent_buffer() argument
5393 read_extent_buffer_to_user(struct extent_buffer *eb, void __user *dstv, unsigned long start, unsigned long len) read_extent_buffer_to_user() argument
5468 memcmp_extent_buffer(struct extent_buffer *eb, const void *ptrv, unsigned long start, unsigned long len) memcmp_extent_buffer() argument
5504 write_extent_buffer(struct extent_buffer *eb, const void *srcv, unsigned long start, unsigned long len) write_extent_buffer() argument
5535 memset_extent_buffer(struct extent_buffer *eb, char c, unsigned long start, unsigned long len) memset_extent_buffer() argument
H A Dextent_map.c85 static u64 range_end(u64 start, u64 len) range_end() argument
87 if (start + len < start) range_end()
89 return start + len; range_end()
98 u64 end = range_end(em->start, em->len); tree_insert()
104 if (em->start < entry->start) tree_insert()
106 else if (em->start >= extent_map_end(entry)) tree_insert()
113 while (parent && em->start >= extent_map_end(entry)) { tree_insert()
118 if (end > entry->start && em->start < extent_map_end(entry)) tree_insert()
123 while (parent && em->start < entry->start) { tree_insert()
128 if (end > entry->start && em->start < extent_map_end(entry)) tree_insert()
155 if (offset < entry->start) __tree_search()
175 while (prev && offset < prev_entry->start) { __tree_search()
209 if (extent_map_end(prev) == next->start && mergable_maps()
230 if (em->start != 0) { try_merge_map()
235 em->start = merge->start; try_merge_map()
267 * @start: logical offset in the file
275 int unpin_extent_cache(struct extent_map_tree *tree, u64 start, u64 len, unpin_extent_cache() argument
283 em = lookup_extent_mapping(tree, start, len); unpin_extent_cache()
285 WARN_ON(!em || em->start != start); unpin_extent_cache()
292 em->mod_start = em->start; unpin_extent_cache()
303 em->mod_start = em->start; unpin_extent_cache()
326 em->mod_start = em->start; setup_extent_mapping()
361 u64 start, u64 len, int strict) __lookup_extent_mapping()
367 u64 end = range_end(start, len); __lookup_extent_mapping()
369 rb_node = __tree_search(&tree->map, start, &prev, &next); __lookup_extent_mapping()
381 if (strict && !(end > em->start && start < extent_map_end(em))) __lookup_extent_mapping()
391 * @start: byte offset to start the search
395 * [start, len] range. There may be additional objects in the tree that
400 u64 start, u64 len) lookup_extent_mapping()
402 return __lookup_extent_mapping(tree, start, len, 1); lookup_extent_mapping()
408 * @start: byte offset to start the search
412 * [start, len] range.
417 u64 start, u64 len) search_extent_mapping()
419 return __lookup_extent_mapping(tree, start, len, 0); search_extent_mapping()
360 __lookup_extent_mapping(struct extent_map_tree *tree, u64 start, u64 len, int strict) __lookup_extent_mapping() argument
399 lookup_extent_mapping(struct extent_map_tree *tree, u64 start, u64 len) lookup_extent_mapping() argument
416 search_extent_mapping(struct extent_map_tree *tree, u64 start, u64 len) search_extent_mapping() argument
H A Dextent_map.h24 u64 start; member in struct:extent_map
54 if (em->start + em->len < em->start) extent_map_end()
56 return em->start + em->len; extent_map_end()
68 u64 start, u64 len);
81 int unpin_extent_cache(struct extent_map_tree *tree, u64 start, u64 len, u64 gen);
84 u64 start, u64 len);
/linux-4.4.14/arch/unicore32/include/asm/
H A Dmemblock.h22 unsigned long start; member in struct:membank
37 #define bank_pfn_start(bank) __phys_to_pfn((bank)->start)
38 #define bank_pfn_end(bank) __phys_to_pfn((bank)->start + (bank)->size)
40 #define bank_phys_start(bank) ((bank)->start)
41 #define bank_phys_end(bank) ((bank)->start + (bank)->size)
H A Dcacheflush.h34 * start addresses should be rounded down, end addresses up.
55 * flush_user_range(start, end, flags)
59 * - start - user start address (inclusive, page aligned)
63 * coherent_kern_range(start, end)
66 * region described by start, end. If you have non-snooping
68 * - start - virtual start address
71 * coherent_user_range(start, end)
74 * region described by start, end. If you have non-snooping
76 * - start - virtual start address
88 * dma_flush_range(start, end)
91 * - start - virtual start address
141 unsigned long start, unsigned long end);
152 #define flush_cache_user_range(vma, start, end) \
153 __cpuc_coherent_user_range((start) & PAGE_MASK, PAGE_ALIGN(end))
165 #define clean_dcache_area(start, size) cpu_dcache_clean_area(start, size)
203 static inline void flush_cache_vmap(unsigned long start, unsigned long end) flush_cache_vmap() argument
207 static inline void flush_cache_vunmap(unsigned long start, unsigned long end) flush_cache_vunmap() argument
/linux-4.4.14/lib/
H A Dfind_bit.c32 unsigned long nbits, unsigned long start, unsigned long invert) _find_next_bit()
36 if (!nbits || start >= nbits) _find_next_bit()
39 tmp = addr[start / BITS_PER_LONG] ^ invert; _find_next_bit()
42 tmp &= BITMAP_FIRST_WORD_MASK(start); _find_next_bit()
43 start = round_down(start, BITS_PER_LONG); _find_next_bit()
46 start += BITS_PER_LONG; _find_next_bit()
47 if (start >= nbits) _find_next_bit()
50 tmp = addr[start / BITS_PER_LONG] ^ invert; _find_next_bit()
53 return min(start + __ffs(tmp), nbits); _find_next_bit()
150 unsigned long nbits, unsigned long start, unsigned long invert) _find_next_bit_le()
154 if (!nbits || start >= nbits) _find_next_bit_le()
157 tmp = addr[start / BITS_PER_LONG] ^ invert; _find_next_bit_le()
160 tmp &= ext2_swab(BITMAP_FIRST_WORD_MASK(start)); _find_next_bit_le()
161 start = round_down(start, BITS_PER_LONG); _find_next_bit_le()
164 start += BITS_PER_LONG; _find_next_bit_le()
165 if (start >= nbits) _find_next_bit_le()
168 tmp = addr[start / BITS_PER_LONG] ^ invert; _find_next_bit_le()
171 return min(start + __ffs(ext2_swab(tmp)), nbits); _find_next_bit_le()
31 _find_next_bit(const unsigned long *addr, unsigned long nbits, unsigned long start, unsigned long invert) _find_next_bit() argument
149 _find_next_bit_le(const unsigned long *addr, unsigned long nbits, unsigned long start, unsigned long invert) _find_next_bit_le() argument
H A Diommu-helper.c20 unsigned long start, unsigned int nr, iommu_area_alloc()
29 index = bitmap_find_next_zero_area(map, size, start, nr, align_mask); iommu_area_alloc()
33 start = index + 1; iommu_area_alloc()
19 iommu_area_alloc(unsigned long *map, unsigned long size, unsigned long start, unsigned int nr, unsigned long shift, unsigned long boundary_size, unsigned long align_mask) iommu_area_alloc() argument
H A Diommu-common.c58 unsigned int start, i; iommu_tbl_pool_init() local
70 start = 0; iommu_tbl_pool_init()
82 iommu->pools[i].start = start; iommu_tbl_pool_init()
83 iommu->pools[i].hint = start; iommu_tbl_pool_init()
84 start += iommu->poolsize; /* start for next pool */ iommu_tbl_pool_init()
85 iommu->pools[i].end = start - 1; iommu_tbl_pool_init()
91 p->start = start; iommu_tbl_pool_init()
92 p->hint = p->start; iommu_tbl_pool_init()
105 unsigned long n, end, start, limit, boundary_size; iommu_tbl_range_alloc() local
137 (*handle >= pool->start) && (*handle < pool->end)) iommu_tbl_range_alloc()
138 start = *handle; iommu_tbl_range_alloc()
140 start = pool->hint; iommu_tbl_range_alloc()
150 if (start >= limit) iommu_tbl_range_alloc()
151 start = pool->start; iommu_tbl_range_alloc()
157 * but on second pass, start at 0 in pool 0. iommu_tbl_range_alloc()
159 if ((start & mask) >= limit || pass > 0) { iommu_tbl_range_alloc()
163 start = pool->start; iommu_tbl_range_alloc()
165 start &= mask; iommu_tbl_range_alloc()
185 n = iommu_area_alloc(iommu->map, limit, start, npages, shift, iommu_tbl_range_alloc()
190 pool->hint = pool->start; iommu_tbl_range_alloc()
199 pool->hint = pool->start; iommu_tbl_range_alloc()
232 unsigned long largepool_start = tbl->large_pool.start; get_pool()
H A Dinterval_tree.c6 #define START(node) ((node)->start)
/linux-4.4.14/arch/arm/mach-imx/devices/
H A Dplatform-imx-dma.c16 .start = iobase, imx_add_imx_dma()
20 .start = irq, imx_add_imx_dma()
24 .start = irq_err, imx_add_imx_dma()
39 .start = iobase, imx_add_imx_sdma()
43 .start = irq, imx_add_imx_sdma()
H A Dplatform-gpio-mxc.c16 .start = iobase, mxc_register_gpio()
20 .start = irq, mxc_register_gpio()
24 .start = irq_high, mxc_register_gpio()
H A Dplatform-mx2-camera.c41 .start = data->iobasecsi, imx_add_mx2_camera()
45 .start = data->irqcsi, imx_add_mx2_camera()
49 .start = data->iobaseemmaprp, imx_add_mx2_camera()
53 .start = data->irqemmaprp, imx_add_mx2_camera()
/linux-4.4.14/arch/arm/mach-iop13xx/
H A Dtpmi.c45 .start = IOP13XX_TPMI_MMR(4), /* tpmi0 starts at dev == 4 */
50 .start = IOP13XX_TPMI_MEM(0),
55 .start = IOP13XX_TPMI_CTRL(0),
60 .start = IOP13XX_TPMI_IOP_CTRL(0),
65 .start = IRQ_IOP13XX_TPMI0_OUT,
73 .start = IOP13XX_TPMI_MMR(1),
78 .start = IOP13XX_TPMI_MEM(1),
83 .start = IOP13XX_TPMI_CTRL(1),
88 .start = IOP13XX_TPMI_IOP_CTRL(1),
93 .start = IRQ_IOP13XX_TPMI1_OUT,
101 .start = IOP13XX_TPMI_MMR(2),
106 .start = IOP13XX_TPMI_MEM(2),
111 .start = IOP13XX_TPMI_CTRL(2),
116 .start = IOP13XX_TPMI_IOP_CTRL(2),
121 .start = IRQ_IOP13XX_TPMI2_OUT,
129 .start = IOP13XX_TPMI_MMR(3),
134 .start = IOP13XX_TPMI_MEM(3),
139 .start = IOP13XX_TPMI_CTRL(3),
144 .start = IOP13XX_TPMI_IOP_CTRL(3),
149 .start = IRQ_IOP13XX_TPMI3_OUT,
/linux-4.4.14/drivers/staging/rtl8188eu/include/
H A Drtw_debug.h120 int proc_get_drv_version(char *page, char **start,
124 int proc_get_write_reg(char *page, char **start,
130 int proc_get_read_reg(char *page, char **start,
137 int proc_get_fwstate(char *page, char **start,
140 int proc_get_sec_info(char *page, char **start,
143 int proc_get_mlmext_state(char *page, char **start,
147 int proc_get_qos_option(char *page, char **start,
150 int proc_get_ht_option(char *page, char **start,
153 int proc_get_rf_info(char *page, char **start,
156 int proc_get_ap_info(char *page, char **start,
160 int proc_get_adapter_state(char *page, char **start,
164 int proc_get_trx_info(char *page, char **start,
168 int proc_get_mac_reg_dump1(char *page, char **start,
172 int proc_get_mac_reg_dump2(char *page, char **start,
176 int proc_get_mac_reg_dump3(char *page, char **start,
180 int proc_get_bb_reg_dump1(char *page, char **start,
184 int proc_get_bb_reg_dump2(char *page, char **start,
188 int proc_get_bb_reg_dump3(char *page, char **start,
192 int proc_get_rf_reg_dump1(char *page, char **start,
196 int proc_get_rf_reg_dump2(char *page, char **start,
200 int proc_get_rf_reg_dump3(char *page, char **start,
204 int proc_get_rf_reg_dump4(char *page, char **start,
210 int proc_get_all_sta_info(char *page, char **start,
216 int proc_get_best_channel(char *page, char **start,
220 int proc_get_rx_signal(char *page, char **start,
227 int proc_get_ht_enable(char *page, char **start,
234 int proc_get_cbw40_enable(char *page, char **start,
241 int proc_get_ampdu_enable(char *page, char **start,
248 int proc_get_rx_stbc(char *page, char **start,
255 int proc_get_two_path_rssi(char *page, char **start,
259 int proc_get_rssi_disp(char *page, char **start,
/linux-4.4.14/arch/x86/include/asm/trace/
H A Dmpx.h68 TP_PROTO(unsigned long start,
70 TP_ARGS(start, end),
73 __field(unsigned long, start)
78 __entry->start = start;
83 (void *)__entry->start,
89 TP_PROTO(unsigned long start, unsigned long end),
90 TP_ARGS(start, end)
94 TP_PROTO(unsigned long start, unsigned long end),
95 TP_ARGS(start, end)
/linux-4.4.14/arch/mips/ath79/
H A Ddev-wmac.c29 /* .start and .end fields are filled dynamically */
32 /* .start and .end fields are filled dynamically */
56 ath79_wmac_resources[0].start = AR913X_WMAC_BASE; ar913x_wmac_setup()
58 ath79_wmac_resources[1].start = ATH79_CPU_IRQ(2); ar913x_wmac_setup()
84 ath79_wmac_resources[0].start = AR933X_WMAC_BASE; ar933x_wmac_setup()
86 ath79_wmac_resources[1].start = ATH79_CPU_IRQ(2); ar933x_wmac_setup()
107 ath79_wmac_resources[0].start = AR934X_WMAC_BASE; ar934x_wmac_setup()
109 ath79_wmac_resources[1].start = ATH79_IP2_IRQ(1); ar934x_wmac_setup()
125 ath79_wmac_resources[0].start = QCA955X_WMAC_BASE; qca955x_wmac_setup()
127 ath79_wmac_resources[1].start = ATH79_IP2_IRQ(1); qca955x_wmac_setup()
H A Ddev-common.c29 .start = AR71XX_UART_BASE,
60 .start = AR933X_UART_BASE,
65 .start = ATH79_MISC_IRQ(3),
105 res.start = AR71XX_RESET_BASE + AR71XX_RESET_REG_WDOG_CTRL; ath79_register_wdt()
106 res.end = res.start + 0x8 - 1; ath79_register_wdt()
116 .start = AR71XX_GPIO_BASE,
120 .start = ATH79_MISC_IRQ(2),
/linux-4.4.14/arch/ia64/include/asm/
H A Dmeminit.h27 u64 start; /* virtual address of beginning of element */ member in struct:rsvd_region
37 extern int filter_rsvd_memory (u64 start, u64 end, void *arg);
38 extern int filter_memory (u64 start, u64 end, void *arg);
43 extern int reserve_elfcorehdr(u64 *start, u64 *end);
52 extern void call_pernode_memory (unsigned long start, unsigned long len, void *func);
54 # define call_pernode_memory(start, len, func) (*func)(start, len, 0)
59 extern int register_active_ranges(u64 start, u64 len, int nid);
65 extern int find_largest_hole(u64 start, u64 end, void *arg);
66 extern int create_mem_map_page_table(u64 start, u64 end, void *arg);
/linux-4.4.14/arch/blackfin/mach-bf548/boards/
H A Dezkit.c48 .start = 0x2C0C0000,
53 .start = IRQ_PG7,
94 .start = IRQ_EPPI0_ERR,
144 .start = IRQ_KEY,
176 .start = CNT_CONFIG,
181 .start = IRQ_CNT,
248 .start = UART0_DLL,
254 .start = PORTE_FER,
260 .start = IRQ_UART0_TX,
265 .start = IRQ_UART0_RX,
270 .start = IRQ_UART0_ERROR,
275 .start = CH_UART0_TX,
280 .start = CH_UART0_RX,
303 .start = UART1_DLL,
309 .start = PORTH_FER,
315 .start = IRQ_UART1_TX,
320 .start = IRQ_UART1_RX,
325 .start = IRQ_UART1_ERROR,
330 .start = CH_UART1_TX,
335 .start = CH_UART1_RX,
341 .start = GPIO_PE10,
346 .start = GPIO_PE9,
374 .start = UART2_DLL,
380 .start = PORTB_FER,
386 .start = IRQ_UART2_TX,
391 .start = IRQ_UART2_RX,
396 .start = IRQ_UART2_ERROR,
401 .start = CH_UART2_TX,
406 .start = CH_UART2_RX,
429 .start = UART3_DLL,
435 .start = PORTB_FER,
441 .start = IRQ_UART3_TX,
446 .start = IRQ_UART3_RX,
451 .start = IRQ_UART3_ERROR,
456 .start = CH_UART3_TX,
461 .start = CH_UART3_RX,
467 .start = GPIO_PB3,
472 .start = GPIO_PB2,
503 .start = 0xFFC00400,
508 .start = IRQ_UART0_RX,
513 .start = CH_UART0_RX,
528 .start = 0xFFC02000,
533 .start = IRQ_UART1_RX,
538 .start = CH_UART1_RX,
553 .start = 0xFFC02100,
558 .start = IRQ_UART2_RX,
563 .start = CH_UART2_RX,
578 .start = 0xFFC03100,
583 .start = IRQ_UART3_RX,
588 .start = CH_UART3_RX,
608 .start = 0x24000000,
613 .start = IRQ_PE8,
640 .start = 0xFFC03C00,
645 .start = IRQ_USB_INT0,
651 .start = IRQ_USB_DMA,
703 .start = SPORT0_TCR1,
708 .start = IRQ_SPORT0_RX,
713 .start = IRQ_SPORT0_ERROR,
737 .start = SPORT1_TCR1,
742 .start = IRQ_SPORT1_RX,
747 .start = IRQ_SPORT1_ERROR,
771 .start = SPORT2_TCR1,
776 .start = IRQ_SPORT2_RX,
781 .start = IRQ_SPORT2_ERROR,
805 .start = SPORT3_TCR1,
810 .start = IRQ_SPORT3_RX,
815 .start = IRQ_SPORT3_ERROR,
846 .start = 0xFFC02A00,
851 .start = IRQ_CAN0_RX,
856 .start = IRQ_CAN0_TX,
861 .start = IRQ_CAN0_ERROR,
883 .start = 0xFFC03200,
888 .start = IRQ_CAN1_RX,
893 .start = IRQ_CAN1_TX,
898 .start = IRQ_CAN1_ERROR,
919 .start = 0xFFC03800,
924 .start = IRQ_ATAPI_ERR,
966 .start = 0xFFC03B00,
971 .start = CH_NFC,
1037 .start = 0x20000000,
1109 .start = PINT0_MASK_SET,
1114 .start = IRQ_PINT0,
1129 .start = PINT1_MASK_SET,
1134 .start = IRQ_PINT1,
1149 .start = PINT2_MASK_SET,
1154 .start = IRQ_PINT2,
1169 .start = PINT3_MASK_SET,
1174 .start = IRQ_PINT3,
1189 .start = PORTA_FER,
1194 .start = IRQ_PA0,
1221 .start = PORTB_FER,
1226 .start = IRQ_PB0,
1253 .start = PORTC_FER,
1258 .start = IRQ_PC0,
1285 .start = PORTD_FER,
1290 .start = IRQ_PD0,
1317 .start = PORTE_FER,
1322 .start = IRQ_PE0,
1349 .start = PORTF_FER,
1354 .start = IRQ_PF0,
1381 .start = PORTG_FER,
1386 .start = IRQ_PG0,
1411 .start = PORTH_FER,
1416 .start = IRQ_PH0,
1441 .start = PORTI_FER,
1446 .start = IRQ_PI0,
1471 .start = PORTJ_FER,
1476 .start = IRQ_PJ0,
1556 .start = SPI0_REGBASE,
1561 .start = CH_SPI0,
1566 .start = IRQ_SPI0,
1575 .start = SPI1_REGBASE,
1580 .start = CH_SPI1,
1585 .start = IRQ_SPI1,
1695 .start = TWI0_REGBASE,
1700 .start = IRQ_TWI0,
1721 .start = TWI1_REGBASE,
1726 .start = IRQ_TWI1,
1863 .start = SPORT##x##_TCR1, \
1868 .start = CH_SPORT##x##_RX, \
1873 .start = CH_SPORT##x##_TX, \
1878 .start = IRQ_SPORT##x##_ERROR, \
H A Dcm_bf548.c55 .start = IRQ_EPPI0_ERR,
105 .start = IRQ_KEY,
133 .start = UART0_DLL,
138 .start = IRQ_UART0_TX,
143 .start = IRQ_UART0_RX,
148 .start = IRQ_UART0_ERROR,
153 .start = CH_UART0_TX,
158 .start = CH_UART0_RX,
181 .start = UART1_DLL,
186 .start = IRQ_UART1_TX,
191 .start = IRQ_UART1_RX,
196 .start = IRQ_UART1_ERROR,
201 .start = CH_UART1_TX,
206 .start = CH_UART1_RX,
212 .start = GPIO_PE10,
217 .start = GPIO_PE9,
245 .start = UART2_DLL,
250 .start = IRQ_UART2_TX,
255 .start = IRQ_UART2_RX,
260 .start = IRQ_UART2_ERROR,
265 .start = CH_UART2_TX,
270 .start = CH_UART2_RX,
293 .start = UART3_DLL,
298 .start = IRQ_UART3_TX,
303 .start = IRQ_UART3_RX,
308 .start = IRQ_UART3_ERROR,
313 .start = CH_UART3_TX,
318 .start = CH_UART3_RX,
324 .start = GPIO_PB3,
329 .start = GPIO_PB2,
360 .start = 0xFFC00400,
365 .start = IRQ_UART0_RX,
370 .start = CH_UART0_RX,
385 .start = 0xFFC02000,
390 .start = IRQ_UART1_RX,
395 .start = CH_UART1_RX,
410 .start = 0xFFC02100,
415 .start = IRQ_UART2_RX,
420 .start = CH_UART2_RX,
435 .start = 0xFFC03100,
440 .start = IRQ_UART3_RX,
445 .start = CH_UART3_RX,
465 .start = 0x24000000,
470 .start = IRQ_PE6,
497 .start = 0xFFC03C00,
502 .start = IRQ_USB_INT0,
508 .start = IRQ_USB_DMA,
560 .start = SPORT0_TCR1,
565 .start = IRQ_SPORT0_RX,
570 .start = IRQ_SPORT0_ERROR,
594 .start = SPORT1_TCR1,
599 .start = IRQ_SPORT1_RX,
604 .start = IRQ_SPORT1_ERROR,
628 .start = SPORT2_TCR1,
633 .start = IRQ_SPORT2_RX,
638 .start = IRQ_SPORT2_ERROR,
662 .start = SPORT3_TCR1,
667 .start = IRQ_SPORT3_RX,
672 .start = IRQ_SPORT3_ERROR,
698 .start = 0xFFC03800,
703 .start = IRQ_ATAPI_ERR,
741 .start = 0xFFC03B00,
746 .start = CH_NFC,
786 .start = 0xFFC02A00,
791 .start = IRQ_CAN0_RX,
796 .start = IRQ_CAN0_TX,
801 .start = IRQ_CAN0_ERROR,
841 .start = 0x20000000,
938 .start = SPI0_REGBASE,
943 .start = CH_SPI0,
948 .start = IRQ_SPI0,
957 .start = SPI1_REGBASE,
962 .start = CH_SPI1,
967 .start = IRQ_SPI1,
1012 .start = TWI0_REGBASE,
1017 .start = IRQ_TWI0,
1038 .start = TWI1_REGBASE,
1043 .start = IRQ_TWI1,
/linux-4.4.14/arch/score/mm/
H A Dcache.c173 unsigned long start, unsigned long end) flush_cache_range()
185 pgdp = pgd_offset(mm, start); flush_cache_range()
186 pudp = pud_offset(pgdp, start); flush_cache_range()
187 pmdp = pmd_offset(pudp, start); flush_cache_range()
188 ptep = pte_offset(pmdp, start); flush_cache_range()
190 while (start <= end) { flush_cache_range()
192 pgdp = pgd_offset(mm, start); flush_cache_range()
193 pudp = pud_offset(pgdp, start); flush_cache_range()
194 pmdp = pmd_offset(pudp, start); flush_cache_range()
195 ptep = pte_offset(pmdp, start); flush_cache_range()
198 start = (start + PAGE_SIZE) & ~(PAGE_SIZE - 1); flush_cache_range()
201 tmpend = (start | (PAGE_SIZE-1)) > end ? flush_cache_range()
202 end : (start | (PAGE_SIZE-1)); flush_cache_range()
204 flush_dcache_range(start, tmpend); flush_cache_range()
206 flush_icache_range(start, tmpend); flush_cache_range()
207 start = (start + PAGE_SIZE) & ~(PAGE_SIZE - 1); flush_cache_range()
246 void flush_dcache_range(unsigned long start, unsigned long end) flush_dcache_range() argument
250 start = start & ~(L1_CACHE_BYTES - 1); flush_dcache_range()
252 size = end - start; flush_dcache_range()
260 : : "r" (start)); flush_dcache_range()
261 start += L1_CACHE_BYTES; flush_dcache_range()
265 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument
268 start = start & ~(L1_CACHE_BYTES - 1); flush_icache_range()
271 size = end - start; flush_icache_range()
277 : : "r" (start)); flush_icache_range()
278 start += L1_CACHE_BYTES; flush_icache_range()
172 flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_cache_range() argument
/linux-4.4.14/arch/hexagon/lib/
H A Dchecksum.c103 int i, start, mid, end, mask; do_csum() local
111 start = 0xF & (16-(((int) ptr) & 0xF)) ; do_csum()
113 start = start & mask ; do_csum()
115 mid = len - start; do_csum()
121 if (start & 1) do_csum()
123 ptr2 = (unsigned short *) &ptr[start & 1]; do_csum()
124 if (start & 2) do_csum()
126 ptr4 = (unsigned int *) &ptr[start & 3]; do_csum()
127 if (start & 4) { do_csum()
133 ptr8 = (u64 *) &ptr[start & 7]; do_csum()
134 if (start & 8) { do_csum()
140 ptr8_o = (u64 *) (ptr + start); do_csum()
141 ptr8_e = (u64 *) (ptr + start + 8); do_csum()
163 ptr4 = (unsigned int *) &ptr[start + (mid * 16) + (end & 8)]; do_csum()
170 ptr2 = (unsigned short *) &ptr[start + (mid * 16) + (end & 12)]; do_csum()
175 sum1 += (u64) ptr[start + (mid * 16) + (end & 14)]; do_csum()
177 ptr8 = (u64 *) &ptr[start + (mid * 16)]; do_csum()
189 if (start & 1) do_csum()
/linux-4.4.14/arch/frv/include/asm/
H A Dcacheflush.h24 #define flush_cache_range(mm, start, end) do {} while(0)
26 #define flush_cache_vmap(start, end) do {} while(0)
27 #define flush_cache_vunmap(start, end) do {} while(0)
35 extern void frv_dcache_writeback(unsigned long start, unsigned long size);
36 extern void frv_cache_invalidate(unsigned long start, unsigned long size);
37 extern void frv_icache_invalidate(unsigned long start, unsigned long size);
38 extern void frv_cache_wback_inv(unsigned long start, unsigned long size);
71 static inline void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument
73 frv_cache_wback_inv(start, end); flush_icache_range()
78 unsigned long start, unsigned long len);
81 unsigned long start, unsigned long len) flush_icache_user_range()
83 frv_cache_wback_inv(start, start + len); flush_icache_user_range()
80 flush_icache_user_range(struct vm_area_struct *vma, struct page *page, unsigned long start, unsigned long len) flush_icache_user_range() argument
H A Dtlbflush.h23 extern void asmlinkage __flush_tlb_page(unsigned long contextid, unsigned long start);
25 unsigned long start, unsigned long end);
42 #define flush_tlb_range(vma,start,end) \
45 __flush_tlb_range((vma)->vm_mm->context.id, start, end); \
59 #define flush_tlb_kernel_range(start, end) flush_tlb_all()
67 #define flush_tlb_range(mm,start,end) BUG()
68 #define flush_tlb_kernel_range(start, end) BUG()
H A Dtermios.h8 start=^Q stop=^S susp=^Z eol=\0
/linux-4.4.14/drivers/infiniband/hw/usnic/
H A Dusnic_uiom_interval_tree.c42 #define START(node) ((node)->start)
45 #define MAKE_NODE(node, start, end, ref_cnt, flags, err, err_out) \
47 node = usnic_uiom_interval_node_alloc(start, \
57 #define MAKE_NODE_AND_APPEND(node, start, end, ref_cnt, flags, err, \
60 MAKE_NODE(node, start, end, \
70 usnic_uiom_interval_node_alloc(long int start, long int last, int ref_cnt, usnic_uiom_interval_node_alloc() argument
78 interval->start = start; usnic_uiom_interval_node_alloc()
94 if (node_a->start < node_b->start) interval_cmp()
96 else if (node_a->start > node_b->start) interval_cmp()
103 find_intervals_intersection_sorted(struct rb_root *root, unsigned long start, find_intervals_intersection_sorted() argument
111 for (node = usnic_uiom_interval_tree_iter_first(root, start, last); find_intervals_intersection_sorted()
113 node = usnic_uiom_interval_tree_iter_next(node, start, last)) find_intervals_intersection_sorted()
119 int usnic_uiom_get_intervals_diff(unsigned long start, unsigned long last, usnic_uiom_get_intervals_diff() argument
126 long int pivot = start; usnic_uiom_get_intervals_diff()
131 find_intervals_intersection_sorted(root, start, last, usnic_uiom_get_intervals_diff()
135 if (pivot < interval->start) { usnic_uiom_get_intervals_diff()
136 MAKE_NODE_AND_APPEND(tmp, pivot, interval->start - 1, usnic_uiom_get_intervals_diff()
139 pivot = interval->start; usnic_uiom_get_intervals_diff()
143 * Invariant: Set [start, pivot] is either in diff_set or root, usnic_uiom_get_intervals_diff()
178 int usnic_uiom_insert_interval(struct rb_root *root, unsigned long start, usnic_uiom_insert_interval() argument
184 unsigned long lpivot = start; usnic_uiom_insert_interval()
189 find_intervals_intersection_sorted(root, start, last, usnic_uiom_insert_interval()
197 istart = interval->start; usnic_uiom_insert_interval()
249 void usnic_uiom_remove_interval(struct rb_root *root, unsigned long start, usnic_uiom_remove_interval() argument
254 for (interval = usnic_uiom_interval_tree_iter_first(root, start, last); usnic_uiom_remove_interval()
257 start, usnic_uiom_remove_interval()
H A Dusnic_uiom_interval_tree.h42 unsigned long start; member in struct:usnic_uiom_interval_node
57 unsigned long start,
61 unsigned long start, unsigned long last);
63 * Inserts {start...last} into {root}. If there are overlaps,
67 unsigned long start, unsigned long last,
70 * Removed {start...last} from {root}. The nodes removed are returned in
75 unsigned long start, unsigned long last,
78 * Returns {start...last} - {root} (relative complement of {start...last} in
81 int usnic_uiom_get_intervals_diff(unsigned long start,
/linux-4.4.14/arch/blackfin/mach-bf527/boards/
H A Dcm_bf527.c44 .start = 0x203C0000,
49 .start = IRQ_PF7,
78 .start = 0xffc03800,
83 .start = IRQ_USB_INT0,
89 .start = IRQ_USB_DMA,
161 .start = NFC_CTL,
166 .start = CH_NFC,
186 .start = 0x20310000, /* IO PORT */
190 .start = 0x20311000, /* Attribute Memory */
194 .start = IRQ_PF4,
198 .start = 6, /* Card Detect PF6 */
231 .start = 0x20300300,
236 .start = IRQ_PF7,
255 .start = 0x203FB800,
260 .start = 0x203FB804,
265 .start = IRQ_PF9,
315 .start = 0x20300000,
319 .start = IRQ_PF7,
455 .start = SPI0_REGBASE,
460 .start = CH_SPI,
465 .start = IRQ_SPI,
510 .start = 0x20000000,
514 .start = (unsigned long)cm_flash_gpios,
535 .start = UART0_THR,
540 .start = IRQ_UART0_TX,
545 .start = IRQ_UART0_RX,
550 .start = IRQ_UART0_ERROR,
555 .start = CH_UART0_TX,
560 .start = CH_UART0_RX,
583 .start = UART1_THR,
588 .start = IRQ_UART1_TX,
593 .start = IRQ_UART1_RX,
598 .start = IRQ_UART1_ERROR,
603 .start = CH_UART1_TX,
608 .start = CH_UART1_RX,
614 .start = GPIO_PF9,
619 .start = GPIO_PF10,
646 .start = 0xFFC00400,
651 .start = IRQ_UART0_RX,
656 .start = CH_UART0_RX,
672 .start = 0xFFC02000,
677 .start = IRQ_UART1_RX,
682 .start = CH_UART1_RX,
702 .start = TWI0_REGBASE,
707 .start = IRQ_TWI,
747 .start = SPORT0_TCR1,
752 .start = IRQ_SPORT0_RX,
757 .start = IRQ_SPORT0_ERROR,
781 .start = SPORT1_TCR1,
786 .start = IRQ_SPORT1_RX,
791 .start = IRQ_SPORT1_ERROR,
H A Dezkit.c49 .start = 0x203C0000,
54 .start = IRQ_PF7,
83 .start = 0xffc03800,
88 .start = IRQ_USB_INT0,
94 .start = IRQ_USB_DMA,
146 .start = IRQ_PPI_ERROR,
170 .start = IRQ_PPI_ERROR,
211 .start = 0x20000000,
255 .start = NFC_CTL,
260 .start = CH_NFC,
280 .start = 0x20310000, /* IO PORT */
284 .start = 0x20311000, /* Attribute Memory */
288 .start = IRQ_PF4,
292 .start = 6, /* Card Detect PF6 */
325 .start = 0x20300300,
330 .start = IRQ_PF7,
349 .start = 0x203FB800,
354 .start = 0x203FB800 + 4,
359 .start = IRQ_PF9,
409 .start = 0x20300000,
413 .start = 1,
416 .start = IRQ_PF7,
516 .start = SPORT##x##_TCR1, \
521 .start = CH_SPORT##x##_RX, \
526 .start = CH_SPORT##x##_TX, \
531 .start = IRQ_SPORT##x##_ERROR, \
669 .start = SPI0_REGBASE,
674 .start = CH_SPI,
679 .start = IRQ_SPI,
700 .start = UART0_THR,
705 .start = IRQ_UART0_TX,
710 .start = IRQ_UART0_RX,
715 .start = IRQ_UART0_ERROR,
720 .start = CH_UART0_TX,
725 .start = CH_UART0_RX,
748 .start = UART1_THR,
753 .start = IRQ_UART1_TX,
758 .start = IRQ_UART1_RX,
763 .start = IRQ_UART1_ERROR,
768 .start = CH_UART1_TX,
773 .start = CH_UART1_RX,
779 .start = GPIO_PF9,
784 .start = GPIO_PF10,
811 .start = 0xFFC00400,
816 .start = IRQ_UART0_RX,
821 .start = CH_UART0_RX,
837 .start = 0xFFC02000,
842 .start = IRQ_UART1_RX,
847 .start = CH_UART1_RX,
867 .start = TWI0_REGBASE,
872 .start = IRQ_TWI,
1006 .start = SPORT0_TCR1,
1011 .start = IRQ_SPORT0_RX,
1016 .start = IRQ_SPORT0_ERROR,
1040 .start = SPORT1_TCR1,
1045 .start = IRQ_SPORT1_RX,
1050 .start = IRQ_SPORT1_ERROR,
1117 .start = CNT_CONFIG,
1122 .start = IRQ_CNT,
H A Dad7160eval.c43 .start = 0xffc03800,
48 .start = IRQ_USB_INT0,
53 .start = IRQ_USB_DMA,
103 .start = IRQ_PPI_ERROR,
141 .start = 0x20000000,
181 .start = NFC_CTL,
186 .start = CH_NFC,
337 .start = SPI0_REGBASE,
342 .start = CH_SPI,
347 .start = IRQ_SPI,
368 .start = UART0_THR,
373 .start = IRQ_UART0_TX,
378 .start = IRQ_UART0_RX,
383 .start = IRQ_UART0_ERROR,
388 .start = CH_UART0_TX,
393 .start = CH_UART0_RX,
416 .start = UART1_THR,
421 .start = IRQ_UART1_TX,
426 .start = IRQ_UART1_RX,
431 .start = IRQ_UART1_ERROR,
436 .start = CH_UART1_TX,
441 .start = CH_UART1_RX,
447 .start = GPIO_PF9,
452 .start = GPIO_PF10,
479 .start = 0xFFC00400,
484 .start = IRQ_UART0_RX,
489 .start = CH_UART0_RX,
505 .start = 0xFFC02000,
510 .start = IRQ_UART1_RX,
515 .start = CH_UART1_RX,
565 .start = TWI0_REGBASE,
570 .start = IRQ_TWI,
601 .start = SPORT0_TCR1,
606 .start = IRQ_SPORT0_RX,
611 .start = IRQ_SPORT0_ERROR,
635 .start = SPORT1_TCR1,
640 .start = IRQ_SPORT1_RX,
645 .start = IRQ_SPORT1_ERROR,
691 .start = CNT_CONFIG,
696 .start = IRQ_CNT,
/linux-4.4.14/arch/alpha/kernel/
H A Dpci.c76 dev->resource[2].start = dev->resource[3].start = 0; quirk_cypress()
80 dev->resource[0].start = 0x170; quirk_cypress()
82 dev->resource[1].start = 0x376; quirk_cypress()
131 resource_size_t start = res->start; pcibios_align_resource() local
134 /* Make sure we start at our min on all hoses */ pcibios_align_resource()
135 if (start - hose->io_space->start < PCIBIOS_MIN_IO) pcibios_align_resource()
136 start = PCIBIOS_MIN_IO + hose->io_space->start; pcibios_align_resource()
141 if (start & 0x300) pcibios_align_resource()
142 start = (start + 0x3ff) & ~0x3ff; pcibios_align_resource()
145 /* Make sure we start at our min on all hoses */ pcibios_align_resource()
146 if (start - hose->mem_space->start < PCIBIOS_MIN_MEM) pcibios_align_resource()
147 start = PCIBIOS_MIN_MEM + hose->mem_space->start; pcibios_align_resource()
167 start = ALIGN(start, alignto); pcibios_align_resource()
169 if (((start / (16*MB)) & 0x7) == 0) { pcibios_align_resource()
170 start &= ~(128*MB - 1); pcibios_align_resource()
171 start += 16*MB; pcibios_align_resource()
172 start = ALIGN(start, alignto); pcibios_align_resource()
174 if (start/(128*MB) != (start + size - 1)/(128*MB)) { pcibios_align_resource()
175 start &= ~(128*MB - 1); pcibios_align_resource()
176 start += (128 + 16)*MB; pcibios_align_resource()
177 start = ALIGN(start, alignto); pcibios_align_resource()
182 return start; pcibios_align_resource()
285 if (r->parent || !r->start || !r->flags) pcibios_claim_one_bus()
329 end = hose->mem_space->start + pci_mem_end; common_init_pci()
335 hose->io_space->start); common_init_pci()
337 hose->mem_space->start); common_init_pci()
/linux-4.4.14/arch/blackfin/mach-bf533/boards/
H A Dip0x.c41 .start = 0x20100000,
45 .start = 0x20100000 + 2,
49 .start = IRQ_PF15,
57 .start = 0x20200000,
61 .start = 0x20200000 + 2,
65 .start = IRQ_PF14,
149 .start = BFIN_UART_THR,
154 .start = IRQ_UART0_TX,
159 .start = IRQ_UART0_RX,
164 .start = IRQ_UART0_ERROR,
169 .start = CH_UART0_TX,
174 .start = CH_UART0_RX,
200 .start = 0xFFC00400,
205 .start = IRQ_UART0_RX,
210 .start = CH_UART0_RX,
228 .start = 0x20300000,
232 .start = 0x20300000 + 2,
236 .start = IRQ_PF11,
H A Dcm_bf533.c109 .start = SPI0_REGBASE,
114 .start = CH_SPI,
119 .start = IRQ_SPI,
161 .start = 0x20200300,
165 .start = IRQ_PF0,
187 .start = 0x20308000,
191 .start = IRQ_PF8,
219 .start = BFIN_UART_THR,
224 .start = IRQ_UART0_TX,
229 .start = IRQ_UART0_RX,
234 .start = IRQ_UART0_ERROR,
239 .start = CH_UART0_TX,
244 .start = CH_UART0_RX,
270 .start = 0xFFC00400,
275 .start = IRQ_UART0_RX,
280 .start = CH_UART0_RX,
299 .start = SPORT0_TCR1,
304 .start = IRQ_SPORT0_RX,
309 .start = IRQ_SPORT0_ERROR,
333 .start = SPORT1_TCR1,
338 .start = IRQ_SPORT1_RX,
343 .start = IRQ_SPORT1_ERROR,
369 .start = 0x20308000,
373 .start = 0x20308004,
377 .start = IRQ_PF4,
409 .start = 0x20300000,
413 .start = IRQ_PF6,
449 .start = 0x20000000,
H A DH8606.c45 .start = 0x20300000,
50 .start = 0x20300004,
55 .start = IRQ_PF10,
82 .start = 0x20300300,
86 .start = IRQ_PROG_INTB,
90 .start = IRQ_PF7,
110 .start = 0x20300000,
114 .start = IRQ_PF10,
197 .start = SPI0_REGBASE,
202 .start = CH_SPI,
207 .start = IRQ_SPI,
236 .start = BFIN_UART_THR,
241 .start = IRQ_UART0_TX,
246 .start = IRQ_UART0_RX,
251 .start = IRQ_UART0_ERROR,
256 .start = CH_UART0_TX,
261 .start = CH_UART0_RX,
287 .start = 0xFFC00400,
292 .start = IRQ_UART0_RX,
297 .start = CH_UART0_RX,
365 .start = 0x20200030,
370 .start = IRQ_PF9,
/linux-4.4.14/arch/x86/boot/compressed/
H A Daslr.c108 unsigned long start; member in struct:mem_vector
118 if (item->start < region->start) mem_contains()
121 if (item->start + item->size > region->start + region->size) mem_contains()
129 if (one->start + one->size <= two->start) mem_overlaps()
132 if (one->start >= two->start + two->size) mem_overlaps()
151 mem_avoid[0].start = unsafe; mem_avoid_init()
159 mem_avoid[1].start = initrd_start; mem_avoid_init()
169 mem_avoid[2].start = cmd_line; mem_avoid_init()
173 mem_avoid[3].start = (unsigned long)free_mem_ptr; mem_avoid_init()
177 mem_avoid[4].start = (unsigned long)free_mem_end_ptr; mem_avoid_init()
197 avoid.start = (unsigned long)ptr; mem_avoid_overlap()
250 region.start = entry->addr; process_e820_entry()
254 if (region.start < minimum) process_e820_entry()
255 region.start = minimum; process_e820_entry()
258 region.start = ALIGN(region.start, CONFIG_PHYSICAL_ALIGN); process_e820_entry()
261 if (region.start > entry->addr + entry->size) process_e820_entry()
265 region.size -= region.start - entry->addr; process_e820_entry()
268 if (region.start + region.size > CONFIG_RANDOMIZE_BASE_MAX_OFFSET) process_e820_entry()
269 region.size = CONFIG_RANDOMIZE_BASE_MAX_OFFSET - region.start; process_e820_entry()
272 for (img.start = region.start, img.size = image_size ; process_e820_entry()
274 img.start += CONFIG_PHYSICAL_ALIGN) { process_e820_entry()
277 slots_append(img.start); process_e820_entry()
/linux-4.4.14/drivers/pnp/
H A Dsystem.c29 resource_size_t start = r->start, end = r->end; reserve_range() local
38 res = request_region(start, end - start + 1, regionid); reserve_range()
40 res = request_mem_region(start, end - start + 1, regionid); reserve_range()
63 if (res->start == 0) reserve_resources_of_dev()
65 if (res->start < 0x100) reserve_resources_of_dev()
75 if (res->end < res->start) reserve_resources_of_dev()
H A Dmanager.c44 "flags %#lx\n", idx, (unsigned long long) res->start, pnp_assign_port()
51 res->start = 0; pnp_assign_port()
60 res->start = rule->min; pnp_assign_port()
61 res->end = res->start + rule->size - 1; pnp_assign_port()
64 res->start += rule->align; pnp_assign_port()
65 res->end = res->start + rule->size - 1; pnp_assign_port()
66 if (res->start > rule->max || !rule->align) { pnp_assign_port()
76 pnp_add_io_resource(dev, res->start, res->end, res->flags); pnp_assign_port()
87 "flags %#lx\n", idx, (unsigned long long) res->start, pnp_assign_mem()
94 res->start = 0; pnp_assign_mem()
111 res->start = rule->min; pnp_assign_mem()
112 res->end = res->start + rule->size - 1; pnp_assign_mem()
115 res->start += rule->align; pnp_assign_mem()
116 res->end = res->start + rule->size - 1; pnp_assign_mem()
117 if (res->start > rule->max || !rule->align) { pnp_assign_mem()
127 pnp_add_mem_resource(dev, res->start, res->end, res->flags); pnp_assign_mem()
144 idx, (int) res->start, res->flags); pnp_assign_irq()
150 res->start = -1; pnp_assign_irq()
160 res->start = find_next_bit(rule->map.bits, PNP_IRQ_NR, 16); pnp_assign_irq()
161 if (res->start < PNP_IRQ_NR) { pnp_assign_irq()
162 res->end = res->start; pnp_assign_irq()
167 res->start = res->end = xtab[i]; pnp_assign_irq()
174 res->start = -1; pnp_assign_irq()
185 pnp_add_irq_resource(dev, res->start, res->flags); pnp_assign_irq()
203 idx, (int) res->start, res->flags); pnp_assign_dma()
209 res->start = -1; pnp_assign_dma()
220 res->start = res->end = xtab[i]; pnp_assign_dma()
230 pnp_add_dma_resource(dev, res->start, res->flags); pnp_assign_dma()
331 * pnp_start_dev - low-level start of the PnP device
/linux-4.4.14/arch/m68k/kernel/
H A Dpcibios.c36 resource_size_t start = res->start; pcibios_align_resource() local
38 if ((res->flags & IORESOURCE_IO) && (start & 0x300)) pcibios_align_resource()
39 start = (start + 0x3ff) & ~0x3ff; pcibios_align_resource()
41 start = (start + align - 1) & ~(align - 1); pcibios_align_resource()
43 return start; pcibios_align_resource()
64 if (!r->start && r->end) { pcibios_enable_device()
/linux-4.4.14/arch/parisc/math-emu/
H A Dfpbits.h53 #define Bitfield_extract(start, length, object) \
54 ((object) >> (HOSTWDSZ - (start) - (length)) & \
57 #define Bitfield_signed_extract(start, length, object) \
58 ((int)((object) << start) >> (HOSTWDSZ - (length)))
60 #define Bitfield_mask(start, len, object) \
61 ((object) & (((unsigned)-1 >> (HOSTWDSZ-len)) << (HOSTWDSZ-start-len)))
63 #define Bitfield_deposit(value,start,len,object) object = \
64 ((object) & ~(((unsigned)-1 >> (HOSTWDSZ-len)) << (HOSTWDSZ-start-len))) | \
65 (((value) & ((unsigned)-1 >> (HOSTWDSZ-len))) << (HOSTWDSZ-start-len))
/linux-4.4.14/arch/m32r/include/asm/
H A Dcacheflush.h13 #define flush_cache_range(vma, start, end) do { } while (0)
20 #define flush_icache_range(start, end) _flush_cache_copyback_all()
26 #define flush_icache_range(start, end) smp_flush_cache_all()
35 #define flush_cache_range(vma, start, end) do { } while (0)
41 #define flush_icache_range(start, end) _flush_cache_all()
49 #define flush_cache_range(vma, start, end) do { } while (0)
55 #define flush_icache_range(start, end) do { } while (0)
61 #define flush_cache_vmap(start, end) do { } while (0)
62 #define flush_cache_vunmap(start, end) do { } while (0)
H A Dtlbflush.h13 * - flush_tlb_range(vma, start, end) flushes a range of pages
14 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages
28 #define flush_tlb_range(vma, start, end) \
29 local_flush_tlb_range(vma, start, end)
30 #define flush_tlb_kernel_range(start, end) local_flush_tlb_all()
35 #define flush_tlb_range(vma, start, end) do { } while (0)
47 #define flush_tlb_range(vma, start, end) \
48 smp_flush_tlb_range(vma, start, end)
49 #define flush_tlb_kernel_range(start, end) smp_flush_tlb_all()
/linux-4.4.14/arch/powerpc/lib/
H A Drheap.c138 blk->start = 0; get_slot()
163 s = blkn->start; attach_free_block()
175 bs = blk->start; attach_free_block()
193 if (before && s != (before->start + before->size)) attach_free_block()
196 if (after && e != after->start) attach_free_block()
221 after->start -= size; attach_free_block()
240 if (blk->start > blkn->start) { attach_taken_block()
329 int rh_attach_region(rh_info_t * info, unsigned long start, int size) rh_attach_region() argument
336 s = start; rh_attach_region()
340 /* Round start up */ rh_attach_region()
350 start = s; rh_attach_region()
359 blk->start = start; rh_attach_region()
370 unsigned long rh_detach_region(rh_info_t * info, unsigned long start, int size) rh_detach_region() argument
381 s = start; rh_detach_region()
385 /* Round start up */ rh_detach_region()
398 bs = blk->start; rh_detach_region()
399 be = blk->start + blk->size; rh_detach_region()
416 /* blk still in free list, with updated start and/or size */ rh_detach_region()
419 blk->start += size; rh_detach_region()
428 newblk->start = e; rh_detach_region()
447 unsigned long start, sp_size; rh_alloc_align() local
463 start = (blk->start + alignment - 1) & ~(alignment - 1); rh_alloc_align()
464 if (start + size <= blk->start + blk->size) rh_alloc_align()
481 sp_size = start - blk->start; rh_alloc_align()
486 spblk->start = blk->start; rh_alloc_align()
492 newblk->start = start; rh_alloc_align()
495 /* blk still in free list, with updated start and size rh_alloc_align()
497 blk->start = start + size; rh_alloc_align()
509 return start; rh_alloc_align()
527 unsigned long rh_alloc_fixed(rh_info_t * info, unsigned long start, int size, const char *owner) rh_alloc_fixed() argument
538 s = start; rh_alloc_fixed()
542 /* Round start up */ rh_alloc_fixed()
555 bs = blk->start; rh_alloc_fixed()
556 be = blk->start + blk->size; rh_alloc_fixed()
571 start = blk->start; rh_alloc_fixed()
574 return start; rh_alloc_fixed()
578 /* blk still in free list, with updated start and/or size */ rh_alloc_fixed()
581 blk->start += size; rh_alloc_fixed()
590 newblk2->start = e; rh_alloc_fixed()
597 newblk1->start = s; rh_alloc_fixed()
601 start = newblk1->start; rh_alloc_fixed()
604 return start; rh_alloc_fixed()
612 int rh_free(rh_info_t * info, unsigned long start) rh_free() argument
622 if (start < blk2->start) rh_free()
627 if (blk == NULL || start > (blk->start + blk->size)) rh_free()
667 stats->start = blk->start; list_for_each()
679 int rh_set_owner(rh_info_t * info, unsigned long start, const char *owner) rh_set_owner() argument
689 if (start < blk2->start) rh_set_owner()
694 if (blk == NULL || start > (blk->start + blk->size)) rh_set_owner()
723 st[i].start, st[i].start + st[i].size, rh_dump()
734 st[i].start, st[i].start + st[i].size, rh_dump()
744 blk, blk->start, blk->start + blk->size, blk->size); rh_dump_blk()
/linux-4.4.14/arch/um/include/asm/
H A Dtlb.h20 unsigned long start; member in struct:mmu_gather
28 if (tlb->start > address) __tlb_remove_tlb_entry()
29 tlb->start = address; __tlb_remove_tlb_entry()
38 tlb->start = TASK_SIZE; init_tlb_gather()
42 tlb->start = 0; init_tlb_gather()
48 tlb_gather_mmu(struct mmu_gather *tlb, struct mm_struct *mm, unsigned long start, unsigned long end) tlb_gather_mmu() argument
51 tlb->start = start; tlb_gather_mmu()
53 tlb->fullmm = !(start | (end+1)); tlb_gather_mmu()
58 extern void flush_tlb_mm_range(struct mm_struct *mm, unsigned long start,
64 flush_tlb_mm_range(tlb->mm, tlb->start, tlb->end); tlb_flush_mmu_tlbonly()
88 tlb_finish_mmu(struct mmu_gather *tlb, unsigned long start, unsigned long end) tlb_finish_mmu() argument
H A Dtlbflush.h19 * - flush_tlb_range(vma, start, end) flushes a range of pages
24 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
28 extern void flush_tlb_kernel_range(unsigned long start, unsigned long end);
/linux-4.4.14/arch/x86/pci/
H A Dbus_numa.c14 if (info->busn.start == bus) x86_find_pci_root_info()
83 info->busn.start = bus_min; alloc_pci_root_info()
94 void update_res(struct pci_root_info *info, resource_size_t start, update_res() argument
100 if (start > end) update_res()
103 if (start == MAX_RESOURCE) update_res()
118 common_start = max(res->start, start); update_res()
123 final_start = min(res->start, start); update_res()
126 res->start = final_start; update_res()
141 res->start = start; update_res()
H A Damd_bus.c70 u64 start; early_root_info_init() local
155 start = reg & 0xfff000; early_root_info_init()
166 node, link, start, end); early_root_info_init()
171 update_res(info, start, end, IORESOURCE_IO, 1); early_root_info_init()
172 subtract_range(range, RANGE_NUM, start, end + 1); early_root_info_init()
182 update_res(info, range[i].start, range[i].end - 1, early_root_info_init()
206 fam10h_mmconf_start = fam10h_mmconf->start; early_root_info_init()
221 start = reg & 0xffffff00; /* 39:16 on 31:8*/ early_root_info_init()
222 start <<= 8; early_root_info_init()
236 node, link, start, end); early_root_info_init()
239 * mmconf range, so need to update start and end. early_root_info_init()
244 if (start >= fam10h_mmconf_start && early_root_info_init()
245 start <= fam10h_mmconf_end) { early_root_info_init()
246 start = fam10h_mmconf_end + 1; early_root_info_init()
256 if (start < fam10h_mmconf_start && early_root_info_init()
260 update_res(info, start, endx, IORESOURCE_MEM, 0); early_root_info_init()
261 subtract_range(range, RANGE_NUM, start, early_root_info_init()
263 printk(KERN_CONT " ==> [%llx, %llx]", start, endx); early_root_info_init()
264 start = fam10h_mmconf_end + 1; early_root_info_init()
268 if (start <= end) { early_root_info_init()
269 printk(KERN_CONT " %s [%llx, %llx]", endx ? "and" : "==>", start, end); early_root_info_init()
277 update_res(info, cap_resource(start), cap_resource(end), early_root_info_init()
279 subtract_range(range, RANGE_NUM, start, end + 1); early_root_info_init()
307 update_res(info, cap_resource(range[i].start), early_root_info_init()
317 busnum = info->busn.start; early_root_info_init()
/linux-4.4.14/arch/xtensa/include/asm/
H A Dsysmem.h17 unsigned long start; member in struct:meminfo
22 * Bank array is sorted by .start.
33 int add_sysmem_bank(unsigned long start, unsigned long end);
/linux-4.4.14/arch/arm/plat-iop/
H A Dadma.c54 .start = IOP3XX_DMA_PHYS_BASE(0),
59 .start = IRQ_DMA0_EOT,
64 .start = IRQ_DMA0_EOC,
69 .start = IRQ_DMA0_ERR,
77 .start = IOP3XX_DMA_PHYS_BASE(1),
82 .start = IRQ_DMA1_EOT,
87 .start = IRQ_DMA1_EOC,
92 .start = IRQ_DMA1_ERR,
101 .start = IOP3XX_AAU_PHYS_BASE,
106 .start = IRQ_AA_EOT,
111 .start = IRQ_AA_EOC,
116 .start = IRQ_AA_ERR,
/linux-4.4.14/arch/arm/mach-ks8695/
H A Ddevices.c42 .start = KS8695_WAN_PA,
48 .start = KS8695_IRQ_WAN_RX_STATUS,
54 .start = KS8695_IRQ_WAN_TX_STATUS,
60 .start = KS8695_IRQ_WAN_LINK,
66 .start = KS8695_MISC_PA,
86 .start = KS8695_LAN_PA,
92 .start = KS8695_IRQ_LAN_RX_STATUS,
98 .start = KS8695_IRQ_LAN_TX_STATUS,
104 .start = KS8695_SWITCH_PA,
124 .start = KS8695_HPNA_PA,
130 .start = KS8695_IRQ_HPNA_RX_STATUS,
136 .start = KS8695_IRQ_HPNA_TX_STATUS,
/linux-4.4.14/arch/blackfin/mach-bf609/boards/
H A Dezkit.c46 .start = 0x2C0C0000,
51 .start = IRQ_PG7,
91 .start = CNT_CONFIG,
96 .start = IRQ_CNT,
149 .start = EMAC0_MACCFG,
155 .start = IRQ_EMAC0_STAT,
217 .start = UART0_REVID,
223 .start = PORTD_FER,
228 .start = PORTD_MUX,
234 .start = IRQ_UART0_TX,
239 .start = IRQ_UART0_RX,
244 .start = IRQ_UART0_STAT,
249 .start = CH_UART0_TX,
254 .start = CH_UART0_RX,
260 .start = GPIO_PD10,
265 .start = GPIO_PD9,
293 .start = UART1_REVID,
299 .start = PORTG_FER_SET,
305 .start = IRQ_UART1_TX,
310 .start = IRQ_UART1_RX,
315 .start = IRQ_UART1_STAT,
320 .start = CH_UART1_TX,
325 .start = CH_UART1_RX,
331 .start = GPIO_PG13,
336 .start = GPIO_PG10,
367 .start = 0xFFC00400,
372 .start = IRQ_UART0_TX,
377 .start = CH_UART0_TX,
392 .start = 0xFFC02000,
397 .start = IRQ_UART1_TX,
402 .start = CH_UART1_TX,
419 .start = 0xFFCC1000,
424 .start = IRQ_USB_STAT,
430 .start = IRQ_USB_DMA,
476 .start = SPORT0_TCR1,
481 .start = IRQ_SPORT0_RX,
486 .start = IRQ_SPORT0_ERROR,
510 .start = SPORT1_TCR1,
515 .start = IRQ_SPORT1_RX,
520 .start = IRQ_SPORT1_ERROR,
544 .start = SPORT2_TCR1,
549 .start = IRQ_SPORT2_RX,
554 .start = IRQ_SPORT2_ERROR,
585 .start = 0xFFC00A00,
590 .start = IRQ_CAN0_RX,
595 .start = IRQ_CAN0_TX,
600 .start = IRQ_CAN0_STAT,
646 .start = 0xFFC03B00,
651 .start = CH_NFC,
730 .start = 0xb0000000,
794 .start = SPORT0_CTL_A,
799 .start = SPORT0_CTL_B,
804 .start = CH_SPORT0_TX,
809 .start = CH_SPORT0_RX,
814 .start = IRQ_SPORT0_TX_STAT,
819 .start = IRQ_SPORT0_RX_STAT,
1214 .start = EPPI2_STAT,
1219 .start = CH_EPPI2_CH0,
1224 .start = IRQ_EPPI2_STAT,
1244 .start = REG_CRC0_CTL,
1249 .start = IRQ_CRC0_DCNTEXP,
1254 .start = CH_MEM_STREAM0_SRC_CRC0,
1259 .start = CH_MEM_STREAM0_DEST_CRC0,
1274 .start = REG_CRC1_CTL,
1279 .start = IRQ_CRC1_DCNTEXP,
1284 .start = CH_MEM_STREAM1_SRC_CRC1,
1289 .start = CH_MEM_STREAM1_DEST_CRC1,
1309 .start = REG_CRC0_CTL,
1314 .start = IRQ_CRC0_DCNTEXP,
1319 .start = CH_MEM_STREAM0_SRC_CRC0,
1365 .start = PINT0_MASK_SET,
1370 .start = IRQ_PINT0,
1385 .start = PINT1_MASK_SET,
1390 .start = IRQ_PINT1,
1405 .start = PINT2_MASK_SET,
1410 .start = IRQ_PINT2,
1425 .start = PINT3_MASK_SET,
1430 .start = IRQ_PINT3,
1445 .start = PINT4_MASK_SET,
1450 .start = IRQ_PINT4,
1465 .start = PINT5_MASK_SET,
1470 .start = IRQ_PINT5,
1485 .start = PORTA_FER,
1490 .start = IRQ_PA0,
1516 .start = PORTB_FER,
1521 .start = IRQ_PB0,
1547 .start = PORTC_FER,
1552 .start = IRQ_PC0,
1578 .start = PORTD_FER,
1583 .start = IRQ_PD0,
1609 .start = PORTE_FER,
1614 .start = IRQ_PE0,
1640 .start = PORTF_FER,
1645 .start = IRQ_PF0,
1671 .start = PORTG_FER,
1676 .start = IRQ_PG0,
1772 .start = SPI0_REGBASE,
1777 .start = CH_SPI0_TX,
1782 .start = CH_SPI0_RX,
1791 .start = SPI1_REGBASE,
1796 .start = CH_SPI1_TX,
1801 .start = CH_SPI1_RX,
1845 .start = TWI0_CLKDIV,
1850 .start = IRQ_TWI0,
1870 .start = TWI1_CLKDIV,
1875 .start = IRQ_TWI1,
/linux-4.4.14/fs/pstore/
H A Dram_core.c33 atomic_t start; member in struct:persistent_ram_buffer
47 return atomic_read(&prz->buffer->start); buffer_start()
50 /* increase and wrap the start pointer, returning the old value */ buffer_start_add_atomic()
57 old = atomic_read(&prz->buffer->start); buffer_start_add_atomic()
61 } while (atomic_cmpxchg(&prz->buffer->start, old, new) != old); buffer_start_add_atomic()
85 /* increase and wrap the start pointer, returning the old value */ buffer_start_add_locked()
94 old = atomic_read(&prz->buffer->start); buffer_start_add_locked()
98 atomic_set(&prz->buffer->start, new); buffer_start_add_locked()
156 unsigned int start, unsigned int count) persistent_ram_update_ecc()
169 block = buffer->data + (start & ~(ecc_block_size - 1)); persistent_ram_update_ecc()
170 par = prz->par_buffer + (start / ecc_block_size) * ecc_size; persistent_ram_update_ecc()
178 } while (block < buffer->data + start + count); persistent_ram_update_ecc()
299 const void *s, unsigned int start, unsigned int count) persistent_ram_update()
302 memcpy(buffer->data + start, s, count); persistent_ram_update()
303 persistent_ram_update_ecc(prz, start, count); persistent_ram_update()
310 size_t start = buffer_start(prz); persistent_ram_save_old() local
325 memcpy(prz->old_log, &buffer->data[start], size - start); persistent_ram_save_old()
326 memcpy(prz->old_log + size - start, &buffer->data[0], start); persistent_ram_save_old()
334 size_t start; persistent_ram_write() local
343 start = buffer_start_add(prz, c); persistent_ram_write()
345 rem = prz->buffer_size - start; persistent_ram_write()
347 persistent_ram_update(prz, s, start, rem); persistent_ram_write()
350 start = 0; persistent_ram_write()
352 persistent_ram_update(prz, s, start, c); persistent_ram_write()
378 atomic_set(&prz->buffer->start, 0); persistent_ram_zap()
383 static void *persistent_ram_vmap(phys_addr_t start, size_t size, persistent_ram_vmap() argument
393 page_start = start - offset_in_page(start); persistent_ram_vmap()
394 page_count = DIV_ROUND_UP(size + offset_in_page(start), PAGE_SIZE); persistent_ram_vmap()
418 static void *persistent_ram_iomap(phys_addr_t start, size_t size, persistent_ram_iomap() argument
423 if (!request_mem_region(start, size, "persistent_ram")) { persistent_ram_iomap()
425 (unsigned long long)size, (unsigned long long)start); persistent_ram_iomap()
433 va = ioremap(start, size); persistent_ram_iomap()
435 va = ioremap_wc(start, size); persistent_ram_iomap()
440 static int persistent_ram_buffer_map(phys_addr_t start, phys_addr_t size, persistent_ram_buffer_map() argument
443 prz->paddr = start; persistent_ram_buffer_map()
446 if (pfn_valid(start >> PAGE_SHIFT)) persistent_ram_buffer_map()
447 prz->vaddr = persistent_ram_vmap(start, size, memtype); persistent_ram_buffer_map()
449 prz->vaddr = persistent_ram_iomap(start, size, memtype); persistent_ram_buffer_map()
453 (unsigned long long)size, (unsigned long long)start); persistent_ram_buffer_map()
457 prz->buffer = prz->vaddr + offset_in_page(start); persistent_ram_buffer_map()
477 pr_info("found existing invalid buffer, size %zu, start %zu\n", persistent_ram_post_init()
480 pr_debug("found existing buffer, size %zu, start %zu\n", persistent_ram_post_init()
514 struct persistent_ram_zone *persistent_ram_new(phys_addr_t start, size_t size, persistent_ram_new() argument
527 ret = persistent_ram_buffer_map(start, size, prz, memtype); persistent_ram_new()
155 persistent_ram_update_ecc(struct persistent_ram_zone *prz, unsigned int start, unsigned int count) persistent_ram_update_ecc() argument
298 persistent_ram_update(struct persistent_ram_zone *prz, const void *s, unsigned int start, unsigned int count) persistent_ram_update() argument
/linux-4.4.14/arch/arm/mach-w90x900/
H A Ddev.c79 .start = NUC900_FLASH_BASE,
99 .start = W90X900_PA_USBEHCIHOST,
104 .start = IRQ_USBH,
127 .start = W90X900_PA_USBOHCIHOST,
132 .start = IRQ_USBH,
154 .start = W90X900_PA_USBDEV,
159 .start = IRQ_USBD,
176 .start = W90X900_PA_EMC,
181 .start = IRQ_EMCTX,
186 .start = IRQ_EMCRX,
220 .start = W90X900_PA_I2C + SPIOFFSET,
225 .start = IRQ_SSP,
273 .start = W90X900_PA_TIMER,
278 .start = IRQ_WDT,
304 .start = W90X900_PA_RTC,
309 .start = IRQ_RTC,
326 .start = W90X900_PA_ADC,
331 .start = IRQ_ADC,
348 .start = W90X900_PA_FMI,
353 .start = IRQ_FMI,
403 .start = W90X900_PA_KPI,
408 .start = IRQ_KPI,
466 .start = W90X900_PA_LCD,
471 .start = IRQ_LCD,
494 .start = W90X900_PA_ACTL,
499 .start = IRQ_ACTL,
/linux-4.4.14/drivers/misc/
H A Dsram.c52 u32 start; member in struct:sram_reserve
90 phys_addr_t start, struct sram_partition *part) sram_add_pool()
99 ret = gen_pool_add_virt(part->pool, (unsigned long)part->base, start, sram_add_pool()
110 phys_addr_t start, struct sram_partition *part) sram_add_export()
115 (unsigned long long)start); sram_add_export()
128 phys_addr_t start) sram_add_partition()
134 part->base = sram->virt_base + block->start; sram_add_partition()
137 ret = sram_add_pool(sram, block, start, part); sram_add_partition()
142 ret = sram_add_export(sram, block, start, part); sram_add_partition()
175 return ra->start - rb->start; sram_reserve_cmp()
213 if (child_res.start < res->start || child_res.end > res->end) { for_each_available_child_of_node()
221 block->start = child_res.start - res->start; for_each_available_child_of_node()
252 block->start, block->start + block->size); for_each_available_child_of_node()
255 block->start, block->start + block->size); for_each_available_child_of_node()
263 rblocks[nblocks - 1].start = size;
282 if (block->start < cur_start) {
285 block->start, cur_start);
293 res->start + block->start);
300 /* current start is in a reserved block, so continue after it */
301 if (block->start == cur_start) {
302 cur_start = block->start + block->size;
311 cur_size = block->start - cur_start;
318 res->start + cur_start, cur_size, -1);
325 cur_start = block->start + block->size;
358 if (!devm_request_mem_region(sram->dev, res->start, size, pdev->name)) { sram_probe()
363 sram->virt_base = devm_ioremap_wc(sram->dev, res->start, size); sram_probe()
89 sram_add_pool(struct sram_dev *sram, struct sram_reserve *block, phys_addr_t start, struct sram_partition *part) sram_add_pool() argument
109 sram_add_export(struct sram_dev *sram, struct sram_reserve *block, phys_addr_t start, struct sram_partition *part) sram_add_export() argument
127 sram_add_partition(struct sram_dev *sram, struct sram_reserve *block, phys_addr_t start) sram_add_partition() argument
/linux-4.4.14/arch/blackfin/mach-bf518/boards/
H A Dtcm-bf518.c60 .start = 0x20000000,
239 .start = SPI0_REGBASE,
244 .start = CH_SPI0,
249 .start = IRQ_SPI0,
274 .start = SPI1_REGBASE,
279 .start = CH_SPI1,
284 .start = IRQ_SPI1,
305 .start = UART0_THR,
310 .start = IRQ_UART0_TX,
315 .start = IRQ_UART0_RX,
320 .start = IRQ_UART0_ERROR,
325 .start = CH_UART0_TX,
330 .start = CH_UART0_RX,
353 .start = UART1_THR,
358 .start = IRQ_UART1_TX,
363 .start = IRQ_UART1_RX,
368 .start = IRQ_UART1_ERROR,
373 .start = CH_UART1_TX,
378 .start = CH_UART1_RX,
404 .start = 0xFFC00400,
409 .start = IRQ_UART0_RX,
414 .start = CH_UART0_RX,
430 .start = 0xFFC02000,
435 .start = IRQ_UART1_RX,
440 .start = CH_UART1_RX,
460 .start = TWI0_REGBASE,
465 .start = IRQ_TWI,
500 .start = SPORT0_TCR1,
505 .start = IRQ_SPORT0_RX,
510 .start = IRQ_SPORT0_ERROR,
534 .start = SPORT1_TCR1,
539 .start = IRQ_SPORT1_RX,
544 .start = IRQ_SPORT1_ERROR,
/linux-4.4.14/arch/um/drivers/
H A Dslip_common.c7 int i, n, size, start; slip_proto_read() local
30 start = slip->pos; slip_proto_read()
32 size = slip_unesc(slip->ibuf[start + i], slip->ibuf,&slip->pos, slip_proto_read()
36 memmove(slip->ibuf, &slip->ibuf[start+i+1], slip_proto_read()
/linux-4.4.14/arch/unicore32/kernel/
H A Dpuv3-core.c52 .start = io_v2p(PKUNITY_USB_BASE),
56 .start = IRQ_USB,
59 .start = IRQ_USB,
84 .start = io_v2p(PKUNITY_SDC_BASE),
89 .start = IRQ_SDC,
97 .start = io_v2p(PKUNITY_UNIGFX_BASE),
105 .start = io_v2p(PKUNITY_RTC_BASE),
110 .start = IRQ_RTCAlarm,
115 .start = IRQ_RTC,
123 .start = io_v2p(PKUNITY_OST_BASE) + 0x80,
131 .start = io_v2p(PKUNITY_UART0_BASE),
136 .start = IRQ_UART0,
144 .start = io_v2p(PKUNITY_UART1_BASE),
149 .start = IRQ_UART1,
157 .start = io_v2p(PKUNITY_UMAL_BASE),
162 .start = IRQ_UMAL,
H A Dsetup.c73 .start = 0,
79 .start = 0,
134 static int __init uc32_add_memory(unsigned long start, unsigned long size) uc32_add_memory() argument
140 "ignoring memory at %#lx\n", start); uc32_add_memory()
145 * Ensure that start/size are aligned to a page boundary. uc32_add_memory()
146 * Size is appropriately rounded down, start is rounded up. uc32_add_memory()
148 size -= start & ~PAGE_MASK; uc32_add_memory()
150 bank->start = PAGE_ALIGN(start); uc32_add_memory()
165 * Pick out the memory size. We look for mem=size@start,
166 * where start and size are "size[KkMm]"
171 unsigned long size, start; early_mem() local
184 start = PHYS_OFFSET; early_mem()
187 start = memparse(endp + 1, NULL); early_mem()
189 uc32_add_memory(start, size); early_mem()
201 kernel_code.start = virt_to_phys(_stext); request_standard_resources()
203 kernel_data.start = virt_to_phys(_sdata); request_standard_resources()
212 res->start = mi->bank[i].start; request_standard_resources()
213 res->end = mi->bank[i].start + mi->bank[i].size - 1; request_standard_resources()
218 if (kernel_code.start >= res->start && request_standard_resources()
221 if (kernel_data.start >= res->start && request_standard_resources()
350 .start = c_start,
/linux-4.4.14/include/asm-generic/
H A Dcacheflush.h14 #define flush_cache_range(vma, start, end) do { } while (0)
20 #define flush_icache_range(start, end) do { } while (0)
23 #define flush_cache_vmap(start, end) do { } while (0)
24 #define flush_cache_vunmap(start, end) do { } while (0)
/linux-4.4.14/arch/powerpc/mm/
H A Dtlb_hash32.c82 * - flush_tlb_range(vma, start, end) flushes a range of pages
83 * - flush_tlb_kernel_range(start, end) flushes kernel pages
90 static void flush_range(struct mm_struct *mm, unsigned long start, flush_range() argument
102 start &= PAGE_MASK; flush_range()
103 if (start >= end) flush_range()
106 pmd = pmd_offset(pud_offset(pgd_offset(mm, start), start), start); flush_range()
108 pmd_end = ((start + PGDIR_SIZE) & PGDIR_MASK) - 1; flush_range()
112 count = ((pmd_end - start) >> PAGE_SHIFT) + 1; flush_range()
113 flush_hash_pages(ctx, start, pmd_val(*pmd), count); flush_range()
117 start = pmd_end + 1; flush_range()
125 void flush_tlb_kernel_range(unsigned long start, unsigned long end) flush_tlb_kernel_range() argument
127 flush_range(&init_mm, start, end); flush_tlb_kernel_range()
175 void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, flush_tlb_range() argument
178 flush_range(vma->vm_mm, start, end); flush_tlb_range()
H A Dinit_64.c158 * represents the start of the section it is within. Note that we have to
166 /* Return the pfn of the start of the section. */ vmemmap_section_start()
175 static int __meminit vmemmap_populated(unsigned long start, int page_size) vmemmap_populated() argument
177 unsigned long end = start + page_size; vmemmap_populated()
178 start = (unsigned long)(pfn_to_page(vmemmap_section_start(start))); vmemmap_populated()
180 for (; start < end; start += (PAGES_PER_SECTION * sizeof(struct page))) vmemmap_populated()
181 if (pfn_valid(page_to_pfn((struct page *)start))) vmemmap_populated()
195 static void __meminit vmemmap_create_mapping(unsigned long start, vmemmap_create_mapping() argument
214 BUG_ON(map_kernel_page(start + i, phys, flags)); vmemmap_create_mapping()
218 static void vmemmap_remove_mapping(unsigned long start, vmemmap_remove_mapping() argument
224 static void __meminit vmemmap_create_mapping(unsigned long start, vmemmap_create_mapping() argument
228 int mapped = htab_bolt_mapping(start, start + page_size, phys, vmemmap_create_mapping()
236 static void vmemmap_remove_mapping(unsigned long start, vmemmap_remove_mapping() argument
239 int mapped = htab_remove_mapping(start, start + page_size, vmemmap_remove_mapping()
281 unsigned long start, vmemmap_list_populate()
293 vmem_back->virt_addr = start; vmemmap_list_populate()
299 int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node) vmemmap_populate() argument
304 start = _ALIGN_DOWN(start, page_size); vmemmap_populate()
306 pr_debug("vmemmap_populate %lx..%lx, node %d\n", start, end, node); vmemmap_populate()
308 for (; start < end; start += page_size) { vmemmap_populate()
311 if (vmemmap_populated(start, page_size)) vmemmap_populate()
318 vmemmap_list_populate(__pa(p), start, node); vmemmap_populate()
321 start, start + page_size, p); vmemmap_populate()
323 vmemmap_create_mapping(start, page_size, __pa(p)); vmemmap_populate()
330 static unsigned long vmemmap_list_free(unsigned long start) vmemmap_list_free() argument
338 if (vmem_back->virt_addr == start) vmemmap_list_free()
362 void __ref vmemmap_free(unsigned long start, unsigned long end) vmemmap_free() argument
366 start = _ALIGN_DOWN(start, page_size); vmemmap_free()
368 pr_debug("vmemmap_free %lx...%lx\n", start, end); vmemmap_free()
370 for (; start < end; start += page_size) { vmemmap_free()
378 if (vmemmap_populated(start, page_size)) vmemmap_free()
381 addr = vmemmap_list_free(start); vmemmap_free()
403 vmemmap_remove_mapping(start, page_size); vmemmap_free()
280 vmemmap_list_populate(unsigned long phys, unsigned long start, int node) vmemmap_list_populate() argument
/linux-4.4.14/arch/sh/kernel/cpu/sh4a/
H A Dpinmux-sh7722.c9 .start = 0xa4050100,
/linux-4.4.14/arch/metag/include/asm/
H A Dcacheflush.h14 void metag_data_cache_flush_all(const void *start);
15 void metag_code_cache_flush_all(const void *start);
24 void metag_data_cache_flush(const void *start, int bytes);
25 void metag_code_cache_flush(const void *start, int bytes);
50 unsigned long start, unsigned long end) flush_cache_range()
76 static inline void flush_cache_vmap(unsigned long start, unsigned long end) flush_cache_vmap() argument
81 static inline void flush_cache_vunmap(unsigned long start, unsigned long end) flush_cache_vunmap() argument
93 #define flush_cache_range(vma, start, end) do { } while (0)
98 #define flush_cache_vmap(start, end) do { } while (0)
99 #define flush_cache_vunmap(start, end) do { } while (0)
157 static inline void cachew_region_op(void *start, unsigned long size, cachew_region_op() argument
160 unsigned long offset = (unsigned long)start & 0x3f; cachew_region_op()
164 start -= offset; cachew_region_op()
168 __builtin_meta2_cachewd(start, op); cachew_region_op()
169 start += 0x40; cachew_region_op()
198 static inline void flush_dcache_region(void *start, unsigned long size) flush_dcache_region() argument
202 cachew_region_op(start, size, CACHEW_FLUSH_L1D_L2); flush_dcache_region()
204 l2c_fence_flush(start + size - 1); flush_dcache_region()
206 metag_data_cache_flush(start, size); flush_dcache_region()
211 static inline void writeback_dcache_region(void *start, unsigned long size) writeback_dcache_region() argument
214 cachew_region_op(start, size, CACHEW_WRITEBACK_L1D_L2); writeback_dcache_region()
215 l2c_fence(start + size - 1); writeback_dcache_region()
220 static inline void invalidate_dcache_region(void *start, unsigned long size) invalidate_dcache_region() argument
223 cachew_region_op(start, size, CACHEW_INVALIDATE_L1D_L2); invalidate_dcache_region()
225 metag_data_cache_flush(start, size); invalidate_dcache_region()
49 flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_cache_range() argument
H A Dtlbflush.h16 * - flush_tlb_range(mm, start, end) flushes a range of pages
17 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages
18 * - flush_tlb_pgtables(mm, start, end) flushes a range of page tables
59 unsigned long start, unsigned long end) flush_tlb_range()
65 unsigned long start, unsigned long end) flush_tlb_pgtables()
70 static inline void flush_tlb_kernel_range(unsigned long start, flush_tlb_kernel_range() argument
58 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_tlb_range() argument
64 flush_tlb_pgtables(struct mm_struct *mm, unsigned long start, unsigned long end) flush_tlb_pgtables() argument
/linux-4.4.14/arch/microblaze/include/asm/
H A Dcacheflush.h60 #define flush_icache_range(start, end) mbc->iflr(start, end);
62 #define invalidate_icache_range(start, end) mbc->iinr(start, end);
71 #define invalidate_dcache_range(start, end) mbc->dinr(start, end);
73 #define flush_dcache_range(start, end) mbc->dflr(start, end);
88 #define flush_cache_vmap(start, end) do { } while (0)
89 #define flush_cache_vunmap(start, end) do { } while (0)
97 #define flush_cache_range(vma, start, len) { \
98 flush_icache_range((unsigned) (start), (unsigned) (start) + (len)); \
99 flush_dcache_range((unsigned) (start), (unsigned) (start) + (len)); \
103 #define flush_cache_range(vma, start, len) do { } while (0)
H A Dtlbflush.h37 unsigned long start, unsigned long end) local_flush_tlb_range()
40 #define flush_tlb_kernel_range(start, end) do { } while (0)
55 unsigned long start, unsigned long end) { }
63 #define flush_tlb_range(mm, start, end) BUG()
64 #define flush_tlb_pgtables(mm, start, end) BUG()
65 #define flush_tlb_kernel_range(start, end) BUG()
36 local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) local_flush_tlb_range() argument
54 flush_tlb_pgtables(struct mm_struct *mm, unsigned long start, unsigned long end) flush_tlb_pgtables() argument
/linux-4.4.14/arch/cris/include/asm/
H A Dcacheflush.h13 #define flush_cache_range(vma, start, end) do { } while (0)
19 #define flush_icache_range(start, end) do { } while (0)
22 #define flush_cache_vmap(start, end) do { } while (0)
23 #define flush_cache_vunmap(start, end) do { } while (0)
/linux-4.4.14/arch/ia64/include/uapi/asm/
H A Dauxvec.h6 * them, start the architecture-specific ones at 32.
/linux-4.4.14/arch/avr32/include/asm/
H A Dtlbflush.h20 * - flush_tlb_range(vma, start, end) flushes a range of pages
21 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages
26 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
30 extern void flush_tlb_kernel_range(unsigned long start, unsigned long end);
/linux-4.4.14/arch/arm/mach-mmp/include/mach/
H A Ddevices.h14 unsigned long start; member in struct:pxa_device_desc
25 .start = _start, \
36 .start = _start, \
47 .start = _start, \
/linux-4.4.14/tools/build/feature/
H A Dtest-libunwind-debug-frame.c7 const char *obj_name, unw_word_t start,
/linux-4.4.14/tools/lib/symbol/
H A Dkallsyms.c13 char type, u64 start)) kallsyms__parse()
26 u64 start; kallsyms__parse() local
37 len = hex2u64(line, &start); kallsyms__parse()
53 err = process_symbol(arg, symbol_name, symbol_type, start); kallsyms__parse()
11 kallsyms__parse(const char *filename, void *arg, int (*process_symbol)(void *arg, const char *name, char type, u64 start)) kallsyms__parse() argument
/linux-4.4.14/arch/x86/kvm/
H A Dmtrr.c146 u64 start; member in struct:fixed_mtrr_segment
151 /* the start position in kvm_mtrr.fixed_ranges[]. */
158 .start = 0x0,
169 .start = 0x80000,
180 .start = 0xc0000,
218 static void fixed_mtrr_seg_unit_range(int seg, int unit, u64 *start, u64 *end) fixed_mtrr_seg_unit_range() argument
223 *start = mtrr_seg->start + unit * unit_size; fixed_mtrr_seg_unit_range()
224 *end = *start + unit_size; fixed_mtrr_seg_unit_range()
232 WARN_ON(mtrr_seg->start + unit * fixed_mtrr_seg_unit_size(seg) fixed_mtrr_seg_unit_range_index()
244 n = (mtrr_seg->end - mtrr_seg->start) >> mtrr_seg->range_shift; fixed_mtrr_seg_end_range_index()
248 static bool fixed_msr_to_range(u32 msr, u64 *start, u64 *end) fixed_msr_to_range() argument
255 fixed_mtrr_seg_unit_range(seg, unit, start, end); fixed_msr_to_range()
276 if (mtrr_seg->start <= addr && addr < mtrr_seg->end) fixed_mtrr_addr_to_seg()
290 index += (addr - mtrr_seg->start) >> mtrr_seg->range_shift; fixed_mtrr_addr_seg_to_range_index()
299 return mtrr_seg->start + ((pos + 1) << mtrr_seg->range_shift); fixed_mtrr_range_end_addr()
302 static void var_mtrr_range(struct kvm_mtrr_range *range, u64 *start, u64 *end) var_mtrr_range() argument
306 *start = range->base & PAGE_MASK; var_mtrr_range()
313 *end = (*start | ~mask) + 1; var_mtrr_range()
319 gfn_t start, end; update_mtrr() local
330 if (fixed_msr_to_range(msr, &start, &end)) { update_mtrr()
334 start = 0x0; update_mtrr()
339 var_mtrr_range(&mtrr_state->var_ranges[index], &start, &end); update_mtrr()
342 kvm_zap_gfn_range(vcpu->kvm, gpa_to_gfn(start), gpa_to_gfn(end)); update_mtrr()
453 u64 start; member in struct:mtrr_iter
460 /* [start, end) is not fully covered in MTRRs? */
489 seg = fixed_mtrr_addr_to_seg(iter->start); mtrr_lookup_fixed_start()
494 index = fixed_mtrr_addr_seg_to_range_index(iter->start, seg); mtrr_lookup_fixed_start()
503 u64 start, end; match_var_range() local
505 var_mtrr_range(range, &start, &end); match_var_range()
506 if (!(start >= iter->end || end <= iter->start)) { match_var_range()
514 iter->partial_map |= iter->start_max < start; match_var_range()
541 iter->start_max = iter->start; mtrr_lookup_var_start()
584 struct kvm_mtrr *mtrr_state, u64 start, u64 end) mtrr_lookup_init()
587 iter->start = start; mtrr_lookup_init()
628 u64 start, end; kvm_mtrr_get_guest_memory_type() local
633 start = gfn_to_gpa(gfn); kvm_mtrr_get_guest_memory_type()
634 end = start + PAGE_SIZE; kvm_mtrr_get_guest_memory_type()
636 mtrr_for_each_mem_type(&iter, mtrr_state, start, end) { kvm_mtrr_get_guest_memory_type()
705 u64 start, end; kvm_mtrr_check_gfn_range_consistency() local
708 start = gfn_to_gpa(gfn); kvm_mtrr_check_gfn_range_consistency()
710 mtrr_for_each_mem_type(&iter, mtrr_state, start, end) { kvm_mtrr_check_gfn_range_consistency()
583 mtrr_lookup_init(struct mtrr_iter *iter, struct kvm_mtrr *mtrr_state, u64 start, u64 end) mtrr_lookup_init() argument
/linux-4.4.14/arch/mips/jazz/
H A Dsetup.c32 .start = 0x00,
37 .start = 0x40,
42 .start = 0x80,
47 .start = 0xc0,
123 .start = JAZZ_SCSI_BASE,
128 .start = JAZZ_SCSI_DMA,
133 .start = JAZZ_SCSI_IRQ,
147 .start = JAZZ_ETHERNET_BASE,
152 .start = JAZZ_ETHERNET_IRQ,
166 .start = 0x70,
171 .start = 8,
/linux-4.4.14/arch/arm64/kernel/
H A Dsys_compat.c32 __do_compat_cache_op(unsigned long start, unsigned long end) __do_compat_cache_op() argument
37 unsigned long chunk = min(PAGE_SIZE, end - start); __do_compat_cache_op()
42 ret = __flush_cache_user_range(start, start + chunk); __do_compat_cache_op()
47 start += chunk; __do_compat_cache_op()
48 } while (start < end); __do_compat_cache_op()
54 do_compat_cache_op(unsigned long start, unsigned long end, int flags) do_compat_cache_op() argument
56 if (end < start || flags) do_compat_cache_op()
59 if (!access_ok(VERIFY_READ, start, end - start)) do_compat_cache_op()
62 return __do_compat_cache_op(start, end); do_compat_cache_op()
/linux-4.4.14/arch/mips/kernel/
H A Dsetup.c91 void __init add_memory_region(phys_addr_t start, phys_addr_t size, long type) add_memory_region() argument
97 if (start + size < start) { add_memory_region()
112 if (start + size < entry->addr) add_memory_region()
115 if (entry->addr + entry->size < start) add_memory_region()
118 top = max(entry->addr + entry->size, start + size); add_memory_region()
119 entry->addr = min(entry->addr, start); add_memory_region()
130 boot_mem_map.map[x].addr = start; add_memory_region()
136 void __init detect_memory_region(phys_addr_t start, phys_addr_t sz_min, phys_addr_t sz_max) detect_memory_region() argument
148 (unsigned long long) start, detect_memory_region()
152 add_memory_region(start, size, BOOT_MEM_RAM); detect_memory_region()
192 unsigned long start = memparse(p, &p); rd_start_early() local
196 if (start < XKPHYS) rd_start_early()
197 start = (int)start; rd_start_early()
199 initrd_start = start; rd_start_early()
200 initrd_end += start; rd_start_early()
226 pr_err("initrd start must be page aligned\n"); init_initrd()
230 pr_err("initrd start < PAGE_OFFSET\n"); init_initrd()
330 unsigned long start, end; bootmem_init() local
335 start = PFN_UP(boot_mem_map.map[i].addr); bootmem_init()
341 if (start < min_low_pfn) bootmem_init()
342 min_low_pfn = start; bootmem_init()
350 if (start >= mapstart) bootmem_init()
352 mapstart = max(reserved_end, start); bootmem_init()
395 unsigned long start, end; bootmem_init() local
397 start = PFN_UP(boot_mem_map.map[i].addr); bootmem_init()
401 if (start <= min_low_pfn) bootmem_init()
402 start = min_low_pfn; bootmem_init()
403 if (start >= end) bootmem_init()
413 if (end <= start) bootmem_init()
417 memblock_add_node(PFN_PHYS(start), PFN_PHYS(end - start), 0); bootmem_init()
424 unsigned long start, end, size; bootmem_init() local
426 start = PFN_UP(boot_mem_map.map[i].addr); bootmem_init()
437 memory_present(0, start, end); bootmem_init()
445 * We are rounding up the start address of usable memory bootmem_init()
448 if (start >= max_low_pfn) bootmem_init()
450 if (start < reserved_end) bootmem_init()
451 start = reserved_end; bootmem_init()
458 if (end <= start) bootmem_init()
460 size = end - start; bootmem_init()
463 free_bootmem(PFN_PHYS(start), size << PAGE_SHIFT); bootmem_init()
464 memory_present(0, start, end); bootmem_init()
507 phys_addr_t start, size; early_parse_mem() local
518 start = 0; early_parse_mem()
521 start = memparse(p + 1, &p); early_parse_mem()
523 add_memory_region(start, size, BOOT_MEM_RAM); early_parse_mem()
537 unsigned long start = boot_mem_map.map[i].addr; early_parse_elfcorehdr() local
540 if (setup_elfcorehdr >= start && setup_elfcorehdr < end) { early_parse_elfcorehdr()
599 crashk_res.start = crash_base; mips_parse_crashkernel()
611 crashk_res.start + 1) >> 20), request_crashkernel()
612 (unsigned long)(crashk_res.start >> 20)); request_crashkernel()
694 if (crashk_res.start != crashk_res.end) arch_mem_init()
695 reserve_bootmem(crashk_res.start, arch_mem_init()
696 crashk_res.end - crashk_res.start + 1, arch_mem_init()
721 code_resource.start = __pa_symbol(&_text); resource_init()
723 data_resource.start = __pa_symbol(&_etext); resource_init()
728 unsigned long start, end; resource_init() local
730 start = boot_mem_map.map[i].addr; resource_init()
732 if (start >= HIGHMEM_START) resource_init()
749 res->start = start; resource_init()
H A Dspinlock_test.c12 ktime_t start, finish; ss_get() local
20 start = ktime_get(); ss_get()
32 *val = ktime_us_delta(finish, start); ss_get()
51 ktime_t start; member in struct:spin_multi_per_thread
69 pt->start = ktime_get(); multi_other()
111 *val = ktime_us_delta(finish, t1.start); multi_get()
/linux-4.4.14/arch/alpha/include/asm/
H A Da.out.h7 /* Assume that start addresses below 4G belong to a TASO application.
9 Worse, we have to notice the start address before swapping to use
/linux-4.4.14/arch/cris/arch-v32/drivers/pci/
H A Dbios.c43 resource_size_t start = res->start; pcibios_align_resource() local
45 if ((res->flags & IORESOURCE_IO) && (start & 0x300)) pcibios_align_resource()
46 start = (start + 0x3ff) & ~0x3ff; pcibios_align_resource()
48 return start; pcibios_align_resource()
65 if (!r->start && r->end) { pcibios_enable_resources()
74 if (dev->resource[PCI_ROM_RESOURCE].start) pcibios_enable_resources()
/linux-4.4.14/fs/befs/
H A Dendian.h77 run.start = le16_to_cpu((__force __le16)n.start); fsrun_to_cpu()
81 run.start = be16_to_cpu((__force __be16)n.start); fsrun_to_cpu()
94 run.start = cpu_to_le16(n.start); cpu_to_fsrun()
98 run.start = cpu_to_be16(n.start); cpu_to_fsrun()
/linux-4.4.14/arch/m68k/mm/
H A Dcache.c78 unsigned long start, end; flush_icache_range() local
79 start = address & ICACHE_SET_MASK; flush_icache_range()
81 if (start > end) { flush_icache_range()
85 flush_cf_icache(start, end); flush_icache_range()
112 unsigned long start, end; flush_icache_user_range() local
113 start = addr & ICACHE_SET_MASK; flush_icache_user_range()
115 if (start > end) { flush_icache_user_range()
119 flush_cf_icache(start, end); flush_icache_user_range()
/linux-4.4.14/arch/mips/vr41xx/common/
H A Drtc.c31 .start = 0x0b0000c0,
36 .start = 0x0b0001c0,
41 .start = ELAPSEDTIME_IRQ,
46 .start = RTCLONG1_IRQ,
54 .start = 0x0f000100,
59 .start = 0x0f000120,
64 .start = ELAPSEDTIME_IRQ,
69 .start = RTCLONG1_IRQ,
/linux-4.4.14/arch/arm/include/asm/hardware/
H A Dmemc.h19 #define video_set_dma(start,end,offset) \
21 memc_write (VDMA_START, (start >> 2)); \
/linux-4.4.14/arch/avr32/kernel/
H A Dsetup.c50 .start = 0,
56 .start = 0,
83 kernel_code.start = __pa(init_mm.start_code); resource_init()
92 mem->start, mem->end); resource_init()
103 res->name, res->start, res->end); resource_init()
108 add_physical_memory(resource_size_t start, resource_size_t end) add_physical_memory() argument
114 if (end < next->start) add_physical_memory()
116 if (start <= next->end) { add_physical_memory()
121 start, end, next->start, next->end); add_physical_memory()
129 start, end); add_physical_memory()
134 new->start = start; add_physical_memory()
143 add_reserved_region(resource_size_t start, resource_size_t end, add_reserved_region() argument
148 if (end < start) add_reserved_region()
156 if (end < next->start) add_reserved_region()
158 if (start <= next->end) add_reserved_region()
163 new->start = start; add_reserved_region()
181 target = ALIGN(mem->start, align); find_free_region()
183 if ((target + size) <= res->start) find_free_region()
196 alloc_reserved_region(resource_size_t *start, resource_size_t size, alloc_reserved_region() argument
209 *start = target; alloc_reserved_region()
287 * Pick out the memory size. We look for mem=size@start,
288 * where start and size are "size[KkMmGg]"
292 resource_size_t size, start; early_mem() local
294 start = system_ram->start; early_mem()
297 start = memparse(p + 1, &p); early_mem()
299 system_ram->start = start; early_mem()
300 system_ram->end = system_ram->start + size - 1; early_mem()
318 unsigned long start, end; parse_tag_mem() local
328 start = tag->u.mem_range.addr; parse_tag_mem()
331 add_physical_memory(start, end); parse_tag_mem()
480 printk(" %08x-%08x\n", res->start, res->end); setup_bootmem()
484 res->start, res->end, res->name); setup_bootmem()
492 first_pfn = PFN_UP(res->start); setup_bootmem()
534 if (res->start > PFN_PHYS(max_pfn)) setup_bootmem()
542 if (res->start >= PFN_PHYS(first_pfn) setup_bootmem()
545 res->start, setup_bootmem()
567 kernel_code.start = __pa(__init_begin); setup_arch()
569 kernel_data.start = __pa(init_mm.end_code); setup_arch()
/linux-4.4.14/arch/arc/include/asm/
H A Dtlbflush.h17 void local_flush_tlb_kernel_range(unsigned long start, unsigned long end);
19 unsigned long start, unsigned long end);
20 void local_flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start,
31 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
34 extern void flush_tlb_kernel_range(unsigned long start, unsigned long end);
37 extern void flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end);
/linux-4.4.14/include/uapi/linux/netfilter/
H A Dnfnetlink_compat.h50 #define NFA_NEST_END(skb, start) \
51 ({ (start)->nfa_len = skb_tail_pointer(skb) - (unsigned char *)(start); \
53 #define NFA_NEST_CANCEL(skb, start) \
54 ({ if (start) \
55 skb_trim(skb, (unsigned char *) (start) - (skb)->data); \
/linux-4.4.14/arch/sh/boards/mach-sh03/
H A Dsetup.c34 .start = 0x1f0,
39 .start = 0x1f0 + 0x206,
44 .start = IRL2_IRQ,
58 .start = 0xa0800000,
92 cf_ide_resources[0].start += (unsigned long)cf_ide_base; sh03_devices_setup()
94 cf_ide_resources[1].start += (unsigned long)cf_ide_base; sh03_devices_setup()
/linux-4.4.14/drivers/firmware/efi/
H A Dfake_mem.c50 if (m1->range.start < m2->range.start) cmp_fake_mem()
52 if (m1->range.start > m2->range.start) cmp_fake_mem()
59 u64 start, end, m_start, m_end, m_attr; efi_fake_memmap() local
73 start = md->phys_addr; efi_fake_memmap()
74 end = start + (md->num_pages << EFI_PAGE_SHIFT) - 1; efi_fake_memmap()
78 m_start = fake_mems[i].range.start; efi_fake_memmap()
81 if (m_start <= start) { efi_fake_memmap()
83 if (start < m_end && m_end < end) efi_fake_memmap()
86 if (start < m_start && m_start < end) { efi_fake_memmap()
118 start = md->phys_addr; efi_fake_memmap()
123 m_start = fake_mems[i].range.start; efi_fake_memmap()
127 if (m_start <= start && end <= m_end) efi_fake_memmap()
130 if (m_start <= start && efi_fake_memmap()
131 (start < m_end && m_end < end)) { efi_fake_memmap()
145 if ((start < m_start && m_start < end) && m_end < end) { efi_fake_memmap()
166 if ((start < m_start && m_start < end) && efi_fake_memmap()
197 u64 start = 0, mem_size = 0, attribute = 0; setup_fake_mem() local
206 start = memparse(p+1, &p); setup_fake_mem()
218 fake_mems[nr_fake_mem].range.start = start; setup_fake_mem()
219 fake_mems[nr_fake_mem].range.end = start + mem_size - 1; setup_fake_mem()
232 fake_mems[i].attribute, fake_mems[i].range.start, setup_fake_mem()
/linux-4.4.14/drivers/gpu/drm/radeon/
H A Dradeon_mem.c42 static struct mem_block *split_block(struct mem_block *p, int start, int size, split_block() argument
45 /* Maybe cut off the start of an existing block */ split_block()
46 if (start > p->start) { split_block()
51 newblock->start = start; split_block()
52 newblock->size = p->size - (start - p->start); split_block()
68 newblock->start = start + size; split_block()
91 int start = (p->start + mask) & ~mask; list_for_each() local
92 if (p->file_priv == NULL && start + size <= p->start + p->size) list_for_each()
93 return split_block(p, start, size, file_priv); list_for_each()
99 static struct mem_block *find_block(struct mem_block *heap, int start) find_block() argument
104 if (p->start == start) find_block()
136 static int init_heap(struct mem_block **heap, int start, int size) init_heap() argument
149 blocks->start = start; init_heap()
246 if (copy_to_user(alloc->region_offset, &block->start, radeon_mem_alloc()
301 return init_heap(heap, initheap->start, initheap->size); radeon_mem_init_heap()
/linux-4.4.14/drivers/media/platform/coda/
H A Dtrace.h91 __field(int, start)
99 __entry->start = meta->start;
104 TP_printk("minor = %d, index = %d, start = 0x%x, end = 0x%x, ctx = %d",
105 __entry->minor, __entry->index, __entry->start, __entry->end,
122 __field(int, start)
129 __entry->start = meta ? meta->start : 0;
134 TP_printk("minor = %d, start = 0x%x, end = 0x%x, ctx = %d",
135 __entry->minor, __entry->start, __entry->end, __entry->ctx)
/linux-4.4.14/drivers/gpu/drm/nouveau/
H A Dnv50_fence.c41 u32 start = mem->start * PAGE_SIZE; nv50_fence_context_new() local
42 u32 limit = start + mem->size - 1; nv50_fence_context_new()
58 .start = start, nv50_fence_context_new()
66 u32 start = bo->bo.mem.start * PAGE_SIZE; nv50_fence_context_new() local
67 u32 limit = start + bo->bo.mem.size - 1; nv50_fence_context_new()
73 .start = start, nv50_fence_context_new()
/linux-4.4.14/arch/unicore32/mm/
H A Dcache-ucv2.S44 * __cpuc_flush_user_range(start, end, flags)
48 * - start - start address (may not be aligned)
83 * __cpuc_coherent_kern_range(start,end)
84 * __cpuc_coherent_user_range(start,end)
90 * - start - virtual start address of region
145 * __cpuc_dma_clean_range(start,end)
146 * - start - virtual start address of region
182 * __cpuc_dma_inv_range(start,end)
183 * __cpuc_dma_flush_range(start,end)
184 * - start - virtual start address of region
/linux-4.4.14/arch/mips/mm/
H A Dtlb-r3k.c81 void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, local_flush_tlb_range() argument
92 cpu_context(cpu, mm) & ASID_MASK, start, end); local_flush_tlb_range()
95 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_range()
100 start &= PAGE_MASK; local_flush_tlb_range()
103 while (start < end) { local_flush_tlb_range()
106 write_c0_entryhi(start | newpid); local_flush_tlb_range()
107 start += PAGE_SIZE; /* BARRIER */ local_flush_tlb_range()
124 void local_flush_tlb_kernel_range(unsigned long start, unsigned long end) local_flush_tlb_kernel_range() argument
129 printk("[tlbrange<%lu,0x%08lx,0x%08lx>]", start, end); local_flush_tlb_kernel_range()
132 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_kernel_range()
136 start &= PAGE_MASK; local_flush_tlb_kernel_range()
140 while (start < end) { local_flush_tlb_kernel_range()
143 write_c0_entryhi(start); local_flush_tlb_kernel_range()
144 start += PAGE_SIZE; /* BARRIER */ local_flush_tlb_kernel_range()
H A Dtlb-r8k.c61 void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, local_flush_tlb_range() argument
72 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_range()
87 start &= PAGE_MASK; local_flush_tlb_range()
90 while (start < end) { local_flush_tlb_range()
93 write_c0_vaddr(start); local_flush_tlb_range()
94 write_c0_entryhi(start); local_flush_tlb_range()
95 start += PAGE_SIZE; local_flush_tlb_range()
111 void local_flush_tlb_kernel_range(unsigned long start, unsigned long end) local_flush_tlb_kernel_range() argument
115 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_kernel_range()
127 start &= PAGE_MASK; local_flush_tlb_kernel_range()
130 while (start < end) { local_flush_tlb_kernel_range()
133 write_c0_vaddr(start); local_flush_tlb_kernel_range()
134 write_c0_entryhi(start); local_flush_tlb_kernel_range()
135 start += PAGE_SIZE; local_flush_tlb_kernel_range()
/linux-4.4.14/arch/arm64/include/asm/
H A Dcacheflush.h36 * Start addresses are inclusive and end addresses are exclusive; start
48 * flush_icache_range(start, end)
51 * region described by start, end.
52 * - start - virtual start address
55 * __flush_cache_user_range(start, end)
58 * region described by start, end.
59 * - start - virtual start address
68 extern void flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end);
69 extern void flush_icache_range(unsigned long start, unsigned long end);
71 extern long __flush_cache_user_range(unsigned long start, unsigned long end);
145 static inline void flush_cache_vmap(unsigned long start, unsigned long end) flush_cache_vmap() argument
149 static inline void flush_cache_vunmap(unsigned long start, unsigned long end) flush_cache_vunmap() argument
H A Dtlbflush.h45 * flush_tlb_range(mm,start,end)
50 * - start - start address (may not be aligned)
108 unsigned long start, unsigned long end, __flush_tlb_range()
114 if ((end - start) > MAX_TLB_RANGE) { __flush_tlb_range()
119 start = asid | (start >> 12); __flush_tlb_range()
123 for (addr = start; addr < end; addr += 1 << (PAGE_SHIFT - 12)) { __flush_tlb_range()
133 unsigned long start, unsigned long end) flush_tlb_range()
135 __flush_tlb_range(vma, start, end, false); flush_tlb_range()
138 static inline void flush_tlb_kernel_range(unsigned long start, unsigned long end) flush_tlb_kernel_range() argument
142 if ((end - start) > MAX_TLB_RANGE) { flush_tlb_kernel_range()
147 start >>= 12; flush_tlb_kernel_range()
151 for (addr = start; addr < end; addr += 1 << (PAGE_SHIFT - 12)) flush_tlb_kernel_range()
107 __flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end, bool last_level) __flush_tlb_range() argument
132 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_tlb_range() argument
/linux-4.4.14/arch/sh/boards/
H A Dboard-sh2007.c34 .start = SMC0_BASE,
39 .start = evt2irq(0x240),
47 .start = SMC1_BASE,
52 .start = evt2irq(0x280),
80 .start = CF_BASE + CF_OFFSET,
85 .start = CF_BASE + CF_OFFSET + 0x206,
90 .start = evt2irq(0x2c0),

Completed in 5114 milliseconds

1234567891011>>