/linux-4.4.14/arch/ia64/scripts/ |
H A D | check-segrel.S | 2 data4 @segrel(start) 4 start: label
|
/linux-4.4.14/kernel/ |
H A D | range.c | 10 int add_range(struct range *range, int az, int nr_range, u64 start, u64 end) add_range() argument 12 if (start >= end) add_range() 19 range[nr_range].start = start; add_range() 28 u64 start, u64 end) add_range_with_merge() 32 if (start >= end) add_range_with_merge() 35 /* get new start/end: */ add_range_with_merge() 42 common_start = max(range[i].start, start); add_range_with_merge() 47 /* new start/end, will add it back at last */ add_range_with_merge() 48 start = min(range[i].start, start); add_range_with_merge() 53 range[nr_range - 1].start = 0; add_range_with_merge() 60 return add_range(range, az, nr_range, start, end); add_range_with_merge() 63 void subtract_range(struct range *range, int az, u64 start, u64 end) subtract_range() argument 67 if (start >= end) subtract_range() 74 if (start <= range[j].start && end >= range[j].end) { subtract_range() 75 range[j].start = 0; subtract_range() 80 if (start <= range[j].start && end < range[j].end && subtract_range() 81 range[j].start < end) { subtract_range() 82 range[j].start = end; subtract_range() 87 if (start > range[j].start && end >= range[j].end && subtract_range() 88 range[j].end > start) { subtract_range() 89 range[j].end = start; subtract_range() 93 if (start > range[j].start && end < range[j].end) { subtract_range() 101 range[i].start = end; subtract_range() 106 range[j].end = start; subtract_range() 117 if (r1->start < r2->start) cmp_range() 119 if (r1->start > r2->start) cmp_range() 139 range[i].start = range[k].start; clean_sort_range() 141 range[k].start = 0; clean_sort_range() 27 add_range_with_merge(struct range *range, int az, int nr_range, u64 start, u64 end) add_range_with_merge() argument
|
H A D | resource.c | 31 .start = 0, 39 .start = 0, 116 width, (unsigned long long) r->start, r_show() 123 .start = r_start, 210 resource_size_t start = new->start; __request_resource() local 214 if (end < start) __request_resource() 216 if (start < root->start) __request_resource() 223 if (!tmp || tmp->start > end) { __request_resource() 230 if (tmp->end < start) __request_resource() 273 tmp->start = 0; __release_child_resources() 336 * Finds the lowest iomem reosurce exists with-in [res->start.res->end) 337 * the caller must specify res->start, res->end, res->flags and "name". 345 resource_size_t start, end; find_next_iomem_res() local 351 start = res->start; find_next_iomem_res() 353 BUG_ON(start >= end); find_next_iomem_res() 365 if (p->start > end) { find_next_iomem_res() 369 if ((p->end >= start) && (p->start < end)) find_next_iomem_res() 377 if (res->start < p->start) find_next_iomem_res() 378 res->start = p->start; find_next_iomem_res() 387 * All the memory ranges which overlap start,end and also match flags and 392 * @start: start addr 395 int walk_iomem_res(char *name, unsigned long flags, u64 start, u64 end, walk_iomem_res() argument 402 res.start = start; walk_iomem_res() 406 while ((res.start < res.end) && walk_iomem_res() 408 ret = (*func)(res.start, res.end, arg); walk_iomem_res() 411 res.start = res.end + 1; walk_iomem_res() 424 int walk_system_ram_res(u64 start, u64 end, void *arg, walk_system_ram_res() argument 431 res.start = start; walk_system_ram_res() 435 while ((res.start < res.end) && walk_system_ram_res() 437 ret = (*func)(res.start, res.end, arg); walk_system_ram_res() 440 res.start = res.end + 1; walk_system_ram_res() 461 res.start = (u64) start_pfn << PAGE_SHIFT; walk_system_ram_range() 465 while ((res.start < res.end) && walk_system_ram_range() 467 pfn = (res.start + PAGE_SIZE - 1) >> PAGE_SHIFT; walk_system_ram_range() 473 res.start = res.end + 1; walk_system_ram_range() 497 * @start: region start address 513 int region_intersects(resource_size_t start, size_t size, const char *name) region_intersects() argument 516 resource_size_t end = start + size - 1; region_intersects() 524 if (start >= p->start && start <= p->end) region_intersects() 526 if (end >= p->start && end <= p->end) region_intersects() 528 if (p->start >= start && p->end <= end) region_intersects() 551 return avail->start; simple_align_resource() 557 if (res->start < min) resource_clip() 558 res->start = min; resource_clip() 575 tmp.start = root->start; __find_resource() 578 * of this->start - 1 to tmp->end below would cause an underflow. __find_resource() 580 if (this && this->start == root->start) { __find_resource() 581 tmp.start = (this == old) ? old->start : this->end + 1; __find_resource() 586 tmp.end = (this == old) ? this->end : this->start - 1; __find_resource() 590 if (tmp.end < tmp.start) __find_resource() 597 avail.start = ALIGN(tmp.start, constraint->align); __find_resource() 600 if (avail.start >= tmp.start) { __find_resource() 602 alloc.start = constraint->alignf(constraint->alignf_data, &avail, __find_resource() 604 alloc.end = alloc.start + size - 1; __find_resource() 606 new->start = alloc.start; __find_resource() 616 tmp.start = this->end + 1; __find_resource() 656 old->start = new.start; reallocate_resource() 667 old->start = new.start; reallocate_resource() 731 * lookup_resource - find an existing resource by a resource start address 733 * @start: resource start address 737 struct resource *lookup_resource(struct resource *root, resource_size_t start) lookup_resource() argument 743 if (res->start == start) lookup_resource() 769 if ((first->start > new->start) || (first->end < new->end)) __insert_resource() 771 if ((first->start == new->start) && (first->end == new->end)) __insert_resource() 777 if (next->start < new->start || next->end > new->end) __insert_resource() 781 if (next->sibling->start > new->end) __insert_resource() 866 if (conflict->start < new->start) insert_resource_expand_to_fit() 867 new->start = conflict->start; insert_resource_expand_to_fit() 876 static int __adjust_resource(struct resource *res, resource_size_t start, __adjust_resource() argument 880 resource_size_t end = start + size - 1; __adjust_resource() 886 if ((start < parent->start) || (end > parent->end)) __adjust_resource() 889 if (res->sibling && (res->sibling->start <= end)) __adjust_resource() 896 if (start <= tmp->end) __adjust_resource() 902 if ((tmp->start < start) || (tmp->end > end)) __adjust_resource() 905 res->start = start; __adjust_resource() 914 * adjust_resource - modify a resource's start and size 916 * @start: new start value 919 * Given an existing resource, change its start and size to match the 923 int adjust_resource(struct resource *res, resource_size_t start, adjust_resource() argument 929 result = __adjust_resource(res, start, size); adjust_resource() 936 resource_size_t start, resource_size_t end, __reserve_region_with_split() 948 res->start = start; __reserve_region_with_split() 964 if (conflict->start <= res->start && __reserve_region_with_split() 972 if (conflict->start > res->start) { __reserve_region_with_split() 974 res->end = conflict->start - 1; __reserve_region_with_split() 982 next_res->start = conflict->end + 1; __reserve_region_with_split() 987 res->start = conflict->end + 1; __reserve_region_with_split() 994 resource_size_t start, resource_size_t end, reserve_region_with_split() 1000 if (root->start > start || root->end < end) { reserve_region_with_split() 1002 (unsigned long long)start, (unsigned long long)end, reserve_region_with_split() 1004 if (start > root->end || end < root->start) reserve_region_with_split() 1009 if (start < root->start) reserve_region_with_split() 1010 start = root->start; reserve_region_with_split() 1012 (unsigned long long)start, reserve_region_with_split() 1018 __reserve_region_with_split(root, start, end, name); reserve_region_with_split() 1034 return res->start; resource_alignment() 1056 * @start: resource start address 1062 resource_size_t start, resource_size_t n, __request_region() 1072 res->start = start; __request_region() 1073 res->end = start + n - 1; __request_region() 1113 * @start: resource start address 1118 void __release_region(struct resource *parent, resource_size_t start, __release_region() argument 1125 end = start + n - 1; __release_region() 1134 if (res->start <= start && res->end >= end) { __release_region() 1139 if (res->start != start || res->end != end) __release_region() 1154 "<%016llx-%016llx>\n", (unsigned long long)start, __release_region() 1163 * @start: resource start address 1181 resource_size_t start, resource_size_t size) release_mem_region_adjustable() 1189 end = start + size - 1; release_mem_region_adjustable() 1190 if ((start < parent->start) || (end > parent->end)) release_mem_region_adjustable() 1200 if (res->start >= end) release_mem_region_adjustable() 1204 if (res->start > start || res->end < end) { release_mem_region_adjustable() 1218 if (res->start == start && res->end == end) { release_mem_region_adjustable() 1223 } else if (res->start == start && res->end != end) { release_mem_region_adjustable() 1224 /* adjust the start */ release_mem_region_adjustable() 1227 } else if (res->start != start && res->end == end) { release_mem_region_adjustable() 1229 ret = __adjust_resource(res, res->start, release_mem_region_adjustable() 1230 start - res->start); release_mem_region_adjustable() 1238 new_res->start = end + 1; release_mem_region_adjustable() 1245 ret = __adjust_resource(res, res->start, release_mem_region_adjustable() 1246 start - res->start); release_mem_region_adjustable() 1337 resource_size_t start; member in struct:region_devres 1345 __release_region(this->parent, this->start, this->n); devm_region_release() 1353 this->start == match->start && this->n == match->n; devm_region_match() 1357 struct resource *parent, resource_size_t start, __devm_request_region() 1369 dr->start = start; __devm_request_region() 1372 res = __request_region(parent, start, n, name, 0); __devm_request_region() 1383 resource_size_t start, resource_size_t n) __devm_release_region() 1385 struct region_devres match_data = { parent, start, n }; __devm_release_region() 1387 __release_region(parent, start, n); __devm_release_region() 1413 res->start = io_start; reserve_setup() 1417 if (request_resource(res->start >= 0x10000 ? &iomem_resource : &ioport_resource, res) == 0) reserve_setup() 1442 if (p->start >= addr + size) iomem_map_sanity_check() 1446 if (PFN_DOWN(p->start) <= PFN_DOWN(addr) && iomem_map_sanity_check() 1498 if (p->start >= addr + size) iomem_is_exclusive() 935 __reserve_region_with_split(struct resource *root, resource_size_t start, resource_size_t end, const char *name) __reserve_region_with_split() argument 993 reserve_region_with_split(struct resource *root, resource_size_t start, resource_size_t end, const char *name) reserve_region_with_split() argument 1061 __request_region(struct resource *parent, resource_size_t start, resource_size_t n, const char *name, int flags) __request_region() argument 1180 release_mem_region_adjustable(struct resource *parent, resource_size_t start, resource_size_t size) release_mem_region_adjustable() argument 1356 __devm_request_region(struct device *dev, struct resource *parent, resource_size_t start, resource_size_t n, const char *name) __devm_request_region() argument 1382 __devm_release_region(struct device *dev, struct resource *parent, resource_size_t start, resource_size_t n) __devm_release_region() argument
|
/linux-4.4.14/tools/perf/ |
H A D | perf-read-vdso.c | 14 void *start, *end; main() local 17 if (find_vdso_map(&start, &end)) main() 20 size = end - start; main() 23 written = fwrite(start, 1, size, stdout); main() 26 start += written; main()
|
/linux-4.4.14/arch/mn10300/include/asm/ |
H A D | cacheflush.h | 24 extern void mn10300_local_icache_inv_page(unsigned long start); 25 extern void mn10300_local_icache_inv_range(unsigned long start, unsigned long end); 26 extern void mn10300_local_icache_inv_range2(unsigned long start, unsigned long size); 28 extern void mn10300_local_dcache_inv_page(unsigned long start); 29 extern void mn10300_local_dcache_inv_range(unsigned long start, unsigned long end); 30 extern void mn10300_local_dcache_inv_range2(unsigned long start, unsigned long size); 32 extern void mn10300_icache_inv_page(unsigned long start); 33 extern void mn10300_icache_inv_range(unsigned long start, unsigned long end); 34 extern void mn10300_icache_inv_range2(unsigned long start, unsigned long size); 36 extern void mn10300_dcache_inv_page(unsigned long start); 37 extern void mn10300_dcache_inv_range(unsigned long start, unsigned long end); 38 extern void mn10300_dcache_inv_range2(unsigned long start, unsigned long size); 41 extern void mn10300_local_dcache_flush_page(unsigned long start); 42 extern void mn10300_local_dcache_flush_range(unsigned long start, unsigned long end); 43 extern void mn10300_local_dcache_flush_range2(unsigned long start, unsigned long size); 45 extern void mn10300_local_dcache_flush_inv_page(unsigned long start); 46 extern void mn10300_local_dcache_flush_inv_range(unsigned long start, unsigned long end); 47 extern void mn10300_local_dcache_flush_inv_range2(unsigned long start, unsigned long size); 49 extern void mn10300_dcache_flush_page(unsigned long start); 50 extern void mn10300_dcache_flush_range(unsigned long start, unsigned long end); 51 extern void mn10300_dcache_flush_range2(unsigned long start, unsigned long size); 53 extern void mn10300_dcache_flush_inv_page(unsigned long start); 54 extern void mn10300_dcache_flush_inv_range(unsigned long start, unsigned long end); 55 extern void mn10300_dcache_flush_inv_range2(unsigned long start, unsigned long size); 58 #define mn10300_local_dcache_flush_page(start) do {} while (0) 59 #define mn10300_local_dcache_flush_range(start, end) do {} while (0) 60 #define mn10300_local_dcache_flush_range2(start, size) do {} while (0) 63 #define mn10300_local_dcache_flush_inv_page(start) \ 64 mn10300_local_dcache_inv_page(start) 65 #define mn10300_local_dcache_flush_inv_range(start, end) \ 66 mn10300_local_dcache_inv_range(start, end) 67 #define mn10300_local_dcache_flush_inv_range2(start, size) \ 68 mn10300_local_dcache_inv_range2(start, size) 70 #define mn10300_dcache_flush_page(start) do {} while (0) 71 #define mn10300_dcache_flush_range(start, end) do {} while (0) 72 #define mn10300_dcache_flush_range2(start, size) do {} while (0) 74 #define mn10300_dcache_flush_inv_page(start) \ 75 mn10300_dcache_inv_page((start)) 76 #define mn10300_dcache_flush_inv_range(start, end) \ 77 mn10300_dcache_inv_range((start), (end)) 78 #define mn10300_dcache_flush_inv_range2(start, size) \ 79 mn10300_dcache_inv_range2((start), (size)) 83 #define mn10300_local_icache_inv_page(start) do {} while (0) 84 #define mn10300_local_icache_inv_range(start, end) do {} while (0) 85 #define mn10300_local_icache_inv_range2(start, size) do {} while (0) 87 #define mn10300_local_dcache_inv_page(start) do {} while (0) 88 #define mn10300_local_dcache_inv_range(start, end) do {} while (0) 89 #define mn10300_local_dcache_inv_range2(start, size) do {} while (0) 91 #define mn10300_local_dcache_flush_inv_page(start) do {} while (0) 93 #define mn10300_local_dcache_flush_inv_range(start, end)do {} while (0) 94 #define mn10300_local_dcache_flush_inv_range2(start, size) do {} while (0) 95 #define mn10300_local_dcache_flush_page(start) do {} while (0) 96 #define mn10300_local_dcache_flush_range(start, end) do {} while (0) 97 #define mn10300_local_dcache_flush_range2(start, size) do {} while (0) 99 #define mn10300_icache_inv_page(start) do {} while (0) 100 #define mn10300_icache_inv_range(start, end) do {} while (0) 101 #define mn10300_icache_inv_range2(start, size) do {} while (0) 103 #define mn10300_dcache_inv_page(start) do {} while (0) 104 #define mn10300_dcache_inv_range(start, end) do {} while (0) 105 #define mn10300_dcache_inv_range2(start, size) do {} while (0) 107 #define mn10300_dcache_flush_inv_page(start) do {} while (0) 109 #define mn10300_dcache_flush_inv_range(start, end) do {} while (0) 110 #define mn10300_dcache_flush_inv_range2(start, size) do {} while (0) 111 #define mn10300_dcache_flush_page(start) do {} while (0) 112 #define mn10300_dcache_flush_range(start, end) do {} while (0) 113 #define mn10300_dcache_flush_range2(start, size) do {} while (0) 122 #define flush_cache_range(mm, start, end) do {} while (0) 124 #define flush_cache_vmap(start, end) do {} while (0) 125 #define flush_cache_vunmap(start, end) do {} while (0) 136 extern void flush_icache_range(unsigned long start, unsigned long end); 143 extern void flush_icache_range(unsigned long start, unsigned long end); 145 #define flush_icache_range(start, end) do {} while (0)
|
H A D | termios.h | 8 start=^Q stop=^S susp=^Z eol=\0
|
/linux-4.4.14/arch/hexagon/mm/ |
H A D | cache.c | 25 #define spanlines(start, end) \ 26 (((end - (start & ~(LINESIZE - 1))) >> LINEBITS) + 1) 28 void flush_dcache_range(unsigned long start, unsigned long end) flush_dcache_range() argument 30 unsigned long lines = spanlines(start, end-1); flush_dcache_range() 33 start &= ~(LINESIZE - 1); flush_dcache_range() 41 : "r" (start) flush_dcache_range() 43 start += LINESIZE; flush_dcache_range() 48 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument 50 unsigned long lines = spanlines(start, end-1); flush_icache_range() 53 start &= ~(LINESIZE - 1); flush_icache_range() 62 : "r" (start) flush_icache_range() 64 start += LINESIZE; flush_icache_range() 73 void hexagon_clean_dcache_range(unsigned long start, unsigned long end) hexagon_clean_dcache_range() argument 75 unsigned long lines = spanlines(start, end-1); hexagon_clean_dcache_range() 78 start &= ~(LINESIZE - 1); hexagon_clean_dcache_range() 86 : "r" (start) hexagon_clean_dcache_range() 88 start += LINESIZE; hexagon_clean_dcache_range() 93 void hexagon_inv_dcache_range(unsigned long start, unsigned long end) hexagon_inv_dcache_range() argument 95 unsigned long lines = spanlines(start, end-1); hexagon_inv_dcache_range() 98 start &= ~(LINESIZE - 1); hexagon_inv_dcache_range() 106 : "r" (start) hexagon_inv_dcache_range() 108 start += LINESIZE; hexagon_inv_dcache_range()
|
H A D | strnlen_user.S | 28 #define start r7 define 49 start = isrc; define 60 tmp1 = memb(start++#1); 65 P1 = cmp.gtu(end,start); 66 mod8 = and(start,#7); 80 dbuf = memd(start); 81 obo = add(start,#1); 88 P0 = cmp.gtu(end,start); 98 if (!P0.new) start = add(obo,tmp1); 101 start = add(start,#8); define 110 start = add(obo,tmp1); define 115 R0 = sub(start,isrc);
|
/linux-4.4.14/mm/ |
H A D | msync.c | 20 * MS_ASYNC does not start I/O (it used to, up to 2.5.67). 26 * Or the application may run fadvise(FADV_DONTNEED) against the fd to start 31 SYSCALL_DEFINE3(msync, unsigned long, start, size_t, len, int, flags) SYSCALL_DEFINE3() 41 if (offset_in_page(start)) SYSCALL_DEFINE3() 47 end = start + len; SYSCALL_DEFINE3() 48 if (end < start) SYSCALL_DEFINE3() 51 if (end == start) SYSCALL_DEFINE3() 54 * If the interval [start,end) covers some unmapped address ranges, SYSCALL_DEFINE3() 58 vma = find_vma(mm, start); SYSCALL_DEFINE3() 63 /* Still start < end. */ SYSCALL_DEFINE3() 67 /* Here start < vma->vm_end. */ SYSCALL_DEFINE3() 68 if (start < vma->vm_start) { SYSCALL_DEFINE3() 69 start = vma->vm_start; SYSCALL_DEFINE3() 70 if (start >= end) SYSCALL_DEFINE3() 74 /* Here vma->vm_start <= start < vma->vm_end. */ SYSCALL_DEFINE3() 81 fstart = (start - vma->vm_start) + SYSCALL_DEFINE3() 83 fend = fstart + (min(end, vma->vm_end) - start) - 1; SYSCALL_DEFINE3() 84 start = vma->vm_end; SYSCALL_DEFINE3() 91 if (error || start >= end) SYSCALL_DEFINE3() 94 vma = find_vma(mm, start); SYSCALL_DEFINE3() 96 if (start >= end) { SYSCALL_DEFINE3()
|
H A D | memtest.c | 36 u64 *p, *start, *end; memtest() local 42 start = __va(start_phys_aligned); memtest() 43 end = start + (size - (start_phys_aligned - start_phys)) / incr; memtest() 47 for (p = start; p < end; p++) memtest() 50 for (p = start; p < end; p++, start_phys_aligned += incr) { memtest() 65 static void __init do_one_pass(u64 pattern, phys_addr_t start, phys_addr_t end) do_one_pass() argument 72 this_start = clamp(this_start, start, end); do_one_pass() 73 this_end = clamp(this_end, start, end); do_one_pass() 99 void __init early_memtest(phys_addr_t start, phys_addr_t end) early_memtest() argument 110 do_one_pass(patterns[idx], start, end); early_memtest()
|
H A D | madvise.c | 48 unsigned long start, unsigned long end, int behavior) madvise_behavior() 87 error = ksm_madvise(vma, start, end, behavior, &new_flags); madvise_behavior() 104 pgoff = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT); madvise_behavior() 105 *prev = vma_merge(mm, *prev, start, end, new_flags, vma->anon_vma, madvise_behavior() 115 if (start != vma->vm_start) { madvise_behavior() 116 error = split_vma(mm, vma, start, 1); madvise_behavior() 140 static int swapin_walk_pmd_entry(pmd_t *pmd, unsigned long start, swapin_walk_pmd_entry() argument 150 for (index = start; index != end; index += PAGE_SIZE) { swapin_walk_pmd_entry() 156 orig_pte = pte_offset_map_lock(vma->vm_mm, pmd, start, &ptl); swapin_walk_pmd_entry() 157 pte = *(orig_pte + ((index - start) / PAGE_SIZE)); swapin_walk_pmd_entry() 176 unsigned long start, unsigned long end) force_swapin_readahead() 184 walk_page_range(start, end, &walk); force_swapin_readahead() 190 unsigned long start, unsigned long end, force_shm_swapin_readahead() 197 for (; start < end; start += PAGE_SIZE) { force_shm_swapin_readahead() 198 index = ((start - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; force_shm_swapin_readahead() 222 unsigned long start, unsigned long end) madvise_willneed() 229 force_swapin_readahead(vma, start, end); madvise_willneed() 235 force_shm_swapin_readahead(vma, start, end, madvise_willneed() 250 start = ((start - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; madvise_willneed() 255 force_page_cache_readahead(file->f_mapping, file, start, end - start); madvise_willneed() 280 unsigned long start, unsigned long end) madvise_dontneed() 286 zap_page_range(vma, start, end - start, NULL); madvise_dontneed() 296 unsigned long start, unsigned long end) madvise_remove() 316 offset = (loff_t)(start - vma->vm_start) madvise_remove() 329 offset, end - start); madvise_remove() 339 static int madvise_hwpoison(int bhv, unsigned long start, unsigned long end) madvise_hwpoison() argument 344 for (; start < end; start += PAGE_SIZE << madvise_hwpoison() 348 ret = get_user_pages_fast(start, 1, 0, &p); madvise_hwpoison() 358 page_to_pfn(p), start); madvise_hwpoison() 365 page_to_pfn(p), start); madvise_hwpoison() 375 unsigned long start, unsigned long end, int behavior) madvise_vma() 379 return madvise_remove(vma, prev, start, end); madvise_vma() 381 return madvise_willneed(vma, prev, start, end); madvise_vma() 383 return madvise_dontneed(vma, prev, start, end); madvise_vma() 385 return madvise_behavior(vma, prev, start, end, behavior); madvise_vma() 451 * -EINVAL - start + len < 0, start is not page-aligned, 460 SYSCALL_DEFINE3(madvise, unsigned long, start, size_t, len_in, int, behavior) SYSCALL_DEFINE3() 472 return madvise_hwpoison(behavior, start, start+len_in); SYSCALL_DEFINE3() 477 if (start & ~PAGE_MASK) SYSCALL_DEFINE3() 485 end = start + len; SYSCALL_DEFINE3() 486 if (end < start) SYSCALL_DEFINE3() 490 if (end == start) SYSCALL_DEFINE3() 500 * If the interval [start,end) covers some unmapped address SYSCALL_DEFINE3() 504 vma = find_vma_prev(current->mm, start, &prev); SYSCALL_DEFINE3() 505 if (vma && start > vma->vm_start) SYSCALL_DEFINE3() 510 /* Still start < end. */ SYSCALL_DEFINE3() 515 /* Here start < (end|vma->vm_end). */ SYSCALL_DEFINE3() 516 if (start < vma->vm_start) { SYSCALL_DEFINE3() 518 start = vma->vm_start; SYSCALL_DEFINE3() 519 if (start >= end) SYSCALL_DEFINE3() 523 /* Here vma->vm_start <= start < (end|vma->vm_end) */ SYSCALL_DEFINE3() 528 /* Here vma->vm_start <= start < tmp <= (end|vma->vm_end). */ SYSCALL_DEFINE3() 529 error = madvise_vma(vma, &prev, start, tmp, behavior); SYSCALL_DEFINE3() 532 start = tmp; SYSCALL_DEFINE3() 533 if (prev && start < prev->vm_end) SYSCALL_DEFINE3() 534 start = prev->vm_end; SYSCALL_DEFINE3() 536 if (start >= end) SYSCALL_DEFINE3() 541 vma = find_vma(current->mm, start); SYSCALL_DEFINE3() 46 madvise_behavior(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) madvise_behavior() argument 175 force_swapin_readahead(struct vm_area_struct *vma, unsigned long start, unsigned long end) force_swapin_readahead() argument 189 force_shm_swapin_readahead(struct vm_area_struct *vma, unsigned long start, unsigned long end, struct address_space *mapping) force_shm_swapin_readahead() argument 220 madvise_willneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_willneed() argument 278 madvise_dontneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_dontneed() argument 294 madvise_remove(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_remove() argument 374 madvise_vma(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) madvise_vma() argument
|
H A D | debug-pagealloc.c | 83 unsigned char *start; check_poison_mem() local 86 start = memchr_inv(mem, PAGE_POISON, bytes); check_poison_mem() 87 if (!start) check_poison_mem() 90 for (end = mem + bytes - 1; end > start; end--) { check_poison_mem() 97 else if (start == end && single_bit_flip(*start, PAGE_POISON)) check_poison_mem() 102 print_hex_dump(KERN_ERR, "", DUMP_PREFIX_ADDRESS, 16, 1, start, check_poison_mem() 103 end - start + 1, 1); check_poison_mem()
|
/linux-4.4.14/arch/mips/bcm63xx/ |
H A D | dev-uart.c | 16 /* start & end filled at runtime */ 20 /* start filled at runtime */ 27 /* start & end filled at runtime */ 31 /* start filled at runtime */ 62 uart0_resources[0].start = bcm63xx_regset_address(RSET_UART0); bcm63xx_uart_register() 63 uart0_resources[0].end = uart0_resources[0].start + bcm63xx_uart_register() 65 uart0_resources[1].start = bcm63xx_get_irq_number(IRQ_UART0); bcm63xx_uart_register() 69 uart1_resources[0].start = bcm63xx_regset_address(RSET_UART1); bcm63xx_uart_register() 70 uart1_resources[0].end = uart1_resources[0].start + bcm63xx_uart_register() 72 uart1_resources[1].start = bcm63xx_get_irq_number(IRQ_UART1); bcm63xx_uart_register()
|
H A D | dev-enet.c | 48 .start = -1, /* filled at runtime */ 53 .start = -1, /* filled at runtime */ 58 .start = -1, /* filled at runtime */ 75 .start = -1, /* filled at runtime */ 80 .start = -1, /* filled at runtime */ 84 .start = -1, /* filled at runtime */ 88 .start = -1, /* filled at runtime */ 107 .start = -1, /* filled at runtime */ 112 .start = -1, /* filled at runtime */ 116 .start = -1, /* filled at runtime */ 120 .start = -1, /* filled at runtime */ 139 /* start & end filled at runtime */ 143 /* start filled at runtime */ 147 /* start filled at runtime */ 172 shared_res[0].start = bcm63xx_regset_address(RSET_ENETDMA); register_shared() 173 shared_res[0].end = shared_res[0].start; register_shared() 186 shared_res[1].start = bcm63xx_regset_address(RSET_ENETDMAC); register_shared() 187 shared_res[1].end = shared_res[1].start; register_shared() 190 shared_res[2].start = bcm63xx_regset_address(RSET_ENETDMAS); register_shared() 191 shared_res[2].end = shared_res[2].start; register_shared() 220 enet0_res[0].start = bcm63xx_regset_address(RSET_ENET0); bcm63xx_enet_register() 221 enet0_res[0].end = enet0_res[0].start; bcm63xx_enet_register() 223 enet0_res[1].start = bcm63xx_get_irq_number(IRQ_ENET0); bcm63xx_enet_register() 224 enet0_res[2].start = bcm63xx_get_irq_number(IRQ_ENET0_RXDMA); bcm63xx_enet_register() 225 enet0_res[3].start = bcm63xx_get_irq_number(IRQ_ENET0_TXDMA); bcm63xx_enet_register() 228 enet1_res[0].start = bcm63xx_regset_address(RSET_ENET1); bcm63xx_enet_register() 229 enet1_res[0].end = enet1_res[0].start; bcm63xx_enet_register() 231 enet1_res[1].start = bcm63xx_get_irq_number(IRQ_ENET1); bcm63xx_enet_register() 232 enet1_res[2].start = bcm63xx_get_irq_number(IRQ_ENET1_RXDMA); bcm63xx_enet_register() 233 enet1_res[3].start = bcm63xx_get_irq_number(IRQ_ENET1_TXDMA); bcm63xx_enet_register() 286 enetsw_res[0].start = bcm63xx_regset_address(RSET_ENETSW); bcm63xx_enetsw_register() 287 enetsw_res[0].end = enetsw_res[0].start; bcm63xx_enetsw_register() 289 enetsw_res[1].start = bcm63xx_get_irq_number(IRQ_ENETSW_RXDMA0); bcm63xx_enetsw_register() 290 enetsw_res[2].start = bcm63xx_get_irq_number(IRQ_ENETSW_TXDMA0); bcm63xx_enetsw_register() 291 if (!enetsw_res[2].start) bcm63xx_enetsw_register() 292 enetsw_res[2].start = -1; bcm63xx_enetsw_register()
|
H A D | dev-hsspi.c | 19 .start = -1, /* filled at runtime */ 24 .start = -1, /* filled at runtime */ 41 spi_resources[0].start = bcm63xx_regset_address(RSET_HSSPI); bcm63xx_hsspi_register() 42 spi_resources[0].end = spi_resources[0].start; bcm63xx_hsspi_register() 44 spi_resources[1].start = bcm63xx_get_irq_number(IRQ_HSSPI); bcm63xx_hsspi_register()
|
H A D | dev-spi.c | 23 .start = -1, /* filled at runtime */ 28 .start = -1, /* filled at runtime */ 44 spi_resources[0].start = bcm63xx_regset_address(RSET_SPI); bcm63xx_spi_register() 45 spi_resources[0].end = spi_resources[0].start; bcm63xx_spi_register() 46 spi_resources[1].start = bcm63xx_get_irq_number(IRQ_SPI); bcm63xx_spi_register()
|
H A D | dev-usb-usbd.c | 47 usbd_resources[0].start = bcm63xx_regset_address(RSET_USBD); bcm63xx_usbd_register() 48 usbd_resources[0].end = usbd_resources[0].start + RSET_USBD_SIZE - 1; bcm63xx_usbd_register() 51 usbd_resources[1].start = bcm63xx_regset_address(RSET_USBDMA); bcm63xx_usbd_register() 52 usbd_resources[1].end = usbd_resources[1].start + RSET_USBDMA_SIZE - 1; bcm63xx_usbd_register() 58 r->start = r->end = bcm63xx_get_irq_number(irq_list[i]); bcm63xx_usbd_register()
|
H A D | dev-pcmcia.c | 22 /* start & end filled at runtime */ 28 .start = BCM_PCMCIA_COMMON_BASE_PA, 33 .start = BCM_PCMCIA_ATTR_BASE_PA, 38 .start = BCM_PCMCIA_IO_BASE_PA, 45 /* start filled at runtime */ 51 .start = BCM_PCMCIA_IO_BASE_PA, 125 pcmcia_resources[0].start = bcm63xx_regset_address(RSET_PCMCIA); bcm63xx_pcmcia_register() 126 pcmcia_resources[0].end = pcmcia_resources[0].start + bcm63xx_pcmcia_register() 128 pcmcia_resources[4].start = bcm63xx_get_irq_number(IRQ_PCMCIA); bcm63xx_pcmcia_register()
|
H A D | dev-rng.c | 16 .start = -1, /* filled at runtime */ 34 rng_resources[0].start = bcm63xx_regset_address(RSET_RNG); bcm63xx_rng_register() 35 rng_resources[0].end = rng_resources[0].start; bcm63xx_rng_register()
|
H A D | dev-wdt.c | 16 .start = -1, /* filled at runtime */ 31 wdt_resources[0].start = bcm63xx_regset_address(RSET_WDT); bcm63xx_wdt_register() 32 wdt_resources[0].end = wdt_resources[0].start; bcm63xx_wdt_register()
|
H A D | dev-dsp.c | 22 .start = -1, /* filled at runtime */ 27 .start = -1, /* filled at runtime */ 47 voip_dsp_resources[0].start = val; bcm63xx_dsp_register() 49 voip_dsp_resources[1].start = pd->ext_irq; bcm63xx_dsp_register()
|
/linux-4.4.14/arch/mn10300/mm/ |
H A D | cache-smp-flush.c | 32 * @start: The address of the page of memory to be flushed. 37 void mn10300_dcache_flush_page(unsigned long start) mn10300_dcache_flush_page() argument 41 start &= ~(PAGE_SIZE-1); mn10300_dcache_flush_page() 44 mn10300_local_dcache_flush_page(start); mn10300_dcache_flush_page() 45 smp_cache_call(SMP_DCACHE_FLUSH_RANGE, start, start + PAGE_SIZE); mn10300_dcache_flush_page() 51 * @start: The start address of the region to be flushed. 54 * Flush a range of addresses in the data cache on all CPUs, between start and 57 void mn10300_dcache_flush_range(unsigned long start, unsigned long end) mn10300_dcache_flush_range() argument 62 mn10300_local_dcache_flush_range(start, end); mn10300_dcache_flush_range() 63 smp_cache_call(SMP_DCACHE_FLUSH_RANGE, start, end); mn10300_dcache_flush_range() 69 * @start: The start address of the region to be flushed. 72 * Flush a range of addresses in the data cache on all CPUs, between start and 73 * start+size-1 inclusive. 75 void mn10300_dcache_flush_range2(unsigned long start, unsigned long size) mn10300_dcache_flush_range2() argument 80 mn10300_local_dcache_flush_range2(start, size); mn10300_dcache_flush_range2() 81 smp_cache_call(SMP_DCACHE_FLUSH_RANGE, start, start + size); mn10300_dcache_flush_range2() 103 * @start: The address of the page of memory to be flushed and invalidated. 108 void mn10300_dcache_flush_inv_page(unsigned long start) mn10300_dcache_flush_inv_page() argument 112 start &= ~(PAGE_SIZE-1); mn10300_dcache_flush_inv_page() 115 mn10300_local_dcache_flush_inv_page(start); mn10300_dcache_flush_inv_page() 116 smp_cache_call(SMP_DCACHE_FLUSH_INV_RANGE, start, start + PAGE_SIZE); mn10300_dcache_flush_inv_page() 123 * @start: The start address of the region to be flushed and invalidated. 127 * between start and end-1 inclusive. 129 void mn10300_dcache_flush_inv_range(unsigned long start, unsigned long end) mn10300_dcache_flush_inv_range() argument 134 mn10300_local_dcache_flush_inv_range(start, end); mn10300_dcache_flush_inv_range() 135 smp_cache_call(SMP_DCACHE_FLUSH_INV_RANGE, start, end); mn10300_dcache_flush_inv_range() 142 * @start: The start address of the region to be flushed and invalidated. 146 * between start and start+size-1 inclusive. 148 void mn10300_dcache_flush_inv_range2(unsigned long start, unsigned long size) mn10300_dcache_flush_inv_range2() argument 153 mn10300_local_dcache_flush_inv_range2(start, size); mn10300_dcache_flush_inv_range2() 154 smp_cache_call(SMP_DCACHE_FLUSH_INV_RANGE, start, start + size); mn10300_dcache_flush_inv_range2()
|
H A D | cache-smp-inv.c | 32 * @start: The address of the page of memory to be invalidated. 37 void mn10300_icache_inv_page(unsigned long start) mn10300_icache_inv_page() argument 41 start &= ~(PAGE_SIZE-1); mn10300_icache_inv_page() 44 mn10300_local_icache_inv_page(start); mn10300_icache_inv_page() 45 smp_cache_call(SMP_ICACHE_INV_RANGE, start, start + PAGE_SIZE); mn10300_icache_inv_page() 51 * @start: The start address of the region to be invalidated. 55 * between start and end-1 inclusive. 57 void mn10300_icache_inv_range(unsigned long start, unsigned long end) mn10300_icache_inv_range() argument 62 mn10300_local_icache_inv_range(start, end); mn10300_icache_inv_range() 63 smp_cache_call(SMP_ICACHE_INV_RANGE, start, end); mn10300_icache_inv_range() 69 * @start: The start address of the region to be invalidated. 73 * between start and start+size-1 inclusive. 75 void mn10300_icache_inv_range2(unsigned long start, unsigned long size) mn10300_icache_inv_range2() argument 80 mn10300_local_icache_inv_range2(start, size); mn10300_icache_inv_range2() 81 smp_cache_call(SMP_ICACHE_INV_RANGE, start, start + size); mn10300_icache_inv_range2() 102 * @start: The address of the page of memory to be invalidated. 107 void mn10300_dcache_inv_page(unsigned long start) mn10300_dcache_inv_page() argument 111 start &= ~(PAGE_SIZE-1); mn10300_dcache_inv_page() 114 mn10300_local_dcache_inv_page(start); mn10300_dcache_inv_page() 115 smp_cache_call(SMP_DCACHE_INV_RANGE, start, start + PAGE_SIZE); mn10300_dcache_inv_page() 121 * @start: The start address of the region to be invalidated. 124 * Invalidate a range of addresses in the data cache on all CPUs, between start 127 void mn10300_dcache_inv_range(unsigned long start, unsigned long end) mn10300_dcache_inv_range() argument 132 mn10300_local_dcache_inv_range(start, end); mn10300_dcache_inv_range() 133 smp_cache_call(SMP_DCACHE_INV_RANGE, start, end); mn10300_dcache_inv_range() 139 * @start: The start address of the region to be invalidated. 142 * Invalidate a range of addresses in the data cache on all CPUs, between start 143 * and start+size-1 inclusive. 145 void mn10300_dcache_inv_range2(unsigned long start, unsigned long size) mn10300_dcache_inv_range2() argument 150 mn10300_local_dcache_inv_range2(start, size); mn10300_dcache_inv_range2() 151 smp_cache_call(SMP_DCACHE_INV_RANGE, start, start + size); mn10300_dcache_inv_range2()
|
H A D | cache-flush-icache.c | 27 unsigned long start = page_to_phys(page); flush_icache_page() local 32 mn10300_local_dcache_flush_page(start); flush_icache_page() 33 mn10300_local_icache_inv_page(start); flush_icache_page() 35 smp_cache_call(SMP_IDCACHE_INV_FLUSH_RANGE, start, start + PAGE_SIZE); flush_icache_page() 43 * @start: The starting virtual address of the page part. 50 static void flush_icache_page_range(unsigned long start, unsigned long end) flush_icache_page_range() argument 60 off = start & ~PAGE_MASK; flush_icache_page_range() 61 size = end - start; flush_icache_page_range() 65 pgd = pgd_offset(current->mm, start); flush_icache_page_range() 69 pud = pud_offset(pgd, start); flush_icache_page_range() 73 pmd = pmd_offset(pud, start); flush_icache_page_range() 77 ppte = pte_offset_map(pmd, start); flush_icache_page_range() 96 smp_cache_call(SMP_IDCACHE_INV_FLUSH_RANGE, start, end); flush_icache_page_range() 101 * @start: The starting virtual address of the region. 108 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument 119 if (start >= end) flush_icache_range() 126 start_page = (start >= 0x80000000UL) ? start : 0x80000000UL; flush_icache_range() 130 if (start_page == start) flush_icache_range() 135 start_page = start & PAGE_MASK; flush_icache_range() 140 flush_icache_page_range(start, end); flush_icache_range() 143 flush_icache_page_range(start, end_page); flush_icache_range()
|
H A D | cache-inv-icache.c | 21 * @start: The starting virtual address of the page part. 28 static void flush_icache_page_range(unsigned long start, unsigned long end) flush_icache_page_range() argument 38 off = start & ~PAGE_MASK; flush_icache_page_range() 39 size = end - start; flush_icache_page_range() 43 pgd = pgd_offset(current->mm, start); flush_icache_page_range() 47 pud = pud_offset(pgd, start); flush_icache_page_range() 51 pmd = pmd_offset(pud, start); flush_icache_page_range() 55 ppte = pte_offset_map(pmd, start); flush_icache_page_range() 72 smp_cache_call(SMP_ICACHE_INV_RANGE, start, end); flush_icache_page_range() 77 * @start: The starting virtual address of the region. 84 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument 95 if (start >= end) flush_icache_range() 102 start_page = (start >= 0x80000000UL) ? start : 0x80000000UL; flush_icache_range() 104 smp_cache_call(SMP_ICACHE_INV_RANGE, start, end); flush_icache_range() 105 if (start_page == start) flush_icache_range() 110 start_page = start & PAGE_MASK; flush_icache_range() 115 flush_icache_page_range(start, end); flush_icache_range() 118 flush_icache_page_range(start, end_page); flush_icache_range()
|
/linux-4.4.14/arch/x86/kernel/ |
H A D | resource.c | 4 static void resource_clip(struct resource *res, resource_size_t start, resource_clip() argument 9 if (res->end < start || res->start > end) resource_clip() 12 if (res->start < start) resource_clip() 13 low = start - res->start; resource_clip() 20 res->end = start - 1; resource_clip() 22 res->start = end + 1; resource_clip()
|
H A D | probe_roms.c | 26 .start = 0xf0000, 33 .start = 0xe0000, 40 .start = 0xc8000, 45 .start = 0, 50 .start = 0, 55 .start = 0, 60 .start = 0, 65 .start = 0, 72 .start = 0xc0000, 126 rom = isa_bus_to_virt(res->start); find_oprom() 160 return ioremap(oprom->start, resource_size(oprom)); pci_map_biosrom() 200 unsigned long start, length, upper; probe_roms() local 205 upper = adapter_rom_resources[0].start; probe_roms() 206 for (start = video_rom_resource.start; start < upper; start += 2048) { probe_roms() 207 rom = isa_bus_to_virt(start); probe_roms() 211 video_rom_resource.start = start; probe_roms() 221 video_rom_resource.end = start + length - 1; probe_roms() 227 start = (video_rom_resource.end + 1 + 2047) & ~2047UL; probe_roms() 228 if (start < upper) probe_roms() 229 start = upper; probe_roms() 233 upper = system_rom_resource.start; probe_roms() 236 rom = isa_bus_to_virt(extension_rom_resource.start); probe_roms() 241 upper = extension_rom_resource.start; probe_roms() 246 for (i = 0; i < ARRAY_SIZE(adapter_rom_resources) && start < upper; start += 2048) { probe_roms() 247 rom = isa_bus_to_virt(start); probe_roms() 258 if (!length || start + length > upper || !romchecksum(rom, length)) probe_roms() 261 adapter_rom_resources[i].start = start; probe_roms() 262 adapter_rom_resources[i].end = start + length - 1; probe_roms() 265 start = adapter_rom_resources[i++].end & ~2047UL; probe_roms()
|
H A D | pci-iommu_table.c | 10 find_dependents_of(struct iommu_table_entry *start, find_dependents_of() argument 19 for (p = start; p < finish; p++) find_dependents_of() 27 void __init sort_iommu_table(struct iommu_table_entry *start, sort_iommu_table() argument 32 for (p = start; p < finish; p++) { sort_iommu_table() 34 q = find_dependents_of(start, finish, p); sort_iommu_table() 49 void __init check_iommu_entries(struct iommu_table_entry *start, check_iommu_entries() argument 55 for (p = start; p < finish; p++) { check_iommu_entries() 56 q = find_dependents_of(start, finish, p); check_iommu_entries() 57 x = find_dependents_of(start, finish, q); check_iommu_entries() 66 for (p = start; p < finish; p++) { check_iommu_entries() 75 inline void check_iommu_entries(struct iommu_table_entry *start, check_iommu_entries() argument
|
H A D | check.c | 73 phys_addr_t start, end; setup_bios_corruption_check() local 94 for_each_free_mem_range(i, NUMA_NO_NODE, MEMBLOCK_NONE, &start, &end, setup_bios_corruption_check() 96 start = clamp_t(phys_addr_t, round_up(start, PAGE_SIZE), setup_bios_corruption_check() 100 if (start >= end) setup_bios_corruption_check() 103 memblock_reserve(start, end - start); setup_bios_corruption_check() 104 scan_areas[num_scan_areas].addr = start; setup_bios_corruption_check() 105 scan_areas[num_scan_areas].size = end - start; setup_bios_corruption_check() 108 memset(__va(start), 0, end - start); setup_bios_corruption_check()
|
/linux-4.4.14/arch/xtensa/mm/ |
H A D | init.c | 43 sysmem.bank[i].start, sysmem.bank[i].end, sysmem_dump() 44 (sysmem.bank[i].end - sysmem.bank[i].start) >> 10); sysmem_dump() 48 * Find bank with maximal .start such that bank.start <= start 50 static inline struct meminfo * __init find_bank(unsigned long start) find_bank() argument 56 if (sysmem.bank[i].start <= start) find_bank() 88 int __init add_sysmem_bank(unsigned long start, unsigned long end) add_sysmem_bank() argument 95 if (start == end || add_sysmem_bank() 96 (start < end) != (PAGE_ALIGN(start) < (end & PAGE_MASK))) { add_sysmem_bank() 98 start, end - start); add_sysmem_bank() 102 start = PAGE_ALIGN(start); add_sysmem_bank() 104 sz = end - start; add_sysmem_bank() 106 it = find_bank(start); add_sysmem_bank() 109 bank_sz = it->end - it->start; add_sysmem_bank() 111 if (it && bank_sz >= start - it->start) { add_sysmem_bank() 112 if (end - it->start > bank_sz) add_sysmem_bank() 123 it->start - start <= sz) { add_sysmem_bank() 124 it->start = start; add_sysmem_bank() 125 if (it->end - it->start < sz) add_sysmem_bank() 132 start, end - start); add_sysmem_bank() 135 it->start = start; add_sysmem_bank() 140 sz = it->end - it->start; add_sysmem_bank() 142 if (sysmem.bank[i].start - it->start <= sz) { add_sysmem_bank() 143 if (sz < sysmem.bank[i].end - it->start) add_sysmem_bank() 154 * mem_reserve(start, end, must_exist) 161 * start Start of region, 170 int __init mem_reserve(unsigned long start, unsigned long end, int must_exist) mem_reserve() argument 177 start = start & PAGE_MASK; mem_reserve() 179 sz = end - start; mem_reserve() 183 it = find_bank(start); mem_reserve() 186 bank_sz = it->end - it->start; mem_reserve() 188 if ((!it || end - it->start > bank_sz) && must_exist) { mem_reserve() 190 start, end); mem_reserve() 194 if (it && start - it->start <= bank_sz) { mem_reserve() 195 if (start == it->start) { mem_reserve() 196 if (end - it->start < bank_sz) { mem_reserve() 197 it->start = end; mem_reserve() 203 it->end = start; mem_reserve() 204 if (end - it->start < bank_sz) mem_reserve() 206 it->start + bank_sz); mem_reserve() 215 if (it->end - start <= sz) { mem_reserve() 219 if (it->start - start < sz) mem_reserve() 220 it->start = end; mem_reserve() 258 pfn = PAGE_ALIGN(sysmem.bank[i].start) >> PAGE_SHIFT; bootmem_init() 279 if (sysmem.bank[i].end - sysmem.bank[i].start >= bootmap_size) { bootmem_init() 280 bootmap_start = sysmem.bank[i].start; bootmem_init() 298 if (sysmem.bank[i].start >> PAGE_SHIFT < max_low_pfn) { bootmem_init() 301 free_bootmem(sysmem.bank[i].start, bootmem_init() 302 end - sysmem.bank[i].start); bootmem_init() 370 void free_initrd_mem(unsigned long start, unsigned long end) free_initrd_mem() argument 373 free_reserved_area((void *)start, (void *)end, -1, "initrd"); free_initrd_mem()
|
/linux-4.4.14/arch/mips/pci/ |
H A D | pci-malta.c | 42 .start = 0x00000000UL, 93 resource_size_t start, end, map, start1, end1, map1, map2, map3, mask; mips_pcibios_init() local 115 start = GT_READ(GT_PCI0M0LD_OFS); mips_pcibios_init() 118 end = (end & GT_PCI_HD_MSK) | (start & ~GT_PCI_HD_MSK); mips_pcibios_init() 124 if (end1 - start1 > end - start) { mips_pcibios_init() 125 start = start1; mips_pcibios_init() 129 mask = ~(start ^ end); mips_pcibios_init() 131 BUG_ON((start & GT_PCI_HD_MSK) != (map & GT_PCI_HD_MSK) && mips_pcibios_init() 133 gt64120_mem_resource.start = start; mips_pcibios_init() 135 gt64120_controller.mem_offset = (start & mask) - (map & mask); mips_pcibios_init() 137 gt64120_mem_resource.start <<= GT_PCI_DCRM_SHF; mips_pcibios_init() 142 start = GT_READ(GT_PCI0IOLD_OFS); mips_pcibios_init() 145 end = (end & GT_PCI_HD_MSK) | (start & ~GT_PCI_HD_MSK); mips_pcibios_init() 146 mask = ~(start ^ end); mips_pcibios_init() 148 BUG_ON((start & GT_PCI_HD_MSK) != (map & GT_PCI_HD_MSK) && mips_pcibios_init() 150 gt64120_io_resource.start = map & mask; mips_pcibios_init() 154 gt64120_io_resource.start <<= GT_PCI_DCRM_SHF; mips_pcibios_init() 172 start = BONITO_PCILO0_BASE; mips_pcibios_init() 176 start = BONITO_PCILO1_BASE; mips_pcibios_init() 181 start = BONITO_PCILO0_BASE; mips_pcibios_init() 184 bonito64_mem_resource.start = start; mips_pcibios_init() 185 bonito64_mem_resource.end = start + mips_pcibios_init() 187 bonito64_controller.mem_offset = start - mips_pcibios_init() 198 MSC_READ(MSC01_PCI_SC2PMBASL, start); mips_pcibios_init() 201 msc_mem_resource.start = start & mask; mips_pcibios_init() 202 msc_mem_resource.end = (start & mask) | ~mask; mips_pcibios_init() 203 msc_controller.mem_offset = (start & mask) - (map & mask); mips_pcibios_init() 205 write_gcr_reg0_base(start); mips_pcibios_init() 209 MSC_READ(MSC01_PCI_SC2PIOBASL, start); mips_pcibios_init() 212 msc_io_resource.start = map & mask; mips_pcibios_init() 217 write_gcr_reg1_base(start); mips_pcibios_init() 222 start = start & mask; mips_pcibios_init() 223 end = start | ~mask; mips_pcibios_init() 224 if ((start >= msc_mem_resource.start && mips_pcibios_init() 225 start <= msc_mem_resource.end) || mips_pcibios_init() 226 (end >= msc_mem_resource.start && mips_pcibios_init() 229 start = max(start, msc_mem_resource.start); mips_pcibios_init() 231 if (start - msc_mem_resource.start >= mips_pcibios_init() 233 msc_mem_resource.end = start - 1; mips_pcibios_init() 235 msc_mem_resource.start = end + 1; mips_pcibios_init() 245 if (controller->io_resource->start < 0x00001000UL) mips_pcibios_init() 246 controller->io_resource->start = 0x00001000UL; mips_pcibios_init()
|
/linux-4.4.14/arch/x86/include/asm/ |
H A D | pat.h | 11 extern int reserve_memtype(u64 start, u64 end, 13 extern int free_memtype(u64 start, u64 end); 18 int io_reserve_memtype(resource_size_t start, resource_size_t end, 21 void io_free_memtype(resource_size_t start, resource_size_t end);
|
H A D | kasan.h | 8 * Compiler uses shadow offset assuming that addresses start 9 * from 0. Kernel addresses don't start from 0, so shadow 11 * 'kernel address space start' >> KASAN_SHADOW_SCALE_SHIFT
|
/linux-4.4.14/include/linux/ |
H A D | range.h | 5 u64 start; member in struct:range 10 u64 start, u64 end); 14 u64 start, u64 end); 16 void subtract_range(struct range *range, int az, u64 start, u64 end);
|
H A D | ioport.h | 19 resource_size_t start; member in struct:resource 46 #define IORESOURCE_STARTALIGN 0x00080000 /* start field is alignment */ 112 .start = (_start), \ 147 resource_size_t start, resource_size_t end, 161 struct resource *lookup_resource(struct resource *root, resource_size_t start); 162 int adjust_resource(struct resource *res, resource_size_t start, 167 return res->end - res->start + 1; resource_size() 180 return r1->start <= r2->start && r1->end >= r2->end; resource_contains() 185 #define request_region(start,n,name) __request_region(&ioport_resource, (start), (n), (name), 0) 186 #define request_muxed_region(start,n,name) __request_region(&ioport_resource, (start), (n), (name), IORESOURCE_MUXED) 187 #define __request_mem_region(start,n,name, excl) __request_region(&iomem_resource, (start), (n), (name), excl) 188 #define request_mem_region(start,n,name) __request_region(&iomem_resource, (start), (n), (name), 0) 189 #define request_mem_region_exclusive(start,n,name) \ 190 __request_region(&iomem_resource, (start), (n), (name), IORESOURCE_EXCLUSIVE) 194 resource_size_t start, 199 #define release_region(start,n) __release_region(&ioport_resource, (start), (n)) 200 #define release_mem_region(start,n) __release_region(&iomem_resource, (start), (n)) 216 #define devm_request_region(dev,start,n,name) \ 217 __devm_request_region(dev, &ioport_resource, (start), (n), (name)) 218 #define devm_request_mem_region(dev,start,n,name) \ 219 __devm_request_region(dev, &iomem_resource, (start), (n), (name)) 222 struct resource *parent, resource_size_t start, 225 #define devm_release_region(dev, start, n) \ 226 __devm_release_region(dev, &ioport_resource, (start), (n)) 227 #define devm_release_mem_region(dev, start, n) \ 228 __devm_release_region(dev, &iomem_resource, (start), (n)) 231 resource_size_t start, resource_size_t n); 239 walk_system_ram_res(u64 start, u64 end, void *arg, 242 walk_iomem_res(char *name, unsigned long flags, u64 start, u64 end, void *arg, 248 return (r1->start <= r2->end && r1->end >= r2->start); resource_overlaps()
|
H A D | interval_tree.h | 8 unsigned long start; /* Start of interval */ member in struct:interval_tree_node 21 unsigned long start, unsigned long last); 25 unsigned long start, unsigned long last);
|
H A D | if_tunnel.h | 13 #define for_each_ip_tunnel_rcu(pos, start) \ 14 for (pos = rcu_dereference(start); pos; pos = rcu_dereference(pos->next))
|
H A D | firmware-map.h | 26 int firmware_map_add_early(u64 start, u64 end, const char *type); 27 int firmware_map_add_hotplug(u64 start, u64 end, const char *type); 28 int firmware_map_remove(u64 start, u64 end, const char *type); 32 static inline int firmware_map_add_early(u64 start, u64 end, const char *type) firmware_map_add_early() argument 37 static inline int firmware_map_add_hotplug(u64 start, u64 end, const char *type) firmware_map_add_hotplug() argument 42 static inline int firmware_map_remove(u64 start, u64 end, const char *type) firmware_map_remove() argument
|
/linux-4.4.14/arch/sparc/include/asm/ |
H A D | tlbflush_32.h | 10 #define flush_tlb_range(vma, start, end) \ 11 sparc32_cachetlb_ops->tlb_range(vma, start, end) 18 static inline void flush_tlb_kernel_range(unsigned long start, flush_tlb_kernel_range() argument
|
/linux-4.4.14/arch/mips/lasat/image/ |
H A D | head.S | 4 .section .text..start, "ax" 15 /* image start and size */ 19 /* start of kernel and entrypoint in uncompressed image */
|
/linux-4.4.14/arch/blackfin/include/asm/ |
H A D | cacheflush.h | 28 #define flush_cache_range(vma, start, end) do { } while (0) 30 #define flush_cache_vmap(start, end) do { } while (0) 31 #define flush_cache_vunmap(start, end) do { } while (0) 34 #define flush_icache_range_others(start, end) \ 35 smp_icache_flush_range_others((start), (end)) 37 #define flush_icache_range_others(start, end) do { } while (0) 40 static inline void flush_icache_range(unsigned start, unsigned end) flush_icache_range() argument 44 blackfin_dcache_flush_range(start, end); flush_icache_range() 47 if (start >= L2_START && end <= L2_START + L2_LENGTH) flush_icache_range() 48 blackfin_dcache_flush_range(start, end); flush_icache_range() 62 blackfin_icache_flush_range(start, end); flush_icache_range() 63 flush_icache_range_others(start, end); flush_icache_range() 67 if (start >= L2_START && end <= L2_START + L2_LENGTH) { flush_icache_range() 68 blackfin_icache_flush_range(start, end); flush_icache_range() 69 flush_icache_range_others(start, end); flush_icache_range() 82 # define invalidate_dcache_range(start,end) blackfin_dcache_invalidate_range((start), (end)) 84 # define invalidate_dcache_range(start,end) do { } while (0) 87 # define flush_dcache_range(start,end) blackfin_dcache_flush_range((start), (end)) 91 # define flush_dcache_range(start,end) do { } while (0)
|
/linux-4.4.14/arch/arm/mm/ |
H A D | cache-feroceon-l2.c | 33 * Cache range operations are initiated by writing the start and 36 * [start:end]. 73 static inline void l2_clean_pa_range(unsigned long start, unsigned long end) l2_clean_pa_range() argument 78 * Make sure 'start' and 'end' reference the same page, as l2_clean_pa_range() 80 * the start address. l2_clean_pa_range() 82 BUG_ON((start ^ end) >> PAGE_SHIFT); l2_clean_pa_range() 84 va_start = l2_get_va(start); l2_clean_pa_range() 85 va_end = va_start + (end - start); l2_clean_pa_range() 104 static inline void l2_inv_pa_range(unsigned long start, unsigned long end) l2_inv_pa_range() argument 109 * Make sure 'start' and 'end' reference the same page, as l2_inv_pa_range() 111 * the start address. l2_inv_pa_range() 113 BUG_ON((start ^ end) >> PAGE_SHIFT); l2_inv_pa_range() 115 va_start = l2_get_va(start); l2_inv_pa_range() 116 va_end = va_start + (end - start); l2_inv_pa_range() 135 * inclusive start and end addresses. 142 static unsigned long calc_range_end(unsigned long start, unsigned long end) calc_range_end() argument 146 BUG_ON(start & (CACHE_LINE_SIZE - 1)); calc_range_end() 150 * Try to process all cache lines between 'start' and 'end'. calc_range_end() 159 if (range_end > start + MAX_RANGE_SIZE) calc_range_end() 160 range_end = start + MAX_RANGE_SIZE; calc_range_end() 165 if (range_end > (start | (PAGE_SIZE - 1)) + 1) calc_range_end() 166 range_end = (start | (PAGE_SIZE - 1)) + 1; calc_range_end() 171 static void feroceon_l2_inv_range(unsigned long start, unsigned long end) feroceon_l2_inv_range() argument 176 if (start & (CACHE_LINE_SIZE - 1)) { feroceon_l2_inv_range() 177 l2_clean_inv_pa(start & ~(CACHE_LINE_SIZE - 1)); feroceon_l2_inv_range() 178 start = (start | (CACHE_LINE_SIZE - 1)) + 1; feroceon_l2_inv_range() 184 if (start < end && end & (CACHE_LINE_SIZE - 1)) { feroceon_l2_inv_range() 190 * Invalidate all full cache lines between 'start' and 'end'. feroceon_l2_inv_range() 192 while (start < end) { feroceon_l2_inv_range() 193 unsigned long range_end = calc_range_end(start, end); feroceon_l2_inv_range() 194 l2_inv_pa_range(start, range_end - CACHE_LINE_SIZE); feroceon_l2_inv_range() 195 start = range_end; feroceon_l2_inv_range() 201 static void feroceon_l2_clean_range(unsigned long start, unsigned long end) feroceon_l2_clean_range() argument 208 start &= ~(CACHE_LINE_SIZE - 1); feroceon_l2_clean_range() 210 while (start != end) { feroceon_l2_clean_range() 211 unsigned long range_end = calc_range_end(start, end); feroceon_l2_clean_range() 212 l2_clean_pa_range(start, range_end - CACHE_LINE_SIZE); feroceon_l2_clean_range() 213 start = range_end; feroceon_l2_clean_range() 220 static void feroceon_l2_flush_range(unsigned long start, unsigned long end) feroceon_l2_flush_range() argument 222 start &= ~(CACHE_LINE_SIZE - 1); feroceon_l2_flush_range() 224 while (start != end) { feroceon_l2_flush_range() 225 unsigned long range_end = calc_range_end(start, end); feroceon_l2_flush_range() 227 l2_clean_pa_range(start, range_end - CACHE_LINE_SIZE); feroceon_l2_flush_range() 228 l2_inv_pa_range(start, range_end - CACHE_LINE_SIZE); feroceon_l2_flush_range() 229 start = range_end; feroceon_l2_flush_range()
|
H A D | cache-v4wt.S | 78 * flush_user_cache_range(start, end, flags) 83 * - start - start address (inclusive, page aligned) 101 * coherent_kern_range(start, end) 104 * region described by start. If you have non-snooping 107 * - start - virtual start address 114 * coherent_user_range(start, end) 117 * region described by start. If you have non-snooping 120 * - start - virtual start address 148 * dma_inv_range(start, end) 151 * May not write back any entries. If 'start' or 'end' 155 * - start - virtual start address 167 * dma_flush_range(start, end) 171 * - start - virtual start address 178 * dma_unmap_area(start, size, dir) 179 * - start - kernel virtual start address 190 * dma_map_area(start, size, dir) 191 * - start - kernel virtual start address
|
H A D | cache-xsc3l2.c | 98 static void xsc3_l2_inv_range(unsigned long start, unsigned long end) xsc3_l2_inv_range() argument 102 if (start == 0 && end == -1ul) { xsc3_l2_inv_range() 112 if (start & (CACHE_LINE_SIZE - 1)) { xsc3_l2_inv_range() 113 vaddr = l2_map_va(start & ~(CACHE_LINE_SIZE - 1), vaddr); xsc3_l2_inv_range() 116 start = (start | (CACHE_LINE_SIZE - 1)) + 1; xsc3_l2_inv_range() 120 * Invalidate all full cache lines between 'start' and 'end'. xsc3_l2_inv_range() 122 while (start < (end & ~(CACHE_LINE_SIZE - 1))) { xsc3_l2_inv_range() 123 vaddr = l2_map_va(start, vaddr); xsc3_l2_inv_range() 125 start += CACHE_LINE_SIZE; xsc3_l2_inv_range() 131 if (start < end) { xsc3_l2_inv_range() 132 vaddr = l2_map_va(start, vaddr); xsc3_l2_inv_range() 142 static void xsc3_l2_clean_range(unsigned long start, unsigned long end) xsc3_l2_clean_range() argument 148 start &= ~(CACHE_LINE_SIZE - 1); xsc3_l2_clean_range() 149 while (start < end) { xsc3_l2_clean_range() 150 vaddr = l2_map_va(start, vaddr); xsc3_l2_clean_range() 152 start += CACHE_LINE_SIZE; xsc3_l2_clean_range() 180 static void xsc3_l2_flush_range(unsigned long start, unsigned long end) xsc3_l2_flush_range() argument 184 if (start == 0 && end == -1ul) { xsc3_l2_flush_range() 191 start &= ~(CACHE_LINE_SIZE - 1); xsc3_l2_flush_range() 192 while (start < end) { xsc3_l2_flush_range() 193 vaddr = l2_map_va(start, vaddr); xsc3_l2_flush_range() 196 start += CACHE_LINE_SIZE; xsc3_l2_flush_range()
|
H A D | pageattr.c | 40 unsigned long start = addr; change_memory_common() local 42 unsigned long end = start + size; change_memory_common() 47 start &= PAGE_MASK; change_memory_common() 48 end = start + size; change_memory_common() 52 if (start < MODULES_VADDR || start >= MODULES_END) change_memory_common() 55 if (end < MODULES_VADDR || start >= MODULES_END) change_memory_common() 61 ret = apply_to_page_range(&init_mm, start, size, change_page_range, change_memory_common() 64 flush_tlb_kernel_range(start, end); change_memory_common()
|
H A D | cache-fa.S | 78 * flush_user_cache_range(start, end, flags) 83 * - start - start address (inclusive, page aligned) 106 * coherent_kern_range(start, end) 109 * region described by start. If you have non-snooping 112 * - start - virtual start address 119 * coherent_user_range(start, end) 122 * region described by start. If you have non-snooping 125 * - start - virtual start address 162 * dma_inv_range(start, end) 165 * May not write back any entries. If 'start' or 'end' 169 * - start - virtual start address 188 * dma_clean_range(start, end) 192 * - start - virtual start address 206 * dma_flush_range(start,end) 207 * - start - virtual start address of region 221 * dma_map_area(start, size, dir) 222 * - start - kernel virtual start address 235 * dma_unmap_area(start, size, dir) 236 * - start - kernel virtual start address
|
H A D | cache-v4wb.S | 101 * flush_user_cache_range(start, end, flags) 106 * - start - start address (inclusive, page aligned) 142 * coherent_kern_range(start, end) 145 * region described by start. If you have non-snooping 148 * - start - virtual start address 155 * coherent_user_range(start, end) 158 * region described by start. If you have non-snooping 161 * - start - virtual start address 178 * dma_inv_range(start, end) 181 * May not write back any entries. If 'start' or 'end' 185 * - start - virtual start address 202 * dma_clean_range(start, end) 206 * - start - virtual start address 219 * dma_flush_range(start, end) 223 * - start - virtual start address 232 * dma_map_area(start, size, dir) 233 * - start - kernel virtual start address 246 * dma_unmap_area(start, size, dir) 247 * - start - kernel virtual start address
|
H A D | cache-v6.S | 89 * v6_flush_cache_range(start, end, flags) 93 * - start - start address (may not be aligned) 104 * v6_coherent_kern_range(start,end) 110 * - start - virtual start address of region 120 * v6_coherent_user_range(start,end) 126 * - start - virtual start address of region 195 * v6_dma_inv_range(start,end) 201 * - start - virtual start address of region 245 * v6_dma_clean_range(start,end) 246 * - start - virtual start address of region 268 * v6_dma_flush_range(start,end) 269 * - start - virtual start address of region 296 * dma_map_area(start, size, dir) 297 * - start - kernel virtual start address 315 * dma_unmap_area(start, size, dir) 316 * - start - kernel virtual start address
|
/linux-4.4.14/tools/perf/util/ |
H A D | svghelper.h | 6 extern void open_svg(const char *filename, int cpus, int rows, u64 start, u64 end); 7 extern void svg_ubox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int merges); 8 extern void svg_lbox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int merges); 9 extern void svg_fbox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int merges); 10 extern void svg_box(int Yslot, u64 start, u64 end, const char *type); 11 extern void svg_blocked(int Yslot, int cpu, u64 start, u64 end, const char *backtrace); 12 extern void svg_running(int Yslot, int cpu, u64 start, u64 end, const char *backtrace); 13 extern void svg_waiting(int Yslot, int cpu, u64 start, u64 end, const char *backtrace); 17 extern void svg_process(int cpu, u64 start, u64 end, int pid, const char *name, const char *backtrace); 18 extern void svg_cstate(int cpu, u64 start, u64 end, int type); 19 extern void svg_pstate(int cpu, u64 start, u64 end, u64 freq); 25 extern void svg_wakeline(u64 start, int row1, int row2, const char *backtrace); 26 extern void svg_partial_wakeline(u64 start, int row1, char *desc1, int row2, char *desc2, const char *backtrace); 27 extern void svg_interrupt(u64 start, int row, const char *backtrace); 28 extern void svg_text(int Yslot, u64 start, const char *text);
|
H A D | find-vdso-map.c | 1 static int find_vdso_map(void **start, void **end) find_vdso_map() argument 18 start, end, &m)) find_vdso_map()
|
H A D | svghelper.c | 86 void open_svg(const char *filename, int cpus, int rows, u64 start, u64 end) open_svg() argument 95 first_time = start; open_svg() 154 void svg_ubox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int merges) svg_ubox() argument 156 double w = time2pixels(end) - time2pixels(start); svg_ubox() 165 time2pixels(start), svg_ubox() 173 void svg_lbox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int merges) svg_lbox() argument 175 double w = time2pixels(end) - time2pixels(start); svg_lbox() 184 time2pixels(start), svg_lbox() 192 void svg_fbox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int merges) svg_fbox() argument 194 double w = time2pixels(end) - time2pixels(start); svg_fbox() 203 time2pixels(start), svg_fbox() 211 void svg_box(int Yslot, u64 start, u64 end, const char *type) svg_box() argument 217 time2pixels(start), time2pixels(end)-time2pixels(start), Yslot * SLOT_MULT, SLOT_HEIGHT, type); svg_box() 221 void svg_blocked(int Yslot, int cpu, u64 start, u64 end, const char *backtrace) svg_blocked() argument 228 time_to_string(end - start)); svg_blocked() 231 svg_box(Yslot, start, end, "blocked"); svg_blocked() 235 void svg_running(int Yslot, int cpu, u64 start, u64 end, const char *backtrace) svg_running() argument 243 if (svg_highlight && end - start > svg_highlight) svg_running() 250 cpu, time_to_string(end - start)); svg_running() 254 time2pixels(start), time2pixels(end)-time2pixels(start), Yslot * SLOT_MULT, SLOT_HEIGHT, svg_running() 257 text_size = (time2pixels(end)-time2pixels(start)); svg_running() 266 time2pixels(start), Yslot * SLOT_MULT + SLOT_HEIGHT - 1, text_size, cpu + 1); svg_running() 289 void svg_waiting(int Yslot, int cpu, u64 start, u64 end, const char *backtrace) svg_waiting() argument 300 if (end-start > 10 * 1000000) /* 10 msec */ svg_waiting() 303 text = time_to_string(end-start); svg_waiting() 305 font_size = 1.0 * (time2pixels(end)-time2pixels(start)); svg_waiting() 312 fprintf(svgfile, "<g transform=\"translate(%.8f,%.8f)\">\n", time2pixels(start), Yslot * SLOT_MULT); svg_waiting() 313 fprintf(svgfile, "<title>#%d waiting %s</title>\n", cpu, time_to_string(end - start)); svg_waiting() 317 time2pixels(end)-time2pixels(start), SLOT_HEIGHT, style); svg_waiting() 383 void svg_process(int cpu, u64 start, u64 end, int pid, const char *name, const char *backtrace) svg_process() argument 391 if (svg_highlight && end - start >= svg_highlight) svg_process() 398 fprintf(svgfile, "<g transform=\"translate(%.8f,%.8f)\">\n", time2pixels(start), cpu2y(cpu)); svg_process() 399 fprintf(svgfile, "<title>%d %s running %s</title>\n", pid, name, time_to_string(end - start)); svg_process() 403 time2pixels(end)-time2pixels(start), SLOT_MULT+SLOT_HEIGHT, type); svg_process() 404 width = time2pixels(end)-time2pixels(start); svg_process() 417 void svg_cstate(int cpu, u64 start, u64 end, int type) svg_cstate() argument 434 time2pixels(start), time2pixels(end)-time2pixels(start), svg_cstate() 437 width = (time2pixels(end)-time2pixels(start))/2.0; svg_cstate() 445 time2pixels(start), cpu2y(cpu)+width, width, type); svg_cstate() 474 void svg_pstate(int cpu, u64 start, u64 end, u64 freq) svg_pstate() argument 487 time2pixels(start), time2pixels(end), height, height); svg_pstate() 489 time2pixels(start), height+0.9, HzToHuman(freq)); svg_pstate() 495 void svg_partial_wakeline(u64 start, int row1, char *desc1, int row2, char *desc2, const char *backtrace) svg_partial_wakeline() argument 515 time2pixels(start), row1 * SLOT_MULT + SLOT_HEIGHT, time2pixels(start), row1 * SLOT_MULT + SLOT_HEIGHT + SLOT_MULT/32); svg_partial_wakeline() 518 time2pixels(start), row1 * SLOT_MULT + SLOT_HEIGHT + SLOT_HEIGHT/48, desc2); svg_partial_wakeline() 522 time2pixels(start), row2 * SLOT_MULT - SLOT_MULT/32, time2pixels(start), row2 * SLOT_MULT); svg_partial_wakeline() 525 time2pixels(start), row2 * SLOT_MULT - SLOT_MULT/32, desc1); svg_partial_wakeline() 530 time2pixels(start), row2 * SLOT_MULT + SLOT_HEIGHT, time2pixels(start), row2 * SLOT_MULT + SLOT_HEIGHT + SLOT_MULT/32); svg_partial_wakeline() 533 time2pixels(start), row2 * SLOT_MULT + SLOT_HEIGHT + SLOT_MULT/48, desc1); svg_partial_wakeline() 537 time2pixels(start), row1 * SLOT_MULT - SLOT_MULT/32, time2pixels(start), row1 * SLOT_MULT); svg_partial_wakeline() 540 time2pixels(start), row1 * SLOT_MULT - SLOT_HEIGHT/32, desc2); svg_partial_wakeline() 548 time2pixels(start), height); svg_partial_wakeline() 553 void svg_wakeline(u64 start, int row1, int row2, const char *backtrace) svg_wakeline() argument 568 time2pixels(start), row1 * SLOT_MULT + SLOT_HEIGHT, time2pixels(start), row2 * SLOT_MULT); svg_wakeline() 571 time2pixels(start), row2 * SLOT_MULT + SLOT_HEIGHT, time2pixels(start), row1 * SLOT_MULT); svg_wakeline() 577 time2pixels(start), height); svg_wakeline() 582 void svg_interrupt(u64 start, int row, const char *backtrace) svg_interrupt() argument 595 time2pixels(start), row * SLOT_MULT); svg_interrupt() 597 time2pixels(start), row * SLOT_MULT + SLOT_HEIGHT); svg_interrupt() 602 void svg_text(int Yslot, u64 start, const char *text) svg_text() argument 608 time2pixels(start), Yslot * SLOT_MULT+SLOT_HEIGHT/2, text); svg_text()
|
/linux-4.4.14/arch/s390/mm/ |
H A D | vmem.c | 24 unsigned long start; member in struct:memory_segment 78 static int vmem_add_mem(unsigned long start, unsigned long size, int ro) vmem_add_mem() argument 80 unsigned long end = start + size; vmem_add_mem() 81 unsigned long address = start; vmem_add_mem() 146 static void vmem_remove_range(unsigned long start, unsigned long size) vmem_remove_range() argument 148 unsigned long end = start + size; vmem_remove_range() 149 unsigned long address = start; vmem_remove_range() 187 flush_tlb_kernel_range(start, end); vmem_remove_range() 193 int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node) vmemmap_populate() argument 195 unsigned long address = start; vmemmap_populate() 202 for (address = start; address < end;) { vmemmap_populate() 264 void vmemmap_free(unsigned long start, unsigned long end) vmemmap_free() argument 276 if (seg->start + seg->size > VMEM_MAX_PHYS || insert_memory_segment() 277 seg->start + seg->size < seg->start) insert_memory_segment() 281 if (seg->start >= tmp->start + tmp->size) insert_memory_segment() 283 if (seg->start + seg->size <= tmp->start) insert_memory_segment() 302 vmem_remove_range(seg->start, seg->size); __remove_shared_memory() 305 int vmem_remove_mapping(unsigned long start, unsigned long size) vmem_remove_mapping() argument 314 if (seg->start == start && seg->size == size) vmem_remove_mapping() 318 if (seg->start != start || seg->size != size) vmem_remove_mapping() 329 int vmem_add_mapping(unsigned long start, unsigned long size) vmem_add_mapping() argument 339 seg->start = start; vmem_add_mapping() 346 ret = vmem_add_mem(start, size, 0); vmem_add_mapping() 369 phys_addr_t start, end; vmem_map_init() local 374 start = reg->base; for_each_memblock() 376 if (start >= ro_end || end <= ro_start) for_each_memblock() 377 vmem_add_mem(start, end - start, 0); for_each_memblock() 378 else if (start >= ro_start && end <= ro_end) for_each_memblock() 379 vmem_add_mem(start, end - start, 1); for_each_memblock() 380 else if (start >= ro_start) { for_each_memblock() 381 vmem_add_mem(start, ro_end - start, 1); for_each_memblock() 384 vmem_add_mem(start, ro_start - start, 0); for_each_memblock() 387 vmem_add_mem(start, ro_start - start, 0); for_each_memblock() 408 seg->start = reg->base; for_each_memblock()
|
H A D | extable.c | 48 void sort_extable(struct exception_table_entry *start, sort_extable() argument 54 /* Normalize entries to being relative to the start of the section */ sort_extable() 55 for (p = start, i = 0; p < finish; p++, i += 8) { sort_extable() 59 sort(start, finish - start, sizeof(*start), cmp_ex, NULL); sort_extable() 61 for (p = start, i = 0; p < finish; p++, i += 8) { sort_extable()
|
/linux-4.4.14/tools/perf/scripts/python/ |
H A D | net_dropmonitor.py | 33 # Invariant: kallsyms[i][0] <= loc for all 0 <= i <= start 35 start, end = -1, len(kallsyms) 36 while end != start + 1: 37 pivot = (start + end) // 2 41 start = pivot 43 # Now (start == -1 or kallsyms[start][0] <= loc) 44 # and (start == len(kallsyms) - 1 or loc < kallsyms[start + 1][0]) 45 if start >= 0: 46 symloc, name = kallsyms[start]
|
/linux-4.4.14/tools/testing/nvdimm/test/ |
H A D | iomap.c | 64 - nfit_res->res->start; __nfit_test_ioremap() 78 - nfit_res->res->start; __wrap_devm_ioremap_nocache() 92 return nfit_res->buf + offset - nfit_res->res->start; __wrap_devm_memremap() 106 return nfit_res->buf + offset - nfit_res->res->start; __wrap_memremap() 163 struct resource *parent, resource_size_t start, nfit_test_request_region() 170 nfit_res = get_nfit_res(start); nfit_test_request_region() 175 if (start + n > nfit_res->res->start nfit_test_request_region() 177 pr_debug("%s: start: %llx n: %llx overflow: %pr\n", nfit_test_request_region() 178 __func__, start, n, nfit_test_request_region() 183 res->start = start; nfit_test_request_region() 184 res->end = start + n - 1; nfit_test_request_region() 193 return __devm_request_region(dev, parent, start, n, name); nfit_test_request_region() 194 return __request_region(parent, start, n, name, flags); nfit_test_request_region() 198 resource_size_t start, resource_size_t n, const char *name, __wrap___request_region() 201 return nfit_test_request_region(NULL, parent, start, n, name, flags); __wrap___request_region() 206 struct resource *parent, resource_size_t start, __wrap___devm_request_region() 211 return nfit_test_request_region(dev, parent, start, n, name, 0); __wrap___devm_request_region() 215 void __wrap___release_region(struct resource *parent, resource_size_t start, __wrap___release_region() argument 222 nfit_res = get_nfit_res(start); __wrap___release_region() 227 if (start != res->start || resource_size(res) != n) __wrap___release_region() 228 pr_info("%s: start: %llx n: %llx mismatch: %pr\n", __wrap___release_region() 229 __func__, start, n, res); __wrap___release_region() 235 __release_region(parent, start, n); __wrap___release_region() 162 nfit_test_request_region(struct device *dev, struct resource *parent, resource_size_t start, resource_size_t n, const char *name, int flags) nfit_test_request_region() argument 197 __wrap___request_region(struct resource *parent, resource_size_t start, resource_size_t n, const char *name, int flags) __wrap___request_region() argument 205 __wrap___devm_request_region(struct device *dev, struct resource *parent, resource_size_t start, resource_size_t n, const char *name) __wrap___devm_request_region() argument
|
/linux-4.4.14/arch/arm/mach-davinci/ |
H A D | devices-da8xx.c | 163 .start = DA8XX_TPCC_BASE, 169 .start = DA8XX_TPTC0_BASE, 175 .start = DA8XX_TPTC1_BASE, 181 .start = IRQ_DA8XX_CCINT0, 186 .start = IRQ_DA8XX_CCERRINT, 194 .start = DA850_TPCC1_BASE, 200 .start = DA850_TPTC2_BASE, 206 .start = IRQ_DA850_CCINT1, 211 .start = IRQ_DA850_CCERRINT1, 266 .start = DA8XX_I2C0_BASE, 271 .start = IRQ_DA8XX_I2CINT0, 286 .start = DA8XX_I2C1_BASE, 291 .start = IRQ_DA8XX_I2CINT1, 322 .start = DA8XX_WDOG_BASE, 355 .start = DA8XX_EMAC_CPPI_PORT_BASE, 360 .start = IRQ_DA8XX_C0_RX_THRESH_PULSE, 365 .start = IRQ_DA8XX_C0_RX_PULSE, 370 .start = IRQ_DA8XX_C0_TX_PULSE, 375 .start = IRQ_DA8XX_C0_MISC_PULSE, 401 .start = DA8XX_EMAC_MDIO_BASE, 428 .start = DAVINCI_DA830_MCASP1_REG_BASE, 435 .start = DAVINCI_DA830_DMA_MCASP1_AXEVT, 442 .start = DAVINCI_DA830_DMA_MCASP1_AREVT, 448 .start = IRQ_DA8XX_MCASPINT, 463 .start = DAVINCI_DA830_MCASP2_REG_BASE, 470 .start = DAVINCI_DA830_DMA_MCASP2_AXEVT, 477 .start = DAVINCI_DA830_DMA_MCASP2_AREVT, 483 .start = IRQ_DA8XX_MCASPINT, 498 .start = DAVINCI_DA8XX_MCASP0_REG_BASE, 505 .start = DAVINCI_DA8XX_DMA_MCASP0_AXEVT, 512 .start = DAVINCI_DA8XX_DMA_MCASP0_AREVT, 518 .start = IRQ_DA8XX_MCASPINT, 561 .start = DA8XX_PRUSS_MEM_BASE, 566 .start = IRQ_DA8XX_EVTOUT0, 571 .start = IRQ_DA8XX_EVTOUT1, 576 .start = IRQ_DA8XX_EVTOUT2, 581 .start = IRQ_DA8XX_EVTOUT3, 586 .start = IRQ_DA8XX_EVTOUT4, 591 .start = IRQ_DA8XX_EVTOUT5, 596 .start = IRQ_DA8XX_EVTOUT6, 601 .start = IRQ_DA8XX_EVTOUT7, 647 .start = DA8XX_LCD_CNTRL_BASE, 652 .start = IRQ_DA8XX_LCDINT, 673 .start = DA8XX_GPIO_BASE, 678 .start = IRQ_DA8XX_GPIO0, 699 .start = DA8XX_MMCSD0_BASE, 704 .start = IRQ_DA8XX_MMCSDINT0, 709 .start = DA8XX_DMA_MMCSD0_RX, 714 .start = DA8XX_DMA_MMCSD0_TX, 736 .start = DA850_MMCSD1_BASE, 741 .start = IRQ_DA850_MMCSDINT0_1, 746 .start = DA850_DMA_MMCSD1_RX, 751 .start = DA850_DMA_MMCSD1_TX, 773 .start = DA8XX_SYSCFG0_BASE + DA8XX_HOST1CFG_REG, 778 .start = DA8XX_SYSCFG0_BASE + DA8XX_CHIPSIG_REG, 783 .start = IRQ_DA8XX_CHIPINT0, 859 .start = DA8XX_RTC_BASE, 864 .start = IRQ_DA8XX_RTC, 869 .start = IRQ_DA8XX_RTC, 902 .start = DA8XX_DDR2_CTL_BASE, 932 .start = DA8XX_SPI0_BASE, 937 .start = IRQ_DA8XX_SPINT0, 942 .start = DA8XX_DMA_SPI0_RX, 947 .start = DA8XX_DMA_SPI0_TX, 955 .start = DA830_SPI1_BASE, 960 .start = IRQ_DA8XX_SPINT1, 965 .start = DA8XX_DMA_SPI1_RX, 970 .start = DA8XX_DMA_SPI1_TX, 1020 da8xx_spi1_resources[0].start = DA850_SPI1_BASE; da8xx_register_spi_bus() 1030 .start = DA850_SATA_BASE, 1035 .start = DA8XX_SYSCFG1_BASE + DA8XX_PWRDN_REG, 1040 .start = IRQ_DA850_SATAINT,
|
H A D | usb.c | 55 .start = DAVINCI_USB_OTG_BASE, 60 .start = IRQ_USBINT, 92 usb_dev.resource[1].start = IRQ_DM646X_USBINT; davinci_setup_usb() 93 usb_dev.resource[2].start = IRQ_DM646X_USBDMAINT; davinci_setup_usb() 103 .start = DA8XX_USB0_BASE, 108 .start = IRQ_DA8XX_USB_INT, 146 .start = DA8XX_USB1_BASE, 151 .start = IRQ_DA8XX_IRQN,
|
/linux-4.4.14/drivers/gpu/drm/exynos/ |
H A D | regs-fimc.h | 27 /* Y 1st frame start address for output DMA */ 29 /* Y 2nd frame start address for output DMA */ 31 /* Y 3rd frame start address for output DMA */ 33 /* Y 4th frame start address for output DMA */ 35 /* Cb 1st frame start address for output DMA */ 37 /* Cb 2nd frame start address for output DMA */ 39 /* Cb 3rd frame start address for output DMA */ 41 /* Cb 4th frame start address for output DMA */ 43 /* Cr 1st frame start address for output DMA */ 45 /* Cr 2nd frame start address for output DMA */ 47 /* Cr 3rd frame start address for output DMA */ 49 /* Cr 4th frame start address for output DMA */ 73 /* Y frame start address for input DMA */ 75 /* Cb frame start address for input DMA */ 77 /* Cr frame start address for input DMA */ 89 /* Y frame start address for input DMA */ 91 /* Cb frame start address for input DMA */ 93 /* Cr frame start address for input DMA */ 123 /* Y 5th frame start address for output DMA */ 125 /* Y 6th frame start address for output DMA */ 127 /* Y 7th frame start address for output DMA */ 129 /* Y 8th frame start address for output DMA */ 131 /* Y 9th frame start address for output DMA */ 133 /* Y 10th frame start address for output DMA */ 135 /* Y 11th frame start address for output DMA */ 137 /* Y 12th frame start address for output DMA */ 139 /* Y 13th frame start address for output DMA */ 141 /* Y 14th frame start address for output DMA */ 143 /* Y 15th frame start address for output DMA */ 145 /* Y 16th frame start address for output DMA */ 147 /* Y 17th frame start address for output DMA */ 149 /* Y 18th frame start address for output DMA */ 151 /* Y 19th frame start address for output DMA */ 153 /* Y 20th frame start address for output DMA */ 155 /* Y 21th frame start address for output DMA */ 157 /* Y 22th frame start address for output DMA */ 159 /* Y 23th frame start address for output DMA */ 161 /* Y 24th frame start address for output DMA */ 163 /* Y 25th frame start address for output DMA */ 165 /* Y 26th frame start address for output DMA */ 167 /* Y 27th frame start address for output DMA */ 169 /* Y 28th frame start address for output DMA */ 171 /* Y 29th frame start address for output DMA */ 173 /* Y 30th frame start address for output DMA */ 175 /* Y 31th frame start address for output DMA */ 177 /* Y 32th frame start address for output DMA */ 180 /* CB 5th frame start address for output DMA */ 182 /* CB 6th frame start address for output DMA */ 184 /* CB 7th frame start address for output DMA */ 186 /* CB 8th frame start address for output DMA */ 188 /* CB 9th frame start address for output DMA */ 190 /* CB 10th frame start address for output DMA */ 192 /* CB 11th frame start address for output DMA */ 194 /* CB 12th frame start address for output DMA */ 196 /* CB 13th frame start address for output DMA */ 198 /* CB 14th frame start address for output DMA */ 200 /* CB 15th frame start address for output DMA */ 202 /* CB 16th frame start address for output DMA */ 204 /* CB 17th frame start address for output DMA */ 206 /* CB 18th frame start address for output DMA */ 208 /* CB 19th frame start address for output DMA */ 210 /* CB 20th frame start address for output DMA */ 212 /* CB 21th frame start address for output DMA */ 214 /* CB 22th frame start address for output DMA */ 216 /* CB 23th frame start address for output DMA */ 218 /* CB 24th frame start address for output DMA */ 220 /* CB 25th frame start address for output DMA */ 222 /* CB 26th frame start address for output DMA */ 224 /* CB 27th frame start address for output DMA */ 226 /* CB 28th frame start address for output DMA */ 228 /* CB 29th frame start address for output DMA */ 230 /* CB 30th frame start address for output DMA */ 232 /* CB 31th frame start address for output DMA */ 234 /* CB 32th frame start address for output DMA */ 237 /* CR 5th frame start address for output DMA */ 239 /* CR 6th frame start address for output DMA */ 241 /* CR 7th frame start address for output DMA */ 243 /* CR 8th frame start address for output DMA */ 245 /* CR 9th frame start address for output DMA */ 247 /* CR 10th frame start address for output DMA */ 249 /* CR 11th frame start address for output DMA */ 251 /* CR 12th frame start address for output DMA */ 253 /* CR 13th frame start address for output DMA */ 255 /* CR 14th frame start address for output DMA */ 257 /* CR 15th frame start address for output DMA */ 259 /* CR 16th frame start address for output DMA */ 261 /* CR 17th frame start address for output DMA */ 263 /* CR 18th frame start address for output DMA */ 265 /* CR 19th frame start address for output DMA */ 267 /* CR 20th frame start address for output DMA */ 269 /* CR 21th frame start address for output DMA */ 271 /* CR 22th frame start address for output DMA */ 273 /* CR 23th frame start address for output DMA */ 275 /* CR 24th frame start address for output DMA */ 277 /* CR 25th frame start address for output DMA */ 279 /* CR 26th frame start address for output DMA */ 281 /* CR 27th frame start address for output DMA */ 283 /* CR 28th frame start address for output DMA */ 285 /* CR 29th frame start address for output DMA */ 287 /* CR 30th frame start address for output DMA */ 289 /* CR 31th frame start address for output DMA */ 291 /* CR 32th frame start address for output DMA */ 297 /* frame start address 1 ~ 4, 5 ~ 32 */
|
/linux-4.4.14/arch/sparc/lib/ |
H A D | user_fixup.c | 21 static unsigned long compute_size(unsigned long start, unsigned long size, unsigned long *offset) compute_size() argument 24 unsigned long end = start + size; compute_size() 26 if (fault_addr < start || fault_addr >= end) { compute_size() 29 *offset = fault_addr - start; compute_size() 58 unsigned long start = (unsigned long) to; copy_in_user_fixup() local 59 unsigned long end = start + size; copy_in_user_fixup() 61 if (fault_addr >= start && fault_addr < end) copy_in_user_fixup() 64 start = (unsigned long) from; copy_in_user_fixup() 65 end = start + size; copy_in_user_fixup() 66 if (fault_addr >= start && fault_addr < end) copy_in_user_fixup()
|
/linux-4.4.14/arch/sh/include/asm/ |
H A D | tlbflush.h | 10 * - flush_tlb_range(vma, start, end) flushes a range of pages 11 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages 16 unsigned long start, 20 extern void local_flush_tlb_kernel_range(unsigned long start, 30 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, 33 extern void flush_tlb_kernel_range(unsigned long start, unsigned long end); 43 #define flush_tlb_range(vma, start, end) \ 44 local_flush_tlb_range(vma, start, end) 46 #define flush_tlb_kernel_range(start, end) \ 47 local_flush_tlb_kernel_range(start, end)
|
H A D | tlb.h | 24 unsigned long start, end; member in struct:mmu_gather 29 tlb->start = TASK_SIZE; init_tlb_gather() 33 tlb->start = 0; init_tlb_gather() 39 tlb_gather_mmu(struct mmu_gather *tlb, struct mm_struct *mm, unsigned long start, unsigned long end) tlb_gather_mmu() argument 42 tlb->start = start; tlb_gather_mmu() 44 tlb->fullmm = !(start | (end+1)); tlb_gather_mmu() 50 tlb_finish_mmu(struct mmu_gather *tlb, unsigned long start, unsigned long end) tlb_finish_mmu() argument 62 if (tlb->start > address) tlb_remove_tlb_entry() 63 tlb->start = address; tlb_remove_tlb_entry() 84 flush_tlb_range(vma, tlb->start, tlb->end); tlb_end_vma()
|
/linux-4.4.14/arch/arm/mach-omap1/ |
H A D | mcbsp.c | 100 .start = OMAP7XX_MCBSP1_BASE, 106 .start = INT_7XX_McBSP1RX, 111 .start = INT_7XX_McBSP1TX, 116 .start = 9, 121 .start = 8, 127 .start = OMAP7XX_MCBSP2_BASE, 133 .start = INT_7XX_McBSP2RX, 138 .start = INT_7XX_McBSP2TX, 143 .start = 11, 148 .start = 10, 177 .start = OMAP1510_MCBSP1_BASE, 183 .start = INT_McBSP1RX, 188 .start = INT_McBSP1TX, 193 .start = 9, 198 .start = 8, 204 .start = OMAP1510_MCBSP2_BASE, 210 .start = INT_1510_SPI_RX, 215 .start = INT_1510_SPI_TX, 220 .start = 17, 225 .start = 16, 231 .start = OMAP1510_MCBSP3_BASE, 237 .start = INT_McBSP3RX, 242 .start = INT_McBSP3TX, 247 .start = 11, 252 .start = 10, 284 .start = OMAP1610_MCBSP1_BASE, 290 .start = INT_McBSP1RX, 295 .start = INT_McBSP1TX, 300 .start = 9, 305 .start = 8, 311 .start = OMAP1610_MCBSP2_BASE, 317 .start = INT_1610_McBSP2_RX, 322 .start = INT_1610_McBSP2_TX, 327 .start = 17, 332 .start = 16, 338 .start = OMAP1610_MCBSP3_BASE, 344 .start = INT_McBSP3RX, 349 .start = INT_McBSP3TX, 354 .start = 11, 359 .start = 10,
|
H A D | gpio7xx.c | 37 .start = OMAP1_MPUIO_VBASE, 42 .start = INT_7XX_MPUIO, 78 .start = OMAP7XX_GPIO1_BASE, 83 .start = INT_7XX_GPIO_BANK1, 117 .start = OMAP7XX_GPIO2_BASE, 122 .start = INT_7XX_GPIO_BANK2, 145 .start = OMAP7XX_GPIO3_BASE, 150 .start = INT_7XX_GPIO_BANK3, 173 .start = OMAP7XX_GPIO4_BASE, 178 .start = INT_7XX_GPIO_BANK4, 201 .start = OMAP7XX_GPIO5_BASE, 206 .start = INT_7XX_GPIO_BANK5, 229 .start = OMAP7XX_GPIO6_BASE, 234 .start = INT_7XX_GPIO_BANK6,
|
/linux-4.4.14/drivers/usb/core/ |
H A D | devices.c | 181 static char *usb_dump_endpoint_descriptor(int speed, char *start, char *end, usb_dump_endpoint_descriptor() argument 187 if (start > end) usb_dump_endpoint_descriptor() 188 return start; usb_dump_endpoint_descriptor() 230 return start; usb_dump_endpoint_descriptor() 241 start += sprintf(start, format_endpt, desc->bEndpointAddress, dir, usb_dump_endpoint_descriptor() 246 return start; usb_dump_endpoint_descriptor() 249 static char *usb_dump_interface_descriptor(char *start, char *end, usb_dump_interface_descriptor() argument 258 if (start > end) usb_dump_interface_descriptor() 259 return start; usb_dump_interface_descriptor() 267 start += sprintf(start, format_iface, usb_dump_interface_descriptor() 277 return start; usb_dump_interface_descriptor() 280 static char *usb_dump_interface(int speed, char *start, char *end, usb_dump_interface() argument 287 start = usb_dump_interface_descriptor(start, end, intfc, iface, setno); usb_dump_interface() 289 if (start > end) usb_dump_interface() 290 return start; usb_dump_interface() 291 start = usb_dump_endpoint_descriptor(speed, usb_dump_interface() 292 start, end, &desc->endpoint[i].desc); usb_dump_interface() 294 return start; usb_dump_interface() 297 static char *usb_dump_iad_descriptor(char *start, char *end, usb_dump_iad_descriptor() argument 300 if (start > end) usb_dump_iad_descriptor() 301 return start; usb_dump_iad_descriptor() 302 start += sprintf(start, format_iad, usb_dump_iad_descriptor() 309 return start; usb_dump_iad_descriptor() 317 static char *usb_dump_config_descriptor(char *start, char *end, usb_dump_config_descriptor() argument 323 if (start > end) usb_dump_config_descriptor() 324 return start; usb_dump_config_descriptor() 329 start += sprintf(start, format_config, usb_dump_config_descriptor() 336 return start; usb_dump_config_descriptor() 339 static char *usb_dump_config(int speed, char *start, char *end, usb_dump_config() argument 346 if (start > end) usb_dump_config() 347 return start; usb_dump_config() 350 return start + sprintf(start, "(null Cfg. desc.)\n"); usb_dump_config() 351 start = usb_dump_config_descriptor(start, end, &config->desc, active, usb_dump_config() 356 start = usb_dump_iad_descriptor(start, end, usb_dump_config() 363 if (start > end) usb_dump_config() 364 return start; usb_dump_config() 365 start = usb_dump_interface(speed, usb_dump_config() 366 start, end, intfc, interface, j); usb_dump_config() 369 return start; usb_dump_config() 375 static char *usb_dump_device_descriptor(char *start, char *end, usb_dump_device_descriptor() argument 381 if (start > end) usb_dump_device_descriptor() 382 return start; usb_dump_device_descriptor() 383 start += sprintf(start, format_device1, usb_dump_device_descriptor() 391 if (start > end) usb_dump_device_descriptor() 392 return start; usb_dump_device_descriptor() 393 start += sprintf(start, format_device2, usb_dump_device_descriptor() 397 return start; usb_dump_device_descriptor() 403 static char *usb_dump_device_strings(char *start, char *end, usb_dump_device_strings() argument 406 if (start > end) usb_dump_device_strings() 407 return start; usb_dump_device_strings() 409 start += sprintf(start, format_string_manufacturer, usb_dump_device_strings() 411 if (start > end) usb_dump_device_strings() 414 start += sprintf(start, format_string_product, dev->product); usb_dump_device_strings() 415 if (start > end) usb_dump_device_strings() 419 start += sprintf(start, format_string_serialnumber, usb_dump_device_strings() 423 return start; usb_dump_device_strings() 426 static char *usb_dump_desc(char *start, char *end, struct usb_device *dev) usb_dump_desc() argument 430 if (start > end) usb_dump_desc() 431 return start; usb_dump_desc() 433 start = usb_dump_device_descriptor(start, end, &dev->descriptor); usb_dump_desc() 435 if (start > end) usb_dump_desc() 436 return start; usb_dump_desc() 438 start = usb_dump_device_strings(start, end, dev); usb_dump_desc() 441 if (start > end) usb_dump_desc() 442 return start; usb_dump_desc() 443 start = usb_dump_config(dev->speed, usb_dump_desc() 444 start, end, dev->config + i, usb_dump_desc() 448 return start; usb_dump_desc() 454 static char *usb_dump_hub_descriptor(char *start, char *end, usb_dump_hub_descriptor() argument 460 if (start > end) usb_dump_hub_descriptor() 461 return start; usb_dump_hub_descriptor() 462 start += sprintf(start, "Interface:"); usb_dump_hub_descriptor() 463 while (leng && start <= end) { usb_dump_hub_descriptor() 464 start += sprintf(start, " %02x", *ptr); usb_dump_hub_descriptor() 467 *start++ = '\n'; usb_dump_hub_descriptor() 468 return start; usb_dump_hub_descriptor() 471 static char *usb_dump_string(char *start, char *end, usb_dump_string() argument 474 if (start > end) usb_dump_string() 475 return start; usb_dump_string() 476 start += sprintf(start, "Interface:"); usb_dump_string() 479 start += sprintf(start, "%s: %.100s ", id, usb_dump_string() 481 return start; usb_dump_string() 580 /* if we can start copying some data to the user */ usb_device_dump()
|
/linux-4.4.14/arch/arm64/mm/ |
H A D | cache.S | 30 * flush_icache_range(start,end) 36 * - start - virtual start address of region 43 * __flush_cache_user_range(start,end) 49 * - start - virtual start address of region 104 * __inval_cache_range(start, end) 105 * - start - start address of region 112 * __dma_inv_range(start, end) 113 * - start - virtual start address of region 123 1: tst x0, x3 // start cache line aligned? 138 * __dma_clean_range(start, end) 139 * - start - virtual start address of region 160 * __dma_flush_range(start, end) 161 * - start - virtual start address of region 177 * __dma_map_area(start, size, dir) 178 * - start - kernel virtual start address 190 * __dma_unmap_area(start, size, dir) 191 * - start - kernel virtual start address
|
H A D | pageattr.c | 42 unsigned long start = addr; change_memory_common() local 44 unsigned long end = start + size; change_memory_common() 49 start &= PAGE_MASK; change_memory_common() 50 end = start + size; change_memory_common() 54 if (start < MODULES_VADDR || start >= MODULES_END) change_memory_common() 66 ret = apply_to_page_range(&init_mm, start, size, change_page_range, change_memory_common() 69 flush_tlb_kernel_range(start, end); change_memory_common()
|
/linux-4.4.14/drivers/pcmcia/ |
H A D | rsrc_iodyn.c | 34 resource_size_t start; pcmcia_align() local 36 start = (res->start & ~data->mask) + data->offset; pcmcia_align() 37 if (start < res->start) pcmcia_align() 38 start += data->mask + 1; pcmcia_align() 42 if (start & 0x300) pcmcia_align() 43 start = (start + 0x3ff) & ~0x3ff; pcmcia_align() 49 if ((res->start + size - 1) >= 1024) pcmcia_align() 50 start = res->end; pcmcia_align() 54 return start; pcmcia_align() 104 if ((s->io[i].res->start & (align-1)) == *base) iodyn_find_io() 125 *base = res->start; iodyn_find_io() 137 if (adjust_resource(s->io[i].res, res->start, iodyn_find_io() 147 try = res->start - num; iodyn_find_io() 150 res->start - num, iodyn_find_io()
|
/linux-4.4.14/arch/nios2/mm/ |
H A D | cacheflush.c | 18 static void __flush_dcache(unsigned long start, unsigned long end) __flush_dcache() argument 22 start &= ~(cpuinfo.dcache_line_size - 1); __flush_dcache() 26 if (end > start + cpuinfo.dcache_size) __flush_dcache() 27 end = start + cpuinfo.dcache_size; __flush_dcache() 29 for (addr = start; addr < end; addr += cpuinfo.dcache_line_size) { __flush_dcache() 37 static void __invalidate_dcache(unsigned long start, unsigned long end) __invalidate_dcache() argument 41 start &= ~(cpuinfo.dcache_line_size - 1); __invalidate_dcache() 45 for (addr = start; addr < end; addr += cpuinfo.dcache_line_size) { __invalidate_dcache() 53 static void __flush_icache(unsigned long start, unsigned long end) __flush_icache() argument 57 start &= ~(cpuinfo.icache_line_size - 1); __flush_icache() 61 if (end > start + cpuinfo.icache_size) __flush_icache() 62 end = start + cpuinfo.icache_size; __flush_icache() 64 for (addr = start; addr < end; addr += cpuinfo.icache_line_size) { __flush_icache() 113 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument 115 __flush_dcache(start, end); flush_icache_range() 116 __flush_icache(start, end); flush_icache_range() 119 void flush_dcache_range(unsigned long start, unsigned long end) flush_dcache_range() argument 121 __flush_dcache(start, end); flush_dcache_range() 122 __flush_icache(start, end); flush_dcache_range() 126 void invalidate_dcache_range(unsigned long start, unsigned long end) invalidate_dcache_range() argument 128 __invalidate_dcache(start, end); invalidate_dcache_range() 132 void flush_cache_range(struct vm_area_struct *vma, unsigned long start, flush_cache_range() argument 135 __flush_dcache(start, end); flush_cache_range() 137 __flush_icache(start, end); flush_cache_range() 142 unsigned long start = (unsigned long) page_address(page); flush_icache_page() local 143 unsigned long end = start + PAGE_SIZE; flush_icache_page() 145 __flush_dcache(start, end); flush_icache_page() 146 __flush_icache(start, end); flush_icache_page() 152 unsigned long start = vmaddr; flush_cache_page() local 153 unsigned long end = start + PAGE_SIZE; flush_cache_page() 155 __flush_dcache(start, end); flush_cache_page() 157 __flush_icache(start, end); flush_cache_page() 167 unsigned long start = (unsigned long)page_address(page); __flush_dcache_page() local 169 __flush_dcache(start, start + PAGE_SIZE); __flush_dcache_page() 191 unsigned long start = (unsigned long)page_address(page); flush_dcache_page() local 193 flush_icache_range(start, start + PAGE_SIZE); flush_dcache_page()
|
/linux-4.4.14/drivers/net/wireless/ti/wlcore/ |
H A D | io.c | 83 if ((addr >= part->mem.start) && wlcore_translate_addr() 84 (addr < part->mem.start + part->mem.size)) wlcore_translate_addr() 85 return addr - part->mem.start; wlcore_translate_addr() 86 else if ((addr >= part->reg.start) && wlcore_translate_addr() 87 (addr < part->reg.start + part->reg.size)) wlcore_translate_addr() 88 return addr - part->reg.start + part->mem.size; wlcore_translate_addr() 89 else if ((addr >= part->mem2.start) && wlcore_translate_addr() 90 (addr < part->mem2.start + part->mem2.size)) wlcore_translate_addr() 91 return addr - part->mem2.start + part->mem.size + wlcore_translate_addr() 93 else if ((addr >= part->mem3.start) && wlcore_translate_addr() 94 (addr < part->mem3.start + part->mem3.size)) wlcore_translate_addr() 95 return addr - part->mem3.start + part->mem.size + wlcore_translate_addr() 117 * ...+----+--> mem.start 121 * 00000000 <--+----+... ...+----+--> mem.start + mem.size 129 * + <--+----+... ...+----+--> reg.start 133 * ...+----+--> reg.start + reg.size 146 p->mem.start, p->mem.size); wlcore_set_partition() 148 p->reg.start, p->reg.size); wlcore_set_partition() 150 p->mem2.start, p->mem2.size); wlcore_set_partition() 152 p->mem3.start, p->mem3.size); wlcore_set_partition() 154 ret = wlcore_raw_write32(wl, HW_PART0_START_ADDR, p->mem.start); wlcore_set_partition() 162 ret = wlcore_raw_write32(wl, HW_PART1_START_ADDR, p->reg.start); wlcore_set_partition() 170 ret = wlcore_raw_write32(wl, HW_PART2_START_ADDR, p->mem2.start); wlcore_set_partition() 183 ret = wlcore_raw_write32(wl, HW_PART3_START_ADDR, p->mem3.start); wlcore_set_partition()
|
/linux-4.4.14/arch/powerpc/include/asm/ |
H A D | sections.h | 25 static inline int overlaps_interrupt_vector_text(unsigned long start, overlaps_interrupt_vector_text() argument 32 return start < (unsigned long)__va(real_end) && overlaps_interrupt_vector_text() 36 static inline int overlaps_kernel_text(unsigned long start, unsigned long end) overlaps_kernel_text() argument 38 return start < (unsigned long)__init_end && overlaps_kernel_text() 42 static inline int overlaps_kvm_tmp(unsigned long start, unsigned long end) overlaps_kvm_tmp() argument 46 return start < (unsigned long)kvm_tmp && overlaps_kvm_tmp()
|
H A D | cacheflush.h | 22 #define flush_cache_range(vma, start, end) do { } while (0) 25 #define flush_cache_vmap(start, end) do { } while (0) 26 #define flush_cache_vunmap(start, end) do { } while (0) 50 extern void flush_dcache_range(unsigned long start, unsigned long stop); 52 extern void clean_dcache_range(unsigned long start, unsigned long stop); 53 extern void invalidate_dcache_range(unsigned long start, unsigned long stop); 56 extern void flush_inval_dcache_range(unsigned long start, unsigned long stop); 57 extern void flush_dcache_phys_range(unsigned long start, unsigned long stop);
|
H A D | rheap.h | 21 unsigned long start; member in struct:_rh_block 41 unsigned long start; member in struct:_rh_stats 60 extern int rh_attach_region(rh_info_t * info, unsigned long start, int size); 63 extern unsigned long rh_detach_region(rh_info_t * info, unsigned long start, int size); 73 extern unsigned long rh_alloc_fixed(rh_info_t * info, unsigned long start, int size, 77 extern int rh_free(rh_info_t * info, unsigned long start); 87 extern int rh_set_owner(rh_info_t * info, unsigned long start, const char *owner);
|
/linux-4.4.14/arch/mips/jz4740/ |
H A D | platform.c | 38 .start = JZ4740_UHC_BASE_ADDR, 43 .start = JZ4740_IRQ_UHC, 68 .start = JZ4740_UDC_BASE_ADDR, 73 .start = JZ4740_IRQ_UDC, 94 .start = JZ4740_MSC_BASE_ADDR, 99 .start = JZ4740_IRQ_MSC, 119 .start = JZ4740_RTC_BASE_ADDR, 124 .start = JZ4740_IRQ_RTC, 140 .start = JZ4740_I2C_BASE_ADDR, 145 .start = JZ4740_IRQ_I2C, 162 .start = JZ4740_EMC_BASE_ADDR, 168 .start = 0x18000000, 174 .start = 0x14000000, 180 .start = 0x0C000000, 186 .start = 0x08000000, 201 .start = JZ4740_LCD_BASE_ADDR, 221 .start = JZ4740_AIC_BASE_ADDR, 243 .start = JZ4740_AIC_BASE_ADDR + 0x80, 259 .start = JZ4740_SADC_BASE_ADDR, 264 .start = JZ4740_IRQ_SADC, 269 .start = JZ4740_IRQ_ADC_BASE, 285 .start = JZ4740_WDT_BASE_ADDR, 307 .start = JZ4740_DMAC_BASE_ADDR, 312 .start = JZ4740_IRQ_DMAC,
|
/linux-4.4.14/arch/arm/mach-pxa/ |
H A D | devices.c | 38 .start = IRQ_PMU, 52 .start = 0x41100000, 57 .start = IRQ_MMC, 62 .start = 21, 67 .start = 22, 103 .start = 0x40600000, 108 .start = IRQ_USB, 141 .start = 0x54100000, 146 .start = IRQ_USB2, 167 .start = 0x44000000, 172 .start = IRQ_LCD, 199 .start = 0x40100000, 203 .start = IRQ_FFUART, 223 .start = 0x40200000, 227 .start = IRQ_BTUART, 247 .start = 0x40700000, 251 .start = IRQ_STUART, 271 .start = 0x41600000, 275 .start = IRQ_HWUART, 298 .start = 0x40301680, 302 .start = IRQ_I2C, 323 .start = 0x40f00180, 327 .start = IRQ_PWRI2C, 343 .start = 0x40400000, 347 .start = IRQ_I2S, 389 .start = IRQ_STUART, 394 .start = IRQ_ICP, 399 .start = 0x40800000, 404 .start = 0x40700000, 409 .start = 17, 414 .start = 18, 438 .start = 0x40900000, 443 .start = IRQ_RTC1Hz, 449 .start = IRQ_RTCAlrm, 472 .start = 0x40500000, 477 .start = IRQ_AC97, 505 .start = 0x40b00000, 520 .start = 0x40c00000, 537 .start = 0x41000000, 542 .start = IRQ_SSP, 548 .start = 13, 554 .start = 14, 575 .start = 0x41400000, 580 .start = IRQ_NSSP, 586 .start = 15, 592 .start = 16, 613 .start = 0x41500000, 618 .start = IRQ_ASSP, 624 .start = 23, 630 .start = 24, 652 .start = 0x50000000, 657 .start = IRQ_CAMERA, 685 .start = 0x4C000000, 690 .start = IRQ_USBH1, 716 .start = 0x41500000, 721 .start = IRQ_KEYPAD, 743 .start = 0x41000000, 748 .start = IRQ_SSP, 754 .start = 13, 760 .start = 14, 781 .start = 0x41700000, 786 .start = IRQ_SSP2, 792 .start = 15, 798 .start = 16, 819 .start = 0x41900000, 824 .start = IRQ_SSP3, 830 .start = 66, 836 .start = 67, 855 .start = 0x40b00000, 870 .start = 0x40c00000, 887 .start = 0x42000000, 892 .start = IRQ_MMC2, 897 .start = 93, 902 .start = 94, 926 .start = 0x42500000, 931 .start = IRQ_MMC3, 936 .start = 100, 941 .start = 101, 965 .start = 0x54000000, 970 .start = IRQ_GCU, 994 .start = 0x40f500c0, 998 .start = IRQ_PWRI2C, 1013 .start = 0x43100000, 1018 .start = IRQ_NAND, 1024 .start = 97, 1030 .start = 99, 1058 .start = 0x41a00000, 1063 .start = IRQ_SSP4, 1069 .start = 2, 1075 .start = 3, 1134 .start = 0x40e00000, 1138 .start = IRQ_GPIO0, 1143 .start = IRQ_GPIO1, 1148 .start = IRQ_GPIO_2_x, 1210 .start = 0x40000000, 1215 .start = IRQ_DMA,
|
/linux-4.4.14/arch/alpha/mm/ |
H A D | numa.c | 48 printk(" memcluster %2d, usage %1lx, start %8lu, end %8lu\n", for_each_mem_cluster() 62 unsigned long start, end; setup_memory_node() local 87 start = cluster->start_pfn; for_each_mem_cluster() 88 end = start + cluster->numpages; for_each_mem_cluster() 90 if (start >= node_pfn_end || end <= node_pfn_start) for_each_mem_cluster() 97 printk(" memcluster %2d, usage %1lx, start %8lu, end %8lu\n", for_each_mem_cluster() 101 if (start < node_pfn_start) for_each_mem_cluster() 102 start = node_pfn_start; for_each_mem_cluster() 106 if (start < node_min_pfn) for_each_mem_cluster() 107 node_min_pfn = start; for_each_mem_cluster() 144 printk(" Detected node memory: start %8lu, end %8lu\n", 158 /* Zone start phys-addr must be 2^(MAX_ORDER-1) aligned. 172 start = cluster->start_pfn; for_each_mem_cluster() 173 end = start + cluster->numpages; for_each_mem_cluster() 175 if (start >= node_max_pfn || end <= node_min_pfn) for_each_mem_cluster() 180 if (start < node_min_pfn) for_each_mem_cluster() 181 start = node_min_pfn; for_each_mem_cluster() 183 if (start < start_kernel_pfn) { for_each_mem_cluster() 190 } else if (start < end_kernel_pfn) for_each_mem_cluster() 191 start = end_kernel_pfn; for_each_mem_cluster() 192 if (end - start >= bootmap_pages) { for_each_mem_cluster() 193 bootmap_start = start; for_each_mem_cluster() 212 start = cluster->start_pfn; for_each_mem_cluster() 215 if (start >= node_max_pfn || end <= node_min_pfn) for_each_mem_cluster() 220 if (start < node_min_pfn) for_each_mem_cluster() 221 start = node_min_pfn; for_each_mem_cluster() 223 if (start < start_kernel_pfn) { for_each_mem_cluster() 225 free_bootmem_node(NODE_DATA(nid), PFN_PHYS(start), for_each_mem_cluster() 227 - PFN_PHYS(start))); for_each_mem_cluster() 229 start, start_kernel_pfn); for_each_mem_cluster() 230 start = end_kernel_pfn; for_each_mem_cluster() 233 } else if (start < end_kernel_pfn) for_each_mem_cluster() 234 start = end_kernel_pfn; for_each_mem_cluster() 235 if (start >= end) for_each_mem_cluster() 238 free_bootmem_node(NODE_DATA(nid), PFN_PHYS(start), PFN_PHYS(end) - PFN_PHYS(start)); for_each_mem_cluster() 239 printk(" freeing pages %ld:%ld\n", start, end); for_each_mem_cluster()
|
/linux-4.4.14/fs/btrfs/tests/ |
H A D | extent-io-tests.c | 28 static noinline int process_page_range(struct inode *inode, u64 start, u64 end, process_page_range() argument 33 unsigned long index = start >> PAGE_CACHE_SHIFT; process_page_range() 59 printk(KERN_ERR "stuck in a loop, start %Lu, end %Lu, nr_pages %lu, ret %d\n", start, end, nr_pages, ret); process_page_range() 75 u64 start, end, test_start; test_find_delalloc() local 113 start = 0; test_find_delalloc() 115 found = find_lock_delalloc_range(inode, &tmp, locked_page, &start, test_find_delalloc() 121 if (start != 0 || end != 4095) { test_find_delalloc() 122 test_msg("Expected start 0 end 4095, got start %Lu end %Lu\n", test_find_delalloc() 123 start, end); test_find_delalloc() 126 unlock_extent(&tmp, start, end); test_find_delalloc() 144 start = test_start; test_find_delalloc() 146 found = find_lock_delalloc_range(inode, &tmp, locked_page, &start, test_find_delalloc() 152 if (start != test_start || end != max_bytes - 1) { test_find_delalloc() 153 test_msg("Expected start %Lu end %Lu, got start %Lu, end " test_find_delalloc() 154 "%Lu\n", test_start, max_bytes - 1, start, end); test_find_delalloc() 157 if (process_page_range(inode, start, end, test_find_delalloc() 162 unlock_extent(&tmp, start, end); test_find_delalloc() 178 start = test_start; test_find_delalloc() 180 found = find_lock_delalloc_range(inode, &tmp, locked_page, &start, test_find_delalloc() 199 start = test_start; test_find_delalloc() 201 found = find_lock_delalloc_range(inode, &tmp, locked_page, &start, test_find_delalloc() 207 if (start != test_start || end != total_dirty - 1) { test_find_delalloc() 208 test_msg("Expected start %Lu end %Lu, got start %Lu end %Lu\n", test_find_delalloc() 209 test_start, total_dirty - 1, start, end); test_find_delalloc() 212 if (process_page_range(inode, start, end, test_find_delalloc() 217 unlock_extent(&tmp, start, end); test_find_delalloc() 234 start = test_start; test_find_delalloc() 242 found = find_lock_delalloc_range(inode, &tmp, locked_page, &start, test_find_delalloc() 248 if (start != test_start && end != test_start + PAGE_CACHE_SIZE - 1) { test_find_delalloc() 249 test_msg("Expected start %Lu end %Lu, got start %Lu end %Lu\n", test_find_delalloc() 250 test_start, test_start + PAGE_CACHE_SIZE - 1, start, test_find_delalloc() 254 if (process_page_range(inode, start, end, PROCESS_TEST_LOCKED | test_find_delalloc()
|
H A D | inode-tests.c | 26 static void insert_extent(struct btrfs_root *root, u64 start, u64 len, insert_extent() argument 45 key.offset = start; insert_extent() 308 if (em->start != 0 || em->len != 5) { test_btrfs_get_extent() 309 test_msg("Unexpected extent wanted start 0 len 5, got start " test_btrfs_get_extent() 310 "%llu len %llu\n", em->start, em->len); test_btrfs_get_extent() 317 offset = em->start + em->len; test_btrfs_get_extent() 329 if (em->start != offset || em->len != 4091) { test_btrfs_get_extent() 330 test_msg("Unexpected extent wanted start %llu len 1, got start " test_btrfs_get_extent() 331 "%llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent() 343 offset = em->start + em->len; test_btrfs_get_extent() 355 if (em->start != offset || em->len != 4) { test_btrfs_get_extent() 356 test_msg("Unexpected extent wanted start %llu len 4, got start " test_btrfs_get_extent() 357 "%llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent() 364 offset = em->start + em->len; test_btrfs_get_extent() 377 if (em->start != offset || em->len != 4095) { test_btrfs_get_extent() 378 test_msg("Unexpected extent wanted start %llu len 4095, got " test_btrfs_get_extent() 379 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent() 386 if (em->orig_start != em->start) { test_btrfs_get_extent() 387 test_msg("Wrong orig offset, want %llu, have %llu\n", em->start, test_btrfs_get_extent() 391 offset = em->start + em->len; test_btrfs_get_extent() 404 if (em->start != offset || em->len != 4096) { test_btrfs_get_extent() 405 test_msg("Unexpected extent wanted start %llu len 4096, got " test_btrfs_get_extent() 406 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent() 413 if (em->orig_start != em->start) { test_btrfs_get_extent() 414 test_msg("Wrong orig offset, want %llu, have %llu\n", em->start, test_btrfs_get_extent() 419 orig_start = em->start; test_btrfs_get_extent() 420 offset = em->start + em->len; test_btrfs_get_extent() 432 if (em->start != offset || em->len != 4096) { test_btrfs_get_extent() 433 test_msg("Unexpected extent wanted start %llu len 4096, got " test_btrfs_get_extent() 434 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent() 441 offset = em->start + em->len; test_btrfs_get_extent() 453 if (em->start != offset || em->len != 8192) { test_btrfs_get_extent() 454 test_msg("Unexpected extent wanted start %llu len 8192, got " test_btrfs_get_extent() 455 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent() 467 disk_bytenr += (em->start - orig_start); test_btrfs_get_extent() 469 test_msg("Wrong block start, want %llu, have %llu\n", test_btrfs_get_extent() 473 offset = em->start + em->len; test_btrfs_get_extent() 486 if (em->start != offset || em->len != 4096) { test_btrfs_get_extent() 487 test_msg("Unexpected extent wanted start %llu len 4096, got " test_btrfs_get_extent() 488 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent() 496 if (em->orig_start != em->start) { test_btrfs_get_extent() 497 test_msg("Wrong orig offset, want %llu, have %llu\n", em->start, test_btrfs_get_extent() 501 offset = em->start + em->len; test_btrfs_get_extent() 514 if (em->start != offset || em->len != 4096) { test_btrfs_get_extent() 515 test_msg("Unexpected extent wanted start %llu len 4096, got " test_btrfs_get_extent() 516 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent() 524 if (em->orig_start != em->start) { test_btrfs_get_extent() 525 test_msg("Wrong orig offset, want %llu, have %llu\n", em->start, test_btrfs_get_extent() 530 orig_start = em->start; test_btrfs_get_extent() 531 offset = em->start + em->len; test_btrfs_get_extent() 543 if (em->start != offset || em->len != 4096) { test_btrfs_get_extent() 544 test_msg("Unexpected extent wanted start %llu len 4096, got " test_btrfs_get_extent() 545 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent() 557 if (em->block_start != (disk_bytenr + (em->start - em->orig_start))) { test_btrfs_get_extent() 558 test_msg("Unexpected block start, wanted %llu, have %llu\n", test_btrfs_get_extent() 559 disk_bytenr + (em->start - em->orig_start), test_btrfs_get_extent() 563 offset = em->start + em->len; test_btrfs_get_extent() 575 if (em->start != offset || em->len != 8192) { test_btrfs_get_extent() 576 test_msg("Unexpected extent wanted start %llu len 8192, got " test_btrfs_get_extent() 577 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent() 590 if (em->block_start != (disk_bytenr + (em->start - em->orig_start))) { test_btrfs_get_extent() 591 test_msg("Unexpected block start, wanted %llu, have %llu\n", test_btrfs_get_extent() 592 disk_bytenr + (em->start - em->orig_start), test_btrfs_get_extent() 596 offset = em->start + em->len; test_btrfs_get_extent() 609 if (em->start != offset || em->len != 8192) { test_btrfs_get_extent() 610 test_msg("Unexpected extent wanted start %llu len 8192, got " test_btrfs_get_extent() 611 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent() 619 if (em->orig_start != em->start) { test_btrfs_get_extent() 621 em->start, em->orig_start); test_btrfs_get_extent() 629 offset = em->start + em->len; test_btrfs_get_extent() 642 if (em->start != offset || em->len != 4096) { test_btrfs_get_extent() 643 test_msg("Unexpected extent wanted start %llu len 4096, got " test_btrfs_get_extent() 644 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent() 652 if (em->orig_start != em->start) { test_btrfs_get_extent() 654 em->start, em->orig_start); test_btrfs_get_extent() 663 orig_start = em->start; test_btrfs_get_extent() 664 offset = em->start + em->len; test_btrfs_get_extent() 676 if (em->start != offset || em->len != 4096) { test_btrfs_get_extent() 677 test_msg("Unexpected extent wanted start %llu len 4096, got " test_btrfs_get_extent() 678 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent() 685 if (em->orig_start != em->start) { test_btrfs_get_extent() 686 test_msg("Wrong orig offset, want %llu, have %llu\n", em->start, test_btrfs_get_extent() 690 offset = em->start + em->len; test_btrfs_get_extent() 699 test_msg("Block start does not match, want %llu got %llu\n", test_btrfs_get_extent() 703 if (em->start != offset || em->len != 8192) { test_btrfs_get_extent() 704 test_msg("Unexpected extent wanted start %llu len 8192, got " test_btrfs_get_extent() 705 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent() 715 em->start, orig_start); test_btrfs_get_extent() 723 offset = em->start + em->len; test_btrfs_get_extent() 736 if (em->start != offset || em->len != 4096) { test_btrfs_get_extent() 737 test_msg("Unexpected extent wanted start %llu len 4096, got " test_btrfs_get_extent() 738 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent() 745 if (em->orig_start != em->start) { test_btrfs_get_extent() 746 test_msg("Wrong orig offset, want %llu, have %llu\n", em->start, test_btrfs_get_extent() 750 offset = em->start + em->len; test_btrfs_get_extent() 767 if (em->start != offset || em->len != 12288) { test_btrfs_get_extent() 768 test_msg("Unexpected extent wanted start %llu len 12288, got " test_btrfs_get_extent() 769 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent() 777 if (em->orig_start != em->start) { test_btrfs_get_extent() 778 test_msg("Wrong orig offset, want %llu, have %llu\n", em->start, test_btrfs_get_extent() 782 offset = em->start + em->len; test_btrfs_get_extent() 794 if (em->start != offset || em->len != 4096) { test_btrfs_get_extent() 795 test_msg("Unexpected extent wanted start %llu len 4096, got " test_btrfs_get_extent() 796 "start %llu len %llu\n", offset, em->start, em->len); test_btrfs_get_extent() 803 if (em->orig_start != em->start) { test_btrfs_get_extent() 804 test_msg("Wrong orig offset, want %llu, have %llu\n", em->start, test_btrfs_get_extent() 874 if (em->start != 0 || em->len != 4096) { test_hole_first() 875 test_msg("Unexpected extent wanted start 0 len 4096, got start " test_hole_first() 876 "%llu len %llu\n", em->start, em->len); test_hole_first() 895 if (em->start != 4096 || em->len != 4096) { test_hole_first() 896 test_msg("Unexpected extent wanted start 4096 len 4096, got " test_hole_first() 897 "start %llu len %llu\n", em->start, em->len); test_hole_first()
|
/linux-4.4.14/drivers/pci/ |
H A D | hotplug-pci.c | 10 int pass, busnr, start = parent->busn_res.start; pci_hp_add_bridge() local 13 for (busnr = start; busnr <= end; busnr++) { pci_hp_add_bridge()
|
H A D | host-bridge.c | 62 region->start = res->start - offset; pcibios_resource_to_bus() 70 return region1->start <= region2->start && region1->end >= region2->end; region_contains() 86 bus_region.start = window->res->start - window->offset; pcibios_bus_to_resource() 95 res->start = region->start + offset; pcibios_bus_to_resource()
|
/linux-4.4.14/arch/x86/mm/ |
H A D | kasan_init_64.c | 17 unsigned long start; map_range() local 20 start = (unsigned long)kasan_mem_to_shadow(pfn_to_kaddr(range->start)); map_range() 28 return vmemmap_populate(start, end + 1, NUMA_NO_NODE); map_range() 31 static void __init clear_pgds(unsigned long start, clear_pgds() argument 34 for (; start < end; start += PGDIR_SIZE) clear_pgds() 35 pgd_clear(pgd_offset_k(start)); clear_pgds() 41 unsigned long start = KASAN_SHADOW_START; kasan_map_early_shadow() local 44 for (i = pgd_index(start); start < end; i++) { kasan_map_early_shadow() 47 start += PGDIR_SIZE; kasan_map_early_shadow()
|
H A D | pat_rbtree.c | 39 static int is_node_overlap(struct memtype *node, u64 start, u64 end) is_node_overlap() argument 41 if (node->start >= end || node->end <= start) is_node_overlap() 75 /* Find the first (lowest start addr) overlapping range from rb tree */ memtype_rb_lowest_match() 77 u64 start, u64 end) memtype_rb_lowest_match() 85 if (get_subtree_max_end(node->rb_left) > start) { memtype_rb_lowest_match() 88 } else if (is_node_overlap(data, start, end)) { memtype_rb_lowest_match() 91 } else if (start >= data->start) { memtype_rb_lowest_match() 102 u64 start, u64 end) memtype_rb_exact_match() 106 match = memtype_rb_lowest_match(root, start, end); memtype_rb_exact_match() 107 while (match != NULL && match->start < end) { memtype_rb_exact_match() 110 if (match->start == start && match->end == end) memtype_rb_exact_match() 124 u64 start, u64 end, memtype_rb_check_conflict() 132 match = memtype_rb_lowest_match(&memtype_rbroot, start, end); memtype_rb_check_conflict() 139 dprintk("Overlap at 0x%Lx-0x%Lx\n", match->start, match->end); memtype_rb_check_conflict() 146 if (match->start >= end) /* Checked all possible matches */ memtype_rb_check_conflict() 149 if (is_node_overlap(match, start, end) && memtype_rb_check_conflict() 164 current->comm, current->pid, start, end, memtype_rb_check_conflict() 180 if (newdata->start <= data->start) memtype_rb_insert() 182 else if (newdata->start > data->start) memtype_rb_insert() 196 err = memtype_rb_check_conflict(&memtype_rbroot, new->start, new->end, rbt_memtype_check_insert() 209 struct memtype *rbt_memtype_erase(u64 start, u64 end) rbt_memtype_erase() argument 213 data = memtype_rb_exact_match(&memtype_rbroot, start, end); rbt_memtype_erase() 76 memtype_rb_lowest_match(struct rb_root *root, u64 start, u64 end) memtype_rb_lowest_match() argument 101 memtype_rb_exact_match(struct rb_root *root, u64 start, u64 end) memtype_rb_exact_match() argument 123 memtype_rb_check_conflict(struct rb_root *root, u64 start, u64 end, enum page_cache_mode reqtype, enum page_cache_mode *newtype) memtype_rb_check_conflict() argument
|
H A D | dump_pagetables.c | 254 * such as the start of vmalloc space etc. note_page() 283 pte_t *start; walk_pte_level() local 286 start = (pte_t *) pmd_page_vaddr(addr); walk_pte_level() 288 prot = pte_flags(*start); walk_pte_level() 291 start++; walk_pte_level() 301 pmd_t *start; walk_pmd_level() local 304 start = (pmd_t *) pud_page_vaddr(addr); walk_pmd_level() 307 if (!pmd_none(*start)) { walk_pmd_level() 308 if (pmd_large(*start) || !pmd_present(*start)) { walk_pmd_level() 309 prot = pmd_flags(*start); walk_pmd_level() 312 walk_pte_level(m, st, *start, walk_pmd_level() 317 start++; walk_pmd_level() 333 pud_t *start; walk_pud_level() local 336 start = (pud_t *) pgd_page_vaddr(addr); walk_pud_level() 340 if (!pud_none(*start)) { walk_pud_level() 341 if (pud_large(*start) || !pud_present(*start)) { walk_pud_level() 342 prot = pud_flags(*start); walk_pud_level() 345 walk_pmd_level(m, st, *start, walk_pud_level() 351 start++; walk_pud_level() 380 pgd_t *start = (pgd_t *) &init_level4_pgt; ptdump_walk_pgd_level_core() local 382 pgd_t *start = swapper_pg_dir; ptdump_walk_pgd_level_core() 389 start = pgd; ptdump_walk_pgd_level_core() 399 if (!pgd_none(*start) && !is_hypervisor_range(i)) { ptdump_walk_pgd_level_core() 400 if (pgd_large(*start) || !pgd_present(*start)) { ptdump_walk_pgd_level_core() 401 prot = pgd_flags(*start); ptdump_walk_pgd_level_core() 404 walk_pud_level(m, &st, *start, ptdump_walk_pgd_level_core() 410 start++; ptdump_walk_pgd_level_core()
|
/linux-4.4.14/arch/blackfin/mach-bf561/boards/ |
H A D | tepla.c | 24 .start = 0x2C000300, 28 .start = IRQ_PROG_INTB, 32 .start = IRQ_PF7, 49 .start = BFIN_UART_THR, 54 .start = IRQ_UART_TX, 59 .start = IRQ_UART_RX, 64 .start = IRQ_UART_ERROR, 69 .start = CH_UART_TX, 74 .start = CH_UART_RX, 100 .start = 0xFFC00400, 105 .start = IRQ_UART0_RX, 110 .start = CH_UART0_RX,
|
H A D | cm_bf561.c | 103 .start = SPI0_REGBASE, 108 .start = CH_SPI, 113 .start = IRQ_SPI, 157 .start = 0x28000300, 161 .start = IRQ_PF0, 183 .start = 0x24008000, 188 .start = IRQ_PF43, 215 .start = 0x24000000, 219 .start = IRQ_PF45, 236 .start = 0x24008000, 240 .start = 0x24008004, 244 .start = IRQ_PF47, 276 .start = BFIN_UART_THR, 281 .start = IRQ_UART_TX, 286 .start = IRQ_UART_RX, 291 .start = IRQ_UART_ERROR, 296 .start = CH_UART_TX, 301 .start = CH_UART_RX, 327 .start = 0xFFC00400, 332 .start = IRQ_UART0_RX, 337 .start = CH_UART0_RX, 361 .start = 0x2400C000, 366 .start = 0x2400D018, 371 .start = PATA_INT, 412 .start = 0x20000000,
|
/linux-4.4.14/arch/c6x/include/asm/ |
H A D | cache.h | 66 extern void enable_caching(unsigned long start, unsigned long end); 67 extern void disable_caching(unsigned long start, unsigned long end); 80 extern void L1P_cache_block_invalidate(unsigned int start, unsigned int end); 81 extern void L1D_cache_block_invalidate(unsigned int start, unsigned int end); 82 extern void L1D_cache_block_writeback_invalidate(unsigned int start, 84 extern void L1D_cache_block_writeback(unsigned int start, unsigned int end); 85 extern void L2_cache_block_invalidate(unsigned int start, unsigned int end); 86 extern void L2_cache_block_writeback(unsigned int start, unsigned int end); 87 extern void L2_cache_block_writeback_invalidate(unsigned int start, 89 extern void L2_cache_block_invalidate_nowait(unsigned int start, 91 extern void L2_cache_block_writeback_nowait(unsigned int start, 94 extern void L2_cache_block_writeback_invalidate_nowait(unsigned int start,
|
/linux-4.4.14/arch/mips/sgi-ip27/ |
H A D | ip27-klconfig.c | 50 lboard_t *find_lboard(lboard_t *start, unsigned char brd_type) find_lboard() argument 53 while (start) { find_lboard() 54 if (start->brd_type == brd_type) find_lboard() 55 return start; find_lboard() 56 start = KLCF_NEXT(start); find_lboard() 62 lboard_t *find_lboard_class(lboard_t *start, unsigned char brd_type) find_lboard_class() argument 65 while (start) { find_lboard_class() 66 if (KLCLASS(start->brd_type) == KLCLASS(brd_type)) find_lboard_class() 67 return start; find_lboard_class() 68 start = KLCF_NEXT(start); find_lboard_class()
|
/linux-4.4.14/arch/mips/include/asm/mach-rc32434/ |
H A D | irq.h | 17 #define GROUP0_IRQ_BASE 8 /* GRP2 IRQ numbers start here */ 18 /* GRP3 IRQ numbers start here */ 20 /* GRP4 IRQ numbers start here */ 22 /* GRP5 IRQ numbers start here */
|
/linux-4.4.14/arch/sh/mm/ |
H A D | cache-sh2.c | 19 static void sh2__flush_wback_region(void *start, int size) sh2__flush_wback_region() argument 24 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); sh2__flush_wback_region() 25 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2__flush_wback_region() 40 static void sh2__flush_purge_region(void *start, int size) sh2__flush_purge_region() argument 45 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); sh2__flush_purge_region() 46 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2__flush_purge_region() 54 static void sh2__flush_invalidate_region(void *start, int size) sh2__flush_invalidate_region() argument 76 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); sh2__flush_invalidate_region() 77 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2__flush_invalidate_region()
|
H A D | numa.c | 27 void __init setup_bootmem_node(int nid, unsigned long start, unsigned long end) setup_bootmem_node() argument 36 start_pfn = PFN_DOWN(start); setup_bootmem_node() 39 pmb_bolt_mapping((unsigned long)__va(start), start, end - start, setup_bootmem_node() 42 memblock_add(start, end - start); setup_bootmem_node()
|
H A D | tlbflush_32.c | 39 void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, local_flush_tlb_range() argument 50 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_range() 60 start &= PAGE_MASK; local_flush_tlb_range() 67 while (start < end) { local_flush_tlb_range() 68 local_flush_tlb_one(asid, start); local_flush_tlb_range() 69 start += PAGE_SIZE; local_flush_tlb_range() 78 void local_flush_tlb_kernel_range(unsigned long start, unsigned long end) local_flush_tlb_kernel_range() argument 85 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_kernel_range() 93 start &= PAGE_MASK; local_flush_tlb_kernel_range() 97 while (start < end) { local_flush_tlb_kernel_range() 98 local_flush_tlb_one(asid, start); local_flush_tlb_kernel_range() 99 start += PAGE_SIZE; local_flush_tlb_kernel_range()
|
H A D | flush-sh4.c | 13 static void sh4__flush_wback_region(void *start, int size) sh4__flush_wback_region() argument 17 aligned_start = register_align(start); sh4__flush_wback_region() 47 static void sh4__flush_purge_region(void *start, int size) sh4__flush_purge_region() argument 51 aligned_start = register_align(start); sh4__flush_purge_region() 77 static void sh4__flush_invalidate_region(void *start, int size) sh4__flush_invalidate_region() argument 81 aligned_start = register_align(start); sh4__flush_invalidate_region()
|
H A D | cache-sh2a.c | 50 static void sh2a__flush_wback_region(void *start, int size) sh2a__flush_wback_region() argument 58 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); sh2a__flush_wback_region() 59 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2a__flush_wback_region() 92 static void sh2a__flush_purge_region(void *start, int size) sh2a__flush_purge_region() argument 98 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); sh2a__flush_purge_region() 99 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2a__flush_purge_region() 122 static void sh2a__flush_invalidate_region(void *start, int size) sh2a__flush_invalidate_region() argument 128 begin = (unsigned long)start & ~(L1_CACHE_BYTES-1); sh2a__flush_invalidate_region() 129 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2a__flush_invalidate_region() 154 unsigned long start, end; sh2a_flush_icache_range() local 158 start = data->addr1 & ~(L1_CACHE_BYTES-1); sh2a_flush_icache_range() 162 sh2a__flush_wback_region((void *)start, end-start); sh2a_flush_icache_range() 170 if (((end - start) >> PAGE_SHIFT) >= MAX_ICACHE_PAGES) { sh2a_flush_icache_range() 174 for (v = start; v < end; v += L1_CACHE_BYTES) sh2a_flush_icache_range()
|
/linux-4.4.14/arch/avr32/mm/ |
H A D | cache.c | 24 void invalidate_dcache_region(void *start, size_t size) invalidate_dcache_region() argument 34 begin = (unsigned long)start; invalidate_dcache_region() 38 flush_dcache_line(start); invalidate_dcache_region() 52 void clean_dcache_region(void *start, size_t size) clean_dcache_region() argument 57 begin = (unsigned long)start & ~(linesz - 1); clean_dcache_region() 58 end = ((unsigned long)start + size + linesz - 1) & ~(linesz - 1); clean_dcache_region() 65 void flush_dcache_region(void *start, size_t size) flush_dcache_region() argument 70 begin = (unsigned long)start & ~(linesz - 1); flush_dcache_region() 71 end = ((unsigned long)start + size + linesz - 1) & ~(linesz - 1); flush_dcache_region() 78 void invalidate_icache_region(void *start, size_t size) invalidate_icache_region() argument 83 begin = (unsigned long)start & ~(linesz - 1); invalidate_icache_region() 84 end = ((unsigned long)start + size + linesz - 1) & ~(linesz - 1); invalidate_icache_region() 90 static inline void __flush_icache_range(unsigned long start, unsigned long end) __flush_icache_range() argument 95 for (v = start; v < end; v += linesz) { __flush_icache_range() 106 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument 111 __flush_icache_range(start & ~(linesz - 1), flush_icache_range()
|
/linux-4.4.14/arch/blackfin/mach-bf537/boards/ |
H A D | minotaur.c | 37 .start = 0x20310000, /* IO PORT */ 41 .start = 0x20311000, /* Attribute Memory */ 45 .start = IRQ_PF4, 49 .start = IRQ_PF6, /* Card Detect PF6 */ 106 .start = 0x20300000, 110 .start = IRQ_PF7, 205 .start = SPI0_REGBASE, 210 .start = CH_SPI, 215 .start = IRQ_SPI, 236 .start = UART0_THR, 241 .start = IRQ_UART0_TX, 246 .start = IRQ_UART0_RX, 251 .start = IRQ_UART0_ERROR, 256 .start = CH_UART0_TX, 261 .start = CH_UART0_RX, 284 .start = UART1_THR, 289 .start = IRQ_UART1_TX, 294 .start = IRQ_UART1_RX, 299 .start = IRQ_UART1_ERROR, 304 .start = CH_UART1_TX, 309 .start = CH_UART1_RX, 335 .start = 0xFFC00400, 340 .start = IRQ_UART0_RX, 345 .start = CH_UART0_RX, 361 .start = 0xFFC02000, 366 .start = IRQ_UART1_RX, 371 .start = CH_UART1_RX, 391 .start = TWI0_REGBASE, 396 .start = IRQ_TWI, 417 .start = SPORT0_TCR1, 422 .start = IRQ_SPORT0_RX, 427 .start = IRQ_SPORT0_ERROR, 451 .start = SPORT1_TCR1, 456 .start = IRQ_SPORT1_RX, 461 .start = IRQ_SPORT1_ERROR,
|
H A D | tcm_bf537.c | 114 .start = SPI0_REGBASE, 119 .start = CH_SPI, 124 .start = IRQ_SPI, 172 .start = 0x20200300, 176 .start = IRQ_PF14, 196 .start = 0x20308000, 200 .start = 0x20308004, 204 .start = IRQ_PG15, 235 .start = 0x20300000, 239 .start = IRQ_PG13, 281 .start = 0x20000000, 285 .start = (unsigned long)cm_flash_gpios, 306 .start = UART0_THR, 311 .start = IRQ_UART0_TX, 316 .start = IRQ_UART0_RX, 321 .start = IRQ_UART0_ERROR, 326 .start = CH_UART0_TX, 331 .start = CH_UART0_RX, 354 .start = UART1_THR, 359 .start = IRQ_UART1_TX, 364 .start = IRQ_UART1_RX, 369 .start = IRQ_UART1_ERROR, 374 .start = CH_UART1_TX, 379 .start = CH_UART1_RX, 405 .start = 0xFFC00400, 410 .start = IRQ_UART0_RX, 415 .start = CH_UART0_RX, 431 .start = 0xFFC02000, 436 .start = IRQ_UART1_RX, 441 .start = CH_UART1_RX, 461 .start = TWI0_REGBASE, 466 .start = IRQ_TWI, 487 .start = SPORT0_TCR1, 492 .start = IRQ_SPORT0_RX, 497 .start = IRQ_SPORT0_ERROR, 521 .start = SPORT1_TCR1, 526 .start = IRQ_SPORT1_RX, 531 .start = IRQ_SPORT1_ERROR, 596 .start = 0x2030C000, 601 .start = 0x2030D018, 606 .start = PATA_INT,
|
H A D | cm_bf537e.c | 114 .start = SPI0_REGBASE, 119 .start = CH_SPI, 124 .start = IRQ_SPI, 160 .start = SPORT0_TCR1, 165 .start = IRQ_SPORT0_ERROR, 190 .start = SPORT1_TCR1, 195 .start = IRQ_SPORT1_ERROR, 237 .start = 0x20200300, 241 .start = IRQ_PF14, 261 .start = 0x20308000, 265 .start = 0x20308004, 269 .start = IRQ_PG15, 300 .start = 0x20300000, 304 .start = IRQ_PG13, 346 .start = 0x20000000, 350 .start = (unsigned long)cm_flash_gpios, 371 .start = UART0_THR, 376 .start = IRQ_UART0_TX, 381 .start = IRQ_UART0_RX, 386 .start = IRQ_UART0_ERROR, 391 .start = CH_UART0_TX, 396 .start = CH_UART0_RX, 405 .start = -1, 413 .start = -1, 437 .start = UART1_THR, 442 .start = IRQ_UART1_TX, 447 .start = IRQ_UART1_RX, 452 .start = IRQ_UART1_ERROR, 457 .start = CH_UART1_TX, 462 .start = CH_UART1_RX, 471 .start = -1, 479 .start = -1, 506 .start = 0xFFC00400, 511 .start = IRQ_UART0_RX, 516 .start = CH_UART0_RX, 531 .start = 0xFFC02000, 536 .start = IRQ_UART1_RX, 541 .start = CH_UART1_RX, 560 .start = TWI0_REGBASE, 565 .start = IRQ_TWI, 593 .start = SPORT0_TCR1, 598 .start = IRQ_SPORT0_RX, 603 .start = IRQ_SPORT0_ERROR, 622 .start = SPORT1_TCR1, 627 .start = IRQ_SPORT1_RX, 632 .start = IRQ_SPORT1_ERROR, 657 .start = SPORT0_TCR1, 662 .start = IRQ_SPORT0_RX, 667 .start = IRQ_SPORT0_TX, 672 .start = IRQ_SPORT0_ERROR, 677 .start = CH_SPORT0_TX, 682 .start = CH_SPORT0_RX, 740 .start = 0x2030C000, 745 .start = 0x2030D018, 750 .start = PATA_INT,
|
H A D | pnav10.c | 36 .start = 0x20310000, /* IO PORT */ 40 .start = 0x20311000, /* Attribute Memory */ 44 .start = IRQ_PF4, 48 .start = 6, /* Card Detect PF6 */ 81 .start = 0x20300300, 86 .start = IRQ_PF7, 138 .start = 0x20300000, 142 .start = IRQ_PF7, 260 .start = SPI0_REGBASE, 265 .start = CH_SPI, 270 .start = IRQ_SPI, 304 .start = UART0_THR, 309 .start = IRQ_UART0_TX, 314 .start = IRQ_UART0_RX, 319 .start = IRQ_UART0_ERROR, 324 .start = CH_UART0_TX, 329 .start = CH_UART0_RX, 352 .start = UART1_THR, 357 .start = IRQ_UART1_TX, 362 .start = IRQ_UART1_RX, 367 .start = IRQ_UART1_ERROR, 372 .start = CH_UART1_TX, 377 .start = CH_UART1_RX, 403 .start = 0xFFC00400, 408 .start = IRQ_UART0_RX, 413 .start = CH_UART0_RX, 429 .start = 0xFFC02000, 434 .start = IRQ_UART1_RX, 439 .start = CH_UART1_RX,
|
H A D | cm_bf537u.c | 114 .start = SPI0_REGBASE, 119 .start = CH_SPI, 124 .start = IRQ_SPI, 172 .start = 0x20200300, 176 .start = IRQ_PF14, 196 .start = 0x20308000, 200 .start = 0x20308004, 204 .start = IRQ_PG15, 235 .start = 0x20200000, 239 .start = IRQ_PH14, 281 .start = 0x20000000, 285 .start = (unsigned long)cm_flash_gpios, 306 .start = UART0_THR, 311 .start = IRQ_UART0_TX, 316 .start = IRQ_UART0_RX, 321 .start = IRQ_UART0_ERROR, 326 .start = CH_UART0_TX, 331 .start = CH_UART0_RX, 354 .start = UART1_THR, 359 .start = IRQ_UART1_TX, 364 .start = IRQ_UART1_RX, 369 .start = IRQ_UART1_ERROR, 374 .start = CH_UART1_TX, 379 .start = CH_UART1_RX, 405 .start = 0xFFC00400, 410 .start = IRQ_UART0_RX, 415 .start = CH_UART0_RX, 430 .start = 0xFFC02000, 435 .start = IRQ_UART1_RX, 440 .start = CH_UART1_RX, 459 .start = TWI0_REGBASE, 464 .start = IRQ_TWI, 485 .start = SPORT0_TCR1, 490 .start = IRQ_SPORT0_RX, 495 .start = IRQ_SPORT0_ERROR, 519 .start = SPORT1_TCR1, 524 .start = IRQ_SPORT1_RX, 529 .start = IRQ_SPORT1_ERROR, 594 .start = 0x2030C000, 599 .start = 0x2030D018, 604 .start = PATA_INT,
|
/linux-4.4.14/arch/blackfin/mach-bf538/boards/ |
H A D | ezkit.c | 47 .start = UART0_THR, 52 .start = IRQ_UART0_TX, 57 .start = IRQ_UART0_RX, 62 .start = IRQ_UART0_ERROR, 67 .start = CH_UART0_TX, 72 .start = CH_UART0_RX, 78 .start = GPIO_PG7, 83 .start = GPIO_PG6, 107 .start = UART1_THR, 112 .start = IRQ_UART1_TX, 117 .start = IRQ_UART1_RX, 122 .start = IRQ_UART1_ERROR, 127 .start = CH_UART1_TX, 132 .start = CH_UART1_RX, 155 .start = UART2_THR, 160 .start = IRQ_UART2_TX, 165 .start = IRQ_UART2_RX, 170 .start = IRQ_UART2_ERROR, 175 .start = CH_UART2_TX, 180 .start = CH_UART2_RX, 206 .start = 0xFFC00400, 211 .start = IRQ_UART0_RX, 216 .start = CH_UART0_RX, 231 .start = 0xFFC02000, 236 .start = IRQ_UART1_RX, 241 .start = CH_UART1_RX, 256 .start = 0xFFC02100, 261 .start = IRQ_UART2_RX, 266 .start = CH_UART2_RX, 284 .start = SPORT0_TCR1, 289 .start = IRQ_SPORT0_RX, 294 .start = IRQ_SPORT0_ERROR, 318 .start = SPORT1_TCR1, 323 .start = IRQ_SPORT1_RX, 328 .start = IRQ_SPORT1_ERROR, 352 .start = SPORT2_TCR1, 357 .start = IRQ_SPORT2_RX, 362 .start = IRQ_SPORT2_ERROR, 386 .start = SPORT3_TCR1, 391 .start = IRQ_SPORT3_RX, 396 .start = IRQ_SPORT3_ERROR, 426 .start = 0xFFC02A00, 431 .start = IRQ_CAN_RX, 436 .start = IRQ_CAN_TX, 441 .start = IRQ_CAN_ERROR, 473 .start = 0x20310300, 477 .start = IRQ_PF0, 552 .start = IRQ_PPI_ERROR, 615 .start = SPI0_REGBASE, 620 .start = CH_SPI0, 625 .start = IRQ_SPI0, 634 .start = SPI1_REGBASE, 639 .start = CH_SPI1, 644 .start = IRQ_SPI1, 653 .start = SPI2_REGBASE, 658 .start = CH_SPI2, 663 .start = IRQ_SPI2, 723 .start = TWI0_REGBASE, 728 .start = IRQ_TWI0, 748 .start = TWI1_REGBASE, 753 .start = IRQ_TWI1, 839 .start = 0x20000000,
|
/linux-4.4.14/arch/microblaze/kernel/cpu/ |
H A D | cache.c | 92 #define CACHE_LOOP_LIMITS(start, end, cache_line_length, cache_size) \ 95 end = min(start + cache_size, end); \ 96 start &= align; \ 119 * start address is cache aligned 124 #define CACHE_RANGE_LOOP_2(start, end, line_length, op) \ 130 count = end - start; \ 136 : : "r" (start), "r" (count), \ 141 #define CACHE_RANGE_LOOP_1(start, end, line_length, op) \ 146 WARN_ON(end < start); \ 152 : : "r" (temp), "r" (start), "r" (end), \ 158 static void __flush_icache_range_msr_irq(unsigned long start, unsigned long end) __flush_icache_range_msr_irq() argument 164 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __flush_icache_range_msr_irq() 165 (unsigned int)start, (unsigned int) end); __flush_icache_range_msr_irq() 167 CACHE_LOOP_LIMITS(start, end, __flush_icache_range_msr_irq() 174 CACHE_RANGE_LOOP_1(start, end, cpuinfo.icache_line_length, wic); __flush_icache_range_msr_irq() 176 for (i = start; i < end; i += cpuinfo.icache_line_length) __flush_icache_range_msr_irq() 184 static void __flush_icache_range_nomsr_irq(unsigned long start, __flush_icache_range_nomsr_irq() argument 191 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __flush_icache_range_nomsr_irq() 192 (unsigned int)start, (unsigned int) end); __flush_icache_range_nomsr_irq() 194 CACHE_LOOP_LIMITS(start, end, __flush_icache_range_nomsr_irq() 201 CACHE_RANGE_LOOP_1(start, end, cpuinfo.icache_line_length, wic); __flush_icache_range_nomsr_irq() 203 for (i = start; i < end; i += cpuinfo.icache_line_length) __flush_icache_range_nomsr_irq() 212 static void __flush_icache_range_noirq(unsigned long start, __flush_icache_range_noirq() argument 218 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __flush_icache_range_noirq() 219 (unsigned int)start, (unsigned int) end); __flush_icache_range_noirq() 221 CACHE_LOOP_LIMITS(start, end, __flush_icache_range_noirq() 224 CACHE_RANGE_LOOP_1(start, end, cpuinfo.icache_line_length, wic); __flush_icache_range_noirq() 226 for (i = start; i < end; i += cpuinfo.icache_line_length) __flush_icache_range_noirq() 376 static void __invalidate_dcache_range_wb(unsigned long start, __invalidate_dcache_range_wb() argument 382 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __invalidate_dcache_range_wb() 383 (unsigned int)start, (unsigned int) end); __invalidate_dcache_range_wb() 385 CACHE_LOOP_LIMITS(start, end, __invalidate_dcache_range_wb() 388 CACHE_RANGE_LOOP_2(start, end, cpuinfo.dcache_line_length, wdc.clear); __invalidate_dcache_range_wb() 390 for (i = start; i < end; i += cpuinfo.dcache_line_length) __invalidate_dcache_range_wb() 396 static void __invalidate_dcache_range_nomsr_wt(unsigned long start, __invalidate_dcache_range_nomsr_wt() argument 402 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __invalidate_dcache_range_nomsr_wt() 403 (unsigned int)start, (unsigned int) end); __invalidate_dcache_range_nomsr_wt() 404 CACHE_LOOP_LIMITS(start, end, __invalidate_dcache_range_nomsr_wt() 408 CACHE_RANGE_LOOP_1(start, end, cpuinfo.dcache_line_length, wdc); __invalidate_dcache_range_nomsr_wt() 410 for (i = start; i < end; i += cpuinfo.dcache_line_length) __invalidate_dcache_range_nomsr_wt() 416 static void __invalidate_dcache_range_msr_irq_wt(unsigned long start, __invalidate_dcache_range_msr_irq_wt() argument 423 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __invalidate_dcache_range_msr_irq_wt() 424 (unsigned int)start, (unsigned int) end); __invalidate_dcache_range_msr_irq_wt() 425 CACHE_LOOP_LIMITS(start, end, __invalidate_dcache_range_msr_irq_wt() 432 CACHE_RANGE_LOOP_1(start, end, cpuinfo.dcache_line_length, wdc); __invalidate_dcache_range_msr_irq_wt() 434 for (i = start; i < end; i += cpuinfo.dcache_line_length) __invalidate_dcache_range_msr_irq_wt() 443 static void __invalidate_dcache_range_nomsr_irq(unsigned long start, __invalidate_dcache_range_nomsr_irq() argument 450 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __invalidate_dcache_range_nomsr_irq() 451 (unsigned int)start, (unsigned int) end); __invalidate_dcache_range_nomsr_irq() 453 CACHE_LOOP_LIMITS(start, end, __invalidate_dcache_range_nomsr_irq() 460 CACHE_RANGE_LOOP_1(start, end, cpuinfo.dcache_line_length, wdc); __invalidate_dcache_range_nomsr_irq() 462 for (i = start; i < end; i += cpuinfo.dcache_line_length) __invalidate_dcache_range_nomsr_irq() 488 static void __flush_dcache_range_wb(unsigned long start, unsigned long end) __flush_dcache_range_wb() argument 493 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __flush_dcache_range_wb() 494 (unsigned int)start, (unsigned int) end); __flush_dcache_range_wb() 496 CACHE_LOOP_LIMITS(start, end, __flush_dcache_range_wb() 499 CACHE_RANGE_LOOP_2(start, end, cpuinfo.dcache_line_length, wdc.flush); __flush_dcache_range_wb() 501 for (i = start; i < end; i += cpuinfo.dcache_line_length) __flush_dcache_range_wb()
|
/linux-4.4.14/arch/c6x/platforms/ |
H A D | cache.c | 132 static void cache_block_operation(unsigned int *start, cache_block_operation() argument 140 - L2_CACHE_ALIGN_LOW((unsigned int) start)) >> 2; cache_block_operation() 143 for (; wcnt; wcnt -= wc, start += wc) { cache_block_operation() 160 imcr_set(bar_reg, L2_CACHE_ALIGN_LOW((unsigned int) start)); cache_block_operation() 177 static void cache_block_operation_nowait(unsigned int *start, cache_block_operation_nowait() argument 185 - L2_CACHE_ALIGN_LOW((unsigned int) start)) >> 2; cache_block_operation_nowait() 188 for (; wcnt; wcnt -= wc, start += wc) { cache_block_operation_nowait() 192 imcr_set(bar_reg, L2_CACHE_ALIGN_LOW((unsigned int) start)); cache_block_operation_nowait() 325 void enable_caching(unsigned long start, unsigned long end) enable_caching() argument 327 unsigned int mar = IMCR_MAR_BASE + ((start >> 24) << 2); enable_caching() 334 void disable_caching(unsigned long start, unsigned long end) disable_caching() argument 336 unsigned int mar = IMCR_MAR_BASE + ((start >> 24) << 2); disable_caching() 347 void L1P_cache_block_invalidate(unsigned int start, unsigned int end) L1P_cache_block_invalidate() argument 349 cache_block_operation((unsigned int *) start, L1P_cache_block_invalidate() 355 void L1D_cache_block_invalidate(unsigned int start, unsigned int end) L1D_cache_block_invalidate() argument 357 cache_block_operation((unsigned int *) start, L1D_cache_block_invalidate() 362 void L1D_cache_block_writeback_invalidate(unsigned int start, unsigned int end) L1D_cache_block_writeback_invalidate() argument 364 cache_block_operation((unsigned int *) start, L1D_cache_block_writeback_invalidate() 369 void L1D_cache_block_writeback(unsigned int start, unsigned int end) L1D_cache_block_writeback() argument 371 cache_block_operation((unsigned int *) start, L1D_cache_block_writeback() 380 void L2_cache_block_invalidate(unsigned int start, unsigned int end) L2_cache_block_invalidate() argument 382 cache_block_operation((unsigned int *) start, L2_cache_block_invalidate() 387 void L2_cache_block_writeback(unsigned int start, unsigned int end) L2_cache_block_writeback() argument 389 cache_block_operation((unsigned int *) start, L2_cache_block_writeback() 394 void L2_cache_block_writeback_invalidate(unsigned int start, unsigned int end) L2_cache_block_writeback_invalidate() argument 396 cache_block_operation((unsigned int *) start, L2_cache_block_writeback_invalidate() 401 void L2_cache_block_invalidate_nowait(unsigned int start, unsigned int end) L2_cache_block_invalidate_nowait() argument 403 cache_block_operation_nowait((unsigned int *) start, L2_cache_block_invalidate_nowait() 408 void L2_cache_block_writeback_nowait(unsigned int start, unsigned int end) L2_cache_block_writeback_nowait() argument 410 cache_block_operation_nowait((unsigned int *) start, L2_cache_block_writeback_nowait() 415 void L2_cache_block_writeback_invalidate_nowait(unsigned int start, L2_cache_block_writeback_invalidate_nowait() argument 418 cache_block_operation_nowait((unsigned int *) start, L2_cache_block_writeback_invalidate_nowait()
|
/linux-4.4.14/arch/cris/boot/rescue/ |
H A D | head_v32.S | 2 * Just get started by jumping to CONFIG_ETRAX_PTABLE_SECTOR to start 17 start: label
|
/linux-4.4.14/arch/nios2/include/asm/ |
H A D | cacheflush.h | 26 extern void flush_cache_range(struct vm_area_struct *vma, unsigned long start, 33 extern void flush_icache_range(unsigned long start, unsigned long end); 36 #define flush_cache_vmap(start, end) flush_dcache_range(start, end) 37 #define flush_cache_vunmap(start, end) flush_dcache_range(start, end) 46 extern void flush_dcache_range(unsigned long start, unsigned long end); 47 extern void invalidate_dcache_range(unsigned long start, unsigned long end);
|
/linux-4.4.14/fs/btrfs/ |
H A D | extent_io.h | 71 u64 start, u64 end, int *page_started, 73 int (*writepage_start_hook)(struct page *page, u64 start, u64 end); 74 int (*writepage_io_hook)(struct page *page, u64 start, u64 end); 81 struct page *page, u64 start, u64 end, 83 int (*writepage_end_io_hook)(struct page *page, u64 start, u64 end, 106 u64 start; member in struct:extent_state 126 u64 start; member in struct:extent_buffer 193 u64 start, u64 len, 202 int lock_extent(struct extent_io_tree *tree, u64 start, u64 end); 203 int lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, 205 int unlock_extent(struct extent_io_tree *tree, u64 start, u64 end); 206 int unlock_extent_cached(struct extent_io_tree *tree, u64 start, u64 end, 208 int try_lock_extent(struct extent_io_tree *tree, u64 start, u64 end); 217 u64 *start, u64 search_end, 221 int test_range_bit(struct extent_io_tree *tree, u64 start, u64 end, 224 int clear_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, 226 int clear_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, 229 int clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, 232 int set_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, 234 int set_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, 237 int set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, 240 int set_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end, 242 int clear_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end, 244 int set_extent_new(struct extent_io_tree *tree, u64 start, u64 end, 246 int set_extent_dirty(struct extent_io_tree *tree, u64 start, u64 end, 248 int clear_extent_dirty(struct extent_io_tree *tree, u64 start, u64 end, 250 int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, 253 int set_extent_delalloc(struct extent_io_tree *tree, u64 start, u64 end, 255 int set_extent_defrag(struct extent_io_tree *tree, u64 start, u64 end, 257 int find_first_extent_bit(struct extent_io_tree *tree, u64 start, 266 u64 start, u64 end, get_extent_t *get_extent, 279 __u64 start, __u64 len, get_extent_t *get_extent); 280 int get_state_private(struct extent_io_tree *tree, u64 start, u64 *private); 284 u64 start); 286 u64 start); 289 u64 start); 296 struct extent_buffer *eb, u64 start, int wait, 300 static inline unsigned long num_extent_pages(u64 start, u64 len) num_extent_pages() argument 302 return ((start + len + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT) - num_extent_pages() 303 (start >> PAGE_CACHE_SHIFT); num_extent_pages() 312 unsigned long start, 315 unsigned long start, 318 unsigned long start, 321 unsigned long start, unsigned long len); 330 unsigned long start, unsigned long len); 341 int extent_range_clear_dirty_for_io(struct inode *inode, u64 start, u64 end); 342 int extent_range_redirty_for_io(struct inode *inode, u64 start, u64 end); 343 int extent_clear_unlock_delalloc(struct inode *inode, u64 start, u64 end, 355 int repair_io_failure(struct inode *inode, u64 start, u64 length, u64 logical, 358 int clean_io_failure(struct inode *inode, u64 start, struct page *page, 360 int end_extent_writepage(struct page *page, int err, u64 start, u64 end); 374 u64 start; member in struct:io_failure_record 383 void btrfs_free_io_failure_record(struct inode *inode, u64 start, u64 end); 384 int btrfs_get_io_failure_record(struct inode *inode, u64 start, u64 end, 396 struct page *locked_page, u64 *start, 400 u64 start);
|
H A D | extent_io.c | 67 pr_err("BTRFS: state leak: start %llu end %llu state %u in tree %d refs %d\n", btrfs_leak_debug_check() 68 state->start, state->end, state->state, btrfs_leak_debug_check() 77 printk(KERN_ERR "BTRFS: buffer leak start %llu len %lu " btrfs_leak_debug_check() 79 eb->start, eb->len, atomic_read(&eb->refs)); btrfs_leak_debug_check() 85 #define btrfs_debug_check_extent_io_range(tree, start, end) \ 86 __btrfs_debug_check_extent_io_range(__func__, (tree), (start), (end)) __btrfs_debug_check_extent_io_range() 88 struct extent_io_tree *tree, u64 start, u64 end) __btrfs_debug_check_extent_io_range() 101 caller, btrfs_ino(inode), isize, start, end); __btrfs_debug_check_extent_io_range() 114 u64 start; member in struct:tree_entry 146 changeset->bytes_changed += state->end - state->start + 1; add_extent_changeset() 147 ret = ulist_add(changeset->range_changed, state->start, state->end, add_extent_changeset() 278 if (offset < entry->start) tree_insert() 310 if (offset < entry->start) __etree_search() 335 while (prev && offset < prev_entry->start) { __etree_search() 394 if (other->end == state->start - 1 && merge_state() 397 state->start = other->start; merge_state() 406 if (other->start == state->end + 1 && merge_state() 446 struct extent_state *state, u64 start, u64 end, insert_state() 453 if (end < start) insert_state() 454 WARN(1, KERN_ERR "BTRFS: end < start %llu %llu\n", insert_state() 455 end, start); insert_state() 456 state->start = start; insert_state() 467 found->start, found->end, start, end); insert_state() 487 * the tree has 'orig' at [orig->start, orig->end]. After calling, there 489 * prealloc: [orig->start, split - 1] 502 prealloc->start = orig->start; split_state() 505 orig->start = split; split_state() 541 u64 range = state->end - state->start + 1; clear_state_bit() 590 * the range [start, end] is inclusive. 594 static int __clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, __clear_extent_bit() argument 607 btrfs_debug_check_extent_io_range(tree, start, end); __clear_extent_bit() 640 cached->start <= start && cached->end > start) { __clear_extent_bit() 653 node = tree_search(tree, start); __clear_extent_bit() 658 if (state->start > end) __clear_extent_bit() 660 WARN_ON(state->end < start); __clear_extent_bit() 685 if (state->start < start) { __clear_extent_bit() 688 err = split_state(tree, state, prealloc, start); __clear_extent_bit() 708 if (state->start <= end && state->end > end) { __clear_extent_bit() 728 start = last_end + 1; __clear_extent_bit() 729 if (start <= end && state && !need_resched()) __clear_extent_bit() 741 if (start > end) __clear_extent_bit() 764 * The range [start, end] is inclusive. 767 static void wait_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, wait_extent_bit() argument 773 btrfs_debug_check_extent_io_range(tree, start, end); wait_extent_bit() 782 node = tree_search(tree, start); wait_extent_bit() 789 if (state->start > end) wait_extent_bit() 793 start = state->start; wait_extent_bit() 799 start = state->end + 1; wait_extent_bit() 801 if (start > end) wait_extent_bit() 821 u64 range = state->end - state->start + 1; set_state_bits() 852 * part of the range already has the desired bits set. The start of the 855 * [start, end] is inclusive This takes the tree lock. 859 __set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, __set_extent_bit() argument 873 btrfs_debug_check_extent_io_range(tree, start, end); __set_extent_bit() 885 if (state->start <= start && state->end > start && __set_extent_bit() 895 node = tree_search_for_insert(tree, start, &p, &parent); __set_extent_bit() 899 err = insert_state(tree, prealloc, start, end, __set_extent_bit() 910 last_start = state->start; __set_extent_bit() 919 if (state->start == start && state->end <= end) { __set_extent_bit() 921 *failed_start = state->start; __set_extent_bit() 931 start = last_end + 1; __set_extent_bit() 933 if (start < end && state && state->start == start && __set_extent_bit() 955 if (state->start < start) { __set_extent_bit() 957 *failed_start = start; __set_extent_bit() 964 err = split_state(tree, state, prealloc, start); __set_extent_bit() 977 start = last_end + 1; __set_extent_bit() 979 if (start < end && state && state->start == start && __set_extent_bit() 992 if (state->start > start) { __set_extent_bit() 1006 err = insert_state(tree, prealloc, start, this_end, __set_extent_bit() 1013 start = this_end + 1; __set_extent_bit() 1022 if (state->start <= end && state->end > end) { __set_extent_bit() 1024 *failed_start = start; __set_extent_bit() 1052 if (start > end) __set_extent_bit() 1060 int set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, set_extent_bit() argument 1064 return __set_extent_bit(tree, start, end, bits, 0, failed_start, set_extent_bit() 1073 * @start: the start offset in bytes 1086 int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, convert_extent_bit() argument 1100 btrfs_debug_check_extent_io_range(tree, start, end); convert_extent_bit() 1119 if (state->start <= start && state->end > start && convert_extent_bit() 1130 node = tree_search_for_insert(tree, start, &p, &parent); convert_extent_bit() 1137 err = insert_state(tree, prealloc, start, end, convert_extent_bit() 1147 last_start = state->start; convert_extent_bit() 1156 if (state->start == start && state->end <= end) { convert_extent_bit() 1162 start = last_end + 1; convert_extent_bit() 1163 if (start < end && state && state->start == start && convert_extent_bit() 1185 if (state->start < start) { convert_extent_bit() 1191 err = split_state(tree, state, prealloc, start); convert_extent_bit() 1204 start = last_end + 1; convert_extent_bit() 1205 if (start < end && state && state->start == start && convert_extent_bit() 1218 if (state->start > start) { convert_extent_bit() 1235 err = insert_state(tree, prealloc, start, this_end, convert_extent_bit() 1241 start = this_end + 1; convert_extent_bit() 1250 if (state->start <= end && state->end > end) { convert_extent_bit() 1278 if (start > end) convert_extent_bit() 1288 int set_extent_dirty(struct extent_io_tree *tree, u64 start, u64 end, set_extent_dirty() argument 1291 return set_extent_bit(tree, start, end, EXTENT_DIRTY, NULL, set_extent_dirty() 1295 int set_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, set_extent_bits() argument 1298 return set_extent_bit(tree, start, end, bits, NULL, set_extent_bits() 1302 int set_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, set_record_extent_bits() argument 1314 return __set_extent_bit(tree, start, end, bits, 0, NULL, NULL, mask, set_record_extent_bits() 1318 int clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, clear_extent_bit() argument 1322 return __clear_extent_bit(tree, start, end, bits, wake, delete, clear_extent_bit() 1326 int clear_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, clear_extent_bits() argument 1334 return clear_extent_bit(tree, start, end, bits, wake, 0, NULL, mask); clear_extent_bits() 1337 int clear_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, clear_record_extent_bits() argument 1347 return __clear_extent_bit(tree, start, end, bits, 0, 0, NULL, mask, clear_record_extent_bits() 1351 int set_extent_delalloc(struct extent_io_tree *tree, u64 start, u64 end, set_extent_delalloc() argument 1354 return set_extent_bit(tree, start, end, set_extent_delalloc() 1359 int set_extent_defrag(struct extent_io_tree *tree, u64 start, u64 end, set_extent_defrag() argument 1362 return set_extent_bit(tree, start, end, set_extent_defrag() 1367 int clear_extent_dirty(struct extent_io_tree *tree, u64 start, u64 end, clear_extent_dirty() argument 1370 return clear_extent_bit(tree, start, end, clear_extent_dirty() 1375 int set_extent_new(struct extent_io_tree *tree, u64 start, u64 end, set_extent_new() argument 1378 return set_extent_bit(tree, start, end, EXTENT_NEW, NULL, set_extent_new() 1382 int set_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end, set_extent_uptodate() argument 1385 return set_extent_bit(tree, start, end, EXTENT_UPTODATE, NULL, set_extent_uptodate() 1389 int clear_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end, clear_extent_uptodate() argument 1392 return clear_extent_bit(tree, start, end, EXTENT_UPTODATE, 0, 0, clear_extent_uptodate() 1397 * either insert or lock state struct between start and end use mask to tell 1400 int lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, lock_extent_bits() argument 1407 err = __set_extent_bit(tree, start, end, EXTENT_LOCKED | bits, lock_extent_bits() 1412 start = failed_start; lock_extent_bits() 1415 WARN_ON(start > end); lock_extent_bits() 1420 int lock_extent(struct extent_io_tree *tree, u64 start, u64 end) lock_extent() argument 1422 return lock_extent_bits(tree, start, end, 0, NULL); lock_extent() 1425 int try_lock_extent(struct extent_io_tree *tree, u64 start, u64 end) try_lock_extent() argument 1430 err = __set_extent_bit(tree, start, end, EXTENT_LOCKED, EXTENT_LOCKED, try_lock_extent() 1433 if (failed_start > start) try_lock_extent() 1434 clear_extent_bit(tree, start, failed_start - 1, try_lock_extent() 1441 int unlock_extent_cached(struct extent_io_tree *tree, u64 start, u64 end, unlock_extent_cached() argument 1444 return clear_extent_bit(tree, start, end, EXTENT_LOCKED, 1, 0, cached, unlock_extent_cached() 1448 int unlock_extent(struct extent_io_tree *tree, u64 start, u64 end) unlock_extent() argument 1450 return clear_extent_bit(tree, start, end, EXTENT_LOCKED, 1, 0, NULL, unlock_extent() 1454 int extent_range_clear_dirty_for_io(struct inode *inode, u64 start, u64 end) extent_range_clear_dirty_for_io() argument 1456 unsigned long index = start >> PAGE_CACHE_SHIFT; extent_range_clear_dirty_for_io() 1470 int extent_range_redirty_for_io(struct inode *inode, u64 start, u64 end) extent_range_redirty_for_io() argument 1472 unsigned long index = start >> PAGE_CACHE_SHIFT; extent_range_redirty_for_io() 1490 static int set_range_writeback(struct extent_io_tree *tree, u64 start, u64 end) set_range_writeback() argument 1492 unsigned long index = start >> PAGE_CACHE_SHIFT; set_range_writeback() 1506 /* find the first state struct with 'bits' set after 'start', and 1508 * nothing was found after 'start' 1512 u64 start, unsigned bits) find_first_extent_bit_state() 1521 node = tree_search(tree, start); find_first_extent_bit_state() 1527 if (state->end >= start && (state->state & bits)) find_first_extent_bit_state() 1545 int find_first_extent_bit(struct extent_io_tree *tree, u64 start, find_first_extent_bit() argument 1556 if (state->end == start - 1 && extent_state_in_tree(state)) { find_first_extent_bit() 1573 state = find_first_extent_bit_state(tree, start, bits); find_first_extent_bit() 1577 *start_ret = state->start; find_first_extent_bit() 1588 * more than 'max_bytes'. start and end are used to return the range, 1593 u64 *start, u64 *end, u64 max_bytes, find_delalloc_range() 1598 u64 cur_start = *start; find_delalloc_range() 1617 if (found && (state->start != cur_start || find_delalloc_range() 1627 *start = state->start; find_delalloc_range() 1635 total_bytes += state->end - state->start + 1; find_delalloc_range() 1648 u64 start, u64 end) __unlock_for_delalloc() 1652 unsigned long index = start >> PAGE_CACHE_SHIFT; __unlock_for_delalloc() 1689 /* the caller is responsible for locking the start index */ lock_delalloc_pages() 1693 /* skip the page at the start index */ lock_delalloc_pages() 1739 * more than 'max_bytes'. start and end are used to return the range, 1745 struct page *locked_page, u64 *start, find_lock_delalloc_range() 1756 /* step one, find a bunch of delalloc bytes starting at start */ find_lock_delalloc_range() 1757 delalloc_start = *start; find_lock_delalloc_range() 1761 if (!found || delalloc_end <= *start) { find_lock_delalloc_range() 1762 *start = delalloc_start; find_lock_delalloc_range() 1769 * start comes from the offset of locked_page. We have to lock find_lock_delalloc_range() 1773 if (delalloc_start < *start) find_lock_delalloc_range() 1774 delalloc_start = *start; find_lock_delalloc_range() 1782 /* step two, lock all the pages after the page that has start */ find_lock_delalloc_range() 1817 *start = delalloc_start; find_lock_delalloc_range() 1823 int extent_clear_unlock_delalloc(struct inode *inode, u64 start, u64 end, extent_clear_unlock_delalloc() argument 1831 unsigned long index = start >> PAGE_CACHE_SHIFT; extent_clear_unlock_delalloc() 1836 clear_extent_bit(tree, start, end, clear_bits, 1, 0, NULL, GFP_NOFS); extent_clear_unlock_delalloc() 1881 u64 *start, u64 search_end, u64 max_bytes, count_range_bits() 1886 u64 cur_start = *start; count_range_bits() 1909 if (state->start > search_end) count_range_bits() 1911 if (contig && found && state->start > last + 1) count_range_bits() 1915 max(cur_start, state->start); count_range_bits() 1919 *start = max(cur_start, state->start); count_range_bits() 1939 static int set_state_private(struct extent_io_tree *tree, u64 start, u64 private) set_state_private() argument 1950 node = tree_search(tree, start); set_state_private() 1956 if (state->start != start) { set_state_private() 1966 int get_state_private(struct extent_io_tree *tree, u64 start, u64 *private) get_state_private() argument 1977 node = tree_search(tree, start); get_state_private() 1983 if (state->start != start) { get_state_private() 1999 int test_range_bit(struct extent_io_tree *tree, u64 start, u64 end, test_range_bit() argument 2007 if (cached && extent_state_in_tree(cached) && cached->start <= start && test_range_bit() 2008 cached->end > start) test_range_bit() 2011 node = tree_search(tree, start); test_range_bit() 2012 while (node && start <= end) { test_range_bit() 2015 if (filled && state->start > start) { test_range_bit() 2020 if (state->start > end) test_range_bit() 2035 start = state->end + 1; test_range_bit() 2036 if (start > end) test_range_bit() 2055 u64 start = page_offset(page); check_page_uptodate() local 2056 u64 end = start + PAGE_CACHE_SIZE - 1; check_page_uptodate() 2057 if (test_range_bit(tree, start, end, EXTENT_UPTODATE, 1, NULL)) check_page_uptodate() 2067 set_state_private(failure_tree, rec->start, 0); free_io_failure() 2068 ret = clear_extent_bits(failure_tree, rec->start, free_io_failure() 2069 rec->start + rec->len - 1, free_io_failure() 2074 ret = clear_extent_bits(&BTRFS_I(inode)->io_tree, rec->start, free_io_failure() 2075 rec->start + rec->len - 1, free_io_failure() 2094 int repair_io_failure(struct inode *inode, u64 start, u64 length, u64 logical, repair_io_failure() argument 2146 btrfs_ino(inode), start, repair_io_failure() 2155 u64 start = eb->start; repair_eb_io_failure() local 2156 unsigned long i, num_pages = num_extent_pages(eb->start, eb->len); repair_eb_io_failure() 2165 ret = repair_io_failure(root->fs_info->btree_inode, start, repair_eb_io_failure() 2166 PAGE_CACHE_SIZE, start, p, repair_eb_io_failure() 2167 start - page_offset(p), mirror_num); repair_eb_io_failure() 2170 start += PAGE_CACHE_SIZE; repair_eb_io_failure() 2180 int clean_io_failure(struct inode *inode, u64 start, struct page *page, clean_io_failure() argument 2197 ret = get_state_private(&BTRFS_I(inode)->io_failure_tree, start, clean_io_failure() 2208 failrec->start); clean_io_failure() 2216 failrec->start, clean_io_failure() 2220 if (state && state->start <= failrec->start && clean_io_failure() 2221 state->end >= failrec->start + failrec->len - 1) { clean_io_failure() 2225 repair_io_failure(inode, start, failrec->len, clean_io_failure() 2243 void btrfs_free_io_failure_record(struct inode *inode, u64 start, u64 end) btrfs_free_io_failure_record() argument 2253 state = find_first_extent_bit_state(failure_tree, start, EXTENT_DIRTY); btrfs_free_io_failure_record() 2255 if (state->start > end) btrfs_free_io_failure_record() 2271 int btrfs_get_io_failure_record(struct inode *inode, u64 start, u64 end, btrfs_get_io_failure_record() argument 2283 ret = get_state_private(failure_tree, start, &private); btrfs_get_io_failure_record() 2289 failrec->start = start; btrfs_get_io_failure_record() 2290 failrec->len = end - start + 1; btrfs_get_io_failure_record() 2296 em = lookup_extent_mapping(em_tree, start, failrec->len); btrfs_get_io_failure_record() 2303 if (em->start > start || em->start + em->len <= start) { btrfs_get_io_failure_record() 2313 logical = start - em->start; btrfs_get_io_failure_record() 2322 pr_debug("Get IO Failure Record: (new) logical=%llu, start=%llu, len=%llu\n", btrfs_get_io_failure_record() 2323 logical, start, failrec->len); btrfs_get_io_failure_record() 2329 ret = set_extent_bits(failure_tree, start, end, btrfs_get_io_failure_record() 2332 ret = set_state_private(failure_tree, start, btrfs_get_io_failure_record() 2336 ret = set_extent_bits(tree, start, end, EXTENT_DAMAGED, btrfs_get_io_failure_record() 2344 pr_debug("Get IO Failure Record: (found) logical=%llu, start=%llu, len=%llu, validation=%d\n", btrfs_get_io_failure_record() 2345 failrec->logical, failrec->start, failrec->len, btrfs_get_io_failure_record() 2466 struct page *page, u64 start, u64 end, bio_readpage_error() 2478 ret = btrfs_get_io_failure_record(inode, start, end, &failrec); bio_readpage_error() 2495 start - page_offset(page), bio_readpage_error() 2519 int end_extent_writepage(struct page *page, int err, u64 start, u64 end) end_extent_writepage() argument 2528 ret = tree->ops->writepage_end_io_hook(page, start, end_extent_writepage() 2555 u64 start; end_bio_extent_writepage() local 2579 start = page_offset(page); bio_for_each_segment_all() 2580 end = start + bvec->bv_offset + bvec->bv_len - 1; bio_for_each_segment_all() 2582 if (end_extent_writepage(page, bio->bi_error, start, end)) bio_for_each_segment_all() 2592 endio_readpage_release_extent(struct extent_io_tree *tree, u64 start, u64 len, endio_readpage_release_extent() argument 2596 u64 end = start + len - 1; endio_readpage_release_extent() 2599 set_extent_uptodate(tree, start, end, &cached, GFP_ATOMIC); endio_readpage_release_extent() 2600 unlock_extent_cached(tree, start, end, &cached, GFP_ATOMIC); endio_readpage_release_extent() 2621 u64 start; end_bio_extent_readpage() local 2656 start = page_offset(page); bio_for_each_segment_all() 2657 end = start + bvec->bv_offset + bvec->bv_len - 1; bio_for_each_segment_all() 2664 page, start, end, bio_for_each_segment_all() 2669 clean_io_failure(inode, start, page, 0); bio_for_each_segment_all() 2690 ret = bio_readpage_error(bio, offset, page, start, end, bio_for_each_segment_all() 2724 endio_readpage_release_extent(tree, start, bio_for_each_segment_all() 2725 end - start + 1, 0); bio_for_each_segment_all() 2727 extent_start = start; bio_for_each_segment_all() 2728 extent_len = end + 1 - start; bio_for_each_segment_all() 2729 } else if (extent_start + extent_len == start) { bio_for_each_segment_all() 2730 extent_len += end + 1 - start; bio_for_each_segment_all() 2734 extent_start = start; bio_for_each_segment_all() 2735 extent_len = end + 1 - start; bio_for_each_segment_all() 2823 u64 start; submit_one_bio() local 2825 start = page_offset(page) + bvec->bv_offset; submit_one_bio() 2833 mirror_num, bio_flags, start); submit_one_bio() 2942 u64 start, u64 len, get_extent_t *get_extent, __get_extent_map() 2949 if (extent_map_in_tree(em) && start >= em->start && __get_extent_map() 2950 start < extent_map_end(em)) { __get_extent_map() 2959 em = get_extent(inode, page, pg_offset, start, len, 0); __get_extent_map() 2982 u64 start = page_offset(page); __do_readpage() local 2983 u64 page_end = start + PAGE_CACHE_SIZE - 1; __do_readpage() 2985 u64 cur = start; __do_readpage() 3008 unlock_extent(tree, start, end); __do_readpage() 3054 extent_offset = cur - em->start; __do_readpage() 3198 u64 start, u64 end, __do_contiguous_readpages() 3211 lock_extent(tree, start, end); __do_contiguous_readpages() 3212 ordered = btrfs_lookup_ordered_range(inode, start, __do_contiguous_readpages() 3213 end - start + 1); __do_contiguous_readpages() 3216 unlock_extent(tree, start, end); __do_contiguous_readpages() 3236 u64 start = 0; __extent_readpages() local 3245 start = page_start; __extent_readpages() 3246 end = start + PAGE_CACHE_SIZE - 1; __extent_readpages() 3252 index - first_index, start, __extent_readpages() 3256 start = page_start; __extent_readpages() 3257 end = start + PAGE_CACHE_SIZE - 1; __extent_readpages() 3264 index - first_index, start, __extent_readpages() 3278 u64 start = page_offset(page); __extent_read_full_page() local 3279 u64 end = start + PAGE_CACHE_SIZE - 1; __extent_read_full_page() 3283 lock_extent(tree, start, end); __extent_read_full_page() 3284 ordered = btrfs_lookup_ordered_extent(inode, start); __extent_read_full_page() 3287 unlock_extent(tree, start, end); __extent_read_full_page() 3407 /* did the fill delalloc function already unlock and start writepage_delalloc() 3427 * helper for __extent_writepage. This calls the writepage start hooks, 3443 u64 start = page_offset(page); __extent_writepage_io() local 3444 u64 page_end = start + PAGE_CACHE_SIZE - 1; __extent_writepage_io() 3446 u64 cur = start; __extent_writepage_io() 3461 ret = tree->ops->writepage_start_hook(page, start, __extent_writepage_io() 3484 if (i_size <= start) { __extent_writepage_io() 3486 tree->ops->writepage_end_io_hook(page, start, __extent_writepage_io() 3509 extent_offset = cur - em->start; __extent_writepage_io() 3601 u64 start = page_offset(page); __extent_writepage() local 3602 u64 page_end = start + PAGE_CACHE_SIZE - 1; __extent_writepage() 3644 ret = writepage_delalloc(inode, page, wbc, epd, start, &nr_written); __extent_writepage() 3663 end_extent_writepage(page, ret, start, page_end); __extent_writepage() 3733 num_pages = num_extent_pages(eb->start, eb->len); lock_extent_buffer_for_io() 3855 u64 offset = eb->start; write_one_eb() 3862 num_pages = num_extent_pages(eb->start, eb->len); write_one_eb() 4013 * back to the start of the file btree_write_cache_pages() 4150 * back to the start of the file extent_write_cache_pages() 4202 u64 start, u64 end, get_extent_t *get_extent, extent_write_locked_range() 4208 unsigned long nr_pages = (end - start + PAGE_CACHE_SIZE) >> extent_write_locked_range() 4222 .range_start = start, extent_write_locked_range() 4226 while (start <= end) { extent_write_locked_range() 4227 page = find_get_page(mapping, start >> PAGE_CACHE_SHIFT); extent_write_locked_range() 4232 tree->ops->writepage_end_io_hook(page, start, extent_write_locked_range() 4233 start + PAGE_CACHE_SIZE - 1, extent_write_locked_range() 4238 start += PAGE_CACHE_SIZE; extent_write_locked_range() 4321 u64 start = page_offset(page); extent_invalidatepage() local 4322 u64 end = start + PAGE_CACHE_SIZE - 1; extent_invalidatepage() 4325 start += ALIGN(offset, blocksize); extent_invalidatepage() 4326 if (start > end) extent_invalidatepage() 4329 lock_extent_bits(tree, start, end, 0, &cached_state); extent_invalidatepage() 4331 clear_extent_bit(tree, start, end, extent_invalidatepage() 4347 u64 start = page_offset(page); try_release_extent_state() local 4348 u64 end = start + PAGE_CACHE_SIZE - 1; try_release_extent_state() 4351 if (test_range_bit(tree, start, end, try_release_extent_state() 4361 ret = clear_extent_bit(tree, start, end, try_release_extent_state() 4386 u64 start = page_offset(page); try_release_extent_mapping() local 4387 u64 end = start + PAGE_CACHE_SIZE - 1; try_release_extent_mapping() 4392 while (start <= end) { try_release_extent_mapping() 4393 len = end - start + 1; try_release_extent_mapping() 4395 em = lookup_extent_mapping(map, start, len); try_release_extent_mapping() 4401 em->start != start) { try_release_extent_mapping() 4406 if (!test_range_bit(tree, em->start, try_release_extent_mapping() 4414 start = extent_map_end(em); try_release_extent_mapping() 4465 __u64 start, __u64 len, get_extent_t *get_extent) extent_fiemap() 4468 u64 off = start; extent_fiemap() 4469 u64 max = start + len; extent_fiemap() 4494 start = round_down(start, BTRFS_I(inode)->root->sectorsize); extent_fiemap() 4495 len = round_up(max, BTRFS_I(inode)->root->sectorsize) - start; extent_fiemap() 4520 * remember the start of the last extent. There are a extent_fiemap() 4531 * extents. so, we trust isize unless the start of the last extent is extent_fiemap() 4539 lock_extent_bits(&BTRFS_I(inode)->io_tree, start, start + len - 1, 0, extent_fiemap() 4542 em = get_extent_skip_holes(inode, start, last_for_get_extent, extent_fiemap() 4555 if (em->start >= max || extent_map_end(em) < off) extent_fiemap() 4564 em_start = max(em->start, off); extent_fiemap() 4567 * record the offset from the start of the extent extent_fiemap() 4573 offset_in_extent = em_start - em->start; extent_fiemap() 4597 (em->start - em->orig_start); extent_fiemap() 4653 unlock_extent_cached(&BTRFS_I(inode)->io_tree, start, start + len - 1, extent_fiemap() 4682 index = num_extent_pages(eb->start, eb->len); btrfs_release_extent_buffer_page() 4733 __alloc_extent_buffer(struct btrfs_fs_info *fs_info, u64 start, __alloc_extent_buffer() argument 4739 eb->start = start; __alloc_extent_buffer() 4775 unsigned long num_pages = num_extent_pages(src->start, src->len); btrfs_clone_extent_buffer() 4777 new = __alloc_extent_buffer(src->fs_info, src->start, src->len); btrfs_clone_extent_buffer() 4801 u64 start) alloc_dummy_extent_buffer() 4819 eb = __alloc_extent_buffer(fs_info, start, len); alloc_dummy_extent_buffer() 4880 num_pages = num_extent_pages(eb->start, eb->len); mark_extent_buffer_accessed() 4890 u64 start) find_extent_buffer() 4896 start >> PAGE_CACHE_SHIFT); find_extent_buffer() 4928 u64 start) alloc_test_extent_buffer() 4933 eb = find_extent_buffer(fs_info, start); alloc_test_extent_buffer() 4936 eb = alloc_dummy_extent_buffer(fs_info, start); alloc_test_extent_buffer() 4946 start >> PAGE_CACHE_SHIFT, eb); alloc_test_extent_buffer() 4950 exists = find_extent_buffer(fs_info, start); alloc_test_extent_buffer() 4974 u64 start) alloc_extent_buffer() 4977 unsigned long num_pages = num_extent_pages(start, len); alloc_extent_buffer() 4979 unsigned long index = start >> PAGE_CACHE_SHIFT; alloc_extent_buffer() 4987 eb = find_extent_buffer(fs_info, start); alloc_extent_buffer() 4991 eb = __alloc_extent_buffer(fs_info, start, len); alloc_extent_buffer() 5048 start >> PAGE_CACHE_SHIFT, eb); alloc_extent_buffer() 5052 exists = find_extent_buffer(fs_info, start); alloc_extent_buffer() 5111 eb->start >> PAGE_CACHE_SHIFT); release_extent_buffer() 5187 num_pages = num_extent_pages(eb->start, eb->len); clear_extent_buffer_dirty() 5221 num_pages = num_extent_pages(eb->start, eb->len); set_extent_buffer_dirty() 5237 num_pages = num_extent_pages(eb->start, eb->len); clear_extent_buffer_uptodate() 5253 num_pages = num_extent_pages(eb->start, eb->len); set_extent_buffer_uptodate() 5267 struct extent_buffer *eb, u64 start, int wait, read_extent_buffer_pages() 5285 if (start) { read_extent_buffer_pages() 5286 WARN_ON(start < eb->start); read_extent_buffer_pages() 5287 start_i = (start >> PAGE_CACHE_SHIFT) - read_extent_buffer_pages() 5288 (eb->start >> PAGE_CACHE_SHIFT); read_extent_buffer_pages() 5293 num_pages = num_extent_pages(eb->start, eb->len); read_extent_buffer_pages() 5363 unsigned long start, read_extent_buffer() 5371 size_t start_offset = eb->start & ((u64)PAGE_CACHE_SIZE - 1); read_extent_buffer() 5372 unsigned long i = (start_offset + start) >> PAGE_CACHE_SHIFT; read_extent_buffer() 5374 WARN_ON(start > eb->len); read_extent_buffer() 5375 WARN_ON(start + len > eb->start + eb->len); read_extent_buffer() 5377 offset = (start_offset + start) & (PAGE_CACHE_SIZE - 1); read_extent_buffer() 5394 unsigned long start, read_extent_buffer_to_user() 5402 size_t start_offset = eb->start & ((u64)PAGE_CACHE_SIZE - 1); read_extent_buffer_to_user() 5403 unsigned long i = (start_offset + start) >> PAGE_CACHE_SHIFT; read_extent_buffer_to_user() 5406 WARN_ON(start > eb->len); read_extent_buffer_to_user() 5407 WARN_ON(start + len > eb->start + eb->len); read_extent_buffer_to_user() 5409 offset = (start_offset + start) & (PAGE_CACHE_SIZE - 1); read_extent_buffer_to_user() 5430 int map_private_extent_buffer(struct extent_buffer *eb, unsigned long start, map_private_extent_buffer() argument 5435 size_t offset = start & (PAGE_CACHE_SIZE - 1); map_private_extent_buffer() 5438 size_t start_offset = eb->start & ((u64)PAGE_CACHE_SIZE - 1); map_private_extent_buffer() 5439 unsigned long i = (start_offset + start) >> PAGE_CACHE_SHIFT; map_private_extent_buffer() 5440 unsigned long end_i = (start_offset + start + min_len - 1) >> map_private_extent_buffer() 5454 if (start + min_len > eb->len) { map_private_extent_buffer() 5455 WARN(1, KERN_ERR "btrfs bad mapping eb start %llu len %lu, " map_private_extent_buffer() 5457 eb->start, eb->len, start, min_len); map_private_extent_buffer() 5469 unsigned long start, memcmp_extent_buffer() 5477 size_t start_offset = eb->start & ((u64)PAGE_CACHE_SIZE - 1); memcmp_extent_buffer() 5478 unsigned long i = (start_offset + start) >> PAGE_CACHE_SHIFT; memcmp_extent_buffer() 5481 WARN_ON(start > eb->len); memcmp_extent_buffer() 5482 WARN_ON(start + len > eb->start + eb->len); memcmp_extent_buffer() 5484 offset = (start_offset + start) & (PAGE_CACHE_SIZE - 1); memcmp_extent_buffer() 5505 unsigned long start, unsigned long len) write_extent_buffer() 5512 size_t start_offset = eb->start & ((u64)PAGE_CACHE_SIZE - 1); write_extent_buffer() 5513 unsigned long i = (start_offset + start) >> PAGE_CACHE_SHIFT; write_extent_buffer() 5515 WARN_ON(start > eb->len); write_extent_buffer() 5516 WARN_ON(start + len > eb->start + eb->len); write_extent_buffer() 5518 offset = (start_offset + start) & (PAGE_CACHE_SIZE - 1); write_extent_buffer() 5536 unsigned long start, unsigned long len) memset_extent_buffer() 5542 size_t start_offset = eb->start & ((u64)PAGE_CACHE_SIZE - 1); memset_extent_buffer() 5543 unsigned long i = (start_offset + start) >> PAGE_CACHE_SHIFT; memset_extent_buffer() 5545 WARN_ON(start > eb->len); memset_extent_buffer() 5546 WARN_ON(start + len > eb->start + eb->len); memset_extent_buffer() 5548 offset = (start_offset + start) & (PAGE_CACHE_SIZE - 1); memset_extent_buffer() 5573 size_t start_offset = dst->start & ((u64)PAGE_CACHE_SIZE - 1); copy_extent_buffer() 5631 size_t start_offset = dst->start & ((u64)PAGE_CACHE_SIZE - 1); memcpy_extent_buffer() 5679 size_t start_offset = dst->start & ((u64)PAGE_CACHE_SIZE - 1); memmove_extent_buffer() 87 __btrfs_debug_check_extent_io_range(const char *caller, struct extent_io_tree *tree, u64 start, u64 end) __btrfs_debug_check_extent_io_range() argument 445 insert_state(struct extent_io_tree *tree, struct extent_state *state, u64 start, u64 end, struct rb_node ***p, struct rb_node **parent, unsigned *bits, struct extent_changeset *changeset) insert_state() argument 1511 find_first_extent_bit_state(struct extent_io_tree *tree, u64 start, unsigned bits) find_first_extent_bit_state() argument 1592 find_delalloc_range(struct extent_io_tree *tree, u64 *start, u64 *end, u64 max_bytes, struct extent_state **cached_state) find_delalloc_range() argument 1646 __unlock_for_delalloc(struct inode *inode, struct page *locked_page, u64 start, u64 end) __unlock_for_delalloc() argument 1743 find_lock_delalloc_range(struct inode *inode, struct extent_io_tree *tree, struct page *locked_page, u64 *start, u64 *end, u64 max_bytes) find_lock_delalloc_range() argument 1880 count_range_bits(struct extent_io_tree *tree, u64 *start, u64 search_end, u64 max_bytes, unsigned bits, int contig) count_range_bits() argument 2465 bio_readpage_error(struct bio *failed_bio, u64 phy_offset, struct page *page, u64 start, u64 end, int failed_mirror) bio_readpage_error() argument 2941 __get_extent_map(struct inode *inode, struct page *page, size_t pg_offset, u64 start, u64 len, get_extent_t *get_extent, struct extent_map **em_cached) __get_extent_map() argument 3196 __do_contiguous_readpages(struct extent_io_tree *tree, struct page *pages[], int nr_pages, u64 start, u64 end, get_extent_t *get_extent, struct extent_map **em_cached, struct bio **bio, int mirror_num, unsigned long *bio_flags, int rw, u64 *prev_em_start) __do_contiguous_readpages() argument 4201 extent_write_locked_range(struct extent_io_tree *tree, struct inode *inode, u64 start, u64 end, get_extent_t *get_extent, int mode) extent_write_locked_range() argument 4464 extent_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, __u64 start, __u64 len, get_extent_t *get_extent) extent_fiemap() argument 4800 alloc_dummy_extent_buffer(struct btrfs_fs_info *fs_info, u64 start) alloc_dummy_extent_buffer() argument 4889 find_extent_buffer(struct btrfs_fs_info *fs_info, u64 start) find_extent_buffer() argument 4927 alloc_test_extent_buffer(struct btrfs_fs_info *fs_info, u64 start) alloc_test_extent_buffer() argument 4973 alloc_extent_buffer(struct btrfs_fs_info *fs_info, u64 start) alloc_extent_buffer() argument 5266 read_extent_buffer_pages(struct extent_io_tree *tree, struct extent_buffer *eb, u64 start, int wait, get_extent_t *get_extent, int mirror_num) read_extent_buffer_pages() argument 5362 read_extent_buffer(struct extent_buffer *eb, void *dstv, unsigned long start, unsigned long len) read_extent_buffer() argument 5393 read_extent_buffer_to_user(struct extent_buffer *eb, void __user *dstv, unsigned long start, unsigned long len) read_extent_buffer_to_user() argument 5468 memcmp_extent_buffer(struct extent_buffer *eb, const void *ptrv, unsigned long start, unsigned long len) memcmp_extent_buffer() argument 5504 write_extent_buffer(struct extent_buffer *eb, const void *srcv, unsigned long start, unsigned long len) write_extent_buffer() argument 5535 memset_extent_buffer(struct extent_buffer *eb, char c, unsigned long start, unsigned long len) memset_extent_buffer() argument
|
H A D | extent_map.c | 85 static u64 range_end(u64 start, u64 len) range_end() argument 87 if (start + len < start) range_end() 89 return start + len; range_end() 98 u64 end = range_end(em->start, em->len); tree_insert() 104 if (em->start < entry->start) tree_insert() 106 else if (em->start >= extent_map_end(entry)) tree_insert() 113 while (parent && em->start >= extent_map_end(entry)) { tree_insert() 118 if (end > entry->start && em->start < extent_map_end(entry)) tree_insert() 123 while (parent && em->start < entry->start) { tree_insert() 128 if (end > entry->start && em->start < extent_map_end(entry)) tree_insert() 155 if (offset < entry->start) __tree_search() 175 while (prev && offset < prev_entry->start) { __tree_search() 209 if (extent_map_end(prev) == next->start && mergable_maps() 230 if (em->start != 0) { try_merge_map() 235 em->start = merge->start; try_merge_map() 267 * @start: logical offset in the file 275 int unpin_extent_cache(struct extent_map_tree *tree, u64 start, u64 len, unpin_extent_cache() argument 283 em = lookup_extent_mapping(tree, start, len); unpin_extent_cache() 285 WARN_ON(!em || em->start != start); unpin_extent_cache() 292 em->mod_start = em->start; unpin_extent_cache() 303 em->mod_start = em->start; unpin_extent_cache() 326 em->mod_start = em->start; setup_extent_mapping() 361 u64 start, u64 len, int strict) __lookup_extent_mapping() 367 u64 end = range_end(start, len); __lookup_extent_mapping() 369 rb_node = __tree_search(&tree->map, start, &prev, &next); __lookup_extent_mapping() 381 if (strict && !(end > em->start && start < extent_map_end(em))) __lookup_extent_mapping() 391 * @start: byte offset to start the search 395 * [start, len] range. There may be additional objects in the tree that 400 u64 start, u64 len) lookup_extent_mapping() 402 return __lookup_extent_mapping(tree, start, len, 1); lookup_extent_mapping() 408 * @start: byte offset to start the search 412 * [start, len] range. 417 u64 start, u64 len) search_extent_mapping() 419 return __lookup_extent_mapping(tree, start, len, 0); search_extent_mapping() 360 __lookup_extent_mapping(struct extent_map_tree *tree, u64 start, u64 len, int strict) __lookup_extent_mapping() argument 399 lookup_extent_mapping(struct extent_map_tree *tree, u64 start, u64 len) lookup_extent_mapping() argument 416 search_extent_mapping(struct extent_map_tree *tree, u64 start, u64 len) search_extent_mapping() argument
|
H A D | extent_map.h | 24 u64 start; member in struct:extent_map 54 if (em->start + em->len < em->start) extent_map_end() 56 return em->start + em->len; extent_map_end() 68 u64 start, u64 len); 81 int unpin_extent_cache(struct extent_map_tree *tree, u64 start, u64 len, u64 gen); 84 u64 start, u64 len);
|
/linux-4.4.14/arch/unicore32/include/asm/ |
H A D | memblock.h | 22 unsigned long start; member in struct:membank 37 #define bank_pfn_start(bank) __phys_to_pfn((bank)->start) 38 #define bank_pfn_end(bank) __phys_to_pfn((bank)->start + (bank)->size) 40 #define bank_phys_start(bank) ((bank)->start) 41 #define bank_phys_end(bank) ((bank)->start + (bank)->size)
|
H A D | cacheflush.h | 34 * start addresses should be rounded down, end addresses up. 55 * flush_user_range(start, end, flags) 59 * - start - user start address (inclusive, page aligned) 63 * coherent_kern_range(start, end) 66 * region described by start, end. If you have non-snooping 68 * - start - virtual start address 71 * coherent_user_range(start, end) 74 * region described by start, end. If you have non-snooping 76 * - start - virtual start address 88 * dma_flush_range(start, end) 91 * - start - virtual start address 141 unsigned long start, unsigned long end); 152 #define flush_cache_user_range(vma, start, end) \ 153 __cpuc_coherent_user_range((start) & PAGE_MASK, PAGE_ALIGN(end)) 165 #define clean_dcache_area(start, size) cpu_dcache_clean_area(start, size) 203 static inline void flush_cache_vmap(unsigned long start, unsigned long end) flush_cache_vmap() argument 207 static inline void flush_cache_vunmap(unsigned long start, unsigned long end) flush_cache_vunmap() argument
|
/linux-4.4.14/lib/ |
H A D | find_bit.c | 32 unsigned long nbits, unsigned long start, unsigned long invert) _find_next_bit() 36 if (!nbits || start >= nbits) _find_next_bit() 39 tmp = addr[start / BITS_PER_LONG] ^ invert; _find_next_bit() 42 tmp &= BITMAP_FIRST_WORD_MASK(start); _find_next_bit() 43 start = round_down(start, BITS_PER_LONG); _find_next_bit() 46 start += BITS_PER_LONG; _find_next_bit() 47 if (start >= nbits) _find_next_bit() 50 tmp = addr[start / BITS_PER_LONG] ^ invert; _find_next_bit() 53 return min(start + __ffs(tmp), nbits); _find_next_bit() 150 unsigned long nbits, unsigned long start, unsigned long invert) _find_next_bit_le() 154 if (!nbits || start >= nbits) _find_next_bit_le() 157 tmp = addr[start / BITS_PER_LONG] ^ invert; _find_next_bit_le() 160 tmp &= ext2_swab(BITMAP_FIRST_WORD_MASK(start)); _find_next_bit_le() 161 start = round_down(start, BITS_PER_LONG); _find_next_bit_le() 164 start += BITS_PER_LONG; _find_next_bit_le() 165 if (start >= nbits) _find_next_bit_le() 168 tmp = addr[start / BITS_PER_LONG] ^ invert; _find_next_bit_le() 171 return min(start + __ffs(ext2_swab(tmp)), nbits); _find_next_bit_le() 31 _find_next_bit(const unsigned long *addr, unsigned long nbits, unsigned long start, unsigned long invert) _find_next_bit() argument 149 _find_next_bit_le(const unsigned long *addr, unsigned long nbits, unsigned long start, unsigned long invert) _find_next_bit_le() argument
|
H A D | iommu-helper.c | 20 unsigned long start, unsigned int nr, iommu_area_alloc() 29 index = bitmap_find_next_zero_area(map, size, start, nr, align_mask); iommu_area_alloc() 33 start = index + 1; iommu_area_alloc() 19 iommu_area_alloc(unsigned long *map, unsigned long size, unsigned long start, unsigned int nr, unsigned long shift, unsigned long boundary_size, unsigned long align_mask) iommu_area_alloc() argument
|
H A D | iommu-common.c | 58 unsigned int start, i; iommu_tbl_pool_init() local 70 start = 0; iommu_tbl_pool_init() 82 iommu->pools[i].start = start; iommu_tbl_pool_init() 83 iommu->pools[i].hint = start; iommu_tbl_pool_init() 84 start += iommu->poolsize; /* start for next pool */ iommu_tbl_pool_init() 85 iommu->pools[i].end = start - 1; iommu_tbl_pool_init() 91 p->start = start; iommu_tbl_pool_init() 92 p->hint = p->start; iommu_tbl_pool_init() 105 unsigned long n, end, start, limit, boundary_size; iommu_tbl_range_alloc() local 137 (*handle >= pool->start) && (*handle < pool->end)) iommu_tbl_range_alloc() 138 start = *handle; iommu_tbl_range_alloc() 140 start = pool->hint; iommu_tbl_range_alloc() 150 if (start >= limit) iommu_tbl_range_alloc() 151 start = pool->start; iommu_tbl_range_alloc() 157 * but on second pass, start at 0 in pool 0. iommu_tbl_range_alloc() 159 if ((start & mask) >= limit || pass > 0) { iommu_tbl_range_alloc() 163 start = pool->start; iommu_tbl_range_alloc() 165 start &= mask; iommu_tbl_range_alloc() 185 n = iommu_area_alloc(iommu->map, limit, start, npages, shift, iommu_tbl_range_alloc() 190 pool->hint = pool->start; iommu_tbl_range_alloc() 199 pool->hint = pool->start; iommu_tbl_range_alloc() 232 unsigned long largepool_start = tbl->large_pool.start; get_pool()
|
H A D | interval_tree.c | 6 #define START(node) ((node)->start)
|
/linux-4.4.14/arch/arm/mach-imx/devices/ |
H A D | platform-imx-dma.c | 16 .start = iobase, imx_add_imx_dma() 20 .start = irq, imx_add_imx_dma() 24 .start = irq_err, imx_add_imx_dma() 39 .start = iobase, imx_add_imx_sdma() 43 .start = irq, imx_add_imx_sdma()
|
H A D | platform-gpio-mxc.c | 16 .start = iobase, mxc_register_gpio() 20 .start = irq, mxc_register_gpio() 24 .start = irq_high, mxc_register_gpio()
|
H A D | platform-mx2-camera.c | 41 .start = data->iobasecsi, imx_add_mx2_camera() 45 .start = data->irqcsi, imx_add_mx2_camera() 49 .start = data->iobaseemmaprp, imx_add_mx2_camera() 53 .start = data->irqemmaprp, imx_add_mx2_camera()
|
/linux-4.4.14/arch/arm/mach-iop13xx/ |
H A D | tpmi.c | 45 .start = IOP13XX_TPMI_MMR(4), /* tpmi0 starts at dev == 4 */ 50 .start = IOP13XX_TPMI_MEM(0), 55 .start = IOP13XX_TPMI_CTRL(0), 60 .start = IOP13XX_TPMI_IOP_CTRL(0), 65 .start = IRQ_IOP13XX_TPMI0_OUT, 73 .start = IOP13XX_TPMI_MMR(1), 78 .start = IOP13XX_TPMI_MEM(1), 83 .start = IOP13XX_TPMI_CTRL(1), 88 .start = IOP13XX_TPMI_IOP_CTRL(1), 93 .start = IRQ_IOP13XX_TPMI1_OUT, 101 .start = IOP13XX_TPMI_MMR(2), 106 .start = IOP13XX_TPMI_MEM(2), 111 .start = IOP13XX_TPMI_CTRL(2), 116 .start = IOP13XX_TPMI_IOP_CTRL(2), 121 .start = IRQ_IOP13XX_TPMI2_OUT, 129 .start = IOP13XX_TPMI_MMR(3), 134 .start = IOP13XX_TPMI_MEM(3), 139 .start = IOP13XX_TPMI_CTRL(3), 144 .start = IOP13XX_TPMI_IOP_CTRL(3), 149 .start = IRQ_IOP13XX_TPMI3_OUT,
|
/linux-4.4.14/drivers/staging/rtl8188eu/include/ |
H A D | rtw_debug.h | 120 int proc_get_drv_version(char *page, char **start, 124 int proc_get_write_reg(char *page, char **start, 130 int proc_get_read_reg(char *page, char **start, 137 int proc_get_fwstate(char *page, char **start, 140 int proc_get_sec_info(char *page, char **start, 143 int proc_get_mlmext_state(char *page, char **start, 147 int proc_get_qos_option(char *page, char **start, 150 int proc_get_ht_option(char *page, char **start, 153 int proc_get_rf_info(char *page, char **start, 156 int proc_get_ap_info(char *page, char **start, 160 int proc_get_adapter_state(char *page, char **start, 164 int proc_get_trx_info(char *page, char **start, 168 int proc_get_mac_reg_dump1(char *page, char **start, 172 int proc_get_mac_reg_dump2(char *page, char **start, 176 int proc_get_mac_reg_dump3(char *page, char **start, 180 int proc_get_bb_reg_dump1(char *page, char **start, 184 int proc_get_bb_reg_dump2(char *page, char **start, 188 int proc_get_bb_reg_dump3(char *page, char **start, 192 int proc_get_rf_reg_dump1(char *page, char **start, 196 int proc_get_rf_reg_dump2(char *page, char **start, 200 int proc_get_rf_reg_dump3(char *page, char **start, 204 int proc_get_rf_reg_dump4(char *page, char **start, 210 int proc_get_all_sta_info(char *page, char **start, 216 int proc_get_best_channel(char *page, char **start, 220 int proc_get_rx_signal(char *page, char **start, 227 int proc_get_ht_enable(char *page, char **start, 234 int proc_get_cbw40_enable(char *page, char **start, 241 int proc_get_ampdu_enable(char *page, char **start, 248 int proc_get_rx_stbc(char *page, char **start, 255 int proc_get_two_path_rssi(char *page, char **start, 259 int proc_get_rssi_disp(char *page, char **start,
|
/linux-4.4.14/arch/x86/include/asm/trace/ |
H A D | mpx.h | 68 TP_PROTO(unsigned long start, 70 TP_ARGS(start, end), 73 __field(unsigned long, start) 78 __entry->start = start; 83 (void *)__entry->start, 89 TP_PROTO(unsigned long start, unsigned long end), 90 TP_ARGS(start, end) 94 TP_PROTO(unsigned long start, unsigned long end), 95 TP_ARGS(start, end)
|
/linux-4.4.14/arch/mips/ath79/ |
H A D | dev-wmac.c | 29 /* .start and .end fields are filled dynamically */ 32 /* .start and .end fields are filled dynamically */ 56 ath79_wmac_resources[0].start = AR913X_WMAC_BASE; ar913x_wmac_setup() 58 ath79_wmac_resources[1].start = ATH79_CPU_IRQ(2); ar913x_wmac_setup() 84 ath79_wmac_resources[0].start = AR933X_WMAC_BASE; ar933x_wmac_setup() 86 ath79_wmac_resources[1].start = ATH79_CPU_IRQ(2); ar933x_wmac_setup() 107 ath79_wmac_resources[0].start = AR934X_WMAC_BASE; ar934x_wmac_setup() 109 ath79_wmac_resources[1].start = ATH79_IP2_IRQ(1); ar934x_wmac_setup() 125 ath79_wmac_resources[0].start = QCA955X_WMAC_BASE; qca955x_wmac_setup() 127 ath79_wmac_resources[1].start = ATH79_IP2_IRQ(1); qca955x_wmac_setup()
|
H A D | dev-common.c | 29 .start = AR71XX_UART_BASE, 60 .start = AR933X_UART_BASE, 65 .start = ATH79_MISC_IRQ(3), 105 res.start = AR71XX_RESET_BASE + AR71XX_RESET_REG_WDOG_CTRL; ath79_register_wdt() 106 res.end = res.start + 0x8 - 1; ath79_register_wdt() 116 .start = AR71XX_GPIO_BASE, 120 .start = ATH79_MISC_IRQ(2),
|
/linux-4.4.14/arch/ia64/include/asm/ |
H A D | meminit.h | 27 u64 start; /* virtual address of beginning of element */ member in struct:rsvd_region 37 extern int filter_rsvd_memory (u64 start, u64 end, void *arg); 38 extern int filter_memory (u64 start, u64 end, void *arg); 43 extern int reserve_elfcorehdr(u64 *start, u64 *end); 52 extern void call_pernode_memory (unsigned long start, unsigned long len, void *func); 54 # define call_pernode_memory(start, len, func) (*func)(start, len, 0) 59 extern int register_active_ranges(u64 start, u64 len, int nid); 65 extern int find_largest_hole(u64 start, u64 end, void *arg); 66 extern int create_mem_map_page_table(u64 start, u64 end, void *arg);
|
/linux-4.4.14/arch/blackfin/mach-bf548/boards/ |
H A D | ezkit.c | 48 .start = 0x2C0C0000, 53 .start = IRQ_PG7, 94 .start = IRQ_EPPI0_ERR, 144 .start = IRQ_KEY, 176 .start = CNT_CONFIG, 181 .start = IRQ_CNT, 248 .start = UART0_DLL, 254 .start = PORTE_FER, 260 .start = IRQ_UART0_TX, 265 .start = IRQ_UART0_RX, 270 .start = IRQ_UART0_ERROR, 275 .start = CH_UART0_TX, 280 .start = CH_UART0_RX, 303 .start = UART1_DLL, 309 .start = PORTH_FER, 315 .start = IRQ_UART1_TX, 320 .start = IRQ_UART1_RX, 325 .start = IRQ_UART1_ERROR, 330 .start = CH_UART1_TX, 335 .start = CH_UART1_RX, 341 .start = GPIO_PE10, 346 .start = GPIO_PE9, 374 .start = UART2_DLL, 380 .start = PORTB_FER, 386 .start = IRQ_UART2_TX, 391 .start = IRQ_UART2_RX, 396 .start = IRQ_UART2_ERROR, 401 .start = CH_UART2_TX, 406 .start = CH_UART2_RX, 429 .start = UART3_DLL, 435 .start = PORTB_FER, 441 .start = IRQ_UART3_TX, 446 .start = IRQ_UART3_RX, 451 .start = IRQ_UART3_ERROR, 456 .start = CH_UART3_TX, 461 .start = CH_UART3_RX, 467 .start = GPIO_PB3, 472 .start = GPIO_PB2, 503 .start = 0xFFC00400, 508 .start = IRQ_UART0_RX, 513 .start = CH_UART0_RX, 528 .start = 0xFFC02000, 533 .start = IRQ_UART1_RX, 538 .start = CH_UART1_RX, 553 .start = 0xFFC02100, 558 .start = IRQ_UART2_RX, 563 .start = CH_UART2_RX, 578 .start = 0xFFC03100, 583 .start = IRQ_UART3_RX, 588 .start = CH_UART3_RX, 608 .start = 0x24000000, 613 .start = IRQ_PE8, 640 .start = 0xFFC03C00, 645 .start = IRQ_USB_INT0, 651 .start = IRQ_USB_DMA, 703 .start = SPORT0_TCR1, 708 .start = IRQ_SPORT0_RX, 713 .start = IRQ_SPORT0_ERROR, 737 .start = SPORT1_TCR1, 742 .start = IRQ_SPORT1_RX, 747 .start = IRQ_SPORT1_ERROR, 771 .start = SPORT2_TCR1, 776 .start = IRQ_SPORT2_RX, 781 .start = IRQ_SPORT2_ERROR, 805 .start = SPORT3_TCR1, 810 .start = IRQ_SPORT3_RX, 815 .start = IRQ_SPORT3_ERROR, 846 .start = 0xFFC02A00, 851 .start = IRQ_CAN0_RX, 856 .start = IRQ_CAN0_TX, 861 .start = IRQ_CAN0_ERROR, 883 .start = 0xFFC03200, 888 .start = IRQ_CAN1_RX, 893 .start = IRQ_CAN1_TX, 898 .start = IRQ_CAN1_ERROR, 919 .start = 0xFFC03800, 924 .start = IRQ_ATAPI_ERR, 966 .start = 0xFFC03B00, 971 .start = CH_NFC, 1037 .start = 0x20000000, 1109 .start = PINT0_MASK_SET, 1114 .start = IRQ_PINT0, 1129 .start = PINT1_MASK_SET, 1134 .start = IRQ_PINT1, 1149 .start = PINT2_MASK_SET, 1154 .start = IRQ_PINT2, 1169 .start = PINT3_MASK_SET, 1174 .start = IRQ_PINT3, 1189 .start = PORTA_FER, 1194 .start = IRQ_PA0, 1221 .start = PORTB_FER, 1226 .start = IRQ_PB0, 1253 .start = PORTC_FER, 1258 .start = IRQ_PC0, 1285 .start = PORTD_FER, 1290 .start = IRQ_PD0, 1317 .start = PORTE_FER, 1322 .start = IRQ_PE0, 1349 .start = PORTF_FER, 1354 .start = IRQ_PF0, 1381 .start = PORTG_FER, 1386 .start = IRQ_PG0, 1411 .start = PORTH_FER, 1416 .start = IRQ_PH0, 1441 .start = PORTI_FER, 1446 .start = IRQ_PI0, 1471 .start = PORTJ_FER, 1476 .start = IRQ_PJ0, 1556 .start = SPI0_REGBASE, 1561 .start = CH_SPI0, 1566 .start = IRQ_SPI0, 1575 .start = SPI1_REGBASE, 1580 .start = CH_SPI1, 1585 .start = IRQ_SPI1, 1695 .start = TWI0_REGBASE, 1700 .start = IRQ_TWI0, 1721 .start = TWI1_REGBASE, 1726 .start = IRQ_TWI1, 1863 .start = SPORT##x##_TCR1, \ 1868 .start = CH_SPORT##x##_RX, \ 1873 .start = CH_SPORT##x##_TX, \ 1878 .start = IRQ_SPORT##x##_ERROR, \
|
H A D | cm_bf548.c | 55 .start = IRQ_EPPI0_ERR, 105 .start = IRQ_KEY, 133 .start = UART0_DLL, 138 .start = IRQ_UART0_TX, 143 .start = IRQ_UART0_RX, 148 .start = IRQ_UART0_ERROR, 153 .start = CH_UART0_TX, 158 .start = CH_UART0_RX, 181 .start = UART1_DLL, 186 .start = IRQ_UART1_TX, 191 .start = IRQ_UART1_RX, 196 .start = IRQ_UART1_ERROR, 201 .start = CH_UART1_TX, 206 .start = CH_UART1_RX, 212 .start = GPIO_PE10, 217 .start = GPIO_PE9, 245 .start = UART2_DLL, 250 .start = IRQ_UART2_TX, 255 .start = IRQ_UART2_RX, 260 .start = IRQ_UART2_ERROR, 265 .start = CH_UART2_TX, 270 .start = CH_UART2_RX, 293 .start = UART3_DLL, 298 .start = IRQ_UART3_TX, 303 .start = IRQ_UART3_RX, 308 .start = IRQ_UART3_ERROR, 313 .start = CH_UART3_TX, 318 .start = CH_UART3_RX, 324 .start = GPIO_PB3, 329 .start = GPIO_PB2, 360 .start = 0xFFC00400, 365 .start = IRQ_UART0_RX, 370 .start = CH_UART0_RX, 385 .start = 0xFFC02000, 390 .start = IRQ_UART1_RX, 395 .start = CH_UART1_RX, 410 .start = 0xFFC02100, 415 .start = IRQ_UART2_RX, 420 .start = CH_UART2_RX, 435 .start = 0xFFC03100, 440 .start = IRQ_UART3_RX, 445 .start = CH_UART3_RX, 465 .start = 0x24000000, 470 .start = IRQ_PE6, 497 .start = 0xFFC03C00, 502 .start = IRQ_USB_INT0, 508 .start = IRQ_USB_DMA, 560 .start = SPORT0_TCR1, 565 .start = IRQ_SPORT0_RX, 570 .start = IRQ_SPORT0_ERROR, 594 .start = SPORT1_TCR1, 599 .start = IRQ_SPORT1_RX, 604 .start = IRQ_SPORT1_ERROR, 628 .start = SPORT2_TCR1, 633 .start = IRQ_SPORT2_RX, 638 .start = IRQ_SPORT2_ERROR, 662 .start = SPORT3_TCR1, 667 .start = IRQ_SPORT3_RX, 672 .start = IRQ_SPORT3_ERROR, 698 .start = 0xFFC03800, 703 .start = IRQ_ATAPI_ERR, 741 .start = 0xFFC03B00, 746 .start = CH_NFC, 786 .start = 0xFFC02A00, 791 .start = IRQ_CAN0_RX, 796 .start = IRQ_CAN0_TX, 801 .start = IRQ_CAN0_ERROR, 841 .start = 0x20000000, 938 .start = SPI0_REGBASE, 943 .start = CH_SPI0, 948 .start = IRQ_SPI0, 957 .start = SPI1_REGBASE, 962 .start = CH_SPI1, 967 .start = IRQ_SPI1, 1012 .start = TWI0_REGBASE, 1017 .start = IRQ_TWI0, 1038 .start = TWI1_REGBASE, 1043 .start = IRQ_TWI1,
|
/linux-4.4.14/arch/score/mm/ |
H A D | cache.c | 173 unsigned long start, unsigned long end) flush_cache_range() 185 pgdp = pgd_offset(mm, start); flush_cache_range() 186 pudp = pud_offset(pgdp, start); flush_cache_range() 187 pmdp = pmd_offset(pudp, start); flush_cache_range() 188 ptep = pte_offset(pmdp, start); flush_cache_range() 190 while (start <= end) { flush_cache_range() 192 pgdp = pgd_offset(mm, start); flush_cache_range() 193 pudp = pud_offset(pgdp, start); flush_cache_range() 194 pmdp = pmd_offset(pudp, start); flush_cache_range() 195 ptep = pte_offset(pmdp, start); flush_cache_range() 198 start = (start + PAGE_SIZE) & ~(PAGE_SIZE - 1); flush_cache_range() 201 tmpend = (start | (PAGE_SIZE-1)) > end ? flush_cache_range() 202 end : (start | (PAGE_SIZE-1)); flush_cache_range() 204 flush_dcache_range(start, tmpend); flush_cache_range() 206 flush_icache_range(start, tmpend); flush_cache_range() 207 start = (start + PAGE_SIZE) & ~(PAGE_SIZE - 1); flush_cache_range() 246 void flush_dcache_range(unsigned long start, unsigned long end) flush_dcache_range() argument 250 start = start & ~(L1_CACHE_BYTES - 1); flush_dcache_range() 252 size = end - start; flush_dcache_range() 260 : : "r" (start)); flush_dcache_range() 261 start += L1_CACHE_BYTES; flush_dcache_range() 265 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument 268 start = start & ~(L1_CACHE_BYTES - 1); flush_icache_range() 271 size = end - start; flush_icache_range() 277 : : "r" (start)); flush_icache_range() 278 start += L1_CACHE_BYTES; flush_icache_range() 172 flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_cache_range() argument
|
/linux-4.4.14/arch/hexagon/lib/ |
H A D | checksum.c | 103 int i, start, mid, end, mask; do_csum() local 111 start = 0xF & (16-(((int) ptr) & 0xF)) ; do_csum() 113 start = start & mask ; do_csum() 115 mid = len - start; do_csum() 121 if (start & 1) do_csum() 123 ptr2 = (unsigned short *) &ptr[start & 1]; do_csum() 124 if (start & 2) do_csum() 126 ptr4 = (unsigned int *) &ptr[start & 3]; do_csum() 127 if (start & 4) { do_csum() 133 ptr8 = (u64 *) &ptr[start & 7]; do_csum() 134 if (start & 8) { do_csum() 140 ptr8_o = (u64 *) (ptr + start); do_csum() 141 ptr8_e = (u64 *) (ptr + start + 8); do_csum() 163 ptr4 = (unsigned int *) &ptr[start + (mid * 16) + (end & 8)]; do_csum() 170 ptr2 = (unsigned short *) &ptr[start + (mid * 16) + (end & 12)]; do_csum() 175 sum1 += (u64) ptr[start + (mid * 16) + (end & 14)]; do_csum() 177 ptr8 = (u64 *) &ptr[start + (mid * 16)]; do_csum() 189 if (start & 1) do_csum()
|
/linux-4.4.14/arch/frv/include/asm/ |
H A D | cacheflush.h | 24 #define flush_cache_range(mm, start, end) do {} while(0) 26 #define flush_cache_vmap(start, end) do {} while(0) 27 #define flush_cache_vunmap(start, end) do {} while(0) 35 extern void frv_dcache_writeback(unsigned long start, unsigned long size); 36 extern void frv_cache_invalidate(unsigned long start, unsigned long size); 37 extern void frv_icache_invalidate(unsigned long start, unsigned long size); 38 extern void frv_cache_wback_inv(unsigned long start, unsigned long size); 71 static inline void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument 73 frv_cache_wback_inv(start, end); flush_icache_range() 78 unsigned long start, unsigned long len); 81 unsigned long start, unsigned long len) flush_icache_user_range() 83 frv_cache_wback_inv(start, start + len); flush_icache_user_range() 80 flush_icache_user_range(struct vm_area_struct *vma, struct page *page, unsigned long start, unsigned long len) flush_icache_user_range() argument
|
H A D | tlbflush.h | 23 extern void asmlinkage __flush_tlb_page(unsigned long contextid, unsigned long start); 25 unsigned long start, unsigned long end); 42 #define flush_tlb_range(vma,start,end) \ 45 __flush_tlb_range((vma)->vm_mm->context.id, start, end); \ 59 #define flush_tlb_kernel_range(start, end) flush_tlb_all() 67 #define flush_tlb_range(mm,start,end) BUG() 68 #define flush_tlb_kernel_range(start, end) BUG()
|
H A D | termios.h | 8 start=^Q stop=^S susp=^Z eol=\0
|
/linux-4.4.14/drivers/infiniband/hw/usnic/ |
H A D | usnic_uiom_interval_tree.c | 42 #define START(node) ((node)->start) 45 #define MAKE_NODE(node, start, end, ref_cnt, flags, err, err_out) \ 47 node = usnic_uiom_interval_node_alloc(start, \ 57 #define MAKE_NODE_AND_APPEND(node, start, end, ref_cnt, flags, err, \ 60 MAKE_NODE(node, start, end, \ 70 usnic_uiom_interval_node_alloc(long int start, long int last, int ref_cnt, usnic_uiom_interval_node_alloc() argument 78 interval->start = start; usnic_uiom_interval_node_alloc() 94 if (node_a->start < node_b->start) interval_cmp() 96 else if (node_a->start > node_b->start) interval_cmp() 103 find_intervals_intersection_sorted(struct rb_root *root, unsigned long start, find_intervals_intersection_sorted() argument 111 for (node = usnic_uiom_interval_tree_iter_first(root, start, last); find_intervals_intersection_sorted() 113 node = usnic_uiom_interval_tree_iter_next(node, start, last)) find_intervals_intersection_sorted() 119 int usnic_uiom_get_intervals_diff(unsigned long start, unsigned long last, usnic_uiom_get_intervals_diff() argument 126 long int pivot = start; usnic_uiom_get_intervals_diff() 131 find_intervals_intersection_sorted(root, start, last, usnic_uiom_get_intervals_diff() 135 if (pivot < interval->start) { usnic_uiom_get_intervals_diff() 136 MAKE_NODE_AND_APPEND(tmp, pivot, interval->start - 1, usnic_uiom_get_intervals_diff() 139 pivot = interval->start; usnic_uiom_get_intervals_diff() 143 * Invariant: Set [start, pivot] is either in diff_set or root, usnic_uiom_get_intervals_diff() 178 int usnic_uiom_insert_interval(struct rb_root *root, unsigned long start, usnic_uiom_insert_interval() argument 184 unsigned long lpivot = start; usnic_uiom_insert_interval() 189 find_intervals_intersection_sorted(root, start, last, usnic_uiom_insert_interval() 197 istart = interval->start; usnic_uiom_insert_interval() 249 void usnic_uiom_remove_interval(struct rb_root *root, unsigned long start, usnic_uiom_remove_interval() argument 254 for (interval = usnic_uiom_interval_tree_iter_first(root, start, last); usnic_uiom_remove_interval() 257 start, usnic_uiom_remove_interval()
|
H A D | usnic_uiom_interval_tree.h | 42 unsigned long start; member in struct:usnic_uiom_interval_node 57 unsigned long start, 61 unsigned long start, unsigned long last); 63 * Inserts {start...last} into {root}. If there are overlaps, 67 unsigned long start, unsigned long last, 70 * Removed {start...last} from {root}. The nodes removed are returned in 75 unsigned long start, unsigned long last, 78 * Returns {start...last} - {root} (relative complement of {start...last} in 81 int usnic_uiom_get_intervals_diff(unsigned long start,
|
/linux-4.4.14/arch/blackfin/mach-bf527/boards/ |
H A D | cm_bf527.c | 44 .start = 0x203C0000, 49 .start = IRQ_PF7, 78 .start = 0xffc03800, 83 .start = IRQ_USB_INT0, 89 .start = IRQ_USB_DMA, 161 .start = NFC_CTL, 166 .start = CH_NFC, 186 .start = 0x20310000, /* IO PORT */ 190 .start = 0x20311000, /* Attribute Memory */ 194 .start = IRQ_PF4, 198 .start = 6, /* Card Detect PF6 */ 231 .start = 0x20300300, 236 .start = IRQ_PF7, 255 .start = 0x203FB800, 260 .start = 0x203FB804, 265 .start = IRQ_PF9, 315 .start = 0x20300000, 319 .start = IRQ_PF7, 455 .start = SPI0_REGBASE, 460 .start = CH_SPI, 465 .start = IRQ_SPI, 510 .start = 0x20000000, 514 .start = (unsigned long)cm_flash_gpios, 535 .start = UART0_THR, 540 .start = IRQ_UART0_TX, 545 .start = IRQ_UART0_RX, 550 .start = IRQ_UART0_ERROR, 555 .start = CH_UART0_TX, 560 .start = CH_UART0_RX, 583 .start = UART1_THR, 588 .start = IRQ_UART1_TX, 593 .start = IRQ_UART1_RX, 598 .start = IRQ_UART1_ERROR, 603 .start = CH_UART1_TX, 608 .start = CH_UART1_RX, 614 .start = GPIO_PF9, 619 .start = GPIO_PF10, 646 .start = 0xFFC00400, 651 .start = IRQ_UART0_RX, 656 .start = CH_UART0_RX, 672 .start = 0xFFC02000, 677 .start = IRQ_UART1_RX, 682 .start = CH_UART1_RX, 702 .start = TWI0_REGBASE, 707 .start = IRQ_TWI, 747 .start = SPORT0_TCR1, 752 .start = IRQ_SPORT0_RX, 757 .start = IRQ_SPORT0_ERROR, 781 .start = SPORT1_TCR1, 786 .start = IRQ_SPORT1_RX, 791 .start = IRQ_SPORT1_ERROR,
|
H A D | ezkit.c | 49 .start = 0x203C0000, 54 .start = IRQ_PF7, 83 .start = 0xffc03800, 88 .start = IRQ_USB_INT0, 94 .start = IRQ_USB_DMA, 146 .start = IRQ_PPI_ERROR, 170 .start = IRQ_PPI_ERROR, 211 .start = 0x20000000, 255 .start = NFC_CTL, 260 .start = CH_NFC, 280 .start = 0x20310000, /* IO PORT */ 284 .start = 0x20311000, /* Attribute Memory */ 288 .start = IRQ_PF4, 292 .start = 6, /* Card Detect PF6 */ 325 .start = 0x20300300, 330 .start = IRQ_PF7, 349 .start = 0x203FB800, 354 .start = 0x203FB800 + 4, 359 .start = IRQ_PF9, 409 .start = 0x20300000, 413 .start = 1, 416 .start = IRQ_PF7, 516 .start = SPORT##x##_TCR1, \ 521 .start = CH_SPORT##x##_RX, \ 526 .start = CH_SPORT##x##_TX, \ 531 .start = IRQ_SPORT##x##_ERROR, \ 669 .start = SPI0_REGBASE, 674 .start = CH_SPI, 679 .start = IRQ_SPI, 700 .start = UART0_THR, 705 .start = IRQ_UART0_TX, 710 .start = IRQ_UART0_RX, 715 .start = IRQ_UART0_ERROR, 720 .start = CH_UART0_TX, 725 .start = CH_UART0_RX, 748 .start = UART1_THR, 753 .start = IRQ_UART1_TX, 758 .start = IRQ_UART1_RX, 763 .start = IRQ_UART1_ERROR, 768 .start = CH_UART1_TX, 773 .start = CH_UART1_RX, 779 .start = GPIO_PF9, 784 .start = GPIO_PF10, 811 .start = 0xFFC00400, 816 .start = IRQ_UART0_RX, 821 .start = CH_UART0_RX, 837 .start = 0xFFC02000, 842 .start = IRQ_UART1_RX, 847 .start = CH_UART1_RX, 867 .start = TWI0_REGBASE, 872 .start = IRQ_TWI, 1006 .start = SPORT0_TCR1, 1011 .start = IRQ_SPORT0_RX, 1016 .start = IRQ_SPORT0_ERROR, 1040 .start = SPORT1_TCR1, 1045 .start = IRQ_SPORT1_RX, 1050 .start = IRQ_SPORT1_ERROR, 1117 .start = CNT_CONFIG, 1122 .start = IRQ_CNT,
|
H A D | ad7160eval.c | 43 .start = 0xffc03800, 48 .start = IRQ_USB_INT0, 53 .start = IRQ_USB_DMA, 103 .start = IRQ_PPI_ERROR, 141 .start = 0x20000000, 181 .start = NFC_CTL, 186 .start = CH_NFC, 337 .start = SPI0_REGBASE, 342 .start = CH_SPI, 347 .start = IRQ_SPI, 368 .start = UART0_THR, 373 .start = IRQ_UART0_TX, 378 .start = IRQ_UART0_RX, 383 .start = IRQ_UART0_ERROR, 388 .start = CH_UART0_TX, 393 .start = CH_UART0_RX, 416 .start = UART1_THR, 421 .start = IRQ_UART1_TX, 426 .start = IRQ_UART1_RX, 431 .start = IRQ_UART1_ERROR, 436 .start = CH_UART1_TX, 441 .start = CH_UART1_RX, 447 .start = GPIO_PF9, 452 .start = GPIO_PF10, 479 .start = 0xFFC00400, 484 .start = IRQ_UART0_RX, 489 .start = CH_UART0_RX, 505 .start = 0xFFC02000, 510 .start = IRQ_UART1_RX, 515 .start = CH_UART1_RX, 565 .start = TWI0_REGBASE, 570 .start = IRQ_TWI, 601 .start = SPORT0_TCR1, 606 .start = IRQ_SPORT0_RX, 611 .start = IRQ_SPORT0_ERROR, 635 .start = SPORT1_TCR1, 640 .start = IRQ_SPORT1_RX, 645 .start = IRQ_SPORT1_ERROR, 691 .start = CNT_CONFIG, 696 .start = IRQ_CNT,
|
/linux-4.4.14/arch/alpha/kernel/ |
H A D | pci.c | 76 dev->resource[2].start = dev->resource[3].start = 0; quirk_cypress() 80 dev->resource[0].start = 0x170; quirk_cypress() 82 dev->resource[1].start = 0x376; quirk_cypress() 131 resource_size_t start = res->start; pcibios_align_resource() local 134 /* Make sure we start at our min on all hoses */ pcibios_align_resource() 135 if (start - hose->io_space->start < PCIBIOS_MIN_IO) pcibios_align_resource() 136 start = PCIBIOS_MIN_IO + hose->io_space->start; pcibios_align_resource() 141 if (start & 0x300) pcibios_align_resource() 142 start = (start + 0x3ff) & ~0x3ff; pcibios_align_resource() 145 /* Make sure we start at our min on all hoses */ pcibios_align_resource() 146 if (start - hose->mem_space->start < PCIBIOS_MIN_MEM) pcibios_align_resource() 147 start = PCIBIOS_MIN_MEM + hose->mem_space->start; pcibios_align_resource() 167 start = ALIGN(start, alignto); pcibios_align_resource() 169 if (((start / (16*MB)) & 0x7) == 0) { pcibios_align_resource() 170 start &= ~(128*MB - 1); pcibios_align_resource() 171 start += 16*MB; pcibios_align_resource() 172 start = ALIGN(start, alignto); pcibios_align_resource() 174 if (start/(128*MB) != (start + size - 1)/(128*MB)) { pcibios_align_resource() 175 start &= ~(128*MB - 1); pcibios_align_resource() 176 start += (128 + 16)*MB; pcibios_align_resource() 177 start = ALIGN(start, alignto); pcibios_align_resource() 182 return start; pcibios_align_resource() 285 if (r->parent || !r->start || !r->flags) pcibios_claim_one_bus() 329 end = hose->mem_space->start + pci_mem_end; common_init_pci() 335 hose->io_space->start); common_init_pci() 337 hose->mem_space->start); common_init_pci()
|
/linux-4.4.14/arch/blackfin/mach-bf533/boards/ |
H A D | ip0x.c | 41 .start = 0x20100000, 45 .start = 0x20100000 + 2, 49 .start = IRQ_PF15, 57 .start = 0x20200000, 61 .start = 0x20200000 + 2, 65 .start = IRQ_PF14, 149 .start = BFIN_UART_THR, 154 .start = IRQ_UART0_TX, 159 .start = IRQ_UART0_RX, 164 .start = IRQ_UART0_ERROR, 169 .start = CH_UART0_TX, 174 .start = CH_UART0_RX, 200 .start = 0xFFC00400, 205 .start = IRQ_UART0_RX, 210 .start = CH_UART0_RX, 228 .start = 0x20300000, 232 .start = 0x20300000 + 2, 236 .start = IRQ_PF11,
|
H A D | cm_bf533.c | 109 .start = SPI0_REGBASE, 114 .start = CH_SPI, 119 .start = IRQ_SPI, 161 .start = 0x20200300, 165 .start = IRQ_PF0, 187 .start = 0x20308000, 191 .start = IRQ_PF8, 219 .start = BFIN_UART_THR, 224 .start = IRQ_UART0_TX, 229 .start = IRQ_UART0_RX, 234 .start = IRQ_UART0_ERROR, 239 .start = CH_UART0_TX, 244 .start = CH_UART0_RX, 270 .start = 0xFFC00400, 275 .start = IRQ_UART0_RX, 280 .start = CH_UART0_RX, 299 .start = SPORT0_TCR1, 304 .start = IRQ_SPORT0_RX, 309 .start = IRQ_SPORT0_ERROR, 333 .start = SPORT1_TCR1, 338 .start = IRQ_SPORT1_RX, 343 .start = IRQ_SPORT1_ERROR, 369 .start = 0x20308000, 373 .start = 0x20308004, 377 .start = IRQ_PF4, 409 .start = 0x20300000, 413 .start = IRQ_PF6, 449 .start = 0x20000000,
|
H A D | H8606.c | 45 .start = 0x20300000, 50 .start = 0x20300004, 55 .start = IRQ_PF10, 82 .start = 0x20300300, 86 .start = IRQ_PROG_INTB, 90 .start = IRQ_PF7, 110 .start = 0x20300000, 114 .start = IRQ_PF10, 197 .start = SPI0_REGBASE, 202 .start = CH_SPI, 207 .start = IRQ_SPI, 236 .start = BFIN_UART_THR, 241 .start = IRQ_UART0_TX, 246 .start = IRQ_UART0_RX, 251 .start = IRQ_UART0_ERROR, 256 .start = CH_UART0_TX, 261 .start = CH_UART0_RX, 287 .start = 0xFFC00400, 292 .start = IRQ_UART0_RX, 297 .start = CH_UART0_RX, 365 .start = 0x20200030, 370 .start = IRQ_PF9,
|
/linux-4.4.14/arch/x86/boot/compressed/ |
H A D | aslr.c | 108 unsigned long start; member in struct:mem_vector 118 if (item->start < region->start) mem_contains() 121 if (item->start + item->size > region->start + region->size) mem_contains() 129 if (one->start + one->size <= two->start) mem_overlaps() 132 if (one->start >= two->start + two->size) mem_overlaps() 151 mem_avoid[0].start = unsafe; mem_avoid_init() 159 mem_avoid[1].start = initrd_start; mem_avoid_init() 169 mem_avoid[2].start = cmd_line; mem_avoid_init() 173 mem_avoid[3].start = (unsigned long)free_mem_ptr; mem_avoid_init() 177 mem_avoid[4].start = (unsigned long)free_mem_end_ptr; mem_avoid_init() 197 avoid.start = (unsigned long)ptr; mem_avoid_overlap() 250 region.start = entry->addr; process_e820_entry() 254 if (region.start < minimum) process_e820_entry() 255 region.start = minimum; process_e820_entry() 258 region.start = ALIGN(region.start, CONFIG_PHYSICAL_ALIGN); process_e820_entry() 261 if (region.start > entry->addr + entry->size) process_e820_entry() 265 region.size -= region.start - entry->addr; process_e820_entry() 268 if (region.start + region.size > CONFIG_RANDOMIZE_BASE_MAX_OFFSET) process_e820_entry() 269 region.size = CONFIG_RANDOMIZE_BASE_MAX_OFFSET - region.start; process_e820_entry() 272 for (img.start = region.start, img.size = image_size ; process_e820_entry() 274 img.start += CONFIG_PHYSICAL_ALIGN) { process_e820_entry() 277 slots_append(img.start); process_e820_entry()
|
/linux-4.4.14/drivers/pnp/ |
H A D | system.c | 29 resource_size_t start = r->start, end = r->end; reserve_range() local 38 res = request_region(start, end - start + 1, regionid); reserve_range() 40 res = request_mem_region(start, end - start + 1, regionid); reserve_range() 63 if (res->start == 0) reserve_resources_of_dev() 65 if (res->start < 0x100) reserve_resources_of_dev() 75 if (res->end < res->start) reserve_resources_of_dev()
|
H A D | manager.c | 44 "flags %#lx\n", idx, (unsigned long long) res->start, pnp_assign_port() 51 res->start = 0; pnp_assign_port() 60 res->start = rule->min; pnp_assign_port() 61 res->end = res->start + rule->size - 1; pnp_assign_port() 64 res->start += rule->align; pnp_assign_port() 65 res->end = res->start + rule->size - 1; pnp_assign_port() 66 if (res->start > rule->max || !rule->align) { pnp_assign_port() 76 pnp_add_io_resource(dev, res->start, res->end, res->flags); pnp_assign_port() 87 "flags %#lx\n", idx, (unsigned long long) res->start, pnp_assign_mem() 94 res->start = 0; pnp_assign_mem() 111 res->start = rule->min; pnp_assign_mem() 112 res->end = res->start + rule->size - 1; pnp_assign_mem() 115 res->start += rule->align; pnp_assign_mem() 116 res->end = res->start + rule->size - 1; pnp_assign_mem() 117 if (res->start > rule->max || !rule->align) { pnp_assign_mem() 127 pnp_add_mem_resource(dev, res->start, res->end, res->flags); pnp_assign_mem() 144 idx, (int) res->start, res->flags); pnp_assign_irq() 150 res->start = -1; pnp_assign_irq() 160 res->start = find_next_bit(rule->map.bits, PNP_IRQ_NR, 16); pnp_assign_irq() 161 if (res->start < PNP_IRQ_NR) { pnp_assign_irq() 162 res->end = res->start; pnp_assign_irq() 167 res->start = res->end = xtab[i]; pnp_assign_irq() 174 res->start = -1; pnp_assign_irq() 185 pnp_add_irq_resource(dev, res->start, res->flags); pnp_assign_irq() 203 idx, (int) res->start, res->flags); pnp_assign_dma() 209 res->start = -1; pnp_assign_dma() 220 res->start = res->end = xtab[i]; pnp_assign_dma() 230 pnp_add_dma_resource(dev, res->start, res->flags); pnp_assign_dma() 331 * pnp_start_dev - low-level start of the PnP device
|
/linux-4.4.14/arch/m68k/kernel/ |
H A D | pcibios.c | 36 resource_size_t start = res->start; pcibios_align_resource() local 38 if ((res->flags & IORESOURCE_IO) && (start & 0x300)) pcibios_align_resource() 39 start = (start + 0x3ff) & ~0x3ff; pcibios_align_resource() 41 start = (start + align - 1) & ~(align - 1); pcibios_align_resource() 43 return start; pcibios_align_resource() 64 if (!r->start && r->end) { pcibios_enable_device()
|
/linux-4.4.14/arch/parisc/math-emu/ |
H A D | fpbits.h | 53 #define Bitfield_extract(start, length, object) \ 54 ((object) >> (HOSTWDSZ - (start) - (length)) & \ 57 #define Bitfield_signed_extract(start, length, object) \ 58 ((int)((object) << start) >> (HOSTWDSZ - (length))) 60 #define Bitfield_mask(start, len, object) \ 61 ((object) & (((unsigned)-1 >> (HOSTWDSZ-len)) << (HOSTWDSZ-start-len))) 63 #define Bitfield_deposit(value,start,len,object) object = \ 64 ((object) & ~(((unsigned)-1 >> (HOSTWDSZ-len)) << (HOSTWDSZ-start-len))) | \ 65 (((value) & ((unsigned)-1 >> (HOSTWDSZ-len))) << (HOSTWDSZ-start-len))
|
/linux-4.4.14/arch/m32r/include/asm/ |
H A D | cacheflush.h | 13 #define flush_cache_range(vma, start, end) do { } while (0) 20 #define flush_icache_range(start, end) _flush_cache_copyback_all() 26 #define flush_icache_range(start, end) smp_flush_cache_all() 35 #define flush_cache_range(vma, start, end) do { } while (0) 41 #define flush_icache_range(start, end) _flush_cache_all() 49 #define flush_cache_range(vma, start, end) do { } while (0) 55 #define flush_icache_range(start, end) do { } while (0) 61 #define flush_cache_vmap(start, end) do { } while (0) 62 #define flush_cache_vunmap(start, end) do { } while (0)
|
H A D | tlbflush.h | 13 * - flush_tlb_range(vma, start, end) flushes a range of pages 14 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages 28 #define flush_tlb_range(vma, start, end) \ 29 local_flush_tlb_range(vma, start, end) 30 #define flush_tlb_kernel_range(start, end) local_flush_tlb_all() 35 #define flush_tlb_range(vma, start, end) do { } while (0) 47 #define flush_tlb_range(vma, start, end) \ 48 smp_flush_tlb_range(vma, start, end) 49 #define flush_tlb_kernel_range(start, end) smp_flush_tlb_all()
|
/linux-4.4.14/arch/powerpc/lib/ |
H A D | rheap.c | 138 blk->start = 0; get_slot() 163 s = blkn->start; attach_free_block() 175 bs = blk->start; attach_free_block() 193 if (before && s != (before->start + before->size)) attach_free_block() 196 if (after && e != after->start) attach_free_block() 221 after->start -= size; attach_free_block() 240 if (blk->start > blkn->start) { attach_taken_block() 329 int rh_attach_region(rh_info_t * info, unsigned long start, int size) rh_attach_region() argument 336 s = start; rh_attach_region() 340 /* Round start up */ rh_attach_region() 350 start = s; rh_attach_region() 359 blk->start = start; rh_attach_region() 370 unsigned long rh_detach_region(rh_info_t * info, unsigned long start, int size) rh_detach_region() argument 381 s = start; rh_detach_region() 385 /* Round start up */ rh_detach_region() 398 bs = blk->start; rh_detach_region() 399 be = blk->start + blk->size; rh_detach_region() 416 /* blk still in free list, with updated start and/or size */ rh_detach_region() 419 blk->start += size; rh_detach_region() 428 newblk->start = e; rh_detach_region() 447 unsigned long start, sp_size; rh_alloc_align() local 463 start = (blk->start + alignment - 1) & ~(alignment - 1); rh_alloc_align() 464 if (start + size <= blk->start + blk->size) rh_alloc_align() 481 sp_size = start - blk->start; rh_alloc_align() 486 spblk->start = blk->start; rh_alloc_align() 492 newblk->start = start; rh_alloc_align() 495 /* blk still in free list, with updated start and size rh_alloc_align() 497 blk->start = start + size; rh_alloc_align() 509 return start; rh_alloc_align() 527 unsigned long rh_alloc_fixed(rh_info_t * info, unsigned long start, int size, const char *owner) rh_alloc_fixed() argument 538 s = start; rh_alloc_fixed() 542 /* Round start up */ rh_alloc_fixed() 555 bs = blk->start; rh_alloc_fixed() 556 be = blk->start + blk->size; rh_alloc_fixed() 571 start = blk->start; rh_alloc_fixed() 574 return start; rh_alloc_fixed() 578 /* blk still in free list, with updated start and/or size */ rh_alloc_fixed() 581 blk->start += size; rh_alloc_fixed() 590 newblk2->start = e; rh_alloc_fixed() 597 newblk1->start = s; rh_alloc_fixed() 601 start = newblk1->start; rh_alloc_fixed() 604 return start; rh_alloc_fixed() 612 int rh_free(rh_info_t * info, unsigned long start) rh_free() argument 622 if (start < blk2->start) rh_free() 627 if (blk == NULL || start > (blk->start + blk->size)) rh_free() 667 stats->start = blk->start; list_for_each() 679 int rh_set_owner(rh_info_t * info, unsigned long start, const char *owner) rh_set_owner() argument 689 if (start < blk2->start) rh_set_owner() 694 if (blk == NULL || start > (blk->start + blk->size)) rh_set_owner() 723 st[i].start, st[i].start + st[i].size, rh_dump() 734 st[i].start, st[i].start + st[i].size, rh_dump() 744 blk, blk->start, blk->start + blk->size, blk->size); rh_dump_blk()
|
/linux-4.4.14/arch/um/include/asm/ |
H A D | tlb.h | 20 unsigned long start; member in struct:mmu_gather 28 if (tlb->start > address) __tlb_remove_tlb_entry() 29 tlb->start = address; __tlb_remove_tlb_entry() 38 tlb->start = TASK_SIZE; init_tlb_gather() 42 tlb->start = 0; init_tlb_gather() 48 tlb_gather_mmu(struct mmu_gather *tlb, struct mm_struct *mm, unsigned long start, unsigned long end) tlb_gather_mmu() argument 51 tlb->start = start; tlb_gather_mmu() 53 tlb->fullmm = !(start | (end+1)); tlb_gather_mmu() 58 extern void flush_tlb_mm_range(struct mm_struct *mm, unsigned long start, 64 flush_tlb_mm_range(tlb->mm, tlb->start, tlb->end); tlb_flush_mmu_tlbonly() 88 tlb_finish_mmu(struct mmu_gather *tlb, unsigned long start, unsigned long end) tlb_finish_mmu() argument
|
H A D | tlbflush.h | 19 * - flush_tlb_range(vma, start, end) flushes a range of pages 24 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, 28 extern void flush_tlb_kernel_range(unsigned long start, unsigned long end);
|
/linux-4.4.14/arch/x86/pci/ |
H A D | bus_numa.c | 14 if (info->busn.start == bus) x86_find_pci_root_info() 83 info->busn.start = bus_min; alloc_pci_root_info() 94 void update_res(struct pci_root_info *info, resource_size_t start, update_res() argument 100 if (start > end) update_res() 103 if (start == MAX_RESOURCE) update_res() 118 common_start = max(res->start, start); update_res() 123 final_start = min(res->start, start); update_res() 126 res->start = final_start; update_res() 141 res->start = start; update_res()
|
H A D | amd_bus.c | 70 u64 start; early_root_info_init() local 155 start = reg & 0xfff000; early_root_info_init() 166 node, link, start, end); early_root_info_init() 171 update_res(info, start, end, IORESOURCE_IO, 1); early_root_info_init() 172 subtract_range(range, RANGE_NUM, start, end + 1); early_root_info_init() 182 update_res(info, range[i].start, range[i].end - 1, early_root_info_init() 206 fam10h_mmconf_start = fam10h_mmconf->start; early_root_info_init() 221 start = reg & 0xffffff00; /* 39:16 on 31:8*/ early_root_info_init() 222 start <<= 8; early_root_info_init() 236 node, link, start, end); early_root_info_init() 239 * mmconf range, so need to update start and end. early_root_info_init() 244 if (start >= fam10h_mmconf_start && early_root_info_init() 245 start <= fam10h_mmconf_end) { early_root_info_init() 246 start = fam10h_mmconf_end + 1; early_root_info_init() 256 if (start < fam10h_mmconf_start && early_root_info_init() 260 update_res(info, start, endx, IORESOURCE_MEM, 0); early_root_info_init() 261 subtract_range(range, RANGE_NUM, start, early_root_info_init() 263 printk(KERN_CONT " ==> [%llx, %llx]", start, endx); early_root_info_init() 264 start = fam10h_mmconf_end + 1; early_root_info_init() 268 if (start <= end) { early_root_info_init() 269 printk(KERN_CONT " %s [%llx, %llx]", endx ? "and" : "==>", start, end); early_root_info_init() 277 update_res(info, cap_resource(start), cap_resource(end), early_root_info_init() 279 subtract_range(range, RANGE_NUM, start, end + 1); early_root_info_init() 307 update_res(info, cap_resource(range[i].start), early_root_info_init() 317 busnum = info->busn.start; early_root_info_init()
|
/linux-4.4.14/arch/xtensa/include/asm/ |
H A D | sysmem.h | 17 unsigned long start; member in struct:meminfo 22 * Bank array is sorted by .start. 33 int add_sysmem_bank(unsigned long start, unsigned long end);
|
/linux-4.4.14/arch/arm/plat-iop/ |
H A D | adma.c | 54 .start = IOP3XX_DMA_PHYS_BASE(0), 59 .start = IRQ_DMA0_EOT, 64 .start = IRQ_DMA0_EOC, 69 .start = IRQ_DMA0_ERR, 77 .start = IOP3XX_DMA_PHYS_BASE(1), 82 .start = IRQ_DMA1_EOT, 87 .start = IRQ_DMA1_EOC, 92 .start = IRQ_DMA1_ERR, 101 .start = IOP3XX_AAU_PHYS_BASE, 106 .start = IRQ_AA_EOT, 111 .start = IRQ_AA_EOC, 116 .start = IRQ_AA_ERR,
|
/linux-4.4.14/arch/arm/mach-ks8695/ |
H A D | devices.c | 42 .start = KS8695_WAN_PA, 48 .start = KS8695_IRQ_WAN_RX_STATUS, 54 .start = KS8695_IRQ_WAN_TX_STATUS, 60 .start = KS8695_IRQ_WAN_LINK, 66 .start = KS8695_MISC_PA, 86 .start = KS8695_LAN_PA, 92 .start = KS8695_IRQ_LAN_RX_STATUS, 98 .start = KS8695_IRQ_LAN_TX_STATUS, 104 .start = KS8695_SWITCH_PA, 124 .start = KS8695_HPNA_PA, 130 .start = KS8695_IRQ_HPNA_RX_STATUS, 136 .start = KS8695_IRQ_HPNA_TX_STATUS,
|
/linux-4.4.14/arch/blackfin/mach-bf609/boards/ |
H A D | ezkit.c | 46 .start = 0x2C0C0000, 51 .start = IRQ_PG7, 91 .start = CNT_CONFIG, 96 .start = IRQ_CNT, 149 .start = EMAC0_MACCFG, 155 .start = IRQ_EMAC0_STAT, 217 .start = UART0_REVID, 223 .start = PORTD_FER, 228 .start = PORTD_MUX, 234 .start = IRQ_UART0_TX, 239 .start = IRQ_UART0_RX, 244 .start = IRQ_UART0_STAT, 249 .start = CH_UART0_TX, 254 .start = CH_UART0_RX, 260 .start = GPIO_PD10, 265 .start = GPIO_PD9, 293 .start = UART1_REVID, 299 .start = PORTG_FER_SET, 305 .start = IRQ_UART1_TX, 310 .start = IRQ_UART1_RX, 315 .start = IRQ_UART1_STAT, 320 .start = CH_UART1_TX, 325 .start = CH_UART1_RX, 331 .start = GPIO_PG13, 336 .start = GPIO_PG10, 367 .start = 0xFFC00400, 372 .start = IRQ_UART0_TX, 377 .start = CH_UART0_TX, 392 .start = 0xFFC02000, 397 .start = IRQ_UART1_TX, 402 .start = CH_UART1_TX, 419 .start = 0xFFCC1000, 424 .start = IRQ_USB_STAT, 430 .start = IRQ_USB_DMA, 476 .start = SPORT0_TCR1, 481 .start = IRQ_SPORT0_RX, 486 .start = IRQ_SPORT0_ERROR, 510 .start = SPORT1_TCR1, 515 .start = IRQ_SPORT1_RX, 520 .start = IRQ_SPORT1_ERROR, 544 .start = SPORT2_TCR1, 549 .start = IRQ_SPORT2_RX, 554 .start = IRQ_SPORT2_ERROR, 585 .start = 0xFFC00A00, 590 .start = IRQ_CAN0_RX, 595 .start = IRQ_CAN0_TX, 600 .start = IRQ_CAN0_STAT, 646 .start = 0xFFC03B00, 651 .start = CH_NFC, 730 .start = 0xb0000000, 794 .start = SPORT0_CTL_A, 799 .start = SPORT0_CTL_B, 804 .start = CH_SPORT0_TX, 809 .start = CH_SPORT0_RX, 814 .start = IRQ_SPORT0_TX_STAT, 819 .start = IRQ_SPORT0_RX_STAT, 1214 .start = EPPI2_STAT, 1219 .start = CH_EPPI2_CH0, 1224 .start = IRQ_EPPI2_STAT, 1244 .start = REG_CRC0_CTL, 1249 .start = IRQ_CRC0_DCNTEXP, 1254 .start = CH_MEM_STREAM0_SRC_CRC0, 1259 .start = CH_MEM_STREAM0_DEST_CRC0, 1274 .start = REG_CRC1_CTL, 1279 .start = IRQ_CRC1_DCNTEXP, 1284 .start = CH_MEM_STREAM1_SRC_CRC1, 1289 .start = CH_MEM_STREAM1_DEST_CRC1, 1309 .start = REG_CRC0_CTL, 1314 .start = IRQ_CRC0_DCNTEXP, 1319 .start = CH_MEM_STREAM0_SRC_CRC0, 1365 .start = PINT0_MASK_SET, 1370 .start = IRQ_PINT0, 1385 .start = PINT1_MASK_SET, 1390 .start = IRQ_PINT1, 1405 .start = PINT2_MASK_SET, 1410 .start = IRQ_PINT2, 1425 .start = PINT3_MASK_SET, 1430 .start = IRQ_PINT3, 1445 .start = PINT4_MASK_SET, 1450 .start = IRQ_PINT4, 1465 .start = PINT5_MASK_SET, 1470 .start = IRQ_PINT5, 1485 .start = PORTA_FER, 1490 .start = IRQ_PA0, 1516 .start = PORTB_FER, 1521 .start = IRQ_PB0, 1547 .start = PORTC_FER, 1552 .start = IRQ_PC0, 1578 .start = PORTD_FER, 1583 .start = IRQ_PD0, 1609 .start = PORTE_FER, 1614 .start = IRQ_PE0, 1640 .start = PORTF_FER, 1645 .start = IRQ_PF0, 1671 .start = PORTG_FER, 1676 .start = IRQ_PG0, 1772 .start = SPI0_REGBASE, 1777 .start = CH_SPI0_TX, 1782 .start = CH_SPI0_RX, 1791 .start = SPI1_REGBASE, 1796 .start = CH_SPI1_TX, 1801 .start = CH_SPI1_RX, 1845 .start = TWI0_CLKDIV, 1850 .start = IRQ_TWI0, 1870 .start = TWI1_CLKDIV, 1875 .start = IRQ_TWI1,
|
/linux-4.4.14/fs/pstore/ |
H A D | ram_core.c | 33 atomic_t start; member in struct:persistent_ram_buffer 47 return atomic_read(&prz->buffer->start); buffer_start() 50 /* increase and wrap the start pointer, returning the old value */ buffer_start_add_atomic() 57 old = atomic_read(&prz->buffer->start); buffer_start_add_atomic() 61 } while (atomic_cmpxchg(&prz->buffer->start, old, new) != old); buffer_start_add_atomic() 85 /* increase and wrap the start pointer, returning the old value */ buffer_start_add_locked() 94 old = atomic_read(&prz->buffer->start); buffer_start_add_locked() 98 atomic_set(&prz->buffer->start, new); buffer_start_add_locked() 156 unsigned int start, unsigned int count) persistent_ram_update_ecc() 169 block = buffer->data + (start & ~(ecc_block_size - 1)); persistent_ram_update_ecc() 170 par = prz->par_buffer + (start / ecc_block_size) * ecc_size; persistent_ram_update_ecc() 178 } while (block < buffer->data + start + count); persistent_ram_update_ecc() 299 const void *s, unsigned int start, unsigned int count) persistent_ram_update() 302 memcpy(buffer->data + start, s, count); persistent_ram_update() 303 persistent_ram_update_ecc(prz, start, count); persistent_ram_update() 310 size_t start = buffer_start(prz); persistent_ram_save_old() local 325 memcpy(prz->old_log, &buffer->data[start], size - start); persistent_ram_save_old() 326 memcpy(prz->old_log + size - start, &buffer->data[0], start); persistent_ram_save_old() 334 size_t start; persistent_ram_write() local 343 start = buffer_start_add(prz, c); persistent_ram_write() 345 rem = prz->buffer_size - start; persistent_ram_write() 347 persistent_ram_update(prz, s, start, rem); persistent_ram_write() 350 start = 0; persistent_ram_write() 352 persistent_ram_update(prz, s, start, c); persistent_ram_write() 378 atomic_set(&prz->buffer->start, 0); persistent_ram_zap() 383 static void *persistent_ram_vmap(phys_addr_t start, size_t size, persistent_ram_vmap() argument 393 page_start = start - offset_in_page(start); persistent_ram_vmap() 394 page_count = DIV_ROUND_UP(size + offset_in_page(start), PAGE_SIZE); persistent_ram_vmap() 418 static void *persistent_ram_iomap(phys_addr_t start, size_t size, persistent_ram_iomap() argument 423 if (!request_mem_region(start, size, "persistent_ram")) { persistent_ram_iomap() 425 (unsigned long long)size, (unsigned long long)start); persistent_ram_iomap() 433 va = ioremap(start, size); persistent_ram_iomap() 435 va = ioremap_wc(start, size); persistent_ram_iomap() 440 static int persistent_ram_buffer_map(phys_addr_t start, phys_addr_t size, persistent_ram_buffer_map() argument 443 prz->paddr = start; persistent_ram_buffer_map() 446 if (pfn_valid(start >> PAGE_SHIFT)) persistent_ram_buffer_map() 447 prz->vaddr = persistent_ram_vmap(start, size, memtype); persistent_ram_buffer_map() 449 prz->vaddr = persistent_ram_iomap(start, size, memtype); persistent_ram_buffer_map() 453 (unsigned long long)size, (unsigned long long)start); persistent_ram_buffer_map() 457 prz->buffer = prz->vaddr + offset_in_page(start); persistent_ram_buffer_map() 477 pr_info("found existing invalid buffer, size %zu, start %zu\n", persistent_ram_post_init() 480 pr_debug("found existing buffer, size %zu, start %zu\n", persistent_ram_post_init() 514 struct persistent_ram_zone *persistent_ram_new(phys_addr_t start, size_t size, persistent_ram_new() argument 527 ret = persistent_ram_buffer_map(start, size, prz, memtype); persistent_ram_new() 155 persistent_ram_update_ecc(struct persistent_ram_zone *prz, unsigned int start, unsigned int count) persistent_ram_update_ecc() argument 298 persistent_ram_update(struct persistent_ram_zone *prz, const void *s, unsigned int start, unsigned int count) persistent_ram_update() argument
|
/linux-4.4.14/arch/arm/mach-w90x900/ |
H A D | dev.c | 79 .start = NUC900_FLASH_BASE, 99 .start = W90X900_PA_USBEHCIHOST, 104 .start = IRQ_USBH, 127 .start = W90X900_PA_USBOHCIHOST, 132 .start = IRQ_USBH, 154 .start = W90X900_PA_USBDEV, 159 .start = IRQ_USBD, 176 .start = W90X900_PA_EMC, 181 .start = IRQ_EMCTX, 186 .start = IRQ_EMCRX, 220 .start = W90X900_PA_I2C + SPIOFFSET, 225 .start = IRQ_SSP, 273 .start = W90X900_PA_TIMER, 278 .start = IRQ_WDT, 304 .start = W90X900_PA_RTC, 309 .start = IRQ_RTC, 326 .start = W90X900_PA_ADC, 331 .start = IRQ_ADC, 348 .start = W90X900_PA_FMI, 353 .start = IRQ_FMI, 403 .start = W90X900_PA_KPI, 408 .start = IRQ_KPI, 466 .start = W90X900_PA_LCD, 471 .start = IRQ_LCD, 494 .start = W90X900_PA_ACTL, 499 .start = IRQ_ACTL,
|
/linux-4.4.14/drivers/misc/ |
H A D | sram.c | 52 u32 start; member in struct:sram_reserve 90 phys_addr_t start, struct sram_partition *part) sram_add_pool() 99 ret = gen_pool_add_virt(part->pool, (unsigned long)part->base, start, sram_add_pool() 110 phys_addr_t start, struct sram_partition *part) sram_add_export() 115 (unsigned long long)start); sram_add_export() 128 phys_addr_t start) sram_add_partition() 134 part->base = sram->virt_base + block->start; sram_add_partition() 137 ret = sram_add_pool(sram, block, start, part); sram_add_partition() 142 ret = sram_add_export(sram, block, start, part); sram_add_partition() 175 return ra->start - rb->start; sram_reserve_cmp() 213 if (child_res.start < res->start || child_res.end > res->end) { for_each_available_child_of_node() 221 block->start = child_res.start - res->start; for_each_available_child_of_node() 252 block->start, block->start + block->size); for_each_available_child_of_node() 255 block->start, block->start + block->size); for_each_available_child_of_node() 263 rblocks[nblocks - 1].start = size; 282 if (block->start < cur_start) { 285 block->start, cur_start); 293 res->start + block->start); 300 /* current start is in a reserved block, so continue after it */ 301 if (block->start == cur_start) { 302 cur_start = block->start + block->size; 311 cur_size = block->start - cur_start; 318 res->start + cur_start, cur_size, -1); 325 cur_start = block->start + block->size; 358 if (!devm_request_mem_region(sram->dev, res->start, size, pdev->name)) { sram_probe() 363 sram->virt_base = devm_ioremap_wc(sram->dev, res->start, size); sram_probe() 89 sram_add_pool(struct sram_dev *sram, struct sram_reserve *block, phys_addr_t start, struct sram_partition *part) sram_add_pool() argument 109 sram_add_export(struct sram_dev *sram, struct sram_reserve *block, phys_addr_t start, struct sram_partition *part) sram_add_export() argument 127 sram_add_partition(struct sram_dev *sram, struct sram_reserve *block, phys_addr_t start) sram_add_partition() argument
|
/linux-4.4.14/arch/blackfin/mach-bf518/boards/ |
H A D | tcm-bf518.c | 60 .start = 0x20000000, 239 .start = SPI0_REGBASE, 244 .start = CH_SPI0, 249 .start = IRQ_SPI0, 274 .start = SPI1_REGBASE, 279 .start = CH_SPI1, 284 .start = IRQ_SPI1, 305 .start = UART0_THR, 310 .start = IRQ_UART0_TX, 315 .start = IRQ_UART0_RX, 320 .start = IRQ_UART0_ERROR, 325 .start = CH_UART0_TX, 330 .start = CH_UART0_RX, 353 .start = UART1_THR, 358 .start = IRQ_UART1_TX, 363 .start = IRQ_UART1_RX, 368 .start = IRQ_UART1_ERROR, 373 .start = CH_UART1_TX, 378 .start = CH_UART1_RX, 404 .start = 0xFFC00400, 409 .start = IRQ_UART0_RX, 414 .start = CH_UART0_RX, 430 .start = 0xFFC02000, 435 .start = IRQ_UART1_RX, 440 .start = CH_UART1_RX, 460 .start = TWI0_REGBASE, 465 .start = IRQ_TWI, 500 .start = SPORT0_TCR1, 505 .start = IRQ_SPORT0_RX, 510 .start = IRQ_SPORT0_ERROR, 534 .start = SPORT1_TCR1, 539 .start = IRQ_SPORT1_RX, 544 .start = IRQ_SPORT1_ERROR,
|
/linux-4.4.14/arch/um/drivers/ |
H A D | slip_common.c | 7 int i, n, size, start; slip_proto_read() local 30 start = slip->pos; slip_proto_read() 32 size = slip_unesc(slip->ibuf[start + i], slip->ibuf,&slip->pos, slip_proto_read() 36 memmove(slip->ibuf, &slip->ibuf[start+i+1], slip_proto_read()
|
/linux-4.4.14/arch/unicore32/kernel/ |
H A D | puv3-core.c | 52 .start = io_v2p(PKUNITY_USB_BASE), 56 .start = IRQ_USB, 59 .start = IRQ_USB, 84 .start = io_v2p(PKUNITY_SDC_BASE), 89 .start = IRQ_SDC, 97 .start = io_v2p(PKUNITY_UNIGFX_BASE), 105 .start = io_v2p(PKUNITY_RTC_BASE), 110 .start = IRQ_RTCAlarm, 115 .start = IRQ_RTC, 123 .start = io_v2p(PKUNITY_OST_BASE) + 0x80, 131 .start = io_v2p(PKUNITY_UART0_BASE), 136 .start = IRQ_UART0, 144 .start = io_v2p(PKUNITY_UART1_BASE), 149 .start = IRQ_UART1, 157 .start = io_v2p(PKUNITY_UMAL_BASE), 162 .start = IRQ_UMAL,
|
H A D | setup.c | 73 .start = 0, 79 .start = 0, 134 static int __init uc32_add_memory(unsigned long start, unsigned long size) uc32_add_memory() argument 140 "ignoring memory at %#lx\n", start); uc32_add_memory() 145 * Ensure that start/size are aligned to a page boundary. uc32_add_memory() 146 * Size is appropriately rounded down, start is rounded up. uc32_add_memory() 148 size -= start & ~PAGE_MASK; uc32_add_memory() 150 bank->start = PAGE_ALIGN(start); uc32_add_memory() 165 * Pick out the memory size. We look for mem=size@start, 166 * where start and size are "size[KkMm]" 171 unsigned long size, start; early_mem() local 184 start = PHYS_OFFSET; early_mem() 187 start = memparse(endp + 1, NULL); early_mem() 189 uc32_add_memory(start, size); early_mem() 201 kernel_code.start = virt_to_phys(_stext); request_standard_resources() 203 kernel_data.start = virt_to_phys(_sdata); request_standard_resources() 212 res->start = mi->bank[i].start; request_standard_resources() 213 res->end = mi->bank[i].start + mi->bank[i].size - 1; request_standard_resources() 218 if (kernel_code.start >= res->start && request_standard_resources() 221 if (kernel_data.start >= res->start && request_standard_resources() 350 .start = c_start,
|
/linux-4.4.14/include/asm-generic/ |
H A D | cacheflush.h | 14 #define flush_cache_range(vma, start, end) do { } while (0) 20 #define flush_icache_range(start, end) do { } while (0) 23 #define flush_cache_vmap(start, end) do { } while (0) 24 #define flush_cache_vunmap(start, end) do { } while (0)
|
/linux-4.4.14/arch/powerpc/mm/ |
H A D | tlb_hash32.c | 82 * - flush_tlb_range(vma, start, end) flushes a range of pages 83 * - flush_tlb_kernel_range(start, end) flushes kernel pages 90 static void flush_range(struct mm_struct *mm, unsigned long start, flush_range() argument 102 start &= PAGE_MASK; flush_range() 103 if (start >= end) flush_range() 106 pmd = pmd_offset(pud_offset(pgd_offset(mm, start), start), start); flush_range() 108 pmd_end = ((start + PGDIR_SIZE) & PGDIR_MASK) - 1; flush_range() 112 count = ((pmd_end - start) >> PAGE_SHIFT) + 1; flush_range() 113 flush_hash_pages(ctx, start, pmd_val(*pmd), count); flush_range() 117 start = pmd_end + 1; flush_range() 125 void flush_tlb_kernel_range(unsigned long start, unsigned long end) flush_tlb_kernel_range() argument 127 flush_range(&init_mm, start, end); flush_tlb_kernel_range() 175 void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, flush_tlb_range() argument 178 flush_range(vma->vm_mm, start, end); flush_tlb_range()
|
H A D | init_64.c | 158 * represents the start of the section it is within. Note that we have to 166 /* Return the pfn of the start of the section. */ vmemmap_section_start() 175 static int __meminit vmemmap_populated(unsigned long start, int page_size) vmemmap_populated() argument 177 unsigned long end = start + page_size; vmemmap_populated() 178 start = (unsigned long)(pfn_to_page(vmemmap_section_start(start))); vmemmap_populated() 180 for (; start < end; start += (PAGES_PER_SECTION * sizeof(struct page))) vmemmap_populated() 181 if (pfn_valid(page_to_pfn((struct page *)start))) vmemmap_populated() 195 static void __meminit vmemmap_create_mapping(unsigned long start, vmemmap_create_mapping() argument 214 BUG_ON(map_kernel_page(start + i, phys, flags)); vmemmap_create_mapping() 218 static void vmemmap_remove_mapping(unsigned long start, vmemmap_remove_mapping() argument 224 static void __meminit vmemmap_create_mapping(unsigned long start, vmemmap_create_mapping() argument 228 int mapped = htab_bolt_mapping(start, start + page_size, phys, vmemmap_create_mapping() 236 static void vmemmap_remove_mapping(unsigned long start, vmemmap_remove_mapping() argument 239 int mapped = htab_remove_mapping(start, start + page_size, vmemmap_remove_mapping() 281 unsigned long start, vmemmap_list_populate() 293 vmem_back->virt_addr = start; vmemmap_list_populate() 299 int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node) vmemmap_populate() argument 304 start = _ALIGN_DOWN(start, page_size); vmemmap_populate() 306 pr_debug("vmemmap_populate %lx..%lx, node %d\n", start, end, node); vmemmap_populate() 308 for (; start < end; start += page_size) { vmemmap_populate() 311 if (vmemmap_populated(start, page_size)) vmemmap_populate() 318 vmemmap_list_populate(__pa(p), start, node); vmemmap_populate() 321 start, start + page_size, p); vmemmap_populate() 323 vmemmap_create_mapping(start, page_size, __pa(p)); vmemmap_populate() 330 static unsigned long vmemmap_list_free(unsigned long start) vmemmap_list_free() argument 338 if (vmem_back->virt_addr == start) vmemmap_list_free() 362 void __ref vmemmap_free(unsigned long start, unsigned long end) vmemmap_free() argument 366 start = _ALIGN_DOWN(start, page_size); vmemmap_free() 368 pr_debug("vmemmap_free %lx...%lx\n", start, end); vmemmap_free() 370 for (; start < end; start += page_size) { vmemmap_free() 378 if (vmemmap_populated(start, page_size)) vmemmap_free() 381 addr = vmemmap_list_free(start); vmemmap_free() 403 vmemmap_remove_mapping(start, page_size); vmemmap_free() 280 vmemmap_list_populate(unsigned long phys, unsigned long start, int node) vmemmap_list_populate() argument
|
/linux-4.4.14/arch/sh/kernel/cpu/sh4a/ |
H A D | pinmux-sh7722.c | 9 .start = 0xa4050100,
|
/linux-4.4.14/arch/metag/include/asm/ |
H A D | cacheflush.h | 14 void metag_data_cache_flush_all(const void *start); 15 void metag_code_cache_flush_all(const void *start); 24 void metag_data_cache_flush(const void *start, int bytes); 25 void metag_code_cache_flush(const void *start, int bytes); 50 unsigned long start, unsigned long end) flush_cache_range() 76 static inline void flush_cache_vmap(unsigned long start, unsigned long end) flush_cache_vmap() argument 81 static inline void flush_cache_vunmap(unsigned long start, unsigned long end) flush_cache_vunmap() argument 93 #define flush_cache_range(vma, start, end) do { } while (0) 98 #define flush_cache_vmap(start, end) do { } while (0) 99 #define flush_cache_vunmap(start, end) do { } while (0) 157 static inline void cachew_region_op(void *start, unsigned long size, cachew_region_op() argument 160 unsigned long offset = (unsigned long)start & 0x3f; cachew_region_op() 164 start -= offset; cachew_region_op() 168 __builtin_meta2_cachewd(start, op); cachew_region_op() 169 start += 0x40; cachew_region_op() 198 static inline void flush_dcache_region(void *start, unsigned long size) flush_dcache_region() argument 202 cachew_region_op(start, size, CACHEW_FLUSH_L1D_L2); flush_dcache_region() 204 l2c_fence_flush(start + size - 1); flush_dcache_region() 206 metag_data_cache_flush(start, size); flush_dcache_region() 211 static inline void writeback_dcache_region(void *start, unsigned long size) writeback_dcache_region() argument 214 cachew_region_op(start, size, CACHEW_WRITEBACK_L1D_L2); writeback_dcache_region() 215 l2c_fence(start + size - 1); writeback_dcache_region() 220 static inline void invalidate_dcache_region(void *start, unsigned long size) invalidate_dcache_region() argument 223 cachew_region_op(start, size, CACHEW_INVALIDATE_L1D_L2); invalidate_dcache_region() 225 metag_data_cache_flush(start, size); invalidate_dcache_region() 49 flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_cache_range() argument
|
H A D | tlbflush.h | 16 * - flush_tlb_range(mm, start, end) flushes a range of pages 17 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages 18 * - flush_tlb_pgtables(mm, start, end) flushes a range of page tables 59 unsigned long start, unsigned long end) flush_tlb_range() 65 unsigned long start, unsigned long end) flush_tlb_pgtables() 70 static inline void flush_tlb_kernel_range(unsigned long start, flush_tlb_kernel_range() argument 58 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_tlb_range() argument 64 flush_tlb_pgtables(struct mm_struct *mm, unsigned long start, unsigned long end) flush_tlb_pgtables() argument
|
/linux-4.4.14/arch/microblaze/include/asm/ |
H A D | cacheflush.h | 60 #define flush_icache_range(start, end) mbc->iflr(start, end); 62 #define invalidate_icache_range(start, end) mbc->iinr(start, end); 71 #define invalidate_dcache_range(start, end) mbc->dinr(start, end); 73 #define flush_dcache_range(start, end) mbc->dflr(start, end); 88 #define flush_cache_vmap(start, end) do { } while (0) 89 #define flush_cache_vunmap(start, end) do { } while (0) 97 #define flush_cache_range(vma, start, len) { \ 98 flush_icache_range((unsigned) (start), (unsigned) (start) + (len)); \ 99 flush_dcache_range((unsigned) (start), (unsigned) (start) + (len)); \ 103 #define flush_cache_range(vma, start, len) do { } while (0)
|
H A D | tlbflush.h | 37 unsigned long start, unsigned long end) local_flush_tlb_range() 40 #define flush_tlb_kernel_range(start, end) do { } while (0) 55 unsigned long start, unsigned long end) { } 63 #define flush_tlb_range(mm, start, end) BUG() 64 #define flush_tlb_pgtables(mm, start, end) BUG() 65 #define flush_tlb_kernel_range(start, end) BUG() 36 local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) local_flush_tlb_range() argument 54 flush_tlb_pgtables(struct mm_struct *mm, unsigned long start, unsigned long end) flush_tlb_pgtables() argument
|
/linux-4.4.14/arch/cris/include/asm/ |
H A D | cacheflush.h | 13 #define flush_cache_range(vma, start, end) do { } while (0) 19 #define flush_icache_range(start, end) do { } while (0) 22 #define flush_cache_vmap(start, end) do { } while (0) 23 #define flush_cache_vunmap(start, end) do { } while (0)
|
/linux-4.4.14/arch/ia64/include/uapi/asm/ |
H A D | auxvec.h | 6 * them, start the architecture-specific ones at 32.
|
/linux-4.4.14/arch/avr32/include/asm/ |
H A D | tlbflush.h | 20 * - flush_tlb_range(vma, start, end) flushes a range of pages 21 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages 26 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, 30 extern void flush_tlb_kernel_range(unsigned long start, unsigned long end);
|
/linux-4.4.14/arch/arm/mach-mmp/include/mach/ |
H A D | devices.h | 14 unsigned long start; member in struct:pxa_device_desc 25 .start = _start, \ 36 .start = _start, \ 47 .start = _start, \
|
/linux-4.4.14/tools/build/feature/ |
H A D | test-libunwind-debug-frame.c | 7 const char *obj_name, unw_word_t start,
|
/linux-4.4.14/tools/lib/symbol/ |
H A D | kallsyms.c | 13 char type, u64 start)) kallsyms__parse() 26 u64 start; kallsyms__parse() local 37 len = hex2u64(line, &start); kallsyms__parse() 53 err = process_symbol(arg, symbol_name, symbol_type, start); kallsyms__parse() 11 kallsyms__parse(const char *filename, void *arg, int (*process_symbol)(void *arg, const char *name, char type, u64 start)) kallsyms__parse() argument
|
/linux-4.4.14/arch/x86/kvm/ |
H A D | mtrr.c | 146 u64 start; member in struct:fixed_mtrr_segment 151 /* the start position in kvm_mtrr.fixed_ranges[]. */ 158 .start = 0x0, 169 .start = 0x80000, 180 .start = 0xc0000, 218 static void fixed_mtrr_seg_unit_range(int seg, int unit, u64 *start, u64 *end) fixed_mtrr_seg_unit_range() argument 223 *start = mtrr_seg->start + unit * unit_size; fixed_mtrr_seg_unit_range() 224 *end = *start + unit_size; fixed_mtrr_seg_unit_range() 232 WARN_ON(mtrr_seg->start + unit * fixed_mtrr_seg_unit_size(seg) fixed_mtrr_seg_unit_range_index() 244 n = (mtrr_seg->end - mtrr_seg->start) >> mtrr_seg->range_shift; fixed_mtrr_seg_end_range_index() 248 static bool fixed_msr_to_range(u32 msr, u64 *start, u64 *end) fixed_msr_to_range() argument 255 fixed_mtrr_seg_unit_range(seg, unit, start, end); fixed_msr_to_range() 276 if (mtrr_seg->start <= addr && addr < mtrr_seg->end) fixed_mtrr_addr_to_seg() 290 index += (addr - mtrr_seg->start) >> mtrr_seg->range_shift; fixed_mtrr_addr_seg_to_range_index() 299 return mtrr_seg->start + ((pos + 1) << mtrr_seg->range_shift); fixed_mtrr_range_end_addr() 302 static void var_mtrr_range(struct kvm_mtrr_range *range, u64 *start, u64 *end) var_mtrr_range() argument 306 *start = range->base & PAGE_MASK; var_mtrr_range() 313 *end = (*start | ~mask) + 1; var_mtrr_range() 319 gfn_t start, end; update_mtrr() local 330 if (fixed_msr_to_range(msr, &start, &end)) { update_mtrr() 334 start = 0x0; update_mtrr() 339 var_mtrr_range(&mtrr_state->var_ranges[index], &start, &end); update_mtrr() 342 kvm_zap_gfn_range(vcpu->kvm, gpa_to_gfn(start), gpa_to_gfn(end)); update_mtrr() 453 u64 start; member in struct:mtrr_iter 460 /* [start, end) is not fully covered in MTRRs? */ 489 seg = fixed_mtrr_addr_to_seg(iter->start); mtrr_lookup_fixed_start() 494 index = fixed_mtrr_addr_seg_to_range_index(iter->start, seg); mtrr_lookup_fixed_start() 503 u64 start, end; match_var_range() local 505 var_mtrr_range(range, &start, &end); match_var_range() 506 if (!(start >= iter->end || end <= iter->start)) { match_var_range() 514 iter->partial_map |= iter->start_max < start; match_var_range() 541 iter->start_max = iter->start; mtrr_lookup_var_start() 584 struct kvm_mtrr *mtrr_state, u64 start, u64 end) mtrr_lookup_init() 587 iter->start = start; mtrr_lookup_init() 628 u64 start, end; kvm_mtrr_get_guest_memory_type() local 633 start = gfn_to_gpa(gfn); kvm_mtrr_get_guest_memory_type() 634 end = start + PAGE_SIZE; kvm_mtrr_get_guest_memory_type() 636 mtrr_for_each_mem_type(&iter, mtrr_state, start, end) { kvm_mtrr_get_guest_memory_type() 705 u64 start, end; kvm_mtrr_check_gfn_range_consistency() local 708 start = gfn_to_gpa(gfn); kvm_mtrr_check_gfn_range_consistency() 710 mtrr_for_each_mem_type(&iter, mtrr_state, start, end) { kvm_mtrr_check_gfn_range_consistency() 583 mtrr_lookup_init(struct mtrr_iter *iter, struct kvm_mtrr *mtrr_state, u64 start, u64 end) mtrr_lookup_init() argument
|
/linux-4.4.14/arch/mips/jazz/ |
H A D | setup.c | 32 .start = 0x00, 37 .start = 0x40, 42 .start = 0x80, 47 .start = 0xc0, 123 .start = JAZZ_SCSI_BASE, 128 .start = JAZZ_SCSI_DMA, 133 .start = JAZZ_SCSI_IRQ, 147 .start = JAZZ_ETHERNET_BASE, 152 .start = JAZZ_ETHERNET_IRQ, 166 .start = 0x70, 171 .start = 8,
|
/linux-4.4.14/arch/arm64/kernel/ |
H A D | sys_compat.c | 32 __do_compat_cache_op(unsigned long start, unsigned long end) __do_compat_cache_op() argument 37 unsigned long chunk = min(PAGE_SIZE, end - start); __do_compat_cache_op() 42 ret = __flush_cache_user_range(start, start + chunk); __do_compat_cache_op() 47 start += chunk; __do_compat_cache_op() 48 } while (start < end); __do_compat_cache_op() 54 do_compat_cache_op(unsigned long start, unsigned long end, int flags) do_compat_cache_op() argument 56 if (end < start || flags) do_compat_cache_op() 59 if (!access_ok(VERIFY_READ, start, end - start)) do_compat_cache_op() 62 return __do_compat_cache_op(start, end); do_compat_cache_op()
|
/linux-4.4.14/arch/mips/kernel/ |
H A D | setup.c | 91 void __init add_memory_region(phys_addr_t start, phys_addr_t size, long type) add_memory_region() argument 97 if (start + size < start) { add_memory_region() 112 if (start + size < entry->addr) add_memory_region() 115 if (entry->addr + entry->size < start) add_memory_region() 118 top = max(entry->addr + entry->size, start + size); add_memory_region() 119 entry->addr = min(entry->addr, start); add_memory_region() 130 boot_mem_map.map[x].addr = start; add_memory_region() 136 void __init detect_memory_region(phys_addr_t start, phys_addr_t sz_min, phys_addr_t sz_max) detect_memory_region() argument 148 (unsigned long long) start, detect_memory_region() 152 add_memory_region(start, size, BOOT_MEM_RAM); detect_memory_region() 192 unsigned long start = memparse(p, &p); rd_start_early() local 196 if (start < XKPHYS) rd_start_early() 197 start = (int)start; rd_start_early() 199 initrd_start = start; rd_start_early() 200 initrd_end += start; rd_start_early() 226 pr_err("initrd start must be page aligned\n"); init_initrd() 230 pr_err("initrd start < PAGE_OFFSET\n"); init_initrd() 330 unsigned long start, end; bootmem_init() local 335 start = PFN_UP(boot_mem_map.map[i].addr); bootmem_init() 341 if (start < min_low_pfn) bootmem_init() 342 min_low_pfn = start; bootmem_init() 350 if (start >= mapstart) bootmem_init() 352 mapstart = max(reserved_end, start); bootmem_init() 395 unsigned long start, end; bootmem_init() local 397 start = PFN_UP(boot_mem_map.map[i].addr); bootmem_init() 401 if (start <= min_low_pfn) bootmem_init() 402 start = min_low_pfn; bootmem_init() 403 if (start >= end) bootmem_init() 413 if (end <= start) bootmem_init() 417 memblock_add_node(PFN_PHYS(start), PFN_PHYS(end - start), 0); bootmem_init() 424 unsigned long start, end, size; bootmem_init() local 426 start = PFN_UP(boot_mem_map.map[i].addr); bootmem_init() 437 memory_present(0, start, end); bootmem_init() 445 * We are rounding up the start address of usable memory bootmem_init() 448 if (start >= max_low_pfn) bootmem_init() 450 if (start < reserved_end) bootmem_init() 451 start = reserved_end; bootmem_init() 458 if (end <= start) bootmem_init() 460 size = end - start; bootmem_init() 463 free_bootmem(PFN_PHYS(start), size << PAGE_SHIFT); bootmem_init() 464 memory_present(0, start, end); bootmem_init() 507 phys_addr_t start, size; early_parse_mem() local 518 start = 0; early_parse_mem() 521 start = memparse(p + 1, &p); early_parse_mem() 523 add_memory_region(start, size, BOOT_MEM_RAM); early_parse_mem() 537 unsigned long start = boot_mem_map.map[i].addr; early_parse_elfcorehdr() local 540 if (setup_elfcorehdr >= start && setup_elfcorehdr < end) { early_parse_elfcorehdr() 599 crashk_res.start = crash_base; mips_parse_crashkernel() 611 crashk_res.start + 1) >> 20), request_crashkernel() 612 (unsigned long)(crashk_res.start >> 20)); request_crashkernel() 694 if (crashk_res.start != crashk_res.end) arch_mem_init() 695 reserve_bootmem(crashk_res.start, arch_mem_init() 696 crashk_res.end - crashk_res.start + 1, arch_mem_init() 721 code_resource.start = __pa_symbol(&_text); resource_init() 723 data_resource.start = __pa_symbol(&_etext); resource_init() 728 unsigned long start, end; resource_init() local 730 start = boot_mem_map.map[i].addr; resource_init() 732 if (start >= HIGHMEM_START) resource_init() 749 res->start = start; resource_init()
|
H A D | spinlock_test.c | 12 ktime_t start, finish; ss_get() local 20 start = ktime_get(); ss_get() 32 *val = ktime_us_delta(finish, start); ss_get() 51 ktime_t start; member in struct:spin_multi_per_thread 69 pt->start = ktime_get(); multi_other() 111 *val = ktime_us_delta(finish, t1.start); multi_get()
|
/linux-4.4.14/arch/alpha/include/asm/ |
H A D | a.out.h | 7 /* Assume that start addresses below 4G belong to a TASO application. 9 Worse, we have to notice the start address before swapping to use
|
/linux-4.4.14/arch/cris/arch-v32/drivers/pci/ |
H A D | bios.c | 43 resource_size_t start = res->start; pcibios_align_resource() local 45 if ((res->flags & IORESOURCE_IO) && (start & 0x300)) pcibios_align_resource() 46 start = (start + 0x3ff) & ~0x3ff; pcibios_align_resource() 48 return start; pcibios_align_resource() 65 if (!r->start && r->end) { pcibios_enable_resources() 74 if (dev->resource[PCI_ROM_RESOURCE].start) pcibios_enable_resources()
|
/linux-4.4.14/fs/befs/ |
H A D | endian.h | 77 run.start = le16_to_cpu((__force __le16)n.start); fsrun_to_cpu() 81 run.start = be16_to_cpu((__force __be16)n.start); fsrun_to_cpu() 94 run.start = cpu_to_le16(n.start); cpu_to_fsrun() 98 run.start = cpu_to_be16(n.start); cpu_to_fsrun()
|
/linux-4.4.14/arch/m68k/mm/ |
H A D | cache.c | 78 unsigned long start, end; flush_icache_range() local 79 start = address & ICACHE_SET_MASK; flush_icache_range() 81 if (start > end) { flush_icache_range() 85 flush_cf_icache(start, end); flush_icache_range() 112 unsigned long start, end; flush_icache_user_range() local 113 start = addr & ICACHE_SET_MASK; flush_icache_user_range() 115 if (start > end) { flush_icache_user_range() 119 flush_cf_icache(start, end); flush_icache_user_range()
|
/linux-4.4.14/arch/mips/vr41xx/common/ |
H A D | rtc.c | 31 .start = 0x0b0000c0, 36 .start = 0x0b0001c0, 41 .start = ELAPSEDTIME_IRQ, 46 .start = RTCLONG1_IRQ, 54 .start = 0x0f000100, 59 .start = 0x0f000120, 64 .start = ELAPSEDTIME_IRQ, 69 .start = RTCLONG1_IRQ,
|
/linux-4.4.14/arch/arm/include/asm/hardware/ |
H A D | memc.h | 19 #define video_set_dma(start,end,offset) \ 21 memc_write (VDMA_START, (start >> 2)); \
|
/linux-4.4.14/arch/avr32/kernel/ |
H A D | setup.c | 50 .start = 0, 56 .start = 0, 83 kernel_code.start = __pa(init_mm.start_code); resource_init() 92 mem->start, mem->end); resource_init() 103 res->name, res->start, res->end); resource_init() 108 add_physical_memory(resource_size_t start, resource_size_t end) add_physical_memory() argument 114 if (end < next->start) add_physical_memory() 116 if (start <= next->end) { add_physical_memory() 121 start, end, next->start, next->end); add_physical_memory() 129 start, end); add_physical_memory() 134 new->start = start; add_physical_memory() 143 add_reserved_region(resource_size_t start, resource_size_t end, add_reserved_region() argument 148 if (end < start) add_reserved_region() 156 if (end < next->start) add_reserved_region() 158 if (start <= next->end) add_reserved_region() 163 new->start = start; add_reserved_region() 181 target = ALIGN(mem->start, align); find_free_region() 183 if ((target + size) <= res->start) find_free_region() 196 alloc_reserved_region(resource_size_t *start, resource_size_t size, alloc_reserved_region() argument 209 *start = target; alloc_reserved_region() 287 * Pick out the memory size. We look for mem=size@start, 288 * where start and size are "size[KkMmGg]" 292 resource_size_t size, start; early_mem() local 294 start = system_ram->start; early_mem() 297 start = memparse(p + 1, &p); early_mem() 299 system_ram->start = start; early_mem() 300 system_ram->end = system_ram->start + size - 1; early_mem() 318 unsigned long start, end; parse_tag_mem() local 328 start = tag->u.mem_range.addr; parse_tag_mem() 331 add_physical_memory(start, end); parse_tag_mem() 480 printk(" %08x-%08x\n", res->start, res->end); setup_bootmem() 484 res->start, res->end, res->name); setup_bootmem() 492 first_pfn = PFN_UP(res->start); setup_bootmem() 534 if (res->start > PFN_PHYS(max_pfn)) setup_bootmem() 542 if (res->start >= PFN_PHYS(first_pfn) setup_bootmem() 545 res->start, setup_bootmem() 567 kernel_code.start = __pa(__init_begin); setup_arch() 569 kernel_data.start = __pa(init_mm.end_code); setup_arch()
|
/linux-4.4.14/arch/arc/include/asm/ |
H A D | tlbflush.h | 17 void local_flush_tlb_kernel_range(unsigned long start, unsigned long end); 19 unsigned long start, unsigned long end); 20 void local_flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start, 31 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, 34 extern void flush_tlb_kernel_range(unsigned long start, unsigned long end); 37 extern void flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end);
|
/linux-4.4.14/include/uapi/linux/netfilter/ |
H A D | nfnetlink_compat.h | 50 #define NFA_NEST_END(skb, start) \ 51 ({ (start)->nfa_len = skb_tail_pointer(skb) - (unsigned char *)(start); \ 53 #define NFA_NEST_CANCEL(skb, start) \ 54 ({ if (start) \ 55 skb_trim(skb, (unsigned char *) (start) - (skb)->data); \
|
/linux-4.4.14/arch/sh/boards/mach-sh03/ |
H A D | setup.c | 34 .start = 0x1f0, 39 .start = 0x1f0 + 0x206, 44 .start = IRL2_IRQ, 58 .start = 0xa0800000, 92 cf_ide_resources[0].start += (unsigned long)cf_ide_base; sh03_devices_setup() 94 cf_ide_resources[1].start += (unsigned long)cf_ide_base; sh03_devices_setup()
|
/linux-4.4.14/drivers/firmware/efi/ |
H A D | fake_mem.c | 50 if (m1->range.start < m2->range.start) cmp_fake_mem() 52 if (m1->range.start > m2->range.start) cmp_fake_mem() 59 u64 start, end, m_start, m_end, m_attr; efi_fake_memmap() local 73 start = md->phys_addr; efi_fake_memmap() 74 end = start + (md->num_pages << EFI_PAGE_SHIFT) - 1; efi_fake_memmap() 78 m_start = fake_mems[i].range.start; efi_fake_memmap() 81 if (m_start <= start) { efi_fake_memmap() 83 if (start < m_end && m_end < end) efi_fake_memmap() 86 if (start < m_start && m_start < end) { efi_fake_memmap() 118 start = md->phys_addr; efi_fake_memmap() 123 m_start = fake_mems[i].range.start; efi_fake_memmap() 127 if (m_start <= start && end <= m_end) efi_fake_memmap() 130 if (m_start <= start && efi_fake_memmap() 131 (start < m_end && m_end < end)) { efi_fake_memmap() 145 if ((start < m_start && m_start < end) && m_end < end) { efi_fake_memmap() 166 if ((start < m_start && m_start < end) && efi_fake_memmap() 197 u64 start = 0, mem_size = 0, attribute = 0; setup_fake_mem() local 206 start = memparse(p+1, &p); setup_fake_mem() 218 fake_mems[nr_fake_mem].range.start = start; setup_fake_mem() 219 fake_mems[nr_fake_mem].range.end = start + mem_size - 1; setup_fake_mem() 232 fake_mems[i].attribute, fake_mems[i].range.start, setup_fake_mem()
|
/linux-4.4.14/drivers/gpu/drm/radeon/ |
H A D | radeon_mem.c | 42 static struct mem_block *split_block(struct mem_block *p, int start, int size, split_block() argument 45 /* Maybe cut off the start of an existing block */ split_block() 46 if (start > p->start) { split_block() 51 newblock->start = start; split_block() 52 newblock->size = p->size - (start - p->start); split_block() 68 newblock->start = start + size; split_block() 91 int start = (p->start + mask) & ~mask; list_for_each() local 92 if (p->file_priv == NULL && start + size <= p->start + p->size) list_for_each() 93 return split_block(p, start, size, file_priv); list_for_each() 99 static struct mem_block *find_block(struct mem_block *heap, int start) find_block() argument 104 if (p->start == start) find_block() 136 static int init_heap(struct mem_block **heap, int start, int size) init_heap() argument 149 blocks->start = start; init_heap() 246 if (copy_to_user(alloc->region_offset, &block->start, radeon_mem_alloc() 301 return init_heap(heap, initheap->start, initheap->size); radeon_mem_init_heap()
|
/linux-4.4.14/drivers/media/platform/coda/ |
H A D | trace.h | 91 __field(int, start) 99 __entry->start = meta->start; 104 TP_printk("minor = %d, index = %d, start = 0x%x, end = 0x%x, ctx = %d", 105 __entry->minor, __entry->index, __entry->start, __entry->end, 122 __field(int, start) 129 __entry->start = meta ? meta->start : 0; 134 TP_printk("minor = %d, start = 0x%x, end = 0x%x, ctx = %d", 135 __entry->minor, __entry->start, __entry->end, __entry->ctx)
|
/linux-4.4.14/drivers/gpu/drm/nouveau/ |
H A D | nv50_fence.c | 41 u32 start = mem->start * PAGE_SIZE; nv50_fence_context_new() local 42 u32 limit = start + mem->size - 1; nv50_fence_context_new() 58 .start = start, nv50_fence_context_new() 66 u32 start = bo->bo.mem.start * PAGE_SIZE; nv50_fence_context_new() local 67 u32 limit = start + bo->bo.mem.size - 1; nv50_fence_context_new() 73 .start = start, nv50_fence_context_new()
|
/linux-4.4.14/arch/unicore32/mm/ |
H A D | cache-ucv2.S | 44 * __cpuc_flush_user_range(start, end, flags) 48 * - start - start address (may not be aligned) 83 * __cpuc_coherent_kern_range(start,end) 84 * __cpuc_coherent_user_range(start,end) 90 * - start - virtual start address of region 145 * __cpuc_dma_clean_range(start,end) 146 * - start - virtual start address of region 182 * __cpuc_dma_inv_range(start,end) 183 * __cpuc_dma_flush_range(start,end) 184 * - start - virtual start address of region
|
/linux-4.4.14/arch/mips/mm/ |
H A D | tlb-r3k.c | 81 void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, local_flush_tlb_range() argument 92 cpu_context(cpu, mm) & ASID_MASK, start, end); local_flush_tlb_range() 95 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_range() 100 start &= PAGE_MASK; local_flush_tlb_range() 103 while (start < end) { local_flush_tlb_range() 106 write_c0_entryhi(start | newpid); local_flush_tlb_range() 107 start += PAGE_SIZE; /* BARRIER */ local_flush_tlb_range() 124 void local_flush_tlb_kernel_range(unsigned long start, unsigned long end) local_flush_tlb_kernel_range() argument 129 printk("[tlbrange<%lu,0x%08lx,0x%08lx>]", start, end); local_flush_tlb_kernel_range() 132 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_kernel_range() 136 start &= PAGE_MASK; local_flush_tlb_kernel_range() 140 while (start < end) { local_flush_tlb_kernel_range() 143 write_c0_entryhi(start); local_flush_tlb_kernel_range() 144 start += PAGE_SIZE; /* BARRIER */ local_flush_tlb_kernel_range()
|
H A D | tlb-r8k.c | 61 void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, local_flush_tlb_range() argument 72 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_range() 87 start &= PAGE_MASK; local_flush_tlb_range() 90 while (start < end) { local_flush_tlb_range() 93 write_c0_vaddr(start); local_flush_tlb_range() 94 write_c0_entryhi(start); local_flush_tlb_range() 95 start += PAGE_SIZE; local_flush_tlb_range() 111 void local_flush_tlb_kernel_range(unsigned long start, unsigned long end) local_flush_tlb_kernel_range() argument 115 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_kernel_range() 127 start &= PAGE_MASK; local_flush_tlb_kernel_range() 130 while (start < end) { local_flush_tlb_kernel_range() 133 write_c0_vaddr(start); local_flush_tlb_kernel_range() 134 write_c0_entryhi(start); local_flush_tlb_kernel_range() 135 start += PAGE_SIZE; local_flush_tlb_kernel_range()
|
/linux-4.4.14/arch/arm64/include/asm/ |
H A D | cacheflush.h | 36 * Start addresses are inclusive and end addresses are exclusive; start 48 * flush_icache_range(start, end) 51 * region described by start, end. 52 * - start - virtual start address 55 * __flush_cache_user_range(start, end) 58 * region described by start, end. 59 * - start - virtual start address 68 extern void flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end); 69 extern void flush_icache_range(unsigned long start, unsigned long end); 71 extern long __flush_cache_user_range(unsigned long start, unsigned long end); 145 static inline void flush_cache_vmap(unsigned long start, unsigned long end) flush_cache_vmap() argument 149 static inline void flush_cache_vunmap(unsigned long start, unsigned long end) flush_cache_vunmap() argument
|
H A D | tlbflush.h | 45 * flush_tlb_range(mm,start,end) 50 * - start - start address (may not be aligned) 108 unsigned long start, unsigned long end, __flush_tlb_range() 114 if ((end - start) > MAX_TLB_RANGE) { __flush_tlb_range() 119 start = asid | (start >> 12); __flush_tlb_range() 123 for (addr = start; addr < end; addr += 1 << (PAGE_SHIFT - 12)) { __flush_tlb_range() 133 unsigned long start, unsigned long end) flush_tlb_range() 135 __flush_tlb_range(vma, start, end, false); flush_tlb_range() 138 static inline void flush_tlb_kernel_range(unsigned long start, unsigned long end) flush_tlb_kernel_range() argument 142 if ((end - start) > MAX_TLB_RANGE) { flush_tlb_kernel_range() 147 start >>= 12; flush_tlb_kernel_range() 151 for (addr = start; addr < end; addr += 1 << (PAGE_SHIFT - 12)) flush_tlb_kernel_range() 107 __flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end, bool last_level) __flush_tlb_range() argument 132 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_tlb_range() argument
|
/linux-4.4.14/arch/sh/boards/ |
H A D | board-sh2007.c | 34 .start = SMC0_BASE, 39 .start = evt2irq(0x240), 47 .start = SMC1_BASE, 52 .start = evt2irq(0x280), 80 .start = CF_BASE + CF_OFFSET, 85 .start = CF_BASE + CF_OFFSET + 0x206, 90 .start = evt2irq(0x2c0),
|