Searched refs:end (Results 1 - 200 of 6683) sorted by relevance

1234567891011>>

/linux-4.4.14/kernel/
H A Drange.c10 int add_range(struct range *range, int az, int nr_range, u64 start, u64 end) add_range() argument
12 if (start >= end) add_range()
20 range[nr_range].end = end; add_range()
28 u64 start, u64 end) add_range_with_merge()
32 if (start >= end) add_range_with_merge()
35 /* get new start/end: */ add_range_with_merge()
39 if (!range[i].end) add_range_with_merge()
43 common_end = min(range[i].end, end); add_range_with_merge()
47 /* new start/end, will add it back at last */ add_range_with_merge()
49 end = max(range[i].end, end); add_range_with_merge()
54 range[nr_range - 1].end = 0; add_range_with_merge()
60 return add_range(range, az, nr_range, start, end); add_range_with_merge()
63 void subtract_range(struct range *range, int az, u64 start, u64 end) subtract_range() argument
67 if (start >= end) subtract_range()
71 if (!range[j].end) subtract_range()
74 if (start <= range[j].start && end >= range[j].end) { subtract_range()
76 range[j].end = 0; subtract_range()
80 if (start <= range[j].start && end < range[j].end && subtract_range()
81 range[j].start < end) { subtract_range()
82 range[j].start = end; subtract_range()
87 if (start > range[j].start && end >= range[j].end && subtract_range()
88 range[j].end > start) { subtract_range()
89 range[j].end = start; subtract_range()
93 if (start > range[j].start && end < range[j].end) { subtract_range()
96 if (range[i].end == 0) subtract_range()
100 range[i].end = range[j].end; subtract_range()
101 range[i].start = end; subtract_range()
106 range[j].end = start; subtract_range()
129 if (range[i].end) clean_sort_range()
132 if (range[j].end) { clean_sort_range()
140 range[i].end = range[k].end; clean_sort_range()
142 range[k].end = 0; clean_sort_range()
147 if (!range[i].end) { clean_sort_range()
27 add_range_with_merge(struct range *range, int az, int nr_range, u64 start, u64 end) add_range_with_merge() argument
H A Dresource.c32 .end = IO_SPACE_LIMIT,
40 .end = -1,
108 int width = root->end < 0x10000 ? 4 : 8; r_show()
117 width, (unsigned long long) r->end, r_show()
211 resource_size_t end = new->end; __request_resource() local
214 if (end < start) __request_resource()
218 if (end > root->end) __request_resource()
223 if (!tmp || tmp->start > end) { __request_resource()
230 if (tmp->end < start) __request_resource()
274 tmp->end = size - 1; __release_child_resources()
336 * Finds the lowest iomem reosurce exists with-in [res->start.res->end)
337 * the caller must specify res->start, res->end, res->flags and "name".
345 resource_size_t start, end; find_next_iomem_res() local
352 end = res->end; find_next_iomem_res()
353 BUG_ON(start >= end); find_next_iomem_res()
365 if (p->start > end) { find_next_iomem_res()
369 if ((p->end >= start) && (p->start < end)) find_next_iomem_res()
379 if (res->end > p->end) find_next_iomem_res()
380 res->end = p->end; find_next_iomem_res()
387 * All the memory ranges which overlap start,end and also match flags and
393 * @end: end addr
395 int walk_iomem_res(char *name, unsigned long flags, u64 start, u64 end, walk_iomem_res() argument
403 res.end = end; walk_iomem_res()
405 orig_end = res.end; walk_iomem_res()
406 while ((res.start < res.end) && walk_iomem_res()
408 ret = (*func)(res.start, res.end, arg); walk_iomem_res()
411 res.start = res.end + 1; walk_iomem_res()
412 res.end = orig_end; walk_iomem_res()
424 int walk_system_ram_res(u64 start, u64 end, void *arg, walk_system_ram_res() argument
432 res.end = end; walk_system_ram_res()
434 orig_end = res.end; walk_system_ram_res()
435 while ((res.start < res.end) && walk_system_ram_res()
437 ret = (*func)(res.start, res.end, arg); walk_system_ram_res()
440 res.start = res.end + 1; walk_system_ram_res()
441 res.end = orig_end; walk_system_ram_res()
462 res.end = ((u64)(start_pfn + nr_pages) << PAGE_SHIFT) - 1; walk_system_ram_range()
464 orig_end = res.end; walk_system_ram_range()
465 while ((res.start < res.end) && walk_system_ram_range()
468 end_pfn = (res.end + 1) >> PAGE_SHIFT; walk_system_ram_range()
473 res.start = res.end + 1; walk_system_ram_range()
474 res.end = orig_end; walk_system_ram_range()
516 resource_size_t end = start + size - 1; region_intersects() local
524 if (start >= p->start && start <= p->end) region_intersects()
526 if (end >= p->start && end <= p->end) region_intersects()
528 if (p->start >= start && p->end <= end) region_intersects()
559 if (res->end > max) resource_clip()
560 res->end = max; resource_clip()
578 * of this->start - 1 to tmp->end below would cause an underflow. __find_resource()
581 tmp.start = (this == old) ? old->start : this->end + 1; __find_resource()
586 tmp.end = (this == old) ? this->end : this->start - 1; __find_resource()
588 tmp.end = root->end; __find_resource()
590 if (tmp.end < tmp.start) __find_resource()
598 avail.end = tmp.end; __find_resource()
604 alloc.end = alloc.start + size - 1; __find_resource()
607 new->end = alloc.end; __find_resource()
612 next: if (!this || this->end == root->end) __find_resource()
616 tmp.start = this->end + 1; __find_resource()
657 old->end = new.end; reallocate_resource()
668 old->end = new.end; reallocate_resource()
769 if ((first->start > new->start) || (first->end < new->end)) __insert_resource()
771 if ((first->start == new->start) && (first->end == new->end)) __insert_resource()
777 if (next->start < new->start || next->end > new->end) __insert_resource()
781 if (next->sibling->start > new->end) __insert_resource()
868 if (conflict->end > new->end) insert_resource_expand_to_fit()
869 new->end = conflict->end; insert_resource_expand_to_fit()
880 resource_size_t end = start + size - 1; __adjust_resource() local
886 if ((start < parent->start) || (end > parent->end)) __adjust_resource()
889 if (res->sibling && (res->sibling->start <= end)) __adjust_resource()
896 if (start <= tmp->end) __adjust_resource()
902 if ((tmp->start < start) || (tmp->end > end)) __adjust_resource()
906 res->end = end; __adjust_resource()
936 resource_size_t start, resource_size_t end, __reserve_region_with_split()
949 res->end = end; __reserve_region_with_split()
965 conflict->end >= res->end) { __reserve_region_with_split()
973 end = res->end; __reserve_region_with_split()
974 res->end = conflict->start - 1; __reserve_region_with_split()
975 if (conflict->end < end) { __reserve_region_with_split()
982 next_res->start = conflict->end + 1; __reserve_region_with_split()
983 next_res->end = end; __reserve_region_with_split()
987 res->start = conflict->end + 1; __reserve_region_with_split()
994 resource_size_t start, resource_size_t end, reserve_region_with_split()
1000 if (root->start > start || root->end < end) { reserve_region_with_split()
1002 (unsigned long long)start, (unsigned long long)end, reserve_region_with_split()
1004 if (start > root->end || end < root->start) reserve_region_with_split()
1007 if (end > root->end) reserve_region_with_split()
1008 end = root->end; reserve_region_with_split()
1013 (unsigned long long)end); reserve_region_with_split()
1018 __reserve_region_with_split(root, start, end, name); reserve_region_with_split()
1073 res->end = start + n - 1; __request_region()
1122 resource_size_t end; __release_region() local
1125 end = start + n - 1; __release_region()
1134 if (res->start <= start && res->end >= end) { __release_region()
1139 if (res->start != start || res->end != end) __release_region()
1155 (unsigned long long)end); __release_region()
1186 resource_size_t end; release_mem_region_adjustable() local
1189 end = start + size - 1; release_mem_region_adjustable()
1190 if ((start < parent->start) || (end > parent->end)) release_mem_region_adjustable()
1200 if (res->start >= end) release_mem_region_adjustable()
1204 if (res->start > start || res->end < end) { release_mem_region_adjustable()
1218 if (res->start == start && res->end == end) { release_mem_region_adjustable()
1223 } else if (res->start == start && res->end != end) { release_mem_region_adjustable()
1225 ret = __adjust_resource(res, end + 1, release_mem_region_adjustable()
1226 res->end - end); release_mem_region_adjustable()
1227 } else if (res->start != start && res->end == end) { release_mem_region_adjustable()
1228 /* adjust the end */ release_mem_region_adjustable()
1238 new_res->start = end + 1; release_mem_region_adjustable()
1239 new_res->end = res->end; release_mem_region_adjustable()
1414 res->end = io_start + io_num - 1; reserve_setup()
1444 if (p->end < addr) iomem_map_sanity_check()
1447 PFN_DOWN(p->end) >= PFN_DOWN(addr + size - 1)) iomem_map_sanity_check()
1500 if (p->end < addr) iomem_is_exclusive()
935 __reserve_region_with_split(struct resource *root, resource_size_t start, resource_size_t end, const char *name) __reserve_region_with_split() argument
993 reserve_region_with_split(struct resource *root, resource_size_t start, resource_size_t end, const char *name) reserve_region_with_split() argument
H A Dcrash_dump.c36 char *end; setup_elfcorehdr() local
39 elfcorehdr_addr = memparse(arg, &end); setup_elfcorehdr()
40 if (*end == '@') { setup_elfcorehdr()
42 elfcorehdr_addr = memparse(end + 1, &end); setup_elfcorehdr()
44 return end > arg ? 0 : -EINVAL; setup_elfcorehdr()
/linux-4.4.14/include/linux/ceph/
H A Ddecode.h14 * void *end pointer to end of buffer (last byte + 1)
50 static inline int ceph_has_room(void **p, void *end, size_t n) ceph_has_room() argument
52 return end >= *p && n <= end - *p; ceph_has_room()
55 #define ceph_decode_need(p, end, n, bad) \
57 if (!likely(ceph_has_room(p, end, n))) \
61 #define ceph_decode_64_safe(p, end, v, bad) \
63 ceph_decode_need(p, end, sizeof(u64), bad); \
66 #define ceph_decode_32_safe(p, end, v, bad) \
68 ceph_decode_need(p, end, sizeof(u32), bad); \
71 #define ceph_decode_16_safe(p, end, v, bad) \
73 ceph_decode_need(p, end, sizeof(u16), bad); \
76 #define ceph_decode_8_safe(p, end, v, bad) \
78 ceph_decode_need(p, end, sizeof(u8), bad); \
82 #define ceph_decode_copy_safe(p, end, pv, n, bad) \
84 ceph_decode_need(p, end, n, bad); \
102 * beyond the "end" pointer provided (-ERANGE)
105 static inline char *ceph_extract_encoded_string(void **p, void *end, ceph_extract_encoded_string() argument
112 ceph_decode_32_safe(&sp, end, len, bad); ceph_extract_encoded_string()
113 if (!ceph_has_room(&sp, end, len)) ceph_extract_encoded_string()
197 static inline void ceph_encode_filepath(void **p, void *end, ceph_encode_filepath() argument
201 BUG_ON(*p + 1 + sizeof(ino) + sizeof(len) + len > end); ceph_encode_filepath()
210 static inline void ceph_encode_string(void **p, void *end, ceph_encode_string() argument
213 BUG_ON(*p + sizeof(len) + len > end); ceph_encode_string()
220 #define ceph_encode_need(p, end, n, bad) \
222 if (!likely(ceph_has_room(p, end, n))) \
226 #define ceph_encode_64_safe(p, end, v, bad) \
228 ceph_encode_need(p, end, sizeof(u64), bad); \
231 #define ceph_encode_32_safe(p, end, v, bad) \
233 ceph_encode_need(p, end, sizeof(u32), bad); \
236 #define ceph_encode_16_safe(p, end, v, bad) \
238 ceph_encode_need(p, end, sizeof(u16), bad); \
241 #define ceph_encode_8_safe(p, end, v, bad) \
243 ceph_encode_need(p, end, sizeof(u8), bad); \
247 #define ceph_encode_copy_safe(p, end, pv, n, bad) \
249 ceph_encode_need(p, end, n, bad); \
252 #define ceph_encode_string_safe(p, end, s, n, bad) \
254 ceph_encode_need(p, end, n, bad); \
255 ceph_encode_string(p, end, s, n); \
/linux-4.4.14/include/linux/
H A Dcirc_buf.h22 /* Return count up to the end of the buffer. Carefully avoid
26 ({int end = (size) - (tail); \
27 int n = ((head) + end) & ((size)-1); \
28 n < end ? n : end;})
30 /* Return space available up to the end of the buffer. */
32 ({int end = (size) - 1 - (head); \
33 int n = (end + (tail)) & ((size)-1); \
34 n <= end ? n : end+1;})
H A Drange.h6 u64 end; member in struct:range
10 u64 start, u64 end);
14 u64 start, u64 end);
16 void subtract_range(struct range *range, int az, u64 start, u64 end);
H A Dfirmware-map.h26 int firmware_map_add_early(u64 start, u64 end, const char *type);
27 int firmware_map_add_hotplug(u64 start, u64 end, const char *type);
28 int firmware_map_remove(u64 start, u64 end, const char *type);
32 static inline int firmware_map_add_early(u64 start, u64 end, const char *type) firmware_map_add_early() argument
37 static inline int firmware_map_add_hotplug(u64 start, u64 end, const char *type) firmware_map_add_hotplug() argument
42 static inline int firmware_map_remove(u64 start, u64 end, const char *type) firmware_map_remove() argument
/linux-4.4.14/arch/x86/kernel/
H A Dresource.c5 resource_size_t end) resource_clip()
9 if (res->end < start || res->start > end) resource_clip()
15 if (res->end > end) resource_clip()
16 high = res->end - end; resource_clip()
20 res->end = start - 1; resource_clip()
22 res->start = end + 1; resource_clip()
4 resource_clip(struct resource *res, resource_size_t start, resource_size_t end) resource_clip() argument
H A Dcheck.c58 char *end; set_corruption_check_size() local
61 size = memparse(arg, &end); set_corruption_check_size()
63 if (*end == '\0') set_corruption_check_size()
73 phys_addr_t start, end; setup_bios_corruption_check() local
94 for_each_free_mem_range(i, NUMA_NO_NODE, MEMBLOCK_NONE, &start, &end, setup_bios_corruption_check()
98 end = clamp_t(phys_addr_t, round_down(end, PAGE_SIZE), setup_bios_corruption_check()
100 if (start >= end) setup_bios_corruption_check()
103 memblock_reserve(start, end - start); setup_bios_corruption_check()
105 scan_areas[num_scan_areas].size = end - start; setup_bios_corruption_check()
108 memset(__va(start), 0, end - start); setup_bios_corruption_check()
H A Dprobe_roms.c27 .end = 0xfffff,
34 .end = 0xeffff,
41 .end = 0,
46 .end = 0,
51 .end = 0,
56 .end = 0,
61 .end = 0,
66 .end = 0,
73 .end = 0xc7fff,
123 if (res->end == 0) find_oprom()
221 video_rom_resource.end = start + length - 1; probe_roms()
227 start = (video_rom_resource.end + 1 + 2047) & ~2047UL; probe_roms()
262 adapter_rom_resources[i].end = start + length - 1; probe_roms()
265 start = adapter_rom_resources[i++].end & ~2047UL; probe_roms()
/linux-4.4.14/arch/blackfin/include/asm/
H A Dcacheflush.h28 #define flush_cache_range(vma, start, end) do { } while (0)
30 #define flush_cache_vmap(start, end) do { } while (0)
31 #define flush_cache_vunmap(start, end) do { } while (0)
34 #define flush_icache_range_others(start, end) \
35 smp_icache_flush_range_others((start), (end))
37 #define flush_icache_range_others(start, end) do { } while (0)
40 static inline void flush_icache_range(unsigned start, unsigned end) flush_icache_range() argument
43 if (end <= physical_mem_end) flush_icache_range()
44 blackfin_dcache_flush_range(start, end); flush_icache_range()
47 if (start >= L2_START && end <= L2_START + L2_LENGTH) flush_icache_range()
48 blackfin_dcache_flush_range(start, end); flush_icache_range()
61 if (end <= physical_mem_end) { flush_icache_range()
62 blackfin_icache_flush_range(start, end); flush_icache_range()
63 flush_icache_range_others(start, end); flush_icache_range()
67 if (start >= L2_START && end <= L2_START + L2_LENGTH) { flush_icache_range()
68 blackfin_icache_flush_range(start, end); flush_icache_range()
69 flush_icache_range_others(start, end); flush_icache_range()
82 # define invalidate_dcache_range(start,end) blackfin_dcache_invalidate_range((start), (end))
84 # define invalidate_dcache_range(start,end) do { } while (0)
87 # define flush_dcache_range(start,end) blackfin_dcache_flush_range((start), (end))
91 # define flush_dcache_range(start,end) do { } while (0)
/linux-4.4.14/drivers/pci/
H A Dhotplug-pci.c11 int end = parent->busn_res.end; pci_hp_add_bridge() local
13 for (busnr = start; busnr <= end; busnr++) { pci_hp_add_bridge()
17 if (busnr-- > end) { pci_hp_add_bridge()
H A Dhost-bridge.c63 region->end = res->end - offset; pcibios_resource_to_bus()
70 return region1->start <= region2->start && region1->end >= region2->end; region_contains()
87 bus_region.end = window->res->end - window->offset; pcibios_bus_to_resource()
96 res->end = region->end + offset; pcibios_bus_to_resource()
/linux-4.4.14/arch/x86/include/asm/
H A Dpat.h11 extern int reserve_memtype(u64 start, u64 end,
13 extern int free_memtype(u64 start, u64 end);
18 int io_reserve_memtype(resource_size_t start, resource_size_t end,
21 void io_free_memtype(resource_size_t start, resource_size_t end);
/linux-4.4.14/arch/powerpc/boot/
H A Dstdlib.h4 unsigned long long int strtoull(const char *ptr, char **end, int base);
H A Dstdlib.c16 unsigned long long int strtoull(const char *ptr, char **end, int base) strtoull() argument
41 if (end) strtoull()
42 *end = (char *)ptr; strtoull()
/linux-4.4.14/arch/sparc/include/asm/
H A Dtlbflush_32.h10 #define flush_tlb_range(vma, start, end) \
11 sparc32_cachetlb_ops->tlb_range(vma, start, end)
19 unsigned long end) flush_tlb_kernel_range()
18 flush_tlb_kernel_range(unsigned long start, unsigned long end) flush_tlb_kernel_range() argument
H A Dcacheflush_32.h12 #define flush_cache_range(vma,start,end) \
13 sparc32_cachetlb_ops->cache_range(vma, start, end)
16 #define flush_icache_range(start, end) do { } while (0)
46 #define flush_cache_vmap(start, end) flush_cache_all()
47 #define flush_cache_vunmap(start, end) flush_cache_all()
H A Dtlbflush_64.h18 void flush_tsb_kernel_range(unsigned long start, unsigned long end);
34 unsigned long start, unsigned long end) flush_tlb_range()
38 void flush_tlb_kernel_range(unsigned long start, unsigned long end);
50 void __flush_tlb_kernel_range(unsigned long start, unsigned long end);
61 void smp_flush_tlb_kernel_range(unsigned long start, unsigned long end);
33 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_tlb_range() argument
/linux-4.4.14/tools/perf/
H A Dperf-read-vdso.c14 void *start, *end; main() local
17 if (find_vdso_map(&start, &end)) main()
20 size = end - start; main()
/linux-4.4.14/arch/s390/lib/
H A Ddelay.c23 * To end the bloody studid and useless discussion about the __delay()
37 u64 end; __udelay_disabled() local
39 end = get_tod_clock() + (usecs << 12); __udelay_disabled()
45 set_clock_comparator(end); __udelay_disabled()
55 u64 clock_saved, end; __udelay_enabled() local
57 end = get_tod_clock_fast() + (usecs << 12); __udelay_enabled()
60 if (end < S390_lowcore.clock_comparator) { __udelay_enabled()
62 set_clock_comparator(end); __udelay_enabled()
67 } while (get_tod_clock_fast() < end); __udelay_enabled()
109 u64 end; udelay_simple() local
111 end = get_tod_clock_fast() + (usecs << 12); udelay_simple()
112 while (get_tod_clock_fast() < end) udelay_simple()
118 u64 end; __ndelay() local
122 end = get_tod_clock_fast() + nsecs; __ndelay()
125 while (get_tod_clock_fast() < end) __ndelay()
/linux-4.4.14/arch/alpha/boot/
H A Dhead.S17 .end __start
26 .end wrent
35 .end wrkgp
85 .end switch_to_osf_pal
94 .end tbi
102 .end halt
123 .end move_stack
/linux-4.4.14/arch/avr32/mm/
H A Dcache.c26 unsigned long v, begin, end, linesz, mask; invalidate_dcache_region() local
35 end = begin + size; invalidate_dcache_region()
41 if (end & mask) { invalidate_dcache_region()
42 flush_dcache_line((void *)end); invalidate_dcache_region()
43 end &= ~mask; invalidate_dcache_region()
47 for (v = begin; v < end; v += linesz) invalidate_dcache_region()
54 unsigned long v, begin, end, linesz; clean_dcache_region() local
58 end = ((unsigned long)start + size + linesz - 1) & ~(linesz - 1); clean_dcache_region()
60 for (v = begin; v < end; v += linesz) clean_dcache_region()
67 unsigned long v, begin, end, linesz; flush_dcache_region() local
71 end = ((unsigned long)start + size + linesz - 1) & ~(linesz - 1); flush_dcache_region()
73 for (v = begin; v < end; v += linesz) flush_dcache_region()
80 unsigned long v, begin, end, linesz; invalidate_icache_region() local
84 end = ((unsigned long)start + size + linesz - 1) & ~(linesz - 1); invalidate_icache_region()
86 for (v = begin; v < end; v += linesz) invalidate_icache_region()
90 static inline void __flush_icache_range(unsigned long start, unsigned long end) __flush_icache_range() argument
95 for (v = start; v < end; v += linesz) { __flush_icache_range()
106 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument
112 (end + linesz - 1) & ~(linesz - 1)); flush_icache_range()
/linux-4.4.14/arch/mips/pci/
H A Dpci-malta.c43 .end = 0x000fffffUL,
93 resource_size_t start, end, map, start1, end1, map1, map2, map3, mask; mips_pcibios_init() local
116 end = GT_READ(GT_PCI0M0HD_OFS); mips_pcibios_init()
118 end = (end & GT_PCI_HD_MSK) | (start & ~GT_PCI_HD_MSK); mips_pcibios_init()
124 if (end1 - start1 > end - start) { mips_pcibios_init()
126 end = end1; mips_pcibios_init()
129 mask = ~(start ^ end); mips_pcibios_init()
134 gt64120_mem_resource.end = end; mips_pcibios_init()
138 gt64120_mem_resource.end <<= GT_PCI_DCRM_SHF; mips_pcibios_init()
139 gt64120_mem_resource.end |= (1 << GT_PCI_DCRM_SHF) - 1; mips_pcibios_init()
143 end = GT_READ(GT_PCI0IOHD_OFS); mips_pcibios_init()
145 end = (end & GT_PCI_HD_MSK) | (start & ~GT_PCI_HD_MSK); mips_pcibios_init()
146 mask = ~(start ^ end); mips_pcibios_init()
151 gt64120_io_resource.end = (map & mask) | ~mask; mips_pcibios_init()
155 gt64120_io_resource.end <<= GT_PCI_DCRM_SHF; mips_pcibios_init()
156 gt64120_io_resource.end |= (1 << GT_PCI_DCRM_SHF) - 1; mips_pcibios_init()
173 end = 1; mips_pcibios_init()
177 end++; mips_pcibios_init()
182 end++; mips_pcibios_init()
185 bonito64_mem_resource.end = start + mips_pcibios_init()
186 BONITO_PCIMAP_WINBASE(end) - 1; mips_pcibios_init()
202 msc_mem_resource.end = (start & mask) | ~mask; mips_pcibios_init()
213 msc_io_resource.end = (map & mask) | ~mask; mips_pcibios_init()
215 ioport_resource.end = ~mask; mips_pcibios_init()
223 end = start | ~mask; mips_pcibios_init()
225 start <= msc_mem_resource.end) || mips_pcibios_init()
226 (end >= msc_mem_resource.start && mips_pcibios_init()
227 end <= msc_mem_resource.end)) { mips_pcibios_init()
230 end = min(end, msc_mem_resource.end); mips_pcibios_init()
232 msc_mem_resource.end - end) mips_pcibios_init()
233 msc_mem_resource.end = start - 1; mips_pcibios_init()
235 msc_mem_resource.start = end + 1; mips_pcibios_init()
248 iomem_resource.end &= 0xfffffffffULL; /* 64 GB */ mips_pcibios_init()
249 ioport_resource.end = controller->io_resource->end; mips_pcibios_init()
/linux-4.4.14/arch/sh/include/asm/
H A Dtlbflush.h10 * - flush_tlb_range(vma, start, end) flushes a range of pages
11 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages
17 unsigned long end);
21 unsigned long end);
31 unsigned long end);
33 extern void flush_tlb_kernel_range(unsigned long start, unsigned long end);
43 #define flush_tlb_range(vma, start, end) \
44 local_flush_tlb_range(vma, start, end)
46 #define flush_tlb_kernel_range(start, end) \
47 local_flush_tlb_kernel_range(start, end)
H A Dtlb.h24 unsigned long start, end; member in struct:mmu_gather
30 tlb->end = 0; init_tlb_gather()
34 tlb->end = TASK_SIZE; init_tlb_gather()
39 tlb_gather_mmu(struct mmu_gather *tlb, struct mm_struct *mm, unsigned long start, unsigned long end) tlb_gather_mmu() argument
43 tlb->end = end; tlb_gather_mmu()
44 tlb->fullmm = !(start | (end+1)); tlb_gather_mmu()
50 tlb_finish_mmu(struct mmu_gather *tlb, unsigned long start, unsigned long end) tlb_finish_mmu() argument
64 if (tlb->end < address + PAGE_SIZE) tlb_remove_tlb_entry()
65 tlb->end = address + PAGE_SIZE; tlb_remove_tlb_entry()
83 if (!tlb->fullmm && tlb->end) { tlb_end_vma()
84 flush_tlb_range(vma, tlb->start, tlb->end); tlb_end_vma()
/linux-4.4.14/fs/btrfs/tests/
H A Dextent-io-tests.c28 static noinline int process_page_range(struct inode *inode, u64 start, u64 end, process_page_range() argument
34 unsigned long end_index = end >> PAGE_CACHE_SHIFT; process_page_range()
59 printk(KERN_ERR "stuck in a loop, start %Lu, end %Lu, nr_pages %lu, ret %d\n", start, end, nr_pages, ret); process_page_range()
75 u64 start, end, test_start; test_find_delalloc() local
114 end = 0; test_find_delalloc()
116 &end, max_bytes); test_find_delalloc()
121 if (start != 0 || end != 4095) { test_find_delalloc()
122 test_msg("Expected start 0 end 4095, got start %Lu end %Lu\n", test_find_delalloc()
123 start, end); test_find_delalloc()
126 unlock_extent(&tmp, start, end); test_find_delalloc()
145 end = 0; test_find_delalloc()
147 &end, max_bytes); test_find_delalloc()
152 if (start != test_start || end != max_bytes - 1) { test_find_delalloc()
153 test_msg("Expected start %Lu end %Lu, got start %Lu, end " test_find_delalloc()
154 "%Lu\n", test_start, max_bytes - 1, start, end); test_find_delalloc()
157 if (process_page_range(inode, start, end, test_find_delalloc()
162 unlock_extent(&tmp, start, end); test_find_delalloc()
179 end = 0; test_find_delalloc()
181 &end, max_bytes); test_find_delalloc()
186 if (end != (u64)-1) { test_find_delalloc()
187 test_msg("Did not return the proper end offset\n"); test_find_delalloc()
200 end = 0; test_find_delalloc()
202 &end, max_bytes); test_find_delalloc()
207 if (start != test_start || end != total_dirty - 1) { test_find_delalloc()
208 test_msg("Expected start %Lu end %Lu, got start %Lu end %Lu\n", test_find_delalloc()
209 test_start, total_dirty - 1, start, end); test_find_delalloc()
212 if (process_page_range(inode, start, end, test_find_delalloc()
217 unlock_extent(&tmp, start, end); test_find_delalloc()
235 end = 0; test_find_delalloc()
243 &end, max_bytes); test_find_delalloc()
248 if (start != test_start && end != test_start + PAGE_CACHE_SIZE - 1) { test_find_delalloc()
249 test_msg("Expected start %Lu end %Lu, got start %Lu end %Lu\n", test_find_delalloc()
251 end); test_find_delalloc()
254 if (process_page_range(inode, start, end, PROCESS_TEST_LOCKED | test_find_delalloc()
/linux-4.4.14/arch/microblaze/kernel/cpu/
H A Dcache.c92 #define CACHE_LOOP_LIMITS(start, end, cache_line_length, cache_size) \
95 end = min(start + cache_size, end); \
120 * end address is not aligned, if end is aligned then I have to subtract
124 #define CACHE_RANGE_LOOP_2(start, end, line_length, op) \
129 end = ((end & align) == end) ? end - line_length : end & align; \
130 count = end - start; \
141 #define CACHE_RANGE_LOOP_1(start, end, line_length, op) \
145 end = ((end & align) == end) ? end - line_length : end & align; \
146 WARN_ON(end < start); \
152 : : "r" (temp), "r" (start), "r" (end), \
158 static void __flush_icache_range_msr_irq(unsigned long start, unsigned long end) __flush_icache_range_msr_irq() argument
164 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __flush_icache_range_msr_irq()
165 (unsigned int)start, (unsigned int) end); __flush_icache_range_msr_irq()
167 CACHE_LOOP_LIMITS(start, end, __flush_icache_range_msr_irq()
174 CACHE_RANGE_LOOP_1(start, end, cpuinfo.icache_line_length, wic); __flush_icache_range_msr_irq()
176 for (i = start; i < end; i += cpuinfo.icache_line_length) __flush_icache_range_msr_irq()
185 unsigned long end) __flush_icache_range_nomsr_irq()
191 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __flush_icache_range_nomsr_irq()
192 (unsigned int)start, (unsigned int) end); __flush_icache_range_nomsr_irq()
194 CACHE_LOOP_LIMITS(start, end, __flush_icache_range_nomsr_irq()
201 CACHE_RANGE_LOOP_1(start, end, cpuinfo.icache_line_length, wic); __flush_icache_range_nomsr_irq()
203 for (i = start; i < end; i += cpuinfo.icache_line_length) __flush_icache_range_nomsr_irq()
213 unsigned long end) __flush_icache_range_noirq()
218 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __flush_icache_range_noirq()
219 (unsigned int)start, (unsigned int) end); __flush_icache_range_noirq()
221 CACHE_LOOP_LIMITS(start, end, __flush_icache_range_noirq()
224 CACHE_RANGE_LOOP_1(start, end, cpuinfo.icache_line_length, wic); __flush_icache_range_noirq()
226 for (i = start; i < end; i += cpuinfo.icache_line_length) __flush_icache_range_noirq()
377 unsigned long end) __invalidate_dcache_range_wb()
382 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __invalidate_dcache_range_wb()
383 (unsigned int)start, (unsigned int) end); __invalidate_dcache_range_wb()
385 CACHE_LOOP_LIMITS(start, end, __invalidate_dcache_range_wb()
388 CACHE_RANGE_LOOP_2(start, end, cpuinfo.dcache_line_length, wdc.clear); __invalidate_dcache_range_wb()
390 for (i = start; i < end; i += cpuinfo.dcache_line_length) __invalidate_dcache_range_wb()
397 unsigned long end) __invalidate_dcache_range_nomsr_wt()
402 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __invalidate_dcache_range_nomsr_wt()
403 (unsigned int)start, (unsigned int) end); __invalidate_dcache_range_nomsr_wt()
404 CACHE_LOOP_LIMITS(start, end, __invalidate_dcache_range_nomsr_wt()
408 CACHE_RANGE_LOOP_1(start, end, cpuinfo.dcache_line_length, wdc); __invalidate_dcache_range_nomsr_wt()
410 for (i = start; i < end; i += cpuinfo.dcache_line_length) __invalidate_dcache_range_nomsr_wt()
417 unsigned long end) __invalidate_dcache_range_msr_irq_wt()
423 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __invalidate_dcache_range_msr_irq_wt()
424 (unsigned int)start, (unsigned int) end); __invalidate_dcache_range_msr_irq_wt()
425 CACHE_LOOP_LIMITS(start, end, __invalidate_dcache_range_msr_irq_wt()
432 CACHE_RANGE_LOOP_1(start, end, cpuinfo.dcache_line_length, wdc); __invalidate_dcache_range_msr_irq_wt()
434 for (i = start; i < end; i += cpuinfo.dcache_line_length) __invalidate_dcache_range_msr_irq_wt()
444 unsigned long end) __invalidate_dcache_range_nomsr_irq()
450 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __invalidate_dcache_range_nomsr_irq()
451 (unsigned int)start, (unsigned int) end); __invalidate_dcache_range_nomsr_irq()
453 CACHE_LOOP_LIMITS(start, end, __invalidate_dcache_range_nomsr_irq()
460 CACHE_RANGE_LOOP_1(start, end, cpuinfo.dcache_line_length, wdc); __invalidate_dcache_range_nomsr_irq()
462 for (i = start; i < end; i += cpuinfo.dcache_line_length) __invalidate_dcache_range_nomsr_irq()
488 static void __flush_dcache_range_wb(unsigned long start, unsigned long end) __flush_dcache_range_wb() argument
493 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __flush_dcache_range_wb()
494 (unsigned int)start, (unsigned int) end); __flush_dcache_range_wb()
496 CACHE_LOOP_LIMITS(start, end, __flush_dcache_range_wb()
499 CACHE_RANGE_LOOP_2(start, end, cpuinfo.dcache_line_length, wdc.flush); __flush_dcache_range_wb()
501 for (i = start; i < end; i += cpuinfo.dcache_line_length) __flush_dcache_range_wb()
184 __flush_icache_range_nomsr_irq(unsigned long start, unsigned long end) __flush_icache_range_nomsr_irq() argument
212 __flush_icache_range_noirq(unsigned long start, unsigned long end) __flush_icache_range_noirq() argument
376 __invalidate_dcache_range_wb(unsigned long start, unsigned long end) __invalidate_dcache_range_wb() argument
396 __invalidate_dcache_range_nomsr_wt(unsigned long start, unsigned long end) __invalidate_dcache_range_nomsr_wt() argument
416 __invalidate_dcache_range_msr_irq_wt(unsigned long start, unsigned long end) __invalidate_dcache_range_msr_irq_wt() argument
443 __invalidate_dcache_range_nomsr_irq(unsigned long start, unsigned long end) __invalidate_dcache_range_nomsr_irq() argument
/linux-4.4.14/arch/powerpc/include/asm/
H A Dsections.h26 unsigned long end) overlaps_interrupt_vector_text()
33 (unsigned long)__va(real_start) < end; overlaps_interrupt_vector_text()
36 static inline int overlaps_kernel_text(unsigned long start, unsigned long end) overlaps_kernel_text() argument
39 (unsigned long)_stext < end; overlaps_kernel_text()
42 static inline int overlaps_kvm_tmp(unsigned long start, unsigned long end) overlaps_kvm_tmp() argument
47 (unsigned long)&kvm_tmp[1024 * 1024] < end; overlaps_kvm_tmp()
25 overlaps_interrupt_vector_text(unsigned long start, unsigned long end) overlaps_interrupt_vector_text() argument
/linux-4.4.14/arch/mips/jz4740/
H A Dplatform.c39 .end = JZ4740_UHC_BASE_ADDR + 0x1000 - 1,
44 .end = JZ4740_IRQ_UHC,
69 .end = JZ4740_UDC_BASE_ADDR + 0x10000 - 1,
74 .end = JZ4740_IRQ_UDC,
95 .end = JZ4740_MSC_BASE_ADDR + 0x1000 - 1,
100 .end = JZ4740_IRQ_MSC,
120 .end = JZ4740_RTC_BASE_ADDR + 0x38 - 1,
125 .end = JZ4740_IRQ_RTC,
141 .end = JZ4740_I2C_BASE_ADDR + 0x1000 - 1,
146 .end = JZ4740_IRQ_I2C,
163 .end = JZ4740_EMC_BASE_ADDR + 0x1000 - 1,
169 .end = 0x180C0000 - 1,
175 .end = 0x140C0000 - 1,
181 .end = 0x0C0C0000 - 1,
187 .end = 0x080C0000 - 1,
202 .end = JZ4740_LCD_BASE_ADDR + 0x1000 - 1,
222 .end = JZ4740_AIC_BASE_ADDR + 0x38 - 1,
244 .end = JZ4740_AIC_BASE_ADDR + 0x88 - 1,
260 .end = JZ4740_SADC_BASE_ADDR + 0x30,
265 .end = JZ4740_IRQ_SADC,
270 .end = JZ4740_IRQ_ADC_BASE,
286 .end = JZ4740_WDT_BASE_ADDR + 0x10 - 1,
308 .end = JZ4740_DMAC_BASE_ADDR + 0x400 - 1,
313 .end = JZ4740_IRQ_DMAC,
/linux-4.4.14/arch/arm/mach-pxa/
H A Ddevices.c39 .end = IRQ_PMU,
53 .end = 0x41100fff,
58 .end = IRQ_MMC,
63 .end = 21,
68 .end = 22,
104 .end = 0x4060ffff,
109 .end = IRQ_USB,
142 .end = 0x54100fff,
147 .end = IRQ_USB2,
168 .end = 0x4400ffff,
173 .end = IRQ_LCD,
200 .end = 0x40100023,
204 .end = IRQ_FFUART,
224 .end = 0x40200023,
228 .end = IRQ_BTUART,
248 .end = 0x40700023,
252 .end = IRQ_STUART,
272 .end = 0x4160002F,
276 .end = IRQ_HWUART,
299 .end = 0x403016a3,
303 .end = IRQ_I2C,
324 .end = 0x40f001a3,
328 .end = IRQ_PWRI2C,
344 .end = 0x40400083,
348 .end = IRQ_I2S,
390 .end = IRQ_STUART,
395 .end = IRQ_ICP,
400 .end = 0x4080001b,
405 .end = 0x40700023,
410 .end = 17,
415 .end = 18,
439 .end = 0x40900000 + 0x3b,
444 .end = IRQ_RTC1Hz,
450 .end = IRQ_RTCAlrm,
473 .end = 0x40500000 + 0xfff,
478 .end = IRQ_AC97,
506 .end = 0x40b0000f,
521 .end = 0x40c0000f,
538 .end = 0x4100001f,
543 .end = IRQ_SSP,
549 .end = 13,
555 .end = 14,
576 .end = 0x4140002f,
581 .end = IRQ_NSSP,
587 .end = 15,
593 .end = 16,
614 .end = 0x4150002f,
619 .end = IRQ_ASSP,
625 .end = 23,
631 .end = 24,
653 .end = 0x50000fff,
658 .end = IRQ_CAMERA,
686 .end = 0x4C00ff6f,
691 .end = IRQ_USBH1,
717 .end = 0x4150004c,
722 .end = IRQ_KEYPAD,
744 .end = 0x4100003f,
749 .end = IRQ_SSP,
755 .end = 13,
761 .end = 14,
782 .end = 0x4170003f,
787 .end = IRQ_SSP2,
793 .end = 15,
799 .end = 16,
820 .end = 0x4190003f,
825 .end = IRQ_SSP3,
831 .end = 66,
837 .end = 67,
856 .end = 0x40b0001f,
871 .end = 0x40c0001f,
888 .end = 0x42000fff,
893 .end = IRQ_MMC2,
898 .end = 93,
903 .end = 94,
927 .end = 0x42500fff,
932 .end = IRQ_MMC3,
937 .end = 100,
942 .end = 101,
966 .end = 0x54000fff,
971 .end = IRQ_GCU,
995 .end = 0x40f500d3,
999 .end = IRQ_PWRI2C,
1014 .end = 0x43100053,
1019 .end = IRQ_NAND,
1025 .end = 97,
1031 .end = 99,
1059 .end = 0x41a0003f,
1064 .end = IRQ_SSP4,
1070 .end = 2,
1076 .end = 3,
1135 .end = 0x40e0ffff,
1139 .end = IRQ_GPIO0,
1144 .end = IRQ_GPIO1,
1149 .end = IRQ_GPIO_2_x,
1211 .end = 0x4000ffff,
1216 .end = IRQ_DMA,
/linux-4.4.14/arch/x86/pci/
H A Damd_bus.c71 u64 end; early_root_info_init() local
159 end = (reg & 0xfff000) | 0xfff; early_root_info_init()
166 node, link, start, end); early_root_info_init()
169 if (end > 0xffff) early_root_info_init()
170 end = 0xffff; early_root_info_init()
171 update_res(info, start, end, IORESOURCE_IO, 1); early_root_info_init()
172 subtract_range(range, RANGE_NUM, start, end + 1); early_root_info_init()
179 if (!range[i].end) early_root_info_init()
182 update_res(info, range[i].start, range[i].end - 1, early_root_info_init()
189 end = cap_resource((0xfdULL<<32) - 1); early_root_info_init()
190 end++; early_root_info_init()
191 add_range(range, RANGE_NUM, 0, 0, end); early_root_info_init()
196 end = (val & 0xffffff800000ULL); early_root_info_init()
197 printk(KERN_INFO "TOM: %016llx aka %lldM\n", end, end>>20); early_root_info_init()
198 if (end < (1ULL<<32)) early_root_info_init()
199 subtract_range(range, RANGE_NUM, 0, end); early_root_info_init()
207 fam10h_mmconf_end = fam10h_mmconf->end; early_root_info_init()
226 end = (reg & 0xffffff00); early_root_info_init()
227 end <<= 8; early_root_info_init()
228 end |= 0xffff; early_root_info_init()
236 node, link, start, end); early_root_info_init()
239 * mmconf range, so need to update start and end. early_root_info_init()
250 if (end >= fam10h_mmconf_start && early_root_info_init()
251 end <= fam10h_mmconf_end) { early_root_info_init()
252 end = fam10h_mmconf_start - 1; early_root_info_init()
257 end > fam10h_mmconf_end) { early_root_info_init()
268 if (start <= end) { early_root_info_init()
269 printk(KERN_CONT " %s [%llx, %llx]", endx ? "and" : "==>", start, end); early_root_info_init()
277 update_res(info, cap_resource(start), cap_resource(end), early_root_info_init()
279 subtract_range(range, RANGE_NUM, start, end + 1); early_root_info_init()
292 end = (val & 0xffffff800000ULL); early_root_info_init()
293 printk(KERN_INFO "TOM2: %016llx aka %lldM\n", end, end>>20); early_root_info_init()
294 subtract_range(range, RANGE_NUM, 1ULL<<32, end); early_root_info_init()
304 if (!range[i].end) early_root_info_init()
308 cap_resource(range[i].end - 1), early_root_info_init()
H A Dbus_numa.c84 info->busn.end = bus_max; alloc_pci_root_info()
95 resource_size_t end, unsigned long flags, int merge) update_res()
100 if (start > end) update_res()
119 common_end = min(res->end, end); update_res()
124 final_end = max(res->end, end); update_res()
127 res->end = final_end; update_res()
142 res->end = end; update_res()
94 update_res(struct pci_root_info *info, resource_size_t start, resource_size_t end, unsigned long flags, int merge) update_res() argument
/linux-4.4.14/arch/blackfin/mach-bf561/boards/
H A Dtepla.c25 .end = 0x2C000320,
29 .end = IRQ_PROG_INTB,
33 .end = IRQ_PF7,
50 .end = BFIN_UART_GCTL+2,
55 .end = IRQ_UART_TX,
60 .end = IRQ_UART_RX,
65 .end = IRQ_UART_ERROR,
70 .end = CH_UART_TX,
75 .end = CH_UART_RX,
101 .end = 0xFFC004FF,
106 .end = IRQ_UART0_RX+1,
111 .end = CH_UART0_RX+1,
H A Dcm_bf561.c104 .end = SPI0_REGBASE + 0xFF,
109 .end = CH_SPI,
114 .end = IRQ_SPI,
158 .end = 0x28000300 + 16,
162 .end = IRQ_PF0,
184 .end = 0x24008000 + 0xFF,
189 .end = IRQ_PF43,
216 .end = 0x24000000 + 0x100,
220 .end = IRQ_PF45,
237 .end = 0x24008000,
241 .end = 0x24008004,
245 .end = IRQ_PF47,
277 .end = BFIN_UART_GCTL+2,
282 .end = IRQ_UART_TX,
287 .end = IRQ_UART_RX,
292 .end = IRQ_UART_ERROR,
297 .end = CH_UART_TX,
302 .end = CH_UART_RX,
328 .end = 0xFFC004FF,
333 .end = IRQ_UART0_RX+1,
338 .end = CH_UART0_RX+1,
362 .end = 0x2400C001F,
367 .end = 0x2400D01B,
372 .end = PATA_INT,
413 .end = 0x207fffff,
H A Dacvilon.c69 .end = 0x20000000 + 0x000fffff,
74 .end = IRQ_PF15,
104 .end = 0x2C000000 + 16,
109 .end = IRQ_PF8,
149 .end = 0x5ffffff,
170 .end = 0x28000000 + 0xFF,
175 .end = IRQ_PF7,
203 .end = BFIN_UART_GCTL + 2,
208 .end = IRQ_UART_TX,
213 .end = IRQ_UART_RX,
218 .end = IRQ_UART_ERROR,
223 .end = CH_UART_TX,
228 .end = CH_UART_RX,
303 .end = 0x24000000 + (1 << MAX(BFIN_NAND_PLAT_CLE, BFIN_NAND_PLAT_ALE)),
378 .end = SPI0_REGBASE + 0xFF,
383 .end = CH_SPI,
388 .end = IRQ_SPI,
435 /* .end = MAX_BLACKFIN_GPIOS - 1, */
436 .end = 32,
/linux-4.4.14/arch/c6x/include/asm/
H A Dcache.h66 extern void enable_caching(unsigned long start, unsigned long end);
67 extern void disable_caching(unsigned long start, unsigned long end);
80 extern void L1P_cache_block_invalidate(unsigned int start, unsigned int end);
81 extern void L1D_cache_block_invalidate(unsigned int start, unsigned int end);
83 unsigned int end);
84 extern void L1D_cache_block_writeback(unsigned int start, unsigned int end);
85 extern void L2_cache_block_invalidate(unsigned int start, unsigned int end);
86 extern void L2_cache_block_writeback(unsigned int start, unsigned int end);
88 unsigned int end);
90 unsigned int end);
92 unsigned int end);
95 unsigned int end);
/linux-4.4.14/arch/m68k/mm/
H A Dcache.c78 unsigned long start, end; flush_icache_range() local
80 end = endaddr & ICACHE_SET_MASK; flush_icache_range()
81 if (start > end) { flush_icache_range()
82 flush_cf_icache(0, end); flush_icache_range()
83 end = ICACHE_MAX_ADDR; flush_icache_range()
85 flush_cf_icache(start, end); flush_icache_range()
112 unsigned long start, end; flush_icache_user_range() local
114 end = (addr + len) & ICACHE_SET_MASK; flush_icache_user_range()
115 if (start > end) { flush_icache_user_range()
116 flush_cf_icache(0, end); flush_icache_user_range()
117 end = ICACHE_MAX_ADDR; flush_icache_user_range()
119 flush_cf_icache(start, end); flush_icache_user_range()
/linux-4.4.14/arch/sparc/lib/
H A Duser_fixup.c24 unsigned long end = start + size; compute_size() local
26 if (fault_addr < start || fault_addr >= end) { compute_size()
30 size = end - fault_addr; compute_size()
59 unsigned long end = start + size; copy_in_user_fixup() local
61 if (fault_addr >= start && fault_addr < end) copy_in_user_fixup()
62 return end - fault_addr; copy_in_user_fixup()
65 end = start + size; copy_in_user_fixup()
66 if (fault_addr >= start && fault_addr < end) copy_in_user_fixup()
67 return end - fault_addr; copy_in_user_fixup()
/linux-4.4.14/mm/
H A Dmemtest.c36 u64 *p, *start, *end; memtest() local
43 end = start + (size - (start_phys_aligned - start_phys)) / incr; memtest()
47 for (p = start; p < end; p++) memtest()
50 for (p = start; p < end; p++, start_phys_aligned += incr) { memtest()
65 static void __init do_one_pass(u64 pattern, phys_addr_t start, phys_addr_t end) do_one_pass() argument
72 this_start = clamp(this_start, start, end); do_one_pass()
73 this_end = clamp(this_end, start, end); do_one_pass()
99 void __init early_memtest(phys_addr_t start, phys_addr_t end) early_memtest() argument
110 do_one_pass(patterns[idx], start, end); early_memtest()
H A Dmsync.c33 unsigned long end; SYSCALL_DEFINE3() local
47 end = start + len; SYSCALL_DEFINE3()
48 if (end < start) SYSCALL_DEFINE3()
51 if (end == start) SYSCALL_DEFINE3()
54 * If the interval [start,end) covers some unmapped address ranges, SYSCALL_DEFINE3()
55 * just ignore them, but return -ENOMEM at the end. SYSCALL_DEFINE3()
63 /* Still start < end. */ SYSCALL_DEFINE3()
70 if (start >= end) SYSCALL_DEFINE3()
83 fend = fstart + (min(end, vma->vm_end) - start) - 1; SYSCALL_DEFINE3()
91 if (error || start >= end) SYSCALL_DEFINE3()
96 if (start >= end) { SYSCALL_DEFINE3()
H A Dpagewalk.c6 static int walk_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, walk_pte_range() argument
18 if (addr == end) walk_pte_range()
27 static int walk_pmd_range(pud_t *pud, unsigned long addr, unsigned long end, walk_pmd_range() argument
37 next = pmd_addr_end(addr, end); walk_pmd_range()
67 } while (pmd++, addr = next, addr != end); walk_pmd_range()
72 static int walk_pud_range(pgd_t *pgd, unsigned long addr, unsigned long end, walk_pud_range() argument
81 next = pud_addr_end(addr, end); walk_pud_range()
93 } while (pud++, addr = next, addr != end); walk_pud_range()
98 static int walk_pgd_range(unsigned long addr, unsigned long end, walk_pgd_range() argument
107 next = pgd_addr_end(addr, end); walk_pgd_range()
119 } while (pgd++, addr = next, addr != end); walk_pgd_range()
126 unsigned long end) hugetlb_entry_end()
129 return boundary < end ? boundary : end; hugetlb_entry_end()
132 static int walk_hugetlb_range(unsigned long addr, unsigned long end, walk_hugetlb_range() argument
143 next = hugetlb_entry_end(h, addr, end); walk_hugetlb_range()
149 } while (addr = next, addr != end); walk_hugetlb_range()
155 static int walk_hugetlb_range(unsigned long addr, unsigned long end, walk_hugetlb_range() argument
164 * Decide whether we really walk over the current vma on [@start, @end)
169 static int walk_page_test(unsigned long start, unsigned long end, walk_page_test() argument
175 return walk->test_walk(start, end, walk); walk_page_test()
188 err = walk->pte_hole(start, end, walk); walk_page_test()
194 static int __walk_page_range(unsigned long start, unsigned long end, __walk_page_range() argument
202 err = walk_hugetlb_range(start, end, walk); __walk_page_range()
204 err = walk_pgd_range(start, end, walk); __walk_page_range()
213 * within the virtual address range [@start, @end). During walking, we can do
219 * end address yet, continue to walk.
239 int walk_page_range(unsigned long start, unsigned long end, walk_page_range() argument
246 if (start >= end) walk_page_range()
258 next = end; walk_page_range()
261 next = min(end, vma->vm_start); walk_page_range()
264 next = min(end, vma->vm_end); walk_page_range()
284 } while (start = next, start < end); walk_page_range()
125 hugetlb_entry_end(struct hstate *h, unsigned long addr, unsigned long end) hugetlb_entry_end() argument
H A Dmprotect.c63 unsigned long addr, unsigned long end, pgprot_t newprot, change_pte_range()
129 } while (pte++, addr += PAGE_SIZE, addr != end); change_pte_range()
137 pud_t *pud, unsigned long addr, unsigned long end, change_pmd_range()
151 next = pmd_addr_end(addr, end); change_pmd_range()
158 mmu_notifier_invalidate_range_start(mm, mni_start, end); change_pmd_range()
183 } while (pmd++, addr = next, addr != end); change_pmd_range()
186 mmu_notifier_invalidate_range_end(mm, mni_start, end); change_pmd_range()
194 pgd_t *pgd, unsigned long addr, unsigned long end, change_pud_range()
203 next = pud_addr_end(addr, end); change_pud_range()
208 } while (pud++, addr = next, addr != end); change_pud_range()
214 unsigned long addr, unsigned long end, pgprot_t newprot, change_protection_range()
223 BUG_ON(addr >= end); change_protection_range()
225 flush_cache_range(vma, addr, end); change_protection_range()
228 next = pgd_addr_end(addr, end); change_protection_range()
233 } while (pgd++, addr = next, addr != end); change_protection_range()
237 flush_tlb_range(vma, start, end); change_protection_range()
244 unsigned long end, pgprot_t newprot, change_protection()
250 pages = hugetlb_change_protection(vma, start, end, newprot); change_protection()
252 pages = change_protection_range(vma, start, end, newprot, dirty_accountable, prot_numa); change_protection()
259 unsigned long start, unsigned long end, unsigned long newflags) mprotect_fixup()
263 long nrpages = (end - start) >> PAGE_SHIFT; mprotect_fixup()
294 *pprev = vma_merge(mm, *pprev, start, end, newflags, mprotect_fixup()
310 if (end != vma->vm_end) { mprotect_fixup()
311 error = split_vma(mm, vma, end, 0); mprotect_fixup()
325 change_protection(vma, start, end, vma->vm_page_prot, mprotect_fixup()
334 populate_vma_page_range(vma, start, end, NULL); mprotect_fixup()
350 unsigned long vm_flags, nstart, end, tmp, reqprot; SYSCALL_DEFINE3() local
363 end = start + len; SYSCALL_DEFINE3()
364 if (end <= start) SYSCALL_DEFINE3()
386 if (vma->vm_start >= end) SYSCALL_DEFINE3()
396 end = vma->vm_end; SYSCALL_DEFINE3()
424 if (tmp > end) SYSCALL_DEFINE3()
425 tmp = end; SYSCALL_DEFINE3()
433 if (nstart >= end) SYSCALL_DEFINE3()
62 change_pte_range(struct vm_area_struct *vma, pmd_t *pmd, unsigned long addr, unsigned long end, pgprot_t newprot, int dirty_accountable, int prot_numa) change_pte_range() argument
136 change_pmd_range(struct vm_area_struct *vma, pud_t *pud, unsigned long addr, unsigned long end, pgprot_t newprot, int dirty_accountable, int prot_numa) change_pmd_range() argument
193 change_pud_range(struct vm_area_struct *vma, pgd_t *pgd, unsigned long addr, unsigned long end, pgprot_t newprot, int dirty_accountable, int prot_numa) change_pud_range() argument
213 change_protection_range(struct vm_area_struct *vma, unsigned long addr, unsigned long end, pgprot_t newprot, int dirty_accountable, int prot_numa) change_protection_range() argument
243 change_protection(struct vm_area_struct *vma, unsigned long start, unsigned long end, pgprot_t newprot, int dirty_accountable, int prot_numa) change_protection() argument
258 mprotect_fixup(struct vm_area_struct *vma, struct vm_area_struct **pprev, unsigned long start, unsigned long end, unsigned long newflags) mprotect_fixup() argument
H A Dmadvise.c48 unsigned long start, unsigned long end, int behavior) madvise_behavior()
87 error = ksm_madvise(vma, start, end, behavior, &new_flags); madvise_behavior()
105 *prev = vma_merge(mm, *prev, start, end, new_flags, vma->anon_vma, madvise_behavior()
121 if (end != vma->vm_end) { madvise_behavior()
122 error = split_vma(mm, vma, end, 0); madvise_behavior()
141 unsigned long end, struct mm_walk *walk) swapin_walk_pmd_entry()
150 for (index = start; index != end; index += PAGE_SIZE) { swapin_walk_pmd_entry()
176 unsigned long start, unsigned long end) force_swapin_readahead()
184 walk_page_range(start, end, &walk); force_swapin_readahead()
190 unsigned long start, unsigned long end, force_shm_swapin_readahead()
197 for (; start < end; start += PAGE_SIZE) { force_shm_swapin_readahead()
222 unsigned long start, unsigned long end) madvise_willneed()
229 force_swapin_readahead(vma, start, end); madvise_willneed()
235 force_shm_swapin_readahead(vma, start, end, madvise_willneed()
251 if (end > vma->vm_end) madvise_willneed()
252 end = vma->vm_end; madvise_willneed()
253 end = ((end - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; madvise_willneed()
255 force_page_cache_readahead(file->f_mapping, file, start, end - start); madvise_willneed()
280 unsigned long start, unsigned long end) madvise_dontneed()
286 zap_page_range(vma, start, end - start, NULL); madvise_dontneed()
296 unsigned long start, unsigned long end) madvise_remove()
329 offset, end - start); madvise_remove()
339 static int madvise_hwpoison(int bhv, unsigned long start, unsigned long end) madvise_hwpoison() argument
344 for (; start < end; start += PAGE_SIZE << madvise_hwpoison()
375 unsigned long start, unsigned long end, int behavior) madvise_vma()
379 return madvise_remove(vma, prev, start, end); madvise_vma()
381 return madvise_willneed(vma, prev, start, end); madvise_vma()
383 return madvise_dontneed(vma, prev, start, end); madvise_vma()
385 return madvise_behavior(vma, prev, start, end, behavior); madvise_vma()
462 unsigned long end, tmp; SYSCALL_DEFINE3() local
485 end = start + len; SYSCALL_DEFINE3()
486 if (end < start) SYSCALL_DEFINE3()
490 if (end == start) SYSCALL_DEFINE3()
500 * If the interval [start,end) covers some unmapped address SYSCALL_DEFINE3()
501 * ranges, just ignore them, but return -ENOMEM at the end. SYSCALL_DEFINE3()
510 /* Still start < end. */ SYSCALL_DEFINE3()
515 /* Here start < (end|vma->vm_end). */ SYSCALL_DEFINE3()
519 if (start >= end) SYSCALL_DEFINE3()
523 /* Here vma->vm_start <= start < (end|vma->vm_end) */ SYSCALL_DEFINE3()
525 if (end < tmp) SYSCALL_DEFINE3()
526 tmp = end; SYSCALL_DEFINE3()
528 /* Here vma->vm_start <= start < tmp <= (end|vma->vm_end). */ SYSCALL_DEFINE3()
536 if (start >= end) SYSCALL_DEFINE3()
46 madvise_behavior(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) madvise_behavior() argument
140 swapin_walk_pmd_entry(pmd_t *pmd, unsigned long start, unsigned long end, struct mm_walk *walk) swapin_walk_pmd_entry() argument
175 force_swapin_readahead(struct vm_area_struct *vma, unsigned long start, unsigned long end) force_swapin_readahead() argument
189 force_shm_swapin_readahead(struct vm_area_struct *vma, unsigned long start, unsigned long end, struct address_space *mapping) force_shm_swapin_readahead() argument
220 madvise_willneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_willneed() argument
278 madvise_dontneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_dontneed() argument
294 madvise_remove(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_remove() argument
374 madvise_vma(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) madvise_vma() argument
H A Ddebug-pagealloc.c84 unsigned char *end; check_poison_mem() local
90 for (end = mem + bytes - 1; end > start; end--) { check_poison_mem()
91 if (*end != PAGE_POISON) check_poison_mem()
97 else if (start == end && single_bit_flip(*start, PAGE_POISON)) check_poison_mem()
103 end - start + 1, 1); check_poison_mem()
/linux-4.4.14/net/ceph/
H A Darmor.c4 int ceph_armor(char *dst, const char *src, const char *end);
5 int ceph_unarmor(char *dst, const char *src, const char *end);
36 int ceph_armor(char *dst, const char *src, const char *end) ceph_armor() argument
41 while (src < end) { ceph_armor()
46 if (src < end) { ceph_armor()
49 if (src < end) { ceph_armor()
74 int ceph_unarmor(char *dst, const char *src, const char *end) ceph_unarmor() argument
78 while (src < end) { ceph_unarmor()
85 if (src + 4 > end) ceph_unarmor()
H A Dosdmap.c55 static int crush_decode_uniform_bucket(void **p, void *end, crush_decode_uniform_bucket() argument
58 dout("crush_decode_uniform_bucket %p to %p\n", *p, end); crush_decode_uniform_bucket()
59 ceph_decode_need(p, end, (1+b->h.size) * sizeof(u32), bad); crush_decode_uniform_bucket()
66 static int crush_decode_list_bucket(void **p, void *end, crush_decode_list_bucket() argument
70 dout("crush_decode_list_bucket %p to %p\n", *p, end); crush_decode_list_bucket()
77 ceph_decode_need(p, end, 2 * b->h.size * sizeof(u32), bad); crush_decode_list_bucket()
87 static int crush_decode_tree_bucket(void **p, void *end, crush_decode_tree_bucket() argument
91 dout("crush_decode_tree_bucket %p to %p\n", *p, end); crush_decode_tree_bucket()
92 ceph_decode_8_safe(p, end, b->num_nodes, bad); crush_decode_tree_bucket()
96 ceph_decode_need(p, end, b->num_nodes * sizeof(u32), bad); crush_decode_tree_bucket()
104 static int crush_decode_straw_bucket(void **p, void *end, crush_decode_straw_bucket() argument
108 dout("crush_decode_straw_bucket %p to %p\n", *p, end); crush_decode_straw_bucket()
115 ceph_decode_need(p, end, 2 * b->h.size * sizeof(u32), bad); crush_decode_straw_bucket()
125 static int crush_decode_straw2_bucket(void **p, void *end, crush_decode_straw2_bucket() argument
129 dout("crush_decode_straw2_bucket %p to %p\n", *p, end); crush_decode_straw2_bucket()
133 ceph_decode_need(p, end, b->h.size * sizeof(u32), bad); crush_decode_straw2_bucket()
141 static int skip_name_map(void **p, void *end) skip_name_map() argument
144 ceph_decode_32_safe(p, end, len ,bad); skip_name_map()
148 ceph_decode_32_safe(p, end, strlen, bad); skip_name_map()
156 static struct crush_map *crush_decode(void *pbyval, void *end) crush_decode() argument
166 dout("crush_decode %p to %p len %d\n", *p, end, (int)(end - *p)); crush_decode()
178 ceph_decode_need(p, end, 4*sizeof(u32), bad); crush_decode()
202 ceph_decode_32_safe(p, end, alg, bad); crush_decode()
208 i, (int)(*p-start), *p, end); crush_decode()
235 ceph_decode_need(p, end, 4*sizeof(u32), bad); crush_decode()
244 b->size, (int)(*p-start), *p, end); crush_decode()
254 ceph_decode_need(p, end, b->size*sizeof(u32), bad); crush_decode()
260 err = crush_decode_uniform_bucket(p, end, crush_decode()
266 err = crush_decode_list_bucket(p, end, crush_decode()
272 err = crush_decode_tree_bucket(p, end, crush_decode()
278 err = crush_decode_straw_bucket(p, end, crush_decode()
284 err = crush_decode_straw2_bucket(p, end, crush_decode()
298 ceph_decode_32_safe(p, end, yes, bad); crush_decode()
301 i, (int)(*p-start), *p, end); crush_decode()
307 i, (int)(*p-start), *p, end); crush_decode()
310 ceph_decode_32_safe(p, end, yes, bad); crush_decode()
324 ceph_decode_copy_safe(p, end, &r->mask, 4, bad); /* 4 u8's */ crush_decode()
325 ceph_decode_need(p, end, r->len*3*sizeof(u32), bad); crush_decode()
335 err = skip_name_map(p, end); crush_decode()
341 ceph_decode_need(p, end, 3*sizeof(u32), done); crush_decode()
352 ceph_decode_need(p, end, sizeof(u32), done); crush_decode()
357 ceph_decode_need(p, end, sizeof(u8), done); crush_decode()
539 static int decode_pool(void **p, void *end, struct ceph_pg_pool_info *pi) decode_pool() argument
545 ceph_decode_need(p, end, 2 + 4, bad); decode_pool()
557 ceph_decode_need(p, end, len, bad); decode_pool()
620 static int decode_pool_names(void **p, void *end, struct ceph_osdmap *map) decode_pool_names() argument
626 ceph_decode_32_safe(p, end, num, bad); decode_pool_names()
629 ceph_decode_64_safe(p, end, pool, bad); decode_pool_names()
630 ceph_decode_32_safe(p, end, len, bad); decode_pool_names()
632 ceph_decode_need(p, end, len, bad); decode_pool_names()
746 static int get_osdmap_client_data_v(void **p, void *end, get_osdmap_client_data_v() argument
751 ceph_decode_8_safe(p, end, struct_v, e_inval); get_osdmap_client_data_v()
755 ceph_decode_8_safe(p, end, struct_compat, e_inval); get_osdmap_client_data_v()
764 ceph_decode_8_safe(p, end, struct_v, e_inval); get_osdmap_client_data_v()
765 ceph_decode_8_safe(p, end, struct_compat, e_inval); get_osdmap_client_data_v()
777 ceph_decode_16_safe(p, end, version, e_inval); get_osdmap_client_data_v()
795 static int __decode_pools(void **p, void *end, struct ceph_osdmap *map, __decode_pools() argument
800 ceph_decode_32_safe(p, end, n, e_inval); __decode_pools()
806 ceph_decode_64_safe(p, end, pool, e_inval); __decode_pools()
823 ret = decode_pool(p, end, pi); __decode_pools()
834 static int decode_pools(void **p, void *end, struct ceph_osdmap *map) decode_pools() argument
836 return __decode_pools(p, end, map, false); decode_pools()
839 static int decode_new_pools(void **p, void *end, struct ceph_osdmap *map) decode_new_pools() argument
841 return __decode_pools(p, end, map, true); decode_new_pools()
844 static int __decode_pg_temp(void **p, void *end, struct ceph_osdmap *map, __decode_pg_temp() argument
849 ceph_decode_32_safe(p, end, n, e_inval); __decode_pg_temp()
855 ret = ceph_decode_pgid(p, end, &pgid); __decode_pg_temp()
859 ceph_decode_32_safe(p, end, len, e_inval); __decode_pg_temp()
867 ceph_decode_need(p, end, len*sizeof(u32), e_inval); __decode_pg_temp()
895 static int decode_pg_temp(void **p, void *end, struct ceph_osdmap *map) decode_pg_temp() argument
897 return __decode_pg_temp(p, end, map, false); decode_pg_temp()
900 static int decode_new_pg_temp(void **p, void *end, struct ceph_osdmap *map) decode_new_pg_temp() argument
902 return __decode_pg_temp(p, end, map, true); decode_new_pg_temp()
905 static int __decode_primary_temp(void **p, void *end, struct ceph_osdmap *map, __decode_primary_temp() argument
910 ceph_decode_32_safe(p, end, n, e_inval); __decode_primary_temp()
916 ret = ceph_decode_pgid(p, end, &pgid); __decode_primary_temp()
920 ceph_decode_32_safe(p, end, osd, e_inval); __decode_primary_temp()
949 static int decode_primary_temp(void **p, void *end, struct ceph_osdmap *map) decode_primary_temp() argument
951 return __decode_primary_temp(p, end, map, false); decode_primary_temp()
954 static int decode_new_primary_temp(void **p, void *end, decode_new_primary_temp() argument
957 return __decode_primary_temp(p, end, map, true); decode_new_primary_temp()
992 static int decode_primary_affinity(void **p, void *end, decode_primary_affinity() argument
997 ceph_decode_32_safe(p, end, len, e_inval); decode_primary_affinity()
1006 ceph_decode_need(p, end, map->max_osd*sizeof(u32), e_inval); decode_primary_affinity()
1022 static int decode_new_primary_affinity(void **p, void *end, decode_new_primary_affinity() argument
1027 ceph_decode_32_safe(p, end, n, e_inval); decode_new_primary_affinity()
1032 ceph_decode_32_safe(p, end, osd, e_inval); decode_new_primary_affinity()
1033 ceph_decode_32_safe(p, end, aff, e_inval); decode_new_primary_affinity()
1051 static int osdmap_decode(void **p, void *end, struct ceph_osdmap *map) osdmap_decode() argument
1060 dout("%s %p to %p len %d\n", __func__, *p, end, (int)(end - *p)); osdmap_decode()
1062 err = get_osdmap_client_data_v(p, end, "full", &struct_v); osdmap_decode()
1067 ceph_decode_need(p, end, sizeof(map->fsid) + sizeof(u32) + osdmap_decode()
1075 err = decode_pools(p, end, map); osdmap_decode()
1080 err = decode_pool_names(p, end, map); osdmap_decode()
1084 ceph_decode_32_safe(p, end, map->pool_max, e_inval); osdmap_decode()
1086 ceph_decode_32_safe(p, end, map->flags, e_inval); osdmap_decode()
1089 ceph_decode_32_safe(p, end, max, e_inval); osdmap_decode()
1097 ceph_decode_need(p, end, 3*sizeof(u32) + osdmap_decode()
1120 err = decode_pg_temp(p, end, map); osdmap_decode()
1126 err = decode_primary_temp(p, end, map); osdmap_decode()
1133 err = decode_primary_affinity(p, end, map); osdmap_decode()
1143 ceph_decode_32_safe(p, end, len, e_inval); osdmap_decode()
1144 map->crush = crush_decode(*p, min(*p + len, end)); osdmap_decode()
1153 *p = end; osdmap_decode()
1162 err, epoch, (int)(*p - start), *p, start, end); osdmap_decode()
1165 start, end - start, true); osdmap_decode()
1172 struct ceph_osdmap *ceph_osdmap_decode(void **p, void *end) ceph_osdmap_decode() argument
1185 ret = osdmap_decode(p, end, map); ceph_osdmap_decode()
1197 struct ceph_osdmap *osdmap_apply_incremental(void **p, void *end, osdmap_apply_incremental() argument
1213 dout("%s %p to %p len %d\n", __func__, *p, end, (int)(end - *p)); osdmap_apply_incremental()
1215 err = get_osdmap_client_data_v(p, end, "inc", &struct_v); osdmap_apply_incremental()
1220 ceph_decode_need(p, end, sizeof(fsid) + sizeof(u32) + sizeof(modified) + osdmap_apply_incremental()
1230 ceph_decode_32_safe(p, end, len, e_inval); osdmap_apply_incremental()
1233 len, *p, end); osdmap_apply_incremental()
1234 return ceph_osdmap_decode(p, min(*p+len, end)); osdmap_apply_incremental()
1238 ceph_decode_32_safe(p, end, len, e_inval); osdmap_apply_incremental()
1240 newcrush = crush_decode(*p, min(*p+len, end)); osdmap_apply_incremental()
1256 ceph_decode_32_safe(p, end, max, e_inval); osdmap_apply_incremental()
1273 err = decode_new_pools(p, end, map); osdmap_apply_incremental()
1278 err = decode_pool_names(p, end, map); osdmap_apply_incremental()
1283 ceph_decode_32_safe(p, end, len, e_inval); osdmap_apply_incremental()
1287 ceph_decode_64_safe(p, end, pool, e_inval); osdmap_apply_incremental()
1294 ceph_decode_32_safe(p, end, len, e_inval); osdmap_apply_incremental()
1298 ceph_decode_32_safe(p, end, osd, e_inval); osdmap_apply_incremental()
1299 ceph_decode_copy_safe(p, end, &addr, sizeof(addr), e_inval); osdmap_apply_incremental()
1308 ceph_decode_32_safe(p, end, len, e_inval); osdmap_apply_incremental()
1312 ceph_decode_32_safe(p, end, osd, e_inval); osdmap_apply_incremental()
1324 ceph_decode_32_safe(p, end, len, e_inval); osdmap_apply_incremental()
1327 ceph_decode_need(p, end, sizeof(u32)*2, e_inval); osdmap_apply_incremental()
1338 err = decode_new_pg_temp(p, end, map); osdmap_apply_incremental()
1344 err = decode_new_primary_temp(p, end, map); osdmap_apply_incremental()
1351 err = decode_new_primary_affinity(p, end, map); osdmap_apply_incremental()
1357 *p = end; osdmap_apply_incremental()
1366 err, epoch, (int)(*p - start), *p, start, end); osdmap_apply_incremental()
1369 start, end - start, true); osdmap_apply_incremental()
H A Dbuffer.c43 int ceph_decode_buffer(struct ceph_buffer **b, void **p, void *end) ceph_decode_buffer() argument
47 ceph_decode_need(p, end, sizeof(u32), bad); ceph_decode_buffer()
50 ceph_decode_need(p, end, len, bad); ceph_decode_buffer()
H A Dauth_none.c42 static int build_request(struct ceph_auth_client *ac, void *buf, void *end) build_request() argument
52 void *buf, void *end) handle_reply()
71 void *p, *end; ceph_auth_none_create_authorizer() local
76 end = p + sizeof(au->buf); ceph_auth_none_create_authorizer()
78 ret = ceph_entity_name_encode(ac->name, &p, end - 8); ceph_auth_none_create_authorizer()
81 ceph_decode_need(&p, end, sizeof(u64), bad2); ceph_auth_none_create_authorizer()
51 handle_reply(struct ceph_auth_client *ac, int result, void *buf, void *end) handle_reply() argument
/linux-4.4.14/arch/sh/mm/
H A Dcache-sh2.c22 unsigned long begin, end; sh2__flush_wback_region() local
25 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2__flush_wback_region()
27 for (v = begin; v < end; v+=L1_CACHE_BYTES) { sh2__flush_wback_region()
43 unsigned long begin, end; sh2__flush_purge_region() local
46 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2__flush_purge_region()
49 for (v = begin; v < end; v+=L1_CACHE_BYTES) sh2__flush_purge_region()
74 unsigned long begin, end; sh2__flush_invalidate_region()
77 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2__flush_invalidate_region()
80 for (v = begin; v < end; v+=L1_CACHE_BYTES) sh2__flush_invalidate_region()
H A Dflush-sh4.c15 reg_size_t aligned_start, v, cnt, end; sh4__flush_wback_region() local
19 end = (aligned_start + size + L1_CACHE_BYTES-1) sh4__flush_wback_region()
21 cnt = (end - v) / L1_CACHE_BYTES; sh4__flush_wback_region()
49 reg_size_t aligned_start, v, cnt, end; sh4__flush_purge_region() local
53 end = (aligned_start + size + L1_CACHE_BYTES-1) sh4__flush_purge_region()
55 cnt = (end - v) / L1_CACHE_BYTES; sh4__flush_purge_region()
79 reg_size_t aligned_start, v, cnt, end; sh4__flush_invalidate_region() local
83 end = (aligned_start + size + L1_CACHE_BYTES-1) sh4__flush_invalidate_region()
85 cnt = (end - v) / L1_CACHE_BYTES; sh4__flush_invalidate_region()
H A Dcache-sh2a.c54 unsigned long begin, end; sh2a__flush_wback_region() local
59 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2a__flush_wback_region()
67 if (((end - begin) >> PAGE_SHIFT) >= MAX_OCACHE_PAGES) { sh2a__flush_wback_region()
69 end = begin + (nr_ways * current_cpu_data.dcache.way_size); sh2a__flush_wback_region()
71 for (v = begin; v < end; v += L1_CACHE_BYTES) { sh2a__flush_wback_region()
79 for (v = begin; v < end; v += L1_CACHE_BYTES) sh2a__flush_wback_region()
95 unsigned long begin, end; sh2a__flush_purge_region() local
99 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2a__flush_purge_region()
105 for (v = begin; v < end; v+=L1_CACHE_BYTES) { sh2a__flush_purge_region()
125 unsigned long begin, end; sh2a__flush_invalidate_region() local
129 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2a__flush_invalidate_region()
136 if (((end - begin) >> PAGE_SHIFT) >= MAX_OCACHE_PAGES) { sh2a__flush_invalidate_region()
140 for (v = begin; v < end; v += L1_CACHE_BYTES) sh2a__flush_invalidate_region()
154 unsigned long start, end; sh2a_flush_icache_range() local
159 end = (data->addr2 + L1_CACHE_BYTES-1) & ~(L1_CACHE_BYTES-1); sh2a_flush_icache_range()
162 sh2a__flush_wback_region((void *)start, end-start); sh2a_flush_icache_range()
170 if (((end - start) >> PAGE_SHIFT) >= MAX_ICACHE_PAGES) { sh2a_flush_icache_range()
174 for (v = start; v < end; v += L1_CACHE_BYTES) sh2a_flush_icache_range()
H A Dnuma.c27 void __init setup_bootmem_node(int nid, unsigned long start, unsigned long end) setup_bootmem_node() argument
37 end_pfn = PFN_DOWN(end); setup_bootmem_node()
39 pmb_bolt_mapping((unsigned long)__va(start), start, end - start, setup_bootmem_node()
42 memblock_add(start, end - start); setup_bootmem_node()
48 SMP_CACHE_BYTES, end)); setup_bootmem_node()
58 PAGE_SIZE, end); setup_bootmem_node()
H A Dtlbflush_32.c40 unsigned long end) local_flush_tlb_range()
50 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_range()
61 end += (PAGE_SIZE - 1); local_flush_tlb_range()
62 end &= PAGE_MASK; local_flush_tlb_range()
67 while (start < end) { local_flush_tlb_range()
78 void local_flush_tlb_kernel_range(unsigned long start, unsigned long end) local_flush_tlb_kernel_range() argument
85 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_kernel_range()
94 end += (PAGE_SIZE - 1); local_flush_tlb_kernel_range()
95 end &= PAGE_MASK; local_flush_tlb_kernel_range()
97 while (start < end) { local_flush_tlb_kernel_range()
39 local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) local_flush_tlb_range() argument
H A Dgup.c75 unsigned long end, int write, struct page **pages, int *nr) gup_pte_range()
113 } while (ptep++, addr += PAGE_SIZE, addr != end); gup_pte_range()
119 static int gup_pmd_range(pud_t pud, unsigned long addr, unsigned long end, gup_pmd_range() argument
129 next = pmd_addr_end(addr, end); gup_pmd_range()
134 } while (pmdp++, addr = next, addr != end); gup_pmd_range()
139 static int gup_pud_range(pgd_t pgd, unsigned long addr, unsigned long end, gup_pud_range() argument
149 next = pud_addr_end(addr, end); gup_pud_range()
154 } while (pudp++, addr = next, addr != end); gup_pud_range()
167 unsigned long addr, len, end; __get_user_pages_fast() local
176 end = start + len; __get_user_pages_fast()
190 next = pgd_addr_end(addr, end); __get_user_pages_fast()
195 } while (pgdp++, addr = next, addr != end); __get_user_pages_fast()
221 unsigned long addr, len, end; get_user_pages_fast() local
230 end = start + len; get_user_pages_fast()
231 if (end < start) get_user_pages_fast()
239 next = pgd_addr_end(addr, end); get_user_pages_fast()
244 } while (pgdp++, addr = next, addr != end); get_user_pages_fast()
247 VM_BUG_ON(nr != (end - start) >> PAGE_SHIFT); get_user_pages_fast()
261 (end - start) >> PAGE_SHIFT, write, 0, pages); get_user_pages_fast()
74 gup_pte_range(pmd_t pmd, unsigned long addr, unsigned long end, int write, struct page **pages, int *nr) gup_pte_range() argument
/linux-4.4.14/arch/arm/mach-davinci/
H A Ddevices-da8xx.c164 .end = DA8XX_TPCC_BASE + SZ_32K - 1,
170 .end = DA8XX_TPTC0_BASE + SZ_1K - 1,
176 .end = DA8XX_TPTC1_BASE + SZ_1K - 1,
195 .end = DA850_TPCC1_BASE + SZ_32K - 1,
201 .end = DA850_TPTC2_BASE + SZ_1K - 1,
267 .end = DA8XX_I2C0_BASE + SZ_4K - 1,
272 .end = IRQ_DA8XX_I2CINT0,
287 .end = DA8XX_I2C1_BASE + SZ_4K - 1,
292 .end = IRQ_DA8XX_I2CINT1,
323 .end = DA8XX_WDOG_BASE + SZ_4K - 1,
356 .end = DA8XX_EMAC_CPPI_PORT_BASE + SZ_16K - 1,
361 .end = IRQ_DA8XX_C0_RX_THRESH_PULSE,
366 .end = IRQ_DA8XX_C0_RX_PULSE,
371 .end = IRQ_DA8XX_C0_TX_PULSE,
376 .end = IRQ_DA8XX_C0_MISC_PULSE,
402 .end = DA8XX_EMAC_MDIO_BASE + SZ_4K - 1,
429 .end = DAVINCI_DA830_MCASP1_REG_BASE + (SZ_1K * 12) - 1,
436 .end = DAVINCI_DA830_DMA_MCASP1_AXEVT,
443 .end = DAVINCI_DA830_DMA_MCASP1_AREVT,
464 .end = DAVINCI_DA830_MCASP2_REG_BASE + (SZ_1K * 12) - 1,
471 .end = DAVINCI_DA830_DMA_MCASP2_AXEVT,
478 .end = DAVINCI_DA830_DMA_MCASP2_AREVT,
499 .end = DAVINCI_DA8XX_MCASP0_REG_BASE + (SZ_1K * 12) - 1,
506 .end = DAVINCI_DA8XX_DMA_MCASP0_AXEVT,
513 .end = DAVINCI_DA8XX_DMA_MCASP0_AREVT,
562 .end = DA8XX_PRUSS_MEM_BASE + 0xFFFF,
567 .end = IRQ_DA8XX_EVTOUT0,
572 .end = IRQ_DA8XX_EVTOUT1,
577 .end = IRQ_DA8XX_EVTOUT2,
582 .end = IRQ_DA8XX_EVTOUT3,
587 .end = IRQ_DA8XX_EVTOUT4,
592 .end = IRQ_DA8XX_EVTOUT5,
597 .end = IRQ_DA8XX_EVTOUT6,
602 .end = IRQ_DA8XX_EVTOUT7,
648 .end = DA8XX_LCD_CNTRL_BASE + SZ_4K - 1,
653 .end = IRQ_DA8XX_LCDINT,
674 .end = DA8XX_GPIO_BASE + SZ_4K - 1,
679 .end = IRQ_DA8XX_GPIO8,
700 .end = DA8XX_MMCSD0_BASE + SZ_4K - 1,
705 .end = IRQ_DA8XX_MMCSDINT0,
710 .end = DA8XX_DMA_MMCSD0_RX,
715 .end = DA8XX_DMA_MMCSD0_TX,
737 .end = DA850_MMCSD1_BASE + SZ_4K - 1,
742 .end = IRQ_DA850_MMCSDINT0_1,
747 .end = DA850_DMA_MMCSD1_RX,
752 .end = DA850_DMA_MMCSD1_TX,
774 .end = DA8XX_SYSCFG0_BASE + DA8XX_HOST1CFG_REG + 3,
779 .end = DA8XX_SYSCFG0_BASE + DA8XX_CHIPSIG_REG + 7,
784 .end = IRQ_DA8XX_CHIPINT0,
860 .end = DA8XX_RTC_BASE + SZ_4K - 1,
865 .end = IRQ_DA8XX_RTC,
870 .end = IRQ_DA8XX_RTC,
903 .end = DA8XX_DDR2_CTL_BASE + SZ_32K - 1,
933 .end = DA8XX_SPI0_BASE + SZ_4K - 1,
938 .end = IRQ_DA8XX_SPINT0,
943 .end = DA8XX_DMA_SPI0_RX,
948 .end = DA8XX_DMA_SPI0_TX,
956 .end = DA830_SPI1_BASE + SZ_4K - 1,
961 .end = IRQ_DA8XX_SPINT1,
966 .end = DA8XX_DMA_SPI1_RX,
971 .end = DA8XX_DMA_SPI1_TX,
1021 da8xx_spi1_resources[0].end = DA850_SPI1_BASE + SZ_4K - 1; da8xx_register_spi_bus()
1031 .end = DA850_SATA_BASE + 0x1fff,
1036 .end = DA8XX_SYSCFG1_BASE + DA8XX_PWRDN_REG + 0x3,
/linux-4.4.14/arch/blackfin/mach-bf538/boards/
H A Dezkit.c48 .end = UART0_GCTL+2,
53 .end = IRQ_UART0_TX,
58 .end = IRQ_UART0_RX,
63 .end = IRQ_UART0_ERROR,
68 .end = CH_UART0_TX,
73 .end = CH_UART0_RX,
79 .end = GPIO_PG7,
84 .end = GPIO_PG6,
108 .end = UART1_GCTL+2,
113 .end = IRQ_UART1_TX,
118 .end = IRQ_UART1_RX,
123 .end = IRQ_UART1_ERROR,
128 .end = CH_UART1_TX,
133 .end = CH_UART1_RX,
156 .end = UART2_GCTL+2,
161 .end = IRQ_UART2_TX,
166 .end = IRQ_UART2_RX,
171 .end = IRQ_UART2_ERROR,
176 .end = CH_UART2_TX,
181 .end = CH_UART2_RX,
207 .end = 0xFFC004FF,
212 .end = IRQ_UART0_RX+1,
217 .end = CH_UART0_RX+1,
232 .end = 0xFFC020FF,
237 .end = IRQ_UART1_RX+1,
242 .end = CH_UART1_RX+1,
257 .end = 0xFFC021FF,
262 .end = IRQ_UART2_RX+1,
267 .end = CH_UART2_RX+1,
285 .end = SPORT0_MRCS3+4,
290 .end = IRQ_SPORT0_RX+1,
295 .end = IRQ_SPORT0_ERROR,
319 .end = SPORT1_MRCS3+4,
324 .end = IRQ_SPORT1_RX+1,
329 .end = IRQ_SPORT1_ERROR,
353 .end = SPORT2_MRCS3+4,
358 .end = IRQ_SPORT2_RX+1,
363 .end = IRQ_SPORT2_ERROR,
387 .end = SPORT3_MRCS3+4,
392 .end = IRQ_SPORT3_RX+1,
397 .end = IRQ_SPORT3_ERROR,
427 .end = 0xFFC02FFF,
432 .end = IRQ_CAN_RX,
437 .end = IRQ_CAN_TX,
442 .end = IRQ_CAN_ERROR,
474 .end = 0x20310300 + 16,
478 .end = IRQ_PF0,
553 .end = IRQ_PPI_ERROR,
616 .end = SPI0_REGBASE + 0xFF,
621 .end = CH_SPI0,
626 .end = IRQ_SPI0,
635 .end = SPI1_REGBASE + 0xFF,
640 .end = CH_SPI1,
645 .end = IRQ_SPI1,
654 .end = SPI2_REGBASE + 0xFF,
659 .end = CH_SPI2,
664 .end = IRQ_SPI2,
724 .end = TWI0_REGBASE + 0xFF,
729 .end = IRQ_TWI0,
749 .end = TWI1_REGBASE + 0xFF,
754 .end = IRQ_TWI1,
841 .end = 0x202fffff,
843 .end = 0x203fffff,
/linux-4.4.14/arch/xtensa/mm/
H A Dinit.c43 sysmem.bank[i].start, sysmem.bank[i].end, sysmem_dump()
44 (sysmem.bank[i].end - sysmem.bank[i].start) >> 10); sysmem_dump()
88 int __init add_sysmem_bank(unsigned long start, unsigned long end) add_sysmem_bank() argument
95 if (start == end || add_sysmem_bank()
96 (start < end) != (PAGE_ALIGN(start) < (end & PAGE_MASK))) { add_sysmem_bank()
98 start, end - start); add_sysmem_bank()
103 end &= PAGE_MASK; add_sysmem_bank()
104 sz = end - start; add_sysmem_bank()
109 bank_sz = it->end - it->start; add_sysmem_bank()
112 if (end - it->start > bank_sz) add_sysmem_bank()
113 it->end = end; add_sysmem_bank()
125 if (it->end - it->start < sz) add_sysmem_bank()
126 it->end = end; add_sysmem_bank()
132 start, end - start); add_sysmem_bank()
136 it->end = end; add_sysmem_bank()
140 sz = it->end - it->start; add_sysmem_bank()
143 if (sz < sysmem.bank[i].end - it->start) add_sysmem_bank()
144 it->end = sysmem.bank[i].end; add_sysmem_bank()
154 * mem_reserve(start, end, must_exist)
162 * end End of region,
170 int __init mem_reserve(unsigned long start, unsigned long end, int must_exist) mem_reserve() argument
178 end = PAGE_ALIGN(end); mem_reserve()
179 sz = end - start; mem_reserve()
186 bank_sz = it->end - it->start; mem_reserve()
188 if ((!it || end - it->start > bank_sz) && must_exist) { mem_reserve()
190 start, end); mem_reserve()
196 if (end - it->start < bank_sz) { mem_reserve()
197 it->start = end; mem_reserve()
203 it->end = start; mem_reserve()
204 if (end - it->start < bank_sz) mem_reserve()
205 return add_sysmem_bank(end, mem_reserve()
215 if (it->end - start <= sz) { mem_reserve()
220 it->start = end; mem_reserve()
261 pfn = PAGE_ALIGN(sysmem.bank[i].end - 1) >> PAGE_SHIFT; bootmem_init()
279 if (sysmem.bank[i].end - sysmem.bank[i].start >= bootmap_size) { bootmem_init()
299 unsigned long end = min(max_low_pfn << PAGE_SHIFT, bootmem_init() local
300 sysmem.bank[i].end); bootmem_init()
302 end - sysmem.bank[i].start); bootmem_init()
370 void free_initrd_mem(unsigned long start, unsigned long end) free_initrd_mem() argument
373 free_reserved_area((void *)start, (void *)end, -1, "initrd"); free_initrd_mem()
/linux-4.4.14/arch/blackfin/mach-bf537/boards/
H A Dminotaur.c38 .end = 0x20312000,
42 .end = 0x20311FFF,
46 .end = IRQ_PF4,
50 .end = IRQ_PF6,
107 .end = 0x20300000 + 0x100,
111 .end = IRQ_PF7,
206 .end = SPI0_REGBASE + 0xFF,
211 .end = CH_SPI,
216 .end = IRQ_SPI,
237 .end = UART0_GCTL+2,
242 .end = IRQ_UART0_TX,
247 .end = IRQ_UART0_RX,
252 .end = IRQ_UART0_ERROR,
257 .end = CH_UART0_TX,
262 .end = CH_UART0_RX,
285 .end = UART1_GCTL+2,
290 .end = IRQ_UART1_TX,
295 .end = IRQ_UART1_RX,
300 .end = IRQ_UART1_ERROR,
305 .end = CH_UART1_TX,
310 .end = CH_UART1_RX,
336 .end = 0xFFC004FF,
341 .end = IRQ_UART0_RX+1,
346 .end = CH_UART0_RX+1,
362 .end = 0xFFC020FF,
367 .end = IRQ_UART1_RX+1,
372 .end = CH_UART1_RX+1,
392 .end = TWI0_REGBASE + 0xFF,
397 .end = IRQ_TWI,
418 .end = SPORT0_MRCS3+4,
423 .end = IRQ_SPORT0_RX+1,
428 .end = IRQ_SPORT0_ERROR,
452 .end = SPORT1_MRCS3+4,
457 .end = IRQ_SPORT1_RX+1,
462 .end = IRQ_SPORT1_ERROR,
H A Dtcm_bf537.c115 .end = SPI0_REGBASE + 0xFF,
120 .end = CH_SPI,
125 .end = IRQ_SPI,
173 .end = 0x20200300 + 16,
177 .end = IRQ_PF14,
197 .end = 0x20308000,
201 .end = 0x20308004,
205 .end = IRQ_PG15,
236 .end = 0x20300000 + 0x100,
240 .end = IRQ_PG13,
282 .end = 0x201fffff,
286 .end = ARRAY_SIZE(cm_flash_gpios),
307 .end = UART0_GCTL+2,
312 .end = IRQ_UART0_TX,
317 .end = IRQ_UART0_RX,
322 .end = IRQ_UART0_ERROR,
327 .end = CH_UART0_TX,
332 .end = CH_UART0_RX,
355 .end = UART1_GCTL+2,
360 .end = IRQ_UART1_TX,
365 .end = IRQ_UART1_RX,
370 .end = IRQ_UART1_ERROR,
375 .end = CH_UART1_TX,
380 .end = CH_UART1_RX,
406 .end = 0xFFC004FF,
411 .end = IRQ_UART0_RX+1,
416 .end = CH_UART0_RX+1,
432 .end = 0xFFC020FF,
437 .end = IRQ_UART1_RX+1,
442 .end = CH_UART1_RX+1,
462 .end = TWI0_REGBASE,
467 .end = IRQ_TWI,
488 .end = SPORT0_MRCS3+4,
493 .end = IRQ_SPORT0_RX+1,
498 .end = IRQ_SPORT0_ERROR,
522 .end = SPORT1_MRCS3+4,
527 .end = IRQ_SPORT1_RX+1,
532 .end = IRQ_SPORT1_ERROR,
597 .end = 0x2030C01F,
602 .end = 0x2030D01B,
607 .end = PATA_INT,
H A Dcm_bf537e.c115 .end = SPI0_REGBASE + 0xFF,
120 .end = CH_SPI,
125 .end = IRQ_SPI,
161 .end = SPORT0_TCR1 + 0xFF,
166 .end = IRQ_SPORT0_ERROR,
191 .end = SPORT1_TCR1 + 0xFF,
196 .end = IRQ_SPORT1_ERROR,
238 .end = 0x20200300 + 16,
242 .end = IRQ_PF14,
262 .end = 0x20308000,
266 .end = 0x20308004,
270 .end = IRQ_PG15,
301 .end = 0x20300000 + 0x100,
305 .end = IRQ_PG13,
347 .end = 0x201fffff,
351 .end = ARRAY_SIZE(cm_flash_gpios),
372 .end = UART0_GCTL+2,
377 .end = IRQ_UART0_TX,
382 .end = IRQ_UART0_RX,
387 .end = IRQ_UART0_ERROR,
392 .end = CH_UART0_TX,
397 .end = CH_UART0_RX,
406 .end = -1,
414 .end = -1,
438 .end = UART1_GCTL+2,
443 .end = IRQ_UART1_TX,
448 .end = IRQ_UART1_RX,
453 .end = IRQ_UART1_ERROR,
458 .end = CH_UART1_TX,
463 .end = CH_UART1_RX,
472 .end = -1,
480 .end = -1,
507 .end = 0xFFC004FF,
512 .end = IRQ_UART0_RX+1,
517 .end = CH_UART0_RX+1,
532 .end = 0xFFC020FF,
537 .end = IRQ_UART1_RX+1,
542 .end = CH_UART1_RX+1,
561 .end = TWI0_REGBASE,
566 .end = IRQ_TWI,
594 .end = SPORT0_MRCS3+4,
599 .end = IRQ_SPORT0_RX+1,
604 .end = IRQ_SPORT0_ERROR,
623 .end = SPORT1_MRCS3+4,
628 .end = IRQ_SPORT1_RX+1,
633 .end = IRQ_SPORT1_ERROR,
658 .end = SPORT0_MRCS3+4,
663 .end = IRQ_SPORT0_RX+1,
668 .end = IRQ_SPORT0_TX+1,
673 .end = IRQ_SPORT0_ERROR,
678 .end = CH_SPORT0_TX,
683 .end = CH_SPORT0_RX,
741 .end = 0x2030C01F,
746 .end = 0x2030D01B,
751 .end = PATA_INT,
H A Dpnav10.c37 .end = 0x20312000,
41 .end = 0x20311FFF,
45 .end = IRQ_PF4,
49 .end = 6,
82 .end = 0x20300300 + 16,
87 .end = IRQ_PF7,
139 .end = 0x20300000 + 0x100,
143 .end = IRQ_PF7,
261 .end = SPI0_REGBASE + 0xFF,
266 .end = CH_SPI,
271 .end = IRQ_SPI,
305 .end = UART0_GCTL+2,
310 .end = IRQ_UART0_TX,
315 .end = IRQ_UART0_RX,
320 .end = IRQ_UART0_ERROR,
325 .end = CH_UART0_TX,
330 .end = CH_UART0_RX,
353 .end = UART1_GCTL+2,
358 .end = IRQ_UART1_TX,
363 .end = IRQ_UART1_RX,
368 .end = IRQ_UART1_ERROR,
373 .end = CH_UART1_TX,
378 .end = CH_UART1_RX,
404 .end = 0xFFC004FF,
409 .end = IRQ_UART0_RX+1,
414 .end = CH_UART0_RX+1,
430 .end = 0xFFC020FF,
435 .end = IRQ_UART1_RX+1,
440 .end = CH_UART1_RX+1,
H A Dcm_bf537u.c115 .end = SPI0_REGBASE + 0xFF,
120 .end = CH_SPI,
125 .end = IRQ_SPI,
173 .end = 0x20200300 + 16,
177 .end = IRQ_PF14,
197 .end = 0x20308000,
201 .end = 0x20308004,
205 .end = IRQ_PG15,
236 .end = 0x20200000 + 0x100,
240 .end = IRQ_PH14,
282 .end = 0x201fffff,
286 .end = ARRAY_SIZE(cm_flash_gpios),
307 .end = UART0_GCTL+2,
312 .end = IRQ_UART0_TX,
317 .end = IRQ_UART0_RX,
322 .end = IRQ_UART0_ERROR,
327 .end = CH_UART0_TX,
332 .end = CH_UART0_RX,
355 .end = UART1_GCTL+2,
360 .end = IRQ_UART1_TX,
365 .end = IRQ_UART1_RX,
370 .end = IRQ_UART1_ERROR,
375 .end = CH_UART1_TX,
380 .end = CH_UART1_RX,
406 .end = 0xFFC004FF,
411 .end = IRQ_UART0_RX+1,
416 .end = CH_UART0_RX+1,
431 .end = 0xFFC020FF,
436 .end = IRQ_UART1_RX+1,
441 .end = CH_UART1_RX+1,
460 .end = TWI0_REGBASE,
465 .end = IRQ_TWI,
486 .end = SPORT0_MRCS3+4,
491 .end = IRQ_SPORT0_RX+1,
496 .end = IRQ_SPORT0_ERROR,
520 .end = SPORT1_MRCS3+4,
525 .end = IRQ_SPORT1_RX+1,
530 .end = IRQ_SPORT1_ERROR,
595 .end = 0x2030C01F,
600 .end = 0x2030D01B,
605 .end = PATA_INT,
/linux-4.4.14/arch/arm/mm/
H A Dcache-feroceon-l2.c34 * end addresses to successive cp15 registers, and process every
36 * [start:end].
73 static inline void l2_clean_pa_range(unsigned long start, unsigned long end) l2_clean_pa_range() argument
78 * Make sure 'start' and 'end' reference the same page, as l2_clean_pa_range()
82 BUG_ON((start ^ end) >> PAGE_SHIFT); l2_clean_pa_range()
85 va_end = va_start + (end - start); l2_clean_pa_range()
104 static inline void l2_inv_pa_range(unsigned long start, unsigned long end) l2_inv_pa_range() argument
109 * Make sure 'start' and 'end' reference the same page, as l2_inv_pa_range()
113 BUG_ON((start ^ end) >> PAGE_SHIFT); l2_inv_pa_range()
116 va_end = va_start + (end - start); l2_inv_pa_range()
133 * Note that the end addresses passed to Linux primitives are
135 * inclusive start and end addresses.
142 static unsigned long calc_range_end(unsigned long start, unsigned long end) calc_range_end() argument
147 BUG_ON(end & (CACHE_LINE_SIZE - 1)); calc_range_end()
150 * Try to process all cache lines between 'start' and 'end'. calc_range_end()
152 range_end = end; calc_range_end()
171 static void feroceon_l2_inv_range(unsigned long start, unsigned long end) feroceon_l2_inv_range() argument
184 if (start < end && end & (CACHE_LINE_SIZE - 1)) { feroceon_l2_inv_range()
185 l2_clean_inv_pa(end & ~(CACHE_LINE_SIZE - 1)); feroceon_l2_inv_range()
186 end &= ~(CACHE_LINE_SIZE - 1); feroceon_l2_inv_range()
190 * Invalidate all full cache lines between 'start' and 'end'. feroceon_l2_inv_range()
192 while (start < end) { feroceon_l2_inv_range()
193 unsigned long range_end = calc_range_end(start, end); feroceon_l2_inv_range()
201 static void feroceon_l2_clean_range(unsigned long start, unsigned long end) feroceon_l2_clean_range() argument
209 end = (end + CACHE_LINE_SIZE - 1) & ~(CACHE_LINE_SIZE - 1); feroceon_l2_clean_range()
210 while (start != end) { feroceon_l2_clean_range()
211 unsigned long range_end = calc_range_end(start, end); feroceon_l2_clean_range()
220 static void feroceon_l2_flush_range(unsigned long start, unsigned long end) feroceon_l2_flush_range() argument
223 end = (end + CACHE_LINE_SIZE - 1) & ~(CACHE_LINE_SIZE - 1); feroceon_l2_flush_range()
224 while (start != end) { feroceon_l2_flush_range()
225 unsigned long range_end = calc_range_end(start, end); feroceon_l2_flush_range()
H A Didmap.c21 static void idmap_add_pmd(pud_t *pud, unsigned long addr, unsigned long end, idmap_add_pmd() argument
46 next = pmd_addr_end(addr, end); idmap_add_pmd()
49 } while (pmd++, addr = next, addr != end); idmap_add_pmd()
52 static void idmap_add_pmd(pud_t *pud, unsigned long addr, unsigned long end, idmap_add_pmd() argument
65 static void idmap_add_pud(pgd_t *pgd, unsigned long addr, unsigned long end, idmap_add_pud() argument
72 next = pud_addr_end(addr, end); idmap_add_pud()
74 } while (pud++, addr = next, addr != end); idmap_add_pud()
80 unsigned long addr, end; identity_mapping_add() local
84 end = virt_to_idmap(text_end); identity_mapping_add()
85 pr_info("Setting up static identity map for 0x%lx - 0x%lx\n", addr, end); identity_mapping_add()
94 next = pgd_addr_end(addr, end); identity_mapping_add()
96 } while (pgd++, addr = next, addr != end); identity_mapping_add()
H A Dcache-v4wt.S78 * flush_user_cache_range(start, end, flags)
84 * - end - end address (exclusive, page aligned)
101 * coherent_kern_range(start, end)
108 * - end - virtual end address
114 * coherent_user_range(start, end)
121 * - end - virtual end address
148 * dma_inv_range(start, end)
151 * May not write back any entries. If 'start' or 'end'
156 * - end - virtual end address
167 * dma_flush_range(start, end)
172 * - end - virtual end address
H A Dcache-fa.S78 * flush_user_cache_range(start, end, flags)
84 * - end - end address (exclusive, page aligned)
106 * coherent_kern_range(start, end)
113 * - end - virtual end address
119 * coherent_user_range(start, end)
126 * - end - virtual end address
162 * dma_inv_range(start, end)
165 * May not write back any entries. If 'start' or 'end'
170 * - end - virtual end address
188 * dma_clean_range(start, end)
193 * - end - virtual end address
206 * dma_flush_range(start,end)
208 * - end - virtual end address of region
H A Dcache-v4wb.S101 * flush_user_cache_range(start, end, flags)
107 * - end - end address (exclusive, page aligned)
142 * coherent_kern_range(start, end)
149 * - end - virtual end address
155 * coherent_user_range(start, end)
162 * - end - virtual end address
178 * dma_inv_range(start, end)
181 * May not write back any entries. If 'start' or 'end'
186 * - end - virtual end address
202 * dma_clean_range(start, end)
207 * - end - virtual end address
219 * dma_flush_range(start, end)
224 * - end - virtual end address
/linux-4.4.14/tools/perf/util/
H A Dsvghelper.h6 extern void open_svg(const char *filename, int cpus, int rows, u64 start, u64 end);
7 extern void svg_ubox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int merges);
8 extern void svg_lbox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int merges);
9 extern void svg_fbox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int merges);
10 extern void svg_box(int Yslot, u64 start, u64 end, const char *type);
11 extern void svg_blocked(int Yslot, int cpu, u64 start, u64 end, const char *backtrace);
12 extern void svg_running(int Yslot, int cpu, u64 start, u64 end, const char *backtrace);
13 extern void svg_waiting(int Yslot, int cpu, u64 start, u64 end, const char *backtrace);
17 extern void svg_process(int cpu, u64 start, u64 end, int pid, const char *name, const char *backtrace);
18 extern void svg_cstate(int cpu, u64 start, u64 end, int type);
19 extern void svg_pstate(int cpu, u64 start, u64 end, u64 freq);
H A Dfind-vdso-map.c1 static int find_vdso_map(void **start, void **end) find_vdso_map() argument
18 start, end, &m)) find_vdso_map()
/linux-4.4.14/arch/nios2/mm/
H A Dcacheflush.c18 static void __flush_dcache(unsigned long start, unsigned long end) __flush_dcache() argument
23 end += (cpuinfo.dcache_line_size - 1); __flush_dcache()
24 end &= ~(cpuinfo.dcache_line_size - 1); __flush_dcache()
26 if (end > start + cpuinfo.dcache_size) __flush_dcache()
27 end = start + cpuinfo.dcache_size; __flush_dcache()
29 for (addr = start; addr < end; addr += cpuinfo.dcache_line_size) { __flush_dcache()
37 static void __invalidate_dcache(unsigned long start, unsigned long end) __invalidate_dcache() argument
42 end += (cpuinfo.dcache_line_size - 1); __invalidate_dcache()
43 end &= ~(cpuinfo.dcache_line_size - 1); __invalidate_dcache()
45 for (addr = start; addr < end; addr += cpuinfo.dcache_line_size) { __invalidate_dcache()
53 static void __flush_icache(unsigned long start, unsigned long end) __flush_icache() argument
58 end += (cpuinfo.icache_line_size - 1); __flush_icache()
59 end &= ~(cpuinfo.icache_line_size - 1); __flush_icache()
61 if (end > start + cpuinfo.icache_size) __flush_icache()
62 end = start + cpuinfo.icache_size; __flush_icache()
64 for (addr = start; addr < end; addr += cpuinfo.icache_line_size) { __flush_icache()
113 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument
115 __flush_dcache(start, end); flush_icache_range()
116 __flush_icache(start, end); flush_icache_range()
119 void flush_dcache_range(unsigned long start, unsigned long end) flush_dcache_range() argument
121 __flush_dcache(start, end); flush_dcache_range()
122 __flush_icache(start, end); flush_dcache_range()
126 void invalidate_dcache_range(unsigned long start, unsigned long end) invalidate_dcache_range() argument
128 __invalidate_dcache(start, end); invalidate_dcache_range()
133 unsigned long end) flush_cache_range()
135 __flush_dcache(start, end); flush_cache_range()
137 __flush_icache(start, end); flush_cache_range()
143 unsigned long end = start + PAGE_SIZE; flush_icache_page() local
145 __flush_dcache(start, end); flush_icache_page()
146 __flush_icache(start, end); flush_icache_page()
153 unsigned long end = start + PAGE_SIZE; flush_cache_page() local
155 __flush_dcache(start, end); flush_cache_page()
157 __flush_icache(start, end); flush_cache_page()
132 flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_cache_range() argument
H A Dioremap.c26 unsigned long end; remap_area_pte() local
32 end = address + size; remap_area_pte()
33 if (end > PMD_SIZE) remap_area_pte()
34 end = PMD_SIZE; remap_area_pte()
35 if (address >= end) remap_area_pte()
47 } while (address && (address < end)); remap_area_pte()
54 unsigned long end; remap_area_pmd() local
57 end = address + size; remap_area_pmd()
58 if (end > PGDIR_SIZE) remap_area_pmd()
59 end = PGDIR_SIZE; remap_area_pmd()
61 if (address >= end) remap_area_pmd()
68 remap_area_pte(pte, address, end - address, address + phys_addr, remap_area_pmd()
72 } while (address && (address < end)); remap_area_pmd()
81 unsigned long end = address + size; remap_area_pages() local
86 if (address >= end) remap_area_pages()
99 if (remap_area_pmd(pmd, address, end - address, remap_area_pages()
105 } while (address && (address < end)); remap_area_pages()
/linux-4.4.14/arch/mips/include/asm/
H A Dtlbflush.h12 * - flush_tlb_range(vma, start, end) flushes a range of pages
13 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages
18 unsigned long start, unsigned long end);
20 unsigned long end);
39 #define flush_tlb_range(vma, vmaddr, end) local_flush_tlb_range(vma, vmaddr, end)
40 #define flush_tlb_kernel_range(vmaddr,end) \
41 local_flush_tlb_kernel_range(vmaddr, end)
/linux-4.4.14/arch/nios2/include/asm/
H A Dcacheflush.h27 unsigned long end);
33 extern void flush_icache_range(unsigned long start, unsigned long end);
36 #define flush_cache_vmap(start, end) flush_dcache_range(start, end)
37 #define flush_cache_vunmap(start, end) flush_dcache_range(start, end)
46 extern void flush_dcache_range(unsigned long start, unsigned long end);
47 extern void invalidate_dcache_range(unsigned long start, unsigned long end);
/linux-4.4.14/arch/cris/mm/
H A Dinit.c38 void free_init_pages(const char *what, unsigned long begin, unsigned long end) free_init_pages() argument
42 for (addr = begin; addr < end; addr += PAGE_SIZE) { free_init_pages()
49 printk(KERN_INFO "Freeing %s: %ldk freed\n", what, (end - begin) >> 10); free_init_pages()
62 void free_initrd_mem(unsigned long start, unsigned long end) free_initrd_mem() argument
66 end); free_initrd_mem()
/linux-4.4.14/arch/arm/mach-imx/devices/
H A Dplatform-imx-dma.c17 .end = iobase + SZ_4K - 1, imx_add_imx_dma()
21 .end = irq, imx_add_imx_dma()
25 .end = irq_err, imx_add_imx_dma()
40 .end = iobase + SZ_16K - 1, imx_add_imx_sdma()
44 .end = irq, imx_add_imx_sdma()
H A Dplatform-gpio-mxc.c17 .end = iobase + iosize - 1, mxc_register_gpio()
21 .end = irq, mxc_register_gpio()
25 .end = irq_high, mxc_register_gpio()
H A Dplatform-mx2-camera.c42 .end = data->iobasecsi + data->iosizecsi - 1, imx_add_mx2_camera()
46 .end = data->irqcsi, imx_add_mx2_camera()
50 .end = data->iobaseemmaprp + data->iosizeemmaprp - 1, imx_add_mx2_camera()
54 .end = data->irqemmaprp, imx_add_mx2_camera()
/linux-4.4.14/arch/arm/mach-iop13xx/
H A Dtpmi.c46 .end = IOP13XX_TPMI_MMR(4) + IOP13XX_TPMI_MMR_SIZE,
51 .end = IOP13XX_TPMI_MEM(0) + IOP13XX_TPMI_MEM_SIZE,
56 .end = IOP13XX_TPMI_CTRL(0) + IOP13XX_TPMI_MEM_CTRL,
61 .end = IOP13XX_TPMI_IOP_CTRL(0) + IOP13XX_TPMI_MEM_CTRL,
66 .end = IRQ_IOP13XX_TPMI0_OUT,
74 .end = IOP13XX_TPMI_MMR(1) + IOP13XX_TPMI_MMR_SIZE,
79 .end = IOP13XX_TPMI_MEM(1) + IOP13XX_TPMI_MEM_SIZE,
84 .end = IOP13XX_TPMI_CTRL(1) + IOP13XX_TPMI_MEM_CTRL,
89 .end = IOP13XX_TPMI_IOP_CTRL(1) + IOP13XX_TPMI_MEM_CTRL,
94 .end = IRQ_IOP13XX_TPMI1_OUT,
102 .end = IOP13XX_TPMI_MMR(2) + IOP13XX_TPMI_MMR_SIZE,
107 .end = IOP13XX_TPMI_MEM(2) + IOP13XX_TPMI_MEM_SIZE,
112 .end = IOP13XX_TPMI_CTRL(2) + IOP13XX_TPMI_MEM_CTRL,
117 .end = IOP13XX_TPMI_IOP_CTRL(2) + IOP13XX_TPMI_MEM_CTRL,
122 .end = IRQ_IOP13XX_TPMI2_OUT,
130 .end = IOP13XX_TPMI_MMR(3) + IOP13XX_TPMI_MMR_SIZE,
135 .end = IOP13XX_TPMI_MEM(3) + IOP13XX_TPMI_MEM_SIZE,
140 .end = IOP13XX_TPMI_CTRL(3) + IOP13XX_TPMI_MEM_CTRL,
145 .end = IOP13XX_TPMI_IOP_CTRL(3) + IOP13XX_TPMI_MEM_CTRL,
150 .end = IRQ_IOP13XX_TPMI3_OUT,
H A Dsetup.c51 .end = IOP13XX_UART0_PHYS + 0x3f,
56 .end = IRQ_IOP13XX_UART0,
64 .end = IOP13XX_UART1_PHYS + 0x3f,
69 .end = IRQ_IOP13XX_UART1,
120 .end = IOP13XX_I2C0_PHYS + 0x18,
125 .end = IRQ_IOP13XX_I2C_0,
133 .end = IOP13XX_I2C1_PHYS + 0x18,
138 .end = IRQ_IOP13XX_I2C_1,
146 .end = IOP13XX_I2C2_PHYS + 0x18,
151 .end = IRQ_IOP13XX_I2C_2,
191 .end = 0,
237 .end = IOP13XX_ADMA_UPPER_PA(0),
242 .end = IRQ_IOP13XX_ADMA0_EOT,
247 .end = IRQ_IOP13XX_ADMA0_EOC,
252 .end = IRQ_IOP13XX_ADMA0_ERR,
260 .end = IOP13XX_ADMA_UPPER_PA(1),
265 .end = IRQ_IOP13XX_ADMA1_EOT,
270 .end = IRQ_IOP13XX_ADMA1_EOC,
275 .end = IRQ_IOP13XX_ADMA1_ERR,
283 .end = IOP13XX_ADMA_UPPER_PA(2),
288 .end = IRQ_IOP13XX_ADMA2_EOT,
293 .end = IRQ_IOP13XX_ADMA2_EOC,
298 .end = IRQ_IOP13XX_ADMA2_ERR,
500 iq8134x_flash_resource.end = iq8134x_flash_resource.start + iop13xx_platform_init()
502 if (iq8134x_flash_resource.end > iq8134x_flash_resource.start) iop13xx_platform_init()
/linux-4.4.14/drivers/pci/hotplug/
H A Drpadlpar_sysfs.c36 char *end; add_slot_store() local
44 end = strchr(drc_name, '\n'); add_slot_store()
45 if (!end) add_slot_store()
46 end = &drc_name[nbytes]; add_slot_store()
47 *end = '\0'; add_slot_store()
68 char *end; remove_slot_store() local
75 end = strchr(drc_name, '\n'); remove_slot_store()
76 if (!end) remove_slot_store()
77 end = &drc_name[nbytes]; remove_slot_store()
78 *end = '\0'; remove_slot_store()
/linux-4.4.14/arch/x86/include/asm/trace/
H A Dmpx.h69 unsigned long end),
70 TP_ARGS(start, end),
74 __field(unsigned long, end)
79 __entry->end = end;
84 (void *)__entry->end
89 TP_PROTO(unsigned long start, unsigned long end),
90 TP_ARGS(start, end)
94 TP_PROTO(unsigned long start, unsigned long end),
95 TP_ARGS(start, end)
/linux-4.4.14/arch/x86/mm/
H A Dkasan_init_64.c18 unsigned long end; map_range() local
21 end = (unsigned long)kasan_mem_to_shadow(pfn_to_kaddr(range->end)); map_range()
24 * end + 1 here is intentional. We check several shadow bytes in advance map_range()
28 return vmemmap_populate(start, end + 1, NUMA_NO_NODE); map_range()
32 unsigned long end) clear_pgds()
34 for (; start < end; start += PGDIR_SIZE) clear_pgds()
42 unsigned long end = KASAN_SHADOW_END; kasan_map_early_shadow() local
44 for (i = pgd_index(start); start < end; i++) { kasan_map_early_shadow()
106 if (pfn_mapped[i].end == 0) kasan_init()
31 clear_pgds(unsigned long start, unsigned long end) clear_pgds() argument
H A Dnuma_emulation.c31 static u64 __init mem_hole_size(u64 start, u64 end) mem_hole_size() argument
34 unsigned long end_pfn = PFN_DOWN(end); mem_hole_size()
42 * Sets up nid to range from @start to @end. The return value is -errno if
59 eb->end = pb->start + size; emu_setup_memblk()
66 if (pb->start >= pb->end) { emu_setup_memblk()
67 WARN_ON_ONCE(pb->start > pb->end); emu_setup_memblk()
72 nid, eb->start, eb->end - 1, (eb->end - eb->start) >> 20); emu_setup_memblk()
129 u64 start, limit, end; for_each_node_mask() local
138 limit = pi->blk[phys_blk].end; for_each_node_mask()
139 end = start + size; for_each_node_mask()
142 end += FAKE_NODE_MIN_SIZE; for_each_node_mask()
148 while (end - start - mem_hole_size(start, end) < size) { for_each_node_mask()
149 end += FAKE_NODE_MIN_SIZE; for_each_node_mask()
150 if (end > limit) { for_each_node_mask()
151 end = limit; for_each_node_mask()
161 if (end < dma32_end && dma32_end - end - for_each_node_mask()
162 mem_hole_size(end, dma32_end) < FAKE_NODE_MIN_SIZE) for_each_node_mask()
163 end = dma32_end; for_each_node_mask()
167 * next node, this one must extend to the end of the for_each_node_mask()
170 if (limit - end - mem_hole_size(end, limit) < size) for_each_node_mask()
171 end = limit; for_each_node_mask()
175 min(end, limit) - start); for_each_node_mask()
184 * Returns the end address of a node so that there is at least `size' amount of
189 u64 end = start + size; find_end_of_node() local
191 while (end - start - mem_hole_size(start, end) < size) { find_end_of_node()
192 end += FAKE_NODE_MIN_SIZE; find_end_of_node()
193 if (end > max_addr) { find_end_of_node()
194 end = max_addr; find_end_of_node()
198 return end; find_end_of_node()
244 u64 start, limit, end; for_each_node_mask() local
253 limit = pi->blk[phys_blk].end; for_each_node_mask()
255 end = find_end_of_node(start, limit, size); for_each_node_mask()
261 if (end < dma32_end && dma32_end - end - for_each_node_mask()
262 mem_hole_size(end, dma32_end) < FAKE_NODE_MIN_SIZE) for_each_node_mask()
263 end = dma32_end; for_each_node_mask()
267 * next node, this one must extend to the end of the for_each_node_mask()
270 if (limit - end - mem_hole_size(end, limit) < size) for_each_node_mask()
271 end = limit; for_each_node_mask()
275 min(end, limit) - start); for_each_node_mask()
H A Dpat_rbtree.c39 static int is_node_overlap(struct memtype *node, u64 start, u64 end) is_node_overlap() argument
41 if (node->start >= end || node->end <= start) is_node_overlap()
59 u64 max_end = data->end, child_max_end; compute_subtree_max_end()
77 u64 start, u64 end) memtype_rb_lowest_match()
88 } else if (is_node_overlap(data, start, end)) { memtype_rb_lowest_match()
102 u64 start, u64 end) memtype_rb_exact_match()
106 match = memtype_rb_lowest_match(root, start, end); memtype_rb_exact_match()
107 while (match != NULL && match->start < end) { memtype_rb_exact_match()
110 if (match->start == start && match->end == end) memtype_rb_exact_match()
124 u64 start, u64 end, memtype_rb_check_conflict()
132 match = memtype_rb_lowest_match(&memtype_rbroot, start, end); memtype_rb_check_conflict()
139 dprintk("Overlap at 0x%Lx-0x%Lx\n", match->start, match->end); memtype_rb_check_conflict()
146 if (match->start >= end) /* Checked all possible matches */ memtype_rb_check_conflict()
149 if (is_node_overlap(match, start, end) && memtype_rb_check_conflict()
164 current->comm, current->pid, start, end, memtype_rb_check_conflict()
178 if (data->subtree_max_end < newdata->end) memtype_rb_insert()
179 data->subtree_max_end = newdata->end; memtype_rb_insert()
186 newdata->subtree_max_end = newdata->end; memtype_rb_insert()
196 err = memtype_rb_check_conflict(&memtype_rbroot, new->start, new->end, rbt_memtype_check_insert()
203 new->subtree_max_end = new->end; rbt_memtype_check_insert()
209 struct memtype *rbt_memtype_erase(u64 start, u64 end) rbt_memtype_erase() argument
213 data = memtype_rb_exact_match(&memtype_rbroot, start, end); rbt_memtype_erase()
76 memtype_rb_lowest_match(struct rb_root *root, u64 start, u64 end) memtype_rb_lowest_match() argument
101 memtype_rb_exact_match(struct rb_root *root, u64 start, u64 end) memtype_rb_exact_match() argument
123 memtype_rb_check_conflict(struct rb_root *root, u64 start, u64 end, enum page_cache_mode reqtype, enum page_cache_mode *newtype) memtype_rb_check_conflict() argument
H A Dinit_64.c60 unsigned long addr, unsigned long end) ident_pmd_init()
63 for (; addr < end; addr += PMD_SIZE) { ident_pmd_init()
71 unsigned long addr, unsigned long end) ident_pud_init()
75 for (; addr < end; addr = next) { ident_pud_init()
80 if (next > end) ident_pud_init()
81 next = end; ident_pud_init()
99 unsigned long addr, unsigned long end) kernel_ident_mapping_init()
105 for (; addr < end; addr = next) { kernel_ident_mapping_init()
110 if (next > end) kernel_ident_mapping_init()
111 next = end; kernel_ident_mapping_init()
166 void sync_global_pgds(unsigned long start, unsigned long end, int removed) sync_global_pgds() argument
170 for (address = start; address <= end; address += PGDIR_SIZE) { sync_global_pgds()
382 unsigned long end = roundup((unsigned long)_brk_end, PMD_SIZE) - 1; cleanup_highmap() local
396 if (vaddr < (unsigned long) _text || vaddr > end) cleanup_highmap()
402 phys_pte_init(pte_t *pte_page, unsigned long addr, unsigned long end, phys_pte_init() argument
406 unsigned long last_map_addr = end; phys_pte_init()
413 if (addr >= end) { phys_pte_init()
447 phys_pmd_init(pmd_t *pmd_page, unsigned long address, unsigned long end, phys_pmd_init() argument
451 unsigned long last_map_addr = end; phys_pmd_init()
461 if (address >= end) { phys_pmd_init()
474 end, prot); phys_pmd_init()
511 last_map_addr = phys_pte_init(pte, address, end, new_prot); phys_pmd_init()
522 phys_pud_init(pud_t *pud_page, unsigned long addr, unsigned long end, phys_pud_init() argument
526 unsigned long last_map_addr = end; phys_pud_init()
535 if (addr >= end) { phys_pud_init()
546 last_map_addr = phys_pmd_init(pmd, addr, end, phys_pud_init()
584 last_map_addr = phys_pmd_init(pmd, addr, end, page_size_mask, phys_pud_init()
600 unsigned long end, kernel_physical_mapping_init()
604 unsigned long next, last_map_addr = end; kernel_physical_mapping_init()
608 end = (unsigned long)__va(end); kernel_physical_mapping_init()
611 for (; start < end; start = next) { kernel_physical_mapping_init()
620 __pa(end), page_size_mask); kernel_physical_mapping_init()
625 last_map_addr = phys_pud_init(pud, __pa(start), __pa(end), kernel_physical_mapping_init()
635 sync_global_pgds(addr, end - 1, 0); kernel_physical_mapping_init()
791 remove_pte_table(pte_t *pte_start, unsigned long addr, unsigned long end, remove_pte_table() argument
800 for (; addr < end; addr = next, pte++) { remove_pte_table()
802 if (next > end) remove_pte_table()
803 next = end; remove_pte_table()
863 remove_pmd_table(pmd_t *pmd_start, unsigned long addr, unsigned long end, remove_pmd_table() argument
872 for (; addr < end; addr = next, pmd++) { remove_pmd_table()
873 next = pmd_addr_end(addr, end); remove_pmd_table()
919 remove_pud_table(pud_t *pud_start, unsigned long addr, unsigned long end, remove_pud_table() argument
928 for (; addr < end; addr = next, pud++) { remove_pud_table()
929 next = pud_addr_end(addr, end); remove_pud_table()
973 /* start and end are both virtual address. */
975 remove_pagetable(unsigned long start, unsigned long end, bool direct) remove_pagetable() argument
983 for (addr = start; addr < end; addr = next) { remove_pagetable()
984 next = pgd_addr_end(addr, end); remove_pagetable()
997 sync_global_pgds(start, end - 1, 1); remove_pagetable()
1002 void __ref vmemmap_free(unsigned long start, unsigned long end) vmemmap_free() argument
1004 remove_pagetable(start, end, false); vmemmap_free()
1009 kernel_physical_mapping_remove(unsigned long start, unsigned long end) kernel_physical_mapping_remove() argument
1012 end = (unsigned long)__va(end); kernel_physical_mapping_remove()
1014 remove_pagetable(start, end, true); kernel_physical_mapping_remove()
1074 unsigned long end = PFN_ALIGN(__stop___ex_table); set_kernel_text_rw() local
1080 start, end); set_kernel_text_rw()
1087 set_memory_rw(start, (end - start) >> PAGE_SHIFT); set_kernel_text_rw()
1093 unsigned long end = PFN_ALIGN(__stop___ex_table); set_kernel_text_ro() local
1099 start, end); set_kernel_text_ro()
1104 set_memory_ro(start, (end - start) >> PAGE_SHIFT); set_kernel_text_ro()
1111 unsigned long end = (unsigned long) &__end_rodata_hpage_align; mark_rodata_ro() local
1117 (end - start) >> 10); mark_rodata_ro()
1118 set_memory_ro(start, (end - start) >> PAGE_SHIFT); mark_rodata_ro()
1128 * split the PMD and the reminder between _brk_end and the end mark_rodata_ro()
1140 printk(KERN_INFO "Testing CPA: undo %lx-%lx\n", start, end); mark_rodata_ro()
1141 set_memory_rw(start, (end-start) >> PAGE_SHIFT); mark_rodata_ro()
1144 set_memory_ro(start, (end-start) >> PAGE_SHIFT); mark_rodata_ro()
1239 unsigned long end, int node) vmemmap_populate_hugepages()
1247 for (addr = start; addr < end; addr = next) { vmemmap_populate_hugepages()
1248 next = pmd_addr_end(addr, end); vmemmap_populate_hugepages()
1295 int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node) vmemmap_populate() argument
1300 err = vmemmap_populate_hugepages(start, end, node); vmemmap_populate()
1302 err = vmemmap_populate_basepages(start, end, node); vmemmap_populate()
1304 sync_global_pgds(start, end - 1, 0); vmemmap_populate()
1313 unsigned long end = (unsigned long)(start_page + size); register_page_bootmem_memmap() local
1321 for (; addr < end; addr = next) { register_page_bootmem_memmap()
1352 next = pmd_addr_end(addr, end); register_page_bootmem_memmap()
59 ident_pmd_init(unsigned long pmd_flag, pmd_t *pmd_page, unsigned long addr, unsigned long end) ident_pmd_init() argument
70 ident_pud_init(struct x86_mapping_info *info, pud_t *pud_page, unsigned long addr, unsigned long end) ident_pud_init() argument
98 kernel_ident_mapping_init(struct x86_mapping_info *info, pgd_t *pgd_page, unsigned long addr, unsigned long end) kernel_ident_mapping_init() argument
599 kernel_physical_mapping_init(unsigned long start, unsigned long end, unsigned long page_size_mask) kernel_physical_mapping_init() argument
1238 vmemmap_populate_hugepages(unsigned long start, unsigned long end, int node) vmemmap_populate_hugepages() argument
H A Dinit.c145 unsigned long end; member in struct:map_range
197 mr[nr_range].end = end_pfn<<PAGE_SHIFT; save_mr()
218 unsigned long end = round_up(mr[i].end, PMD_SIZE); adjust_range_page_size_mask() local
221 if ((end >> PAGE_SHIFT) > max_low_pfn) adjust_range_page_size_mask()
225 if (memblock_is_region_memory(start, end - start)) adjust_range_page_size_mask()
231 unsigned long end = round_up(mr[i].end, PUD_SIZE); adjust_range_page_size_mask() local
233 if (memblock_is_region_memory(start, end - start)) adjust_range_page_size_mask()
266 unsigned long end) split_mem_range()
272 limit_pfn = PFN_DOWN(end); split_mem_range()
345 if (mr[i].end != mr[i+1].start || split_mem_range()
358 mr[i].start, mr[i].end - 1, split_mem_range()
386 (end_pfn <= pfn_mapped[i].end)) pfn_range_is_mapped()
398 unsigned long end) init_memory_mapping()
405 start, end - 1); init_memory_mapping()
408 nr_range = split_mem_range(mr, 0, start, end); init_memory_mapping()
411 ret = kernel_physical_mapping_init(mr[i].start, mr[i].end, init_memory_mapping()
442 u64 end = clamp_val(PFN_PHYS(end_pfn), r_start, r_end); init_range_memory_mapping() local
443 if (start >= end) init_range_memory_mapping()
451 min(end, (u64)pgt_buf_top<<PAGE_SHIFT); init_range_memory_mapping()
452 init_memory_mapping(start, end); init_range_memory_mapping()
453 mapped_ram_size += end - start; init_range_memory_mapping()
482 * @map_end: end address of the target memory range
486 * will be allocated at the end of the memory, and we map the
497 /* xen has big range in reserved near end of ram, skip it at first.*/ memory_map_top_down()
508 * We start from the top (end of memory) and go to the bottom. memory_map_top_down()
510 * end of RAM in [min_pfn_mapped, max_pfn_mapped) used as new pages memory_map_top_down()
535 * @map_end: end address of the target memory range
557 * end of RAM in [min_pfn_mapped, max_pfn_mapped) used as new pages memory_map_bottom_up()
579 unsigned long end; init_mem_mapping() local
584 end = max_pfn << PAGE_SHIFT; init_mem_mapping()
586 end = max_low_pfn << PAGE_SHIFT; init_mem_mapping()
602 * [kernel_end, end) to make memory above the kernel be mapped init_mem_mapping()
606 memory_map_bottom_up(kernel_end, end); init_mem_mapping()
609 memory_map_top_down(ISA_END_ADDRESS, end); init_mem_mapping()
648 void free_init_pages(char *what, unsigned long begin, unsigned long end) free_init_pages() argument
654 end_aligned = end & PAGE_MASK; free_init_pages()
656 if (WARN_ON(begin_aligned != begin || end_aligned != end)) { free_init_pages()
658 end = end_aligned; free_init_pages()
661 if (begin >= end) free_init_pages()
671 begin, end - 1); free_init_pages()
672 set_memory_np(begin, (end - begin) >> PAGE_SHIFT); free_init_pages()
679 set_memory_nx(begin, (end - begin) >> PAGE_SHIFT); free_init_pages()
680 set_memory_rw(begin, (end - begin) >> PAGE_SHIFT); free_init_pages()
682 free_reserved_area((void *)begin, (void *)end, POISON_FREE_INITMEM, what); free_init_pages()
694 void __init free_initrd_mem(unsigned long start, unsigned long end) free_initrd_mem() argument
704 * end could be not aligned, and We can not align that, free_initrd_mem()
706 * We already reserve the end partial page before in free_initrd_mem()
712 free_init_pages("initrd", start, PAGE_ALIGN(end)); free_initrd_mem()
264 split_mem_range(struct map_range *mr, int nr_range, unsigned long start, unsigned long end) split_mem_range() argument
397 init_memory_mapping(unsigned long start, unsigned long end) init_memory_mapping() argument
H A Dpat_internal.h11 u64 end; member in struct:memtype
33 extern struct memtype *rbt_memtype_erase(u64 start, u64 end);
40 static inline struct memtype *rbt_memtype_erase(u64 start, u64 end) rbt_memtype_erase() argument
/linux-4.4.14/arch/mips/ath79/
H A Ddev-wmac.c29 /* .start and .end fields are filled dynamically */
32 /* .start and .end fields are filled dynamically */
57 ath79_wmac_resources[0].end = AR913X_WMAC_BASE + AR913X_WMAC_SIZE - 1; ar913x_wmac_setup()
59 ath79_wmac_resources[1].end = ATH79_CPU_IRQ(2); ar913x_wmac_setup()
85 ath79_wmac_resources[0].end = AR933X_WMAC_BASE + AR933X_WMAC_SIZE - 1; ar933x_wmac_setup()
87 ath79_wmac_resources[1].end = ATH79_CPU_IRQ(2); ar933x_wmac_setup()
108 ath79_wmac_resources[0].end = AR934X_WMAC_BASE + AR934X_WMAC_SIZE - 1; ar934x_wmac_setup()
110 ath79_wmac_resources[1].end = ATH79_IP2_IRQ(1); ar934x_wmac_setup()
126 ath79_wmac_resources[0].end = QCA955X_WMAC_BASE + QCA955X_WMAC_SIZE - 1; qca955x_wmac_setup()
128 ath79_wmac_resources[1].end = ATH79_IP2_IRQ(1); qca955x_wmac_setup()
/linux-4.4.14/arch/hexagon/mm/
H A Dcache.c25 #define spanlines(start, end) \
26 (((end - (start & ~(LINESIZE - 1))) >> LINEBITS) + 1)
28 void flush_dcache_range(unsigned long start, unsigned long end) flush_dcache_range() argument
30 unsigned long lines = spanlines(start, end-1); flush_dcache_range()
48 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument
50 unsigned long lines = spanlines(start, end-1); flush_icache_range()
73 void hexagon_clean_dcache_range(unsigned long start, unsigned long end) hexagon_clean_dcache_range() argument
75 unsigned long lines = spanlines(start, end-1); hexagon_clean_dcache_range()
93 void hexagon_inv_dcache_range(unsigned long start, unsigned long end) hexagon_inv_dcache_range() argument
95 unsigned long lines = spanlines(start, end-1); hexagon_inv_dcache_range()
/linux-4.4.14/arch/blackfin/mach-bf548/boards/
H A Dezkit.c49 .end = 0x2C0C0000 + 0xfffff,
54 .end = IRQ_PG7,
95 .end = IRQ_EPPI0_ERR,
145 .end = IRQ_KEY,
177 .end = CNT_CONFIG + 0xff,
182 .end = IRQ_CNT,
249 .end = UART0_RBR+2,
255 .end = PORTE_FER+2,
261 .end = IRQ_UART0_TX,
266 .end = IRQ_UART0_RX,
271 .end = IRQ_UART0_ERROR,
276 .end = CH_UART0_TX,
281 .end = CH_UART0_RX,
304 .end = UART1_RBR+2,
310 .end = PORTH_FER+2,
316 .end = IRQ_UART1_TX,
321 .end = IRQ_UART1_RX,
326 .end = IRQ_UART1_ERROR,
331 .end = CH_UART1_TX,
336 .end = CH_UART1_RX,
342 .end = GPIO_PE10,
347 .end = GPIO_PE9,
375 .end = UART2_RBR+2,
381 .end = PORTB_FER+2,
387 .end = IRQ_UART2_TX,
392 .end = IRQ_UART2_RX,
397 .end = IRQ_UART2_ERROR,
402 .end = CH_UART2_TX,
407 .end = CH_UART2_RX,
430 .end = UART3_RBR+2,
436 .end = PORTB_FER+2,
442 .end = IRQ_UART3_TX,
447 .end = IRQ_UART3_RX,
452 .end = IRQ_UART3_ERROR,
457 .end = CH_UART3_TX,
462 .end = CH_UART3_RX,
468 .end = GPIO_PB3,
473 .end = GPIO_PB2,
504 .end = 0xFFC004FF,
509 .end = IRQ_UART0_RX+1,
514 .end = CH_UART0_RX+1,
529 .end = 0xFFC020FF,
534 .end = IRQ_UART1_RX+1,
539 .end = CH_UART1_RX+1,
554 .end = 0xFFC021FF,
559 .end = IRQ_UART2_RX+1,
564 .end = CH_UART2_RX+1,
579 .end = 0xFFC031FF,
584 .end = IRQ_UART3_RX+1,
589 .end = CH_UART3_RX+1,
609 .end = 0x24000000 + 0xFF,
614 .end = IRQ_PE8,
641 .end = 0xFFC040FF,
646 .end = IRQ_USB_INT0,
652 .end = IRQ_USB_DMA,
704 .end = SPORT0_MRCS3+4,
709 .end = IRQ_SPORT0_RX+1,
714 .end = IRQ_SPORT0_ERROR,
738 .end = SPORT1_MRCS3+4,
743 .end = IRQ_SPORT1_RX+1,
748 .end = IRQ_SPORT1_ERROR,
772 .end = SPORT2_MRCS3+4,
777 .end = IRQ_SPORT2_RX+1,
782 .end = IRQ_SPORT2_ERROR,
806 .end = SPORT3_MRCS3+4,
811 .end = IRQ_SPORT3_RX+1,
816 .end = IRQ_SPORT3_ERROR,
847 .end = 0xFFC02FFF,
852 .end = IRQ_CAN0_RX,
857 .end = IRQ_CAN0_TX,
862 .end = IRQ_CAN0_ERROR,
884 .end = 0xFFC037FF,
889 .end = IRQ_CAN1_RX,
894 .end = IRQ_CAN1_TX,
899 .end = IRQ_CAN1_ERROR,
920 .end = 0xFFC0386F,
925 .end = IRQ_ATAPI_ERR,
967 .end = 0xFFC03B4F,
972 .end = CH_NFC,
1038 .end = 0x21ffffff,
1110 .end = PINT0_LATCH + 3,
1115 .end = IRQ_PINT0,
1130 .end = PINT1_LATCH + 3,
1135 .end = IRQ_PINT1,
1150 .end = PINT2_LATCH + 3,
1155 .end = IRQ_PINT2,
1170 .end = PINT3_LATCH + 3,
1175 .end = IRQ_PINT3,
1190 .end = PORTA_MUX + 3,
1195 .end = IRQ_PA0,
1222 .end = PORTB_MUX + 3,
1227 .end = IRQ_PB0,
1254 .end = PORTC_MUX + 3,
1259 .end = IRQ_PC0,
1286 .end = PORTD_MUX + 3,
1291 .end = IRQ_PD0,
1318 .end = PORTE_MUX + 3,
1323 .end = IRQ_PE0,
1350 .end = PORTF_MUX + 3,
1355 .end = IRQ_PF0,
1382 .end = PORTG_MUX + 3,
1387 .end = IRQ_PG0,
1412 .end = PORTH_MUX + 3,
1417 .end = IRQ_PH0,
1442 .end = PORTI_MUX + 3,
1447 .end = IRQ_PI0,
1472 .end = PORTJ_MUX + 3,
1477 .end = IRQ_PJ0,
1557 .end = SPI0_REGBASE + 0xFF,
1562 .end = CH_SPI0,
1567 .end = IRQ_SPI0,
1576 .end = SPI1_REGBASE + 0xFF,
1581 .end = CH_SPI1,
1586 .end = IRQ_SPI1,
1696 .end = TWI0_REGBASE + 0xFF,
1701 .end = IRQ_TWI0,
1722 .end = TWI1_REGBASE + 0xFF,
1727 .end = IRQ_TWI1,
1864 .end = SPORT##x##_TCR1, \
1869 .end = CH_SPORT##x##_RX, \
1874 .end = CH_SPORT##x##_TX, \
1879 .end = IRQ_SPORT##x##_ERROR, \
H A Dcm_bf548.c56 .end = IRQ_EPPI0_ERR,
106 .end = IRQ_KEY,
134 .end = UART0_RBR+2,
139 .end = IRQ_UART0_TX,
144 .end = IRQ_UART0_RX,
149 .end = IRQ_UART0_ERROR,
154 .end = CH_UART0_TX,
159 .end = CH_UART0_RX,
182 .end = UART1_RBR+2,
187 .end = IRQ_UART1_TX,
192 .end = IRQ_UART1_RX,
197 .end = IRQ_UART1_ERROR,
202 .end = CH_UART1_TX,
207 .end = CH_UART1_RX,
213 .end = GPIO_PE10,
218 .end = GPIO_PE9,
246 .end = UART2_RBR+2,
251 .end = IRQ_UART2_TX,
256 .end = IRQ_UART2_RX,
261 .end = IRQ_UART2_ERROR,
266 .end = CH_UART2_TX,
271 .end = CH_UART2_RX,
294 .end = UART3_RBR+2,
299 .end = IRQ_UART3_TX,
304 .end = IRQ_UART3_RX,
309 .end = IRQ_UART3_ERROR,
314 .end = CH_UART3_TX,
319 .end = CH_UART3_RX,
325 .end = GPIO_PB3,
330 .end = GPIO_PB2,
361 .end = 0xFFC004FF,
366 .end = IRQ_UART0_RX+1,
371 .end = CH_UART0_RX+1,
386 .end = 0xFFC020FF,
391 .end = IRQ_UART1_RX+1,
396 .end = CH_UART1_RX+1,
411 .end = 0xFFC021FF,
416 .end = IRQ_UART2_RX+1,
421 .end = CH_UART2_RX+1,
436 .end = 0xFFC031FF,
441 .end = IRQ_UART3_RX+1,
446 .end = CH_UART3_RX+1,
466 .end = 0x24000000 + 0xFF,
471 .end = IRQ_PE6,
498 .end = 0xFFC040FF,
503 .end = IRQ_USB_INT0,
509 .end = IRQ_USB_DMA,
561 .end = SPORT0_MRCS3+4,
566 .end = IRQ_SPORT0_RX+1,
571 .end = IRQ_SPORT0_ERROR,
595 .end = SPORT1_MRCS3+4,
600 .end = IRQ_SPORT1_RX+1,
605 .end = IRQ_SPORT1_ERROR,
629 .end = SPORT2_MRCS3+4,
634 .end = IRQ_SPORT2_RX+1,
639 .end = IRQ_SPORT2_ERROR,
663 .end = SPORT3_MRCS3+4,
668 .end = IRQ_SPORT3_RX+1,
673 .end = IRQ_SPORT3_ERROR,
699 .end = 0xFFC0386F,
704 .end = IRQ_ATAPI_ERR,
742 .end = 0xFFC03B4F,
747 .end = CH_NFC,
787 .end = 0xFFC02FFF,
792 .end = IRQ_CAN0_RX,
797 .end = IRQ_CAN0_TX,
802 .end = IRQ_CAN0_ERROR,
842 .end = 0x207fffff,
939 .end = SPI0_REGBASE + 0xFF,
944 .end = CH_SPI0,
949 .end = IRQ_SPI0,
958 .end = SPI1_REGBASE + 0xFF,
963 .end = CH_SPI1,
968 .end = IRQ_SPI1,
1013 .end = TWI0_REGBASE + 0xFF,
1018 .end = IRQ_TWI0,
1039 .end = TWI1_REGBASE + 0xFF,
1044 .end = IRQ_TWI1,
/linux-4.4.14/arch/alpha/mm/
H A Dnuma.c48 printk(" memcluster %2d, usage %1lx, start %8lu, end %8lu\n", for_each_mem_cluster()
62 unsigned long start, end; setup_memory_node() local
88 end = start + cluster->numpages; for_each_mem_cluster()
90 if (start >= node_pfn_end || end <= node_pfn_start) for_each_mem_cluster()
97 printk(" memcluster %2d, usage %1lx, start %8lu, end %8lu\n", for_each_mem_cluster()
103 if (end > node_pfn_end) for_each_mem_cluster()
104 end = node_pfn_end; for_each_mem_cluster()
108 if (end > node_max_pfn) for_each_mem_cluster()
109 node_max_pfn = end; for_each_mem_cluster()
144 printk(" Detected node memory: start %8lu, end %8lu\n",
173 end = start + cluster->numpages; for_each_mem_cluster()
175 if (start >= node_max_pfn || end <= node_min_pfn) for_each_mem_cluster()
178 if (end > node_max_pfn) for_each_mem_cluster()
179 end = node_max_pfn; for_each_mem_cluster()
184 if (end > end_kernel_pfn for_each_mem_cluster()
185 && end - end_kernel_pfn >= bootmap_pages) { for_each_mem_cluster()
188 } else if (end > start_kernel_pfn) for_each_mem_cluster()
189 end = start_kernel_pfn; for_each_mem_cluster()
192 if (end - start >= bootmap_pages) { for_each_mem_cluster()
213 end = cluster->start_pfn + cluster->numpages; for_each_mem_cluster()
215 if (start >= node_max_pfn || end <= node_min_pfn) for_each_mem_cluster()
218 if (end > node_max_pfn) for_each_mem_cluster()
219 end = node_max_pfn; for_each_mem_cluster()
224 if (end > end_kernel_pfn) { for_each_mem_cluster()
231 } else if (end > start_kernel_pfn) for_each_mem_cluster()
232 end = start_kernel_pfn; for_each_mem_cluster()
235 if (start >= end) for_each_mem_cluster()
238 free_bootmem_node(NODE_DATA(nid), PFN_PHYS(start), PFN_PHYS(end) - PFN_PHYS(start)); for_each_mem_cluster()
239 printk(" freeing pages %ld:%ld\n", start, end); for_each_mem_cluster()
275 printk("initrd extends beyond end of memory " setup_memory()
/linux-4.4.14/arch/mips/bcm63xx/
H A Ddev-enet.c49 .end = -1, /* filled at runtime */
54 .end = -1, /* filled at runtime */
59 .end = -1, /* filled at runtime */
76 .end = -1, /* filled at runtime */
108 .end = -1, /* filled at runtime */
139 /* start & end filled at runtime */
173 shared_res[0].end = shared_res[0].start; register_shared()
175 shared_res[0].end += (RSET_6345_ENETDMA_SIZE) - 1; register_shared()
177 shared_res[0].end += (RSET_ENETDMA_SIZE) - 1; register_shared()
187 shared_res[1].end = shared_res[1].start; register_shared()
188 shared_res[1].end += RSET_ENETDMAC_SIZE(chan_count) - 1; register_shared()
191 shared_res[2].end = shared_res[2].start; register_shared()
192 shared_res[2].end += RSET_ENETDMAS_SIZE(chan_count) - 1; register_shared()
221 enet0_res[0].end = enet0_res[0].start; bcm63xx_enet_register()
222 enet0_res[0].end += RSET_ENET_SIZE - 1; bcm63xx_enet_register()
229 enet1_res[0].end = enet1_res[0].start; bcm63xx_enet_register()
230 enet1_res[0].end += RSET_ENET_SIZE - 1; bcm63xx_enet_register()
287 enetsw_res[0].end = enetsw_res[0].start; bcm63xx_enetsw_register()
288 enetsw_res[0].end += RSET_ENETSW_SIZE - 1; bcm63xx_enetsw_register()
H A Ddev-hsspi.c20 .end = -1, /* filled at runtime */
42 spi_resources[0].end = spi_resources[0].start; bcm63xx_hsspi_register()
43 spi_resources[0].end += RSET_HSSPI_SIZE - 1; bcm63xx_hsspi_register()
H A Ddev-rng.c17 .end = -1, /* filled at runtime */
35 rng_resources[0].end = rng_resources[0].start; bcm63xx_rng_register()
36 rng_resources[0].end += RSET_RNG_SIZE - 1; bcm63xx_rng_register()
H A Ddev-wdt.c17 .end = -1, /* filled at runtime */
32 wdt_resources[0].end = wdt_resources[0].start; bcm63xx_wdt_register()
33 wdt_resources[0].end += RSET_WDT_SIZE - 1; bcm63xx_wdt_register()
H A Ddev-spi.c24 .end = -1, /* filled at runtime */
45 spi_resources[0].end = spi_resources[0].start; bcm63xx_spi_register()
50 spi_resources[0].end += BCM_6348_RSET_SPI_SIZE - 1; bcm63xx_spi_register()
56 spi_resources[0].end += BCM_6358_RSET_SPI_SIZE - 1; bcm63xx_spi_register()
H A Ddev-uart.c16 /* start & end filled at runtime */
27 /* start & end filled at runtime */
63 uart0_resources[0].end = uart0_resources[0].start + bcm63xx_uart_register()
70 uart1_resources[0].end = uart1_resources[0].start + bcm63xx_uart_register()
H A Ddev-usb-usbd.c48 usbd_resources[0].end = usbd_resources[0].start + RSET_USBD_SIZE - 1; bcm63xx_usbd_register()
52 usbd_resources[1].end = usbd_resources[1].start + RSET_USBDMA_SIZE - 1; bcm63xx_usbd_register()
58 r->start = r->end = bcm63xx_get_irq_number(irq_list[i]); bcm63xx_usbd_register()
/linux-4.4.14/arch/arm64/mm/
H A Dcache.S30 * flush_icache_range(start,end)
37 * - end - virtual end address of region
43 * __flush_cache_user_range(start,end)
50 * - end - virtual end address of region
104 * __inval_cache_range(start, end)
106 * - end - end address of region
112 * __dma_inv_range(start, end)
114 * - end - virtual end address of region
119 tst x1, x3 // end cache line aligned?
138 * __dma_clean_range(start, end)
140 * - end - virtual end address of region
160 * __dma_flush_range(start, end)
162 * - end - virtual end address of region
H A Dkasan_init.c27 unsigned long end) kasan_early_pte_populate()
40 } while (pte++, addr = next, addr != end && pte_none(*pte)); kasan_early_pte_populate()
45 unsigned long end) kasan_early_pmd_populate()
55 next = pmd_addr_end(addr, end); kasan_early_pmd_populate()
57 } while (pmd++, addr = next, addr != end && pmd_none(*pmd)); kasan_early_pmd_populate()
62 unsigned long end) kasan_early_pud_populate()
72 next = pud_addr_end(addr, end); kasan_early_pud_populate()
74 } while (pud++, addr = next, addr != end && pud_none(*pud)); kasan_early_pud_populate()
80 unsigned long end = KASAN_SHADOW_END; kasan_map_early_shadow() local
86 next = pgd_addr_end(addr, end); kasan_map_early_shadow()
88 } while (pgd++, addr = next, addr != end); kasan_map_early_shadow()
100 unsigned long end) clear_pgds()
107 for (; start < end; start += PGDIR_SIZE) clear_pgds()
142 void *end = (void *)__phys_to_virt(reg->base + reg->size); for_each_memblock() local
144 if (start >= end) for_each_memblock()
148 * end + 1 here is intentional. We check several shadow bytes in for_each_memblock()
154 (unsigned long)kasan_mem_to_shadow(end) + 1, for_each_memblock()
26 kasan_early_pte_populate(pmd_t *pmd, unsigned long addr, unsigned long end) kasan_early_pte_populate() argument
43 kasan_early_pmd_populate(pud_t *pud, unsigned long addr, unsigned long end) kasan_early_pmd_populate() argument
60 kasan_early_pud_populate(pgd_t *pgd, unsigned long addr, unsigned long end) kasan_early_pud_populate() argument
99 clear_pgds(unsigned long start, unsigned long end) clear_pgds() argument
/linux-4.4.14/arch/um/include/asm/
H A Dtlb.h21 unsigned long end; member in struct:mmu_gather
30 if (tlb->end < address + PAGE_SIZE) __tlb_remove_tlb_entry()
31 tlb->end = address + PAGE_SIZE; __tlb_remove_tlb_entry()
39 tlb->end = 0; init_tlb_gather()
43 tlb->end = TASK_SIZE; init_tlb_gather()
48 tlb_gather_mmu(struct mmu_gather *tlb, struct mm_struct *mm, unsigned long start, unsigned long end) tlb_gather_mmu() argument
52 tlb->end = end; tlb_gather_mmu()
53 tlb->fullmm = !(start | (end+1)); tlb_gather_mmu()
59 unsigned long end);
64 flush_tlb_mm_range(tlb->mm, tlb->start, tlb->end); tlb_flush_mmu_tlbonly()
84 * Called at the end of the shootdown operation to free up any resources
88 tlb_finish_mmu(struct mmu_gather *tlb, unsigned long start, unsigned long end) tlb_finish_mmu() argument
H A Dtlbflush.h19 * - flush_tlb_range(vma, start, end) flushes a range of pages
25 unsigned long end);
28 extern void flush_tlb_kernel_range(unsigned long start, unsigned long end);
/linux-4.4.14/lib/
H A Dioremap.c54 unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pte_range()
67 } while (pte++, addr += PAGE_SIZE, addr != end); ioremap_pte_range()
72 unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pmd_range()
82 next = pmd_addr_end(addr, end); ioremap_pmd_range()
93 } while (pmd++, addr = next, addr != end); ioremap_pmd_range()
98 unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pud_range()
108 next = pud_addr_end(addr, end); ioremap_pud_range()
119 } while (pud++, addr = next, addr != end); ioremap_pud_range()
124 unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_page_range()
131 BUG_ON(addr >= end); ioremap_page_range()
137 next = pgd_addr_end(addr, end); ioremap_page_range()
141 } while (pgd++, addr = next, addr != end); ioremap_page_range()
143 flush_cache_vmap(start, end); ioremap_page_range()
53 ioremap_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pte_range() argument
71 ioremap_pmd_range(pud_t *pud, unsigned long addr, unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pmd_range() argument
97 ioremap_pud_range(pgd_t *pgd, unsigned long addr, unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pud_range() argument
123 ioremap_page_range(unsigned long addr, unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_page_range() argument
H A Dvsprintf.c45 * @endp: A pointer to the end of the parsed string will be placed here
70 * @endp: A pointer to the end of the parsed string will be placed here
84 * @endp: A pointer to the end of the parsed string will be placed here
101 * @endp: A pointer to the end of the parsed string will be placed here
392 char *number(char *buf, char *end, unsigned long long num, number() argument
454 if (buf < end) number()
461 if (buf < end) number()
468 if (buf < end) number()
473 if (buf < end) number()
483 if (buf < end) number()
490 if (buf < end) number()
496 if (buf < end) number()
502 if (buf < end) number()
511 char *string(char *buf, char *end, const char *s, struct printf_spec spec) string() argument
522 if (buf < end) string()
528 if (buf < end) string()
533 if (buf < end) string()
541 static void widen(char *buf, char *end, unsigned len, unsigned spaces) widen() argument
544 if (buf >= end) /* nowhere to put anything */ widen()
546 size = end - buf; widen()
560 char *dentry_name(char *buf, char *end, const struct dentry *d, struct printf_spec spec, dentry_name() argument
596 if (buf < end) dentry_name()
604 widen(buf - n, end, n, spaces); dentry_name()
608 if (buf < end) dentry_name()
617 char *symbol_string(char *buf, char *end, void *ptr, symbol_string() argument
637 return string(buf, end, sym, spec); symbol_string()
643 return number(buf, end, value, spec); symbol_string()
648 char *resource_string(char *buf, char *end, struct resource *res, resource_string() argument
731 if (res->start != res->end) { resource_string()
733 p = number(p, pend, res->end, *specp); resource_string()
752 return string(buf, end, sym, spec); resource_string()
756 char *hex_string(char *buf, char *end, u8 *addr, struct printf_spec spec, hex_string() argument
769 return string(buf, end, NULL, spec); hex_string()
790 if (buf < end) hex_string()
793 if (buf < end) hex_string()
798 if (buf < end) hex_string()
808 char *bitmap_string(char *buf, char *end, unsigned long *bitmap, bitmap_string() argument
834 if (buf < end) bitmap_string()
841 buf = number(buf, end, val, spec); bitmap_string()
849 char *bitmap_list_string(char *buf, char *end, unsigned long *bitmap, bitmap_list_string() argument
868 if (buf < end) bitmap_list_string()
874 buf = number(buf, end, rbot, spec); bitmap_list_string()
876 if (buf < end) bitmap_list_string()
880 buf = number(buf, end, rtop, spec); bitmap_list_string()
889 char *mac_address_string(char *buf, char *end, u8 *addr, mac_address_string() argument
923 return string(buf, end, mac_addr, spec); mac_address_string()
1076 char *ip6_addr_string(char *buf, char *end, const u8 *addr, ip6_addr_string() argument
1086 return string(buf, end, ip6_addr, spec); ip6_addr_string()
1090 char *ip4_addr_string(char *buf, char *end, const u8 *addr, ip4_addr_string() argument
1097 return string(buf, end, ip4_addr, spec); ip4_addr_string()
1101 char *ip6_addr_string_sa(char *buf, char *end, const struct sockaddr_in6 *sa, ip6_addr_string_sa() argument
1159 return string(buf, end, ip6_addr, spec); ip6_addr_string_sa()
1163 char *ip4_addr_string_sa(char *buf, char *end, const struct sockaddr_in *sa, ip4_addr_string_sa() argument
1194 return string(buf, end, ip4_addr, spec); ip4_addr_string_sa()
1198 char *escaped_string(char *buf, char *end, u8 *addr, struct printf_spec spec, escaped_string() argument
1210 return string(buf, end, NULL, spec); /* NULL pointer */ escaped_string()
1252 buf += string_escape_mem(addr, len, buf, buf < end ? end - buf : 0, flags, NULL); escaped_string()
1258 char *uuid_string(char *buf, char *end, const u8 *addr, uuid_string() argument
1301 return string(buf, end, uuid, spec); uuid_string()
1305 char *netdev_feature_string(char *buf, char *end, const u8 *addr, netdev_feature_string() argument
1313 return number(buf, end, *(const netdev_features_t *)addr, spec); netdev_feature_string()
1317 char *address_val(char *buf, char *end, const void *addr, address_val() argument
1337 return number(buf, end, num, spec); address_val()
1341 char *clock(char *buf, char *end, struct clk *clk, struct printf_spec spec, clock() argument
1345 return string(buf, end, NULL, spec); clock()
1349 return number(buf, end, clk_get_rate(clk), spec); clock()
1354 return string(buf, end, __clk_get_name(clk), spec); clock()
1359 return number(buf, end, (unsigned long)clk, spec); clock()
1459 char *pointer(const char *fmt, char *buf, char *end, void *ptr, pointer() argument
1471 return string(buf, end, "(null)", spec); pointer()
1482 return symbol_string(buf, end, ptr, spec, fmt); pointer()
1485 return resource_string(buf, end, ptr, spec, fmt); pointer()
1487 return hex_string(buf, end, ptr, spec, fmt); pointer()
1491 return bitmap_list_string(buf, end, ptr, spec, fmt); pointer()
1493 return bitmap_string(buf, end, ptr, spec, fmt); pointer()
1499 return mac_address_string(buf, end, ptr, spec, fmt); pointer()
1511 return ip6_addr_string(buf, end, ptr, spec, fmt); pointer()
1513 return ip4_addr_string(buf, end, ptr, spec, fmt); pointer()
1523 return ip4_addr_string_sa(buf, end, &sa->v4, spec, fmt); pointer()
1525 return ip6_addr_string_sa(buf, end, &sa->v6, spec, fmt); pointer()
1527 return string(buf, end, "(invalid address)", spec); pointer()
1532 return escaped_string(buf, end, ptr, spec, fmt); pointer()
1534 return uuid_string(buf, end, ptr, spec, fmt); pointer()
1540 buf += vsnprintf(buf, end > buf ? end - buf : 0, pointer()
1554 return string(buf, end, "pK-error", spec); pointer()
1590 return netdev_feature_string(buf, end, ptr, spec); pointer()
1594 return address_val(buf, end, ptr, spec, fmt); pointer()
1596 return dentry_name(buf, end, ptr, spec, fmt); pointer()
1598 return clock(buf, end, ptr, spec, fmt); pointer()
1600 return dentry_name(buf, end, pointer()
1611 return number(buf, end, (unsigned long) ptr, spec); pointer()
1840 char *str, *end; vsnprintf() local
1849 end = buf + size; vsnprintf()
1851 /* Make sure end is always >= buf */ vsnprintf()
1852 if (end < buf) { vsnprintf()
1853 end = ((void *)-1); vsnprintf()
1854 size = end - buf; vsnprintf()
1866 if (str < end) { vsnprintf()
1867 if (copy > end - str) vsnprintf()
1868 copy = end - str; vsnprintf()
1888 if (str < end) vsnprintf()
1895 if (str < end) vsnprintf()
1899 if (str < end) vsnprintf()
1907 str = string(str, end, va_arg(args, char *), spec); vsnprintf()
1911 str = pointer(fmt, str, end, va_arg(args, void *), vsnprintf()
1918 if (str < end) vsnprintf()
1973 str = number(str, end, num, spec); vsnprintf()
1979 if (str < end) vsnprintf()
1982 end[-1] = '\0'; vsnprintf()
2143 char *str, *end; vbin_printf() local
2146 end = (char *)(bin_buf + size); vbin_printf()
2154 if (str + sizeof(type) <= end) { \ vbin_printf()
2162 if (str + sizeof(type) <= end) \ vbin_printf()
2197 if (str + len < end) vbin_printf()
2271 char *str, *end; bstr_printf() local
2278 end = buf + size; bstr_printf()
2295 /* Make sure end is always >= buf */ bstr_printf()
2296 if (end < buf) { bstr_printf()
2297 end = ((void *)-1); bstr_printf()
2298 size = end - buf; bstr_printf()
2310 if (str < end) { bstr_printf()
2311 if (copy > end - str) bstr_printf()
2312 copy = end - str; bstr_printf()
2332 if (str < end) bstr_printf()
2338 if (str < end) bstr_printf()
2342 if (str < end) bstr_printf()
2352 str = string(str, end, (char *)str_arg, spec); bstr_printf()
2357 str = pointer(fmt, str, end, get_arg(void *), spec); bstr_printf()
2363 if (str < end) bstr_printf()
2408 str = number(str, end, num, spec); bstr_printf()
2415 if (str < end) bstr_printf()
2418 end[-1] = '\0'; bstr_printf()
/linux-4.4.14/arch/blackfin/mach-bf527/boards/
H A Dcm_bf527.c45 .end = 0x203C0000 + 0x000fffff,
50 .end = IRQ_PF7,
79 .end = 0xffc03cff,
84 .end = IRQ_USB_INT0,
90 .end = IRQ_USB_DMA,
162 .end = NFC_DATA_RD + 2,
167 .end = CH_NFC,
187 .end = 0x20312000,
191 .end = 0x20311FFF,
195 .end = IRQ_PF4,
199 .end = 6,
232 .end = 0x20300300 + 16,
237 .end = IRQ_PF7,
256 .end = 0x203FB800 + 1,
261 .end = 0x203FB804 + 1,
266 .end = IRQ_PF9,
316 .end = 0x20300000 + 0x100,
320 .end = IRQ_PF7,
456 .end = SPI0_REGBASE + 0xFF,
461 .end = CH_SPI,
466 .end = IRQ_SPI,
511 .end = 0x201fffff,
515 .end = ARRAY_SIZE(cm_flash_gpios),
536 .end = UART0_GCTL+2,
541 .end = IRQ_UART0_TX,
546 .end = IRQ_UART0_RX,
551 .end = IRQ_UART0_ERROR,
556 .end = CH_UART0_TX,
561 .end = CH_UART0_RX,
584 .end = UART1_GCTL+2,
589 .end = IRQ_UART1_TX,
594 .end = IRQ_UART1_RX,
599 .end = IRQ_UART1_ERROR,
604 .end = CH_UART1_TX,
609 .end = CH_UART1_RX,
615 .end = GPIO_PF9,
620 .end = GPIO_PF10,
647 .end = 0xFFC004FF,
652 .end = IRQ_UART0_RX+1,
657 .end = CH_UART0_RX+1,
673 .end = 0xFFC020FF,
678 .end = IRQ_UART1_RX+1,
683 .end = CH_UART1_RX+1,
703 .end = TWI0_REGBASE,
708 .end = IRQ_TWI,
748 .end = SPORT0_MRCS3+4,
753 .end = IRQ_SPORT0_RX+1,
758 .end = IRQ_SPORT0_ERROR,
782 .end = SPORT1_MRCS3+4,
787 .end = IRQ_SPORT1_RX+1,
792 .end = IRQ_SPORT1_ERROR,
H A Dezkit.c50 .end = 0x203C0000 + 0x000fffff,
55 .end = IRQ_PF7,
84 .end = 0xffc03cff,
89 .end = IRQ_USB_INT0,
95 .end = IRQ_USB_DMA,
147 .end = IRQ_PPI_ERROR,
171 .end = IRQ_PPI_ERROR,
212 .end = 0x203fffff,
256 .end = NFC_DATA_RD + 2,
261 .end = CH_NFC,
281 .end = 0x20312000,
285 .end = 0x20311FFF,
289 .end = IRQ_PF4,
293 .end = 6,
326 .end = 0x20300300 + 16,
331 .end = IRQ_PF7,
350 .end = 0x203FB800 + 1,
355 .end = 0x203FB800 + 5,
360 .end = IRQ_PF9,
410 .end = 0x20300000 + 0x100,
417 .end = IRQ_PF7,
517 .end = SPORT##x##_TCR1, \
522 .end = CH_SPORT##x##_RX, \
527 .end = CH_SPORT##x##_TX, \
532 .end = IRQ_SPORT##x##_ERROR, \
670 .end = SPI0_REGBASE + 0xFF,
675 .end = CH_SPI,
680 .end = IRQ_SPI,
701 .end = UART0_GCTL+2,
706 .end = IRQ_UART0_TX,
711 .end = IRQ_UART0_RX,
716 .end = IRQ_UART0_ERROR,
721 .end = CH_UART0_TX,
726 .end = CH_UART0_RX,
749 .end = UART1_GCTL+2,
754 .end = IRQ_UART1_TX,
759 .end = IRQ_UART1_RX,
764 .end = IRQ_UART1_ERROR,
769 .end = CH_UART1_TX,
774 .end = CH_UART1_RX,
780 .end = GPIO_PF9,
785 .end = GPIO_PF10,
812 .end = 0xFFC004FF,
817 .end = IRQ_UART0_RX+1,
822 .end = CH_UART0_RX+1,
838 .end = 0xFFC020FF,
843 .end = IRQ_UART1_RX+1,
848 .end = CH_UART1_RX+1,
868 .end = TWI0_REGBASE,
873 .end = IRQ_TWI,
1007 .end = SPORT0_MRCS3+4,
1012 .end = IRQ_SPORT0_RX+1,
1017 .end = IRQ_SPORT0_ERROR,
1041 .end = SPORT1_MRCS3+4,
1046 .end = IRQ_SPORT1_RX+1,
1051 .end = IRQ_SPORT1_ERROR,
1118 .end = CNT_CONFIG + 0xff,
1123 .end = IRQ_CNT,
H A Dad7160eval.c44 .end = 0xffc03cff,
49 .end = IRQ_USB_INT0,
54 .end = IRQ_USB_DMA,
104 .end = IRQ_PPI_ERROR,
142 .end = 0x203fffff,
182 .end = NFC_DATA_RD + 2,
187 .end = CH_NFC,
338 .end = SPI0_REGBASE + 0xFF,
343 .end = CH_SPI,
348 .end = IRQ_SPI,
369 .end = UART0_GCTL+2,
374 .end = IRQ_UART0_TX,
379 .end = IRQ_UART0_RX,
384 .end = IRQ_UART0_ERROR,
389 .end = CH_UART0_TX,
394 .end = CH_UART0_RX,
417 .end = UART1_GCTL+2,
422 .end = IRQ_UART1_TX,
427 .end = IRQ_UART1_RX,
432 .end = IRQ_UART1_ERROR,
437 .end = CH_UART1_TX,
442 .end = CH_UART1_RX,
448 .end = GPIO_PF9,
453 .end = GPIO_PF10,
480 .end = 0xFFC004FF,
485 .end = IRQ_UART0_RX+1,
490 .end = CH_UART0_RX+1,
506 .end = 0xFFC020FF,
511 .end = IRQ_UART1_RX+1,
516 .end = CH_UART1_RX+1,
566 .end = TWI0_REGBASE,
571 .end = IRQ_TWI,
602 .end = SPORT0_MRCS3+4,
607 .end = IRQ_SPORT0_RX+1,
612 .end = IRQ_SPORT0_ERROR,
636 .end = SPORT1_MRCS3+4,
641 .end = IRQ_SPORT1_RX+1,
646 .end = IRQ_SPORT1_ERROR,
692 .end = CNT_CONFIG + 0xff,
697 .end = IRQ_CNT,
H A Dezbrd.c43 .end = 0xffc03cff,
48 .end = IRQ_USB_INT0,
54 .end = IRQ_USB_DMA,
126 .end = 0x203fffff,
170 .end = NFC_DATA_RD + 2,
175 .end = CH_NFC,
384 .end = SPI0_REGBASE + 0xFF,
389 .end = CH_SPI,
394 .end = IRQ_SPI,
415 .end = UART0_GCTL+2,
420 .end = IRQ_UART0_TX,
425 .end = IRQ_UART0_RX,
430 .end = IRQ_UART0_ERROR,
435 .end = CH_UART0_TX,
440 .end = CH_UART0_RX,
463 .end = UART1_GCTL+2,
468 .end = IRQ_UART1_TX,
473 .end = IRQ_UART1_RX,
478 .end = IRQ_UART1_ERROR,
483 .end = CH_UART1_TX,
488 .end = CH_UART1_RX,
494 .end = GPIO_PG0,
499 .end = GPIO_PF10,
526 .end = 0xFFC004FF,
531 .end = IRQ_UART0_RX+1,
536 .end = CH_UART0_RX+1,
552 .end = 0xFFC020FF,
557 .end = IRQ_UART1_RX+1,
562 .end = CH_UART1_RX+1,
582 .end = TWI0_REGBASE,
587 .end = IRQ_TWI,
622 .end = SPORT0_MRCS3+4,
627 .end = IRQ_SPORT0_RX+1,
632 .end = IRQ_SPORT0_ERROR,
656 .end = SPORT1_MRCS3+4,
661 .end = IRQ_SPORT1_RX+1,
666 .end = IRQ_SPORT1_ERROR,
745 .end = IRQ_PPI_ERROR,
H A Dtll6527m.c51 .end = 0xffc03cff,
56 .end = IRQ_USB_INT0,
61 .end = IRQ_USB_DMA,
119 .end = IRQ_PPI_ERROR,
164 .end = 0x201fffff,
168 .end = ARRAY_SIZE(tll6527m_flash_gpios),
459 .end = SPI0_REGBASE + 0xFF,
464 .end = CH_SPI,
469 .end = IRQ_SPI,
490 .end = UART0_GCTL+2,
495 .end = IRQ_UART0_TX,
500 .end = IRQ_UART0_RX,
505 .end = IRQ_UART0_ERROR,
510 .end = CH_UART0_TX,
515 .end = CH_UART0_RX,
539 .end = UART1_GCTL+2,
544 .end = IRQ_UART1_TX,
549 .end = IRQ_UART1_RX,
554 .end = IRQ_UART1_ERROR,
559 .end = CH_UART1_TX,
564 .end = CH_UART1_RX,
570 .end = GPIO_PF9,
575 .end = GPIO_PF10,
603 .end = 0xFFC004FF,
608 .end = IRQ_UART0_RX+1,
613 .end = CH_UART0_RX+1,
629 .end = 0xFFC020FF,
634 .end = IRQ_UART1_RX+1,
639 .end = CH_UART1_RX+1,
659 .end = TWI0_REGBASE,
664 .end = IRQ_TWI,
725 .end = SPORT0_MRCS3+4,
730 .end = IRQ_SPORT0_RX+1,
735 .end = IRQ_SPORT0_ERROR,
760 .end = SPORT1_MRCS3+4,
765 .end = IRQ_SPORT1_RX+1,
770 .end = IRQ_SPORT1_ERROR,
/linux-4.4.14/arch/blackfin/mach-bf533/boards/
H A Dip0x.c42 .end = 0x20100000 + 1,
46 .end = 0x20100000 + 3,
50 .end = IRQ_PF15,
58 .end = 0x20200000 + 1,
62 .end = 0x20200000 + 3,
66 .end = IRQ_PF14,
150 .end = BFIN_UART_GCTL+2,
155 .end = IRQ_UART0_TX,
160 .end = IRQ_UART0_RX,
165 .end = IRQ_UART0_ERROR,
170 .end = CH_UART0_TX,
175 .end = CH_UART0_RX,
201 .end = 0xFFC004FF,
206 .end = IRQ_UART0_RX+1,
211 .end = CH_UART0_RX+1,
229 .end = 0x20300000 + 1,
233 .end = 0x20300000 + 3,
237 .end = IRQ_PF11,
H A Dcm_bf533.c110 .end = SPI0_REGBASE + 0xFF,
115 .end = CH_SPI,
120 .end = IRQ_SPI,
162 .end = 0x20200300 + 16,
166 .end = IRQ_PF0,
188 .end = 0x20308000 + 0xFF,
192 .end = IRQ_PF8,
220 .end = BFIN_UART_GCTL+2,
225 .end = IRQ_UART0_TX,
230 .end = IRQ_UART0_RX,
235 .end = IRQ_UART0_ERROR,
240 .end = CH_UART0_TX,
245 .end = CH_UART0_RX,
271 .end = 0xFFC004FF,
276 .end = IRQ_UART0_RX+1,
281 .end = CH_UART0_RX+1,
300 .end = SPORT0_MRCS3+4,
305 .end = IRQ_SPORT0_RX+1,
310 .end = IRQ_SPORT0_ERROR,
334 .end = SPORT1_MRCS3+4,
339 .end = IRQ_SPORT1_RX+1,
344 .end = IRQ_SPORT1_ERROR,
370 .end = 0x20308000,
374 .end = 0x20308004,
378 .end = IRQ_PF4,
410 .end = 0x20300000 + 0x100,
414 .end = IRQ_PF6,
450 .end = 0x201fffff,
H A DH8606.c46 .end = 0x20300002,
51 .end = 0x20300006,
56 .end = IRQ_PF10,
83 .end = 0x20300300 + 16,
87 .end = IRQ_PROG_INTB,
91 .end = IRQ_PF7,
111 .end = 0x20300000 + 0x100,
115 .end = IRQ_PF10,
198 .end = SPI0_REGBASE + 0xFF,
203 .end = CH_SPI,
208 .end = IRQ_SPI,
237 .end = BFIN_UART_GCTL+2,
242 .end = IRQ_UART0_TX,
247 .end = IRQ_UART0_RX,
252 .end = IRQ_UART0_ERROR,
257 .end = CH_UART0_TX,
262 .end = CH_UART0_RX,
288 .end = 0xFFC004FF,
293 .end = IRQ_UART0_RX+1,
298 .end = CH_UART0_RX+1,
366 .end = 0x20300030 + 2,
371 .end = IRQ_PF9,
H A Dblackstamp.c58 .end = 0x20300300 + 16,
62 .end = IRQ_PF3,
155 .end = SPI0_REGBASE + 0xFF,
160 .end = CH_SPI,
165 .end = IRQ_SPI,
193 .end = BFIN_UART_GCTL+2,
198 .end = IRQ_UART0_TX,
203 .end = IRQ_UART0_RX,
208 .end = IRQ_UART0_ERROR,
213 .end = CH_UART0_TX,
218 .end = CH_UART0_RX,
244 .end = 0xFFC004FF,
249 .end = IRQ_UART0_RX+1,
254 .end = CH_UART0_RX+1,
273 .end = SPORT0_MRCS3+4,
278 .end = IRQ_SPORT0_RX+1,
283 .end = IRQ_SPORT0_ERROR,
307 .end = SPORT1_MRCS3+4,
312 .end = IRQ_SPORT1_RX+1,
317 .end = IRQ_SPORT1_ERROR,
H A Dstamp.c58 .end = 0x20300300 + 16,
62 .end = IRQ_PF7,
82 .end = 0x20300000 + 0x100,
89 .end = IRQ_PF10,
129 .end = 0x203fffff,
133 .end = 0x7BB07BB0, /* AMBCTL1 setting when accessing flash */
261 .end = SPI0_REGBASE + 0xFF,
266 .end = CH_SPI,
271 .end = IRQ_SPI,
299 .end = BFIN_UART_GCTL+2,
304 .end = IRQ_UART0_TX,
309 .end = IRQ_UART0_RX,
314 .end = IRQ_UART0_ERROR,
319 .end = CH_UART0_TX,
324 .end = CH_UART0_RX,
350 .end = 0xFFC004FF,
355 .end = IRQ_UART0_RX+1,
360 .end = CH_UART0_RX+1,
379 .end = SPORT0_MRCS3+4,
384 .end = IRQ_SPORT0_RX+1,
389 .end = IRQ_SPORT0_ERROR,
413 .end = SPORT1_MRCS3+4,
418 .end = IRQ_SPORT1_RX+1,
423 .end = IRQ_SPORT1_ERROR,
449 .end = SPORT0_MRCS3+4,
454 .end = IRQ_SPORT0_TX+1,
459 .end = IRQ_SPORT0_RX+1,
464 .end = IRQ_SPORT0_ERROR,
469 .end = CH_SPORT0_TX,
474 .end = CH_SPORT0_RX,
616 .end = SPORT##x##_TCR1, \
621 .end = CH_SPORT##x##_RX, \
626 .end = CH_SPORT##x##_TX, \
631 .end = IRQ_SPORT##x##_ERROR, \
/linux-4.4.14/arch/c6x/platforms/
H A Dcache.c133 unsigned int *end, cache_block_operation()
139 (L2_CACHE_ALIGN_CNT((unsigned int) end) cache_block_operation()
178 unsigned int *end, cache_block_operation_nowait()
184 (L2_CACHE_ALIGN_CNT((unsigned int) end) cache_block_operation_nowait()
325 void enable_caching(unsigned long start, unsigned long end) enable_caching() argument
328 unsigned int mar_e = IMCR_MAR_BASE + ((end >> 24) << 2); enable_caching()
334 void disable_caching(unsigned long start, unsigned long end) disable_caching() argument
337 unsigned int mar_e = IMCR_MAR_BASE + ((end >> 24) << 2); disable_caching()
347 void L1P_cache_block_invalidate(unsigned int start, unsigned int end) L1P_cache_block_invalidate() argument
350 (unsigned int *) end, L1P_cache_block_invalidate()
355 void L1D_cache_block_invalidate(unsigned int start, unsigned int end) L1D_cache_block_invalidate() argument
358 (unsigned int *) end, L1D_cache_block_invalidate()
362 void L1D_cache_block_writeback_invalidate(unsigned int start, unsigned int end) L1D_cache_block_writeback_invalidate() argument
365 (unsigned int *) end, L1D_cache_block_writeback_invalidate()
369 void L1D_cache_block_writeback(unsigned int start, unsigned int end) L1D_cache_block_writeback() argument
372 (unsigned int *) end, L1D_cache_block_writeback()
380 void L2_cache_block_invalidate(unsigned int start, unsigned int end) L2_cache_block_invalidate() argument
383 (unsigned int *) end, L2_cache_block_invalidate()
387 void L2_cache_block_writeback(unsigned int start, unsigned int end) L2_cache_block_writeback() argument
390 (unsigned int *) end, L2_cache_block_writeback()
394 void L2_cache_block_writeback_invalidate(unsigned int start, unsigned int end) L2_cache_block_writeback_invalidate() argument
397 (unsigned int *) end, L2_cache_block_writeback_invalidate()
401 void L2_cache_block_invalidate_nowait(unsigned int start, unsigned int end) L2_cache_block_invalidate_nowait() argument
404 (unsigned int *) end, L2_cache_block_invalidate_nowait()
408 void L2_cache_block_writeback_nowait(unsigned int start, unsigned int end) L2_cache_block_writeback_nowait() argument
411 (unsigned int *) end, L2_cache_block_writeback_nowait()
416 unsigned int end) L2_cache_block_writeback_invalidate_nowait()
419 (unsigned int *) end, L2_cache_block_writeback_invalidate_nowait()
132 cache_block_operation(unsigned int *start, unsigned int *end, unsigned int bar_reg, unsigned int wc_reg) cache_block_operation() argument
177 cache_block_operation_nowait(unsigned int *start, unsigned int *end, unsigned int bar_reg, unsigned int wc_reg) cache_block_operation_nowait() argument
415 L2_cache_block_writeback_invalidate_nowait(unsigned int start, unsigned int end) L2_cache_block_writeback_invalidate_nowait() argument
/linux-4.4.14/arch/mn10300/include/asm/
H A Dcacheflush.h25 extern void mn10300_local_icache_inv_range(unsigned long start, unsigned long end);
29 extern void mn10300_local_dcache_inv_range(unsigned long start, unsigned long end);
33 extern void mn10300_icache_inv_range(unsigned long start, unsigned long end);
37 extern void mn10300_dcache_inv_range(unsigned long start, unsigned long end);
42 extern void mn10300_local_dcache_flush_range(unsigned long start, unsigned long end);
46 extern void mn10300_local_dcache_flush_inv_range(unsigned long start, unsigned long end);
50 extern void mn10300_dcache_flush_range(unsigned long start, unsigned long end);
54 extern void mn10300_dcache_flush_inv_range(unsigned long start, unsigned long end);
59 #define mn10300_local_dcache_flush_range(start, end) do {} while (0)
65 #define mn10300_local_dcache_flush_inv_range(start, end) \
66 mn10300_local_dcache_inv_range(start, end)
71 #define mn10300_dcache_flush_range(start, end) do {} while (0)
76 #define mn10300_dcache_flush_inv_range(start, end) \
77 mn10300_dcache_inv_range((start), (end))
84 #define mn10300_local_icache_inv_range(start, end) do {} while (0)
88 #define mn10300_local_dcache_inv_range(start, end) do {} while (0)
93 #define mn10300_local_dcache_flush_inv_range(start, end)do {} while (0)
96 #define mn10300_local_dcache_flush_range(start, end) do {} while (0)
100 #define mn10300_icache_inv_range(start, end) do {} while (0)
104 #define mn10300_dcache_inv_range(start, end) do {} while (0)
109 #define mn10300_dcache_flush_inv_range(start, end) do {} while (0)
112 #define mn10300_dcache_flush_range(start, end) do {} while (0)
122 #define flush_cache_range(mm, start, end) do {} while (0)
124 #define flush_cache_vmap(start, end) do {} while (0)
125 #define flush_cache_vunmap(start, end) do {} while (0)
136 extern void flush_icache_range(unsigned long start, unsigned long end);
143 extern void flush_icache_range(unsigned long start, unsigned long end);
145 #define flush_icache_range(start, end) do {} while (0)
/linux-4.4.14/arch/m32r/include/asm/
H A Dcacheflush.h13 #define flush_cache_range(vma, start, end) do { } while (0)
20 #define flush_icache_range(start, end) _flush_cache_copyback_all()
26 #define flush_icache_range(start, end) smp_flush_cache_all()
35 #define flush_cache_range(vma, start, end) do { } while (0)
41 #define flush_icache_range(start, end) _flush_cache_all()
49 #define flush_cache_range(vma, start, end) do { } while (0)
55 #define flush_icache_range(start, end) do { } while (0)
61 #define flush_cache_vmap(start, end) do { } while (0)
62 #define flush_cache_vunmap(start, end) do { } while (0)
H A Dtlbflush.h13 * - flush_tlb_range(vma, start, end) flushes a range of pages
14 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages
28 #define flush_tlb_range(vma, start, end) \
29 local_flush_tlb_range(vma, start, end)
30 #define flush_tlb_kernel_range(start, end) local_flush_tlb_all()
35 #define flush_tlb_range(vma, start, end) do { } while (0)
47 #define flush_tlb_range(vma, start, end) \
48 smp_flush_tlb_range(vma, start, end)
49 #define flush_tlb_kernel_range(start, end) smp_flush_tlb_all()
/linux-4.4.14/arch/arm/plat-iop/
H A Dadma.c55 .end = IOP3XX_DMA_UPPER_PA(0),
60 .end = IRQ_DMA0_EOT,
65 .end = IRQ_DMA0_EOC,
70 .end = IRQ_DMA0_ERR,
78 .end = IOP3XX_DMA_UPPER_PA(1),
83 .end = IRQ_DMA1_EOT,
88 .end = IRQ_DMA1_EOC,
93 .end = IRQ_DMA1_ERR,
102 .end = IOP3XX_AAU_UPPER_PA,
107 .end = IRQ_AA_EOT,
112 .end = IRQ_AA_EOC,
117 .end = IRQ_AA_ERR,
/linux-4.4.14/arch/arm/mach-ks8695/
H A Ddevices.c43 .end = KS8695_WAN_PA + 0x00ff,
49 .end = KS8695_IRQ_WAN_RX_STATUS,
55 .end = KS8695_IRQ_WAN_TX_STATUS,
61 .end = KS8695_IRQ_WAN_LINK,
67 .end = KS8695_MISC_PA + 0x1f,
87 .end = KS8695_LAN_PA + 0x00ff,
93 .end = KS8695_IRQ_LAN_RX_STATUS,
99 .end = KS8695_IRQ_LAN_TX_STATUS,
105 .end = KS8695_SWITCH_PA + 0x4f,
125 .end = KS8695_HPNA_PA + 0x00ff,
131 .end = KS8695_IRQ_HPNA_RX_STATUS,
137 .end = KS8695_IRQ_HPNA_TX_STATUS,
/linux-4.4.14/arch/mips/mm/
H A Dioremap.c22 phys_addr_t end; remap_area_pte() local
28 end = address + size; remap_area_pte()
29 if (end > PMD_SIZE) remap_area_pte()
30 end = PMD_SIZE; remap_area_pte()
31 BUG_ON(address >= end); remap_area_pte()
42 } while (address && (address < end)); remap_area_pte()
48 phys_addr_t end; remap_area_pmd() local
51 end = address + size; remap_area_pmd()
52 if (end > PGDIR_SIZE) remap_area_pmd()
53 end = PGDIR_SIZE; remap_area_pmd()
55 BUG_ON(address >= end); remap_area_pmd()
60 remap_area_pte(pte, address, end - address, address + phys_addr, flags); remap_area_pmd()
63 } while (address && (address < end)); remap_area_pmd()
72 unsigned long end = address + size; remap_area_pages() local
77 BUG_ON(address >= end); remap_area_pages()
89 if (remap_area_pmd(pmd, address, end - address, remap_area_pages()
95 } while (address && (address < end)); remap_area_pages()
H A Dpgtable-64.c18 unsigned long *p, *end; pgd_init() local
28 end = p + PTRS_PER_PGD; pgd_init()
40 } while (p != end); pgd_init()
46 unsigned long *p, *end; pmd_init() local
49 end = p + PTRS_PER_PMD; pmd_init()
61 } while (p != end); pmd_init()
H A Dgup.c37 static int gup_pte_range(pmd_t pmd, unsigned long addr, unsigned long end, gup_pte_range() argument
57 } while (ptep++, addr += PAGE_SIZE, addr != end); gup_pte_range()
71 static int gup_huge_pmd(pmd_t pmd, unsigned long addr, unsigned long end, gup_huge_pmd() argument
95 } while (addr += PAGE_SIZE, addr != end); gup_huge_pmd()
101 static int gup_pmd_range(pud_t pud, unsigned long addr, unsigned long end, gup_pmd_range() argument
111 next = pmd_addr_end(addr, end); gup_pmd_range()
132 } while (pmdp++, addr = next, addr != end); gup_pmd_range()
137 static int gup_huge_pud(pud_t pud, unsigned long addr, unsigned long end, gup_huge_pud() argument
161 } while (addr += PAGE_SIZE, addr != end); gup_huge_pud()
167 static int gup_pud_range(pgd_t pgd, unsigned long addr, unsigned long end, gup_pud_range() argument
177 next = pud_addr_end(addr, end); gup_pud_range()
187 } while (pudp++, addr = next, addr != end); gup_pud_range()
200 unsigned long addr, len, end; __get_user_pages_fast() local
209 end = start + len; __get_user_pages_fast()
236 next = pgd_addr_end(addr, end); __get_user_pages_fast()
241 } while (pgdp++, addr = next, addr != end); __get_user_pages_fast()
267 unsigned long addr, len, end; get_user_pages_fast() local
276 end = start + len; get_user_pages_fast()
277 if (end < start || cpu_has_dc_aliases) get_user_pages_fast()
286 next = pgd_addr_end(addr, end); get_user_pages_fast()
291 } while (pgdp++, addr = next, addr != end); get_user_pages_fast()
294 VM_BUG_ON(nr != (end - start) >> PAGE_SHIFT); get_user_pages_fast()
305 (end - start) >> PAGE_SHIFT, get_user_pages_fast()
H A Dsc-r5k.c27 unsigned long end = start + scache_size; blast_r5000_scache() local
29 while(start < end) { blast_r5000_scache()
37 unsigned long end, a; r5k_dma_cache_inv_sc() local
52 end = (addr + size - 1) & ~(SC_PAGE - 1); r5k_dma_cache_inv_sc()
53 while (a <= end) { r5k_dma_cache_inv_sc()
H A Dsc-rm7k.c44 unsigned long end, a; rm7k_sc_wback_inv() local
57 end = (addr + size - 1) & ~(tc_pagesize - 1); rm7k_sc_wback_inv()
60 if (a == end) rm7k_sc_wback_inv()
68 unsigned long end, a; rm7k_sc_inv() local
81 end = (addr + size - 1) & ~(tc_pagesize - 1); rm7k_sc_inv()
84 if (a == end) rm7k_sc_inv()
93 unsigned long end = start + tcache_size; blast_rm7k_tcache() local
97 while (start < end) { blast_rm7k_tcache()
188 unsigned long flags, addr, begin, end, pow2; __probe_tcache() local
192 end = begin + (8 * 1024 * 1024); __probe_tcache()
200 for (addr = begin; addr <= end; addr = (begin + pow2)) { __probe_tcache()
213 for (addr = begin + (512 * 1024); addr <= end; addr = begin + pow2) { __probe_tcache()
H A Dtlb-r3k.c82 unsigned long end) local_flush_tlb_range()
92 cpu_context(cpu, mm) & ASID_MASK, start, end); local_flush_tlb_range()
95 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_range()
101 end += PAGE_SIZE - 1; local_flush_tlb_range()
102 end &= PAGE_MASK; local_flush_tlb_range()
103 while (start < end) { local_flush_tlb_range()
124 void local_flush_tlb_kernel_range(unsigned long start, unsigned long end) local_flush_tlb_kernel_range() argument
129 printk("[tlbrange<%lu,0x%08lx,0x%08lx>]", start, end); local_flush_tlb_kernel_range()
132 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_kernel_range()
137 end += PAGE_SIZE - 1; local_flush_tlb_kernel_range()
138 end &= PAGE_MASK; local_flush_tlb_kernel_range()
140 while (start < end) { local_flush_tlb_kernel_range()
81 local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) local_flush_tlb_range() argument
/linux-4.4.14/arch/mn10300/mm/
H A Dcache-flush-icache.c44 * @end: The ending virtual address of the page part.
50 static void flush_icache_page_range(unsigned long start, unsigned long end) flush_icache_page_range() argument
61 size = end - start; flush_icache_page_range()
96 smp_cache_call(SMP_IDCACHE_INV_FLUSH_RANGE, start, end); flush_icache_page_range()
102 * @end: The ending virtual address of the region.
108 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument
115 if (end > 0x80000000UL) { flush_icache_range()
117 if (end > 0xa0000000UL) { flush_icache_range()
118 end = 0xa0000000UL; flush_icache_range()
119 if (start >= end) flush_icache_range()
127 mn10300_local_dcache_flush_range(start_page, end); flush_icache_range()
128 mn10300_local_icache_inv_range(start_page, end); flush_icache_range()
129 smp_cache_call(SMP_IDCACHE_INV_FLUSH_RANGE, start_page, end); flush_icache_range()
132 end = start_page; flush_icache_range()
136 end_page = (end - 1) & PAGE_MASK; flush_icache_range()
140 flush_icache_page_range(start, end); flush_icache_range()
144 flush_icache_page_range(end_page, end); flush_icache_range()
H A Dcache-inv-icache.c22 * @end: The ending virtual address of the page part.
28 static void flush_icache_page_range(unsigned long start, unsigned long end) flush_icache_page_range() argument
39 size = end - start; flush_icache_page_range()
72 smp_cache_call(SMP_ICACHE_INV_RANGE, start, end); flush_icache_page_range()
78 * @end: The ending virtual address of the region.
84 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument
91 if (end > 0x80000000UL) { flush_icache_range()
93 if (end > 0xa0000000UL) { flush_icache_range()
94 end = 0xa0000000UL; flush_icache_range()
95 if (start >= end) flush_icache_range()
103 mn10300_icache_inv_range(start_page, end); flush_icache_range()
104 smp_cache_call(SMP_ICACHE_INV_RANGE, start, end); flush_icache_range()
107 end = start_page; flush_icache_range()
111 end_page = (end - 1) & PAGE_MASK; flush_icache_range()
115 flush_icache_page_range(start, end); flush_icache_range()
119 flush_icache_page_range(end_page, end); flush_icache_range()
H A Dcache-smp-flush.c52 * @end: The end address of the region to be flushed.
55 * end-1 inclusive.
57 void mn10300_dcache_flush_range(unsigned long start, unsigned long end) mn10300_dcache_flush_range() argument
62 mn10300_local_dcache_flush_range(start, end); mn10300_dcache_flush_range()
63 smp_cache_call(SMP_DCACHE_FLUSH_RANGE, start, end); mn10300_dcache_flush_range()
124 * @end: The end address of the region to be flushed and invalidated.
127 * between start and end-1 inclusive.
129 void mn10300_dcache_flush_inv_range(unsigned long start, unsigned long end) mn10300_dcache_flush_inv_range() argument
134 mn10300_local_dcache_flush_inv_range(start, end); mn10300_dcache_flush_inv_range()
135 smp_cache_call(SMP_DCACHE_FLUSH_INV_RANGE, start, end); mn10300_dcache_flush_inv_range()
H A Dcache-smp-inv.c52 * @end: The end address of the region to be invalidated.
55 * between start and end-1 inclusive.
57 void mn10300_icache_inv_range(unsigned long start, unsigned long end) mn10300_icache_inv_range() argument
62 mn10300_local_icache_inv_range(start, end); mn10300_icache_inv_range()
63 smp_cache_call(SMP_ICACHE_INV_RANGE, start, end); mn10300_icache_inv_range()
122 * @end: The end address of the region to be invalidated.
125 * and end-1 inclusive.
127 void mn10300_dcache_inv_range(unsigned long start, unsigned long end) mn10300_dcache_inv_range() argument
132 mn10300_local_dcache_inv_range(start, end); mn10300_dcache_inv_range()
133 smp_cache_call(SMP_DCACHE_INV_RANGE, start, end); mn10300_dcache_inv_range()
H A Dcache.c47 asmlinkage long sys_cacheflush(unsigned long start, unsigned long end) sys_cacheflush() argument
49 if (end < start) sys_cacheflush()
52 flush_icache_range(start, end); sys_cacheflush()
/linux-4.4.14/arch/blackfin/mach-bf609/boards/
H A Dezkit.c47 .end = 0x2C0C0000 + 0xfffff,
52 .end = IRQ_PG7,
92 .end = CNT_CONFIG + 0xff,
97 .end = IRQ_CNT,
150 .end = EMAC0_MACCFG + 0x1274,
156 .end = IRQ_EMAC0_STAT,
218 .end = UART0_RXDIV+4,
224 .end = PORTD_FER+2,
229 .end = PORTD_MUX+3,
235 .end = IRQ_UART0_TX,
240 .end = IRQ_UART0_RX,
245 .end = IRQ_UART0_STAT,
250 .end = CH_UART0_TX,
255 .end = CH_UART0_RX,
261 .end = GPIO_PD10,
266 .end = GPIO_PD9,
294 .end = UART1_RXDIV+4,
300 .end = PORTG_FER_SET+2,
306 .end = IRQ_UART1_TX,
311 .end = IRQ_UART1_RX,
316 .end = IRQ_UART1_STAT,
321 .end = CH_UART1_TX,
326 .end = CH_UART1_RX,
332 .end = GPIO_PG13,
337 .end = GPIO_PG10,
368 .end = 0xFFC004FF,
373 .end = IRQ_UART0_TX+1,
378 .end = CH_UART0_TX+1,
393 .end = 0xFFC020FF,
398 .end = IRQ_UART1_TX+1,
403 .end = CH_UART1_TX+1,
420 .end = 0xFFCC1398,
425 .end = IRQ_USB_STAT,
431 .end = IRQ_USB_DMA,
477 .end = SPORT0_MRCS3+4,
482 .end = IRQ_SPORT0_RX+1,
487 .end = IRQ_SPORT0_ERROR,
511 .end = SPORT1_MRCS3+4,
516 .end = IRQ_SPORT1_RX+1,
521 .end = IRQ_SPORT1_ERROR,
545 .end = SPORT2_MRCS3+4,
550 .end = IRQ_SPORT2_RX+1,
555 .end = IRQ_SPORT2_ERROR,
586 .end = 0xFFC00FFF,
591 .end = IRQ_CAN0_RX,
596 .end = IRQ_CAN0_TX,
601 .end = IRQ_CAN0_STAT,
647 .end = 0xFFC03B4F,
652 .end = CH_NFC,
731 .end = 0xb0ffffff,
795 .end = SPORT0_CTL_A,
800 .end = SPORT0_CTL_B,
805 .end = CH_SPORT0_TX,
810 .end = CH_SPORT0_RX,
815 .end = IRQ_SPORT0_TX_STAT,
820 .end = IRQ_SPORT0_RX_STAT,
1215 .end = EPPI2_STAT,
1220 .end = CH_EPPI2_CH0,
1225 .end = IRQ_EPPI2_STAT,
1245 .end = REG_CRC0_REVID+4,
1250 .end = IRQ_CRC0_DCNTEXP,
1255 .end = CH_MEM_STREAM0_SRC_CRC0,
1260 .end = CH_MEM_STREAM0_DEST_CRC0,
1275 .end = REG_CRC1_REVID+4,
1280 .end = IRQ_CRC1_DCNTEXP,
1285 .end = CH_MEM_STREAM1_SRC_CRC1,
1290 .end = CH_MEM_STREAM1_DEST_CRC1,
1310 .end = REG_CRC0_REVID+4,
1315 .end = IRQ_CRC0_DCNTEXP,
1320 .end = CH_MEM_STREAM0_SRC_CRC0,
1366 .end = PINT0_LATCH + 3,
1371 .end = IRQ_PINT0,
1386 .end = PINT1_LATCH + 3,
1391 .end = IRQ_PINT1,
1406 .end = PINT2_LATCH + 3,
1411 .end = IRQ_PINT2,
1426 .end = PINT3_LATCH + 3,
1431 .end = IRQ_PINT3,
1446 .end = PINT4_LATCH + 3,
1451 .end = IRQ_PINT4,
1466 .end = PINT5_LATCH + 3,
1471 .end = IRQ_PINT5,
1486 .end = PORTA_MUX + 3,
1491 .end = IRQ_PA0,
1517 .end = PORTB_MUX + 3,
1522 .end = IRQ_PB0,
1548 .end = PORTC_MUX + 3,
1553 .end = IRQ_PC0,
1579 .end = PORTD_MUX + 3,
1584 .end = IRQ_PD0,
1610 .end = PORTE_MUX + 3,
1615 .end = IRQ_PE0,
1641 .end = PORTF_MUX + 3,
1646 .end = IRQ_PF0,
1672 .end = PORTG_MUX + 3,
1677 .end = IRQ_PG0,
1773 .end = SPI0_REGBASE + 0xFF,
1778 .end = CH_SPI0_TX,
1783 .end = CH_SPI0_RX,
1792 .end = SPI1_REGBASE + 0xFF,
1797 .end = CH_SPI1_TX,
1802 .end = CH_SPI1_RX,
1846 .end = TWI0_CLKDIV + 0xFF,
1851 .end = IRQ_TWI0,
1871 .end = TWI1_CLKDIV + 0xFF,
1876 .end = IRQ_TWI1,
/linux-4.4.14/arch/arm/mach-w90x900/
H A Ddev.c80 .end = NUC900_FLASH_BASE + NUC900_FLASH_SIZE - 1,
100 .end = W90X900_PA_USBEHCIHOST + W90X900_SZ_USBEHCIHOST - 1,
105 .end = IRQ_USBH,
128 .end = W90X900_PA_USBOHCIHOST + W90X900_SZ_USBOHCIHOST - 1,
133 .end = IRQ_USBH,
155 .end = W90X900_PA_USBDEV + W90X900_SZ_USBDEV - 1,
160 .end = IRQ_USBD,
177 .end = W90X900_PA_EMC + W90X900_SZ_EMC - 1,
182 .end = IRQ_EMCTX,
187 .end = IRQ_EMCRX,
221 .end = W90X900_PA_I2C + SPIOFFSET + SPIOREG_SIZE - 1,
226 .end = IRQ_SSP,
274 .end = W90X900_PA_TIMER + W90X900_SZ_TIMER - 1,
279 .end = IRQ_WDT,
305 .end = W90X900_PA_RTC + 0xff,
310 .end = IRQ_RTC,
327 .end = W90X900_PA_ADC + W90X900_SZ_ADC-1,
332 .end = IRQ_ADC,
349 .end = W90X900_PA_FMI + W90X900_SZ_FMI - 1,
354 .end = IRQ_FMI,
404 .end = W90X900_PA_KPI + W90X900_SZ_KPI - 1,
409 .end = IRQ_KPI,
467 .end = W90X900_PA_LCD + W90X900_SZ_LCD - 1,
472 .end = IRQ_LCD,
495 .end = W90X900_PA_ACTL + W90X900_SZ_ACTL - 1,
500 .end = IRQ_ACTL,
/linux-4.4.14/arch/blackfin/mach-bf518/boards/
H A Dtcm-bf518.c61 .end = 0x201fffff,
240 .end = SPI0_REGBASE + 0xFF,
245 .end = CH_SPI0,
250 .end = IRQ_SPI0,
275 .end = SPI1_REGBASE + 0xFF,
280 .end = CH_SPI1,
285 .end = IRQ_SPI1,
306 .end = UART0_GCTL+2,
311 .end = IRQ_UART0_TX,
316 .end = IRQ_UART0_RX,
321 .end = IRQ_UART0_ERROR,
326 .end = CH_UART0_TX,
331 .end = CH_UART0_RX,
354 .end = UART1_GCTL+2,
359 .end = IRQ_UART1_TX,
364 .end = IRQ_UART1_RX,
369 .end = IRQ_UART1_ERROR,
374 .end = CH_UART1_TX,
379 .end = CH_UART1_RX,
405 .end = 0xFFC004FF,
410 .end = IRQ_UART0_RX+1,
415 .end = CH_UART0_RX+1,
431 .end = 0xFFC020FF,
436 .end = IRQ_UART1_RX+1,
441 .end = CH_UART1_RX+1,
461 .end = TWI0_REGBASE,
466 .end = IRQ_TWI,
501 .end = SPORT0_MRCS3+4,
506 .end = IRQ_SPORT0_RX+1,
511 .end = IRQ_SPORT0_ERROR,
535 .end = SPORT1_MRCS3+4,
540 .end = IRQ_SPORT1_RX+1,
545 .end = IRQ_SPORT1_ERROR,
H A Dezbrd.c65 .end = 0x202fffff,
67 .end = 0x203fffff,
305 .end = SPI0_REGBASE + 0xFF,
310 .end = CH_SPI0,
315 .end = IRQ_SPI0,
340 .end = SPI1_REGBASE + 0xFF,
345 .end = CH_SPI1,
350 .end = IRQ_SPI1,
371 .end = UART0_GCTL+2,
376 .end = IRQ_UART0_TX,
381 .end = IRQ_UART0_RX,
386 .end = IRQ_UART0_ERROR,
391 .end = CH_UART0_TX,
396 .end = CH_UART0_RX,
419 .end = UART1_GCTL+2,
424 .end = IRQ_UART1_TX,
429 .end = IRQ_UART1_RX,
434 .end = IRQ_UART1_ERROR,
439 .end = CH_UART1_TX,
444 .end = CH_UART1_RX,
470 .end = 0xFFC004FF,
475 .end = IRQ_UART0_RX+1,
480 .end = CH_UART0_RX+1,
496 .end = 0xFFC020FF,
501 .end = IRQ_UART1_RX+1,
506 .end = CH_UART1_RX+1,
534 .end = TWI0_REGBASE,
539 .end = IRQ_TWI,
579 .end = SPORT0_MRCS3+4,
584 .end = IRQ_SPORT0_RX+1,
589 .end = IRQ_SPORT0_ERROR,
613 .end = SPORT1_MRCS3+4,
618 .end = IRQ_SPORT1_RX+1,
623 .end = IRQ_SPORT1_ERROR,
/linux-4.4.14/arch/alpha/lib/
H A Dclear_page.S39 .end clear_page
/linux-4.4.14/include/asm-generic/
H A Dcacheflush.h14 #define flush_cache_range(vma, start, end) do { } while (0)
20 #define flush_icache_range(start, end) do { } while (0)
23 #define flush_cache_vmap(start, end) do { } while (0)
24 #define flush_cache_vunmap(start, end) do { } while (0)
/linux-4.4.14/arch/powerpc/perf/req-gen/
H A D_begin.h13 #define REQUEST_END CAT2_STR(REQ_GEN_PREFIX, _request-end.h)
/linux-4.4.14/arch/sh/kernel/cpu/sh4a/
H A Dpinmux-sh7722.c10 .end = 0xa405018f,
/linux-4.4.14/arch/mips/cavium-octeon/
H A Dcsrc-octeon.c140 u64 cur, end, inc; __udelay() local
145 end = cur + inc; __udelay()
147 while (end > cur) __udelay()
154 u64 cur, end, inc; __ndelay() local
159 end = cur + inc; __ndelay()
161 while (end > cur) __ndelay()
168 u64 cur, end; __delay() local
171 end = cur + loops; __delay()
173 while (end > cur) __delay()
189 u64 cur, end; octeon_io_clk_delay() local
193 end = count * rdiv; octeon_io_clk_delay()
197 : [cnt] "+r" (end) octeon_io_clk_delay()
201 end = cur + end; octeon_io_clk_delay()
203 end = cur + count; octeon_io_clk_delay()
205 while (end > cur) octeon_io_clk_delay()
/linux-4.4.14/arch/cris/include/asm/
H A Dcacheflush.h13 #define flush_cache_range(vma, start, end) do { } while (0)
19 #define flush_icache_range(start, end) do { } while (0)
22 #define flush_cache_vmap(start, end) do { } while (0)
23 #define flush_cache_vunmap(start, end) do { } while (0)
H A Dtlbflush.h16 * - flush_tlb_range(mm, start, end) flushes a range of pages
29 static inline void flush_tlb_range(struct vm_area_struct * vma, unsigned long start, unsigned long end) flush_tlb_range() argument
39 #define flush_tlb_kernel_range(start, end) flush_tlb_all()
/linux-4.4.14/arch/m32r/boot/compressed/
H A Dvmlinux.lds.S27 end = . ; define
/linux-4.4.14/arch/avr32/include/asm/
H A Dtlbflush.h20 * - flush_tlb_range(vma, start, end) flushes a range of pages
21 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages
27 unsigned long end);
30 extern void flush_tlb_kernel_range(unsigned long start, unsigned long end);
/linux-4.4.14/tools/build/feature/
H A Dtest-libunwind-debug-frame.c8 unw_word_t end);
/linux-4.4.14/tools/include/linux/
H A Dlist.h12 * @end: last element in the range to delete from the list.
17 struct list_head *end) list_del_range()
19 begin->prev->next = end->next; list_del_range()
20 end->next->prev = begin->prev; list_del_range()
16 list_del_range(struct list_head *begin, struct list_head *end) list_del_range() argument
/linux-4.4.14/drivers/net/ethernet/stmicro/stmmac/
H A Ddescs_com.h36 static inline void ehn_desc_rx_set_on_ring(struct dma_desc *p, int end) ehn_desc_rx_set_on_ring() argument
39 if (end) ehn_desc_rx_set_on_ring()
43 static inline void ehn_desc_tx_set_on_ring(struct dma_desc *p, int end) ehn_desc_tx_set_on_ring() argument
45 if (end) ehn_desc_tx_set_on_ring()
64 static inline void ndesc_rx_set_on_ring(struct dma_desc *p, int end) ndesc_rx_set_on_ring() argument
67 if (end) ndesc_rx_set_on_ring()
71 static inline void ndesc_tx_set_on_ring(struct dma_desc *p, int end) ndesc_tx_set_on_ring() argument
73 if (end) ndesc_tx_set_on_ring()
94 static inline void ehn_desc_rx_set_on_chain(struct dma_desc *p, int end) ehn_desc_rx_set_on_chain() argument
99 static inline void ehn_desc_tx_set_on_chain(struct dma_desc *p, int end) ehn_desc_tx_set_on_chain() argument
115 static inline void ndesc_rx_set_on_chain(struct dma_desc *p, int end) ndesc_rx_set_on_chain() argument
/linux-4.4.14/arch/mips/jazz/
H A Dsetup.c33 .end = 0x1f,
38 .end = 0x5f,
43 .end = 0x8f,
48 .end = 0xdf,
124 .end = JAZZ_SCSI_BASE + 31,
129 .end = JAZZ_SCSI_DMA,
134 .end = JAZZ_SCSI_IRQ,
148 .end = JAZZ_ETHERNET_BASE + 0xff,
153 .end = JAZZ_ETHERNET_IRQ,
167 .end = 0x71,
172 .end = 8,
/linux-4.4.14/arch/arc/mm/
H A Dioremap.c19 unsigned long end; ioremap() local
22 end = paddr + size - 1; ioremap()
23 if (!size || (end < paddr)) ioremap()
46 unsigned long off, end; ioremap_prot() local
50 end = paddr + size - 1; ioremap_prot()
51 if ((!size) || (end < paddr)) ioremap_prot()
54 /* An early platform driver might end up here */ ioremap_prot()
64 size = PAGE_ALIGN(end + 1) - paddr; ioremap_prot()
/linux-4.4.14/arch/mips/vr41xx/common/
H A Drtc.c32 .end = 0x0b0000df,
37 .end = 0x0b0001df,
42 .end = ELAPSEDTIME_IRQ,
47 .end = RTCLONG1_IRQ,
55 .end = 0x0f00011f,
60 .end = 0x0f00013f,
65 .end = ELAPSEDTIME_IRQ,
70 .end = RTCLONG1_IRQ,
H A Dgiu.c33 .end = 0x0b00011f,
38 .end = 0x0b0002e3,
43 .end = GIUINT_IRQ,
51 .end = 0x0f00015f,
56 .end = GIUINT_IRQ,
64 .end = 0x0f000167,
69 .end = GIUINT_IRQ,
/linux-4.4.14/arch/arm/include/asm/hardware/
H A Dmemc.h19 #define video_set_dma(start,end,offset) \
22 memc_write (VDMA_END, (end - VDMA_XFERSIZE) >> 2); \
/linux-4.4.14/arch/arc/include/asm/
H A Dtlbflush.h17 void local_flush_tlb_kernel_range(unsigned long start, unsigned long end);
19 unsigned long start, unsigned long end);
21 unsigned long end);
32 unsigned long end);
34 extern void flush_tlb_kernel_range(unsigned long start, unsigned long end);
37 extern void flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end);
/linux-4.4.14/arch/metag/include/asm/
H A Dtlbflush.h16 * - flush_tlb_range(mm, start, end) flushes a range of pages
17 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages
18 * - flush_tlb_pgtables(mm, start, end) flushes a range of page tables
59 unsigned long start, unsigned long end) flush_tlb_range()
65 unsigned long start, unsigned long end) flush_tlb_pgtables()
71 unsigned long end) flush_tlb_kernel_range()
58 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_tlb_range() argument
64 flush_tlb_pgtables(struct mm_struct *mm, unsigned long start, unsigned long end) flush_tlb_pgtables() argument
70 flush_tlb_kernel_range(unsigned long start, unsigned long end) flush_tlb_kernel_range() argument
/linux-4.4.14/arch/microblaze/include/asm/
H A Dtlbflush.h37 unsigned long start, unsigned long end) local_flush_tlb_range()
40 #define flush_tlb_kernel_range(start, end) do { } while (0)
55 unsigned long start, unsigned long end) { }
63 #define flush_tlb_range(mm, start, end) BUG()
64 #define flush_tlb_pgtables(mm, start, end) BUG()
65 #define flush_tlb_kernel_range(start, end) BUG()
36 local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) local_flush_tlb_range() argument
54 flush_tlb_pgtables(struct mm_struct *mm, unsigned long start, unsigned long end) flush_tlb_pgtables() argument
/linux-4.4.14/arch/frv/include/asm/
H A Dtlbflush.h25 unsigned long start, unsigned long end);
42 #define flush_tlb_range(vma,start,end) \
45 __flush_tlb_range((vma)->vm_mm->context.id, start, end); \
59 #define flush_tlb_kernel_range(start, end) flush_tlb_all()
67 #define flush_tlb_range(mm,start,end) BUG()
68 #define flush_tlb_kernel_range(start, end) BUG()
/linux-4.4.14/mm/kasan/
H A Dkasan_init.c47 unsigned long end) zero_pte_populate()
55 while (addr + PAGE_SIZE <= end) { zero_pte_populate()
63 unsigned long end) zero_pmd_populate()
69 next = pmd_addr_end(addr, end); zero_pmd_populate()
71 if (IS_ALIGNED(addr, PMD_SIZE) && end - addr >= PMD_SIZE) { zero_pmd_populate()
81 } while (pmd++, addr = next, addr != end); zero_pmd_populate()
85 unsigned long end) zero_pud_populate()
91 next = pud_addr_end(addr, end); zero_pud_populate()
92 if (IS_ALIGNED(addr, PUD_SIZE) && end - addr >= PUD_SIZE) { zero_pud_populate()
106 } while (pud++, addr = next, addr != end); zero_pud_populate()
113 * @shadow_end - end of the memory range to populate
119 unsigned long end = (unsigned long)shadow_end; kasan_populate_zero_shadow() local
124 next = pgd_addr_end(addr, end); kasan_populate_zero_shadow()
126 if (IS_ALIGNED(addr, PGDIR_SIZE) && end - addr >= PGDIR_SIZE) { kasan_populate_zero_shadow()
151 } while (pgd++, addr = next, addr != end); kasan_populate_zero_shadow()
46 zero_pte_populate(pmd_t *pmd, unsigned long addr, unsigned long end) zero_pte_populate() argument
62 zero_pmd_populate(pud_t *pud, unsigned long addr, unsigned long end) zero_pmd_populate() argument
84 zero_pud_populate(pgd_t *pgd, unsigned long addr, unsigned long end) zero_pud_populate() argument
/linux-4.4.14/arch/arm64/include/asm/
H A Dcacheflush.h36 * Start addresses are inclusive and end addresses are exclusive; start
37 * addresses should be rounded down, end addresses up.
48 * flush_icache_range(start, end)
51 * region described by start, end.
53 * - end - virtual end address
55 * __flush_cache_user_range(start, end)
58 * region described by start, end.
60 * - end - virtual end address
68 extern void flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end);
69 extern void flush_icache_range(unsigned long start, unsigned long end);
71 extern long __flush_cache_user_range(unsigned long start, unsigned long end);
145 static inline void flush_cache_vmap(unsigned long start, unsigned long end) flush_cache_vmap() argument
149 static inline void flush_cache_vunmap(unsigned long start, unsigned long end) flush_cache_vunmap() argument
H A Dtlbflush.h33 * addresses are inclusive and end addresses are exclusive; it is safe to
45 * flush_tlb_range(mm,start,end)
51 * - end - end address (exclusive, may not be aligned)
108 unsigned long start, unsigned long end, __flush_tlb_range()
114 if ((end - start) > MAX_TLB_RANGE) { __flush_tlb_range()
120 end = asid | (end >> 12); __flush_tlb_range()
123 for (addr = start; addr < end; addr += 1 << (PAGE_SHIFT - 12)) { __flush_tlb_range()
133 unsigned long start, unsigned long end) flush_tlb_range()
135 __flush_tlb_range(vma, start, end, false); flush_tlb_range()
138 static inline void flush_tlb_kernel_range(unsigned long start, unsigned long end) flush_tlb_kernel_range() argument
142 if ((end - start) > MAX_TLB_RANGE) { flush_tlb_kernel_range()
148 end >>= 12; flush_tlb_kernel_range()
151 for (addr = start; addr < end; addr += 1 << (PAGE_SHIFT - 12)) flush_tlb_kernel_range()
107 __flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end, bool last_level) __flush_tlb_range() argument
132 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_tlb_range() argument
/linux-4.4.14/arch/arm/include/asm/
H A Doutercache.h50 * @end: end physical address, exclusive
52 static inline void outer_inv_range(phys_addr_t start, phys_addr_t end) outer_inv_range() argument
55 outer_cache.inv_range(start, end); outer_inv_range()
61 * @end: end physical address, exclusive
63 static inline void outer_clean_range(phys_addr_t start, phys_addr_t end) outer_clean_range() argument
66 outer_cache.clean_range(start, end); outer_clean_range()
72 * @end: end physical address, exclusive
74 static inline void outer_flush_range(phys_addr_t start, phys_addr_t end) outer_flush_range() argument
77 outer_cache.flush_range(start, end); outer_flush_range()
120 static inline void outer_inv_range(phys_addr_t start, phys_addr_t end) outer_inv_range() argument
122 static inline void outer_clean_range(phys_addr_t start, phys_addr_t end) outer_clean_range() argument
124 static inline void outer_flush_range(phys_addr_t start, phys_addr_t end) outer_flush_range() argument
/linux-4.4.14/drivers/media/platform/coda/
H A Dtrace.h92 __field(int, end)
100 __entry->end = meta->end;
104 TP_printk("minor = %d, index = %d, start = 0x%x, end = 0x%x, ctx = %d",
105 __entry->minor, __entry->index, __entry->start, __entry->end,
123 __field(int, end)
130 __entry->end = meta ? meta->end : 0;
134 TP_printk("minor = %d, start = 0x%x, end = 0x%x, ctx = %d",
135 __entry->minor, __entry->start, __entry->end, __entry->ctx)
/linux-4.4.14/fs/ceph/
H A Dmdsmap.c51 struct ceph_mdsmap *ceph_mdsmap_decode(void **p, void *end) ceph_mdsmap_decode() argument
63 ceph_decode_16_safe(p, end, version, bad); ceph_mdsmap_decode()
69 ceph_decode_need(p, end, 8*sizeof(u32) + sizeof(u64), bad); ceph_mdsmap_decode()
97 ceph_decode_need(p, end, sizeof(u64)*2 + 1 + sizeof(u32), bad); ceph_mdsmap_decode()
104 ceph_decode_need(p, end, ceph_mdsmap_decode()
116 ceph_decode_32_safe(p, end, namelen, bad); ceph_mdsmap_decode()
119 ceph_decode_32_safe(p, end, num_export_targets, bad); ceph_mdsmap_decode()
155 ceph_decode_32_safe(p, end, n, bad); ceph_mdsmap_decode()
160 ceph_decode_need(p, end, sizeof(u64)*(n+1), bad); ceph_mdsmap_decode()
175 start, end - start, true); ceph_mdsmap_decode()
/linux-4.4.14/arch/x86/kvm/
H A Dmtrr.c147 u64 end; member in struct:fixed_mtrr_segment
159 .end = 0x80000,
170 .end = 0xc0000,
181 .end = 0x100000,
218 static void fixed_mtrr_seg_unit_range(int seg, int unit, u64 *start, u64 *end) fixed_mtrr_seg_unit_range() argument
224 *end = *start + unit_size; fixed_mtrr_seg_unit_range()
225 WARN_ON(*end > mtrr_seg->end); fixed_mtrr_seg_unit_range()
233 > mtrr_seg->end); fixed_mtrr_seg_unit_range_index()
244 n = (mtrr_seg->end - mtrr_seg->start) >> mtrr_seg->range_shift; fixed_mtrr_seg_end_range_index()
248 static bool fixed_msr_to_range(u32 msr, u64 *start, u64 *end) fixed_msr_to_range() argument
255 fixed_mtrr_seg_unit_range(seg, unit, start, end); fixed_msr_to_range()
276 if (mtrr_seg->start <= addr && addr < mtrr_seg->end) fixed_mtrr_addr_to_seg()
302 static void var_mtrr_range(struct kvm_mtrr_range *range, u64 *start, u64 *end) var_mtrr_range() argument
313 *end = (*start | ~mask) + 1; var_mtrr_range()
319 gfn_t start, end; update_mtrr() local
330 if (fixed_msr_to_range(msr, &start, &end)) { update_mtrr()
335 end = ~0ULL; update_mtrr()
339 var_mtrr_range(&mtrr_state->var_ranges[index], &start, &end); update_mtrr()
342 kvm_zap_gfn_range(vcpu->kvm, gpa_to_gfn(start), gpa_to_gfn(end)); update_mtrr()
454 u64 end; member in struct:mtrr_iter
460 /* [start, end) is not fully covered in MTRRs? */
503 u64 start, end; match_var_range() local
505 var_mtrr_range(range, &start, &end); match_var_range()
506 if (!(start >= iter->end || end <= iter->start)) { match_var_range()
512 * [looker->start_max, looker->end). match_var_range()
517 iter->start_max = max(iter->start_max, end); match_var_range()
533 iter->partial_map |= iter->start_max < iter->end; __mtrr_lookup_var_next()
550 if (fixed_mtrr_range_end_addr(iter->seg, iter->index) >= iter->end) { mtrr_lookup_fixed_next()
584 struct kvm_mtrr *mtrr_state, u64 start, u64 end) mtrr_lookup_init()
588 iter->end = end; mtrr_lookup_init()
628 u64 start, end; kvm_mtrr_get_guest_memory_type() local
634 end = start + PAGE_SIZE; kvm_mtrr_get_guest_memory_type()
636 mtrr_for_each_mem_type(&iter, mtrr_state, start, end) { kvm_mtrr_get_guest_memory_type()
705 u64 start, end; kvm_mtrr_check_gfn_range_consistency() local
709 end = gfn_to_gpa(gfn + page_num); kvm_mtrr_check_gfn_range_consistency()
710 mtrr_for_each_mem_type(&iter, mtrr_state, start, end) { kvm_mtrr_check_gfn_range_consistency()
583 mtrr_lookup_init(struct mtrr_iter *iter, struct kvm_mtrr *mtrr_state, u64 start, u64 end) mtrr_lookup_init() argument
/linux-4.4.14/sound/firewire/oxfw/
H A Doxfw-stream.c49 goto end; set_rate()
54 end: set_rate()
139 goto end; start_stream()
147 goto end; start_stream()
156 goto end; start_stream()
161 goto end; start_stream()
166 goto end; start_stream()
173 goto end; start_stream()
181 end: start_stream()
229 goto end; snd_oxfw_stream_init_simplex()
235 goto end; snd_oxfw_stream_init_simplex()
250 end: snd_oxfw_stream_init_simplex()
282 goto end; snd_oxfw_stream_start_simplex()
290 goto end; snd_oxfw_stream_start_simplex()
298 goto end; snd_oxfw_stream_start_simplex()
308 goto end; snd_oxfw_stream_start_simplex()
317 goto end; snd_oxfw_stream_start_simplex()
327 goto end; snd_oxfw_stream_start_simplex()
339 end: snd_oxfw_stream_start_simplex()
402 goto end; snd_oxfw_stream_get_current_formation()
405 goto end; snd_oxfw_stream_get_current_formation()
409 end: snd_oxfw_stream_get_current_formation()
512 goto end; assume_stream_formats()
519 goto end; assume_stream_formats()
524 goto end; assume_stream_formats()
542 goto end; assume_stream_formats()
549 end: assume_stream_formats()
579 goto end; fill_stream_formats()
585 goto end; fill_stream_formats()
624 end: fill_stream_formats()
643 goto end; snd_oxfw_stream_discover()
646 goto end; snd_oxfw_stream_discover()
653 goto end; snd_oxfw_stream_discover()
675 goto end; snd_oxfw_stream_discover()
690 end: snd_oxfw_stream_discover()
709 goto end; snd_oxfw_stream_lock_try()
716 end: snd_oxfw_stream_lock_try()
726 goto end; snd_oxfw_stream_lock_release()
729 end: snd_oxfw_stream_lock_release()
/linux-4.4.14/arch/unicore32/kernel/
H A Dpuv3-core.c53 .end = io_v2p(PKUNITY_USB_BASE) + 0x3ff,
85 .end = io_v2p(PKUNITY_SDC_BASE) + 0xfff,
90 .end = IRQ_SDC,
98 .end = io_v2p(PKUNITY_UNIGFX_BASE) + 0xfff,
106 .end = io_v2p(PKUNITY_RTC_BASE) + 0xff,
111 .end = IRQ_RTCAlarm,
116 .end = IRQ_RTC,
124 .end = io_v2p(PKUNITY_OST_BASE) + 0xff,
132 .end = io_v2p(PKUNITY_UART0_BASE) + 0xff,
137 .end = IRQ_UART0,
145 .end = io_v2p(PKUNITY_UART1_BASE) + 0xff,
150 .end = IRQ_UART1,
158 .end = io_v2p(PKUNITY_UMAL_BASE) + 0x1fff,
163 .end = IRQ_UMAL,
/linux-4.4.14/arch/unicore32/mm/
H A Dcache-ucv2.S44 * __cpuc_flush_user_range(start, end, flags)
49 * - end - end address (exclusive, may not be aligned)
83 * __cpuc_coherent_kern_range(start,end)
84 * __cpuc_coherent_user_range(start,end)
91 * - end - virtual end address of region
145 * __cpuc_dma_clean_range(start,end)
147 * - end - virtual end address of region
182 * __cpuc_dma_inv_range(start,end)
183 * __cpuc_dma_flush_range(start,end)
185 * - end - virtual end address of region
H A Dinit.c72 struct page *page, *end; for_each_bank() local
78 end = pfn_to_page(pfn2 - 1) + 1; for_each_bank()
93 } while (page < end); for_each_bank()
115 unsigned long start, end; for_each_bank() local
118 end = bank_pfn_end(bank); for_each_bank()
122 if (*max_high < end) for_each_bank()
123 *max_high = end; for_each_bank()
126 if (*max_low < end) for_each_bank()
127 *max_low = end; for_each_bank()
158 unsigned long end = memblock_region_memory_end_pfn(reg); for_each_memblock() local
160 if (end >= end_pfn) for_each_memblock()
161 end = end_pfn; for_each_memblock()
162 if (start >= end) for_each_memblock()
165 free_bootmem(__pfn_to_phys(start), (end - start) << PAGE_SHIFT); for_each_memblock()
171 unsigned long end = memblock_region_reserved_end_pfn(reg); for_each_memblock() local
173 if (end >= end_pfn) for_each_memblock()
174 end = end_pfn; for_each_memblock()
175 if (start >= end) for_each_memblock()
179 (end - start) << PAGE_SHIFT, BOOTMEM_DEFAULT); for_each_memblock()
208 unsigned long end = memblock_region_memory_end_pfn(reg); for_each_memblock() local
211 unsigned long low_end = min(end, max_low); for_each_memblock()
330 * round start upwards and end downwards. free_memmap()
369 * memmap entries are valid from the bank end aligned to for_each_bank()
438 void free_initrd_mem(unsigned long start, unsigned long end) free_initrd_mem() argument
441 free_reserved_area((void *)start, (void *)end, -1, "initrd"); free_initrd_mem()
/linux-4.4.14/fs/hfsplus/
H A Dbitmap.c24 __be32 *pptr, *curr, *end; hfsplus_block_allocate() local
46 end = pptr + PAGE_CACHE_BITS / 32; hfsplus_block_allocate()
48 end = pptr + ((size + 31) & (PAGE_CACHE_BITS - 1)) / 32; hfsplus_block_allocate()
64 while (curr < end) { hfsplus_block_allocate()
88 end = pptr + PAGE_CACHE_BITS / 32; hfsplus_block_allocate()
90 end = pptr + ((size + 31) & (PAGE_CACHE_BITS - 1)) / 32; hfsplus_block_allocate()
117 while (curr < end) { hfsplus_block_allocate()
139 end = pptr + PAGE_CACHE_BITS / 32; hfsplus_block_allocate()
142 /* do any partial u32 at end */ hfsplus_block_allocate()
168 __be32 *pptr, *curr, *end; hfsplus_block_free() local
189 end = pptr + PAGE_CACHE_BITS / 32; hfsplus_block_free()
208 while (curr < end) { hfsplus_block_free()
223 end = pptr + PAGE_CACHE_BITS / 32; hfsplus_block_free()
226 /* do any partial u32 at end */ hfsplus_block_free()
/linux-4.4.14/arch/powerpc/mm/
H A Dtlb_hash32.c63 * Called at the end of a mmu_gather operation to make sure the
82 * - flush_tlb_range(vma, start, end) flushes a range of pages
83 * - flush_tlb_kernel_range(start, end) flushes kernel pages
91 unsigned long end) flush_range()
103 if (start >= end) flush_range()
105 end = (end - 1) | ~PAGE_MASK; flush_range()
109 if (pmd_end > end) flush_range()
110 pmd_end = end; flush_range()
115 if (pmd_end == end) flush_range()
125 void flush_tlb_kernel_range(unsigned long start, unsigned long end) flush_tlb_kernel_range() argument
127 flush_range(&init_mm, start, end); flush_tlb_kernel_range()
176 unsigned long end) flush_tlb_range()
178 flush_range(vma->vm_mm, start, end); flush_tlb_range()
90 flush_range(struct mm_struct *mm, unsigned long start, unsigned long end) flush_range() argument
175 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_tlb_range() argument
/linux-4.4.14/drivers/staging/lustre/lustre/include/
H A Dinterval_tree.h56 __u64 end; member in struct:interval_node::interval_node_extent
72 return node->in_extent.end; interval_high()
76 __u64 start, __u64 end) interval_set()
78 LASSERT(start <= end); interval_set()
80 node->in_extent.end = end; interval_set()
81 node->in_max_high = end; interval_set()
75 interval_set(struct interval_node *node, __u64 start, __u64 end) interval_set() argument
/linux-4.4.14/arch/sh/boards/
H A Dboard-sh2007.c35 .end = SMC0_BASE + 0xff,
40 .end = evt2irq(0x240),
48 .end = SMC1_BASE + 0xff,
53 .end = evt2irq(0x280),
81 .end = CF_BASE + CF_OFFSET + 0x0f,
86 .end = CF_BASE + CF_OFFSET + 0x20f,
91 .end = evt2irq(0x2c0),
/linux-4.4.14/arch/ia64/include/asm/
H A Dmeminit.h28 u64 end; /* virtual address of end of element + 1 */ member in struct:rsvd_region
37 extern int filter_rsvd_memory (u64 start, u64 end, void *arg);
38 extern int filter_memory (u64 start, u64 end, void *arg);
43 extern int reserve_elfcorehdr(u64 *start, u64 *end);
65 extern int find_largest_hole(u64 start, u64 end, void *arg);
66 extern int create_mem_map_page_table(u64 start, u64 end, void *arg);
H A Dpatch.h21 extern void ia64_patch_mckinley_e9 (unsigned long start, unsigned long end);
22 extern void ia64_patch_vtop (unsigned long start, unsigned long end);
24 extern void ia64_patch_rse (unsigned long start, unsigned long end);
/linux-4.4.14/arch/unicore32/include/asm/
H A Dcacheflush.h33 * Start addresses are inclusive and end addresses are exclusive;
34 * start addresses should be rounded down, end addresses up.
55 * flush_user_range(start, end, flags)
60 * - end - user end address (exclusive, page aligned)
63 * coherent_kern_range(start, end)
66 * region described by start, end. If you have non-snooping
69 * - end - virtual end address
71 * coherent_user_range(start, end)
74 * region described by start, end. If you have non-snooping
77 * - end - virtual end address
88 * dma_flush_range(start, end)
92 * - end - virtual end address
141 unsigned long start, unsigned long end);
152 #define flush_cache_user_range(vma, start, end) \
153 __cpuc_coherent_user_range((start) & PAGE_MASK, PAGE_ALIGN(end))
203 static inline void flush_cache_vmap(unsigned long start, unsigned long end) flush_cache_vmap() argument
207 static inline void flush_cache_vunmap(unsigned long start, unsigned long end) flush_cache_vunmap() argument
/linux-4.4.14/sound/firewire/fireworks/
H A Dfireworks_stream.c32 goto end; init_stream()
39 end: init_stream()
65 goto end; start_stream()
79 goto end; start_stream()
85 goto end; start_stream()
93 goto end; start_stream()
101 end: start_stream()
170 goto end; snd_efw_stream_init_duplex()
196 goto end; snd_efw_stream_init_duplex()
205 end: snd_efw_stream_init_duplex()
222 goto end; snd_efw_stream_start_duplex()
226 goto end; snd_efw_stream_start_duplex()
243 goto end; snd_efw_stream_start_duplex()
254 goto end; snd_efw_stream_start_duplex()
269 goto end; snd_efw_stream_start_duplex()
275 goto end; snd_efw_stream_start_duplex()
288 end: snd_efw_stream_start_duplex()
357 goto end; snd_efw_stream_lock_try()
364 end: snd_efw_stream_lock_try()
374 goto end; snd_efw_stream_lock_release()
377 end: snd_efw_stream_lock_release()
/linux-4.4.14/drivers/block/drbd/
H A Ddrbd_interval.c6 * interval_end - return end of @node
12 return this->end; interval_end()
16 * compute_subtree_last - compute end of @node
18 * The end of an interval is the highest (start + (size >> 9)) value of this
19 * node and of its children. Called for @node and its parents whenever the end
41 sector_t, end, compute_subtree_last);
59 if (here->end < this_end) drbd_insert_interval()
60 here->end = this_end; drbd_insert_interval()
73 this->end = this_end; drbd_insert_interval()
138 sector_t end = sector + (size >> 9); drbd_find_overlap() local
150 } else if (here->sector < end && drbd_find_overlap()
166 sector_t end = sector + (size >> 9); drbd_next_overlap() local
174 if (i->sector >= end) drbd_next_overlap()
/linux-4.4.14/arch/m68k/sun3x/
H A Ddvma.c86 unsigned long end; dvma_map_cpu() local
92 end = PAGE_ALIGN(vaddr + len); dvma_map_cpu()
109 if((end & PGDIR_MASK) > (vaddr & PGDIR_MASK)) dvma_map_cpu()
112 end2 = end; dvma_map_cpu()
142 } while(vaddr < end); dvma_map_cpu()
154 unsigned long end, index; dvma_map_iommu() local
157 end = ((baddr+len) >> DVMA_PAGE_SHIFT); dvma_map_iommu()
160 end++; dvma_map_iommu()
162 for(; index < end ; index++) { dvma_map_iommu()
176 for(index = (baddr >> DVMA_PAGE_SHIFT); index < end; index++) dvma_map_iommu()
186 int index, end; dvma_unmap_iommu() local
190 end = (DVMA_PAGE_ALIGN(baddr+len) >> DVMA_PAGE_SHIFT); dvma_unmap_iommu()
192 for(; index < end ; index++) { dvma_unmap_iommu()
/linux-4.4.14/arch/mips/sni/
H A Dpcit.c64 .end = 0x71,
69 .end = 8,
87 .end = 0x03bfffffUL,
95 .end = 0x1f,
100 .end = 0x5f,
105 .end = 0x6f,
110 .end = 0x8f,
115 .end = 0xdf,
120 .end = 0xcfb,
125 .end = 0xcff,
146 .end = 0x1fbfffffUL,
265 ioport_resource.end = sni_io_resource.end; sni_pcit_init()
H A Da20r.c46 .end = 0x1c081fff,
60 .end = 0x18000004,
65 .end = 0x18010004,
70 .end = 0x1ff00020,
75 .end = 22,
92 .end = 0x190fffff,
97 .end = 19,
111 .end = 0x1c0700ff,
116 .end = 20,
/linux-4.4.14/arch/sh/drivers/superhyway/
H A Dops-sh4-202.c31 .end = PHYS_EMI_CBLOCK + 0x00300000 - 1,
36 .end = PHYS_EMI_DBLOCK + 0x08000000 - 1,
50 .end = PHYS_FEMI_CBLOCK + 0x00100000 - 1,
55 .end = PHYS_FEMI_DBLOCK + 0x08000000 - 1,
69 .end = P4SEGADDR(0x1e7ffff8 + (sizeof(u32) * 2) - 1),
74 .end = PHYS_EPBR_BLOCK + 0x00a00000 - 1,
87 .end = PHYS_DMAC_BLOCK + 0x00100000 - 1,
100 .end = P4SEGADDR(0x1ffffff8 + (sizeof(u32) * 2) - 1),
105 .end = PHYS_PBR_BLOCK + 0x00400000 - (sizeof(u32) * 2) - 1,
/linux-4.4.14/arch/arm/mach-dove/
H A Dmpp.c21 int end; member in struct:dove_mpp_grp
28 .end = 39,
32 .end = 45,
36 .end = 51,
40 .end = 61,
44 .end = 63,
50 static void __init dove_mpp_gpio_mode(int start, int end, int gpio_mode) dove_mpp_gpio_mode() argument
54 for (i = start; i <= end; i++) dove_mpp_gpio_mode()
141 dove_mpp_grp[num].end, gpio_mode); dove_mpp_conf_grp()
/linux-4.4.14/tools/testing/selftests/timers/
H A Draw_skew.c72 long long diff_timespec(struct timespec start, struct timespec end) diff_timespec() argument
77 end_ns = ts_to_nsec(end); diff_timespec()
83 struct timespec start, mid, end; get_monotonic_and_raw() local
92 clock_gettime(CLOCK_MONOTONIC, &end); get_monotonic_and_raw()
94 newdiff = diff_timespec(start, end); get_monotonic_and_raw()
98 tmp = (ts_to_nsec(start) + ts_to_nsec(end))/2; get_monotonic_and_raw()
106 struct timespec mon, raw, start, end; main() local
130 end = mon; main()
135 interval = diff_timespec(start, end); main()
/linux-4.4.14/arch/sparc/mm/
H A Dgup.c21 unsigned long end, int write, struct page **pages, int *nr) gup_pte_range()
64 } while (ptep++, addr += PAGE_SIZE, addr != end); gup_pte_range()
70 unsigned long end, int write, struct page **pages, gup_huge_pmd()
92 } while (addr += PAGE_SIZE, addr != end); gup_huge_pmd()
118 static int gup_pmd_range(pud_t pud, unsigned long addr, unsigned long end, gup_pmd_range() argument
128 next = pmd_addr_end(addr, end); gup_pmd_range()
138 } while (pmdp++, addr = next, addr != end); gup_pmd_range()
143 static int gup_pud_range(pgd_t pgd, unsigned long addr, unsigned long end, gup_pud_range() argument
153 next = pud_addr_end(addr, end); gup_pud_range()
158 } while (pudp++, addr = next, addr != end); gup_pud_range()
167 unsigned long addr, len, end; __get_user_pages_fast() local
175 end = start + len; __get_user_pages_fast()
182 next = pgd_addr_end(addr, end); __get_user_pages_fast()
187 } while (pgdp++, addr = next, addr != end); __get_user_pages_fast()
197 unsigned long addr, len, end; get_user_pages_fast() local
205 end = start + len; get_user_pages_fast()
230 next = pgd_addr_end(addr, end); get_user_pages_fast()
235 } while (pgdp++, addr = next, addr != end); get_user_pages_fast()
239 VM_BUG_ON(nr != (end - start) >> PAGE_SHIFT); get_user_pages_fast()
253 (end - start) >> PAGE_SHIFT, write, 0, pages); get_user_pages_fast()
20 gup_pte_range(pmd_t pmd, unsigned long addr, unsigned long end, int write, struct page **pages, int *nr) gup_pte_range() argument
69 gup_huge_pmd(pmd_t *pmdp, pmd_t pmd, unsigned long addr, unsigned long end, int write, struct page **pages, int *nr) gup_huge_pmd() argument
/linux-4.4.14/Documentation/watchdog/src/
H A Dwatchdog-test.c60 goto end; main()
66 goto end; main()
72 goto end; main()
82 goto end; main()
95 end: main()
/linux-4.4.14/arch/s390/mm/
H A Dvmem.c80 unsigned long end = start + size; vmem_add_mem() local
88 while (address < end) { vmem_add_mem()
99 !(address & ~PUD_MASK) && (address + PUD_SIZE <= end)) { vmem_add_mem()
116 !(address & ~PMD_MASK) && (address + PMD_SIZE <= end)) { vmem_add_mem()
148 unsigned long end = start + size; vmem_remove_range() local
157 while (address < end) { vmem_remove_range()
187 flush_tlb_kernel_range(start, end); vmem_remove_range()
193 int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node) vmemmap_populate() argument
202 for (address = start; address < end;) { vmemmap_populate()
264 void vmemmap_free(unsigned long start, unsigned long end) vmemmap_free() argument
369 phys_addr_t start, end; vmem_map_init() local
375 end = reg->base + reg->size - 1; for_each_memblock()
376 if (start >= ro_end || end <= ro_start) for_each_memblock()
377 vmem_add_mem(start, end - start, 0); for_each_memblock()
378 else if (start >= ro_start && end <= ro_end) for_each_memblock()
379 vmem_add_mem(start, end - start, 1); for_each_memblock()
382 vmem_add_mem(ro_end, end - ro_end, 0); for_each_memblock()
383 } else if (end < ro_end) { for_each_memblock()
385 vmem_add_mem(ro_start, end - ro_start, 1); for_each_memblock()
389 vmem_add_mem(ro_end, end - ro_end, 0); for_each_memblock()
/linux-4.4.14/arch/score/include/asm/
H A Dcacheflush.h10 unsigned long start, unsigned long end);
15 extern void flush_icache_range(unsigned long start, unsigned long end);
16 extern void flush_dcache_range(unsigned long start, unsigned long end);
25 #define flush_cache_vmap(start, end) do {} while (0)
26 #define flush_cache_vunmap(start, end) do {} while (0)
/linux-4.4.14/arch/sh/boards/mach-sh03/
H A Dsetup.c35 .end = 0x1f0 + 8,
40 .end = 0x1f0 +8 + 0x206 + 8,
59 .end = 0xa0800000,
93 cf_ide_resources[0].end += (unsigned long)cf_ide_base; sh03_devices_setup()
95 cf_ide_resources[1].end += (unsigned long)cf_ide_base; sh03_devices_setup()
/linux-4.4.14/drivers/firmware/efi/
H A Dfake_mem.c59 u64 start, end, m_start, m_end, m_attr; efi_fake_memmap() local
74 end = start + (md->num_pages << EFI_PAGE_SHIFT) - 1; efi_fake_memmap()
79 m_end = fake_mems[i].range.end; efi_fake_memmap()
83 if (start < m_end && m_end < end) efi_fake_memmap()
86 if (start < m_start && m_start < end) { efi_fake_memmap()
88 if (m_end < end) efi_fake_memmap()
91 if (end <= m_end) efi_fake_memmap()
119 end = md->phys_addr + (md->num_pages << EFI_PAGE_SHIFT) - 1; efi_fake_memmap()
124 m_end = fake_mems[i].range.end; efi_fake_memmap()
127 if (m_start <= start && end <= m_end) efi_fake_memmap()
131 (start < m_end && m_end < end)) { efi_fake_memmap()
141 md->num_pages = (end - md->phys_addr + 1) >> efi_fake_memmap()
145 if ((start < m_start && m_start < end) && m_end < end) { efi_fake_memmap()
162 md->num_pages = (end - m_end) >> efi_fake_memmap()
166 if ((start < m_start && m_start < end) && efi_fake_memmap()
167 (end <= m_end)) { efi_fake_memmap()
176 md->num_pages = (end - md->phys_addr + 1) >> efi_fake_memmap()
219 fake_mems[nr_fake_mem].range.end = start + mem_size - 1; setup_fake_mem()
233 fake_mems[i].range.end); setup_fake_mem()
/linux-4.4.14/sound/firewire/dice/
H A Ddice-stream.c71 goto end; keep_resources()
83 end: keep_resources()
108 goto end; start_stream()
138 goto end; start_stream()
155 goto end; start_stream()
162 end: start_stream()
173 goto end; get_sync_mode()
189 end: get_sync_mode()
201 goto end; snd_dice_stream_start_duplex()
205 goto end; snd_dice_stream_start_duplex()
223 goto end; snd_dice_stream_start_duplex()
240 goto end; snd_dice_stream_start_duplex()
248 goto end; snd_dice_stream_start_duplex()
255 goto end; snd_dice_stream_start_duplex()
263 goto end; snd_dice_stream_start_duplex()
275 end: snd_dice_stream_start_duplex()
306 goto end; init_stream()
314 end: init_stream()
343 goto end; snd_dice_stream_init_duplex()
348 end: snd_dice_stream_init_duplex()
/linux-4.4.14/tools/perf/scripts/python/
H A Dnet_dropmonitor.py34 # kallsyms[i][0] > loc for all end <= i < len(kallsyms)
35 start, end = -1, len(kallsyms)
36 while end != start + 1:
37 pivot = (start + end) // 2
39 end = pivot
/linux-4.4.14/arch/arc/kernel/
H A Dunwind.c172 const void *end, signed ptrType);
198 const u8 *end = header_start + header_size; init_unwind_table() local
210 || (void *)read_pointer(&ptr, end, header_start[1]) != table_start init_unwind_table()
212 || read_pointer(&ptr, end, header_start[2]) <= 0 init_unwind_table()
463 static uleb128_t get_uleb128(const u8 **pcur, const u8 *end) get_uleb128() argument
469 for (shift = 0, value = 0; cur < end; shift += 7) { get_uleb128()
472 cur = end + 1; get_uleb128()
484 static sleb128_t get_sleb128(const u8 **pcur, const u8 *end) get_sleb128() argument
490 for (shift = 0, value = 0; cur < end; shift += 7) { get_sleb128()
493 cur = end + 1; get_sleb128()
531 static unsigned long read_pointer(const u8 **pLoc, const void *end, read_pointer() argument
549 if (end < (const void *)(ptr.p16u + 1)) read_pointer()
558 if (end < (const void *)(ptr.p32u + 1)) read_pointer()
571 if (end < (const void *)(ptr.pul + 1)) read_pointer()
577 value = ptrType & DW_EH_PE_signed ? get_sleb128(&ptr.p8, end) read_pointer()
578 : get_uleb128(&ptr.p8, end); read_pointer()
579 if ((const void *)ptr.p8 > end) read_pointer()
609 const u8 *end = (const u8 *)(cie + 1) + *cie; fde_pointer_type() local
618 ptr = memchr(aug, 0, end - ptr); fde_pointer_type()
623 get_uleb128(&ptr, end); /* skip code alignment */ fde_pointer_type()
624 get_sleb128(&ptr, end); /* skip data alignment */ fde_pointer_type()
626 version <= 1 ? (void) ++ptr : (void)get_uleb128(&ptr, end); fde_pointer_type()
627 len = get_uleb128(&ptr, end); /* augmentation length */ fde_pointer_type()
629 if (ptr + len < ptr || ptr + len > end) fde_pointer_type()
632 end = ptr + len; fde_pointer_type()
634 if (ptr >= end) fde_pointer_type()
643 if (!read_pointer(&ptr, end, ptrType) fde_pointer_type()
644 || ptr > end) fde_pointer_type()
698 static int processCFI(const u8 *start, const u8 *end, unsigned long targetLoc, processCFI() argument
717 for (ptr.p8 = start; result && ptr.p8 < end;) { processCFI()
729 state->loc = read_pointer(&ptr.p8, end, processCFI()
737 result = ptr.p8 < end processCFI()
744 result = ptr.p8 <= end + 2 processCFI()
750 result = ptr.p8 <= end + 4 processCFI()
754 value = get_uleb128(&ptr.p8, end); processCFI()
757 get_uleb128(&ptr.p8, end), state); processCFI()
760 value = get_uleb128(&ptr.p8, end); processCFI()
762 get_uleb128(&ptr.p8, end), state); processCFI()
765 value = get_uleb128(&ptr.p8, end); processCFI()
767 get_sleb128(&ptr.p8, end), state); processCFI()
770 value = get_uleb128(&ptr.p8, end); processCFI()
772 get_sleb128(&ptr.p8, end), state); processCFI()
780 set_rule(get_uleb128(&ptr.p8, end), Nowhere, 0, processCFI()
785 value = get_uleb128(&ptr.p8, end); processCFI()
788 get_uleb128(&ptr.p8, end), state); processCFI()
814 processCFI(start, end, 0, ptrType, processCFI()
822 state->cfa.reg = get_uleb128(&ptr.p8, end); processCFI()
826 state->cfa.offs = get_uleb128(&ptr.p8, end); processCFI()
831 state->cfa.reg = get_uleb128(&ptr.p8, end); processCFI()
834 state->cfa.offs = get_sleb128(&ptr.p8, end) processCFI()
839 state->cfa.reg = get_uleb128(&ptr.p8, end); processCFI()
845 get_uleb128(&ptr.p8, end); processCFI()
848 value = get_uleb128(&ptr.p8, end); processCFI()
852 end), processCFI()
869 set_rule(value, Memory, get_uleb128(&ptr.p8, end), processCFI()
878 if (ptr.p8 > end) processCFI()
884 return result && ptr.p8 == end && (targetLoc == 0 || ( processCFI()
887 never reaches the end of the function. processCFI()
897 const u8 *ptr = NULL, *end = NULL; arc_unwind() local
952 end = hdr + table->hdrsz; arc_unwind()
953 if (tableSize && read_pointer(&ptr, end, hdr[1]) arc_unwind()
955 && (i = read_pointer(&ptr, end, hdr[2])) > 0 arc_unwind()
956 && i == (end - ptr) / (2 * tableSize) arc_unwind()
957 && !((end - ptr) % (2 * tableSize))) { arc_unwind()
1016 end = (const u8 *)(cie + 1) + *cie; arc_unwind()
1021 while (++ptr < end && *ptr) { arc_unwind()
1038 if (ptr >= end || *ptr) arc_unwind()
1045 state.codeAlign = get_uleb128(&ptr, end); arc_unwind()
1047 state.dataAlign = get_sleb128(&ptr, end); arc_unwind()
1048 if (state.codeAlign == 0 || state.dataAlign == 0 || ptr >= end) arc_unwind()
1053 end); arc_unwind()
1061 uleb128_t augSize = get_uleb128(&ptr, end); arc_unwind()
1065 if (ptr > end || retAddrReg >= ARRAY_SIZE(reg_info) arc_unwind()
1075 state.cieEnd = end; arc_unwind()
1076 end = (const u8 *)(fde + 1) + *fde; arc_unwind()
1079 uleb128_t augSize = get_uleb128(&ptr, end); arc_unwind()
1081 if ((ptr += augSize) > end) arc_unwind()
1139 if (!processCFI(ptr, end, pc, ptrType, &state) arc_unwind()
/linux-4.4.14/fs/btrfs/
H A Dextent_io.h71 u64 start, u64 end, int *page_started,
73 int (*writepage_start_hook)(struct page *page, u64 start, u64 end);
74 int (*writepage_io_hook)(struct page *page, u64 start, u64 end);
81 struct page *page, u64 start, u64 end,
83 int (*writepage_end_io_hook)(struct page *page, u64 start, u64 end,
107 u64 end; /* inclusive */ member in struct:extent_state
202 int lock_extent(struct extent_io_tree *tree, u64 start, u64 end);
203 int lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
205 int unlock_extent(struct extent_io_tree *tree, u64 start, u64 end);
206 int unlock_extent_cached(struct extent_io_tree *tree, u64 start, u64 end,
208 int try_lock_extent(struct extent_io_tree *tree, u64 start, u64 end);
221 int test_range_bit(struct extent_io_tree *tree, u64 start, u64 end,
224 int clear_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
226 int clear_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
229 int clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
232 int set_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
234 int set_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
237 int set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
240 int set_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end,
242 int clear_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end,
244 int set_extent_new(struct extent_io_tree *tree, u64 start, u64 end,
246 int set_extent_dirty(struct extent_io_tree *tree, u64 start, u64 end,
248 int clear_extent_dirty(struct extent_io_tree *tree, u64 start, u64 end,
250 int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
253 int set_extent_delalloc(struct extent_io_tree *tree, u64 start, u64 end,
255 int set_extent_defrag(struct extent_io_tree *tree, u64 start, u64 end,
266 u64 start, u64 end, get_extent_t *get_extent,
341 int extent_range_clear_dirty_for_io(struct inode *inode, u64 start, u64 end);
342 int extent_range_redirty_for_io(struct inode *inode, u64 start, u64 end);
343 int extent_clear_unlock_delalloc(struct inode *inode, u64 start, u64 end,
360 int end_extent_writepage(struct page *page, int err, u64 start, u64 end);
383 void btrfs_free_io_failure_record(struct inode *inode, u64 start, u64 end);
384 int btrfs_get_io_failure_record(struct inode *inode, u64 start, u64 end,
397 u64 *end, u64 max_bytes);
/linux-4.4.14/arch/m68k/include/asm/
H A Dcacheflush_mm.h30 static inline void clear_cf_icache(unsigned long start, unsigned long end) clear_cf_icache() argument
39 static inline void clear_cf_dcache(unsigned long start, unsigned long end) clear_cf_dcache() argument
48 static inline void clear_cf_bcache(unsigned long start, unsigned long end) clear_cf_bcache() argument
59 * The start and end addresses are cache line numbers not memory addresses.
61 static inline void flush_cf_icache(unsigned long start, unsigned long end) flush_cf_icache() argument
65 for (set = start; set <= end; set += (0x10 - 3)) { flush_cf_icache()
79 static inline void flush_cf_dcache(unsigned long start, unsigned long end) flush_cf_dcache() argument
83 for (set = start; set <= end; set += (0x10 - 3)) { flush_cf_dcache()
97 static inline void flush_cf_bcache(unsigned long start, unsigned long end) flush_cf_bcache() argument
101 for (set = start; set <= end; set += (0x10 - 3)) { flush_cf_bcache()
192 #define flush_cache_vmap(start, end) flush_cache_all()
193 #define flush_cache_vunmap(start, end) flush_cache_all()
207 unsigned long end) flush_cache_range()
225 unsigned long addr, start, end; __flush_page_to_ram() local
228 end = (addr + PAGE_SIZE - 1) & ICACHE_SET_MASK; __flush_page_to_ram()
229 if (start > end) { __flush_page_to_ram()
230 flush_cf_bcache(0, end); __flush_page_to_ram()
231 end = ICACHE_MAX_ADDR; __flush_page_to_ram()
233 flush_cf_bcache(start, end); __flush_page_to_ram()
205 flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_cache_range() argument
/linux-4.4.14/arch/x86/purgatory/
H A Dpurgatory.c44 struct sha_region *ptr, *end; verify_sha256_digest() local
49 end = &sha_regions[sizeof(sha_regions)/sizeof(sha_regions[0])]; verify_sha256_digest()
50 for (ptr = sha_regions; ptr < end; ptr++) verify_sha256_digest()
/linux-4.4.14/arch/sh/boards/mach-se/7206/
H A Dsetup.c22 .end = PA_SMSC + 0x300 + 0x020 - 1,
27 .end = 64,
57 .end = PA_LED,
/linux-4.4.14/arch/m68k/sun3/
H A Ddvma.c49 unsigned long end; dvma_map_iommu() local
54 end = vaddr + len; dvma_map_iommu()
56 while(vaddr < end) { dvma_map_iommu()
/linux-4.4.14/arch/ia64/kernel/
H A Dpatch.c100 ia64_patch_vtop (unsigned long start, unsigned long end) ia64_patch_vtop() argument
105 while (offp < (s32 *) end) { ia64_patch_vtop()
123 ia64_patch_rse (unsigned long start, unsigned long end) ia64_patch_rse() argument
128 while (offp < (s32 *) end) { ia64_patch_rse()
141 ia64_patch_mckinley_e9 (unsigned long start, unsigned long end) ia64_patch_mckinley_e9() argument
158 while (offp < (s32 *) end) { ia64_patch_mckinley_e9()
172 patch_fsyscall_table (unsigned long start, unsigned long end) patch_fsyscall_table() argument
178 while (offp < (s32 *) end) { patch_fsyscall_table()
189 patch_brl_fsys_bubble_down (unsigned long start, unsigned long end) patch_brl_fsys_bubble_down() argument
195 while (offp < (s32 *) end) { patch_brl_fsys_bubble_down()
221 s32 * end = (s32 *) __end___phys_stack_reg_patchlist; ia64_patch_phys_stack_reg() local
228 while (offp < end) { ia64_patch_phys_stack_reg()
/linux-4.4.14/arch/avr32/kernel/
H A Dsetup.c51 .end = 0,
57 .end = 0,
92 mem->start, mem->end); resource_init()
103 res->name, res->start, res->end); resource_init()
108 add_physical_memory(resource_size_t start, resource_size_t end) add_physical_memory() argument
114 if (end < next->start) add_physical_memory()
116 if (start <= next->end) { add_physical_memory()
121 start, end, next->start, next->end); add_physical_memory()
129 start, end); add_physical_memory()
135 new->end = end; add_physical_memory()
143 add_reserved_region(resource_size_t start, resource_size_t end, add_reserved_region() argument
148 if (end < start) add_reserved_region()
156 if (end < next->start) add_reserved_region()
158 if (start <= next->end) add_reserved_region()
164 new->end = end; add_reserved_region()
185 if (target <= res->end) find_free_region()
186 target = ALIGN(res->end + 1, align); find_free_region()
189 if ((target + size) > (mem->end + 1)) find_free_region()
190 return mem->end + 1; find_free_region()
205 if (target <= mem->end) { alloc_reserved_region()
300 system_ram->end = system_ram->start + size - 1; early_mem()
318 unsigned long start, end; parse_tag_mem() local
329 end = tag->u.mem_range.addr + tag->u.mem_range.size - 1; parse_tag_mem()
331 add_physical_memory(start, end); parse_tag_mem()
480 printk(" %08x-%08x\n", res->start, res->end); setup_bootmem()
484 res->start, res->end, res->name); setup_bootmem()
493 max_low_pfn = max_pfn = PFN_DOWN(res->end + 1); setup_bootmem()
543 && res->end < PFN_PHYS(max_pfn)) setup_bootmem()
568 kernel_code.end = __pa(init_mm.end_code - 1); setup_arch()
570 kernel_data.end = __pa(init_mm.brk - 1); setup_arch()

Completed in 6938 milliseconds

1234567891011>>