/linux-4.4.14/kernel/ |
H A D | range.c | 10 int add_range(struct range *range, int az, int nr_range, u64 start, u64 end) add_range() argument 12 if (start >= end) add_range() 20 range[nr_range].end = end; add_range() 28 u64 start, u64 end) add_range_with_merge() 32 if (start >= end) add_range_with_merge() 35 /* get new start/end: */ add_range_with_merge() 39 if (!range[i].end) add_range_with_merge() 43 common_end = min(range[i].end, end); add_range_with_merge() 47 /* new start/end, will add it back at last */ add_range_with_merge() 49 end = max(range[i].end, end); add_range_with_merge() 54 range[nr_range - 1].end = 0; add_range_with_merge() 60 return add_range(range, az, nr_range, start, end); add_range_with_merge() 63 void subtract_range(struct range *range, int az, u64 start, u64 end) subtract_range() argument 67 if (start >= end) subtract_range() 71 if (!range[j].end) subtract_range() 74 if (start <= range[j].start && end >= range[j].end) { subtract_range() 76 range[j].end = 0; subtract_range() 80 if (start <= range[j].start && end < range[j].end && subtract_range() 81 range[j].start < end) { subtract_range() 82 range[j].start = end; subtract_range() 87 if (start > range[j].start && end >= range[j].end && subtract_range() 88 range[j].end > start) { subtract_range() 89 range[j].end = start; subtract_range() 93 if (start > range[j].start && end < range[j].end) { subtract_range() 96 if (range[i].end == 0) subtract_range() 100 range[i].end = range[j].end; subtract_range() 101 range[i].start = end; subtract_range() 106 range[j].end = start; subtract_range() 129 if (range[i].end) clean_sort_range() 132 if (range[j].end) { clean_sort_range() 140 range[i].end = range[k].end; clean_sort_range() 142 range[k].end = 0; clean_sort_range() 147 if (!range[i].end) { clean_sort_range() 27 add_range_with_merge(struct range *range, int az, int nr_range, u64 start, u64 end) add_range_with_merge() argument
|
H A D | resource.c | 32 .end = IO_SPACE_LIMIT, 40 .end = -1, 108 int width = root->end < 0x10000 ? 4 : 8; r_show() 117 width, (unsigned long long) r->end, r_show() 211 resource_size_t end = new->end; __request_resource() local 214 if (end < start) __request_resource() 218 if (end > root->end) __request_resource() 223 if (!tmp || tmp->start > end) { __request_resource() 230 if (tmp->end < start) __request_resource() 274 tmp->end = size - 1; __release_child_resources() 336 * Finds the lowest iomem reosurce exists with-in [res->start.res->end) 337 * the caller must specify res->start, res->end, res->flags and "name". 345 resource_size_t start, end; find_next_iomem_res() local 352 end = res->end; find_next_iomem_res() 353 BUG_ON(start >= end); find_next_iomem_res() 365 if (p->start > end) { find_next_iomem_res() 369 if ((p->end >= start) && (p->start < end)) find_next_iomem_res() 379 if (res->end > p->end) find_next_iomem_res() 380 res->end = p->end; find_next_iomem_res() 387 * All the memory ranges which overlap start,end and also match flags and 393 * @end: end addr 395 int walk_iomem_res(char *name, unsigned long flags, u64 start, u64 end, walk_iomem_res() argument 403 res.end = end; walk_iomem_res() 405 orig_end = res.end; walk_iomem_res() 406 while ((res.start < res.end) && walk_iomem_res() 408 ret = (*func)(res.start, res.end, arg); walk_iomem_res() 411 res.start = res.end + 1; walk_iomem_res() 412 res.end = orig_end; walk_iomem_res() 424 int walk_system_ram_res(u64 start, u64 end, void *arg, walk_system_ram_res() argument 432 res.end = end; walk_system_ram_res() 434 orig_end = res.end; walk_system_ram_res() 435 while ((res.start < res.end) && walk_system_ram_res() 437 ret = (*func)(res.start, res.end, arg); walk_system_ram_res() 440 res.start = res.end + 1; walk_system_ram_res() 441 res.end = orig_end; walk_system_ram_res() 462 res.end = ((u64)(start_pfn + nr_pages) << PAGE_SHIFT) - 1; walk_system_ram_range() 464 orig_end = res.end; walk_system_ram_range() 465 while ((res.start < res.end) && walk_system_ram_range() 468 end_pfn = (res.end + 1) >> PAGE_SHIFT; walk_system_ram_range() 473 res.start = res.end + 1; walk_system_ram_range() 474 res.end = orig_end; walk_system_ram_range() 516 resource_size_t end = start + size - 1; region_intersects() local 524 if (start >= p->start && start <= p->end) region_intersects() 526 if (end >= p->start && end <= p->end) region_intersects() 528 if (p->start >= start && p->end <= end) region_intersects() 559 if (res->end > max) resource_clip() 560 res->end = max; resource_clip() 578 * of this->start - 1 to tmp->end below would cause an underflow. __find_resource() 581 tmp.start = (this == old) ? old->start : this->end + 1; __find_resource() 586 tmp.end = (this == old) ? this->end : this->start - 1; __find_resource() 588 tmp.end = root->end; __find_resource() 590 if (tmp.end < tmp.start) __find_resource() 598 avail.end = tmp.end; __find_resource() 604 alloc.end = alloc.start + size - 1; __find_resource() 607 new->end = alloc.end; __find_resource() 612 next: if (!this || this->end == root->end) __find_resource() 616 tmp.start = this->end + 1; __find_resource() 657 old->end = new.end; reallocate_resource() 668 old->end = new.end; reallocate_resource() 769 if ((first->start > new->start) || (first->end < new->end)) __insert_resource() 771 if ((first->start == new->start) && (first->end == new->end)) __insert_resource() 777 if (next->start < new->start || next->end > new->end) __insert_resource() 781 if (next->sibling->start > new->end) __insert_resource() 868 if (conflict->end > new->end) insert_resource_expand_to_fit() 869 new->end = conflict->end; insert_resource_expand_to_fit() 880 resource_size_t end = start + size - 1; __adjust_resource() local 886 if ((start < parent->start) || (end > parent->end)) __adjust_resource() 889 if (res->sibling && (res->sibling->start <= end)) __adjust_resource() 896 if (start <= tmp->end) __adjust_resource() 902 if ((tmp->start < start) || (tmp->end > end)) __adjust_resource() 906 res->end = end; __adjust_resource() 936 resource_size_t start, resource_size_t end, __reserve_region_with_split() 949 res->end = end; __reserve_region_with_split() 965 conflict->end >= res->end) { __reserve_region_with_split() 973 end = res->end; __reserve_region_with_split() 974 res->end = conflict->start - 1; __reserve_region_with_split() 975 if (conflict->end < end) { __reserve_region_with_split() 982 next_res->start = conflict->end + 1; __reserve_region_with_split() 983 next_res->end = end; __reserve_region_with_split() 987 res->start = conflict->end + 1; __reserve_region_with_split() 994 resource_size_t start, resource_size_t end, reserve_region_with_split() 1000 if (root->start > start || root->end < end) { reserve_region_with_split() 1002 (unsigned long long)start, (unsigned long long)end, reserve_region_with_split() 1004 if (start > root->end || end < root->start) reserve_region_with_split() 1007 if (end > root->end) reserve_region_with_split() 1008 end = root->end; reserve_region_with_split() 1013 (unsigned long long)end); reserve_region_with_split() 1018 __reserve_region_with_split(root, start, end, name); reserve_region_with_split() 1073 res->end = start + n - 1; __request_region() 1122 resource_size_t end; __release_region() local 1125 end = start + n - 1; __release_region() 1134 if (res->start <= start && res->end >= end) { __release_region() 1139 if (res->start != start || res->end != end) __release_region() 1155 (unsigned long long)end); __release_region() 1186 resource_size_t end; release_mem_region_adjustable() local 1189 end = start + size - 1; release_mem_region_adjustable() 1190 if ((start < parent->start) || (end > parent->end)) release_mem_region_adjustable() 1200 if (res->start >= end) release_mem_region_adjustable() 1204 if (res->start > start || res->end < end) { release_mem_region_adjustable() 1218 if (res->start == start && res->end == end) { release_mem_region_adjustable() 1223 } else if (res->start == start && res->end != end) { release_mem_region_adjustable() 1225 ret = __adjust_resource(res, end + 1, release_mem_region_adjustable() 1226 res->end - end); release_mem_region_adjustable() 1227 } else if (res->start != start && res->end == end) { release_mem_region_adjustable() 1228 /* adjust the end */ release_mem_region_adjustable() 1238 new_res->start = end + 1; release_mem_region_adjustable() 1239 new_res->end = res->end; release_mem_region_adjustable() 1414 res->end = io_start + io_num - 1; reserve_setup() 1444 if (p->end < addr) iomem_map_sanity_check() 1447 PFN_DOWN(p->end) >= PFN_DOWN(addr + size - 1)) iomem_map_sanity_check() 1500 if (p->end < addr) iomem_is_exclusive() 935 __reserve_region_with_split(struct resource *root, resource_size_t start, resource_size_t end, const char *name) __reserve_region_with_split() argument 993 reserve_region_with_split(struct resource *root, resource_size_t start, resource_size_t end, const char *name) reserve_region_with_split() argument
|
H A D | crash_dump.c | 36 char *end; setup_elfcorehdr() local 39 elfcorehdr_addr = memparse(arg, &end); setup_elfcorehdr() 40 if (*end == '@') { setup_elfcorehdr() 42 elfcorehdr_addr = memparse(end + 1, &end); setup_elfcorehdr() 44 return end > arg ? 0 : -EINVAL; setup_elfcorehdr()
|
/linux-4.4.14/include/linux/ceph/ |
H A D | decode.h | 14 * void *end pointer to end of buffer (last byte + 1) 50 static inline int ceph_has_room(void **p, void *end, size_t n) ceph_has_room() argument 52 return end >= *p && n <= end - *p; ceph_has_room() 55 #define ceph_decode_need(p, end, n, bad) \ 57 if (!likely(ceph_has_room(p, end, n))) \ 61 #define ceph_decode_64_safe(p, end, v, bad) \ 63 ceph_decode_need(p, end, sizeof(u64), bad); \ 66 #define ceph_decode_32_safe(p, end, v, bad) \ 68 ceph_decode_need(p, end, sizeof(u32), bad); \ 71 #define ceph_decode_16_safe(p, end, v, bad) \ 73 ceph_decode_need(p, end, sizeof(u16), bad); \ 76 #define ceph_decode_8_safe(p, end, v, bad) \ 78 ceph_decode_need(p, end, sizeof(u8), bad); \ 82 #define ceph_decode_copy_safe(p, end, pv, n, bad) \ 84 ceph_decode_need(p, end, n, bad); \ 102 * beyond the "end" pointer provided (-ERANGE) 105 static inline char *ceph_extract_encoded_string(void **p, void *end, ceph_extract_encoded_string() argument 112 ceph_decode_32_safe(&sp, end, len, bad); ceph_extract_encoded_string() 113 if (!ceph_has_room(&sp, end, len)) ceph_extract_encoded_string() 197 static inline void ceph_encode_filepath(void **p, void *end, ceph_encode_filepath() argument 201 BUG_ON(*p + 1 + sizeof(ino) + sizeof(len) + len > end); ceph_encode_filepath() 210 static inline void ceph_encode_string(void **p, void *end, ceph_encode_string() argument 213 BUG_ON(*p + sizeof(len) + len > end); ceph_encode_string() 220 #define ceph_encode_need(p, end, n, bad) \ 222 if (!likely(ceph_has_room(p, end, n))) \ 226 #define ceph_encode_64_safe(p, end, v, bad) \ 228 ceph_encode_need(p, end, sizeof(u64), bad); \ 231 #define ceph_encode_32_safe(p, end, v, bad) \ 233 ceph_encode_need(p, end, sizeof(u32), bad); \ 236 #define ceph_encode_16_safe(p, end, v, bad) \ 238 ceph_encode_need(p, end, sizeof(u16), bad); \ 241 #define ceph_encode_8_safe(p, end, v, bad) \ 243 ceph_encode_need(p, end, sizeof(u8), bad); \ 247 #define ceph_encode_copy_safe(p, end, pv, n, bad) \ 249 ceph_encode_need(p, end, n, bad); \ 252 #define ceph_encode_string_safe(p, end, s, n, bad) \ 254 ceph_encode_need(p, end, n, bad); \ 255 ceph_encode_string(p, end, s, n); \
|
/linux-4.4.14/include/linux/ |
H A D | circ_buf.h | 22 /* Return count up to the end of the buffer. Carefully avoid 26 ({int end = (size) - (tail); \ 27 int n = ((head) + end) & ((size)-1); \ 28 n < end ? n : end;}) 30 /* Return space available up to the end of the buffer. */ 32 ({int end = (size) - 1 - (head); \ 33 int n = (end + (tail)) & ((size)-1); \ 34 n <= end ? n : end+1;})
|
H A D | range.h | 6 u64 end; member in struct:range 10 u64 start, u64 end); 14 u64 start, u64 end); 16 void subtract_range(struct range *range, int az, u64 start, u64 end);
|
H A D | firmware-map.h | 26 int firmware_map_add_early(u64 start, u64 end, const char *type); 27 int firmware_map_add_hotplug(u64 start, u64 end, const char *type); 28 int firmware_map_remove(u64 start, u64 end, const char *type); 32 static inline int firmware_map_add_early(u64 start, u64 end, const char *type) firmware_map_add_early() argument 37 static inline int firmware_map_add_hotplug(u64 start, u64 end, const char *type) firmware_map_add_hotplug() argument 42 static inline int firmware_map_remove(u64 start, u64 end, const char *type) firmware_map_remove() argument
|
/linux-4.4.14/arch/x86/kernel/ |
H A D | resource.c | 5 resource_size_t end) resource_clip() 9 if (res->end < start || res->start > end) resource_clip() 15 if (res->end > end) resource_clip() 16 high = res->end - end; resource_clip() 20 res->end = start - 1; resource_clip() 22 res->start = end + 1; resource_clip() 4 resource_clip(struct resource *res, resource_size_t start, resource_size_t end) resource_clip() argument
|
H A D | check.c | 58 char *end; set_corruption_check_size() local 61 size = memparse(arg, &end); set_corruption_check_size() 63 if (*end == '\0') set_corruption_check_size() 73 phys_addr_t start, end; setup_bios_corruption_check() local 94 for_each_free_mem_range(i, NUMA_NO_NODE, MEMBLOCK_NONE, &start, &end, setup_bios_corruption_check() 98 end = clamp_t(phys_addr_t, round_down(end, PAGE_SIZE), setup_bios_corruption_check() 100 if (start >= end) setup_bios_corruption_check() 103 memblock_reserve(start, end - start); setup_bios_corruption_check() 105 scan_areas[num_scan_areas].size = end - start; setup_bios_corruption_check() 108 memset(__va(start), 0, end - start); setup_bios_corruption_check()
|
H A D | probe_roms.c | 27 .end = 0xfffff, 34 .end = 0xeffff, 41 .end = 0, 46 .end = 0, 51 .end = 0, 56 .end = 0, 61 .end = 0, 66 .end = 0, 73 .end = 0xc7fff, 123 if (res->end == 0) find_oprom() 221 video_rom_resource.end = start + length - 1; probe_roms() 227 start = (video_rom_resource.end + 1 + 2047) & ~2047UL; probe_roms() 262 adapter_rom_resources[i].end = start + length - 1; probe_roms() 265 start = adapter_rom_resources[i++].end & ~2047UL; probe_roms()
|
/linux-4.4.14/arch/blackfin/include/asm/ |
H A D | cacheflush.h | 28 #define flush_cache_range(vma, start, end) do { } while (0) 30 #define flush_cache_vmap(start, end) do { } while (0) 31 #define flush_cache_vunmap(start, end) do { } while (0) 34 #define flush_icache_range_others(start, end) \ 35 smp_icache_flush_range_others((start), (end)) 37 #define flush_icache_range_others(start, end) do { } while (0) 40 static inline void flush_icache_range(unsigned start, unsigned end) flush_icache_range() argument 43 if (end <= physical_mem_end) flush_icache_range() 44 blackfin_dcache_flush_range(start, end); flush_icache_range() 47 if (start >= L2_START && end <= L2_START + L2_LENGTH) flush_icache_range() 48 blackfin_dcache_flush_range(start, end); flush_icache_range() 61 if (end <= physical_mem_end) { flush_icache_range() 62 blackfin_icache_flush_range(start, end); flush_icache_range() 63 flush_icache_range_others(start, end); flush_icache_range() 67 if (start >= L2_START && end <= L2_START + L2_LENGTH) { flush_icache_range() 68 blackfin_icache_flush_range(start, end); flush_icache_range() 69 flush_icache_range_others(start, end); flush_icache_range() 82 # define invalidate_dcache_range(start,end) blackfin_dcache_invalidate_range((start), (end)) 84 # define invalidate_dcache_range(start,end) do { } while (0) 87 # define flush_dcache_range(start,end) blackfin_dcache_flush_range((start), (end)) 91 # define flush_dcache_range(start,end) do { } while (0)
|
/linux-4.4.14/drivers/pci/ |
H A D | hotplug-pci.c | 11 int end = parent->busn_res.end; pci_hp_add_bridge() local 13 for (busnr = start; busnr <= end; busnr++) { pci_hp_add_bridge() 17 if (busnr-- > end) { pci_hp_add_bridge()
|
H A D | host-bridge.c | 63 region->end = res->end - offset; pcibios_resource_to_bus() 70 return region1->start <= region2->start && region1->end >= region2->end; region_contains() 87 bus_region.end = window->res->end - window->offset; pcibios_bus_to_resource() 96 res->end = region->end + offset; pcibios_bus_to_resource()
|
/linux-4.4.14/arch/x86/include/asm/ |
H A D | pat.h | 11 extern int reserve_memtype(u64 start, u64 end, 13 extern int free_memtype(u64 start, u64 end); 18 int io_reserve_memtype(resource_size_t start, resource_size_t end, 21 void io_free_memtype(resource_size_t start, resource_size_t end);
|
/linux-4.4.14/arch/powerpc/boot/ |
H A D | stdlib.h | 4 unsigned long long int strtoull(const char *ptr, char **end, int base);
|
H A D | stdlib.c | 16 unsigned long long int strtoull(const char *ptr, char **end, int base) strtoull() argument 41 if (end) strtoull() 42 *end = (char *)ptr; strtoull()
|
/linux-4.4.14/arch/sparc/include/asm/ |
H A D | tlbflush_32.h | 10 #define flush_tlb_range(vma, start, end) \ 11 sparc32_cachetlb_ops->tlb_range(vma, start, end) 19 unsigned long end) flush_tlb_kernel_range() 18 flush_tlb_kernel_range(unsigned long start, unsigned long end) flush_tlb_kernel_range() argument
|
H A D | cacheflush_32.h | 12 #define flush_cache_range(vma,start,end) \ 13 sparc32_cachetlb_ops->cache_range(vma, start, end) 16 #define flush_icache_range(start, end) do { } while (0) 46 #define flush_cache_vmap(start, end) flush_cache_all() 47 #define flush_cache_vunmap(start, end) flush_cache_all()
|
H A D | tlbflush_64.h | 18 void flush_tsb_kernel_range(unsigned long start, unsigned long end); 34 unsigned long start, unsigned long end) flush_tlb_range() 38 void flush_tlb_kernel_range(unsigned long start, unsigned long end); 50 void __flush_tlb_kernel_range(unsigned long start, unsigned long end); 61 void smp_flush_tlb_kernel_range(unsigned long start, unsigned long end); 33 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_tlb_range() argument
|
/linux-4.4.14/tools/perf/ |
H A D | perf-read-vdso.c | 14 void *start, *end; main() local 17 if (find_vdso_map(&start, &end)) main() 20 size = end - start; main()
|
/linux-4.4.14/arch/s390/lib/ |
H A D | delay.c | 23 * To end the bloody studid and useless discussion about the __delay() 37 u64 end; __udelay_disabled() local 39 end = get_tod_clock() + (usecs << 12); __udelay_disabled() 45 set_clock_comparator(end); __udelay_disabled() 55 u64 clock_saved, end; __udelay_enabled() local 57 end = get_tod_clock_fast() + (usecs << 12); __udelay_enabled() 60 if (end < S390_lowcore.clock_comparator) { __udelay_enabled() 62 set_clock_comparator(end); __udelay_enabled() 67 } while (get_tod_clock_fast() < end); __udelay_enabled() 109 u64 end; udelay_simple() local 111 end = get_tod_clock_fast() + (usecs << 12); udelay_simple() 112 while (get_tod_clock_fast() < end) udelay_simple() 118 u64 end; __ndelay() local 122 end = get_tod_clock_fast() + nsecs; __ndelay() 125 while (get_tod_clock_fast() < end) __ndelay()
|
/linux-4.4.14/arch/alpha/boot/ |
H A D | head.S | 17 .end __start 26 .end wrent 35 .end wrkgp 85 .end switch_to_osf_pal 94 .end tbi 102 .end halt 123 .end move_stack
|
/linux-4.4.14/arch/avr32/mm/ |
H A D | cache.c | 26 unsigned long v, begin, end, linesz, mask; invalidate_dcache_region() local 35 end = begin + size; invalidate_dcache_region() 41 if (end & mask) { invalidate_dcache_region() 42 flush_dcache_line((void *)end); invalidate_dcache_region() 43 end &= ~mask; invalidate_dcache_region() 47 for (v = begin; v < end; v += linesz) invalidate_dcache_region() 54 unsigned long v, begin, end, linesz; clean_dcache_region() local 58 end = ((unsigned long)start + size + linesz - 1) & ~(linesz - 1); clean_dcache_region() 60 for (v = begin; v < end; v += linesz) clean_dcache_region() 67 unsigned long v, begin, end, linesz; flush_dcache_region() local 71 end = ((unsigned long)start + size + linesz - 1) & ~(linesz - 1); flush_dcache_region() 73 for (v = begin; v < end; v += linesz) flush_dcache_region() 80 unsigned long v, begin, end, linesz; invalidate_icache_region() local 84 end = ((unsigned long)start + size + linesz - 1) & ~(linesz - 1); invalidate_icache_region() 86 for (v = begin; v < end; v += linesz) invalidate_icache_region() 90 static inline void __flush_icache_range(unsigned long start, unsigned long end) __flush_icache_range() argument 95 for (v = start; v < end; v += linesz) { __flush_icache_range() 106 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument 112 (end + linesz - 1) & ~(linesz - 1)); flush_icache_range()
|
/linux-4.4.14/arch/mips/pci/ |
H A D | pci-malta.c | 43 .end = 0x000fffffUL, 93 resource_size_t start, end, map, start1, end1, map1, map2, map3, mask; mips_pcibios_init() local 116 end = GT_READ(GT_PCI0M0HD_OFS); mips_pcibios_init() 118 end = (end & GT_PCI_HD_MSK) | (start & ~GT_PCI_HD_MSK); mips_pcibios_init() 124 if (end1 - start1 > end - start) { mips_pcibios_init() 126 end = end1; mips_pcibios_init() 129 mask = ~(start ^ end); mips_pcibios_init() 134 gt64120_mem_resource.end = end; mips_pcibios_init() 138 gt64120_mem_resource.end <<= GT_PCI_DCRM_SHF; mips_pcibios_init() 139 gt64120_mem_resource.end |= (1 << GT_PCI_DCRM_SHF) - 1; mips_pcibios_init() 143 end = GT_READ(GT_PCI0IOHD_OFS); mips_pcibios_init() 145 end = (end & GT_PCI_HD_MSK) | (start & ~GT_PCI_HD_MSK); mips_pcibios_init() 146 mask = ~(start ^ end); mips_pcibios_init() 151 gt64120_io_resource.end = (map & mask) | ~mask; mips_pcibios_init() 155 gt64120_io_resource.end <<= GT_PCI_DCRM_SHF; mips_pcibios_init() 156 gt64120_io_resource.end |= (1 << GT_PCI_DCRM_SHF) - 1; mips_pcibios_init() 173 end = 1; mips_pcibios_init() 177 end++; mips_pcibios_init() 182 end++; mips_pcibios_init() 185 bonito64_mem_resource.end = start + mips_pcibios_init() 186 BONITO_PCIMAP_WINBASE(end) - 1; mips_pcibios_init() 202 msc_mem_resource.end = (start & mask) | ~mask; mips_pcibios_init() 213 msc_io_resource.end = (map & mask) | ~mask; mips_pcibios_init() 215 ioport_resource.end = ~mask; mips_pcibios_init() 223 end = start | ~mask; mips_pcibios_init() 225 start <= msc_mem_resource.end) || mips_pcibios_init() 226 (end >= msc_mem_resource.start && mips_pcibios_init() 227 end <= msc_mem_resource.end)) { mips_pcibios_init() 230 end = min(end, msc_mem_resource.end); mips_pcibios_init() 232 msc_mem_resource.end - end) mips_pcibios_init() 233 msc_mem_resource.end = start - 1; mips_pcibios_init() 235 msc_mem_resource.start = end + 1; mips_pcibios_init() 248 iomem_resource.end &= 0xfffffffffULL; /* 64 GB */ mips_pcibios_init() 249 ioport_resource.end = controller->io_resource->end; mips_pcibios_init()
|
/linux-4.4.14/arch/sh/include/asm/ |
H A D | tlbflush.h | 10 * - flush_tlb_range(vma, start, end) flushes a range of pages 11 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages 17 unsigned long end); 21 unsigned long end); 31 unsigned long end); 33 extern void flush_tlb_kernel_range(unsigned long start, unsigned long end); 43 #define flush_tlb_range(vma, start, end) \ 44 local_flush_tlb_range(vma, start, end) 46 #define flush_tlb_kernel_range(start, end) \ 47 local_flush_tlb_kernel_range(start, end)
|
H A D | tlb.h | 24 unsigned long start, end; member in struct:mmu_gather 30 tlb->end = 0; init_tlb_gather() 34 tlb->end = TASK_SIZE; init_tlb_gather() 39 tlb_gather_mmu(struct mmu_gather *tlb, struct mm_struct *mm, unsigned long start, unsigned long end) tlb_gather_mmu() argument 43 tlb->end = end; tlb_gather_mmu() 44 tlb->fullmm = !(start | (end+1)); tlb_gather_mmu() 50 tlb_finish_mmu(struct mmu_gather *tlb, unsigned long start, unsigned long end) tlb_finish_mmu() argument 64 if (tlb->end < address + PAGE_SIZE) tlb_remove_tlb_entry() 65 tlb->end = address + PAGE_SIZE; tlb_remove_tlb_entry() 83 if (!tlb->fullmm && tlb->end) { tlb_end_vma() 84 flush_tlb_range(vma, tlb->start, tlb->end); tlb_end_vma()
|
/linux-4.4.14/fs/btrfs/tests/ |
H A D | extent-io-tests.c | 28 static noinline int process_page_range(struct inode *inode, u64 start, u64 end, process_page_range() argument 34 unsigned long end_index = end >> PAGE_CACHE_SHIFT; process_page_range() 59 printk(KERN_ERR "stuck in a loop, start %Lu, end %Lu, nr_pages %lu, ret %d\n", start, end, nr_pages, ret); process_page_range() 75 u64 start, end, test_start; test_find_delalloc() local 114 end = 0; test_find_delalloc() 116 &end, max_bytes); test_find_delalloc() 121 if (start != 0 || end != 4095) { test_find_delalloc() 122 test_msg("Expected start 0 end 4095, got start %Lu end %Lu\n", test_find_delalloc() 123 start, end); test_find_delalloc() 126 unlock_extent(&tmp, start, end); test_find_delalloc() 145 end = 0; test_find_delalloc() 147 &end, max_bytes); test_find_delalloc() 152 if (start != test_start || end != max_bytes - 1) { test_find_delalloc() 153 test_msg("Expected start %Lu end %Lu, got start %Lu, end " test_find_delalloc() 154 "%Lu\n", test_start, max_bytes - 1, start, end); test_find_delalloc() 157 if (process_page_range(inode, start, end, test_find_delalloc() 162 unlock_extent(&tmp, start, end); test_find_delalloc() 179 end = 0; test_find_delalloc() 181 &end, max_bytes); test_find_delalloc() 186 if (end != (u64)-1) { test_find_delalloc() 187 test_msg("Did not return the proper end offset\n"); test_find_delalloc() 200 end = 0; test_find_delalloc() 202 &end, max_bytes); test_find_delalloc() 207 if (start != test_start || end != total_dirty - 1) { test_find_delalloc() 208 test_msg("Expected start %Lu end %Lu, got start %Lu end %Lu\n", test_find_delalloc() 209 test_start, total_dirty - 1, start, end); test_find_delalloc() 212 if (process_page_range(inode, start, end, test_find_delalloc() 217 unlock_extent(&tmp, start, end); test_find_delalloc() 235 end = 0; test_find_delalloc() 243 &end, max_bytes); test_find_delalloc() 248 if (start != test_start && end != test_start + PAGE_CACHE_SIZE - 1) { test_find_delalloc() 249 test_msg("Expected start %Lu end %Lu, got start %Lu end %Lu\n", test_find_delalloc() 251 end); test_find_delalloc() 254 if (process_page_range(inode, start, end, PROCESS_TEST_LOCKED | test_find_delalloc()
|
/linux-4.4.14/arch/microblaze/kernel/cpu/ |
H A D | cache.c | 92 #define CACHE_LOOP_LIMITS(start, end, cache_line_length, cache_size) \ 95 end = min(start + cache_size, end); \ 120 * end address is not aligned, if end is aligned then I have to subtract 124 #define CACHE_RANGE_LOOP_2(start, end, line_length, op) \ 129 end = ((end & align) == end) ? end - line_length : end & align; \ 130 count = end - start; \ 141 #define CACHE_RANGE_LOOP_1(start, end, line_length, op) \ 145 end = ((end & align) == end) ? end - line_length : end & align; \ 146 WARN_ON(end < start); \ 152 : : "r" (temp), "r" (start), "r" (end), \ 158 static void __flush_icache_range_msr_irq(unsigned long start, unsigned long end) __flush_icache_range_msr_irq() argument 164 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __flush_icache_range_msr_irq() 165 (unsigned int)start, (unsigned int) end); __flush_icache_range_msr_irq() 167 CACHE_LOOP_LIMITS(start, end, __flush_icache_range_msr_irq() 174 CACHE_RANGE_LOOP_1(start, end, cpuinfo.icache_line_length, wic); __flush_icache_range_msr_irq() 176 for (i = start; i < end; i += cpuinfo.icache_line_length) __flush_icache_range_msr_irq() 185 unsigned long end) __flush_icache_range_nomsr_irq() 191 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __flush_icache_range_nomsr_irq() 192 (unsigned int)start, (unsigned int) end); __flush_icache_range_nomsr_irq() 194 CACHE_LOOP_LIMITS(start, end, __flush_icache_range_nomsr_irq() 201 CACHE_RANGE_LOOP_1(start, end, cpuinfo.icache_line_length, wic); __flush_icache_range_nomsr_irq() 203 for (i = start; i < end; i += cpuinfo.icache_line_length) __flush_icache_range_nomsr_irq() 213 unsigned long end) __flush_icache_range_noirq() 218 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __flush_icache_range_noirq() 219 (unsigned int)start, (unsigned int) end); __flush_icache_range_noirq() 221 CACHE_LOOP_LIMITS(start, end, __flush_icache_range_noirq() 224 CACHE_RANGE_LOOP_1(start, end, cpuinfo.icache_line_length, wic); __flush_icache_range_noirq() 226 for (i = start; i < end; i += cpuinfo.icache_line_length) __flush_icache_range_noirq() 377 unsigned long end) __invalidate_dcache_range_wb() 382 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __invalidate_dcache_range_wb() 383 (unsigned int)start, (unsigned int) end); __invalidate_dcache_range_wb() 385 CACHE_LOOP_LIMITS(start, end, __invalidate_dcache_range_wb() 388 CACHE_RANGE_LOOP_2(start, end, cpuinfo.dcache_line_length, wdc.clear); __invalidate_dcache_range_wb() 390 for (i = start; i < end; i += cpuinfo.dcache_line_length) __invalidate_dcache_range_wb() 397 unsigned long end) __invalidate_dcache_range_nomsr_wt() 402 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __invalidate_dcache_range_nomsr_wt() 403 (unsigned int)start, (unsigned int) end); __invalidate_dcache_range_nomsr_wt() 404 CACHE_LOOP_LIMITS(start, end, __invalidate_dcache_range_nomsr_wt() 408 CACHE_RANGE_LOOP_1(start, end, cpuinfo.dcache_line_length, wdc); __invalidate_dcache_range_nomsr_wt() 410 for (i = start; i < end; i += cpuinfo.dcache_line_length) __invalidate_dcache_range_nomsr_wt() 417 unsigned long end) __invalidate_dcache_range_msr_irq_wt() 423 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __invalidate_dcache_range_msr_irq_wt() 424 (unsigned int)start, (unsigned int) end); __invalidate_dcache_range_msr_irq_wt() 425 CACHE_LOOP_LIMITS(start, end, __invalidate_dcache_range_msr_irq_wt() 432 CACHE_RANGE_LOOP_1(start, end, cpuinfo.dcache_line_length, wdc); __invalidate_dcache_range_msr_irq_wt() 434 for (i = start; i < end; i += cpuinfo.dcache_line_length) __invalidate_dcache_range_msr_irq_wt() 444 unsigned long end) __invalidate_dcache_range_nomsr_irq() 450 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __invalidate_dcache_range_nomsr_irq() 451 (unsigned int)start, (unsigned int) end); __invalidate_dcache_range_nomsr_irq() 453 CACHE_LOOP_LIMITS(start, end, __invalidate_dcache_range_nomsr_irq() 460 CACHE_RANGE_LOOP_1(start, end, cpuinfo.dcache_line_length, wdc); __invalidate_dcache_range_nomsr_irq() 462 for (i = start; i < end; i += cpuinfo.dcache_line_length) __invalidate_dcache_range_nomsr_irq() 488 static void __flush_dcache_range_wb(unsigned long start, unsigned long end) __flush_dcache_range_wb() argument 493 pr_debug("%s: start 0x%x, end 0x%x\n", __func__, __flush_dcache_range_wb() 494 (unsigned int)start, (unsigned int) end); __flush_dcache_range_wb() 496 CACHE_LOOP_LIMITS(start, end, __flush_dcache_range_wb() 499 CACHE_RANGE_LOOP_2(start, end, cpuinfo.dcache_line_length, wdc.flush); __flush_dcache_range_wb() 501 for (i = start; i < end; i += cpuinfo.dcache_line_length) __flush_dcache_range_wb() 184 __flush_icache_range_nomsr_irq(unsigned long start, unsigned long end) __flush_icache_range_nomsr_irq() argument 212 __flush_icache_range_noirq(unsigned long start, unsigned long end) __flush_icache_range_noirq() argument 376 __invalidate_dcache_range_wb(unsigned long start, unsigned long end) __invalidate_dcache_range_wb() argument 396 __invalidate_dcache_range_nomsr_wt(unsigned long start, unsigned long end) __invalidate_dcache_range_nomsr_wt() argument 416 __invalidate_dcache_range_msr_irq_wt(unsigned long start, unsigned long end) __invalidate_dcache_range_msr_irq_wt() argument 443 __invalidate_dcache_range_nomsr_irq(unsigned long start, unsigned long end) __invalidate_dcache_range_nomsr_irq() argument
|
/linux-4.4.14/arch/powerpc/include/asm/ |
H A D | sections.h | 26 unsigned long end) overlaps_interrupt_vector_text() 33 (unsigned long)__va(real_start) < end; overlaps_interrupt_vector_text() 36 static inline int overlaps_kernel_text(unsigned long start, unsigned long end) overlaps_kernel_text() argument 39 (unsigned long)_stext < end; overlaps_kernel_text() 42 static inline int overlaps_kvm_tmp(unsigned long start, unsigned long end) overlaps_kvm_tmp() argument 47 (unsigned long)&kvm_tmp[1024 * 1024] < end; overlaps_kvm_tmp() 25 overlaps_interrupt_vector_text(unsigned long start, unsigned long end) overlaps_interrupt_vector_text() argument
|
/linux-4.4.14/arch/mips/jz4740/ |
H A D | platform.c | 39 .end = JZ4740_UHC_BASE_ADDR + 0x1000 - 1, 44 .end = JZ4740_IRQ_UHC, 69 .end = JZ4740_UDC_BASE_ADDR + 0x10000 - 1, 74 .end = JZ4740_IRQ_UDC, 95 .end = JZ4740_MSC_BASE_ADDR + 0x1000 - 1, 100 .end = JZ4740_IRQ_MSC, 120 .end = JZ4740_RTC_BASE_ADDR + 0x38 - 1, 125 .end = JZ4740_IRQ_RTC, 141 .end = JZ4740_I2C_BASE_ADDR + 0x1000 - 1, 146 .end = JZ4740_IRQ_I2C, 163 .end = JZ4740_EMC_BASE_ADDR + 0x1000 - 1, 169 .end = 0x180C0000 - 1, 175 .end = 0x140C0000 - 1, 181 .end = 0x0C0C0000 - 1, 187 .end = 0x080C0000 - 1, 202 .end = JZ4740_LCD_BASE_ADDR + 0x1000 - 1, 222 .end = JZ4740_AIC_BASE_ADDR + 0x38 - 1, 244 .end = JZ4740_AIC_BASE_ADDR + 0x88 - 1, 260 .end = JZ4740_SADC_BASE_ADDR + 0x30, 265 .end = JZ4740_IRQ_SADC, 270 .end = JZ4740_IRQ_ADC_BASE, 286 .end = JZ4740_WDT_BASE_ADDR + 0x10 - 1, 308 .end = JZ4740_DMAC_BASE_ADDR + 0x400 - 1, 313 .end = JZ4740_IRQ_DMAC,
|
/linux-4.4.14/arch/arm/mach-pxa/ |
H A D | devices.c | 39 .end = IRQ_PMU, 53 .end = 0x41100fff, 58 .end = IRQ_MMC, 63 .end = 21, 68 .end = 22, 104 .end = 0x4060ffff, 109 .end = IRQ_USB, 142 .end = 0x54100fff, 147 .end = IRQ_USB2, 168 .end = 0x4400ffff, 173 .end = IRQ_LCD, 200 .end = 0x40100023, 204 .end = IRQ_FFUART, 224 .end = 0x40200023, 228 .end = IRQ_BTUART, 248 .end = 0x40700023, 252 .end = IRQ_STUART, 272 .end = 0x4160002F, 276 .end = IRQ_HWUART, 299 .end = 0x403016a3, 303 .end = IRQ_I2C, 324 .end = 0x40f001a3, 328 .end = IRQ_PWRI2C, 344 .end = 0x40400083, 348 .end = IRQ_I2S, 390 .end = IRQ_STUART, 395 .end = IRQ_ICP, 400 .end = 0x4080001b, 405 .end = 0x40700023, 410 .end = 17, 415 .end = 18, 439 .end = 0x40900000 + 0x3b, 444 .end = IRQ_RTC1Hz, 450 .end = IRQ_RTCAlrm, 473 .end = 0x40500000 + 0xfff, 478 .end = IRQ_AC97, 506 .end = 0x40b0000f, 521 .end = 0x40c0000f, 538 .end = 0x4100001f, 543 .end = IRQ_SSP, 549 .end = 13, 555 .end = 14, 576 .end = 0x4140002f, 581 .end = IRQ_NSSP, 587 .end = 15, 593 .end = 16, 614 .end = 0x4150002f, 619 .end = IRQ_ASSP, 625 .end = 23, 631 .end = 24, 653 .end = 0x50000fff, 658 .end = IRQ_CAMERA, 686 .end = 0x4C00ff6f, 691 .end = IRQ_USBH1, 717 .end = 0x4150004c, 722 .end = IRQ_KEYPAD, 744 .end = 0x4100003f, 749 .end = IRQ_SSP, 755 .end = 13, 761 .end = 14, 782 .end = 0x4170003f, 787 .end = IRQ_SSP2, 793 .end = 15, 799 .end = 16, 820 .end = 0x4190003f, 825 .end = IRQ_SSP3, 831 .end = 66, 837 .end = 67, 856 .end = 0x40b0001f, 871 .end = 0x40c0001f, 888 .end = 0x42000fff, 893 .end = IRQ_MMC2, 898 .end = 93, 903 .end = 94, 927 .end = 0x42500fff, 932 .end = IRQ_MMC3, 937 .end = 100, 942 .end = 101, 966 .end = 0x54000fff, 971 .end = IRQ_GCU, 995 .end = 0x40f500d3, 999 .end = IRQ_PWRI2C, 1014 .end = 0x43100053, 1019 .end = IRQ_NAND, 1025 .end = 97, 1031 .end = 99, 1059 .end = 0x41a0003f, 1064 .end = IRQ_SSP4, 1070 .end = 2, 1076 .end = 3, 1135 .end = 0x40e0ffff, 1139 .end = IRQ_GPIO0, 1144 .end = IRQ_GPIO1, 1149 .end = IRQ_GPIO_2_x, 1211 .end = 0x4000ffff, 1216 .end = IRQ_DMA,
|
/linux-4.4.14/arch/x86/pci/ |
H A D | amd_bus.c | 71 u64 end; early_root_info_init() local 159 end = (reg & 0xfff000) | 0xfff; early_root_info_init() 166 node, link, start, end); early_root_info_init() 169 if (end > 0xffff) early_root_info_init() 170 end = 0xffff; early_root_info_init() 171 update_res(info, start, end, IORESOURCE_IO, 1); early_root_info_init() 172 subtract_range(range, RANGE_NUM, start, end + 1); early_root_info_init() 179 if (!range[i].end) early_root_info_init() 182 update_res(info, range[i].start, range[i].end - 1, early_root_info_init() 189 end = cap_resource((0xfdULL<<32) - 1); early_root_info_init() 190 end++; early_root_info_init() 191 add_range(range, RANGE_NUM, 0, 0, end); early_root_info_init() 196 end = (val & 0xffffff800000ULL); early_root_info_init() 197 printk(KERN_INFO "TOM: %016llx aka %lldM\n", end, end>>20); early_root_info_init() 198 if (end < (1ULL<<32)) early_root_info_init() 199 subtract_range(range, RANGE_NUM, 0, end); early_root_info_init() 207 fam10h_mmconf_end = fam10h_mmconf->end; early_root_info_init() 226 end = (reg & 0xffffff00); early_root_info_init() 227 end <<= 8; early_root_info_init() 228 end |= 0xffff; early_root_info_init() 236 node, link, start, end); early_root_info_init() 239 * mmconf range, so need to update start and end. early_root_info_init() 250 if (end >= fam10h_mmconf_start && early_root_info_init() 251 end <= fam10h_mmconf_end) { early_root_info_init() 252 end = fam10h_mmconf_start - 1; early_root_info_init() 257 end > fam10h_mmconf_end) { early_root_info_init() 268 if (start <= end) { early_root_info_init() 269 printk(KERN_CONT " %s [%llx, %llx]", endx ? "and" : "==>", start, end); early_root_info_init() 277 update_res(info, cap_resource(start), cap_resource(end), early_root_info_init() 279 subtract_range(range, RANGE_NUM, start, end + 1); early_root_info_init() 292 end = (val & 0xffffff800000ULL); early_root_info_init() 293 printk(KERN_INFO "TOM2: %016llx aka %lldM\n", end, end>>20); early_root_info_init() 294 subtract_range(range, RANGE_NUM, 1ULL<<32, end); early_root_info_init() 304 if (!range[i].end) early_root_info_init() 308 cap_resource(range[i].end - 1), early_root_info_init()
|
H A D | bus_numa.c | 84 info->busn.end = bus_max; alloc_pci_root_info() 95 resource_size_t end, unsigned long flags, int merge) update_res() 100 if (start > end) update_res() 119 common_end = min(res->end, end); update_res() 124 final_end = max(res->end, end); update_res() 127 res->end = final_end; update_res() 142 res->end = end; update_res() 94 update_res(struct pci_root_info *info, resource_size_t start, resource_size_t end, unsigned long flags, int merge) update_res() argument
|
/linux-4.4.14/arch/blackfin/mach-bf561/boards/ |
H A D | tepla.c | 25 .end = 0x2C000320, 29 .end = IRQ_PROG_INTB, 33 .end = IRQ_PF7, 50 .end = BFIN_UART_GCTL+2, 55 .end = IRQ_UART_TX, 60 .end = IRQ_UART_RX, 65 .end = IRQ_UART_ERROR, 70 .end = CH_UART_TX, 75 .end = CH_UART_RX, 101 .end = 0xFFC004FF, 106 .end = IRQ_UART0_RX+1, 111 .end = CH_UART0_RX+1,
|
H A D | cm_bf561.c | 104 .end = SPI0_REGBASE + 0xFF, 109 .end = CH_SPI, 114 .end = IRQ_SPI, 158 .end = 0x28000300 + 16, 162 .end = IRQ_PF0, 184 .end = 0x24008000 + 0xFF, 189 .end = IRQ_PF43, 216 .end = 0x24000000 + 0x100, 220 .end = IRQ_PF45, 237 .end = 0x24008000, 241 .end = 0x24008004, 245 .end = IRQ_PF47, 277 .end = BFIN_UART_GCTL+2, 282 .end = IRQ_UART_TX, 287 .end = IRQ_UART_RX, 292 .end = IRQ_UART_ERROR, 297 .end = CH_UART_TX, 302 .end = CH_UART_RX, 328 .end = 0xFFC004FF, 333 .end = IRQ_UART0_RX+1, 338 .end = CH_UART0_RX+1, 362 .end = 0x2400C001F, 367 .end = 0x2400D01B, 372 .end = PATA_INT, 413 .end = 0x207fffff,
|
H A D | acvilon.c | 69 .end = 0x20000000 + 0x000fffff, 74 .end = IRQ_PF15, 104 .end = 0x2C000000 + 16, 109 .end = IRQ_PF8, 149 .end = 0x5ffffff, 170 .end = 0x28000000 + 0xFF, 175 .end = IRQ_PF7, 203 .end = BFIN_UART_GCTL + 2, 208 .end = IRQ_UART_TX, 213 .end = IRQ_UART_RX, 218 .end = IRQ_UART_ERROR, 223 .end = CH_UART_TX, 228 .end = CH_UART_RX, 303 .end = 0x24000000 + (1 << MAX(BFIN_NAND_PLAT_CLE, BFIN_NAND_PLAT_ALE)), 378 .end = SPI0_REGBASE + 0xFF, 383 .end = CH_SPI, 388 .end = IRQ_SPI, 435 /* .end = MAX_BLACKFIN_GPIOS - 1, */ 436 .end = 32,
|
/linux-4.4.14/arch/c6x/include/asm/ |
H A D | cache.h | 66 extern void enable_caching(unsigned long start, unsigned long end); 67 extern void disable_caching(unsigned long start, unsigned long end); 80 extern void L1P_cache_block_invalidate(unsigned int start, unsigned int end); 81 extern void L1D_cache_block_invalidate(unsigned int start, unsigned int end); 83 unsigned int end); 84 extern void L1D_cache_block_writeback(unsigned int start, unsigned int end); 85 extern void L2_cache_block_invalidate(unsigned int start, unsigned int end); 86 extern void L2_cache_block_writeback(unsigned int start, unsigned int end); 88 unsigned int end); 90 unsigned int end); 92 unsigned int end); 95 unsigned int end);
|
/linux-4.4.14/arch/m68k/mm/ |
H A D | cache.c | 78 unsigned long start, end; flush_icache_range() local 80 end = endaddr & ICACHE_SET_MASK; flush_icache_range() 81 if (start > end) { flush_icache_range() 82 flush_cf_icache(0, end); flush_icache_range() 83 end = ICACHE_MAX_ADDR; flush_icache_range() 85 flush_cf_icache(start, end); flush_icache_range() 112 unsigned long start, end; flush_icache_user_range() local 114 end = (addr + len) & ICACHE_SET_MASK; flush_icache_user_range() 115 if (start > end) { flush_icache_user_range() 116 flush_cf_icache(0, end); flush_icache_user_range() 117 end = ICACHE_MAX_ADDR; flush_icache_user_range() 119 flush_cf_icache(start, end); flush_icache_user_range()
|
/linux-4.4.14/arch/sparc/lib/ |
H A D | user_fixup.c | 24 unsigned long end = start + size; compute_size() local 26 if (fault_addr < start || fault_addr >= end) { compute_size() 30 size = end - fault_addr; compute_size() 59 unsigned long end = start + size; copy_in_user_fixup() local 61 if (fault_addr >= start && fault_addr < end) copy_in_user_fixup() 62 return end - fault_addr; copy_in_user_fixup() 65 end = start + size; copy_in_user_fixup() 66 if (fault_addr >= start && fault_addr < end) copy_in_user_fixup() 67 return end - fault_addr; copy_in_user_fixup()
|
/linux-4.4.14/mm/ |
H A D | memtest.c | 36 u64 *p, *start, *end; memtest() local 43 end = start + (size - (start_phys_aligned - start_phys)) / incr; memtest() 47 for (p = start; p < end; p++) memtest() 50 for (p = start; p < end; p++, start_phys_aligned += incr) { memtest() 65 static void __init do_one_pass(u64 pattern, phys_addr_t start, phys_addr_t end) do_one_pass() argument 72 this_start = clamp(this_start, start, end); do_one_pass() 73 this_end = clamp(this_end, start, end); do_one_pass() 99 void __init early_memtest(phys_addr_t start, phys_addr_t end) early_memtest() argument 110 do_one_pass(patterns[idx], start, end); early_memtest()
|
H A D | msync.c | 33 unsigned long end; SYSCALL_DEFINE3() local 47 end = start + len; SYSCALL_DEFINE3() 48 if (end < start) SYSCALL_DEFINE3() 51 if (end == start) SYSCALL_DEFINE3() 54 * If the interval [start,end) covers some unmapped address ranges, SYSCALL_DEFINE3() 55 * just ignore them, but return -ENOMEM at the end. SYSCALL_DEFINE3() 63 /* Still start < end. */ SYSCALL_DEFINE3() 70 if (start >= end) SYSCALL_DEFINE3() 83 fend = fstart + (min(end, vma->vm_end) - start) - 1; SYSCALL_DEFINE3() 91 if (error || start >= end) SYSCALL_DEFINE3() 96 if (start >= end) { SYSCALL_DEFINE3()
|
H A D | pagewalk.c | 6 static int walk_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, walk_pte_range() argument 18 if (addr == end) walk_pte_range() 27 static int walk_pmd_range(pud_t *pud, unsigned long addr, unsigned long end, walk_pmd_range() argument 37 next = pmd_addr_end(addr, end); walk_pmd_range() 67 } while (pmd++, addr = next, addr != end); walk_pmd_range() 72 static int walk_pud_range(pgd_t *pgd, unsigned long addr, unsigned long end, walk_pud_range() argument 81 next = pud_addr_end(addr, end); walk_pud_range() 93 } while (pud++, addr = next, addr != end); walk_pud_range() 98 static int walk_pgd_range(unsigned long addr, unsigned long end, walk_pgd_range() argument 107 next = pgd_addr_end(addr, end); walk_pgd_range() 119 } while (pgd++, addr = next, addr != end); walk_pgd_range() 126 unsigned long end) hugetlb_entry_end() 129 return boundary < end ? boundary : end; hugetlb_entry_end() 132 static int walk_hugetlb_range(unsigned long addr, unsigned long end, walk_hugetlb_range() argument 143 next = hugetlb_entry_end(h, addr, end); walk_hugetlb_range() 149 } while (addr = next, addr != end); walk_hugetlb_range() 155 static int walk_hugetlb_range(unsigned long addr, unsigned long end, walk_hugetlb_range() argument 164 * Decide whether we really walk over the current vma on [@start, @end) 169 static int walk_page_test(unsigned long start, unsigned long end, walk_page_test() argument 175 return walk->test_walk(start, end, walk); walk_page_test() 188 err = walk->pte_hole(start, end, walk); walk_page_test() 194 static int __walk_page_range(unsigned long start, unsigned long end, __walk_page_range() argument 202 err = walk_hugetlb_range(start, end, walk); __walk_page_range() 204 err = walk_pgd_range(start, end, walk); __walk_page_range() 213 * within the virtual address range [@start, @end). During walking, we can do 219 * end address yet, continue to walk. 239 int walk_page_range(unsigned long start, unsigned long end, walk_page_range() argument 246 if (start >= end) walk_page_range() 258 next = end; walk_page_range() 261 next = min(end, vma->vm_start); walk_page_range() 264 next = min(end, vma->vm_end); walk_page_range() 284 } while (start = next, start < end); walk_page_range() 125 hugetlb_entry_end(struct hstate *h, unsigned long addr, unsigned long end) hugetlb_entry_end() argument
|
H A D | mprotect.c | 63 unsigned long addr, unsigned long end, pgprot_t newprot, change_pte_range() 129 } while (pte++, addr += PAGE_SIZE, addr != end); change_pte_range() 137 pud_t *pud, unsigned long addr, unsigned long end, change_pmd_range() 151 next = pmd_addr_end(addr, end); change_pmd_range() 158 mmu_notifier_invalidate_range_start(mm, mni_start, end); change_pmd_range() 183 } while (pmd++, addr = next, addr != end); change_pmd_range() 186 mmu_notifier_invalidate_range_end(mm, mni_start, end); change_pmd_range() 194 pgd_t *pgd, unsigned long addr, unsigned long end, change_pud_range() 203 next = pud_addr_end(addr, end); change_pud_range() 208 } while (pud++, addr = next, addr != end); change_pud_range() 214 unsigned long addr, unsigned long end, pgprot_t newprot, change_protection_range() 223 BUG_ON(addr >= end); change_protection_range() 225 flush_cache_range(vma, addr, end); change_protection_range() 228 next = pgd_addr_end(addr, end); change_protection_range() 233 } while (pgd++, addr = next, addr != end); change_protection_range() 237 flush_tlb_range(vma, start, end); change_protection_range() 244 unsigned long end, pgprot_t newprot, change_protection() 250 pages = hugetlb_change_protection(vma, start, end, newprot); change_protection() 252 pages = change_protection_range(vma, start, end, newprot, dirty_accountable, prot_numa); change_protection() 259 unsigned long start, unsigned long end, unsigned long newflags) mprotect_fixup() 263 long nrpages = (end - start) >> PAGE_SHIFT; mprotect_fixup() 294 *pprev = vma_merge(mm, *pprev, start, end, newflags, mprotect_fixup() 310 if (end != vma->vm_end) { mprotect_fixup() 311 error = split_vma(mm, vma, end, 0); mprotect_fixup() 325 change_protection(vma, start, end, vma->vm_page_prot, mprotect_fixup() 334 populate_vma_page_range(vma, start, end, NULL); mprotect_fixup() 350 unsigned long vm_flags, nstart, end, tmp, reqprot; SYSCALL_DEFINE3() local 363 end = start + len; SYSCALL_DEFINE3() 364 if (end <= start) SYSCALL_DEFINE3() 386 if (vma->vm_start >= end) SYSCALL_DEFINE3() 396 end = vma->vm_end; SYSCALL_DEFINE3() 424 if (tmp > end) SYSCALL_DEFINE3() 425 tmp = end; SYSCALL_DEFINE3() 433 if (nstart >= end) SYSCALL_DEFINE3() 62 change_pte_range(struct vm_area_struct *vma, pmd_t *pmd, unsigned long addr, unsigned long end, pgprot_t newprot, int dirty_accountable, int prot_numa) change_pte_range() argument 136 change_pmd_range(struct vm_area_struct *vma, pud_t *pud, unsigned long addr, unsigned long end, pgprot_t newprot, int dirty_accountable, int prot_numa) change_pmd_range() argument 193 change_pud_range(struct vm_area_struct *vma, pgd_t *pgd, unsigned long addr, unsigned long end, pgprot_t newprot, int dirty_accountable, int prot_numa) change_pud_range() argument 213 change_protection_range(struct vm_area_struct *vma, unsigned long addr, unsigned long end, pgprot_t newprot, int dirty_accountable, int prot_numa) change_protection_range() argument 243 change_protection(struct vm_area_struct *vma, unsigned long start, unsigned long end, pgprot_t newprot, int dirty_accountable, int prot_numa) change_protection() argument 258 mprotect_fixup(struct vm_area_struct *vma, struct vm_area_struct **pprev, unsigned long start, unsigned long end, unsigned long newflags) mprotect_fixup() argument
|
H A D | madvise.c | 48 unsigned long start, unsigned long end, int behavior) madvise_behavior() 87 error = ksm_madvise(vma, start, end, behavior, &new_flags); madvise_behavior() 105 *prev = vma_merge(mm, *prev, start, end, new_flags, vma->anon_vma, madvise_behavior() 121 if (end != vma->vm_end) { madvise_behavior() 122 error = split_vma(mm, vma, end, 0); madvise_behavior() 141 unsigned long end, struct mm_walk *walk) swapin_walk_pmd_entry() 150 for (index = start; index != end; index += PAGE_SIZE) { swapin_walk_pmd_entry() 176 unsigned long start, unsigned long end) force_swapin_readahead() 184 walk_page_range(start, end, &walk); force_swapin_readahead() 190 unsigned long start, unsigned long end, force_shm_swapin_readahead() 197 for (; start < end; start += PAGE_SIZE) { force_shm_swapin_readahead() 222 unsigned long start, unsigned long end) madvise_willneed() 229 force_swapin_readahead(vma, start, end); madvise_willneed() 235 force_shm_swapin_readahead(vma, start, end, madvise_willneed() 251 if (end > vma->vm_end) madvise_willneed() 252 end = vma->vm_end; madvise_willneed() 253 end = ((end - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; madvise_willneed() 255 force_page_cache_readahead(file->f_mapping, file, start, end - start); madvise_willneed() 280 unsigned long start, unsigned long end) madvise_dontneed() 286 zap_page_range(vma, start, end - start, NULL); madvise_dontneed() 296 unsigned long start, unsigned long end) madvise_remove() 329 offset, end - start); madvise_remove() 339 static int madvise_hwpoison(int bhv, unsigned long start, unsigned long end) madvise_hwpoison() argument 344 for (; start < end; start += PAGE_SIZE << madvise_hwpoison() 375 unsigned long start, unsigned long end, int behavior) madvise_vma() 379 return madvise_remove(vma, prev, start, end); madvise_vma() 381 return madvise_willneed(vma, prev, start, end); madvise_vma() 383 return madvise_dontneed(vma, prev, start, end); madvise_vma() 385 return madvise_behavior(vma, prev, start, end, behavior); madvise_vma() 462 unsigned long end, tmp; SYSCALL_DEFINE3() local 485 end = start + len; SYSCALL_DEFINE3() 486 if (end < start) SYSCALL_DEFINE3() 490 if (end == start) SYSCALL_DEFINE3() 500 * If the interval [start,end) covers some unmapped address SYSCALL_DEFINE3() 501 * ranges, just ignore them, but return -ENOMEM at the end. SYSCALL_DEFINE3() 510 /* Still start < end. */ SYSCALL_DEFINE3() 515 /* Here start < (end|vma->vm_end). */ SYSCALL_DEFINE3() 519 if (start >= end) SYSCALL_DEFINE3() 523 /* Here vma->vm_start <= start < (end|vma->vm_end) */ SYSCALL_DEFINE3() 525 if (end < tmp) SYSCALL_DEFINE3() 526 tmp = end; SYSCALL_DEFINE3() 528 /* Here vma->vm_start <= start < tmp <= (end|vma->vm_end). */ SYSCALL_DEFINE3() 536 if (start >= end) SYSCALL_DEFINE3() 46 madvise_behavior(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) madvise_behavior() argument 140 swapin_walk_pmd_entry(pmd_t *pmd, unsigned long start, unsigned long end, struct mm_walk *walk) swapin_walk_pmd_entry() argument 175 force_swapin_readahead(struct vm_area_struct *vma, unsigned long start, unsigned long end) force_swapin_readahead() argument 189 force_shm_swapin_readahead(struct vm_area_struct *vma, unsigned long start, unsigned long end, struct address_space *mapping) force_shm_swapin_readahead() argument 220 madvise_willneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_willneed() argument 278 madvise_dontneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_dontneed() argument 294 madvise_remove(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_remove() argument 374 madvise_vma(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) madvise_vma() argument
|
H A D | debug-pagealloc.c | 84 unsigned char *end; check_poison_mem() local 90 for (end = mem + bytes - 1; end > start; end--) { check_poison_mem() 91 if (*end != PAGE_POISON) check_poison_mem() 97 else if (start == end && single_bit_flip(*start, PAGE_POISON)) check_poison_mem() 103 end - start + 1, 1); check_poison_mem()
|
/linux-4.4.14/net/ceph/ |
H A D | armor.c | 4 int ceph_armor(char *dst, const char *src, const char *end); 5 int ceph_unarmor(char *dst, const char *src, const char *end); 36 int ceph_armor(char *dst, const char *src, const char *end) ceph_armor() argument 41 while (src < end) { ceph_armor() 46 if (src < end) { ceph_armor() 49 if (src < end) { ceph_armor() 74 int ceph_unarmor(char *dst, const char *src, const char *end) ceph_unarmor() argument 78 while (src < end) { ceph_unarmor() 85 if (src + 4 > end) ceph_unarmor()
|
H A D | osdmap.c | 55 static int crush_decode_uniform_bucket(void **p, void *end, crush_decode_uniform_bucket() argument 58 dout("crush_decode_uniform_bucket %p to %p\n", *p, end); crush_decode_uniform_bucket() 59 ceph_decode_need(p, end, (1+b->h.size) * sizeof(u32), bad); crush_decode_uniform_bucket() 66 static int crush_decode_list_bucket(void **p, void *end, crush_decode_list_bucket() argument 70 dout("crush_decode_list_bucket %p to %p\n", *p, end); crush_decode_list_bucket() 77 ceph_decode_need(p, end, 2 * b->h.size * sizeof(u32), bad); crush_decode_list_bucket() 87 static int crush_decode_tree_bucket(void **p, void *end, crush_decode_tree_bucket() argument 91 dout("crush_decode_tree_bucket %p to %p\n", *p, end); crush_decode_tree_bucket() 92 ceph_decode_8_safe(p, end, b->num_nodes, bad); crush_decode_tree_bucket() 96 ceph_decode_need(p, end, b->num_nodes * sizeof(u32), bad); crush_decode_tree_bucket() 104 static int crush_decode_straw_bucket(void **p, void *end, crush_decode_straw_bucket() argument 108 dout("crush_decode_straw_bucket %p to %p\n", *p, end); crush_decode_straw_bucket() 115 ceph_decode_need(p, end, 2 * b->h.size * sizeof(u32), bad); crush_decode_straw_bucket() 125 static int crush_decode_straw2_bucket(void **p, void *end, crush_decode_straw2_bucket() argument 129 dout("crush_decode_straw2_bucket %p to %p\n", *p, end); crush_decode_straw2_bucket() 133 ceph_decode_need(p, end, b->h.size * sizeof(u32), bad); crush_decode_straw2_bucket() 141 static int skip_name_map(void **p, void *end) skip_name_map() argument 144 ceph_decode_32_safe(p, end, len ,bad); skip_name_map() 148 ceph_decode_32_safe(p, end, strlen, bad); skip_name_map() 156 static struct crush_map *crush_decode(void *pbyval, void *end) crush_decode() argument 166 dout("crush_decode %p to %p len %d\n", *p, end, (int)(end - *p)); crush_decode() 178 ceph_decode_need(p, end, 4*sizeof(u32), bad); crush_decode() 202 ceph_decode_32_safe(p, end, alg, bad); crush_decode() 208 i, (int)(*p-start), *p, end); crush_decode() 235 ceph_decode_need(p, end, 4*sizeof(u32), bad); crush_decode() 244 b->size, (int)(*p-start), *p, end); crush_decode() 254 ceph_decode_need(p, end, b->size*sizeof(u32), bad); crush_decode() 260 err = crush_decode_uniform_bucket(p, end, crush_decode() 266 err = crush_decode_list_bucket(p, end, crush_decode() 272 err = crush_decode_tree_bucket(p, end, crush_decode() 278 err = crush_decode_straw_bucket(p, end, crush_decode() 284 err = crush_decode_straw2_bucket(p, end, crush_decode() 298 ceph_decode_32_safe(p, end, yes, bad); crush_decode() 301 i, (int)(*p-start), *p, end); crush_decode() 307 i, (int)(*p-start), *p, end); crush_decode() 310 ceph_decode_32_safe(p, end, yes, bad); crush_decode() 324 ceph_decode_copy_safe(p, end, &r->mask, 4, bad); /* 4 u8's */ crush_decode() 325 ceph_decode_need(p, end, r->len*3*sizeof(u32), bad); crush_decode() 335 err = skip_name_map(p, end); crush_decode() 341 ceph_decode_need(p, end, 3*sizeof(u32), done); crush_decode() 352 ceph_decode_need(p, end, sizeof(u32), done); crush_decode() 357 ceph_decode_need(p, end, sizeof(u8), done); crush_decode() 539 static int decode_pool(void **p, void *end, struct ceph_pg_pool_info *pi) decode_pool() argument 545 ceph_decode_need(p, end, 2 + 4, bad); decode_pool() 557 ceph_decode_need(p, end, len, bad); decode_pool() 620 static int decode_pool_names(void **p, void *end, struct ceph_osdmap *map) decode_pool_names() argument 626 ceph_decode_32_safe(p, end, num, bad); decode_pool_names() 629 ceph_decode_64_safe(p, end, pool, bad); decode_pool_names() 630 ceph_decode_32_safe(p, end, len, bad); decode_pool_names() 632 ceph_decode_need(p, end, len, bad); decode_pool_names() 746 static int get_osdmap_client_data_v(void **p, void *end, get_osdmap_client_data_v() argument 751 ceph_decode_8_safe(p, end, struct_v, e_inval); get_osdmap_client_data_v() 755 ceph_decode_8_safe(p, end, struct_compat, e_inval); get_osdmap_client_data_v() 764 ceph_decode_8_safe(p, end, struct_v, e_inval); get_osdmap_client_data_v() 765 ceph_decode_8_safe(p, end, struct_compat, e_inval); get_osdmap_client_data_v() 777 ceph_decode_16_safe(p, end, version, e_inval); get_osdmap_client_data_v() 795 static int __decode_pools(void **p, void *end, struct ceph_osdmap *map, __decode_pools() argument 800 ceph_decode_32_safe(p, end, n, e_inval); __decode_pools() 806 ceph_decode_64_safe(p, end, pool, e_inval); __decode_pools() 823 ret = decode_pool(p, end, pi); __decode_pools() 834 static int decode_pools(void **p, void *end, struct ceph_osdmap *map) decode_pools() argument 836 return __decode_pools(p, end, map, false); decode_pools() 839 static int decode_new_pools(void **p, void *end, struct ceph_osdmap *map) decode_new_pools() argument 841 return __decode_pools(p, end, map, true); decode_new_pools() 844 static int __decode_pg_temp(void **p, void *end, struct ceph_osdmap *map, __decode_pg_temp() argument 849 ceph_decode_32_safe(p, end, n, e_inval); __decode_pg_temp() 855 ret = ceph_decode_pgid(p, end, &pgid); __decode_pg_temp() 859 ceph_decode_32_safe(p, end, len, e_inval); __decode_pg_temp() 867 ceph_decode_need(p, end, len*sizeof(u32), e_inval); __decode_pg_temp() 895 static int decode_pg_temp(void **p, void *end, struct ceph_osdmap *map) decode_pg_temp() argument 897 return __decode_pg_temp(p, end, map, false); decode_pg_temp() 900 static int decode_new_pg_temp(void **p, void *end, struct ceph_osdmap *map) decode_new_pg_temp() argument 902 return __decode_pg_temp(p, end, map, true); decode_new_pg_temp() 905 static int __decode_primary_temp(void **p, void *end, struct ceph_osdmap *map, __decode_primary_temp() argument 910 ceph_decode_32_safe(p, end, n, e_inval); __decode_primary_temp() 916 ret = ceph_decode_pgid(p, end, &pgid); __decode_primary_temp() 920 ceph_decode_32_safe(p, end, osd, e_inval); __decode_primary_temp() 949 static int decode_primary_temp(void **p, void *end, struct ceph_osdmap *map) decode_primary_temp() argument 951 return __decode_primary_temp(p, end, map, false); decode_primary_temp() 954 static int decode_new_primary_temp(void **p, void *end, decode_new_primary_temp() argument 957 return __decode_primary_temp(p, end, map, true); decode_new_primary_temp() 992 static int decode_primary_affinity(void **p, void *end, decode_primary_affinity() argument 997 ceph_decode_32_safe(p, end, len, e_inval); decode_primary_affinity() 1006 ceph_decode_need(p, end, map->max_osd*sizeof(u32), e_inval); decode_primary_affinity() 1022 static int decode_new_primary_affinity(void **p, void *end, decode_new_primary_affinity() argument 1027 ceph_decode_32_safe(p, end, n, e_inval); decode_new_primary_affinity() 1032 ceph_decode_32_safe(p, end, osd, e_inval); decode_new_primary_affinity() 1033 ceph_decode_32_safe(p, end, aff, e_inval); decode_new_primary_affinity() 1051 static int osdmap_decode(void **p, void *end, struct ceph_osdmap *map) osdmap_decode() argument 1060 dout("%s %p to %p len %d\n", __func__, *p, end, (int)(end - *p)); osdmap_decode() 1062 err = get_osdmap_client_data_v(p, end, "full", &struct_v); osdmap_decode() 1067 ceph_decode_need(p, end, sizeof(map->fsid) + sizeof(u32) + osdmap_decode() 1075 err = decode_pools(p, end, map); osdmap_decode() 1080 err = decode_pool_names(p, end, map); osdmap_decode() 1084 ceph_decode_32_safe(p, end, map->pool_max, e_inval); osdmap_decode() 1086 ceph_decode_32_safe(p, end, map->flags, e_inval); osdmap_decode() 1089 ceph_decode_32_safe(p, end, max, e_inval); osdmap_decode() 1097 ceph_decode_need(p, end, 3*sizeof(u32) + osdmap_decode() 1120 err = decode_pg_temp(p, end, map); osdmap_decode() 1126 err = decode_primary_temp(p, end, map); osdmap_decode() 1133 err = decode_primary_affinity(p, end, map); osdmap_decode() 1143 ceph_decode_32_safe(p, end, len, e_inval); osdmap_decode() 1144 map->crush = crush_decode(*p, min(*p + len, end)); osdmap_decode() 1153 *p = end; osdmap_decode() 1162 err, epoch, (int)(*p - start), *p, start, end); osdmap_decode() 1165 start, end - start, true); osdmap_decode() 1172 struct ceph_osdmap *ceph_osdmap_decode(void **p, void *end) ceph_osdmap_decode() argument 1185 ret = osdmap_decode(p, end, map); ceph_osdmap_decode() 1197 struct ceph_osdmap *osdmap_apply_incremental(void **p, void *end, osdmap_apply_incremental() argument 1213 dout("%s %p to %p len %d\n", __func__, *p, end, (int)(end - *p)); osdmap_apply_incremental() 1215 err = get_osdmap_client_data_v(p, end, "inc", &struct_v); osdmap_apply_incremental() 1220 ceph_decode_need(p, end, sizeof(fsid) + sizeof(u32) + sizeof(modified) + osdmap_apply_incremental() 1230 ceph_decode_32_safe(p, end, len, e_inval); osdmap_apply_incremental() 1233 len, *p, end); osdmap_apply_incremental() 1234 return ceph_osdmap_decode(p, min(*p+len, end)); osdmap_apply_incremental() 1238 ceph_decode_32_safe(p, end, len, e_inval); osdmap_apply_incremental() 1240 newcrush = crush_decode(*p, min(*p+len, end)); osdmap_apply_incremental() 1256 ceph_decode_32_safe(p, end, max, e_inval); osdmap_apply_incremental() 1273 err = decode_new_pools(p, end, map); osdmap_apply_incremental() 1278 err = decode_pool_names(p, end, map); osdmap_apply_incremental() 1283 ceph_decode_32_safe(p, end, len, e_inval); osdmap_apply_incremental() 1287 ceph_decode_64_safe(p, end, pool, e_inval); osdmap_apply_incremental() 1294 ceph_decode_32_safe(p, end, len, e_inval); osdmap_apply_incremental() 1298 ceph_decode_32_safe(p, end, osd, e_inval); osdmap_apply_incremental() 1299 ceph_decode_copy_safe(p, end, &addr, sizeof(addr), e_inval); osdmap_apply_incremental() 1308 ceph_decode_32_safe(p, end, len, e_inval); osdmap_apply_incremental() 1312 ceph_decode_32_safe(p, end, osd, e_inval); osdmap_apply_incremental() 1324 ceph_decode_32_safe(p, end, len, e_inval); osdmap_apply_incremental() 1327 ceph_decode_need(p, end, sizeof(u32)*2, e_inval); osdmap_apply_incremental() 1338 err = decode_new_pg_temp(p, end, map); osdmap_apply_incremental() 1344 err = decode_new_primary_temp(p, end, map); osdmap_apply_incremental() 1351 err = decode_new_primary_affinity(p, end, map); osdmap_apply_incremental() 1357 *p = end; osdmap_apply_incremental() 1366 err, epoch, (int)(*p - start), *p, start, end); osdmap_apply_incremental() 1369 start, end - start, true); osdmap_apply_incremental()
|
H A D | buffer.c | 43 int ceph_decode_buffer(struct ceph_buffer **b, void **p, void *end) ceph_decode_buffer() argument 47 ceph_decode_need(p, end, sizeof(u32), bad); ceph_decode_buffer() 50 ceph_decode_need(p, end, len, bad); ceph_decode_buffer()
|
H A D | auth_none.c | 42 static int build_request(struct ceph_auth_client *ac, void *buf, void *end) build_request() argument 52 void *buf, void *end) handle_reply() 71 void *p, *end; ceph_auth_none_create_authorizer() local 76 end = p + sizeof(au->buf); ceph_auth_none_create_authorizer() 78 ret = ceph_entity_name_encode(ac->name, &p, end - 8); ceph_auth_none_create_authorizer() 81 ceph_decode_need(&p, end, sizeof(u64), bad2); ceph_auth_none_create_authorizer() 51 handle_reply(struct ceph_auth_client *ac, int result, void *buf, void *end) handle_reply() argument
|
/linux-4.4.14/arch/sh/mm/ |
H A D | cache-sh2.c | 22 unsigned long begin, end; sh2__flush_wback_region() local 25 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2__flush_wback_region() 27 for (v = begin; v < end; v+=L1_CACHE_BYTES) { sh2__flush_wback_region() 43 unsigned long begin, end; sh2__flush_purge_region() local 46 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2__flush_purge_region() 49 for (v = begin; v < end; v+=L1_CACHE_BYTES) sh2__flush_purge_region() 74 unsigned long begin, end; sh2__flush_invalidate_region() 77 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2__flush_invalidate_region() 80 for (v = begin; v < end; v+=L1_CACHE_BYTES) sh2__flush_invalidate_region()
|
H A D | flush-sh4.c | 15 reg_size_t aligned_start, v, cnt, end; sh4__flush_wback_region() local 19 end = (aligned_start + size + L1_CACHE_BYTES-1) sh4__flush_wback_region() 21 cnt = (end - v) / L1_CACHE_BYTES; sh4__flush_wback_region() 49 reg_size_t aligned_start, v, cnt, end; sh4__flush_purge_region() local 53 end = (aligned_start + size + L1_CACHE_BYTES-1) sh4__flush_purge_region() 55 cnt = (end - v) / L1_CACHE_BYTES; sh4__flush_purge_region() 79 reg_size_t aligned_start, v, cnt, end; sh4__flush_invalidate_region() local 83 end = (aligned_start + size + L1_CACHE_BYTES-1) sh4__flush_invalidate_region() 85 cnt = (end - v) / L1_CACHE_BYTES; sh4__flush_invalidate_region()
|
H A D | cache-sh2a.c | 54 unsigned long begin, end; sh2a__flush_wback_region() local 59 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2a__flush_wback_region() 67 if (((end - begin) >> PAGE_SHIFT) >= MAX_OCACHE_PAGES) { sh2a__flush_wback_region() 69 end = begin + (nr_ways * current_cpu_data.dcache.way_size); sh2a__flush_wback_region() 71 for (v = begin; v < end; v += L1_CACHE_BYTES) { sh2a__flush_wback_region() 79 for (v = begin; v < end; v += L1_CACHE_BYTES) sh2a__flush_wback_region() 95 unsigned long begin, end; sh2a__flush_purge_region() local 99 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2a__flush_purge_region() 105 for (v = begin; v < end; v+=L1_CACHE_BYTES) { sh2a__flush_purge_region() 125 unsigned long begin, end; sh2a__flush_invalidate_region() local 129 end = ((unsigned long)start + size + L1_CACHE_BYTES-1) sh2a__flush_invalidate_region() 136 if (((end - begin) >> PAGE_SHIFT) >= MAX_OCACHE_PAGES) { sh2a__flush_invalidate_region() 140 for (v = begin; v < end; v += L1_CACHE_BYTES) sh2a__flush_invalidate_region() 154 unsigned long start, end; sh2a_flush_icache_range() local 159 end = (data->addr2 + L1_CACHE_BYTES-1) & ~(L1_CACHE_BYTES-1); sh2a_flush_icache_range() 162 sh2a__flush_wback_region((void *)start, end-start); sh2a_flush_icache_range() 170 if (((end - start) >> PAGE_SHIFT) >= MAX_ICACHE_PAGES) { sh2a_flush_icache_range() 174 for (v = start; v < end; v += L1_CACHE_BYTES) sh2a_flush_icache_range()
|
H A D | numa.c | 27 void __init setup_bootmem_node(int nid, unsigned long start, unsigned long end) setup_bootmem_node() argument 37 end_pfn = PFN_DOWN(end); setup_bootmem_node() 39 pmb_bolt_mapping((unsigned long)__va(start), start, end - start, setup_bootmem_node() 42 memblock_add(start, end - start); setup_bootmem_node() 48 SMP_CACHE_BYTES, end)); setup_bootmem_node() 58 PAGE_SIZE, end); setup_bootmem_node()
|
H A D | tlbflush_32.c | 40 unsigned long end) local_flush_tlb_range() 50 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_range() 61 end += (PAGE_SIZE - 1); local_flush_tlb_range() 62 end &= PAGE_MASK; local_flush_tlb_range() 67 while (start < end) { local_flush_tlb_range() 78 void local_flush_tlb_kernel_range(unsigned long start, unsigned long end) local_flush_tlb_kernel_range() argument 85 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_kernel_range() 94 end += (PAGE_SIZE - 1); local_flush_tlb_kernel_range() 95 end &= PAGE_MASK; local_flush_tlb_kernel_range() 97 while (start < end) { local_flush_tlb_kernel_range() 39 local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) local_flush_tlb_range() argument
|
H A D | gup.c | 75 unsigned long end, int write, struct page **pages, int *nr) gup_pte_range() 113 } while (ptep++, addr += PAGE_SIZE, addr != end); gup_pte_range() 119 static int gup_pmd_range(pud_t pud, unsigned long addr, unsigned long end, gup_pmd_range() argument 129 next = pmd_addr_end(addr, end); gup_pmd_range() 134 } while (pmdp++, addr = next, addr != end); gup_pmd_range() 139 static int gup_pud_range(pgd_t pgd, unsigned long addr, unsigned long end, gup_pud_range() argument 149 next = pud_addr_end(addr, end); gup_pud_range() 154 } while (pudp++, addr = next, addr != end); gup_pud_range() 167 unsigned long addr, len, end; __get_user_pages_fast() local 176 end = start + len; __get_user_pages_fast() 190 next = pgd_addr_end(addr, end); __get_user_pages_fast() 195 } while (pgdp++, addr = next, addr != end); __get_user_pages_fast() 221 unsigned long addr, len, end; get_user_pages_fast() local 230 end = start + len; get_user_pages_fast() 231 if (end < start) get_user_pages_fast() 239 next = pgd_addr_end(addr, end); get_user_pages_fast() 244 } while (pgdp++, addr = next, addr != end); get_user_pages_fast() 247 VM_BUG_ON(nr != (end - start) >> PAGE_SHIFT); get_user_pages_fast() 261 (end - start) >> PAGE_SHIFT, write, 0, pages); get_user_pages_fast() 74 gup_pte_range(pmd_t pmd, unsigned long addr, unsigned long end, int write, struct page **pages, int *nr) gup_pte_range() argument
|
/linux-4.4.14/arch/arm/mach-davinci/ |
H A D | devices-da8xx.c | 164 .end = DA8XX_TPCC_BASE + SZ_32K - 1, 170 .end = DA8XX_TPTC0_BASE + SZ_1K - 1, 176 .end = DA8XX_TPTC1_BASE + SZ_1K - 1, 195 .end = DA850_TPCC1_BASE + SZ_32K - 1, 201 .end = DA850_TPTC2_BASE + SZ_1K - 1, 267 .end = DA8XX_I2C0_BASE + SZ_4K - 1, 272 .end = IRQ_DA8XX_I2CINT0, 287 .end = DA8XX_I2C1_BASE + SZ_4K - 1, 292 .end = IRQ_DA8XX_I2CINT1, 323 .end = DA8XX_WDOG_BASE + SZ_4K - 1, 356 .end = DA8XX_EMAC_CPPI_PORT_BASE + SZ_16K - 1, 361 .end = IRQ_DA8XX_C0_RX_THRESH_PULSE, 366 .end = IRQ_DA8XX_C0_RX_PULSE, 371 .end = IRQ_DA8XX_C0_TX_PULSE, 376 .end = IRQ_DA8XX_C0_MISC_PULSE, 402 .end = DA8XX_EMAC_MDIO_BASE + SZ_4K - 1, 429 .end = DAVINCI_DA830_MCASP1_REG_BASE + (SZ_1K * 12) - 1, 436 .end = DAVINCI_DA830_DMA_MCASP1_AXEVT, 443 .end = DAVINCI_DA830_DMA_MCASP1_AREVT, 464 .end = DAVINCI_DA830_MCASP2_REG_BASE + (SZ_1K * 12) - 1, 471 .end = DAVINCI_DA830_DMA_MCASP2_AXEVT, 478 .end = DAVINCI_DA830_DMA_MCASP2_AREVT, 499 .end = DAVINCI_DA8XX_MCASP0_REG_BASE + (SZ_1K * 12) - 1, 506 .end = DAVINCI_DA8XX_DMA_MCASP0_AXEVT, 513 .end = DAVINCI_DA8XX_DMA_MCASP0_AREVT, 562 .end = DA8XX_PRUSS_MEM_BASE + 0xFFFF, 567 .end = IRQ_DA8XX_EVTOUT0, 572 .end = IRQ_DA8XX_EVTOUT1, 577 .end = IRQ_DA8XX_EVTOUT2, 582 .end = IRQ_DA8XX_EVTOUT3, 587 .end = IRQ_DA8XX_EVTOUT4, 592 .end = IRQ_DA8XX_EVTOUT5, 597 .end = IRQ_DA8XX_EVTOUT6, 602 .end = IRQ_DA8XX_EVTOUT7, 648 .end = DA8XX_LCD_CNTRL_BASE + SZ_4K - 1, 653 .end = IRQ_DA8XX_LCDINT, 674 .end = DA8XX_GPIO_BASE + SZ_4K - 1, 679 .end = IRQ_DA8XX_GPIO8, 700 .end = DA8XX_MMCSD0_BASE + SZ_4K - 1, 705 .end = IRQ_DA8XX_MMCSDINT0, 710 .end = DA8XX_DMA_MMCSD0_RX, 715 .end = DA8XX_DMA_MMCSD0_TX, 737 .end = DA850_MMCSD1_BASE + SZ_4K - 1, 742 .end = IRQ_DA850_MMCSDINT0_1, 747 .end = DA850_DMA_MMCSD1_RX, 752 .end = DA850_DMA_MMCSD1_TX, 774 .end = DA8XX_SYSCFG0_BASE + DA8XX_HOST1CFG_REG + 3, 779 .end = DA8XX_SYSCFG0_BASE + DA8XX_CHIPSIG_REG + 7, 784 .end = IRQ_DA8XX_CHIPINT0, 860 .end = DA8XX_RTC_BASE + SZ_4K - 1, 865 .end = IRQ_DA8XX_RTC, 870 .end = IRQ_DA8XX_RTC, 903 .end = DA8XX_DDR2_CTL_BASE + SZ_32K - 1, 933 .end = DA8XX_SPI0_BASE + SZ_4K - 1, 938 .end = IRQ_DA8XX_SPINT0, 943 .end = DA8XX_DMA_SPI0_RX, 948 .end = DA8XX_DMA_SPI0_TX, 956 .end = DA830_SPI1_BASE + SZ_4K - 1, 961 .end = IRQ_DA8XX_SPINT1, 966 .end = DA8XX_DMA_SPI1_RX, 971 .end = DA8XX_DMA_SPI1_TX, 1021 da8xx_spi1_resources[0].end = DA850_SPI1_BASE + SZ_4K - 1; da8xx_register_spi_bus() 1031 .end = DA850_SATA_BASE + 0x1fff, 1036 .end = DA8XX_SYSCFG1_BASE + DA8XX_PWRDN_REG + 0x3,
|
/linux-4.4.14/arch/blackfin/mach-bf538/boards/ |
H A D | ezkit.c | 48 .end = UART0_GCTL+2, 53 .end = IRQ_UART0_TX, 58 .end = IRQ_UART0_RX, 63 .end = IRQ_UART0_ERROR, 68 .end = CH_UART0_TX, 73 .end = CH_UART0_RX, 79 .end = GPIO_PG7, 84 .end = GPIO_PG6, 108 .end = UART1_GCTL+2, 113 .end = IRQ_UART1_TX, 118 .end = IRQ_UART1_RX, 123 .end = IRQ_UART1_ERROR, 128 .end = CH_UART1_TX, 133 .end = CH_UART1_RX, 156 .end = UART2_GCTL+2, 161 .end = IRQ_UART2_TX, 166 .end = IRQ_UART2_RX, 171 .end = IRQ_UART2_ERROR, 176 .end = CH_UART2_TX, 181 .end = CH_UART2_RX, 207 .end = 0xFFC004FF, 212 .end = IRQ_UART0_RX+1, 217 .end = CH_UART0_RX+1, 232 .end = 0xFFC020FF, 237 .end = IRQ_UART1_RX+1, 242 .end = CH_UART1_RX+1, 257 .end = 0xFFC021FF, 262 .end = IRQ_UART2_RX+1, 267 .end = CH_UART2_RX+1, 285 .end = SPORT0_MRCS3+4, 290 .end = IRQ_SPORT0_RX+1, 295 .end = IRQ_SPORT0_ERROR, 319 .end = SPORT1_MRCS3+4, 324 .end = IRQ_SPORT1_RX+1, 329 .end = IRQ_SPORT1_ERROR, 353 .end = SPORT2_MRCS3+4, 358 .end = IRQ_SPORT2_RX+1, 363 .end = IRQ_SPORT2_ERROR, 387 .end = SPORT3_MRCS3+4, 392 .end = IRQ_SPORT3_RX+1, 397 .end = IRQ_SPORT3_ERROR, 427 .end = 0xFFC02FFF, 432 .end = IRQ_CAN_RX, 437 .end = IRQ_CAN_TX, 442 .end = IRQ_CAN_ERROR, 474 .end = 0x20310300 + 16, 478 .end = IRQ_PF0, 553 .end = IRQ_PPI_ERROR, 616 .end = SPI0_REGBASE + 0xFF, 621 .end = CH_SPI0, 626 .end = IRQ_SPI0, 635 .end = SPI1_REGBASE + 0xFF, 640 .end = CH_SPI1, 645 .end = IRQ_SPI1, 654 .end = SPI2_REGBASE + 0xFF, 659 .end = CH_SPI2, 664 .end = IRQ_SPI2, 724 .end = TWI0_REGBASE + 0xFF, 729 .end = IRQ_TWI0, 749 .end = TWI1_REGBASE + 0xFF, 754 .end = IRQ_TWI1, 841 .end = 0x202fffff, 843 .end = 0x203fffff,
|
/linux-4.4.14/arch/xtensa/mm/ |
H A D | init.c | 43 sysmem.bank[i].start, sysmem.bank[i].end, sysmem_dump() 44 (sysmem.bank[i].end - sysmem.bank[i].start) >> 10); sysmem_dump() 88 int __init add_sysmem_bank(unsigned long start, unsigned long end) add_sysmem_bank() argument 95 if (start == end || add_sysmem_bank() 96 (start < end) != (PAGE_ALIGN(start) < (end & PAGE_MASK))) { add_sysmem_bank() 98 start, end - start); add_sysmem_bank() 103 end &= PAGE_MASK; add_sysmem_bank() 104 sz = end - start; add_sysmem_bank() 109 bank_sz = it->end - it->start; add_sysmem_bank() 112 if (end - it->start > bank_sz) add_sysmem_bank() 113 it->end = end; add_sysmem_bank() 125 if (it->end - it->start < sz) add_sysmem_bank() 126 it->end = end; add_sysmem_bank() 132 start, end - start); add_sysmem_bank() 136 it->end = end; add_sysmem_bank() 140 sz = it->end - it->start; add_sysmem_bank() 143 if (sz < sysmem.bank[i].end - it->start) add_sysmem_bank() 144 it->end = sysmem.bank[i].end; add_sysmem_bank() 154 * mem_reserve(start, end, must_exist) 162 * end End of region, 170 int __init mem_reserve(unsigned long start, unsigned long end, int must_exist) mem_reserve() argument 178 end = PAGE_ALIGN(end); mem_reserve() 179 sz = end - start; mem_reserve() 186 bank_sz = it->end - it->start; mem_reserve() 188 if ((!it || end - it->start > bank_sz) && must_exist) { mem_reserve() 190 start, end); mem_reserve() 196 if (end - it->start < bank_sz) { mem_reserve() 197 it->start = end; mem_reserve() 203 it->end = start; mem_reserve() 204 if (end - it->start < bank_sz) mem_reserve() 205 return add_sysmem_bank(end, mem_reserve() 215 if (it->end - start <= sz) { mem_reserve() 220 it->start = end; mem_reserve() 261 pfn = PAGE_ALIGN(sysmem.bank[i].end - 1) >> PAGE_SHIFT; bootmem_init() 279 if (sysmem.bank[i].end - sysmem.bank[i].start >= bootmap_size) { bootmem_init() 299 unsigned long end = min(max_low_pfn << PAGE_SHIFT, bootmem_init() local 300 sysmem.bank[i].end); bootmem_init() 302 end - sysmem.bank[i].start); bootmem_init() 370 void free_initrd_mem(unsigned long start, unsigned long end) free_initrd_mem() argument 373 free_reserved_area((void *)start, (void *)end, -1, "initrd"); free_initrd_mem()
|
/linux-4.4.14/arch/blackfin/mach-bf537/boards/ |
H A D | minotaur.c | 38 .end = 0x20312000, 42 .end = 0x20311FFF, 46 .end = IRQ_PF4, 50 .end = IRQ_PF6, 107 .end = 0x20300000 + 0x100, 111 .end = IRQ_PF7, 206 .end = SPI0_REGBASE + 0xFF, 211 .end = CH_SPI, 216 .end = IRQ_SPI, 237 .end = UART0_GCTL+2, 242 .end = IRQ_UART0_TX, 247 .end = IRQ_UART0_RX, 252 .end = IRQ_UART0_ERROR, 257 .end = CH_UART0_TX, 262 .end = CH_UART0_RX, 285 .end = UART1_GCTL+2, 290 .end = IRQ_UART1_TX, 295 .end = IRQ_UART1_RX, 300 .end = IRQ_UART1_ERROR, 305 .end = CH_UART1_TX, 310 .end = CH_UART1_RX, 336 .end = 0xFFC004FF, 341 .end = IRQ_UART0_RX+1, 346 .end = CH_UART0_RX+1, 362 .end = 0xFFC020FF, 367 .end = IRQ_UART1_RX+1, 372 .end = CH_UART1_RX+1, 392 .end = TWI0_REGBASE + 0xFF, 397 .end = IRQ_TWI, 418 .end = SPORT0_MRCS3+4, 423 .end = IRQ_SPORT0_RX+1, 428 .end = IRQ_SPORT0_ERROR, 452 .end = SPORT1_MRCS3+4, 457 .end = IRQ_SPORT1_RX+1, 462 .end = IRQ_SPORT1_ERROR,
|
H A D | tcm_bf537.c | 115 .end = SPI0_REGBASE + 0xFF, 120 .end = CH_SPI, 125 .end = IRQ_SPI, 173 .end = 0x20200300 + 16, 177 .end = IRQ_PF14, 197 .end = 0x20308000, 201 .end = 0x20308004, 205 .end = IRQ_PG15, 236 .end = 0x20300000 + 0x100, 240 .end = IRQ_PG13, 282 .end = 0x201fffff, 286 .end = ARRAY_SIZE(cm_flash_gpios), 307 .end = UART0_GCTL+2, 312 .end = IRQ_UART0_TX, 317 .end = IRQ_UART0_RX, 322 .end = IRQ_UART0_ERROR, 327 .end = CH_UART0_TX, 332 .end = CH_UART0_RX, 355 .end = UART1_GCTL+2, 360 .end = IRQ_UART1_TX, 365 .end = IRQ_UART1_RX, 370 .end = IRQ_UART1_ERROR, 375 .end = CH_UART1_TX, 380 .end = CH_UART1_RX, 406 .end = 0xFFC004FF, 411 .end = IRQ_UART0_RX+1, 416 .end = CH_UART0_RX+1, 432 .end = 0xFFC020FF, 437 .end = IRQ_UART1_RX+1, 442 .end = CH_UART1_RX+1, 462 .end = TWI0_REGBASE, 467 .end = IRQ_TWI, 488 .end = SPORT0_MRCS3+4, 493 .end = IRQ_SPORT0_RX+1, 498 .end = IRQ_SPORT0_ERROR, 522 .end = SPORT1_MRCS3+4, 527 .end = IRQ_SPORT1_RX+1, 532 .end = IRQ_SPORT1_ERROR, 597 .end = 0x2030C01F, 602 .end = 0x2030D01B, 607 .end = PATA_INT,
|
H A D | cm_bf537e.c | 115 .end = SPI0_REGBASE + 0xFF, 120 .end = CH_SPI, 125 .end = IRQ_SPI, 161 .end = SPORT0_TCR1 + 0xFF, 166 .end = IRQ_SPORT0_ERROR, 191 .end = SPORT1_TCR1 + 0xFF, 196 .end = IRQ_SPORT1_ERROR, 238 .end = 0x20200300 + 16, 242 .end = IRQ_PF14, 262 .end = 0x20308000, 266 .end = 0x20308004, 270 .end = IRQ_PG15, 301 .end = 0x20300000 + 0x100, 305 .end = IRQ_PG13, 347 .end = 0x201fffff, 351 .end = ARRAY_SIZE(cm_flash_gpios), 372 .end = UART0_GCTL+2, 377 .end = IRQ_UART0_TX, 382 .end = IRQ_UART0_RX, 387 .end = IRQ_UART0_ERROR, 392 .end = CH_UART0_TX, 397 .end = CH_UART0_RX, 406 .end = -1, 414 .end = -1, 438 .end = UART1_GCTL+2, 443 .end = IRQ_UART1_TX, 448 .end = IRQ_UART1_RX, 453 .end = IRQ_UART1_ERROR, 458 .end = CH_UART1_TX, 463 .end = CH_UART1_RX, 472 .end = -1, 480 .end = -1, 507 .end = 0xFFC004FF, 512 .end = IRQ_UART0_RX+1, 517 .end = CH_UART0_RX+1, 532 .end = 0xFFC020FF, 537 .end = IRQ_UART1_RX+1, 542 .end = CH_UART1_RX+1, 561 .end = TWI0_REGBASE, 566 .end = IRQ_TWI, 594 .end = SPORT0_MRCS3+4, 599 .end = IRQ_SPORT0_RX+1, 604 .end = IRQ_SPORT0_ERROR, 623 .end = SPORT1_MRCS3+4, 628 .end = IRQ_SPORT1_RX+1, 633 .end = IRQ_SPORT1_ERROR, 658 .end = SPORT0_MRCS3+4, 663 .end = IRQ_SPORT0_RX+1, 668 .end = IRQ_SPORT0_TX+1, 673 .end = IRQ_SPORT0_ERROR, 678 .end = CH_SPORT0_TX, 683 .end = CH_SPORT0_RX, 741 .end = 0x2030C01F, 746 .end = 0x2030D01B, 751 .end = PATA_INT,
|
H A D | pnav10.c | 37 .end = 0x20312000, 41 .end = 0x20311FFF, 45 .end = IRQ_PF4, 49 .end = 6, 82 .end = 0x20300300 + 16, 87 .end = IRQ_PF7, 139 .end = 0x20300000 + 0x100, 143 .end = IRQ_PF7, 261 .end = SPI0_REGBASE + 0xFF, 266 .end = CH_SPI, 271 .end = IRQ_SPI, 305 .end = UART0_GCTL+2, 310 .end = IRQ_UART0_TX, 315 .end = IRQ_UART0_RX, 320 .end = IRQ_UART0_ERROR, 325 .end = CH_UART0_TX, 330 .end = CH_UART0_RX, 353 .end = UART1_GCTL+2, 358 .end = IRQ_UART1_TX, 363 .end = IRQ_UART1_RX, 368 .end = IRQ_UART1_ERROR, 373 .end = CH_UART1_TX, 378 .end = CH_UART1_RX, 404 .end = 0xFFC004FF, 409 .end = IRQ_UART0_RX+1, 414 .end = CH_UART0_RX+1, 430 .end = 0xFFC020FF, 435 .end = IRQ_UART1_RX+1, 440 .end = CH_UART1_RX+1,
|
H A D | cm_bf537u.c | 115 .end = SPI0_REGBASE + 0xFF, 120 .end = CH_SPI, 125 .end = IRQ_SPI, 173 .end = 0x20200300 + 16, 177 .end = IRQ_PF14, 197 .end = 0x20308000, 201 .end = 0x20308004, 205 .end = IRQ_PG15, 236 .end = 0x20200000 + 0x100, 240 .end = IRQ_PH14, 282 .end = 0x201fffff, 286 .end = ARRAY_SIZE(cm_flash_gpios), 307 .end = UART0_GCTL+2, 312 .end = IRQ_UART0_TX, 317 .end = IRQ_UART0_RX, 322 .end = IRQ_UART0_ERROR, 327 .end = CH_UART0_TX, 332 .end = CH_UART0_RX, 355 .end = UART1_GCTL+2, 360 .end = IRQ_UART1_TX, 365 .end = IRQ_UART1_RX, 370 .end = IRQ_UART1_ERROR, 375 .end = CH_UART1_TX, 380 .end = CH_UART1_RX, 406 .end = 0xFFC004FF, 411 .end = IRQ_UART0_RX+1, 416 .end = CH_UART0_RX+1, 431 .end = 0xFFC020FF, 436 .end = IRQ_UART1_RX+1, 441 .end = CH_UART1_RX+1, 460 .end = TWI0_REGBASE, 465 .end = IRQ_TWI, 486 .end = SPORT0_MRCS3+4, 491 .end = IRQ_SPORT0_RX+1, 496 .end = IRQ_SPORT0_ERROR, 520 .end = SPORT1_MRCS3+4, 525 .end = IRQ_SPORT1_RX+1, 530 .end = IRQ_SPORT1_ERROR, 595 .end = 0x2030C01F, 600 .end = 0x2030D01B, 605 .end = PATA_INT,
|
/linux-4.4.14/arch/arm/mm/ |
H A D | cache-feroceon-l2.c | 34 * end addresses to successive cp15 registers, and process every 36 * [start:end]. 73 static inline void l2_clean_pa_range(unsigned long start, unsigned long end) l2_clean_pa_range() argument 78 * Make sure 'start' and 'end' reference the same page, as l2_clean_pa_range() 82 BUG_ON((start ^ end) >> PAGE_SHIFT); l2_clean_pa_range() 85 va_end = va_start + (end - start); l2_clean_pa_range() 104 static inline void l2_inv_pa_range(unsigned long start, unsigned long end) l2_inv_pa_range() argument 109 * Make sure 'start' and 'end' reference the same page, as l2_inv_pa_range() 113 BUG_ON((start ^ end) >> PAGE_SHIFT); l2_inv_pa_range() 116 va_end = va_start + (end - start); l2_inv_pa_range() 133 * Note that the end addresses passed to Linux primitives are 135 * inclusive start and end addresses. 142 static unsigned long calc_range_end(unsigned long start, unsigned long end) calc_range_end() argument 147 BUG_ON(end & (CACHE_LINE_SIZE - 1)); calc_range_end() 150 * Try to process all cache lines between 'start' and 'end'. calc_range_end() 152 range_end = end; calc_range_end() 171 static void feroceon_l2_inv_range(unsigned long start, unsigned long end) feroceon_l2_inv_range() argument 184 if (start < end && end & (CACHE_LINE_SIZE - 1)) { feroceon_l2_inv_range() 185 l2_clean_inv_pa(end & ~(CACHE_LINE_SIZE - 1)); feroceon_l2_inv_range() 186 end &= ~(CACHE_LINE_SIZE - 1); feroceon_l2_inv_range() 190 * Invalidate all full cache lines between 'start' and 'end'. feroceon_l2_inv_range() 192 while (start < end) { feroceon_l2_inv_range() 193 unsigned long range_end = calc_range_end(start, end); feroceon_l2_inv_range() 201 static void feroceon_l2_clean_range(unsigned long start, unsigned long end) feroceon_l2_clean_range() argument 209 end = (end + CACHE_LINE_SIZE - 1) & ~(CACHE_LINE_SIZE - 1); feroceon_l2_clean_range() 210 while (start != end) { feroceon_l2_clean_range() 211 unsigned long range_end = calc_range_end(start, end); feroceon_l2_clean_range() 220 static void feroceon_l2_flush_range(unsigned long start, unsigned long end) feroceon_l2_flush_range() argument 223 end = (end + CACHE_LINE_SIZE - 1) & ~(CACHE_LINE_SIZE - 1); feroceon_l2_flush_range() 224 while (start != end) { feroceon_l2_flush_range() 225 unsigned long range_end = calc_range_end(start, end); feroceon_l2_flush_range()
|
H A D | idmap.c | 21 static void idmap_add_pmd(pud_t *pud, unsigned long addr, unsigned long end, idmap_add_pmd() argument 46 next = pmd_addr_end(addr, end); idmap_add_pmd() 49 } while (pmd++, addr = next, addr != end); idmap_add_pmd() 52 static void idmap_add_pmd(pud_t *pud, unsigned long addr, unsigned long end, idmap_add_pmd() argument 65 static void idmap_add_pud(pgd_t *pgd, unsigned long addr, unsigned long end, idmap_add_pud() argument 72 next = pud_addr_end(addr, end); idmap_add_pud() 74 } while (pud++, addr = next, addr != end); idmap_add_pud() 80 unsigned long addr, end; identity_mapping_add() local 84 end = virt_to_idmap(text_end); identity_mapping_add() 85 pr_info("Setting up static identity map for 0x%lx - 0x%lx\n", addr, end); identity_mapping_add() 94 next = pgd_addr_end(addr, end); identity_mapping_add() 96 } while (pgd++, addr = next, addr != end); identity_mapping_add()
|
H A D | cache-v4wt.S | 78 * flush_user_cache_range(start, end, flags) 84 * - end - end address (exclusive, page aligned) 101 * coherent_kern_range(start, end) 108 * - end - virtual end address 114 * coherent_user_range(start, end) 121 * - end - virtual end address 148 * dma_inv_range(start, end) 151 * May not write back any entries. If 'start' or 'end' 156 * - end - virtual end address 167 * dma_flush_range(start, end) 172 * - end - virtual end address
|
H A D | cache-fa.S | 78 * flush_user_cache_range(start, end, flags) 84 * - end - end address (exclusive, page aligned) 106 * coherent_kern_range(start, end) 113 * - end - virtual end address 119 * coherent_user_range(start, end) 126 * - end - virtual end address 162 * dma_inv_range(start, end) 165 * May not write back any entries. If 'start' or 'end' 170 * - end - virtual end address 188 * dma_clean_range(start, end) 193 * - end - virtual end address 206 * dma_flush_range(start,end) 208 * - end - virtual end address of region
|
H A D | cache-v4wb.S | 101 * flush_user_cache_range(start, end, flags) 107 * - end - end address (exclusive, page aligned) 142 * coherent_kern_range(start, end) 149 * - end - virtual end address 155 * coherent_user_range(start, end) 162 * - end - virtual end address 178 * dma_inv_range(start, end) 181 * May not write back any entries. If 'start' or 'end' 186 * - end - virtual end address 202 * dma_clean_range(start, end) 207 * - end - virtual end address 219 * dma_flush_range(start, end) 224 * - end - virtual end address
|
/linux-4.4.14/tools/perf/util/ |
H A D | svghelper.h | 6 extern void open_svg(const char *filename, int cpus, int rows, u64 start, u64 end); 7 extern void svg_ubox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int merges); 8 extern void svg_lbox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int merges); 9 extern void svg_fbox(int Yslot, u64 start, u64 end, double height, const char *type, int fd, int err, int merges); 10 extern void svg_box(int Yslot, u64 start, u64 end, const char *type); 11 extern void svg_blocked(int Yslot, int cpu, u64 start, u64 end, const char *backtrace); 12 extern void svg_running(int Yslot, int cpu, u64 start, u64 end, const char *backtrace); 13 extern void svg_waiting(int Yslot, int cpu, u64 start, u64 end, const char *backtrace); 17 extern void svg_process(int cpu, u64 start, u64 end, int pid, const char *name, const char *backtrace); 18 extern void svg_cstate(int cpu, u64 start, u64 end, int type); 19 extern void svg_pstate(int cpu, u64 start, u64 end, u64 freq);
|
H A D | find-vdso-map.c | 1 static int find_vdso_map(void **start, void **end) find_vdso_map() argument 18 start, end, &m)) find_vdso_map()
|
/linux-4.4.14/arch/nios2/mm/ |
H A D | cacheflush.c | 18 static void __flush_dcache(unsigned long start, unsigned long end) __flush_dcache() argument 23 end += (cpuinfo.dcache_line_size - 1); __flush_dcache() 24 end &= ~(cpuinfo.dcache_line_size - 1); __flush_dcache() 26 if (end > start + cpuinfo.dcache_size) __flush_dcache() 27 end = start + cpuinfo.dcache_size; __flush_dcache() 29 for (addr = start; addr < end; addr += cpuinfo.dcache_line_size) { __flush_dcache() 37 static void __invalidate_dcache(unsigned long start, unsigned long end) __invalidate_dcache() argument 42 end += (cpuinfo.dcache_line_size - 1); __invalidate_dcache() 43 end &= ~(cpuinfo.dcache_line_size - 1); __invalidate_dcache() 45 for (addr = start; addr < end; addr += cpuinfo.dcache_line_size) { __invalidate_dcache() 53 static void __flush_icache(unsigned long start, unsigned long end) __flush_icache() argument 58 end += (cpuinfo.icache_line_size - 1); __flush_icache() 59 end &= ~(cpuinfo.icache_line_size - 1); __flush_icache() 61 if (end > start + cpuinfo.icache_size) __flush_icache() 62 end = start + cpuinfo.icache_size; __flush_icache() 64 for (addr = start; addr < end; addr += cpuinfo.icache_line_size) { __flush_icache() 113 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument 115 __flush_dcache(start, end); flush_icache_range() 116 __flush_icache(start, end); flush_icache_range() 119 void flush_dcache_range(unsigned long start, unsigned long end) flush_dcache_range() argument 121 __flush_dcache(start, end); flush_dcache_range() 122 __flush_icache(start, end); flush_dcache_range() 126 void invalidate_dcache_range(unsigned long start, unsigned long end) invalidate_dcache_range() argument 128 __invalidate_dcache(start, end); invalidate_dcache_range() 133 unsigned long end) flush_cache_range() 135 __flush_dcache(start, end); flush_cache_range() 137 __flush_icache(start, end); flush_cache_range() 143 unsigned long end = start + PAGE_SIZE; flush_icache_page() local 145 __flush_dcache(start, end); flush_icache_page() 146 __flush_icache(start, end); flush_icache_page() 153 unsigned long end = start + PAGE_SIZE; flush_cache_page() local 155 __flush_dcache(start, end); flush_cache_page() 157 __flush_icache(start, end); flush_cache_page() 132 flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_cache_range() argument
|
H A D | ioremap.c | 26 unsigned long end; remap_area_pte() local 32 end = address + size; remap_area_pte() 33 if (end > PMD_SIZE) remap_area_pte() 34 end = PMD_SIZE; remap_area_pte() 35 if (address >= end) remap_area_pte() 47 } while (address && (address < end)); remap_area_pte() 54 unsigned long end; remap_area_pmd() local 57 end = address + size; remap_area_pmd() 58 if (end > PGDIR_SIZE) remap_area_pmd() 59 end = PGDIR_SIZE; remap_area_pmd() 61 if (address >= end) remap_area_pmd() 68 remap_area_pte(pte, address, end - address, address + phys_addr, remap_area_pmd() 72 } while (address && (address < end)); remap_area_pmd() 81 unsigned long end = address + size; remap_area_pages() local 86 if (address >= end) remap_area_pages() 99 if (remap_area_pmd(pmd, address, end - address, remap_area_pages() 105 } while (address && (address < end)); remap_area_pages()
|
/linux-4.4.14/arch/mips/include/asm/ |
H A D | tlbflush.h | 12 * - flush_tlb_range(vma, start, end) flushes a range of pages 13 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages 18 unsigned long start, unsigned long end); 20 unsigned long end); 39 #define flush_tlb_range(vma, vmaddr, end) local_flush_tlb_range(vma, vmaddr, end) 40 #define flush_tlb_kernel_range(vmaddr,end) \ 41 local_flush_tlb_kernel_range(vmaddr, end)
|
/linux-4.4.14/arch/nios2/include/asm/ |
H A D | cacheflush.h | 27 unsigned long end); 33 extern void flush_icache_range(unsigned long start, unsigned long end); 36 #define flush_cache_vmap(start, end) flush_dcache_range(start, end) 37 #define flush_cache_vunmap(start, end) flush_dcache_range(start, end) 46 extern void flush_dcache_range(unsigned long start, unsigned long end); 47 extern void invalidate_dcache_range(unsigned long start, unsigned long end);
|
/linux-4.4.14/arch/cris/mm/ |
H A D | init.c | 38 void free_init_pages(const char *what, unsigned long begin, unsigned long end) free_init_pages() argument 42 for (addr = begin; addr < end; addr += PAGE_SIZE) { free_init_pages() 49 printk(KERN_INFO "Freeing %s: %ldk freed\n", what, (end - begin) >> 10); free_init_pages() 62 void free_initrd_mem(unsigned long start, unsigned long end) free_initrd_mem() argument 66 end); free_initrd_mem()
|
/linux-4.4.14/arch/arm/mach-imx/devices/ |
H A D | platform-imx-dma.c | 17 .end = iobase + SZ_4K - 1, imx_add_imx_dma() 21 .end = irq, imx_add_imx_dma() 25 .end = irq_err, imx_add_imx_dma() 40 .end = iobase + SZ_16K - 1, imx_add_imx_sdma() 44 .end = irq, imx_add_imx_sdma()
|
H A D | platform-gpio-mxc.c | 17 .end = iobase + iosize - 1, mxc_register_gpio() 21 .end = irq, mxc_register_gpio() 25 .end = irq_high, mxc_register_gpio()
|
H A D | platform-mx2-camera.c | 42 .end = data->iobasecsi + data->iosizecsi - 1, imx_add_mx2_camera() 46 .end = data->irqcsi, imx_add_mx2_camera() 50 .end = data->iobaseemmaprp + data->iosizeemmaprp - 1, imx_add_mx2_camera() 54 .end = data->irqemmaprp, imx_add_mx2_camera()
|
/linux-4.4.14/arch/arm/mach-iop13xx/ |
H A D | tpmi.c | 46 .end = IOP13XX_TPMI_MMR(4) + IOP13XX_TPMI_MMR_SIZE, 51 .end = IOP13XX_TPMI_MEM(0) + IOP13XX_TPMI_MEM_SIZE, 56 .end = IOP13XX_TPMI_CTRL(0) + IOP13XX_TPMI_MEM_CTRL, 61 .end = IOP13XX_TPMI_IOP_CTRL(0) + IOP13XX_TPMI_MEM_CTRL, 66 .end = IRQ_IOP13XX_TPMI0_OUT, 74 .end = IOP13XX_TPMI_MMR(1) + IOP13XX_TPMI_MMR_SIZE, 79 .end = IOP13XX_TPMI_MEM(1) + IOP13XX_TPMI_MEM_SIZE, 84 .end = IOP13XX_TPMI_CTRL(1) + IOP13XX_TPMI_MEM_CTRL, 89 .end = IOP13XX_TPMI_IOP_CTRL(1) + IOP13XX_TPMI_MEM_CTRL, 94 .end = IRQ_IOP13XX_TPMI1_OUT, 102 .end = IOP13XX_TPMI_MMR(2) + IOP13XX_TPMI_MMR_SIZE, 107 .end = IOP13XX_TPMI_MEM(2) + IOP13XX_TPMI_MEM_SIZE, 112 .end = IOP13XX_TPMI_CTRL(2) + IOP13XX_TPMI_MEM_CTRL, 117 .end = IOP13XX_TPMI_IOP_CTRL(2) + IOP13XX_TPMI_MEM_CTRL, 122 .end = IRQ_IOP13XX_TPMI2_OUT, 130 .end = IOP13XX_TPMI_MMR(3) + IOP13XX_TPMI_MMR_SIZE, 135 .end = IOP13XX_TPMI_MEM(3) + IOP13XX_TPMI_MEM_SIZE, 140 .end = IOP13XX_TPMI_CTRL(3) + IOP13XX_TPMI_MEM_CTRL, 145 .end = IOP13XX_TPMI_IOP_CTRL(3) + IOP13XX_TPMI_MEM_CTRL, 150 .end = IRQ_IOP13XX_TPMI3_OUT,
|
H A D | setup.c | 51 .end = IOP13XX_UART0_PHYS + 0x3f, 56 .end = IRQ_IOP13XX_UART0, 64 .end = IOP13XX_UART1_PHYS + 0x3f, 69 .end = IRQ_IOP13XX_UART1, 120 .end = IOP13XX_I2C0_PHYS + 0x18, 125 .end = IRQ_IOP13XX_I2C_0, 133 .end = IOP13XX_I2C1_PHYS + 0x18, 138 .end = IRQ_IOP13XX_I2C_1, 146 .end = IOP13XX_I2C2_PHYS + 0x18, 151 .end = IRQ_IOP13XX_I2C_2, 191 .end = 0, 237 .end = IOP13XX_ADMA_UPPER_PA(0), 242 .end = IRQ_IOP13XX_ADMA0_EOT, 247 .end = IRQ_IOP13XX_ADMA0_EOC, 252 .end = IRQ_IOP13XX_ADMA0_ERR, 260 .end = IOP13XX_ADMA_UPPER_PA(1), 265 .end = IRQ_IOP13XX_ADMA1_EOT, 270 .end = IRQ_IOP13XX_ADMA1_EOC, 275 .end = IRQ_IOP13XX_ADMA1_ERR, 283 .end = IOP13XX_ADMA_UPPER_PA(2), 288 .end = IRQ_IOP13XX_ADMA2_EOT, 293 .end = IRQ_IOP13XX_ADMA2_EOC, 298 .end = IRQ_IOP13XX_ADMA2_ERR, 500 iq8134x_flash_resource.end = iq8134x_flash_resource.start + iop13xx_platform_init() 502 if (iq8134x_flash_resource.end > iq8134x_flash_resource.start) iop13xx_platform_init()
|
/linux-4.4.14/drivers/pci/hotplug/ |
H A D | rpadlpar_sysfs.c | 36 char *end; add_slot_store() local 44 end = strchr(drc_name, '\n'); add_slot_store() 45 if (!end) add_slot_store() 46 end = &drc_name[nbytes]; add_slot_store() 47 *end = '\0'; add_slot_store() 68 char *end; remove_slot_store() local 75 end = strchr(drc_name, '\n'); remove_slot_store() 76 if (!end) remove_slot_store() 77 end = &drc_name[nbytes]; remove_slot_store() 78 *end = '\0'; remove_slot_store()
|
/linux-4.4.14/arch/x86/include/asm/trace/ |
H A D | mpx.h | 69 unsigned long end), 70 TP_ARGS(start, end), 74 __field(unsigned long, end) 79 __entry->end = end; 84 (void *)__entry->end 89 TP_PROTO(unsigned long start, unsigned long end), 90 TP_ARGS(start, end) 94 TP_PROTO(unsigned long start, unsigned long end), 95 TP_ARGS(start, end)
|
/linux-4.4.14/arch/x86/mm/ |
H A D | kasan_init_64.c | 18 unsigned long end; map_range() local 21 end = (unsigned long)kasan_mem_to_shadow(pfn_to_kaddr(range->end)); map_range() 24 * end + 1 here is intentional. We check several shadow bytes in advance map_range() 28 return vmemmap_populate(start, end + 1, NUMA_NO_NODE); map_range() 32 unsigned long end) clear_pgds() 34 for (; start < end; start += PGDIR_SIZE) clear_pgds() 42 unsigned long end = KASAN_SHADOW_END; kasan_map_early_shadow() local 44 for (i = pgd_index(start); start < end; i++) { kasan_map_early_shadow() 106 if (pfn_mapped[i].end == 0) kasan_init() 31 clear_pgds(unsigned long start, unsigned long end) clear_pgds() argument
|
H A D | numa_emulation.c | 31 static u64 __init mem_hole_size(u64 start, u64 end) mem_hole_size() argument 34 unsigned long end_pfn = PFN_DOWN(end); mem_hole_size() 42 * Sets up nid to range from @start to @end. The return value is -errno if 59 eb->end = pb->start + size; emu_setup_memblk() 66 if (pb->start >= pb->end) { emu_setup_memblk() 67 WARN_ON_ONCE(pb->start > pb->end); emu_setup_memblk() 72 nid, eb->start, eb->end - 1, (eb->end - eb->start) >> 20); emu_setup_memblk() 129 u64 start, limit, end; for_each_node_mask() local 138 limit = pi->blk[phys_blk].end; for_each_node_mask() 139 end = start + size; for_each_node_mask() 142 end += FAKE_NODE_MIN_SIZE; for_each_node_mask() 148 while (end - start - mem_hole_size(start, end) < size) { for_each_node_mask() 149 end += FAKE_NODE_MIN_SIZE; for_each_node_mask() 150 if (end > limit) { for_each_node_mask() 151 end = limit; for_each_node_mask() 161 if (end < dma32_end && dma32_end - end - for_each_node_mask() 162 mem_hole_size(end, dma32_end) < FAKE_NODE_MIN_SIZE) for_each_node_mask() 163 end = dma32_end; for_each_node_mask() 167 * next node, this one must extend to the end of the for_each_node_mask() 170 if (limit - end - mem_hole_size(end, limit) < size) for_each_node_mask() 171 end = limit; for_each_node_mask() 175 min(end, limit) - start); for_each_node_mask() 184 * Returns the end address of a node so that there is at least `size' amount of 189 u64 end = start + size; find_end_of_node() local 191 while (end - start - mem_hole_size(start, end) < size) { find_end_of_node() 192 end += FAKE_NODE_MIN_SIZE; find_end_of_node() 193 if (end > max_addr) { find_end_of_node() 194 end = max_addr; find_end_of_node() 198 return end; find_end_of_node() 244 u64 start, limit, end; for_each_node_mask() local 253 limit = pi->blk[phys_blk].end; for_each_node_mask() 255 end = find_end_of_node(start, limit, size); for_each_node_mask() 261 if (end < dma32_end && dma32_end - end - for_each_node_mask() 262 mem_hole_size(end, dma32_end) < FAKE_NODE_MIN_SIZE) for_each_node_mask() 263 end = dma32_end; for_each_node_mask() 267 * next node, this one must extend to the end of the for_each_node_mask() 270 if (limit - end - mem_hole_size(end, limit) < size) for_each_node_mask() 271 end = limit; for_each_node_mask() 275 min(end, limit) - start); for_each_node_mask()
|
H A D | pat_rbtree.c | 39 static int is_node_overlap(struct memtype *node, u64 start, u64 end) is_node_overlap() argument 41 if (node->start >= end || node->end <= start) is_node_overlap() 59 u64 max_end = data->end, child_max_end; compute_subtree_max_end() 77 u64 start, u64 end) memtype_rb_lowest_match() 88 } else if (is_node_overlap(data, start, end)) { memtype_rb_lowest_match() 102 u64 start, u64 end) memtype_rb_exact_match() 106 match = memtype_rb_lowest_match(root, start, end); memtype_rb_exact_match() 107 while (match != NULL && match->start < end) { memtype_rb_exact_match() 110 if (match->start == start && match->end == end) memtype_rb_exact_match() 124 u64 start, u64 end, memtype_rb_check_conflict() 132 match = memtype_rb_lowest_match(&memtype_rbroot, start, end); memtype_rb_check_conflict() 139 dprintk("Overlap at 0x%Lx-0x%Lx\n", match->start, match->end); memtype_rb_check_conflict() 146 if (match->start >= end) /* Checked all possible matches */ memtype_rb_check_conflict() 149 if (is_node_overlap(match, start, end) && memtype_rb_check_conflict() 164 current->comm, current->pid, start, end, memtype_rb_check_conflict() 178 if (data->subtree_max_end < newdata->end) memtype_rb_insert() 179 data->subtree_max_end = newdata->end; memtype_rb_insert() 186 newdata->subtree_max_end = newdata->end; memtype_rb_insert() 196 err = memtype_rb_check_conflict(&memtype_rbroot, new->start, new->end, rbt_memtype_check_insert() 203 new->subtree_max_end = new->end; rbt_memtype_check_insert() 209 struct memtype *rbt_memtype_erase(u64 start, u64 end) rbt_memtype_erase() argument 213 data = memtype_rb_exact_match(&memtype_rbroot, start, end); rbt_memtype_erase() 76 memtype_rb_lowest_match(struct rb_root *root, u64 start, u64 end) memtype_rb_lowest_match() argument 101 memtype_rb_exact_match(struct rb_root *root, u64 start, u64 end) memtype_rb_exact_match() argument 123 memtype_rb_check_conflict(struct rb_root *root, u64 start, u64 end, enum page_cache_mode reqtype, enum page_cache_mode *newtype) memtype_rb_check_conflict() argument
|
H A D | init_64.c | 60 unsigned long addr, unsigned long end) ident_pmd_init() 63 for (; addr < end; addr += PMD_SIZE) { ident_pmd_init() 71 unsigned long addr, unsigned long end) ident_pud_init() 75 for (; addr < end; addr = next) { ident_pud_init() 80 if (next > end) ident_pud_init() 81 next = end; ident_pud_init() 99 unsigned long addr, unsigned long end) kernel_ident_mapping_init() 105 for (; addr < end; addr = next) { kernel_ident_mapping_init() 110 if (next > end) kernel_ident_mapping_init() 111 next = end; kernel_ident_mapping_init() 166 void sync_global_pgds(unsigned long start, unsigned long end, int removed) sync_global_pgds() argument 170 for (address = start; address <= end; address += PGDIR_SIZE) { sync_global_pgds() 382 unsigned long end = roundup((unsigned long)_brk_end, PMD_SIZE) - 1; cleanup_highmap() local 396 if (vaddr < (unsigned long) _text || vaddr > end) cleanup_highmap() 402 phys_pte_init(pte_t *pte_page, unsigned long addr, unsigned long end, phys_pte_init() argument 406 unsigned long last_map_addr = end; phys_pte_init() 413 if (addr >= end) { phys_pte_init() 447 phys_pmd_init(pmd_t *pmd_page, unsigned long address, unsigned long end, phys_pmd_init() argument 451 unsigned long last_map_addr = end; phys_pmd_init() 461 if (address >= end) { phys_pmd_init() 474 end, prot); phys_pmd_init() 511 last_map_addr = phys_pte_init(pte, address, end, new_prot); phys_pmd_init() 522 phys_pud_init(pud_t *pud_page, unsigned long addr, unsigned long end, phys_pud_init() argument 526 unsigned long last_map_addr = end; phys_pud_init() 535 if (addr >= end) { phys_pud_init() 546 last_map_addr = phys_pmd_init(pmd, addr, end, phys_pud_init() 584 last_map_addr = phys_pmd_init(pmd, addr, end, page_size_mask, phys_pud_init() 600 unsigned long end, kernel_physical_mapping_init() 604 unsigned long next, last_map_addr = end; kernel_physical_mapping_init() 608 end = (unsigned long)__va(end); kernel_physical_mapping_init() 611 for (; start < end; start = next) { kernel_physical_mapping_init() 620 __pa(end), page_size_mask); kernel_physical_mapping_init() 625 last_map_addr = phys_pud_init(pud, __pa(start), __pa(end), kernel_physical_mapping_init() 635 sync_global_pgds(addr, end - 1, 0); kernel_physical_mapping_init() 791 remove_pte_table(pte_t *pte_start, unsigned long addr, unsigned long end, remove_pte_table() argument 800 for (; addr < end; addr = next, pte++) { remove_pte_table() 802 if (next > end) remove_pte_table() 803 next = end; remove_pte_table() 863 remove_pmd_table(pmd_t *pmd_start, unsigned long addr, unsigned long end, remove_pmd_table() argument 872 for (; addr < end; addr = next, pmd++) { remove_pmd_table() 873 next = pmd_addr_end(addr, end); remove_pmd_table() 919 remove_pud_table(pud_t *pud_start, unsigned long addr, unsigned long end, remove_pud_table() argument 928 for (; addr < end; addr = next, pud++) { remove_pud_table() 929 next = pud_addr_end(addr, end); remove_pud_table() 973 /* start and end are both virtual address. */ 975 remove_pagetable(unsigned long start, unsigned long end, bool direct) remove_pagetable() argument 983 for (addr = start; addr < end; addr = next) { remove_pagetable() 984 next = pgd_addr_end(addr, end); remove_pagetable() 997 sync_global_pgds(start, end - 1, 1); remove_pagetable() 1002 void __ref vmemmap_free(unsigned long start, unsigned long end) vmemmap_free() argument 1004 remove_pagetable(start, end, false); vmemmap_free() 1009 kernel_physical_mapping_remove(unsigned long start, unsigned long end) kernel_physical_mapping_remove() argument 1012 end = (unsigned long)__va(end); kernel_physical_mapping_remove() 1014 remove_pagetable(start, end, true); kernel_physical_mapping_remove() 1074 unsigned long end = PFN_ALIGN(__stop___ex_table); set_kernel_text_rw() local 1080 start, end); set_kernel_text_rw() 1087 set_memory_rw(start, (end - start) >> PAGE_SHIFT); set_kernel_text_rw() 1093 unsigned long end = PFN_ALIGN(__stop___ex_table); set_kernel_text_ro() local 1099 start, end); set_kernel_text_ro() 1104 set_memory_ro(start, (end - start) >> PAGE_SHIFT); set_kernel_text_ro() 1111 unsigned long end = (unsigned long) &__end_rodata_hpage_align; mark_rodata_ro() local 1117 (end - start) >> 10); mark_rodata_ro() 1118 set_memory_ro(start, (end - start) >> PAGE_SHIFT); mark_rodata_ro() 1128 * split the PMD and the reminder between _brk_end and the end mark_rodata_ro() 1140 printk(KERN_INFO "Testing CPA: undo %lx-%lx\n", start, end); mark_rodata_ro() 1141 set_memory_rw(start, (end-start) >> PAGE_SHIFT); mark_rodata_ro() 1144 set_memory_ro(start, (end-start) >> PAGE_SHIFT); mark_rodata_ro() 1239 unsigned long end, int node) vmemmap_populate_hugepages() 1247 for (addr = start; addr < end; addr = next) { vmemmap_populate_hugepages() 1248 next = pmd_addr_end(addr, end); vmemmap_populate_hugepages() 1295 int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node) vmemmap_populate() argument 1300 err = vmemmap_populate_hugepages(start, end, node); vmemmap_populate() 1302 err = vmemmap_populate_basepages(start, end, node); vmemmap_populate() 1304 sync_global_pgds(start, end - 1, 0); vmemmap_populate() 1313 unsigned long end = (unsigned long)(start_page + size); register_page_bootmem_memmap() local 1321 for (; addr < end; addr = next) { register_page_bootmem_memmap() 1352 next = pmd_addr_end(addr, end); register_page_bootmem_memmap() 59 ident_pmd_init(unsigned long pmd_flag, pmd_t *pmd_page, unsigned long addr, unsigned long end) ident_pmd_init() argument 70 ident_pud_init(struct x86_mapping_info *info, pud_t *pud_page, unsigned long addr, unsigned long end) ident_pud_init() argument 98 kernel_ident_mapping_init(struct x86_mapping_info *info, pgd_t *pgd_page, unsigned long addr, unsigned long end) kernel_ident_mapping_init() argument 599 kernel_physical_mapping_init(unsigned long start, unsigned long end, unsigned long page_size_mask) kernel_physical_mapping_init() argument 1238 vmemmap_populate_hugepages(unsigned long start, unsigned long end, int node) vmemmap_populate_hugepages() argument
|
H A D | init.c | 145 unsigned long end; member in struct:map_range 197 mr[nr_range].end = end_pfn<<PAGE_SHIFT; save_mr() 218 unsigned long end = round_up(mr[i].end, PMD_SIZE); adjust_range_page_size_mask() local 221 if ((end >> PAGE_SHIFT) > max_low_pfn) adjust_range_page_size_mask() 225 if (memblock_is_region_memory(start, end - start)) adjust_range_page_size_mask() 231 unsigned long end = round_up(mr[i].end, PUD_SIZE); adjust_range_page_size_mask() local 233 if (memblock_is_region_memory(start, end - start)) adjust_range_page_size_mask() 266 unsigned long end) split_mem_range() 272 limit_pfn = PFN_DOWN(end); split_mem_range() 345 if (mr[i].end != mr[i+1].start || split_mem_range() 358 mr[i].start, mr[i].end - 1, split_mem_range() 386 (end_pfn <= pfn_mapped[i].end)) pfn_range_is_mapped() 398 unsigned long end) init_memory_mapping() 405 start, end - 1); init_memory_mapping() 408 nr_range = split_mem_range(mr, 0, start, end); init_memory_mapping() 411 ret = kernel_physical_mapping_init(mr[i].start, mr[i].end, init_memory_mapping() 442 u64 end = clamp_val(PFN_PHYS(end_pfn), r_start, r_end); init_range_memory_mapping() local 443 if (start >= end) init_range_memory_mapping() 451 min(end, (u64)pgt_buf_top<<PAGE_SHIFT); init_range_memory_mapping() 452 init_memory_mapping(start, end); init_range_memory_mapping() 453 mapped_ram_size += end - start; init_range_memory_mapping() 482 * @map_end: end address of the target memory range 486 * will be allocated at the end of the memory, and we map the 497 /* xen has big range in reserved near end of ram, skip it at first.*/ memory_map_top_down() 508 * We start from the top (end of memory) and go to the bottom. memory_map_top_down() 510 * end of RAM in [min_pfn_mapped, max_pfn_mapped) used as new pages memory_map_top_down() 535 * @map_end: end address of the target memory range 557 * end of RAM in [min_pfn_mapped, max_pfn_mapped) used as new pages memory_map_bottom_up() 579 unsigned long end; init_mem_mapping() local 584 end = max_pfn << PAGE_SHIFT; init_mem_mapping() 586 end = max_low_pfn << PAGE_SHIFT; init_mem_mapping() 602 * [kernel_end, end) to make memory above the kernel be mapped init_mem_mapping() 606 memory_map_bottom_up(kernel_end, end); init_mem_mapping() 609 memory_map_top_down(ISA_END_ADDRESS, end); init_mem_mapping() 648 void free_init_pages(char *what, unsigned long begin, unsigned long end) free_init_pages() argument 654 end_aligned = end & PAGE_MASK; free_init_pages() 656 if (WARN_ON(begin_aligned != begin || end_aligned != end)) { free_init_pages() 658 end = end_aligned; free_init_pages() 661 if (begin >= end) free_init_pages() 671 begin, end - 1); free_init_pages() 672 set_memory_np(begin, (end - begin) >> PAGE_SHIFT); free_init_pages() 679 set_memory_nx(begin, (end - begin) >> PAGE_SHIFT); free_init_pages() 680 set_memory_rw(begin, (end - begin) >> PAGE_SHIFT); free_init_pages() 682 free_reserved_area((void *)begin, (void *)end, POISON_FREE_INITMEM, what); free_init_pages() 694 void __init free_initrd_mem(unsigned long start, unsigned long end) free_initrd_mem() argument 704 * end could be not aligned, and We can not align that, free_initrd_mem() 706 * We already reserve the end partial page before in free_initrd_mem() 712 free_init_pages("initrd", start, PAGE_ALIGN(end)); free_initrd_mem() 264 split_mem_range(struct map_range *mr, int nr_range, unsigned long start, unsigned long end) split_mem_range() argument 397 init_memory_mapping(unsigned long start, unsigned long end) init_memory_mapping() argument
|
H A D | pat_internal.h | 11 u64 end; member in struct:memtype 33 extern struct memtype *rbt_memtype_erase(u64 start, u64 end); 40 static inline struct memtype *rbt_memtype_erase(u64 start, u64 end) rbt_memtype_erase() argument
|
/linux-4.4.14/arch/mips/ath79/ |
H A D | dev-wmac.c | 29 /* .start and .end fields are filled dynamically */ 32 /* .start and .end fields are filled dynamically */ 57 ath79_wmac_resources[0].end = AR913X_WMAC_BASE + AR913X_WMAC_SIZE - 1; ar913x_wmac_setup() 59 ath79_wmac_resources[1].end = ATH79_CPU_IRQ(2); ar913x_wmac_setup() 85 ath79_wmac_resources[0].end = AR933X_WMAC_BASE + AR933X_WMAC_SIZE - 1; ar933x_wmac_setup() 87 ath79_wmac_resources[1].end = ATH79_CPU_IRQ(2); ar933x_wmac_setup() 108 ath79_wmac_resources[0].end = AR934X_WMAC_BASE + AR934X_WMAC_SIZE - 1; ar934x_wmac_setup() 110 ath79_wmac_resources[1].end = ATH79_IP2_IRQ(1); ar934x_wmac_setup() 126 ath79_wmac_resources[0].end = QCA955X_WMAC_BASE + QCA955X_WMAC_SIZE - 1; qca955x_wmac_setup() 128 ath79_wmac_resources[1].end = ATH79_IP2_IRQ(1); qca955x_wmac_setup()
|
/linux-4.4.14/arch/hexagon/mm/ |
H A D | cache.c | 25 #define spanlines(start, end) \ 26 (((end - (start & ~(LINESIZE - 1))) >> LINEBITS) + 1) 28 void flush_dcache_range(unsigned long start, unsigned long end) flush_dcache_range() argument 30 unsigned long lines = spanlines(start, end-1); flush_dcache_range() 48 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument 50 unsigned long lines = spanlines(start, end-1); flush_icache_range() 73 void hexagon_clean_dcache_range(unsigned long start, unsigned long end) hexagon_clean_dcache_range() argument 75 unsigned long lines = spanlines(start, end-1); hexagon_clean_dcache_range() 93 void hexagon_inv_dcache_range(unsigned long start, unsigned long end) hexagon_inv_dcache_range() argument 95 unsigned long lines = spanlines(start, end-1); hexagon_inv_dcache_range()
|
/linux-4.4.14/arch/blackfin/mach-bf548/boards/ |
H A D | ezkit.c | 49 .end = 0x2C0C0000 + 0xfffff, 54 .end = IRQ_PG7, 95 .end = IRQ_EPPI0_ERR, 145 .end = IRQ_KEY, 177 .end = CNT_CONFIG + 0xff, 182 .end = IRQ_CNT, 249 .end = UART0_RBR+2, 255 .end = PORTE_FER+2, 261 .end = IRQ_UART0_TX, 266 .end = IRQ_UART0_RX, 271 .end = IRQ_UART0_ERROR, 276 .end = CH_UART0_TX, 281 .end = CH_UART0_RX, 304 .end = UART1_RBR+2, 310 .end = PORTH_FER+2, 316 .end = IRQ_UART1_TX, 321 .end = IRQ_UART1_RX, 326 .end = IRQ_UART1_ERROR, 331 .end = CH_UART1_TX, 336 .end = CH_UART1_RX, 342 .end = GPIO_PE10, 347 .end = GPIO_PE9, 375 .end = UART2_RBR+2, 381 .end = PORTB_FER+2, 387 .end = IRQ_UART2_TX, 392 .end = IRQ_UART2_RX, 397 .end = IRQ_UART2_ERROR, 402 .end = CH_UART2_TX, 407 .end = CH_UART2_RX, 430 .end = UART3_RBR+2, 436 .end = PORTB_FER+2, 442 .end = IRQ_UART3_TX, 447 .end = IRQ_UART3_RX, 452 .end = IRQ_UART3_ERROR, 457 .end = CH_UART3_TX, 462 .end = CH_UART3_RX, 468 .end = GPIO_PB3, 473 .end = GPIO_PB2, 504 .end = 0xFFC004FF, 509 .end = IRQ_UART0_RX+1, 514 .end = CH_UART0_RX+1, 529 .end = 0xFFC020FF, 534 .end = IRQ_UART1_RX+1, 539 .end = CH_UART1_RX+1, 554 .end = 0xFFC021FF, 559 .end = IRQ_UART2_RX+1, 564 .end = CH_UART2_RX+1, 579 .end = 0xFFC031FF, 584 .end = IRQ_UART3_RX+1, 589 .end = CH_UART3_RX+1, 609 .end = 0x24000000 + 0xFF, 614 .end = IRQ_PE8, 641 .end = 0xFFC040FF, 646 .end = IRQ_USB_INT0, 652 .end = IRQ_USB_DMA, 704 .end = SPORT0_MRCS3+4, 709 .end = IRQ_SPORT0_RX+1, 714 .end = IRQ_SPORT0_ERROR, 738 .end = SPORT1_MRCS3+4, 743 .end = IRQ_SPORT1_RX+1, 748 .end = IRQ_SPORT1_ERROR, 772 .end = SPORT2_MRCS3+4, 777 .end = IRQ_SPORT2_RX+1, 782 .end = IRQ_SPORT2_ERROR, 806 .end = SPORT3_MRCS3+4, 811 .end = IRQ_SPORT3_RX+1, 816 .end = IRQ_SPORT3_ERROR, 847 .end = 0xFFC02FFF, 852 .end = IRQ_CAN0_RX, 857 .end = IRQ_CAN0_TX, 862 .end = IRQ_CAN0_ERROR, 884 .end = 0xFFC037FF, 889 .end = IRQ_CAN1_RX, 894 .end = IRQ_CAN1_TX, 899 .end = IRQ_CAN1_ERROR, 920 .end = 0xFFC0386F, 925 .end = IRQ_ATAPI_ERR, 967 .end = 0xFFC03B4F, 972 .end = CH_NFC, 1038 .end = 0x21ffffff, 1110 .end = PINT0_LATCH + 3, 1115 .end = IRQ_PINT0, 1130 .end = PINT1_LATCH + 3, 1135 .end = IRQ_PINT1, 1150 .end = PINT2_LATCH + 3, 1155 .end = IRQ_PINT2, 1170 .end = PINT3_LATCH + 3, 1175 .end = IRQ_PINT3, 1190 .end = PORTA_MUX + 3, 1195 .end = IRQ_PA0, 1222 .end = PORTB_MUX + 3, 1227 .end = IRQ_PB0, 1254 .end = PORTC_MUX + 3, 1259 .end = IRQ_PC0, 1286 .end = PORTD_MUX + 3, 1291 .end = IRQ_PD0, 1318 .end = PORTE_MUX + 3, 1323 .end = IRQ_PE0, 1350 .end = PORTF_MUX + 3, 1355 .end = IRQ_PF0, 1382 .end = PORTG_MUX + 3, 1387 .end = IRQ_PG0, 1412 .end = PORTH_MUX + 3, 1417 .end = IRQ_PH0, 1442 .end = PORTI_MUX + 3, 1447 .end = IRQ_PI0, 1472 .end = PORTJ_MUX + 3, 1477 .end = IRQ_PJ0, 1557 .end = SPI0_REGBASE + 0xFF, 1562 .end = CH_SPI0, 1567 .end = IRQ_SPI0, 1576 .end = SPI1_REGBASE + 0xFF, 1581 .end = CH_SPI1, 1586 .end = IRQ_SPI1, 1696 .end = TWI0_REGBASE + 0xFF, 1701 .end = IRQ_TWI0, 1722 .end = TWI1_REGBASE + 0xFF, 1727 .end = IRQ_TWI1, 1864 .end = SPORT##x##_TCR1, \ 1869 .end = CH_SPORT##x##_RX, \ 1874 .end = CH_SPORT##x##_TX, \ 1879 .end = IRQ_SPORT##x##_ERROR, \
|
H A D | cm_bf548.c | 56 .end = IRQ_EPPI0_ERR, 106 .end = IRQ_KEY, 134 .end = UART0_RBR+2, 139 .end = IRQ_UART0_TX, 144 .end = IRQ_UART0_RX, 149 .end = IRQ_UART0_ERROR, 154 .end = CH_UART0_TX, 159 .end = CH_UART0_RX, 182 .end = UART1_RBR+2, 187 .end = IRQ_UART1_TX, 192 .end = IRQ_UART1_RX, 197 .end = IRQ_UART1_ERROR, 202 .end = CH_UART1_TX, 207 .end = CH_UART1_RX, 213 .end = GPIO_PE10, 218 .end = GPIO_PE9, 246 .end = UART2_RBR+2, 251 .end = IRQ_UART2_TX, 256 .end = IRQ_UART2_RX, 261 .end = IRQ_UART2_ERROR, 266 .end = CH_UART2_TX, 271 .end = CH_UART2_RX, 294 .end = UART3_RBR+2, 299 .end = IRQ_UART3_TX, 304 .end = IRQ_UART3_RX, 309 .end = IRQ_UART3_ERROR, 314 .end = CH_UART3_TX, 319 .end = CH_UART3_RX, 325 .end = GPIO_PB3, 330 .end = GPIO_PB2, 361 .end = 0xFFC004FF, 366 .end = IRQ_UART0_RX+1, 371 .end = CH_UART0_RX+1, 386 .end = 0xFFC020FF, 391 .end = IRQ_UART1_RX+1, 396 .end = CH_UART1_RX+1, 411 .end = 0xFFC021FF, 416 .end = IRQ_UART2_RX+1, 421 .end = CH_UART2_RX+1, 436 .end = 0xFFC031FF, 441 .end = IRQ_UART3_RX+1, 446 .end = CH_UART3_RX+1, 466 .end = 0x24000000 + 0xFF, 471 .end = IRQ_PE6, 498 .end = 0xFFC040FF, 503 .end = IRQ_USB_INT0, 509 .end = IRQ_USB_DMA, 561 .end = SPORT0_MRCS3+4, 566 .end = IRQ_SPORT0_RX+1, 571 .end = IRQ_SPORT0_ERROR, 595 .end = SPORT1_MRCS3+4, 600 .end = IRQ_SPORT1_RX+1, 605 .end = IRQ_SPORT1_ERROR, 629 .end = SPORT2_MRCS3+4, 634 .end = IRQ_SPORT2_RX+1, 639 .end = IRQ_SPORT2_ERROR, 663 .end = SPORT3_MRCS3+4, 668 .end = IRQ_SPORT3_RX+1, 673 .end = IRQ_SPORT3_ERROR, 699 .end = 0xFFC0386F, 704 .end = IRQ_ATAPI_ERR, 742 .end = 0xFFC03B4F, 747 .end = CH_NFC, 787 .end = 0xFFC02FFF, 792 .end = IRQ_CAN0_RX, 797 .end = IRQ_CAN0_TX, 802 .end = IRQ_CAN0_ERROR, 842 .end = 0x207fffff, 939 .end = SPI0_REGBASE + 0xFF, 944 .end = CH_SPI0, 949 .end = IRQ_SPI0, 958 .end = SPI1_REGBASE + 0xFF, 963 .end = CH_SPI1, 968 .end = IRQ_SPI1, 1013 .end = TWI0_REGBASE + 0xFF, 1018 .end = IRQ_TWI0, 1039 .end = TWI1_REGBASE + 0xFF, 1044 .end = IRQ_TWI1,
|
/linux-4.4.14/arch/alpha/mm/ |
H A D | numa.c | 48 printk(" memcluster %2d, usage %1lx, start %8lu, end %8lu\n", for_each_mem_cluster() 62 unsigned long start, end; setup_memory_node() local 88 end = start + cluster->numpages; for_each_mem_cluster() 90 if (start >= node_pfn_end || end <= node_pfn_start) for_each_mem_cluster() 97 printk(" memcluster %2d, usage %1lx, start %8lu, end %8lu\n", for_each_mem_cluster() 103 if (end > node_pfn_end) for_each_mem_cluster() 104 end = node_pfn_end; for_each_mem_cluster() 108 if (end > node_max_pfn) for_each_mem_cluster() 109 node_max_pfn = end; for_each_mem_cluster() 144 printk(" Detected node memory: start %8lu, end %8lu\n", 173 end = start + cluster->numpages; for_each_mem_cluster() 175 if (start >= node_max_pfn || end <= node_min_pfn) for_each_mem_cluster() 178 if (end > node_max_pfn) for_each_mem_cluster() 179 end = node_max_pfn; for_each_mem_cluster() 184 if (end > end_kernel_pfn for_each_mem_cluster() 185 && end - end_kernel_pfn >= bootmap_pages) { for_each_mem_cluster() 188 } else if (end > start_kernel_pfn) for_each_mem_cluster() 189 end = start_kernel_pfn; for_each_mem_cluster() 192 if (end - start >= bootmap_pages) { for_each_mem_cluster() 213 end = cluster->start_pfn + cluster->numpages; for_each_mem_cluster() 215 if (start >= node_max_pfn || end <= node_min_pfn) for_each_mem_cluster() 218 if (end > node_max_pfn) for_each_mem_cluster() 219 end = node_max_pfn; for_each_mem_cluster() 224 if (end > end_kernel_pfn) { for_each_mem_cluster() 231 } else if (end > start_kernel_pfn) for_each_mem_cluster() 232 end = start_kernel_pfn; for_each_mem_cluster() 235 if (start >= end) for_each_mem_cluster() 238 free_bootmem_node(NODE_DATA(nid), PFN_PHYS(start), PFN_PHYS(end) - PFN_PHYS(start)); for_each_mem_cluster() 239 printk(" freeing pages %ld:%ld\n", start, end); for_each_mem_cluster() 275 printk("initrd extends beyond end of memory " setup_memory()
|
/linux-4.4.14/arch/mips/bcm63xx/ |
H A D | dev-enet.c | 49 .end = -1, /* filled at runtime */ 54 .end = -1, /* filled at runtime */ 59 .end = -1, /* filled at runtime */ 76 .end = -1, /* filled at runtime */ 108 .end = -1, /* filled at runtime */ 139 /* start & end filled at runtime */ 173 shared_res[0].end = shared_res[0].start; register_shared() 175 shared_res[0].end += (RSET_6345_ENETDMA_SIZE) - 1; register_shared() 177 shared_res[0].end += (RSET_ENETDMA_SIZE) - 1; register_shared() 187 shared_res[1].end = shared_res[1].start; register_shared() 188 shared_res[1].end += RSET_ENETDMAC_SIZE(chan_count) - 1; register_shared() 191 shared_res[2].end = shared_res[2].start; register_shared() 192 shared_res[2].end += RSET_ENETDMAS_SIZE(chan_count) - 1; register_shared() 221 enet0_res[0].end = enet0_res[0].start; bcm63xx_enet_register() 222 enet0_res[0].end += RSET_ENET_SIZE - 1; bcm63xx_enet_register() 229 enet1_res[0].end = enet1_res[0].start; bcm63xx_enet_register() 230 enet1_res[0].end += RSET_ENET_SIZE - 1; bcm63xx_enet_register() 287 enetsw_res[0].end = enetsw_res[0].start; bcm63xx_enetsw_register() 288 enetsw_res[0].end += RSET_ENETSW_SIZE - 1; bcm63xx_enetsw_register()
|
H A D | dev-hsspi.c | 20 .end = -1, /* filled at runtime */ 42 spi_resources[0].end = spi_resources[0].start; bcm63xx_hsspi_register() 43 spi_resources[0].end += RSET_HSSPI_SIZE - 1; bcm63xx_hsspi_register()
|
H A D | dev-rng.c | 17 .end = -1, /* filled at runtime */ 35 rng_resources[0].end = rng_resources[0].start; bcm63xx_rng_register() 36 rng_resources[0].end += RSET_RNG_SIZE - 1; bcm63xx_rng_register()
|
H A D | dev-wdt.c | 17 .end = -1, /* filled at runtime */ 32 wdt_resources[0].end = wdt_resources[0].start; bcm63xx_wdt_register() 33 wdt_resources[0].end += RSET_WDT_SIZE - 1; bcm63xx_wdt_register()
|
H A D | dev-spi.c | 24 .end = -1, /* filled at runtime */ 45 spi_resources[0].end = spi_resources[0].start; bcm63xx_spi_register() 50 spi_resources[0].end += BCM_6348_RSET_SPI_SIZE - 1; bcm63xx_spi_register() 56 spi_resources[0].end += BCM_6358_RSET_SPI_SIZE - 1; bcm63xx_spi_register()
|
H A D | dev-uart.c | 16 /* start & end filled at runtime */ 27 /* start & end filled at runtime */ 63 uart0_resources[0].end = uart0_resources[0].start + bcm63xx_uart_register() 70 uart1_resources[0].end = uart1_resources[0].start + bcm63xx_uart_register()
|
H A D | dev-usb-usbd.c | 48 usbd_resources[0].end = usbd_resources[0].start + RSET_USBD_SIZE - 1; bcm63xx_usbd_register() 52 usbd_resources[1].end = usbd_resources[1].start + RSET_USBDMA_SIZE - 1; bcm63xx_usbd_register() 58 r->start = r->end = bcm63xx_get_irq_number(irq_list[i]); bcm63xx_usbd_register()
|
/linux-4.4.14/arch/arm64/mm/ |
H A D | cache.S | 30 * flush_icache_range(start,end) 37 * - end - virtual end address of region 43 * __flush_cache_user_range(start,end) 50 * - end - virtual end address of region 104 * __inval_cache_range(start, end) 106 * - end - end address of region 112 * __dma_inv_range(start, end) 114 * - end - virtual end address of region 119 tst x1, x3 // end cache line aligned? 138 * __dma_clean_range(start, end) 140 * - end - virtual end address of region 160 * __dma_flush_range(start, end) 162 * - end - virtual end address of region
|
H A D | kasan_init.c | 27 unsigned long end) kasan_early_pte_populate() 40 } while (pte++, addr = next, addr != end && pte_none(*pte)); kasan_early_pte_populate() 45 unsigned long end) kasan_early_pmd_populate() 55 next = pmd_addr_end(addr, end); kasan_early_pmd_populate() 57 } while (pmd++, addr = next, addr != end && pmd_none(*pmd)); kasan_early_pmd_populate() 62 unsigned long end) kasan_early_pud_populate() 72 next = pud_addr_end(addr, end); kasan_early_pud_populate() 74 } while (pud++, addr = next, addr != end && pud_none(*pud)); kasan_early_pud_populate() 80 unsigned long end = KASAN_SHADOW_END; kasan_map_early_shadow() local 86 next = pgd_addr_end(addr, end); kasan_map_early_shadow() 88 } while (pgd++, addr = next, addr != end); kasan_map_early_shadow() 100 unsigned long end) clear_pgds() 107 for (; start < end; start += PGDIR_SIZE) clear_pgds() 142 void *end = (void *)__phys_to_virt(reg->base + reg->size); for_each_memblock() local 144 if (start >= end) for_each_memblock() 148 * end + 1 here is intentional. We check several shadow bytes in for_each_memblock() 154 (unsigned long)kasan_mem_to_shadow(end) + 1, for_each_memblock() 26 kasan_early_pte_populate(pmd_t *pmd, unsigned long addr, unsigned long end) kasan_early_pte_populate() argument 43 kasan_early_pmd_populate(pud_t *pud, unsigned long addr, unsigned long end) kasan_early_pmd_populate() argument 60 kasan_early_pud_populate(pgd_t *pgd, unsigned long addr, unsigned long end) kasan_early_pud_populate() argument 99 clear_pgds(unsigned long start, unsigned long end) clear_pgds() argument
|
/linux-4.4.14/arch/um/include/asm/ |
H A D | tlb.h | 21 unsigned long end; member in struct:mmu_gather 30 if (tlb->end < address + PAGE_SIZE) __tlb_remove_tlb_entry() 31 tlb->end = address + PAGE_SIZE; __tlb_remove_tlb_entry() 39 tlb->end = 0; init_tlb_gather() 43 tlb->end = TASK_SIZE; init_tlb_gather() 48 tlb_gather_mmu(struct mmu_gather *tlb, struct mm_struct *mm, unsigned long start, unsigned long end) tlb_gather_mmu() argument 52 tlb->end = end; tlb_gather_mmu() 53 tlb->fullmm = !(start | (end+1)); tlb_gather_mmu() 59 unsigned long end); 64 flush_tlb_mm_range(tlb->mm, tlb->start, tlb->end); tlb_flush_mmu_tlbonly() 84 * Called at the end of the shootdown operation to free up any resources 88 tlb_finish_mmu(struct mmu_gather *tlb, unsigned long start, unsigned long end) tlb_finish_mmu() argument
|
H A D | tlbflush.h | 19 * - flush_tlb_range(vma, start, end) flushes a range of pages 25 unsigned long end); 28 extern void flush_tlb_kernel_range(unsigned long start, unsigned long end);
|
/linux-4.4.14/lib/ |
H A D | ioremap.c | 54 unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pte_range() 67 } while (pte++, addr += PAGE_SIZE, addr != end); ioremap_pte_range() 72 unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pmd_range() 82 next = pmd_addr_end(addr, end); ioremap_pmd_range() 93 } while (pmd++, addr = next, addr != end); ioremap_pmd_range() 98 unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pud_range() 108 next = pud_addr_end(addr, end); ioremap_pud_range() 119 } while (pud++, addr = next, addr != end); ioremap_pud_range() 124 unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_page_range() 131 BUG_ON(addr >= end); ioremap_page_range() 137 next = pgd_addr_end(addr, end); ioremap_page_range() 141 } while (pgd++, addr = next, addr != end); ioremap_page_range() 143 flush_cache_vmap(start, end); ioremap_page_range() 53 ioremap_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pte_range() argument 71 ioremap_pmd_range(pud_t *pud, unsigned long addr, unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pmd_range() argument 97 ioremap_pud_range(pgd_t *pgd, unsigned long addr, unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pud_range() argument 123 ioremap_page_range(unsigned long addr, unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_page_range() argument
|
H A D | vsprintf.c | 45 * @endp: A pointer to the end of the parsed string will be placed here 70 * @endp: A pointer to the end of the parsed string will be placed here 84 * @endp: A pointer to the end of the parsed string will be placed here 101 * @endp: A pointer to the end of the parsed string will be placed here 392 char *number(char *buf, char *end, unsigned long long num, number() argument 454 if (buf < end) number() 461 if (buf < end) number() 468 if (buf < end) number() 473 if (buf < end) number() 483 if (buf < end) number() 490 if (buf < end) number() 496 if (buf < end) number() 502 if (buf < end) number() 511 char *string(char *buf, char *end, const char *s, struct printf_spec spec) string() argument 522 if (buf < end) string() 528 if (buf < end) string() 533 if (buf < end) string() 541 static void widen(char *buf, char *end, unsigned len, unsigned spaces) widen() argument 544 if (buf >= end) /* nowhere to put anything */ widen() 546 size = end - buf; widen() 560 char *dentry_name(char *buf, char *end, const struct dentry *d, struct printf_spec spec, dentry_name() argument 596 if (buf < end) dentry_name() 604 widen(buf - n, end, n, spaces); dentry_name() 608 if (buf < end) dentry_name() 617 char *symbol_string(char *buf, char *end, void *ptr, symbol_string() argument 637 return string(buf, end, sym, spec); symbol_string() 643 return number(buf, end, value, spec); symbol_string() 648 char *resource_string(char *buf, char *end, struct resource *res, resource_string() argument 731 if (res->start != res->end) { resource_string() 733 p = number(p, pend, res->end, *specp); resource_string() 752 return string(buf, end, sym, spec); resource_string() 756 char *hex_string(char *buf, char *end, u8 *addr, struct printf_spec spec, hex_string() argument 769 return string(buf, end, NULL, spec); hex_string() 790 if (buf < end) hex_string() 793 if (buf < end) hex_string() 798 if (buf < end) hex_string() 808 char *bitmap_string(char *buf, char *end, unsigned long *bitmap, bitmap_string() argument 834 if (buf < end) bitmap_string() 841 buf = number(buf, end, val, spec); bitmap_string() 849 char *bitmap_list_string(char *buf, char *end, unsigned long *bitmap, bitmap_list_string() argument 868 if (buf < end) bitmap_list_string() 874 buf = number(buf, end, rbot, spec); bitmap_list_string() 876 if (buf < end) bitmap_list_string() 880 buf = number(buf, end, rtop, spec); bitmap_list_string() 889 char *mac_address_string(char *buf, char *end, u8 *addr, mac_address_string() argument 923 return string(buf, end, mac_addr, spec); mac_address_string() 1076 char *ip6_addr_string(char *buf, char *end, const u8 *addr, ip6_addr_string() argument 1086 return string(buf, end, ip6_addr, spec); ip6_addr_string() 1090 char *ip4_addr_string(char *buf, char *end, const u8 *addr, ip4_addr_string() argument 1097 return string(buf, end, ip4_addr, spec); ip4_addr_string() 1101 char *ip6_addr_string_sa(char *buf, char *end, const struct sockaddr_in6 *sa, ip6_addr_string_sa() argument 1159 return string(buf, end, ip6_addr, spec); ip6_addr_string_sa() 1163 char *ip4_addr_string_sa(char *buf, char *end, const struct sockaddr_in *sa, ip4_addr_string_sa() argument 1194 return string(buf, end, ip4_addr, spec); ip4_addr_string_sa() 1198 char *escaped_string(char *buf, char *end, u8 *addr, struct printf_spec spec, escaped_string() argument 1210 return string(buf, end, NULL, spec); /* NULL pointer */ escaped_string() 1252 buf += string_escape_mem(addr, len, buf, buf < end ? end - buf : 0, flags, NULL); escaped_string() 1258 char *uuid_string(char *buf, char *end, const u8 *addr, uuid_string() argument 1301 return string(buf, end, uuid, spec); uuid_string() 1305 char *netdev_feature_string(char *buf, char *end, const u8 *addr, netdev_feature_string() argument 1313 return number(buf, end, *(const netdev_features_t *)addr, spec); netdev_feature_string() 1317 char *address_val(char *buf, char *end, const void *addr, address_val() argument 1337 return number(buf, end, num, spec); address_val() 1341 char *clock(char *buf, char *end, struct clk *clk, struct printf_spec spec, clock() argument 1345 return string(buf, end, NULL, spec); clock() 1349 return number(buf, end, clk_get_rate(clk), spec); clock() 1354 return string(buf, end, __clk_get_name(clk), spec); clock() 1359 return number(buf, end, (unsigned long)clk, spec); clock() 1459 char *pointer(const char *fmt, char *buf, char *end, void *ptr, pointer() argument 1471 return string(buf, end, "(null)", spec); pointer() 1482 return symbol_string(buf, end, ptr, spec, fmt); pointer() 1485 return resource_string(buf, end, ptr, spec, fmt); pointer() 1487 return hex_string(buf, end, ptr, spec, fmt); pointer() 1491 return bitmap_list_string(buf, end, ptr, spec, fmt); pointer() 1493 return bitmap_string(buf, end, ptr, spec, fmt); pointer() 1499 return mac_address_string(buf, end, ptr, spec, fmt); pointer() 1511 return ip6_addr_string(buf, end, ptr, spec, fmt); pointer() 1513 return ip4_addr_string(buf, end, ptr, spec, fmt); pointer() 1523 return ip4_addr_string_sa(buf, end, &sa->v4, spec, fmt); pointer() 1525 return ip6_addr_string_sa(buf, end, &sa->v6, spec, fmt); pointer() 1527 return string(buf, end, "(invalid address)", spec); pointer() 1532 return escaped_string(buf, end, ptr, spec, fmt); pointer() 1534 return uuid_string(buf, end, ptr, spec, fmt); pointer() 1540 buf += vsnprintf(buf, end > buf ? end - buf : 0, pointer() 1554 return string(buf, end, "pK-error", spec); pointer() 1590 return netdev_feature_string(buf, end, ptr, spec); pointer() 1594 return address_val(buf, end, ptr, spec, fmt); pointer() 1596 return dentry_name(buf, end, ptr, spec, fmt); pointer() 1598 return clock(buf, end, ptr, spec, fmt); pointer() 1600 return dentry_name(buf, end, pointer() 1611 return number(buf, end, (unsigned long) ptr, spec); pointer() 1840 char *str, *end; vsnprintf() local 1849 end = buf + size; vsnprintf() 1851 /* Make sure end is always >= buf */ vsnprintf() 1852 if (end < buf) { vsnprintf() 1853 end = ((void *)-1); vsnprintf() 1854 size = end - buf; vsnprintf() 1866 if (str < end) { vsnprintf() 1867 if (copy > end - str) vsnprintf() 1868 copy = end - str; vsnprintf() 1888 if (str < end) vsnprintf() 1895 if (str < end) vsnprintf() 1899 if (str < end) vsnprintf() 1907 str = string(str, end, va_arg(args, char *), spec); vsnprintf() 1911 str = pointer(fmt, str, end, va_arg(args, void *), vsnprintf() 1918 if (str < end) vsnprintf() 1973 str = number(str, end, num, spec); vsnprintf() 1979 if (str < end) vsnprintf() 1982 end[-1] = '\0'; vsnprintf() 2143 char *str, *end; vbin_printf() local 2146 end = (char *)(bin_buf + size); vbin_printf() 2154 if (str + sizeof(type) <= end) { \ vbin_printf() 2162 if (str + sizeof(type) <= end) \ vbin_printf() 2197 if (str + len < end) vbin_printf() 2271 char *str, *end; bstr_printf() local 2278 end = buf + size; bstr_printf() 2295 /* Make sure end is always >= buf */ bstr_printf() 2296 if (end < buf) { bstr_printf() 2297 end = ((void *)-1); bstr_printf() 2298 size = end - buf; bstr_printf() 2310 if (str < end) { bstr_printf() 2311 if (copy > end - str) bstr_printf() 2312 copy = end - str; bstr_printf() 2332 if (str < end) bstr_printf() 2338 if (str < end) bstr_printf() 2342 if (str < end) bstr_printf() 2352 str = string(str, end, (char *)str_arg, spec); bstr_printf() 2357 str = pointer(fmt, str, end, get_arg(void *), spec); bstr_printf() 2363 if (str < end) bstr_printf() 2408 str = number(str, end, num, spec); bstr_printf() 2415 if (str < end) bstr_printf() 2418 end[-1] = '\0'; bstr_printf()
|
/linux-4.4.14/arch/blackfin/mach-bf527/boards/ |
H A D | cm_bf527.c | 45 .end = 0x203C0000 + 0x000fffff, 50 .end = IRQ_PF7, 79 .end = 0xffc03cff, 84 .end = IRQ_USB_INT0, 90 .end = IRQ_USB_DMA, 162 .end = NFC_DATA_RD + 2, 167 .end = CH_NFC, 187 .end = 0x20312000, 191 .end = 0x20311FFF, 195 .end = IRQ_PF4, 199 .end = 6, 232 .end = 0x20300300 + 16, 237 .end = IRQ_PF7, 256 .end = 0x203FB800 + 1, 261 .end = 0x203FB804 + 1, 266 .end = IRQ_PF9, 316 .end = 0x20300000 + 0x100, 320 .end = IRQ_PF7, 456 .end = SPI0_REGBASE + 0xFF, 461 .end = CH_SPI, 466 .end = IRQ_SPI, 511 .end = 0x201fffff, 515 .end = ARRAY_SIZE(cm_flash_gpios), 536 .end = UART0_GCTL+2, 541 .end = IRQ_UART0_TX, 546 .end = IRQ_UART0_RX, 551 .end = IRQ_UART0_ERROR, 556 .end = CH_UART0_TX, 561 .end = CH_UART0_RX, 584 .end = UART1_GCTL+2, 589 .end = IRQ_UART1_TX, 594 .end = IRQ_UART1_RX, 599 .end = IRQ_UART1_ERROR, 604 .end = CH_UART1_TX, 609 .end = CH_UART1_RX, 615 .end = GPIO_PF9, 620 .end = GPIO_PF10, 647 .end = 0xFFC004FF, 652 .end = IRQ_UART0_RX+1, 657 .end = CH_UART0_RX+1, 673 .end = 0xFFC020FF, 678 .end = IRQ_UART1_RX+1, 683 .end = CH_UART1_RX+1, 703 .end = TWI0_REGBASE, 708 .end = IRQ_TWI, 748 .end = SPORT0_MRCS3+4, 753 .end = IRQ_SPORT0_RX+1, 758 .end = IRQ_SPORT0_ERROR, 782 .end = SPORT1_MRCS3+4, 787 .end = IRQ_SPORT1_RX+1, 792 .end = IRQ_SPORT1_ERROR,
|
H A D | ezkit.c | 50 .end = 0x203C0000 + 0x000fffff, 55 .end = IRQ_PF7, 84 .end = 0xffc03cff, 89 .end = IRQ_USB_INT0, 95 .end = IRQ_USB_DMA, 147 .end = IRQ_PPI_ERROR, 171 .end = IRQ_PPI_ERROR, 212 .end = 0x203fffff, 256 .end = NFC_DATA_RD + 2, 261 .end = CH_NFC, 281 .end = 0x20312000, 285 .end = 0x20311FFF, 289 .end = IRQ_PF4, 293 .end = 6, 326 .end = 0x20300300 + 16, 331 .end = IRQ_PF7, 350 .end = 0x203FB800 + 1, 355 .end = 0x203FB800 + 5, 360 .end = IRQ_PF9, 410 .end = 0x20300000 + 0x100, 417 .end = IRQ_PF7, 517 .end = SPORT##x##_TCR1, \ 522 .end = CH_SPORT##x##_RX, \ 527 .end = CH_SPORT##x##_TX, \ 532 .end = IRQ_SPORT##x##_ERROR, \ 670 .end = SPI0_REGBASE + 0xFF, 675 .end = CH_SPI, 680 .end = IRQ_SPI, 701 .end = UART0_GCTL+2, 706 .end = IRQ_UART0_TX, 711 .end = IRQ_UART0_RX, 716 .end = IRQ_UART0_ERROR, 721 .end = CH_UART0_TX, 726 .end = CH_UART0_RX, 749 .end = UART1_GCTL+2, 754 .end = IRQ_UART1_TX, 759 .end = IRQ_UART1_RX, 764 .end = IRQ_UART1_ERROR, 769 .end = CH_UART1_TX, 774 .end = CH_UART1_RX, 780 .end = GPIO_PF9, 785 .end = GPIO_PF10, 812 .end = 0xFFC004FF, 817 .end = IRQ_UART0_RX+1, 822 .end = CH_UART0_RX+1, 838 .end = 0xFFC020FF, 843 .end = IRQ_UART1_RX+1, 848 .end = CH_UART1_RX+1, 868 .end = TWI0_REGBASE, 873 .end = IRQ_TWI, 1007 .end = SPORT0_MRCS3+4, 1012 .end = IRQ_SPORT0_RX+1, 1017 .end = IRQ_SPORT0_ERROR, 1041 .end = SPORT1_MRCS3+4, 1046 .end = IRQ_SPORT1_RX+1, 1051 .end = IRQ_SPORT1_ERROR, 1118 .end = CNT_CONFIG + 0xff, 1123 .end = IRQ_CNT,
|
H A D | ad7160eval.c | 44 .end = 0xffc03cff, 49 .end = IRQ_USB_INT0, 54 .end = IRQ_USB_DMA, 104 .end = IRQ_PPI_ERROR, 142 .end = 0x203fffff, 182 .end = NFC_DATA_RD + 2, 187 .end = CH_NFC, 338 .end = SPI0_REGBASE + 0xFF, 343 .end = CH_SPI, 348 .end = IRQ_SPI, 369 .end = UART0_GCTL+2, 374 .end = IRQ_UART0_TX, 379 .end = IRQ_UART0_RX, 384 .end = IRQ_UART0_ERROR, 389 .end = CH_UART0_TX, 394 .end = CH_UART0_RX, 417 .end = UART1_GCTL+2, 422 .end = IRQ_UART1_TX, 427 .end = IRQ_UART1_RX, 432 .end = IRQ_UART1_ERROR, 437 .end = CH_UART1_TX, 442 .end = CH_UART1_RX, 448 .end = GPIO_PF9, 453 .end = GPIO_PF10, 480 .end = 0xFFC004FF, 485 .end = IRQ_UART0_RX+1, 490 .end = CH_UART0_RX+1, 506 .end = 0xFFC020FF, 511 .end = IRQ_UART1_RX+1, 516 .end = CH_UART1_RX+1, 566 .end = TWI0_REGBASE, 571 .end = IRQ_TWI, 602 .end = SPORT0_MRCS3+4, 607 .end = IRQ_SPORT0_RX+1, 612 .end = IRQ_SPORT0_ERROR, 636 .end = SPORT1_MRCS3+4, 641 .end = IRQ_SPORT1_RX+1, 646 .end = IRQ_SPORT1_ERROR, 692 .end = CNT_CONFIG + 0xff, 697 .end = IRQ_CNT,
|
H A D | ezbrd.c | 43 .end = 0xffc03cff, 48 .end = IRQ_USB_INT0, 54 .end = IRQ_USB_DMA, 126 .end = 0x203fffff, 170 .end = NFC_DATA_RD + 2, 175 .end = CH_NFC, 384 .end = SPI0_REGBASE + 0xFF, 389 .end = CH_SPI, 394 .end = IRQ_SPI, 415 .end = UART0_GCTL+2, 420 .end = IRQ_UART0_TX, 425 .end = IRQ_UART0_RX, 430 .end = IRQ_UART0_ERROR, 435 .end = CH_UART0_TX, 440 .end = CH_UART0_RX, 463 .end = UART1_GCTL+2, 468 .end = IRQ_UART1_TX, 473 .end = IRQ_UART1_RX, 478 .end = IRQ_UART1_ERROR, 483 .end = CH_UART1_TX, 488 .end = CH_UART1_RX, 494 .end = GPIO_PG0, 499 .end = GPIO_PF10, 526 .end = 0xFFC004FF, 531 .end = IRQ_UART0_RX+1, 536 .end = CH_UART0_RX+1, 552 .end = 0xFFC020FF, 557 .end = IRQ_UART1_RX+1, 562 .end = CH_UART1_RX+1, 582 .end = TWI0_REGBASE, 587 .end = IRQ_TWI, 622 .end = SPORT0_MRCS3+4, 627 .end = IRQ_SPORT0_RX+1, 632 .end = IRQ_SPORT0_ERROR, 656 .end = SPORT1_MRCS3+4, 661 .end = IRQ_SPORT1_RX+1, 666 .end = IRQ_SPORT1_ERROR, 745 .end = IRQ_PPI_ERROR,
|
H A D | tll6527m.c | 51 .end = 0xffc03cff, 56 .end = IRQ_USB_INT0, 61 .end = IRQ_USB_DMA, 119 .end = IRQ_PPI_ERROR, 164 .end = 0x201fffff, 168 .end = ARRAY_SIZE(tll6527m_flash_gpios), 459 .end = SPI0_REGBASE + 0xFF, 464 .end = CH_SPI, 469 .end = IRQ_SPI, 490 .end = UART0_GCTL+2, 495 .end = IRQ_UART0_TX, 500 .end = IRQ_UART0_RX, 505 .end = IRQ_UART0_ERROR, 510 .end = CH_UART0_TX, 515 .end = CH_UART0_RX, 539 .end = UART1_GCTL+2, 544 .end = IRQ_UART1_TX, 549 .end = IRQ_UART1_RX, 554 .end = IRQ_UART1_ERROR, 559 .end = CH_UART1_TX, 564 .end = CH_UART1_RX, 570 .end = GPIO_PF9, 575 .end = GPIO_PF10, 603 .end = 0xFFC004FF, 608 .end = IRQ_UART0_RX+1, 613 .end = CH_UART0_RX+1, 629 .end = 0xFFC020FF, 634 .end = IRQ_UART1_RX+1, 639 .end = CH_UART1_RX+1, 659 .end = TWI0_REGBASE, 664 .end = IRQ_TWI, 725 .end = SPORT0_MRCS3+4, 730 .end = IRQ_SPORT0_RX+1, 735 .end = IRQ_SPORT0_ERROR, 760 .end = SPORT1_MRCS3+4, 765 .end = IRQ_SPORT1_RX+1, 770 .end = IRQ_SPORT1_ERROR,
|
/linux-4.4.14/arch/blackfin/mach-bf533/boards/ |
H A D | ip0x.c | 42 .end = 0x20100000 + 1, 46 .end = 0x20100000 + 3, 50 .end = IRQ_PF15, 58 .end = 0x20200000 + 1, 62 .end = 0x20200000 + 3, 66 .end = IRQ_PF14, 150 .end = BFIN_UART_GCTL+2, 155 .end = IRQ_UART0_TX, 160 .end = IRQ_UART0_RX, 165 .end = IRQ_UART0_ERROR, 170 .end = CH_UART0_TX, 175 .end = CH_UART0_RX, 201 .end = 0xFFC004FF, 206 .end = IRQ_UART0_RX+1, 211 .end = CH_UART0_RX+1, 229 .end = 0x20300000 + 1, 233 .end = 0x20300000 + 3, 237 .end = IRQ_PF11,
|
H A D | cm_bf533.c | 110 .end = SPI0_REGBASE + 0xFF, 115 .end = CH_SPI, 120 .end = IRQ_SPI, 162 .end = 0x20200300 + 16, 166 .end = IRQ_PF0, 188 .end = 0x20308000 + 0xFF, 192 .end = IRQ_PF8, 220 .end = BFIN_UART_GCTL+2, 225 .end = IRQ_UART0_TX, 230 .end = IRQ_UART0_RX, 235 .end = IRQ_UART0_ERROR, 240 .end = CH_UART0_TX, 245 .end = CH_UART0_RX, 271 .end = 0xFFC004FF, 276 .end = IRQ_UART0_RX+1, 281 .end = CH_UART0_RX+1, 300 .end = SPORT0_MRCS3+4, 305 .end = IRQ_SPORT0_RX+1, 310 .end = IRQ_SPORT0_ERROR, 334 .end = SPORT1_MRCS3+4, 339 .end = IRQ_SPORT1_RX+1, 344 .end = IRQ_SPORT1_ERROR, 370 .end = 0x20308000, 374 .end = 0x20308004, 378 .end = IRQ_PF4, 410 .end = 0x20300000 + 0x100, 414 .end = IRQ_PF6, 450 .end = 0x201fffff,
|
H A D | H8606.c | 46 .end = 0x20300002, 51 .end = 0x20300006, 56 .end = IRQ_PF10, 83 .end = 0x20300300 + 16, 87 .end = IRQ_PROG_INTB, 91 .end = IRQ_PF7, 111 .end = 0x20300000 + 0x100, 115 .end = IRQ_PF10, 198 .end = SPI0_REGBASE + 0xFF, 203 .end = CH_SPI, 208 .end = IRQ_SPI, 237 .end = BFIN_UART_GCTL+2, 242 .end = IRQ_UART0_TX, 247 .end = IRQ_UART0_RX, 252 .end = IRQ_UART0_ERROR, 257 .end = CH_UART0_TX, 262 .end = CH_UART0_RX, 288 .end = 0xFFC004FF, 293 .end = IRQ_UART0_RX+1, 298 .end = CH_UART0_RX+1, 366 .end = 0x20300030 + 2, 371 .end = IRQ_PF9,
|
H A D | blackstamp.c | 58 .end = 0x20300300 + 16, 62 .end = IRQ_PF3, 155 .end = SPI0_REGBASE + 0xFF, 160 .end = CH_SPI, 165 .end = IRQ_SPI, 193 .end = BFIN_UART_GCTL+2, 198 .end = IRQ_UART0_TX, 203 .end = IRQ_UART0_RX, 208 .end = IRQ_UART0_ERROR, 213 .end = CH_UART0_TX, 218 .end = CH_UART0_RX, 244 .end = 0xFFC004FF, 249 .end = IRQ_UART0_RX+1, 254 .end = CH_UART0_RX+1, 273 .end = SPORT0_MRCS3+4, 278 .end = IRQ_SPORT0_RX+1, 283 .end = IRQ_SPORT0_ERROR, 307 .end = SPORT1_MRCS3+4, 312 .end = IRQ_SPORT1_RX+1, 317 .end = IRQ_SPORT1_ERROR,
|
H A D | stamp.c | 58 .end = 0x20300300 + 16, 62 .end = IRQ_PF7, 82 .end = 0x20300000 + 0x100, 89 .end = IRQ_PF10, 129 .end = 0x203fffff, 133 .end = 0x7BB07BB0, /* AMBCTL1 setting when accessing flash */ 261 .end = SPI0_REGBASE + 0xFF, 266 .end = CH_SPI, 271 .end = IRQ_SPI, 299 .end = BFIN_UART_GCTL+2, 304 .end = IRQ_UART0_TX, 309 .end = IRQ_UART0_RX, 314 .end = IRQ_UART0_ERROR, 319 .end = CH_UART0_TX, 324 .end = CH_UART0_RX, 350 .end = 0xFFC004FF, 355 .end = IRQ_UART0_RX+1, 360 .end = CH_UART0_RX+1, 379 .end = SPORT0_MRCS3+4, 384 .end = IRQ_SPORT0_RX+1, 389 .end = IRQ_SPORT0_ERROR, 413 .end = SPORT1_MRCS3+4, 418 .end = IRQ_SPORT1_RX+1, 423 .end = IRQ_SPORT1_ERROR, 449 .end = SPORT0_MRCS3+4, 454 .end = IRQ_SPORT0_TX+1, 459 .end = IRQ_SPORT0_RX+1, 464 .end = IRQ_SPORT0_ERROR, 469 .end = CH_SPORT0_TX, 474 .end = CH_SPORT0_RX, 616 .end = SPORT##x##_TCR1, \ 621 .end = CH_SPORT##x##_RX, \ 626 .end = CH_SPORT##x##_TX, \ 631 .end = IRQ_SPORT##x##_ERROR, \
|
/linux-4.4.14/arch/c6x/platforms/ |
H A D | cache.c | 133 unsigned int *end, cache_block_operation() 139 (L2_CACHE_ALIGN_CNT((unsigned int) end) cache_block_operation() 178 unsigned int *end, cache_block_operation_nowait() 184 (L2_CACHE_ALIGN_CNT((unsigned int) end) cache_block_operation_nowait() 325 void enable_caching(unsigned long start, unsigned long end) enable_caching() argument 328 unsigned int mar_e = IMCR_MAR_BASE + ((end >> 24) << 2); enable_caching() 334 void disable_caching(unsigned long start, unsigned long end) disable_caching() argument 337 unsigned int mar_e = IMCR_MAR_BASE + ((end >> 24) << 2); disable_caching() 347 void L1P_cache_block_invalidate(unsigned int start, unsigned int end) L1P_cache_block_invalidate() argument 350 (unsigned int *) end, L1P_cache_block_invalidate() 355 void L1D_cache_block_invalidate(unsigned int start, unsigned int end) L1D_cache_block_invalidate() argument 358 (unsigned int *) end, L1D_cache_block_invalidate() 362 void L1D_cache_block_writeback_invalidate(unsigned int start, unsigned int end) L1D_cache_block_writeback_invalidate() argument 365 (unsigned int *) end, L1D_cache_block_writeback_invalidate() 369 void L1D_cache_block_writeback(unsigned int start, unsigned int end) L1D_cache_block_writeback() argument 372 (unsigned int *) end, L1D_cache_block_writeback() 380 void L2_cache_block_invalidate(unsigned int start, unsigned int end) L2_cache_block_invalidate() argument 383 (unsigned int *) end, L2_cache_block_invalidate() 387 void L2_cache_block_writeback(unsigned int start, unsigned int end) L2_cache_block_writeback() argument 390 (unsigned int *) end, L2_cache_block_writeback() 394 void L2_cache_block_writeback_invalidate(unsigned int start, unsigned int end) L2_cache_block_writeback_invalidate() argument 397 (unsigned int *) end, L2_cache_block_writeback_invalidate() 401 void L2_cache_block_invalidate_nowait(unsigned int start, unsigned int end) L2_cache_block_invalidate_nowait() argument 404 (unsigned int *) end, L2_cache_block_invalidate_nowait() 408 void L2_cache_block_writeback_nowait(unsigned int start, unsigned int end) L2_cache_block_writeback_nowait() argument 411 (unsigned int *) end, L2_cache_block_writeback_nowait() 416 unsigned int end) L2_cache_block_writeback_invalidate_nowait() 419 (unsigned int *) end, L2_cache_block_writeback_invalidate_nowait() 132 cache_block_operation(unsigned int *start, unsigned int *end, unsigned int bar_reg, unsigned int wc_reg) cache_block_operation() argument 177 cache_block_operation_nowait(unsigned int *start, unsigned int *end, unsigned int bar_reg, unsigned int wc_reg) cache_block_operation_nowait() argument 415 L2_cache_block_writeback_invalidate_nowait(unsigned int start, unsigned int end) L2_cache_block_writeback_invalidate_nowait() argument
|
/linux-4.4.14/arch/mn10300/include/asm/ |
H A D | cacheflush.h | 25 extern void mn10300_local_icache_inv_range(unsigned long start, unsigned long end); 29 extern void mn10300_local_dcache_inv_range(unsigned long start, unsigned long end); 33 extern void mn10300_icache_inv_range(unsigned long start, unsigned long end); 37 extern void mn10300_dcache_inv_range(unsigned long start, unsigned long end); 42 extern void mn10300_local_dcache_flush_range(unsigned long start, unsigned long end); 46 extern void mn10300_local_dcache_flush_inv_range(unsigned long start, unsigned long end); 50 extern void mn10300_dcache_flush_range(unsigned long start, unsigned long end); 54 extern void mn10300_dcache_flush_inv_range(unsigned long start, unsigned long end); 59 #define mn10300_local_dcache_flush_range(start, end) do {} while (0) 65 #define mn10300_local_dcache_flush_inv_range(start, end) \ 66 mn10300_local_dcache_inv_range(start, end) 71 #define mn10300_dcache_flush_range(start, end) do {} while (0) 76 #define mn10300_dcache_flush_inv_range(start, end) \ 77 mn10300_dcache_inv_range((start), (end)) 84 #define mn10300_local_icache_inv_range(start, end) do {} while (0) 88 #define mn10300_local_dcache_inv_range(start, end) do {} while (0) 93 #define mn10300_local_dcache_flush_inv_range(start, end)do {} while (0) 96 #define mn10300_local_dcache_flush_range(start, end) do {} while (0) 100 #define mn10300_icache_inv_range(start, end) do {} while (0) 104 #define mn10300_dcache_inv_range(start, end) do {} while (0) 109 #define mn10300_dcache_flush_inv_range(start, end) do {} while (0) 112 #define mn10300_dcache_flush_range(start, end) do {} while (0) 122 #define flush_cache_range(mm, start, end) do {} while (0) 124 #define flush_cache_vmap(start, end) do {} while (0) 125 #define flush_cache_vunmap(start, end) do {} while (0) 136 extern void flush_icache_range(unsigned long start, unsigned long end); 143 extern void flush_icache_range(unsigned long start, unsigned long end); 145 #define flush_icache_range(start, end) do {} while (0)
|
/linux-4.4.14/arch/m32r/include/asm/ |
H A D | cacheflush.h | 13 #define flush_cache_range(vma, start, end) do { } while (0) 20 #define flush_icache_range(start, end) _flush_cache_copyback_all() 26 #define flush_icache_range(start, end) smp_flush_cache_all() 35 #define flush_cache_range(vma, start, end) do { } while (0) 41 #define flush_icache_range(start, end) _flush_cache_all() 49 #define flush_cache_range(vma, start, end) do { } while (0) 55 #define flush_icache_range(start, end) do { } while (0) 61 #define flush_cache_vmap(start, end) do { } while (0) 62 #define flush_cache_vunmap(start, end) do { } while (0)
|
H A D | tlbflush.h | 13 * - flush_tlb_range(vma, start, end) flushes a range of pages 14 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages 28 #define flush_tlb_range(vma, start, end) \ 29 local_flush_tlb_range(vma, start, end) 30 #define flush_tlb_kernel_range(start, end) local_flush_tlb_all() 35 #define flush_tlb_range(vma, start, end) do { } while (0) 47 #define flush_tlb_range(vma, start, end) \ 48 smp_flush_tlb_range(vma, start, end) 49 #define flush_tlb_kernel_range(start, end) smp_flush_tlb_all()
|
/linux-4.4.14/arch/arm/plat-iop/ |
H A D | adma.c | 55 .end = IOP3XX_DMA_UPPER_PA(0), 60 .end = IRQ_DMA0_EOT, 65 .end = IRQ_DMA0_EOC, 70 .end = IRQ_DMA0_ERR, 78 .end = IOP3XX_DMA_UPPER_PA(1), 83 .end = IRQ_DMA1_EOT, 88 .end = IRQ_DMA1_EOC, 93 .end = IRQ_DMA1_ERR, 102 .end = IOP3XX_AAU_UPPER_PA, 107 .end = IRQ_AA_EOT, 112 .end = IRQ_AA_EOC, 117 .end = IRQ_AA_ERR,
|
/linux-4.4.14/arch/arm/mach-ks8695/ |
H A D | devices.c | 43 .end = KS8695_WAN_PA + 0x00ff, 49 .end = KS8695_IRQ_WAN_RX_STATUS, 55 .end = KS8695_IRQ_WAN_TX_STATUS, 61 .end = KS8695_IRQ_WAN_LINK, 67 .end = KS8695_MISC_PA + 0x1f, 87 .end = KS8695_LAN_PA + 0x00ff, 93 .end = KS8695_IRQ_LAN_RX_STATUS, 99 .end = KS8695_IRQ_LAN_TX_STATUS, 105 .end = KS8695_SWITCH_PA + 0x4f, 125 .end = KS8695_HPNA_PA + 0x00ff, 131 .end = KS8695_IRQ_HPNA_RX_STATUS, 137 .end = KS8695_IRQ_HPNA_TX_STATUS,
|
/linux-4.4.14/arch/mips/mm/ |
H A D | ioremap.c | 22 phys_addr_t end; remap_area_pte() local 28 end = address + size; remap_area_pte() 29 if (end > PMD_SIZE) remap_area_pte() 30 end = PMD_SIZE; remap_area_pte() 31 BUG_ON(address >= end); remap_area_pte() 42 } while (address && (address < end)); remap_area_pte() 48 phys_addr_t end; remap_area_pmd() local 51 end = address + size; remap_area_pmd() 52 if (end > PGDIR_SIZE) remap_area_pmd() 53 end = PGDIR_SIZE; remap_area_pmd() 55 BUG_ON(address >= end); remap_area_pmd() 60 remap_area_pte(pte, address, end - address, address + phys_addr, flags); remap_area_pmd() 63 } while (address && (address < end)); remap_area_pmd() 72 unsigned long end = address + size; remap_area_pages() local 77 BUG_ON(address >= end); remap_area_pages() 89 if (remap_area_pmd(pmd, address, end - address, remap_area_pages() 95 } while (address && (address < end)); remap_area_pages()
|
H A D | pgtable-64.c | 18 unsigned long *p, *end; pgd_init() local 28 end = p + PTRS_PER_PGD; pgd_init() 40 } while (p != end); pgd_init() 46 unsigned long *p, *end; pmd_init() local 49 end = p + PTRS_PER_PMD; pmd_init() 61 } while (p != end); pmd_init()
|
H A D | gup.c | 37 static int gup_pte_range(pmd_t pmd, unsigned long addr, unsigned long end, gup_pte_range() argument 57 } while (ptep++, addr += PAGE_SIZE, addr != end); gup_pte_range() 71 static int gup_huge_pmd(pmd_t pmd, unsigned long addr, unsigned long end, gup_huge_pmd() argument 95 } while (addr += PAGE_SIZE, addr != end); gup_huge_pmd() 101 static int gup_pmd_range(pud_t pud, unsigned long addr, unsigned long end, gup_pmd_range() argument 111 next = pmd_addr_end(addr, end); gup_pmd_range() 132 } while (pmdp++, addr = next, addr != end); gup_pmd_range() 137 static int gup_huge_pud(pud_t pud, unsigned long addr, unsigned long end, gup_huge_pud() argument 161 } while (addr += PAGE_SIZE, addr != end); gup_huge_pud() 167 static int gup_pud_range(pgd_t pgd, unsigned long addr, unsigned long end, gup_pud_range() argument 177 next = pud_addr_end(addr, end); gup_pud_range() 187 } while (pudp++, addr = next, addr != end); gup_pud_range() 200 unsigned long addr, len, end; __get_user_pages_fast() local 209 end = start + len; __get_user_pages_fast() 236 next = pgd_addr_end(addr, end); __get_user_pages_fast() 241 } while (pgdp++, addr = next, addr != end); __get_user_pages_fast() 267 unsigned long addr, len, end; get_user_pages_fast() local 276 end = start + len; get_user_pages_fast() 277 if (end < start || cpu_has_dc_aliases) get_user_pages_fast() 286 next = pgd_addr_end(addr, end); get_user_pages_fast() 291 } while (pgdp++, addr = next, addr != end); get_user_pages_fast() 294 VM_BUG_ON(nr != (end - start) >> PAGE_SHIFT); get_user_pages_fast() 305 (end - start) >> PAGE_SHIFT, get_user_pages_fast()
|
H A D | sc-r5k.c | 27 unsigned long end = start + scache_size; blast_r5000_scache() local 29 while(start < end) { blast_r5000_scache() 37 unsigned long end, a; r5k_dma_cache_inv_sc() local 52 end = (addr + size - 1) & ~(SC_PAGE - 1); r5k_dma_cache_inv_sc() 53 while (a <= end) { r5k_dma_cache_inv_sc()
|
H A D | sc-rm7k.c | 44 unsigned long end, a; rm7k_sc_wback_inv() local 57 end = (addr + size - 1) & ~(tc_pagesize - 1); rm7k_sc_wback_inv() 60 if (a == end) rm7k_sc_wback_inv() 68 unsigned long end, a; rm7k_sc_inv() local 81 end = (addr + size - 1) & ~(tc_pagesize - 1); rm7k_sc_inv() 84 if (a == end) rm7k_sc_inv() 93 unsigned long end = start + tcache_size; blast_rm7k_tcache() local 97 while (start < end) { blast_rm7k_tcache() 188 unsigned long flags, addr, begin, end, pow2; __probe_tcache() local 192 end = begin + (8 * 1024 * 1024); __probe_tcache() 200 for (addr = begin; addr <= end; addr = (begin + pow2)) { __probe_tcache() 213 for (addr = begin + (512 * 1024); addr <= end; addr = begin + pow2) { __probe_tcache()
|
H A D | tlb-r3k.c | 82 unsigned long end) local_flush_tlb_range() 92 cpu_context(cpu, mm) & ASID_MASK, start, end); local_flush_tlb_range() 95 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_range() 101 end += PAGE_SIZE - 1; local_flush_tlb_range() 102 end &= PAGE_MASK; local_flush_tlb_range() 103 while (start < end) { local_flush_tlb_range() 124 void local_flush_tlb_kernel_range(unsigned long start, unsigned long end) local_flush_tlb_kernel_range() argument 129 printk("[tlbrange<%lu,0x%08lx,0x%08lx>]", start, end); local_flush_tlb_kernel_range() 132 size = (end - start + (PAGE_SIZE - 1)) >> PAGE_SHIFT; local_flush_tlb_kernel_range() 137 end += PAGE_SIZE - 1; local_flush_tlb_kernel_range() 138 end &= PAGE_MASK; local_flush_tlb_kernel_range() 140 while (start < end) { local_flush_tlb_kernel_range() 81 local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) local_flush_tlb_range() argument
|
/linux-4.4.14/arch/mn10300/mm/ |
H A D | cache-flush-icache.c | 44 * @end: The ending virtual address of the page part. 50 static void flush_icache_page_range(unsigned long start, unsigned long end) flush_icache_page_range() argument 61 size = end - start; flush_icache_page_range() 96 smp_cache_call(SMP_IDCACHE_INV_FLUSH_RANGE, start, end); flush_icache_page_range() 102 * @end: The ending virtual address of the region. 108 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument 115 if (end > 0x80000000UL) { flush_icache_range() 117 if (end > 0xa0000000UL) { flush_icache_range() 118 end = 0xa0000000UL; flush_icache_range() 119 if (start >= end) flush_icache_range() 127 mn10300_local_dcache_flush_range(start_page, end); flush_icache_range() 128 mn10300_local_icache_inv_range(start_page, end); flush_icache_range() 129 smp_cache_call(SMP_IDCACHE_INV_FLUSH_RANGE, start_page, end); flush_icache_range() 132 end = start_page; flush_icache_range() 136 end_page = (end - 1) & PAGE_MASK; flush_icache_range() 140 flush_icache_page_range(start, end); flush_icache_range() 144 flush_icache_page_range(end_page, end); flush_icache_range()
|
H A D | cache-inv-icache.c | 22 * @end: The ending virtual address of the page part. 28 static void flush_icache_page_range(unsigned long start, unsigned long end) flush_icache_page_range() argument 39 size = end - start; flush_icache_page_range() 72 smp_cache_call(SMP_ICACHE_INV_RANGE, start, end); flush_icache_page_range() 78 * @end: The ending virtual address of the region. 84 void flush_icache_range(unsigned long start, unsigned long end) flush_icache_range() argument 91 if (end > 0x80000000UL) { flush_icache_range() 93 if (end > 0xa0000000UL) { flush_icache_range() 94 end = 0xa0000000UL; flush_icache_range() 95 if (start >= end) flush_icache_range() 103 mn10300_icache_inv_range(start_page, end); flush_icache_range() 104 smp_cache_call(SMP_ICACHE_INV_RANGE, start, end); flush_icache_range() 107 end = start_page; flush_icache_range() 111 end_page = (end - 1) & PAGE_MASK; flush_icache_range() 115 flush_icache_page_range(start, end); flush_icache_range() 119 flush_icache_page_range(end_page, end); flush_icache_range()
|
H A D | cache-smp-flush.c | 52 * @end: The end address of the region to be flushed. 55 * end-1 inclusive. 57 void mn10300_dcache_flush_range(unsigned long start, unsigned long end) mn10300_dcache_flush_range() argument 62 mn10300_local_dcache_flush_range(start, end); mn10300_dcache_flush_range() 63 smp_cache_call(SMP_DCACHE_FLUSH_RANGE, start, end); mn10300_dcache_flush_range() 124 * @end: The end address of the region to be flushed and invalidated. 127 * between start and end-1 inclusive. 129 void mn10300_dcache_flush_inv_range(unsigned long start, unsigned long end) mn10300_dcache_flush_inv_range() argument 134 mn10300_local_dcache_flush_inv_range(start, end); mn10300_dcache_flush_inv_range() 135 smp_cache_call(SMP_DCACHE_FLUSH_INV_RANGE, start, end); mn10300_dcache_flush_inv_range()
|
H A D | cache-smp-inv.c | 52 * @end: The end address of the region to be invalidated. 55 * between start and end-1 inclusive. 57 void mn10300_icache_inv_range(unsigned long start, unsigned long end) mn10300_icache_inv_range() argument 62 mn10300_local_icache_inv_range(start, end); mn10300_icache_inv_range() 63 smp_cache_call(SMP_ICACHE_INV_RANGE, start, end); mn10300_icache_inv_range() 122 * @end: The end address of the region to be invalidated. 125 * and end-1 inclusive. 127 void mn10300_dcache_inv_range(unsigned long start, unsigned long end) mn10300_dcache_inv_range() argument 132 mn10300_local_dcache_inv_range(start, end); mn10300_dcache_inv_range() 133 smp_cache_call(SMP_DCACHE_INV_RANGE, start, end); mn10300_dcache_inv_range()
|
H A D | cache.c | 47 asmlinkage long sys_cacheflush(unsigned long start, unsigned long end) sys_cacheflush() argument 49 if (end < start) sys_cacheflush() 52 flush_icache_range(start, end); sys_cacheflush()
|
/linux-4.4.14/arch/blackfin/mach-bf609/boards/ |
H A D | ezkit.c | 47 .end = 0x2C0C0000 + 0xfffff, 52 .end = IRQ_PG7, 92 .end = CNT_CONFIG + 0xff, 97 .end = IRQ_CNT, 150 .end = EMAC0_MACCFG + 0x1274, 156 .end = IRQ_EMAC0_STAT, 218 .end = UART0_RXDIV+4, 224 .end = PORTD_FER+2, 229 .end = PORTD_MUX+3, 235 .end = IRQ_UART0_TX, 240 .end = IRQ_UART0_RX, 245 .end = IRQ_UART0_STAT, 250 .end = CH_UART0_TX, 255 .end = CH_UART0_RX, 261 .end = GPIO_PD10, 266 .end = GPIO_PD9, 294 .end = UART1_RXDIV+4, 300 .end = PORTG_FER_SET+2, 306 .end = IRQ_UART1_TX, 311 .end = IRQ_UART1_RX, 316 .end = IRQ_UART1_STAT, 321 .end = CH_UART1_TX, 326 .end = CH_UART1_RX, 332 .end = GPIO_PG13, 337 .end = GPIO_PG10, 368 .end = 0xFFC004FF, 373 .end = IRQ_UART0_TX+1, 378 .end = CH_UART0_TX+1, 393 .end = 0xFFC020FF, 398 .end = IRQ_UART1_TX+1, 403 .end = CH_UART1_TX+1, 420 .end = 0xFFCC1398, 425 .end = IRQ_USB_STAT, 431 .end = IRQ_USB_DMA, 477 .end = SPORT0_MRCS3+4, 482 .end = IRQ_SPORT0_RX+1, 487 .end = IRQ_SPORT0_ERROR, 511 .end = SPORT1_MRCS3+4, 516 .end = IRQ_SPORT1_RX+1, 521 .end = IRQ_SPORT1_ERROR, 545 .end = SPORT2_MRCS3+4, 550 .end = IRQ_SPORT2_RX+1, 555 .end = IRQ_SPORT2_ERROR, 586 .end = 0xFFC00FFF, 591 .end = IRQ_CAN0_RX, 596 .end = IRQ_CAN0_TX, 601 .end = IRQ_CAN0_STAT, 647 .end = 0xFFC03B4F, 652 .end = CH_NFC, 731 .end = 0xb0ffffff, 795 .end = SPORT0_CTL_A, 800 .end = SPORT0_CTL_B, 805 .end = CH_SPORT0_TX, 810 .end = CH_SPORT0_RX, 815 .end = IRQ_SPORT0_TX_STAT, 820 .end = IRQ_SPORT0_RX_STAT, 1215 .end = EPPI2_STAT, 1220 .end = CH_EPPI2_CH0, 1225 .end = IRQ_EPPI2_STAT, 1245 .end = REG_CRC0_REVID+4, 1250 .end = IRQ_CRC0_DCNTEXP, 1255 .end = CH_MEM_STREAM0_SRC_CRC0, 1260 .end = CH_MEM_STREAM0_DEST_CRC0, 1275 .end = REG_CRC1_REVID+4, 1280 .end = IRQ_CRC1_DCNTEXP, 1285 .end = CH_MEM_STREAM1_SRC_CRC1, 1290 .end = CH_MEM_STREAM1_DEST_CRC1, 1310 .end = REG_CRC0_REVID+4, 1315 .end = IRQ_CRC0_DCNTEXP, 1320 .end = CH_MEM_STREAM0_SRC_CRC0, 1366 .end = PINT0_LATCH + 3, 1371 .end = IRQ_PINT0, 1386 .end = PINT1_LATCH + 3, 1391 .end = IRQ_PINT1, 1406 .end = PINT2_LATCH + 3, 1411 .end = IRQ_PINT2, 1426 .end = PINT3_LATCH + 3, 1431 .end = IRQ_PINT3, 1446 .end = PINT4_LATCH + 3, 1451 .end = IRQ_PINT4, 1466 .end = PINT5_LATCH + 3, 1471 .end = IRQ_PINT5, 1486 .end = PORTA_MUX + 3, 1491 .end = IRQ_PA0, 1517 .end = PORTB_MUX + 3, 1522 .end = IRQ_PB0, 1548 .end = PORTC_MUX + 3, 1553 .end = IRQ_PC0, 1579 .end = PORTD_MUX + 3, 1584 .end = IRQ_PD0, 1610 .end = PORTE_MUX + 3, 1615 .end = IRQ_PE0, 1641 .end = PORTF_MUX + 3, 1646 .end = IRQ_PF0, 1672 .end = PORTG_MUX + 3, 1677 .end = IRQ_PG0, 1773 .end = SPI0_REGBASE + 0xFF, 1778 .end = CH_SPI0_TX, 1783 .end = CH_SPI0_RX, 1792 .end = SPI1_REGBASE + 0xFF, 1797 .end = CH_SPI1_TX, 1802 .end = CH_SPI1_RX, 1846 .end = TWI0_CLKDIV + 0xFF, 1851 .end = IRQ_TWI0, 1871 .end = TWI1_CLKDIV + 0xFF, 1876 .end = IRQ_TWI1,
|
/linux-4.4.14/arch/arm/mach-w90x900/ |
H A D | dev.c | 80 .end = NUC900_FLASH_BASE + NUC900_FLASH_SIZE - 1, 100 .end = W90X900_PA_USBEHCIHOST + W90X900_SZ_USBEHCIHOST - 1, 105 .end = IRQ_USBH, 128 .end = W90X900_PA_USBOHCIHOST + W90X900_SZ_USBOHCIHOST - 1, 133 .end = IRQ_USBH, 155 .end = W90X900_PA_USBDEV + W90X900_SZ_USBDEV - 1, 160 .end = IRQ_USBD, 177 .end = W90X900_PA_EMC + W90X900_SZ_EMC - 1, 182 .end = IRQ_EMCTX, 187 .end = IRQ_EMCRX, 221 .end = W90X900_PA_I2C + SPIOFFSET + SPIOREG_SIZE - 1, 226 .end = IRQ_SSP, 274 .end = W90X900_PA_TIMER + W90X900_SZ_TIMER - 1, 279 .end = IRQ_WDT, 305 .end = W90X900_PA_RTC + 0xff, 310 .end = IRQ_RTC, 327 .end = W90X900_PA_ADC + W90X900_SZ_ADC-1, 332 .end = IRQ_ADC, 349 .end = W90X900_PA_FMI + W90X900_SZ_FMI - 1, 354 .end = IRQ_FMI, 404 .end = W90X900_PA_KPI + W90X900_SZ_KPI - 1, 409 .end = IRQ_KPI, 467 .end = W90X900_PA_LCD + W90X900_SZ_LCD - 1, 472 .end = IRQ_LCD, 495 .end = W90X900_PA_ACTL + W90X900_SZ_ACTL - 1, 500 .end = IRQ_ACTL,
|
/linux-4.4.14/arch/blackfin/mach-bf518/boards/ |
H A D | tcm-bf518.c | 61 .end = 0x201fffff, 240 .end = SPI0_REGBASE + 0xFF, 245 .end = CH_SPI0, 250 .end = IRQ_SPI0, 275 .end = SPI1_REGBASE + 0xFF, 280 .end = CH_SPI1, 285 .end = IRQ_SPI1, 306 .end = UART0_GCTL+2, 311 .end = IRQ_UART0_TX, 316 .end = IRQ_UART0_RX, 321 .end = IRQ_UART0_ERROR, 326 .end = CH_UART0_TX, 331 .end = CH_UART0_RX, 354 .end = UART1_GCTL+2, 359 .end = IRQ_UART1_TX, 364 .end = IRQ_UART1_RX, 369 .end = IRQ_UART1_ERROR, 374 .end = CH_UART1_TX, 379 .end = CH_UART1_RX, 405 .end = 0xFFC004FF, 410 .end = IRQ_UART0_RX+1, 415 .end = CH_UART0_RX+1, 431 .end = 0xFFC020FF, 436 .end = IRQ_UART1_RX+1, 441 .end = CH_UART1_RX+1, 461 .end = TWI0_REGBASE, 466 .end = IRQ_TWI, 501 .end = SPORT0_MRCS3+4, 506 .end = IRQ_SPORT0_RX+1, 511 .end = IRQ_SPORT0_ERROR, 535 .end = SPORT1_MRCS3+4, 540 .end = IRQ_SPORT1_RX+1, 545 .end = IRQ_SPORT1_ERROR,
|
H A D | ezbrd.c | 65 .end = 0x202fffff, 67 .end = 0x203fffff, 305 .end = SPI0_REGBASE + 0xFF, 310 .end = CH_SPI0, 315 .end = IRQ_SPI0, 340 .end = SPI1_REGBASE + 0xFF, 345 .end = CH_SPI1, 350 .end = IRQ_SPI1, 371 .end = UART0_GCTL+2, 376 .end = IRQ_UART0_TX, 381 .end = IRQ_UART0_RX, 386 .end = IRQ_UART0_ERROR, 391 .end = CH_UART0_TX, 396 .end = CH_UART0_RX, 419 .end = UART1_GCTL+2, 424 .end = IRQ_UART1_TX, 429 .end = IRQ_UART1_RX, 434 .end = IRQ_UART1_ERROR, 439 .end = CH_UART1_TX, 444 .end = CH_UART1_RX, 470 .end = 0xFFC004FF, 475 .end = IRQ_UART0_RX+1, 480 .end = CH_UART0_RX+1, 496 .end = 0xFFC020FF, 501 .end = IRQ_UART1_RX+1, 506 .end = CH_UART1_RX+1, 534 .end = TWI0_REGBASE, 539 .end = IRQ_TWI, 579 .end = SPORT0_MRCS3+4, 584 .end = IRQ_SPORT0_RX+1, 589 .end = IRQ_SPORT0_ERROR, 613 .end = SPORT1_MRCS3+4, 618 .end = IRQ_SPORT1_RX+1, 623 .end = IRQ_SPORT1_ERROR,
|
/linux-4.4.14/arch/alpha/lib/ |
H A D | clear_page.S | 39 .end clear_page
|
/linux-4.4.14/include/asm-generic/ |
H A D | cacheflush.h | 14 #define flush_cache_range(vma, start, end) do { } while (0) 20 #define flush_icache_range(start, end) do { } while (0) 23 #define flush_cache_vmap(start, end) do { } while (0) 24 #define flush_cache_vunmap(start, end) do { } while (0)
|
/linux-4.4.14/arch/powerpc/perf/req-gen/ |
H A D | _begin.h | 13 #define REQUEST_END CAT2_STR(REQ_GEN_PREFIX, _request-end.h)
|
/linux-4.4.14/arch/sh/kernel/cpu/sh4a/ |
H A D | pinmux-sh7722.c | 10 .end = 0xa405018f,
|
/linux-4.4.14/arch/mips/cavium-octeon/ |
H A D | csrc-octeon.c | 140 u64 cur, end, inc; __udelay() local 145 end = cur + inc; __udelay() 147 while (end > cur) __udelay() 154 u64 cur, end, inc; __ndelay() local 159 end = cur + inc; __ndelay() 161 while (end > cur) __ndelay() 168 u64 cur, end; __delay() local 171 end = cur + loops; __delay() 173 while (end > cur) __delay() 189 u64 cur, end; octeon_io_clk_delay() local 193 end = count * rdiv; octeon_io_clk_delay() 197 : [cnt] "+r" (end) octeon_io_clk_delay() 201 end = cur + end; octeon_io_clk_delay() 203 end = cur + count; octeon_io_clk_delay() 205 while (end > cur) octeon_io_clk_delay()
|
/linux-4.4.14/arch/cris/include/asm/ |
H A D | cacheflush.h | 13 #define flush_cache_range(vma, start, end) do { } while (0) 19 #define flush_icache_range(start, end) do { } while (0) 22 #define flush_cache_vmap(start, end) do { } while (0) 23 #define flush_cache_vunmap(start, end) do { } while (0)
|
H A D | tlbflush.h | 16 * - flush_tlb_range(mm, start, end) flushes a range of pages 29 static inline void flush_tlb_range(struct vm_area_struct * vma, unsigned long start, unsigned long end) flush_tlb_range() argument 39 #define flush_tlb_kernel_range(start, end) flush_tlb_all()
|
/linux-4.4.14/arch/m32r/boot/compressed/ |
H A D | vmlinux.lds.S | 27 end = . ; define
|
/linux-4.4.14/arch/avr32/include/asm/ |
H A D | tlbflush.h | 20 * - flush_tlb_range(vma, start, end) flushes a range of pages 21 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages 27 unsigned long end); 30 extern void flush_tlb_kernel_range(unsigned long start, unsigned long end);
|
/linux-4.4.14/tools/build/feature/ |
H A D | test-libunwind-debug-frame.c | 8 unw_word_t end);
|
/linux-4.4.14/tools/include/linux/ |
H A D | list.h | 12 * @end: last element in the range to delete from the list. 17 struct list_head *end) list_del_range() 19 begin->prev->next = end->next; list_del_range() 20 end->next->prev = begin->prev; list_del_range() 16 list_del_range(struct list_head *begin, struct list_head *end) list_del_range() argument
|
/linux-4.4.14/drivers/net/ethernet/stmicro/stmmac/ |
H A D | descs_com.h | 36 static inline void ehn_desc_rx_set_on_ring(struct dma_desc *p, int end) ehn_desc_rx_set_on_ring() argument 39 if (end) ehn_desc_rx_set_on_ring() 43 static inline void ehn_desc_tx_set_on_ring(struct dma_desc *p, int end) ehn_desc_tx_set_on_ring() argument 45 if (end) ehn_desc_tx_set_on_ring() 64 static inline void ndesc_rx_set_on_ring(struct dma_desc *p, int end) ndesc_rx_set_on_ring() argument 67 if (end) ndesc_rx_set_on_ring() 71 static inline void ndesc_tx_set_on_ring(struct dma_desc *p, int end) ndesc_tx_set_on_ring() argument 73 if (end) ndesc_tx_set_on_ring() 94 static inline void ehn_desc_rx_set_on_chain(struct dma_desc *p, int end) ehn_desc_rx_set_on_chain() argument 99 static inline void ehn_desc_tx_set_on_chain(struct dma_desc *p, int end) ehn_desc_tx_set_on_chain() argument 115 static inline void ndesc_rx_set_on_chain(struct dma_desc *p, int end) ndesc_rx_set_on_chain() argument
|
/linux-4.4.14/arch/mips/jazz/ |
H A D | setup.c | 33 .end = 0x1f, 38 .end = 0x5f, 43 .end = 0x8f, 48 .end = 0xdf, 124 .end = JAZZ_SCSI_BASE + 31, 129 .end = JAZZ_SCSI_DMA, 134 .end = JAZZ_SCSI_IRQ, 148 .end = JAZZ_ETHERNET_BASE + 0xff, 153 .end = JAZZ_ETHERNET_IRQ, 167 .end = 0x71, 172 .end = 8,
|
/linux-4.4.14/arch/arc/mm/ |
H A D | ioremap.c | 19 unsigned long end; ioremap() local 22 end = paddr + size - 1; ioremap() 23 if (!size || (end < paddr)) ioremap() 46 unsigned long off, end; ioremap_prot() local 50 end = paddr + size - 1; ioremap_prot() 51 if ((!size) || (end < paddr)) ioremap_prot() 54 /* An early platform driver might end up here */ ioremap_prot() 64 size = PAGE_ALIGN(end + 1) - paddr; ioremap_prot()
|
/linux-4.4.14/arch/mips/vr41xx/common/ |
H A D | rtc.c | 32 .end = 0x0b0000df, 37 .end = 0x0b0001df, 42 .end = ELAPSEDTIME_IRQ, 47 .end = RTCLONG1_IRQ, 55 .end = 0x0f00011f, 60 .end = 0x0f00013f, 65 .end = ELAPSEDTIME_IRQ, 70 .end = RTCLONG1_IRQ,
|
H A D | giu.c | 33 .end = 0x0b00011f, 38 .end = 0x0b0002e3, 43 .end = GIUINT_IRQ, 51 .end = 0x0f00015f, 56 .end = GIUINT_IRQ, 64 .end = 0x0f000167, 69 .end = GIUINT_IRQ,
|
/linux-4.4.14/arch/arm/include/asm/hardware/ |
H A D | memc.h | 19 #define video_set_dma(start,end,offset) \ 22 memc_write (VDMA_END, (end - VDMA_XFERSIZE) >> 2); \
|
/linux-4.4.14/arch/arc/include/asm/ |
H A D | tlbflush.h | 17 void local_flush_tlb_kernel_range(unsigned long start, unsigned long end); 19 unsigned long start, unsigned long end); 21 unsigned long end); 32 unsigned long end); 34 extern void flush_tlb_kernel_range(unsigned long start, unsigned long end); 37 extern void flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end);
|
/linux-4.4.14/arch/metag/include/asm/ |
H A D | tlbflush.h | 16 * - flush_tlb_range(mm, start, end) flushes a range of pages 17 * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages 18 * - flush_tlb_pgtables(mm, start, end) flushes a range of page tables 59 unsigned long start, unsigned long end) flush_tlb_range() 65 unsigned long start, unsigned long end) flush_tlb_pgtables() 71 unsigned long end) flush_tlb_kernel_range() 58 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_tlb_range() argument 64 flush_tlb_pgtables(struct mm_struct *mm, unsigned long start, unsigned long end) flush_tlb_pgtables() argument 70 flush_tlb_kernel_range(unsigned long start, unsigned long end) flush_tlb_kernel_range() argument
|
/linux-4.4.14/arch/microblaze/include/asm/ |
H A D | tlbflush.h | 37 unsigned long start, unsigned long end) local_flush_tlb_range() 40 #define flush_tlb_kernel_range(start, end) do { } while (0) 55 unsigned long start, unsigned long end) { } 63 #define flush_tlb_range(mm, start, end) BUG() 64 #define flush_tlb_pgtables(mm, start, end) BUG() 65 #define flush_tlb_kernel_range(start, end) BUG() 36 local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) local_flush_tlb_range() argument 54 flush_tlb_pgtables(struct mm_struct *mm, unsigned long start, unsigned long end) flush_tlb_pgtables() argument
|
/linux-4.4.14/arch/frv/include/asm/ |
H A D | tlbflush.h | 25 unsigned long start, unsigned long end); 42 #define flush_tlb_range(vma,start,end) \ 45 __flush_tlb_range((vma)->vm_mm->context.id, start, end); \ 59 #define flush_tlb_kernel_range(start, end) flush_tlb_all() 67 #define flush_tlb_range(mm,start,end) BUG() 68 #define flush_tlb_kernel_range(start, end) BUG()
|
/linux-4.4.14/mm/kasan/ |
H A D | kasan_init.c | 47 unsigned long end) zero_pte_populate() 55 while (addr + PAGE_SIZE <= end) { zero_pte_populate() 63 unsigned long end) zero_pmd_populate() 69 next = pmd_addr_end(addr, end); zero_pmd_populate() 71 if (IS_ALIGNED(addr, PMD_SIZE) && end - addr >= PMD_SIZE) { zero_pmd_populate() 81 } while (pmd++, addr = next, addr != end); zero_pmd_populate() 85 unsigned long end) zero_pud_populate() 91 next = pud_addr_end(addr, end); zero_pud_populate() 92 if (IS_ALIGNED(addr, PUD_SIZE) && end - addr >= PUD_SIZE) { zero_pud_populate() 106 } while (pud++, addr = next, addr != end); zero_pud_populate() 113 * @shadow_end - end of the memory range to populate 119 unsigned long end = (unsigned long)shadow_end; kasan_populate_zero_shadow() local 124 next = pgd_addr_end(addr, end); kasan_populate_zero_shadow() 126 if (IS_ALIGNED(addr, PGDIR_SIZE) && end - addr >= PGDIR_SIZE) { kasan_populate_zero_shadow() 151 } while (pgd++, addr = next, addr != end); kasan_populate_zero_shadow() 46 zero_pte_populate(pmd_t *pmd, unsigned long addr, unsigned long end) zero_pte_populate() argument 62 zero_pmd_populate(pud_t *pud, unsigned long addr, unsigned long end) zero_pmd_populate() argument 84 zero_pud_populate(pgd_t *pgd, unsigned long addr, unsigned long end) zero_pud_populate() argument
|
/linux-4.4.14/arch/arm64/include/asm/ |
H A D | cacheflush.h | 36 * Start addresses are inclusive and end addresses are exclusive; start 37 * addresses should be rounded down, end addresses up. 48 * flush_icache_range(start, end) 51 * region described by start, end. 53 * - end - virtual end address 55 * __flush_cache_user_range(start, end) 58 * region described by start, end. 60 * - end - virtual end address 68 extern void flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end); 69 extern void flush_icache_range(unsigned long start, unsigned long end); 71 extern long __flush_cache_user_range(unsigned long start, unsigned long end); 145 static inline void flush_cache_vmap(unsigned long start, unsigned long end) flush_cache_vmap() argument 149 static inline void flush_cache_vunmap(unsigned long start, unsigned long end) flush_cache_vunmap() argument
|
H A D | tlbflush.h | 33 * addresses are inclusive and end addresses are exclusive; it is safe to 45 * flush_tlb_range(mm,start,end) 51 * - end - end address (exclusive, may not be aligned) 108 unsigned long start, unsigned long end, __flush_tlb_range() 114 if ((end - start) > MAX_TLB_RANGE) { __flush_tlb_range() 120 end = asid | (end >> 12); __flush_tlb_range() 123 for (addr = start; addr < end; addr += 1 << (PAGE_SHIFT - 12)) { __flush_tlb_range() 133 unsigned long start, unsigned long end) flush_tlb_range() 135 __flush_tlb_range(vma, start, end, false); flush_tlb_range() 138 static inline void flush_tlb_kernel_range(unsigned long start, unsigned long end) flush_tlb_kernel_range() argument 142 if ((end - start) > MAX_TLB_RANGE) { flush_tlb_kernel_range() 148 end >>= 12; flush_tlb_kernel_range() 151 for (addr = start; addr < end; addr += 1 << (PAGE_SHIFT - 12)) flush_tlb_kernel_range() 107 __flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end, bool last_level) __flush_tlb_range() argument 132 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_tlb_range() argument
|
/linux-4.4.14/arch/arm/include/asm/ |
H A D | outercache.h | 50 * @end: end physical address, exclusive 52 static inline void outer_inv_range(phys_addr_t start, phys_addr_t end) outer_inv_range() argument 55 outer_cache.inv_range(start, end); outer_inv_range() 61 * @end: end physical address, exclusive 63 static inline void outer_clean_range(phys_addr_t start, phys_addr_t end) outer_clean_range() argument 66 outer_cache.clean_range(start, end); outer_clean_range() 72 * @end: end physical address, exclusive 74 static inline void outer_flush_range(phys_addr_t start, phys_addr_t end) outer_flush_range() argument 77 outer_cache.flush_range(start, end); outer_flush_range() 120 static inline void outer_inv_range(phys_addr_t start, phys_addr_t end) outer_inv_range() argument 122 static inline void outer_clean_range(phys_addr_t start, phys_addr_t end) outer_clean_range() argument 124 static inline void outer_flush_range(phys_addr_t start, phys_addr_t end) outer_flush_range() argument
|
/linux-4.4.14/drivers/media/platform/coda/ |
H A D | trace.h | 92 __field(int, end) 100 __entry->end = meta->end; 104 TP_printk("minor = %d, index = %d, start = 0x%x, end = 0x%x, ctx = %d", 105 __entry->minor, __entry->index, __entry->start, __entry->end, 123 __field(int, end) 130 __entry->end = meta ? meta->end : 0; 134 TP_printk("minor = %d, start = 0x%x, end = 0x%x, ctx = %d", 135 __entry->minor, __entry->start, __entry->end, __entry->ctx)
|
/linux-4.4.14/fs/ceph/ |
H A D | mdsmap.c | 51 struct ceph_mdsmap *ceph_mdsmap_decode(void **p, void *end) ceph_mdsmap_decode() argument 63 ceph_decode_16_safe(p, end, version, bad); ceph_mdsmap_decode() 69 ceph_decode_need(p, end, 8*sizeof(u32) + sizeof(u64), bad); ceph_mdsmap_decode() 97 ceph_decode_need(p, end, sizeof(u64)*2 + 1 + sizeof(u32), bad); ceph_mdsmap_decode() 104 ceph_decode_need(p, end, ceph_mdsmap_decode() 116 ceph_decode_32_safe(p, end, namelen, bad); ceph_mdsmap_decode() 119 ceph_decode_32_safe(p, end, num_export_targets, bad); ceph_mdsmap_decode() 155 ceph_decode_32_safe(p, end, n, bad); ceph_mdsmap_decode() 160 ceph_decode_need(p, end, sizeof(u64)*(n+1), bad); ceph_mdsmap_decode() 175 start, end - start, true); ceph_mdsmap_decode()
|
/linux-4.4.14/arch/x86/kvm/ |
H A D | mtrr.c | 147 u64 end; member in struct:fixed_mtrr_segment 159 .end = 0x80000, 170 .end = 0xc0000, 181 .end = 0x100000, 218 static void fixed_mtrr_seg_unit_range(int seg, int unit, u64 *start, u64 *end) fixed_mtrr_seg_unit_range() argument 224 *end = *start + unit_size; fixed_mtrr_seg_unit_range() 225 WARN_ON(*end > mtrr_seg->end); fixed_mtrr_seg_unit_range() 233 > mtrr_seg->end); fixed_mtrr_seg_unit_range_index() 244 n = (mtrr_seg->end - mtrr_seg->start) >> mtrr_seg->range_shift; fixed_mtrr_seg_end_range_index() 248 static bool fixed_msr_to_range(u32 msr, u64 *start, u64 *end) fixed_msr_to_range() argument 255 fixed_mtrr_seg_unit_range(seg, unit, start, end); fixed_msr_to_range() 276 if (mtrr_seg->start <= addr && addr < mtrr_seg->end) fixed_mtrr_addr_to_seg() 302 static void var_mtrr_range(struct kvm_mtrr_range *range, u64 *start, u64 *end) var_mtrr_range() argument 313 *end = (*start | ~mask) + 1; var_mtrr_range() 319 gfn_t start, end; update_mtrr() local 330 if (fixed_msr_to_range(msr, &start, &end)) { update_mtrr() 335 end = ~0ULL; update_mtrr() 339 var_mtrr_range(&mtrr_state->var_ranges[index], &start, &end); update_mtrr() 342 kvm_zap_gfn_range(vcpu->kvm, gpa_to_gfn(start), gpa_to_gfn(end)); update_mtrr() 454 u64 end; member in struct:mtrr_iter 460 /* [start, end) is not fully covered in MTRRs? */ 503 u64 start, end; match_var_range() local 505 var_mtrr_range(range, &start, &end); match_var_range() 506 if (!(start >= iter->end || end <= iter->start)) { match_var_range() 512 * [looker->start_max, looker->end). match_var_range() 517 iter->start_max = max(iter->start_max, end); match_var_range() 533 iter->partial_map |= iter->start_max < iter->end; __mtrr_lookup_var_next() 550 if (fixed_mtrr_range_end_addr(iter->seg, iter->index) >= iter->end) { mtrr_lookup_fixed_next() 584 struct kvm_mtrr *mtrr_state, u64 start, u64 end) mtrr_lookup_init() 588 iter->end = end; mtrr_lookup_init() 628 u64 start, end; kvm_mtrr_get_guest_memory_type() local 634 end = start + PAGE_SIZE; kvm_mtrr_get_guest_memory_type() 636 mtrr_for_each_mem_type(&iter, mtrr_state, start, end) { kvm_mtrr_get_guest_memory_type() 705 u64 start, end; kvm_mtrr_check_gfn_range_consistency() local 709 end = gfn_to_gpa(gfn + page_num); kvm_mtrr_check_gfn_range_consistency() 710 mtrr_for_each_mem_type(&iter, mtrr_state, start, end) { kvm_mtrr_check_gfn_range_consistency() 583 mtrr_lookup_init(struct mtrr_iter *iter, struct kvm_mtrr *mtrr_state, u64 start, u64 end) mtrr_lookup_init() argument
|
/linux-4.4.14/sound/firewire/oxfw/ |
H A D | oxfw-stream.c | 49 goto end; set_rate() 54 end: set_rate() 139 goto end; start_stream() 147 goto end; start_stream() 156 goto end; start_stream() 161 goto end; start_stream() 166 goto end; start_stream() 173 goto end; start_stream() 181 end: start_stream() 229 goto end; snd_oxfw_stream_init_simplex() 235 goto end; snd_oxfw_stream_init_simplex() 250 end: snd_oxfw_stream_init_simplex() 282 goto end; snd_oxfw_stream_start_simplex() 290 goto end; snd_oxfw_stream_start_simplex() 298 goto end; snd_oxfw_stream_start_simplex() 308 goto end; snd_oxfw_stream_start_simplex() 317 goto end; snd_oxfw_stream_start_simplex() 327 goto end; snd_oxfw_stream_start_simplex() 339 end: snd_oxfw_stream_start_simplex() 402 goto end; snd_oxfw_stream_get_current_formation() 405 goto end; snd_oxfw_stream_get_current_formation() 409 end: snd_oxfw_stream_get_current_formation() 512 goto end; assume_stream_formats() 519 goto end; assume_stream_formats() 524 goto end; assume_stream_formats() 542 goto end; assume_stream_formats() 549 end: assume_stream_formats() 579 goto end; fill_stream_formats() 585 goto end; fill_stream_formats() 624 end: fill_stream_formats() 643 goto end; snd_oxfw_stream_discover() 646 goto end; snd_oxfw_stream_discover() 653 goto end; snd_oxfw_stream_discover() 675 goto end; snd_oxfw_stream_discover() 690 end: snd_oxfw_stream_discover() 709 goto end; snd_oxfw_stream_lock_try() 716 end: snd_oxfw_stream_lock_try() 726 goto end; snd_oxfw_stream_lock_release() 729 end: snd_oxfw_stream_lock_release()
|
/linux-4.4.14/arch/unicore32/kernel/ |
H A D | puv3-core.c | 53 .end = io_v2p(PKUNITY_USB_BASE) + 0x3ff, 85 .end = io_v2p(PKUNITY_SDC_BASE) + 0xfff, 90 .end = IRQ_SDC, 98 .end = io_v2p(PKUNITY_UNIGFX_BASE) + 0xfff, 106 .end = io_v2p(PKUNITY_RTC_BASE) + 0xff, 111 .end = IRQ_RTCAlarm, 116 .end = IRQ_RTC, 124 .end = io_v2p(PKUNITY_OST_BASE) + 0xff, 132 .end = io_v2p(PKUNITY_UART0_BASE) + 0xff, 137 .end = IRQ_UART0, 145 .end = io_v2p(PKUNITY_UART1_BASE) + 0xff, 150 .end = IRQ_UART1, 158 .end = io_v2p(PKUNITY_UMAL_BASE) + 0x1fff, 163 .end = IRQ_UMAL,
|
/linux-4.4.14/arch/unicore32/mm/ |
H A D | cache-ucv2.S | 44 * __cpuc_flush_user_range(start, end, flags) 49 * - end - end address (exclusive, may not be aligned) 83 * __cpuc_coherent_kern_range(start,end) 84 * __cpuc_coherent_user_range(start,end) 91 * - end - virtual end address of region 145 * __cpuc_dma_clean_range(start,end) 147 * - end - virtual end address of region 182 * __cpuc_dma_inv_range(start,end) 183 * __cpuc_dma_flush_range(start,end) 185 * - end - virtual end address of region
|
H A D | init.c | 72 struct page *page, *end; for_each_bank() local 78 end = pfn_to_page(pfn2 - 1) + 1; for_each_bank() 93 } while (page < end); for_each_bank() 115 unsigned long start, end; for_each_bank() local 118 end = bank_pfn_end(bank); for_each_bank() 122 if (*max_high < end) for_each_bank() 123 *max_high = end; for_each_bank() 126 if (*max_low < end) for_each_bank() 127 *max_low = end; for_each_bank() 158 unsigned long end = memblock_region_memory_end_pfn(reg); for_each_memblock() local 160 if (end >= end_pfn) for_each_memblock() 161 end = end_pfn; for_each_memblock() 162 if (start >= end) for_each_memblock() 165 free_bootmem(__pfn_to_phys(start), (end - start) << PAGE_SHIFT); for_each_memblock() 171 unsigned long end = memblock_region_reserved_end_pfn(reg); for_each_memblock() local 173 if (end >= end_pfn) for_each_memblock() 174 end = end_pfn; for_each_memblock() 175 if (start >= end) for_each_memblock() 179 (end - start) << PAGE_SHIFT, BOOTMEM_DEFAULT); for_each_memblock() 208 unsigned long end = memblock_region_memory_end_pfn(reg); for_each_memblock() local 211 unsigned long low_end = min(end, max_low); for_each_memblock() 330 * round start upwards and end downwards. free_memmap() 369 * memmap entries are valid from the bank end aligned to for_each_bank() 438 void free_initrd_mem(unsigned long start, unsigned long end) free_initrd_mem() argument 441 free_reserved_area((void *)start, (void *)end, -1, "initrd"); free_initrd_mem()
|
/linux-4.4.14/fs/hfsplus/ |
H A D | bitmap.c | 24 __be32 *pptr, *curr, *end; hfsplus_block_allocate() local 46 end = pptr + PAGE_CACHE_BITS / 32; hfsplus_block_allocate() 48 end = pptr + ((size + 31) & (PAGE_CACHE_BITS - 1)) / 32; hfsplus_block_allocate() 64 while (curr < end) { hfsplus_block_allocate() 88 end = pptr + PAGE_CACHE_BITS / 32; hfsplus_block_allocate() 90 end = pptr + ((size + 31) & (PAGE_CACHE_BITS - 1)) / 32; hfsplus_block_allocate() 117 while (curr < end) { hfsplus_block_allocate() 139 end = pptr + PAGE_CACHE_BITS / 32; hfsplus_block_allocate() 142 /* do any partial u32 at end */ hfsplus_block_allocate() 168 __be32 *pptr, *curr, *end; hfsplus_block_free() local 189 end = pptr + PAGE_CACHE_BITS / 32; hfsplus_block_free() 208 while (curr < end) { hfsplus_block_free() 223 end = pptr + PAGE_CACHE_BITS / 32; hfsplus_block_free() 226 /* do any partial u32 at end */ hfsplus_block_free()
|
/linux-4.4.14/arch/powerpc/mm/ |
H A D | tlb_hash32.c | 63 * Called at the end of a mmu_gather operation to make sure the 82 * - flush_tlb_range(vma, start, end) flushes a range of pages 83 * - flush_tlb_kernel_range(start, end) flushes kernel pages 91 unsigned long end) flush_range() 103 if (start >= end) flush_range() 105 end = (end - 1) | ~PAGE_MASK; flush_range() 109 if (pmd_end > end) flush_range() 110 pmd_end = end; flush_range() 115 if (pmd_end == end) flush_range() 125 void flush_tlb_kernel_range(unsigned long start, unsigned long end) flush_tlb_kernel_range() argument 127 flush_range(&init_mm, start, end); flush_tlb_kernel_range() 176 unsigned long end) flush_tlb_range() 178 flush_range(vma->vm_mm, start, end); flush_tlb_range() 90 flush_range(struct mm_struct *mm, unsigned long start, unsigned long end) flush_range() argument 175 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_tlb_range() argument
|
/linux-4.4.14/drivers/staging/lustre/lustre/include/ |
H A D | interval_tree.h | 56 __u64 end; member in struct:interval_node::interval_node_extent 72 return node->in_extent.end; interval_high() 76 __u64 start, __u64 end) interval_set() 78 LASSERT(start <= end); interval_set() 80 node->in_extent.end = end; interval_set() 81 node->in_max_high = end; interval_set() 75 interval_set(struct interval_node *node, __u64 start, __u64 end) interval_set() argument
|
/linux-4.4.14/arch/sh/boards/ |
H A D | board-sh2007.c | 35 .end = SMC0_BASE + 0xff, 40 .end = evt2irq(0x240), 48 .end = SMC1_BASE + 0xff, 53 .end = evt2irq(0x280), 81 .end = CF_BASE + CF_OFFSET + 0x0f, 86 .end = CF_BASE + CF_OFFSET + 0x20f, 91 .end = evt2irq(0x2c0),
|
/linux-4.4.14/arch/ia64/include/asm/ |
H A D | meminit.h | 28 u64 end; /* virtual address of end of element + 1 */ member in struct:rsvd_region 37 extern int filter_rsvd_memory (u64 start, u64 end, void *arg); 38 extern int filter_memory (u64 start, u64 end, void *arg); 43 extern int reserve_elfcorehdr(u64 *start, u64 *end); 65 extern int find_largest_hole(u64 start, u64 end, void *arg); 66 extern int create_mem_map_page_table(u64 start, u64 end, void *arg);
|
H A D | patch.h | 21 extern void ia64_patch_mckinley_e9 (unsigned long start, unsigned long end); 22 extern void ia64_patch_vtop (unsigned long start, unsigned long end); 24 extern void ia64_patch_rse (unsigned long start, unsigned long end);
|
/linux-4.4.14/arch/unicore32/include/asm/ |
H A D | cacheflush.h | 33 * Start addresses are inclusive and end addresses are exclusive; 34 * start addresses should be rounded down, end addresses up. 55 * flush_user_range(start, end, flags) 60 * - end - user end address (exclusive, page aligned) 63 * coherent_kern_range(start, end) 66 * region described by start, end. If you have non-snooping 69 * - end - virtual end address 71 * coherent_user_range(start, end) 74 * region described by start, end. If you have non-snooping 77 * - end - virtual end address 88 * dma_flush_range(start, end) 92 * - end - virtual end address 141 unsigned long start, unsigned long end); 152 #define flush_cache_user_range(vma, start, end) \ 153 __cpuc_coherent_user_range((start) & PAGE_MASK, PAGE_ALIGN(end)) 203 static inline void flush_cache_vmap(unsigned long start, unsigned long end) flush_cache_vmap() argument 207 static inline void flush_cache_vunmap(unsigned long start, unsigned long end) flush_cache_vunmap() argument
|
/linux-4.4.14/sound/firewire/fireworks/ |
H A D | fireworks_stream.c | 32 goto end; init_stream() 39 end: init_stream() 65 goto end; start_stream() 79 goto end; start_stream() 85 goto end; start_stream() 93 goto end; start_stream() 101 end: start_stream() 170 goto end; snd_efw_stream_init_duplex() 196 goto end; snd_efw_stream_init_duplex() 205 end: snd_efw_stream_init_duplex() 222 goto end; snd_efw_stream_start_duplex() 226 goto end; snd_efw_stream_start_duplex() 243 goto end; snd_efw_stream_start_duplex() 254 goto end; snd_efw_stream_start_duplex() 269 goto end; snd_efw_stream_start_duplex() 275 goto end; snd_efw_stream_start_duplex() 288 end: snd_efw_stream_start_duplex() 357 goto end; snd_efw_stream_lock_try() 364 end: snd_efw_stream_lock_try() 374 goto end; snd_efw_stream_lock_release() 377 end: snd_efw_stream_lock_release()
|
/linux-4.4.14/drivers/block/drbd/ |
H A D | drbd_interval.c | 6 * interval_end - return end of @node 12 return this->end; interval_end() 16 * compute_subtree_last - compute end of @node 18 * The end of an interval is the highest (start + (size >> 9)) value of this 19 * node and of its children. Called for @node and its parents whenever the end 41 sector_t, end, compute_subtree_last); 59 if (here->end < this_end) drbd_insert_interval() 60 here->end = this_end; drbd_insert_interval() 73 this->end = this_end; drbd_insert_interval() 138 sector_t end = sector + (size >> 9); drbd_find_overlap() local 150 } else if (here->sector < end && drbd_find_overlap() 166 sector_t end = sector + (size >> 9); drbd_next_overlap() local 174 if (i->sector >= end) drbd_next_overlap()
|
/linux-4.4.14/arch/m68k/sun3x/ |
H A D | dvma.c | 86 unsigned long end; dvma_map_cpu() local 92 end = PAGE_ALIGN(vaddr + len); dvma_map_cpu() 109 if((end & PGDIR_MASK) > (vaddr & PGDIR_MASK)) dvma_map_cpu() 112 end2 = end; dvma_map_cpu() 142 } while(vaddr < end); dvma_map_cpu() 154 unsigned long end, index; dvma_map_iommu() local 157 end = ((baddr+len) >> DVMA_PAGE_SHIFT); dvma_map_iommu() 160 end++; dvma_map_iommu() 162 for(; index < end ; index++) { dvma_map_iommu() 176 for(index = (baddr >> DVMA_PAGE_SHIFT); index < end; index++) dvma_map_iommu() 186 int index, end; dvma_unmap_iommu() local 190 end = (DVMA_PAGE_ALIGN(baddr+len) >> DVMA_PAGE_SHIFT); dvma_unmap_iommu() 192 for(; index < end ; index++) { dvma_unmap_iommu()
|
/linux-4.4.14/arch/mips/sni/ |
H A D | pcit.c | 64 .end = 0x71, 69 .end = 8, 87 .end = 0x03bfffffUL, 95 .end = 0x1f, 100 .end = 0x5f, 105 .end = 0x6f, 110 .end = 0x8f, 115 .end = 0xdf, 120 .end = 0xcfb, 125 .end = 0xcff, 146 .end = 0x1fbfffffUL, 265 ioport_resource.end = sni_io_resource.end; sni_pcit_init()
|
H A D | a20r.c | 46 .end = 0x1c081fff, 60 .end = 0x18000004, 65 .end = 0x18010004, 70 .end = 0x1ff00020, 75 .end = 22, 92 .end = 0x190fffff, 97 .end = 19, 111 .end = 0x1c0700ff, 116 .end = 20,
|
/linux-4.4.14/arch/sh/drivers/superhyway/ |
H A D | ops-sh4-202.c | 31 .end = PHYS_EMI_CBLOCK + 0x00300000 - 1, 36 .end = PHYS_EMI_DBLOCK + 0x08000000 - 1, 50 .end = PHYS_FEMI_CBLOCK + 0x00100000 - 1, 55 .end = PHYS_FEMI_DBLOCK + 0x08000000 - 1, 69 .end = P4SEGADDR(0x1e7ffff8 + (sizeof(u32) * 2) - 1), 74 .end = PHYS_EPBR_BLOCK + 0x00a00000 - 1, 87 .end = PHYS_DMAC_BLOCK + 0x00100000 - 1, 100 .end = P4SEGADDR(0x1ffffff8 + (sizeof(u32) * 2) - 1), 105 .end = PHYS_PBR_BLOCK + 0x00400000 - (sizeof(u32) * 2) - 1,
|
/linux-4.4.14/arch/arm/mach-dove/ |
H A D | mpp.c | 21 int end; member in struct:dove_mpp_grp 28 .end = 39, 32 .end = 45, 36 .end = 51, 40 .end = 61, 44 .end = 63, 50 static void __init dove_mpp_gpio_mode(int start, int end, int gpio_mode) dove_mpp_gpio_mode() argument 54 for (i = start; i <= end; i++) dove_mpp_gpio_mode() 141 dove_mpp_grp[num].end, gpio_mode); dove_mpp_conf_grp()
|
/linux-4.4.14/tools/testing/selftests/timers/ |
H A D | raw_skew.c | 72 long long diff_timespec(struct timespec start, struct timespec end) diff_timespec() argument 77 end_ns = ts_to_nsec(end); diff_timespec() 83 struct timespec start, mid, end; get_monotonic_and_raw() local 92 clock_gettime(CLOCK_MONOTONIC, &end); get_monotonic_and_raw() 94 newdiff = diff_timespec(start, end); get_monotonic_and_raw() 98 tmp = (ts_to_nsec(start) + ts_to_nsec(end))/2; get_monotonic_and_raw() 106 struct timespec mon, raw, start, end; main() local 130 end = mon; main() 135 interval = diff_timespec(start, end); main()
|
/linux-4.4.14/arch/sparc/mm/ |
H A D | gup.c | 21 unsigned long end, int write, struct page **pages, int *nr) gup_pte_range() 64 } while (ptep++, addr += PAGE_SIZE, addr != end); gup_pte_range() 70 unsigned long end, int write, struct page **pages, gup_huge_pmd() 92 } while (addr += PAGE_SIZE, addr != end); gup_huge_pmd() 118 static int gup_pmd_range(pud_t pud, unsigned long addr, unsigned long end, gup_pmd_range() argument 128 next = pmd_addr_end(addr, end); gup_pmd_range() 138 } while (pmdp++, addr = next, addr != end); gup_pmd_range() 143 static int gup_pud_range(pgd_t pgd, unsigned long addr, unsigned long end, gup_pud_range() argument 153 next = pud_addr_end(addr, end); gup_pud_range() 158 } while (pudp++, addr = next, addr != end); gup_pud_range() 167 unsigned long addr, len, end; __get_user_pages_fast() local 175 end = start + len; __get_user_pages_fast() 182 next = pgd_addr_end(addr, end); __get_user_pages_fast() 187 } while (pgdp++, addr = next, addr != end); __get_user_pages_fast() 197 unsigned long addr, len, end; get_user_pages_fast() local 205 end = start + len; get_user_pages_fast() 230 next = pgd_addr_end(addr, end); get_user_pages_fast() 235 } while (pgdp++, addr = next, addr != end); get_user_pages_fast() 239 VM_BUG_ON(nr != (end - start) >> PAGE_SHIFT); get_user_pages_fast() 253 (end - start) >> PAGE_SHIFT, write, 0, pages); get_user_pages_fast() 20 gup_pte_range(pmd_t pmd, unsigned long addr, unsigned long end, int write, struct page **pages, int *nr) gup_pte_range() argument 69 gup_huge_pmd(pmd_t *pmdp, pmd_t pmd, unsigned long addr, unsigned long end, int write, struct page **pages, int *nr) gup_huge_pmd() argument
|
/linux-4.4.14/Documentation/watchdog/src/ |
H A D | watchdog-test.c | 60 goto end; main() 66 goto end; main() 72 goto end; main() 82 goto end; main() 95 end: main()
|
/linux-4.4.14/arch/s390/mm/ |
H A D | vmem.c | 80 unsigned long end = start + size; vmem_add_mem() local 88 while (address < end) { vmem_add_mem() 99 !(address & ~PUD_MASK) && (address + PUD_SIZE <= end)) { vmem_add_mem() 116 !(address & ~PMD_MASK) && (address + PMD_SIZE <= end)) { vmem_add_mem() 148 unsigned long end = start + size; vmem_remove_range() local 157 while (address < end) { vmem_remove_range() 187 flush_tlb_kernel_range(start, end); vmem_remove_range() 193 int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node) vmemmap_populate() argument 202 for (address = start; address < end;) { vmemmap_populate() 264 void vmemmap_free(unsigned long start, unsigned long end) vmemmap_free() argument 369 phys_addr_t start, end; vmem_map_init() local 375 end = reg->base + reg->size - 1; for_each_memblock() 376 if (start >= ro_end || end <= ro_start) for_each_memblock() 377 vmem_add_mem(start, end - start, 0); for_each_memblock() 378 else if (start >= ro_start && end <= ro_end) for_each_memblock() 379 vmem_add_mem(start, end - start, 1); for_each_memblock() 382 vmem_add_mem(ro_end, end - ro_end, 0); for_each_memblock() 383 } else if (end < ro_end) { for_each_memblock() 385 vmem_add_mem(ro_start, end - ro_start, 1); for_each_memblock() 389 vmem_add_mem(ro_end, end - ro_end, 0); for_each_memblock()
|
/linux-4.4.14/arch/score/include/asm/ |
H A D | cacheflush.h | 10 unsigned long start, unsigned long end); 15 extern void flush_icache_range(unsigned long start, unsigned long end); 16 extern void flush_dcache_range(unsigned long start, unsigned long end); 25 #define flush_cache_vmap(start, end) do {} while (0) 26 #define flush_cache_vunmap(start, end) do {} while (0)
|
/linux-4.4.14/arch/sh/boards/mach-sh03/ |
H A D | setup.c | 35 .end = 0x1f0 + 8, 40 .end = 0x1f0 +8 + 0x206 + 8, 59 .end = 0xa0800000, 93 cf_ide_resources[0].end += (unsigned long)cf_ide_base; sh03_devices_setup() 95 cf_ide_resources[1].end += (unsigned long)cf_ide_base; sh03_devices_setup()
|
/linux-4.4.14/drivers/firmware/efi/ |
H A D | fake_mem.c | 59 u64 start, end, m_start, m_end, m_attr; efi_fake_memmap() local 74 end = start + (md->num_pages << EFI_PAGE_SHIFT) - 1; efi_fake_memmap() 79 m_end = fake_mems[i].range.end; efi_fake_memmap() 83 if (start < m_end && m_end < end) efi_fake_memmap() 86 if (start < m_start && m_start < end) { efi_fake_memmap() 88 if (m_end < end) efi_fake_memmap() 91 if (end <= m_end) efi_fake_memmap() 119 end = md->phys_addr + (md->num_pages << EFI_PAGE_SHIFT) - 1; efi_fake_memmap() 124 m_end = fake_mems[i].range.end; efi_fake_memmap() 127 if (m_start <= start && end <= m_end) efi_fake_memmap() 131 (start < m_end && m_end < end)) { efi_fake_memmap() 141 md->num_pages = (end - md->phys_addr + 1) >> efi_fake_memmap() 145 if ((start < m_start && m_start < end) && m_end < end) { efi_fake_memmap() 162 md->num_pages = (end - m_end) >> efi_fake_memmap() 166 if ((start < m_start && m_start < end) && efi_fake_memmap() 167 (end <= m_end)) { efi_fake_memmap() 176 md->num_pages = (end - md->phys_addr + 1) >> efi_fake_memmap() 219 fake_mems[nr_fake_mem].range.end = start + mem_size - 1; setup_fake_mem() 233 fake_mems[i].range.end); setup_fake_mem()
|
/linux-4.4.14/sound/firewire/dice/ |
H A D | dice-stream.c | 71 goto end; keep_resources() 83 end: keep_resources() 108 goto end; start_stream() 138 goto end; start_stream() 155 goto end; start_stream() 162 end: start_stream() 173 goto end; get_sync_mode() 189 end: get_sync_mode() 201 goto end; snd_dice_stream_start_duplex() 205 goto end; snd_dice_stream_start_duplex() 223 goto end; snd_dice_stream_start_duplex() 240 goto end; snd_dice_stream_start_duplex() 248 goto end; snd_dice_stream_start_duplex() 255 goto end; snd_dice_stream_start_duplex() 263 goto end; snd_dice_stream_start_duplex() 275 end: snd_dice_stream_start_duplex() 306 goto end; init_stream() 314 end: init_stream() 343 goto end; snd_dice_stream_init_duplex() 348 end: snd_dice_stream_init_duplex()
|
/linux-4.4.14/tools/perf/scripts/python/ |
H A D | net_dropmonitor.py | 34 # kallsyms[i][0] > loc for all end <= i < len(kallsyms) 35 start, end = -1, len(kallsyms) 36 while end != start + 1: 37 pivot = (start + end) // 2 39 end = pivot
|
/linux-4.4.14/arch/arc/kernel/ |
H A D | unwind.c | 172 const void *end, signed ptrType); 198 const u8 *end = header_start + header_size; init_unwind_table() local 210 || (void *)read_pointer(&ptr, end, header_start[1]) != table_start init_unwind_table() 212 || read_pointer(&ptr, end, header_start[2]) <= 0 init_unwind_table() 463 static uleb128_t get_uleb128(const u8 **pcur, const u8 *end) get_uleb128() argument 469 for (shift = 0, value = 0; cur < end; shift += 7) { get_uleb128() 472 cur = end + 1; get_uleb128() 484 static sleb128_t get_sleb128(const u8 **pcur, const u8 *end) get_sleb128() argument 490 for (shift = 0, value = 0; cur < end; shift += 7) { get_sleb128() 493 cur = end + 1; get_sleb128() 531 static unsigned long read_pointer(const u8 **pLoc, const void *end, read_pointer() argument 549 if (end < (const void *)(ptr.p16u + 1)) read_pointer() 558 if (end < (const void *)(ptr.p32u + 1)) read_pointer() 571 if (end < (const void *)(ptr.pul + 1)) read_pointer() 577 value = ptrType & DW_EH_PE_signed ? get_sleb128(&ptr.p8, end) read_pointer() 578 : get_uleb128(&ptr.p8, end); read_pointer() 579 if ((const void *)ptr.p8 > end) read_pointer() 609 const u8 *end = (const u8 *)(cie + 1) + *cie; fde_pointer_type() local 618 ptr = memchr(aug, 0, end - ptr); fde_pointer_type() 623 get_uleb128(&ptr, end); /* skip code alignment */ fde_pointer_type() 624 get_sleb128(&ptr, end); /* skip data alignment */ fde_pointer_type() 626 version <= 1 ? (void) ++ptr : (void)get_uleb128(&ptr, end); fde_pointer_type() 627 len = get_uleb128(&ptr, end); /* augmentation length */ fde_pointer_type() 629 if (ptr + len < ptr || ptr + len > end) fde_pointer_type() 632 end = ptr + len; fde_pointer_type() 634 if (ptr >= end) fde_pointer_type() 643 if (!read_pointer(&ptr, end, ptrType) fde_pointer_type() 644 || ptr > end) fde_pointer_type() 698 static int processCFI(const u8 *start, const u8 *end, unsigned long targetLoc, processCFI() argument 717 for (ptr.p8 = start; result && ptr.p8 < end;) { processCFI() 729 state->loc = read_pointer(&ptr.p8, end, processCFI() 737 result = ptr.p8 < end processCFI() 744 result = ptr.p8 <= end + 2 processCFI() 750 result = ptr.p8 <= end + 4 processCFI() 754 value = get_uleb128(&ptr.p8, end); processCFI() 757 get_uleb128(&ptr.p8, end), state); processCFI() 760 value = get_uleb128(&ptr.p8, end); processCFI() 762 get_uleb128(&ptr.p8, end), state); processCFI() 765 value = get_uleb128(&ptr.p8, end); processCFI() 767 get_sleb128(&ptr.p8, end), state); processCFI() 770 value = get_uleb128(&ptr.p8, end); processCFI() 772 get_sleb128(&ptr.p8, end), state); processCFI() 780 set_rule(get_uleb128(&ptr.p8, end), Nowhere, 0, processCFI() 785 value = get_uleb128(&ptr.p8, end); processCFI() 788 get_uleb128(&ptr.p8, end), state); processCFI() 814 processCFI(start, end, 0, ptrType, processCFI() 822 state->cfa.reg = get_uleb128(&ptr.p8, end); processCFI() 826 state->cfa.offs = get_uleb128(&ptr.p8, end); processCFI() 831 state->cfa.reg = get_uleb128(&ptr.p8, end); processCFI() 834 state->cfa.offs = get_sleb128(&ptr.p8, end) processCFI() 839 state->cfa.reg = get_uleb128(&ptr.p8, end); processCFI() 845 get_uleb128(&ptr.p8, end); processCFI() 848 value = get_uleb128(&ptr.p8, end); processCFI() 852 end), processCFI() 869 set_rule(value, Memory, get_uleb128(&ptr.p8, end), processCFI() 878 if (ptr.p8 > end) processCFI() 884 return result && ptr.p8 == end && (targetLoc == 0 || ( processCFI() 887 never reaches the end of the function. processCFI() 897 const u8 *ptr = NULL, *end = NULL; arc_unwind() local 952 end = hdr + table->hdrsz; arc_unwind() 953 if (tableSize && read_pointer(&ptr, end, hdr[1]) arc_unwind() 955 && (i = read_pointer(&ptr, end, hdr[2])) > 0 arc_unwind() 956 && i == (end - ptr) / (2 * tableSize) arc_unwind() 957 && !((end - ptr) % (2 * tableSize))) { arc_unwind() 1016 end = (const u8 *)(cie + 1) + *cie; arc_unwind() 1021 while (++ptr < end && *ptr) { arc_unwind() 1038 if (ptr >= end || *ptr) arc_unwind() 1045 state.codeAlign = get_uleb128(&ptr, end); arc_unwind() 1047 state.dataAlign = get_sleb128(&ptr, end); arc_unwind() 1048 if (state.codeAlign == 0 || state.dataAlign == 0 || ptr >= end) arc_unwind() 1053 end); arc_unwind() 1061 uleb128_t augSize = get_uleb128(&ptr, end); arc_unwind() 1065 if (ptr > end || retAddrReg >= ARRAY_SIZE(reg_info) arc_unwind() 1075 state.cieEnd = end; arc_unwind() 1076 end = (const u8 *)(fde + 1) + *fde; arc_unwind() 1079 uleb128_t augSize = get_uleb128(&ptr, end); arc_unwind() 1081 if ((ptr += augSize) > end) arc_unwind() 1139 if (!processCFI(ptr, end, pc, ptrType, &state) arc_unwind()
|
/linux-4.4.14/fs/btrfs/ |
H A D | extent_io.h | 71 u64 start, u64 end, int *page_started, 73 int (*writepage_start_hook)(struct page *page, u64 start, u64 end); 74 int (*writepage_io_hook)(struct page *page, u64 start, u64 end); 81 struct page *page, u64 start, u64 end, 83 int (*writepage_end_io_hook)(struct page *page, u64 start, u64 end, 107 u64 end; /* inclusive */ member in struct:extent_state 202 int lock_extent(struct extent_io_tree *tree, u64 start, u64 end); 203 int lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, 205 int unlock_extent(struct extent_io_tree *tree, u64 start, u64 end); 206 int unlock_extent_cached(struct extent_io_tree *tree, u64 start, u64 end, 208 int try_lock_extent(struct extent_io_tree *tree, u64 start, u64 end); 221 int test_range_bit(struct extent_io_tree *tree, u64 start, u64 end, 224 int clear_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, 226 int clear_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, 229 int clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, 232 int set_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, 234 int set_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, 237 int set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, 240 int set_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end, 242 int clear_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end, 244 int set_extent_new(struct extent_io_tree *tree, u64 start, u64 end, 246 int set_extent_dirty(struct extent_io_tree *tree, u64 start, u64 end, 248 int clear_extent_dirty(struct extent_io_tree *tree, u64 start, u64 end, 250 int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, 253 int set_extent_delalloc(struct extent_io_tree *tree, u64 start, u64 end, 255 int set_extent_defrag(struct extent_io_tree *tree, u64 start, u64 end, 266 u64 start, u64 end, get_extent_t *get_extent, 341 int extent_range_clear_dirty_for_io(struct inode *inode, u64 start, u64 end); 342 int extent_range_redirty_for_io(struct inode *inode, u64 start, u64 end); 343 int extent_clear_unlock_delalloc(struct inode *inode, u64 start, u64 end, 360 int end_extent_writepage(struct page *page, int err, u64 start, u64 end); 383 void btrfs_free_io_failure_record(struct inode *inode, u64 start, u64 end); 384 int btrfs_get_io_failure_record(struct inode *inode, u64 start, u64 end, 397 u64 *end, u64 max_bytes);
|
/linux-4.4.14/arch/m68k/include/asm/ |
H A D | cacheflush_mm.h | 30 static inline void clear_cf_icache(unsigned long start, unsigned long end) clear_cf_icache() argument 39 static inline void clear_cf_dcache(unsigned long start, unsigned long end) clear_cf_dcache() argument 48 static inline void clear_cf_bcache(unsigned long start, unsigned long end) clear_cf_bcache() argument 59 * The start and end addresses are cache line numbers not memory addresses. 61 static inline void flush_cf_icache(unsigned long start, unsigned long end) flush_cf_icache() argument 65 for (set = start; set <= end; set += (0x10 - 3)) { flush_cf_icache() 79 static inline void flush_cf_dcache(unsigned long start, unsigned long end) flush_cf_dcache() argument 83 for (set = start; set <= end; set += (0x10 - 3)) { flush_cf_dcache() 97 static inline void flush_cf_bcache(unsigned long start, unsigned long end) flush_cf_bcache() argument 101 for (set = start; set <= end; set += (0x10 - 3)) { flush_cf_bcache() 192 #define flush_cache_vmap(start, end) flush_cache_all() 193 #define flush_cache_vunmap(start, end) flush_cache_all() 207 unsigned long end) flush_cache_range() 225 unsigned long addr, start, end; __flush_page_to_ram() local 228 end = (addr + PAGE_SIZE - 1) & ICACHE_SET_MASK; __flush_page_to_ram() 229 if (start > end) { __flush_page_to_ram() 230 flush_cf_bcache(0, end); __flush_page_to_ram() 231 end = ICACHE_MAX_ADDR; __flush_page_to_ram() 233 flush_cf_bcache(start, end); __flush_page_to_ram() 205 flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) flush_cache_range() argument
|
/linux-4.4.14/arch/x86/purgatory/ |
H A D | purgatory.c | 44 struct sha_region *ptr, *end; verify_sha256_digest() local 49 end = &sha_regions[sizeof(sha_regions)/sizeof(sha_regions[0])]; verify_sha256_digest() 50 for (ptr = sha_regions; ptr < end; ptr++) verify_sha256_digest()
|
/linux-4.4.14/arch/sh/boards/mach-se/7206/ |
H A D | setup.c | 22 .end = PA_SMSC + 0x300 + 0x020 - 1, 27 .end = 64, 57 .end = PA_LED,
|
/linux-4.4.14/arch/m68k/sun3/ |
H A D | dvma.c | 49 unsigned long end; dvma_map_iommu() local 54 end = vaddr + len; dvma_map_iommu() 56 while(vaddr < end) { dvma_map_iommu()
|
/linux-4.4.14/arch/ia64/kernel/ |
H A D | patch.c | 100 ia64_patch_vtop (unsigned long start, unsigned long end) ia64_patch_vtop() argument 105 while (offp < (s32 *) end) { ia64_patch_vtop() 123 ia64_patch_rse (unsigned long start, unsigned long end) ia64_patch_rse() argument 128 while (offp < (s32 *) end) { ia64_patch_rse() 141 ia64_patch_mckinley_e9 (unsigned long start, unsigned long end) ia64_patch_mckinley_e9() argument 158 while (offp < (s32 *) end) { ia64_patch_mckinley_e9() 172 patch_fsyscall_table (unsigned long start, unsigned long end) patch_fsyscall_table() argument 178 while (offp < (s32 *) end) { patch_fsyscall_table() 189 patch_brl_fsys_bubble_down (unsigned long start, unsigned long end) patch_brl_fsys_bubble_down() argument 195 while (offp < (s32 *) end) { patch_brl_fsys_bubble_down() 221 s32 * end = (s32 *) __end___phys_stack_reg_patchlist; ia64_patch_phys_stack_reg() local 228 while (offp < end) { ia64_patch_phys_stack_reg()
|
/linux-4.4.14/arch/avr32/kernel/ |
H A D | setup.c | 51 .end = 0, 57 .end = 0, 92 mem->start, mem->end); resource_init() 103 res->name, res->start, res->end); resource_init() 108 add_physical_memory(resource_size_t start, resource_size_t end) add_physical_memory() argument 114 if (end < next->start) add_physical_memory() 116 if (start <= next->end) { add_physical_memory() 121 start, end, next->start, next->end); add_physical_memory() 129 start, end); add_physical_memory() 135 new->end = end; add_physical_memory() 143 add_reserved_region(resource_size_t start, resource_size_t end, add_reserved_region() argument 148 if (end < start) add_reserved_region() 156 if (end < next->start) add_reserved_region() 158 if (start <= next->end) add_reserved_region() 164 new->end = end; add_reserved_region() 185 if (target <= res->end) find_free_region() 186 target = ALIGN(res->end + 1, align); find_free_region() 189 if ((target + size) > (mem->end + 1)) find_free_region() 190 return mem->end + 1; find_free_region() 205 if (target <= mem->end) { alloc_reserved_region() 300 system_ram->end = system_ram->start + size - 1; early_mem() 318 unsigned long start, end; parse_tag_mem() local 329 end = tag->u.mem_range.addr + tag->u.mem_range.size - 1; parse_tag_mem() 331 add_physical_memory(start, end); parse_tag_mem() 480 printk(" %08x-%08x\n", res->start, res->end); setup_bootmem() 484 res->start, res->end, res->name); setup_bootmem() 493 max_low_pfn = max_pfn = PFN_DOWN(res->end + 1); setup_bootmem() 543 && res->end < PFN_PHYS(max_pfn)) setup_bootmem() 568 kernel_code.end = __pa(init_mm.end_code - 1); setup_arch() 570 kernel_data.end = __pa(init_mm.brk - 1); setup_arch()
|