Lines Matching refs:size
31 int ioremap_change_attr(unsigned long vaddr, unsigned long size, in ioremap_change_attr() argument
34 unsigned long nrpages = size >> PAGE_SHIFT; in ioremap_change_attr()
83 unsigned long size, enum page_cache_mode pcm, void *caller) in __ioremap_caller() argument
88 const unsigned long unaligned_size = size; in __ioremap_caller()
97 last_addr = phys_addr + size - 1; in __ioremap_caller()
98 if (!size || last_addr < phys_addr) in __ioremap_caller()
118 ram_region = region_is_ram(phys_addr, size); in __ioremap_caller()
139 size = PAGE_ALIGN(last_addr+1) - phys_addr; in __ioremap_caller()
141 retval = reserve_memtype(phys_addr, (u64)phys_addr + size, in __ioremap_caller()
149 if (!is_new_memtype_allowed(phys_addr, size, pcm, new_pcm)) { in __ioremap_caller()
153 (unsigned long long)(phys_addr + size), in __ioremap_caller()
182 area = get_vm_area_caller(size, VM_IOREMAP, caller); in __ioremap_caller()
188 if (kernel_map_sync_memtype(phys_addr, size, pcm)) in __ioremap_caller()
191 if (ioremap_page_range(vaddr, vaddr + size, phys_addr, prot)) in __ioremap_caller()
208 free_memtype(phys_addr, phys_addr + size); in __ioremap_caller()
233 void __iomem *ioremap_nocache(resource_size_t phys_addr, unsigned long size) in ioremap_nocache() argument
244 return __ioremap_caller(phys_addr, size, pcm, in ioremap_nocache()
259 void __iomem *ioremap_wc(resource_size_t phys_addr, unsigned long size) in ioremap_wc() argument
262 return __ioremap_caller(phys_addr, size, _PAGE_CACHE_MODE_WC, in ioremap_wc()
265 return ioremap_nocache(phys_addr, size); in ioremap_wc()
269 void __iomem *ioremap_cache(resource_size_t phys_addr, unsigned long size) in ioremap_cache() argument
271 return __ioremap_caller(phys_addr, size, _PAGE_CACHE_MODE_WB, in ioremap_cache()
276 void __iomem *ioremap_prot(resource_size_t phys_addr, unsigned long size, in ioremap_prot() argument
279 return __ioremap_caller(phys_addr, size, in ioremap_prot()