Lines Matching refs:size

170 extern void __iomem * __ioremap(phys_addr_t offset, phys_addr_t size, unsigned long flags);
178 static inline void __iomem * __ioremap_mode(phys_addr_t offset, unsigned long size, in __ioremap_mode() argument
181 void __iomem *addr = plat_ioremap(offset, size, flags); in __ioremap_mode()
199 __builtin_constant_p(size) && __builtin_constant_p(flags)) { in __ioremap_mode()
202 phys_addr = fixup_bigphys_addr(offset, size); in __ioremap_mode()
205 last_addr = phys_addr + size - 1; in __ioremap_mode()
206 if (!size || last_addr < phys_addr) in __ioremap_mode()
219 return __ioremap(offset, size, flags); in __ioremap_mode()
235 #define ioremap(offset, size) \ argument
236 __ioremap_mode((offset), (size), _CACHE_UNCACHED)
257 #define ioremap_nocache(offset, size) \ argument
258 __ioremap_mode((offset), (size), _CACHE_UNCACHED)
275 #define ioremap_cachable(offset, size) \ argument
276 __ioremap_mode((offset), (size), _page_cachable_default)
284 #define ioremap_cacheable_cow(offset, size) \ argument
285 __ioremap_mode((offset), (size), _CACHE_CACHABLE_COW)
286 #define ioremap_uncached_accelerated(offset, size) \ argument
287 __ioremap_mode((offset), (size), _CACHE_UNCACHED_ACCELERATED)
589 extern void (*_dma_cache_wback_inv)(unsigned long start, unsigned long size);
590 extern void (*_dma_cache_wback)(unsigned long start, unsigned long size);
591 extern void (*_dma_cache_inv)(unsigned long start, unsigned long size);
593 #define dma_cache_wback_inv(start, size) _dma_cache_wback_inv(start, size) argument
594 #define dma_cache_wback(start, size) _dma_cache_wback(start, size) argument
595 #define dma_cache_inv(start, size) _dma_cache_inv(start, size) argument
599 #define dma_cache_wback_inv(start,size) \ argument
600 do { (void) (start); (void) (size); } while (0)
601 #define dma_cache_wback(start,size) \ argument
602 do { (void) (start); (void) (size); } while (0)
603 #define dma_cache_inv(start,size) \ argument
604 do { (void) (start); (void) (size); } while (0)