/linux-4.4.14/arch/x86/include/asm/ |
H A D | shmparam.h | 4 #define SHMLBA PAGE_SIZE /* attach addr a multiple of this */ macro
|
/linux-4.4.14/arch/alpha/include/asm/ |
H A D | shmparam.h | 4 #define SHMLBA PAGE_SIZE /* attach addr a multiple of this */ macro
|
/linux-4.4.14/include/uapi/asm-generic/ |
H A D | shmparam.h | 4 #define SHMLBA PAGE_SIZE /* attach addr a multiple of this */ macro
|
/linux-4.4.14/arch/powerpc/include/asm/ |
H A D | shmparam.h | 4 #define SHMLBA PAGE_SIZE /* attach addr a multiple of this */ macro
|
/linux-4.4.14/arch/sparc/include/asm/ |
H A D | shmparam_32.h | 7 #define SHMLBA (vac_cache_size ? vac_cache_size : PAGE_SIZE) macro
|
H A D | shmparam_64.h | 8 #define SHMLBA ((PAGE_SIZE > L1DCACHE_SIZE) ? PAGE_SIZE : L1DCACHE_SIZE) macro
|
/linux-4.4.14/arch/mn10300/include/asm/ |
H A D | shmparam.h | 4 #define SHMLBA PAGE_SIZE /* attach addr a multiple of this */ macro
|
/linux-4.4.14/arch/frv/include/asm/ |
H A D | shmparam.h | 4 #define SHMLBA PAGE_SIZE /* attach addr a multiple of this */ macro
|
/linux-4.4.14/arch/m32r/include/asm/ |
H A D | shmparam.h | 4 #define SHMLBA PAGE_SIZE /* attach addr a multiple of this */ macro
|
/linux-4.4.14/arch/avr32/include/asm/ |
H A D | shmparam.h | 4 #define SHMLBA PAGE_SIZE /* attach addr a multiple of this */ macro
|
/linux-4.4.14/arch/parisc/include/asm/ |
H A D | shmparam.h | 4 #define SHMLBA PAGE_SIZE /* attach addr a multiple of this */ macro
|
/linux-4.4.14/arch/cris/include/asm/ |
H A D | shmparam.h | 6 #define SHMLBA PAGE_SIZE /* attach addr a multiple of this */ macro
|
/linux-4.4.14/arch/ia64/include/asm/ |
H A D | shmparam.h | 5 * SHMLBA controls minimum alignment at which shared memory segments 8 * To reduce the chance of this, we set SHMLBA to 1MB. --davidm 00/12/20 10 #define SHMLBA (1024*1024) macro
|
/linux-4.4.14/arch/s390/include/asm/ |
H A D | shmparam.h | 9 #define SHMLBA PAGE_SIZE /* attach addr a multiple of this */ macro
|
/linux-4.4.14/arch/arc/include/asm/ |
H A D | shmparam.h | 13 #define SHMLBA (2 * PAGE_SIZE) macro 15 /* Enforce SHMLBA in shmat */
|
/linux-4.4.14/arch/arm/include/asm/ |
H A D | shmparam.h | 9 #define SHMLBA (4 * PAGE_SIZE) /* attach addr a multiple of this */ macro 12 * Enforce SHMLBA in shmat
|
H A D | cacheflush.h | 20 #define CACHE_COLOUR(vaddr) ((vaddr & (SHMLBA - 1)) >> PAGE_SHIFT)
|
/linux-4.4.14/arch/arc/mm/ |
H A D | mmap.c | 20 ((((addr) + SHMLBA - 1) & ~(SHMLBA - 1)) + \ 21 (((pgoff) << PAGE_SHIFT) & (SHMLBA - 1))) 28 * SHMLBA bytes. 51 (addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)) arch_get_unmapped_area() 75 info.align_mask = do_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; arch_get_unmapped_area()
|
/linux-4.4.14/arch/mips/include/asm/ |
H A D | shmparam.h | 11 #define SHMLBA 0x40000 /* attach addr a multiple of this */ macro
|
/linux-4.4.14/arch/xtensa/kernel/ |
H A D | syscall.c | 40 ((((addr) + SHMLBA - 1) & ~(SHMLBA - 1)) + \ 41 (((pgoff) << PAGE_SHIFT) & (SHMLBA - 1))) 48 err = do_shmat(shmid, shmaddr, shmflg, &ret, SHMLBA); xtensa_shmat() 71 ((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))) arch_get_unmapped_area()
|
/linux-4.4.14/arch/arm/mm/ |
H A D | mmap.c | 15 ((((addr)+SHMLBA-1)&~(SHMLBA-1)) + \ 16 (((pgoff)<<PAGE_SHIFT) & (SHMLBA-1))) 49 * SHMLBA bytes. 76 (addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)) arch_get_unmapped_area() 100 info.align_mask = do_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; arch_get_unmapped_area() 130 (addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1)) arch_get_unmapped_area_topdown() 151 info.align_mask = do_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; arch_get_unmapped_area_topdown()
|
H A D | copypage-v6.c | 23 #if SHMLBA > 16384
|
/linux-4.4.14/arch/xtensa/include/asm/ |
H A D | shmparam.h | 19 #define SHMLBA ((PAGE_SIZE > DCACHE_WAY_SIZE)? PAGE_SIZE : DCACHE_WAY_SIZE) macro
|
/linux-4.4.14/arch/sh/include/asm/ |
H A D | shmparam.h | 18 #define SHMLBA 0x4000 /* attach addr a multiple of this */ macro
|
/linux-4.4.14/arch/nios2/include/asm/ |
H A D | shmparam.h | 19 #define SHMLBA CONFIG_NIOS2_DCACHE_SIZE macro
|
/linux-4.4.14/arch/sparc/kernel/ |
H A D | sys_sparc_64.c | 80 unsigned long base = (addr+SHMLBA-1)&~(SHMLBA-1); COLOR_ALIGN() 81 unsigned long off = (pgoff<<PAGE_SHIFT) & (SHMLBA-1); COLOR_ALIGN() 99 ((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))) arch_get_unmapped_area() 129 info.align_mask = do_color_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; arch_get_unmapped_area() 163 ((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))) arch_get_unmapped_area_topdown() 192 info.align_mask = do_color_align ? (PAGE_MASK & (SHMLBA - 1)) : 0; arch_get_unmapped_area_topdown() 386 err = do_shmat(first, ptr, (int)second, &raddr, SHMLBA); SYSCALL_DEFINE6()
|
H A D | sys_sparc_32.c | 48 ((addr - (pgoff << PAGE_SHIFT)) & (SHMLBA - 1))) arch_get_unmapped_area() 64 (PAGE_MASK & (SHMLBA - 1)) : 0; arch_get_unmapped_area()
|
/linux-4.4.14/ipc/ |
H A D | syscall.c | 76 second, &raddr, SHMLBA); SYSCALL_DEFINE6()
|
H A D | compat.c | 325 #define COMPAT_SHMLBA SHMLBA
|
H A D | shm.c | 1248 err = do_shmat(shmid, shmaddr, shmflg, &ret, SHMLBA); SYSCALL_DEFINE3()
|
/linux-4.4.14/include/uapi/linux/ |
H A D | shm.h | 49 #define SHM_RND 020000 /* round attach address to SHMLBA boundary */
|
/linux-4.4.14/arch/ia64/kernel/ |
H A D | sys_ia64.c | 56 align_mask = PAGE_MASK & (SHMLBA - 1); arch_get_unmapped_area()
|
/linux-4.4.14/arch/unicore32/include/asm/ |
H A D | cacheflush.h | 19 #define CACHE_COLOUR(vaddr) ((vaddr & (SHMLBA - 1)) >> PAGE_SHIFT)
|
/linux-4.4.14/drivers/gpu/drm/ |
H A D | drm_bufs.c | 100 if (shm && (SHMLBA > PAGE_SIZE)) { drm_map_handle() 101 int bits = ilog2(SHMLBA >> PAGE_SHIFT) + 1; drm_map_handle() 103 /* For shared memory, we have to preserve the SHMLBA drm_map_handle() 111 * Therefore, make sure the SHMLBA relevant bits of the drm_map_handle()
|
/linux-4.4.14/mm/ |
H A D | vmalloc.c | 1778 ret = __vmalloc_node(size, SHMLBA, vmalloc_user()
|