/linux-4.1.27/tools/lib/lockdep/uinclude/linux/ |
H A D | rbtree_augmented.h | 1 #define __always_inline macro
|
/linux-4.1.27/arch/arm64/include/asm/ |
H A D | bitrev.h | 3 static __always_inline __attribute_const__ u32 __arch_bitrev32(u32 x) __arch_bitrev32() 9 static __always_inline __attribute_const__ u16 __arch_bitrev16(u16 x) __arch_bitrev16() 14 static __always_inline __attribute_const__ u8 __arch_bitrev8(u8 x) __arch_bitrev8()
|
H A D | jump_label.h | 29 static __always_inline bool arch_static_branch(struct static_key *key) arch_static_branch()
|
H A D | stackprotector.h | 26 static __always_inline void boot_init_stack_canary(void) boot_init_stack_canary()
|
H A D | arch_timer.h | 35 static __always_inline arch_timer_reg_write_cp15() 61 static __always_inline arch_timer_reg_read_cp15()
|
H A D | insn.h | 221 static __always_inline bool aarch64_insn_is_##abbr(u32 code) \ 223 static __always_inline u32 aarch64_insn_get_##abbr##_value(void) \
|
/linux-4.1.27/arch/arm/include/asm/ |
H A D | bitrev.h | 4 static __always_inline __attribute_const__ u32 __arch_bitrev32(u32 x) __arch_bitrev32() 10 static __always_inline __attribute_const__ u16 __arch_bitrev16(u16 x) __arch_bitrev16() 15 static __always_inline __attribute_const__ u8 __arch_bitrev8(u8 x) __arch_bitrev8()
|
H A D | jump_label.h | 16 static __always_inline bool arch_static_branch(struct static_key *key) arch_static_branch()
|
H A D | stacktrace.h | 17 static __always_inline arm_get_current_stackframe()
|
H A D | stackprotector.h | 26 static __always_inline void boot_init_stack_canary(void) boot_init_stack_canary()
|
H A D | arch_timer.h | 20 static __always_inline arch_timer_reg_write_cp15() 46 static __always_inline arch_timer_reg_read_cp15()
|
/linux-4.1.27/include/asm-generic/ |
H A D | preempt.h | 8 static __always_inline int preempt_count(void) preempt_count() 13 static __always_inline int *preempt_count_ptr(void) preempt_count_ptr() 18 static __always_inline void preempt_count_set(int pc) preempt_count_set() 34 static __always_inline void set_preempt_need_resched(void) set_preempt_need_resched() 38 static __always_inline void clear_preempt_need_resched(void) clear_preempt_need_resched() 42 static __always_inline bool test_preempt_need_resched(void) test_preempt_need_resched() 51 static __always_inline void __preempt_count_add(int val) __preempt_count_add() 56 static __always_inline void __preempt_count_sub(int val) __preempt_count_sub() 61 static __always_inline bool __preempt_count_dec_and_test(void) __preempt_count_dec_and_test() 74 static __always_inline bool should_resched(int preempt_offset) should_resched()
|
H A D | fixmap.h | 29 static __always_inline unsigned long fix_to_virt(const unsigned int idx) fix_to_virt()
|
/linux-4.1.27/arch/alpha/include/asm/ |
H A D | compiler.h | 14 #undef __always_inline macro 15 #define __always_inline inline __attribute__((always_inline)) macro
|
H A D | xchg.h | 108 static __always_inline unsigned long ____xchg() 240 static __always_inline unsigned long ____cmpxchg()
|
/linux-4.1.27/tools/include/linux/ |
H A D | compiler.h | 4 #ifndef __always_inline 5 # define __always_inline inline __attribute__((always_inline)) macro
|
/linux-4.1.27/arch/x86/include/asm/ |
H A D | preempt.h | 20 static __always_inline int preempt_count(void) preempt_count() 25 static __always_inline void preempt_count_set(int pc) preempt_count_set() 51 static __always_inline void set_preempt_need_resched(void) set_preempt_need_resched() 56 static __always_inline void clear_preempt_need_resched(void) clear_preempt_need_resched() 61 static __always_inline bool test_preempt_need_resched(void) test_preempt_need_resched() 70 static __always_inline void __preempt_count_add(int val) __preempt_count_add() 75 static __always_inline void __preempt_count_sub(int val) __preempt_count_sub() 85 static __always_inline bool __preempt_count_dec_and_test(void) __preempt_count_dec_and_test() 93 static __always_inline bool should_resched(int preempt_offset) should_resched()
|
H A D | current.h | 12 static __always_inline struct task_struct *get_current(void) get_current()
|
H A D | dmi.h | 10 static __always_inline __init void *dmi_alloc(unsigned len) dmi_alloc()
|
H A D | jump_label.h | 19 static __always_inline bool arch_static_branch(struct static_key *key) arch_static_branch()
|
H A D | smap.h | 49 static __always_inline void clac(void) clac() 55 static __always_inline void stac(void) stac()
|
H A D | uaccess_64.h | 26 static __always_inline __must_check unsigned long copy_user_generic() 51 static __always_inline __must_check __copy_from_user_nocheck() 94 static __always_inline __must_check __copy_from_user() 101 static __always_inline __must_check __copy_to_user_nocheck() 144 static __always_inline __must_check __copy_to_user() 151 static __always_inline __must_check __copy_in_user() 204 static __must_check __always_inline int __copy_from_user_inatomic() 210 static __must_check __always_inline int __copy_to_user_inatomic()
|
H A D | spinlock.h | 43 static __always_inline bool static_key_false(struct static_key *key); 53 static __always_inline void __ticket_lock_spinning(arch_spinlock_t *lock, __ticket_lock_spinning() 84 static __always_inline int arch_spin_value_unlocked(arch_spinlock_t lock) arch_spin_value_unlocked() 102 static __always_inline void arch_spin_lock(arch_spinlock_t *lock) arch_spin_lock() 127 static __always_inline int arch_spin_trylock(arch_spinlock_t *lock) arch_spin_trylock() 142 static __always_inline void arch_spin_unlock(arch_spinlock_t *lock) arch_spin_unlock() 176 static __always_inline void arch_spin_lock_flags(arch_spinlock_t *lock, arch_spin_lock_flags()
|
H A D | pvclock.h | 62 static __always_inline pvclock_get_nsec_offset() 70 static __always_inline __pvclock_read_cycles()
|
H A D | uaccess_32.h | 43 static __always_inline unsigned long __must_check __copy_to_user_inatomic() 81 static __always_inline unsigned long __must_check __copy_to_user() 88 static __always_inline unsigned long __copy_from_user_inatomic() 136 static __always_inline unsigned long __copy_from_user() 158 static __always_inline unsigned long __copy_from_user_nocache(void *to, __copy_from_user_nocache() 180 static __always_inline unsigned long __copy_from_user_inatomic_nocache()
|
H A D | string_32.h | 32 static __always_inline void *__memcpy(void *to, const void *from, size_t n) __memcpy() 51 static __always_inline void *__constant_memcpy(void *to, const void *from, __constant_memcpy() 226 static __always_inline __constant_c_memset() 256 static __always_inline __constant_c_and_count_memset()
|
H A D | tsc.h | 35 static __always_inline cycles_t vget_cycles(void) vget_cycles()
|
H A D | bitops.h | 8 * __always_inline to avoid problems with older gcc's inlining heuristics. 71 static __always_inline void set_bit() 109 static __always_inline void clear_bit() 216 static __always_inline int test_and_set_bit_lock() 308 static __always_inline int constant_test_bit(long nr, const volatile unsigned long *addr) constant_test_bit() 479 static __always_inline int fls64(__u64 x) fls64()
|
H A D | string_64.h | 9 static __always_inline void *__inline_memcpy(void *to, const void *from, size_t n) __inline_memcpy()
|
H A D | barrier.h | 99 static __always_inline void rdtsc_barrier(void) rdtsc_barrier()
|
H A D | fpu-internal.h | 116 static __always_inline __pure bool use_eager_fpu(void) use_eager_fpu() 121 static __always_inline __pure bool use_xsaveopt(void) use_xsaveopt() 126 static __always_inline __pure bool use_xsave(void) use_xsave() 131 static __always_inline __pure bool use_fxsr(void) use_fxsr()
|
H A D | stackprotector.h | 58 static __always_inline void boot_init_stack_canary(void) boot_init_stack_canary()
|
H A D | msr.h | 113 static __always_inline unsigned long long __native_read_tsc(void) __native_read_tsc()
|
H A D | cpufeature.h | 414 static __always_inline __pure bool __static_cpu_has(u16 bit) __static_cpu_has() 497 static __always_inline __pure bool _static_cpu_has_safe(u16 bit) _static_cpu_has_safe()
|
H A D | percpu.h | 520 static __always_inline int x86_this_cpu_constant_test_bit(unsigned int nr, x86_this_cpu_constant_test_bit()
|
H A D | paravirt.h | 715 static __always_inline void __ticket_lock_spinning(struct arch_spinlock *lock, __ticket_lock_spinning() 721 static __always_inline void __ticket_unlock_kick(struct arch_spinlock *lock, __ticket_unlock_kick()
|
/linux-4.1.27/include/asm-generic/bitops/ |
H A D | builtin-__ffs.h | 10 static __always_inline unsigned long __ffs(unsigned long word) __ffs()
|
H A D | builtin-fls.h | 11 static __always_inline int fls(int x) fls()
|
H A D | fls64.h | 18 static __always_inline int fls64(__u64 x) fls64() 26 static __always_inline int fls64(__u64 x) fls64()
|
H A D | builtin-__fls.h | 10 static __always_inline unsigned long __fls(unsigned long word) __fls()
|
H A D | builtin-ffs.h | 12 static __always_inline int ffs(int x) ffs()
|
H A D | fls.h | 12 static __always_inline int fls(int x) fls()
|
H A D | __ffs.h | 12 static __always_inline unsigned long __ffs(unsigned long word) __ffs()
|
H A D | __fls.h | 12 static __always_inline unsigned long __fls(unsigned long word) __fls()
|
/linux-4.1.27/arch/powerpc/include/asm/ |
H A D | cmpxchg.h | 15 static __always_inline unsigned long __xchg_u32() 40 static __always_inline unsigned long __xchg_u32_local() 58 static __always_inline unsigned long __xchg_u64() 77 static __always_inline unsigned long __xchg_u64_local() 101 static __always_inline unsigned long __xchg() 116 static __always_inline unsigned long __xchg_local() 149 static __always_inline unsigned long __cmpxchg_u32() 172 static __always_inline unsigned long __cmpxchg_u32_local() 195 static __always_inline unsigned long __cmpxchg_u64() 217 static __always_inline unsigned long __cmpxchg_u64_local() 243 static __always_inline unsigned long __cmpxchg() 259 static __always_inline unsigned long __cmpxchg_local()
|
H A D | jump_label.h | 21 static __always_inline bool arch_static_branch(struct static_key *key) arch_static_branch()
|
H A D | spinlock.h | 57 static __always_inline int arch_spin_value_unlocked(arch_spinlock_t lock) arch_spin_value_unlocked()
|
/linux-4.1.27/arch/sparc/include/asm/ |
H A D | jump_label.h | 10 static __always_inline bool arch_static_branch(struct static_key *key) arch_static_branch()
|
/linux-4.1.27/tools/include/asm-generic/bitops/ |
H A D | atomic.h | 16 static __always_inline int test_bit(unsigned int nr, const unsigned long *addr) test_bit()
|
H A D | __ffs.h | 12 static __always_inline unsigned long __ffs(unsigned long word) __ffs()
|
/linux-4.1.27/arch/s390/include/asm/ |
H A D | jump_label.h | 15 static __always_inline bool arch_static_branch(struct static_key *key) arch_static_branch()
|
H A D | irq.h | 79 static __always_inline void inc_irq_stat(enum interruption_class irq) inc_irq_stat()
|
/linux-4.1.27/arch/sh/include/asm/ |
H A D | stackprotector.h | 15 static __always_inline void boot_init_stack_canary(void) boot_init_stack_canary()
|
H A D | uaccess.h | 114 static __always_inline unsigned long __copy_from_user() 120 static __always_inline unsigned long __must_check __copy_to_user()
|
H A D | unaligned-sh4a.h | 34 static __always_inline u32 sh4a_get_unaligned_cpu32(const u8 *p) sh4a_get_unaligned_cpu32()
|
H A D | dwarf.h | 207 static __always_inline unsigned long dwarf_read_arch_reg(unsigned int reg) dwarf_read_arch_reg()
|
/linux-4.1.27/include/linux/ |
H A D | bottom_half.h | 10 static __always_inline void __local_bh_disable_ip(unsigned long ip, unsigned int cnt) __local_bh_disable_ip()
|
H A D | mm_inline.h | 25 static __always_inline void add_page_to_lru_list(struct page *page, add_page_to_lru_list() 34 static __always_inline void del_page_from_lru_list(struct page *page, del_page_from_lru_list() 65 static __always_inline enum lru_list page_off_lru(struct page *page) page_off_lru() 89 static __always_inline enum lru_list page_lru(struct page *page) page_lru()
|
H A D | jump_label.h | 123 static __always_inline bool static_key_false(struct static_key *key) static_key_false() 128 static __always_inline bool static_key_true(struct static_key *key) static_key_true() 157 static __always_inline void jump_label_init(void) jump_label_init() 162 static __always_inline bool static_key_false(struct static_key *key) static_key_false() 169 static __always_inline bool static_key_true(struct static_key *key) static_key_true()
|
H A D | slab.h | 246 static __always_inline int kmalloc_index(size_t size) kmalloc_index() 297 static __always_inline void *__kmalloc_node(size_t size, gfp_t flags, int node) __kmalloc_node() 302 static __always_inline void *kmem_cache_alloc_node(struct kmem_cache *s, gfp_t flags, int node) kmem_cache_alloc_node() 316 static __always_inline void * kmem_cache_alloc_node_trace() 326 static __always_inline void *kmem_cache_alloc_trace(struct kmem_cache *s, kmem_cache_alloc_trace() 335 static __always_inline void * kmem_cache_alloc_node_trace() 352 static __always_inline void * kmalloc_order_trace() 359 static __always_inline void *kmalloc_large(size_t size, gfp_t flags) kmalloc_large() 418 static __always_inline void *kmalloc(size_t size, gfp_t flags) kmalloc() 443 static __always_inline int kmalloc_size(int n) kmalloc_size() 458 static __always_inline void *kmalloc_node(size_t size, gfp_t flags, int node) kmalloc_node()
|
H A D | hash.h | 51 static __always_inline u64 hash_64(u64 val, unsigned int bits) hash_64()
|
H A D | compiler.h | 199 static __always_inline void __read_once_size(const volatile void *p, void *res, int size) __read_once_size() 213 static __always_inline void __write_once_size(volatile void *p, void *res, int size) __write_once_size() 331 #ifndef __always_inline 332 #define __always_inline inline macro 463 # define nokprobe_inline __always_inline
|
H A D | math64.h | 117 static __always_inline u32 __iter_div_u64_rem()
|
H A D | rbtree_augmented.h | 136 static __always_inline struct rb_node * __rb_erase_augmented() 233 static __always_inline void rb_erase_augmented()
|
H A D | memcontrol.h | 522 static __always_inline struct kmem_cache * memcg_kmem_get_cache() 539 static __always_inline void memcg_kmem_put_cache(struct kmem_cache *cachep) memcg_kmem_put_cache() 545 static __always_inline struct mem_cgroup *mem_cgroup_from_kmem(void *ptr) mem_cgroup_from_kmem()
|
H A D | time64.h | 182 static __always_inline void timespec64_add_ns(struct timespec64 *a, u64 ns) timespec64_add_ns()
|
H A D | nodemask.h | 111 * freeable items not being freed). So we must use __always_inline here 113 * this situation they will also need to be annotated as __always_inline 116 static __always_inline void __node_set(int node, volatile nodemask_t *dstp) __node_set()
|
H A D | radix-tree.h | 340 static __always_inline void ** radix_tree_iter_init() 394 static __always_inline long radix_tree_chunk_size() 411 static __always_inline void ** radix_tree_next_slot()
|
H A D | time.h | 233 static __always_inline void timespec_add_ns(struct timespec *a, u64 ns) timespec_add_ns()
|
H A D | async_tx.h | 30 #define __async_inline __always_inline
|
H A D | compiler-gcc.h | 245 #define __always_inline inline __attribute__((always_inline)) macro
|
H A D | perf_event.h | 770 static __always_inline void perf_sw_event() 784 static __always_inline void perf_sw_event_sched()
|
H A D | quota.h | 488 static __always_inline unsigned dquot_state_types(unsigned flags, unsigned flag) dquot_state_types()
|
H A D | sched.h | 2949 static __always_inline bool need_resched(void) need_resched()
|
H A D | mm.h | 923 static __always_inline void *lowmem_page_address(const struct page *page) lowmem_page_address()
|
/linux-4.1.27/arch/ia64/include/asm/ |
H A D | spinlock.h | 40 static __always_inline void __ticket_spin_lock(arch_spinlock_t *lock) __ticket_spin_lock() 60 static __always_inline int __ticket_spin_trylock(arch_spinlock_t *lock) __ticket_spin_trylock() 69 static __always_inline void __ticket_spin_unlock(arch_spinlock_t *lock) __ticket_spin_unlock() 77 static __always_inline void __ticket_spin_unlock_wait(arch_spinlock_t *lock) __ticket_spin_unlock_wait() 105 static __always_inline int arch_spin_value_unlocked(arch_spinlock_t lock) arch_spin_value_unlocked() 121 static __always_inline void arch_spin_lock(arch_spinlock_t *lock) arch_spin_lock() 126 static __always_inline int arch_spin_trylock(arch_spinlock_t *lock) arch_spin_trylock() 131 static __always_inline void arch_spin_unlock(arch_spinlock_t *lock) arch_spin_unlock() 136 static __always_inline void arch_spin_lock_flags(arch_spinlock_t *lock, arch_spin_lock_flags() 152 static __always_inline void arch_read_lock_flags() 203 static __always_inline void arch_write_lock_flags()
|
/linux-4.1.27/arch/xtensa/include/asm/ |
H A D | stacktrace.h | 20 static __always_inline unsigned long *stack_pointer(struct task_struct *task) stack_pointer()
|
H A D | fixmap.h | 60 static __always_inline unsigned long fix_to_virt(const unsigned int idx) fix_to_virt()
|
/linux-4.1.27/arch/mips/include/asm/ |
H A D | jump_label.h | 29 static __always_inline bool arch_static_branch(struct static_key *key) arch_static_branch()
|
H A D | stackprotector.h | 28 static __always_inline void boot_init_stack_canary(void) boot_init_stack_canary()
|
H A D | stacktrace.h | 23 static __always_inline void prepare_frametrace(struct pt_regs *regs) prepare_frametrace()
|
/linux-4.1.27/mm/kasan/ |
H A D | kasan.c | 69 static __always_inline bool memory_is_poisoned_1(unsigned long addr) memory_is_poisoned_1() 81 static __always_inline bool memory_is_poisoned_2(unsigned long addr) memory_is_poisoned_2() 98 static __always_inline bool memory_is_poisoned_4(unsigned long addr) memory_is_poisoned_4() 115 static __always_inline bool memory_is_poisoned_8(unsigned long addr) memory_is_poisoned_8() 132 static __always_inline bool memory_is_poisoned_16(unsigned long addr) memory_is_poisoned_16() 152 static __always_inline unsigned long bytes_is_zero(const u8 *start, bytes_is_zero() 165 static __always_inline unsigned long memory_is_zero(const void *start, memory_is_zero() 194 static __always_inline bool memory_is_poisoned_n(unsigned long addr, memory_is_poisoned_n() 213 static __always_inline bool memory_is_poisoned(unsigned long addr, size_t size) memory_is_poisoned() 236 static __always_inline void check_memory_region(unsigned long addr, check_memory_region()
|
/linux-4.1.27/drivers/infiniband/hw/mthca/ |
H A D | mthca_wqe.h | 116 static __always_inline void mthca_set_data_seg(struct mthca_data_seg *dseg, mthca_set_data_seg() 124 static __always_inline void mthca_set_data_seg_inval(struct mthca_data_seg *dseg) mthca_set_data_seg_inval()
|
H A D | mthca_qp.c | 1563 static __always_inline void set_raddr_seg(struct mthca_raddr_seg *rseg, set_raddr_seg() 1571 static __always_inline void set_atomic_seg(struct mthca_atomic_seg *aseg, set_atomic_seg()
|
/linux-4.1.27/arch/m32r/include/asm/ |
H A D | cmpxchg.h | 16 static __always_inline unsigned long __xchg() 70 static __always_inline unsigned long __xchg_local()
|
/linux-4.1.27/arch/openrisc/include/asm/ |
H A D | fixmap.h | 64 static __always_inline unsigned long fix_to_virt(const unsigned int idx) fix_to_virt()
|
/linux-4.1.27/net/core/ |
H A D | secure_seq.c | 19 static __always_inline void net_secret_init(void) net_secret_init()
|
H A D | flow_dissector.c | 272 static __always_inline void __flow_hash_secret_init(void) __flow_hash_secret_init() 277 static __always_inline u32 __flow_hash_3words(u32 a, u32 b, u32 c) __flow_hash_3words()
|
/linux-4.1.27/arch/x86/vdso/ |
H A D | vclock_gettime.c | 230 notrace static int __always_inline do_realtime(struct timespec *ts) do_realtime() 251 notrace static int __always_inline do_monotonic(struct timespec *ts) do_monotonic()
|
/linux-4.1.27/arch/m68k/include/asm/ |
H A D | uaccess_mm.h | 238 static __always_inline unsigned long __constant_copy_from_user() 319 static __always_inline unsigned long __constant_copy_to_user()
|
/linux-4.1.27/drivers/clocksource/ |
H A D | arm_arch_timer.c | 78 static __always_inline arch_timer_reg_write() 107 static __always_inline arch_timer_reg_read() 140 static __always_inline irqreturn_t timer_handler(const int access, timer_handler() 184 static __always_inline void timer_set_mode(const int access, int mode, timer_set_mode() 224 static __always_inline void set_next_event(const int access, unsigned long evt, set_next_event()
|
/linux-4.1.27/arch/sh/mm/ |
H A D | pmb.c | 72 static __always_inline unsigned long mk_pmb_entry(unsigned int entry) mk_pmb_entry() 77 static __always_inline unsigned long mk_pmb_addr(unsigned int entry) mk_pmb_addr() 82 static __always_inline unsigned long mk_pmb_data(unsigned int entry) mk_pmb_data() 87 static __always_inline unsigned int pmb_ppn_in_range(unsigned long ppn) pmb_ppn_in_range() 99 static __always_inline unsigned long pmb_cache_flags(void) pmb_cache_flags()
|
/linux-4.1.27/lib/ |
H A D | rbtree.c | 72 static __always_inline void __rb_insert() 201 static __always_inline void ____rb_erase_color() 404 * This instantiates the same __always_inline functions as in the non-augmented
|
H A D | radix-tree.c | 150 static __always_inline unsigned long radix_tree_find_next_bit()
|
/linux-4.1.27/kernel/locking/ |
H A D | qrwlock.c | 33 static __always_inline void rspin_until_writer_unlock()
|
H A D | mutex.c | 109 static __always_inline void ww_mutex_lock_acquired(struct ww_mutex *ww, ww_mutex_lock_acquired() 156 static __always_inline void ww_mutex_set_context_fastpath() 200 static __always_inline void ww_mutex_set_context_slowpath() 504 static __always_inline int __sched __mutex_lock_common()
|
/linux-4.1.27/mm/ |
H A D | slab.h | 233 static __always_inline int memcg_charge_slab(struct kmem_cache *s, memcg_charge_slab() 243 static __always_inline void memcg_uncharge_slab(struct kmem_cache *s, int order) memcg_uncharge_slab()
|
H A D | swap.c | 117 static __always_inline put_unrefcounted_compound_page() 165 static __always_inline put_refcounted_compound_page()
|
H A D | slob.c | 426 static __always_inline void * __do_kmalloc_node()
|
H A D | gup.c | 578 static __always_inline long __get_user_pages_locked(struct task_struct *tsk, __get_user_pages_locked() 715 __always_inline long __get_user_pages_unlocked(struct task_struct *tsk, struct mm_struct *mm, __get_user_pages_unlocked()
|
H A D | slab.c | 3152 static __always_inline void * slab_alloc_node() 3210 static __always_inline void * __do_cache_alloc() 3234 static __always_inline void * __do_cache_alloc() 3242 static __always_inline void * slab_alloc() 3482 static __always_inline void * __do_kmalloc_node() 3513 static __always_inline void *__do_kmalloc(size_t size, gfp_t flags, __do_kmalloc()
|
H A D | slub.c | 339 static __always_inline void slab_lock(struct page *page) slab_lock() 344 static __always_inline void slab_unlock(struct page *page) slab_unlock() 2425 static __always_inline void *slab_alloc_node(struct kmem_cache *s, slab_alloc_node() 2512 static __always_inline void *slab_alloc(struct kmem_cache *s, slab_alloc() 2701 static __always_inline void slab_free(struct kmem_cache *s, slab_free()
|
H A D | slab_common.c | 1072 static __always_inline void *__do_krealloc(const void *p, size_t new_size, __do_krealloc()
|
/linux-4.1.27/lib/xz/ |
H A D | xz_dec_lzma2.c | 478 static __always_inline void rc_normalize(struct rc_dec *rc) rc_normalize() 497 static __always_inline int rc_bit(struct rc_dec *rc, uint16_t *prob) rc_bit() 519 static __always_inline uint32_t rc_bittree(struct rc_dec *rc, rc_bittree() 535 static __always_inline void rc_bittree_reverse(struct rc_dec *rc, rc_bittree_reverse()
|
/linux-4.1.27/kernel/trace/ |
H A D | trace.h | 472 static __always_inline int trace_get_context_bit(void) trace_get_context_bit() 490 static __always_inline int trace_test_and_set_recursion(int start, int max) trace_test_and_set_recursion() 510 static __always_inline void trace_clear_recursion(int bit) trace_clear_recursion()
|
H A D | ring_buffer.c | 2680 static __always_inline int trace_recursive_lock() 2705 static __always_inline void trace_recursive_unlock() 2815 static __always_inline void rb_wakeups()
|
H A D | ftrace.c | 1210 static bool __always_inline ftrace_hash_empty(struct ftrace_hash *hash) ftrace_hash_empty()
|
/linux-4.1.27/drivers/md/ |
H A D | dm-switch.c | 360 static __always_inline unsigned long parse_hex(const char **string) parse_hex()
|
H A D | dm-cache-target.c | 557 __always_inline
|
/linux-4.1.27/arch/x86/kernel/cpu/ |
H A D | common.c | 279 static __always_inline void setup_smep(struct cpuinfo_x86 *c) setup_smep() 292 static __always_inline void setup_smap(struct cpuinfo_x86 *c) setup_smap()
|
/linux-4.1.27/kernel/time/ |
H A D | timekeeping.c | 413 static __always_inline u64 __ktime_get_fast_ns(struct tk_fast *tkf) __ktime_get_fast_ns() 1508 static __always_inline void timekeeping_apply_adjustment(struct timekeeper *tk, timekeeping_apply_adjustment() 1590 static __always_inline void timekeeping_freqadjust(struct timekeeper *tk, timekeeping_freqadjust()
|
/linux-4.1.27/drivers/md/bcache/ |
H A D | bset.h | 395 static __always_inline int64_t bkey_cmp(const struct bkey *l, bkey_cmp()
|
/linux-4.1.27/kernel/sched/ |
H A D | cputime.c | 257 static __always_inline bool steal_account_process_tick(void) steal_account_process_tick()
|
H A D | sched.h | 978 static __always_inline bool static_branch_##name(struct static_key *key) \ 1393 static __always_inline arch_scale_freq_capacity()
|
H A D | fair.c | 431 static __always_inline 2436 static __always_inline u64 decay_load(u64 val, u64 n) decay_load() 2521 static __always_inline int __update_entity_runnable_avg(u64 now, int cpu, __update_entity_runnable_avg() 3144 static __always_inline void return_cfs_rq_runtime(struct cfs_rq *cfs_rq); 3563 static __always_inline account_cfs_rq_runtime() 3910 static __always_inline void return_cfs_rq_runtime(struct cfs_rq *cfs_rq) return_cfs_rq_runtime() 4130 static __always_inline void return_cfs_rq_runtime(struct cfs_rq *cfs_rq) {} return_cfs_rq_runtime()
|
/linux-4.1.27/drivers/base/ |
H A D | devres.c | 84 static __always_inline struct devres * alloc_dr(dr_release_t release, alloc_dr()
|
/linux-4.1.27/fs/ |
H A D | namei.c | 682 static __always_inline void set_root(struct nameidata *nd) set_root() 689 static __always_inline unsigned set_root_rcu(struct nameidata *nd) set_root_rcu() 870 static __always_inline int follow_link()
|
/linux-4.1.27/arch/x86/kvm/ |
H A D | vmx.c | 1379 static __always_inline unsigned long vmcs_readl(unsigned long field) vmcs_readl() 1388 static __always_inline u16 vmcs_read16(unsigned long field) vmcs_read16() 1393 static __always_inline u32 vmcs_read32(unsigned long field) vmcs_read32() 1398 static __always_inline u64 vmcs_read64(unsigned long field) vmcs_read64()
|
H A D | emulate.c | 643 static __always_inline int __linearize(struct x86_emulate_ctxt *ctxt, __linearize() 833 static __always_inline int do_insn_fetch_bytes(struct x86_emulate_ctxt *ctxt, do_insn_fetch_bytes()
|
/linux-4.1.27/scripts/ |
H A D | kernel-doc | 2084 $prototype =~ s/^__always_inline +//;
|
H A D | checkpatch.pl | 300 our $Inline = qr{inline|__always_inline|noinline|__inline|__inline__};
|
/linux-4.1.27/drivers/infiniband/hw/mlx4/ |
H A D | qp.c | 2448 static __always_inline void set_raddr_seg(struct mlx4_wqe_raddr_seg *rseg, set_raddr_seg()
|
/linux-4.1.27/drivers/infiniband/hw/mlx5/ |
H A D | qp.c | 1842 static __always_inline void set_raddr_seg(struct mlx5_wqe_raddr_seg *rseg, set_raddr_seg()
|
/linux-4.1.27/drivers/net/wireless/brcm80211/brcmfmac/ |
H A D | cfg80211.c | 2931 static __always_inline void brcmf_delay(u32 ms) brcmf_delay()
|