/linux-4.4.14/tools/lib/lockdep/uinclude/linux/ |
H A D | rbtree_augmented.h | 1 #define __always_inline macro
|
/linux-4.4.14/arch/arm64/include/asm/ |
H A D | bitrev.h | 3 static __always_inline __attribute_const__ u32 __arch_bitrev32(u32 x) __arch_bitrev32() 9 static __always_inline __attribute_const__ u16 __arch_bitrev16(u16 x) __arch_bitrev16() 14 static __always_inline __attribute_const__ u8 __arch_bitrev8(u8 x) __arch_bitrev8()
|
H A D | jump_label.h | 29 static __always_inline bool arch_static_branch(struct static_key *key, bool branch) arch_static_branch() 43 static __always_inline bool arch_static_branch_jump(struct static_key *key, bool branch) arch_static_branch_jump()
|
H A D | stackprotector.h | 26 static __always_inline void boot_init_stack_canary(void) boot_init_stack_canary()
|
H A D | arch_timer.h | 35 static __always_inline arch_timer_reg_write_cp15() 61 static __always_inline arch_timer_reg_read_cp15()
|
H A D | insn.h | 221 static __always_inline bool aarch64_insn_is_##abbr(u32 code) \ 223 static __always_inline u32 aarch64_insn_get_##abbr##_value(void) \
|
/linux-4.4.14/arch/arm/include/asm/ |
H A D | bitrev.h | 4 static __always_inline __attribute_const__ u32 __arch_bitrev32(u32 x) __arch_bitrev32() 10 static __always_inline __attribute_const__ u16 __arch_bitrev16(u16 x) __arch_bitrev16() 15 static __always_inline __attribute_const__ u8 __arch_bitrev8(u8 x) __arch_bitrev8()
|
H A D | jump_label.h | 11 static __always_inline bool arch_static_branch(struct static_key *key, bool branch) arch_static_branch() 25 static __always_inline bool arch_static_branch_jump(struct static_key *key, bool branch) arch_static_branch_jump()
|
H A D | stacktrace.h | 17 static __always_inline arm_get_current_stackframe()
|
H A D | stackprotector.h | 26 static __always_inline void boot_init_stack_canary(void) boot_init_stack_canary()
|
H A D | arch_timer.h | 20 static __always_inline arch_timer_reg_write_cp15() 46 static __always_inline arch_timer_reg_read_cp15()
|
/linux-4.4.14/include/asm-generic/ |
H A D | preempt.h | 8 static __always_inline int preempt_count(void) preempt_count() 13 static __always_inline int *preempt_count_ptr(void) preempt_count_ptr() 18 static __always_inline void preempt_count_set(int pc) preempt_count_set() 34 static __always_inline void set_preempt_need_resched(void) set_preempt_need_resched() 38 static __always_inline void clear_preempt_need_resched(void) clear_preempt_need_resched() 42 static __always_inline bool test_preempt_need_resched(void) test_preempt_need_resched() 51 static __always_inline void __preempt_count_add(int val) __preempt_count_add() 56 static __always_inline void __preempt_count_sub(int val) __preempt_count_sub() 61 static __always_inline bool __preempt_count_dec_and_test(void) __preempt_count_dec_and_test() 74 static __always_inline bool should_resched(int preempt_offset) should_resched()
|
H A D | qspinlock.h | 28 static __always_inline int queued_spin_is_locked(struct qspinlock *lock) queued_spin_is_locked() 66 static __always_inline int queued_spin_value_unlocked(struct qspinlock lock) queued_spin_value_unlocked() 76 static __always_inline int queued_spin_is_contended(struct qspinlock *lock) queued_spin_is_contended() 85 static __always_inline int queued_spin_trylock(struct qspinlock *lock) queued_spin_trylock() 99 static __always_inline void queued_spin_lock(struct qspinlock *lock) queued_spin_lock() 114 static __always_inline void queued_spin_unlock(struct qspinlock *lock) queued_spin_unlock() 140 static __always_inline bool virt_spin_lock(struct qspinlock *lock) virt_spin_lock()
|
H A D | fixmap.h | 29 static __always_inline unsigned long fix_to_virt(const unsigned int idx) fix_to_virt()
|
/linux-4.4.14/arch/alpha/include/asm/ |
H A D | compiler.h | 14 #undef __always_inline macro 15 #define __always_inline inline __attribute__((always_inline)) macro
|
H A D | xchg.h | 108 static __always_inline unsigned long ____xchg() 240 static __always_inline unsigned long ____cmpxchg()
|
/linux-4.4.14/arch/x86/include/asm/ |
H A D | preempt.h | 20 static __always_inline int preempt_count(void) preempt_count() 25 static __always_inline void preempt_count_set(int pc) preempt_count_set() 48 static __always_inline void set_preempt_need_resched(void) set_preempt_need_resched() 53 static __always_inline void clear_preempt_need_resched(void) clear_preempt_need_resched() 58 static __always_inline bool test_preempt_need_resched(void) test_preempt_need_resched() 67 static __always_inline void __preempt_count_add(int val) __preempt_count_add() 72 static __always_inline void __preempt_count_sub(int val) __preempt_count_sub() 82 static __always_inline bool __preempt_count_dec_and_test(void) __preempt_count_dec_and_test() 90 static __always_inline bool should_resched(int preempt_offset) should_resched()
|
H A D | current.h | 12 static __always_inline struct task_struct *get_current(void) get_current()
|
H A D | atomic.h | 25 static __always_inline int atomic_read(const atomic_t *v) atomic_read() 37 static __always_inline void atomic_set(atomic_t *v, int i) atomic_set() 49 static __always_inline void atomic_add(int i, atomic_t *v) atomic_add() 63 static __always_inline void atomic_sub(int i, atomic_t *v) atomic_sub() 79 static __always_inline int atomic_sub_and_test(int i, atomic_t *v) atomic_sub_and_test() 90 static __always_inline void atomic_inc(atomic_t *v) atomic_inc() 102 static __always_inline void atomic_dec(atomic_t *v) atomic_dec() 116 static __always_inline int atomic_dec_and_test(atomic_t *v) atomic_dec_and_test() 129 static __always_inline int atomic_inc_and_test(atomic_t *v) atomic_inc_and_test() 143 static __always_inline int atomic_add_negative(int i, atomic_t *v) atomic_add_negative() 155 static __always_inline int atomic_add_return(int i, atomic_t *v) atomic_add_return() 167 static __always_inline int atomic_sub_return(int i, atomic_t *v) atomic_sub_return() 175 static __always_inline int atomic_cmpxchg(atomic_t *v, int old, int new) atomic_cmpxchg() 209 static __always_inline int __atomic_add_unless(atomic_t *v, int a, int u) __atomic_add_unless() 231 static __always_inline short int atomic_inc_short(short int *v) atomic_inc_short()
|
H A D | jump_label.h | 19 static __always_inline bool arch_static_branch(struct static_key *key, bool branch) arch_static_branch() 34 static __always_inline bool arch_static_branch_jump(struct static_key *key, bool branch) arch_static_branch_jump()
|
H A D | dmi.h | 10 static __always_inline __init void *dmi_alloc(unsigned len) dmi_alloc()
|
H A D | arch_hweight.h | 24 static __always_inline unsigned int __arch_hweight32(unsigned int w) __arch_hweight32() 52 static __always_inline unsigned long __arch_hweight64(__u64 w) __arch_hweight64()
|
H A D | pvclock.h | 62 static __always_inline pvclock_get_nsec_offset() 70 static __always_inline __pvclock_read_cycles()
|
H A D | smap.h | 49 static __always_inline void clac(void) clac() 55 static __always_inline void stac(void) stac()
|
H A D | uaccess_64.h | 26 static __always_inline __must_check unsigned long copy_user_generic() 51 static __always_inline __must_check __copy_from_user_nocheck() 94 static __always_inline __must_check __copy_from_user() 101 static __always_inline __must_check __copy_to_user_nocheck() 144 static __always_inline __must_check __copy_to_user() 151 static __always_inline __must_check __copy_in_user() 204 static __must_check __always_inline int __copy_from_user_inatomic() 210 static __must_check __always_inline int __copy_to_user_inatomic()
|
H A D | spinlock.h | 43 static __always_inline bool static_key_false(struct static_key *key); 57 static __always_inline void __ticket_lock_spinning(arch_spinlock_t *lock, __ticket_lock_spinning() 88 static __always_inline int arch_spin_value_unlocked(arch_spinlock_t lock) arch_spin_value_unlocked() 106 static __always_inline void arch_spin_lock(arch_spinlock_t *lock) arch_spin_lock() 131 static __always_inline int arch_spin_trylock(arch_spinlock_t *lock) arch_spin_trylock() 146 static __always_inline void arch_spin_unlock(arch_spinlock_t *lock) arch_spin_unlock() 180 static __always_inline void arch_spin_lock_flags(arch_spinlock_t *lock, arch_spin_lock_flags()
|
H A D | atomic64_64.h | 43 static __always_inline void atomic64_add(long i, atomic64_t *v) atomic64_add() 84 static __always_inline void atomic64_inc(atomic64_t *v) atomic64_inc() 97 static __always_inline void atomic64_dec(atomic64_t *v) atomic64_dec() 151 static __always_inline long atomic64_add_return(long i, atomic64_t *v) atomic64_add_return()
|
H A D | uaccess_32.h | 43 static __always_inline unsigned long __must_check __copy_to_user_inatomic() 86 static __always_inline unsigned long __must_check __copy_to_user() 93 static __always_inline unsigned long __copy_from_user_inatomic() 142 static __always_inline unsigned long __copy_from_user() 164 static __always_inline unsigned long __copy_from_user_nocache(void *to, __copy_from_user_nocache() 186 static __always_inline unsigned long __copy_from_user_inatomic_nocache()
|
H A D | string_32.h | 32 static __always_inline void *__memcpy(void *to, const void *from, size_t n) __memcpy() 51 static __always_inline void *__constant_memcpy(void *to, const void *from, __constant_memcpy() 226 static __always_inline __constant_c_memset() 256 static __always_inline __constant_c_and_count_memset()
|
H A D | bitops.h | 8 * __always_inline to avoid problems with older gcc's inlining heuristics. 71 static __always_inline void set_bit() 109 static __always_inline void clear_bit() 216 static __always_inline int test_and_set_bit_lock() 308 static __always_inline int constant_test_bit(long nr, const volatile unsigned long *addr) constant_test_bit() 479 static __always_inline int fls64(__u64 x) fls64()
|
H A D | string_64.h | 9 static __always_inline void *__inline_memcpy(void *to, const void *from, size_t n) __inline_memcpy()
|
H A D | paravirt.h | 693 static __always_inline void pv_queued_spin_lock_slowpath(struct qspinlock *lock, pv_queued_spin_lock_slowpath() 699 static __always_inline void pv_queued_spin_unlock(struct qspinlock *lock) pv_queued_spin_unlock() 704 static __always_inline void pv_wait(u8 *ptr, u8 val) pv_wait() 709 static __always_inline void pv_kick(int cpu) pv_kick() 716 static __always_inline void __ticket_lock_spinning(struct arch_spinlock *lock, __ticket_lock_spinning() 722 static __always_inline void __ticket_unlock_kick(struct arch_spinlock *lock, __ticket_unlock_kick()
|
H A D | stackprotector.h | 60 static __always_inline void boot_init_stack_canary(void) boot_init_stack_canary()
|
H A D | msr.h | 120 static __always_inline unsigned long long rdtsc(void) rdtsc() 137 static __always_inline unsigned long long rdtsc_ordered(void) rdtsc_ordered()
|
H A D | cpufeature.h | 421 static __always_inline __pure bool __static_cpu_has(u16 bit) __static_cpu_has() 504 static __always_inline __pure bool _static_cpu_has_safe(u16 bit) _static_cpu_has_safe()
|
H A D | processor.h | 560 static __always_inline void rep_nop(void) rep_nop() 565 static __always_inline void cpu_relax(void) cpu_relax()
|
H A D | percpu.h | 520 static __always_inline int x86_this_cpu_constant_test_bit(unsigned int nr, x86_this_cpu_constant_test_bit()
|
/linux-4.4.14/include/asm-generic/bitops/ |
H A D | builtin-__ffs.h | 10 static __always_inline unsigned long __ffs(unsigned long word) __ffs()
|
H A D | builtin-fls.h | 11 static __always_inline int fls(int x) fls()
|
H A D | fls64.h | 18 static __always_inline int fls64(__u64 x) fls64() 26 static __always_inline int fls64(__u64 x) fls64()
|
H A D | builtin-__fls.h | 10 static __always_inline unsigned long __fls(unsigned long word) __fls()
|
H A D | builtin-ffs.h | 12 static __always_inline int ffs(int x) ffs()
|
H A D | fls.h | 12 static __always_inline int fls(int x) fls()
|
H A D | __ffs.h | 12 static __always_inline unsigned long __ffs(unsigned long word) __ffs()
|
H A D | __fls.h | 12 static __always_inline unsigned long __fls(unsigned long word) __fls()
|
/linux-4.4.14/arch/sparc/include/asm/ |
H A D | jump_label.h | 10 static __always_inline bool arch_static_branch(struct static_key *key, bool branch) arch_static_branch() 26 static __always_inline bool arch_static_branch_jump(struct static_key *key, bool branch) arch_static_branch_jump()
|
/linux-4.4.14/arch/powerpc/include/asm/ |
H A D | cmpxchg.h | 15 static __always_inline unsigned long __xchg_u32() 40 static __always_inline unsigned long __xchg_u32_local() 58 static __always_inline unsigned long __xchg_u64() 77 static __always_inline unsigned long __xchg_u64_local() 101 static __always_inline unsigned long __xchg() 116 static __always_inline unsigned long __xchg_local() 148 static __always_inline unsigned long __cmpxchg_u32() 171 static __always_inline unsigned long __cmpxchg_u32_local() 194 static __always_inline unsigned long __cmpxchg_u64() 216 static __always_inline unsigned long __cmpxchg_u64_local() 242 static __always_inline unsigned long __cmpxchg() 258 static __always_inline unsigned long __cmpxchg_local()
|
H A D | jump_label.h | 21 static __always_inline bool arch_static_branch(struct static_key *key, bool branch) arch_static_branch() 35 static __always_inline bool arch_static_branch_jump(struct static_key *key, bool branch) arch_static_branch_jump()
|
H A D | spinlock.h | 55 static __always_inline int arch_spin_value_unlocked(arch_spinlock_t lock) arch_spin_value_unlocked()
|
/linux-4.4.14/arch/s390/include/asm/ |
H A D | jump_label.h | 15 static __always_inline bool arch_static_branch(struct static_key *key, bool branch) arch_static_branch() 29 static __always_inline bool arch_static_branch_jump(struct static_key *key, bool branch) arch_static_branch_jump()
|
H A D | irq.h | 78 static __always_inline void inc_irq_stat(enum interruption_class irq) inc_irq_stat()
|
/linux-4.4.14/tools/include/asm-generic/bitops/ |
H A D | atomic.h | 16 static __always_inline int test_bit(unsigned int nr, const unsigned long *addr) test_bit()
|
H A D | __ffs.h | 12 static __always_inline unsigned long __ffs(unsigned long word) __ffs()
|
/linux-4.4.14/drivers/misc/mic/scif/ |
H A D | scif_map.h | 23 static __always_inline void * scif_alloc_coherent() 43 static __always_inline void scif_free_coherent() 57 static __always_inline int scif_map_single() 78 static __always_inline void scif_unmap_single() 90 static __always_inline void * scif_ioremap() 104 static __always_inline void scif_iounmap() 114 static __always_inline int scif_map_page()
|
H A D | scif_nodeqp.c | 284 static __always_inline void scif_send_msg_intr() 710 static __always_inline void scif_init() 737 static __always_inline void scif_exit() 755 static __always_inline void scif_exit_ack() 776 static __always_inline void scif_node_add() 894 static __always_inline void scif_node_add_ack() 946 static __always_inline void scif_node_add_nack() 964 static __always_inline void scif_node_remove() 980 static __always_inline void scif_node_remove_ack() 995 static __always_inline void scif_get_node_info_resp()
|
H A D | scif_fd.c | 71 static __always_inline void scif_err_debug(int err, const char *str) scif_err_debug()
|
/linux-4.4.14/arch/mips/include/asm/ |
H A D | jump_label.h | 29 static __always_inline bool arch_static_branch(struct static_key *key, bool branch) arch_static_branch() 43 static __always_inline bool arch_static_branch_jump(struct static_key *key, bool branch) arch_static_branch_jump()
|
H A D | stackprotector.h | 28 static __always_inline void boot_init_stack_canary(void) boot_init_stack_canary()
|
H A D | stacktrace.h | 23 static __always_inline void prepare_frametrace(struct pt_regs *regs) prepare_frametrace()
|
/linux-4.4.14/include/linux/ |
H A D | bottom_half.h | 9 static __always_inline void __local_bh_disable_ip(unsigned long ip, unsigned int cnt) __local_bh_disable_ip()
|
H A D | spinlock.h | 289 static __always_inline raw_spinlock_t *spinlock_check(spinlock_t *lock) spinlock_check() 300 static __always_inline void spin_lock(spinlock_t *lock) spin_lock() 305 static __always_inline void spin_lock_bh(spinlock_t *lock) spin_lock_bh() 310 static __always_inline int spin_trylock(spinlock_t *lock) spin_trylock() 330 static __always_inline void spin_lock_irq(spinlock_t *lock) spin_lock_irq() 345 static __always_inline void spin_unlock(spinlock_t *lock) spin_unlock() 350 static __always_inline void spin_unlock_bh(spinlock_t *lock) spin_unlock_bh() 355 static __always_inline void spin_unlock_irq(spinlock_t *lock) spin_unlock_irq() 360 static __always_inline void spin_unlock_irqrestore(spinlock_t *lock, unsigned long flags) spin_unlock_irqrestore() 365 static __always_inline int spin_trylock_bh(spinlock_t *lock) spin_trylock_bh() 370 static __always_inline int spin_trylock_irq(spinlock_t *lock) spin_trylock_irq() 380 static __always_inline void spin_unlock_wait(spinlock_t *lock) spin_unlock_wait() 385 static __always_inline int spin_is_locked(spinlock_t *lock) spin_is_locked() 390 static __always_inline int spin_is_contended(spinlock_t *lock) spin_is_contended() 395 static __always_inline int spin_can_lock(spinlock_t *lock) spin_can_lock()
|
H A D | mm_inline.h | 25 static __always_inline void add_page_to_lru_list(struct page *page, add_page_to_lru_list() 34 static __always_inline void del_page_from_lru_list(struct page *page, del_page_from_lru_list() 65 static __always_inline enum lru_list page_off_lru(struct page *page) page_off_lru() 89 static __always_inline enum lru_list page_lru(struct page *page) page_lru()
|
H A D | rbtree_latch.h | 67 static __always_inline struct latch_tree_node * __lt_from_rb() 73 static __always_inline void __lt_insert() 97 static __always_inline void __lt_erase() 103 static __always_inline struct latch_tree_node * __lt_find() 141 static __always_inline void latch_tree_insert() 168 static __always_inline void latch_tree_erase() 197 static __always_inline struct latch_tree_node * latch_tree_find()
|
H A D | uaccess.h | 7 static __always_inline void pagefault_disabled_inc(void) pagefault_disabled_inc() 12 static __always_inline void pagefault_disabled_dec(void) pagefault_disabled_dec()
|
H A D | slab.h | 264 static __always_inline int kmalloc_index(size_t size) kmalloc_index() 325 static __always_inline void *__kmalloc_node(size_t size, gfp_t flags, int node) __kmalloc_node() 330 static __always_inline void *kmem_cache_alloc_node(struct kmem_cache *s, gfp_t flags, int node) kmem_cache_alloc_node() 344 static __always_inline void * kmem_cache_alloc_node_trace() 354 static __always_inline void *kmem_cache_alloc_trace(struct kmem_cache *s, kmem_cache_alloc_trace() 363 static __always_inline void * kmem_cache_alloc_node_trace() 380 static __always_inline void * kmalloc_order_trace() 387 static __always_inline void *kmalloc_large(size_t size, gfp_t flags) kmalloc_large() 446 static __always_inline void *kmalloc(size_t size, gfp_t flags) kmalloc() 471 static __always_inline int kmalloc_size(int n) kmalloc_size() 486 static __always_inline void *kmalloc_node(size_t size, gfp_t flags, int node) kmalloc_node()
|
H A D | jump_label.h | 131 static __always_inline bool static_key_false(struct static_key *key) static_key_false() 136 static __always_inline bool static_key_true(struct static_key *key) static_key_true() 165 static __always_inline void jump_label_init(void) jump_label_init() 170 static __always_inline bool static_key_false(struct static_key *key) static_key_false() 177 static __always_inline bool static_key_true(struct static_key *key) static_key_true()
|
H A D | compiler.h | 215 static __always_inline __read_once_size() 234 static __always_inline __read_once_size_nocheck() 241 static __always_inline void __write_once_size(volatile void *p, void *res, int size) __write_once_size() 378 #ifndef __always_inline 379 #define __always_inline inline macro 533 # define nokprobe_inline __always_inline
|
H A D | hash.h | 51 static __always_inline u64 hash_64(u64 val, unsigned int bits) hash_64()
|
H A D | rbtree_augmented.h | 136 static __always_inline struct rb_node * __rb_erase_augmented() 240 static __always_inline void rb_erase_augmented()
|
H A D | time64.h | 219 static __always_inline void timespec64_add_ns(struct timespec64 *a, u64 ns) timespec64_add_ns()
|
H A D | memcontrol.h | 791 static __always_inline int memcg_kmem_charge(struct page *page, memcg_kmem_charge() 804 static __always_inline void memcg_kmem_uncharge(struct page *page, int order) memcg_kmem_uncharge() 817 static __always_inline struct kmem_cache * memcg_kmem_get_cache() 825 static __always_inline void memcg_kmem_put_cache(struct kmem_cache *cachep) memcg_kmem_put_cache()
|
H A D | nodemask.h | 111 * freeable items not being freed). So we must use __always_inline here 113 * this situation they will also need to be annotated as __always_inline 116 static __always_inline void __node_set(int node, volatile nodemask_t *dstp) __node_set()
|
H A D | radix-tree.h | 340 static __always_inline void ** radix_tree_iter_init() 394 static __always_inline long radix_tree_chunk_size() 411 static __always_inline void ** radix_tree_next_slot()
|
H A D | time.h | 233 static __always_inline void timespec_add_ns(struct timespec *a, u64 ns) timespec_add_ns()
|
H A D | async_tx.h | 30 #define __async_inline __always_inline
|
H A D | bitops.h | 78 static __always_inline unsigned long hweight_long(unsigned long w) hweight_long()
|
H A D | math64.h | 117 static __always_inline u32 __iter_div_u64_rem()
|
H A D | jiffies.h | 354 static __always_inline unsigned long msecs_to_jiffies(const unsigned int m) msecs_to_jiffies() 401 static __always_inline unsigned long usecs_to_jiffies(const unsigned int u) usecs_to_jiffies()
|
H A D | compiler-gcc.h | 83 #define __always_inline inline __attribute__((always_inline)) macro
|
H A D | perf_event.h | 886 static __always_inline void perf_sw_event() 900 static __always_inline void perf_sw_event_sched() 913 static __always_inline bool perf_sw_migrate_enabled()
|
H A D | bitmap.h | 298 static __always_inline int bitmap_weight(const unsigned long *src, unsigned int nbits) bitmap_weight()
|
H A D | quota.h | 488 static __always_inline unsigned dquot_state_types(unsigned flags, unsigned flag) dquot_state_types()
|
H A D | mm.h | 906 static __always_inline void *lowmem_page_address(const struct page *page) lowmem_page_address()
|
H A D | sched.h | 3052 static __always_inline bool need_resched(void) need_resched()
|
/linux-4.4.14/arch/sh/include/asm/ |
H A D | stackprotector.h | 15 static __always_inline void boot_init_stack_canary(void) boot_init_stack_canary()
|
H A D | uaccess.h | 114 static __always_inline unsigned long __copy_from_user() 120 static __always_inline unsigned long __must_check __copy_to_user()
|
H A D | unaligned-sh4a.h | 34 static __always_inline u32 sh4a_get_unaligned_cpu32(const u8 *p) sh4a_get_unaligned_cpu32()
|
H A D | dwarf.h | 207 static __always_inline unsigned long dwarf_read_arch_reg(unsigned int reg) dwarf_read_arch_reg()
|
/linux-4.4.14/arch/ia64/include/asm/ |
H A D | spinlock.h | 40 static __always_inline void __ticket_spin_lock(arch_spinlock_t *lock) __ticket_spin_lock() 60 static __always_inline int __ticket_spin_trylock(arch_spinlock_t *lock) __ticket_spin_trylock() 69 static __always_inline void __ticket_spin_unlock(arch_spinlock_t *lock) __ticket_spin_unlock() 77 static __always_inline void __ticket_spin_unlock_wait(arch_spinlock_t *lock) __ticket_spin_unlock_wait() 105 static __always_inline int arch_spin_value_unlocked(arch_spinlock_t lock) arch_spin_value_unlocked() 121 static __always_inline void arch_spin_lock(arch_spinlock_t *lock) arch_spin_lock() 126 static __always_inline int arch_spin_trylock(arch_spinlock_t *lock) arch_spin_trylock() 131 static __always_inline void arch_spin_unlock(arch_spinlock_t *lock) arch_spin_unlock() 136 static __always_inline void arch_spin_lock_flags(arch_spinlock_t *lock, arch_spin_lock_flags() 152 static __always_inline void arch_read_lock_flags() 203 static __always_inline void arch_write_lock_flags()
|
/linux-4.4.14/arch/mips/vdso/ |
H A D | gettimeofday.c | 23 static __always_inline int do_realtime_coarse(struct timespec *ts, do_realtime_coarse() 38 static __always_inline int do_monotonic_coarse(struct timespec *ts, do_monotonic_coarse() 63 static __always_inline u64 read_r4k_count(void) read_r4k_count() 81 static __always_inline u64 read_gic_count(const union mips_vdso_data *data) read_gic_count() 97 static __always_inline u64 get_ns(const union mips_vdso_data *data) get_ns() 124 static __always_inline int do_realtime(struct timespec *ts, do_realtime() 146 static __always_inline int do_monotonic(struct timespec *ts, do_monotonic()
|
/linux-4.4.14/tools/include/linux/ |
H A D | compiler.h | 8 #ifndef __always_inline 9 # define __always_inline inline __attribute__((always_inline)) macro 62 static __always_inline void __read_once_size(const volatile void *p, void *res, int size) __read_once_size() 76 static __always_inline void __write_once_size(volatile void *p, void *res, int size) __write_once_size()
|
H A D | rbtree_augmented.h | 139 static __always_inline struct rb_node * __rb_erase_augmented() 236 static __always_inline void rb_erase_augmented()
|
/linux-4.4.14/mm/kasan/ |
H A D | kasan.c | 70 static __always_inline bool memory_is_poisoned_1(unsigned long addr) memory_is_poisoned_1() 82 static __always_inline bool memory_is_poisoned_2(unsigned long addr) memory_is_poisoned_2() 104 static __always_inline bool memory_is_poisoned_4(unsigned long addr) memory_is_poisoned_4() 126 static __always_inline bool memory_is_poisoned_8(unsigned long addr) memory_is_poisoned_8() 148 static __always_inline bool memory_is_poisoned_16(unsigned long addr) memory_is_poisoned_16() 172 static __always_inline unsigned long bytes_is_zero(const u8 *start, bytes_is_zero() 185 static __always_inline unsigned long memory_is_zero(const void *start, memory_is_zero() 214 static __always_inline bool memory_is_poisoned_n(unsigned long addr, memory_is_poisoned_n() 233 static __always_inline bool memory_is_poisoned(unsigned long addr, size_t size) memory_is_poisoned() 256 static __always_inline void check_memory_region(unsigned long addr, check_memory_region()
|
/linux-4.4.14/arch/xtensa/include/asm/ |
H A D | stacktrace.h | 20 static __always_inline unsigned long *stack_pointer(struct task_struct *task) stack_pointer()
|
H A D | fixmap.h | 60 static __always_inline unsigned long fix_to_virt(const unsigned int idx) fix_to_virt()
|
/linux-4.4.14/kernel/locking/ |
H A D | qspinlock.c | 158 static __always_inline void clear_pending_set_locked(struct qspinlock *lock) clear_pending_set_locked() 175 static __always_inline u32 xchg_tail(struct qspinlock *lock, u32 tail) xchg_tail() 190 static __always_inline void clear_pending_set_locked(struct qspinlock *lock) clear_pending_set_locked() 205 static __always_inline u32 xchg_tail(struct qspinlock *lock, u32 tail) xchg_tail() 227 static __always_inline void set_locked(struct qspinlock *lock) set_locked() 240 static __always_inline void __pv_init_node(struct mcs_spinlock *node) { } __pv_wait_node() 241 static __always_inline void __pv_wait_node(struct mcs_spinlock *node) { } __pv_kick_node() 242 static __always_inline void __pv_kick_node(struct qspinlock *lock, __pv_kick_node() 244 static __always_inline void __pv_wait_head(struct qspinlock *lock, __pv_wait_head()
|
H A D | qrwlock.c | 53 static __always_inline void rspin_until_writer_unlock()
|
H A D | mutex.c | 109 static __always_inline void ww_mutex_lock_acquired(struct ww_mutex *ww, ww_mutex_lock_acquired() 156 static __always_inline void ww_mutex_set_context_fastpath() 200 static __always_inline void ww_mutex_set_context_slowpath() 507 static __always_inline int __sched __mutex_lock_common()
|
/linux-4.4.14/drivers/infiniband/hw/mthca/ |
H A D | mthca_wqe.h | 116 static __always_inline void mthca_set_data_seg(struct mthca_data_seg *dseg, mthca_set_data_seg() 124 static __always_inline void mthca_set_data_seg_inval(struct mthca_data_seg *dseg) mthca_set_data_seg_inval()
|
H A D | mthca_qp.c | 1563 static __always_inline void set_raddr_seg(struct mthca_raddr_seg *rseg, set_raddr_seg() 1571 static __always_inline void set_atomic_seg(struct mthca_atomic_seg *aseg, set_atomic_seg()
|
/linux-4.4.14/arch/m32r/include/asm/ |
H A D | cmpxchg.h | 16 static __always_inline unsigned long __xchg() 70 static __always_inline unsigned long __xchg_local()
|
/linux-4.4.14/net/core/ |
H A D | secure_seq.c | 19 static __always_inline void net_secret_init(void) net_secret_init()
|
H A D | flow_dissector.c | 505 static __always_inline void __flow_hash_secret_init(void) __flow_hash_secret_init() 510 static __always_inline u32 __flow_hash_words(const u32 *words, u32 length, __flow_hash_words()
|
/linux-4.4.14/arch/openrisc/include/asm/ |
H A D | fixmap.h | 64 static __always_inline unsigned long fix_to_virt(const unsigned int idx) fix_to_virt()
|
/linux-4.4.14/arch/x86/entry/vdso/ |
H A D | vclock_gettime.c | 218 notrace static int __always_inline do_realtime(struct timespec *ts) do_realtime() 239 notrace static int __always_inline do_monotonic(struct timespec *ts) do_monotonic()
|
/linux-4.4.14/arch/m68k/include/asm/ |
H A D | uaccess_mm.h | 238 static __always_inline unsigned long __constant_copy_from_user() 319 static __always_inline unsigned long __constant_copy_to_user()
|
/linux-4.4.14/drivers/clocksource/ |
H A D | arm_arch_timer.c | 78 static __always_inline arch_timer_reg_write() 107 static __always_inline arch_timer_reg_read() 140 static __always_inline irqreturn_t timer_handler(const int access, timer_handler() 184 static __always_inline int timer_shutdown(const int access, timer_shutdown() 216 static __always_inline void set_next_event(const int access, unsigned long evt, set_next_event()
|
/linux-4.4.14/arch/sh/mm/ |
H A D | pmb.c | 72 static __always_inline unsigned long mk_pmb_entry(unsigned int entry) mk_pmb_entry() 77 static __always_inline unsigned long mk_pmb_addr(unsigned int entry) mk_pmb_addr() 82 static __always_inline unsigned long mk_pmb_data(unsigned int entry) mk_pmb_data() 87 static __always_inline unsigned int pmb_ppn_in_range(unsigned long ppn) pmb_ppn_in_range() 99 static __always_inline unsigned long pmb_cache_flags(void) pmb_cache_flags()
|
/linux-4.4.14/tools/lib/ |
H A D | rbtree.c | 71 static __always_inline void __rb_insert() 200 static __always_inline void ____rb_erase_color() 400 * This instantiates the same __always_inline functions as in the non-augmented
|
/linux-4.4.14/arch/x86/include/asm/fpu/ |
H A D | internal.h | 58 static __always_inline __pure bool use_eager_fpu(void) use_eager_fpu() 63 static __always_inline __pure bool use_xsaveopt(void) use_xsaveopt() 68 static __always_inline __pure bool use_xsave(void) use_xsave() 73 static __always_inline __pure bool use_fxsr(void) use_fxsr()
|
/linux-4.4.14/lib/ |
H A D | rbtree.c | 96 static __always_inline void __rb_insert() 227 static __always_inline void ____rb_erase_color() 436 * This instantiates the same __always_inline functions as in the non-augmented
|
H A D | radix-tree.c | 151 static __always_inline unsigned long radix_tree_find_next_bit()
|
/linux-4.4.14/lib/xz/ |
H A D | xz_dec_lzma2.c | 478 static __always_inline void rc_normalize(struct rc_dec *rc) rc_normalize() 497 static __always_inline int rc_bit(struct rc_dec *rc, uint16_t *prob) rc_bit() 519 static __always_inline uint32_t rc_bittree(struct rc_dec *rc, rc_bittree() 535 static __always_inline void rc_bittree_reverse(struct rc_dec *rc, rc_bittree_reverse()
|
/linux-4.4.14/mm/ |
H A D | userfaultfd.c | 142 static __always_inline ssize_t __mcopy_atomic(struct mm_struct *dst_mm, __mcopy_atomic()
|
H A D | list_lru.c | 66 static __always_inline struct mem_cgroup *mem_cgroup_from_kmem(void *ptr) mem_cgroup_from_kmem()
|
H A D | slab.h | 239 static __always_inline int memcg_charge_slab(struct page *page, memcg_charge_slab()
|
H A D | swap.c | 118 static __always_inline put_unrefcounted_compound_page() 165 static __always_inline put_refcounted_compound_page()
|
H A D | slob.c | 426 static __always_inline void * __do_kmalloc_node()
|
H A D | gup.c | 621 static __always_inline long __get_user_pages_locked(struct task_struct *tsk, __get_user_pages_locked() 758 __always_inline long __get_user_pages_unlocked(struct task_struct *tsk, struct mm_struct *mm, __get_user_pages_unlocked()
|
H A D | slab.c | 3143 static __always_inline void * slab_alloc_node() 3201 static __always_inline void * __do_cache_alloc() 3225 static __always_inline void * __do_cache_alloc() 3233 static __always_inline void * slab_alloc() 3486 static __always_inline void * __do_kmalloc_node() 3517 static __always_inline void *__do_kmalloc(size_t size, gfp_t flags, __do_kmalloc()
|
H A D | slub.c | 339 static __always_inline void slab_lock(struct page *page) slab_lock() 344 static __always_inline void slab_unlock(struct page *page) slab_unlock() 2482 static __always_inline void *slab_alloc_node(struct kmem_cache *s, slab_alloc_node() 2569 static __always_inline void *slab_alloc(struct kmem_cache *s, slab_alloc() 2764 static __always_inline void slab_free(struct kmem_cache *s, struct page *page, slab_free()
|
H A D | slab_common.c | 1181 static __always_inline void *__do_krealloc(const void *p, size_t new_size, __do_krealloc()
|
/linux-4.4.14/samples/bpf/ |
H A D | sockex3_kern.c | 138 static __always_inline void parse_ip_proto(struct __sk_buff *skb, parse_ip_proto()
|
/linux-4.4.14/kernel/sched/ |
H A D | sched.h | 995 static __always_inline bool static_branch_##name(struct static_key *key) \ 1400 static __always_inline arch_scale_freq_capacity() 1408 static __always_inline arch_scale_cpu_capacity()
|
H A D | cputime.c | 257 static __always_inline bool steal_account_process_tick(void) steal_account_process_tick()
|
H A D | fair.c | 426 static __always_inline 2469 static __always_inline u64 decay_load(u64 val, u64 n) decay_load() 2559 static __always_inline int __update_load_avg() 3078 static __always_inline void return_cfs_rq_runtime(struct cfs_rq *cfs_rq); 3487 static __always_inline account_cfs_rq_runtime() 3836 static __always_inline void return_cfs_rq_runtime(struct cfs_rq *cfs_rq) return_cfs_rq_runtime() 4043 static __always_inline void return_cfs_rq_runtime(struct cfs_rq *cfs_rq) {} return_cfs_rq_runtime()
|
H A D | core.c | 2457 static __always_inline void fire_sched_in_preempt_notifiers(struct task_struct *curr) fire_sched_in_preempt_notifiers() 2473 static __always_inline void fire_sched_out_preempt_notifiers()
|
/linux-4.4.14/kernel/trace/ |
H A D | trace.h | 497 static __always_inline int trace_get_context_bit(void) trace_get_context_bit() 515 static __always_inline int trace_test_and_set_recursion(int start, int max) trace_test_and_set_recursion() 535 static __always_inline void trace_clear_recursion(int bit) trace_clear_recursion()
|
H A D | ring_buffer.c | 2545 static __always_inline void rb_wakeups() 2610 static __always_inline int trace_recursive_lock() 2635 static __always_inline void trace_recursive_unlock()
|
H A D | ftrace.c | 1220 static bool __always_inline ftrace_hash_empty(struct ftrace_hash *hash) ftrace_hash_empty()
|
/linux-4.4.14/drivers/md/ |
H A D | dm-switch.c | 360 static __always_inline unsigned long parse_hex(const char **string) parse_hex()
|
H A D | dm-cache-target.c | 606 __always_inline
|
/linux-4.4.14/arch/x86/entry/ |
H A D | common.c | 364 __always_inline void do_syscall_32_irqs_on(struct pt_regs *regs) do_syscall_32_irqs_on()
|
/linux-4.4.14/arch/x86/kernel/cpu/ |
H A D | common.c | 261 static __always_inline void setup_smep(struct cpuinfo_x86 *c) setup_smep() 274 static __always_inline void setup_smap(struct cpuinfo_x86 *c) setup_smap()
|
/linux-4.4.14/fs/ |
H A D | userfaultfd.c | 666 static __always_inline void wake_userfault(struct userfaultfd_ctx *ctx, wake_userfault() 697 static __always_inline int validate_range(struct mm_struct *mm, validate_range()
|
H A D | namei.c | 981 static __always_inline get_link()
|
/linux-4.4.14/kernel/time/ |
H A D | timekeeping.c | 377 static __always_inline u64 __ktime_get_fast_ns(struct tk_fast *tkf) __ktime_get_fast_ns() 1507 static __always_inline void timekeeping_apply_adjustment(struct timekeeper *tk, timekeeping_apply_adjustment() 1589 static __always_inline void timekeeping_freqadjust(struct timekeeper *tk, timekeeping_freqadjust()
|
/linux-4.4.14/drivers/md/bcache/ |
H A D | bset.h | 395 static __always_inline int64_t bkey_cmp(const struct bkey *l, bkey_cmp()
|
/linux-4.4.14/drivers/base/ |
H A D | devres.c | 84 static __always_inline struct devres * alloc_dr(dr_release_t release, alloc_dr()
|
/linux-4.4.14/kernel/ |
H A D | module.c | 123 static __always_inline unsigned long __mod_tree_val(struct latch_tree_node *n) __mod_tree_val() 134 static __always_inline unsigned long __mod_tree_size(struct latch_tree_node *n) __mod_tree_size() 145 static __always_inline bool mod_tree_less() 151 static __always_inline int mod_tree_comp()
|
/linux-4.4.14/scripts/ |
H A D | kernel-doc | 2171 $prototype =~ s/^__always_inline +//;
|
H A D | checkpatch.pl | 308 our $Inline = qr{inline|__always_inline|noinline|__inline|__inline__};
|
/linux-4.4.14/arch/x86/kvm/ |
H A D | emulate.c | 640 static __always_inline int __linearize(struct x86_emulate_ctxt *ctxt, __linearize() 830 static __always_inline int do_insn_fetch_bytes(struct x86_emulate_ctxt *ctxt, do_insn_fetch_bytes()
|
H A D | vmx.c | 1453 static __always_inline unsigned long vmcs_readl(unsigned long field) vmcs_readl() 1462 static __always_inline u16 vmcs_read16(unsigned long field) vmcs_read16() 1467 static __always_inline u32 vmcs_read32(unsigned long field) vmcs_read32() 1472 static __always_inline u64 vmcs_read64(unsigned long field) vmcs_read64()
|
/linux-4.4.14/sound/soc/ |
H A D | soc-dapm.c | 183 static __always_inline void dapm_widget_invalidate_paths( dapm_widget_invalidate_paths() 1074 static __always_inline int is_connected_ep(struct snd_soc_dapm_widget *widget, is_connected_ep()
|
/linux-4.4.14/drivers/infiniband/hw/mlx4/ |
H A D | qp.c | 2556 static __always_inline void set_raddr_seg(struct mlx4_wqe_raddr_seg *rseg, set_raddr_seg()
|
/linux-4.4.14/drivers/infiniband/hw/mlx5/ |
H A D | qp.c | 1830 static __always_inline void set_raddr_seg(struct mlx5_wqe_raddr_seg *rseg, set_raddr_seg()
|
/linux-4.4.14/drivers/net/wireless/brcm80211/brcmfmac/ |
H A D | cfg80211.c | 3016 static __always_inline void brcmf_delay(u32 ms) brcmf_delay()
|