__always_inline 151 arch/x86/include/asm/barrier.h static __always_inline void rdtsc_barrier(void) __always_inline 71 arch/x86/include/asm/bitops.h static __always_inline void __always_inline 109 arch/x86/include/asm/bitops.h static __always_inline void __always_inline 216 arch/x86/include/asm/bitops.h static __always_inline int __always_inline 308 arch/x86/include/asm/bitops.h static __always_inline int constant_test_bit(long nr, const volatile unsigned long *addr) __always_inline 479 arch/x86/include/asm/bitops.h static __always_inline int fls64(__u64 x) __always_inline 396 arch/x86/include/asm/cpufeature.h static __always_inline __pure bool __static_cpu_has(u16 bit) __always_inline 476 arch/x86/include/asm/cpufeature.h static __always_inline __pure bool _static_cpu_has_safe(u16 bit) __always_inline 12 arch/x86/include/asm/current.h static __always_inline struct task_struct *get_current(void) __always_inline 10 arch/x86/include/asm/dmi.h static __always_inline __init void *dmi_alloc(unsigned len) __always_inline 88 arch/x86/include/asm/fpu-internal.h static __always_inline __pure bool use_eager_fpu(void) __always_inline 93 arch/x86/include/asm/fpu-internal.h static __always_inline __pure bool use_xsaveopt(void) __always_inline 98 arch/x86/include/asm/fpu-internal.h static __always_inline __pure bool use_xsave(void) __always_inline 103 arch/x86/include/asm/fpu-internal.h static __always_inline __pure bool use_fxsr(void) __always_inline 19 arch/x86/include/asm/jump_label.h static __always_inline bool arch_static_branch(struct static_key *key) __always_inline 113 arch/x86/include/asm/msr.h static __always_inline unsigned long long __native_read_tsc(void) __always_inline 715 arch/x86/include/asm/paravirt.h static __always_inline void __ticket_lock_spinning(struct arch_spinlock *lock, __always_inline 721 arch/x86/include/asm/paravirt.h static __always_inline void __ticket_unlock_kick(struct arch_spinlock *lock, __always_inline 491 arch/x86/include/asm/percpu.h static __always_inline int x86_this_cpu_constant_test_bit(unsigned int nr, __always_inline 82 arch/x86/include/asm/pgtable-2level.h static __always_inline pgoff_t pte_to_pgoff(pte_t pte) __always_inline 90 arch/x86/include/asm/pgtable-2level.h static __always_inline pte_t pgoff_to_pte(pgoff_t off) __always_inline 20 arch/x86/include/asm/preempt.h static __always_inline int preempt_count(void) __always_inline 25 arch/x86/include/asm/preempt.h static __always_inline void preempt_count_set(int pc) __always_inline 54 arch/x86/include/asm/preempt.h static __always_inline void set_preempt_need_resched(void) __always_inline 59 arch/x86/include/asm/preempt.h static __always_inline void clear_preempt_need_resched(void) __always_inline 64 arch/x86/include/asm/preempt.h static __always_inline bool test_preempt_need_resched(void) __always_inline 73 arch/x86/include/asm/preempt.h static __always_inline void __preempt_count_add(int val) __always_inline 78 arch/x86/include/asm/preempt.h static __always_inline void __preempt_count_sub(int val) __always_inline 88 arch/x86/include/asm/preempt.h static __always_inline bool __preempt_count_dec_and_test(void) __always_inline 96 arch/x86/include/asm/preempt.h static __always_inline bool should_resched(void) __always_inline 62 arch/x86/include/asm/pvclock.h static __always_inline __always_inline 70 arch/x86/include/asm/pvclock.h static __always_inline __always_inline 61 arch/x86/include/asm/smap.h static __always_inline void clac(void) __always_inline 67 arch/x86/include/asm/smap.h static __always_inline void stac(void) __always_inline 43 arch/x86/include/asm/spinlock.h static __always_inline bool static_key_false(struct static_key *key); __always_inline 53 arch/x86/include/asm/spinlock.h static __always_inline void __ticket_lock_spinning(arch_spinlock_t *lock, __always_inline 64 arch/x86/include/asm/spinlock.h static __always_inline int arch_spin_value_unlocked(arch_spinlock_t lock) __always_inline 82 arch/x86/include/asm/spinlock.h static __always_inline void arch_spin_lock(arch_spinlock_t *lock) __always_inline 104 arch/x86/include/asm/spinlock.h static __always_inline int arch_spin_trylock(arch_spinlock_t *lock) __always_inline 146 arch/x86/include/asm/spinlock.h static __always_inline void arch_spin_unlock(arch_spinlock_t *lock) __always_inline 178 arch/x86/include/asm/spinlock.h static __always_inline void arch_spin_lock_flags(arch_spinlock_t *lock, __always_inline 58 arch/x86/include/asm/stackprotector.h static __always_inline void boot_init_stack_canary(void) __always_inline 32 arch/x86/include/asm/string_32.h static __always_inline void *__memcpy(void *to, const void *from, size_t n) __always_inline 51 arch/x86/include/asm/string_32.h static __always_inline void *__constant_memcpy(void *to, const void *from, __always_inline 226 arch/x86/include/asm/string_32.h static __always_inline __always_inline 256 arch/x86/include/asm/string_32.h static __always_inline __always_inline 9 arch/x86/include/asm/string_64.h static __always_inline void *__inline_memcpy(void *to, const void *from, size_t n) __always_inline 35 arch/x86/include/asm/tsc.h static __always_inline cycles_t vget_cycles(void) __always_inline 43 arch/x86/include/asm/uaccess_32.h static __always_inline unsigned long __must_check __always_inline 81 arch/x86/include/asm/uaccess_32.h static __always_inline unsigned long __must_check __always_inline 88 arch/x86/include/asm/uaccess_32.h static __always_inline unsigned long __always_inline 136 arch/x86/include/asm/uaccess_32.h static __always_inline unsigned long __always_inline 158 arch/x86/include/asm/uaccess_32.h static __always_inline unsigned long __copy_from_user_nocache(void *to, __always_inline 180 arch/x86/include/asm/uaccess_32.h static __always_inline unsigned long __always_inline 26 arch/x86/include/asm/uaccess_64.h static __always_inline __must_check unsigned long __always_inline 51 arch/x86/include/asm/uaccess_64.h static __always_inline __must_check __always_inline 94 arch/x86/include/asm/uaccess_64.h static __always_inline __must_check __always_inline 101 arch/x86/include/asm/uaccess_64.h static __always_inline __must_check __always_inline 144 arch/x86/include/asm/uaccess_64.h static __always_inline __must_check __always_inline 151 arch/x86/include/asm/uaccess_64.h static __always_inline __must_check __always_inline 204 arch/x86/include/asm/uaccess_64.h static __must_check __always_inline int __always_inline 210 arch/x86/include/asm/uaccess_64.h static __must_check __always_inline int __always_inline 12 include/asm-generic/bitops/__ffs.h static __always_inline unsigned long __ffs(unsigned long word) __always_inline 12 include/asm-generic/bitops/__fls.h static __always_inline unsigned long __fls(unsigned long word) __always_inline 10 include/asm-generic/bitops/builtin-__ffs.h static __always_inline unsigned long __ffs(unsigned long word) __always_inline 10 include/asm-generic/bitops/builtin-__fls.h static __always_inline unsigned long __fls(unsigned long word) __always_inline 12 include/asm-generic/bitops/builtin-ffs.h static __always_inline int ffs(int x) __always_inline 11 include/asm-generic/bitops/builtin-fls.h static __always_inline int fls(int x) __always_inline 12 include/asm-generic/bitops/fls.h static __always_inline int fls(int x) __always_inline 18 include/asm-generic/bitops/fls64.h static __always_inline int fls64(__u64 x) __always_inline 26 include/asm-generic/bitops/fls64.h static __always_inline int fls64(__u64 x) __always_inline 29 include/asm-generic/fixmap.h static __always_inline unsigned long fix_to_virt(const unsigned int idx) __always_inline 8 include/asm-generic/preempt.h static __always_inline int preempt_count(void) __always_inline 13 include/asm-generic/preempt.h static __always_inline int *preempt_count_ptr(void) __always_inline 18 include/asm-generic/preempt.h static __always_inline void preempt_count_set(int pc) __always_inline 37 include/asm-generic/preempt.h static __always_inline void set_preempt_need_resched(void) __always_inline 41 include/asm-generic/preempt.h static __always_inline void clear_preempt_need_resched(void) __always_inline 45 include/asm-generic/preempt.h static __always_inline bool test_preempt_need_resched(void) __always_inline 54 include/asm-generic/preempt.h static __always_inline void __preempt_count_add(int val) __always_inline 59 include/asm-generic/preempt.h static __always_inline void __preempt_count_sub(int val) __always_inline 64 include/asm-generic/preempt.h static __always_inline bool __preempt_count_dec_and_test(void) __always_inline 77 include/asm-generic/preempt.h static __always_inline bool should_resched(void) __always_inline 30 include/linux/async_tx.h #define __async_inline __always_inline __always_inline 10 include/linux/bottom_half.h static __always_inline void __local_bh_disable_ip(unsigned long ip, unsigned int cnt) __always_inline 265 include/linux/compiler.h #ifndef __always_inline __always_inline 386 include/linux/compiler.h # define nokprobe_inline __always_inline __always_inline 36 include/linux/hash.h static __always_inline u64 hash_64(u64 val, unsigned int bits) __always_inline 112 include/linux/jump_label.h static __always_inline bool static_key_false(struct static_key *key) __always_inline 117 include/linux/jump_label.h static __always_inline bool static_key_true(struct static_key *key) __always_inline 146 include/linux/jump_label.h static __always_inline void jump_label_init(void) __always_inline 151 include/linux/jump_label.h static __always_inline bool static_key_false(struct static_key *key) __always_inline 158 include/linux/jump_label.h static __always_inline bool static_key_true(struct static_key *key) __always_inline 117 include/linux/math64.h static __always_inline u32 __always_inline 505 include/linux/memcontrol.h static __always_inline struct kmem_cache * __always_inline 933 include/linux/mm.h static __always_inline void *lowmem_page_address(const struct page *page) __always_inline 25 include/linux/mm_inline.h static __always_inline void add_page_to_lru_list(struct page *page, __always_inline 34 include/linux/mm_inline.h static __always_inline void del_page_from_lru_list(struct page *page, __always_inline 65 include/linux/mm_inline.h static __always_inline enum lru_list page_off_lru(struct page *page) __always_inline 89 include/linux/mm_inline.h static __always_inline enum lru_list page_lru(struct page *page) __always_inline 111 include/linux/nodemask.h static __always_inline void __node_set(int node, volatile nodemask_t *dstp) __always_inline 684 include/linux/perf_event.h static __always_inline void __always_inline 340 include/linux/radix-tree.h static __always_inline void ** __always_inline 378 include/linux/radix-tree.h static __always_inline unsigned __always_inline 395 include/linux/radix-tree.h static __always_inline void ** __always_inline 136 include/linux/rbtree_augmented.h static __always_inline struct rb_node * __always_inline 233 include/linux/rbtree_augmented.h static __always_inline void __always_inline 2886 include/linux/sched.h static __always_inline bool need_resched(void) __always_inline 247 include/linux/slab.h static __always_inline int kmalloc_index(size_t size) __always_inline 297 include/linux/slab.h static __always_inline void *__kmalloc_node(size_t size, gfp_t flags, int node) __always_inline 302 include/linux/slab.h static __always_inline void *kmem_cache_alloc_node(struct kmem_cache *s, gfp_t flags, int node) __always_inline 316 include/linux/slab.h static __always_inline void * __always_inline 326 include/linux/slab.h static __always_inline void *kmem_cache_alloc_trace(struct kmem_cache *s, __always_inline 332 include/linux/slab.h static __always_inline void * __always_inline 346 include/linux/slab.h static __always_inline void * __always_inline 353 include/linux/slab.h static __always_inline void *kmalloc_large(size_t size, gfp_t flags) __always_inline 412 include/linux/slab.h static __always_inline void *kmalloc(size_t size, gfp_t flags) __always_inline 437 include/linux/slab.h static __always_inline int kmalloc_size(int n) __always_inline 452 include/linux/slab.h static __always_inline void *kmalloc_node(size_t size, gfp_t flags, int node) __always_inline 222 include/linux/time.h static __always_inline void timespec_add_ns(struct timespec *a, u64 ns) __always_inline 182 include/linux/time64.h static __always_inline void timespec64_add_ns(struct timespec64 *a, u64 ns)