__always_inline   151 arch/x86/include/asm/barrier.h static __always_inline void rdtsc_barrier(void)
__always_inline    71 arch/x86/include/asm/bitops.h static __always_inline void
__always_inline   109 arch/x86/include/asm/bitops.h static __always_inline void
__always_inline   216 arch/x86/include/asm/bitops.h static __always_inline int
__always_inline   308 arch/x86/include/asm/bitops.h static __always_inline int constant_test_bit(long nr, const volatile unsigned long *addr)
__always_inline   479 arch/x86/include/asm/bitops.h static __always_inline int fls64(__u64 x)
__always_inline   396 arch/x86/include/asm/cpufeature.h static __always_inline __pure bool __static_cpu_has(u16 bit)
__always_inline   476 arch/x86/include/asm/cpufeature.h static __always_inline __pure bool _static_cpu_has_safe(u16 bit)
__always_inline    12 arch/x86/include/asm/current.h static __always_inline struct task_struct *get_current(void)
__always_inline    10 arch/x86/include/asm/dmi.h static __always_inline __init void *dmi_alloc(unsigned len)
__always_inline    88 arch/x86/include/asm/fpu-internal.h static __always_inline __pure bool use_eager_fpu(void)
__always_inline    93 arch/x86/include/asm/fpu-internal.h static __always_inline __pure bool use_xsaveopt(void)
__always_inline    98 arch/x86/include/asm/fpu-internal.h static __always_inline __pure bool use_xsave(void)
__always_inline   103 arch/x86/include/asm/fpu-internal.h static __always_inline __pure bool use_fxsr(void)
__always_inline    19 arch/x86/include/asm/jump_label.h static __always_inline bool arch_static_branch(struct static_key *key)
__always_inline   113 arch/x86/include/asm/msr.h static __always_inline unsigned long long __native_read_tsc(void)
__always_inline   715 arch/x86/include/asm/paravirt.h static __always_inline void __ticket_lock_spinning(struct arch_spinlock *lock,
__always_inline   721 arch/x86/include/asm/paravirt.h static __always_inline void __ticket_unlock_kick(struct arch_spinlock *lock,
__always_inline   491 arch/x86/include/asm/percpu.h static __always_inline int x86_this_cpu_constant_test_bit(unsigned int nr,
__always_inline    82 arch/x86/include/asm/pgtable-2level.h static __always_inline pgoff_t pte_to_pgoff(pte_t pte)
__always_inline    90 arch/x86/include/asm/pgtable-2level.h static __always_inline pte_t pgoff_to_pte(pgoff_t off)
__always_inline    20 arch/x86/include/asm/preempt.h static __always_inline int preempt_count(void)
__always_inline    25 arch/x86/include/asm/preempt.h static __always_inline void preempt_count_set(int pc)
__always_inline    54 arch/x86/include/asm/preempt.h static __always_inline void set_preempt_need_resched(void)
__always_inline    59 arch/x86/include/asm/preempt.h static __always_inline void clear_preempt_need_resched(void)
__always_inline    64 arch/x86/include/asm/preempt.h static __always_inline bool test_preempt_need_resched(void)
__always_inline    73 arch/x86/include/asm/preempt.h static __always_inline void __preempt_count_add(int val)
__always_inline    78 arch/x86/include/asm/preempt.h static __always_inline void __preempt_count_sub(int val)
__always_inline    88 arch/x86/include/asm/preempt.h static __always_inline bool __preempt_count_dec_and_test(void)
__always_inline    96 arch/x86/include/asm/preempt.h static __always_inline bool should_resched(void)
__always_inline    62 arch/x86/include/asm/pvclock.h static __always_inline
__always_inline    70 arch/x86/include/asm/pvclock.h static __always_inline
__always_inline    61 arch/x86/include/asm/smap.h static __always_inline void clac(void)
__always_inline    67 arch/x86/include/asm/smap.h static __always_inline void stac(void)
__always_inline    43 arch/x86/include/asm/spinlock.h static __always_inline bool static_key_false(struct static_key *key);
__always_inline    53 arch/x86/include/asm/spinlock.h static __always_inline void __ticket_lock_spinning(arch_spinlock_t *lock,
__always_inline    64 arch/x86/include/asm/spinlock.h static __always_inline int arch_spin_value_unlocked(arch_spinlock_t lock)
__always_inline    82 arch/x86/include/asm/spinlock.h static __always_inline void arch_spin_lock(arch_spinlock_t *lock)
__always_inline   104 arch/x86/include/asm/spinlock.h static __always_inline int arch_spin_trylock(arch_spinlock_t *lock)
__always_inline   146 arch/x86/include/asm/spinlock.h static __always_inline void arch_spin_unlock(arch_spinlock_t *lock)
__always_inline   178 arch/x86/include/asm/spinlock.h static __always_inline void arch_spin_lock_flags(arch_spinlock_t *lock,
__always_inline    58 arch/x86/include/asm/stackprotector.h static __always_inline void boot_init_stack_canary(void)
__always_inline    32 arch/x86/include/asm/string_32.h static __always_inline void *__memcpy(void *to, const void *from, size_t n)
__always_inline    51 arch/x86/include/asm/string_32.h static __always_inline void *__constant_memcpy(void *to, const void *from,
__always_inline   226 arch/x86/include/asm/string_32.h static __always_inline
__always_inline   256 arch/x86/include/asm/string_32.h static __always_inline
__always_inline     9 arch/x86/include/asm/string_64.h static __always_inline void *__inline_memcpy(void *to, const void *from, size_t n)
__always_inline    35 arch/x86/include/asm/tsc.h static __always_inline cycles_t vget_cycles(void)
__always_inline    43 arch/x86/include/asm/uaccess_32.h static __always_inline unsigned long __must_check
__always_inline    81 arch/x86/include/asm/uaccess_32.h static __always_inline unsigned long __must_check
__always_inline    88 arch/x86/include/asm/uaccess_32.h static __always_inline unsigned long
__always_inline   136 arch/x86/include/asm/uaccess_32.h static __always_inline unsigned long
__always_inline   158 arch/x86/include/asm/uaccess_32.h static __always_inline unsigned long __copy_from_user_nocache(void *to,
__always_inline   180 arch/x86/include/asm/uaccess_32.h static __always_inline unsigned long
__always_inline    26 arch/x86/include/asm/uaccess_64.h static __always_inline __must_check unsigned long
__always_inline    51 arch/x86/include/asm/uaccess_64.h static __always_inline __must_check
__always_inline    94 arch/x86/include/asm/uaccess_64.h static __always_inline __must_check
__always_inline   101 arch/x86/include/asm/uaccess_64.h static __always_inline __must_check
__always_inline   144 arch/x86/include/asm/uaccess_64.h static __always_inline __must_check
__always_inline   151 arch/x86/include/asm/uaccess_64.h static __always_inline __must_check
__always_inline   204 arch/x86/include/asm/uaccess_64.h static __must_check __always_inline int
__always_inline   210 arch/x86/include/asm/uaccess_64.h static __must_check __always_inline int
__always_inline    12 include/asm-generic/bitops/__ffs.h static __always_inline unsigned long __ffs(unsigned long word)
__always_inline    12 include/asm-generic/bitops/__fls.h static __always_inline unsigned long __fls(unsigned long word)
__always_inline    10 include/asm-generic/bitops/builtin-__ffs.h static __always_inline unsigned long __ffs(unsigned long word)
__always_inline    10 include/asm-generic/bitops/builtin-__fls.h static __always_inline unsigned long __fls(unsigned long word)
__always_inline    12 include/asm-generic/bitops/builtin-ffs.h static __always_inline int ffs(int x)
__always_inline    11 include/asm-generic/bitops/builtin-fls.h static __always_inline int fls(int x)
__always_inline    12 include/asm-generic/bitops/fls.h static __always_inline int fls(int x)
__always_inline    18 include/asm-generic/bitops/fls64.h static __always_inline int fls64(__u64 x)
__always_inline    26 include/asm-generic/bitops/fls64.h static __always_inline int fls64(__u64 x)
__always_inline    29 include/asm-generic/fixmap.h static __always_inline unsigned long fix_to_virt(const unsigned int idx)
__always_inline     8 include/asm-generic/preempt.h static __always_inline int preempt_count(void)
__always_inline    13 include/asm-generic/preempt.h static __always_inline int *preempt_count_ptr(void)
__always_inline    18 include/asm-generic/preempt.h static __always_inline void preempt_count_set(int pc)
__always_inline    37 include/asm-generic/preempt.h static __always_inline void set_preempt_need_resched(void)
__always_inline    41 include/asm-generic/preempt.h static __always_inline void clear_preempt_need_resched(void)
__always_inline    45 include/asm-generic/preempt.h static __always_inline bool test_preempt_need_resched(void)
__always_inline    54 include/asm-generic/preempt.h static __always_inline void __preempt_count_add(int val)
__always_inline    59 include/asm-generic/preempt.h static __always_inline void __preempt_count_sub(int val)
__always_inline    64 include/asm-generic/preempt.h static __always_inline bool __preempt_count_dec_and_test(void)
__always_inline    77 include/asm-generic/preempt.h static __always_inline bool should_resched(void)
__always_inline    30 include/linux/async_tx.h #define __async_inline __always_inline
__always_inline    10 include/linux/bottom_half.h static __always_inline void __local_bh_disable_ip(unsigned long ip, unsigned int cnt)
__always_inline   265 include/linux/compiler.h #ifndef __always_inline
__always_inline   386 include/linux/compiler.h # define nokprobe_inline	__always_inline
__always_inline    36 include/linux/hash.h static __always_inline u64 hash_64(u64 val, unsigned int bits)
__always_inline   112 include/linux/jump_label.h static __always_inline bool static_key_false(struct static_key *key)
__always_inline   117 include/linux/jump_label.h static __always_inline bool static_key_true(struct static_key *key)
__always_inline   146 include/linux/jump_label.h static __always_inline void jump_label_init(void)
__always_inline   151 include/linux/jump_label.h static __always_inline bool static_key_false(struct static_key *key)
__always_inline   158 include/linux/jump_label.h static __always_inline bool static_key_true(struct static_key *key)
__always_inline   117 include/linux/math64.h static __always_inline u32
__always_inline   505 include/linux/memcontrol.h static __always_inline struct kmem_cache *
__always_inline   933 include/linux/mm.h static __always_inline void *lowmem_page_address(const struct page *page)
__always_inline    25 include/linux/mm_inline.h static __always_inline void add_page_to_lru_list(struct page *page,
__always_inline    34 include/linux/mm_inline.h static __always_inline void del_page_from_lru_list(struct page *page,
__always_inline    65 include/linux/mm_inline.h static __always_inline enum lru_list page_off_lru(struct page *page)
__always_inline    89 include/linux/mm_inline.h static __always_inline enum lru_list page_lru(struct page *page)
__always_inline   111 include/linux/nodemask.h static __always_inline void __node_set(int node, volatile nodemask_t *dstp)
__always_inline   684 include/linux/perf_event.h static __always_inline void
__always_inline   340 include/linux/radix-tree.h static __always_inline void **
__always_inline   378 include/linux/radix-tree.h static __always_inline unsigned
__always_inline   395 include/linux/radix-tree.h static __always_inline void **
__always_inline   136 include/linux/rbtree_augmented.h static __always_inline struct rb_node *
__always_inline   233 include/linux/rbtree_augmented.h static __always_inline void
__always_inline  2886 include/linux/sched.h static __always_inline bool need_resched(void)
__always_inline   247 include/linux/slab.h static __always_inline int kmalloc_index(size_t size)
__always_inline   297 include/linux/slab.h static __always_inline void *__kmalloc_node(size_t size, gfp_t flags, int node)
__always_inline   302 include/linux/slab.h static __always_inline void *kmem_cache_alloc_node(struct kmem_cache *s, gfp_t flags, int node)
__always_inline   316 include/linux/slab.h static __always_inline void *
__always_inline   326 include/linux/slab.h static __always_inline void *kmem_cache_alloc_trace(struct kmem_cache *s,
__always_inline   332 include/linux/slab.h static __always_inline void *
__always_inline   346 include/linux/slab.h static __always_inline void *
__always_inline   353 include/linux/slab.h static __always_inline void *kmalloc_large(size_t size, gfp_t flags)
__always_inline   412 include/linux/slab.h static __always_inline void *kmalloc(size_t size, gfp_t flags)
__always_inline   437 include/linux/slab.h static __always_inline int kmalloc_size(int n)
__always_inline   452 include/linux/slab.h static __always_inline void *kmalloc_node(size_t size, gfp_t flags, int node)
__always_inline   222 include/linux/time.h static __always_inline void timespec_add_ns(struct timespec *a, u64 ns)
__always_inline   182 include/linux/time64.h static __always_inline void timespec64_add_ns(struct timespec64 *a, u64 ns)