msr 138 arch/x86/include/asm/apic.h u64 msr; msr 143 arch/x86/include/asm/apic.h rdmsrl(APIC_BASE_MSR + (reg >> 4), msr); msr 144 arch/x86/include/asm/apic.h return (u32)msr; msr 178 arch/x86/include/asm/apic.h u64 msr; msr 183 arch/x86/include/asm/apic.h rdmsrl(MSR_IA32_APICBASE, msr); msr 184 arch/x86/include/asm/apic.h if (msr & X2APIC_ENABLE) msr 684 arch/x86/include/asm/kvm_host.h int (*set_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr); msr 855 arch/x86/include/asm/kvm_host.h int kvm_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr); msr 882 arch/x86/include/asm/kvm_host.h int kvm_get_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata); msr 883 arch/x86/include/asm/kvm_host.h int kvm_set_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr); msr 973 arch/x86/include/asm/kvm_host.h static inline unsigned long read_msr(unsigned long msr) msr 977 arch/x86/include/asm/kvm_host.h rdmsrl(msr, value); msr 1066 arch/x86/include/asm/kvm_host.h void kvm_define_shared_msr(unsigned index, u32 msr); msr 1088 arch/x86/include/asm/kvm_host.h bool kvm_pmu_msr(struct kvm_vcpu *vcpu, u32 msr); msr 1089 arch/x86/include/asm/kvm_host.h int kvm_pmu_get_msr(struct kvm_vcpu *vcpu, u32 msr, u64 *data); msr 4 arch/x86/include/asm/microcode.h #define native_rdmsr(msr, val1, val2) \ msr 6 arch/x86/include/asm/microcode.h u64 __val = native_read_msr((msr)); \ msr 11 arch/x86/include/asm/microcode.h #define native_wrmsr(msr, low, high) \ msr 12 arch/x86/include/asm/microcode.h native_write_msr(msr, low, high) msr 14 arch/x86/include/asm/microcode.h #define native_wrmsrl(msr, val) \ msr 15 arch/x86/include/asm/microcode.h native_write_msr((msr), \ msr 24 arch/x86/include/asm/msr.h struct msr reg; msr 25 arch/x86/include/asm/msr.h struct msr *msrs; msr 60 arch/x86/include/asm/msr.h static inline unsigned long long native_read_msr(unsigned int msr) msr 64 arch/x86/include/asm/msr.h asm volatile("rdmsr" : EAX_EDX_RET(val, low, high) : "c" (msr)); msr 68 arch/x86/include/asm/msr.h static inline unsigned long long native_read_msr_safe(unsigned int msr, msr 80 arch/x86/include/asm/msr.h : "c" (msr), [fault] "i" (-EIO)); msr 84 arch/x86/include/asm/msr.h static inline void native_write_msr(unsigned int msr, msr 87 arch/x86/include/asm/msr.h asm volatile("wrmsr" : : "c" (msr), "a"(low), "d" (high) : "memory"); msr 91 arch/x86/include/asm/msr.h notrace static inline int native_write_msr_safe(unsigned int msr, msr 102 arch/x86/include/asm/msr.h : "c" (msr), "0" (low), "d" (high), msr 140 arch/x86/include/asm/msr.h #define rdmsr(msr, low, high) \ msr 142 arch/x86/include/asm/msr.h u64 __val = native_read_msr((msr)); \ msr 147 arch/x86/include/asm/msr.h static inline void wrmsr(unsigned msr, unsigned low, unsigned high) msr 149 arch/x86/include/asm/msr.h native_write_msr(msr, low, high); msr 152 arch/x86/include/asm/msr.h #define rdmsrl(msr, val) \ msr 153 arch/x86/include/asm/msr.h ((val) = native_read_msr((msr))) msr 155 arch/x86/include/asm/msr.h #define wrmsrl(msr, val) \ msr 156 arch/x86/include/asm/msr.h native_write_msr((msr), (u32)((u64)(val)), (u32)((u64)(val) >> 32)) msr 159 arch/x86/include/asm/msr.h static inline int wrmsr_safe(unsigned msr, unsigned low, unsigned high) msr 161 arch/x86/include/asm/msr.h return native_write_msr_safe(msr, low, high); msr 165 arch/x86/include/asm/msr.h #define rdmsr_safe(msr, low, high) \ msr 168 arch/x86/include/asm/msr.h u64 __val = native_read_msr_safe((msr), &__err); \ msr 174 arch/x86/include/asm/msr.h static inline int rdmsrl_safe(unsigned msr, unsigned long long *p) msr 178 arch/x86/include/asm/msr.h *p = native_read_msr_safe(msr, &err); msr 208 arch/x86/include/asm/msr.h #define wrmsrl_safe(msr, val) wrmsr_safe((msr), (u32)(val), \ msr 215 arch/x86/include/asm/msr.h struct msr *msrs_alloc(void); msr 216 arch/x86/include/asm/msr.h void msrs_free(struct msr *msrs); msr 217 arch/x86/include/asm/msr.h int msr_set_bit(u32 msr, u8 bit); msr 218 arch/x86/include/asm/msr.h int msr_clear_bit(u32 msr, u8 bit); msr 225 arch/x86/include/asm/msr.h void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs); msr 226 arch/x86/include/asm/msr.h void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs); msr 255 arch/x86/include/asm/msr.h struct msr *msrs) msr 260 arch/x86/include/asm/msr.h struct msr *msrs) msr 126 arch/x86/include/asm/paravirt.h static inline u64 paravirt_read_msr(unsigned msr, int *err) msr 128 arch/x86/include/asm/paravirt.h return PVOP_CALL2(u64, pv_cpu_ops.read_msr, msr, err); msr 131 arch/x86/include/asm/paravirt.h static inline int paravirt_write_msr(unsigned msr, unsigned low, unsigned high) msr 133 arch/x86/include/asm/paravirt.h return PVOP_CALL3(int, pv_cpu_ops.write_msr, msr, low, high); msr 137 arch/x86/include/asm/paravirt.h #define rdmsr(msr, val1, val2) \ msr 140 arch/x86/include/asm/paravirt.h u64 _l = paravirt_read_msr(msr, &_err); \ msr 145 arch/x86/include/asm/paravirt.h #define wrmsr(msr, val1, val2) \ msr 147 arch/x86/include/asm/paravirt.h paravirt_write_msr(msr, val1, val2); \ msr 150 arch/x86/include/asm/paravirt.h #define rdmsrl(msr, val) \ msr 153 arch/x86/include/asm/paravirt.h val = paravirt_read_msr(msr, &_err); \ msr 156 arch/x86/include/asm/paravirt.h #define wrmsrl(msr, val) wrmsr(msr, (u32)((u64)(val)), ((u64)(val))>>32) msr 157 arch/x86/include/asm/paravirt.h #define wrmsr_safe(msr, a, b) paravirt_write_msr(msr, a, b) msr 160 arch/x86/include/asm/paravirt.h #define rdmsr_safe(msr, a, b) \ msr 163 arch/x86/include/asm/paravirt.h u64 _l = paravirt_read_msr(msr, &_err); \ msr 169 arch/x86/include/asm/paravirt.h static inline int rdmsrl_safe(unsigned msr, unsigned long long *p) msr 173 arch/x86/include/asm/paravirt.h *p = paravirt_read_msr(msr, &err); msr 156 arch/x86/include/asm/paravirt_types.h u64 (*read_msr)(unsigned int msr, int *err); msr 157 arch/x86/include/asm/paravirt_types.h int (*write_msr)(unsigned int msr, unsigned low, unsigned high); msr 254 arch/x86/include/asm/perf_event.h unsigned msr; msr 316 include/linux/isdn.h int msr; /* Modem status register */ msr 829 include/uapi/linux/kvm.h __u32 msr;