msr               138 arch/x86/include/asm/apic.h 	u64 msr;
msr               143 arch/x86/include/asm/apic.h 	rdmsrl(APIC_BASE_MSR + (reg >> 4), msr);
msr               144 arch/x86/include/asm/apic.h 	return (u32)msr;
msr               178 arch/x86/include/asm/apic.h 	u64 msr;
msr               183 arch/x86/include/asm/apic.h 	rdmsrl(MSR_IA32_APICBASE, msr);
msr               184 arch/x86/include/asm/apic.h 	if (msr & X2APIC_ENABLE)
msr               684 arch/x86/include/asm/kvm_host.h 	int (*set_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr);
msr               855 arch/x86/include/asm/kvm_host.h int kvm_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr);
msr               882 arch/x86/include/asm/kvm_host.h int kvm_get_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata);
msr               883 arch/x86/include/asm/kvm_host.h int kvm_set_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr);
msr               973 arch/x86/include/asm/kvm_host.h static inline unsigned long read_msr(unsigned long msr)
msr               977 arch/x86/include/asm/kvm_host.h 	rdmsrl(msr, value);
msr              1066 arch/x86/include/asm/kvm_host.h void kvm_define_shared_msr(unsigned index, u32 msr);
msr              1088 arch/x86/include/asm/kvm_host.h bool kvm_pmu_msr(struct kvm_vcpu *vcpu, u32 msr);
msr              1089 arch/x86/include/asm/kvm_host.h int kvm_pmu_get_msr(struct kvm_vcpu *vcpu, u32 msr, u64 *data);
msr                 4 arch/x86/include/asm/microcode.h #define native_rdmsr(msr, val1, val2)			\
msr                 6 arch/x86/include/asm/microcode.h 	u64 __val = native_read_msr((msr));		\
msr                11 arch/x86/include/asm/microcode.h #define native_wrmsr(msr, low, high)			\
msr                12 arch/x86/include/asm/microcode.h 	native_write_msr(msr, low, high)
msr                14 arch/x86/include/asm/microcode.h #define native_wrmsrl(msr, val)				\
msr                15 arch/x86/include/asm/microcode.h 	native_write_msr((msr),				\
msr                24 arch/x86/include/asm/msr.h 	struct msr reg;
msr                25 arch/x86/include/asm/msr.h 	struct msr *msrs;
msr                60 arch/x86/include/asm/msr.h static inline unsigned long long native_read_msr(unsigned int msr)
msr                64 arch/x86/include/asm/msr.h 	asm volatile("rdmsr" : EAX_EDX_RET(val, low, high) : "c" (msr));
msr                68 arch/x86/include/asm/msr.h static inline unsigned long long native_read_msr_safe(unsigned int msr,
msr                80 arch/x86/include/asm/msr.h 		     : "c" (msr), [fault] "i" (-EIO));
msr                84 arch/x86/include/asm/msr.h static inline void native_write_msr(unsigned int msr,
msr                87 arch/x86/include/asm/msr.h 	asm volatile("wrmsr" : : "c" (msr), "a"(low), "d" (high) : "memory");
msr                91 arch/x86/include/asm/msr.h notrace static inline int native_write_msr_safe(unsigned int msr,
msr               102 arch/x86/include/asm/msr.h 		     : "c" (msr), "0" (low), "d" (high),
msr               140 arch/x86/include/asm/msr.h #define rdmsr(msr, low, high)					\
msr               142 arch/x86/include/asm/msr.h 	u64 __val = native_read_msr((msr));			\
msr               147 arch/x86/include/asm/msr.h static inline void wrmsr(unsigned msr, unsigned low, unsigned high)
msr               149 arch/x86/include/asm/msr.h 	native_write_msr(msr, low, high);
msr               152 arch/x86/include/asm/msr.h #define rdmsrl(msr, val)			\
msr               153 arch/x86/include/asm/msr.h 	((val) = native_read_msr((msr)))
msr               155 arch/x86/include/asm/msr.h #define wrmsrl(msr, val)						\
msr               156 arch/x86/include/asm/msr.h 	native_write_msr((msr), (u32)((u64)(val)), (u32)((u64)(val) >> 32))
msr               159 arch/x86/include/asm/msr.h static inline int wrmsr_safe(unsigned msr, unsigned low, unsigned high)
msr               161 arch/x86/include/asm/msr.h 	return native_write_msr_safe(msr, low, high);
msr               165 arch/x86/include/asm/msr.h #define rdmsr_safe(msr, low, high)				\
msr               168 arch/x86/include/asm/msr.h 	u64 __val = native_read_msr_safe((msr), &__err);	\
msr               174 arch/x86/include/asm/msr.h static inline int rdmsrl_safe(unsigned msr, unsigned long long *p)
msr               178 arch/x86/include/asm/msr.h 	*p = native_read_msr_safe(msr, &err);
msr               208 arch/x86/include/asm/msr.h #define wrmsrl_safe(msr, val) wrmsr_safe((msr), (u32)(val),		\
msr               215 arch/x86/include/asm/msr.h struct msr *msrs_alloc(void);
msr               216 arch/x86/include/asm/msr.h void msrs_free(struct msr *msrs);
msr               217 arch/x86/include/asm/msr.h int msr_set_bit(u32 msr, u8 bit);
msr               218 arch/x86/include/asm/msr.h int msr_clear_bit(u32 msr, u8 bit);
msr               225 arch/x86/include/asm/msr.h void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs);
msr               226 arch/x86/include/asm/msr.h void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr *msrs);
msr               255 arch/x86/include/asm/msr.h 				struct msr *msrs)
msr               260 arch/x86/include/asm/msr.h 				struct msr *msrs)
msr               126 arch/x86/include/asm/paravirt.h static inline u64 paravirt_read_msr(unsigned msr, int *err)
msr               128 arch/x86/include/asm/paravirt.h 	return PVOP_CALL2(u64, pv_cpu_ops.read_msr, msr, err);
msr               131 arch/x86/include/asm/paravirt.h static inline int paravirt_write_msr(unsigned msr, unsigned low, unsigned high)
msr               133 arch/x86/include/asm/paravirt.h 	return PVOP_CALL3(int, pv_cpu_ops.write_msr, msr, low, high);
msr               137 arch/x86/include/asm/paravirt.h #define rdmsr(msr, val1, val2)			\
msr               140 arch/x86/include/asm/paravirt.h 	u64 _l = paravirt_read_msr(msr, &_err);	\
msr               145 arch/x86/include/asm/paravirt.h #define wrmsr(msr, val1, val2)			\
msr               147 arch/x86/include/asm/paravirt.h 	paravirt_write_msr(msr, val1, val2);	\
msr               150 arch/x86/include/asm/paravirt.h #define rdmsrl(msr, val)			\
msr               153 arch/x86/include/asm/paravirt.h 	val = paravirt_read_msr(msr, &_err);	\
msr               156 arch/x86/include/asm/paravirt.h #define wrmsrl(msr, val)	wrmsr(msr, (u32)((u64)(val)), ((u64)(val))>>32)
msr               157 arch/x86/include/asm/paravirt.h #define wrmsr_safe(msr, a, b)	paravirt_write_msr(msr, a, b)
msr               160 arch/x86/include/asm/paravirt.h #define rdmsr_safe(msr, a, b)			\
msr               163 arch/x86/include/asm/paravirt.h 	u64 _l = paravirt_read_msr(msr, &_err);	\
msr               169 arch/x86/include/asm/paravirt.h static inline int rdmsrl_safe(unsigned msr, unsigned long long *p)
msr               173 arch/x86/include/asm/paravirt.h 	*p = paravirt_read_msr(msr, &err);
msr               156 arch/x86/include/asm/paravirt_types.h 	u64 (*read_msr)(unsigned int msr, int *err);
msr               157 arch/x86/include/asm/paravirt_types.h 	int (*write_msr)(unsigned int msr, unsigned low, unsigned high);
msr               254 arch/x86/include/asm/perf_event.h 	unsigned msr;
msr               316 include/linux/isdn.h   int                   msr;             /* Modem status register          */
msr               829 include/uapi/linux/kvm.h 	__u32 msr;