fpu 205 arch/x86/include/asm/fpu-internal.h static inline void fpu_fxsave(struct fpu *fpu) fpu 208 arch/x86/include/asm/fpu-internal.h asm volatile( "fxsave %[fx]" : [fx] "=m" (fpu->state->fxsave)); fpu 210 arch/x86/include/asm/fpu-internal.h asm volatile("fxsaveq %0" : "=m" (fpu->state->fxsave)); fpu 234 arch/x86/include/asm/fpu-internal.h : "=m" (fpu->state->fxsave) fpu 235 arch/x86/include/asm/fpu-internal.h : [fx] "R" (&fpu->state->fxsave)); fpu 243 arch/x86/include/asm/fpu-internal.h static inline int fpu_save_init(struct fpu *fpu) fpu 246 arch/x86/include/asm/fpu-internal.h fpu_xsave(fpu); fpu 251 arch/x86/include/asm/fpu-internal.h if (!(fpu->state->xsave.xsave_hdr.xstate_bv & XSTATE_FP)) fpu 254 arch/x86/include/asm/fpu-internal.h fpu_fxsave(fpu); fpu 257 arch/x86/include/asm/fpu-internal.h : [fx] "=m" (fpu->state->fsave)); fpu 269 arch/x86/include/asm/fpu-internal.h if (unlikely(fpu->state->fxsave.swd & X87_FSW_ES)) { fpu 278 arch/x86/include/asm/fpu-internal.h return fpu_save_init(&tsk->thread.fpu); fpu 281 arch/x86/include/asm/fpu-internal.h static inline int fpu_restore_checking(struct fpu *fpu) fpu 284 arch/x86/include/asm/fpu-internal.h return fpu_xrstor_checking(&fpu->state->xsave); fpu 286 arch/x86/include/asm/fpu-internal.h return fxrstor_checking(&fpu->state->fxsave); fpu 288 arch/x86/include/asm/fpu-internal.h return frstor_checking(&fpu->state->fsave); fpu 301 arch/x86/include/asm/fpu-internal.h : : [addr] "m" (tsk->thread.fpu.has_fpu)); fpu 304 arch/x86/include/asm/fpu-internal.h return fpu_restore_checking(&tsk->thread.fpu); fpu 314 arch/x86/include/asm/fpu-internal.h return tsk->thread.fpu.has_fpu; fpu 320 arch/x86/include/asm/fpu-internal.h tsk->thread.fpu.has_fpu = 0; fpu 327 arch/x86/include/asm/fpu-internal.h tsk->thread.fpu.has_fpu = 1; fpu 416 arch/x86/include/asm/fpu-internal.h cpu == new->thread.fpu.last_cpu; fpu 421 arch/x86/include/asm/fpu-internal.h fpu_switch_t fpu; fpu 427 arch/x86/include/asm/fpu-internal.h fpu.preload = tsk_used_math(new) && (use_eager_fpu() || fpu 432 arch/x86/include/asm/fpu-internal.h old->thread.fpu.last_cpu = cpu; fpu 433 arch/x86/include/asm/fpu-internal.h old->thread.fpu.has_fpu = 0; /* But leave fpu_owner_task! */ fpu 436 arch/x86/include/asm/fpu-internal.h if (fpu.preload) { fpu 439 arch/x86/include/asm/fpu-internal.h prefetch(new->thread.fpu.state); fpu 444 arch/x86/include/asm/fpu-internal.h old->thread.fpu.last_cpu = ~0; fpu 445 arch/x86/include/asm/fpu-internal.h if (fpu.preload) { fpu 448 arch/x86/include/asm/fpu-internal.h fpu.preload = 0; fpu 450 arch/x86/include/asm/fpu-internal.h prefetch(new->thread.fpu.state); fpu 454 arch/x86/include/asm/fpu-internal.h return fpu; fpu 463 arch/x86/include/asm/fpu-internal.h static inline void switch_fpu_finish(struct task_struct *new, fpu_switch_t fpu) fpu 465 arch/x86/include/asm/fpu-internal.h if (fpu.preload) { fpu 513 arch/x86/include/asm/fpu-internal.h xsave_state_booting(&tsk->thread.fpu.state->xsave, -1); fpu 515 arch/x86/include/asm/fpu-internal.h xsave_state(&tsk->thread.fpu.state->xsave, -1); fpu 517 arch/x86/include/asm/fpu-internal.h fpu_fxsave(&tsk->thread.fpu); fpu 544 arch/x86/include/asm/fpu-internal.h return tsk->thread.fpu.state->fxsave.cwd; fpu 546 arch/x86/include/asm/fpu-internal.h return (unsigned short)tsk->thread.fpu.state->fsave.cwd; fpu 553 arch/x86/include/asm/fpu-internal.h return tsk->thread.fpu.state->fxsave.swd; fpu 555 arch/x86/include/asm/fpu-internal.h return (unsigned short)tsk->thread.fpu.state->fsave.swd; fpu 562 arch/x86/include/asm/fpu-internal.h return tsk->thread.fpu.state->fxsave.mxcsr; fpu 568 arch/x86/include/asm/fpu-internal.h static bool fpu_allocated(struct fpu *fpu) fpu 570 arch/x86/include/asm/fpu-internal.h return fpu->state != NULL; fpu 573 arch/x86/include/asm/fpu-internal.h static inline int fpu_alloc(struct fpu *fpu) fpu 575 arch/x86/include/asm/fpu-internal.h if (fpu_allocated(fpu)) fpu 577 arch/x86/include/asm/fpu-internal.h fpu->state = kmem_cache_alloc(task_xstate_cachep, GFP_KERNEL); fpu 578 arch/x86/include/asm/fpu-internal.h if (!fpu->state) fpu 580 arch/x86/include/asm/fpu-internal.h WARN_ON((unsigned long)fpu->state & 15); fpu 584 arch/x86/include/asm/fpu-internal.h static inline void fpu_free(struct fpu *fpu) fpu 586 arch/x86/include/asm/fpu-internal.h if (fpu->state) { fpu 587 arch/x86/include/asm/fpu-internal.h kmem_cache_free(task_xstate_cachep, fpu->state); fpu 588 arch/x86/include/asm/fpu-internal.h fpu->state = NULL; fpu 595 arch/x86/include/asm/fpu-internal.h memset(&dst->thread.fpu.state->xsave, 0, xstate_size); fpu 598 arch/x86/include/asm/fpu-internal.h struct fpu *dfpu = &dst->thread.fpu; fpu 599 arch/x86/include/asm/fpu-internal.h struct fpu *sfpu = &src->thread.fpu; fpu 22 arch/x86/include/asm/i387.h extern void fpu_finit(struct fpu *fpu); fpu 97 arch/x86/include/asm/i387.h return current->thread.fpu.has_fpu; fpu 394 arch/x86/include/asm/kvm_host.h struct fpu guest_fpu; fpu 500 arch/x86/include/asm/processor.h struct fpu fpu; fpu 196 arch/x86/include/asm/xsave.h static inline void fpu_xsave(struct fpu *fpu) fpu 198 arch/x86/include/asm/xsave.h xsave_state(&fpu->state->xsave, -1); fpu 42 include/linux/elfcore.h static inline int elf_core_copy_task_fpregs(struct task_struct *t, struct pt_regs *regs, elf_fpregset_t *fpu) fpu 45 include/linux/elfcore.h return ELF_CORE_COPY_FPREGS(t, fpu); fpu 47 include/linux/elfcore.h return dump_fpu(regs, fpu); fpu 612 include/linux/kvm_host.h int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu); fpu 613 include/linux/kvm_host.h int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu);