/linux-6.1.9/arch/powerpc/kernel/ptrace/ |
D | ptrace-vsx.c | 32 buf[32] = target->thread.fp_state.fpscr; in fpr_get() 60 buf[32] = target->thread.fp_state.fpscr; in fpr_set() 69 target->thread.fp_state.fpscr = buf[32]; in fpr_set() 109 buf[i] = target->thread.fp_state.fpr[i][TS_VSRLOWOFFSET]; in vsr_get() 139 buf[i] = target->thread.fp_state.fpr[i][TS_VSRLOWOFFSET]; in vsr_set() 145 target->thread.fp_state.fpr[i][TS_VSRLOWOFFSET] = buf[i]; in vsr_set()
|
D | ptrace-fpu.c | 23 *data = ((u32 *)child->thread.fp_state.fpr)[fpidx]; in ptrace_get_fpr() 27 *data = child->thread.fp_state.fpscr; in ptrace_get_fpr() 49 ((u32 *)child->thread.fp_state.fpr)[fpidx] = data; in ptrace_put_fpr() 53 child->thread.fp_state.fpscr = data; in ptrace_put_fpr()
|
D | ptrace32.c | 97 tmp = ((unsigned int *)child->thread.fp_state.fpr) in compat_arch_ptrace() 138 tmp = child->thread.fp_state.fpr[numReg - PT_FPR0][0]; in compat_arch_ptrace() 197 ((unsigned int *)child->thread.fp_state.fpr) in compat_arch_ptrace() 240 tmp = &child->thread.fp_state.fpr[numReg - PT_FPR0][0]; in compat_arch_ptrace()
|
D | ptrace-novsx.c | 30 return membuf_write(&to, &target->thread.fp_state, 33 * sizeof(u64)); in fpr_get() 60 &target->thread.fp_state, 0, -1); in fpr_set()
|
/linux-6.1.9/arch/arm64/kvm/ |
D | fpsimd.c | 83 vcpu->arch.fp_state = FP_STATE_HOST_OWNED; in kvm_arch_vcpu_load_fp() 104 vcpu->arch.fp_state = FP_STATE_FREE; in kvm_arch_vcpu_load_fp() 120 vcpu->arch.fp_state = FP_STATE_FREE; in kvm_arch_vcpu_ctxflush_fp() 134 if (vcpu->arch.fp_state == FP_STATE_GUEST_OWNED) { in kvm_arch_vcpu_ctxsync_fp() 177 if (vcpu->arch.fp_state == FP_STATE_GUEST_OWNED) { in kvm_arch_vcpu_put_fp()
|
/linux-6.1.9/arch/powerpc/kernel/ |
D | signal.h | 58 unsafe_put_user(__t->thread.fp_state.fpscr, &buf[i], label); \ 67 unsafe_put_user(__t->thread.fp_state.fpr[i][TS_VSRLOWOFFSET], \ 78 unsafe_get_user(__t->thread.fp_state.fpscr, &buf[i], label); \ 87 unsafe_get_user(__t->thread.fp_state.fpr[i][TS_VSRLOWOFFSET], \ 135 unsafe_copy_to_user(to, (task)->thread.fp_state.fpr, \ 139 unsafe_copy_from_user((task)->thread.fp_state.fpr, from, \ 145 return __copy_to_user(to, task->thread.fp_state.fpr, in copy_fpr_to_user() 152 return __copy_from_user(task->thread.fp_state.fpr, from, in copy_fpr_from_user()
|
D | signal.c | 38 buf[i] = task->thread.fp_state.fpscr; in copy_fpr_to_user() 52 task->thread.fp_state.fpscr = buf[i]; in copy_fpr_from_user() 65 buf[i] = task->thread.fp_state.fpr[i][TS_VSRLOWOFFSET]; in copy_vsx_to_user() 78 task->thread.fp_state.fpr[i][TS_VSRLOWOFFSET] = buf[i]; in copy_vsx_from_user()
|
D | signal_32.c | 527 current->thread.fp_state.fpr[i][TS_VSRLOWOFFSET] = 0; in restore_user_regs() 628 current->thread.fp_state.fpr[i][TS_VSRLOWOFFSET] = 0; in restore_tm_user_regs() 699 load_fp_state(¤t->thread.fp_state); in restore_tm_user_regs() 800 tsk->thread.fp_state.fpscr = 0; /* turn off all fp exceptions */ in handle_rt_signal32() 891 tsk->thread.fp_state.fpscr = 0; /* turn off all fp exceptions */ in handle_signal32()
|
D | signal_64.c | 423 tsk->thread.fp_state.fpr[i][TS_VSRLOWOFFSET] = 0; in __unsafe_restore_sigcontext() 560 tsk->thread.fp_state.fpr[i][TS_VSRLOWOFFSET] = 0; in restore_tm_sigcontexts() 601 load_fp_state(&tsk->thread.fp_state); in restore_tm_sigcontexts() 921 tsk->thread.fp_state.fpscr = 0; in handle_rt_signal64()
|
D | process.c | 465 load_fp_state(¤t->thread.fp_state); in do_restore_fp() 938 memcpy(&thr->ckfp_state, &thr->fp_state, in tm_reclaim_thread() 1919 memset(¤t->thread.fp_state, 0, sizeof(current->thread.fp_state)); in start_thread()
|
D | traps.c | 928 vdst = (u8 *)¤t->thread.fp_state.fpr[t][0]; in p9_hmi_special_emu() 1201 code = __parse_fpscr(current->thread.fp_state.fpscr); in parse_fpe() 1460 code = __parse_fpscr(current->thread.fp_state.fpscr); in emulate_math()
|
/linux-6.1.9/arch/arm/include/asm/ |
D | fpstate.h | 67 union fp_state { union 75 #define FP_SIZE (sizeof(union fp_state) / sizeof(int))
|
D | thread_info.h | 71 union fp_state fpstate __attribute__((aligned(8)));
|
/linux-6.1.9/arch/ia64/kernel/ |
D | traps.c | 236 fp_state_t fp_state; in fp_emulate() local 242 memset(&fp_state, 0, sizeof(fp_state_t)); in fp_emulate() 249 fp_state.bitmask_low64 = 0xfc0; /* bit6..bit11 */ in fp_emulate() 251 fp_state.fp_state_low_volatile = (fp_state_low_volatile_t *) ®s->f6; in fp_emulate() 266 (unsigned long *) ifs, &fp_state); in fp_emulate()
|
/linux-6.1.9/arch/arm64/kvm/hyp/include/hyp/ |
D | switch.h | 43 return vcpu->arch.fp_state == FP_STATE_GUEST_OWNED; in guest_owns_fp_regs() 202 if (vcpu->arch.fp_state == FP_STATE_HOST_OWNED) in kvm_hyp_handle_fpsimd() 215 vcpu->arch.fp_state = FP_STATE_GUEST_OWNED; in kvm_hyp_handle_fpsimd()
|
/linux-6.1.9/arch/ia64/include/asm/ |
D | fpswa.h | 60 unsigned long *ifs, fp_state_t *fp_state);
|
/linux-6.1.9/arch/arm/nwfpe/ |
D | fpa11.c | 73 void nwfpe_init_fpa(union fp_state *fp) in nwfpe_init_fpa()
|
D | fpa11.h | 83 extern void nwfpe_init_fpa(union fp_state *fp);
|
D | fpmodule.c | 76 if (sizeof(FPA11) > sizeof(union fp_state)) { in fpe_init()
|
/linux-6.1.9/arch/arm64/kvm/hyp/nvhe/ |
D | switch.c | 102 if (vcpu_has_sve(vcpu) && (vcpu->arch.fp_state == FP_STATE_GUEST_OWNED)) in __deactivate_traps() 314 if (vcpu->arch.fp_state == FP_STATE_GUEST_OWNED) in __kvm_vcpu_run()
|
/linux-6.1.9/arch/powerpc/include/asm/ |
D | processor.h | 79 #define TS_FPR(i) fp_state.fpr[i][TS_FPROFFSET] 168 struct thread_fp_state fp_state; member
|
D | sfp-machine.h | 128 #define __FPU_FPSCR (current->thread.fp_state.fpscr)
|
/linux-6.1.9/arch/arm64/kvm/hyp/vhe/ |
D | switch.c | 170 if (vcpu->arch.fp_state == FP_STATE_GUEST_OWNED) in __kvm_vcpu_run_vhe()
|
/linux-6.1.9/arch/arm/kernel/ |
D | ptrace.c | 761 0, sizeof(union fp_state), in arch_ptrace() 768 0, sizeof(union fp_state), in arch_ptrace()
|
D | process.c | 228 memset(&thread->fpstate, 0, sizeof(union fp_state)); in flush_thread()
|