Searched refs:csr_read (Results 1 – 18 of 18) sorted by relevance
13 context->scratch = csr_read(CSR_SCRATCH); in suspend_save_csrs()14 context->tvec = csr_read(CSR_TVEC); in suspend_save_csrs()15 context->ie = csr_read(CSR_IE); in suspend_save_csrs()28 context->satp = csr_read(CSR_SATP); in suspend_save_csrs()
31 cpu_mfr_info->vendor_id = csr_read(CSR_MVENDORID); in riscv_fill_cpu_mfr_info()32 cpu_mfr_info->arch_id = csr_read(CSR_MARCHID); in riscv_fill_cpu_mfr_info()33 cpu_mfr_info->imp_id = csr_read(CSR_MIMPID); in riscv_fill_cpu_mfr_info()
91 ci->mvendorid = csr_read(CSR_MVENDORID); in riscv_cpuinfo_starting()92 ci->marchid = csr_read(CSR_MARCHID); in riscv_cpuinfo_starting()93 ci->mimpid = csr_read(CSR_MIMPID); in riscv_cpuinfo_starting()
99 unsigned long tmp = csr_read(CSR_STATUS); in compat_mode_detect()103 (csr_read(CSR_STATUS) & SR_UXL) == SR_UXL_32; in compat_mode_detect()
243 unsigned long addr = csr_read(mtval); in handle_misaligned_load()326 unsigned long addr = csr_read(mtval); in handle_misaligned_store()
699 csr->vsie = csr_read(CSR_VSIE); in kvm_riscv_vcpu_sync_interrupts()702 hvip = csr_read(CSR_HVIP); in kvm_riscv_vcpu_sync_interrupts()865 csr->vsstatus = csr_read(CSR_VSSTATUS); in kvm_arch_vcpu_put()866 csr->vsie = csr_read(CSR_VSIE); in kvm_arch_vcpu_put()867 csr->vstvec = csr_read(CSR_VSTVEC); in kvm_arch_vcpu_put()868 csr->vsscratch = csr_read(CSR_VSSCRATCH); in kvm_arch_vcpu_put()869 csr->vsepc = csr_read(CSR_VSEPC); in kvm_arch_vcpu_put()870 csr->vscause = csr_read(CSR_VSCAUSE); in kvm_arch_vcpu_put()871 csr->vstval = csr_read(CSR_VSTVAL); in kvm_arch_vcpu_put()872 csr->hvip = csr_read(CSR_HVIP); in kvm_arch_vcpu_put()[all …]
28 old = csr_read(CSR_HGATP); in kvm_riscv_gstage_vmid_detect()30 vmid_bits = csr_read(CSR_HGATP); in kvm_riscv_gstage_vmid_detect()
138 unsigned long vsstatus = csr_read(CSR_VSSTATUS); in kvm_riscv_vcpu_trap_redirect()162 vcpu->arch.guest_context.sepc = csr_read(CSR_VSTVEC); in kvm_riscv_vcpu_trap_redirect()
331 t->next_cycles = csr_read(CSR_VSTIMECMP); in kvm_riscv_vcpu_timer_sync()332 t->next_cycles |= (u64)csr_read(CSR_VSTIMECMPH) << 32; in kvm_riscv_vcpu_timer_sync()334 t->next_cycles = csr_read(CSR_VSTIMECMP); in kvm_riscv_vcpu_timer_sync()
768 if ((csr_read(CSR_HGATP) >> HGATP_MODE_SHIFT) == HGATP_MODE_SV57X4) { in kvm_riscv_gstage_mode_detect()776 if ((csr_read(CSR_HGATP) >> HGATP_MODE_SHIFT) == HGATP_MODE_SV48X4) { in kvm_riscv_gstage_mode_detect()
53 return csr_read(CSR_TIME); in get_cycles()59 return csr_read(CSR_TIMEH); in get_cycles_hi()
16 return csr_read(CSR_STATUS); in arch_local_save_flags()
355 #define csr_read(csr) \ macro
231 old = csr_read(CSR_SATP); in asids_init()234 asid_bits = (csr_read(CSR_SATP) >> SATP_ASID_SHIFT) & SATP_ASID_MASK; in asids_init()
125 pfn = csr_read(CSR_SATP) & SATP_PPN; in vmalloc_fault()
71 return csr_read(CSR_TIME); in __arch_get_hw_counter()
24 __val = csr_read(__csr_num); \ in csr_read_num()
591 overflow = csr_read(CSR_SSCOUNTOVF); in pmu_sbi_ovf_handler()