/linux-5.19.10/include/kvm/ |
D | arm_pmu.h | 51 u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu, u64 select_idx); 52 void kvm_pmu_set_counter_value(struct kvm_vcpu *vcpu, u64 select_idx, u64 val); 53 u64 kvm_pmu_valid_counter_mask(struct kvm_vcpu *vcpu); 54 u64 kvm_pmu_get_pmceid(struct kvm_vcpu *vcpu, bool pmceid1); 55 void kvm_pmu_vcpu_init(struct kvm_vcpu *vcpu); 56 void kvm_pmu_vcpu_reset(struct kvm_vcpu *vcpu); 57 void kvm_pmu_vcpu_destroy(struct kvm_vcpu *vcpu); 58 void kvm_pmu_disable_counter_mask(struct kvm_vcpu *vcpu, u64 val); 59 void kvm_pmu_enable_counter_mask(struct kvm_vcpu *vcpu, u64 val); 60 void kvm_pmu_flush_hwstate(struct kvm_vcpu *vcpu); [all …]
|
D | arm_arch_timer.h | 27 struct kvm_vcpu *vcpu; 64 int kvm_timer_enable(struct kvm_vcpu *vcpu); 65 int kvm_timer_vcpu_reset(struct kvm_vcpu *vcpu); 66 void kvm_timer_vcpu_init(struct kvm_vcpu *vcpu); 67 void kvm_timer_sync_user(struct kvm_vcpu *vcpu); 68 bool kvm_timer_should_notify_user(struct kvm_vcpu *vcpu); 69 void kvm_timer_update_run(struct kvm_vcpu *vcpu); 70 void kvm_timer_vcpu_terminate(struct kvm_vcpu *vcpu); 72 u64 kvm_arm_timer_get_reg(struct kvm_vcpu *, u64 regid); 73 int kvm_arm_timer_set_reg(struct kvm_vcpu *, u64 regid, u64 value); [all …]
|
/linux-5.19.10/arch/powerpc/include/asm/ |
D | kvm_ppc.h | 61 extern int kvmppc_vcpu_run(struct kvm_vcpu *vcpu); 62 extern int __kvmppc_vcpu_run(struct kvm_vcpu *vcpu); 65 extern void kvmppc_dump_vcpu(struct kvm_vcpu *vcpu); 66 extern int kvmppc_handle_load(struct kvm_vcpu *vcpu, 69 extern int kvmppc_handle_loads(struct kvm_vcpu *vcpu, 72 extern int kvmppc_handle_vsx_load(struct kvm_vcpu *vcpu, 75 extern int kvmppc_handle_vmx_load(struct kvm_vcpu *vcpu, 77 extern int kvmppc_handle_vmx_store(struct kvm_vcpu *vcpu, 79 extern int kvmppc_handle_store(struct kvm_vcpu *vcpu, 82 extern int kvmppc_handle_vsx_store(struct kvm_vcpu *vcpu, [all …]
|
D | kvm_book3s.h | 78 struct kvm_vcpu *runnable_threads[MAX_SMT_THREADS]; 85 struct kvm_vcpu *runner; 145 extern void kvmppc_mmu_pte_flush(struct kvm_vcpu *vcpu, ulong ea, ulong ea_mask); 146 extern void kvmppc_mmu_pte_vflush(struct kvm_vcpu *vcpu, u64 vp, u64 vp_mask); 147 extern void kvmppc_mmu_pte_pflush(struct kvm_vcpu *vcpu, ulong pa_start, ulong pa_end); 148 extern void kvmppc_set_msr(struct kvm_vcpu *vcpu, u64 new_msr); 149 extern void kvmppc_mmu_book3s_64_init(struct kvm_vcpu *vcpu); 150 extern void kvmppc_mmu_book3s_32_init(struct kvm_vcpu *vcpu); 151 extern void kvmppc_mmu_book3s_hv_init(struct kvm_vcpu *vcpu); 152 extern int kvmppc_mmu_map_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte, [all …]
|
D | kvm_booke.h | 26 static inline void kvmppc_set_gpr(struct kvm_vcpu *vcpu, int num, ulong val) in kvmppc_set_gpr() 31 static inline ulong kvmppc_get_gpr(struct kvm_vcpu *vcpu, int num) in kvmppc_get_gpr() 36 static inline void kvmppc_set_cr(struct kvm_vcpu *vcpu, u32 val) in kvmppc_set_cr() 41 static inline u32 kvmppc_get_cr(struct kvm_vcpu *vcpu) in kvmppc_get_cr() 46 static inline void kvmppc_set_xer(struct kvm_vcpu *vcpu, ulong val) in kvmppc_set_xer() 51 static inline ulong kvmppc_get_xer(struct kvm_vcpu *vcpu) in kvmppc_get_xer() 56 static inline bool kvmppc_need_byteswap(struct kvm_vcpu *vcpu) in kvmppc_need_byteswap() 62 static inline void kvmppc_set_ctr(struct kvm_vcpu *vcpu, ulong val) in kvmppc_set_ctr() 67 static inline ulong kvmppc_get_ctr(struct kvm_vcpu *vcpu) in kvmppc_get_ctr() 72 static inline void kvmppc_set_lr(struct kvm_vcpu *vcpu, ulong val) in kvmppc_set_lr() [all …]
|
/linux-5.19.10/arch/x86/kvm/ |
D | lapic.h | 50 struct kvm_vcpu *vcpu; 72 int kvm_create_lapic(struct kvm_vcpu *vcpu, int timer_advance_ns); 73 void kvm_free_lapic(struct kvm_vcpu *vcpu); 75 int kvm_apic_has_interrupt(struct kvm_vcpu *vcpu); 76 int kvm_apic_accept_pic_intr(struct kvm_vcpu *vcpu); 77 int kvm_get_apic_interrupt(struct kvm_vcpu *vcpu); 78 int kvm_apic_accept_events(struct kvm_vcpu *vcpu); 79 void kvm_lapic_reset(struct kvm_vcpu *vcpu, bool init_event); 80 u64 kvm_lapic_get_cr8(struct kvm_vcpu *vcpu); 81 void kvm_lapic_set_tpr(struct kvm_vcpu *vcpu, unsigned long cr8); [all …]
|
D | x86.h | 61 void kvm_service_local_tlb_flush_requests(struct kvm_vcpu *vcpu); 62 int kvm_check_nested_events(struct kvm_vcpu *vcpu); 64 static inline void kvm_clear_exception_queue(struct kvm_vcpu *vcpu) in kvm_clear_exception_queue() 70 static inline void kvm_queue_interrupt(struct kvm_vcpu *vcpu, u8 vector, in kvm_queue_interrupt() 78 static inline void kvm_clear_interrupt_queue(struct kvm_vcpu *vcpu) in kvm_clear_interrupt_queue() 83 static inline bool kvm_event_needs_reinjection(struct kvm_vcpu *vcpu) in kvm_event_needs_reinjection() 94 static inline bool is_protmode(struct kvm_vcpu *vcpu) in is_protmode() 99 static inline int is_long_mode(struct kvm_vcpu *vcpu) in is_long_mode() 108 static inline bool is_64_bit_mode(struct kvm_vcpu *vcpu) in is_64_bit_mode() 120 static inline bool is_64_bit_hypercall(struct kvm_vcpu *vcpu) in is_64_bit_hypercall() [all …]
|
D | xen.h | 17 int __kvm_xen_has_interrupt(struct kvm_vcpu *vcpu); 18 void kvm_xen_inject_pending_events(struct kvm_vcpu *vcpu); 19 int kvm_xen_vcpu_set_attr(struct kvm_vcpu *vcpu, struct kvm_xen_vcpu_attr *data); 20 int kvm_xen_vcpu_get_attr(struct kvm_vcpu *vcpu, struct kvm_xen_vcpu_attr *data); 24 int kvm_xen_write_hypercall_page(struct kvm_vcpu *vcpu, u64 data); 28 void kvm_xen_init_vcpu(struct kvm_vcpu *vcpu); 29 void kvm_xen_destroy_vcpu(struct kvm_vcpu *vcpu); 49 static inline int kvm_xen_has_interrupt(struct kvm_vcpu *vcpu) in kvm_xen_has_interrupt() 59 static inline bool kvm_xen_has_pending_events(struct kvm_vcpu *vcpu) in kvm_xen_has_pending_events() 65 static inline bool kvm_xen_timer_enabled(struct kvm_vcpu *vcpu) in kvm_xen_timer_enabled() [all …]
|
D | hyperv.h | 58 static inline struct kvm_vcpu_hv *to_hv_vcpu(struct kvm_vcpu *vcpu) in to_hv_vcpu() 63 static inline struct kvm_vcpu_hv_synic *to_hv_synic(struct kvm_vcpu *vcpu) in to_hv_synic() 70 static inline struct kvm_vcpu *hv_synic_to_vcpu(struct kvm_vcpu_hv_synic *synic) in hv_synic_to_vcpu() 77 static inline struct kvm_hv_syndbg *to_hv_syndbg(struct kvm_vcpu *vcpu) in to_hv_syndbg() 82 static inline u32 kvm_hv_get_vpindex(struct kvm_vcpu *vcpu) in kvm_hv_get_vpindex() 89 int kvm_hv_set_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 data, bool host); 90 int kvm_hv_get_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata, bool host); 92 static inline bool kvm_hv_hypercall_enabled(struct kvm_vcpu *vcpu) in kvm_hv_hypercall_enabled() 97 int kvm_hv_hypercall(struct kvm_vcpu *vcpu); 101 void kvm_hv_synic_send_eoi(struct kvm_vcpu *vcpu, int vector); [all …]
|
D | cpuid.h | 14 void kvm_update_cpuid_runtime(struct kvm_vcpu *vcpu); 15 void kvm_update_pv_runtime(struct kvm_vcpu *vcpu); 16 struct kvm_cpuid_entry2 *kvm_find_cpuid_entry(struct kvm_vcpu *vcpu, 21 int kvm_vcpu_ioctl_set_cpuid(struct kvm_vcpu *vcpu, 24 int kvm_vcpu_ioctl_set_cpuid2(struct kvm_vcpu *vcpu, 27 int kvm_vcpu_ioctl_get_cpuid2(struct kvm_vcpu *vcpu, 30 bool kvm_cpuid(struct kvm_vcpu *vcpu, u32 *eax, u32 *ebx, 35 int cpuid_query_maxphyaddr(struct kvm_vcpu *vcpu); 36 u64 kvm_vcpu_reserved_gpa_bits_raw(struct kvm_vcpu *vcpu); 38 static inline int cpuid_maxphyaddr(struct kvm_vcpu *vcpu) in cpuid_maxphyaddr() [all …]
|
D | pmu.h | 8 #define pmu_to_vcpu(pmu) (container_of((pmu), struct kvm_vcpu, arch.pmu)) 28 struct kvm_pmc *(*rdpmc_ecx_to_pmc)(struct kvm_vcpu *vcpu, 30 struct kvm_pmc *(*msr_idx_to_pmc)(struct kvm_vcpu *vcpu, u32 msr); 31 bool (*is_valid_rdpmc_ecx)(struct kvm_vcpu *vcpu, unsigned int idx); 32 bool (*is_valid_msr)(struct kvm_vcpu *vcpu, u32 msr); 33 int (*get_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr_info); 34 int (*set_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr_info); 35 void (*refresh)(struct kvm_vcpu *vcpu); 36 void (*init)(struct kvm_vcpu *vcpu); 37 void (*reset)(struct kvm_vcpu *vcpu); [all …]
|
D | kvm_cache_regs.h | 19 static __always_inline unsigned long kvm_##lname##_read(struct kvm_vcpu *vcpu)\ 23 static __always_inline void kvm_##lname##_write(struct kvm_vcpu *vcpu, \ 53 static inline bool kvm_register_is_available(struct kvm_vcpu *vcpu, in BUILD_KVM_GPR_ACCESSORS() 59 static inline bool kvm_register_is_dirty(struct kvm_vcpu *vcpu, in kvm_register_is_dirty() 65 static inline void kvm_register_mark_available(struct kvm_vcpu *vcpu, in kvm_register_mark_available() 71 static inline void kvm_register_mark_dirty(struct kvm_vcpu *vcpu, in kvm_register_mark_dirty() 83 static inline unsigned long kvm_register_read_raw(struct kvm_vcpu *vcpu, int reg) in kvm_register_read_raw() 94 static inline void kvm_register_write_raw(struct kvm_vcpu *vcpu, int reg, in kvm_register_write_raw() 104 static inline unsigned long kvm_rip_read(struct kvm_vcpu *vcpu) in kvm_rip_read() 109 static inline void kvm_rip_write(struct kvm_vcpu *vcpu, unsigned long val) in kvm_rip_write() [all …]
|
/linux-5.19.10/arch/x86/include/asm/ |
D | kvm_host.h | 428 unsigned long (*get_guest_pgd)(struct kvm_vcpu *vcpu); 429 u64 (*get_pdptr)(struct kvm_vcpu *vcpu, int index); 430 int (*page_fault)(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault); 431 void (*inject_page_fault)(struct kvm_vcpu *vcpu, 433 gpa_t (*gva_to_gpa)(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, 436 int (*sync_page)(struct kvm_vcpu *vcpu, 438 void (*invlpg)(struct kvm_vcpu *vcpu, gva_t gva, hpa_t root_hpa); 492 struct kvm_vcpu *vcpu; 585 struct kvm_vcpu *vcpu; 755 int (*complete_userspace_io)(struct kvm_vcpu *vcpu); [all …]
|
/linux-5.19.10/arch/s390/kvm/ |
D | kvm-s390.h | 60 static inline void kvm_s390_set_cpuflags(struct kvm_vcpu *vcpu, u32 flags) in kvm_s390_set_cpuflags() 65 static inline void kvm_s390_clear_cpuflags(struct kvm_vcpu *vcpu, u32 flags) in kvm_s390_clear_cpuflags() 70 static inline bool kvm_s390_test_cpuflags(struct kvm_vcpu *vcpu, u32 flags) in kvm_s390_test_cpuflags() 75 static inline int is_vcpu_stopped(struct kvm_vcpu *vcpu) in is_vcpu_stopped() 80 static inline int is_vcpu_idle(struct kvm_vcpu *vcpu) in is_vcpu_idle() 97 static inline u32 kvm_s390_get_prefix(struct kvm_vcpu *vcpu) in kvm_s390_get_prefix() 102 static inline void kvm_s390_set_prefix(struct kvm_vcpu *vcpu, u32 prefix) in kvm_s390_set_prefix() 111 static inline u64 kvm_s390_get_base_disp_s(struct kvm_vcpu *vcpu, u8 *ar) in kvm_s390_get_base_disp_s() 122 static inline void kvm_s390_get_base_disp_sse(struct kvm_vcpu *vcpu, in kvm_s390_get_base_disp_sse() 140 static inline void kvm_s390_get_regs_rre(struct kvm_vcpu *vcpu, int *r1, int *r2) in kvm_s390_get_regs_rre() [all …]
|
D | gaccess.h | 44 static inline unsigned long kvm_s390_real_to_abs(struct kvm_vcpu *vcpu, in kvm_s390_real_to_abs() 86 static inline unsigned long kvm_s390_logical_to_effective(struct kvm_vcpu *vcpu, in kvm_s390_logical_to_effective() 122 struct kvm_vcpu *__vcpu = (vcpu); \ 149 int write_guest_lc(struct kvm_vcpu *vcpu, unsigned long gra, void *data, in write_guest_lc() 175 int read_guest_lc(struct kvm_vcpu *vcpu, unsigned long gra, void *data, in read_guest_lc() 189 int guest_translate_address_with_key(struct kvm_vcpu *vcpu, unsigned long gva, u8 ar, 193 int check_gva_range(struct kvm_vcpu *vcpu, unsigned long gva, u8 ar, 202 int access_guest_with_key(struct kvm_vcpu *vcpu, unsigned long ga, u8 ar, 206 int access_guest_real(struct kvm_vcpu *vcpu, unsigned long gra, 256 int write_guest_with_key(struct kvm_vcpu *vcpu, unsigned long ga, u8 ar, in write_guest_with_key() [all …]
|
/linux-5.19.10/arch/arm64/include/asm/ |
D | kvm_emulate.h | 36 bool kvm_condition_valid32(const struct kvm_vcpu *vcpu); 37 void kvm_skip_instr32(struct kvm_vcpu *vcpu); 39 void kvm_inject_undefined(struct kvm_vcpu *vcpu); 40 void kvm_inject_vabt(struct kvm_vcpu *vcpu); 41 void kvm_inject_dabt(struct kvm_vcpu *vcpu, unsigned long addr); 42 void kvm_inject_pabt(struct kvm_vcpu *vcpu, unsigned long addr); 43 void kvm_inject_size_fault(struct kvm_vcpu *vcpu); 45 void kvm_vcpu_wfi(struct kvm_vcpu *vcpu); 48 static __always_inline bool vcpu_el1_is_32bit(struct kvm_vcpu *vcpu) in vcpu_el1_is_32bit() 53 static __always_inline bool vcpu_el1_is_32bit(struct kvm_vcpu *vcpu) in vcpu_el1_is_32bit() [all …]
|
D | kvm_host.h | 73 int kvm_reset_vcpu(struct kvm_vcpu *vcpu); 74 void kvm_arm_vcpu_destroy(struct kvm_vcpu *vcpu); 270 struct kvm_vcpu *__hyp_running_vcpu; 506 u64 vcpu_read_sys_reg(const struct kvm_vcpu *vcpu, int reg); 507 void vcpu_write_sys_reg(struct kvm_vcpu *vcpu, u64 val, int reg); 616 unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu); 617 int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *indices); 618 int kvm_arm_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg); 619 int kvm_arm_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg); 621 unsigned long kvm_arm_num_sys_reg_descs(struct kvm_vcpu *vcpu); [all …]
|
/linux-5.19.10/arch/powerpc/kernel/ |
D | asm-offsets.c | 154 OFFSET(THREAD_KVM_VCPU, thread_struct, kvm_vcpu); in main() 356 OFFSET(VCPU_HOST_STACK, kvm_vcpu, arch.host_stack); in main() 357 OFFSET(VCPU_HOST_PID, kvm_vcpu, arch.host_pid); in main() 358 OFFSET(VCPU_GUEST_PID, kvm_vcpu, arch.pid); in main() 359 OFFSET(VCPU_GPRS, kvm_vcpu, arch.regs.gpr); in main() 360 OFFSET(VCPU_VRSAVE, kvm_vcpu, arch.vrsave); in main() 361 OFFSET(VCPU_FPRS, kvm_vcpu, arch.fp.fpr); in main() 363 OFFSET(VCPU_VRS, kvm_vcpu, arch.vr.vr); in main() 365 OFFSET(VCPU_XER, kvm_vcpu, arch.regs.xer); in main() 366 OFFSET(VCPU_CTR, kvm_vcpu, arch.regs.ctr); in main() [all …]
|
/linux-5.19.10/arch/arm64/kvm/vgic/ |
D | vgic-mmio.h | 14 unsigned long (*read)(struct kvm_vcpu *vcpu, gpa_t addr, 20 void (*write)(struct kvm_vcpu *vcpu, gpa_t addr, 26 unsigned long (*uaccess_read)(struct kvm_vcpu *vcpu, gpa_t addr, 29 int (*uaccess_write)(struct kvm_vcpu *vcpu, gpa_t addr, 112 unsigned long vgic_mmio_read_raz(struct kvm_vcpu *vcpu, 115 unsigned long vgic_mmio_read_rao(struct kvm_vcpu *vcpu, 118 void vgic_mmio_write_wi(struct kvm_vcpu *vcpu, gpa_t addr, 121 int vgic_mmio_uaccess_write_wi(struct kvm_vcpu *vcpu, gpa_t addr, 124 unsigned long vgic_mmio_read_group(struct kvm_vcpu *vcpu, gpa_t addr, 127 void vgic_mmio_write_group(struct kvm_vcpu *vcpu, gpa_t addr, [all …]
|
D | vgic.h | 101 static inline u32 vgic_get_implementation_rev(struct kvm_vcpu *vcpu) in vgic_get_implementation_rev() 156 struct kvm_vcpu *vcpu; 165 vgic_get_mmio_region(struct kvm_vcpu *vcpu, struct vgic_io_device *iodev, 167 struct vgic_irq *vgic_get_irq(struct kvm *kvm, struct kvm_vcpu *vcpu, 184 void vgic_v2_fold_lr_state(struct kvm_vcpu *vcpu); 185 void vgic_v2_populate_lr(struct kvm_vcpu *vcpu, struct vgic_irq *irq, int lr); 186 void vgic_v2_clear_lr(struct kvm_vcpu *vcpu, int lr); 187 void vgic_v2_set_underflow(struct kvm_vcpu *vcpu); 188 void vgic_v2_set_npie(struct kvm_vcpu *vcpu); 190 int vgic_v2_dist_uaccess(struct kvm_vcpu *vcpu, bool is_write, [all …]
|
/linux-5.19.10/arch/powerpc/kvm/ |
D | booke.h | 65 void kvmppc_set_msr(struct kvm_vcpu *vcpu, u32 new_msr); 66 void kvmppc_mmu_msr_notify(struct kvm_vcpu *vcpu, u32 old_msr); 68 void kvmppc_set_epcr(struct kvm_vcpu *vcpu, u32 new_epcr); 69 void kvmppc_set_tcr(struct kvm_vcpu *vcpu, u32 new_tcr); 70 void kvmppc_set_tsr_bits(struct kvm_vcpu *vcpu, u32 tsr_bits); 71 void kvmppc_clr_tsr_bits(struct kvm_vcpu *vcpu, u32 tsr_bits); 73 int kvmppc_booke_emulate_op(struct kvm_vcpu *vcpu, 75 int kvmppc_booke_emulate_mfspr(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val); 76 int kvmppc_booke_emulate_mtspr(struct kvm_vcpu *vcpu, int sprn, ulong spr_val); 79 void kvmppc_load_guest_spe(struct kvm_vcpu *vcpu); [all …]
|
/linux-5.19.10/arch/x86/kvm/vmx/ |
D | vmx.h | 110 bool intel_pmu_lbr_is_compatible(struct kvm_vcpu *vcpu); 111 bool intel_pmu_lbr_is_enabled(struct kvm_vcpu *vcpu); 113 int intel_pmu_create_guest_lbr_event(struct kvm_vcpu *vcpu); 114 void vmx_passthrough_lbr_msrs(struct kvm_vcpu *vcpu); 261 struct kvm_vcpu vcpu; 386 bool nested_vmx_allowed(struct kvm_vcpu *vcpu); 387 void vmx_vcpu_load_vmcs(struct kvm_vcpu *vcpu, int cpu, 392 void vmx_prepare_switch_to_guest(struct kvm_vcpu *vcpu); 395 int vmx_get_cpl(struct kvm_vcpu *vcpu); 396 bool vmx_emulation_required(struct kvm_vcpu *vcpu); [all …]
|
D | nested.h | 19 void vmx_leave_nested(struct kvm_vcpu *vcpu); 22 __init int nested_vmx_hardware_setup(int (*exit_handlers[])(struct kvm_vcpu *)); 24 void nested_vmx_free_vcpu(struct kvm_vcpu *vcpu); 25 enum nvmx_vmentry_status nested_vmx_enter_non_root_mode(struct kvm_vcpu *vcpu, 27 bool nested_vmx_reflect_vmexit(struct kvm_vcpu *vcpu); 28 void nested_vmx_vmexit(struct kvm_vcpu *vcpu, u32 vm_exit_reason, 30 void nested_sync_vmcs12_to_shadow(struct kvm_vcpu *vcpu); 31 int vmx_set_vmx_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 data); 33 int get_vmx_mem_address(struct kvm_vcpu *vcpu, unsigned long exit_qualification, 35 void nested_vmx_pmu_refresh(struct kvm_vcpu *vcpu, [all …]
|
/linux-5.19.10/arch/mips/include/asm/ |
D | kvm_host.h | 294 int (*vcpu_run)(struct kvm_vcpu *vcpu); 719 int (*handle_cop_unusable)(struct kvm_vcpu *vcpu); 720 int (*handle_tlb_mod)(struct kvm_vcpu *vcpu); 721 int (*handle_tlb_ld_miss)(struct kvm_vcpu *vcpu); 722 int (*handle_tlb_st_miss)(struct kvm_vcpu *vcpu); 723 int (*handle_addr_err_st)(struct kvm_vcpu *vcpu); 724 int (*handle_addr_err_ld)(struct kvm_vcpu *vcpu); 725 int (*handle_syscall)(struct kvm_vcpu *vcpu); 726 int (*handle_res_inst)(struct kvm_vcpu *vcpu); 727 int (*handle_break)(struct kvm_vcpu *vcpu); [all …]
|
/linux-5.19.10/arch/riscv/include/asm/ |
D | kvm_host.h | 234 static inline void kvm_arch_sched_in(struct kvm_vcpu *vcpu, int cpu) {} in kvm_arch_sched_in() 259 void kvm_riscv_local_tlb_sanitize(struct kvm_vcpu *vcpu); 261 void kvm_riscv_fence_i_process(struct kvm_vcpu *vcpu); 262 void kvm_riscv_hfence_gvma_vmid_all_process(struct kvm_vcpu *vcpu); 263 void kvm_riscv_hfence_vvma_all_process(struct kvm_vcpu *vcpu); 264 void kvm_riscv_hfence_process(struct kvm_vcpu *vcpu); 288 int kvm_riscv_gstage_map(struct kvm_vcpu *vcpu, 293 void kvm_riscv_gstage_update_hgatp(struct kvm_vcpu *vcpu); 302 void kvm_riscv_gstage_vmid_update(struct kvm_vcpu *vcpu); 306 void kvm_riscv_vcpu_wfi(struct kvm_vcpu *vcpu); [all …]
|