Home
last modified time | relevance | path

Searched refs:kvm_vcpu (Results 1 – 25 of 244) sorted by relevance

12345678910

/linux-5.19.10/include/kvm/
Darm_pmu.h51 u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu, u64 select_idx);
52 void kvm_pmu_set_counter_value(struct kvm_vcpu *vcpu, u64 select_idx, u64 val);
53 u64 kvm_pmu_valid_counter_mask(struct kvm_vcpu *vcpu);
54 u64 kvm_pmu_get_pmceid(struct kvm_vcpu *vcpu, bool pmceid1);
55 void kvm_pmu_vcpu_init(struct kvm_vcpu *vcpu);
56 void kvm_pmu_vcpu_reset(struct kvm_vcpu *vcpu);
57 void kvm_pmu_vcpu_destroy(struct kvm_vcpu *vcpu);
58 void kvm_pmu_disable_counter_mask(struct kvm_vcpu *vcpu, u64 val);
59 void kvm_pmu_enable_counter_mask(struct kvm_vcpu *vcpu, u64 val);
60 void kvm_pmu_flush_hwstate(struct kvm_vcpu *vcpu);
[all …]
Darm_arch_timer.h27 struct kvm_vcpu *vcpu;
64 int kvm_timer_enable(struct kvm_vcpu *vcpu);
65 int kvm_timer_vcpu_reset(struct kvm_vcpu *vcpu);
66 void kvm_timer_vcpu_init(struct kvm_vcpu *vcpu);
67 void kvm_timer_sync_user(struct kvm_vcpu *vcpu);
68 bool kvm_timer_should_notify_user(struct kvm_vcpu *vcpu);
69 void kvm_timer_update_run(struct kvm_vcpu *vcpu);
70 void kvm_timer_vcpu_terminate(struct kvm_vcpu *vcpu);
72 u64 kvm_arm_timer_get_reg(struct kvm_vcpu *, u64 regid);
73 int kvm_arm_timer_set_reg(struct kvm_vcpu *, u64 regid, u64 value);
[all …]
/linux-5.19.10/arch/powerpc/include/asm/
Dkvm_ppc.h61 extern int kvmppc_vcpu_run(struct kvm_vcpu *vcpu);
62 extern int __kvmppc_vcpu_run(struct kvm_vcpu *vcpu);
65 extern void kvmppc_dump_vcpu(struct kvm_vcpu *vcpu);
66 extern int kvmppc_handle_load(struct kvm_vcpu *vcpu,
69 extern int kvmppc_handle_loads(struct kvm_vcpu *vcpu,
72 extern int kvmppc_handle_vsx_load(struct kvm_vcpu *vcpu,
75 extern int kvmppc_handle_vmx_load(struct kvm_vcpu *vcpu,
77 extern int kvmppc_handle_vmx_store(struct kvm_vcpu *vcpu,
79 extern int kvmppc_handle_store(struct kvm_vcpu *vcpu,
82 extern int kvmppc_handle_vsx_store(struct kvm_vcpu *vcpu,
[all …]
Dkvm_book3s.h78 struct kvm_vcpu *runnable_threads[MAX_SMT_THREADS];
85 struct kvm_vcpu *runner;
145 extern void kvmppc_mmu_pte_flush(struct kvm_vcpu *vcpu, ulong ea, ulong ea_mask);
146 extern void kvmppc_mmu_pte_vflush(struct kvm_vcpu *vcpu, u64 vp, u64 vp_mask);
147 extern void kvmppc_mmu_pte_pflush(struct kvm_vcpu *vcpu, ulong pa_start, ulong pa_end);
148 extern void kvmppc_set_msr(struct kvm_vcpu *vcpu, u64 new_msr);
149 extern void kvmppc_mmu_book3s_64_init(struct kvm_vcpu *vcpu);
150 extern void kvmppc_mmu_book3s_32_init(struct kvm_vcpu *vcpu);
151 extern void kvmppc_mmu_book3s_hv_init(struct kvm_vcpu *vcpu);
152 extern int kvmppc_mmu_map_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte,
[all …]
Dkvm_booke.h26 static inline void kvmppc_set_gpr(struct kvm_vcpu *vcpu, int num, ulong val) in kvmppc_set_gpr()
31 static inline ulong kvmppc_get_gpr(struct kvm_vcpu *vcpu, int num) in kvmppc_get_gpr()
36 static inline void kvmppc_set_cr(struct kvm_vcpu *vcpu, u32 val) in kvmppc_set_cr()
41 static inline u32 kvmppc_get_cr(struct kvm_vcpu *vcpu) in kvmppc_get_cr()
46 static inline void kvmppc_set_xer(struct kvm_vcpu *vcpu, ulong val) in kvmppc_set_xer()
51 static inline ulong kvmppc_get_xer(struct kvm_vcpu *vcpu) in kvmppc_get_xer()
56 static inline bool kvmppc_need_byteswap(struct kvm_vcpu *vcpu) in kvmppc_need_byteswap()
62 static inline void kvmppc_set_ctr(struct kvm_vcpu *vcpu, ulong val) in kvmppc_set_ctr()
67 static inline ulong kvmppc_get_ctr(struct kvm_vcpu *vcpu) in kvmppc_get_ctr()
72 static inline void kvmppc_set_lr(struct kvm_vcpu *vcpu, ulong val) in kvmppc_set_lr()
[all …]
/linux-5.19.10/arch/x86/kvm/
Dlapic.h50 struct kvm_vcpu *vcpu;
72 int kvm_create_lapic(struct kvm_vcpu *vcpu, int timer_advance_ns);
73 void kvm_free_lapic(struct kvm_vcpu *vcpu);
75 int kvm_apic_has_interrupt(struct kvm_vcpu *vcpu);
76 int kvm_apic_accept_pic_intr(struct kvm_vcpu *vcpu);
77 int kvm_get_apic_interrupt(struct kvm_vcpu *vcpu);
78 int kvm_apic_accept_events(struct kvm_vcpu *vcpu);
79 void kvm_lapic_reset(struct kvm_vcpu *vcpu, bool init_event);
80 u64 kvm_lapic_get_cr8(struct kvm_vcpu *vcpu);
81 void kvm_lapic_set_tpr(struct kvm_vcpu *vcpu, unsigned long cr8);
[all …]
Dx86.h61 void kvm_service_local_tlb_flush_requests(struct kvm_vcpu *vcpu);
62 int kvm_check_nested_events(struct kvm_vcpu *vcpu);
64 static inline void kvm_clear_exception_queue(struct kvm_vcpu *vcpu) in kvm_clear_exception_queue()
70 static inline void kvm_queue_interrupt(struct kvm_vcpu *vcpu, u8 vector, in kvm_queue_interrupt()
78 static inline void kvm_clear_interrupt_queue(struct kvm_vcpu *vcpu) in kvm_clear_interrupt_queue()
83 static inline bool kvm_event_needs_reinjection(struct kvm_vcpu *vcpu) in kvm_event_needs_reinjection()
94 static inline bool is_protmode(struct kvm_vcpu *vcpu) in is_protmode()
99 static inline int is_long_mode(struct kvm_vcpu *vcpu) in is_long_mode()
108 static inline bool is_64_bit_mode(struct kvm_vcpu *vcpu) in is_64_bit_mode()
120 static inline bool is_64_bit_hypercall(struct kvm_vcpu *vcpu) in is_64_bit_hypercall()
[all …]
Dxen.h17 int __kvm_xen_has_interrupt(struct kvm_vcpu *vcpu);
18 void kvm_xen_inject_pending_events(struct kvm_vcpu *vcpu);
19 int kvm_xen_vcpu_set_attr(struct kvm_vcpu *vcpu, struct kvm_xen_vcpu_attr *data);
20 int kvm_xen_vcpu_get_attr(struct kvm_vcpu *vcpu, struct kvm_xen_vcpu_attr *data);
24 int kvm_xen_write_hypercall_page(struct kvm_vcpu *vcpu, u64 data);
28 void kvm_xen_init_vcpu(struct kvm_vcpu *vcpu);
29 void kvm_xen_destroy_vcpu(struct kvm_vcpu *vcpu);
49 static inline int kvm_xen_has_interrupt(struct kvm_vcpu *vcpu) in kvm_xen_has_interrupt()
59 static inline bool kvm_xen_has_pending_events(struct kvm_vcpu *vcpu) in kvm_xen_has_pending_events()
65 static inline bool kvm_xen_timer_enabled(struct kvm_vcpu *vcpu) in kvm_xen_timer_enabled()
[all …]
Dhyperv.h58 static inline struct kvm_vcpu_hv *to_hv_vcpu(struct kvm_vcpu *vcpu) in to_hv_vcpu()
63 static inline struct kvm_vcpu_hv_synic *to_hv_synic(struct kvm_vcpu *vcpu) in to_hv_synic()
70 static inline struct kvm_vcpu *hv_synic_to_vcpu(struct kvm_vcpu_hv_synic *synic) in hv_synic_to_vcpu()
77 static inline struct kvm_hv_syndbg *to_hv_syndbg(struct kvm_vcpu *vcpu) in to_hv_syndbg()
82 static inline u32 kvm_hv_get_vpindex(struct kvm_vcpu *vcpu) in kvm_hv_get_vpindex()
89 int kvm_hv_set_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 data, bool host);
90 int kvm_hv_get_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata, bool host);
92 static inline bool kvm_hv_hypercall_enabled(struct kvm_vcpu *vcpu) in kvm_hv_hypercall_enabled()
97 int kvm_hv_hypercall(struct kvm_vcpu *vcpu);
101 void kvm_hv_synic_send_eoi(struct kvm_vcpu *vcpu, int vector);
[all …]
Dcpuid.h14 void kvm_update_cpuid_runtime(struct kvm_vcpu *vcpu);
15 void kvm_update_pv_runtime(struct kvm_vcpu *vcpu);
16 struct kvm_cpuid_entry2 *kvm_find_cpuid_entry(struct kvm_vcpu *vcpu,
21 int kvm_vcpu_ioctl_set_cpuid(struct kvm_vcpu *vcpu,
24 int kvm_vcpu_ioctl_set_cpuid2(struct kvm_vcpu *vcpu,
27 int kvm_vcpu_ioctl_get_cpuid2(struct kvm_vcpu *vcpu,
30 bool kvm_cpuid(struct kvm_vcpu *vcpu, u32 *eax, u32 *ebx,
35 int cpuid_query_maxphyaddr(struct kvm_vcpu *vcpu);
36 u64 kvm_vcpu_reserved_gpa_bits_raw(struct kvm_vcpu *vcpu);
38 static inline int cpuid_maxphyaddr(struct kvm_vcpu *vcpu) in cpuid_maxphyaddr()
[all …]
Dpmu.h8 #define pmu_to_vcpu(pmu) (container_of((pmu), struct kvm_vcpu, arch.pmu))
28 struct kvm_pmc *(*rdpmc_ecx_to_pmc)(struct kvm_vcpu *vcpu,
30 struct kvm_pmc *(*msr_idx_to_pmc)(struct kvm_vcpu *vcpu, u32 msr);
31 bool (*is_valid_rdpmc_ecx)(struct kvm_vcpu *vcpu, unsigned int idx);
32 bool (*is_valid_msr)(struct kvm_vcpu *vcpu, u32 msr);
33 int (*get_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr_info);
34 int (*set_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr_info);
35 void (*refresh)(struct kvm_vcpu *vcpu);
36 void (*init)(struct kvm_vcpu *vcpu);
37 void (*reset)(struct kvm_vcpu *vcpu);
[all …]
Dkvm_cache_regs.h19 static __always_inline unsigned long kvm_##lname##_read(struct kvm_vcpu *vcpu)\
23 static __always_inline void kvm_##lname##_write(struct kvm_vcpu *vcpu, \
53 static inline bool kvm_register_is_available(struct kvm_vcpu *vcpu, in BUILD_KVM_GPR_ACCESSORS()
59 static inline bool kvm_register_is_dirty(struct kvm_vcpu *vcpu, in kvm_register_is_dirty()
65 static inline void kvm_register_mark_available(struct kvm_vcpu *vcpu, in kvm_register_mark_available()
71 static inline void kvm_register_mark_dirty(struct kvm_vcpu *vcpu, in kvm_register_mark_dirty()
83 static inline unsigned long kvm_register_read_raw(struct kvm_vcpu *vcpu, int reg) in kvm_register_read_raw()
94 static inline void kvm_register_write_raw(struct kvm_vcpu *vcpu, int reg, in kvm_register_write_raw()
104 static inline unsigned long kvm_rip_read(struct kvm_vcpu *vcpu) in kvm_rip_read()
109 static inline void kvm_rip_write(struct kvm_vcpu *vcpu, unsigned long val) in kvm_rip_write()
[all …]
/linux-5.19.10/arch/x86/include/asm/
Dkvm_host.h428 unsigned long (*get_guest_pgd)(struct kvm_vcpu *vcpu);
429 u64 (*get_pdptr)(struct kvm_vcpu *vcpu, int index);
430 int (*page_fault)(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault);
431 void (*inject_page_fault)(struct kvm_vcpu *vcpu,
433 gpa_t (*gva_to_gpa)(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu,
436 int (*sync_page)(struct kvm_vcpu *vcpu,
438 void (*invlpg)(struct kvm_vcpu *vcpu, gva_t gva, hpa_t root_hpa);
492 struct kvm_vcpu *vcpu;
585 struct kvm_vcpu *vcpu;
755 int (*complete_userspace_io)(struct kvm_vcpu *vcpu);
[all …]
/linux-5.19.10/arch/s390/kvm/
Dkvm-s390.h60 static inline void kvm_s390_set_cpuflags(struct kvm_vcpu *vcpu, u32 flags) in kvm_s390_set_cpuflags()
65 static inline void kvm_s390_clear_cpuflags(struct kvm_vcpu *vcpu, u32 flags) in kvm_s390_clear_cpuflags()
70 static inline bool kvm_s390_test_cpuflags(struct kvm_vcpu *vcpu, u32 flags) in kvm_s390_test_cpuflags()
75 static inline int is_vcpu_stopped(struct kvm_vcpu *vcpu) in is_vcpu_stopped()
80 static inline int is_vcpu_idle(struct kvm_vcpu *vcpu) in is_vcpu_idle()
97 static inline u32 kvm_s390_get_prefix(struct kvm_vcpu *vcpu) in kvm_s390_get_prefix()
102 static inline void kvm_s390_set_prefix(struct kvm_vcpu *vcpu, u32 prefix) in kvm_s390_set_prefix()
111 static inline u64 kvm_s390_get_base_disp_s(struct kvm_vcpu *vcpu, u8 *ar) in kvm_s390_get_base_disp_s()
122 static inline void kvm_s390_get_base_disp_sse(struct kvm_vcpu *vcpu, in kvm_s390_get_base_disp_sse()
140 static inline void kvm_s390_get_regs_rre(struct kvm_vcpu *vcpu, int *r1, int *r2) in kvm_s390_get_regs_rre()
[all …]
Dgaccess.h44 static inline unsigned long kvm_s390_real_to_abs(struct kvm_vcpu *vcpu, in kvm_s390_real_to_abs()
86 static inline unsigned long kvm_s390_logical_to_effective(struct kvm_vcpu *vcpu, in kvm_s390_logical_to_effective()
122 struct kvm_vcpu *__vcpu = (vcpu); \
149 int write_guest_lc(struct kvm_vcpu *vcpu, unsigned long gra, void *data, in write_guest_lc()
175 int read_guest_lc(struct kvm_vcpu *vcpu, unsigned long gra, void *data, in read_guest_lc()
189 int guest_translate_address_with_key(struct kvm_vcpu *vcpu, unsigned long gva, u8 ar,
193 int check_gva_range(struct kvm_vcpu *vcpu, unsigned long gva, u8 ar,
202 int access_guest_with_key(struct kvm_vcpu *vcpu, unsigned long ga, u8 ar,
206 int access_guest_real(struct kvm_vcpu *vcpu, unsigned long gra,
256 int write_guest_with_key(struct kvm_vcpu *vcpu, unsigned long ga, u8 ar, in write_guest_with_key()
[all …]
/linux-5.19.10/arch/arm64/include/asm/
Dkvm_emulate.h36 bool kvm_condition_valid32(const struct kvm_vcpu *vcpu);
37 void kvm_skip_instr32(struct kvm_vcpu *vcpu);
39 void kvm_inject_undefined(struct kvm_vcpu *vcpu);
40 void kvm_inject_vabt(struct kvm_vcpu *vcpu);
41 void kvm_inject_dabt(struct kvm_vcpu *vcpu, unsigned long addr);
42 void kvm_inject_pabt(struct kvm_vcpu *vcpu, unsigned long addr);
43 void kvm_inject_size_fault(struct kvm_vcpu *vcpu);
45 void kvm_vcpu_wfi(struct kvm_vcpu *vcpu);
48 static __always_inline bool vcpu_el1_is_32bit(struct kvm_vcpu *vcpu) in vcpu_el1_is_32bit()
53 static __always_inline bool vcpu_el1_is_32bit(struct kvm_vcpu *vcpu) in vcpu_el1_is_32bit()
[all …]
Dkvm_host.h73 int kvm_reset_vcpu(struct kvm_vcpu *vcpu);
74 void kvm_arm_vcpu_destroy(struct kvm_vcpu *vcpu);
270 struct kvm_vcpu *__hyp_running_vcpu;
506 u64 vcpu_read_sys_reg(const struct kvm_vcpu *vcpu, int reg);
507 void vcpu_write_sys_reg(struct kvm_vcpu *vcpu, u64 val, int reg);
616 unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu);
617 int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *indices);
618 int kvm_arm_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg);
619 int kvm_arm_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg);
621 unsigned long kvm_arm_num_sys_reg_descs(struct kvm_vcpu *vcpu);
[all …]
/linux-5.19.10/arch/powerpc/kernel/
Dasm-offsets.c154 OFFSET(THREAD_KVM_VCPU, thread_struct, kvm_vcpu); in main()
356 OFFSET(VCPU_HOST_STACK, kvm_vcpu, arch.host_stack); in main()
357 OFFSET(VCPU_HOST_PID, kvm_vcpu, arch.host_pid); in main()
358 OFFSET(VCPU_GUEST_PID, kvm_vcpu, arch.pid); in main()
359 OFFSET(VCPU_GPRS, kvm_vcpu, arch.regs.gpr); in main()
360 OFFSET(VCPU_VRSAVE, kvm_vcpu, arch.vrsave); in main()
361 OFFSET(VCPU_FPRS, kvm_vcpu, arch.fp.fpr); in main()
363 OFFSET(VCPU_VRS, kvm_vcpu, arch.vr.vr); in main()
365 OFFSET(VCPU_XER, kvm_vcpu, arch.regs.xer); in main()
366 OFFSET(VCPU_CTR, kvm_vcpu, arch.regs.ctr); in main()
[all …]
/linux-5.19.10/arch/arm64/kvm/vgic/
Dvgic-mmio.h14 unsigned long (*read)(struct kvm_vcpu *vcpu, gpa_t addr,
20 void (*write)(struct kvm_vcpu *vcpu, gpa_t addr,
26 unsigned long (*uaccess_read)(struct kvm_vcpu *vcpu, gpa_t addr,
29 int (*uaccess_write)(struct kvm_vcpu *vcpu, gpa_t addr,
112 unsigned long vgic_mmio_read_raz(struct kvm_vcpu *vcpu,
115 unsigned long vgic_mmio_read_rao(struct kvm_vcpu *vcpu,
118 void vgic_mmio_write_wi(struct kvm_vcpu *vcpu, gpa_t addr,
121 int vgic_mmio_uaccess_write_wi(struct kvm_vcpu *vcpu, gpa_t addr,
124 unsigned long vgic_mmio_read_group(struct kvm_vcpu *vcpu, gpa_t addr,
127 void vgic_mmio_write_group(struct kvm_vcpu *vcpu, gpa_t addr,
[all …]
Dvgic.h101 static inline u32 vgic_get_implementation_rev(struct kvm_vcpu *vcpu) in vgic_get_implementation_rev()
156 struct kvm_vcpu *vcpu;
165 vgic_get_mmio_region(struct kvm_vcpu *vcpu, struct vgic_io_device *iodev,
167 struct vgic_irq *vgic_get_irq(struct kvm *kvm, struct kvm_vcpu *vcpu,
184 void vgic_v2_fold_lr_state(struct kvm_vcpu *vcpu);
185 void vgic_v2_populate_lr(struct kvm_vcpu *vcpu, struct vgic_irq *irq, int lr);
186 void vgic_v2_clear_lr(struct kvm_vcpu *vcpu, int lr);
187 void vgic_v2_set_underflow(struct kvm_vcpu *vcpu);
188 void vgic_v2_set_npie(struct kvm_vcpu *vcpu);
190 int vgic_v2_dist_uaccess(struct kvm_vcpu *vcpu, bool is_write,
[all …]
/linux-5.19.10/arch/powerpc/kvm/
Dbooke.h65 void kvmppc_set_msr(struct kvm_vcpu *vcpu, u32 new_msr);
66 void kvmppc_mmu_msr_notify(struct kvm_vcpu *vcpu, u32 old_msr);
68 void kvmppc_set_epcr(struct kvm_vcpu *vcpu, u32 new_epcr);
69 void kvmppc_set_tcr(struct kvm_vcpu *vcpu, u32 new_tcr);
70 void kvmppc_set_tsr_bits(struct kvm_vcpu *vcpu, u32 tsr_bits);
71 void kvmppc_clr_tsr_bits(struct kvm_vcpu *vcpu, u32 tsr_bits);
73 int kvmppc_booke_emulate_op(struct kvm_vcpu *vcpu,
75 int kvmppc_booke_emulate_mfspr(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val);
76 int kvmppc_booke_emulate_mtspr(struct kvm_vcpu *vcpu, int sprn, ulong spr_val);
79 void kvmppc_load_guest_spe(struct kvm_vcpu *vcpu);
[all …]
/linux-5.19.10/arch/x86/kvm/vmx/
Dvmx.h110 bool intel_pmu_lbr_is_compatible(struct kvm_vcpu *vcpu);
111 bool intel_pmu_lbr_is_enabled(struct kvm_vcpu *vcpu);
113 int intel_pmu_create_guest_lbr_event(struct kvm_vcpu *vcpu);
114 void vmx_passthrough_lbr_msrs(struct kvm_vcpu *vcpu);
261 struct kvm_vcpu vcpu;
386 bool nested_vmx_allowed(struct kvm_vcpu *vcpu);
387 void vmx_vcpu_load_vmcs(struct kvm_vcpu *vcpu, int cpu,
392 void vmx_prepare_switch_to_guest(struct kvm_vcpu *vcpu);
395 int vmx_get_cpl(struct kvm_vcpu *vcpu);
396 bool vmx_emulation_required(struct kvm_vcpu *vcpu);
[all …]
Dnested.h19 void vmx_leave_nested(struct kvm_vcpu *vcpu);
22 __init int nested_vmx_hardware_setup(int (*exit_handlers[])(struct kvm_vcpu *));
24 void nested_vmx_free_vcpu(struct kvm_vcpu *vcpu);
25 enum nvmx_vmentry_status nested_vmx_enter_non_root_mode(struct kvm_vcpu *vcpu,
27 bool nested_vmx_reflect_vmexit(struct kvm_vcpu *vcpu);
28 void nested_vmx_vmexit(struct kvm_vcpu *vcpu, u32 vm_exit_reason,
30 void nested_sync_vmcs12_to_shadow(struct kvm_vcpu *vcpu);
31 int vmx_set_vmx_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 data);
33 int get_vmx_mem_address(struct kvm_vcpu *vcpu, unsigned long exit_qualification,
35 void nested_vmx_pmu_refresh(struct kvm_vcpu *vcpu,
[all …]
/linux-5.19.10/arch/mips/include/asm/
Dkvm_host.h294 int (*vcpu_run)(struct kvm_vcpu *vcpu);
719 int (*handle_cop_unusable)(struct kvm_vcpu *vcpu);
720 int (*handle_tlb_mod)(struct kvm_vcpu *vcpu);
721 int (*handle_tlb_ld_miss)(struct kvm_vcpu *vcpu);
722 int (*handle_tlb_st_miss)(struct kvm_vcpu *vcpu);
723 int (*handle_addr_err_st)(struct kvm_vcpu *vcpu);
724 int (*handle_addr_err_ld)(struct kvm_vcpu *vcpu);
725 int (*handle_syscall)(struct kvm_vcpu *vcpu);
726 int (*handle_res_inst)(struct kvm_vcpu *vcpu);
727 int (*handle_break)(struct kvm_vcpu *vcpu);
[all …]
/linux-5.19.10/arch/riscv/include/asm/
Dkvm_host.h234 static inline void kvm_arch_sched_in(struct kvm_vcpu *vcpu, int cpu) {} in kvm_arch_sched_in()
259 void kvm_riscv_local_tlb_sanitize(struct kvm_vcpu *vcpu);
261 void kvm_riscv_fence_i_process(struct kvm_vcpu *vcpu);
262 void kvm_riscv_hfence_gvma_vmid_all_process(struct kvm_vcpu *vcpu);
263 void kvm_riscv_hfence_vvma_all_process(struct kvm_vcpu *vcpu);
264 void kvm_riscv_hfence_process(struct kvm_vcpu *vcpu);
288 int kvm_riscv_gstage_map(struct kvm_vcpu *vcpu,
293 void kvm_riscv_gstage_update_hgatp(struct kvm_vcpu *vcpu);
302 void kvm_riscv_gstage_vmid_update(struct kvm_vcpu *vcpu);
306 void kvm_riscv_vcpu_wfi(struct kvm_vcpu *vcpu);
[all …]

12345678910