/linux-6.1.9/tools/testing/selftests/kvm/ |
D | access_tracking_perf_test.c | 131 int vcpu_idx = vcpu_args->vcpu_idx; in mark_vcpu_memory_idle() local 141 if (overlap_memory_access && vcpu_idx) in mark_vcpu_memory_idle() 173 vcpu_idx, no_pfn, pages); in mark_vcpu_memory_idle() 192 vcpu_idx, still_idle, pages); in mark_vcpu_memory_idle() 227 int vcpu_idx = vcpu_args->vcpu_idx; in vcpu_thread_main() local 241 vcpu_last_completed_iteration[vcpu_idx] = current_iteration; in vcpu_thread_main() 245 static void spin_wait_for_vcpu(int vcpu_idx, int target_iteration) in spin_wait_for_vcpu() argument 247 while (READ_ONCE(vcpu_last_completed_iteration[vcpu_idx]) != in spin_wait_for_vcpu()
|
D | dirty_log_perf_test.c | 73 int vcpu_idx = vcpu_args->vcpu_idx; in vcpu_worker() local 96 pr_debug("Got sync event from vCPU %d\n", vcpu_idx); in vcpu_worker() 97 vcpu_last_completed_iteration[vcpu_idx] = current_iteration; in vcpu_worker() 99 vcpu_idx, vcpu_last_completed_iteration[vcpu_idx]); in vcpu_worker() 105 vcpu_idx, current_iteration, ts_diff.tv_sec, in vcpu_worker() 109 vcpu_idx, current_iteration, ts_diff.tv_sec, in vcpu_worker() 122 avg = timespec_div(total, vcpu_last_completed_iteration[vcpu_idx]); in vcpu_worker() 124 vcpu_idx, pages_count, vcpu_last_completed_iteration[vcpu_idx], in vcpu_worker()
|
D | steal_time.c | 81 static void steal_time_dump(struct kvm_vm *vm, uint32_t vcpu_idx) in steal_time_dump() argument 83 struct kvm_steal_time *st = addr_gva2hva(vm, (ulong)st_gva[vcpu_idx]); in steal_time_dump() 86 pr_info("VCPU%d:\n", vcpu_idx); in steal_time_dump() 196 static void steal_time_dump(struct kvm_vm *vm, uint32_t vcpu_idx) in steal_time_dump() argument 198 struct st_time *st = addr_gva2hva(vm, (ulong)st_gva[vcpu_idx]); in steal_time_dump() 200 pr_info("VCPU%d:\n", vcpu_idx); in steal_time_dump()
|
D | demand_paging_test.c | 48 int vcpu_idx = vcpu_args->vcpu_idx; in vcpu_worker() local 66 PER_VCPU_DEBUG("vCPU %d execution time: %ld.%.9lds\n", vcpu_idx, in vcpu_worker()
|
/linux-6.1.9/tools/testing/selftests/kvm/lib/ |
D | perf_test_util.c | 21 int vcpu_idx; member 45 void perf_test_guest_code(uint32_t vcpu_idx) in perf_test_guest_code() argument 48 struct perf_test_vcpu_args *vcpu_args = &pta->vcpu_args[vcpu_idx]; in perf_test_guest_code() 57 GUEST_ASSERT(vcpu_args->vcpu_idx == vcpu_idx); in perf_test_guest_code() 86 vcpu_args->vcpu_idx = i; in perf_test_setup_vcpus() 258 vcpu_thread_fn(&perf_test_args.vcpu_args[vcpu->vcpu_idx]); in vcpu_thread_main() 274 vcpu->vcpu_idx = i; in perf_test_start_vcpu_threads()
|
/linux-6.1.9/tools/testing/selftests/kvm/aarch64/ |
D | arch_timer.c | 215 unsigned int vcpu_idx = (unsigned long)arg; in test_vcpu_run() local 217 struct kvm_vcpu *vcpu = vcpus[vcpu_idx]; in test_vcpu_run() 219 struct test_vcpu_shared_data *shared_data = &vcpu_shared_data[vcpu_idx]; in test_vcpu_run() 225 set_bit(vcpu_idx, vcpu_done_map); in test_vcpu_run() 238 vcpu_idx, in test_vcpu_run() 266 static int test_migrate_vcpu(unsigned int vcpu_idx) in test_migrate_vcpu() argument 275 pr_debug("Migrating vCPU: %u to pCPU: %u\n", vcpu_idx, new_pcpu); in test_migrate_vcpu() 277 ret = pthread_setaffinity_np(pt_vcpu_run[vcpu_idx], in test_migrate_vcpu() 283 vcpu_idx, new_pcpu, ret); in test_migrate_vcpu()
|
/linux-6.1.9/arch/arm64/kvm/ |
D | trace_arm.h | 93 TP_PROTO(unsigned int type, int vcpu_idx, int irq_num, int level), 94 TP_ARGS(type, vcpu_idx, irq_num, level), 98 __field( int, vcpu_idx ) 105 __entry->vcpu_idx = vcpu_idx; 114 __entry->type, __entry->vcpu_idx, __entry->irq_num, __entry->level)
|
D | arm.c | 1068 unsigned int irq_type, vcpu_idx, irq_num; in kvm_vm_ioctl_irq_line() local 1074 vcpu_idx = (irq >> KVM_ARM_IRQ_VCPU_SHIFT) & KVM_ARM_IRQ_VCPU_MASK; in kvm_vm_ioctl_irq_line() 1075 …vcpu_idx += ((irq >> KVM_ARM_IRQ_VCPU2_SHIFT) & KVM_ARM_IRQ_VCPU2_MASK) * (KVM_ARM_IRQ_VCPU_MASK +… in kvm_vm_ioctl_irq_line() 1078 trace_kvm_irq_line(irq_type, vcpu_idx, irq_num, irq_level->level); in kvm_vm_ioctl_irq_line() 1085 if (vcpu_idx >= nrcpus) in kvm_vm_ioctl_irq_line() 1088 vcpu = kvm_get_vcpu(kvm, vcpu_idx); in kvm_vm_ioctl_irq_line() 1100 if (vcpu_idx >= nrcpus) in kvm_vm_ioctl_irq_line() 1103 vcpu = kvm_get_vcpu(kvm, vcpu_idx); in kvm_vm_ioctl_irq_line()
|
/linux-6.1.9/arch/x86/kvm/ |
D | xen.c | 119 e.vcpu_idx = vcpu->vcpu_idx; in kvm_xen_inject_timer_irqs() 1064 set_bit(vcpu->vcpu_idx, vcpu->kvm->arch.xen.poll_mask); in kvm_xen_schedop_poll() 1085 clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.xen.poll_mask); in kvm_xen_schedop_poll() 1327 test_and_clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.xen.poll_mask)) { in kvm_xen_check_poller() 1351 int vcpu_idx, idx, rc; in kvm_xen_set_evtchn_fast() local 1353 vcpu_idx = READ_ONCE(xe->vcpu_idx); in kvm_xen_set_evtchn_fast() 1354 if (vcpu_idx >= 0) in kvm_xen_set_evtchn_fast() 1355 vcpu = kvm_get_vcpu(kvm, vcpu_idx); in kvm_xen_set_evtchn_fast() 1360 WRITE_ONCE(xe->vcpu_idx, vcpu->vcpu_idx); in kvm_xen_set_evtchn_fast() 1556 e->xen_evtchn.vcpu_idx = vcpu->vcpu_idx; in kvm_xen_setup_evtchn() [all …]
|
D | hyperv.h | 89 return hv_vcpu ? hv_vcpu->vp_index : vcpu->vcpu_idx; in kvm_hv_get_vpindex()
|
D | lapic.c | 1234 int vcpu_idx; in kvm_bitmap_or_dest_vcpus() local 1246 vcpu_idx = dest_vcpu[i]->vcpu->vcpu_idx; in kvm_bitmap_or_dest_vcpus() 1247 __set_bit(vcpu_idx, vcpu_bitmap); in kvm_bitmap_or_dest_vcpus()
|
D | hyperv.c | 955 hv_vcpu->vp_index = vcpu->vcpu_idx; in kvm_hv_vcpu_init() 1448 if (hv_vcpu->vp_index == vcpu->vcpu_idx) in kvm_hv_set_msr() 1450 else if (new_vp_index == vcpu->vcpu_idx) in kvm_hv_set_msr()
|
D | x86.c | 11959 if (kvmclock_periodic_sync && vcpu->vcpu_idx == 0) in kvm_arch_vcpu_postcreate()
|
/linux-6.1.9/tools/testing/selftests/kvm/include/ |
D | perf_test_util.h | 29 int vcpu_idx; member
|
/linux-6.1.9/virt/kvm/ |
D | pfncache.c | 50 __set_bit(gpc->vcpu->vcpu_idx, vcpu_bitmap); in gfn_to_pfn_cache_invalidate_start()
|
D | kvm_main.c | 3949 vcpu->vcpu_idx = atomic_read(&kvm->online_vcpus); in kvm_vm_ioctl_create_vcpu() 3950 r = xa_insert(&kvm->vcpu_array, vcpu->vcpu_idx, vcpu, GFP_KERNEL_ACCOUNT); in kvm_vm_ioctl_create_vcpu() 3959 xa_erase(&kvm->vcpu_array, vcpu->vcpu_idx); in kvm_vm_ioctl_create_vcpu()
|
/linux-6.1.9/arch/s390/kvm/ |
D | interrupt.c | 428 set_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask); in __set_cpu_idle() 434 clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask); in __unset_cpu_idle() 3066 int vcpu_idx, online_vcpus = atomic_read(&kvm->online_vcpus); in __airqs_kick_single_vcpu() local 3071 for_each_set_bit(vcpu_idx, kvm->arch.idle_mask, online_vcpus) { in __airqs_kick_single_vcpu() 3072 vcpu = kvm_get_vcpu(kvm, vcpu_idx); in __airqs_kick_single_vcpu() 3078 if (test_and_set_bit(vcpu_idx, gi->kicked_mask)) in __airqs_kick_single_vcpu()
|
D | kvm-s390.h | 82 return test_bit(vcpu->vcpu_idx, vcpu->kvm->arch.idle_mask); in is_vcpu_idle()
|
D | kvm-s390.c | 3397 unsigned long vcpu_idx; in sca_switch_to_extended() local 3415 kvm_for_each_vcpu(vcpu_idx, vcpu, kvm) { in sca_switch_to_extended() 3819 clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.gisa_int.kicked_mask); in kvm_arch_vcpu_runnable() 4527 clear_bit(vcpu->vcpu_idx, vcpu->kvm->arch.gisa_int.kicked_mask); in vcpu_pre_run()
|
/linux-6.1.9/include/linux/ |
D | kvm_host.h | 316 int vcpu_idx; /* index in kvm->vcpus array */ member 617 int vcpu_idx; member
|
/linux-6.1.9/arch/riscv/kvm/ |
D | vcpu.c | 204 if (vcpu->vcpu_idx != 0) in kvm_arch_vcpu_postcreate()
|