/linux-5.19.10/tools/virtio/ringtest/ |
D | virtio_ring_0_9.c | 41 struct guest { struct 52 } guest; argument 78 guest.avail_idx = 0; in alloc_ring() 79 guest.kicked_avail_idx = -1; in alloc_ring() 80 guest.last_used_idx = 0; in alloc_ring() 83 guest.free_head = 0; in alloc_ring() 89 guest.num_free = ring_size; in alloc_ring() 107 if (!guest.num_free) in add_inbuf() 111 head = (ring_size - 1) & (guest.avail_idx++); in add_inbuf() 113 head = guest.free_head; in add_inbuf() [all …]
|
D | ring.c | 59 struct guest { struct 65 } guest; argument 92 guest.avail_idx = 0; in alloc_ring() 93 guest.kicked_avail_idx = -1; in alloc_ring() 94 guest.last_used_idx = 0; in alloc_ring() 103 guest.num_free = ring_size; in alloc_ring() 116 if (!guest.num_free) in add_inbuf() 119 guest.num_free--; in add_inbuf() 120 head = (ring_size - 1) & (guest.avail_idx++); in add_inbuf() 145 unsigned head = (ring_size - 1) & guest.last_used_idx; in get_buf() [all …]
|
/linux-5.19.10/drivers/misc/cxl/ |
D | of.c | 88 afu->guest->handle = addr; in read_phys_addr() 91 afu->guest->p2n_phys += addr; in read_phys_addr() 92 afu->guest->p2n_size = size; in read_phys_addr() 133 if (read_handle(afu_np, &afu->guest->handle)) in cxl_of_read_afu_handle() 135 pr_devel("AFU handle: 0x%.16llx\n", afu->guest->handle); in cxl_of_read_afu_handle() 190 read_prop_dword(np, "ibm,max-ints-per-process", &afu->guest->max_ints); in cxl_of_read_afu_properties() 191 afu->irqs_max = afu->guest->max_ints; in cxl_of_read_afu_properties() 269 pr_devel("AFU handle: %#llx\n", afu->guest->handle); in cxl_of_read_afu_properties() 271 afu->guest->p2n_phys, afu->guest->p2n_size); in cxl_of_read_afu_properties() 301 adapter->guest->irq_avail = kcalloc(nranges, sizeof(struct irq_avail), in read_adapter_irq_config() [all …]
|
D | guest.c | 117 rc = cxl_h_collect_vpd_adapter(adapter->guest->handle, in guest_collect_vpd() 120 rc = cxl_h_collect_vpd(afu->guest->handle, 0, in guest_collect_vpd() 158 return cxl_h_collect_int_info(ctx->afu->guest->handle, ctx->process_token, info); in guest_get_irq_info() 186 rc = cxl_h_read_error_state(afu->guest->handle, &state); in afu_read_error_state() 203 rc = cxl_h_get_fn_error_interrupt(afu->guest->handle, &serr); in guest_slice_irq_err() 214 rc = cxl_h_ack_fn_error_interrupt(afu->guest->handle, serr); in guest_slice_irq_err() 228 for (i = 0; i < adapter->guest->irq_nranges; i++) { in irq_alloc_range() 229 cur = &adapter->guest->irq_avail[i]; in irq_alloc_range() 252 for (i = 0; i < adapter->guest->irq_nranges; i++) { in irq_free_range() 253 cur = &adapter->guest->irq_avail[i]; in irq_free_range() [all …]
|
D | flash.c | 191 unwa->unit_address = cpu_to_be64(adapter->guest->handle); in update_devicetree() 245 header->vendor = cpu_to_be16(adapter->guest->vendor); in handle_image() 246 header->device = cpu_to_be16(adapter->guest->device); in handle_image() 247 header->subsystem_vendor = cpu_to_be16(adapter->guest->subsystem_vendor); in handle_image() 248 header->subsystem = cpu_to_be16(adapter->guest->subsystem); in handle_image() 312 rc = fct(adapter->guest->handle, virt_to_phys(le), entries, in handle_image() 335 cxl_h_reset_adapter(adapter->guest->handle); in transfer_image() 344 cxl_h_reset_adapter(adapter->guest->handle); in transfer_image() 353 cxl_h_reset_adapter(adapter->guest->handle); in transfer_image() 501 cxl_h_reset_adapter(adapter->guest->handle); in device_close() [all …]
|
/linux-5.19.10/arch/mips/include/asm/ |
D | cpu-features.h | 649 #define cpu_guest_has_conf1 (cpu_data[0].guest.conf & (1 << 1)) 652 #define cpu_guest_has_conf2 (cpu_data[0].guest.conf & (1 << 2)) 655 #define cpu_guest_has_conf3 (cpu_data[0].guest.conf & (1 << 3)) 658 #define cpu_guest_has_conf4 (cpu_data[0].guest.conf & (1 << 4)) 661 #define cpu_guest_has_conf5 (cpu_data[0].guest.conf & (1 << 5)) 664 #define cpu_guest_has_conf6 (cpu_data[0].guest.conf & (1 << 6)) 667 #define cpu_guest_has_conf7 (cpu_data[0].guest.conf & (1 << 7)) 670 #define cpu_guest_has_fpu (cpu_data[0].guest.options & MIPS_CPU_FPU) 673 #define cpu_guest_has_watch (cpu_data[0].guest.options & MIPS_CPU_WATCH) 676 #define cpu_guest_has_contextconfig (cpu_data[0].guest.options & MIPS_CPU_CTXTC) [all …]
|
/linux-5.19.10/Documentation/virt/kvm/x86/ |
D | running-nested-guests.rst | 7 A nested guest is the ability to run a guest inside another guest (it 9 example is a KVM guest that in turn runs on a KVM guest (the rest of 33 - L1 – level-1 guest; a VM running on L0; also called the "guest 36 - L2 – level-2 guest; a VM running on L1, this is the "nested guest" 46 (guest hypervisor), L3 (nested guest). 61 Provider, using nested KVM lets you rent a large enough "guest 62 hypervisor" (level-1 guest). This in turn allows you to create 66 - Live migration of "guest hypervisors" and their nested guests, for 139 .. note:: If you suspect your L2 (i.e. nested guest) is running slower, 144 Starting a nested guest (x86) [all …]
|
D | mmu.rst | 8 for presenting a standard x86 mmu to the guest, while translating guest 14 the guest should not be able to determine that it is running 19 the guest must not be able to touch host memory not assigned 28 Linux memory management code must be in control of guest memory 32 report writes to guest memory to enable live migration 47 gfn guest frame number 48 gpa guest physical address 49 gva guest virtual address 50 ngpa nested guest physical address 51 ngva nested guest virtual address [all …]
|
D | amd-memory-encryption.rst | 52 The SEV guest key management is handled by a separate processor called the AMD 55 encrypting bootstrap code, snapshot, migrating and debugging the guest. For more 102 context. To create the encryption context, user must provide a guest policy, 113 __u32 policy; /* guest's policy */ 115 … __u64 dh_uaddr; /* userspace address pointing to the guest owner's PDH key */ 118 … __u64 session_addr; /* userspace address which points to the guest session information */ 133 of the memory contents that can be sent to the guest owner as an attestation 153 data encrypted by the KVM_SEV_LAUNCH_UPDATE_DATA command. The guest owner may 154 wait to provide the guest with confidential information until it can verify the 155 measurement. Since the guest owner knows the initial contents of the guest at [all …]
|
D | cpuid.rst | 9 A guest running on a kvm host, can check some of its features using 12 a guest. 65 KVM_FEATURE_PV_UNHALT 7 guest checks this feature bit 69 KVM_FEATURE_PV_TLB_FLUSH 9 guest checks this feature bit 77 KVM_FEATURE_PV_SEND_IPI 11 guest checks this feature bit 85 KVM_FEATURE_PV_SCHED_YIELD 13 guest checks this feature bit 89 KVM_FEATURE_ASYNC_PF_INT 14 guest checks this feature bit 95 KVM_FEATURE_MSI_EXT_DEST_ID 15 guest checks this feature bit 99 KVM_FEATURE_HC_MAP_GPA_RANGE 16 guest checks this feature bit before 103 KVM_FEATURE_MIGRATION_CONTROL 17 guest checks this feature bit before [all …]
|
D | hypercalls.rst | 54 :Purpose: Trigger guest exit so that the host can check for pending 70 :Purpose: Expose hypercall availability to the guest. On x86 platforms, cpuid 81 :Purpose: To enable communication between the hypervisor and guest there is a 83 The guest can map this shared page to access its supervisor register 93 A vcpu of a paravirtualized guest that is busywaiting in guest 98 same guest can wakeup the sleeping vcpu by issuing KVM_HC_KICK_CPU hypercall, 107 :Purpose: Hypercall used to synchronize host and guest clocks. 111 a0: guest physical address where host copies 130 * tsc: guest TSC value used to calculate sec/nsec pair 133 The hypercall lets a guest compute a precise timestamp across [all …]
|
D | msr.rst | 25 in guest RAM. This memory is expected to hold a copy of the following 40 guest has to check version before and after grabbing 64 guest RAM, plus an enable bit in bit 0. This memory is expected to hold 87 guest has to check version before and after grabbing 127 coordinated between the guest and the hypervisor. Availability 139 | | | guest vcpu has been paused by | 196 which must be in guest RAM and must be zeroed. This memory is expected 221 a token that will be used to notify the guest when missing page becomes 225 is currently supported, when set, it indicates that the guest is dealing 232 as regular page fault, guest must reset 'flags' to '0' before it does [all …]
|
/linux-5.19.10/tools/perf/Documentation/ |
D | perf-kvm.txt | 6 perf-kvm - Tool to trace/measure kvm guest os 11 'perf kvm' [--host] [--guest] [--guestmount=<path> 14 'perf kvm' [--host] [--guest] [--guestkallsyms=<path> --guestmodules=<path> 23 a performance counter profile of guest os in realtime 28 default behavior of perf kvm as --guest, so if neither --host nor --guest 29 is input, the perf data file name is perf.data.guest. If --host is input, 31 perf.data.host, please input --host --no-guest. The behaviors are shown as 33 Default('') -> perf.data.guest 35 --guest -> perf.data.guest 36 --host --guest -> perf.data.kvm [all …]
|
/linux-5.19.10/Documentation/x86/ |
D | tdx.rst | 7 Intel's Trust Domain Extensions (TDX) protect confidential guest VMs from 8 the host and physical attacks by isolating the guest register state and by 9 encrypting the guest memory. In TDX, a special module running in a special 10 mode sits between the host and the guest and manages the guest/host 13 Since the host cannot directly access guest registers or memory, much 14 normal functionality of a hypervisor must be moved into the guest. This is 16 guest kernel. A #VE is handled entirely inside the guest kernel, but some 20 guest to the hypervisor or the TDX module. 64 indicates a bug in the guest. The guest may try to handle the #GP with a 70 The "just works" MSRs do not need any special guest handling. They might [all …]
|
/linux-5.19.10/tools/testing/vsock/ |
D | README | 3 These tests exercise net/vmw_vsock/ host<->guest sockets for VMware, KVM, and 16 3. Install the kernel and tests inside the guest. 17 4. Boot the guest and ensure that the AF_VSOCK transport is enabled. 21 # host=server, guest=client 25 (guest)# $TEST_BINARY --mode=client \ 30 # host=client, guest=server 31 (guest)# $TEST_BINARY --mode=server \
|
/linux-5.19.10/Documentation/ABI/testing/ |
D | sysfs-hypervisor-xen | 6 Type of guest: 7 "Xen": standard guest type on arm 8 "HVM": fully virtualized guest (x86) 9 "PV": paravirtualized guest (x86) 10 "PVH": fully virtualized guest without legacy emulation (x86) 22 "self" The guest can profile itself 23 "hv" The guest can profile itself and, if it is 25 "all" The guest can profile itself, the hypervisor
|
D | sysfs-driver-pciback | 7 the format of DDDD:BB:DD.F-REG:SIZE:MASK will allow the guest 14 will allow the guest to read and write to the configuration 23 MSI, MSI-X) set by a connected guest. It is meant to be set 24 only when the guest is a stubdomain hosting device model (qemu) 27 to a PV guest. The device is automatically removed from this
|
/linux-5.19.10/Documentation/virt/kvm/s390/ |
D | s390-pv.rst | 10 access VM state like guest memory or guest registers. Instead, the 15 Each guest starts in non-protected mode and then may make a request to 16 transition into protected mode. On transition, KVM registers the guest 20 The Ultravisor will secure and decrypt the guest's boot memory 22 starts/stops and injected interrupts while the guest is running. 24 As access to the guest's state, such as the SIE state description, is 29 reduce exposed guest state. 40 field (offset 0x54). If the guest cpu is not enabled for the interrupt 50 access to the guest memory. 84 instruction text, in order not to leak guest instruction text. [all …]
|
/linux-5.19.10/arch/x86/xen/ |
D | Kconfig | 7 bool "Xen guest support" 19 bool "Xen PV guest support" 28 Support running as a Xen PV guest. 60 bool "Xen PVHVM guest support" 64 Support running as a Xen PVHVM guest. 80 bool "Xen PVH guest support" 85 Support for running as a Xen PVH guest. 94 Support running as a Xen Dom0 guest.
|
/linux-5.19.10/Documentation/arm64/ |
D | perf.rst | 34 For the guest this attribute will exclude EL1. Please note that EL2 is 35 never counted within a guest. 48 guest/host transitions. 50 For the guest this attribute has no effect. Please note that EL2 is 51 never counted within a guest. 57 These attributes exclude the KVM host and guest, respectively. 62 The KVM guest may run at EL0 (userspace) and EL1 (kernel). 66 must enable/disable counting on the entry and exit to the guest. This is 70 exiting the guest we disable/enable the event as appropriate based on the 74 for exclude_host. Upon entering and exiting the guest we modify the event [all …]
|
/linux-5.19.10/Documentation/ABI/stable/ |
D | sysfs-hypervisor-xen | 33 Space separated list of supported guest system types. Each type 40 <major>: major guest interface version 41 <minor>: minor guest interface version 43 "x86_32": 32 bit x86 guest without PAE 44 "x86_32p": 32 bit x86 guest with PAE 45 "x86_64": 64 bit x86 guest 46 "armv7l": 32 bit arm guest 47 "aarch64": 64 bit arm guest 64 Features the Xen hypervisor supports for the guest as defined 96 UUID of the guest as known to the Xen hypervisor.
|
/linux-5.19.10/tools/virtio/virtio-trace/ |
D | README | 4 Trace agent is a user tool for sending trace data of a guest to a Host in low 48 For example, if a guest use three CPUs, the names are 83 example, if a guest use three CPUs, chardev names should be trace-path-cpu0, 86 3) Boot the guest 87 You can find some chardev in /dev/virtio-ports/ in the guest. 93 0) Build trace agent in a guest 96 1) Enable ftrace in the guest 100 2) Run trace agent in the guest 104 option, trace data are output via stdout in the guest. 109 the guest will stop by specification of chardev in QEMU. This blocking mode may [all …]
|
/linux-5.19.10/Documentation/virt/coco/ |
D | sev-guest.rst | 10 The SEV API is a set of ioctls that are used by the guest or hypervisor 22 This section describes ioctls that is used for querying the SEV guest report 30 hypervisor or guest. The ioctl can be used inside the guest or the 40 The guest ioctl should be issued on a file descriptor of the /dev/sev-guest device. 47 the guests message sequence counter. If guest driver fails to increment message 68 :Type: guest ioctl 83 :Type: guest ioctl 88 The derived key can be used by the guest for any purpose, such as sealing keys 102 :Type: guest ioctl 139 expected contents of a CPUID page as part of guest initialization will be [all …]
|
/linux-5.19.10/Documentation/virt/kvm/arm/ |
D | hypercalls.rst | 11 This means that a guest booted on two different versions of KVM can observe 12 two different "firmware" revisions. This could cause issues if a given guest 15 guest. 28 and power-off to the guest. 40 offered by KVM to the guest via a HVC call. The workaround is described 48 guest is unknown. 51 available to the guest and required for the mitigation. 54 is available to the guest, but it is not needed on this VCPU. 58 offered by KVM to the guest via a HVC call. The workaround is described 83 bitmap is translated to the services that are available to the guest. [all …]
|
/linux-5.19.10/Documentation/filesystems/ |
D | virtiofs.rst | 6 virtiofs: virtio-fs host<->guest shared file system 14 VIRTIO "virtio-fs" device for guest<->host file system sharing. It allows a 15 guest to mount a directory that has been exported on the host. 24 expose the storage network to the guest. The virtio-fs device was designed to 28 guest and host to increase performance and provide semantics that are not 37 guest# mount -t virtiofs myfs /mnt 60 client. The guest acts as the FUSE client while the host acts as the FUSE 65 response portion of the buffer is filled in by the host and the guest handles
|