/linux-6.6.21/arch/arm64/kvm/ |
D | pvtime.c | 16 u64 base = vcpu->arch.steal.base; in kvm_update_stolen_time() 17 u64 last_steal = vcpu->arch.steal.last_steal; in kvm_update_stolen_time() 19 u64 steal = 0; in kvm_update_stolen_time() local 26 if (!kvm_get_guest(kvm, base + offset, steal)) { in kvm_update_stolen_time() 27 steal = le64_to_cpu(steal); in kvm_update_stolen_time() 28 vcpu->arch.steal.last_steal = READ_ONCE(current->sched_info.run_delay); in kvm_update_stolen_time() 29 steal += vcpu->arch.steal.last_steal - last_steal; in kvm_update_stolen_time() 30 kvm_put_guest(kvm, base + offset, cpu_to_le64(steal)); in kvm_update_stolen_time() 43 if (vcpu->arch.steal.base != INVALID_GPA) in kvm_hypercall_pv_features() 55 u64 base = vcpu->arch.steal.base; in kvm_init_stolen_time() [all …]
|
/linux-6.6.21/drivers/media/pci/ivtv/ |
D | ivtv-queue.c | 110 int ivtv_queue_move(struct ivtv_stream *s, struct ivtv_queue *from, struct ivtv_queue *steal, in ivtv_queue_move() argument 126 bytes_steal = (from_free && steal) ? steal->length : 0; in ivtv_queue_move() 132 while (steal && bytes_available < needed_bytes) { in ivtv_queue_move() 133 struct ivtv_buffer *buf = list_entry(steal->list.prev, struct ivtv_buffer, list); in ivtv_queue_move() 141 list_move_tail(steal->list.prev, &from->list); in ivtv_queue_move() 143 steal->buffers--; in ivtv_queue_move() 144 steal->length -= s->buf_size; in ivtv_queue_move() 145 steal->bytesused -= buf->bytesused - buf->readpos; in ivtv_queue_move() 150 if (list_empty(&steal->list)) in ivtv_queue_move() 152 buf = list_entry(steal->list.prev, struct ivtv_buffer, list); in ivtv_queue_move()
|
D | ivtv-queue.h | 63 int ivtv_queue_move(struct ivtv_stream *s, struct ivtv_queue *from, struct ivtv_queue *steal,
|
/linux-6.6.21/kernel/sched/ |
D | cputime.c | 257 u64 steal; in steal_account_process_time() local 259 steal = paravirt_steal_clock(smp_processor_id()); in steal_account_process_time() 260 steal -= this_rq()->prev_steal_time; in steal_account_process_time() 261 steal = min(steal, maxtime); in steal_account_process_time() 262 account_steal_time(steal); in steal_account_process_time() 263 this_rq()->prev_steal_time += steal; in steal_account_process_time() 265 return steal; in steal_account_process_time() 489 u64 cputime, steal; in account_process_tick() local 500 steal = steal_account_process_time(ULONG_MAX); in account_process_tick() 502 if (steal >= cputime) in account_process_tick() [all …]
|
D | core.c | 700 s64 __maybe_unused steal = 0, irq_delta = 0; in update_rq_clock_task() local 730 steal = paravirt_steal_clock(cpu_of(rq)); in update_rq_clock_task() 731 steal -= rq->prev_steal_time_rq; in update_rq_clock_task() 733 if (unlikely(steal > delta)) in update_rq_clock_task() 734 steal = delta; in update_rq_clock_task() 736 rq->prev_steal_time_rq += steal; in update_rq_clock_task() 737 delta -= steal; in update_rq_clock_task() 744 if ((irq_delta + steal) && sched_feat(NONTASK_CAPACITY)) in update_rq_clock_task() 745 update_irq_load_avg(rq, irq_delta + steal); in update_rq_clock_task()
|
/linux-6.6.21/fs/proc/ |
D | stat.c | 85 u64 user, nice, system, idle, iowait, irq, softirq, steal; in show_stat() local 93 irq = softirq = steal = 0; in show_stat() 112 steal += cpustat[CPUTIME_STEAL]; in show_stat() 134 seq_put_decimal_ull(p, " ", nsec_to_clock_t(steal)); in show_stat() 153 steal = cpustat[CPUTIME_STEAL]; in show_stat() 164 seq_put_decimal_ull(p, " ", nsec_to_clock_t(steal)); in show_stat()
|
/linux-6.6.21/arch/x86/kernel/cpu/ |
D | vmware.c | 220 struct vmware_steal_time *steal = &per_cpu(vmw_steal_time, cpu); in vmware_steal_clock() local 224 clock = READ_ONCE(steal->clock); in vmware_steal_clock() 229 initial_high = READ_ONCE(steal->clock_high); in vmware_steal_clock() 232 low = READ_ONCE(steal->clock_low); in vmware_steal_clock() 235 high = READ_ONCE(steal->clock_high); in vmware_steal_clock()
|
/linux-6.6.21/arch/s390/kernel/ |
D | vtime.c | 207 u64 steal, avg_steal; in vtime_flush() local 212 steal = S390_lowcore.steal_timer; in vtime_flush() 214 if ((s64) steal > 0) { in vtime_flush() 216 account_steal_time(cputime_to_nsecs(steal)); in vtime_flush() 217 avg_steal += steal; in vtime_flush()
|
/linux-6.6.21/kernel/bpf/ |
D | bpf_lru_list.c | 442 int steal, first_steal; in bpf_common_lru_pop_free() local 473 steal = first_steal; in bpf_common_lru_pop_free() 475 steal_loc_l = per_cpu_ptr(clru->local_list, steal); in bpf_common_lru_pop_free() 485 steal = get_next_cpu(steal); in bpf_common_lru_pop_free() 486 } while (!node && steal != first_steal); in bpf_common_lru_pop_free() 488 loc_l->next_steal = steal; in bpf_common_lru_pop_free()
|
/linux-6.6.21/drivers/gpu/drm/radeon/ |
D | radeon_object.c | 547 int steal; in radeon_bo_get_surface_reg() local 560 steal = -1; in radeon_bo_get_surface_reg() 569 steal = i; in radeon_bo_get_surface_reg() 574 if (steal == -1) in radeon_bo_get_surface_reg() 577 reg = &rdev->surface_regs[steal]; in radeon_bo_get_surface_reg() 580 DRM_DEBUG("stealing surface reg %d from %p\n", steal, old_object); in radeon_bo_get_surface_reg() 583 i = steal; in radeon_bo_get_surface_reg()
|
/linux-6.6.21/tools/testing/selftests/kvm/ |
D | steal_time.c | 49 WRITE_ONCE(guest_stolen_time[cpu], st->steal); in guest_code() 56 WRITE_ONCE(guest_stolen_time[cpu], st->steal); in guest_code() 87 pr_info(" steal: %lld\n", st->steal); in steal_time_dump()
|
/linux-6.6.21/arch/x86/include/uapi/asm/ |
D | kvm_para.h | 63 __u64 steal; member
|
/linux-6.6.21/arch/x86/kernel/ |
D | kvm.c | 405 u64 steal; in kvm_steal_clock() local 413 steal = src->steal; in kvm_steal_clock() 417 return steal; in kvm_steal_clock()
|
/linux-6.6.21/Documentation/translations/zh_CN/admin-guide/ |
D | cpu-load.rst | 12 avg-cpu: %user %nice %system %iowait %steal %idle
|
/linux-6.6.21/Documentation/translations/zh_TW/admin-guide/ |
D | cpu-load.rst | 18 avg-cpu: %user %nice %system %iowait %steal %idle
|
/linux-6.6.21/fs/btrfs/ |
D | space-info.h | 163 bool steal; member
|
D | space-info.c | 956 if (!ticket->steal) in steal_from_global_rsv() 1689 ticket.steal = can_steal(flush); in __reserve_bytes()
|
/linux-6.6.21/arch/arm64/include/asm/ |
D | kvm_host.h | 590 } steal; member 1033 vcpu_arch->steal.base = INVALID_GPA; in kvm_arm_pvtime_vcpu_init() 1038 return (vcpu_arch->steal.base != INVALID_GPA); in kvm_arm_is_pvtime_enabled()
|
/linux-6.6.21/Documentation/admin-guide/ |
D | cpu-load.rst | 12 avg-cpu: %user %nice %system %iowait %steal %idle
|
/linux-6.6.21/lib/raid6/ |
D | altivec.uc | 21 * you can just "steal" the vec unit with enable_kernel_altivec() (but
|
/linux-6.6.21/Documentation/virt/kvm/x86/ |
D | cpuid.rst | 58 KVM_FEATURE_STEAL_TIME 5 steal time can be enabled by
|
D | msr.rst | 269 __u64 steal; 296 steal: 299 reported as steal time.
|
/linux-6.6.21/Documentation/arch/x86/ |
D | intel_txt.rst | 161 attempt to crash the system to gain control on reboot and steal
|
/linux-6.6.21/Documentation/target/ |
D | tcmu-design.rst | 192 command(a.k.a steal the original command's entry).
|
/linux-6.6.21/arch/x86/kvm/ |
D | x86.c | 3525 u64 steal; in record_steal_time() local 3602 unsafe_get_user(steal, &st->steal, out); in record_steal_time() 3603 steal += current->sched_info.run_delay - in record_steal_time() 3606 unsafe_put_user(steal, &st->steal, out); in record_steal_time()
|