Home
last modified time | relevance | path

Searched refs:steal (Results 1 – 25 of 35) sorted by relevance

12

/linux-6.6.21/arch/arm64/kvm/
Dpvtime.c16 u64 base = vcpu->arch.steal.base; in kvm_update_stolen_time()
17 u64 last_steal = vcpu->arch.steal.last_steal; in kvm_update_stolen_time()
19 u64 steal = 0; in kvm_update_stolen_time() local
26 if (!kvm_get_guest(kvm, base + offset, steal)) { in kvm_update_stolen_time()
27 steal = le64_to_cpu(steal); in kvm_update_stolen_time()
28 vcpu->arch.steal.last_steal = READ_ONCE(current->sched_info.run_delay); in kvm_update_stolen_time()
29 steal += vcpu->arch.steal.last_steal - last_steal; in kvm_update_stolen_time()
30 kvm_put_guest(kvm, base + offset, cpu_to_le64(steal)); in kvm_update_stolen_time()
43 if (vcpu->arch.steal.base != INVALID_GPA) in kvm_hypercall_pv_features()
55 u64 base = vcpu->arch.steal.base; in kvm_init_stolen_time()
[all …]
/linux-6.6.21/drivers/media/pci/ivtv/
Divtv-queue.c110 int ivtv_queue_move(struct ivtv_stream *s, struct ivtv_queue *from, struct ivtv_queue *steal, in ivtv_queue_move() argument
126 bytes_steal = (from_free && steal) ? steal->length : 0; in ivtv_queue_move()
132 while (steal && bytes_available < needed_bytes) { in ivtv_queue_move()
133 struct ivtv_buffer *buf = list_entry(steal->list.prev, struct ivtv_buffer, list); in ivtv_queue_move()
141 list_move_tail(steal->list.prev, &from->list); in ivtv_queue_move()
143 steal->buffers--; in ivtv_queue_move()
144 steal->length -= s->buf_size; in ivtv_queue_move()
145 steal->bytesused -= buf->bytesused - buf->readpos; in ivtv_queue_move()
150 if (list_empty(&steal->list)) in ivtv_queue_move()
152 buf = list_entry(steal->list.prev, struct ivtv_buffer, list); in ivtv_queue_move()
Divtv-queue.h63 int ivtv_queue_move(struct ivtv_stream *s, struct ivtv_queue *from, struct ivtv_queue *steal,
/linux-6.6.21/kernel/sched/
Dcputime.c257 u64 steal; in steal_account_process_time() local
259 steal = paravirt_steal_clock(smp_processor_id()); in steal_account_process_time()
260 steal -= this_rq()->prev_steal_time; in steal_account_process_time()
261 steal = min(steal, maxtime); in steal_account_process_time()
262 account_steal_time(steal); in steal_account_process_time()
263 this_rq()->prev_steal_time += steal; in steal_account_process_time()
265 return steal; in steal_account_process_time()
489 u64 cputime, steal; in account_process_tick() local
500 steal = steal_account_process_time(ULONG_MAX); in account_process_tick()
502 if (steal >= cputime) in account_process_tick()
[all …]
Dcore.c700 s64 __maybe_unused steal = 0, irq_delta = 0; in update_rq_clock_task() local
730 steal = paravirt_steal_clock(cpu_of(rq)); in update_rq_clock_task()
731 steal -= rq->prev_steal_time_rq; in update_rq_clock_task()
733 if (unlikely(steal > delta)) in update_rq_clock_task()
734 steal = delta; in update_rq_clock_task()
736 rq->prev_steal_time_rq += steal; in update_rq_clock_task()
737 delta -= steal; in update_rq_clock_task()
744 if ((irq_delta + steal) && sched_feat(NONTASK_CAPACITY)) in update_rq_clock_task()
745 update_irq_load_avg(rq, irq_delta + steal); in update_rq_clock_task()
/linux-6.6.21/fs/proc/
Dstat.c85 u64 user, nice, system, idle, iowait, irq, softirq, steal; in show_stat() local
93 irq = softirq = steal = 0; in show_stat()
112 steal += cpustat[CPUTIME_STEAL]; in show_stat()
134 seq_put_decimal_ull(p, " ", nsec_to_clock_t(steal)); in show_stat()
153 steal = cpustat[CPUTIME_STEAL]; in show_stat()
164 seq_put_decimal_ull(p, " ", nsec_to_clock_t(steal)); in show_stat()
/linux-6.6.21/arch/x86/kernel/cpu/
Dvmware.c220 struct vmware_steal_time *steal = &per_cpu(vmw_steal_time, cpu); in vmware_steal_clock() local
224 clock = READ_ONCE(steal->clock); in vmware_steal_clock()
229 initial_high = READ_ONCE(steal->clock_high); in vmware_steal_clock()
232 low = READ_ONCE(steal->clock_low); in vmware_steal_clock()
235 high = READ_ONCE(steal->clock_high); in vmware_steal_clock()
/linux-6.6.21/arch/s390/kernel/
Dvtime.c207 u64 steal, avg_steal; in vtime_flush() local
212 steal = S390_lowcore.steal_timer; in vtime_flush()
214 if ((s64) steal > 0) { in vtime_flush()
216 account_steal_time(cputime_to_nsecs(steal)); in vtime_flush()
217 avg_steal += steal; in vtime_flush()
/linux-6.6.21/kernel/bpf/
Dbpf_lru_list.c442 int steal, first_steal; in bpf_common_lru_pop_free() local
473 steal = first_steal; in bpf_common_lru_pop_free()
475 steal_loc_l = per_cpu_ptr(clru->local_list, steal); in bpf_common_lru_pop_free()
485 steal = get_next_cpu(steal); in bpf_common_lru_pop_free()
486 } while (!node && steal != first_steal); in bpf_common_lru_pop_free()
488 loc_l->next_steal = steal; in bpf_common_lru_pop_free()
/linux-6.6.21/drivers/gpu/drm/radeon/
Dradeon_object.c547 int steal; in radeon_bo_get_surface_reg() local
560 steal = -1; in radeon_bo_get_surface_reg()
569 steal = i; in radeon_bo_get_surface_reg()
574 if (steal == -1) in radeon_bo_get_surface_reg()
577 reg = &rdev->surface_regs[steal]; in radeon_bo_get_surface_reg()
580 DRM_DEBUG("stealing surface reg %d from %p\n", steal, old_object); in radeon_bo_get_surface_reg()
583 i = steal; in radeon_bo_get_surface_reg()
/linux-6.6.21/tools/testing/selftests/kvm/
Dsteal_time.c49 WRITE_ONCE(guest_stolen_time[cpu], st->steal); in guest_code()
56 WRITE_ONCE(guest_stolen_time[cpu], st->steal); in guest_code()
87 pr_info(" steal: %lld\n", st->steal); in steal_time_dump()
/linux-6.6.21/arch/x86/include/uapi/asm/
Dkvm_para.h63 __u64 steal; member
/linux-6.6.21/arch/x86/kernel/
Dkvm.c405 u64 steal; in kvm_steal_clock() local
413 steal = src->steal; in kvm_steal_clock()
417 return steal; in kvm_steal_clock()
/linux-6.6.21/Documentation/translations/zh_CN/admin-guide/
Dcpu-load.rst12 avg-cpu: %user %nice %system %iowait %steal %idle
/linux-6.6.21/Documentation/translations/zh_TW/admin-guide/
Dcpu-load.rst18 avg-cpu: %user %nice %system %iowait %steal %idle
/linux-6.6.21/fs/btrfs/
Dspace-info.h163 bool steal; member
Dspace-info.c956 if (!ticket->steal) in steal_from_global_rsv()
1689 ticket.steal = can_steal(flush); in __reserve_bytes()
/linux-6.6.21/arch/arm64/include/asm/
Dkvm_host.h590 } steal; member
1033 vcpu_arch->steal.base = INVALID_GPA; in kvm_arm_pvtime_vcpu_init()
1038 return (vcpu_arch->steal.base != INVALID_GPA); in kvm_arm_is_pvtime_enabled()
/linux-6.6.21/Documentation/admin-guide/
Dcpu-load.rst12 avg-cpu: %user %nice %system %iowait %steal %idle
/linux-6.6.21/lib/raid6/
Daltivec.uc21 * you can just "steal" the vec unit with enable_kernel_altivec() (but
/linux-6.6.21/Documentation/virt/kvm/x86/
Dcpuid.rst58 KVM_FEATURE_STEAL_TIME 5 steal time can be enabled by
Dmsr.rst269 __u64 steal;
296 steal:
299 reported as steal time.
/linux-6.6.21/Documentation/arch/x86/
Dintel_txt.rst161 attempt to crash the system to gain control on reboot and steal
/linux-6.6.21/Documentation/target/
Dtcmu-design.rst192 command(a.k.a steal the original command's entry).
/linux-6.6.21/arch/x86/kvm/
Dx86.c3525 u64 steal; in record_steal_time() local
3602 unsafe_get_user(steal, &st->steal, out); in record_steal_time()
3603 steal += current->sched_info.run_delay - in record_steal_time()
3606 unsafe_put_user(steal, &st->steal, out); in record_steal_time()

12