/linux/arch/arm64/kvm/ |
A D | pvtime.c | 16 u64 base = vcpu->arch.steal.base; in kvm_update_stolen_time() 17 u64 last_steal = vcpu->arch.steal.last_steal; in kvm_update_stolen_time() 19 u64 steal = 0; in kvm_update_stolen_time() local 26 if (!kvm_get_guest(kvm, base + offset, steal)) { in kvm_update_stolen_time() 27 steal = le64_to_cpu(steal); in kvm_update_stolen_time() 29 steal += vcpu->arch.steal.last_steal - last_steal; in kvm_update_stolen_time() 43 if (vcpu->arch.steal.base != GPA_INVALID) in kvm_hypercall_pv_features() 55 u64 base = vcpu->arch.steal.base; in kvm_init_stolen_time() 92 if (vcpu->arch.steal.base != GPA_INVALID) in kvm_arm_pvtime_set_attr() 102 vcpu->arch.steal.base = ipa; in kvm_arm_pvtime_set_attr() [all …]
|
/linux/drivers/media/pci/ivtv/ |
A D | ivtv-queue.c | 110 int ivtv_queue_move(struct ivtv_stream *s, struct ivtv_queue *from, struct ivtv_queue *steal, in ivtv_queue_move() argument 126 bytes_steal = (from_free && steal) ? steal->length : 0; in ivtv_queue_move() 132 while (steal && bytes_available < needed_bytes) { in ivtv_queue_move() 133 struct ivtv_buffer *buf = list_entry(steal->list.prev, struct ivtv_buffer, list); in ivtv_queue_move() 141 list_move_tail(steal->list.prev, &from->list); in ivtv_queue_move() 143 steal->buffers--; in ivtv_queue_move() 144 steal->length -= s->buf_size; in ivtv_queue_move() 145 steal->bytesused -= buf->bytesused - buf->readpos; in ivtv_queue_move() 150 if (list_empty(&steal->list)) in ivtv_queue_move() 152 buf = list_entry(steal->list.prev, struct ivtv_buffer, list); in ivtv_queue_move()
|
A D | ivtv-queue.h | 62 int ivtv_queue_move(struct ivtv_stream *s, struct ivtv_queue *from, struct ivtv_queue *steal,
|
/linux/kernel/sched/ |
A D | cputime.c | 239 u64 steal; in steal_account_process_time() local 242 steal -= this_rq()->prev_steal_time; in steal_account_process_time() 243 steal = min(steal, maxtime); in steal_account_process_time() 244 account_steal_time(steal); in steal_account_process_time() 247 return steal; in steal_account_process_time() 471 u64 cputime, steal; in account_process_tick() local 484 if (steal >= cputime) in account_process_tick() 487 cputime -= steal; in account_process_tick() 503 u64 cputime, steal; in account_idle_ticks() local 513 if (steal >= cputime) in account_idle_ticks() [all …]
|
A D | core.c | 620 s64 __maybe_unused steal = 0, irq_delta = 0; in update_rq_clock_task() local 648 steal = paravirt_steal_clock(cpu_of(rq)); in update_rq_clock_task() 649 steal -= rq->prev_steal_time_rq; in update_rq_clock_task() 651 if (unlikely(steal > delta)) in update_rq_clock_task() 652 steal = delta; in update_rq_clock_task() 654 rq->prev_steal_time_rq += steal; in update_rq_clock_task() 655 delta -= steal; in update_rq_clock_task() 662 if ((irq_delta + steal) && sched_feat(NONTASK_CAPACITY)) in update_rq_clock_task() 663 update_irq_load_avg(rq, irq_delta + steal); in update_rq_clock_task()
|
/linux/fs/proc/ |
A D | stat.c | 111 u64 user, nice, system, idle, iowait, irq, softirq, steal; in show_stat() local 119 irq = softirq = steal = 0; in show_stat() 138 steal += cpustat[CPUTIME_STEAL]; in show_stat() 160 seq_put_decimal_ull(p, " ", nsec_to_clock_t(steal)); in show_stat() 179 steal = cpustat[CPUTIME_STEAL]; in show_stat() 190 seq_put_decimal_ull(p, " ", nsec_to_clock_t(steal)); in show_stat()
|
/linux/arch/x86/kernel/cpu/ |
A D | vmware.c | 220 struct vmware_steal_time *steal = &per_cpu(vmw_steal_time, cpu); in vmware_steal_clock() local 224 clock = READ_ONCE(steal->clock); in vmware_steal_clock() 229 initial_high = READ_ONCE(steal->clock_high); in vmware_steal_clock() 232 low = READ_ONCE(steal->clock_low); in vmware_steal_clock() 235 high = READ_ONCE(steal->clock_high); in vmware_steal_clock()
|
/linux/arch/s390/kernel/ |
A D | vtime.c | 208 u64 steal, avg_steal; in vtime_flush() local 213 steal = S390_lowcore.steal_timer; in vtime_flush() 215 if ((s64) steal > 0) { in vtime_flush() 217 account_steal_time(cputime_to_nsecs(steal)); in vtime_flush() 218 avg_steal += steal; in vtime_flush()
|
/linux/kernel/bpf/ |
A D | bpf_lru_list.c | 437 int steal, first_steal; in bpf_common_lru_pop_free() local 468 steal = first_steal; in bpf_common_lru_pop_free() 470 steal_loc_l = per_cpu_ptr(clru->local_list, steal); in bpf_common_lru_pop_free() 480 steal = get_next_cpu(steal); in bpf_common_lru_pop_free() 481 } while (!node && steal != first_steal); in bpf_common_lru_pop_free() 483 loc_l->next_steal = steal; in bpf_common_lru_pop_free()
|
/linux/drivers/gpu/drm/radeon/ |
A D | radeon_object.c | 561 int steal; in radeon_bo_get_surface_reg() local 575 steal = -1; in radeon_bo_get_surface_reg() 584 steal = i; in radeon_bo_get_surface_reg() 589 if (steal == -1) in radeon_bo_get_surface_reg() 592 reg = &rdev->surface_regs[steal]; in radeon_bo_get_surface_reg() 595 DRM_DEBUG("stealing surface reg %d from %p\n", steal, old_object); in radeon_bo_get_surface_reg() 598 i = steal; in radeon_bo_get_surface_reg()
|
/linux/tools/testing/selftests/kvm/ |
A D | steal_time.c | 49 WRITE_ONCE(guest_stolen_time[cpu], st->steal); in guest_code() 56 WRITE_ONCE(guest_stolen_time[cpu], st->steal); in guest_code() 91 pr_info(" steal: %lld\n", st->steal); in steal_time_dump()
|
/linux/arch/x86/kernel/ |
A D | kvm.c | 392 u64 steal; in kvm_steal_clock() local 400 steal = src->steal; in kvm_steal_clock() 404 return steal; in kvm_steal_clock()
|
/linux/arch/x86/include/uapi/asm/ |
A D | kvm_para.h | 63 __u64 steal; member
|
/linux/arch/arm64/include/asm/ |
A D | kvm_host.h | 387 } steal; member 695 vcpu_arch->steal.base = GPA_INVALID; in kvm_arm_pvtime_vcpu_init() 700 return (vcpu_arch->steal.base != GPA_INVALID); in kvm_arm_is_pvtime_enabled()
|
/linux/Documentation/translations/zh_CN/admin-guide/ |
A D | cpu-load.rst | 12 avg-cpu: %user %nice %system %iowait %steal %idle
|
/linux/Documentation/translations/zh_TW/admin-guide/ |
A D | cpu-load.rst | 18 avg-cpu: %user %nice %system %iowait %steal %idle
|
/linux/fs/btrfs/ |
A D | space-info.h | 75 bool steal; member
|
A D | space-info.c | 902 if (!aborted && ticket->steal && in maybe_fail_all_tickets() 1514 ticket.steal = (flush == BTRFS_RESERVE_FLUSH_ALL_STEAL); in __reserve_bytes()
|
/linux/Documentation/admin-guide/ |
A D | cpu-load.rst | 12 avg-cpu: %user %nice %system %iowait %steal %idle
|
/linux/lib/raid6/ |
A D | altivec.uc | 21 * you can just "steal" the vec unit with enable_kernel_altivec() (but
|
/linux/Documentation/virt/kvm/ |
A D | cpuid.rst | 58 KVM_FEATURE_STEAL_TIME 5 steal time can be enabled by
|
A D | msr.rst | 269 __u64 steal; 296 steal: 299 reported as steal time.
|
/linux/Documentation/x86/ |
A D | intel_txt.rst | 161 attempt to crash the system to gain control on reboot and steal
|
/linux/Documentation/target/ |
A D | tcmu-design.rst | 192 command(a.k.a steal the original command's entry).
|
/linux/Documentation/vm/ |
A D | frontswap.rst | 155 Dickins has observed that frontswap could probably steal one of
|