Searched refs:steal (Results 1 - 20 of 20) sorted by relevance

/linux-master/arch/arm64/kvm/
H A Dpvtime.c16 u64 base = vcpu->arch.steal.base;
17 u64 last_steal = vcpu->arch.steal.last_steal;
19 u64 steal = 0; local
26 if (!kvm_get_guest(kvm, base + offset, steal)) {
27 steal = le64_to_cpu(steal);
28 vcpu->arch.steal.last_steal = READ_ONCE(current->sched_info.run_delay);
29 steal += vcpu->arch.steal.last_steal - last_steal;
30 kvm_put_guest(kvm, base + offset, cpu_to_le64(steal));
[all...]
/linux-master/arch/riscv/kernel/
H A Dparavirt.c42 early_param("no-steal-acc", parse_no_stealacc);
66 pr_warn("Failed to disable steal-time shmem");
68 pr_warn("Failed to set steal-time shmem");
95 __le64 steal; local
98 * Check the sequence field before and after reading the steal
104 steal = READ_ONCE(st->steal);
109 return le64_to_cpu(steal);
132 pr_info("Computing paravirt steal-time\n");
/linux-master/drivers/media/pci/ivtv/
H A Divtv-queue.c95 If 'steal' != NULL, then buffers may also taken from that queue if
99 also cleared if buffers need to be taken from the 'steal' queue and
104 bytesused value. For the 'steal' queue the total available buffer
110 int ivtv_queue_move(struct ivtv_stream *s, struct ivtv_queue *from, struct ivtv_queue *steal, argument
126 bytes_steal = (from_free && steal) ? steal->length : 0;
132 while (steal && bytes_available < needed_bytes) {
133 struct ivtv_buffer *buf = list_entry(steal->list.prev, struct ivtv_buffer, list);
136 /* move buffers from the tail of the 'steal' queue to the tail of the
141 list_move_tail(steal
[all...]
H A Divtv-queue.h63 int ivtv_queue_move(struct ivtv_stream *s, struct ivtv_queue *from, struct ivtv_queue *steal,
/linux-master/fs/proc/
H A Dstat.c85 u64 user, nice, system, idle, iowait, irq, softirq, steal; local
93 irq = softirq = steal = 0;
112 steal += cpustat[CPUTIME_STEAL];
134 seq_put_decimal_ull(p, " ", nsec_to_clock_t(steal));
153 steal = cpustat[CPUTIME_STEAL];
164 seq_put_decimal_ull(p, " ", nsec_to_clock_t(steal));
/linux-master/arch/riscv/kvm/
H A Dvcpu_sbi_sta.c34 u64 steal; local
57 offsetof(struct sbi_sta_struct, steal));
69 steal = le64_to_cpu(steal_le);
71 steal += vcpu->arch.sta.last_steal - last_steal;
72 WARN_ON(put_user(cpu_to_le64(steal), steal_ptr));
/linux-master/kernel/sched/
H A Dcputime.c257 u64 steal; local
259 steal = paravirt_steal_clock(smp_processor_id());
260 steal -= this_rq()->prev_steal_time;
261 steal = min(steal, maxtime);
262 account_steal_time(steal);
263 this_rq()->prev_steal_time += steal;
265 return steal;
272 * Account how much elapsed time was spent in steal, irq, or softirq time.
383 * once, including some ticks of steal, ir
489 u64 cputime, steal; local
521 u64 cputime, steal; local
[all...]
H A Dcore.c701 s64 __maybe_unused steal = 0, irq_delta = 0; local
731 steal = paravirt_steal_clock(cpu_of(rq));
732 steal -= rq->prev_steal_time_rq;
734 if (unlikely(steal > delta))
735 steal = delta;
737 rq->prev_steal_time_rq += steal;
738 delta -= steal;
745 if ((irq_delta + steal) && sched_feat(NONTASK_CAPACITY))
746 update_irq_load_avg(rq, irq_delta + steal);
7484 * which excludes things like IRQ and steal
[all...]
/linux-master/tools/testing/selftests/kvm/
H A Dsteal_time.c3 * steal/stolen time test
51 WRITE_ONCE(guest_stolen_time[cpu], st->steal);
58 WRITE_ONCE(guest_stolen_time[cpu], st->steal);
89 pr_info(" steal: %lld\n", st->steal);
218 uint64_t steal; member in struct:sta_struct
257 WRITE_ONCE(guest_stolen_time[cpu], st->steal);
264 WRITE_ONCE(guest_stolen_time[cpu], st->steal);
297 pr_info(" steal: %"PRIu64"\n", st->steal);
[all...]
/linux-master/arch/x86/kernel/cpu/
H A Dvmware.c130 static bool steal_acc __initdata = true; /* steal time accounting */
144 early_param("no-steal-acc", parse_no_stealacc);
208 * vmware_steal_clock() - read the per-cpu steal clock
209 * @cpu: the cpu number whose steal clock we want to read
211 * The function reads the steal clock if we are on a 64-bit system, otherwise
216 * The steal clock reading in ns.
220 struct vmware_steal_time *steal = &per_cpu(vmw_steal_time, cpu); local
224 clock = READ_ONCE(steal->clock);
229 initial_high = READ_ONCE(steal->clock_high);
232 low = READ_ONCE(steal
[all...]
/linux-master/arch/x86/include/uapi/asm/
H A Dkvm_para.h63 __u64 steal; member in struct:kvm_steal_time
/linux-master/arch/s390/kernel/
H A Dvtime.c207 u64 steal, avg_steal; local
212 steal = S390_lowcore.steal_timer;
214 if ((s64) steal > 0) {
216 account_steal_time(cputime_to_nsecs(steal));
217 avg_steal += steal;
/linux-master/kernel/bpf/
H A Dbpf_lru_list.c442 int steal, first_steal; local
473 steal = first_steal;
475 steal_loc_l = per_cpu_ptr(clru->local_list, steal);
485 steal = get_next_cpu(steal);
486 } while (!node && steal != first_steal);
488 loc_l->next_steal = steal;
/linux-master/drivers/gpu/drm/radeon/
H A Dradeon_object.c545 int steal; local
558 steal = -1;
567 steal = i;
572 if (steal == -1)
575 reg = &rdev->surface_regs[steal];
578 DRM_DEBUG("stealing surface reg %d from %p\n", steal, old_object);
581 i = steal;
/linux-master/fs/btrfs/
H A Dspace-info.h56 * Pretty much the same as FLUSH_ALL, but can also steal space from
173 bool steal; member in struct:reserve_ticket
H A Dspace-info.c974 if (!ticket->steal)
1422 * Attempt to steal from the global rsv if we can, except if the fs was
1425 * success to the caller if we can steal from the global rsv - this is
1708 ticket.steal = can_steal(flush);
/linux-master/arch/riscv/include/asm/
H A Dsbi.h256 /* SBI STA (steal-time accounting) extension */
264 __le64 steal; member in struct:sbi_sta_struct
/linux-master/arch/x86/kernel/
H A Dkvm.c66 early_param("no-steal-acc", parse_no_stealacc);
406 u64 steal; local
414 steal = src->steal;
418 return steal;
/linux-master/arch/arm64/include/asm/
H A Dkvm_host.h689 } steal; member in struct:kvm_vcpu_arch
1157 vcpu_arch->steal.base = INVALID_GPA;
1162 return (vcpu_arch->steal.base != INVALID_GPA);
/linux-master/arch/x86/kvm/
H A Dx86.c3681 u64 steal; local
3758 unsafe_get_user(steal, &st->steal, out);
3759 steal += current->sched_info.run_delay -
3762 unsafe_put_user(steal, &st->steal, out);

Completed in 431 milliseconds