/linux-master/arch/arm64/kvm/ |
H A D | pvtime.c | 16 u64 base = vcpu->arch.steal.base; 17 u64 last_steal = vcpu->arch.steal.last_steal; 19 u64 steal = 0; local 26 if (!kvm_get_guest(kvm, base + offset, steal)) { 27 steal = le64_to_cpu(steal); 28 vcpu->arch.steal.last_steal = READ_ONCE(current->sched_info.run_delay); 29 steal += vcpu->arch.steal.last_steal - last_steal; 30 kvm_put_guest(kvm, base + offset, cpu_to_le64(steal)); [all...] |
/linux-master/arch/riscv/kernel/ |
H A D | paravirt.c | 42 early_param("no-steal-acc", parse_no_stealacc); 66 pr_warn("Failed to disable steal-time shmem"); 68 pr_warn("Failed to set steal-time shmem"); 95 __le64 steal; local 98 * Check the sequence field before and after reading the steal 104 steal = READ_ONCE(st->steal); 109 return le64_to_cpu(steal); 132 pr_info("Computing paravirt steal-time\n");
|
/linux-master/drivers/media/pci/ivtv/ |
H A D | ivtv-queue.c | 95 If 'steal' != NULL, then buffers may also taken from that queue if 99 also cleared if buffers need to be taken from the 'steal' queue and 104 bytesused value. For the 'steal' queue the total available buffer 110 int ivtv_queue_move(struct ivtv_stream *s, struct ivtv_queue *from, struct ivtv_queue *steal, argument 126 bytes_steal = (from_free && steal) ? steal->length : 0; 132 while (steal && bytes_available < needed_bytes) { 133 struct ivtv_buffer *buf = list_entry(steal->list.prev, struct ivtv_buffer, list); 136 /* move buffers from the tail of the 'steal' queue to the tail of the 141 list_move_tail(steal [all...] |
H A D | ivtv-queue.h | 63 int ivtv_queue_move(struct ivtv_stream *s, struct ivtv_queue *from, struct ivtv_queue *steal,
|
/linux-master/fs/proc/ |
H A D | stat.c | 85 u64 user, nice, system, idle, iowait, irq, softirq, steal; local 93 irq = softirq = steal = 0; 112 steal += cpustat[CPUTIME_STEAL]; 134 seq_put_decimal_ull(p, " ", nsec_to_clock_t(steal)); 153 steal = cpustat[CPUTIME_STEAL]; 164 seq_put_decimal_ull(p, " ", nsec_to_clock_t(steal));
|
/linux-master/arch/riscv/kvm/ |
H A D | vcpu_sbi_sta.c | 34 u64 steal; local 57 offsetof(struct sbi_sta_struct, steal)); 69 steal = le64_to_cpu(steal_le); 71 steal += vcpu->arch.sta.last_steal - last_steal; 72 WARN_ON(put_user(cpu_to_le64(steal), steal_ptr));
|
/linux-master/kernel/sched/ |
H A D | cputime.c | 257 u64 steal; local 259 steal = paravirt_steal_clock(smp_processor_id()); 260 steal -= this_rq()->prev_steal_time; 261 steal = min(steal, maxtime); 262 account_steal_time(steal); 263 this_rq()->prev_steal_time += steal; 265 return steal; 272 * Account how much elapsed time was spent in steal, irq, or softirq time. 383 * once, including some ticks of steal, ir 489 u64 cputime, steal; local 521 u64 cputime, steal; local [all...] |
H A D | core.c | 701 s64 __maybe_unused steal = 0, irq_delta = 0; local 731 steal = paravirt_steal_clock(cpu_of(rq)); 732 steal -= rq->prev_steal_time_rq; 734 if (unlikely(steal > delta)) 735 steal = delta; 737 rq->prev_steal_time_rq += steal; 738 delta -= steal; 745 if ((irq_delta + steal) && sched_feat(NONTASK_CAPACITY)) 746 update_irq_load_avg(rq, irq_delta + steal); 7484 * which excludes things like IRQ and steal [all...] |
/linux-master/tools/testing/selftests/kvm/ |
H A D | steal_time.c | 3 * steal/stolen time test 51 WRITE_ONCE(guest_stolen_time[cpu], st->steal); 58 WRITE_ONCE(guest_stolen_time[cpu], st->steal); 89 pr_info(" steal: %lld\n", st->steal); 218 uint64_t steal; member in struct:sta_struct 257 WRITE_ONCE(guest_stolen_time[cpu], st->steal); 264 WRITE_ONCE(guest_stolen_time[cpu], st->steal); 297 pr_info(" steal: %"PRIu64"\n", st->steal); [all...] |
/linux-master/arch/x86/kernel/cpu/ |
H A D | vmware.c | 130 static bool steal_acc __initdata = true; /* steal time accounting */ 144 early_param("no-steal-acc", parse_no_stealacc); 208 * vmware_steal_clock() - read the per-cpu steal clock 209 * @cpu: the cpu number whose steal clock we want to read 211 * The function reads the steal clock if we are on a 64-bit system, otherwise 216 * The steal clock reading in ns. 220 struct vmware_steal_time *steal = &per_cpu(vmw_steal_time, cpu); local 224 clock = READ_ONCE(steal->clock); 229 initial_high = READ_ONCE(steal->clock_high); 232 low = READ_ONCE(steal [all...] |
/linux-master/arch/x86/include/uapi/asm/ |
H A D | kvm_para.h | 63 __u64 steal; member in struct:kvm_steal_time
|
/linux-master/arch/s390/kernel/ |
H A D | vtime.c | 207 u64 steal, avg_steal; local 212 steal = S390_lowcore.steal_timer; 214 if ((s64) steal > 0) { 216 account_steal_time(cputime_to_nsecs(steal)); 217 avg_steal += steal;
|
/linux-master/kernel/bpf/ |
H A D | bpf_lru_list.c | 442 int steal, first_steal; local 473 steal = first_steal; 475 steal_loc_l = per_cpu_ptr(clru->local_list, steal); 485 steal = get_next_cpu(steal); 486 } while (!node && steal != first_steal); 488 loc_l->next_steal = steal;
|
/linux-master/drivers/gpu/drm/radeon/ |
H A D | radeon_object.c | 545 int steal; local 558 steal = -1; 567 steal = i; 572 if (steal == -1) 575 reg = &rdev->surface_regs[steal]; 578 DRM_DEBUG("stealing surface reg %d from %p\n", steal, old_object); 581 i = steal;
|
/linux-master/fs/btrfs/ |
H A D | space-info.h | 56 * Pretty much the same as FLUSH_ALL, but can also steal space from 173 bool steal; member in struct:reserve_ticket
|
H A D | space-info.c | 974 if (!ticket->steal) 1422 * Attempt to steal from the global rsv if we can, except if the fs was 1425 * success to the caller if we can steal from the global rsv - this is 1708 ticket.steal = can_steal(flush);
|
/linux-master/arch/riscv/include/asm/ |
H A D | sbi.h | 256 /* SBI STA (steal-time accounting) extension */ 264 __le64 steal; member in struct:sbi_sta_struct
|
/linux-master/arch/x86/kernel/ |
H A D | kvm.c | 66 early_param("no-steal-acc", parse_no_stealacc); 406 u64 steal; local 414 steal = src->steal; 418 return steal;
|
/linux-master/arch/arm64/include/asm/ |
H A D | kvm_host.h | 689 } steal; member in struct:kvm_vcpu_arch 1157 vcpu_arch->steal.base = INVALID_GPA; 1162 return (vcpu_arch->steal.base != INVALID_GPA);
|
/linux-master/arch/x86/kvm/ |
H A D | x86.c | 3681 u64 steal; local 3758 unsafe_get_user(steal, &st->steal, out); 3759 steal += current->sched_info.run_delay - 3762 unsafe_put_user(steal, &st->steal, out);
|