/linux-master/arch/powerpc/lib/ |
H A D | qspinlock.c | 19 u8 sleepy; /* 1 if the previous vCPU was preempted or 283 /* Called inside spin_begin(). Returns whether or not the vCPU was preempted. */ 288 bool preempted = false; local 307 preempted = true; 327 return preempted; 333 return preempted; 336 /* Called inside spin_begin(). Returns whether or not the vCPU was preempted. */ 342 /* Called inside spin_begin(). Returns whether or not the vCPU was preempted. */ 379 bool preempted = false; local 388 * If the previous waiter was preempted i 471 bool preempted = false; local 610 bool preempted; local [all...] |
/linux-master/arch/x86/kernel/ |
H A D | asm-offsets_64.c | 24 OFFSET(KVM_STEAL_TIME_preempted, kvm_steal_time, preempted);
|
H A D | kvm.c | 664 * The local vCPU is never preempted, so we do not explicitly 669 state = READ_ONCE(src->preempted); 671 if (try_cmpxchg(&src->preempted, &state, 789 return !!(src->preempted & KVM_VCPU_PREEMPTED); 1074 * preferred over native qspinlock when vCPU is preempted.
|
/linux-master/arch/x86/include/uapi/asm/ |
H A D | kvm_para.h | 66 __u8 preempted; member in struct:kvm_steal_time
|
/linux-master/tools/testing/selftests/kvm/ |
H A D | steal_time.c | 37 GUEST_ASSERT_EQ(READ_ONCE(st->preempted), 0); 92 pr_info(" preempted: %d\n", st->preempted); 219 uint8_t preempted; member in struct:sta_struct 240 GUEST_ASSERT(READ_ONCE(st->preempted) == 0); 298 pr_info(" preempted: %d\n", st->preempted);
|
/linux-master/drivers/gpu/drm/i915/gvt/ |
H A D | execlist.h | 102 u32 preempted :1; member in struct:execlist_context_status_format::__anon631::__anon632
|
H A D | execlist.c | 337 * ----> lite-restore + preempted 351 status.lite_restore = status.preempted = 1;
|
/linux-master/arch/x86/kvm/ |
H A D | xen.h | 181 * If the vCPU wasn't preempted but took a normal exit for 187 if (WARN_ON_ONCE(!vcpu->preempted))
|
H A D | x86.c | 3726 "+m" (st->preempted)); 3732 vcpu->arch.st.preempted = 0; 3745 unsafe_put_user(0, &st->preempted, out); 3746 vcpu->arch.st.preempted = 0; 5069 static const u8 preempted = KVM_VCPU_PREEMPTED; local 5073 * The vCPU can be marked preempted if and only if the VM-Exit was on 5077 * preempted if and only if the VM-Exit was due to a host interrupt. 5088 if (vcpu->arch.st.preempted) 5103 BUILD_BUG_ON(sizeof(st->preempted) != sizeof(preempted)); [all...] |
/linux-master/arch/arc/kernel/ |
H A D | entry-arcv2.S | 203 ; Exception in U mode, preempted in kernel, Intr taken (K mode), orig 230 ; entry was via Exception in DS which got preempted in kernel).
|
/linux-master/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_cmdbuf.c | 50 * @preempted: List of preempted command buffers. 57 struct list_head preempted; member in struct:vmw_cmdbuf_context 325 INIT_LIST_HEAD(&ctx->preempted); 406 list_add_tail(&entry->list, &ctx->preempted); 587 /* Move preempted command buffers to the preempted queue. */ 591 * Add the preempted queue after the command buffer 594 list_splice_init(&ctx->preempted, restart_head[i].prev); 624 * @check_preempted: Check also the preempted queu [all...] |
/linux-master/arch/x86/kvm/vmx/ |
H A D | posted_intr.c | 76 * Clear SN if it was set due to being preempted. Again, do 207 * Set SN when the vCPU is preempted. Note, the vCPU can both be seen 208 * as blocking and preempted, e.g. if it's preempted between setting 211 if (vcpu->preempted)
|
/linux-master/arch/mips/kernel/ |
H A D | unaligned.c | 463 unsigned int wd, preempted; local 489 * whether we get preempted in order to avoid 496 preempted = test_thread_flag(TIF_USEDMSA); 512 preempted = 0; 515 } while (preempted);
|
/linux-master/arch/riscv/include/asm/ |
H A D | sbi.h | 265 u8 preempted; member in struct:sbi_sta_struct
|
/linux-master/arch/mips/include/asm/octeon/ |
H A D | cvmx-pko-defs.h | 621 uint64_t preempted:1; member in struct:cvmx_pko_mem_debug6::cvmx_pko_mem_debug6_s 635 uint64_t preempted:1; 678 uint64_t preempted:1; member in struct:cvmx_pko_mem_debug6::cvmx_pko_mem_debug6_cn52xx 694 uint64_t preempted:1;
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_debugfs.c | 1880 bool preempted = true; local 1887 preempted = false; 1905 if (preempted && (&job->hw_fence) == fence) 1906 /* mark the job as preempted */ 1956 DRM_INFO("ring %d was preempted\n", ring->idx);
|
/linux-master/arch/m68k/ifpsp060/ |
H A D | iskeleton.S | 199 | preempted until _060_unlock_page(), so that the page stays mapped.
|
/linux-master/virt/kvm/ |
H A D | kvm_main.c | 496 vcpu->preempted = false; 3991 * (preempted lock holder), indicated by @in_spin_loop. 4000 * to preempted lock-holder could result in wrong VCPU selection and CPU 4050 * vCPU was preempted in kernel mode. All architectures except x86 (or more 4078 * currently running, because it got preempted by something 4103 if (READ_ONCE(vcpu->preempted) && yield_to_kernel_mode && 5693 * can be preempted, but the task cannot be frozen until it has dropped 6324 WRITE_ONCE(vcpu->preempted, false); 6338 WRITE_ONCE(vcpu->preempted, true);
|
/linux-master/include/linux/ |
H A D | kvm_host.h | 385 bool preempted; member in struct:kvm_vcpu
|
/linux-master/arch/x86/include/asm/ |
H A D | kvm_host.h | 897 u8 preempted; member in struct:kvm_vcpu_arch::__anon14 1008 /* be preempted when it's in kernel-mode(cpl=0) */
|