Lines Matching refs:lock
102 * kvm->lock --> kvm->slots_lock --> kvm->irq_lock
828 * each cache's lock. There are relatively few caches in existence at
1202 mutex_init(&kvm->lock);
1328 * We do not need to take the kvm->lock here, because nobody else
1848 * slots_arch_lock in kvm_swap_active_memslots() and re-acquiring the lock
1930 * (and without a lock), a window would exist between effecting the
3991 * (preempted lock holder), indicated by @in_spin_loop.
4000 * to preempted lock-holder could result in wrong VCPU selection and CPU
4001 * burning. Giving priority for a potential lock-holder increases lock
4080 * VCPU is holding the lock that we need and will release it.
4253 mutex_lock(&kvm->lock);
4255 mutex_unlock(&kvm->lock);
4261 mutex_unlock(&kvm->lock);
4266 mutex_unlock(&kvm->lock);
4295 mutex_lock(&kvm->lock);
4298 /* Ensure that lockdep knows vcpu->mutex is taken *inside* kvm->lock */
4331 mutex_unlock(&kvm->lock);
4340 mutex_unlock(&kvm->lock);
4349 mutex_lock(&kvm->lock);
4351 mutex_unlock(&kvm->lock);
4726 mutex_lock(&kvm->lock);
4729 mutex_unlock(&kvm->lock);
4803 mutex_lock(&kvm->lock);
4806 mutex_unlock(&kvm->lock);
4811 mutex_unlock(&kvm->lock);
4820 mutex_lock(&kvm->lock);
4824 mutex_unlock(&kvm->lock);
4932 mutex_lock(&kvm->lock);
4942 mutex_unlock(&kvm->lock);