Lines Matching refs:mmu_lock

26 		lockdep_assert_held_read(&kvm->mmu_lock);
28 lockdep_assert_held_write(&kvm->mmu_lock);
112 * role.invalid are protected by mmu_lock.
114 lockdep_assert_held(&kvm->mmu_lock);
154 ({ lockdep_assert_held(&(_kvm)->mmu_lock); }), _root; \
164 ({ lockdep_assert_held(&(_kvm)->mmu_lock); }), _root; \
168 * Iterate over all TDP MMU roots. Requires that mmu_lock be held for write,
169 * the implication being that any flow that holds mmu_lock for read is
171 * Holding mmu_lock for write obviates the need for RCU protection as the list
240 read_lock(&kvm->mmu_lock);
255 * root has been invalidated, which requires holding mmu_lock for write.
279 read_unlock(&kvm->mmu_lock);
283 * mmu_lock, and the root can't be freed as this vCPU holds a reference.
378 * mmu_lock ensures the SPTE can't be made present.
390 * modified by a different vCPU outside of mmu_lock.
399 * task can zap/remove the SPTE as mmu_lock is held for
564 lockdep_assert_held_read(&kvm->mmu_lock);
568 * does not hold the mmu_lock. On failure, i.e. if a different logical
628 lockdep_assert_held_write(&kvm->mmu_lock);
691 if (need_resched() || rwlock_needbreak(&kvm->mmu_lock)) {
698 cond_resched_rwlock_read(&kvm->mmu_lock);
700 cond_resched_rwlock_write(&kvm->mmu_lock);
756 * mmu_lock. When handling an unmap/release mmu_notifier command, KVM
758 * callback. Dropping mmu_lock with an unreachable root would result
772 * preempt models) or mmu_lock contention (full or real-time models).
831 lockdep_assert_held_write(&kvm->mmu_lock);
874 lockdep_assert_held_write(&kvm->mmu_lock);
889 * all that expensive and mmu_lock is already held, which means the
897 lockdep_assert_held_write(&kvm->mmu_lock);
910 read_lock(&kvm->mmu_lock);
938 read_unlock(&kvm->mmu_lock);
944 * zapping is done separately so that it happens with mmu_lock with read,
945 * whereas invalidating roots must be done with mmu_lock held for write (unless
956 * mmu_lock must be held for write to ensure that a root doesn't become
968 lockdep_assert_held_write(&kvm->mmu_lock);
971 * As above, mmu_lock isn't held when destroying the VM! There can't
1356 lockdep_assert_held_read(&kvm->mmu_lock);
1408 read_unlock(&kvm->mmu_lock);
1410 write_unlock(&kvm->mmu_lock);
1416 read_lock(&kvm->mmu_lock);
1418 write_lock(&kvm->mmu_lock);
1607 lockdep_assert_held_read(&kvm->mmu_lock);
1622 lockdep_assert_held_write(&kvm->mmu_lock);
1732 lockdep_assert_held_read(&kvm->mmu_lock);
1785 lockdep_assert_held_write(&kvm->mmu_lock);
1848 * outside of mmu_lock.