Lines Matching defs:sched_lock
120 struct mtx sched_lock;
689 thread0.td_lock = &sched_lock;
691 mtx_init(&sched_lock, "sched lock", NULL, MTX_SPIN | MTX_RECURSE);
797 childtd->td_lock = &sched_lock;
977 if (td->td_lock != &sched_lock) {
978 mtx_lock_spin(&sched_lock);
1028 MPASS(newtd->td_lock == &sched_lock);
1040 lock_profile_release_lock(&sched_lock.lock_object);
1052 lock_profile_obtain_lock_success(&sched_lock.lock_object,
1083 sched_lock.mtx_lock = (uintptr_t)td;
1085 MPASS(td->td_lock == &sched_lock);
1115 mtx_assert(&sched_lock, MA_OWNED);
1234 mtx_assert(&sched_lock, MA_OWNED);
1287 if (td->td_lock != &sched_lock) {
1288 mtx_lock_spin(&sched_lock);
1289 thread_lock_set(td, &sched_lock);
1380 if (td->td_lock != &sched_lock) {
1381 mtx_lock_spin(&sched_lock);
1382 thread_lock_set(td, &sched_lock);
1418 mtx_assert(&sched_lock, MA_OWNED);
1444 mtx_assert(&sched_lock, MA_OWNED);
1646 mtx_lock_spin(&sched_lock);
1648 mtx_unlock_spin(&sched_lock);
1661 * spin lock (sched_lock) held in fork_trampoline(). Since we've
1668 mtx_lock_spin(&sched_lock);
1673 lock_profile_release_lock(&sched_lock.lock_object);
1674 MPASS(td->td_lock == &sched_lock);
1676 mtx_assert(&sched_lock, MA_OWNED);
1687 * non-nested critical section with sched_lock held but not recursed.
1690 sched_lock.mtx_lock = (uintptr_t)td;
1691 lock_profile_obtain_lock_success(&sched_lock.lock_object,