Lines Matching defs:rnh_lock

104  * lock (rnh_lock) and the per-entry lock (rt_lock); the latter is a mutex that
119 * In cases where both rnh_lock and rt_lock must be held, the former must be
121 * that rnh_lock be acquired first before rt_lock, but in case both must be
128 * - Routing table lock (rnh_lock).
147 * - Always written/modified with both rnh_lock and rt_lock held.
149 * - May be read freely with rnh_lock held, else must hold rt_lock
152 * - In the event rnh_lock is not acquired, or is not possible to be
160 * When rnh_lock is held rt_setgate(), rt_setif(), and rtsetifa() are
176 * INSERTION of an entry into the radix tree holds the rnh_lock, checks
180 * LOOKUP of an entry holds the rnh_lock and bumps up the reference count
184 * REMOVAL of an entry from the radix tree holds the rnh_lock, removes the
208 lck_mtx_t *rnh_lock = &rnh_lock_data;
272 /* List of trash route entries protected by rnh_lock */
378 * or deleted. Protected by rnh_lock.
414 * Set the ifscope of the primary interface; caller holds rnh_lock.
426 * Return the ifscope of the primary interface; caller holds rnh_lock.
749 lck_mtx_init(rnh_lock, rnh_lock_grp, rnh_lock_attr);
755 lck_mtx_lock(rnh_lock);
757 lck_mtx_unlock(rnh_lock);
827 lck_mtx_assert(rnh_lock, LCK_MTX_ASSERT_NOTOWNED);
828 lck_mtx_lock(rnh_lock);
830 lck_mtx_unlock(rnh_lock);
836 lck_mtx_assert(rnh_lock, LCK_MTX_ASSERT_NOTOWNED);
837 lck_mtx_lock(rnh_lock);
839 lck_mtx_unlock(rnh_lock);
944 lck_mtx_assert(rnh_lock, LCK_MTX_ASSERT_NOTOWNED);
945 lck_mtx_lock(rnh_lock);
947 lck_mtx_unlock(rnh_lock);
956 lck_mtx_assert(rnh_lock, LCK_MTX_ASSERT_NOTOWNED);
957 lck_mtx_lock(rnh_lock);
959 lck_mtx_unlock(rnh_lock);
990 * trying to acquire the global rnh_lock. If we are called with
991 * rnh_lock held, then we already have exclusive access; otherwise
996 * Note that we check it again below after grabbing rnh_lock,
998 * the race, grabs the rnh_lock first, and bumps up the reference
1006 lck_mtx_lock(rnh_lock);
1022 lck_mtx_assert(rnh_lock, LCK_MTX_ASSERT_OWNED);
1125 lck_mtx_unlock(rnh_lock);
1200 lck_mtx_assert(rnh_lock, LCK_MTX_ASSERT_OWNED);
1241 lck_mtx_assert(rnh_lock, LCK_MTX_ASSERT_NOTOWNED);
1242 lck_mtx_lock(rnh_lock);
1275 * going down recently. Holding rnh_lock here prevents the
1366 lck_mtx_unlock(rnh_lock);
1399 lck_mtx_lock(rnh_lock);
1401 lck_mtx_unlock(rnh_lock);
1434 lck_mtx_assert(rnh_lock, LCK_MTX_ASSERT_OWNED);
1500 * Holding rnh_lock here prevents the possibility of ifa from
1614 lck_mtx_assert(rnh_lock, LCK_MTX_ASSERT_OWNED);
1793 * process below. Since rnh_lock is held at this point, the
1935 * rnh_lock is held here, so rt_key and
1975 * parent's references). rnh_lock is still held, which prevents
2070 lck_mtx_assert(rnh_lock, LCK_MTX_ASSERT_NOTOWNED);
2071 lck_mtx_lock(rnh_lock);
2073 lck_mtx_unlock(rnh_lock);
2083 lck_mtx_assert(rnh_lock, LCK_MTX_ASSERT_NOTOWNED);
2084 lck_mtx_lock(rnh_lock);
2087 lck_mtx_unlock(rnh_lock);
2104 lck_mtx_assert(rnh_lock, LCK_MTX_ASSERT_OWNED);
2111 * rnh_lock here prevents another thread from calling
2145 lck_mtx_assert(rnh_lock, LCK_MTX_ASSERT_OWNED);
2201 * Safe to drop rt_lock and use rt_key, since holding rnh_lock here
2234 lck_mtx_assert(rnh_lock, LCK_MTX_ASSERT_OWNED);
2287 * since holding rnh_lock here prevents another thread
2349 * we're holding rnh_lock now, so rt_gateway and rt_ifp of gwrt
2486 lck_mtx_assert(rnh_lock, LCK_MTX_ASSERT_OWNED);
2643 lck_mtx_assert(rnh_lock, LCK_MTX_ASSERT_OWNED);
2716 * we're holding rnh_lock now, so rt_ifp should not change.
2827 lck_mtx_assert(rnh_lock, LCK_MTX_ASSERT_NOTOWNED);
2828 lck_mtx_lock(rnh_lock);
2830 lck_mtx_unlock(rnh_lock);
2846 * Holding rnh_lock here prevents the possibility of ifa from
2912 lck_mtx_unlock(rnh_lock);
3072 lck_mtx_lock(rnh_lock);
3094 lck_mtx_unlock(rnh_lock);
3410 lck_mtx_lock(rnh_lock);
3431 lck_mtx_unlock(rnh_lock);
3444 lck_mtx_unlock(rnh_lock);
3556 lck_mtx_lock(rnh_lock);
3560 lck_mtx_unlock(rnh_lock);