Lines Matching refs:lock

40 void	IOLockInitWithState( IOLock * lock, IOLockState state)
43 lck_mtx_lock( lock);
51 void IOLockFree( IOLock * lock)
53 lck_mtx_free( lock, IOLockGroup);
56 lck_mtx_t * IOLockGetMachLock( IOLock * lock)
58 return( (lck_mtx_t *)lock);
61 int IOLockSleep( IOLock * lock, void *event, UInt32 interType)
63 return (int) lck_mtx_sleep(lock, LCK_SLEEP_DEFAULT, (event_t) event, (wait_interrupt_t) interType);
66 int IOLockSleepDeadline( IOLock * lock, void *event,
69 return (int) lck_mtx_sleep_deadline(lock, LCK_SLEEP_DEFAULT, (event_t) event,
73 void IOLockWakeup(IOLock * lock, void *event, bool oneThread)
88 _IORecursiveLock * lock;
93 lock = IONew( _IORecursiveLock, 1 );
94 if( !lock )
97 lock->mutex = lck_mtx_alloc_init( lockGroup, LCK_ATTR_NULL );
98 if( lock->mutex ) {
99 lock->group = lockGroup;
100 lock->thread = 0;
101 lock->count = 0;
103 IODelete( lock, _IORecursiveLock, 1 );
104 lock = 0;
107 return( (IORecursiveLock *) lock );
118 _IORecursiveLock * lock = (_IORecursiveLock *)_lock;
120 lck_mtx_free( lock->mutex, lock->group );
121 IODelete( lock, _IORecursiveLock, 1 );
124 lck_mtx_t * IORecursiveLockGetMachLock( IORecursiveLock * lock )
126 return( lock->mutex );
131 _IORecursiveLock * lock = (_IORecursiveLock *)_lock;
133 if( lock->thread == IOThreadSelf())
134 lock->count++;
136 lck_mtx_lock( lock->mutex );
137 assert( lock->thread == 0 );
138 assert( lock->count == 0 );
139 lock->thread = IOThreadSelf();
140 lock->count = 1;
146 _IORecursiveLock * lock = (_IORecursiveLock *)_lock;
148 if( lock->thread == IOThreadSelf()) {
149 lock->count++;
152 if( lck_mtx_try_lock( lock->mutex )) {
153 assert( lock->thread == 0 );
154 assert( lock->count == 0 );
155 lock->thread = IOThreadSelf();
156 lock->count = 1;
165 _IORecursiveLock * lock = (_IORecursiveLock *)_lock;
167 assert( lock->thread == IOThreadSelf() );
169 if( 0 == (--lock->count)) {
170 lock->thread = 0;
171 lck_mtx_unlock( lock->mutex );
177 _IORecursiveLock * lock = (_IORecursiveLock *)_lock;
179 return( lock->thread == IOThreadSelf());
184 _IORecursiveLock * lock = (_IORecursiveLock *)_lock;
185 UInt32 count = lock->count;
188 assert(lock->thread == IOThreadSelf());
190 lock->count = 0;
191 lock->thread = 0;
192 res = lck_mtx_sleep(lock->mutex, LCK_SLEEP_DEFAULT, (event_t) event, (wait_interrupt_t) interType);
194 // Must re-establish the recursive lock no matter why we woke up
196 assert(lock->thread == 0);
197 assert(lock->count == 0);
198 lock->thread = IOThreadSelf();
199 lock->count = count;
206 _IORecursiveLock * lock = (_IORecursiveLock *)_lock;
207 UInt32 count = lock->count;
210 assert(lock->thread == IOThreadSelf());
212 lock->count = 0;
213 lock->thread = 0;
214 res = lck_mtx_sleep_deadline(lock->mutex, LCK_SLEEP_DEFAULT, (event_t) event,
217 // Must re-establish the recursive lock no matter why we woke up
219 assert(lock->thread == 0);
220 assert(lock->count == 0);
221 lock->thread = IOThreadSelf();
222 lock->count = count;
232 * Complex (read/write) lock operations
240 void IORWLockFree( IORWLock * lock)
242 lck_rw_free( lock, IOLockGroup);
245 lck_rw_t * IORWLockGetMachLock( IORWLock * lock)
247 return( (lck_rw_t *)lock);
260 void IOSimpleLockInit( IOSimpleLock * lock)
262 lck_spin_init( lock, IOLockGroup, LCK_ATTR_NULL);
265 void IOSimpleLockFree( IOSimpleLock * lock )
267 lck_spin_free( lock, IOLockGroup);
270 lck_spin_t * IOSimpleLockGetMachLock( IOSimpleLock * lock)
272 return( (lck_spin_t *)lock);