/barrelfish-2018-10-04/lib/barrelfish/ |
H A D | event_mutex.c | 28 em->locked = false; 46 if (!em->locked) { 48 em->locked = true; 69 if (!em->locked) { 71 em->locked = true; 80 assert(em->locked); 96 assert(em->locked); 112 em->locked = false;
|
H A D | thread_sync.c | 165 mutex->locked = 0; 187 if (mutex->locked > 0) { 191 mutex->locked = 1; 217 if (mutex->locked > 0 222 mutex->locked++; 248 if (mutex->locked > 0) { 257 if (mutex->locked > 0) { 261 mutex->locked = 1; 289 assert_disabled(mutex->locked > 0); 291 if(mutex->locked [all...] |
/barrelfish-2018-10-04/usr/tests/octopus/ |
H A D | d2locksimple.c | 29 static bool locked = false; variable 36 if (!locked) { 40 locked = true; 45 locked = false;
|
/barrelfish-2018-10-04/lib/phoenix/ |
H A D | mcs.c | 37 uintptr_t locked; member in struct:mcs_lock_priv 62 priv->locked = 0; 85 assert (priv->locked == 0); 97 /* NOTE: this ordering is important-- if locked after next assignment, 99 set_and_flush(priv->locked, 1); 102 while (atomic_read(&priv->locked)) { asm("":::"memory"); } 127 set_and_flush(priv->next->locked, 0);
|
/barrelfish-2018-10-04/include/barrelfish/ |
H A D | event_mutex.h | 29 bool locked; member in struct:event_mutex
|
H A D | thread_sync.h | 27 volatile int locked; member in struct:thread_mutex 34 { .locked = 0, .queue = NULL, .lock = 0, .holder = NULL }
|
/barrelfish-2018-10-04/include/mdb/ |
H A D | types.h | 27 bool locked:1, in_delete:1; member in struct:mdbnode
|
/barrelfish-2018-10-04/kernel/include/ |
H A D | distcaps.h | 29 if (dcap->mdbnode.locked || dcap->mdbnode.in_delete) {
|
/barrelfish-2018-10-04/usr/monitor/capops/ |
H A D | delete.c | 38 delete_result__rx(errval_t status, struct delete_st *del_st, bool locked) argument 40 DEBUG_CAPOPS("%s: status=%s, locked=%d\n", __FUNCTION__, err_getcode(status), locked); 43 if (locked) { 115 bool locked = true; local 129 locked = false; 143 delete_result__rx(err, del_st, locked); 409 bool locked = false; local 417 // If cap is already locked, just enqueue for retry 419 DEBUG_CAPOPS("%s: from cnode_delete(): cap already locked, queuin [all...] |
H A D | revoke.c | 46 bool locked); 120 bool locked) 125 if (locked) { 118 revoke_result__rx(errval_t result, struct revoke_master_st *st, bool locked) argument
|
/barrelfish-2018-10-04/lib/posixcompat/ |
H A D | pthreads.c | 36 int locked; member in struct:pthread_mutex 228 (*mutex)->locked = 0; 257 (*mutex)->locked++; 275 if((*mutex)->locked == 0) { 280 (*mutex)->locked--; 300 (*mutex)->locked++;
|
/barrelfish-2018-10-04/lib/openssl-1.0.0d/crypto/engine/ |
H A D | eng_int.h | 138 int engine_free_util(ENGINE *e, int locked);
|
H A D | eng_lib.c | 105 int engine_free_util(ENGINE *e, int locked) argument 115 if(locked)
|
/barrelfish-2018-10-04/kernel/ |
H A D | cap_delete.c | 50 if (distcap_is_in_delete(cte) || cte->mdbnode.locked) { 51 // locked or already in process of being deleted 451 // Foreign capabilities are rarely locked, since they can be deleted 452 // immediately. The only time a foreign capability is locked is during 457 // capability is locked on the owner. 458 cte->mdbnode.locked = false; 604 if (cte->mdbnode.locked) { 732 if (cte->mdbnode.locked) { 748 if (cte->mdbnode.locked) {
|
H A D | monitor.c | 242 pred->mdbnode.locked = lock; 248 succ->mdbnode.locked = lock; 264 if (target->mdbnode.locked) {
|
H A D | capabilities.c | 1250 CP_ATTR(locked); 1257 dest->mdbnode.locked = false; 1800 CP_ATTR(locked);
|
/barrelfish-2018-10-04/lib/openssl-1.0.0d/crypto/bio/ |
H A D | b_sock.c | 129 int locked = 0; local 150 locked = 1; 169 if (locked)
|
/barrelfish-2018-10-04/include/vm/ |
H A D | vm_fault.c | 296 int ahead, behind, cluster_offset, error, locked, rv; local 366 if (fs.vp == NULL /* avoid locked vnode leak */ && 647 locked = VOP_ISLOCKED(vp); 649 if (locked != LK_EXCLUSIVE) 650 locked = LK_SHARED; 652 error = vget(vp, locked | LK_CANRECURSE | 658 error = vget(vp, locked | LK_RETRY | 1296 * The source and destination maps must be locked for write.
|
H A D | vm_mmap.c | 758 boolean_t locked; local 849 locked = VM_OBJECT_TRYWLOCK(object); 851 if (!locked) { 1117 * will be memory resident and locked (wired) upon return
|
H A D | uma_core.c | 635 * bucket The free/alloc bucket with items, cpu queue must be locked. 787 * Drain the cached buckets from a zone. Expects a locked zone on entry. 1896 bool locked; local 1925 locked = false; 1928 locked = true; 1931 if (locked) 1944 bool locked; local 1959 locked = false; 1962 locked = true; 1966 if (locked) [all...] |
/barrelfish-2018-10-04/usr/eclipseclp/Contrib/ |
H A D | xml_pp.pl | 148 possible_operator( (locked) ).
|
/barrelfish-2018-10-04/lib/lwip/src/ |
H A D | sys_arch.c | 61 if(lwip_mutex->locked == 0) { 62 printf("mutex NOT locked, called %p, last %p, locked by %p\n", __builtin_return_address(0), last_user, last_locker);
|
/barrelfish-2018-10-04/lib/lwip2/src/ |
H A D | sys_arch.c | 71 if(lwip_mutex->locked == 0) { 72 printf("mutex NOT locked, called %p, last %p, locked by %p\n", __builtin_return_address(0), last_user, last_locker);
|
/barrelfish-2018-10-04/usr/eclipseclp/Kernel/lib/ |
H A D | util.pl | 274 get_module_info(Module, locked, Locked), 275 ( Locked == on -> writeln(".\t% (locked)") ; writeln(.) ),
|
/barrelfish-2018-10-04/usr/drivers/solarflair/sfxge/common/ |
H A D | efx_ev.c | 842 unsigned int locked; local 884 locked = (eep->ee_index == 0) ? 1 : 0; 887 eep->ee_index, &dword, locked);
|