/freebsd-11-stable/contrib/apr-util/include/ |
H A D | apr_anylock.h | 51 #define APR_ANYLOCK_LOCK(lck) \ 52 (((lck)->type == apr_anylock_none) \ 54 : (((lck)->type == apr_anylock_threadmutex) \ 55 ? apr_thread_mutex_lock((lck)->lock.tm) \ 56 : (((lck)->type == apr_anylock_procmutex) \ 57 ? apr_proc_mutex_lock((lck)->lock.pm) \ 58 : (((lck)->type == apr_anylock_readlock) \ 59 ? apr_thread_rwlock_rdlock((lck)->lock.rw) \ 60 : (((lck)->type == apr_anylock_writelock) \ 61 ? apr_thread_rwlock_wrlock((lck) [all...] |
/freebsd-11-stable/lib/libc/gen/ |
H A D | _spinlock_stub.c | 45 _atomic_lock_stub(volatile long *lck __unused) 52 _spinlock(spinlock_t *lck) argument 55 ((void (*)(spinlock_t *lck))__libc_interposing[INTERPOS_spinlock]) 56 (lck); 62 _spinunlock(spinlock_t *lck) argument 65 ((void (*)(spinlock_t *lck))__libc_interposing[INTERPOS_spinunlock]) 66 (lck); 71 __libc_spinlock_stub(spinlock_t *lck __unused) 76 __libc_spinunlock_stub(spinlock_t *lck __unused)
|
/freebsd-11-stable/contrib/llvm-project/openmp/runtime/src/ |
H A D | kmp_lock.cpp | 72 static kmp_int32 __kmp_get_tas_lock_owner(kmp_tas_lock_t *lck) { argument 73 return KMP_LOCK_STRIP(KMP_ATOMIC_LD_RLX(&lck->lk.poll)) - 1; 76 static inline bool __kmp_is_tas_lock_nestable(kmp_tas_lock_t *lck) { argument 77 return lck->lk.depth_locked != -1; 81 __kmp_acquire_tas_lock_timed_template(kmp_tas_lock_t *lck, kmp_int32 gtid) { argument 85 kmp_uint32 curr = KMP_LOCK_STRIP(lck->lk.poll); 87 __kmp_printf("LOCK CONTENTION: %p\n", lck); 94 if (KMP_ATOMIC_LD_RLX(&lck->lk.poll) == tas_free && 95 __kmp_atomic_compare_store_acq(&lck->lk.poll, tas_free, tas_busy)) { 96 KMP_FSYNC_ACQUIRED(lck); 113 __kmp_acquire_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 119 __kmp_acquire_tas_lock_with_checks(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 132 __kmp_test_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 143 __kmp_test_tas_lock_with_checks(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 153 __kmp_release_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 165 __kmp_release_tas_lock_with_checks(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 183 __kmp_init_tas_lock(kmp_tas_lock_t *lck) argument 187 __kmp_destroy_tas_lock(kmp_tas_lock_t *lck) argument 189 __kmp_destroy_tas_lock_with_checks(kmp_tas_lock_t *lck) argument 203 __kmp_acquire_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 217 __kmp_acquire_nested_tas_lock_with_checks(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 226 __kmp_test_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 242 __kmp_test_nested_tas_lock_with_checks(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 251 __kmp_release_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 262 __kmp_release_nested_tas_lock_with_checks(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 278 __kmp_init_nested_tas_lock(kmp_tas_lock_t *lck) argument 283 __kmp_destroy_nested_tas_lock(kmp_tas_lock_t *lck) argument 288 __kmp_destroy_nested_tas_lock_with_checks(kmp_tas_lock_t *lck) argument 309 __kmp_get_futex_lock_owner(kmp_futex_lock_t *lck) argument 313 __kmp_is_futex_lock_nestable(kmp_futex_lock_t *lck) argument 318 __kmp_acquire_futex_lock_timed_template(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 399 __kmp_acquire_futex_lock(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 405 __kmp_acquire_futex_lock_with_checks(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 418 __kmp_test_futex_lock(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 427 __kmp_test_futex_lock_with_checks(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 437 __kmp_release_futex_lock(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 469 __kmp_release_futex_lock_with_checks(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 487 __kmp_init_futex_lock(kmp_futex_lock_t *lck) argument 491 __kmp_destroy_futex_lock(kmp_futex_lock_t *lck) argument 493 __kmp_destroy_futex_lock_with_checks(kmp_futex_lock_t *lck) argument 507 __kmp_acquire_nested_futex_lock(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 521 __kmp_acquire_nested_futex_lock_with_checks(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 530 __kmp_test_nested_futex_lock(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 546 __kmp_test_nested_futex_lock_with_checks(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 555 __kmp_release_nested_futex_lock(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 566 __kmp_release_nested_futex_lock_with_checks(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 582 __kmp_init_nested_futex_lock(kmp_futex_lock_t *lck) argument 587 __kmp_destroy_nested_futex_lock(kmp_futex_lock_t *lck) argument 592 __kmp_destroy_nested_futex_lock_with_checks(kmp_futex_lock_t *lck) argument 608 __kmp_get_ticket_lock_owner(kmp_ticket_lock_t *lck) argument 614 __kmp_is_ticket_lock_nestable(kmp_ticket_lock_t *lck) argument 625 __kmp_acquire_ticket_lock_timed_template(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 645 __kmp_acquire_ticket_lock(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 651 __kmp_acquire_ticket_lock_with_checks(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 676 __kmp_test_ticket_lock(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 692 __kmp_test_ticket_lock_with_checks(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 716 __kmp_release_ticket_lock(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 731 __kmp_release_ticket_lock_with_checks(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 756 __kmp_init_ticket_lock(kmp_ticket_lock_t *lck) argument 773 __kmp_destroy_ticket_lock(kmp_ticket_lock_t *lck) argument 787 __kmp_destroy_ticket_lock_with_checks(kmp_ticket_lock_t *lck) argument 808 __kmp_acquire_nested_ticket_lock(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 826 __kmp_acquire_nested_ticket_lock_with_checks(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 843 __kmp_test_nested_ticket_lock(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 864 __kmp_test_nested_ticket_lock_with_checks(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 881 __kmp_release_nested_ticket_lock(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 894 __kmp_release_nested_ticket_lock_with_checks(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 917 __kmp_init_nested_ticket_lock(kmp_ticket_lock_t *lck) argument 924 __kmp_destroy_nested_ticket_lock(kmp_ticket_lock_t *lck) argument 931 __kmp_destroy_nested_ticket_lock_with_checks(kmp_ticket_lock_t *lck) argument 952 __kmp_get_ticket_lock_location(kmp_ticket_lock_t *lck) argument 956 __kmp_set_ticket_lock_location(kmp_ticket_lock_t *lck, const ident_t *loc) argument 961 __kmp_get_ticket_lock_flags(kmp_ticket_lock_t *lck) argument 965 __kmp_set_ticket_lock_flags(kmp_ticket_lock_t *lck, kmp_lock_flags_t flags) argument 1039 __kmp_dump_queuing_lock(kmp_info_t *this_thr, kmp_int32 gtid, kmp_queuing_lock_t *lck, kmp_int32 head_id, kmp_int32 tail_id) argument 1075 __kmp_get_queuing_lock_owner(kmp_queuing_lock_t *lck) argument 1079 __kmp_is_queuing_lock_nestable(kmp_queuing_lock_t *lck) argument 1088 __kmp_acquire_queuing_lock_timed_template(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1281 __kmp_acquire_queuing_lock(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1289 __kmp_acquire_queuing_lock_with_checks(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1308 __kmp_test_queuing_lock(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1341 __kmp_test_queuing_lock_with_checks(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1359 __kmp_release_queuing_lock(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1502 __kmp_release_queuing_lock_with_checks(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1522 __kmp_init_queuing_lock(kmp_queuing_lock_t *lck) argument 1535 __kmp_destroy_queuing_lock(kmp_queuing_lock_t *lck) argument 1546 __kmp_destroy_queuing_lock_with_checks(kmp_queuing_lock_t *lck) argument 1562 __kmp_acquire_nested_queuing_lock(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1580 __kmp_acquire_nested_queuing_lock_with_checks(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1592 __kmp_test_nested_queuing_lock(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1610 __kmp_test_nested_queuing_lock_with_checks(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1622 __kmp_release_nested_queuing_lock(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1636 __kmp_release_nested_queuing_lock_with_checks(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1655 __kmp_init_nested_queuing_lock(kmp_queuing_lock_t *lck) argument 1660 __kmp_destroy_nested_queuing_lock(kmp_queuing_lock_t *lck) argument 1666 __kmp_destroy_nested_queuing_lock_with_checks(kmp_queuing_lock_t *lck) argument 1682 __kmp_get_queuing_lock_location(kmp_queuing_lock_t *lck) argument 1686 __kmp_set_queuing_lock_location(kmp_queuing_lock_t *lck, const ident_t *loc) argument 1691 __kmp_get_queuing_lock_flags(kmp_queuing_lock_t *lck) argument 1695 __kmp_set_queuing_lock_flags(kmp_queuing_lock_t *lck, kmp_lock_flags_t flags) argument 1823 kmp_adaptive_lock_info_t *lck = &liveLocks; local 1837 __kmp_remember_lock(kmp_adaptive_lock_info_t *lck) argument 1852 __kmp_forget_lock(kmp_adaptive_lock_info_t *lck) argument 1863 __kmp_zero_speculative_stats(kmp_adaptive_lock_info_t *lck) argument 1869 __kmp_add_stats(kmp_adaptive_lock_statistics_t *t, kmp_adaptive_lock_info_t *lck) argument 1881 __kmp_accumulate_speculative_stats(kmp_adaptive_lock_info_t *lck) argument 1910 kmp_adaptive_lock_info_t *lck; local 1967 __kmp_is_unlocked_queuing_lock(kmp_queuing_lock_t *lck) argument 1985 __kmp_update_badness_after_success(kmp_adaptive_lock_t *lck) argument 1992 __kmp_step_badness(kmp_adaptive_lock_t *lck) argument 2002 __kmp_should_speculate(kmp_adaptive_lock_t *lck, kmp_int32 gtid) argument 2012 __kmp_test_adaptive_lock_only(kmp_adaptive_lock_t *lck, kmp_int32 gtid) argument 2060 __kmp_test_adaptive_lock(kmp_adaptive_lock_t *lck, kmp_int32 gtid) argument 2079 __kmp_test_adaptive_lock_with_checks(kmp_adaptive_lock_t *lck, kmp_int32 gtid) argument 2105 __kmp_acquire_adaptive_lock(kmp_adaptive_lock_t *lck, kmp_int32 gtid) argument 2138 __kmp_acquire_adaptive_lock_with_checks(kmp_adaptive_lock_t *lck, kmp_int32 gtid) argument 2153 __kmp_release_adaptive_lock(kmp_adaptive_lock_t *lck, kmp_int32 gtid) argument 2168 __kmp_release_adaptive_lock_with_checks(kmp_adaptive_lock_t *lck, kmp_int32 gtid) argument 2186 __kmp_init_adaptive_lock(kmp_adaptive_lock_t *lck) argument 2199 __kmp_destroy_adaptive_lock(kmp_adaptive_lock_t *lck) argument 2207 __kmp_destroy_adaptive_lock_with_checks(kmp_adaptive_lock_t *lck) argument 2224 __kmp_get_drdpa_lock_owner(kmp_drdpa_lock_t *lck) argument 2228 __kmp_is_drdpa_lock_nestable(kmp_drdpa_lock_t *lck) argument 2233 __kmp_acquire_drdpa_lock_timed_template(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2365 __kmp_acquire_drdpa_lock(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2371 __kmp_acquire_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2390 __kmp_test_drdpa_lock(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2417 __kmp_test_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2435 __kmp_release_drdpa_lock(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2449 __kmp_release_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2470 __kmp_init_drdpa_lock(kmp_drdpa_lock_t *lck) argument 2487 __kmp_destroy_drdpa_lock(kmp_drdpa_lock_t *lck) argument 2507 __kmp_destroy_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck) argument 2523 __kmp_acquire_nested_drdpa_lock(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2540 __kmp_acquire_nested_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2552 __kmp_test_nested_drdpa_lock(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2570 __kmp_test_nested_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2582 __kmp_release_nested_drdpa_lock(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2595 __kmp_release_nested_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2614 __kmp_init_nested_drdpa_lock(kmp_drdpa_lock_t *lck) argument 2619 __kmp_destroy_nested_drdpa_lock(kmp_drdpa_lock_t *lck) argument 2624 __kmp_destroy_nested_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck) argument 2640 __kmp_get_drdpa_lock_location(kmp_drdpa_lock_t *lck) argument 2644 __kmp_set_drdpa_lock_location(kmp_drdpa_lock_t *lck, const ident_t *loc) argument 2649 __kmp_get_drdpa_lock_flags(kmp_drdpa_lock_t *lck) argument 2653 __kmp_set_drdpa_lock_flags(kmp_drdpa_lock_t *lck, kmp_lock_flags_t flags) argument 2698 __kmp_init_direct_lock(kmp_dyna_lock_t *lck, kmp_dyna_lockseq_t seq) argument 2717 __kmp_destroy_hle_lock(kmp_dyna_lock_t *lck) argument 2719 __kmp_destroy_hle_lock_with_checks(kmp_dyna_lock_t *lck) argument 2723 __kmp_acquire_hle_lock(kmp_dyna_lock_t *lck, kmp_int32 gtid) argument 2737 __kmp_acquire_hle_lock_with_checks(kmp_dyna_lock_t *lck, kmp_int32 gtid) argument 2742 __kmp_release_hle_lock(kmp_dyna_lock_t *lck, kmp_int32 gtid) argument 2750 __kmp_release_hle_lock_with_checks(kmp_dyna_lock_t *lck, kmp_int32 gtid) argument 2755 __kmp_test_hle_lock(kmp_dyna_lock_t *lck, kmp_int32 gtid) argument 2759 __kmp_test_hle_lock_with_checks(kmp_dyna_lock_t *lck, kmp_int32 gtid) argument 2764 __kmp_init_rtm_lock(kmp_queuing_lock_t *lck) argument 2768 __kmp_destroy_rtm_lock(kmp_queuing_lock_t *lck) argument 2772 __kmp_destroy_rtm_lock_with_checks(kmp_queuing_lock_t *lck) argument 2776 __kmp_acquire_rtm_lock(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 2798 __kmp_acquire_rtm_lock_with_checks(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 2803 __kmp_release_rtm_lock(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 2814 __kmp_release_rtm_lock_with_checks(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 2819 __kmp_test_rtm_lock(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 2833 __kmp_test_rtm_lock_with_checks(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 3028 kmp_indirect_lock_t *lck; local 3086 kmp_indirect_lock_t *lck = NULL; local 3192 __kmp_get_user_lock_owner(kmp_user_lock_p lck, kmp_uint32 seq) argument 3361 __kmp_init_tas_lock_with_checks(kmp_tas_lock_t *lck) argument 3365 __kmp_init_nested_tas_lock_with_checks(kmp_tas_lock_t *lck) argument 3370 __kmp_init_futex_lock_with_checks(kmp_futex_lock_t *lck) argument 3374 __kmp_init_nested_futex_lock_with_checks(kmp_futex_lock_t *lck) argument 3379 __kmp_is_ticket_lock_initialized(kmp_ticket_lock_t *lck) argument 3383 __kmp_init_ticket_lock_with_checks(kmp_ticket_lock_t *lck) argument 3387 __kmp_init_nested_ticket_lock_with_checks(kmp_ticket_lock_t *lck) argument 3391 __kmp_is_queuing_lock_initialized(kmp_queuing_lock_t *lck) argument 3395 __kmp_init_queuing_lock_with_checks(kmp_queuing_lock_t *lck) argument 3400 __kmp_init_nested_queuing_lock_with_checks(kmp_queuing_lock_t *lck) argument 3405 __kmp_init_adaptive_lock_with_checks(kmp_adaptive_lock_t *lck) argument 3410 __kmp_is_drdpa_lock_initialized(kmp_drdpa_lock_t *lck) argument 3414 __kmp_init_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck) argument 3418 __kmp_init_nested_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck) argument 3681 __kmp_lock_table_insert(kmp_user_lock_p lck) argument 3742 kmp_user_lock_p lck; local 3788 __kmp_user_lock_free(void **user_lock, kmp_int32 gtid, kmp_user_lock_p lck) argument 3807 kmp_user_lock_p lck = NULL; local 3875 kmp_user_lock_p lck = local [all...] |
H A D | tsan_annotations.h | 85 #define ANNOTATE_RWLOCK_CREATE(lck) \ 86 AnnotateRWLockCreate(__FILE__, __LINE__, (uptr)lck) 87 #define ANNOTATE_RWLOCK_RELEASED(lck) \ 88 AnnotateRWLockAcquired(__FILE__, __LINE__, (uptr)lck, 1) 89 #define ANNOTATE_RWLOCK_ACQUIRED(lck) \ 90 AnnotateRWLockReleased(__FILE__, __LINE__, (uptr)lck, 1) 104 #define ANNOTATE_RWLOCK_CREATE(lck) 105 #define ANNOTATE_RWLOCK_RELEASED(lck) 106 #define ANNOTATE_RWLOCK_ACQUIRED(lck) 120 #define ANNOTATE_QUEUING_CREATE(lck) [all...] |
H A D | kmp_lock.h | 144 extern int __kmp_acquire_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 145 extern int __kmp_test_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 146 extern int __kmp_release_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 147 extern void __kmp_init_tas_lock(kmp_tas_lock_t *lck); 148 extern void __kmp_destroy_tas_lock(kmp_tas_lock_t *lck); 150 extern int __kmp_acquire_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 151 extern int __kmp_test_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 152 extern int __kmp_release_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 153 extern void __kmp_init_nested_tas_lock(kmp_tas_lock_t *lck); 154 extern void __kmp_destroy_nested_tas_lock(kmp_tas_lock_t *lck); 523 __kmp_acquire_bootstrap_lock(kmp_bootstrap_lock_t *lck) argument 527 __kmp_test_bootstrap_lock(kmp_bootstrap_lock_t *lck) argument 531 __kmp_release_bootstrap_lock(kmp_bootstrap_lock_t *lck) argument 535 __kmp_init_bootstrap_lock(kmp_bootstrap_lock_t *lck) argument 539 __kmp_destroy_bootstrap_lock(kmp_bootstrap_lock_t *lck) argument 555 __kmp_acquire_lock(kmp_lock_t *lck, kmp_int32 gtid) argument 559 __kmp_test_lock(kmp_lock_t *lck, kmp_int32 gtid) argument 563 __kmp_release_lock(kmp_lock_t *lck, kmp_int32 gtid) argument 567 __kmp_init_lock(kmp_lock_t *lck) argument 571 __kmp_destroy_lock(kmp_lock_t *lck) argument 627 __kmp_get_user_lock_owner(kmp_user_lock_p lck) argument 668 __kmp_acquire_user_lock_with_checks(kmp_user_lock_p lck, kmp_int32 gtid) argument 683 __kmp_test_user_lock_with_checks(kmp_user_lock_p lck, kmp_int32 gtid) argument 701 __kmp_test_user_lock_with_checks(kmp_user_lock_p lck, kmp_int32 gtid) argument 711 __kmp_release_user_lock_with_checks(kmp_user_lock_p lck, kmp_int32 gtid) argument 719 __kmp_init_user_lock_with_checks(kmp_user_lock_p lck) argument 728 __kmp_destroy_user_lock(kmp_user_lock_p lck) argument 735 __kmp_destroy_user_lock_with_checks(kmp_user_lock_p lck) argument 780 __kmp_acquire_nested_user_lock_with_checks(kmp_user_lock_p lck, kmp_int32 gtid, int *depth) argument 791 __kmp_test_nested_user_lock_with_checks(kmp_user_lock_p lck, kmp_int32 gtid) argument 820 __kmp_test_nested_user_lock_with_checks(kmp_user_lock_p lck, kmp_int32 gtid) argument 831 __kmp_release_nested_user_lock_with_checks(kmp_user_lock_p lck, kmp_int32 gtid) argument 840 __kmp_init_nested_user_lock_with_checks(kmp_user_lock_p lck) argument 848 __kmp_destroy_nested_user_lock_with_checks(kmp_user_lock_p lck) argument 872 __kmp_get_user_lock_location(kmp_user_lock_p lck) argument 883 __kmp_set_user_lock_location(kmp_user_lock_p lck, const ident_t *loc) argument 895 __kmp_set_user_lock_flags(kmp_user_lock_p lck, kmp_lock_flags_t flags) argument [all...] |
H A D | kmp_csupport.cpp | 860 ompt_wait_id_t lck; local 865 lck = (ompt_wait_id_t)(uintptr_t)&team->t.t_ordered.dt.t_value; 867 th->th.ompt_thread_info.wait_id = lck; 874 ompt_mutex_ordered, omp_lock_hint_none, kmp_mutex_impl_spin, lck, 894 ompt_mutex_ordered, (ompt_wait_id_t)(uintptr_t)lck, codeptr_ra); 949 kmp_indirect_lock_t **lck; local 950 lck = (kmp_indirect_lock_t **)crit; 960 int status = KMP_COMPARE_AND_STORE_PTR(lck, nullptr, ilk); 969 KMP_DEBUG_ASSERT(*lck != NULL); 1093 kmp_user_lock_p lck local 1372 kmp_user_lock_p lck; local 1502 kmp_user_lock_p lck; local 2388 kmp_user_lock_p lck; local 2402 kmp_user_lock_p lck; local 2582 kmp_user_lock_p lck; local 3015 kmp_user_lock_p lck; local 3212 kmp_user_lock_p lck; local 3272 kmp_user_lock_p lck; local 3361 __kmpc_reduce_nowait(ident_t *loc, kmp_int32 global_tid, kmp_int32 num_vars, size_t reduce_size, void *reduce_data, void (*reduce_func)(void *lhs_data, void *rhs_data), kmp_critical_name *lck) argument 3520 __kmpc_end_reduce_nowait(ident_t *loc, kmp_int32 global_tid, kmp_critical_name *lck) argument 3588 __kmpc_reduce(ident_t *loc, kmp_int32 global_tid, kmp_int32 num_vars, size_t reduce_size, void *reduce_data, void (*reduce_func)(void *lhs_data, void *rhs_data), kmp_critical_name *lck) argument 3709 __kmpc_end_reduce(ident_t *loc, kmp_int32 global_tid, kmp_critical_name *lck) argument [all...] |
H A D | kmp_error.cpp | 244 __kmp_check_sync( int gtid, enum cons_type ct, ident_t const * ident, kmp_user_lock_p lck, kmp_uint32 seq ) argument 246 __kmp_check_sync( int gtid, enum cons_type ct, ident_t const * ident, kmp_user_lock_p lck ) 292 if (lck != NULL && 293 __kmp_get_user_lock_owner(lck, seq) == 296 if (lck != NULL && 297 __kmp_get_user_lock_owner(lck) == 303 while (index != 0 && p->stack_data[index].name != lck) { 330 __kmp_push_sync( int gtid, enum cons_type ct, ident_t const * ident, kmp_user_lock_p lck, kmp_uint32 seq ) argument 332 __kmp_push_sync( int gtid, enum cons_type ct, ident_t const * ident, kmp_user_lock_p lck ) 341 __kmp_check_sync(gtid, ct, ident, lck, se [all...] |
H A D | kmp_atomic.h | 361 static inline void __kmp_acquire_atomic_lock(kmp_atomic_lock_t *lck, argument 366 ompt_mutex_atomic, 0, kmp_mutex_impl_queuing, (ompt_wait_id_t)(uintptr_t)lck, 371 __kmp_acquire_queuing_lock(lck, gtid); 376 ompt_mutex_atomic, (ompt_wait_id_t)(uintptr_t)lck, OMPT_GET_RETURN_ADDRESS(0)); 381 static inline int __kmp_test_atomic_lock(kmp_atomic_lock_t *lck, argument 383 return __kmp_test_queuing_lock(lck, gtid); 386 static inline void __kmp_release_atomic_lock(kmp_atomic_lock_t *lck, argument 388 __kmp_release_queuing_lock(lck, gtid); 392 ompt_mutex_atomic, (ompt_wait_id_t)(uintptr_t)lck, OMPT_GET_RETURN_ADDRESS(0)); 397 static inline void __kmp_init_atomic_lock(kmp_atomic_lock_t *lck) { argument 401 __kmp_destroy_atomic_lock(kmp_atomic_lock_t *lck) argument [all...] |
H A D | kmp_dispatch.cpp | 1198 kmp_lock_t *lck = th->th.th_dispatch->th_steal_lock; local 1199 KMP_DEBUG_ASSERT(lck != NULL); 1201 __kmp_acquire_lock(lck, gtid); 1205 __kmp_release_lock(lck, gtid); 1243 lck = other_threads[victimIdx]->th.th_dispatch->th_steal_lock; 1244 KMP_ASSERT(lck != NULL); 1245 __kmp_acquire_lock(lck, gtid); 1249 __kmp_release_lock(lck, gtid); 1264 __kmp_release_lock(lck, gtid); 2074 kmp_lock_t *lck local [all...] |
H A D | kmp.h | 3758 kmp_critical_name *lck); 3760 kmp_critical_name *lck); 3764 kmp_critical_name *lck); 3766 kmp_critical_name *lck); 3773 kmp_critical_name *lck);
|
H A D | kmp_runtime.cpp | 537 static void __kmp_reset_lock(kmp_bootstrap_lock_t *lck) { argument 539 __kmp_init_bootstrap_lock(lck); // make the lock released 8047 kmp_critical_name *lck) { 8049 // Default reduction method: critical construct ( lck != NULL, like in current 8063 KMP_DEBUG_ASSERT(lck); // it would be nice to test ( lck != 0 ) 8157 KMP_ASSERT(lck); // lck should be != 0
|
/freebsd-11-stable/lib/libthr/thread/ |
H A D | thr_spinlock.c | 56 static void init_spinlock(spinlock_t *lck); 64 __thr_spinunlock(spinlock_t *lck) argument 68 _extra = lck->thr_extra; 73 __thr_spinlock(spinlock_t *lck) argument 81 if (lck->thr_extra == NULL) 82 init_spinlock(lck); 83 _extra = lck->thr_extra; 88 init_spinlock(spinlock_t *lck) argument 93 if ((lck->thr_extra == NULL) && (spinlock_count < MAX_SPINLOCKS)) { 94 lck [all...] |
H A D | thr_pspinlock.c | 56 struct pthread_spinlock *lck; local 61 lck = malloc(sizeof(struct pthread_spinlock)); 62 if (lck == NULL) 64 *lock = lck; 66 lck = __thr_pshared_offpage(lock, 1); 67 if (lck == NULL) 73 _thr_umutex_init(&lck->s_lock); 101 struct pthread_spinlock *lck; local 105 lck = *lock == THR_PSHARED_PTR ? __thr_pshared_offpage(lock, 0) : *lock; 106 if (lck 115 struct pthread_spinlock *lck; local 145 struct pthread_spinlock *lck; local [all...] |
H A D | thr_private.h | 594 #define THR_UMUTEX_TRYLOCK(thrd, lck) \ 595 _thr_umutex_trylock((lck), TID(thrd)) 597 #define THR_UMUTEX_LOCK(thrd, lck) \ 598 _thr_umutex_lock((lck), TID(thrd)) 600 #define THR_UMUTEX_TIMEDLOCK(thrd, lck, timo) \ 601 _thr_umutex_timedlock((lck), TID(thrd), (timo)) 603 #define THR_UMUTEX_UNLOCK(thrd, lck) \ 604 _thr_umutex_unlock((lck), TID(thrd)) 606 #define THR_LOCK_ACQUIRE(thrd, lck) \ 609 _thr_umutex_lock(lck, TI [all...] |
/freebsd-11-stable/crypto/openssl/crypto/asn1/ |
H A D | tasn_utl.c | 106 int *lck, ret; local 113 lck = offset2ptr(*pval, aux->ref_offset); 115 *lck = 1; 118 ret = CRYPTO_add(lck, op, aux->ref_lock); 120 fprintf(stderr, "%s: Reference Count: %d\n", it->sname, *lck);
|
H A D | asn1t.h | 180 # define ASN1_SEQUENCE_ref(tname, cb, lck) \ 181 static const ASN1_AUX tname##_aux = {NULL, ASN1_AFLG_REFCOUNT, offsetof(tname, references), lck, cb, 0}; \
|
/freebsd-11-stable/crypto/openssl/ssl/ |
H A D | ssl_sess.c | 148 static int remove_session_lock(SSL_CTX *ctx, SSL_SESSION *c, int lck); 823 static int remove_session_lock(SSL_CTX *ctx, SSL_SESSION *c, int lck) argument 829 if (lck) 837 if (lck)
|
/freebsd-11-stable/contrib/subversion/subversion/libsvn_wc/ |
H A D | status.c | 677 svn_lock_t *lck = svn_lock_create(result_pool); 678 lck->path = stat->repos_relpath; 679 lck->token = info->lock->token; 680 lck->owner = info->lock->owner; 681 lck->comment = info->lock->comment; 682 lck->creation_date = info->lock->date; 683 stat->lock = lck; 674 svn_lock_t *lck = svn_lock_create(result_pool); local
|
/freebsd-11-stable/sys/kern/ |
H A D | kern_event.c | 83 #define KQ_GLOBAL_LOCK(lck, haslck) do { \ 85 mtx_lock(lck); \ 88 #define KQ_GLOBAL_UNLOCK(lck, haslck) do { \ 90 mtx_unlock(lck); \
|
/freebsd-11-stable/sys/contrib/octeon-sdk/ |
H A D | cvmx-l2c-defs.h | 4073 uint64_t lck : 1; /**< SW attempted to LOCK DOWN the last available set of member in struct:cvmx_l2c_int_stat::cvmx_l2c_int_stat_s 4113 uint64_t lck : 1;
|