/freebsd-13-stable/contrib/apr-util/include/ |
H A D | apr_anylock.h | 51 #define APR_ANYLOCK_LOCK(lck) \ 52 (((lck)->type == apr_anylock_none) \ 54 : (((lck)->type == apr_anylock_threadmutex) \ 55 ? apr_thread_mutex_lock((lck)->lock.tm) \ 56 : (((lck)->type == apr_anylock_procmutex) \ 57 ? apr_proc_mutex_lock((lck)->lock.pm) \ 58 : (((lck)->type == apr_anylock_readlock) \ 59 ? apr_thread_rwlock_rdlock((lck)->lock.rw) \ 60 : (((lck)->type == apr_anylock_writelock) \ 61 ? apr_thread_rwlock_wrlock((lck) [all...] |
/freebsd-13-stable/lib/libc/gen/ |
H A D | _spinlock_stub.c | 47 _atomic_lock_stub(volatile long *lck __unused) 54 _spinlock(spinlock_t *lck) argument 57 ((void (*)(spinlock_t *lck))__libc_interposing[INTERPOS_spinlock]) 58 (lck); 64 _spinunlock(spinlock_t *lck) argument 67 ((void (*)(spinlock_t *lck))__libc_interposing[INTERPOS_spinunlock]) 68 (lck); 73 __libc_spinlock_stub(spinlock_t *lck __unused) 78 __libc_spinunlock_stub(spinlock_t *lck __unused)
|
/freebsd-13-stable/contrib/llvm-project/openmp/runtime/src/ |
H A D | kmp_lock.cpp | 72 static kmp_int32 __kmp_get_tas_lock_owner(kmp_tas_lock_t *lck) { argument 73 return KMP_LOCK_STRIP(KMP_ATOMIC_LD_RLX(&lck->lk.poll)) - 1; 76 static inline bool __kmp_is_tas_lock_nestable(kmp_tas_lock_t *lck) { argument 77 return lck->lk.depth_locked != -1; 81 __kmp_acquire_tas_lock_timed_template(kmp_tas_lock_t *lck, kmp_int32 gtid) { argument 85 kmp_uint32 curr = KMP_LOCK_STRIP(lck->lk.poll); 87 __kmp_printf("LOCK CONTENTION: %p\n", lck); 94 if (KMP_ATOMIC_LD_RLX(&lck->lk.poll) == tas_free && 95 __kmp_atomic_compare_store_acq(&lck->lk.poll, tas_free, tas_busy)) { 96 KMP_FSYNC_ACQUIRED(lck); 113 __kmp_acquire_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 119 __kmp_acquire_tas_lock_with_checks(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 132 __kmp_test_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 143 __kmp_test_tas_lock_with_checks(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 153 __kmp_release_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 165 __kmp_release_tas_lock_with_checks(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 183 __kmp_init_tas_lock(kmp_tas_lock_t *lck) argument 187 __kmp_destroy_tas_lock(kmp_tas_lock_t *lck) argument 189 __kmp_destroy_tas_lock_with_checks(kmp_tas_lock_t *lck) argument 203 __kmp_acquire_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 217 __kmp_acquire_nested_tas_lock_with_checks(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 226 __kmp_test_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 242 __kmp_test_nested_tas_lock_with_checks(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 251 __kmp_release_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 262 __kmp_release_nested_tas_lock_with_checks(kmp_tas_lock_t *lck, kmp_int32 gtid) argument 278 __kmp_init_nested_tas_lock(kmp_tas_lock_t *lck) argument 283 __kmp_destroy_nested_tas_lock(kmp_tas_lock_t *lck) argument 288 __kmp_destroy_nested_tas_lock_with_checks(kmp_tas_lock_t *lck) argument 309 __kmp_get_futex_lock_owner(kmp_futex_lock_t *lck) argument 313 __kmp_is_futex_lock_nestable(kmp_futex_lock_t *lck) argument 318 __kmp_acquire_futex_lock_timed_template(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 399 __kmp_acquire_futex_lock(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 405 __kmp_acquire_futex_lock_with_checks(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 418 __kmp_test_futex_lock(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 427 __kmp_test_futex_lock_with_checks(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 437 __kmp_release_futex_lock(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 469 __kmp_release_futex_lock_with_checks(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 487 __kmp_init_futex_lock(kmp_futex_lock_t *lck) argument 491 __kmp_destroy_futex_lock(kmp_futex_lock_t *lck) argument 493 __kmp_destroy_futex_lock_with_checks(kmp_futex_lock_t *lck) argument 507 __kmp_acquire_nested_futex_lock(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 521 __kmp_acquire_nested_futex_lock_with_checks(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 530 __kmp_test_nested_futex_lock(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 546 __kmp_test_nested_futex_lock_with_checks(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 555 __kmp_release_nested_futex_lock(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 566 __kmp_release_nested_futex_lock_with_checks(kmp_futex_lock_t *lck, kmp_int32 gtid) argument 582 __kmp_init_nested_futex_lock(kmp_futex_lock_t *lck) argument 587 __kmp_destroy_nested_futex_lock(kmp_futex_lock_t *lck) argument 592 __kmp_destroy_nested_futex_lock_with_checks(kmp_futex_lock_t *lck) argument 608 __kmp_get_ticket_lock_owner(kmp_ticket_lock_t *lck) argument 614 __kmp_is_ticket_lock_nestable(kmp_ticket_lock_t *lck) argument 625 __kmp_acquire_ticket_lock_timed_template(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 645 __kmp_acquire_ticket_lock(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 651 __kmp_acquire_ticket_lock_with_checks(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 676 __kmp_test_ticket_lock(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 692 __kmp_test_ticket_lock_with_checks(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 716 __kmp_release_ticket_lock(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 731 __kmp_release_ticket_lock_with_checks(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 756 __kmp_init_ticket_lock(kmp_ticket_lock_t *lck) argument 773 __kmp_destroy_ticket_lock(kmp_ticket_lock_t *lck) argument 787 __kmp_destroy_ticket_lock_with_checks(kmp_ticket_lock_t *lck) argument 808 __kmp_acquire_nested_ticket_lock(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 826 __kmp_acquire_nested_ticket_lock_with_checks(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 843 __kmp_test_nested_ticket_lock(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 864 __kmp_test_nested_ticket_lock_with_checks(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 881 __kmp_release_nested_ticket_lock(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 894 __kmp_release_nested_ticket_lock_with_checks(kmp_ticket_lock_t *lck, kmp_int32 gtid) argument 917 __kmp_init_nested_ticket_lock(kmp_ticket_lock_t *lck) argument 924 __kmp_destroy_nested_ticket_lock(kmp_ticket_lock_t *lck) argument 931 __kmp_destroy_nested_ticket_lock_with_checks(kmp_ticket_lock_t *lck) argument 952 __kmp_get_ticket_lock_location(kmp_ticket_lock_t *lck) argument 956 __kmp_set_ticket_lock_location(kmp_ticket_lock_t *lck, const ident_t *loc) argument 961 __kmp_get_ticket_lock_flags(kmp_ticket_lock_t *lck) argument 965 __kmp_set_ticket_lock_flags(kmp_ticket_lock_t *lck, kmp_lock_flags_t flags) argument 1039 __kmp_dump_queuing_lock(kmp_info_t *this_thr, kmp_int32 gtid, kmp_queuing_lock_t *lck, kmp_int32 head_id, kmp_int32 tail_id) argument 1075 __kmp_get_queuing_lock_owner(kmp_queuing_lock_t *lck) argument 1079 __kmp_is_queuing_lock_nestable(kmp_queuing_lock_t *lck) argument 1088 __kmp_acquire_queuing_lock_timed_template(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1284 __kmp_acquire_queuing_lock(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1292 __kmp_acquire_queuing_lock_with_checks(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1311 __kmp_test_queuing_lock(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1344 __kmp_test_queuing_lock_with_checks(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1362 __kmp_release_queuing_lock(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1505 __kmp_release_queuing_lock_with_checks(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1525 __kmp_init_queuing_lock(kmp_queuing_lock_t *lck) argument 1538 __kmp_destroy_queuing_lock(kmp_queuing_lock_t *lck) argument 1549 __kmp_destroy_queuing_lock_with_checks(kmp_queuing_lock_t *lck) argument 1565 __kmp_acquire_nested_queuing_lock(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1583 __kmp_acquire_nested_queuing_lock_with_checks(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1595 __kmp_test_nested_queuing_lock(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1613 __kmp_test_nested_queuing_lock_with_checks(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1625 __kmp_release_nested_queuing_lock(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1639 __kmp_release_nested_queuing_lock_with_checks(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 1658 __kmp_init_nested_queuing_lock(kmp_queuing_lock_t *lck) argument 1663 __kmp_destroy_nested_queuing_lock(kmp_queuing_lock_t *lck) argument 1669 __kmp_destroy_nested_queuing_lock_with_checks(kmp_queuing_lock_t *lck) argument 1685 __kmp_get_queuing_lock_location(kmp_queuing_lock_t *lck) argument 1689 __kmp_set_queuing_lock_location(kmp_queuing_lock_t *lck, const ident_t *loc) argument 1694 __kmp_get_queuing_lock_flags(kmp_queuing_lock_t *lck) argument 1698 __kmp_set_queuing_lock_flags(kmp_queuing_lock_t *lck, kmp_lock_flags_t flags) argument 1827 kmp_adaptive_lock_info_t *lck = &liveLocks; local 1841 __kmp_remember_lock(kmp_adaptive_lock_info_t *lck) argument 1856 __kmp_forget_lock(kmp_adaptive_lock_info_t *lck) argument 1867 __kmp_zero_speculative_stats(kmp_adaptive_lock_info_t *lck) argument 1873 __kmp_add_stats(kmp_adaptive_lock_statistics_t *t, kmp_adaptive_lock_info_t *lck) argument 1885 __kmp_accumulate_speculative_stats(kmp_adaptive_lock_info_t *lck) argument 1914 kmp_adaptive_lock_info_t *lck; local 1971 __kmp_is_unlocked_queuing_lock(kmp_queuing_lock_t *lck) argument 1989 __kmp_update_badness_after_success(kmp_adaptive_lock_t *lck) argument 1996 __kmp_step_badness(kmp_adaptive_lock_t *lck) argument 2006 __kmp_should_speculate(kmp_adaptive_lock_t *lck, kmp_int32 gtid) argument 2016 __kmp_test_adaptive_lock_only(kmp_adaptive_lock_t *lck, kmp_int32 gtid) argument 2064 __kmp_test_adaptive_lock(kmp_adaptive_lock_t *lck, kmp_int32 gtid) argument 2083 __kmp_test_adaptive_lock_with_checks(kmp_adaptive_lock_t *lck, kmp_int32 gtid) argument 2109 __kmp_acquire_adaptive_lock(kmp_adaptive_lock_t *lck, kmp_int32 gtid) argument 2142 __kmp_acquire_adaptive_lock_with_checks(kmp_adaptive_lock_t *lck, kmp_int32 gtid) argument 2157 __kmp_release_adaptive_lock(kmp_adaptive_lock_t *lck, kmp_int32 gtid) argument 2172 __kmp_release_adaptive_lock_with_checks(kmp_adaptive_lock_t *lck, kmp_int32 gtid) argument 2190 __kmp_init_adaptive_lock(kmp_adaptive_lock_t *lck) argument 2203 __kmp_destroy_adaptive_lock(kmp_adaptive_lock_t *lck) argument 2211 __kmp_destroy_adaptive_lock_with_checks(kmp_adaptive_lock_t *lck) argument 2228 __kmp_get_drdpa_lock_owner(kmp_drdpa_lock_t *lck) argument 2232 __kmp_is_drdpa_lock_nestable(kmp_drdpa_lock_t *lck) argument 2237 __kmp_acquire_drdpa_lock_timed_template(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2369 __kmp_acquire_drdpa_lock(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2375 __kmp_acquire_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2394 __kmp_test_drdpa_lock(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2421 __kmp_test_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2439 __kmp_release_drdpa_lock(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2453 __kmp_release_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2474 __kmp_init_drdpa_lock(kmp_drdpa_lock_t *lck) argument 2491 __kmp_destroy_drdpa_lock(kmp_drdpa_lock_t *lck) argument 2511 __kmp_destroy_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck) argument 2527 __kmp_acquire_nested_drdpa_lock(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2544 __kmp_acquire_nested_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2556 __kmp_test_nested_drdpa_lock(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2574 __kmp_test_nested_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2586 __kmp_release_nested_drdpa_lock(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2599 __kmp_release_nested_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck, kmp_int32 gtid) argument 2618 __kmp_init_nested_drdpa_lock(kmp_drdpa_lock_t *lck) argument 2623 __kmp_destroy_nested_drdpa_lock(kmp_drdpa_lock_t *lck) argument 2628 __kmp_destroy_nested_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck) argument 2644 __kmp_get_drdpa_lock_location(kmp_drdpa_lock_t *lck) argument 2648 __kmp_set_drdpa_lock_location(kmp_drdpa_lock_t *lck, const ident_t *loc) argument 2653 __kmp_get_drdpa_lock_flags(kmp_drdpa_lock_t *lck) argument 2657 __kmp_set_drdpa_lock_flags(kmp_drdpa_lock_t *lck, kmp_lock_flags_t flags) argument 2702 __kmp_init_direct_lock(kmp_dyna_lock_t *lck, kmp_dyna_lockseq_t seq) argument 2721 __kmp_destroy_hle_lock(kmp_dyna_lock_t *lck) argument 2723 __kmp_destroy_hle_lock_with_checks(kmp_dyna_lock_t *lck) argument 2727 __kmp_acquire_hle_lock(kmp_dyna_lock_t *lck, kmp_int32 gtid) argument 2741 __kmp_acquire_hle_lock_with_checks(kmp_dyna_lock_t *lck, kmp_int32 gtid) argument 2746 __kmp_release_hle_lock(kmp_dyna_lock_t *lck, kmp_int32 gtid) argument 2754 __kmp_release_hle_lock_with_checks(kmp_dyna_lock_t *lck, kmp_int32 gtid) argument 2759 __kmp_test_hle_lock(kmp_dyna_lock_t *lck, kmp_int32 gtid) argument 2763 __kmp_test_hle_lock_with_checks(kmp_dyna_lock_t *lck, kmp_int32 gtid) argument 2768 __kmp_init_rtm_lock(kmp_queuing_lock_t *lck) argument 2772 __kmp_destroy_rtm_lock(kmp_queuing_lock_t *lck) argument 2776 __kmp_destroy_rtm_lock_with_checks(kmp_queuing_lock_t *lck) argument 2780 __kmp_acquire_rtm_lock(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 2802 __kmp_acquire_rtm_lock_with_checks(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 2807 __kmp_release_rtm_lock(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 2818 __kmp_release_rtm_lock_with_checks(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 2823 __kmp_test_rtm_lock(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 2837 __kmp_test_rtm_lock_with_checks(kmp_queuing_lock_t *lck, kmp_int32 gtid) argument 3032 kmp_indirect_lock_t *lck; local 3090 kmp_indirect_lock_t *lck = NULL; local 3196 __kmp_get_user_lock_owner(kmp_user_lock_p lck, kmp_uint32 seq) argument 3365 __kmp_init_tas_lock_with_checks(kmp_tas_lock_t *lck) argument 3369 __kmp_init_nested_tas_lock_with_checks(kmp_tas_lock_t *lck) argument 3374 __kmp_init_futex_lock_with_checks(kmp_futex_lock_t *lck) argument 3378 __kmp_init_nested_futex_lock_with_checks(kmp_futex_lock_t *lck) argument 3383 __kmp_is_ticket_lock_initialized(kmp_ticket_lock_t *lck) argument 3387 __kmp_init_ticket_lock_with_checks(kmp_ticket_lock_t *lck) argument 3391 __kmp_init_nested_ticket_lock_with_checks(kmp_ticket_lock_t *lck) argument 3395 __kmp_is_queuing_lock_initialized(kmp_queuing_lock_t *lck) argument 3399 __kmp_init_queuing_lock_with_checks(kmp_queuing_lock_t *lck) argument 3404 __kmp_init_nested_queuing_lock_with_checks(kmp_queuing_lock_t *lck) argument 3409 __kmp_init_adaptive_lock_with_checks(kmp_adaptive_lock_t *lck) argument 3414 __kmp_is_drdpa_lock_initialized(kmp_drdpa_lock_t *lck) argument 3418 __kmp_init_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck) argument 3422 __kmp_init_nested_drdpa_lock_with_checks(kmp_drdpa_lock_t *lck) argument 3685 __kmp_lock_table_insert(kmp_user_lock_p lck) argument 3746 kmp_user_lock_p lck; local 3792 __kmp_user_lock_free(void **user_lock, kmp_int32 gtid, kmp_user_lock_p lck) argument 3811 kmp_user_lock_p lck = NULL; local 3879 kmp_user_lock_p lck = local [all...] |
H A D | tsan_annotations.h | 85 #define ANNOTATE_RWLOCK_CREATE(lck) \ 86 AnnotateRWLockCreate(__FILE__, __LINE__, (uptr)lck) 87 #define ANNOTATE_RWLOCK_RELEASED(lck) \ 88 AnnotateRWLockAcquired(__FILE__, __LINE__, (uptr)lck, 1) 89 #define ANNOTATE_RWLOCK_ACQUIRED(lck) \ 90 AnnotateRWLockReleased(__FILE__, __LINE__, (uptr)lck, 1) 104 #define ANNOTATE_RWLOCK_CREATE(lck) 105 #define ANNOTATE_RWLOCK_RELEASED(lck) 106 #define ANNOTATE_RWLOCK_ACQUIRED(lck) 120 #define ANNOTATE_QUEUING_CREATE(lck) [all...] |
H A D | kmp_lock.h | 144 extern int __kmp_acquire_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 145 extern int __kmp_test_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 146 extern int __kmp_release_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 147 extern void __kmp_init_tas_lock(kmp_tas_lock_t *lck); 148 extern void __kmp_destroy_tas_lock(kmp_tas_lock_t *lck); 150 extern int __kmp_acquire_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 151 extern int __kmp_test_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 152 extern int __kmp_release_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 153 extern void __kmp_init_nested_tas_lock(kmp_tas_lock_t *lck); 154 extern void __kmp_destroy_nested_tas_lock(kmp_tas_lock_t *lck); 523 __kmp_acquire_bootstrap_lock(kmp_bootstrap_lock_t *lck) argument 527 __kmp_test_bootstrap_lock(kmp_bootstrap_lock_t *lck) argument 531 __kmp_release_bootstrap_lock(kmp_bootstrap_lock_t *lck) argument 535 __kmp_init_bootstrap_lock(kmp_bootstrap_lock_t *lck) argument 539 __kmp_destroy_bootstrap_lock(kmp_bootstrap_lock_t *lck) argument 555 __kmp_acquire_lock(kmp_lock_t *lck, kmp_int32 gtid) argument 559 __kmp_test_lock(kmp_lock_t *lck, kmp_int32 gtid) argument 563 __kmp_release_lock(kmp_lock_t *lck, kmp_int32 gtid) argument 567 __kmp_init_lock(kmp_lock_t *lck) argument 571 __kmp_destroy_lock(kmp_lock_t *lck) argument 627 __kmp_get_user_lock_owner(kmp_user_lock_p lck) argument 668 __kmp_acquire_user_lock_with_checks(kmp_user_lock_p lck, kmp_int32 gtid) argument 683 __kmp_test_user_lock_with_checks(kmp_user_lock_p lck, kmp_int32 gtid) argument 701 __kmp_test_user_lock_with_checks(kmp_user_lock_p lck, kmp_int32 gtid) argument 711 __kmp_release_user_lock_with_checks(kmp_user_lock_p lck, kmp_int32 gtid) argument 719 __kmp_init_user_lock_with_checks(kmp_user_lock_p lck) argument 728 __kmp_destroy_user_lock(kmp_user_lock_p lck) argument 735 __kmp_destroy_user_lock_with_checks(kmp_user_lock_p lck) argument 780 __kmp_acquire_nested_user_lock_with_checks(kmp_user_lock_p lck, kmp_int32 gtid, int *depth) argument 791 __kmp_test_nested_user_lock_with_checks(kmp_user_lock_p lck, kmp_int32 gtid) argument 820 __kmp_test_nested_user_lock_with_checks(kmp_user_lock_p lck, kmp_int32 gtid) argument 831 __kmp_release_nested_user_lock_with_checks(kmp_user_lock_p lck, kmp_int32 gtid) argument 840 __kmp_init_nested_user_lock_with_checks(kmp_user_lock_p lck) argument 848 __kmp_destroy_nested_user_lock_with_checks(kmp_user_lock_p lck) argument 872 __kmp_get_user_lock_location(kmp_user_lock_p lck) argument 883 __kmp_set_user_lock_location(kmp_user_lock_p lck, const ident_t *loc) argument 895 __kmp_set_user_lock_flags(kmp_user_lock_p lck, kmp_lock_flags_t flags) argument [all...] |
H A D | kmp_csupport.cpp | 860 ompt_wait_id_t lck; local 865 lck = (ompt_wait_id_t)(uintptr_t)&team->t.t_ordered.dt.t_value; 867 th->th.ompt_thread_info.wait_id = lck; 874 ompt_mutex_ordered, omp_lock_hint_none, kmp_mutex_impl_spin, lck, 894 ompt_mutex_ordered, (ompt_wait_id_t)(uintptr_t)lck, codeptr_ra); 949 kmp_indirect_lock_t **lck; local 950 lck = (kmp_indirect_lock_t **)crit; 960 int status = KMP_COMPARE_AND_STORE_PTR(lck, nullptr, ilk); 969 KMP_DEBUG_ASSERT(*lck != NULL); 1093 kmp_user_lock_p lck local 1372 kmp_user_lock_p lck; local 1502 kmp_user_lock_p lck; local 2388 kmp_user_lock_p lck; local 2402 kmp_user_lock_p lck; local 2582 kmp_user_lock_p lck; local 3015 kmp_user_lock_p lck; local 3212 kmp_user_lock_p lck; local 3272 kmp_user_lock_p lck; local 3361 __kmpc_reduce_nowait(ident_t *loc, kmp_int32 global_tid, kmp_int32 num_vars, size_t reduce_size, void *reduce_data, void (*reduce_func)(void *lhs_data, void *rhs_data), kmp_critical_name *lck) argument 3520 __kmpc_end_reduce_nowait(ident_t *loc, kmp_int32 global_tid, kmp_critical_name *lck) argument 3588 __kmpc_reduce(ident_t *loc, kmp_int32 global_tid, kmp_int32 num_vars, size_t reduce_size, void *reduce_data, void (*reduce_func)(void *lhs_data, void *rhs_data), kmp_critical_name *lck) argument 3709 __kmpc_end_reduce(ident_t *loc, kmp_int32 global_tid, kmp_critical_name *lck) argument [all...] |
H A D | kmp_error.cpp | 244 __kmp_check_sync( int gtid, enum cons_type ct, ident_t const * ident, kmp_user_lock_p lck, kmp_uint32 seq ) argument 246 __kmp_check_sync( int gtid, enum cons_type ct, ident_t const * ident, kmp_user_lock_p lck ) 292 if (lck != NULL && 293 __kmp_get_user_lock_owner(lck, seq) == 296 if (lck != NULL && 297 __kmp_get_user_lock_owner(lck) == 303 while (index != 0 && p->stack_data[index].name != lck) { 330 __kmp_push_sync( int gtid, enum cons_type ct, ident_t const * ident, kmp_user_lock_p lck, kmp_uint32 seq ) argument 332 __kmp_push_sync( int gtid, enum cons_type ct, ident_t const * ident, kmp_user_lock_p lck ) 341 __kmp_check_sync(gtid, ct, ident, lck, se [all...] |
H A D | kmp_atomic.h | 361 static inline void __kmp_acquire_atomic_lock(kmp_atomic_lock_t *lck, argument 366 ompt_mutex_atomic, 0, kmp_mutex_impl_queuing, (ompt_wait_id_t)(uintptr_t)lck, 371 __kmp_acquire_queuing_lock(lck, gtid); 376 ompt_mutex_atomic, (ompt_wait_id_t)(uintptr_t)lck, OMPT_GET_RETURN_ADDRESS(0)); 381 static inline int __kmp_test_atomic_lock(kmp_atomic_lock_t *lck, argument 383 return __kmp_test_queuing_lock(lck, gtid); 386 static inline void __kmp_release_atomic_lock(kmp_atomic_lock_t *lck, argument 388 __kmp_release_queuing_lock(lck, gtid); 392 ompt_mutex_atomic, (ompt_wait_id_t)(uintptr_t)lck, OMPT_GET_RETURN_ADDRESS(0)); 397 static inline void __kmp_init_atomic_lock(kmp_atomic_lock_t *lck) { argument 401 __kmp_destroy_atomic_lock(kmp_atomic_lock_t *lck) argument [all...] |
H A D | kmp_dispatch.cpp | 1198 kmp_lock_t *lck = pr->u.p.th_steal_lock; local 1199 KMP_DEBUG_ASSERT(lck != NULL); 1201 __kmp_acquire_lock(lck, gtid); 1205 __kmp_release_lock(lck, gtid); 1244 lck = victim->u.p.th_steal_lock; 1245 KMP_ASSERT(lck != NULL); 1246 __kmp_acquire_lock(lck, gtid); 1250 __kmp_release_lock(lck, gtid); 1265 __kmp_release_lock(lck, gtid); 2080 kmp_lock_t *lck local [all...] |
H A D | kmp.h | 3754 kmp_critical_name *lck); 3756 kmp_critical_name *lck); 3760 kmp_critical_name *lck); 3762 kmp_critical_name *lck); 3769 kmp_critical_name *lck);
|
H A D | kmp_runtime.cpp | 537 static void __kmp_reset_lock(kmp_bootstrap_lock_t *lck) { argument 539 __kmp_init_bootstrap_lock(lck); // make the lock released 8047 kmp_critical_name *lck) { 8049 // Default reduction method: critical construct ( lck != NULL, like in current 8063 KMP_DEBUG_ASSERT(lck); // it would be nice to test ( lck != 0 ) 8157 KMP_ASSERT(lck); // lck should be != 0
|
/freebsd-13-stable/lib/libthr/thread/ |
H A D | thr_spinlock.c | 58 static void init_spinlock(spinlock_t *lck); 66 __thr_spinunlock(spinlock_t *lck) argument 70 _extra = lck->thr_extra; 75 __thr_spinlock(spinlock_t *lck) argument 83 if (lck->thr_extra == NULL) 84 init_spinlock(lck); 85 _extra = lck->thr_extra; 90 init_spinlock(spinlock_t *lck) argument 95 if ((lck->thr_extra == NULL) && (spinlock_count < MAX_SPINLOCKS)) { 96 lck [all...] |
H A D | thr_pspinlock.c | 58 struct pthread_spinlock *lck; local 63 lck = malloc(sizeof(struct pthread_spinlock)); 64 if (lck == NULL) 66 *lock = lck; 68 lck = __thr_pshared_offpage(lock, 1); 69 if (lck == NULL) 75 _thr_umutex_init(&lck->s_lock); 103 struct pthread_spinlock *lck; local 107 lck = *lock == THR_PSHARED_PTR ? __thr_pshared_offpage(lock, 0) : *lock; 108 if (lck 117 struct pthread_spinlock *lck; local 147 struct pthread_spinlock *lck; local [all...] |
H A D | thr_private.h | 603 #define THR_UMUTEX_TRYLOCK(thrd, lck) \ 604 _thr_umutex_trylock((lck), TID(thrd)) 606 #define THR_UMUTEX_LOCK(thrd, lck) \ 607 _thr_umutex_lock((lck), TID(thrd)) 609 #define THR_UMUTEX_TIMEDLOCK(thrd, lck, timo) \ 610 _thr_umutex_timedlock((lck), TID(thrd), (timo)) 612 #define THR_UMUTEX_UNLOCK(thrd, lck) \ 613 _thr_umutex_unlock((lck), TID(thrd)) 615 #define THR_LOCK_ACQUIRE(thrd, lck) \ 618 _thr_umutex_lock(lck, TI [all...] |
/freebsd-13-stable/crypto/openssl/crypto/asn1/ |
H A D | tasn_utl.c | 61 CRYPTO_REF_COUNT *lck; local 71 lck = offset2ptr(*pval, aux->ref_offset); 76 *lck = ret = 1; 84 if (!CRYPTO_UP_REF(lck, &ret, *lock)) 88 if (!CRYPTO_DOWN_REF(lck, &ret, *lock))
|
/freebsd-13-stable/crypto/openssl/ssl/ |
H A D | ssl_sess.c | 21 static int remove_session_lock(SSL_CTX *ctx, SSL_SESSION *c, int lck); 719 static int remove_session_lock(SSL_CTX *ctx, SSL_SESSION *c, int lck) argument 725 if (lck) 734 if (lck)
|
/freebsd-13-stable/sys/dev/cxgbe/ |
H A D | t4_mp_ring.c | 262 ring_can_drain_t can_drain, struct malloc_type *mt, struct mtx *lck, 284 r->cons_lock = lck; 261 mp_ring_alloc(struct mp_ring **pr, int size, void *cookie, ring_drain_t drain, ring_can_drain_t can_drain, struct malloc_type *mt, struct mtx *lck, int flags) argument
|
/freebsd-13-stable/contrib/subversion/subversion/libsvn_wc/ |
H A D | status.c | 677 svn_lock_t *lck = svn_lock_create(result_pool); 678 lck->path = stat->repos_relpath; 679 lck->token = info->lock->token; 680 lck->owner = info->lock->owner; 681 lck->comment = info->lock->comment; 682 lck->creation_date = info->lock->date; 683 stat->lock = lck; 674 svn_lock_t *lck = svn_lock_create(result_pool); local
|
/freebsd-13-stable/sys/kern/ |
H A D | kern_event.c | 90 #define KQ_GLOBAL_LOCK(lck, haslck) do { \ 92 mtx_lock(lck); \ 95 #define KQ_GLOBAL_UNLOCK(lck, haslck) do { \ 97 mtx_unlock(lck); \
|
/freebsd-13-stable/sys/contrib/octeon-sdk/ |
H A D | cvmx-l2c-defs.h | 4073 uint64_t lck : 1; /**< SW attempted to LOCK DOWN the last available set of member in struct:cvmx_l2c_int_stat::cvmx_l2c_int_stat_s 4113 uint64_t lck : 1;
|