• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /freebsd-12-stable/contrib/llvm-project/openmp/runtime/src/

Lines Matching refs:tid

47     enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
57 gtid, team->t.t_id, tid, bt));
69 if (!KMP_MASTER_TID(tid)) {
73 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(0, team),
99 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(i, team),
126 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(i, team),
142 gtid, team->t.t_id, tid, team->t.t_id, &team_bar->b_arrived,
148 gtid, team->t.t_id, tid, bt));
154 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
160 if (KMP_MASTER_TID(tid)) {
171 gtid, team->t.t_id, tid, bt));
201 gtid, team->t.t_id, tid, other_threads[i]->th.th_info.ds.ds_gtid,
247 tid = __kmp_tid_from_gtid(gtid);
254 gtid, team->t.t_id, tid, &thr_bar->b_go, KMP_INIT_BARRIER_STATE));
260 gtid, team->t.t_id, tid, bt));
265 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
268 bt, this_thr, gtid, tid, reduce USE_ITT_BUILD_ARG(itt_sync_obj));
272 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
275 bt, this_thr, gtid, tid, reduce USE_ITT_BUILD_ARG(itt_sync_obj));
279 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
282 bt, this_thr, gtid, tid, propagate_icvs USE_ITT_BUILD_ARG(itt_sync_obj));
286 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
289 bt, this_thr, gtid, tid, propagate_icvs USE_ITT_BUILD_ARG(itt_sync_obj));
295 int tid, void (*reduce)(void *, void *)
310 gtid, team->t.t_id, tid, bt));
322 child_tid = (tid << branch_bits) + 1;
339 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(child_tid, team),
356 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(child_tid, team),
372 if (!KMP_MASTER_TID(tid)) { // Worker threads
373 kmp_int32 parent_tid = (tid - 1) >> branch_bits;
378 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(parent_tid, team),
397 gtid, team->t.t_id, tid, team->t.t_id,
402 gtid, team->t.t_id, tid, bt));
406 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
419 tid)) { // Handle fork barrier workers who aren't part of a team yet
450 tid = __kmp_tid_from_gtid(gtid);
455 team->t.t_id, tid, &thr_bar->b_go, KMP_INIT_BARRIER_STATE));
462 gtid, team->t.t_id, tid, bt));
465 child_tid = (tid << branch_bits) + 1;
496 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(child_tid, team),
509 gtid, team->t.t_id, tid, bt));
515 int tid, void (*reduce)(void *, void *)
531 gtid, team->t.t_id, tid, bt));
549 if (((tid >> level) & (branch_factor - 1)) != 0) {
550 kmp_int32 parent_tid = tid & ~((1 << (level + branch_bits)) - 1);
555 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(parent_tid, team),
572 for (child = 1, child_tid = tid + (1 << level);
587 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(child_tid, team),
604 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(child_tid, team),
618 if (KMP_MASTER_TID(tid)) {
626 gtid, team->t.t_id, tid, team->t.t_id,
631 gtid, team->t.t_id, tid, bt));
637 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
655 if (KMP_MASTER_TID(tid)) { // master
660 gtid, team->t.t_id, tid, bt));
664 &team->t.t_implicit_task_taskdata[tid].td_icvs);
697 tid = __kmp_tid_from_gtid(gtid);
702 gtid, team->t.t_id, tid, &thr_bar->b_go, KMP_INIT_BARRIER_STATE));
711 offset < num_threads && (((tid >> level) & (branch_factor - 1)) == 0);
729 child_tid = tid + (child << level);
732 if (((tid >> level) & (branch_factor - 1)) != 0)
737 for (child = 1, child_tid = tid + (1 << level);
768 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(child_tid, team),
780 !KMP_MASTER_TID(tid)) { // copy ICVs locally to final dest
781 __kmp_init_implicit_task(team->t.t_ident, team->t.t_threads[tid], team, tid,
783 copy_icvs(&team->t.t_implicit_task_taskdata[tid].td_icvs,
790 gtid, team->t.t_id, tid, bt));
806 int tid, kmp_team_t *team) {
811 bool tid_changed = tid != thr_bar->old_tid;
822 tid)) { // if not master, find parent thread in hierarchy
831 } else if ((rem = tid % thr_bar->skip_per_level[d + 1]) !=
834 thr_bar->parent_tid = tid - rem;
841 thr_bar->offset = 7 - (tid - thr_bar->parent_tid - 1);
842 thr_bar->old_tid = tid;
859 if (thr_bar->leaf_kids && (kmp_uint32)tid + thr_bar->leaf_kids + 1 > nproc)
860 thr_bar->leaf_kids = nproc - tid - 1;
869 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
890 gtid, team->t.t_id, tid, bt));
900 (void)__kmp_init_hierarchical_barrier_thread(bt, thr_bar, nproc, gtid, tid,
912 KMP_MASTER_TID(tid)
917 gtid, team->t.t_id, tid));
924 for (child_tid = tid + 1; child_tid <= tid + thr_bar->leaf_kids;
928 gtid, team->t.t_id, tid,
945 kmp_uint32 last = tid + thr_bar->skip_per_level[d + 1],
949 for (child_tid = tid + skip; child_tid < (int)last; child_tid += skip) {
955 gtid, team->t.t_id, tid,
964 gtid, team->t.t_id, tid,
978 kmp_uint32 last = tid + thr_bar->skip_per_level[d + 1],
982 for (child_tid = tid + skip; child_tid < (int)last; child_tid += skip) {
988 gtid, team->t.t_id, tid,
997 gtid, team->t.t_id, tid,
1012 if (!KMP_MASTER_TID(tid)) { // worker threads release parent in hierarchy
1015 gtid, team->t.t_id, tid,
1039 gtid, team->t.t_id, tid, team->t.t_id,
1045 gtid, team->t.t_id, tid, bt));
1049 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,
1057 if (KMP_MASTER_TID(tid)) {
1062 gtid, team->t.t_id, tid, bt));
1099 tid = __kmp_tid_from_gtid(gtid);
1104 gtid, team->t.t_id, tid, &thr_bar->b_go, KMP_INIT_BARRIER_STATE));
1128 tid, team);
1135 __kmp_init_implicit_task(team->t.t_ident, team->t.t_threads[tid], team, tid,
1138 tid)) { // master already has copy in final destination; copy
1140 &team->t.t_implicit_task_taskdata[tid].td_icvs);
1146 copy_icvs(&team->t.t_implicit_task_taskdata[tid].td_icvs,
1154 copy_icvs(&team->t.t_implicit_task_taskdata[tid].td_icvs,
1166 if (KMP_MASTER_TID(tid)) { // do a flat release
1182 gtid, team->t.t_id, tid,
1203 last = tid + thr_bar->skip_per_level[1];
1206 for (child_tid = tid + 1 + old_leaf_kids; child_tid < (int)last;
1214 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(child_tid, team),
1230 last = tid + thr_bar->skip_per_level[d + 1];
1234 for (child_tid = tid + skip; child_tid < (int)last; child_tid += skip) {
1239 gtid, team->t.t_id, tid,
1251 if (propagate_icvs && !KMP_MASTER_TID(tid))
1253 copy_icvs(&team->t.t_implicit_task_taskdata[tid].td_icvs,
1259 gtid, team->t.t_id, tid, bt));
1297 int tid = __kmp_tid_from_gtid(gtid);
1361 team->t.t_implicit_task_taskdata[tid].td_icvs.bt_intervals;
1363 team->t.t_implicit_task_taskdata[tid].td_icvs.bt_set;
1365 this_thr->th.th_team_bt_intervals = KMP_BLOCKTIME_INTERVAL(team, tid);
1375 if (KMP_MASTER_TID(tid)) { // Master counter is stored in team structure.
1386 if (KMP_MASTER_TID(tid) && __kmp_tasking_mode != tskm_immediate_exec)
1392 bt, this_thr, gtid, tid, reduce USE_ITT_BUILD_ARG(itt_sync_obj));
1398 __kmp_hyper_barrier_gather(bt, this_thr, gtid, tid,
1404 bt, this_thr, gtid, tid, reduce USE_ITT_BUILD_ARG(itt_sync_obj));
1410 __kmp_tree_barrier_gather(bt, this_thr, gtid, tid,
1415 __kmp_linear_barrier_gather(bt, this_thr, gtid, tid,
1423 if (KMP_MASTER_TID(tid)) {
1504 bt, this_thr, gtid, tid, FALSE USE_ITT_BUILD_ARG(itt_sync_obj));
1509 __kmp_hyper_barrier_release(bt, this_thr, gtid, tid,
1515 bt, this_thr, gtid, tid, FALSE USE_ITT_BUILD_ARG(itt_sync_obj));
1520 __kmp_tree_barrier_release(bt, this_thr, gtid, tid,
1525 __kmp_linear_barrier_release(bt, this_thr, gtid, tid,
1609 int tid = __kmp_tid_from_gtid(gtid);
1611 if (KMP_MASTER_TID(tid)) {
1629 int tid = __kmp_tid_from_gtid(gtid);
1639 __kmp_hyper_barrier_release(bt, this_thr, gtid, tid,
1644 __kmp_hierarchical_barrier_release(bt, this_thr, gtid, tid,
1650 __kmp_tree_barrier_release(bt, this_thr, gtid, tid,
1655 __kmp_linear_barrier_release(bt, this_thr, gtid, tid,
1674 int tid;
1692 tid = __kmp_tid_from_gtid(gtid);
1709 KMP_DEBUG_ASSERT(this_thr == team->t.t_threads[tid]);
1711 gtid, team_id, tid));
1747 team_id, tid));
1769 team->t.t_implicit_task_taskdata[tid].td_icvs.bt_intervals;
1771 team->t.t_implicit_task_taskdata[tid].td_icvs.bt_set;
1773 this_thr->th.th_team_bt_intervals = KMP_BLOCKTIME_INTERVAL(team, tid);
1785 __kmp_hyper_barrier_gather(bs_forkjoin_barrier, this_thr, gtid, tid,
1790 __kmp_hierarchical_barrier_gather(bs_forkjoin_barrier, this_thr, gtid, tid,
1796 __kmp_tree_barrier_gather(bs_forkjoin_barrier, this_thr, gtid, tid,
1801 __kmp_linear_barrier_gather(bs_forkjoin_barrier, this_thr, gtid, tid,
1810 if (KMP_MASTER_TID(tid)) {
1886 if (KMP_MASTER_TID(tid)) {
1890 gtid, team_id, tid, nproc));
1897 ("__kmp_join_barrier: T#%d(%d:%d) leaving\n", gtid, team_id, tid));
1904 void __kmp_fork_barrier(int gtid, int tid) {
1908 kmp_team_t *team = (tid == 0) ? this_thr->th.th_team : NULL;
1916 (team != NULL) ? team->t.t_id : -1, tid));
1919 if (KMP_MASTER_TID(tid)) {
1963 team->t.t_implicit_task_taskdata[tid].td_icvs.bt_intervals;
1965 team->t.t_implicit_task_taskdata[tid].td_icvs.bt_set;
1967 this_thr->th.th_team_bt_intervals = KMP_BLOCKTIME_INTERVAL(team, tid);
1975 __kmp_hyper_barrier_release(bs_forkjoin_barrier, this_thr, gtid, tid,
1980 __kmp_hierarchical_barrier_release(bs_forkjoin_barrier, this_thr, gtid, tid,
1986 __kmp_tree_barrier_release(bs_forkjoin_barrier, this_thr, gtid, tid,
1991 __kmp_linear_barrier_release(bs_forkjoin_barrier, this_thr, gtid, tid,
2034 if (!KMP_MASTER_TID(tid)) {
2051 tid = __kmp_tid_from_gtid(gtid);
2062 if (!KMP_MASTER_TID(tid)) { // master thread already has ICVs
2064 // task for this tid.
2066 ("__kmp_fork_barrier: T#%d(%d) is PULLing ICVs\n", gtid, tid));
2067 __kmp_init_implicit_task(team->t.t_ident, team->t.t_threads[tid], team,
2068 tid, FALSE);
2069 copy_icvs(&team->t.t_implicit_task_taskdata[tid].td_icvs,
2111 if (!KMP_MASTER_TID(tid))
2116 if (!KMP_MASTER_TID(tid)) {
2125 team->t.t_id, tid));