reference, declaration → definition definition → references, declarations, derived classes, virtual overrides reference to multiple definitions → definitions unreferenced |
1438 master_th = __kmp_threads[gtid]; // AC: potentially unsafe, not in sync with 1440 parent_team = master_th->th.th_team; 1441 master_tid = master_th->th.th_info.ds.ds_tid; 1442 master_this_cons = master_th->th.th_local.this_construct; 1443 root = master_th->th.th_root; 1445 master_set_numthreads = master_th->th.th_set_nproc; 1466 teams_level = master_th->th.th_teams_level; 1468 p_hot_teams = &master_th->th.th_hot_teams; 1492 master_th->th.ompt_thread_info.state = ompt_state_overhead; 1496 master_th->th.th_ident = loc; 1498 if (master_th->th.th_teams_microtask && ap && 1515 if (parent_team == master_th->th.th_serial_team) { 1527 __ompt_lw_taskteam_init(&lw_taskteam, master_th, gtid, 1531 __ompt_lw_taskteam_link(&lw_taskteam, master_th, 0); 1535 implicit_task_data = OMPT_CUR_TASK_DATA(master_th); 1537 OMPT_CUR_TASK_INFO(master_th) 1540 ompt_scope_begin, OMPT_CUR_TEAM_DATA(master_th), 1542 OMPT_CUR_TASK_INFO(master_th)->thread_num, ompt_task_implicit); 1546 master_th->th.ompt_thread_info.state = ompt_state_work_parallel; 1569 OMPT_CUR_TASK_INFO(master_th)->frame.exit_frame = ompt_data_none; 1573 OMPT_CUR_TASK_INFO(master_th)->thread_num, ompt_task_implicit); 1575 ompt_parallel_data = *OMPT_CUR_TEAM_DATA(master_th); 1576 __ompt_lw_taskteam_unlink(master_th); 1579 &ompt_parallel_data, OMPT_CUR_TASK_DATA(master_th), 1583 master_th->th.ompt_thread_info.state = ompt_state_overhead; 1594 parent_team->t.t_def_allocator = master_th->th.th_def_allocator; // save 1599 __ompt_lw_taskteam_init(&lw_taskteam, master_th, gtid, 1601 __ompt_lw_taskteam_link(&lw_taskteam, master_th, 1, true); 1607 if (master_set_numthreads < master_th->th.th_teams_size.nth) { 1616 master_th->th.th_set_nproc = 0; 1660 master_th->th.th_current_task->td_icvs.max_active_levels) { 1676 if ((get__max_active_levels(master_th) == 1 && 1705 master_th->th.th_set_nproc = 0; 1725 master_th->th.th_serial_team->t.t_ident = loc; 1728 master_th->th.th_serial_team->t.t_level--; 1739 __ompt_lw_taskteam_init(&lw_taskteam, master_th, gtid, 1742 __ompt_lw_taskteam_link(&lw_taskteam, master_th, 0); 1745 task_info = OMPT_CUR_TASK_INFO(master_th); 1748 OMPT_CUR_TASK_INFO(master_th) 1751 ompt_scope_begin, OMPT_CUR_TEAM_DATA(master_th), 1753 OMPT_CUR_TASK_INFO(master_th)->thread_num, 1758 master_th->th.ompt_thread_info.state = ompt_state_work_parallel; 1782 OMPT_CUR_TASK_INFO(master_th)->thread_num, 1785 ompt_parallel_data = *OMPT_CUR_TEAM_DATA(master_th); 1786 __ompt_lw_taskteam_unlink(master_th); 1793 master_th->th.ompt_thread_info.state = ompt_state_overhead; 1799 team = master_th->th.th_team; 1825 ompt_task_info_t *task_info = OMPT_CUR_TASK_INFO(master_th); 1829 OMPT_CUR_TASK_INFO(master_th)->thread_num, ompt_task_initial); 1837 master_th->th.ompt_thread_info.state = ompt_state_overhead; 1859 __ompt_lw_taskteam_init(&lw_taskteam, master_th, gtid, 1861 __ompt_lw_taskteam_link(&lw_taskteam, master_th, 0); 1863 task_info = OMPT_CUR_TASK_INFO(master_th); 1867 implicit_task_data = OMPT_CUR_TASK_DATA(master_th); 1870 ompt_scope_begin, OMPT_CUR_TEAM_DATA(master_th), 1873 OMPT_CUR_TASK_INFO(master_th) 1878 master_th->th.ompt_thread_info.state = ompt_state_work_parallel; 1901 OMPT_CUR_TASK_INFO(master_th)->thread_num, 1905 ompt_parallel_data = *OMPT_CUR_TEAM_DATA(master_th); 1906 __ompt_lw_taskteam_unlink(master_th); 1913 master_th->th.ompt_thread_info.state = ompt_state_overhead; 1920 __ompt_lw_taskteam_init(&lwt, master_th, gtid, &ompt_parallel_data, 1924 __ompt_lw_taskteam_link(&lwt, master_th, 1); 1951 master_th->th.th_current_task->td_flags.executing = 0; 1953 if (!master_th->th.th_teams_microtask || level > teams_level) { 1959 int nthreads_icv = master_th->th.th_current_task->td_icvs.nproc; 1968 kmp_proc_bind_t proc_bind = master_th->th.th_set_proc_bind; 1971 if (master_th->th.th_current_task->td_icvs.proc_bind == proc_bind_false) { 1977 proc_bind = master_th->th.th_current_task->td_icvs.proc_bind; 1985 master_th->th.th_current_task->td_icvs.proc_bind)) { 1991 master_th->th.th_set_proc_bind = proc_bind_default; 1995 copy_icvs(&new_icvs, &master_th->th.th_current_task->td_icvs); 2011 argc USE_NESTED_HOT_ARG(master_th)); 2020 &master_th->th.th_current_task->td_icvs, 2021 argc USE_NESTED_HOT_ARG(master_th)); 2038 if (!master_th->th.th_teams_microtask || level > teams_level) { 2055 KMP_CHECK_UPDATE(team->t.t_def_allocator, master_th->th.th_def_allocator); 2071 if (active_level || master_th->th.th_task_team) { 2074 if (master_th->th.th_task_state_top >= 2075 master_th->th.th_task_state_stack_sz) { // increase size 2076 kmp_uint32 new_size = 2 * master_th->th.th_task_state_stack_sz; 2080 for (i = 0; i < master_th->th.th_task_state_stack_sz; ++i) { 2081 new_stack[i] = master_th->th.th_task_state_memo_stack[i]; 2083 for (i = master_th->th.th_task_state_stack_sz; i < new_size; 2087 old_stack = master_th->th.th_task_state_memo_stack; 2088 master_th->th.th_task_state_memo_stack = new_stack; 2089 master_th->th.th_task_state_stack_sz = new_size; 2093 master_th->th 2094 .th_task_state_memo_stack[master_th->th.th_task_state_top] = 2095 master_th->th.th_task_state; 2096 master_th->th.th_task_state_top++; 2098 if (master_th->th.th_hot_teams && 2100 team == master_th->th.th_hot_teams[active_level].hot_team) { 2102 master_th->th.th_task_state = 2103 master_th->th 2104 .th_task_state_memo_stack[master_th->th.th_task_state_top]; 2107 master_th->th.th_task_state = 0; 2154 __kmp_fork_team_threads(root, team, master_th, gtid); 2156 &master_th->th.th_current_task->td_icvs, loc); 2159 master_th->th.ompt_thread_info.state = ompt_state_work_parallel; 2166 && !master_th->th.th_teams_microtask) { // not in teams construct 2175 master_th->th.th_frame_time = tmp_time; 2250 master_th->th.ompt_thread_info.state = ompt_state_overhead;