|
reference, declaration → definition
definition → references, declarations, derived classes, virtual overrides
reference to multiple definitions → definitions
unreferenced
|
References
projects/openmp/runtime/src/kmp.h 1149 kmp_uint32 eax;
1150 kmp_uint32 ebx;
1151 kmp_uint32 ecx;
1152 kmp_uint32 edx;
1212 static inline void __kmp_load_mxcsr(const kmp_uint32 *p) { _mm_setcsr(*p); }
1213 static inline void __kmp_store_mxcsr(kmp_uint32 *p) { *p = _mm_getcsr(); }
1565 kmp_uint32 ordered_lower;
1566 kmp_uint32 ordered_upper;
1675 volatile kmp_uint32 iteration;
1676 volatile kmp_uint32 num_done;
1677 volatile kmp_uint32 ordered_iteration;
1697 volatile kmp_uint32 buffer_index;
1699 volatile kmp_uint32 *doacross_flags; // shared array of iteration flags (0/1)
1724 volatile kmp_uint32 *th_doacross_flags; // pointer to shared array of flags
1839 kmp_uint32 *skip_per_level;
1840 kmp_uint32 my_level;
1843 kmp_uint32 depth;
1847 kmp_uint32 nproc;
2185 kmp_uint32 nelements;
2186 kmp_uint32 nconflicts;
2274 kmp_uint32 td_taskwait_counter;
2318 kmp_uint32 td_deque_head; // Head of deque (will wrap)
2319 kmp_uint32 td_deque_tail; // Tail of deque (will wrap)
2362 volatile kmp_uint32
2492 volatile kmp_uint32 th_spin_here; /* thread-local location for spinning */
2507 kmp_uint32 th_task_state_top; // Top element of th_task_state_memo_stack
2508 kmp_uint32 th_task_state_stack_sz; // Size of th_task_state_memo_stack
2509 kmp_uint32 th_reap_state; // Non-zero indicates thread is not
2566 typedef struct kmp_base_data { volatile kmp_uint32 t_value; } kmp_base_data_t;
2646 kmp_uint32 t_mxcsr;
2833 extern kmp_uint32 __kmp_barrier_gather_bb_dflt;
2834 extern kmp_uint32 __kmp_barrier_release_bb_dflt;
2837 extern kmp_uint32 __kmp_barrier_gather_branch_bits[bs_last_barrier];
2838 extern kmp_uint32 __kmp_barrier_release_branch_bits[bs_last_barrier];
2894 extern kmp_uint32 __kmp_yield_init;
2895 extern kmp_uint32 __kmp_yield_next;
2959 extern kmp_uint32 __kmp_init_mxcsr; /* init thread's mxscr */
3002 kmp_uint32 max_soft_retries;
3005 kmp_uint32 max_badness;
3222 enum sched_type schedule, kmp_uint32 lb,
3223 kmp_uint32 ub, kmp_int32 st,
3237 kmp_int32 *p_last, kmp_uint32 *p_lb,
3238 kmp_uint32 *p_ub, kmp_int32 *p_st);
3258 enum sched_type schedule, kmp_uint32 lb,
3259 kmp_uint32 ub, kmp_int32 st,
3276 extern kmp_uint32 __kmp_eq_4(kmp_uint32 value, kmp_uint32 checker);
3276 extern kmp_uint32 __kmp_eq_4(kmp_uint32 value, kmp_uint32 checker);
3276 extern kmp_uint32 __kmp_eq_4(kmp_uint32 value, kmp_uint32 checker);
3277 extern kmp_uint32 __kmp_neq_4(kmp_uint32 value, kmp_uint32 checker);
3277 extern kmp_uint32 __kmp_neq_4(kmp_uint32 value, kmp_uint32 checker);
3277 extern kmp_uint32 __kmp_neq_4(kmp_uint32 value, kmp_uint32 checker);
3278 extern kmp_uint32 __kmp_lt_4(kmp_uint32 value, kmp_uint32 checker);
3278 extern kmp_uint32 __kmp_lt_4(kmp_uint32 value, kmp_uint32 checker);
3278 extern kmp_uint32 __kmp_lt_4(kmp_uint32 value, kmp_uint32 checker);
3279 extern kmp_uint32 __kmp_ge_4(kmp_uint32 value, kmp_uint32 checker);
3279 extern kmp_uint32 __kmp_ge_4(kmp_uint32 value, kmp_uint32 checker);
3279 extern kmp_uint32 __kmp_ge_4(kmp_uint32 value, kmp_uint32 checker);
3280 extern kmp_uint32 __kmp_le_4(kmp_uint32 value, kmp_uint32 checker);
3280 extern kmp_uint32 __kmp_le_4(kmp_uint32 value, kmp_uint32 checker);
3280 extern kmp_uint32 __kmp_le_4(kmp_uint32 value, kmp_uint32 checker);
3281 extern kmp_uint32 __kmp_wait_4(kmp_uint32 volatile *spinner, kmp_uint32 checker,
3281 extern kmp_uint32 __kmp_wait_4(kmp_uint32 volatile *spinner, kmp_uint32 checker,
3281 extern kmp_uint32 __kmp_wait_4(kmp_uint32 volatile *spinner, kmp_uint32 checker,
3282 kmp_uint32 (*pred)(kmp_uint32, kmp_uint32),
3282 kmp_uint32 (*pred)(kmp_uint32, kmp_uint32),
3282 kmp_uint32 (*pred)(kmp_uint32, kmp_uint32),
3284 extern void __kmp_wait_4_ptr(void *spinner, kmp_uint32 checker,
3285 kmp_uint32 (*pred)(void *, kmp_uint32), void *obj);
3285 kmp_uint32 (*pred)(void *, kmp_uint32), void *obj);
3356 extern void __kmp_get_hierarchy(kmp_uint32 nproc, kmp_bstate_t *thr_bar);
projects/openmp/runtime/src/kmp_affinity.cpp 28 void __kmp_get_hierarchy(kmp_uint32 nproc, kmp_bstate_t *thr_bar) {
29 kmp_uint32 depth;
projects/openmp/runtime/src/kmp_affinity.h 648 static const kmp_uint32 maxLeaves = 4;
649 static const kmp_uint32 minBranch = 4;
655 kmp_uint32 maxLevels;
661 kmp_uint32 depth;
662 kmp_uint32 base_num_threads;
672 kmp_uint32 *numPerLevel;
673 kmp_uint32 *skipPerLevel;
721 for (kmp_uint32 i = 0; i < maxLevels;
745 kmp_uint32 branch = minBranch;
750 for (kmp_uint32 d = 0; d < depth - 1; ++d) { // optimize hierarchy width
767 for (kmp_uint32 i = 1; i < depth; ++i)
770 for (kmp_uint32 i = depth; i < maxLevels; ++i)
777 void resize(kmp_uint32 nproc) {
791 kmp_uint32 old_sz = skipPerLevel[depth - 1];
792 kmp_uint32 incs = 0, old_maxLevels = maxLevels;
794 for (kmp_uint32 i = depth; i < maxLevels && nproc > old_sz; ++i) {
809 kmp_uint32 *old_numPerLevel = numPerLevel;
810 kmp_uint32 *old_skipPerLevel = skipPerLevel;
817 for (kmp_uint32 i = 0; i < old_maxLevels;
824 for (kmp_uint32 i = old_maxLevels; i < maxLevels;
835 for (kmp_uint32 i = old_maxLevels; i < maxLevels; ++i)
projects/openmp/runtime/src/kmp_atomic.h 501 void __kmpc_atomic_fixed4u_div(ident_t *id_ref, int gtid, kmp_uint32 *lhs,
502 kmp_uint32 rhs);
511 void __kmpc_atomic_fixed4u_shr(ident_t *id_ref, int gtid, kmp_uint32 *lhs,
512 kmp_uint32 rhs);
722 void __kmpc_atomic_fixed4u_div_rev(ident_t *id_ref, int gtid, kmp_uint32 *lhs,
723 kmp_uint32 rhs);
728 void __kmpc_atomic_fixed4u_shr_rev(ident_t *id_ref, int gtid, kmp_uint32 *lhs,
729 kmp_uint32 rhs);
1155 kmp_uint32 __kmpc_atomic_fixed4u_div_cpt(ident_t *id_ref, int gtid,
1156 kmp_uint32 *lhs, kmp_uint32 rhs,
1156 kmp_uint32 *lhs, kmp_uint32 rhs,
1166 kmp_uint32 __kmpc_atomic_fixed4u_shr_cpt(ident_t *id_ref, int gtid,
1167 kmp_uint32 *lhs, kmp_uint32 rhs,
1167 kmp_uint32 *lhs, kmp_uint32 rhs,
1439 kmp_uint32 __kmpc_atomic_fixed4u_div_cpt_rev(ident_t *id_ref, int gtid,
1440 kmp_uint32 *lhs, kmp_uint32 rhs,
1440 kmp_uint32 *lhs, kmp_uint32 rhs,
1448 kmp_uint32 __kmpc_atomic_fixed4u_shr_cpt_rev(ident_t *id_ref, int gtid,
1449 kmp_uint32 *lhs, kmp_uint32 rhs,
1449 kmp_uint32 *lhs, kmp_uint32 rhs,
projects/openmp/runtime/src/kmp_barrier.cpp 161 kmp_uint32 nproc = this_thr->th.th_team_nproc;
300 kmp_uint32 nproc = this_thr->th.th_team_nproc;
301 kmp_uint32 branch_bits = __kmp_barrier_gather_branch_bits[bt];
302 kmp_uint32 branch_factor = 1 << branch_bits;
303 kmp_uint32 child;
304 kmp_uint32 child_tid;
407 kmp_uint32 nproc;
408 kmp_uint32 branch_bits = __kmp_barrier_release_branch_bits[bt];
409 kmp_uint32 branch_factor = 1 << branch_bits;
410 kmp_uint32 child;
411 kmp_uint32 child_tid;
518 kmp_uint32 num_threads = this_thr->th.th_team_nproc;
519 kmp_uint32 branch_bits = __kmp_barrier_gather_branch_bits[bt];
520 kmp_uint32 branch_factor = 1 << branch_bits;
521 kmp_uint32 offset;
522 kmp_uint32 level;
542 kmp_uint32 child;
543 kmp_uint32 child_tid;
636 kmp_uint32 num_threads;
637 kmp_uint32 branch_bits = __kmp_barrier_release_branch_bits[bt];
638 kmp_uint32 branch_factor = 1 << branch_bits;
639 kmp_uint32 child;
640 kmp_uint32 child_tid;
641 kmp_uint32 offset;
642 kmp_uint32 level;
798 kmp_uint32 nproc, int gtid,
816 kmp_uint32 d = 0;
819 kmp_uint32 rem;
867 kmp_uint32 nproc = this_thr->th.th_team_nproc;
933 for (kmp_uint32 d = 1; d < thr_bar->my_level;
935 kmp_uint32 last = tid + thr_bar->skip_per_level[d + 1],
966 for (kmp_uint32 d = 0; d < thr_bar->my_level;
968 kmp_uint32 last = tid + thr_bar->skip_per_level[d + 1],
1044 kmp_uint32 nproc;
1153 kmp_uint32 last;
1221 kmp_uint32 skip = thr_bar->skip_per_level[d];
projects/openmp/runtime/src/kmp_csupport.cpp 3873 kmp_uint32 *flags;
3989 kmp_uint32 flag;
4082 kmp_uint32 flag;
projects/openmp/runtime/src/kmp_dispatch.cpp 767 kmp_uint32 my_buffer_index;
892 __kmp_wait<kmp_uint32>(&sh->buffer_index, my_buffer_index,
893 __kmp_eq<kmp_uint32> USE_ITT_BUILD_ARG(NULL));
2148 kmp_uint32 team_id;
2149 kmp_uint32 nteams;
2293 enum sched_type schedule, kmp_uint32 lb,
2294 kmp_uint32 ub, kmp_int32 st, kmp_int32 chunk) {
2299 __kmp_dispatch_init<kmp_uint32>(loc, gtid, schedule, lb, ub, st, chunk, true);
2351 kmp_uint32 lb, kmp_uint32 ub, kmp_int32 st,
2351 kmp_uint32 lb, kmp_uint32 ub, kmp_int32 st,
2357 __kmp_dist_get_bounds<kmp_uint32>(loc, gtid, p_last, &lb, &ub, st);
2358 __kmp_dispatch_init<kmp_uint32>(loc, gtid, schedule, lb, ub, st, chunk, true);
2415 kmp_uint32 *p_lb, kmp_uint32 *p_ub,
2415 kmp_uint32 *p_lb, kmp_uint32 *p_ub,
2420 return __kmp_dispatch_next<kmp_uint32>(loc, gtid, p_last, p_lb, p_ub, p_st
2468 __kmp_dispatch_finish<kmp_uint32>(gtid, loc);
2482 __kmp_dispatch_finish<kmp_uint32>(gtid, loc);
2496 kmp_uint32 __kmp_eq_4(kmp_uint32 value, kmp_uint32 checker) {
2496 kmp_uint32 __kmp_eq_4(kmp_uint32 value, kmp_uint32 checker) {
2496 kmp_uint32 __kmp_eq_4(kmp_uint32 value, kmp_uint32 checker) {
2500 kmp_uint32 __kmp_neq_4(kmp_uint32 value, kmp_uint32 checker) {
2500 kmp_uint32 __kmp_neq_4(kmp_uint32 value, kmp_uint32 checker) {
2500 kmp_uint32 __kmp_neq_4(kmp_uint32 value, kmp_uint32 checker) {
2504 kmp_uint32 __kmp_lt_4(kmp_uint32 value, kmp_uint32 checker) {
2504 kmp_uint32 __kmp_lt_4(kmp_uint32 value, kmp_uint32 checker) {
2504 kmp_uint32 __kmp_lt_4(kmp_uint32 value, kmp_uint32 checker) {
2508 kmp_uint32 __kmp_ge_4(kmp_uint32 value, kmp_uint32 checker) {
2508 kmp_uint32 __kmp_ge_4(kmp_uint32 value, kmp_uint32 checker) {
2508 kmp_uint32 __kmp_ge_4(kmp_uint32 value, kmp_uint32 checker) {
2512 kmp_uint32 __kmp_le_4(kmp_uint32 value, kmp_uint32 checker) {
2512 kmp_uint32 __kmp_le_4(kmp_uint32 value, kmp_uint32 checker) {
2512 kmp_uint32 __kmp_le_4(kmp_uint32 value, kmp_uint32 checker) {
2516 kmp_uint32
2517 __kmp_wait_4(volatile kmp_uint32 *spinner, kmp_uint32 checker,
2517 __kmp_wait_4(volatile kmp_uint32 *spinner, kmp_uint32 checker,
2518 kmp_uint32 (*pred)(kmp_uint32, kmp_uint32),
2518 kmp_uint32 (*pred)(kmp_uint32, kmp_uint32),
2518 kmp_uint32 (*pred)(kmp_uint32, kmp_uint32),
2522 volatile kmp_uint32 *spin = spinner;
2523 kmp_uint32 check = checker;
2524 kmp_uint32 spins;
2525 kmp_uint32 (*f)(kmp_uint32, kmp_uint32) = pred;
2525 kmp_uint32 (*f)(kmp_uint32, kmp_uint32) = pred;
2525 kmp_uint32 (*f)(kmp_uint32, kmp_uint32) = pred;
2526 kmp_uint32 r;
2543 void __kmp_wait_4_ptr(void *spinner, kmp_uint32 checker,
2544 kmp_uint32 (*pred)(void *, kmp_uint32),
2544 kmp_uint32 (*pred)(void *, kmp_uint32),
2549 kmp_uint32 check = checker;
2550 kmp_uint32 spins;
2551 kmp_uint32 (*f)(void *, kmp_uint32) = pred;
2551 kmp_uint32 (*f)(void *, kmp_uint32) = pred;
2578 enum sched_type schedule, kmp_uint32 lb,
2579 kmp_uint32 ub, kmp_int32 st, kmp_int32 chunk,
2581 __kmp_dispatch_init<kmp_uint32>(loc, gtid, schedule, lb, ub, st, chunk,
2602 __kmp_dispatch_finish_chunk<kmp_uint32>(gtid, loc);
2610 __kmp_dispatch_finish_chunk<kmp_uint32>(gtid, loc);
projects/openmp/runtime/src/kmp_dispatch.h 137 kmp_uint32 ordered_bumped;
141 kmp_uint32 type_size;
173 volatile kmp_uint32 buffer_index;
175 kmp_uint32 *doacross_flags; // array of iteration flags (0/1)
262 template <typename T> kmp_uint32 __kmp_ge(T value, T checker) {
265 template <typename T> kmp_uint32 __kmp_eq(T value, T checker) {
290 kmp_uint32 (*pred)(UT, UT) USE_ITT_BUILD_ARG(void *obj)) {
294 kmp_uint32 spins;
295 kmp_uint32 (*f)(UT, UT) = pred;
projects/openmp/runtime/src/kmp_error.cpp 244 __kmp_check_sync( int gtid, enum cons_type ct, ident_t const * ident, kmp_user_lock_p lck, kmp_uint32 seq )
330 __kmp_push_sync( int gtid, enum cons_type ct, ident_t const * ident, kmp_user_lock_p lck, kmp_uint32 seq )
projects/openmp/runtime/src/kmp_error.h 35 kmp_user_lock_p name, kmp_uint32);
44 kmp_user_lock_p name, kmp_uint32);
projects/openmp/runtime/src/kmp_global.cpp 76 kmp_uint32 __kmp_barrier_gather_bb_dflt = 2;
78 kmp_uint32 __kmp_barrier_release_bb_dflt = 2;
86 kmp_uint32 __kmp_barrier_gather_branch_bits[bs_last_barrier] = {0};
87 kmp_uint32 __kmp_barrier_release_branch_bits[bs_last_barrier] = {0};
185 kmp_uint32 __kmp_init_mxcsr = 0;
408 kmp_uint32 __kmp_yield_init = KMP_INIT_WAIT;
409 kmp_uint32 __kmp_yield_next = KMP_NEXT_WAIT;
projects/openmp/runtime/src/kmp_itt.inl 528 kmp_uint32 impl = 0;
projects/openmp/runtime/src/kmp_lock.cpp 47 kmp_uint32 x, y;
54 kmp_uint32 z = (x - y);
100 kmp_uint32 spins;
619 static kmp_uint32 __kmp_bakery_check(void *now_serving, kmp_uint32 my_ticket) {
619 static kmp_uint32 __kmp_bakery_check(void *now_serving, kmp_uint32 my_ticket) {
627 kmp_uint32 my_ticket = std::atomic_fetch_add_explicit(
677 kmp_uint32 my_ticket = std::atomic_load_explicit(&lck->lk.next_ticket,
682 kmp_uint32 next_ticket = my_ticket + 1;
717 kmp_uint32 distance = std::atomic_load_explicit(&lck->lk.next_ticket,
1093 volatile kmp_uint32 *spin_here_p;
1993 kmp_uint32 newBadness = (lck->lk.adaptive.badness << 1) | 1;
2004 kmp_uint32 badness = lck->lk.adaptive.badness;
2005 kmp_uint32 attempts = lck->lk.adaptive.acquire_attempts;
2021 kmp_uint32 status = _xbegin();
2251 kmp_uint32 spins;
2290 kmp_uint32 num_polls = TCR_4(lck->lk.num_polls);
2311 kmp_uint32 old_num_polls = num_polls;
2324 kmp_uint32 i;
2684 kmp_uint32 i;
2712 static inline kmp_uint32 swap4(kmp_uint32 volatile *p, kmp_uint32 v) {
2712 static inline kmp_uint32 swap4(kmp_uint32 volatile *p, kmp_uint32 v) {
2712 static inline kmp_uint32 swap4(kmp_uint32 volatile *p, kmp_uint32 v) {
3005 static kmp_uint32 __kmp_indirect_lock_size[KMP_NUM_I_LOCKS] = {0};
3135 kmp_uint32 gtid = __kmp_entry_gtid();
3192 kmp_int32 __kmp_get_user_lock_owner(kmp_user_lock_p lck, kmp_uint32 seq) {
projects/openmp/runtime/src/kmp_lock.h 68 typedef kmp_uint32 kmp_lock_flags_t;
73 typedef kmp_uint32 kmp_lock_index_t;
340 kmp_uint32 volatile badness;
341 kmp_uint32 volatile acquire_attempts;
343 kmp_uint32 max_badness;
344 kmp_uint32 max_soft_retries;
370 volatile kmp_uint32
372 volatile kmp_uint32
452 kmp_uint32 num_polls; // must be power of 2
474 volatile kmp_uint32 owner_id; // (gtid+1) of owning thread, 0 if unlocked
1083 typedef kmp_uint32 kmp_dyna_lock_t;
1248 extern kmp_int32 __kmp_get_user_lock_owner(kmp_user_lock_p, kmp_uint32);
1260 kmp_uint32 step; // current step
1261 kmp_uint32 max_backoff; // upper bound of outer delay loop
1262 kmp_uint32 min_tick; // size of inner delay loop in ticks (machine-dependent)
projects/openmp/runtime/src/kmp_runtime.cpp 1072 kmp_uint32 mxcsr;
1107 kmp_uint32 mxcsr;
2076 kmp_uint32 new_size = 2 * master_th->th.th_task_state_stack_sz;
2078 kmp_uint32 i;
5407 volatile kmp_uint32 *state = &th->th.th_reap_state;
projects/openmp/runtime/src/kmp_sched.cpp 84 kmp_uint32 tid;
85 kmp_uint32 nth;
431 kmp_uint32 tid;
432 kmp_uint32 nth;
433 kmp_uint32 team_id;
434 kmp_uint32 nteams;
673 kmp_uint32 team_id;
674 kmp_uint32 nteams;
812 kmp_uint32 *plower, kmp_uint32 *pupper,
812 kmp_uint32 *plower, kmp_uint32 *pupper,
815 __kmp_for_static_init<kmp_uint32>(loc, gtid, schedtype, plastiter, plower,
896 kmp_uint32 *plower, kmp_uint32 *pupper,
896 kmp_uint32 *plower, kmp_uint32 *pupper,
897 kmp_uint32 *pupperD, kmp_int32 *pstride,
899 __kmp_dist_for_static_init<kmp_uint32>(loc, gtid, schedule, plastiter, plower,
969 kmp_uint32 *p_lb, kmp_uint32 *p_ub,
969 kmp_uint32 *p_lb, kmp_uint32 *p_ub,
973 __kmp_team_static_init<kmp_uint32>(loc, gtid, p_last, p_lb, p_ub, p_st, incr,
projects/openmp/runtime/src/kmp_settings.cpp 4189 kmp_uint32 max_backoff = __kmp_spin_backoff_params.max_backoff;
4190 kmp_uint32 min_tick = __kmp_spin_backoff_params.min_tick;
projects/openmp/runtime/src/kmp_tasking.cpp 2578 kmp_uint32 tail;
3394 kmp_uint32 spins;
3584 std::atomic<kmp_uint32> *spin = RCAST(
4032 kmp_uint32 *lower = RCAST(kmp_uint32 *, task->shareds);
4051 kmp_uint32 *upper = RCAST(kmp_uint32 *, task->shareds) + 1;
projects/openmp/runtime/src/kmp_utility.cpp 146 kmp_uint32 t, data[4];
projects/openmp/runtime/src/kmp_wait_release.h 171 kmp_uint32 spins;
518 typedef kmp_uint32 flag_t;
555 kmp_uint32
571 kmp_info_t *get_waiter(kmp_uint32 i) {
578 kmp_uint32 get_num_waiters() { return num_waiting_threads; }
652 kmp_uint32
668 kmp_info_t *get_waiter(kmp_uint32 i) {
675 kmp_uint32 get_num_waiters() { return num_waiting_threads; }
739 class kmp_flag_32 : public kmp_basic_flag<kmp_uint32> {
741 kmp_flag_32(std::atomic<kmp_uint32> *p) : kmp_basic_flag<kmp_uint32>(p) {}
742 kmp_flag_32(std::atomic<kmp_uint32> *p, kmp_info_t *thr)
744 kmp_flag_32(std::atomic<kmp_uint32> *p, kmp_uint32 c)
744 kmp_flag_32(std::atomic<kmp_uint32> *p, kmp_uint32 c)
813 kmp_uint32 num_waiting_threads;
814 kmp_uint32
832 kmp_flag_oncore(volatile kmp_uint64 *p, kmp_uint32 idx)
835 kmp_flag_oncore(volatile kmp_uint64 *p, kmp_uint64 c, kmp_uint32 idx,
841 kmp_info_t *get_waiter(kmp_uint32 i) {
845 kmp_uint32 get_num_waiters() { return num_waiting_threads; }