|
reference, declaration → definition
definition → references, declarations, derived classes, virtual overrides
reference to multiple definitions → definitions
unreferenced
|
References
projects/openmp/runtime/src/kmp.h 945 kmp_uint64 pool_size;
946 kmp_uint64 pool_used;
1043 extern kmp_uint64 __kmp_ticks_per_msec;
1168 kmp_uint64 frequency; // Nominal CPU frequency in Hz.
1600 kmp_uint64 ordered_lower;
1601 kmp_uint64 ordered_upper;
1685 volatile kmp_uint64 iteration;
1686 volatile kmp_uint64 num_done;
1687 volatile kmp_uint64 ordered_iteration;
1836 volatile kmp_uint64 b_go; // STATE => task should proceed (hierarchical)
1837 KMP_ALIGN_CACHE volatile kmp_uint64
1846 kmp_uint64 leaf_state;
1874 kmp_uint64 b_arrived; /* STATE => task reached synch point. */
2070 extern kmp_uint64 __kmp_taskloop_min_tasks;
2446 kmp_uint64 th_team_bt_intervals;
2472 kmp_uint64 th_bar_arrive_time; /* arrival to barrier timestamp */
2473 kmp_uint64 th_bar_min_time; /* minimum arrival time at the barrier */
2474 kmp_uint64 th_frame_time; /* frame timestamp */
2626 kmp_uint64 t_region_time; // region begin timestamp
3229 enum sched_type schedule, kmp_uint64 lb,
3230 kmp_uint64 ub, kmp_int64 st,
3243 kmp_int32 *p_last, kmp_uint64 *p_lb,
3244 kmp_uint64 *p_ub, kmp_int64 *p_st);
3266 enum sched_type schedule, kmp_uint64 lb,
3267 kmp_uint64 ub, kmp_int64 st,
3572 extern kmp_uint64 __kmp_hardware_timestamp(void);
3716 kmp_int32 if_val, kmp_uint64 *lb,
3717 kmp_uint64 *ub, kmp_int64 st, kmp_int32 nogroup,
3718 kmp_int32 sched, kmp_uint64 grainsize,
3785 KMP_EXPORT kmp_uint64 __kmpc_get_taskid();
3786 KMP_EXPORT kmp_uint64 __kmpc_get_parent_taskid();
projects/openmp/runtime/src/kmp_alloc.cpp 1475 kmp_uint64 used =
1531 kmp_uint64 used =
1612 kmp_uint64 used =
1621 kmp_uint64 used =
projects/openmp/runtime/src/kmp_atomic.h 520 void __kmpc_atomic_fixed8u_div(ident_t *id_ref, int gtid, kmp_uint64 *lhs,
521 kmp_uint64 rhs);
530 void __kmpc_atomic_fixed8u_shr(ident_t *id_ref, int gtid, kmp_uint64 *lhs,
531 kmp_uint64 rhs);
734 void __kmpc_atomic_fixed8u_div_rev(ident_t *id_ref, int gtid, kmp_uint64 *lhs,
735 kmp_uint64 rhs);
740 void __kmpc_atomic_fixed8u_shr_rev(ident_t *id_ref, int gtid, kmp_uint64 *lhs,
741 kmp_uint64 rhs);
1177 kmp_uint64 __kmpc_atomic_fixed8u_div_cpt(ident_t *id_ref, int gtid,
1178 kmp_uint64 *lhs, kmp_uint64 rhs,
1178 kmp_uint64 *lhs, kmp_uint64 rhs,
1188 kmp_uint64 __kmpc_atomic_fixed8u_shr_cpt(ident_t *id_ref, int gtid,
1189 kmp_uint64 *lhs, kmp_uint64 rhs,
1189 kmp_uint64 *lhs, kmp_uint64 rhs,
1457 kmp_uint64 __kmpc_atomic_fixed8u_div_cpt_rev(ident_t *id_ref, int gtid,
1458 kmp_uint64 *lhs, kmp_uint64 rhs,
1458 kmp_uint64 *lhs, kmp_uint64 rhs,
1466 kmp_uint64 __kmpc_atomic_fixed8u_shr_cpt_rev(ident_t *id_ref, int gtid,
1467 kmp_uint64 *lhs, kmp_uint64 rhs,
1467 kmp_uint64 *lhs, kmp_uint64 rhs,
projects/openmp/runtime/src/kmp_barrier.cpp 90 kmp_uint64 new_state = team_bar->b_arrived + KMP_BARRIER_STATE_BUMP;
305 kmp_uint64 new_state;
517 kmp_uint64 new_state = KMP_BARRIER_UNUSED_STATE;
869 kmp_uint64 new_state;
904 kmp_uint64 leaf_state =
1116 kmp_uint64 old_leaf_state = thr_bar->leaf_state;
1447 kmp_uint64 cur_time = __itt_get_timestamp();
1465 kmp_uint64 delta = cur_time - this_thr->th.th_bar_arrive_time;
1832 kmp_uint64 cur_time = __itt_get_timestamp();
1849 kmp_uint64 delta = cur_time - this_thr->th.th_bar_arrive_time;
projects/openmp/runtime/src/kmp_csupport.cpp 3826 kmp_uint64 __kmpc_get_taskid() {
3840 kmp_uint64 __kmpc_get_parent_taskid() {
projects/openmp/runtime/src/kmp_dispatch.cpp 101 kmp_uint64 *cur_chunk,
844 kmp_uint64 cur_chunk = chunk;
911 kmp_uint64 schedtype = 0;
2319 enum sched_type schedule, kmp_uint64 lb,
2320 kmp_uint64 ub, kmp_int64 st, kmp_int64 chunk) {
2325 __kmp_dispatch_init<kmp_uint64>(loc, gtid, schedule, lb, ub, st, chunk, true);
2375 kmp_uint64 lb, kmp_uint64 ub, kmp_int64 st,
2375 kmp_uint64 lb, kmp_uint64 ub, kmp_int64 st,
2381 __kmp_dist_get_bounds<kmp_uint64>(loc, gtid, p_last, &lb, &ub, st);
2382 __kmp_dispatch_init<kmp_uint64>(loc, gtid, schedule, lb, ub, st, chunk, true);
2448 kmp_uint64 *p_lb, kmp_uint64 *p_ub,
2448 kmp_uint64 *p_lb, kmp_uint64 *p_ub,
2453 return __kmp_dispatch_next<kmp_uint64>(loc, gtid, p_last, p_lb, p_ub, p_st
2475 __kmp_dispatch_finish<kmp_uint64>(gtid, loc);
2489 __kmp_dispatch_finish<kmp_uint64>(gtid, loc);
2594 enum sched_type schedule, kmp_uint64 lb,
2595 kmp_uint64 ub, kmp_int64 st, kmp_int64 chunk,
2597 __kmp_dispatch_init<kmp_uint64>(loc, gtid, schedule, lb, ub, st, chunk,
2606 __kmp_dispatch_finish_chunk<kmp_uint64>(gtid, loc);
2614 __kmp_dispatch_finish_chunk<kmp_uint64>(gtid, loc);
projects/openmp/runtime/src/kmp_dispatch.h 51 kmp_uint64 *cur_chunk,
projects/openmp/runtime/src/kmp_global.cpp 288 kmp_uint64 __kmp_taskloop_min_tasks = 0;
projects/openmp/runtime/src/kmp_itt.h 71 __kmp_inline void __kmp_itt_metadata_imbalance(int gtid, kmp_uint64 begin,
72 kmp_uint64 end,
73 kmp_uint64 imbalance,
74 kmp_uint64 reduction);
77 __kmp_inline void __kmp_itt_metadata_loop(ident_t *loc, kmp_uint64 sched_type,
78 kmp_uint64 iterations,
79 kmp_uint64 chunk);
projects/openmp/runtime/src/kmp_itt.inl 321 LINKAGE void __kmp_itt_metadata_imbalance(int gtid, kmp_uint64 begin,
322 kmp_uint64 end, kmp_uint64 imbalance,
322 kmp_uint64 end, kmp_uint64 imbalance,
323 kmp_uint64 reduction) {
338 kmp_uint64 imbalance_data[4];
350 LINKAGE void __kmp_itt_metadata_loop(ident_t *loc, kmp_uint64 sched_type,
351 kmp_uint64 iterations, kmp_uint64 chunk) {
351 kmp_uint64 iterations, kmp_uint64 chunk) {
383 kmp_uint64 loop_data[5];
412 kmp_uint64 single_data[2];
500 kmp_uint64 counter =
projects/openmp/runtime/src/kmp_lock.cpp 2234 kmp_uint64 ticket = KMP_ATOMIC_INC(&lck->lk.next_ticket);
2235 kmp_uint64 mask = lck->lk.mask; // atomic load
2236 std::atomic<kmp_uint64> *polls = lck->lk.polls;
2289 std::atomic<kmp_uint64> *old_polls = polls;
2309 kmp_uint64 num_waiting = TCR_8(lck->lk.next_ticket) - ticket - 1;
2393 kmp_uint64 ticket = lck->lk.next_ticket; // atomic load
2394 std::atomic<kmp_uint64> *polls = lck->lk.polls;
2395 kmp_uint64 mask = lck->lk.mask; // atomic load
2397 kmp_uint64 next_ticket = ticket + 1;
2438 kmp_uint64 ticket = lck->lk.now_serving + 1; // non-atomic load
2439 std::atomic<kmp_uint64> *polls = lck->lk.polls; // atomic load
2440 kmp_uint64 mask = lck->lk.mask; // atomic load
2677 static inline bool before(kmp_uint64 a, kmp_uint64 b) {
2677 static inline bool before(kmp_uint64 a, kmp_uint64 b) {
2686 kmp_uint64 goal = __kmp_tsc() + boff->min_tick;
projects/openmp/runtime/src/kmp_lock.h 448 std::atomic<std::atomic<kmp_uint64> *> polls;
449 std::atomic<kmp_uint64> mask; // is 2**num_polls-1 for mod op
450 kmp_uint64 cleanup_ticket; // thread with cleanup ticket
451 std::atomic<kmp_uint64> *old_polls; // will deallocate old_polls
458 std::atomic<kmp_uint64> next_ticket;
473 kmp_uint64 now_serving; // doesn't have to be volatile
projects/openmp/runtime/src/kmp_os.h 195 typedef kmp_uint64 kmp_uint;
projects/openmp/runtime/src/kmp_runtime.cpp 2171 kmp_uint64 tmp_time = 0;
projects/openmp/runtime/src/kmp_sched.cpp 383 kmp_uint64 cur_chunk = chunk;
845 kmp_uint64 *plower, kmp_uint64 *pupper,
845 kmp_uint64 *plower, kmp_uint64 *pupper,
848 __kmp_for_static_init<kmp_uint64>(loc, gtid, schedtype, plastiter, plower,
920 kmp_uint64 *plower, kmp_uint64 *pupper,
920 kmp_uint64 *plower, kmp_uint64 *pupper,
921 kmp_uint64 *pupperD, kmp_int64 *pstride,
923 __kmp_dist_for_static_init<kmp_uint64>(loc, gtid, schedule, plastiter, plower,
993 kmp_uint64 *p_lb, kmp_uint64 *p_ub,
993 kmp_uint64 *p_lb, kmp_uint64 *p_ub,
997 __kmp_team_static_init<kmp_uint64>(loc, gtid, p_last, p_lb, p_ub, p_st, incr,
projects/openmp/runtime/src/kmp_settings.cpp 372 kmp_uint64 uint = *out;
554 kmp_uint64 value) {
1216 kmp_uint64 tmp_dflt = 0;
projects/openmp/runtime/src/kmp_str.cpp 704 kmp_uint64 *out, // O: Parsed number.
projects/openmp/runtime/src/kmp_str.h 117 void __kmp_str_to_uint(char const *str, kmp_uint64 *out, char const **error);
projects/openmp/runtime/src/kmp_tasking.cpp 1529 kmp_uint64 cur_time;
3970 kmp_taskloop_bounds_t(kmp_task_t *_task, kmp_uint64 *lb, kmp_uint64 *ub)
3970 kmp_taskloop_bounds_t(kmp_task_t *_task, kmp_uint64 *lb, kmp_uint64 *ub)
3982 kmp_uint64 get_lb() const {
4003 kmp_uint64 get_ub() const {
4024 void set_lb(kmp_uint64 lb) {
4035 kmp_uint64 *lower = RCAST(kmp_uint64 *, task->shareds);
4043 void set_ub(kmp_uint64 ub) {
4054 kmp_uint64 *upper = RCAST(kmp_uint64 *, task->shareds) + 1;
4080 kmp_uint64 *lb, kmp_uint64 *ub, kmp_int64 st,
4080 kmp_uint64 *lb, kmp_uint64 *ub, kmp_int64 st,
4081 kmp_uint64 ub_glob, kmp_uint64 num_tasks,
4081 kmp_uint64 ub_glob, kmp_uint64 num_tasks,
4082 kmp_uint64 grainsize, kmp_uint64 extras,
4082 kmp_uint64 grainsize, kmp_uint64 extras,
4083 kmp_uint64 tc,
4093 kmp_uint64 lower = task_bounds.get_lb();
4094 kmp_uint64 upper = task_bounds.get_ub();
4095 kmp_uint64 i;
4111 kmp_uint64 chunk_minus_1;
4173 kmp_uint64 *lb;
4174 kmp_uint64 *ub;
4177 kmp_uint64 ub_glob;
4178 kmp_uint64 num_tasks;
4179 kmp_uint64 grainsize;
4180 kmp_uint64 extras;
4181 kmp_uint64 tc;
4182 kmp_uint64 num_t_min;
4188 void __kmp_taskloop_recur(ident_t *, int, kmp_task_t *, kmp_uint64 *,
4189 kmp_uint64 *, kmp_int64, kmp_uint64, kmp_uint64,
4189 kmp_uint64 *, kmp_int64, kmp_uint64, kmp_uint64,
4189 kmp_uint64 *, kmp_int64, kmp_uint64, kmp_uint64,
4190 kmp_uint64, kmp_uint64, kmp_uint64, kmp_uint64,
4190 kmp_uint64, kmp_uint64, kmp_uint64, kmp_uint64,
4190 kmp_uint64, kmp_uint64, kmp_uint64, kmp_uint64,
4190 kmp_uint64, kmp_uint64, kmp_uint64, kmp_uint64,
4201 kmp_uint64 *lb = p->lb;
4202 kmp_uint64 *ub = p->ub;
4206 kmp_uint64 ub_glob = p->ub_glob;
4207 kmp_uint64 num_tasks = p->num_tasks;
4208 kmp_uint64 grainsize = p->grainsize;
4209 kmp_uint64 extras = p->extras;
4210 kmp_uint64 tc = p->tc;
4211 kmp_uint64 num_t_min = p->num_t_min;
4261 kmp_uint64 *lb, kmp_uint64 *ub, kmp_int64 st,
4261 kmp_uint64 *lb, kmp_uint64 *ub, kmp_int64 st,
4262 kmp_uint64 ub_glob, kmp_uint64 num_tasks,
4262 kmp_uint64 ub_glob, kmp_uint64 num_tasks,
4263 kmp_uint64 grainsize, kmp_uint64 extras,
4263 kmp_uint64 grainsize, kmp_uint64 extras,
4264 kmp_uint64 tc, kmp_uint64 num_t_min,
4264 kmp_uint64 tc, kmp_uint64 num_t_min,
4279 kmp_uint64 lower = *lb;
4293 kmp_uint64 lb1, ub0, tc0, tc1, ext0, ext1;
4294 kmp_uint64 gr_size0 = grainsize;
4295 kmp_uint64 n_tsk0 = num_tasks >> 1; // num_tasks/2 to execute
4296 kmp_uint64 n_tsk1 = num_tasks - n_tsk0; // to schedule as a task
4383 kmp_uint64 *lb, kmp_uint64 *ub, kmp_int64 st, int nogroup,
4383 kmp_uint64 *lb, kmp_uint64 *ub, kmp_int64 st, int nogroup,
4384 int sched, kmp_uint64 grainsize, void *task_dup) {
4398 kmp_uint64 tc;
4400 kmp_uint64 lower = task_bounds.get_lb();
4401 kmp_uint64 upper = task_bounds.get_ub();
4402 kmp_uint64 ub_glob = upper; // global upper used to calc lastprivate flag
4403 kmp_uint64 num_tasks = 0, extras = 0;
4404 kmp_uint64 num_tasks_min = __kmp_taskloop_min_tasks;
projects/openmp/runtime/src/kmp_utility.cpp 93 static kmp_uint64 __kmp_parse_frequency( // R: Frequency in Hz.
99 kmp_uint64 result = 0; /* Zero is a better unknown value than all ones. */
projects/openmp/runtime/src/kmp_wait_release.h 176 kmp_uint64 poll_count;
177 kmp_uint64 hibernate_goal;
533 typedef kmp_uint64 flag_t;
768 class kmp_flag_64 : public kmp_basic_flag_native<kmp_uint64> {
770 kmp_flag_64(volatile kmp_uint64 *p) : kmp_basic_flag_native<kmp_uint64>(p) {}
771 kmp_flag_64(volatile kmp_uint64 *p, kmp_info_t *thr)
773 kmp_flag_64(volatile kmp_uint64 *p, kmp_uint64 c)
773 kmp_flag_64(volatile kmp_uint64 *p, kmp_uint64 c)
810 class kmp_flag_oncore : public kmp_flag_native<kmp_uint64> {
811 kmp_uint64 checker;
824 unsigned char &byteref(volatile kmp_uint64 *loc, size_t offset) {
829 kmp_flag_oncore(volatile kmp_uint64 *p)
832 kmp_flag_oncore(volatile kmp_uint64 *p, kmp_uint32 idx)
835 kmp_flag_oncore(volatile kmp_uint64 *p, kmp_uint64 c, kmp_uint32 idx,
835 kmp_flag_oncore(volatile kmp_uint64 *p, kmp_uint64 c, kmp_uint32 idx,
850 bool done_check_val(kmp_uint64 old_loc) {
873 kmp_uint64 mask = 0;
878 kmp_uint64 set_sleeping() {
881 kmp_uint64 unset_sleeping() {
884 bool is_sleeping_val(kmp_uint64 old_loc) {
projects/openmp/runtime/src/z_Linux_util.cpp 93 kmp_uint64 __kmp_ticks_per_msec = 1000000;
1972 kmp_uint64 __kmp_now_nsec() {
1975 kmp_uint64 nsec = (kmp_uint64)KMP_NSEC_PER_SEC * (kmp_uint64)t.tv_sec +
1983 kmp_uint64 now, nsec2, diff;
1984 kmp_uint64 delay = 100000; // 50~100 usec on most machines.
1985 kmp_uint64 nsec = __kmp_now_nsec();
1986 kmp_uint64 goal = __kmp_hardware_timestamp() + delay;
1992 kmp_uint64 tpms = (kmp_uint64)(1e6 * (delay + (now - goal)) / diff);