Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 1 | /* |
| 2 | * kmp_global.c -- KPTS global variables for runtime support library |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 3 | */ |
| 4 | |
| 5 | |
| 6 | //===----------------------------------------------------------------------===// |
| 7 | // |
| 8 | // The LLVM Compiler Infrastructure |
| 9 | // |
| 10 | // This file is dual licensed under the MIT and the University of Illinois Open |
| 11 | // Source Licenses. See LICENSE.txt for details. |
| 12 | // |
| 13 | //===----------------------------------------------------------------------===// |
| 14 | |
| 15 | |
| 16 | #include "kmp.h" |
| 17 | |
| 18 | #ifdef KMP_SETVERSION |
| 19 | char __kmp_setversion_string[] = VERSION_STRING; |
| 20 | #endif |
| 21 | |
| 22 | kmp_key_t __kmp_gtid_threadprivate_key; |
| 23 | |
| 24 | kmp_cpuinfo_t __kmp_cpuinfo = { 0 }; // Not initialized |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 25 | |
Jim Cownie | 4cc4bb4 | 2014-10-07 16:25:50 +0000 | [diff] [blame] | 26 | #if KMP_STATS_ENABLED |
| 27 | #include "kmp_stats.h" |
| 28 | // lock for modifying the global __kmp_stats_list |
| 29 | kmp_tas_lock_t __kmp_stats_lock = KMP_TAS_LOCK_INITIALIZER(__kmp_stats_lock); |
| 30 | |
| 31 | // global list of per thread stats, the head is a sentinel node which accumulates all stats produced before __kmp_create_worker is called. |
| 32 | kmp_stats_list __kmp_stats_list; |
| 33 | |
| 34 | // thread local pointer to stats node within list |
| 35 | __thread kmp_stats_list* __kmp_stats_thread_ptr = &__kmp_stats_list; |
| 36 | |
| 37 | // gives reference tick for all events (considered the 0 tick) |
| 38 | tsc_tick_count __kmp_stats_start_time; |
| 39 | #endif |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 40 | |
| 41 | /* ----------------------------------------------------- */ |
| 42 | /* INITIALIZATION VARIABLES */ |
| 43 | /* they are syncronized to write during init, but read anytime */ |
| 44 | volatile int __kmp_init_serial = FALSE; |
| 45 | volatile int __kmp_init_gtid = FALSE; |
| 46 | volatile int __kmp_init_common = FALSE; |
| 47 | volatile int __kmp_init_middle = FALSE; |
| 48 | volatile int __kmp_init_parallel = FALSE; |
| 49 | volatile int __kmp_init_monitor = 0; /* 1 - launched, 2 - actually started (Windows* OS only) */ |
| 50 | volatile int __kmp_init_user_locks = FALSE; |
| 51 | |
| 52 | /* list of address of allocated caches for commons */ |
| 53 | kmp_cached_addr_t *__kmp_threadpriv_cache_list = NULL; |
| 54 | |
| 55 | int __kmp_init_counter = 0; |
| 56 | int __kmp_root_counter = 0; |
| 57 | int __kmp_version = 0; |
| 58 | |
| 59 | volatile kmp_uint32 __kmp_team_counter = 0; |
| 60 | volatile kmp_uint32 __kmp_task_counter = 0; |
| 61 | |
| 62 | unsigned int __kmp_init_wait = KMP_DEFAULT_INIT_WAIT; /* initial number of spin-tests */ |
| 63 | unsigned int __kmp_next_wait = KMP_DEFAULT_NEXT_WAIT; /* susequent number of spin-tests */ |
| 64 | |
| 65 | size_t __kmp_stksize = KMP_DEFAULT_STKSIZE; |
| 66 | size_t __kmp_monitor_stksize = 0; // auto adjust |
| 67 | size_t __kmp_stkoffset = KMP_DEFAULT_STKOFFSET; |
Jim Cownie | 4cc4bb4 | 2014-10-07 16:25:50 +0000 | [diff] [blame] | 68 | int __kmp_stkpadding = KMP_MIN_STKPADDING; |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 69 | |
| 70 | size_t __kmp_malloc_pool_incr = KMP_DEFAULT_MALLOC_POOL_INCR; |
| 71 | |
| 72 | /* Barrier method defaults, settings, and strings */ |
| 73 | /* branch factor = 2^branch_bits (only relevant for tree and hyper barrier types) */ |
| 74 | #if KMP_ARCH_X86_64 |
| 75 | kmp_uint32 __kmp_barrier_gather_bb_dflt = 2; /* branch_factor = 4 */ /* hyper2: C78980 */ |
| 76 | kmp_uint32 __kmp_barrier_release_bb_dflt = 2; /* branch_factor = 4 */ /* hyper2: C78980 */ |
| 77 | #else |
| 78 | kmp_uint32 __kmp_barrier_gather_bb_dflt = 2; /* branch_factor = 4 */ /* communication in core for MIC */ |
| 79 | kmp_uint32 __kmp_barrier_release_bb_dflt = 2; /* branch_factor = 4 */ /* communication in core for MIC */ |
| 80 | #endif // KMP_ARCH_X86_64 |
| 81 | #if KMP_ARCH_X86_64 |
| 82 | kmp_bar_pat_e __kmp_barrier_gather_pat_dflt = bp_hyper_bar; /* hyper2: C78980 */ |
| 83 | kmp_bar_pat_e __kmp_barrier_release_pat_dflt = bp_hyper_bar; /* hyper2: C78980 */ |
| 84 | #else |
| 85 | kmp_bar_pat_e __kmp_barrier_gather_pat_dflt = bp_linear_bar; |
| 86 | kmp_bar_pat_e __kmp_barrier_release_pat_dflt = bp_linear_bar; |
| 87 | #endif |
| 88 | kmp_uint32 __kmp_barrier_gather_branch_bits [ bs_last_barrier ] = { 0 }; |
| 89 | kmp_uint32 __kmp_barrier_release_branch_bits [ bs_last_barrier ] = { 0 }; |
| 90 | kmp_bar_pat_e __kmp_barrier_gather_pattern [ bs_last_barrier ] = { bp_linear_bar }; |
| 91 | kmp_bar_pat_e __kmp_barrier_release_pattern [ bs_last_barrier ] = { bp_linear_bar }; |
| 92 | char const *__kmp_barrier_branch_bit_env_name [ bs_last_barrier ] = |
| 93 | { "KMP_PLAIN_BARRIER", "KMP_FORKJOIN_BARRIER" |
| 94 | #if KMP_FAST_REDUCTION_BARRIER |
| 95 | , "KMP_REDUCTION_BARRIER" |
| 96 | #endif // KMP_FAST_REDUCTION_BARRIER |
| 97 | }; |
| 98 | char const *__kmp_barrier_pattern_env_name [ bs_last_barrier ] = |
| 99 | { "KMP_PLAIN_BARRIER_PATTERN", "KMP_FORKJOIN_BARRIER_PATTERN" |
| 100 | #if KMP_FAST_REDUCTION_BARRIER |
| 101 | , "KMP_REDUCTION_BARRIER_PATTERN" |
| 102 | #endif // KMP_FAST_REDUCTION_BARRIER |
| 103 | }; |
| 104 | char const *__kmp_barrier_type_name [ bs_last_barrier ] = |
| 105 | { "plain", "forkjoin" |
| 106 | #if KMP_FAST_REDUCTION_BARRIER |
| 107 | , "reduction" |
| 108 | #endif // KMP_FAST_REDUCTION_BARRIER |
| 109 | }; |
Jim Cownie | 4cc4bb4 | 2014-10-07 16:25:50 +0000 | [diff] [blame] | 110 | char const *__kmp_barrier_pattern_name [ bp_last_bar ] = { "linear", "tree", "hyper", "hierarchical" }; |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 111 | |
| 112 | |
| 113 | int __kmp_allThreadsSpecified = 0; |
| 114 | size_t __kmp_align_alloc = CACHE_LINE; |
| 115 | |
| 116 | |
| 117 | int __kmp_generate_warnings = kmp_warnings_low; |
| 118 | int __kmp_reserve_warn = 0; |
| 119 | int __kmp_xproc = 0; |
| 120 | int __kmp_avail_proc = 0; |
| 121 | size_t __kmp_sys_min_stksize = KMP_MIN_STKSIZE; |
| 122 | int __kmp_sys_max_nth = KMP_MAX_NTH; |
| 123 | int __kmp_max_nth = 0; |
| 124 | int __kmp_threads_capacity = 0; |
| 125 | int __kmp_dflt_team_nth = 0; |
| 126 | int __kmp_dflt_team_nth_ub = 0; |
| 127 | int __kmp_tp_capacity = 0; |
| 128 | int __kmp_tp_cached = 0; |
| 129 | int __kmp_dflt_nested = FALSE; |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 130 | int __kmp_dflt_max_active_levels = KMP_MAX_ACTIVE_LEVELS_LIMIT; /* max_active_levels limit */ |
Jim Cownie | 4cc4bb4 | 2014-10-07 16:25:50 +0000 | [diff] [blame] | 131 | #if KMP_NESTED_HOT_TEAMS |
| 132 | int __kmp_hot_teams_mode = 0; /* 0 - free extra threads when reduced */ |
| 133 | /* 1 - keep extra threads when reduced */ |
| 134 | int __kmp_hot_teams_max_level = 1; /* nesting level of hot teams */ |
| 135 | #endif |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 136 | enum library_type __kmp_library = library_none; |
| 137 | enum sched_type __kmp_sched = kmp_sch_default; /* scheduling method for runtime scheduling */ |
| 138 | enum sched_type __kmp_static = kmp_sch_static_greedy; /* default static scheduling method */ |
| 139 | enum sched_type __kmp_guided = kmp_sch_guided_iterative_chunked; /* default guided scheduling method */ |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 140 | enum sched_type __kmp_auto = kmp_sch_guided_analytical_chunked; /* default auto scheduling method */ |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 141 | int __kmp_dflt_blocktime = KMP_DEFAULT_BLOCKTIME; |
| 142 | int __kmp_monitor_wakeups = KMP_MIN_MONITOR_WAKEUPS; |
| 143 | int __kmp_bt_intervals = KMP_INTERVALS_FROM_BLOCKTIME( KMP_DEFAULT_BLOCKTIME, KMP_MIN_MONITOR_WAKEUPS ); |
| 144 | #ifdef KMP_ADJUST_BLOCKTIME |
| 145 | int __kmp_zero_bt = FALSE; |
| 146 | #endif /* KMP_ADJUST_BLOCKTIME */ |
Andrey Churbanov | f696c82 | 2015-01-27 16:55:43 +0000 | [diff] [blame] | 147 | #ifdef KMP_DFLT_NTH_CORES |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 148 | int __kmp_ncores = 0; |
Andrey Churbanov | f696c82 | 2015-01-27 16:55:43 +0000 | [diff] [blame] | 149 | #endif |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 150 | int __kmp_chunk = 0; |
| 151 | int __kmp_abort_delay = 0; |
| 152 | #if KMP_OS_LINUX && defined(KMP_TDATA_GTID) |
| 153 | int __kmp_gtid_mode = 3; /* use __declspec(thread) TLS to store gtid */ |
| 154 | int __kmp_adjust_gtid_mode = FALSE; |
| 155 | #elif KMP_OS_WINDOWS |
| 156 | int __kmp_gtid_mode = 2; /* use TLS functions to store gtid */ |
| 157 | int __kmp_adjust_gtid_mode = FALSE; |
| 158 | #else |
| 159 | int __kmp_gtid_mode = 0; /* select method to get gtid based on #threads */ |
| 160 | int __kmp_adjust_gtid_mode = TRUE; |
| 161 | #endif /* KMP_OS_LINUX && defined(KMP_TDATA_GTID) */ |
| 162 | #ifdef KMP_TDATA_GTID |
| 163 | #if KMP_OS_WINDOWS |
| 164 | __declspec(thread) int __kmp_gtid = KMP_GTID_DNE; |
| 165 | #else |
| 166 | __thread int __kmp_gtid = KMP_GTID_DNE; |
| 167 | #endif /* KMP_OS_WINDOWS - workaround because Intel(R) Many Integrated Core compiler 20110316 doesn't accept __declspec */ |
| 168 | #endif /* KMP_TDATA_GTID */ |
| 169 | int __kmp_tls_gtid_min = INT_MAX; |
| 170 | int __kmp_foreign_tp = TRUE; |
| 171 | #if KMP_ARCH_X86 || KMP_ARCH_X86_64 |
| 172 | int __kmp_inherit_fp_control = TRUE; |
| 173 | kmp_int16 __kmp_init_x87_fpu_control_word = 0; |
| 174 | kmp_uint32 __kmp_init_mxcsr = 0; |
| 175 | #endif /* KMP_ARCH_X86 || KMP_ARCH_X86_64 */ |
| 176 | |
| 177 | #ifdef USE_LOAD_BALANCE |
| 178 | double __kmp_load_balance_interval = 1.0; |
| 179 | #endif /* USE_LOAD_BALANCE */ |
| 180 | |
| 181 | kmp_nested_nthreads_t __kmp_nested_nth = { NULL, 0, 0 }; |
| 182 | |
| 183 | #if KMP_USE_ADAPTIVE_LOCKS |
| 184 | |
| 185 | kmp_adaptive_backoff_params_t __kmp_adaptive_backoff_params = { 1, 1024 }; // TODO: tune it! |
| 186 | |
| 187 | #if KMP_DEBUG_ADAPTIVE_LOCKS |
| 188 | char * __kmp_speculative_statsfile = "-"; |
| 189 | #endif |
| 190 | |
| 191 | #endif // KMP_USE_ADAPTIVE_LOCKS |
| 192 | |
| 193 | #if OMP_40_ENABLED |
| 194 | int __kmp_display_env = FALSE; |
| 195 | int __kmp_display_env_verbose = FALSE; |
Jim Cownie | 181b4bb | 2013-12-23 17:28:57 +0000 | [diff] [blame] | 196 | int __kmp_omp_cancellation = FALSE; |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 197 | #endif |
| 198 | |
| 199 | /* map OMP 3.0 schedule types with our internal schedule types */ |
| 200 | enum sched_type __kmp_sch_map[ kmp_sched_upper - kmp_sched_lower_ext + kmp_sched_upper_std - kmp_sched_lower - 2 ] = { |
| 201 | kmp_sch_static_chunked, // ==> kmp_sched_static = 1 |
| 202 | kmp_sch_dynamic_chunked, // ==> kmp_sched_dynamic = 2 |
| 203 | kmp_sch_guided_chunked, // ==> kmp_sched_guided = 3 |
| 204 | kmp_sch_auto, // ==> kmp_sched_auto = 4 |
| 205 | kmp_sch_trapezoidal // ==> kmp_sched_trapezoidal = 101 |
| 206 | // will likely not used, introduced here just to debug the code |
| 207 | // of public intel extension schedules |
| 208 | }; |
| 209 | |
| 210 | #if KMP_OS_LINUX |
| 211 | enum clock_function_type __kmp_clock_function; |
| 212 | int __kmp_clock_function_param; |
| 213 | #endif /* KMP_OS_LINUX */ |
| 214 | |
Andrey Churbanov | 613edeb | 2015-02-20 18:14:43 +0000 | [diff] [blame] | 215 | #if KMP_ARCH_X86_64 && (KMP_OS_LINUX || KMP_OS_WINDOWS) |
| 216 | enum mic_type __kmp_mic_type = non_mic; |
| 217 | #endif |
| 218 | |
Alp Toker | 98758b0 | 2014-03-02 04:12:06 +0000 | [diff] [blame] | 219 | #if KMP_AFFINITY_SUPPORTED |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 220 | |
Andrey Churbanov | 7daf980 | 2015-01-27 16:52:57 +0000 | [diff] [blame] | 221 | # if KMP_GROUP_AFFINITY |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 222 | |
| 223 | int __kmp_num_proc_groups = 1; |
| 224 | |
| 225 | kmp_GetActiveProcessorCount_t __kmp_GetActiveProcessorCount = NULL; |
| 226 | kmp_GetActiveProcessorGroupCount_t __kmp_GetActiveProcessorGroupCount = NULL; |
| 227 | kmp_GetThreadGroupAffinity_t __kmp_GetThreadGroupAffinity = NULL; |
| 228 | kmp_SetThreadGroupAffinity_t __kmp_SetThreadGroupAffinity = NULL; |
| 229 | |
Andrey Churbanov | 7daf980 | 2015-01-27 16:52:57 +0000 | [diff] [blame] | 230 | # endif /* KMP_GROUP_AFFINITY */ |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 231 | |
| 232 | size_t __kmp_affin_mask_size = 0; |
| 233 | enum affinity_type __kmp_affinity_type = affinity_default; |
| 234 | enum affinity_gran __kmp_affinity_gran = affinity_gran_default; |
| 235 | int __kmp_affinity_gran_levels = -1; |
| 236 | int __kmp_affinity_dups = TRUE; |
| 237 | enum affinity_top_method __kmp_affinity_top_method = affinity_top_method_default; |
| 238 | int __kmp_affinity_compact = 0; |
| 239 | int __kmp_affinity_offset = 0; |
| 240 | int __kmp_affinity_verbose = FALSE; |
| 241 | int __kmp_affinity_warnings = TRUE; |
| 242 | int __kmp_affinity_respect_mask = affinity_respect_mask_default; |
| 243 | char * __kmp_affinity_proclist = NULL; |
| 244 | kmp_affin_mask_t *__kmp_affinity_masks = NULL; |
| 245 | unsigned __kmp_affinity_num_masks = 0; |
| 246 | |
| 247 | char const * __kmp_cpuinfo_file = NULL; |
| 248 | |
Alp Toker | 98758b0 | 2014-03-02 04:12:06 +0000 | [diff] [blame] | 249 | #endif /* KMP_AFFINITY_SUPPORTED */ |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 250 | |
| 251 | #if OMP_40_ENABLED |
| 252 | kmp_nested_proc_bind_t __kmp_nested_proc_bind = { NULL, 0, 0 }; |
| 253 | int __kmp_affinity_num_places = 0; |
| 254 | #endif |
| 255 | |
Andrey Churbanov | 1287557 | 2015-03-10 09:00:36 +0000 | [diff] [blame] | 256 | int __kmp_place_num_cores = 0; |
| 257 | int __kmp_place_num_threads_per_core = 0; |
| 258 | int __kmp_place_core_offset = 0; |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 259 | |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 260 | kmp_tasking_mode_t __kmp_tasking_mode = tskm_task_teams; |
| 261 | |
| 262 | /* This check ensures that the compiler is passing the correct data type |
| 263 | * for the flags formal parameter of the function kmpc_omp_task_alloc(). |
| 264 | * If the type is not a 4-byte type, then give an error message about |
| 265 | * a non-positive length array pointing here. If that happens, the |
| 266 | * kmp_tasking_flags_t structure must be redefined to have exactly 32 bits. |
| 267 | */ |
| 268 | KMP_BUILD_ASSERT( sizeof(kmp_tasking_flags_t) == 4 ); |
| 269 | |
| 270 | kmp_int32 __kmp_task_stealing_constraint = 1; /* Constrain task stealing by default */ |
| 271 | |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 272 | #ifdef DEBUG_SUSPEND |
| 273 | int __kmp_suspend_count = 0; |
| 274 | #endif |
| 275 | |
| 276 | int __kmp_settings = FALSE; |
| 277 | int __kmp_duplicate_library_ok = 0; |
| 278 | #if USE_ITT_BUILD |
| 279 | int __kmp_forkjoin_frames = 1; |
| 280 | int __kmp_forkjoin_frames_mode = 0; |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 281 | #endif |
| 282 | PACKED_REDUCTION_METHOD_T __kmp_force_reduction_method = reduction_method_not_defined; |
| 283 | int __kmp_determ_red = FALSE; |
| 284 | |
| 285 | #ifdef KMP_DEBUG |
| 286 | int kmp_a_debug = 0; |
| 287 | int kmp_b_debug = 0; |
| 288 | int kmp_c_debug = 0; |
| 289 | int kmp_d_debug = 0; |
| 290 | int kmp_e_debug = 0; |
| 291 | int kmp_f_debug = 0; |
| 292 | int kmp_diag = 0; |
| 293 | #endif |
| 294 | |
| 295 | /* For debug information logging using rotating buffer */ |
| 296 | int __kmp_debug_buf = FALSE; /* TRUE means use buffer, FALSE means print to stderr */ |
| 297 | int __kmp_debug_buf_lines = KMP_DEBUG_BUF_LINES_INIT; /* Lines of debug stored in buffer */ |
| 298 | int __kmp_debug_buf_chars = KMP_DEBUG_BUF_CHARS_INIT; /* Characters allowed per line in buffer */ |
| 299 | int __kmp_debug_buf_atomic = FALSE; /* TRUE means use atomic update of buffer entry pointer */ |
| 300 | |
| 301 | char *__kmp_debug_buffer = NULL; /* Debug buffer itself */ |
| 302 | int __kmp_debug_count = 0; /* Counter for number of lines printed in buffer so far */ |
| 303 | int __kmp_debug_buf_warn_chars = 0; /* Keep track of char increase recommended in warnings */ |
| 304 | /* end rotating debug buffer */ |
| 305 | |
| 306 | #ifdef KMP_DEBUG |
| 307 | int __kmp_par_range; /* +1 => only go par for constructs in range */ |
| 308 | /* -1 => only go par for constructs outside range */ |
| 309 | char __kmp_par_range_routine[KMP_PAR_RANGE_ROUTINE_LEN] = { '\0' }; |
| 310 | char __kmp_par_range_filename[KMP_PAR_RANGE_FILENAME_LEN] = { '\0' }; |
| 311 | int __kmp_par_range_lb = 0; |
| 312 | int __kmp_par_range_ub = INT_MAX; |
| 313 | #endif /* KMP_DEBUG */ |
| 314 | |
| 315 | /* For printing out dynamic storage map for threads and teams */ |
| 316 | int __kmp_storage_map = FALSE; /* True means print storage map for threads and teams */ |
| 317 | int __kmp_storage_map_verbose = FALSE; /* True means storage map includes placement info */ |
| 318 | int __kmp_storage_map_verbose_specified = FALSE; |
| 319 | /* Initialize the library data structures when we fork a child process, defaults to TRUE */ |
| 320 | int __kmp_need_register_atfork = TRUE; /* At initialization, call pthread_atfork to install fork handler */ |
| 321 | int __kmp_need_register_atfork_specified = TRUE; |
| 322 | |
| 323 | |
| 324 | int __kmp_env_chunk = FALSE; /* KMP_CHUNK specified? */ |
| 325 | int __kmp_env_stksize = FALSE; /* KMP_STACKSIZE specified? */ |
| 326 | int __kmp_env_omp_stksize = FALSE; /* OMP_STACKSIZE specified? */ |
| 327 | int __kmp_env_all_threads = FALSE;/* KMP_ALL_THREADS or KMP_MAX_THREADS specified? */ |
| 328 | int __kmp_env_omp_all_threads = FALSE;/* OMP_THREAD_LIMIT specified? */ |
| 329 | int __kmp_env_blocktime = FALSE; /* KMP_BLOCKTIME specified? */ |
| 330 | int __kmp_env_checks = FALSE; /* KMP_CHECKS specified? */ |
| 331 | int __kmp_env_consistency_check = FALSE; /* KMP_CONSISTENCY_CHECK specified? */ |
| 332 | |
| 333 | kmp_uint32 __kmp_yield_init = KMP_INIT_WAIT; |
| 334 | kmp_uint32 __kmp_yield_next = KMP_NEXT_WAIT; |
| 335 | kmp_uint32 __kmp_yielding_on = 1; |
Jim Cownie | 3051f97 | 2014-08-07 10:12:54 +0000 | [diff] [blame] | 336 | #if KMP_OS_CNK |
| 337 | kmp_uint32 __kmp_yield_cycle = 0; |
| 338 | #else |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 339 | kmp_uint32 __kmp_yield_cycle = 1; /* Yield-cycle is on by default */ |
Jim Cownie | 3051f97 | 2014-08-07 10:12:54 +0000 | [diff] [blame] | 340 | #endif |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 341 | kmp_int32 __kmp_yield_on_count = 10; /* By default, yielding is on for 10 monitor periods. */ |
| 342 | kmp_int32 __kmp_yield_off_count = 1; /* By default, yielding is off for 1 monitor periods. */ |
| 343 | /* ----------------------------------------------------- */ |
| 344 | |
| 345 | |
| 346 | /* ------------------------------------------------------ */ |
| 347 | /* STATE mostly syncronized with global lock */ |
| 348 | /* data written to rarely by masters, read often by workers */ |
| 349 | /* |
| 350 | * SHALL WE EDIT THE COMMENT BELOW IN SOME WAY? |
| 351 | * TODO: None of this global padding stuff works consistently because |
| 352 | * the order of declaration is not necessarily correlated to storage order. |
| 353 | * To fix this, all the important globals must be put in a big structure |
| 354 | * instead. |
| 355 | */ |
| 356 | KMP_ALIGN_CACHE |
| 357 | kmp_info_t **__kmp_threads = NULL; |
| 358 | kmp_root_t **__kmp_root = NULL; |
| 359 | |
| 360 | /* data read/written to often by masters */ |
| 361 | KMP_ALIGN_CACHE |
| 362 | volatile int __kmp_nth = 0; |
| 363 | volatile int __kmp_all_nth = 0; |
| 364 | int __kmp_thread_pool_nth = 0; |
| 365 | volatile kmp_info_t *__kmp_thread_pool = NULL; |
| 366 | volatile kmp_team_t *__kmp_team_pool = NULL; |
| 367 | |
| 368 | KMP_ALIGN_CACHE |
| 369 | volatile int __kmp_thread_pool_active_nth = 0; |
| 370 | |
| 371 | /* ------------------------------------------------- |
| 372 | * GLOBAL/ROOT STATE */ |
| 373 | KMP_ALIGN_CACHE |
| 374 | kmp_global_t __kmp_global = {{ 0 }}; |
| 375 | |
| 376 | /* ----------------------------------------------- */ |
Alp Toker | 8f2d3f0 | 2014-02-24 10:40:15 +0000 | [diff] [blame] | 377 | /* GLOBAL SYNCHRONIZATION LOCKS */ |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 378 | /* TODO verify the need for these locks and if they need to be global */ |
Jim Cownie | 4cc4bb4 | 2014-10-07 16:25:50 +0000 | [diff] [blame] | 379 | |
| 380 | #if KMP_USE_INTERNODE_ALIGNMENT |
| 381 | /* Multinode systems have larger cache line granularity which can cause |
| 382 | * false sharing if the alignment is not large enough for these locks */ |
| 383 | KMP_ALIGN_CACHE_INTERNODE |
| 384 | |
| 385 | kmp_bootstrap_lock_t __kmp_initz_lock = KMP_BOOTSTRAP_LOCK_INITIALIZER( __kmp_initz_lock ); /* Control initializations */ |
| 386 | KMP_ALIGN_CACHE_INTERNODE |
| 387 | kmp_bootstrap_lock_t __kmp_forkjoin_lock; /* control fork/join access */ |
| 388 | KMP_ALIGN_CACHE_INTERNODE |
| 389 | kmp_bootstrap_lock_t __kmp_exit_lock; /* exit() is not always thread-safe */ |
| 390 | KMP_ALIGN_CACHE_INTERNODE |
| 391 | kmp_bootstrap_lock_t __kmp_monitor_lock; /* control monitor thread creation */ |
| 392 | KMP_ALIGN_CACHE_INTERNODE |
| 393 | kmp_bootstrap_lock_t __kmp_tp_cached_lock; /* used for the hack to allow threadprivate cache and __kmp_threads expansion to co-exist */ |
| 394 | |
| 395 | KMP_ALIGN_CACHE_INTERNODE |
| 396 | kmp_lock_t __kmp_global_lock; /* Control OS/global access */ |
| 397 | KMP_ALIGN_CACHE_INTERNODE |
| 398 | kmp_queuing_lock_t __kmp_dispatch_lock; /* Control dispatch access */ |
| 399 | KMP_ALIGN_CACHE_INTERNODE |
| 400 | kmp_lock_t __kmp_debug_lock; /* Control I/O access for KMP_DEBUG */ |
| 401 | #else |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 402 | KMP_ALIGN_CACHE |
| 403 | |
| 404 | kmp_bootstrap_lock_t __kmp_initz_lock = KMP_BOOTSTRAP_LOCK_INITIALIZER( __kmp_initz_lock ); /* Control initializations */ |
| 405 | kmp_bootstrap_lock_t __kmp_forkjoin_lock; /* control fork/join access */ |
| 406 | kmp_bootstrap_lock_t __kmp_exit_lock; /* exit() is not always thread-safe */ |
| 407 | kmp_bootstrap_lock_t __kmp_monitor_lock; /* control monitor thread creation */ |
| 408 | kmp_bootstrap_lock_t __kmp_tp_cached_lock; /* used for the hack to allow threadprivate cache and __kmp_threads expansion to co-exist */ |
| 409 | |
| 410 | KMP_ALIGN(128) |
| 411 | kmp_lock_t __kmp_global_lock; /* Control OS/global access */ |
| 412 | KMP_ALIGN(128) |
| 413 | kmp_queuing_lock_t __kmp_dispatch_lock; /* Control dispatch access */ |
| 414 | KMP_ALIGN(128) |
| 415 | kmp_lock_t __kmp_debug_lock; /* Control I/O access for KMP_DEBUG */ |
Jim Cownie | 4cc4bb4 | 2014-10-07 16:25:50 +0000 | [diff] [blame] | 416 | #endif |
Jim Cownie | 5e8470a | 2013-09-27 10:38:44 +0000 | [diff] [blame] | 417 | |
| 418 | /* ----------------------------------------------- */ |
| 419 | |
| 420 | #if KMP_HANDLE_SIGNALS |
| 421 | /* |
| 422 | Signal handling is disabled by default, because it confuses users: In case of sigsegv |
| 423 | (or other trouble) in user code signal handler catches the signal, which then "appears" in |
| 424 | the monitor thread (when the monitor executes raise() function). Users see signal in the |
| 425 | monitor thread and blame OpenMP RTL. |
| 426 | |
| 427 | Grant said signal handling required on some older OSes (Irix?) supported by KAI, because |
| 428 | bad applications hung but not aborted. Currently it is not a problem for Linux* OS, OS X* and |
| 429 | Windows* OS. |
| 430 | |
| 431 | Grant: Found new hangs for EL4, EL5, and a Fedora Core machine. So I'm putting |
| 432 | the default back for now to see if that fixes hangs on those machines. |
| 433 | |
| 434 | 2010-04013 Lev: It was a bug in Fortran RTL. Fortran RTL prints a kind of stack backtrace |
| 435 | when program is aborting, but the code is not signal-safe. When multiple signals raised at |
| 436 | the same time (which occurs in dynamic negative tests because all the worker threads detects |
| 437 | the same error), Fortran RTL may hang. The bug finally fixed in Fortran RTL library provided |
| 438 | by Steve R., and will be available soon. |
| 439 | */ |
| 440 | int __kmp_handle_signals = FALSE; |
| 441 | #endif |
| 442 | |
| 443 | /* ----------------------------------------------- */ |
| 444 | #ifdef BUILD_TV |
| 445 | kmp_key_t __kmp_tv_key = 0; |
| 446 | #endif |
| 447 | |
| 448 | /* ------------------------------------------------------------------------ */ |
| 449 | /* ------------------------------------------------------------------------ */ |
| 450 | |
| 451 | #ifdef DEBUG_SUSPEND |
| 452 | int |
| 453 | get_suspend_count_( void ) { |
| 454 | int count = __kmp_suspend_count; |
| 455 | __kmp_suspend_count = 0; |
| 456 | return count; |
| 457 | } |
| 458 | void |
| 459 | set_suspend_count_( int * value ) { |
| 460 | __kmp_suspend_count = *value; |
| 461 | } |
| 462 | #endif |
| 463 | |
| 464 | // Symbols for MS mutual detection. |
| 465 | int _You_must_link_with_exactly_one_OpenMP_library = 1; |
| 466 | int _You_must_link_with_Intel_OpenMP_library = 1; |
| 467 | #if KMP_OS_WINDOWS && ( KMP_VERSION_MAJOR > 4 ) |
| 468 | int _You_must_link_with_Microsoft_OpenMP_library = 1; |
| 469 | #endif |
| 470 | |
| 471 | // end of file // |