14 #include "kmp_affinity.h"
15 #if KMP_USE_HIER_SCHED
16 #include "kmp_dispatch_hier.h"
19 kmp_key_t __kmp_gtid_threadprivate_key;
21 #if KMP_ARCH_X86 || KMP_ARCH_X86_64
22 kmp_cpuinfo_t __kmp_cpuinfo = {0};
26 #include "kmp_stats.h"
28 kmp_tas_lock_t __kmp_stats_lock;
32 kmp_stats_list *__kmp_stats_list;
35 KMP_THREAD_LOCAL kmp_stats_list *__kmp_stats_thread_ptr = NULL;
38 tsc_tick_count __kmp_stats_start_time;
44 volatile int __kmp_init_serial = FALSE;
45 volatile int __kmp_init_gtid = FALSE;
46 volatile int __kmp_init_common = FALSE;
47 volatile int __kmp_init_middle = FALSE;
48 volatile int __kmp_init_parallel = FALSE;
49 volatile int __kmp_init_hidden_helper = FALSE;
50 volatile int __kmp_init_hidden_helper_threads = FALSE;
51 volatile int __kmp_hidden_helper_team_done = FALSE;
53 volatile int __kmp_init_monitor =
56 volatile int __kmp_init_user_locks = FALSE;
59 kmp_cached_addr_t *__kmp_threadpriv_cache_list = NULL;
61 int __kmp_init_counter = 0;
62 int __kmp_root_counter = 0;
63 int __kmp_version = 0;
65 std::atomic<kmp_int32> __kmp_team_counter = ATOMIC_VAR_INIT(0);
66 std::atomic<kmp_int32> __kmp_task_counter = ATOMIC_VAR_INIT(0);
68 size_t __kmp_stksize = KMP_DEFAULT_STKSIZE;
70 size_t __kmp_monitor_stksize = 0;
72 size_t __kmp_stkoffset = KMP_DEFAULT_STKOFFSET;
73 int __kmp_stkpadding = KMP_MIN_STKPADDING;
75 size_t __kmp_malloc_pool_incr = KMP_DEFAULT_MALLOC_POOL_INCR;
79 kmp_uint32 __kmp_barrier_gather_bb_dflt = 2;
81 kmp_uint32 __kmp_barrier_release_bb_dflt = 2;
84 kmp_bar_pat_e __kmp_barrier_gather_pat_dflt = bp_hyper_bar;
86 kmp_bar_pat_e __kmp_barrier_release_pat_dflt = bp_hyper_bar;
89 kmp_uint32 __kmp_barrier_gather_branch_bits[bs_last_barrier] = {0};
90 kmp_uint32 __kmp_barrier_release_branch_bits[bs_last_barrier] = {0};
91 kmp_bar_pat_e __kmp_barrier_gather_pattern[bs_last_barrier] = {bp_linear_bar};
92 kmp_bar_pat_e __kmp_barrier_release_pattern[bs_last_barrier] = {bp_linear_bar};
93 char const *__kmp_barrier_branch_bit_env_name[bs_last_barrier] = {
94 "KMP_PLAIN_BARRIER",
"KMP_FORKJOIN_BARRIER"
95 #if KMP_FAST_REDUCTION_BARRIER
97 "KMP_REDUCTION_BARRIER"
100 char const *__kmp_barrier_pattern_env_name[bs_last_barrier] = {
101 "KMP_PLAIN_BARRIER_PATTERN",
"KMP_FORKJOIN_BARRIER_PATTERN"
102 #if KMP_FAST_REDUCTION_BARRIER
104 "KMP_REDUCTION_BARRIER_PATTERN"
107 char const *__kmp_barrier_type_name[bs_last_barrier] = {
"plain",
"forkjoin"
108 #if KMP_FAST_REDUCTION_BARRIER
113 char const *__kmp_barrier_pattern_name[bp_last_bar] = {
114 "linear",
"tree",
"hyper",
"hierarchical",
"dist"};
116 int __kmp_allThreadsSpecified = 0;
117 size_t __kmp_align_alloc = CACHE_LINE;
119 int __kmp_generate_warnings = kmp_warnings_low;
120 int __kmp_reserve_warn = 0;
122 int __kmp_avail_proc = 0;
123 size_t __kmp_sys_min_stksize = KMP_MIN_STKSIZE;
124 int __kmp_sys_max_nth = KMP_MAX_NTH;
125 int __kmp_max_nth = 0;
126 int __kmp_cg_max_nth = 0;
127 int __kmp_teams_max_nth = 0;
128 int __kmp_threads_capacity = 0;
129 int __kmp_dflt_team_nth = 0;
130 int __kmp_dflt_team_nth_ub = 0;
131 int __kmp_tp_capacity = 0;
132 int __kmp_tp_cached = 0;
133 int __kmp_dispatch_num_buffers = KMP_DFLT_DISP_NUM_BUFF;
134 int __kmp_dflt_max_active_levels = 1;
135 bool __kmp_dflt_max_active_levels_set =
false;
136 #if KMP_NESTED_HOT_TEAMS
137 int __kmp_hot_teams_mode = 0;
139 int __kmp_hot_teams_max_level = 1;
141 enum library_type __kmp_library = library_none;
145 kmp_sch_static_greedy;
147 kmp_sch_guided_iterative_chunked;
149 kmp_sch_guided_analytical_chunked;
150 #if KMP_USE_HIER_SCHED
151 int __kmp_dispatch_hand_threading = 0;
152 int __kmp_hier_max_units[kmp_hier_layer_e::LAYER_LAST + 1];
153 int __kmp_hier_threads_per[kmp_hier_layer_e::LAYER_LAST + 1];
154 kmp_hier_sched_env_t __kmp_hier_scheds = {0, 0, NULL, NULL, NULL};
156 int __kmp_dflt_blocktime = KMP_DEFAULT_BLOCKTIME;
158 int __kmp_monitor_wakeups = KMP_MIN_MONITOR_WAKEUPS;
159 int __kmp_bt_intervals = KMP_INTERVALS_FROM_BLOCKTIME(KMP_DEFAULT_BLOCKTIME,
160 KMP_MIN_MONITOR_WAKEUPS);
162 #ifdef KMP_ADJUST_BLOCKTIME
163 int __kmp_zero_bt = FALSE;
165 #ifdef KMP_DFLT_NTH_CORES
166 int __kmp_ncores = 0;
169 int __kmp_force_monotonic = 0;
170 int __kmp_abort_delay = 0;
171 #if KMP_OS_LINUX && defined(KMP_TDATA_GTID)
172 int __kmp_gtid_mode = 3;
173 int __kmp_adjust_gtid_mode = FALSE;
175 int __kmp_gtid_mode = 2;
176 int __kmp_adjust_gtid_mode = FALSE;
178 int __kmp_gtid_mode = 0;
179 int __kmp_adjust_gtid_mode = TRUE;
181 #ifdef KMP_TDATA_GTID
182 KMP_THREAD_LOCAL
int __kmp_gtid = KMP_GTID_DNE;
184 int __kmp_tls_gtid_min = INT_MAX;
185 int __kmp_foreign_tp = TRUE;
186 #if KMP_ARCH_X86 || KMP_ARCH_X86_64
187 int __kmp_inherit_fp_control = TRUE;
188 kmp_int16 __kmp_init_x87_fpu_control_word = 0;
189 kmp_uint32 __kmp_init_mxcsr = 0;
192 #ifdef USE_LOAD_BALANCE
193 double __kmp_load_balance_interval = 1.0;
196 kmp_nested_nthreads_t __kmp_nested_nth = {NULL, 0, 0};
198 #if KMP_USE_ADAPTIVE_LOCKS
200 kmp_adaptive_backoff_params_t __kmp_adaptive_backoff_params = {
203 #if KMP_DEBUG_ADAPTIVE_LOCKS
204 const char *__kmp_speculative_statsfile =
"-";
209 int __kmp_display_env = FALSE;
210 int __kmp_display_env_verbose = FALSE;
211 int __kmp_omp_cancellation = FALSE;
212 int __kmp_nteams = 0;
213 int __kmp_teams_thread_limit = 0;
215 #if KMP_HAVE_MWAIT || KMP_HAVE_UMWAIT
216 int __kmp_user_level_mwait = FALSE;
217 int __kmp_umwait_enabled = FALSE;
218 int __kmp_mwait_enabled = FALSE;
219 int __kmp_mwait_hints = 0;
223 enum sched_type __kmp_sch_map[kmp_sched_upper - kmp_sched_lower_ext +
224 kmp_sched_upper_std - kmp_sched_lower - 2] = {
225 kmp_sch_static_chunked,
226 kmp_sch_dynamic_chunked,
235 enum clock_function_type __kmp_clock_function;
236 int __kmp_clock_function_param;
239 #if KMP_MIC_SUPPORTED
240 enum mic_type __kmp_mic_type = non_mic;
243 #if KMP_AFFINITY_SUPPORTED
245 KMPAffinity *__kmp_affinity_dispatch = NULL;
248 int __kmp_hwloc_error = FALSE;
249 hwloc_topology_t __kmp_hwloc_topology = NULL;
253 #if KMP_GROUP_AFFINITY
254 int __kmp_num_proc_groups = 1;
256 kmp_GetActiveProcessorCount_t __kmp_GetActiveProcessorCount = NULL;
257 kmp_GetActiveProcessorGroupCount_t __kmp_GetActiveProcessorGroupCount = NULL;
258 kmp_GetThreadGroupAffinity_t __kmp_GetThreadGroupAffinity = NULL;
259 kmp_SetThreadGroupAffinity_t __kmp_SetThreadGroupAffinity = NULL;
262 size_t __kmp_affin_mask_size = 0;
263 enum affinity_type __kmp_affinity_type = affinity_default;
264 kmp_hw_t __kmp_affinity_gran = KMP_HW_UNKNOWN;
265 int __kmp_affinity_gran_levels = -1;
266 int __kmp_affinity_dups = TRUE;
267 enum affinity_top_method __kmp_affinity_top_method =
268 affinity_top_method_default;
269 int __kmp_affinity_compact = 0;
270 int __kmp_affinity_offset = 0;
271 int __kmp_affinity_verbose = FALSE;
272 int __kmp_affinity_warnings = TRUE;
273 int __kmp_affinity_respect_mask = affinity_respect_mask_default;
274 char *__kmp_affinity_proclist = NULL;
275 kmp_affin_mask_t *__kmp_affinity_masks = NULL;
276 unsigned __kmp_affinity_num_masks = 0;
278 char *__kmp_cpuinfo_file = NULL;
282 kmp_nested_proc_bind_t __kmp_nested_proc_bind = {NULL, 0, 0};
283 int __kmp_affinity_num_places = 0;
284 int __kmp_display_affinity = FALSE;
285 char *__kmp_affinity_format = NULL;
287 kmp_int32 __kmp_default_device = 0;
289 kmp_tasking_mode_t __kmp_tasking_mode = tskm_task_teams;
290 kmp_int32 __kmp_max_task_priority = 0;
291 kmp_uint64 __kmp_taskloop_min_tasks = 0;
293 int __kmp_memkind_available = 0;
294 omp_allocator_handle_t
const omp_null_allocator = NULL;
295 omp_allocator_handle_t
const omp_default_mem_alloc =
296 (omp_allocator_handle_t
const)1;
297 omp_allocator_handle_t
const omp_large_cap_mem_alloc =
298 (omp_allocator_handle_t
const)2;
299 omp_allocator_handle_t
const omp_const_mem_alloc =
300 (omp_allocator_handle_t
const)3;
301 omp_allocator_handle_t
const omp_high_bw_mem_alloc =
302 (omp_allocator_handle_t
const)4;
303 omp_allocator_handle_t
const omp_low_lat_mem_alloc =
304 (omp_allocator_handle_t
const)5;
305 omp_allocator_handle_t
const omp_cgroup_mem_alloc =
306 (omp_allocator_handle_t
const)6;
307 omp_allocator_handle_t
const omp_pteam_mem_alloc =
308 (omp_allocator_handle_t
const)7;
309 omp_allocator_handle_t
const omp_thread_mem_alloc =
310 (omp_allocator_handle_t
const)8;
312 omp_allocator_handle_t
const llvm_omp_target_host_mem_alloc =
313 (omp_allocator_handle_t
const)100;
314 omp_allocator_handle_t
const llvm_omp_target_shared_mem_alloc =
315 (omp_allocator_handle_t
const)101;
316 omp_allocator_handle_t
const llvm_omp_target_device_mem_alloc =
317 (omp_allocator_handle_t
const)102;
318 omp_allocator_handle_t
const kmp_max_mem_alloc =
319 (omp_allocator_handle_t
const)1024;
320 omp_allocator_handle_t __kmp_def_allocator = omp_default_mem_alloc;
322 omp_memspace_handle_t
const omp_default_mem_space =
323 (omp_memspace_handle_t
const)0;
324 omp_memspace_handle_t
const omp_large_cap_mem_space =
325 (omp_memspace_handle_t
const)1;
326 omp_memspace_handle_t
const omp_const_mem_space =
327 (omp_memspace_handle_t
const)2;
328 omp_memspace_handle_t
const omp_high_bw_mem_space =
329 (omp_memspace_handle_t
const)3;
330 omp_memspace_handle_t
const omp_low_lat_mem_space =
331 (omp_memspace_handle_t
const)4;
333 omp_memspace_handle_t
const llvm_omp_target_host_mem_space =
334 (omp_memspace_handle_t
const)100;
335 omp_memspace_handle_t
const llvm_omp_target_shared_mem_space =
336 (omp_memspace_handle_t
const)101;
337 omp_memspace_handle_t
const llvm_omp_target_device_mem_space =
338 (omp_memspace_handle_t
const)102;
345 KMP_BUILD_ASSERT(
sizeof(kmp_tasking_flags_t) == 4);
347 int __kmp_task_stealing_constraint = 1;
348 int __kmp_enable_task_throttling = 1;
351 int __kmp_suspend_count = 0;
354 int __kmp_settings = FALSE;
355 int __kmp_duplicate_library_ok = 0;
357 int __kmp_forkjoin_frames = 1;
358 int __kmp_forkjoin_frames_mode = 3;
360 PACKED_REDUCTION_METHOD_T __kmp_force_reduction_method =
361 reduction_method_not_defined;
362 int __kmp_determ_red = FALSE;
375 int __kmp_debug_buf =
377 int __kmp_debug_buf_lines =
378 KMP_DEBUG_BUF_LINES_INIT;
379 int __kmp_debug_buf_chars =
380 KMP_DEBUG_BUF_CHARS_INIT;
381 int __kmp_debug_buf_atomic =
384 char *__kmp_debug_buffer = NULL;
385 std::atomic<int> __kmp_debug_count =
387 int __kmp_debug_buf_warn_chars =
394 char __kmp_par_range_routine[KMP_PAR_RANGE_ROUTINE_LEN] = {
'\0'};
395 char __kmp_par_range_filename[KMP_PAR_RANGE_FILENAME_LEN] = {
'\0'};
396 int __kmp_par_range_lb = 0;
397 int __kmp_par_range_ub = INT_MAX;
401 int __kmp_storage_map =
403 int __kmp_storage_map_verbose =
405 int __kmp_storage_map_verbose_specified = FALSE;
408 int __kmp_need_register_atfork =
410 int __kmp_need_register_atfork_specified = TRUE;
412 int __kmp_env_stksize = FALSE;
413 int __kmp_env_blocktime = FALSE;
414 int __kmp_env_checks = FALSE;
415 int __kmp_env_consistency_check = FALSE;
421 kmp_int32 __kmp_use_yield = 1;
423 kmp_int32 __kmp_use_yield_exp_set = 0;
425 kmp_uint32 __kmp_yield_init = KMP_INIT_WAIT;
426 kmp_uint32 __kmp_yield_next = KMP_NEXT_WAIT;
435 kmp_info_t **__kmp_threads = NULL;
436 kmp_root_t **__kmp_root = NULL;
440 volatile int __kmp_nth = 0;
441 volatile int __kmp_all_nth = 0;
442 volatile kmp_info_t *__kmp_thread_pool = NULL;
443 volatile kmp_team_t *__kmp_team_pool = NULL;
446 std::atomic<int> __kmp_thread_pool_active_nth = ATOMIC_VAR_INIT(0);
451 kmp_global_t __kmp_global;
457 #if KMP_USE_INTERNODE_ALIGNMENT
460 KMP_ALIGN_CACHE_INTERNODE
462 KMP_BOOTSTRAP_LOCK_INIT(__kmp_initz_lock);
463 KMP_ALIGN_CACHE_INTERNODE
464 KMP_BOOTSTRAP_LOCK_INIT(__kmp_forkjoin_lock);
465 KMP_ALIGN_CACHE_INTERNODE
466 KMP_BOOTSTRAP_LOCK_INIT(__kmp_exit_lock);
469 KMP_ALIGN_CACHE_INTERNODE
470 KMP_BOOTSTRAP_LOCK_INIT(__kmp_monitor_lock);
474 KMP_ALIGN_CACHE_INTERNODE
475 KMP_BOOTSTRAP_LOCK_INIT(__kmp_tp_cached_lock);
477 KMP_ALIGN_CACHE_INTERNODE
478 KMP_LOCK_INIT(__kmp_global_lock);
479 KMP_ALIGN_CACHE_INTERNODE
480 kmp_queuing_lock_t __kmp_dispatch_lock;
481 KMP_ALIGN_CACHE_INTERNODE
482 KMP_LOCK_INIT(__kmp_debug_lock);
486 KMP_BOOTSTRAP_LOCK_INIT(__kmp_initz_lock);
487 KMP_BOOTSTRAP_LOCK_INIT(__kmp_forkjoin_lock);
488 KMP_BOOTSTRAP_LOCK_INIT(__kmp_exit_lock);
491 KMP_BOOTSTRAP_LOCK_INIT(__kmp_monitor_lock);
495 KMP_BOOTSTRAP_LOCK_INIT(__kmp_tp_cached_lock);
498 KMP_LOCK_INIT(__kmp_global_lock);
500 kmp_queuing_lock_t __kmp_dispatch_lock;
502 KMP_LOCK_INIT(__kmp_debug_lock);
507 #if KMP_HANDLE_SIGNALS
527 int __kmp_handle_signals = FALSE;
531 int get_suspend_count_(
void) {
532 int count = __kmp_suspend_count;
533 __kmp_suspend_count = 0;
536 void set_suspend_count_(
int *value) { __kmp_suspend_count = *value; }
540 int _You_must_link_with_exactly_one_OpenMP_library = 1;
541 int _You_must_link_with_Intel_OpenMP_library = 1;
542 #if KMP_OS_WINDOWS && (KMP_VERSION_MAJOR > 4)
543 int _You_must_link_with_Microsoft_OpenMP_library = 1;
546 kmp_target_offload_kind_t __kmp_target_offload = tgt_default;
549 kmp_pause_status_t __kmp_pause_status = kmp_not_paused;
552 int __kmp_nesting_mode = 0;
553 int __kmp_nesting_mode_nlevels = 1;
554 int *__kmp_nesting_nth_level;