14 #include "kmp_affinity.h"
15 #if KMP_USE_HIER_SCHED
16 #include "kmp_dispatch_hier.h"
19 kmp_key_t __kmp_gtid_threadprivate_key;
21 #if KMP_ARCH_X86 || KMP_ARCH_X86_64
22 kmp_cpuinfo_t __kmp_cpuinfo = {0};
26 #include "kmp_stats.h"
28 kmp_tas_lock_t __kmp_stats_lock;
32 kmp_stats_list *__kmp_stats_list;
35 KMP_THREAD_LOCAL kmp_stats_list *__kmp_stats_thread_ptr = NULL;
38 tsc_tick_count __kmp_stats_start_time;
44 volatile int __kmp_init_serial = FALSE;
45 volatile int __kmp_init_gtid = FALSE;
46 volatile int __kmp_init_common = FALSE;
47 volatile int __kmp_init_middle = FALSE;
48 volatile int __kmp_init_parallel = FALSE;
50 volatile int __kmp_init_monitor =
53 volatile int __kmp_init_user_locks = FALSE;
56 kmp_cached_addr_t *__kmp_threadpriv_cache_list = NULL;
58 int __kmp_init_counter = 0;
59 int __kmp_root_counter = 0;
60 int __kmp_version = 0;
62 std::atomic<kmp_int32> __kmp_team_counter = ATOMIC_VAR_INIT(0);
63 std::atomic<kmp_int32> __kmp_task_counter = ATOMIC_VAR_INIT(0);
65 unsigned int __kmp_init_wait =
66 KMP_DEFAULT_INIT_WAIT;
67 unsigned int __kmp_next_wait =
68 KMP_DEFAULT_NEXT_WAIT;
70 size_t __kmp_stksize = KMP_DEFAULT_STKSIZE;
72 size_t __kmp_monitor_stksize = 0;
74 size_t __kmp_stkoffset = KMP_DEFAULT_STKOFFSET;
75 int __kmp_stkpadding = KMP_MIN_STKPADDING;
77 size_t __kmp_malloc_pool_incr = KMP_DEFAULT_MALLOC_POOL_INCR;
81 kmp_uint32 __kmp_barrier_gather_bb_dflt = 2;
83 kmp_uint32 __kmp_barrier_release_bb_dflt = 2;
86 kmp_bar_pat_e __kmp_barrier_gather_pat_dflt = bp_hyper_bar;
88 kmp_bar_pat_e __kmp_barrier_release_pat_dflt = bp_hyper_bar;
91 kmp_uint32 __kmp_barrier_gather_branch_bits[bs_last_barrier] = {0};
92 kmp_uint32 __kmp_barrier_release_branch_bits[bs_last_barrier] = {0};
93 kmp_bar_pat_e __kmp_barrier_gather_pattern[bs_last_barrier] = {bp_linear_bar};
94 kmp_bar_pat_e __kmp_barrier_release_pattern[bs_last_barrier] = {bp_linear_bar};
95 char const *__kmp_barrier_branch_bit_env_name[bs_last_barrier] = {
96 "KMP_PLAIN_BARRIER",
"KMP_FORKJOIN_BARRIER"
97 #if KMP_FAST_REDUCTION_BARRIER
99 "KMP_REDUCTION_BARRIER"
100 #endif // KMP_FAST_REDUCTION_BARRIER
102 char const *__kmp_barrier_pattern_env_name[bs_last_barrier] = {
103 "KMP_PLAIN_BARRIER_PATTERN",
"KMP_FORKJOIN_BARRIER_PATTERN"
104 #if KMP_FAST_REDUCTION_BARRIER
106 "KMP_REDUCTION_BARRIER_PATTERN"
107 #endif // KMP_FAST_REDUCTION_BARRIER
109 char const *__kmp_barrier_type_name[bs_last_barrier] = {
"plain",
"forkjoin"
110 #if KMP_FAST_REDUCTION_BARRIER
113 #endif // KMP_FAST_REDUCTION_BARRIER
115 char const *__kmp_barrier_pattern_name[bp_last_bar] = {
"linear",
"tree",
116 "hyper",
"hierarchical"};
118 int __kmp_allThreadsSpecified = 0;
119 size_t __kmp_align_alloc = CACHE_LINE;
121 int __kmp_generate_warnings = kmp_warnings_low;
122 int __kmp_reserve_warn = 0;
124 int __kmp_avail_proc = 0;
125 size_t __kmp_sys_min_stksize = KMP_MIN_STKSIZE;
126 int __kmp_sys_max_nth = KMP_MAX_NTH;
127 int __kmp_max_nth = 0;
128 int __kmp_cg_max_nth = 0;
129 int __kmp_teams_max_nth = 0;
130 int __kmp_threads_capacity = 0;
131 int __kmp_dflt_team_nth = 0;
132 int __kmp_dflt_team_nth_ub = 0;
133 int __kmp_tp_capacity = 0;
134 int __kmp_tp_cached = 0;
135 int __kmp_dflt_nested = FALSE;
136 int __kmp_dispatch_num_buffers = KMP_DFLT_DISP_NUM_BUFF;
137 int __kmp_dflt_max_active_levels =
138 KMP_MAX_ACTIVE_LEVELS_LIMIT;
139 #if KMP_NESTED_HOT_TEAMS
140 int __kmp_hot_teams_mode = 0;
142 int __kmp_hot_teams_max_level = 1;
144 enum library_type __kmp_library = library_none;
148 kmp_sch_static_greedy;
150 kmp_sch_guided_iterative_chunked;
152 kmp_sch_guided_analytical_chunked;
153 #if KMP_USE_HIER_SCHED
154 int __kmp_dispatch_hand_threading = 0;
155 int __kmp_hier_max_units[kmp_hier_layer_e::LAYER_LAST + 1];
156 int __kmp_hier_threads_per[kmp_hier_layer_e::LAYER_LAST + 1];
157 kmp_hier_sched_env_t __kmp_hier_scheds = {0, 0, NULL, NULL, NULL};
159 int __kmp_dflt_blocktime = KMP_DEFAULT_BLOCKTIME;
161 int __kmp_monitor_wakeups = KMP_MIN_MONITOR_WAKEUPS;
162 int __kmp_bt_intervals = KMP_INTERVALS_FROM_BLOCKTIME(KMP_DEFAULT_BLOCKTIME,
163 KMP_MIN_MONITOR_WAKEUPS);
165 #ifdef KMP_ADJUST_BLOCKTIME
166 int __kmp_zero_bt = FALSE;
168 #ifdef KMP_DFLT_NTH_CORES
169 int __kmp_ncores = 0;
172 int __kmp_abort_delay = 0;
173 #if KMP_OS_LINUX && defined(KMP_TDATA_GTID)
174 int __kmp_gtid_mode = 3;
175 int __kmp_adjust_gtid_mode = FALSE;
177 int __kmp_gtid_mode = 2;
178 int __kmp_adjust_gtid_mode = FALSE;
180 int __kmp_gtid_mode = 0;
181 int __kmp_adjust_gtid_mode = TRUE;
183 #ifdef KMP_TDATA_GTID
184 KMP_THREAD_LOCAL
int __kmp_gtid = KMP_GTID_DNE;
186 int __kmp_tls_gtid_min = INT_MAX;
187 int __kmp_foreign_tp = TRUE;
188 #if KMP_ARCH_X86 || KMP_ARCH_X86_64
189 int __kmp_inherit_fp_control = TRUE;
190 kmp_int16 __kmp_init_x87_fpu_control_word = 0;
191 kmp_uint32 __kmp_init_mxcsr = 0;
194 #ifdef USE_LOAD_BALANCE
195 double __kmp_load_balance_interval = 1.0;
198 kmp_nested_nthreads_t __kmp_nested_nth = {NULL, 0, 0};
200 #if KMP_USE_ADAPTIVE_LOCKS
202 kmp_adaptive_backoff_params_t __kmp_adaptive_backoff_params = {
205 #if KMP_DEBUG_ADAPTIVE_LOCKS
206 const char *__kmp_speculative_statsfile =
"-";
209 #endif // KMP_USE_ADAPTIVE_LOCKS
212 int __kmp_display_env = FALSE;
213 int __kmp_display_env_verbose = FALSE;
214 int __kmp_omp_cancellation = FALSE;
218 enum sched_type __kmp_sch_map[kmp_sched_upper - kmp_sched_lower_ext +
219 kmp_sched_upper_std - kmp_sched_lower - 2] = {
220 kmp_sch_static_chunked,
221 kmp_sch_dynamic_chunked,
230 enum clock_function_type __kmp_clock_function;
231 int __kmp_clock_function_param;
234 #if KMP_MIC_SUPPORTED
235 enum mic_type __kmp_mic_type = non_mic;
238 #if KMP_AFFINITY_SUPPORTED
240 KMPAffinity *__kmp_affinity_dispatch = NULL;
243 int __kmp_hwloc_error = FALSE;
244 hwloc_topology_t __kmp_hwloc_topology = NULL;
245 int __kmp_numa_detected = FALSE;
246 int __kmp_tile_depth = 0;
250 #if KMP_GROUP_AFFINITY
251 int __kmp_num_proc_groups = 1;
253 kmp_GetActiveProcessorCount_t __kmp_GetActiveProcessorCount = NULL;
254 kmp_GetActiveProcessorGroupCount_t __kmp_GetActiveProcessorGroupCount = NULL;
255 kmp_GetThreadGroupAffinity_t __kmp_GetThreadGroupAffinity = NULL;
256 kmp_SetThreadGroupAffinity_t __kmp_SetThreadGroupAffinity = NULL;
259 size_t __kmp_affin_mask_size = 0;
260 enum affinity_type __kmp_affinity_type = affinity_default;
261 enum affinity_gran __kmp_affinity_gran = affinity_gran_default;
262 int __kmp_affinity_gran_levels = -1;
263 int __kmp_affinity_dups = TRUE;
264 enum affinity_top_method __kmp_affinity_top_method =
265 affinity_top_method_default;
266 int __kmp_affinity_compact = 0;
267 int __kmp_affinity_offset = 0;
268 int __kmp_affinity_verbose = FALSE;
269 int __kmp_affinity_warnings = TRUE;
270 int __kmp_affinity_respect_mask = affinity_respect_mask_default;
271 char *__kmp_affinity_proclist = NULL;
272 kmp_affin_mask_t *__kmp_affinity_masks = NULL;
273 unsigned __kmp_affinity_num_masks = 0;
275 char *__kmp_cpuinfo_file = NULL;
280 kmp_nested_proc_bind_t __kmp_nested_proc_bind = {NULL, 0, 0};
281 int __kmp_affinity_num_places = 0;
285 int __kmp_display_affinity = FALSE;
286 char *__kmp_affinity_format = NULL;
287 #endif // OMP_50_ENABLED
289 kmp_hws_item_t __kmp_hws_socket = {0, 0};
290 kmp_hws_item_t __kmp_hws_node = {0, 0};
291 kmp_hws_item_t __kmp_hws_tile = {0, 0};
292 kmp_hws_item_t __kmp_hws_core = {0, 0};
293 kmp_hws_item_t __kmp_hws_proc = {0, 0};
294 int __kmp_hws_requested = 0;
295 int __kmp_hws_abs_flag = 0;
298 kmp_int32 __kmp_default_device = 0;
301 kmp_tasking_mode_t __kmp_tasking_mode = tskm_task_teams;
303 kmp_int32 __kmp_max_task_priority = 0;
304 kmp_uint64 __kmp_taskloop_min_tasks = 0;
308 int __kmp_memkind_available = 0;
309 int __kmp_hbw_mem_available = 0;
310 const omp_allocator_t *OMP_NULL_ALLOCATOR = NULL;
311 const omp_allocator_t *omp_default_mem_alloc = (
const omp_allocator_t *)1;
312 const omp_allocator_t *omp_large_cap_mem_alloc = (
const omp_allocator_t *)2;
313 const omp_allocator_t *omp_const_mem_alloc = (
const omp_allocator_t *)3;
314 const omp_allocator_t *omp_high_bw_mem_alloc = (
const omp_allocator_t *)4;
315 const omp_allocator_t *omp_low_lat_mem_alloc = (
const omp_allocator_t *)5;
316 const omp_allocator_t *omp_cgroup_mem_alloc = (
const omp_allocator_t *)6;
317 const omp_allocator_t *omp_pteam_mem_alloc = (
const omp_allocator_t *)7;
318 const omp_allocator_t *omp_thread_mem_alloc = (
const omp_allocator_t *)8;
319 void *
const *__kmp_def_allocator = omp_default_mem_alloc;
327 KMP_BUILD_ASSERT(
sizeof(kmp_tasking_flags_t) == 4);
329 int __kmp_task_stealing_constraint = 1;
332 int __kmp_suspend_count = 0;
335 int __kmp_settings = FALSE;
336 int __kmp_duplicate_library_ok = 0;
338 int __kmp_forkjoin_frames = 1;
339 int __kmp_forkjoin_frames_mode = 3;
341 PACKED_REDUCTION_METHOD_T __kmp_force_reduction_method =
342 reduction_method_not_defined;
343 int __kmp_determ_red = FALSE;
356 int __kmp_debug_buf =
358 int __kmp_debug_buf_lines =
359 KMP_DEBUG_BUF_LINES_INIT;
360 int __kmp_debug_buf_chars =
361 KMP_DEBUG_BUF_CHARS_INIT;
362 int __kmp_debug_buf_atomic =
365 char *__kmp_debug_buffer = NULL;
366 std::atomic<int> __kmp_debug_count =
368 int __kmp_debug_buf_warn_chars =
375 char __kmp_par_range_routine[KMP_PAR_RANGE_ROUTINE_LEN] = {
'\0'};
376 char __kmp_par_range_filename[KMP_PAR_RANGE_FILENAME_LEN] = {
'\0'};
377 int __kmp_par_range_lb = 0;
378 int __kmp_par_range_ub = INT_MAX;
382 int __kmp_storage_map =
384 int __kmp_storage_map_verbose =
386 int __kmp_storage_map_verbose_specified = FALSE;
389 int __kmp_need_register_atfork =
391 int __kmp_need_register_atfork_specified = TRUE;
393 int __kmp_env_stksize = FALSE;
394 int __kmp_env_blocktime = FALSE;
395 int __kmp_env_checks = FALSE;
396 int __kmp_env_consistency_check = FALSE;
398 kmp_uint32 __kmp_yield_init = KMP_INIT_WAIT;
399 kmp_uint32 __kmp_yield_next = KMP_NEXT_WAIT;
402 kmp_uint32 __kmp_yielding_on = 1;
405 kmp_uint32 __kmp_yield_cycle = 0;
407 kmp_uint32 __kmp_yield_cycle = 1;
409 kmp_int32 __kmp_yield_on_count =
411 kmp_int32 __kmp_yield_off_count =
421 kmp_info_t **__kmp_threads = NULL;
422 kmp_root_t **__kmp_root = NULL;
426 volatile int __kmp_nth = 0;
427 volatile int __kmp_all_nth = 0;
428 int __kmp_thread_pool_nth = 0;
429 volatile kmp_info_t *__kmp_thread_pool = NULL;
430 volatile kmp_team_t *__kmp_team_pool = NULL;
433 std::atomic<int> __kmp_thread_pool_active_nth = ATOMIC_VAR_INIT(0);
438 kmp_global_t __kmp_global = {{0}};
444 #if KMP_USE_INTERNODE_ALIGNMENT
447 KMP_ALIGN_CACHE_INTERNODE
449 KMP_BOOTSTRAP_LOCK_INIT(__kmp_initz_lock);
450 KMP_ALIGN_CACHE_INTERNODE
451 KMP_BOOTSTRAP_LOCK_INIT(__kmp_forkjoin_lock);
452 KMP_ALIGN_CACHE_INTERNODE
453 KMP_BOOTSTRAP_LOCK_INIT(__kmp_exit_lock);
456 KMP_ALIGN_CACHE_INTERNODE
457 KMP_BOOTSTRAP_LOCK_INIT(__kmp_monitor_lock);
461 KMP_ALIGN_CACHE_INTERNODE
462 KMP_BOOTSTRAP_LOCK_INIT(__kmp_tp_cached_lock);
464 KMP_ALIGN_CACHE_INTERNODE
465 KMP_LOCK_INIT(__kmp_global_lock);
466 KMP_ALIGN_CACHE_INTERNODE
467 kmp_queuing_lock_t __kmp_dispatch_lock;
468 KMP_ALIGN_CACHE_INTERNODE
469 KMP_LOCK_INIT(__kmp_debug_lock);
473 KMP_BOOTSTRAP_LOCK_INIT(__kmp_initz_lock);
474 KMP_BOOTSTRAP_LOCK_INIT(__kmp_forkjoin_lock);
475 KMP_BOOTSTRAP_LOCK_INIT(__kmp_exit_lock);
478 KMP_BOOTSTRAP_LOCK_INIT(__kmp_monitor_lock);
482 KMP_BOOTSTRAP_LOCK_INIT(__kmp_tp_cached_lock);
485 KMP_LOCK_INIT(__kmp_global_lock);
487 kmp_queuing_lock_t __kmp_dispatch_lock;
489 KMP_LOCK_INIT(__kmp_debug_lock);
494 #if KMP_HANDLE_SIGNALS
514 int __kmp_handle_signals = FALSE;
518 int get_suspend_count_(
void) {
519 int count = __kmp_suspend_count;
520 __kmp_suspend_count = 0;
523 void set_suspend_count_(
int *value) { __kmp_suspend_count = *value; }
527 int _You_must_link_with_exactly_one_OpenMP_library = 1;
528 int _You_must_link_with_Intel_OpenMP_library = 1;
529 #if KMP_OS_WINDOWS && (KMP_VERSION_MAJOR > 4)
530 int _You_must_link_with_Microsoft_OpenMP_library = 1;
534 kmp_target_offload_kind_t __kmp_target_offload = tgt_default;
537 kmp_pause_status_t __kmp_pause_status = kmp_not_paused;
538 #endif // OMP_50_ENABLED