mirror of
https://github.com/jart/cosmopolitan.git
synced 2025-01-31 03:27:39 +00:00
178 lines
7.3 KiB
C++
178 lines
7.3 KiB
C++
/*
|
|
* ompt-specific.h - header of OMPT internal functions implementation
|
|
*/
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
//
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#ifndef OMPT_SPECIFIC_H
|
|
#define OMPT_SPECIFIC_H
|
|
|
|
#include "kmp.h"
|
|
|
|
#if OMPT_SUPPORT
|
|
/*****************************************************************************
|
|
* forward declarations
|
|
****************************************************************************/
|
|
|
|
/// Entrypoint used by libomptarget to register callbacks in libomp, if not
|
|
/// done already
|
|
void __ompt_force_initialization();
|
|
|
|
void __ompt_team_assign_id(kmp_team_t *team, ompt_data_t ompt_pid);
|
|
void __ompt_thread_assign_wait_id(void *variable);
|
|
|
|
void __ompt_lw_taskteam_init(ompt_lw_taskteam_t *lwt, kmp_info_t *thr, int gtid,
|
|
ompt_data_t *ompt_pid, void *codeptr);
|
|
|
|
void __ompt_lw_taskteam_link(ompt_lw_taskteam_t *lwt, kmp_info_t *thr,
|
|
int on_heap, bool always = false);
|
|
|
|
void __ompt_lw_taskteam_unlink(kmp_info_t *thr);
|
|
|
|
ompt_team_info_t *__ompt_get_teaminfo(int depth, int *size);
|
|
|
|
ompt_data_t *__ompt_get_task_data();
|
|
|
|
ompt_data_t *__ompt_get_target_task_data();
|
|
|
|
ompt_task_info_t *__ompt_get_task_info_object(int depth);
|
|
|
|
int __ompt_get_parallel_info_internal(int ancestor_level,
|
|
ompt_data_t **parallel_data,
|
|
int *team_size);
|
|
|
|
int __ompt_get_task_info_internal(int ancestor_level, int *type,
|
|
ompt_data_t **task_data,
|
|
ompt_frame_t **task_frame,
|
|
ompt_data_t **parallel_data, int *thread_num);
|
|
|
|
ompt_data_t *__ompt_get_thread_data_internal();
|
|
|
|
/*
|
|
* Unused currently
|
|
static uint64_t __ompt_get_get_unique_id_internal();
|
|
*/
|
|
|
|
ompt_sync_region_t __ompt_get_barrier_kind(enum barrier_type, kmp_info_t *);
|
|
|
|
/*****************************************************************************
|
|
* macros
|
|
****************************************************************************/
|
|
|
|
#define OMPT_CUR_TASK_INFO(thr) (&((thr)->th.th_current_task->ompt_task_info))
|
|
#define OMPT_CUR_TASK_DATA(thr) \
|
|
(&((thr)->th.th_current_task->ompt_task_info.task_data))
|
|
#define OMPT_CUR_TEAM_INFO(thr) (&((thr)->th.th_team->t.ompt_team_info))
|
|
#define OMPT_CUR_TEAM_DATA(thr) \
|
|
(&((thr)->th.th_team->t.ompt_team_info.parallel_data))
|
|
|
|
#define OMPT_HAVE_WEAK_ATTRIBUTE KMP_HAVE_WEAK_ATTRIBUTE
|
|
#define OMPT_HAVE_PSAPI KMP_HAVE_PSAPI
|
|
#define OMPT_STR_MATCH(haystack, needle) __kmp_str_match(haystack, 0, needle)
|
|
|
|
inline void *__ompt_load_return_address(int gtid) {
|
|
kmp_info_t *thr = __kmp_threads[gtid];
|
|
void *return_address = thr->th.ompt_thread_info.return_address;
|
|
thr->th.ompt_thread_info.return_address = NULL;
|
|
return return_address;
|
|
}
|
|
|
|
/*#define OMPT_STORE_RETURN_ADDRESS(gtid) \
|
|
if (ompt_enabled.enabled && gtid >= 0 && __kmp_threads[gtid] && \
|
|
!__kmp_threads[gtid]->th.ompt_thread_info.return_address) \
|
|
__kmp_threads[gtid]->th.ompt_thread_info.return_address = \
|
|
__builtin_return_address(0)*/
|
|
#define OMPT_STORE_RETURN_ADDRESS(gtid) \
|
|
OmptReturnAddressGuard ReturnAddressGuard{gtid, __builtin_return_address(0)};
|
|
#define OMPT_LOAD_RETURN_ADDRESS(gtid) __ompt_load_return_address(gtid)
|
|
#define OMPT_LOAD_OR_GET_RETURN_ADDRESS(gtid) \
|
|
((ompt_enabled.enabled && gtid >= 0 && __kmp_threads[gtid] && \
|
|
__kmp_threads[gtid]->th.ompt_thread_info.return_address) \
|
|
? __ompt_load_return_address(gtid) \
|
|
: __builtin_return_address(0))
|
|
|
|
#define OMPT_GET_DISPATCH_CHUNK(chunk, lb, ub, incr) \
|
|
do { \
|
|
if (incr > 0) { \
|
|
chunk.start = static_cast<uint64_t>(lb); \
|
|
chunk.iterations = static_cast<uint64_t>(((ub) - (lb)) / (incr) + 1); \
|
|
} else { \
|
|
chunk.start = static_cast<uint64_t>(ub); \
|
|
chunk.iterations = static_cast<uint64_t>(((lb) - (ub)) / -(incr) + 1); \
|
|
} \
|
|
} while (0)
|
|
|
|
//******************************************************************************
|
|
// inline functions
|
|
//******************************************************************************
|
|
|
|
inline kmp_info_t *ompt_get_thread_gtid(int gtid) {
|
|
return (gtid >= 0) ? __kmp_thread_from_gtid(gtid) : NULL;
|
|
}
|
|
|
|
inline kmp_info_t *ompt_get_thread() {
|
|
int gtid = __kmp_get_gtid();
|
|
return ompt_get_thread_gtid(gtid);
|
|
}
|
|
|
|
inline void ompt_set_thread_state(kmp_info_t *thread, ompt_state_t state) {
|
|
if (thread)
|
|
thread->th.ompt_thread_info.state = state;
|
|
}
|
|
|
|
inline const char *ompt_get_runtime_version() {
|
|
return &__kmp_version_lib_ver[KMP_VERSION_MAGIC_LEN];
|
|
}
|
|
|
|
class OmptReturnAddressGuard {
|
|
private:
|
|
bool SetAddress{false};
|
|
int Gtid;
|
|
|
|
public:
|
|
OmptReturnAddressGuard(int Gtid, void *ReturnAddress) : Gtid(Gtid) {
|
|
if (ompt_enabled.enabled && Gtid >= 0 && __kmp_threads[Gtid] &&
|
|
!__kmp_threads[Gtid]->th.ompt_thread_info.return_address) {
|
|
SetAddress = true;
|
|
__kmp_threads[Gtid]->th.ompt_thread_info.return_address = ReturnAddress;
|
|
}
|
|
}
|
|
~OmptReturnAddressGuard() {
|
|
if (SetAddress)
|
|
__kmp_threads[Gtid]->th.ompt_thread_info.return_address = NULL;
|
|
}
|
|
};
|
|
|
|
#endif // OMPT_SUPPORT
|
|
|
|
// macros providing the OMPT callbacks for reduction clause
|
|
#if OMPT_SUPPORT && OMPT_OPTIONAL
|
|
#define OMPT_REDUCTION_DECL(this_thr, gtid) \
|
|
ompt_data_t *my_task_data = OMPT_CUR_TASK_DATA(this_thr); \
|
|
ompt_data_t *my_parallel_data = OMPT_CUR_TEAM_DATA(this_thr); \
|
|
void *return_address = OMPT_LOAD_RETURN_ADDRESS(gtid);
|
|
#define OMPT_REDUCTION_BEGIN \
|
|
if (ompt_enabled.enabled && ompt_enabled.ompt_callback_reduction) { \
|
|
ompt_callbacks.ompt_callback(ompt_callback_reduction)( \
|
|
ompt_sync_region_reduction, ompt_scope_begin, my_parallel_data, \
|
|
my_task_data, return_address); \
|
|
}
|
|
#define OMPT_REDUCTION_END \
|
|
if (ompt_enabled.enabled && ompt_enabled.ompt_callback_reduction) { \
|
|
ompt_callbacks.ompt_callback(ompt_callback_reduction)( \
|
|
ompt_sync_region_reduction, ompt_scope_end, my_parallel_data, \
|
|
my_task_data, return_address); \
|
|
}
|
|
#else // OMPT_SUPPORT && OMPT_OPTIONAL
|
|
#define OMPT_REDUCTION_DECL(this_thr, gtid)
|
|
#define OMPT_REDUCTION_BEGIN
|
|
#define OMPT_REDUCTION_END
|
|
#endif // ! OMPT_SUPPORT && OMPT_OPTIONAL
|
|
|
|
#endif
|