123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798 |
- /*
- * kmp_threadprivate.cpp -- OpenMP threadprivate support library
- */
- //===----------------------------------------------------------------------===//
- //
- // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
- // See https://llvm.org/LICENSE.txt for license information.
- // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
- //
- //===----------------------------------------------------------------------===//
- #include "kmp.h"
- #include "kmp_i18n.h"
- #include "kmp_itt.h"
- #define USE_CHECKS_COMMON
- #define KMP_INLINE_SUBR 1
- void kmp_threadprivate_insert_private_data(int gtid, void *pc_addr,
- void *data_addr, size_t pc_size);
- struct private_common *kmp_threadprivate_insert(int gtid, void *pc_addr,
- void *data_addr,
- size_t pc_size);
- struct shared_table __kmp_threadprivate_d_table;
- static
- #ifdef KMP_INLINE_SUBR
- __forceinline
- #endif
- struct private_common *
- __kmp_threadprivate_find_task_common(struct common_table *tbl, int gtid,
- void *pc_addr)
- {
- struct private_common *tn;
- #ifdef KMP_TASK_COMMON_DEBUG
- KC_TRACE(10, ("__kmp_threadprivate_find_task_common: thread#%d, called with "
- "address %p\n",
- gtid, pc_addr));
- dump_list();
- #endif
- for (tn = tbl->data[KMP_HASH(pc_addr)]; tn; tn = tn->next) {
- if (tn->gbl_addr == pc_addr) {
- #ifdef KMP_TASK_COMMON_DEBUG
- KC_TRACE(10, ("__kmp_threadprivate_find_task_common: thread#%d, found "
- "node %p on list\n",
- gtid, pc_addr));
- #endif
- return tn;
- }
- }
- return 0;
- }
- static
- #ifdef KMP_INLINE_SUBR
- __forceinline
- #endif
- struct shared_common *
- __kmp_find_shared_task_common(struct shared_table *tbl, int gtid,
- void *pc_addr) {
- struct shared_common *tn;
- for (tn = tbl->data[KMP_HASH(pc_addr)]; tn; tn = tn->next) {
- if (tn->gbl_addr == pc_addr) {
- #ifdef KMP_TASK_COMMON_DEBUG
- KC_TRACE(
- 10,
- ("__kmp_find_shared_task_common: thread#%d, found node %p on list\n",
- gtid, pc_addr));
- #endif
- return tn;
- }
- }
- return 0;
- }
- // Create a template for the data initialized storage. Either the template is
- // NULL indicating zero fill, or the template is a copy of the original data.
- static struct private_data *__kmp_init_common_data(void *pc_addr,
- size_t pc_size) {
- struct private_data *d;
- size_t i;
- char *p;
- d = (struct private_data *)__kmp_allocate(sizeof(struct private_data));
- /*
- d->data = 0; // AC: commented out because __kmp_allocate zeroes the
- memory
- d->next = 0;
- */
- d->size = pc_size;
- d->more = 1;
- p = (char *)pc_addr;
- for (i = pc_size; i > 0; --i) {
- if (*p++ != '\0') {
- d->data = __kmp_allocate(pc_size);
- KMP_MEMCPY(d->data, pc_addr, pc_size);
- break;
- }
- }
- return d;
- }
- // Initialize the data area from the template.
- static void __kmp_copy_common_data(void *pc_addr, struct private_data *d) {
- char *addr = (char *)pc_addr;
- for (size_t offset = 0; d != 0; d = d->next) {
- for (int i = d->more; i > 0; --i) {
- if (d->data == 0)
- memset(&addr[offset], '\0', d->size);
- else
- KMP_MEMCPY(&addr[offset], d->data, d->size);
- offset += d->size;
- }
- }
- }
- /* we are called from __kmp_serial_initialize() with __kmp_initz_lock held. */
- void __kmp_common_initialize(void) {
- if (!TCR_4(__kmp_init_common)) {
- int q;
- #ifdef KMP_DEBUG
- int gtid;
- #endif
- __kmp_threadpriv_cache_list = NULL;
- #ifdef KMP_DEBUG
- /* verify the uber masters were initialized */
- for (gtid = 0; gtid < __kmp_threads_capacity; gtid++)
- if (__kmp_root[gtid]) {
- KMP_DEBUG_ASSERT(__kmp_root[gtid]->r.r_uber_thread);
- for (q = 0; q < KMP_HASH_TABLE_SIZE; ++q)
- KMP_DEBUG_ASSERT(
- !__kmp_root[gtid]->r.r_uber_thread->th.th_pri_common->data[q]);
- /* __kmp_root[ gitd ]-> r.r_uber_thread ->
- * th.th_pri_common -> data[ q ] = 0;*/
- }
- #endif /* KMP_DEBUG */
- for (q = 0; q < KMP_HASH_TABLE_SIZE; ++q)
- __kmp_threadprivate_d_table.data[q] = 0;
- TCW_4(__kmp_init_common, TRUE);
- }
- }
- /* Call all destructors for threadprivate data belonging to all threads.
- Currently unused! */
- void __kmp_common_destroy(void) {
- if (TCR_4(__kmp_init_common)) {
- int q;
- TCW_4(__kmp_init_common, FALSE);
- for (q = 0; q < KMP_HASH_TABLE_SIZE; ++q) {
- int gtid;
- struct private_common *tn;
- struct shared_common *d_tn;
- /* C++ destructors need to be called once per thread before exiting.
- Don't call destructors for primary thread though unless we used copy
- constructor */
- for (d_tn = __kmp_threadprivate_d_table.data[q]; d_tn;
- d_tn = d_tn->next) {
- if (d_tn->is_vec) {
- if (d_tn->dt.dtorv != 0) {
- for (gtid = 0; gtid < __kmp_all_nth; ++gtid) {
- if (__kmp_threads[gtid]) {
- if ((__kmp_foreign_tp) ? (!KMP_INITIAL_GTID(gtid))
- : (!KMP_UBER_GTID(gtid))) {
- tn = __kmp_threadprivate_find_task_common(
- __kmp_threads[gtid]->th.th_pri_common, gtid,
- d_tn->gbl_addr);
- if (tn) {
- (*d_tn->dt.dtorv)(tn->par_addr, d_tn->vec_len);
- }
- }
- }
- }
- if (d_tn->obj_init != 0) {
- (*d_tn->dt.dtorv)(d_tn->obj_init, d_tn->vec_len);
- }
- }
- } else {
- if (d_tn->dt.dtor != 0) {
- for (gtid = 0; gtid < __kmp_all_nth; ++gtid) {
- if (__kmp_threads[gtid]) {
- if ((__kmp_foreign_tp) ? (!KMP_INITIAL_GTID(gtid))
- : (!KMP_UBER_GTID(gtid))) {
- tn = __kmp_threadprivate_find_task_common(
- __kmp_threads[gtid]->th.th_pri_common, gtid,
- d_tn->gbl_addr);
- if (tn) {
- (*d_tn->dt.dtor)(tn->par_addr);
- }
- }
- }
- }
- if (d_tn->obj_init != 0) {
- (*d_tn->dt.dtor)(d_tn->obj_init);
- }
- }
- }
- }
- __kmp_threadprivate_d_table.data[q] = 0;
- }
- }
- }
- /* Call all destructors for threadprivate data belonging to this thread */
- void __kmp_common_destroy_gtid(int gtid) {
- struct private_common *tn;
- struct shared_common *d_tn;
- if (!TCR_4(__kmp_init_gtid)) {
- // This is possible when one of multiple roots initiates early library
- // termination in a sequential region while other teams are active, and its
- // child threads are about to end.
- return;
- }
- KC_TRACE(10, ("__kmp_common_destroy_gtid: T#%d called\n", gtid));
- if ((__kmp_foreign_tp) ? (!KMP_INITIAL_GTID(gtid)) : (!KMP_UBER_GTID(gtid))) {
- if (TCR_4(__kmp_init_common)) {
- /* Cannot do this here since not all threads have destroyed their data */
- /* TCW_4(__kmp_init_common, FALSE); */
- for (tn = __kmp_threads[gtid]->th.th_pri_head; tn; tn = tn->link) {
- d_tn = __kmp_find_shared_task_common(&__kmp_threadprivate_d_table, gtid,
- tn->gbl_addr);
- if (d_tn == NULL)
- continue;
- if (d_tn->is_vec) {
- if (d_tn->dt.dtorv != 0) {
- (void)(*d_tn->dt.dtorv)(tn->par_addr, d_tn->vec_len);
- }
- if (d_tn->obj_init != 0) {
- (void)(*d_tn->dt.dtorv)(d_tn->obj_init, d_tn->vec_len);
- }
- } else {
- if (d_tn->dt.dtor != 0) {
- (void)(*d_tn->dt.dtor)(tn->par_addr);
- }
- if (d_tn->obj_init != 0) {
- (void)(*d_tn->dt.dtor)(d_tn->obj_init);
- }
- }
- }
- KC_TRACE(30, ("__kmp_common_destroy_gtid: T#%d threadprivate destructors "
- "complete\n",
- gtid));
- }
- }
- }
- #ifdef KMP_TASK_COMMON_DEBUG
- static void dump_list(void) {
- int p, q;
- for (p = 0; p < __kmp_all_nth; ++p) {
- if (!__kmp_threads[p])
- continue;
- for (q = 0; q < KMP_HASH_TABLE_SIZE; ++q) {
- if (__kmp_threads[p]->th.th_pri_common->data[q]) {
- struct private_common *tn;
- KC_TRACE(10, ("\tdump_list: gtid:%d addresses\n", p));
- for (tn = __kmp_threads[p]->th.th_pri_common->data[q]; tn;
- tn = tn->next) {
- KC_TRACE(10,
- ("\tdump_list: THREADPRIVATE: Serial %p -> Parallel %p\n",
- tn->gbl_addr, tn->par_addr));
- }
- }
- }
- }
- }
- #endif /* KMP_TASK_COMMON_DEBUG */
- // NOTE: this routine is to be called only from the serial part of the program.
- void kmp_threadprivate_insert_private_data(int gtid, void *pc_addr,
- void *data_addr, size_t pc_size) {
- struct shared_common **lnk_tn, *d_tn;
- KMP_DEBUG_ASSERT(__kmp_threads[gtid] &&
- __kmp_threads[gtid]->th.th_root->r.r_active == 0);
- d_tn = __kmp_find_shared_task_common(&__kmp_threadprivate_d_table, gtid,
- pc_addr);
- if (d_tn == 0) {
- d_tn = (struct shared_common *)__kmp_allocate(sizeof(struct shared_common));
- d_tn->gbl_addr = pc_addr;
- d_tn->pod_init = __kmp_init_common_data(data_addr, pc_size);
- /*
- d_tn->obj_init = 0; // AC: commented out because __kmp_allocate
- zeroes the memory
- d_tn->ct.ctor = 0;
- d_tn->cct.cctor = 0;;
- d_tn->dt.dtor = 0;
- d_tn->is_vec = FALSE;
- d_tn->vec_len = 0L;
- */
- d_tn->cmn_size = pc_size;
- __kmp_acquire_lock(&__kmp_global_lock, gtid);
- lnk_tn = &(__kmp_threadprivate_d_table.data[KMP_HASH(pc_addr)]);
- d_tn->next = *lnk_tn;
- *lnk_tn = d_tn;
- __kmp_release_lock(&__kmp_global_lock, gtid);
- }
- }
- struct private_common *kmp_threadprivate_insert(int gtid, void *pc_addr,
- void *data_addr,
- size_t pc_size) {
- struct private_common *tn, **tt;
- struct shared_common *d_tn;
- /* +++++++++ START OF CRITICAL SECTION +++++++++ */
- __kmp_acquire_lock(&__kmp_global_lock, gtid);
- tn = (struct private_common *)__kmp_allocate(sizeof(struct private_common));
- tn->gbl_addr = pc_addr;
- d_tn = __kmp_find_shared_task_common(
- &__kmp_threadprivate_d_table, gtid,
- pc_addr); /* Only the MASTER data table exists. */
- if (d_tn != 0) {
- /* This threadprivate variable has already been seen. */
- if (d_tn->pod_init == 0 && d_tn->obj_init == 0) {
- d_tn->cmn_size = pc_size;
- if (d_tn->is_vec) {
- if (d_tn->ct.ctorv != 0) {
- /* Construct from scratch so no prototype exists */
- d_tn->obj_init = 0;
- } else if (d_tn->cct.cctorv != 0) {
- /* Now data initialize the prototype since it was previously
- * registered */
- d_tn->obj_init = (void *)__kmp_allocate(d_tn->cmn_size);
- (void)(*d_tn->cct.cctorv)(d_tn->obj_init, pc_addr, d_tn->vec_len);
- } else {
- d_tn->pod_init = __kmp_init_common_data(data_addr, d_tn->cmn_size);
- }
- } else {
- if (d_tn->ct.ctor != 0) {
- /* Construct from scratch so no prototype exists */
- d_tn->obj_init = 0;
- } else if (d_tn->cct.cctor != 0) {
- /* Now data initialize the prototype since it was previously
- registered */
- d_tn->obj_init = (void *)__kmp_allocate(d_tn->cmn_size);
- (void)(*d_tn->cct.cctor)(d_tn->obj_init, pc_addr);
- } else {
- d_tn->pod_init = __kmp_init_common_data(data_addr, d_tn->cmn_size);
- }
- }
- }
- } else {
- struct shared_common **lnk_tn;
- d_tn = (struct shared_common *)__kmp_allocate(sizeof(struct shared_common));
- d_tn->gbl_addr = pc_addr;
- d_tn->cmn_size = pc_size;
- d_tn->pod_init = __kmp_init_common_data(data_addr, pc_size);
- /*
- d_tn->obj_init = 0; // AC: commented out because __kmp_allocate
- zeroes the memory
- d_tn->ct.ctor = 0;
- d_tn->cct.cctor = 0;
- d_tn->dt.dtor = 0;
- d_tn->is_vec = FALSE;
- d_tn->vec_len = 0L;
- */
- lnk_tn = &(__kmp_threadprivate_d_table.data[KMP_HASH(pc_addr)]);
- d_tn->next = *lnk_tn;
- *lnk_tn = d_tn;
- }
- tn->cmn_size = d_tn->cmn_size;
- if ((__kmp_foreign_tp) ? (KMP_INITIAL_GTID(gtid)) : (KMP_UBER_GTID(gtid))) {
- tn->par_addr = (void *)pc_addr;
- } else {
- tn->par_addr = (void *)__kmp_allocate(tn->cmn_size);
- }
- __kmp_release_lock(&__kmp_global_lock, gtid);
- /* +++++++++ END OF CRITICAL SECTION +++++++++ */
- #ifdef USE_CHECKS_COMMON
- if (pc_size > d_tn->cmn_size) {
- KC_TRACE(
- 10, ("__kmp_threadprivate_insert: THREADPRIVATE: %p (%" KMP_UINTPTR_SPEC
- " ,%" KMP_UINTPTR_SPEC ")\n",
- pc_addr, pc_size, d_tn->cmn_size));
- KMP_FATAL(TPCommonBlocksInconsist);
- }
- #endif /* USE_CHECKS_COMMON */
- tt = &(__kmp_threads[gtid]->th.th_pri_common->data[KMP_HASH(pc_addr)]);
- #ifdef KMP_TASK_COMMON_DEBUG
- if (*tt != 0) {
- KC_TRACE(
- 10,
- ("__kmp_threadprivate_insert: WARNING! thread#%d: collision on %p\n",
- gtid, pc_addr));
- }
- #endif
- tn->next = *tt;
- *tt = tn;
- #ifdef KMP_TASK_COMMON_DEBUG
- KC_TRACE(10,
- ("__kmp_threadprivate_insert: thread#%d, inserted node %p on list\n",
- gtid, pc_addr));
- dump_list();
- #endif
- /* Link the node into a simple list */
- tn->link = __kmp_threads[gtid]->th.th_pri_head;
- __kmp_threads[gtid]->th.th_pri_head = tn;
- if ((__kmp_foreign_tp) ? (KMP_INITIAL_GTID(gtid)) : (KMP_UBER_GTID(gtid)))
- return tn;
- /* if C++ object with copy constructor, use it;
- * else if C++ object with constructor, use it for the non-primary thread
- copies only;
- * else use pod_init and memcpy
- *
- * C++ constructors need to be called once for each non-primary thread on
- * allocate
- * C++ copy constructors need to be called once for each thread on allocate */
- /* C++ object with constructors/destructors; don't call constructors for
- primary thread though */
- if (d_tn->is_vec) {
- if (d_tn->ct.ctorv != 0) {
- (void)(*d_tn->ct.ctorv)(tn->par_addr, d_tn->vec_len);
- } else if (d_tn->cct.cctorv != 0) {
- (void)(*d_tn->cct.cctorv)(tn->par_addr, d_tn->obj_init, d_tn->vec_len);
- } else if (tn->par_addr != tn->gbl_addr) {
- __kmp_copy_common_data(tn->par_addr, d_tn->pod_init);
- }
- } else {
- if (d_tn->ct.ctor != 0) {
- (void)(*d_tn->ct.ctor)(tn->par_addr);
- } else if (d_tn->cct.cctor != 0) {
- (void)(*d_tn->cct.cctor)(tn->par_addr, d_tn->obj_init);
- } else if (tn->par_addr != tn->gbl_addr) {
- __kmp_copy_common_data(tn->par_addr, d_tn->pod_init);
- }
- }
- /* !BUILD_OPENMP_C
- if (tn->par_addr != tn->gbl_addr)
- __kmp_copy_common_data( tn->par_addr, d_tn->pod_init ); */
- return tn;
- }
- /* ------------------------------------------------------------------------ */
- /* We are currently parallel, and we know the thread id. */
- /* ------------------------------------------------------------------------ */
- /*!
- @ingroup THREADPRIVATE
- @param loc source location information
- @param data pointer to data being privatized
- @param ctor pointer to constructor function for data
- @param cctor pointer to copy constructor function for data
- @param dtor pointer to destructor function for data
- Register constructors and destructors for thread private data.
- This function is called when executing in parallel, when we know the thread id.
- */
- void __kmpc_threadprivate_register(ident_t *loc, void *data, kmpc_ctor ctor,
- kmpc_cctor cctor, kmpc_dtor dtor) {
- struct shared_common *d_tn, **lnk_tn;
- KC_TRACE(10, ("__kmpc_threadprivate_register: called\n"));
- #ifdef USE_CHECKS_COMMON
- /* copy constructor must be zero for current code gen (Nov 2002 - jph) */
- KMP_ASSERT(cctor == 0);
- #endif /* USE_CHECKS_COMMON */
- /* Only the global data table exists. */
- d_tn = __kmp_find_shared_task_common(&__kmp_threadprivate_d_table, -1, data);
- if (d_tn == 0) {
- d_tn = (struct shared_common *)__kmp_allocate(sizeof(struct shared_common));
- d_tn->gbl_addr = data;
- d_tn->ct.ctor = ctor;
- d_tn->cct.cctor = cctor;
- d_tn->dt.dtor = dtor;
- /*
- d_tn->is_vec = FALSE; // AC: commented out because __kmp_allocate
- zeroes the memory
- d_tn->vec_len = 0L;
- d_tn->obj_init = 0;
- d_tn->pod_init = 0;
- */
- lnk_tn = &(__kmp_threadprivate_d_table.data[KMP_HASH(data)]);
- d_tn->next = *lnk_tn;
- *lnk_tn = d_tn;
- }
- }
- void *__kmpc_threadprivate(ident_t *loc, kmp_int32 global_tid, void *data,
- size_t size) {
- void *ret;
- struct private_common *tn;
- KC_TRACE(10, ("__kmpc_threadprivate: T#%d called\n", global_tid));
- #ifdef USE_CHECKS_COMMON
- if (!__kmp_init_serial)
- KMP_FATAL(RTLNotInitialized);
- #endif /* USE_CHECKS_COMMON */
- if (!__kmp_threads[global_tid]->th.th_root->r.r_active && !__kmp_foreign_tp) {
- /* The parallel address will NEVER overlap with the data_address */
- /* dkp: 3rd arg to kmp_threadprivate_insert_private_data() is the
- * data_address; use data_address = data */
- KC_TRACE(20, ("__kmpc_threadprivate: T#%d inserting private data\n",
- global_tid));
- kmp_threadprivate_insert_private_data(global_tid, data, data, size);
- ret = data;
- } else {
- KC_TRACE(
- 50,
- ("__kmpc_threadprivate: T#%d try to find private data at address %p\n",
- global_tid, data));
- tn = __kmp_threadprivate_find_task_common(
- __kmp_threads[global_tid]->th.th_pri_common, global_tid, data);
- if (tn) {
- KC_TRACE(20, ("__kmpc_threadprivate: T#%d found data\n", global_tid));
- #ifdef USE_CHECKS_COMMON
- if ((size_t)size > tn->cmn_size) {
- KC_TRACE(10, ("THREADPRIVATE: %p (%" KMP_UINTPTR_SPEC
- " ,%" KMP_UINTPTR_SPEC ")\n",
- data, size, tn->cmn_size));
- KMP_FATAL(TPCommonBlocksInconsist);
- }
- #endif /* USE_CHECKS_COMMON */
- } else {
- /* The parallel address will NEVER overlap with the data_address */
- /* dkp: 3rd arg to kmp_threadprivate_insert() is the data_address; use
- * data_address = data */
- KC_TRACE(20, ("__kmpc_threadprivate: T#%d inserting data\n", global_tid));
- tn = kmp_threadprivate_insert(global_tid, data, data, size);
- }
- ret = tn->par_addr;
- }
- KC_TRACE(10, ("__kmpc_threadprivate: T#%d exiting; return value = %p\n",
- global_tid, ret));
- return ret;
- }
- static kmp_cached_addr_t *__kmp_find_cache(void *data) {
- kmp_cached_addr_t *ptr = __kmp_threadpriv_cache_list;
- while (ptr && ptr->data != data)
- ptr = ptr->next;
- return ptr;
- }
- /*!
- @ingroup THREADPRIVATE
- @param loc source location information
- @param global_tid global thread number
- @param data pointer to data to privatize
- @param size size of data to privatize
- @param cache pointer to cache
- @return pointer to private storage
- Allocate private storage for threadprivate data.
- */
- void *
- __kmpc_threadprivate_cached(ident_t *loc,
- kmp_int32 global_tid, // gtid.
- void *data, // Pointer to original global variable.
- size_t size, // Size of original global variable.
- void ***cache) {
- KC_TRACE(10, ("__kmpc_threadprivate_cached: T#%d called with cache: %p, "
- "address: %p, size: %" KMP_SIZE_T_SPEC "\n",
- global_tid, *cache, data, size));
- if (TCR_PTR(*cache) == 0) {
- __kmp_acquire_lock(&__kmp_global_lock, global_tid);
- if (TCR_PTR(*cache) == 0) {
- __kmp_acquire_bootstrap_lock(&__kmp_tp_cached_lock);
- // Compiler often passes in NULL cache, even if it's already been created
- void **my_cache;
- kmp_cached_addr_t *tp_cache_addr;
- // Look for an existing cache
- tp_cache_addr = __kmp_find_cache(data);
- if (!tp_cache_addr) { // Cache was never created; do it now
- __kmp_tp_cached = 1;
- KMP_ITT_IGNORE(my_cache = (void **)__kmp_allocate(
- sizeof(void *) * __kmp_tp_capacity +
- sizeof(kmp_cached_addr_t)););
- // No need to zero the allocated memory; __kmp_allocate does that.
- KC_TRACE(50, ("__kmpc_threadprivate_cached: T#%d allocated cache at "
- "address %p\n",
- global_tid, my_cache));
- /* TODO: free all this memory in __kmp_common_destroy using
- * __kmp_threadpriv_cache_list */
- /* Add address of mycache to linked list for cleanup later */
- tp_cache_addr = (kmp_cached_addr_t *)&my_cache[__kmp_tp_capacity];
- tp_cache_addr->addr = my_cache;
- tp_cache_addr->data = data;
- tp_cache_addr->compiler_cache = cache;
- tp_cache_addr->next = __kmp_threadpriv_cache_list;
- __kmp_threadpriv_cache_list = tp_cache_addr;
- } else { // A cache was already created; use it
- my_cache = tp_cache_addr->addr;
- tp_cache_addr->compiler_cache = cache;
- }
- KMP_MB();
- TCW_PTR(*cache, my_cache);
- __kmp_release_bootstrap_lock(&__kmp_tp_cached_lock);
- KMP_MB();
- }
- __kmp_release_lock(&__kmp_global_lock, global_tid);
- }
- void *ret;
- if ((ret = TCR_PTR((*cache)[global_tid])) == 0) {
- ret = __kmpc_threadprivate(loc, global_tid, data, (size_t)size);
- TCW_PTR((*cache)[global_tid], ret);
- }
- KC_TRACE(10,
- ("__kmpc_threadprivate_cached: T#%d exiting; return value = %p\n",
- global_tid, ret));
- return ret;
- }
- // This function should only be called when both __kmp_tp_cached_lock and
- // kmp_forkjoin_lock are held.
- void __kmp_threadprivate_resize_cache(int newCapacity) {
- KC_TRACE(10, ("__kmp_threadprivate_resize_cache: called with size: %d\n",
- newCapacity));
- kmp_cached_addr_t *ptr = __kmp_threadpriv_cache_list;
- while (ptr) {
- if (ptr->data) { // this location has an active cache; resize it
- void **my_cache;
- KMP_ITT_IGNORE(my_cache =
- (void **)__kmp_allocate(sizeof(void *) * newCapacity +
- sizeof(kmp_cached_addr_t)););
- // No need to zero the allocated memory; __kmp_allocate does that.
- KC_TRACE(50, ("__kmp_threadprivate_resize_cache: allocated cache at %p\n",
- my_cache));
- // Now copy old cache into new cache
- void **old_cache = ptr->addr;
- for (int i = 0; i < __kmp_tp_capacity; ++i) {
- my_cache[i] = old_cache[i];
- }
- // Add address of new my_cache to linked list for cleanup later
- kmp_cached_addr_t *tp_cache_addr;
- tp_cache_addr = (kmp_cached_addr_t *)&my_cache[newCapacity];
- tp_cache_addr->addr = my_cache;
- tp_cache_addr->data = ptr->data;
- tp_cache_addr->compiler_cache = ptr->compiler_cache;
- tp_cache_addr->next = __kmp_threadpriv_cache_list;
- __kmp_threadpriv_cache_list = tp_cache_addr;
- // Copy new cache to compiler's location: We can copy directly
- // to (*compiler_cache) if compiler guarantees it will keep
- // using the same location for the cache. This is not yet true
- // for some compilers, in which case we have to check if
- // compiler_cache is still pointing at old cache, and if so, we
- // can point it at the new cache with an atomic compare&swap
- // operation. (Old method will always work, but we should shift
- // to new method (commented line below) when Intel and Clang
- // compilers use new method.)
- (void)KMP_COMPARE_AND_STORE_PTR(tp_cache_addr->compiler_cache, old_cache,
- my_cache);
- // TCW_PTR(*(tp_cache_addr->compiler_cache), my_cache);
- // If the store doesn't happen here, the compiler's old behavior will
- // inevitably call __kmpc_threadprivate_cache with a new location for the
- // cache, and that function will store the resized cache there at that
- // point.
- // Nullify old cache's data pointer so we skip it next time
- ptr->data = NULL;
- }
- ptr = ptr->next;
- }
- // After all caches are resized, update __kmp_tp_capacity to the new size
- *(volatile int *)&__kmp_tp_capacity = newCapacity;
- }
- /*!
- @ingroup THREADPRIVATE
- @param loc source location information
- @param data pointer to data being privatized
- @param ctor pointer to constructor function for data
- @param cctor pointer to copy constructor function for data
- @param dtor pointer to destructor function for data
- @param vector_length length of the vector (bytes or elements?)
- Register vector constructors and destructors for thread private data.
- */
- void __kmpc_threadprivate_register_vec(ident_t *loc, void *data,
- kmpc_ctor_vec ctor, kmpc_cctor_vec cctor,
- kmpc_dtor_vec dtor,
- size_t vector_length) {
- struct shared_common *d_tn, **lnk_tn;
- KC_TRACE(10, ("__kmpc_threadprivate_register_vec: called\n"));
- #ifdef USE_CHECKS_COMMON
- /* copy constructor must be zero for current code gen (Nov 2002 - jph) */
- KMP_ASSERT(cctor == 0);
- #endif /* USE_CHECKS_COMMON */
- d_tn = __kmp_find_shared_task_common(
- &__kmp_threadprivate_d_table, -1,
- data); /* Only the global data table exists. */
- if (d_tn == 0) {
- d_tn = (struct shared_common *)__kmp_allocate(sizeof(struct shared_common));
- d_tn->gbl_addr = data;
- d_tn->ct.ctorv = ctor;
- d_tn->cct.cctorv = cctor;
- d_tn->dt.dtorv = dtor;
- d_tn->is_vec = TRUE;
- d_tn->vec_len = (size_t)vector_length;
- // d_tn->obj_init = 0; // AC: __kmp_allocate zeroes the memory
- // d_tn->pod_init = 0;
- lnk_tn = &(__kmp_threadprivate_d_table.data[KMP_HASH(data)]);
- d_tn->next = *lnk_tn;
- *lnk_tn = d_tn;
- }
- }
- void __kmp_cleanup_threadprivate_caches() {
- kmp_cached_addr_t *ptr = __kmp_threadpriv_cache_list;
- while (ptr) {
- void **cache = ptr->addr;
- __kmp_threadpriv_cache_list = ptr->next;
- if (*ptr->compiler_cache)
- *ptr->compiler_cache = NULL;
- ptr->compiler_cache = NULL;
- ptr->data = NULL;
- ptr->addr = NULL;
- ptr->next = NULL;
- // Threadprivate data pointed at by cache entries are destroyed at end of
- // __kmp_launch_thread with __kmp_common_destroy_gtid.
- __kmp_free(cache); // implicitly frees ptr too
- ptr = __kmp_threadpriv_cache_list;
- }
- }
|