2011-11-08 19:13:41 +08:00
|
|
|
/* Copyright (C) 2008, 2009, 2011 Free Software Foundation, Inc.
|
|
|
|
Contributed by Richard Henderson <rth@redhat.com>.
|
|
|
|
|
|
|
|
This file is part of the GNU Transactional Memory Library (libitm).
|
|
|
|
|
|
|
|
Libitm is free software; you can redistribute it and/or modify it
|
|
|
|
under the terms of the GNU General Public License as published by
|
|
|
|
the Free Software Foundation; either version 3 of the License, or
|
|
|
|
(at your option) any later version.
|
|
|
|
|
|
|
|
Libitm is distributed in the hope that it will be useful, but WITHOUT ANY
|
|
|
|
WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
|
|
|
|
FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
|
|
more details.
|
|
|
|
|
|
|
|
Under Section 7 of GPL version 3, you are granted additional
|
|
|
|
permissions described in the GCC Runtime Library Exception, version
|
|
|
|
3.1, as published by the Free Software Foundation.
|
|
|
|
|
|
|
|
You should have received a copy of the GNU General Public License and
|
|
|
|
a copy of the GCC Runtime Library Exception along with this program;
|
|
|
|
see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
|
|
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
|
|
|
|
#include "libitm_i.h"
|
|
|
|
#include <pthread.h>
|
|
|
|
|
|
|
|
|
|
|
|
using namespace GTM;
|
|
|
|
|
|
|
|
#if !defined(HAVE_ARCH_GTM_THREAD) || !defined(HAVE_ARCH_GTM_THREAD_DISP)
|
|
|
|
extern __thread gtm_thread_tls _gtm_thr_tls;
|
|
|
|
#endif
|
|
|
|
|
|
|
|
gtm_rwlock GTM::gtm_thread::serial_lock;
|
|
|
|
gtm_thread *GTM::gtm_thread::list_of_threads = 0;
|
|
|
|
unsigned GTM::gtm_thread::number_of_threads = 0;
|
|
|
|
|
|
|
|
gtm_stmlock GTM::gtm_stmlock_array[LOCK_ARRAY_SIZE];
|
2011-12-14 03:11:25 +08:00
|
|
|
atomic<gtm_version> GTM::gtm_clock;
|
2011-11-08 19:13:41 +08:00
|
|
|
|
|
|
|
/* ??? Move elsewhere when we figure out library initialization. */
|
|
|
|
uint64_t GTM::gtm_spin_count_var = 1000;
|
|
|
|
|
2011-12-14 03:11:25 +08:00
|
|
|
#ifdef HAVE_64BIT_SYNC_BUILTINS
|
|
|
|
static atomic<_ITM_transactionId_t> global_tid;
|
|
|
|
#else
|
2011-11-08 19:13:41 +08:00
|
|
|
static _ITM_transactionId_t global_tid;
|
2011-12-14 03:11:25 +08:00
|
|
|
static pthread_mutex_t global_tid_lock = PTHREAD_MUTEX_INITIALIZER;
|
|
|
|
#endif
|
|
|
|
|
2011-11-08 19:13:41 +08:00
|
|
|
|
|
|
|
// Provides a on-thread-exit callback used to release per-thread data.
|
|
|
|
static pthread_key_t thr_release_key;
|
|
|
|
static pthread_once_t thr_release_once = PTHREAD_ONCE_INIT;
|
|
|
|
|
|
|
|
|
|
|
|
/* Allocate a transaction structure. */
|
|
|
|
void *
|
|
|
|
GTM::gtm_thread::operator new (size_t s)
|
|
|
|
{
|
|
|
|
void *tx;
|
|
|
|
|
|
|
|
assert(s == sizeof(gtm_thread));
|
|
|
|
|
|
|
|
tx = xmalloc (sizeof (gtm_thread), true);
|
|
|
|
memset (tx, 0, sizeof (gtm_thread));
|
|
|
|
|
|
|
|
return tx;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Free the given transaction. Raises an error if the transaction is still
|
|
|
|
in use. */
|
|
|
|
void
|
|
|
|
GTM::gtm_thread::operator delete(void *tx)
|
|
|
|
{
|
|
|
|
free(tx);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
thread_exit_handler(void *)
|
|
|
|
{
|
|
|
|
gtm_thread *thr = gtm_thr();
|
|
|
|
if (thr)
|
|
|
|
delete thr;
|
|
|
|
set_gtm_thr(0);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
thread_exit_init()
|
|
|
|
{
|
|
|
|
if (pthread_key_create(&thr_release_key, thread_exit_handler))
|
|
|
|
GTM_fatal("Creating thread release TLS key failed.");
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
GTM::gtm_thread::~gtm_thread()
|
|
|
|
{
|
|
|
|
if (nesting > 0)
|
|
|
|
GTM_fatal("Thread exit while a transaction is still active.");
|
|
|
|
|
|
|
|
// Deregister this transaction.
|
|
|
|
serial_lock.write_lock ();
|
|
|
|
gtm_thread **prev = &list_of_threads;
|
|
|
|
for (; *prev; prev = &(*prev)->next_thread)
|
|
|
|
{
|
|
|
|
if (*prev == this)
|
|
|
|
{
|
|
|
|
*prev = (*prev)->next_thread;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
number_of_threads--;
|
|
|
|
number_of_threads_changed(number_of_threads + 1, number_of_threads);
|
|
|
|
serial_lock.write_unlock ();
|
|
|
|
}
|
|
|
|
|
|
|
|
GTM::gtm_thread::gtm_thread ()
|
|
|
|
{
|
|
|
|
// This object's memory has been set to zero by operator new, so no need
|
|
|
|
// to initialize any of the other primitive-type members that do not have
|
|
|
|
// constructors.
|
2011-12-14 03:11:25 +08:00
|
|
|
shared_state.store(-1, memory_order_relaxed);
|
2011-11-08 19:13:41 +08:00
|
|
|
|
|
|
|
// Register this transaction with the list of all threads' transactions.
|
|
|
|
serial_lock.write_lock ();
|
|
|
|
next_thread = list_of_threads;
|
|
|
|
list_of_threads = this;
|
|
|
|
number_of_threads++;
|
|
|
|
number_of_threads_changed(number_of_threads - 1, number_of_threads);
|
|
|
|
serial_lock.write_unlock ();
|
|
|
|
|
|
|
|
if (pthread_once(&thr_release_once, thread_exit_init))
|
|
|
|
GTM_fatal("Initializing thread release TLS key failed.");
|
|
|
|
// Any non-null value is sufficient to trigger destruction of this
|
|
|
|
// transaction when the current thread terminates.
|
|
|
|
if (pthread_setspecific(thr_release_key, this))
|
|
|
|
GTM_fatal("Setting thread release TLS key failed.");
|
|
|
|
}
|
|
|
|
|
2011-12-14 03:11:25 +08:00
|
|
|
static inline uint32_t
|
|
|
|
choose_code_path(uint32_t prop, abi_dispatch *disp)
|
2011-11-08 19:13:41 +08:00
|
|
|
{
|
|
|
|
if ((prop & pr_uninstrumentedCode) && disp->can_run_uninstrumented_code())
|
|
|
|
return a_runUninstrumentedCode;
|
|
|
|
else
|
|
|
|
return a_runInstrumentedCode;
|
|
|
|
}
|
|
|
|
|
|
|
|
uint32_t
|
|
|
|
GTM::gtm_thread::begin_transaction (uint32_t prop, const gtm_jmpbuf *jb)
|
|
|
|
{
|
|
|
|
static const _ITM_transactionId_t tid_block_size = 1 << 16;
|
|
|
|
|
|
|
|
gtm_thread *tx;
|
|
|
|
abi_dispatch *disp;
|
|
|
|
uint32_t ret;
|
|
|
|
|
|
|
|
// ??? pr_undoLogCode is not properly defined in the ABI. Are barriers
|
|
|
|
// omitted because they are not necessary (e.g., a transaction on thread-
|
|
|
|
// local data) or because the compiler thinks that some kind of global
|
|
|
|
// synchronization might perform better?
|
|
|
|
if (unlikely(prop & pr_undoLogCode))
|
|
|
|
GTM_fatal("pr_undoLogCode not supported");
|
|
|
|
|
|
|
|
tx = gtm_thr();
|
|
|
|
if (unlikely(tx == NULL))
|
|
|
|
{
|
|
|
|
// Create the thread object. The constructor will also set up automatic
|
|
|
|
// deletion on thread termination.
|
|
|
|
tx = new gtm_thread();
|
|
|
|
set_gtm_thr(tx);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (tx->nesting > 0)
|
|
|
|
{
|
|
|
|
// This is a nested transaction.
|
|
|
|
// Check prop compatibility:
|
|
|
|
// The ABI requires pr_hasNoFloatUpdate, pr_hasNoVectorUpdate,
|
|
|
|
// pr_hasNoIrrevocable, pr_aWBarriersOmitted, pr_RaRBarriersOmitted, and
|
|
|
|
// pr_hasNoSimpleReads to hold for the full dynamic scope of a
|
|
|
|
// transaction. We could check that these are set for the nested
|
|
|
|
// transaction if they are also set for the parent transaction, but the
|
|
|
|
// ABI does not require these flags to be set if they could be set,
|
|
|
|
// so the check could be too strict.
|
|
|
|
// ??? For pr_readOnly, lexical or dynamic scope is unspecified.
|
|
|
|
|
|
|
|
if (prop & pr_hasNoAbort)
|
|
|
|
{
|
|
|
|
// We can use flat nesting, so elide this transaction.
|
|
|
|
if (!(prop & pr_instrumentedCode))
|
|
|
|
{
|
|
|
|
if (!(tx->state & STATE_SERIAL) ||
|
|
|
|
!(tx->state & STATE_IRREVOCABLE))
|
|
|
|
tx->serialirr_mode();
|
|
|
|
}
|
|
|
|
// Increment nesting level after checking that we have a method that
|
|
|
|
// allows us to continue.
|
|
|
|
tx->nesting++;
|
|
|
|
return choose_code_path(prop, abi_disp());
|
|
|
|
}
|
|
|
|
|
|
|
|
// The transaction might abort, so use closed nesting if possible.
|
|
|
|
// pr_hasNoAbort has lexical scope, so the compiler should really have
|
|
|
|
// generated an instrumented code path.
|
|
|
|
assert(prop & pr_instrumentedCode);
|
|
|
|
|
|
|
|
// Create a checkpoint of the current transaction.
|
|
|
|
gtm_transaction_cp *cp = tx->parent_txns.push();
|
|
|
|
cp->save(tx);
|
|
|
|
new (&tx->alloc_actions) aa_tree<uintptr_t, gtm_alloc_action>();
|
|
|
|
|
|
|
|
// Check whether the current method actually supports closed nesting.
|
|
|
|
// If we can switch to another one, do so.
|
|
|
|
// If not, we assume that actual aborts are infrequent, and rather
|
|
|
|
// restart in _ITM_abortTransaction when we really have to.
|
|
|
|
disp = abi_disp();
|
|
|
|
if (!disp->closed_nesting())
|
|
|
|
{
|
|
|
|
// ??? Should we elide the transaction if there is no alternative
|
|
|
|
// method that supports closed nesting? If we do, we need to set
|
|
|
|
// some flag to prevent _ITM_abortTransaction from aborting the
|
|
|
|
// wrong transaction (i.e., some parent transaction).
|
|
|
|
abi_dispatch *cn_disp = disp->closed_nesting_alternative();
|
|
|
|
if (cn_disp)
|
|
|
|
{
|
|
|
|
disp = cn_disp;
|
|
|
|
set_abi_disp(disp);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// Outermost transaction
|
|
|
|
disp = tx->decide_begin_dispatch (prop);
|
|
|
|
if (disp == dispatch_serialirr() || disp == dispatch_serial())
|
|
|
|
{
|
|
|
|
tx->state = STATE_SERIAL;
|
|
|
|
if (disp == dispatch_serialirr())
|
|
|
|
tx->state |= STATE_IRREVOCABLE;
|
|
|
|
serial_lock.write_lock ();
|
|
|
|
}
|
|
|
|
else
|
|
|
|
serial_lock.read_lock (tx);
|
|
|
|
|
|
|
|
set_abi_disp (disp);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Initialization that is common for outermost and nested transactions.
|
|
|
|
tx->prop = prop;
|
|
|
|
tx->nesting++;
|
|
|
|
|
|
|
|
tx->jb = *jb;
|
|
|
|
|
|
|
|
// As long as we have not exhausted a previously allocated block of TIDs,
|
|
|
|
// we can avoid an atomic operation on a shared cacheline.
|
|
|
|
if (tx->local_tid & (tid_block_size - 1))
|
|
|
|
tx->id = tx->local_tid++;
|
|
|
|
else
|
|
|
|
{
|
|
|
|
#ifdef HAVE_64BIT_SYNC_BUILTINS
|
libitm: Fixed conversion to C++11 atomics.
libitm/
* beginend.cc (GTM::gtm_thread::begin_transaction): Add comment.
(GTM::gtm_thread::try_commit): Changed memory order.
* config/linux/alpha/futex_bits.h (sys_futex0): Take atomic int
as parameter.
* config/linux/x86/futex_bits.h (sys_futex0): Same.
* config/linux/sh/futex_bits.h (sys_futex0): Same.
* config/linux/futex_bits.h (sys_futex0): Same.
* config/linux/futex.cc (futex_wait, futex_wake): Same.
* config/linux/futex.h (futex_wait, futex_wake): Same.
* config/linux/rwlock.h (gtm_rwlock::writers,
gtm_rwlock::writer_readers, gtm_rwlock::readers): Change to atomic
ints.
* config/linux/rwlock.cc (gtm_rwlock::read_lock,
gtm_rwlock::write_lock_generic, gtm_rwlock::read_unlock,
gtm_rwlock::write_unlock): Fix memory orders and fences.
* config/posix/rwlock.cc (gtm_rwlock::read_lock,
gtm_rwlock::write_lock_generic, gtm_rwlock::read_unlock,
gtm_rwlock::write_unlock): Same.
* config/linux/rwlock.h (gtm_rwlock::summary): Change to atomic int.
* method-gl.cc (gl_mg::init, gl_wt_dispatch::memtransfer_static,
gl_wt_dispatch::memset_static, gl_wt_dispatch::begin_or_restart):
Add comments.
(gl_wt_dispatch::pre_write, gl_wt_dispatch::validate,
gl_wt_dispatch::load, gl_wt_dispatch::store,
gl_wt_dispatch::try_commit, gl_wt_dispatch::rollback): Fix memory
orders and fences. Add comments.
From-SVN: r182674
2011-12-24 09:42:20 +08:00
|
|
|
// We don't really care which block of TIDs we get but only that we
|
|
|
|
// acquire one atomically; therefore, relaxed memory order is
|
|
|
|
// sufficient.
|
2011-12-14 03:11:25 +08:00
|
|
|
tx->id = global_tid.fetch_add(tid_block_size, memory_order_relaxed);
|
2011-11-08 19:13:41 +08:00
|
|
|
tx->local_tid = tx->id + 1;
|
|
|
|
#else
|
|
|
|
pthread_mutex_lock (&global_tid_lock);
|
|
|
|
global_tid += tid_block_size;
|
|
|
|
tx->id = global_tid;
|
|
|
|
tx->local_tid = tx->id + 1;
|
|
|
|
pthread_mutex_unlock (&global_tid_lock);
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
// Run dispatch-specific restart code. Retry until we succeed.
|
|
|
|
GTM::gtm_restart_reason rr;
|
|
|
|
while ((rr = disp->begin_or_restart()) != NO_RESTART)
|
|
|
|
{
|
|
|
|
tx->decide_retry_strategy(rr);
|
|
|
|
disp = abi_disp();
|
|
|
|
}
|
|
|
|
|
|
|
|
// Determine the code path to run. Only irrevocable transactions cannot be
|
|
|
|
// restarted, so all other transactions need to save live variables.
|
|
|
|
ret = choose_code_path(prop, disp);
|
|
|
|
if (!(tx->state & STATE_IRREVOCABLE))
|
|
|
|
ret |= a_saveLiveVariables;
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void
|
|
|
|
GTM::gtm_transaction_cp::save(gtm_thread* tx)
|
|
|
|
{
|
|
|
|
// Save everything that we might have to restore on restarts or aborts.
|
|
|
|
jb = tx->jb;
|
|
|
|
undolog_size = tx->undolog.size();
|
|
|
|
memcpy(&alloc_actions, &tx->alloc_actions, sizeof(alloc_actions));
|
|
|
|
user_actions_size = tx->user_actions.size();
|
|
|
|
id = tx->id;
|
|
|
|
prop = tx->prop;
|
|
|
|
cxa_catch_count = tx->cxa_catch_count;
|
|
|
|
cxa_unthrown = tx->cxa_unthrown;
|
|
|
|
disp = abi_disp();
|
|
|
|
nesting = tx->nesting;
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
GTM::gtm_transaction_cp::commit(gtm_thread* tx)
|
|
|
|
{
|
|
|
|
// Restore state that is not persistent across commits. Exception handling,
|
|
|
|
// information, nesting level, and any logs do not need to be restored on
|
|
|
|
// commits of nested transactions. Allocation actions must be committed
|
|
|
|
// before committing the snapshot.
|
|
|
|
tx->jb = jb;
|
|
|
|
memcpy(&tx->alloc_actions, &alloc_actions, sizeof(alloc_actions));
|
|
|
|
tx->id = id;
|
|
|
|
tx->prop = prop;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void
|
|
|
|
GTM::gtm_thread::rollback (gtm_transaction_cp *cp, bool aborting)
|
|
|
|
{
|
|
|
|
// The undo log is special in that it used for both thread-local and shared
|
|
|
|
// data. Because of the latter, we have to roll it back before any
|
|
|
|
// dispatch-specific rollback (which handles synchronization with other
|
|
|
|
// transactions).
|
|
|
|
rollback_undolog (cp ? cp->undolog_size : 0);
|
|
|
|
|
|
|
|
// Perform dispatch-specific rollback.
|
|
|
|
abi_disp()->rollback (cp);
|
|
|
|
|
|
|
|
// Roll back all actions that are supposed to happen around the transaction.
|
|
|
|
rollback_user_actions (cp ? cp->user_actions_size : 0);
|
|
|
|
commit_allocations (true, (cp ? &cp->alloc_actions : 0));
|
|
|
|
revert_cpp_exceptions (cp);
|
|
|
|
|
|
|
|
if (cp)
|
|
|
|
{
|
|
|
|
// We do not yet handle restarts of nested transactions. To do that, we
|
|
|
|
// would have to restore some state (jb, id, prop, nesting) not to the
|
|
|
|
// checkpoint but to the transaction that was started from this
|
|
|
|
// checkpoint (e.g., nesting = cp->nesting + 1);
|
|
|
|
assert(aborting);
|
|
|
|
// Roll back the rest of the state to the checkpoint.
|
|
|
|
jb = cp->jb;
|
|
|
|
id = cp->id;
|
|
|
|
prop = cp->prop;
|
|
|
|
if (cp->disp != abi_disp())
|
|
|
|
set_abi_disp(cp->disp);
|
|
|
|
memcpy(&alloc_actions, &cp->alloc_actions, sizeof(alloc_actions));
|
|
|
|
nesting = cp->nesting;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// Roll back to the outermost transaction.
|
|
|
|
// Restore the jump buffer and transaction properties, which we will
|
|
|
|
// need for the longjmp used to restart or abort the transaction.
|
|
|
|
if (parent_txns.size() > 0)
|
|
|
|
{
|
|
|
|
jb = parent_txns[0].jb;
|
|
|
|
id = parent_txns[0].id;
|
|
|
|
prop = parent_txns[0].prop;
|
|
|
|
}
|
|
|
|
// Reset the transaction. Do not reset this->state, which is handled by
|
|
|
|
// the callers. Note that if we are not aborting, we reset the
|
|
|
|
// transaction to the point after having executed begin_transaction
|
|
|
|
// (we will return from it), so the nesting level must be one, not zero.
|
|
|
|
nesting = (aborting ? 0 : 1);
|
|
|
|
parent_txns.clear();
|
|
|
|
}
|
|
|
|
|
|
|
|
if (this->eh_in_flight)
|
|
|
|
{
|
|
|
|
_Unwind_DeleteException ((_Unwind_Exception *) this->eh_in_flight);
|
|
|
|
this->eh_in_flight = NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void ITM_REGPARM
|
|
|
|
_ITM_abortTransaction (_ITM_abortReason reason)
|
|
|
|
{
|
|
|
|
gtm_thread *tx = gtm_thr();
|
|
|
|
|
|
|
|
assert (reason == userAbort || reason == (userAbort | outerAbort));
|
|
|
|
assert ((tx->prop & pr_hasNoAbort) == 0);
|
|
|
|
|
|
|
|
if (tx->state & gtm_thread::STATE_IRREVOCABLE)
|
|
|
|
abort ();
|
|
|
|
|
|
|
|
// Roll back to innermost transaction.
|
|
|
|
if (tx->parent_txns.size() > 0 && !(reason & outerAbort))
|
|
|
|
{
|
|
|
|
// If the current method does not support closed nesting but we are
|
|
|
|
// nested and must only roll back the innermost transaction, then
|
|
|
|
// restart with a method that supports closed nesting.
|
|
|
|
abi_dispatch *disp = abi_disp();
|
|
|
|
if (!disp->closed_nesting())
|
|
|
|
tx->restart(RESTART_CLOSED_NESTING);
|
|
|
|
|
|
|
|
// The innermost transaction is a closed nested transaction.
|
|
|
|
gtm_transaction_cp *cp = tx->parent_txns.pop();
|
|
|
|
uint32_t longjmp_prop = tx->prop;
|
|
|
|
gtm_jmpbuf longjmp_jb = tx->jb;
|
|
|
|
|
|
|
|
tx->rollback (cp, true);
|
|
|
|
|
|
|
|
// Jump to nested transaction (use the saved jump buffer).
|
2011-12-01 06:29:33 +08:00
|
|
|
GTM_longjmp (a_abortTransaction | a_restoreLiveVariables,
|
|
|
|
&longjmp_jb, longjmp_prop);
|
2011-11-08 19:13:41 +08:00
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// There is no nested transaction or an abort of the outermost
|
|
|
|
// transaction was requested, so roll back to the outermost transaction.
|
|
|
|
tx->rollback (0, true);
|
|
|
|
|
|
|
|
// Aborting an outermost transaction finishes execution of the whole
|
|
|
|
// transaction. Therefore, reset transaction state.
|
|
|
|
if (tx->state & gtm_thread::STATE_SERIAL)
|
|
|
|
gtm_thread::serial_lock.write_unlock ();
|
|
|
|
else
|
|
|
|
gtm_thread::serial_lock.read_unlock (tx);
|
|
|
|
tx->state = 0;
|
|
|
|
|
2011-12-01 06:29:33 +08:00
|
|
|
GTM_longjmp (a_abortTransaction | a_restoreLiveVariables,
|
|
|
|
&tx->jb, tx->prop);
|
2011-11-08 19:13:41 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
bool
|
|
|
|
GTM::gtm_thread::trycommit ()
|
|
|
|
{
|
|
|
|
nesting--;
|
|
|
|
|
|
|
|
// Skip any real commit for elided transactions.
|
|
|
|
if (nesting > 0 && (parent_txns.size() == 0 ||
|
|
|
|
nesting > parent_txns[parent_txns.size() - 1].nesting))
|
|
|
|
return true;
|
|
|
|
|
|
|
|
if (nesting > 0)
|
|
|
|
{
|
|
|
|
// Commit of a closed-nested transaction. Remove one checkpoint and add
|
|
|
|
// any effects of this transaction to the parent transaction.
|
|
|
|
gtm_transaction_cp *cp = parent_txns.pop();
|
|
|
|
commit_allocations(false, &cp->alloc_actions);
|
|
|
|
cp->commit(this);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Commit of an outermost transaction.
|
|
|
|
gtm_word priv_time = 0;
|
|
|
|
if (abi_disp()->trycommit (priv_time))
|
|
|
|
{
|
|
|
|
// The transaction is now inactive. Everything that we still have to do
|
|
|
|
// will not synchronize with other transactions anymore.
|
|
|
|
if (state & gtm_thread::STATE_SERIAL)
|
2011-12-24 09:42:48 +08:00
|
|
|
{
|
|
|
|
gtm_thread::serial_lock.write_unlock ();
|
|
|
|
// There are no other active transactions, so there's no need to
|
|
|
|
// enforce privatization safety.
|
|
|
|
priv_time = 0;
|
|
|
|
}
|
2011-11-08 19:13:41 +08:00
|
|
|
else
|
|
|
|
gtm_thread::serial_lock.read_unlock (this);
|
|
|
|
state = 0;
|
|
|
|
|
|
|
|
// We can commit the undo log after dispatch-specific commit and after
|
|
|
|
// making the transaction inactive because we only have to reset
|
|
|
|
// gtm_thread state.
|
|
|
|
commit_undolog ();
|
|
|
|
// Reset further transaction state.
|
|
|
|
cxa_catch_count = 0;
|
|
|
|
cxa_unthrown = NULL;
|
|
|
|
restart_total = 0;
|
|
|
|
|
|
|
|
// Ensure privatization safety, if necessary.
|
|
|
|
if (priv_time)
|
|
|
|
{
|
libitm: Fixed conversion to C++11 atomics.
libitm/
* beginend.cc (GTM::gtm_thread::begin_transaction): Add comment.
(GTM::gtm_thread::try_commit): Changed memory order.
* config/linux/alpha/futex_bits.h (sys_futex0): Take atomic int
as parameter.
* config/linux/x86/futex_bits.h (sys_futex0): Same.
* config/linux/sh/futex_bits.h (sys_futex0): Same.
* config/linux/futex_bits.h (sys_futex0): Same.
* config/linux/futex.cc (futex_wait, futex_wake): Same.
* config/linux/futex.h (futex_wait, futex_wake): Same.
* config/linux/rwlock.h (gtm_rwlock::writers,
gtm_rwlock::writer_readers, gtm_rwlock::readers): Change to atomic
ints.
* config/linux/rwlock.cc (gtm_rwlock::read_lock,
gtm_rwlock::write_lock_generic, gtm_rwlock::read_unlock,
gtm_rwlock::write_unlock): Fix memory orders and fences.
* config/posix/rwlock.cc (gtm_rwlock::read_lock,
gtm_rwlock::write_lock_generic, gtm_rwlock::read_unlock,
gtm_rwlock::write_unlock): Same.
* config/linux/rwlock.h (gtm_rwlock::summary): Change to atomic int.
* method-gl.cc (gl_mg::init, gl_wt_dispatch::memtransfer_static,
gl_wt_dispatch::memset_static, gl_wt_dispatch::begin_or_restart):
Add comments.
(gl_wt_dispatch::pre_write, gl_wt_dispatch::validate,
gl_wt_dispatch::load, gl_wt_dispatch::store,
gl_wt_dispatch::try_commit, gl_wt_dispatch::rollback): Fix memory
orders and fences. Add comments.
From-SVN: r182674
2011-12-24 09:42:20 +08:00
|
|
|
// There must be a seq_cst fence between the following loads of the
|
|
|
|
// other transactions' shared_state and the dispatch-specific stores
|
|
|
|
// that signal updates by this transaction (e.g., lock
|
|
|
|
// acquisitions). This ensures that if we read prior to other
|
|
|
|
// reader transactions setting their shared_state to 0, then those
|
|
|
|
// readers will observe our updates. We can reuse the seq_cst fence
|
|
|
|
// in serial_lock.read_unlock() however, so we don't need another
|
|
|
|
// one here.
|
2011-11-08 19:13:41 +08:00
|
|
|
// TODO Don't just spin but also block using cond vars / futexes
|
|
|
|
// here. Should probably be integrated with the serial lock code.
|
|
|
|
for (gtm_thread *it = gtm_thread::list_of_threads; it != 0;
|
|
|
|
it = it->next_thread)
|
|
|
|
{
|
|
|
|
if (it == this) continue;
|
libitm: Fixed conversion to C++11 atomics.
libitm/
* beginend.cc (GTM::gtm_thread::begin_transaction): Add comment.
(GTM::gtm_thread::try_commit): Changed memory order.
* config/linux/alpha/futex_bits.h (sys_futex0): Take atomic int
as parameter.
* config/linux/x86/futex_bits.h (sys_futex0): Same.
* config/linux/sh/futex_bits.h (sys_futex0): Same.
* config/linux/futex_bits.h (sys_futex0): Same.
* config/linux/futex.cc (futex_wait, futex_wake): Same.
* config/linux/futex.h (futex_wait, futex_wake): Same.
* config/linux/rwlock.h (gtm_rwlock::writers,
gtm_rwlock::writer_readers, gtm_rwlock::readers): Change to atomic
ints.
* config/linux/rwlock.cc (gtm_rwlock::read_lock,
gtm_rwlock::write_lock_generic, gtm_rwlock::read_unlock,
gtm_rwlock::write_unlock): Fix memory orders and fences.
* config/posix/rwlock.cc (gtm_rwlock::read_lock,
gtm_rwlock::write_lock_generic, gtm_rwlock::read_unlock,
gtm_rwlock::write_unlock): Same.
* config/linux/rwlock.h (gtm_rwlock::summary): Change to atomic int.
* method-gl.cc (gl_mg::init, gl_wt_dispatch::memtransfer_static,
gl_wt_dispatch::memset_static, gl_wt_dispatch::begin_or_restart):
Add comments.
(gl_wt_dispatch::pre_write, gl_wt_dispatch::validate,
gl_wt_dispatch::load, gl_wt_dispatch::store,
gl_wt_dispatch::try_commit, gl_wt_dispatch::rollback): Fix memory
orders and fences. Add comments.
From-SVN: r182674
2011-12-24 09:42:20 +08:00
|
|
|
// We need to load other threads' shared_state using acquire
|
|
|
|
// semantics (matching the release semantics of the respective
|
|
|
|
// updates). This is necessary to ensure that the other
|
|
|
|
// threads' memory accesses happen before our actions that
|
|
|
|
// assume privatization safety.
|
|
|
|
// TODO Are there any platform-specific optimizations (e.g.,
|
|
|
|
// merging barriers)?
|
|
|
|
while (it->shared_state.load(memory_order_acquire) < priv_time)
|
2011-11-08 19:13:41 +08:00
|
|
|
cpu_relax();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// After ensuring privatization safety, we execute potentially
|
|
|
|
// privatizing actions (e.g., calling free()). User actions are first.
|
|
|
|
commit_user_actions ();
|
|
|
|
commit_allocations (false, 0);
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
void ITM_NORETURN
|
2011-12-24 09:42:35 +08:00
|
|
|
GTM::gtm_thread::restart (gtm_restart_reason r, bool finish_serial_upgrade)
|
2011-11-08 19:13:41 +08:00
|
|
|
{
|
|
|
|
// Roll back to outermost transaction. Do not reset transaction state because
|
|
|
|
// we will continue executing this transaction.
|
|
|
|
rollback ();
|
2011-12-24 09:42:35 +08:00
|
|
|
|
|
|
|
// If we have to restart while an upgrade of the serial lock is happening,
|
|
|
|
// we need to finish this here, after rollback (to ensure privatization
|
|
|
|
// safety despite undo writes) and before deciding about the retry strategy
|
|
|
|
// (which could switch to/from serial mode).
|
|
|
|
if (finish_serial_upgrade)
|
|
|
|
gtm_thread::serial_lock.write_upgrade_finish(this);
|
|
|
|
|
2011-11-08 19:13:41 +08:00
|
|
|
decide_retry_strategy (r);
|
|
|
|
|
|
|
|
// Run dispatch-specific restart code. Retry until we succeed.
|
|
|
|
abi_dispatch* disp = abi_disp();
|
|
|
|
GTM::gtm_restart_reason rr;
|
|
|
|
while ((rr = disp->begin_or_restart()) != NO_RESTART)
|
|
|
|
{
|
|
|
|
decide_retry_strategy(rr);
|
|
|
|
disp = abi_disp();
|
|
|
|
}
|
|
|
|
|
2011-12-01 06:29:33 +08:00
|
|
|
GTM_longjmp (choose_code_path(prop, disp) | a_restoreLiveVariables,
|
|
|
|
&jb, prop);
|
2011-11-08 19:13:41 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
void ITM_REGPARM
|
|
|
|
_ITM_commitTransaction(void)
|
|
|
|
{
|
|
|
|
gtm_thread *tx = gtm_thr();
|
|
|
|
if (!tx->trycommit ())
|
|
|
|
tx->restart (RESTART_VALIDATE_COMMIT);
|
|
|
|
}
|
|
|
|
|
|
|
|
void ITM_REGPARM
|
|
|
|
_ITM_commitTransactionEH(void *exc_ptr)
|
|
|
|
{
|
|
|
|
gtm_thread *tx = gtm_thr();
|
|
|
|
if (!tx->trycommit ())
|
|
|
|
{
|
|
|
|
tx->eh_in_flight = exc_ptr;
|
|
|
|
tx->restart (RESTART_VALIDATE_COMMIT);
|
|
|
|
}
|
|
|
|
}
|