mirror of git://gcc.gnu.org/git/gcc.git
				
				
				
			
		
			
				
	
	
		
			749 lines
		
	
	
		
			26 KiB
		
	
	
	
		
			C++
		
	
	
	
			
		
		
	
	
			749 lines
		
	
	
		
			26 KiB
		
	
	
	
		
			C++
		
	
	
	
| /* Copyright (C) 2008-2019 Free Software Foundation, Inc.
 | |
|    Contributed by Richard Henderson <rth@redhat.com>.
 | |
| 
 | |
|    This file is part of the GNU Transactional Memory Library (libitm).
 | |
| 
 | |
|    Libitm is free software; you can redistribute it and/or modify it
 | |
|    under the terms of the GNU General Public License as published by
 | |
|    the Free Software Foundation; either version 3 of the License, or
 | |
|    (at your option) any later version.
 | |
| 
 | |
|    Libitm is distributed in the hope that it will be useful, but WITHOUT ANY
 | |
|    WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
 | |
|    FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
 | |
|    more details.
 | |
| 
 | |
|    Under Section 7 of GPL version 3, you are granted additional
 | |
|    permissions described in the GCC Runtime Library Exception, version
 | |
|    3.1, as published by the Free Software Foundation.
 | |
| 
 | |
|    You should have received a copy of the GNU General Public License and
 | |
|    a copy of the GCC Runtime Library Exception along with this program;
 | |
|    see the files COPYING3 and COPYING.RUNTIME respectively.  If not, see
 | |
|    <http://www.gnu.org/licenses/>.  */
 | |
| 
 | |
| #include "libitm_i.h"
 | |
| #include <pthread.h>
 | |
| 
 | |
| 
 | |
| using namespace GTM;
 | |
| 
 | |
| #if !defined(HAVE_ARCH_GTM_THREAD) || !defined(HAVE_ARCH_GTM_THREAD_DISP)
 | |
| extern __thread gtm_thread_tls _gtm_thr_tls;
 | |
| #endif
 | |
| 
 | |
| // Put this at the start of a cacheline so that serial_lock's writers and
 | |
| // htm_fastpath fields are on the same cacheline, so that HW transactions
 | |
| // only have to pay one cacheline capacity to monitor both.
 | |
| gtm_rwlock GTM::gtm_thread::serial_lock
 | |
|   __attribute__((aligned(HW_CACHELINE_SIZE)));
 | |
| gtm_thread *GTM::gtm_thread::list_of_threads = 0;
 | |
| unsigned GTM::gtm_thread::number_of_threads = 0;
 | |
| 
 | |
| /* ??? Move elsewhere when we figure out library initialization.  */
 | |
| uint64_t GTM::gtm_spin_count_var = 1000;
 | |
| 
 | |
| #ifdef HAVE_64BIT_SYNC_BUILTINS
 | |
| static atomic<_ITM_transactionId_t> global_tid;
 | |
| #else
 | |
| static _ITM_transactionId_t global_tid;
 | |
| static pthread_mutex_t global_tid_lock = PTHREAD_MUTEX_INITIALIZER;
 | |
| #endif
 | |
| 
 | |
| 
 | |
| // Provides a on-thread-exit callback used to release per-thread data.
 | |
| static pthread_key_t thr_release_key;
 | |
| static pthread_once_t thr_release_once = PTHREAD_ONCE_INIT;
 | |
| 
 | |
| /* Allocate a transaction structure.  */
 | |
| void *
 | |
| GTM::gtm_thread::operator new (size_t s)
 | |
| {
 | |
|   void *tx;
 | |
| 
 | |
|   assert(s == sizeof(gtm_thread));
 | |
| 
 | |
|   tx = xmalloc (sizeof (gtm_thread), true);
 | |
|   memset (tx, 0, sizeof (gtm_thread));
 | |
| 
 | |
|   return tx;
 | |
| }
 | |
| 
 | |
| /* Free the given transaction. Raises an error if the transaction is still
 | |
|    in use.  */
 | |
| void
 | |
| GTM::gtm_thread::operator delete(void *tx)
 | |
| {
 | |
|   free(tx);
 | |
| }
 | |
| 
 | |
| static void
 | |
| thread_exit_handler(void *)
 | |
| {
 | |
|   gtm_thread *thr = gtm_thr();
 | |
|   if (thr)
 | |
|     delete thr;
 | |
|   set_gtm_thr(0);
 | |
| }
 | |
| 
 | |
| static void
 | |
| thread_exit_init()
 | |
| {
 | |
|   if (pthread_key_create(&thr_release_key, thread_exit_handler))
 | |
|     GTM_fatal("Creating thread release TLS key failed.");
 | |
| }
 | |
| 
 | |
| 
 | |
| GTM::gtm_thread::~gtm_thread()
 | |
| {
 | |
|   if (nesting > 0)
 | |
|     GTM_fatal("Thread exit while a transaction is still active.");
 | |
| 
 | |
|   // Deregister this transaction.
 | |
|   serial_lock.write_lock ();
 | |
|   gtm_thread **prev = &list_of_threads;
 | |
|   for (; *prev; prev = &(*prev)->next_thread)
 | |
|     {
 | |
|       if (*prev == this)
 | |
| 	{
 | |
| 	  *prev = (*prev)->next_thread;
 | |
| 	  break;
 | |
| 	}
 | |
|     }
 | |
|   number_of_threads--;
 | |
|   number_of_threads_changed(number_of_threads + 1, number_of_threads);
 | |
|   serial_lock.write_unlock ();
 | |
| }
 | |
| 
 | |
| GTM::gtm_thread::gtm_thread ()
 | |
| {
 | |
|   // This object's memory has been set to zero by operator new, so no need
 | |
|   // to initialize any of the other primitive-type members that do not have
 | |
|   // constructors.
 | |
|   shared_state.store(-1, memory_order_relaxed);
 | |
| 
 | |
|   // Register this transaction with the list of all threads' transactions.
 | |
|   serial_lock.write_lock ();
 | |
|   next_thread = list_of_threads;
 | |
|   list_of_threads = this;
 | |
|   number_of_threads++;
 | |
|   number_of_threads_changed(number_of_threads - 1, number_of_threads);
 | |
|   serial_lock.write_unlock ();
 | |
| 
 | |
|   init_cpp_exceptions ();
 | |
| 
 | |
|   if (pthread_once(&thr_release_once, thread_exit_init))
 | |
|     GTM_fatal("Initializing thread release TLS key failed.");
 | |
|   // Any non-null value is sufficient to trigger destruction of this
 | |
|   // transaction when the current thread terminates.
 | |
|   if (pthread_setspecific(thr_release_key, this))
 | |
|     GTM_fatal("Setting thread release TLS key failed.");
 | |
| }
 | |
| 
 | |
| static inline uint32_t
 | |
| choose_code_path(uint32_t prop, abi_dispatch *disp)
 | |
| {
 | |
|   if ((prop & pr_uninstrumentedCode) && disp->can_run_uninstrumented_code())
 | |
|     return a_runUninstrumentedCode;
 | |
|   else
 | |
|     return a_runInstrumentedCode;
 | |
| }
 | |
| 
 | |
| #ifdef TARGET_BEGIN_TRANSACTION_ATTRIBUTE
 | |
| /* This macro can be used to define target specific attributes for this
 | |
|    function.  For example, S/390 requires floating point to be disabled in
 | |
|    begin_transaction.  */
 | |
| TARGET_BEGIN_TRANSACTION_ATTRIBUTE
 | |
| #endif
 | |
| uint32_t
 | |
| GTM::gtm_thread::begin_transaction (uint32_t prop, const gtm_jmpbuf *jb)
 | |
| {
 | |
|   static const _ITM_transactionId_t tid_block_size = 1 << 16;
 | |
| 
 | |
|   gtm_thread *tx;
 | |
|   abi_dispatch *disp;
 | |
|   uint32_t ret;
 | |
| 
 | |
|   // ??? pr_undoLogCode is not properly defined in the ABI. Are barriers
 | |
|   // omitted because they are not necessary (e.g., a transaction on thread-
 | |
|   // local data) or because the compiler thinks that some kind of global
 | |
|   // synchronization might perform better?
 | |
|   if (unlikely(prop & pr_undoLogCode))
 | |
|     GTM_fatal("pr_undoLogCode not supported");
 | |
| 
 | |
| #ifdef USE_HTM_FASTPATH
 | |
|   // HTM fastpath.  Only chosen in the absence of transaction_cancel to allow
 | |
|   // using an uninstrumented code path.
 | |
|   // The fastpath is enabled only by dispatch_htm's method group, which uses
 | |
|   // serial-mode methods as fallback.  Serial-mode transactions cannot execute
 | |
|   // concurrently with HW transactions because the latter monitor the serial
 | |
|   // lock's writer flag and thus abort if another thread is or becomes a
 | |
|   // serial transaction.  Therefore, if the fastpath is enabled, then a
 | |
|   // transaction is not executing as a HW transaction iff the serial lock is
 | |
|   // write-locked.  Also, HW transactions monitor the fastpath control
 | |
|   // variable, so that they will only execute if dispatch_htm is still the
 | |
|   // current method group.  This allows us to use htm_fastpath and the serial
 | |
|   // lock's writers flag to reliable determine whether the current thread runs
 | |
|   // a HW transaction, and thus we do not need to maintain this information in
 | |
|   // per-thread state.
 | |
|   // If an uninstrumented code path is not available, we can still run
 | |
|   // instrumented code from a HW transaction because the HTM fastpath kicks
 | |
|   // in early in both begin and commit, and the transaction is not canceled.
 | |
|   // HW transactions might get requests to switch to serial-irrevocable mode,
 | |
|   // but these can be ignored because the HTM provides all necessary
 | |
|   // correctness guarantees.  Transactions cannot detect whether they are
 | |
|   // indeed in serial mode, and HW transactions should never need serial mode
 | |
|   // for any internal changes (e.g., they never abort visibly to the STM code
 | |
|   // and thus do not trigger the standard retry handling).
 | |
| #ifndef HTM_CUSTOM_FASTPATH
 | |
|   if (likely(serial_lock.get_htm_fastpath() && (prop & pr_hasNoAbort)))
 | |
|     {
 | |
|       // Note that the snapshot of htm_fastpath that we take here could be
 | |
|       // outdated, and a different method group than dispatch_htm may have
 | |
|       // been chosen in the meantime.  Therefore, take care not not touch
 | |
|       // anything besides the serial lock, which is independent of method
 | |
|       // groups.
 | |
|       for (uint32_t t = serial_lock.get_htm_fastpath(); t; t--)
 | |
| 	{
 | |
| 	  uint32_t ret = htm_begin();
 | |
| 	  if (htm_begin_success(ret))
 | |
| 	    {
 | |
| 	      // We are executing a transaction now.
 | |
| 	      // Monitor the writer flag in the serial-mode lock, and abort
 | |
| 	      // if there is an active or waiting serial-mode transaction.
 | |
| 	      // Also checks that htm_fastpath is still nonzero and thus
 | |
| 	      // HW transactions are allowed to run.
 | |
| 	      // Note that this can also happen due to an enclosing
 | |
| 	      // serial-mode transaction; we handle this case below.
 | |
| 	      if (unlikely(serial_lock.htm_fastpath_disabled()))
 | |
| 		htm_abort();
 | |
| 	      else
 | |
| 		// We do not need to set a_saveLiveVariables because of HTM.
 | |
| 		return (prop & pr_uninstrumentedCode) ?
 | |
| 		    a_runUninstrumentedCode : a_runInstrumentedCode;
 | |
| 	    }
 | |
| 	  // The transaction has aborted.  Don't retry if it's unlikely that
 | |
| 	  // retrying the transaction will be successful.
 | |
| 	  if (!htm_abort_should_retry(ret))
 | |
| 	    break;
 | |
| 	  // Check whether the HTM fastpath has been disabled.
 | |
| 	  if (!serial_lock.get_htm_fastpath())
 | |
| 	    break;
 | |
| 	  // Wait until any concurrent serial-mode transactions have finished.
 | |
| 	  // This is an empty critical section, but won't be elided.
 | |
| 	  if (serial_lock.htm_fastpath_disabled())
 | |
| 	    {
 | |
| 	      tx = gtm_thr();
 | |
| 	      if (unlikely(tx == NULL))
 | |
| 	        {
 | |
| 	          // See below.
 | |
| 	          tx = new gtm_thread();
 | |
| 	          set_gtm_thr(tx);
 | |
| 	        }
 | |
| 	      // Check whether there is an enclosing serial-mode transaction;
 | |
| 	      // if so, we just continue as a nested transaction and don't
 | |
| 	      // try to use the HTM fastpath.  This case can happen when an
 | |
| 	      // outermost relaxed transaction calls unsafe code that starts
 | |
| 	      // a transaction.
 | |
| 	      if (tx->nesting > 0)
 | |
| 		break;
 | |
| 	      // Another thread is running a serial-mode transaction.  Wait.
 | |
| 	      serial_lock.read_lock(tx);
 | |
| 	      serial_lock.read_unlock(tx);
 | |
| 	      // TODO We should probably reset the retry count t here, unless
 | |
| 	      // we have retried so often that we should go serial to avoid
 | |
| 	      // starvation.
 | |
| 	    }
 | |
| 	}
 | |
|     }
 | |
| #else
 | |
|   // If we have a custom HTM fastpath in ITM_beginTransaction, we implement
 | |
|   // just the retry policy here.  We communicate with the custom fastpath
 | |
|   // through additional property bits and return codes, and either transfer
 | |
|   // control back to the custom fastpath or run the fallback mechanism.  The
 | |
|   // fastpath synchronization algorithm itself is the same.
 | |
|   // pr_HTMRetryableAbort states that a HW transaction started by the custom
 | |
|   // HTM fastpath aborted, and that we thus have to decide whether to retry
 | |
|   // the fastpath (returning a_tryHTMFastPath) or just proceed with the
 | |
|   // fallback method.
 | |
|   if (likely(serial_lock.get_htm_fastpath() && (prop & pr_HTMRetryableAbort)))
 | |
|     {
 | |
|       tx = gtm_thr();
 | |
|       if (unlikely(tx == NULL))
 | |
|         {
 | |
|           // See below.
 | |
|           tx = new gtm_thread();
 | |
|           set_gtm_thr(tx);
 | |
|         }
 | |
|       // If this is the first abort, reset the retry count.  We abuse
 | |
|       // restart_total for the retry count, which is fine because our only
 | |
|       // other fallback will use serial transactions, which don't use
 | |
|       // restart_total but will reset it when committing.
 | |
|       if (!(prop & pr_HTMRetriedAfterAbort))
 | |
| 	tx->restart_total = gtm_thread::serial_lock.get_htm_fastpath();
 | |
| 
 | |
|       if (--tx->restart_total > 0)
 | |
| 	{
 | |
| 	  // Wait until any concurrent serial-mode transactions have finished.
 | |
| 	  // Essentially the same code as above.
 | |
| 	  if (!serial_lock.get_htm_fastpath())
 | |
| 	    goto stop_custom_htm_fastpath;
 | |
| 	  if (serial_lock.htm_fastpath_disabled())
 | |
| 	    {
 | |
| 	      if (tx->nesting > 0)
 | |
| 		goto stop_custom_htm_fastpath;
 | |
| 	      serial_lock.read_lock(tx);
 | |
| 	      serial_lock.read_unlock(tx);
 | |
| 	    }
 | |
| 	  // Let ITM_beginTransaction retry the custom HTM fastpath.
 | |
| 	  return a_tryHTMFastPath;
 | |
| 	}
 | |
|     }
 | |
|  stop_custom_htm_fastpath:
 | |
| #endif
 | |
| #endif
 | |
| 
 | |
|   tx = gtm_thr();
 | |
|   if (unlikely(tx == NULL))
 | |
|     {
 | |
|       // Create the thread object. The constructor will also set up automatic
 | |
|       // deletion on thread termination.
 | |
|       tx = new gtm_thread();
 | |
|       set_gtm_thr(tx);
 | |
|     }
 | |
| 
 | |
|   if (tx->nesting > 0)
 | |
|     {
 | |
|       // This is a nested transaction.
 | |
|       // Check prop compatibility:
 | |
|       // The ABI requires pr_hasNoFloatUpdate, pr_hasNoVectorUpdate,
 | |
|       // pr_hasNoIrrevocable, pr_aWBarriersOmitted, pr_RaRBarriersOmitted, and
 | |
|       // pr_hasNoSimpleReads to hold for the full dynamic scope of a
 | |
|       // transaction. We could check that these are set for the nested
 | |
|       // transaction if they are also set for the parent transaction, but the
 | |
|       // ABI does not require these flags to be set if they could be set,
 | |
|       // so the check could be too strict.
 | |
|       // ??? For pr_readOnly, lexical or dynamic scope is unspecified.
 | |
| 
 | |
|       if (prop & pr_hasNoAbort)
 | |
| 	{
 | |
| 	  // We can use flat nesting, so elide this transaction.
 | |
| 	  if (!(prop & pr_instrumentedCode))
 | |
| 	    {
 | |
| 	      if (!(tx->state & STATE_SERIAL) ||
 | |
| 		  !(tx->state & STATE_IRREVOCABLE))
 | |
| 		tx->serialirr_mode();
 | |
| 	    }
 | |
| 	  // Increment nesting level after checking that we have a method that
 | |
| 	  // allows us to continue.
 | |
| 	  tx->nesting++;
 | |
| 	  return choose_code_path(prop, abi_disp());
 | |
| 	}
 | |
| 
 | |
|       // The transaction might abort, so use closed nesting if possible.
 | |
|       // pr_hasNoAbort has lexical scope, so the compiler should really have
 | |
|       // generated an instrumented code path.
 | |
|       assert(prop & pr_instrumentedCode);
 | |
| 
 | |
|       // Create a checkpoint of the current transaction.
 | |
|       gtm_transaction_cp *cp = tx->parent_txns.push();
 | |
|       cp->save(tx);
 | |
|       new (&tx->alloc_actions) aa_tree<uintptr_t, gtm_alloc_action>();
 | |
| 
 | |
|       // Check whether the current method actually supports closed nesting.
 | |
|       // If we can switch to another one, do so.
 | |
|       // If not, we assume that actual aborts are infrequent, and rather
 | |
|       // restart in _ITM_abortTransaction when we really have to.
 | |
|       disp = abi_disp();
 | |
|       if (!disp->closed_nesting())
 | |
| 	{
 | |
| 	  // ??? Should we elide the transaction if there is no alternative
 | |
| 	  // method that supports closed nesting? If we do, we need to set
 | |
| 	  // some flag to prevent _ITM_abortTransaction from aborting the
 | |
| 	  // wrong transaction (i.e., some parent transaction).
 | |
| 	  abi_dispatch *cn_disp = disp->closed_nesting_alternative();
 | |
| 	  if (cn_disp)
 | |
| 	    {
 | |
| 	      disp = cn_disp;
 | |
| 	      set_abi_disp(disp);
 | |
| 	    }
 | |
| 	}
 | |
|     }
 | |
|   else
 | |
|     {
 | |
|       // Outermost transaction
 | |
|       disp = tx->decide_begin_dispatch (prop);
 | |
|       set_abi_disp (disp);
 | |
|     }
 | |
| 
 | |
|   // Initialization that is common for outermost and nested transactions.
 | |
|   tx->prop = prop;
 | |
|   tx->nesting++;
 | |
| 
 | |
|   tx->jb = *jb;
 | |
| 
 | |
|   // As long as we have not exhausted a previously allocated block of TIDs,
 | |
|   // we can avoid an atomic operation on a shared cacheline.
 | |
|   if (tx->local_tid & (tid_block_size - 1))
 | |
|     tx->id = tx->local_tid++;
 | |
|   else
 | |
|     {
 | |
| #ifdef HAVE_64BIT_SYNC_BUILTINS
 | |
|       // We don't really care which block of TIDs we get but only that we
 | |
|       // acquire one atomically; therefore, relaxed memory order is
 | |
|       // sufficient.
 | |
|       tx->id = global_tid.fetch_add(tid_block_size, memory_order_relaxed);
 | |
|       tx->local_tid = tx->id + 1;
 | |
| #else
 | |
|       pthread_mutex_lock (&global_tid_lock);
 | |
|       global_tid += tid_block_size;
 | |
|       tx->id = global_tid;
 | |
|       tx->local_tid = tx->id + 1;
 | |
|       pthread_mutex_unlock (&global_tid_lock);
 | |
| #endif
 | |
|     }
 | |
| 
 | |
|   // Log the number of uncaught exceptions if we might have to roll back this
 | |
|   // state.
 | |
|   if (tx->cxa_uncaught_count_ptr != 0)
 | |
|     tx->cxa_uncaught_count = *tx->cxa_uncaught_count_ptr;
 | |
| 
 | |
|   // Run dispatch-specific restart code. Retry until we succeed.
 | |
|   GTM::gtm_restart_reason rr;
 | |
|   while ((rr = disp->begin_or_restart()) != NO_RESTART)
 | |
|     {
 | |
|       tx->decide_retry_strategy(rr);
 | |
|       disp = abi_disp();
 | |
|     }
 | |
| 
 | |
|   // Determine the code path to run. Only irrevocable transactions cannot be
 | |
|   // restarted, so all other transactions need to save live variables.
 | |
|   ret = choose_code_path(prop, disp);
 | |
|   if (!(tx->state & STATE_IRREVOCABLE))
 | |
|     ret |= a_saveLiveVariables;
 | |
|   return ret;
 | |
| }
 | |
| 
 | |
| 
 | |
| void
 | |
| GTM::gtm_transaction_cp::save(gtm_thread* tx)
 | |
| {
 | |
|   // Save everything that we might have to restore on restarts or aborts.
 | |
|   jb = tx->jb;
 | |
|   undolog_size = tx->undolog.size();
 | |
| 
 | |
|   /* FIXME!  Assignment of an aatree like alloc_actions is unsafe; if either
 | |
|    *this or *tx is destroyed, the other ends up pointing to a freed node.  */
 | |
| #pragma GCC diagnostic warning "-Wdeprecated-copy"
 | |
|   alloc_actions = tx->alloc_actions;
 | |
| 
 | |
|   user_actions_size = tx->user_actions.size();
 | |
|   id = tx->id;
 | |
|   prop = tx->prop;
 | |
|   cxa_catch_count = tx->cxa_catch_count;
 | |
|   cxa_uncaught_count = tx->cxa_uncaught_count;
 | |
|   disp = abi_disp();
 | |
|   nesting = tx->nesting;
 | |
| }
 | |
| 
 | |
| void
 | |
| GTM::gtm_transaction_cp::commit(gtm_thread* tx)
 | |
| {
 | |
|   // Restore state that is not persistent across commits. Exception handling,
 | |
|   // information, nesting level, and any logs do not need to be restored on
 | |
|   // commits of nested transactions. Allocation actions must be committed
 | |
|   // before committing the snapshot.
 | |
|   tx->jb = jb;
 | |
|   tx->alloc_actions = alloc_actions;
 | |
|   tx->id = id;
 | |
|   tx->prop = prop;
 | |
| }
 | |
| 
 | |
| 
 | |
| void
 | |
| GTM::gtm_thread::rollback (gtm_transaction_cp *cp, bool aborting)
 | |
| {
 | |
|   // The undo log is special in that it used for both thread-local and shared
 | |
|   // data. Because of the latter, we have to roll it back before any
 | |
|   // dispatch-specific rollback (which handles synchronization with other
 | |
|   // transactions).
 | |
|   undolog.rollback (this, cp ? cp->undolog_size : 0);
 | |
| 
 | |
|   // Perform dispatch-specific rollback.
 | |
|   abi_disp()->rollback (cp);
 | |
| 
 | |
|   // Roll back all actions that are supposed to happen around the transaction.
 | |
|   rollback_user_actions (cp ? cp->user_actions_size : 0);
 | |
|   commit_allocations (true, (cp ? &cp->alloc_actions : 0));
 | |
|   revert_cpp_exceptions (cp);
 | |
| 
 | |
|   if (cp)
 | |
|     {
 | |
|       // We do not yet handle restarts of nested transactions. To do that, we
 | |
|       // would have to restore some state (jb, id, prop, nesting) not to the
 | |
|       // checkpoint but to the transaction that was started from this
 | |
|       // checkpoint (e.g., nesting = cp->nesting + 1);
 | |
|       assert(aborting);
 | |
|       // Roll back the rest of the state to the checkpoint.
 | |
|       jb = cp->jb;
 | |
|       id = cp->id;
 | |
|       prop = cp->prop;
 | |
|       if (cp->disp != abi_disp())
 | |
| 	set_abi_disp(cp->disp);
 | |
|       alloc_actions = cp->alloc_actions;
 | |
|       nesting = cp->nesting;
 | |
|     }
 | |
|   else
 | |
|     {
 | |
|       // Roll back to the outermost transaction.
 | |
|       // Restore the jump buffer and transaction properties, which we will
 | |
|       // need for the longjmp used to restart or abort the transaction.
 | |
|       if (parent_txns.size() > 0)
 | |
| 	{
 | |
| 	  jb = parent_txns[0].jb;
 | |
| 	  id = parent_txns[0].id;
 | |
| 	  prop = parent_txns[0].prop;
 | |
| 	}
 | |
|       // Reset the transaction. Do not reset this->state, which is handled by
 | |
|       // the callers. Note that if we are not aborting, we reset the
 | |
|       // transaction to the point after having executed begin_transaction
 | |
|       // (we will return from it), so the nesting level must be one, not zero.
 | |
|       nesting = (aborting ? 0 : 1);
 | |
|       parent_txns.clear();
 | |
|     }
 | |
| 
 | |
|   if (this->eh_in_flight)
 | |
|     {
 | |
|       _Unwind_DeleteException ((_Unwind_Exception *) this->eh_in_flight);
 | |
|       this->eh_in_flight = NULL;
 | |
|     }
 | |
| }
 | |
| 
 | |
| void ITM_REGPARM
 | |
| _ITM_abortTransaction (_ITM_abortReason reason)
 | |
| {
 | |
|   gtm_thread *tx = gtm_thr();
 | |
| 
 | |
|   assert (reason == userAbort || reason == (userAbort | outerAbort));
 | |
|   assert ((tx->prop & pr_hasNoAbort) == 0);
 | |
| 
 | |
|   if (tx->state & gtm_thread::STATE_IRREVOCABLE)
 | |
|     abort ();
 | |
| 
 | |
|   // Roll back to innermost transaction.
 | |
|   if (tx->parent_txns.size() > 0 && !(reason & outerAbort))
 | |
|     {
 | |
|       // If the current method does not support closed nesting but we are
 | |
|       // nested and must only roll back the innermost transaction, then
 | |
|       // restart with a method that supports closed nesting.
 | |
|       abi_dispatch *disp = abi_disp();
 | |
|       if (!disp->closed_nesting())
 | |
| 	tx->restart(RESTART_CLOSED_NESTING);
 | |
| 
 | |
|       // The innermost transaction is a closed nested transaction.
 | |
|       gtm_transaction_cp *cp = tx->parent_txns.pop();
 | |
|       uint32_t longjmp_prop = tx->prop;
 | |
|       gtm_jmpbuf longjmp_jb = tx->jb;
 | |
| 
 | |
|       tx->rollback (cp, true);
 | |
| 
 | |
|       // Jump to nested transaction (use the saved jump buffer).
 | |
|       GTM_longjmp (a_abortTransaction | a_restoreLiveVariables,
 | |
| 		   &longjmp_jb, longjmp_prop);
 | |
|     }
 | |
|   else
 | |
|     {
 | |
|       // There is no nested transaction or an abort of the outermost
 | |
|       // transaction was requested, so roll back to the outermost transaction.
 | |
|       tx->rollback (0, true);
 | |
| 
 | |
|       // Aborting an outermost transaction finishes execution of the whole
 | |
|       // transaction. Therefore, reset transaction state.
 | |
|       if (tx->state & gtm_thread::STATE_SERIAL)
 | |
| 	gtm_thread::serial_lock.write_unlock ();
 | |
|       else
 | |
| 	gtm_thread::serial_lock.read_unlock (tx);
 | |
|       tx->state = 0;
 | |
| 
 | |
|       GTM_longjmp (a_abortTransaction | a_restoreLiveVariables,
 | |
| 		   &tx->jb, tx->prop);
 | |
|     }
 | |
| }
 | |
| 
 | |
| bool
 | |
| GTM::gtm_thread::trycommit ()
 | |
| {
 | |
|   nesting--;
 | |
| 
 | |
|   // Skip any real commit for elided transactions.
 | |
|   if (nesting > 0 && (parent_txns.size() == 0 ||
 | |
|       nesting > parent_txns[parent_txns.size() - 1].nesting))
 | |
|     return true;
 | |
| 
 | |
|   if (nesting > 0)
 | |
|     {
 | |
|       // Commit of a closed-nested transaction. Remove one checkpoint and add
 | |
|       // any effects of this transaction to the parent transaction.
 | |
|       gtm_transaction_cp *cp = parent_txns.pop();
 | |
|       commit_allocations(false, &cp->alloc_actions);
 | |
|       cp->commit(this);
 | |
|       return true;
 | |
|     }
 | |
| 
 | |
|   // Commit of an outermost transaction.
 | |
|   gtm_word priv_time = 0;
 | |
|   if (abi_disp()->trycommit (priv_time))
 | |
|     {
 | |
|       // The transaction is now finished but we will still access some shared
 | |
|       // data if we have to ensure privatization safety.
 | |
|       bool do_read_unlock = false;
 | |
|       if (state & gtm_thread::STATE_SERIAL)
 | |
|         {
 | |
|           gtm_thread::serial_lock.write_unlock ();
 | |
|           // There are no other active transactions, so there's no need to
 | |
|           // enforce privatization safety.
 | |
|           priv_time = 0;
 | |
|         }
 | |
|       else
 | |
| 	{
 | |
| 	  // If we have to ensure privatization safety, we must not yet
 | |
| 	  // release the read lock and become inactive because (1) we still
 | |
| 	  // have to go through the list of all transactions, which can be
 | |
| 	  // modified by serial mode threads, and (2) we interpret each
 | |
| 	  // transactions' shared_state in the context of what we believe to
 | |
| 	  // be the current method group (and serial mode transactions can
 | |
| 	  // change the method group).  Therefore, if we have to ensure
 | |
| 	  // privatization safety, delay becoming inactive but set a maximum
 | |
| 	  // snapshot time (we have committed and thus have an empty snapshot,
 | |
| 	  // so it will always be most recent).  Use release MO so that this
 | |
| 	  // synchronizes with other threads observing our snapshot time.
 | |
| 	  if (priv_time)
 | |
| 	    {
 | |
| 	      do_read_unlock = true;
 | |
| 	      shared_state.store((~(typeof gtm_thread::shared_state)0) - 1,
 | |
| 		  memory_order_release);
 | |
| 	    }
 | |
| 	  else
 | |
| 	    gtm_thread::serial_lock.read_unlock (this);
 | |
| 	}
 | |
|       state = 0;
 | |
| 
 | |
|       // We can commit the undo log after dispatch-specific commit and after
 | |
|       // making the transaction inactive because we only have to reset
 | |
|       // gtm_thread state.
 | |
|       undolog.commit ();
 | |
|       // Reset further transaction state.
 | |
|       cxa_catch_count = 0;
 | |
|       restart_total = 0;
 | |
| 
 | |
|       // Ensure privatization safety, if necessary.
 | |
|       if (priv_time)
 | |
| 	{
 | |
|           // There must be a seq_cst fence between the following loads of the
 | |
|           // other transactions' shared_state and the dispatch-specific stores
 | |
|           // that signal updates by this transaction (e.g., lock
 | |
|           // acquisitions).  This ensures that if we read prior to other
 | |
|           // reader transactions setting their shared_state to 0, then those
 | |
|           // readers will observe our updates.  We can reuse the seq_cst fence
 | |
|           // in serial_lock.read_unlock() if we performed that; if not, we
 | |
| 	  // issue the fence.
 | |
| 	  if (do_read_unlock)
 | |
| 	    atomic_thread_fence (memory_order_seq_cst);
 | |
| 	  // TODO Don't just spin but also block using cond vars / futexes
 | |
| 	  // here. Should probably be integrated with the serial lock code.
 | |
| 	  for (gtm_thread *it = gtm_thread::list_of_threads; it != 0;
 | |
| 	      it = it->next_thread)
 | |
| 	    {
 | |
| 	      if (it == this) continue;
 | |
| 	      // We need to load other threads' shared_state using acquire
 | |
| 	      // semantics (matching the release semantics of the respective
 | |
| 	      // updates).  This is necessary to ensure that the other
 | |
| 	      // threads' memory accesses happen before our actions that
 | |
| 	      // assume privatization safety.
 | |
| 	      // TODO Are there any platform-specific optimizations (e.g.,
 | |
| 	      // merging barriers)?
 | |
| 	      while (it->shared_state.load(memory_order_acquire) < priv_time)
 | |
| 		cpu_relax();
 | |
| 	    }
 | |
| 	}
 | |
| 
 | |
|       // After ensuring privatization safety, we are now truly inactive and
 | |
|       // thus can release the read lock.  We will also execute potentially
 | |
|       // privatizing actions (e.g., calling free()).  User actions are first.
 | |
|       if (do_read_unlock)
 | |
| 	gtm_thread::serial_lock.read_unlock (this);
 | |
|       commit_user_actions ();
 | |
|       commit_allocations (false, 0);
 | |
| 
 | |
|       return true;
 | |
|     }
 | |
|   return false;
 | |
| }
 | |
| 
 | |
| void ITM_NORETURN
 | |
| GTM::gtm_thread::restart (gtm_restart_reason r, bool finish_serial_upgrade)
 | |
| {
 | |
|   // Roll back to outermost transaction. Do not reset transaction state because
 | |
|   // we will continue executing this transaction.
 | |
|   rollback ();
 | |
| 
 | |
|   // If we have to restart while an upgrade of the serial lock is happening,
 | |
|   // we need to finish this here, after rollback (to ensure privatization
 | |
|   // safety despite undo writes) and before deciding about the retry strategy
 | |
|   // (which could switch to/from serial mode).
 | |
|   if (finish_serial_upgrade)
 | |
|     gtm_thread::serial_lock.write_upgrade_finish(this);
 | |
| 
 | |
|   decide_retry_strategy (r);
 | |
| 
 | |
|   // Run dispatch-specific restart code. Retry until we succeed.
 | |
|   abi_dispatch* disp = abi_disp();
 | |
|   GTM::gtm_restart_reason rr;
 | |
|   while ((rr = disp->begin_or_restart()) != NO_RESTART)
 | |
|     {
 | |
|       decide_retry_strategy(rr);
 | |
|       disp = abi_disp();
 | |
|     }
 | |
| 
 | |
|   GTM_longjmp (choose_code_path(prop, disp) | a_restoreLiveVariables,
 | |
| 	       &jb, prop);
 | |
| }
 | |
| 
 | |
| void ITM_REGPARM
 | |
| _ITM_commitTransaction(void)
 | |
| {
 | |
| #if defined(USE_HTM_FASTPATH)
 | |
|   // HTM fastpath.  If we are not executing a HW transaction, then we will be
 | |
|   // a serial-mode transaction.  If we are, then there will be no other
 | |
|   // concurrent serial-mode transaction.
 | |
|   // See gtm_thread::begin_transaction.
 | |
|   if (likely(!gtm_thread::serial_lock.htm_fastpath_disabled()))
 | |
|     {
 | |
|       htm_commit();
 | |
|       return;
 | |
|     }
 | |
| #endif
 | |
|   gtm_thread *tx = gtm_thr();
 | |
|   if (!tx->trycommit ())
 | |
|     tx->restart (RESTART_VALIDATE_COMMIT);
 | |
| }
 | |
| 
 | |
| void ITM_REGPARM
 | |
| _ITM_commitTransactionEH(void *exc_ptr)
 | |
| {
 | |
| #if defined(USE_HTM_FASTPATH)
 | |
|   // See _ITM_commitTransaction.
 | |
|   if (likely(!gtm_thread::serial_lock.htm_fastpath_disabled()))
 | |
|     {
 | |
|       htm_commit();
 | |
|       return;
 | |
|     }
 | |
| #endif
 | |
|   gtm_thread *tx = gtm_thr();
 | |
|   if (!tx->trycommit ())
 | |
|     {
 | |
|       tx->eh_in_flight = exc_ptr;
 | |
|       tx->restart (RESTART_VALIDATE_COMMIT);
 | |
|     }
 | |
| }
 |