mirror of
https://gcc.gnu.org/git/gcc.git
synced 2024-12-13 22:03:43 +08:00
1024 lines
30 KiB
C
1024 lines
30 KiB
C
/* Copyright (C) 2005-2022 Free Software Foundation, Inc.
|
|
Contributed by Richard Henderson <rth@redhat.com>.
|
|
|
|
This file is part of the GNU Offloading and Multi Processing Library
|
|
(libgomp).
|
|
|
|
Libgomp is free software; you can redistribute it and/or modify it
|
|
under the terms of the GNU General Public License as published by
|
|
the Free Software Foundation; either version 3, or (at your option)
|
|
any later version.
|
|
|
|
Libgomp is distributed in the hope that it will be useful, but WITHOUT ANY
|
|
WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
|
|
FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
more details.
|
|
|
|
Under Section 7 of GPL version 3, you are granted additional
|
|
permissions described in the GCC Runtime Library Exception, version
|
|
3.1, as published by the Free Software Foundation.
|
|
|
|
You should have received a copy of the GNU General Public License and
|
|
a copy of the GCC Runtime Library Exception along with this program;
|
|
see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
/* This file handles the LOOP (FOR/DO) construct. */
|
|
|
|
#include <limits.h>
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
#include "libgomp.h"
|
|
|
|
ialias (GOMP_loop_ull_runtime_next)
|
|
ialias_redirect (GOMP_taskgroup_reduction_register)
|
|
|
|
typedef unsigned long long gomp_ull;
|
|
|
|
/* Initialize the given work share construct from the given arguments. */
|
|
|
|
static inline void
|
|
gomp_loop_ull_init (struct gomp_work_share *ws, bool up, gomp_ull start,
|
|
gomp_ull end, gomp_ull incr, enum gomp_schedule_type sched,
|
|
gomp_ull chunk_size)
|
|
{
|
|
ws->sched = sched;
|
|
ws->chunk_size_ull = chunk_size;
|
|
/* Canonicalize loops that have zero iterations to ->next == ->end. */
|
|
ws->end_ull = ((up && start > end) || (!up && start < end))
|
|
? start : end;
|
|
ws->incr_ull = incr;
|
|
ws->next_ull = start;
|
|
ws->mode = 0;
|
|
if (sched == GFS_DYNAMIC)
|
|
{
|
|
ws->chunk_size_ull *= incr;
|
|
|
|
#if defined HAVE_SYNC_BUILTINS && defined __LP64__
|
|
{
|
|
/* For dynamic scheduling prepare things to make each iteration
|
|
faster. */
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
struct gomp_team *team = thr->ts.team;
|
|
long nthreads = team ? team->nthreads : 1;
|
|
|
|
if (__builtin_expect (up, 1))
|
|
{
|
|
/* Cheap overflow protection. */
|
|
if (__builtin_expect ((nthreads | ws->chunk_size_ull)
|
|
< 1ULL << (sizeof (gomp_ull)
|
|
* __CHAR_BIT__ / 2 - 1), 1))
|
|
ws->mode = ws->end_ull < (__LONG_LONG_MAX__ * 2ULL + 1
|
|
- (nthreads + 1) * ws->chunk_size_ull);
|
|
}
|
|
/* Cheap overflow protection. */
|
|
else if (__builtin_expect ((nthreads | -ws->chunk_size_ull)
|
|
< 1ULL << (sizeof (gomp_ull)
|
|
* __CHAR_BIT__ / 2 - 1), 1))
|
|
ws->mode = ws->end_ull > ((nthreads + 1) * -ws->chunk_size_ull
|
|
- (__LONG_LONG_MAX__ * 2ULL + 1));
|
|
}
|
|
#endif
|
|
}
|
|
if (!up)
|
|
ws->mode |= 2;
|
|
}
|
|
|
|
/* The *_start routines are called when first encountering a loop construct
|
|
that is not bound directly to a parallel construct. The first thread
|
|
that arrives will create the work-share construct; subsequent threads
|
|
will see the construct exists and allocate work from it.
|
|
|
|
START, END, INCR are the bounds of the loop; due to the restrictions of
|
|
OpenMP, these values must be the same in every thread. This is not
|
|
verified (nor is it entirely verifiable, since START is not necessarily
|
|
retained intact in the work-share data structure). CHUNK_SIZE is the
|
|
scheduling parameter; again this must be identical in all threads.
|
|
|
|
Returns true if there's any work for this thread to perform. If so,
|
|
*ISTART and *IEND are filled with the bounds of the iteration block
|
|
allocated to this thread. Returns false if all work was assigned to
|
|
other threads prior to this thread's arrival. */
|
|
|
|
static bool
|
|
gomp_loop_ull_static_start (bool up, gomp_ull start, gomp_ull end,
|
|
gomp_ull incr, gomp_ull chunk_size,
|
|
gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
|
|
thr->ts.static_trip = 0;
|
|
if (gomp_work_share_start (0))
|
|
{
|
|
gomp_loop_ull_init (thr->ts.work_share, up, start, end, incr,
|
|
GFS_STATIC, chunk_size);
|
|
gomp_work_share_init_done ();
|
|
}
|
|
|
|
return !gomp_iter_ull_static_next (istart, iend);
|
|
}
|
|
|
|
static bool
|
|
gomp_loop_ull_dynamic_start (bool up, gomp_ull start, gomp_ull end,
|
|
gomp_ull incr, gomp_ull chunk_size,
|
|
gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
bool ret;
|
|
|
|
if (gomp_work_share_start (0))
|
|
{
|
|
gomp_loop_ull_init (thr->ts.work_share, up, start, end, incr,
|
|
GFS_DYNAMIC, chunk_size);
|
|
gomp_work_share_init_done ();
|
|
}
|
|
|
|
#if defined HAVE_SYNC_BUILTINS && defined __LP64__
|
|
ret = gomp_iter_ull_dynamic_next (istart, iend);
|
|
#else
|
|
gomp_mutex_lock (&thr->ts.work_share->lock);
|
|
ret = gomp_iter_ull_dynamic_next_locked (istart, iend);
|
|
gomp_mutex_unlock (&thr->ts.work_share->lock);
|
|
#endif
|
|
|
|
return ret;
|
|
}
|
|
|
|
static bool
|
|
gomp_loop_ull_guided_start (bool up, gomp_ull start, gomp_ull end,
|
|
gomp_ull incr, gomp_ull chunk_size,
|
|
gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
bool ret;
|
|
|
|
if (gomp_work_share_start (0))
|
|
{
|
|
gomp_loop_ull_init (thr->ts.work_share, up, start, end, incr,
|
|
GFS_GUIDED, chunk_size);
|
|
gomp_work_share_init_done ();
|
|
}
|
|
|
|
#if defined HAVE_SYNC_BUILTINS && defined __LP64__
|
|
ret = gomp_iter_ull_guided_next (istart, iend);
|
|
#else
|
|
gomp_mutex_lock (&thr->ts.work_share->lock);
|
|
ret = gomp_iter_ull_guided_next_locked (istart, iend);
|
|
gomp_mutex_unlock (&thr->ts.work_share->lock);
|
|
#endif
|
|
|
|
return ret;
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_runtime_start (bool up, gomp_ull start, gomp_ull end,
|
|
gomp_ull incr, gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
struct gomp_task_icv *icv = gomp_icv (false);
|
|
switch (icv->run_sched_var & ~GFS_MONOTONIC)
|
|
{
|
|
case GFS_STATIC:
|
|
return gomp_loop_ull_static_start (up, start, end, incr,
|
|
icv->run_sched_chunk_size,
|
|
istart, iend);
|
|
case GFS_DYNAMIC:
|
|
return gomp_loop_ull_dynamic_start (up, start, end, incr,
|
|
icv->run_sched_chunk_size,
|
|
istart, iend);
|
|
case GFS_GUIDED:
|
|
return gomp_loop_ull_guided_start (up, start, end, incr,
|
|
icv->run_sched_chunk_size,
|
|
istart, iend);
|
|
case GFS_AUTO:
|
|
/* For now map to schedule(static), later on we could play with feedback
|
|
driven choice. */
|
|
return gomp_loop_ull_static_start (up, start, end, incr,
|
|
0, istart, iend);
|
|
default:
|
|
abort ();
|
|
}
|
|
}
|
|
|
|
static long
|
|
gomp_adjust_sched (long sched, gomp_ull *chunk_size)
|
|
{
|
|
sched &= ~GFS_MONOTONIC;
|
|
switch (sched)
|
|
{
|
|
case GFS_STATIC:
|
|
case GFS_DYNAMIC:
|
|
case GFS_GUIDED:
|
|
return sched;
|
|
/* GFS_RUNTIME is used for runtime schedule without monotonic
|
|
or nonmonotonic modifiers on the clause.
|
|
GFS_RUNTIME|GFS_MONOTONIC for runtime schedule with monotonic
|
|
modifier. */
|
|
case GFS_RUNTIME:
|
|
/* GFS_AUTO is used for runtime schedule with nonmonotonic
|
|
modifier. */
|
|
case GFS_AUTO:
|
|
{
|
|
struct gomp_task_icv *icv = gomp_icv (false);
|
|
sched = icv->run_sched_var & ~GFS_MONOTONIC;
|
|
switch (sched)
|
|
{
|
|
case GFS_STATIC:
|
|
case GFS_DYNAMIC:
|
|
case GFS_GUIDED:
|
|
*chunk_size = icv->run_sched_chunk_size;
|
|
break;
|
|
case GFS_AUTO:
|
|
sched = GFS_STATIC;
|
|
*chunk_size = 0;
|
|
break;
|
|
default:
|
|
abort ();
|
|
}
|
|
return sched;
|
|
}
|
|
default:
|
|
abort ();
|
|
}
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_start (bool up, gomp_ull start, gomp_ull end,
|
|
gomp_ull incr, long sched, gomp_ull chunk_size,
|
|
gomp_ull *istart, gomp_ull *iend,
|
|
uintptr_t *reductions, void **mem)
|
|
{
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
|
|
thr->ts.static_trip = 0;
|
|
if (reductions)
|
|
gomp_workshare_taskgroup_start ();
|
|
if (gomp_work_share_start (0))
|
|
{
|
|
sched = gomp_adjust_sched (sched, &chunk_size);
|
|
gomp_loop_ull_init (thr->ts.work_share, up, start, end, incr,
|
|
sched, chunk_size);
|
|
if (reductions)
|
|
{
|
|
GOMP_taskgroup_reduction_register (reductions);
|
|
thr->task->taskgroup->workshare = true;
|
|
thr->ts.work_share->task_reductions = reductions;
|
|
}
|
|
if (mem)
|
|
{
|
|
uintptr_t size = (uintptr_t) *mem;
|
|
#define INLINE_ORDERED_TEAM_IDS_OFF \
|
|
((offsetof (struct gomp_work_share, inline_ordered_team_ids) \
|
|
+ __alignof__ (long long) - 1) & ~(__alignof__ (long long) - 1))
|
|
if (size > (sizeof (struct gomp_work_share)
|
|
- INLINE_ORDERED_TEAM_IDS_OFF))
|
|
*mem
|
|
= (void *) (thr->ts.work_share->ordered_team_ids
|
|
= gomp_malloc_cleared (size));
|
|
else
|
|
*mem = memset (((char *) thr->ts.work_share)
|
|
+ INLINE_ORDERED_TEAM_IDS_OFF, '\0', size);
|
|
}
|
|
gomp_work_share_init_done ();
|
|
}
|
|
else
|
|
{
|
|
if (reductions)
|
|
{
|
|
uintptr_t *first_reductions = thr->ts.work_share->task_reductions;
|
|
gomp_workshare_task_reduction_register (reductions,
|
|
first_reductions);
|
|
}
|
|
if (mem)
|
|
{
|
|
if ((offsetof (struct gomp_work_share, inline_ordered_team_ids)
|
|
& (__alignof__ (long long) - 1)) == 0)
|
|
*mem = (void *) thr->ts.work_share->ordered_team_ids;
|
|
else
|
|
{
|
|
uintptr_t p = (uintptr_t) thr->ts.work_share->ordered_team_ids;
|
|
p += __alignof__ (long long) - 1;
|
|
p &= ~(__alignof__ (long long) - 1);
|
|
*mem = (void *) p;
|
|
}
|
|
}
|
|
}
|
|
|
|
return ialias_call (GOMP_loop_ull_runtime_next) (istart, iend);
|
|
}
|
|
|
|
/* The *_ordered_*_start routines are similar. The only difference is that
|
|
this work-share construct is initialized to expect an ORDERED section. */
|
|
|
|
static bool
|
|
gomp_loop_ull_ordered_static_start (bool up, gomp_ull start, gomp_ull end,
|
|
gomp_ull incr, gomp_ull chunk_size,
|
|
gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
|
|
thr->ts.static_trip = 0;
|
|
if (gomp_work_share_start (1))
|
|
{
|
|
gomp_loop_ull_init (thr->ts.work_share, up, start, end, incr,
|
|
GFS_STATIC, chunk_size);
|
|
gomp_ordered_static_init ();
|
|
gomp_work_share_init_done ();
|
|
}
|
|
|
|
return !gomp_iter_ull_static_next (istart, iend);
|
|
}
|
|
|
|
static bool
|
|
gomp_loop_ull_ordered_dynamic_start (bool up, gomp_ull start, gomp_ull end,
|
|
gomp_ull incr, gomp_ull chunk_size,
|
|
gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
bool ret;
|
|
|
|
if (gomp_work_share_start (1))
|
|
{
|
|
gomp_loop_ull_init (thr->ts.work_share, up, start, end, incr,
|
|
GFS_DYNAMIC, chunk_size);
|
|
gomp_mutex_lock (&thr->ts.work_share->lock);
|
|
gomp_work_share_init_done ();
|
|
}
|
|
else
|
|
gomp_mutex_lock (&thr->ts.work_share->lock);
|
|
|
|
ret = gomp_iter_ull_dynamic_next_locked (istart, iend);
|
|
if (ret)
|
|
gomp_ordered_first ();
|
|
gomp_mutex_unlock (&thr->ts.work_share->lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static bool
|
|
gomp_loop_ull_ordered_guided_start (bool up, gomp_ull start, gomp_ull end,
|
|
gomp_ull incr, gomp_ull chunk_size,
|
|
gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
bool ret;
|
|
|
|
if (gomp_work_share_start (1))
|
|
{
|
|
gomp_loop_ull_init (thr->ts.work_share, up, start, end, incr,
|
|
GFS_GUIDED, chunk_size);
|
|
gomp_mutex_lock (&thr->ts.work_share->lock);
|
|
gomp_work_share_init_done ();
|
|
}
|
|
else
|
|
gomp_mutex_lock (&thr->ts.work_share->lock);
|
|
|
|
ret = gomp_iter_ull_guided_next_locked (istart, iend);
|
|
if (ret)
|
|
gomp_ordered_first ();
|
|
gomp_mutex_unlock (&thr->ts.work_share->lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_ordered_runtime_start (bool up, gomp_ull start, gomp_ull end,
|
|
gomp_ull incr, gomp_ull *istart,
|
|
gomp_ull *iend)
|
|
{
|
|
struct gomp_task_icv *icv = gomp_icv (false);
|
|
switch (icv->run_sched_var & ~GFS_MONOTONIC)
|
|
{
|
|
case GFS_STATIC:
|
|
return gomp_loop_ull_ordered_static_start (up, start, end, incr,
|
|
icv->run_sched_chunk_size,
|
|
istart, iend);
|
|
case GFS_DYNAMIC:
|
|
return gomp_loop_ull_ordered_dynamic_start (up, start, end, incr,
|
|
icv->run_sched_chunk_size,
|
|
istart, iend);
|
|
case GFS_GUIDED:
|
|
return gomp_loop_ull_ordered_guided_start (up, start, end, incr,
|
|
icv->run_sched_chunk_size,
|
|
istart, iend);
|
|
case GFS_AUTO:
|
|
/* For now map to schedule(static), later on we could play with feedback
|
|
driven choice. */
|
|
return gomp_loop_ull_ordered_static_start (up, start, end, incr,
|
|
0, istart, iend);
|
|
default:
|
|
abort ();
|
|
}
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_ordered_start (bool up, gomp_ull start, gomp_ull end,
|
|
gomp_ull incr, long sched, gomp_ull chunk_size,
|
|
gomp_ull *istart, gomp_ull *iend,
|
|
uintptr_t *reductions, void **mem)
|
|
{
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
size_t ordered = 1;
|
|
bool ret;
|
|
|
|
thr->ts.static_trip = 0;
|
|
if (reductions)
|
|
gomp_workshare_taskgroup_start ();
|
|
if (mem)
|
|
ordered += (uintptr_t) *mem;
|
|
if (gomp_work_share_start (ordered))
|
|
{
|
|
sched = gomp_adjust_sched (sched, &chunk_size);
|
|
gomp_loop_ull_init (thr->ts.work_share, up, start, end, incr,
|
|
sched, chunk_size);
|
|
if (reductions)
|
|
{
|
|
GOMP_taskgroup_reduction_register (reductions);
|
|
thr->task->taskgroup->workshare = true;
|
|
thr->ts.work_share->task_reductions = reductions;
|
|
}
|
|
if (sched == GFS_STATIC)
|
|
gomp_ordered_static_init ();
|
|
else
|
|
gomp_mutex_lock (&thr->ts.work_share->lock);
|
|
gomp_work_share_init_done ();
|
|
}
|
|
else
|
|
{
|
|
if (reductions)
|
|
{
|
|
uintptr_t *first_reductions = thr->ts.work_share->task_reductions;
|
|
gomp_workshare_task_reduction_register (reductions,
|
|
first_reductions);
|
|
}
|
|
sched = thr->ts.work_share->sched;
|
|
if (sched != GFS_STATIC)
|
|
gomp_mutex_lock (&thr->ts.work_share->lock);
|
|
}
|
|
|
|
if (mem)
|
|
{
|
|
uintptr_t p
|
|
= (uintptr_t) (thr->ts.work_share->ordered_team_ids
|
|
+ (thr->ts.team ? thr->ts.team->nthreads : 1));
|
|
p += __alignof__ (long long) - 1;
|
|
p &= ~(__alignof__ (long long) - 1);
|
|
*mem = (void *) p;
|
|
}
|
|
|
|
switch (sched)
|
|
{
|
|
case GFS_STATIC:
|
|
case GFS_AUTO:
|
|
return !gomp_iter_ull_static_next (istart, iend);
|
|
case GFS_DYNAMIC:
|
|
ret = gomp_iter_ull_dynamic_next_locked (istart, iend);
|
|
break;
|
|
case GFS_GUIDED:
|
|
ret = gomp_iter_ull_guided_next_locked (istart, iend);
|
|
break;
|
|
default:
|
|
abort ();
|
|
}
|
|
|
|
if (ret)
|
|
gomp_ordered_first ();
|
|
gomp_mutex_unlock (&thr->ts.work_share->lock);
|
|
return ret;
|
|
}
|
|
|
|
/* The *_doacross_*_start routines are similar. The only difference is that
|
|
this work-share construct is initialized to expect an ORDERED(N) - DOACROSS
|
|
section, and the worksharing loop iterates always from 0 to COUNTS[0] - 1
|
|
and other COUNTS array elements tell the library number of iterations
|
|
in the ordered inner loops. */
|
|
|
|
static bool
|
|
gomp_loop_ull_doacross_static_start (unsigned ncounts, gomp_ull *counts,
|
|
gomp_ull chunk_size, gomp_ull *istart,
|
|
gomp_ull *iend)
|
|
{
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
|
|
thr->ts.static_trip = 0;
|
|
if (gomp_work_share_start (0))
|
|
{
|
|
gomp_loop_ull_init (thr->ts.work_share, true, 0, counts[0], 1,
|
|
GFS_STATIC, chunk_size);
|
|
gomp_doacross_ull_init (ncounts, counts, chunk_size, 0);
|
|
gomp_work_share_init_done ();
|
|
}
|
|
|
|
return !gomp_iter_ull_static_next (istart, iend);
|
|
}
|
|
|
|
static bool
|
|
gomp_loop_ull_doacross_dynamic_start (unsigned ncounts, gomp_ull *counts,
|
|
gomp_ull chunk_size, gomp_ull *istart,
|
|
gomp_ull *iend)
|
|
{
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
bool ret;
|
|
|
|
if (gomp_work_share_start (0))
|
|
{
|
|
gomp_loop_ull_init (thr->ts.work_share, true, 0, counts[0], 1,
|
|
GFS_DYNAMIC, chunk_size);
|
|
gomp_doacross_ull_init (ncounts, counts, chunk_size, 0);
|
|
gomp_work_share_init_done ();
|
|
}
|
|
|
|
#if defined HAVE_SYNC_BUILTINS && defined __LP64__
|
|
ret = gomp_iter_ull_dynamic_next (istart, iend);
|
|
#else
|
|
gomp_mutex_lock (&thr->ts.work_share->lock);
|
|
ret = gomp_iter_ull_dynamic_next_locked (istart, iend);
|
|
gomp_mutex_unlock (&thr->ts.work_share->lock);
|
|
#endif
|
|
|
|
return ret;
|
|
}
|
|
|
|
static bool
|
|
gomp_loop_ull_doacross_guided_start (unsigned ncounts, gomp_ull *counts,
|
|
gomp_ull chunk_size, gomp_ull *istart,
|
|
gomp_ull *iend)
|
|
{
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
bool ret;
|
|
|
|
if (gomp_work_share_start (0))
|
|
{
|
|
gomp_loop_ull_init (thr->ts.work_share, true, 0, counts[0], 1,
|
|
GFS_GUIDED, chunk_size);
|
|
gomp_doacross_ull_init (ncounts, counts, chunk_size, 0);
|
|
gomp_work_share_init_done ();
|
|
}
|
|
|
|
#if defined HAVE_SYNC_BUILTINS && defined __LP64__
|
|
ret = gomp_iter_ull_guided_next (istart, iend);
|
|
#else
|
|
gomp_mutex_lock (&thr->ts.work_share->lock);
|
|
ret = gomp_iter_ull_guided_next_locked (istart, iend);
|
|
gomp_mutex_unlock (&thr->ts.work_share->lock);
|
|
#endif
|
|
|
|
return ret;
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_doacross_runtime_start (unsigned ncounts, gomp_ull *counts,
|
|
gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
struct gomp_task_icv *icv = gomp_icv (false);
|
|
switch (icv->run_sched_var & ~GFS_MONOTONIC)
|
|
{
|
|
case GFS_STATIC:
|
|
return gomp_loop_ull_doacross_static_start (ncounts, counts,
|
|
icv->run_sched_chunk_size,
|
|
istart, iend);
|
|
case GFS_DYNAMIC:
|
|
return gomp_loop_ull_doacross_dynamic_start (ncounts, counts,
|
|
icv->run_sched_chunk_size,
|
|
istart, iend);
|
|
case GFS_GUIDED:
|
|
return gomp_loop_ull_doacross_guided_start (ncounts, counts,
|
|
icv->run_sched_chunk_size,
|
|
istart, iend);
|
|
case GFS_AUTO:
|
|
/* For now map to schedule(static), later on we could play with feedback
|
|
driven choice. */
|
|
return gomp_loop_ull_doacross_static_start (ncounts, counts,
|
|
0, istart, iend);
|
|
default:
|
|
abort ();
|
|
}
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_doacross_start (unsigned ncounts, gomp_ull *counts,
|
|
long sched, gomp_ull chunk_size,
|
|
gomp_ull *istart, gomp_ull *iend,
|
|
uintptr_t *reductions, void **mem)
|
|
{
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
|
|
thr->ts.static_trip = 0;
|
|
if (reductions)
|
|
gomp_workshare_taskgroup_start ();
|
|
if (gomp_work_share_start (0))
|
|
{
|
|
size_t extra = 0;
|
|
if (mem)
|
|
extra = (uintptr_t) *mem;
|
|
sched = gomp_adjust_sched (sched, &chunk_size);
|
|
gomp_loop_ull_init (thr->ts.work_share, true, 0, counts[0], 1,
|
|
sched, chunk_size);
|
|
gomp_doacross_ull_init (ncounts, counts, chunk_size, extra);
|
|
if (reductions)
|
|
{
|
|
GOMP_taskgroup_reduction_register (reductions);
|
|
thr->task->taskgroup->workshare = true;
|
|
thr->ts.work_share->task_reductions = reductions;
|
|
}
|
|
gomp_work_share_init_done ();
|
|
}
|
|
else
|
|
{
|
|
if (reductions)
|
|
{
|
|
uintptr_t *first_reductions = thr->ts.work_share->task_reductions;
|
|
gomp_workshare_task_reduction_register (reductions,
|
|
first_reductions);
|
|
}
|
|
sched = thr->ts.work_share->sched;
|
|
}
|
|
|
|
if (mem)
|
|
*mem = thr->ts.work_share->doacross->extra;
|
|
|
|
return ialias_call (GOMP_loop_ull_runtime_next) (istart, iend);
|
|
}
|
|
|
|
/* The *_next routines are called when the thread completes processing of
|
|
the iteration block currently assigned to it. If the work-share
|
|
construct is bound directly to a parallel construct, then the iteration
|
|
bounds may have been set up before the parallel. In which case, this
|
|
may be the first iteration for the thread.
|
|
|
|
Returns true if there is work remaining to be performed; *ISTART and
|
|
*IEND are filled with a new iteration block. Returns false if all work
|
|
has been assigned. */
|
|
|
|
static bool
|
|
gomp_loop_ull_static_next (gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return !gomp_iter_ull_static_next (istart, iend);
|
|
}
|
|
|
|
static bool
|
|
gomp_loop_ull_dynamic_next (gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
bool ret;
|
|
|
|
#if defined HAVE_SYNC_BUILTINS && defined __LP64__
|
|
ret = gomp_iter_ull_dynamic_next (istart, iend);
|
|
#else
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
gomp_mutex_lock (&thr->ts.work_share->lock);
|
|
ret = gomp_iter_ull_dynamic_next_locked (istart, iend);
|
|
gomp_mutex_unlock (&thr->ts.work_share->lock);
|
|
#endif
|
|
|
|
return ret;
|
|
}
|
|
|
|
static bool
|
|
gomp_loop_ull_guided_next (gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
bool ret;
|
|
|
|
#if defined HAVE_SYNC_BUILTINS && defined __LP64__
|
|
ret = gomp_iter_ull_guided_next (istart, iend);
|
|
#else
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
gomp_mutex_lock (&thr->ts.work_share->lock);
|
|
ret = gomp_iter_ull_guided_next_locked (istart, iend);
|
|
gomp_mutex_unlock (&thr->ts.work_share->lock);
|
|
#endif
|
|
|
|
return ret;
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_runtime_next (gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
|
|
switch (thr->ts.work_share->sched)
|
|
{
|
|
case GFS_STATIC:
|
|
case GFS_AUTO:
|
|
return gomp_loop_ull_static_next (istart, iend);
|
|
case GFS_DYNAMIC:
|
|
return gomp_loop_ull_dynamic_next (istart, iend);
|
|
case GFS_GUIDED:
|
|
return gomp_loop_ull_guided_next (istart, iend);
|
|
default:
|
|
abort ();
|
|
}
|
|
}
|
|
|
|
/* The *_ordered_*_next routines are called when the thread completes
|
|
processing of the iteration block currently assigned to it.
|
|
|
|
Returns true if there is work remaining to be performed; *ISTART and
|
|
*IEND are filled with a new iteration block. Returns false if all work
|
|
has been assigned. */
|
|
|
|
static bool
|
|
gomp_loop_ull_ordered_static_next (gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
int test;
|
|
|
|
gomp_ordered_sync ();
|
|
gomp_mutex_lock (&thr->ts.work_share->lock);
|
|
test = gomp_iter_ull_static_next (istart, iend);
|
|
if (test >= 0)
|
|
gomp_ordered_static_next ();
|
|
gomp_mutex_unlock (&thr->ts.work_share->lock);
|
|
|
|
return test == 0;
|
|
}
|
|
|
|
static bool
|
|
gomp_loop_ull_ordered_dynamic_next (gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
bool ret;
|
|
|
|
gomp_ordered_sync ();
|
|
gomp_mutex_lock (&thr->ts.work_share->lock);
|
|
ret = gomp_iter_ull_dynamic_next_locked (istart, iend);
|
|
if (ret)
|
|
gomp_ordered_next ();
|
|
else
|
|
gomp_ordered_last ();
|
|
gomp_mutex_unlock (&thr->ts.work_share->lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static bool
|
|
gomp_loop_ull_ordered_guided_next (gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
bool ret;
|
|
|
|
gomp_ordered_sync ();
|
|
gomp_mutex_lock (&thr->ts.work_share->lock);
|
|
ret = gomp_iter_ull_guided_next_locked (istart, iend);
|
|
if (ret)
|
|
gomp_ordered_next ();
|
|
else
|
|
gomp_ordered_last ();
|
|
gomp_mutex_unlock (&thr->ts.work_share->lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_ordered_runtime_next (gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
|
|
switch (thr->ts.work_share->sched)
|
|
{
|
|
case GFS_STATIC:
|
|
case GFS_AUTO:
|
|
return gomp_loop_ull_ordered_static_next (istart, iend);
|
|
case GFS_DYNAMIC:
|
|
return gomp_loop_ull_ordered_dynamic_next (istart, iend);
|
|
case GFS_GUIDED:
|
|
return gomp_loop_ull_ordered_guided_next (istart, iend);
|
|
default:
|
|
abort ();
|
|
}
|
|
}
|
|
|
|
/* We use static functions above so that we're sure that the "runtime"
|
|
function can defer to the proper routine without interposition. We
|
|
export the static function with a strong alias when possible, or with
|
|
a wrapper function otherwise. */
|
|
|
|
#ifdef HAVE_ATTRIBUTE_ALIAS
|
|
extern __typeof(gomp_loop_ull_static_start) GOMP_loop_ull_static_start
|
|
__attribute__((alias ("gomp_loop_ull_static_start")));
|
|
extern __typeof(gomp_loop_ull_dynamic_start) GOMP_loop_ull_dynamic_start
|
|
__attribute__((alias ("gomp_loop_ull_dynamic_start")));
|
|
extern __typeof(gomp_loop_ull_guided_start) GOMP_loop_ull_guided_start
|
|
__attribute__((alias ("gomp_loop_ull_guided_start")));
|
|
extern __typeof(gomp_loop_ull_dynamic_start) GOMP_loop_ull_nonmonotonic_dynamic_start
|
|
__attribute__((alias ("gomp_loop_ull_dynamic_start")));
|
|
extern __typeof(gomp_loop_ull_guided_start) GOMP_loop_ull_nonmonotonic_guided_start
|
|
__attribute__((alias ("gomp_loop_ull_guided_start")));
|
|
extern __typeof(GOMP_loop_ull_runtime_start) GOMP_loop_ull_nonmonotonic_runtime_start
|
|
__attribute__((alias ("GOMP_loop_ull_runtime_start")));
|
|
extern __typeof(GOMP_loop_ull_runtime_start) GOMP_loop_ull_maybe_nonmonotonic_runtime_start
|
|
__attribute__((alias ("GOMP_loop_ull_runtime_start")));
|
|
|
|
extern __typeof(gomp_loop_ull_ordered_static_start) GOMP_loop_ull_ordered_static_start
|
|
__attribute__((alias ("gomp_loop_ull_ordered_static_start")));
|
|
extern __typeof(gomp_loop_ull_ordered_dynamic_start) GOMP_loop_ull_ordered_dynamic_start
|
|
__attribute__((alias ("gomp_loop_ull_ordered_dynamic_start")));
|
|
extern __typeof(gomp_loop_ull_ordered_guided_start) GOMP_loop_ull_ordered_guided_start
|
|
__attribute__((alias ("gomp_loop_ull_ordered_guided_start")));
|
|
|
|
extern __typeof(gomp_loop_ull_doacross_static_start) GOMP_loop_ull_doacross_static_start
|
|
__attribute__((alias ("gomp_loop_ull_doacross_static_start")));
|
|
extern __typeof(gomp_loop_ull_doacross_dynamic_start) GOMP_loop_ull_doacross_dynamic_start
|
|
__attribute__((alias ("gomp_loop_ull_doacross_dynamic_start")));
|
|
extern __typeof(gomp_loop_ull_doacross_guided_start) GOMP_loop_ull_doacross_guided_start
|
|
__attribute__((alias ("gomp_loop_ull_doacross_guided_start")));
|
|
|
|
extern __typeof(gomp_loop_ull_static_next) GOMP_loop_ull_static_next
|
|
__attribute__((alias ("gomp_loop_ull_static_next")));
|
|
extern __typeof(gomp_loop_ull_dynamic_next) GOMP_loop_ull_dynamic_next
|
|
__attribute__((alias ("gomp_loop_ull_dynamic_next")));
|
|
extern __typeof(gomp_loop_ull_guided_next) GOMP_loop_ull_guided_next
|
|
__attribute__((alias ("gomp_loop_ull_guided_next")));
|
|
extern __typeof(gomp_loop_ull_dynamic_next) GOMP_loop_ull_nonmonotonic_dynamic_next
|
|
__attribute__((alias ("gomp_loop_ull_dynamic_next")));
|
|
extern __typeof(gomp_loop_ull_guided_next) GOMP_loop_ull_nonmonotonic_guided_next
|
|
__attribute__((alias ("gomp_loop_ull_guided_next")));
|
|
extern __typeof(GOMP_loop_ull_runtime_next) GOMP_loop_ull_nonmonotonic_runtime_next
|
|
__attribute__((alias ("GOMP_loop_ull_runtime_next")));
|
|
extern __typeof(GOMP_loop_ull_runtime_next) GOMP_loop_ull_maybe_nonmonotonic_runtime_next
|
|
__attribute__((alias ("GOMP_loop_ull_runtime_next")));
|
|
|
|
extern __typeof(gomp_loop_ull_ordered_static_next) GOMP_loop_ull_ordered_static_next
|
|
__attribute__((alias ("gomp_loop_ull_ordered_static_next")));
|
|
extern __typeof(gomp_loop_ull_ordered_dynamic_next) GOMP_loop_ull_ordered_dynamic_next
|
|
__attribute__((alias ("gomp_loop_ull_ordered_dynamic_next")));
|
|
extern __typeof(gomp_loop_ull_ordered_guided_next) GOMP_loop_ull_ordered_guided_next
|
|
__attribute__((alias ("gomp_loop_ull_ordered_guided_next")));
|
|
#else
|
|
bool
|
|
GOMP_loop_ull_static_start (bool up, gomp_ull start, gomp_ull end,
|
|
gomp_ull incr, gomp_ull chunk_size,
|
|
gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_static_start (up, start, end, incr, chunk_size, istart,
|
|
iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_dynamic_start (bool up, gomp_ull start, gomp_ull end,
|
|
gomp_ull incr, gomp_ull chunk_size,
|
|
gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_dynamic_start (up, start, end, incr, chunk_size, istart,
|
|
iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_guided_start (bool up, gomp_ull start, gomp_ull end,
|
|
gomp_ull incr, gomp_ull chunk_size,
|
|
gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_guided_start (up, start, end, incr, chunk_size, istart,
|
|
iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_nonmonotonic_dynamic_start (bool up, gomp_ull start,
|
|
gomp_ull end, gomp_ull incr,
|
|
gomp_ull chunk_size,
|
|
gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_dynamic_start (up, start, end, incr, chunk_size, istart,
|
|
iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_nonmonotonic_guided_start (bool up, gomp_ull start, gomp_ull end,
|
|
gomp_ull incr, gomp_ull chunk_size,
|
|
gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_guided_start (up, start, end, incr, chunk_size, istart,
|
|
iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_nonmonotonic_runtime_start (bool up, gomp_ull start,
|
|
gomp_ull end, gomp_ull incr,
|
|
gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return GOMP_loop_ull_runtime_start (up, start, end, incr, istart, iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_maybe_nonmonotonic_runtime_start (bool up, gomp_ull start,
|
|
gomp_ull end, gomp_ull incr,
|
|
gomp_ull *istart,
|
|
gomp_ull *iend)
|
|
{
|
|
return GOMP_loop_ull_runtime_start (up, start, end, incr, istart, iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_ordered_static_start (bool up, gomp_ull start, gomp_ull end,
|
|
gomp_ull incr, gomp_ull chunk_size,
|
|
gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_ordered_static_start (up, start, end, incr, chunk_size,
|
|
istart, iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_ordered_dynamic_start (bool up, gomp_ull start, gomp_ull end,
|
|
gomp_ull incr, gomp_ull chunk_size,
|
|
gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_ordered_dynamic_start (up, start, end, incr, chunk_size,
|
|
istart, iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_ordered_guided_start (bool up, gomp_ull start, gomp_ull end,
|
|
gomp_ull incr, gomp_ull chunk_size,
|
|
gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_ordered_guided_start (up, start, end, incr, chunk_size,
|
|
istart, iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_doacross_static_start (unsigned ncounts, gomp_ull *counts,
|
|
gomp_ull chunk_size, gomp_ull *istart,
|
|
gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_doacross_static_start (ncounts, counts, chunk_size,
|
|
istart, iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_doacross_dynamic_start (unsigned ncounts, gomp_ull *counts,
|
|
gomp_ull chunk_size, gomp_ull *istart,
|
|
gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_doacross_dynamic_start (ncounts, counts, chunk_size,
|
|
istart, iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_doacross_guided_start (unsigned ncounts, gomp_ull *counts,
|
|
gomp_ull chunk_size, gomp_ull *istart,
|
|
gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_doacross_guided_start (ncounts, counts, chunk_size,
|
|
istart, iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_static_next (gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_static_next (istart, iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_dynamic_next (gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_dynamic_next (istart, iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_guided_next (gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_guided_next (istart, iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_nonmonotonic_dynamic_next (gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_dynamic_next (istart, iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_nonmonotonic_guided_next (gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_guided_next (istart, iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_nonmonotonic_runtime_next (gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return GOMP_loop_ull_runtime_next (istart, iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_maybe_nonmonotonic_runtime_next (gomp_ull *istart,
|
|
gomp_ull *iend)
|
|
{
|
|
return GOMP_loop_ull_runtime_next (istart, iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_ordered_static_next (gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_ordered_static_next (istart, iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_ordered_dynamic_next (gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_ordered_dynamic_next (istart, iend);
|
|
}
|
|
|
|
bool
|
|
GOMP_loop_ull_ordered_guided_next (gomp_ull *istart, gomp_ull *iend)
|
|
{
|
|
return gomp_loop_ull_ordered_guided_next (istart, iend);
|
|
}
|
|
#endif
|