
This implements the OpenMP low-latency memory allocator for AMD GCN using the small per-team LDS memory (Local Data Store). Since addresses can now refer to LDS space, the "Global" address space is no-longer compatible. This patch therefore switches the backend to use entirely "Flat" addressing (which supports both memories). A future patch will re-enable "global" instructions for cases where it is known to be safe to do so. gcc/ChangeLog: * config/gcn/gcn-builtins.def (DISPATCH_PTR): New built-in. * config/gcn/gcn.cc (gcn_init_machine_status): Disable global addressing. (gcn_expand_builtin_1): Implement GCN_BUILTIN_DISPATCH_PTR. libgomp/ChangeLog: * config/gcn/libgomp-gcn.h (TEAM_ARENA_START): Move to here. (TEAM_ARENA_FREE): Likewise. (TEAM_ARENA_END): Likewise. (GCN_LOWLAT_HEAP): New. * config/gcn/team.c (LITTLEENDIAN_CPU): New, and import hsa.h. (__gcn_lowlat_init): New prototype. (gomp_gcn_enter_kernel): Initialize the low-latency heap. * libgomp.h (TEAM_ARENA_START): Move to libgomp.h. (TEAM_ARENA_FREE): Likewise. (TEAM_ARENA_END): Likewise. * plugin/plugin-gcn.c (lowlat_size): New variable. (print_kernel_dispatch): Label the group_segment_size purpose. (init_environment_variables): Read GOMP_GCN_LOWLAT_POOL. (create_kernel_dispatch): Pass low-latency head allocation to kernel. (run_kernel): Use shadow; don't assume values. * testsuite/libgomp.c/omp_alloc-traits.c: Enable for amdgcn. * config/gcn/allocator.c: New file. * libgomp.texi: Document low-latency implementation details.
232 lines
7.5 KiB
C
232 lines
7.5 KiB
C
/* Copyright (C) 2017-2023 Free Software Foundation, Inc.
|
|
Contributed by Mentor Embedded.
|
|
|
|
This file is part of the GNU Offloading and Multi Processing Library
|
|
(libgomp).
|
|
|
|
Libgomp is free software; you can redistribute it and/or modify it
|
|
under the terms of the GNU General Public License as published by
|
|
the Free Software Foundation; either version 3, or (at your option)
|
|
any later version.
|
|
|
|
Libgomp is distributed in the hope that it will be useful, but WITHOUT ANY
|
|
WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
|
|
FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
more details.
|
|
|
|
Under Section 7 of GPL version 3, you are granted additional
|
|
permissions described in the GCC Runtime Library Exception, version
|
|
3.1, as published by the Free Software Foundation.
|
|
|
|
You should have received a copy of the GNU General Public License and
|
|
a copy of the GCC Runtime Library Exception along with this program;
|
|
see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
/* This file handles maintenance of threads on AMD GCN. */
|
|
|
|
#include "libgomp.h"
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
|
|
#define LITTLEENDIAN_CPU
|
|
#include "hsa.h"
|
|
|
|
/* Defined in basic-allocator.c via config/amdgcn/allocator.c. */
|
|
void __gcn_lowlat_init (void *heap, size_t size);
|
|
|
|
static void gomp_thread_start (struct gomp_thread_pool *);
|
|
extern void build_indirect_map (void);
|
|
|
|
/* This externally visible function handles target region entry. It
|
|
sets up a per-team thread pool and transfers control by returning to
|
|
the kernel in the master thread or gomp_thread_start in other threads.
|
|
|
|
The name of this function is part of the interface with the compiler: for
|
|
each OpenMP kernel the compiler configures the stack, then calls here.
|
|
|
|
Likewise, gomp_gcn_exit_kernel is called during the kernel epilogue. */
|
|
|
|
void
|
|
gomp_gcn_enter_kernel (void)
|
|
{
|
|
int threadid = __builtin_gcn_dim_pos (1);
|
|
|
|
/* Initialize indirect function support. */
|
|
build_indirect_map ();
|
|
|
|
if (threadid == 0)
|
|
{
|
|
int numthreads = __builtin_gcn_dim_size (1);
|
|
int teamid = __builtin_gcn_dim_pos(0);
|
|
|
|
/* Set up the global state.
|
|
Every team will do this, but that should be harmless. */
|
|
gomp_global_icv.nthreads_var = 16;
|
|
gomp_global_icv.thread_limit_var = numthreads;
|
|
/* Starting additional threads is not supported. */
|
|
gomp_global_icv.dyn_var = true;
|
|
|
|
/* Initialize the team arena for optimized memory allocation.
|
|
The arena has been allocated on the host side, and the address
|
|
passed in via the kernargs. Each team takes a small slice of it. */
|
|
struct kernargs_abi *kernargs =
|
|
(struct kernargs_abi*) __builtin_gcn_kernarg_ptr ();
|
|
void *team_arena = ((void*)kernargs->arena_ptr
|
|
+ kernargs->arena_size_per_team * teamid);
|
|
void * __lds *arena_start = (void * __lds *)TEAM_ARENA_START;
|
|
void * __lds *arena_free = (void * __lds *)TEAM_ARENA_FREE;
|
|
void * __lds *arena_end = (void * __lds *)TEAM_ARENA_END;
|
|
*arena_start = team_arena;
|
|
*arena_free = team_arena;
|
|
*arena_end = team_arena + kernargs->arena_size_per_team;
|
|
|
|
/* Initialize the low-latency heap. The header is the size. */
|
|
void __lds *lowlat = (void __lds *)GCN_LOWLAT_HEAP;
|
|
hsa_kernel_dispatch_packet_t *queue_ptr = __builtin_gcn_dispatch_ptr ();
|
|
__gcn_lowlat_init ((void*)(uintptr_t)(void __flat*)lowlat,
|
|
queue_ptr->group_segment_size - GCN_LOWLAT_HEAP);
|
|
|
|
/* Allocate and initialize the team-local-storage data. */
|
|
struct gomp_thread *thrs = team_malloc_cleared (sizeof (*thrs)
|
|
* numthreads);
|
|
set_gcn_thrs (thrs);
|
|
|
|
/* Allocate and initialize a pool of threads in the team.
|
|
The threads are already running, of course, we just need to manage
|
|
the communication between them. */
|
|
struct gomp_thread_pool *pool = team_malloc (sizeof (*pool));
|
|
pool->threads = team_malloc (sizeof (void *) * numthreads);
|
|
for (int tid = 0; tid < numthreads; tid++)
|
|
pool->threads[tid] = &thrs[tid];
|
|
pool->threads_size = numthreads;
|
|
pool->threads_used = numthreads;
|
|
pool->threads_busy = 1;
|
|
pool->last_team = NULL;
|
|
gomp_simple_barrier_init (&pool->threads_dock, numthreads);
|
|
thrs->thread_pool = pool;
|
|
|
|
asm ("s_barrier" ::: "memory");
|
|
return; /* Return to kernel. */
|
|
}
|
|
else
|
|
{
|
|
asm ("s_barrier" ::: "memory");
|
|
gomp_thread_start (gcn_thrs ()[0].thread_pool);
|
|
/* gomp_thread_start does not return. */
|
|
}
|
|
}
|
|
|
|
void
|
|
gomp_gcn_exit_kernel (void)
|
|
{
|
|
gomp_free_thread (gcn_thrs ());
|
|
team_free (gcn_thrs ());
|
|
}
|
|
|
|
/* This function contains the idle loop in which a thread waits
|
|
to be called up to become part of a team. */
|
|
|
|
static void
|
|
gomp_thread_start (struct gomp_thread_pool *pool)
|
|
{
|
|
struct gomp_thread *thr = gomp_thread ();
|
|
|
|
gomp_sem_init (&thr->release, 0);
|
|
thr->thread_pool = pool;
|
|
|
|
/* The loop exits only when "fn" is assigned "gomp_free_pool_helper",
|
|
which contains "s_endpgm", or an infinite no-op loop is
|
|
suspected (this happens when the thread master crashes). */
|
|
int nul_limit = 99;
|
|
do
|
|
{
|
|
gomp_simple_barrier_wait (&pool->threads_dock);
|
|
if (!thr->fn)
|
|
{
|
|
if (nul_limit-- > 0)
|
|
continue;
|
|
else
|
|
{
|
|
const char msg[] = ("team master not responding;"
|
|
" slave thread aborting");
|
|
write (2, msg, sizeof (msg)-1);
|
|
abort();
|
|
}
|
|
}
|
|
thr->fn (thr->data);
|
|
thr->fn = NULL;
|
|
|
|
struct gomp_task *task = thr->task;
|
|
gomp_team_barrier_wait_final (&thr->ts.team->barrier);
|
|
gomp_finish_task (task);
|
|
}
|
|
while (1);
|
|
}
|
|
|
|
/* Launch a team. */
|
|
|
|
void
|
|
gomp_team_start (void (*fn) (void *), void *data, unsigned nthreads,
|
|
unsigned flags, struct gomp_team *team,
|
|
struct gomp_taskgroup *taskgroup)
|
|
{
|
|
struct gomp_thread *thr, *nthr;
|
|
struct gomp_task *task;
|
|
struct gomp_task_icv *icv;
|
|
struct gomp_thread_pool *pool;
|
|
unsigned long nthreads_var;
|
|
|
|
thr = gomp_thread ();
|
|
pool = thr->thread_pool;
|
|
task = thr->task;
|
|
icv = task ? &task->icv : &gomp_global_icv;
|
|
|
|
/* Always save the previous state, even if this isn't a nested team.
|
|
In particular, we should save any work share state from an outer
|
|
orphaned work share construct. */
|
|
team->prev_ts = thr->ts;
|
|
|
|
thr->ts.team = team;
|
|
thr->ts.team_id = 0;
|
|
++thr->ts.level;
|
|
if (nthreads > 1)
|
|
++thr->ts.active_level;
|
|
thr->ts.work_share = &team->work_shares[0];
|
|
thr->ts.last_work_share = NULL;
|
|
thr->ts.single_count = 0;
|
|
thr->ts.static_trip = 0;
|
|
thr->task = &team->implicit_task[0];
|
|
nthreads_var = icv->nthreads_var;
|
|
gomp_init_task (thr->task, task, icv);
|
|
team->implicit_task[0].icv.nthreads_var = nthreads_var;
|
|
team->implicit_task[0].taskgroup = taskgroup;
|
|
|
|
if (nthreads == 1)
|
|
return;
|
|
|
|
/* Release existing idle threads. */
|
|
for (unsigned i = 1; i < nthreads; ++i)
|
|
{
|
|
nthr = pool->threads[i];
|
|
nthr->ts.team = team;
|
|
nthr->ts.work_share = &team->work_shares[0];
|
|
nthr->ts.last_work_share = NULL;
|
|
nthr->ts.team_id = i;
|
|
nthr->ts.level = team->prev_ts.level + 1;
|
|
nthr->ts.active_level = thr->ts.active_level;
|
|
nthr->ts.single_count = 0;
|
|
nthr->ts.static_trip = 0;
|
|
nthr->task = &team->implicit_task[i];
|
|
gomp_init_task (nthr->task, task, icv);
|
|
team->implicit_task[i].icv.nthreads_var = nthreads_var;
|
|
team->implicit_task[i].taskgroup = taskgroup;
|
|
nthr->fn = fn;
|
|
nthr->data = data;
|
|
team->ordered_release[i] = &nthr->release;
|
|
}
|
|
|
|
gomp_simple_barrier_wait (&pool->threads_dock);
|
|
}
|
|
|
|
#include "../../team.c"
|