aboutsummaryrefslogtreecommitdiffstats
path: root/gcc-4.4.3/libgomp/team.c
diff options
context:
space:
mode:
Diffstat (limited to 'gcc-4.4.3/libgomp/team.c')
-rw-r--r--gcc-4.4.3/libgomp/team.c555
1 files changed, 0 insertions, 555 deletions
diff --git a/gcc-4.4.3/libgomp/team.c b/gcc-4.4.3/libgomp/team.c
deleted file mode 100644
index 44ffd5609..000000000
--- a/gcc-4.4.3/libgomp/team.c
+++ /dev/null
@@ -1,555 +0,0 @@
-/* Copyright (C) 2005, 2006, 2007, 2008, 2009 Free Software Foundation, Inc.
- Contributed by Richard Henderson <rth@redhat.com>.
-
- This file is part of the GNU OpenMP Library (libgomp).
-
- Libgomp is free software; you can redistribute it and/or modify it
- under the terms of the GNU General Public License as published by
- the Free Software Foundation; either version 3, or (at your option)
- any later version.
-
- Libgomp is distributed in the hope that it will be useful, but WITHOUT ANY
- WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
- FOR A PARTICULAR PURPOSE. See the GNU General Public License for
- more details.
-
- Under Section 7 of GPL version 3, you are granted additional
- permissions described in the GCC Runtime Library Exception, version
- 3.1, as published by the Free Software Foundation.
-
- You should have received a copy of the GNU General Public License and
- a copy of the GCC Runtime Library Exception along with this program;
- see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
- <http://www.gnu.org/licenses/>. */
-
-/* This file handles the maintainence of threads in response to team
- creation and termination. */
-
-#include "libgomp.h"
-#include <stdlib.h>
-#include <string.h>
-
-/* This attribute contains PTHREAD_CREATE_DETACHED. */
-pthread_attr_t gomp_thread_attr;
-
-/* This key is for the thread destructor. */
-pthread_key_t gomp_thread_destructor;
-
-
-/* This is the libgomp per-thread data structure. */
-#ifdef HAVE_TLS
-__thread struct gomp_thread gomp_tls_data;
-#else
-pthread_key_t gomp_tls_key;
-#endif
-
-
-/* This structure is used to communicate across pthread_create. */
-
-struct gomp_thread_start_data
-{
- void (*fn) (void *);
- void *fn_data;
- struct gomp_team_state ts;
- struct gomp_task *task;
- struct gomp_thread_pool *thread_pool;
- bool nested;
-};
-
-
-/* This function is a pthread_create entry point. This contains the idle
- loop in which a thread waits to be called up to become part of a team. */
-
-static void *
-gomp_thread_start (void *xdata)
-{
- struct gomp_thread_start_data *data = xdata;
- struct gomp_thread *thr;
- struct gomp_thread_pool *pool;
- void (*local_fn) (void *);
- void *local_data;
-
-#ifdef HAVE_TLS
- thr = &gomp_tls_data;
-#else
- struct gomp_thread local_thr;
- thr = &local_thr;
- pthread_setspecific (gomp_tls_key, thr);
-#endif
- gomp_sem_init (&thr->release, 0);
-
- /* Extract what we need from data. */
- local_fn = data->fn;
- local_data = data->fn_data;
- thr->thread_pool = data->thread_pool;
- thr->ts = data->ts;
- thr->task = data->task;
-
- thr->ts.team->ordered_release[thr->ts.team_id] = &thr->release;
-
- /* Make thread pool local. */
- pool = thr->thread_pool;
-
- if (data->nested)
- {
- struct gomp_team *team = thr->ts.team;
- struct gomp_task *task = thr->task;
-
- gomp_barrier_wait (&team->barrier);
-
- local_fn (local_data);
- gomp_team_barrier_wait (&team->barrier);
- gomp_finish_task (task);
- gomp_barrier_wait_last (&team->barrier);
- }
- else
- {
- pool->threads[thr->ts.team_id] = thr;
-
- gomp_barrier_wait (&pool->threads_dock);
- do
- {
- struct gomp_team *team = thr->ts.team;
- struct gomp_task *task = thr->task;
-
- local_fn (local_data);
- gomp_team_barrier_wait (&team->barrier);
- gomp_finish_task (task);
-
- gomp_barrier_wait (&pool->threads_dock);
-
- local_fn = thr->fn;
- local_data = thr->data;
- thr->fn = NULL;
- }
- while (local_fn);
- }
-
- gomp_sem_destroy (&thr->release);
- return NULL;
-}
-
-
-/* Create a new team data structure. */
-
-struct gomp_team *
-gomp_new_team (unsigned nthreads)
-{
- struct gomp_team *team;
- size_t size;
- int i;
-
- size = sizeof (*team) + nthreads * (sizeof (team->ordered_release[0])
- + sizeof (team->implicit_task[0]));
- team = gomp_malloc (size);
-
- team->work_share_chunk = 8;
-#ifdef HAVE_SYNC_BUILTINS
- team->single_count = 0;
-#else
- gomp_mutex_init (&team->work_share_list_free_lock);
-#endif
- gomp_init_work_share (&team->work_shares[0], false, nthreads);
- team->work_shares[0].next_alloc = NULL;
- team->work_share_list_free = NULL;
- team->work_share_list_alloc = &team->work_shares[1];
- for (i = 1; i < 7; i++)
- team->work_shares[i].next_free = &team->work_shares[i + 1];
- team->work_shares[i].next_free = NULL;
-
- team->nthreads = nthreads;
- gomp_barrier_init (&team->barrier, nthreads);
-
- gomp_sem_init (&team->master_release, 0);
- team->ordered_release = (void *) &team->implicit_task[nthreads];
- team->ordered_release[0] = &team->master_release;
-
- gomp_mutex_init (&team->task_lock);
- team->task_queue = NULL;
- team->task_count = 0;
- team->task_running_count = 0;
-
- return team;
-}
-
-
-/* Free a team data structure. */
-
-static void
-free_team (struct gomp_team *team)
-{
- gomp_barrier_destroy (&team->barrier);
- gomp_mutex_destroy (&team->task_lock);
- free (team);
-}
-
-/* Allocate and initialize a thread pool. */
-
-static struct gomp_thread_pool *gomp_new_thread_pool (void)
-{
- struct gomp_thread_pool *pool
- = gomp_malloc (sizeof(struct gomp_thread_pool));
- pool->threads = NULL;
- pool->threads_size = 0;
- pool->threads_used = 0;
- pool->last_team = NULL;
- return pool;
-}
-
-static void
-gomp_free_pool_helper (void *thread_pool)
-{
- struct gomp_thread_pool *pool
- = (struct gomp_thread_pool *) thread_pool;
- gomp_barrier_wait_last (&pool->threads_dock);
- gomp_sem_destroy (&gomp_thread ()->release);
- pthread_exit (NULL);
-}
-
-/* Free a thread pool and release its threads. */
-
-static void
-gomp_free_thread (void *arg __attribute__((unused)))
-{
- struct gomp_thread *thr = gomp_thread ();
- struct gomp_thread_pool *pool = thr->thread_pool;
- if (pool)
- {
- if (pool->threads_used > 0)
- {
- int i;
- for (i = 1; i < pool->threads_used; i++)
- {
- struct gomp_thread *nthr = pool->threads[i];
- nthr->fn = gomp_free_pool_helper;
- nthr->data = pool;
- }
- /* This barrier undocks threads docked on pool->threads_dock. */
- gomp_barrier_wait (&pool->threads_dock);
- /* And this waits till all threads have called gomp_barrier_wait_last
- in gomp_free_pool_helper. */
- gomp_barrier_wait (&pool->threads_dock);
- /* Now it is safe to destroy the barrier and free the pool. */
- gomp_barrier_destroy (&pool->threads_dock);
- }
- free (pool->threads);
- if (pool->last_team)
- free_team (pool->last_team);
- free (pool);
- thr->thread_pool = NULL;
- }
- if (thr->task != NULL)
- {
- struct gomp_task *task = thr->task;
- gomp_end_task ();
- free (task);
- }
-}
-
-/* Launch a team. */
-
-void
-gomp_team_start (void (*fn) (void *), void *data, unsigned nthreads,
- struct gomp_team *team)
-{
- struct gomp_thread_start_data *start_data;
- struct gomp_thread *thr, *nthr;
- struct gomp_task *task;
- struct gomp_task_icv *icv;
- bool nested;
- struct gomp_thread_pool *pool;
- unsigned i, n, old_threads_used = 0;
- pthread_attr_t thread_attr, *attr;
-
- thr = gomp_thread ();
- nested = thr->ts.team != NULL;
- if (__builtin_expect (thr->thread_pool == NULL, 0))
- {
- thr->thread_pool = gomp_new_thread_pool ();
- pthread_setspecific (gomp_thread_destructor, thr);
- }
- pool = thr->thread_pool;
- task = thr->task;
- icv = task ? &task->icv : &gomp_global_icv;
-
- /* Always save the previous state, even if this isn't a nested team.
- In particular, we should save any work share state from an outer
- orphaned work share construct. */
- team->prev_ts = thr->ts;
-
- thr->ts.team = team;
- thr->ts.team_id = 0;
- ++thr->ts.level;
- if (nthreads > 1)
- ++thr->ts.active_level;
- thr->ts.work_share = &team->work_shares[0];
- thr->ts.last_work_share = NULL;
-#ifdef HAVE_SYNC_BUILTINS
- thr->ts.single_count = 0;
-#endif
- thr->ts.static_trip = 0;
- thr->task = &team->implicit_task[0];
- gomp_init_task (thr->task, task, icv);
-
- if (nthreads == 1)
- return;
-
- i = 1;
-
- /* We only allow the reuse of idle threads for non-nested PARALLEL
- regions. This appears to be implied by the semantics of
- threadprivate variables, but perhaps that's reading too much into
- things. Certainly it does prevent any locking problems, since
- only the initial program thread will modify gomp_threads. */
- if (!nested)
- {
- old_threads_used = pool->threads_used;
-
- if (nthreads <= old_threads_used)
- n = nthreads;
- else if (old_threads_used == 0)
- {
- n = 0;
- gomp_barrier_init (&pool->threads_dock, nthreads);
- }
- else
- {
- n = old_threads_used;
-
- /* Increase the barrier threshold to make sure all new
- threads arrive before the team is released. */
- gomp_barrier_reinit (&pool->threads_dock, nthreads);
- }
-
- /* Not true yet, but soon will be. We're going to release all
- threads from the dock, and those that aren't part of the
- team will exit. */
- pool->threads_used = nthreads;
-
- /* Release existing idle threads. */
- for (; i < n; ++i)
- {
- nthr = pool->threads[i];
- nthr->ts.team = team;
- nthr->ts.work_share = &team->work_shares[0];
- nthr->ts.last_work_share = NULL;
- nthr->ts.team_id = i;
- nthr->ts.level = team->prev_ts.level + 1;
- nthr->ts.active_level = thr->ts.active_level;
-#ifdef HAVE_SYNC_BUILTINS
- nthr->ts.single_count = 0;
-#endif
- nthr->ts.static_trip = 0;
- nthr->task = &team->implicit_task[i];
- gomp_init_task (nthr->task, task, icv);
- nthr->fn = fn;
- nthr->data = data;
- team->ordered_release[i] = &nthr->release;
- }
-
- if (i == nthreads)
- goto do_release;
-
- /* If necessary, expand the size of the gomp_threads array. It is
- expected that changes in the number of threads are rare, thus we
- make no effort to expand gomp_threads_size geometrically. */
- if (nthreads >= pool->threads_size)
- {
- pool->threads_size = nthreads + 1;
- pool->threads
- = gomp_realloc (pool->threads,
- pool->threads_size
- * sizeof (struct gomp_thread_data *));
- }
- }
-
- if (__builtin_expect (nthreads > old_threads_used, 0))
- {
- long diff = (long) nthreads - (long) old_threads_used;
-
- if (old_threads_used == 0)
- --diff;
-
-#ifdef HAVE_SYNC_BUILTINS
- __sync_fetch_and_add (&gomp_managed_threads, diff);
-#else
- gomp_mutex_lock (&gomp_remaining_threads_lock);
- gomp_managed_threads += diff;
- gomp_mutex_unlock (&gomp_remaining_threads_lock);
-#endif
- }
-
- attr = &gomp_thread_attr;
- if (__builtin_expect (gomp_cpu_affinity != NULL, 0))
- {
- size_t stacksize;
- pthread_attr_init (&thread_attr);
- pthread_attr_setdetachstate (&thread_attr, PTHREAD_CREATE_DETACHED);
- if (! pthread_attr_getstacksize (&gomp_thread_attr, &stacksize))
- pthread_attr_setstacksize (&thread_attr, stacksize);
- attr = &thread_attr;
- }
-
- start_data = gomp_alloca (sizeof (struct gomp_thread_start_data)
- * (nthreads-i));
-
- /* Launch new threads. */
- for (; i < nthreads; ++i, ++start_data)
- {
- pthread_t pt;
- int err;
-
- start_data->fn = fn;
- start_data->fn_data = data;
- start_data->ts.team = team;
- start_data->ts.work_share = &team->work_shares[0];
- start_data->ts.last_work_share = NULL;
- start_data->ts.team_id = i;
- start_data->ts.level = team->prev_ts.level + 1;
- start_data->ts.active_level = thr->ts.active_level;
-#ifdef HAVE_SYNC_BUILTINS
- start_data->ts.single_count = 0;
-#endif
- start_data->ts.static_trip = 0;
- start_data->task = &team->implicit_task[i];
- gomp_init_task (start_data->task, task, icv);
- start_data->thread_pool = pool;
- start_data->nested = nested;
-
- if (gomp_cpu_affinity != NULL)
- gomp_init_thread_affinity (attr);
-
- err = pthread_create (&pt, attr, gomp_thread_start, start_data);
- if (err != 0)
- gomp_fatal ("Thread creation failed: %s", strerror (err));
- }
-
- if (__builtin_expect (gomp_cpu_affinity != NULL, 0))
- pthread_attr_destroy (&thread_attr);
-
- do_release:
- gomp_barrier_wait (nested ? &team->barrier : &pool->threads_dock);
-
- /* Decrease the barrier threshold to match the number of threads
- that should arrive back at the end of this team. The extra
- threads should be exiting. Note that we arrange for this test
- to never be true for nested teams. */
- if (__builtin_expect (nthreads < old_threads_used, 0))
- {
- long diff = (long) nthreads - (long) old_threads_used;
-
- gomp_barrier_reinit (&pool->threads_dock, nthreads);
-
-#ifdef HAVE_SYNC_BUILTINS
- __sync_fetch_and_add (&gomp_managed_threads, diff);
-#else
- gomp_mutex_lock (&gomp_remaining_threads_lock);
- gomp_managed_threads += diff;
- gomp_mutex_unlock (&gomp_remaining_threads_lock);
-#endif
- }
-}
-
-
-/* Terminate the current team. This is only to be called by the master
- thread. We assume that we must wait for the other threads. */
-
-void
-gomp_team_end (void)
-{
- struct gomp_thread *thr = gomp_thread ();
- struct gomp_team *team = thr->ts.team;
-
- /* This barrier handles all pending explicit threads. */
- gomp_team_barrier_wait (&team->barrier);
- gomp_fini_work_share (thr->ts.work_share);
-
- gomp_end_task ();
- thr->ts = team->prev_ts;
-
- if (__builtin_expect (thr->ts.team != NULL, 0))
- {
-#ifdef HAVE_SYNC_BUILTINS
- __sync_fetch_and_add (&gomp_managed_threads, 1L - team->nthreads);
-#else
- gomp_mutex_lock (&gomp_remaining_threads_lock);
- gomp_managed_threads -= team->nthreads - 1L;
- gomp_mutex_unlock (&gomp_remaining_threads_lock);
-#endif
- /* This barrier has gomp_barrier_wait_last counterparts
- and ensures the team can be safely destroyed. */
- gomp_barrier_wait (&team->barrier);
- }
-
- if (__builtin_expect (team->work_shares[0].next_alloc != NULL, 0))
- {
- struct gomp_work_share *ws = team->work_shares[0].next_alloc;
- do
- {
- struct gomp_work_share *next_ws = ws->next_alloc;
- free (ws);
- ws = next_ws;
- }
- while (ws != NULL);
- }
- gomp_sem_destroy (&team->master_release);
-#ifndef HAVE_SYNC_BUILTINS
- gomp_mutex_destroy (&team->work_share_list_free_lock);
-#endif
-
- if (__builtin_expect (thr->ts.team != NULL, 0)
- || __builtin_expect (team->nthreads == 1, 0))
- free_team (team);
- else
- {
- struct gomp_thread_pool *pool = thr->thread_pool;
- if (pool->last_team)
- free_team (pool->last_team);
- pool->last_team = team;
- }
-}
-
-
-/* Constructors for this file. */
-
-static void __attribute__((constructor))
-initialize_team (void)
-{
- struct gomp_thread *thr;
-
-#ifndef HAVE_TLS
- static struct gomp_thread initial_thread_tls_data;
-
- pthread_key_create (&gomp_tls_key, NULL);
- pthread_setspecific (gomp_tls_key, &initial_thread_tls_data);
-#endif
-
- if (pthread_key_create (&gomp_thread_destructor, gomp_free_thread) != 0)
- gomp_fatal ("could not create thread pool destructor.");
-
-#ifdef HAVE_TLS
- thr = &gomp_tls_data;
-#else
- thr = &initial_thread_tls_data;
-#endif
- gomp_sem_init (&thr->release, 0);
-}
-
-static void __attribute__((destructor))
-team_destructor (void)
-{
- /* Without this dlclose on libgomp could lead to subsequent
- crashes. */
- pthread_key_delete (gomp_thread_destructor);
-}
-
-struct gomp_task_icv *
-gomp_new_icv (void)
-{
- struct gomp_thread *thr = gomp_thread ();
- struct gomp_task *task = gomp_malloc (sizeof (struct gomp_task));
- gomp_init_task (task, NULL, &gomp_global_icv);
- thr->task = task;
- pthread_setspecific (gomp_thread_destructor, thr);
- return &task->icv;
-}