[1] | 1 | /* Copyright (C) 2005, 2006, 2007, 2008, 2009 Free Software Foundation, Inc. |
---|
| 2 | Contributed by Richard Henderson <rth@redhat.com>. |
---|
| 3 | |
---|
| 4 | This file is part of the GNU OpenMP Library (libgomp). |
---|
| 5 | |
---|
| 6 | Libgomp is free software; you can redistribute it and/or modify it |
---|
| 7 | under the terms of the GNU General Public License as published by |
---|
| 8 | the Free Software Foundation; either version 3, or (at your option) |
---|
| 9 | any later version. |
---|
| 10 | |
---|
| 11 | Libgomp is distributed in the hope that it will be useful, but WITHOUT ANY |
---|
| 12 | WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS |
---|
| 13 | FOR A PARTICULAR PURPOSE. See the GNU General Public License for |
---|
| 14 | more details. |
---|
| 15 | |
---|
| 16 | Under Section 7 of GPL version 3, you are granted additional |
---|
| 17 | permissions described in the GCC Runtime Library Exception, version |
---|
| 18 | 3.1, as published by the Free Software Foundation. |
---|
| 19 | |
---|
| 20 | You should have received a copy of the GNU General Public License and |
---|
| 21 | a copy of the GCC Runtime Library Exception along with this program; |
---|
| 22 | see the files COPYING3 and COPYING.RUNTIME respectively. If not, see |
---|
| 23 | <http://www.gnu.org/licenses/>. */ |
---|
| 24 | |
---|
| 25 | /* This file handles the maintainence of threads in response to team |
---|
| 26 | creation and termination. */ |
---|
| 27 | |
---|
| 28 | #include <gomp/libgomp.h> |
---|
| 29 | #include <stdlib.h> |
---|
| 30 | #include <string.h> |
---|
| 31 | |
---|
| 32 | /* This attribute contains PTHREAD_CREATE_DETACHED. */ |
---|
| 33 | pthread_attr_t gomp_thread_attr; |
---|
| 34 | |
---|
| 35 | /* This key is for the thread destructor. */ |
---|
| 36 | pthread_key_t gomp_thread_destructor; |
---|
| 37 | |
---|
| 38 | |
---|
| 39 | /* This is the libgomp per-thread data structure. */ |
---|
| 40 | #ifdef HAVE_TLS |
---|
| 41 | __thread struct gomp_thread gomp_tls_data; |
---|
| 42 | #else |
---|
| 43 | pthread_key_t gomp_tls_key; |
---|
| 44 | #endif |
---|
| 45 | |
---|
| 46 | |
---|
| 47 | /* This structure is used to communicate across pthread_create. */ |
---|
| 48 | |
---|
| 49 | struct gomp_thread_start_data |
---|
| 50 | { |
---|
| 51 | void (*fn) (void *); |
---|
| 52 | void *fn_data; |
---|
| 53 | struct gomp_team_state ts; |
---|
| 54 | struct gomp_task *task; |
---|
| 55 | struct gomp_thread_pool *thread_pool; |
---|
| 56 | bool nested; |
---|
| 57 | }; |
---|
| 58 | |
---|
| 59 | |
---|
| 60 | /* This function is a pthread_create entry point. This contains the idle |
---|
| 61 | loop in which a thread waits to be called up to become part of a team. */ |
---|
| 62 | |
---|
| 63 | static void * |
---|
| 64 | gomp_thread_start (void *xdata) |
---|
| 65 | { |
---|
| 66 | struct gomp_thread_start_data *data = xdata; |
---|
| 67 | struct gomp_thread *thr; |
---|
| 68 | struct gomp_thread_pool *pool; |
---|
| 69 | void (*local_fn) (void *); |
---|
| 70 | void *local_data; |
---|
| 71 | |
---|
| 72 | #ifdef HAVE_TLS |
---|
| 73 | thr = &gomp_tls_data; |
---|
| 74 | #else |
---|
| 75 | struct gomp_thread local_thr; |
---|
| 76 | thr = &local_thr; |
---|
| 77 | pthread_setspecific (gomp_tls_key, thr); |
---|
| 78 | #endif |
---|
| 79 | |
---|
| 80 | gomp_sem_init (&thr->release, 0); |
---|
| 81 | |
---|
| 82 | /* Extract what we need from data. */ |
---|
| 83 | local_fn = data->fn; |
---|
| 84 | local_data = data->fn_data; |
---|
| 85 | thr->thread_pool = data->thread_pool; |
---|
| 86 | thr->ts = data->ts; |
---|
| 87 | thr->task = data->task; |
---|
| 88 | |
---|
| 89 | thr->ts.team->ordered_release[thr->ts.team_id] = &thr->release; |
---|
| 90 | |
---|
| 91 | /* Make thread pool local. */ |
---|
| 92 | pool = thr->thread_pool; |
---|
| 93 | |
---|
| 94 | if (data->nested) |
---|
| 95 | { |
---|
| 96 | struct gomp_team *team = thr->ts.team; |
---|
| 97 | struct gomp_task *task = thr->task; |
---|
| 98 | |
---|
| 99 | gomp_barrier_wait (&team->barrier); |
---|
| 100 | |
---|
| 101 | local_fn (local_data); |
---|
| 102 | gomp_team_barrier_wait (&team->barrier); |
---|
| 103 | gomp_finish_task (task); |
---|
| 104 | gomp_barrier_wait_last (&team->barrier); |
---|
| 105 | } |
---|
| 106 | else |
---|
| 107 | { |
---|
| 108 | pool->threads[thr->ts.team_id] = thr; |
---|
| 109 | |
---|
| 110 | gomp_barrier_wait (&pool->threads_dock); |
---|
| 111 | do |
---|
| 112 | { |
---|
| 113 | struct gomp_team *team = thr->ts.team; |
---|
| 114 | struct gomp_task *task = thr->task; |
---|
| 115 | |
---|
| 116 | local_fn (local_data); |
---|
| 117 | gomp_team_barrier_wait (&team->barrier); |
---|
| 118 | gomp_finish_task (task); |
---|
| 119 | |
---|
| 120 | gomp_barrier_wait (&pool->threads_dock); |
---|
| 121 | |
---|
| 122 | local_fn = thr->fn; |
---|
| 123 | local_data = thr->data; |
---|
| 124 | thr->fn = NULL; |
---|
| 125 | } |
---|
| 126 | while (local_fn); |
---|
| 127 | } |
---|
| 128 | |
---|
| 129 | gomp_sem_destroy (&thr->release); |
---|
| 130 | return NULL; |
---|
| 131 | } |
---|
| 132 | |
---|
| 133 | |
---|
| 134 | /* Create a new team data structure. */ |
---|
| 135 | |
---|
| 136 | struct gomp_team * |
---|
| 137 | gomp_new_team (unsigned nthreads) |
---|
| 138 | { |
---|
| 139 | struct gomp_team *team; |
---|
| 140 | size_t size; |
---|
| 141 | int i; |
---|
| 142 | |
---|
| 143 | size = sizeof (*team) + nthreads * (sizeof (team->ordered_release[0]) |
---|
| 144 | + sizeof (team->implicit_task[0])); |
---|
| 145 | team = gomp_malloc (size); |
---|
| 146 | |
---|
| 147 | team->work_share_chunk = 8; |
---|
| 148 | #ifdef HAVE_SYNC_BUILTINS |
---|
| 149 | team->single_count = 0; |
---|
| 150 | #else |
---|
| 151 | gomp_mutex_init (&team->work_share_list_free_lock); |
---|
| 152 | #endif |
---|
| 153 | gomp_init_work_share (&team->work_shares[0], false, nthreads); |
---|
| 154 | team->work_shares[0].next_alloc = NULL; |
---|
| 155 | team->work_share_list_free = NULL; |
---|
| 156 | team->work_share_list_alloc = &team->work_shares[1]; |
---|
| 157 | for (i = 1; i < 7; i++) |
---|
| 158 | team->work_shares[i].next_free = &team->work_shares[i + 1]; |
---|
| 159 | team->work_shares[i].next_free = NULL; |
---|
| 160 | |
---|
| 161 | team->nthreads = nthreads; |
---|
| 162 | gomp_barrier_init (&team->barrier, nthreads); |
---|
| 163 | |
---|
| 164 | gomp_sem_init (&team->master_release, 0); |
---|
| 165 | team->ordered_release = (void *) &team->implicit_task[nthreads]; |
---|
| 166 | team->ordered_release[0] = &team->master_release; |
---|
| 167 | |
---|
| 168 | gomp_mutex_init (&team->task_lock); |
---|
| 169 | team->task_queue = NULL; |
---|
| 170 | team->task_count = 0; |
---|
| 171 | team->task_running_count = 0; |
---|
| 172 | |
---|
| 173 | return team; |
---|
| 174 | } |
---|
| 175 | |
---|
| 176 | |
---|
| 177 | /* Free a team data structure. */ |
---|
| 178 | |
---|
| 179 | static void |
---|
| 180 | free_team (struct gomp_team *team) |
---|
| 181 | { |
---|
| 182 | gomp_barrier_destroy (&team->barrier); |
---|
| 183 | gomp_mutex_destroy (&team->task_lock); |
---|
| 184 | free (team); |
---|
| 185 | } |
---|
| 186 | |
---|
| 187 | /* Allocate and initialize a thread pool. */ |
---|
| 188 | |
---|
| 189 | static struct gomp_thread_pool *gomp_new_thread_pool (void) |
---|
| 190 | { |
---|
| 191 | struct gomp_thread_pool *pool |
---|
| 192 | = gomp_malloc (sizeof(struct gomp_thread_pool)); |
---|
| 193 | pool->threads = NULL; |
---|
| 194 | pool->threads_size = 0; |
---|
| 195 | pool->threads_used = 0; |
---|
| 196 | pool->last_team = NULL; |
---|
| 197 | return pool; |
---|
| 198 | } |
---|
| 199 | |
---|
| 200 | static void |
---|
| 201 | gomp_free_pool_helper (void *thread_pool) |
---|
| 202 | { |
---|
| 203 | struct gomp_thread_pool *pool |
---|
| 204 | = (struct gomp_thread_pool *) thread_pool; |
---|
| 205 | |
---|
| 206 | gomp_barrier_wait_last (&pool->threads_dock); |
---|
| 207 | gomp_sem_destroy (&gomp_thread ()->release); |
---|
| 208 | pthread_exit (NULL); |
---|
| 209 | } |
---|
| 210 | |
---|
| 211 | /* Free a thread pool and release its threads. */ |
---|
| 212 | |
---|
| 213 | static void |
---|
| 214 | gomp_free_thread (void *arg __attribute__((unused))) |
---|
| 215 | { |
---|
| 216 | struct gomp_thread *thr; |
---|
| 217 | struct gomp_thread_pool *pool; |
---|
| 218 | |
---|
| 219 | #ifndef HAVE_TLS |
---|
| 220 | pthread_setspecific(gomp_tls_key, arg); |
---|
| 221 | #endif |
---|
| 222 | |
---|
| 223 | thr = gomp_thread (); |
---|
| 224 | pool = thr->thread_pool; |
---|
| 225 | |
---|
| 226 | if (pool) |
---|
| 227 | { |
---|
| 228 | if (pool->threads_used > 0) |
---|
| 229 | { |
---|
| 230 | int i; |
---|
| 231 | for (i = 1; i < (int)pool->threads_used; i++) |
---|
| 232 | { |
---|
| 233 | struct gomp_thread *nthr = pool->threads[i]; |
---|
| 234 | nthr->fn = gomp_free_pool_helper; |
---|
| 235 | nthr->data = pool; |
---|
| 236 | } |
---|
| 237 | /* This barrier undocks threads docked on pool->threads_dock. */ |
---|
| 238 | gomp_barrier_wait (&pool->threads_dock); |
---|
| 239 | /* And this waits till all threads have called gomp_barrier_wait_last |
---|
| 240 | in gomp_free_pool_helper. */ |
---|
| 241 | gomp_barrier_wait (&pool->threads_dock); |
---|
| 242 | /* Now it is safe to destroy the barrier and free the pool. */ |
---|
| 243 | gomp_barrier_destroy (&pool->threads_dock); |
---|
| 244 | } |
---|
| 245 | free (pool->threads); |
---|
| 246 | if (pool->last_team) |
---|
| 247 | free_team (pool->last_team); |
---|
| 248 | free (pool); |
---|
| 249 | thr->thread_pool = NULL; |
---|
| 250 | } |
---|
| 251 | if (thr->task != NULL) |
---|
| 252 | { |
---|
| 253 | struct gomp_task *task = thr->task; |
---|
| 254 | gomp_end_task (); |
---|
| 255 | free (task); |
---|
| 256 | } |
---|
| 257 | |
---|
| 258 | #ifndef HAVE_TLS |
---|
| 259 | pthread_key_delete(gomp_tls_key); |
---|
| 260 | #endif |
---|
| 261 | } |
---|
| 262 | |
---|
| 263 | /* Launch a team. */ |
---|
| 264 | |
---|
| 265 | void |
---|
| 266 | gomp_team_start (void (*fn) (void *), void *data, unsigned nthreads, |
---|
| 267 | struct gomp_team *team) |
---|
| 268 | { |
---|
| 269 | struct gomp_thread_start_data *start_data; |
---|
| 270 | struct gomp_thread *thr, *nthr; |
---|
| 271 | struct gomp_task *task; |
---|
| 272 | struct gomp_task_icv *icv; |
---|
| 273 | bool nested; |
---|
| 274 | struct gomp_thread_pool *pool; |
---|
| 275 | unsigned i, n, old_threads_used = 0; |
---|
| 276 | pthread_attr_t thread_attr, *attr; |
---|
| 277 | |
---|
| 278 | thr = gomp_thread (); |
---|
| 279 | nested = thr->ts.team != NULL; |
---|
| 280 | |
---|
| 281 | if (__builtin_expect (thr->thread_pool == NULL, 0)) |
---|
| 282 | { |
---|
| 283 | thr->thread_pool = gomp_new_thread_pool (); |
---|
| 284 | pthread_setspecific (gomp_thread_destructor, thr); |
---|
| 285 | } |
---|
| 286 | pool = thr->thread_pool; |
---|
| 287 | task = thr->task; |
---|
| 288 | icv = task ? &task->icv : &gomp_global_icv; |
---|
| 289 | |
---|
| 290 | /* Always save the previous state, even if this isn't a nested team. |
---|
| 291 | In particular, we should save any work share state from an outer |
---|
| 292 | orphaned work share construct. */ |
---|
| 293 | team->prev_ts = thr->ts; |
---|
| 294 | |
---|
| 295 | thr->ts.team = team; |
---|
| 296 | thr->ts.team_id = 0; |
---|
| 297 | ++thr->ts.level; |
---|
| 298 | if (nthreads > 1) |
---|
| 299 | ++thr->ts.active_level; |
---|
| 300 | thr->ts.work_share = &team->work_shares[0]; |
---|
| 301 | thr->ts.last_work_share = NULL; |
---|
| 302 | #ifdef HAVE_SYNC_BUILTINS |
---|
| 303 | thr->ts.single_count = 0; |
---|
| 304 | #endif |
---|
| 305 | thr->ts.static_trip = 0; |
---|
| 306 | thr->task = &team->implicit_task[0]; |
---|
| 307 | gomp_init_task (thr->task, task, icv); |
---|
| 308 | |
---|
| 309 | if (nthreads == 1) |
---|
| 310 | return; |
---|
| 311 | |
---|
| 312 | i = 1; |
---|
| 313 | |
---|
| 314 | /* We only allow the reuse of idle threads for non-nested PARALLEL |
---|
| 315 | regions. This appears to be implied by the semantics of |
---|
| 316 | threadprivate variables, but perhaps that's reading too much into |
---|
| 317 | things. Certainly it does prevent any locking problems, since |
---|
| 318 | only the initial program thread will modify gomp_threads. */ |
---|
| 319 | if (!nested) |
---|
| 320 | { |
---|
| 321 | old_threads_used = pool->threads_used; |
---|
| 322 | |
---|
| 323 | if (nthreads <= old_threads_used) |
---|
| 324 | n = nthreads; |
---|
| 325 | else if (old_threads_used == 0) |
---|
| 326 | { |
---|
| 327 | n = 0; |
---|
| 328 | gomp_barrier_init (&pool->threads_dock, nthreads); |
---|
| 329 | } |
---|
| 330 | else |
---|
| 331 | { |
---|
| 332 | n = old_threads_used; |
---|
| 333 | |
---|
| 334 | /* Increase the barrier threshold to make sure all new |
---|
| 335 | threads arrive before the team is released. */ |
---|
| 336 | gomp_barrier_reinit (&pool->threads_dock, nthreads); |
---|
| 337 | } |
---|
| 338 | |
---|
| 339 | /* Not true yet, but soon will be. We're going to release all |
---|
| 340 | threads from the dock, and those that aren't part of the |
---|
| 341 | team will exit. */ |
---|
| 342 | pool->threads_used = nthreads; |
---|
| 343 | |
---|
| 344 | /* Release existing idle threads. */ |
---|
| 345 | for (; i < n; ++i) |
---|
| 346 | { |
---|
| 347 | nthr = pool->threads[i]; |
---|
| 348 | nthr->ts.team = team; |
---|
| 349 | nthr->ts.work_share = &team->work_shares[0]; |
---|
| 350 | nthr->ts.last_work_share = NULL; |
---|
| 351 | nthr->ts.team_id = i; |
---|
| 352 | nthr->ts.level = team->prev_ts.level + 1; |
---|
| 353 | nthr->ts.active_level = thr->ts.active_level; |
---|
| 354 | #ifdef HAVE_SYNC_BUILTINS |
---|
| 355 | nthr->ts.single_count = 0; |
---|
| 356 | #endif |
---|
| 357 | nthr->ts.static_trip = 0; |
---|
| 358 | nthr->task = &team->implicit_task[i]; |
---|
| 359 | gomp_init_task (nthr->task, task, icv); |
---|
| 360 | nthr->fn = fn; |
---|
| 361 | nthr->data = data; |
---|
| 362 | team->ordered_release[i] = &nthr->release; |
---|
| 363 | } |
---|
| 364 | |
---|
| 365 | if (i == nthreads) |
---|
| 366 | goto do_release; |
---|
| 367 | |
---|
| 368 | /* If necessary, expand the size of the gomp_threads array. It is |
---|
| 369 | expected that changes in the number of threads are rare, thus we |
---|
| 370 | make no effort to expand gomp_threads_size geometrically. */ |
---|
| 371 | if (nthreads >= pool->threads_size) |
---|
| 372 | { |
---|
| 373 | pool->threads_size = nthreads + 1; |
---|
| 374 | pool->threads |
---|
| 375 | = gomp_realloc (pool->threads, |
---|
| 376 | pool->threads_size |
---|
| 377 | * sizeof (struct gomp_thread_data *)); |
---|
| 378 | } |
---|
| 379 | } |
---|
| 380 | |
---|
| 381 | if (__builtin_expect (nthreads > old_threads_used, 0)) |
---|
| 382 | { |
---|
| 383 | long diff = (long) nthreads - (long) old_threads_used; |
---|
| 384 | |
---|
| 385 | if (old_threads_used == 0) |
---|
| 386 | --diff; |
---|
| 387 | |
---|
| 388 | #ifdef HAVE_SYNC_BUILTINS |
---|
| 389 | __sync_fetch_and_add (&gomp_managed_threads, diff); |
---|
| 390 | #else |
---|
| 391 | gomp_mutex_lock (&gomp_remaining_threads_lock); |
---|
| 392 | gomp_managed_threads += diff; |
---|
| 393 | gomp_mutex_unlock (&gomp_remaining_threads_lock); |
---|
| 394 | #endif |
---|
| 395 | } |
---|
| 396 | |
---|
| 397 | attr = &gomp_thread_attr; |
---|
| 398 | if (__builtin_expect (gomp_cpu_affinity != NULL, 0)) |
---|
| 399 | { |
---|
| 400 | size_t stacksize; |
---|
| 401 | pthread_attr_init (&thread_attr); |
---|
| 402 | pthread_attr_setdetachstate (&thread_attr, PTHREAD_CREATE_DETACHED); |
---|
| 403 | if (! pthread_attr_getstacksize (&gomp_thread_attr, &stacksize)) |
---|
| 404 | pthread_attr_setstacksize (&thread_attr, stacksize); |
---|
| 405 | attr = &thread_attr; |
---|
| 406 | } |
---|
| 407 | |
---|
| 408 | start_data = gomp_alloca (sizeof (struct gomp_thread_start_data) |
---|
| 409 | * (nthreads-i)); |
---|
| 410 | |
---|
| 411 | /* Launch new threads. */ |
---|
| 412 | int n_master; |
---|
| 413 | int error = pthread_attr_getcpuid_np(&n_master); |
---|
| 414 | if (error != 0) { |
---|
| 415 | fprintf(stderr, "ERROR: %s: failed to get cpu id\n", __FUNCTION__); |
---|
| 416 | n_master = -1; |
---|
| 417 | } |
---|
| 418 | |
---|
| 419 | for (; i < nthreads; ++i, ++start_data ) |
---|
| 420 | { |
---|
| 421 | pthread_t pt; |
---|
| 422 | int err; |
---|
| 423 | |
---|
| 424 | start_data->fn = fn; |
---|
| 425 | start_data->fn_data = data; |
---|
| 426 | start_data->ts.team = team; |
---|
| 427 | start_data->ts.work_share = &team->work_shares[0]; |
---|
| 428 | start_data->ts.last_work_share = NULL; |
---|
| 429 | start_data->ts.team_id = i; |
---|
| 430 | start_data->ts.level = team->prev_ts.level + 1; |
---|
| 431 | start_data->ts.active_level = thr->ts.active_level; |
---|
| 432 | #ifdef HAVE_SYNC_BUILTINS |
---|
| 433 | start_data->ts.single_count = 0; |
---|
| 434 | #endif |
---|
| 435 | start_data->ts.static_trip = 0; |
---|
| 436 | start_data->task = &team->implicit_task[i]; |
---|
| 437 | gomp_init_task (start_data->task, task, icv); |
---|
| 438 | start_data->thread_pool = pool; |
---|
| 439 | start_data->nested = nested; |
---|
| 440 | |
---|
| 441 | if (gomp_cpu_affinity != NULL) |
---|
| 442 | gomp_init_thread_affinity (attr); |
---|
| 443 | |
---|
| 444 | if (i==(unsigned)n_master) |
---|
| 445 | pthread_attr_setcpuid_np(attr, 0, NULL); |
---|
| 446 | else |
---|
| 447 | pthread_attr_setcpuid_np(attr, i, NULL); |
---|
| 448 | |
---|
| 449 | err = pthread_create (&pt, attr, gomp_thread_start, start_data); |
---|
| 450 | if (err != 0) |
---|
| 451 | gomp_fatal ("Thread creation failed: %s", strerror (err)); |
---|
| 452 | } |
---|
| 453 | |
---|
| 454 | if (__builtin_expect (gomp_cpu_affinity != NULL, 0)) |
---|
| 455 | pthread_attr_destroy (&thread_attr); |
---|
| 456 | |
---|
| 457 | do_release: |
---|
| 458 | gomp_barrier_wait (nested ? &team->barrier : &pool->threads_dock); |
---|
| 459 | |
---|
| 460 | /* Decrease the barrier threshold to match the number of threads |
---|
| 461 | that should arrive back at the end of this team. The extra |
---|
| 462 | threads should be exiting. Note that we arrange for this test |
---|
| 463 | to never be true for nested teams. */ |
---|
| 464 | if (__builtin_expect (nthreads < old_threads_used, 0)) |
---|
| 465 | { |
---|
| 466 | long diff = (long) nthreads - (long) old_threads_used; |
---|
| 467 | |
---|
| 468 | gomp_barrier_reinit (&pool->threads_dock, nthreads); |
---|
| 469 | |
---|
| 470 | #ifdef HAVE_SYNC_BUILTINS |
---|
| 471 | __sync_fetch_and_add (&gomp_managed_threads, diff); |
---|
| 472 | #else |
---|
| 473 | gomp_mutex_lock (&gomp_remaining_threads_lock); |
---|
| 474 | gomp_managed_threads += diff; |
---|
| 475 | gomp_mutex_unlock (&gomp_remaining_threads_lock); |
---|
| 476 | #endif |
---|
| 477 | } |
---|
| 478 | } |
---|
| 479 | |
---|
| 480 | |
---|
| 481 | /* Terminate the current team. This is only to be called by the master |
---|
| 482 | thread. We assume that we must wait for the other threads. */ |
---|
| 483 | |
---|
| 484 | void |
---|
| 485 | gomp_team_end (void) |
---|
| 486 | { |
---|
| 487 | struct gomp_thread *thr = gomp_thread (); |
---|
| 488 | struct gomp_team *team = thr->ts.team; |
---|
| 489 | |
---|
| 490 | /* This barrier handles all pending explicit threads. */ |
---|
| 491 | gomp_team_barrier_wait (&team->barrier); |
---|
| 492 | gomp_fini_work_share (thr->ts.work_share); |
---|
| 493 | |
---|
| 494 | gomp_end_task (); |
---|
| 495 | thr->ts = team->prev_ts; |
---|
| 496 | |
---|
| 497 | if (__builtin_expect (thr->ts.team != NULL, 0)) |
---|
| 498 | { |
---|
| 499 | #ifdef HAVE_SYNC_BUILTINS |
---|
| 500 | __sync_fetch_and_add (&gomp_managed_threads, 1L - team->nthreads); |
---|
| 501 | #else |
---|
| 502 | gomp_mutex_lock (&gomp_remaining_threads_lock); |
---|
| 503 | gomp_managed_threads -= team->nthreads - 1L; |
---|
| 504 | gomp_mutex_unlock (&gomp_remaining_threads_lock); |
---|
| 505 | #endif |
---|
| 506 | /* This barrier has gomp_barrier_wait_last counterparts |
---|
| 507 | and ensures the team can be safely destroyed. */ |
---|
| 508 | gomp_barrier_wait (&team->barrier); |
---|
| 509 | } |
---|
| 510 | |
---|
| 511 | if (__builtin_expect (team->work_shares[0].next_alloc != NULL, 0)) |
---|
| 512 | { |
---|
| 513 | struct gomp_work_share *ws = team->work_shares[0].next_alloc; |
---|
| 514 | do |
---|
| 515 | { |
---|
| 516 | struct gomp_work_share *next_ws = ws->next_alloc; |
---|
| 517 | free (ws); |
---|
| 518 | ws = next_ws; |
---|
| 519 | } |
---|
| 520 | while (ws != NULL); |
---|
| 521 | } |
---|
| 522 | gomp_sem_destroy (&team->master_release); |
---|
| 523 | #ifndef HAVE_SYNC_BUILTINS |
---|
| 524 | gomp_mutex_destroy (&team->work_share_list_free_lock); |
---|
| 525 | #endif |
---|
| 526 | |
---|
| 527 | if (__builtin_expect (thr->ts.team != NULL, 0) |
---|
| 528 | || __builtin_expect (team->nthreads == 1, 0)) |
---|
| 529 | free_team (team); |
---|
| 530 | else |
---|
| 531 | { |
---|
| 532 | struct gomp_thread_pool *pool = thr->thread_pool; |
---|
| 533 | if (pool->last_team) |
---|
| 534 | free_team (pool->last_team); |
---|
| 535 | pool->last_team = team; |
---|
| 536 | } |
---|
| 537 | } |
---|
| 538 | |
---|
| 539 | /* Constructors for this file. */ |
---|
| 540 | void initialize_team (void) |
---|
| 541 | { |
---|
| 542 | struct gomp_thread *thr; |
---|
| 543 | int err; |
---|
| 544 | |
---|
| 545 | #ifndef HAVE_TLS |
---|
| 546 | static struct gomp_thread initial_thread_tls_data; |
---|
| 547 | |
---|
| 548 | if((err=pthread_key_create (&gomp_tls_key, NULL))) |
---|
| 549 | gomp_fatal ("could not create gomp_tls_key"); |
---|
| 550 | |
---|
| 551 | |
---|
| 552 | if((err=pthread_setspecific (gomp_tls_key, &initial_thread_tls_data))) |
---|
| 553 | gomp_fatal ("could not seting initial thread tls data"); |
---|
| 554 | #endif |
---|
| 555 | |
---|
| 556 | if (pthread_key_create (&gomp_thread_destructor, gomp_free_thread) != 0) |
---|
| 557 | gomp_fatal ("could not create thread pool destructor."); |
---|
| 558 | |
---|
| 559 | #ifdef HAVE_TLS |
---|
| 560 | thr = &gomp_tls_data; |
---|
| 561 | #else |
---|
| 562 | thr = &initial_thread_tls_data; |
---|
| 563 | #endif |
---|
| 564 | gomp_sem_init (&thr->release, 0); |
---|
| 565 | } |
---|
| 566 | |
---|
| 567 | void team_destructor (void) |
---|
| 568 | { |
---|
| 569 | /* Without this dlclose on libgomp could lead to subsequent |
---|
| 570 | crashes. */ |
---|
| 571 | //pthread_key_delete (gomp_thread_destructor); |
---|
| 572 | } |
---|
| 573 | |
---|
| 574 | struct gomp_task_icv * |
---|
| 575 | gomp_new_icv (void) |
---|
| 576 | { |
---|
| 577 | struct gomp_thread *thr = gomp_thread (); |
---|
| 578 | struct gomp_task *task = gomp_malloc (sizeof (struct gomp_task)); |
---|
| 579 | gomp_init_task (task, NULL, &gomp_global_icv); |
---|
| 580 | thr->task = task; |
---|
| 581 | pthread_setspecific (gomp_thread_destructor, thr); |
---|
| 582 | return &task->icv; |
---|
| 583 | } |
---|