145
|
1 /* Copyright (C) 2005-2020 Free Software Foundation, Inc.
|
0
|
2 Contributed by Richard Henderson <rth@redhat.com>.
|
|
3
|
111
|
4 This file is part of the GNU Offloading and Multi Processing Library
|
|
5 (libgomp).
|
0
|
6
|
|
7 Libgomp is free software; you can redistribute it and/or modify it
|
|
8 under the terms of the GNU General Public License as published by
|
|
9 the Free Software Foundation; either version 3, or (at your option)
|
|
10 any later version.
|
|
11
|
|
12 Libgomp is distributed in the hope that it will be useful, but WITHOUT ANY
|
|
13 WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
|
|
14 FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
15 more details.
|
|
16
|
|
17 Under Section 7 of GPL version 3, you are granted additional
|
|
18 permissions described in the GCC Runtime Library Exception, version
|
|
19 3.1, as published by the Free Software Foundation.
|
|
20
|
|
21 You should have received a copy of the GNU General Public License and
|
|
22 a copy of the GCC Runtime Library Exception along with this program;
|
|
23 see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
|
|
24 <http://www.gnu.org/licenses/>. */
|
|
25
|
|
26 /* This file handles the (bare) PARALLEL construct. */
|
|
27
|
|
28 #include "libgomp.h"
|
|
29 #include <limits.h>
|
|
30
|
|
31
|
|
32 /* Determine the number of threads to be launched for a PARALLEL construct.
|
|
33 This algorithm is explicitly described in OpenMP 3.0 section 2.4.1.
|
|
34 SPECIFIED is a combination of the NUM_THREADS clause and the IF clause.
|
|
35 If the IF clause is false, SPECIFIED is forced to 1. When NUM_THREADS
|
|
36 is not present, SPECIFIED is 0. */
|
|
37
|
|
38 unsigned
|
|
39 gomp_resolve_num_threads (unsigned specified, unsigned count)
|
|
40 {
|
111
|
41 struct gomp_thread *thr = gomp_thread ();
|
0
|
42 struct gomp_task_icv *icv;
|
|
43 unsigned threads_requested, max_num_threads, num_threads;
|
111
|
44 unsigned long busy;
|
|
45 struct gomp_thread_pool *pool;
|
0
|
46
|
|
47 icv = gomp_icv (false);
|
|
48
|
|
49 if (specified == 1)
|
|
50 return 1;
|
111
|
51 else if (thr->ts.active_level >= 1 && !icv->nest_var)
|
0
|
52 return 1;
|
111
|
53 else if (thr->ts.active_level >= gomp_max_active_levels_var)
|
0
|
54 return 1;
|
|
55
|
|
56 /* If NUM_THREADS not specified, use nthreads_var. */
|
|
57 if (specified == 0)
|
|
58 threads_requested = icv->nthreads_var;
|
|
59 else
|
|
60 threads_requested = specified;
|
|
61
|
|
62 max_num_threads = threads_requested;
|
|
63
|
|
64 /* If dynamic threads are enabled, bound the number of threads
|
|
65 that we launch. */
|
|
66 if (icv->dyn_var)
|
|
67 {
|
|
68 unsigned dyn = gomp_dynamic_max_threads ();
|
|
69 if (dyn < max_num_threads)
|
|
70 max_num_threads = dyn;
|
|
71
|
|
72 /* Optimization for parallel sections. */
|
|
73 if (count && count < max_num_threads)
|
|
74 max_num_threads = count;
|
|
75 }
|
|
76
|
111
|
77 /* UINT_MAX stands for infinity. */
|
|
78 if (__builtin_expect (icv->thread_limit_var == UINT_MAX, 1)
|
0
|
79 || max_num_threads == 1)
|
|
80 return max_num_threads;
|
|
81
|
111
|
82 /* The threads_busy counter lives in thread_pool, if there
|
|
83 isn't a thread_pool yet, there must be just one thread
|
|
84 in the contention group. If thr->team is NULL, this isn't
|
|
85 nested parallel, so there is just one thread in the
|
|
86 contention group as well, no need to handle it atomically. */
|
|
87 pool = thr->thread_pool;
|
|
88 if (thr->ts.team == NULL || pool == NULL)
|
|
89 {
|
|
90 num_threads = max_num_threads;
|
|
91 if (num_threads > icv->thread_limit_var)
|
|
92 num_threads = icv->thread_limit_var;
|
|
93 if (pool)
|
|
94 pool->threads_busy = num_threads;
|
|
95 return num_threads;
|
|
96 }
|
|
97
|
0
|
98 #ifdef HAVE_SYNC_BUILTINS
|
|
99 do
|
|
100 {
|
111
|
101 busy = pool->threads_busy;
|
0
|
102 num_threads = max_num_threads;
|
111
|
103 if (icv->thread_limit_var - busy + 1 < num_threads)
|
|
104 num_threads = icv->thread_limit_var - busy + 1;
|
0
|
105 }
|
111
|
106 while (__sync_val_compare_and_swap (&pool->threads_busy,
|
|
107 busy, busy + num_threads - 1)
|
|
108 != busy);
|
0
|
109 #else
|
111
|
110 gomp_mutex_lock (&gomp_managed_threads_lock);
|
0
|
111 num_threads = max_num_threads;
|
111
|
112 busy = pool->threads_busy;
|
|
113 if (icv->thread_limit_var - busy + 1 < num_threads)
|
|
114 num_threads = icv->thread_limit_var - busy + 1;
|
|
115 pool->threads_busy += num_threads - 1;
|
|
116 gomp_mutex_unlock (&gomp_managed_threads_lock);
|
0
|
117 #endif
|
|
118
|
|
119 return num_threads;
|
|
120 }
|
|
121
|
|
122 void
|
|
123 GOMP_parallel_start (void (*fn) (void *), void *data, unsigned num_threads)
|
|
124 {
|
|
125 num_threads = gomp_resolve_num_threads (num_threads, 0);
|
145
|
126 gomp_team_start (fn, data, num_threads, 0, gomp_new_team (num_threads),
|
|
127 NULL);
|
0
|
128 }
|
|
129
|
|
130 void
|
|
131 GOMP_parallel_end (void)
|
|
132 {
|
111
|
133 struct gomp_task_icv *icv = gomp_icv (false);
|
|
134 if (__builtin_expect (icv->thread_limit_var != UINT_MAX, 0))
|
0
|
135 {
|
|
136 struct gomp_thread *thr = gomp_thread ();
|
|
137 struct gomp_team *team = thr->ts.team;
|
111
|
138 unsigned int nthreads = team ? team->nthreads : 1;
|
|
139 gomp_team_end ();
|
|
140 if (nthreads > 1)
|
0
|
141 {
|
111
|
142 /* If not nested, there is just one thread in the
|
|
143 contention group left, no need for atomicity. */
|
|
144 if (thr->ts.team == NULL)
|
|
145 thr->thread_pool->threads_busy = 1;
|
|
146 else
|
|
147 {
|
0
|
148 #ifdef HAVE_SYNC_BUILTINS
|
111
|
149 __sync_fetch_and_add (&thr->thread_pool->threads_busy,
|
|
150 1UL - nthreads);
|
0
|
151 #else
|
111
|
152 gomp_mutex_lock (&gomp_managed_threads_lock);
|
|
153 thr->thread_pool->threads_busy -= nthreads - 1;
|
|
154 gomp_mutex_unlock (&gomp_managed_threads_lock);
|
0
|
155 #endif
|
111
|
156 }
|
0
|
157 }
|
|
158 }
|
111
|
159 else
|
|
160 gomp_team_end ();
|
|
161 }
|
|
162 ialias (GOMP_parallel_end)
|
|
163
|
|
164 void
|
145
|
165 GOMP_parallel (void (*fn) (void *), void *data, unsigned num_threads,
|
|
166 unsigned int flags)
|
111
|
167 {
|
|
168 num_threads = gomp_resolve_num_threads (num_threads, 0);
|
145
|
169 gomp_team_start (fn, data, num_threads, flags, gomp_new_team (num_threads),
|
|
170 NULL);
|
111
|
171 fn (data);
|
|
172 ialias_call (GOMP_parallel_end) ();
|
0
|
173 }
|
|
174
|
145
|
175 unsigned
|
|
176 GOMP_parallel_reductions (void (*fn) (void *), void *data,
|
|
177 unsigned num_threads, unsigned int flags)
|
|
178 {
|
|
179 struct gomp_taskgroup *taskgroup;
|
|
180 num_threads = gomp_resolve_num_threads (num_threads, 0);
|
|
181 uintptr_t *rdata = *(uintptr_t **)data;
|
|
182 taskgroup = gomp_parallel_reduction_register (rdata, num_threads);
|
|
183 gomp_team_start (fn, data, num_threads, flags, gomp_new_team (num_threads),
|
|
184 taskgroup);
|
|
185 fn (data);
|
|
186 ialias_call (GOMP_parallel_end) ();
|
|
187 gomp_sem_destroy (&taskgroup->taskgroup_sem);
|
|
188 free (taskgroup);
|
|
189 return num_threads;
|
|
190 }
|
|
191
|
111
|
192 bool
|
|
193 GOMP_cancellation_point (int which)
|
|
194 {
|
|
195 if (!gomp_cancel_var)
|
|
196 return false;
|
|
197
|
|
198 struct gomp_thread *thr = gomp_thread ();
|
|
199 struct gomp_team *team = thr->ts.team;
|
|
200 if (which & (GOMP_CANCEL_LOOP | GOMP_CANCEL_SECTIONS))
|
|
201 {
|
|
202 if (team == NULL)
|
|
203 return false;
|
|
204 return team->work_share_cancelled != 0;
|
|
205 }
|
|
206 else if (which & GOMP_CANCEL_TASKGROUP)
|
|
207 {
|
145
|
208 if (thr->task->taskgroup)
|
|
209 {
|
|
210 if (thr->task->taskgroup->cancelled)
|
|
211 return true;
|
|
212 if (thr->task->taskgroup->workshare
|
|
213 && thr->task->taskgroup->prev
|
|
214 && thr->task->taskgroup->prev->cancelled)
|
|
215 return true;
|
|
216 }
|
111
|
217 /* FALLTHRU into the GOMP_CANCEL_PARALLEL case,
|
|
218 as #pragma omp cancel parallel also cancels all explicit
|
|
219 tasks. */
|
|
220 }
|
|
221 if (team)
|
|
222 return gomp_team_barrier_cancelled (&team->barrier);
|
|
223 return false;
|
|
224 }
|
|
225 ialias (GOMP_cancellation_point)
|
|
226
|
|
227 bool
|
|
228 GOMP_cancel (int which, bool do_cancel)
|
|
229 {
|
|
230 if (!gomp_cancel_var)
|
|
231 return false;
|
|
232
|
|
233 if (!do_cancel)
|
|
234 return ialias_call (GOMP_cancellation_point) (which);
|
|
235
|
|
236 struct gomp_thread *thr = gomp_thread ();
|
|
237 struct gomp_team *team = thr->ts.team;
|
|
238 if (which & (GOMP_CANCEL_LOOP | GOMP_CANCEL_SECTIONS))
|
|
239 {
|
|
240 /* In orphaned worksharing region, all we want to cancel
|
|
241 is current thread. */
|
|
242 if (team != NULL)
|
|
243 team->work_share_cancelled = 1;
|
|
244 return true;
|
|
245 }
|
|
246 else if (which & GOMP_CANCEL_TASKGROUP)
|
|
247 {
|
145
|
248 if (thr->task->taskgroup)
|
111
|
249 {
|
145
|
250 struct gomp_taskgroup *taskgroup = thr->task->taskgroup;
|
|
251 if (taskgroup->workshare && taskgroup->prev)
|
|
252 taskgroup = taskgroup->prev;
|
|
253 if (!taskgroup->cancelled)
|
|
254 {
|
|
255 gomp_mutex_lock (&team->task_lock);
|
|
256 taskgroup->cancelled = true;
|
|
257 gomp_mutex_unlock (&team->task_lock);
|
|
258 }
|
111
|
259 }
|
|
260 return true;
|
|
261 }
|
|
262 team->team_cancelled = 1;
|
|
263 gomp_team_barrier_cancel (team);
|
|
264 return true;
|
|
265 }
|
0
|
266
|
|
267 /* The public OpenMP API for thread and team related inquiries. */
|
|
268
|
|
269 int
|
|
270 omp_get_num_threads (void)
|
|
271 {
|
|
272 struct gomp_team *team = gomp_thread ()->ts.team;
|
|
273 return team ? team->nthreads : 1;
|
|
274 }
|
|
275
|
|
276 int
|
|
277 omp_get_thread_num (void)
|
|
278 {
|
|
279 return gomp_thread ()->ts.team_id;
|
|
280 }
|
|
281
|
|
282 /* This wasn't right for OpenMP 2.5. Active region used to be non-zero
|
|
283 when the IF clause doesn't evaluate to false, starting with OpenMP 3.0
|
|
284 it is non-zero with more than one thread in the team. */
|
|
285
|
|
286 int
|
|
287 omp_in_parallel (void)
|
|
288 {
|
|
289 return gomp_thread ()->ts.active_level > 0;
|
|
290 }
|
|
291
|
|
292 int
|
|
293 omp_get_level (void)
|
|
294 {
|
|
295 return gomp_thread ()->ts.level;
|
|
296 }
|
|
297
|
|
298 int
|
|
299 omp_get_ancestor_thread_num (int level)
|
|
300 {
|
|
301 struct gomp_team_state *ts = &gomp_thread ()->ts;
|
|
302 if (level < 0 || level > ts->level)
|
|
303 return -1;
|
|
304 for (level = ts->level - level; level > 0; --level)
|
|
305 ts = &ts->team->prev_ts;
|
|
306 return ts->team_id;
|
|
307 }
|
|
308
|
|
309 int
|
|
310 omp_get_team_size (int level)
|
|
311 {
|
|
312 struct gomp_team_state *ts = &gomp_thread ()->ts;
|
|
313 if (level < 0 || level > ts->level)
|
|
314 return -1;
|
|
315 for (level = ts->level - level; level > 0; --level)
|
|
316 ts = &ts->team->prev_ts;
|
|
317 if (ts->team == NULL)
|
|
318 return 1;
|
|
319 else
|
|
320 return ts->team->nthreads;
|
|
321 }
|
|
322
|
|
323 int
|
|
324 omp_get_active_level (void)
|
|
325 {
|
|
326 return gomp_thread ()->ts.active_level;
|
|
327 }
|
|
328
|
|
329 ialias (omp_get_num_threads)
|
|
330 ialias (omp_get_thread_num)
|
|
331 ialias (omp_in_parallel)
|
|
332 ialias (omp_get_level)
|
|
333 ialias (omp_get_ancestor_thread_num)
|
|
334 ialias (omp_get_team_size)
|
|
335 ialias (omp_get_active_level)
|