1 | /* |
2 | Copyright (c) 2005-2019 Intel Corporation |
3 | |
4 | Licensed under the Apache License, Version 2.0 (the "License"); |
5 | you may not use this file except in compliance with the License. |
6 | You may obtain a copy of the License at |
7 | |
8 | http://www.apache.org/licenses/LICENSE-2.0 |
9 | |
10 | Unless required by applicable law or agreed to in writing, software |
11 | distributed under the License is distributed on an "AS IS" BASIS, |
12 | WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
13 | See the License for the specific language governing permissions and |
14 | limitations under the License. |
15 | */ |
16 | |
17 | // undefine __TBB_CPF_BUILD to simulate user's setup |
18 | #undef __TBB_CPF_BUILD |
19 | |
20 | #include "tbb/tbb_config.h" |
21 | #include "harness.h" |
22 | |
23 | #if __TBB_SCHEDULER_OBSERVER |
24 | #include "tbb/task_scheduler_observer.h" |
25 | #include "tbb/task_scheduler_init.h" |
26 | #include "tbb/atomic.h" |
27 | #include "tbb/task.h" |
28 | #include "tbb/enumerable_thread_specific.h" |
29 | #include "../tbb/tls.h" |
30 | #include "tbb/tick_count.h" |
31 | #include "harness_barrier.h" |
32 | |
33 | #if _MSC_VER && __TBB_NO_IMPLICIT_LINKAGE |
34 | // plays around __TBB_NO_IMPLICIT_LINKAGE. __TBB_LIB_NAME should be defined (in makefiles) |
35 | #pragma comment(lib, __TBB_STRING(__TBB_LIB_NAME)) |
36 | #endif |
37 | |
38 | const int MaxFlagIndex = sizeof(uintptr_t)*8-1; |
39 | |
40 | struct ObserverStats { |
41 | tbb::atomic<int> m_entries; |
42 | tbb::atomic<int> m_exits; |
43 | tbb::atomic<int> m_workerEntries; |
44 | tbb::atomic<int> m_workerExits; |
45 | |
46 | void Reset () { |
47 | m_entries = m_exits = m_workerEntries = m_workerExits = 0; |
48 | } |
49 | |
50 | void operator += ( const ObserverStats& s ) { |
51 | m_entries += s.m_entries; |
52 | m_exits += s.m_exits; |
53 | m_workerEntries += s.m_workerEntries; |
54 | m_workerExits += s.m_workerExits; |
55 | } |
56 | }; |
57 | |
58 | struct ThreadState { |
59 | uintptr_t m_flags; |
60 | tbb::task_scheduler_observer *m_dyingObserver; |
61 | bool m_isMaster; |
62 | ThreadState() { reset(); } |
63 | void reset() { |
64 | m_flags = 0; |
65 | m_dyingObserver = NULL; |
66 | m_isMaster = false; |
67 | } |
68 | static ThreadState &get(); |
69 | }; |
70 | |
71 | tbb::enumerable_thread_specific<ThreadState> theLocalState; |
72 | tbb::internal::tls<intptr_t> theThreadPrivate; |
73 | |
74 | ThreadState &ThreadState::get() { |
75 | bool exists; |
76 | ThreadState& state = theLocalState.local(exists); |
77 | // ETS will not detect that a thread was allocated with the same id as a destroyed thread |
78 | if( exists && theThreadPrivate.get() == 0 ) state.reset(); |
79 | theThreadPrivate = 1; // mark thread constructed |
80 | return state; |
81 | } |
82 | |
83 | static ObserverStats theStats; |
84 | static tbb::atomic<int> theNumObservers; |
85 | |
86 | const int P = min( tbb::task_scheduler_init::default_num_threads(), (int)sizeof(int) * CHAR_BIT ); |
87 | |
88 | enum TestMode { |
89 | //! Ensure timely workers destruction in order to guarantee all exit notification are fired. |
90 | tmSynchronized = 1, |
91 | //! Use local observer. |
92 | tmLocalObservation = 2, |
93 | //! Observer causes autoinitialization of the scheduler |
94 | tmAutoinitialization = 4 |
95 | }; |
96 | |
97 | uintptr_t theTestMode, |
98 | thePrevMode = 0; |
99 | |
100 | class MyObserver : public tbb::task_scheduler_observer, public ObserverStats { |
101 | uintptr_t m_flag; |
102 | tbb::atomic<bool> m_dying; |
103 | |
104 | void on_scheduler_entry( bool is_worker ) __TBB_override { |
105 | ThreadState& state = ThreadState::get(); |
106 | ASSERT( is_worker==!state.m_isMaster, NULL ); |
107 | if ( thePrevMode & tmSynchronized ) { |
108 | ASSERT( !(state.m_flags & m_flag), "Observer repeatedly invoked for the same thread" ); |
109 | if ( theTestMode & tmLocalObservation ) |
110 | ASSERT( !state.m_flags, "Observer locality breached" ); |
111 | } |
112 | if ( m_dying && theTestMode & tmLocalObservation ) { |
113 | // In case of local observation a worker may enter the arena after |
114 | // the wait for lagging on_entry calls in the MyObserver destructor |
115 | // succeeds but before its base class tbb::task_scheduler_observer |
116 | // destructor removes it from the internal list maintained by the |
117 | // task scheduler. This will result in on_entry notification without, |
118 | // subsequent on_exit as the observer is likely to be destroyed before |
119 | // the worker discovers that the arena is empty and leaves it. |
120 | // |
121 | // To prevent statistics distortion, ignore the notifications for |
122 | // observers about to be destroyed. |
123 | ASSERT( !state.m_dyingObserver || state.m_dyingObserver != this || thePrevMode & tmSynchronized, NULL ); |
124 | state.m_dyingObserver = this; |
125 | return; |
126 | } |
127 | state.m_dyingObserver = NULL; |
128 | ++m_entries; |
129 | state.m_flags |= m_flag; |
130 | if ( is_worker ) |
131 | ++m_workerEntries; |
132 | } |
133 | void on_scheduler_exit( bool is_worker ) __TBB_override { |
134 | ThreadState& state = ThreadState::get(); |
135 | ASSERT( is_worker==!state.m_isMaster, NULL ); |
136 | if ( m_dying && state.m_dyingObserver ) { |
137 | ASSERT( state.m_dyingObserver == this, "Exit without entry (for a dying observer)" ); |
138 | state.m_dyingObserver = NULL; |
139 | return; |
140 | } |
141 | ASSERT( state.m_flags & m_flag, "Exit without entry" ); |
142 | state.m_flags &= ~m_flag; |
143 | ++m_exits; |
144 | if ( is_worker ) |
145 | ++m_workerExits; |
146 | } |
147 | public: |
148 | MyObserver( uintptr_t flag ) |
149 | : tbb::task_scheduler_observer(theTestMode & tmLocalObservation ? true : false) |
150 | , m_flag(flag) |
151 | { |
152 | ++theNumObservers; |
153 | Reset(); |
154 | m_dying = false; |
155 | // Local observer causes automatic scheduler initialization |
156 | // in the current thread, so here, we must postpone the activation. |
157 | if ( !(theTestMode & tmLocalObservation)) |
158 | observe(true); |
159 | } |
160 | |
161 | ~MyObserver () { |
162 | m_dying = true; |
163 | ASSERT( m_exits <= m_entries, NULL ); |
164 | if ( theTestMode & tmSynchronized ) { |
165 | tbb::tick_count t0 = tbb::tick_count::now(); |
166 | while ( m_exits < m_entries && (tbb::tick_count::now() - t0).seconds() < 5 ) |
167 | Harness::Sleep(10); |
168 | if ( m_exits < m_entries ) |
169 | REPORT( "Warning: Entry/exit count mismatch (%d, %d). Observer is broken or machine is overloaded.\n" , (int)m_entries, (int)m_exits ); |
170 | } |
171 | theStats += *this; |
172 | --theNumObservers; |
173 | // it is recommended to disable observation before destructor of the base class starts, |
174 | // otherwise it can lead to concurrent notification callback on partly destroyed object, |
175 | // which in turn can harm (in addition) if derived class has new virtual methods. |
176 | // This class has no, and for test purposes we rely on implementation failsafe mechanism. |
177 | //observe(false); |
178 | } |
179 | }; // class MyObserver |
180 | |
181 | Harness::SpinBarrier theGlobalBarrier; |
182 | bool theGlobalBarrierActive = true; |
183 | |
184 | class FibTask : public tbb::task { |
185 | const int N; |
186 | uintptr_t m_flag; |
187 | MyObserver &m_observer; |
188 | public: |
189 | FibTask( int n, uintptr_t flags, MyObserver &obs ) : N(n), m_flag(flags), m_observer(obs) {} |
190 | |
191 | tbb::task* execute() __TBB_override { |
192 | ThreadState& s = ThreadState::get(); |
193 | ASSERT( !(~s.m_flags & m_flag), NULL ); |
194 | if( N < 2 ) |
195 | return NULL; |
196 | bool globalBarrierActive = false; |
197 | if ( s.m_isMaster ) { |
198 | if ( theGlobalBarrierActive ) { |
199 | // This is the root task. Its N is equal to the number of threads. |
200 | // Spawn a task for each worker. |
201 | set_ref_count(N); |
202 | for ( int i = 1; i < N; ++i ) |
203 | spawn( *new( allocate_child() ) FibTask(20, m_flag, m_observer) ); |
204 | if ( theTestMode & tmSynchronized ) { |
205 | theGlobalBarrier.wait(); |
206 | ASSERT( m_observer.m_entries >= N, "Wrong number of on_entry calls after the first barrier" ); |
207 | // All the spawned tasks have been stolen by workers. |
208 | // Now wait for workers to spawn some more tasks for this thread to steal back. |
209 | theGlobalBarrier.wait(); |
210 | ASSERT( !theGlobalBarrierActive, "Workers are expected to have reset this flag" ); |
211 | } |
212 | else |
213 | theGlobalBarrierActive = false; |
214 | wait_for_all(); |
215 | return NULL; |
216 | } |
217 | } |
218 | else { |
219 | if ( theGlobalBarrierActive ) { |
220 | if ( theTestMode & tmSynchronized ) { |
221 | theGlobalBarrier.wait(); |
222 | globalBarrierActive = true; |
223 | } |
224 | theGlobalBarrierActive = false; |
225 | } |
226 | } |
227 | set_ref_count(3); |
228 | spawn( *new( allocate_child() ) FibTask(N-1, m_flag, m_observer) ); |
229 | spawn( *new( allocate_child() ) FibTask(N-2, m_flag, m_observer) ); |
230 | if ( globalBarrierActive ) { |
231 | // It's the first task executed by a worker. Release the master thread. |
232 | theGlobalBarrier.wait(); |
233 | } |
234 | wait_for_all(); |
235 | return NULL; |
236 | } |
237 | }; // class FibTask |
238 | |
239 | Harness::SpinBarrier theMasterBarrier; |
240 | |
241 | class TestBody { |
242 | int m_numThreads; |
243 | public: |
244 | TestBody( int numThreads ) : m_numThreads(numThreads) {} |
245 | |
246 | void operator()( int i ) const { |
247 | ThreadState &state = ThreadState::get(); |
248 | ASSERT( !state.m_isMaster, "should be newly initialized thread" ); |
249 | state.m_isMaster = true; |
250 | uintptr_t f = i <= MaxFlagIndex ? 1<<i : 0; |
251 | MyObserver o(f); |
252 | if ( theTestMode & tmSynchronized ) |
253 | theMasterBarrier.wait(); |
254 | // when mode is local observation but not synchronized and when num threads == default |
255 | if ( theTestMode & tmAutoinitialization ) |
256 | o.observe(true); // test autoinitialization can be done by observer |
257 | // Observer in enabled state must outlive the scheduler to ensure that |
258 | // all exit notifications are called. |
259 | tbb::task_scheduler_init init(m_numThreads); |
260 | // when local & non-autoinitialized observation mode |
261 | if ( theTestMode & tmLocalObservation ) |
262 | o.observe(true); |
263 | for ( int j = 0; j < 2; ++j ) { |
264 | tbb::task &t = *new( tbb::task::allocate_root() ) FibTask(m_numThreads, f, o); |
265 | tbb::task::spawn_root_and_wait(t); |
266 | thePrevMode = theTestMode; |
267 | } |
268 | } |
269 | }; // class TestBody |
270 | |
271 | void TestObserver( int M, int T, uintptr_t testMode ) { |
272 | theLocalState.clear(); |
273 | theStats.Reset(); |
274 | theGlobalBarrierActive = true; |
275 | theTestMode = testMode; |
276 | NativeParallelFor( M, TestBody(T) ); |
277 | // When T (number of threads in arena, i.e. master + workers) is less than P |
278 | // (hardware concurrency), more than T-1 workers can visit the same arena. This |
279 | // is possible in case of imbalance or when other arenas are activated/deactivated |
280 | // concurrently). |
281 | ASSERT( !theNumObservers, "Unexpected alive observer(s)" ); |
282 | REMARK( "Entries %d / %d, exits %d\n" , (int)theStats.m_entries, (int)theStats.m_workerEntries, (int)theStats.m_exits ); |
283 | if ( testMode & tmSynchronized ) { |
284 | if ( testMode & tmLocalObservation ) { |
285 | ASSERT( theStats.m_entries >= M * T, "Too few on_entry calls" ); |
286 | ASSERT( theStats.m_workerEntries >= M * (T - 1), "Too few worker entries" ); |
287 | } |
288 | else { |
289 | ASSERT( theStats.m_entries >= M * M * T, "Too few on_entry calls" ); |
290 | ASSERT( theStats.m_entries <= M * (P + 1), "Too many on_entry calls" ); |
291 | ASSERT( theStats.m_workerEntries >= M * M * (T - 1), "Too few worker entries" ); |
292 | ASSERT( theStats.m_workerEntries <= M * (P - 1), "Too many worker entries" ); |
293 | } |
294 | ASSERT( theStats.m_entries == theStats.m_exits, "Entries/exits mismatch" ); |
295 | } |
296 | else { |
297 | ASSERT( theStats.m_entries >= M, "Too few on_entry calls" ); |
298 | ASSERT( theStats.m_exits >= M || (testMode & tmAutoinitialization), "Too few on_exit calls" ); |
299 | if ( !(testMode & tmLocalObservation) ) { |
300 | ASSERT( theStats.m_entries <= M * M * P, "Too many on_entry calls" ); |
301 | ASSERT( theStats.m_exits <= M * M * T, "Too many on_exit calls" ); |
302 | } |
303 | ASSERT( theStats.m_entries >= theStats.m_exits, "More exits than entries" ); |
304 | } |
305 | } |
306 | |
307 | int TestMain () { |
308 | if ( P < 2 ) |
309 | return Harness::Skipped; |
310 | theNumObservers = 0; |
311 | // Fully- and under-utilized mode |
312 | for ( int M = 1; M < P; M <<= 1 ) { |
313 | if ( M > P/2 ) { |
314 | ASSERT( P & (P-1), "Can get here only in case of non power of two cores" ); |
315 | M = P/2; |
316 | if ( M==1 || (M & (M-1)) ) |
317 | break; // Already tested this configuration |
318 | } |
319 | int T = P / M; |
320 | ASSERT( T > 1, NULL ); |
321 | REMARK( "Masters: %d; Arena size: %d\n" , M, T ); |
322 | theMasterBarrier.initialize(M); |
323 | theGlobalBarrier.initialize(M * T); |
324 | TestObserver(M, T, 0); |
325 | TestObserver(M, T, tmSynchronized | tmLocalObservation ); |
326 | // keep tmAutoInitialization the last, as it does not release worker threads |
327 | TestObserver(M, T, tmLocalObservation | ( T==P? tmAutoinitialization : 0) ); |
328 | } |
329 | // Oversubscribed mode |
330 | for ( int i = 0; i < 4; ++i ) { |
331 | REMARK( "Masters: %d; Arena size: %d\n" , P-1, P ); |
332 | TestObserver(P-1, P, 0); |
333 | TestObserver(P-1, P, tmLocalObservation); |
334 | } |
335 | Harness::Sleep(20); |
336 | return Harness::Done; |
337 | } |
338 | |
339 | #else /* !__TBB_SCHEDULER_OBSERVER */ |
340 | |
341 | int TestMain () { |
342 | return Harness::Skipped; |
343 | } |
344 | #endif /* !__TBB_SCHEDULER_OBSERVER */ |
345 | |