| 1 | /* |
| 2 | Copyright (c) 2005-2019 Intel Corporation |
| 3 | |
| 4 | Licensed under the Apache License, Version 2.0 (the "License"); |
| 5 | you may not use this file except in compliance with the License. |
| 6 | You may obtain a copy of the License at |
| 7 | |
| 8 | http://www.apache.org/licenses/LICENSE-2.0 |
| 9 | |
| 10 | Unless required by applicable law or agreed to in writing, software |
| 11 | distributed under the License is distributed on an "AS IS" BASIS, |
| 12 | WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| 13 | See the License for the specific language governing permissions and |
| 14 | limitations under the License. |
| 15 | */ |
| 16 | |
| 17 | // undefine __TBB_CPF_BUILD to simulate user's setup |
| 18 | #undef __TBB_CPF_BUILD |
| 19 | |
| 20 | #include "tbb/tbb_config.h" |
| 21 | #include "harness.h" |
| 22 | |
| 23 | #if __TBB_SCHEDULER_OBSERVER |
| 24 | #include "tbb/task_scheduler_observer.h" |
| 25 | #include "tbb/task_scheduler_init.h" |
| 26 | #include "tbb/atomic.h" |
| 27 | #include "tbb/task.h" |
| 28 | #include "tbb/enumerable_thread_specific.h" |
| 29 | #include "../tbb/tls.h" |
| 30 | #include "tbb/tick_count.h" |
| 31 | #include "harness_barrier.h" |
| 32 | |
| 33 | #if _MSC_VER && __TBB_NO_IMPLICIT_LINKAGE |
| 34 | // plays around __TBB_NO_IMPLICIT_LINKAGE. __TBB_LIB_NAME should be defined (in makefiles) |
| 35 | #pragma comment(lib, __TBB_STRING(__TBB_LIB_NAME)) |
| 36 | #endif |
| 37 | |
| 38 | const int MaxFlagIndex = sizeof(uintptr_t)*8-1; |
| 39 | |
| 40 | struct ObserverStats { |
| 41 | tbb::atomic<int> m_entries; |
| 42 | tbb::atomic<int> m_exits; |
| 43 | tbb::atomic<int> m_workerEntries; |
| 44 | tbb::atomic<int> m_workerExits; |
| 45 | |
| 46 | void Reset () { |
| 47 | m_entries = m_exits = m_workerEntries = m_workerExits = 0; |
| 48 | } |
| 49 | |
| 50 | void operator += ( const ObserverStats& s ) { |
| 51 | m_entries += s.m_entries; |
| 52 | m_exits += s.m_exits; |
| 53 | m_workerEntries += s.m_workerEntries; |
| 54 | m_workerExits += s.m_workerExits; |
| 55 | } |
| 56 | }; |
| 57 | |
| 58 | struct ThreadState { |
| 59 | uintptr_t m_flags; |
| 60 | tbb::task_scheduler_observer *m_dyingObserver; |
| 61 | bool m_isMaster; |
| 62 | ThreadState() { reset(); } |
| 63 | void reset() { |
| 64 | m_flags = 0; |
| 65 | m_dyingObserver = NULL; |
| 66 | m_isMaster = false; |
| 67 | } |
| 68 | static ThreadState &get(); |
| 69 | }; |
| 70 | |
| 71 | tbb::enumerable_thread_specific<ThreadState> theLocalState; |
| 72 | tbb::internal::tls<intptr_t> theThreadPrivate; |
| 73 | |
| 74 | ThreadState &ThreadState::get() { |
| 75 | bool exists; |
| 76 | ThreadState& state = theLocalState.local(exists); |
| 77 | // ETS will not detect that a thread was allocated with the same id as a destroyed thread |
| 78 | if( exists && theThreadPrivate.get() == 0 ) state.reset(); |
| 79 | theThreadPrivate = 1; // mark thread constructed |
| 80 | return state; |
| 81 | } |
| 82 | |
| 83 | static ObserverStats theStats; |
| 84 | static tbb::atomic<int> theNumObservers; |
| 85 | |
| 86 | const int P = min( tbb::task_scheduler_init::default_num_threads(), (int)sizeof(int) * CHAR_BIT ); |
| 87 | |
| 88 | enum TestMode { |
| 89 | //! Ensure timely workers destruction in order to guarantee all exit notification are fired. |
| 90 | tmSynchronized = 1, |
| 91 | //! Use local observer. |
| 92 | tmLocalObservation = 2, |
| 93 | //! Observer causes autoinitialization of the scheduler |
| 94 | tmAutoinitialization = 4 |
| 95 | }; |
| 96 | |
| 97 | uintptr_t theTestMode, |
| 98 | thePrevMode = 0; |
| 99 | |
| 100 | class MyObserver : public tbb::task_scheduler_observer, public ObserverStats { |
| 101 | uintptr_t m_flag; |
| 102 | tbb::atomic<bool> m_dying; |
| 103 | |
| 104 | void on_scheduler_entry( bool is_worker ) __TBB_override { |
| 105 | ThreadState& state = ThreadState::get(); |
| 106 | ASSERT( is_worker==!state.m_isMaster, NULL ); |
| 107 | if ( thePrevMode & tmSynchronized ) { |
| 108 | ASSERT( !(state.m_flags & m_flag), "Observer repeatedly invoked for the same thread" ); |
| 109 | if ( theTestMode & tmLocalObservation ) |
| 110 | ASSERT( !state.m_flags, "Observer locality breached" ); |
| 111 | } |
| 112 | if ( m_dying && theTestMode & tmLocalObservation ) { |
| 113 | // In case of local observation a worker may enter the arena after |
| 114 | // the wait for lagging on_entry calls in the MyObserver destructor |
| 115 | // succeeds but before its base class tbb::task_scheduler_observer |
| 116 | // destructor removes it from the internal list maintained by the |
| 117 | // task scheduler. This will result in on_entry notification without, |
| 118 | // subsequent on_exit as the observer is likely to be destroyed before |
| 119 | // the worker discovers that the arena is empty and leaves it. |
| 120 | // |
| 121 | // To prevent statistics distortion, ignore the notifications for |
| 122 | // observers about to be destroyed. |
| 123 | ASSERT( !state.m_dyingObserver || state.m_dyingObserver != this || thePrevMode & tmSynchronized, NULL ); |
| 124 | state.m_dyingObserver = this; |
| 125 | return; |
| 126 | } |
| 127 | state.m_dyingObserver = NULL; |
| 128 | ++m_entries; |
| 129 | state.m_flags |= m_flag; |
| 130 | if ( is_worker ) |
| 131 | ++m_workerEntries; |
| 132 | } |
| 133 | void on_scheduler_exit( bool is_worker ) __TBB_override { |
| 134 | ThreadState& state = ThreadState::get(); |
| 135 | ASSERT( is_worker==!state.m_isMaster, NULL ); |
| 136 | if ( m_dying && state.m_dyingObserver ) { |
| 137 | ASSERT( state.m_dyingObserver == this, "Exit without entry (for a dying observer)" ); |
| 138 | state.m_dyingObserver = NULL; |
| 139 | return; |
| 140 | } |
| 141 | ASSERT( state.m_flags & m_flag, "Exit without entry" ); |
| 142 | state.m_flags &= ~m_flag; |
| 143 | ++m_exits; |
| 144 | if ( is_worker ) |
| 145 | ++m_workerExits; |
| 146 | } |
| 147 | public: |
| 148 | MyObserver( uintptr_t flag ) |
| 149 | : tbb::task_scheduler_observer(theTestMode & tmLocalObservation ? true : false) |
| 150 | , m_flag(flag) |
| 151 | { |
| 152 | ++theNumObservers; |
| 153 | Reset(); |
| 154 | m_dying = false; |
| 155 | // Local observer causes automatic scheduler initialization |
| 156 | // in the current thread, so here, we must postpone the activation. |
| 157 | if ( !(theTestMode & tmLocalObservation)) |
| 158 | observe(true); |
| 159 | } |
| 160 | |
| 161 | ~MyObserver () { |
| 162 | m_dying = true; |
| 163 | ASSERT( m_exits <= m_entries, NULL ); |
| 164 | if ( theTestMode & tmSynchronized ) { |
| 165 | tbb::tick_count t0 = tbb::tick_count::now(); |
| 166 | while ( m_exits < m_entries && (tbb::tick_count::now() - t0).seconds() < 5 ) |
| 167 | Harness::Sleep(10); |
| 168 | if ( m_exits < m_entries ) |
| 169 | REPORT( "Warning: Entry/exit count mismatch (%d, %d). Observer is broken or machine is overloaded.\n" , (int)m_entries, (int)m_exits ); |
| 170 | } |
| 171 | theStats += *this; |
| 172 | --theNumObservers; |
| 173 | // it is recommended to disable observation before destructor of the base class starts, |
| 174 | // otherwise it can lead to concurrent notification callback on partly destroyed object, |
| 175 | // which in turn can harm (in addition) if derived class has new virtual methods. |
| 176 | // This class has no, and for test purposes we rely on implementation failsafe mechanism. |
| 177 | //observe(false); |
| 178 | } |
| 179 | }; // class MyObserver |
| 180 | |
| 181 | Harness::SpinBarrier theGlobalBarrier; |
| 182 | bool theGlobalBarrierActive = true; |
| 183 | |
| 184 | class FibTask : public tbb::task { |
| 185 | const int N; |
| 186 | uintptr_t m_flag; |
| 187 | MyObserver &m_observer; |
| 188 | public: |
| 189 | FibTask( int n, uintptr_t flags, MyObserver &obs ) : N(n), m_flag(flags), m_observer(obs) {} |
| 190 | |
| 191 | tbb::task* execute() __TBB_override { |
| 192 | ThreadState& s = ThreadState::get(); |
| 193 | ASSERT( !(~s.m_flags & m_flag), NULL ); |
| 194 | if( N < 2 ) |
| 195 | return NULL; |
| 196 | bool globalBarrierActive = false; |
| 197 | if ( s.m_isMaster ) { |
| 198 | if ( theGlobalBarrierActive ) { |
| 199 | // This is the root task. Its N is equal to the number of threads. |
| 200 | // Spawn a task for each worker. |
| 201 | set_ref_count(N); |
| 202 | for ( int i = 1; i < N; ++i ) |
| 203 | spawn( *new( allocate_child() ) FibTask(20, m_flag, m_observer) ); |
| 204 | if ( theTestMode & tmSynchronized ) { |
| 205 | theGlobalBarrier.wait(); |
| 206 | ASSERT( m_observer.m_entries >= N, "Wrong number of on_entry calls after the first barrier" ); |
| 207 | // All the spawned tasks have been stolen by workers. |
| 208 | // Now wait for workers to spawn some more tasks for this thread to steal back. |
| 209 | theGlobalBarrier.wait(); |
| 210 | ASSERT( !theGlobalBarrierActive, "Workers are expected to have reset this flag" ); |
| 211 | } |
| 212 | else |
| 213 | theGlobalBarrierActive = false; |
| 214 | wait_for_all(); |
| 215 | return NULL; |
| 216 | } |
| 217 | } |
| 218 | else { |
| 219 | if ( theGlobalBarrierActive ) { |
| 220 | if ( theTestMode & tmSynchronized ) { |
| 221 | theGlobalBarrier.wait(); |
| 222 | globalBarrierActive = true; |
| 223 | } |
| 224 | theGlobalBarrierActive = false; |
| 225 | } |
| 226 | } |
| 227 | set_ref_count(3); |
| 228 | spawn( *new( allocate_child() ) FibTask(N-1, m_flag, m_observer) ); |
| 229 | spawn( *new( allocate_child() ) FibTask(N-2, m_flag, m_observer) ); |
| 230 | if ( globalBarrierActive ) { |
| 231 | // It's the first task executed by a worker. Release the master thread. |
| 232 | theGlobalBarrier.wait(); |
| 233 | } |
| 234 | wait_for_all(); |
| 235 | return NULL; |
| 236 | } |
| 237 | }; // class FibTask |
| 238 | |
| 239 | Harness::SpinBarrier theMasterBarrier; |
| 240 | |
| 241 | class TestBody { |
| 242 | int m_numThreads; |
| 243 | public: |
| 244 | TestBody( int numThreads ) : m_numThreads(numThreads) {} |
| 245 | |
| 246 | void operator()( int i ) const { |
| 247 | ThreadState &state = ThreadState::get(); |
| 248 | ASSERT( !state.m_isMaster, "should be newly initialized thread" ); |
| 249 | state.m_isMaster = true; |
| 250 | uintptr_t f = i <= MaxFlagIndex ? 1<<i : 0; |
| 251 | MyObserver o(f); |
| 252 | if ( theTestMode & tmSynchronized ) |
| 253 | theMasterBarrier.wait(); |
| 254 | // when mode is local observation but not synchronized and when num threads == default |
| 255 | if ( theTestMode & tmAutoinitialization ) |
| 256 | o.observe(true); // test autoinitialization can be done by observer |
| 257 | // Observer in enabled state must outlive the scheduler to ensure that |
| 258 | // all exit notifications are called. |
| 259 | tbb::task_scheduler_init init(m_numThreads); |
| 260 | // when local & non-autoinitialized observation mode |
| 261 | if ( theTestMode & tmLocalObservation ) |
| 262 | o.observe(true); |
| 263 | for ( int j = 0; j < 2; ++j ) { |
| 264 | tbb::task &t = *new( tbb::task::allocate_root() ) FibTask(m_numThreads, f, o); |
| 265 | tbb::task::spawn_root_and_wait(t); |
| 266 | thePrevMode = theTestMode; |
| 267 | } |
| 268 | } |
| 269 | }; // class TestBody |
| 270 | |
| 271 | void TestObserver( int M, int T, uintptr_t testMode ) { |
| 272 | theLocalState.clear(); |
| 273 | theStats.Reset(); |
| 274 | theGlobalBarrierActive = true; |
| 275 | theTestMode = testMode; |
| 276 | NativeParallelFor( M, TestBody(T) ); |
| 277 | // When T (number of threads in arena, i.e. master + workers) is less than P |
| 278 | // (hardware concurrency), more than T-1 workers can visit the same arena. This |
| 279 | // is possible in case of imbalance or when other arenas are activated/deactivated |
| 280 | // concurrently). |
| 281 | ASSERT( !theNumObservers, "Unexpected alive observer(s)" ); |
| 282 | REMARK( "Entries %d / %d, exits %d\n" , (int)theStats.m_entries, (int)theStats.m_workerEntries, (int)theStats.m_exits ); |
| 283 | if ( testMode & tmSynchronized ) { |
| 284 | if ( testMode & tmLocalObservation ) { |
| 285 | ASSERT( theStats.m_entries >= M * T, "Too few on_entry calls" ); |
| 286 | ASSERT( theStats.m_workerEntries >= M * (T - 1), "Too few worker entries" ); |
| 287 | } |
| 288 | else { |
| 289 | ASSERT( theStats.m_entries >= M * M * T, "Too few on_entry calls" ); |
| 290 | ASSERT( theStats.m_entries <= M * (P + 1), "Too many on_entry calls" ); |
| 291 | ASSERT( theStats.m_workerEntries >= M * M * (T - 1), "Too few worker entries" ); |
| 292 | ASSERT( theStats.m_workerEntries <= M * (P - 1), "Too many worker entries" ); |
| 293 | } |
| 294 | ASSERT( theStats.m_entries == theStats.m_exits, "Entries/exits mismatch" ); |
| 295 | } |
| 296 | else { |
| 297 | ASSERT( theStats.m_entries >= M, "Too few on_entry calls" ); |
| 298 | ASSERT( theStats.m_exits >= M || (testMode & tmAutoinitialization), "Too few on_exit calls" ); |
| 299 | if ( !(testMode & tmLocalObservation) ) { |
| 300 | ASSERT( theStats.m_entries <= M * M * P, "Too many on_entry calls" ); |
| 301 | ASSERT( theStats.m_exits <= M * M * T, "Too many on_exit calls" ); |
| 302 | } |
| 303 | ASSERT( theStats.m_entries >= theStats.m_exits, "More exits than entries" ); |
| 304 | } |
| 305 | } |
| 306 | |
| 307 | int TestMain () { |
| 308 | if ( P < 2 ) |
| 309 | return Harness::Skipped; |
| 310 | theNumObservers = 0; |
| 311 | // Fully- and under-utilized mode |
| 312 | for ( int M = 1; M < P; M <<= 1 ) { |
| 313 | if ( M > P/2 ) { |
| 314 | ASSERT( P & (P-1), "Can get here only in case of non power of two cores" ); |
| 315 | M = P/2; |
| 316 | if ( M==1 || (M & (M-1)) ) |
| 317 | break; // Already tested this configuration |
| 318 | } |
| 319 | int T = P / M; |
| 320 | ASSERT( T > 1, NULL ); |
| 321 | REMARK( "Masters: %d; Arena size: %d\n" , M, T ); |
| 322 | theMasterBarrier.initialize(M); |
| 323 | theGlobalBarrier.initialize(M * T); |
| 324 | TestObserver(M, T, 0); |
| 325 | TestObserver(M, T, tmSynchronized | tmLocalObservation ); |
| 326 | // keep tmAutoInitialization the last, as it does not release worker threads |
| 327 | TestObserver(M, T, tmLocalObservation | ( T==P? tmAutoinitialization : 0) ); |
| 328 | } |
| 329 | // Oversubscribed mode |
| 330 | for ( int i = 0; i < 4; ++i ) { |
| 331 | REMARK( "Masters: %d; Arena size: %d\n" , P-1, P ); |
| 332 | TestObserver(P-1, P, 0); |
| 333 | TestObserver(P-1, P, tmLocalObservation); |
| 334 | } |
| 335 | Harness::Sleep(20); |
| 336 | return Harness::Done; |
| 337 | } |
| 338 | |
| 339 | #else /* !__TBB_SCHEDULER_OBSERVER */ |
| 340 | |
| 341 | int TestMain () { |
| 342 | return Harness::Skipped; |
| 343 | } |
| 344 | #endif /* !__TBB_SCHEDULER_OBSERVER */ |
| 345 | |