| 1 | /* | 
|---|
| 2 | Copyright (c) 2005-2019 Intel Corporation | 
|---|
| 3 |  | 
|---|
| 4 | Licensed under the Apache License, Version 2.0 (the "License"); | 
|---|
| 5 | you may not use this file except in compliance with the License. | 
|---|
| 6 | You may obtain a copy of the License at | 
|---|
| 7 |  | 
|---|
| 8 | http://www.apache.org/licenses/LICENSE-2.0 | 
|---|
| 9 |  | 
|---|
| 10 | Unless required by applicable law or agreed to in writing, software | 
|---|
| 11 | distributed under the License is distributed on an "AS IS" BASIS, | 
|---|
| 12 | WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | 
|---|
| 13 | See the License for the specific language governing permissions and | 
|---|
| 14 | limitations under the License. | 
|---|
| 15 | */ | 
|---|
| 16 |  | 
|---|
| 17 | /* Common part for the partitioner whitebox tests */ | 
|---|
| 18 |  | 
|---|
| 19 | #include <typeinfo> | 
|---|
| 20 |  | 
|---|
| 21 | #include "tbb/tbb_thread.h" | 
|---|
| 22 | #include "tbb/enumerable_thread_specific.h" | 
|---|
| 23 |  | 
|---|
| 24 | #include "string.h" | 
|---|
| 25 | #include "harness_assert.h" | 
|---|
| 26 | #include "test_partitioner.h" | 
|---|
| 27 | #include <numeric> | 
|---|
| 28 |  | 
|---|
| 29 | #if TBB_USE_DEBUG | 
|---|
| 30 | // reducing number of simulations due to test timeout | 
|---|
| 31 | const size_t max_simulated_threads = 256; | 
|---|
| 32 | #else | 
|---|
| 33 | const size_t max_simulated_threads = 640; | 
|---|
| 34 | #endif | 
|---|
| 35 |  | 
|---|
| 36 | typedef tbb::enumerable_thread_specific<size_t> ThreadNumsType; | 
|---|
| 37 | size_t g_threadNumInitialValue = 10; | 
|---|
| 38 | ThreadNumsType g_threadNums(g_threadNumInitialValue); | 
|---|
| 39 |  | 
|---|
| 40 | namespace whitebox_simulation { | 
|---|
| 41 | size_t whitebox_thread_index = 0; | 
|---|
| 42 | test_partitioner_utils::BinaryTree reference_tree; | 
|---|
| 43 | } | 
|---|
| 44 |  | 
|---|
| 45 | // simulate a subset of task.h | 
|---|
| 46 | namespace tbb { | 
|---|
| 47 | namespace internal { | 
|---|
| 48 | typedef unsigned short affinity_id; | 
|---|
| 49 | } | 
|---|
| 50 | class fake_task { | 
|---|
| 51 | public: | 
|---|
| 52 | typedef internal::affinity_id affinity_id; | 
|---|
| 53 | void set_affinity(affinity_id a) { my_affinity = a; } | 
|---|
| 54 | affinity_id affinity() const { return my_affinity; } | 
|---|
| 55 | void set_parent(fake_task* p) { my_parent = p; } | 
|---|
| 56 | fake_task *parent() const { return my_parent; } | 
|---|
| 57 | bool is_stolen_task() const { return false; } | 
|---|
| 58 | intptr_t ref_count() const { return 1; } | 
|---|
| 59 | bool is_cancelled() const { return false; } | 
|---|
| 60 | static void spawn(fake_task &) {} // for legacy in partitioner.h | 
|---|
| 61 | virtual fake_task* execute() = 0; // enables dynamic_cast | 
|---|
| 62 |  | 
|---|
| 63 | fake_task() : my_parent(0), my_affinity(0) {} | 
|---|
| 64 | virtual ~fake_task() {} | 
|---|
| 65 | private: | 
|---|
| 66 | fake_task *my_parent; | 
|---|
| 67 | affinity_id my_affinity; | 
|---|
| 68 | }; | 
|---|
| 69 | namespace task_arena { | 
|---|
| 70 | static const int not_initialized = -2;//should match corresponding value in task_arena.h | 
|---|
| 71 | }//namespace task_arena | 
|---|
| 72 | namespace this_task_arena { | 
|---|
| 73 | inline int current_thread_index() { return (int)whitebox_simulation::whitebox_thread_index; } | 
|---|
| 74 | } | 
|---|
| 75 | }//namespace tbb | 
|---|
| 76 |  | 
|---|
| 77 | #define __TBB_task_H | 
|---|
| 78 | #define __TBB_task_arena_H | 
|---|
| 79 | #define get_initial_auto_partitioner_divisor my_get_initial_auto_partitioner_divisor | 
|---|
| 80 | #define affinity_partitioner_base_v3 my_affinity_partitioner_base_v3 | 
|---|
| 81 | #define task fake_task | 
|---|
| 82 | #define __TBB_STATIC_THRESHOLD 0 | 
|---|
| 83 | #include "tbb/partitioner.h" | 
|---|
| 84 | #undef __TBB_STATIC_THRESHOLD | 
|---|
| 85 | #undef task | 
|---|
| 86 | #undef affinity_partitioner_base_v3 | 
|---|
| 87 | #undef get_initial_auto_partitioner_divisor | 
|---|
| 88 |  | 
|---|
| 89 | // replace library functions to simulate concurrency | 
|---|
| 90 | namespace tbb { | 
|---|
| 91 | namespace internal { | 
|---|
| 92 | size_t my_get_initial_auto_partitioner_divisor() { | 
|---|
| 93 | const size_t X_FACTOR = 4; | 
|---|
| 94 | return X_FACTOR * g_threadNums.local(); | 
|---|
| 95 | } | 
|---|
| 96 |  | 
|---|
| 97 | void* __TBB_EXPORTED_FUNC NFS_Allocate( size_t n_element, size_t element_size, void* hint ); | 
|---|
| 98 | void __TBB_EXPORTED_FUNC NFS_Free( void* ); | 
|---|
| 99 |  | 
|---|
| 100 | void my_affinity_partitioner_base_v3::resize( unsigned factor ) { | 
|---|
| 101 | // Check factor to avoid asking for number of workers while there might be no arena. | 
|---|
| 102 | size_t new_size = factor ? factor * g_threadNums.local() : 0; | 
|---|
| 103 | if (new_size != my_size) { | 
|---|
| 104 | if (my_array) { | 
|---|
| 105 | NFS_Free(my_array); | 
|---|
| 106 | // Following two assignments must be done here for sake of exception safety. | 
|---|
| 107 | my_array = NULL; | 
|---|
| 108 | my_size = 0; | 
|---|
| 109 | } | 
|---|
| 110 | if (new_size) { | 
|---|
| 111 | my_array = static_cast<affinity_id*>(NFS_Allocate(new_size, sizeof(affinity_id), NULL )); | 
|---|
| 112 | memset(my_array, 0, sizeof(affinity_id) * new_size); | 
|---|
| 113 | my_size = new_size; | 
|---|
| 114 | } | 
|---|
| 115 | } | 
|---|
| 116 | } | 
|---|
| 117 |  | 
|---|
| 118 | } //namespace internal | 
|---|
| 119 | // simulate a subset of parallel_for | 
|---|
| 120 | namespace interface9 { | 
|---|
| 121 | namespace internal { | 
|---|
| 122 |  | 
|---|
| 123 | // parallel_for algorithm that executes sequentially | 
|---|
| 124 | template<typename Range, typename Body, typename Partitioner> | 
|---|
| 125 | class start_for : public fake_task { | 
|---|
| 126 | Range my_range; | 
|---|
| 127 | Body my_body; | 
|---|
| 128 | typename Partitioner::task_partition_type my_partition; | 
|---|
| 129 | size_t m_executedBegin, m_executedEnd; | 
|---|
| 130 | bool m_firstTimeRun; | 
|---|
| 131 | size_t m_joinedBegin, m_joinedEnd; | 
|---|
| 132 | test_partitioner_utils::BinaryTree* m_tree; | 
|---|
| 133 | public: | 
|---|
| 134 | start_for( const Range& range, const Body& body, Partitioner& partitioner, | 
|---|
| 135 | test_partitioner_utils::BinaryTree* tree ) : | 
|---|
| 136 | my_range(range), my_body(body), my_partition(partitioner), | 
|---|
| 137 | m_executedBegin(0), m_executedEnd(0), m_firstTimeRun(true), | 
|---|
| 138 | m_joinedBegin(/* grows left */ range.end()), m_joinedEnd(range.end()), m_tree(tree) | 
|---|
| 139 | { | 
|---|
| 140 | if (m_tree) { | 
|---|
| 141 | m_tree->push_node( test_partitioner_utils::make_node(my_range.begin(), my_range.end(), affinity()) ); | 
|---|
| 142 | } | 
|---|
| 143 | } | 
|---|
| 144 | //! Splitting constructor used to generate children. | 
|---|
| 145 | /** parent_ becomes left child.  Newly constructed object is right child. */ | 
|---|
| 146 | start_for( start_for& parent_, typename Partitioner::split_type& split_obj) : | 
|---|
| 147 | my_range(parent_.my_range, split_obj), | 
|---|
| 148 | my_body(parent_.my_body), | 
|---|
| 149 | my_partition(parent_.my_partition, split_obj), | 
|---|
| 150 | m_executedBegin(0), m_executedEnd(0), m_firstTimeRun(true), | 
|---|
| 151 | m_joinedBegin(/* grows left */ my_range.end()), m_joinedEnd(my_range.end()), | 
|---|
| 152 | m_tree(parent_.m_tree) | 
|---|
| 153 | { | 
|---|
| 154 | set_parent(parent_.parent()); | 
|---|
| 155 | my_partition.set_affinity(*this); | 
|---|
| 156 |  | 
|---|
| 157 | if (m_tree) { | 
|---|
| 158 | // collecting splitting statistics | 
|---|
| 159 | m_tree->push_node( test_partitioner_utils::make_node(my_range.begin(), | 
|---|
| 160 | my_range.end(), | 
|---|
| 161 | affinity()) ); | 
|---|
| 162 | m_tree->push_node( test_partitioner_utils::make_node(parent_.my_range.begin(), | 
|---|
| 163 | parent_.my_range.end(), | 
|---|
| 164 | parent_.affinity()) ); | 
|---|
| 165 | } | 
|---|
| 166 | } | 
|---|
| 167 | //! Construct right child from the given range as response to the demand. | 
|---|
| 168 | /** parent_ remains left child.  Newly constructed object is right child. */ | 
|---|
| 169 | start_for( start_for& parent_, const Range& r, depth_t d ) : | 
|---|
| 170 | my_range(r), | 
|---|
| 171 | my_body(parent_.my_body), | 
|---|
| 172 | my_partition(parent_.my_partition, tbb::split()), | 
|---|
| 173 | m_executedBegin(0), m_executedEnd(0), m_firstTimeRun(true), | 
|---|
| 174 | m_joinedBegin(/* grows left */ r.end()), m_joinedEnd(r.end()), | 
|---|
| 175 | m_tree(parent_.m_tree) | 
|---|
| 176 | { | 
|---|
| 177 | set_parent(parent_.parent()); | 
|---|
| 178 | my_partition.set_affinity(*this); | 
|---|
| 179 | my_partition.align_depth( d ); | 
|---|
| 180 | } | 
|---|
| 181 | fake_task* execute() __TBB_override { | 
|---|
| 182 | my_partition.check_being_stolen( *this ); | 
|---|
| 183 | size_t origBegin = my_range.begin(); | 
|---|
| 184 | size_t origEnd = my_range.end(); | 
|---|
| 185 |  | 
|---|
| 186 | my_partition.execute(*this, my_range); | 
|---|
| 187 |  | 
|---|
| 188 | ASSERT(m_executedEnd == m_joinedBegin, "Non-continuous execution"); | 
|---|
| 189 | m_executedEnd = m_joinedEnd; | 
|---|
| 190 |  | 
|---|
| 191 | ASSERT(origBegin == m_executedBegin && origEnd == m_executedEnd, | 
|---|
| 192 | "Not all iterations were processed"); | 
|---|
| 193 | return NULL; | 
|---|
| 194 | } | 
|---|
| 195 | //! Run body for range, serves as callback for partitioner | 
|---|
| 196 | void run_body( Range &r ) { | 
|---|
| 197 | if( r.is_ensure_non_emptiness() ) | 
|---|
| 198 | ASSERT( !r.empty(), "Empty ranges are not allowed"); | 
|---|
| 199 | my_body(r); | 
|---|
| 200 | if (m_firstTimeRun) { | 
|---|
| 201 | m_firstTimeRun = false; | 
|---|
| 202 | m_executedBegin = m_executedEnd = r.begin(); | 
|---|
| 203 | } | 
|---|
| 204 | ASSERT(m_executedBegin <= r.begin() && m_executedEnd <= r.end(), | 
|---|
| 205 | "Non-continuous execution"); | 
|---|
| 206 | m_executedEnd = r.end(); | 
|---|
| 207 | } | 
|---|
| 208 | //! spawn right task, serves as callback for partitioner | 
|---|
| 209 | void offer_work(typename Partitioner::split_type& split_obj) { | 
|---|
| 210 | start_for sibling(*this, split_obj); | 
|---|
| 211 | sibling.execute(); | 
|---|
| 212 | join(sibling.m_executedBegin, sibling.m_executedEnd); | 
|---|
| 213 | } | 
|---|
| 214 | //! spawn right task, serves as callback for partitioner | 
|---|
| 215 | void offer_work(const Range& r, depth_t d = 0) { | 
|---|
| 216 | start_for sibling(*this, r, d); | 
|---|
| 217 | sibling.execute(); | 
|---|
| 218 | join(sibling.m_executedBegin, sibling.m_executedEnd); | 
|---|
| 219 | } | 
|---|
| 220 | void join(size_t siblingExecutedBegin, size_t siblingExecutedEnd) { | 
|---|
| 221 | ASSERT(siblingExecutedEnd == m_joinedBegin, "?"); | 
|---|
| 222 | m_joinedBegin = siblingExecutedBegin; | 
|---|
| 223 | } | 
|---|
| 224 | }; | 
|---|
| 225 |  | 
|---|
| 226 | } //namespace internal | 
|---|
| 227 | } //namespace interfaceX | 
|---|
| 228 | } //namespace tbb | 
|---|
| 229 |  | 
|---|
| 230 | namespace whitebox_simulation { | 
|---|
| 231 | using namespace tbb::interface9::internal; | 
|---|
| 232 | template<typename Range, typename Body, typename Partitioner> | 
|---|
| 233 | void parallel_for( const Range& range, const Body& body, Partitioner& partitioner, | 
|---|
| 234 | test_partitioner_utils::BinaryTree* tree = NULL) { | 
|---|
| 235 | if (!range.empty()) { | 
|---|
| 236 | flag_task parent; | 
|---|
| 237 | start_for<Range, Body, Partitioner> start(range, body, partitioner, tree); | 
|---|
| 238 | start.set_parent(&parent); | 
|---|
| 239 | start.execute(); | 
|---|
| 240 | } | 
|---|
| 241 | } | 
|---|
| 242 |  | 
|---|
| 243 | } //namespace whitebox_simulation | 
|---|
| 244 |  | 
|---|
| 245 | template <typename Range, typename Body, typename Partitioner> | 
|---|
| 246 | void test_case(Range& range, const Body& body, Partitioner& partitioner, | 
|---|
| 247 | test_partitioner_utils::BinaryTree* tree = NULL) { | 
|---|
| 248 | whitebox_simulation::parallel_for(range, body, partitioner, tree); | 
|---|
| 249 | } | 
|---|
| 250 |  | 
|---|
| 251 | // Functions generate size for range objects used in tests | 
|---|
| 252 | template <typename T> | 
|---|
| 253 | size_t default_range_size_generator(T* factor, unsigned index, size_t thread_num) { | 
|---|
| 254 | return size_t(factor[index] * thread_num); | 
|---|
| 255 | } | 
|---|
| 256 |  | 
|---|
| 257 | size_t shifted_left_range_size_generator(size_t* factor, unsigned index, size_t thread_num) { | 
|---|
| 258 | return factor[index] * thread_num - 1; | 
|---|
| 259 | } | 
|---|
| 260 |  | 
|---|
| 261 | size_t shifted_right_range_size_generator(size_t* factor, unsigned index, size_t thread_num) { | 
|---|
| 262 | return factor[index] * thread_num + 1; | 
|---|
| 263 | } | 
|---|
| 264 |  | 
|---|
| 265 | size_t max_range_size_generator(size_t*, unsigned, size_t) { | 
|---|
| 266 | return size_t(-1); | 
|---|
| 267 | } | 
|---|
| 268 |  | 
|---|
| 269 | size_t simple_size_generator(size_t*, unsigned index, size_t) { | 
|---|
| 270 | return index; | 
|---|
| 271 | } | 
|---|
| 272 |  | 
|---|
| 273 | namespace uniform_iterations_distribution { | 
|---|
| 274 |  | 
|---|
| 275 | /* | 
|---|
| 276 | * Test checks uniform distribution of range's iterations among all tasks just after | 
|---|
| 277 | * work distribution phase has been completed and just before work balancing phase has been started | 
|---|
| 278 | */ | 
|---|
| 279 |  | 
|---|
| 280 | using namespace test_partitioner_utils; | 
|---|
| 281 |  | 
|---|
| 282 | class ParallelTestBody { | 
|---|
| 283 | public: | 
|---|
| 284 | struct use_case_settings_t; | 
|---|
| 285 |  | 
|---|
| 286 | typedef void (*CheckerFuncType)(const char*, size_t, const use_case_settings_t*, const RangeStatisticData&); | 
|---|
| 287 |  | 
|---|
| 288 | struct use_case_settings_t { | 
|---|
| 289 | size_t thread_num;                         // number of threads used during current use case | 
|---|
| 290 | unsigned factors_array_len;                // size of 'factors' array | 
|---|
| 291 | size_t range_begin;                        // beginning of range iterations | 
|---|
| 292 | bool provide_feedback;                     // 'true' if range should give feedback | 
|---|
| 293 | bool ensure_non_empty_size;                // don't allow empty size ranges | 
|---|
| 294 |  | 
|---|
| 295 | size_t above_threads_size_tolerance;       // allowed value for number of created ranges | 
|---|
| 296 | // when initial size of the range was greater or | 
|---|
| 297 | // equal to number of threads | 
|---|
| 298 |  | 
|---|
| 299 | size_t below_threads_size_tolerance;       // allowed value for number of created ranges | 
|---|
| 300 | // when initial size of the range was less than | 
|---|
| 301 | // number of threads | 
|---|
| 302 |  | 
|---|
| 303 | size_t between_min_max_ranges_tolerance;   // allowed value for difference of iterations | 
|---|
| 304 | // between bigger and lesser ranges | 
|---|
| 305 |  | 
|---|
| 306 | CheckerFuncType checker;                   // checker function for a particular test case | 
|---|
| 307 | }; | 
|---|
| 308 |  | 
|---|
| 309 | ParallelTestBody(size_t parallel_group_thread_starting_index) | 
|---|
| 310 | : m_parallel_group_thread_starting_index(parallel_group_thread_starting_index) { } | 
|---|
| 311 |  | 
|---|
| 312 | void operator()(size_t) const { ASSERT( false, "Empty ParallelTestBody called"); } | 
|---|
| 313 |  | 
|---|
| 314 | static void uniform_distribution_checker(const char* rangeName, size_t rangeSize, const use_case_settings_t* settings, | 
|---|
| 315 | const RangeStatisticData& stat) | 
|---|
| 316 | { | 
|---|
| 317 | // Checking that all threads were given a task | 
|---|
| 318 | if (rangeSize >= settings->thread_num) { | 
|---|
| 319 | uint64_t disparity = | 
|---|
| 320 | max(stat.m_rangeNum, settings->thread_num) - min(stat.m_rangeNum, settings->thread_num); | 
|---|
| 321 | if (disparity > settings->above_threads_size_tolerance) { | 
|---|
| 322 | REPORT( "ERROR: '%s (f=%d|e=%d)': |#ranges(%llu)-#threads(%llu)|=%llu > %llu=tolerance\n", | 
|---|
| 323 | rangeName, int(settings->provide_feedback), int(settings->ensure_non_empty_size), stat.m_rangeNum, | 
|---|
| 324 | settings->thread_num, disparity, uint64_t(settings->above_threads_size_tolerance)); | 
|---|
| 325 | ASSERT(disparity <= settings->above_threads_size_tolerance, "Incorrect number of range " | 
|---|
| 326 | "objects was created before work balancing phase started"); | 
|---|
| 327 | } | 
|---|
| 328 | } else if (settings->ensure_non_empty_size && rangeSize != 0) { | 
|---|
| 329 | uint64_t disparity = max(stat.m_rangeNum, rangeSize) - min(stat.m_rangeNum, rangeSize); | 
|---|
| 330 | if (disparity > settings->below_threads_size_tolerance ) { | 
|---|
| 331 | REPORT( "ERROR: '%s (f=%d|e=%d)': |#ranges-range size|=%llu > %llu=tolerance\n", | 
|---|
| 332 | rangeName, int(settings->provide_feedback), int(settings->ensure_non_empty_size), | 
|---|
| 333 | disparity, uint64_t(settings->below_threads_size_tolerance)); | 
|---|
| 334 | ASSERT(disparity <= settings->below_threads_size_tolerance, "Incorrect number of range objects" | 
|---|
| 335 | " was created before work balancing phase started"); | 
|---|
| 336 | } | 
|---|
| 337 | } | 
|---|
| 338 | // Checking difference between min and max number of range iterations | 
|---|
| 339 | size_t diff = stat.m_maxRangeSize - stat.m_minRangeSize; | 
|---|
| 340 | if (diff > settings->between_min_max_ranges_tolerance) { | 
|---|
| 341 | REPORT( "ERROR: '%s (f=%d|e=%d)': range size difference=%llu > %llu=tolerance\n", | 
|---|
| 342 | rangeName, int(settings->provide_feedback), int(settings->ensure_non_empty_size), | 
|---|
| 343 | uint64_t(diff), uint64_t(settings->between_min_max_ranges_tolerance)); | 
|---|
| 344 | ASSERT(diff <= settings->between_min_max_ranges_tolerance, "Uniform iteration distribution error"); | 
|---|
| 345 | } | 
|---|
| 346 | } | 
|---|
| 347 | // Checker for test cases where ranges don't provide feedback during proportional split to | 
|---|
| 348 | // partitioner and differ from tbb::blocked_range implementation in their splitting algorithm | 
|---|
| 349 | static void nonuniform_distribution_checker(const char* rangeName, size_t rangeSize, const use_case_settings_t* settings, | 
|---|
| 350 | const RangeStatisticData& stat) | 
|---|
| 351 | { | 
|---|
| 352 | if (stat.m_rangeNum > settings->thread_num) { | 
|---|
| 353 | REPORT( "ERROR: '%s (f=%d|e=%d)': %llu=#ranges > #threads=%llu\n", | 
|---|
| 354 | rangeName, int(settings->provide_feedback), int(settings->ensure_non_empty_size), | 
|---|
| 355 | uint64_t(stat.m_rangeNum), uint64_t(settings->thread_num)); | 
|---|
| 356 | ASSERT(stat.m_rangeNum <= settings->thread_num, | 
|---|
| 357 | "Incorrect number of range objects was created before work balancing phase started"); | 
|---|
| 358 | } | 
|---|
| 359 | // Checking difference between min and max number of range iterations | 
|---|
| 360 | size_t diff = stat.m_maxRangeSize - stat.m_minRangeSize; | 
|---|
| 361 | if (diff > rangeSize) { | 
|---|
| 362 | REPORT( "ERROR: '%s (f=%d|e=%d)': range size difference=%llu > %llu=initial range size\n", | 
|---|
| 363 | rangeName, int(settings->provide_feedback), int(settings->ensure_non_empty_size), | 
|---|
| 364 | uint64_t(diff), uint64_t(rangeSize)); | 
|---|
| 365 | ASSERT(diff <= rangeSize, "Iteration distribution error"); | 
|---|
| 366 | } | 
|---|
| 367 | } | 
|---|
| 368 |  | 
|---|
| 369 | protected: | 
|---|
| 370 | size_t m_parallel_group_thread_starting_index; // starting index of thread | 
|---|
| 371 |  | 
|---|
| 372 | template <typename Range, typename Partitioner, typename T> | 
|---|
| 373 | void test(use_case_settings_t& settings, T factors[], size_t (*rsgFunc)(T*, unsigned, size_t) | 
|---|
| 374 | = &default_range_size_generator<T>) const | 
|---|
| 375 | { | 
|---|
| 376 | for (unsigned i = 0; i < settings.factors_array_len; ++i) { | 
|---|
| 377 | size_t range_end = rsgFunc(factors, i, settings.thread_num); | 
|---|
| 378 | RangeStatisticData stat = { /*range num=*/ 0, /*minimal size of range=*/ 0, | 
|---|
| 379 | /*maximal size of range=*/ 0, /*minimal size of range was not rewritten yet=*/ false }; | 
|---|
| 380 | Range range = Range(settings.range_begin, range_end, &stat, settings.provide_feedback, | 
|---|
| 381 | settings.ensure_non_empty_size); | 
|---|
| 382 | Partitioner my_partitioner; | 
|---|
| 383 | test_case(range, SimpleBody(), my_partitioner, NULL); | 
|---|
| 384 | size_t range_size = range_end - settings.range_begin; | 
|---|
| 385 | const char* rangeName = typeid(range).name(); | 
|---|
| 386 | settings.checker(rangeName, range_size, &settings, stat); | 
|---|
| 387 | } | 
|---|
| 388 | } | 
|---|
| 389 | }; | 
|---|
| 390 |  | 
|---|
| 391 | template <typename ParallelTestBody> | 
|---|
| 392 | void test() { | 
|---|
| 393 | size_t hw_threads_num = tbb::tbb_thread::hardware_concurrency(); | 
|---|
| 394 | size_t threadsToRunOn = std::min<size_t>(max_simulated_threads, hw_threads_num); | 
|---|
| 395 |  | 
|---|
| 396 | size_t parallel_group_thread_starting_index = 1; | 
|---|
| 397 | while( parallel_group_thread_starting_index <= max_simulated_threads - threadsToRunOn ) { | 
|---|
| 398 | NativeParallelFor(threadsToRunOn, ParallelTestBody(parallel_group_thread_starting_index)); | 
|---|
| 399 | parallel_group_thread_starting_index += threadsToRunOn; | 
|---|
| 400 | } | 
|---|
| 401 | NativeParallelFor(max_simulated_threads - parallel_group_thread_starting_index, | 
|---|
| 402 | ParallelTestBody(parallel_group_thread_starting_index)); | 
|---|
| 403 | } | 
|---|
| 404 |  | 
|---|
| 405 | namespace task_affinity_whitebox { | 
|---|
| 406 | size_t range_begin = 0; | 
|---|
| 407 | size_t range_end = 20; | 
|---|
| 408 | } | 
|---|
| 409 |  | 
|---|
| 410 | template<typename Partitioner> | 
|---|
| 411 | void check_tree(const test_partitioner_utils::BinaryTree&); | 
|---|
| 412 |  | 
|---|
| 413 | template<> | 
|---|
| 414 | void check_tree<tbb::affinity_partitioner>(const test_partitioner_utils::BinaryTree& tree) { | 
|---|
| 415 | ASSERT(tree == whitebox_simulation::reference_tree, | 
|---|
| 416 | "affinity_partitioner distributes tasks differently from run to run"); | 
|---|
| 417 | } | 
|---|
| 418 |  | 
|---|
| 419 | template<> | 
|---|
| 420 | void check_tree<tbb::static_partitioner>(const test_partitioner_utils::BinaryTree& tree) { | 
|---|
| 421 | std::vector<test_partitioner_utils::TreeNode* > tree_leafs; | 
|---|
| 422 | tree.fill_leafs(tree_leafs); | 
|---|
| 423 | typedef std::vector<size_t> Slots; | 
|---|
| 424 | Slots affinity_slots(tree_leafs.size() + 1, 0); | 
|---|
| 425 |  | 
|---|
| 426 | for (std::vector<test_partitioner_utils::TreeNode*>::iterator i = tree_leafs.begin(); i != tree_leafs.end(); ++i) { | 
|---|
| 427 | affinity_slots[(*i)->m_affinity]++; | 
|---|
| 428 | if ((*i)->m_affinity == 0) | 
|---|
| 429 | ASSERT((*i)->m_range_begin == task_affinity_whitebox::range_begin, | 
|---|
| 430 | "Task with affinity 0 was executed with wrong range"); | 
|---|
| 431 | } | 
|---|
| 432 |  | 
|---|
| 433 | typedef std::iterator_traits<Slots::iterator>::difference_type slots_difference_type; | 
|---|
| 434 | ASSERT(std::count(affinity_slots.begin(), affinity_slots.end(), size_t(0)) == slots_difference_type(1), | 
|---|
| 435 | "static_partitioner incorrectly distributed tasks by threads"); | 
|---|
| 436 | ASSERT(std::count(affinity_slots.begin(), affinity_slots.end(), size_t(1)) == slots_difference_type(g_threadNums.local()), | 
|---|
| 437 | "static_partitioner incorrectly distributed tasks by threads"); | 
|---|
| 438 | ASSERT(affinity_slots[tbb::this_task_arena::current_thread_index() + 1] == 0, | 
|---|
| 439 | "static_partitioner incorrectly assigns task with 0 affinity"); | 
|---|
| 440 | ASSERT(std::accumulate(affinity_slots.begin(), affinity_slots.end(), size_t(0)) == g_threadNums.local(), | 
|---|
| 441 | "static_partitioner has created more tasks than the number of threads"); | 
|---|
| 442 | } | 
|---|
| 443 |  | 
|---|
| 444 | template<typename Partitioner> | 
|---|
| 445 | void test_task_affinity() { | 
|---|
| 446 | using namespace task_affinity_whitebox; | 
|---|
| 447 | test_partitioner_utils::SimpleBody body; | 
|---|
| 448 | for (size_t p = 1; p <= 50; ++p) { | 
|---|
| 449 | g_threadNums.local() = p; | 
|---|
| 450 | whitebox_simulation::whitebox_thread_index = 0; | 
|---|
| 451 | test_partitioner_utils::TestRanges::BlockedRange range(range_begin, range_end, /*statData*/NULL, | 
|---|
| 452 | /*provide_feedback*/false, /*ensure_non_empty_size*/false); | 
|---|
| 453 | Partitioner partitioner; | 
|---|
| 454 | whitebox_simulation::reference_tree = test_partitioner_utils::BinaryTree(); | 
|---|
| 455 | whitebox_simulation::parallel_for(range, body, partitioner, &(whitebox_simulation::reference_tree)); | 
|---|
| 456 | while (whitebox_simulation::whitebox_thread_index < p) { | 
|---|
| 457 | test_partitioner_utils::BinaryTree tree; | 
|---|
| 458 | whitebox_simulation::parallel_for(range, body, partitioner, &tree); | 
|---|
| 459 | check_tree<Partitioner>(tree); | 
|---|
| 460 | whitebox_simulation::whitebox_thread_index++; | 
|---|
| 461 | } | 
|---|
| 462 | range_begin++; | 
|---|
| 463 | range_end += 2; | 
|---|
| 464 | } | 
|---|
| 465 | } | 
|---|
| 466 |  | 
|---|
| 467 | } /* namespace uniform_iterations_distribution */ | 
|---|
| 468 |  | 
|---|