1 | // The template and inlines for the -*- C++ -*- internal _Array helper class. |
2 | |
3 | // Copyright (C) 1997-2018 Free Software Foundation, Inc. |
4 | // |
5 | // This file is part of the GNU ISO C++ Library. This library is free |
6 | // software; you can redistribute it and/or modify it under the |
7 | // terms of the GNU General Public License as published by the |
8 | // Free Software Foundation; either version 3, or (at your option) |
9 | // any later version. |
10 | |
11 | // This library is distributed in the hope that it will be useful, |
12 | // but WITHOUT ANY WARRANTY; without even the implied warranty of |
13 | // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
14 | // GNU General Public License for more details. |
15 | |
16 | // Under Section 7 of GPL version 3, you are granted additional |
17 | // permissions described in the GCC Runtime Library Exception, version |
18 | // 3.1, as published by the Free Software Foundation. |
19 | |
20 | // You should have received a copy of the GNU General Public License and |
21 | // a copy of the GCC Runtime Library Exception along with this program; |
22 | // see the files COPYING3 and COPYING.RUNTIME respectively. If not, see |
23 | // <http://www.gnu.org/licenses/>. |
24 | |
25 | /** @file bits/valarray_array.h |
26 | * This is an internal header file, included by other library headers. |
27 | * Do not attempt to use it directly. @headername{valarray} |
28 | */ |
29 | |
30 | // Written by Gabriel Dos Reis <Gabriel.Dos-Reis@DPTMaths.ENS-Cachan.Fr> |
31 | |
32 | #ifndef _VALARRAY_ARRAY_H |
33 | #define _VALARRAY_ARRAY_H 1 |
34 | |
35 | #pragma GCC system_header |
36 | |
37 | #include <bits/c++config.h> |
38 | #include <bits/cpp_type_traits.h> |
39 | #include <cstdlib> |
40 | #include <new> |
41 | |
42 | namespace std _GLIBCXX_VISIBILITY(default) |
43 | { |
44 | _GLIBCXX_BEGIN_NAMESPACE_VERSION |
45 | |
46 | // |
47 | // Helper functions on raw pointers |
48 | // |
49 | |
50 | // We get memory by the old fashion way |
51 | inline void* |
52 | __valarray_get_memory(size_t __n) |
53 | { return operator new(__n); } |
54 | |
55 | template<typename _Tp> |
56 | inline _Tp*__restrict__ |
57 | __valarray_get_storage(size_t __n) |
58 | { |
59 | return static_cast<_Tp*__restrict__> |
60 | (std::__valarray_get_memory(__n * sizeof(_Tp))); |
61 | } |
62 | |
63 | // Return memory to the system |
64 | inline void |
65 | __valarray_release_memory(void* __p) |
66 | { operator delete(__p); } |
67 | |
68 | // Turn a raw-memory into an array of _Tp filled with _Tp() |
69 | // This is required in 'valarray<T> v(n);' |
70 | template<typename _Tp, bool> |
71 | struct _Array_default_ctor |
72 | { |
73 | // Please note that this isn't exception safe. But |
74 | // valarrays aren't required to be exception safe. |
75 | inline static void |
76 | _S_do_it(_Tp* __b, _Tp* __e) |
77 | { |
78 | while (__b != __e) |
79 | new(__b++) _Tp(); |
80 | } |
81 | }; |
82 | |
83 | template<typename _Tp> |
84 | struct _Array_default_ctor<_Tp, true> |
85 | { |
86 | // For fundamental types, it suffices to say 'memset()' |
87 | inline static void |
88 | _S_do_it(_Tp* __b, _Tp* __e) |
89 | { __builtin_memset(__b, 0, (__e - __b) * sizeof(_Tp)); } |
90 | }; |
91 | |
92 | template<typename _Tp> |
93 | inline void |
94 | __valarray_default_construct(_Tp* __b, _Tp* __e) |
95 | { |
96 | _Array_default_ctor<_Tp, __is_scalar<_Tp>::__value>::_S_do_it(__b, __e); |
97 | } |
98 | |
99 | // Turn a raw-memory into an array of _Tp filled with __t |
100 | // This is the required in valarray<T> v(n, t). Also |
101 | // used in valarray<>::resize(). |
102 | template<typename _Tp, bool> |
103 | struct _Array_init_ctor |
104 | { |
105 | // Please note that this isn't exception safe. But |
106 | // valarrays aren't required to be exception safe. |
107 | inline static void |
108 | _S_do_it(_Tp* __b, _Tp* __e, const _Tp __t) |
109 | { |
110 | while (__b != __e) |
111 | new(__b++) _Tp(__t); |
112 | } |
113 | }; |
114 | |
115 | template<typename _Tp> |
116 | struct _Array_init_ctor<_Tp, true> |
117 | { |
118 | inline static void |
119 | _S_do_it(_Tp* __b, _Tp* __e, const _Tp __t) |
120 | { |
121 | while (__b != __e) |
122 | *__b++ = __t; |
123 | } |
124 | }; |
125 | |
126 | template<typename _Tp> |
127 | inline void |
128 | __valarray_fill_construct(_Tp* __b, _Tp* __e, const _Tp __t) |
129 | { |
130 | _Array_init_ctor<_Tp, __is_trivial(_Tp)>::_S_do_it(__b, __e, __t); |
131 | } |
132 | |
133 | // |
134 | // copy-construct raw array [__o, *) from plain array [__b, __e) |
135 | // We can't just say 'memcpy()' |
136 | // |
137 | template<typename _Tp, bool> |
138 | struct _Array_copy_ctor |
139 | { |
140 | // Please note that this isn't exception safe. But |
141 | // valarrays aren't required to be exception safe. |
142 | inline static void |
143 | _S_do_it(const _Tp* __b, const _Tp* __e, _Tp* __restrict__ __o) |
144 | { |
145 | while (__b != __e) |
146 | new(__o++) _Tp(*__b++); |
147 | } |
148 | }; |
149 | |
150 | template<typename _Tp> |
151 | struct _Array_copy_ctor<_Tp, true> |
152 | { |
153 | inline static void |
154 | _S_do_it(const _Tp* __b, const _Tp* __e, _Tp* __restrict__ __o) |
155 | { |
156 | if (__b) |
157 | __builtin_memcpy(__o, __b, (__e - __b) * sizeof(_Tp)); |
158 | } |
159 | }; |
160 | |
161 | template<typename _Tp> |
162 | inline void |
163 | __valarray_copy_construct(const _Tp* __b, const _Tp* __e, |
164 | _Tp* __restrict__ __o) |
165 | { |
166 | _Array_copy_ctor<_Tp, __is_trivial(_Tp)>::_S_do_it(__b, __e, __o); |
167 | } |
168 | |
169 | // copy-construct raw array [__o, *) from strided array __a[<__n : __s>] |
170 | template<typename _Tp> |
171 | inline void |
172 | __valarray_copy_construct (const _Tp* __restrict__ __a, size_t __n, |
173 | size_t __s, _Tp* __restrict__ __o) |
174 | { |
175 | if (__is_trivial(_Tp)) |
176 | while (__n--) |
177 | { |
178 | *__o++ = *__a; |
179 | __a += __s; |
180 | } |
181 | else |
182 | while (__n--) |
183 | { |
184 | new(__o++) _Tp(*__a); |
185 | __a += __s; |
186 | } |
187 | } |
188 | |
189 | // copy-construct raw array [__o, *) from indexed array __a[__i[<__n>]] |
190 | template<typename _Tp> |
191 | inline void |
192 | __valarray_copy_construct (const _Tp* __restrict__ __a, |
193 | const size_t* __restrict__ __i, |
194 | _Tp* __restrict__ __o, size_t __n) |
195 | { |
196 | if (__is_trivial(_Tp)) |
197 | while (__n--) |
198 | *__o++ = __a[*__i++]; |
199 | else |
200 | while (__n--) |
201 | new (__o++) _Tp(__a[*__i++]); |
202 | } |
203 | |
204 | // Do the necessary cleanup when we're done with arrays. |
205 | template<typename _Tp> |
206 | inline void |
207 | __valarray_destroy_elements(_Tp* __b, _Tp* __e) |
208 | { |
209 | if (!__is_trivial(_Tp)) |
210 | while (__b != __e) |
211 | { |
212 | __b->~_Tp(); |
213 | ++__b; |
214 | } |
215 | } |
216 | |
217 | // Fill a plain array __a[<__n>] with __t |
218 | template<typename _Tp> |
219 | inline void |
220 | __valarray_fill(_Tp* __restrict__ __a, size_t __n, const _Tp& __t) |
221 | { |
222 | while (__n--) |
223 | *__a++ = __t; |
224 | } |
225 | |
226 | // fill strided array __a[<__n-1 : __s>] with __t |
227 | template<typename _Tp> |
228 | inline void |
229 | __valarray_fill(_Tp* __restrict__ __a, size_t __n, |
230 | size_t __s, const _Tp& __t) |
231 | { |
232 | for (size_t __i = 0; __i < __n; ++__i, __a += __s) |
233 | *__a = __t; |
234 | } |
235 | |
236 | // fill indirect array __a[__i[<__n>]] with __i |
237 | template<typename _Tp> |
238 | inline void |
239 | __valarray_fill(_Tp* __restrict__ __a, const size_t* __restrict__ __i, |
240 | size_t __n, const _Tp& __t) |
241 | { |
242 | for (size_t __j = 0; __j < __n; ++__j, ++__i) |
243 | __a[*__i] = __t; |
244 | } |
245 | |
246 | // copy plain array __a[<__n>] in __b[<__n>] |
247 | // For non-fundamental types, it is wrong to say 'memcpy()' |
248 | template<typename _Tp, bool> |
249 | struct _Array_copier |
250 | { |
251 | inline static void |
252 | _S_do_it(const _Tp* __restrict__ __a, size_t __n, _Tp* __restrict__ __b) |
253 | { |
254 | while(__n--) |
255 | *__b++ = *__a++; |
256 | } |
257 | }; |
258 | |
259 | template<typename _Tp> |
260 | struct _Array_copier<_Tp, true> |
261 | { |
262 | inline static void |
263 | _S_do_it(const _Tp* __restrict__ __a, size_t __n, _Tp* __restrict__ __b) |
264 | { |
265 | if (__n != 0) |
266 | __builtin_memcpy(__b, __a, __n * sizeof (_Tp)); |
267 | } |
268 | }; |
269 | |
270 | // Copy a plain array __a[<__n>] into a play array __b[<>] |
271 | template<typename _Tp> |
272 | inline void |
273 | __valarray_copy(const _Tp* __restrict__ __a, size_t __n, |
274 | _Tp* __restrict__ __b) |
275 | { |
276 | _Array_copier<_Tp, __is_trivial(_Tp)>::_S_do_it(__a, __n, __b); |
277 | } |
278 | |
279 | // Copy strided array __a[<__n : __s>] in plain __b[<__n>] |
280 | template<typename _Tp> |
281 | inline void |
282 | __valarray_copy(const _Tp* __restrict__ __a, size_t __n, size_t __s, |
283 | _Tp* __restrict__ __b) |
284 | { |
285 | for (size_t __i = 0; __i < __n; ++__i, ++__b, __a += __s) |
286 | *__b = *__a; |
287 | } |
288 | |
289 | // Copy a plain array __a[<__n>] into a strided array __b[<__n : __s>] |
290 | template<typename _Tp> |
291 | inline void |
292 | __valarray_copy(const _Tp* __restrict__ __a, _Tp* __restrict__ __b, |
293 | size_t __n, size_t __s) |
294 | { |
295 | for (size_t __i = 0; __i < __n; ++__i, ++__a, __b += __s) |
296 | *__b = *__a; |
297 | } |
298 | |
299 | // Copy strided array __src[<__n : __s1>] into another |
300 | // strided array __dst[< : __s2>]. Their sizes must match. |
301 | template<typename _Tp> |
302 | inline void |
303 | __valarray_copy(const _Tp* __restrict__ __src, size_t __n, size_t __s1, |
304 | _Tp* __restrict__ __dst, size_t __s2) |
305 | { |
306 | for (size_t __i = 0; __i < __n; ++__i) |
307 | __dst[__i * __s2] = __src[__i * __s1]; |
308 | } |
309 | |
310 | // Copy an indexed array __a[__i[<__n>]] in plain array __b[<__n>] |
311 | template<typename _Tp> |
312 | inline void |
313 | __valarray_copy(const _Tp* __restrict__ __a, |
314 | const size_t* __restrict__ __i, |
315 | _Tp* __restrict__ __b, size_t __n) |
316 | { |
317 | for (size_t __j = 0; __j < __n; ++__j, ++__b, ++__i) |
318 | *__b = __a[*__i]; |
319 | } |
320 | |
321 | // Copy a plain array __a[<__n>] in an indexed array __b[__i[<__n>]] |
322 | template<typename _Tp> |
323 | inline void |
324 | __valarray_copy(const _Tp* __restrict__ __a, size_t __n, |
325 | _Tp* __restrict__ __b, const size_t* __restrict__ __i) |
326 | { |
327 | for (size_t __j = 0; __j < __n; ++__j, ++__a, ++__i) |
328 | __b[*__i] = *__a; |
329 | } |
330 | |
331 | // Copy the __n first elements of an indexed array __src[<__i>] into |
332 | // another indexed array __dst[<__j>]. |
333 | template<typename _Tp> |
334 | inline void |
335 | __valarray_copy(const _Tp* __restrict__ __src, size_t __n, |
336 | const size_t* __restrict__ __i, |
337 | _Tp* __restrict__ __dst, const size_t* __restrict__ __j) |
338 | { |
339 | for (size_t __k = 0; __k < __n; ++__k) |
340 | __dst[*__j++] = __src[*__i++]; |
341 | } |
342 | |
343 | // |
344 | // Compute the sum of elements in range [__f, __l) which must not be empty. |
345 | // This is a naive algorithm. It suffers from cancelling. |
346 | // In the future try to specialize for _Tp = float, double, long double |
347 | // using a more accurate algorithm. |
348 | // |
349 | template<typename _Tp> |
350 | inline _Tp |
351 | __valarray_sum(const _Tp* __f, const _Tp* __l) |
352 | { |
353 | _Tp __r = *__f++; |
354 | while (__f != __l) |
355 | __r += *__f++; |
356 | return __r; |
357 | } |
358 | |
359 | // Compute the product of all elements in range [__f, __l) |
360 | template<typename _Tp> |
361 | inline _Tp |
362 | __valarray_product(const _Tp* __f, const _Tp* __l) |
363 | { |
364 | _Tp __r = _Tp(1); |
365 | while (__f != __l) |
366 | __r = __r * *__f++; |
367 | return __r; |
368 | } |
369 | |
370 | // Compute the min/max of an array-expression |
371 | template<typename _Ta> |
372 | inline typename _Ta::value_type |
373 | __valarray_min(const _Ta& __a) |
374 | { |
375 | size_t __s = __a.size(); |
376 | typedef typename _Ta::value_type _Value_type; |
377 | _Value_type __r = __s == 0 ? _Value_type() : __a[0]; |
378 | for (size_t __i = 1; __i < __s; ++__i) |
379 | { |
380 | _Value_type __t = __a[__i]; |
381 | if (__t < __r) |
382 | __r = __t; |
383 | } |
384 | return __r; |
385 | } |
386 | |
387 | template<typename _Ta> |
388 | inline typename _Ta::value_type |
389 | __valarray_max(const _Ta& __a) |
390 | { |
391 | size_t __s = __a.size(); |
392 | typedef typename _Ta::value_type _Value_type; |
393 | _Value_type __r = __s == 0 ? _Value_type() : __a[0]; |
394 | for (size_t __i = 1; __i < __s; ++__i) |
395 | { |
396 | _Value_type __t = __a[__i]; |
397 | if (__t > __r) |
398 | __r = __t; |
399 | } |
400 | return __r; |
401 | } |
402 | |
403 | // |
404 | // Helper class _Array, first layer of valarray abstraction. |
405 | // All operations on valarray should be forwarded to this class |
406 | // whenever possible. -- gdr |
407 | // |
408 | |
409 | template<typename _Tp> |
410 | struct _Array |
411 | { |
412 | explicit _Array(size_t); |
413 | explicit _Array(_Tp* const __restrict__); |
414 | explicit _Array(const valarray<_Tp>&); |
415 | _Array(const _Tp* __restrict__, size_t); |
416 | |
417 | _Tp* begin() const; |
418 | |
419 | _Tp* const __restrict__ _M_data; |
420 | }; |
421 | |
422 | |
423 | // Copy-construct plain array __b[<__n>] from indexed array __a[__i[<__n>]] |
424 | template<typename _Tp> |
425 | inline void |
426 | __valarray_copy_construct(_Array<_Tp> __a, _Array<size_t> __i, |
427 | _Array<_Tp> __b, size_t __n) |
428 | { std::__valarray_copy_construct(__a._M_data, __i._M_data, |
429 | __b._M_data, __n); } |
430 | |
431 | // Copy-construct plain array __b[<__n>] from strided array __a[<__n : __s>] |
432 | template<typename _Tp> |
433 | inline void |
434 | __valarray_copy_construct(_Array<_Tp> __a, size_t __n, size_t __s, |
435 | _Array<_Tp> __b) |
436 | { std::__valarray_copy_construct(__a._M_data, __n, __s, __b._M_data); } |
437 | |
438 | template<typename _Tp> |
439 | inline void |
440 | __valarray_fill (_Array<_Tp> __a, size_t __n, const _Tp& __t) |
441 | { std::__valarray_fill(__a._M_data, __n, __t); } |
442 | |
443 | template<typename _Tp> |
444 | inline void |
445 | __valarray_fill(_Array<_Tp> __a, size_t __n, size_t __s, const _Tp& __t) |
446 | { std::__valarray_fill(__a._M_data, __n, __s, __t); } |
447 | |
448 | template<typename _Tp> |
449 | inline void |
450 | __valarray_fill(_Array<_Tp> __a, _Array<size_t> __i, |
451 | size_t __n, const _Tp& __t) |
452 | { std::__valarray_fill(__a._M_data, __i._M_data, __n, __t); } |
453 | |
454 | // Copy a plain array __a[<__n>] into a play array __b[<>] |
455 | template<typename _Tp> |
456 | inline void |
457 | __valarray_copy(_Array<_Tp> __a, size_t __n, _Array<_Tp> __b) |
458 | { std::__valarray_copy(__a._M_data, __n, __b._M_data); } |
459 | |
460 | // Copy strided array __a[<__n : __s>] in plain __b[<__n>] |
461 | template<typename _Tp> |
462 | inline void |
463 | __valarray_copy(_Array<_Tp> __a, size_t __n, size_t __s, _Array<_Tp> __b) |
464 | { std::__valarray_copy(__a._M_data, __n, __s, __b._M_data); } |
465 | |
466 | // Copy a plain array __a[<__n>] into a strided array __b[<__n : __s>] |
467 | template<typename _Tp> |
468 | inline void |
469 | __valarray_copy(_Array<_Tp> __a, _Array<_Tp> __b, size_t __n, size_t __s) |
470 | { __valarray_copy(__a._M_data, __b._M_data, __n, __s); } |
471 | |
472 | // Copy strided array __src[<__n : __s1>] into another |
473 | // strided array __dst[< : __s2>]. Their sizes must match. |
474 | template<typename _Tp> |
475 | inline void |
476 | __valarray_copy(_Array<_Tp> __a, size_t __n, size_t __s1, |
477 | _Array<_Tp> __b, size_t __s2) |
478 | { std::__valarray_copy(__a._M_data, __n, __s1, __b._M_data, __s2); } |
479 | |
480 | // Copy an indexed array __a[__i[<__n>]] in plain array __b[<__n>] |
481 | template<typename _Tp> |
482 | inline void |
483 | __valarray_copy(_Array<_Tp> __a, _Array<size_t> __i, |
484 | _Array<_Tp> __b, size_t __n) |
485 | { std::__valarray_copy(__a._M_data, __i._M_data, __b._M_data, __n); } |
486 | |
487 | // Copy a plain array __a[<__n>] in an indexed array __b[__i[<__n>]] |
488 | template<typename _Tp> |
489 | inline void |
490 | __valarray_copy(_Array<_Tp> __a, size_t __n, _Array<_Tp> __b, |
491 | _Array<size_t> __i) |
492 | { std::__valarray_copy(__a._M_data, __n, __b._M_data, __i._M_data); } |
493 | |
494 | // Copy the __n first elements of an indexed array __src[<__i>] into |
495 | // another indexed array __dst[<__j>]. |
496 | template<typename _Tp> |
497 | inline void |
498 | __valarray_copy(_Array<_Tp> __src, size_t __n, _Array<size_t> __i, |
499 | _Array<_Tp> __dst, _Array<size_t> __j) |
500 | { |
501 | std::__valarray_copy(__src._M_data, __n, __i._M_data, |
502 | __dst._M_data, __j._M_data); |
503 | } |
504 | |
505 | template<typename _Tp> |
506 | inline |
507 | _Array<_Tp>::_Array(size_t __n) |
508 | : _M_data(__valarray_get_storage<_Tp>(__n)) |
509 | { std::__valarray_default_construct(_M_data, _M_data + __n); } |
510 | |
511 | template<typename _Tp> |
512 | inline |
513 | _Array<_Tp>::_Array(_Tp* const __restrict__ __p) |
514 | : _M_data (__p) {} |
515 | |
516 | template<typename _Tp> |
517 | inline |
518 | _Array<_Tp>::_Array(const valarray<_Tp>& __v) |
519 | : _M_data (__v._M_data) {} |
520 | |
521 | template<typename _Tp> |
522 | inline |
523 | _Array<_Tp>::_Array(const _Tp* __restrict__ __b, size_t __s) |
524 | : _M_data(__valarray_get_storage<_Tp>(__s)) |
525 | { std::__valarray_copy_construct(__b, __s, _M_data); } |
526 | |
527 | template<typename _Tp> |
528 | inline _Tp* |
529 | _Array<_Tp>::begin () const |
530 | { return _M_data; } |
531 | |
532 | #define _DEFINE_ARRAY_FUNCTION(_Op, _Name) \ |
533 | template<typename _Tp> \ |
534 | inline void \ |
535 | _Array_augmented_##_Name(_Array<_Tp> __a, size_t __n, const _Tp& __t) \ |
536 | { \ |
537 | for (_Tp* __p = __a._M_data; __p < __a._M_data + __n; ++__p) \ |
538 | *__p _Op##= __t; \ |
539 | } \ |
540 | \ |
541 | template<typename _Tp> \ |
542 | inline void \ |
543 | _Array_augmented_##_Name(_Array<_Tp> __a, size_t __n, _Array<_Tp> __b) \ |
544 | { \ |
545 | _Tp* __p = __a._M_data; \ |
546 | for (_Tp* __q = __b._M_data; __q < __b._M_data + __n; ++__p, ++__q) \ |
547 | *__p _Op##= *__q; \ |
548 | } \ |
549 | \ |
550 | template<typename _Tp, class _Dom> \ |
551 | void \ |
552 | _Array_augmented_##_Name(_Array<_Tp> __a, \ |
553 | const _Expr<_Dom, _Tp>& __e, size_t __n) \ |
554 | { \ |
555 | _Tp* __p(__a._M_data); \ |
556 | for (size_t __i = 0; __i < __n; ++__i, ++__p) \ |
557 | *__p _Op##= __e[__i]; \ |
558 | } \ |
559 | \ |
560 | template<typename _Tp> \ |
561 | inline void \ |
562 | _Array_augmented_##_Name(_Array<_Tp> __a, size_t __n, size_t __s, \ |
563 | _Array<_Tp> __b) \ |
564 | { \ |
565 | _Tp* __q(__b._M_data); \ |
566 | for (_Tp* __p = __a._M_data; __p < __a._M_data + __s * __n; \ |
567 | __p += __s, ++__q) \ |
568 | *__p _Op##= *__q; \ |
569 | } \ |
570 | \ |
571 | template<typename _Tp> \ |
572 | inline void \ |
573 | _Array_augmented_##_Name(_Array<_Tp> __a, _Array<_Tp> __b, \ |
574 | size_t __n, size_t __s) \ |
575 | { \ |
576 | _Tp* __q(__b._M_data); \ |
577 | for (_Tp* __p = __a._M_data; __p < __a._M_data + __n; \ |
578 | ++__p, __q += __s) \ |
579 | *__p _Op##= *__q; \ |
580 | } \ |
581 | \ |
582 | template<typename _Tp, class _Dom> \ |
583 | void \ |
584 | _Array_augmented_##_Name(_Array<_Tp> __a, size_t __s, \ |
585 | const _Expr<_Dom, _Tp>& __e, size_t __n) \ |
586 | { \ |
587 | _Tp* __p(__a._M_data); \ |
588 | for (size_t __i = 0; __i < __n; ++__i, __p += __s) \ |
589 | *__p _Op##= __e[__i]; \ |
590 | } \ |
591 | \ |
592 | template<typename _Tp> \ |
593 | inline void \ |
594 | _Array_augmented_##_Name(_Array<_Tp> __a, _Array<size_t> __i, \ |
595 | _Array<_Tp> __b, size_t __n) \ |
596 | { \ |
597 | _Tp* __q(__b._M_data); \ |
598 | for (size_t* __j = __i._M_data; __j < __i._M_data + __n; \ |
599 | ++__j, ++__q) \ |
600 | __a._M_data[*__j] _Op##= *__q; \ |
601 | } \ |
602 | \ |
603 | template<typename _Tp> \ |
604 | inline void \ |
605 | _Array_augmented_##_Name(_Array<_Tp> __a, size_t __n, \ |
606 | _Array<_Tp> __b, _Array<size_t> __i) \ |
607 | { \ |
608 | _Tp* __p(__a._M_data); \ |
609 | for (size_t* __j = __i._M_data; __j<__i._M_data + __n; \ |
610 | ++__j, ++__p) \ |
611 | *__p _Op##= __b._M_data[*__j]; \ |
612 | } \ |
613 | \ |
614 | template<typename _Tp, class _Dom> \ |
615 | void \ |
616 | _Array_augmented_##_Name(_Array<_Tp> __a, _Array<size_t> __i, \ |
617 | const _Expr<_Dom, _Tp>& __e, size_t __n) \ |
618 | { \ |
619 | size_t* __j(__i._M_data); \ |
620 | for (size_t __k = 0; __k<__n; ++__k, ++__j) \ |
621 | __a._M_data[*__j] _Op##= __e[__k]; \ |
622 | } \ |
623 | \ |
624 | template<typename _Tp> \ |
625 | void \ |
626 | _Array_augmented_##_Name(_Array<_Tp> __a, _Array<bool> __m, \ |
627 | _Array<_Tp> __b, size_t __n) \ |
628 | { \ |
629 | bool* __ok(__m._M_data); \ |
630 | _Tp* __p(__a._M_data); \ |
631 | for (_Tp* __q = __b._M_data; __q < __b._M_data + __n; \ |
632 | ++__q, ++__ok, ++__p) \ |
633 | { \ |
634 | while (! *__ok) \ |
635 | { \ |
636 | ++__ok; \ |
637 | ++__p; \ |
638 | } \ |
639 | *__p _Op##= *__q; \ |
640 | } \ |
641 | } \ |
642 | \ |
643 | template<typename _Tp> \ |
644 | void \ |
645 | _Array_augmented_##_Name(_Array<_Tp> __a, size_t __n, \ |
646 | _Array<_Tp> __b, _Array<bool> __m) \ |
647 | { \ |
648 | bool* __ok(__m._M_data); \ |
649 | _Tp* __q(__b._M_data); \ |
650 | for (_Tp* __p = __a._M_data; __p < __a._M_data + __n; \ |
651 | ++__p, ++__ok, ++__q) \ |
652 | { \ |
653 | while (! *__ok) \ |
654 | { \ |
655 | ++__ok; \ |
656 | ++__q; \ |
657 | } \ |
658 | *__p _Op##= *__q; \ |
659 | } \ |
660 | } \ |
661 | \ |
662 | template<typename _Tp, class _Dom> \ |
663 | void \ |
664 | _Array_augmented_##_Name(_Array<_Tp> __a, _Array<bool> __m, \ |
665 | const _Expr<_Dom, _Tp>& __e, size_t __n) \ |
666 | { \ |
667 | bool* __ok(__m._M_data); \ |
668 | _Tp* __p(__a._M_data); \ |
669 | for (size_t __i = 0; __i < __n; ++__i, ++__ok, ++__p) \ |
670 | { \ |
671 | while (! *__ok) \ |
672 | { \ |
673 | ++__ok; \ |
674 | ++__p; \ |
675 | } \ |
676 | *__p _Op##= __e[__i]; \ |
677 | } \ |
678 | } |
679 | |
680 | _DEFINE_ARRAY_FUNCTION(+, __plus) |
681 | _DEFINE_ARRAY_FUNCTION(-, __minus) |
682 | _DEFINE_ARRAY_FUNCTION(*, __multiplies) |
683 | _DEFINE_ARRAY_FUNCTION(/, __divides) |
684 | _DEFINE_ARRAY_FUNCTION(%, __modulus) |
685 | _DEFINE_ARRAY_FUNCTION(^, __bitwise_xor) |
686 | _DEFINE_ARRAY_FUNCTION(|, __bitwise_or) |
687 | _DEFINE_ARRAY_FUNCTION(&, __bitwise_and) |
688 | _DEFINE_ARRAY_FUNCTION(<<, __shift_left) |
689 | _DEFINE_ARRAY_FUNCTION(>>, __shift_right) |
690 | |
691 | #undef _DEFINE_ARRAY_FUNCTION |
692 | |
693 | _GLIBCXX_END_NAMESPACE_VERSION |
694 | } // namespace |
695 | |
696 | # include <bits/valarray_array.tcc> |
697 | |
698 | #endif /* _ARRAY_H */ |
699 | |