1 | /* |
2 | * Copyright (c) 2001, 2019, Oracle and/or its affiliates. All rights reserved. |
3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
4 | * |
5 | * This code is free software; you can redistribute it and/or modify it |
6 | * under the terms of the GNU General Public License version 2 only, as |
7 | * published by the Free Software Foundation. |
8 | * |
9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
12 | * version 2 for more details (a copy is included in the LICENSE file that |
13 | * accompanied this code). |
14 | * |
15 | * You should have received a copy of the GNU General Public License version |
16 | * 2 along with this work; if not, write to the Free Software Foundation, |
17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
18 | * |
19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
20 | * or visit www.oracle.com if you need additional information or have any |
21 | * questions. |
22 | * |
23 | */ |
24 | |
25 | #ifndef SHARE_GC_SHARED_PTRQUEUE_HPP |
26 | #define SHARE_GC_SHARED_PTRQUEUE_HPP |
27 | |
28 | #include "memory/padded.hpp" |
29 | #include "utilities/align.hpp" |
30 | #include "utilities/debug.hpp" |
31 | #include "utilities/lockFreeStack.hpp" |
32 | #include "utilities/sizes.hpp" |
33 | |
34 | class Mutex; |
35 | class Monitor; |
36 | |
37 | // There are various techniques that require threads to be able to log |
38 | // addresses. For example, a generational write barrier might log |
39 | // the addresses of modified old-generation objects. This type supports |
40 | // this operation. |
41 | |
42 | class BufferNode; |
43 | class PtrQueueSet; |
44 | class PtrQueue { |
45 | friend class VMStructs; |
46 | |
47 | // Noncopyable - not defined. |
48 | PtrQueue(const PtrQueue&); |
49 | PtrQueue& operator=(const PtrQueue&); |
50 | |
51 | // The ptr queue set to which this queue belongs. |
52 | PtrQueueSet* const _qset; |
53 | |
54 | // Whether updates should be logged. |
55 | bool _active; |
56 | |
57 | // The (byte) index at which an object was last enqueued. Starts at |
58 | // capacity_in_bytes (indicating an empty buffer) and goes towards zero. |
59 | // Value is always pointer-size aligned. |
60 | size_t _index; |
61 | |
62 | // Size of the current buffer, in bytes. |
63 | // Value is always pointer-size aligned. |
64 | size_t _capacity_in_bytes; |
65 | |
66 | static const size_t _element_size = sizeof(void*); |
67 | |
68 | // Get the capacity, in bytes. The capacity must have been set. |
69 | size_t capacity_in_bytes() const { |
70 | assert(_capacity_in_bytes > 0, "capacity not set" ); |
71 | return _capacity_in_bytes; |
72 | } |
73 | |
74 | static size_t byte_index_to_index(size_t ind) { |
75 | assert(is_aligned(ind, _element_size), "precondition" ); |
76 | return ind / _element_size; |
77 | } |
78 | |
79 | static size_t index_to_byte_index(size_t ind) { |
80 | return ind * _element_size; |
81 | } |
82 | |
83 | protected: |
84 | // The buffer. |
85 | void** _buf; |
86 | |
87 | size_t index() const { |
88 | return byte_index_to_index(_index); |
89 | } |
90 | |
91 | void set_index(size_t new_index) { |
92 | size_t byte_index = index_to_byte_index(new_index); |
93 | assert(byte_index <= capacity_in_bytes(), "precondition" ); |
94 | _index = byte_index; |
95 | } |
96 | |
97 | size_t capacity() const { |
98 | return byte_index_to_index(capacity_in_bytes()); |
99 | } |
100 | |
101 | PtrQueueSet* qset() const { return _qset; } |
102 | |
103 | // Process queue entries and release resources. |
104 | void flush_impl(); |
105 | |
106 | // Process (some of) the buffer and leave it in place for further use, |
107 | // or enqueue the buffer and allocate a new one. |
108 | virtual void handle_completed_buffer() = 0; |
109 | |
110 | void allocate_buffer(); |
111 | |
112 | // Enqueue the current buffer in the qset and allocate a new buffer. |
113 | void enqueue_completed_buffer(); |
114 | |
115 | // Initialize this queue to contain a null buffer, and be part of the |
116 | // given PtrQueueSet. |
117 | PtrQueue(PtrQueueSet* qset, bool active = false); |
118 | |
119 | // Requires queue flushed. |
120 | ~PtrQueue(); |
121 | |
122 | public: |
123 | |
124 | // Forcibly set empty. |
125 | void reset() { |
126 | if (_buf != NULL) { |
127 | _index = capacity_in_bytes(); |
128 | } |
129 | } |
130 | |
131 | void enqueue(volatile void* ptr) { |
132 | enqueue((void*)(ptr)); |
133 | } |
134 | |
135 | // Enqueues the given "obj". |
136 | void enqueue(void* ptr) { |
137 | if (!_active) return; |
138 | else enqueue_known_active(ptr); |
139 | } |
140 | |
141 | void handle_zero_index(); |
142 | |
143 | void enqueue_known_active(void* ptr); |
144 | |
145 | // Return the size of the in-use region. |
146 | size_t size() const { |
147 | size_t result = 0; |
148 | if (_buf != NULL) { |
149 | assert(_index <= capacity_in_bytes(), "Invariant" ); |
150 | result = byte_index_to_index(capacity_in_bytes() - _index); |
151 | } |
152 | return result; |
153 | } |
154 | |
155 | bool is_empty() const { |
156 | return _buf == NULL || capacity_in_bytes() == _index; |
157 | } |
158 | |
159 | // Set the "active" property of the queue to "b". An enqueue to an |
160 | // inactive thread is a no-op. Setting a queue to inactive resets its |
161 | // log to the empty state. |
162 | void set_active(bool b) { |
163 | _active = b; |
164 | if (!b && _buf != NULL) { |
165 | reset(); |
166 | } else if (b && _buf != NULL) { |
167 | assert(index() == capacity(), |
168 | "invariant: queues are empty when activated." ); |
169 | } |
170 | } |
171 | |
172 | bool is_active() const { return _active; } |
173 | |
174 | // To support compiler. |
175 | |
176 | protected: |
177 | template<typename Derived> |
178 | static ByteSize byte_offset_of_index() { |
179 | return byte_offset_of(Derived, _index); |
180 | } |
181 | |
182 | static ByteSize byte_width_of_index() { return in_ByteSize(sizeof(size_t)); } |
183 | |
184 | template<typename Derived> |
185 | static ByteSize byte_offset_of_buf() { |
186 | return byte_offset_of(Derived, _buf); |
187 | } |
188 | |
189 | static ByteSize byte_width_of_buf() { return in_ByteSize(_element_size); } |
190 | |
191 | template<typename Derived> |
192 | static ByteSize byte_offset_of_active() { |
193 | return byte_offset_of(Derived, _active); |
194 | } |
195 | |
196 | static ByteSize byte_width_of_active() { return in_ByteSize(sizeof(bool)); } |
197 | |
198 | }; |
199 | |
200 | class BufferNode { |
201 | size_t _index; |
202 | BufferNode* volatile _next; |
203 | void* _buffer[1]; // Pseudo flexible array member. |
204 | |
205 | BufferNode() : _index(0), _next(NULL) { } |
206 | ~BufferNode() { } |
207 | |
208 | static size_t buffer_offset() { |
209 | return offset_of(BufferNode, _buffer); |
210 | } |
211 | |
212 | static BufferNode* volatile* next_ptr(BufferNode& bn) { return &bn._next; } |
213 | |
214 | AIX_ONLY(public:) // xlC 12 on AIX doesn't implement C++ DR45. |
215 | // Allocate a new BufferNode with the "buffer" having size elements. |
216 | static BufferNode* allocate(size_t size); |
217 | |
218 | // Free a BufferNode. |
219 | static void deallocate(BufferNode* node); |
220 | |
221 | public: |
222 | typedef LockFreeStack<BufferNode, &next_ptr> Stack; |
223 | |
224 | BufferNode* next() const { return _next; } |
225 | void set_next(BufferNode* n) { _next = n; } |
226 | size_t index() const { return _index; } |
227 | void set_index(size_t i) { _index = i; } |
228 | |
229 | // Return the BufferNode containing the buffer, after setting its index. |
230 | static BufferNode* make_node_from_buffer(void** buffer, size_t index) { |
231 | BufferNode* node = |
232 | reinterpret_cast<BufferNode*>( |
233 | reinterpret_cast<char*>(buffer) - buffer_offset()); |
234 | node->set_index(index); |
235 | return node; |
236 | } |
237 | |
238 | // Return the buffer for node. |
239 | static void** make_buffer_from_node(BufferNode *node) { |
240 | // &_buffer[0] might lead to index out of bounds warnings. |
241 | return reinterpret_cast<void**>( |
242 | reinterpret_cast<char*>(node) + buffer_offset()); |
243 | } |
244 | |
245 | class Allocator; // Free-list based allocator. |
246 | class TestSupport; // Unit test support. |
247 | }; |
248 | |
249 | // Allocation is based on a lock-free free list of nodes, linked through |
250 | // BufferNode::_next (see BufferNode::Stack). To solve the ABA problem, |
251 | // popping a node from the free list is performed within a GlobalCounter |
252 | // critical section, and pushing nodes onto the free list is done after |
253 | // a GlobalCounter synchronization associated with the nodes to be pushed. |
254 | // This is documented behavior so that other parts of the node life-cycle |
255 | // can depend on and make use of it too. |
256 | class BufferNode::Allocator { |
257 | friend class TestSupport; |
258 | |
259 | // Since we don't expect many instances, and measured >15% speedup |
260 | // on stress gtest, padding seems like a good tradeoff here. |
261 | #define DECLARE_PADDED_MEMBER(Id, Type, Name) \ |
262 | Type Name; DEFINE_PAD_MINUS_SIZE(Id, DEFAULT_CACHE_LINE_SIZE, sizeof(Type)) |
263 | |
264 | const size_t _buffer_size; |
265 | char _name[DEFAULT_CACHE_LINE_SIZE - sizeof(size_t)]; // Use name as padding. |
266 | DECLARE_PADDED_MEMBER(1, Stack, _pending_list); |
267 | DECLARE_PADDED_MEMBER(2, Stack, _free_list); |
268 | DECLARE_PADDED_MEMBER(3, volatile size_t, _pending_count); |
269 | DECLARE_PADDED_MEMBER(4, volatile size_t, _free_count); |
270 | DECLARE_PADDED_MEMBER(5, volatile bool, _transfer_lock); |
271 | |
272 | #undef DECLARE_PADDED_MEMBER |
273 | |
274 | void delete_list(BufferNode* list); |
275 | bool try_transfer_pending(); |
276 | |
277 | public: |
278 | Allocator(const char* name, size_t buffer_size); |
279 | ~Allocator(); |
280 | |
281 | const char* name() const { return _name; } |
282 | size_t buffer_size() const { return _buffer_size; } |
283 | size_t free_count() const; |
284 | BufferNode* allocate(); |
285 | void release(BufferNode* node); |
286 | |
287 | // Deallocate some of the available buffers. remove_goal is the target |
288 | // number to remove. Returns the number actually deallocated, which may |
289 | // be less than the goal if there were fewer available. |
290 | size_t reduce_free_list(size_t remove_goal); |
291 | }; |
292 | |
293 | // A PtrQueueSet represents resources common to a set of pointer queues. |
294 | // In particular, the individual queues allocate buffers from this shared |
295 | // set, and return completed buffers to the set. |
296 | class PtrQueueSet { |
297 | BufferNode::Allocator* _allocator; |
298 | |
299 | Monitor* _cbl_mon; // Protects the fields below. |
300 | BufferNode* _completed_buffers_head; |
301 | BufferNode* _completed_buffers_tail; |
302 | volatile size_t _n_completed_buffers; |
303 | |
304 | size_t _process_completed_buffers_threshold; |
305 | volatile bool _process_completed_buffers; |
306 | |
307 | // If true, notify_all on _cbl_mon when the threshold is reached. |
308 | bool _notify_when_complete; |
309 | |
310 | void assert_completed_buffers_list_len_correct_locked() NOT_DEBUG_RETURN; |
311 | |
312 | protected: |
313 | bool _all_active; |
314 | |
315 | // Create an empty ptr queue set. |
316 | PtrQueueSet(bool notify_when_complete = false); |
317 | ~PtrQueueSet(); |
318 | |
319 | // Because of init-order concerns, we can't pass these as constructor |
320 | // arguments. |
321 | void initialize(Monitor* cbl_mon, BufferNode::Allocator* allocator); |
322 | |
323 | // For (unlocked!) iteration over the completed buffers. |
324 | BufferNode* completed_buffers_head() const { return _completed_buffers_head; } |
325 | |
326 | // Deallocate all of the completed buffers. |
327 | void abandon_completed_buffers(); |
328 | |
329 | public: |
330 | |
331 | // Return the buffer for a BufferNode of size buffer_size(). |
332 | void** allocate_buffer(); |
333 | |
334 | // Return an empty buffer to the free list. The node is required |
335 | // to have been allocated with a size of buffer_size(). |
336 | void deallocate_buffer(BufferNode* node); |
337 | |
338 | // A completed buffer is a buffer the mutator is finished with, and |
339 | // is ready to be processed by the collector. It need not be full. |
340 | |
341 | // Adds node to the completed buffer list. |
342 | void enqueue_completed_buffer(BufferNode* node); |
343 | |
344 | // If the number of completed buffers is > stop_at, then remove and |
345 | // return a completed buffer from the list. Otherwise, return NULL. |
346 | BufferNode* get_completed_buffer(size_t stop_at = 0); |
347 | |
348 | bool process_completed_buffers() { return _process_completed_buffers; } |
349 | void set_process_completed_buffers(bool x) { _process_completed_buffers = x; } |
350 | |
351 | bool is_active() { return _all_active; } |
352 | |
353 | size_t buffer_size() const { |
354 | return _allocator->buffer_size(); |
355 | } |
356 | |
357 | // Get/Set the number of completed buffers that triggers log processing. |
358 | // Log processing should be done when the number of buffers exceeds the |
359 | // threshold. |
360 | void set_process_completed_buffers_threshold(size_t sz) { |
361 | _process_completed_buffers_threshold = sz; |
362 | } |
363 | size_t process_completed_buffers_threshold() const { |
364 | return _process_completed_buffers_threshold; |
365 | } |
366 | static const size_t ProcessCompletedBuffersThresholdNever = ~size_t(0); |
367 | |
368 | size_t completed_buffers_num() const { return _n_completed_buffers; } |
369 | |
370 | void merge_bufferlists(PtrQueueSet* src); |
371 | |
372 | // Notify the consumer if the number of buffers crossed the threshold |
373 | void notify_if_necessary(); |
374 | }; |
375 | |
376 | #endif // SHARE_GC_SHARED_PTRQUEUE_HPP |
377 | |