1/*
2 * Copyright (c) 2018, 2019, Red Hat, Inc. All rights reserved.
3 *
4 * This code is free software; you can redistribute it and/or modify it
5 * under the terms of the GNU General Public License version 2 only, as
6 * published by the Free Software Foundation.
7 *
8 * This code is distributed in the hope that it will be useful, but WITHOUT
9 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
10 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
11 * version 2 for more details (a copy is included in the LICENSE file that
12 * accompanied this code).
13 *
14 * You should have received a copy of the GNU General Public License version
15 * 2 along with this work; if not, write to the Free Software Foundation,
16 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
17 *
18 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
19 * or visit www.oracle.com if you need additional information or have any
20 * questions.
21 *
22 */
23
24#include "precompiled.hpp"
25
26#include "classfile/classLoaderData.hpp"
27#include "classfile/classLoaderDataGraph.hpp"
28#include "gc/shared/referenceProcessor.hpp"
29#include "gc/shared/referenceProcessorPhaseTimes.hpp"
30#include "gc/shared/workgroup.hpp"
31#include "gc/shared/weakProcessor.inline.hpp"
32#include "gc/shenandoah/shenandoahBarrierSet.hpp"
33#include "gc/shenandoah/shenandoahClosures.inline.hpp"
34#include "gc/shenandoah/shenandoahCodeRoots.hpp"
35#include "gc/shenandoah/shenandoahCollectionSet.hpp"
36#include "gc/shenandoah/shenandoahCollectorPolicy.hpp"
37#include "gc/shenandoah/shenandoahFreeSet.hpp"
38#include "gc/shenandoah/shenandoahPhaseTimings.hpp"
39#include "gc/shenandoah/shenandoahHeap.inline.hpp"
40#include "gc/shenandoah/shenandoahHeapRegionSet.inline.hpp"
41#include "gc/shenandoah/shenandoahHeuristics.hpp"
42#include "gc/shenandoah/shenandoahMarkingContext.inline.hpp"
43#include "gc/shenandoah/shenandoahOopClosures.inline.hpp"
44#include "gc/shenandoah/shenandoahRootProcessor.inline.hpp"
45#include "gc/shenandoah/shenandoahStringDedup.hpp"
46#include "gc/shenandoah/shenandoahTaskqueue.inline.hpp"
47#include "gc/shenandoah/shenandoahTimingTracker.hpp"
48#include "gc/shenandoah/shenandoahTraversalGC.hpp"
49#include "gc/shenandoah/shenandoahUtils.hpp"
50#include "gc/shenandoah/shenandoahVerifier.hpp"
51
52#include "memory/iterator.hpp"
53#include "memory/metaspace.hpp"
54#include "memory/resourceArea.hpp"
55#include "memory/universe.hpp"
56
57/**
58 * NOTE: We are using the SATB buffer in thread.hpp and satbMarkQueue.hpp, however, it is not an SATB algorithm.
59 * We're using the buffer as generic oop buffer to enqueue new values in concurrent oop stores, IOW, the algorithm
60 * is incremental-update-based.
61 *
62 * NOTE on interaction with TAMS: we want to avoid traversing new objects for
63 * several reasons:
64 * - We will not reclaim them in this cycle anyway, because they are not in the
65 * cset
66 * - It makes up for the bulk of work during final-pause
67 * - It also shortens the concurrent cycle because we don't need to
68 * pointlessly traverse through newly allocated objects.
69 * - As a nice side-effect, it solves the I-U termination problem (mutators
70 * cannot outrun the GC by allocating like crazy)
71 * - It is an easy way to achieve MWF. What MWF does is to also enqueue the
72 * target object of stores if it's new. Treating new objects live implicitely
73 * achieves the same, but without extra barriers. I think the effect of
74 * shortened final-pause (mentioned above) is the main advantage of MWF. In
75 * particular, we will not see the head of a completely new long linked list
76 * in final-pause and end up traversing huge chunks of the heap there.
77 * - We don't need to see/update the fields of new objects either, because they
78 * are either still null, or anything that's been stored into them has been
79 * evacuated+enqueued before (and will thus be treated later).
80 *
81 * We achieve this by setting TAMS for each region, and everything allocated
82 * beyond TAMS will be 'implicitely marked'.
83 *
84 * Gotchas:
85 * - While we want new objects to be implicitely marked, we don't want to count
86 * them alive. Otherwise the next cycle wouldn't pick them up and consider
87 * them for cset. This means that we need to protect such regions from
88 * getting accidentally thrashed at the end of traversal cycle. This is why I
89 * keep track of alloc-regions and check is_alloc_region() in the trashing
90 * code.
91 * - We *need* to traverse through evacuated objects. Those objects are
92 * pre-existing, and any references in them point to interesting objects that
93 * we need to see. We also want to count them as live, because we just
94 * determined that they are alive :-) I achieve this by upping TAMS
95 * concurrently for every gclab/gc-shared alloc before publishing the
96 * evacuated object. This way, the GC threads will not consider such objects
97 * implictely marked, and traverse through them as normal.
98 */
99class ShenandoahTraversalSATBBufferClosure : public SATBBufferClosure {
100private:
101 ShenandoahObjToScanQueue* _queue;
102 ShenandoahTraversalGC* _traversal_gc;
103 ShenandoahHeap* const _heap;
104
105public:
106 ShenandoahTraversalSATBBufferClosure(ShenandoahObjToScanQueue* q) :
107 _queue(q),
108 _heap(ShenandoahHeap::heap())
109 { }
110
111 void do_buffer(void** buffer, size_t size) {
112 for (size_t i = 0; i < size; ++i) {
113 oop* p = (oop*) &buffer[i];
114 oop obj = RawAccess<>::oop_load(p);
115 shenandoah_assert_not_forwarded(p, obj);
116 if (_heap->marking_context()->mark(obj)) {
117 _queue->push(ShenandoahMarkTask(obj));
118 }
119 }
120 }
121};
122
123class ShenandoahTraversalSATBThreadsClosure : public ThreadClosure {
124private:
125 ShenandoahTraversalSATBBufferClosure* _satb_cl;
126
127public:
128 ShenandoahTraversalSATBThreadsClosure(ShenandoahTraversalSATBBufferClosure* satb_cl) :
129 _satb_cl(satb_cl) {}
130
131 void do_thread(Thread* thread) {
132 ShenandoahThreadLocalData::satb_mark_queue(thread).apply_closure_and_empty(_satb_cl);
133 }
134};
135
136// Like CLDToOopClosure, but clears has_modified_oops, so that we can record modified CLDs during traversal
137// and remark them later during final-traversal.
138class ShenandoahMarkCLDClosure : public CLDClosure {
139private:
140 OopClosure* _cl;
141public:
142 ShenandoahMarkCLDClosure(OopClosure* cl) : _cl(cl) {}
143 void do_cld(ClassLoaderData* cld) {
144 cld->oops_do(_cl, true, true);
145 }
146};
147
148// Like CLDToOopClosure, but only process modified CLDs
149class ShenandoahRemarkCLDClosure : public CLDClosure {
150private:
151 OopClosure* _cl;
152public:
153 ShenandoahRemarkCLDClosure(OopClosure* cl) : _cl(cl) {}
154 void do_cld(ClassLoaderData* cld) {
155 if (cld->has_modified_oops()) {
156 cld->oops_do(_cl, true, true);
157 }
158 }
159};
160
161class ShenandoahInitTraversalCollectionTask : public AbstractGangTask {
162private:
163 ShenandoahCSetRootScanner* _rp;
164 ShenandoahHeap* _heap;
165 ShenandoahCsetCodeRootsIterator* _cset_coderoots;
166 ShenandoahStringDedupRoots _dedup_roots;
167
168public:
169 ShenandoahInitTraversalCollectionTask(ShenandoahCSetRootScanner* rp) :
170 AbstractGangTask("Shenandoah Init Traversal Collection"),
171 _rp(rp),
172 _heap(ShenandoahHeap::heap()) {}
173
174 void work(uint worker_id) {
175 ShenandoahParallelWorkerSession worker_session(worker_id);
176
177 ShenandoahEvacOOMScope oom_evac_scope;
178 ShenandoahObjToScanQueueSet* queues = _heap->traversal_gc()->task_queues();
179 ShenandoahObjToScanQueue* q = queues->queue(worker_id);
180
181 bool process_refs = _heap->process_references();
182 bool unload_classes = _heap->unload_classes();
183 ReferenceProcessor* rp = NULL;
184 if (process_refs) {
185 rp = _heap->ref_processor();
186 }
187
188 // Step 1: Process ordinary GC roots.
189 {
190 ShenandoahTraversalClosure roots_cl(q, rp);
191 ShenandoahMarkCLDClosure cld_cl(&roots_cl);
192 MarkingCodeBlobClosure code_cl(&roots_cl, CodeBlobToOopClosure::FixRelocations);
193 if (unload_classes) {
194 _rp->roots_do(worker_id, &roots_cl, NULL, &code_cl);
195 } else {
196 _rp->roots_do(worker_id, &roots_cl, &cld_cl, &code_cl);
197 }
198
199 AlwaysTrueClosure is_alive;
200 _dedup_roots.oops_do(&is_alive, &roots_cl, worker_id);
201 }
202 }
203};
204
205class ShenandoahConcurrentTraversalCollectionTask : public AbstractGangTask {
206private:
207 ShenandoahTaskTerminator* _terminator;
208 ShenandoahHeap* _heap;
209public:
210 ShenandoahConcurrentTraversalCollectionTask(ShenandoahTaskTerminator* terminator) :
211 AbstractGangTask("Shenandoah Concurrent Traversal Collection"),
212 _terminator(terminator),
213 _heap(ShenandoahHeap::heap()) {}
214
215 void work(uint worker_id) {
216 ShenandoahConcurrentWorkerSession worker_session(worker_id);
217 ShenandoahSuspendibleThreadSetJoiner stsj(ShenandoahSuspendibleWorkers);
218 ShenandoahEvacOOMScope oom_evac_scope;
219 ShenandoahTraversalGC* traversal_gc = _heap->traversal_gc();
220
221 // Drain all outstanding work in queues.
222 traversal_gc->main_loop(worker_id, _terminator, true);
223 }
224};
225
226class ShenandoahFinalTraversalCollectionTask : public AbstractGangTask {
227private:
228 ShenandoahAllRootScanner* _rp;
229 ShenandoahTaskTerminator* _terminator;
230 ShenandoahHeap* _heap;
231public:
232 ShenandoahFinalTraversalCollectionTask(ShenandoahAllRootScanner* rp, ShenandoahTaskTerminator* terminator) :
233 AbstractGangTask("Shenandoah Final Traversal Collection"),
234 _rp(rp),
235 _terminator(terminator),
236 _heap(ShenandoahHeap::heap()) {}
237
238 void work(uint worker_id) {
239 ShenandoahParallelWorkerSession worker_session(worker_id);
240
241 ShenandoahEvacOOMScope oom_evac_scope;
242 ShenandoahTraversalGC* traversal_gc = _heap->traversal_gc();
243
244 ShenandoahObjToScanQueueSet* queues = traversal_gc->task_queues();
245 ShenandoahObjToScanQueue* q = queues->queue(worker_id);
246
247 bool process_refs = _heap->process_references();
248 bool unload_classes = _heap->unload_classes();
249 ReferenceProcessor* rp = NULL;
250 if (process_refs) {
251 rp = _heap->ref_processor();
252 }
253
254 // Step 0: Drain outstanding SATB queues.
255 // NOTE: we piggy-back draining of remaining thread SATB buffers on the final root scan below.
256 ShenandoahTraversalSATBBufferClosure satb_cl(q);
257 {
258 // Process remaining finished SATB buffers.
259 SATBMarkQueueSet& satb_mq_set = ShenandoahBarrierSet::satb_mark_queue_set();
260 while (satb_mq_set.apply_closure_to_completed_buffer(&satb_cl));
261 // Process remaining threads SATB buffers below.
262 }
263
264 // Step 1: Process GC roots.
265 // For oops in code roots, they are marked, evacuated, enqueued for further traversal,
266 // and the references to the oops are updated during init pause. New nmethods are handled
267 // in similar way during nmethod-register process. Therefore, we don't need to rescan code
268 // roots here.
269 if (!_heap->is_degenerated_gc_in_progress()) {
270 ShenandoahTraversalClosure roots_cl(q, rp);
271 ShenandoahTraversalSATBThreadsClosure tc(&satb_cl);
272 if (unload_classes) {
273 ShenandoahRemarkCLDClosure remark_cld_cl(&roots_cl);
274 _rp->strong_roots_do(worker_id, &roots_cl, &remark_cld_cl, NULL, &tc);
275 } else {
276 CLDToOopClosure cld_cl(&roots_cl, ClassLoaderData::_claim_strong);
277 _rp->roots_do(worker_id, &roots_cl, &cld_cl, NULL, &tc);
278 }
279 } else {
280 ShenandoahTraversalDegenClosure roots_cl(q, rp);
281 ShenandoahTraversalSATBThreadsClosure tc(&satb_cl);
282 if (unload_classes) {
283 ShenandoahRemarkCLDClosure remark_cld_cl(&roots_cl);
284 _rp->strong_roots_do(worker_id, &roots_cl, &remark_cld_cl, NULL, &tc);
285 } else {
286 CLDToOopClosure cld_cl(&roots_cl, ClassLoaderData::_claim_strong);
287 _rp->roots_do(worker_id, &roots_cl, &cld_cl, NULL, &tc);
288 }
289 }
290
291 {
292 ShenandoahWorkerTimings *worker_times = _heap->phase_timings()->worker_times();
293 ShenandoahWorkerTimingsTracker timer(worker_times, ShenandoahPhaseTimings::FinishQueues, worker_id);
294
295 // Step 3: Finally drain all outstanding work in queues.
296 traversal_gc->main_loop(worker_id, _terminator, false);
297 }
298
299 }
300};
301
302ShenandoahTraversalGC::ShenandoahTraversalGC(ShenandoahHeap* heap, size_t num_regions) :
303 _heap(heap),
304 _task_queues(new ShenandoahObjToScanQueueSet(heap->max_workers())),
305 _traversal_set(ShenandoahHeapRegionSet()) {
306
307 // Traversal does not support concurrent code root scanning
308 FLAG_SET_DEFAULT(ShenandoahConcurrentScanCodeRoots, false);
309
310 uint num_queues = heap->max_workers();
311 for (uint i = 0; i < num_queues; ++i) {
312 ShenandoahObjToScanQueue* task_queue = new ShenandoahObjToScanQueue();
313 task_queue->initialize();
314 _task_queues->register_queue(i, task_queue);
315 }
316}
317
318ShenandoahTraversalGC::~ShenandoahTraversalGC() {
319}
320
321void ShenandoahTraversalGC::prepare_regions() {
322 size_t num_regions = _heap->num_regions();
323 ShenandoahMarkingContext* const ctx = _heap->marking_context();
324 for (size_t i = 0; i < num_regions; i++) {
325 ShenandoahHeapRegion* region = _heap->get_region(i);
326 if (_heap->is_bitmap_slice_committed(region)) {
327 if (_traversal_set.is_in(i)) {
328 ctx->capture_top_at_mark_start(region);
329 region->clear_live_data();
330 assert(ctx->is_bitmap_clear_range(region->bottom(), region->end()), "bitmap for traversal regions must be cleared");
331 } else {
332 // Everything outside the traversal set is always considered live.
333 ctx->reset_top_at_mark_start(region);
334 }
335 } else {
336 // FreeSet may contain uncommitted empty regions, once they are recommitted,
337 // their TAMS may have old values, so reset them here.
338 ctx->reset_top_at_mark_start(region);
339 }
340 }
341}
342
343void ShenandoahTraversalGC::prepare() {
344 _heap->collection_set()->clear();
345 assert(_heap->collection_set()->count() == 0, "collection set not clear");
346
347 {
348 ShenandoahGCPhase phase(ShenandoahPhaseTimings::traversal_gc_make_parsable);
349 _heap->make_parsable(true);
350 }
351
352 if (UseTLAB) {
353 ShenandoahGCPhase phase(ShenandoahPhaseTimings::traversal_gc_resize_tlabs);
354 _heap->resize_tlabs();
355 }
356
357 assert(_heap->marking_context()->is_bitmap_clear(), "need clean mark bitmap");
358 assert(!_heap->marking_context()->is_complete(), "should not be complete");
359
360 ShenandoahFreeSet* free_set = _heap->free_set();
361 ShenandoahCollectionSet* collection_set = _heap->collection_set();
362
363 // Find collection set
364 _heap->heuristics()->choose_collection_set(collection_set);
365 prepare_regions();
366
367 // Rebuild free set
368 free_set->rebuild();
369
370 log_info(gc, ergo)("Collectable Garbage: " SIZE_FORMAT "M, " SIZE_FORMAT "M CSet, " SIZE_FORMAT " CSet regions",
371 collection_set->garbage() / M, collection_set->live_data() / M, collection_set->count());
372}
373
374void ShenandoahTraversalGC::init_traversal_collection() {
375 assert(ShenandoahSafepoint::is_at_shenandoah_safepoint(), "STW traversal GC");
376
377 if (ShenandoahVerify) {
378 _heap->verifier()->verify_before_traversal();
379 }
380
381 if (VerifyBeforeGC) {
382 Universe::verify();
383 }
384
385 {
386 ShenandoahGCPhase phase_prepare(ShenandoahPhaseTimings::traversal_gc_prepare);
387 ShenandoahHeapLocker lock(_heap->lock());
388 prepare();
389 }
390
391 _heap->set_concurrent_traversal_in_progress(true);
392
393 bool process_refs = _heap->process_references();
394 if (process_refs) {
395 ReferenceProcessor* rp = _heap->ref_processor();
396 rp->enable_discovery(true /*verify_no_refs*/);
397 rp->setup_policy(_heap->soft_ref_policy()->should_clear_all_soft_refs());
398 }
399
400 {
401 ShenandoahGCPhase phase_work(ShenandoahPhaseTimings::init_traversal_gc_work);
402 assert(_task_queues->is_empty(), "queues must be empty before traversal GC");
403 TASKQUEUE_STATS_ONLY(_task_queues->reset_taskqueue_stats());
404
405#if COMPILER2_OR_JVMCI
406 DerivedPointerTable::clear();
407#endif
408
409 {
410 uint nworkers = _heap->workers()->active_workers();
411 task_queues()->reserve(nworkers);
412 ShenandoahCSetRootScanner rp(nworkers, ShenandoahPhaseTimings::init_traversal_gc_work);
413 ShenandoahInitTraversalCollectionTask traversal_task(&rp);
414 _heap->workers()->run_task(&traversal_task);
415 }
416
417#if COMPILER2_OR_JVMCI
418 DerivedPointerTable::update_pointers();
419#endif
420 }
421
422 if (ShenandoahPacing) {
423 _heap->pacer()->setup_for_traversal();
424 }
425}
426
427void ShenandoahTraversalGC::main_loop(uint w, ShenandoahTaskTerminator* t, bool sts_yield) {
428 ShenandoahObjToScanQueue* q = task_queues()->queue(w);
429
430 // Initialize live data.
431 jushort* ld = _heap->get_liveness_cache(w);
432
433 ReferenceProcessor* rp = NULL;
434 if (_heap->process_references()) {
435 rp = _heap->ref_processor();
436 }
437 {
438 if (!_heap->is_degenerated_gc_in_progress()) {
439 if (_heap->unload_classes()) {
440 if (ShenandoahStringDedup::is_enabled()) {
441 ShenandoahTraversalMetadataDedupClosure cl(q, rp);
442 main_loop_work<ShenandoahTraversalMetadataDedupClosure>(&cl, ld, w, t, sts_yield);
443 } else {
444 ShenandoahTraversalMetadataClosure cl(q, rp);
445 main_loop_work<ShenandoahTraversalMetadataClosure>(&cl, ld, w, t, sts_yield);
446 }
447 } else {
448 if (ShenandoahStringDedup::is_enabled()) {
449 ShenandoahTraversalDedupClosure cl(q, rp);
450 main_loop_work<ShenandoahTraversalDedupClosure>(&cl, ld, w, t, sts_yield);
451 } else {
452 ShenandoahTraversalClosure cl(q, rp);
453 main_loop_work<ShenandoahTraversalClosure>(&cl, ld, w, t, sts_yield);
454 }
455 }
456 } else {
457 if (_heap->unload_classes()) {
458 if (ShenandoahStringDedup::is_enabled()) {
459 ShenandoahTraversalMetadataDedupDegenClosure cl(q, rp);
460 main_loop_work<ShenandoahTraversalMetadataDedupDegenClosure>(&cl, ld, w, t, sts_yield);
461 } else {
462 ShenandoahTraversalMetadataDegenClosure cl(q, rp);
463 main_loop_work<ShenandoahTraversalMetadataDegenClosure>(&cl, ld, w, t, sts_yield);
464 }
465 } else {
466 if (ShenandoahStringDedup::is_enabled()) {
467 ShenandoahTraversalDedupDegenClosure cl(q, rp);
468 main_loop_work<ShenandoahTraversalDedupDegenClosure>(&cl, ld, w, t, sts_yield);
469 } else {
470 ShenandoahTraversalDegenClosure cl(q, rp);
471 main_loop_work<ShenandoahTraversalDegenClosure>(&cl, ld, w, t, sts_yield);
472 }
473 }
474 }
475 }
476
477 _heap->flush_liveness_cache(w);
478}
479
480template <class T>
481void ShenandoahTraversalGC::main_loop_work(T* cl, jushort* live_data, uint worker_id, ShenandoahTaskTerminator* terminator, bool sts_yield) {
482 ShenandoahObjToScanQueueSet* queues = task_queues();
483 ShenandoahObjToScanQueue* q = queues->queue(worker_id);
484 ShenandoahConcurrentMark* conc_mark = _heap->concurrent_mark();
485
486 uintx stride = ShenandoahMarkLoopStride;
487
488 ShenandoahMarkTask task;
489
490 // Process outstanding queues, if any.
491 q = queues->claim_next();
492 while (q != NULL) {
493 if (_heap->check_cancelled_gc_and_yield(sts_yield)) {
494 return;
495 }
496
497 for (uint i = 0; i < stride; i++) {
498 if (q->pop(task)) {
499 conc_mark->do_task<T>(q, cl, live_data, &task);
500 } else {
501 assert(q->is_empty(), "Must be empty");
502 q = queues->claim_next();
503 break;
504 }
505 }
506 }
507
508 if (check_and_handle_cancelled_gc(terminator, sts_yield)) return;
509
510 // Normal loop.
511 q = queues->queue(worker_id);
512
513 ShenandoahTraversalSATBBufferClosure drain_satb(q);
514 SATBMarkQueueSet& satb_mq_set = ShenandoahBarrierSet::satb_mark_queue_set();
515
516 while (true) {
517 if (check_and_handle_cancelled_gc(terminator, sts_yield)) return;
518
519 while (satb_mq_set.completed_buffers_num() > 0) {
520 satb_mq_set.apply_closure_to_completed_buffer(&drain_satb);
521 }
522
523 uint work = 0;
524 for (uint i = 0; i < stride; i++) {
525 if (q->pop(task) ||
526 queues->steal(worker_id, task)) {
527 conc_mark->do_task<T>(q, cl, live_data, &task);
528 work++;
529 } else {
530 break;
531 }
532 }
533
534 if (work == 0) {
535 // No more work, try to terminate
536 ShenandoahEvacOOMScopeLeaver oom_scope_leaver;
537 ShenandoahSuspendibleThreadSetLeaver stsl(sts_yield && ShenandoahSuspendibleWorkers);
538 ShenandoahTerminationTimingsTracker term_tracker(worker_id);
539 ShenandoahTerminatorTerminator tt(_heap);
540
541 if (terminator->offer_termination(&tt)) return;
542 }
543 }
544}
545
546bool ShenandoahTraversalGC::check_and_handle_cancelled_gc(ShenandoahTaskTerminator* terminator, bool sts_yield) {
547 if (_heap->cancelled_gc()) {
548 return true;
549 }
550 return false;
551}
552
553void ShenandoahTraversalGC::concurrent_traversal_collection() {
554 ShenandoahGCPhase phase_work(ShenandoahPhaseTimings::conc_traversal);
555 if (!_heap->cancelled_gc()) {
556 uint nworkers = _heap->workers()->active_workers();
557 task_queues()->reserve(nworkers);
558 ShenandoahTerminationTracker tracker(ShenandoahPhaseTimings::conc_traversal_termination);
559
560 ShenandoahTaskTerminator terminator(nworkers, task_queues());
561 ShenandoahConcurrentTraversalCollectionTask task(&terminator);
562 _heap->workers()->run_task(&task);
563 }
564
565 if (!_heap->cancelled_gc() && ShenandoahPreclean && _heap->process_references()) {
566 preclean_weak_refs();
567 }
568}
569
570void ShenandoahTraversalGC::final_traversal_collection() {
571 _heap->make_parsable(true);
572
573 if (!_heap->cancelled_gc()) {
574#if COMPILER2_OR_JVMCI
575 DerivedPointerTable::clear();
576#endif
577 ShenandoahGCPhase phase_work(ShenandoahPhaseTimings::final_traversal_gc_work);
578 uint nworkers = _heap->workers()->active_workers();
579 task_queues()->reserve(nworkers);
580
581 // Finish traversal
582 ShenandoahAllRootScanner rp(nworkers, ShenandoahPhaseTimings::final_traversal_gc_work);
583 ShenandoahTerminationTracker term(ShenandoahPhaseTimings::final_traversal_gc_termination);
584
585 ShenandoahTaskTerminator terminator(nworkers, task_queues());
586 ShenandoahFinalTraversalCollectionTask task(&rp, &terminator);
587 _heap->workers()->run_task(&task);
588#if COMPILER2_OR_JVMCI
589 DerivedPointerTable::update_pointers();
590#endif
591 }
592
593 if (!_heap->cancelled_gc() && _heap->process_references()) {
594 weak_refs_work();
595 }
596
597 if (!_heap->cancelled_gc()) {
598 fixup_roots();
599 if (_heap->unload_classes()) {
600 _heap->unload_classes_and_cleanup_tables(false);
601 }
602 }
603
604 if (!_heap->cancelled_gc()) {
605 assert(_task_queues->is_empty(), "queues must be empty after traversal GC");
606 TASKQUEUE_STATS_ONLY(_task_queues->print_taskqueue_stats());
607 TASKQUEUE_STATS_ONLY(_task_queues->reset_taskqueue_stats());
608
609 // No more marking expected
610 _heap->mark_complete_marking_context();
611
612 // Resize metaspace
613 MetaspaceGC::compute_new_size();
614
615 // Still good? We can now trash the cset, and make final verification
616 {
617 ShenandoahGCPhase phase_cleanup(ShenandoahPhaseTimings::traversal_gc_cleanup);
618 ShenandoahHeapLocker lock(_heap->lock());
619
620 // Trash everything
621 // Clear immediate garbage regions.
622 size_t num_regions = _heap->num_regions();
623
624 ShenandoahHeapRegionSet* traversal_regions = traversal_set();
625 ShenandoahFreeSet* free_regions = _heap->free_set();
626 ShenandoahMarkingContext* const ctx = _heap->marking_context();
627 free_regions->clear();
628 for (size_t i = 0; i < num_regions; i++) {
629 ShenandoahHeapRegion* r = _heap->get_region(i);
630 bool not_allocated = ctx->top_at_mark_start(r) == r->top();
631
632 bool candidate = traversal_regions->is_in(r) && !r->has_live() && not_allocated;
633 if (r->is_humongous_start() && candidate) {
634 // Trash humongous.
635 HeapWord* humongous_obj = r->bottom();
636 assert(!ctx->is_marked(oop(humongous_obj)), "must not be marked");
637 r->make_trash_immediate();
638 while (i + 1 < num_regions && _heap->get_region(i + 1)->is_humongous_continuation()) {
639 i++;
640 r = _heap->get_region(i);
641 assert(r->is_humongous_continuation(), "must be humongous continuation");
642 r->make_trash_immediate();
643 }
644 } else if (!r->is_empty() && candidate) {
645 // Trash regular.
646 assert(!r->is_humongous(), "handled above");
647 assert(!r->is_trash(), "must not already be trashed");
648 r->make_trash_immediate();
649 }
650 }
651 _heap->collection_set()->clear();
652 _heap->free_set()->rebuild();
653 reset();
654 }
655
656 assert(_task_queues->is_empty(), "queues must be empty after traversal GC");
657 _heap->set_concurrent_traversal_in_progress(false);
658 assert(!_heap->cancelled_gc(), "must not be cancelled when getting out here");
659
660 if (ShenandoahVerify) {
661 _heap->verifier()->verify_after_traversal();
662 }
663
664 if (VerifyAfterGC) {
665 Universe::verify();
666 }
667 }
668}
669
670class ShenandoahTraversalFixRootsClosure : public OopClosure {
671private:
672 template <class T>
673 inline void do_oop_work(T* p) {
674 T o = RawAccess<>::oop_load(p);
675 if (!CompressedOops::is_null(o)) {
676 oop obj = CompressedOops::decode_not_null(o);
677 oop forw = ShenandoahBarrierSet::resolve_forwarded_not_null(obj);
678 if (!oopDesc::equals_raw(obj, forw)) {
679 RawAccess<IS_NOT_NULL>::oop_store(p, forw);
680 }
681 }
682 }
683
684public:
685 inline void do_oop(oop* p) { do_oop_work(p); }
686 inline void do_oop(narrowOop* p) { do_oop_work(p); }
687};
688
689class ShenandoahTraversalFixRootsTask : public AbstractGangTask {
690private:
691 ShenandoahRootUpdater* _rp;
692
693public:
694 ShenandoahTraversalFixRootsTask(ShenandoahRootUpdater* rp) :
695 AbstractGangTask("Shenandoah traversal fix roots"),
696 _rp(rp) {
697 assert(ShenandoahHeap::heap()->has_forwarded_objects(), "Must be");
698 }
699
700 void work(uint worker_id) {
701 ShenandoahParallelWorkerSession worker_session(worker_id);
702 ShenandoahTraversalFixRootsClosure cl;
703 ShenandoahForwardedIsAliveClosure is_alive;
704 _rp->roots_do<ShenandoahForwardedIsAliveClosure, ShenandoahTraversalFixRootsClosure>(worker_id, &is_alive, &cl);
705 }
706};
707
708void ShenandoahTraversalGC::fixup_roots() {
709#if COMPILER2_OR_JVMCI
710 DerivedPointerTable::clear();
711#endif
712 ShenandoahRootUpdater rp(_heap->workers()->active_workers(), ShenandoahPhaseTimings::final_traversal_update_roots, true /* update code cache */);
713 ShenandoahTraversalFixRootsTask update_roots_task(&rp);
714 _heap->workers()->run_task(&update_roots_task);
715#if COMPILER2_OR_JVMCI
716 DerivedPointerTable::update_pointers();
717#endif
718}
719
720void ShenandoahTraversalGC::reset() {
721 _task_queues->clear();
722}
723
724ShenandoahObjToScanQueueSet* ShenandoahTraversalGC::task_queues() {
725 return _task_queues;
726}
727
728class ShenandoahTraversalCancelledGCYieldClosure : public YieldClosure {
729private:
730 ShenandoahHeap* const _heap;
731public:
732 ShenandoahTraversalCancelledGCYieldClosure() : _heap(ShenandoahHeap::heap()) {};
733 virtual bool should_return() { return _heap->cancelled_gc(); }
734};
735
736class ShenandoahTraversalPrecleanCompleteGCClosure : public VoidClosure {
737public:
738 void do_void() {
739 ShenandoahHeap* sh = ShenandoahHeap::heap();
740 ShenandoahTraversalGC* traversal_gc = sh->traversal_gc();
741 assert(sh->process_references(), "why else would we be here?");
742 ShenandoahTaskTerminator terminator(1, traversal_gc->task_queues());
743 shenandoah_assert_rp_isalive_installed();
744 traversal_gc->main_loop((uint) 0, &terminator, true);
745 }
746};
747
748class ShenandoahTraversalKeepAliveUpdateClosure : public OopClosure {
749private:
750 ShenandoahObjToScanQueue* _queue;
751 Thread* _thread;
752 ShenandoahTraversalGC* _traversal_gc;
753 ShenandoahMarkingContext* const _mark_context;
754
755 template <class T>
756 inline void do_oop_work(T* p) {
757 _traversal_gc->process_oop<T, false /* string dedup */, false /* degen */>(p, _thread, _queue, _mark_context);
758 }
759
760public:
761 ShenandoahTraversalKeepAliveUpdateClosure(ShenandoahObjToScanQueue* q) :
762 _queue(q), _thread(Thread::current()),
763 _traversal_gc(ShenandoahHeap::heap()->traversal_gc()),
764 _mark_context(ShenandoahHeap::heap()->marking_context()) {}
765
766 void do_oop(narrowOop* p) { do_oop_work(p); }
767 void do_oop(oop* p) { do_oop_work(p); }
768};
769
770class ShenandoahTraversalKeepAliveUpdateDegenClosure : public OopClosure {
771private:
772 ShenandoahObjToScanQueue* _queue;
773 Thread* _thread;
774 ShenandoahTraversalGC* _traversal_gc;
775 ShenandoahMarkingContext* const _mark_context;
776
777 template <class T>
778 inline void do_oop_work(T* p) {
779 _traversal_gc->process_oop<T, false /* string dedup */, true /* degen */>(p, _thread, _queue, _mark_context);
780 }
781
782public:
783 ShenandoahTraversalKeepAliveUpdateDegenClosure(ShenandoahObjToScanQueue* q) :
784 _queue(q), _thread(Thread::current()),
785 _traversal_gc(ShenandoahHeap::heap()->traversal_gc()),
786 _mark_context(ShenandoahHeap::heap()->marking_context()) {}
787
788 void do_oop(narrowOop* p) { do_oop_work(p); }
789 void do_oop(oop* p) { do_oop_work(p); }
790};
791
792class ShenandoahTraversalSingleThreadKeepAliveUpdateClosure : public OopClosure {
793private:
794 ShenandoahObjToScanQueue* _queue;
795 Thread* _thread;
796 ShenandoahTraversalGC* _traversal_gc;
797 ShenandoahMarkingContext* const _mark_context;
798
799 template <class T>
800 inline void do_oop_work(T* p) {
801 ShenandoahEvacOOMScope evac_scope;
802 _traversal_gc->process_oop<T, false /* string dedup */, false /* degen */>(p, _thread, _queue, _mark_context);
803 }
804
805public:
806 ShenandoahTraversalSingleThreadKeepAliveUpdateClosure(ShenandoahObjToScanQueue* q) :
807 _queue(q), _thread(Thread::current()),
808 _traversal_gc(ShenandoahHeap::heap()->traversal_gc()),
809 _mark_context(ShenandoahHeap::heap()->marking_context()) {}
810
811 void do_oop(narrowOop* p) { do_oop_work(p); }
812 void do_oop(oop* p) { do_oop_work(p); }
813};
814
815class ShenandoahTraversalSingleThreadKeepAliveUpdateDegenClosure : public OopClosure {
816private:
817 ShenandoahObjToScanQueue* _queue;
818 Thread* _thread;
819 ShenandoahTraversalGC* _traversal_gc;
820 ShenandoahMarkingContext* const _mark_context;
821
822 template <class T>
823 inline void do_oop_work(T* p) {
824 ShenandoahEvacOOMScope evac_scope;
825 _traversal_gc->process_oop<T, false /* string dedup */, true /* degen */>(p, _thread, _queue, _mark_context);
826 }
827
828public:
829 ShenandoahTraversalSingleThreadKeepAliveUpdateDegenClosure(ShenandoahObjToScanQueue* q) :
830 _queue(q), _thread(Thread::current()),
831 _traversal_gc(ShenandoahHeap::heap()->traversal_gc()),
832 _mark_context(ShenandoahHeap::heap()->marking_context()) {}
833
834 void do_oop(narrowOop* p) { do_oop_work(p); }
835 void do_oop(oop* p) { do_oop_work(p); }
836};
837
838class ShenandoahTraversalPrecleanTask : public AbstractGangTask {
839private:
840 ReferenceProcessor* _rp;
841
842public:
843 ShenandoahTraversalPrecleanTask(ReferenceProcessor* rp) :
844 AbstractGangTask("Precleaning task"),
845 _rp(rp) {}
846
847 void work(uint worker_id) {
848 assert(worker_id == 0, "The code below is single-threaded, only one worker is expected");
849 ShenandoahParallelWorkerSession worker_session(worker_id);
850 ShenandoahSuspendibleThreadSetJoiner stsj(ShenandoahSuspendibleWorkers);
851 ShenandoahEvacOOMScope oom_evac_scope;
852
853 ShenandoahHeap* sh = ShenandoahHeap::heap();
854
855 ShenandoahObjToScanQueue* q = sh->traversal_gc()->task_queues()->queue(worker_id);
856
857 ShenandoahForwardedIsAliveClosure is_alive;
858 ShenandoahTraversalCancelledGCYieldClosure yield;
859 ShenandoahTraversalPrecleanCompleteGCClosure complete_gc;
860 ShenandoahTraversalKeepAliveUpdateClosure keep_alive(q);
861 ResourceMark rm;
862 _rp->preclean_discovered_references(&is_alive, &keep_alive,
863 &complete_gc, &yield,
864 NULL);
865 }
866};
867
868void ShenandoahTraversalGC::preclean_weak_refs() {
869 // Pre-cleaning weak references before diving into STW makes sense at the
870 // end of concurrent mark. This will filter out the references which referents
871 // are alive. Note that ReferenceProcessor already filters out these on reference
872 // discovery, and the bulk of work is done here. This phase processes leftovers
873 // that missed the initial filtering, i.e. when referent was marked alive after
874 // reference was discovered by RP.
875
876 assert(_heap->process_references(), "sanity");
877 assert(!_heap->is_degenerated_gc_in_progress(), "must be in concurrent non-degenerated phase");
878
879 // Shortcut if no references were discovered to avoid winding up threads.
880 ReferenceProcessor* rp = _heap->ref_processor();
881 if (!rp->has_discovered_references()) {
882 return;
883 }
884
885 ReferenceProcessorMTDiscoveryMutator fix_mt_discovery(rp, false);
886
887 shenandoah_assert_rp_isalive_not_installed();
888 ShenandoahForwardedIsAliveClosure is_alive;
889 ReferenceProcessorIsAliveMutator fix_isalive(rp, &is_alive);
890
891 assert(task_queues()->is_empty(), "Should be empty");
892
893 // Execute precleaning in the worker thread: it will give us GCLABs, String dedup
894 // queues and other goodies. When upstream ReferenceProcessor starts supporting
895 // parallel precleans, we can extend this to more threads.
896 ShenandoahPushWorkerScope scope(_heap->workers(), 1, /* check_workers = */ false);
897
898 WorkGang* workers = _heap->workers();
899 uint nworkers = workers->active_workers();
900 assert(nworkers == 1, "This code uses only a single worker");
901 task_queues()->reserve(nworkers);
902
903 ShenandoahTraversalPrecleanTask task(rp);
904 workers->run_task(&task);
905
906 assert(_heap->cancelled_gc() || task_queues()->is_empty(), "Should be empty");
907}
908
909// Weak Reference Closures
910class ShenandoahTraversalDrainMarkingStackClosure: public VoidClosure {
911 uint _worker_id;
912 ShenandoahTaskTerminator* _terminator;
913 bool _reset_terminator;
914
915public:
916 ShenandoahTraversalDrainMarkingStackClosure(uint worker_id, ShenandoahTaskTerminator* t, bool reset_terminator = false):
917 _worker_id(worker_id),
918 _terminator(t),
919 _reset_terminator(reset_terminator) {
920 }
921
922 void do_void() {
923 assert(ShenandoahSafepoint::is_at_shenandoah_safepoint(), "Must be at a safepoint");
924
925 ShenandoahHeap* sh = ShenandoahHeap::heap();
926 ShenandoahTraversalGC* traversal_gc = sh->traversal_gc();
927 assert(sh->process_references(), "why else would we be here?");
928 shenandoah_assert_rp_isalive_installed();
929
930 traversal_gc->main_loop(_worker_id, _terminator, false);
931
932 if (_reset_terminator) {
933 _terminator->reset_for_reuse();
934 }
935 }
936};
937
938class ShenandoahTraversalSingleThreadedDrainMarkingStackClosure: public VoidClosure {
939 uint _worker_id;
940 ShenandoahTaskTerminator* _terminator;
941 bool _reset_terminator;
942
943public:
944 ShenandoahTraversalSingleThreadedDrainMarkingStackClosure(uint worker_id, ShenandoahTaskTerminator* t, bool reset_terminator = false):
945 _worker_id(worker_id),
946 _terminator(t),
947 _reset_terminator(reset_terminator) {
948 }
949
950 void do_void() {
951 assert(ShenandoahSafepoint::is_at_shenandoah_safepoint(), "Must be at a safepoint");
952
953 ShenandoahHeap* sh = ShenandoahHeap::heap();
954 ShenandoahTraversalGC* traversal_gc = sh->traversal_gc();
955 assert(sh->process_references(), "why else would we be here?");
956 shenandoah_assert_rp_isalive_installed();
957
958 ShenandoahEvacOOMScope evac_scope;
959 traversal_gc->main_loop(_worker_id, _terminator, false);
960
961 if (_reset_terminator) {
962 _terminator->reset_for_reuse();
963 }
964 }
965};
966
967void ShenandoahTraversalGC::weak_refs_work() {
968 assert(_heap->process_references(), "sanity");
969
970 ShenandoahPhaseTimings::Phase phase_root = ShenandoahPhaseTimings::weakrefs;
971
972 ShenandoahGCPhase phase(phase_root);
973
974 ReferenceProcessor* rp = _heap->ref_processor();
975
976 // NOTE: We cannot shortcut on has_discovered_references() here, because
977 // we will miss marking JNI Weak refs then, see implementation in
978 // ReferenceProcessor::process_discovered_references.
979 weak_refs_work_doit();
980
981 rp->verify_no_references_recorded();
982 assert(!rp->discovery_enabled(), "Post condition");
983
984}
985
986class ShenandoahTraversalRefProcTaskProxy : public AbstractGangTask {
987private:
988 AbstractRefProcTaskExecutor::ProcessTask& _proc_task;
989 ShenandoahTaskTerminator* _terminator;
990
991public:
992 ShenandoahTraversalRefProcTaskProxy(AbstractRefProcTaskExecutor::ProcessTask& proc_task,
993 ShenandoahTaskTerminator* t) :
994 AbstractGangTask("Process reference objects in parallel"),
995 _proc_task(proc_task),
996 _terminator(t) {
997 }
998
999 void work(uint worker_id) {
1000 ShenandoahEvacOOMScope oom_evac_scope;
1001 assert(ShenandoahSafepoint::is_at_shenandoah_safepoint(), "Must be at a safepoint");
1002 ShenandoahHeap* heap = ShenandoahHeap::heap();
1003 ShenandoahTraversalDrainMarkingStackClosure complete_gc(worker_id, _terminator);
1004
1005 ShenandoahForwardedIsAliveClosure is_alive;
1006 if (!heap->is_degenerated_gc_in_progress()) {
1007 ShenandoahTraversalKeepAliveUpdateClosure keep_alive(heap->traversal_gc()->task_queues()->queue(worker_id));
1008 _proc_task.work(worker_id, is_alive, keep_alive, complete_gc);
1009 } else {
1010 ShenandoahTraversalKeepAliveUpdateDegenClosure keep_alive(heap->traversal_gc()->task_queues()->queue(worker_id));
1011 _proc_task.work(worker_id, is_alive, keep_alive, complete_gc);
1012 }
1013 }
1014};
1015
1016class ShenandoahTraversalRefProcTaskExecutor : public AbstractRefProcTaskExecutor {
1017private:
1018 WorkGang* _workers;
1019
1020public:
1021 ShenandoahTraversalRefProcTaskExecutor(WorkGang* workers) : _workers(workers) {}
1022
1023 // Executes a task using worker threads.
1024 void execute(ProcessTask& task, uint ergo_workers) {
1025 assert(ShenandoahSafepoint::is_at_shenandoah_safepoint(), "Must be at a safepoint");
1026
1027 ShenandoahHeap* heap = ShenandoahHeap::heap();
1028 ShenandoahTraversalGC* traversal_gc = heap->traversal_gc();
1029 ShenandoahPushWorkerQueuesScope scope(_workers,
1030 traversal_gc->task_queues(),
1031 ergo_workers,
1032 /* do_check = */ false);
1033 uint nworkers = _workers->active_workers();
1034 traversal_gc->task_queues()->reserve(nworkers);
1035 ShenandoahTaskTerminator terminator(nworkers, traversal_gc->task_queues());
1036 ShenandoahTraversalRefProcTaskProxy proc_task_proxy(task, &terminator);
1037 _workers->run_task(&proc_task_proxy);
1038 }
1039};
1040
1041void ShenandoahTraversalGC::weak_refs_work_doit() {
1042 ReferenceProcessor* rp = _heap->ref_processor();
1043
1044 ShenandoahPhaseTimings::Phase phase_process = ShenandoahPhaseTimings::weakrefs_process;
1045
1046 shenandoah_assert_rp_isalive_not_installed();
1047 ShenandoahForwardedIsAliveClosure is_alive;
1048 ReferenceProcessorIsAliveMutator fix_isalive(rp, &is_alive);
1049
1050 WorkGang* workers = _heap->workers();
1051 uint nworkers = workers->active_workers();
1052
1053 rp->setup_policy(_heap->soft_ref_policy()->should_clear_all_soft_refs());
1054 rp->set_active_mt_degree(nworkers);
1055
1056 assert(task_queues()->is_empty(), "Should be empty");
1057
1058 // complete_gc and keep_alive closures instantiated here are only needed for
1059 // single-threaded path in RP. They share the queue 0 for tracking work, which
1060 // simplifies implementation. Since RP may decide to call complete_gc several
1061 // times, we need to be able to reuse the terminator.
1062 uint serial_worker_id = 0;
1063 ShenandoahTaskTerminator terminator(1, task_queues());
1064 ShenandoahTraversalSingleThreadedDrainMarkingStackClosure complete_gc(serial_worker_id, &terminator, /* reset_terminator = */ true);
1065 ShenandoahPushWorkerQueuesScope scope(workers, task_queues(), 1, /* do_check = */ false);
1066
1067 ShenandoahTraversalRefProcTaskExecutor executor(workers);
1068
1069 ReferenceProcessorPhaseTimes pt(_heap->gc_timer(), rp->num_queues());
1070 if (!_heap->is_degenerated_gc_in_progress()) {
1071 ShenandoahTraversalSingleThreadKeepAliveUpdateClosure keep_alive(task_queues()->queue(serial_worker_id));
1072 rp->process_discovered_references(&is_alive, &keep_alive,
1073 &complete_gc, &executor,
1074 &pt);
1075 } else {
1076 ShenandoahTraversalSingleThreadKeepAliveUpdateDegenClosure keep_alive(task_queues()->queue(serial_worker_id));
1077 rp->process_discovered_references(&is_alive, &keep_alive,
1078 &complete_gc, &executor,
1079 &pt);
1080 }
1081
1082 pt.print_all_references();
1083 assert(task_queues()->is_empty() || _heap->cancelled_gc(), "Should be empty");
1084}
1085