| 1 | /* |
| 2 | * Copyright (c) 2002, 2019, Oracle and/or its affiliates. All rights reserved. |
| 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
| 4 | * |
| 5 | * This code is free software; you can redistribute it and/or modify it |
| 6 | * under the terms of the GNU General Public License version 2 only, as |
| 7 | * published by the Free Software Foundation. |
| 8 | * |
| 9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
| 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| 11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
| 12 | * version 2 for more details (a copy is included in the LICENSE file that |
| 13 | * accompanied this code). |
| 14 | * |
| 15 | * You should have received a copy of the GNU General Public License version |
| 16 | * 2 along with this work; if not, write to the Free Software Foundation, |
| 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
| 18 | * |
| 19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
| 20 | * or visit www.oracle.com if you need additional information or have any |
| 21 | * questions. |
| 22 | * |
| 23 | */ |
| 24 | |
| 25 | #ifndef SHARE_GC_PARALLEL_PSPROMOTIONMANAGER_INLINE_HPP |
| 26 | #define SHARE_GC_PARALLEL_PSPROMOTIONMANAGER_INLINE_HPP |
| 27 | |
| 28 | #include "gc/parallel/parallelScavengeHeap.hpp" |
| 29 | #include "gc/parallel/parMarkBitMap.inline.hpp" |
| 30 | #include "gc/parallel/psOldGen.hpp" |
| 31 | #include "gc/parallel/psPromotionLAB.inline.hpp" |
| 32 | #include "gc/parallel/psPromotionManager.hpp" |
| 33 | #include "gc/parallel/psScavenge.inline.hpp" |
| 34 | #include "gc/shared/taskqueue.inline.hpp" |
| 35 | #include "logging/log.hpp" |
| 36 | #include "memory/iterator.inline.hpp" |
| 37 | #include "oops/access.inline.hpp" |
| 38 | #include "oops/oop.inline.hpp" |
| 39 | |
| 40 | inline PSPromotionManager* PSPromotionManager::manager_array(uint index) { |
| 41 | assert(_manager_array != NULL, "access of NULL manager_array" ); |
| 42 | assert(index <= ParallelGCThreads, "out of range manager_array access" ); |
| 43 | return &_manager_array[index]; |
| 44 | } |
| 45 | |
| 46 | template <class T> |
| 47 | inline void PSPromotionManager::push_depth(T* p) { |
| 48 | claimed_stack_depth()->push(p); |
| 49 | } |
| 50 | |
| 51 | template <class T> |
| 52 | inline void PSPromotionManager::claim_or_forward_internal_depth(T* p) { |
| 53 | if (p != NULL) { // XXX: error if p != NULL here |
| 54 | oop o = RawAccess<IS_NOT_NULL>::oop_load(p); |
| 55 | if (o->is_forwarded()) { |
| 56 | o = o->forwardee(); |
| 57 | // Card mark |
| 58 | if (PSScavenge::is_obj_in_young(o)) { |
| 59 | PSScavenge::card_table()->inline_write_ref_field_gc(p, o); |
| 60 | } |
| 61 | RawAccess<IS_NOT_NULL>::oop_store(p, o); |
| 62 | } else { |
| 63 | push_depth(p); |
| 64 | } |
| 65 | } |
| 66 | } |
| 67 | |
| 68 | template <class T> |
| 69 | inline void PSPromotionManager::claim_or_forward_depth(T* p) { |
| 70 | assert(should_scavenge(p, true), "revisiting object?" ); |
| 71 | assert(ParallelScavengeHeap::heap()->is_in(p), "pointer outside heap" ); |
| 72 | |
| 73 | claim_or_forward_internal_depth(p); |
| 74 | } |
| 75 | |
| 76 | inline void PSPromotionManager::promotion_trace_event(oop new_obj, oop old_obj, |
| 77 | size_t obj_size, |
| 78 | uint age, bool tenured, |
| 79 | const PSPromotionLAB* lab) { |
| 80 | // Skip if memory allocation failed |
| 81 | if (new_obj != NULL) { |
| 82 | const ParallelScavengeTracer* gc_tracer = PSScavenge::gc_tracer(); |
| 83 | |
| 84 | if (lab != NULL) { |
| 85 | // Promotion of object through newly allocated PLAB |
| 86 | if (gc_tracer->should_report_promotion_in_new_plab_event()) { |
| 87 | size_t obj_bytes = obj_size * HeapWordSize; |
| 88 | size_t lab_size = lab->capacity(); |
| 89 | gc_tracer->report_promotion_in_new_plab_event(old_obj->klass(), obj_bytes, |
| 90 | age, tenured, lab_size); |
| 91 | } |
| 92 | } else { |
| 93 | // Promotion of object directly to heap |
| 94 | if (gc_tracer->should_report_promotion_outside_plab_event()) { |
| 95 | size_t obj_bytes = obj_size * HeapWordSize; |
| 96 | gc_tracer->report_promotion_outside_plab_event(old_obj->klass(), obj_bytes, |
| 97 | age, tenured); |
| 98 | } |
| 99 | } |
| 100 | } |
| 101 | } |
| 102 | |
| 103 | class PSPushContentsClosure: public BasicOopIterateClosure { |
| 104 | PSPromotionManager* _pm; |
| 105 | public: |
| 106 | PSPushContentsClosure(PSPromotionManager* pm) : BasicOopIterateClosure(PSScavenge::reference_processor()), _pm(pm) {} |
| 107 | |
| 108 | template <typename T> void do_oop_nv(T* p) { |
| 109 | if (PSScavenge::should_scavenge(p)) { |
| 110 | _pm->claim_or_forward_depth(p); |
| 111 | } |
| 112 | } |
| 113 | |
| 114 | virtual void do_oop(oop* p) { do_oop_nv(p); } |
| 115 | virtual void do_oop(narrowOop* p) { do_oop_nv(p); } |
| 116 | |
| 117 | // Don't use the oop verification code in the oop_oop_iterate framework. |
| 118 | debug_only(virtual bool should_verify_oops() { return false; }) |
| 119 | }; |
| 120 | |
| 121 | // |
| 122 | // This closure specialization will override the one that is defined in |
| 123 | // instanceRefKlass.inline.cpp. It swaps the order of oop_oop_iterate and |
| 124 | // oop_oop_iterate_ref_processing. Unfortunately G1 and Parallel behaves |
| 125 | // significantly better (especially in the Derby benchmark) using opposite |
| 126 | // order of these function calls. |
| 127 | // |
| 128 | template <> |
| 129 | inline void InstanceRefKlass::oop_oop_iterate_reverse<oop, PSPushContentsClosure>(oop obj, PSPushContentsClosure* closure) { |
| 130 | oop_oop_iterate_ref_processing<oop>(obj, closure); |
| 131 | InstanceKlass::oop_oop_iterate_reverse<oop>(obj, closure); |
| 132 | } |
| 133 | |
| 134 | template <> |
| 135 | inline void InstanceRefKlass::oop_oop_iterate_reverse<narrowOop, PSPushContentsClosure>(oop obj, PSPushContentsClosure* closure) { |
| 136 | oop_oop_iterate_ref_processing<narrowOop>(obj, closure); |
| 137 | InstanceKlass::oop_oop_iterate_reverse<narrowOop>(obj, closure); |
| 138 | } |
| 139 | |
| 140 | inline void PSPromotionManager::push_contents(oop obj) { |
| 141 | if (!obj->klass()->is_typeArray_klass()) { |
| 142 | PSPushContentsClosure pcc(this); |
| 143 | obj->oop_iterate_backwards(&pcc); |
| 144 | } |
| 145 | } |
| 146 | // |
| 147 | // This method is pretty bulky. It would be nice to split it up |
| 148 | // into smaller submethods, but we need to be careful not to hurt |
| 149 | // performance. |
| 150 | // |
| 151 | template<bool promote_immediately> |
| 152 | inline oop PSPromotionManager::copy_to_survivor_space(oop o) { |
| 153 | assert(should_scavenge(&o), "Sanity" ); |
| 154 | |
| 155 | oop new_obj = NULL; |
| 156 | |
| 157 | // NOTE! We must be very careful with any methods that access the mark |
| 158 | // in o. There may be multiple threads racing on it, and it may be forwarded |
| 159 | // at any time. Do not use oop methods for accessing the mark! |
| 160 | markOop test_mark = o->mark_raw(); |
| 161 | |
| 162 | // The same test as "o->is_forwarded()" |
| 163 | if (!test_mark->is_marked()) { |
| 164 | bool new_obj_is_tenured = false; |
| 165 | size_t new_obj_size = o->size(); |
| 166 | |
| 167 | // Find the objects age, MT safe. |
| 168 | uint age = (test_mark->has_displaced_mark_helper() /* o->has_displaced_mark() */) ? |
| 169 | test_mark->displaced_mark_helper()->age() : test_mark->age(); |
| 170 | |
| 171 | if (!promote_immediately) { |
| 172 | // Try allocating obj in to-space (unless too old) |
| 173 | if (age < PSScavenge::tenuring_threshold()) { |
| 174 | new_obj = (oop) _young_lab.allocate(new_obj_size); |
| 175 | if (new_obj == NULL && !_young_gen_is_full) { |
| 176 | // Do we allocate directly, or flush and refill? |
| 177 | if (new_obj_size > (YoungPLABSize / 2)) { |
| 178 | // Allocate this object directly |
| 179 | new_obj = (oop)young_space()->cas_allocate(new_obj_size); |
| 180 | promotion_trace_event(new_obj, o, new_obj_size, age, false, NULL); |
| 181 | } else { |
| 182 | // Flush and fill |
| 183 | _young_lab.flush(); |
| 184 | |
| 185 | HeapWord* lab_base = young_space()->cas_allocate(YoungPLABSize); |
| 186 | if (lab_base != NULL) { |
| 187 | _young_lab.initialize(MemRegion(lab_base, YoungPLABSize)); |
| 188 | // Try the young lab allocation again. |
| 189 | new_obj = (oop) _young_lab.allocate(new_obj_size); |
| 190 | promotion_trace_event(new_obj, o, new_obj_size, age, false, &_young_lab); |
| 191 | } else { |
| 192 | _young_gen_is_full = true; |
| 193 | } |
| 194 | } |
| 195 | } |
| 196 | } |
| 197 | } |
| 198 | |
| 199 | // Otherwise try allocating obj tenured |
| 200 | if (new_obj == NULL) { |
| 201 | #ifndef PRODUCT |
| 202 | if (ParallelScavengeHeap::heap()->promotion_should_fail()) { |
| 203 | return oop_promotion_failed(o, test_mark); |
| 204 | } |
| 205 | #endif // #ifndef PRODUCT |
| 206 | |
| 207 | new_obj = (oop) _old_lab.allocate(new_obj_size); |
| 208 | new_obj_is_tenured = true; |
| 209 | |
| 210 | if (new_obj == NULL) { |
| 211 | if (!_old_gen_is_full) { |
| 212 | // Do we allocate directly, or flush and refill? |
| 213 | if (new_obj_size > (OldPLABSize / 2)) { |
| 214 | // Allocate this object directly |
| 215 | new_obj = (oop)old_gen()->cas_allocate(new_obj_size); |
| 216 | promotion_trace_event(new_obj, o, new_obj_size, age, true, NULL); |
| 217 | } else { |
| 218 | // Flush and fill |
| 219 | _old_lab.flush(); |
| 220 | |
| 221 | HeapWord* lab_base = old_gen()->cas_allocate(OldPLABSize); |
| 222 | if(lab_base != NULL) { |
| 223 | #ifdef ASSERT |
| 224 | // Delay the initialization of the promotion lab (plab). |
| 225 | // This exposes uninitialized plabs to card table processing. |
| 226 | if (GCWorkerDelayMillis > 0) { |
| 227 | os::sleep(Thread::current(), GCWorkerDelayMillis, false); |
| 228 | } |
| 229 | #endif |
| 230 | _old_lab.initialize(MemRegion(lab_base, OldPLABSize)); |
| 231 | // Try the old lab allocation again. |
| 232 | new_obj = (oop) _old_lab.allocate(new_obj_size); |
| 233 | promotion_trace_event(new_obj, o, new_obj_size, age, true, &_old_lab); |
| 234 | } |
| 235 | } |
| 236 | } |
| 237 | |
| 238 | // This is the promotion failed test, and code handling. |
| 239 | // The code belongs here for two reasons. It is slightly |
| 240 | // different than the code below, and cannot share the |
| 241 | // CAS testing code. Keeping the code here also minimizes |
| 242 | // the impact on the common case fast path code. |
| 243 | |
| 244 | if (new_obj == NULL) { |
| 245 | _old_gen_is_full = true; |
| 246 | return oop_promotion_failed(o, test_mark); |
| 247 | } |
| 248 | } |
| 249 | } |
| 250 | |
| 251 | assert(new_obj != NULL, "allocation should have succeeded" ); |
| 252 | |
| 253 | // Copy obj |
| 254 | Copy::aligned_disjoint_words((HeapWord*)o, (HeapWord*)new_obj, new_obj_size); |
| 255 | |
| 256 | // Now we have to CAS in the header. |
| 257 | // Make copy visible to threads reading the forwardee. |
| 258 | if (o->cas_forward_to(new_obj, test_mark, memory_order_release)) { |
| 259 | // We won any races, we "own" this object. |
| 260 | assert(new_obj == o->forwardee(), "Sanity" ); |
| 261 | |
| 262 | // Increment age if obj still in new generation. Now that |
| 263 | // we're dealing with a markOop that cannot change, it is |
| 264 | // okay to use the non mt safe oop methods. |
| 265 | if (!new_obj_is_tenured) { |
| 266 | new_obj->incr_age(); |
| 267 | assert(young_space()->contains(new_obj), "Attempt to push non-promoted obj" ); |
| 268 | } |
| 269 | |
| 270 | // Do the size comparison first with new_obj_size, which we |
| 271 | // already have. Hopefully, only a few objects are larger than |
| 272 | // _min_array_size_for_chunking, and most of them will be arrays. |
| 273 | // So, the is->objArray() test would be very infrequent. |
| 274 | if (new_obj_size > _min_array_size_for_chunking && |
| 275 | new_obj->is_objArray() && |
| 276 | PSChunkLargeArrays) { |
| 277 | // we'll chunk it |
| 278 | oop* const masked_o = mask_chunked_array_oop(o); |
| 279 | push_depth(masked_o); |
| 280 | TASKQUEUE_STATS_ONLY(++_arrays_chunked; ++_masked_pushes); |
| 281 | } else { |
| 282 | // we'll just push its contents |
| 283 | push_contents(new_obj); |
| 284 | } |
| 285 | } else { |
| 286 | // We lost, someone else "owns" this object |
| 287 | guarantee(o->is_forwarded(), "Object must be forwarded if the cas failed." ); |
| 288 | |
| 289 | // Try to deallocate the space. If it was directly allocated we cannot |
| 290 | // deallocate it, so we have to test. If the deallocation fails, |
| 291 | // overwrite with a filler object. |
| 292 | if (new_obj_is_tenured) { |
| 293 | if (!_old_lab.unallocate_object((HeapWord*) new_obj, new_obj_size)) { |
| 294 | CollectedHeap::fill_with_object((HeapWord*) new_obj, new_obj_size); |
| 295 | } |
| 296 | } else if (!_young_lab.unallocate_object((HeapWord*) new_obj, new_obj_size)) { |
| 297 | CollectedHeap::fill_with_object((HeapWord*) new_obj, new_obj_size); |
| 298 | } |
| 299 | |
| 300 | // don't update this before the unallocation! |
| 301 | // Using acquire though consume would be accurate for accessing new_obj. |
| 302 | new_obj = o->forwardee_acquire(); |
| 303 | } |
| 304 | } else { |
| 305 | assert(o->is_forwarded(), "Sanity" ); |
| 306 | new_obj = o->forwardee_acquire(); |
| 307 | } |
| 308 | |
| 309 | // This code must come after the CAS test, or it will print incorrect |
| 310 | // information. |
| 311 | log_develop_trace(gc, scavenge)("{%s %s " PTR_FORMAT " -> " PTR_FORMAT " (%d)}" , |
| 312 | should_scavenge(&new_obj) ? "copying" : "tenuring" , |
| 313 | new_obj->klass()->internal_name(), p2i((void *)o), p2i((void *)new_obj), new_obj->size()); |
| 314 | |
| 315 | return new_obj; |
| 316 | } |
| 317 | |
| 318 | // Attempt to "claim" oop at p via CAS, push the new obj if successful |
| 319 | // This version tests the oop* to make sure it is within the heap before |
| 320 | // attempting marking. |
| 321 | template <class T, bool promote_immediately> |
| 322 | inline void PSPromotionManager::copy_and_push_safe_barrier(T* p) { |
| 323 | assert(should_scavenge(p, true), "revisiting object?" ); |
| 324 | |
| 325 | oop o = RawAccess<IS_NOT_NULL>::oop_load(p); |
| 326 | oop new_obj = o->is_forwarded() |
| 327 | ? o->forwardee() |
| 328 | : copy_to_survivor_space<promote_immediately>(o); |
| 329 | |
| 330 | // This code must come after the CAS test, or it will print incorrect |
| 331 | // information. |
| 332 | if (log_develop_is_enabled(Trace, gc, scavenge) && o->is_forwarded()) { |
| 333 | log_develop_trace(gc, scavenge)("{%s %s " PTR_FORMAT " -> " PTR_FORMAT " (%d)}" , |
| 334 | "forwarding" , |
| 335 | new_obj->klass()->internal_name(), p2i((void *)o), p2i((void *)new_obj), new_obj->size()); |
| 336 | } |
| 337 | |
| 338 | RawAccess<IS_NOT_NULL>::oop_store(p, new_obj); |
| 339 | |
| 340 | // We cannot mark without test, as some code passes us pointers |
| 341 | // that are outside the heap. These pointers are either from roots |
| 342 | // or from metadata. |
| 343 | if ((!PSScavenge::is_obj_in_young((HeapWord*)p)) && |
| 344 | ParallelScavengeHeap::heap()->is_in_reserved(p)) { |
| 345 | if (PSScavenge::is_obj_in_young(new_obj)) { |
| 346 | PSScavenge::card_table()->inline_write_ref_field_gc(p, new_obj); |
| 347 | } |
| 348 | } |
| 349 | } |
| 350 | |
| 351 | inline void PSPromotionManager::process_popped_location_depth(StarTask p) { |
| 352 | if (is_oop_masked(p)) { |
| 353 | assert(PSChunkLargeArrays, "invariant" ); |
| 354 | oop const old = unmask_chunked_array_oop(p); |
| 355 | process_array_chunk(old); |
| 356 | } else { |
| 357 | if (p.is_narrow()) { |
| 358 | assert(UseCompressedOops, "Error" ); |
| 359 | copy_and_push_safe_barrier<narrowOop, /*promote_immediately=*/false>(p); |
| 360 | } else { |
| 361 | copy_and_push_safe_barrier<oop, /*promote_immediately=*/false>(p); |
| 362 | } |
| 363 | } |
| 364 | } |
| 365 | |
| 366 | inline bool PSPromotionManager::steal_depth(int queue_num, StarTask& t) { |
| 367 | return stack_array_depth()->steal(queue_num, t); |
| 368 | } |
| 369 | |
| 370 | #if TASKQUEUE_STATS |
| 371 | void PSPromotionManager::record_steal(StarTask& p) { |
| 372 | if (is_oop_masked(p)) { |
| 373 | ++_masked_steals; |
| 374 | } |
| 375 | } |
| 376 | #endif // TASKQUEUE_STATS |
| 377 | |
| 378 | #endif // SHARE_GC_PARALLEL_PSPROMOTIONMANAGER_INLINE_HPP |
| 379 | |