| 1 | /* |
| 2 | * Copyright (c) 2000, 2019, Oracle and/or its affiliates. All rights reserved. |
| 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
| 4 | * |
| 5 | * This code is free software; you can redistribute it and/or modify it |
| 6 | * under the terms of the GNU General Public License version 2 only, as |
| 7 | * published by the Free Software Foundation. |
| 8 | * |
| 9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
| 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| 11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
| 12 | * version 2 for more details (a copy is included in the LICENSE file that |
| 13 | * accompanied this code). |
| 14 | * |
| 15 | * You should have received a copy of the GNU General Public License version |
| 16 | * 2 along with this work; if not, write to the Free Software Foundation, |
| 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
| 18 | * |
| 19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
| 20 | * or visit www.oracle.com if you need additional information or have any |
| 21 | * questions. |
| 22 | * |
| 23 | */ |
| 24 | |
| 25 | #ifndef SHARE_GC_SHARED_GENCOLLECTEDHEAP_HPP |
| 26 | #define SHARE_GC_SHARED_GENCOLLECTEDHEAP_HPP |
| 27 | |
| 28 | #include "gc/shared/collectedHeap.hpp" |
| 29 | #include "gc/shared/generation.hpp" |
| 30 | #include "gc/shared/oopStorageParState.hpp" |
| 31 | #include "gc/shared/softRefGenPolicy.hpp" |
| 32 | |
| 33 | class AdaptiveSizePolicy; |
| 34 | class CardTableRS; |
| 35 | class GCPolicyCounters; |
| 36 | class GenerationSpec; |
| 37 | class StrongRootsScope; |
| 38 | class SubTasksDone; |
| 39 | class WorkGang; |
| 40 | |
| 41 | // A "GenCollectedHeap" is a CollectedHeap that uses generational |
| 42 | // collection. It has two generations, young and old. |
| 43 | class GenCollectedHeap : public CollectedHeap { |
| 44 | friend class Generation; |
| 45 | friend class DefNewGeneration; |
| 46 | friend class TenuredGeneration; |
| 47 | friend class ConcurrentMarkSweepGeneration; |
| 48 | friend class CMSCollector; |
| 49 | friend class GenMarkSweep; |
| 50 | friend class VM_GenCollectForAllocation; |
| 51 | friend class VM_GenCollectFull; |
| 52 | friend class VM_GenCollectFullConcurrent; |
| 53 | friend class VM_GC_HeapInspection; |
| 54 | friend class VM_HeapDumper; |
| 55 | friend class HeapInspection; |
| 56 | friend class GCCauseSetter; |
| 57 | friend class VMStructs; |
| 58 | public: |
| 59 | friend class VM_PopulateDumpSharedSpace; |
| 60 | |
| 61 | enum GenerationType { |
| 62 | YoungGen, |
| 63 | OldGen |
| 64 | }; |
| 65 | |
| 66 | protected: |
| 67 | Generation* _young_gen; |
| 68 | Generation* _old_gen; |
| 69 | |
| 70 | private: |
| 71 | GenerationSpec* _young_gen_spec; |
| 72 | GenerationSpec* _old_gen_spec; |
| 73 | |
| 74 | // The singleton CardTable Remembered Set. |
| 75 | CardTableRS* _rem_set; |
| 76 | |
| 77 | SoftRefGenPolicy _soft_ref_gen_policy; |
| 78 | |
| 79 | // The sizing of the heap is controlled by a sizing policy. |
| 80 | AdaptiveSizePolicy* _size_policy; |
| 81 | |
| 82 | GCPolicyCounters* _gc_policy_counters; |
| 83 | |
| 84 | // Indicates that the most recent previous incremental collection failed. |
| 85 | // The flag is cleared when an action is taken that might clear the |
| 86 | // condition that caused that incremental collection to fail. |
| 87 | bool _incremental_collection_failed; |
| 88 | |
| 89 | // In support of ExplicitGCInvokesConcurrent functionality |
| 90 | unsigned int _full_collections_completed; |
| 91 | |
| 92 | // Collects the given generation. |
| 93 | void collect_generation(Generation* gen, bool full, size_t size, bool is_tlab, |
| 94 | bool run_verification, bool clear_soft_refs, |
| 95 | bool restore_marks_for_biased_locking); |
| 96 | |
| 97 | // Reserve aligned space for the heap as needed by the contained generations. |
| 98 | char* allocate(size_t alignment, ReservedSpace* heap_rs); |
| 99 | |
| 100 | // Initialize ("weak") refs processing support |
| 101 | void ref_processing_init(); |
| 102 | |
| 103 | protected: |
| 104 | |
| 105 | // The set of potentially parallel tasks in root scanning. |
| 106 | enum GCH_strong_roots_tasks { |
| 107 | GCH_PS_Universe_oops_do, |
| 108 | GCH_PS_JNIHandles_oops_do, |
| 109 | GCH_PS_ObjectSynchronizer_oops_do, |
| 110 | GCH_PS_FlatProfiler_oops_do, |
| 111 | GCH_PS_Management_oops_do, |
| 112 | GCH_PS_SystemDictionary_oops_do, |
| 113 | GCH_PS_ClassLoaderDataGraph_oops_do, |
| 114 | GCH_PS_jvmti_oops_do, |
| 115 | GCH_PS_CodeCache_oops_do, |
| 116 | AOT_ONLY(GCH_PS_aot_oops_do COMMA) |
| 117 | JVMCI_ONLY(GCH_PS_jvmci_oops_do COMMA) |
| 118 | GCH_PS_younger_gens, |
| 119 | // Leave this one last. |
| 120 | GCH_PS_NumElements |
| 121 | }; |
| 122 | |
| 123 | // Data structure for claiming the (potentially) parallel tasks in |
| 124 | // (gen-specific) roots processing. |
| 125 | SubTasksDone* _process_strong_tasks; |
| 126 | |
| 127 | GCMemoryManager* _young_manager; |
| 128 | GCMemoryManager* _old_manager; |
| 129 | |
| 130 | // Helper functions for allocation |
| 131 | HeapWord* attempt_allocation(size_t size, |
| 132 | bool is_tlab, |
| 133 | bool first_only); |
| 134 | |
| 135 | // Helper function for two callbacks below. |
| 136 | // Considers collection of the first max_level+1 generations. |
| 137 | void do_collection(bool full, |
| 138 | bool clear_all_soft_refs, |
| 139 | size_t size, |
| 140 | bool is_tlab, |
| 141 | GenerationType max_generation); |
| 142 | |
| 143 | // Callback from VM_GenCollectForAllocation operation. |
| 144 | // This function does everything necessary/possible to satisfy an |
| 145 | // allocation request that failed in the youngest generation that should |
| 146 | // have handled it (including collection, expansion, etc.) |
| 147 | HeapWord* satisfy_failed_allocation(size_t size, bool is_tlab); |
| 148 | |
| 149 | // Callback from VM_GenCollectFull operation. |
| 150 | // Perform a full collection of the first max_level+1 generations. |
| 151 | virtual void do_full_collection(bool clear_all_soft_refs); |
| 152 | void do_full_collection(bool clear_all_soft_refs, GenerationType max_generation); |
| 153 | |
| 154 | // Does the "cause" of GC indicate that |
| 155 | // we absolutely __must__ clear soft refs? |
| 156 | bool must_clear_all_soft_refs(); |
| 157 | |
| 158 | GenCollectedHeap(Generation::Name young, |
| 159 | Generation::Name old, |
| 160 | const char* policy_counters_name); |
| 161 | |
| 162 | public: |
| 163 | |
| 164 | // Returns JNI_OK on success |
| 165 | virtual jint initialize(); |
| 166 | virtual CardTableRS* create_rem_set(const MemRegion& reserved_region); |
| 167 | |
| 168 | void initialize_size_policy(size_t init_eden_size, |
| 169 | size_t init_promo_size, |
| 170 | size_t init_survivor_size); |
| 171 | |
| 172 | // Does operations required after initialization has been done. |
| 173 | void post_initialize(); |
| 174 | |
| 175 | Generation* young_gen() const { return _young_gen; } |
| 176 | Generation* old_gen() const { return _old_gen; } |
| 177 | |
| 178 | bool is_young_gen(const Generation* gen) const { return gen == _young_gen; } |
| 179 | bool is_old_gen(const Generation* gen) const { return gen == _old_gen; } |
| 180 | |
| 181 | GenerationSpec* young_gen_spec() const; |
| 182 | GenerationSpec* old_gen_spec() const; |
| 183 | |
| 184 | virtual SoftRefPolicy* soft_ref_policy() { return &_soft_ref_gen_policy; } |
| 185 | |
| 186 | // Adaptive size policy |
| 187 | virtual AdaptiveSizePolicy* size_policy() { |
| 188 | return _size_policy; |
| 189 | } |
| 190 | |
| 191 | // Performance Counter support |
| 192 | GCPolicyCounters* counters() { return _gc_policy_counters; } |
| 193 | |
| 194 | size_t capacity() const; |
| 195 | size_t used() const; |
| 196 | |
| 197 | // Save the "used_region" for both generations. |
| 198 | void save_used_regions(); |
| 199 | |
| 200 | size_t max_capacity() const; |
| 201 | |
| 202 | HeapWord* mem_allocate(size_t size, bool* gc_overhead_limit_was_exceeded); |
| 203 | |
| 204 | // We may support a shared contiguous allocation area, if the youngest |
| 205 | // generation does. |
| 206 | bool supports_inline_contig_alloc() const; |
| 207 | HeapWord* volatile* top_addr() const; |
| 208 | HeapWord** end_addr() const; |
| 209 | |
| 210 | // Perform a full collection of the heap; intended for use in implementing |
| 211 | // "System.gc". This implies as full a collection as the CollectedHeap |
| 212 | // supports. Caller does not hold the Heap_lock on entry. |
| 213 | virtual void collect(GCCause::Cause cause); |
| 214 | |
| 215 | // The same as above but assume that the caller holds the Heap_lock. |
| 216 | void collect_locked(GCCause::Cause cause); |
| 217 | |
| 218 | // Perform a full collection of generations up to and including max_generation. |
| 219 | // Mostly used for testing purposes. Caller does not hold the Heap_lock on entry. |
| 220 | void collect(GCCause::Cause cause, GenerationType max_generation); |
| 221 | |
| 222 | // Returns "TRUE" iff "p" points into the committed areas of the heap. |
| 223 | // The methods is_in() and is_in_youngest() may be expensive to compute |
| 224 | // in general, so, to prevent their inadvertent use in product jvm's, we |
| 225 | // restrict their use to assertion checking or verification only. |
| 226 | bool is_in(const void* p) const; |
| 227 | |
| 228 | // Returns true if the reference is to an object in the reserved space |
| 229 | // for the young generation. |
| 230 | // Assumes the the young gen address range is less than that of the old gen. |
| 231 | bool is_in_young(oop p); |
| 232 | |
| 233 | #ifdef ASSERT |
| 234 | bool is_in_partial_collection(const void* p); |
| 235 | #endif |
| 236 | |
| 237 | // Optimized nmethod scanning support routines |
| 238 | virtual void register_nmethod(nmethod* nm); |
| 239 | virtual void unregister_nmethod(nmethod* nm); |
| 240 | virtual void verify_nmethod(nmethod* nm); |
| 241 | virtual void flush_nmethod(nmethod* nm); |
| 242 | |
| 243 | void prune_scavengable_nmethods(); |
| 244 | |
| 245 | // Iteration functions. |
| 246 | void oop_iterate(OopIterateClosure* cl); |
| 247 | void object_iterate(ObjectClosure* cl); |
| 248 | void safe_object_iterate(ObjectClosure* cl); |
| 249 | Space* space_containing(const void* addr) const; |
| 250 | |
| 251 | // A CollectedHeap is divided into a dense sequence of "blocks"; that is, |
| 252 | // each address in the (reserved) heap is a member of exactly |
| 253 | // one block. The defining characteristic of a block is that it is |
| 254 | // possible to find its size, and thus to progress forward to the next |
| 255 | // block. (Blocks may be of different sizes.) Thus, blocks may |
| 256 | // represent Java objects, or they might be free blocks in a |
| 257 | // free-list-based heap (or subheap), as long as the two kinds are |
| 258 | // distinguishable and the size of each is determinable. |
| 259 | |
| 260 | // Returns the address of the start of the "block" that contains the |
| 261 | // address "addr". We say "blocks" instead of "object" since some heaps |
| 262 | // may not pack objects densely; a chunk may either be an object or a |
| 263 | // non-object. |
| 264 | virtual HeapWord* block_start(const void* addr) const; |
| 265 | |
| 266 | // Requires "addr" to be the start of a block, and returns "TRUE" iff |
| 267 | // the block is an object. Assumes (and verifies in non-product |
| 268 | // builds) that addr is in the allocated part of the heap and is |
| 269 | // the start of a chunk. |
| 270 | virtual bool block_is_obj(const HeapWord* addr) const; |
| 271 | |
| 272 | // Section on TLAB's. |
| 273 | virtual bool supports_tlab_allocation() const; |
| 274 | virtual size_t tlab_capacity(Thread* thr) const; |
| 275 | virtual size_t tlab_used(Thread* thr) const; |
| 276 | virtual size_t unsafe_max_tlab_alloc(Thread* thr) const; |
| 277 | virtual HeapWord* allocate_new_tlab(size_t min_size, |
| 278 | size_t requested_size, |
| 279 | size_t* actual_size); |
| 280 | |
| 281 | // The "requestor" generation is performing some garbage collection |
| 282 | // action for which it would be useful to have scratch space. The |
| 283 | // requestor promises to allocate no more than "max_alloc_words" in any |
| 284 | // older generation (via promotion say.) Any blocks of space that can |
| 285 | // be provided are returned as a list of ScratchBlocks, sorted by |
| 286 | // decreasing size. |
| 287 | ScratchBlock* gather_scratch(Generation* requestor, size_t max_alloc_words); |
| 288 | // Allow each generation to reset any scratch space that it has |
| 289 | // contributed as it needs. |
| 290 | void release_scratch(); |
| 291 | |
| 292 | // Ensure parsability: override |
| 293 | virtual void ensure_parsability(bool retire_tlabs); |
| 294 | |
| 295 | // Time in ms since the longest time a collector ran in |
| 296 | // in any generation. |
| 297 | virtual jlong millis_since_last_gc(); |
| 298 | |
| 299 | // Total number of full collections completed. |
| 300 | unsigned int total_full_collections_completed() { |
| 301 | assert(_full_collections_completed <= _total_full_collections, |
| 302 | "Can't complete more collections than were started" ); |
| 303 | return _full_collections_completed; |
| 304 | } |
| 305 | |
| 306 | // Update above counter, as appropriate, at the end of a stop-world GC cycle |
| 307 | unsigned int update_full_collections_completed(); |
| 308 | // Update above counter, as appropriate, at the end of a concurrent GC cycle |
| 309 | unsigned int update_full_collections_completed(unsigned int count); |
| 310 | |
| 311 | // Update "time of last gc" for all generations to "now". |
| 312 | void update_time_of_last_gc(jlong now) { |
| 313 | _young_gen->update_time_of_last_gc(now); |
| 314 | _old_gen->update_time_of_last_gc(now); |
| 315 | } |
| 316 | |
| 317 | // Update the gc statistics for each generation. |
| 318 | void update_gc_stats(Generation* current_generation, bool full) { |
| 319 | _old_gen->update_gc_stats(current_generation, full); |
| 320 | } |
| 321 | |
| 322 | bool no_gc_in_progress() { return !is_gc_active(); } |
| 323 | |
| 324 | // Override. |
| 325 | void prepare_for_verify(); |
| 326 | |
| 327 | // Override. |
| 328 | void verify(VerifyOption option); |
| 329 | |
| 330 | // Override. |
| 331 | virtual void print_on(outputStream* st) const; |
| 332 | virtual void print_gc_threads_on(outputStream* st) const; |
| 333 | virtual void gc_threads_do(ThreadClosure* tc) const; |
| 334 | virtual void print_tracing_info() const; |
| 335 | |
| 336 | void print_heap_change(size_t young_prev_used, size_t old_prev_used) const; |
| 337 | |
| 338 | // The functions below are helper functions that a subclass of |
| 339 | // "CollectedHeap" can use in the implementation of its virtual |
| 340 | // functions. |
| 341 | |
| 342 | class GenClosure : public StackObj { |
| 343 | public: |
| 344 | virtual void do_generation(Generation* gen) = 0; |
| 345 | }; |
| 346 | |
| 347 | // Apply "cl.do_generation" to all generations in the heap |
| 348 | // If "old_to_young" determines the order. |
| 349 | void generation_iterate(GenClosure* cl, bool old_to_young); |
| 350 | |
| 351 | // Return "true" if all generations have reached the |
| 352 | // maximal committed limit that they can reach, without a garbage |
| 353 | // collection. |
| 354 | virtual bool is_maximal_no_gc() const; |
| 355 | |
| 356 | // This function returns the CardTableRS object that allows us to scan |
| 357 | // generations in a fully generational heap. |
| 358 | CardTableRS* rem_set() { return _rem_set; } |
| 359 | |
| 360 | // Convenience function to be used in situations where the heap type can be |
| 361 | // asserted to be this type. |
| 362 | static GenCollectedHeap* heap(); |
| 363 | |
| 364 | // The ScanningOption determines which of the roots |
| 365 | // the closure is applied to: |
| 366 | // "SO_None" does none; |
| 367 | enum ScanningOption { |
| 368 | SO_None = 0x0, |
| 369 | SO_AllCodeCache = 0x8, |
| 370 | SO_ScavengeCodeCache = 0x10 |
| 371 | }; |
| 372 | |
| 373 | protected: |
| 374 | void process_roots(StrongRootsScope* scope, |
| 375 | ScanningOption so, |
| 376 | OopClosure* strong_roots, |
| 377 | CLDClosure* strong_cld_closure, |
| 378 | CLDClosure* weak_cld_closure, |
| 379 | CodeBlobToOopClosure* code_roots); |
| 380 | |
| 381 | // Accessor for memory state verification support |
| 382 | NOT_PRODUCT( |
| 383 | virtual size_t skip_header_HeapWords() { return 0; } |
| 384 | ) |
| 385 | |
| 386 | virtual void gc_prologue(bool full); |
| 387 | virtual void gc_epilogue(bool full); |
| 388 | |
| 389 | public: |
| 390 | void young_process_roots(StrongRootsScope* scope, |
| 391 | OopsInGenClosure* root_closure, |
| 392 | OopsInGenClosure* old_gen_closure, |
| 393 | CLDClosure* cld_closure); |
| 394 | |
| 395 | void full_process_roots(StrongRootsScope* scope, |
| 396 | bool is_adjust_phase, |
| 397 | ScanningOption so, |
| 398 | bool only_strong_roots, |
| 399 | OopsInGenClosure* root_closure, |
| 400 | CLDClosure* cld_closure); |
| 401 | |
| 402 | // Apply "root_closure" to all the weak roots of the system. |
| 403 | // These include JNI weak roots, string table, |
| 404 | // and referents of reachable weak refs. |
| 405 | void gen_process_weak_roots(OopClosure* root_closure); |
| 406 | |
| 407 | // Set the saved marks of generations, if that makes sense. |
| 408 | // In particular, if any generation might iterate over the oops |
| 409 | // in other generations, it should call this method. |
| 410 | void save_marks(); |
| 411 | |
| 412 | // Returns "true" iff no allocations have occurred since the last |
| 413 | // call to "save_marks". |
| 414 | bool no_allocs_since_save_marks(); |
| 415 | |
| 416 | // Returns true if an incremental collection is likely to fail. |
| 417 | // We optionally consult the young gen, if asked to do so; |
| 418 | // otherwise we base our answer on whether the previous incremental |
| 419 | // collection attempt failed with no corrective action as of yet. |
| 420 | bool incremental_collection_will_fail(bool consult_young) { |
| 421 | // The first disjunct remembers if an incremental collection failed, even |
| 422 | // when we thought (second disjunct) that it would not. |
| 423 | return incremental_collection_failed() || |
| 424 | (consult_young && !_young_gen->collection_attempt_is_safe()); |
| 425 | } |
| 426 | |
| 427 | // If a generation bails out of an incremental collection, |
| 428 | // it sets this flag. |
| 429 | bool incremental_collection_failed() const { |
| 430 | return _incremental_collection_failed; |
| 431 | } |
| 432 | void set_incremental_collection_failed() { |
| 433 | _incremental_collection_failed = true; |
| 434 | } |
| 435 | void clear_incremental_collection_failed() { |
| 436 | _incremental_collection_failed = false; |
| 437 | } |
| 438 | |
| 439 | // Promotion of obj into gen failed. Try to promote obj to higher |
| 440 | // gens in ascending order; return the new location of obj if successful. |
| 441 | // Otherwise, try expand-and-allocate for obj in both the young and old |
| 442 | // generation; return the new location of obj if successful. Otherwise, return NULL. |
| 443 | oop handle_failed_promotion(Generation* old_gen, |
| 444 | oop obj, |
| 445 | size_t obj_size); |
| 446 | |
| 447 | |
| 448 | private: |
| 449 | // Return true if an allocation should be attempted in the older generation |
| 450 | // if it fails in the younger generation. Return false, otherwise. |
| 451 | bool should_try_older_generation_allocation(size_t word_size) const; |
| 452 | |
| 453 | // Try to allocate space by expanding the heap. |
| 454 | HeapWord* expand_heap_and_allocate(size_t size, bool is_tlab); |
| 455 | |
| 456 | HeapWord* mem_allocate_work(size_t size, |
| 457 | bool is_tlab, |
| 458 | bool* gc_overhead_limit_was_exceeded); |
| 459 | |
| 460 | // Override |
| 461 | void check_for_non_bad_heap_word_value(HeapWord* addr, |
| 462 | size_t size) PRODUCT_RETURN; |
| 463 | |
| 464 | #if INCLUDE_SERIALGC |
| 465 | // For use by mark-sweep. As implemented, mark-sweep-compact is global |
| 466 | // in an essential way: compaction is performed across generations, by |
| 467 | // iterating over spaces. |
| 468 | void prepare_for_compaction(); |
| 469 | #endif |
| 470 | |
| 471 | // Perform a full collection of the generations up to and including max_generation. |
| 472 | // This is the low level interface used by the public versions of |
| 473 | // collect() and collect_locked(). Caller holds the Heap_lock on entry. |
| 474 | void collect_locked(GCCause::Cause cause, GenerationType max_generation); |
| 475 | |
| 476 | // Save the tops of the spaces in all generations |
| 477 | void record_gen_tops_before_GC() PRODUCT_RETURN; |
| 478 | |
| 479 | // Return true if we need to perform full collection. |
| 480 | bool should_do_full_collection(size_t size, bool full, |
| 481 | bool is_tlab, GenerationType max_gen) const; |
| 482 | }; |
| 483 | |
| 484 | #endif // SHARE_GC_SHARED_GENCOLLECTEDHEAP_HPP |
| 485 | |