| 1 | /* |
| 2 | * Copyright (c) 2001, 2019, Oracle and/or its affiliates. All rights reserved. |
| 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
| 4 | * |
| 5 | * This code is free software; you can redistribute it and/or modify it |
| 6 | * under the terms of the GNU General Public License version 2 only, as |
| 7 | * published by the Free Software Foundation. |
| 8 | * |
| 9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
| 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| 11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
| 12 | * version 2 for more details (a copy is included in the LICENSE file that |
| 13 | * accompanied this code). |
| 14 | * |
| 15 | * You should have received a copy of the GNU General Public License version |
| 16 | * 2 along with this work; if not, write to the Free Software Foundation, |
| 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
| 18 | * |
| 19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
| 20 | * or visit www.oracle.com if you need additional information or have any |
| 21 | * questions. |
| 22 | * |
| 23 | */ |
| 24 | |
| 25 | #ifndef SHARE_GC_PARALLEL_PARALLELSCAVENGEHEAP_HPP |
| 26 | #define SHARE_GC_PARALLEL_PARALLELSCAVENGEHEAP_HPP |
| 27 | |
| 28 | #include "gc/parallel/objectStartArray.hpp" |
| 29 | #include "gc/parallel/psGCAdaptivePolicyCounters.hpp" |
| 30 | #include "gc/parallel/psOldGen.hpp" |
| 31 | #include "gc/parallel/psYoungGen.hpp" |
| 32 | #include "gc/shared/cardTableBarrierSet.hpp" |
| 33 | #include "gc/shared/collectedHeap.hpp" |
| 34 | #include "gc/shared/gcPolicyCounters.hpp" |
| 35 | #include "gc/shared/gcWhen.hpp" |
| 36 | #include "gc/shared/referenceProcessor.hpp" |
| 37 | #include "gc/shared/softRefPolicy.hpp" |
| 38 | #include "gc/shared/strongRootsScope.hpp" |
| 39 | #include "logging/log.hpp" |
| 40 | #include "memory/metaspace.hpp" |
| 41 | #include "utilities/growableArray.hpp" |
| 42 | #include "utilities/ostream.hpp" |
| 43 | |
| 44 | class AdjoiningGenerations; |
| 45 | class GCHeapSummary; |
| 46 | class GCTaskManager; |
| 47 | class MemoryManager; |
| 48 | class MemoryPool; |
| 49 | class PSAdaptiveSizePolicy; |
| 50 | class PSCardTable; |
| 51 | class PSHeapSummary; |
| 52 | |
| 53 | class ParallelScavengeHeap : public CollectedHeap { |
| 54 | friend class VMStructs; |
| 55 | private: |
| 56 | static PSYoungGen* _young_gen; |
| 57 | static PSOldGen* _old_gen; |
| 58 | |
| 59 | // Sizing policy for entire heap |
| 60 | static PSAdaptiveSizePolicy* _size_policy; |
| 61 | static PSGCAdaptivePolicyCounters* _gc_policy_counters; |
| 62 | |
| 63 | SoftRefPolicy _soft_ref_policy; |
| 64 | |
| 65 | // Collection of generations that are adjacent in the |
| 66 | // space reserved for the heap. |
| 67 | AdjoiningGenerations* _gens; |
| 68 | unsigned int _death_march_count; |
| 69 | |
| 70 | // The task manager |
| 71 | static GCTaskManager* _gc_task_manager; |
| 72 | |
| 73 | GCMemoryManager* _young_manager; |
| 74 | GCMemoryManager* _old_manager; |
| 75 | |
| 76 | MemoryPool* _eden_pool; |
| 77 | MemoryPool* _survivor_pool; |
| 78 | MemoryPool* _old_pool; |
| 79 | |
| 80 | virtual void initialize_serviceability(); |
| 81 | |
| 82 | void trace_heap(GCWhen::Type when, const GCTracer* tracer); |
| 83 | |
| 84 | protected: |
| 85 | static inline size_t total_invocations(); |
| 86 | HeapWord* allocate_new_tlab(size_t min_size, size_t requested_size, size_t* actual_size); |
| 87 | |
| 88 | inline bool should_alloc_in_eden(size_t size) const; |
| 89 | inline void death_march_check(HeapWord* const result, size_t size); |
| 90 | HeapWord* mem_allocate_old_gen(size_t size); |
| 91 | |
| 92 | public: |
| 93 | ParallelScavengeHeap() : |
| 94 | CollectedHeap(), |
| 95 | _gens(NULL), |
| 96 | _death_march_count(0), |
| 97 | _young_manager(NULL), |
| 98 | _old_manager(NULL), |
| 99 | _eden_pool(NULL), |
| 100 | _survivor_pool(NULL), |
| 101 | _old_pool(NULL) { } |
| 102 | |
| 103 | // For use by VM operations |
| 104 | enum CollectionType { |
| 105 | Scavenge, |
| 106 | MarkSweep |
| 107 | }; |
| 108 | |
| 109 | virtual Name kind() const { |
| 110 | return CollectedHeap::Parallel; |
| 111 | } |
| 112 | |
| 113 | virtual const char* name() const { |
| 114 | return "Parallel" ; |
| 115 | } |
| 116 | |
| 117 | virtual SoftRefPolicy* soft_ref_policy() { return &_soft_ref_policy; } |
| 118 | |
| 119 | virtual GrowableArray<GCMemoryManager*> memory_managers(); |
| 120 | virtual GrowableArray<MemoryPool*> memory_pools(); |
| 121 | |
| 122 | static PSYoungGen* young_gen() { return _young_gen; } |
| 123 | static PSOldGen* old_gen() { return _old_gen; } |
| 124 | |
| 125 | virtual PSAdaptiveSizePolicy* size_policy() { return _size_policy; } |
| 126 | |
| 127 | static PSGCAdaptivePolicyCounters* gc_policy_counters() { return _gc_policy_counters; } |
| 128 | |
| 129 | static ParallelScavengeHeap* heap(); |
| 130 | |
| 131 | static GCTaskManager* const gc_task_manager() { return _gc_task_manager; } |
| 132 | |
| 133 | CardTableBarrierSet* barrier_set(); |
| 134 | PSCardTable* card_table(); |
| 135 | |
| 136 | AdjoiningGenerations* gens() { return _gens; } |
| 137 | |
| 138 | // Returns JNI_OK on success |
| 139 | virtual jint initialize(); |
| 140 | |
| 141 | void post_initialize(); |
| 142 | void update_counters(); |
| 143 | |
| 144 | size_t capacity() const; |
| 145 | size_t used() const; |
| 146 | |
| 147 | // Return "true" if all generations have reached the |
| 148 | // maximal committed limit that they can reach, without a garbage |
| 149 | // collection. |
| 150 | virtual bool is_maximal_no_gc() const; |
| 151 | |
| 152 | virtual void register_nmethod(nmethod* nm); |
| 153 | virtual void unregister_nmethod(nmethod* nm); |
| 154 | virtual void verify_nmethod(nmethod* nm); |
| 155 | virtual void flush_nmethod(nmethod* nm); |
| 156 | |
| 157 | void prune_scavengable_nmethods(); |
| 158 | |
| 159 | size_t max_capacity() const; |
| 160 | |
| 161 | // Whether p is in the allocated part of the heap |
| 162 | bool is_in(const void* p) const; |
| 163 | |
| 164 | bool is_in_reserved(const void* p) const; |
| 165 | |
| 166 | bool is_in_young(oop p); // reserved part |
| 167 | bool is_in_old(oop p); // reserved part |
| 168 | |
| 169 | // Memory allocation. "gc_time_limit_was_exceeded" will |
| 170 | // be set to true if the adaptive size policy determine that |
| 171 | // an excessive amount of time is being spent doing collections |
| 172 | // and caused a NULL to be returned. If a NULL is not returned, |
| 173 | // "gc_time_limit_was_exceeded" has an undefined meaning. |
| 174 | HeapWord* mem_allocate(size_t size, bool* gc_overhead_limit_was_exceeded); |
| 175 | |
| 176 | // Allocation attempt(s) during a safepoint. It should never be called |
| 177 | // to allocate a new TLAB as this allocation might be satisfied out |
| 178 | // of the old generation. |
| 179 | HeapWord* failed_mem_allocate(size_t size); |
| 180 | |
| 181 | // Support for System.gc() |
| 182 | void collect(GCCause::Cause cause); |
| 183 | |
| 184 | // These also should be called by the vm thread at a safepoint (e.g., from a |
| 185 | // VM operation). |
| 186 | // |
| 187 | // The first collects the young generation only, unless the scavenge fails; it |
| 188 | // will then attempt a full gc. The second collects the entire heap; if |
| 189 | // maximum_compaction is true, it will compact everything and clear all soft |
| 190 | // references. |
| 191 | inline void invoke_scavenge(); |
| 192 | |
| 193 | // Perform a full collection |
| 194 | virtual void do_full_collection(bool clear_all_soft_refs); |
| 195 | |
| 196 | bool supports_inline_contig_alloc() const { return !UseNUMA; } |
| 197 | |
| 198 | HeapWord* volatile* top_addr() const { return !UseNUMA ? young_gen()->top_addr() : (HeapWord* volatile*)-1; } |
| 199 | HeapWord** end_addr() const { return !UseNUMA ? young_gen()->end_addr() : (HeapWord**)-1; } |
| 200 | |
| 201 | void ensure_parsability(bool retire_tlabs); |
| 202 | void resize_all_tlabs(); |
| 203 | |
| 204 | bool supports_tlab_allocation() const { return true; } |
| 205 | |
| 206 | size_t tlab_capacity(Thread* thr) const; |
| 207 | size_t tlab_used(Thread* thr) const; |
| 208 | size_t unsafe_max_tlab_alloc(Thread* thr) const; |
| 209 | |
| 210 | void object_iterate(ObjectClosure* cl); |
| 211 | void safe_object_iterate(ObjectClosure* cl) { object_iterate(cl); } |
| 212 | |
| 213 | HeapWord* block_start(const void* addr) const; |
| 214 | bool block_is_obj(const HeapWord* addr) const; |
| 215 | |
| 216 | jlong millis_since_last_gc(); |
| 217 | |
| 218 | void prepare_for_verify(); |
| 219 | PSHeapSummary create_ps_heap_summary(); |
| 220 | virtual void print_on(outputStream* st) const; |
| 221 | virtual void print_on_error(outputStream* st) const; |
| 222 | virtual void print_gc_threads_on(outputStream* st) const; |
| 223 | virtual void gc_threads_do(ThreadClosure* tc) const; |
| 224 | virtual void print_tracing_info() const; |
| 225 | |
| 226 | void verify(VerifyOption option /* ignored */); |
| 227 | |
| 228 | // Resize the young generation. The reserved space for the |
| 229 | // generation may be expanded in preparation for the resize. |
| 230 | void resize_young_gen(size_t eden_size, size_t survivor_size); |
| 231 | |
| 232 | // Resize the old generation. The reserved space for the |
| 233 | // generation may be expanded in preparation for the resize. |
| 234 | void resize_old_gen(size_t desired_free_space); |
| 235 | |
| 236 | // Save the tops of the spaces in all generations |
| 237 | void record_gen_tops_before_GC() PRODUCT_RETURN; |
| 238 | |
| 239 | // Mangle the unused parts of all spaces in the heap |
| 240 | void gen_mangle_unused_area() PRODUCT_RETURN; |
| 241 | |
| 242 | // Call these in sequential code around the processing of strong roots. |
| 243 | class ParStrongRootsScope : public MarkScope { |
| 244 | public: |
| 245 | ParStrongRootsScope(); |
| 246 | ~ParStrongRootsScope(); |
| 247 | }; |
| 248 | |
| 249 | GCMemoryManager* old_gc_manager() const { return _old_manager; } |
| 250 | GCMemoryManager* young_gc_manager() const { return _young_manager; } |
| 251 | }; |
| 252 | |
| 253 | // Simple class for storing info about the heap at the start of GC, to be used |
| 254 | // after GC for comparison/printing. |
| 255 | class PreGCValues { |
| 256 | public: |
| 257 | PreGCValues(ParallelScavengeHeap* heap) : |
| 258 | _heap_used(heap->used()), |
| 259 | _young_gen_used(heap->young_gen()->used_in_bytes()), |
| 260 | _old_gen_used(heap->old_gen()->used_in_bytes()), |
| 261 | _metadata_used(MetaspaceUtils::used_bytes()) { }; |
| 262 | |
| 263 | size_t heap_used() const { return _heap_used; } |
| 264 | size_t young_gen_used() const { return _young_gen_used; } |
| 265 | size_t old_gen_used() const { return _old_gen_used; } |
| 266 | size_t metadata_used() const { return _metadata_used; } |
| 267 | |
| 268 | private: |
| 269 | size_t _heap_used; |
| 270 | size_t _young_gen_used; |
| 271 | size_t _old_gen_used; |
| 272 | size_t _metadata_used; |
| 273 | }; |
| 274 | |
| 275 | // Class that can be used to print information about the |
| 276 | // adaptive size policy at intervals specified by |
| 277 | // AdaptiveSizePolicyOutputInterval. Only print information |
| 278 | // if an adaptive size policy is in use. |
| 279 | class AdaptiveSizePolicyOutput : AllStatic { |
| 280 | static bool enabled() { |
| 281 | return UseParallelGC && |
| 282 | UseAdaptiveSizePolicy && |
| 283 | log_is_enabled(Debug, gc, ergo); |
| 284 | } |
| 285 | public: |
| 286 | static void print() { |
| 287 | if (enabled()) { |
| 288 | ParallelScavengeHeap::heap()->size_policy()->print(); |
| 289 | } |
| 290 | } |
| 291 | |
| 292 | static void print(AdaptiveSizePolicy* size_policy, uint count) { |
| 293 | bool do_print = |
| 294 | enabled() && |
| 295 | (AdaptiveSizePolicyOutputInterval > 0) && |
| 296 | (count % AdaptiveSizePolicyOutputInterval) == 0; |
| 297 | |
| 298 | if (do_print) { |
| 299 | size_policy->print(); |
| 300 | } |
| 301 | } |
| 302 | }; |
| 303 | |
| 304 | #endif // SHARE_GC_PARALLEL_PARALLELSCAVENGEHEAP_HPP |
| 305 | |