1 | /* |
2 | * Copyright (c) 2014, 2018, Oracle and/or its affiliates. All rights reserved. |
3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
4 | * |
5 | * This code is free software; you can redistribute it and/or modify it |
6 | * under the terms of the GNU General Public License version 2 only, as |
7 | * published by the Free Software Foundation. |
8 | * |
9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
12 | * version 2 for more details (a copy is included in the LICENSE file that |
13 | * accompanied this code). |
14 | * |
15 | * You should have received a copy of the GNU General Public License version |
16 | * 2 along with this work; if not, write to the Free Software Foundation, |
17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
18 | * |
19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
20 | * or visit www.oracle.com if you need additional information or have any |
21 | * questions. |
22 | * |
23 | */ |
24 | |
25 | #include "precompiled.hpp" |
26 | |
27 | #include "gc/shared/blockOffsetTable.inline.hpp" |
28 | #include "gc/shared/cardGeneration.inline.hpp" |
29 | #include "gc/shared/cardTableRS.hpp" |
30 | #include "gc/shared/gcLocker.hpp" |
31 | #include "gc/shared/genCollectedHeap.hpp" |
32 | #include "gc/shared/genOopClosures.inline.hpp" |
33 | #include "gc/shared/generationSpec.hpp" |
34 | #include "gc/shared/space.inline.hpp" |
35 | #include "memory/iterator.hpp" |
36 | #include "memory/memRegion.hpp" |
37 | #include "logging/log.hpp" |
38 | #include "runtime/java.hpp" |
39 | |
40 | CardGeneration::CardGeneration(ReservedSpace rs, |
41 | size_t initial_byte_size, |
42 | CardTableRS* remset) : |
43 | Generation(rs, initial_byte_size), _rs(remset), |
44 | _shrink_factor(0), _min_heap_delta_bytes(), _capacity_at_prologue(), |
45 | _used_at_prologue() |
46 | { |
47 | HeapWord* start = (HeapWord*)rs.base(); |
48 | size_t reserved_byte_size = rs.size(); |
49 | assert((uintptr_t(start) & 3) == 0, "bad alignment" ); |
50 | assert((reserved_byte_size & 3) == 0, "bad alignment" ); |
51 | MemRegion reserved_mr(start, heap_word_size(reserved_byte_size)); |
52 | _bts = new BlockOffsetSharedArray(reserved_mr, |
53 | heap_word_size(initial_byte_size)); |
54 | MemRegion committed_mr(start, heap_word_size(initial_byte_size)); |
55 | _rs->resize_covered_region(committed_mr); |
56 | if (_bts == NULL) { |
57 | vm_exit_during_initialization("Could not allocate a BlockOffsetArray" ); |
58 | } |
59 | |
60 | // Verify that the start and end of this generation is the start of a card. |
61 | // If this wasn't true, a single card could span more than on generation, |
62 | // which would cause problems when we commit/uncommit memory, and when we |
63 | // clear and dirty cards. |
64 | guarantee(_rs->is_aligned(reserved_mr.start()), "generation must be card aligned" ); |
65 | if (reserved_mr.end() != GenCollectedHeap::heap()->reserved_region().end()) { |
66 | // Don't check at the very end of the heap as we'll assert that we're probing off |
67 | // the end if we try. |
68 | guarantee(_rs->is_aligned(reserved_mr.end()), "generation must be card aligned" ); |
69 | } |
70 | _min_heap_delta_bytes = MinHeapDeltaBytes; |
71 | _capacity_at_prologue = initial_byte_size; |
72 | _used_at_prologue = 0; |
73 | } |
74 | |
75 | bool CardGeneration::grow_by(size_t bytes) { |
76 | assert_correct_size_change_locking(); |
77 | bool result = _virtual_space.expand_by(bytes); |
78 | if (result) { |
79 | size_t new_word_size = |
80 | heap_word_size(_virtual_space.committed_size()); |
81 | MemRegion mr(space()->bottom(), new_word_size); |
82 | // Expand card table |
83 | GenCollectedHeap::heap()->rem_set()->resize_covered_region(mr); |
84 | // Expand shared block offset array |
85 | _bts->resize(new_word_size); |
86 | |
87 | // Fix for bug #4668531 |
88 | if (ZapUnusedHeapArea) { |
89 | MemRegion mangle_region(space()->end(), |
90 | (HeapWord*)_virtual_space.high()); |
91 | SpaceMangler::mangle_region(mangle_region); |
92 | } |
93 | |
94 | // Expand space -- also expands space's BOT |
95 | // (which uses (part of) shared array above) |
96 | space()->set_end((HeapWord*)_virtual_space.high()); |
97 | |
98 | // update the space and generation capacity counters |
99 | update_counters(); |
100 | |
101 | size_t new_mem_size = _virtual_space.committed_size(); |
102 | size_t old_mem_size = new_mem_size - bytes; |
103 | log_trace(gc, heap)("Expanding %s from " SIZE_FORMAT "K by " SIZE_FORMAT "K to " SIZE_FORMAT "K" , |
104 | name(), old_mem_size/K, bytes/K, new_mem_size/K); |
105 | } |
106 | return result; |
107 | } |
108 | |
109 | bool CardGeneration::expand(size_t bytes, size_t expand_bytes) { |
110 | assert_locked_or_safepoint(Heap_lock); |
111 | if (bytes == 0) { |
112 | return true; // That's what grow_by(0) would return |
113 | } |
114 | size_t aligned_bytes = ReservedSpace::page_align_size_up(bytes); |
115 | if (aligned_bytes == 0){ |
116 | // The alignment caused the number of bytes to wrap. An expand_by(0) will |
117 | // return true with the implication that an expansion was done when it |
118 | // was not. A call to expand implies a best effort to expand by "bytes" |
119 | // but not a guarantee. Align down to give a best effort. This is likely |
120 | // the most that the generation can expand since it has some capacity to |
121 | // start with. |
122 | aligned_bytes = ReservedSpace::page_align_size_down(bytes); |
123 | } |
124 | size_t aligned_expand_bytes = ReservedSpace::page_align_size_up(expand_bytes); |
125 | bool success = false; |
126 | if (aligned_expand_bytes > aligned_bytes) { |
127 | success = grow_by(aligned_expand_bytes); |
128 | } |
129 | if (!success) { |
130 | success = grow_by(aligned_bytes); |
131 | } |
132 | if (!success) { |
133 | success = grow_to_reserved(); |
134 | } |
135 | if (success && GCLocker::is_active_and_needs_gc()) { |
136 | log_trace(gc, heap)("Garbage collection disabled, expanded heap instead" ); |
137 | } |
138 | |
139 | return success; |
140 | } |
141 | |
142 | bool CardGeneration::grow_to_reserved() { |
143 | assert_correct_size_change_locking(); |
144 | bool success = true; |
145 | const size_t remaining_bytes = _virtual_space.uncommitted_size(); |
146 | if (remaining_bytes > 0) { |
147 | success = grow_by(remaining_bytes); |
148 | DEBUG_ONLY(if (!success) log_warning(gc)("grow to reserved failed" );) |
149 | } |
150 | return success; |
151 | } |
152 | |
153 | void CardGeneration::shrink(size_t bytes) { |
154 | assert_correct_size_change_locking(); |
155 | |
156 | size_t size = ReservedSpace::page_align_size_down(bytes); |
157 | if (size == 0) { |
158 | return; |
159 | } |
160 | |
161 | // Shrink committed space |
162 | _virtual_space.shrink_by(size); |
163 | // Shrink space; this also shrinks the space's BOT |
164 | space()->set_end((HeapWord*) _virtual_space.high()); |
165 | size_t new_word_size = heap_word_size(space()->capacity()); |
166 | // Shrink the shared block offset array |
167 | _bts->resize(new_word_size); |
168 | MemRegion mr(space()->bottom(), new_word_size); |
169 | // Shrink the card table |
170 | GenCollectedHeap::heap()->rem_set()->resize_covered_region(mr); |
171 | |
172 | size_t new_mem_size = _virtual_space.committed_size(); |
173 | size_t old_mem_size = new_mem_size + size; |
174 | log_trace(gc, heap)("Shrinking %s from " SIZE_FORMAT "K to " SIZE_FORMAT "K" , |
175 | name(), old_mem_size/K, new_mem_size/K); |
176 | } |
177 | |
178 | // No young generation references, clear this generation's cards. |
179 | void CardGeneration::clear_remembered_set() { |
180 | _rs->clear(reserved()); |
181 | } |
182 | |
183 | // Objects in this generation may have moved, invalidate this |
184 | // generation's cards. |
185 | void CardGeneration::invalidate_remembered_set() { |
186 | _rs->invalidate(used_region()); |
187 | } |
188 | |
189 | void CardGeneration::compute_new_size() { |
190 | assert(_shrink_factor <= 100, "invalid shrink factor" ); |
191 | size_t current_shrink_factor = _shrink_factor; |
192 | _shrink_factor = 0; |
193 | |
194 | // We don't have floating point command-line arguments |
195 | // Note: argument processing ensures that MinHeapFreeRatio < 100. |
196 | const double minimum_free_percentage = MinHeapFreeRatio / 100.0; |
197 | const double maximum_used_percentage = 1.0 - minimum_free_percentage; |
198 | |
199 | // Compute some numbers about the state of the heap. |
200 | const size_t used_after_gc = used(); |
201 | const size_t capacity_after_gc = capacity(); |
202 | |
203 | const double min_tmp = used_after_gc / maximum_used_percentage; |
204 | size_t minimum_desired_capacity = (size_t)MIN2(min_tmp, double(max_uintx)); |
205 | // Don't shrink less than the initial generation size |
206 | minimum_desired_capacity = MAX2(minimum_desired_capacity, initial_size()); |
207 | assert(used_after_gc <= minimum_desired_capacity, "sanity check" ); |
208 | |
209 | const size_t free_after_gc = free(); |
210 | const double free_percentage = ((double)free_after_gc) / capacity_after_gc; |
211 | log_trace(gc, heap)("CardGeneration::compute_new_size:" ); |
212 | log_trace(gc, heap)(" minimum_free_percentage: %6.2f maximum_used_percentage: %6.2f" , |
213 | minimum_free_percentage, |
214 | maximum_used_percentage); |
215 | log_trace(gc, heap)(" free_after_gc : %6.1fK used_after_gc : %6.1fK capacity_after_gc : %6.1fK" , |
216 | free_after_gc / (double) K, |
217 | used_after_gc / (double) K, |
218 | capacity_after_gc / (double) K); |
219 | log_trace(gc, heap)(" free_percentage: %6.2f" , free_percentage); |
220 | |
221 | if (capacity_after_gc < minimum_desired_capacity) { |
222 | // If we have less free space than we want then expand |
223 | size_t expand_bytes = minimum_desired_capacity - capacity_after_gc; |
224 | // Don't expand unless it's significant |
225 | if (expand_bytes >= _min_heap_delta_bytes) { |
226 | expand(expand_bytes, 0); // safe if expansion fails |
227 | } |
228 | log_trace(gc, heap)(" expanding: minimum_desired_capacity: %6.1fK expand_bytes: %6.1fK _min_heap_delta_bytes: %6.1fK" , |
229 | minimum_desired_capacity / (double) K, |
230 | expand_bytes / (double) K, |
231 | _min_heap_delta_bytes / (double) K); |
232 | return; |
233 | } |
234 | |
235 | // No expansion, now see if we want to shrink |
236 | size_t shrink_bytes = 0; |
237 | // We would never want to shrink more than this |
238 | size_t max_shrink_bytes = capacity_after_gc - minimum_desired_capacity; |
239 | |
240 | if (MaxHeapFreeRatio < 100) { |
241 | const double maximum_free_percentage = MaxHeapFreeRatio / 100.0; |
242 | const double minimum_used_percentage = 1.0 - maximum_free_percentage; |
243 | const double max_tmp = used_after_gc / minimum_used_percentage; |
244 | size_t maximum_desired_capacity = (size_t)MIN2(max_tmp, double(max_uintx)); |
245 | maximum_desired_capacity = MAX2(maximum_desired_capacity, initial_size()); |
246 | log_trace(gc, heap)(" maximum_free_percentage: %6.2f minimum_used_percentage: %6.2f" , |
247 | maximum_free_percentage, minimum_used_percentage); |
248 | log_trace(gc, heap)(" _capacity_at_prologue: %6.1fK minimum_desired_capacity: %6.1fK maximum_desired_capacity: %6.1fK" , |
249 | _capacity_at_prologue / (double) K, |
250 | minimum_desired_capacity / (double) K, |
251 | maximum_desired_capacity / (double) K); |
252 | assert(minimum_desired_capacity <= maximum_desired_capacity, |
253 | "sanity check" ); |
254 | |
255 | if (capacity_after_gc > maximum_desired_capacity) { |
256 | // Capacity too large, compute shrinking size |
257 | shrink_bytes = capacity_after_gc - maximum_desired_capacity; |
258 | if (ShrinkHeapInSteps) { |
259 | // If ShrinkHeapInSteps is true (the default), |
260 | // we don't want to shrink all the way back to initSize if people call |
261 | // System.gc(), because some programs do that between "phases" and then |
262 | // we'd just have to grow the heap up again for the next phase. So we |
263 | // damp the shrinking: 0% on the first call, 10% on the second call, 40% |
264 | // on the third call, and 100% by the fourth call. But if we recompute |
265 | // size without shrinking, it goes back to 0%. |
266 | shrink_bytes = shrink_bytes / 100 * current_shrink_factor; |
267 | if (current_shrink_factor == 0) { |
268 | _shrink_factor = 10; |
269 | } else { |
270 | _shrink_factor = MIN2(current_shrink_factor * 4, (size_t) 100); |
271 | } |
272 | } |
273 | assert(shrink_bytes <= max_shrink_bytes, "invalid shrink size" ); |
274 | log_trace(gc, heap)(" shrinking: initSize: %.1fK maximum_desired_capacity: %.1fK" , |
275 | initial_size() / (double) K, maximum_desired_capacity / (double) K); |
276 | log_trace(gc, heap)(" shrink_bytes: %.1fK current_shrink_factor: " SIZE_FORMAT " new shrink factor: " SIZE_FORMAT " _min_heap_delta_bytes: %.1fK" , |
277 | shrink_bytes / (double) K, |
278 | current_shrink_factor, |
279 | _shrink_factor, |
280 | _min_heap_delta_bytes / (double) K); |
281 | } |
282 | } |
283 | |
284 | if (capacity_after_gc > _capacity_at_prologue) { |
285 | // We might have expanded for promotions, in which case we might want to |
286 | // take back that expansion if there's room after GC. That keeps us from |
287 | // stretching the heap with promotions when there's plenty of room. |
288 | size_t expansion_for_promotion = capacity_after_gc - _capacity_at_prologue; |
289 | expansion_for_promotion = MIN2(expansion_for_promotion, max_shrink_bytes); |
290 | // We have two shrinking computations, take the largest |
291 | shrink_bytes = MAX2(shrink_bytes, expansion_for_promotion); |
292 | assert(shrink_bytes <= max_shrink_bytes, "invalid shrink size" ); |
293 | log_trace(gc, heap)(" aggressive shrinking: _capacity_at_prologue: %.1fK capacity_after_gc: %.1fK expansion_for_promotion: %.1fK shrink_bytes: %.1fK" , |
294 | capacity_after_gc / (double) K, |
295 | _capacity_at_prologue / (double) K, |
296 | expansion_for_promotion / (double) K, |
297 | shrink_bytes / (double) K); |
298 | } |
299 | // Don't shrink unless it's significant |
300 | if (shrink_bytes >= _min_heap_delta_bytes) { |
301 | shrink(shrink_bytes); |
302 | } |
303 | } |
304 | |
305 | // Currently nothing to do. |
306 | void CardGeneration::prepare_for_verify() {} |
307 | |
308 | void CardGeneration::space_iterate(SpaceClosure* blk, |
309 | bool usedOnly) { |
310 | blk->do_space(space()); |
311 | } |
312 | |
313 | void CardGeneration::younger_refs_iterate(OopsInGenClosure* blk, uint n_threads) { |
314 | blk->set_generation(this); |
315 | younger_refs_in_space_iterate(space(), blk, n_threads); |
316 | blk->reset_generation(); |
317 | } |
318 | |