| 1 | /* |
| 2 | * Copyright (c) 2014, 2018, Oracle and/or its affiliates. All rights reserved. |
| 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
| 4 | * |
| 5 | * This code is free software; you can redistribute it and/or modify it |
| 6 | * under the terms of the GNU General Public License version 2 only, as |
| 7 | * published by the Free Software Foundation. |
| 8 | * |
| 9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
| 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| 11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
| 12 | * version 2 for more details (a copy is included in the LICENSE file that |
| 13 | * accompanied this code). |
| 14 | * |
| 15 | * You should have received a copy of the GNU General Public License version |
| 16 | * 2 along with this work; if not, write to the Free Software Foundation, |
| 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
| 18 | * |
| 19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
| 20 | * or visit www.oracle.com if you need additional information or have any |
| 21 | * questions. |
| 22 | * |
| 23 | */ |
| 24 | |
| 25 | #include "precompiled.hpp" |
| 26 | #include "gc/g1/g1PageBasedVirtualSpace.hpp" |
| 27 | #include "gc/shared/workgroup.hpp" |
| 28 | #include "oops/markOop.hpp" |
| 29 | #include "oops/oop.inline.hpp" |
| 30 | #include "runtime/atomic.hpp" |
| 31 | #include "runtime/os.inline.hpp" |
| 32 | #include "services/memTracker.hpp" |
| 33 | #include "utilities/align.hpp" |
| 34 | #include "utilities/bitMap.inline.hpp" |
| 35 | |
| 36 | G1PageBasedVirtualSpace::G1PageBasedVirtualSpace(ReservedSpace rs, size_t used_size, size_t page_size) : |
| 37 | _low_boundary(NULL), _high_boundary(NULL), _tail_size(0), _page_size(0), |
| 38 | _committed(mtGC), _dirty(mtGC), _special(false), _executable(false) { |
| 39 | initialize_with_page_size(rs, used_size, page_size); |
| 40 | } |
| 41 | |
| 42 | void G1PageBasedVirtualSpace::initialize_with_page_size(ReservedSpace rs, size_t used_size, size_t page_size) { |
| 43 | guarantee(rs.is_reserved(), "Given reserved space must have been reserved already." ); |
| 44 | |
| 45 | vmassert(_low_boundary == NULL, "VirtualSpace already initialized" ); |
| 46 | vmassert(page_size > 0, "Page size must be non-zero." ); |
| 47 | |
| 48 | guarantee(is_aligned(rs.base(), page_size), |
| 49 | "Reserved space base " PTR_FORMAT " is not aligned to requested page size " SIZE_FORMAT, p2i(rs.base()), page_size); |
| 50 | guarantee(is_aligned(used_size, os::vm_page_size()), |
| 51 | "Given used reserved space size needs to be OS page size aligned (%d bytes) but is " SIZE_FORMAT, os::vm_page_size(), used_size); |
| 52 | guarantee(used_size <= rs.size(), |
| 53 | "Used size of reserved space " SIZE_FORMAT " bytes is smaller than reservation at " SIZE_FORMAT " bytes" , used_size, rs.size()); |
| 54 | guarantee(is_aligned(rs.size(), page_size), |
| 55 | "Expected that the virtual space is size aligned, but " SIZE_FORMAT " is not aligned to page size " SIZE_FORMAT, rs.size(), page_size); |
| 56 | |
| 57 | _low_boundary = rs.base(); |
| 58 | _high_boundary = _low_boundary + used_size; |
| 59 | |
| 60 | _special = rs.special(); |
| 61 | _executable = rs.executable(); |
| 62 | |
| 63 | _page_size = page_size; |
| 64 | |
| 65 | vmassert(_committed.size() == 0, "virtual space initialized more than once" ); |
| 66 | BitMap::idx_t size_in_pages = rs.size() / page_size; |
| 67 | _committed.initialize(size_in_pages); |
| 68 | if (_special) { |
| 69 | _dirty.initialize(size_in_pages); |
| 70 | } |
| 71 | |
| 72 | _tail_size = used_size % _page_size; |
| 73 | } |
| 74 | |
| 75 | G1PageBasedVirtualSpace::~G1PageBasedVirtualSpace() { |
| 76 | // This does not release memory it never reserved. |
| 77 | // Caller must release via rs.release(); |
| 78 | _low_boundary = NULL; |
| 79 | _high_boundary = NULL; |
| 80 | _special = false; |
| 81 | _executable = false; |
| 82 | _page_size = 0; |
| 83 | _tail_size = 0; |
| 84 | } |
| 85 | |
| 86 | size_t G1PageBasedVirtualSpace::committed_size() const { |
| 87 | size_t result = _committed.count_one_bits() * _page_size; |
| 88 | // The last page might not be in full. |
| 89 | if (is_last_page_partial() && _committed.at(_committed.size() - 1)) { |
| 90 | result -= _page_size - _tail_size; |
| 91 | } |
| 92 | return result; |
| 93 | } |
| 94 | |
| 95 | size_t G1PageBasedVirtualSpace::reserved_size() const { |
| 96 | return pointer_delta(_high_boundary, _low_boundary, sizeof(char)); |
| 97 | } |
| 98 | |
| 99 | size_t G1PageBasedVirtualSpace::uncommitted_size() const { |
| 100 | return reserved_size() - committed_size(); |
| 101 | } |
| 102 | |
| 103 | void G1PageBasedVirtualSpace::commit_and_set_special() { |
| 104 | commit_internal(addr_to_page_index(_low_boundary), addr_to_page_index(_high_boundary)); |
| 105 | _special = true; |
| 106 | _dirty.initialize(reserved_size()/_page_size); |
| 107 | } |
| 108 | |
| 109 | size_t G1PageBasedVirtualSpace::addr_to_page_index(char* addr) const { |
| 110 | return (addr - _low_boundary) / _page_size; |
| 111 | } |
| 112 | |
| 113 | bool G1PageBasedVirtualSpace::is_area_committed(size_t start_page, size_t size_in_pages) const { |
| 114 | size_t end_page = start_page + size_in_pages; |
| 115 | return _committed.get_next_zero_offset(start_page, end_page) >= end_page; |
| 116 | } |
| 117 | |
| 118 | bool G1PageBasedVirtualSpace::is_area_uncommitted(size_t start_page, size_t size_in_pages) const { |
| 119 | size_t end_page = start_page + size_in_pages; |
| 120 | return _committed.get_next_one_offset(start_page, end_page) >= end_page; |
| 121 | } |
| 122 | |
| 123 | char* G1PageBasedVirtualSpace::page_start(size_t index) const { |
| 124 | return _low_boundary + index * _page_size; |
| 125 | } |
| 126 | |
| 127 | bool G1PageBasedVirtualSpace::is_after_last_page(size_t index) const { |
| 128 | guarantee(index <= _committed.size(), |
| 129 | "Given boundary page " SIZE_FORMAT " is beyond managed page count " SIZE_FORMAT, index, _committed.size()); |
| 130 | return index == _committed.size(); |
| 131 | } |
| 132 | |
| 133 | void G1PageBasedVirtualSpace::commit_preferred_pages(size_t start, size_t num_pages) { |
| 134 | vmassert(num_pages > 0, "No full pages to commit" ); |
| 135 | vmassert(start + num_pages <= _committed.size(), |
| 136 | "Tried to commit area from page " SIZE_FORMAT " to page " SIZE_FORMAT " " |
| 137 | "that is outside of managed space of " SIZE_FORMAT " pages" , |
| 138 | start, start + num_pages, _committed.size()); |
| 139 | |
| 140 | char* start_addr = page_start(start); |
| 141 | size_t size = num_pages * _page_size; |
| 142 | |
| 143 | os::commit_memory_or_exit(start_addr, size, _page_size, _executable, |
| 144 | err_msg("Failed to commit area from " PTR_FORMAT " to " PTR_FORMAT " of length " SIZE_FORMAT "." , |
| 145 | p2i(start_addr), p2i(start_addr + size), size)); |
| 146 | } |
| 147 | |
| 148 | void G1PageBasedVirtualSpace::commit_tail() { |
| 149 | vmassert(_tail_size > 0, "The size of the tail area must be > 0 when reaching here" ); |
| 150 | |
| 151 | char* const aligned_end_address = align_down(_high_boundary, _page_size); |
| 152 | os::commit_memory_or_exit(aligned_end_address, _tail_size, os::vm_page_size(), _executable, |
| 153 | err_msg("Failed to commit tail area from " PTR_FORMAT " to " PTR_FORMAT " of length " SIZE_FORMAT "." , |
| 154 | p2i(aligned_end_address), p2i(_high_boundary), _tail_size)); |
| 155 | } |
| 156 | |
| 157 | void G1PageBasedVirtualSpace::commit_internal(size_t start_page, size_t end_page) { |
| 158 | guarantee(start_page < end_page, |
| 159 | "Given start page " SIZE_FORMAT " is larger or equal to end page " SIZE_FORMAT, start_page, end_page); |
| 160 | guarantee(end_page <= _committed.size(), |
| 161 | "Given end page " SIZE_FORMAT " is beyond end of managed page amount of " SIZE_FORMAT, end_page, _committed.size()); |
| 162 | |
| 163 | size_t pages = end_page - start_page; |
| 164 | bool need_to_commit_tail = is_after_last_page(end_page) && is_last_page_partial(); |
| 165 | |
| 166 | // If we have to commit some (partial) tail area, decrease the amount of pages to avoid |
| 167 | // committing that in the full-page commit code. |
| 168 | if (need_to_commit_tail) { |
| 169 | pages--; |
| 170 | } |
| 171 | |
| 172 | if (pages > 0) { |
| 173 | commit_preferred_pages(start_page, pages); |
| 174 | } |
| 175 | |
| 176 | if (need_to_commit_tail) { |
| 177 | commit_tail(); |
| 178 | } |
| 179 | } |
| 180 | |
| 181 | char* G1PageBasedVirtualSpace::bounded_end_addr(size_t end_page) const { |
| 182 | return MIN2(_high_boundary, page_start(end_page)); |
| 183 | } |
| 184 | |
| 185 | void G1PageBasedVirtualSpace::pretouch_internal(size_t start_page, size_t end_page) { |
| 186 | guarantee(start_page < end_page, |
| 187 | "Given start page " SIZE_FORMAT " is larger or equal to end page " SIZE_FORMAT, start_page, end_page); |
| 188 | |
| 189 | os::pretouch_memory(page_start(start_page), bounded_end_addr(end_page), _page_size); |
| 190 | } |
| 191 | |
| 192 | bool G1PageBasedVirtualSpace::commit(size_t start_page, size_t size_in_pages) { |
| 193 | // We need to make sure to commit all pages covered by the given area. |
| 194 | guarantee(is_area_uncommitted(start_page, size_in_pages), "Specified area is not uncommitted" ); |
| 195 | |
| 196 | bool zero_filled = true; |
| 197 | size_t end_page = start_page + size_in_pages; |
| 198 | |
| 199 | if (_special) { |
| 200 | // Check for dirty pages and update zero_filled if any found. |
| 201 | if (_dirty.get_next_one_offset(start_page, end_page) < end_page) { |
| 202 | zero_filled = false; |
| 203 | _dirty.clear_range(start_page, end_page); |
| 204 | } |
| 205 | } else { |
| 206 | commit_internal(start_page, end_page); |
| 207 | } |
| 208 | _committed.set_range(start_page, end_page); |
| 209 | |
| 210 | return zero_filled; |
| 211 | } |
| 212 | |
| 213 | void G1PageBasedVirtualSpace::uncommit_internal(size_t start_page, size_t end_page) { |
| 214 | guarantee(start_page < end_page, |
| 215 | "Given start page " SIZE_FORMAT " is larger or equal to end page " SIZE_FORMAT, start_page, end_page); |
| 216 | |
| 217 | char* start_addr = page_start(start_page); |
| 218 | os::uncommit_memory(start_addr, pointer_delta(bounded_end_addr(end_page), start_addr, sizeof(char))); |
| 219 | } |
| 220 | |
| 221 | void G1PageBasedVirtualSpace::uncommit(size_t start_page, size_t size_in_pages) { |
| 222 | guarantee(is_area_committed(start_page, size_in_pages), "checking" ); |
| 223 | |
| 224 | size_t end_page = start_page + size_in_pages; |
| 225 | if (_special) { |
| 226 | // Mark that memory is dirty. If committed again the memory might |
| 227 | // need to be cleared explicitly. |
| 228 | _dirty.set_range(start_page, end_page); |
| 229 | } else { |
| 230 | uncommit_internal(start_page, end_page); |
| 231 | } |
| 232 | |
| 233 | _committed.clear_range(start_page, end_page); |
| 234 | } |
| 235 | |
| 236 | class G1PretouchTask : public AbstractGangTask { |
| 237 | private: |
| 238 | char* volatile _cur_addr; |
| 239 | char* const _start_addr; |
| 240 | char* const _end_addr; |
| 241 | size_t _page_size; |
| 242 | public: |
| 243 | G1PretouchTask(char* start_address, char* end_address, size_t page_size) : |
| 244 | AbstractGangTask("G1 PreTouch" ), |
| 245 | _cur_addr(start_address), |
| 246 | _start_addr(start_address), |
| 247 | _end_addr(end_address), |
| 248 | _page_size(0) { |
| 249 | #ifdef LINUX |
| 250 | _page_size = UseTransparentHugePages ? (size_t)os::vm_page_size(): page_size; |
| 251 | #else |
| 252 | _page_size = page_size; |
| 253 | #endif |
| 254 | } |
| 255 | |
| 256 | virtual void work(uint worker_id) { |
| 257 | size_t const actual_chunk_size = MAX2(chunk_size(), _page_size); |
| 258 | while (true) { |
| 259 | char* touch_addr = Atomic::add(actual_chunk_size, &_cur_addr) - actual_chunk_size; |
| 260 | if (touch_addr < _start_addr || touch_addr >= _end_addr) { |
| 261 | break; |
| 262 | } |
| 263 | char* end_addr = touch_addr + MIN2(actual_chunk_size, pointer_delta(_end_addr, touch_addr, sizeof(char))); |
| 264 | os::pretouch_memory(touch_addr, end_addr, _page_size); |
| 265 | } |
| 266 | } |
| 267 | |
| 268 | static size_t chunk_size() { return PreTouchParallelChunkSize; } |
| 269 | }; |
| 270 | |
| 271 | void G1PageBasedVirtualSpace::pretouch(size_t start_page, size_t size_in_pages, WorkGang* pretouch_gang) { |
| 272 | G1PretouchTask cl(page_start(start_page), bounded_end_addr(start_page + size_in_pages), _page_size); |
| 273 | |
| 274 | if (pretouch_gang != NULL) { |
| 275 | size_t num_chunks = MAX2((size_t)1, size_in_pages * _page_size / MAX2(G1PretouchTask::chunk_size(), _page_size)); |
| 276 | |
| 277 | uint num_workers = MIN2((uint)num_chunks, pretouch_gang->active_workers()); |
| 278 | log_debug(gc, heap)("Running %s with %u workers for " SIZE_FORMAT " work units pre-touching " SIZE_FORMAT "B." , |
| 279 | cl.name(), num_workers, num_chunks, size_in_pages * _page_size); |
| 280 | pretouch_gang->run_task(&cl, num_workers); |
| 281 | } else { |
| 282 | log_debug(gc, heap)("Running %s pre-touching " SIZE_FORMAT "B." , |
| 283 | cl.name(), size_in_pages * _page_size); |
| 284 | cl.work(0); |
| 285 | } |
| 286 | } |
| 287 | |
| 288 | bool G1PageBasedVirtualSpace::contains(const void* p) const { |
| 289 | return _low_boundary <= (const char*) p && (const char*) p < _high_boundary; |
| 290 | } |
| 291 | |
| 292 | #ifndef PRODUCT |
| 293 | void G1PageBasedVirtualSpace::print_on(outputStream* out) { |
| 294 | out->print ("Virtual space:" ); |
| 295 | if (_special) out->print(" (pinned in memory)" ); |
| 296 | out->cr(); |
| 297 | out->print_cr(" - committed: " SIZE_FORMAT, committed_size()); |
| 298 | out->print_cr(" - reserved: " SIZE_FORMAT, reserved_size()); |
| 299 | out->print_cr(" - preferred page size: " SIZE_FORMAT, _page_size); |
| 300 | out->print_cr(" - [low_b, high_b]: [" PTR_FORMAT ", " PTR_FORMAT "]" , p2i(_low_boundary), p2i(_high_boundary)); |
| 301 | } |
| 302 | |
| 303 | void G1PageBasedVirtualSpace::print() { |
| 304 | print_on(tty); |
| 305 | } |
| 306 | #endif |
| 307 | |