| 1 | /* |
| 2 | * Copyright (c) 2014, 2019, Oracle and/or its affiliates. All rights reserved. |
| 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
| 4 | * |
| 5 | * This code is free software; you can redistribute it and/or modify it |
| 6 | * under the terms of the GNU General Public License version 2 only, as |
| 7 | * published by the Free Software Foundation. |
| 8 | * |
| 9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
| 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| 11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
| 12 | * version 2 for more details (a copy is included in the LICENSE file that |
| 13 | * accompanied this code). |
| 14 | * |
| 15 | * You should have received a copy of the GNU General Public License version |
| 16 | * 2 along with this work; if not, write to the Free Software Foundation, |
| 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
| 18 | * |
| 19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
| 20 | * or visit www.oracle.com if you need additional information or have any |
| 21 | * questions. |
| 22 | * |
| 23 | */ |
| 24 | |
| 25 | #ifndef SHARE_SERVICES_MALLOCSITETABLE_HPP |
| 26 | #define SHARE_SERVICES_MALLOCSITETABLE_HPP |
| 27 | |
| 28 | #if INCLUDE_NMT |
| 29 | |
| 30 | #include "memory/allocation.hpp" |
| 31 | #include "runtime/atomic.hpp" |
| 32 | #include "services/allocationSite.hpp" |
| 33 | #include "services/mallocTracker.hpp" |
| 34 | #include "services/nmtCommon.hpp" |
| 35 | #include "utilities/nativeCallStack.hpp" |
| 36 | |
| 37 | // MallocSite represents a code path that eventually calls |
| 38 | // os::malloc() to allocate memory |
| 39 | class MallocSite : public AllocationSite<MemoryCounter> { |
| 40 | public: |
| 41 | MallocSite() : |
| 42 | AllocationSite<MemoryCounter>(NativeCallStack::empty_stack(), mtNone) {} |
| 43 | |
| 44 | MallocSite(const NativeCallStack& stack, MEMFLAGS flags) : |
| 45 | AllocationSite<MemoryCounter>(stack, flags) {} |
| 46 | |
| 47 | |
| 48 | void allocate(size_t size) { data()->allocate(size); } |
| 49 | void deallocate(size_t size) { data()->deallocate(size); } |
| 50 | |
| 51 | // Memory allocated from this code path |
| 52 | size_t size() const { return peek()->size(); } |
| 53 | // The number of calls were made |
| 54 | size_t count() const { return peek()->count(); } |
| 55 | }; |
| 56 | |
| 57 | // Malloc site hashtable entry |
| 58 | class MallocSiteHashtableEntry : public CHeapObj<mtNMT> { |
| 59 | private: |
| 60 | MallocSite _malloc_site; |
| 61 | MallocSiteHashtableEntry* volatile _next; |
| 62 | |
| 63 | public: |
| 64 | MallocSiteHashtableEntry() : _next(NULL) { } |
| 65 | |
| 66 | MallocSiteHashtableEntry(NativeCallStack stack, MEMFLAGS flags): |
| 67 | _malloc_site(stack, flags), _next(NULL) { |
| 68 | assert(flags != mtNone, "Expect a real memory type" ); |
| 69 | } |
| 70 | |
| 71 | inline const MallocSiteHashtableEntry* next() const { |
| 72 | return _next; |
| 73 | } |
| 74 | |
| 75 | // Insert an entry atomically. |
| 76 | // Return true if the entry is inserted successfully. |
| 77 | // The operation can be failed due to contention from other thread. |
| 78 | bool atomic_insert(MallocSiteHashtableEntry* entry); |
| 79 | |
| 80 | void set_callsite(const MallocSite& site) { |
| 81 | _malloc_site = site; |
| 82 | } |
| 83 | |
| 84 | inline const MallocSite* peek() const { return &_malloc_site; } |
| 85 | inline MallocSite* data() { return &_malloc_site; } |
| 86 | |
| 87 | inline long hash() const { return _malloc_site.hash(); } |
| 88 | inline bool equals(const NativeCallStack& stack) const { |
| 89 | return _malloc_site.equals(stack); |
| 90 | } |
| 91 | // Allocation/deallocation on this allocation site |
| 92 | inline void allocate(size_t size) { _malloc_site.allocate(size); } |
| 93 | inline void deallocate(size_t size) { _malloc_site.deallocate(size); } |
| 94 | // Memory counters |
| 95 | inline size_t size() const { return _malloc_site.size(); } |
| 96 | inline size_t count() const { return _malloc_site.count(); } |
| 97 | }; |
| 98 | |
| 99 | // The walker walks every entry on MallocSiteTable |
| 100 | class MallocSiteWalker : public StackObj { |
| 101 | public: |
| 102 | virtual bool do_malloc_site(const MallocSite* e) { return false; } |
| 103 | }; |
| 104 | |
| 105 | /* |
| 106 | * Native memory tracking call site table. |
| 107 | * The table is only needed when detail tracking is enabled. |
| 108 | */ |
| 109 | class MallocSiteTable : AllStatic { |
| 110 | private: |
| 111 | // The number of hash bucket in this hashtable. The number should |
| 112 | // be tuned if malloc activities changed significantly. |
| 113 | // The statistics data can be obtained via Jcmd |
| 114 | // jcmd <pid> VM.native_memory statistics. |
| 115 | |
| 116 | // Currently, (number of buckets / number of entires) ratio is |
| 117 | // about 1 / 6 |
| 118 | enum { |
| 119 | table_base_size = 128, // The base size is calculated from statistics to give |
| 120 | // table ratio around 1:6 |
| 121 | table_size = (table_base_size * NMT_TrackingStackDepth - 1) |
| 122 | }; |
| 123 | |
| 124 | |
| 125 | // This is a very special lock, that allows multiple shared accesses (sharedLock), but |
| 126 | // once exclusive access (exclusiveLock) is requested, all shared accesses are |
| 127 | // rejected forever. |
| 128 | class AccessLock : public StackObj { |
| 129 | enum LockState { |
| 130 | NoLock, |
| 131 | SharedLock, |
| 132 | ExclusiveLock |
| 133 | }; |
| 134 | |
| 135 | private: |
| 136 | // A very large negative number. The only possibility to "overflow" |
| 137 | // this number is when there are more than -min_jint threads in |
| 138 | // this process, which is not going to happen in foreseeable future. |
| 139 | const static int _MAGIC_ = min_jint; |
| 140 | |
| 141 | LockState _lock_state; |
| 142 | volatile int* _lock; |
| 143 | public: |
| 144 | AccessLock(volatile int* lock) : |
| 145 | _lock_state(NoLock), _lock(lock) { |
| 146 | } |
| 147 | |
| 148 | ~AccessLock() { |
| 149 | if (_lock_state == SharedLock) { |
| 150 | Atomic::dec(_lock); |
| 151 | } |
| 152 | } |
| 153 | // Acquire shared lock. |
| 154 | // Return true if shared access is granted. |
| 155 | inline bool sharedLock() { |
| 156 | jint res = Atomic::add(1, _lock); |
| 157 | if (res < 0) { |
| 158 | Atomic::dec(_lock); |
| 159 | return false; |
| 160 | } |
| 161 | _lock_state = SharedLock; |
| 162 | return true; |
| 163 | } |
| 164 | // Acquire exclusive lock |
| 165 | void exclusiveLock(); |
| 166 | }; |
| 167 | |
| 168 | public: |
| 169 | static bool initialize(); |
| 170 | static void shutdown(); |
| 171 | |
| 172 | NOT_PRODUCT(static int access_peak_count() { return _peak_count; }) |
| 173 | |
| 174 | // Number of hash buckets |
| 175 | static inline int hash_buckets() { return (int)table_size; } |
| 176 | |
| 177 | // Access and copy a call stack from this table. Shared lock should be |
| 178 | // acquired before access the entry. |
| 179 | static inline bool access_stack(NativeCallStack& stack, size_t bucket_idx, |
| 180 | size_t pos_idx) { |
| 181 | AccessLock locker(&_access_count); |
| 182 | if (locker.sharedLock()) { |
| 183 | NOT_PRODUCT(_peak_count = MAX2(_peak_count, _access_count);) |
| 184 | MallocSite* site = malloc_site(bucket_idx, pos_idx); |
| 185 | if (site != NULL) { |
| 186 | stack = *site->call_stack(); |
| 187 | return true; |
| 188 | } |
| 189 | } |
| 190 | return false; |
| 191 | } |
| 192 | |
| 193 | // Record a new allocation from specified call path. |
| 194 | // Return true if the allocation is recorded successfully, bucket_idx |
| 195 | // and pos_idx are also updated to indicate the entry where the allocation |
| 196 | // information was recorded. |
| 197 | // Return false only occurs under rare scenarios: |
| 198 | // 1. out of memory |
| 199 | // 2. overflow hash bucket |
| 200 | static inline bool allocation_at(const NativeCallStack& stack, size_t size, |
| 201 | size_t* bucket_idx, size_t* pos_idx, MEMFLAGS flags) { |
| 202 | AccessLock locker(&_access_count); |
| 203 | if (locker.sharedLock()) { |
| 204 | NOT_PRODUCT(_peak_count = MAX2(_peak_count, _access_count);) |
| 205 | MallocSite* site = lookup_or_add(stack, bucket_idx, pos_idx, flags); |
| 206 | if (site != NULL) site->allocate(size); |
| 207 | return site != NULL; |
| 208 | } |
| 209 | return false; |
| 210 | } |
| 211 | |
| 212 | // Record memory deallocation. bucket_idx and pos_idx indicate where the allocation |
| 213 | // information was recorded. |
| 214 | static inline bool deallocation_at(size_t size, size_t bucket_idx, size_t pos_idx) { |
| 215 | AccessLock locker(&_access_count); |
| 216 | if (locker.sharedLock()) { |
| 217 | NOT_PRODUCT(_peak_count = MAX2(_peak_count, _access_count);) |
| 218 | MallocSite* site = malloc_site(bucket_idx, pos_idx); |
| 219 | if (site != NULL) { |
| 220 | site->deallocate(size); |
| 221 | return true; |
| 222 | } |
| 223 | } |
| 224 | return false; |
| 225 | } |
| 226 | |
| 227 | // Walk this table. |
| 228 | static bool walk_malloc_site(MallocSiteWalker* walker); |
| 229 | |
| 230 | private: |
| 231 | static MallocSiteHashtableEntry* new_entry(const NativeCallStack& key, MEMFLAGS flags); |
| 232 | static void reset(); |
| 233 | |
| 234 | // Delete a bucket linked list |
| 235 | static void delete_linked_list(MallocSiteHashtableEntry* head); |
| 236 | |
| 237 | static MallocSite* lookup_or_add(const NativeCallStack& key, size_t* bucket_idx, size_t* pos_idx, MEMFLAGS flags); |
| 238 | static MallocSite* malloc_site(size_t bucket_idx, size_t pos_idx); |
| 239 | static bool walk(MallocSiteWalker* walker); |
| 240 | |
| 241 | static inline unsigned int hash_to_index(unsigned int hash) { |
| 242 | return (hash % table_size); |
| 243 | } |
| 244 | |
| 245 | static inline const NativeCallStack* hash_entry_allocation_stack() { |
| 246 | assert(_hash_entry_allocation_stack != NULL, "Must be set" ); |
| 247 | return _hash_entry_allocation_stack; |
| 248 | } |
| 249 | |
| 250 | static inline const MallocSiteHashtableEntry* hash_entry_allocation_site() { |
| 251 | assert(_hash_entry_allocation_site != NULL, "Must be set" ); |
| 252 | return _hash_entry_allocation_site; |
| 253 | } |
| 254 | |
| 255 | private: |
| 256 | // Counter for counting concurrent access |
| 257 | static volatile int _access_count; |
| 258 | |
| 259 | // The callsite hashtable. It has to be a static table, |
| 260 | // since malloc call can come from C runtime linker. |
| 261 | static MallocSiteHashtableEntry* _table[table_size]; |
| 262 | static const NativeCallStack* _hash_entry_allocation_stack; |
| 263 | static const MallocSiteHashtableEntry* _hash_entry_allocation_site; |
| 264 | |
| 265 | |
| 266 | NOT_PRODUCT(static int _peak_count;) |
| 267 | }; |
| 268 | |
| 269 | #endif // INCLUDE_NMT |
| 270 | #endif // SHARE_SERVICES_MALLOCSITETABLE_HPP |
| 271 | |