| 1 | // Licensed to the .NET Foundation under one or more agreements. | 
|---|
| 2 | // The .NET Foundation licenses this file to you under the MIT license. | 
|---|
| 3 | // See the LICENSE file in the project root for more information. | 
|---|
| 4 |  | 
|---|
| 5 | #ifndef _ALLOC_H_ | 
|---|
| 6 | #define _ALLOC_H_ | 
|---|
| 7 |  | 
|---|
| 8 | #if !defined(_HOST_H_) | 
|---|
| 9 | #include "host.h" | 
|---|
| 10 | #endif // defined(_HOST_H_) | 
|---|
| 11 |  | 
|---|
| 12 | // CompMemKind values are used to tag memory allocations performed via | 
|---|
| 13 | // the compiler's allocator so that the memory usage of various compiler | 
|---|
| 14 | // components can be tracked separately (when MEASURE_MEM_ALLOC is defined). | 
|---|
| 15 |  | 
|---|
| 16 | enum CompMemKind | 
|---|
| 17 | { | 
|---|
| 18 | #define CompMemKindMacro(kind) CMK_##kind, | 
|---|
| 19 | #include "compmemkind.h" | 
|---|
| 20 | CMK_Count | 
|---|
| 21 | }; | 
|---|
| 22 |  | 
|---|
| 23 | class ArenaAllocator | 
|---|
| 24 | { | 
|---|
| 25 | private: | 
|---|
| 26 | ArenaAllocator(const ArenaAllocator& other) = delete; | 
|---|
| 27 | ArenaAllocator& operator=(const ArenaAllocator& other) = delete; | 
|---|
| 28 | ArenaAllocator& operator=(ArenaAllocator&& other) = delete; | 
|---|
| 29 |  | 
|---|
| 30 | struct PageDescriptor | 
|---|
| 31 | { | 
|---|
| 32 | PageDescriptor* m_next; | 
|---|
| 33 |  | 
|---|
| 34 | size_t m_pageBytes; // # of bytes allocated | 
|---|
| 35 | size_t m_usedBytes; // # of bytes actually used. (This is only valid when we've allocated a new page.) | 
|---|
| 36 | // See ArenaAllocator::allocateNewPage. | 
|---|
| 37 |  | 
|---|
| 38 | BYTE m_contents[]; | 
|---|
| 39 | }; | 
|---|
| 40 |  | 
|---|
| 41 | enum | 
|---|
| 42 | { | 
|---|
| 43 | DEFAULT_PAGE_SIZE = 0x10000, | 
|---|
| 44 | }; | 
|---|
| 45 |  | 
|---|
| 46 | PageDescriptor* m_firstPage; | 
|---|
| 47 | PageDescriptor* m_lastPage; | 
|---|
| 48 |  | 
|---|
| 49 | // These two pointers (when non-null) will always point into 'm_lastPage'. | 
|---|
| 50 | BYTE* m_nextFreeByte; | 
|---|
| 51 | BYTE* m_lastFreeByte; | 
|---|
| 52 |  | 
|---|
| 53 | void* allocateNewPage(size_t size); | 
|---|
| 54 |  | 
|---|
| 55 | static void* allocateHostMemory(size_t size, size_t* pActualSize); | 
|---|
| 56 | static void freeHostMemory(void* block, size_t size); | 
|---|
| 57 |  | 
|---|
| 58 | #if MEASURE_MEM_ALLOC | 
|---|
| 59 | struct MemStats | 
|---|
| 60 | { | 
|---|
| 61 | unsigned allocCnt;                 // # of allocs | 
|---|
| 62 | UINT64   allocSz;                  // total size of those alloc. | 
|---|
| 63 | UINT64   allocSzMax;               // Maximum single allocation. | 
|---|
| 64 | UINT64   allocSzByKind[CMK_Count]; // Classified by "kind". | 
|---|
| 65 | UINT64   nraTotalSizeAlloc; | 
|---|
| 66 | UINT64   nraTotalSizeUsed; | 
|---|
| 67 |  | 
|---|
| 68 | static const char* s_CompMemKindNames[]; // Names of the kinds. | 
|---|
| 69 |  | 
|---|
| 70 | void AddAlloc(size_t sz, CompMemKind cmk) | 
|---|
| 71 | { | 
|---|
| 72 | allocCnt += 1; | 
|---|
| 73 | allocSz += sz; | 
|---|
| 74 | if (sz > allocSzMax) | 
|---|
| 75 | { | 
|---|
| 76 | allocSzMax = sz; | 
|---|
| 77 | } | 
|---|
| 78 | allocSzByKind[cmk] += sz; | 
|---|
| 79 | } | 
|---|
| 80 |  | 
|---|
| 81 | void Print(FILE* f);       // Print these stats to file. | 
|---|
| 82 | void PrintByKind(FILE* f); // Do just the by-kind histogram part. | 
|---|
| 83 | }; | 
|---|
| 84 |  | 
|---|
| 85 | struct AggregateMemStats : public MemStats | 
|---|
| 86 | { | 
|---|
| 87 | unsigned nMethods; | 
|---|
| 88 |  | 
|---|
| 89 | void Add(const MemStats& ms) | 
|---|
| 90 | { | 
|---|
| 91 | nMethods++; | 
|---|
| 92 | allocCnt += ms.allocCnt; | 
|---|
| 93 | allocSz += ms.allocSz; | 
|---|
| 94 | allocSzMax = max(allocSzMax, ms.allocSzMax); | 
|---|
| 95 | for (int i = 0; i < CMK_Count; i++) | 
|---|
| 96 | { | 
|---|
| 97 | allocSzByKind[i] += ms.allocSzByKind[i]; | 
|---|
| 98 | } | 
|---|
| 99 | nraTotalSizeAlloc += ms.nraTotalSizeAlloc; | 
|---|
| 100 | nraTotalSizeUsed += ms.nraTotalSizeUsed; | 
|---|
| 101 | } | 
|---|
| 102 |  | 
|---|
| 103 | void Print(FILE* f); // Print these stats to file. | 
|---|
| 104 | }; | 
|---|
| 105 |  | 
|---|
| 106 | public: | 
|---|
| 107 | struct MemStatsAllocator | 
|---|
| 108 | { | 
|---|
| 109 | ArenaAllocator* m_arena; | 
|---|
| 110 | CompMemKind     m_kind; | 
|---|
| 111 |  | 
|---|
| 112 | void* allocateMemory(size_t sz) | 
|---|
| 113 | { | 
|---|
| 114 | m_arena->m_stats.AddAlloc(sz, m_kind); | 
|---|
| 115 | return m_arena->allocateMemory(sz); | 
|---|
| 116 | } | 
|---|
| 117 | }; | 
|---|
| 118 |  | 
|---|
| 119 | private: | 
|---|
| 120 | static CritSecObject     s_statsLock; // This lock protects the data structures below. | 
|---|
| 121 | static MemStats          s_maxStats;  // Stats for the allocator with the largest amount allocated. | 
|---|
| 122 | static AggregateMemStats s_aggStats;  // Aggregates statistics for all allocators. | 
|---|
| 123 |  | 
|---|
| 124 | MemStats          m_stats; | 
|---|
| 125 | MemStatsAllocator m_statsAllocators[CMK_Count]; | 
|---|
| 126 |  | 
|---|
| 127 | public: | 
|---|
| 128 | MemStatsAllocator* getMemStatsAllocator(CompMemKind kind); | 
|---|
| 129 | void finishMemStats(); | 
|---|
| 130 | void dumpMemStats(FILE* file); | 
|---|
| 131 |  | 
|---|
| 132 | static void dumpMaxMemStats(FILE* file); | 
|---|
| 133 | static void dumpAggregateMemStats(FILE* file); | 
|---|
| 134 | #endif // MEASURE_MEM_ALLOC | 
|---|
| 135 |  | 
|---|
| 136 | public: | 
|---|
| 137 | ArenaAllocator(); | 
|---|
| 138 |  | 
|---|
| 139 | // NOTE: it would be nice to have a destructor on this type to ensure that any value that | 
|---|
| 140 | //       goes out of scope is either uninitialized or has been torn down via a call to | 
|---|
| 141 | //       destroy(), but this interacts badly in methods that use SEH. #3058 tracks | 
|---|
| 142 | //       revisiting EH in the JIT; such a destructor could be added if SEH is removed | 
|---|
| 143 | //       as part of that work. | 
|---|
| 144 |  | 
|---|
| 145 | void destroy(); | 
|---|
| 146 |  | 
|---|
| 147 | inline void* allocateMemory(size_t sz); | 
|---|
| 148 |  | 
|---|
| 149 | size_t getTotalBytesAllocated(); | 
|---|
| 150 | size_t getTotalBytesUsed(); | 
|---|
| 151 |  | 
|---|
| 152 | static bool   bypassHostAllocator(); | 
|---|
| 153 | static size_t getDefaultPageSize(); | 
|---|
| 154 | }; | 
|---|
| 155 |  | 
|---|
| 156 | //------------------------------------------------------------------------ | 
|---|
| 157 | // ArenaAllocator::allocateMemory: | 
|---|
| 158 | //    Allocates memory using an `ArenaAllocator`. | 
|---|
| 159 | // | 
|---|
| 160 | // Arguments: | 
|---|
| 161 | //    size - The number of bytes to allocate. | 
|---|
| 162 | // | 
|---|
| 163 | // Return Value: | 
|---|
| 164 | //    A pointer to the allocated memory. | 
|---|
| 165 | // | 
|---|
| 166 | // Note: | 
|---|
| 167 | //    The DEBUG version of the method has some abilities that the release | 
|---|
| 168 | //    version does not: it may inject faults into the allocator and | 
|---|
| 169 | //    seeds all allocations with a specified pattern to help catch | 
|---|
| 170 | //    use-before-init problems. | 
|---|
| 171 | // | 
|---|
| 172 | inline void* ArenaAllocator::allocateMemory(size_t size) | 
|---|
| 173 | { | 
|---|
| 174 | assert(size != 0); | 
|---|
| 175 |  | 
|---|
| 176 | // Ensure that we always allocate in pointer sized increments. | 
|---|
| 177 | size = roundUp(size, sizeof(size_t)); | 
|---|
| 178 |  | 
|---|
| 179 | #if defined(DEBUG) | 
|---|
| 180 | if (JitConfig.ShouldInjectFault() != 0) | 
|---|
| 181 | { | 
|---|
| 182 | // Force the underlying memory allocator (either the OS or the CLR hoster) | 
|---|
| 183 | // to allocate the memory. Any fault injection will kick in. | 
|---|
| 184 | void* p = ClrAllocInProcessHeap(0, S_SIZE_T(1)); | 
|---|
| 185 | if (p != nullptr) | 
|---|
| 186 | { | 
|---|
| 187 | ClrFreeInProcessHeap(0, p); | 
|---|
| 188 | } | 
|---|
| 189 | else | 
|---|
| 190 | { | 
|---|
| 191 | NOMEM(); // Throw! | 
|---|
| 192 | } | 
|---|
| 193 | } | 
|---|
| 194 | #endif | 
|---|
| 195 |  | 
|---|
| 196 | void* block = m_nextFreeByte; | 
|---|
| 197 | m_nextFreeByte += size; | 
|---|
| 198 |  | 
|---|
| 199 | if (m_nextFreeByte > m_lastFreeByte) | 
|---|
| 200 | { | 
|---|
| 201 | block = allocateNewPage(size); | 
|---|
| 202 | } | 
|---|
| 203 |  | 
|---|
| 204 | #if defined(DEBUG) | 
|---|
| 205 | memset(block, UninitializedWord<char>(nullptr), size); | 
|---|
| 206 | #endif | 
|---|
| 207 |  | 
|---|
| 208 | return block; | 
|---|
| 209 | } | 
|---|
| 210 |  | 
|---|
| 211 | // Allows general purpose code (e.g. collection classes) to allocate | 
|---|
| 212 | // memory of a pre-determined kind via an arena allocator. | 
|---|
| 213 |  | 
|---|
| 214 | class CompAllocator | 
|---|
| 215 | { | 
|---|
| 216 | #if MEASURE_MEM_ALLOC | 
|---|
| 217 | ArenaAllocator::MemStatsAllocator* m_arena; | 
|---|
| 218 | #else | 
|---|
| 219 | ArenaAllocator* m_arena; | 
|---|
| 220 | #endif | 
|---|
| 221 |  | 
|---|
| 222 | public: | 
|---|
| 223 | CompAllocator(ArenaAllocator* arena, CompMemKind cmk) | 
|---|
| 224 | #if MEASURE_MEM_ALLOC | 
|---|
| 225 | : m_arena(arena->getMemStatsAllocator(cmk)) | 
|---|
| 226 | #else | 
|---|
| 227 | : m_arena(arena) | 
|---|
| 228 | #endif | 
|---|
| 229 | { | 
|---|
| 230 | } | 
|---|
| 231 |  | 
|---|
| 232 | // Allocate a block of memory suitable to store `count` objects of type `T`. | 
|---|
| 233 | // Zero-length allocations are not allowed. | 
|---|
| 234 | template <typename T> | 
|---|
| 235 | T* allocate(size_t count) | 
|---|
| 236 | { | 
|---|
| 237 | // Ensure that count * sizeof(T) does not overflow. | 
|---|
| 238 | if (count > (SIZE_MAX / sizeof(T))) | 
|---|
| 239 | { | 
|---|
| 240 | NOMEM(); | 
|---|
| 241 | } | 
|---|
| 242 |  | 
|---|
| 243 | void* p = m_arena->allocateMemory(count * sizeof(T)); | 
|---|
| 244 |  | 
|---|
| 245 | // Ensure that the allocator returned sizeof(size_t) aligned memory. | 
|---|
| 246 | assert((size_t(p) & (sizeof(size_t) - 1)) == 0); | 
|---|
| 247 |  | 
|---|
| 248 | return static_cast<T*>(p); | 
|---|
| 249 | } | 
|---|
| 250 |  | 
|---|
| 251 | // Deallocate a block of memory previously allocated by `allocate`. | 
|---|
| 252 | // The arena allocator does not release memory so this doesn't do anything. | 
|---|
| 253 | void deallocate(void* p) | 
|---|
| 254 | { | 
|---|
| 255 | } | 
|---|
| 256 | }; | 
|---|
| 257 |  | 
|---|
| 258 | // Global operator new overloads that work with CompAllocator | 
|---|
| 259 |  | 
|---|
| 260 | inline void* __cdecl operator new(size_t n, CompAllocator alloc) | 
|---|
| 261 | { | 
|---|
| 262 | return alloc.allocate<char>(n); | 
|---|
| 263 | } | 
|---|
| 264 |  | 
|---|
| 265 | inline void* __cdecl operator new[](size_t n, CompAllocator alloc) | 
|---|
| 266 | { | 
|---|
| 267 | return alloc.allocate<char>(n); | 
|---|
| 268 | } | 
|---|
| 269 |  | 
|---|
| 270 | // A CompAllocator wrapper that implements IAllocator and allows zero-length | 
|---|
| 271 | // memory allocations (the arena allocator does not support zero-length | 
|---|
| 272 | // allocation). | 
|---|
| 273 |  | 
|---|
| 274 | class CompIAllocator : public IAllocator | 
|---|
| 275 | { | 
|---|
| 276 | CompAllocator m_alloc; | 
|---|
| 277 | char          m_zeroLenAllocTarg; | 
|---|
| 278 |  | 
|---|
| 279 | public: | 
|---|
| 280 | CompIAllocator(CompAllocator alloc) : m_alloc(alloc) | 
|---|
| 281 | { | 
|---|
| 282 | } | 
|---|
| 283 |  | 
|---|
| 284 | // Allocates a block of memory at least `sz` in size. | 
|---|
| 285 | virtual void* Alloc(size_t sz) override | 
|---|
| 286 | { | 
|---|
| 287 | if (sz == 0) | 
|---|
| 288 | { | 
|---|
| 289 | return &m_zeroLenAllocTarg; | 
|---|
| 290 | } | 
|---|
| 291 | else | 
|---|
| 292 | { | 
|---|
| 293 | return m_alloc.allocate<char>(sz); | 
|---|
| 294 | } | 
|---|
| 295 | } | 
|---|
| 296 |  | 
|---|
| 297 | // Allocates a block of memory at least `elems * elemSize` in size. | 
|---|
| 298 | virtual void* ArrayAlloc(size_t elems, size_t elemSize) override | 
|---|
| 299 | { | 
|---|
| 300 | if ((elems == 0) || (elemSize == 0)) | 
|---|
| 301 | { | 
|---|
| 302 | return &m_zeroLenAllocTarg; | 
|---|
| 303 | } | 
|---|
| 304 | else | 
|---|
| 305 | { | 
|---|
| 306 | // Ensure that elems * elemSize does not overflow. | 
|---|
| 307 | if (elems > (SIZE_MAX / elemSize)) | 
|---|
| 308 | { | 
|---|
| 309 | NOMEM(); | 
|---|
| 310 | } | 
|---|
| 311 |  | 
|---|
| 312 | return m_alloc.allocate<char>(elems * elemSize); | 
|---|
| 313 | } | 
|---|
| 314 | } | 
|---|
| 315 |  | 
|---|
| 316 | // Frees the block of memory pointed to by p. | 
|---|
| 317 | virtual void Free(void* p) override | 
|---|
| 318 | { | 
|---|
| 319 | m_alloc.deallocate(p); | 
|---|
| 320 | } | 
|---|
| 321 | }; | 
|---|
| 322 |  | 
|---|
| 323 | #endif // _ALLOC_H_ | 
|---|
| 324 |  | 
|---|