| 1 | #define	JEMALLOC_CHUNK_MMAP_C_ | 
|---|
| 2 | #include "jemalloc/internal/jemalloc_internal.h" | 
|---|
| 3 |  | 
|---|
| 4 | /******************************************************************************/ | 
|---|
| 5 |  | 
|---|
| 6 | static void * | 
|---|
| 7 | chunk_alloc_mmap_slow(size_t size, size_t alignment, bool *zero, bool *commit) | 
|---|
| 8 | { | 
|---|
| 9 | void *ret; | 
|---|
| 10 | size_t alloc_size; | 
|---|
| 11 |  | 
|---|
| 12 | alloc_size = size + alignment - PAGE; | 
|---|
| 13 | /* Beware size_t wrap-around. */ | 
|---|
| 14 | if (alloc_size < size) | 
|---|
| 15 | return (NULL); | 
|---|
| 16 | do { | 
|---|
| 17 | void *pages; | 
|---|
| 18 | size_t leadsize; | 
|---|
| 19 | pages = pages_map(NULL, alloc_size, commit); | 
|---|
| 20 | if (pages == NULL) | 
|---|
| 21 | return (NULL); | 
|---|
| 22 | leadsize = ALIGNMENT_CEILING((uintptr_t)pages, alignment) - | 
|---|
| 23 | (uintptr_t)pages; | 
|---|
| 24 | ret = pages_trim(pages, alloc_size, leadsize, size, commit); | 
|---|
| 25 | } while (ret == NULL); | 
|---|
| 26 |  | 
|---|
| 27 | assert(ret != NULL); | 
|---|
| 28 | *zero = true; | 
|---|
| 29 | return (ret); | 
|---|
| 30 | } | 
|---|
| 31 |  | 
|---|
| 32 | void * | 
|---|
| 33 | chunk_alloc_mmap(void *new_addr, size_t size, size_t alignment, bool *zero, | 
|---|
| 34 | bool *commit) | 
|---|
| 35 | { | 
|---|
| 36 | void *ret; | 
|---|
| 37 | size_t offset; | 
|---|
| 38 |  | 
|---|
| 39 | /* | 
|---|
| 40 | * Ideally, there would be a way to specify alignment to mmap() (like | 
|---|
| 41 | * NetBSD has), but in the absence of such a feature, we have to work | 
|---|
| 42 | * hard to efficiently create aligned mappings.  The reliable, but | 
|---|
| 43 | * slow method is to create a mapping that is over-sized, then trim the | 
|---|
| 44 | * excess.  However, that always results in one or two calls to | 
|---|
| 45 | * pages_unmap(). | 
|---|
| 46 | * | 
|---|
| 47 | * Optimistically try mapping precisely the right amount before falling | 
|---|
| 48 | * back to the slow method, with the expectation that the optimistic | 
|---|
| 49 | * approach works most of the time. | 
|---|
| 50 | */ | 
|---|
| 51 |  | 
|---|
| 52 | assert(alignment != 0); | 
|---|
| 53 | assert((alignment & chunksize_mask) == 0); | 
|---|
| 54 |  | 
|---|
| 55 | ret = pages_map(new_addr, size, commit); | 
|---|
| 56 | if (ret == NULL || ret == new_addr) | 
|---|
| 57 | return (ret); | 
|---|
| 58 | assert(new_addr == NULL); | 
|---|
| 59 | offset = ALIGNMENT_ADDR2OFFSET(ret, alignment); | 
|---|
| 60 | if (offset != 0) { | 
|---|
| 61 | pages_unmap(ret, size); | 
|---|
| 62 | return (chunk_alloc_mmap_slow(size, alignment, zero, commit)); | 
|---|
| 63 | } | 
|---|
| 64 |  | 
|---|
| 65 | assert(ret != NULL); | 
|---|
| 66 | *zero = true; | 
|---|
| 67 | return (ret); | 
|---|
| 68 | } | 
|---|
| 69 |  | 
|---|
| 70 | bool | 
|---|
| 71 | chunk_dalloc_mmap(void *chunk, size_t size) | 
|---|
| 72 | { | 
|---|
| 73 |  | 
|---|
| 74 | if (config_munmap) | 
|---|
| 75 | pages_unmap(chunk, size); | 
|---|
| 76 |  | 
|---|
| 77 | return (!config_munmap); | 
|---|
| 78 | } | 
|---|
| 79 |  | 
|---|