1 | #include "jemalloc/internal/jemalloc_preamble.h" |
2 | |
3 | #include "jemalloc/internal/hook.h" |
4 | |
5 | #include "jemalloc/internal/atomic.h" |
6 | #include "jemalloc/internal/mutex.h" |
7 | #include "jemalloc/internal/seq.h" |
8 | |
9 | typedef struct hooks_internal_s hooks_internal_t; |
10 | struct hooks_internal_s { |
11 | hooks_t hooks; |
12 | bool in_use; |
13 | }; |
14 | |
15 | seq_define(hooks_internal_t, hooks) |
16 | |
17 | static atomic_u_t nhooks = ATOMIC_INIT(0); |
18 | static seq_hooks_t hooks[HOOK_MAX]; |
19 | static malloc_mutex_t hooks_mu; |
20 | |
21 | bool |
22 | hook_boot() { |
23 | return malloc_mutex_init(&hooks_mu, "hooks" , WITNESS_RANK_HOOK, |
24 | malloc_mutex_rank_exclusive); |
25 | } |
26 | |
27 | static void * |
28 | hook_install_locked(hooks_t *to_install) { |
29 | hooks_internal_t hooks_internal; |
30 | for (int i = 0; i < HOOK_MAX; i++) { |
31 | bool success = seq_try_load_hooks(&hooks_internal, &hooks[i]); |
32 | /* We hold mu; no concurrent access. */ |
33 | assert(success); |
34 | if (!hooks_internal.in_use) { |
35 | hooks_internal.hooks = *to_install; |
36 | hooks_internal.in_use = true; |
37 | seq_store_hooks(&hooks[i], &hooks_internal); |
38 | atomic_store_u(&nhooks, |
39 | atomic_load_u(&nhooks, ATOMIC_RELAXED) + 1, |
40 | ATOMIC_RELAXED); |
41 | return &hooks[i]; |
42 | } |
43 | } |
44 | return NULL; |
45 | } |
46 | |
47 | void * |
48 | hook_install(tsdn_t *tsdn, hooks_t *to_install) { |
49 | malloc_mutex_lock(tsdn, &hooks_mu); |
50 | void *ret = hook_install_locked(to_install); |
51 | if (ret != NULL) { |
52 | tsd_global_slow_inc(tsdn); |
53 | } |
54 | malloc_mutex_unlock(tsdn, &hooks_mu); |
55 | return ret; |
56 | } |
57 | |
58 | static void |
59 | hook_remove_locked(seq_hooks_t *to_remove) { |
60 | hooks_internal_t hooks_internal; |
61 | bool success = seq_try_load_hooks(&hooks_internal, to_remove); |
62 | /* We hold mu; no concurrent access. */ |
63 | assert(success); |
64 | /* Should only remove hooks that were added. */ |
65 | assert(hooks_internal.in_use); |
66 | hooks_internal.in_use = false; |
67 | seq_store_hooks(to_remove, &hooks_internal); |
68 | atomic_store_u(&nhooks, atomic_load_u(&nhooks, ATOMIC_RELAXED) - 1, |
69 | ATOMIC_RELAXED); |
70 | } |
71 | |
72 | void |
73 | hook_remove(tsdn_t *tsdn, void *opaque) { |
74 | if (config_debug) { |
75 | char *hooks_begin = (char *)&hooks[0]; |
76 | char *hooks_end = (char *)&hooks[HOOK_MAX]; |
77 | char *hook = (char *)opaque; |
78 | assert(hooks_begin <= hook && hook < hooks_end |
79 | && (hook - hooks_begin) % sizeof(seq_hooks_t) == 0); |
80 | } |
81 | malloc_mutex_lock(tsdn, &hooks_mu); |
82 | hook_remove_locked((seq_hooks_t *)opaque); |
83 | tsd_global_slow_dec(tsdn); |
84 | malloc_mutex_unlock(tsdn, &hooks_mu); |
85 | } |
86 | |
87 | #define FOR_EACH_HOOK_BEGIN(hooks_internal_ptr) \ |
88 | for (int for_each_hook_counter = 0; \ |
89 | for_each_hook_counter < HOOK_MAX; \ |
90 | for_each_hook_counter++) { \ |
91 | bool for_each_hook_success = seq_try_load_hooks( \ |
92 | (hooks_internal_ptr), &hooks[for_each_hook_counter]); \ |
93 | if (!for_each_hook_success) { \ |
94 | continue; \ |
95 | } \ |
96 | if (!(hooks_internal_ptr)->in_use) { \ |
97 | continue; \ |
98 | } |
99 | #define FOR_EACH_HOOK_END \ |
100 | } |
101 | |
102 | static bool * |
103 | hook_reentrantp() { |
104 | /* |
105 | * We prevent user reentrancy within hooks. This is basically just a |
106 | * thread-local bool that triggers an early-exit. |
107 | * |
108 | * We don't fold in_hook into reentrancy. There are two reasons for |
109 | * this: |
110 | * - Right now, we turn on reentrancy during things like extent hook |
111 | * execution. Allocating during extent hooks is not officially |
112 | * supported, but we don't want to break it for the time being. These |
113 | * sorts of allocations should probably still be hooked, though. |
114 | * - If a hook allocates, we may want it to be relatively fast (after |
115 | * all, it executes on every allocator operation). Turning on |
116 | * reentrancy is a fairly heavyweight mode (disabling tcache, |
117 | * redirecting to arena 0, etc.). It's possible we may one day want |
118 | * to turn on reentrant mode here, if it proves too difficult to keep |
119 | * this working. But that's fairly easy for us to see; OTOH, people |
120 | * not using hooks because they're too slow is easy for us to miss. |
121 | * |
122 | * The tricky part is |
123 | * that this code might get invoked even if we don't have access to tsd. |
124 | * This function mimics getting a pointer to thread-local data, except |
125 | * that it might secretly return a pointer to some global data if we |
126 | * know that the caller will take the early-exit path. |
127 | * If we return a bool that indicates that we are reentrant, then the |
128 | * caller will go down the early exit path, leaving the global |
129 | * untouched. |
130 | */ |
131 | static bool in_hook_global = true; |
132 | tsdn_t *tsdn = tsdn_fetch(); |
133 | tcache_t *tcache = tsdn_tcachep_get(tsdn); |
134 | if (tcache != NULL) { |
135 | return &tcache->in_hook; |
136 | } |
137 | return &in_hook_global; |
138 | } |
139 | |
140 | #define HOOK_PROLOGUE \ |
141 | if (likely(atomic_load_u(&nhooks, ATOMIC_RELAXED) == 0)) { \ |
142 | return; \ |
143 | } \ |
144 | bool *in_hook = hook_reentrantp(); \ |
145 | if (*in_hook) { \ |
146 | return; \ |
147 | } \ |
148 | *in_hook = true; |
149 | |
150 | #define HOOK_EPILOGUE \ |
151 | *in_hook = false; |
152 | |
153 | void |
154 | hook_invoke_alloc(hook_alloc_t type, void *result, uintptr_t result_raw, |
155 | uintptr_t args_raw[3]) { |
156 | HOOK_PROLOGUE |
157 | |
158 | hooks_internal_t hook; |
159 | FOR_EACH_HOOK_BEGIN(&hook) |
160 | hook_alloc h = hook.hooks.alloc_hook; |
161 | if (h != NULL) { |
162 | h(hook.hooks.extra, type, result, result_raw, args_raw); |
163 | } |
164 | FOR_EACH_HOOK_END |
165 | |
166 | HOOK_EPILOGUE |
167 | } |
168 | |
169 | void |
170 | hook_invoke_dalloc(hook_dalloc_t type, void *address, uintptr_t args_raw[3]) { |
171 | HOOK_PROLOGUE |
172 | hooks_internal_t hook; |
173 | FOR_EACH_HOOK_BEGIN(&hook) |
174 | hook_dalloc h = hook.hooks.dalloc_hook; |
175 | if (h != NULL) { |
176 | h(hook.hooks.extra, type, address, args_raw); |
177 | } |
178 | FOR_EACH_HOOK_END |
179 | HOOK_EPILOGUE |
180 | } |
181 | |
182 | void |
183 | hook_invoke_expand(hook_expand_t type, void *address, size_t old_usize, |
184 | size_t new_usize, uintptr_t result_raw, uintptr_t args_raw[4]) { |
185 | HOOK_PROLOGUE |
186 | hooks_internal_t hook; |
187 | FOR_EACH_HOOK_BEGIN(&hook) |
188 | hook_expand h = hook.hooks.expand_hook; |
189 | if (h != NULL) { |
190 | h(hook.hooks.extra, type, address, old_usize, new_usize, |
191 | result_raw, args_raw); |
192 | } |
193 | FOR_EACH_HOOK_END |
194 | HOOK_EPILOGUE |
195 | } |
196 | |