1 | /* |
2 | * Copyright (c) 2018, Oracle and/or its affiliates. All rights reserved. |
3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
4 | * |
5 | * This code is free software; you can redistribute it and/or modify it |
6 | * under the terms of the GNU General Public License version 2 only, as |
7 | * published by the Free Software Foundation. |
8 | * |
9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
12 | * version 2 for more details (a copy is included in the LICENSE file that |
13 | * accompanied this code). |
14 | * |
15 | * You should have received a copy of the GNU General Public License version |
16 | * 2 along with this work; if not, write to the Free Software Foundation, |
17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
18 | * |
19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
20 | * or visit www.oracle.com if you need additional information or have any |
21 | * questions. |
22 | * |
23 | */ |
24 | |
25 | #include "precompiled.hpp" |
26 | #include "gc/shared/barrierSet.hpp" |
27 | #include "gc/shared/barrierSetAssembler.hpp" |
28 | #include "gc/shared/barrierSetNMethod.hpp" |
29 | #include "gc/shared/collectedHeap.hpp" |
30 | #include "interpreter/interp_masm.hpp" |
31 | #include "memory/universe.hpp" |
32 | #include "runtime/jniHandles.hpp" |
33 | #include "runtime/sharedRuntime.hpp" |
34 | #include "runtime/thread.hpp" |
35 | |
36 | #define __ masm-> |
37 | |
38 | void BarrierSetAssembler::load_at(MacroAssembler* masm, DecoratorSet decorators, BasicType type, |
39 | Register dst, Address src, Register tmp1, Register tmp_thread) { |
40 | bool in_heap = (decorators & IN_HEAP) != 0; |
41 | bool in_native = (decorators & IN_NATIVE) != 0; |
42 | bool is_not_null = (decorators & IS_NOT_NULL) != 0; |
43 | bool atomic = (decorators & MO_RELAXED) != 0; |
44 | |
45 | switch (type) { |
46 | case T_OBJECT: |
47 | case T_ARRAY: { |
48 | if (in_heap) { |
49 | #ifdef _LP64 |
50 | if (UseCompressedOops) { |
51 | __ movl(dst, src); |
52 | if (is_not_null) { |
53 | __ decode_heap_oop_not_null(dst); |
54 | } else { |
55 | __ decode_heap_oop(dst); |
56 | } |
57 | } else |
58 | #endif |
59 | { |
60 | __ movptr(dst, src); |
61 | } |
62 | } else { |
63 | assert(in_native, "why else?" ); |
64 | __ movptr(dst, src); |
65 | } |
66 | break; |
67 | } |
68 | case T_BOOLEAN: __ load_unsigned_byte(dst, src); break; |
69 | case T_BYTE: __ load_signed_byte(dst, src); break; |
70 | case T_CHAR: __ load_unsigned_short(dst, src); break; |
71 | case T_SHORT: __ load_signed_short(dst, src); break; |
72 | case T_INT: __ movl (dst, src); break; |
73 | case T_ADDRESS: __ movptr(dst, src); break; |
74 | case T_FLOAT: |
75 | assert(dst == noreg, "only to ftos" ); |
76 | __ load_float(src); |
77 | break; |
78 | case T_DOUBLE: |
79 | assert(dst == noreg, "only to dtos" ); |
80 | __ load_double(src); |
81 | break; |
82 | case T_LONG: |
83 | assert(dst == noreg, "only to ltos" ); |
84 | #ifdef _LP64 |
85 | __ movq(rax, src); |
86 | #else |
87 | if (atomic) { |
88 | __ fild_d(src); // Must load atomically |
89 | __ subptr(rsp,2*wordSize); // Make space for store |
90 | __ fistp_d(Address(rsp,0)); |
91 | __ pop(rax); |
92 | __ pop(rdx); |
93 | } else { |
94 | __ movl(rax, src); |
95 | __ movl(rdx, src.plus_disp(wordSize)); |
96 | } |
97 | #endif |
98 | break; |
99 | default: Unimplemented(); |
100 | } |
101 | } |
102 | |
103 | void BarrierSetAssembler::store_at(MacroAssembler* masm, DecoratorSet decorators, BasicType type, |
104 | Address dst, Register val, Register tmp1, Register tmp2) { |
105 | bool in_heap = (decorators & IN_HEAP) != 0; |
106 | bool in_native = (decorators & IN_NATIVE) != 0; |
107 | bool is_not_null = (decorators & IS_NOT_NULL) != 0; |
108 | bool atomic = (decorators & MO_RELAXED) != 0; |
109 | |
110 | switch (type) { |
111 | case T_OBJECT: |
112 | case T_ARRAY: { |
113 | if (in_heap) { |
114 | if (val == noreg) { |
115 | assert(!is_not_null, "inconsistent access" ); |
116 | #ifdef _LP64 |
117 | if (UseCompressedOops) { |
118 | __ movl(dst, (int32_t)NULL_WORD); |
119 | } else { |
120 | __ movslq(dst, (int32_t)NULL_WORD); |
121 | } |
122 | #else |
123 | __ movl(dst, (int32_t)NULL_WORD); |
124 | #endif |
125 | } else { |
126 | #ifdef _LP64 |
127 | if (UseCompressedOops) { |
128 | assert(!dst.uses(val), "not enough registers" ); |
129 | if (is_not_null) { |
130 | __ encode_heap_oop_not_null(val); |
131 | } else { |
132 | __ encode_heap_oop(val); |
133 | } |
134 | __ movl(dst, val); |
135 | } else |
136 | #endif |
137 | { |
138 | __ movptr(dst, val); |
139 | } |
140 | } |
141 | } else { |
142 | assert(in_native, "why else?" ); |
143 | assert(val != noreg, "not supported" ); |
144 | __ movptr(dst, val); |
145 | } |
146 | break; |
147 | } |
148 | case T_BOOLEAN: |
149 | __ andl(val, 0x1); // boolean is true if LSB is 1 |
150 | __ movb(dst, val); |
151 | break; |
152 | case T_BYTE: |
153 | __ movb(dst, val); |
154 | break; |
155 | case T_SHORT: |
156 | __ movw(dst, val); |
157 | break; |
158 | case T_CHAR: |
159 | __ movw(dst, val); |
160 | break; |
161 | case T_INT: |
162 | __ movl(dst, val); |
163 | break; |
164 | case T_LONG: |
165 | assert(val == noreg, "only tos" ); |
166 | #ifdef _LP64 |
167 | __ movq(dst, rax); |
168 | #else |
169 | if (atomic) { |
170 | __ push(rdx); |
171 | __ push(rax); // Must update atomically with FIST |
172 | __ fild_d(Address(rsp,0)); // So load into FPU register |
173 | __ fistp_d(dst); // and put into memory atomically |
174 | __ addptr(rsp, 2*wordSize); |
175 | } else { |
176 | __ movptr(dst, rax); |
177 | __ movptr(dst.plus_disp(wordSize), rdx); |
178 | } |
179 | #endif |
180 | break; |
181 | case T_FLOAT: |
182 | assert(val == noreg, "only tos" ); |
183 | __ store_float(dst); |
184 | break; |
185 | case T_DOUBLE: |
186 | assert(val == noreg, "only tos" ); |
187 | __ store_double(dst); |
188 | break; |
189 | case T_ADDRESS: |
190 | __ movptr(dst, val); |
191 | break; |
192 | default: Unimplemented(); |
193 | } |
194 | } |
195 | |
196 | #ifndef _LP64 |
197 | void BarrierSetAssembler::obj_equals(MacroAssembler* masm, |
198 | Address obj1, jobject obj2) { |
199 | __ cmpoop_raw(obj1, obj2); |
200 | } |
201 | |
202 | void BarrierSetAssembler::obj_equals(MacroAssembler* masm, |
203 | Register obj1, jobject obj2) { |
204 | __ cmpoop_raw(obj1, obj2); |
205 | } |
206 | #endif |
207 | void BarrierSetAssembler::obj_equals(MacroAssembler* masm, |
208 | Register obj1, Address obj2) { |
209 | __ cmpptr(obj1, obj2); |
210 | } |
211 | |
212 | void BarrierSetAssembler::obj_equals(MacroAssembler* masm, |
213 | Register obj1, Register obj2) { |
214 | __ cmpptr(obj1, obj2); |
215 | } |
216 | |
217 | void BarrierSetAssembler::try_resolve_jobject_in_native(MacroAssembler* masm, Register jni_env, |
218 | Register obj, Register tmp, Label& slowpath) { |
219 | __ clear_jweak_tag(obj); |
220 | __ movptr(obj, Address(obj, 0)); |
221 | } |
222 | |
223 | void BarrierSetAssembler::tlab_allocate(MacroAssembler* masm, |
224 | Register thread, Register obj, |
225 | Register var_size_in_bytes, |
226 | int con_size_in_bytes, |
227 | Register t1, |
228 | Register t2, |
229 | Label& slow_case) { |
230 | assert_different_registers(obj, t1, t2); |
231 | assert_different_registers(obj, var_size_in_bytes, t1); |
232 | Register end = t2; |
233 | if (!thread->is_valid()) { |
234 | #ifdef _LP64 |
235 | thread = r15_thread; |
236 | #else |
237 | assert(t1->is_valid(), "need temp reg" ); |
238 | thread = t1; |
239 | __ get_thread(thread); |
240 | #endif |
241 | } |
242 | |
243 | __ verify_tlab(); |
244 | |
245 | __ movptr(obj, Address(thread, JavaThread::tlab_top_offset())); |
246 | if (var_size_in_bytes == noreg) { |
247 | __ lea(end, Address(obj, con_size_in_bytes)); |
248 | } else { |
249 | __ lea(end, Address(obj, var_size_in_bytes, Address::times_1)); |
250 | } |
251 | __ cmpptr(end, Address(thread, JavaThread::tlab_end_offset())); |
252 | __ jcc(Assembler::above, slow_case); |
253 | |
254 | // update the tlab top pointer |
255 | __ movptr(Address(thread, JavaThread::tlab_top_offset()), end); |
256 | |
257 | // recover var_size_in_bytes if necessary |
258 | if (var_size_in_bytes == end) { |
259 | __ subptr(var_size_in_bytes, obj); |
260 | } |
261 | __ verify_tlab(); |
262 | } |
263 | |
264 | // Defines obj, preserves var_size_in_bytes |
265 | void BarrierSetAssembler::eden_allocate(MacroAssembler* masm, |
266 | Register thread, Register obj, |
267 | Register var_size_in_bytes, |
268 | int con_size_in_bytes, |
269 | Register t1, |
270 | Label& slow_case) { |
271 | assert(obj == rax, "obj must be in rax, for cmpxchg" ); |
272 | assert_different_registers(obj, var_size_in_bytes, t1); |
273 | if (!Universe::heap()->supports_inline_contig_alloc()) { |
274 | __ jmp(slow_case); |
275 | } else { |
276 | Register end = t1; |
277 | Label retry; |
278 | __ bind(retry); |
279 | ExternalAddress heap_top((address) Universe::heap()->top_addr()); |
280 | __ movptr(obj, heap_top); |
281 | if (var_size_in_bytes == noreg) { |
282 | __ lea(end, Address(obj, con_size_in_bytes)); |
283 | } else { |
284 | __ lea(end, Address(obj, var_size_in_bytes, Address::times_1)); |
285 | } |
286 | // if end < obj then we wrapped around => object too long => slow case |
287 | __ cmpptr(end, obj); |
288 | __ jcc(Assembler::below, slow_case); |
289 | __ cmpptr(end, ExternalAddress((address) Universe::heap()->end_addr())); |
290 | __ jcc(Assembler::above, slow_case); |
291 | // Compare obj with the top addr, and if still equal, store the new top addr in |
292 | // end at the address of the top addr pointer. Sets ZF if was equal, and clears |
293 | // it otherwise. Use lock prefix for atomicity on MPs. |
294 | __ locked_cmpxchgptr(end, heap_top); |
295 | __ jcc(Assembler::notEqual, retry); |
296 | incr_allocated_bytes(masm, thread, var_size_in_bytes, con_size_in_bytes, thread->is_valid() ? noreg : t1); |
297 | } |
298 | } |
299 | |
300 | void BarrierSetAssembler::incr_allocated_bytes(MacroAssembler* masm, Register thread, |
301 | Register var_size_in_bytes, |
302 | int con_size_in_bytes, |
303 | Register t1) { |
304 | if (!thread->is_valid()) { |
305 | #ifdef _LP64 |
306 | thread = r15_thread; |
307 | #else |
308 | assert(t1->is_valid(), "need temp reg" ); |
309 | thread = t1; |
310 | __ get_thread(thread); |
311 | #endif |
312 | } |
313 | |
314 | #ifdef _LP64 |
315 | if (var_size_in_bytes->is_valid()) { |
316 | __ addq(Address(thread, in_bytes(JavaThread::allocated_bytes_offset())), var_size_in_bytes); |
317 | } else { |
318 | __ addq(Address(thread, in_bytes(JavaThread::allocated_bytes_offset())), con_size_in_bytes); |
319 | } |
320 | #else |
321 | if (var_size_in_bytes->is_valid()) { |
322 | __ addl(Address(thread, in_bytes(JavaThread::allocated_bytes_offset())), var_size_in_bytes); |
323 | } else { |
324 | __ addl(Address(thread, in_bytes(JavaThread::allocated_bytes_offset())), con_size_in_bytes); |
325 | } |
326 | __ adcl(Address(thread, in_bytes(JavaThread::allocated_bytes_offset())+4), 0); |
327 | #endif |
328 | } |
329 | |
330 | void BarrierSetAssembler::nmethod_entry_barrier(MacroAssembler* masm) { |
331 | BarrierSetNMethod* bs_nm = BarrierSet::barrier_set()->barrier_set_nmethod(); |
332 | if (bs_nm == NULL) { |
333 | return; |
334 | } |
335 | #ifndef _LP64 |
336 | ShouldNotReachHere(); |
337 | #else |
338 | Label continuation; |
339 | Register thread = LP64_ONLY(r15_thread); |
340 | Address disarmed_addr(thread, in_bytes(bs_nm->thread_disarmed_offset())); |
341 | __ align(8); |
342 | __ cmpl(disarmed_addr, 0); |
343 | __ jcc(Assembler::equal, continuation); |
344 | __ call(RuntimeAddress(StubRoutines::x86::method_entry_barrier())); |
345 | __ bind(continuation); |
346 | #endif |
347 | } |
348 | |
349 | void BarrierSetAssembler::c2i_entry_barrier(MacroAssembler* masm) { |
350 | BarrierSetNMethod* bs = BarrierSet::barrier_set()->barrier_set_nmethod(); |
351 | if (bs == NULL) { |
352 | return; |
353 | } |
354 | |
355 | Label bad_call; |
356 | __ cmpptr(rbx, 0); // rbx contains the incoming method for c2i adapters. |
357 | __ jcc(Assembler::equal, bad_call); |
358 | |
359 | // Pointer chase to the method holder to find out if the method is concurrently unloading. |
360 | Label method_live; |
361 | __ load_method_holder_cld(rscratch1, rbx); |
362 | |
363 | // Is it a strong CLD? |
364 | __ movl(rscratch2, Address(rscratch1, ClassLoaderData::keep_alive_offset())); |
365 | __ cmpptr(rscratch2, 0); |
366 | __ jcc(Assembler::greater, method_live); |
367 | |
368 | // Is it a weak but alive CLD? |
369 | __ movptr(rscratch1, Address(rscratch1, ClassLoaderData::holder_offset())); |
370 | __ resolve_weak_handle(rscratch1, rscratch2); |
371 | __ cmpptr(rscratch1, 0); |
372 | __ jcc(Assembler::notEqual, method_live); |
373 | |
374 | __ bind(bad_call); |
375 | __ jump(RuntimeAddress(SharedRuntime::get_handle_wrong_method_stub())); |
376 | __ bind(method_live); |
377 | } |
378 | |