| 1 | /* | 
|---|
| 2 | * Copyright (c) 2003, 2018, Oracle and/or its affiliates. All rights reserved. | 
|---|
| 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. | 
|---|
| 4 | * | 
|---|
| 5 | * This code is free software; you can redistribute it and/or modify it | 
|---|
| 6 | * under the terms of the GNU General Public License version 2 only, as | 
|---|
| 7 | * published by the Free Software Foundation. | 
|---|
| 8 | * | 
|---|
| 9 | * This code is distributed in the hope that it will be useful, but WITHOUT | 
|---|
| 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | 
|---|
| 11 | * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License | 
|---|
| 12 | * version 2 for more details (a copy is included in the LICENSE file that | 
|---|
| 13 | * accompanied this code). | 
|---|
| 14 | * | 
|---|
| 15 | * You should have received a copy of the GNU General Public License version | 
|---|
| 16 | * 2 along with this work; if not, write to the Free Software Foundation, | 
|---|
| 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. | 
|---|
| 18 | * | 
|---|
| 19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA | 
|---|
| 20 | * or visit www.oracle.com if you need additional information or have any | 
|---|
| 21 | * questions. | 
|---|
| 22 | * | 
|---|
| 23 | */ | 
|---|
| 24 |  | 
|---|
| 25 | #include "precompiled.hpp" | 
|---|
| 26 | #include "asm/macroAssembler.hpp" | 
|---|
| 27 | #include "code/vtableStubs.hpp" | 
|---|
| 28 | #include "interp_masm_x86.hpp" | 
|---|
| 29 | #include "memory/resourceArea.hpp" | 
|---|
| 30 | #include "oops/compiledICHolder.hpp" | 
|---|
| 31 | #include "oops/instanceKlass.hpp" | 
|---|
| 32 | #include "oops/klassVtable.hpp" | 
|---|
| 33 | #include "runtime/sharedRuntime.hpp" | 
|---|
| 34 | #include "vmreg_x86.inline.hpp" | 
|---|
| 35 | #ifdef COMPILER2 | 
|---|
| 36 | #include "opto/runtime.hpp" | 
|---|
| 37 | #endif | 
|---|
| 38 |  | 
|---|
| 39 | // machine-dependent part of VtableStubs: create VtableStub of correct size and | 
|---|
| 40 | // initialize its code | 
|---|
| 41 |  | 
|---|
| 42 | #define __ masm-> | 
|---|
| 43 |  | 
|---|
| 44 | #ifndef PRODUCT | 
|---|
| 45 | extern "C"void bad_compiled_vtable_index(JavaThread* thread, oop receiver, int index); | 
|---|
| 46 | #endif | 
|---|
| 47 |  | 
|---|
| 48 | VtableStub* VtableStubs::create_vtable_stub(int vtable_index) { | 
|---|
| 49 | // Read "A word on VtableStub sizing" in share/code/vtableStubs.hpp for details on stub sizing. | 
|---|
| 50 | const int stub_code_length = code_size_limit(true); | 
|---|
| 51 | VtableStub* s = new(stub_code_length) VtableStub(true, vtable_index); | 
|---|
| 52 | // Can be NULL if there is no free space in the code cache. | 
|---|
| 53 | if (s == NULL) { | 
|---|
| 54 | return NULL; | 
|---|
| 55 | } | 
|---|
| 56 |  | 
|---|
| 57 | // Count unused bytes in instruction sequences of variable size. | 
|---|
| 58 | // We add them to the computed buffer size in order to avoid | 
|---|
| 59 | // overflow in subsequently generated stubs. | 
|---|
| 60 | address   start_pc; | 
|---|
| 61 | int       slop_bytes = 0; | 
|---|
| 62 | int       slop_delta = 0; | 
|---|
| 63 | // No variance was detected in vtable stub sizes. Setting index_dependent_slop == 0 will unveil any deviation from this observation. | 
|---|
| 64 | const int index_dependent_slop     = 0; | 
|---|
| 65 |  | 
|---|
| 66 | ResourceMark    rm; | 
|---|
| 67 | CodeBuffer      cb(s->entry_point(), stub_code_length); | 
|---|
| 68 | MacroAssembler* masm = new MacroAssembler(&cb); | 
|---|
| 69 |  | 
|---|
| 70 | #if (!defined(PRODUCT) && defined(COMPILER2)) | 
|---|
| 71 | if (CountCompiledCalls) { | 
|---|
| 72 | __ incrementl(ExternalAddress((address) SharedRuntime::nof_megamorphic_calls_addr())); | 
|---|
| 73 | } | 
|---|
| 74 | #endif | 
|---|
| 75 |  | 
|---|
| 76 | // get receiver (need to skip return address on top of stack) | 
|---|
| 77 | assert(VtableStub::receiver_location() == j_rarg0->as_VMReg(), "receiver expected in j_rarg0"); | 
|---|
| 78 |  | 
|---|
| 79 | // Free registers (non-args) are rax, rbx | 
|---|
| 80 |  | 
|---|
| 81 | // get receiver klass | 
|---|
| 82 | address npe_addr = __ pc(); | 
|---|
| 83 | __ load_klass(rax, j_rarg0); | 
|---|
| 84 |  | 
|---|
| 85 | #ifndef PRODUCT | 
|---|
| 86 | if (DebugVtables) { | 
|---|
| 87 | Label L; | 
|---|
| 88 | start_pc = __ pc(); | 
|---|
| 89 | // check offset vs vtable length | 
|---|
| 90 | __ cmpl(Address(rax, Klass::vtable_length_offset()), vtable_index*vtableEntry::size()); | 
|---|
| 91 | slop_delta  = 12 - (__ pc() - start_pc);  // cmpl varies in length, depending on data | 
|---|
| 92 | slop_bytes += slop_delta; | 
|---|
| 93 | assert(slop_delta >= 0, "negative slop(%d) encountered, adjust code size estimate!", slop_delta); | 
|---|
| 94 |  | 
|---|
| 95 | __ jcc(Assembler::greater, L); | 
|---|
| 96 | __ movl(rbx, vtable_index); | 
|---|
| 97 | // VTABLE TODO: find upper bound for call_VM length. | 
|---|
| 98 | start_pc = __ pc(); | 
|---|
| 99 | __ call_VM(noreg, CAST_FROM_FN_PTR(address, bad_compiled_vtable_index), j_rarg0, rbx); | 
|---|
| 100 | slop_delta  = 480 - (__ pc() - start_pc); | 
|---|
| 101 | slop_bytes += slop_delta; | 
|---|
| 102 | assert(slop_delta >= 0, "negative slop(%d) encountered, adjust code size estimate!", slop_delta); | 
|---|
| 103 | __ bind(L); | 
|---|
| 104 | } | 
|---|
| 105 | #endif // PRODUCT | 
|---|
| 106 |  | 
|---|
| 107 | const Register method = rbx; | 
|---|
| 108 |  | 
|---|
| 109 | // load Method* and target address | 
|---|
| 110 | start_pc = __ pc(); | 
|---|
| 111 | __ lookup_virtual_method(rax, vtable_index, method); | 
|---|
| 112 | slop_delta  = 8 - (int)(__ pc() - start_pc); | 
|---|
| 113 | slop_bytes += slop_delta; | 
|---|
| 114 | assert(slop_delta >= 0, "negative slop(%d) encountered, adjust code size estimate!", slop_delta); | 
|---|
| 115 |  | 
|---|
| 116 | #ifndef PRODUCT | 
|---|
| 117 | if (DebugVtables) { | 
|---|
| 118 | Label L; | 
|---|
| 119 | __ cmpptr(method, (int32_t)NULL_WORD); | 
|---|
| 120 | __ jcc(Assembler::equal, L); | 
|---|
| 121 | __ cmpptr(Address(method, Method::from_compiled_offset()), (int32_t)NULL_WORD); | 
|---|
| 122 | __ jcc(Assembler::notZero, L); | 
|---|
| 123 | __ stop( "Vtable entry is NULL"); | 
|---|
| 124 | __ bind(L); | 
|---|
| 125 | } | 
|---|
| 126 | #endif // PRODUCT | 
|---|
| 127 |  | 
|---|
| 128 | // rax: receiver klass | 
|---|
| 129 | // method (rbx): Method* | 
|---|
| 130 | // rcx: receiver | 
|---|
| 131 | address ame_addr = __ pc(); | 
|---|
| 132 | __ jmp( Address(rbx, Method::from_compiled_offset())); | 
|---|
| 133 |  | 
|---|
| 134 | masm->flush(); | 
|---|
| 135 | slop_bytes += index_dependent_slop; // add'l slop for size variance due to large itable offsets | 
|---|
| 136 | bookkeeping(masm, tty, s, npe_addr, ame_addr, true, vtable_index, slop_bytes, index_dependent_slop); | 
|---|
| 137 |  | 
|---|
| 138 | return s; | 
|---|
| 139 | } | 
|---|
| 140 |  | 
|---|
| 141 |  | 
|---|
| 142 | VtableStub* VtableStubs::create_itable_stub(int itable_index) { | 
|---|
| 143 | // Read "A word on VtableStub sizing" in share/code/vtableStubs.hpp for details on stub sizing. | 
|---|
| 144 | const int stub_code_length = code_size_limit(false); | 
|---|
| 145 | VtableStub* s = new(stub_code_length) VtableStub(false, itable_index); | 
|---|
| 146 | // Can be NULL if there is no free space in the code cache. | 
|---|
| 147 | if (s == NULL) { | 
|---|
| 148 | return NULL; | 
|---|
| 149 | } | 
|---|
| 150 | // Count unused bytes in instruction sequences of variable size. | 
|---|
| 151 | // We add them to the computed buffer size in order to avoid | 
|---|
| 152 | // overflow in subsequently generated stubs. | 
|---|
| 153 | address   start_pc; | 
|---|
| 154 | int       slop_bytes = 0; | 
|---|
| 155 | int       slop_delta = 0; | 
|---|
| 156 | const int index_dependent_slop = (itable_index == 0) ? 4 :     // code size change with transition from 8-bit to 32-bit constant (@index == 16). | 
|---|
| 157 | (itable_index < 16) ? 3 : 0;  // index == 0 generates even shorter code. | 
|---|
| 158 |  | 
|---|
| 159 | ResourceMark    rm; | 
|---|
| 160 | CodeBuffer      cb(s->entry_point(), stub_code_length); | 
|---|
| 161 | MacroAssembler *masm = new MacroAssembler(&cb); | 
|---|
| 162 |  | 
|---|
| 163 | #if (!defined(PRODUCT) && defined(COMPILER2)) | 
|---|
| 164 | if (CountCompiledCalls) { | 
|---|
| 165 | __ incrementl(ExternalAddress((address) SharedRuntime::nof_megamorphic_calls_addr())); | 
|---|
| 166 | } | 
|---|
| 167 | #endif // PRODUCT | 
|---|
| 168 |  | 
|---|
| 169 | // Entry arguments: | 
|---|
| 170 | //  rax: CompiledICHolder | 
|---|
| 171 | //  j_rarg0: Receiver | 
|---|
| 172 |  | 
|---|
| 173 | // Most registers are in use; we'll use rax, rbx, r10, r11 | 
|---|
| 174 | // (various calling sequences use r[cd]x, r[sd]i, r[89]; stay away from them) | 
|---|
| 175 | const Register recv_klass_reg     = r10; | 
|---|
| 176 | const Register holder_klass_reg   = rax; // declaring interface klass (DECC) | 
|---|
| 177 | const Register resolved_klass_reg = rbx; // resolved interface klass (REFC) | 
|---|
| 178 | const Register temp_reg           = r11; | 
|---|
| 179 |  | 
|---|
| 180 | const Register icholder_reg = rax; | 
|---|
| 181 | __ movptr(resolved_klass_reg, Address(icholder_reg, CompiledICHolder::holder_klass_offset())); | 
|---|
| 182 | __ movptr(holder_klass_reg,   Address(icholder_reg, CompiledICHolder::holder_metadata_offset())); | 
|---|
| 183 |  | 
|---|
| 184 | Label L_no_such_interface; | 
|---|
| 185 |  | 
|---|
| 186 | // get receiver klass (also an implicit null-check) | 
|---|
| 187 | assert(VtableStub::receiver_location() == j_rarg0->as_VMReg(), "receiver expected in j_rarg0"); | 
|---|
| 188 | address npe_addr = __ pc(); | 
|---|
| 189 | __ load_klass(recv_klass_reg, j_rarg0); | 
|---|
| 190 |  | 
|---|
| 191 | start_pc = __ pc(); | 
|---|
| 192 |  | 
|---|
| 193 | // Receiver subtype check against REFC. | 
|---|
| 194 | // Destroys recv_klass_reg value. | 
|---|
| 195 | __ lookup_interface_method(// inputs: rec. class, interface | 
|---|
| 196 | recv_klass_reg, resolved_klass_reg, noreg, | 
|---|
| 197 | // outputs:  scan temp. reg1, scan temp. reg2 | 
|---|
| 198 | recv_klass_reg, temp_reg, | 
|---|
| 199 | L_no_such_interface, | 
|---|
| 200 | /*return_method=*/false); | 
|---|
| 201 |  | 
|---|
| 202 | const ptrdiff_t  typecheckSize = __ pc() - start_pc; | 
|---|
| 203 | start_pc = __ pc(); | 
|---|
| 204 |  | 
|---|
| 205 | // Get selected method from declaring class and itable index | 
|---|
| 206 | const Register method = rbx; | 
|---|
| 207 | __ load_klass(recv_klass_reg, j_rarg0);   // restore recv_klass_reg | 
|---|
| 208 | __ lookup_interface_method(// inputs: rec. class, interface, itable index | 
|---|
| 209 | recv_klass_reg, holder_klass_reg, itable_index, | 
|---|
| 210 | // outputs: method, scan temp. reg | 
|---|
| 211 | method, temp_reg, | 
|---|
| 212 | L_no_such_interface); | 
|---|
| 213 |  | 
|---|
| 214 | const ptrdiff_t  lookupSize = __ pc() - start_pc; | 
|---|
| 215 |  | 
|---|
| 216 | // We expect we need index_dependent_slop extra bytes. Reason: | 
|---|
| 217 | // The emitted code in lookup_interface_method changes when itable_index exceeds 15. | 
|---|
| 218 | // For linux, a very narrow estimate would be 112, but Solaris requires some more space (130). | 
|---|
| 219 | const ptrdiff_t estimate = 136; | 
|---|
| 220 | const ptrdiff_t codesize = typecheckSize + lookupSize + index_dependent_slop; | 
|---|
| 221 | slop_delta  = (int)(estimate - codesize); | 
|---|
| 222 | slop_bytes += slop_delta; | 
|---|
| 223 | assert(slop_delta >= 0, "itable #%d: Code size estimate (%d) for lookup_interface_method too small, required: %d", itable_index, (int)estimate, (int)codesize); | 
|---|
| 224 |  | 
|---|
| 225 | // If we take a trap while this arg is on the stack we will not | 
|---|
| 226 | // be able to walk the stack properly. This is not an issue except | 
|---|
| 227 | // when there are mistakes in this assembly code that could generate | 
|---|
| 228 | // a spurious fault. Ask me how I know... | 
|---|
| 229 |  | 
|---|
| 230 | // method (rbx): Method* | 
|---|
| 231 | // j_rarg0: receiver | 
|---|
| 232 |  | 
|---|
| 233 | #ifdef ASSERT | 
|---|
| 234 | if (DebugVtables) { | 
|---|
| 235 | Label L2; | 
|---|
| 236 | __ cmpptr(method, (int32_t)NULL_WORD); | 
|---|
| 237 | __ jcc(Assembler::equal, L2); | 
|---|
| 238 | __ cmpptr(Address(method, Method::from_compiled_offset()), (int32_t)NULL_WORD); | 
|---|
| 239 | __ jcc(Assembler::notZero, L2); | 
|---|
| 240 | __ stop( "compiler entrypoint is null"); | 
|---|
| 241 | __ bind(L2); | 
|---|
| 242 | } | 
|---|
| 243 | #endif // ASSERT | 
|---|
| 244 |  | 
|---|
| 245 | address ame_addr = __ pc(); | 
|---|
| 246 | __ jmp(Address(method, Method::from_compiled_offset())); | 
|---|
| 247 |  | 
|---|
| 248 | __ bind(L_no_such_interface); | 
|---|
| 249 | // Handle IncompatibleClassChangeError in itable stubs. | 
|---|
| 250 | // More detailed error message. | 
|---|
| 251 | // We force resolving of the call site by jumping to the "handle | 
|---|
| 252 | // wrong method" stub, and so let the interpreter runtime do all the | 
|---|
| 253 | // dirty work. | 
|---|
| 254 | __ jump(RuntimeAddress(SharedRuntime::get_handle_wrong_method_stub())); | 
|---|
| 255 |  | 
|---|
| 256 | masm->flush(); | 
|---|
| 257 | slop_bytes += index_dependent_slop; // add'l slop for size variance due to large itable offsets | 
|---|
| 258 | bookkeeping(masm, tty, s, npe_addr, ame_addr, false, itable_index, slop_bytes, index_dependent_slop); | 
|---|
| 259 |  | 
|---|
| 260 | return s; | 
|---|
| 261 | } | 
|---|
| 262 |  | 
|---|
| 263 | int VtableStub::pd_code_alignment() { | 
|---|
| 264 | // x86 cache line size is 64 bytes, but we want to limit alignment loss. | 
|---|
| 265 | const unsigned int icache_line_size = wordSize; | 
|---|
| 266 | return icache_line_size; | 
|---|
| 267 | } | 
|---|
| 268 |  | 
|---|