| 1 | /* | 
|---|
| 2 | * Copyright (c) 2001, 2019, Oracle and/or its affiliates. All rights reserved. | 
|---|
| 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. | 
|---|
| 4 | * | 
|---|
| 5 | * This code is free software; you can redistribute it and/or modify it | 
|---|
| 6 | * under the terms of the GNU General Public License version 2 only, as | 
|---|
| 7 | * published by the Free Software Foundation. | 
|---|
| 8 | * | 
|---|
| 9 | * This code is distributed in the hope that it will be useful, but WITHOUT | 
|---|
| 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | 
|---|
| 11 | * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License | 
|---|
| 12 | * version 2 for more details (a copy is included in the LICENSE file that | 
|---|
| 13 | * accompanied this code). | 
|---|
| 14 | * | 
|---|
| 15 | * You should have received a copy of the GNU General Public License version | 
|---|
| 16 | * 2 along with this work; if not, write to the Free Software Foundation, | 
|---|
| 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. | 
|---|
| 18 | * | 
|---|
| 19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA | 
|---|
| 20 | * or visit www.oracle.com if you need additional information or have any | 
|---|
| 21 | * questions. | 
|---|
| 22 | * | 
|---|
| 23 | */ | 
|---|
| 24 |  | 
|---|
| 25 | #include "precompiled.hpp" | 
|---|
| 26 | #include "ci/ciUtilities.hpp" | 
|---|
| 27 | #include "compiler/compileLog.hpp" | 
|---|
| 28 | #include "gc/shared/barrierSet.hpp" | 
|---|
| 29 | #include "gc/shared/c2/barrierSetC2.hpp" | 
|---|
| 30 | #include "interpreter/interpreter.hpp" | 
|---|
| 31 | #include "memory/resourceArea.hpp" | 
|---|
| 32 | #include "opto/addnode.hpp" | 
|---|
| 33 | #include "opto/castnode.hpp" | 
|---|
| 34 | #include "opto/convertnode.hpp" | 
|---|
| 35 | #include "opto/graphKit.hpp" | 
|---|
| 36 | #include "opto/idealKit.hpp" | 
|---|
| 37 | #include "opto/intrinsicnode.hpp" | 
|---|
| 38 | #include "opto/locknode.hpp" | 
|---|
| 39 | #include "opto/machnode.hpp" | 
|---|
| 40 | #include "opto/opaquenode.hpp" | 
|---|
| 41 | #include "opto/parse.hpp" | 
|---|
| 42 | #include "opto/rootnode.hpp" | 
|---|
| 43 | #include "opto/runtime.hpp" | 
|---|
| 44 | #include "runtime/deoptimization.hpp" | 
|---|
| 45 | #include "runtime/sharedRuntime.hpp" | 
|---|
| 46 |  | 
|---|
| 47 | //----------------------------GraphKit----------------------------------------- | 
|---|
| 48 | // Main utility constructor. | 
|---|
| 49 | GraphKit::GraphKit(JVMState* jvms) | 
|---|
| 50 | : Phase(Phase::Parser), | 
|---|
| 51 | _env(C->env()), | 
|---|
| 52 | _gvn(*C->initial_gvn()), | 
|---|
| 53 | _barrier_set(BarrierSet::barrier_set()->barrier_set_c2()) | 
|---|
| 54 | { | 
|---|
| 55 | _exceptions = jvms->map()->next_exception(); | 
|---|
| 56 | if (_exceptions != NULL)  jvms->map()->set_next_exception(NULL); | 
|---|
| 57 | set_jvms(jvms); | 
|---|
| 58 | } | 
|---|
| 59 |  | 
|---|
| 60 | // Private constructor for parser. | 
|---|
| 61 | GraphKit::GraphKit() | 
|---|
| 62 | : Phase(Phase::Parser), | 
|---|
| 63 | _env(C->env()), | 
|---|
| 64 | _gvn(*C->initial_gvn()), | 
|---|
| 65 | _barrier_set(BarrierSet::barrier_set()->barrier_set_c2()) | 
|---|
| 66 | { | 
|---|
| 67 | _exceptions = NULL; | 
|---|
| 68 | set_map(NULL); | 
|---|
| 69 | debug_only(_sp = -99); | 
|---|
| 70 | debug_only(set_bci(-99)); | 
|---|
| 71 | } | 
|---|
| 72 |  | 
|---|
| 73 |  | 
|---|
| 74 |  | 
|---|
| 75 | //---------------------------clean_stack--------------------------------------- | 
|---|
| 76 | // Clear away rubbish from the stack area of the JVM state. | 
|---|
| 77 | // This destroys any arguments that may be waiting on the stack. | 
|---|
| 78 | void GraphKit::clean_stack(int from_sp) { | 
|---|
| 79 | SafePointNode* map      = this->map(); | 
|---|
| 80 | JVMState*      jvms     = this->jvms(); | 
|---|
| 81 | int            stk_size = jvms->stk_size(); | 
|---|
| 82 | int            stkoff   = jvms->stkoff(); | 
|---|
| 83 | Node*          top      = this->top(); | 
|---|
| 84 | for (int i = from_sp; i < stk_size; i++) { | 
|---|
| 85 | if (map->in(stkoff + i) != top) { | 
|---|
| 86 | map->set_req(stkoff + i, top); | 
|---|
| 87 | } | 
|---|
| 88 | } | 
|---|
| 89 | } | 
|---|
| 90 |  | 
|---|
| 91 |  | 
|---|
| 92 | //--------------------------------sync_jvms----------------------------------- | 
|---|
| 93 | // Make sure our current jvms agrees with our parse state. | 
|---|
| 94 | JVMState* GraphKit::sync_jvms() const { | 
|---|
| 95 | JVMState* jvms = this->jvms(); | 
|---|
| 96 | jvms->set_bci(bci());       // Record the new bci in the JVMState | 
|---|
| 97 | jvms->set_sp(sp());         // Record the new sp in the JVMState | 
|---|
| 98 | assert(jvms_in_sync(), "jvms is now in sync"); | 
|---|
| 99 | return jvms; | 
|---|
| 100 | } | 
|---|
| 101 |  | 
|---|
| 102 | //--------------------------------sync_jvms_for_reexecute--------------------- | 
|---|
| 103 | // Make sure our current jvms agrees with our parse state.  This version | 
|---|
| 104 | // uses the reexecute_sp for reexecuting bytecodes. | 
|---|
| 105 | JVMState* GraphKit::sync_jvms_for_reexecute() { | 
|---|
| 106 | JVMState* jvms = this->jvms(); | 
|---|
| 107 | jvms->set_bci(bci());          // Record the new bci in the JVMState | 
|---|
| 108 | jvms->set_sp(reexecute_sp());  // Record the new sp in the JVMState | 
|---|
| 109 | return jvms; | 
|---|
| 110 | } | 
|---|
| 111 |  | 
|---|
| 112 | #ifdef ASSERT | 
|---|
| 113 | bool GraphKit::jvms_in_sync() const { | 
|---|
| 114 | Parse* parse = is_Parse(); | 
|---|
| 115 | if (parse == NULL) { | 
|---|
| 116 | if (bci() !=      jvms()->bci())          return false; | 
|---|
| 117 | if (sp()  != (int)jvms()->sp())           return false; | 
|---|
| 118 | return true; | 
|---|
| 119 | } | 
|---|
| 120 | if (jvms()->method() != parse->method())    return false; | 
|---|
| 121 | if (jvms()->bci()    != parse->bci())       return false; | 
|---|
| 122 | int jvms_sp = jvms()->sp(); | 
|---|
| 123 | if (jvms_sp          != parse->sp())        return false; | 
|---|
| 124 | int jvms_depth = jvms()->depth(); | 
|---|
| 125 | if (jvms_depth       != parse->depth())     return false; | 
|---|
| 126 | return true; | 
|---|
| 127 | } | 
|---|
| 128 |  | 
|---|
| 129 | // Local helper checks for special internal merge points | 
|---|
| 130 | // used to accumulate and merge exception states. | 
|---|
| 131 | // They are marked by the region's in(0) edge being the map itself. | 
|---|
| 132 | // Such merge points must never "escape" into the parser at large, | 
|---|
| 133 | // until they have been handed to gvn.transform. | 
|---|
| 134 | static bool is_hidden_merge(Node* reg) { | 
|---|
| 135 | if (reg == NULL)  return false; | 
|---|
| 136 | if (reg->is_Phi()) { | 
|---|
| 137 | reg = reg->in(0); | 
|---|
| 138 | if (reg == NULL)  return false; | 
|---|
| 139 | } | 
|---|
| 140 | return reg->is_Region() && reg->in(0) != NULL && reg->in(0)->is_Root(); | 
|---|
| 141 | } | 
|---|
| 142 |  | 
|---|
| 143 | void GraphKit::verify_map() const { | 
|---|
| 144 | if (map() == NULL)  return;  // null map is OK | 
|---|
| 145 | assert(map()->req() <= jvms()->endoff(), "no extra garbage on map"); | 
|---|
| 146 | assert(!map()->has_exceptions(), "call add_exception_states_from 1st"); | 
|---|
| 147 | assert(!is_hidden_merge(control()), "call use_exception_state, not set_map"); | 
|---|
| 148 | } | 
|---|
| 149 |  | 
|---|
| 150 | void GraphKit::verify_exception_state(SafePointNode* ex_map) { | 
|---|
| 151 | assert(ex_map->next_exception() == NULL, "not already part of a chain"); | 
|---|
| 152 | assert(has_saved_ex_oop(ex_map), "every exception state has an ex_oop"); | 
|---|
| 153 | } | 
|---|
| 154 | #endif | 
|---|
| 155 |  | 
|---|
| 156 | //---------------------------stop_and_kill_map--------------------------------- | 
|---|
| 157 | // Set _map to NULL, signalling a stop to further bytecode execution. | 
|---|
| 158 | // First smash the current map's control to a constant, to mark it dead. | 
|---|
| 159 | void GraphKit::stop_and_kill_map() { | 
|---|
| 160 | SafePointNode* dead_map = stop(); | 
|---|
| 161 | if (dead_map != NULL) { | 
|---|
| 162 | dead_map->disconnect_inputs(NULL, C); // Mark the map as killed. | 
|---|
| 163 | assert(dead_map->is_killed(), "must be so marked"); | 
|---|
| 164 | } | 
|---|
| 165 | } | 
|---|
| 166 |  | 
|---|
| 167 |  | 
|---|
| 168 | //--------------------------------stopped-------------------------------------- | 
|---|
| 169 | // Tell if _map is NULL, or control is top. | 
|---|
| 170 | bool GraphKit::stopped() { | 
|---|
| 171 | if (map() == NULL)           return true; | 
|---|
| 172 | else if (control() == top()) return true; | 
|---|
| 173 | else                         return false; | 
|---|
| 174 | } | 
|---|
| 175 |  | 
|---|
| 176 |  | 
|---|
| 177 | //-----------------------------has_ex_handler---------------------------------- | 
|---|
| 178 | // Tell if this method or any caller method has exception handlers. | 
|---|
| 179 | bool GraphKit::has_ex_handler() { | 
|---|
| 180 | for (JVMState* jvmsp = jvms(); jvmsp != NULL; jvmsp = jvmsp->caller()) { | 
|---|
| 181 | if (jvmsp->has_method() && jvmsp->method()->has_exception_handlers()) { | 
|---|
| 182 | return true; | 
|---|
| 183 | } | 
|---|
| 184 | } | 
|---|
| 185 | return false; | 
|---|
| 186 | } | 
|---|
| 187 |  | 
|---|
| 188 | //------------------------------save_ex_oop------------------------------------ | 
|---|
| 189 | // Save an exception without blowing stack contents or other JVM state. | 
|---|
| 190 | void GraphKit::set_saved_ex_oop(SafePointNode* ex_map, Node* ex_oop) { | 
|---|
| 191 | assert(!has_saved_ex_oop(ex_map), "clear ex-oop before setting again"); | 
|---|
| 192 | ex_map->add_req(ex_oop); | 
|---|
| 193 | debug_only(verify_exception_state(ex_map)); | 
|---|
| 194 | } | 
|---|
| 195 |  | 
|---|
| 196 | inline static Node* common_saved_ex_oop(SafePointNode* ex_map, bool clear_it) { | 
|---|
| 197 | assert(GraphKit::has_saved_ex_oop(ex_map), "ex_oop must be there"); | 
|---|
| 198 | Node* ex_oop = ex_map->in(ex_map->req()-1); | 
|---|
| 199 | if (clear_it)  ex_map->del_req(ex_map->req()-1); | 
|---|
| 200 | return ex_oop; | 
|---|
| 201 | } | 
|---|
| 202 |  | 
|---|
| 203 | //-----------------------------saved_ex_oop------------------------------------ | 
|---|
| 204 | // Recover a saved exception from its map. | 
|---|
| 205 | Node* GraphKit::saved_ex_oop(SafePointNode* ex_map) { | 
|---|
| 206 | return common_saved_ex_oop(ex_map, false); | 
|---|
| 207 | } | 
|---|
| 208 |  | 
|---|
| 209 | //--------------------------clear_saved_ex_oop--------------------------------- | 
|---|
| 210 | // Erase a previously saved exception from its map. | 
|---|
| 211 | Node* GraphKit::clear_saved_ex_oop(SafePointNode* ex_map) { | 
|---|
| 212 | return common_saved_ex_oop(ex_map, true); | 
|---|
| 213 | } | 
|---|
| 214 |  | 
|---|
| 215 | #ifdef ASSERT | 
|---|
| 216 | //---------------------------has_saved_ex_oop---------------------------------- | 
|---|
| 217 | // Erase a previously saved exception from its map. | 
|---|
| 218 | bool GraphKit::has_saved_ex_oop(SafePointNode* ex_map) { | 
|---|
| 219 | return ex_map->req() == ex_map->jvms()->endoff()+1; | 
|---|
| 220 | } | 
|---|
| 221 | #endif | 
|---|
| 222 |  | 
|---|
| 223 | //-------------------------make_exception_state-------------------------------- | 
|---|
| 224 | // Turn the current JVM state into an exception state, appending the ex_oop. | 
|---|
| 225 | SafePointNode* GraphKit::make_exception_state(Node* ex_oop) { | 
|---|
| 226 | sync_jvms(); | 
|---|
| 227 | SafePointNode* ex_map = stop();  // do not manipulate this map any more | 
|---|
| 228 | set_saved_ex_oop(ex_map, ex_oop); | 
|---|
| 229 | return ex_map; | 
|---|
| 230 | } | 
|---|
| 231 |  | 
|---|
| 232 |  | 
|---|
| 233 | //--------------------------add_exception_state-------------------------------- | 
|---|
| 234 | // Add an exception to my list of exceptions. | 
|---|
| 235 | void GraphKit::add_exception_state(SafePointNode* ex_map) { | 
|---|
| 236 | if (ex_map == NULL || ex_map->control() == top()) { | 
|---|
| 237 | return; | 
|---|
| 238 | } | 
|---|
| 239 | #ifdef ASSERT | 
|---|
| 240 | verify_exception_state(ex_map); | 
|---|
| 241 | if (has_exceptions()) { | 
|---|
| 242 | assert(ex_map->jvms()->same_calls_as(_exceptions->jvms()), "all collected exceptions must come from the same place"); | 
|---|
| 243 | } | 
|---|
| 244 | #endif | 
|---|
| 245 |  | 
|---|
| 246 | // If there is already an exception of exactly this type, merge with it. | 
|---|
| 247 | // In particular, null-checks and other low-level exceptions common up here. | 
|---|
| 248 | Node*       ex_oop  = saved_ex_oop(ex_map); | 
|---|
| 249 | const Type* ex_type = _gvn.type(ex_oop); | 
|---|
| 250 | if (ex_oop == top()) { | 
|---|
| 251 | // No action needed. | 
|---|
| 252 | return; | 
|---|
| 253 | } | 
|---|
| 254 | assert(ex_type->isa_instptr(), "exception must be an instance"); | 
|---|
| 255 | for (SafePointNode* e2 = _exceptions; e2 != NULL; e2 = e2->next_exception()) { | 
|---|
| 256 | const Type* ex_type2 = _gvn.type(saved_ex_oop(e2)); | 
|---|
| 257 | // We check sp also because call bytecodes can generate exceptions | 
|---|
| 258 | // both before and after arguments are popped! | 
|---|
| 259 | if (ex_type2 == ex_type | 
|---|
| 260 | && e2->_jvms->sp() == ex_map->_jvms->sp()) { | 
|---|
| 261 | combine_exception_states(ex_map, e2); | 
|---|
| 262 | return; | 
|---|
| 263 | } | 
|---|
| 264 | } | 
|---|
| 265 |  | 
|---|
| 266 | // No pre-existing exception of the same type.  Chain it on the list. | 
|---|
| 267 | push_exception_state(ex_map); | 
|---|
| 268 | } | 
|---|
| 269 |  | 
|---|
| 270 | //-----------------------add_exception_states_from----------------------------- | 
|---|
| 271 | void GraphKit::add_exception_states_from(JVMState* jvms) { | 
|---|
| 272 | SafePointNode* ex_map = jvms->map()->next_exception(); | 
|---|
| 273 | if (ex_map != NULL) { | 
|---|
| 274 | jvms->map()->set_next_exception(NULL); | 
|---|
| 275 | for (SafePointNode* next_map; ex_map != NULL; ex_map = next_map) { | 
|---|
| 276 | next_map = ex_map->next_exception(); | 
|---|
| 277 | ex_map->set_next_exception(NULL); | 
|---|
| 278 | add_exception_state(ex_map); | 
|---|
| 279 | } | 
|---|
| 280 | } | 
|---|
| 281 | } | 
|---|
| 282 |  | 
|---|
| 283 | //-----------------------transfer_exceptions_into_jvms------------------------- | 
|---|
| 284 | JVMState* GraphKit::transfer_exceptions_into_jvms() { | 
|---|
| 285 | if (map() == NULL) { | 
|---|
| 286 | // We need a JVMS to carry the exceptions, but the map has gone away. | 
|---|
| 287 | // Create a scratch JVMS, cloned from any of the exception states... | 
|---|
| 288 | if (has_exceptions()) { | 
|---|
| 289 | _map = _exceptions; | 
|---|
| 290 | _map = clone_map(); | 
|---|
| 291 | _map->set_next_exception(NULL); | 
|---|
| 292 | clear_saved_ex_oop(_map); | 
|---|
| 293 | debug_only(verify_map()); | 
|---|
| 294 | } else { | 
|---|
| 295 | // ...or created from scratch | 
|---|
| 296 | JVMState* jvms = new (C) JVMState(_method, NULL); | 
|---|
| 297 | jvms->set_bci(_bci); | 
|---|
| 298 | jvms->set_sp(_sp); | 
|---|
| 299 | jvms->set_map(new SafePointNode(TypeFunc::Parms, jvms)); | 
|---|
| 300 | set_jvms(jvms); | 
|---|
| 301 | for (uint i = 0; i < map()->req(); i++)  map()->init_req(i, top()); | 
|---|
| 302 | set_all_memory(top()); | 
|---|
| 303 | while (map()->req() < jvms->endoff())  map()->add_req(top()); | 
|---|
| 304 | } | 
|---|
| 305 | // (This is a kludge, in case you didn't notice.) | 
|---|
| 306 | set_control(top()); | 
|---|
| 307 | } | 
|---|
| 308 | JVMState* jvms = sync_jvms(); | 
|---|
| 309 | assert(!jvms->map()->has_exceptions(), "no exceptions on this map yet"); | 
|---|
| 310 | jvms->map()->set_next_exception(_exceptions); | 
|---|
| 311 | _exceptions = NULL;   // done with this set of exceptions | 
|---|
| 312 | return jvms; | 
|---|
| 313 | } | 
|---|
| 314 |  | 
|---|
| 315 | static inline void add_n_reqs(Node* dstphi, Node* srcphi) { | 
|---|
| 316 | assert(is_hidden_merge(dstphi), "must be a special merge node"); | 
|---|
| 317 | assert(is_hidden_merge(srcphi), "must be a special merge node"); | 
|---|
| 318 | uint limit = srcphi->req(); | 
|---|
| 319 | for (uint i = PhiNode::Input; i < limit; i++) { | 
|---|
| 320 | dstphi->add_req(srcphi->in(i)); | 
|---|
| 321 | } | 
|---|
| 322 | } | 
|---|
| 323 | static inline void add_one_req(Node* dstphi, Node* src) { | 
|---|
| 324 | assert(is_hidden_merge(dstphi), "must be a special merge node"); | 
|---|
| 325 | assert(!is_hidden_merge(src), "must not be a special merge node"); | 
|---|
| 326 | dstphi->add_req(src); | 
|---|
| 327 | } | 
|---|
| 328 |  | 
|---|
| 329 | //-----------------------combine_exception_states------------------------------ | 
|---|
| 330 | // This helper function combines exception states by building phis on a | 
|---|
| 331 | // specially marked state-merging region.  These regions and phis are | 
|---|
| 332 | // untransformed, and can build up gradually.  The region is marked by | 
|---|
| 333 | // having a control input of its exception map, rather than NULL.  Such | 
|---|
| 334 | // regions do not appear except in this function, and in use_exception_state. | 
|---|
| 335 | void GraphKit::combine_exception_states(SafePointNode* ex_map, SafePointNode* phi_map) { | 
|---|
| 336 | if (failing())  return;  // dying anyway... | 
|---|
| 337 | JVMState* ex_jvms = ex_map->_jvms; | 
|---|
| 338 | assert(ex_jvms->same_calls_as(phi_map->_jvms), "consistent call chains"); | 
|---|
| 339 | assert(ex_jvms->stkoff() == phi_map->_jvms->stkoff(), "matching locals"); | 
|---|
| 340 | assert(ex_jvms->sp() == phi_map->_jvms->sp(), "matching stack sizes"); | 
|---|
| 341 | assert(ex_jvms->monoff() == phi_map->_jvms->monoff(), "matching JVMS"); | 
|---|
| 342 | assert(ex_jvms->scloff() == phi_map->_jvms->scloff(), "matching scalar replaced objects"); | 
|---|
| 343 | assert(ex_map->req() == phi_map->req(), "matching maps"); | 
|---|
| 344 | uint tos = ex_jvms->stkoff() + ex_jvms->sp(); | 
|---|
| 345 | Node*         hidden_merge_mark = root(); | 
|---|
| 346 | Node*         region  = phi_map->control(); | 
|---|
| 347 | MergeMemNode* phi_mem = phi_map->merged_memory(); | 
|---|
| 348 | MergeMemNode* ex_mem  = ex_map->merged_memory(); | 
|---|
| 349 | if (region->in(0) != hidden_merge_mark) { | 
|---|
| 350 | // The control input is not (yet) a specially-marked region in phi_map. | 
|---|
| 351 | // Make it so, and build some phis. | 
|---|
| 352 | region = new RegionNode(2); | 
|---|
| 353 | _gvn.set_type(region, Type::CONTROL); | 
|---|
| 354 | region->set_req(0, hidden_merge_mark);  // marks an internal ex-state | 
|---|
| 355 | region->init_req(1, phi_map->control()); | 
|---|
| 356 | phi_map->set_control(region); | 
|---|
| 357 | Node* io_phi = PhiNode::make(region, phi_map->i_o(), Type::ABIO); | 
|---|
| 358 | record_for_igvn(io_phi); | 
|---|
| 359 | _gvn.set_type(io_phi, Type::ABIO); | 
|---|
| 360 | phi_map->set_i_o(io_phi); | 
|---|
| 361 | for (MergeMemStream mms(phi_mem); mms.next_non_empty(); ) { | 
|---|
| 362 | Node* m = mms.memory(); | 
|---|
| 363 | Node* m_phi = PhiNode::make(region, m, Type::MEMORY, mms.adr_type(C)); | 
|---|
| 364 | record_for_igvn(m_phi); | 
|---|
| 365 | _gvn.set_type(m_phi, Type::MEMORY); | 
|---|
| 366 | mms.set_memory(m_phi); | 
|---|
| 367 | } | 
|---|
| 368 | } | 
|---|
| 369 |  | 
|---|
| 370 | // Either or both of phi_map and ex_map might already be converted into phis. | 
|---|
| 371 | Node* ex_control = ex_map->control(); | 
|---|
| 372 | // if there is special marking on ex_map also, we add multiple edges from src | 
|---|
| 373 | bool add_multiple = (ex_control->in(0) == hidden_merge_mark); | 
|---|
| 374 | // how wide was the destination phi_map, originally? | 
|---|
| 375 | uint orig_width = region->req(); | 
|---|
| 376 |  | 
|---|
| 377 | if (add_multiple) { | 
|---|
| 378 | add_n_reqs(region, ex_control); | 
|---|
| 379 | add_n_reqs(phi_map->i_o(), ex_map->i_o()); | 
|---|
| 380 | } else { | 
|---|
| 381 | // ex_map has no merges, so we just add single edges everywhere | 
|---|
| 382 | add_one_req(region, ex_control); | 
|---|
| 383 | add_one_req(phi_map->i_o(), ex_map->i_o()); | 
|---|
| 384 | } | 
|---|
| 385 | for (MergeMemStream mms(phi_mem, ex_mem); mms.next_non_empty2(); ) { | 
|---|
| 386 | if (mms.is_empty()) { | 
|---|
| 387 | // get a copy of the base memory, and patch some inputs into it | 
|---|
| 388 | const TypePtr* adr_type = mms.adr_type(C); | 
|---|
| 389 | Node* phi = mms.force_memory()->as_Phi()->slice_memory(adr_type); | 
|---|
| 390 | assert(phi->as_Phi()->region() == mms.base_memory()->in(0), ""); | 
|---|
| 391 | mms.set_memory(phi); | 
|---|
| 392 | // Prepare to append interesting stuff onto the newly sliced phi: | 
|---|
| 393 | while (phi->req() > orig_width)  phi->del_req(phi->req()-1); | 
|---|
| 394 | } | 
|---|
| 395 | // Append stuff from ex_map: | 
|---|
| 396 | if (add_multiple) { | 
|---|
| 397 | add_n_reqs(mms.memory(), mms.memory2()); | 
|---|
| 398 | } else { | 
|---|
| 399 | add_one_req(mms.memory(), mms.memory2()); | 
|---|
| 400 | } | 
|---|
| 401 | } | 
|---|
| 402 | uint limit = ex_map->req(); | 
|---|
| 403 | for (uint i = TypeFunc::Parms; i < limit; i++) { | 
|---|
| 404 | // Skip everything in the JVMS after tos.  (The ex_oop follows.) | 
|---|
| 405 | if (i == tos)  i = ex_jvms->monoff(); | 
|---|
| 406 | Node* src = ex_map->in(i); | 
|---|
| 407 | Node* dst = phi_map->in(i); | 
|---|
| 408 | if (src != dst) { | 
|---|
| 409 | PhiNode* phi; | 
|---|
| 410 | if (dst->in(0) != region) { | 
|---|
| 411 | dst = phi = PhiNode::make(region, dst, _gvn.type(dst)); | 
|---|
| 412 | record_for_igvn(phi); | 
|---|
| 413 | _gvn.set_type(phi, phi->type()); | 
|---|
| 414 | phi_map->set_req(i, dst); | 
|---|
| 415 | // Prepare to append interesting stuff onto the new phi: | 
|---|
| 416 | while (dst->req() > orig_width)  dst->del_req(dst->req()-1); | 
|---|
| 417 | } else { | 
|---|
| 418 | assert(dst->is_Phi(), "nobody else uses a hidden region"); | 
|---|
| 419 | phi = dst->as_Phi(); | 
|---|
| 420 | } | 
|---|
| 421 | if (add_multiple && src->in(0) == ex_control) { | 
|---|
| 422 | // Both are phis. | 
|---|
| 423 | add_n_reqs(dst, src); | 
|---|
| 424 | } else { | 
|---|
| 425 | while (dst->req() < region->req())  add_one_req(dst, src); | 
|---|
| 426 | } | 
|---|
| 427 | const Type* srctype = _gvn.type(src); | 
|---|
| 428 | if (phi->type() != srctype) { | 
|---|
| 429 | const Type* dsttype = phi->type()->meet_speculative(srctype); | 
|---|
| 430 | if (phi->type() != dsttype) { | 
|---|
| 431 | phi->set_type(dsttype); | 
|---|
| 432 | _gvn.set_type(phi, dsttype); | 
|---|
| 433 | } | 
|---|
| 434 | } | 
|---|
| 435 | } | 
|---|
| 436 | } | 
|---|
| 437 | phi_map->merge_replaced_nodes_with(ex_map); | 
|---|
| 438 | } | 
|---|
| 439 |  | 
|---|
| 440 | //--------------------------use_exception_state-------------------------------- | 
|---|
| 441 | Node* GraphKit::use_exception_state(SafePointNode* phi_map) { | 
|---|
| 442 | if (failing()) { stop(); return top(); } | 
|---|
| 443 | Node* region = phi_map->control(); | 
|---|
| 444 | Node* hidden_merge_mark = root(); | 
|---|
| 445 | assert(phi_map->jvms()->map() == phi_map, "sanity: 1-1 relation"); | 
|---|
| 446 | Node* ex_oop = clear_saved_ex_oop(phi_map); | 
|---|
| 447 | if (region->in(0) == hidden_merge_mark) { | 
|---|
| 448 | // Special marking for internal ex-states.  Process the phis now. | 
|---|
| 449 | region->set_req(0, region);  // now it's an ordinary region | 
|---|
| 450 | set_jvms(phi_map->jvms());   // ...so now we can use it as a map | 
|---|
| 451 | // Note: Setting the jvms also sets the bci and sp. | 
|---|
| 452 | set_control(_gvn.transform(region)); | 
|---|
| 453 | uint tos = jvms()->stkoff() + sp(); | 
|---|
| 454 | for (uint i = 1; i < tos; i++) { | 
|---|
| 455 | Node* x = phi_map->in(i); | 
|---|
| 456 | if (x->in(0) == region) { | 
|---|
| 457 | assert(x->is_Phi(), "expected a special phi"); | 
|---|
| 458 | phi_map->set_req(i, _gvn.transform(x)); | 
|---|
| 459 | } | 
|---|
| 460 | } | 
|---|
| 461 | for (MergeMemStream mms(merged_memory()); mms.next_non_empty(); ) { | 
|---|
| 462 | Node* x = mms.memory(); | 
|---|
| 463 | if (x->in(0) == region) { | 
|---|
| 464 | assert(x->is_Phi(), "nobody else uses a hidden region"); | 
|---|
| 465 | mms.set_memory(_gvn.transform(x)); | 
|---|
| 466 | } | 
|---|
| 467 | } | 
|---|
| 468 | if (ex_oop->in(0) == region) { | 
|---|
| 469 | assert(ex_oop->is_Phi(), "expected a special phi"); | 
|---|
| 470 | ex_oop = _gvn.transform(ex_oop); | 
|---|
| 471 | } | 
|---|
| 472 | } else { | 
|---|
| 473 | set_jvms(phi_map->jvms()); | 
|---|
| 474 | } | 
|---|
| 475 |  | 
|---|
| 476 | assert(!is_hidden_merge(phi_map->control()), "hidden ex. states cleared"); | 
|---|
| 477 | assert(!is_hidden_merge(phi_map->i_o()), "hidden ex. states cleared"); | 
|---|
| 478 | return ex_oop; | 
|---|
| 479 | } | 
|---|
| 480 |  | 
|---|
| 481 | //---------------------------------java_bc------------------------------------- | 
|---|
| 482 | Bytecodes::Code GraphKit::java_bc() const { | 
|---|
| 483 | ciMethod* method = this->method(); | 
|---|
| 484 | int       bci    = this->bci(); | 
|---|
| 485 | if (method != NULL && bci != InvocationEntryBci) | 
|---|
| 486 | return method->java_code_at_bci(bci); | 
|---|
| 487 | else | 
|---|
| 488 | return Bytecodes::_illegal; | 
|---|
| 489 | } | 
|---|
| 490 |  | 
|---|
| 491 | void GraphKit::uncommon_trap_if_should_post_on_exceptions(Deoptimization::DeoptReason reason, | 
|---|
| 492 | bool must_throw) { | 
|---|
| 493 | // if the exception capability is set, then we will generate code | 
|---|
| 494 | // to check the JavaThread.should_post_on_exceptions flag to see | 
|---|
| 495 | // if we actually need to report exception events (for this | 
|---|
| 496 | // thread).  If we don't need to report exception events, we will | 
|---|
| 497 | // take the normal fast path provided by add_exception_events.  If | 
|---|
| 498 | // exception event reporting is enabled for this thread, we will | 
|---|
| 499 | // take the uncommon_trap in the BuildCutout below. | 
|---|
| 500 |  | 
|---|
| 501 | // first must access the should_post_on_exceptions_flag in this thread's JavaThread | 
|---|
| 502 | Node* jthread = _gvn.transform(new ThreadLocalNode()); | 
|---|
| 503 | Node* adr = basic_plus_adr(top(), jthread, in_bytes(JavaThread::should_post_on_exceptions_flag_offset())); | 
|---|
| 504 | Node* should_post_flag = make_load(control(), adr, TypeInt::INT, T_INT, Compile::AliasIdxRaw, MemNode::unordered); | 
|---|
| 505 |  | 
|---|
| 506 | // Test the should_post_on_exceptions_flag vs. 0 | 
|---|
| 507 | Node* chk = _gvn.transform( new CmpINode(should_post_flag, intcon(0)) ); | 
|---|
| 508 | Node* tst = _gvn.transform( new BoolNode(chk, BoolTest::eq) ); | 
|---|
| 509 |  | 
|---|
| 510 | // Branch to slow_path if should_post_on_exceptions_flag was true | 
|---|
| 511 | { BuildCutout unless(this, tst, PROB_MAX); | 
|---|
| 512 | // Do not try anything fancy if we're notifying the VM on every throw. | 
|---|
| 513 | // Cf. case Bytecodes::_athrow in parse2.cpp. | 
|---|
| 514 | uncommon_trap(reason, Deoptimization::Action_none, | 
|---|
| 515 | (ciKlass*)NULL, (char*)NULL, must_throw); | 
|---|
| 516 | } | 
|---|
| 517 |  | 
|---|
| 518 | } | 
|---|
| 519 |  | 
|---|
| 520 | //------------------------------builtin_throw---------------------------------- | 
|---|
| 521 | void GraphKit::builtin_throw(Deoptimization::DeoptReason reason, Node* arg) { | 
|---|
| 522 | bool must_throw = true; | 
|---|
| 523 |  | 
|---|
| 524 | if (env()->jvmti_can_post_on_exceptions()) { | 
|---|
| 525 | // check if we must post exception events, take uncommon trap if so | 
|---|
| 526 | uncommon_trap_if_should_post_on_exceptions(reason, must_throw); | 
|---|
| 527 | // here if should_post_on_exceptions is false | 
|---|
| 528 | // continue on with the normal codegen | 
|---|
| 529 | } | 
|---|
| 530 |  | 
|---|
| 531 | // If this particular condition has not yet happened at this | 
|---|
| 532 | // bytecode, then use the uncommon trap mechanism, and allow for | 
|---|
| 533 | // a future recompilation if several traps occur here. | 
|---|
| 534 | // If the throw is hot, try to use a more complicated inline mechanism | 
|---|
| 535 | // which keeps execution inside the compiled code. | 
|---|
| 536 | bool treat_throw_as_hot = false; | 
|---|
| 537 | ciMethodData* md = method()->method_data(); | 
|---|
| 538 |  | 
|---|
| 539 | if (ProfileTraps) { | 
|---|
| 540 | if (too_many_traps(reason)) { | 
|---|
| 541 | treat_throw_as_hot = true; | 
|---|
| 542 | } | 
|---|
| 543 | // (If there is no MDO at all, assume it is early in | 
|---|
| 544 | // execution, and that any deopts are part of the | 
|---|
| 545 | // startup transient, and don't need to be remembered.) | 
|---|
| 546 |  | 
|---|
| 547 | // Also, if there is a local exception handler, treat all throws | 
|---|
| 548 | // as hot if there has been at least one in this method. | 
|---|
| 549 | if (C->trap_count(reason) != 0 | 
|---|
| 550 | && method()->method_data()->trap_count(reason) != 0 | 
|---|
| 551 | && has_ex_handler()) { | 
|---|
| 552 | treat_throw_as_hot = true; | 
|---|
| 553 | } | 
|---|
| 554 | } | 
|---|
| 555 |  | 
|---|
| 556 | // If this throw happens frequently, an uncommon trap might cause | 
|---|
| 557 | // a performance pothole.  If there is a local exception handler, | 
|---|
| 558 | // and if this particular bytecode appears to be deoptimizing often, | 
|---|
| 559 | // let us handle the throw inline, with a preconstructed instance. | 
|---|
| 560 | // Note:   If the deopt count has blown up, the uncommon trap | 
|---|
| 561 | // runtime is going to flush this nmethod, not matter what. | 
|---|
| 562 | if (treat_throw_as_hot | 
|---|
| 563 | && (!StackTraceInThrowable || OmitStackTraceInFastThrow)) { | 
|---|
| 564 | // If the throw is local, we use a pre-existing instance and | 
|---|
| 565 | // punt on the backtrace.  This would lead to a missing backtrace | 
|---|
| 566 | // (a repeat of 4292742) if the backtrace object is ever asked | 
|---|
| 567 | // for its backtrace. | 
|---|
| 568 | // Fixing this remaining case of 4292742 requires some flavor of | 
|---|
| 569 | // escape analysis.  Leave that for the future. | 
|---|
| 570 | ciInstance* ex_obj = NULL; | 
|---|
| 571 | switch (reason) { | 
|---|
| 572 | case Deoptimization::Reason_null_check: | 
|---|
| 573 | ex_obj = env()->NullPointerException_instance(); | 
|---|
| 574 | break; | 
|---|
| 575 | case Deoptimization::Reason_div0_check: | 
|---|
| 576 | ex_obj = env()->ArithmeticException_instance(); | 
|---|
| 577 | break; | 
|---|
| 578 | case Deoptimization::Reason_range_check: | 
|---|
| 579 | ex_obj = env()->ArrayIndexOutOfBoundsException_instance(); | 
|---|
| 580 | break; | 
|---|
| 581 | case Deoptimization::Reason_class_check: | 
|---|
| 582 | if (java_bc() == Bytecodes::_aastore) { | 
|---|
| 583 | ex_obj = env()->ArrayStoreException_instance(); | 
|---|
| 584 | } else { | 
|---|
| 585 | ex_obj = env()->ClassCastException_instance(); | 
|---|
| 586 | } | 
|---|
| 587 | break; | 
|---|
| 588 | default: | 
|---|
| 589 | break; | 
|---|
| 590 | } | 
|---|
| 591 | if (failing()) { stop(); return; }  // exception allocation might fail | 
|---|
| 592 | if (ex_obj != NULL) { | 
|---|
| 593 | // Cheat with a preallocated exception object. | 
|---|
| 594 | if (C->log() != NULL) | 
|---|
| 595 | C->log()->elem( "hot_throw preallocated='1' reason='%s'", | 
|---|
| 596 | Deoptimization::trap_reason_name(reason)); | 
|---|
| 597 | const TypeInstPtr* ex_con  = TypeInstPtr::make(ex_obj); | 
|---|
| 598 | Node*              ex_node = _gvn.transform(ConNode::make(ex_con)); | 
|---|
| 599 |  | 
|---|
| 600 | // Clear the detail message of the preallocated exception object. | 
|---|
| 601 | // Weblogic sometimes mutates the detail message of exceptions | 
|---|
| 602 | // using reflection. | 
|---|
| 603 | int offset = java_lang_Throwable::get_detailMessage_offset(); | 
|---|
| 604 | const TypePtr* adr_typ = ex_con->add_offset(offset); | 
|---|
| 605 |  | 
|---|
| 606 | Node *adr = basic_plus_adr(ex_node, ex_node, offset); | 
|---|
| 607 | const TypeOopPtr* val_type = TypeOopPtr::make_from_klass(env()->String_klass()); | 
|---|
| 608 | Node *store = access_store_at(ex_node, adr, adr_typ, null(), val_type, T_OBJECT, IN_HEAP); | 
|---|
| 609 |  | 
|---|
| 610 | add_exception_state(make_exception_state(ex_node)); | 
|---|
| 611 | return; | 
|---|
| 612 | } | 
|---|
| 613 | } | 
|---|
| 614 |  | 
|---|
| 615 | // %%% Maybe add entry to OptoRuntime which directly throws the exc.? | 
|---|
| 616 | // It won't be much cheaper than bailing to the interp., since we'll | 
|---|
| 617 | // have to pass up all the debug-info, and the runtime will have to | 
|---|
| 618 | // create the stack trace. | 
|---|
| 619 |  | 
|---|
| 620 | // Usual case:  Bail to interpreter. | 
|---|
| 621 | // Reserve the right to recompile if we haven't seen anything yet. | 
|---|
| 622 |  | 
|---|
| 623 | ciMethod* m = Deoptimization::reason_is_speculate(reason) ? C->method() : NULL; | 
|---|
| 624 | Deoptimization::DeoptAction action = Deoptimization::Action_maybe_recompile; | 
|---|
| 625 | if (treat_throw_as_hot | 
|---|
| 626 | && (method()->method_data()->trap_recompiled_at(bci(), m) | 
|---|
| 627 | || C->too_many_traps(reason))) { | 
|---|
| 628 | // We cannot afford to take more traps here.  Suffer in the interpreter. | 
|---|
| 629 | if (C->log() != NULL) | 
|---|
| 630 | C->log()->elem( "hot_throw preallocated='0' reason='%s' mcount='%d'", | 
|---|
| 631 | Deoptimization::trap_reason_name(reason), | 
|---|
| 632 | C->trap_count(reason)); | 
|---|
| 633 | action = Deoptimization::Action_none; | 
|---|
| 634 | } | 
|---|
| 635 |  | 
|---|
| 636 | // "must_throw" prunes the JVM state to include only the stack, if there | 
|---|
| 637 | // are no local exception handlers.  This should cut down on register | 
|---|
| 638 | // allocation time and code size, by drastically reducing the number | 
|---|
| 639 | // of in-edges on the call to the uncommon trap. | 
|---|
| 640 |  | 
|---|
| 641 | uncommon_trap(reason, action, (ciKlass*)NULL, (char*)NULL, must_throw); | 
|---|
| 642 | } | 
|---|
| 643 |  | 
|---|
| 644 |  | 
|---|
| 645 | //----------------------------PreserveJVMState--------------------------------- | 
|---|
| 646 | PreserveJVMState::PreserveJVMState(GraphKit* kit, bool clone_map) { | 
|---|
| 647 | debug_only(kit->verify_map()); | 
|---|
| 648 | _kit    = kit; | 
|---|
| 649 | _map    = kit->map();   // preserve the map | 
|---|
| 650 | _sp     = kit->sp(); | 
|---|
| 651 | kit->set_map(clone_map ? kit->clone_map() : NULL); | 
|---|
| 652 | #ifdef ASSERT | 
|---|
| 653 | _bci    = kit->bci(); | 
|---|
| 654 | Parse* parser = kit->is_Parse(); | 
|---|
| 655 | int block = (parser == NULL || parser->block() == NULL) ? -1 : parser->block()->rpo(); | 
|---|
| 656 | _block  = block; | 
|---|
| 657 | #endif | 
|---|
| 658 | } | 
|---|
| 659 | PreserveJVMState::~PreserveJVMState() { | 
|---|
| 660 | GraphKit* kit = _kit; | 
|---|
| 661 | #ifdef ASSERT | 
|---|
| 662 | assert(kit->bci() == _bci, "bci must not shift"); | 
|---|
| 663 | Parse* parser = kit->is_Parse(); | 
|---|
| 664 | int block = (parser == NULL || parser->block() == NULL) ? -1 : parser->block()->rpo(); | 
|---|
| 665 | assert(block == _block, "block must not shift"); | 
|---|
| 666 | #endif | 
|---|
| 667 | kit->set_map(_map); | 
|---|
| 668 | kit->set_sp(_sp); | 
|---|
| 669 | } | 
|---|
| 670 |  | 
|---|
| 671 |  | 
|---|
| 672 | //-----------------------------BuildCutout------------------------------------- | 
|---|
| 673 | BuildCutout::BuildCutout(GraphKit* kit, Node* p, float prob, float cnt) | 
|---|
| 674 | : PreserveJVMState(kit) | 
|---|
| 675 | { | 
|---|
| 676 | assert(p->is_Con() || p->is_Bool(), "test must be a bool"); | 
|---|
| 677 | SafePointNode* outer_map = _map;   // preserved map is caller's | 
|---|
| 678 | SafePointNode* inner_map = kit->map(); | 
|---|
| 679 | IfNode* iff = kit->create_and_map_if(outer_map->control(), p, prob, cnt); | 
|---|
| 680 | outer_map->set_control(kit->gvn().transform( new IfTrueNode(iff) )); | 
|---|
| 681 | inner_map->set_control(kit->gvn().transform( new IfFalseNode(iff) )); | 
|---|
| 682 | } | 
|---|
| 683 | BuildCutout::~BuildCutout() { | 
|---|
| 684 | GraphKit* kit = _kit; | 
|---|
| 685 | assert(kit->stopped(), "cutout code must stop, throw, return, etc."); | 
|---|
| 686 | } | 
|---|
| 687 |  | 
|---|
| 688 | //---------------------------PreserveReexecuteState---------------------------- | 
|---|
| 689 | PreserveReexecuteState::PreserveReexecuteState(GraphKit* kit) { | 
|---|
| 690 | assert(!kit->stopped(), "must call stopped() before"); | 
|---|
| 691 | _kit    =    kit; | 
|---|
| 692 | _sp     =    kit->sp(); | 
|---|
| 693 | _reexecute = kit->jvms()->_reexecute; | 
|---|
| 694 | } | 
|---|
| 695 | PreserveReexecuteState::~PreserveReexecuteState() { | 
|---|
| 696 | if (_kit->stopped()) return; | 
|---|
| 697 | _kit->jvms()->_reexecute = _reexecute; | 
|---|
| 698 | _kit->set_sp(_sp); | 
|---|
| 699 | } | 
|---|
| 700 |  | 
|---|
| 701 | //------------------------------clone_map-------------------------------------- | 
|---|
| 702 | // Implementation of PreserveJVMState | 
|---|
| 703 | // | 
|---|
| 704 | // Only clone_map(...) here. If this function is only used in the | 
|---|
| 705 | // PreserveJVMState class we may want to get rid of this extra | 
|---|
| 706 | // function eventually and do it all there. | 
|---|
| 707 |  | 
|---|
| 708 | SafePointNode* GraphKit::clone_map() { | 
|---|
| 709 | if (map() == NULL)  return NULL; | 
|---|
| 710 |  | 
|---|
| 711 | // Clone the memory edge first | 
|---|
| 712 | Node* mem = MergeMemNode::make(map()->memory()); | 
|---|
| 713 | gvn().set_type_bottom(mem); | 
|---|
| 714 |  | 
|---|
| 715 | SafePointNode *clonemap = (SafePointNode*)map()->clone(); | 
|---|
| 716 | JVMState* jvms = this->jvms(); | 
|---|
| 717 | JVMState* clonejvms = jvms->clone_shallow(C); | 
|---|
| 718 | clonemap->set_memory(mem); | 
|---|
| 719 | clonemap->set_jvms(clonejvms); | 
|---|
| 720 | clonejvms->set_map(clonemap); | 
|---|
| 721 | record_for_igvn(clonemap); | 
|---|
| 722 | gvn().set_type_bottom(clonemap); | 
|---|
| 723 | return clonemap; | 
|---|
| 724 | } | 
|---|
| 725 |  | 
|---|
| 726 |  | 
|---|
| 727 | //-----------------------------set_map_clone----------------------------------- | 
|---|
| 728 | void GraphKit::set_map_clone(SafePointNode* m) { | 
|---|
| 729 | _map = m; | 
|---|
| 730 | _map = clone_map(); | 
|---|
| 731 | _map->set_next_exception(NULL); | 
|---|
| 732 | debug_only(verify_map()); | 
|---|
| 733 | } | 
|---|
| 734 |  | 
|---|
| 735 |  | 
|---|
| 736 | //----------------------------kill_dead_locals--------------------------------- | 
|---|
| 737 | // Detect any locals which are known to be dead, and force them to top. | 
|---|
| 738 | void GraphKit::kill_dead_locals() { | 
|---|
| 739 | // Consult the liveness information for the locals.  If any | 
|---|
| 740 | // of them are unused, then they can be replaced by top().  This | 
|---|
| 741 | // should help register allocation time and cut down on the size | 
|---|
| 742 | // of the deoptimization information. | 
|---|
| 743 |  | 
|---|
| 744 | // This call is made from many of the bytecode handling | 
|---|
| 745 | // subroutines called from the Big Switch in do_one_bytecode. | 
|---|
| 746 | // Every bytecode which might include a slow path is responsible | 
|---|
| 747 | // for killing its dead locals.  The more consistent we | 
|---|
| 748 | // are about killing deads, the fewer useless phis will be | 
|---|
| 749 | // constructed for them at various merge points. | 
|---|
| 750 |  | 
|---|
| 751 | // bci can be -1 (InvocationEntryBci).  We return the entry | 
|---|
| 752 | // liveness for the method. | 
|---|
| 753 |  | 
|---|
| 754 | if (method() == NULL || method()->code_size() == 0) { | 
|---|
| 755 | // We are building a graph for a call to a native method. | 
|---|
| 756 | // All locals are live. | 
|---|
| 757 | return; | 
|---|
| 758 | } | 
|---|
| 759 |  | 
|---|
| 760 | ResourceMark rm; | 
|---|
| 761 |  | 
|---|
| 762 | // Consult the liveness information for the locals.  If any | 
|---|
| 763 | // of them are unused, then they can be replaced by top().  This | 
|---|
| 764 | // should help register allocation time and cut down on the size | 
|---|
| 765 | // of the deoptimization information. | 
|---|
| 766 | MethodLivenessResult live_locals = method()->liveness_at_bci(bci()); | 
|---|
| 767 |  | 
|---|
| 768 | int len = (int)live_locals.size(); | 
|---|
| 769 | assert(len <= jvms()->loc_size(), "too many live locals"); | 
|---|
| 770 | for (int local = 0; local < len; local++) { | 
|---|
| 771 | if (!live_locals.at(local)) { | 
|---|
| 772 | set_local(local, top()); | 
|---|
| 773 | } | 
|---|
| 774 | } | 
|---|
| 775 | } | 
|---|
| 776 |  | 
|---|
| 777 | #ifdef ASSERT | 
|---|
| 778 | //-------------------------dead_locals_are_killed------------------------------ | 
|---|
| 779 | // Return true if all dead locals are set to top in the map. | 
|---|
| 780 | // Used to assert "clean" debug info at various points. | 
|---|
| 781 | bool GraphKit::dead_locals_are_killed() { | 
|---|
| 782 | if (method() == NULL || method()->code_size() == 0) { | 
|---|
| 783 | // No locals need to be dead, so all is as it should be. | 
|---|
| 784 | return true; | 
|---|
| 785 | } | 
|---|
| 786 |  | 
|---|
| 787 | // Make sure somebody called kill_dead_locals upstream. | 
|---|
| 788 | ResourceMark rm; | 
|---|
| 789 | for (JVMState* jvms = this->jvms(); jvms != NULL; jvms = jvms->caller()) { | 
|---|
| 790 | if (jvms->loc_size() == 0)  continue;  // no locals to consult | 
|---|
| 791 | SafePointNode* map = jvms->map(); | 
|---|
| 792 | ciMethod* method = jvms->method(); | 
|---|
| 793 | int       bci    = jvms->bci(); | 
|---|
| 794 | if (jvms == this->jvms()) { | 
|---|
| 795 | bci = this->bci();  // it might not yet be synched | 
|---|
| 796 | } | 
|---|
| 797 | MethodLivenessResult live_locals = method->liveness_at_bci(bci); | 
|---|
| 798 | int len = (int)live_locals.size(); | 
|---|
| 799 | if (!live_locals.is_valid() || len == 0) | 
|---|
| 800 | // This method is trivial, or is poisoned by a breakpoint. | 
|---|
| 801 | return true; | 
|---|
| 802 | assert(len == jvms->loc_size(), "live map consistent with locals map"); | 
|---|
| 803 | for (int local = 0; local < len; local++) { | 
|---|
| 804 | if (!live_locals.at(local) && map->local(jvms, local) != top()) { | 
|---|
| 805 | if (PrintMiscellaneous && (Verbose || WizardMode)) { | 
|---|
| 806 | tty->print_cr( "Zombie local %d: ", local); | 
|---|
| 807 | jvms->dump(); | 
|---|
| 808 | } | 
|---|
| 809 | return false; | 
|---|
| 810 | } | 
|---|
| 811 | } | 
|---|
| 812 | } | 
|---|
| 813 | return true; | 
|---|
| 814 | } | 
|---|
| 815 |  | 
|---|
| 816 | #endif //ASSERT | 
|---|
| 817 |  | 
|---|
| 818 | // Helper function for enforcing certain bytecodes to reexecute if | 
|---|
| 819 | // deoptimization happens | 
|---|
| 820 | static bool should_reexecute_implied_by_bytecode(JVMState *jvms, bool is_anewarray) { | 
|---|
| 821 | ciMethod* cur_method = jvms->method(); | 
|---|
| 822 | int       cur_bci   = jvms->bci(); | 
|---|
| 823 | if (cur_method != NULL && cur_bci != InvocationEntryBci) { | 
|---|
| 824 | Bytecodes::Code code = cur_method->java_code_at_bci(cur_bci); | 
|---|
| 825 | return Interpreter::bytecode_should_reexecute(code) || | 
|---|
| 826 | (is_anewarray && code == Bytecodes::_multianewarray); | 
|---|
| 827 | // Reexecute _multianewarray bytecode which was replaced with | 
|---|
| 828 | // sequence of [a]newarray. See Parse::do_multianewarray(). | 
|---|
| 829 | // | 
|---|
| 830 | // Note: interpreter should not have it set since this optimization | 
|---|
| 831 | // is limited by dimensions and guarded by flag so in some cases | 
|---|
| 832 | // multianewarray() runtime calls will be generated and | 
|---|
| 833 | // the bytecode should not be reexecutes (stack will not be reset). | 
|---|
| 834 | } else | 
|---|
| 835 | return false; | 
|---|
| 836 | } | 
|---|
| 837 |  | 
|---|
| 838 | // Helper function for adding JVMState and debug information to node | 
|---|
| 839 | void GraphKit::add_safepoint_edges(SafePointNode* call, bool must_throw) { | 
|---|
| 840 | // Add the safepoint edges to the call (or other safepoint). | 
|---|
| 841 |  | 
|---|
| 842 | // Make sure dead locals are set to top.  This | 
|---|
| 843 | // should help register allocation time and cut down on the size | 
|---|
| 844 | // of the deoptimization information. | 
|---|
| 845 | assert(dead_locals_are_killed(), "garbage in debug info before safepoint"); | 
|---|
| 846 |  | 
|---|
| 847 | // Walk the inline list to fill in the correct set of JVMState's | 
|---|
| 848 | // Also fill in the associated edges for each JVMState. | 
|---|
| 849 |  | 
|---|
| 850 | // If the bytecode needs to be reexecuted we need to put | 
|---|
| 851 | // the arguments back on the stack. | 
|---|
| 852 | const bool should_reexecute = jvms()->should_reexecute(); | 
|---|
| 853 | JVMState* youngest_jvms = should_reexecute ? sync_jvms_for_reexecute() : sync_jvms(); | 
|---|
| 854 |  | 
|---|
| 855 | // NOTE: set_bci (called from sync_jvms) might reset the reexecute bit to | 
|---|
| 856 | // undefined if the bci is different.  This is normal for Parse but it | 
|---|
| 857 | // should not happen for LibraryCallKit because only one bci is processed. | 
|---|
| 858 | assert(!is_LibraryCallKit() || (jvms()->should_reexecute() == should_reexecute), | 
|---|
| 859 | "in LibraryCallKit the reexecute bit should not change"); | 
|---|
| 860 |  | 
|---|
| 861 | // If we are guaranteed to throw, we can prune everything but the | 
|---|
| 862 | // input to the current bytecode. | 
|---|
| 863 | bool can_prune_locals = false; | 
|---|
| 864 | uint stack_slots_not_pruned = 0; | 
|---|
| 865 | int inputs = 0, depth = 0; | 
|---|
| 866 | if (must_throw) { | 
|---|
| 867 | assert(method() == youngest_jvms->method(), "sanity"); | 
|---|
| 868 | if (compute_stack_effects(inputs, depth)) { | 
|---|
| 869 | can_prune_locals = true; | 
|---|
| 870 | stack_slots_not_pruned = inputs; | 
|---|
| 871 | } | 
|---|
| 872 | } | 
|---|
| 873 |  | 
|---|
| 874 | if (env()->should_retain_local_variables()) { | 
|---|
| 875 | // At any safepoint, this method can get breakpointed, which would | 
|---|
| 876 | // then require an immediate deoptimization. | 
|---|
| 877 | can_prune_locals = false;  // do not prune locals | 
|---|
| 878 | stack_slots_not_pruned = 0; | 
|---|
| 879 | } | 
|---|
| 880 |  | 
|---|
| 881 | // do not scribble on the input jvms | 
|---|
| 882 | JVMState* out_jvms = youngest_jvms->clone_deep(C); | 
|---|
| 883 | call->set_jvms(out_jvms); // Start jvms list for call node | 
|---|
| 884 |  | 
|---|
| 885 | // For a known set of bytecodes, the interpreter should reexecute them if | 
|---|
| 886 | // deoptimization happens. We set the reexecute state for them here | 
|---|
| 887 | if (out_jvms->is_reexecute_undefined() && //don't change if already specified | 
|---|
| 888 | should_reexecute_implied_by_bytecode(out_jvms, call->is_AllocateArray())) { | 
|---|
| 889 | out_jvms->set_should_reexecute(true); //NOTE: youngest_jvms not changed | 
|---|
| 890 | } | 
|---|
| 891 |  | 
|---|
| 892 | // Presize the call: | 
|---|
| 893 | DEBUG_ONLY(uint non_debug_edges = call->req()); | 
|---|
| 894 | call->add_req_batch(top(), youngest_jvms->debug_depth()); | 
|---|
| 895 | assert(call->req() == non_debug_edges + youngest_jvms->debug_depth(), ""); | 
|---|
| 896 |  | 
|---|
| 897 | // Set up edges so that the call looks like this: | 
|---|
| 898 | //  Call [state:] ctl io mem fptr retadr | 
|---|
| 899 | //       [parms:] parm0 ... parmN | 
|---|
| 900 | //       [root:]  loc0 ... locN stk0 ... stkSP mon0 obj0 ... monN objN | 
|---|
| 901 | //    [...mid:]   loc0 ... locN stk0 ... stkSP mon0 obj0 ... monN objN [...] | 
|---|
| 902 | //       [young:] loc0 ... locN stk0 ... stkSP mon0 obj0 ... monN objN | 
|---|
| 903 | // Note that caller debug info precedes callee debug info. | 
|---|
| 904 |  | 
|---|
| 905 | // Fill pointer walks backwards from "young:" to "root:" in the diagram above: | 
|---|
| 906 | uint debug_ptr = call->req(); | 
|---|
| 907 |  | 
|---|
| 908 | // Loop over the map input edges associated with jvms, add them | 
|---|
| 909 | // to the call node, & reset all offsets to match call node array. | 
|---|
| 910 | for (JVMState* in_jvms = youngest_jvms; in_jvms != NULL; ) { | 
|---|
| 911 | uint debug_end   = debug_ptr; | 
|---|
| 912 | uint debug_start = debug_ptr - in_jvms->debug_size(); | 
|---|
| 913 | debug_ptr = debug_start;  // back up the ptr | 
|---|
| 914 |  | 
|---|
| 915 | uint p = debug_start;  // walks forward in [debug_start, debug_end) | 
|---|
| 916 | uint j, k, l; | 
|---|
| 917 | SafePointNode* in_map = in_jvms->map(); | 
|---|
| 918 | out_jvms->set_map(call); | 
|---|
| 919 |  | 
|---|
| 920 | if (can_prune_locals) { | 
|---|
| 921 | assert(in_jvms->method() == out_jvms->method(), "sanity"); | 
|---|
| 922 | // If the current throw can reach an exception handler in this JVMS, | 
|---|
| 923 | // then we must keep everything live that can reach that handler. | 
|---|
| 924 | // As a quick and dirty approximation, we look for any handlers at all. | 
|---|
| 925 | if (in_jvms->method()->has_exception_handlers()) { | 
|---|
| 926 | can_prune_locals = false; | 
|---|
| 927 | } | 
|---|
| 928 | } | 
|---|
| 929 |  | 
|---|
| 930 | // Add the Locals | 
|---|
| 931 | k = in_jvms->locoff(); | 
|---|
| 932 | l = in_jvms->loc_size(); | 
|---|
| 933 | out_jvms->set_locoff(p); | 
|---|
| 934 | if (!can_prune_locals) { | 
|---|
| 935 | for (j = 0; j < l; j++) | 
|---|
| 936 | call->set_req(p++, in_map->in(k+j)); | 
|---|
| 937 | } else { | 
|---|
| 938 | p += l;  // already set to top above by add_req_batch | 
|---|
| 939 | } | 
|---|
| 940 |  | 
|---|
| 941 | // Add the Expression Stack | 
|---|
| 942 | k = in_jvms->stkoff(); | 
|---|
| 943 | l = in_jvms->sp(); | 
|---|
| 944 | out_jvms->set_stkoff(p); | 
|---|
| 945 | if (!can_prune_locals) { | 
|---|
| 946 | for (j = 0; j < l; j++) | 
|---|
| 947 | call->set_req(p++, in_map->in(k+j)); | 
|---|
| 948 | } else if (can_prune_locals && stack_slots_not_pruned != 0) { | 
|---|
| 949 | // Divide stack into {S0,...,S1}, where S0 is set to top. | 
|---|
| 950 | uint s1 = stack_slots_not_pruned; | 
|---|
| 951 | stack_slots_not_pruned = 0;  // for next iteration | 
|---|
| 952 | if (s1 > l)  s1 = l; | 
|---|
| 953 | uint s0 = l - s1; | 
|---|
| 954 | p += s0;  // skip the tops preinstalled by add_req_batch | 
|---|
| 955 | for (j = s0; j < l; j++) | 
|---|
| 956 | call->set_req(p++, in_map->in(k+j)); | 
|---|
| 957 | } else { | 
|---|
| 958 | p += l;  // already set to top above by add_req_batch | 
|---|
| 959 | } | 
|---|
| 960 |  | 
|---|
| 961 | // Add the Monitors | 
|---|
| 962 | k = in_jvms->monoff(); | 
|---|
| 963 | l = in_jvms->mon_size(); | 
|---|
| 964 | out_jvms->set_monoff(p); | 
|---|
| 965 | for (j = 0; j < l; j++) | 
|---|
| 966 | call->set_req(p++, in_map->in(k+j)); | 
|---|
| 967 |  | 
|---|
| 968 | // Copy any scalar object fields. | 
|---|
| 969 | k = in_jvms->scloff(); | 
|---|
| 970 | l = in_jvms->scl_size(); | 
|---|
| 971 | out_jvms->set_scloff(p); | 
|---|
| 972 | for (j = 0; j < l; j++) | 
|---|
| 973 | call->set_req(p++, in_map->in(k+j)); | 
|---|
| 974 |  | 
|---|
| 975 | // Finish the new jvms. | 
|---|
| 976 | out_jvms->set_endoff(p); | 
|---|
| 977 |  | 
|---|
| 978 | assert(out_jvms->endoff()     == debug_end, "fill ptr must match"); | 
|---|
| 979 | assert(out_jvms->depth()      == in_jvms->depth(), "depth must match"); | 
|---|
| 980 | assert(out_jvms->loc_size()   == in_jvms->loc_size(), "size must match"); | 
|---|
| 981 | assert(out_jvms->mon_size()   == in_jvms->mon_size(), "size must match"); | 
|---|
| 982 | assert(out_jvms->scl_size()   == in_jvms->scl_size(), "size must match"); | 
|---|
| 983 | assert(out_jvms->debug_size() == in_jvms->debug_size(), "size must match"); | 
|---|
| 984 |  | 
|---|
| 985 | // Update the two tail pointers in parallel. | 
|---|
| 986 | out_jvms = out_jvms->caller(); | 
|---|
| 987 | in_jvms  = in_jvms->caller(); | 
|---|
| 988 | } | 
|---|
| 989 |  | 
|---|
| 990 | assert(debug_ptr == non_debug_edges, "debug info must fit exactly"); | 
|---|
| 991 |  | 
|---|
| 992 | // Test the correctness of JVMState::debug_xxx accessors: | 
|---|
| 993 | assert(call->jvms()->debug_start() == non_debug_edges, ""); | 
|---|
| 994 | assert(call->jvms()->debug_end()   == call->req(), ""); | 
|---|
| 995 | assert(call->jvms()->debug_depth() == call->req() - non_debug_edges, ""); | 
|---|
| 996 | } | 
|---|
| 997 |  | 
|---|
| 998 | bool GraphKit::compute_stack_effects(int& inputs, int& depth) { | 
|---|
| 999 | Bytecodes::Code code = java_bc(); | 
|---|
| 1000 | if (code == Bytecodes::_wide) { | 
|---|
| 1001 | code = method()->java_code_at_bci(bci() + 1); | 
|---|
| 1002 | } | 
|---|
| 1003 |  | 
|---|
| 1004 | BasicType rtype = T_ILLEGAL; | 
|---|
| 1005 | int       rsize = 0; | 
|---|
| 1006 |  | 
|---|
| 1007 | if (code != Bytecodes::_illegal) { | 
|---|
| 1008 | depth = Bytecodes::depth(code); // checkcast=0, athrow=-1 | 
|---|
| 1009 | rtype = Bytecodes::result_type(code); // checkcast=P, athrow=V | 
|---|
| 1010 | if (rtype < T_CONFLICT) | 
|---|
| 1011 | rsize = type2size[rtype]; | 
|---|
| 1012 | } | 
|---|
| 1013 |  | 
|---|
| 1014 | switch (code) { | 
|---|
| 1015 | case Bytecodes::_illegal: | 
|---|
| 1016 | return false; | 
|---|
| 1017 |  | 
|---|
| 1018 | case Bytecodes::_ldc: | 
|---|
| 1019 | case Bytecodes::_ldc_w: | 
|---|
| 1020 | case Bytecodes::_ldc2_w: | 
|---|
| 1021 | inputs = 0; | 
|---|
| 1022 | break; | 
|---|
| 1023 |  | 
|---|
| 1024 | case Bytecodes::_dup:         inputs = 1;  break; | 
|---|
| 1025 | case Bytecodes::_dup_x1:      inputs = 2;  break; | 
|---|
| 1026 | case Bytecodes::_dup_x2:      inputs = 3;  break; | 
|---|
| 1027 | case Bytecodes::_dup2:        inputs = 2;  break; | 
|---|
| 1028 | case Bytecodes::_dup2_x1:     inputs = 3;  break; | 
|---|
| 1029 | case Bytecodes::_dup2_x2:     inputs = 4;  break; | 
|---|
| 1030 | case Bytecodes::_swap:        inputs = 2;  break; | 
|---|
| 1031 | case Bytecodes::_arraylength: inputs = 1;  break; | 
|---|
| 1032 |  | 
|---|
| 1033 | case Bytecodes::_getstatic: | 
|---|
| 1034 | case Bytecodes::_putstatic: | 
|---|
| 1035 | case Bytecodes::_getfield: | 
|---|
| 1036 | case Bytecodes::_putfield: | 
|---|
| 1037 | { | 
|---|
| 1038 | bool ignored_will_link; | 
|---|
| 1039 | ciField* field = method()->get_field_at_bci(bci(), ignored_will_link); | 
|---|
| 1040 | int      size  = field->type()->size(); | 
|---|
| 1041 | bool is_get = (depth >= 0), is_static = (depth & 1); | 
|---|
| 1042 | inputs = (is_static ? 0 : 1); | 
|---|
| 1043 | if (is_get) { | 
|---|
| 1044 | depth = size - inputs; | 
|---|
| 1045 | } else { | 
|---|
| 1046 | inputs += size;        // putxxx pops the value from the stack | 
|---|
| 1047 | depth = - inputs; | 
|---|
| 1048 | } | 
|---|
| 1049 | } | 
|---|
| 1050 | break; | 
|---|
| 1051 |  | 
|---|
| 1052 | case Bytecodes::_invokevirtual: | 
|---|
| 1053 | case Bytecodes::_invokespecial: | 
|---|
| 1054 | case Bytecodes::_invokestatic: | 
|---|
| 1055 | case Bytecodes::_invokedynamic: | 
|---|
| 1056 | case Bytecodes::_invokeinterface: | 
|---|
| 1057 | { | 
|---|
| 1058 | bool ignored_will_link; | 
|---|
| 1059 | ciSignature* declared_signature = NULL; | 
|---|
| 1060 | ciMethod* ignored_callee = method()->get_method_at_bci(bci(), ignored_will_link, &declared_signature); | 
|---|
| 1061 | assert(declared_signature != NULL, "cannot be null"); | 
|---|
| 1062 | inputs   = declared_signature->arg_size_for_bc(code); | 
|---|
| 1063 | int size = declared_signature->return_type()->size(); | 
|---|
| 1064 | depth = size - inputs; | 
|---|
| 1065 | } | 
|---|
| 1066 | break; | 
|---|
| 1067 |  | 
|---|
| 1068 | case Bytecodes::_multianewarray: | 
|---|
| 1069 | { | 
|---|
| 1070 | ciBytecodeStream iter(method()); | 
|---|
| 1071 | iter.reset_to_bci(bci()); | 
|---|
| 1072 | iter.next(); | 
|---|
| 1073 | inputs = iter.get_dimensions(); | 
|---|
| 1074 | assert(rsize == 1, ""); | 
|---|
| 1075 | depth = rsize - inputs; | 
|---|
| 1076 | } | 
|---|
| 1077 | break; | 
|---|
| 1078 |  | 
|---|
| 1079 | case Bytecodes::_ireturn: | 
|---|
| 1080 | case Bytecodes::_lreturn: | 
|---|
| 1081 | case Bytecodes::_freturn: | 
|---|
| 1082 | case Bytecodes::_dreturn: | 
|---|
| 1083 | case Bytecodes::_areturn: | 
|---|
| 1084 | assert(rsize == -depth, ""); | 
|---|
| 1085 | inputs = rsize; | 
|---|
| 1086 | break; | 
|---|
| 1087 |  | 
|---|
| 1088 | case Bytecodes::_jsr: | 
|---|
| 1089 | case Bytecodes::_jsr_w: | 
|---|
| 1090 | inputs = 0; | 
|---|
| 1091 | depth  = 1;                  // S.B. depth=1, not zero | 
|---|
| 1092 | break; | 
|---|
| 1093 |  | 
|---|
| 1094 | default: | 
|---|
| 1095 | // bytecode produces a typed result | 
|---|
| 1096 | inputs = rsize - depth; | 
|---|
| 1097 | assert(inputs >= 0, ""); | 
|---|
| 1098 | break; | 
|---|
| 1099 | } | 
|---|
| 1100 |  | 
|---|
| 1101 | #ifdef ASSERT | 
|---|
| 1102 | // spot check | 
|---|
| 1103 | int outputs = depth + inputs; | 
|---|
| 1104 | assert(outputs >= 0, "sanity"); | 
|---|
| 1105 | switch (code) { | 
|---|
| 1106 | case Bytecodes::_checkcast: assert(inputs == 1 && outputs == 1, ""); break; | 
|---|
| 1107 | case Bytecodes::_athrow:    assert(inputs == 1 && outputs == 0, ""); break; | 
|---|
| 1108 | case Bytecodes::_aload_0:   assert(inputs == 0 && outputs == 1, ""); break; | 
|---|
| 1109 | case Bytecodes::_return:    assert(inputs == 0 && outputs == 0, ""); break; | 
|---|
| 1110 | case Bytecodes::_drem:      assert(inputs == 4 && outputs == 2, ""); break; | 
|---|
| 1111 | default:                    break; | 
|---|
| 1112 | } | 
|---|
| 1113 | #endif //ASSERT | 
|---|
| 1114 |  | 
|---|
| 1115 | return true; | 
|---|
| 1116 | } | 
|---|
| 1117 |  | 
|---|
| 1118 |  | 
|---|
| 1119 |  | 
|---|
| 1120 | //------------------------------basic_plus_adr--------------------------------- | 
|---|
| 1121 | Node* GraphKit::basic_plus_adr(Node* base, Node* ptr, Node* offset) { | 
|---|
| 1122 | // short-circuit a common case | 
|---|
| 1123 | if (offset == intcon(0))  return ptr; | 
|---|
| 1124 | return _gvn.transform( new AddPNode(base, ptr, offset) ); | 
|---|
| 1125 | } | 
|---|
| 1126 |  | 
|---|
| 1127 | Node* GraphKit::ConvI2L(Node* offset) { | 
|---|
| 1128 | // short-circuit a common case | 
|---|
| 1129 | jint offset_con = find_int_con(offset, Type::OffsetBot); | 
|---|
| 1130 | if (offset_con != Type::OffsetBot) { | 
|---|
| 1131 | return longcon((jlong) offset_con); | 
|---|
| 1132 | } | 
|---|
| 1133 | return _gvn.transform( new ConvI2LNode(offset)); | 
|---|
| 1134 | } | 
|---|
| 1135 |  | 
|---|
| 1136 | Node* GraphKit::ConvI2UL(Node* offset) { | 
|---|
| 1137 | juint offset_con = (juint) find_int_con(offset, Type::OffsetBot); | 
|---|
| 1138 | if (offset_con != (juint) Type::OffsetBot) { | 
|---|
| 1139 | return longcon((julong) offset_con); | 
|---|
| 1140 | } | 
|---|
| 1141 | Node* conv = _gvn.transform( new ConvI2LNode(offset)); | 
|---|
| 1142 | Node* mask = _gvn.transform(ConLNode::make((julong) max_juint)); | 
|---|
| 1143 | return _gvn.transform( new AndLNode(conv, mask) ); | 
|---|
| 1144 | } | 
|---|
| 1145 |  | 
|---|
| 1146 | Node* GraphKit::ConvL2I(Node* offset) { | 
|---|
| 1147 | // short-circuit a common case | 
|---|
| 1148 | jlong offset_con = find_long_con(offset, (jlong)Type::OffsetBot); | 
|---|
| 1149 | if (offset_con != (jlong)Type::OffsetBot) { | 
|---|
| 1150 | return intcon((int) offset_con); | 
|---|
| 1151 | } | 
|---|
| 1152 | return _gvn.transform( new ConvL2INode(offset)); | 
|---|
| 1153 | } | 
|---|
| 1154 |  | 
|---|
| 1155 | //-------------------------load_object_klass----------------------------------- | 
|---|
| 1156 | Node* GraphKit::load_object_klass(Node* obj) { | 
|---|
| 1157 | // Special-case a fresh allocation to avoid building nodes: | 
|---|
| 1158 | Node* akls = AllocateNode::Ideal_klass(obj, &_gvn); | 
|---|
| 1159 | if (akls != NULL)  return akls; | 
|---|
| 1160 | Node* k_adr = basic_plus_adr(obj, oopDesc::klass_offset_in_bytes()); | 
|---|
| 1161 | return _gvn.transform(LoadKlassNode::make(_gvn, NULL, immutable_memory(), k_adr, TypeInstPtr::KLASS)); | 
|---|
| 1162 | } | 
|---|
| 1163 |  | 
|---|
| 1164 | //-------------------------load_array_length----------------------------------- | 
|---|
| 1165 | Node* GraphKit::load_array_length(Node* array) { | 
|---|
| 1166 | // Special-case a fresh allocation to avoid building nodes: | 
|---|
| 1167 | AllocateArrayNode* alloc = AllocateArrayNode::Ideal_array_allocation(array, &_gvn); | 
|---|
| 1168 | Node *alen; | 
|---|
| 1169 | if (alloc == NULL) { | 
|---|
| 1170 | Node *r_adr = basic_plus_adr(array, arrayOopDesc::length_offset_in_bytes()); | 
|---|
| 1171 | alen = _gvn.transform( new LoadRangeNode(0, immutable_memory(), r_adr, TypeInt::POS)); | 
|---|
| 1172 | } else { | 
|---|
| 1173 | alen = alloc->Ideal_length(); | 
|---|
| 1174 | Node* ccast = alloc->make_ideal_length(_gvn.type(array)->is_oopptr(), &_gvn); | 
|---|
| 1175 | if (ccast != alen) { | 
|---|
| 1176 | alen = _gvn.transform(ccast); | 
|---|
| 1177 | } | 
|---|
| 1178 | } | 
|---|
| 1179 | return alen; | 
|---|
| 1180 | } | 
|---|
| 1181 |  | 
|---|
| 1182 | //------------------------------do_null_check---------------------------------- | 
|---|
| 1183 | // Helper function to do a NULL pointer check.  Returned value is | 
|---|
| 1184 | // the incoming address with NULL casted away.  You are allowed to use the | 
|---|
| 1185 | // not-null value only if you are control dependent on the test. | 
|---|
| 1186 | #ifndef PRODUCT | 
|---|
| 1187 | extern int explicit_null_checks_inserted, | 
|---|
| 1188 | explicit_null_checks_elided; | 
|---|
| 1189 | #endif | 
|---|
| 1190 | Node* GraphKit::null_check_common(Node* value, BasicType type, | 
|---|
| 1191 | // optional arguments for variations: | 
|---|
| 1192 | bool assert_null, | 
|---|
| 1193 | Node* *null_control, | 
|---|
| 1194 | bool speculative) { | 
|---|
| 1195 | assert(!assert_null || null_control == NULL, "not both at once"); | 
|---|
| 1196 | if (stopped())  return top(); | 
|---|
| 1197 | NOT_PRODUCT(explicit_null_checks_inserted++); | 
|---|
| 1198 |  | 
|---|
| 1199 | // Construct NULL check | 
|---|
| 1200 | Node *chk = NULL; | 
|---|
| 1201 | switch(type) { | 
|---|
| 1202 | case T_LONG   : chk = new CmpLNode(value, _gvn.zerocon(T_LONG)); break; | 
|---|
| 1203 | case T_INT    : chk = new CmpINode(value, _gvn.intcon(0)); break; | 
|---|
| 1204 | case T_ARRAY  : // fall through | 
|---|
| 1205 | type = T_OBJECT;  // simplify further tests | 
|---|
| 1206 | case T_OBJECT : { | 
|---|
| 1207 | const Type *t = _gvn.type( value ); | 
|---|
| 1208 |  | 
|---|
| 1209 | const TypeOopPtr* tp = t->isa_oopptr(); | 
|---|
| 1210 | if (tp != NULL && tp->klass() != NULL && !tp->klass()->is_loaded() | 
|---|
| 1211 | // Only for do_null_check, not any of its siblings: | 
|---|
| 1212 | && !assert_null && null_control == NULL) { | 
|---|
| 1213 | // Usually, any field access or invocation on an unloaded oop type | 
|---|
| 1214 | // will simply fail to link, since the statically linked class is | 
|---|
| 1215 | // likely also to be unloaded.  However, in -Xcomp mode, sometimes | 
|---|
| 1216 | // the static class is loaded but the sharper oop type is not. | 
|---|
| 1217 | // Rather than checking for this obscure case in lots of places, | 
|---|
| 1218 | // we simply observe that a null check on an unloaded class | 
|---|
| 1219 | // will always be followed by a nonsense operation, so we | 
|---|
| 1220 | // can just issue the uncommon trap here. | 
|---|
| 1221 | // Our access to the unloaded class will only be correct | 
|---|
| 1222 | // after it has been loaded and initialized, which requires | 
|---|
| 1223 | // a trip through the interpreter. | 
|---|
| 1224 | #ifndef PRODUCT | 
|---|
| 1225 | if (WizardMode) { tty->print( "Null check of unloaded "); tp->klass()->print(); tty->cr(); } | 
|---|
| 1226 | #endif | 
|---|
| 1227 | uncommon_trap(Deoptimization::Reason_unloaded, | 
|---|
| 1228 | Deoptimization::Action_reinterpret, | 
|---|
| 1229 | tp->klass(), "!loaded"); | 
|---|
| 1230 | return top(); | 
|---|
| 1231 | } | 
|---|
| 1232 |  | 
|---|
| 1233 | if (assert_null) { | 
|---|
| 1234 | // See if the type is contained in NULL_PTR. | 
|---|
| 1235 | // If so, then the value is already null. | 
|---|
| 1236 | if (t->higher_equal(TypePtr::NULL_PTR)) { | 
|---|
| 1237 | NOT_PRODUCT(explicit_null_checks_elided++); | 
|---|
| 1238 | return value;           // Elided null assert quickly! | 
|---|
| 1239 | } | 
|---|
| 1240 | } else { | 
|---|
| 1241 | // See if mixing in the NULL pointer changes type. | 
|---|
| 1242 | // If so, then the NULL pointer was not allowed in the original | 
|---|
| 1243 | // type.  In other words, "value" was not-null. | 
|---|
| 1244 | if (t->meet(TypePtr::NULL_PTR) != t->remove_speculative()) { | 
|---|
| 1245 | // same as: if (!TypePtr::NULL_PTR->higher_equal(t)) ... | 
|---|
| 1246 | NOT_PRODUCT(explicit_null_checks_elided++); | 
|---|
| 1247 | return value;           // Elided null check quickly! | 
|---|
| 1248 | } | 
|---|
| 1249 | } | 
|---|
| 1250 | chk = new CmpPNode( value, null() ); | 
|---|
| 1251 | break; | 
|---|
| 1252 | } | 
|---|
| 1253 |  | 
|---|
| 1254 | default: | 
|---|
| 1255 | fatal( "unexpected type: %s", type2name(type)); | 
|---|
| 1256 | } | 
|---|
| 1257 | assert(chk != NULL, "sanity check"); | 
|---|
| 1258 | chk = _gvn.transform(chk); | 
|---|
| 1259 |  | 
|---|
| 1260 | BoolTest::mask btest = assert_null ? BoolTest::eq : BoolTest::ne; | 
|---|
| 1261 | BoolNode *btst = new BoolNode( chk, btest); | 
|---|
| 1262 | Node   *tst = _gvn.transform( btst ); | 
|---|
| 1263 |  | 
|---|
| 1264 | //----------- | 
|---|
| 1265 | // if peephole optimizations occurred, a prior test existed. | 
|---|
| 1266 | // If a prior test existed, maybe it dominates as we can avoid this test. | 
|---|
| 1267 | if (tst != btst && type == T_OBJECT) { | 
|---|
| 1268 | // At this point we want to scan up the CFG to see if we can | 
|---|
| 1269 | // find an identical test (and so avoid this test altogether). | 
|---|
| 1270 | Node *cfg = control(); | 
|---|
| 1271 | int depth = 0; | 
|---|
| 1272 | while( depth < 16 ) {       // Limit search depth for speed | 
|---|
| 1273 | if( cfg->Opcode() == Op_IfTrue && | 
|---|
| 1274 | cfg->in(0)->in(1) == tst ) { | 
|---|
| 1275 | // Found prior test.  Use "cast_not_null" to construct an identical | 
|---|
| 1276 | // CastPP (and hence hash to) as already exists for the prior test. | 
|---|
| 1277 | // Return that casted value. | 
|---|
| 1278 | if (assert_null) { | 
|---|
| 1279 | replace_in_map(value, null()); | 
|---|
| 1280 | return null();  // do not issue the redundant test | 
|---|
| 1281 | } | 
|---|
| 1282 | Node *oldcontrol = control(); | 
|---|
| 1283 | set_control(cfg); | 
|---|
| 1284 | Node *res = cast_not_null(value); | 
|---|
| 1285 | set_control(oldcontrol); | 
|---|
| 1286 | NOT_PRODUCT(explicit_null_checks_elided++); | 
|---|
| 1287 | return res; | 
|---|
| 1288 | } | 
|---|
| 1289 | cfg = IfNode::up_one_dom(cfg, /*linear_only=*/ true); | 
|---|
| 1290 | if (cfg == NULL)  break;  // Quit at region nodes | 
|---|
| 1291 | depth++; | 
|---|
| 1292 | } | 
|---|
| 1293 | } | 
|---|
| 1294 |  | 
|---|
| 1295 | //----------- | 
|---|
| 1296 | // Branch to failure if null | 
|---|
| 1297 | float ok_prob = PROB_MAX;  // a priori estimate:  nulls never happen | 
|---|
| 1298 | Deoptimization::DeoptReason reason; | 
|---|
| 1299 | if (assert_null) { | 
|---|
| 1300 | reason = Deoptimization::reason_null_assert(speculative); | 
|---|
| 1301 | } else if (type == T_OBJECT) { | 
|---|
| 1302 | reason = Deoptimization::reason_null_check(speculative); | 
|---|
| 1303 | } else { | 
|---|
| 1304 | reason = Deoptimization::Reason_div0_check; | 
|---|
| 1305 | } | 
|---|
| 1306 | // %%% Since Reason_unhandled is not recorded on a per-bytecode basis, | 
|---|
| 1307 | // ciMethodData::has_trap_at will return a conservative -1 if any | 
|---|
| 1308 | // must-be-null assertion has failed.  This could cause performance | 
|---|
| 1309 | // problems for a method after its first do_null_assert failure. | 
|---|
| 1310 | // Consider using 'Reason_class_check' instead? | 
|---|
| 1311 |  | 
|---|
| 1312 | // To cause an implicit null check, we set the not-null probability | 
|---|
| 1313 | // to the maximum (PROB_MAX).  For an explicit check the probability | 
|---|
| 1314 | // is set to a smaller value. | 
|---|
| 1315 | if (null_control != NULL || too_many_traps(reason)) { | 
|---|
| 1316 | // probability is less likely | 
|---|
| 1317 | ok_prob =  PROB_LIKELY_MAG(3); | 
|---|
| 1318 | } else if (!assert_null && | 
|---|
| 1319 | (ImplicitNullCheckThreshold > 0) && | 
|---|
| 1320 | method() != NULL && | 
|---|
| 1321 | (method()->method_data()->trap_count(reason) | 
|---|
| 1322 | >= (uint)ImplicitNullCheckThreshold)) { | 
|---|
| 1323 | ok_prob =  PROB_LIKELY_MAG(3); | 
|---|
| 1324 | } | 
|---|
| 1325 |  | 
|---|
| 1326 | if (null_control != NULL) { | 
|---|
| 1327 | IfNode* iff = create_and_map_if(control(), tst, ok_prob, COUNT_UNKNOWN); | 
|---|
| 1328 | Node* null_true = _gvn.transform( new IfFalseNode(iff)); | 
|---|
| 1329 | set_control(      _gvn.transform( new IfTrueNode(iff))); | 
|---|
| 1330 | #ifndef PRODUCT | 
|---|
| 1331 | if (null_true == top()) { | 
|---|
| 1332 | explicit_null_checks_elided++; | 
|---|
| 1333 | } | 
|---|
| 1334 | #endif | 
|---|
| 1335 | (*null_control) = null_true; | 
|---|
| 1336 | } else { | 
|---|
| 1337 | BuildCutout unless(this, tst, ok_prob); | 
|---|
| 1338 | // Check for optimizer eliding test at parse time | 
|---|
| 1339 | if (stopped()) { | 
|---|
| 1340 | // Failure not possible; do not bother making uncommon trap. | 
|---|
| 1341 | NOT_PRODUCT(explicit_null_checks_elided++); | 
|---|
| 1342 | } else if (assert_null) { | 
|---|
| 1343 | uncommon_trap(reason, | 
|---|
| 1344 | Deoptimization::Action_make_not_entrant, | 
|---|
| 1345 | NULL, "assert_null"); | 
|---|
| 1346 | } else { | 
|---|
| 1347 | replace_in_map(value, zerocon(type)); | 
|---|
| 1348 | builtin_throw(reason); | 
|---|
| 1349 | } | 
|---|
| 1350 | } | 
|---|
| 1351 |  | 
|---|
| 1352 | // Must throw exception, fall-thru not possible? | 
|---|
| 1353 | if (stopped()) { | 
|---|
| 1354 | return top();               // No result | 
|---|
| 1355 | } | 
|---|
| 1356 |  | 
|---|
| 1357 | if (assert_null) { | 
|---|
| 1358 | // Cast obj to null on this path. | 
|---|
| 1359 | replace_in_map(value, zerocon(type)); | 
|---|
| 1360 | return zerocon(type); | 
|---|
| 1361 | } | 
|---|
| 1362 |  | 
|---|
| 1363 | // Cast obj to not-null on this path, if there is no null_control. | 
|---|
| 1364 | // (If there is a null_control, a non-null value may come back to haunt us.) | 
|---|
| 1365 | if (type == T_OBJECT) { | 
|---|
| 1366 | Node* cast = cast_not_null(value, false); | 
|---|
| 1367 | if (null_control == NULL || (*null_control) == top()) | 
|---|
| 1368 | replace_in_map(value, cast); | 
|---|
| 1369 | value = cast; | 
|---|
| 1370 | } | 
|---|
| 1371 |  | 
|---|
| 1372 | return value; | 
|---|
| 1373 | } | 
|---|
| 1374 |  | 
|---|
| 1375 |  | 
|---|
| 1376 | //------------------------------cast_not_null---------------------------------- | 
|---|
| 1377 | // Cast obj to not-null on this path | 
|---|
| 1378 | Node* GraphKit::cast_not_null(Node* obj, bool do_replace_in_map) { | 
|---|
| 1379 | const Type *t = _gvn.type(obj); | 
|---|
| 1380 | const Type *t_not_null = t->join_speculative(TypePtr::NOTNULL); | 
|---|
| 1381 | // Object is already not-null? | 
|---|
| 1382 | if( t == t_not_null ) return obj; | 
|---|
| 1383 |  | 
|---|
| 1384 | Node *cast = new CastPPNode(obj,t_not_null); | 
|---|
| 1385 | cast->init_req(0, control()); | 
|---|
| 1386 | cast = _gvn.transform( cast ); | 
|---|
| 1387 |  | 
|---|
| 1388 | // Scan for instances of 'obj' in the current JVM mapping. | 
|---|
| 1389 | // These instances are known to be not-null after the test. | 
|---|
| 1390 | if (do_replace_in_map) | 
|---|
| 1391 | replace_in_map(obj, cast); | 
|---|
| 1392 |  | 
|---|
| 1393 | return cast;                  // Return casted value | 
|---|
| 1394 | } | 
|---|
| 1395 |  | 
|---|
| 1396 | // Sometimes in intrinsics, we implicitly know an object is not null | 
|---|
| 1397 | // (there's no actual null check) so we can cast it to not null. In | 
|---|
| 1398 | // the course of optimizations, the input to the cast can become null. | 
|---|
| 1399 | // In that case that data path will die and we need the control path | 
|---|
| 1400 | // to become dead as well to keep the graph consistent. So we have to | 
|---|
| 1401 | // add a check for null for which one branch can't be taken. It uses | 
|---|
| 1402 | // an Opaque4 node that will cause the check to be removed after loop | 
|---|
| 1403 | // opts so the test goes away and the compiled code doesn't execute a | 
|---|
| 1404 | // useless check. | 
|---|
| 1405 | Node* GraphKit::must_be_not_null(Node* value, bool do_replace_in_map) { | 
|---|
| 1406 | Node* chk = _gvn.transform(new CmpPNode(value, null())); | 
|---|
| 1407 | Node *tst = _gvn.transform(new BoolNode(chk, BoolTest::ne)); | 
|---|
| 1408 | Node* opaq = _gvn.transform(new Opaque4Node(C, tst, intcon(1))); | 
|---|
| 1409 | IfNode *iff = new IfNode(control(), opaq, PROB_MAX, COUNT_UNKNOWN); | 
|---|
| 1410 | _gvn.set_type(iff, iff->Value(&_gvn)); | 
|---|
| 1411 | Node *if_f = _gvn.transform(new IfFalseNode(iff)); | 
|---|
| 1412 | Node *frame = _gvn.transform(new ParmNode(C->start(), TypeFunc::FramePtr)); | 
|---|
| 1413 | Node *halt = _gvn.transform(new HaltNode(if_f, frame)); | 
|---|
| 1414 | C->root()->add_req(halt); | 
|---|
| 1415 | Node *if_t = _gvn.transform(new IfTrueNode(iff)); | 
|---|
| 1416 | set_control(if_t); | 
|---|
| 1417 | return cast_not_null(value, do_replace_in_map); | 
|---|
| 1418 | } | 
|---|
| 1419 |  | 
|---|
| 1420 |  | 
|---|
| 1421 | //--------------------------replace_in_map------------------------------------- | 
|---|
| 1422 | void GraphKit::replace_in_map(Node* old, Node* neww) { | 
|---|
| 1423 | if (old == neww) { | 
|---|
| 1424 | return; | 
|---|
| 1425 | } | 
|---|
| 1426 |  | 
|---|
| 1427 | map()->replace_edge(old, neww); | 
|---|
| 1428 |  | 
|---|
| 1429 | // Note: This operation potentially replaces any edge | 
|---|
| 1430 | // on the map.  This includes locals, stack, and monitors | 
|---|
| 1431 | // of the current (innermost) JVM state. | 
|---|
| 1432 |  | 
|---|
| 1433 | // don't let inconsistent types from profiling escape this | 
|---|
| 1434 | // method | 
|---|
| 1435 |  | 
|---|
| 1436 | const Type* told = _gvn.type(old); | 
|---|
| 1437 | const Type* tnew = _gvn.type(neww); | 
|---|
| 1438 |  | 
|---|
| 1439 | if (!tnew->higher_equal(told)) { | 
|---|
| 1440 | return; | 
|---|
| 1441 | } | 
|---|
| 1442 |  | 
|---|
| 1443 | map()->record_replaced_node(old, neww); | 
|---|
| 1444 | } | 
|---|
| 1445 |  | 
|---|
| 1446 |  | 
|---|
| 1447 | //============================================================================= | 
|---|
| 1448 | //--------------------------------memory--------------------------------------- | 
|---|
| 1449 | Node* GraphKit::memory(uint alias_idx) { | 
|---|
| 1450 | MergeMemNode* mem = merged_memory(); | 
|---|
| 1451 | Node* p = mem->memory_at(alias_idx); | 
|---|
| 1452 | _gvn.set_type(p, Type::MEMORY);  // must be mapped | 
|---|
| 1453 | return p; | 
|---|
| 1454 | } | 
|---|
| 1455 |  | 
|---|
| 1456 | //-----------------------------reset_memory------------------------------------ | 
|---|
| 1457 | Node* GraphKit::reset_memory() { | 
|---|
| 1458 | Node* mem = map()->memory(); | 
|---|
| 1459 | // do not use this node for any more parsing! | 
|---|
| 1460 | debug_only( map()->set_memory((Node*)NULL) ); | 
|---|
| 1461 | return _gvn.transform( mem ); | 
|---|
| 1462 | } | 
|---|
| 1463 |  | 
|---|
| 1464 | //------------------------------set_all_memory--------------------------------- | 
|---|
| 1465 | void GraphKit::set_all_memory(Node* newmem) { | 
|---|
| 1466 | Node* mergemem = MergeMemNode::make(newmem); | 
|---|
| 1467 | gvn().set_type_bottom(mergemem); | 
|---|
| 1468 | map()->set_memory(mergemem); | 
|---|
| 1469 | } | 
|---|
| 1470 |  | 
|---|
| 1471 | //------------------------------set_all_memory_call---------------------------- | 
|---|
| 1472 | void GraphKit::set_all_memory_call(Node* call, bool separate_io_proj) { | 
|---|
| 1473 | Node* newmem = _gvn.transform( new ProjNode(call, TypeFunc::Memory, separate_io_proj) ); | 
|---|
| 1474 | set_all_memory(newmem); | 
|---|
| 1475 | } | 
|---|
| 1476 |  | 
|---|
| 1477 | //============================================================================= | 
|---|
| 1478 | // | 
|---|
| 1479 | // parser factory methods for MemNodes | 
|---|
| 1480 | // | 
|---|
| 1481 | // These are layered on top of the factory methods in LoadNode and StoreNode, | 
|---|
| 1482 | // and integrate with the parser's memory state and _gvn engine. | 
|---|
| 1483 | // | 
|---|
| 1484 |  | 
|---|
| 1485 | // factory methods in "int adr_idx" | 
|---|
| 1486 | Node* GraphKit::make_load(Node* ctl, Node* adr, const Type* t, BasicType bt, | 
|---|
| 1487 | int adr_idx, | 
|---|
| 1488 | MemNode::MemOrd mo, | 
|---|
| 1489 | LoadNode::ControlDependency control_dependency, | 
|---|
| 1490 | bool require_atomic_access, | 
|---|
| 1491 | bool unaligned, | 
|---|
| 1492 | bool mismatched, | 
|---|
| 1493 | bool unsafe) { | 
|---|
| 1494 | assert(adr_idx != Compile::AliasIdxTop, "use other make_load factory"); | 
|---|
| 1495 | const TypePtr* adr_type = NULL; // debug-mode-only argument | 
|---|
| 1496 | debug_only(adr_type = C->get_adr_type(adr_idx)); | 
|---|
| 1497 | Node* mem = memory(adr_idx); | 
|---|
| 1498 | Node* ld; | 
|---|
| 1499 | if (require_atomic_access && bt == T_LONG) { | 
|---|
| 1500 | ld = LoadLNode::make_atomic(ctl, mem, adr, adr_type, t, mo, control_dependency, unaligned, mismatched, unsafe); | 
|---|
| 1501 | } else if (require_atomic_access && bt == T_DOUBLE) { | 
|---|
| 1502 | ld = LoadDNode::make_atomic(ctl, mem, adr, adr_type, t, mo, control_dependency, unaligned, mismatched, unsafe); | 
|---|
| 1503 | } else { | 
|---|
| 1504 | ld = LoadNode::make(_gvn, ctl, mem, adr, adr_type, t, bt, mo, control_dependency, unaligned, mismatched, unsafe); | 
|---|
| 1505 | } | 
|---|
| 1506 | ld = _gvn.transform(ld); | 
|---|
| 1507 | if (((bt == T_OBJECT) && C->do_escape_analysis()) || C->eliminate_boxing()) { | 
|---|
| 1508 | // Improve graph before escape analysis and boxing elimination. | 
|---|
| 1509 | record_for_igvn(ld); | 
|---|
| 1510 | } | 
|---|
| 1511 | return ld; | 
|---|
| 1512 | } | 
|---|
| 1513 |  | 
|---|
| 1514 | Node* GraphKit::store_to_memory(Node* ctl, Node* adr, Node *val, BasicType bt, | 
|---|
| 1515 | int adr_idx, | 
|---|
| 1516 | MemNode::MemOrd mo, | 
|---|
| 1517 | bool require_atomic_access, | 
|---|
| 1518 | bool unaligned, | 
|---|
| 1519 | bool mismatched, | 
|---|
| 1520 | bool unsafe) { | 
|---|
| 1521 | assert(adr_idx != Compile::AliasIdxTop, "use other store_to_memory factory"); | 
|---|
| 1522 | const TypePtr* adr_type = NULL; | 
|---|
| 1523 | debug_only(adr_type = C->get_adr_type(adr_idx)); | 
|---|
| 1524 | Node *mem = memory(adr_idx); | 
|---|
| 1525 | Node* st; | 
|---|
| 1526 | if (require_atomic_access && bt == T_LONG) { | 
|---|
| 1527 | st = StoreLNode::make_atomic(ctl, mem, adr, adr_type, val, mo); | 
|---|
| 1528 | } else if (require_atomic_access && bt == T_DOUBLE) { | 
|---|
| 1529 | st = StoreDNode::make_atomic(ctl, mem, adr, adr_type, val, mo); | 
|---|
| 1530 | } else { | 
|---|
| 1531 | st = StoreNode::make(_gvn, ctl, mem, adr, adr_type, val, bt, mo); | 
|---|
| 1532 | } | 
|---|
| 1533 | if (unaligned) { | 
|---|
| 1534 | st->as_Store()->set_unaligned_access(); | 
|---|
| 1535 | } | 
|---|
| 1536 | if (mismatched) { | 
|---|
| 1537 | st->as_Store()->set_mismatched_access(); | 
|---|
| 1538 | } | 
|---|
| 1539 | if (unsafe) { | 
|---|
| 1540 | st->as_Store()->set_unsafe_access(); | 
|---|
| 1541 | } | 
|---|
| 1542 | st = _gvn.transform(st); | 
|---|
| 1543 | set_memory(st, adr_idx); | 
|---|
| 1544 | // Back-to-back stores can only remove intermediate store with DU info | 
|---|
| 1545 | // so push on worklist for optimizer. | 
|---|
| 1546 | if (mem->req() > MemNode::Address && adr == mem->in(MemNode::Address)) | 
|---|
| 1547 | record_for_igvn(st); | 
|---|
| 1548 |  | 
|---|
| 1549 | return st; | 
|---|
| 1550 | } | 
|---|
| 1551 |  | 
|---|
| 1552 | Node* GraphKit::access_store_at(Node* obj, | 
|---|
| 1553 | Node* adr, | 
|---|
| 1554 | const TypePtr* adr_type, | 
|---|
| 1555 | Node* val, | 
|---|
| 1556 | const Type* val_type, | 
|---|
| 1557 | BasicType bt, | 
|---|
| 1558 | DecoratorSet decorators) { | 
|---|
| 1559 | // Transformation of a value which could be NULL pointer (CastPP #NULL) | 
|---|
| 1560 | // could be delayed during Parse (for example, in adjust_map_after_if()). | 
|---|
| 1561 | // Execute transformation here to avoid barrier generation in such case. | 
|---|
| 1562 | if (_gvn.type(val) == TypePtr::NULL_PTR) { | 
|---|
| 1563 | val = _gvn.makecon(TypePtr::NULL_PTR); | 
|---|
| 1564 | } | 
|---|
| 1565 |  | 
|---|
| 1566 | if (stopped()) { | 
|---|
| 1567 | return top(); // Dead path ? | 
|---|
| 1568 | } | 
|---|
| 1569 |  | 
|---|
| 1570 | assert(val != NULL, "not dead path"); | 
|---|
| 1571 |  | 
|---|
| 1572 | C2AccessValuePtr addr(adr, adr_type); | 
|---|
| 1573 | C2AccessValue value(val, val_type); | 
|---|
| 1574 | C2ParseAccess access(this, decorators | C2_WRITE_ACCESS, bt, obj, addr); | 
|---|
| 1575 | if (access.is_raw()) { | 
|---|
| 1576 | return _barrier_set->BarrierSetC2::store_at(access, value); | 
|---|
| 1577 | } else { | 
|---|
| 1578 | return _barrier_set->store_at(access, value); | 
|---|
| 1579 | } | 
|---|
| 1580 | } | 
|---|
| 1581 |  | 
|---|
| 1582 | Node* GraphKit::access_load_at(Node* obj,   // containing obj | 
|---|
| 1583 | Node* adr,   // actual adress to store val at | 
|---|
| 1584 | const TypePtr* adr_type, | 
|---|
| 1585 | const Type* val_type, | 
|---|
| 1586 | BasicType bt, | 
|---|
| 1587 | DecoratorSet decorators) { | 
|---|
| 1588 | if (stopped()) { | 
|---|
| 1589 | return top(); // Dead path ? | 
|---|
| 1590 | } | 
|---|
| 1591 |  | 
|---|
| 1592 | C2AccessValuePtr addr(adr, adr_type); | 
|---|
| 1593 | C2ParseAccess access(this, decorators | C2_READ_ACCESS, bt, obj, addr); | 
|---|
| 1594 | if (access.is_raw()) { | 
|---|
| 1595 | return _barrier_set->BarrierSetC2::load_at(access, val_type); | 
|---|
| 1596 | } else { | 
|---|
| 1597 | return _barrier_set->load_at(access, val_type); | 
|---|
| 1598 | } | 
|---|
| 1599 | } | 
|---|
| 1600 |  | 
|---|
| 1601 | Node* GraphKit::access_load(Node* adr,   // actual adress to load val at | 
|---|
| 1602 | const Type* val_type, | 
|---|
| 1603 | BasicType bt, | 
|---|
| 1604 | DecoratorSet decorators) { | 
|---|
| 1605 | if (stopped()) { | 
|---|
| 1606 | return top(); // Dead path ? | 
|---|
| 1607 | } | 
|---|
| 1608 |  | 
|---|
| 1609 | C2AccessValuePtr addr(adr, NULL); | 
|---|
| 1610 | C2ParseAccess access(this, decorators | C2_READ_ACCESS, bt, NULL, addr); | 
|---|
| 1611 | if (access.is_raw()) { | 
|---|
| 1612 | return _barrier_set->BarrierSetC2::load_at(access, val_type); | 
|---|
| 1613 | } else { | 
|---|
| 1614 | return _barrier_set->load_at(access, val_type); | 
|---|
| 1615 | } | 
|---|
| 1616 | } | 
|---|
| 1617 |  | 
|---|
| 1618 | Node* GraphKit::access_atomic_cmpxchg_val_at(Node* obj, | 
|---|
| 1619 | Node* adr, | 
|---|
| 1620 | const TypePtr* adr_type, | 
|---|
| 1621 | int alias_idx, | 
|---|
| 1622 | Node* expected_val, | 
|---|
| 1623 | Node* new_val, | 
|---|
| 1624 | const Type* value_type, | 
|---|
| 1625 | BasicType bt, | 
|---|
| 1626 | DecoratorSet decorators) { | 
|---|
| 1627 | C2AccessValuePtr addr(adr, adr_type); | 
|---|
| 1628 | C2AtomicParseAccess access(this, decorators | C2_READ_ACCESS | C2_WRITE_ACCESS, | 
|---|
| 1629 | bt, obj, addr, alias_idx); | 
|---|
| 1630 | if (access.is_raw()) { | 
|---|
| 1631 | return _barrier_set->BarrierSetC2::atomic_cmpxchg_val_at(access, expected_val, new_val, value_type); | 
|---|
| 1632 | } else { | 
|---|
| 1633 | return _barrier_set->atomic_cmpxchg_val_at(access, expected_val, new_val, value_type); | 
|---|
| 1634 | } | 
|---|
| 1635 | } | 
|---|
| 1636 |  | 
|---|
| 1637 | Node* GraphKit::access_atomic_cmpxchg_bool_at(Node* obj, | 
|---|
| 1638 | Node* adr, | 
|---|
| 1639 | const TypePtr* adr_type, | 
|---|
| 1640 | int alias_idx, | 
|---|
| 1641 | Node* expected_val, | 
|---|
| 1642 | Node* new_val, | 
|---|
| 1643 | const Type* value_type, | 
|---|
| 1644 | BasicType bt, | 
|---|
| 1645 | DecoratorSet decorators) { | 
|---|
| 1646 | C2AccessValuePtr addr(adr, adr_type); | 
|---|
| 1647 | C2AtomicParseAccess access(this, decorators | C2_READ_ACCESS | C2_WRITE_ACCESS, | 
|---|
| 1648 | bt, obj, addr, alias_idx); | 
|---|
| 1649 | if (access.is_raw()) { | 
|---|
| 1650 | return _barrier_set->BarrierSetC2::atomic_cmpxchg_bool_at(access, expected_val, new_val, value_type); | 
|---|
| 1651 | } else { | 
|---|
| 1652 | return _barrier_set->atomic_cmpxchg_bool_at(access, expected_val, new_val, value_type); | 
|---|
| 1653 | } | 
|---|
| 1654 | } | 
|---|
| 1655 |  | 
|---|
| 1656 | Node* GraphKit::access_atomic_xchg_at(Node* obj, | 
|---|
| 1657 | Node* adr, | 
|---|
| 1658 | const TypePtr* adr_type, | 
|---|
| 1659 | int alias_idx, | 
|---|
| 1660 | Node* new_val, | 
|---|
| 1661 | const Type* value_type, | 
|---|
| 1662 | BasicType bt, | 
|---|
| 1663 | DecoratorSet decorators) { | 
|---|
| 1664 | C2AccessValuePtr addr(adr, adr_type); | 
|---|
| 1665 | C2AtomicParseAccess access(this, decorators | C2_READ_ACCESS | C2_WRITE_ACCESS, | 
|---|
| 1666 | bt, obj, addr, alias_idx); | 
|---|
| 1667 | if (access.is_raw()) { | 
|---|
| 1668 | return _barrier_set->BarrierSetC2::atomic_xchg_at(access, new_val, value_type); | 
|---|
| 1669 | } else { | 
|---|
| 1670 | return _barrier_set->atomic_xchg_at(access, new_val, value_type); | 
|---|
| 1671 | } | 
|---|
| 1672 | } | 
|---|
| 1673 |  | 
|---|
| 1674 | Node* GraphKit::access_atomic_add_at(Node* obj, | 
|---|
| 1675 | Node* adr, | 
|---|
| 1676 | const TypePtr* adr_type, | 
|---|
| 1677 | int alias_idx, | 
|---|
| 1678 | Node* new_val, | 
|---|
| 1679 | const Type* value_type, | 
|---|
| 1680 | BasicType bt, | 
|---|
| 1681 | DecoratorSet decorators) { | 
|---|
| 1682 | C2AccessValuePtr addr(adr, adr_type); | 
|---|
| 1683 | C2AtomicParseAccess access(this, decorators | C2_READ_ACCESS | C2_WRITE_ACCESS, bt, obj, addr, alias_idx); | 
|---|
| 1684 | if (access.is_raw()) { | 
|---|
| 1685 | return _barrier_set->BarrierSetC2::atomic_add_at(access, new_val, value_type); | 
|---|
| 1686 | } else { | 
|---|
| 1687 | return _barrier_set->atomic_add_at(access, new_val, value_type); | 
|---|
| 1688 | } | 
|---|
| 1689 | } | 
|---|
| 1690 |  | 
|---|
| 1691 | void GraphKit::access_clone(Node* src, Node* dst, Node* size, bool is_array) { | 
|---|
| 1692 | return _barrier_set->clone(this, src, dst, size, is_array); | 
|---|
| 1693 | } | 
|---|
| 1694 |  | 
|---|
| 1695 | Node* GraphKit::access_resolve(Node* n, DecoratorSet decorators) { | 
|---|
| 1696 | // Use stronger ACCESS_WRITE|ACCESS_READ by default. | 
|---|
| 1697 | if ((decorators & (ACCESS_READ | ACCESS_WRITE)) == 0) { | 
|---|
| 1698 | decorators |= ACCESS_READ | ACCESS_WRITE; | 
|---|
| 1699 | } | 
|---|
| 1700 | return _barrier_set->resolve(this, n, decorators); | 
|---|
| 1701 | } | 
|---|
| 1702 |  | 
|---|
| 1703 | //-------------------------array_element_address------------------------- | 
|---|
| 1704 | Node* GraphKit::array_element_address(Node* ary, Node* idx, BasicType elembt, | 
|---|
| 1705 | const TypeInt* sizetype, Node* ctrl) { | 
|---|
| 1706 | uint shift  = exact_log2(type2aelembytes(elembt)); | 
|---|
| 1707 | uint  = arrayOopDesc::base_offset_in_bytes(elembt); | 
|---|
| 1708 |  | 
|---|
| 1709 | // short-circuit a common case (saves lots of confusing waste motion) | 
|---|
| 1710 | jint idx_con = find_int_con(idx, -1); | 
|---|
| 1711 | if (idx_con >= 0) { | 
|---|
| 1712 | intptr_t offset = header + ((intptr_t)idx_con << shift); | 
|---|
| 1713 | return basic_plus_adr(ary, offset); | 
|---|
| 1714 | } | 
|---|
| 1715 |  | 
|---|
| 1716 | // must be correct type for alignment purposes | 
|---|
| 1717 | Node* base  = basic_plus_adr(ary, header); | 
|---|
| 1718 | idx = Compile::conv_I2X_index(&_gvn, idx, sizetype, ctrl); | 
|---|
| 1719 | Node* scale = _gvn.transform( new LShiftXNode(idx, intcon(shift)) ); | 
|---|
| 1720 | return basic_plus_adr(ary, base, scale); | 
|---|
| 1721 | } | 
|---|
| 1722 |  | 
|---|
| 1723 | //-------------------------load_array_element------------------------- | 
|---|
| 1724 | Node* GraphKit::load_array_element(Node* ctl, Node* ary, Node* idx, const TypeAryPtr* arytype) { | 
|---|
| 1725 | const Type* elemtype = arytype->elem(); | 
|---|
| 1726 | BasicType elembt = elemtype->array_element_basic_type(); | 
|---|
| 1727 | Node* adr = array_element_address(ary, idx, elembt, arytype->size()); | 
|---|
| 1728 | if (elembt == T_NARROWOOP) { | 
|---|
| 1729 | elembt = T_OBJECT; // To satisfy switch in LoadNode::make() | 
|---|
| 1730 | } | 
|---|
| 1731 | Node* ld = make_load(ctl, adr, elemtype, elembt, arytype, MemNode::unordered); | 
|---|
| 1732 | return ld; | 
|---|
| 1733 | } | 
|---|
| 1734 |  | 
|---|
| 1735 | //-------------------------set_arguments_for_java_call------------------------- | 
|---|
| 1736 | // Arguments (pre-popped from the stack) are taken from the JVMS. | 
|---|
| 1737 | void GraphKit::set_arguments_for_java_call(CallJavaNode* call) { | 
|---|
| 1738 | // Add the call arguments: | 
|---|
| 1739 | uint nargs = call->method()->arg_size(); | 
|---|
| 1740 | for (uint i = 0; i < nargs; i++) { | 
|---|
| 1741 | Node* arg = argument(i); | 
|---|
| 1742 | call->init_req(i + TypeFunc::Parms, arg); | 
|---|
| 1743 | } | 
|---|
| 1744 | } | 
|---|
| 1745 |  | 
|---|
| 1746 | //---------------------------set_edges_for_java_call--------------------------- | 
|---|
| 1747 | // Connect a newly created call into the current JVMS. | 
|---|
| 1748 | // A return value node (if any) is returned from set_edges_for_java_call. | 
|---|
| 1749 | void GraphKit::set_edges_for_java_call(CallJavaNode* call, bool must_throw, bool separate_io_proj) { | 
|---|
| 1750 |  | 
|---|
| 1751 | // Add the predefined inputs: | 
|---|
| 1752 | call->init_req( TypeFunc::Control, control() ); | 
|---|
| 1753 | call->init_req( TypeFunc::I_O    , i_o() ); | 
|---|
| 1754 | call->init_req( TypeFunc::Memory , reset_memory() ); | 
|---|
| 1755 | call->init_req( TypeFunc::FramePtr, frameptr() ); | 
|---|
| 1756 | call->init_req( TypeFunc::ReturnAdr, top() ); | 
|---|
| 1757 |  | 
|---|
| 1758 | add_safepoint_edges(call, must_throw); | 
|---|
| 1759 |  | 
|---|
| 1760 | Node* xcall = _gvn.transform(call); | 
|---|
| 1761 |  | 
|---|
| 1762 | if (xcall == top()) { | 
|---|
| 1763 | set_control(top()); | 
|---|
| 1764 | return; | 
|---|
| 1765 | } | 
|---|
| 1766 | assert(xcall == call, "call identity is stable"); | 
|---|
| 1767 |  | 
|---|
| 1768 | // Re-use the current map to produce the result. | 
|---|
| 1769 |  | 
|---|
| 1770 | set_control(_gvn.transform(new ProjNode(call, TypeFunc::Control))); | 
|---|
| 1771 | set_i_o(    _gvn.transform(new ProjNode(call, TypeFunc::I_O    , separate_io_proj))); | 
|---|
| 1772 | set_all_memory_call(xcall, separate_io_proj); | 
|---|
| 1773 |  | 
|---|
| 1774 | //return xcall;   // no need, caller already has it | 
|---|
| 1775 | } | 
|---|
| 1776 |  | 
|---|
| 1777 | Node* GraphKit::set_results_for_java_call(CallJavaNode* call, bool separate_io_proj, bool deoptimize) { | 
|---|
| 1778 | if (stopped())  return top();  // maybe the call folded up? | 
|---|
| 1779 |  | 
|---|
| 1780 | // Capture the return value, if any. | 
|---|
| 1781 | Node* ret; | 
|---|
| 1782 | if (call->method() == NULL || | 
|---|
| 1783 | call->method()->return_type()->basic_type() == T_VOID) | 
|---|
| 1784 | ret = top(); | 
|---|
| 1785 | else  ret = _gvn.transform(new ProjNode(call, TypeFunc::Parms)); | 
|---|
| 1786 |  | 
|---|
| 1787 | // Note:  Since any out-of-line call can produce an exception, | 
|---|
| 1788 | // we always insert an I_O projection from the call into the result. | 
|---|
| 1789 |  | 
|---|
| 1790 | make_slow_call_ex(call, env()->Throwable_klass(), separate_io_proj, deoptimize); | 
|---|
| 1791 |  | 
|---|
| 1792 | if (separate_io_proj) { | 
|---|
| 1793 | // The caller requested separate projections be used by the fall | 
|---|
| 1794 | // through and exceptional paths, so replace the projections for | 
|---|
| 1795 | // the fall through path. | 
|---|
| 1796 | set_i_o(_gvn.transform( new ProjNode(call, TypeFunc::I_O) )); | 
|---|
| 1797 | set_all_memory(_gvn.transform( new ProjNode(call, TypeFunc::Memory) )); | 
|---|
| 1798 | } | 
|---|
| 1799 | return ret; | 
|---|
| 1800 | } | 
|---|
| 1801 |  | 
|---|
| 1802 | //--------------------set_predefined_input_for_runtime_call-------------------- | 
|---|
| 1803 | // Reading and setting the memory state is way conservative here. | 
|---|
| 1804 | // The real problem is that I am not doing real Type analysis on memory, | 
|---|
| 1805 | // so I cannot distinguish card mark stores from other stores.  Across a GC | 
|---|
| 1806 | // point the Store Barrier and the card mark memory has to agree.  I cannot | 
|---|
| 1807 | // have a card mark store and its barrier split across the GC point from | 
|---|
| 1808 | // either above or below.  Here I get that to happen by reading ALL of memory. | 
|---|
| 1809 | // A better answer would be to separate out card marks from other memory. | 
|---|
| 1810 | // For now, return the input memory state, so that it can be reused | 
|---|
| 1811 | // after the call, if this call has restricted memory effects. | 
|---|
| 1812 | Node* GraphKit::set_predefined_input_for_runtime_call(SafePointNode* call, Node* narrow_mem) { | 
|---|
| 1813 | // Set fixed predefined input arguments | 
|---|
| 1814 | Node* memory = reset_memory(); | 
|---|
| 1815 | Node* m = narrow_mem == NULL ? memory : narrow_mem; | 
|---|
| 1816 | call->init_req( TypeFunc::Control,   control()  ); | 
|---|
| 1817 | call->init_req( TypeFunc::I_O,       top()      ); // does no i/o | 
|---|
| 1818 | call->init_req( TypeFunc::Memory,    m          ); // may gc ptrs | 
|---|
| 1819 | call->init_req( TypeFunc::FramePtr,  frameptr() ); | 
|---|
| 1820 | call->init_req( TypeFunc::ReturnAdr, top()      ); | 
|---|
| 1821 | return memory; | 
|---|
| 1822 | } | 
|---|
| 1823 |  | 
|---|
| 1824 | //-------------------set_predefined_output_for_runtime_call-------------------- | 
|---|
| 1825 | // Set control and memory (not i_o) from the call. | 
|---|
| 1826 | // If keep_mem is not NULL, use it for the output state, | 
|---|
| 1827 | // except for the RawPtr output of the call, if hook_mem is TypeRawPtr::BOTTOM. | 
|---|
| 1828 | // If hook_mem is NULL, this call produces no memory effects at all. | 
|---|
| 1829 | // If hook_mem is a Java-visible memory slice (such as arraycopy operands), | 
|---|
| 1830 | // then only that memory slice is taken from the call. | 
|---|
| 1831 | // In the last case, we must put an appropriate memory barrier before | 
|---|
| 1832 | // the call, so as to create the correct anti-dependencies on loads | 
|---|
| 1833 | // preceding the call. | 
|---|
| 1834 | void GraphKit::set_predefined_output_for_runtime_call(Node* call, | 
|---|
| 1835 | Node* keep_mem, | 
|---|
| 1836 | const TypePtr* hook_mem) { | 
|---|
| 1837 | // no i/o | 
|---|
| 1838 | set_control(_gvn.transform( new ProjNode(call,TypeFunc::Control) )); | 
|---|
| 1839 | if (keep_mem) { | 
|---|
| 1840 | // First clone the existing memory state | 
|---|
| 1841 | set_all_memory(keep_mem); | 
|---|
| 1842 | if (hook_mem != NULL) { | 
|---|
| 1843 | // Make memory for the call | 
|---|
| 1844 | Node* mem = _gvn.transform( new ProjNode(call, TypeFunc::Memory) ); | 
|---|
| 1845 | // Set the RawPtr memory state only.  This covers all the heap top/GC stuff | 
|---|
| 1846 | // We also use hook_mem to extract specific effects from arraycopy stubs. | 
|---|
| 1847 | set_memory(mem, hook_mem); | 
|---|
| 1848 | } | 
|---|
| 1849 | // ...else the call has NO memory effects. | 
|---|
| 1850 |  | 
|---|
| 1851 | // Make sure the call advertises its memory effects precisely. | 
|---|
| 1852 | // This lets us build accurate anti-dependences in gcm.cpp. | 
|---|
| 1853 | assert(C->alias_type(call->adr_type()) == C->alias_type(hook_mem), | 
|---|
| 1854 | "call node must be constructed correctly"); | 
|---|
| 1855 | } else { | 
|---|
| 1856 | assert(hook_mem == NULL, ""); | 
|---|
| 1857 | // This is not a "slow path" call; all memory comes from the call. | 
|---|
| 1858 | set_all_memory_call(call); | 
|---|
| 1859 | } | 
|---|
| 1860 | } | 
|---|
| 1861 |  | 
|---|
| 1862 | // Keep track of MergeMems feeding into other MergeMems | 
|---|
| 1863 | static void add_mergemem_users_to_worklist(Unique_Node_List& wl, Node* mem) { | 
|---|
| 1864 | if (!mem->is_MergeMem()) { | 
|---|
| 1865 | return; | 
|---|
| 1866 | } | 
|---|
| 1867 | for (SimpleDUIterator i(mem); i.has_next(); i.next()) { | 
|---|
| 1868 | Node* use = i.get(); | 
|---|
| 1869 | if (use->is_MergeMem()) { | 
|---|
| 1870 | wl.push(use); | 
|---|
| 1871 | } | 
|---|
| 1872 | } | 
|---|
| 1873 | } | 
|---|
| 1874 |  | 
|---|
| 1875 | // Replace the call with the current state of the kit. | 
|---|
| 1876 | void GraphKit::replace_call(CallNode* call, Node* result, bool do_replaced_nodes) { | 
|---|
| 1877 | JVMState* ejvms = NULL; | 
|---|
| 1878 | if (has_exceptions()) { | 
|---|
| 1879 | ejvms = transfer_exceptions_into_jvms(); | 
|---|
| 1880 | } | 
|---|
| 1881 |  | 
|---|
| 1882 | ReplacedNodes replaced_nodes = map()->replaced_nodes(); | 
|---|
| 1883 | ReplacedNodes replaced_nodes_exception; | 
|---|
| 1884 | Node* ex_ctl = top(); | 
|---|
| 1885 |  | 
|---|
| 1886 | SafePointNode* final_state = stop(); | 
|---|
| 1887 |  | 
|---|
| 1888 | // Find all the needed outputs of this call | 
|---|
| 1889 | CallProjections callprojs; | 
|---|
| 1890 | call->extract_projections(&callprojs, true); | 
|---|
| 1891 |  | 
|---|
| 1892 | Unique_Node_List wl; | 
|---|
| 1893 | Node* init_mem = call->in(TypeFunc::Memory); | 
|---|
| 1894 | Node* final_mem = final_state->in(TypeFunc::Memory); | 
|---|
| 1895 | Node* final_ctl = final_state->in(TypeFunc::Control); | 
|---|
| 1896 | Node* final_io = final_state->in(TypeFunc::I_O); | 
|---|
| 1897 |  | 
|---|
| 1898 | // Replace all the old call edges with the edges from the inlining result | 
|---|
| 1899 | if (callprojs.fallthrough_catchproj != NULL) { | 
|---|
| 1900 | C->gvn_replace_by(callprojs.fallthrough_catchproj, final_ctl); | 
|---|
| 1901 | } | 
|---|
| 1902 | if (callprojs.fallthrough_memproj != NULL) { | 
|---|
| 1903 | if (final_mem->is_MergeMem()) { | 
|---|
| 1904 | // Parser's exits MergeMem was not transformed but may be optimized | 
|---|
| 1905 | final_mem = _gvn.transform(final_mem); | 
|---|
| 1906 | } | 
|---|
| 1907 | C->gvn_replace_by(callprojs.fallthrough_memproj,   final_mem); | 
|---|
| 1908 | add_mergemem_users_to_worklist(wl, final_mem); | 
|---|
| 1909 | } | 
|---|
| 1910 | if (callprojs.fallthrough_ioproj != NULL) { | 
|---|
| 1911 | C->gvn_replace_by(callprojs.fallthrough_ioproj,    final_io); | 
|---|
| 1912 | } | 
|---|
| 1913 |  | 
|---|
| 1914 | // Replace the result with the new result if it exists and is used | 
|---|
| 1915 | if (callprojs.resproj != NULL && result != NULL) { | 
|---|
| 1916 | C->gvn_replace_by(callprojs.resproj, result); | 
|---|
| 1917 | } | 
|---|
| 1918 |  | 
|---|
| 1919 | if (ejvms == NULL) { | 
|---|
| 1920 | // No exception edges to simply kill off those paths | 
|---|
| 1921 | if (callprojs.catchall_catchproj != NULL) { | 
|---|
| 1922 | C->gvn_replace_by(callprojs.catchall_catchproj, C->top()); | 
|---|
| 1923 | } | 
|---|
| 1924 | if (callprojs.catchall_memproj != NULL) { | 
|---|
| 1925 | C->gvn_replace_by(callprojs.catchall_memproj,   C->top()); | 
|---|
| 1926 | } | 
|---|
| 1927 | if (callprojs.catchall_ioproj != NULL) { | 
|---|
| 1928 | C->gvn_replace_by(callprojs.catchall_ioproj,    C->top()); | 
|---|
| 1929 | } | 
|---|
| 1930 | // Replace the old exception object with top | 
|---|
| 1931 | if (callprojs.exobj != NULL) { | 
|---|
| 1932 | C->gvn_replace_by(callprojs.exobj, C->top()); | 
|---|
| 1933 | } | 
|---|
| 1934 | } else { | 
|---|
| 1935 | GraphKit ekit(ejvms); | 
|---|
| 1936 |  | 
|---|
| 1937 | // Load my combined exception state into the kit, with all phis transformed: | 
|---|
| 1938 | SafePointNode* ex_map = ekit.combine_and_pop_all_exception_states(); | 
|---|
| 1939 | replaced_nodes_exception = ex_map->replaced_nodes(); | 
|---|
| 1940 |  | 
|---|
| 1941 | Node* ex_oop = ekit.use_exception_state(ex_map); | 
|---|
| 1942 |  | 
|---|
| 1943 | if (callprojs.catchall_catchproj != NULL) { | 
|---|
| 1944 | C->gvn_replace_by(callprojs.catchall_catchproj, ekit.control()); | 
|---|
| 1945 | ex_ctl = ekit.control(); | 
|---|
| 1946 | } | 
|---|
| 1947 | if (callprojs.catchall_memproj != NULL) { | 
|---|
| 1948 | Node* ex_mem = ekit.reset_memory(); | 
|---|
| 1949 | C->gvn_replace_by(callprojs.catchall_memproj,   ex_mem); | 
|---|
| 1950 | add_mergemem_users_to_worklist(wl, ex_mem); | 
|---|
| 1951 | } | 
|---|
| 1952 | if (callprojs.catchall_ioproj != NULL) { | 
|---|
| 1953 | C->gvn_replace_by(callprojs.catchall_ioproj,    ekit.i_o()); | 
|---|
| 1954 | } | 
|---|
| 1955 |  | 
|---|
| 1956 | // Replace the old exception object with the newly created one | 
|---|
| 1957 | if (callprojs.exobj != NULL) { | 
|---|
| 1958 | C->gvn_replace_by(callprojs.exobj, ex_oop); | 
|---|
| 1959 | } | 
|---|
| 1960 | } | 
|---|
| 1961 |  | 
|---|
| 1962 | // Disconnect the call from the graph | 
|---|
| 1963 | call->disconnect_inputs(NULL, C); | 
|---|
| 1964 | C->gvn_replace_by(call, C->top()); | 
|---|
| 1965 |  | 
|---|
| 1966 | // Clean up any MergeMems that feed other MergeMems since the | 
|---|
| 1967 | // optimizer doesn't like that. | 
|---|
| 1968 | while (wl.size() > 0) { | 
|---|
| 1969 | _gvn.transform(wl.pop()); | 
|---|
| 1970 | } | 
|---|
| 1971 |  | 
|---|
| 1972 | if (callprojs.fallthrough_catchproj != NULL && !final_ctl->is_top() && do_replaced_nodes) { | 
|---|
| 1973 | replaced_nodes.apply(C, final_ctl); | 
|---|
| 1974 | } | 
|---|
| 1975 | if (!ex_ctl->is_top() && do_replaced_nodes) { | 
|---|
| 1976 | replaced_nodes_exception.apply(C, ex_ctl); | 
|---|
| 1977 | } | 
|---|
| 1978 | } | 
|---|
| 1979 |  | 
|---|
| 1980 |  | 
|---|
| 1981 | //------------------------------increment_counter------------------------------ | 
|---|
| 1982 | // for statistics: increment a VM counter by 1 | 
|---|
| 1983 |  | 
|---|
| 1984 | void GraphKit::increment_counter(address counter_addr) { | 
|---|
| 1985 | Node* adr1 = makecon(TypeRawPtr::make(counter_addr)); | 
|---|
| 1986 | increment_counter(adr1); | 
|---|
| 1987 | } | 
|---|
| 1988 |  | 
|---|
| 1989 | void GraphKit::increment_counter(Node* counter_addr) { | 
|---|
| 1990 | int adr_type = Compile::AliasIdxRaw; | 
|---|
| 1991 | Node* ctrl = control(); | 
|---|
| 1992 | Node* cnt  = make_load(ctrl, counter_addr, TypeInt::INT, T_INT, adr_type, MemNode::unordered); | 
|---|
| 1993 | Node* incr = _gvn.transform(new AddINode(cnt, _gvn.intcon(1))); | 
|---|
| 1994 | store_to_memory(ctrl, counter_addr, incr, T_INT, adr_type, MemNode::unordered); | 
|---|
| 1995 | } | 
|---|
| 1996 |  | 
|---|
| 1997 |  | 
|---|
| 1998 | //------------------------------uncommon_trap---------------------------------- | 
|---|
| 1999 | // Bail out to the interpreter in mid-method.  Implemented by calling the | 
|---|
| 2000 | // uncommon_trap blob.  This helper function inserts a runtime call with the | 
|---|
| 2001 | // right debug info. | 
|---|
| 2002 | void GraphKit::uncommon_trap(int trap_request, | 
|---|
| 2003 | ciKlass* klass, const char* , | 
|---|
| 2004 | bool must_throw, | 
|---|
| 2005 | bool keep_exact_action) { | 
|---|
| 2006 | if (failing())  stop(); | 
|---|
| 2007 | if (stopped())  return; // trap reachable? | 
|---|
| 2008 |  | 
|---|
| 2009 | // Note:  If ProfileTraps is true, and if a deopt. actually | 
|---|
| 2010 | // occurs here, the runtime will make sure an MDO exists.  There is | 
|---|
| 2011 | // no need to call method()->ensure_method_data() at this point. | 
|---|
| 2012 |  | 
|---|
| 2013 | // Set the stack pointer to the right value for reexecution: | 
|---|
| 2014 | set_sp(reexecute_sp()); | 
|---|
| 2015 |  | 
|---|
| 2016 | #ifdef ASSERT | 
|---|
| 2017 | if (!must_throw) { | 
|---|
| 2018 | // Make sure the stack has at least enough depth to execute | 
|---|
| 2019 | // the current bytecode. | 
|---|
| 2020 | int inputs, ignored_depth; | 
|---|
| 2021 | if (compute_stack_effects(inputs, ignored_depth)) { | 
|---|
| 2022 | assert(sp() >= inputs, "must have enough JVMS stack to execute %s: sp=%d, inputs=%d", | 
|---|
| 2023 | Bytecodes::name(java_bc()), sp(), inputs); | 
|---|
| 2024 | } | 
|---|
| 2025 | } | 
|---|
| 2026 | #endif | 
|---|
| 2027 |  | 
|---|
| 2028 | Deoptimization::DeoptReason reason = Deoptimization::trap_request_reason(trap_request); | 
|---|
| 2029 | Deoptimization::DeoptAction action = Deoptimization::trap_request_action(trap_request); | 
|---|
| 2030 |  | 
|---|
| 2031 | switch (action) { | 
|---|
| 2032 | case Deoptimization::Action_maybe_recompile: | 
|---|
| 2033 | case Deoptimization::Action_reinterpret: | 
|---|
| 2034 | // Temporary fix for 6529811 to allow virtual calls to be sure they | 
|---|
| 2035 | // get the chance to go from mono->bi->mega | 
|---|
| 2036 | if (!keep_exact_action && | 
|---|
| 2037 | Deoptimization::trap_request_index(trap_request) < 0 && | 
|---|
| 2038 | too_many_recompiles(reason)) { | 
|---|
| 2039 | // This BCI is causing too many recompilations. | 
|---|
| 2040 | if (C->log() != NULL) { | 
|---|
| 2041 | C->log()->elem( "observe that='trap_action_change' reason='%s' from='%s' to='none'", | 
|---|
| 2042 | Deoptimization::trap_reason_name(reason), | 
|---|
| 2043 | Deoptimization::trap_action_name(action)); | 
|---|
| 2044 | } | 
|---|
| 2045 | action = Deoptimization::Action_none; | 
|---|
| 2046 | trap_request = Deoptimization::make_trap_request(reason, action); | 
|---|
| 2047 | } else { | 
|---|
| 2048 | C->set_trap_can_recompile(true); | 
|---|
| 2049 | } | 
|---|
| 2050 | break; | 
|---|
| 2051 | case Deoptimization::Action_make_not_entrant: | 
|---|
| 2052 | C->set_trap_can_recompile(true); | 
|---|
| 2053 | break; | 
|---|
| 2054 | case Deoptimization::Action_none: | 
|---|
| 2055 | case Deoptimization::Action_make_not_compilable: | 
|---|
| 2056 | break; | 
|---|
| 2057 | default: | 
|---|
| 2058 | #ifdef ASSERT | 
|---|
| 2059 | fatal( "unknown action %d: %s", action, Deoptimization::trap_action_name(action)); | 
|---|
| 2060 | #endif | 
|---|
| 2061 | break; | 
|---|
| 2062 | } | 
|---|
| 2063 |  | 
|---|
| 2064 | if (TraceOptoParse) { | 
|---|
| 2065 | char buf[100]; | 
|---|
| 2066 | tty->print_cr( "Uncommon trap %s at bci:%d", | 
|---|
| 2067 | Deoptimization::format_trap_request(buf, sizeof(buf), | 
|---|
| 2068 | trap_request), bci()); | 
|---|
| 2069 | } | 
|---|
| 2070 |  | 
|---|
| 2071 | CompileLog* log = C->log(); | 
|---|
| 2072 | if (log != NULL) { | 
|---|
| 2073 | int kid = (klass == NULL)? -1: log->identify(klass); | 
|---|
| 2074 | log->begin_elem( "uncommon_trap bci='%d'", bci()); | 
|---|
| 2075 | char buf[100]; | 
|---|
| 2076 | log->print( " %s", Deoptimization::format_trap_request(buf, sizeof(buf), | 
|---|
| 2077 | trap_request)); | 
|---|
| 2078 | if (kid >= 0)         log->print( " klass='%d'", kid); | 
|---|
| 2079 | if (comment != NULL)  log->print( " comment='%s'", comment); | 
|---|
| 2080 | log->end_elem(); | 
|---|
| 2081 | } | 
|---|
| 2082 |  | 
|---|
| 2083 | // Make sure any guarding test views this path as very unlikely | 
|---|
| 2084 | Node *i0 = control()->in(0); | 
|---|
| 2085 | if (i0 != NULL && i0->is_If()) {        // Found a guarding if test? | 
|---|
| 2086 | IfNode *iff = i0->as_If(); | 
|---|
| 2087 | float f = iff->_prob;   // Get prob | 
|---|
| 2088 | if (control()->Opcode() == Op_IfTrue) { | 
|---|
| 2089 | if (f > PROB_UNLIKELY_MAG(4)) | 
|---|
| 2090 | iff->_prob = PROB_MIN; | 
|---|
| 2091 | } else { | 
|---|
| 2092 | if (f < PROB_LIKELY_MAG(4)) | 
|---|
| 2093 | iff->_prob = PROB_MAX; | 
|---|
| 2094 | } | 
|---|
| 2095 | } | 
|---|
| 2096 |  | 
|---|
| 2097 | // Clear out dead values from the debug info. | 
|---|
| 2098 | kill_dead_locals(); | 
|---|
| 2099 |  | 
|---|
| 2100 | // Now insert the uncommon trap subroutine call | 
|---|
| 2101 | address call_addr = SharedRuntime::uncommon_trap_blob()->entry_point(); | 
|---|
| 2102 | const TypePtr* no_memory_effects = NULL; | 
|---|
| 2103 | // Pass the index of the class to be loaded | 
|---|
| 2104 | Node* call = make_runtime_call(RC_NO_LEAF | RC_UNCOMMON | | 
|---|
| 2105 | (must_throw ? RC_MUST_THROW : 0), | 
|---|
| 2106 | OptoRuntime::uncommon_trap_Type(), | 
|---|
| 2107 | call_addr, "uncommon_trap", no_memory_effects, | 
|---|
| 2108 | intcon(trap_request)); | 
|---|
| 2109 | assert(call->as_CallStaticJava()->uncommon_trap_request() == trap_request, | 
|---|
| 2110 | "must extract request correctly from the graph"); | 
|---|
| 2111 | assert(trap_request != 0, "zero value reserved by uncommon_trap_request"); | 
|---|
| 2112 |  | 
|---|
| 2113 | call->set_req(TypeFunc::ReturnAdr, returnadr()); | 
|---|
| 2114 | // The debug info is the only real input to this call. | 
|---|
| 2115 |  | 
|---|
| 2116 | // Halt-and-catch fire here.  The above call should never return! | 
|---|
| 2117 | HaltNode* halt = new HaltNode(control(), frameptr()); | 
|---|
| 2118 | _gvn.set_type_bottom(halt); | 
|---|
| 2119 | root()->add_req(halt); | 
|---|
| 2120 |  | 
|---|
| 2121 | stop_and_kill_map(); | 
|---|
| 2122 | } | 
|---|
| 2123 |  | 
|---|
| 2124 |  | 
|---|
| 2125 | //--------------------------just_allocated_object------------------------------ | 
|---|
| 2126 | // Report the object that was just allocated. | 
|---|
| 2127 | // It must be the case that there are no intervening safepoints. | 
|---|
| 2128 | // We use this to determine if an object is so "fresh" that | 
|---|
| 2129 | // it does not require card marks. | 
|---|
| 2130 | Node* GraphKit::just_allocated_object(Node* current_control) { | 
|---|
| 2131 | Node* ctrl = current_control; | 
|---|
| 2132 | // Object::<init> is invoked after allocation, most of invoke nodes | 
|---|
| 2133 | // will be reduced, but a region node is kept in parse time, we check | 
|---|
| 2134 | // the pattern and skip the region node if it degraded to a copy. | 
|---|
| 2135 | if (ctrl != NULL && ctrl->is_Region() && ctrl->req() == 2 && | 
|---|
| 2136 | ctrl->as_Region()->is_copy()) { | 
|---|
| 2137 | ctrl = ctrl->as_Region()->is_copy(); | 
|---|
| 2138 | } | 
|---|
| 2139 | if (C->recent_alloc_ctl() == ctrl) { | 
|---|
| 2140 | return C->recent_alloc_obj(); | 
|---|
| 2141 | } | 
|---|
| 2142 | return NULL; | 
|---|
| 2143 | } | 
|---|
| 2144 |  | 
|---|
| 2145 |  | 
|---|
| 2146 | void GraphKit::round_double_arguments(ciMethod* dest_method) { | 
|---|
| 2147 | // (Note:  TypeFunc::make has a cache that makes this fast.) | 
|---|
| 2148 | const TypeFunc* tf    = TypeFunc::make(dest_method); | 
|---|
| 2149 | int             nargs = tf->domain()->cnt() - TypeFunc::Parms; | 
|---|
| 2150 | for (int j = 0; j < nargs; j++) { | 
|---|
| 2151 | const Type *targ = tf->domain()->field_at(j + TypeFunc::Parms); | 
|---|
| 2152 | if( targ->basic_type() == T_DOUBLE ) { | 
|---|
| 2153 | // If any parameters are doubles, they must be rounded before | 
|---|
| 2154 | // the call, dstore_rounding does gvn.transform | 
|---|
| 2155 | Node *arg = argument(j); | 
|---|
| 2156 | arg = dstore_rounding(arg); | 
|---|
| 2157 | set_argument(j, arg); | 
|---|
| 2158 | } | 
|---|
| 2159 | } | 
|---|
| 2160 | } | 
|---|
| 2161 |  | 
|---|
| 2162 | /** | 
|---|
| 2163 | * Record profiling data exact_kls for Node n with the type system so | 
|---|
| 2164 | * that it can propagate it (speculation) | 
|---|
| 2165 | * | 
|---|
| 2166 | * @param n          node that the type applies to | 
|---|
| 2167 | * @param exact_kls  type from profiling | 
|---|
| 2168 | * @param maybe_null did profiling see null? | 
|---|
| 2169 | * | 
|---|
| 2170 | * @return           node with improved type | 
|---|
| 2171 | */ | 
|---|
| 2172 | Node* GraphKit::record_profile_for_speculation(Node* n, ciKlass* exact_kls, ProfilePtrKind ptr_kind) { | 
|---|
| 2173 | const Type* current_type = _gvn.type(n); | 
|---|
| 2174 | assert(UseTypeSpeculation, "type speculation must be on"); | 
|---|
| 2175 |  | 
|---|
| 2176 | const TypePtr* speculative = current_type->speculative(); | 
|---|
| 2177 |  | 
|---|
| 2178 | // Should the klass from the profile be recorded in the speculative type? | 
|---|
| 2179 | if (current_type->would_improve_type(exact_kls, jvms()->depth())) { | 
|---|
| 2180 | const TypeKlassPtr* tklass = TypeKlassPtr::make(exact_kls); | 
|---|
| 2181 | const TypeOopPtr* xtype = tklass->as_instance_type(); | 
|---|
| 2182 | assert(xtype->klass_is_exact(), "Should be exact"); | 
|---|
| 2183 | // Any reason to believe n is not null (from this profiling or a previous one)? | 
|---|
| 2184 | assert(ptr_kind != ProfileAlwaysNull, "impossible here"); | 
|---|
| 2185 | const TypePtr* ptr = (ptr_kind == ProfileMaybeNull && current_type->speculative_maybe_null()) ? TypePtr::BOTTOM : TypePtr::NOTNULL; | 
|---|
| 2186 | // record the new speculative type's depth | 
|---|
| 2187 | speculative = xtype->cast_to_ptr_type(ptr->ptr())->is_ptr(); | 
|---|
| 2188 | speculative = speculative->with_inline_depth(jvms()->depth()); | 
|---|
| 2189 | } else if (current_type->would_improve_ptr(ptr_kind)) { | 
|---|
| 2190 | // Profiling report that null was never seen so we can change the | 
|---|
| 2191 | // speculative type to non null ptr. | 
|---|
| 2192 | if (ptr_kind == ProfileAlwaysNull) { | 
|---|
| 2193 | speculative = TypePtr::NULL_PTR; | 
|---|
| 2194 | } else { | 
|---|
| 2195 | assert(ptr_kind == ProfileNeverNull, "nothing else is an improvement"); | 
|---|
| 2196 | const TypePtr* ptr = TypePtr::NOTNULL; | 
|---|
| 2197 | if (speculative != NULL) { | 
|---|
| 2198 | speculative = speculative->cast_to_ptr_type(ptr->ptr())->is_ptr(); | 
|---|
| 2199 | } else { | 
|---|
| 2200 | speculative = ptr; | 
|---|
| 2201 | } | 
|---|
| 2202 | } | 
|---|
| 2203 | } | 
|---|
| 2204 |  | 
|---|
| 2205 | if (speculative != current_type->speculative()) { | 
|---|
| 2206 | // Build a type with a speculative type (what we think we know | 
|---|
| 2207 | // about the type but will need a guard when we use it) | 
|---|
| 2208 | const TypeOopPtr* spec_type = TypeOopPtr::make(TypePtr::BotPTR, Type::OffsetBot, TypeOopPtr::InstanceBot, speculative); | 
|---|
| 2209 | // We're changing the type, we need a new CheckCast node to carry | 
|---|
| 2210 | // the new type. The new type depends on the control: what | 
|---|
| 2211 | // profiling tells us is only valid from here as far as we can | 
|---|
| 2212 | // tell. | 
|---|
| 2213 | Node* cast = new CheckCastPPNode(control(), n, current_type->remove_speculative()->join_speculative(spec_type)); | 
|---|
| 2214 | cast = _gvn.transform(cast); | 
|---|
| 2215 | replace_in_map(n, cast); | 
|---|
| 2216 | n = cast; | 
|---|
| 2217 | } | 
|---|
| 2218 |  | 
|---|
| 2219 | return n; | 
|---|
| 2220 | } | 
|---|
| 2221 |  | 
|---|
| 2222 | /** | 
|---|
| 2223 | * Record profiling data from receiver profiling at an invoke with the | 
|---|
| 2224 | * type system so that it can propagate it (speculation) | 
|---|
| 2225 | * | 
|---|
| 2226 | * @param n  receiver node | 
|---|
| 2227 | * | 
|---|
| 2228 | * @return   node with improved type | 
|---|
| 2229 | */ | 
|---|
| 2230 | Node* GraphKit::record_profiled_receiver_for_speculation(Node* n) { | 
|---|
| 2231 | if (!UseTypeSpeculation) { | 
|---|
| 2232 | return n; | 
|---|
| 2233 | } | 
|---|
| 2234 | ciKlass* exact_kls = profile_has_unique_klass(); | 
|---|
| 2235 | ProfilePtrKind ptr_kind = ProfileMaybeNull; | 
|---|
| 2236 | if ((java_bc() == Bytecodes::_checkcast || | 
|---|
| 2237 | java_bc() == Bytecodes::_instanceof || | 
|---|
| 2238 | java_bc() == Bytecodes::_aastore) && | 
|---|
| 2239 | method()->method_data()->is_mature()) { | 
|---|
| 2240 | ciProfileData* data = method()->method_data()->bci_to_data(bci()); | 
|---|
| 2241 | if (data != NULL) { | 
|---|
| 2242 | if (!data->as_BitData()->null_seen()) { | 
|---|
| 2243 | ptr_kind = ProfileNeverNull; | 
|---|
| 2244 | } else { | 
|---|
| 2245 | assert(data->is_ReceiverTypeData(), "bad profile data type"); | 
|---|
| 2246 | ciReceiverTypeData* call = (ciReceiverTypeData*)data->as_ReceiverTypeData(); | 
|---|
| 2247 | uint i = 0; | 
|---|
| 2248 | for (; i < call->row_limit(); i++) { | 
|---|
| 2249 | ciKlass* receiver = call->receiver(i); | 
|---|
| 2250 | if (receiver != NULL) { | 
|---|
| 2251 | break; | 
|---|
| 2252 | } | 
|---|
| 2253 | } | 
|---|
| 2254 | ptr_kind = (i == call->row_limit()) ? ProfileAlwaysNull : ProfileMaybeNull; | 
|---|
| 2255 | } | 
|---|
| 2256 | } | 
|---|
| 2257 | } | 
|---|
| 2258 | return record_profile_for_speculation(n, exact_kls, ptr_kind); | 
|---|
| 2259 | } | 
|---|
| 2260 |  | 
|---|
| 2261 | /** | 
|---|
| 2262 | * Record profiling data from argument profiling at an invoke with the | 
|---|
| 2263 | * type system so that it can propagate it (speculation) | 
|---|
| 2264 | * | 
|---|
| 2265 | * @param dest_method  target method for the call | 
|---|
| 2266 | * @param bc           what invoke bytecode is this? | 
|---|
| 2267 | */ | 
|---|
| 2268 | void GraphKit::record_profiled_arguments_for_speculation(ciMethod* dest_method, Bytecodes::Code bc) { | 
|---|
| 2269 | if (!UseTypeSpeculation) { | 
|---|
| 2270 | return; | 
|---|
| 2271 | } | 
|---|
| 2272 | const TypeFunc* tf    = TypeFunc::make(dest_method); | 
|---|
| 2273 | int             nargs = tf->domain()->cnt() - TypeFunc::Parms; | 
|---|
| 2274 | int skip = Bytecodes::has_receiver(bc) ? 1 : 0; | 
|---|
| 2275 | for (int j = skip, i = 0; j < nargs && i < TypeProfileArgsLimit; j++) { | 
|---|
| 2276 | const Type *targ = tf->domain()->field_at(j + TypeFunc::Parms); | 
|---|
| 2277 | if (targ->basic_type() == T_OBJECT || targ->basic_type() == T_ARRAY) { | 
|---|
| 2278 | ProfilePtrKind ptr_kind = ProfileMaybeNull; | 
|---|
| 2279 | ciKlass* better_type = NULL; | 
|---|
| 2280 | if (method()->argument_profiled_type(bci(), i, better_type, ptr_kind)) { | 
|---|
| 2281 | record_profile_for_speculation(argument(j), better_type, ptr_kind); | 
|---|
| 2282 | } | 
|---|
| 2283 | i++; | 
|---|
| 2284 | } | 
|---|
| 2285 | } | 
|---|
| 2286 | } | 
|---|
| 2287 |  | 
|---|
| 2288 | /** | 
|---|
| 2289 | * Record profiling data from parameter profiling at an invoke with | 
|---|
| 2290 | * the type system so that it can propagate it (speculation) | 
|---|
| 2291 | */ | 
|---|
| 2292 | void GraphKit::record_profiled_parameters_for_speculation() { | 
|---|
| 2293 | if (!UseTypeSpeculation) { | 
|---|
| 2294 | return; | 
|---|
| 2295 | } | 
|---|
| 2296 | for (int i = 0, j = 0; i < method()->arg_size() ; i++) { | 
|---|
| 2297 | if (_gvn.type(local(i))->isa_oopptr()) { | 
|---|
| 2298 | ProfilePtrKind ptr_kind = ProfileMaybeNull; | 
|---|
| 2299 | ciKlass* better_type = NULL; | 
|---|
| 2300 | if (method()->parameter_profiled_type(j, better_type, ptr_kind)) { | 
|---|
| 2301 | record_profile_for_speculation(local(i), better_type, ptr_kind); | 
|---|
| 2302 | } | 
|---|
| 2303 | j++; | 
|---|
| 2304 | } | 
|---|
| 2305 | } | 
|---|
| 2306 | } | 
|---|
| 2307 |  | 
|---|
| 2308 | /** | 
|---|
| 2309 | * Record profiling data from return value profiling at an invoke with | 
|---|
| 2310 | * the type system so that it can propagate it (speculation) | 
|---|
| 2311 | */ | 
|---|
| 2312 | void GraphKit::record_profiled_return_for_speculation() { | 
|---|
| 2313 | if (!UseTypeSpeculation) { | 
|---|
| 2314 | return; | 
|---|
| 2315 | } | 
|---|
| 2316 | ProfilePtrKind ptr_kind = ProfileMaybeNull; | 
|---|
| 2317 | ciKlass* better_type = NULL; | 
|---|
| 2318 | if (method()->return_profiled_type(bci(), better_type, ptr_kind)) { | 
|---|
| 2319 | // If profiling reports a single type for the return value, | 
|---|
| 2320 | // feed it to the type system so it can propagate it as a | 
|---|
| 2321 | // speculative type | 
|---|
| 2322 | record_profile_for_speculation(stack(sp()-1), better_type, ptr_kind); | 
|---|
| 2323 | } | 
|---|
| 2324 | } | 
|---|
| 2325 |  | 
|---|
| 2326 | void GraphKit::round_double_result(ciMethod* dest_method) { | 
|---|
| 2327 | // A non-strict method may return a double value which has an extended | 
|---|
| 2328 | // exponent, but this must not be visible in a caller which is 'strict' | 
|---|
| 2329 | // If a strict caller invokes a non-strict callee, round a double result | 
|---|
| 2330 |  | 
|---|
| 2331 | BasicType result_type = dest_method->return_type()->basic_type(); | 
|---|
| 2332 | assert( method() != NULL, "must have caller context"); | 
|---|
| 2333 | if( result_type == T_DOUBLE && method()->is_strict() && !dest_method->is_strict() ) { | 
|---|
| 2334 | // Destination method's return value is on top of stack | 
|---|
| 2335 | // dstore_rounding() does gvn.transform | 
|---|
| 2336 | Node *result = pop_pair(); | 
|---|
| 2337 | result = dstore_rounding(result); | 
|---|
| 2338 | push_pair(result); | 
|---|
| 2339 | } | 
|---|
| 2340 | } | 
|---|
| 2341 |  | 
|---|
| 2342 | // rounding for strict float precision conformance | 
|---|
| 2343 | Node* GraphKit::precision_rounding(Node* n) { | 
|---|
| 2344 | return UseStrictFP && _method->flags().is_strict() | 
|---|
| 2345 | && UseSSE == 0 && Matcher::strict_fp_requires_explicit_rounding | 
|---|
| 2346 | ? _gvn.transform( new RoundFloatNode(0, n) ) | 
|---|
| 2347 | : n; | 
|---|
| 2348 | } | 
|---|
| 2349 |  | 
|---|
| 2350 | // rounding for strict double precision conformance | 
|---|
| 2351 | Node* GraphKit::dprecision_rounding(Node *n) { | 
|---|
| 2352 | return UseStrictFP && _method->flags().is_strict() | 
|---|
| 2353 | && UseSSE <= 1 && Matcher::strict_fp_requires_explicit_rounding | 
|---|
| 2354 | ? _gvn.transform( new RoundDoubleNode(0, n) ) | 
|---|
| 2355 | : n; | 
|---|
| 2356 | } | 
|---|
| 2357 |  | 
|---|
| 2358 | // rounding for non-strict double stores | 
|---|
| 2359 | Node* GraphKit::dstore_rounding(Node* n) { | 
|---|
| 2360 | return Matcher::strict_fp_requires_explicit_rounding | 
|---|
| 2361 | && UseSSE <= 1 | 
|---|
| 2362 | ? _gvn.transform( new RoundDoubleNode(0, n) ) | 
|---|
| 2363 | : n; | 
|---|
| 2364 | } | 
|---|
| 2365 |  | 
|---|
| 2366 | //============================================================================= | 
|---|
| 2367 | // Generate a fast path/slow path idiom.  Graph looks like: | 
|---|
| 2368 | // [foo] indicates that 'foo' is a parameter | 
|---|
| 2369 | // | 
|---|
| 2370 | //              [in]     NULL | 
|---|
| 2371 | //                 \    / | 
|---|
| 2372 | //                  CmpP | 
|---|
| 2373 | //                  Bool ne | 
|---|
| 2374 | //                   If | 
|---|
| 2375 | //                  /  \ | 
|---|
| 2376 | //              True    False-<2> | 
|---|
| 2377 | //              / | | 
|---|
| 2378 | //             /  cast_not_null | 
|---|
| 2379 | //           Load  |    |   ^ | 
|---|
| 2380 | //        [fast_test]   |   | | 
|---|
| 2381 | // gvn to   opt_test    |   | | 
|---|
| 2382 | //          /    \      |  <1> | 
|---|
| 2383 | //      True     False  | | 
|---|
| 2384 | //        |         \\  | | 
|---|
| 2385 | //   [slow_call]     \[fast_result] | 
|---|
| 2386 | //    Ctl   Val       \      \ | 
|---|
| 2387 | //     |               \      \ | 
|---|
| 2388 | //    Catch       <1>   \      \ | 
|---|
| 2389 | //   /    \        ^     \      \ | 
|---|
| 2390 | //  Ex    No_Ex    |      \      \ | 
|---|
| 2391 | //  |       \   \  |       \ <2>  \ | 
|---|
| 2392 | //  ...      \  [slow_res] |  |    \   [null_result] | 
|---|
| 2393 | //            \         \--+--+---  |  | | 
|---|
| 2394 | //             \           | /    \ | / | 
|---|
| 2395 | //              --------Region     Phi | 
|---|
| 2396 | // | 
|---|
| 2397 | //============================================================================= | 
|---|
| 2398 | // Code is structured as a series of driver functions all called 'do_XXX' that | 
|---|
| 2399 | // call a set of helper functions.  Helper functions first, then drivers. | 
|---|
| 2400 |  | 
|---|
| 2401 | //------------------------------null_check_oop--------------------------------- | 
|---|
| 2402 | // Null check oop.  Set null-path control into Region in slot 3. | 
|---|
| 2403 | // Make a cast-not-nullness use the other not-null control.  Return cast. | 
|---|
| 2404 | Node* GraphKit::null_check_oop(Node* value, Node* *null_control, | 
|---|
| 2405 | bool never_see_null, | 
|---|
| 2406 | bool safe_for_replace, | 
|---|
| 2407 | bool speculative) { | 
|---|
| 2408 | // Initial NULL check taken path | 
|---|
| 2409 | (*null_control) = top(); | 
|---|
| 2410 | Node* cast = null_check_common(value, T_OBJECT, false, null_control, speculative); | 
|---|
| 2411 |  | 
|---|
| 2412 | // Generate uncommon_trap: | 
|---|
| 2413 | if (never_see_null && (*null_control) != top()) { | 
|---|
| 2414 | // If we see an unexpected null at a check-cast we record it and force a | 
|---|
| 2415 | // recompile; the offending check-cast will be compiled to handle NULLs. | 
|---|
| 2416 | // If we see more than one offending BCI, then all checkcasts in the | 
|---|
| 2417 | // method will be compiled to handle NULLs. | 
|---|
| 2418 | PreserveJVMState pjvms(this); | 
|---|
| 2419 | set_control(*null_control); | 
|---|
| 2420 | replace_in_map(value, null()); | 
|---|
| 2421 | Deoptimization::DeoptReason reason = Deoptimization::reason_null_check(speculative); | 
|---|
| 2422 | uncommon_trap(reason, | 
|---|
| 2423 | Deoptimization::Action_make_not_entrant); | 
|---|
| 2424 | (*null_control) = top();    // NULL path is dead | 
|---|
| 2425 | } | 
|---|
| 2426 | if ((*null_control) == top() && safe_for_replace) { | 
|---|
| 2427 | replace_in_map(value, cast); | 
|---|
| 2428 | } | 
|---|
| 2429 |  | 
|---|
| 2430 | // Cast away null-ness on the result | 
|---|
| 2431 | return cast; | 
|---|
| 2432 | } | 
|---|
| 2433 |  | 
|---|
| 2434 | //------------------------------opt_iff---------------------------------------- | 
|---|
| 2435 | // Optimize the fast-check IfNode.  Set the fast-path region slot 2. | 
|---|
| 2436 | // Return slow-path control. | 
|---|
| 2437 | Node* GraphKit::opt_iff(Node* region, Node* iff) { | 
|---|
| 2438 | IfNode *opt_iff = _gvn.transform(iff)->as_If(); | 
|---|
| 2439 |  | 
|---|
| 2440 | // Fast path taken; set region slot 2 | 
|---|
| 2441 | Node *fast_taken = _gvn.transform( new IfFalseNode(opt_iff) ); | 
|---|
| 2442 | region->init_req(2,fast_taken); // Capture fast-control | 
|---|
| 2443 |  | 
|---|
| 2444 | // Fast path not-taken, i.e. slow path | 
|---|
| 2445 | Node *slow_taken = _gvn.transform( new IfTrueNode(opt_iff) ); | 
|---|
| 2446 | return slow_taken; | 
|---|
| 2447 | } | 
|---|
| 2448 |  | 
|---|
| 2449 | //-----------------------------make_runtime_call------------------------------- | 
|---|
| 2450 | Node* GraphKit::make_runtime_call(int flags, | 
|---|
| 2451 | const TypeFunc* call_type, address call_addr, | 
|---|
| 2452 | const char* call_name, | 
|---|
| 2453 | const TypePtr* adr_type, | 
|---|
| 2454 | // The following parms are all optional. | 
|---|
| 2455 | // The first NULL ends the list. | 
|---|
| 2456 | Node* parm0, Node* parm1, | 
|---|
| 2457 | Node* parm2, Node* parm3, | 
|---|
| 2458 | Node* parm4, Node* parm5, | 
|---|
| 2459 | Node* parm6, Node* parm7) { | 
|---|
| 2460 | assert(call_addr != NULL, "must not call NULL targets"); | 
|---|
| 2461 |  | 
|---|
| 2462 | // Slow-path call | 
|---|
| 2463 | bool is_leaf = !(flags & RC_NO_LEAF); | 
|---|
| 2464 | bool has_io  = (!is_leaf && !(flags & RC_NO_IO)); | 
|---|
| 2465 | if (call_name == NULL) { | 
|---|
| 2466 | assert(!is_leaf, "must supply name for leaf"); | 
|---|
| 2467 | call_name = OptoRuntime::stub_name(call_addr); | 
|---|
| 2468 | } | 
|---|
| 2469 | CallNode* call; | 
|---|
| 2470 | if (!is_leaf) { | 
|---|
| 2471 | call = new CallStaticJavaNode(call_type, call_addr, call_name, | 
|---|
| 2472 | bci(), adr_type); | 
|---|
| 2473 | } else if (flags & RC_NO_FP) { | 
|---|
| 2474 | call = new CallLeafNoFPNode(call_type, call_addr, call_name, adr_type); | 
|---|
| 2475 | } else { | 
|---|
| 2476 | call = new CallLeafNode(call_type, call_addr, call_name, adr_type); | 
|---|
| 2477 | } | 
|---|
| 2478 |  | 
|---|
| 2479 | // The following is similar to set_edges_for_java_call, | 
|---|
| 2480 | // except that the memory effects of the call are restricted to AliasIdxRaw. | 
|---|
| 2481 |  | 
|---|
| 2482 | // Slow path call has no side-effects, uses few values | 
|---|
| 2483 | bool wide_in  = !(flags & RC_NARROW_MEM); | 
|---|
| 2484 | bool wide_out = (C->get_alias_index(adr_type) == Compile::AliasIdxBot); | 
|---|
| 2485 |  | 
|---|
| 2486 | Node* prev_mem = NULL; | 
|---|
| 2487 | if (wide_in) { | 
|---|
| 2488 | prev_mem = set_predefined_input_for_runtime_call(call); | 
|---|
| 2489 | } else { | 
|---|
| 2490 | assert(!wide_out, "narrow in => narrow out"); | 
|---|
| 2491 | Node* narrow_mem = memory(adr_type); | 
|---|
| 2492 | prev_mem = set_predefined_input_for_runtime_call(call, narrow_mem); | 
|---|
| 2493 | } | 
|---|
| 2494 |  | 
|---|
| 2495 | // Hook each parm in order.  Stop looking at the first NULL. | 
|---|
| 2496 | if (parm0 != NULL) { call->init_req(TypeFunc::Parms+0, parm0); | 
|---|
| 2497 | if (parm1 != NULL) { call->init_req(TypeFunc::Parms+1, parm1); | 
|---|
| 2498 | if (parm2 != NULL) { call->init_req(TypeFunc::Parms+2, parm2); | 
|---|
| 2499 | if (parm3 != NULL) { call->init_req(TypeFunc::Parms+3, parm3); | 
|---|
| 2500 | if (parm4 != NULL) { call->init_req(TypeFunc::Parms+4, parm4); | 
|---|
| 2501 | if (parm5 != NULL) { call->init_req(TypeFunc::Parms+5, parm5); | 
|---|
| 2502 | if (parm6 != NULL) { call->init_req(TypeFunc::Parms+6, parm6); | 
|---|
| 2503 | if (parm7 != NULL) { call->init_req(TypeFunc::Parms+7, parm7); | 
|---|
| 2504 | /* close each nested if ===> */  } } } } } } } } | 
|---|
| 2505 | assert(call->in(call->req()-1) != NULL, "must initialize all parms"); | 
|---|
| 2506 |  | 
|---|
| 2507 | if (!is_leaf) { | 
|---|
| 2508 | // Non-leaves can block and take safepoints: | 
|---|
| 2509 | add_safepoint_edges(call, ((flags & RC_MUST_THROW) != 0)); | 
|---|
| 2510 | } | 
|---|
| 2511 | // Non-leaves can throw exceptions: | 
|---|
| 2512 | if (has_io) { | 
|---|
| 2513 | call->set_req(TypeFunc::I_O, i_o()); | 
|---|
| 2514 | } | 
|---|
| 2515 |  | 
|---|
| 2516 | if (flags & RC_UNCOMMON) { | 
|---|
| 2517 | // Set the count to a tiny probability.  Cf. Estimate_Block_Frequency. | 
|---|
| 2518 | // (An "if" probability corresponds roughly to an unconditional count. | 
|---|
| 2519 | // Sort of.) | 
|---|
| 2520 | call->set_cnt(PROB_UNLIKELY_MAG(4)); | 
|---|
| 2521 | } | 
|---|
| 2522 |  | 
|---|
| 2523 | Node* c = _gvn.transform(call); | 
|---|
| 2524 | assert(c == call, "cannot disappear"); | 
|---|
| 2525 |  | 
|---|
| 2526 | if (wide_out) { | 
|---|
| 2527 | // Slow path call has full side-effects. | 
|---|
| 2528 | set_predefined_output_for_runtime_call(call); | 
|---|
| 2529 | } else { | 
|---|
| 2530 | // Slow path call has few side-effects, and/or sets few values. | 
|---|
| 2531 | set_predefined_output_for_runtime_call(call, prev_mem, adr_type); | 
|---|
| 2532 | } | 
|---|
| 2533 |  | 
|---|
| 2534 | if (has_io) { | 
|---|
| 2535 | set_i_o(_gvn.transform(new ProjNode(call, TypeFunc::I_O))); | 
|---|
| 2536 | } | 
|---|
| 2537 | return call; | 
|---|
| 2538 |  | 
|---|
| 2539 | } | 
|---|
| 2540 |  | 
|---|
| 2541 | //------------------------------merge_memory----------------------------------- | 
|---|
| 2542 | // Merge memory from one path into the current memory state. | 
|---|
| 2543 | void GraphKit::merge_memory(Node* new_mem, Node* region, int new_path) { | 
|---|
| 2544 | for (MergeMemStream mms(merged_memory(), new_mem->as_MergeMem()); mms.next_non_empty2(); ) { | 
|---|
| 2545 | Node* old_slice = mms.force_memory(); | 
|---|
| 2546 | Node* new_slice = mms.memory2(); | 
|---|
| 2547 | if (old_slice != new_slice) { | 
|---|
| 2548 | PhiNode* phi; | 
|---|
| 2549 | if (old_slice->is_Phi() && old_slice->as_Phi()->region() == region) { | 
|---|
| 2550 | if (mms.is_empty()) { | 
|---|
| 2551 | // clone base memory Phi's inputs for this memory slice | 
|---|
| 2552 | assert(old_slice == mms.base_memory(), "sanity"); | 
|---|
| 2553 | phi = PhiNode::make(region, NULL, Type::MEMORY, mms.adr_type(C)); | 
|---|
| 2554 | _gvn.set_type(phi, Type::MEMORY); | 
|---|
| 2555 | for (uint i = 1; i < phi->req(); i++) { | 
|---|
| 2556 | phi->init_req(i, old_slice->in(i)); | 
|---|
| 2557 | } | 
|---|
| 2558 | } else { | 
|---|
| 2559 | phi = old_slice->as_Phi(); // Phi was generated already | 
|---|
| 2560 | } | 
|---|
| 2561 | } else { | 
|---|
| 2562 | phi = PhiNode::make(region, old_slice, Type::MEMORY, mms.adr_type(C)); | 
|---|
| 2563 | _gvn.set_type(phi, Type::MEMORY); | 
|---|
| 2564 | } | 
|---|
| 2565 | phi->set_req(new_path, new_slice); | 
|---|
| 2566 | mms.set_memory(phi); | 
|---|
| 2567 | } | 
|---|
| 2568 | } | 
|---|
| 2569 | } | 
|---|
| 2570 |  | 
|---|
| 2571 | //------------------------------make_slow_call_ex------------------------------ | 
|---|
| 2572 | // Make the exception handler hookups for the slow call | 
|---|
| 2573 | void GraphKit::make_slow_call_ex(Node* call, ciInstanceKlass* ex_klass, bool separate_io_proj, bool deoptimize) { | 
|---|
| 2574 | if (stopped())  return; | 
|---|
| 2575 |  | 
|---|
| 2576 | // Make a catch node with just two handlers:  fall-through and catch-all | 
|---|
| 2577 | Node* i_o  = _gvn.transform( new ProjNode(call, TypeFunc::I_O, separate_io_proj) ); | 
|---|
| 2578 | Node* catc = _gvn.transform( new CatchNode(control(), i_o, 2) ); | 
|---|
| 2579 | Node* norm = _gvn.transform( new CatchProjNode(catc, CatchProjNode::fall_through_index, CatchProjNode::no_handler_bci) ); | 
|---|
| 2580 | Node* excp = _gvn.transform( new CatchProjNode(catc, CatchProjNode::catch_all_index,    CatchProjNode::no_handler_bci) ); | 
|---|
| 2581 |  | 
|---|
| 2582 | { PreserveJVMState pjvms(this); | 
|---|
| 2583 | set_control(excp); | 
|---|
| 2584 | set_i_o(i_o); | 
|---|
| 2585 |  | 
|---|
| 2586 | if (excp != top()) { | 
|---|
| 2587 | if (deoptimize) { | 
|---|
| 2588 | // Deoptimize if an exception is caught. Don't construct exception state in this case. | 
|---|
| 2589 | uncommon_trap(Deoptimization::Reason_unhandled, | 
|---|
| 2590 | Deoptimization::Action_none); | 
|---|
| 2591 | } else { | 
|---|
| 2592 | // Create an exception state also. | 
|---|
| 2593 | // Use an exact type if the caller has a specific exception. | 
|---|
| 2594 | const Type* ex_type = TypeOopPtr::make_from_klass_unique(ex_klass)->cast_to_ptr_type(TypePtr::NotNull); | 
|---|
| 2595 | Node*       ex_oop  = new CreateExNode(ex_type, control(), i_o); | 
|---|
| 2596 | add_exception_state(make_exception_state(_gvn.transform(ex_oop))); | 
|---|
| 2597 | } | 
|---|
| 2598 | } | 
|---|
| 2599 | } | 
|---|
| 2600 |  | 
|---|
| 2601 | // Get the no-exception control from the CatchNode. | 
|---|
| 2602 | set_control(norm); | 
|---|
| 2603 | } | 
|---|
| 2604 |  | 
|---|
| 2605 | static IfNode* gen_subtype_check_compare(Node* ctrl, Node* in1, Node* in2, BoolTest::mask test, float p, PhaseGVN* gvn, BasicType bt) { | 
|---|
| 2606 | Node* cmp = NULL; | 
|---|
| 2607 | switch(bt) { | 
|---|
| 2608 | case T_INT: cmp = new CmpINode(in1, in2); break; | 
|---|
| 2609 | case T_ADDRESS: cmp = new CmpPNode(in1, in2); break; | 
|---|
| 2610 | default: fatal( "unexpected comparison type %s", type2name(bt)); | 
|---|
| 2611 | } | 
|---|
| 2612 | gvn->transform(cmp); | 
|---|
| 2613 | Node* bol = gvn->transform(new BoolNode(cmp, test)); | 
|---|
| 2614 | IfNode* iff = new IfNode(ctrl, bol, p, COUNT_UNKNOWN); | 
|---|
| 2615 | gvn->transform(iff); | 
|---|
| 2616 | if (!bol->is_Con()) gvn->record_for_igvn(iff); | 
|---|
| 2617 | return iff; | 
|---|
| 2618 | } | 
|---|
| 2619 |  | 
|---|
| 2620 |  | 
|---|
| 2621 | //-------------------------------gen_subtype_check----------------------------- | 
|---|
| 2622 | // Generate a subtyping check.  Takes as input the subtype and supertype. | 
|---|
| 2623 | // Returns 2 values: sets the default control() to the true path and returns | 
|---|
| 2624 | // the false path.  Only reads invariant memory; sets no (visible) memory. | 
|---|
| 2625 | // The PartialSubtypeCheckNode sets the hidden 1-word cache in the encoding | 
|---|
| 2626 | // but that's not exposed to the optimizer.  This call also doesn't take in an | 
|---|
| 2627 | // Object; if you wish to check an Object you need to load the Object's class | 
|---|
| 2628 | // prior to coming here. | 
|---|
| 2629 | Node* Phase::gen_subtype_check(Node* subklass, Node* superklass, Node** ctrl, MergeMemNode* mem, PhaseGVN* gvn) { | 
|---|
| 2630 | Compile* C = gvn->C; | 
|---|
| 2631 |  | 
|---|
| 2632 | if ((*ctrl)->is_top()) { | 
|---|
| 2633 | return C->top(); | 
|---|
| 2634 | } | 
|---|
| 2635 |  | 
|---|
| 2636 | // Fast check for identical types, perhaps identical constants. | 
|---|
| 2637 | // The types can even be identical non-constants, in cases | 
|---|
| 2638 | // involving Array.newInstance, Object.clone, etc. | 
|---|
| 2639 | if (subklass == superklass) | 
|---|
| 2640 | return C->top();             // false path is dead; no test needed. | 
|---|
| 2641 |  | 
|---|
| 2642 | if (gvn->type(superklass)->singleton()) { | 
|---|
| 2643 | ciKlass* superk = gvn->type(superklass)->is_klassptr()->klass(); | 
|---|
| 2644 | ciKlass* subk   = gvn->type(subklass)->is_klassptr()->klass(); | 
|---|
| 2645 |  | 
|---|
| 2646 | // In the common case of an exact superklass, try to fold up the | 
|---|
| 2647 | // test before generating code.  You may ask, why not just generate | 
|---|
| 2648 | // the code and then let it fold up?  The answer is that the generated | 
|---|
| 2649 | // code will necessarily include null checks, which do not always | 
|---|
| 2650 | // completely fold away.  If they are also needless, then they turn | 
|---|
| 2651 | // into a performance loss.  Example: | 
|---|
| 2652 | //    Foo[] fa = blah(); Foo x = fa[0]; fa[1] = x; | 
|---|
| 2653 | // Here, the type of 'fa' is often exact, so the store check | 
|---|
| 2654 | // of fa[1]=x will fold up, without testing the nullness of x. | 
|---|
| 2655 | switch (C->static_subtype_check(superk, subk)) { | 
|---|
| 2656 | case Compile::SSC_always_false: | 
|---|
| 2657 | { | 
|---|
| 2658 | Node* always_fail = *ctrl; | 
|---|
| 2659 | *ctrl = gvn->C->top(); | 
|---|
| 2660 | return always_fail; | 
|---|
| 2661 | } | 
|---|
| 2662 | case Compile::SSC_always_true: | 
|---|
| 2663 | return C->top(); | 
|---|
| 2664 | case Compile::SSC_easy_test: | 
|---|
| 2665 | { | 
|---|
| 2666 | // Just do a direct pointer compare and be done. | 
|---|
| 2667 | IfNode* iff = gen_subtype_check_compare(*ctrl, subklass, superklass, BoolTest::eq, PROB_STATIC_FREQUENT, gvn, T_ADDRESS); | 
|---|
| 2668 | *ctrl = gvn->transform(new IfTrueNode(iff)); | 
|---|
| 2669 | return gvn->transform(new IfFalseNode(iff)); | 
|---|
| 2670 | } | 
|---|
| 2671 | case Compile::SSC_full_test: | 
|---|
| 2672 | break; | 
|---|
| 2673 | default: | 
|---|
| 2674 | ShouldNotReachHere(); | 
|---|
| 2675 | } | 
|---|
| 2676 | } | 
|---|
| 2677 |  | 
|---|
| 2678 | // %%% Possible further optimization:  Even if the superklass is not exact, | 
|---|
| 2679 | // if the subklass is the unique subtype of the superklass, the check | 
|---|
| 2680 | // will always succeed.  We could leave a dependency behind to ensure this. | 
|---|
| 2681 |  | 
|---|
| 2682 | // First load the super-klass's check-offset | 
|---|
| 2683 | Node *p1 = gvn->transform(new AddPNode(superklass, superklass, gvn->MakeConX(in_bytes(Klass::super_check_offset_offset())))); | 
|---|
| 2684 | Node* m = mem->memory_at(C->get_alias_index(gvn->type(p1)->is_ptr())); | 
|---|
| 2685 | Node *chk_off = gvn->transform(new LoadINode(NULL, m, p1, gvn->type(p1)->is_ptr(), TypeInt::INT, MemNode::unordered)); | 
|---|
| 2686 | int cacheoff_con = in_bytes(Klass::secondary_super_cache_offset()); | 
|---|
| 2687 | bool might_be_cache = (gvn->find_int_con(chk_off, cacheoff_con) == cacheoff_con); | 
|---|
| 2688 |  | 
|---|
| 2689 | // Load from the sub-klass's super-class display list, or a 1-word cache of | 
|---|
| 2690 | // the secondary superclass list, or a failing value with a sentinel offset | 
|---|
| 2691 | // if the super-klass is an interface or exceptionally deep in the Java | 
|---|
| 2692 | // hierarchy and we have to scan the secondary superclass list the hard way. | 
|---|
| 2693 | // Worst-case type is a little odd: NULL is allowed as a result (usually | 
|---|
| 2694 | // klass loads can never produce a NULL). | 
|---|
| 2695 | Node *chk_off_X = chk_off; | 
|---|
| 2696 | #ifdef _LP64 | 
|---|
| 2697 | chk_off_X = gvn->transform(new ConvI2LNode(chk_off_X)); | 
|---|
| 2698 | #endif | 
|---|
| 2699 | Node *p2 = gvn->transform(new AddPNode(subklass,subklass,chk_off_X)); | 
|---|
| 2700 | // For some types like interfaces the following loadKlass is from a 1-word | 
|---|
| 2701 | // cache which is mutable so can't use immutable memory.  Other | 
|---|
| 2702 | // types load from the super-class display table which is immutable. | 
|---|
| 2703 | m = mem->memory_at(C->get_alias_index(gvn->type(p2)->is_ptr())); | 
|---|
| 2704 | Node *kmem = might_be_cache ? m : C->immutable_memory(); | 
|---|
| 2705 | Node *nkls = gvn->transform(LoadKlassNode::make(*gvn, NULL, kmem, p2, gvn->type(p2)->is_ptr(), TypeKlassPtr::OBJECT_OR_NULL)); | 
|---|
| 2706 |  | 
|---|
| 2707 | // Compile speed common case: ARE a subtype and we canNOT fail | 
|---|
| 2708 | if( superklass == nkls ) | 
|---|
| 2709 | return C->top();             // false path is dead; no test needed. | 
|---|
| 2710 |  | 
|---|
| 2711 | // See if we get an immediate positive hit.  Happens roughly 83% of the | 
|---|
| 2712 | // time.  Test to see if the value loaded just previously from the subklass | 
|---|
| 2713 | // is exactly the superklass. | 
|---|
| 2714 | IfNode *iff1 = gen_subtype_check_compare(*ctrl, superklass, nkls, BoolTest::eq, PROB_LIKELY(0.83f), gvn, T_ADDRESS); | 
|---|
| 2715 | Node *iftrue1 = gvn->transform( new IfTrueNode (iff1)); | 
|---|
| 2716 | *ctrl = gvn->transform(new IfFalseNode(iff1)); | 
|---|
| 2717 |  | 
|---|
| 2718 | // Compile speed common case: Check for being deterministic right now.  If | 
|---|
| 2719 | // chk_off is a constant and not equal to cacheoff then we are NOT a | 
|---|
| 2720 | // subklass.  In this case we need exactly the 1 test above and we can | 
|---|
| 2721 | // return those results immediately. | 
|---|
| 2722 | if (!might_be_cache) { | 
|---|
| 2723 | Node* not_subtype_ctrl = *ctrl; | 
|---|
| 2724 | *ctrl = iftrue1; // We need exactly the 1 test above | 
|---|
| 2725 | return not_subtype_ctrl; | 
|---|
| 2726 | } | 
|---|
| 2727 |  | 
|---|
| 2728 | // Gather the various success & failures here | 
|---|
| 2729 | RegionNode *r_ok_subtype = new RegionNode(4); | 
|---|
| 2730 | gvn->record_for_igvn(r_ok_subtype); | 
|---|
| 2731 | RegionNode *r_not_subtype = new RegionNode(3); | 
|---|
| 2732 | gvn->record_for_igvn(r_not_subtype); | 
|---|
| 2733 |  | 
|---|
| 2734 | r_ok_subtype->init_req(1, iftrue1); | 
|---|
| 2735 |  | 
|---|
| 2736 | // Check for immediate negative hit.  Happens roughly 11% of the time (which | 
|---|
| 2737 | // is roughly 63% of the remaining cases).  Test to see if the loaded | 
|---|
| 2738 | // check-offset points into the subklass display list or the 1-element | 
|---|
| 2739 | // cache.  If it points to the display (and NOT the cache) and the display | 
|---|
| 2740 | // missed then it's not a subtype. | 
|---|
| 2741 | Node *cacheoff = gvn->intcon(cacheoff_con); | 
|---|
| 2742 | IfNode *iff2 = gen_subtype_check_compare(*ctrl, chk_off, cacheoff, BoolTest::ne, PROB_LIKELY(0.63f), gvn, T_INT); | 
|---|
| 2743 | r_not_subtype->init_req(1, gvn->transform(new IfTrueNode (iff2))); | 
|---|
| 2744 | *ctrl = gvn->transform(new IfFalseNode(iff2)); | 
|---|
| 2745 |  | 
|---|
| 2746 | // Check for self.  Very rare to get here, but it is taken 1/3 the time. | 
|---|
| 2747 | // No performance impact (too rare) but allows sharing of secondary arrays | 
|---|
| 2748 | // which has some footprint reduction. | 
|---|
| 2749 | IfNode *iff3 = gen_subtype_check_compare(*ctrl, subklass, superklass, BoolTest::eq, PROB_LIKELY(0.36f), gvn, T_ADDRESS); | 
|---|
| 2750 | r_ok_subtype->init_req(2, gvn->transform(new IfTrueNode(iff3))); | 
|---|
| 2751 | *ctrl = gvn->transform(new IfFalseNode(iff3)); | 
|---|
| 2752 |  | 
|---|
| 2753 | // -- Roads not taken here: -- | 
|---|
| 2754 | // We could also have chosen to perform the self-check at the beginning | 
|---|
| 2755 | // of this code sequence, as the assembler does.  This would not pay off | 
|---|
| 2756 | // the same way, since the optimizer, unlike the assembler, can perform | 
|---|
| 2757 | // static type analysis to fold away many successful self-checks. | 
|---|
| 2758 | // Non-foldable self checks work better here in second position, because | 
|---|
| 2759 | // the initial primary superclass check subsumes a self-check for most | 
|---|
| 2760 | // types.  An exception would be a secondary type like array-of-interface, | 
|---|
| 2761 | // which does not appear in its own primary supertype display. | 
|---|
| 2762 | // Finally, we could have chosen to move the self-check into the | 
|---|
| 2763 | // PartialSubtypeCheckNode, and from there out-of-line in a platform | 
|---|
| 2764 | // dependent manner.  But it is worthwhile to have the check here, | 
|---|
| 2765 | // where it can be perhaps be optimized.  The cost in code space is | 
|---|
| 2766 | // small (register compare, branch). | 
|---|
| 2767 |  | 
|---|
| 2768 | // Now do a linear scan of the secondary super-klass array.  Again, no real | 
|---|
| 2769 | // performance impact (too rare) but it's gotta be done. | 
|---|
| 2770 | // Since the code is rarely used, there is no penalty for moving it | 
|---|
| 2771 | // out of line, and it can only improve I-cache density. | 
|---|
| 2772 | // The decision to inline or out-of-line this final check is platform | 
|---|
| 2773 | // dependent, and is found in the AD file definition of PartialSubtypeCheck. | 
|---|
| 2774 | Node* psc = gvn->transform( | 
|---|
| 2775 | new PartialSubtypeCheckNode(*ctrl, subklass, superklass)); | 
|---|
| 2776 |  | 
|---|
| 2777 | IfNode *iff4 = gen_subtype_check_compare(*ctrl, psc, gvn->zerocon(T_OBJECT), BoolTest::ne, PROB_FAIR, gvn, T_ADDRESS); | 
|---|
| 2778 | r_not_subtype->init_req(2, gvn->transform(new IfTrueNode (iff4))); | 
|---|
| 2779 | r_ok_subtype ->init_req(3, gvn->transform(new IfFalseNode(iff4))); | 
|---|
| 2780 |  | 
|---|
| 2781 | // Return false path; set default control to true path. | 
|---|
| 2782 | *ctrl = gvn->transform(r_ok_subtype); | 
|---|
| 2783 | return gvn->transform(r_not_subtype); | 
|---|
| 2784 | } | 
|---|
| 2785 |  | 
|---|
| 2786 | // Profile-driven exact type check: | 
|---|
| 2787 | Node* GraphKit::type_check_receiver(Node* receiver, ciKlass* klass, | 
|---|
| 2788 | float prob, | 
|---|
| 2789 | Node* *casted_receiver) { | 
|---|
| 2790 | const TypeKlassPtr* tklass = TypeKlassPtr::make(klass); | 
|---|
| 2791 | Node* recv_klass = load_object_klass(receiver); | 
|---|
| 2792 | Node* want_klass = makecon(tklass); | 
|---|
| 2793 | Node* cmp = _gvn.transform( new CmpPNode(recv_klass, want_klass) ); | 
|---|
| 2794 | Node* bol = _gvn.transform( new BoolNode(cmp, BoolTest::eq) ); | 
|---|
| 2795 | IfNode* iff = create_and_xform_if(control(), bol, prob, COUNT_UNKNOWN); | 
|---|
| 2796 | set_control( _gvn.transform( new IfTrueNode (iff) )); | 
|---|
| 2797 | Node* fail = _gvn.transform( new IfFalseNode(iff) ); | 
|---|
| 2798 |  | 
|---|
| 2799 | const TypeOopPtr* recv_xtype = tklass->as_instance_type(); | 
|---|
| 2800 | assert(recv_xtype->klass_is_exact(), ""); | 
|---|
| 2801 |  | 
|---|
| 2802 | // Subsume downstream occurrences of receiver with a cast to | 
|---|
| 2803 | // recv_xtype, since now we know what the type will be. | 
|---|
| 2804 | Node* cast = new CheckCastPPNode(control(), receiver, recv_xtype); | 
|---|
| 2805 | (*casted_receiver) = _gvn.transform(cast); | 
|---|
| 2806 | // (User must make the replace_in_map call.) | 
|---|
| 2807 |  | 
|---|
| 2808 | return fail; | 
|---|
| 2809 | } | 
|---|
| 2810 |  | 
|---|
| 2811 | //------------------------------subtype_check_receiver------------------------- | 
|---|
| 2812 | Node* GraphKit::subtype_check_receiver(Node* receiver, ciKlass* klass, | 
|---|
| 2813 | Node** casted_receiver) { | 
|---|
| 2814 | const TypeKlassPtr* tklass = TypeKlassPtr::make(klass); | 
|---|
| 2815 | Node* recv_klass = load_object_klass(receiver); | 
|---|
| 2816 | Node* want_klass = makecon(tklass); | 
|---|
| 2817 |  | 
|---|
| 2818 | Node* slow_ctl = gen_subtype_check(recv_klass, want_klass); | 
|---|
| 2819 |  | 
|---|
| 2820 | // Cast receiver after successful check | 
|---|
| 2821 | const TypeOopPtr* recv_type = tklass->cast_to_exactness(false)->is_klassptr()->as_instance_type(); | 
|---|
| 2822 | Node* cast = new CheckCastPPNode(control(), receiver, recv_type); | 
|---|
| 2823 | (*casted_receiver) = _gvn.transform(cast); | 
|---|
| 2824 |  | 
|---|
| 2825 | return slow_ctl; | 
|---|
| 2826 | } | 
|---|
| 2827 |  | 
|---|
| 2828 | //------------------------------seems_never_null------------------------------- | 
|---|
| 2829 | // Use null_seen information if it is available from the profile. | 
|---|
| 2830 | // If we see an unexpected null at a type check we record it and force a | 
|---|
| 2831 | // recompile; the offending check will be recompiled to handle NULLs. | 
|---|
| 2832 | // If we see several offending BCIs, then all checks in the | 
|---|
| 2833 | // method will be recompiled. | 
|---|
| 2834 | bool GraphKit::seems_never_null(Node* obj, ciProfileData* data, bool& speculating) { | 
|---|
| 2835 | speculating = !_gvn.type(obj)->speculative_maybe_null(); | 
|---|
| 2836 | Deoptimization::DeoptReason reason = Deoptimization::reason_null_check(speculating); | 
|---|
| 2837 | if (UncommonNullCast               // Cutout for this technique | 
|---|
| 2838 | && obj != null()               // And not the -Xcomp stupid case? | 
|---|
| 2839 | && !too_many_traps(reason) | 
|---|
| 2840 | ) { | 
|---|
| 2841 | if (speculating) { | 
|---|
| 2842 | return true; | 
|---|
| 2843 | } | 
|---|
| 2844 | if (data == NULL) | 
|---|
| 2845 | // Edge case:  no mature data.  Be optimistic here. | 
|---|
| 2846 | return true; | 
|---|
| 2847 | // If the profile has not seen a null, assume it won't happen. | 
|---|
| 2848 | assert(java_bc() == Bytecodes::_checkcast || | 
|---|
| 2849 | java_bc() == Bytecodes::_instanceof || | 
|---|
| 2850 | java_bc() == Bytecodes::_aastore, "MDO must collect null_seen bit here"); | 
|---|
| 2851 | return !data->as_BitData()->null_seen(); | 
|---|
| 2852 | } | 
|---|
| 2853 | speculating = false; | 
|---|
| 2854 | return false; | 
|---|
| 2855 | } | 
|---|
| 2856 |  | 
|---|
| 2857 | void GraphKit::guard_klass_being_initialized(Node* klass) { | 
|---|
| 2858 | int init_state_off = in_bytes(InstanceKlass::init_state_offset()); | 
|---|
| 2859 | Node* adr = basic_plus_adr(top(), klass, init_state_off); | 
|---|
| 2860 | Node* init_state = LoadNode::make(_gvn, NULL, immutable_memory(), adr, | 
|---|
| 2861 | adr->bottom_type()->is_ptr(), TypeInt::BYTE, | 
|---|
| 2862 | T_BYTE, MemNode::unordered); | 
|---|
| 2863 | init_state = _gvn.transform(init_state); | 
|---|
| 2864 |  | 
|---|
| 2865 | Node* being_initialized_state = makecon(TypeInt::make(InstanceKlass::being_initialized)); | 
|---|
| 2866 |  | 
|---|
| 2867 | Node* chk = _gvn.transform(new CmpINode(being_initialized_state, init_state)); | 
|---|
| 2868 | Node* tst = _gvn.transform(new BoolNode(chk, BoolTest::eq)); | 
|---|
| 2869 |  | 
|---|
| 2870 | { BuildCutout unless(this, tst, PROB_MAX); | 
|---|
| 2871 | uncommon_trap(Deoptimization::Reason_initialized, Deoptimization::Action_reinterpret); | 
|---|
| 2872 | } | 
|---|
| 2873 | } | 
|---|
| 2874 |  | 
|---|
| 2875 | void GraphKit::guard_init_thread(Node* klass) { | 
|---|
| 2876 | int init_thread_off = in_bytes(InstanceKlass::init_thread_offset()); | 
|---|
| 2877 | Node* adr = basic_plus_adr(top(), klass, init_thread_off); | 
|---|
| 2878 |  | 
|---|
| 2879 | Node* init_thread = LoadNode::make(_gvn, NULL, immutable_memory(), adr, | 
|---|
| 2880 | adr->bottom_type()->is_ptr(), TypePtr::NOTNULL, | 
|---|
| 2881 | T_ADDRESS, MemNode::unordered); | 
|---|
| 2882 | init_thread = _gvn.transform(init_thread); | 
|---|
| 2883 |  | 
|---|
| 2884 | Node* cur_thread = _gvn.transform(new ThreadLocalNode()); | 
|---|
| 2885 |  | 
|---|
| 2886 | Node* chk = _gvn.transform(new CmpPNode(cur_thread, init_thread)); | 
|---|
| 2887 | Node* tst = _gvn.transform(new BoolNode(chk, BoolTest::eq)); | 
|---|
| 2888 |  | 
|---|
| 2889 | { BuildCutout unless(this, tst, PROB_MAX); | 
|---|
| 2890 | uncommon_trap(Deoptimization::Reason_uninitialized, Deoptimization::Action_none); | 
|---|
| 2891 | } | 
|---|
| 2892 | } | 
|---|
| 2893 |  | 
|---|
| 2894 | void GraphKit::clinit_barrier(ciInstanceKlass* ik, ciMethod* context) { | 
|---|
| 2895 | if (ik->is_being_initialized()) { | 
|---|
| 2896 | if (C->needs_clinit_barrier(ik, context)) { | 
|---|
| 2897 | Node* klass = makecon(TypeKlassPtr::make(ik)); | 
|---|
| 2898 | guard_klass_being_initialized(klass); | 
|---|
| 2899 | guard_init_thread(klass); | 
|---|
| 2900 | insert_mem_bar(Op_MemBarCPUOrder); | 
|---|
| 2901 | } | 
|---|
| 2902 | } else if (ik->is_initialized()) { | 
|---|
| 2903 | return; // no barrier needed | 
|---|
| 2904 | } else { | 
|---|
| 2905 | uncommon_trap(Deoptimization::Reason_uninitialized, | 
|---|
| 2906 | Deoptimization::Action_reinterpret, | 
|---|
| 2907 | NULL); | 
|---|
| 2908 | } | 
|---|
| 2909 | } | 
|---|
| 2910 |  | 
|---|
| 2911 | //------------------------maybe_cast_profiled_receiver------------------------- | 
|---|
| 2912 | // If the profile has seen exactly one type, narrow to exactly that type. | 
|---|
| 2913 | // Subsequent type checks will always fold up. | 
|---|
| 2914 | Node* GraphKit::maybe_cast_profiled_receiver(Node* not_null_obj, | 
|---|
| 2915 | ciKlass* require_klass, | 
|---|
| 2916 | ciKlass* spec_klass, | 
|---|
| 2917 | bool safe_for_replace) { | 
|---|
| 2918 | if (!UseTypeProfile || !TypeProfileCasts) return NULL; | 
|---|
| 2919 |  | 
|---|
| 2920 | Deoptimization::DeoptReason reason = Deoptimization::reason_class_check(spec_klass != NULL); | 
|---|
| 2921 |  | 
|---|
| 2922 | // Make sure we haven't already deoptimized from this tactic. | 
|---|
| 2923 | if (too_many_traps_or_recompiles(reason)) | 
|---|
| 2924 | return NULL; | 
|---|
| 2925 |  | 
|---|
| 2926 | // (No, this isn't a call, but it's enough like a virtual call | 
|---|
| 2927 | // to use the same ciMethod accessor to get the profile info...) | 
|---|
| 2928 | // If we have a speculative type use it instead of profiling (which | 
|---|
| 2929 | // may not help us) | 
|---|
| 2930 | ciKlass* exact_kls = spec_klass == NULL ? profile_has_unique_klass() : spec_klass; | 
|---|
| 2931 | if (exact_kls != NULL) {// no cast failures here | 
|---|
| 2932 | if (require_klass == NULL || | 
|---|
| 2933 | C->static_subtype_check(require_klass, exact_kls) == Compile::SSC_always_true) { | 
|---|
| 2934 | // If we narrow the type to match what the type profile sees or | 
|---|
| 2935 | // the speculative type, we can then remove the rest of the | 
|---|
| 2936 | // cast. | 
|---|
| 2937 | // This is a win, even if the exact_kls is very specific, | 
|---|
| 2938 | // because downstream operations, such as method calls, | 
|---|
| 2939 | // will often benefit from the sharper type. | 
|---|
| 2940 | Node* exact_obj = not_null_obj; // will get updated in place... | 
|---|
| 2941 | Node* slow_ctl  = type_check_receiver(exact_obj, exact_kls, 1.0, | 
|---|
| 2942 | &exact_obj); | 
|---|
| 2943 | { PreserveJVMState pjvms(this); | 
|---|
| 2944 | set_control(slow_ctl); | 
|---|
| 2945 | uncommon_trap_exact(reason, Deoptimization::Action_maybe_recompile); | 
|---|
| 2946 | } | 
|---|
| 2947 | if (safe_for_replace) { | 
|---|
| 2948 | replace_in_map(not_null_obj, exact_obj); | 
|---|
| 2949 | } | 
|---|
| 2950 | return exact_obj; | 
|---|
| 2951 | } | 
|---|
| 2952 | // assert(ssc == Compile::SSC_always_true)... except maybe the profile lied to us. | 
|---|
| 2953 | } | 
|---|
| 2954 |  | 
|---|
| 2955 | return NULL; | 
|---|
| 2956 | } | 
|---|
| 2957 |  | 
|---|
| 2958 | /** | 
|---|
| 2959 | * Cast obj to type and emit guard unless we had too many traps here | 
|---|
| 2960 | * already | 
|---|
| 2961 | * | 
|---|
| 2962 | * @param obj       node being casted | 
|---|
| 2963 | * @param type      type to cast the node to | 
|---|
| 2964 | * @param not_null  true if we know node cannot be null | 
|---|
| 2965 | */ | 
|---|
| 2966 | Node* GraphKit::maybe_cast_profiled_obj(Node* obj, | 
|---|
| 2967 | ciKlass* type, | 
|---|
| 2968 | bool not_null) { | 
|---|
| 2969 | if (stopped()) { | 
|---|
| 2970 | return obj; | 
|---|
| 2971 | } | 
|---|
| 2972 |  | 
|---|
| 2973 | // type == NULL if profiling tells us this object is always null | 
|---|
| 2974 | if (type != NULL) { | 
|---|
| 2975 | Deoptimization::DeoptReason class_reason = Deoptimization::Reason_speculate_class_check; | 
|---|
| 2976 | Deoptimization::DeoptReason null_reason = Deoptimization::Reason_speculate_null_check; | 
|---|
| 2977 |  | 
|---|
| 2978 | if (!too_many_traps_or_recompiles(null_reason) && | 
|---|
| 2979 | !too_many_traps_or_recompiles(class_reason)) { | 
|---|
| 2980 | Node* not_null_obj = NULL; | 
|---|
| 2981 | // not_null is true if we know the object is not null and | 
|---|
| 2982 | // there's no need for a null check | 
|---|
| 2983 | if (!not_null) { | 
|---|
| 2984 | Node* null_ctl = top(); | 
|---|
| 2985 | not_null_obj = null_check_oop(obj, &null_ctl, true, true, true); | 
|---|
| 2986 | assert(null_ctl->is_top(), "no null control here"); | 
|---|
| 2987 | } else { | 
|---|
| 2988 | not_null_obj = obj; | 
|---|
| 2989 | } | 
|---|
| 2990 |  | 
|---|
| 2991 | Node* exact_obj = not_null_obj; | 
|---|
| 2992 | ciKlass* exact_kls = type; | 
|---|
| 2993 | Node* slow_ctl  = type_check_receiver(exact_obj, exact_kls, 1.0, | 
|---|
| 2994 | &exact_obj); | 
|---|
| 2995 | { | 
|---|
| 2996 | PreserveJVMState pjvms(this); | 
|---|
| 2997 | set_control(slow_ctl); | 
|---|
| 2998 | uncommon_trap_exact(class_reason, Deoptimization::Action_maybe_recompile); | 
|---|
| 2999 | } | 
|---|
| 3000 | replace_in_map(not_null_obj, exact_obj); | 
|---|
| 3001 | obj = exact_obj; | 
|---|
| 3002 | } | 
|---|
| 3003 | } else { | 
|---|
| 3004 | if (!too_many_traps_or_recompiles(Deoptimization::Reason_null_assert)) { | 
|---|
| 3005 | Node* exact_obj = null_assert(obj); | 
|---|
| 3006 | replace_in_map(obj, exact_obj); | 
|---|
| 3007 | obj = exact_obj; | 
|---|
| 3008 | } | 
|---|
| 3009 | } | 
|---|
| 3010 | return obj; | 
|---|
| 3011 | } | 
|---|
| 3012 |  | 
|---|
| 3013 | //-------------------------------gen_instanceof-------------------------------- | 
|---|
| 3014 | // Generate an instance-of idiom.  Used by both the instance-of bytecode | 
|---|
| 3015 | // and the reflective instance-of call. | 
|---|
| 3016 | Node* GraphKit::gen_instanceof(Node* obj, Node* superklass, bool safe_for_replace) { | 
|---|
| 3017 | kill_dead_locals();           // Benefit all the uncommon traps | 
|---|
| 3018 | assert( !stopped(), "dead parse path should be checked in callers"); | 
|---|
| 3019 | assert(!TypePtr::NULL_PTR->higher_equal(_gvn.type(superklass)->is_klassptr()), | 
|---|
| 3020 | "must check for not-null not-dead klass in callers"); | 
|---|
| 3021 |  | 
|---|
| 3022 | // Make the merge point | 
|---|
| 3023 | enum { _obj_path = 1, _fail_path, _null_path, PATH_LIMIT }; | 
|---|
| 3024 | RegionNode* region = new RegionNode(PATH_LIMIT); | 
|---|
| 3025 | Node*       phi    = new PhiNode(region, TypeInt::BOOL); | 
|---|
| 3026 | C->set_has_split_ifs(true); // Has chance for split-if optimization | 
|---|
| 3027 |  | 
|---|
| 3028 | ciProfileData* data = NULL; | 
|---|
| 3029 | if (java_bc() == Bytecodes::_instanceof) {  // Only for the bytecode | 
|---|
| 3030 | data = method()->method_data()->bci_to_data(bci()); | 
|---|
| 3031 | } | 
|---|
| 3032 | bool speculative_not_null = false; | 
|---|
| 3033 | bool never_see_null = (ProfileDynamicTypes  // aggressive use of profile | 
|---|
| 3034 | && seems_never_null(obj, data, speculative_not_null)); | 
|---|
| 3035 |  | 
|---|
| 3036 | // Null check; get casted pointer; set region slot 3 | 
|---|
| 3037 | Node* null_ctl = top(); | 
|---|
| 3038 | Node* not_null_obj = null_check_oop(obj, &null_ctl, never_see_null, safe_for_replace, speculative_not_null); | 
|---|
| 3039 |  | 
|---|
| 3040 | // If not_null_obj is dead, only null-path is taken | 
|---|
| 3041 | if (stopped()) {              // Doing instance-of on a NULL? | 
|---|
| 3042 | set_control(null_ctl); | 
|---|
| 3043 | return intcon(0); | 
|---|
| 3044 | } | 
|---|
| 3045 | region->init_req(_null_path, null_ctl); | 
|---|
| 3046 | phi   ->init_req(_null_path, intcon(0)); // Set null path value | 
|---|
| 3047 | if (null_ctl == top()) { | 
|---|
| 3048 | // Do this eagerly, so that pattern matches like is_diamond_phi | 
|---|
| 3049 | // will work even during parsing. | 
|---|
| 3050 | assert(_null_path == PATH_LIMIT-1, "delete last"); | 
|---|
| 3051 | region->del_req(_null_path); | 
|---|
| 3052 | phi   ->del_req(_null_path); | 
|---|
| 3053 | } | 
|---|
| 3054 |  | 
|---|
| 3055 | // Do we know the type check always succeed? | 
|---|
| 3056 | bool known_statically = false; | 
|---|
| 3057 | if (_gvn.type(superklass)->singleton()) { | 
|---|
| 3058 | ciKlass* superk = _gvn.type(superklass)->is_klassptr()->klass(); | 
|---|
| 3059 | ciKlass* subk = _gvn.type(obj)->is_oopptr()->klass(); | 
|---|
| 3060 | if (subk != NULL && subk->is_loaded()) { | 
|---|
| 3061 | int static_res = C->static_subtype_check(superk, subk); | 
|---|
| 3062 | known_statically = (static_res == Compile::SSC_always_true || static_res == Compile::SSC_always_false); | 
|---|
| 3063 | } | 
|---|
| 3064 | } | 
|---|
| 3065 |  | 
|---|
| 3066 | if (!known_statically) { | 
|---|
| 3067 | const TypeOopPtr* obj_type = _gvn.type(obj)->is_oopptr(); | 
|---|
| 3068 | // We may not have profiling here or it may not help us. If we | 
|---|
| 3069 | // have a speculative type use it to perform an exact cast. | 
|---|
| 3070 | ciKlass* spec_obj_type = obj_type->speculative_type(); | 
|---|
| 3071 | if (spec_obj_type != NULL || (ProfileDynamicTypes && data != NULL)) { | 
|---|
| 3072 | Node* cast_obj = maybe_cast_profiled_receiver(not_null_obj, NULL, spec_obj_type, safe_for_replace); | 
|---|
| 3073 | if (stopped()) {            // Profile disagrees with this path. | 
|---|
| 3074 | set_control(null_ctl);    // Null is the only remaining possibility. | 
|---|
| 3075 | return intcon(0); | 
|---|
| 3076 | } | 
|---|
| 3077 | if (cast_obj != NULL) { | 
|---|
| 3078 | not_null_obj = cast_obj; | 
|---|
| 3079 | } | 
|---|
| 3080 | } | 
|---|
| 3081 | } | 
|---|
| 3082 |  | 
|---|
| 3083 | // Load the object's klass | 
|---|
| 3084 | Node* obj_klass = load_object_klass(not_null_obj); | 
|---|
| 3085 |  | 
|---|
| 3086 | // Generate the subtype check | 
|---|
| 3087 | Node* not_subtype_ctrl = gen_subtype_check(obj_klass, superklass); | 
|---|
| 3088 |  | 
|---|
| 3089 | // Plug in the success path to the general merge in slot 1. | 
|---|
| 3090 | region->init_req(_obj_path, control()); | 
|---|
| 3091 | phi   ->init_req(_obj_path, intcon(1)); | 
|---|
| 3092 |  | 
|---|
| 3093 | // Plug in the failing path to the general merge in slot 2. | 
|---|
| 3094 | region->init_req(_fail_path, not_subtype_ctrl); | 
|---|
| 3095 | phi   ->init_req(_fail_path, intcon(0)); | 
|---|
| 3096 |  | 
|---|
| 3097 | // Return final merged results | 
|---|
| 3098 | set_control( _gvn.transform(region) ); | 
|---|
| 3099 | record_for_igvn(region); | 
|---|
| 3100 |  | 
|---|
| 3101 | // If we know the type check always succeeds then we don't use the | 
|---|
| 3102 | // profiling data at this bytecode. Don't lose it, feed it to the | 
|---|
| 3103 | // type system as a speculative type. | 
|---|
| 3104 | if (safe_for_replace) { | 
|---|
| 3105 | Node* casted_obj = record_profiled_receiver_for_speculation(obj); | 
|---|
| 3106 | replace_in_map(obj, casted_obj); | 
|---|
| 3107 | } | 
|---|
| 3108 |  | 
|---|
| 3109 | return _gvn.transform(phi); | 
|---|
| 3110 | } | 
|---|
| 3111 |  | 
|---|
| 3112 | //-------------------------------gen_checkcast--------------------------------- | 
|---|
| 3113 | // Generate a checkcast idiom.  Used by both the checkcast bytecode and the | 
|---|
| 3114 | // array store bytecode.  Stack must be as-if BEFORE doing the bytecode so the | 
|---|
| 3115 | // uncommon-trap paths work.  Adjust stack after this call. | 
|---|
| 3116 | // If failure_control is supplied and not null, it is filled in with | 
|---|
| 3117 | // the control edge for the cast failure.  Otherwise, an appropriate | 
|---|
| 3118 | // uncommon trap or exception is thrown. | 
|---|
| 3119 | Node* GraphKit::gen_checkcast(Node *obj, Node* superklass, | 
|---|
| 3120 | Node* *failure_control) { | 
|---|
| 3121 | kill_dead_locals();           // Benefit all the uncommon traps | 
|---|
| 3122 | const TypeKlassPtr *tk = _gvn.type(superklass)->is_klassptr(); | 
|---|
| 3123 | const Type *toop = TypeOopPtr::make_from_klass(tk->klass()); | 
|---|
| 3124 |  | 
|---|
| 3125 | // Fast cutout:  Check the case that the cast is vacuously true. | 
|---|
| 3126 | // This detects the common cases where the test will short-circuit | 
|---|
| 3127 | // away completely.  We do this before we perform the null check, | 
|---|
| 3128 | // because if the test is going to turn into zero code, we don't | 
|---|
| 3129 | // want a residual null check left around.  (Causes a slowdown, | 
|---|
| 3130 | // for example, in some objArray manipulations, such as a[i]=a[j].) | 
|---|
| 3131 | if (tk->singleton()) { | 
|---|
| 3132 | const TypeOopPtr* objtp = _gvn.type(obj)->isa_oopptr(); | 
|---|
| 3133 | if (objtp != NULL && objtp->klass() != NULL) { | 
|---|
| 3134 | switch (C->static_subtype_check(tk->klass(), objtp->klass())) { | 
|---|
| 3135 | case Compile::SSC_always_true: | 
|---|
| 3136 | // If we know the type check always succeed then we don't use | 
|---|
| 3137 | // the profiling data at this bytecode. Don't lose it, feed it | 
|---|
| 3138 | // to the type system as a speculative type. | 
|---|
| 3139 | return record_profiled_receiver_for_speculation(obj); | 
|---|
| 3140 | case Compile::SSC_always_false: | 
|---|
| 3141 | // It needs a null check because a null will *pass* the cast check. | 
|---|
| 3142 | // A non-null value will always produce an exception. | 
|---|
| 3143 | return null_assert(obj); | 
|---|
| 3144 | } | 
|---|
| 3145 | } | 
|---|
| 3146 | } | 
|---|
| 3147 |  | 
|---|
| 3148 | ciProfileData* data = NULL; | 
|---|
| 3149 | bool safe_for_replace = false; | 
|---|
| 3150 | if (failure_control == NULL) {        // use MDO in regular case only | 
|---|
| 3151 | assert(java_bc() == Bytecodes::_aastore || | 
|---|
| 3152 | java_bc() == Bytecodes::_checkcast, | 
|---|
| 3153 | "interpreter profiles type checks only for these BCs"); | 
|---|
| 3154 | data = method()->method_data()->bci_to_data(bci()); | 
|---|
| 3155 | safe_for_replace = true; | 
|---|
| 3156 | } | 
|---|
| 3157 |  | 
|---|
| 3158 | // Make the merge point | 
|---|
| 3159 | enum { _obj_path = 1, _null_path, PATH_LIMIT }; | 
|---|
| 3160 | RegionNode* region = new RegionNode(PATH_LIMIT); | 
|---|
| 3161 | Node*       phi    = new PhiNode(region, toop); | 
|---|
| 3162 | C->set_has_split_ifs(true); // Has chance for split-if optimization | 
|---|
| 3163 |  | 
|---|
| 3164 | // Use null-cast information if it is available | 
|---|
| 3165 | bool speculative_not_null = false; | 
|---|
| 3166 | bool never_see_null = ((failure_control == NULL)  // regular case only | 
|---|
| 3167 | && seems_never_null(obj, data, speculative_not_null)); | 
|---|
| 3168 |  | 
|---|
| 3169 | // Null check; get casted pointer; set region slot 3 | 
|---|
| 3170 | Node* null_ctl = top(); | 
|---|
| 3171 | Node* not_null_obj = null_check_oop(obj, &null_ctl, never_see_null, safe_for_replace, speculative_not_null); | 
|---|
| 3172 |  | 
|---|
| 3173 | // If not_null_obj is dead, only null-path is taken | 
|---|
| 3174 | if (stopped()) {              // Doing instance-of on a NULL? | 
|---|
| 3175 | set_control(null_ctl); | 
|---|
| 3176 | return null(); | 
|---|
| 3177 | } | 
|---|
| 3178 | region->init_req(_null_path, null_ctl); | 
|---|
| 3179 | phi   ->init_req(_null_path, null());  // Set null path value | 
|---|
| 3180 | if (null_ctl == top()) { | 
|---|
| 3181 | // Do this eagerly, so that pattern matches like is_diamond_phi | 
|---|
| 3182 | // will work even during parsing. | 
|---|
| 3183 | assert(_null_path == PATH_LIMIT-1, "delete last"); | 
|---|
| 3184 | region->del_req(_null_path); | 
|---|
| 3185 | phi   ->del_req(_null_path); | 
|---|
| 3186 | } | 
|---|
| 3187 |  | 
|---|
| 3188 | Node* cast_obj = NULL; | 
|---|
| 3189 | if (tk->klass_is_exact()) { | 
|---|
| 3190 | // The following optimization tries to statically cast the speculative type of the object | 
|---|
| 3191 | // (for example obtained during profiling) to the type of the superklass and then do a | 
|---|
| 3192 | // dynamic check that the type of the object is what we expect. To work correctly | 
|---|
| 3193 | // for checkcast and aastore the type of superklass should be exact. | 
|---|
| 3194 | const TypeOopPtr* obj_type = _gvn.type(obj)->is_oopptr(); | 
|---|
| 3195 | // We may not have profiling here or it may not help us. If we have | 
|---|
| 3196 | // a speculative type use it to perform an exact cast. | 
|---|
| 3197 | ciKlass* spec_obj_type = obj_type->speculative_type(); | 
|---|
| 3198 | if (spec_obj_type != NULL || data != NULL) { | 
|---|
| 3199 | cast_obj = maybe_cast_profiled_receiver(not_null_obj, tk->klass(), spec_obj_type, safe_for_replace); | 
|---|
| 3200 | if (cast_obj != NULL) { | 
|---|
| 3201 | if (failure_control != NULL) // failure is now impossible | 
|---|
| 3202 | (*failure_control) = top(); | 
|---|
| 3203 | // adjust the type of the phi to the exact klass: | 
|---|
| 3204 | phi->raise_bottom_type(_gvn.type(cast_obj)->meet_speculative(TypePtr::NULL_PTR)); | 
|---|
| 3205 | } | 
|---|
| 3206 | } | 
|---|
| 3207 | } | 
|---|
| 3208 |  | 
|---|
| 3209 | if (cast_obj == NULL) { | 
|---|
| 3210 | // Load the object's klass | 
|---|
| 3211 | Node* obj_klass = load_object_klass(not_null_obj); | 
|---|
| 3212 |  | 
|---|
| 3213 | // Generate the subtype check | 
|---|
| 3214 | Node* not_subtype_ctrl = gen_subtype_check( obj_klass, superklass ); | 
|---|
| 3215 |  | 
|---|
| 3216 | // Plug in success path into the merge | 
|---|
| 3217 | cast_obj = _gvn.transform(new CheckCastPPNode(control(), not_null_obj, toop)); | 
|---|
| 3218 | // Failure path ends in uncommon trap (or may be dead - failure impossible) | 
|---|
| 3219 | if (failure_control == NULL) { | 
|---|
| 3220 | if (not_subtype_ctrl != top()) { // If failure is possible | 
|---|
| 3221 | PreserveJVMState pjvms(this); | 
|---|
| 3222 | set_control(not_subtype_ctrl); | 
|---|
| 3223 | builtin_throw(Deoptimization::Reason_class_check, obj_klass); | 
|---|
| 3224 | } | 
|---|
| 3225 | } else { | 
|---|
| 3226 | (*failure_control) = not_subtype_ctrl; | 
|---|
| 3227 | } | 
|---|
| 3228 | } | 
|---|
| 3229 |  | 
|---|
| 3230 | region->init_req(_obj_path, control()); | 
|---|
| 3231 | phi   ->init_req(_obj_path, cast_obj); | 
|---|
| 3232 |  | 
|---|
| 3233 | // A merge of NULL or Casted-NotNull obj | 
|---|
| 3234 | Node* res = _gvn.transform(phi); | 
|---|
| 3235 |  | 
|---|
| 3236 | // Note I do NOT always 'replace_in_map(obj,result)' here. | 
|---|
| 3237 | //  if( tk->klass()->can_be_primary_super()  ) | 
|---|
| 3238 | // This means that if I successfully store an Object into an array-of-String | 
|---|
| 3239 | // I 'forget' that the Object is really now known to be a String.  I have to | 
|---|
| 3240 | // do this because we don't have true union types for interfaces - if I store | 
|---|
| 3241 | // a Baz into an array-of-Interface and then tell the optimizer it's an | 
|---|
| 3242 | // Interface, I forget that it's also a Baz and cannot do Baz-like field | 
|---|
| 3243 | // references to it.  FIX THIS WHEN UNION TYPES APPEAR! | 
|---|
| 3244 | //  replace_in_map( obj, res ); | 
|---|
| 3245 |  | 
|---|
| 3246 | // Return final merged results | 
|---|
| 3247 | set_control( _gvn.transform(region) ); | 
|---|
| 3248 | record_for_igvn(region); | 
|---|
| 3249 |  | 
|---|
| 3250 | return record_profiled_receiver_for_speculation(res); | 
|---|
| 3251 | } | 
|---|
| 3252 |  | 
|---|
| 3253 | //------------------------------next_monitor----------------------------------- | 
|---|
| 3254 | // What number should be given to the next monitor? | 
|---|
| 3255 | int GraphKit::next_monitor() { | 
|---|
| 3256 | int current = jvms()->monitor_depth()* C->sync_stack_slots(); | 
|---|
| 3257 | int next = current + C->sync_stack_slots(); | 
|---|
| 3258 | // Keep the toplevel high water mark current: | 
|---|
| 3259 | if (C->fixed_slots() < next)  C->set_fixed_slots(next); | 
|---|
| 3260 | return current; | 
|---|
| 3261 | } | 
|---|
| 3262 |  | 
|---|
| 3263 | //------------------------------insert_mem_bar--------------------------------- | 
|---|
| 3264 | // Memory barrier to avoid floating things around | 
|---|
| 3265 | // The membar serves as a pinch point between both control and all memory slices. | 
|---|
| 3266 | Node* GraphKit::insert_mem_bar(int opcode, Node* precedent) { | 
|---|
| 3267 | MemBarNode* mb = MemBarNode::make(C, opcode, Compile::AliasIdxBot, precedent); | 
|---|
| 3268 | mb->init_req(TypeFunc::Control, control()); | 
|---|
| 3269 | mb->init_req(TypeFunc::Memory,  reset_memory()); | 
|---|
| 3270 | Node* membar = _gvn.transform(mb); | 
|---|
| 3271 | set_control(_gvn.transform(new ProjNode(membar, TypeFunc::Control))); | 
|---|
| 3272 | set_all_memory_call(membar); | 
|---|
| 3273 | return membar; | 
|---|
| 3274 | } | 
|---|
| 3275 |  | 
|---|
| 3276 | //-------------------------insert_mem_bar_volatile---------------------------- | 
|---|
| 3277 | // Memory barrier to avoid floating things around | 
|---|
| 3278 | // The membar serves as a pinch point between both control and memory(alias_idx). | 
|---|
| 3279 | // If you want to make a pinch point on all memory slices, do not use this | 
|---|
| 3280 | // function (even with AliasIdxBot); use insert_mem_bar() instead. | 
|---|
| 3281 | Node* GraphKit::insert_mem_bar_volatile(int opcode, int alias_idx, Node* precedent) { | 
|---|
| 3282 | // When Parse::do_put_xxx updates a volatile field, it appends a series | 
|---|
| 3283 | // of MemBarVolatile nodes, one for *each* volatile field alias category. | 
|---|
| 3284 | // The first membar is on the same memory slice as the field store opcode. | 
|---|
| 3285 | // This forces the membar to follow the store.  (Bug 6500685 broke this.) | 
|---|
| 3286 | // All the other membars (for other volatile slices, including AliasIdxBot, | 
|---|
| 3287 | // which stands for all unknown volatile slices) are control-dependent | 
|---|
| 3288 | // on the first membar.  This prevents later volatile loads or stores | 
|---|
| 3289 | // from sliding up past the just-emitted store. | 
|---|
| 3290 |  | 
|---|
| 3291 | MemBarNode* mb = MemBarNode::make(C, opcode, alias_idx, precedent); | 
|---|
| 3292 | mb->set_req(TypeFunc::Control,control()); | 
|---|
| 3293 | if (alias_idx == Compile::AliasIdxBot) { | 
|---|
| 3294 | mb->set_req(TypeFunc::Memory, merged_memory()->base_memory()); | 
|---|
| 3295 | } else { | 
|---|
| 3296 | assert(!(opcode == Op_Initialize && alias_idx != Compile::AliasIdxRaw), "fix caller"); | 
|---|
| 3297 | mb->set_req(TypeFunc::Memory, memory(alias_idx)); | 
|---|
| 3298 | } | 
|---|
| 3299 | Node* membar = _gvn.transform(mb); | 
|---|
| 3300 | set_control(_gvn.transform(new ProjNode(membar, TypeFunc::Control))); | 
|---|
| 3301 | if (alias_idx == Compile::AliasIdxBot) { | 
|---|
| 3302 | merged_memory()->set_base_memory(_gvn.transform(new ProjNode(membar, TypeFunc::Memory))); | 
|---|
| 3303 | } else { | 
|---|
| 3304 | set_memory(_gvn.transform(new ProjNode(membar, TypeFunc::Memory)),alias_idx); | 
|---|
| 3305 | } | 
|---|
| 3306 | return membar; | 
|---|
| 3307 | } | 
|---|
| 3308 |  | 
|---|
| 3309 | void GraphKit::insert_store_load_for_barrier() { | 
|---|
| 3310 | Node* mem = reset_memory(); | 
|---|
| 3311 | MemBarNode* mb = MemBarNode::make(C, Op_MemBarVolatile, Compile::AliasIdxRaw); | 
|---|
| 3312 | mb->init_req(TypeFunc::Control, control()); | 
|---|
| 3313 | mb->init_req(TypeFunc::Memory, mem); | 
|---|
| 3314 | Node* membar = _gvn.transform(mb); | 
|---|
| 3315 | set_control(_gvn.transform(new ProjNode(membar, TypeFunc::Control))); | 
|---|
| 3316 | Node* newmem = _gvn.transform(new ProjNode(membar, TypeFunc::Memory)); | 
|---|
| 3317 | set_all_memory(mem); | 
|---|
| 3318 | set_memory(newmem, Compile::AliasIdxRaw); | 
|---|
| 3319 | } | 
|---|
| 3320 |  | 
|---|
| 3321 | //------------------------------shared_lock------------------------------------ | 
|---|
| 3322 | // Emit locking code. | 
|---|
| 3323 | FastLockNode* GraphKit::shared_lock(Node* obj) { | 
|---|
| 3324 | // bci is either a monitorenter bc or InvocationEntryBci | 
|---|
| 3325 | // %%% SynchronizationEntryBCI is redundant; use InvocationEntryBci in interfaces | 
|---|
| 3326 | assert(SynchronizationEntryBCI == InvocationEntryBci, ""); | 
|---|
| 3327 |  | 
|---|
| 3328 | if( !GenerateSynchronizationCode ) | 
|---|
| 3329 | return NULL;                // Not locking things? | 
|---|
| 3330 | if (stopped())                // Dead monitor? | 
|---|
| 3331 | return NULL; | 
|---|
| 3332 |  | 
|---|
| 3333 | assert(dead_locals_are_killed(), "should kill locals before sync. point"); | 
|---|
| 3334 |  | 
|---|
| 3335 | obj = access_resolve(obj, ACCESS_READ | ACCESS_WRITE); | 
|---|
| 3336 |  | 
|---|
| 3337 | // Box the stack location | 
|---|
| 3338 | Node* box = _gvn.transform(new BoxLockNode(next_monitor())); | 
|---|
| 3339 | Node* mem = reset_memory(); | 
|---|
| 3340 |  | 
|---|
| 3341 | FastLockNode * flock = _gvn.transform(new FastLockNode(0, obj, box) )->as_FastLock(); | 
|---|
| 3342 | if (UseBiasedLocking && PrintPreciseBiasedLockingStatistics) { | 
|---|
| 3343 | // Create the counters for this fast lock. | 
|---|
| 3344 | flock->create_lock_counter(sync_jvms()); // sync_jvms used to get current bci | 
|---|
| 3345 | } | 
|---|
| 3346 |  | 
|---|
| 3347 | // Create the rtm counters for this fast lock if needed. | 
|---|
| 3348 | flock->create_rtm_lock_counter(sync_jvms()); // sync_jvms used to get current bci | 
|---|
| 3349 |  | 
|---|
| 3350 | // Add monitor to debug info for the slow path.  If we block inside the | 
|---|
| 3351 | // slow path and de-opt, we need the monitor hanging around | 
|---|
| 3352 | map()->push_monitor( flock ); | 
|---|
| 3353 |  | 
|---|
| 3354 | const TypeFunc *tf = LockNode::lock_type(); | 
|---|
| 3355 | LockNode *lock = new LockNode(C, tf); | 
|---|
| 3356 |  | 
|---|
| 3357 | lock->init_req( TypeFunc::Control, control() ); | 
|---|
| 3358 | lock->init_req( TypeFunc::Memory , mem ); | 
|---|
| 3359 | lock->init_req( TypeFunc::I_O    , top() )     ;   // does no i/o | 
|---|
| 3360 | lock->init_req( TypeFunc::FramePtr, frameptr() ); | 
|---|
| 3361 | lock->init_req( TypeFunc::ReturnAdr, top() ); | 
|---|
| 3362 |  | 
|---|
| 3363 | lock->init_req(TypeFunc::Parms + 0, obj); | 
|---|
| 3364 | lock->init_req(TypeFunc::Parms + 1, box); | 
|---|
| 3365 | lock->init_req(TypeFunc::Parms + 2, flock); | 
|---|
| 3366 | add_safepoint_edges(lock); | 
|---|
| 3367 |  | 
|---|
| 3368 | lock = _gvn.transform( lock )->as_Lock(); | 
|---|
| 3369 |  | 
|---|
| 3370 | // lock has no side-effects, sets few values | 
|---|
| 3371 | set_predefined_output_for_runtime_call(lock, mem, TypeRawPtr::BOTTOM); | 
|---|
| 3372 |  | 
|---|
| 3373 | insert_mem_bar(Op_MemBarAcquireLock); | 
|---|
| 3374 |  | 
|---|
| 3375 | // Add this to the worklist so that the lock can be eliminated | 
|---|
| 3376 | record_for_igvn(lock); | 
|---|
| 3377 |  | 
|---|
| 3378 | #ifndef PRODUCT | 
|---|
| 3379 | if (PrintLockStatistics) { | 
|---|
| 3380 | // Update the counter for this lock.  Don't bother using an atomic | 
|---|
| 3381 | // operation since we don't require absolute accuracy. | 
|---|
| 3382 | lock->create_lock_counter(map()->jvms()); | 
|---|
| 3383 | increment_counter(lock->counter()->addr()); | 
|---|
| 3384 | } | 
|---|
| 3385 | #endif | 
|---|
| 3386 |  | 
|---|
| 3387 | return flock; | 
|---|
| 3388 | } | 
|---|
| 3389 |  | 
|---|
| 3390 |  | 
|---|
| 3391 | //------------------------------shared_unlock---------------------------------- | 
|---|
| 3392 | // Emit unlocking code. | 
|---|
| 3393 | void GraphKit::shared_unlock(Node* box, Node* obj) { | 
|---|
| 3394 | // bci is either a monitorenter bc or InvocationEntryBci | 
|---|
| 3395 | // %%% SynchronizationEntryBCI is redundant; use InvocationEntryBci in interfaces | 
|---|
| 3396 | assert(SynchronizationEntryBCI == InvocationEntryBci, ""); | 
|---|
| 3397 |  | 
|---|
| 3398 | if( !GenerateSynchronizationCode ) | 
|---|
| 3399 | return; | 
|---|
| 3400 | if (stopped()) {               // Dead monitor? | 
|---|
| 3401 | map()->pop_monitor();        // Kill monitor from debug info | 
|---|
| 3402 | return; | 
|---|
| 3403 | } | 
|---|
| 3404 |  | 
|---|
| 3405 | // Memory barrier to avoid floating things down past the locked region | 
|---|
| 3406 | insert_mem_bar(Op_MemBarReleaseLock); | 
|---|
| 3407 |  | 
|---|
| 3408 | const TypeFunc *tf = OptoRuntime::complete_monitor_exit_Type(); | 
|---|
| 3409 | UnlockNode *unlock = new UnlockNode(C, tf); | 
|---|
| 3410 | #ifdef ASSERT | 
|---|
| 3411 | unlock->set_dbg_jvms(sync_jvms()); | 
|---|
| 3412 | #endif | 
|---|
| 3413 | uint raw_idx = Compile::AliasIdxRaw; | 
|---|
| 3414 | unlock->init_req( TypeFunc::Control, control() ); | 
|---|
| 3415 | unlock->init_req( TypeFunc::Memory , memory(raw_idx) ); | 
|---|
| 3416 | unlock->init_req( TypeFunc::I_O    , top() )     ;   // does no i/o | 
|---|
| 3417 | unlock->init_req( TypeFunc::FramePtr, frameptr() ); | 
|---|
| 3418 | unlock->init_req( TypeFunc::ReturnAdr, top() ); | 
|---|
| 3419 |  | 
|---|
| 3420 | unlock->init_req(TypeFunc::Parms + 0, obj); | 
|---|
| 3421 | unlock->init_req(TypeFunc::Parms + 1, box); | 
|---|
| 3422 | unlock = _gvn.transform(unlock)->as_Unlock(); | 
|---|
| 3423 |  | 
|---|
| 3424 | Node* mem = reset_memory(); | 
|---|
| 3425 |  | 
|---|
| 3426 | // unlock has no side-effects, sets few values | 
|---|
| 3427 | set_predefined_output_for_runtime_call(unlock, mem, TypeRawPtr::BOTTOM); | 
|---|
| 3428 |  | 
|---|
| 3429 | // Kill monitor from debug info | 
|---|
| 3430 | map()->pop_monitor( ); | 
|---|
| 3431 | } | 
|---|
| 3432 |  | 
|---|
| 3433 | //-------------------------------get_layout_helper----------------------------- | 
|---|
| 3434 | // If the given klass is a constant or known to be an array, | 
|---|
| 3435 | // fetch the constant layout helper value into constant_value | 
|---|
| 3436 | // and return (Node*)NULL.  Otherwise, load the non-constant | 
|---|
| 3437 | // layout helper value, and return the node which represents it. | 
|---|
| 3438 | // This two-faced routine is useful because allocation sites | 
|---|
| 3439 | // almost always feature constant types. | 
|---|
| 3440 | Node* GraphKit::get_layout_helper(Node* klass_node, jint& constant_value) { | 
|---|
| 3441 | const TypeKlassPtr* inst_klass = _gvn.type(klass_node)->isa_klassptr(); | 
|---|
| 3442 | if (!StressReflectiveCode && inst_klass != NULL) { | 
|---|
| 3443 | ciKlass* klass = inst_klass->klass(); | 
|---|
| 3444 | bool    xklass = inst_klass->klass_is_exact(); | 
|---|
| 3445 | if (xklass || klass->is_array_klass()) { | 
|---|
| 3446 | jint lhelper = klass->layout_helper(); | 
|---|
| 3447 | if (lhelper != Klass::_lh_neutral_value) { | 
|---|
| 3448 | constant_value = lhelper; | 
|---|
| 3449 | return (Node*) NULL; | 
|---|
| 3450 | } | 
|---|
| 3451 | } | 
|---|
| 3452 | } | 
|---|
| 3453 | constant_value = Klass::_lh_neutral_value;  // put in a known value | 
|---|
| 3454 | Node* lhp = basic_plus_adr(klass_node, klass_node, in_bytes(Klass::layout_helper_offset())); | 
|---|
| 3455 | return make_load(NULL, lhp, TypeInt::INT, T_INT, MemNode::unordered); | 
|---|
| 3456 | } | 
|---|
| 3457 |  | 
|---|
| 3458 | // We just put in an allocate/initialize with a big raw-memory effect. | 
|---|
| 3459 | // Hook selected additional alias categories on the initialization. | 
|---|
| 3460 | static void hook_memory_on_init(GraphKit& kit, int alias_idx, | 
|---|
| 3461 | MergeMemNode* init_in_merge, | 
|---|
| 3462 | Node* init_out_raw) { | 
|---|
| 3463 | DEBUG_ONLY(Node* init_in_raw = init_in_merge->base_memory()); | 
|---|
| 3464 | assert(init_in_merge->memory_at(alias_idx) == init_in_raw, ""); | 
|---|
| 3465 |  | 
|---|
| 3466 | Node* prevmem = kit.memory(alias_idx); | 
|---|
| 3467 | init_in_merge->set_memory_at(alias_idx, prevmem); | 
|---|
| 3468 | kit.set_memory(init_out_raw, alias_idx); | 
|---|
| 3469 | } | 
|---|
| 3470 |  | 
|---|
| 3471 | //---------------------------set_output_for_allocation------------------------- | 
|---|
| 3472 | Node* GraphKit::set_output_for_allocation(AllocateNode* alloc, | 
|---|
| 3473 | const TypeOopPtr* oop_type, | 
|---|
| 3474 | bool deoptimize_on_exception) { | 
|---|
| 3475 | int rawidx = Compile::AliasIdxRaw; | 
|---|
| 3476 | alloc->set_req( TypeFunc::FramePtr, frameptr() ); | 
|---|
| 3477 | add_safepoint_edges(alloc); | 
|---|
| 3478 | Node* allocx = _gvn.transform(alloc); | 
|---|
| 3479 | set_control( _gvn.transform(new ProjNode(allocx, TypeFunc::Control) ) ); | 
|---|
| 3480 | // create memory projection for i_o | 
|---|
| 3481 | set_memory ( _gvn.transform( new ProjNode(allocx, TypeFunc::Memory, true) ), rawidx ); | 
|---|
| 3482 | make_slow_call_ex(allocx, env()->Throwable_klass(), true, deoptimize_on_exception); | 
|---|
| 3483 |  | 
|---|
| 3484 | // create a memory projection as for the normal control path | 
|---|
| 3485 | Node* malloc = _gvn.transform(new ProjNode(allocx, TypeFunc::Memory)); | 
|---|
| 3486 | set_memory(malloc, rawidx); | 
|---|
| 3487 |  | 
|---|
| 3488 | // a normal slow-call doesn't change i_o, but an allocation does | 
|---|
| 3489 | // we create a separate i_o projection for the normal control path | 
|---|
| 3490 | set_i_o(_gvn.transform( new ProjNode(allocx, TypeFunc::I_O, false) ) ); | 
|---|
| 3491 | Node* rawoop = _gvn.transform( new ProjNode(allocx, TypeFunc::Parms) ); | 
|---|
| 3492 |  | 
|---|
| 3493 | // put in an initialization barrier | 
|---|
| 3494 | InitializeNode* init = insert_mem_bar_volatile(Op_Initialize, rawidx, | 
|---|
| 3495 | rawoop)->as_Initialize(); | 
|---|
| 3496 | assert(alloc->initialization() == init, "2-way macro link must work"); | 
|---|
| 3497 | assert(init ->allocation()     == alloc, "2-way macro link must work"); | 
|---|
| 3498 | { | 
|---|
| 3499 | // Extract memory strands which may participate in the new object's | 
|---|
| 3500 | // initialization, and source them from the new InitializeNode. | 
|---|
| 3501 | // This will allow us to observe initializations when they occur, | 
|---|
| 3502 | // and link them properly (as a group) to the InitializeNode. | 
|---|
| 3503 | assert(init->in(InitializeNode::Memory) == malloc, ""); | 
|---|
| 3504 | MergeMemNode* minit_in = MergeMemNode::make(malloc); | 
|---|
| 3505 | init->set_req(InitializeNode::Memory, minit_in); | 
|---|
| 3506 | record_for_igvn(minit_in); // fold it up later, if possible | 
|---|
| 3507 | Node* minit_out = memory(rawidx); | 
|---|
| 3508 | assert(minit_out->is_Proj() && minit_out->in(0) == init, ""); | 
|---|
| 3509 | // Add an edge in the MergeMem for the header fields so an access | 
|---|
| 3510 | // to one of those has correct memory state | 
|---|
| 3511 | set_memory(minit_out, C->get_alias_index(oop_type->add_offset(oopDesc::mark_offset_in_bytes()))); | 
|---|
| 3512 | set_memory(minit_out, C->get_alias_index(oop_type->add_offset(oopDesc::klass_offset_in_bytes()))); | 
|---|
| 3513 | if (oop_type->isa_aryptr()) { | 
|---|
| 3514 | const TypePtr* telemref = oop_type->add_offset(Type::OffsetBot); | 
|---|
| 3515 | int            elemidx  = C->get_alias_index(telemref); | 
|---|
| 3516 | hook_memory_on_init(*this, elemidx, minit_in, minit_out); | 
|---|
| 3517 | } else if (oop_type->isa_instptr()) { | 
|---|
| 3518 | ciInstanceKlass* ik = oop_type->klass()->as_instance_klass(); | 
|---|
| 3519 | for (int i = 0, len = ik->nof_nonstatic_fields(); i < len; i++) { | 
|---|
| 3520 | ciField* field = ik->nonstatic_field_at(i); | 
|---|
| 3521 | if (field->offset() >= TrackedInitializationLimit * HeapWordSize) | 
|---|
| 3522 | continue;  // do not bother to track really large numbers of fields | 
|---|
| 3523 | // Find (or create) the alias category for this field: | 
|---|
| 3524 | int fieldidx = C->alias_type(field)->index(); | 
|---|
| 3525 | hook_memory_on_init(*this, fieldidx, minit_in, minit_out); | 
|---|
| 3526 | } | 
|---|
| 3527 | } | 
|---|
| 3528 | } | 
|---|
| 3529 |  | 
|---|
| 3530 | // Cast raw oop to the real thing... | 
|---|
| 3531 | Node* javaoop = new CheckCastPPNode(control(), rawoop, oop_type); | 
|---|
| 3532 | javaoop = _gvn.transform(javaoop); | 
|---|
| 3533 | C->set_recent_alloc(control(), javaoop); | 
|---|
| 3534 | assert(just_allocated_object(control()) == javaoop, "just allocated"); | 
|---|
| 3535 |  | 
|---|
| 3536 | #ifdef ASSERT | 
|---|
| 3537 | { // Verify that the AllocateNode::Ideal_allocation recognizers work: | 
|---|
| 3538 | assert(AllocateNode::Ideal_allocation(rawoop, &_gvn) == alloc, | 
|---|
| 3539 | "Ideal_allocation works"); | 
|---|
| 3540 | assert(AllocateNode::Ideal_allocation(javaoop, &_gvn) == alloc, | 
|---|
| 3541 | "Ideal_allocation works"); | 
|---|
| 3542 | if (alloc->is_AllocateArray()) { | 
|---|
| 3543 | assert(AllocateArrayNode::Ideal_array_allocation(rawoop, &_gvn) == alloc->as_AllocateArray(), | 
|---|
| 3544 | "Ideal_allocation works"); | 
|---|
| 3545 | assert(AllocateArrayNode::Ideal_array_allocation(javaoop, &_gvn) == alloc->as_AllocateArray(), | 
|---|
| 3546 | "Ideal_allocation works"); | 
|---|
| 3547 | } else { | 
|---|
| 3548 | assert(alloc->in(AllocateNode::ALength)->is_top(), "no length, please"); | 
|---|
| 3549 | } | 
|---|
| 3550 | } | 
|---|
| 3551 | #endif //ASSERT | 
|---|
| 3552 |  | 
|---|
| 3553 | return javaoop; | 
|---|
| 3554 | } | 
|---|
| 3555 |  | 
|---|
| 3556 | //---------------------------new_instance-------------------------------------- | 
|---|
| 3557 | // This routine takes a klass_node which may be constant (for a static type) | 
|---|
| 3558 | // or may be non-constant (for reflective code).  It will work equally well | 
|---|
| 3559 | // for either, and the graph will fold nicely if the optimizer later reduces | 
|---|
| 3560 | // the type to a constant. | 
|---|
| 3561 | // The optional arguments are for specialized use by intrinsics: | 
|---|
| 3562 | //  - If 'extra_slow_test' if not null is an extra condition for the slow-path. | 
|---|
| 3563 | //  - If 'return_size_val', report the the total object size to the caller. | 
|---|
| 3564 | //  - deoptimize_on_exception controls how Java exceptions are handled (rethrow vs deoptimize) | 
|---|
| 3565 | Node* GraphKit::new_instance(Node* klass_node, | 
|---|
| 3566 | Node* , | 
|---|
| 3567 | Node* *return_size_val, | 
|---|
| 3568 | bool deoptimize_on_exception) { | 
|---|
| 3569 | // Compute size in doublewords | 
|---|
| 3570 | // The size is always an integral number of doublewords, represented | 
|---|
| 3571 | // as a positive bytewise size stored in the klass's layout_helper. | 
|---|
| 3572 | // The layout_helper also encodes (in a low bit) the need for a slow path. | 
|---|
| 3573 | jint  layout_con = Klass::_lh_neutral_value; | 
|---|
| 3574 | Node* layout_val = get_layout_helper(klass_node, layout_con); | 
|---|
| 3575 | int   layout_is_con = (layout_val == NULL); | 
|---|
| 3576 |  | 
|---|
| 3577 | if (extra_slow_test == NULL)  extra_slow_test = intcon(0); | 
|---|
| 3578 | // Generate the initial go-slow test.  It's either ALWAYS (return a | 
|---|
| 3579 | // Node for 1) or NEVER (return a NULL) or perhaps (in the reflective | 
|---|
| 3580 | // case) a computed value derived from the layout_helper. | 
|---|
| 3581 | Node* initial_slow_test = NULL; | 
|---|
| 3582 | if (layout_is_con) { | 
|---|
| 3583 | assert(!StressReflectiveCode, "stress mode does not use these paths"); | 
|---|
| 3584 | bool must_go_slow = Klass::layout_helper_needs_slow_path(layout_con); | 
|---|
| 3585 | initial_slow_test = must_go_slow ? intcon(1) : extra_slow_test; | 
|---|
| 3586 | } else {   // reflective case | 
|---|
| 3587 | // This reflective path is used by Unsafe.allocateInstance. | 
|---|
| 3588 | // (It may be stress-tested by specifying StressReflectiveCode.) | 
|---|
| 3589 | // Basically, we want to get into the VM is there's an illegal argument. | 
|---|
| 3590 | Node* bit = intcon(Klass::_lh_instance_slow_path_bit); | 
|---|
| 3591 | initial_slow_test = _gvn.transform( new AndINode(layout_val, bit) ); | 
|---|
| 3592 | if (extra_slow_test != intcon(0)) { | 
|---|
| 3593 | initial_slow_test = _gvn.transform( new OrINode(initial_slow_test, extra_slow_test) ); | 
|---|
| 3594 | } | 
|---|
| 3595 | // (Macro-expander will further convert this to a Bool, if necessary.) | 
|---|
| 3596 | } | 
|---|
| 3597 |  | 
|---|
| 3598 | // Find the size in bytes.  This is easy; it's the layout_helper. | 
|---|
| 3599 | // The size value must be valid even if the slow path is taken. | 
|---|
| 3600 | Node* size = NULL; | 
|---|
| 3601 | if (layout_is_con) { | 
|---|
| 3602 | size = MakeConX(Klass::layout_helper_size_in_bytes(layout_con)); | 
|---|
| 3603 | } else {   // reflective case | 
|---|
| 3604 | // This reflective path is used by clone and Unsafe.allocateInstance. | 
|---|
| 3605 | size = ConvI2X(layout_val); | 
|---|
| 3606 |  | 
|---|
| 3607 | // Clear the low bits to extract layout_helper_size_in_bytes: | 
|---|
| 3608 | assert((int)Klass::_lh_instance_slow_path_bit < BytesPerLong, "clear bit"); | 
|---|
| 3609 | Node* mask = MakeConX(~ (intptr_t)right_n_bits(LogBytesPerLong)); | 
|---|
| 3610 | size = _gvn.transform( new AndXNode(size, mask) ); | 
|---|
| 3611 | } | 
|---|
| 3612 | if (return_size_val != NULL) { | 
|---|
| 3613 | (*return_size_val) = size; | 
|---|
| 3614 | } | 
|---|
| 3615 |  | 
|---|
| 3616 | // This is a precise notnull oop of the klass. | 
|---|
| 3617 | // (Actually, it need not be precise if this is a reflective allocation.) | 
|---|
| 3618 | // It's what we cast the result to. | 
|---|
| 3619 | const TypeKlassPtr* tklass = _gvn.type(klass_node)->isa_klassptr(); | 
|---|
| 3620 | if (!tklass)  tklass = TypeKlassPtr::OBJECT; | 
|---|
| 3621 | const TypeOopPtr* oop_type = tklass->as_instance_type(); | 
|---|
| 3622 |  | 
|---|
| 3623 | // Now generate allocation code | 
|---|
| 3624 |  | 
|---|
| 3625 | // The entire memory state is needed for slow path of the allocation | 
|---|
| 3626 | // since GC and deoptimization can happened. | 
|---|
| 3627 | Node *mem = reset_memory(); | 
|---|
| 3628 | set_all_memory(mem); // Create new memory state | 
|---|
| 3629 |  | 
|---|
| 3630 | AllocateNode* alloc = new AllocateNode(C, AllocateNode::alloc_type(Type::TOP), | 
|---|
| 3631 | control(), mem, i_o(), | 
|---|
| 3632 | size, klass_node, | 
|---|
| 3633 | initial_slow_test); | 
|---|
| 3634 |  | 
|---|
| 3635 | return set_output_for_allocation(alloc, oop_type, deoptimize_on_exception); | 
|---|
| 3636 | } | 
|---|
| 3637 |  | 
|---|
| 3638 | //-------------------------------new_array------------------------------------- | 
|---|
| 3639 | // helper for both newarray and anewarray | 
|---|
| 3640 | // The 'length' parameter is (obviously) the length of the array. | 
|---|
| 3641 | // See comments on new_instance for the meaning of the other arguments. | 
|---|
| 3642 | Node* GraphKit::new_array(Node* klass_node,     // array klass (maybe variable) | 
|---|
| 3643 | Node* length,         // number of array elements | 
|---|
| 3644 | int   nargs,          // number of arguments to push back for uncommon trap | 
|---|
| 3645 | Node* *return_size_val, | 
|---|
| 3646 | bool deoptimize_on_exception) { | 
|---|
| 3647 | jint  layout_con = Klass::_lh_neutral_value; | 
|---|
| 3648 | Node* layout_val = get_layout_helper(klass_node, layout_con); | 
|---|
| 3649 | int   layout_is_con = (layout_val == NULL); | 
|---|
| 3650 |  | 
|---|
| 3651 | if (!layout_is_con && !StressReflectiveCode && | 
|---|
| 3652 | !too_many_traps(Deoptimization::Reason_class_check)) { | 
|---|
| 3653 | // This is a reflective array creation site. | 
|---|
| 3654 | // Optimistically assume that it is a subtype of Object[], | 
|---|
| 3655 | // so that we can fold up all the address arithmetic. | 
|---|
| 3656 | layout_con = Klass::array_layout_helper(T_OBJECT); | 
|---|
| 3657 | Node* cmp_lh = _gvn.transform( new CmpINode(layout_val, intcon(layout_con)) ); | 
|---|
| 3658 | Node* bol_lh = _gvn.transform( new BoolNode(cmp_lh, BoolTest::eq) ); | 
|---|
| 3659 | { BuildCutout unless(this, bol_lh, PROB_MAX); | 
|---|
| 3660 | inc_sp(nargs); | 
|---|
| 3661 | uncommon_trap(Deoptimization::Reason_class_check, | 
|---|
| 3662 | Deoptimization::Action_maybe_recompile); | 
|---|
| 3663 | } | 
|---|
| 3664 | layout_val = NULL; | 
|---|
| 3665 | layout_is_con = true; | 
|---|
| 3666 | } | 
|---|
| 3667 |  | 
|---|
| 3668 | // Generate the initial go-slow test.  Make sure we do not overflow | 
|---|
| 3669 | // if length is huge (near 2Gig) or negative!  We do not need | 
|---|
| 3670 | // exact double-words here, just a close approximation of needed | 
|---|
| 3671 | // double-words.  We can't add any offset or rounding bits, lest we | 
|---|
| 3672 | // take a size -1 of bytes and make it positive.  Use an unsigned | 
|---|
| 3673 | // compare, so negative sizes look hugely positive. | 
|---|
| 3674 | int fast_size_limit = FastAllocateSizeLimit; | 
|---|
| 3675 | if (layout_is_con) { | 
|---|
| 3676 | assert(!StressReflectiveCode, "stress mode does not use these paths"); | 
|---|
| 3677 | // Increase the size limit if we have exact knowledge of array type. | 
|---|
| 3678 | int log2_esize = Klass::layout_helper_log2_element_size(layout_con); | 
|---|
| 3679 | fast_size_limit <<= (LogBytesPerLong - log2_esize); | 
|---|
| 3680 | } | 
|---|
| 3681 |  | 
|---|
| 3682 | Node* initial_slow_cmp  = _gvn.transform( new CmpUNode( length, intcon( fast_size_limit ) ) ); | 
|---|
| 3683 | Node* initial_slow_test = _gvn.transform( new BoolNode( initial_slow_cmp, BoolTest::gt ) ); | 
|---|
| 3684 |  | 
|---|
| 3685 | // --- Size Computation --- | 
|---|
| 3686 | // array_size = round_to_heap(array_header + (length << elem_shift)); | 
|---|
| 3687 | // where round_to_heap(x) == align_to(x, MinObjAlignmentInBytes) | 
|---|
| 3688 | // and align_to(x, y) == ((x + y-1) & ~(y-1)) | 
|---|
| 3689 | // The rounding mask is strength-reduced, if possible. | 
|---|
| 3690 | int round_mask = MinObjAlignmentInBytes - 1; | 
|---|
| 3691 | Node*  = NULL; | 
|---|
| 3692 | int     = arrayOopDesc::base_offset_in_bytes(T_BYTE); | 
|---|
| 3693 | // (T_BYTE has the weakest alignment and size restrictions...) | 
|---|
| 3694 | if (layout_is_con) { | 
|---|
| 3695 | int       hsize  = Klass::layout_helper_header_size(layout_con); | 
|---|
| 3696 | int       eshift = Klass::layout_helper_log2_element_size(layout_con); | 
|---|
| 3697 | BasicType etype  = Klass::layout_helper_element_type(layout_con); | 
|---|
| 3698 | if ((round_mask & ~right_n_bits(eshift)) == 0) | 
|---|
| 3699 | round_mask = 0;  // strength-reduce it if it goes away completely | 
|---|
| 3700 | assert((hsize & right_n_bits(eshift)) == 0, "hsize is pre-rounded"); | 
|---|
| 3701 | assert(header_size_min <= hsize, "generic minimum is smallest"); | 
|---|
| 3702 | header_size_min = hsize; | 
|---|
| 3703 | header_size = intcon(hsize + round_mask); | 
|---|
| 3704 | } else { | 
|---|
| 3705 | Node* hss   = intcon(Klass::_lh_header_size_shift); | 
|---|
| 3706 | Node* hsm   = intcon(Klass::_lh_header_size_mask); | 
|---|
| 3707 | Node* hsize = _gvn.transform( new URShiftINode(layout_val, hss) ); | 
|---|
| 3708 | hsize       = _gvn.transform( new AndINode(hsize, hsm) ); | 
|---|
| 3709 | Node* mask  = intcon(round_mask); | 
|---|
| 3710 | header_size = _gvn.transform( new AddINode(hsize, mask) ); | 
|---|
| 3711 | } | 
|---|
| 3712 |  | 
|---|
| 3713 | Node* elem_shift = NULL; | 
|---|
| 3714 | if (layout_is_con) { | 
|---|
| 3715 | int eshift = Klass::layout_helper_log2_element_size(layout_con); | 
|---|
| 3716 | if (eshift != 0) | 
|---|
| 3717 | elem_shift = intcon(eshift); | 
|---|
| 3718 | } else { | 
|---|
| 3719 | // There is no need to mask or shift this value. | 
|---|
| 3720 | // The semantics of LShiftINode include an implicit mask to 0x1F. | 
|---|
| 3721 | assert(Klass::_lh_log2_element_size_shift == 0, "use shift in place"); | 
|---|
| 3722 | elem_shift = layout_val; | 
|---|
| 3723 | } | 
|---|
| 3724 |  | 
|---|
| 3725 | // Transition to native address size for all offset calculations: | 
|---|
| 3726 | Node* lengthx = ConvI2X(length); | 
|---|
| 3727 | Node*  = ConvI2X(header_size); | 
|---|
| 3728 | #ifdef _LP64 | 
|---|
| 3729 | { const TypeInt* tilen = _gvn.find_int_type(length); | 
|---|
| 3730 | if (tilen != NULL && tilen->_lo < 0) { | 
|---|
| 3731 | // Add a manual constraint to a positive range.  Cf. array_element_address. | 
|---|
| 3732 | jint size_max = fast_size_limit; | 
|---|
| 3733 | if (size_max > tilen->_hi)  size_max = tilen->_hi; | 
|---|
| 3734 | const TypeInt* tlcon = TypeInt::make(0, size_max, Type::WidenMin); | 
|---|
| 3735 |  | 
|---|
| 3736 | // Only do a narrow I2L conversion if the range check passed. | 
|---|
| 3737 | IfNode* iff = new IfNode(control(), initial_slow_test, PROB_MIN, COUNT_UNKNOWN); | 
|---|
| 3738 | _gvn.transform(iff); | 
|---|
| 3739 | RegionNode* region = new RegionNode(3); | 
|---|
| 3740 | _gvn.set_type(region, Type::CONTROL); | 
|---|
| 3741 | lengthx = new PhiNode(region, TypeLong::LONG); | 
|---|
| 3742 | _gvn.set_type(lengthx, TypeLong::LONG); | 
|---|
| 3743 |  | 
|---|
| 3744 | // Range check passed. Use ConvI2L node with narrow type. | 
|---|
| 3745 | Node* passed = IfFalse(iff); | 
|---|
| 3746 | region->init_req(1, passed); | 
|---|
| 3747 | // Make I2L conversion control dependent to prevent it from | 
|---|
| 3748 | // floating above the range check during loop optimizations. | 
|---|
| 3749 | lengthx->init_req(1, C->constrained_convI2L(&_gvn, length, tlcon, passed)); | 
|---|
| 3750 |  | 
|---|
| 3751 | // Range check failed. Use ConvI2L with wide type because length may be invalid. | 
|---|
| 3752 | region->init_req(2, IfTrue(iff)); | 
|---|
| 3753 | lengthx->init_req(2, ConvI2X(length)); | 
|---|
| 3754 |  | 
|---|
| 3755 | set_control(region); | 
|---|
| 3756 | record_for_igvn(region); | 
|---|
| 3757 | record_for_igvn(lengthx); | 
|---|
| 3758 | } | 
|---|
| 3759 | } | 
|---|
| 3760 | #endif | 
|---|
| 3761 |  | 
|---|
| 3762 | // Combine header size (plus rounding) and body size.  Then round down. | 
|---|
| 3763 | // This computation cannot overflow, because it is used only in two | 
|---|
| 3764 | // places, one where the length is sharply limited, and the other | 
|---|
| 3765 | // after a successful allocation. | 
|---|
| 3766 | Node* abody = lengthx; | 
|---|
| 3767 | if (elem_shift != NULL) | 
|---|
| 3768 | abody     = _gvn.transform( new LShiftXNode(lengthx, elem_shift) ); | 
|---|
| 3769 | Node* size  = _gvn.transform( new AddXNode(headerx, abody) ); | 
|---|
| 3770 | if (round_mask != 0) { | 
|---|
| 3771 | Node* mask = MakeConX(~round_mask); | 
|---|
| 3772 | size       = _gvn.transform( new AndXNode(size, mask) ); | 
|---|
| 3773 | } | 
|---|
| 3774 | // else if round_mask == 0, the size computation is self-rounding | 
|---|
| 3775 |  | 
|---|
| 3776 | if (return_size_val != NULL) { | 
|---|
| 3777 | // This is the size | 
|---|
| 3778 | (*return_size_val) = size; | 
|---|
| 3779 | } | 
|---|
| 3780 |  | 
|---|
| 3781 | // Now generate allocation code | 
|---|
| 3782 |  | 
|---|
| 3783 | // The entire memory state is needed for slow path of the allocation | 
|---|
| 3784 | // since GC and deoptimization can happened. | 
|---|
| 3785 | Node *mem = reset_memory(); | 
|---|
| 3786 | set_all_memory(mem); // Create new memory state | 
|---|
| 3787 |  | 
|---|
| 3788 | if (initial_slow_test->is_Bool()) { | 
|---|
| 3789 | // Hide it behind a CMoveI, or else PhaseIdealLoop::split_up will get sick. | 
|---|
| 3790 | initial_slow_test = initial_slow_test->as_Bool()->as_int_value(&_gvn); | 
|---|
| 3791 | } | 
|---|
| 3792 |  | 
|---|
| 3793 | // Create the AllocateArrayNode and its result projections | 
|---|
| 3794 | AllocateArrayNode* alloc | 
|---|
| 3795 | = new AllocateArrayNode(C, AllocateArrayNode::alloc_type(TypeInt::INT), | 
|---|
| 3796 | control(), mem, i_o(), | 
|---|
| 3797 | size, klass_node, | 
|---|
| 3798 | initial_slow_test, | 
|---|
| 3799 | length); | 
|---|
| 3800 |  | 
|---|
| 3801 | // Cast to correct type.  Note that the klass_node may be constant or not, | 
|---|
| 3802 | // and in the latter case the actual array type will be inexact also. | 
|---|
| 3803 | // (This happens via a non-constant argument to inline_native_newArray.) | 
|---|
| 3804 | // In any case, the value of klass_node provides the desired array type. | 
|---|
| 3805 | const TypeInt* length_type = _gvn.find_int_type(length); | 
|---|
| 3806 | const TypeOopPtr* ary_type = _gvn.type(klass_node)->is_klassptr()->as_instance_type(); | 
|---|
| 3807 | if (ary_type->isa_aryptr() && length_type != NULL) { | 
|---|
| 3808 | // Try to get a better type than POS for the size | 
|---|
| 3809 | ary_type = ary_type->is_aryptr()->cast_to_size(length_type); | 
|---|
| 3810 | } | 
|---|
| 3811 |  | 
|---|
| 3812 | Node* javaoop = set_output_for_allocation(alloc, ary_type, deoptimize_on_exception); | 
|---|
| 3813 |  | 
|---|
| 3814 | // Cast length on remaining path to be as narrow as possible | 
|---|
| 3815 | if (map()->find_edge(length) >= 0) { | 
|---|
| 3816 | Node* ccast = alloc->make_ideal_length(ary_type, &_gvn); | 
|---|
| 3817 | if (ccast != length) { | 
|---|
| 3818 | _gvn.set_type_bottom(ccast); | 
|---|
| 3819 | record_for_igvn(ccast); | 
|---|
| 3820 | replace_in_map(length, ccast); | 
|---|
| 3821 | } | 
|---|
| 3822 | } | 
|---|
| 3823 |  | 
|---|
| 3824 | return javaoop; | 
|---|
| 3825 | } | 
|---|
| 3826 |  | 
|---|
| 3827 | // The following "Ideal_foo" functions are placed here because they recognize | 
|---|
| 3828 | // the graph shapes created by the functions immediately above. | 
|---|
| 3829 |  | 
|---|
| 3830 | //---------------------------Ideal_allocation---------------------------------- | 
|---|
| 3831 | // Given an oop pointer or raw pointer, see if it feeds from an AllocateNode. | 
|---|
| 3832 | AllocateNode* AllocateNode::Ideal_allocation(Node* ptr, PhaseTransform* phase) { | 
|---|
| 3833 | if (ptr == NULL) {     // reduce dumb test in callers | 
|---|
| 3834 | return NULL; | 
|---|
| 3835 | } | 
|---|
| 3836 |  | 
|---|
| 3837 | BarrierSetC2* bs = BarrierSet::barrier_set()->barrier_set_c2(); | 
|---|
| 3838 | ptr = bs->step_over_gc_barrier(ptr); | 
|---|
| 3839 |  | 
|---|
| 3840 | if (ptr->is_CheckCastPP()) { // strip only one raw-to-oop cast | 
|---|
| 3841 | ptr = ptr->in(1); | 
|---|
| 3842 | if (ptr == NULL) return NULL; | 
|---|
| 3843 | } | 
|---|
| 3844 | // Return NULL for allocations with several casts: | 
|---|
| 3845 | //   j.l.reflect.Array.newInstance(jobject, jint) | 
|---|
| 3846 | //   Object.clone() | 
|---|
| 3847 | // to keep more precise type from last cast. | 
|---|
| 3848 | if (ptr->is_Proj()) { | 
|---|
| 3849 | Node* allo = ptr->in(0); | 
|---|
| 3850 | if (allo != NULL && allo->is_Allocate()) { | 
|---|
| 3851 | return allo->as_Allocate(); | 
|---|
| 3852 | } | 
|---|
| 3853 | } | 
|---|
| 3854 | // Report failure to match. | 
|---|
| 3855 | return NULL; | 
|---|
| 3856 | } | 
|---|
| 3857 |  | 
|---|
| 3858 | // Fancy version which also strips off an offset (and reports it to caller). | 
|---|
| 3859 | AllocateNode* AllocateNode::Ideal_allocation(Node* ptr, PhaseTransform* phase, | 
|---|
| 3860 | intptr_t& offset) { | 
|---|
| 3861 | Node* base = AddPNode::Ideal_base_and_offset(ptr, phase, offset); | 
|---|
| 3862 | if (base == NULL)  return NULL; | 
|---|
| 3863 | return Ideal_allocation(base, phase); | 
|---|
| 3864 | } | 
|---|
| 3865 |  | 
|---|
| 3866 | // Trace Initialize <- Proj[Parm] <- Allocate | 
|---|
| 3867 | AllocateNode* InitializeNode::allocation() { | 
|---|
| 3868 | Node* rawoop = in(InitializeNode::RawAddress); | 
|---|
| 3869 | if (rawoop->is_Proj()) { | 
|---|
| 3870 | Node* alloc = rawoop->in(0); | 
|---|
| 3871 | if (alloc->is_Allocate()) { | 
|---|
| 3872 | return alloc->as_Allocate(); | 
|---|
| 3873 | } | 
|---|
| 3874 | } | 
|---|
| 3875 | return NULL; | 
|---|
| 3876 | } | 
|---|
| 3877 |  | 
|---|
| 3878 | // Trace Allocate -> Proj[Parm] -> Initialize | 
|---|
| 3879 | InitializeNode* AllocateNode::initialization() { | 
|---|
| 3880 | ProjNode* rawoop = proj_out_or_null(AllocateNode::RawAddress); | 
|---|
| 3881 | if (rawoop == NULL)  return NULL; | 
|---|
| 3882 | for (DUIterator_Fast imax, i = rawoop->fast_outs(imax); i < imax; i++) { | 
|---|
| 3883 | Node* init = rawoop->fast_out(i); | 
|---|
| 3884 | if (init->is_Initialize()) { | 
|---|
| 3885 | assert(init->as_Initialize()->allocation() == this, "2-way link"); | 
|---|
| 3886 | return init->as_Initialize(); | 
|---|
| 3887 | } | 
|---|
| 3888 | } | 
|---|
| 3889 | return NULL; | 
|---|
| 3890 | } | 
|---|
| 3891 |  | 
|---|
| 3892 | //----------------------------- loop predicates --------------------------- | 
|---|
| 3893 |  | 
|---|
| 3894 | //------------------------------add_predicate_impl---------------------------- | 
|---|
| 3895 | void GraphKit::add_predicate_impl(Deoptimization::DeoptReason reason, int nargs) { | 
|---|
| 3896 | // Too many traps seen? | 
|---|
| 3897 | if (too_many_traps(reason)) { | 
|---|
| 3898 | #ifdef ASSERT | 
|---|
| 3899 | if (TraceLoopPredicate) { | 
|---|
| 3900 | int tc = C->trap_count(reason); | 
|---|
| 3901 | tty->print( "too many traps=%s tcount=%d in ", | 
|---|
| 3902 | Deoptimization::trap_reason_name(reason), tc); | 
|---|
| 3903 | method()->print(); // which method has too many predicate traps | 
|---|
| 3904 | tty->cr(); | 
|---|
| 3905 | } | 
|---|
| 3906 | #endif | 
|---|
| 3907 | // We cannot afford to take more traps here, | 
|---|
| 3908 | // do not generate predicate. | 
|---|
| 3909 | return; | 
|---|
| 3910 | } | 
|---|
| 3911 |  | 
|---|
| 3912 | Node *cont    = _gvn.intcon(1); | 
|---|
| 3913 | Node* opq     = _gvn.transform(new Opaque1Node(C, cont)); | 
|---|
| 3914 | Node *bol     = _gvn.transform(new Conv2BNode(opq)); | 
|---|
| 3915 | IfNode* iff   = create_and_map_if(control(), bol, PROB_MAX, COUNT_UNKNOWN); | 
|---|
| 3916 | Node* iffalse = _gvn.transform(new IfFalseNode(iff)); | 
|---|
| 3917 | C->add_predicate_opaq(opq); | 
|---|
| 3918 | { | 
|---|
| 3919 | PreserveJVMState pjvms(this); | 
|---|
| 3920 | set_control(iffalse); | 
|---|
| 3921 | inc_sp(nargs); | 
|---|
| 3922 | uncommon_trap(reason, Deoptimization::Action_maybe_recompile); | 
|---|
| 3923 | } | 
|---|
| 3924 | Node* iftrue = _gvn.transform(new IfTrueNode(iff)); | 
|---|
| 3925 | set_control(iftrue); | 
|---|
| 3926 | } | 
|---|
| 3927 |  | 
|---|
| 3928 | //------------------------------add_predicate--------------------------------- | 
|---|
| 3929 | void GraphKit::add_predicate(int nargs) { | 
|---|
| 3930 | if (UseLoopPredicate) { | 
|---|
| 3931 | add_predicate_impl(Deoptimization::Reason_predicate, nargs); | 
|---|
| 3932 | } | 
|---|
| 3933 | if (UseProfiledLoopPredicate) { | 
|---|
| 3934 | add_predicate_impl(Deoptimization::Reason_profile_predicate, nargs); | 
|---|
| 3935 | } | 
|---|
| 3936 | // loop's limit check predicate should be near the loop. | 
|---|
| 3937 | add_predicate_impl(Deoptimization::Reason_loop_limit_check, nargs); | 
|---|
| 3938 | } | 
|---|
| 3939 |  | 
|---|
| 3940 | void GraphKit::sync_kit(IdealKit& ideal) { | 
|---|
| 3941 | set_all_memory(ideal.merged_memory()); | 
|---|
| 3942 | set_i_o(ideal.i_o()); | 
|---|
| 3943 | set_control(ideal.ctrl()); | 
|---|
| 3944 | } | 
|---|
| 3945 |  | 
|---|
| 3946 | void GraphKit::final_sync(IdealKit& ideal) { | 
|---|
| 3947 | // Final sync IdealKit and graphKit. | 
|---|
| 3948 | sync_kit(ideal); | 
|---|
| 3949 | } | 
|---|
| 3950 |  | 
|---|
| 3951 | Node* GraphKit::load_String_length(Node* str, bool set_ctrl) { | 
|---|
| 3952 | Node* len = load_array_length(load_String_value(str, set_ctrl)); | 
|---|
| 3953 | Node* coder = load_String_coder(str, set_ctrl); | 
|---|
| 3954 | // Divide length by 2 if coder is UTF16 | 
|---|
| 3955 | return _gvn.transform(new RShiftINode(len, coder)); | 
|---|
| 3956 | } | 
|---|
| 3957 |  | 
|---|
| 3958 | Node* GraphKit::load_String_value(Node* str, bool set_ctrl) { | 
|---|
| 3959 | int value_offset = java_lang_String::value_offset_in_bytes(); | 
|---|
| 3960 | const TypeInstPtr* string_type = TypeInstPtr::make(TypePtr::NotNull, C->env()->String_klass(), | 
|---|
| 3961 | false, NULL, 0); | 
|---|
| 3962 | const TypePtr* value_field_type = string_type->add_offset(value_offset); | 
|---|
| 3963 | const TypeAryPtr* value_type = TypeAryPtr::make(TypePtr::NotNull, | 
|---|
| 3964 | TypeAry::make(TypeInt::BYTE, TypeInt::POS), | 
|---|
| 3965 | ciTypeArrayKlass::make(T_BYTE), true, 0); | 
|---|
| 3966 | Node* p = basic_plus_adr(str, str, value_offset); | 
|---|
| 3967 | Node* load = access_load_at(str, p, value_field_type, value_type, T_OBJECT, | 
|---|
| 3968 | IN_HEAP | (set_ctrl ? C2_CONTROL_DEPENDENT_LOAD : 0) | MO_UNORDERED); | 
|---|
| 3969 | return load; | 
|---|
| 3970 | } | 
|---|
| 3971 |  | 
|---|
| 3972 | Node* GraphKit::load_String_coder(Node* str, bool set_ctrl) { | 
|---|
| 3973 | if (!CompactStrings) { | 
|---|
| 3974 | return intcon(java_lang_String::CODER_UTF16); | 
|---|
| 3975 | } | 
|---|
| 3976 | int coder_offset = java_lang_String::coder_offset_in_bytes(); | 
|---|
| 3977 | const TypeInstPtr* string_type = TypeInstPtr::make(TypePtr::NotNull, C->env()->String_klass(), | 
|---|
| 3978 | false, NULL, 0); | 
|---|
| 3979 | const TypePtr* coder_field_type = string_type->add_offset(coder_offset); | 
|---|
| 3980 |  | 
|---|
| 3981 | Node* p = basic_plus_adr(str, str, coder_offset); | 
|---|
| 3982 | Node* load = access_load_at(str, p, coder_field_type, TypeInt::BYTE, T_BYTE, | 
|---|
| 3983 | IN_HEAP | (set_ctrl ? C2_CONTROL_DEPENDENT_LOAD : 0) | MO_UNORDERED); | 
|---|
| 3984 | return load; | 
|---|
| 3985 | } | 
|---|
| 3986 |  | 
|---|
| 3987 | void GraphKit::store_String_value(Node* str, Node* value) { | 
|---|
| 3988 | int value_offset = java_lang_String::value_offset_in_bytes(); | 
|---|
| 3989 | const TypeInstPtr* string_type = TypeInstPtr::make(TypePtr::NotNull, C->env()->String_klass(), | 
|---|
| 3990 | false, NULL, 0); | 
|---|
| 3991 | const TypePtr* value_field_type = string_type->add_offset(value_offset); | 
|---|
| 3992 |  | 
|---|
| 3993 | access_store_at(str,  basic_plus_adr(str, value_offset), value_field_type, | 
|---|
| 3994 | value, TypeAryPtr::BYTES, T_OBJECT, IN_HEAP | MO_UNORDERED); | 
|---|
| 3995 | } | 
|---|
| 3996 |  | 
|---|
| 3997 | void GraphKit::store_String_coder(Node* str, Node* value) { | 
|---|
| 3998 | int coder_offset = java_lang_String::coder_offset_in_bytes(); | 
|---|
| 3999 | const TypeInstPtr* string_type = TypeInstPtr::make(TypePtr::NotNull, C->env()->String_klass(), | 
|---|
| 4000 | false, NULL, 0); | 
|---|
| 4001 | const TypePtr* coder_field_type = string_type->add_offset(coder_offset); | 
|---|
| 4002 |  | 
|---|
| 4003 | access_store_at(str, basic_plus_adr(str, coder_offset), coder_field_type, | 
|---|
| 4004 | value, TypeInt::BYTE, T_BYTE, IN_HEAP | MO_UNORDERED); | 
|---|
| 4005 | } | 
|---|
| 4006 |  | 
|---|
| 4007 | // Capture src and dst memory state with a MergeMemNode | 
|---|
| 4008 | Node* GraphKit::capture_memory(const TypePtr* src_type, const TypePtr* dst_type) { | 
|---|
| 4009 | if (src_type == dst_type) { | 
|---|
| 4010 | // Types are equal, we don't need a MergeMemNode | 
|---|
| 4011 | return memory(src_type); | 
|---|
| 4012 | } | 
|---|
| 4013 | MergeMemNode* merge = MergeMemNode::make(map()->memory()); | 
|---|
| 4014 | record_for_igvn(merge); // fold it up later, if possible | 
|---|
| 4015 | int src_idx = C->get_alias_index(src_type); | 
|---|
| 4016 | int dst_idx = C->get_alias_index(dst_type); | 
|---|
| 4017 | merge->set_memory_at(src_idx, memory(src_idx)); | 
|---|
| 4018 | merge->set_memory_at(dst_idx, memory(dst_idx)); | 
|---|
| 4019 | return merge; | 
|---|
| 4020 | } | 
|---|
| 4021 |  | 
|---|
| 4022 | Node* GraphKit::compress_string(Node* src, const TypeAryPtr* src_type, Node* dst, Node* count) { | 
|---|
| 4023 | assert(Matcher::match_rule_supported(Op_StrCompressedCopy), "Intrinsic not supported"); | 
|---|
| 4024 | assert(src_type == TypeAryPtr::BYTES || src_type == TypeAryPtr::CHARS, "invalid source type"); | 
|---|
| 4025 | // If input and output memory types differ, capture both states to preserve | 
|---|
| 4026 | // the dependency between preceding and subsequent loads/stores. | 
|---|
| 4027 | // For example, the following program: | 
|---|
| 4028 | //  StoreB | 
|---|
| 4029 | //  compress_string | 
|---|
| 4030 | //  LoadB | 
|---|
| 4031 | // has this memory graph (use->def): | 
|---|
| 4032 | //  LoadB -> compress_string -> CharMem | 
|---|
| 4033 | //             ... -> StoreB -> ByteMem | 
|---|
| 4034 | // The intrinsic hides the dependency between LoadB and StoreB, causing | 
|---|
| 4035 | // the load to read from memory not containing the result of the StoreB. | 
|---|
| 4036 | // The correct memory graph should look like this: | 
|---|
| 4037 | //  LoadB -> compress_string -> MergeMem(CharMem, StoreB(ByteMem)) | 
|---|
| 4038 | Node* mem = capture_memory(src_type, TypeAryPtr::BYTES); | 
|---|
| 4039 | StrCompressedCopyNode* str = new StrCompressedCopyNode(control(), mem, src, dst, count); | 
|---|
| 4040 | Node* res_mem = _gvn.transform(new SCMemProjNode(str)); | 
|---|
| 4041 | set_memory(res_mem, TypeAryPtr::BYTES); | 
|---|
| 4042 | return str; | 
|---|
| 4043 | } | 
|---|
| 4044 |  | 
|---|
| 4045 | void GraphKit::inflate_string(Node* src, Node* dst, const TypeAryPtr* dst_type, Node* count) { | 
|---|
| 4046 | assert(Matcher::match_rule_supported(Op_StrInflatedCopy), "Intrinsic not supported"); | 
|---|
| 4047 | assert(dst_type == TypeAryPtr::BYTES || dst_type == TypeAryPtr::CHARS, "invalid dest type"); | 
|---|
| 4048 | // Capture src and dst memory (see comment in 'compress_string'). | 
|---|
| 4049 | Node* mem = capture_memory(TypeAryPtr::BYTES, dst_type); | 
|---|
| 4050 | StrInflatedCopyNode* str = new StrInflatedCopyNode(control(), mem, src, dst, count); | 
|---|
| 4051 | set_memory(_gvn.transform(str), dst_type); | 
|---|
| 4052 | } | 
|---|
| 4053 |  | 
|---|
| 4054 | void GraphKit::inflate_string_slow(Node* src, Node* dst, Node* start, Node* count) { | 
|---|
| 4055 | /** | 
|---|
| 4056 | * int i_char = start; | 
|---|
| 4057 | * for (int i_byte = 0; i_byte < count; i_byte++) { | 
|---|
| 4058 | *   dst[i_char++] = (char)(src[i_byte] & 0xff); | 
|---|
| 4059 | * } | 
|---|
| 4060 | */ | 
|---|
| 4061 | src = access_resolve(src, ACCESS_READ); | 
|---|
| 4062 | dst = access_resolve(dst, ACCESS_WRITE); | 
|---|
| 4063 | add_predicate(); | 
|---|
| 4064 | RegionNode* head = new RegionNode(3); | 
|---|
| 4065 | head->init_req(1, control()); | 
|---|
| 4066 | gvn().set_type(head, Type::CONTROL); | 
|---|
| 4067 | record_for_igvn(head); | 
|---|
| 4068 |  | 
|---|
| 4069 | Node* i_byte = new PhiNode(head, TypeInt::INT); | 
|---|
| 4070 | i_byte->init_req(1, intcon(0)); | 
|---|
| 4071 | gvn().set_type(i_byte, TypeInt::INT); | 
|---|
| 4072 | record_for_igvn(i_byte); | 
|---|
| 4073 |  | 
|---|
| 4074 | Node* i_char = new PhiNode(head, TypeInt::INT); | 
|---|
| 4075 | i_char->init_req(1, start); | 
|---|
| 4076 | gvn().set_type(i_char, TypeInt::INT); | 
|---|
| 4077 | record_for_igvn(i_char); | 
|---|
| 4078 |  | 
|---|
| 4079 | Node* mem = PhiNode::make(head, memory(TypeAryPtr::BYTES), Type::MEMORY, TypeAryPtr::BYTES); | 
|---|
| 4080 | gvn().set_type(mem, Type::MEMORY); | 
|---|
| 4081 | record_for_igvn(mem); | 
|---|
| 4082 | set_control(head); | 
|---|
| 4083 | set_memory(mem, TypeAryPtr::BYTES); | 
|---|
| 4084 | Node* ch = load_array_element(control(), src, i_byte, TypeAryPtr::BYTES); | 
|---|
| 4085 | Node* st = store_to_memory(control(), array_element_address(dst, i_char, T_BYTE), | 
|---|
| 4086 | AndI(ch, intcon(0xff)), T_CHAR, TypeAryPtr::BYTES, MemNode::unordered, | 
|---|
| 4087 | false, false, true /* mismatched */); | 
|---|
| 4088 |  | 
|---|
| 4089 | IfNode* iff = create_and_map_if(head, Bool(CmpI(i_byte, count), BoolTest::lt), PROB_FAIR, COUNT_UNKNOWN); | 
|---|
| 4090 | head->init_req(2, IfTrue(iff)); | 
|---|
| 4091 | mem->init_req(2, st); | 
|---|
| 4092 | i_byte->init_req(2, AddI(i_byte, intcon(1))); | 
|---|
| 4093 | i_char->init_req(2, AddI(i_char, intcon(2))); | 
|---|
| 4094 |  | 
|---|
| 4095 | set_control(IfFalse(iff)); | 
|---|
| 4096 | set_memory(st, TypeAryPtr::BYTES); | 
|---|
| 4097 | } | 
|---|
| 4098 |  | 
|---|
| 4099 | Node* GraphKit::make_constant_from_field(ciField* field, Node* obj) { | 
|---|
| 4100 | if (!field->is_constant()) { | 
|---|
| 4101 | return NULL; // Field not marked as constant. | 
|---|
| 4102 | } | 
|---|
| 4103 | ciInstance* holder = NULL; | 
|---|
| 4104 | if (!field->is_static()) { | 
|---|
| 4105 | ciObject* const_oop = obj->bottom_type()->is_oopptr()->const_oop(); | 
|---|
| 4106 | if (const_oop != NULL && const_oop->is_instance()) { | 
|---|
| 4107 | holder = const_oop->as_instance(); | 
|---|
| 4108 | } | 
|---|
| 4109 | } | 
|---|
| 4110 | const Type* con_type = Type::make_constant_from_field(field, holder, field->layout_type(), | 
|---|
| 4111 | /*is_unsigned_load=*/false); | 
|---|
| 4112 | if (con_type != NULL) { | 
|---|
| 4113 | return makecon(con_type); | 
|---|
| 4114 | } | 
|---|
| 4115 | return NULL; | 
|---|
| 4116 | } | 
|---|
| 4117 |  | 
|---|