1 | /* |
2 | * Copyright (c) 1997, 2019, Oracle and/or its affiliates. All rights reserved. |
3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
4 | * |
5 | * This code is free software; you can redistribute it and/or modify it |
6 | * under the terms of the GNU General Public License version 2 only, as |
7 | * published by the Free Software Foundation. |
8 | * |
9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
12 | * version 2 for more details (a copy is included in the LICENSE file that |
13 | * accompanied this code). |
14 | * |
15 | * You should have received a copy of the GNU General Public License version |
16 | * 2 along with this work; if not, write to the Free Software Foundation, |
17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
18 | * |
19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
20 | * or visit www.oracle.com if you need additional information or have any |
21 | * questions. |
22 | * |
23 | */ |
24 | |
25 | #ifndef SHARE_OPTO_PARSE_HPP |
26 | #define SHARE_OPTO_PARSE_HPP |
27 | |
28 | #include "ci/ciMethodData.hpp" |
29 | #include "ci/ciTypeFlow.hpp" |
30 | #include "compiler/methodLiveness.hpp" |
31 | #include "libadt/vectset.hpp" |
32 | #include "oops/generateOopMap.hpp" |
33 | #include "opto/graphKit.hpp" |
34 | #include "opto/subnode.hpp" |
35 | |
36 | class BytecodeParseHistogram; |
37 | class InlineTree; |
38 | class Parse; |
39 | class SwitchRange; |
40 | |
41 | |
42 | //------------------------------InlineTree------------------------------------- |
43 | class InlineTree : public ResourceObj { |
44 | friend class VMStructs; |
45 | |
46 | Compile* C; // cache |
47 | JVMState* _caller_jvms; // state of caller |
48 | ciMethod* _method; // method being called by the caller_jvms |
49 | InlineTree* _caller_tree; |
50 | uint _count_inline_bcs; // Accumulated count of inlined bytecodes |
51 | // Call-site count / interpreter invocation count, scaled recursively. |
52 | // Always between 0.0 and 1.0. Represents the percentage of the method's |
53 | // total execution time used at this call site. |
54 | const float _site_invoke_ratio; |
55 | const int _max_inline_level; // the maximum inline level for this sub-tree (may be adjusted) |
56 | float compute_callee_frequency( int caller_bci ) const; |
57 | |
58 | GrowableArray<InlineTree*> _subtrees; |
59 | |
60 | bool pass_initial_checks(ciMethod* caller_method, int caller_bci, ciMethod* callee_method); |
61 | |
62 | void print_impl(outputStream* stj, int indent) const PRODUCT_RETURN; |
63 | const char* _msg; |
64 | protected: |
65 | InlineTree(Compile* C, |
66 | const InlineTree* caller_tree, |
67 | ciMethod* callee_method, |
68 | JVMState* caller_jvms, |
69 | int caller_bci, |
70 | float site_invoke_ratio, |
71 | int max_inline_level); |
72 | InlineTree *build_inline_tree_for_callee(ciMethod* callee_method, |
73 | JVMState* caller_jvms, |
74 | int caller_bci); |
75 | bool try_to_inline(ciMethod* callee_method, |
76 | ciMethod* caller_method, |
77 | int caller_bci, |
78 | JVMState* jvms, |
79 | ciCallProfile& profile, |
80 | WarmCallInfo* wci_result, |
81 | bool& should_delay); |
82 | bool should_inline(ciMethod* callee_method, |
83 | ciMethod* caller_method, |
84 | int caller_bci, |
85 | ciCallProfile& profile, |
86 | WarmCallInfo* wci_result); |
87 | bool should_not_inline(ciMethod* callee_method, |
88 | ciMethod* caller_method, |
89 | JVMState* jvms, |
90 | WarmCallInfo* wci_result); |
91 | bool is_not_reached(ciMethod* callee_method, |
92 | ciMethod* caller_method, |
93 | int caller_bci, |
94 | ciCallProfile& profile); |
95 | void print_inlining(ciMethod* callee_method, int caller_bci, |
96 | ciMethod* caller_method, bool success) const; |
97 | |
98 | InlineTree* caller_tree() const { return _caller_tree; } |
99 | InlineTree* callee_at(int bci, ciMethod* m) const; |
100 | int inline_level() const { return stack_depth(); } |
101 | int stack_depth() const { return _caller_jvms ? _caller_jvms->depth() : 0; } |
102 | const char* msg() const { return _msg; } |
103 | void set_msg(const char* msg) { _msg = msg; } |
104 | public: |
105 | static const char* check_can_parse(ciMethod* callee); |
106 | |
107 | static InlineTree* build_inline_tree_root(); |
108 | static InlineTree* find_subtree_from_root(InlineTree* root, JVMState* jvms, ciMethod* callee); |
109 | |
110 | // For temporary (stack-allocated, stateless) ilts: |
111 | InlineTree(Compile* c, ciMethod* callee_method, JVMState* caller_jvms, float site_invoke_ratio, int max_inline_level); |
112 | |
113 | // See if it is OK to inline. |
114 | // The receiver is the inline tree for the caller. |
115 | // |
116 | // The result is a temperature indication. If it is hot or cold, |
117 | // inlining is immediate or undesirable. Otherwise, the info block |
118 | // returned is newly allocated and may be enqueued. |
119 | // |
120 | // If the method is inlinable, a new inline subtree is created on the fly, |
121 | // and may be accessed by find_subtree_from_root. |
122 | // The call_method is the dest_method for a special or static invocation. |
123 | // The call_method is an optimized virtual method candidate otherwise. |
124 | WarmCallInfo* ok_to_inline(ciMethod *call_method, JVMState* caller_jvms, ciCallProfile& profile, WarmCallInfo* wci, bool& should_delay); |
125 | |
126 | // Information about inlined method |
127 | JVMState* caller_jvms() const { return _caller_jvms; } |
128 | ciMethod *method() const { return _method; } |
129 | int caller_bci() const { return _caller_jvms ? _caller_jvms->bci() : InvocationEntryBci; } |
130 | uint count_inline_bcs() const { return _count_inline_bcs; } |
131 | float site_invoke_ratio() const { return _site_invoke_ratio; }; |
132 | |
133 | #ifndef PRODUCT |
134 | private: |
135 | uint _count_inlines; // Count of inlined methods |
136 | public: |
137 | // Debug information collected during parse |
138 | uint count_inlines() const { return _count_inlines; }; |
139 | #endif |
140 | GrowableArray<InlineTree*> subtrees() { return _subtrees; } |
141 | |
142 | void print_value_on(outputStream* st) const PRODUCT_RETURN; |
143 | |
144 | bool _forced_inline; // Inlining was forced by CompilerOracle, ciReplay or annotation |
145 | bool forced_inline() const { return _forced_inline; } |
146 | // Count number of nodes in this subtree |
147 | int count() const; |
148 | // Dump inlining replay data to the stream. |
149 | void dump_replay_data(outputStream* out); |
150 | }; |
151 | |
152 | |
153 | //----------------------------------------------------------------------------- |
154 | //------------------------------Parse------------------------------------------ |
155 | // Parse bytecodes, build a Graph |
156 | class Parse : public GraphKit { |
157 | public: |
158 | // Per-block information needed by the parser: |
159 | class Block { |
160 | private: |
161 | ciTypeFlow::Block* _flow; |
162 | int _pred_count; // how many predecessors in CFG? |
163 | int _preds_parsed; // how many of these have been parsed? |
164 | uint _count; // how many times executed? Currently only set by _goto's |
165 | bool _is_parsed; // has this block been parsed yet? |
166 | bool _is_handler; // is this block an exception handler? |
167 | bool _has_merged_backedge; // does this block have merged backedge? |
168 | SafePointNode* _start_map; // all values flowing into this block |
169 | MethodLivenessResult _live_locals; // lazily initialized liveness bitmap |
170 | bool _has_predicates; // Were predicates added before parsing of the loop head? |
171 | |
172 | int _num_successors; // Includes only normal control flow. |
173 | int _all_successors; // Include exception paths also. |
174 | Block** _successors; |
175 | |
176 | public: |
177 | |
178 | // Set up the block data structure itself. |
179 | Block(Parse* outer, int rpo); |
180 | |
181 | // Set up the block's relations to other blocks. |
182 | void init_graph(Parse* outer); |
183 | |
184 | ciTypeFlow::Block* flow() const { return _flow; } |
185 | int pred_count() const { return _pred_count; } |
186 | int preds_parsed() const { return _preds_parsed; } |
187 | bool is_parsed() const { return _is_parsed; } |
188 | bool is_handler() const { return _is_handler; } |
189 | void set_count( uint x ) { _count = x; } |
190 | uint count() const { return _count; } |
191 | |
192 | SafePointNode* start_map() const { assert(is_merged(),"" ); return _start_map; } |
193 | void set_start_map(SafePointNode* m) { assert(!is_merged(), "" ); _start_map = m; } |
194 | |
195 | // True after any predecessor flows control into this block |
196 | bool is_merged() const { return _start_map != NULL; } |
197 | |
198 | #ifdef ASSERT |
199 | // True after backedge predecessor flows control into this block |
200 | bool has_merged_backedge() const { return _has_merged_backedge; } |
201 | void mark_merged_backedge(Block* pred) { |
202 | assert(is_SEL_head(), "should be loop head" ); |
203 | if (pred != NULL && is_SEL_backedge(pred)) { |
204 | assert(is_parsed(), "block should be parsed before merging backedges" ); |
205 | _has_merged_backedge = true; |
206 | } |
207 | } |
208 | #endif |
209 | |
210 | // True when all non-exception predecessors have been parsed. |
211 | bool is_ready() const { return preds_parsed() == pred_count(); } |
212 | |
213 | bool has_predicates() const { return _has_predicates; } |
214 | void set_has_predicates() { _has_predicates = true; } |
215 | |
216 | int num_successors() const { return _num_successors; } |
217 | int all_successors() const { return _all_successors; } |
218 | Block* successor_at(int i) const { |
219 | assert((uint)i < (uint)all_successors(), "" ); |
220 | return _successors[i]; |
221 | } |
222 | Block* successor_for_bci(int bci); |
223 | |
224 | int start() const { return flow()->start(); } |
225 | int limit() const { return flow()->limit(); } |
226 | int rpo() const { return flow()->rpo(); } |
227 | int start_sp() const { return flow()->stack_size(); } |
228 | |
229 | bool is_loop_head() const { return flow()->is_loop_head(); } |
230 | bool is_SEL_head() const { return flow()->is_single_entry_loop_head(); } |
231 | bool is_SEL_backedge(Block* pred) const{ return is_SEL_head() && pred->rpo() >= rpo(); } |
232 | bool is_invariant_local(uint i) const { |
233 | const JVMState* jvms = start_map()->jvms(); |
234 | if (!jvms->is_loc(i) || flow()->outer()->has_irreducible_entry()) return false; |
235 | return flow()->is_invariant_local(i - jvms->locoff()); |
236 | } |
237 | bool can_elide_SEL_phi(uint i) const { assert(is_SEL_head(),"" ); return is_invariant_local(i); } |
238 | |
239 | const Type* peek(int off=0) const { return stack_type_at(start_sp() - (off+1)); } |
240 | |
241 | const Type* stack_type_at(int i) const; |
242 | const Type* local_type_at(int i) const; |
243 | static const Type* get_type(ciType* t) { return Type::get_typeflow_type(t); } |
244 | |
245 | bool has_trap_at(int bci) const { return flow()->has_trap() && flow()->trap_bci() == bci; } |
246 | |
247 | // Call this just before parsing a block. |
248 | void mark_parsed() { |
249 | assert(!_is_parsed, "must parse each block exactly once" ); |
250 | _is_parsed = true; |
251 | } |
252 | |
253 | // Return the phi/region input index for the "current" pred, |
254 | // and bump the pred number. For historical reasons these index |
255 | // numbers are handed out in descending order. The last index is |
256 | // always PhiNode::Input (i.e., 1). The value returned is known |
257 | // as a "path number" because it distinguishes by which path we are |
258 | // entering the block. |
259 | int next_path_num() { |
260 | assert(preds_parsed() < pred_count(), "too many preds?" ); |
261 | return pred_count() - _preds_parsed++; |
262 | } |
263 | |
264 | // Add a previously unaccounted predecessor to this block. |
265 | // This operates by increasing the size of the block's region |
266 | // and all its phi nodes (if any). The value returned is a |
267 | // path number ("pnum"). |
268 | int add_new_path(); |
269 | |
270 | // Initialize me by recording the parser's map. My own map must be NULL. |
271 | void record_state(Parse* outer); |
272 | }; |
273 | |
274 | #ifndef PRODUCT |
275 | // BytecodeParseHistogram collects number of bytecodes parsed, nodes constructed, and transformations. |
276 | class BytecodeParseHistogram : public ResourceObj { |
277 | private: |
278 | enum BPHType { |
279 | BPH_transforms, |
280 | BPH_values |
281 | }; |
282 | static bool _initialized; |
283 | static uint _bytecodes_parsed [Bytecodes::number_of_codes]; |
284 | static uint _nodes_constructed[Bytecodes::number_of_codes]; |
285 | static uint _nodes_transformed[Bytecodes::number_of_codes]; |
286 | static uint _new_values [Bytecodes::number_of_codes]; |
287 | |
288 | Bytecodes::Code _initial_bytecode; |
289 | int _initial_node_count; |
290 | int _initial_transforms; |
291 | int _initial_values; |
292 | |
293 | Parse *_parser; |
294 | Compile *_compiler; |
295 | |
296 | // Initialization |
297 | static void reset(); |
298 | |
299 | // Return info being collected, select with global flag 'BytecodeParseInfo' |
300 | int current_count(BPHType info_selector); |
301 | |
302 | public: |
303 | BytecodeParseHistogram(Parse *p, Compile *c); |
304 | static bool initialized(); |
305 | |
306 | // Record info when starting to parse one bytecode |
307 | void set_initial_state( Bytecodes::Code bc ); |
308 | // Record results of parsing one bytecode |
309 | void record_change(); |
310 | |
311 | // Profile printing |
312 | static void print(float cutoff = 0.01F); // cutoff in percent |
313 | }; |
314 | |
315 | public: |
316 | // Record work done during parsing |
317 | BytecodeParseHistogram* _parse_histogram; |
318 | void set_parse_histogram(BytecodeParseHistogram *bph) { _parse_histogram = bph; } |
319 | BytecodeParseHistogram* parse_histogram() { return _parse_histogram; } |
320 | #endif |
321 | |
322 | private: |
323 | friend class Block; |
324 | |
325 | // Variables which characterize this compilation as a whole: |
326 | |
327 | JVMState* _caller; // JVMS which carries incoming args & state. |
328 | float _expected_uses; // expected number of calls to this code |
329 | float _prof_factor; // discount applied to my profile counts |
330 | int _depth; // Inline tree depth, for debug printouts |
331 | const TypeFunc*_tf; // My kind of function type |
332 | int _entry_bci; // the osr bci or InvocationEntryBci |
333 | |
334 | ciTypeFlow* _flow; // Results of previous flow pass. |
335 | Block* _blocks; // Array of basic-block structs. |
336 | int _block_count; // Number of elements in _blocks. |
337 | |
338 | GraphKit _exits; // Record all normal returns and throws here. |
339 | bool _wrote_final; // Did we write a final field? |
340 | bool _wrote_volatile; // Did we write a volatile field? |
341 | bool _wrote_stable; // Did we write a @Stable field? |
342 | bool _wrote_fields; // Did we write any field? |
343 | bool _count_invocations; // update and test invocation counter |
344 | bool _method_data_update; // update method data oop |
345 | Node* _alloc_with_final; // An allocation node with final field |
346 | |
347 | // Variables which track Java semantics during bytecode parsing: |
348 | |
349 | Block* _block; // block currently getting parsed |
350 | ciBytecodeStream _iter; // stream of this method's bytecodes |
351 | |
352 | const FastLockNode* _synch_lock; // FastLockNode for synchronized method |
353 | |
354 | #ifndef PRODUCT |
355 | int _max_switch_depth; // Debugging SwitchRanges. |
356 | int _est_switch_depth; // Debugging SwitchRanges. |
357 | #endif |
358 | |
359 | bool _first_return; // true if return is the first to be parsed |
360 | bool _replaced_nodes_for_exceptions; // needs processing of replaced nodes in exception paths? |
361 | uint _new_idx; // any node with _idx above were new during this parsing. Used to trim the replaced nodes list. |
362 | |
363 | public: |
364 | // Constructor |
365 | Parse(JVMState* caller, ciMethod* parse_method, float expected_uses); |
366 | |
367 | virtual Parse* is_Parse() const { return (Parse*)this; } |
368 | |
369 | // Accessors. |
370 | JVMState* caller() const { return _caller; } |
371 | float expected_uses() const { return _expected_uses; } |
372 | float prof_factor() const { return _prof_factor; } |
373 | int depth() const { return _depth; } |
374 | const TypeFunc* tf() const { return _tf; } |
375 | // entry_bci() -- see osr_bci, etc. |
376 | |
377 | ciTypeFlow* flow() const { return _flow; } |
378 | // blocks() -- see rpo_at, start_block, etc. |
379 | int block_count() const { return _block_count; } |
380 | |
381 | GraphKit& exits() { return _exits; } |
382 | bool wrote_final() const { return _wrote_final; } |
383 | void set_wrote_final(bool z) { _wrote_final = z; } |
384 | bool wrote_volatile() const { return _wrote_volatile; } |
385 | void set_wrote_volatile(bool z) { _wrote_volatile = z; } |
386 | bool wrote_stable() const { return _wrote_stable; } |
387 | void set_wrote_stable(bool z) { _wrote_stable = z; } |
388 | bool wrote_fields() const { return _wrote_fields; } |
389 | void set_wrote_fields(bool z) { _wrote_fields = z; } |
390 | bool count_invocations() const { return _count_invocations; } |
391 | bool method_data_update() const { return _method_data_update; } |
392 | Node* alloc_with_final() const { return _alloc_with_final; } |
393 | void set_alloc_with_final(Node* n) { |
394 | assert((_alloc_with_final == NULL) || (_alloc_with_final == n), "different init objects?" ); |
395 | _alloc_with_final = n; |
396 | } |
397 | |
398 | Block* block() const { return _block; } |
399 | ciBytecodeStream& iter() { return _iter; } |
400 | Bytecodes::Code bc() const { return _iter.cur_bc(); } |
401 | |
402 | void set_block(Block* b) { _block = b; } |
403 | |
404 | // Derived accessors: |
405 | bool is_normal_parse() const { return _entry_bci == InvocationEntryBci; } |
406 | bool is_osr_parse() const { return _entry_bci != InvocationEntryBci; } |
407 | int osr_bci() const { assert(is_osr_parse(),"" ); return _entry_bci; } |
408 | |
409 | void set_parse_bci(int bci); |
410 | |
411 | // Must this parse be aborted? |
412 | bool failing() { return C->failing(); } |
413 | |
414 | Block* rpo_at(int rpo) { |
415 | assert(0 <= rpo && rpo < _block_count, "oob" ); |
416 | return &_blocks[rpo]; |
417 | } |
418 | Block* start_block() { |
419 | return rpo_at(flow()->start_block()->rpo()); |
420 | } |
421 | // Can return NULL if the flow pass did not complete a block. |
422 | Block* successor_for_bci(int bci) { |
423 | return block()->successor_for_bci(bci); |
424 | } |
425 | |
426 | private: |
427 | // Create a JVMS & map for the initial state of this method. |
428 | SafePointNode* create_entry_map(); |
429 | |
430 | // OSR helpers |
431 | Node *fetch_interpreter_state(int index, BasicType bt, Node *local_addrs, Node *local_addrs_base); |
432 | Node* check_interpreter_type(Node* l, const Type* type, SafePointNode* &bad_type_exit); |
433 | void load_interpreter_state(Node* osr_buf); |
434 | |
435 | // Functions for managing basic blocks: |
436 | void init_blocks(); |
437 | void load_state_from(Block* b); |
438 | void store_state_to(Block* b) { b->record_state(this); } |
439 | |
440 | // Parse all the basic blocks. |
441 | void do_all_blocks(); |
442 | |
443 | // Parse the current basic block |
444 | void do_one_block(); |
445 | |
446 | // Raise an error if we get a bad ciTypeFlow CFG. |
447 | void handle_missing_successor(int bci); |
448 | |
449 | // first actions (before BCI 0) |
450 | void do_method_entry(); |
451 | |
452 | // implementation of monitorenter/monitorexit |
453 | void do_monitor_enter(); |
454 | void do_monitor_exit(); |
455 | |
456 | // Eagerly create phie throughout the state, to cope with back edges. |
457 | void ensure_phis_everywhere(); |
458 | |
459 | // Merge the current mapping into the basic block starting at bci |
460 | void merge( int target_bci); |
461 | // Same as plain merge, except that it allocates a new path number. |
462 | void merge_new_path( int target_bci); |
463 | // Merge the current mapping into an exception handler. |
464 | void merge_exception(int target_bci); |
465 | // Helper: Merge the current mapping into the given basic block |
466 | void merge_common(Block* target, int pnum); |
467 | // Helper functions for merging individual cells. |
468 | PhiNode *ensure_phi( int idx, bool nocreate = false); |
469 | PhiNode *ensure_memory_phi(int idx, bool nocreate = false); |
470 | // Helper to merge the current memory state into the given basic block |
471 | void merge_memory_edges(MergeMemNode* n, int pnum, bool nophi); |
472 | |
473 | // Parse this bytecode, and alter the Parsers JVM->Node mapping |
474 | void do_one_bytecode(); |
475 | |
476 | // helper function to generate array store check |
477 | void array_store_check(); |
478 | // Helper function to generate array load |
479 | void array_load(BasicType etype); |
480 | // Helper function to generate array store |
481 | void array_store(BasicType etype); |
482 | // Helper function to compute array addressing |
483 | Node* array_addressing(BasicType type, int vals, const Type* *result2=NULL); |
484 | |
485 | void clinit_deopt(); |
486 | |
487 | void rtm_deopt(); |
488 | |
489 | // Pass current map to exits |
490 | void return_current(Node* value); |
491 | |
492 | // Register finalizers on return from Object.<init> |
493 | void call_register_finalizer(); |
494 | |
495 | // Insert a compiler safepoint into the graph |
496 | void add_safepoint(); |
497 | |
498 | // Insert a compiler safepoint into the graph, if there is a back-branch. |
499 | void maybe_add_safepoint(int target_bci) { |
500 | if (UseLoopSafepoints && target_bci <= bci()) { |
501 | add_safepoint(); |
502 | } |
503 | } |
504 | |
505 | // Note: Intrinsic generation routines may be found in library_call.cpp. |
506 | |
507 | // Helper function to setup Ideal Call nodes |
508 | void do_call(); |
509 | |
510 | // Helper function to uncommon-trap or bailout for non-compilable call-sites |
511 | bool can_not_compile_call_site(ciMethod *dest_method, ciInstanceKlass *klass); |
512 | |
513 | // Helper function to setup for type-profile based inlining |
514 | bool prepare_type_profile_inline(ciInstanceKlass* prof_klass, ciMethod* prof_method); |
515 | |
516 | // Helper functions for type checking bytecodes: |
517 | void do_checkcast(); |
518 | void do_instanceof(); |
519 | |
520 | // Helper functions for shifting & arithmetic |
521 | void modf(); |
522 | void modd(); |
523 | void l2f(); |
524 | |
525 | void do_irem(); |
526 | |
527 | // implementation of _get* and _put* bytecodes |
528 | void do_getstatic() { do_field_access(true, false); } |
529 | void do_getfield () { do_field_access(true, true); } |
530 | void do_putstatic() { do_field_access(false, false); } |
531 | void do_putfield () { do_field_access(false, true); } |
532 | |
533 | // common code for making initial checks and forming addresses |
534 | void do_field_access(bool is_get, bool is_field); |
535 | |
536 | // common code for actually performing the load or store |
537 | void do_get_xxx(Node* obj, ciField* field, bool is_field); |
538 | void do_put_xxx(Node* obj, ciField* field, bool is_field); |
539 | |
540 | // implementation of object creation bytecodes |
541 | void do_new(); |
542 | void do_newarray(BasicType elemtype); |
543 | void do_anewarray(); |
544 | void do_multianewarray(); |
545 | Node* expand_multianewarray(ciArrayKlass* array_klass, Node* *lengths, int ndimensions, int nargs); |
546 | |
547 | // implementation of jsr/ret |
548 | void do_jsr(); |
549 | void do_ret(); |
550 | |
551 | float dynamic_branch_prediction(float &cnt, BoolTest::mask btest, Node* test); |
552 | float branch_prediction(float &cnt, BoolTest::mask btest, int target_bci, Node* test); |
553 | bool seems_never_taken(float prob) const; |
554 | bool path_is_suitable_for_uncommon_trap(float prob) const; |
555 | bool seems_stable_comparison() const; |
556 | |
557 | void do_ifnull(BoolTest::mask btest, Node* c); |
558 | void do_if(BoolTest::mask btest, Node* c); |
559 | int repush_if_args(); |
560 | void adjust_map_after_if(BoolTest::mask btest, Node* c, float prob, |
561 | Block* path, Block* other_path); |
562 | void sharpen_type_after_if(BoolTest::mask btest, |
563 | Node* con, const Type* tcon, |
564 | Node* val, const Type* tval); |
565 | void maybe_add_predicate_after_if(Block* path); |
566 | IfNode* jump_if_fork_int(Node* a, Node* b, BoolTest::mask mask, float prob, float cnt); |
567 | Node* jump_if_join(Node* iffalse, Node* iftrue); |
568 | void jump_if_true_fork(IfNode *ifNode, int dest_bci_if_true, int prof_table_index, bool unc); |
569 | void jump_if_false_fork(IfNode *ifNode, int dest_bci_if_false, int prof_table_index, bool unc); |
570 | void jump_if_always_fork(int dest_bci_if_true, int prof_table_index, bool unc); |
571 | |
572 | friend class SwitchRange; |
573 | void do_tableswitch(); |
574 | void do_lookupswitch(); |
575 | void jump_switch_ranges(Node* a, SwitchRange* lo, SwitchRange* hi, int depth = 0); |
576 | bool create_jump_tables(Node* a, SwitchRange* lo, SwitchRange* hi); |
577 | void linear_search_switch_ranges(Node* key_val, SwitchRange*& lo, SwitchRange*& hi); |
578 | |
579 | void decrement_age(); |
580 | // helper functions for methodData style profiling |
581 | void test_counter_against_threshold(Node* cnt, int limit); |
582 | void increment_and_test_invocation_counter(int limit); |
583 | void test_for_osr_md_counter_at(ciMethodData* md, ciProfileData* data, ByteSize offset, int limit); |
584 | Node* method_data_addressing(ciMethodData* md, ciProfileData* data, ByteSize offset, Node* idx = NULL, uint stride = 0); |
585 | void increment_md_counter_at(ciMethodData* md, ciProfileData* data, ByteSize offset, Node* idx = NULL, uint stride = 0); |
586 | void set_md_flag_at(ciMethodData* md, ciProfileData* data, int flag_constant); |
587 | |
588 | void profile_method_entry(); |
589 | void profile_taken_branch(int target_bci, bool force_update = false); |
590 | void profile_not_taken_branch(bool force_update = false); |
591 | void profile_call(Node* receiver); |
592 | void profile_generic_call(); |
593 | void profile_receiver_type(Node* receiver); |
594 | void profile_ret(int target_bci); |
595 | void profile_null_checkcast(); |
596 | void profile_switch_case(int table_index); |
597 | |
598 | // helper function for call statistics |
599 | void count_compiled_calls(bool at_method_entry, bool is_inline) PRODUCT_RETURN; |
600 | |
601 | Node_Notes* make_node_notes(Node_Notes* caller_nn); |
602 | |
603 | // Helper functions for handling normal and abnormal exits. |
604 | void build_exits(); |
605 | |
606 | // Fix up all exceptional control flow exiting a single bytecode. |
607 | void do_exceptions(); |
608 | |
609 | // Fix up all exiting control flow at the end of the parse. |
610 | void do_exits(); |
611 | |
612 | // Add Catch/CatchProjs |
613 | // The call is either a Java call or the VM's rethrow stub |
614 | void catch_call_exceptions(ciExceptionHandlerStream&); |
615 | |
616 | // Handle all exceptions thrown by the inlined method. |
617 | // Also handles exceptions for individual bytecodes. |
618 | void catch_inline_exceptions(SafePointNode* ex_map); |
619 | |
620 | // Merge the given map into correct exceptional exit state. |
621 | // Assumes that there is no applicable local handler. |
622 | void throw_to_exit(SafePointNode* ex_map); |
623 | |
624 | // Use speculative type to optimize CmpP node |
625 | Node* optimize_cmp_with_klass(Node* c); |
626 | |
627 | public: |
628 | #ifndef PRODUCT |
629 | // Handle PrintOpto, etc. |
630 | void show_parse_info(); |
631 | void dump_map_adr_mem() const; |
632 | static void print_statistics(); // Print some performance counters |
633 | void dump(); |
634 | void dump_bci(int bci); |
635 | #endif |
636 | }; |
637 | |
638 | #endif // SHARE_OPTO_PARSE_HPP |
639 | |