1 | /* |
2 | * Copyright (c) 2018, 2019, Oracle and/or its affiliates. All rights reserved. |
3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
4 | * |
5 | * This code is free software; you can redistribute it and/or modify it |
6 | * under the terms of the GNU General Public License version 2 only, as |
7 | * published by the Free Software Foundation. |
8 | * |
9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
12 | * version 2 for more details (a copy is included in the LICENSE file that |
13 | * accompanied this code). |
14 | * |
15 | * You should have received a copy of the GNU General Public License version |
16 | * 2 along with this work; if not, write to the Free Software Foundation, |
17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
18 | * |
19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
20 | * or visit www.oracle.com if you need additional information or have any |
21 | * questions. |
22 | * |
23 | */ |
24 | |
25 | #ifndef SHARE_GC_SHARED_C2_BARRIERSETC2_HPP |
26 | #define SHARE_GC_SHARED_C2_BARRIERSETC2_HPP |
27 | |
28 | #include "memory/allocation.hpp" |
29 | #include "oops/accessDecorators.hpp" |
30 | #include "opto/loopnode.hpp" |
31 | #include "opto/matcher.hpp" |
32 | #include "opto/memnode.hpp" |
33 | #include "utilities/globalDefinitions.hpp" |
34 | |
35 | // This means the access is mismatched. This means the value of an access |
36 | // is not equivalent to the value pointed to by the address. |
37 | const DecoratorSet C2_MISMATCHED = DECORATOR_LAST << 1; |
38 | // The access may not be aligned to its natural size. |
39 | const DecoratorSet C2_UNALIGNED = DECORATOR_LAST << 2; |
40 | // The atomic cmpxchg is weak, meaning that spurious false negatives are allowed, |
41 | // but never false positives. |
42 | const DecoratorSet C2_WEAK_CMPXCHG = DECORATOR_LAST << 3; |
43 | // This denotes that a load has control dependency. |
44 | const DecoratorSet C2_CONTROL_DEPENDENT_LOAD = DECORATOR_LAST << 4; |
45 | // This denotes that a load that must be pinned. |
46 | const DecoratorSet C2_PINNED_LOAD = DECORATOR_LAST << 5; |
47 | // This denotes that the access is produced from the sun.misc.Unsafe intrinsics. |
48 | const DecoratorSet C2_UNSAFE_ACCESS = DECORATOR_LAST << 6; |
49 | // This denotes that the access mutates state. |
50 | const DecoratorSet C2_WRITE_ACCESS = DECORATOR_LAST << 7; |
51 | // This denotes that the access reads state. |
52 | const DecoratorSet C2_READ_ACCESS = DECORATOR_LAST << 8; |
53 | // A nearby allocation? |
54 | const DecoratorSet C2_TIGHTLY_COUPLED_ALLOC = DECORATOR_LAST << 9; |
55 | // Loads and stores from an arraycopy being optimized |
56 | const DecoratorSet C2_ARRAY_COPY = DECORATOR_LAST << 10; |
57 | |
58 | class Compile; |
59 | class ConnectionGraph; |
60 | class GraphKit; |
61 | class IdealKit; |
62 | class Node; |
63 | class PhaseGVN; |
64 | class PhaseIdealLoop; |
65 | class PhaseMacroExpand; |
66 | class Type; |
67 | class TypePtr; |
68 | class Unique_Node_List; |
69 | |
70 | // This class wraps a node and a type. |
71 | class C2AccessValue: public StackObj { |
72 | protected: |
73 | Node* _node; |
74 | const Type* _type; |
75 | |
76 | public: |
77 | C2AccessValue(Node* node, const Type* type) : |
78 | _node(node), |
79 | _type(type) {} |
80 | |
81 | Node* node() const { return _node; } |
82 | const Type* type() const { return _type; } |
83 | |
84 | void set_node(Node* node) { _node = node; } |
85 | }; |
86 | |
87 | // This class wraps a node and a pointer type. |
88 | class C2AccessValuePtr: public C2AccessValue { |
89 | |
90 | public: |
91 | C2AccessValuePtr(Node* node, const TypePtr* type) : |
92 | C2AccessValue(node, reinterpret_cast<const Type*>(type)) {} |
93 | |
94 | const TypePtr* type() const { return reinterpret_cast<const TypePtr*>(_type); } |
95 | }; |
96 | |
97 | // This class wraps a bunch of context parameters thare are passed around in the |
98 | // BarrierSetC2 backend hierarchy, for loads and stores, to reduce boiler plate. |
99 | class C2Access: public StackObj { |
100 | protected: |
101 | DecoratorSet _decorators; |
102 | BasicType _type; |
103 | Node* _base; |
104 | C2AccessValuePtr& _addr; |
105 | Node* _raw_access; |
106 | |
107 | void fixup_decorators(); |
108 | |
109 | public: |
110 | C2Access(DecoratorSet decorators, |
111 | BasicType type, Node* base, C2AccessValuePtr& addr) : |
112 | _decorators(decorators), |
113 | _type(type), |
114 | _base(base), |
115 | _addr(addr), |
116 | _raw_access(NULL) |
117 | {} |
118 | |
119 | DecoratorSet decorators() const { return _decorators; } |
120 | Node* base() const { return _base; } |
121 | C2AccessValuePtr& addr() const { return _addr; } |
122 | BasicType type() const { return _type; } |
123 | bool is_oop() const { return _type == T_OBJECT || _type == T_ARRAY; } |
124 | bool is_raw() const { return (_decorators & AS_RAW) != 0; } |
125 | Node* raw_access() const { return _raw_access; } |
126 | |
127 | void set_raw_access(Node* raw_access) { _raw_access = raw_access; } |
128 | virtual void set_memory() {} // no-op for normal accesses, but not for atomic accesses. |
129 | |
130 | MemNode::MemOrd mem_node_mo() const; |
131 | bool needs_cpu_membar() const; |
132 | |
133 | virtual PhaseGVN& gvn() const = 0; |
134 | virtual bool is_parse_access() const { return false; } |
135 | virtual bool is_opt_access() const { return false; } |
136 | }; |
137 | |
138 | // C2Access for parse time calls to the BarrierSetC2 backend. |
139 | class C2ParseAccess: public C2Access { |
140 | protected: |
141 | GraphKit* _kit; |
142 | |
143 | void* barrier_set_state() const; |
144 | |
145 | public: |
146 | C2ParseAccess(GraphKit* kit, DecoratorSet decorators, |
147 | BasicType type, Node* base, C2AccessValuePtr& addr) : |
148 | C2Access(decorators, type, base, addr), |
149 | _kit(kit) { |
150 | fixup_decorators(); |
151 | } |
152 | |
153 | GraphKit* kit() const { return _kit; } |
154 | |
155 | template <typename T> |
156 | T barrier_set_state_as() const { |
157 | return reinterpret_cast<T>(barrier_set_state()); |
158 | } |
159 | |
160 | virtual PhaseGVN& gvn() const; |
161 | virtual bool is_parse_access() const { return true; } |
162 | }; |
163 | |
164 | // This class wraps a bunch of context parameters thare are passed around in the |
165 | // BarrierSetC2 backend hierarchy, for atomic accesses, to reduce boiler plate. |
166 | class C2AtomicParseAccess: public C2ParseAccess { |
167 | Node* _memory; |
168 | uint _alias_idx; |
169 | bool _needs_pinning; |
170 | |
171 | public: |
172 | C2AtomicParseAccess(GraphKit* kit, DecoratorSet decorators, BasicType type, |
173 | Node* base, C2AccessValuePtr& addr, uint alias_idx) : |
174 | C2ParseAccess(kit, decorators, type, base, addr), |
175 | _memory(NULL), |
176 | _alias_idx(alias_idx), |
177 | _needs_pinning(true) {} |
178 | |
179 | // Set the memory node based on the current memory slice. |
180 | virtual void set_memory(); |
181 | |
182 | Node* memory() const { return _memory; } |
183 | uint alias_idx() const { return _alias_idx; } |
184 | bool needs_pinning() const { return _needs_pinning; } |
185 | |
186 | void set_needs_pinning(bool value) { _needs_pinning = value; } |
187 | }; |
188 | |
189 | // C2Access for optimization time calls to the BarrierSetC2 backend. |
190 | class C2OptAccess: public C2Access { |
191 | PhaseGVN& _gvn; |
192 | MergeMemNode* _mem; |
193 | Node* _ctl; |
194 | |
195 | public: |
196 | C2OptAccess(PhaseGVN& gvn, Node* ctl, MergeMemNode* mem, DecoratorSet decorators, |
197 | BasicType type, Node* base, C2AccessValuePtr& addr) : |
198 | C2Access(decorators, type, base, addr), |
199 | _gvn(gvn), _mem(mem), _ctl(ctl) { |
200 | fixup_decorators(); |
201 | } |
202 | |
203 | |
204 | MergeMemNode* mem() const { return _mem; } |
205 | Node* ctl() const { return _ctl; } |
206 | // void set_mem(Node* mem) { _mem = mem; } |
207 | void set_ctl(Node* ctl) { _ctl = ctl; } |
208 | |
209 | virtual PhaseGVN& gvn() const { return _gvn; } |
210 | virtual bool is_opt_access() const { return true; } |
211 | }; |
212 | |
213 | |
214 | // This is the top-level class for the backend of the Access API in C2. |
215 | // The top-level class is responsible for performing raw accesses. The |
216 | // various GC barrier sets inherit from the BarrierSetC2 class to sprinkle |
217 | // barriers into the accesses. |
218 | class BarrierSetC2: public CHeapObj<mtGC> { |
219 | protected: |
220 | virtual void resolve_address(C2Access& access) const; |
221 | virtual Node* store_at_resolved(C2Access& access, C2AccessValue& val) const; |
222 | virtual Node* load_at_resolved(C2Access& access, const Type* val_type) const; |
223 | |
224 | virtual Node* atomic_cmpxchg_val_at_resolved(C2AtomicParseAccess& access, Node* expected_val, |
225 | Node* new_val, const Type* val_type) const; |
226 | virtual Node* atomic_cmpxchg_bool_at_resolved(C2AtomicParseAccess& access, Node* expected_val, |
227 | Node* new_val, const Type* value_type) const; |
228 | virtual Node* atomic_xchg_at_resolved(C2AtomicParseAccess& access, Node* new_val, const Type* val_type) const; |
229 | virtual Node* atomic_add_at_resolved(C2AtomicParseAccess& access, Node* new_val, const Type* val_type) const; |
230 | void pin_atomic_op(C2AtomicParseAccess& access) const; |
231 | |
232 | public: |
233 | // This is the entry-point for the backend to perform accesses through the Access API. |
234 | virtual Node* store_at(C2Access& access, C2AccessValue& val) const; |
235 | virtual Node* load_at(C2Access& access, const Type* val_type) const; |
236 | |
237 | virtual Node* atomic_cmpxchg_val_at(C2AtomicParseAccess& access, Node* expected_val, |
238 | Node* new_val, const Type* val_type) const; |
239 | virtual Node* atomic_cmpxchg_bool_at(C2AtomicParseAccess& access, Node* expected_val, |
240 | Node* new_val, const Type* val_type) const; |
241 | virtual Node* atomic_xchg_at(C2AtomicParseAccess& access, Node* new_val, const Type* value_type) const; |
242 | virtual Node* atomic_add_at(C2AtomicParseAccess& access, Node* new_val, const Type* value_type) const; |
243 | |
244 | virtual void clone(GraphKit* kit, Node* src, Node* dst, Node* size, bool is_array) const; |
245 | |
246 | virtual Node* resolve(GraphKit* kit, Node* n, DecoratorSet decorators) const { return n; } |
247 | |
248 | virtual Node* obj_allocate(PhaseMacroExpand* macro, Node* ctrl, Node* mem, Node* toobig_false, Node* size_in_bytes, |
249 | Node*& i_o, Node*& needgc_ctrl, |
250 | Node*& fast_oop_ctrl, Node*& fast_oop_rawmem, |
251 | intx prefetch_lines) const; |
252 | |
253 | virtual Node* ideal_node(PhaseGVN* phase, Node* n, bool can_reshape) const { return NULL; } |
254 | virtual Node* identity_node(PhaseGVN* phase, Node* n) const { return n; } |
255 | |
256 | // These are general helper methods used by C2 |
257 | enum ArrayCopyPhase { |
258 | Parsing, |
259 | Optimization, |
260 | Expansion |
261 | }; |
262 | |
263 | virtual bool array_copy_requires_gc_barriers(bool tightly_coupled_alloc, BasicType type, bool is_clone, ArrayCopyPhase phase) const { return false; } |
264 | virtual void clone_barrier_at_expansion(ArrayCopyNode* ac, Node* call, PhaseIterGVN& igvn) const; |
265 | |
266 | // Support for GC barriers emitted during parsing |
267 | virtual bool has_load_barriers() const { return false; } |
268 | virtual bool is_gc_barrier_node(Node* node) const { return false; } |
269 | virtual Node* step_over_gc_barrier(Node* c) const { return c; } |
270 | virtual Node* step_over_gc_barrier_ctrl(Node* c) const { return c; } |
271 | |
272 | // Support for macro expanded GC barriers |
273 | virtual void register_potential_barrier_node(Node* node) const { } |
274 | virtual void unregister_potential_barrier_node(Node* node) const { } |
275 | virtual void eliminate_gc_barrier(PhaseMacroExpand* macro, Node* node) const { } |
276 | virtual void enqueue_useful_gc_barrier(PhaseIterGVN* igvn, Node* node) const {} |
277 | virtual void eliminate_useless_gc_barriers(Unique_Node_List &useful, Compile* C) const {} |
278 | |
279 | // Allow barrier sets to have shared state that is preserved across a compilation unit. |
280 | // This could for example comprise macro nodes to be expanded during macro expansion. |
281 | virtual void* create_barrier_state(Arena* comp_arena) const { return NULL; } |
282 | // If the BarrierSetC2 state has barrier nodes in its compilation |
283 | // unit state to be expanded later, then now is the time to do so. |
284 | virtual bool expand_barriers(Compile* C, PhaseIterGVN& igvn) const { return false; } |
285 | virtual bool optimize_loops(PhaseIdealLoop* phase, LoopOptsMode mode, VectorSet& visited, Node_Stack& nstack, Node_List& worklist) const { return false; } |
286 | virtual bool strip_mined_loops_expanded(LoopOptsMode mode) const { return false; } |
287 | virtual bool is_gc_specific_loop_opts_pass(LoopOptsMode mode) const { return false; } |
288 | |
289 | virtual bool has_special_unique_user(const Node* node) const { return false; } |
290 | virtual bool needs_anti_dependence_check(const Node* node) const { return true; } |
291 | |
292 | virtual void barrier_insertion_phase(Compile* C, PhaseIterGVN &igvn) const { } |
293 | |
294 | enum CompilePhase { |
295 | BeforeOptimize, |
296 | BeforeLateInsertion, |
297 | BeforeMacroExpand, |
298 | BeforeCodeGen |
299 | }; |
300 | |
301 | virtual bool flatten_gc_alias_type(const TypePtr*& adr_type) const { return false; } |
302 | #ifdef ASSERT |
303 | virtual bool verify_gc_alias_type(const TypePtr* adr_type, int offset) const { return false; } |
304 | virtual void verify_gc_barriers(Compile* compile, CompilePhase phase) const {} |
305 | #endif |
306 | |
307 | virtual bool final_graph_reshaping(Compile* compile, Node* n, uint opcode) const { return false; } |
308 | |
309 | virtual bool escape_add_to_con_graph(ConnectionGraph* conn_graph, PhaseGVN* gvn, Unique_Node_List* delayed_worklist, Node* n, uint opcode) const { return false; } |
310 | virtual bool escape_add_final_edges(ConnectionGraph* conn_graph, PhaseGVN* gvn, Node* n, uint opcode) const { return false; } |
311 | virtual bool escape_has_out_with_unsafe_object(Node* n) const { return false; } |
312 | virtual bool escape_is_barrier_node(Node* n) const { return false; } |
313 | |
314 | virtual bool matcher_find_shared_visit(Matcher* matcher, Matcher::MStack& mstack, Node* n, uint opcode, bool& mem_op, int& mem_addr_idx) const { return false; }; |
315 | virtual bool matcher_find_shared_post_visit(Matcher* matcher, Node* n, uint opcode) const { return false; }; |
316 | virtual bool matcher_is_store_load_barrier(Node* x, uint xop) const { return false; } |
317 | |
318 | virtual void igvn_add_users_to_worklist(PhaseIterGVN* igvn, Node* use) const { } |
319 | virtual void ccp_analyze(PhaseCCP* ccp, Unique_Node_List& worklist, Node* use) const { } |
320 | |
321 | virtual Node* split_if_pre(PhaseIdealLoop* phase, Node* n) const { return NULL; } |
322 | virtual bool build_loop_late_post(PhaseIdealLoop* phase, Node* n) const { return false; } |
323 | virtual bool sink_node(PhaseIdealLoop* phase, Node* n, Node* x, Node* x_ctrl, Node* n_ctrl) const { return false; } |
324 | }; |
325 | |
326 | #endif // SHARE_GC_SHARED_C2_BARRIERSETC2_HPP |
327 | |