File: | jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp |
Warning: | line 751, column 26 Called C++ object pointer is null |
Press '?' to see keyboard shortcuts
Keyboard shortcuts:
1 | /* | ||||
2 | * Copyright (c) 2018, 2020, Oracle and/or its affiliates. All rights reserved. | ||||
3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. | ||||
4 | * | ||||
5 | * This code is free software; you can redistribute it and/or modify it | ||||
6 | * under the terms of the GNU General Public License version 2 only, as | ||||
7 | * published by the Free Software Foundation. | ||||
8 | * | ||||
9 | * This code is distributed in the hope that it will be useful, but WITHOUT | ||||
10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | ||||
11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License | ||||
12 | * version 2 for more details (a copy is included in the LICENSE file that | ||||
13 | * accompanied this code). | ||||
14 | * | ||||
15 | * You should have received a copy of the GNU General Public License version | ||||
16 | * 2 along with this work; if not, write to the Free Software Foundation, | ||||
17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. | ||||
18 | * | ||||
19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA | ||||
20 | * or visit www.oracle.com if you need additional information or have any | ||||
21 | * questions. | ||||
22 | * | ||||
23 | */ | ||||
24 | |||||
25 | #include "precompiled.hpp" | ||||
26 | #include "classfile/javaClasses.hpp" | ||||
27 | #include "gc/g1/c2/g1BarrierSetC2.hpp" | ||||
28 | #include "gc/g1/g1BarrierSet.hpp" | ||||
29 | #include "gc/g1/g1BarrierSetRuntime.hpp" | ||||
30 | #include "gc/g1/g1CardTable.hpp" | ||||
31 | #include "gc/g1/g1ThreadLocalData.hpp" | ||||
32 | #include "gc/g1/heapRegion.hpp" | ||||
33 | #include "opto/arraycopynode.hpp" | ||||
34 | #include "opto/compile.hpp" | ||||
35 | #include "opto/escape.hpp" | ||||
36 | #include "opto/graphKit.hpp" | ||||
37 | #include "opto/idealKit.hpp" | ||||
38 | #include "opto/macro.hpp" | ||||
39 | #include "opto/rootnode.hpp" | ||||
40 | #include "opto/type.hpp" | ||||
41 | #include "utilities/macros.hpp" | ||||
42 | |||||
43 | const TypeFunc *G1BarrierSetC2::write_ref_field_pre_entry_Type() { | ||||
44 | const Type **fields = TypeTuple::fields(2); | ||||
45 | fields[TypeFunc::Parms+0] = TypeInstPtr::NOTNULL; // original field value | ||||
46 | fields[TypeFunc::Parms+1] = TypeRawPtr::NOTNULL; // thread | ||||
47 | const TypeTuple *domain = TypeTuple::make(TypeFunc::Parms+2, fields); | ||||
48 | |||||
49 | // create result type (range) | ||||
50 | fields = TypeTuple::fields(0); | ||||
51 | const TypeTuple *range = TypeTuple::make(TypeFunc::Parms+0, fields); | ||||
52 | |||||
53 | return TypeFunc::make(domain, range); | ||||
54 | } | ||||
55 | |||||
56 | const TypeFunc *G1BarrierSetC2::write_ref_field_post_entry_Type() { | ||||
57 | const Type **fields = TypeTuple::fields(2); | ||||
58 | fields[TypeFunc::Parms+0] = TypeRawPtr::NOTNULL; // Card addr | ||||
59 | fields[TypeFunc::Parms+1] = TypeRawPtr::NOTNULL; // thread | ||||
60 | const TypeTuple *domain = TypeTuple::make(TypeFunc::Parms+2, fields); | ||||
61 | |||||
62 | // create result type (range) | ||||
63 | fields = TypeTuple::fields(0); | ||||
64 | const TypeTuple *range = TypeTuple::make(TypeFunc::Parms, fields); | ||||
65 | |||||
66 | return TypeFunc::make(domain, range); | ||||
67 | } | ||||
68 | |||||
69 | #define __ ideal. | ||||
70 | /* | ||||
71 | * Determine if the G1 pre-barrier can be removed. The pre-barrier is | ||||
72 | * required by SATB to make sure all objects live at the start of the | ||||
73 | * marking are kept alive, all reference updates need to any previous | ||||
74 | * reference stored before writing. | ||||
75 | * | ||||
76 | * If the previous value is NULL there is no need to save the old value. | ||||
77 | * References that are NULL are filtered during runtime by the barrier | ||||
78 | * code to avoid unnecessary queuing. | ||||
79 | * | ||||
80 | * However in the case of newly allocated objects it might be possible to | ||||
81 | * prove that the reference about to be overwritten is NULL during compile | ||||
82 | * time and avoid adding the barrier code completely. | ||||
83 | * | ||||
84 | * The compiler needs to determine that the object in which a field is about | ||||
85 | * to be written is newly allocated, and that no prior store to the same field | ||||
86 | * has happened since the allocation. | ||||
87 | * | ||||
88 | * Returns true if the pre-barrier can be removed | ||||
89 | */ | ||||
90 | bool G1BarrierSetC2::g1_can_remove_pre_barrier(GraphKit* kit, | ||||
91 | PhaseTransform* phase, | ||||
92 | Node* adr, | ||||
93 | BasicType bt, | ||||
94 | uint adr_idx) const { | ||||
95 | intptr_t offset = 0; | ||||
96 | Node* base = AddPNode::Ideal_base_and_offset(adr, phase, offset); | ||||
97 | AllocateNode* alloc = AllocateNode::Ideal_allocation(base, phase); | ||||
98 | |||||
99 | if (offset == Type::OffsetBot) { | ||||
100 | return false; // cannot unalias unless there are precise offsets | ||||
101 | } | ||||
102 | |||||
103 | if (alloc == NULL__null) { | ||||
104 | return false; // No allocation found | ||||
105 | } | ||||
106 | |||||
107 | intptr_t size_in_bytes = type2aelembytes(bt); | ||||
108 | |||||
109 | Node* mem = kit->memory(adr_idx); // start searching here... | ||||
110 | |||||
111 | for (int cnt = 0; cnt < 50; cnt++) { | ||||
112 | |||||
113 | if (mem->is_Store()) { | ||||
114 | |||||
115 | Node* st_adr = mem->in(MemNode::Address); | ||||
116 | intptr_t st_offset = 0; | ||||
117 | Node* st_base = AddPNode::Ideal_base_and_offset(st_adr, phase, st_offset); | ||||
118 | |||||
119 | if (st_base == NULL__null) { | ||||
120 | break; // inscrutable pointer | ||||
121 | } | ||||
122 | |||||
123 | // Break we have found a store with same base and offset as ours so break | ||||
124 | if (st_base == base && st_offset == offset) { | ||||
125 | break; | ||||
126 | } | ||||
127 | |||||
128 | if (st_offset != offset && st_offset != Type::OffsetBot) { | ||||
129 | const int MAX_STORE = BytesPerLong; | ||||
130 | if (st_offset >= offset + size_in_bytes || | ||||
131 | st_offset <= offset - MAX_STORE || | ||||
132 | st_offset <= offset - mem->as_Store()->memory_size()) { | ||||
133 | // Success: The offsets are provably independent. | ||||
134 | // (You may ask, why not just test st_offset != offset and be done? | ||||
135 | // The answer is that stores of different sizes can co-exist | ||||
136 | // in the same sequence of RawMem effects. We sometimes initialize | ||||
137 | // a whole 'tile' of array elements with a single jint or jlong.) | ||||
138 | mem = mem->in(MemNode::Memory); | ||||
139 | continue; // advance through independent store memory | ||||
140 | } | ||||
141 | } | ||||
142 | |||||
143 | if (st_base != base | ||||
144 | && MemNode::detect_ptr_independence(base, alloc, st_base, | ||||
145 | AllocateNode::Ideal_allocation(st_base, phase), | ||||
146 | phase)) { | ||||
147 | // Success: The bases are provably independent. | ||||
148 | mem = mem->in(MemNode::Memory); | ||||
149 | continue; // advance through independent store memory | ||||
150 | } | ||||
151 | } else if (mem->is_Proj() && mem->in(0)->is_Initialize()) { | ||||
152 | |||||
153 | InitializeNode* st_init = mem->in(0)->as_Initialize(); | ||||
154 | AllocateNode* st_alloc = st_init->allocation(); | ||||
155 | |||||
156 | // Make sure that we are looking at the same allocation site. | ||||
157 | // The alloc variable is guaranteed to not be null here from earlier check. | ||||
158 | if (alloc == st_alloc) { | ||||
159 | // Check that the initialization is storing NULL so that no previous store | ||||
160 | // has been moved up and directly write a reference | ||||
161 | Node* captured_store = st_init->find_captured_store(offset, | ||||
162 | type2aelembytes(T_OBJECT), | ||||
163 | phase); | ||||
164 | if (captured_store == NULL__null || captured_store == st_init->zero_memory()) { | ||||
165 | return true; | ||||
166 | } | ||||
167 | } | ||||
168 | } | ||||
169 | |||||
170 | // Unless there is an explicit 'continue', we must bail out here, | ||||
171 | // because 'mem' is an inscrutable memory state (e.g., a call). | ||||
172 | break; | ||||
173 | } | ||||
174 | |||||
175 | return false; | ||||
176 | } | ||||
177 | |||||
178 | // G1 pre/post barriers | ||||
179 | void G1BarrierSetC2::pre_barrier(GraphKit* kit, | ||||
180 | bool do_load, | ||||
181 | Node* ctl, | ||||
182 | Node* obj, | ||||
183 | Node* adr, | ||||
184 | uint alias_idx, | ||||
185 | Node* val, | ||||
186 | const TypeOopPtr* val_type, | ||||
187 | Node* pre_val, | ||||
188 | BasicType bt) const { | ||||
189 | // Some sanity checks | ||||
190 | // Note: val is unused in this routine. | ||||
191 | |||||
192 | if (do_load) { | ||||
193 | // We need to generate the load of the previous value | ||||
194 | assert(obj != NULL, "must have a base")do { if (!(obj != __null)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 194, "assert(" "obj != __null" ") failed", "must have a base" ); ::breakpoint(); } } while (0); | ||||
195 | assert(adr != NULL, "where are loading from?")do { if (!(adr != __null)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 195, "assert(" "adr != __null" ") failed", "where are loading from?" ); ::breakpoint(); } } while (0); | ||||
196 | assert(pre_val == NULL, "loaded already?")do { if (!(pre_val == __null)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 196, "assert(" "pre_val == __null" ") failed", "loaded already?" ); ::breakpoint(); } } while (0); | ||||
197 | assert(val_type != NULL, "need a type")do { if (!(val_type != __null)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 197, "assert(" "val_type != __null" ") failed", "need a type" ); ::breakpoint(); } } while (0); | ||||
198 | |||||
199 | if (use_ReduceInitialCardMarks() | ||||
200 | && g1_can_remove_pre_barrier(kit, &kit->gvn(), adr, bt, alias_idx)) { | ||||
201 | return; | ||||
202 | } | ||||
203 | |||||
204 | } else { | ||||
205 | // In this case both val_type and alias_idx are unused. | ||||
206 | assert(pre_val != NULL, "must be loaded already")do { if (!(pre_val != __null)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 206, "assert(" "pre_val != __null" ") failed", "must be loaded already" ); ::breakpoint(); } } while (0); | ||||
207 | // Nothing to be done if pre_val is null. | ||||
208 | if (pre_val->bottom_type() == TypePtr::NULL_PTR) return; | ||||
209 | assert(pre_val->bottom_type()->basic_type() == T_OBJECT, "or we shouldn't be here")do { if (!(pre_val->bottom_type()->basic_type() == T_OBJECT )) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 209, "assert(" "pre_val->bottom_type()->basic_type() == T_OBJECT" ") failed", "or we shouldn't be here"); ::breakpoint(); } } while (0); | ||||
210 | } | ||||
211 | assert(bt == T_OBJECT, "or we shouldn't be here")do { if (!(bt == T_OBJECT)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 211, "assert(" "bt == T_OBJECT" ") failed", "or we shouldn't be here" ); ::breakpoint(); } } while (0); | ||||
212 | |||||
213 | IdealKit ideal(kit, true); | ||||
214 | |||||
215 | Node* tls = __ thread(); // ThreadLocalStorage | ||||
216 | |||||
217 | Node* no_base = __ top(); | ||||
218 | Node* zero = __ ConI(0); | ||||
219 | Node* zeroX = __ ConX(0); | ||||
220 | |||||
221 | float likely = PROB_LIKELY(0.999)((float) (0.999)); | ||||
222 | float unlikely = PROB_UNLIKELY(0.999)(1.0f - (float)(0.999)); | ||||
223 | |||||
224 | BasicType active_type = in_bytes(SATBMarkQueue::byte_width_of_active()) == 4 ? T_INT : T_BYTE; | ||||
225 | assert(in_bytes(SATBMarkQueue::byte_width_of_active()) == 4 || in_bytes(SATBMarkQueue::byte_width_of_active()) == 1, "flag width")do { if (!(in_bytes(SATBMarkQueue::byte_width_of_active()) == 4 || in_bytes(SATBMarkQueue::byte_width_of_active()) == 1)) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 225, "assert(" "in_bytes(SATBMarkQueue::byte_width_of_active()) == 4 || in_bytes(SATBMarkQueue::byte_width_of_active()) == 1" ") failed", "flag width"); ::breakpoint(); } } while (0); | ||||
226 | |||||
227 | // Offsets into the thread | ||||
228 | const int marking_offset = in_bytes(G1ThreadLocalData::satb_mark_queue_active_offset()); | ||||
229 | const int index_offset = in_bytes(G1ThreadLocalData::satb_mark_queue_index_offset()); | ||||
230 | const int buffer_offset = in_bytes(G1ThreadLocalData::satb_mark_queue_buffer_offset()); | ||||
231 | |||||
232 | // Now the actual pointers into the thread | ||||
233 | Node* marking_adr = __ AddP(no_base, tls, __ ConX(marking_offset)); | ||||
234 | Node* buffer_adr = __ AddP(no_base, tls, __ ConX(buffer_offset)); | ||||
235 | Node* index_adr = __ AddP(no_base, tls, __ ConX(index_offset)); | ||||
236 | |||||
237 | // Now some of the values | ||||
238 | Node* marking = __ load(__ ctrl(), marking_adr, TypeInt::INT, active_type, Compile::AliasIdxRaw); | ||||
239 | |||||
240 | // if (!marking) | ||||
241 | __ if_then(marking, BoolTest::ne, zero, unlikely); { | ||||
242 | BasicType index_bt = TypeX_XTypeLong::LONG->basic_type(); | ||||
243 | assert(sizeof(size_t) == type2aelembytes(index_bt), "Loading G1 SATBMarkQueue::_index with wrong size.")do { if (!(sizeof(size_t) == type2aelembytes(index_bt))) { (* g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 243, "assert(" "sizeof(size_t) == type2aelembytes(index_bt)" ") failed", "Loading G1 SATBMarkQueue::_index with wrong size." ); ::breakpoint(); } } while (0); | ||||
244 | Node* index = __ load(__ ctrl(), index_adr, TypeX_XTypeLong::LONG, index_bt, Compile::AliasIdxRaw); | ||||
245 | |||||
246 | if (do_load) { | ||||
247 | // load original value | ||||
248 | // alias_idx correct?? | ||||
249 | pre_val = __ load(__ ctrl(), adr, val_type, bt, alias_idx); | ||||
250 | } | ||||
251 | |||||
252 | // if (pre_val != NULL) | ||||
253 | __ if_then(pre_val, BoolTest::ne, kit->null()); { | ||||
254 | Node* buffer = __ load(__ ctrl(), buffer_adr, TypeRawPtr::NOTNULL, T_ADDRESS, Compile::AliasIdxRaw); | ||||
255 | |||||
256 | // is the queue for this thread full? | ||||
257 | __ if_then(index, BoolTest::ne, zeroX, likely); { | ||||
258 | |||||
259 | // decrement the index | ||||
260 | Node* next_index = kit->gvn().transform(new SubXNodeSubLNode(index, __ ConX(sizeof(intptr_t)))); | ||||
261 | |||||
262 | // Now get the buffer location we will log the previous value into and store it | ||||
263 | Node *log_addr = __ AddP(no_base, buffer, next_index); | ||||
264 | __ store(__ ctrl(), log_addr, pre_val, T_OBJECT, Compile::AliasIdxRaw, MemNode::unordered); | ||||
265 | // update the index | ||||
266 | __ store(__ ctrl(), index_adr, next_index, index_bt, Compile::AliasIdxRaw, MemNode::unordered); | ||||
267 | |||||
268 | } __ else_(); { | ||||
269 | |||||
270 | // logging buffer is full, call the runtime | ||||
271 | const TypeFunc *tf = write_ref_field_pre_entry_Type(); | ||||
272 | __ make_leaf_call(tf, CAST_FROM_FN_PTR(address, G1BarrierSetRuntime::write_ref_field_pre_entry)((address)((address_word)(G1BarrierSetRuntime::write_ref_field_pre_entry ))), "write_ref_field_pre_entry", pre_val, tls); | ||||
273 | } __ end_if(); // (!index) | ||||
274 | } __ end_if(); // (pre_val != NULL) | ||||
275 | } __ end_if(); // (!marking) | ||||
276 | |||||
277 | // Final sync IdealKit and GraphKit. | ||||
278 | kit->final_sync(ideal); | ||||
279 | } | ||||
280 | |||||
281 | /* | ||||
282 | * G1 similar to any GC with a Young Generation requires a way to keep track of | ||||
283 | * references from Old Generation to Young Generation to make sure all live | ||||
284 | * objects are found. G1 also requires to keep track of object references | ||||
285 | * between different regions to enable evacuation of old regions, which is done | ||||
286 | * as part of mixed collections. References are tracked in remembered sets and | ||||
287 | * is continuously updated as reference are written to with the help of the | ||||
288 | * post-barrier. | ||||
289 | * | ||||
290 | * To reduce the number of updates to the remembered set the post-barrier | ||||
291 | * filters updates to fields in objects located in the Young Generation, | ||||
292 | * the same region as the reference, when the NULL is being written or | ||||
293 | * if the card is already marked as dirty by an earlier write. | ||||
294 | * | ||||
295 | * Under certain circumstances it is possible to avoid generating the | ||||
296 | * post-barrier completely if it is possible during compile time to prove | ||||
297 | * the object is newly allocated and that no safepoint exists between the | ||||
298 | * allocation and the store. | ||||
299 | * | ||||
300 | * In the case of slow allocation the allocation code must handle the barrier | ||||
301 | * as part of the allocation in the case the allocated object is not located | ||||
302 | * in the nursery; this would happen for humongous objects. | ||||
303 | * | ||||
304 | * Returns true if the post barrier can be removed | ||||
305 | */ | ||||
306 | bool G1BarrierSetC2::g1_can_remove_post_barrier(GraphKit* kit, | ||||
307 | PhaseTransform* phase, Node* store, | ||||
308 | Node* adr) const { | ||||
309 | intptr_t offset = 0; | ||||
310 | Node* base = AddPNode::Ideal_base_and_offset(adr, phase, offset); | ||||
311 | AllocateNode* alloc = AllocateNode::Ideal_allocation(base, phase); | ||||
312 | |||||
313 | if (offset == Type::OffsetBot) { | ||||
314 | return false; // cannot unalias unless there are precise offsets | ||||
315 | } | ||||
316 | |||||
317 | if (alloc == NULL__null) { | ||||
318 | return false; // No allocation found | ||||
319 | } | ||||
320 | |||||
321 | // Start search from Store node | ||||
322 | Node* mem = store->in(MemNode::Control); | ||||
323 | if (mem->is_Proj() && mem->in(0)->is_Initialize()) { | ||||
324 | |||||
325 | InitializeNode* st_init = mem->in(0)->as_Initialize(); | ||||
326 | AllocateNode* st_alloc = st_init->allocation(); | ||||
327 | |||||
328 | // Make sure we are looking at the same allocation | ||||
329 | if (alloc == st_alloc) { | ||||
330 | return true; | ||||
331 | } | ||||
332 | } | ||||
333 | |||||
334 | return false; | ||||
335 | } | ||||
336 | |||||
337 | // | ||||
338 | // Update the card table and add card address to the queue | ||||
339 | // | ||||
340 | void G1BarrierSetC2::g1_mark_card(GraphKit* kit, | ||||
341 | IdealKit& ideal, | ||||
342 | Node* card_adr, | ||||
343 | Node* oop_store, | ||||
344 | uint oop_alias_idx, | ||||
345 | Node* index, | ||||
346 | Node* index_adr, | ||||
347 | Node* buffer, | ||||
348 | const TypeFunc* tf) const { | ||||
349 | Node* zero = __ ConI(0); | ||||
350 | Node* zeroX = __ ConX(0); | ||||
351 | Node* no_base = __ top(); | ||||
352 | BasicType card_bt = T_BYTE; | ||||
353 | // Smash zero into card. MUST BE ORDERED WRT TO STORE | ||||
354 | __ storeCM(__ ctrl(), card_adr, zero, oop_store, oop_alias_idx, card_bt, Compile::AliasIdxRaw); | ||||
355 | |||||
356 | // Now do the queue work | ||||
357 | __ if_then(index, BoolTest::ne, zeroX); { | ||||
358 | |||||
359 | Node* next_index = kit->gvn().transform(new SubXNodeSubLNode(index, __ ConX(sizeof(intptr_t)))); | ||||
360 | Node* log_addr = __ AddP(no_base, buffer, next_index); | ||||
361 | |||||
362 | // Order, see storeCM. | ||||
363 | __ store(__ ctrl(), log_addr, card_adr, T_ADDRESS, Compile::AliasIdxRaw, MemNode::unordered); | ||||
364 | __ store(__ ctrl(), index_adr, next_index, TypeX_XTypeLong::LONG->basic_type(), Compile::AliasIdxRaw, MemNode::unordered); | ||||
365 | |||||
366 | } __ else_(); { | ||||
367 | __ make_leaf_call(tf, CAST_FROM_FN_PTR(address, G1BarrierSetRuntime::write_ref_field_post_entry)((address)((address_word)(G1BarrierSetRuntime::write_ref_field_post_entry ))), "write_ref_field_post_entry", card_adr, __ thread()); | ||||
368 | } __ end_if(); | ||||
369 | |||||
370 | } | ||||
371 | |||||
372 | void G1BarrierSetC2::post_barrier(GraphKit* kit, | ||||
373 | Node* ctl, | ||||
374 | Node* oop_store, | ||||
375 | Node* obj, | ||||
376 | Node* adr, | ||||
377 | uint alias_idx, | ||||
378 | Node* val, | ||||
379 | BasicType bt, | ||||
380 | bool use_precise) const { | ||||
381 | // If we are writing a NULL then we need no post barrier | ||||
382 | |||||
383 | if (val != NULL__null && val->is_Con() && val->bottom_type() == TypePtr::NULL_PTR) { | ||||
384 | // Must be NULL | ||||
385 | const Type* t = val->bottom_type(); | ||||
386 | assert(t == Type::TOP || t == TypePtr::NULL_PTR, "must be NULL")do { if (!(t == Type::TOP || t == TypePtr::NULL_PTR)) { (*g_assert_poison ) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 386, "assert(" "t == Type::TOP || t == TypePtr::NULL_PTR" ") failed" , "must be NULL"); ::breakpoint(); } } while (0); | ||||
387 | // No post barrier if writing NULLx | ||||
388 | return; | ||||
389 | } | ||||
390 | |||||
391 | if (use_ReduceInitialCardMarks() && obj == kit->just_allocated_object(kit->control())) { | ||||
392 | // We can skip marks on a freshly-allocated object in Eden. | ||||
393 | // Keep this code in sync with new_deferred_store_barrier() in runtime.cpp. | ||||
394 | // That routine informs GC to take appropriate compensating steps, | ||||
395 | // upon a slow-path allocation, so as to make this card-mark | ||||
396 | // elision safe. | ||||
397 | return; | ||||
398 | } | ||||
399 | |||||
400 | if (use_ReduceInitialCardMarks() | ||||
401 | && g1_can_remove_post_barrier(kit, &kit->gvn(), oop_store, adr)) { | ||||
402 | return; | ||||
403 | } | ||||
404 | |||||
405 | if (!use_precise) { | ||||
406 | // All card marks for a (non-array) instance are in one place: | ||||
407 | adr = obj; | ||||
408 | } | ||||
409 | // (Else it's an array (or unknown), and we want more precise card marks.) | ||||
410 | assert(adr != NULL, "")do { if (!(adr != __null)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 410, "assert(" "adr != __null" ") failed", ""); ::breakpoint (); } } while (0); | ||||
411 | |||||
412 | IdealKit ideal(kit, true); | ||||
413 | |||||
414 | Node* tls = __ thread(); // ThreadLocalStorage | ||||
415 | |||||
416 | Node* no_base = __ top(); | ||||
417 | float likely = PROB_LIKELY_MAG(3)(1.0f-(1e-3f)); | ||||
418 | float unlikely = PROB_UNLIKELY_MAG(3)(1e-3f); | ||||
419 | Node* young_card = __ ConI((jint)G1CardTable::g1_young_card_val()); | ||||
420 | Node* dirty_card = __ ConI((jint)G1CardTable::dirty_card_val()); | ||||
421 | Node* zeroX = __ ConX(0); | ||||
422 | |||||
423 | const TypeFunc *tf = write_ref_field_post_entry_Type(); | ||||
424 | |||||
425 | // Offsets into the thread | ||||
426 | const int index_offset = in_bytes(G1ThreadLocalData::dirty_card_queue_index_offset()); | ||||
427 | const int buffer_offset = in_bytes(G1ThreadLocalData::dirty_card_queue_buffer_offset()); | ||||
428 | |||||
429 | // Pointers into the thread | ||||
430 | |||||
431 | Node* buffer_adr = __ AddP(no_base, tls, __ ConX(buffer_offset)); | ||||
432 | Node* index_adr = __ AddP(no_base, tls, __ ConX(index_offset)); | ||||
433 | |||||
434 | // Now some values | ||||
435 | // Use ctrl to avoid hoisting these values past a safepoint, which could | ||||
436 | // potentially reset these fields in the JavaThread. | ||||
437 | Node* index = __ load(__ ctrl(), index_adr, TypeX_XTypeLong::LONG, TypeX_XTypeLong::LONG->basic_type(), Compile::AliasIdxRaw); | ||||
438 | Node* buffer = __ load(__ ctrl(), buffer_adr, TypeRawPtr::NOTNULL, T_ADDRESS, Compile::AliasIdxRaw); | ||||
439 | |||||
440 | // Convert the store obj pointer to an int prior to doing math on it | ||||
441 | // Must use ctrl to prevent "integerized oop" existing across safepoint | ||||
442 | Node* cast = __ CastPX(__ ctrl(), adr); | ||||
443 | |||||
444 | // Divide pointer by card size | ||||
445 | Node* card_offset = __ URShiftX( cast, __ ConI(CardTable::card_shift()) ); | ||||
446 | |||||
447 | // Combine card table base and card offset | ||||
448 | Node* card_adr = __ AddP(no_base, byte_map_base_node(kit), card_offset ); | ||||
449 | |||||
450 | // If we know the value being stored does it cross regions? | ||||
451 | |||||
452 | if (val != NULL__null) { | ||||
453 | // Does the store cause us to cross regions? | ||||
454 | |||||
455 | // Should be able to do an unsigned compare of region_size instead of | ||||
456 | // and extra shift. Do we have an unsigned compare?? | ||||
457 | // Node* region_size = __ ConI(1 << HeapRegion::LogOfHRGrainBytes); | ||||
458 | Node* xor_res = __ URShiftX ( __ XorX( cast, __ CastPX(__ ctrl(), val)), __ ConI(HeapRegion::LogOfHRGrainBytes)); | ||||
459 | |||||
460 | // if (xor_res == 0) same region so skip | ||||
461 | __ if_then(xor_res, BoolTest::ne, zeroX, likely); { | ||||
462 | |||||
463 | // No barrier if we are storing a NULL | ||||
464 | __ if_then(val, BoolTest::ne, kit->null(), likely); { | ||||
465 | |||||
466 | // Ok must mark the card if not already dirty | ||||
467 | |||||
468 | // load the original value of the card | ||||
469 | Node* card_val = __ load(__ ctrl(), card_adr, TypeInt::INT, T_BYTE, Compile::AliasIdxRaw); | ||||
470 | |||||
471 | __ if_then(card_val, BoolTest::ne, young_card, unlikely); { | ||||
472 | kit->sync_kit(ideal); | ||||
473 | kit->insert_mem_bar(Op_MemBarVolatile, oop_store); | ||||
474 | __ sync_kit(kit); | ||||
475 | |||||
476 | Node* card_val_reload = __ load(__ ctrl(), card_adr, TypeInt::INT, T_BYTE, Compile::AliasIdxRaw); | ||||
477 | __ if_then(card_val_reload, BoolTest::ne, dirty_card); { | ||||
478 | g1_mark_card(kit, ideal, card_adr, oop_store, alias_idx, index, index_adr, buffer, tf); | ||||
479 | } __ end_if(); | ||||
480 | } __ end_if(); | ||||
481 | } __ end_if(); | ||||
482 | } __ end_if(); | ||||
483 | } else { | ||||
484 | // The Object.clone() intrinsic uses this path if !ReduceInitialCardMarks. | ||||
485 | // We don't need a barrier here if the destination is a newly allocated object | ||||
486 | // in Eden. Otherwise, GC verification breaks because we assume that cards in Eden | ||||
487 | // are set to 'g1_young_gen' (see G1CardTable::verify_g1_young_region()). | ||||
488 | assert(!use_ReduceInitialCardMarks(), "can only happen with card marking")do { if (!(!use_ReduceInitialCardMarks())) { (*g_assert_poison ) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 488, "assert(" "!use_ReduceInitialCardMarks()" ") failed", "can only happen with card marking" ); ::breakpoint(); } } while (0); | ||||
489 | Node* card_val = __ load(__ ctrl(), card_adr, TypeInt::INT, T_BYTE, Compile::AliasIdxRaw); | ||||
490 | __ if_then(card_val, BoolTest::ne, young_card); { | ||||
491 | g1_mark_card(kit, ideal, card_adr, oop_store, alias_idx, index, index_adr, buffer, tf); | ||||
492 | } __ end_if(); | ||||
493 | } | ||||
494 | |||||
495 | // Final sync IdealKit and GraphKit. | ||||
496 | kit->final_sync(ideal); | ||||
497 | } | ||||
498 | |||||
499 | // Helper that guards and inserts a pre-barrier. | ||||
500 | void G1BarrierSetC2::insert_pre_barrier(GraphKit* kit, Node* base_oop, Node* offset, | ||||
501 | Node* pre_val, bool need_mem_bar) const { | ||||
502 | // We could be accessing the referent field of a reference object. If so, when G1 | ||||
503 | // is enabled, we need to log the value in the referent field in an SATB buffer. | ||||
504 | // This routine performs some compile time filters and generates suitable | ||||
505 | // runtime filters that guard the pre-barrier code. | ||||
506 | // Also add memory barrier for non volatile load from the referent field | ||||
507 | // to prevent commoning of loads across safepoint. | ||||
508 | |||||
509 | // Some compile time checks. | ||||
510 | |||||
511 | // If offset is a constant, is it java_lang_ref_Reference::_reference_offset? | ||||
512 | const TypeXTypeLong* otype = offset->find_intptr_t_typefind_long_type(); | ||||
513 | if (otype != NULL__null && otype->is_con() && | ||||
514 | otype->get_con() != java_lang_ref_Reference::referent_offset()) { | ||||
515 | // Constant offset but not the reference_offset so just return | ||||
516 | return; | ||||
517 | } | ||||
518 | |||||
519 | // We only need to generate the runtime guards for instances. | ||||
520 | const TypeOopPtr* btype = base_oop->bottom_type()->isa_oopptr(); | ||||
521 | if (btype != NULL__null) { | ||||
522 | if (btype->isa_aryptr()) { | ||||
523 | // Array type so nothing to do | ||||
524 | return; | ||||
525 | } | ||||
526 | |||||
527 | const TypeInstPtr* itype = btype->isa_instptr(); | ||||
528 | if (itype != NULL__null) { | ||||
529 | // Can the klass of base_oop be statically determined to be | ||||
530 | // _not_ a sub-class of Reference and _not_ Object? | ||||
531 | ciKlass* klass = itype->klass(); | ||||
532 | if ( klass->is_loaded() && | ||||
533 | !klass->is_subtype_of(kit->env()->Reference_klass()) && | ||||
534 | !kit->env()->Object_klass()->is_subtype_of(klass)) { | ||||
535 | return; | ||||
536 | } | ||||
537 | } | ||||
538 | } | ||||
539 | |||||
540 | // The compile time filters did not reject base_oop/offset so | ||||
541 | // we need to generate the following runtime filters | ||||
542 | // | ||||
543 | // if (offset == java_lang_ref_Reference::_reference_offset) { | ||||
544 | // if (instance_of(base, java.lang.ref.Reference)) { | ||||
545 | // pre_barrier(_, pre_val, ...); | ||||
546 | // } | ||||
547 | // } | ||||
548 | |||||
549 | float likely = PROB_LIKELY( 0.999)((float) (0.999)); | ||||
550 | float unlikely = PROB_UNLIKELY(0.999)(1.0f - (float)(0.999)); | ||||
551 | |||||
552 | IdealKit ideal(kit); | ||||
553 | |||||
554 | Node* referent_off = __ ConX(java_lang_ref_Reference::referent_offset()); | ||||
555 | |||||
556 | __ if_then(offset, BoolTest::eq, referent_off, unlikely); { | ||||
557 | // Update graphKit memory and control from IdealKit. | ||||
558 | kit->sync_kit(ideal); | ||||
559 | |||||
560 | Node* ref_klass_con = kit->makecon(TypeKlassPtr::make(kit->env()->Reference_klass())); | ||||
561 | Node* is_instof = kit->gen_instanceof(base_oop, ref_klass_con); | ||||
562 | |||||
563 | // Update IdealKit memory and control from graphKit. | ||||
564 | __ sync_kit(kit); | ||||
565 | |||||
566 | Node* one = __ ConI(1); | ||||
567 | // is_instof == 0 if base_oop == NULL | ||||
568 | __ if_then(is_instof, BoolTest::eq, one, unlikely); { | ||||
569 | |||||
570 | // Update graphKit from IdeakKit. | ||||
571 | kit->sync_kit(ideal); | ||||
572 | |||||
573 | // Use the pre-barrier to record the value in the referent field | ||||
574 | pre_barrier(kit, false /* do_load */, | ||||
575 | __ ctrl(), | ||||
576 | NULL__null /* obj */, NULL__null /* adr */, max_juint /* alias_idx */, NULL__null /* val */, NULL__null /* val_type */, | ||||
577 | pre_val /* pre_val */, | ||||
578 | T_OBJECT); | ||||
579 | if (need_mem_bar) { | ||||
580 | // Add memory barrier to prevent commoning reads from this field | ||||
581 | // across safepoint since GC can change its value. | ||||
582 | kit->insert_mem_bar(Op_MemBarCPUOrder); | ||||
583 | } | ||||
584 | // Update IdealKit from graphKit. | ||||
585 | __ sync_kit(kit); | ||||
586 | |||||
587 | } __ end_if(); // _ref_type != ref_none | ||||
588 | } __ end_if(); // offset == referent_offset | ||||
589 | |||||
590 | // Final sync IdealKit and GraphKit. | ||||
591 | kit->final_sync(ideal); | ||||
592 | } | ||||
593 | |||||
594 | #undef __ | ||||
595 | |||||
596 | Node* G1BarrierSetC2::load_at_resolved(C2Access& access, const Type* val_type) const { | ||||
597 | DecoratorSet decorators = access.decorators(); | ||||
598 | Node* adr = access.addr().node(); | ||||
599 | Node* obj = access.base(); | ||||
600 | |||||
601 | bool anonymous = (decorators & C2_UNSAFE_ACCESS) != 0; | ||||
602 | bool mismatched = (decorators & C2_MISMATCHED) != 0; | ||||
603 | bool unknown = (decorators & ON_UNKNOWN_OOP_REF) != 0; | ||||
604 | bool in_heap = (decorators & IN_HEAP) != 0; | ||||
605 | bool in_native = (decorators & IN_NATIVE) != 0; | ||||
606 | bool on_weak = (decorators & ON_WEAK_OOP_REF) != 0; | ||||
607 | bool on_phantom = (decorators & ON_PHANTOM_OOP_REF) != 0; | ||||
608 | bool is_unordered = (decorators & MO_UNORDERED) != 0; | ||||
609 | bool no_keepalive = (decorators & AS_NO_KEEPALIVE) != 0; | ||||
610 | bool is_mixed = !in_heap && !in_native; | ||||
611 | bool need_cpu_mem_bar = !is_unordered || mismatched || is_mixed; | ||||
612 | |||||
613 | Node* top = Compile::current()->top(); | ||||
614 | Node* offset = adr->is_AddP() ? adr->in(AddPNode::Offset) : top; | ||||
615 | Node* load = CardTableBarrierSetC2::load_at_resolved(access, val_type); | ||||
616 | |||||
617 | // If we are reading the value of the referent field of a Reference | ||||
618 | // object (either by using Unsafe directly or through reflection) | ||||
619 | // then, if G1 is enabled, we need to record the referent in an | ||||
620 | // SATB log buffer using the pre-barrier mechanism. | ||||
621 | // Also we need to add memory barrier to prevent commoning reads | ||||
622 | // from this field across safepoint since GC can change its value. | ||||
623 | bool need_read_barrier = (((on_weak || on_phantom) && !no_keepalive) || | ||||
624 | (in_heap && unknown && offset != top && obj != top)); | ||||
625 | |||||
626 | if (!access.is_oop() || !need_read_barrier) { | ||||
627 | return load; | ||||
628 | } | ||||
629 | |||||
630 | assert(access.is_parse_access(), "entry not supported at optimization time")do { if (!(access.is_parse_access())) { (*g_assert_poison) = 'X' ;; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 630, "assert(" "access.is_parse_access()" ") failed", "entry not supported at optimization time" ); ::breakpoint(); } } while (0); | ||||
631 | C2ParseAccess& parse_access = static_cast<C2ParseAccess&>(access); | ||||
632 | GraphKit* kit = parse_access.kit(); | ||||
633 | |||||
634 | if (on_weak || on_phantom) { | ||||
635 | // Use the pre-barrier to record the value in the referent field | ||||
636 | pre_barrier(kit, false /* do_load */, | ||||
637 | kit->control(), | ||||
638 | NULL__null /* obj */, NULL__null /* adr */, max_juint /* alias_idx */, NULL__null /* val */, NULL__null /* val_type */, | ||||
639 | load /* pre_val */, T_OBJECT); | ||||
640 | // Add memory barrier to prevent commoning reads from this field | ||||
641 | // across safepoint since GC can change its value. | ||||
642 | kit->insert_mem_bar(Op_MemBarCPUOrder); | ||||
643 | } else if (unknown) { | ||||
644 | // We do not require a mem bar inside pre_barrier if need_mem_bar | ||||
645 | // is set: the barriers would be emitted by us. | ||||
646 | insert_pre_barrier(kit, obj, offset, load, !need_cpu_mem_bar); | ||||
647 | } | ||||
648 | |||||
649 | return load; | ||||
650 | } | ||||
651 | |||||
652 | bool G1BarrierSetC2::is_gc_barrier_node(Node* node) const { | ||||
653 | if (CardTableBarrierSetC2::is_gc_barrier_node(node)) { | ||||
654 | return true; | ||||
655 | } | ||||
656 | if (node->Opcode() != Op_CallLeaf) { | ||||
657 | return false; | ||||
658 | } | ||||
659 | CallLeafNode *call = node->as_CallLeaf(); | ||||
660 | if (call->_name == NULL__null) { | ||||
661 | return false; | ||||
662 | } | ||||
663 | |||||
664 | return strcmp(call->_name, "write_ref_field_pre_entry") == 0 || strcmp(call->_name, "write_ref_field_post_entry") == 0; | ||||
665 | } | ||||
666 | |||||
667 | void G1BarrierSetC2::eliminate_gc_barrier(PhaseMacroExpand* macro, Node* node) const { | ||||
668 | assert(node->Opcode() == Op_CastP2X, "ConvP2XNode required")do { if (!(node->Opcode() == Op_CastP2X)) { (*g_assert_poison ) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 668, "assert(" "node->Opcode() == Op_CastP2X" ") failed" , "ConvP2XNode required"); ::breakpoint(); } } while (0); | ||||
669 | assert(node->outcnt() <= 2, "expects 1 or 2 users: Xor and URShift nodes")do { if (!(node->outcnt() <= 2)) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 669, "assert(" "node->outcnt() <= 2" ") failed", "expects 1 or 2 users: Xor and URShift nodes" ); ::breakpoint(); } } while (0); | ||||
670 | // It could be only one user, URShift node, in Object.clone() intrinsic | ||||
671 | // but the new allocation is passed to arraycopy stub and it could not | ||||
672 | // be scalar replaced. So we don't check the case. | ||||
673 | |||||
674 | // An other case of only one user (Xor) is when the value check for NULL | ||||
675 | // in G1 post barrier is folded after CCP so the code which used URShift | ||||
676 | // is removed. | ||||
677 | |||||
678 | // Take Region node before eliminating post barrier since it also | ||||
679 | // eliminates CastP2X node when it has only one user. | ||||
680 | Node* this_region = node->in(0); | ||||
681 | assert(this_region != NULL, "")do { if (!(this_region != __null)) { (*g_assert_poison) = 'X' ;; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 681, "assert(" "this_region != __null" ") failed", ""); ::breakpoint (); } } while (0); | ||||
682 | |||||
683 | // Remove G1 post barrier. | ||||
684 | |||||
685 | // Search for CastP2X->Xor->URShift->Cmp path which | ||||
686 | // checks if the store done to a different from the value's region. | ||||
687 | // And replace Cmp with #0 (false) to collapse G1 post barrier. | ||||
688 | Node* xorx = node->find_out_with(Op_XorXOp_XorL); | ||||
689 | if (xorx != NULL__null) { | ||||
690 | Node* shift = xorx->unique_out(); | ||||
691 | Node* cmpx = shift->unique_out(); | ||||
692 | assert(cmpx->is_Cmp() && cmpx->unique_out()->is_Bool() &&do { if (!(cmpx->is_Cmp() && cmpx->unique_out() ->is_Bool() && cmpx->unique_out()->as_Bool() ->_test._test == BoolTest::ne)) { (*g_assert_poison) = 'X' ;; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 694, "assert(" "cmpx->is_Cmp() && cmpx->unique_out()->is_Bool() && cmpx->unique_out()->as_Bool()->_test._test == BoolTest::ne" ") failed", "missing region check in G1 post barrier"); ::breakpoint (); } } while (0) | ||||
693 | cmpx->unique_out()->as_Bool()->_test._test == BoolTest::ne,do { if (!(cmpx->is_Cmp() && cmpx->unique_out() ->is_Bool() && cmpx->unique_out()->as_Bool() ->_test._test == BoolTest::ne)) { (*g_assert_poison) = 'X' ;; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 694, "assert(" "cmpx->is_Cmp() && cmpx->unique_out()->is_Bool() && cmpx->unique_out()->as_Bool()->_test._test == BoolTest::ne" ") failed", "missing region check in G1 post barrier"); ::breakpoint (); } } while (0) | ||||
694 | "missing region check in G1 post barrier")do { if (!(cmpx->is_Cmp() && cmpx->unique_out() ->is_Bool() && cmpx->unique_out()->as_Bool() ->_test._test == BoolTest::ne)) { (*g_assert_poison) = 'X' ;; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 694, "assert(" "cmpx->is_Cmp() && cmpx->unique_out()->is_Bool() && cmpx->unique_out()->as_Bool()->_test._test == BoolTest::ne" ") failed", "missing region check in G1 post barrier"); ::breakpoint (); } } while (0); | ||||
695 | macro->replace_node(cmpx, macro->makecon(TypeInt::CC_EQ)); | ||||
696 | |||||
697 | // Remove G1 pre barrier. | ||||
698 | |||||
699 | // Search "if (marking != 0)" check and set it to "false". | ||||
700 | // There is no G1 pre barrier if previous stored value is NULL | ||||
701 | // (for example, after initialization). | ||||
702 | if (this_region->is_Region() && this_region->req() == 3) { | ||||
703 | int ind = 1; | ||||
704 | if (!this_region->in(ind)->is_IfFalse()) { | ||||
705 | ind = 2; | ||||
706 | } | ||||
707 | if (this_region->in(ind)->is_IfFalse() && | ||||
708 | this_region->in(ind)->in(0)->Opcode() == Op_If) { | ||||
709 | Node* bol = this_region->in(ind)->in(0)->in(1); | ||||
710 | assert(bol->is_Bool(), "")do { if (!(bol->is_Bool())) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 710, "assert(" "bol->is_Bool()" ") failed", ""); ::breakpoint (); } } while (0); | ||||
711 | cmpx = bol->in(1); | ||||
712 | if (bol->as_Bool()->_test._test == BoolTest::ne && | ||||
713 | cmpx->is_Cmp() && cmpx->in(2) == macro->intcon(0) && | ||||
714 | cmpx->in(1)->is_Load()) { | ||||
715 | Node* adr = cmpx->in(1)->as_Load()->in(MemNode::Address); | ||||
716 | const int marking_offset = in_bytes(G1ThreadLocalData::satb_mark_queue_active_offset()); | ||||
717 | if (adr->is_AddP() && adr->in(AddPNode::Base) == macro->top() && | ||||
718 | adr->in(AddPNode::Address)->Opcode() == Op_ThreadLocal && | ||||
719 | adr->in(AddPNode::Offset) == macro->MakeConXlongcon(marking_offset)) { | ||||
720 | macro->replace_node(cmpx, macro->makecon(TypeInt::CC_EQ)); | ||||
721 | } | ||||
722 | } | ||||
723 | } | ||||
724 | } | ||||
725 | } else { | ||||
726 | assert(!use_ReduceInitialCardMarks(), "can only happen with card marking")do { if (!(!use_ReduceInitialCardMarks())) { (*g_assert_poison ) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 726, "assert(" "!use_ReduceInitialCardMarks()" ") failed", "can only happen with card marking" ); ::breakpoint(); } } while (0); | ||||
727 | // This is a G1 post barrier emitted by the Object.clone() intrinsic. | ||||
728 | // Search for the CastP2X->URShiftX->AddP->LoadB->Cmp path which checks if the card | ||||
729 | // is marked as young_gen and replace the Cmp with 0 (false) to collapse the barrier. | ||||
730 | Node* shift = node->find_out_with(Op_URShiftXOp_URShiftL); | ||||
731 | assert(shift != NULL, "missing G1 post barrier")do { if (!(shift != __null)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 731, "assert(" "shift != __null" ") failed", "missing G1 post barrier" ); ::breakpoint(); } } while (0); | ||||
732 | Node* addp = shift->unique_out(); | ||||
733 | Node* load = addp->find_out_with(Op_LoadB); | ||||
734 | assert(load != NULL, "missing G1 post barrier")do { if (!(load != __null)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 734, "assert(" "load != __null" ") failed", "missing G1 post barrier" ); ::breakpoint(); } } while (0); | ||||
735 | Node* cmpx = load->unique_out(); | ||||
736 | assert(cmpx->is_Cmp() && cmpx->unique_out()->is_Bool() &&do { if (!(cmpx->is_Cmp() && cmpx->unique_out() ->is_Bool() && cmpx->unique_out()->as_Bool() ->_test._test == BoolTest::ne)) { (*g_assert_poison) = 'X' ;; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 738, "assert(" "cmpx->is_Cmp() && cmpx->unique_out()->is_Bool() && cmpx->unique_out()->as_Bool()->_test._test == BoolTest::ne" ") failed", "missing card value check in G1 post barrier"); :: breakpoint(); } } while (0) | ||||
737 | cmpx->unique_out()->as_Bool()->_test._test == BoolTest::ne,do { if (!(cmpx->is_Cmp() && cmpx->unique_out() ->is_Bool() && cmpx->unique_out()->as_Bool() ->_test._test == BoolTest::ne)) { (*g_assert_poison) = 'X' ;; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 738, "assert(" "cmpx->is_Cmp() && cmpx->unique_out()->is_Bool() && cmpx->unique_out()->as_Bool()->_test._test == BoolTest::ne" ") failed", "missing card value check in G1 post barrier"); :: breakpoint(); } } while (0) | ||||
738 | "missing card value check in G1 post barrier")do { if (!(cmpx->is_Cmp() && cmpx->unique_out() ->is_Bool() && cmpx->unique_out()->as_Bool() ->_test._test == BoolTest::ne)) { (*g_assert_poison) = 'X' ;; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 738, "assert(" "cmpx->is_Cmp() && cmpx->unique_out()->is_Bool() && cmpx->unique_out()->as_Bool()->_test._test == BoolTest::ne" ") failed", "missing card value check in G1 post barrier"); :: breakpoint(); } } while (0); | ||||
739 | macro->replace_node(cmpx, macro->makecon(TypeInt::CC_EQ)); | ||||
740 | // There is no G1 pre barrier in this case | ||||
741 | } | ||||
742 | // Now CastP2X can be removed since it is used only on dead path | ||||
743 | // which currently still alive until igvn optimize it. | ||||
744 | assert(node->outcnt() == 0 || node->unique_out()->Opcode() == Op_URShiftX, "")do { if (!(node->outcnt() == 0 || node->unique_out()-> Opcode() == Op_URShiftL)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 744, "assert(" "node->outcnt() == 0 || node->unique_out()->Opcode() == Op_URShiftL" ") failed", ""); ::breakpoint(); } } while (0); | ||||
745 | macro->replace_node(node, macro->top()); | ||||
746 | } | ||||
747 | |||||
748 | Node* G1BarrierSetC2::step_over_gc_barrier(Node* c) const { | ||||
749 | if (!use_ReduceInitialCardMarks() && | ||||
| |||||
750 | c != NULL__null && c->is_Region() && c->req() == 3) { | ||||
751 | for (uint i = 1; i < c->req(); i++) { | ||||
| |||||
752 | if (c->in(i) != NULL__null && c->in(i)->is_Region() && | ||||
753 | c->in(i)->req() == 3) { | ||||
754 | Node* r = c->in(i); | ||||
755 | for (uint j = 1; j < r->req(); j++) { | ||||
756 | if (r->in(j) != NULL__null && r->in(j)->is_Proj() && | ||||
757 | r->in(j)->in(0) != NULL__null && | ||||
758 | r->in(j)->in(0)->Opcode() == Op_CallLeaf && | ||||
759 | r->in(j)->in(0)->as_Call()->entry_point() == CAST_FROM_FN_PTR(address, G1BarrierSetRuntime::write_ref_field_post_entry)((address)((address_word)(G1BarrierSetRuntime::write_ref_field_post_entry )))) { | ||||
760 | Node* call = r->in(j)->in(0); | ||||
761 | c = c->in(i
| ||||
762 | if (c != NULL__null && c->Opcode() != Op_Parm) { | ||||
763 | c = c->in(0); | ||||
764 | if (c != NULL__null) { | ||||
765 | c = c->in(0); | ||||
766 | assert(call->in(0) == NULL ||do { if (!(call->in(0) == __null || call->in(0)->in( 0) == __null || call->in(0)->in(0)->in(0) == __null || call->in(0)->in(0)->in(0)->in(0) == __null || call ->in(0)->in(0)->in(0)->in(0)->in(0) == __null || c == call->in(0)->in(0)->in(0)->in(0)->in(0)) ) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 771, "assert(" "call->in(0) == __null || call->in(0)->in(0) == __null || call->in(0)->in(0)->in(0) == __null || call->in(0)->in(0)->in(0)->in(0) == __null || call->in(0)->in(0)->in(0)->in(0)->in(0) == __null || c == call->in(0)->in(0)->in(0)->in(0)->in(0)" ") failed", "bad barrier shape"); ::breakpoint(); } } while ( 0) | ||||
767 | call->in(0)->in(0) == NULL ||do { if (!(call->in(0) == __null || call->in(0)->in( 0) == __null || call->in(0)->in(0)->in(0) == __null || call->in(0)->in(0)->in(0)->in(0) == __null || call ->in(0)->in(0)->in(0)->in(0)->in(0) == __null || c == call->in(0)->in(0)->in(0)->in(0)->in(0)) ) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 771, "assert(" "call->in(0) == __null || call->in(0)->in(0) == __null || call->in(0)->in(0)->in(0) == __null || call->in(0)->in(0)->in(0)->in(0) == __null || call->in(0)->in(0)->in(0)->in(0)->in(0) == __null || c == call->in(0)->in(0)->in(0)->in(0)->in(0)" ") failed", "bad barrier shape"); ::breakpoint(); } } while ( 0) | ||||
768 | call->in(0)->in(0)->in(0) == NULL ||do { if (!(call->in(0) == __null || call->in(0)->in( 0) == __null || call->in(0)->in(0)->in(0) == __null || call->in(0)->in(0)->in(0)->in(0) == __null || call ->in(0)->in(0)->in(0)->in(0)->in(0) == __null || c == call->in(0)->in(0)->in(0)->in(0)->in(0)) ) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 771, "assert(" "call->in(0) == __null || call->in(0)->in(0) == __null || call->in(0)->in(0)->in(0) == __null || call->in(0)->in(0)->in(0)->in(0) == __null || call->in(0)->in(0)->in(0)->in(0)->in(0) == __null || c == call->in(0)->in(0)->in(0)->in(0)->in(0)" ") failed", "bad barrier shape"); ::breakpoint(); } } while ( 0) | ||||
769 | call->in(0)->in(0)->in(0)->in(0) == NULL ||do { if (!(call->in(0) == __null || call->in(0)->in( 0) == __null || call->in(0)->in(0)->in(0) == __null || call->in(0)->in(0)->in(0)->in(0) == __null || call ->in(0)->in(0)->in(0)->in(0)->in(0) == __null || c == call->in(0)->in(0)->in(0)->in(0)->in(0)) ) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 771, "assert(" "call->in(0) == __null || call->in(0)->in(0) == __null || call->in(0)->in(0)->in(0) == __null || call->in(0)->in(0)->in(0)->in(0) == __null || call->in(0)->in(0)->in(0)->in(0)->in(0) == __null || c == call->in(0)->in(0)->in(0)->in(0)->in(0)" ") failed", "bad barrier shape"); ::breakpoint(); } } while ( 0) | ||||
770 | call->in(0)->in(0)->in(0)->in(0)->in(0) == NULL ||do { if (!(call->in(0) == __null || call->in(0)->in( 0) == __null || call->in(0)->in(0)->in(0) == __null || call->in(0)->in(0)->in(0)->in(0) == __null || call ->in(0)->in(0)->in(0)->in(0)->in(0) == __null || c == call->in(0)->in(0)->in(0)->in(0)->in(0)) ) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 771, "assert(" "call->in(0) == __null || call->in(0)->in(0) == __null || call->in(0)->in(0)->in(0) == __null || call->in(0)->in(0)->in(0)->in(0) == __null || call->in(0)->in(0)->in(0)->in(0)->in(0) == __null || c == call->in(0)->in(0)->in(0)->in(0)->in(0)" ") failed", "bad barrier shape"); ::breakpoint(); } } while ( 0) | ||||
771 | c == call->in(0)->in(0)->in(0)->in(0)->in(0), "bad barrier shape")do { if (!(call->in(0) == __null || call->in(0)->in( 0) == __null || call->in(0)->in(0)->in(0) == __null || call->in(0)->in(0)->in(0)->in(0) == __null || call ->in(0)->in(0)->in(0)->in(0)->in(0) == __null || c == call->in(0)->in(0)->in(0)->in(0)->in(0)) ) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 771, "assert(" "call->in(0) == __null || call->in(0)->in(0) == __null || call->in(0)->in(0)->in(0) == __null || call->in(0)->in(0)->in(0)->in(0) == __null || call->in(0)->in(0)->in(0)->in(0)->in(0) == __null || c == call->in(0)->in(0)->in(0)->in(0)->in(0)" ") failed", "bad barrier shape"); ::breakpoint(); } } while ( 0); | ||||
772 | return c; | ||||
773 | } | ||||
774 | } | ||||
775 | } | ||||
776 | } | ||||
777 | } | ||||
778 | } | ||||
779 | } | ||||
780 | return c; | ||||
781 | } | ||||
782 | |||||
783 | #ifdef ASSERT1 | ||||
784 | void G1BarrierSetC2::verify_gc_barriers(Compile* compile, CompilePhase phase) const { | ||||
785 | if (phase != BarrierSetC2::BeforeCodeGen) { | ||||
786 | return; | ||||
787 | } | ||||
788 | // Verify G1 pre-barriers | ||||
789 | const int marking_offset = in_bytes(G1ThreadLocalData::satb_mark_queue_active_offset()); | ||||
790 | |||||
791 | Unique_Node_List visited; | ||||
792 | Node_List worklist; | ||||
793 | // We're going to walk control flow backwards starting from the Root | ||||
794 | worklist.push(compile->root()); | ||||
795 | while (worklist.size() > 0) { | ||||
796 | Node* x = worklist.pop(); | ||||
797 | if (x == NULL__null || x == compile->top()) continue; | ||||
798 | if (visited.member(x)) { | ||||
799 | continue; | ||||
800 | } else { | ||||
801 | visited.push(x); | ||||
802 | } | ||||
803 | |||||
804 | if (x->is_Region()) { | ||||
805 | for (uint i = 1; i < x->req(); i++) { | ||||
806 | worklist.push(x->in(i)); | ||||
807 | } | ||||
808 | } else { | ||||
809 | worklist.push(x->in(0)); | ||||
810 | // We are looking for the pattern: | ||||
811 | // /->ThreadLocal | ||||
812 | // If->Bool->CmpI->LoadB->AddP->ConL(marking_offset) | ||||
813 | // \->ConI(0) | ||||
814 | // We want to verify that the If and the LoadB have the same control | ||||
815 | // See GraphKit::g1_write_barrier_pre() | ||||
816 | if (x->is_If()) { | ||||
817 | IfNode *iff = x->as_If(); | ||||
818 | if (iff->in(1)->is_Bool() && iff->in(1)->in(1)->is_Cmp()) { | ||||
819 | CmpNode *cmp = iff->in(1)->in(1)->as_Cmp(); | ||||
820 | if (cmp->Opcode() == Op_CmpI && cmp->in(2)->is_Con() && cmp->in(2)->bottom_type()->is_int()->get_con() == 0 | ||||
821 | && cmp->in(1)->is_Load()) { | ||||
822 | LoadNode* load = cmp->in(1)->as_Load(); | ||||
823 | if (load->Opcode() == Op_LoadB && load->in(2)->is_AddP() && load->in(2)->in(2)->Opcode() == Op_ThreadLocal | ||||
824 | && load->in(2)->in(3)->is_Con() | ||||
825 | && load->in(2)->in(3)->bottom_type()->is_intptr_tis_long()->get_con() == marking_offset) { | ||||
826 | |||||
827 | Node* if_ctrl = iff->in(0); | ||||
828 | Node* load_ctrl = load->in(0); | ||||
829 | |||||
830 | if (if_ctrl != load_ctrl) { | ||||
831 | // Skip possible CProj->NeverBranch in infinite loops | ||||
832 | if ((if_ctrl->is_Proj() && if_ctrl->Opcode() == Op_CProj) | ||||
833 | && (if_ctrl->in(0)->is_MultiBranch() && if_ctrl->in(0)->Opcode() == Op_NeverBranch)) { | ||||
834 | if_ctrl = if_ctrl->in(0)->in(0); | ||||
835 | } | ||||
836 | } | ||||
837 | assert(load_ctrl != NULL && if_ctrl == load_ctrl, "controls must match")do { if (!(load_ctrl != __null && if_ctrl == load_ctrl )) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/gc/g1/c2/g1BarrierSetC2.cpp" , 837, "assert(" "load_ctrl != __null && if_ctrl == load_ctrl" ") failed", "controls must match"); ::breakpoint(); } } while (0); | ||||
838 | } | ||||
839 | } | ||||
840 | } | ||||
841 | } | ||||
842 | } | ||||
843 | } | ||||
844 | } | ||||
845 | #endif | ||||
846 | |||||
847 | bool G1BarrierSetC2::escape_add_to_con_graph(ConnectionGraph* conn_graph, PhaseGVN* gvn, Unique_Node_List* delayed_worklist, Node* n, uint opcode) const { | ||||
848 | if (opcode == Op_StoreP) { | ||||
849 | Node* adr = n->in(MemNode::Address); | ||||
850 | const Type* adr_type = gvn->type(adr); | ||||
851 | // Pointer stores in G1 barriers looks like unsafe access. | ||||
852 | // Ignore such stores to be able scalar replace non-escaping | ||||
853 | // allocations. | ||||
854 | if (adr_type->isa_rawptr() && adr->is_AddP()) { | ||||
855 | Node* base = conn_graph->get_addp_base(adr); | ||||
856 | if (base->Opcode() == Op_LoadP && | ||||
857 | base->in(MemNode::Address)->is_AddP()) { | ||||
858 | adr = base->in(MemNode::Address); | ||||
859 | Node* tls = conn_graph->get_addp_base(adr); | ||||
860 | if (tls->Opcode() == Op_ThreadLocal) { | ||||
861 | int offs = (int) gvn->find_intptr_t_confind_long_con(adr->in(AddPNode::Offset), Type::OffsetBot); | ||||
862 | const int buf_offset = in_bytes(G1ThreadLocalData::satb_mark_queue_buffer_offset()); | ||||
863 | if (offs == buf_offset) { | ||||
864 | return true; // G1 pre barrier previous oop value store. | ||||
865 | } | ||||
866 | if (offs == in_bytes(G1ThreadLocalData::dirty_card_queue_buffer_offset())) { | ||||
867 | return true; // G1 post barrier card address store. | ||||
868 | } | ||||
869 | } | ||||
870 | } | ||||
871 | } | ||||
872 | } | ||||
873 | return false; | ||||
874 | } |
1 | /* |
2 | * Copyright (c) 1997, 2021, Oracle and/or its affiliates. All rights reserved. |
3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
4 | * |
5 | * This code is free software; you can redistribute it and/or modify it |
6 | * under the terms of the GNU General Public License version 2 only, as |
7 | * published by the Free Software Foundation. |
8 | * |
9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
12 | * version 2 for more details (a copy is included in the LICENSE file that |
13 | * accompanied this code). |
14 | * |
15 | * You should have received a copy of the GNU General Public License version |
16 | * 2 along with this work; if not, write to the Free Software Foundation, |
17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
18 | * |
19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
20 | * or visit www.oracle.com if you need additional information or have any |
21 | * questions. |
22 | * |
23 | */ |
24 | |
25 | #ifndef SHARE_OPTO_NODE_HPP |
26 | #define SHARE_OPTO_NODE_HPP |
27 | |
28 | #include "libadt/vectset.hpp" |
29 | #include "opto/compile.hpp" |
30 | #include "opto/type.hpp" |
31 | #include "utilities/copy.hpp" |
32 | |
33 | // Portions of code courtesy of Clifford Click |
34 | |
35 | // Optimization - Graph Style |
36 | |
37 | |
38 | class AbstractLockNode; |
39 | class AddNode; |
40 | class AddPNode; |
41 | class AliasInfo; |
42 | class AllocateArrayNode; |
43 | class AllocateNode; |
44 | class ArrayCopyNode; |
45 | class BaseCountedLoopNode; |
46 | class BaseCountedLoopEndNode; |
47 | class BlackholeNode; |
48 | class Block; |
49 | class BoolNode; |
50 | class BoxLockNode; |
51 | class CMoveNode; |
52 | class CallDynamicJavaNode; |
53 | class CallJavaNode; |
54 | class CallLeafNode; |
55 | class CallLeafNoFPNode; |
56 | class CallNode; |
57 | class CallRuntimeNode; |
58 | class CallNativeNode; |
59 | class CallStaticJavaNode; |
60 | class CastFFNode; |
61 | class CastDDNode; |
62 | class CastVVNode; |
63 | class CastIINode; |
64 | class CastLLNode; |
65 | class CatchNode; |
66 | class CatchProjNode; |
67 | class CheckCastPPNode; |
68 | class ClearArrayNode; |
69 | class CmpNode; |
70 | class CodeBuffer; |
71 | class ConstraintCastNode; |
72 | class ConNode; |
73 | class CompareAndSwapNode; |
74 | class CompareAndExchangeNode; |
75 | class CountedLoopNode; |
76 | class CountedLoopEndNode; |
77 | class DecodeNarrowPtrNode; |
78 | class DecodeNNode; |
79 | class DecodeNKlassNode; |
80 | class EncodeNarrowPtrNode; |
81 | class EncodePNode; |
82 | class EncodePKlassNode; |
83 | class FastLockNode; |
84 | class FastUnlockNode; |
85 | class HaltNode; |
86 | class IfNode; |
87 | class IfProjNode; |
88 | class IfFalseNode; |
89 | class IfTrueNode; |
90 | class InitializeNode; |
91 | class JVMState; |
92 | class JumpNode; |
93 | class JumpProjNode; |
94 | class LoadNode; |
95 | class LoadStoreNode; |
96 | class LoadStoreConditionalNode; |
97 | class LockNode; |
98 | class LongCountedLoopNode; |
99 | class LongCountedLoopEndNode; |
100 | class LoopNode; |
101 | class LShiftNode; |
102 | class MachBranchNode; |
103 | class MachCallDynamicJavaNode; |
104 | class MachCallJavaNode; |
105 | class MachCallLeafNode; |
106 | class MachCallNode; |
107 | class MachCallNativeNode; |
108 | class MachCallRuntimeNode; |
109 | class MachCallStaticJavaNode; |
110 | class MachConstantBaseNode; |
111 | class MachConstantNode; |
112 | class MachGotoNode; |
113 | class MachIfNode; |
114 | class MachJumpNode; |
115 | class MachNode; |
116 | class MachNullCheckNode; |
117 | class MachProjNode; |
118 | class MachReturnNode; |
119 | class MachSafePointNode; |
120 | class MachSpillCopyNode; |
121 | class MachTempNode; |
122 | class MachMergeNode; |
123 | class MachMemBarNode; |
124 | class Matcher; |
125 | class MemBarNode; |
126 | class MemBarStoreStoreNode; |
127 | class MemNode; |
128 | class MergeMemNode; |
129 | class MoveNode; |
130 | class MulNode; |
131 | class MultiNode; |
132 | class MultiBranchNode; |
133 | class NeverBranchNode; |
134 | class Opaque1Node; |
135 | class OuterStripMinedLoopNode; |
136 | class OuterStripMinedLoopEndNode; |
137 | class Node; |
138 | class Node_Array; |
139 | class Node_List; |
140 | class Node_Stack; |
141 | class OopMap; |
142 | class ParmNode; |
143 | class PCTableNode; |
144 | class PhaseCCP; |
145 | class PhaseGVN; |
146 | class PhaseIterGVN; |
147 | class PhaseRegAlloc; |
148 | class PhaseTransform; |
149 | class PhaseValues; |
150 | class PhiNode; |
151 | class Pipeline; |
152 | class ProjNode; |
153 | class RangeCheckNode; |
154 | class RegMask; |
155 | class RegionNode; |
156 | class RootNode; |
157 | class SafePointNode; |
158 | class SafePointScalarObjectNode; |
159 | class StartNode; |
160 | class State; |
161 | class StoreNode; |
162 | class SubNode; |
163 | class SubTypeCheckNode; |
164 | class Type; |
165 | class TypeNode; |
166 | class UnlockNode; |
167 | class VectorNode; |
168 | class LoadVectorNode; |
169 | class LoadVectorMaskedNode; |
170 | class StoreVectorMaskedNode; |
171 | class LoadVectorGatherNode; |
172 | class StoreVectorNode; |
173 | class StoreVectorScatterNode; |
174 | class VectorMaskCmpNode; |
175 | class VectorUnboxNode; |
176 | class VectorSet; |
177 | class VectorReinterpretNode; |
178 | class ShiftVNode; |
179 | |
180 | // The type of all node counts and indexes. |
181 | // It must hold at least 16 bits, but must also be fast to load and store. |
182 | // This type, if less than 32 bits, could limit the number of possible nodes. |
183 | // (To make this type platform-specific, move to globalDefinitions_xxx.hpp.) |
184 | typedef unsigned int node_idx_t; |
185 | |
186 | |
187 | #ifndef OPTO_DU_ITERATOR_ASSERT1 |
188 | #ifdef ASSERT1 |
189 | #define OPTO_DU_ITERATOR_ASSERT1 1 |
190 | #else |
191 | #define OPTO_DU_ITERATOR_ASSERT1 0 |
192 | #endif |
193 | #endif //OPTO_DU_ITERATOR_ASSERT |
194 | |
195 | #if OPTO_DU_ITERATOR_ASSERT1 |
196 | class DUIterator; |
197 | class DUIterator_Fast; |
198 | class DUIterator_Last; |
199 | #else |
200 | typedef uint DUIterator; |
201 | typedef Node** DUIterator_Fast; |
202 | typedef Node** DUIterator_Last; |
203 | #endif |
204 | |
205 | // Node Sentinel |
206 | #define NodeSentinel(Node*)-1 (Node*)-1 |
207 | |
208 | // Unknown count frequency |
209 | #define COUNT_UNKNOWN(-1.0f) (-1.0f) |
210 | |
211 | //------------------------------Node------------------------------------------- |
212 | // Nodes define actions in the program. They create values, which have types. |
213 | // They are both vertices in a directed graph and program primitives. Nodes |
214 | // are labeled; the label is the "opcode", the primitive function in the lambda |
215 | // calculus sense that gives meaning to the Node. Node inputs are ordered (so |
216 | // that "a-b" is different from "b-a"). The inputs to a Node are the inputs to |
217 | // the Node's function. These inputs also define a Type equation for the Node. |
218 | // Solving these Type equations amounts to doing dataflow analysis. |
219 | // Control and data are uniformly represented in the graph. Finally, Nodes |
220 | // have a unique dense integer index which is used to index into side arrays |
221 | // whenever I have phase-specific information. |
222 | |
223 | class Node { |
224 | friend class VMStructs; |
225 | |
226 | // Lots of restrictions on cloning Nodes |
227 | NONCOPYABLE(Node)Node(Node const&) = delete; Node& operator=(Node const &) = delete; |
228 | |
229 | public: |
230 | friend class Compile; |
231 | #if OPTO_DU_ITERATOR_ASSERT1 |
232 | friend class DUIterator_Common; |
233 | friend class DUIterator; |
234 | friend class DUIterator_Fast; |
235 | friend class DUIterator_Last; |
236 | #endif |
237 | |
238 | // Because Nodes come and go, I define an Arena of Node structures to pull |
239 | // from. This should allow fast access to node creation & deletion. This |
240 | // field is a local cache of a value defined in some "program fragment" for |
241 | // which these Nodes are just a part of. |
242 | |
243 | inline void* operator new(size_t x) throw() { |
244 | Compile* C = Compile::current(); |
245 | Node* n = (Node*)C->node_arena()->AmallocWords(x); |
246 | return (void*)n; |
247 | } |
248 | |
249 | // Delete is a NOP |
250 | void operator delete( void *ptr ) {} |
251 | // Fancy destructor; eagerly attempt to reclaim Node numberings and storage |
252 | void destruct(PhaseValues* phase); |
253 | |
254 | // Create a new Node. Required is the number is of inputs required for |
255 | // semantic correctness. |
256 | Node( uint required ); |
257 | |
258 | // Create a new Node with given input edges. |
259 | // This version requires use of the "edge-count" new. |
260 | // E.g. new (C,3) FooNode( C, NULL, left, right ); |
261 | Node( Node *n0 ); |
262 | Node( Node *n0, Node *n1 ); |
263 | Node( Node *n0, Node *n1, Node *n2 ); |
264 | Node( Node *n0, Node *n1, Node *n2, Node *n3 ); |
265 | Node( Node *n0, Node *n1, Node *n2, Node *n3, Node *n4 ); |
266 | Node( Node *n0, Node *n1, Node *n2, Node *n3, Node *n4, Node *n5 ); |
267 | Node( Node *n0, Node *n1, Node *n2, Node *n3, |
268 | Node *n4, Node *n5, Node *n6 ); |
269 | |
270 | // Clone an inherited Node given only the base Node type. |
271 | Node* clone() const; |
272 | |
273 | // Clone a Node, immediately supplying one or two new edges. |
274 | // The first and second arguments, if non-null, replace in(1) and in(2), |
275 | // respectively. |
276 | Node* clone_with_data_edge(Node* in1, Node* in2 = NULL__null) const { |
277 | Node* nn = clone(); |
278 | if (in1 != NULL__null) nn->set_req(1, in1); |
279 | if (in2 != NULL__null) nn->set_req(2, in2); |
280 | return nn; |
281 | } |
282 | |
283 | private: |
284 | // Shared setup for the above constructors. |
285 | // Handles all interactions with Compile::current. |
286 | // Puts initial values in all Node fields except _idx. |
287 | // Returns the initial value for _idx, which cannot |
288 | // be initialized by assignment. |
289 | inline int Init(int req); |
290 | |
291 | //----------------- input edge handling |
292 | protected: |
293 | friend class PhaseCFG; // Access to address of _in array elements |
294 | Node **_in; // Array of use-def references to Nodes |
295 | Node **_out; // Array of def-use references to Nodes |
296 | |
297 | // Input edges are split into two categories. Required edges are required |
298 | // for semantic correctness; order is important and NULLs are allowed. |
299 | // Precedence edges are used to help determine execution order and are |
300 | // added, e.g., for scheduling purposes. They are unordered and not |
301 | // duplicated; they have no embedded NULLs. Edges from 0 to _cnt-1 |
302 | // are required, from _cnt to _max-1 are precedence edges. |
303 | node_idx_t _cnt; // Total number of required Node inputs. |
304 | |
305 | node_idx_t _max; // Actual length of input array. |
306 | |
307 | // Output edges are an unordered list of def-use edges which exactly |
308 | // correspond to required input edges which point from other nodes |
309 | // to this one. Thus the count of the output edges is the number of |
310 | // users of this node. |
311 | node_idx_t _outcnt; // Total number of Node outputs. |
312 | |
313 | node_idx_t _outmax; // Actual length of output array. |
314 | |
315 | // Grow the actual input array to the next larger power-of-2 bigger than len. |
316 | void grow( uint len ); |
317 | // Grow the output array to the next larger power-of-2 bigger than len. |
318 | void out_grow( uint len ); |
319 | |
320 | public: |
321 | // Each Node is assigned a unique small/dense number. This number is used |
322 | // to index into auxiliary arrays of data and bit vectors. |
323 | // The field _idx is declared constant to defend against inadvertent assignments, |
324 | // since it is used by clients as a naked field. However, the field's value can be |
325 | // changed using the set_idx() method. |
326 | // |
327 | // The PhaseRenumberLive phase renumbers nodes based on liveness information. |
328 | // Therefore, it updates the value of the _idx field. The parse-time _idx is |
329 | // preserved in _parse_idx. |
330 | const node_idx_t _idx; |
331 | DEBUG_ONLY(const node_idx_t _parse_idx;)const node_idx_t _parse_idx; |
332 | // IGV node identifier. Two nodes, possibly in different compilation phases, |
333 | // have the same IGV identifier if (and only if) they are the very same node |
334 | // (same memory address) or one is "derived" from the other (by e.g. |
335 | // renumbering or matching). This identifier makes it possible to follow the |
336 | // entire lifetime of a node in IGV even if its C2 identifier (_idx) changes. |
337 | NOT_PRODUCT(node_idx_t _igv_idx;)node_idx_t _igv_idx; |
338 | |
339 | // Get the (read-only) number of input edges |
340 | uint req() const { return _cnt; } |
341 | uint len() const { return _max; } |
342 | // Get the (read-only) number of output edges |
343 | uint outcnt() const { return _outcnt; } |
344 | |
345 | #if OPTO_DU_ITERATOR_ASSERT1 |
346 | // Iterate over the out-edges of this node. Deletions are illegal. |
347 | inline DUIterator outs() const; |
348 | // Use this when the out array might have changed to suppress asserts. |
349 | inline DUIterator& refresh_out_pos(DUIterator& i) const; |
350 | // Does the node have an out at this position? (Used for iteration.) |
351 | inline bool has_out(DUIterator& i) const; |
352 | inline Node* out(DUIterator& i) const; |
353 | // Iterate over the out-edges of this node. All changes are illegal. |
354 | inline DUIterator_Fast fast_outs(DUIterator_Fast& max) const; |
355 | inline Node* fast_out(DUIterator_Fast& i) const; |
356 | // Iterate over the out-edges of this node, deleting one at a time. |
357 | inline DUIterator_Last last_outs(DUIterator_Last& min) const; |
358 | inline Node* last_out(DUIterator_Last& i) const; |
359 | // The inline bodies of all these methods are after the iterator definitions. |
360 | #else |
361 | // Iterate over the out-edges of this node. Deletions are illegal. |
362 | // This iteration uses integral indexes, to decouple from array reallocations. |
363 | DUIterator outs() const { return 0; } |
364 | // Use this when the out array might have changed to suppress asserts. |
365 | DUIterator refresh_out_pos(DUIterator i) const { return i; } |
366 | |
367 | // Reference to the i'th output Node. Error if out of bounds. |
368 | Node* out(DUIterator i) const { assert(i < _outcnt, "oob")do { if (!(i < _outcnt)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 368, "assert(" "i < _outcnt" ") failed", "oob"); ::breakpoint (); } } while (0); return _out[i]; } |
369 | // Does the node have an out at this position? (Used for iteration.) |
370 | bool has_out(DUIterator i) const { return i < _outcnt; } |
371 | |
372 | // Iterate over the out-edges of this node. All changes are illegal. |
373 | // This iteration uses a pointer internal to the out array. |
374 | DUIterator_Fast fast_outs(DUIterator_Fast& max) const { |
375 | Node** out = _out; |
376 | // Assign a limit pointer to the reference argument: |
377 | max = out + (ptrdiff_t)_outcnt; |
378 | // Return the base pointer: |
379 | return out; |
380 | } |
381 | Node* fast_out(DUIterator_Fast i) const { return *i; } |
382 | // Iterate over the out-edges of this node, deleting one at a time. |
383 | // This iteration uses a pointer internal to the out array. |
384 | DUIterator_Last last_outs(DUIterator_Last& min) const { |
385 | Node** out = _out; |
386 | // Assign a limit pointer to the reference argument: |
387 | min = out; |
388 | // Return the pointer to the start of the iteration: |
389 | return out + (ptrdiff_t)_outcnt - 1; |
390 | } |
391 | Node* last_out(DUIterator_Last i) const { return *i; } |
392 | #endif |
393 | |
394 | // Reference to the i'th input Node. Error if out of bounds. |
395 | Node* in(uint i) const { assert(i < _max, "oob: i=%d, _max=%d", i, _max)do { if (!(i < _max)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 395, "assert(" "i < _max" ") failed", "oob: i=%d, _max=%d" , i, _max); ::breakpoint(); } } while (0); return _in[i]; } |
396 | // Reference to the i'th input Node. NULL if out of bounds. |
397 | Node* lookup(uint i) const { return ((i < _max) ? _in[i] : NULL__null); } |
398 | // Reference to the i'th output Node. Error if out of bounds. |
399 | // Use this accessor sparingly. We are going trying to use iterators instead. |
400 | Node* raw_out(uint i) const { assert(i < _outcnt,"oob")do { if (!(i < _outcnt)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 400, "assert(" "i < _outcnt" ") failed", "oob"); ::breakpoint (); } } while (0); return _out[i]; } |
401 | // Return the unique out edge. |
402 | Node* unique_out() const { assert(_outcnt==1,"not unique")do { if (!(_outcnt==1)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 402, "assert(" "_outcnt==1" ") failed", "not unique"); ::breakpoint (); } } while (0); return _out[0]; } |
403 | // Delete out edge at position 'i' by moving last out edge to position 'i' |
404 | void raw_del_out(uint i) { |
405 | assert(i < _outcnt,"oob")do { if (!(i < _outcnt)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 405, "assert(" "i < _outcnt" ") failed", "oob"); ::breakpoint (); } } while (0); |
406 | assert(_outcnt > 0,"oob")do { if (!(_outcnt > 0)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 406, "assert(" "_outcnt > 0" ") failed", "oob"); ::breakpoint (); } } while (0); |
407 | #if OPTO_DU_ITERATOR_ASSERT1 |
408 | // Record that a change happened here. |
409 | debug_only(_last_del = _out[i]; ++_del_tick)_last_del = _out[i]; ++_del_tick; |
410 | #endif |
411 | _out[i] = _out[--_outcnt]; |
412 | // Smash the old edge so it can't be used accidentally. |
413 | debug_only(_out[_outcnt] = (Node *)(uintptr_t)0xdeadbeef)_out[_outcnt] = (Node *)(uintptr_t)0xdeadbeef; |
414 | } |
415 | |
416 | #ifdef ASSERT1 |
417 | bool is_dead() const; |
418 | #define is_not_dead(n)((n) == __null || !VerifyIterativeGVN || !((n)->is_dead()) ) ((n) == NULL__null || !VerifyIterativeGVN || !((n)->is_dead())) |
419 | bool is_reachable_from_root() const; |
420 | #endif |
421 | // Check whether node has become unreachable |
422 | bool is_unreachable(PhaseIterGVN &igvn) const; |
423 | |
424 | // Set a required input edge, also updates corresponding output edge |
425 | void add_req( Node *n ); // Append a NEW required input |
426 | void add_req( Node *n0, Node *n1 ) { |
427 | add_req(n0); add_req(n1); } |
428 | void add_req( Node *n0, Node *n1, Node *n2 ) { |
429 | add_req(n0); add_req(n1); add_req(n2); } |
430 | void add_req_batch( Node* n, uint m ); // Append m NEW required inputs (all n). |
431 | void del_req( uint idx ); // Delete required edge & compact |
432 | void del_req_ordered( uint idx ); // Delete required edge & compact with preserved order |
433 | void ins_req( uint i, Node *n ); // Insert a NEW required input |
434 | void set_req( uint i, Node *n ) { |
435 | assert( is_not_dead(n), "can not use dead node")do { if (!(((n) == __null || !VerifyIterativeGVN || !((n)-> is_dead())))) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 435, "assert(" "((n) == __null || !VerifyIterativeGVN || !((n)->is_dead()))" ") failed", "can not use dead node"); ::breakpoint(); } } while (0); |
436 | assert( i < _cnt, "oob: i=%d, _cnt=%d", i, _cnt)do { if (!(i < _cnt)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 436, "assert(" "i < _cnt" ") failed", "oob: i=%d, _cnt=%d" , i, _cnt); ::breakpoint(); } } while (0); |
437 | assert( !VerifyHashTableKeys || _hash_lock == 0,do { if (!(!VerifyHashTableKeys || _hash_lock == 0)) { (*g_assert_poison ) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 438, "assert(" "!VerifyHashTableKeys || _hash_lock == 0" ") failed" , "remove node from hash table before modifying it"); ::breakpoint (); } } while (0) |
438 | "remove node from hash table before modifying it")do { if (!(!VerifyHashTableKeys || _hash_lock == 0)) { (*g_assert_poison ) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 438, "assert(" "!VerifyHashTableKeys || _hash_lock == 0" ") failed" , "remove node from hash table before modifying it"); ::breakpoint (); } } while (0); |
439 | Node** p = &_in[i]; // cache this._in, across the del_out call |
440 | if (*p != NULL__null) (*p)->del_out((Node *)this); |
441 | (*p) = n; |
442 | if (n != NULL__null) n->add_out((Node *)this); |
443 | Compile::current()->record_modified_node(this); |
444 | } |
445 | // Light version of set_req() to init inputs after node creation. |
446 | void init_req( uint i, Node *n ) { |
447 | assert( i == 0 && this == n ||do { if (!(i == 0 && this == n || ((n) == __null || ! VerifyIterativeGVN || !((n)->is_dead())))) { (*g_assert_poison ) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 448, "assert(" "i == 0 && this == n || ((n) == __null || !VerifyIterativeGVN || !((n)->is_dead()))" ") failed", "can not use dead node"); ::breakpoint(); } } while (0) |
448 | is_not_dead(n), "can not use dead node")do { if (!(i == 0 && this == n || ((n) == __null || ! VerifyIterativeGVN || !((n)->is_dead())))) { (*g_assert_poison ) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 448, "assert(" "i == 0 && this == n || ((n) == __null || !VerifyIterativeGVN || !((n)->is_dead()))" ") failed", "can not use dead node"); ::breakpoint(); } } while (0); |
449 | assert( i < _cnt, "oob")do { if (!(i < _cnt)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 449, "assert(" "i < _cnt" ") failed", "oob"); ::breakpoint (); } } while (0); |
450 | assert( !VerifyHashTableKeys || _hash_lock == 0,do { if (!(!VerifyHashTableKeys || _hash_lock == 0)) { (*g_assert_poison ) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 451, "assert(" "!VerifyHashTableKeys || _hash_lock == 0" ") failed" , "remove node from hash table before modifying it"); ::breakpoint (); } } while (0) |
451 | "remove node from hash table before modifying it")do { if (!(!VerifyHashTableKeys || _hash_lock == 0)) { (*g_assert_poison ) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 451, "assert(" "!VerifyHashTableKeys || _hash_lock == 0" ") failed" , "remove node from hash table before modifying it"); ::breakpoint (); } } while (0); |
452 | assert( _in[i] == NULL, "sanity")do { if (!(_in[i] == __null)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 452, "assert(" "_in[i] == __null" ") failed", "sanity"); :: breakpoint(); } } while (0); |
453 | _in[i] = n; |
454 | if (n != NULL__null) n->add_out((Node *)this); |
455 | Compile::current()->record_modified_node(this); |
456 | } |
457 | // Find first occurrence of n among my edges: |
458 | int find_edge(Node* n); |
459 | int find_prec_edge(Node* n) { |
460 | for (uint i = req(); i < len(); i++) { |
461 | if (_in[i] == n) return i; |
462 | if (_in[i] == NULL__null) { |
463 | DEBUG_ONLY( while ((++i) < len()) assert(_in[i] == NULL, "Gap in prec edges!"); )while ((++i) < len()) do { if (!(_in[i] == __null)) { (*g_assert_poison ) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 463, "assert(" "_in[i] == __null" ") failed", "Gap in prec edges!" ); ::breakpoint(); } } while (0); |
464 | break; |
465 | } |
466 | } |
467 | return -1; |
468 | } |
469 | int replace_edge(Node* old, Node* neww, PhaseGVN* gvn = NULL__null); |
470 | int replace_edges_in_range(Node* old, Node* neww, int start, int end, PhaseGVN* gvn); |
471 | // NULL out all inputs to eliminate incoming Def-Use edges. |
472 | void disconnect_inputs(Compile* C); |
473 | |
474 | // Quickly, return true if and only if I am Compile::current()->top(). |
475 | bool is_top() const { |
476 | assert((this == (Node*) Compile::current()->top()) == (_out == NULL), "")do { if (!((this == (Node*) Compile::current()->top()) == ( _out == __null))) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 476, "assert(" "(this == (Node*) Compile::current()->top()) == (_out == __null)" ") failed", ""); ::breakpoint(); } } while (0); |
477 | return (_out == NULL__null); |
478 | } |
479 | // Reaffirm invariants for is_top. (Only from Compile::set_cached_top_node.) |
480 | void setup_is_top(); |
481 | |
482 | // Strip away casting. (It is depth-limited.) |
483 | Node* uncast(bool keep_deps = false) const; |
484 | // Return whether two Nodes are equivalent, after stripping casting. |
485 | bool eqv_uncast(const Node* n, bool keep_deps = false) const { |
486 | return (this->uncast(keep_deps) == n->uncast(keep_deps)); |
487 | } |
488 | |
489 | // Find out of current node that matches opcode. |
490 | Node* find_out_with(int opcode); |
491 | // Return true if the current node has an out that matches opcode. |
492 | bool has_out_with(int opcode); |
493 | // Return true if the current node has an out that matches any of the opcodes. |
494 | bool has_out_with(int opcode1, int opcode2, int opcode3, int opcode4); |
495 | |
496 | private: |
497 | static Node* uncast_helper(const Node* n, bool keep_deps); |
498 | |
499 | // Add an output edge to the end of the list |
500 | void add_out( Node *n ) { |
501 | if (is_top()) return; |
502 | if( _outcnt == _outmax ) out_grow(_outcnt); |
503 | _out[_outcnt++] = n; |
504 | } |
505 | // Delete an output edge |
506 | void del_out( Node *n ) { |
507 | if (is_top()) return; |
508 | Node** outp = &_out[_outcnt]; |
509 | // Find and remove n |
510 | do { |
511 | assert(outp > _out, "Missing Def-Use edge")do { if (!(outp > _out)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 511, "assert(" "outp > _out" ") failed", "Missing Def-Use edge" ); ::breakpoint(); } } while (0); |
512 | } while (*--outp != n); |
513 | *outp = _out[--_outcnt]; |
514 | // Smash the old edge so it can't be used accidentally. |
515 | debug_only(_out[_outcnt] = (Node *)(uintptr_t)0xdeadbeef)_out[_outcnt] = (Node *)(uintptr_t)0xdeadbeef; |
516 | // Record that a change happened here. |
517 | #if OPTO_DU_ITERATOR_ASSERT1 |
518 | debug_only(_last_del = n; ++_del_tick)_last_del = n; ++_del_tick; |
519 | #endif |
520 | } |
521 | // Close gap after removing edge. |
522 | void close_prec_gap_at(uint gap) { |
523 | assert(_cnt <= gap && gap < _max, "no valid prec edge")do { if (!(_cnt <= gap && gap < _max)) { (*g_assert_poison ) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 523, "assert(" "_cnt <= gap && gap < _max" ") failed" , "no valid prec edge"); ::breakpoint(); } } while (0); |
524 | uint i = gap; |
525 | Node *last = NULL__null; |
526 | for (; i < _max-1; ++i) { |
527 | Node *next = _in[i+1]; |
528 | if (next == NULL__null) break; |
529 | last = next; |
530 | } |
531 | _in[gap] = last; // Move last slot to empty one. |
532 | _in[i] = NULL__null; // NULL out last slot. |
533 | } |
534 | |
535 | public: |
536 | // Globally replace this node by a given new node, updating all uses. |
537 | void replace_by(Node* new_node); |
538 | // Globally replace this node by a given new node, updating all uses |
539 | // and cutting input edges of old node. |
540 | void subsume_by(Node* new_node, Compile* c) { |
541 | replace_by(new_node); |
542 | disconnect_inputs(c); |
543 | } |
544 | void set_req_X(uint i, Node *n, PhaseIterGVN *igvn); |
545 | void set_req_X(uint i, Node *n, PhaseGVN *gvn); |
546 | // Find the one non-null required input. RegionNode only |
547 | Node *nonnull_req() const; |
548 | // Add or remove precedence edges |
549 | void add_prec( Node *n ); |
550 | void rm_prec( uint i ); |
551 | |
552 | // Note: prec(i) will not necessarily point to n if edge already exists. |
553 | void set_prec( uint i, Node *n ) { |
554 | assert(i < _max, "oob: i=%d, _max=%d", i, _max)do { if (!(i < _max)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 554, "assert(" "i < _max" ") failed", "oob: i=%d, _max=%d" , i, _max); ::breakpoint(); } } while (0); |
555 | assert(is_not_dead(n), "can not use dead node")do { if (!(((n) == __null || !VerifyIterativeGVN || !((n)-> is_dead())))) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 555, "assert(" "((n) == __null || !VerifyIterativeGVN || !((n)->is_dead()))" ") failed", "can not use dead node"); ::breakpoint(); } } while (0); |
556 | assert(i >= _cnt, "not a precedence edge")do { if (!(i >= _cnt)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 556, "assert(" "i >= _cnt" ") failed", "not a precedence edge" ); ::breakpoint(); } } while (0); |
557 | // Avoid spec violation: duplicated prec edge. |
558 | if (_in[i] == n) return; |
559 | if (n == NULL__null || find_prec_edge(n) != -1) { |
560 | rm_prec(i); |
561 | return; |
562 | } |
563 | if (_in[i] != NULL__null) _in[i]->del_out((Node *)this); |
564 | _in[i] = n; |
565 | n->add_out((Node *)this); |
566 | } |
567 | |
568 | // Set this node's index, used by cisc_version to replace current node |
569 | void set_idx(uint new_idx) { |
570 | const node_idx_t* ref = &_idx; |
571 | *(node_idx_t*)ref = new_idx; |
572 | } |
573 | // Swap input edge order. (Edge indexes i1 and i2 are usually 1 and 2.) |
574 | void swap_edges(uint i1, uint i2) { |
575 | debug_only(uint check_hash = (VerifyHashTableKeys && _hash_lock) ? hash() : NO_HASH)uint check_hash = (VerifyHashTableKeys && _hash_lock) ? hash() : NO_HASH; |
576 | // Def-Use info is unchanged |
577 | Node* n1 = in(i1); |
578 | Node* n2 = in(i2); |
579 | _in[i1] = n2; |
580 | _in[i2] = n1; |
581 | // If this node is in the hash table, make sure it doesn't need a rehash. |
582 | assert(check_hash == NO_HASH || check_hash == hash(), "edge swap must preserve hash code")do { if (!(check_hash == NO_HASH || check_hash == hash())) { ( *g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 582, "assert(" "check_hash == NO_HASH || check_hash == hash()" ") failed", "edge swap must preserve hash code"); ::breakpoint (); } } while (0); |
583 | } |
584 | |
585 | // Iterators over input Nodes for a Node X are written as: |
586 | // for( i = 0; i < X.req(); i++ ) ... X[i] ... |
587 | // NOTE: Required edges can contain embedded NULL pointers. |
588 | |
589 | //----------------- Other Node Properties |
590 | |
591 | // Generate class IDs for (some) ideal nodes so that it is possible to determine |
592 | // the type of a node using a non-virtual method call (the method is_<Node>() below). |
593 | // |
594 | // A class ID of an ideal node is a set of bits. In a class ID, a single bit determines |
595 | // the type of the node the ID represents; another subset of an ID's bits are reserved |
596 | // for the superclasses of the node represented by the ID. |
597 | // |
598 | // By design, if A is a supertype of B, A.is_B() returns true and B.is_A() |
599 | // returns false. A.is_A() returns true. |
600 | // |
601 | // If two classes, A and B, have the same superclass, a different bit of A's class id |
602 | // is reserved for A's type than for B's type. That bit is specified by the third |
603 | // parameter in the macro DEFINE_CLASS_ID. |
604 | // |
605 | // By convention, classes with deeper hierarchy are declared first. Moreover, |
606 | // classes with the same hierarchy depth are sorted by usage frequency. |
607 | // |
608 | // The query method masks the bits to cut off bits of subclasses and then compares |
609 | // the result with the class id (see the macro DEFINE_CLASS_QUERY below). |
610 | // |
611 | // Class_MachCall=30, ClassMask_MachCall=31 |
612 | // 12 8 4 0 |
613 | // 0 0 0 0 0 0 0 0 1 1 1 1 0 |
614 | // | | | | |
615 | // | | | Bit_Mach=2 |
616 | // | | Bit_MachReturn=4 |
617 | // | Bit_MachSafePoint=8 |
618 | // Bit_MachCall=16 |
619 | // |
620 | // Class_CountedLoop=56, ClassMask_CountedLoop=63 |
621 | // 12 8 4 0 |
622 | // 0 0 0 0 0 0 0 1 1 1 0 0 0 |
623 | // | | | |
624 | // | | Bit_Region=8 |
625 | // | Bit_Loop=16 |
626 | // Bit_CountedLoop=32 |
627 | |
628 | #define DEFINE_CLASS_ID(cl, supcl, subn) \ |
629 | Bit_##cl = (Class_##supcl == 0) ? 1 << subn : (Bit_##supcl) << (1 + subn) , \ |
630 | Class_##cl = Class_##supcl + Bit_##cl , \ |
631 | ClassMask_##cl = ((Bit_##cl << 1) - 1) , |
632 | |
633 | // This enum is used only for C2 ideal and mach nodes with is_<node>() methods |
634 | // so that its values fit into 32 bits. |
635 | enum NodeClasses { |
636 | Bit_Node = 0x00000000, |
637 | Class_Node = 0x00000000, |
638 | ClassMask_Node = 0xFFFFFFFF, |
639 | |
640 | DEFINE_CLASS_ID(Multi, Node, 0) |
641 | DEFINE_CLASS_ID(SafePoint, Multi, 0) |
642 | DEFINE_CLASS_ID(Call, SafePoint, 0) |
643 | DEFINE_CLASS_ID(CallJava, Call, 0) |
644 | DEFINE_CLASS_ID(CallStaticJava, CallJava, 0) |
645 | DEFINE_CLASS_ID(CallDynamicJava, CallJava, 1) |
646 | DEFINE_CLASS_ID(CallRuntime, Call, 1) |
647 | DEFINE_CLASS_ID(CallLeaf, CallRuntime, 0) |
648 | DEFINE_CLASS_ID(CallLeafNoFP, CallLeaf, 0) |
649 | DEFINE_CLASS_ID(Allocate, Call, 2) |
650 | DEFINE_CLASS_ID(AllocateArray, Allocate, 0) |
651 | DEFINE_CLASS_ID(AbstractLock, Call, 3) |
652 | DEFINE_CLASS_ID(Lock, AbstractLock, 0) |
653 | DEFINE_CLASS_ID(Unlock, AbstractLock, 1) |
654 | DEFINE_CLASS_ID(ArrayCopy, Call, 4) |
655 | DEFINE_CLASS_ID(CallNative, Call, 5) |
656 | DEFINE_CLASS_ID(MultiBranch, Multi, 1) |
657 | DEFINE_CLASS_ID(PCTable, MultiBranch, 0) |
658 | DEFINE_CLASS_ID(Catch, PCTable, 0) |
659 | DEFINE_CLASS_ID(Jump, PCTable, 1) |
660 | DEFINE_CLASS_ID(If, MultiBranch, 1) |
661 | DEFINE_CLASS_ID(BaseCountedLoopEnd, If, 0) |
662 | DEFINE_CLASS_ID(CountedLoopEnd, BaseCountedLoopEnd, 0) |
663 | DEFINE_CLASS_ID(LongCountedLoopEnd, BaseCountedLoopEnd, 1) |
664 | DEFINE_CLASS_ID(RangeCheck, If, 1) |
665 | DEFINE_CLASS_ID(OuterStripMinedLoopEnd, If, 2) |
666 | DEFINE_CLASS_ID(NeverBranch, MultiBranch, 2) |
667 | DEFINE_CLASS_ID(Start, Multi, 2) |
668 | DEFINE_CLASS_ID(MemBar, Multi, 3) |
669 | DEFINE_CLASS_ID(Initialize, MemBar, 0) |
670 | DEFINE_CLASS_ID(MemBarStoreStore, MemBar, 1) |
671 | |
672 | DEFINE_CLASS_ID(Mach, Node, 1) |
673 | DEFINE_CLASS_ID(MachReturn, Mach, 0) |
674 | DEFINE_CLASS_ID(MachSafePoint, MachReturn, 0) |
675 | DEFINE_CLASS_ID(MachCall, MachSafePoint, 0) |
676 | DEFINE_CLASS_ID(MachCallJava, MachCall, 0) |
677 | DEFINE_CLASS_ID(MachCallStaticJava, MachCallJava, 0) |
678 | DEFINE_CLASS_ID(MachCallDynamicJava, MachCallJava, 1) |
679 | DEFINE_CLASS_ID(MachCallRuntime, MachCall, 1) |
680 | DEFINE_CLASS_ID(MachCallLeaf, MachCallRuntime, 0) |
681 | DEFINE_CLASS_ID(MachCallNative, MachCall, 2) |
682 | DEFINE_CLASS_ID(MachBranch, Mach, 1) |
683 | DEFINE_CLASS_ID(MachIf, MachBranch, 0) |
684 | DEFINE_CLASS_ID(MachGoto, MachBranch, 1) |
685 | DEFINE_CLASS_ID(MachNullCheck, MachBranch, 2) |
686 | DEFINE_CLASS_ID(MachSpillCopy, Mach, 2) |
687 | DEFINE_CLASS_ID(MachTemp, Mach, 3) |
688 | DEFINE_CLASS_ID(MachConstantBase, Mach, 4) |
689 | DEFINE_CLASS_ID(MachConstant, Mach, 5) |
690 | DEFINE_CLASS_ID(MachJump, MachConstant, 0) |
691 | DEFINE_CLASS_ID(MachMerge, Mach, 6) |
692 | DEFINE_CLASS_ID(MachMemBar, Mach, 7) |
693 | |
694 | DEFINE_CLASS_ID(Type, Node, 2) |
695 | DEFINE_CLASS_ID(Phi, Type, 0) |
696 | DEFINE_CLASS_ID(ConstraintCast, Type, 1) |
697 | DEFINE_CLASS_ID(CastII, ConstraintCast, 0) |
698 | DEFINE_CLASS_ID(CheckCastPP, ConstraintCast, 1) |
699 | DEFINE_CLASS_ID(CastLL, ConstraintCast, 2) |
700 | DEFINE_CLASS_ID(CastFF, ConstraintCast, 3) |
701 | DEFINE_CLASS_ID(CastDD, ConstraintCast, 4) |
702 | DEFINE_CLASS_ID(CastVV, ConstraintCast, 5) |
703 | DEFINE_CLASS_ID(CMove, Type, 3) |
704 | DEFINE_CLASS_ID(SafePointScalarObject, Type, 4) |
705 | DEFINE_CLASS_ID(DecodeNarrowPtr, Type, 5) |
706 | DEFINE_CLASS_ID(DecodeN, DecodeNarrowPtr, 0) |
707 | DEFINE_CLASS_ID(DecodeNKlass, DecodeNarrowPtr, 1) |
708 | DEFINE_CLASS_ID(EncodeNarrowPtr, Type, 6) |
709 | DEFINE_CLASS_ID(EncodeP, EncodeNarrowPtr, 0) |
710 | DEFINE_CLASS_ID(EncodePKlass, EncodeNarrowPtr, 1) |
711 | DEFINE_CLASS_ID(Vector, Type, 7) |
712 | DEFINE_CLASS_ID(VectorMaskCmp, Vector, 0) |
713 | DEFINE_CLASS_ID(VectorUnbox, Vector, 1) |
714 | DEFINE_CLASS_ID(VectorReinterpret, Vector, 2) |
715 | DEFINE_CLASS_ID(ShiftV, Vector, 3) |
716 | |
717 | DEFINE_CLASS_ID(Proj, Node, 3) |
718 | DEFINE_CLASS_ID(CatchProj, Proj, 0) |
719 | DEFINE_CLASS_ID(JumpProj, Proj, 1) |
720 | DEFINE_CLASS_ID(IfProj, Proj, 2) |
721 | DEFINE_CLASS_ID(IfTrue, IfProj, 0) |
722 | DEFINE_CLASS_ID(IfFalse, IfProj, 1) |
723 | DEFINE_CLASS_ID(Parm, Proj, 4) |
724 | DEFINE_CLASS_ID(MachProj, Proj, 5) |
725 | |
726 | DEFINE_CLASS_ID(Mem, Node, 4) |
727 | DEFINE_CLASS_ID(Load, Mem, 0) |
728 | DEFINE_CLASS_ID(LoadVector, Load, 0) |
729 | DEFINE_CLASS_ID(LoadVectorGather, LoadVector, 0) |
730 | DEFINE_CLASS_ID(LoadVectorMasked, LoadVector, 1) |
731 | DEFINE_CLASS_ID(Store, Mem, 1) |
732 | DEFINE_CLASS_ID(StoreVector, Store, 0) |
733 | DEFINE_CLASS_ID(StoreVectorScatter, StoreVector, 0) |
734 | DEFINE_CLASS_ID(StoreVectorMasked, StoreVector, 1) |
735 | DEFINE_CLASS_ID(LoadStore, Mem, 2) |
736 | DEFINE_CLASS_ID(LoadStoreConditional, LoadStore, 0) |
737 | DEFINE_CLASS_ID(CompareAndSwap, LoadStoreConditional, 0) |
738 | DEFINE_CLASS_ID(CompareAndExchangeNode, LoadStore, 1) |
739 | |
740 | DEFINE_CLASS_ID(Region, Node, 5) |
741 | DEFINE_CLASS_ID(Loop, Region, 0) |
742 | DEFINE_CLASS_ID(Root, Loop, 0) |
743 | DEFINE_CLASS_ID(BaseCountedLoop, Loop, 1) |
744 | DEFINE_CLASS_ID(CountedLoop, BaseCountedLoop, 0) |
745 | DEFINE_CLASS_ID(LongCountedLoop, BaseCountedLoop, 1) |
746 | DEFINE_CLASS_ID(OuterStripMinedLoop, Loop, 2) |
747 | |
748 | DEFINE_CLASS_ID(Sub, Node, 6) |
749 | DEFINE_CLASS_ID(Cmp, Sub, 0) |
750 | DEFINE_CLASS_ID(FastLock, Cmp, 0) |
751 | DEFINE_CLASS_ID(FastUnlock, Cmp, 1) |
752 | DEFINE_CLASS_ID(SubTypeCheck,Cmp, 2) |
753 | |
754 | DEFINE_CLASS_ID(MergeMem, Node, 7) |
755 | DEFINE_CLASS_ID(Bool, Node, 8) |
756 | DEFINE_CLASS_ID(AddP, Node, 9) |
757 | DEFINE_CLASS_ID(BoxLock, Node, 10) |
758 | DEFINE_CLASS_ID(Add, Node, 11) |
759 | DEFINE_CLASS_ID(Mul, Node, 12) |
760 | DEFINE_CLASS_ID(ClearArray, Node, 14) |
761 | DEFINE_CLASS_ID(Halt, Node, 15) |
762 | DEFINE_CLASS_ID(Opaque1, Node, 16) |
763 | DEFINE_CLASS_ID(Move, Node, 17) |
764 | DEFINE_CLASS_ID(LShift, Node, 18) |
765 | |
766 | _max_classes = ClassMask_Move |
767 | }; |
768 | #undef DEFINE_CLASS_ID |
769 | |
770 | // Flags are sorted by usage frequency. |
771 | enum NodeFlags { |
772 | Flag_is_Copy = 1 << 0, // should be first bit to avoid shift |
773 | Flag_rematerialize = 1 << 1, |
774 | Flag_needs_anti_dependence_check = 1 << 2, |
775 | Flag_is_macro = 1 << 3, |
776 | Flag_is_Con = 1 << 4, |
777 | Flag_is_cisc_alternate = 1 << 5, |
778 | Flag_is_dead_loop_safe = 1 << 6, |
779 | Flag_may_be_short_branch = 1 << 7, |
780 | Flag_avoid_back_to_back_before = 1 << 8, |
781 | Flag_avoid_back_to_back_after = 1 << 9, |
782 | Flag_has_call = 1 << 10, |
783 | Flag_is_reduction = 1 << 11, |
784 | Flag_is_scheduled = 1 << 12, |
785 | Flag_has_vector_mask_set = 1 << 13, |
786 | Flag_is_expensive = 1 << 14, |
787 | Flag_is_predicated_vector = 1 << 15, |
788 | Flag_for_post_loop_opts_igvn = 1 << 16, |
789 | _last_flag = Flag_for_post_loop_opts_igvn |
790 | }; |
791 | |
792 | class PD; |
793 | |
794 | private: |
795 | juint _class_id; |
796 | juint _flags; |
797 | |
798 | static juint max_flags(); |
799 | |
800 | protected: |
801 | // These methods should be called from constructors only. |
802 | void init_class_id(juint c) { |
803 | _class_id = c; // cast out const |
804 | } |
805 | void init_flags(uint fl) { |
806 | assert(fl <= max_flags(), "invalid node flag")do { if (!(fl <= max_flags())) { (*g_assert_poison) = 'X'; ; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 806, "assert(" "fl <= max_flags()" ") failed", "invalid node flag" ); ::breakpoint(); } } while (0); |
807 | _flags |= fl; |
808 | } |
809 | void clear_flag(uint fl) { |
810 | assert(fl <= max_flags(), "invalid node flag")do { if (!(fl <= max_flags())) { (*g_assert_poison) = 'X'; ; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 810, "assert(" "fl <= max_flags()" ") failed", "invalid node flag" ); ::breakpoint(); } } while (0); |
811 | _flags &= ~fl; |
812 | } |
813 | |
814 | public: |
815 | const juint class_id() const { return _class_id; } |
816 | |
817 | const juint flags() const { return _flags; } |
818 | |
819 | void add_flag(juint fl) { init_flags(fl); } |
820 | |
821 | void remove_flag(juint fl) { clear_flag(fl); } |
822 | |
823 | // Return a dense integer opcode number |
824 | virtual int Opcode() const; |
825 | |
826 | // Virtual inherited Node size |
827 | virtual uint size_of() const; |
828 | |
829 | // Other interesting Node properties |
830 | #define DEFINE_CLASS_QUERY(type) \ |
831 | bool is_##type() const { \ |
832 | return ((_class_id & ClassMask_##type) == Class_##type); \ |
833 | } \ |
834 | type##Node *as_##type() const { \ |
835 | assert(is_##type(), "invalid node class: %s", Name())do { if (!(is_##type())) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 835, "assert(" "is_##type()" ") failed", "invalid node class: %s" , Name()); ::breakpoint(); } } while (0); \ |
836 | return (type##Node*)this; \ |
837 | } \ |
838 | type##Node* isa_##type() const { \ |
839 | return (is_##type()) ? as_##type() : NULL__null; \ |
840 | } |
841 | |
842 | DEFINE_CLASS_QUERY(AbstractLock) |
843 | DEFINE_CLASS_QUERY(Add) |
844 | DEFINE_CLASS_QUERY(AddP) |
845 | DEFINE_CLASS_QUERY(Allocate) |
846 | DEFINE_CLASS_QUERY(AllocateArray) |
847 | DEFINE_CLASS_QUERY(ArrayCopy) |
848 | DEFINE_CLASS_QUERY(BaseCountedLoop) |
849 | DEFINE_CLASS_QUERY(BaseCountedLoopEnd) |
850 | DEFINE_CLASS_QUERY(Bool) |
851 | DEFINE_CLASS_QUERY(BoxLock) |
852 | DEFINE_CLASS_QUERY(Call) |
853 | DEFINE_CLASS_QUERY(CallNative) |
854 | DEFINE_CLASS_QUERY(CallDynamicJava) |
855 | DEFINE_CLASS_QUERY(CallJava) |
856 | DEFINE_CLASS_QUERY(CallLeaf) |
857 | DEFINE_CLASS_QUERY(CallLeafNoFP) |
858 | DEFINE_CLASS_QUERY(CallRuntime) |
859 | DEFINE_CLASS_QUERY(CallStaticJava) |
860 | DEFINE_CLASS_QUERY(Catch) |
861 | DEFINE_CLASS_QUERY(CatchProj) |
862 | DEFINE_CLASS_QUERY(CheckCastPP) |
863 | DEFINE_CLASS_QUERY(CastII) |
864 | DEFINE_CLASS_QUERY(CastLL) |
865 | DEFINE_CLASS_QUERY(ConstraintCast) |
866 | DEFINE_CLASS_QUERY(ClearArray) |
867 | DEFINE_CLASS_QUERY(CMove) |
868 | DEFINE_CLASS_QUERY(Cmp) |
869 | DEFINE_CLASS_QUERY(CountedLoop) |
870 | DEFINE_CLASS_QUERY(CountedLoopEnd) |
871 | DEFINE_CLASS_QUERY(DecodeNarrowPtr) |
872 | DEFINE_CLASS_QUERY(DecodeN) |
873 | DEFINE_CLASS_QUERY(DecodeNKlass) |
874 | DEFINE_CLASS_QUERY(EncodeNarrowPtr) |
875 | DEFINE_CLASS_QUERY(EncodeP) |
876 | DEFINE_CLASS_QUERY(EncodePKlass) |
877 | DEFINE_CLASS_QUERY(FastLock) |
878 | DEFINE_CLASS_QUERY(FastUnlock) |
879 | DEFINE_CLASS_QUERY(Halt) |
880 | DEFINE_CLASS_QUERY(If) |
881 | DEFINE_CLASS_QUERY(RangeCheck) |
882 | DEFINE_CLASS_QUERY(IfProj) |
883 | DEFINE_CLASS_QUERY(IfFalse) |
884 | DEFINE_CLASS_QUERY(IfTrue) |
885 | DEFINE_CLASS_QUERY(Initialize) |
886 | DEFINE_CLASS_QUERY(Jump) |
887 | DEFINE_CLASS_QUERY(JumpProj) |
888 | DEFINE_CLASS_QUERY(LongCountedLoop) |
889 | DEFINE_CLASS_QUERY(LongCountedLoopEnd) |
890 | DEFINE_CLASS_QUERY(Load) |
891 | DEFINE_CLASS_QUERY(LoadStore) |
892 | DEFINE_CLASS_QUERY(LoadStoreConditional) |
893 | DEFINE_CLASS_QUERY(Lock) |
894 | DEFINE_CLASS_QUERY(Loop) |
895 | DEFINE_CLASS_QUERY(LShift) |
896 | DEFINE_CLASS_QUERY(Mach) |
897 | DEFINE_CLASS_QUERY(MachBranch) |
898 | DEFINE_CLASS_QUERY(MachCall) |
899 | DEFINE_CLASS_QUERY(MachCallNative) |
900 | DEFINE_CLASS_QUERY(MachCallDynamicJava) |
901 | DEFINE_CLASS_QUERY(MachCallJava) |
902 | DEFINE_CLASS_QUERY(MachCallLeaf) |
903 | DEFINE_CLASS_QUERY(MachCallRuntime) |
904 | DEFINE_CLASS_QUERY(MachCallStaticJava) |
905 | DEFINE_CLASS_QUERY(MachConstantBase) |
906 | DEFINE_CLASS_QUERY(MachConstant) |
907 | DEFINE_CLASS_QUERY(MachGoto) |
908 | DEFINE_CLASS_QUERY(MachIf) |
909 | DEFINE_CLASS_QUERY(MachJump) |
910 | DEFINE_CLASS_QUERY(MachNullCheck) |
911 | DEFINE_CLASS_QUERY(MachProj) |
912 | DEFINE_CLASS_QUERY(MachReturn) |
913 | DEFINE_CLASS_QUERY(MachSafePoint) |
914 | DEFINE_CLASS_QUERY(MachSpillCopy) |
915 | DEFINE_CLASS_QUERY(MachTemp) |
916 | DEFINE_CLASS_QUERY(MachMemBar) |
917 | DEFINE_CLASS_QUERY(MachMerge) |
918 | DEFINE_CLASS_QUERY(Mem) |
919 | DEFINE_CLASS_QUERY(MemBar) |
920 | DEFINE_CLASS_QUERY(MemBarStoreStore) |
921 | DEFINE_CLASS_QUERY(MergeMem) |
922 | DEFINE_CLASS_QUERY(Move) |
923 | DEFINE_CLASS_QUERY(Mul) |
924 | DEFINE_CLASS_QUERY(Multi) |
925 | DEFINE_CLASS_QUERY(MultiBranch) |
926 | DEFINE_CLASS_QUERY(Opaque1) |
927 | DEFINE_CLASS_QUERY(OuterStripMinedLoop) |
928 | DEFINE_CLASS_QUERY(OuterStripMinedLoopEnd) |
929 | DEFINE_CLASS_QUERY(Parm) |
930 | DEFINE_CLASS_QUERY(PCTable) |
931 | DEFINE_CLASS_QUERY(Phi) |
932 | DEFINE_CLASS_QUERY(Proj) |
933 | DEFINE_CLASS_QUERY(Region) |
934 | DEFINE_CLASS_QUERY(Root) |
935 | DEFINE_CLASS_QUERY(SafePoint) |
936 | DEFINE_CLASS_QUERY(SafePointScalarObject) |
937 | DEFINE_CLASS_QUERY(Start) |
938 | DEFINE_CLASS_QUERY(Store) |
939 | DEFINE_CLASS_QUERY(Sub) |
940 | DEFINE_CLASS_QUERY(SubTypeCheck) |
941 | DEFINE_CLASS_QUERY(Type) |
942 | DEFINE_CLASS_QUERY(Vector) |
943 | DEFINE_CLASS_QUERY(VectorMaskCmp) |
944 | DEFINE_CLASS_QUERY(VectorUnbox) |
945 | DEFINE_CLASS_QUERY(VectorReinterpret); |
946 | DEFINE_CLASS_QUERY(LoadVector) |
947 | DEFINE_CLASS_QUERY(LoadVectorGather) |
948 | DEFINE_CLASS_QUERY(StoreVector) |
949 | DEFINE_CLASS_QUERY(StoreVectorScatter) |
950 | DEFINE_CLASS_QUERY(ShiftV) |
951 | DEFINE_CLASS_QUERY(Unlock) |
952 | |
953 | #undef DEFINE_CLASS_QUERY |
954 | |
955 | // duplicate of is_MachSpillCopy() |
956 | bool is_SpillCopy () const { |
957 | return ((_class_id & ClassMask_MachSpillCopy) == Class_MachSpillCopy); |
958 | } |
959 | |
960 | bool is_Con () const { return (_flags & Flag_is_Con) != 0; } |
961 | // The data node which is safe to leave in dead loop during IGVN optimization. |
962 | bool is_dead_loop_safe() const; |
963 | |
964 | // is_Copy() returns copied edge index (0 or 1) |
965 | uint is_Copy() const { return (_flags & Flag_is_Copy); } |
966 | |
967 | virtual bool is_CFG() const { return false; } |
968 | |
969 | // If this node is control-dependent on a test, can it be |
970 | // rerouted to a dominating equivalent test? This is usually |
971 | // true of non-CFG nodes, but can be false for operations which |
972 | // depend for their correct sequencing on more than one test. |
973 | // (In that case, hoisting to a dominating test may silently |
974 | // skip some other important test.) |
975 | virtual bool depends_only_on_test() const { assert(!is_CFG(), "")do { if (!(!is_CFG())) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 975, "assert(" "!is_CFG()" ") failed", ""); ::breakpoint(); } } while (0); return true; }; |
976 | |
977 | // When building basic blocks, I need to have a notion of block beginning |
978 | // Nodes, next block selector Nodes (block enders), and next block |
979 | // projections. These calls need to work on their machine equivalents. The |
980 | // Ideal beginning Nodes are RootNode, RegionNode and StartNode. |
981 | bool is_block_start() const { |
982 | if ( is_Region() ) |
983 | return this == (const Node*)in(0); |
984 | else |
985 | return is_Start(); |
986 | } |
987 | |
988 | // The Ideal control projection Nodes are IfTrue/IfFalse, JumpProjNode, Root, |
989 | // Goto and Return. This call also returns the block ending Node. |
990 | virtual const Node *is_block_proj() const; |
991 | |
992 | // The node is a "macro" node which needs to be expanded before matching |
993 | bool is_macro() const { return (_flags & Flag_is_macro) != 0; } |
994 | // The node is expensive: the best control is set during loop opts |
995 | bool is_expensive() const { return (_flags & Flag_is_expensive) != 0 && in(0) != NULL__null; } |
996 | |
997 | // An arithmetic node which accumulates a data in a loop. |
998 | // It must have the loop's phi as input and provide a def to the phi. |
999 | bool is_reduction() const { return (_flags & Flag_is_reduction) != 0; } |
1000 | |
1001 | bool is_predicated_vector() const { return (_flags & Flag_is_predicated_vector) != 0; } |
1002 | |
1003 | // The node is a CountedLoopEnd with a mask annotation so as to emit a restore context |
1004 | bool has_vector_mask_set() const { return (_flags & Flag_has_vector_mask_set) != 0; } |
1005 | |
1006 | // Used in lcm to mark nodes that have scheduled |
1007 | bool is_scheduled() const { return (_flags & Flag_is_scheduled) != 0; } |
1008 | |
1009 | bool for_post_loop_opts_igvn() const { return (_flags & Flag_for_post_loop_opts_igvn) != 0; } |
1010 | |
1011 | //----------------- Optimization |
1012 | |
1013 | // Get the worst-case Type output for this Node. |
1014 | virtual const class Type *bottom_type() const; |
1015 | |
1016 | // If we find a better type for a node, try to record it permanently. |
1017 | // Return true if this node actually changed. |
1018 | // Be sure to do the hash_delete game in the "rehash" variant. |
1019 | void raise_bottom_type(const Type* new_type); |
1020 | |
1021 | // Get the address type with which this node uses and/or defs memory, |
1022 | // or NULL if none. The address type is conservatively wide. |
1023 | // Returns non-null for calls, membars, loads, stores, etc. |
1024 | // Returns TypePtr::BOTTOM if the node touches memory "broadly". |
1025 | virtual const class TypePtr *adr_type() const { return NULL__null; } |
1026 | |
1027 | // Return an existing node which computes the same function as this node. |
1028 | // The optimistic combined algorithm requires this to return a Node which |
1029 | // is a small number of steps away (e.g., one of my inputs). |
1030 | virtual Node* Identity(PhaseGVN* phase); |
1031 | |
1032 | // Return the set of values this Node can take on at runtime. |
1033 | virtual const Type* Value(PhaseGVN* phase) const; |
1034 | |
1035 | // Return a node which is more "ideal" than the current node. |
1036 | // The invariants on this call are subtle. If in doubt, read the |
1037 | // treatise in node.cpp above the default implemention AND TEST WITH |
1038 | // +VerifyIterativeGVN! |
1039 | virtual Node *Ideal(PhaseGVN *phase, bool can_reshape); |
1040 | |
1041 | // Some nodes have specific Ideal subgraph transformations only if they are |
1042 | // unique users of specific nodes. Such nodes should be put on IGVN worklist |
1043 | // for the transformations to happen. |
1044 | bool has_special_unique_user() const; |
1045 | |
1046 | // Skip Proj and CatchProj nodes chains. Check for Null and Top. |
1047 | Node* find_exact_control(Node* ctrl); |
1048 | |
1049 | // Check if 'this' node dominates or equal to 'sub'. |
1050 | bool dominates(Node* sub, Node_List &nlist); |
1051 | |
1052 | protected: |
1053 | bool remove_dead_region(PhaseGVN *phase, bool can_reshape); |
1054 | public: |
1055 | |
1056 | // See if there is valid pipeline info |
1057 | static const Pipeline *pipeline_class(); |
1058 | virtual const Pipeline *pipeline() const; |
1059 | |
1060 | // Compute the latency from the def to this instruction of the ith input node |
1061 | uint latency(uint i); |
1062 | |
1063 | // Hash & compare functions, for pessimistic value numbering |
1064 | |
1065 | // If the hash function returns the special sentinel value NO_HASH, |
1066 | // the node is guaranteed never to compare equal to any other node. |
1067 | // If we accidentally generate a hash with value NO_HASH the node |
1068 | // won't go into the table and we'll lose a little optimization. |
1069 | static const uint NO_HASH = 0; |
1070 | virtual uint hash() const; |
1071 | virtual bool cmp( const Node &n ) const; |
1072 | |
1073 | // Operation appears to be iteratively computed (such as an induction variable) |
1074 | // It is possible for this operation to return false for a loop-varying |
1075 | // value, if it appears (by local graph inspection) to be computed by a simple conditional. |
1076 | bool is_iteratively_computed(); |
1077 | |
1078 | // Determine if a node is a counted loop induction variable. |
1079 | // NOTE: The method is defined in "loopnode.cpp". |
1080 | bool is_cloop_ind_var() const; |
1081 | |
1082 | // Return a node with opcode "opc" and same inputs as "this" if one can |
1083 | // be found; Otherwise return NULL; |
1084 | Node* find_similar(int opc); |
1085 | |
1086 | // Return the unique control out if only one. Null if none or more than one. |
1087 | Node* unique_ctrl_out() const; |
1088 | |
1089 | // Set control or add control as precedence edge |
1090 | void ensure_control_or_add_prec(Node* c); |
1091 | |
1092 | //----------------- Code Generation |
1093 | |
1094 | // Ideal register class for Matching. Zero means unmatched instruction |
1095 | // (these are cloned instead of converted to machine nodes). |
1096 | virtual uint ideal_reg() const; |
1097 | |
1098 | static const uint NotAMachineReg; // must be > max. machine register |
1099 | |
1100 | // Do we Match on this edge index or not? Generally false for Control |
1101 | // and true for everything else. Weird for calls & returns. |
1102 | virtual uint match_edge(uint idx) const; |
1103 | |
1104 | // Register class output is returned in |
1105 | virtual const RegMask &out_RegMask() const; |
1106 | // Register class input is expected in |
1107 | virtual const RegMask &in_RegMask(uint) const; |
1108 | // Should we clone rather than spill this instruction? |
1109 | bool rematerialize() const; |
1110 | |
1111 | // Return JVM State Object if this Node carries debug info, or NULL otherwise |
1112 | virtual JVMState* jvms() const; |
1113 | |
1114 | // Print as assembly |
1115 | virtual void format( PhaseRegAlloc *, outputStream* st = tty ) const; |
1116 | // Emit bytes starting at parameter 'ptr' |
1117 | // Bump 'ptr' by the number of output bytes |
1118 | virtual void emit(CodeBuffer &cbuf, PhaseRegAlloc *ra_) const; |
1119 | // Size of instruction in bytes |
1120 | virtual uint size(PhaseRegAlloc *ra_) const; |
1121 | |
1122 | // Convenience function to extract an integer constant from a node. |
1123 | // If it is not an integer constant (either Con, CastII, or Mach), |
1124 | // return value_if_unknown. |
1125 | jint find_int_con(jint value_if_unknown) const { |
1126 | const TypeInt* t = find_int_type(); |
1127 | return (t != NULL__null && t->is_con()) ? t->get_con() : value_if_unknown; |
1128 | } |
1129 | // Return the constant, knowing it is an integer constant already |
1130 | jint get_int() const { |
1131 | const TypeInt* t = find_int_type(); |
1132 | guarantee(t != NULL, "must be con")do { if (!(t != __null)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1132, "guarantee(" "t != NULL" ") failed", "must be con"); :: breakpoint(); } } while (0); |
1133 | return t->get_con(); |
1134 | } |
1135 | // Here's where the work is done. Can produce non-constant int types too. |
1136 | const TypeInt* find_int_type() const; |
1137 | const TypeInteger* find_integer_type(BasicType bt) const; |
1138 | |
1139 | // Same thing for long (and intptr_t, via type.hpp): |
1140 | jlong get_long() const { |
1141 | const TypeLong* t = find_long_type(); |
1142 | guarantee(t != NULL, "must be con")do { if (!(t != __null)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1142, "guarantee(" "t != NULL" ") failed", "must be con"); :: breakpoint(); } } while (0); |
1143 | return t->get_con(); |
1144 | } |
1145 | jlong find_long_con(jint value_if_unknown) const { |
1146 | const TypeLong* t = find_long_type(); |
1147 | return (t != NULL__null && t->is_con()) ? t->get_con() : value_if_unknown; |
1148 | } |
1149 | const TypeLong* find_long_type() const; |
1150 | |
1151 | jlong get_integer_as_long(BasicType bt) const { |
1152 | const TypeInteger* t = find_integer_type(bt); |
1153 | guarantee(t != NULL, "must be con")do { if (!(t != __null)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1153, "guarantee(" "t != NULL" ") failed", "must be con"); :: breakpoint(); } } while (0); |
1154 | return t->get_con_as_long(bt); |
1155 | } |
1156 | const TypePtr* get_ptr_type() const; |
1157 | |
1158 | // These guys are called by code generated by ADLC: |
1159 | intptr_t get_ptr() const; |
1160 | intptr_t get_narrowcon() const; |
1161 | jdouble getd() const; |
1162 | jfloat getf() const; |
1163 | |
1164 | // Nodes which are pinned into basic blocks |
1165 | virtual bool pinned() const { return false; } |
1166 | |
1167 | // Nodes which use memory without consuming it, hence need antidependences |
1168 | // More specifically, needs_anti_dependence_check returns true iff the node |
1169 | // (a) does a load, and (b) does not perform a store (except perhaps to a |
1170 | // stack slot or some other unaliased location). |
1171 | bool needs_anti_dependence_check() const; |
1172 | |
1173 | // Return which operand this instruction may cisc-spill. In other words, |
1174 | // return operand position that can convert from reg to memory access |
1175 | virtual int cisc_operand() const { return AdlcVMDeps::Not_cisc_spillable; } |
1176 | bool is_cisc_alternate() const { return (_flags & Flag_is_cisc_alternate) != 0; } |
1177 | |
1178 | // Whether this is a memory-writing machine node. |
1179 | bool is_memory_writer() const { return is_Mach() && bottom_type()->has_memory(); } |
1180 | |
1181 | //----------------- Printing, etc |
1182 | #ifndef PRODUCT |
1183 | private: |
1184 | int _indent; |
1185 | |
1186 | public: |
1187 | void set_indent(int indent) { _indent = indent; } |
1188 | |
1189 | private: |
1190 | static bool add_to_worklist(Node* n, Node_List* worklist, Arena* old_arena, VectorSet* old_space, VectorSet* new_space); |
1191 | public: |
1192 | Node* find(int idx, bool only_ctrl = false); // Search the graph for the given idx. |
1193 | Node* find_ctrl(int idx); // Search control ancestors for the given idx. |
1194 | void dump() const { dump("\n"); } // Print this node. |
1195 | void dump(const char* suffix, bool mark = false, outputStream *st = tty) const; // Print this node. |
1196 | void dump(int depth) const; // Print this node, recursively to depth d |
1197 | void dump_ctrl(int depth) const; // Print control nodes, to depth d |
1198 | void dump_comp() const; // Print this node in compact representation. |
1199 | // Print this node in compact representation. |
1200 | void dump_comp(const char* suffix, outputStream *st = tty) const; |
1201 | virtual void dump_req(outputStream *st = tty) const; // Print required-edge info |
1202 | virtual void dump_prec(outputStream *st = tty) const; // Print precedence-edge info |
1203 | virtual void dump_out(outputStream *st = tty) const; // Print the output edge info |
1204 | virtual void dump_spec(outputStream *st) const {}; // Print per-node info |
1205 | // Print compact per-node info |
1206 | virtual void dump_compact_spec(outputStream *st) const { dump_spec(st); } |
1207 | void dump_related() const; // Print related nodes (depends on node at hand). |
1208 | // Print related nodes up to given depths for input and output nodes. |
1209 | void dump_related(uint d_in, uint d_out) const; |
1210 | void dump_related_compact() const; // Print related nodes in compact representation. |
1211 | // Collect related nodes. |
1212 | virtual void related(GrowableArray<Node*> *in_rel, GrowableArray<Node*> *out_rel, bool compact) const; |
1213 | // Collect nodes starting from this node, explicitly including/excluding control and data links. |
1214 | void collect_nodes(GrowableArray<Node*> *ns, int d, bool ctrl, bool data) const; |
1215 | |
1216 | // Node collectors, to be used in implementations of Node::rel(). |
1217 | // Collect the entire data input graph. Include control inputs if requested. |
1218 | void collect_nodes_in_all_data(GrowableArray<Node*> *ns, bool ctrl) const; |
1219 | // Collect the entire control input graph. Include data inputs if requested. |
1220 | void collect_nodes_in_all_ctrl(GrowableArray<Node*> *ns, bool data) const; |
1221 | // Collect the entire output graph until hitting and including control nodes. |
1222 | void collect_nodes_out_all_ctrl_boundary(GrowableArray<Node*> *ns) const; |
1223 | |
1224 | void verify_edges(Unique_Node_List &visited); // Verify bi-directional edges |
1225 | static void verify(int verify_depth, VectorSet& visited, Node_List& worklist); |
1226 | |
1227 | // This call defines a class-unique string used to identify class instances |
1228 | virtual const char *Name() const; |
1229 | |
1230 | void dump_format(PhaseRegAlloc *ra) const; // debug access to MachNode::format(...) |
1231 | // RegMask Print Functions |
1232 | void dump_in_regmask(int idx) { in_RegMask(idx).dump(); } |
1233 | void dump_out_regmask() { out_RegMask().dump(); } |
1234 | static bool in_dump() { return Compile::current()->_in_dump_cnt > 0; } |
1235 | void fast_dump() const { |
1236 | tty->print("%4d: %-17s", _idx, Name()); |
1237 | for (uint i = 0; i < len(); i++) |
1238 | if (in(i)) |
1239 | tty->print(" %4d", in(i)->_idx); |
1240 | else |
1241 | tty->print(" NULL"); |
1242 | tty->print("\n"); |
1243 | } |
1244 | #endif |
1245 | #ifdef ASSERT1 |
1246 | void verify_construction(); |
1247 | bool verify_jvms(const JVMState* jvms) const; |
1248 | int _debug_idx; // Unique value assigned to every node. |
1249 | int debug_idx() const { return _debug_idx; } |
1250 | void set_debug_idx( int debug_idx ) { _debug_idx = debug_idx; } |
1251 | |
1252 | Node* _debug_orig; // Original version of this, if any. |
1253 | Node* debug_orig() const { return _debug_orig; } |
1254 | void set_debug_orig(Node* orig); // _debug_orig = orig |
1255 | void dump_orig(outputStream *st, bool print_key = true) const; |
1256 | |
1257 | int _hash_lock; // Barrier to modifications of nodes in the hash table |
1258 | void enter_hash_lock() { ++_hash_lock; assert(_hash_lock < 99, "in too many hash tables?")do { if (!(_hash_lock < 99)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1258, "assert(" "_hash_lock < 99" ") failed", "in too many hash tables?" ); ::breakpoint(); } } while (0); } |
1259 | void exit_hash_lock() { --_hash_lock; assert(_hash_lock >= 0, "mispaired hash locks")do { if (!(_hash_lock >= 0)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1259, "assert(" "_hash_lock >= 0" ") failed", "mispaired hash locks" ); ::breakpoint(); } } while (0); } |
1260 | |
1261 | static void init_NodeProperty(); |
1262 | |
1263 | #if OPTO_DU_ITERATOR_ASSERT1 |
1264 | const Node* _last_del; // The last deleted node. |
1265 | uint _del_tick; // Bumped when a deletion happens.. |
1266 | #endif |
1267 | #endif |
1268 | }; |
1269 | |
1270 | inline bool not_a_node(const Node* n) { |
1271 | if (n == NULL__null) return true; |
1272 | if (((intptr_t)n & 1) != 0) return true; // uninitialized, etc. |
1273 | if (*(address*)n == badAddress((address)::badAddressVal)) return true; // kill by Node::destruct |
1274 | return false; |
1275 | } |
1276 | |
1277 | //----------------------------------------------------------------------------- |
1278 | // Iterators over DU info, and associated Node functions. |
1279 | |
1280 | #if OPTO_DU_ITERATOR_ASSERT1 |
1281 | |
1282 | // Common code for assertion checking on DU iterators. |
1283 | class DUIterator_Common { |
1284 | #ifdef ASSERT1 |
1285 | protected: |
1286 | bool _vdui; // cached value of VerifyDUIterators |
1287 | const Node* _node; // the node containing the _out array |
1288 | uint _outcnt; // cached node->_outcnt |
1289 | uint _del_tick; // cached node->_del_tick |
1290 | Node* _last; // last value produced by the iterator |
1291 | |
1292 | void sample(const Node* node); // used by c'tor to set up for verifies |
1293 | void verify(const Node* node, bool at_end_ok = false); |
1294 | void verify_resync(); |
1295 | void reset(const DUIterator_Common& that); |
1296 | |
1297 | // The VDUI_ONLY macro protects code conditionalized on VerifyDUIterators |
1298 | #define I_VDUI_ONLY(i,x) { if ((i)._vdui) { x; } } |
1299 | #else |
1300 | #define I_VDUI_ONLY(i,x) { } |
1301 | #endif //ASSERT |
1302 | }; |
1303 | |
1304 | #define VDUI_ONLY(x) I_VDUI_ONLY(*this, x) |
1305 | |
1306 | // Default DU iterator. Allows appends onto the out array. |
1307 | // Allows deletion from the out array only at the current point. |
1308 | // Usage: |
1309 | // for (DUIterator i = x->outs(); x->has_out(i); i++) { |
1310 | // Node* y = x->out(i); |
1311 | // ... |
1312 | // } |
1313 | // Compiles in product mode to a unsigned integer index, which indexes |
1314 | // onto a repeatedly reloaded base pointer of x->_out. The loop predicate |
1315 | // also reloads x->_outcnt. If you delete, you must perform "--i" just |
1316 | // before continuing the loop. You must delete only the last-produced |
1317 | // edge. You must delete only a single copy of the last-produced edge, |
1318 | // or else you must delete all copies at once (the first time the edge |
1319 | // is produced by the iterator). |
1320 | class DUIterator : public DUIterator_Common { |
1321 | friend class Node; |
1322 | |
1323 | // This is the index which provides the product-mode behavior. |
1324 | // Whatever the product-mode version of the system does to the |
1325 | // DUI index is done to this index. All other fields in |
1326 | // this class are used only for assertion checking. |
1327 | uint _idx; |
1328 | |
1329 | #ifdef ASSERT1 |
1330 | uint _refresh_tick; // Records the refresh activity. |
1331 | |
1332 | void sample(const Node* node); // Initialize _refresh_tick etc. |
1333 | void verify(const Node* node, bool at_end_ok = false); |
1334 | void verify_increment(); // Verify an increment operation. |
1335 | void verify_resync(); // Verify that we can back up over a deletion. |
1336 | void verify_finish(); // Verify that the loop terminated properly. |
1337 | void refresh(); // Resample verification info. |
1338 | void reset(const DUIterator& that); // Resample after assignment. |
1339 | #endif |
1340 | |
1341 | DUIterator(const Node* node, int dummy_to_avoid_conversion) |
1342 | { _idx = 0; debug_only(sample(node))sample(node); } |
1343 | |
1344 | public: |
1345 | // initialize to garbage; clear _vdui to disable asserts |
1346 | DUIterator() |
1347 | { /*initialize to garbage*/ debug_only(_vdui = false)_vdui = false; } |
1348 | |
1349 | DUIterator(const DUIterator& that) |
1350 | { _idx = that._idx; debug_only(_vdui = false; reset(that))_vdui = false; reset(that); } |
1351 | |
1352 | void operator++(int dummy_to_specify_postfix_op) |
1353 | { _idx++; VDUI_ONLY(verify_increment()); } |
1354 | |
1355 | void operator--() |
1356 | { VDUI_ONLY(verify_resync()); --_idx; } |
1357 | |
1358 | ~DUIterator() |
1359 | { VDUI_ONLY(verify_finish()); } |
1360 | |
1361 | void operator=(const DUIterator& that) |
1362 | { _idx = that._idx; debug_only(reset(that))reset(that); } |
1363 | }; |
1364 | |
1365 | DUIterator Node::outs() const |
1366 | { return DUIterator(this, 0); } |
1367 | DUIterator& Node::refresh_out_pos(DUIterator& i) const |
1368 | { I_VDUI_ONLY(i, i.refresh()); return i; } |
1369 | bool Node::has_out(DUIterator& i) const |
1370 | { I_VDUI_ONLY(i, i.verify(this,true));return i._idx < _outcnt; } |
1371 | Node* Node::out(DUIterator& i) const |
1372 | { I_VDUI_ONLY(i, i.verify(this)); return debug_only(i._last=)i._last= _out[i._idx]; } |
1373 | |
1374 | |
1375 | // Faster DU iterator. Disallows insertions into the out array. |
1376 | // Allows deletion from the out array only at the current point. |
1377 | // Usage: |
1378 | // for (DUIterator_Fast imax, i = x->fast_outs(imax); i < imax; i++) { |
1379 | // Node* y = x->fast_out(i); |
1380 | // ... |
1381 | // } |
1382 | // Compiles in product mode to raw Node** pointer arithmetic, with |
1383 | // no reloading of pointers from the original node x. If you delete, |
1384 | // you must perform "--i; --imax" just before continuing the loop. |
1385 | // If you delete multiple copies of the same edge, you must decrement |
1386 | // imax, but not i, multiple times: "--i, imax -= num_edges". |
1387 | class DUIterator_Fast : public DUIterator_Common { |
1388 | friend class Node; |
1389 | friend class DUIterator_Last; |
1390 | |
1391 | // This is the pointer which provides the product-mode behavior. |
1392 | // Whatever the product-mode version of the system does to the |
1393 | // DUI pointer is done to this pointer. All other fields in |
1394 | // this class are used only for assertion checking. |
1395 | Node** _outp; |
1396 | |
1397 | #ifdef ASSERT1 |
1398 | void verify(const Node* node, bool at_end_ok = false); |
1399 | void verify_limit(); |
1400 | void verify_resync(); |
1401 | void verify_relimit(uint n); |
1402 | void reset(const DUIterator_Fast& that); |
1403 | #endif |
1404 | |
1405 | // Note: offset must be signed, since -1 is sometimes passed |
1406 | DUIterator_Fast(const Node* node, ptrdiff_t offset) |
1407 | { _outp = node->_out + offset; debug_only(sample(node))sample(node); } |
1408 | |
1409 | public: |
1410 | // initialize to garbage; clear _vdui to disable asserts |
1411 | DUIterator_Fast() |
1412 | { /*initialize to garbage*/ debug_only(_vdui = false)_vdui = false; } |
1413 | |
1414 | DUIterator_Fast(const DUIterator_Fast& that) |
1415 | { _outp = that._outp; debug_only(_vdui = false; reset(that))_vdui = false; reset(that); } |
1416 | |
1417 | void operator++(int dummy_to_specify_postfix_op) |
1418 | { _outp++; VDUI_ONLY(verify(_node, true)); } |
1419 | |
1420 | void operator--() |
1421 | { VDUI_ONLY(verify_resync()); --_outp; } |
1422 | |
1423 | void operator-=(uint n) // applied to the limit only |
1424 | { _outp -= n; VDUI_ONLY(verify_relimit(n)); } |
1425 | |
1426 | bool operator<(DUIterator_Fast& limit) { |
1427 | I_VDUI_ONLY(*this, this->verify(_node, true)); |
1428 | I_VDUI_ONLY(limit, limit.verify_limit()); |
1429 | return _outp < limit._outp; |
1430 | } |
1431 | |
1432 | void operator=(const DUIterator_Fast& that) |
1433 | { _outp = that._outp; debug_only(reset(that))reset(that); } |
1434 | }; |
1435 | |
1436 | DUIterator_Fast Node::fast_outs(DUIterator_Fast& imax) const { |
1437 | // Assign a limit pointer to the reference argument: |
1438 | imax = DUIterator_Fast(this, (ptrdiff_t)_outcnt); |
1439 | // Return the base pointer: |
1440 | return DUIterator_Fast(this, 0); |
1441 | } |
1442 | Node* Node::fast_out(DUIterator_Fast& i) const { |
1443 | I_VDUI_ONLY(i, i.verify(this)); |
1444 | return debug_only(i._last=)i._last= *i._outp; |
1445 | } |
1446 | |
1447 | |
1448 | // Faster DU iterator. Requires each successive edge to be removed. |
1449 | // Does not allow insertion of any edges. |
1450 | // Usage: |
1451 | // for (DUIterator_Last imin, i = x->last_outs(imin); i >= imin; i -= num_edges) { |
1452 | // Node* y = x->last_out(i); |
1453 | // ... |
1454 | // } |
1455 | // Compiles in product mode to raw Node** pointer arithmetic, with |
1456 | // no reloading of pointers from the original node x. |
1457 | class DUIterator_Last : private DUIterator_Fast { |
1458 | friend class Node; |
1459 | |
1460 | #ifdef ASSERT1 |
1461 | void verify(const Node* node, bool at_end_ok = false); |
1462 | void verify_limit(); |
1463 | void verify_step(uint num_edges); |
1464 | #endif |
1465 | |
1466 | // Note: offset must be signed, since -1 is sometimes passed |
1467 | DUIterator_Last(const Node* node, ptrdiff_t offset) |
1468 | : DUIterator_Fast(node, offset) { } |
1469 | |
1470 | void operator++(int dummy_to_specify_postfix_op) {} // do not use |
1471 | void operator<(int) {} // do not use |
1472 | |
1473 | public: |
1474 | DUIterator_Last() { } |
1475 | // initialize to garbage |
1476 | |
1477 | DUIterator_Last(const DUIterator_Last& that) = default; |
1478 | |
1479 | void operator--() |
1480 | { _outp--; VDUI_ONLY(verify_step(1)); } |
1481 | |
1482 | void operator-=(uint n) |
1483 | { _outp -= n; VDUI_ONLY(verify_step(n)); } |
1484 | |
1485 | bool operator>=(DUIterator_Last& limit) { |
1486 | I_VDUI_ONLY(*this, this->verify(_node, true)); |
1487 | I_VDUI_ONLY(limit, limit.verify_limit()); |
1488 | return _outp >= limit._outp; |
1489 | } |
1490 | |
1491 | DUIterator_Last& operator=(const DUIterator_Last& that) = default; |
1492 | }; |
1493 | |
1494 | DUIterator_Last Node::last_outs(DUIterator_Last& imin) const { |
1495 | // Assign a limit pointer to the reference argument: |
1496 | imin = DUIterator_Last(this, 0); |
1497 | // Return the initial pointer: |
1498 | return DUIterator_Last(this, (ptrdiff_t)_outcnt - 1); |
1499 | } |
1500 | Node* Node::last_out(DUIterator_Last& i) const { |
1501 | I_VDUI_ONLY(i, i.verify(this)); |
1502 | return debug_only(i._last=)i._last= *i._outp; |
1503 | } |
1504 | |
1505 | #endif //OPTO_DU_ITERATOR_ASSERT |
1506 | |
1507 | #undef I_VDUI_ONLY |
1508 | #undef VDUI_ONLY |
1509 | |
1510 | // An Iterator that truly follows the iterator pattern. Doesn't |
1511 | // support deletion but could be made to. |
1512 | // |
1513 | // for (SimpleDUIterator i(n); i.has_next(); i.next()) { |
1514 | // Node* m = i.get(); |
1515 | // |
1516 | class SimpleDUIterator : public StackObj { |
1517 | private: |
1518 | Node* node; |
1519 | DUIterator_Fast i; |
1520 | DUIterator_Fast imax; |
1521 | public: |
1522 | SimpleDUIterator(Node* n): node(n), i(n->fast_outs(imax)) {} |
1523 | bool has_next() { return i < imax; } |
1524 | void next() { i++; } |
1525 | Node* get() { return node->fast_out(i); } |
1526 | }; |
1527 | |
1528 | |
1529 | //----------------------------------------------------------------------------- |
1530 | // Map dense integer indices to Nodes. Uses classic doubling-array trick. |
1531 | // Abstractly provides an infinite array of Node*'s, initialized to NULL. |
1532 | // Note that the constructor just zeros things, and since I use Arena |
1533 | // allocation I do not need a destructor to reclaim storage. |
1534 | class Node_Array : public ResourceObj { |
1535 | friend class VMStructs; |
1536 | protected: |
1537 | Arena* _a; // Arena to allocate in |
1538 | uint _max; |
1539 | Node** _nodes; |
1540 | void grow( uint i ); // Grow array node to fit |
1541 | public: |
1542 | Node_Array(Arena* a, uint max = OptoNodeListSize) : _a(a), _max(max) { |
1543 | _nodes = NEW_ARENA_ARRAY(a, Node*, max)(Node**) (a)->Amalloc((max) * sizeof(Node*)); |
1544 | clear(); |
1545 | } |
1546 | |
1547 | Node_Array(Node_Array* na) : _a(na->_a), _max(na->_max), _nodes(na->_nodes) {} |
1548 | Node *operator[] ( uint i ) const // Lookup, or NULL for not mapped |
1549 | { return (i<_max) ? _nodes[i] : (Node*)NULL__null; } |
1550 | Node* at(uint i) const { assert(i<_max,"oob")do { if (!(i<_max)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1550, "assert(" "i<_max" ") failed", "oob"); ::breakpoint (); } } while (0); return _nodes[i]; } |
1551 | Node** adr() { return _nodes; } |
1552 | // Extend the mapping: index i maps to Node *n. |
1553 | void map( uint i, Node *n ) { if( i>=_max ) grow(i); _nodes[i] = n; } |
1554 | void insert( uint i, Node *n ); |
1555 | void remove( uint i ); // Remove, preserving order |
1556 | // Clear all entries in _nodes to NULL but keep storage |
1557 | void clear() { |
1558 | Copy::zero_to_bytes(_nodes, _max * sizeof(Node*)); |
1559 | } |
1560 | |
1561 | uint Size() const { return _max; } |
1562 | void dump() const; |
1563 | }; |
1564 | |
1565 | class Node_List : public Node_Array { |
1566 | friend class VMStructs; |
1567 | uint _cnt; |
1568 | public: |
1569 | Node_List(uint max = OptoNodeListSize) : Node_Array(Thread::current()->resource_area(), max), _cnt(0) {} |
1570 | Node_List(Arena *a, uint max = OptoNodeListSize) : Node_Array(a, max), _cnt(0) {} |
1571 | bool contains(const Node* n) const { |
1572 | for (uint e = 0; e < size(); e++) { |
1573 | if (at(e) == n) return true; |
1574 | } |
1575 | return false; |
1576 | } |
1577 | void insert( uint i, Node *n ) { Node_Array::insert(i,n); _cnt++; } |
1578 | void remove( uint i ) { Node_Array::remove(i); _cnt--; } |
1579 | void push( Node *b ) { map(_cnt++,b); } |
1580 | void yank( Node *n ); // Find and remove |
1581 | Node *pop() { return _nodes[--_cnt]; } |
1582 | void clear() { _cnt = 0; Node_Array::clear(); } // retain storage |
1583 | void copy(const Node_List& from) { |
1584 | if (from._max > _max) { |
1585 | grow(from._max); |
1586 | } |
1587 | _cnt = from._cnt; |
1588 | Copy::conjoint_words_to_higher((HeapWord*)&from._nodes[0], (HeapWord*)&_nodes[0], from._max * sizeof(Node*)); |
1589 | } |
1590 | |
1591 | uint size() const { return _cnt; } |
1592 | void dump() const; |
1593 | void dump_simple() const; |
1594 | }; |
1595 | |
1596 | //------------------------------Unique_Node_List------------------------------- |
1597 | class Unique_Node_List : public Node_List { |
1598 | friend class VMStructs; |
1599 | VectorSet _in_worklist; |
1600 | uint _clock_index; // Index in list where to pop from next |
1601 | public: |
1602 | Unique_Node_List() : Node_List(), _clock_index(0) {} |
1603 | Unique_Node_List(Arena *a) : Node_List(a), _in_worklist(a), _clock_index(0) {} |
1604 | |
1605 | void remove( Node *n ); |
1606 | bool member( Node *n ) { return _in_worklist.test(n->_idx) != 0; } |
1607 | VectorSet& member_set(){ return _in_worklist; } |
1608 | |
1609 | void push(Node* b) { |
1610 | if( !_in_worklist.test_set(b->_idx) ) |
1611 | Node_List::push(b); |
1612 | } |
1613 | Node *pop() { |
1614 | if( _clock_index >= size() ) _clock_index = 0; |
1615 | Node *b = at(_clock_index); |
1616 | map( _clock_index, Node_List::pop()); |
1617 | if (size() != 0) _clock_index++; // Always start from 0 |
1618 | _in_worklist.remove(b->_idx); |
1619 | return b; |
1620 | } |
1621 | Node *remove(uint i) { |
1622 | Node *b = Node_List::at(i); |
1623 | _in_worklist.remove(b->_idx); |
1624 | map(i,Node_List::pop()); |
1625 | return b; |
1626 | } |
1627 | void yank(Node *n) { |
1628 | _in_worklist.remove(n->_idx); |
1629 | Node_List::yank(n); |
1630 | } |
1631 | void clear() { |
1632 | _in_worklist.clear(); // Discards storage but grows automatically |
1633 | Node_List::clear(); |
1634 | _clock_index = 0; |
1635 | } |
1636 | |
1637 | // Used after parsing to remove useless nodes before Iterative GVN |
1638 | void remove_useless_nodes(VectorSet& useful); |
1639 | |
1640 | bool contains(const Node* n) const { |
1641 | fatal("use faster member() instead")do { (*g_assert_poison) = 'X';; report_fatal(INTERNAL_ERROR, "/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1641, "use faster member() instead"); ::breakpoint(); } while (0); |
1642 | return false; |
1643 | } |
1644 | |
1645 | #ifndef PRODUCT |
1646 | void print_set() const { _in_worklist.print(); } |
1647 | #endif |
1648 | }; |
1649 | |
1650 | // Inline definition of Compile::record_for_igvn must be deferred to this point. |
1651 | inline void Compile::record_for_igvn(Node* n) { |
1652 | _for_igvn->push(n); |
1653 | } |
1654 | |
1655 | //------------------------------Node_Stack------------------------------------- |
1656 | class Node_Stack { |
1657 | friend class VMStructs; |
1658 | protected: |
1659 | struct INode { |
1660 | Node *node; // Processed node |
1661 | uint indx; // Index of next node's child |
1662 | }; |
1663 | INode *_inode_top; // tos, stack grows up |
1664 | INode *_inode_max; // End of _inodes == _inodes + _max |
1665 | INode *_inodes; // Array storage for the stack |
1666 | Arena *_a; // Arena to allocate in |
1667 | void grow(); |
1668 | public: |
1669 | Node_Stack(int size) { |
1670 | size_t max = (size > OptoNodeListSize) ? size : OptoNodeListSize; |
1671 | _a = Thread::current()->resource_area(); |
1672 | _inodes = NEW_ARENA_ARRAY( _a, INode, max )(INode*) (_a)->Amalloc((max) * sizeof(INode)); |
1673 | _inode_max = _inodes + max; |
1674 | _inode_top = _inodes - 1; // stack is empty |
1675 | } |
1676 | |
1677 | Node_Stack(Arena *a, int size) : _a(a) { |
1678 | size_t max = (size > OptoNodeListSize) ? size : OptoNodeListSize; |
1679 | _inodes = NEW_ARENA_ARRAY( _a, INode, max )(INode*) (_a)->Amalloc((max) * sizeof(INode)); |
1680 | _inode_max = _inodes + max; |
1681 | _inode_top = _inodes - 1; // stack is empty |
1682 | } |
1683 | |
1684 | void pop() { |
1685 | assert(_inode_top >= _inodes, "node stack underflow")do { if (!(_inode_top >= _inodes)) { (*g_assert_poison) = 'X' ;; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1685, "assert(" "_inode_top >= _inodes" ") failed", "node stack underflow" ); ::breakpoint(); } } while (0); |
1686 | --_inode_top; |
1687 | } |
1688 | void push(Node *n, uint i) { |
1689 | ++_inode_top; |
1690 | if (_inode_top >= _inode_max) grow(); |
1691 | INode *top = _inode_top; // optimization |
1692 | top->node = n; |
1693 | top->indx = i; |
1694 | } |
1695 | Node *node() const { |
1696 | return _inode_top->node; |
1697 | } |
1698 | Node* node_at(uint i) const { |
1699 | assert(_inodes + i <= _inode_top, "in range")do { if (!(_inodes + i <= _inode_top)) { (*g_assert_poison ) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1699, "assert(" "_inodes + i <= _inode_top" ") failed", "in range" ); ::breakpoint(); } } while (0); |
1700 | return _inodes[i].node; |
1701 | } |
1702 | uint index() const { |
1703 | return _inode_top->indx; |
1704 | } |
1705 | uint index_at(uint i) const { |
1706 | assert(_inodes + i <= _inode_top, "in range")do { if (!(_inodes + i <= _inode_top)) { (*g_assert_poison ) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1706, "assert(" "_inodes + i <= _inode_top" ") failed", "in range" ); ::breakpoint(); } } while (0); |
1707 | return _inodes[i].indx; |
1708 | } |
1709 | void set_node(Node *n) { |
1710 | _inode_top->node = n; |
1711 | } |
1712 | void set_index(uint i) { |
1713 | _inode_top->indx = i; |
1714 | } |
1715 | uint size_max() const { return (uint)pointer_delta(_inode_max, _inodes, sizeof(INode)); } // Max size |
1716 | uint size() const { return (uint)pointer_delta((_inode_top+1), _inodes, sizeof(INode)); } // Current size |
1717 | bool is_nonempty() const { return (_inode_top >= _inodes); } |
1718 | bool is_empty() const { return (_inode_top < _inodes); } |
1719 | void clear() { _inode_top = _inodes - 1; } // retain storage |
1720 | |
1721 | // Node_Stack is used to map nodes. |
1722 | Node* find(uint idx) const; |
1723 | }; |
1724 | |
1725 | |
1726 | //-----------------------------Node_Notes-------------------------------------- |
1727 | // Debugging or profiling annotations loosely and sparsely associated |
1728 | // with some nodes. See Compile::node_notes_at for the accessor. |
1729 | class Node_Notes { |
1730 | friend class VMStructs; |
1731 | JVMState* _jvms; |
1732 | |
1733 | public: |
1734 | Node_Notes(JVMState* jvms = NULL__null) { |
1735 | _jvms = jvms; |
1736 | } |
1737 | |
1738 | JVMState* jvms() { return _jvms; } |
1739 | void set_jvms(JVMState* x) { _jvms = x; } |
1740 | |
1741 | // True if there is nothing here. |
1742 | bool is_clear() { |
1743 | return (_jvms == NULL__null); |
1744 | } |
1745 | |
1746 | // Make there be nothing here. |
1747 | void clear() { |
1748 | _jvms = NULL__null; |
1749 | } |
1750 | |
1751 | // Make a new, clean node notes. |
1752 | static Node_Notes* make(Compile* C) { |
1753 | Node_Notes* nn = NEW_ARENA_ARRAY(C->comp_arena(), Node_Notes, 1)(Node_Notes*) (C->comp_arena())->Amalloc((1) * sizeof(Node_Notes )); |
1754 | nn->clear(); |
1755 | return nn; |
1756 | } |
1757 | |
1758 | Node_Notes* clone(Compile* C) { |
1759 | Node_Notes* nn = NEW_ARENA_ARRAY(C->comp_arena(), Node_Notes, 1)(Node_Notes*) (C->comp_arena())->Amalloc((1) * sizeof(Node_Notes )); |
1760 | (*nn) = (*this); |
1761 | return nn; |
1762 | } |
1763 | |
1764 | // Absorb any information from source. |
1765 | bool update_from(Node_Notes* source) { |
1766 | bool changed = false; |
1767 | if (source != NULL__null) { |
1768 | if (source->jvms() != NULL__null) { |
1769 | set_jvms(source->jvms()); |
1770 | changed = true; |
1771 | } |
1772 | } |
1773 | return changed; |
1774 | } |
1775 | }; |
1776 | |
1777 | // Inlined accessors for Compile::node_nodes that require the preceding class: |
1778 | inline Node_Notes* |
1779 | Compile::locate_node_notes(GrowableArray<Node_Notes*>* arr, |
1780 | int idx, bool can_grow) { |
1781 | assert(idx >= 0, "oob")do { if (!(idx >= 0)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1781, "assert(" "idx >= 0" ") failed", "oob"); ::breakpoint (); } } while (0); |
1782 | int block_idx = (idx >> _log2_node_notes_block_size); |
1783 | int grow_by = (block_idx - (arr == NULL__null? 0: arr->length())); |
1784 | if (grow_by >= 0) { |
1785 | if (!can_grow) return NULL__null; |
1786 | grow_node_notes(arr, grow_by + 1); |
1787 | } |
1788 | if (arr == NULL__null) return NULL__null; |
1789 | // (Every element of arr is a sub-array of length _node_notes_block_size.) |
1790 | return arr->at(block_idx) + (idx & (_node_notes_block_size-1)); |
1791 | } |
1792 | |
1793 | inline bool |
1794 | Compile::set_node_notes_at(int idx, Node_Notes* value) { |
1795 | if (value == NULL__null || value->is_clear()) |
1796 | return false; // nothing to write => write nothing |
1797 | Node_Notes* loc = locate_node_notes(_node_note_array, idx, true); |
1798 | assert(loc != NULL, "")do { if (!(loc != __null)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1798, "assert(" "loc != __null" ") failed", ""); ::breakpoint (); } } while (0); |
1799 | return loc->update_from(value); |
1800 | } |
1801 | |
1802 | |
1803 | //------------------------------TypeNode--------------------------------------- |
1804 | // Node with a Type constant. |
1805 | class TypeNode : public Node { |
1806 | protected: |
1807 | virtual uint hash() const; // Check the type |
1808 | virtual bool cmp( const Node &n ) const; |
1809 | virtual uint size_of() const; // Size is bigger |
1810 | const Type* const _type; |
1811 | public: |
1812 | void set_type(const Type* t) { |
1813 | assert(t != NULL, "sanity")do { if (!(t != __null)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1813, "assert(" "t != __null" ") failed", "sanity"); ::breakpoint (); } } while (0); |
1814 | debug_only(uint check_hash = (VerifyHashTableKeys && _hash_lock) ? hash() : NO_HASH)uint check_hash = (VerifyHashTableKeys && _hash_lock) ? hash() : NO_HASH; |
1815 | *(const Type**)&_type = t; // cast away const-ness |
1816 | // If this node is in the hash table, make sure it doesn't need a rehash. |
1817 | assert(check_hash == NO_HASH || check_hash == hash(), "type change must preserve hash code")do { if (!(check_hash == NO_HASH || check_hash == hash())) { ( *g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1817, "assert(" "check_hash == NO_HASH || check_hash == hash()" ") failed", "type change must preserve hash code"); ::breakpoint (); } } while (0); |
1818 | } |
1819 | const Type* type() const { assert(_type != NULL, "sanity")do { if (!(_type != __null)) { (*g_assert_poison) = 'X';; report_vm_error ("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1819, "assert(" "_type != __null" ") failed", "sanity"); :: breakpoint(); } } while (0); return _type; }; |
1820 | TypeNode( const Type *t, uint required ) : Node(required), _type(t) { |
1821 | init_class_id(Class_Type); |
1822 | } |
1823 | virtual const Type* Value(PhaseGVN* phase) const; |
1824 | virtual const Type *bottom_type() const; |
1825 | virtual uint ideal_reg() const; |
1826 | #ifndef PRODUCT |
1827 | virtual void dump_spec(outputStream *st) const; |
1828 | virtual void dump_compact_spec(outputStream *st) const; |
1829 | #endif |
1830 | }; |
1831 | |
1832 | #include "opto/opcodes.hpp" |
1833 | |
1834 | #define Op_IL(op)inline int Op_op(BasicType bt) { do { if (!(bt == T_INT || bt == T_LONG)) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1834, "assert(" "bt == T_INT || bt == T_LONG" ") failed", "only for int or longs" ); ::breakpoint(); } } while (0); if (bt == T_INT) { return Op_opI ; } return Op_opL; } \ |
1835 | inline int Op_ ## op(BasicType bt) { \ |
1836 | assert(bt == T_INT || bt == T_LONG, "only for int or longs")do { if (!(bt == T_INT || bt == T_LONG)) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1836, "assert(" "bt == T_INT || bt == T_LONG" ") failed", "only for int or longs" ); ::breakpoint(); } } while (0); \ |
1837 | if (bt == T_INT) { \ |
1838 | return Op_## op ## I; \ |
1839 | } \ |
1840 | return Op_## op ## L; \ |
1841 | } |
1842 | |
1843 | Op_IL(Add)inline int Op_Add(BasicType bt) { do { if (!(bt == T_INT || bt == T_LONG)) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1843, "assert(" "bt == T_INT || bt == T_LONG" ") failed", "only for int or longs" ); ::breakpoint(); } } while (0); if (bt == T_INT) { return Op_AddI ; } return Op_AddL; } |
1844 | Op_IL(Sub)inline int Op_Sub(BasicType bt) { do { if (!(bt == T_INT || bt == T_LONG)) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1844, "assert(" "bt == T_INT || bt == T_LONG" ") failed", "only for int or longs" ); ::breakpoint(); } } while (0); if (bt == T_INT) { return Op_SubI ; } return Op_SubL; } |
1845 | Op_IL(Mul)inline int Op_Mul(BasicType bt) { do { if (!(bt == T_INT || bt == T_LONG)) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1845, "assert(" "bt == T_INT || bt == T_LONG" ") failed", "only for int or longs" ); ::breakpoint(); } } while (0); if (bt == T_INT) { return Op_MulI ; } return Op_MulL; } |
1846 | Op_IL(URShift)inline int Op_URShift(BasicType bt) { do { if (!(bt == T_INT || bt == T_LONG)) { (*g_assert_poison) = 'X';; report_vm_error( "/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1846, "assert(" "bt == T_INT || bt == T_LONG" ") failed", "only for int or longs" ); ::breakpoint(); } } while (0); if (bt == T_INT) { return Op_URShiftI ; } return Op_URShiftL; } |
1847 | Op_IL(LShift)inline int Op_LShift(BasicType bt) { do { if (!(bt == T_INT || bt == T_LONG)) { (*g_assert_poison) = 'X';; report_vm_error( "/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1847, "assert(" "bt == T_INT || bt == T_LONG" ") failed", "only for int or longs" ); ::breakpoint(); } } while (0); if (bt == T_INT) { return Op_LShiftI ; } return Op_LShiftL; } |
1848 | Op_IL(Xor)inline int Op_Xor(BasicType bt) { do { if (!(bt == T_INT || bt == T_LONG)) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1848, "assert(" "bt == T_INT || bt == T_LONG" ") failed", "only for int or longs" ); ::breakpoint(); } } while (0); if (bt == T_INT) { return Op_XorI ; } return Op_XorL; } |
1849 | Op_IL(Cmp)inline int Op_Cmp(BasicType bt) { do { if (!(bt == T_INT || bt == T_LONG)) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1849, "assert(" "bt == T_INT || bt == T_LONG" ") failed", "only for int or longs" ); ::breakpoint(); } } while (0); if (bt == T_INT) { return Op_CmpI ; } return Op_CmpL; } |
1850 | |
1851 | inline int Op_Cmp_unsigned(BasicType bt) { |
1852 | assert(bt == T_INT || bt == T_LONG, "only for int or longs")do { if (!(bt == T_INT || bt == T_LONG)) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1852, "assert(" "bt == T_INT || bt == T_LONG" ") failed", "only for int or longs" ); ::breakpoint(); } } while (0); |
1853 | if (bt == T_INT) { |
1854 | return Op_CmpU; |
1855 | } |
1856 | return Op_CmpUL; |
1857 | } |
1858 | |
1859 | inline int Op_Cast(BasicType bt) { |
1860 | assert(bt == T_INT || bt == T_LONG, "only for int or longs")do { if (!(bt == T_INT || bt == T_LONG)) { (*g_assert_poison) = 'X';; report_vm_error("/home/daniel/Projects/java/jdk/src/hotspot/share/opto/node.hpp" , 1860, "assert(" "bt == T_INT || bt == T_LONG" ") failed", "only for int or longs" ); ::breakpoint(); } } while (0); |
1861 | if (bt == T_INT) { |
1862 | return Op_CastII; |
1863 | } |
1864 | return Op_CastLL; |
1865 | } |
1866 | |
1867 | #endif // SHARE_OPTO_NODE_HPP |