1 /*
2 * Copyright (c) 2018, 2026, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
24
25 #include "code/vmreg.inline.hpp"
26 #include "gc/shared/barrierSet.hpp"
27 #include "gc/shared/c2/barrierSetC2.hpp"
28 #include "gc/shared/tlab_globals.hpp"
29 #include "opto/arraycopynode.hpp"
30 #include "opto/block.hpp"
31 #include "opto/convertnode.hpp"
32 #include "opto/graphKit.hpp"
33 #include "opto/idealKit.hpp"
34 #include "opto/macro.hpp"
35 #include "opto/narrowptrnode.hpp"
36 #include "opto/output.hpp"
37 #include "opto/regalloc.hpp"
38 #include "opto/runtime.hpp"
39 #include "utilities/macros.hpp"
40 #include CPU_HEADER(gc/shared/barrierSetAssembler)
41
42 // By default this is a no-op.
43 void BarrierSetC2::resolve_address(C2Access& access) const { }
44
45 void* C2ParseAccess::barrier_set_state() const {
46 return _kit->barrier_set_state();
47 }
48
49 PhaseGVN& C2ParseAccess::gvn() const { return _kit->gvn(); }
50
51 Node* C2ParseAccess::control() const {
52 return _ctl == nullptr ? _kit->control() : _ctl;
53 }
54
55 bool C2Access::needs_cpu_membar() const {
56 bool mismatched = (_decorators & C2_MISMATCHED) != 0;
57 bool is_unordered = (_decorators & MO_UNORDERED) != 0;
58
59 bool anonymous = (_decorators & C2_UNSAFE_ACCESS) != 0;
60 bool in_heap = (_decorators & IN_HEAP) != 0;
61 bool in_native = (_decorators & IN_NATIVE) != 0;
62 bool is_mixed = !in_heap && !in_native;
63
64 bool is_write = (_decorators & C2_WRITE_ACCESS) != 0;
65 bool is_read = (_decorators & C2_READ_ACCESS) != 0;
66 bool is_atomic = is_read && is_write;
67
68 if (is_atomic) {
69 // Atomics always need to be wrapped in CPU membars
70 return true;
71 }
72
73 if (anonymous) {
74 // We will need memory barriers unless we can determine a unique
75 // alias category for this reference. (Note: If for some reason
76 // the barriers get omitted and the unsafe reference begins to "pollute"
77 // the alias analysis of the rest of the graph, either Compile::can_alias
78 // or Compile::must_alias will throw a diagnostic assert.)
79 if (is_mixed || !is_unordered || (mismatched && !_addr.type()->isa_aryptr())) {
80 return true;
81 }
82 } else {
83 assert(!is_mixed, "not unsafe");
84 }
85
86 return false;
87 }
88
89 static BarrierSetC2State* barrier_set_state() {
90 return reinterpret_cast<BarrierSetC2State*>(Compile::current()->barrier_set_state());
91 }
92
93 RegMask& BarrierStubC2::live() const {
94 return *barrier_set_state()->live(_node);
95 }
96
97 BarrierStubC2::BarrierStubC2(const MachNode* node)
98 : _node(node),
99 _entry(),
100 _continuation(),
101 _preserve(live()) {}
102
103 Label* BarrierStubC2::entry() {
104 // The _entry will never be bound when in_scratch_emit_size() is true.
105 // However, we still need to return a label that is not bound now, but
106 // will eventually be bound. Any eventually bound label will do, as it
107 // will only act as a placeholder, so we return the _continuation label.
108 return Compile::current()->output()->in_scratch_emit_size() ? &_continuation : &_entry;
109 }
110
111 Label* BarrierStubC2::continuation() {
112 return &_continuation;
113 }
114
115 uint8_t BarrierStubC2::barrier_data() const {
116 return _node->barrier_data();
117 }
118
119 void BarrierStubC2::preserve(Register r) {
120 const VMReg vm_reg = r->as_VMReg();
121 assert(vm_reg->is_Register(), "r must be a general-purpose register");
122 _preserve.insert(OptoReg::as_OptoReg(vm_reg));
123 }
124
125 void BarrierStubC2::dont_preserve(Register r) {
126 VMReg vm_reg = r->as_VMReg();
127 assert(vm_reg->is_Register(), "r must be a general-purpose register");
128 // Subtract the given register and all its sub-registers (e.g. {R11, R11_H}
129 // for r11 in aarch64).
130 do {
131 _preserve.remove(OptoReg::as_OptoReg(vm_reg));
132 vm_reg = vm_reg->next();
133 } while (vm_reg->is_Register() && !vm_reg->is_concrete());
134 }
135
136 const RegMask& BarrierStubC2::preserve_set() const {
137 return _preserve;
138 }
139
140 Node* BarrierSetC2::store_at_resolved(C2Access& access, C2AccessValue& val) const {
141 DecoratorSet decorators = access.decorators();
142
143 bool mismatched = (decorators & C2_MISMATCHED) != 0;
144 bool unaligned = (decorators & C2_UNALIGNED) != 0;
145 bool unsafe = (decorators & C2_UNSAFE_ACCESS) != 0;
146 bool requires_atomic_access = (decorators & MO_UNORDERED) == 0;
147
148 MemNode::MemOrd mo = access.mem_node_mo();
149
150 Node* store;
151 BasicType bt = access.type();
152 if (access.is_parse_access()) {
153 C2ParseAccess& parse_access = static_cast<C2ParseAccess&>(access);
154
155 GraphKit* kit = parse_access.kit();
156 store = kit->store_to_memory(kit->control(), access.addr().node(), val.node(), bt,
157 mo, requires_atomic_access, unaligned, mismatched,
158 unsafe, access.barrier_data());
159 } else {
160 assert(access.is_opt_access(), "either parse or opt access");
161 C2OptAccess& opt_access = static_cast<C2OptAccess&>(access);
162 Node* ctl = opt_access.ctl();
163 MergeMemNode* mm = opt_access.mem();
164 PhaseGVN& gvn = opt_access.gvn();
165 const TypePtr* adr_type = access.addr().type();
166 int alias = gvn.C->get_alias_index(adr_type);
167 Node* mem = mm->memory_at(alias);
168
169 StoreNode* st = StoreNode::make(gvn, ctl, mem, access.addr().node(), adr_type, val.node(), bt, mo, requires_atomic_access);
170 if (unaligned) {
171 st->set_unaligned_access();
172 }
173 if (mismatched) {
174 st->set_mismatched_access();
175 }
176 st->set_barrier_data(access.barrier_data());
177 store = gvn.transform(st);
178 if (store == st) {
179 mm->set_memory_at(alias, st);
180 }
181 }
182 access.set_raw_access(store);
183
184 return store;
185 }
186
187 Node* BarrierSetC2::load_at_resolved(C2Access& access, const Type* val_type) const {
188 DecoratorSet decorators = access.decorators();
189
190 Node* adr = access.addr().node();
191 const TypePtr* adr_type = access.addr().type();
192
193 bool mismatched = (decorators & C2_MISMATCHED) != 0;
194 bool requires_atomic_access = (decorators & MO_UNORDERED) == 0;
195 bool unaligned = (decorators & C2_UNALIGNED) != 0;
196 bool control_dependent = (decorators & C2_CONTROL_DEPENDENT_LOAD) != 0;
197 bool unknown_control = (decorators & C2_UNKNOWN_CONTROL_LOAD) != 0;
198 bool unsafe = (decorators & C2_UNSAFE_ACCESS) != 0;
199 bool immutable = (decorators & C2_IMMUTABLE_MEMORY) != 0;
200
201 MemNode::MemOrd mo = access.mem_node_mo();
202 LoadNode::ControlDependency dep = unknown_control ? LoadNode::UnknownControl : LoadNode::DependsOnlyOnTest;
203
204 Node* load;
205 if (access.is_parse_access()) {
206 C2ParseAccess& parse_access = static_cast<C2ParseAccess&>(access);
207 GraphKit* kit = parse_access.kit();
208 Node* control = control_dependent ? parse_access.control() : nullptr;
209
210 if (immutable) {
211 Compile* C = Compile::current();
212 Node* mem = kit->immutable_memory();
213 load = LoadNode::make(kit->gvn(), control, mem, adr,
214 adr_type, val_type, access.type(), mo, dep, requires_atomic_access,
215 unaligned, mismatched, unsafe, access.barrier_data());
216 load = kit->gvn().transform(load);
217 } else {
218 load = kit->make_load(control, adr, val_type, access.type(), mo,
219 dep, requires_atomic_access, unaligned, mismatched, unsafe,
220 access.barrier_data());
221 }
222 } else {
223 assert(access.is_opt_access(), "either parse or opt access");
224 C2OptAccess& opt_access = static_cast<C2OptAccess&>(access);
225 Node* control = control_dependent ? opt_access.ctl() : nullptr;
226 MergeMemNode* mm = opt_access.mem();
227 PhaseGVN& gvn = opt_access.gvn();
228 Node* mem = mm->memory_at(gvn.C->get_alias_index(adr_type));
229 load = LoadNode::make(gvn, control, mem, adr, adr_type, val_type, access.type(), mo, dep,
230 requires_atomic_access, unaligned, mismatched, unsafe, access.barrier_data());
231 load = gvn.transform(load);
232 }
233 access.set_raw_access(load);
234
235 return load;
236 }
237
238 class C2AccessFence: public StackObj {
239 C2Access& _access;
240 Node* _leading_membar;
241
242 public:
243 C2AccessFence(C2Access& access) :
244 _access(access), _leading_membar(nullptr) {
245 GraphKit* kit = nullptr;
246 if (access.is_parse_access()) {
247 C2ParseAccess& parse_access = static_cast<C2ParseAccess&>(access);
248 kit = parse_access.kit();
249 }
250 DecoratorSet decorators = access.decorators();
251
252 bool is_write = (decorators & C2_WRITE_ACCESS) != 0;
253 bool is_read = (decorators & C2_READ_ACCESS) != 0;
254 bool is_atomic = is_read && is_write;
255
256 bool is_volatile = (decorators & MO_SEQ_CST) != 0;
257 bool is_release = (decorators & MO_RELEASE) != 0;
258
259 if (is_atomic) {
260 assert(kit != nullptr, "unsupported at optimization time");
261 // Memory-model-wise, a LoadStore acts like a little synchronized
262 // block, so needs barriers on each side. These don't translate
263 // into actual barriers on most machines, but we still need rest of
264 // compiler to respect ordering.
265 if (is_release) {
266 _leading_membar = kit->insert_mem_bar(Op_MemBarRelease);
267 } else if (is_volatile) {
268 if (support_IRIW_for_not_multiple_copy_atomic_cpu) {
269 _leading_membar = kit->insert_mem_bar(Op_MemBarVolatile);
270 } else {
271 _leading_membar = kit->insert_mem_bar(Op_MemBarRelease);
272 }
273 }
274 } else if (is_write) {
275 // If reference is volatile, prevent following memory ops from
276 // floating down past the volatile write. Also prevents commoning
277 // another volatile read.
278 if (is_volatile || is_release) {
279 assert(kit != nullptr, "unsupported at optimization time");
280 _leading_membar = kit->insert_mem_bar(Op_MemBarRelease);
281 }
282 } else {
283 // Memory barrier to prevent normal and 'unsafe' accesses from
284 // bypassing each other. Happens after null checks, so the
285 // exception paths do not take memory state from the memory barrier,
286 // so there's no problems making a strong assert about mixing users
287 // of safe & unsafe memory.
288 if (is_volatile && support_IRIW_for_not_multiple_copy_atomic_cpu) {
289 assert(kit != nullptr, "unsupported at optimization time");
290 _leading_membar = kit->insert_mem_bar(Op_MemBarVolatile);
291 }
292 }
293
294 if (access.needs_cpu_membar()) {
295 assert(kit != nullptr, "unsupported at optimization time");
296 kit->insert_mem_bar(Op_MemBarCPUOrder);
297 }
298
299 if (is_atomic) {
300 // 4984716: MemBars must be inserted before this
301 // memory node in order to avoid a false
302 // dependency which will confuse the scheduler.
303 access.set_memory();
304 }
305 }
306
307 ~C2AccessFence() {
308 GraphKit* kit = nullptr;
309 if (_access.is_parse_access()) {
310 C2ParseAccess& parse_access = static_cast<C2ParseAccess&>(_access);
311 kit = parse_access.kit();
312 }
313 DecoratorSet decorators = _access.decorators();
314
315 bool is_write = (decorators & C2_WRITE_ACCESS) != 0;
316 bool is_read = (decorators & C2_READ_ACCESS) != 0;
317 bool is_atomic = is_read && is_write;
318
319 bool is_volatile = (decorators & MO_SEQ_CST) != 0;
320 bool is_acquire = (decorators & MO_ACQUIRE) != 0;
321
322 // If reference is volatile, prevent following volatiles ops from
323 // floating up before the volatile access.
324 if (_access.needs_cpu_membar()) {
325 kit->insert_mem_bar(Op_MemBarCPUOrder);
326 }
327
328 if (is_atomic) {
329 assert(kit != nullptr, "unsupported at optimization time");
330 if (is_acquire || is_volatile) {
331 Node* n = _access.raw_access();
332 Node* mb = kit->insert_mem_bar(Op_MemBarAcquire, n);
333 if (_leading_membar != nullptr) {
334 MemBarNode::set_load_store_pair(_leading_membar->as_MemBar(), mb->as_MemBar());
335 }
336 }
337 } else if (is_write) {
338 // If not multiple copy atomic, we do the MemBarVolatile before the load.
339 if (is_volatile && !support_IRIW_for_not_multiple_copy_atomic_cpu) {
340 assert(kit != nullptr, "unsupported at optimization time");
341 Node* n = _access.raw_access();
342 Node* mb = kit->insert_mem_bar(Op_MemBarVolatile, n); // Use fat membar
343 if (_leading_membar != nullptr) {
344 MemBarNode::set_store_pair(_leading_membar->as_MemBar(), mb->as_MemBar());
345 }
346 }
347 } else {
348 if (is_volatile || is_acquire) {
349 assert(kit != nullptr, "unsupported at optimization time");
350 Node* n = _access.raw_access();
351 assert(_leading_membar == nullptr || support_IRIW_for_not_multiple_copy_atomic_cpu, "no leading membar expected");
352 Node* mb = kit->insert_mem_bar(Op_MemBarAcquire, n);
353 mb->as_MemBar()->set_trailing_load();
354 }
355 }
356 }
357 };
358
359 Node* BarrierSetC2::store_at(C2Access& access, C2AccessValue& val) const {
360 C2AccessFence fence(access);
361 resolve_address(access);
362 return store_at_resolved(access, val);
363 }
364
365 Node* BarrierSetC2::load_at(C2Access& access, const Type* val_type) const {
366 C2AccessFence fence(access);
367 resolve_address(access);
368 return load_at_resolved(access, val_type);
369 }
370
371 MemNode::MemOrd C2Access::mem_node_mo() const {
372 bool is_write = (_decorators & C2_WRITE_ACCESS) != 0;
373 bool is_read = (_decorators & C2_READ_ACCESS) != 0;
374 if ((_decorators & MO_SEQ_CST) != 0) {
375 if (is_write && is_read) {
376 // For atomic operations
377 return MemNode::seqcst;
378 } else if (is_write) {
379 return MemNode::release;
380 } else {
381 assert(is_read, "what else?");
382 return MemNode::acquire;
383 }
384 } else if ((_decorators & MO_RELEASE) != 0) {
385 return MemNode::release;
386 } else if ((_decorators & MO_ACQUIRE) != 0) {
387 return MemNode::acquire;
388 } else if (is_write) {
389 // Volatile fields need releasing stores.
390 // Non-volatile fields also need releasing stores if they hold an
391 // object reference, because the object reference might point to
392 // a freshly created object.
393 // Conservatively release stores of object references.
394 return StoreNode::release_if_reference(_type);
395 } else {
396 return MemNode::unordered;
397 }
398 }
399
400 void C2Access::fixup_decorators() {
401 bool default_mo = (_decorators & MO_DECORATOR_MASK) == 0;
402 bool is_unordered = (_decorators & MO_UNORDERED) != 0 || default_mo;
403 bool anonymous = (_decorators & C2_UNSAFE_ACCESS) != 0;
404
405 bool is_read = (_decorators & C2_READ_ACCESS) != 0;
406 bool is_write = (_decorators & C2_WRITE_ACCESS) != 0;
407
408 if (AlwaysAtomicAccesses && is_unordered) {
409 _decorators &= ~MO_DECORATOR_MASK; // clear the MO bits
410 _decorators |= MO_RELAXED; // Force the MO_RELAXED decorator with AlwaysAtomicAccess
411 }
412
413 _decorators = AccessInternal::decorator_fixup(_decorators, _type);
414
415 if (is_read && !is_write && anonymous) {
416 // To be valid, unsafe loads may depend on other conditions than
417 // the one that guards them: pin the Load node
418 _decorators |= C2_CONTROL_DEPENDENT_LOAD;
419 _decorators |= C2_UNKNOWN_CONTROL_LOAD;
420 const TypePtr* adr_type = _addr.type();
421 Node* adr = _addr.node();
422 if (!needs_cpu_membar() && adr_type->isa_instptr()) {
423 assert(adr_type->meet(TypePtr::NULL_PTR) != adr_type->remove_speculative(), "should be not null");
424 intptr_t offset = Type::OffsetBot;
425 AddPNode::Ideal_base_and_offset(adr, &gvn(), offset);
426 if (offset >= 0) {
427 int s = Klass::layout_helper_size_in_bytes(adr_type->isa_instptr()->instance_klass()->layout_helper());
428 if (offset < s) {
429 // Guaranteed to be a valid access, no need to pin it
430 _decorators ^= C2_CONTROL_DEPENDENT_LOAD;
431 _decorators ^= C2_UNKNOWN_CONTROL_LOAD;
432 }
433 }
434 }
435 }
436 }
437
438 //--------------------------- atomic operations---------------------------------
439
440 void BarrierSetC2::pin_atomic_op(C2AtomicParseAccess& access) const {
441 // SCMemProjNodes represent the memory state of a LoadStore. Their
442 // main role is to prevent LoadStore nodes from being optimized away
443 // when their results aren't used.
444 assert(access.is_parse_access(), "entry not supported at optimization time");
445 C2ParseAccess& parse_access = static_cast<C2ParseAccess&>(access);
446 GraphKit* kit = parse_access.kit();
447 Node* load_store = access.raw_access();
448 assert(load_store != nullptr, "must pin atomic op");
449 Node* proj = kit->gvn().transform(new SCMemProjNode(load_store));
450 kit->set_memory(proj, access.alias_idx());
451 }
452
453 void C2AtomicParseAccess::set_memory() {
454 Node *mem = _kit->memory(_alias_idx);
455 _memory = mem;
456 }
457
458 Node* BarrierSetC2::atomic_cmpxchg_val_at_resolved(C2AtomicParseAccess& access, Node* expected_val,
459 Node* new_val, const Type* value_type) const {
460 GraphKit* kit = access.kit();
461 MemNode::MemOrd mo = access.mem_node_mo();
462 Node* mem = access.memory();
463
464 Node* adr = access.addr().node();
465 const TypePtr* adr_type = access.addr().type();
466
467 Node* load_store = nullptr;
468
469 if (access.is_oop()) {
470 #ifdef _LP64
471 if (adr->bottom_type()->is_ptr_to_narrowoop()) {
472 Node *newval_enc = kit->gvn().transform(new EncodePNode(new_val, new_val->bottom_type()->make_narrowoop()));
473 Node *oldval_enc = kit->gvn().transform(new EncodePNode(expected_val, expected_val->bottom_type()->make_narrowoop()));
474 load_store = new CompareAndExchangeNNode(kit->control(), mem, adr, newval_enc, oldval_enc, adr_type, value_type->make_narrowoop(), mo);
475 } else
476 #endif
477 {
478 load_store = new CompareAndExchangePNode(kit->control(), mem, adr, new_val, expected_val, adr_type, value_type->is_oopptr(), mo);
479 }
480 } else {
481 switch (access.type()) {
482 case T_BYTE: {
483 load_store = new CompareAndExchangeBNode(kit->control(), mem, adr, new_val, expected_val, adr_type, mo);
484 break;
485 }
486 case T_SHORT: {
487 load_store = new CompareAndExchangeSNode(kit->control(), mem, adr, new_val, expected_val, adr_type, mo);
488 break;
489 }
490 case T_INT: {
491 load_store = new CompareAndExchangeINode(kit->control(), mem, adr, new_val, expected_val, adr_type, mo);
492 break;
493 }
494 case T_LONG: {
495 load_store = new CompareAndExchangeLNode(kit->control(), mem, adr, new_val, expected_val, adr_type, mo);
496 break;
497 }
498 default:
499 ShouldNotReachHere();
500 }
501 }
502
503 load_store->as_LoadStore()->set_barrier_data(access.barrier_data());
504 load_store = kit->gvn().transform(load_store);
505
506 access.set_raw_access(load_store);
507 pin_atomic_op(access);
508
509 #ifdef _LP64
510 if (access.is_oop() && adr->bottom_type()->is_ptr_to_narrowoop()) {
511 return kit->gvn().transform(new DecodeNNode(load_store, load_store->get_ptr_type()));
512 }
513 #endif
514
515 return load_store;
516 }
517
518 Node* BarrierSetC2::atomic_cmpxchg_bool_at_resolved(C2AtomicParseAccess& access, Node* expected_val,
519 Node* new_val, const Type* value_type) const {
520 GraphKit* kit = access.kit();
521 DecoratorSet decorators = access.decorators();
522 MemNode::MemOrd mo = access.mem_node_mo();
523 Node* mem = access.memory();
524 bool is_weak_cas = (decorators & C2_WEAK_CMPXCHG) != 0;
525 Node* load_store = nullptr;
526 Node* adr = access.addr().node();
527
528 if (access.is_oop()) {
529 #ifdef _LP64
530 if (adr->bottom_type()->is_ptr_to_narrowoop()) {
531 Node *newval_enc = kit->gvn().transform(new EncodePNode(new_val, new_val->bottom_type()->make_narrowoop()));
532 Node *oldval_enc = kit->gvn().transform(new EncodePNode(expected_val, expected_val->bottom_type()->make_narrowoop()));
533 if (is_weak_cas) {
534 load_store = new WeakCompareAndSwapNNode(kit->control(), mem, adr, newval_enc, oldval_enc, mo);
535 } else {
536 load_store = new CompareAndSwapNNode(kit->control(), mem, adr, newval_enc, oldval_enc, mo);
537 }
538 } else
539 #endif
540 {
541 if (is_weak_cas) {
542 load_store = new WeakCompareAndSwapPNode(kit->control(), mem, adr, new_val, expected_val, mo);
543 } else {
544 load_store = new CompareAndSwapPNode(kit->control(), mem, adr, new_val, expected_val, mo);
545 }
546 }
547 } else {
548 switch(access.type()) {
549 case T_BYTE: {
550 if (is_weak_cas) {
551 load_store = new WeakCompareAndSwapBNode(kit->control(), mem, adr, new_val, expected_val, mo);
552 } else {
553 load_store = new CompareAndSwapBNode(kit->control(), mem, adr, new_val, expected_val, mo);
554 }
555 break;
556 }
557 case T_SHORT: {
558 if (is_weak_cas) {
559 load_store = new WeakCompareAndSwapSNode(kit->control(), mem, adr, new_val, expected_val, mo);
560 } else {
561 load_store = new CompareAndSwapSNode(kit->control(), mem, adr, new_val, expected_val, mo);
562 }
563 break;
564 }
565 case T_INT: {
566 if (is_weak_cas) {
567 load_store = new WeakCompareAndSwapINode(kit->control(), mem, adr, new_val, expected_val, mo);
568 } else {
569 load_store = new CompareAndSwapINode(kit->control(), mem, adr, new_val, expected_val, mo);
570 }
571 break;
572 }
573 case T_LONG: {
574 if (is_weak_cas) {
575 load_store = new WeakCompareAndSwapLNode(kit->control(), mem, adr, new_val, expected_val, mo);
576 } else {
577 load_store = new CompareAndSwapLNode(kit->control(), mem, adr, new_val, expected_val, mo);
578 }
579 break;
580 }
581 default:
582 ShouldNotReachHere();
583 }
584 }
585
586 load_store->as_LoadStore()->set_barrier_data(access.barrier_data());
587 load_store = kit->gvn().transform(load_store);
588
589 access.set_raw_access(load_store);
590 pin_atomic_op(access);
591
592 return load_store;
593 }
594
595 Node* BarrierSetC2::atomic_xchg_at_resolved(C2AtomicParseAccess& access, Node* new_val, const Type* value_type) const {
596 GraphKit* kit = access.kit();
597 Node* mem = access.memory();
598 Node* adr = access.addr().node();
599 const TypePtr* adr_type = access.addr().type();
600 Node* load_store = nullptr;
601
602 if (access.is_oop()) {
603 #ifdef _LP64
604 if (adr->bottom_type()->is_ptr_to_narrowoop()) {
605 Node *newval_enc = kit->gvn().transform(new EncodePNode(new_val, new_val->bottom_type()->make_narrowoop()));
606 load_store = kit->gvn().transform(new GetAndSetNNode(kit->control(), mem, adr, newval_enc, adr_type, value_type->make_narrowoop()));
607 } else
608 #endif
609 {
610 load_store = new GetAndSetPNode(kit->control(), mem, adr, new_val, adr_type, value_type->is_oopptr());
611 }
612 } else {
613 switch (access.type()) {
614 case T_BYTE:
615 load_store = new GetAndSetBNode(kit->control(), mem, adr, new_val, adr_type);
616 break;
617 case T_SHORT:
618 load_store = new GetAndSetSNode(kit->control(), mem, adr, new_val, adr_type);
619 break;
620 case T_INT:
621 load_store = new GetAndSetINode(kit->control(), mem, adr, new_val, adr_type);
622 break;
623 case T_LONG:
624 load_store = new GetAndSetLNode(kit->control(), mem, adr, new_val, adr_type);
625 break;
626 default:
627 ShouldNotReachHere();
628 }
629 }
630
631 load_store->as_LoadStore()->set_barrier_data(access.barrier_data());
632 load_store = kit->gvn().transform(load_store);
633
634 access.set_raw_access(load_store);
635 pin_atomic_op(access);
636
637 #ifdef _LP64
638 if (access.is_oop() && adr->bottom_type()->is_ptr_to_narrowoop()) {
639 return kit->gvn().transform(new DecodeNNode(load_store, load_store->get_ptr_type()));
640 }
641 #endif
642
643 return load_store;
644 }
645
646 Node* BarrierSetC2::atomic_add_at_resolved(C2AtomicParseAccess& access, Node* new_val, const Type* value_type) const {
647 Node* load_store = nullptr;
648 GraphKit* kit = access.kit();
649 Node* adr = access.addr().node();
650 const TypePtr* adr_type = access.addr().type();
651 Node* mem = access.memory();
652
653 switch(access.type()) {
654 case T_BYTE:
655 load_store = new GetAndAddBNode(kit->control(), mem, adr, new_val, adr_type);
656 break;
657 case T_SHORT:
658 load_store = new GetAndAddSNode(kit->control(), mem, adr, new_val, adr_type);
659 break;
660 case T_INT:
661 load_store = new GetAndAddINode(kit->control(), mem, adr, new_val, adr_type);
662 break;
663 case T_LONG:
664 load_store = new GetAndAddLNode(kit->control(), mem, adr, new_val, adr_type);
665 break;
666 default:
667 ShouldNotReachHere();
668 }
669
670 load_store->as_LoadStore()->set_barrier_data(access.barrier_data());
671 load_store = kit->gvn().transform(load_store);
672
673 access.set_raw_access(load_store);
674 pin_atomic_op(access);
675
676 return load_store;
677 }
678
679 Node* BarrierSetC2::atomic_cmpxchg_val_at(C2AtomicParseAccess& access, Node* expected_val,
680 Node* new_val, const Type* value_type) const {
681 C2AccessFence fence(access);
682 resolve_address(access);
683 return atomic_cmpxchg_val_at_resolved(access, expected_val, new_val, value_type);
684 }
685
686 Node* BarrierSetC2::atomic_cmpxchg_bool_at(C2AtomicParseAccess& access, Node* expected_val,
687 Node* new_val, const Type* value_type) const {
688 C2AccessFence fence(access);
689 resolve_address(access);
690 return atomic_cmpxchg_bool_at_resolved(access, expected_val, new_val, value_type);
691 }
692
693 Node* BarrierSetC2::atomic_xchg_at(C2AtomicParseAccess& access, Node* new_val, const Type* value_type) const {
694 C2AccessFence fence(access);
695 resolve_address(access);
696 return atomic_xchg_at_resolved(access, new_val, value_type);
697 }
698
699 Node* BarrierSetC2::atomic_add_at(C2AtomicParseAccess& access, Node* new_val, const Type* value_type) const {
700 C2AccessFence fence(access);
701 resolve_address(access);
702 return atomic_add_at_resolved(access, new_val, value_type);
703 }
704
705 int BarrierSetC2::arraycopy_payload_base_offset(bool is_array) {
706 // Exclude the header but include array length to copy by 8 bytes words.
707 // Can't use base_offset_in_bytes(bt) since basic type is unknown.
708 int base_off = is_array ? arrayOopDesc::length_offset_in_bytes() :
709 instanceOopDesc::base_offset_in_bytes();
710 // base_off:
711 // 8 - 32-bit VM or 64-bit VM, compact headers
712 // 12 - 64-bit VM, compressed klass
713 // 16 - 64-bit VM, normal klass
714 if (base_off % BytesPerLong != 0) {
715 assert(!UseCompactObjectHeaders, "");
716 if (is_array) {
717 // Exclude length to copy by 8 bytes words.
718 base_off += sizeof(int);
719 } else {
720 // Include klass to copy by 8 bytes words.
721 base_off = instanceOopDesc::klass_offset_in_bytes();
722 }
723 assert(base_off % BytesPerLong == 0, "expect 8 bytes alignment");
724 }
725 return base_off;
726 }
727
728 void BarrierSetC2::clone(GraphKit* kit, Node* src_base, Node* dst_base, Node* size, bool is_array) const {
729 int base_off = arraycopy_payload_base_offset(is_array);
730 Node* payload_size = size;
731 Node* offset = kit->MakeConX(base_off);
732 payload_size = kit->gvn().transform(new SubXNode(payload_size, offset));
733 if (is_array) {
734 // Ensure the array payload size is rounded up to the next BytesPerLong
735 // multiple when converting to double-words. This is necessary because array
736 // size does not include object alignment padding, so it might not be a
737 // multiple of BytesPerLong for sub-long element types.
738 payload_size = kit->gvn().transform(new AddXNode(payload_size, kit->MakeConX(BytesPerLong - 1)));
739 }
740 payload_size = kit->gvn().transform(new URShiftXNode(payload_size, kit->intcon(LogBytesPerLong)));
741 ArrayCopyNode* ac = ArrayCopyNode::make(kit, false, src_base, offset, dst_base, offset, payload_size, true, false);
742 if (is_array) {
743 ac->set_clone_array();
744 } else {
745 ac->set_clone_inst();
746 }
747 Node* n = kit->gvn().transform(ac);
748 if (n == ac) {
749 const TypePtr* raw_adr_type = TypeRawPtr::BOTTOM;
750 ac->set_adr_type(TypeRawPtr::BOTTOM);
751 kit->set_predefined_output_for_runtime_call(ac, ac->in(TypeFunc::Memory), raw_adr_type);
752 } else {
753 kit->set_all_memory(n);
754 }
755 }
756
757 Node* BarrierSetC2::obj_allocate(PhaseMacroExpand* macro, Node* mem, Node* toobig_false, Node* size_in_bytes,
758 Node*& i_o, Node*& needgc_ctrl,
759 Node*& fast_oop_ctrl, Node*& fast_oop_rawmem,
760 intx prefetch_lines) const {
761 assert(UseTLAB, "Only for TLAB enabled allocations");
762
763 Node* thread = macro->transform_later(new ThreadLocalNode());
764 Node* tlab_top_adr = macro->off_heap_plus_addr(thread, in_bytes(JavaThread::tlab_top_offset()));
765 Node* tlab_end_adr = macro->off_heap_plus_addr(thread, in_bytes(JavaThread::tlab_end_offset()));
766
767 // Load TLAB end.
768 //
769 // Note: We set the control input on "tlab_end" and "old_tlab_top" to work around
770 // a bug where these values were being moved across
771 // a safepoint. These are not oops, so they cannot be include in the oop
772 // map, but they can be changed by a GC. The proper way to fix this would
773 // be to set the raw memory state when generating a SafepointNode. However
774 // this will require extensive changes to the loop optimization in order to
775 // prevent a degradation of the optimization.
776 // See comment in memnode.hpp, around line 227 in class LoadPNode.
777 Node* tlab_end = macro->make_load_raw(toobig_false, mem, tlab_end_adr, 0, TypeRawPtr::BOTTOM, T_ADDRESS);
778
779 // Load the TLAB top.
780 Node* old_tlab_top = new LoadPNode(toobig_false, mem, tlab_top_adr, TypeRawPtr::BOTTOM, TypeRawPtr::BOTTOM, MemNode::unordered);
781 macro->transform_later(old_tlab_top);
782
783 // Add to heap top to get a new TLAB top
784 Node* new_tlab_top = AddPNode::make_off_heap(old_tlab_top, size_in_bytes);
785 macro->transform_later(new_tlab_top);
786
787 // Check against TLAB end
788 Node* tlab_full = new CmpPNode(new_tlab_top, tlab_end);
789 macro->transform_later(tlab_full);
790
791 Node* needgc_bol = new BoolNode(tlab_full, BoolTest::ge);
792 macro->transform_later(needgc_bol);
793 IfNode* needgc_iff = new IfNode(toobig_false, needgc_bol, PROB_UNLIKELY_MAG(4), COUNT_UNKNOWN);
794 macro->transform_later(needgc_iff);
795
796 // Plug the failing-heap-space-need-gc test into the slow-path region
797 Node* needgc_true = new IfTrueNode(needgc_iff);
798 macro->transform_later(needgc_true);
799 needgc_ctrl = needgc_true;
800
801 // No need for a GC.
802 Node* needgc_false = new IfFalseNode(needgc_iff);
803 macro->transform_later(needgc_false);
804
805 // Fast path:
806 i_o = macro->prefetch_allocation(i_o, needgc_false, mem,
807 old_tlab_top, new_tlab_top, prefetch_lines);
808
809 // Store the modified TLAB top back down.
810 Node* store_tlab_top = new StorePNode(needgc_false, mem, tlab_top_adr,
811 TypeRawPtr::BOTTOM, new_tlab_top, MemNode::unordered);
812 macro->transform_later(store_tlab_top);
813
814 fast_oop_ctrl = needgc_false;
815 fast_oop_rawmem = store_tlab_top;
816 return old_tlab_top;
817 }
818
819 const TypeFunc* BarrierSetC2::_clone_type_Type = nullptr;
820
821 void BarrierSetC2::make_clone_type() {
822 assert(BarrierSetC2::_clone_type_Type == nullptr, "should be");
823 // Create input type (domain)
824 int argcnt = NOT_LP64(3) LP64_ONLY(4);
825 const Type** const domain_fields = TypeTuple::fields(argcnt);
826 int argp = TypeFunc::Parms;
827 domain_fields[argp++] = TypeInstPtr::NOTNULL; // src
828 domain_fields[argp++] = TypeInstPtr::NOTNULL; // dst
829 domain_fields[argp++] = TypeX_X; // size lower
830 LP64_ONLY(domain_fields[argp++] = Type::HALF); // size upper
831 assert(argp == TypeFunc::Parms+argcnt, "correct decoding");
832 const TypeTuple* const domain = TypeTuple::make(TypeFunc::Parms + argcnt, domain_fields);
833
834 // Create result type (range)
835 const Type** const range_fields = TypeTuple::fields(0);
836 const TypeTuple* const range = TypeTuple::make(TypeFunc::Parms + 0, range_fields);
837
838 BarrierSetC2::_clone_type_Type = TypeFunc::make(domain, range);
839 }
840
841 inline const TypeFunc* BarrierSetC2::clone_type() {
842 assert(BarrierSetC2::_clone_type_Type != nullptr, "should be initialized");
843 return BarrierSetC2::_clone_type_Type;
844 }
845
846 #define XTOP LP64_ONLY(COMMA phase->top())
847
848 void BarrierSetC2::clone_in_runtime(PhaseMacroExpand* phase, ArrayCopyNode* ac,
849 address clone_addr, const char* clone_name) const {
850 Node* const ctrl = ac->in(TypeFunc::Control);
851 Node* const mem = ac->in(TypeFunc::Memory);
852 Node* const src = ac->in(ArrayCopyNode::Src);
853 Node* const dst = ac->in(ArrayCopyNode::Dest);
854 Node* const size = ac->in(ArrayCopyNode::Length);
855
856 assert(size->bottom_type()->base() == Type_X,
857 "Should be of object size type (int for 32 bits, long for 64 bits)");
858
859 // The native clone we are calling here expects the object size in words.
860 // Add header/offset size to payload size to get object size.
861 Node* const base_offset = phase->MakeConX(arraycopy_payload_base_offset(ac->is_clone_array()) >> LogBytesPerLong);
862 Node* const full_size = phase->transform_later(new AddXNode(size, base_offset));
863 // HeapAccess<>::clone expects size in heap words.
864 // For 64-bits platforms, this is a no-operation.
865 // For 32-bits platforms, we need to multiply full_size by HeapWordsPerLong (2).
866 Node* const full_size_in_heap_words = phase->transform_later(new LShiftXNode(full_size, phase->intcon(LogHeapWordsPerLong)));
867
868 Node* const call = phase->make_leaf_call(ctrl,
869 mem,
870 clone_type(),
871 clone_addr,
872 clone_name,
873 TypeRawPtr::BOTTOM,
874 src, dst, full_size_in_heap_words XTOP);
875 phase->transform_later(call);
876 phase->replace_node(ac, call);
877 }
878
879 void BarrierSetC2::clone_at_expansion(PhaseMacroExpand* phase, ArrayCopyNode* ac) const {
880 Node* ctrl = ac->in(TypeFunc::Control);
881 Node* mem = ac->in(TypeFunc::Memory);
882 Node* src = ac->in(ArrayCopyNode::Src);
883 Node* src_offset = ac->in(ArrayCopyNode::SrcPos);
884 Node* dest = ac->in(ArrayCopyNode::Dest);
885 Node* dest_offset = ac->in(ArrayCopyNode::DestPos);
886 Node* length = ac->in(ArrayCopyNode::Length);
887
888 Node* payload_src = phase->basic_plus_adr(src, src_offset);
889 Node* payload_dst = phase->basic_plus_adr(dest, dest_offset);
890
891 const char* copyfunc_name = "arraycopy";
892 address copyfunc_addr = phase->basictype2arraycopy(T_LONG, nullptr, nullptr, true, copyfunc_name, true);
893
894 const TypePtr* raw_adr_type = TypeRawPtr::BOTTOM;
895 const TypeFunc* call_type = OptoRuntime::fast_arraycopy_Type();
896
897 Node* call = phase->make_leaf_call(ctrl, mem, call_type, copyfunc_addr, copyfunc_name, raw_adr_type, payload_src, payload_dst, length XTOP);
898 phase->transform_later(call);
899
900 phase->replace_node(ac, call);
901 }
902
903 #undef XTOP
904
905 static bool block_has_safepoint(const Block* block, uint from, uint to) {
906 for (uint i = from; i < to; i++) {
907 if (block->get_node(i)->is_MachSafePoint()) {
908 // Safepoint found
909 return true;
910 }
911 }
912
913 // Safepoint not found
914 return false;
915 }
916
917 static bool block_has_safepoint(const Block* block) {
918 return block_has_safepoint(block, 0, block->number_of_nodes());
919 }
920
921 static uint block_index(const Block* block, const Node* node) {
922 for (uint j = 0; j < block->number_of_nodes(); ++j) {
923 if (block->get_node(j) == node) {
924 return j;
925 }
926 }
927 ShouldNotReachHere();
928 return 0;
929 }
930
931 // Look through various node aliases
932 static const Node* look_through_node(const Node* node) {
933 while (node != nullptr) {
934 const Node* new_node = node;
935 if (node->is_Mach()) {
936 const MachNode* const node_mach = node->as_Mach();
937 if (node_mach->ideal_Opcode() == Op_CheckCastPP) {
938 new_node = node->in(1);
939 }
940 if (node_mach->is_SpillCopy()) {
941 new_node = node->in(1);
942 }
943 }
944 if (new_node == node || new_node == nullptr) {
945 break;
946 } else {
947 node = new_node;
948 }
949 }
950
951 return node;
952 }
953
954 // Whether the given offset is undefined.
955 static bool is_undefined(intptr_t offset) {
956 return offset == Type::OffsetTop;
957 }
958
959 // Whether the given offset is unknown.
960 static bool is_unknown(intptr_t offset) {
961 return offset == Type::OffsetBot;
962 }
963
964 // Whether the given offset is concrete (defined and compile-time known).
965 static bool is_concrete(intptr_t offset) {
966 return !is_undefined(offset) && !is_unknown(offset);
967 }
968
969 // Compute base + offset components of the memory address accessed by mach.
970 // Return a node representing the base address, or null if the base cannot be
971 // found or the offset is undefined or a concrete negative value. If a non-null
972 // base is returned, the offset is a concrete, nonnegative value or unknown.
973 static const Node* get_base_and_offset(const MachNode* mach, intptr_t& offset) {
974 const TypePtr* adr_type = nullptr;
975 offset = 0;
976 const Node* base = mach->get_base_and_disp(offset, adr_type);
977
978 if (base == nullptr || base == NodeSentinel) {
979 return nullptr;
980 }
981
982 if (offset == 0 && base->is_Mach() && base->as_Mach()->ideal_Opcode() == Op_AddP) {
983 // The memory address is computed by 'base' and fed to 'mach' via an
984 // indirect memory operand (indicated by offset == 0). The ultimate base and
985 // offset can be fetched directly from the inputs and Ideal type of 'base'.
986 const TypeOopPtr* oopptr = base->bottom_type()->isa_oopptr();
987 if (oopptr == nullptr) return nullptr;
988 offset = oopptr->offset();
989 // Even if 'base' is not an Ideal AddP node anymore, Matcher::ReduceInst()
990 // guarantees that the base address is still available at the same slot.
991 base = base->in(AddPNode::Base);
992 assert(base != nullptr, "");
993 }
994
995 if (is_undefined(offset) || (is_concrete(offset) && offset < 0)) {
996 return nullptr;
997 }
998
999 return look_through_node(base);
1000 }
1001
1002 // Whether a phi node corresponds to an array allocation.
1003 // This test is incomplete: in some edge cases, it might return false even
1004 // though the node does correspond to an array allocation.
1005 static bool is_array_allocation(const Node* phi) {
1006 precond(phi->is_Phi());
1007 // Check whether phi has a successor cast (CheckCastPP) to Java array pointer,
1008 // possibly below spill copies and other cast nodes. Limit the exploration to
1009 // a single path from the phi node consisting of these node types.
1010 const Node* current = phi;
1011 while (true) {
1012 const Node* next = nullptr;
1013 for (DUIterator_Fast imax, i = current->fast_outs(imax); i < imax; i++) {
1014 if (!current->fast_out(i)->isa_Mach()) {
1015 continue;
1016 }
1017 const MachNode* succ = current->fast_out(i)->as_Mach();
1018 if (succ->ideal_Opcode() == Op_CheckCastPP) {
1019 if (succ->get_ptr_type()->isa_aryptr()) {
1020 // Cast to Java array pointer: phi corresponds to an array allocation.
1021 return true;
1022 }
1023 // Other cast: record as candidate for further exploration.
1024 next = succ;
1025 } else if (succ->is_SpillCopy() && next == nullptr) {
1026 // Spill copy, and no better candidate found: record as candidate.
1027 next = succ;
1028 }
1029 }
1030 if (next == nullptr) {
1031 // No evidence found that phi corresponds to an array allocation, and no
1032 // candidates available to continue exploring.
1033 return false;
1034 }
1035 // Continue exploring from the best candidate found.
1036 current = next;
1037 }
1038 ShouldNotReachHere();
1039 }
1040
1041 bool BarrierSetC2::is_allocation(const Node* node) {
1042 assert(node->is_Phi(), "expected phi node");
1043 if (node->req() != 3) {
1044 return false;
1045 }
1046 const Node* const fast_node = node->in(2);
1047 if (!fast_node->is_Mach()) {
1048 return false;
1049 }
1050 const MachNode* const fast_mach = fast_node->as_Mach();
1051 if (fast_mach->ideal_Opcode() != Op_LoadP) {
1052 return false;
1053 }
1054 intptr_t offset;
1055 const Node* const base = get_base_and_offset(fast_mach, offset);
1056 if (base == nullptr || !base->is_Mach() || !is_concrete(offset)) {
1057 return false;
1058 }
1059 const MachNode* const base_mach = base->as_Mach();
1060 if (base_mach->ideal_Opcode() != Op_ThreadLocal) {
1061 return false;
1062 }
1063 return offset == in_bytes(Thread::tlab_top_offset());
1064 }
1065
1066 void BarrierSetC2::elide_dominated_barriers(Node_List& accesses, Node_List& access_dominators) const {
1067 Compile* const C = Compile::current();
1068 PhaseCFG* const cfg = C->cfg();
1069
1070 for (uint i = 0; i < accesses.size(); i++) {
1071 MachNode* const access = accesses.at(i)->as_Mach();
1072 intptr_t access_offset;
1073 const Node* const access_obj = get_base_and_offset(access, access_offset);
1074 Block* const access_block = cfg->get_block_for_node(access);
1075 const uint access_index = block_index(access_block, access);
1076
1077 if (access_obj == nullptr) {
1078 // No information available
1079 continue;
1080 }
1081
1082 for (uint j = 0; j < access_dominators.size(); j++) {
1083 const Node* const mem = access_dominators.at(j);
1084 if (mem->is_Phi()) {
1085 assert(is_allocation(mem), "expected allocation phi node");
1086 if (mem != access_obj) {
1087 continue;
1088 }
1089 if (is_unknown(access_offset) && !is_array_allocation(mem)) {
1090 // The accessed address has an unknown offset, but the allocated
1091 // object cannot be determined to be an array. Avoid eliding in this
1092 // case, to be on the safe side.
1093 continue;
1094 }
1095 assert((is_concrete(access_offset) && access_offset >= 0) || (is_unknown(access_offset) && is_array_allocation(mem)),
1096 "candidate allocation-dominated access offsets must be either concrete and nonnegative, or unknown (for array allocations only)");
1097 } else {
1098 // Access node
1099 const MachNode* const mem_mach = mem->as_Mach();
1100 intptr_t mem_offset;
1101 const Node* const mem_obj = get_base_and_offset(mem_mach, mem_offset);
1102
1103 if (mem_obj == nullptr ||
1104 !is_concrete(access_offset) ||
1105 !is_concrete(mem_offset)) {
1106 // No information available
1107 continue;
1108 }
1109
1110 if (mem_obj != access_obj || mem_offset != access_offset) {
1111 // Not the same addresses, not a candidate
1112 continue;
1113 }
1114 assert(is_concrete(access_offset) && access_offset >= 0,
1115 "candidate non-allocation-dominated access offsets must be concrete and nonnegative");
1116 }
1117
1118 Block* mem_block = cfg->get_block_for_node(mem);
1119 const uint mem_index = block_index(mem_block, mem);
1120
1121 if (access_block == mem_block) {
1122 // Earlier accesses in the same block
1123 if (mem_index < access_index && !block_has_safepoint(mem_block, mem_index + 1, access_index)) {
1124 elide_dominated_barrier(access);
1125 }
1126 } else if (mem_block->dominates(access_block)) {
1127 // Dominating block? Look around for safepoints
1128 ResourceMark rm;
1129 Block_List stack;
1130 VectorSet visited;
1131 stack.push(access_block);
1132 bool safepoint_found = block_has_safepoint(access_block);
1133 while (!safepoint_found && stack.size() > 0) {
1134 const Block* const block = stack.pop();
1135 if (visited.test_set(block->_pre_order)) {
1136 continue;
1137 }
1138 if (block_has_safepoint(block)) {
1139 safepoint_found = true;
1140 break;
1141 }
1142 if (block == mem_block) {
1143 continue;
1144 }
1145
1146 // Push predecessor blocks
1147 for (uint p = 1; p < block->num_preds(); ++p) {
1148 Block* const pred = cfg->get_block_for_node(block->pred(p));
1149 stack.push(pred);
1150 }
1151 }
1152
1153 if (!safepoint_found) {
1154 elide_dominated_barrier(access);
1155 }
1156 }
1157 }
1158 }
1159 }
1160
1161 void BarrierSetC2::compute_liveness_at_stubs() const {
1162 ResourceMark rm;
1163 Compile* const C = Compile::current();
1164 Arena* const A = Thread::current()->resource_area();
1165 PhaseCFG* const cfg = C->cfg();
1166 PhaseRegAlloc* const regalloc = C->regalloc();
1167 RegMask* const live = NEW_ARENA_ARRAY(A, RegMask, cfg->number_of_blocks() * sizeof(RegMask));
1168 BarrierSetAssembler* const bs = BarrierSet::barrier_set()->barrier_set_assembler();
1169 BarrierSetC2State* bs_state = barrier_set_state();
1170 Block_List worklist;
1171
1172 for (uint i = 0; i < cfg->number_of_blocks(); ++i) {
1173 new ((void*)(live + i)) RegMask();
1174 worklist.push(cfg->get_block(i));
1175 }
1176
1177 while (worklist.size() > 0) {
1178 const Block* const block = worklist.pop();
1179 RegMask& old_live = live[block->_pre_order];
1180 RegMask new_live;
1181
1182 // Initialize to union of successors
1183 for (uint i = 0; i < block->_num_succs; i++) {
1184 const uint succ_id = block->_succs[i]->_pre_order;
1185 new_live.or_with(live[succ_id]);
1186 }
1187
1188 // Walk block backwards, computing liveness
1189 for (int i = block->number_of_nodes() - 1; i >= 0; --i) {
1190 const Node* const node = block->get_node(i);
1191
1192 // If this node tracks out-liveness, update it
1193 if (!bs_state->needs_livein_data()) {
1194 RegMask* const regs = bs_state->live(node);
1195 if (regs != nullptr) {
1196 regs->or_with(new_live);
1197 }
1198 }
1199
1200 // Remove def bits
1201 const OptoReg::Name first = bs->refine_register(node, regalloc->get_reg_first(node));
1202 const OptoReg::Name second = bs->refine_register(node, regalloc->get_reg_second(node));
1203 if (first != OptoReg::Bad) {
1204 new_live.remove(first);
1205 }
1206 if (second != OptoReg::Bad) {
1207 new_live.remove(second);
1208 }
1209
1210 // Add use bits
1211 for (uint j = 1; j < node->req(); ++j) {
1212 const Node* const use = node->in(j);
1213 const OptoReg::Name first = bs->refine_register(use, regalloc->get_reg_first(use));
1214 const OptoReg::Name second = bs->refine_register(use, regalloc->get_reg_second(use));
1215 if (first != OptoReg::Bad) {
1216 new_live.insert(first);
1217 }
1218 if (second != OptoReg::Bad) {
1219 new_live.insert(second);
1220 }
1221 }
1222
1223 // If this node tracks in-liveness, update it
1224 if (bs_state->needs_livein_data()) {
1225 RegMask* const regs = bs_state->live(node);
1226 if (regs != nullptr) {
1227 regs->or_with(new_live);
1228 }
1229 }
1230 }
1231
1232 // Now at block top, see if we have any changes
1233 new_live.subtract(old_live);
1234 if (!new_live.is_empty()) {
1235 // Liveness has refined, update and propagate to prior blocks
1236 old_live.or_with(new_live);
1237 for (uint i = 1; i < block->num_preds(); ++i) {
1238 Block* const pred = cfg->get_block_for_node(block->pred(i));
1239 worklist.push(pred);
1240 }
1241 }
1242 }
1243 }