1 /* 2 * Copyright (c) 2015, 2022, Red Hat, Inc. All rights reserved. 3 * Copyright Amazon.com Inc. or its affiliates. All Rights Reserved. 4 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 5 * 6 * This code is free software; you can redistribute it and/or modify it 7 * under the terms of the GNU General Public License version 2 only, as 8 * published by the Free Software Foundation. 9 * 10 * This code is distributed in the hope that it will be useful, but WITHOUT 11 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 12 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 13 * version 2 for more details (a copy is included in the LICENSE file that 14 * accompanied this code). 15 * 16 * You should have received a copy of the GNU General Public License version 17 * 2 along with this work; if not, write to the Free Software Foundation, 18 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 19 * 20 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 21 * or visit www.oracle.com if you need additional information or have any 22 * questions. 23 * 24 */ 25 26 #ifndef SHARE_GC_SHENANDOAH_SHENANDOAHBARRIERSET_INLINE_HPP 27 #define SHARE_GC_SHENANDOAH_SHENANDOAHBARRIERSET_INLINE_HPP 28 29 #include "gc/shenandoah/shenandoahBarrierSet.hpp" 30 31 #include "gc/shared/accessBarrierSupport.inline.hpp" 32 #include "gc/shared/cardTable.hpp" 33 #include "gc/shenandoah/shenandoahAsserts.hpp" 34 #include "gc/shenandoah/shenandoahCardTable.hpp" 35 #include "gc/shenandoah/shenandoahCollectionSet.inline.hpp" 36 #include "gc/shenandoah/shenandoahEvacOOMHandler.inline.hpp" 37 #include "gc/shenandoah/shenandoahForwarding.inline.hpp" 38 #include "gc/shenandoah/shenandoahGeneration.hpp" 39 #include "gc/shenandoah/shenandoahHeap.inline.hpp" 40 #include "gc/shenandoah/shenandoahHeapRegion.hpp" 41 #include "gc/shenandoah/shenandoahMarkingContext.inline.hpp" 42 #include "gc/shenandoah/shenandoahThreadLocalData.hpp" 43 #include "gc/shenandoah/mode/shenandoahMode.hpp" 44 #include "memory/iterator.inline.hpp" 45 #include "oops/oop.inline.hpp" 46 47 inline oop ShenandoahBarrierSet::resolve_forwarded_not_null(oop p) { 48 return ShenandoahForwarding::get_forwardee(p); 49 } 50 51 inline oop ShenandoahBarrierSet::resolve_forwarded(oop p) { 52 if (p != nullptr) { 53 return resolve_forwarded_not_null(p); 54 } else { 55 return p; 56 } 57 } 58 59 inline oop ShenandoahBarrierSet::resolve_forwarded_not_null_mutator(oop p) { 60 return ShenandoahForwarding::get_forwardee_mutator(p); 61 } 62 63 template <class T> 64 inline oop ShenandoahBarrierSet::load_reference_barrier_mutator(oop obj, T* load_addr) { 65 assert(ShenandoahLoadRefBarrier, "should be enabled"); 66 shenandoah_assert_in_cset(load_addr, obj); 67 68 oop fwd = resolve_forwarded_not_null_mutator(obj); 69 if (obj == fwd) { 70 assert(_heap->is_evacuation_in_progress(), 71 "evac should be in progress"); 72 Thread* const t = Thread::current(); 73 ShenandoahEvacOOMScope scope(t); 74 fwd = _heap->evacuate_object(obj, t); 75 } 76 77 if (load_addr != nullptr && fwd != obj) { 78 // Since we are here and we know the load address, update the reference. 79 ShenandoahHeap::atomic_update_oop(fwd, load_addr, obj); 80 } 81 82 return fwd; 83 } 84 85 inline oop ShenandoahBarrierSet::load_reference_barrier(oop obj) { 86 if (!ShenandoahLoadRefBarrier) { 87 return obj; 88 } 89 if (_heap->has_forwarded_objects() && 90 _heap->in_collection_set(obj)) { // Subsumes null-check 91 assert(obj != nullptr, "cset check must have subsumed null-check"); 92 oop fwd = resolve_forwarded_not_null(obj); 93 if (obj == fwd && _heap->is_evacuation_in_progress()) { 94 Thread* t = Thread::current(); 95 ShenandoahEvacOOMScope oom_evac_scope(t); 96 return _heap->evacuate_object(obj, t); 97 } 98 return fwd; 99 } 100 return obj; 101 } 102 103 template <class T> 104 inline oop ShenandoahBarrierSet::load_reference_barrier(DecoratorSet decorators, oop obj, T* load_addr) { 105 if (obj == nullptr) { 106 return nullptr; 107 } 108 109 // Prevent resurrection of unreachable phantom (i.e. weak-native) references. 110 if ((decorators & ON_PHANTOM_OOP_REF) != 0 && 111 _heap->is_concurrent_weak_root_in_progress() && 112 _heap->is_in_active_generation(obj) && 113 !_heap->marking_context()->is_marked(obj)) { 114 return nullptr; 115 } 116 117 // Prevent resurrection of unreachable weak references. 118 if ((decorators & ON_WEAK_OOP_REF) != 0 && 119 _heap->is_concurrent_weak_root_in_progress() && 120 _heap->is_in_active_generation(obj) && 121 !_heap->marking_context()->is_marked_strong(obj)) { 122 return nullptr; 123 } 124 125 // Prevent resurrection of unreachable objects that are visited during 126 // concurrent class-unloading. 127 if ((decorators & AS_NO_KEEPALIVE) != 0 && 128 _heap->is_evacuation_in_progress() && 129 !_heap->marking_context()->is_marked(obj)) { 130 return obj; 131 } 132 133 oop fwd = load_reference_barrier(obj); 134 if (load_addr != nullptr && fwd != obj) { 135 // Since we are here and we know the load address, update the reference. 136 ShenandoahHeap::atomic_update_oop(fwd, load_addr, obj); 137 } 138 139 return fwd; 140 } 141 142 inline void ShenandoahBarrierSet::enqueue(oop obj) { 143 assert(obj != nullptr, "checked by caller"); 144 assert(_satb_mark_queue_set.is_active(), "only get here when SATB active"); 145 146 // Filter marked objects before hitting the SATB queues. The same predicate would 147 // be used by SATBMQ::filter to eliminate already marked objects downstream, but 148 // filtering here helps to avoid wasteful SATB queueing work to begin with. 149 if (!_heap->requires_marking(obj)) return; 150 151 SATBMarkQueue& queue = ShenandoahThreadLocalData::satb_mark_queue(Thread::current()); 152 _satb_mark_queue_set.enqueue_known_active(queue, obj); 153 } 154 155 template <DecoratorSet decorators, typename T> 156 inline void ShenandoahBarrierSet::satb_barrier(T *field) { 157 if (HasDecorator<decorators, IS_DEST_UNINITIALIZED>::value || 158 HasDecorator<decorators, AS_NO_KEEPALIVE>::value) { 159 return; 160 } 161 if (ShenandoahSATBBarrier && _heap->is_concurrent_mark_in_progress()) { 162 T heap_oop = RawAccess<>::oop_load(field); 163 if (!CompressedOops::is_null(heap_oop)) { 164 enqueue(CompressedOops::decode(heap_oop)); 165 } 166 } 167 } 168 169 inline void ShenandoahBarrierSet::satb_enqueue(oop value) { 170 if (value != nullptr && ShenandoahSATBBarrier && _heap->is_concurrent_mark_in_progress()) { 171 enqueue(value); 172 } 173 } 174 175 inline void ShenandoahBarrierSet::keep_alive_if_weak(DecoratorSet decorators, oop value) { 176 assert((decorators & ON_UNKNOWN_OOP_REF) == 0, "Reference strength must be known"); 177 const bool on_strong_oop_ref = (decorators & ON_STRONG_OOP_REF) != 0; 178 const bool peek = (decorators & AS_NO_KEEPALIVE) != 0; 179 if (!peek && !on_strong_oop_ref) { 180 satb_enqueue(value); 181 } 182 } 183 184 template <DecoratorSet decorators, typename T> 185 inline void ShenandoahBarrierSet::write_ref_field_post(T* field) { 186 assert(ShenandoahCardBarrier, "Should have been checked by caller"); 187 volatile CardTable::CardValue* byte = card_table()->byte_for(field); 188 *byte = CardTable::dirty_card_val(); 189 } 190 191 template <typename T> 192 inline oop ShenandoahBarrierSet::oop_load(DecoratorSet decorators, T* addr) { 193 oop value = RawAccess<>::oop_load(addr); 194 value = load_reference_barrier(decorators, value, addr); 195 keep_alive_if_weak(decorators, value); 196 return value; 197 } 198 199 template <typename T> 200 inline oop ShenandoahBarrierSet::oop_cmpxchg(DecoratorSet decorators, T* addr, oop compare_value, oop new_value) { 201 oop res; 202 oop expected = compare_value; 203 do { 204 compare_value = expected; 205 res = RawAccess<>::oop_atomic_cmpxchg(addr, compare_value, new_value); 206 expected = res; 207 } while ((compare_value != expected) && (resolve_forwarded(compare_value) == resolve_forwarded(expected))); 208 209 // Note: We don't need a keep-alive-barrier here. We already enqueue any loaded reference for SATB anyway, 210 // because it must be the previous value. 211 res = load_reference_barrier(decorators, res, static_cast<T*>(nullptr)); 212 satb_enqueue(res); 213 return res; 214 } 215 216 template <typename T> 217 inline oop ShenandoahBarrierSet::oop_xchg(DecoratorSet decorators, T* addr, oop new_value) { 218 oop previous = RawAccess<>::oop_atomic_xchg(addr, new_value); 219 // Note: We don't need a keep-alive-barrier here. We already enqueue any loaded reference for SATB anyway, 220 // because it must be the previous value. 221 previous = load_reference_barrier<T>(decorators, previous, static_cast<T*>(nullptr)); 222 satb_enqueue(previous); 223 return previous; 224 } 225 226 template <DecoratorSet decorators, typename BarrierSetT> 227 template <typename T> 228 inline oop ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_load_not_in_heap(T* addr) { 229 assert((decorators & ON_UNKNOWN_OOP_REF) == 0, "must be absent"); 230 ShenandoahBarrierSet* const bs = ShenandoahBarrierSet::barrier_set(); 231 return bs->oop_load(decorators, addr); 232 } 233 234 template <DecoratorSet decorators, typename BarrierSetT> 235 template <typename T> 236 inline oop ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_load_in_heap(T* addr) { 237 assert((decorators & ON_UNKNOWN_OOP_REF) == 0, "must be absent"); 238 ShenandoahBarrierSet* const bs = ShenandoahBarrierSet::barrier_set(); 239 return bs->oop_load(decorators, addr); 240 } 241 242 template <DecoratorSet decorators, typename BarrierSetT> 243 inline oop ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_load_in_heap_at(oop base, ptrdiff_t offset) { 244 ShenandoahBarrierSet* const bs = ShenandoahBarrierSet::barrier_set(); 245 DecoratorSet resolved_decorators = AccessBarrierSupport::resolve_possibly_unknown_oop_ref_strength<decorators>(base, offset); 246 return bs->oop_load(resolved_decorators, AccessInternal::oop_field_addr<decorators>(base, offset)); 247 } 248 249 template <DecoratorSet decorators, typename BarrierSetT> 250 template <typename T> 251 inline void ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_store_common(T* addr, oop value) { 252 shenandoah_assert_marked_if(nullptr, value, 253 !CompressedOops::is_null(value) && ShenandoahHeap::heap()->is_evacuation_in_progress() 254 && !(ShenandoahHeap::heap()->active_generation()->is_young() 255 && ShenandoahHeap::heap()->heap_region_containing(value)->is_old())); 256 shenandoah_assert_not_in_cset_if(addr, value, value != nullptr && !ShenandoahHeap::heap()->cancelled_gc()); 257 ShenandoahBarrierSet* const bs = ShenandoahBarrierSet::barrier_set(); 258 bs->satb_barrier<decorators>(addr); 259 Raw::oop_store(addr, value); 260 } 261 262 template <DecoratorSet decorators, typename BarrierSetT> 263 template <typename T> 264 inline void ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_store_not_in_heap(T* addr, oop value) { 265 oop_store_common(addr, value); 266 } 267 268 template <DecoratorSet decorators, typename BarrierSetT> 269 template <typename T> 270 inline void ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_store_in_heap(T* addr, oop value) { 271 shenandoah_assert_not_in_cset_loc_except(addr, ShenandoahHeap::heap()->cancelled_gc()); 272 shenandoah_assert_not_forwarded_except (addr, value, value == nullptr || ShenandoahHeap::heap()->cancelled_gc() || !ShenandoahHeap::heap()->is_concurrent_mark_in_progress()); 273 274 oop_store_common(addr, value); 275 if (ShenandoahCardBarrier) { 276 ShenandoahBarrierSet* bs = ShenandoahBarrierSet::barrier_set(); 277 bs->write_ref_field_post<decorators>(addr); 278 } 279 } 280 281 template <DecoratorSet decorators, typename BarrierSetT> 282 inline void ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_store_in_heap_at(oop base, ptrdiff_t offset, oop value) { 283 oop_store_in_heap(AccessInternal::oop_field_addr<decorators>(base, offset), value); 284 } 285 286 template <DecoratorSet decorators, typename BarrierSetT> 287 template <typename T> 288 inline oop ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_atomic_cmpxchg_not_in_heap(T* addr, oop compare_value, oop new_value) { 289 assert((decorators & (AS_NO_KEEPALIVE | ON_UNKNOWN_OOP_REF)) == 0, "must be absent"); 290 ShenandoahBarrierSet* bs = ShenandoahBarrierSet::barrier_set(); 291 return bs->oop_cmpxchg(decorators, addr, compare_value, new_value); 292 } 293 294 template <DecoratorSet decorators, typename BarrierSetT> 295 template <typename T> 296 inline oop ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_atomic_cmpxchg_in_heap(T* addr, oop compare_value, oop new_value) { 297 assert((decorators & (AS_NO_KEEPALIVE | ON_UNKNOWN_OOP_REF)) == 0, "must be absent"); 298 ShenandoahBarrierSet* bs = ShenandoahBarrierSet::barrier_set(); 299 oop result = bs->oop_cmpxchg(decorators, addr, compare_value, new_value); 300 if (ShenandoahCardBarrier) { 301 bs->write_ref_field_post<decorators>(addr); 302 } 303 return result; 304 } 305 306 template <DecoratorSet decorators, typename BarrierSetT> 307 inline oop ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_atomic_cmpxchg_in_heap_at(oop base, ptrdiff_t offset, oop compare_value, oop new_value) { 308 assert((decorators & AS_NO_KEEPALIVE) == 0, "must be absent"); 309 ShenandoahBarrierSet* bs = ShenandoahBarrierSet::barrier_set(); 310 DecoratorSet resolved_decorators = AccessBarrierSupport::resolve_possibly_unknown_oop_ref_strength<decorators>(base, offset); 311 auto addr = AccessInternal::oop_field_addr<decorators>(base, offset); 312 oop result = bs->oop_cmpxchg(resolved_decorators, addr, compare_value, new_value); 313 if (ShenandoahCardBarrier) { 314 bs->write_ref_field_post<decorators>(addr); 315 } 316 return result; 317 } 318 319 template <DecoratorSet decorators, typename BarrierSetT> 320 template <typename T> 321 inline oop ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_atomic_xchg_not_in_heap(T* addr, oop new_value) { 322 assert((decorators & (AS_NO_KEEPALIVE | ON_UNKNOWN_OOP_REF)) == 0, "must be absent"); 323 ShenandoahBarrierSet* bs = ShenandoahBarrierSet::barrier_set(); 324 return bs->oop_xchg(decorators, addr, new_value); 325 } 326 327 template <DecoratorSet decorators, typename BarrierSetT> 328 template <typename T> 329 inline oop ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_atomic_xchg_in_heap(T* addr, oop new_value) { 330 assert((decorators & (AS_NO_KEEPALIVE | ON_UNKNOWN_OOP_REF)) == 0, "must be absent"); 331 ShenandoahBarrierSet* bs = ShenandoahBarrierSet::barrier_set(); 332 oop result = bs->oop_xchg(decorators, addr, new_value); 333 if (ShenandoahCardBarrier) { 334 bs->write_ref_field_post<decorators>(addr); 335 } 336 return result; 337 } 338 339 template <DecoratorSet decorators, typename BarrierSetT> 340 inline oop ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_atomic_xchg_in_heap_at(oop base, ptrdiff_t offset, oop new_value) { 341 assert((decorators & AS_NO_KEEPALIVE) == 0, "must be absent"); 342 ShenandoahBarrierSet* bs = ShenandoahBarrierSet::barrier_set(); 343 DecoratorSet resolved_decorators = AccessBarrierSupport::resolve_possibly_unknown_oop_ref_strength<decorators>(base, offset); 344 auto addr = AccessInternal::oop_field_addr<decorators>(base, offset); 345 oop result = bs->oop_xchg(resolved_decorators, addr, new_value); 346 if (ShenandoahCardBarrier) { 347 bs->write_ref_field_post<decorators>(addr); 348 } 349 return result; 350 } 351 352 // Clone barrier support 353 template <DecoratorSet decorators, typename BarrierSetT> 354 void ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::clone_in_heap(oop src, oop dst, size_t size) { 355 if (ShenandoahCloneBarrier) { 356 ShenandoahBarrierSet::barrier_set()->clone_barrier_runtime(src); 357 } 358 Raw::clone(src, dst, size); 359 } 360 361 template <DecoratorSet decorators, typename BarrierSetT> 362 template <typename T> 363 bool ShenandoahBarrierSet::AccessBarrier<decorators, BarrierSetT>::oop_arraycopy_in_heap(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw, 364 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw, 365 size_t length) { 366 T* src = arrayOopDesc::obj_offset_to_raw(src_obj, src_offset_in_bytes, src_raw); 367 T* dst = arrayOopDesc::obj_offset_to_raw(dst_obj, dst_offset_in_bytes, dst_raw); 368 369 ShenandoahBarrierSet* bs = ShenandoahBarrierSet::barrier_set(); 370 bs->arraycopy_barrier(src, dst, length); 371 bool result = Raw::oop_arraycopy_in_heap(src_obj, src_offset_in_bytes, src_raw, dst_obj, dst_offset_in_bytes, dst_raw, length); 372 if (ShenandoahCardBarrier) { 373 bs->write_ref_array((HeapWord*) dst, length); 374 } 375 return result; 376 } 377 378 template <class T, bool HAS_FWD, bool EVAC, bool ENQUEUE> 379 void ShenandoahBarrierSet::arraycopy_work(T* src, size_t count) { 380 // Young cycles are allowed to run when old marking is in progress. When old marking is in progress, 381 // this barrier will be called with ENQUEUE=true and HAS_FWD=false, even though the young generation 382 // may have forwarded objects. In this case, the `arraycopy_work` is first called with HAS_FWD=true and 383 // ENQUEUE=false. 384 assert(HAS_FWD == _heap->has_forwarded_objects() || (_heap->gc_state() & ShenandoahHeap::OLD_MARKING) != 0, 385 "Forwarded object status is sane"); 386 // This function cannot be called to handle marking and evacuation at the same time (they operate on 387 // different sides of the copy). 388 assert((HAS_FWD || EVAC) != ENQUEUE, "Cannot evacuate and mark both sides of copy."); 389 390 Thread* thread = Thread::current(); 391 SATBMarkQueue& queue = ShenandoahThreadLocalData::satb_mark_queue(thread); 392 ShenandoahMarkingContext* ctx = _heap->marking_context(); 393 const ShenandoahCollectionSet* const cset = _heap->collection_set(); 394 T* end = src + count; 395 for (T* elem_ptr = src; elem_ptr < end; elem_ptr++) { 396 T o = RawAccess<>::oop_load(elem_ptr); 397 if (!CompressedOops::is_null(o)) { 398 oop obj = CompressedOops::decode_not_null(o); 399 if (HAS_FWD && cset->is_in(obj)) { 400 oop fwd = resolve_forwarded_not_null(obj); 401 if (EVAC && obj == fwd) { 402 fwd = _heap->evacuate_object(obj, thread); 403 } 404 shenandoah_assert_forwarded_except(elem_ptr, obj, _heap->cancelled_gc()); 405 ShenandoahHeap::atomic_update_oop(fwd, elem_ptr, o); 406 } 407 if (ENQUEUE && !ctx->is_marked_strong_or_old(obj)) { 408 _satb_mark_queue_set.enqueue_known_active(queue, obj); 409 } 410 } 411 } 412 } 413 414 template <class T> 415 void ShenandoahBarrierSet::arraycopy_barrier(T* src, T* dst, size_t count) { 416 if (count == 0) { 417 // No elements to copy, no need for barrier 418 return; 419 } 420 421 int gc_state = _heap->gc_state(); 422 if ((gc_state & ShenandoahHeap::EVACUATION) != 0) { 423 arraycopy_evacuation(src, count); 424 } else if ((gc_state & ShenandoahHeap::UPDATEREFS) != 0) { 425 arraycopy_update(src, count); 426 } 427 428 if (_heap->mode()->is_generational()) { 429 assert(ShenandoahSATBBarrier, "Generational mode assumes SATB mode"); 430 if ((gc_state & ShenandoahHeap::YOUNG_MARKING) != 0) { 431 arraycopy_marking(src, dst, count, false); 432 } 433 if ((gc_state & ShenandoahHeap::OLD_MARKING) != 0) { 434 arraycopy_marking(src, dst, count, true); 435 } 436 } else if ((gc_state & ShenandoahHeap::MARKING) != 0) { 437 arraycopy_marking(src, dst, count, false); 438 } 439 } 440 441 template <class T> 442 void ShenandoahBarrierSet::arraycopy_marking(T* src, T* dst, size_t count, bool is_old_marking) { 443 assert(_heap->is_concurrent_mark_in_progress(), "only during marking"); 444 /* 445 * Note that an old-gen object is considered live if it is live at the start of OLD marking or if it is promoted 446 * following the start of OLD marking. 447 * 448 * 1. Every object promoted following the start of OLD marking will be above TAMS within its old-gen region 449 * 2. Every object live at the start of OLD marking will be referenced from a "root" or it will be referenced from 450 * another live OLD-gen object. With regards to old-gen, roots include stack locations and all of live young-gen. 451 * All root references to old-gen are identified during a bootstrap young collection. All references from other 452 * old-gen objects will be marked during the traversal of all old objects, or will be marked by the SATB barrier. 453 * 454 * During old-gen marking (which is interleaved with young-gen collections), call arraycopy_work() if: 455 * 456 * 1. The overwritten array resides in old-gen and it is below TAMS within its old-gen region 457 * 2. Do not call arraycopy_work for any array residing in young-gen because young-gen collection is idle at this time 458 * 459 * During young-gen marking, call arraycopy_work() if: 460 * 461 * 1. The overwritten array resides in young-gen and is below TAMS within its young-gen region 462 * 2. Additionally, if array resides in old-gen, regardless of its relationship to TAMS because this old-gen array 463 * may hold references to young-gen 464 */ 465 if (ShenandoahSATBBarrier) { 466 T* array = dst; 467 HeapWord* array_addr = reinterpret_cast<HeapWord*>(array); 468 ShenandoahHeapRegion* r = _heap->heap_region_containing(array_addr); 469 if (is_old_marking) { 470 // Generational, old marking 471 assert(_heap->mode()->is_generational(), "Invariant"); 472 if (r->is_old() && (array_addr < _heap->marking_context()->top_at_mark_start(r))) { 473 arraycopy_work<T, false, false, true>(array, count); 474 } 475 } else if (_heap->mode()->is_generational()) { 476 // Generational, young marking 477 if (r->is_old() || (array_addr < _heap->marking_context()->top_at_mark_start(r))) { 478 arraycopy_work<T, false, false, true>(array, count); 479 } 480 } else if (array_addr < _heap->marking_context()->top_at_mark_start(r)) { 481 // Non-generational, marking 482 arraycopy_work<T, false, false, true>(array, count); 483 } 484 } 485 } 486 487 inline bool ShenandoahBarrierSet::need_bulk_update(HeapWord* ary) { 488 return ary < _heap->heap_region_containing(ary)->get_update_watermark(); 489 } 490 491 template <class T> 492 void ShenandoahBarrierSet::arraycopy_evacuation(T* src, size_t count) { 493 assert(_heap->is_evacuation_in_progress(), "only during evacuation"); 494 if (need_bulk_update(reinterpret_cast<HeapWord*>(src))) { 495 ShenandoahEvacOOMScope oom_evac; 496 arraycopy_work<T, true, true, false>(src, count); 497 } 498 } 499 500 template <class T> 501 void ShenandoahBarrierSet::arraycopy_update(T* src, size_t count) { 502 assert(_heap->is_update_refs_in_progress(), "only during update-refs"); 503 if (need_bulk_update(reinterpret_cast<HeapWord*>(src))) { 504 arraycopy_work<T, true, false, false>(src, count); 505 } 506 } 507 508 #endif // SHARE_GC_SHENANDOAH_SHENANDOAHBARRIERSET_INLINE_HPP --- EOF ---