1 /* 2 * Copyright (c) 2015, 2022, Red Hat, Inc. All rights reserved. 3 * Copyright Amazon.com Inc. or its affiliates. All Rights Reserved. 4 * Copyright (c) 2025, Oracle and/or its affiliates. All rights reserved. 5 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 6 * 7 * This code is free software; you can redistribute it and/or modify it 8 * under the terms of the GNU General Public License version 2 only, as 9 * published by the Free Software Foundation. 10 * 11 * This code is distributed in the hope that it will be useful, but WITHOUT 12 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 13 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 14 * version 2 for more details (a copy is included in the LICENSE file that 15 * accompanied this code). 16 * 17 * You should have received a copy of the GNU General Public License version 18 * 2 along with this work; if not, write to the Free Software Foundation, 19 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 20 * 21 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 22 * or visit www.oracle.com if you need additional information or have any 23 * questions. 24 * 25 */ 26 27 #ifndef SHARE_GC_SHENANDOAH_SHENANDOAHMARK_INLINE_HPP 28 #define SHARE_GC_SHENANDOAH_SHENANDOAHMARK_INLINE_HPP 29 30 #include "gc/shenandoah/shenandoahMark.hpp" 31 32 #include "gc/shared/continuationGCSupport.inline.hpp" 33 #include "gc/shenandoah/shenandoahAgeCensus.hpp" 34 #include "gc/shenandoah/shenandoahAsserts.hpp" 35 #include "gc/shenandoah/shenandoahBarrierSet.inline.hpp" 36 #include "gc/shenandoah/shenandoahClosures.inline.hpp" 37 #include "gc/shenandoah/shenandoahHeap.inline.hpp" 38 #include "gc/shenandoah/shenandoahMarkingContext.inline.hpp" 39 #include "gc/shenandoah/shenandoahOldGeneration.hpp" 40 #include "gc/shenandoah/shenandoahScanRemembered.inline.hpp" 41 #include "gc/shenandoah/shenandoahStringDedup.inline.hpp" 42 #include "gc/shenandoah/shenandoahTaskqueue.inline.hpp" 43 #include "gc/shenandoah/shenandoahUtils.hpp" 44 #include "memory/iterator.inline.hpp" 45 #include "oops/compressedOops.inline.hpp" 46 #include "oops/oop.inline.hpp" 47 #include "runtime/prefetch.inline.hpp" 48 #include "utilities/devirtualizer.inline.hpp" 49 #include "utilities/powerOfTwo.hpp" 50 51 template <StringDedupMode STRING_DEDUP> 52 void ShenandoahMark::dedup_string(oop obj, StringDedup::Requests* const req) { 53 if (STRING_DEDUP == ENQUEUE_DEDUP) { 54 if (ShenandoahStringDedup::is_candidate(obj)) { 55 req->add(obj); 56 } 57 } else if (STRING_DEDUP == ALWAYS_DEDUP) { 58 if (ShenandoahStringDedup::is_string_candidate(obj) && 59 !ShenandoahStringDedup::dedup_requested(obj)) { 60 req->add(obj); 61 } 62 } 63 } 64 65 template <class T, ShenandoahGenerationType GENERATION, StringDedupMode STRING_DEDUP> 66 void ShenandoahMark::do_task(ShenandoahObjToScanQueue* q, T* cl, ShenandoahLiveData* live_data, StringDedup::Requests* const req, ShenandoahMarkTask* task, uint worker_id) { 67 oop obj = task->obj(); 68 69 shenandoah_assert_not_forwarded(nullptr, obj); 70 shenandoah_assert_marked(nullptr, obj); 71 shenandoah_assert_not_in_cset_except(nullptr, obj, ShenandoahHeap::heap()->cancelled_gc()); 72 73 // Are we in weak subgraph scan? 74 bool weak = task->is_weak(); 75 cl->set_weak(weak); 76 77 if (task->is_not_chunked()) { 78 if (obj->is_instance()) { 79 // Case 1: Normal oop, process as usual. 80 if (ContinuationGCSupport::relativize_stack_chunk(obj)) { 81 // Loom doesn't support mixing of weak marking and strong marking of 82 // stack chunks. 83 cl->set_weak(false); 84 } 85 86 obj->oop_iterate(cl); 87 dedup_string<STRING_DEDUP>(obj, req); 88 } else if (obj->is_objArray()) { 89 // Case 2: Object array instance and no chunk is set. Must be the first 90 // time we visit it, start the chunked processing. 91 do_chunked_array_start<T>(q, cl, obj, weak); 92 } else { 93 // Case 3: Primitive array. Do nothing, no oops there. We use the same 94 // performance tweak TypeArrayKlass::oop_oop_iterate_impl is using: 95 // We skip iterating over the klass pointer since we know that 96 // Universe::TypeArrayKlass never moves. 97 assert (obj->is_typeArray(), "should be type array"); 98 } 99 // Count liveness the last: push the outstanding work to the queues first 100 // Avoid double-counting objects that are visited twice due to upgrade 101 // from final- to strong mark. 102 if (task->count_liveness()) { 103 count_liveness<GENERATION>(live_data, obj, worker_id); 104 } 105 } else { 106 // Case 4: Array chunk, has sensible chunk id. Process it. 107 do_chunked_array<T>(q, cl, obj, task->chunk(), task->pow(), weak); 108 } 109 } 110 111 template <ShenandoahGenerationType GENERATION> 112 inline void ShenandoahMark::count_liveness(ShenandoahLiveData* live_data, oop obj, uint worker_id) { 113 const ShenandoahHeap* const heap = ShenandoahHeap::heap(); 114 const size_t region_idx = heap->heap_region_index_containing(obj); 115 ShenandoahHeapRegion* const region = heap->get_region(region_idx); 116 const size_t size = obj->size(); 117 118 // Age census for objects in the young generation 119 if (GENERATION == YOUNG || (GENERATION == GLOBAL && region->is_young())) { 120 assert(heap->mode()->is_generational(), "Only if generational"); 121 if (ShenandoahGenerationalAdaptiveTenuring && !ShenandoahGenerationalCensusAtEvac) { 122 assert(region->is_young(), "Only for young objects"); 123 uint age = ShenandoahHeap::get_object_age(obj); 124 ShenandoahAgeCensus* const census = ShenandoahGenerationalHeap::heap()->age_census(); 125 CENSUS_NOISE(census->add(age, region->age(), region->youth(), size, worker_id);) 126 NO_CENSUS_NOISE(census->add(age, region->age(), size, worker_id);) 127 } 128 } 129 130 if (!region->is_humongous_start()) { 131 assert(!region->is_humongous(), "Cannot have continuations here"); 132 assert(region->is_affiliated(), "Do not count live data within Free Regular Region %zu", region_idx); 133 ShenandoahLiveData cur = live_data[region_idx]; 134 size_t new_val = size + cur; 135 if (new_val >= SHENANDOAH_LIVEDATA_MAX) { 136 // overflow, flush to region data 137 region->increase_live_data_gc_words(new_val); 138 live_data[region_idx] = 0; 139 } else { 140 // still good, remember in locals 141 live_data[region_idx] = (ShenandoahLiveData) new_val; 142 } 143 } else { 144 shenandoah_assert_in_correct_region(nullptr, obj); 145 size_t num_regions = ShenandoahHeapRegion::required_regions(size * HeapWordSize); 146 147 assert(region->is_affiliated(), "Do not count live data within FREE Humongous Start Region %zu", region_idx); 148 for (size_t i = region_idx; i < region_idx + num_regions; i++) { 149 ShenandoahHeapRegion* chain_reg = heap->get_region(i); 150 assert(chain_reg->is_humongous(), "Expecting a humongous region"); 151 assert(chain_reg->is_affiliated(), "Do not count live data within FREE Humongous Continuation Region %zu", i); 152 chain_reg->increase_live_data_gc_words(chain_reg->used() >> LogHeapWordSize); 153 } 154 } 155 } 156 157 template <class T> 158 inline void ShenandoahMark::do_chunked_array_start(ShenandoahObjToScanQueue* q, T* cl, oop obj, bool weak) { 159 assert(obj->is_objArray(), "expect object array"); 160 objArrayOop array = objArrayOop(obj); 161 int len = array->length(); 162 163 // Mark objArray klass metadata 164 if (Devirtualizer::do_metadata(cl)) { 165 Devirtualizer::do_klass(cl, array->klass()); 166 } 167 168 if (len <= (int) ObjArrayMarkingStride*2) { 169 // A few slices only, process directly 170 array->oop_iterate_range(cl, 0, len); 171 } else { 172 int bits = log2i_graceful(len); 173 // Compensate for non-power-of-two arrays, cover the array in excess: 174 if (len != (1 << bits)) bits++; 175 176 // Only allow full chunks on the queue. This frees do_chunked_array() from checking from/to 177 // boundaries against array->length(), touching the array header on every chunk. 178 // 179 // To do this, we cut the prefix in full-sized chunks, and submit them on the queue. 180 // If the array is not divided in chunk sizes, then there would be an irregular tail, 181 // which we will process separately. 182 183 int last_idx = 0; 184 185 int chunk = 1; 186 int pow = bits; 187 188 // Handle overflow 189 if (pow >= 31) { 190 assert (pow == 31, "sanity"); 191 pow--; 192 chunk = 2; 193 last_idx = (1 << pow); 194 bool pushed = q->push(ShenandoahMarkTask(array, true, weak, 1, pow)); 195 assert(pushed, "overflow queue should always succeed pushing"); 196 } 197 198 // Split out tasks, as suggested in ShenandoahMarkTask docs. Record the last 199 // successful right boundary to figure out the irregular tail. 200 while ((1 << pow) > (int)ObjArrayMarkingStride && 201 (chunk*2 < ShenandoahMarkTask::chunk_size())) { 202 pow--; 203 int left_chunk = chunk*2 - 1; 204 int right_chunk = chunk*2; 205 int left_chunk_end = left_chunk * (1 << pow); 206 if (left_chunk_end < len) { 207 bool pushed = q->push(ShenandoahMarkTask(array, true, weak, left_chunk, pow)); 208 assert(pushed, "overflow queue should always succeed pushing"); 209 chunk = right_chunk; 210 last_idx = left_chunk_end; 211 } else { 212 chunk = left_chunk; 213 } 214 } 215 216 // Process the irregular tail, if present 217 int from = last_idx; 218 if (from < len) { 219 array->oop_iterate_range(cl, from, len); 220 } 221 } 222 } 223 224 template <class T> 225 inline void ShenandoahMark::do_chunked_array(ShenandoahObjToScanQueue* q, T* cl, oop obj, int chunk, int pow, bool weak) { 226 assert(obj->is_objArray(), "expect object array"); 227 objArrayOop array = objArrayOop(obj); 228 229 assert (ObjArrayMarkingStride > 0, "sanity"); 230 231 // Split out tasks, as suggested in ShenandoahMarkTask docs. Avoid pushing tasks that 232 // are known to start beyond the array. 233 while ((1 << pow) > (int)ObjArrayMarkingStride && (chunk*2 < ShenandoahMarkTask::chunk_size())) { 234 pow--; 235 chunk *= 2; 236 bool pushed = q->push(ShenandoahMarkTask(array, true, weak, chunk - 1, pow)); 237 assert(pushed, "overflow queue should always succeed pushing"); 238 } 239 240 int chunk_size = 1 << pow; 241 242 int from = (chunk - 1) * chunk_size; 243 int to = chunk * chunk_size; 244 245 #ifdef ASSERT 246 int len = array->length(); 247 assert (0 <= from && from < len, "from is sane: %d/%d", from, len); 248 assert (0 < to && to <= len, "to is sane: %d/%d", to, len); 249 #endif 250 251 array->oop_iterate_range(cl, from, to); 252 } 253 254 template <ShenandoahGenerationType GENERATION> 255 class ShenandoahSATBBufferClosure : public SATBBufferClosure { 256 private: 257 ShenandoahObjToScanQueue* _queue; 258 ShenandoahObjToScanQueue* _old_queue; 259 ShenandoahHeap* _heap; 260 ShenandoahMarkingContext* const _mark_context; 261 public: 262 ShenandoahSATBBufferClosure(ShenandoahObjToScanQueue* q, ShenandoahObjToScanQueue* old_q) : 263 _queue(q), 264 _old_queue(old_q), 265 _heap(ShenandoahHeap::heap()), 266 _mark_context(_heap->marking_context()) 267 { 268 } 269 270 void do_buffer(void **buffer, size_t size) { 271 assert(size == 0 || !_heap->has_forwarded_objects() || _heap->is_concurrent_old_mark_in_progress(), "Forwarded objects are not expected here"); 272 for (size_t i = 0; i < size; ++i) { 273 oop *p = (oop *) &buffer[i]; 274 ShenandoahMark::mark_through_ref<oop, GENERATION>(p, _queue, _old_queue, _mark_context, false); 275 } 276 } 277 }; 278 279 template<ShenandoahGenerationType GENERATION> 280 bool ShenandoahMark::in_generation(ShenandoahHeap* const heap, oop obj) { 281 // Each in-line expansion of in_generation() resolves GENERATION at compile time. 282 if (GENERATION == YOUNG) { 283 return heap->is_in_young(obj); 284 } 285 286 if (GENERATION == OLD) { 287 return heap->is_in_old(obj); 288 } 289 290 assert((GENERATION == GLOBAL || GENERATION == NON_GEN), "Unexpected generation type"); 291 assert(heap->is_in(obj), "Object must be in heap"); 292 return true; 293 } 294 295 template<class T, ShenandoahGenerationType GENERATION> 296 inline void ShenandoahMark::mark_through_ref(T *p, ShenandoahObjToScanQueue* q, ShenandoahObjToScanQueue* old_q, ShenandoahMarkingContext* const mark_context, bool weak) { 297 // Note: This is a very hot code path, so the code should be conditional on GENERATION template 298 // parameter where possible, in order to generate the most efficient code. 299 300 T o = RawAccess<>::oop_load(p); 301 if (!CompressedOops::is_null(o)) { 302 oop obj = CompressedOops::decode_not_null(o); 303 304 ShenandoahGenerationalHeap* heap = ShenandoahGenerationalHeap::heap(); 305 shenandoah_assert_not_forwarded(p, obj); 306 shenandoah_assert_not_in_cset_except(p, obj, heap->cancelled_gc()); 307 if (in_generation<GENERATION>(heap, obj)) { 308 mark_ref(q, mark_context, weak, obj); 309 shenandoah_assert_marked(p, obj); 310 if (GENERATION == YOUNG && heap->is_in_old(p)) { 311 // Mark card as dirty because remembered set scanning still finds interesting pointer. 312 heap->old_generation()->mark_card_as_dirty((HeapWord*)p); 313 } else if (GENERATION == GLOBAL && heap->is_in_old(p) && heap->is_in_young(obj)) { 314 // Mark card as dirty because GLOBAL marking finds interesting pointer. 315 heap->old_generation()->mark_card_as_dirty((HeapWord*)p); 316 } 317 } else if (old_q != nullptr) { 318 // Young mark, bootstrapping old_q or concurrent with old_q marking. 319 mark_ref(old_q, mark_context, weak, obj); 320 shenandoah_assert_marked(p, obj); 321 } else if (GENERATION == OLD) { 322 // Old mark, found a young pointer. 323 if (heap->is_in(p)) { 324 assert(heap->is_in_young(obj), "Expected young object."); 325 heap->old_generation()->mark_card_as_dirty(p); 326 } 327 } 328 } 329 } 330 331 template<> 332 inline void ShenandoahMark::mark_through_ref<oop, ShenandoahGenerationType::NON_GEN>(oop *p, ShenandoahObjToScanQueue* q, ShenandoahObjToScanQueue* old_q, ShenandoahMarkingContext* const mark_context, bool weak) { 333 mark_non_generational_ref(p, q, mark_context, weak); 334 } 335 336 template<> 337 inline void ShenandoahMark::mark_through_ref<narrowOop, ShenandoahGenerationType::NON_GEN>(narrowOop *p, ShenandoahObjToScanQueue* q, ShenandoahObjToScanQueue* old_q, ShenandoahMarkingContext* const mark_context, bool weak) { 338 mark_non_generational_ref(p, q, mark_context, weak); 339 } 340 341 template<class T> 342 inline void ShenandoahMark::mark_non_generational_ref(T* p, ShenandoahObjToScanQueue* q, 343 ShenandoahMarkingContext* const mark_context, bool weak) { 344 oop o = RawAccess<>::oop_load(p); 345 if (!CompressedOops::is_null(o)) { 346 oop obj = CompressedOops::decode_not_null(o); 347 348 shenandoah_assert_not_forwarded(p, obj); 349 shenandoah_assert_not_in_cset_except(p, obj, ShenandoahHeap::heap()->cancelled_gc()); 350 351 mark_ref(q, mark_context, weak, obj); 352 353 shenandoah_assert_marked(p, obj); 354 } 355 } 356 357 inline void ShenandoahMark::mark_ref(ShenandoahObjToScanQueue* q, 358 ShenandoahMarkingContext* const mark_context, 359 bool weak, oop obj) { 360 bool skip_live = false; 361 bool marked; 362 if (weak) { 363 marked = mark_context->mark_weak(obj); 364 } else { 365 marked = mark_context->mark_strong(obj, /* was_upgraded = */ skip_live); 366 } 367 if (marked) { 368 bool pushed = q->push(ShenandoahMarkTask(obj, skip_live, weak)); 369 assert(pushed, "overflow queue should always succeed pushing"); 370 } 371 } 372 373 ShenandoahObjToScanQueueSet* ShenandoahMark::task_queues() const { 374 return _task_queues; 375 } 376 377 ShenandoahObjToScanQueue* ShenandoahMark::get_queue(uint index) const { 378 return _task_queues->queue(index); 379 } 380 381 ShenandoahObjToScanQueue* ShenandoahMark::get_old_queue(uint index) const { 382 if (_old_gen_task_queues != nullptr) { 383 return _old_gen_task_queues->queue(index); 384 } 385 return nullptr; 386 } 387 388 #endif // SHARE_GC_SHENANDOAH_SHENANDOAHMARK_INLINE_HPP