1 /*
2 * Copyright Amazon.com Inc. or its affiliates. All Rights Reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
24
25
26 #include "precompiled.hpp"
27
28 #include "gc/shenandoah/heuristics/shenandoahOldHeuristics.hpp"
29 #include "gc/shenandoah/shenandoahAsserts.hpp"
30 #include "gc/shenandoah/shenandoahCardTable.hpp"
31 #include "gc/shenandoah/shenandoahCollectorPolicy.hpp"
32 #include "gc/shenandoah/shenandoahFreeSet.hpp"
33 #include "gc/shenandoah/shenandoahGenerationalHeap.hpp"
34 #include "gc/shenandoah/shenandoahHeap.hpp"
35 #include "gc/shenandoah/shenandoahHeap.inline.hpp"
36 #include "gc/shenandoah/shenandoahHeapRegion.hpp"
37 #include "gc/shenandoah/shenandoahHeapRegionClosures.hpp"
38 #include "gc/shenandoah/shenandoahMonitoringSupport.hpp"
39 #include "gc/shenandoah/shenandoahOldGeneration.hpp"
40 #include "gc/shenandoah/shenandoahOopClosures.inline.hpp"
41 #include "gc/shenandoah/shenandoahReferenceProcessor.hpp"
42 #include "gc/shenandoah/shenandoahScanRemembered.inline.hpp"
43 #include "gc/shenandoah/shenandoahUtils.hpp"
44 #include "gc/shenandoah/shenandoahWorkerPolicy.hpp"
45 #include "gc/shenandoah/shenandoahYoungGeneration.hpp"
46 #include "runtime/threads.hpp"
47 #include "utilities/events.hpp"
48
49 class ShenandoahFlushAllSATB : public ThreadClosure {
50 private:
51 SATBMarkQueueSet& _satb_qset;
52
53 public:
54 explicit ShenandoahFlushAllSATB(SATBMarkQueueSet& satb_qset) :
55 _satb_qset(satb_qset) {}
56
57 void do_thread(Thread* thread) override {
58 // Transfer any partial buffer to the qset for completed buffer processing.
59 _satb_qset.flush_queue(ShenandoahThreadLocalData::satb_mark_queue(thread));
60 }
61 };
62
63 class ShenandoahProcessOldSATB : public SATBBufferClosure {
64 private:
65 ShenandoahObjToScanQueue* _queue;
66 ShenandoahHeap* _heap;
67 ShenandoahMarkingContext* const _mark_context;
68 size_t _trashed_oops;
69
70 public:
71 explicit ShenandoahProcessOldSATB(ShenandoahObjToScanQueue* q) :
72 _queue(q),
73 _heap(ShenandoahHeap::heap()),
74 _mark_context(_heap->marking_context()),
75 _trashed_oops(0) {}
76
77 void do_buffer(void** buffer, size_t size) override {
78 assert(size == 0 || !_heap->has_forwarded_objects() || _heap->is_concurrent_old_mark_in_progress(), "Forwarded objects are not expected here");
79 for (size_t i = 0; i < size; ++i) {
80 oop *p = (oop *) &buffer[i];
81 ShenandoahHeapRegion* region = _heap->heap_region_containing(*p);
82 if (region->is_old() && region->is_active()) {
83 ShenandoahMark::mark_through_ref<oop, OLD>(p, _queue, nullptr, _mark_context, false);
84 } else {
85 _trashed_oops++;
86 }
87 }
88 }
89
90 size_t trashed_oops() const {
91 return _trashed_oops;
92 }
93 };
94
95 class ShenandoahPurgeSATBTask : public WorkerTask {
96 private:
97 ShenandoahObjToScanQueueSet* _mark_queues;
98 // Keep track of the number of oops that are not transferred to mark queues.
99 // This is volatile because workers update it, but the vm thread reads it.
100 volatile size_t _trashed_oops;
101
102 public:
103 explicit ShenandoahPurgeSATBTask(ShenandoahObjToScanQueueSet* queues) :
104 WorkerTask("Purge SATB"),
105 _mark_queues(queues),
106 _trashed_oops(0) {
107 Threads::change_thread_claim_token();
108 }
109
110 ~ShenandoahPurgeSATBTask() {
111 if (_trashed_oops > 0) {
112 log_debug(gc)("Purged " SIZE_FORMAT " oops from old generation SATB buffers", _trashed_oops);
113 }
114 }
115
116 void work(uint worker_id) override {
117 ShenandoahParallelWorkerSession worker_session(worker_id);
118 ShenandoahSATBMarkQueueSet &satb_queues = ShenandoahBarrierSet::satb_mark_queue_set();
119 ShenandoahFlushAllSATB flusher(satb_queues);
120 Threads::possibly_parallel_threads_do(true /* is_par */, &flusher);
121
122 ShenandoahObjToScanQueue* mark_queue = _mark_queues->queue(worker_id);
123 ShenandoahProcessOldSATB processor(mark_queue);
124 while (satb_queues.apply_closure_to_completed_buffer(&processor)) {}
125
126 Atomic::add(&_trashed_oops, processor.trashed_oops());
127 }
128 };
129
130 class ShenandoahTransferOldSATBTask : public WorkerTask {
131 ShenandoahSATBMarkQueueSet& _satb_queues;
132 ShenandoahObjToScanQueueSet* _mark_queues;
133 // Keep track of the number of oops that are not transferred to mark queues.
134 // This is volatile because workers update it, but the control thread reads it.
135 volatile size_t _trashed_oops;
136
137 public:
138 explicit ShenandoahTransferOldSATBTask(ShenandoahSATBMarkQueueSet& satb_queues, ShenandoahObjToScanQueueSet* mark_queues) :
139 WorkerTask("Transfer SATB"),
140 _satb_queues(satb_queues),
141 _mark_queues(mark_queues),
142 _trashed_oops(0) {}
143
144 ~ShenandoahTransferOldSATBTask() {
145 if (_trashed_oops > 0) {
146 log_debug(gc)("Purged %zu oops from old generation SATB buffers", _trashed_oops);
147 }
148 }
149
150 void work(uint worker_id) override {
151 ShenandoahObjToScanQueue* mark_queue = _mark_queues->queue(worker_id);
152 ShenandoahProcessOldSATB processor(mark_queue);
153 while (_satb_queues.apply_closure_to_completed_buffer(&processor)) {}
154
155 Atomic::add(&_trashed_oops, processor.trashed_oops());
156 }
157 };
158
159 class ShenandoahConcurrentCoalesceAndFillTask : public WorkerTask {
160 private:
161 uint _nworkers;
162 ShenandoahHeapRegion** _coalesce_and_fill_region_array;
163 uint _coalesce_and_fill_region_count;
164 volatile bool _is_preempted;
165
166 public:
167 ShenandoahConcurrentCoalesceAndFillTask(uint nworkers,
168 ShenandoahHeapRegion** coalesce_and_fill_region_array,
169 uint region_count) :
170 WorkerTask("Shenandoah Concurrent Coalesce and Fill"),
171 _nworkers(nworkers),
172 _coalesce_and_fill_region_array(coalesce_and_fill_region_array),
173 _coalesce_and_fill_region_count(region_count),
174 _is_preempted(false) {
175 }
176
177 void work(uint worker_id) override {
178 ShenandoahWorkerTimingsTracker timer(ShenandoahPhaseTimings::conc_coalesce_and_fill, ShenandoahPhaseTimings::ScanClusters, worker_id);
179 for (uint region_idx = worker_id; region_idx < _coalesce_and_fill_region_count; region_idx += _nworkers) {
180 ShenandoahHeapRegion* r = _coalesce_and_fill_region_array[region_idx];
181 if (r->is_humongous()) {
182 // There is only one object in this region and it is not garbage,
183 // so no need to coalesce or fill.
184 continue;
185 }
186
187 if (!r->oop_coalesce_and_fill(true)) {
188 // Coalesce and fill has been preempted
189 Atomic::store(&_is_preempted, true);
190 return;
191 }
192 }
193 }
194
195 // Value returned from is_completed() is only valid after all worker thread have terminated.
196 bool is_completed() {
197 return !Atomic::load(&_is_preempted);
198 }
199 };
200
201 ShenandoahOldGeneration::ShenandoahOldGeneration(uint max_queues, size_t max_capacity)
202 : ShenandoahGeneration(OLD, max_queues, max_capacity),
203 _coalesce_and_fill_region_array(NEW_C_HEAP_ARRAY(ShenandoahHeapRegion*, ShenandoahHeap::heap()->num_regions(), mtGC)),
204 _old_heuristics(nullptr),
205 _region_balance(0),
206 _promoted_reserve(0),
207 _promoted_expended(0),
208 _promotion_potential(0),
209 _pad_for_promote_in_place(0),
210 _promotable_humongous_regions(0),
211 _promotable_regular_regions(0),
212 _is_parsable(true),
213 _card_scan(nullptr),
214 _state(WAITING_FOR_BOOTSTRAP),
215 _growth_before_compaction(INITIAL_GROWTH_BEFORE_COMPACTION),
216 _min_growth_before_compaction ((ShenandoahMinOldGenGrowthPercent * FRACTIONAL_DENOMINATOR) / 100)
217 {
218 _live_bytes_after_last_mark = ShenandoahHeap::heap()->capacity() * INITIAL_LIVE_FRACTION / FRACTIONAL_DENOMINATOR;
219 // Always clear references for old generation
220 ref_processor()->set_soft_reference_policy(true);
221
222 if (ShenandoahCardBarrier) {
223 ShenandoahCardTable* card_table = ShenandoahBarrierSet::barrier_set()->card_table();
224 size_t card_count = card_table->cards_required(ShenandoahHeap::heap()->reserved_region().word_size());
225 auto rs = new ShenandoahDirectCardMarkRememberedSet(card_table, card_count);
226 _card_scan = new ShenandoahScanRemembered(rs);
227 }
228 }
229
230 void ShenandoahOldGeneration::set_promoted_reserve(size_t new_val) {
231 shenandoah_assert_heaplocked_or_safepoint();
232 _promoted_reserve = new_val;
233 }
234
235 size_t ShenandoahOldGeneration::get_promoted_reserve() const {
236 return _promoted_reserve;
237 }
238
239 void ShenandoahOldGeneration::augment_promoted_reserve(size_t increment) {
240 shenandoah_assert_heaplocked_or_safepoint();
241 _promoted_reserve += increment;
242 }
243
244 void ShenandoahOldGeneration::reset_promoted_expended() {
245 shenandoah_assert_heaplocked_or_safepoint();
246 Atomic::store(&_promoted_expended, (size_t) 0);
247 }
248
249 size_t ShenandoahOldGeneration::expend_promoted(size_t increment) {
250 shenandoah_assert_heaplocked_or_safepoint();
251 assert(get_promoted_expended() + increment <= get_promoted_reserve(), "Do not expend more promotion than budgeted");
252 return Atomic::add(&_promoted_expended, increment);
253 }
254
255 size_t ShenandoahOldGeneration::unexpend_promoted(size_t decrement) {
256 return Atomic::sub(&_promoted_expended, decrement);
257 }
258
259 size_t ShenandoahOldGeneration::get_promoted_expended() const {
260 return Atomic::load(&_promoted_expended);
261 }
262
263 bool ShenandoahOldGeneration::can_allocate(const ShenandoahAllocRequest &req) const {
264 assert(req.type() != ShenandoahAllocRequest::_alloc_gclab, "GCLAB pertains only to young-gen memory");
265
266 const size_t requested_bytes = req.size() * HeapWordSize;
267 // The promotion reserve may also be used for evacuations. If we can promote this object,
268 // then we can also evacuate it.
269 if (can_promote(requested_bytes)) {
270 // The promotion reserve should be able to accommodate this request. The request
271 // might still fail if alignment with the card table increases the size. The request
272 // may also fail if the heap is badly fragmented and the free set cannot find room for it.
273 return true;
274 }
275
276 if (req.type() == ShenandoahAllocRequest::_alloc_plab) {
277 // The promotion reserve cannot accommodate this plab request. Check if we still have room for
278 // evacuations. Note that we cannot really know how much of the plab will be used for evacuations,
279 // so here we only check that some evacuation reserve still exists.
280 return get_evacuation_reserve() > 0;
281 }
282
283 // This is a shared allocation request. We've already checked that it can't be promoted, so if
284 // it is a promotion, we return false. Otherwise, it is a shared evacuation request, and we allow
285 // the allocation to proceed.
286 return !req.is_promotion();
287 }
288
289 void
290 ShenandoahOldGeneration::configure_plab_for_current_thread(const ShenandoahAllocRequest &req) {
291 // Note: Even when a mutator is performing a promotion outside a LAB, we use a 'shared_gc' request.
292 if (req.is_gc_alloc()) {
293 const size_t actual_size = req.actual_size() * HeapWordSize;
294 if (req.type() == ShenandoahAllocRequest::_alloc_plab) {
295 // We've created a new plab. Now we configure it whether it will be used for promotions
296 // and evacuations - or just evacuations.
297 Thread* thread = Thread::current();
298 ShenandoahThreadLocalData::reset_plab_promoted(thread);
299
300 // The actual size of the allocation may be larger than the requested bytes (due to alignment on card boundaries).
301 // If this puts us over our promotion budget, we need to disable future PLAB promotions for this thread.
302 if (can_promote(actual_size)) {
303 // Assume the entirety of this PLAB will be used for promotion. This prevents promotion from overreach.
304 // When we retire this plab, we'll unexpend what we don't really use.
305 expend_promoted(actual_size);
306 ShenandoahThreadLocalData::enable_plab_promotions(thread);
307 ShenandoahThreadLocalData::set_plab_actual_size(thread, actual_size);
308 } else {
309 // Disable promotions in this thread because entirety of this PLAB must be available to hold old-gen evacuations.
310 ShenandoahThreadLocalData::disable_plab_promotions(thread);
311 ShenandoahThreadLocalData::set_plab_actual_size(thread, 0);
312 }
313 } else if (req.is_promotion()) {
314 // Shared promotion.
315 expend_promoted(actual_size);
316 }
317 }
318 }
319
320 size_t ShenandoahOldGeneration::get_live_bytes_after_last_mark() const {
321 return _live_bytes_after_last_mark;
322 }
323
324 void ShenandoahOldGeneration::set_live_bytes_after_last_mark(size_t bytes) {
325 if (bytes == 0) {
326 // Restart search for best old-gen size to the initial state
327 _live_bytes_after_last_mark = ShenandoahHeap::heap()->capacity() * INITIAL_LIVE_FRACTION / FRACTIONAL_DENOMINATOR;
328 _growth_before_compaction = INITIAL_GROWTH_BEFORE_COMPACTION;
329 } else {
330 _live_bytes_after_last_mark = bytes;
331 _growth_before_compaction /= 2;
332 if (_growth_before_compaction < _min_growth_before_compaction) {
333 _growth_before_compaction = _min_growth_before_compaction;
334 }
335 }
336 }
337
338 void ShenandoahOldGeneration::handle_failed_transfer() {
339 _old_heuristics->trigger_cannot_expand();
340 }
341
342 size_t ShenandoahOldGeneration::usage_trigger_threshold() const {
343 size_t result = _live_bytes_after_last_mark + (_live_bytes_after_last_mark * _growth_before_compaction) / FRACTIONAL_DENOMINATOR;
344 return result;
345 }
346
347 bool ShenandoahOldGeneration::contains(ShenandoahAffiliation affiliation) const {
348 return affiliation == OLD_GENERATION;
349 }
350 bool ShenandoahOldGeneration::contains(ShenandoahHeapRegion* region) const {
351 return region->is_old();
352 }
353
354 void ShenandoahOldGeneration::parallel_heap_region_iterate(ShenandoahHeapRegionClosure* cl) {
355 ShenandoahIncludeRegionClosure<OLD_GENERATION> old_regions_cl(cl);
356 ShenandoahHeap::heap()->parallel_heap_region_iterate(&old_regions_cl);
357 }
358
359 void ShenandoahOldGeneration::heap_region_iterate(ShenandoahHeapRegionClosure* cl) {
360 ShenandoahIncludeRegionClosure<OLD_GENERATION> old_regions_cl(cl);
361 ShenandoahHeap::heap()->heap_region_iterate(&old_regions_cl);
362 }
363
364 void ShenandoahOldGeneration::set_concurrent_mark_in_progress(bool in_progress) {
365 ShenandoahHeap::heap()->set_concurrent_old_mark_in_progress(in_progress);
366 }
367
368 bool ShenandoahOldGeneration::is_concurrent_mark_in_progress() {
369 return ShenandoahHeap::heap()->is_concurrent_old_mark_in_progress();
370 }
371
372 void ShenandoahOldGeneration::cancel_marking() {
373 if (is_concurrent_mark_in_progress()) {
374 log_debug(gc)("Abandon SATB buffers");
375 ShenandoahBarrierSet::satb_mark_queue_set().abandon_partial_marking();
376 }
377
378 ShenandoahGeneration::cancel_marking();
379 }
380
381 void ShenandoahOldGeneration::cancel_gc() {
382 shenandoah_assert_safepoint();
383 if (is_idle()) {
384 #ifdef ASSERT
385 validate_waiting_for_bootstrap();
386 #endif
387 } else {
388 log_info(gc)("Terminating old gc cycle.");
389 // Stop marking
390 cancel_marking();
391 // Stop tracking old regions
392 abandon_collection_candidates();
393 // Remove old generation access to young generation mark queues
394 ShenandoahHeap::heap()->young_generation()->set_old_gen_task_queues(nullptr);
395 // Transition to IDLE now.
396 transition_to(ShenandoahOldGeneration::WAITING_FOR_BOOTSTRAP);
397 }
398 }
399
400 void ShenandoahOldGeneration::prepare_gc() {
401 // Now that we have made the old generation parsable, it is safe to reset the mark bitmap.
402 assert(state() != FILLING, "Cannot reset old without making it parsable");
403
404 ShenandoahGeneration::prepare_gc();
405 }
406
407 bool ShenandoahOldGeneration::entry_coalesce_and_fill() {
408 ShenandoahHeap* const heap = ShenandoahHeap::heap();
409
410 static const char* msg = "Coalescing and filling (Old)";
411 ShenandoahConcurrentPhase gc_phase(msg, ShenandoahPhaseTimings::conc_coalesce_and_fill);
412
413 TraceCollectorStats tcs(heap->monitoring_support()->concurrent_collection_counters());
414 EventMark em("%s", msg);
415 ShenandoahWorkerScope scope(heap->workers(),
416 ShenandoahWorkerPolicy::calc_workers_for_conc_marking(),
417 msg);
418
419 return coalesce_and_fill();
420 }
421
422 // Make the old generation regions parsable, so they can be safely
423 // scanned when looking for objects in memory indicated by dirty cards.
424 bool ShenandoahOldGeneration::coalesce_and_fill() {
425 transition_to(FILLING);
426
427 // This code will see the same set of regions to fill on each resumption as it did
428 // on the initial run. That's okay because each region keeps track of its own coalesce
429 // and fill state. Regions that were filled on a prior attempt will not try to fill again.
430 uint coalesce_and_fill_regions_count = _old_heuristics->get_coalesce_and_fill_candidates(_coalesce_and_fill_region_array);
431 assert(coalesce_and_fill_regions_count <= ShenandoahHeap::heap()->num_regions(), "Sanity");
432 if (coalesce_and_fill_regions_count == 0) {
433 // No regions need to be filled.
434 abandon_collection_candidates();
435 return true;
436 }
437
438 ShenandoahHeap* const heap = ShenandoahHeap::heap();
439 WorkerThreads* workers = heap->workers();
440 uint nworkers = workers->active_workers();
441 ShenandoahConcurrentCoalesceAndFillTask task(nworkers, _coalesce_and_fill_region_array, coalesce_and_fill_regions_count);
442
443 log_debug(gc)("Starting (or resuming) coalesce-and-fill of " UINT32_FORMAT " old heap regions", coalesce_and_fill_regions_count);
444 workers->run_task(&task);
445 if (task.is_completed()) {
446 // We no longer need to track regions that need to be coalesced and filled.
447 abandon_collection_candidates();
448 return true;
449 } else {
450 // Coalesce-and-fill has been preempted. We'll finish that effort in the future. Do not invoke
451 // ShenandoahGeneration::prepare_gc() until coalesce-and-fill is done because it resets the mark bitmap
452 // and invokes set_mark_incomplete(). Coalesce-and-fill depends on the mark bitmap.
453 log_debug(gc)("Suspending coalesce-and-fill of old heap regions");
454 return false;
455 }
456 }
457
458 void ShenandoahOldGeneration::concurrent_transfer_pointers_from_satb() const {
459 const ShenandoahHeap* heap = ShenandoahHeap::heap();
460 assert(heap->is_concurrent_old_mark_in_progress(), "Only necessary during old marking.");
461 log_debug(gc)("Transfer SATB buffers");
462
463 // Step 1. All threads need to 'complete' partially filled, thread local SATB buffers. This
464 // is accomplished in ShenandoahConcurrentGC::complete_abbreviated_cycle using a Handshake
465 // operation.
466 // Step 2. Use worker threads to transfer oops from old, active regions in the completed
467 // SATB buffers to old generation mark queues.
468 ShenandoahSATBMarkQueueSet& satb_queues = ShenandoahBarrierSet::satb_mark_queue_set();
469 ShenandoahTransferOldSATBTask transfer_task(satb_queues, task_queues());
470 heap->workers()->run_task(&transfer_task);
471 }
472
473 void ShenandoahOldGeneration::transfer_pointers_from_satb() const {
474 const ShenandoahHeap* heap = ShenandoahHeap::heap();
475 assert(heap->is_concurrent_old_mark_in_progress(), "Only necessary during old marking.");
476 log_debug(gc)("Transfer SATB buffers");
477 ShenandoahPurgeSATBTask purge_satb_task(task_queues());
478 heap->workers()->run_task(&purge_satb_task);
479 }
480
481 bool ShenandoahOldGeneration::contains(oop obj) const {
482 return ShenandoahHeap::heap()->is_in_old(obj);
483 }
484
485 void ShenandoahOldGeneration::prepare_regions_and_collection_set(bool concurrent) {
486 ShenandoahHeap* heap = ShenandoahHeap::heap();
487 assert(!heap->is_full_gc_in_progress(), "Only for concurrent and degenerated GC");
488
489 {
490 ShenandoahGCPhase phase(concurrent ?
491 ShenandoahPhaseTimings::final_update_region_states :
492 ShenandoahPhaseTimings::degen_gc_final_update_region_states);
493 ShenandoahFinalMarkUpdateRegionStateClosure cl(complete_marking_context());
494
495 parallel_heap_region_iterate(&cl);
496 heap->assert_pinned_region_status();
497 }
498
499 {
500 // This doesn't actually choose a collection set, but prepares a list of
501 // regions as 'candidates' for inclusion in a mixed collection.
502 ShenandoahGCPhase phase(concurrent ?
503 ShenandoahPhaseTimings::choose_cset :
504 ShenandoahPhaseTimings::degen_gc_choose_cset);
505 ShenandoahHeapLocker locker(heap->lock());
506 _old_heuristics->prepare_for_old_collections();
507 }
508
509 {
510 // Though we did not choose a collection set above, we still may have
511 // freed up immediate garbage regions so proceed with rebuilding the free set.
512 ShenandoahGCPhase phase(concurrent ?
513 ShenandoahPhaseTimings::final_rebuild_freeset :
514 ShenandoahPhaseTimings::degen_gc_final_rebuild_freeset);
515 ShenandoahHeapLocker locker(heap->lock());
516 size_t cset_young_regions, cset_old_regions;
517 size_t first_old, last_old, num_old;
518 heap->free_set()->prepare_to_rebuild(cset_young_regions, cset_old_regions, first_old, last_old, num_old);
519 // This is just old-gen completion. No future budgeting required here. The only reason to rebuild the freeset here
520 // is in case there was any immediate old garbage identified.
521 heap->free_set()->finish_rebuild(cset_young_regions, cset_old_regions, num_old);
522 }
523 }
524
525 const char* ShenandoahOldGeneration::state_name(State state) {
526 switch (state) {
527 case WAITING_FOR_BOOTSTRAP: return "Waiting for Bootstrap";
528 case FILLING: return "Coalescing";
529 case BOOTSTRAPPING: return "Bootstrapping";
530 case MARKING: return "Marking";
531 case EVACUATING: return "Evacuating";
532 case EVACUATING_AFTER_GLOBAL: return "Evacuating (G)";
533 default:
534 ShouldNotReachHere();
535 return "Unknown";
536 }
537 }
538
539 void ShenandoahOldGeneration::transition_to(State new_state) {
540 if (_state != new_state) {
541 log_debug(gc, thread)("Old generation transition from %s to %s", state_name(_state), state_name(new_state));
542 EventMark event("Old was %s, now is %s", state_name(_state), state_name(new_state));
543 validate_transition(new_state);
544 _state = new_state;
545 }
546 }
547
548 #ifdef ASSERT
549 // This diagram depicts the expected state transitions for marking the old generation
550 // and preparing for old collections. When a young generation cycle executes, the
551 // remembered set scan must visit objects in old regions. Visiting an object which
552 // has become dead on previous old cycles will result in crashes. To avoid visiting
553 // such objects, the remembered set scan will use the old generation mark bitmap when
554 // possible. It is _not_ possible to use the old generation bitmap when old marking
555 // is active (bitmap is not complete). For this reason, the old regions are made
556 // parsable _before_ the old generation bitmap is reset. The diagram does not depict
557 // cancellation of old collections by global or full collections.
558 //
559 // When a global collection supersedes an old collection, the global mark still
560 // "completes" the old mark bitmap. Subsequent remembered set scans may use the
561 // old generation mark bitmap, but any uncollected old regions must still be made parsable
562 // before the next old generation cycle begins. For this reason, a global collection may
563 // create mixed collection candidates and coalesce and fill candidates and will put
564 // the old generation in the respective states (EVACUATING or FILLING). After a Full GC,
565 // the mark bitmaps are all reset, all regions are parsable and the mark context will
566 // not be "complete". After a Full GC, remembered set scans will _not_ use the mark bitmap
567 // and we expect the old generation to be waiting for bootstrap.
568 //
569 // +-----------------+
570 // +------------> | FILLING | <---+
571 // | +--------> | | |
572 // | | +-----------------+ |
573 // | | | |
574 // | | | Filling Complete | <-> A global collection may
575 // | | v | move the old generation
576 // | | +-----------------+ | directly from waiting for
577 // +-- |-- |--------> | WAITING | | bootstrap to filling or
578 // | | | +---- | FOR BOOTSTRAP | ----+ evacuating. It may also
579 // | | | | +-----------------+ move from filling to waiting
580 // | | | | | for bootstrap.
581 // | | | | | Reset Bitmap
582 // | | | | v
583 // | | | | +-----------------+ +----------------------+
584 // | | | | | BOOTSTRAP | <-> | YOUNG GC |
585 // | | | | | | | (RSet Parses Region) |
586 // | | | | +-----------------+ +----------------------+
587 // | | | | |
588 // | | | | | Old Marking
589 // | | | | v
590 // | | | | +-----------------+ +----------------------+
591 // | | | | | MARKING | <-> | YOUNG GC |
592 // | | +--------- | | | (RSet Parses Region) |
593 // | | | +-----------------+ +----------------------+
594 // | | | |
595 // | | | | Has Evacuation Candidates
596 // | | | v
597 // | | | +-----------------+ +--------------------+
598 // | | +---> | EVACUATING | <-> | YOUNG GC |
599 // | +------------- | | | (RSet Uses Bitmap) |
600 // | +-----------------+ +--------------------+
601 // | |
602 // | | Global Cycle Coalesces and Fills Old Regions
603 // | v
604 // | +-----------------+ +--------------------+
605 // +----------------- | EVACUATING | <-> | YOUNG GC |
606 // | AFTER GLOBAL | | (RSet Uses Bitmap) |
607 // +-----------------+ +--------------------+
608 //
609 //
610 void ShenandoahOldGeneration::validate_transition(State new_state) {
611 ShenandoahGenerationalHeap* heap = ShenandoahGenerationalHeap::heap();
612 switch (new_state) {
613 case FILLING:
614 assert(_state != BOOTSTRAPPING, "Cannot begin making old regions parsable after bootstrapping");
615 assert(is_mark_complete(), "Cannot begin filling without first completing marking, state is '%s'", state_name(_state));
616 assert(_old_heuristics->has_coalesce_and_fill_candidates(), "Cannot begin filling without something to fill.");
617 break;
618 case WAITING_FOR_BOOTSTRAP:
619 // GC cancellation can send us back here from any state.
620 validate_waiting_for_bootstrap();
621 break;
622 case BOOTSTRAPPING:
623 assert(_state == WAITING_FOR_BOOTSTRAP, "Cannot reset bitmap without making old regions parsable, state is '%s'", state_name(_state));
624 assert(_old_heuristics->unprocessed_old_collection_candidates() == 0, "Cannot bootstrap with mixed collection candidates");
625 assert(!heap->is_prepare_for_old_mark_in_progress(), "Cannot still be making old regions parsable.");
626 break;
627 case MARKING:
628 assert(_state == BOOTSTRAPPING, "Must have finished bootstrapping before marking, state is '%s'", state_name(_state));
629 assert(heap->young_generation()->old_gen_task_queues() != nullptr, "Young generation needs old mark queues.");
630 assert(heap->is_concurrent_old_mark_in_progress(), "Should be marking old now.");
631 break;
632 case EVACUATING_AFTER_GLOBAL:
633 assert(_state == EVACUATING, "Must have been evacuating, state is '%s'", state_name(_state));
634 break;
635 case EVACUATING:
636 assert(_state == WAITING_FOR_BOOTSTRAP || _state == MARKING, "Cannot have old collection candidates without first marking, state is '%s'", state_name(_state));
637 assert(_old_heuristics->unprocessed_old_collection_candidates() > 0, "Must have collection candidates here.");
638 break;
639 default:
640 fatal("Unknown new state");
641 }
642 }
643
644 bool ShenandoahOldGeneration::validate_waiting_for_bootstrap() {
645 ShenandoahHeap* heap = ShenandoahHeap::heap();
646 assert(!heap->is_concurrent_old_mark_in_progress(), "Cannot become ready for bootstrap during old mark.");
647 assert(heap->young_generation()->old_gen_task_queues() == nullptr, "Cannot become ready for bootstrap when still setup for bootstrapping.");
648 assert(!is_concurrent_mark_in_progress(), "Cannot be marking in IDLE");
649 assert(!heap->young_generation()->is_bootstrap_cycle(), "Cannot have old mark queues if IDLE");
650 assert(!_old_heuristics->has_coalesce_and_fill_candidates(), "Cannot have coalesce and fill candidates in IDLE");
651 assert(_old_heuristics->unprocessed_old_collection_candidates() == 0, "Cannot have mixed collection candidates in IDLE");
652 return true;
653 }
654 #endif
655
656 ShenandoahHeuristics* ShenandoahOldGeneration::initialize_heuristics(ShenandoahMode* gc_mode) {
657 _old_heuristics = new ShenandoahOldHeuristics(this, ShenandoahGenerationalHeap::heap());
658 _old_heuristics->set_guaranteed_gc_interval(ShenandoahGuaranteedOldGCInterval);
659 _heuristics = _old_heuristics;
660 return _heuristics;
661 }
662
663 void ShenandoahOldGeneration::record_success_concurrent(bool abbreviated) {
664 heuristics()->record_success_concurrent();
665 ShenandoahHeap::heap()->shenandoah_policy()->record_success_old();
666 }
667
668 void ShenandoahOldGeneration::handle_failed_evacuation() {
669 if (_failed_evacuation.try_set()) {
670 log_debug(gc)("Old gen evac failure.");
671 }
672 }
673
674 void ShenandoahOldGeneration::handle_failed_promotion(Thread* thread, size_t size) {
675 // We squelch excessive reports to reduce noise in logs.
676 const size_t MaxReportsPerEpoch = 4;
677 static size_t last_report_epoch = 0;
678 static size_t epoch_report_count = 0;
679 auto heap = ShenandoahGenerationalHeap::heap();
680
681 size_t promotion_reserve;
682 size_t promotion_expended;
683
684 const size_t gc_id = heap->control_thread()->get_gc_id();
685
686 if ((gc_id != last_report_epoch) || (epoch_report_count++ < MaxReportsPerEpoch)) {
687 {
688 // Promotion failures should be very rare. Invest in providing useful diagnostic info.
689 ShenandoahHeapLocker locker(heap->lock());
690 promotion_reserve = get_promoted_reserve();
691 promotion_expended = get_promoted_expended();
692 }
693 PLAB* const plab = ShenandoahThreadLocalData::plab(thread);
694 const size_t words_remaining = (plab == nullptr)? 0: plab->words_remaining();
695 const char* promote_enabled = ShenandoahThreadLocalData::allow_plab_promotions(thread)? "enabled": "disabled";
696
697 log_info(gc, ergo)("Promotion failed, size " SIZE_FORMAT ", has plab? %s, PLAB remaining: " SIZE_FORMAT
698 ", plab promotions %s, promotion reserve: " SIZE_FORMAT ", promotion expended: " SIZE_FORMAT
699 ", old capacity: " SIZE_FORMAT ", old_used: " SIZE_FORMAT ", old unaffiliated regions: " SIZE_FORMAT,
700 size * HeapWordSize, plab == nullptr? "no": "yes",
701 words_remaining * HeapWordSize, promote_enabled, promotion_reserve, promotion_expended,
702 max_capacity(), used(), free_unaffiliated_regions());
703
704 if ((gc_id == last_report_epoch) && (epoch_report_count >= MaxReportsPerEpoch)) {
705 log_debug(gc, ergo)("Squelching additional promotion failure reports for current epoch");
706 } else if (gc_id != last_report_epoch) {
707 last_report_epoch = gc_id;
708 epoch_report_count = 1;
709 }
710 }
711 }
712
713 void ShenandoahOldGeneration::handle_evacuation(HeapWord* obj, size_t words, bool promotion) {
714 // Only register the copy of the object that won the evacuation race.
715 _card_scan->register_object_without_lock(obj);
716
717 // Mark the entire range of the evacuated object as dirty. At next remembered set scan,
718 // we will clear dirty bits that do not hold interesting pointers. It's more efficient to
719 // do this in batch, in a background GC thread than to try to carefully dirty only cards
720 // that hold interesting pointers right now.
721 _card_scan->mark_range_as_dirty(obj, words);
722
723 if (promotion) {
724 // This evacuation was a promotion, track this as allocation against old gen
725 increase_allocated(words * HeapWordSize);
726 }
727 }
728
729 bool ShenandoahOldGeneration::has_unprocessed_collection_candidates() {
730 return _old_heuristics->unprocessed_old_collection_candidates() > 0;
731 }
732
733 size_t ShenandoahOldGeneration::unprocessed_collection_candidates_live_memory() {
734 return _old_heuristics->unprocessed_old_collection_candidates_live_memory();
735 }
736
737 void ShenandoahOldGeneration::abandon_collection_candidates() {
738 _old_heuristics->abandon_collection_candidates();
739 }
740
741 void ShenandoahOldGeneration::prepare_for_mixed_collections_after_global_gc() {
742 assert(is_mark_complete(), "Expected old generation mark to be complete after global cycle.");
743 _old_heuristics->prepare_for_old_collections();
744 log_info(gc, ergo)("After choosing global collection set, mixed candidates: " UINT32_FORMAT ", coalescing candidates: " SIZE_FORMAT,
745 _old_heuristics->unprocessed_old_collection_candidates(),
746 _old_heuristics->coalesce_and_fill_candidates_count());
747 }
748
749 void ShenandoahOldGeneration::parallel_heap_region_iterate_free(ShenandoahHeapRegionClosure* cl) {
750 // Iterate over old and free regions (exclude young).
751 ShenandoahExcludeRegionClosure<YOUNG_GENERATION> exclude_cl(cl);
752 ShenandoahGeneration::parallel_heap_region_iterate_free(&exclude_cl);
753 }
754
755 void ShenandoahOldGeneration::set_parsable(bool parsable) {
756 _is_parsable = parsable;
757 if (_is_parsable) {
758 // The current state would have been chosen during final mark of the global
759 // collection, _before_ any decisions about class unloading have been made.
760 //
761 // After unloading classes, we have made the old generation regions parsable.
762 // We can skip filling or transition to a state that knows everything has
763 // already been filled.
764 switch (state()) {
765 case ShenandoahOldGeneration::EVACUATING:
766 transition_to(ShenandoahOldGeneration::EVACUATING_AFTER_GLOBAL);
767 break;
768 case ShenandoahOldGeneration::FILLING:
769 assert(_old_heuristics->unprocessed_old_collection_candidates() == 0, "Expected no mixed collection candidates");
770 assert(_old_heuristics->coalesce_and_fill_candidates_count() > 0, "Expected coalesce and fill candidates");
771 // When the heuristic put the old generation in this state, it didn't know
772 // that we would unload classes and make everything parsable. But, we know
773 // that now so we can override this state.
774 abandon_collection_candidates();
775 transition_to(ShenandoahOldGeneration::WAITING_FOR_BOOTSTRAP);
776 break;
777 default:
778 // We can get here during a full GC. The full GC will cancel anything
779 // happening in the old generation and return it to the waiting for bootstrap
780 // state. The full GC will then record that the old regions are parsable
781 // after rebuilding the remembered set.
782 assert(is_idle(), "Unexpected state %s at end of global GC", state_name());
783 break;
784 }
785 }
786 }
787
788 void ShenandoahOldGeneration::complete_mixed_evacuations() {
789 assert(is_doing_mixed_evacuations(), "Mixed evacuations should be in progress");
790 if (!_old_heuristics->has_coalesce_and_fill_candidates()) {
791 // No candidate regions to coalesce and fill
792 transition_to(ShenandoahOldGeneration::WAITING_FOR_BOOTSTRAP);
793 return;
794 }
795
796 if (state() == ShenandoahOldGeneration::EVACUATING) {
797 transition_to(ShenandoahOldGeneration::FILLING);
798 return;
799 }
800
801 // Here, we have no more candidates for mixed collections. The candidates for coalescing
802 // and filling have already been processed during the global cycle, so there is nothing
803 // more to do.
804 assert(state() == ShenandoahOldGeneration::EVACUATING_AFTER_GLOBAL, "Should be evacuating after a global cycle");
805 abandon_collection_candidates();
806 transition_to(ShenandoahOldGeneration::WAITING_FOR_BOOTSTRAP);
807 }
808
809 void ShenandoahOldGeneration::abandon_mixed_evacuations() {
810 switch(state()) {
811 case ShenandoahOldGeneration::EVACUATING:
812 transition_to(ShenandoahOldGeneration::FILLING);
813 break;
814 case ShenandoahOldGeneration::EVACUATING_AFTER_GLOBAL:
815 abandon_collection_candidates();
816 transition_to(ShenandoahOldGeneration::WAITING_FOR_BOOTSTRAP);
817 break;
818 default:
819 log_warning(gc)("Abandon mixed evacuations in unexpected state: %s", state_name(state()));
820 ShouldNotReachHere();
821 break;
822 }
823 }
824
825 void ShenandoahOldGeneration::clear_cards_for(ShenandoahHeapRegion* region) {
826 _card_scan->mark_range_as_empty(region->bottom(), pointer_delta(region->end(), region->bottom()));
827 }
828
829 void ShenandoahOldGeneration::mark_card_as_dirty(void* location) {
830 _card_scan->mark_card_as_dirty((HeapWord*)location);
831 }