1 /*
  2  * Copyright (c) 2013, 2019, Red Hat, Inc. All rights reserved.
  3  * Copyright Amazon.com Inc. or its affiliates. All Rights Reserved.
  4  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
  5  *
  6  * This code is free software; you can redistribute it and/or modify it
  7  * under the terms of the GNU General Public License version 2 only, as
  8  * published by the Free Software Foundation.
  9  *
 10  * This code is distributed in the hope that it will be useful, but WITHOUT
 11  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
 12  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
 13  * version 2 for more details (a copy is included in the LICENSE file that
 14  * accompanied this code).
 15  *
 16  * You should have received a copy of the GNU General Public License version
 17  * 2 along with this work; if not, write to the Free Software Foundation,
 18  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
 19  *
 20  * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
 21  * or visit www.oracle.com if you need additional information or have any
 22  * questions.
 23  *
 24  */
 25 
 26 #ifndef SHARE_GC_SHENANDOAH_SHENANDOAHHEAPREGION_HPP
 27 #define SHARE_GC_SHENANDOAH_SHENANDOAHHEAPREGION_HPP
 28 
 29 #include "gc/shared/gc_globals.hpp"
 30 #include "gc/shared/spaceDecorator.hpp"
 31 #include "gc/shenandoah/shenandoahAffiliation.hpp"
 32 #include "gc/shenandoah/shenandoahAgeCensus.hpp"
 33 #include "gc/shenandoah/shenandoahAllocRequest.hpp"
 34 #include "gc/shenandoah/shenandoahAsserts.hpp"
 35 #include "gc/shenandoah/shenandoahHeap.hpp"
 36 #include "gc/shenandoah/shenandoahPadding.hpp"
 37 #include "utilities/sizes.hpp"
 38 
 39 class VMStructs;
 40 class ShenandoahHeapRegionStateConstant;
 41 
 42 class ShenandoahHeapRegion {
 43   friend class VMStructs;
 44   friend class ShenandoahHeapRegionStateConstant;
 45 private:
 46   /*
 47     Region state is described by a state machine. Transitions are guarded by
 48     heap lock, which allows changing the state of several regions atomically.
 49     Region states can be logically aggregated in groups.
 50 
 51       "Empty":
 52       .................................................................
 53       .                                                               .
 54       .                                                               .
 55       .         Uncommitted  <-------  Committed <------------------------\
 56       .              |                     |                          .   |
 57       .              \---------v-----------/                          .   |
 58       .                        |                                      .   |
 59       .........................|.......................................   |
 60                                |                                          |
 61       "Active":                |                                          |
 62       .........................|.......................................   |
 63       .                        |                                      .   |
 64       .      /-----------------^-------------------\                  .   |
 65       .      |                                     |                  .   |
 66       .      v                                     v    "Humongous":  .   |
 67       .   Regular ---\-----\     ..................O................  .   |
 68       .     |  ^     |     |     .                 |               .  .   |
 69       .     |  |     |     |     .                 *---------\     .  .   |
 70       .     v  |     |     |     .                 v         v     .  .   |
 71       .    Pinned  Cset    |     .  HStart <--> H/Start   H/Cont   .  .   |
 72       .       ^    / |     |     .  Pinned         v         |     .  .   |
 73       .       |   /  |     |     .                 *<--------/     .  .   |
 74       .       |  v   |     |     .                 |               .  .   |
 75       .  CsetPinned  |     |     ..................O................  .   |
 76       .              |     |                       |                  .   |
 77       .              \-----\---v-------------------/                  .   |
 78       .                        |                                      .   |
 79       .........................|.......................................   |
 80                                |                                          |
 81       "Trash":                 |                                          |
 82       .........................|.......................................   |
 83       .                        |                                      .   |
 84       .                        v                                      .   |
 85       .                      Trash ---------------------------------------/
 86       .                                                               .
 87       .                                                               .
 88       .................................................................
 89 
 90     Transition from "Empty" to "Active" is first allocation. It can go from {Uncommitted, Committed}
 91     to {Regular, "Humongous"}. The allocation may happen in Regular regions too, but not in Humongous.
 92 
 93     Transition from "Active" to "Trash" is reclamation. It can go from CSet during the normal cycle,
 94     and from {Regular, "Humongous"} for immediate reclamation. The existence of Trash state allows
 95     quick reclamation without actual cleaning up.
 96 
 97     Transition from "Trash" to "Empty" is recycling. It cleans up the regions and corresponding metadata.
 98     Can be done asynchronously and in bulk.
 99 
100     Note how internal transitions disallow logic bugs:
101       a) No region can go Empty, unless properly reclaimed/recycled;
102       b) No region can go Uncommitted, unless reclaimed/recycled first;
103       c) Only Regular regions can go to CSet;
104       d) Pinned cannot go Trash, thus it could never be reclaimed until unpinned;
105       e) Pinned cannot go CSet, thus it never moves;
106       f) Humongous cannot be used for regular allocations;
107       g) Humongous cannot go CSet, thus it never moves;
108       h) Humongous start can go pinned, and thus can be protected from moves (humongous continuations should
109          follow associated humongous starts, not pinnable/movable by themselves);
110       i) Empty cannot go Trash, avoiding useless work;
111       j) ...
112    */
113 
114   enum RegionState {
115     _empty_uncommitted,       // region is empty and has memory uncommitted
116     _empty_committed,         // region is empty and has memory committed
117     _regular,                 // region is for regular allocations
118     _humongous_start,         // region is the humongous start
119     _humongous_cont,          // region is the humongous continuation
120     _pinned_humongous_start,  // region is both humongous start and pinned
121     _cset,                    // region is in collection set
122     _pinned,                  // region is pinned
123     _pinned_cset,             // region is pinned and in cset (evac failure path)
124     _trash,                   // region contains only trash
125     _REGION_STATES_NUM        // last
126   };
127 
128 public:
129   static const char* region_state_to_string(RegionState s) {
130     switch (s) {
131       case _empty_uncommitted:       return "Empty Uncommitted";
132       case _empty_committed:         return "Empty Committed";
133       case _regular:                 return "Regular";
134       case _humongous_start:         return "Humongous Start";
135       case _humongous_cont:          return "Humongous Continuation";
136       case _pinned_humongous_start:  return "Humongous Start, Pinned";
137       case _cset:                    return "Collection Set";
138       case _pinned:                  return "Pinned";
139       case _pinned_cset:             return "Collection Set, Pinned";
140       case _trash:                   return "Trash";
141       default:
142         ShouldNotReachHere();
143         return "";
144     }
145   }
146 
147 private:
148   // This method protects from accidental changes in enum order:
149   int region_state_to_ordinal(RegionState s) const {
150     switch (s) {
151       case _empty_uncommitted:      return 0;
152       case _empty_committed:        return 1;
153       case _regular:                return 2;
154       case _humongous_start:        return 3;
155       case _humongous_cont:         return 4;
156       case _cset:                   return 5;
157       case _pinned:                 return 6;
158       case _trash:                  return 7;
159       case _pinned_cset:            return 8;
160       case _pinned_humongous_start: return 9;
161       default:
162         ShouldNotReachHere();
163         return -1;
164     }
165   }
166 
167   void report_illegal_transition(const char* method);
168 
169 public:
170   static int region_states_num() {
171     return _REGION_STATES_NUM;
172   }
173 
174   // Allowed transitions from the outside code:
175   void make_regular_allocation(ShenandoahAffiliation affiliation);
176   void make_affiliated_maybe();
177   void make_regular_bypass();
178   void make_humongous_start();
179   void make_humongous_cont();
180   void make_humongous_start_bypass(ShenandoahAffiliation affiliation);
181   void make_humongous_cont_bypass(ShenandoahAffiliation affiliation);
182   void make_pinned();
183   void make_unpinned();
184   void make_cset();
185   void make_trash();
186   void make_trash_immediate();
187   void make_empty();
188   void make_uncommitted();
189   void make_committed_bypass();
190 
191   // Individual states:
192   bool is_empty_uncommitted()      const { return _state == _empty_uncommitted; }
193   bool is_empty_committed()        const { return _state == _empty_committed; }
194   bool is_regular()                const { return _state == _regular; }
195   bool is_humongous_continuation() const { return _state == _humongous_cont; }
196 
197   // Participation in logical groups:
198   bool is_empty()                  const { return is_empty_committed() || is_empty_uncommitted(); }
199   bool is_active()                 const { return !is_empty() && !is_trash(); }
200   bool is_trash()                  const { return _state == _trash; }
201   bool is_humongous_start()        const { return _state == _humongous_start || _state == _pinned_humongous_start; }
202   bool is_humongous()              const { return is_humongous_start() || is_humongous_continuation(); }
203   bool is_committed()              const { return !is_empty_uncommitted(); }
204   bool is_cset()                   const { return _state == _cset   || _state == _pinned_cset; }
205   bool is_pinned()                 const { return _state == _pinned || _state == _pinned_cset || _state == _pinned_humongous_start; }
206   bool is_regular_pinned()         const { return _state == _pinned; }
207 
208   inline bool is_young() const;
209   inline bool is_old() const;
210   inline bool is_affiliated() const;
211 
212   // Macro-properties:
213   bool is_alloc_allowed()          const { return is_empty() || is_regular() || _state == _pinned; }
214   bool is_stw_move_allowed()       const { return is_regular() || _state == _cset || (ShenandoahHumongousMoves && _state == _humongous_start); }
215 
216   RegionState state()              const { return _state; }
217   int  state_ordinal()             const { return region_state_to_ordinal(_state); }
218 
219   void record_pin();
220   void record_unpin();
221   size_t pin_count() const;
222 
223 private:
224   static size_t RegionCount;
225   static size_t RegionSizeBytes;
226   static size_t RegionSizeWords;
227   static size_t RegionSizeBytesShift;
228   static size_t RegionSizeWordsShift;
229   static size_t RegionSizeBytesMask;
230   static size_t RegionSizeWordsMask;
231   static size_t MaxTLABSizeBytes;
232   static size_t MaxTLABSizeWords;
233 
234   // Never updated fields
235   size_t const _index;
236   HeapWord* const _bottom;
237   HeapWord* const _end;
238 
239   // Rarely updated fields
240   HeapWord* _new_top;
241   double _empty_time;
242 
243   HeapWord* _top_before_promoted;
244 
245   // Seldom updated fields
246   RegionState _state;
247   HeapWord* _coalesce_and_fill_boundary; // for old regions not selected as collection set candidates.
248 
249   // Frequently updated fields
250   HeapWord* _top;
251 
252   size_t _tlab_allocs;
253   size_t _gclab_allocs;
254   size_t _plab_allocs;
255 
256   volatile size_t _live_data;
257   volatile size_t _critical_pins;
258 
259   HeapWord* volatile _update_watermark;
260 
261   uint _age;
262   CENSUS_NOISE(uint _youth;)   // tracks epochs of retrograde ageing (rejuvenation)
263 
264 public:
265   ShenandoahHeapRegion(HeapWord* start, size_t index, bool committed);
266 
267   static const size_t MIN_NUM_REGIONS = 10;
268 
269   // Return adjusted max heap size
270   static size_t setup_sizes(size_t max_heap_size);
271 
272   double empty_time() {
273     return _empty_time;
274   }
275 
276   inline static size_t required_regions(size_t bytes) {
277     return (bytes + ShenandoahHeapRegion::region_size_bytes() - 1) >> ShenandoahHeapRegion::region_size_bytes_shift();
278   }
279 
280   inline static bool requires_humongous(size_t words) {
281     return words > ShenandoahHeapRegion::RegionSizeWords;
282   }
283 
284   inline static size_t region_count() {
285     return ShenandoahHeapRegion::RegionCount;
286   }
287 
288   inline static size_t region_size_bytes() {
289     return ShenandoahHeapRegion::RegionSizeBytes;
290   }
291 
292   inline static size_t region_size_words() {
293     return ShenandoahHeapRegion::RegionSizeWords;
294   }
295 
296   inline static size_t region_size_bytes_shift() {
297     return ShenandoahHeapRegion::RegionSizeBytesShift;
298   }
299 
300   inline static size_t region_size_words_shift() {
301     return ShenandoahHeapRegion::RegionSizeWordsShift;
302   }
303 
304   inline static size_t region_size_bytes_mask() {
305     return ShenandoahHeapRegion::RegionSizeBytesMask;
306   }
307 
308   inline static size_t region_size_words_mask() {
309     return ShenandoahHeapRegion::RegionSizeWordsMask;
310   }
311 
312   // Convert to jint with sanity checking
313   inline static jint region_size_bytes_jint() {
314     assert (ShenandoahHeapRegion::RegionSizeBytes <= (size_t)max_jint, "sanity");
315     return (jint)ShenandoahHeapRegion::RegionSizeBytes;
316   }
317 
318   // Convert to jint with sanity checking
319   inline static jint region_size_words_jint() {
320     assert (ShenandoahHeapRegion::RegionSizeWords <= (size_t)max_jint, "sanity");
321     return (jint)ShenandoahHeapRegion::RegionSizeWords;
322   }
323 
324   // Convert to jint with sanity checking
325   inline static jint region_size_bytes_shift_jint() {
326     assert (ShenandoahHeapRegion::RegionSizeBytesShift <= (size_t)max_jint, "sanity");
327     return (jint)ShenandoahHeapRegion::RegionSizeBytesShift;
328   }
329 
330   // Convert to jint with sanity checking
331   inline static jint region_size_words_shift_jint() {
332     assert (ShenandoahHeapRegion::RegionSizeWordsShift <= (size_t)max_jint, "sanity");
333     return (jint)ShenandoahHeapRegion::RegionSizeWordsShift;
334   }
335 
336   inline static size_t max_tlab_size_bytes() {
337     return ShenandoahHeapRegion::MaxTLABSizeBytes;
338   }
339 
340   inline static size_t max_tlab_size_words() {
341     return ShenandoahHeapRegion::MaxTLABSizeWords;
342   }
343 
344   inline size_t index() const {
345     return _index;
346   }
347 
348   inline void save_top_before_promote();
349   inline HeapWord* get_top_before_promote() const { return _top_before_promoted; }
350   inline void restore_top_before_promote();
351   inline size_t garbage_before_padded_for_promote() const;
352 
353   // If next available memory is not aligned on address that is multiple of alignment, fill the empty space
354   // so that returned object is aligned on an address that is a multiple of alignment_in_bytes.  Requested
355   // size is in words.  It is assumed that this->is_old().  A pad object is allocated, filled, and registered
356   // if necessary to assure the new allocation is properly aligned.  Return nullptr if memory is not available.
357   inline HeapWord* allocate_aligned(size_t word_size, ShenandoahAllocRequest &req, size_t alignment_in_bytes);
358 
359   // Allocation (return nullptr if full)
360   inline HeapWord* allocate(size_t word_size, const ShenandoahAllocRequest& req);
361 
362   inline void clear_live_data();
363   void set_live_data(size_t s);
364 
365   // Increase live data for newly allocated region
366   inline void increase_live_data_alloc_words(size_t s);
367 
368   // Increase live data for region scanned with GC
369   inline void increase_live_data_gc_words(size_t s);
370 
371   inline bool has_live() const;
372   inline size_t get_live_data_bytes() const;
373   inline size_t get_live_data_words() const;
374 
375   inline size_t garbage() const;
376 
377   void print_on(outputStream* st) const;
378 
379   void recycle();
380 
381   inline void begin_preemptible_coalesce_and_fill() {
382     _coalesce_and_fill_boundary = _bottom;
383   }
384 
385   inline void end_preemptible_coalesce_and_fill() {
386     _coalesce_and_fill_boundary = _end;
387   }
388 
389   inline void suspend_coalesce_and_fill(HeapWord* next_focus) {
390     _coalesce_and_fill_boundary = next_focus;
391   }
392 
393   inline HeapWord* resume_coalesce_and_fill() {
394     return _coalesce_and_fill_boundary;
395   }
396 
397   // Coalesce contiguous spans of garbage objects by filling header and registering start locations with remembered set.
398   // This is used by old-gen GC following concurrent marking to make old-gen HeapRegions parsable. Old regions must be
399   // parsable because the mark bitmap is not reliable during the concurrent old mark.
400   // Return true iff region is completely coalesced and filled.  Returns false if cancelled before task is complete.
401   bool oop_coalesce_and_fill(bool cancellable);
402 
403   // Invoke closure on every reference contained within the humongous object that spans this humongous
404   // region if the reference is contained within a DIRTY card and the reference is no more than words following
405   // start within the humongous object.
406   void oop_iterate_humongous_slice_dirty(OopIterateClosure* cl, HeapWord* start, size_t words, bool write_table) const;
407 
408   // Invoke closure on every reference contained within the humongous object starting from start and
409   // ending at start + words.
410   void oop_iterate_humongous_slice_all(OopIterateClosure* cl, HeapWord* start, size_t words) const;
411 
412   HeapWord* block_start(const void* p) const;
413   size_t block_size(const HeapWord* p) const;
414   bool block_is_obj(const HeapWord* p) const { return p < top(); }
415 
416   // Find humongous start region that this region belongs to
417   ShenandoahHeapRegion* humongous_start_region() const;
418 
419   HeapWord* top() const         { return _top;     }
420   void set_top(HeapWord* v)     { _top = v;        }
421 
422   HeapWord* new_top() const     { return _new_top; }
423   void set_new_top(HeapWord* v) { _new_top = v;    }
424 
425   HeapWord* bottom() const      { return _bottom;  }
426   HeapWord* end() const         { return _end;     }
427 
428   size_t capacity() const       { return byte_size(bottom(), end()); }
429   size_t used() const           { return byte_size(bottom(), top()); }
430   size_t used_before_promote() const { return byte_size(bottom(), get_top_before_promote()); }
431   size_t free() const           { return byte_size(top(),    end()); }
432 
433   // Does this region contain this address?
434   bool contains(HeapWord* p) const {
435     return (bottom() <= p) && (p < top());
436   }
437 
438   inline void adjust_alloc_metadata(ShenandoahAllocRequest::Type type, size_t);
439   void reset_alloc_metadata();
440   size_t get_shared_allocs() const;
441   size_t get_tlab_allocs() const;
442   size_t get_gclab_allocs() const;
443   size_t get_plab_allocs() const;
444 
445   inline HeapWord* get_update_watermark() const;
446   inline void set_update_watermark(HeapWord* w);
447   inline void set_update_watermark_at_safepoint(HeapWord* w);
448 
449   inline ShenandoahAffiliation affiliation() const;
450   inline const char* affiliation_name() const;
451 
452   void set_affiliation(ShenandoahAffiliation new_affiliation);
453 
454   // Region ageing and rejuvenation
455   uint age() const { return _age; }
456   CENSUS_NOISE(uint youth() const { return _youth; })
457 
458   void increment_age() {
459     const uint max_age = markWord::max_age;
460     assert(_age <= max_age, "Error");
461     if (_age++ >= max_age) {
462       _age = max_age;   // clamp
463     }
464   }
465 
466   void reset_age() {
467     CENSUS_NOISE(_youth += _age;)
468     _age = 0;
469   }
470 
471   CENSUS_NOISE(void clear_youth() { _youth = 0; })
472 
473 private:
474   void decrement_humongous_waste() const;
475   void do_commit();
476   void do_uncommit();
477 
478   inline void internal_increase_live_data(size_t s);
479 
480   void set_state(RegionState to);
481 };
482 
483 #endif // SHARE_GC_SHENANDOAH_SHENANDOAHHEAPREGION_HPP