1 /*
  2  * Copyright (c) 2015, 2019, Red Hat, Inc. All rights reserved.
  3  * Copyright Amazon.com Inc. or its affiliates. All Rights Reserved.
  4  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
  5  *
  6  * This code is free software; you can redistribute it and/or modify it
  7  * under the terms of the GNU General Public License version 2 only, as
  8  * published by the Free Software Foundation.
  9  *
 10  * This code is distributed in the hope that it will be useful, but WITHOUT
 11  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
 12  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
 13  * version 2 for more details (a copy is included in the LICENSE file that
 14  * accompanied this code).
 15  *
 16  * You should have received a copy of the GNU General Public License version
 17  * 2 along with this work; if not, write to the Free Software Foundation,
 18  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
 19  *
 20  * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
 21  * or visit www.oracle.com if you need additional information or have any
 22  * questions.
 23  *
 24  */
 25 
 26 #ifndef SHARE_GC_SHENANDOAH_SHENANDOAHHEAPREGION_INLINE_HPP
 27 #define SHARE_GC_SHENANDOAH_SHENANDOAHHEAPREGION_INLINE_HPP
 28 
 29 #include "gc/shenandoah/shenandoahGenerationalHeap.hpp"
 30 #include "gc/shenandoah/shenandoahHeapRegion.hpp"
 31 #include "gc/shenandoah/shenandoahHeap.inline.hpp"
 32 #include "gc/shenandoah/shenandoahOldGeneration.hpp"
 33 #include "gc/shenandoah/shenandoahPacer.inline.hpp"
 34 #include "runtime/atomic.hpp"
 35 
 36 HeapWord* ShenandoahHeapRegion::allocate_aligned(size_t size, ShenandoahAllocRequest &req, size_t alignment_in_bytes) {
 37   shenandoah_assert_heaplocked_or_safepoint();
 38   assert(req.is_lab_alloc(), "allocate_aligned() only applies to LAB allocations");
 39   assert(is_object_aligned(size), "alloc size breaks alignment: " SIZE_FORMAT, size);
 40   assert(is_old(), "aligned allocations are only taken from OLD regions to support PLABs");
 41   assert(is_aligned(alignment_in_bytes, HeapWordSize), "Expect hea word alignment");
 42 
 43   HeapWord* orig_top = top();
 44   size_t alignment_in_words = alignment_in_bytes / HeapWordSize;
 45 
 46   // unalignment_words is the amount by which current top() exceeds the desired alignment point.  We subtract this amount
 47   // from alignment_in_words to determine padding required to next alignment point.
 48 
 49   HeapWord* aligned_obj = (HeapWord*) align_up(orig_top, alignment_in_bytes);
 50   size_t pad_words = aligned_obj - orig_top;
 51   if ((pad_words > 0) && (pad_words < ShenandoahHeap::min_fill_size())) {
 52     pad_words += alignment_in_words;
 53     aligned_obj += alignment_in_words;
 54   }
 55 
 56   if (pointer_delta(end(), aligned_obj) < size) {
 57     // Shrink size to fit within available space and align it
 58     size = pointer_delta(end(), aligned_obj);
 59     size = align_down(size, alignment_in_words);
 60   }
 61 
 62   // Both originally requested size and adjusted size must be properly aligned
 63   assert (is_aligned(size, alignment_in_words), "Size must be multiple of alignment constraint");
 64   if (size >= req.min_size()) {
 65     // Even if req.min_size() may not be a multiple of card size, we know that size is.
 66     if (pad_words > 0) {
 67       assert(pad_words >= ShenandoahHeap::min_fill_size(), "pad_words expanded above to meet size constraint");
 68       ShenandoahHeap::fill_with_object(orig_top, pad_words);
 69       ShenandoahGenerationalHeap::heap()->old_generation()->card_scan()->register_object(orig_top);
 70     }
 71 
 72     make_regular_allocation(req.affiliation());
 73     adjust_alloc_metadata(req.type(), size);
 74 
 75     HeapWord* new_top = aligned_obj + size;
 76     assert(new_top <= end(), "PLAB cannot span end of heap region");
 77     set_top(new_top);
 78     req.set_actual_size(size);
 79     req.set_waste(pad_words);
 80     assert(is_object_aligned(new_top), "new top breaks alignment: " PTR_FORMAT, p2i(new_top));
 81     assert(is_aligned(aligned_obj, alignment_in_bytes), "obj is not aligned: " PTR_FORMAT, p2i(aligned_obj));
 82     return aligned_obj;
 83   } else {
 84     // The aligned size that fits in this region is smaller than min_size, so don't align top and don't allocate.  Return failure.
 85     return nullptr;
 86   }
 87 }
 88 
 89 HeapWord* ShenandoahHeapRegion::allocate(size_t size, ShenandoahAllocRequest req) {
 90   shenandoah_assert_heaplocked_or_safepoint();
 91   assert(is_object_aligned(size), "alloc size breaks alignment: " SIZE_FORMAT, size);
 92 
 93   HeapWord* obj = top();
 94   if (pointer_delta(end(), obj) >= size) {
 95     make_regular_allocation(req.affiliation());
 96     adjust_alloc_metadata(req.type(), size);
 97 
 98     HeapWord* new_top = obj + size;
 99     set_top(new_top);
100 
101     assert(is_object_aligned(new_top), "new top breaks alignment: " PTR_FORMAT, p2i(new_top));
102     assert(is_object_aligned(obj),     "obj is not aligned: "       PTR_FORMAT, p2i(obj));
103 
104     return obj;
105   } else {
106     return nullptr;
107   }
108 }
109 
110 inline void ShenandoahHeapRegion::adjust_alloc_metadata(ShenandoahAllocRequest::Type type, size_t size) {
111   switch (type) {
112     case ShenandoahAllocRequest::_alloc_shared:
113     case ShenandoahAllocRequest::_alloc_shared_gc:
114       // Counted implicitly by tlab/gclab allocs
115       break;
116     case ShenandoahAllocRequest::_alloc_tlab:
117       _tlab_allocs += size;
118       break;
119     case ShenandoahAllocRequest::_alloc_gclab:
120       _gclab_allocs += size;
121       break;
122     case ShenandoahAllocRequest::_alloc_plab:
123       _plab_allocs += size;
124       break;
125     default:
126       ShouldNotReachHere();
127   }
128 }
129 
130 inline void ShenandoahHeapRegion::increase_live_data_alloc_words(size_t s) {
131   internal_increase_live_data(s);
132 }
133 
134 inline void ShenandoahHeapRegion::increase_live_data_gc_words(size_t s) {
135   internal_increase_live_data(s);
136   if (ShenandoahPacing) {
137     ShenandoahHeap::heap()->pacer()->report_mark(s);
138   }
139 }
140 
141 inline void ShenandoahHeapRegion::internal_increase_live_data(size_t s) {
142   size_t new_live_data = Atomic::add(&_live_data, s, memory_order_relaxed);
143 }
144 
145 inline void ShenandoahHeapRegion::clear_live_data() {
146   Atomic::store(&_live_data, (size_t)0);
147 }
148 
149 inline size_t ShenandoahHeapRegion::get_live_data_words() const {
150   return Atomic::load(&_live_data);
151 }
152 
153 inline size_t ShenandoahHeapRegion::get_live_data_bytes() const {
154   return get_live_data_words() * HeapWordSize;
155 }
156 
157 inline bool ShenandoahHeapRegion::has_live() const {
158   return get_live_data_words() != 0;
159 }
160 
161 inline size_t ShenandoahHeapRegion::garbage() const {
162   assert(used() >= get_live_data_bytes(),
163          "Live Data must be a subset of used() live: " SIZE_FORMAT " used: " SIZE_FORMAT,
164          get_live_data_bytes(), used());
165 
166   size_t result = used() - get_live_data_bytes();
167   return result;
168 }
169 
170 inline size_t ShenandoahHeapRegion::garbage_before_padded_for_promote() const {
171   assert(get_top_before_promote() != nullptr, "top before promote should not equal null");
172   size_t used_before_promote = byte_size(bottom(), get_top_before_promote());
173   assert(used_before_promote >= get_live_data_bytes(),
174          "Live Data must be a subset of used before promotion live: " SIZE_FORMAT " used: " SIZE_FORMAT,
175          get_live_data_bytes(), used_before_promote);
176   size_t result = used_before_promote - get_live_data_bytes();
177   return result;
178 
179 }
180 
181 inline HeapWord* ShenandoahHeapRegion::get_update_watermark() const {
182   HeapWord* watermark = Atomic::load_acquire(&_update_watermark);
183   assert(bottom() <= watermark && watermark <= top(), "within bounds");
184   return watermark;
185 }
186 
187 inline void ShenandoahHeapRegion::set_update_watermark(HeapWord* w) {
188   assert(bottom() <= w && w <= top(), "within bounds");
189   Atomic::release_store(&_update_watermark, w);
190 }
191 
192 // Fast version that avoids synchronization, only to be used at safepoints.
193 inline void ShenandoahHeapRegion::set_update_watermark_at_safepoint(HeapWord* w) {
194   assert(bottom() <= w && w <= top(), "within bounds");
195   assert(SafepointSynchronize::is_at_safepoint(), "Should be at Shenandoah safepoint");
196   _update_watermark = w;
197 }
198 
199 inline ShenandoahAffiliation ShenandoahHeapRegion::affiliation() const {
200   return ShenandoahHeap::heap()->region_affiliation(this);
201 }
202 
203 inline const char* ShenandoahHeapRegion::affiliation_name() const {
204   return shenandoah_affiliation_name(affiliation());
205 }
206 
207 inline bool ShenandoahHeapRegion::is_young() const {
208   return affiliation() == YOUNG_GENERATION;
209 }
210 
211 inline bool ShenandoahHeapRegion::is_old() const {
212   return affiliation() == OLD_GENERATION;
213 }
214 
215 inline bool ShenandoahHeapRegion::is_affiliated() const {
216   return affiliation() != FREE;
217 }
218 
219 inline void ShenandoahHeapRegion::save_top_before_promote() {
220   _top_before_promoted = _top;
221 }
222 
223 inline void ShenandoahHeapRegion::restore_top_before_promote() {
224   _top = _top_before_promoted;
225   _top_before_promoted = nullptr;
226  }
227 
228 
229 #endif // SHARE_GC_SHENANDOAH_SHENANDOAHHEAPREGION_INLINE_HPP