1 /*
  2  * Copyright (c) 2019, 2024, Oracle and/or its affiliates. All rights reserved.
  3  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
  4  *
  5  * This code is free software; you can redistribute it and/or modify it
  6  * under the terms of the GNU General Public License version 2 only, as
  7  * published by the Free Software Foundation.
  8  *
  9  * This code is distributed in the hope that it will be useful, but WITHOUT
 10  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
 11  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
 12  * version 2 for more details (a copy is included in the LICENSE file that
 13  * accompanied this code).
 14  *
 15  * You should have received a copy of the GNU General Public License version
 16  * 2 along with this work; if not, write to the Free Software Foundation,
 17  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
 18  *
 19  * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
 20  * or visit www.oracle.com if you need additional information or have any
 21  * questions.
 22  *
 23  */
 24 
 25 #ifndef SHARE_CDS_ARCHIVEUTILS_HPP
 26 #define SHARE_CDS_ARCHIVEUTILS_HPP
 27 
 28 #include "cds/cds_globals.hpp"
 29 #include "cds/serializeClosure.hpp"
 30 #include "logging/log.hpp"
 31 #include "memory/metaspace.hpp"
 32 #include "memory/virtualspace.hpp"
 33 #include "utilities/bitMap.hpp"
 34 #include "utilities/exceptions.hpp"
 35 #include "utilities/macros.hpp"
 36 #include "runtime/nonJavaThread.hpp"
 37 #include "runtime/semaphore.hpp"
 38 
 39 class BootstrapInfo;
 40 class ReservedSpace;
 41 class VirtualSpace;
 42 
 43 template<class E> class Array;
 44 template<class E> class GrowableArray;
 45 
 46 // ArchivePtrMarker is used to mark the location of pointers embedded in a CDS archive. E.g., when an
 47 // InstanceKlass k is dumped, we mark the location of the k->_name pointer by effectively calling
 48 // mark_pointer(/*ptr_loc=*/&k->_name). It's required that (_prt_base <= ptr_loc < _ptr_end). _ptr_base is
 49 // fixed, but _ptr_end can be expanded as more objects are dumped.
 50 class ArchivePtrMarker : AllStatic {
 51   static CHeapBitMap*  _ptrmap;
 52   static CHeapBitMap*  _rw_ptrmap;
 53   static CHeapBitMap*  _ro_ptrmap;
 54   static VirtualSpace* _vs;
 55 
 56   // Once _ptrmap is compacted, we don't allow bit marking anymore. This is to
 57   // avoid unintentional copy operations after the bitmap has been finalized and written.
 58   static bool         _compacted;
 59 
 60   static address* ptr_base() { return (address*)_vs->low();  } // committed lower bound (inclusive)
 61   static address* ptr_end()  { return (address*)_vs->high(); } // committed upper bound (exclusive)
 62 
 63 public:
 64   static void initialize(CHeapBitMap* ptrmap, VirtualSpace* vs);
 65   static void initialize_rw_ro_maps(CHeapBitMap* rw_ptrmap, CHeapBitMap* ro_ptrmap);
 66   static void mark_pointer(address* ptr_loc);
 67   static void clear_pointer(address* ptr_loc);
 68   static void compact(address relocatable_base, address relocatable_end);
 69   static void compact(size_t max_non_null_offset);
 70 
 71   template <typename T>
 72   static void mark_pointer(T* ptr_loc) {
 73     mark_pointer((address*)ptr_loc);
 74   }
 75 
 76   template <typename T>
 77   static void set_and_mark_pointer(T* ptr_loc, T ptr_value) {
 78     *ptr_loc = ptr_value;
 79     mark_pointer(ptr_loc);
 80   }
 81 
 82   static CHeapBitMap* ptrmap() {
 83     return _ptrmap;
 84   }
 85 
 86   static CHeapBitMap* rw_ptrmap() {
 87     return _rw_ptrmap;
 88   }
 89 
 90   static CHeapBitMap* ro_ptrmap() {
 91     return _ro_ptrmap;
 92   }
 93 
 94   static void reset_map_and_vs() {
 95     _ptrmap = nullptr;
 96     _rw_ptrmap = nullptr;
 97     _ro_ptrmap = nullptr;
 98     _vs = nullptr;
 99   }
100 };
101 
102 // SharedDataRelocator is used to shift pointers in the CDS archive.
103 //
104 // The CDS archive is basically a contiguous block of memory (divided into several regions)
105 // that contains multiple objects. The objects may contain direct pointers that point to other objects
106 // within the archive (e.g., InstanceKlass::_name points to a Symbol in the archive). During dumping, we
107 // built a bitmap that marks the locations of all these pointers (using ArchivePtrMarker, see comments above).
108 //
109 // The contents of the archive assumes that it’s mapped at the default SharedBaseAddress (e.g. 0x800000000).
110 // If the archive ends up being mapped at a different address (e.g. 0x810000000), SharedDataRelocator
111 // is used to shift each marked pointer by a delta (0x10000000 in this example), so that it points to
112 // the actually mapped location of the target object.
113 class SharedDataRelocator: public BitMapClosure {
114   // for all (address** p), where (is_marked(p) && _patch_base <= p && p < _patch_end) { *p += delta; }
115 
116   // Patch all pointers within this region that are marked.
117   address* _patch_base;
118   address* _patch_end;
119 
120   // Before patching, all pointers must point to this region.
121   address _valid_old_base;
122   address _valid_old_end;
123 
124   // After patching, all pointers must point to this region.
125   address _valid_new_base;
126   address _valid_new_end;
127 
128   // How much to relocate for each pointer.
129   intx _delta;
130 
131  public:
132   SharedDataRelocator(address* patch_base, address* patch_end,
133                       address valid_old_base, address valid_old_end,
134                       address valid_new_base, address valid_new_end, intx delta) :
135     _patch_base(patch_base), _patch_end(patch_end),
136     _valid_old_base(valid_old_base), _valid_old_end(valid_old_end),
137     _valid_new_base(valid_new_base), _valid_new_end(valid_new_end),
138     _delta(delta) {
139     log_debug(cds, reloc)("SharedDataRelocator::_patch_base     = " PTR_FORMAT, p2i(_patch_base));
140     log_debug(cds, reloc)("SharedDataRelocator::_patch_end      = " PTR_FORMAT, p2i(_patch_end));
141     log_debug(cds, reloc)("SharedDataRelocator::_valid_old_base = " PTR_FORMAT, p2i(_valid_old_base));
142     log_debug(cds, reloc)("SharedDataRelocator::_valid_old_end  = " PTR_FORMAT, p2i(_valid_old_end));
143     log_debug(cds, reloc)("SharedDataRelocator::_valid_new_base = " PTR_FORMAT, p2i(_valid_new_base));
144     log_debug(cds, reloc)("SharedDataRelocator::_valid_new_end  = " PTR_FORMAT, p2i(_valid_new_end));
145   }
146 
147   bool do_bit(size_t offset);
148 };
149 
150 class DumpRegion {
151 private:
152   const char* _name;
153   char* _base;
154   char* _top;
155   char* _end;
156   uintx _max_delta;
157   bool _is_packed;
158   ReservedSpace* _rs;
159   VirtualSpace* _vs;
160 
161   void commit_to(char* newtop);
162 
163 public:
164   DumpRegion(const char* name, uintx max_delta = 0)
165     : _name(name), _base(nullptr), _top(nullptr), _end(nullptr),
166       _max_delta(max_delta), _is_packed(false),
167       _rs(nullptr), _vs(nullptr) {}
168 
169   char* expand_top_to(char* newtop);
170   char* allocate(size_t num_bytes, size_t alignment = 0);
171 
172   void append_intptr_t(intptr_t n, bool need_to_mark = false) NOT_CDS_RETURN;
173 
174   char* base()      const { return _base;        }
175   char* top()       const { return _top;         }
176   char* end()       const { return _end;         }
177   size_t reserved() const { return _end - _base; }
178   size_t used()     const { return _top - _base; }
179   bool is_packed()  const { return _is_packed;   }
180   bool is_allocatable() const {
181     return !is_packed() && _base != nullptr;
182   }
183 
184   void print(size_t total_bytes) const;
185   void print_out_of_space_msg(const char* failing_region, size_t needed_bytes);
186 
187   void init(ReservedSpace* rs, VirtualSpace* vs);
188 
189   void pack(DumpRegion* next = nullptr);
190 
191   bool contains(char* p) {
192     return base() <= p && p < top();
193   }
194 };
195 
196 // Closure for serializing initialization data out to a data area to be
197 // written to the shared file.
198 
199 class WriteClosure : public SerializeClosure {
200 private:
201   DumpRegion* _dump_region;
202 
203 public:
204   WriteClosure(DumpRegion* r) {
205     _dump_region = r;
206   }
207 
208   void do_ptr(void** p);
209 
210   void do_u4(u4* p) {
211     _dump_region->append_intptr_t((intptr_t)(*p));
212   }
213 
214   void do_int(int* p) {
215     _dump_region->append_intptr_t((intptr_t)(*p));
216   }
217 
218   void do_bool(bool *p) {
219     _dump_region->append_intptr_t((intptr_t)(*p));
220   }
221 
222   void do_tag(int tag) {
223     _dump_region->append_intptr_t((intptr_t)tag);
224   }
225 
226   char* region_top() {
227     return _dump_region->top();
228   }
229 
230   bool reading() const { return false; }
231 };
232 
233 // Closure for serializing initialization data in from a data area
234 // (ptr_array) read from the shared file.
235 
236 class ReadClosure : public SerializeClosure {
237 private:
238   intptr_t** _ptr_array;
239   intptr_t _base_address;
240   inline intptr_t nextPtr() {
241     return *(*_ptr_array)++;
242   }
243 
244 public:
245   ReadClosure(intptr_t** ptr_array, intptr_t base_address) :
246     _ptr_array(ptr_array), _base_address(base_address) {}
247 
248   void do_ptr(void** p);
249   void do_u4(u4* p);
250   void do_int(int* p);
251   void do_bool(bool *p);
252   void do_tag(int tag);
253   bool reading() const { return true; }
254   char* region_top() { return nullptr; }
255 };
256 
257 class ArchiveUtils {
258 public:
259   static const uintx MAX_SHARED_DELTA = 0x7FFFFFFF;
260   static void log_to_classlist(BootstrapInfo* bootstrap_specifier, TRAPS) NOT_CDS_RETURN;
261   static bool has_aot_initialized_mirror(InstanceKlass* src_ik);
262   template <typename T> static Array<T>* archive_array(GrowableArray<T>* tmp_array);
263 
264   // The following functions translate between a u4 offset and an address in the
265   // the range of the mapped CDS archive (e.g., Metaspace::is_in_shared_metaspace()).
266   // Since the first 16 bytes in this range are dummy data (see ArchiveBuilder::reserve_buffer()),
267   // we know that offset 0 never represents a valid object. As a result, an offset of 0
268   // is used to encode a nullptr.
269   //
270   // Use the "archived_address_or_null" variants if a nullptr may be encoded.
271 
272   // offset must represent an object of type T in the mapped shared space. Return
273   // a direct pointer to this object.
274   template <typename T> T static offset_to_archived_address(u4 offset) {
275     assert(offset != 0, "sanity");
276     T p = (T)(SharedBaseAddress + offset);
277     assert(Metaspace::is_in_shared_metaspace(p), "must be");
278     return p;
279   }
280 
281   template <typename T> T static offset_to_archived_address_or_null(u4 offset) {
282     if (offset == 0) {
283       return nullptr;
284     } else {
285       return offset_to_archived_address<T>(offset);
286     }
287   }
288 
289   // p must be an archived object. Get its offset from SharedBaseAddress
290   template <typename T> static u4 archived_address_to_offset(T p) {
291     uintx pn = (uintx)p;
292     uintx base = (uintx)SharedBaseAddress;
293     assert(Metaspace::is_in_shared_metaspace(p), "must be");
294     assert(pn > base, "sanity"); // No valid object is stored at 0 offset from SharedBaseAddress
295     uintx offset = pn - base;
296     assert(offset <= MAX_SHARED_DELTA, "range check");
297     return static_cast<u4>(offset);
298   }
299 
300   template <typename T> static u4 archived_address_or_null_to_offset(T p) {
301     if (p == nullptr) {
302       return 0;
303     } else {
304       return archived_address_to_offset<T>(p);
305     }
306   }
307 };
308 
309 class HeapRootSegments {
310 private:
311   size_t _base_offset;
312   size_t _count;
313   int _roots_count;
314   int _max_size_in_bytes;
315   int _max_size_in_elems;
316 
317 public:
318   size_t base_offset() { return _base_offset; }
319   size_t count() { return _count; }
320   int roots_count() { return _roots_count; }
321   int max_size_in_bytes() { return _max_size_in_bytes; }
322   int max_size_in_elems() { return _max_size_in_elems; }
323 
324   size_t size_in_bytes(size_t seg_idx);
325   int size_in_elems(size_t seg_idx);
326   size_t segment_offset(size_t seg_idx);
327 
328   // Trivial copy assignments are allowed to copy the entire object representation.
329   // We also inline this class into archive header. Therefore, it is important to make
330   // sure any gaps in object representation are initialized to zeroes. This is why
331   // constructors memset before doing field assignments.
332   HeapRootSegments() {
333     memset(this, 0, sizeof(*this));
334   }
335   HeapRootSegments(size_t base_offset, int roots_count, int max_size_in_bytes, int max_size_in_elems) {
336     memset(this, 0, sizeof(*this));
337     _base_offset = base_offset;
338     _count = (roots_count + max_size_in_elems - 1) / max_size_in_elems;
339     _roots_count = roots_count;
340     _max_size_in_bytes = max_size_in_bytes;
341     _max_size_in_elems = max_size_in_elems;
342   }
343 
344   // This class is trivially copyable and assignable.
345   HeapRootSegments(const HeapRootSegments&) = default;
346   HeapRootSegments& operator=(const HeapRootSegments&) = default;
347 };
348 
349 class ArchiveWorkers;
350 
351 // A task to be worked on by worker threads
352 class ArchiveWorkerTask : public CHeapObj<mtInternal> {
353   friend class ArchiveWorkers;
354 private:
355   const char* _name;
356   int _max_chunks;
357   volatile int _chunk;
358 
359   void run();
360 
361   void configure_max_chunks(int max_chunks);
362 
363 public:
364   ArchiveWorkerTask(const char* name) :
365       _name(name), _max_chunks(0), _chunk(0) {}
366   const char* name() const { return _name; }
367   virtual void work(int chunk, int max_chunks) = 0;
368 };
369 
370 class ArchiveWorkerThread : public NamedThread {
371   friend class ArchiveWorkers;
372 private:
373   ArchiveWorkers* const _pool;
374 
375   void post_run() override;
376 
377 public:
378   ArchiveWorkerThread(ArchiveWorkers* pool);
379   const char* type_name() const override { return "Archive Worker Thread"; }
380   void run() override;
381 };
382 
383 // Special archive workers. The goal for this implementation is to startup fast,
384 // distribute spiky workloads efficiently, and shutdown immediately after use.
385 // This makes the implementation quite different from the normal GC worker pool.
386 class ArchiveWorkers : public StackObj {
387   friend class ArchiveWorkerThread;
388 private:
389   // Target number of chunks per worker. This should be large enough to even
390   // out work imbalance, and small enough to keep bookkeeping overheads low.
391   static constexpr int CHUNKS_PER_WORKER = 4;
392   static int max_workers();
393 
394   Semaphore _end_semaphore;
395 
396   int _num_workers;
397   int _started_workers;
398   int _finish_tokens;
399 
400   typedef enum { UNUSED, WORKING, SHUTDOWN } State;
401   volatile State _state;
402 
403   ArchiveWorkerTask* _task;
404 
405   void run_as_worker();
406   void start_worker_if_needed();
407 
408   void run_task_single(ArchiveWorkerTask* task);
409   void run_task_multi(ArchiveWorkerTask* task);
410 
411   bool is_parallel();
412 
413 public:
414   ArchiveWorkers();
415   ~ArchiveWorkers();
416   void run_task(ArchiveWorkerTask* task);
417 };
418 
419 #endif // SHARE_CDS_ARCHIVEUTILS_HPP