1 /* 2 * Copyright (c) 2019, 2025, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #ifndef SHARE_CDS_ARCHIVEUTILS_HPP 26 #define SHARE_CDS_ARCHIVEUTILS_HPP 27 28 #include "cds/cds_globals.hpp" 29 #include "cds/serializeClosure.hpp" 30 #include "logging/log.hpp" 31 #include "memory/metaspace.hpp" 32 #include "memory/virtualspace.hpp" 33 #include "utilities/bitMap.hpp" 34 #include "utilities/exceptions.hpp" 35 #include "utilities/macros.hpp" 36 #include "runtime/nonJavaThread.hpp" 37 #include "runtime/semaphore.hpp" 38 39 class BootstrapInfo; 40 class ReservedSpace; 41 class VirtualSpace; 42 43 template<class E> class Array; 44 template<class E> class GrowableArray; 45 46 // ArchivePtrMarker is used to mark the location of pointers embedded in a CDS archive. E.g., when an 47 // InstanceKlass k is dumped, we mark the location of the k->_name pointer by effectively calling 48 // mark_pointer(/*ptr_loc=*/&k->_name). It's required that (_prt_base <= ptr_loc < _ptr_end). _ptr_base is 49 // fixed, but _ptr_end can be expanded as more objects are dumped. 50 class ArchivePtrMarker : AllStatic { 51 static CHeapBitMap* _ptrmap; 52 static CHeapBitMap* _rw_ptrmap; 53 static CHeapBitMap* _ro_ptrmap; 54 static CHeapBitMap* _cc_ptrmap; 55 static VirtualSpace* _vs; 56 57 // Once _ptrmap is compacted, we don't allow bit marking anymore. This is to 58 // avoid unintentional copy operations after the bitmap has been finalized and written. 59 static bool _compacted; 60 61 static address* ptr_base() { return (address*)_vs->low(); } // committed lower bound (inclusive) 62 static address* ptr_end() { return (address*)_vs->high(); } // committed upper bound (exclusive) 63 64 public: 65 static void initialize(CHeapBitMap* ptrmap, VirtualSpace* vs); 66 static void initialize_rw_ro_cc_maps(CHeapBitMap* rw_ptrmap, CHeapBitMap* ro_ptrmap, CHeapBitMap* cc_ptrmap); 67 static void mark_pointer(address* ptr_loc); 68 static void clear_pointer(address* ptr_loc); 69 static void compact(address relocatable_base, address relocatable_end); 70 static void compact(size_t max_non_null_offset); 71 72 template <typename T> 73 static void mark_pointer(T* ptr_loc) { 74 mark_pointer((address*)ptr_loc); 75 } 76 77 template <typename T> 78 static void set_and_mark_pointer(T* ptr_loc, T ptr_value) { 79 *ptr_loc = ptr_value; 80 mark_pointer(ptr_loc); 81 } 82 83 static CHeapBitMap* ptrmap() { 84 return _ptrmap; 85 } 86 87 static CHeapBitMap* rw_ptrmap() { 88 return _rw_ptrmap; 89 } 90 91 static CHeapBitMap* ro_ptrmap() { 92 return _ro_ptrmap; 93 } 94 95 static CHeapBitMap* cc_ptrmap() { 96 return _cc_ptrmap; 97 } 98 99 static void reset_map_and_vs() { 100 _ptrmap = nullptr; 101 _rw_ptrmap = nullptr; 102 _ro_ptrmap = nullptr; 103 _cc_ptrmap = nullptr; 104 _vs = nullptr; 105 } 106 }; 107 108 // SharedDataRelocator is used to shift pointers in the CDS archive. 109 // 110 // The CDS archive is basically a contiguous block of memory (divided into several regions) 111 // that contains multiple objects. The objects may contain direct pointers that point to other objects 112 // within the archive (e.g., InstanceKlass::_name points to a Symbol in the archive). During dumping, we 113 // built a bitmap that marks the locations of all these pointers (using ArchivePtrMarker, see comments above). 114 // 115 // The contents of the archive assumes that it’s mapped at the default SharedBaseAddress (e.g. 0x800000000). 116 // If the archive ends up being mapped at a different address (e.g. 0x810000000), SharedDataRelocator 117 // is used to shift each marked pointer by a delta (0x10000000 in this example), so that it points to 118 // the actually mapped location of the target object. 119 class SharedDataRelocator: public BitMapClosure { 120 // for all (address** p), where (is_marked(p) && _patch_base <= p && p < _patch_end) { *p += delta; } 121 122 // Patch all pointers within this region that are marked. 123 address* _patch_base; 124 address* _patch_end; 125 126 // Before patching, all pointers must point to this region. 127 address _valid_old_base; 128 address _valid_old_end; 129 130 // After patching, all pointers must point to this region. 131 address _valid_new_base; 132 address _valid_new_end; 133 134 // How much to relocate for each pointer. 135 intx _delta; 136 137 public: 138 SharedDataRelocator(address* patch_base, address* patch_end, 139 address valid_old_base, address valid_old_end, 140 address valid_new_base, address valid_new_end, intx delta) : 141 _patch_base(patch_base), _patch_end(patch_end), 142 _valid_old_base(valid_old_base), _valid_old_end(valid_old_end), 143 _valid_new_base(valid_new_base), _valid_new_end(valid_new_end), 144 _delta(delta) { 145 log_debug(cds, reloc)("SharedDataRelocator::_patch_base = " PTR_FORMAT, p2i(_patch_base)); 146 log_debug(cds, reloc)("SharedDataRelocator::_patch_end = " PTR_FORMAT, p2i(_patch_end)); 147 log_debug(cds, reloc)("SharedDataRelocator::_valid_old_base = " PTR_FORMAT, p2i(_valid_old_base)); 148 log_debug(cds, reloc)("SharedDataRelocator::_valid_old_end = " PTR_FORMAT, p2i(_valid_old_end)); 149 log_debug(cds, reloc)("SharedDataRelocator::_valid_new_base = " PTR_FORMAT, p2i(_valid_new_base)); 150 log_debug(cds, reloc)("SharedDataRelocator::_valid_new_end = " PTR_FORMAT, p2i(_valid_new_end)); 151 } 152 153 bool do_bit(size_t offset); 154 }; 155 156 class DumpRegion { 157 private: 158 const char* _name; 159 char* _base; 160 char* _top; 161 char* _end; 162 uintx _max_delta; 163 bool _is_packed; 164 ReservedSpace* _rs; 165 VirtualSpace* _vs; 166 167 void commit_to(char* newtop); 168 169 public: 170 DumpRegion(const char* name, uintx max_delta = 0) 171 : _name(name), _base(nullptr), _top(nullptr), _end(nullptr), 172 _max_delta(max_delta), _is_packed(false), 173 _rs(nullptr), _vs(nullptr) {} 174 175 char* expand_top_to(char* newtop); 176 char* allocate(size_t num_bytes, size_t alignment = 0); 177 178 void append_intptr_t(intptr_t n, bool need_to_mark = false) NOT_CDS_RETURN; 179 180 char* base() const { return _base; } 181 char* top() const { return _top; } 182 char* end() const { return _end; } 183 size_t reserved() const { return _end - _base; } 184 size_t used() const { return _top - _base; } 185 bool is_packed() const { return _is_packed; } 186 bool is_allocatable() const { 187 return !is_packed() && _base != nullptr; 188 } 189 190 bool is_empty() const { return _base == _top; } 191 192 void print(size_t total_bytes) const; 193 void print_out_of_space_msg(const char* failing_region, size_t needed_bytes); 194 195 void init(ReservedSpace* rs, VirtualSpace* vs); 196 197 void pack(DumpRegion* next = nullptr); 198 199 bool contains(char* p) { 200 return base() <= p && p < top(); 201 } 202 }; 203 204 // Closure for serializing initialization data out to a data area to be 205 // written to the shared file. 206 207 class WriteClosure : public SerializeClosure { 208 private: 209 DumpRegion* _dump_region; 210 211 public: 212 WriteClosure(DumpRegion* r) { 213 _dump_region = r; 214 } 215 216 void do_ptr(void** p); 217 218 void do_u4(u4* p) { 219 _dump_region->append_intptr_t((intptr_t)(*p)); 220 } 221 222 void do_int(int* p) { 223 _dump_region->append_intptr_t((intptr_t)(*p)); 224 } 225 226 void do_bool(bool *p) { 227 _dump_region->append_intptr_t((intptr_t)(*p)); 228 } 229 230 void do_tag(int tag) { 231 _dump_region->append_intptr_t((intptr_t)tag); 232 } 233 234 char* region_top() { 235 return _dump_region->top(); 236 } 237 238 bool reading() const { return false; } 239 }; 240 241 // Closure for serializing initialization data in from a data area 242 // (ptr_array) read from the shared file. 243 244 class ReadClosure : public SerializeClosure { 245 private: 246 intptr_t** _ptr_array; 247 intptr_t _base_address; 248 inline intptr_t nextPtr() { 249 return *(*_ptr_array)++; 250 } 251 252 public: 253 ReadClosure(intptr_t** ptr_array, intptr_t base_address) : 254 _ptr_array(ptr_array), _base_address(base_address) {} 255 256 void do_ptr(void** p); 257 void do_u4(u4* p); 258 void do_int(int* p); 259 void do_bool(bool *p); 260 void do_tag(int tag); 261 bool reading() const { return true; } 262 char* region_top() { return nullptr; } 263 }; 264 265 class ArchiveUtils { 266 template <typename T> static Array<T>* archive_non_ptr_array(GrowableArray<T>* tmp_array); 267 template <typename T> static Array<T>* archive_ptr_array(GrowableArray<T>* tmp_array); 268 269 public: 270 static const uintx MAX_SHARED_DELTA = 0x7FFFFFFF; 271 static void log_to_classlist(BootstrapInfo* bootstrap_specifier, TRAPS) NOT_CDS_RETURN; 272 static bool has_aot_initialized_mirror(InstanceKlass* src_ik); 273 274 template <typename T, ENABLE_IF(!std::is_pointer<T>::value)> 275 static Array<T>* archive_array(GrowableArray<T>* tmp_array) { 276 return archive_non_ptr_array(tmp_array); 277 } 278 279 template <typename T, ENABLE_IF(std::is_pointer<T>::value)> 280 static Array<T>* archive_array(GrowableArray<T>* tmp_array) { 281 return archive_ptr_array(tmp_array); 282 } 283 284 static const char* builtin_loader_name_or_null(oop loader); // "boot", "platform", "app", or nullptr 285 static const char* builtin_loader_name(oop loader); // "boot", "platform", or "app". Asserts if not a built-in-loader. 286 287 static bool builtin_loader_from_type(const char* loader_type, oop* value_ret); 288 static oop builtin_loader_from_type(int loader_type); 289 290 // The following functions translate between a u4 offset and an address in the 291 // the range of the mapped CDS archive (e.g., Metaspace::is_in_shared_metaspace()). 292 // Since the first 16 bytes in this range are dummy data (see ArchiveBuilder::reserve_buffer()), 293 // we know that offset 0 never represents a valid object. As a result, an offset of 0 294 // is used to encode a nullptr. 295 // 296 // Use the "archived_address_or_null" variants if a nullptr may be encoded. 297 298 // offset must represent an object of type T in the mapped shared space. Return 299 // a direct pointer to this object. 300 template <typename T> T static offset_to_archived_address(u4 offset) { 301 assert(offset != 0, "sanity"); 302 T p = (T)(SharedBaseAddress + offset); 303 assert(Metaspace::is_in_shared_metaspace(p), "must be"); 304 return p; 305 } 306 307 template <typename T> T static offset_to_archived_address_or_null(u4 offset) { 308 if (offset == 0) { 309 return nullptr; 310 } else { 311 return offset_to_archived_address<T>(offset); 312 } 313 } 314 315 // p must be an archived object. Get its offset from SharedBaseAddress 316 template <typename T> static u4 archived_address_to_offset(T p) { 317 uintx pn = (uintx)p; 318 uintx base = (uintx)SharedBaseAddress; 319 assert(Metaspace::is_in_shared_metaspace(p), "must be"); 320 assert(pn > base, "sanity"); // No valid object is stored at 0 offset from SharedBaseAddress 321 uintx offset = pn - base; 322 assert(offset <= MAX_SHARED_DELTA, "range check"); 323 return static_cast<u4>(offset); 324 } 325 326 template <typename T> static u4 archived_address_or_null_to_offset(T p) { 327 if (p == nullptr) { 328 return 0; 329 } else { 330 return archived_address_to_offset<T>(p); 331 } 332 } 333 }; 334 335 class HeapRootSegments { 336 private: 337 size_t _base_offset; 338 size_t _count; 339 int _roots_count; 340 int _max_size_in_bytes; 341 int _max_size_in_elems; 342 343 public: 344 size_t base_offset() { return _base_offset; } 345 size_t count() { return _count; } 346 int roots_count() { return _roots_count; } 347 int max_size_in_bytes() { return _max_size_in_bytes; } 348 int max_size_in_elems() { return _max_size_in_elems; } 349 350 size_t size_in_bytes(size_t seg_idx); 351 int size_in_elems(size_t seg_idx); 352 size_t segment_offset(size_t seg_idx); 353 354 // Trivial copy assignments are allowed to copy the entire object representation. 355 // We also inline this class into archive header. Therefore, it is important to make 356 // sure any gaps in object representation are initialized to zeroes. This is why 357 // constructors memset before doing field assignments. 358 HeapRootSegments() { 359 memset(this, 0, sizeof(*this)); 360 } 361 HeapRootSegments(size_t base_offset, int roots_count, int max_size_in_bytes, int max_size_in_elems) { 362 memset(this, 0, sizeof(*this)); 363 _base_offset = base_offset; 364 _count = (roots_count + max_size_in_elems - 1) / max_size_in_elems; 365 _roots_count = roots_count; 366 _max_size_in_bytes = max_size_in_bytes; 367 _max_size_in_elems = max_size_in_elems; 368 } 369 370 // This class is trivially copyable and assignable. 371 HeapRootSegments(const HeapRootSegments&) = default; 372 HeapRootSegments& operator=(const HeapRootSegments&) = default; 373 }; 374 375 class ArchiveWorkers; 376 377 // A task to be worked on by worker threads 378 class ArchiveWorkerTask : public CHeapObj<mtInternal> { 379 friend class ArchiveWorkers; 380 private: 381 const char* _name; 382 int _max_chunks; 383 volatile int _chunk; 384 385 void run(); 386 387 void configure_max_chunks(int max_chunks); 388 389 public: 390 ArchiveWorkerTask(const char* name) : 391 _name(name), _max_chunks(0), _chunk(0) {} 392 const char* name() const { return _name; } 393 virtual void work(int chunk, int max_chunks) = 0; 394 }; 395 396 class ArchiveWorkerThread : public NamedThread { 397 friend class ArchiveWorkers; 398 private: 399 ArchiveWorkers* const _pool; 400 401 void post_run() override; 402 403 public: 404 ArchiveWorkerThread(ArchiveWorkers* pool); 405 const char* type_name() const override { return "Archive Worker Thread"; } 406 void run() override; 407 }; 408 409 // Special archive workers. The goal for this implementation is to startup fast, 410 // distribute spiky workloads efficiently, and shutdown immediately after use. 411 // This makes the implementation quite different from the normal GC worker pool. 412 class ArchiveWorkers : public StackObj { 413 friend class ArchiveWorkerThread; 414 private: 415 // Target number of chunks per worker. This should be large enough to even 416 // out work imbalance, and small enough to keep bookkeeping overheads low. 417 static constexpr int CHUNKS_PER_WORKER = 4; 418 static int max_workers(); 419 420 Semaphore _end_semaphore; 421 422 int _num_workers; 423 int _started_workers; 424 int _finish_tokens; 425 426 typedef enum { UNUSED, WORKING, SHUTDOWN } State; 427 volatile State _state; 428 429 ArchiveWorkerTask* _task; 430 431 void run_as_worker(); 432 void start_worker_if_needed(); 433 434 void run_task_single(ArchiveWorkerTask* task); 435 void run_task_multi(ArchiveWorkerTask* task); 436 437 bool is_parallel(); 438 439 public: 440 ArchiveWorkers(); 441 ~ArchiveWorkers(); 442 void run_task(ArchiveWorkerTask* task); 443 }; 444 445 #endif // SHARE_CDS_ARCHIVEUTILS_HPP