129 bool _read_only;
130 bool _has_embedded_pointer;
131 FollowMode _follow_mode;
132 int _size_in_bytes;
133 int _id; // Each object has a unique serial ID, starting from zero. The ID is assigned
134 // when the object is added into _source_objs.
135 MetaspaceObj::Type _msotype;
136 address _source_addr; // The source object to be copied.
137 address _buffered_addr; // The copy of this object insider the buffer.
138 public:
139 SourceObjInfo(MetaspaceClosure::Ref* ref, bool read_only, FollowMode follow_mode) :
140 _ptrmap_start(0), _ptrmap_end(0), _read_only(read_only), _has_embedded_pointer(false), _follow_mode(follow_mode),
141 _size_in_bytes(ref->size() * BytesPerWord), _id(0), _msotype(ref->msotype()),
142 _source_addr(ref->obj()) {
143 if (follow_mode == point_to_it) {
144 _buffered_addr = ref->obj();
145 } else {
146 _buffered_addr = nullptr;
147 }
148 }
149
150 // This constructor is only used for regenerated objects (created by LambdaFormInvokers, etc).
151 // src = address of a Method or InstanceKlass that has been regenerated.
152 // renegerated_obj_info = info for the regenerated version of src.
153 SourceObjInfo(address src, SourceObjInfo* renegerated_obj_info) :
154 _ptrmap_start(0), _ptrmap_end(0), _read_only(false),
155 _follow_mode(renegerated_obj_info->_follow_mode),
156 _size_in_bytes(0), _msotype(renegerated_obj_info->_msotype),
157 _source_addr(src), _buffered_addr(renegerated_obj_info->_buffered_addr) {}
158
159 bool should_copy() const { return _follow_mode == make_a_copy; }
160 void set_buffered_addr(address addr) {
161 assert(should_copy(), "must be");
162 assert(_buffered_addr == nullptr, "cannot be copied twice");
163 assert(addr != nullptr, "must be a valid copy");
164 _buffered_addr = addr;
165 }
166 void set_ptrmap_start(uintx v) { _ptrmap_start = v; }
167 void set_ptrmap_end(uintx v) { _ptrmap_end = v; }
168 uintx ptrmap_start() const { return _ptrmap_start; } // inclusive
195 GrowableArray<SourceObjInfo*>* objs() const { return _objs; }
196
197 void append(SourceObjInfo* src_info);
198 void remember_embedded_pointer(SourceObjInfo* pointing_obj, MetaspaceClosure::Ref* ref);
199 void relocate(int i, ArchiveBuilder* builder);
200
201 // convenience accessor
202 SourceObjInfo* at(int i) const { return objs()->at(i); }
203 };
204
205 class CDSMapLogger;
206
207 static const int INITIAL_TABLE_SIZE = 15889;
208 static const int MAX_TABLE_SIZE = 1000000;
209
210 ReservedSpace _shared_rs;
211 VirtualSpace _shared_vs;
212
213 DumpRegion _rw_region;
214 DumpRegion _ro_region;
215
216 // Combined bitmap to track pointers in both RW and RO regions. This is updated
217 // as objects are copied into RW and RO.
218 CHeapBitMap _ptrmap;
219
220 // _ptrmap is split into these two bitmaps which are written into the archive.
221 CHeapBitMap _rw_ptrmap; // marks pointers in the RW region
222 CHeapBitMap _ro_ptrmap; // marks pointers in the RO region
223
224 SourceObjList _rw_src_objs; // objs to put in rw region
225 SourceObjList _ro_src_objs; // objs to put in ro region
226 ResizeableResourceHashtable<address, SourceObjInfo, AnyObj::C_HEAP, mtClassShared> _src_obj_table;
227 ResizeableResourceHashtable<address, address, AnyObj::C_HEAP, mtClassShared> _buffered_to_src_table;
228 GrowableArray<Klass*>* _klasses;
229 GrowableArray<Symbol*>* _symbols;
230 unsigned int _entropy_seed;
231
232 // statistics
233 DumpAllocStats _alloc_stats;
234 size_t _total_heap_region_size;
235
236 void print_region_stats(FileMapInfo *map_info, ArchiveHeapInfo* heap_info);
237 void print_bitmap_region_stats(size_t size, size_t total_size);
238 void print_heap_region_stats(ArchiveHeapInfo* heap_info, size_t total_size);
239
240 // For global access.
241 static ArchiveBuilder* _current;
242
243 public:
244 // Use this when you allocate space outside of ArchiveBuilder::dump_{rw,ro}_region.
245 // These are usually for misc tables that are allocated in the RO space.
246 class OtherROAllocMark {
247 char* _oldtop;
248 public:
249 OtherROAllocMark() {
250 _oldtop = _current->_ro_region.top();
251 }
252 ~OtherROAllocMark();
253 };
254
255 private:
256 FollowMode get_follow_mode(MetaspaceClosure::Ref *ref);
257
258 void iterate_sorted_roots(MetaspaceClosure* it);
259 void sort_klasses();
260 static int compare_symbols_by_address(Symbol** a, Symbol** b);
261 static int compare_klass_by_name(Klass** a, Klass** b);
262
263 void make_shallow_copies(DumpRegion *dump_region, const SourceObjList* src_objs);
264 void make_shallow_copy(DumpRegion *dump_region, SourceObjInfo* src_info);
265
266 void relocate_embedded_pointers(SourceObjList* src_objs);
267
268 bool is_excluded(Klass* k);
269 void clean_up_src_obj_table();
270
271 protected:
272 virtual void iterate_roots(MetaspaceClosure* it) = 0;
273
274 // Conservative estimate for number of bytes needed for:
275 size_t _estimated_metaspaceobj_bytes; // all archived MetaspaceObj's.
276 size_t _estimated_hashtable_bytes; // symbol table and dictionaries
277
278 static const int _total_dump_regions = 2;
279
280 size_t estimate_archive_size();
281
349
350 template <typename T>
351 T offset_to_buffered(u4 offset) const {
352 return (T)offset_to_buffered_address(offset);
353 }
354
355 public:
356 ArchiveBuilder();
357 ~ArchiveBuilder();
358
359 int entropy();
360 void gather_klasses_and_symbols();
361 void gather_source_objs();
362 bool gather_klass_and_symbol(MetaspaceClosure::Ref* ref, bool read_only);
363 bool gather_one_source_obj(MetaspaceClosure::Ref* ref, bool read_only);
364 void remember_embedded_pointer_in_enclosing_obj(MetaspaceClosure::Ref* ref);
365 static void serialize_dynamic_archivable_items(SerializeClosure* soc);
366
367 DumpRegion* rw_region() { return &_rw_region; }
368 DumpRegion* ro_region() { return &_ro_region; }
369
370 static char* rw_region_alloc(size_t num_bytes) {
371 return current()->rw_region()->allocate(num_bytes);
372 }
373 static char* ro_region_alloc(size_t num_bytes) {
374 return current()->ro_region()->allocate(num_bytes);
375 }
376
377 template <typename T>
378 static Array<T>* new_ro_array(int length) {
379 size_t byte_size = Array<T>::byte_sizeof(length, sizeof(T));
380 Array<T>* array = (Array<T>*)ro_region_alloc(byte_size);
381 array->initialize(length);
382 return array;
383 }
384
385 template <typename T>
386 static Array<T>* new_rw_array(int length) {
387 size_t byte_size = Array<T>::byte_sizeof(length, sizeof(T));
388 Array<T>* array = (Array<T>*)rw_region_alloc(byte_size);
389 array->initialize(length);
390 return array;
391 }
392
393 template <typename T>
394 static size_t ro_array_bytesize(int length) {
395 size_t byte_size = Array<T>::byte_sizeof(length, sizeof(T));
396 return align_up(byte_size, SharedSpaceObjectAlignment);
397 }
398
399 char* ro_strdup(const char* s);
400
401 static int compare_src_objs(SourceObjInfo** a, SourceObjInfo** b);
402 void sort_metadata_objs();
403 void dump_rw_metadata();
404 void dump_ro_metadata();
405 void relocate_metaspaceobj_embedded_pointers();
406 void record_regenerated_object(address orig_src_obj, address regen_src_obj);
407 void make_klasses_shareable();
408 void relocate_to_requested();
409 void write_archive(FileMapInfo* mapinfo, ArchiveHeapInfo* heap_info);
410 void write_region(FileMapInfo* mapinfo, int region_idx, DumpRegion* dump_region,
411 bool read_only, bool allow_exec);
412
413 void write_pointer_in_buffer(address* ptr_location, address src_addr);
414 template <typename T> void write_pointer_in_buffer(T* ptr_location, T src_addr) {
415 write_pointer_in_buffer((address*)ptr_location, (address)src_addr);
416 }
417
418 void mark_and_relocate_to_buffered_addr(address* ptr_location);
419 template <typename T> void mark_and_relocate_to_buffered_addr(T ptr_location) {
420 mark_and_relocate_to_buffered_addr((address*)ptr_location);
421 }
422
423 bool has_been_buffered(address src_addr) const;
424 template <typename T> bool has_been_buffered(T src_addr) const {
425 return has_been_buffered((address)src_addr);
426 }
427
428 address get_buffered_addr(address src_addr) const;
429 template <typename T> T get_buffered_addr(T src_addr) const {
430 return (T)get_buffered_addr((address)src_addr);
431 }
432
433 address get_source_addr(address buffered_addr) const;
434 template <typename T> T get_source_addr(T buffered_addr) const {
435 return (T)get_source_addr((address)buffered_addr);
436 }
437
438 // All klasses and symbols that will be copied into the archive
439 GrowableArray<Klass*>* klasses() const { return _klasses; }
440 GrowableArray<Symbol*>* symbols() const { return _symbols; }
441
442 static bool is_active() {
443 return (_current != nullptr);
444 }
445
446 static ArchiveBuilder* current() {
447 assert(_current != nullptr, "ArchiveBuilder must be active");
448 return _current;
449 }
450
451 static DumpAllocStats* alloc_stats() {
452 return &(current()->_alloc_stats);
453 }
454
455 static CompactHashtableStats* symbol_stats() {
456 return alloc_stats()->symbol_stats();
457 }
458
459 static CompactHashtableStats* string_stats() {
460 return alloc_stats()->string_stats();
461 }
462
463 narrowKlass get_requested_narrow_klass(Klass* k);
|
129 bool _read_only;
130 bool _has_embedded_pointer;
131 FollowMode _follow_mode;
132 int _size_in_bytes;
133 int _id; // Each object has a unique serial ID, starting from zero. The ID is assigned
134 // when the object is added into _source_objs.
135 MetaspaceObj::Type _msotype;
136 address _source_addr; // The source object to be copied.
137 address _buffered_addr; // The copy of this object insider the buffer.
138 public:
139 SourceObjInfo(MetaspaceClosure::Ref* ref, bool read_only, FollowMode follow_mode) :
140 _ptrmap_start(0), _ptrmap_end(0), _read_only(read_only), _has_embedded_pointer(false), _follow_mode(follow_mode),
141 _size_in_bytes(ref->size() * BytesPerWord), _id(0), _msotype(ref->msotype()),
142 _source_addr(ref->obj()) {
143 if (follow_mode == point_to_it) {
144 _buffered_addr = ref->obj();
145 } else {
146 _buffered_addr = nullptr;
147 }
148 }
149 SourceObjInfo(address src, address buf) {
150 _source_addr = src;
151 _buffered_addr = buf;
152 }
153
154 // This constructor is only used for regenerated objects (created by LambdaFormInvokers, etc).
155 // src = address of a Method or InstanceKlass that has been regenerated.
156 // renegerated_obj_info = info for the regenerated version of src.
157 SourceObjInfo(address src, SourceObjInfo* renegerated_obj_info) :
158 _ptrmap_start(0), _ptrmap_end(0), _read_only(false),
159 _follow_mode(renegerated_obj_info->_follow_mode),
160 _size_in_bytes(0), _msotype(renegerated_obj_info->_msotype),
161 _source_addr(src), _buffered_addr(renegerated_obj_info->_buffered_addr) {}
162
163 bool should_copy() const { return _follow_mode == make_a_copy; }
164 void set_buffered_addr(address addr) {
165 assert(should_copy(), "must be");
166 assert(_buffered_addr == nullptr, "cannot be copied twice");
167 assert(addr != nullptr, "must be a valid copy");
168 _buffered_addr = addr;
169 }
170 void set_ptrmap_start(uintx v) { _ptrmap_start = v; }
171 void set_ptrmap_end(uintx v) { _ptrmap_end = v; }
172 uintx ptrmap_start() const { return _ptrmap_start; } // inclusive
199 GrowableArray<SourceObjInfo*>* objs() const { return _objs; }
200
201 void append(SourceObjInfo* src_info);
202 void remember_embedded_pointer(SourceObjInfo* pointing_obj, MetaspaceClosure::Ref* ref);
203 void relocate(int i, ArchiveBuilder* builder);
204
205 // convenience accessor
206 SourceObjInfo* at(int i) const { return objs()->at(i); }
207 };
208
209 class CDSMapLogger;
210
211 static const int INITIAL_TABLE_SIZE = 15889;
212 static const int MAX_TABLE_SIZE = 1000000;
213
214 ReservedSpace _shared_rs;
215 VirtualSpace _shared_vs;
216
217 DumpRegion _rw_region;
218 DumpRegion _ro_region;
219 DumpRegion _cc_region;
220
221 // Combined bitmap to track pointers in both RW and RO regions. This is updated
222 // as objects are copied into RW and RO.
223 CHeapBitMap _ptrmap;
224
225 // _ptrmap is split into these two bitmaps which are written into the archive.
226 CHeapBitMap _rw_ptrmap; // marks pointers in the RW region
227 CHeapBitMap _ro_ptrmap; // marks pointers in the RO region
228 CHeapBitMap _cc_ptrmap; // marks pointers in the CC region
229
230 SourceObjList _rw_src_objs; // objs to put in rw region
231 SourceObjList _ro_src_objs; // objs to put in ro region
232 ResizeableResourceHashtable<address, SourceObjInfo, AnyObj::C_HEAP, mtClassShared> _src_obj_table;
233 ResizeableResourceHashtable<address, address, AnyObj::C_HEAP, mtClassShared> _buffered_to_src_table;
234 GrowableArray<Klass*>* _klasses;
235 GrowableArray<Symbol*>* _symbols;
236 unsigned int _entropy_seed;
237
238 // statistics
239 DumpAllocStats _alloc_stats;
240 size_t _total_heap_region_size;
241
242 void print_region_stats(FileMapInfo *map_info, ArchiveHeapInfo* heap_info);
243 void print_bitmap_region_stats(size_t size, size_t total_size);
244 void print_heap_region_stats(ArchiveHeapInfo* heap_info, size_t total_size);
245
246 // For global access.
247 static ArchiveBuilder* _current;
248
249 public:
250 // Use this when you allocate space outside of ArchiveBuilder::dump_{rw,ro}_region.
251 // These are usually for misc tables that are allocated in the RO space.
252 class OtherROAllocMark {
253 char* _oldtop;
254 public:
255 OtherROAllocMark() {
256 _oldtop = _current->_ro_region.top();
257 }
258 ~OtherROAllocMark();
259 };
260
261 private:
262 FollowMode get_follow_mode(MetaspaceClosure::Ref *ref);
263
264 void iterate_sorted_roots(MetaspaceClosure* it);
265 void sort_klasses();
266 static int compare_symbols_by_address(Symbol** a, Symbol** b);
267 static int compare_klass_by_name(Klass** a, Klass** b);
268 void update_hidden_class_loader_type(InstanceKlass* ik) NOT_CDS_JAVA_HEAP_RETURN;
269
270 void make_shallow_copies(DumpRegion *dump_region, const SourceObjList* src_objs);
271 void make_shallow_copy(DumpRegion *dump_region, SourceObjInfo* src_info);
272
273 void relocate_embedded_pointers(SourceObjList* src_objs);
274
275 bool is_excluded(Klass* k);
276 void clean_up_src_obj_table();
277
278 protected:
279 virtual void iterate_roots(MetaspaceClosure* it) = 0;
280
281 // Conservative estimate for number of bytes needed for:
282 size_t _estimated_metaspaceobj_bytes; // all archived MetaspaceObj's.
283 size_t _estimated_hashtable_bytes; // symbol table and dictionaries
284
285 static const int _total_dump_regions = 2;
286
287 size_t estimate_archive_size();
288
356
357 template <typename T>
358 T offset_to_buffered(u4 offset) const {
359 return (T)offset_to_buffered_address(offset);
360 }
361
362 public:
363 ArchiveBuilder();
364 ~ArchiveBuilder();
365
366 int entropy();
367 void gather_klasses_and_symbols();
368 void gather_source_objs();
369 bool gather_klass_and_symbol(MetaspaceClosure::Ref* ref, bool read_only);
370 bool gather_one_source_obj(MetaspaceClosure::Ref* ref, bool read_only);
371 void remember_embedded_pointer_in_enclosing_obj(MetaspaceClosure::Ref* ref);
372 static void serialize_dynamic_archivable_items(SerializeClosure* soc);
373
374 DumpRegion* rw_region() { return &_rw_region; }
375 DumpRegion* ro_region() { return &_ro_region; }
376 DumpRegion* cc_region() { return &_cc_region; }
377
378 void start_cc_region();
379 void end_cc_region();
380
381 static char* rw_region_alloc(size_t num_bytes) {
382 return current()->rw_region()->allocate(num_bytes);
383 }
384 static char* ro_region_alloc(size_t num_bytes) {
385 return current()->ro_region()->allocate(num_bytes);
386 }
387 static char* cc_region_alloc(size_t num_bytes) {
388 return current()->cc_region()->allocate(num_bytes);
389 }
390
391 template <typename T>
392 static Array<T>* new_ro_array(int length) {
393 size_t byte_size = Array<T>::byte_sizeof(length, sizeof(T));
394 Array<T>* array = (Array<T>*)ro_region_alloc(byte_size);
395 array->initialize(length);
396 return array;
397 }
398
399 template <typename T>
400 static Array<T>* new_rw_array(int length) {
401 size_t byte_size = Array<T>::byte_sizeof(length, sizeof(T));
402 Array<T>* array = (Array<T>*)rw_region_alloc(byte_size);
403 array->initialize(length);
404 return array;
405 }
406
407 template <typename T>
408 static size_t ro_array_bytesize(int length) {
409 size_t byte_size = Array<T>::byte_sizeof(length, sizeof(T));
410 return align_up(byte_size, SharedSpaceObjectAlignment);
411 }
412
413 char* ro_strdup(const char* s);
414
415 static int compare_src_objs(SourceObjInfo** a, SourceObjInfo** b);
416 void sort_metadata_objs();
417 void dump_rw_metadata();
418 void dump_ro_metadata();
419 void relocate_metaspaceobj_embedded_pointers();
420 void record_regenerated_object(address orig_src_obj, address regen_src_obj);
421 void make_klasses_shareable();
422 void make_training_data_shareable();
423 void relocate_to_requested();
424 void write_archive(FileMapInfo* mapinfo, ArchiveHeapInfo* heap_info);
425 void write_region(FileMapInfo* mapinfo, int region_idx, DumpRegion* dump_region,
426 bool read_only, bool allow_exec);
427
428 void write_pointer_in_buffer(address* ptr_location, address src_addr);
429 template <typename T> void write_pointer_in_buffer(T* ptr_location, T src_addr) {
430 write_pointer_in_buffer((address*)ptr_location, (address)src_addr);
431 }
432
433 void mark_and_relocate_to_buffered_addr(address* ptr_location);
434 template <typename T> void mark_and_relocate_to_buffered_addr(T ptr_location) {
435 mark_and_relocate_to_buffered_addr((address*)ptr_location);
436 }
437
438 bool has_been_archived(address src_addr) const;
439
440 bool has_been_buffered(address src_addr) const;
441 template <typename T> bool has_been_buffered(T src_addr) const {
442 return has_been_buffered((address)src_addr);
443 }
444 address get_buffered_addr(address src_addr) const;
445 template <typename T> T get_buffered_addr(T src_addr) const {
446 CDS_ONLY(return (T)get_buffered_addr((address)src_addr);)
447 NOT_CDS(return nullptr;)
448 }
449
450 address get_source_addr(address buffered_addr) const;
451 template <typename T> T get_source_addr(T buffered_addr) const {
452 return (T)get_source_addr((address)buffered_addr);
453 }
454
455 // All klasses and symbols that will be copied into the archive
456 GrowableArray<Klass*>* klasses() const { return _klasses; }
457 GrowableArray<Symbol*>* symbols() const { return _symbols; }
458
459 static bool is_active() {
460 CDS_ONLY(return (_current != nullptr));
461 NOT_CDS(return false;)
462 }
463
464 static ArchiveBuilder* current() {
465 assert(_current != nullptr, "ArchiveBuilder must be active");
466 return _current;
467 }
468
469 static DumpAllocStats* alloc_stats() {
470 return &(current()->_alloc_stats);
471 }
472
473 static CompactHashtableStats* symbol_stats() {
474 return alloc_stats()->symbol_stats();
475 }
476
477 static CompactHashtableStats* string_stats() {
478 return alloc_stats()->string_stats();
479 }
480
481 narrowKlass get_requested_narrow_klass(Klass* k);
|