1 /* 2 * Copyright (c) 2005, 2024, Oracle and/or its affiliates. All rights reserved. 3 * Copyright (c) 2023, Alibaba Group Holding Limited. All rights reserved. 4 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 5 * 6 * This code is free software; you can redistribute it and/or modify it 7 * under the terms of the GNU General Public License version 2 only, as 8 * published by the Free Software Foundation. 9 * 10 * This code is distributed in the hope that it will be useful, but WITHOUT 11 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 12 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 13 * version 2 for more details (a copy is included in the LICENSE file that 14 * accompanied this code). 15 * 16 * You should have received a copy of the GNU General Public License version 17 * 2 along with this work; if not, write to the Free Software Foundation, 18 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 19 * 20 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 21 * or visit www.oracle.com if you need additional information or have any 22 * questions. 23 * 24 */ 25 26 #include "precompiled.hpp" 27 #include "classfile/classLoaderData.inline.hpp" 28 #include "classfile/classLoaderDataGraph.hpp" 29 #include "classfile/javaClasses.inline.hpp" 30 #include "classfile/symbolTable.hpp" 31 #include "classfile/vmClasses.hpp" 32 #include "classfile/vmSymbols.hpp" 33 #include "gc/shared/gcLocker.hpp" 34 #include "gc/shared/gcVMOperations.hpp" 35 #include "gc/shared/workerThread.hpp" 36 #include "jfr/jfrEvents.hpp" 37 #include "jvm.h" 38 #include "memory/allocation.inline.hpp" 39 #include "memory/resourceArea.hpp" 40 #include "memory/universe.hpp" 41 #include "oops/fieldStreams.inline.hpp" 42 #include "oops/klass.inline.hpp" 43 #include "oops/objArrayKlass.hpp" 44 #include "oops/objArrayOop.inline.hpp" 45 #include "oops/flatArrayKlass.hpp" 46 #include "oops/flatArrayOop.inline.hpp" 47 #include "oops/oop.inline.hpp" 48 #include "oops/typeArrayOop.inline.hpp" 49 #include "runtime/continuationWrapper.inline.hpp" 50 #include "runtime/fieldDescriptor.inline.hpp" 51 #include "runtime/frame.inline.hpp" 52 #include "runtime/handles.inline.hpp" 53 #include "runtime/javaCalls.hpp" 54 #include "runtime/javaThread.inline.hpp" 55 #include "runtime/jniHandles.hpp" 56 #include "runtime/os.hpp" 57 #include "runtime/threads.hpp" 58 #include "runtime/threadSMR.hpp" 59 #include "runtime/vframe.hpp" 60 #include "runtime/vmOperations.hpp" 61 #include "runtime/vmThread.hpp" 62 #include "runtime/timerTrace.hpp" 63 #include "services/heapDumper.hpp" 64 #include "services/heapDumperCompression.hpp" 65 #include "services/threadService.hpp" 66 #include "utilities/checkedCast.hpp" 67 #include "utilities/macros.hpp" 68 #include "utilities/ostream.hpp" 69 #ifdef LINUX 70 #include "os_linux.hpp" 71 #endif 72 73 /* 74 * HPROF binary format - description copied from: 75 * src/share/demo/jvmti/hprof/hprof_io.c 76 * 77 * 78 * header "JAVA PROFILE 1.0.2" (0-terminated) 79 * 80 * u4 size of identifiers. Identifiers are used to represent 81 * UTF8 strings, objects, stack traces, etc. They usually 82 * have the same size as host pointers. 83 * u4 high word 84 * u4 low word number of milliseconds since 0:00 GMT, 1/1/70 85 * [record]* a sequence of records. 86 * 87 * 88 * Record format: 89 * 90 * u1 a TAG denoting the type of the record 91 * u4 number of *microseconds* since the time stamp in the 92 * header. (wraps around in a little more than an hour) 93 * u4 number of bytes *remaining* in the record. Note that 94 * this number excludes the tag and the length field itself. 95 * [u1]* BODY of the record (a sequence of bytes) 96 * 97 * 98 * The following TAGs are supported: 99 * 100 * TAG BODY notes 101 *---------------------------------------------------------- 102 * HPROF_UTF8 a UTF8-encoded name 103 * 104 * id name ID 105 * [u1]* UTF8 characters (no trailing zero) 106 * 107 * HPROF_LOAD_CLASS a newly loaded class 108 * 109 * u4 class serial number (> 0) 110 * id class object ID 111 * u4 stack trace serial number 112 * id class name ID 113 * 114 * HPROF_UNLOAD_CLASS an unloading class 115 * 116 * u4 class serial_number 117 * 118 * HPROF_FRAME a Java stack frame 119 * 120 * id stack frame ID 121 * id method name ID 122 * id method signature ID 123 * id source file name ID 124 * u4 class serial number 125 * i4 line number. >0: normal 126 * -1: unknown 127 * -2: compiled method 128 * -3: native method 129 * 130 * HPROF_TRACE a Java stack trace 131 * 132 * u4 stack trace serial number 133 * u4 thread serial number 134 * u4 number of frames 135 * [id]* stack frame IDs 136 * 137 * 138 * HPROF_ALLOC_SITES a set of heap allocation sites, obtained after GC 139 * 140 * u2 flags 0x0001: incremental vs. complete 141 * 0x0002: sorted by allocation vs. live 142 * 0x0004: whether to force a GC 143 * u4 cutoff ratio 144 * u4 total live bytes 145 * u4 total live instances 146 * u8 total bytes allocated 147 * u8 total instances allocated 148 * u4 number of sites that follow 149 * [u1 is_array: 0: normal object 150 * 2: object array 151 * 4: boolean array 152 * 5: char array 153 * 6: float array 154 * 7: double array 155 * 8: byte array 156 * 9: short array 157 * 10: int array 158 * 11: long array 159 * u4 class serial number (may be zero during startup) 160 * u4 stack trace serial number 161 * u4 number of bytes alive 162 * u4 number of instances alive 163 * u4 number of bytes allocated 164 * u4]* number of instance allocated 165 * 166 * HPROF_START_THREAD a newly started thread. 167 * 168 * u4 thread serial number (> 0) 169 * id thread object ID 170 * u4 stack trace serial number 171 * id thread name ID 172 * id thread group name ID 173 * id thread group parent name ID 174 * 175 * HPROF_END_THREAD a terminating thread. 176 * 177 * u4 thread serial number 178 * 179 * HPROF_HEAP_SUMMARY heap summary 180 * 181 * u4 total live bytes 182 * u4 total live instances 183 * u8 total bytes allocated 184 * u8 total instances allocated 185 * 186 * HPROF_HEAP_DUMP denote a heap dump 187 * 188 * [heap dump sub-records]* 189 * 190 * There are four kinds of heap dump sub-records: 191 * 192 * u1 sub-record type 193 * 194 * HPROF_GC_ROOT_UNKNOWN unknown root 195 * 196 * id object ID 197 * 198 * HPROF_GC_ROOT_THREAD_OBJ thread object 199 * 200 * id thread object ID (may be 0 for a 201 * thread newly attached through JNI) 202 * u4 thread sequence number 203 * u4 stack trace sequence number 204 * 205 * HPROF_GC_ROOT_JNI_GLOBAL JNI global ref root 206 * 207 * id object ID 208 * id JNI global ref ID 209 * 210 * HPROF_GC_ROOT_JNI_LOCAL JNI local ref 211 * 212 * id object ID 213 * u4 thread serial number 214 * u4 frame # in stack trace (-1 for empty) 215 * 216 * HPROF_GC_ROOT_JAVA_FRAME Java stack frame 217 * 218 * id object ID 219 * u4 thread serial number 220 * u4 frame # in stack trace (-1 for empty) 221 * 222 * HPROF_GC_ROOT_NATIVE_STACK Native stack 223 * 224 * id object ID 225 * u4 thread serial number 226 * 227 * HPROF_GC_ROOT_STICKY_CLASS System class 228 * 229 * id object ID 230 * 231 * HPROF_GC_ROOT_THREAD_BLOCK Reference from thread block 232 * 233 * id object ID 234 * u4 thread serial number 235 * 236 * HPROF_GC_ROOT_MONITOR_USED Busy monitor 237 * 238 * id object ID 239 * 240 * HPROF_GC_CLASS_DUMP dump of a class object 241 * 242 * id class object ID 243 * u4 stack trace serial number 244 * id super class object ID 245 * id class loader object ID 246 * id signers object ID 247 * id protection domain object ID 248 * id reserved 249 * id reserved 250 * 251 * u4 instance size (in bytes) 252 * 253 * u2 size of constant pool 254 * [u2, constant pool index, 255 * ty, type 256 * 2: object 257 * 4: boolean 258 * 5: char 259 * 6: float 260 * 7: double 261 * 8: byte 262 * 9: short 263 * 10: int 264 * 11: long 265 * vl]* and value 266 * 267 * u2 number of static fields 268 * [id, static field name, 269 * ty, type, 270 * vl]* and value 271 * 272 * u2 number of inst. fields (not inc. super) 273 * [id, instance field name, 274 * ty]* type 275 * 276 * HPROF_GC_INSTANCE_DUMP dump of a normal object 277 * 278 * id object ID 279 * u4 stack trace serial number 280 * id class object ID 281 * u4 number of bytes that follow 282 * [vl]* instance field values (class, followed 283 * by super, super's super ...) 284 * 285 * HPROF_GC_OBJ_ARRAY_DUMP dump of an object array 286 * 287 * id array object ID 288 * u4 stack trace serial number 289 * u4 number of elements 290 * id array class ID 291 * [id]* elements 292 * 293 * HPROF_GC_PRIM_ARRAY_DUMP dump of a primitive array 294 * 295 * id array object ID 296 * u4 stack trace serial number 297 * u4 number of elements 298 * u1 element type 299 * 4: boolean array 300 * 5: char array 301 * 6: float array 302 * 7: double array 303 * 8: byte array 304 * 9: short array 305 * 10: int array 306 * 11: long array 307 * [u1]* elements 308 * 309 * HPROF_CPU_SAMPLES a set of sample traces of running threads 310 * 311 * u4 total number of samples 312 * u4 # of traces 313 * [u4 # of samples 314 * u4]* stack trace serial number 315 * 316 * HPROF_CONTROL_SETTINGS the settings of on/off switches 317 * 318 * u4 0x00000001: alloc traces on/off 319 * 0x00000002: cpu sampling on/off 320 * u2 stack trace depth 321 * 322 * HPROF_FLAT_ARRAYS list of flat arrays 323 * 324 * [flat array sub-records]* 325 * 326 * HPROF_FLAT_ARRAY flat array 327 * 328 * id array object ID (dumped as HPROF_GC_PRIM_ARRAY_DUMP) 329 * id element class ID (dumped by HPROF_GC_CLASS_DUMP) 330 * 331 * HPROF_INLINED_FIELDS decribes inlined fields 332 * 333 * [class with inlined fields sub-records]* 334 * 335 * HPROF_CLASS_WITH_INLINED_FIELDS 336 * 337 * id class ID (dumped as HPROF_GC_CLASS_DUMP) 338 * 339 * u2 number of instance inlined fields (not including super) 340 * [u2, inlined field index, 341 * u2, synthetic field count, 342 * id, original field name, 343 * id]* inlined field class ID (dumped by HPROF_GC_CLASS_DUMP) 344 * 345 * When the header is "JAVA PROFILE 1.0.2" a heap dump can optionally 346 * be generated as a sequence of heap dump segments. This sequence is 347 * terminated by an end record. The additional tags allowed by format 348 * "JAVA PROFILE 1.0.2" are: 349 * 350 * HPROF_HEAP_DUMP_SEGMENT denote a heap dump segment 351 * 352 * [heap dump sub-records]* 353 * The same sub-record types allowed by HPROF_HEAP_DUMP 354 * 355 * HPROF_HEAP_DUMP_END denotes the end of a heap dump 356 * 357 */ 358 359 360 // HPROF tags 361 362 enum hprofTag : u1 { 363 // top-level records 364 HPROF_UTF8 = 0x01, 365 HPROF_LOAD_CLASS = 0x02, 366 HPROF_UNLOAD_CLASS = 0x03, 367 HPROF_FRAME = 0x04, 368 HPROF_TRACE = 0x05, 369 HPROF_ALLOC_SITES = 0x06, 370 HPROF_HEAP_SUMMARY = 0x07, 371 HPROF_START_THREAD = 0x0A, 372 HPROF_END_THREAD = 0x0B, 373 HPROF_HEAP_DUMP = 0x0C, 374 HPROF_CPU_SAMPLES = 0x0D, 375 HPROF_CONTROL_SETTINGS = 0x0E, 376 377 // 1.0.2 record types 378 HPROF_HEAP_DUMP_SEGMENT = 0x1C, 379 HPROF_HEAP_DUMP_END = 0x2C, 380 381 // inlined object support 382 HPROF_FLAT_ARRAYS = 0x12, 383 HPROF_INLINED_FIELDS = 0x13, 384 // inlined object subrecords 385 HPROF_FLAT_ARRAY = 0x01, 386 HPROF_CLASS_WITH_INLINED_FIELDS = 0x01, 387 388 // field types 389 HPROF_ARRAY_OBJECT = 0x01, 390 HPROF_NORMAL_OBJECT = 0x02, 391 HPROF_BOOLEAN = 0x04, 392 HPROF_CHAR = 0x05, 393 HPROF_FLOAT = 0x06, 394 HPROF_DOUBLE = 0x07, 395 HPROF_BYTE = 0x08, 396 HPROF_SHORT = 0x09, 397 HPROF_INT = 0x0A, 398 HPROF_LONG = 0x0B, 399 400 // data-dump sub-records 401 HPROF_GC_ROOT_UNKNOWN = 0xFF, 402 HPROF_GC_ROOT_JNI_GLOBAL = 0x01, 403 HPROF_GC_ROOT_JNI_LOCAL = 0x02, 404 HPROF_GC_ROOT_JAVA_FRAME = 0x03, 405 HPROF_GC_ROOT_NATIVE_STACK = 0x04, 406 HPROF_GC_ROOT_STICKY_CLASS = 0x05, 407 HPROF_GC_ROOT_THREAD_BLOCK = 0x06, 408 HPROF_GC_ROOT_MONITOR_USED = 0x07, 409 HPROF_GC_ROOT_THREAD_OBJ = 0x08, 410 HPROF_GC_CLASS_DUMP = 0x20, 411 HPROF_GC_INSTANCE_DUMP = 0x21, 412 HPROF_GC_OBJ_ARRAY_DUMP = 0x22, 413 HPROF_GC_PRIM_ARRAY_DUMP = 0x23 414 }; 415 416 // Default stack trace ID (used for dummy HPROF_TRACE record) 417 enum { 418 STACK_TRACE_ID = 1, 419 INITIAL_CLASS_COUNT = 200 420 }; 421 422 423 class AbstractDumpWriter; 424 425 class InlinedObjects { 426 427 struct ClassInlinedFields { 428 const Klass *klass; 429 uintx base_index; // base index of the inlined field names (1st field has index base_index+1). 430 ClassInlinedFields(const Klass *klass = nullptr, uintx base_index = 0) : klass(klass), base_index(base_index) {} 431 432 // For GrowableArray::find_sorted(). 433 static int compare(const ClassInlinedFields& a, const ClassInlinedFields& b) { 434 return a.klass - b.klass; 435 } 436 // For GrowableArray::sort(). 437 static int compare(ClassInlinedFields* a, ClassInlinedFields* b) { 438 return compare(*a, *b); 439 } 440 }; 441 442 uintx _min_string_id; 443 uintx _max_string_id; 444 445 GrowableArray<ClassInlinedFields> *_inlined_field_map; 446 447 // counters for classes with inlined fields and for the fields 448 int _classes_count; 449 int _inlined_fields_count; 450 451 static InlinedObjects *_instance; 452 453 static void inlined_field_names_callback(InlinedObjects* _this, const Klass *klass, uintx base_index, int count); 454 455 GrowableArray<oop> *_flat_arrays; 456 457 public: 458 InlinedObjects() 459 : _min_string_id(0), _max_string_id(0), 460 _inlined_field_map(nullptr), 461 _classes_count(0), _inlined_fields_count(0), 462 _flat_arrays(nullptr) { 463 } 464 465 static InlinedObjects* get_instance() { 466 return _instance; 467 } 468 469 void init(); 470 void release(); 471 472 void dump_inlined_field_names(AbstractDumpWriter *writer); 473 474 uintx get_base_index_for(Klass* k); 475 uintx get_next_string_id(uintx id); 476 477 void dump_classed_with_inlined_fields(AbstractDumpWriter* writer); 478 479 void add_flat_array(oop array); 480 void dump_flat_arrays(AbstractDumpWriter* writer); 481 482 }; 483 484 InlinedObjects *InlinedObjects::_instance = nullptr; 485 486 487 // Supports I/O operations for a dump 488 // Base class for dump and parallel dump 489 class AbstractDumpWriter : public CHeapObj<mtInternal> { 490 protected: 491 enum { 492 io_buffer_max_size = 1*M, 493 dump_segment_header_size = 9 494 }; 495 496 char* _buffer; // internal buffer 497 size_t _size; 498 size_t _pos; 499 500 bool _in_dump_segment; // Are we currently in a dump segment? 501 bool _is_huge_sub_record; // Are we writing a sub-record larger than the buffer size? 502 DEBUG_ONLY(size_t _sub_record_left;) // The bytes not written for the current sub-record. 503 DEBUG_ONLY(bool _sub_record_ended;) // True if we have called the end_sub_record(). 504 505 char* buffer() const { return _buffer; } 506 size_t buffer_size() const { return _size; } 507 void set_position(size_t pos) { _pos = pos; } 508 509 // Can be called if we have enough room in the buffer. 510 void write_fast(const void* s, size_t len); 511 512 // Returns true if we have enough room in the buffer for 'len' bytes. 513 bool can_write_fast(size_t len); 514 515 void write_address(address a); 516 517 public: 518 AbstractDumpWriter() : 519 _buffer(nullptr), 520 _size(io_buffer_max_size), 521 _pos(0), 522 _in_dump_segment(false) { } 523 524 // Total number of bytes written to the disk 525 virtual julong bytes_written() const = 0; 526 // Return non-null if error occurred 527 virtual char const* error() const = 0; 528 529 size_t position() const { return _pos; } 530 // writer functions 531 virtual void write_raw(const void* s, size_t len); 532 void write_u1(u1 x); 533 void write_u2(u2 x); 534 void write_u4(u4 x); 535 void write_u8(u8 x); 536 void write_objectID(oop o); 537 void write_rootID(oop* p); 538 void write_symbolID(Symbol* o); 539 void write_classID(Klass* k); 540 void write_id(u4 x); 541 542 // Start a new sub-record. Starts a new heap dump segment if needed. 543 void start_sub_record(u1 tag, u4 len); 544 // Ends the current sub-record. 545 void end_sub_record(); 546 // Finishes the current dump segment if not already finished. 547 void finish_dump_segment(); 548 // Flush internal buffer to persistent storage 549 virtual void flush() = 0; 550 }; 551 552 void AbstractDumpWriter::write_fast(const void* s, size_t len) { 553 assert(!_in_dump_segment || (_sub_record_left >= len), "sub-record too large"); 554 assert(buffer_size() - position() >= len, "Must fit"); 555 debug_only(_sub_record_left -= len); 556 memcpy(buffer() + position(), s, len); 557 set_position(position() + len); 558 } 559 560 bool AbstractDumpWriter::can_write_fast(size_t len) { 561 return buffer_size() - position() >= len; 562 } 563 564 // write raw bytes 565 void AbstractDumpWriter::write_raw(const void* s, size_t len) { 566 assert(!_in_dump_segment || (_sub_record_left >= len), "sub-record too large"); 567 debug_only(_sub_record_left -= len); 568 569 // flush buffer to make room. 570 while (len > buffer_size() - position()) { 571 assert(!_in_dump_segment || _is_huge_sub_record, 572 "Cannot overflow in non-huge sub-record."); 573 size_t to_write = buffer_size() - position(); 574 memcpy(buffer() + position(), s, to_write); 575 s = (void*) ((char*) s + to_write); 576 len -= to_write; 577 set_position(position() + to_write); 578 flush(); 579 } 580 581 memcpy(buffer() + position(), s, len); 582 set_position(position() + len); 583 } 584 585 // Makes sure we inline the fast write into the write_u* functions. This is a big speedup. 586 #define WRITE_KNOWN_TYPE(p, len) do { if (can_write_fast((len))) write_fast((p), (len)); \ 587 else write_raw((p), (len)); } while (0) 588 589 void AbstractDumpWriter::write_u1(u1 x) { 590 WRITE_KNOWN_TYPE(&x, 1); 591 } 592 593 void AbstractDumpWriter::write_u2(u2 x) { 594 u2 v; 595 Bytes::put_Java_u2((address)&v, x); 596 WRITE_KNOWN_TYPE(&v, 2); 597 } 598 599 void AbstractDumpWriter::write_u4(u4 x) { 600 u4 v; 601 Bytes::put_Java_u4((address)&v, x); 602 WRITE_KNOWN_TYPE(&v, 4); 603 } 604 605 void AbstractDumpWriter::write_u8(u8 x) { 606 u8 v; 607 Bytes::put_Java_u8((address)&v, x); 608 WRITE_KNOWN_TYPE(&v, 8); 609 } 610 611 void AbstractDumpWriter::write_address(address a) { 612 #ifdef _LP64 613 write_u8((u8)a); 614 #else 615 write_u4((u4)a); 616 #endif 617 } 618 619 void AbstractDumpWriter::write_objectID(oop o) { 620 write_address(cast_from_oop<address>(o)); 621 } 622 623 void AbstractDumpWriter::write_rootID(oop* p) { 624 write_address((address)p); 625 } 626 627 void AbstractDumpWriter::write_symbolID(Symbol* s) { 628 write_address((address)((uintptr_t)s)); 629 } 630 631 void AbstractDumpWriter::write_id(u4 x) { 632 #ifdef _LP64 633 write_u8((u8) x); 634 #else 635 write_u4(x); 636 #endif 637 } 638 639 // We use java mirror as the class ID 640 void AbstractDumpWriter::write_classID(Klass* k) { 641 write_objectID(k->java_mirror()); 642 } 643 644 void AbstractDumpWriter::finish_dump_segment() { 645 if (_in_dump_segment) { 646 assert(_sub_record_left == 0, "Last sub-record not written completely"); 647 assert(_sub_record_ended, "sub-record must have ended"); 648 649 // Fix up the dump segment length if we haven't written a huge sub-record last 650 // (in which case the segment length was already set to the correct value initially). 651 if (!_is_huge_sub_record) { 652 assert(position() > dump_segment_header_size, "Dump segment should have some content"); 653 Bytes::put_Java_u4((address) (buffer() + 5), 654 (u4) (position() - dump_segment_header_size)); 655 } else { 656 // Finish process huge sub record 657 // Set _is_huge_sub_record to false so the parallel dump writer can flush data to file. 658 _is_huge_sub_record = false; 659 } 660 661 _in_dump_segment = false; 662 flush(); 663 } 664 } 665 666 void AbstractDumpWriter::start_sub_record(u1 tag, u4 len) { 667 if (!_in_dump_segment) { 668 if (position() > 0) { 669 flush(); 670 } 671 672 assert(position() == 0 && buffer_size() > dump_segment_header_size, "Must be at the start"); 673 674 write_u1(HPROF_HEAP_DUMP_SEGMENT); 675 write_u4(0); // timestamp 676 // Will be fixed up later if we add more sub-records. If this is a huge sub-record, 677 // this is already the correct length, since we don't add more sub-records. 678 write_u4(len); 679 assert(Bytes::get_Java_u4((address)(buffer() + 5)) == len, "Inconsistent size!"); 680 _in_dump_segment = true; 681 _is_huge_sub_record = len > buffer_size() - dump_segment_header_size; 682 } else if (_is_huge_sub_record || (len > buffer_size() - position())) { 683 // This object will not fit in completely or the last sub-record was huge. 684 // Finish the current segment and try again. 685 finish_dump_segment(); 686 start_sub_record(tag, len); 687 688 return; 689 } 690 691 debug_only(_sub_record_left = len); 692 debug_only(_sub_record_ended = false); 693 694 write_u1(tag); 695 } 696 697 void AbstractDumpWriter::end_sub_record() { 698 assert(_in_dump_segment, "must be in dump segment"); 699 assert(_sub_record_left == 0, "sub-record not written completely"); 700 assert(!_sub_record_ended, "Must not have ended yet"); 701 debug_only(_sub_record_ended = true); 702 } 703 704 // Supports I/O operations for a dump 705 706 class DumpWriter : public AbstractDumpWriter { 707 private: 708 FileWriter* _writer; 709 AbstractCompressor* _compressor; 710 size_t _bytes_written; 711 char* _error; 712 // Compression support 713 char* _out_buffer; 714 size_t _out_size; 715 size_t _out_pos; 716 char* _tmp_buffer; 717 size_t _tmp_size; 718 719 private: 720 void do_compress(); 721 722 public: 723 DumpWriter(const char* path, bool overwrite, AbstractCompressor* compressor); 724 ~DumpWriter(); 725 julong bytes_written() const override { return (julong) _bytes_written; } 726 void set_bytes_written(julong bytes_written) { _bytes_written = bytes_written; } 727 char const* error() const override { return _error; } 728 void set_error(const char* error) { _error = (char*)error; } 729 bool has_error() const { return _error != nullptr; } 730 const char* get_file_path() const { return _writer->get_file_path(); } 731 AbstractCompressor* compressor() { return _compressor; } 732 void set_compressor(AbstractCompressor* p) { _compressor = p; } 733 bool is_overwrite() const { return _writer->is_overwrite(); } 734 int get_fd() const { return _writer->get_fd(); } 735 736 void flush() override; 737 }; 738 739 DumpWriter::DumpWriter(const char* path, bool overwrite, AbstractCompressor* compressor) : 740 AbstractDumpWriter(), 741 _writer(new (std::nothrow) FileWriter(path, overwrite)), 742 _compressor(compressor), 743 _bytes_written(0), 744 _error(nullptr), 745 _out_buffer(nullptr), 746 _out_size(0), 747 _out_pos(0), 748 _tmp_buffer(nullptr), 749 _tmp_size(0) { 750 _error = (char*)_writer->open_writer(); 751 if (_error == nullptr) { 752 _buffer = (char*)os::malloc(io_buffer_max_size, mtInternal); 753 if (compressor != nullptr) { 754 _error = (char*)_compressor->init(io_buffer_max_size, &_out_size, &_tmp_size); 755 if (_error == nullptr) { 756 if (_out_size > 0) { 757 _out_buffer = (char*)os::malloc(_out_size, mtInternal); 758 } 759 if (_tmp_size > 0) { 760 _tmp_buffer = (char*)os::malloc(_tmp_size, mtInternal); 761 } 762 } 763 } 764 } 765 // initialize internal buffer 766 _pos = 0; 767 _size = io_buffer_max_size; 768 } 769 770 DumpWriter::~DumpWriter(){ 771 if (_buffer != nullptr) { 772 os::free(_buffer); 773 } 774 if (_out_buffer != nullptr) { 775 os::free(_out_buffer); 776 } 777 if (_tmp_buffer != nullptr) { 778 os::free(_tmp_buffer); 779 } 780 if (_writer != nullptr) { 781 delete _writer; 782 } 783 _bytes_written = -1; 784 } 785 786 // flush any buffered bytes to the file 787 void DumpWriter::flush() { 788 if (_pos <= 0) { 789 return; 790 } 791 if (has_error()) { 792 _pos = 0; 793 return; 794 } 795 char* result = nullptr; 796 if (_compressor == nullptr) { 797 result = (char*)_writer->write_buf(_buffer, _pos); 798 _bytes_written += _pos; 799 } else { 800 do_compress(); 801 if (!has_error()) { 802 result = (char*)_writer->write_buf(_out_buffer, _out_pos); 803 _bytes_written += _out_pos; 804 } 805 } 806 _pos = 0; // reset pos to make internal buffer available 807 808 if (result != nullptr) { 809 set_error(result); 810 } 811 } 812 813 void DumpWriter::do_compress() { 814 const char* msg = _compressor->compress(_buffer, _pos, _out_buffer, _out_size, 815 _tmp_buffer, _tmp_size, &_out_pos); 816 817 if (msg != nullptr) { 818 set_error(msg); 819 } 820 } 821 822 class DumperClassCacheTable; 823 class DumperClassCacheTableEntry; 824 825 // Support class with a collection of functions used when dumping the heap 826 class DumperSupport : AllStatic { 827 public: 828 829 // write a header of the given type 830 static void write_header(AbstractDumpWriter* writer, hprofTag tag, u4 len); 831 832 // returns hprof tag for the given type signature 833 static hprofTag sig2tag(Symbol* sig); 834 // returns hprof tag for the given basic type 835 static hprofTag type2tag(BasicType type); 836 // Returns the size of the data to write. 837 static u4 sig2size(Symbol* sig); 838 839 // calculates the total size of the all fields of the given class. 840 static u4 instance_size(InstanceKlass* ik, DumperClassCacheTableEntry* class_cache_entry = nullptr); 841 842 // dump a jfloat 843 static void dump_float(AbstractDumpWriter* writer, jfloat f); 844 // dump a jdouble 845 static void dump_double(AbstractDumpWriter* writer, jdouble d); 846 // dumps the raw value of the given field 847 static void dump_field_value(AbstractDumpWriter* writer, char type, oop obj, int offset); 848 // returns the size of the static fields; also counts the static fields 849 static u4 get_static_fields_size(InstanceKlass* ik, u2& field_count); 850 // dumps static fields of the given class 851 static void dump_static_fields(AbstractDumpWriter* writer, Klass* k); 852 // dump the raw values of the instance fields of the given identity or inlined object; 853 // for identity objects offset is 0 and 'klass' is o->klass(), 854 // for inlined objects offset is the offset in the holder object, 'klass' is inlined object class 855 static void dump_instance_fields(AbstractDumpWriter* writer, oop o, int offset, DumperClassCacheTable* class_cache, DumperClassCacheTableEntry* class_cache_entry); 856 // dump the raw values of the instance fields of the given inlined object; 857 // dump_instance_fields wrapper for inlined objects 858 static void dump_inlined_object_fields(AbstractDumpWriter* writer, oop o, int offset, DumperClassCacheTable* class_cache, DumperClassCacheTableEntry* class_cache_entry); 859 860 // get the count of the instance fields for a given class 861 static u2 get_instance_fields_count(InstanceKlass* ik); 862 // dumps the definition of the instance fields for a given class 863 static void dump_instance_field_descriptors(AbstractDumpWriter* writer, InstanceKlass* k, uintx *inlined_fields_index = nullptr); 864 // creates HPROF_GC_INSTANCE_DUMP record for the given object 865 static void dump_instance(AbstractDumpWriter* writer, oop o, DumperClassCacheTable* class_cache); 866 // creates HPROF_GC_CLASS_DUMP record for the given instance class 867 static void dump_instance_class(AbstractDumpWriter* writer, Klass* k); 868 // creates HPROF_GC_CLASS_DUMP record for a given array class 869 static void dump_array_class(AbstractDumpWriter* writer, Klass* k); 870 871 // creates HPROF_GC_OBJ_ARRAY_DUMP record for the given object array 872 static void dump_object_array(AbstractDumpWriter* writer, objArrayOop array); 873 // creates HPROF_GC_PRIM_ARRAY_DUMP record for the given flat array 874 static void dump_flat_array(AbstractDumpWriter* writer, flatArrayOop array, DumperClassCacheTable* class_cache); 875 // creates HPROF_GC_PRIM_ARRAY_DUMP record for the given type array 876 static void dump_prim_array(AbstractDumpWriter* writer, typeArrayOop array); 877 // create HPROF_FRAME record for the given method and bci 878 static void dump_stack_frame(AbstractDumpWriter* writer, int frame_serial_num, int class_serial_num, Method* m, int bci); 879 880 // check if we need to truncate an array 881 static int calculate_array_max_length(AbstractDumpWriter* writer, arrayOop array, short header_size); 882 // extended version to dump flat arrays as primitive arrays; 883 // type_size specifies size of the inlined objects. 884 static int calculate_array_max_length(AbstractDumpWriter* writer, arrayOop array, int type_size, short header_size); 885 886 // fixes up the current dump record and writes HPROF_HEAP_DUMP_END record 887 static void end_of_dump(AbstractDumpWriter* writer); 888 889 static oop mask_dormant_archived_object(oop o, oop ref_obj) { 890 if (o != nullptr && o->klass()->java_mirror_no_keepalive() == nullptr) { 891 // Ignore this object since the corresponding java mirror is not loaded. 892 // Might be a dormant archive object. 893 report_dormant_archived_object(o, ref_obj); 894 return nullptr; 895 } else { 896 return o; 897 } 898 } 899 900 // helper methods for inlined fields. 901 static bool is_inlined_field(const fieldDescriptor& fld) { 902 return fld.is_flat(); 903 } 904 static InlineKlass* get_inlined_field_klass(const fieldDescriptor& fld) { 905 assert(is_inlined_field(fld), "must be inlined field"); 906 InstanceKlass* holder_klass = fld.field_holder(); 907 return InlineKlass::cast(holder_klass->get_inline_type_field_klass(fld.index())); 908 } 909 910 static void report_dormant_archived_object(oop o, oop ref_obj) { 911 if (log_is_enabled(Trace, cds, heap)) { 912 ResourceMark rm; 913 if (ref_obj != nullptr) { 914 log_trace(cds, heap)("skipped dormant archived object " INTPTR_FORMAT " (%s) referenced by " INTPTR_FORMAT " (%s)", 915 p2i(o), o->klass()->external_name(), 916 p2i(ref_obj), ref_obj->klass()->external_name()); 917 } else { 918 log_trace(cds, heap)("skipped dormant archived object " INTPTR_FORMAT " (%s)", 919 p2i(o), o->klass()->external_name()); 920 } 921 } 922 } 923 }; 924 925 // Hash table of klasses to the klass metadata. This should greatly improve the 926 // hash dumping performance. This hash table is supposed to be used by a single 927 // thread only. 928 // 929 class DumperClassCacheTableEntry : public CHeapObj<mtServiceability> { 930 friend class DumperClassCacheTable; 931 private: 932 GrowableArray<char> _sigs_start; 933 GrowableArray<int> _offsets; 934 GrowableArray<InlineKlass*> _inline_klasses; 935 u4 _instance_size; 936 int _entries; 937 938 public: 939 DumperClassCacheTableEntry() : _instance_size(0), _entries(0) {}; 940 941 int field_count() { return _entries; } 942 char sig_start(int field_idx) { return _sigs_start.at(field_idx); } 943 void push_sig_start_inlined() { _sigs_start.push('Q'); } 944 bool is_inlined(int field_idx){ return _sigs_start.at(field_idx) == 'Q'; } 945 InlineKlass* inline_klass(int field_idx) { assert(is_inlined(field_idx), "Not inlined"); return _inline_klasses.at(field_idx); } 946 int offset(int field_idx) { return _offsets.at(field_idx); } 947 u4 instance_size() { return _instance_size; } 948 }; 949 950 class DumperClassCacheTable { 951 private: 952 // ResourceHashtable SIZE is specified at compile time so we 953 // use 1031 which is the first prime after 1024. 954 static constexpr size_t TABLE_SIZE = 1031; 955 956 // Maintain the cache for N classes. This limits memory footprint 957 // impact, regardless of how many classes we have in the dump. 958 // This also improves look up performance by keeping the statically 959 // sized table from overloading. 960 static constexpr int CACHE_TOP = 256; 961 962 typedef ResourceHashtable<InstanceKlass*, DumperClassCacheTableEntry*, 963 TABLE_SIZE, AnyObj::C_HEAP, mtServiceability> PtrTable; 964 PtrTable* _ptrs; 965 966 // Single-slot cache to handle the major case of objects of the same 967 // class back-to-back, e.g. from T[]. 968 InstanceKlass* _last_ik; 969 DumperClassCacheTableEntry* _last_entry; 970 971 void unlink_all(PtrTable* table) { 972 class CleanupEntry: StackObj { 973 public: 974 bool do_entry(InstanceKlass*& key, DumperClassCacheTableEntry*& entry) { 975 delete entry; 976 return true; 977 } 978 } cleanup; 979 table->unlink(&cleanup); 980 } 981 982 public: 983 DumperClassCacheTableEntry* lookup_or_create(InstanceKlass* ik) { 984 if (_last_ik == ik) { 985 return _last_entry; 986 } 987 988 DumperClassCacheTableEntry* entry; 989 DumperClassCacheTableEntry** from_cache = _ptrs->get(ik); 990 if (from_cache == nullptr) { 991 entry = new DumperClassCacheTableEntry(); 992 for (HierarchicalFieldStream<JavaFieldStream> fld(ik); !fld.done(); fld.next()) { 993 if (!fld.access_flags().is_static()) { 994 InlineKlass* inlineKlass = nullptr; 995 if (DumperSupport::is_inlined_field(fld.field_descriptor())) { 996 inlineKlass = DumperSupport::get_inlined_field_klass(fld.field_descriptor()); 997 entry->push_sig_start_inlined(); 998 entry->_instance_size += DumperSupport::instance_size(inlineKlass); 999 } else { 1000 Symbol* sig = fld.signature(); 1001 entry->_sigs_start.push(sig->char_at(0)); 1002 entry->_instance_size += DumperSupport::sig2size(sig); 1003 } 1004 entry->_inline_klasses.push(inlineKlass); 1005 entry->_offsets.push(fld.offset()); 1006 entry->_entries++; 1007 } 1008 } 1009 1010 if (_ptrs->number_of_entries() >= CACHE_TOP) { 1011 // We do not track the individual hit rates for table entries. 1012 // Purge the entire table, and let the cache catch up with new 1013 // distribution. 1014 unlink_all(_ptrs); 1015 } 1016 1017 _ptrs->put(ik, entry); 1018 } else { 1019 entry = *from_cache; 1020 } 1021 1022 // Remember for single-slot cache. 1023 _last_ik = ik; 1024 _last_entry = entry; 1025 1026 return entry; 1027 } 1028 1029 DumperClassCacheTable() : _ptrs(new (mtServiceability) PtrTable), _last_ik(nullptr), _last_entry(nullptr) {} 1030 1031 ~DumperClassCacheTable() { 1032 unlink_all(_ptrs); 1033 delete _ptrs; 1034 } 1035 }; 1036 1037 // write a header of the given type 1038 void DumperSupport:: write_header(AbstractDumpWriter* writer, hprofTag tag, u4 len) { 1039 writer->write_u1(tag); 1040 writer->write_u4(0); // current ticks 1041 writer->write_u4(len); 1042 } 1043 1044 // returns hprof tag for the given type signature 1045 hprofTag DumperSupport::sig2tag(Symbol* sig) { 1046 switch (sig->char_at(0)) { 1047 case JVM_SIGNATURE_CLASS : return HPROF_NORMAL_OBJECT; 1048 case JVM_SIGNATURE_ARRAY : return HPROF_NORMAL_OBJECT; 1049 case JVM_SIGNATURE_BYTE : return HPROF_BYTE; 1050 case JVM_SIGNATURE_CHAR : return HPROF_CHAR; 1051 case JVM_SIGNATURE_FLOAT : return HPROF_FLOAT; 1052 case JVM_SIGNATURE_DOUBLE : return HPROF_DOUBLE; 1053 case JVM_SIGNATURE_INT : return HPROF_INT; 1054 case JVM_SIGNATURE_LONG : return HPROF_LONG; 1055 case JVM_SIGNATURE_SHORT : return HPROF_SHORT; 1056 case JVM_SIGNATURE_BOOLEAN : return HPROF_BOOLEAN; 1057 default : ShouldNotReachHere(); /* to shut up compiler */ return HPROF_BYTE; 1058 } 1059 } 1060 1061 hprofTag DumperSupport::type2tag(BasicType type) { 1062 switch (type) { 1063 case T_BYTE : return HPROF_BYTE; 1064 case T_CHAR : return HPROF_CHAR; 1065 case T_FLOAT : return HPROF_FLOAT; 1066 case T_DOUBLE : return HPROF_DOUBLE; 1067 case T_INT : return HPROF_INT; 1068 case T_LONG : return HPROF_LONG; 1069 case T_SHORT : return HPROF_SHORT; 1070 case T_BOOLEAN : return HPROF_BOOLEAN; 1071 default : ShouldNotReachHere(); /* to shut up compiler */ return HPROF_BYTE; 1072 } 1073 } 1074 1075 u4 DumperSupport::sig2size(Symbol* sig) { 1076 switch (sig->char_at(0)) { 1077 case JVM_SIGNATURE_CLASS: 1078 case JVM_SIGNATURE_ARRAY: return sizeof(address); 1079 case JVM_SIGNATURE_BOOLEAN: 1080 case JVM_SIGNATURE_BYTE: return 1; 1081 case JVM_SIGNATURE_SHORT: 1082 case JVM_SIGNATURE_CHAR: return 2; 1083 case JVM_SIGNATURE_INT: 1084 case JVM_SIGNATURE_FLOAT: return 4; 1085 case JVM_SIGNATURE_LONG: 1086 case JVM_SIGNATURE_DOUBLE: return 8; 1087 default: ShouldNotReachHere(); /* to shut up compiler */ return 0; 1088 } 1089 } 1090 1091 template<typename T, typename F> T bit_cast(F from) { // replace with the real thing when we can use c++20 1092 T to; 1093 static_assert(sizeof(to) == sizeof(from), "must be of the same size"); 1094 memcpy(&to, &from, sizeof(to)); 1095 return to; 1096 } 1097 1098 // dump a jfloat 1099 void DumperSupport::dump_float(AbstractDumpWriter* writer, jfloat f) { 1100 if (g_isnan(f)) { 1101 writer->write_u4(0x7fc00000); // collapsing NaNs 1102 } else { 1103 writer->write_u4(bit_cast<u4>(f)); 1104 } 1105 } 1106 1107 // dump a jdouble 1108 void DumperSupport::dump_double(AbstractDumpWriter* writer, jdouble d) { 1109 if (g_isnan(d)) { 1110 writer->write_u8(0x7ff80000ull << 32); // collapsing NaNs 1111 } else { 1112 writer->write_u8(bit_cast<u8>(d)); 1113 } 1114 } 1115 1116 1117 // dumps the raw value of the given field 1118 void DumperSupport::dump_field_value(AbstractDumpWriter* writer, char type, oop obj, int offset) { 1119 switch (type) { 1120 case JVM_SIGNATURE_CLASS : 1121 case JVM_SIGNATURE_ARRAY : { 1122 oop o = obj->obj_field_access<ON_UNKNOWN_OOP_REF | AS_NO_KEEPALIVE>(offset); 1123 o = mask_dormant_archived_object(o, obj); 1124 assert(oopDesc::is_oop_or_null(o), "Expected an oop or nullptr at " PTR_FORMAT, p2i(o)); 1125 writer->write_objectID(o); 1126 break; 1127 } 1128 case JVM_SIGNATURE_BYTE : { 1129 jbyte b = obj->byte_field(offset); 1130 writer->write_u1(b); 1131 break; 1132 } 1133 case JVM_SIGNATURE_CHAR : { 1134 jchar c = obj->char_field(offset); 1135 writer->write_u2(c); 1136 break; 1137 } 1138 case JVM_SIGNATURE_SHORT : { 1139 jshort s = obj->short_field(offset); 1140 writer->write_u2(s); 1141 break; 1142 } 1143 case JVM_SIGNATURE_FLOAT : { 1144 jfloat f = obj->float_field(offset); 1145 dump_float(writer, f); 1146 break; 1147 } 1148 case JVM_SIGNATURE_DOUBLE : { 1149 jdouble d = obj->double_field(offset); 1150 dump_double(writer, d); 1151 break; 1152 } 1153 case JVM_SIGNATURE_INT : { 1154 jint i = obj->int_field(offset); 1155 writer->write_u4(i); 1156 break; 1157 } 1158 case JVM_SIGNATURE_LONG : { 1159 jlong l = obj->long_field(offset); 1160 writer->write_u8(l); 1161 break; 1162 } 1163 case JVM_SIGNATURE_BOOLEAN : { 1164 jboolean b = obj->bool_field(offset); 1165 writer->write_u1(b); 1166 break; 1167 } 1168 default : { 1169 ShouldNotReachHere(); 1170 break; 1171 } 1172 } 1173 } 1174 1175 // calculates the total size of the all fields of the given class. 1176 u4 DumperSupport::instance_size(InstanceKlass* ik, DumperClassCacheTableEntry* class_cache_entry) { 1177 if (class_cache_entry != nullptr) { 1178 return class_cache_entry->instance_size(); 1179 } else { 1180 u4 size = 0; 1181 for (HierarchicalFieldStream<JavaFieldStream> fld(ik); !fld.done(); fld.next()) { 1182 if (!fld.access_flags().is_static()) { 1183 if (is_inlined_field(fld.field_descriptor())) { 1184 size += instance_size(get_inlined_field_klass(fld.field_descriptor())); 1185 } else { 1186 size += sig2size(fld.signature()); 1187 } 1188 } 1189 } 1190 return size; 1191 } 1192 } 1193 1194 u4 DumperSupport::get_static_fields_size(InstanceKlass* ik, u2& field_count) { 1195 field_count = 0; 1196 u4 size = 0; 1197 1198 for (JavaFieldStream fldc(ik); !fldc.done(); fldc.next()) { 1199 if (fldc.access_flags().is_static()) { 1200 assert(!is_inlined_field(fldc.field_descriptor()), "static fields cannot be inlined"); 1201 1202 field_count++; 1203 size += sig2size(fldc.signature()); 1204 } 1205 } 1206 1207 // Add in resolved_references which is referenced by the cpCache 1208 // The resolved_references is an array per InstanceKlass holding the 1209 // strings and other oops resolved from the constant pool. 1210 oop resolved_references = ik->constants()->resolved_references_or_null(); 1211 if (resolved_references != nullptr) { 1212 field_count++; 1213 size += sizeof(address); 1214 1215 // Add in the resolved_references of the used previous versions of the class 1216 // in the case of RedefineClasses 1217 InstanceKlass* prev = ik->previous_versions(); 1218 while (prev != nullptr && prev->constants()->resolved_references_or_null() != nullptr) { 1219 field_count++; 1220 size += sizeof(address); 1221 prev = prev->previous_versions(); 1222 } 1223 } 1224 1225 // We write the value itself plus a name and a one byte type tag per field. 1226 return checked_cast<u4>(size + field_count * (sizeof(address) + 1)); 1227 } 1228 1229 // dumps static fields of the given class 1230 void DumperSupport::dump_static_fields(AbstractDumpWriter* writer, Klass* k) { 1231 InstanceKlass* ik = InstanceKlass::cast(k); 1232 1233 // dump the field descriptors and raw values 1234 for (JavaFieldStream fld(ik); !fld.done(); fld.next()) { 1235 if (fld.access_flags().is_static()) { 1236 assert(!is_inlined_field(fld.field_descriptor()), "static fields cannot be inlined"); 1237 1238 Symbol* sig = fld.signature(); 1239 1240 writer->write_symbolID(fld.name()); // name 1241 writer->write_u1(sig2tag(sig)); // type 1242 1243 // value 1244 dump_field_value(writer, sig->char_at(0), ik->java_mirror(), fld.offset()); 1245 } 1246 } 1247 1248 // Add resolved_references for each class that has them 1249 oop resolved_references = ik->constants()->resolved_references_or_null(); 1250 if (resolved_references != nullptr) { 1251 writer->write_symbolID(vmSymbols::resolved_references_name()); // name 1252 writer->write_u1(sig2tag(vmSymbols::object_array_signature())); // type 1253 writer->write_objectID(resolved_references); 1254 1255 // Also write any previous versions 1256 InstanceKlass* prev = ik->previous_versions(); 1257 while (prev != nullptr && prev->constants()->resolved_references_or_null() != nullptr) { 1258 writer->write_symbolID(vmSymbols::resolved_references_name()); // name 1259 writer->write_u1(sig2tag(vmSymbols::object_array_signature())); // type 1260 writer->write_objectID(prev->constants()->resolved_references()); 1261 prev = prev->previous_versions(); 1262 } 1263 } 1264 } 1265 1266 // dump the raw values of the instance fields of the given identity or inlined object; 1267 // for identity objects offset is 0 and 'klass' is o->klass(), 1268 // for inlined objects offset is the offset in the holder object, 'klass' is inlined object class. 1269 void DumperSupport::dump_instance_fields(AbstractDumpWriter* writer, oop o, int offset, DumperClassCacheTable* class_cache, DumperClassCacheTableEntry* class_cache_entry) { 1270 assert(class_cache_entry != nullptr, "Pre-condition: must be provided"); 1271 for (int idx = 0; idx < class_cache_entry->field_count(); idx++) { 1272 if (class_cache_entry->is_inlined(idx)) { 1273 InlineKlass* field_klass = class_cache_entry->inline_klass(idx); 1274 int fields_offset = offset + (class_cache_entry->offset(idx) - field_klass->first_field_offset()); 1275 DumperClassCacheTableEntry* inline_class_cache_entry = class_cache->lookup_or_create(field_klass); 1276 dump_inlined_object_fields(writer, o, fields_offset, class_cache, inline_class_cache_entry); 1277 } else { 1278 dump_field_value(writer, class_cache_entry->sig_start(idx), o, class_cache_entry->offset(idx)); 1279 } 1280 } 1281 } 1282 1283 void DumperSupport::dump_inlined_object_fields(AbstractDumpWriter* writer, oop o, int offset, DumperClassCacheTable* class_cache, DumperClassCacheTableEntry* class_cache_entry) { 1284 // the object is inlined, so all its fields are stored without headers. 1285 dump_instance_fields(writer, o, offset, class_cache, class_cache_entry); 1286 } 1287 1288 // gets the count of the instance fields for a given class 1289 u2 DumperSupport::get_instance_fields_count(InstanceKlass* ik) { 1290 u2 field_count = 0; 1291 1292 for (JavaFieldStream fldc(ik); !fldc.done(); fldc.next()) { 1293 if (!fldc.access_flags().is_static()) { 1294 if (is_inlined_field(fldc.field_descriptor())) { 1295 // add "synthetic" fields for inlined fields. 1296 field_count += get_instance_fields_count(get_inlined_field_klass(fldc.field_descriptor())); 1297 } else { 1298 field_count++; 1299 } 1300 } 1301 } 1302 1303 return field_count; 1304 } 1305 1306 // dumps the definition of the instance fields for a given class 1307 // inlined_fields_id is not-nullptr for inlined fields (to get synthetic field name IDs 1308 // by using InlinedObjects::get_next_string_id()). 1309 void DumperSupport::dump_instance_field_descriptors(AbstractDumpWriter* writer, InstanceKlass* ik, uintx* inlined_fields_id) { 1310 // inlined_fields_id != nullptr means ik is a class of inlined field. 1311 // Inlined field id pointer for this class; lazyly initialized 1312 // if the class has inlined field(s) and the caller didn't provide inlined_fields_id. 1313 uintx *this_klass_inlined_fields_id = inlined_fields_id; 1314 uintx inlined_id = 0; 1315 1316 // dump the field descriptors 1317 for (JavaFieldStream fld(ik); !fld.done(); fld.next()) { 1318 if (!fld.access_flags().is_static()) { 1319 if (is_inlined_field(fld.field_descriptor())) { 1320 // dump "synthetic" fields for inlined fields. 1321 if (this_klass_inlined_fields_id == nullptr) { 1322 inlined_id = InlinedObjects::get_instance()->get_base_index_for(ik); 1323 this_klass_inlined_fields_id = &inlined_id; 1324 } 1325 dump_instance_field_descriptors(writer, get_inlined_field_klass(fld.field_descriptor()), this_klass_inlined_fields_id); 1326 } else { 1327 Symbol* sig = fld.signature(); 1328 Symbol* name = nullptr; 1329 // Use inlined_fields_id provided by caller. 1330 if (inlined_fields_id != nullptr) { 1331 uintx name_id = InlinedObjects::get_instance()->get_next_string_id(*inlined_fields_id); 1332 1333 // name_id == 0 is returned on error. use original field signature. 1334 if (name_id != 0) { 1335 *inlined_fields_id = name_id; 1336 name = reinterpret_cast<Symbol*>(name_id); 1337 } 1338 } 1339 if (name == nullptr) { 1340 name = fld.name(); 1341 } 1342 1343 writer->write_symbolID(name); // name 1344 writer->write_u1(sig2tag(sig)); // type 1345 } 1346 } 1347 } 1348 } 1349 1350 // creates HPROF_GC_INSTANCE_DUMP record for the given object 1351 void DumperSupport::dump_instance(AbstractDumpWriter* writer, oop o, DumperClassCacheTable* class_cache) { 1352 InstanceKlass* ik = InstanceKlass::cast(o->klass()); 1353 1354 DumperClassCacheTableEntry* cache_entry = class_cache->lookup_or_create(ik); 1355 1356 u4 is = instance_size(ik, cache_entry); 1357 u4 size = 1 + sizeof(address) + 4 + sizeof(address) + 4 + is; 1358 1359 writer->start_sub_record(HPROF_GC_INSTANCE_DUMP, size); 1360 writer->write_objectID(o); 1361 writer->write_u4(STACK_TRACE_ID); 1362 1363 // class ID 1364 writer->write_classID(ik); 1365 1366 // number of bytes that follow 1367 writer->write_u4(is); 1368 1369 // field values 1370 dump_instance_fields(writer, o, 0, class_cache, cache_entry); 1371 1372 writer->end_sub_record(); 1373 } 1374 1375 // creates HPROF_GC_CLASS_DUMP record for the given instance class 1376 void DumperSupport::dump_instance_class(AbstractDumpWriter* writer, Klass* k) { 1377 InstanceKlass* ik = InstanceKlass::cast(k); 1378 1379 // We can safepoint and do a heap dump at a point where we have a Klass, 1380 // but no java mirror class has been setup for it. So we need to check 1381 // that the class is at least loaded, to avoid crash from a null mirror. 1382 if (!ik->is_loaded()) { 1383 return; 1384 } 1385 1386 u2 static_fields_count = 0; 1387 u4 static_size = get_static_fields_size(ik, static_fields_count); 1388 u2 instance_fields_count = get_instance_fields_count(ik); 1389 u4 instance_fields_size = instance_fields_count * (sizeof(address) + 1); 1390 u4 size = checked_cast<u4>(1 + sizeof(address) + 4 + 6 * sizeof(address) + 4 + 2 + 2 + static_size + 2 + instance_fields_size); 1391 1392 writer->start_sub_record(HPROF_GC_CLASS_DUMP, size); 1393 1394 // class ID 1395 writer->write_classID(ik); 1396 writer->write_u4(STACK_TRACE_ID); 1397 1398 // super class ID 1399 InstanceKlass* java_super = ik->java_super(); 1400 if (java_super == nullptr) { 1401 writer->write_objectID(oop(nullptr)); 1402 } else { 1403 writer->write_classID(java_super); 1404 } 1405 1406 writer->write_objectID(ik->class_loader()); 1407 writer->write_objectID(ik->signers()); 1408 writer->write_objectID(ik->protection_domain()); 1409 1410 // reserved 1411 writer->write_objectID(oop(nullptr)); 1412 writer->write_objectID(oop(nullptr)); 1413 1414 // instance size 1415 writer->write_u4(HeapWordSize * ik->size_helper()); 1416 1417 // size of constant pool - ignored by HAT 1.1 1418 writer->write_u2(0); 1419 1420 // static fields 1421 writer->write_u2(static_fields_count); 1422 dump_static_fields(writer, ik); 1423 1424 // description of instance fields 1425 writer->write_u2(instance_fields_count); 1426 dump_instance_field_descriptors(writer, ik); 1427 1428 writer->end_sub_record(); 1429 } 1430 1431 // creates HPROF_GC_CLASS_DUMP record for the given array class 1432 void DumperSupport::dump_array_class(AbstractDumpWriter* writer, Klass* k) { 1433 InstanceKlass* ik = nullptr; // bottom class for object arrays, null for primitive type arrays 1434 if (k->is_objArray_klass()) { 1435 Klass *bk = ObjArrayKlass::cast(k)->bottom_klass(); 1436 assert(bk != nullptr, "checking"); 1437 if (bk->is_instance_klass()) { 1438 ik = InstanceKlass::cast(bk); 1439 } 1440 } 1441 1442 u4 size = 1 + sizeof(address) + 4 + 6 * sizeof(address) + 4 + 2 + 2 + 2; 1443 writer->start_sub_record(HPROF_GC_CLASS_DUMP, size); 1444 writer->write_classID(k); 1445 writer->write_u4(STACK_TRACE_ID); 1446 1447 // super class of array classes is java.lang.Object 1448 InstanceKlass* java_super = k->java_super(); 1449 assert(java_super != nullptr, "checking"); 1450 writer->write_classID(java_super); 1451 1452 writer->write_objectID(ik == nullptr ? oop(nullptr) : ik->class_loader()); 1453 writer->write_objectID(ik == nullptr ? oop(nullptr) : ik->signers()); 1454 writer->write_objectID(ik == nullptr ? oop(nullptr) : ik->protection_domain()); 1455 1456 writer->write_objectID(oop(nullptr)); // reserved 1457 writer->write_objectID(oop(nullptr)); 1458 writer->write_u4(0); // instance size 1459 writer->write_u2(0); // constant pool 1460 writer->write_u2(0); // static fields 1461 writer->write_u2(0); // instance fields 1462 1463 writer->end_sub_record(); 1464 1465 } 1466 1467 // Hprof uses an u4 as record length field, 1468 // which means we need to truncate arrays that are too long. 1469 int DumperSupport::calculate_array_max_length(AbstractDumpWriter* writer, arrayOop array, int type_size, short header_size) { 1470 int length = array->length(); 1471 1472 size_t length_in_bytes = (size_t)length * type_size; 1473 uint max_bytes = max_juint - header_size; 1474 1475 if (length_in_bytes > max_bytes) { 1476 length = max_bytes / type_size; 1477 length_in_bytes = (size_t)length * type_size; 1478 1479 BasicType type = ArrayKlass::cast(array->klass())->element_type(); 1480 warning("cannot dump array of type %s[] with length %d; truncating to length %d", 1481 type2name_tab[type], array->length(), length); 1482 } 1483 return length; 1484 } 1485 1486 int DumperSupport::calculate_array_max_length(AbstractDumpWriter* writer, arrayOop array, short header_size) { 1487 BasicType type = ArrayKlass::cast(array->klass())->element_type(); 1488 assert((type >= T_BOOLEAN && type <= T_OBJECT) || type == T_PRIMITIVE_OBJECT, "invalid array element type"); 1489 int type_size; 1490 if (type == T_OBJECT) { 1491 type_size = sizeof(address); 1492 } else if (type == T_PRIMITIVE_OBJECT) { 1493 // TODO: FIXME 1494 fatal("Not supported yet"); // FIXME: JDK-8325678 1495 } else { 1496 type_size = type2aelembytes(type); 1497 } 1498 1499 return calculate_array_max_length(writer, array, type_size, header_size); 1500 } 1501 1502 // creates HPROF_GC_OBJ_ARRAY_DUMP record for the given object array 1503 void DumperSupport::dump_object_array(AbstractDumpWriter* writer, objArrayOop array) { 1504 // sizeof(u1) + 2 * sizeof(u4) + sizeof(objectID) + sizeof(classID) 1505 short header_size = 1 + 2 * 4 + 2 * sizeof(address); 1506 int length = calculate_array_max_length(writer, array, header_size); 1507 u4 size = checked_cast<u4>(header_size + length * sizeof(address)); 1508 1509 writer->start_sub_record(HPROF_GC_OBJ_ARRAY_DUMP, size); 1510 writer->write_objectID(array); 1511 writer->write_u4(STACK_TRACE_ID); 1512 writer->write_u4(length); 1513 1514 // array class ID 1515 writer->write_classID(array->klass()); 1516 1517 // [id]* elements 1518 for (int index = 0; index < length; index++) { 1519 oop o = array->obj_at(index); 1520 o = mask_dormant_archived_object(o, array); 1521 writer->write_objectID(o); 1522 } 1523 1524 writer->end_sub_record(); 1525 } 1526 1527 // creates HPROF_GC_PRIM_ARRAY_DUMP record for the given flat array 1528 void DumperSupport::dump_flat_array(AbstractDumpWriter* writer, flatArrayOop array, DumperClassCacheTable* class_cache) { 1529 FlatArrayKlass* array_klass = FlatArrayKlass::cast(array->klass()); 1530 InlineKlass* element_klass = array_klass->element_klass(); 1531 int element_size = instance_size(element_klass); 1532 /* id array object ID 1533 * u4 stack trace serial number 1534 * u4 number of elements 1535 * u1 element type 1536 */ 1537 short header_size = 1 + sizeof(address) + 2 * 4 + 1; 1538 1539 // TODO: use T_SHORT/T_INT/T_LONG if needed to avoid truncation 1540 BasicType type = T_BYTE; 1541 int type_size = type2aelembytes(type); 1542 int length = calculate_array_max_length(writer, array, element_size, header_size); 1543 u4 length_in_bytes = (u4)(length * element_size); 1544 u4 size = header_size + length_in_bytes; 1545 1546 writer->start_sub_record(HPROF_GC_PRIM_ARRAY_DUMP, size); 1547 writer->write_objectID(array); 1548 writer->write_u4(STACK_TRACE_ID); 1549 // TODO: round up array length for T_SHORT/T_INT/T_LONG 1550 writer->write_u4(length * element_size); 1551 writer->write_u1(type2tag(type)); 1552 1553 for (int index = 0; index < length; index++) { 1554 // need offset in the holder to read inlined object. calculate it from flatArrayOop::value_at_addr() 1555 int offset = (int)((address)array->value_at_addr(index, array_klass->layout_helper()) 1556 - cast_from_oop<address>(array)); 1557 DumperClassCacheTableEntry* class_cache_entry = class_cache->lookup_or_create(element_klass); 1558 dump_inlined_object_fields(writer, array, offset, class_cache, class_cache_entry); 1559 } 1560 1561 // TODO: write padding bytes for T_SHORT/T_INT/T_LONG 1562 1563 InlinedObjects::get_instance()->add_flat_array(array); 1564 1565 writer->end_sub_record(); 1566 } 1567 1568 #define WRITE_ARRAY(Array, Type, Size, Length) \ 1569 for (int i = 0; i < Length; i++) { writer->write_##Size((Size)Array->Type##_at(i)); } 1570 1571 // creates HPROF_GC_PRIM_ARRAY_DUMP record for the given type array 1572 void DumperSupport::dump_prim_array(AbstractDumpWriter* writer, typeArrayOop array) { 1573 BasicType type = TypeArrayKlass::cast(array->klass())->element_type(); 1574 // 2 * sizeof(u1) + 2 * sizeof(u4) + sizeof(objectID) 1575 short header_size = 2 * 1 + 2 * 4 + sizeof(address); 1576 1577 int length = calculate_array_max_length(writer, array, header_size); 1578 int type_size = type2aelembytes(type); 1579 u4 length_in_bytes = (u4)length * type_size; 1580 u4 size = header_size + length_in_bytes; 1581 1582 writer->start_sub_record(HPROF_GC_PRIM_ARRAY_DUMP, size); 1583 writer->write_objectID(array); 1584 writer->write_u4(STACK_TRACE_ID); 1585 writer->write_u4(length); 1586 writer->write_u1(type2tag(type)); 1587 1588 // nothing to copy 1589 if (length == 0) { 1590 writer->end_sub_record(); 1591 return; 1592 } 1593 1594 // If the byte ordering is big endian then we can copy most types directly 1595 1596 switch (type) { 1597 case T_INT : { 1598 if (Endian::is_Java_byte_ordering_different()) { 1599 WRITE_ARRAY(array, int, u4, length); 1600 } else { 1601 writer->write_raw(array->int_at_addr(0), length_in_bytes); 1602 } 1603 break; 1604 } 1605 case T_BYTE : { 1606 writer->write_raw(array->byte_at_addr(0), length_in_bytes); 1607 break; 1608 } 1609 case T_CHAR : { 1610 if (Endian::is_Java_byte_ordering_different()) { 1611 WRITE_ARRAY(array, char, u2, length); 1612 } else { 1613 writer->write_raw(array->char_at_addr(0), length_in_bytes); 1614 } 1615 break; 1616 } 1617 case T_SHORT : { 1618 if (Endian::is_Java_byte_ordering_different()) { 1619 WRITE_ARRAY(array, short, u2, length); 1620 } else { 1621 writer->write_raw(array->short_at_addr(0), length_in_bytes); 1622 } 1623 break; 1624 } 1625 case T_BOOLEAN : { 1626 if (Endian::is_Java_byte_ordering_different()) { 1627 WRITE_ARRAY(array, bool, u1, length); 1628 } else { 1629 writer->write_raw(array->bool_at_addr(0), length_in_bytes); 1630 } 1631 break; 1632 } 1633 case T_LONG : { 1634 if (Endian::is_Java_byte_ordering_different()) { 1635 WRITE_ARRAY(array, long, u8, length); 1636 } else { 1637 writer->write_raw(array->long_at_addr(0), length_in_bytes); 1638 } 1639 break; 1640 } 1641 1642 // handle float/doubles in a special value to ensure than NaNs are 1643 // written correctly. TO DO: Check if we can avoid this on processors that 1644 // use IEEE 754. 1645 1646 case T_FLOAT : { 1647 for (int i = 0; i < length; i++) { 1648 dump_float(writer, array->float_at(i)); 1649 } 1650 break; 1651 } 1652 case T_DOUBLE : { 1653 for (int i = 0; i < length; i++) { 1654 dump_double(writer, array->double_at(i)); 1655 } 1656 break; 1657 } 1658 default : ShouldNotReachHere(); 1659 } 1660 1661 writer->end_sub_record(); 1662 } 1663 1664 // create a HPROF_FRAME record of the given Method* and bci 1665 void DumperSupport::dump_stack_frame(AbstractDumpWriter* writer, 1666 int frame_serial_num, 1667 int class_serial_num, 1668 Method* m, 1669 int bci) { 1670 int line_number; 1671 if (m->is_native()) { 1672 line_number = -3; // native frame 1673 } else { 1674 line_number = m->line_number_from_bci(bci); 1675 } 1676 1677 write_header(writer, HPROF_FRAME, 4*oopSize + 2*sizeof(u4)); 1678 writer->write_id(frame_serial_num); // frame serial number 1679 writer->write_symbolID(m->name()); // method's name 1680 writer->write_symbolID(m->signature()); // method's signature 1681 1682 assert(m->method_holder()->is_instance_klass(), "not InstanceKlass"); 1683 writer->write_symbolID(m->method_holder()->source_file_name()); // source file name 1684 writer->write_u4(class_serial_num); // class serial number 1685 writer->write_u4((u4) line_number); // line number 1686 } 1687 1688 1689 class InlinedFieldNameDumper : public LockedClassesDo { 1690 public: 1691 typedef void (*Callback)(InlinedObjects *owner, const Klass *klass, uintx base_index, int count); 1692 1693 private: 1694 AbstractDumpWriter* _writer; 1695 InlinedObjects *_owner; 1696 Callback _callback; 1697 uintx _index; 1698 1699 void dump_inlined_field_names(GrowableArray<Symbol*>* super_names, Symbol* field_name, InlineKlass* klass) { 1700 super_names->push(field_name); 1701 for (HierarchicalFieldStream<JavaFieldStream> fld(klass); !fld.done(); fld.next()) { 1702 if (!fld.access_flags().is_static()) { 1703 if (DumperSupport::is_inlined_field(fld.field_descriptor())) { 1704 dump_inlined_field_names(super_names, fld.name(), DumperSupport::get_inlined_field_klass(fld.field_descriptor())); 1705 } else { 1706 // get next string ID. 1707 uintx next_index = _owner->get_next_string_id(_index); 1708 if (next_index == 0) { 1709 // something went wrong (overflow?) 1710 // stop generation; the rest of inlined objects will have original field names. 1711 return; 1712 } 1713 _index = next_index; 1714 1715 // Calculate length. 1716 int len = fld.name()->utf8_length(); 1717 for (GrowableArrayIterator<Symbol*> it = super_names->begin(); it != super_names->end(); ++it) { 1718 len += (*it)->utf8_length() + 1; // +1 for ".". 1719 } 1720 1721 DumperSupport::write_header(_writer, HPROF_UTF8, oopSize + len); 1722 _writer->write_symbolID(reinterpret_cast<Symbol*>(_index)); 1723 // Write the string value. 1724 // 1) super_names. 1725 for (GrowableArrayIterator<Symbol*> it = super_names->begin(); it != super_names->end(); ++it) { 1726 _writer->write_raw((*it)->bytes(), (*it)->utf8_length()); 1727 _writer->write_u1('.'); 1728 } 1729 // 2) field name. 1730 _writer->write_raw(fld.name()->bytes(), fld.name()->utf8_length()); 1731 } 1732 } 1733 } 1734 super_names->pop(); 1735 } 1736 1737 void dump_inlined_field_names(Symbol* field_name, InlineKlass* field_klass) { 1738 GrowableArray<Symbol*> super_names(4, mtServiceability); 1739 dump_inlined_field_names(&super_names, field_name, field_klass); 1740 } 1741 1742 public: 1743 InlinedFieldNameDumper(AbstractDumpWriter* writer, InlinedObjects* owner, Callback callback) 1744 : _writer(writer), _owner(owner), _callback(callback), _index(0) { 1745 } 1746 1747 void do_klass(Klass* k) { 1748 if (!k->is_instance_klass()) { 1749 return; 1750 } 1751 InstanceKlass* ik = InstanceKlass::cast(k); 1752 // if (ik->has_inline_type_fields()) { 1753 // return; 1754 // } 1755 1756 uintx base_index = _index; 1757 int count = 0; 1758 1759 for (HierarchicalFieldStream<JavaFieldStream> fld(ik); !fld.done(); fld.next()) { 1760 if (!fld.access_flags().is_static()) { 1761 if (DumperSupport::is_inlined_field(fld.field_descriptor())) { 1762 dump_inlined_field_names(fld.name(), DumperSupport::get_inlined_field_klass(fld.field_descriptor())); 1763 count++; 1764 } 1765 } 1766 } 1767 1768 if (count != 0) { 1769 _callback(_owner, k, base_index, count); 1770 } 1771 } 1772 }; 1773 1774 class InlinedFieldsDumper : public LockedClassesDo { 1775 private: 1776 AbstractDumpWriter* _writer; 1777 1778 public: 1779 InlinedFieldsDumper(AbstractDumpWriter* writer) : _writer(writer) {} 1780 1781 void do_klass(Klass* k) { 1782 if (!k->is_instance_klass()) { 1783 return; 1784 } 1785 InstanceKlass* ik = InstanceKlass::cast(k); 1786 // if (ik->has_inline_type_fields()) { 1787 // return; 1788 // } 1789 1790 // We can be at a point where java mirror does not exist yet. 1791 // So we need to check that the class is at least loaded, to avoid crash from a null mirror. 1792 if (!ik->is_loaded()) { 1793 return; 1794 } 1795 1796 u2 inlined_count = 0; 1797 for (HierarchicalFieldStream<JavaFieldStream> fld(ik); !fld.done(); fld.next()) { 1798 if (!fld.access_flags().is_static()) { 1799 if (DumperSupport::is_inlined_field(fld.field_descriptor())) { 1800 inlined_count++; 1801 } 1802 } 1803 } 1804 if (inlined_count != 0) { 1805 _writer->write_u1(HPROF_CLASS_WITH_INLINED_FIELDS); 1806 1807 // class ID 1808 _writer->write_classID(ik); 1809 // number of inlined fields 1810 _writer->write_u2(inlined_count); 1811 u2 index = 0; 1812 for (HierarchicalFieldStream<JavaFieldStream> fld(ik); !fld.done(); fld.next()) { 1813 if (!fld.access_flags().is_static()) { 1814 if (DumperSupport::is_inlined_field(fld.field_descriptor())) { 1815 // inlined field index 1816 _writer->write_u2(index); 1817 // synthetic field count 1818 u2 field_count = DumperSupport::get_instance_fields_count(DumperSupport::get_inlined_field_klass(fld.field_descriptor())); 1819 _writer->write_u2(field_count); 1820 // original field name 1821 _writer->write_symbolID(fld.name()); 1822 // inlined field class ID 1823 _writer->write_classID(DumperSupport::get_inlined_field_klass(fld.field_descriptor())); 1824 1825 index += field_count; 1826 } else { 1827 index++; 1828 } 1829 } 1830 } 1831 } 1832 } 1833 }; 1834 1835 1836 void InlinedObjects::init() { 1837 _instance = this; 1838 1839 struct Closure : public SymbolClosure { 1840 uintx _min_id = max_uintx; 1841 uintx _max_id = 0; 1842 Closure() : _min_id(max_uintx), _max_id(0) {} 1843 1844 void do_symbol(Symbol** p) { 1845 uintx val = reinterpret_cast<uintx>(*p); 1846 if (val < _min_id) { 1847 _min_id = val; 1848 } 1849 if (val > _max_id) { 1850 _max_id = val; 1851 } 1852 } 1853 } closure; 1854 1855 SymbolTable::symbols_do(&closure); 1856 1857 _min_string_id = closure._min_id; 1858 _max_string_id = closure._max_id; 1859 } 1860 1861 void InlinedObjects::release() { 1862 _instance = nullptr; 1863 1864 if (_inlined_field_map != nullptr) { 1865 delete _inlined_field_map; 1866 _inlined_field_map = nullptr; 1867 } 1868 if (_flat_arrays != nullptr) { 1869 delete _flat_arrays; 1870 _flat_arrays = nullptr; 1871 } 1872 } 1873 1874 void InlinedObjects::inlined_field_names_callback(InlinedObjects* _this, const Klass* klass, uintx base_index, int count) { 1875 if (_this->_inlined_field_map == nullptr) { 1876 _this->_inlined_field_map = new (mtServiceability) GrowableArray<ClassInlinedFields>(100, mtServiceability); 1877 } 1878 _this->_inlined_field_map->append(ClassInlinedFields(klass, base_index)); 1879 1880 // counters for dumping classes with inlined fields 1881 _this->_classes_count++; 1882 _this->_inlined_fields_count += count; 1883 } 1884 1885 void InlinedObjects::dump_inlined_field_names(AbstractDumpWriter* writer) { 1886 InlinedFieldNameDumper nameDumper(writer, this, inlined_field_names_callback); 1887 ClassLoaderDataGraph::classes_do(&nameDumper); 1888 1889 if (_inlined_field_map != nullptr) { 1890 // prepare the map for get_base_index_for(). 1891 _inlined_field_map->sort(ClassInlinedFields::compare); 1892 } 1893 } 1894 1895 uintx InlinedObjects::get_base_index_for(Klass* k) { 1896 if (_inlined_field_map != nullptr) { 1897 bool found = false; 1898 int idx = _inlined_field_map->find_sorted<ClassInlinedFields, ClassInlinedFields::compare>(ClassInlinedFields(k, 0), found); 1899 if (found) { 1900 return _inlined_field_map->at(idx).base_index; 1901 } 1902 } 1903 1904 // return max_uintx, so get_next_string_id returns 0. 1905 return max_uintx; 1906 } 1907 1908 uintx InlinedObjects::get_next_string_id(uintx id) { 1909 if (++id == _min_string_id) { 1910 return _max_string_id + 1; 1911 } 1912 return id; 1913 } 1914 1915 void InlinedObjects::dump_classed_with_inlined_fields(AbstractDumpWriter* writer) { 1916 if (_classes_count != 0) { 1917 // Record for each class contains tag(u1), class ID and count(u2) 1918 // for each inlined field index(u2), synthetic fields count(u2), original field name and class ID 1919 int size = _classes_count * (1 + sizeof(address) + 2) 1920 + _inlined_fields_count * (2 + 2 + sizeof(address) + sizeof(address)); 1921 DumperSupport::write_header(writer, HPROF_INLINED_FIELDS, (u4)size); 1922 1923 InlinedFieldsDumper dumper(writer); 1924 ClassLoaderDataGraph::classes_do(&dumper); 1925 } 1926 } 1927 1928 void InlinedObjects::add_flat_array(oop array) { 1929 if (_flat_arrays == nullptr) { 1930 _flat_arrays = new (mtServiceability) GrowableArray<oop>(100, mtServiceability); 1931 } 1932 _flat_arrays->append(array); 1933 } 1934 1935 void InlinedObjects::dump_flat_arrays(AbstractDumpWriter* writer) { 1936 if (_flat_arrays != nullptr) { 1937 // For each flat array the record contains tag (u1), object ID and class ID. 1938 int size = _flat_arrays->length() * (1 + sizeof(address) + sizeof(address)); 1939 1940 DumperSupport::write_header(writer, HPROF_FLAT_ARRAYS, (u4)size); 1941 for (GrowableArrayIterator<oop> it = _flat_arrays->begin(); it != _flat_arrays->end(); ++it) { 1942 flatArrayOop array = flatArrayOop(*it); 1943 FlatArrayKlass* array_klass = FlatArrayKlass::cast(array->klass()); 1944 InlineKlass* element_klass = array_klass->element_klass(); 1945 writer->write_u1(HPROF_FLAT_ARRAY); 1946 writer->write_objectID(array); 1947 writer->write_classID(element_klass); 1948 } 1949 } 1950 } 1951 1952 1953 // Support class used to generate HPROF_UTF8 records from the entries in the 1954 // SymbolTable. 1955 1956 class SymbolTableDumper : public SymbolClosure { 1957 private: 1958 AbstractDumpWriter* _writer; 1959 AbstractDumpWriter* writer() const { return _writer; } 1960 public: 1961 SymbolTableDumper(AbstractDumpWriter* writer) { _writer = writer; } 1962 void do_symbol(Symbol** p); 1963 }; 1964 1965 void SymbolTableDumper::do_symbol(Symbol** p) { 1966 ResourceMark rm; 1967 Symbol* sym = *p; 1968 int len = sym->utf8_length(); 1969 if (len > 0) { 1970 char* s = sym->as_utf8(); 1971 DumperSupport::write_header(writer(), HPROF_UTF8, oopSize + len); 1972 writer()->write_symbolID(sym); 1973 writer()->write_raw(s, len); 1974 } 1975 } 1976 1977 // Support class used to generate HPROF_GC_CLASS_DUMP records 1978 1979 class ClassDumper : public KlassClosure { 1980 private: 1981 AbstractDumpWriter* _writer; 1982 AbstractDumpWriter* writer() const { return _writer; } 1983 1984 public: 1985 ClassDumper(AbstractDumpWriter* writer) : _writer(writer) {} 1986 1987 void do_klass(Klass* k) { 1988 if (k->is_instance_klass()) { 1989 DumperSupport::dump_instance_class(writer(), k); 1990 } else { 1991 DumperSupport::dump_array_class(writer(), k); 1992 } 1993 } 1994 }; 1995 1996 // Support class used to generate HPROF_GC_ROOT_JNI_LOCAL records 1997 1998 class JNILocalsDumper : public OopClosure { 1999 private: 2000 AbstractDumpWriter* _writer; 2001 u4 _thread_serial_num; 2002 int _frame_num; 2003 AbstractDumpWriter* writer() const { return _writer; } 2004 public: 2005 JNILocalsDumper(AbstractDumpWriter* writer, u4 thread_serial_num) { 2006 _writer = writer; 2007 _thread_serial_num = thread_serial_num; 2008 _frame_num = -1; // default - empty stack 2009 } 2010 void set_frame_number(int n) { _frame_num = n; } 2011 void do_oop(oop* obj_p); 2012 void do_oop(narrowOop* obj_p) { ShouldNotReachHere(); } 2013 }; 2014 2015 void JNILocalsDumper::do_oop(oop* obj_p) { 2016 // ignore null handles 2017 oop o = *obj_p; 2018 if (o != nullptr) { 2019 u4 size = 1 + sizeof(address) + 4 + 4; 2020 writer()->start_sub_record(HPROF_GC_ROOT_JNI_LOCAL, size); 2021 writer()->write_objectID(o); 2022 writer()->write_u4(_thread_serial_num); 2023 writer()->write_u4((u4)_frame_num); 2024 writer()->end_sub_record(); 2025 } 2026 } 2027 2028 2029 // Support class used to generate HPROF_GC_ROOT_JNI_GLOBAL records 2030 2031 class JNIGlobalsDumper : public OopClosure { 2032 private: 2033 AbstractDumpWriter* _writer; 2034 AbstractDumpWriter* writer() const { return _writer; } 2035 2036 public: 2037 JNIGlobalsDumper(AbstractDumpWriter* writer) { 2038 _writer = writer; 2039 } 2040 void do_oop(oop* obj_p); 2041 void do_oop(narrowOop* obj_p) { ShouldNotReachHere(); } 2042 }; 2043 2044 void JNIGlobalsDumper::do_oop(oop* obj_p) { 2045 oop o = NativeAccess<AS_NO_KEEPALIVE>::oop_load(obj_p); 2046 2047 // ignore these 2048 if (o == nullptr) return; 2049 // we ignore global ref to symbols and other internal objects 2050 if (o->is_instance() || o->is_objArray() || o->is_typeArray()) { 2051 u4 size = 1 + 2 * sizeof(address); 2052 writer()->start_sub_record(HPROF_GC_ROOT_JNI_GLOBAL, size); 2053 writer()->write_objectID(o); 2054 writer()->write_rootID(obj_p); // global ref ID 2055 writer()->end_sub_record(); 2056 } 2057 }; 2058 2059 // Support class used to generate HPROF_GC_ROOT_STICKY_CLASS records 2060 2061 class StickyClassDumper : public KlassClosure { 2062 private: 2063 AbstractDumpWriter* _writer; 2064 AbstractDumpWriter* writer() const { return _writer; } 2065 public: 2066 StickyClassDumper(AbstractDumpWriter* writer) { 2067 _writer = writer; 2068 } 2069 void do_klass(Klass* k) { 2070 if (k->is_instance_klass()) { 2071 InstanceKlass* ik = InstanceKlass::cast(k); 2072 u4 size = 1 + sizeof(address); 2073 writer()->start_sub_record(HPROF_GC_ROOT_STICKY_CLASS, size); 2074 writer()->write_classID(ik); 2075 writer()->end_sub_record(); 2076 } 2077 } 2078 }; 2079 2080 // Support class used to generate HPROF_GC_ROOT_JAVA_FRAME records. 2081 2082 class JavaStackRefDumper : public StackObj { 2083 private: 2084 AbstractDumpWriter* _writer; 2085 u4 _thread_serial_num; 2086 int _frame_num; 2087 AbstractDumpWriter* writer() const { return _writer; } 2088 public: 2089 JavaStackRefDumper(AbstractDumpWriter* writer, u4 thread_serial_num) 2090 : _writer(writer), _thread_serial_num(thread_serial_num), _frame_num(-1) // default - empty stack 2091 { 2092 } 2093 2094 void set_frame_number(int n) { _frame_num = n; } 2095 2096 void dump_java_stack_refs(StackValueCollection* values); 2097 }; 2098 2099 void JavaStackRefDumper::dump_java_stack_refs(StackValueCollection* values) { 2100 for (int index = 0; index < values->size(); index++) { 2101 if (values->at(index)->type() == T_OBJECT) { 2102 oop o = values->obj_at(index)(); 2103 if (o != nullptr) { 2104 u4 size = 1 + sizeof(address) + 4 + 4; 2105 writer()->start_sub_record(HPROF_GC_ROOT_JAVA_FRAME, size); 2106 writer()->write_objectID(o); 2107 writer()->write_u4(_thread_serial_num); 2108 writer()->write_u4((u4)_frame_num); 2109 writer()->end_sub_record(); 2110 } 2111 } 2112 } 2113 } 2114 2115 // Class to collect, store and dump thread-related data: 2116 // - HPROF_TRACE and HPROF_FRAME records; 2117 // - HPROF_GC_ROOT_THREAD_OBJ/HPROF_GC_ROOT_JAVA_FRAME/HPROF_GC_ROOT_JNI_LOCAL subrecords. 2118 class ThreadDumper : public CHeapObj<mtInternal> { 2119 public: 2120 enum class ThreadType { Platform, MountedVirtual, UnmountedVirtual }; 2121 2122 private: 2123 ThreadType _thread_type; 2124 JavaThread* _java_thread; 2125 oop _thread_oop; 2126 2127 GrowableArray<StackFrameInfo*>* _frames; 2128 // non-null if the thread is OOM thread 2129 Method* _oome_constructor; 2130 int _thread_serial_num; 2131 int _start_frame_serial_num; 2132 2133 vframe* get_top_frame() const; 2134 2135 public: 2136 static bool should_dump_pthread(JavaThread* thread) { 2137 return thread->threadObj() != nullptr && !thread->is_exiting() && !thread->is_hidden_from_external_view(); 2138 } 2139 2140 static bool should_dump_vthread(oop vt) { 2141 return java_lang_VirtualThread::state(vt) != java_lang_VirtualThread::NEW 2142 && java_lang_VirtualThread::state(vt) != java_lang_VirtualThread::TERMINATED; 2143 } 2144 2145 static bool is_vthread_mounted(oop vt) { 2146 // The code should be consistent with the "mounted virtual thread" case 2147 // (VM_HeapDumper::dump_stack_traces(), ThreadDumper::get_top_frame()). 2148 // I.e. virtual thread is mounted if its carrierThread is not null 2149 // and is_vthread_mounted() for the carrier thread returns true. 2150 oop carrier_thread = java_lang_VirtualThread::carrier_thread(vt); 2151 if (carrier_thread == nullptr) { 2152 return false; 2153 } 2154 JavaThread* java_thread = java_lang_Thread::thread(carrier_thread); 2155 return java_thread->is_vthread_mounted(); 2156 } 2157 2158 ThreadDumper(ThreadType thread_type, JavaThread* java_thread, oop thread_oop); 2159 2160 // affects frame_count 2161 void add_oom_frame(Method* oome_constructor) { 2162 assert(_start_frame_serial_num == 0, "add_oom_frame cannot be called after init_serial_nums"); 2163 _oome_constructor = oome_constructor; 2164 } 2165 2166 void init_serial_nums(volatile int* thread_counter, volatile int* frame_counter) { 2167 assert(_start_frame_serial_num == 0, "already initialized"); 2168 _thread_serial_num = Atomic::fetch_then_add(thread_counter, 1); 2169 _start_frame_serial_num = Atomic::fetch_then_add(frame_counter, frame_count()); 2170 } 2171 2172 bool oom_thread() const { 2173 return _oome_constructor != nullptr; 2174 } 2175 2176 int frame_count() const { 2177 return _frames->length() + (oom_thread() ? 1 : 0); 2178 } 2179 2180 u4 thread_serial_num() const { 2181 return (u4)_thread_serial_num; 2182 } 2183 2184 u4 stack_trace_serial_num() const { 2185 return (u4)(_thread_serial_num + STACK_TRACE_ID); 2186 } 2187 2188 // writes HPROF_TRACE and HPROF_FRAME records 2189 // returns number of dumped frames 2190 void dump_stack_traces(AbstractDumpWriter* writer, GrowableArray<Klass*>* klass_map); 2191 2192 // writes HPROF_GC_ROOT_THREAD_OBJ subrecord 2193 void dump_thread_obj(AbstractDumpWriter* writer); 2194 2195 // Walk the stack of the thread. 2196 // Dumps a HPROF_GC_ROOT_JAVA_FRAME subrecord for each local 2197 // Dumps a HPROF_GC_ROOT_JNI_LOCAL subrecord for each JNI local 2198 void dump_stack_refs(AbstractDumpWriter* writer); 2199 2200 }; 2201 2202 ThreadDumper::ThreadDumper(ThreadType thread_type, JavaThread* java_thread, oop thread_oop) 2203 : _thread_type(thread_type), _java_thread(java_thread), _thread_oop(thread_oop), 2204 _oome_constructor(nullptr), 2205 _thread_serial_num(0), _start_frame_serial_num(0) 2206 { 2207 // sanity checks 2208 if (_thread_type == ThreadType::UnmountedVirtual) { 2209 assert(_java_thread == nullptr, "sanity"); 2210 assert(_thread_oop != nullptr, "sanity"); 2211 } else { 2212 assert(_java_thread != nullptr, "sanity"); 2213 assert(_thread_oop != nullptr, "sanity"); 2214 } 2215 2216 _frames = new (mtServiceability) GrowableArray<StackFrameInfo*>(10, mtServiceability); 2217 bool stop_at_vthread_entry = _thread_type == ThreadType::MountedVirtual; 2218 2219 // vframes are resource allocated 2220 Thread* current_thread = Thread::current(); 2221 ResourceMark rm(current_thread); 2222 HandleMark hm(current_thread); 2223 2224 for (vframe* vf = get_top_frame(); vf != nullptr; vf = vf->sender()) { 2225 if (stop_at_vthread_entry && vf->is_vthread_entry()) { 2226 break; 2227 } 2228 if (vf->is_java_frame()) { 2229 javaVFrame* jvf = javaVFrame::cast(vf); 2230 _frames->append(new StackFrameInfo(jvf, false)); 2231 } else { 2232 // ignore non-Java frames 2233 } 2234 } 2235 } 2236 2237 void ThreadDumper::dump_stack_traces(AbstractDumpWriter* writer, GrowableArray<Klass*>* klass_map) { 2238 assert(_thread_serial_num != 0 && _start_frame_serial_num != 0, "serial_nums are not initialized"); 2239 2240 // write HPROF_FRAME records for this thread's stack trace 2241 int depth = _frames->length(); 2242 int frame_serial_num = _start_frame_serial_num; 2243 2244 if (oom_thread()) { 2245 // OOM thread 2246 // write fake frame that makes it look like the thread, which caused OOME, 2247 // is in the OutOfMemoryError zero-parameter constructor 2248 int oome_serial_num = klass_map->find(_oome_constructor->method_holder()); 2249 // the class serial number starts from 1 2250 assert(oome_serial_num > 0, "OutOfMemoryError class not found"); 2251 DumperSupport::dump_stack_frame(writer, ++frame_serial_num, oome_serial_num, _oome_constructor, 0); 2252 depth++; 2253 } 2254 2255 for (int j = 0; j < _frames->length(); j++) { 2256 StackFrameInfo* frame = _frames->at(j); 2257 Method* m = frame->method(); 2258 int class_serial_num = klass_map->find(m->method_holder()); 2259 // the class serial number starts from 1 2260 assert(class_serial_num > 0, "class not found"); 2261 DumperSupport::dump_stack_frame(writer, ++frame_serial_num, class_serial_num, m, frame->bci()); 2262 } 2263 2264 // write HPROF_TRACE record for the thread 2265 DumperSupport::write_header(writer, HPROF_TRACE, checked_cast<u4>(3 * sizeof(u4) + depth * oopSize)); 2266 writer->write_u4(stack_trace_serial_num()); // stack trace serial number 2267 writer->write_u4(thread_serial_num()); // thread serial number 2268 writer->write_u4((u4)depth); // frame count (including oom frame) 2269 for (int j = 1; j <= depth; j++) { 2270 writer->write_id(_start_frame_serial_num + j); 2271 } 2272 } 2273 2274 void ThreadDumper::dump_thread_obj(AbstractDumpWriter * writer) { 2275 assert(_thread_serial_num != 0 && _start_frame_serial_num != 0, "serial_num is not initialized"); 2276 2277 u4 size = 1 + sizeof(address) + 4 + 4; 2278 writer->start_sub_record(HPROF_GC_ROOT_THREAD_OBJ, size); 2279 writer->write_objectID(_thread_oop); 2280 writer->write_u4(thread_serial_num()); // thread serial number 2281 writer->write_u4(stack_trace_serial_num()); // stack trace serial number 2282 writer->end_sub_record(); 2283 } 2284 2285 void ThreadDumper::dump_stack_refs(AbstractDumpWriter * writer) { 2286 assert(_thread_serial_num != 0 && _start_frame_serial_num != 0, "serial_num is not initialized"); 2287 2288 JNILocalsDumper blk(writer, thread_serial_num()); 2289 if (_thread_type == ThreadType::Platform) { 2290 if (!_java_thread->has_last_Java_frame()) { 2291 // no last java frame but there may be JNI locals 2292 _java_thread->active_handles()->oops_do(&blk); 2293 return; 2294 } 2295 } 2296 2297 JavaStackRefDumper java_ref_dumper(writer, thread_serial_num()); 2298 2299 // vframes are resource allocated 2300 Thread* current_thread = Thread::current(); 2301 ResourceMark rm(current_thread); 2302 HandleMark hm(current_thread); 2303 2304 bool stopAtVthreadEntry = _thread_type == ThreadType::MountedVirtual; 2305 frame* last_entry_frame = nullptr; 2306 bool is_top_frame = true; 2307 int depth = 0; 2308 if (oom_thread()) { 2309 depth++; 2310 } 2311 2312 for (vframe* vf = get_top_frame(); vf != nullptr; vf = vf->sender()) { 2313 if (stopAtVthreadEntry && vf->is_vthread_entry()) { 2314 break; 2315 } 2316 2317 if (vf->is_java_frame()) { 2318 javaVFrame* jvf = javaVFrame::cast(vf); 2319 if (!(jvf->method()->is_native())) { 2320 java_ref_dumper.set_frame_number(depth); 2321 java_ref_dumper.dump_java_stack_refs(jvf->locals()); 2322 java_ref_dumper.dump_java_stack_refs(jvf->expressions()); 2323 } else { 2324 // native frame 2325 blk.set_frame_number(depth); 2326 if (is_top_frame) { 2327 // JNI locals for the top frame. 2328 assert(_java_thread != nullptr, "impossible for unmounted vthread"); 2329 _java_thread->active_handles()->oops_do(&blk); 2330 } else { 2331 if (last_entry_frame != nullptr) { 2332 // JNI locals for the entry frame 2333 assert(last_entry_frame->is_entry_frame(), "checking"); 2334 last_entry_frame->entry_frame_call_wrapper()->handles()->oops_do(&blk); 2335 } 2336 } 2337 } 2338 last_entry_frame = nullptr; 2339 // increment only for Java frames 2340 depth++; 2341 } else { 2342 // externalVFrame - for an entry frame then we report the JNI locals 2343 // when we find the corresponding javaVFrame 2344 frame* fr = vf->frame_pointer(); 2345 assert(fr != nullptr, "sanity check"); 2346 if (fr->is_entry_frame()) { 2347 last_entry_frame = fr; 2348 } 2349 } 2350 is_top_frame = false; 2351 } 2352 assert(depth == frame_count(), "total number of Java frames not matched"); 2353 } 2354 2355 vframe* ThreadDumper::get_top_frame() const { 2356 if (_thread_type == ThreadType::UnmountedVirtual) { 2357 ContinuationWrapper cont(java_lang_VirtualThread::continuation(_thread_oop)); 2358 if (cont.is_empty()) { 2359 return nullptr; 2360 } 2361 assert(!cont.is_mounted(), "sanity check"); 2362 stackChunkOop chunk = cont.last_nonempty_chunk(); 2363 if (chunk == nullptr || chunk->is_empty()) { 2364 return nullptr; 2365 } 2366 2367 RegisterMap reg_map(cont.continuation(), RegisterMap::UpdateMap::include); 2368 frame fr = chunk->top_frame(®_map); 2369 vframe* vf = vframe::new_vframe(&fr, ®_map, nullptr); // don't need JavaThread 2370 return vf; 2371 } 2372 2373 RegisterMap reg_map(_java_thread, 2374 RegisterMap::UpdateMap::include, 2375 RegisterMap::ProcessFrames::include, 2376 RegisterMap::WalkContinuation::skip); 2377 switch (_thread_type) { 2378 case ThreadType::Platform: 2379 if (!_java_thread->has_last_Java_frame()) { 2380 return nullptr; 2381 } 2382 return _java_thread->is_vthread_mounted() 2383 ? _java_thread->carrier_last_java_vframe(®_map) 2384 : _java_thread->platform_thread_last_java_vframe(®_map); 2385 2386 case ThreadType::MountedVirtual: 2387 return _java_thread->last_java_vframe(®_map); 2388 2389 default: // make compilers happy 2390 break; 2391 } 2392 ShouldNotReachHere(); 2393 return nullptr; 2394 } 2395 2396 // Callback to dump thread-related data for unmounted virtual threads; 2397 // implemented by VM_HeapDumper. 2398 class UnmountedVThreadDumper { 2399 public: 2400 virtual void dump_vthread(oop vt, AbstractDumpWriter* segment_writer) = 0; 2401 }; 2402 2403 // Support class used when iterating over the heap. 2404 class HeapObjectDumper : public ObjectClosure { 2405 private: 2406 AbstractDumpWriter* _writer; 2407 AbstractDumpWriter* writer() { return _writer; } 2408 UnmountedVThreadDumper* _vthread_dumper; 2409 2410 DumperClassCacheTable _class_cache; 2411 2412 public: 2413 HeapObjectDumper(AbstractDumpWriter* writer, UnmountedVThreadDumper* vthread_dumper) 2414 : _writer(writer), _vthread_dumper(vthread_dumper) {} 2415 2416 // called for each object in the heap 2417 void do_object(oop o); 2418 }; 2419 2420 void HeapObjectDumper::do_object(oop o) { 2421 // skip classes as these emitted as HPROF_GC_CLASS_DUMP records 2422 if (o->klass() == vmClasses::Class_klass()) { 2423 if (!java_lang_Class::is_primitive(o)) { 2424 return; 2425 } 2426 } 2427 2428 if (DumperSupport::mask_dormant_archived_object(o, nullptr) == nullptr) { 2429 return; 2430 } 2431 2432 if (o->is_instance()) { 2433 // create a HPROF_GC_INSTANCE record for each object 2434 DumperSupport::dump_instance(writer(), o, &_class_cache); 2435 // If we encounter an unmounted virtual thread it needs to be dumped explicitly 2436 // (mounted virtual threads are dumped with their carriers). 2437 if (java_lang_VirtualThread::is_instance(o) 2438 && ThreadDumper::should_dump_vthread(o) && !ThreadDumper::is_vthread_mounted(o)) { 2439 _vthread_dumper->dump_vthread(o, writer()); 2440 } 2441 } else if (o->is_objArray()) { 2442 // create a HPROF_GC_OBJ_ARRAY_DUMP record for each object array 2443 DumperSupport::dump_object_array(writer(), objArrayOop(o)); 2444 } else if (o->is_flatArray()) { 2445 DumperSupport::dump_flat_array(writer(), flatArrayOop(o), &_class_cache); 2446 } else if (o->is_typeArray()) { 2447 // create a HPROF_GC_PRIM_ARRAY_DUMP record for each type array 2448 DumperSupport::dump_prim_array(writer(), typeArrayOop(o)); 2449 } 2450 } 2451 2452 // The dumper controller for parallel heap dump 2453 class DumperController : public CHeapObj<mtInternal> { 2454 private: 2455 Monitor* _lock; 2456 Mutex* _global_writer_lock; 2457 2458 const uint _dumper_number; 2459 uint _complete_number; 2460 2461 bool _started; // VM dumper started and acquired global writer lock 2462 2463 public: 2464 DumperController(uint number) : 2465 // _lock and _global_writer_lock are used for synchronization between GC worker threads inside safepoint, 2466 // so we lock with _no_safepoint_check_flag. 2467 // signal_start() acquires _lock when global writer is locked, 2468 // its rank must be less than _global_writer_lock rank. 2469 _lock(new (std::nothrow) PaddedMonitor(Mutex::nosafepoint - 1, "DumperController_lock")), 2470 _global_writer_lock(new (std::nothrow) Mutex(Mutex::nosafepoint, "DumpWriter_lock")), 2471 _dumper_number(number), 2472 _complete_number(0), 2473 _started(false) 2474 {} 2475 2476 ~DumperController() { 2477 delete _lock; 2478 delete _global_writer_lock; 2479 } 2480 2481 // parallel (non VM) dumpers must wait until VM dumper acquires global writer lock 2482 void wait_for_start_signal() { 2483 MonitorLocker ml(_lock, Mutex::_no_safepoint_check_flag); 2484 while (_started == false) { 2485 ml.wait(); 2486 } 2487 } 2488 2489 void signal_start() { 2490 MonitorLocker ml(_lock, Mutex::_no_safepoint_check_flag); 2491 _started = true; 2492 ml.notify_all(); 2493 } 2494 2495 void lock_global_writer() { 2496 _global_writer_lock->lock_without_safepoint_check(); 2497 } 2498 2499 void unlock_global_writer() { 2500 _global_writer_lock->unlock(); 2501 } 2502 2503 void dumper_complete(DumpWriter* local_writer, DumpWriter* global_writer) { 2504 MonitorLocker ml(_lock, Mutex::_no_safepoint_check_flag); 2505 _complete_number++; 2506 // propagate local error to global if any 2507 if (local_writer->has_error()) { 2508 global_writer->set_error(local_writer->error()); 2509 } 2510 ml.notify(); 2511 } 2512 2513 void wait_all_dumpers_complete() { 2514 MonitorLocker ml(_lock, Mutex::_no_safepoint_check_flag); 2515 while (_complete_number != _dumper_number) { 2516 ml.wait(); 2517 } 2518 } 2519 }; 2520 2521 // DumpMerger merges separate dump files into a complete one 2522 class DumpMerger : public StackObj { 2523 private: 2524 DumpWriter* _writer; 2525 InlinedObjects* _inlined_objects; 2526 const char* _path; 2527 bool _has_error; 2528 int _dump_seq; 2529 2530 private: 2531 void merge_file(const char* path); 2532 void merge_done(); 2533 void set_error(const char* msg); 2534 2535 public: 2536 DumpMerger(const char* path, DumpWriter* writer, InlinedObjects* inlined_objects, int dump_seq) : 2537 _writer(writer), 2538 _inlined_objects(inlined_objects), 2539 _path(path), 2540 _has_error(_writer->has_error()), 2541 _dump_seq(dump_seq) {} 2542 2543 void do_merge(); 2544 2545 // returns path for the parallel DumpWriter (resource allocated) 2546 static char* get_writer_path(const char* base_path, int seq); 2547 2548 }; 2549 2550 char* DumpMerger::get_writer_path(const char* base_path, int seq) { 2551 // approximate required buffer size 2552 size_t buf_size = strlen(base_path) 2553 + 2 // ".p" 2554 + 10 // number (that's enough for 2^32 parallel dumpers) 2555 + 1; // '\0' 2556 2557 char* path = NEW_RESOURCE_ARRAY(char, buf_size); 2558 memset(path, 0, buf_size); 2559 2560 os::snprintf(path, buf_size, "%s.p%d", base_path, seq); 2561 2562 return path; 2563 } 2564 2565 2566 void DumpMerger::merge_done() { 2567 // Writes the HPROF_HEAP_DUMP_END record. 2568 if (!_has_error) { 2569 DumperSupport::end_of_dump(_writer); 2570 _inlined_objects->dump_flat_arrays(_writer); 2571 _writer->flush(); 2572 _inlined_objects->release(); 2573 } 2574 _dump_seq = 0; //reset 2575 } 2576 2577 void DumpMerger::set_error(const char* msg) { 2578 assert(msg != nullptr, "sanity check"); 2579 log_error(heapdump)("%s (file: %s)", msg, _path); 2580 _writer->set_error(msg); 2581 _has_error = true; 2582 } 2583 2584 #ifdef LINUX 2585 // Merge segmented heap files via sendfile, it's more efficient than the 2586 // read+write combination, which would require transferring data to and from 2587 // user space. 2588 void DumpMerger::merge_file(const char* path) { 2589 TraceTime timer("Merge segmented heap file directly", TRACETIME_LOG(Info, heapdump)); 2590 2591 int segment_fd = os::open(path, O_RDONLY, 0); 2592 if (segment_fd == -1) { 2593 set_error("Can not open segmented heap file during merging"); 2594 return; 2595 } 2596 2597 struct stat st; 2598 if (os::stat(path, &st) != 0) { 2599 ::close(segment_fd); 2600 set_error("Can not get segmented heap file size during merging"); 2601 return; 2602 } 2603 2604 // A successful call to sendfile may write fewer bytes than requested; the 2605 // caller should be prepared to retry the call if there were unsent bytes. 2606 jlong offset = 0; 2607 while (offset < st.st_size) { 2608 int ret = os::Linux::sendfile(_writer->get_fd(), segment_fd, &offset, st.st_size); 2609 if (ret == -1) { 2610 ::close(segment_fd); 2611 set_error("Failed to merge segmented heap file"); 2612 return; 2613 } 2614 } 2615 2616 // As sendfile variant does not call the write method of the global writer, 2617 // bytes_written is also incorrect for this variant, we need to explicitly 2618 // accumulate bytes_written for the global writer in this case 2619 julong accum = _writer->bytes_written() + st.st_size; 2620 _writer->set_bytes_written(accum); 2621 ::close(segment_fd); 2622 } 2623 #else 2624 // Generic implementation using read+write 2625 void DumpMerger::merge_file(const char* path) { 2626 TraceTime timer("Merge segmented heap file", TRACETIME_LOG(Info, heapdump)); 2627 2628 fileStream segment_fs(path, "rb"); 2629 if (!segment_fs.is_open()) { 2630 set_error("Can not open segmented heap file during merging"); 2631 return; 2632 } 2633 2634 jlong total = 0; 2635 size_t cnt = 0; 2636 char read_buf[4096]; 2637 while ((cnt = segment_fs.read(read_buf, 1, 4096)) != 0) { 2638 _writer->write_raw(read_buf, cnt); 2639 total += cnt; 2640 } 2641 2642 _writer->flush(); 2643 if (segment_fs.fileSize() != total) { 2644 set_error("Merged heap dump is incomplete"); 2645 } 2646 } 2647 #endif 2648 2649 void DumpMerger::do_merge() { 2650 TraceTime timer("Merge heap files complete", TRACETIME_LOG(Info, heapdump)); 2651 2652 // Since contents in segmented heap file were already zipped, we don't need to zip 2653 // them again during merging. 2654 AbstractCompressor* saved_compressor = _writer->compressor(); 2655 _writer->set_compressor(nullptr); 2656 2657 // Merge the content of the remaining files into base file. Regardless of whether 2658 // the merge process is successful or not, these segmented files will be deleted. 2659 for (int i = 0; i < _dump_seq; i++) { 2660 ResourceMark rm; 2661 const char* path = get_writer_path(_path, i); 2662 if (!_has_error) { 2663 merge_file(path); 2664 } 2665 // Delete selected segmented heap file nevertheless 2666 if (remove(path) != 0) { 2667 log_info(heapdump)("Removal of segment file (%d) failed (%d)", i, errno); 2668 } 2669 } 2670 2671 // restore compressor for further use 2672 _writer->set_compressor(saved_compressor); 2673 merge_done(); 2674 } 2675 2676 // The VM operation wraps DumpMerger so that it could be performed by VM thread 2677 class VM_HeapDumpMerge : public VM_Operation { 2678 private: 2679 DumpMerger* _merger; 2680 public: 2681 VM_HeapDumpMerge(DumpMerger* merger) : _merger(merger) {} 2682 VMOp_Type type() const { return VMOp_HeapDumpMerge; } 2683 // heap dump merge could happen outside safepoint 2684 virtual bool evaluate_at_safepoint() const { return false; } 2685 void doit() { 2686 _merger->do_merge(); 2687 } 2688 }; 2689 2690 // The VM operation that performs the heap dump 2691 class VM_HeapDumper : public VM_GC_Operation, public WorkerTask, public UnmountedVThreadDumper { 2692 private: 2693 static VM_HeapDumper* _global_dumper; 2694 static DumpWriter* _global_writer; 2695 DumpWriter* _local_writer; 2696 JavaThread* _oome_thread; 2697 Method* _oome_constructor; 2698 bool _gc_before_heap_dump; 2699 GrowableArray<Klass*>* _klass_map; 2700 2701 ThreadDumper** _thread_dumpers; // platform, carrier and mounted virtual threads 2702 int _thread_dumpers_count; 2703 volatile int _thread_serial_num; 2704 volatile int _frame_serial_num; 2705 2706 volatile int _dump_seq; 2707 2708 // Inlined object support. 2709 InlinedObjects _inlined_objects; 2710 2711 // parallel heap dump support 2712 uint _num_dumper_threads; 2713 DumperController* _dumper_controller; 2714 ParallelObjectIterator* _poi; 2715 2716 // Dumper id of VMDumper thread. 2717 static const int VMDumperId = 0; 2718 // VM dumper dumps both heap and non-heap data, other dumpers dump heap-only data. 2719 static bool is_vm_dumper(int dumper_id) { return dumper_id == VMDumperId; } 2720 // the 1st dumper calling get_next_dumper_id becomes VM dumper 2721 int get_next_dumper_id() { 2722 return Atomic::fetch_then_add(&_dump_seq, 1); 2723 } 2724 2725 // accessors and setters 2726 static VM_HeapDumper* dumper() { assert(_global_dumper != nullptr, "Error"); return _global_dumper; } 2727 static DumpWriter* writer() { assert(_global_writer != nullptr, "Error"); return _global_writer; } 2728 2729 void set_global_dumper() { 2730 assert(_global_dumper == nullptr, "Error"); 2731 _global_dumper = this; 2732 } 2733 void set_global_writer() { 2734 assert(_global_writer == nullptr, "Error"); 2735 _global_writer = _local_writer; 2736 } 2737 void clear_global_dumper() { _global_dumper = nullptr; } 2738 void clear_global_writer() { _global_writer = nullptr; } 2739 2740 bool skip_operation() const; 2741 2742 // writes a HPROF_LOAD_CLASS record to global writer 2743 static void do_load_class(Klass* k); 2744 2745 // HPROF_GC_ROOT_THREAD_OBJ records for platform and mounted virtual threads 2746 void dump_threads(AbstractDumpWriter* writer); 2747 2748 void add_class_serial_number(Klass* k, int serial_num) { 2749 _klass_map->at_put_grow(serial_num, k); 2750 } 2751 2752 bool is_oom_thread(JavaThread* thread) const { 2753 return thread == _oome_thread && _oome_constructor != nullptr; 2754 } 2755 2756 // HPROF_TRACE and HPROF_FRAME records for platform and mounted virtual threads 2757 void dump_stack_traces(AbstractDumpWriter* writer); 2758 2759 public: 2760 VM_HeapDumper(DumpWriter* writer, bool gc_before_heap_dump, bool oome, uint num_dump_threads) : 2761 VM_GC_Operation(0 /* total collections, dummy, ignored */, 2762 GCCause::_heap_dump /* GC Cause */, 2763 0 /* total full collections, dummy, ignored */, 2764 gc_before_heap_dump), 2765 WorkerTask("dump heap") { 2766 _local_writer = writer; 2767 _gc_before_heap_dump = gc_before_heap_dump; 2768 _klass_map = new (mtServiceability) GrowableArray<Klass*>(INITIAL_CLASS_COUNT, mtServiceability); 2769 2770 _thread_dumpers = nullptr; 2771 _thread_dumpers_count = 0; 2772 _thread_serial_num = 1; 2773 _frame_serial_num = 1; 2774 2775 _dump_seq = VMDumperId; 2776 _num_dumper_threads = num_dump_threads; 2777 _dumper_controller = nullptr; 2778 _poi = nullptr; 2779 if (oome) { 2780 assert(!Thread::current()->is_VM_thread(), "Dump from OutOfMemoryError cannot be called by the VMThread"); 2781 // get OutOfMemoryError zero-parameter constructor 2782 InstanceKlass* oome_ik = vmClasses::OutOfMemoryError_klass(); 2783 _oome_constructor = oome_ik->find_method(vmSymbols::object_initializer_name(), 2784 vmSymbols::void_method_signature()); 2785 // get thread throwing OOME when generating the heap dump at OOME 2786 _oome_thread = JavaThread::current(); 2787 } else { 2788 _oome_thread = nullptr; 2789 _oome_constructor = nullptr; 2790 } 2791 } 2792 2793 ~VM_HeapDumper() { 2794 if (_thread_dumpers != nullptr) { 2795 for (int i = 0; i < _thread_dumpers_count; i++) { 2796 delete _thread_dumpers[i]; 2797 } 2798 FREE_C_HEAP_ARRAY(ThreadDumper*, _thread_dumpers); 2799 } 2800 2801 if (_dumper_controller != nullptr) { 2802 delete _dumper_controller; 2803 _dumper_controller = nullptr; 2804 } 2805 delete _klass_map; 2806 } 2807 int dump_seq() { return _dump_seq; } 2808 bool is_parallel_dump() { return _num_dumper_threads > 1; } 2809 void prepare_parallel_dump(WorkerThreads* workers); 2810 2811 InlinedObjects* inlined_objects() { return &_inlined_objects; } 2812 2813 VMOp_Type type() const { return VMOp_HeapDumper; } 2814 virtual bool doit_prologue(); 2815 void doit(); 2816 void work(uint worker_id); 2817 2818 // UnmountedVThreadDumper implementation 2819 void dump_vthread(oop vt, AbstractDumpWriter* segment_writer); 2820 }; 2821 2822 VM_HeapDumper* VM_HeapDumper::_global_dumper = nullptr; 2823 DumpWriter* VM_HeapDumper::_global_writer = nullptr; 2824 2825 bool VM_HeapDumper::skip_operation() const { 2826 return false; 2827 } 2828 2829 // fixes up the current dump record and writes HPROF_HEAP_DUMP_END record 2830 void DumperSupport::end_of_dump(AbstractDumpWriter* writer) { 2831 writer->finish_dump_segment(); 2832 2833 writer->write_u1(HPROF_HEAP_DUMP_END); 2834 writer->write_u4(0); 2835 writer->write_u4(0); 2836 } 2837 2838 // writes a HPROF_LOAD_CLASS record for the class 2839 void VM_HeapDumper::do_load_class(Klass* k) { 2840 static u4 class_serial_num = 0; 2841 2842 // len of HPROF_LOAD_CLASS record 2843 u4 remaining = 2*oopSize + 2*sizeof(u4); 2844 2845 DumperSupport::write_header(writer(), HPROF_LOAD_CLASS, remaining); 2846 2847 // class serial number is just a number 2848 writer()->write_u4(++class_serial_num); 2849 2850 // class ID 2851 writer()->write_classID(k); 2852 2853 // add the Klass* and class serial number pair 2854 dumper()->add_class_serial_number(k, class_serial_num); 2855 2856 writer()->write_u4(STACK_TRACE_ID); 2857 2858 // class name ID 2859 Symbol* name = k->name(); 2860 writer()->write_symbolID(name); 2861 } 2862 2863 // Write a HPROF_GC_ROOT_THREAD_OBJ record for platform/carrier and mounted virtual threads. 2864 // Then walk the stack so that locals and JNI locals are dumped. 2865 void VM_HeapDumper::dump_threads(AbstractDumpWriter* writer) { 2866 for (int i = 0; i < _thread_dumpers_count; i++) { 2867 _thread_dumpers[i]->dump_thread_obj(writer); 2868 _thread_dumpers[i]->dump_stack_refs(writer); 2869 } 2870 } 2871 2872 bool VM_HeapDumper::doit_prologue() { 2873 if (_gc_before_heap_dump && UseZGC) { 2874 // ZGC cannot perform a synchronous GC cycle from within the VM thread. 2875 // So ZCollectedHeap::collect_as_vm_thread() is a noop. To respect the 2876 // _gc_before_heap_dump flag a synchronous GC cycle is performed from 2877 // the caller thread in the prologue. 2878 Universe::heap()->collect(GCCause::_heap_dump); 2879 } 2880 return VM_GC_Operation::doit_prologue(); 2881 } 2882 2883 void VM_HeapDumper::prepare_parallel_dump(WorkerThreads* workers) { 2884 uint num_active_workers = workers != nullptr ? workers->active_workers() : 0; 2885 uint num_requested_dump_threads = _num_dumper_threads; 2886 // check if we can dump in parallel based on requested and active threads 2887 if (num_active_workers <= 1 || num_requested_dump_threads <= 1) { 2888 _num_dumper_threads = 1; 2889 } else { 2890 _num_dumper_threads = clamp(num_requested_dump_threads, 2U, num_active_workers); 2891 } 2892 _dumper_controller = new (std::nothrow) DumperController(_num_dumper_threads); 2893 bool can_parallel = _num_dumper_threads > 1; 2894 log_info(heapdump)("Requested dump threads %u, active dump threads %u, " 2895 "actual dump threads %u, parallelism %s", 2896 num_requested_dump_threads, num_active_workers, 2897 _num_dumper_threads, can_parallel ? "true" : "false"); 2898 } 2899 2900 // The VM operation that dumps the heap. The dump consists of the following 2901 // records: 2902 // 2903 // HPROF_HEADER 2904 // [HPROF_UTF8]* 2905 // [HPROF_LOAD_CLASS]* 2906 // [[HPROF_FRAME]*|HPROF_TRACE]* 2907 // [HPROF_GC_CLASS_DUMP]* 2908 // [HPROF_HEAP_DUMP_SEGMENT]* 2909 // HPROF_HEAP_DUMP_END 2910 // 2911 // The HPROF_TRACE records represent the stack traces where the heap dump 2912 // is generated and a "dummy trace" record which does not include 2913 // any frames. The dummy trace record is used to be referenced as the 2914 // unknown object alloc site. 2915 // 2916 // Each HPROF_HEAP_DUMP_SEGMENT record has a length followed by sub-records. 2917 // To allow the heap dump be generated in a single pass we remember the position 2918 // of the dump length and fix it up after all sub-records have been written. 2919 // To generate the sub-records we iterate over the heap, writing 2920 // HPROF_GC_INSTANCE_DUMP, HPROF_GC_OBJ_ARRAY_DUMP, and HPROF_GC_PRIM_ARRAY_DUMP 2921 // records as we go. Once that is done we write records for some of the GC 2922 // roots. 2923 2924 void VM_HeapDumper::doit() { 2925 2926 CollectedHeap* ch = Universe::heap(); 2927 2928 ch->ensure_parsability(false); // must happen, even if collection does 2929 // not happen (e.g. due to GCLocker) 2930 2931 if (_gc_before_heap_dump) { 2932 if (GCLocker::is_active()) { 2933 warning("GC locker is held; pre-heapdump GC was skipped"); 2934 } else { 2935 ch->collect_as_vm_thread(GCCause::_heap_dump); 2936 } 2937 } 2938 2939 // At this point we should be the only dumper active, so 2940 // the following should be safe. 2941 set_global_dumper(); 2942 set_global_writer(); 2943 2944 WorkerThreads* workers = ch->safepoint_workers(); 2945 prepare_parallel_dump(workers); 2946 2947 if (!is_parallel_dump()) { 2948 work(VMDumperId); 2949 } else { 2950 ParallelObjectIterator poi(_num_dumper_threads); 2951 _poi = &poi; 2952 workers->run_task(this, _num_dumper_threads); 2953 _poi = nullptr; 2954 } 2955 2956 // Now we clear the global variables, so that a future dumper can run. 2957 clear_global_dumper(); 2958 clear_global_writer(); 2959 } 2960 2961 void VM_HeapDumper::work(uint worker_id) { 2962 // VM Dumper works on all non-heap data dumping and part of heap iteration. 2963 int dumper_id = get_next_dumper_id(); 2964 2965 if (is_vm_dumper(dumper_id)) { 2966 // lock global writer, it will be unlocked after VM Dumper finishes with non-heap data 2967 _dumper_controller->lock_global_writer(); 2968 _dumper_controller->signal_start(); 2969 } else { 2970 _dumper_controller->wait_for_start_signal(); 2971 } 2972 2973 if (is_vm_dumper(dumper_id)) { 2974 TraceTime timer("Dump non-objects", TRACETIME_LOG(Info, heapdump)); 2975 // Write the file header - we always use 1.0.2 2976 const char* header = "JAVA PROFILE 1.0.2"; 2977 2978 // header is few bytes long - no chance to overflow int 2979 writer()->write_raw(header, strlen(header) + 1); // NUL terminated 2980 writer()->write_u4(oopSize); 2981 // timestamp is current time in ms 2982 writer()->write_u8(os::javaTimeMillis()); 2983 // HPROF_UTF8 records 2984 SymbolTableDumper sym_dumper(writer()); 2985 SymbolTable::symbols_do(&sym_dumper); 2986 2987 // HPROF_UTF8 records for inlined field names. 2988 inlined_objects()->init(); 2989 inlined_objects()->dump_inlined_field_names(writer()); 2990 2991 // HPROF_INLINED_FIELDS 2992 inlined_objects()->dump_classed_with_inlined_fields(writer()); 2993 2994 // write HPROF_LOAD_CLASS records 2995 { 2996 LockedClassesDo locked_load_classes(&do_load_class); 2997 ClassLoaderDataGraph::classes_do(&locked_load_classes); 2998 } 2999 3000 // write HPROF_FRAME and HPROF_TRACE records 3001 // this must be called after _klass_map is built when iterating the classes above. 3002 dump_stack_traces(writer()); 3003 3004 // unlock global writer, so parallel dumpers can dump stack traces of unmounted virtual threads 3005 _dumper_controller->unlock_global_writer(); 3006 } 3007 3008 // HPROF_HEAP_DUMP/HPROF_HEAP_DUMP_SEGMENT starts here 3009 3010 ResourceMark rm; 3011 // share global compressor, local DumpWriter is not responsible for its life cycle 3012 DumpWriter segment_writer(DumpMerger::get_writer_path(writer()->get_file_path(), dumper_id), 3013 writer()->is_overwrite(), writer()->compressor()); 3014 if (!segment_writer.has_error()) { 3015 if (is_vm_dumper(dumper_id)) { 3016 // dump some non-heap subrecords to heap dump segment 3017 TraceTime timer("Dump non-objects (part 2)", TRACETIME_LOG(Info, heapdump)); 3018 // Writes HPROF_GC_CLASS_DUMP records 3019 ClassDumper class_dumper(&segment_writer); 3020 ClassLoaderDataGraph::classes_do(&class_dumper); 3021 3022 // HPROF_GC_ROOT_THREAD_OBJ + frames + jni locals 3023 dump_threads(&segment_writer); 3024 3025 // HPROF_GC_ROOT_JNI_GLOBAL 3026 JNIGlobalsDumper jni_dumper(&segment_writer); 3027 JNIHandles::oops_do(&jni_dumper); 3028 // technically not jni roots, but global roots 3029 // for things like preallocated throwable backtraces 3030 Universe::vm_global()->oops_do(&jni_dumper); 3031 // HPROF_GC_ROOT_STICKY_CLASS 3032 // These should be classes in the null class loader data, and not all classes 3033 // if !ClassUnloading 3034 StickyClassDumper stiky_class_dumper(&segment_writer); 3035 ClassLoaderData::the_null_class_loader_data()->classes_do(&stiky_class_dumper); 3036 } 3037 3038 // Heap iteration. 3039 // writes HPROF_GC_INSTANCE_DUMP records. 3040 // After each sub-record is written check_segment_length will be invoked 3041 // to check if the current segment exceeds a threshold. If so, a new 3042 // segment is started. 3043 // The HPROF_GC_CLASS_DUMP and HPROF_GC_INSTANCE_DUMP are the vast bulk 3044 // of the heap dump. 3045 3046 TraceTime timer(is_parallel_dump() ? "Dump heap objects in parallel" : "Dump heap objects", TRACETIME_LOG(Info, heapdump)); 3047 HeapObjectDumper obj_dumper(&segment_writer, this); 3048 if (!is_parallel_dump()) { 3049 Universe::heap()->object_iterate(&obj_dumper); 3050 } else { 3051 // == Parallel dump 3052 _poi->object_iterate(&obj_dumper, worker_id); 3053 } 3054 3055 segment_writer.finish_dump_segment(); 3056 segment_writer.flush(); 3057 } 3058 3059 _dumper_controller->dumper_complete(&segment_writer, writer()); 3060 3061 if (is_vm_dumper(dumper_id)) { 3062 _dumper_controller->wait_all_dumpers_complete(); 3063 3064 // flush global writer 3065 writer()->flush(); 3066 3067 // At this point, all fragments of the heapdump have been written to separate files. 3068 // We need to merge them into a complete heapdump and write HPROF_HEAP_DUMP_END at that time. 3069 } 3070 } 3071 3072 void VM_HeapDumper::dump_stack_traces(AbstractDumpWriter* writer) { 3073 // write a HPROF_TRACE record without any frames to be referenced as object alloc sites 3074 DumperSupport::write_header(writer, HPROF_TRACE, 3 * sizeof(u4)); 3075 writer->write_u4((u4)STACK_TRACE_ID); 3076 writer->write_u4(0); // thread number 3077 writer->write_u4(0); // frame count 3078 3079 // max number if every platform thread is carrier with mounted virtual thread 3080 _thread_dumpers = NEW_C_HEAP_ARRAY(ThreadDumper*, Threads::number_of_threads() * 2, mtInternal); 3081 3082 for (JavaThreadIteratorWithHandle jtiwh; JavaThread * thread = jtiwh.next(); ) { 3083 if (ThreadDumper::should_dump_pthread(thread)) { 3084 bool add_oom_frame = is_oom_thread(thread); 3085 3086 oop mounted_vt = thread->is_vthread_mounted() ? thread->vthread() : nullptr; 3087 if (mounted_vt != nullptr && !ThreadDumper::should_dump_vthread(mounted_vt)) { 3088 mounted_vt = nullptr; 3089 } 3090 3091 // mounted vthread (if any) 3092 if (mounted_vt != nullptr) { 3093 ThreadDumper* thread_dumper = new ThreadDumper(ThreadDumper::ThreadType::MountedVirtual, thread, mounted_vt); 3094 _thread_dumpers[_thread_dumpers_count++] = thread_dumper; 3095 if (add_oom_frame) { 3096 thread_dumper->add_oom_frame(_oome_constructor); 3097 // we add oom frame to the VT stack, don't add it to the carrier thread stack 3098 add_oom_frame = false; 3099 } 3100 thread_dumper->init_serial_nums(&_thread_serial_num, &_frame_serial_num); 3101 thread_dumper->dump_stack_traces(writer, _klass_map); 3102 } 3103 3104 // platform or carrier thread 3105 ThreadDumper* thread_dumper = new ThreadDumper(ThreadDumper::ThreadType::Platform, thread, thread->threadObj()); 3106 _thread_dumpers[_thread_dumpers_count++] = thread_dumper; 3107 if (add_oom_frame) { 3108 thread_dumper->add_oom_frame(_oome_constructor); 3109 } 3110 thread_dumper->init_serial_nums(&_thread_serial_num, &_frame_serial_num); 3111 thread_dumper->dump_stack_traces(writer, _klass_map); 3112 } 3113 } 3114 } 3115 3116 void VM_HeapDumper::dump_vthread(oop vt, AbstractDumpWriter* segment_writer) { 3117 // unmounted vthread has no JavaThread 3118 ThreadDumper thread_dumper(ThreadDumper::ThreadType::UnmountedVirtual, nullptr, vt); 3119 thread_dumper.init_serial_nums(&_thread_serial_num, &_frame_serial_num); 3120 3121 // write HPROF_TRACE/HPROF_FRAME records to global writer 3122 _dumper_controller->lock_global_writer(); 3123 thread_dumper.dump_stack_traces(writer(), _klass_map); 3124 _dumper_controller->unlock_global_writer(); 3125 3126 // write HPROF_GC_ROOT_THREAD_OBJ/HPROF_GC_ROOT_JAVA_FRAME/HPROF_GC_ROOT_JNI_LOCAL subrecord 3127 // to segment writer 3128 thread_dumper.dump_thread_obj(segment_writer); 3129 thread_dumper.dump_stack_refs(segment_writer); 3130 } 3131 3132 // dump the heap to given path. 3133 int HeapDumper::dump(const char* path, outputStream* out, int compression, bool overwrite, uint num_dump_threads) { 3134 assert(path != nullptr && strlen(path) > 0, "path missing"); 3135 3136 // print message in interactive case 3137 if (out != nullptr) { 3138 out->print_cr("Dumping heap to %s ...", path); 3139 timer()->start(); 3140 } 3141 // create JFR event 3142 EventHeapDump event; 3143 3144 AbstractCompressor* compressor = nullptr; 3145 3146 if (compression > 0) { 3147 compressor = new (std::nothrow) GZipCompressor(compression); 3148 3149 if (compressor == nullptr) { 3150 set_error("Could not allocate gzip compressor"); 3151 return -1; 3152 } 3153 } 3154 3155 DumpWriter writer(path, overwrite, compressor); 3156 3157 if (writer.error() != nullptr) { 3158 set_error(writer.error()); 3159 if (out != nullptr) { 3160 out->print_cr("Unable to create %s: %s", path, 3161 (error() != nullptr) ? error() : "reason unknown"); 3162 } 3163 return -1; 3164 } 3165 3166 // generate the segmented heap dump into separate files 3167 VM_HeapDumper dumper(&writer, _gc_before_heap_dump, _oome, num_dump_threads); 3168 VMThread::execute(&dumper); 3169 3170 // record any error that the writer may have encountered 3171 set_error(writer.error()); 3172 3173 // Heap dump process is done in two phases 3174 // 3175 // Phase 1: Concurrent threads directly write heap data to multiple heap files. 3176 // This is done by VM_HeapDumper, which is performed within safepoint. 3177 // 3178 // Phase 2: Merge multiple heap files into one complete heap dump file. 3179 // This is done by DumpMerger, which is performed outside safepoint 3180 3181 DumpMerger merger(path, &writer, dumper.inlined_objects(), dumper.dump_seq()); 3182 Thread* current_thread = Thread::current(); 3183 if (current_thread->is_AttachListener_thread()) { 3184 // perform heapdump file merge operation in the current thread prevents us 3185 // from occupying the VM Thread, which in turn affects the occurrence of 3186 // GC and other VM operations. 3187 merger.do_merge(); 3188 } else { 3189 // otherwise, performs it by VM thread 3190 VM_HeapDumpMerge op(&merger); 3191 VMThread::execute(&op); 3192 } 3193 if (writer.error() != nullptr) { 3194 set_error(writer.error()); 3195 } 3196 3197 // emit JFR event 3198 if (error() == nullptr) { 3199 event.set_destination(path); 3200 event.set_gcBeforeDump(_gc_before_heap_dump); 3201 event.set_size(writer.bytes_written()); 3202 event.set_onOutOfMemoryError(_oome); 3203 event.set_overwrite(overwrite); 3204 event.set_compression(compression); 3205 event.commit(); 3206 } else { 3207 log_debug(cds, heap)("Error %s while dumping heap", error()); 3208 } 3209 3210 // print message in interactive case 3211 if (out != nullptr) { 3212 timer()->stop(); 3213 if (error() == nullptr) { 3214 out->print_cr("Heap dump file created [" JULONG_FORMAT " bytes in %3.3f secs]", 3215 writer.bytes_written(), timer()->seconds()); 3216 } else { 3217 out->print_cr("Dump file is incomplete: %s", writer.error()); 3218 } 3219 } 3220 3221 if (compressor != nullptr) { 3222 delete compressor; 3223 } 3224 return (writer.error() == nullptr) ? 0 : -1; 3225 } 3226 3227 // stop timer (if still active), and free any error string we might be holding 3228 HeapDumper::~HeapDumper() { 3229 if (timer()->is_active()) { 3230 timer()->stop(); 3231 } 3232 set_error(nullptr); 3233 } 3234 3235 3236 // returns the error string (resource allocated), or null 3237 char* HeapDumper::error_as_C_string() const { 3238 if (error() != nullptr) { 3239 char* str = NEW_RESOURCE_ARRAY(char, strlen(error())+1); 3240 strcpy(str, error()); 3241 return str; 3242 } else { 3243 return nullptr; 3244 } 3245 } 3246 3247 // set the error string 3248 void HeapDumper::set_error(char const* error) { 3249 if (_error != nullptr) { 3250 os::free(_error); 3251 } 3252 if (error == nullptr) { 3253 _error = nullptr; 3254 } else { 3255 _error = os::strdup(error); 3256 assert(_error != nullptr, "allocation failure"); 3257 } 3258 } 3259 3260 // Called by out-of-memory error reporting by a single Java thread 3261 // outside of a JVM safepoint 3262 void HeapDumper::dump_heap_from_oome() { 3263 HeapDumper::dump_heap(true); 3264 } 3265 3266 // Called by error reporting by a single Java thread outside of a JVM safepoint, 3267 // or by heap dumping by the VM thread during a (GC) safepoint. Thus, these various 3268 // callers are strictly serialized and guaranteed not to interfere below. For more 3269 // general use, however, this method will need modification to prevent 3270 // inteference when updating the static variables base_path and dump_file_seq below. 3271 void HeapDumper::dump_heap() { 3272 HeapDumper::dump_heap(false); 3273 } 3274 3275 void HeapDumper::dump_heap(bool oome) { 3276 static char base_path[JVM_MAXPATHLEN] = {'\0'}; 3277 static uint dump_file_seq = 0; 3278 char* my_path; 3279 const int max_digit_chars = 20; 3280 3281 const char* dump_file_name = "java_pid"; 3282 const char* dump_file_ext = HeapDumpGzipLevel > 0 ? ".hprof.gz" : ".hprof"; 3283 3284 // The dump file defaults to java_pid<pid>.hprof in the current working 3285 // directory. HeapDumpPath=<file> can be used to specify an alternative 3286 // dump file name or a directory where dump file is created. 3287 if (dump_file_seq == 0) { // first time in, we initialize base_path 3288 // Calculate potentially longest base path and check if we have enough 3289 // allocated statically. 3290 const size_t total_length = 3291 (HeapDumpPath == nullptr ? 0 : strlen(HeapDumpPath)) + 3292 strlen(os::file_separator()) + max_digit_chars + 3293 strlen(dump_file_name) + strlen(dump_file_ext) + 1; 3294 if (total_length > sizeof(base_path)) { 3295 warning("Cannot create heap dump file. HeapDumpPath is too long."); 3296 return; 3297 } 3298 3299 bool use_default_filename = true; 3300 if (HeapDumpPath == nullptr || HeapDumpPath[0] == '\0') { 3301 // HeapDumpPath=<file> not specified 3302 } else { 3303 strcpy(base_path, HeapDumpPath); 3304 // check if the path is a directory (must exist) 3305 DIR* dir = os::opendir(base_path); 3306 if (dir == nullptr) { 3307 use_default_filename = false; 3308 } else { 3309 // HeapDumpPath specified a directory. We append a file separator 3310 // (if needed). 3311 os::closedir(dir); 3312 size_t fs_len = strlen(os::file_separator()); 3313 if (strlen(base_path) >= fs_len) { 3314 char* end = base_path; 3315 end += (strlen(base_path) - fs_len); 3316 if (strcmp(end, os::file_separator()) != 0) { 3317 strcat(base_path, os::file_separator()); 3318 } 3319 } 3320 } 3321 } 3322 // If HeapDumpPath wasn't a file name then we append the default name 3323 if (use_default_filename) { 3324 const size_t dlen = strlen(base_path); // if heap dump dir specified 3325 jio_snprintf(&base_path[dlen], sizeof(base_path)-dlen, "%s%d%s", 3326 dump_file_name, os::current_process_id(), dump_file_ext); 3327 } 3328 const size_t len = strlen(base_path) + 1; 3329 my_path = (char*)os::malloc(len, mtInternal); 3330 if (my_path == nullptr) { 3331 warning("Cannot create heap dump file. Out of system memory."); 3332 return; 3333 } 3334 strncpy(my_path, base_path, len); 3335 } else { 3336 // Append a sequence number id for dumps following the first 3337 const size_t len = strlen(base_path) + max_digit_chars + 2; // for '.' and \0 3338 my_path = (char*)os::malloc(len, mtInternal); 3339 if (my_path == nullptr) { 3340 warning("Cannot create heap dump file. Out of system memory."); 3341 return; 3342 } 3343 jio_snprintf(my_path, len, "%s.%d", base_path, dump_file_seq); 3344 } 3345 dump_file_seq++; // increment seq number for next time we dump 3346 3347 HeapDumper dumper(false /* no GC before heap dump */, 3348 oome /* pass along out-of-memory-error flag */); 3349 dumper.dump(my_path, tty, HeapDumpGzipLevel); 3350 os::free(my_path); 3351 }