1 /*
   2  * Copyright (c) 2005, 2025, Oracle and/or its affiliates. All rights reserved.
   3  * Copyright (c) 2023, Alibaba Group Holding Limited. All rights reserved.
   4  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
   5  *
   6  * This code is free software; you can redistribute it and/or modify it
   7  * under the terms of the GNU General Public License version 2 only, as
   8  * published by the Free Software Foundation.
   9  *
  10  * This code is distributed in the hope that it will be useful, but WITHOUT
  11  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
  12  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
  13  * version 2 for more details (a copy is included in the LICENSE file that
  14  * accompanied this code).
  15  *
  16  * You should have received a copy of the GNU General Public License version
  17  * 2 along with this work; if not, write to the Free Software Foundation,
  18  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
  19  *
  20  * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
  21  * or visit www.oracle.com if you need additional information or have any
  22  * questions.
  23  *
  24  */
  25 
  26 #include "classfile/classLoaderData.inline.hpp"
  27 #include "classfile/classLoaderDataGraph.hpp"
  28 #include "classfile/javaClasses.inline.hpp"
  29 #include "classfile/symbolTable.hpp"
  30 #include "classfile/vmClasses.hpp"
  31 #include "classfile/vmSymbols.hpp"
  32 #include "gc/shared/gcLocker.hpp"
  33 #include "gc/shared/gcVMOperations.hpp"
  34 #include "gc/shared/workerThread.hpp"
  35 #include "jfr/jfrEvents.hpp"
  36 #include "jvm.h"
  37 #include "memory/allocation.inline.hpp"
  38 #include "memory/resourceArea.hpp"
  39 #include "memory/universe.hpp"
  40 #include "oops/fieldStreams.inline.hpp"
  41 #include "oops/klass.inline.hpp"
  42 #include "oops/objArrayKlass.hpp"
  43 #include "oops/objArrayOop.inline.hpp"
  44 #include "oops/oop.inline.hpp"
  45 #include "oops/typeArrayOop.inline.hpp"
  46 #include "runtime/continuationWrapper.inline.hpp"
  47 #include "runtime/frame.inline.hpp"
  48 #include "runtime/handles.inline.hpp"
  49 #include "runtime/javaCalls.hpp"
  50 #include "runtime/javaThread.inline.hpp"
  51 #include "runtime/jniHandles.hpp"
  52 #include "runtime/os.hpp"
  53 #include "runtime/threads.hpp"
  54 #include "runtime/threadSMR.hpp"
  55 #include "runtime/vframe.hpp"
  56 #include "runtime/vmOperations.hpp"
  57 #include "runtime/vmThread.hpp"
  58 #include "runtime/timerTrace.hpp"
  59 #include "services/heapDumper.hpp"
  60 #include "services/heapDumperCompression.hpp"
  61 #include "services/threadService.hpp"
  62 #include "utilities/checkedCast.hpp"
  63 #include "utilities/macros.hpp"
  64 #include "utilities/ostream.hpp"
  65 #ifdef LINUX
  66 #include "os_linux.hpp"
  67 #endif
  68 
  69 /*
  70  * HPROF binary format - description copied from:
  71  *   src/share/demo/jvmti/hprof/hprof_io.c
  72  *
  73  *
  74  *  header    "JAVA PROFILE 1.0.2" (0-terminated)
  75  *
  76  *  u4        size of identifiers. Identifiers are used to represent
  77  *            UTF8 strings, objects, stack traces, etc. They usually
  78  *            have the same size as host pointers.
  79  * u4         high word
  80  * u4         low word    number of milliseconds since 0:00 GMT, 1/1/70
  81  * [record]*  a sequence of records.
  82  *
  83  *
  84  * Record format:
  85  *
  86  * u1         a TAG denoting the type of the record
  87  * u4         number of *microseconds* since the time stamp in the
  88  *            header. (wraps around in a little more than an hour)
  89  * u4         number of bytes *remaining* in the record. Note that
  90  *            this number excludes the tag and the length field itself.
  91  * [u1]*      BODY of the record (a sequence of bytes)
  92  *
  93  *
  94  * The following TAGs are supported:
  95  *
  96  * TAG           BODY       notes
  97  *----------------------------------------------------------
  98  * HPROF_UTF8               a UTF8-encoded name
  99  *
 100  *               id         name ID
 101  *               [u1]*      UTF8 characters (no trailing zero)
 102  *
 103  * HPROF_LOAD_CLASS         a newly loaded class
 104  *
 105  *                u4        class serial number (> 0)
 106  *                id        class object ID
 107  *                u4        stack trace serial number
 108  *                id        class name ID
 109  *
 110  * HPROF_UNLOAD_CLASS       an unloading class
 111  *
 112  *                u4        class serial_number
 113  *
 114  * HPROF_FRAME              a Java stack frame
 115  *
 116  *                id        stack frame ID
 117  *                id        method name ID
 118  *                id        method signature ID
 119  *                id        source file name ID
 120  *                u4        class serial number
 121  *                i4        line number. >0: normal
 122  *                                       -1: unknown
 123  *                                       -2: compiled method
 124  *                                       -3: native method
 125  *
 126  * HPROF_TRACE              a Java stack trace
 127  *
 128  *               u4         stack trace serial number
 129  *               u4         thread serial number
 130  *               u4         number of frames
 131  *               [id]*      stack frame IDs
 132  *
 133  *
 134  * HPROF_ALLOC_SITES        a set of heap allocation sites, obtained after GC
 135  *
 136  *               u2         flags 0x0001: incremental vs. complete
 137  *                                0x0002: sorted by allocation vs. live
 138  *                                0x0004: whether to force a GC
 139  *               u4         cutoff ratio
 140  *               u4         total live bytes
 141  *               u4         total live instances
 142  *               u8         total bytes allocated
 143  *               u8         total instances allocated
 144  *               u4         number of sites that follow
 145  *               [u1        is_array: 0:  normal object
 146  *                                    2:  object array
 147  *                                    4:  boolean array
 148  *                                    5:  char array
 149  *                                    6:  float array
 150  *                                    7:  double array
 151  *                                    8:  byte array
 152  *                                    9:  short array
 153  *                                    10: int array
 154  *                                    11: long array
 155  *                u4        class serial number (may be zero during startup)
 156  *                u4        stack trace serial number
 157  *                u4        number of bytes alive
 158  *                u4        number of instances alive
 159  *                u4        number of bytes allocated
 160  *                u4]*      number of instance allocated
 161  *
 162  * HPROF_START_THREAD       a newly started thread.
 163  *
 164  *               u4         thread serial number (> 0)
 165  *               id         thread object ID
 166  *               u4         stack trace serial number
 167  *               id         thread name ID
 168  *               id         thread group name ID
 169  *               id         thread group parent name ID
 170  *
 171  * HPROF_END_THREAD         a terminating thread.
 172  *
 173  *               u4         thread serial number
 174  *
 175  * HPROF_HEAP_SUMMARY       heap summary
 176  *
 177  *               u4         total live bytes
 178  *               u4         total live instances
 179  *               u8         total bytes allocated
 180  *               u8         total instances allocated
 181  *
 182  * HPROF_HEAP_DUMP          denote a heap dump
 183  *
 184  *               [heap dump sub-records]*
 185  *
 186  *                          There are four kinds of heap dump sub-records:
 187  *
 188  *               u1         sub-record type
 189  *
 190  *               HPROF_GC_ROOT_UNKNOWN         unknown root
 191  *
 192  *                          id         object ID
 193  *
 194  *               HPROF_GC_ROOT_THREAD_OBJ      thread object
 195  *
 196  *                          id         thread object ID  (may be 0 for a
 197  *                                     thread newly attached through JNI)
 198  *                          u4         thread sequence number
 199  *                          u4         stack trace sequence number
 200  *
 201  *               HPROF_GC_ROOT_JNI_GLOBAL      JNI global ref root
 202  *
 203  *                          id         object ID
 204  *                          id         JNI global ref ID
 205  *
 206  *               HPROF_GC_ROOT_JNI_LOCAL       JNI local ref
 207  *
 208  *                          id         object ID
 209  *                          u4         thread serial number
 210  *                          u4         frame # in stack trace (-1 for empty)
 211  *
 212  *               HPROF_GC_ROOT_JAVA_FRAME      Java stack frame
 213  *
 214  *                          id         object ID
 215  *                          u4         thread serial number
 216  *                          u4         frame # in stack trace (-1 for empty)
 217  *
 218  *               HPROF_GC_ROOT_NATIVE_STACK    Native stack
 219  *
 220  *                          id         object ID
 221  *                          u4         thread serial number
 222  *
 223  *               HPROF_GC_ROOT_STICKY_CLASS    System class
 224  *
 225  *                          id         object ID
 226  *
 227  *               HPROF_GC_ROOT_THREAD_BLOCK    Reference from thread block
 228  *
 229  *                          id         object ID
 230  *                          u4         thread serial number
 231  *
 232  *               HPROF_GC_ROOT_MONITOR_USED    Busy monitor
 233  *
 234  *                          id         object ID
 235  *
 236  *               HPROF_GC_CLASS_DUMP           dump of a class object
 237  *
 238  *                          id         class object ID
 239  *                          u4         stack trace serial number
 240  *                          id         super class object ID
 241  *                          id         class loader object ID
 242  *                          id         signers object ID
 243  *                          id         protection domain object ID
 244  *                          id         reserved
 245  *                          id         reserved
 246  *
 247  *                          u4         instance size (in bytes)
 248  *
 249  *                          u2         size of constant pool
 250  *                          [u2,       constant pool index,
 251  *                           ty,       type
 252  *                                     2:  object
 253  *                                     4:  boolean
 254  *                                     5:  char
 255  *                                     6:  float
 256  *                                     7:  double
 257  *                                     8:  byte
 258  *                                     9:  short
 259  *                                     10: int
 260  *                                     11: long
 261  *                           vl]*      and value
 262  *
 263  *                          u2         number of static fields
 264  *                          [id,       static field name,
 265  *                           ty,       type,
 266  *                           vl]*      and value
 267  *
 268  *                          u2         number of inst. fields (not inc. super)
 269  *                          [id,       instance field name,
 270  *                           ty]*      type
 271  *
 272  *               HPROF_GC_INSTANCE_DUMP        dump of a normal object
 273  *
 274  *                          id         object ID
 275  *                          u4         stack trace serial number
 276  *                          id         class object ID
 277  *                          u4         number of bytes that follow
 278  *                          [vl]*      instance field values (class, followed
 279  *                                     by super, super's super ...)
 280  *
 281  *               HPROF_GC_OBJ_ARRAY_DUMP       dump of an object array
 282  *
 283  *                          id         array object ID
 284  *                          u4         stack trace serial number
 285  *                          u4         number of elements
 286  *                          id         array class ID
 287  *                          [id]*      elements
 288  *
 289  *               HPROF_GC_PRIM_ARRAY_DUMP      dump of a primitive array
 290  *
 291  *                          id         array object ID
 292  *                          u4         stack trace serial number
 293  *                          u4         number of elements
 294  *                          u1         element type
 295  *                                     4:  boolean array
 296  *                                     5:  char array
 297  *                                     6:  float array
 298  *                                     7:  double array
 299  *                                     8:  byte array
 300  *                                     9:  short array
 301  *                                     10: int array
 302  *                                     11: long array
 303  *                          [u1]*      elements
 304  *
 305  * HPROF_CPU_SAMPLES        a set of sample traces of running threads
 306  *
 307  *                u4        total number of samples
 308  *                u4        # of traces
 309  *               [u4        # of samples
 310  *                u4]*      stack trace serial number
 311  *
 312  * HPROF_CONTROL_SETTINGS   the settings of on/off switches
 313  *
 314  *                u4        0x00000001: alloc traces on/off
 315  *                          0x00000002: cpu sampling on/off
 316  *                u2        stack trace depth
 317  *
 318  *
 319  * When the header is "JAVA PROFILE 1.0.2" a heap dump can optionally
 320  * be generated as a sequence of heap dump segments. This sequence is
 321  * terminated by an end record. The additional tags allowed by format
 322  * "JAVA PROFILE 1.0.2" are:
 323  *
 324  * HPROF_HEAP_DUMP_SEGMENT  denote a heap dump segment
 325  *
 326  *               [heap dump sub-records]*
 327  *               The same sub-record types allowed by HPROF_HEAP_DUMP
 328  *
 329  * HPROF_HEAP_DUMP_END      denotes the end of a heap dump
 330  *
 331  */
 332 
 333 
 334 // HPROF tags
 335 
 336 enum hprofTag : u1 {
 337   // top-level records
 338   HPROF_UTF8                    = 0x01,
 339   HPROF_LOAD_CLASS              = 0x02,
 340   HPROF_UNLOAD_CLASS            = 0x03,
 341   HPROF_FRAME                   = 0x04,
 342   HPROF_TRACE                   = 0x05,
 343   HPROF_ALLOC_SITES             = 0x06,
 344   HPROF_HEAP_SUMMARY            = 0x07,
 345   HPROF_START_THREAD            = 0x0A,
 346   HPROF_END_THREAD              = 0x0B,
 347   HPROF_HEAP_DUMP               = 0x0C,
 348   HPROF_CPU_SAMPLES             = 0x0D,
 349   HPROF_CONTROL_SETTINGS        = 0x0E,
 350 
 351   // 1.0.2 record types
 352   HPROF_HEAP_DUMP_SEGMENT       = 0x1C,
 353   HPROF_HEAP_DUMP_END           = 0x2C,
 354 
 355   // field types
 356   HPROF_ARRAY_OBJECT            = 0x01,
 357   HPROF_NORMAL_OBJECT           = 0x02,
 358   HPROF_BOOLEAN                 = 0x04,
 359   HPROF_CHAR                    = 0x05,
 360   HPROF_FLOAT                   = 0x06,
 361   HPROF_DOUBLE                  = 0x07,
 362   HPROF_BYTE                    = 0x08,
 363   HPROF_SHORT                   = 0x09,
 364   HPROF_INT                     = 0x0A,
 365   HPROF_LONG                    = 0x0B,
 366 
 367   // data-dump sub-records
 368   HPROF_GC_ROOT_UNKNOWN         = 0xFF,
 369   HPROF_GC_ROOT_JNI_GLOBAL      = 0x01,
 370   HPROF_GC_ROOT_JNI_LOCAL       = 0x02,
 371   HPROF_GC_ROOT_JAVA_FRAME      = 0x03,
 372   HPROF_GC_ROOT_NATIVE_STACK    = 0x04,
 373   HPROF_GC_ROOT_STICKY_CLASS    = 0x05,
 374   HPROF_GC_ROOT_THREAD_BLOCK    = 0x06,
 375   HPROF_GC_ROOT_MONITOR_USED    = 0x07,
 376   HPROF_GC_ROOT_THREAD_OBJ      = 0x08,
 377   HPROF_GC_CLASS_DUMP           = 0x20,
 378   HPROF_GC_INSTANCE_DUMP        = 0x21,
 379   HPROF_GC_OBJ_ARRAY_DUMP       = 0x22,
 380   HPROF_GC_PRIM_ARRAY_DUMP      = 0x23
 381 };
 382 
 383 // Default stack trace ID (used for dummy HPROF_TRACE record)
 384 enum {
 385   STACK_TRACE_ID = 1,
 386   INITIAL_CLASS_COUNT = 200
 387 };
 388 
 389 // Supports I/O operations for a dump
 390 // Base class for dump and parallel dump
 391 class AbstractDumpWriter : public CHeapObj<mtInternal> {
 392  protected:
 393   enum {
 394     io_buffer_max_size = 1*M,
 395     dump_segment_header_size = 9
 396   };
 397 
 398   char* _buffer;    // internal buffer
 399   size_t _size;
 400   size_t _pos;
 401 
 402   bool _in_dump_segment; // Are we currently in a dump segment?
 403   bool _is_huge_sub_record; // Are we writing a sub-record larger than the buffer size?
 404   DEBUG_ONLY(size_t _sub_record_left;) // The bytes not written for the current sub-record.
 405   DEBUG_ONLY(bool _sub_record_ended;) // True if we have called the end_sub_record().
 406 
 407   char* buffer() const                          { return _buffer; }
 408   size_t buffer_size() const                    { return _size; }
 409   void set_position(size_t pos)                 { _pos = pos; }
 410 
 411   // Can be called if we have enough room in the buffer.
 412   void write_fast(const void* s, size_t len);
 413 
 414   // Returns true if we have enough room in the buffer for 'len' bytes.
 415   bool can_write_fast(size_t len);
 416 
 417   void write_address(address a);
 418 
 419  public:
 420   AbstractDumpWriter() :
 421     _buffer(nullptr),
 422     _size(io_buffer_max_size),
 423     _pos(0),
 424     _in_dump_segment(false) { }
 425 
 426   // Total number of bytes written to the disk
 427   virtual julong bytes_written() const = 0;
 428   // Return non-null if error occurred
 429   virtual char const* error() const = 0;
 430 
 431   size_t position() const                       { return _pos; }
 432   // writer functions
 433   virtual void write_raw(const void* s, size_t len);
 434   void write_u1(u1 x);
 435   void write_u2(u2 x);
 436   void write_u4(u4 x);
 437   void write_u8(u8 x);
 438   void write_objectID(oop o);
 439   void write_rootID(oop* p);
 440   void write_symbolID(Symbol* o);
 441   void write_classID(Klass* k);
 442   void write_id(u4 x);
 443 
 444   // Start a new sub-record. Starts a new heap dump segment if needed.
 445   void start_sub_record(u1 tag, u4 len);
 446   // Ends the current sub-record.
 447   void end_sub_record();
 448   // Finishes the current dump segment if not already finished.
 449   void finish_dump_segment();
 450   // Flush internal buffer to persistent storage
 451   virtual void flush() = 0;
 452 };
 453 
 454 void AbstractDumpWriter::write_fast(const void* s, size_t len) {
 455   assert(!_in_dump_segment || (_sub_record_left >= len), "sub-record too large");
 456   assert(buffer_size() - position() >= len, "Must fit");
 457   debug_only(_sub_record_left -= len);
 458   memcpy(buffer() + position(), s, len);
 459   set_position(position() + len);
 460 }
 461 
 462 bool AbstractDumpWriter::can_write_fast(size_t len) {
 463   return buffer_size() - position() >= len;
 464 }
 465 
 466 // write raw bytes
 467 void AbstractDumpWriter::write_raw(const void* s, size_t len) {
 468   assert(!_in_dump_segment || (_sub_record_left >= len), "sub-record too large");
 469   debug_only(_sub_record_left -= len);
 470 
 471   // flush buffer to make room.
 472   while (len > buffer_size() - position()) {
 473     assert(!_in_dump_segment || _is_huge_sub_record,
 474            "Cannot overflow in non-huge sub-record.");
 475     size_t to_write = buffer_size() - position();
 476     memcpy(buffer() + position(), s, to_write);
 477     s = (void*) ((char*) s + to_write);
 478     len -= to_write;
 479     set_position(position() + to_write);
 480     flush();
 481   }
 482 
 483   memcpy(buffer() + position(), s, len);
 484   set_position(position() + len);
 485 }
 486 
 487 // Makes sure we inline the fast write into the write_u* functions. This is a big speedup.
 488 #define WRITE_KNOWN_TYPE(p, len) do { if (can_write_fast((len))) write_fast((p), (len)); \
 489                                       else write_raw((p), (len)); } while (0)
 490 
 491 void AbstractDumpWriter::write_u1(u1 x) {
 492   WRITE_KNOWN_TYPE(&x, 1);
 493 }
 494 
 495 void AbstractDumpWriter::write_u2(u2 x) {
 496   u2 v;
 497   Bytes::put_Java_u2((address)&v, x);
 498   WRITE_KNOWN_TYPE(&v, 2);
 499 }
 500 
 501 void AbstractDumpWriter::write_u4(u4 x) {
 502   u4 v;
 503   Bytes::put_Java_u4((address)&v, x);
 504   WRITE_KNOWN_TYPE(&v, 4);
 505 }
 506 
 507 void AbstractDumpWriter::write_u8(u8 x) {
 508   u8 v;
 509   Bytes::put_Java_u8((address)&v, x);
 510   WRITE_KNOWN_TYPE(&v, 8);
 511 }
 512 
 513 void AbstractDumpWriter::write_address(address a) {
 514 #ifdef _LP64
 515   write_u8((u8)a);
 516 #else
 517   write_u4((u4)a);
 518 #endif
 519 }
 520 
 521 void AbstractDumpWriter::write_objectID(oop o) {
 522   write_address(cast_from_oop<address>(o));
 523 }
 524 
 525 void AbstractDumpWriter::write_rootID(oop* p) {
 526   write_address((address)p);
 527 }
 528 
 529 void AbstractDumpWriter::write_symbolID(Symbol* s) {
 530   write_address((address)((uintptr_t)s));
 531 }
 532 
 533 void AbstractDumpWriter::write_id(u4 x) {
 534 #ifdef _LP64
 535   write_u8((u8) x);
 536 #else
 537   write_u4(x);
 538 #endif
 539 }
 540 
 541 // We use java mirror as the class ID
 542 void AbstractDumpWriter::write_classID(Klass* k) {
 543   write_objectID(k->java_mirror());
 544 }
 545 
 546 void AbstractDumpWriter::finish_dump_segment() {
 547   if (_in_dump_segment) {
 548     assert(_sub_record_left == 0, "Last sub-record not written completely");
 549     assert(_sub_record_ended, "sub-record must have ended");
 550 
 551     // Fix up the dump segment length if we haven't written a huge sub-record last
 552     // (in which case the segment length was already set to the correct value initially).
 553     if (!_is_huge_sub_record) {
 554       assert(position() > dump_segment_header_size, "Dump segment should have some content");
 555       Bytes::put_Java_u4((address) (buffer() + 5),
 556                          (u4) (position() - dump_segment_header_size));
 557     } else {
 558       // Finish process huge sub record
 559       // Set _is_huge_sub_record to false so the parallel dump writer can flush data to file.
 560       _is_huge_sub_record = false;
 561     }
 562 
 563     _in_dump_segment = false;
 564     flush();
 565   }
 566 }
 567 
 568 void AbstractDumpWriter::start_sub_record(u1 tag, u4 len) {
 569   if (!_in_dump_segment) {
 570     if (position() > 0) {
 571       flush();
 572     }
 573 
 574     assert(position() == 0 && buffer_size() > dump_segment_header_size, "Must be at the start");
 575 
 576     write_u1(HPROF_HEAP_DUMP_SEGMENT);
 577     write_u4(0); // timestamp
 578     // Will be fixed up later if we add more sub-records.  If this is a huge sub-record,
 579     // this is already the correct length, since we don't add more sub-records.
 580     write_u4(len);
 581     assert(Bytes::get_Java_u4((address)(buffer() + 5)) == len, "Inconsistent size!");
 582     _in_dump_segment = true;
 583     _is_huge_sub_record = len > buffer_size() - dump_segment_header_size;
 584   } else if (_is_huge_sub_record || (len > buffer_size() - position())) {
 585     // This object will not fit in completely or the last sub-record was huge.
 586     // Finish the current segment and try again.
 587     finish_dump_segment();
 588     start_sub_record(tag, len);
 589 
 590     return;
 591   }
 592 
 593   debug_only(_sub_record_left = len);
 594   debug_only(_sub_record_ended = false);
 595 
 596   write_u1(tag);
 597 }
 598 
 599 void AbstractDumpWriter::end_sub_record() {
 600   assert(_in_dump_segment, "must be in dump segment");
 601   assert(_sub_record_left == 0, "sub-record not written completely");
 602   assert(!_sub_record_ended, "Must not have ended yet");
 603   debug_only(_sub_record_ended = true);
 604 }
 605 
 606 // Supports I/O operations for a dump
 607 
 608 class DumpWriter : public AbstractDumpWriter {
 609 private:
 610   FileWriter* _writer;
 611   AbstractCompressor* _compressor;
 612   size_t _bytes_written;
 613   char* _error;
 614   // Compression support
 615   char* _out_buffer;
 616   size_t _out_size;
 617   size_t _out_pos;
 618   char* _tmp_buffer;
 619   size_t _tmp_size;
 620 
 621 private:
 622   void do_compress();
 623 
 624 public:
 625   DumpWriter(const char* path, bool overwrite, AbstractCompressor* compressor);
 626   ~DumpWriter();
 627   julong bytes_written() const override        { return (julong) _bytes_written; }
 628   char const* error() const override           { return _error; }
 629   void set_error(const char* error)            { _error = (char*)error; }
 630   bool has_error() const                       { return _error != nullptr; }
 631   const char* get_file_path() const            { return _writer->get_file_path(); }
 632   AbstractCompressor* compressor()             { return _compressor; }
 633   bool is_overwrite() const                    { return _writer->is_overwrite(); }
 634 
 635   void flush() override;
 636 
 637 private:
 638   // internals for DumpMerger
 639   friend class DumpMerger;
 640   void set_bytes_written(julong bytes_written) { _bytes_written = bytes_written; }
 641   int get_fd() const                           { return _writer->get_fd(); }
 642   void set_compressor(AbstractCompressor* p)   { _compressor = p; }
 643 };
 644 
 645 DumpWriter::DumpWriter(const char* path, bool overwrite, AbstractCompressor* compressor) :
 646   AbstractDumpWriter(),
 647   _writer(new (std::nothrow) FileWriter(path, overwrite)),
 648   _compressor(compressor),
 649   _bytes_written(0),
 650   _error(nullptr),
 651   _out_buffer(nullptr),
 652   _out_size(0),
 653   _out_pos(0),
 654   _tmp_buffer(nullptr),
 655   _tmp_size(0) {
 656   _error = (char*)_writer->open_writer();
 657   if (_error == nullptr) {
 658     _buffer = (char*)os::malloc(io_buffer_max_size, mtInternal);
 659     if (compressor != nullptr) {
 660       _error = (char*)_compressor->init(io_buffer_max_size, &_out_size, &_tmp_size);
 661       if (_error == nullptr) {
 662         if (_out_size > 0) {
 663           _out_buffer = (char*)os::malloc(_out_size, mtInternal);
 664         }
 665         if (_tmp_size > 0) {
 666           _tmp_buffer = (char*)os::malloc(_tmp_size, mtInternal);
 667         }
 668       }
 669     }
 670   }
 671   // initialize internal buffer
 672   _pos = 0;
 673   _size = io_buffer_max_size;
 674 }
 675 
 676 DumpWriter::~DumpWriter(){
 677   if (_buffer != nullptr) {
 678     os::free(_buffer);
 679   }
 680   if (_out_buffer != nullptr) {
 681     os::free(_out_buffer);
 682   }
 683   if (_tmp_buffer != nullptr) {
 684     os::free(_tmp_buffer);
 685   }
 686   if (_writer != nullptr) {
 687     delete _writer;
 688   }
 689   _bytes_written = -1;
 690 }
 691 
 692 // flush any buffered bytes to the file
 693 void DumpWriter::flush() {
 694   if (_pos <= 0) {
 695     return;
 696   }
 697   if (has_error()) {
 698     _pos = 0;
 699     return;
 700   }
 701   char* result = nullptr;
 702   if (_compressor == nullptr) {
 703     result = (char*)_writer->write_buf(_buffer, _pos);
 704     _bytes_written += _pos;
 705   } else {
 706     do_compress();
 707     if (!has_error()) {
 708       result = (char*)_writer->write_buf(_out_buffer, _out_pos);
 709       _bytes_written += _out_pos;
 710     }
 711   }
 712   _pos = 0; // reset pos to make internal buffer available
 713 
 714   if (result != nullptr) {
 715     set_error(result);
 716   }
 717 }
 718 
 719 void DumpWriter::do_compress() {
 720   const char* msg = _compressor->compress(_buffer, _pos, _out_buffer, _out_size,
 721                                           _tmp_buffer, _tmp_size, &_out_pos);
 722 
 723   if (msg != nullptr) {
 724     set_error(msg);
 725   }
 726 }
 727 
 728 class DumperClassCacheTable;
 729 class DumperClassCacheTableEntry;
 730 
 731 // Support class with a collection of functions used when dumping the heap
 732 class DumperSupport : AllStatic {
 733  public:
 734 
 735   // write a header of the given type
 736   static void write_header(AbstractDumpWriter* writer, hprofTag tag, u4 len);
 737 
 738   // returns hprof tag for the given type signature
 739   static hprofTag sig2tag(Symbol* sig);
 740   // returns hprof tag for the given basic type
 741   static hprofTag type2tag(BasicType type);
 742   // Returns the size of the data to write.
 743   static u4 sig2size(Symbol* sig);
 744 
 745   // returns the size of the instance of the given class
 746   static u4 instance_size(InstanceKlass* ik, DumperClassCacheTableEntry* class_cache_entry = nullptr);
 747 
 748   // dump a jfloat
 749   static void dump_float(AbstractDumpWriter* writer, jfloat f);
 750   // dump a jdouble
 751   static void dump_double(AbstractDumpWriter* writer, jdouble d);
 752   // dumps the raw value of the given field
 753   static void dump_field_value(AbstractDumpWriter* writer, char type, oop obj, int offset);
 754   // returns the size of the static fields; also counts the static fields
 755   static u4 get_static_fields_size(InstanceKlass* ik, u2& field_count);
 756   // dumps static fields of the given class
 757   static void dump_static_fields(AbstractDumpWriter* writer, Klass* k);
 758   // dump the raw values of the instance fields of the given object
 759   static void dump_instance_fields(AbstractDumpWriter* writer, oop o, DumperClassCacheTableEntry* class_cache_entry);
 760   // get the count of the instance fields for a given class
 761   static u2 get_instance_fields_count(InstanceKlass* ik);
 762   // dumps the definition of the instance fields for a given class
 763   static void dump_instance_field_descriptors(AbstractDumpWriter* writer, Klass* k);
 764   // creates HPROF_GC_INSTANCE_DUMP record for the given object
 765   static void dump_instance(AbstractDumpWriter* writer, oop o, DumperClassCacheTable* class_cache);
 766   // creates HPROF_GC_CLASS_DUMP record for the given instance class
 767   static void dump_instance_class(AbstractDumpWriter* writer, Klass* k);
 768   // creates HPROF_GC_CLASS_DUMP record for a given array class
 769   static void dump_array_class(AbstractDumpWriter* writer, Klass* k);
 770 
 771   // creates HPROF_GC_OBJ_ARRAY_DUMP record for the given object array
 772   static void dump_object_array(AbstractDumpWriter* writer, objArrayOop array);
 773   // creates HPROF_GC_PRIM_ARRAY_DUMP record for the given type array
 774   static void dump_prim_array(AbstractDumpWriter* writer, typeArrayOop array);
 775   // create HPROF_FRAME record for the given method and bci
 776   static void dump_stack_frame(AbstractDumpWriter* writer, int frame_serial_num, int class_serial_num, Method* m, int bci);
 777 
 778   // check if we need to truncate an array
 779   static int calculate_array_max_length(AbstractDumpWriter* writer, arrayOop array, short header_size);
 780 
 781   // fixes up the current dump record and writes HPROF_HEAP_DUMP_END record
 782   static void end_of_dump(AbstractDumpWriter* writer);
 783 
 784   static oop mask_dormant_archived_object(oop o, oop ref_obj) {
 785     if (o != nullptr && o->klass()->java_mirror_no_keepalive() == nullptr) {
 786       // Ignore this object since the corresponding java mirror is not loaded.
 787       // Might be a dormant archive object.
 788       report_dormant_archived_object(o, ref_obj);
 789       return nullptr;
 790     } else {
 791       return o;
 792     }
 793   }
 794 
 795   static void report_dormant_archived_object(oop o, oop ref_obj) {
 796     if (log_is_enabled(Trace, cds, heap)) {
 797       ResourceMark rm;
 798       if (ref_obj != nullptr) {
 799         log_trace(cds, heap)("skipped dormant archived object " INTPTR_FORMAT " (%s) referenced by " INTPTR_FORMAT " (%s)",
 800                   p2i(o), o->klass()->external_name(),
 801                   p2i(ref_obj), ref_obj->klass()->external_name());
 802       } else {
 803         log_trace(cds, heap)("skipped dormant archived object " INTPTR_FORMAT " (%s)",
 804                   p2i(o), o->klass()->external_name());
 805       }
 806     }
 807   }
 808 };
 809 
 810 // Hash table of klasses to the klass metadata. This should greatly improve the
 811 // hash dumping performance. This hash table is supposed to be used by a single
 812 // thread only.
 813 //
 814 class DumperClassCacheTableEntry : public CHeapObj<mtServiceability> {
 815   friend class DumperClassCacheTable;
 816 private:
 817   GrowableArray<char> _sigs_start;
 818   GrowableArray<int> _offsets;
 819   u4 _instance_size;
 820   int _entries;
 821 
 822 public:
 823   DumperClassCacheTableEntry() : _instance_size(0), _entries(0) {};
 824 
 825   int field_count()             { return _entries; }
 826   char sig_start(int field_idx) { return _sigs_start.at(field_idx); }
 827   int offset(int field_idx)     { return _offsets.at(field_idx); }
 828   u4 instance_size()            { return _instance_size; }
 829 };
 830 
 831 class DumperClassCacheTable {
 832 private:
 833   // ResourceHashtable SIZE is specified at compile time so we
 834   // use 1031 which is the first prime after 1024.
 835   static constexpr size_t TABLE_SIZE = 1031;
 836 
 837   // Maintain the cache for N classes. This limits memory footprint
 838   // impact, regardless of how many classes we have in the dump.
 839   // This also improves look up performance by keeping the statically
 840   // sized table from overloading.
 841   static constexpr int CACHE_TOP = 256;
 842 
 843   typedef ResourceHashtable<InstanceKlass*, DumperClassCacheTableEntry*,
 844                             TABLE_SIZE, AnyObj::C_HEAP, mtServiceability> PtrTable;
 845   PtrTable* _ptrs;
 846 
 847   // Single-slot cache to handle the major case of objects of the same
 848   // class back-to-back, e.g. from T[].
 849   InstanceKlass* _last_ik;
 850   DumperClassCacheTableEntry* _last_entry;
 851 
 852   void unlink_all(PtrTable* table) {
 853     class CleanupEntry: StackObj {
 854     public:
 855       bool do_entry(InstanceKlass*& key, DumperClassCacheTableEntry*& entry) {
 856         delete entry;
 857         return true;
 858       }
 859     } cleanup;
 860     table->unlink(&cleanup);
 861   }
 862 
 863 public:
 864   DumperClassCacheTableEntry* lookup_or_create(InstanceKlass* ik) {
 865     if (_last_ik == ik) {
 866       return _last_entry;
 867     }
 868 
 869     DumperClassCacheTableEntry* entry;
 870     DumperClassCacheTableEntry** from_cache = _ptrs->get(ik);
 871     if (from_cache == nullptr) {
 872       entry = new DumperClassCacheTableEntry();
 873       for (HierarchicalFieldStream<JavaFieldStream> fld(ik); !fld.done(); fld.next()) {
 874         if (!fld.access_flags().is_static()) {
 875           Symbol* sig = fld.signature();
 876           entry->_sigs_start.push(sig->char_at(0));
 877           entry->_offsets.push(fld.offset());
 878           entry->_entries++;
 879           entry->_instance_size += DumperSupport::sig2size(sig);
 880         }
 881       }
 882 
 883       if (_ptrs->number_of_entries() >= CACHE_TOP) {
 884         // We do not track the individual hit rates for table entries.
 885         // Purge the entire table, and let the cache catch up with new
 886         // distribution.
 887         unlink_all(_ptrs);
 888       }
 889 
 890       _ptrs->put(ik, entry);
 891     } else {
 892       entry = *from_cache;
 893     }
 894 
 895     // Remember for single-slot cache.
 896     _last_ik = ik;
 897     _last_entry = entry;
 898 
 899     return entry;
 900   }
 901 
 902   DumperClassCacheTable() : _ptrs(new (mtServiceability) PtrTable), _last_ik(nullptr), _last_entry(nullptr) {}
 903 
 904   ~DumperClassCacheTable() {
 905     unlink_all(_ptrs);
 906     delete _ptrs;
 907   }
 908 };
 909 
 910 // write a header of the given type
 911 void DumperSupport:: write_header(AbstractDumpWriter* writer, hprofTag tag, u4 len) {
 912   writer->write_u1(tag);
 913   writer->write_u4(0);                  // current ticks
 914   writer->write_u4(len);
 915 }
 916 
 917 // returns hprof tag for the given type signature
 918 hprofTag DumperSupport::sig2tag(Symbol* sig) {
 919   switch (sig->char_at(0)) {
 920     case JVM_SIGNATURE_CLASS    : return HPROF_NORMAL_OBJECT;
 921     case JVM_SIGNATURE_ARRAY    : return HPROF_NORMAL_OBJECT;
 922     case JVM_SIGNATURE_BYTE     : return HPROF_BYTE;
 923     case JVM_SIGNATURE_CHAR     : return HPROF_CHAR;
 924     case JVM_SIGNATURE_FLOAT    : return HPROF_FLOAT;
 925     case JVM_SIGNATURE_DOUBLE   : return HPROF_DOUBLE;
 926     case JVM_SIGNATURE_INT      : return HPROF_INT;
 927     case JVM_SIGNATURE_LONG     : return HPROF_LONG;
 928     case JVM_SIGNATURE_SHORT    : return HPROF_SHORT;
 929     case JVM_SIGNATURE_BOOLEAN  : return HPROF_BOOLEAN;
 930     default : ShouldNotReachHere(); /* to shut up compiler */ return HPROF_BYTE;
 931   }
 932 }
 933 
 934 hprofTag DumperSupport::type2tag(BasicType type) {
 935   switch (type) {
 936     case T_BYTE     : return HPROF_BYTE;
 937     case T_CHAR     : return HPROF_CHAR;
 938     case T_FLOAT    : return HPROF_FLOAT;
 939     case T_DOUBLE   : return HPROF_DOUBLE;
 940     case T_INT      : return HPROF_INT;
 941     case T_LONG     : return HPROF_LONG;
 942     case T_SHORT    : return HPROF_SHORT;
 943     case T_BOOLEAN  : return HPROF_BOOLEAN;
 944     default : ShouldNotReachHere(); /* to shut up compiler */ return HPROF_BYTE;
 945   }
 946 }
 947 
 948 u4 DumperSupport::sig2size(Symbol* sig) {
 949   switch (sig->char_at(0)) {
 950     case JVM_SIGNATURE_CLASS:
 951     case JVM_SIGNATURE_ARRAY: return sizeof(address);
 952     case JVM_SIGNATURE_BOOLEAN:
 953     case JVM_SIGNATURE_BYTE: return 1;
 954     case JVM_SIGNATURE_SHORT:
 955     case JVM_SIGNATURE_CHAR: return 2;
 956     case JVM_SIGNATURE_INT:
 957     case JVM_SIGNATURE_FLOAT: return 4;
 958     case JVM_SIGNATURE_LONG:
 959     case JVM_SIGNATURE_DOUBLE: return 8;
 960     default: ShouldNotReachHere(); /* to shut up compiler */ return 0;
 961   }
 962 }
 963 
 964 template<typename T, typename F> T bit_cast(F from) { // replace with the real thing when we can use c++20
 965   T to;
 966   static_assert(sizeof(to) == sizeof(from), "must be of the same size");
 967   memcpy(&to, &from, sizeof(to));
 968   return to;
 969 }
 970 
 971 // dump a jfloat
 972 void DumperSupport::dump_float(AbstractDumpWriter* writer, jfloat f) {
 973   if (g_isnan(f)) {
 974     writer->write_u4(0x7fc00000); // collapsing NaNs
 975   } else {
 976     writer->write_u4(bit_cast<u4>(f));
 977   }
 978 }
 979 
 980 // dump a jdouble
 981 void DumperSupport::dump_double(AbstractDumpWriter* writer, jdouble d) {
 982   if (g_isnan(d)) {
 983     writer->write_u8(0x7ff80000ull << 32); // collapsing NaNs
 984   } else {
 985     writer->write_u8(bit_cast<u8>(d));
 986   }
 987 }
 988 
 989 // dumps the raw value of the given field
 990 void DumperSupport::dump_field_value(AbstractDumpWriter* writer, char type, oop obj, int offset) {
 991   switch (type) {
 992     case JVM_SIGNATURE_CLASS :
 993     case JVM_SIGNATURE_ARRAY : {
 994       oop o = obj->obj_field_access<ON_UNKNOWN_OOP_REF | AS_NO_KEEPALIVE>(offset);
 995       o = mask_dormant_archived_object(o, obj);
 996       assert(oopDesc::is_oop_or_null(o), "Expected an oop or nullptr at " PTR_FORMAT, p2i(o));
 997       writer->write_objectID(o);
 998       break;
 999     }
1000     case JVM_SIGNATURE_BYTE : {
1001       jbyte b = obj->byte_field(offset);
1002       writer->write_u1(b);
1003       break;
1004     }
1005     case JVM_SIGNATURE_CHAR : {
1006       jchar c = obj->char_field(offset);
1007       writer->write_u2(c);
1008       break;
1009     }
1010     case JVM_SIGNATURE_SHORT : {
1011       jshort s = obj->short_field(offset);
1012       writer->write_u2(s);
1013       break;
1014     }
1015     case JVM_SIGNATURE_FLOAT : {
1016       jfloat f = obj->float_field(offset);
1017       dump_float(writer, f);
1018       break;
1019     }
1020     case JVM_SIGNATURE_DOUBLE : {
1021       jdouble d = obj->double_field(offset);
1022       dump_double(writer, d);
1023       break;
1024     }
1025     case JVM_SIGNATURE_INT : {
1026       jint i = obj->int_field(offset);
1027       writer->write_u4(i);
1028       break;
1029     }
1030     case JVM_SIGNATURE_LONG : {
1031       jlong l = obj->long_field(offset);
1032       writer->write_u8(l);
1033       break;
1034     }
1035     case JVM_SIGNATURE_BOOLEAN : {
1036       jboolean b = obj->bool_field(offset);
1037       writer->write_u1(b);
1038       break;
1039     }
1040     default : {
1041       ShouldNotReachHere();
1042       break;
1043     }
1044   }
1045 }
1046 
1047 // returns the size of the instance of the given class
1048 u4 DumperSupport::instance_size(InstanceKlass* ik, DumperClassCacheTableEntry* class_cache_entry) {
1049   if (class_cache_entry != nullptr) {
1050     return class_cache_entry->instance_size();
1051   } else {
1052     u4 size = 0;
1053     for (HierarchicalFieldStream<JavaFieldStream> fld(ik); !fld.done(); fld.next()) {
1054       if (!fld.access_flags().is_static()) {
1055         size += sig2size(fld.signature());
1056       }
1057     }
1058     return size;
1059   }
1060 }
1061 
1062 u4 DumperSupport::get_static_fields_size(InstanceKlass* ik, u2& field_count) {
1063   field_count = 0;
1064   u4 size = 0;
1065 
1066   for (JavaFieldStream fldc(ik); !fldc.done(); fldc.next()) {
1067     if (fldc.access_flags().is_static()) {
1068       field_count++;
1069       size += sig2size(fldc.signature());
1070     }
1071   }
1072 
1073   // Add in resolved_references which is referenced by the cpCache
1074   // The resolved_references is an array per InstanceKlass holding the
1075   // strings and other oops resolved from the constant pool.
1076   oop resolved_references = ik->constants()->resolved_references_or_null();
1077   if (resolved_references != nullptr) {
1078     field_count++;
1079     size += sizeof(address);
1080 
1081     // Add in the resolved_references of the used previous versions of the class
1082     // in the case of RedefineClasses
1083     InstanceKlass* prev = ik->previous_versions();
1084     while (prev != nullptr && prev->constants()->resolved_references_or_null() != nullptr) {
1085       field_count++;
1086       size += sizeof(address);
1087       prev = prev->previous_versions();
1088     }
1089   }
1090 
1091   // Also provide a pointer to the init_lock if present, so there aren't unreferenced int[0]
1092   // arrays.
1093   oop init_lock = ik->init_lock();
1094   if (init_lock != nullptr) {
1095     field_count++;
1096     size += sizeof(address);
1097   }
1098 
1099   // We write the value itself plus a name and a one byte type tag per field.
1100   return checked_cast<u4>(size + field_count * (sizeof(address) + 1));
1101 }
1102 
1103 // dumps static fields of the given class
1104 void DumperSupport::dump_static_fields(AbstractDumpWriter* writer, Klass* k) {
1105   InstanceKlass* ik = InstanceKlass::cast(k);
1106 
1107   // dump the field descriptors and raw values
1108   for (JavaFieldStream fld(ik); !fld.done(); fld.next()) {
1109     if (fld.access_flags().is_static()) {
1110       Symbol* sig = fld.signature();
1111 
1112       writer->write_symbolID(fld.name());   // name
1113       writer->write_u1(sig2tag(sig));       // type
1114 
1115       // value
1116       dump_field_value(writer, sig->char_at(0), ik->java_mirror(), fld.offset());
1117     }
1118   }
1119 
1120   // Add resolved_references for each class that has them
1121   oop resolved_references = ik->constants()->resolved_references_or_null();
1122   if (resolved_references != nullptr) {
1123     writer->write_symbolID(vmSymbols::resolved_references_name());  // name
1124     writer->write_u1(sig2tag(vmSymbols::object_array_signature())); // type
1125     writer->write_objectID(resolved_references);
1126 
1127     // Also write any previous versions
1128     InstanceKlass* prev = ik->previous_versions();
1129     while (prev != nullptr && prev->constants()->resolved_references_or_null() != nullptr) {
1130       writer->write_symbolID(vmSymbols::resolved_references_name());  // name
1131       writer->write_u1(sig2tag(vmSymbols::object_array_signature())); // type
1132       writer->write_objectID(prev->constants()->resolved_references());
1133       prev = prev->previous_versions();
1134     }
1135   }
1136 
1137   // Add init lock to the end if the class is not yet initialized
1138   oop init_lock = ik->init_lock();
1139   if (init_lock != nullptr) {
1140     writer->write_symbolID(vmSymbols::init_lock_name());         // name
1141     writer->write_u1(sig2tag(vmSymbols::int_array_signature())); // type
1142     writer->write_objectID(init_lock);
1143   }
1144 }
1145 
1146 // dump the raw values of the instance fields of the given object
1147 void DumperSupport::dump_instance_fields(AbstractDumpWriter* writer, oop o, DumperClassCacheTableEntry* class_cache_entry) {
1148   assert(class_cache_entry != nullptr, "Pre-condition: must be provided");
1149   for (int idx = 0; idx < class_cache_entry->field_count(); idx++) {
1150     dump_field_value(writer, class_cache_entry->sig_start(idx), o, class_cache_entry->offset(idx));
1151   }
1152 }
1153 
1154 // dumps the definition of the instance fields for a given class
1155 u2 DumperSupport::get_instance_fields_count(InstanceKlass* ik) {
1156   u2 field_count = 0;
1157 
1158   for (JavaFieldStream fldc(ik); !fldc.done(); fldc.next()) {
1159     if (!fldc.access_flags().is_static()) field_count++;
1160   }
1161 
1162   return field_count;
1163 }
1164 
1165 // dumps the definition of the instance fields for a given class
1166 void DumperSupport::dump_instance_field_descriptors(AbstractDumpWriter* writer, Klass* k) {
1167   InstanceKlass* ik = InstanceKlass::cast(k);
1168 
1169   // dump the field descriptors
1170   for (JavaFieldStream fld(ik); !fld.done(); fld.next()) {
1171     if (!fld.access_flags().is_static()) {
1172       Symbol* sig = fld.signature();
1173 
1174       writer->write_symbolID(fld.name());   // name
1175       writer->write_u1(sig2tag(sig));       // type
1176     }
1177   }
1178 }
1179 
1180 // creates HPROF_GC_INSTANCE_DUMP record for the given object
1181 void DumperSupport::dump_instance(AbstractDumpWriter* writer, oop o, DumperClassCacheTable* class_cache) {
1182   InstanceKlass* ik = InstanceKlass::cast(o->klass());
1183 
1184   DumperClassCacheTableEntry* cache_entry = class_cache->lookup_or_create(ik);
1185 
1186   u4 is = instance_size(ik, cache_entry);
1187   u4 size = 1 + sizeof(address) + 4 + sizeof(address) + 4 + is;
1188 
1189   writer->start_sub_record(HPROF_GC_INSTANCE_DUMP, size);
1190   writer->write_objectID(o);
1191   writer->write_u4(STACK_TRACE_ID);
1192 
1193   // class ID
1194   writer->write_classID(ik);
1195 
1196   // number of bytes that follow
1197   writer->write_u4(is);
1198 
1199   // field values
1200   dump_instance_fields(writer, o, cache_entry);
1201 
1202   writer->end_sub_record();
1203 }
1204 
1205 // creates HPROF_GC_CLASS_DUMP record for the given instance class
1206 void DumperSupport::dump_instance_class(AbstractDumpWriter* writer, Klass* k) {
1207   InstanceKlass* ik = InstanceKlass::cast(k);
1208 
1209   // We can safepoint and do a heap dump at a point where we have a Klass,
1210   // but no java mirror class has been setup for it. So we need to check
1211   // that the class is at least loaded, to avoid crash from a null mirror.
1212   if (!ik->is_loaded()) {
1213     return;
1214   }
1215 
1216   u2 static_fields_count = 0;
1217   u4 static_size = get_static_fields_size(ik, static_fields_count);
1218   u2 instance_fields_count = get_instance_fields_count(ik);
1219   u4 instance_fields_size = instance_fields_count * (sizeof(address) + 1);
1220   u4 size = checked_cast<u4>(1 + sizeof(address) + 4 + 6 * sizeof(address) + 4 + 2 + 2 + static_size + 2 + instance_fields_size);
1221 
1222   writer->start_sub_record(HPROF_GC_CLASS_DUMP, size);
1223 
1224   // class ID
1225   writer->write_classID(ik);
1226   writer->write_u4(STACK_TRACE_ID);
1227 
1228   // super class ID
1229   InstanceKlass* java_super = ik->java_super();
1230   if (java_super == nullptr) {
1231     writer->write_objectID(oop(nullptr));
1232   } else {
1233     writer->write_classID(java_super);
1234   }
1235 
1236   writer->write_objectID(ik->class_loader());
1237   writer->write_objectID(ik->signers());
1238   writer->write_objectID(ik->protection_domain());
1239 
1240   // reserved
1241   writer->write_objectID(oop(nullptr));
1242   writer->write_objectID(oop(nullptr));
1243 
1244   // instance size
1245   writer->write_u4(DumperSupport::instance_size(ik));
1246 
1247   // size of constant pool - ignored by HAT 1.1
1248   writer->write_u2(0);
1249 
1250   // static fields
1251   writer->write_u2(static_fields_count);
1252   dump_static_fields(writer, ik);
1253 
1254   // description of instance fields
1255   writer->write_u2(instance_fields_count);
1256   dump_instance_field_descriptors(writer, ik);
1257 
1258   writer->end_sub_record();
1259 }
1260 
1261 // creates HPROF_GC_CLASS_DUMP record for the given array class
1262 void DumperSupport::dump_array_class(AbstractDumpWriter* writer, Klass* k) {
1263   InstanceKlass* ik = nullptr; // bottom class for object arrays, null for primitive type arrays
1264   if (k->is_objArray_klass()) {
1265     Klass *bk = ObjArrayKlass::cast(k)->bottom_klass();
1266     assert(bk != nullptr, "checking");
1267     if (bk->is_instance_klass()) {
1268       ik = InstanceKlass::cast(bk);
1269     }
1270   }
1271 
1272   u4 size = 1 + sizeof(address) + 4 + 6 * sizeof(address) + 4 + 2 + 2 + 2;
1273   writer->start_sub_record(HPROF_GC_CLASS_DUMP, size);
1274   writer->write_classID(k);
1275   writer->write_u4(STACK_TRACE_ID);
1276 
1277   // super class of array classes is java.lang.Object
1278   InstanceKlass* java_super = k->java_super();
1279   assert(java_super != nullptr, "checking");
1280   writer->write_classID(java_super);
1281 
1282   writer->write_objectID(ik == nullptr ? oop(nullptr) : ik->class_loader());
1283   writer->write_objectID(ik == nullptr ? oop(nullptr) : ik->signers());
1284   writer->write_objectID(ik == nullptr ? oop(nullptr) : ik->protection_domain());
1285 
1286   writer->write_objectID(oop(nullptr));    // reserved
1287   writer->write_objectID(oop(nullptr));
1288   writer->write_u4(0);             // instance size
1289   writer->write_u2(0);             // constant pool
1290   writer->write_u2(0);             // static fields
1291   writer->write_u2(0);             // instance fields
1292 
1293   writer->end_sub_record();
1294 
1295 }
1296 
1297 // Hprof uses an u4 as record length field,
1298 // which means we need to truncate arrays that are too long.
1299 int DumperSupport::calculate_array_max_length(AbstractDumpWriter* writer, arrayOop array, short header_size) {
1300   BasicType type = ArrayKlass::cast(array->klass())->element_type();
1301   assert(type >= T_BOOLEAN && type <= T_OBJECT, "invalid array element type");
1302 
1303   int length = array->length();
1304 
1305   int type_size;
1306   if (type == T_OBJECT) {
1307     type_size = sizeof(address);
1308   } else {
1309     type_size = type2aelembytes(type);
1310   }
1311 
1312   size_t length_in_bytes = (size_t)length * type_size;
1313   uint max_bytes = max_juint - header_size;
1314 
1315   if (length_in_bytes > max_bytes) {
1316     length = max_bytes / type_size;
1317     length_in_bytes = (size_t)length * type_size;
1318 
1319     warning("cannot dump array of type %s[] with length %d; truncating to length %d",
1320             type2name_tab[type], array->length(), length);
1321   }
1322   return length;
1323 }
1324 
1325 // creates HPROF_GC_OBJ_ARRAY_DUMP record for the given object array
1326 void DumperSupport::dump_object_array(AbstractDumpWriter* writer, objArrayOop array) {
1327   // sizeof(u1) + 2 * sizeof(u4) + sizeof(objectID) + sizeof(classID)
1328   short header_size = 1 + 2 * 4 + 2 * sizeof(address);
1329   int length = calculate_array_max_length(writer, array, header_size);
1330   u4 size = checked_cast<u4>(header_size + length * sizeof(address));
1331 
1332   writer->start_sub_record(HPROF_GC_OBJ_ARRAY_DUMP, size);
1333   writer->write_objectID(array);
1334   writer->write_u4(STACK_TRACE_ID);
1335   writer->write_u4(length);
1336 
1337   // array class ID
1338   writer->write_classID(array->klass());
1339 
1340   // [id]* elements
1341   for (int index = 0; index < length; index++) {
1342     oop o = array->obj_at(index);
1343     o = mask_dormant_archived_object(o, array);
1344     writer->write_objectID(o);
1345   }
1346 
1347   writer->end_sub_record();
1348 }
1349 
1350 #define WRITE_ARRAY(Array, Type, Size, Length) \
1351   for (int i = 0; i < Length; i++) { writer->write_##Size((Size)Array->Type##_at(i)); }
1352 
1353 // creates HPROF_GC_PRIM_ARRAY_DUMP record for the given type array
1354 void DumperSupport::dump_prim_array(AbstractDumpWriter* writer, typeArrayOop array) {
1355   BasicType type = TypeArrayKlass::cast(array->klass())->element_type();
1356   // 2 * sizeof(u1) + 2 * sizeof(u4) + sizeof(objectID)
1357   short header_size = 2 * 1 + 2 * 4 + sizeof(address);
1358 
1359   int length = calculate_array_max_length(writer, array, header_size);
1360   int type_size = type2aelembytes(type);
1361   u4 length_in_bytes = (u4)length * type_size;
1362   u4 size = header_size + length_in_bytes;
1363 
1364   writer->start_sub_record(HPROF_GC_PRIM_ARRAY_DUMP, size);
1365   writer->write_objectID(array);
1366   writer->write_u4(STACK_TRACE_ID);
1367   writer->write_u4(length);
1368   writer->write_u1(type2tag(type));
1369 
1370   // nothing to copy
1371   if (length == 0) {
1372     writer->end_sub_record();
1373     return;
1374   }
1375 
1376   // If the byte ordering is big endian then we can copy most types directly
1377 
1378   switch (type) {
1379     case T_INT : {
1380       if (Endian::is_Java_byte_ordering_different()) {
1381         WRITE_ARRAY(array, int, u4, length);
1382       } else {
1383         writer->write_raw(array->int_at_addr(0), length_in_bytes);
1384       }
1385       break;
1386     }
1387     case T_BYTE : {
1388       writer->write_raw(array->byte_at_addr(0), length_in_bytes);
1389       break;
1390     }
1391     case T_CHAR : {
1392       if (Endian::is_Java_byte_ordering_different()) {
1393         WRITE_ARRAY(array, char, u2, length);
1394       } else {
1395         writer->write_raw(array->char_at_addr(0), length_in_bytes);
1396       }
1397       break;
1398     }
1399     case T_SHORT : {
1400       if (Endian::is_Java_byte_ordering_different()) {
1401         WRITE_ARRAY(array, short, u2, length);
1402       } else {
1403         writer->write_raw(array->short_at_addr(0), length_in_bytes);
1404       }
1405       break;
1406     }
1407     case T_BOOLEAN : {
1408       if (Endian::is_Java_byte_ordering_different()) {
1409         WRITE_ARRAY(array, bool, u1, length);
1410       } else {
1411         writer->write_raw(array->bool_at_addr(0), length_in_bytes);
1412       }
1413       break;
1414     }
1415     case T_LONG : {
1416       if (Endian::is_Java_byte_ordering_different()) {
1417         WRITE_ARRAY(array, long, u8, length);
1418       } else {
1419         writer->write_raw(array->long_at_addr(0), length_in_bytes);
1420       }
1421       break;
1422     }
1423 
1424     // handle float/doubles in a special value to ensure than NaNs are
1425     // written correctly. TO DO: Check if we can avoid this on processors that
1426     // use IEEE 754.
1427 
1428     case T_FLOAT : {
1429       for (int i = 0; i < length; i++) {
1430         dump_float(writer, array->float_at(i));
1431       }
1432       break;
1433     }
1434     case T_DOUBLE : {
1435       for (int i = 0; i < length; i++) {
1436         dump_double(writer, array->double_at(i));
1437       }
1438       break;
1439     }
1440     default : ShouldNotReachHere();
1441   }
1442 
1443   writer->end_sub_record();
1444 }
1445 
1446 // create a HPROF_FRAME record of the given Method* and bci
1447 void DumperSupport::dump_stack_frame(AbstractDumpWriter* writer,
1448                                      int frame_serial_num,
1449                                      int class_serial_num,
1450                                      Method* m,
1451                                      int bci) {
1452   int line_number;
1453   if (m->is_native()) {
1454     line_number = -3;  // native frame
1455   } else {
1456     line_number = m->line_number_from_bci(bci);
1457   }
1458 
1459   write_header(writer, HPROF_FRAME, 4*oopSize + 2*sizeof(u4));
1460   writer->write_id(frame_serial_num);               // frame serial number
1461   writer->write_symbolID(m->name());                // method's name
1462   writer->write_symbolID(m->signature());           // method's signature
1463 
1464   assert(m->method_holder()->is_instance_klass(), "not InstanceKlass");
1465   writer->write_symbolID(m->method_holder()->source_file_name());  // source file name
1466   writer->write_u4(class_serial_num);               // class serial number
1467   writer->write_u4((u4) line_number);               // line number
1468 }
1469 
1470 
1471 // Support class used to generate HPROF_UTF8 records from the entries in the
1472 // SymbolTable.
1473 
1474 class SymbolTableDumper : public SymbolClosure {
1475  private:
1476   AbstractDumpWriter* _writer;
1477   AbstractDumpWriter* writer() const                { return _writer; }
1478  public:
1479   SymbolTableDumper(AbstractDumpWriter* writer)     { _writer = writer; }
1480   void do_symbol(Symbol** p);
1481 };
1482 
1483 void SymbolTableDumper::do_symbol(Symbol** p) {
1484   ResourceMark rm;
1485   Symbol* sym = *p;
1486   int len = sym->utf8_length();
1487   if (len > 0) {
1488     char* s = sym->as_utf8();
1489     DumperSupport::write_header(writer(), HPROF_UTF8, oopSize + len);
1490     writer()->write_symbolID(sym);
1491     writer()->write_raw(s, len);
1492   }
1493 }
1494 
1495 // Support class used to generate HPROF_GC_CLASS_DUMP records
1496 
1497 class ClassDumper : public KlassClosure {
1498  private:
1499   AbstractDumpWriter* _writer;
1500   AbstractDumpWriter* writer() const { return _writer; }
1501 
1502  public:
1503   ClassDumper(AbstractDumpWriter* writer) : _writer(writer) {}
1504 
1505   void do_klass(Klass* k) {
1506     if (k->is_instance_klass()) {
1507       DumperSupport::dump_instance_class(writer(), k);
1508     } else {
1509       DumperSupport::dump_array_class(writer(), k);
1510     }
1511   }
1512 };
1513 
1514 // Support class used to generate HPROF_LOAD_CLASS records
1515 
1516 class LoadedClassDumper : public LockedClassesDo {
1517  private:
1518   AbstractDumpWriter* _writer;
1519   GrowableArray<Klass*>* _klass_map;
1520   u4 _class_serial_num;
1521   AbstractDumpWriter* writer() const { return _writer; }
1522   void add_class_serial_number(Klass* k, int serial_num) {
1523     _klass_map->at_put_grow(serial_num, k);
1524   }
1525  public:
1526   LoadedClassDumper(AbstractDumpWriter* writer, GrowableArray<Klass*>* klass_map)
1527     : _writer(writer), _klass_map(klass_map), _class_serial_num(0) {}
1528 
1529   void do_klass(Klass* k) {
1530     // len of HPROF_LOAD_CLASS record
1531     u4 remaining = 2 * oopSize + 2 * sizeof(u4);
1532     DumperSupport::write_header(writer(), HPROF_LOAD_CLASS, remaining);
1533     // class serial number is just a number
1534     writer()->write_u4(++_class_serial_num);
1535     // class ID
1536     writer()->write_classID(k);
1537     // add the Klass* and class serial number pair
1538     add_class_serial_number(k, _class_serial_num);
1539     writer()->write_u4(STACK_TRACE_ID);
1540     // class name ID
1541     Symbol* name = k->name();
1542     writer()->write_symbolID(name);
1543   }
1544 };
1545 
1546 // Support class used to generate HPROF_GC_ROOT_JNI_LOCAL records
1547 
1548 class JNILocalsDumper : public OopClosure {
1549  private:
1550   AbstractDumpWriter* _writer;
1551   u4 _thread_serial_num;
1552   int _frame_num;
1553   AbstractDumpWriter* writer() const                { return _writer; }
1554  public:
1555   JNILocalsDumper(AbstractDumpWriter* writer, u4 thread_serial_num) {
1556     _writer = writer;
1557     _thread_serial_num = thread_serial_num;
1558     _frame_num = -1;  // default - empty stack
1559   }
1560   void set_frame_number(int n) { _frame_num = n; }
1561   void do_oop(oop* obj_p);
1562   void do_oop(narrowOop* obj_p) { ShouldNotReachHere(); }
1563 };
1564 
1565 void JNILocalsDumper::do_oop(oop* obj_p) {
1566   // ignore null handles
1567   oop o = *obj_p;
1568   if (o != nullptr) {
1569     u4 size = 1 + sizeof(address) + 4 + 4;
1570     writer()->start_sub_record(HPROF_GC_ROOT_JNI_LOCAL, size);
1571     writer()->write_objectID(o);
1572     writer()->write_u4(_thread_serial_num);
1573     writer()->write_u4((u4)_frame_num);
1574     writer()->end_sub_record();
1575   }
1576 }
1577 
1578 
1579 // Support class used to generate HPROF_GC_ROOT_JNI_GLOBAL records
1580 
1581 class JNIGlobalsDumper : public OopClosure {
1582  private:
1583   AbstractDumpWriter* _writer;
1584   AbstractDumpWriter* writer() const                { return _writer; }
1585 
1586  public:
1587   JNIGlobalsDumper(AbstractDumpWriter* writer) {
1588     _writer = writer;
1589   }
1590   void do_oop(oop* obj_p);
1591   void do_oop(narrowOop* obj_p) { ShouldNotReachHere(); }
1592 };
1593 
1594 void JNIGlobalsDumper::do_oop(oop* obj_p) {
1595   oop o = NativeAccess<AS_NO_KEEPALIVE>::oop_load(obj_p);
1596 
1597   // ignore these
1598   if (o == nullptr) return;
1599   // we ignore global ref to symbols and other internal objects
1600   if (o->is_instance() || o->is_objArray() || o->is_typeArray()) {
1601     u4 size = 1 + 2 * sizeof(address);
1602     writer()->start_sub_record(HPROF_GC_ROOT_JNI_GLOBAL, size);
1603     writer()->write_objectID(o);
1604     writer()->write_rootID(obj_p);      // global ref ID
1605     writer()->end_sub_record();
1606   }
1607 };
1608 
1609 // Support class used to generate HPROF_GC_ROOT_STICKY_CLASS records
1610 
1611 class StickyClassDumper : public KlassClosure {
1612  private:
1613   AbstractDumpWriter* _writer;
1614   AbstractDumpWriter* writer() const                { return _writer; }
1615  public:
1616   StickyClassDumper(AbstractDumpWriter* writer) {
1617     _writer = writer;
1618   }
1619   void do_klass(Klass* k) {
1620     if (k->is_instance_klass()) {
1621       InstanceKlass* ik = InstanceKlass::cast(k);
1622       u4 size = 1 + sizeof(address);
1623       writer()->start_sub_record(HPROF_GC_ROOT_STICKY_CLASS, size);
1624       writer()->write_classID(ik);
1625       writer()->end_sub_record();
1626     }
1627   }
1628 };
1629 
1630 // Support class used to generate HPROF_GC_ROOT_JAVA_FRAME records.
1631 
1632 class JavaStackRefDumper : public StackObj {
1633 private:
1634   AbstractDumpWriter* _writer;
1635   u4 _thread_serial_num;
1636   int _frame_num;
1637   AbstractDumpWriter* writer() const { return _writer; }
1638 public:
1639   JavaStackRefDumper(AbstractDumpWriter* writer, u4 thread_serial_num)
1640       : _writer(writer), _thread_serial_num(thread_serial_num), _frame_num(-1) // default - empty stack
1641   {
1642   }
1643 
1644   void set_frame_number(int n) { _frame_num = n; }
1645 
1646   void dump_java_stack_refs(StackValueCollection* values);
1647 };
1648 
1649 void JavaStackRefDumper::dump_java_stack_refs(StackValueCollection* values) {
1650   for (int index = 0; index < values->size(); index++) {
1651     if (values->at(index)->type() == T_OBJECT) {
1652       oop o = values->obj_at(index)();
1653       if (o != nullptr) {
1654         u4 size = 1 + sizeof(address) + 4 + 4;
1655         writer()->start_sub_record(HPROF_GC_ROOT_JAVA_FRAME, size);
1656         writer()->write_objectID(o);
1657         writer()->write_u4(_thread_serial_num);
1658         writer()->write_u4((u4)_frame_num);
1659         writer()->end_sub_record();
1660       }
1661     }
1662   }
1663 }
1664 
1665 // Class to collect, store and dump thread-related data:
1666 // - HPROF_TRACE and HPROF_FRAME records;
1667 // - HPROF_GC_ROOT_THREAD_OBJ/HPROF_GC_ROOT_JAVA_FRAME/HPROF_GC_ROOT_JNI_LOCAL subrecords.
1668 class ThreadDumper : public CHeapObj<mtInternal> {
1669 public:
1670   enum class ThreadType { Platform, MountedVirtual, UnmountedVirtual };
1671 
1672 private:
1673   ThreadType _thread_type;
1674   JavaThread* _java_thread;
1675   oop _thread_oop;
1676 
1677   GrowableArray<StackFrameInfo*>* _frames;
1678   // non-null if the thread is OOM thread
1679   Method* _oome_constructor;
1680   int _thread_serial_num;
1681   int _start_frame_serial_num;
1682 
1683   vframe* get_top_frame() const;
1684 
1685 public:
1686   static bool should_dump_pthread(JavaThread* thread) {
1687     return thread->threadObj() != nullptr && !thread->is_exiting() && !thread->is_hidden_from_external_view();
1688   }
1689 
1690   static bool should_dump_vthread(oop vt) {
1691     return java_lang_VirtualThread::state(vt) != java_lang_VirtualThread::NEW
1692         && java_lang_VirtualThread::state(vt) != java_lang_VirtualThread::TERMINATED;
1693   }
1694 
1695   static bool is_vthread_mounted(oop vt) {
1696     // The code should be consistent with the "mounted virtual thread" case
1697     // (VM_HeapDumper::dump_stack_traces(), ThreadDumper::get_top_frame()).
1698     // I.e. virtual thread is mounted if its carrierThread is not null
1699     // and is_vthread_mounted() for the carrier thread returns true.
1700     oop carrier_thread = java_lang_VirtualThread::carrier_thread(vt);
1701     if (carrier_thread == nullptr) {
1702       return false;
1703     }
1704     JavaThread* java_thread = java_lang_Thread::thread(carrier_thread);
1705     return java_thread->is_vthread_mounted();
1706   }
1707 
1708   ThreadDumper(ThreadType thread_type, JavaThread* java_thread, oop thread_oop);
1709   ~ThreadDumper() {
1710     for (int index = 0; index < _frames->length(); index++) {
1711       delete _frames->at(index);
1712     }
1713     delete _frames;
1714   }
1715 
1716   // affects frame_count
1717   void add_oom_frame(Method* oome_constructor) {
1718     assert(_start_frame_serial_num == 0, "add_oom_frame cannot be called after init_serial_nums");
1719     _oome_constructor = oome_constructor;
1720   }
1721 
1722   void init_serial_nums(volatile int* thread_counter, volatile int* frame_counter) {
1723     assert(_start_frame_serial_num == 0, "already initialized");
1724     _thread_serial_num = Atomic::fetch_then_add(thread_counter, 1);
1725     _start_frame_serial_num = Atomic::fetch_then_add(frame_counter, frame_count());
1726   }
1727 
1728   bool oom_thread() const {
1729     return _oome_constructor != nullptr;
1730   }
1731 
1732   int frame_count() const {
1733     return _frames->length() + (oom_thread() ? 1 : 0);
1734   }
1735 
1736   u4 thread_serial_num() const {
1737     return (u4)_thread_serial_num;
1738   }
1739 
1740   u4 stack_trace_serial_num() const {
1741     return (u4)(_thread_serial_num + STACK_TRACE_ID);
1742   }
1743 
1744   // writes HPROF_TRACE and HPROF_FRAME records
1745   // returns number of dumped frames
1746   void dump_stack_traces(AbstractDumpWriter* writer, GrowableArray<Klass*>* klass_map);
1747 
1748   // writes HPROF_GC_ROOT_THREAD_OBJ subrecord
1749   void dump_thread_obj(AbstractDumpWriter* writer);
1750 
1751   // Walk the stack of the thread.
1752   // Dumps a HPROF_GC_ROOT_JAVA_FRAME subrecord for each local
1753   // Dumps a HPROF_GC_ROOT_JNI_LOCAL subrecord for each JNI local
1754   void dump_stack_refs(AbstractDumpWriter* writer);
1755 
1756 };
1757 
1758 ThreadDumper::ThreadDumper(ThreadType thread_type, JavaThread* java_thread, oop thread_oop)
1759     : _thread_type(thread_type), _java_thread(java_thread), _thread_oop(thread_oop),
1760       _oome_constructor(nullptr),
1761       _thread_serial_num(0), _start_frame_serial_num(0)
1762 {
1763   // sanity checks
1764   if (_thread_type == ThreadType::UnmountedVirtual) {
1765     assert(_java_thread == nullptr, "sanity");
1766     assert(_thread_oop != nullptr, "sanity");
1767   } else {
1768     assert(_java_thread != nullptr, "sanity");
1769     assert(_thread_oop != nullptr, "sanity");
1770   }
1771 
1772   _frames = new (mtServiceability) GrowableArray<StackFrameInfo*>(10, mtServiceability);
1773   bool stop_at_vthread_entry = _thread_type == ThreadType::MountedVirtual;
1774 
1775   // vframes are resource allocated
1776   Thread* current_thread = Thread::current();
1777   ResourceMark rm(current_thread);
1778   HandleMark hm(current_thread);
1779 
1780   for (vframe* vf = get_top_frame(); vf != nullptr; vf = vf->sender()) {
1781     if (stop_at_vthread_entry && vf->is_vthread_entry()) {
1782       break;
1783     }
1784     if (vf->is_java_frame()) {
1785       javaVFrame* jvf = javaVFrame::cast(vf);
1786       _frames->append(new StackFrameInfo(jvf, false));
1787     } else {
1788       // ignore non-Java frames
1789     }
1790   }
1791 }
1792 
1793 void ThreadDumper::dump_stack_traces(AbstractDumpWriter* writer, GrowableArray<Klass*>* klass_map) {
1794   assert(_thread_serial_num != 0 && _start_frame_serial_num != 0, "serial_nums are not initialized");
1795 
1796   // write HPROF_FRAME records for this thread's stack trace
1797   int depth = _frames->length();
1798   int frame_serial_num = _start_frame_serial_num;
1799 
1800   if (oom_thread()) {
1801     // OOM thread
1802     // write fake frame that makes it look like the thread, which caused OOME,
1803     // is in the OutOfMemoryError zero-parameter constructor
1804     int oome_serial_num = klass_map->find(_oome_constructor->method_holder());
1805     // the class serial number starts from 1
1806     assert(oome_serial_num > 0, "OutOfMemoryError class not found");
1807     DumperSupport::dump_stack_frame(writer, ++frame_serial_num, oome_serial_num, _oome_constructor, 0);
1808     depth++;
1809   }
1810 
1811   for (int j = 0; j < _frames->length(); j++) {
1812     StackFrameInfo* frame = _frames->at(j);
1813     Method* m = frame->method();
1814     int class_serial_num = klass_map->find(m->method_holder());
1815     // the class serial number starts from 1
1816     assert(class_serial_num > 0, "class not found");
1817     DumperSupport::dump_stack_frame(writer, ++frame_serial_num, class_serial_num, m, frame->bci());
1818   }
1819 
1820   // write HPROF_TRACE record for the thread
1821   DumperSupport::write_header(writer, HPROF_TRACE, checked_cast<u4>(3 * sizeof(u4) + depth * oopSize));
1822   writer->write_u4(stack_trace_serial_num());   // stack trace serial number
1823   writer->write_u4(thread_serial_num());        // thread serial number
1824   writer->write_u4((u4)depth);                  // frame count (including oom frame)
1825   for (int j = 1; j <= depth; j++) {
1826     writer->write_id(_start_frame_serial_num + j);
1827   }
1828 }
1829 
1830 void ThreadDumper::dump_thread_obj(AbstractDumpWriter * writer) {
1831   assert(_thread_serial_num != 0 && _start_frame_serial_num != 0, "serial_num is not initialized");
1832 
1833   u4 size = 1 + sizeof(address) + 4 + 4;
1834   writer->start_sub_record(HPROF_GC_ROOT_THREAD_OBJ, size);
1835   writer->write_objectID(_thread_oop);
1836   writer->write_u4(thread_serial_num());      // thread serial number
1837   writer->write_u4(stack_trace_serial_num()); // stack trace serial number
1838   writer->end_sub_record();
1839 }
1840 
1841 void ThreadDumper::dump_stack_refs(AbstractDumpWriter * writer) {
1842   assert(_thread_serial_num != 0 && _start_frame_serial_num != 0, "serial_num is not initialized");
1843 
1844   JNILocalsDumper blk(writer, thread_serial_num());
1845   if (_thread_type == ThreadType::Platform) {
1846     if (!_java_thread->has_last_Java_frame()) {
1847       // no last java frame but there may be JNI locals
1848       _java_thread->active_handles()->oops_do(&blk);
1849       return;
1850     }
1851   }
1852 
1853   JavaStackRefDumper java_ref_dumper(writer, thread_serial_num());
1854 
1855   // vframes are resource allocated
1856   Thread* current_thread = Thread::current();
1857   ResourceMark rm(current_thread);
1858   HandleMark hm(current_thread);
1859 
1860   bool stopAtVthreadEntry = _thread_type == ThreadType::MountedVirtual;
1861   frame* last_entry_frame = nullptr;
1862   bool is_top_frame = true;
1863   int depth = 0;
1864   if (oom_thread()) {
1865     depth++;
1866   }
1867 
1868   for (vframe* vf = get_top_frame(); vf != nullptr; vf = vf->sender()) {
1869     if (stopAtVthreadEntry && vf->is_vthread_entry()) {
1870       break;
1871     }
1872 
1873     if (vf->is_java_frame()) {
1874       javaVFrame* jvf = javaVFrame::cast(vf);
1875       if (!(jvf->method()->is_native())) {
1876         java_ref_dumper.set_frame_number(depth);
1877         java_ref_dumper.dump_java_stack_refs(jvf->locals());
1878         java_ref_dumper.dump_java_stack_refs(jvf->expressions());
1879       } else {
1880         // native frame
1881         blk.set_frame_number(depth);
1882         if (is_top_frame) {
1883           // JNI locals for the top frame if mounted
1884           assert(_java_thread != nullptr || jvf->method()->is_synchronized()
1885                  || jvf->method()->is_object_wait0(), "impossible for unmounted vthread");
1886           if (_java_thread != nullptr) {
1887             _java_thread->active_handles()->oops_do(&blk);
1888           }
1889         } else {
1890           if (last_entry_frame != nullptr) {
1891             // JNI locals for the entry frame
1892             assert(last_entry_frame->is_entry_frame(), "checking");
1893             last_entry_frame->entry_frame_call_wrapper()->handles()->oops_do(&blk);
1894           }
1895         }
1896       }
1897       last_entry_frame = nullptr;
1898       // increment only for Java frames
1899       depth++;
1900     } else {
1901       // externalVFrame - for an entry frame then we report the JNI locals
1902       // when we find the corresponding javaVFrame
1903       frame* fr = vf->frame_pointer();
1904       assert(fr != nullptr, "sanity check");
1905       if (fr->is_entry_frame()) {
1906         last_entry_frame = fr;
1907       }
1908     }
1909   is_top_frame = false;
1910   }
1911   assert(depth == frame_count(), "total number of Java frames not matched");
1912 }
1913 
1914 vframe* ThreadDumper::get_top_frame() const {
1915   if (_thread_type == ThreadType::UnmountedVirtual) {
1916     ContinuationWrapper cont(java_lang_VirtualThread::continuation(_thread_oop));
1917     if (cont.is_empty()) {
1918       return nullptr;
1919     }
1920     assert(!cont.is_mounted(), "sanity check");
1921     stackChunkOop chunk = cont.last_nonempty_chunk();
1922     if (chunk == nullptr || chunk->is_empty()) {
1923       return nullptr;
1924     }
1925 
1926     RegisterMap reg_map(cont.continuation(), RegisterMap::UpdateMap::include);
1927     frame fr = chunk->top_frame(&reg_map);
1928     vframe* vf = vframe::new_vframe(&fr, &reg_map, nullptr); // don't need JavaThread
1929     return vf;
1930   }
1931 
1932   RegisterMap reg_map(_java_thread,
1933       RegisterMap::UpdateMap::include,
1934       RegisterMap::ProcessFrames::include,
1935       RegisterMap::WalkContinuation::skip);
1936   switch (_thread_type) {
1937   case ThreadType::Platform:
1938     if (!_java_thread->has_last_Java_frame()) {
1939       return nullptr;
1940     }
1941     return _java_thread->is_vthread_mounted()
1942         ? _java_thread->carrier_last_java_vframe(&reg_map)
1943         : _java_thread->platform_thread_last_java_vframe(&reg_map);
1944 
1945   case ThreadType::MountedVirtual:
1946     return _java_thread->last_java_vframe(&reg_map);
1947 
1948   default: // make compilers happy
1949       break;
1950   }
1951   ShouldNotReachHere();
1952   return nullptr;
1953 }
1954 
1955 // Callback to dump thread-related data for unmounted virtual threads;
1956 // implemented by VM_HeapDumper.
1957 class UnmountedVThreadDumper {
1958  public:
1959   virtual void dump_vthread(oop vt, AbstractDumpWriter* segment_writer) = 0;
1960 };
1961 
1962 // Support class used when iterating over the heap.
1963 class HeapObjectDumper : public ObjectClosure {
1964  private:
1965   AbstractDumpWriter* _writer;
1966   AbstractDumpWriter* writer()                  { return _writer; }
1967   UnmountedVThreadDumper* _vthread_dumper;
1968 
1969   DumperClassCacheTable _class_cache;
1970 
1971  public:
1972   HeapObjectDumper(AbstractDumpWriter* writer, UnmountedVThreadDumper* vthread_dumper)
1973     : _writer(writer), _vthread_dumper(vthread_dumper) {}
1974 
1975   // called for each object in the heap
1976   void do_object(oop o);
1977 };
1978 
1979 void HeapObjectDumper::do_object(oop o) {
1980   // skip classes as these emitted as HPROF_GC_CLASS_DUMP records
1981   if (o->klass() == vmClasses::Class_klass()) {
1982     if (!java_lang_Class::is_primitive(o)) {
1983       return;
1984     }
1985   }
1986 
1987   if (DumperSupport::mask_dormant_archived_object(o, nullptr) == nullptr) {
1988     return;
1989   }
1990 
1991   if (o->is_instance()) {
1992     // create a HPROF_GC_INSTANCE record for each object
1993     DumperSupport::dump_instance(writer(), o, &_class_cache);
1994     // If we encounter an unmounted virtual thread it needs to be dumped explicitly
1995     // (mounted virtual threads are dumped with their carriers).
1996     if (java_lang_VirtualThread::is_instance(o)
1997         && ThreadDumper::should_dump_vthread(o) && !ThreadDumper::is_vthread_mounted(o)) {
1998       _vthread_dumper->dump_vthread(o, writer());
1999     }
2000   } else if (o->is_objArray()) {
2001     // create a HPROF_GC_OBJ_ARRAY_DUMP record for each object array
2002     DumperSupport::dump_object_array(writer(), objArrayOop(o));
2003   } else if (o->is_typeArray()) {
2004     // create a HPROF_GC_PRIM_ARRAY_DUMP record for each type array
2005     DumperSupport::dump_prim_array(writer(), typeArrayOop(o));
2006   }
2007 }
2008 
2009 // The dumper controller for parallel heap dump
2010 class DumperController : public CHeapObj<mtInternal> {
2011  private:
2012    Monitor* _lock;
2013    Mutex* _global_writer_lock;
2014 
2015    const uint   _dumper_number;
2016    uint   _complete_number;
2017 
2018    bool   _started; // VM dumper started and acquired global writer lock
2019 
2020  public:
2021    DumperController(uint number) :
2022      // _lock and _global_writer_lock are used for synchronization between GC worker threads inside safepoint,
2023      // so we lock with _no_safepoint_check_flag.
2024      // signal_start() acquires _lock when global writer is locked,
2025      // its rank must be less than _global_writer_lock rank.
2026      _lock(new (std::nothrow) PaddedMonitor(Mutex::nosafepoint - 1, "DumperController_lock")),
2027      _global_writer_lock(new (std::nothrow) Mutex(Mutex::nosafepoint, "DumpWriter_lock")),
2028      _dumper_number(number),
2029      _complete_number(0),
2030      _started(false)
2031    {}
2032 
2033    ~DumperController() {
2034      delete _lock;
2035      delete _global_writer_lock;
2036    }
2037 
2038    // parallel (non VM) dumpers must wait until VM dumper acquires global writer lock
2039    void wait_for_start_signal() {
2040      MonitorLocker ml(_lock, Mutex::_no_safepoint_check_flag);
2041      while (_started == false) {
2042        ml.wait();
2043      }
2044    }
2045 
2046    void signal_start() {
2047      MonitorLocker ml(_lock, Mutex::_no_safepoint_check_flag);
2048      _started = true;
2049      ml.notify_all();
2050    }
2051 
2052    void lock_global_writer() {
2053      _global_writer_lock->lock_without_safepoint_check();
2054    }
2055 
2056    void unlock_global_writer() {
2057      _global_writer_lock->unlock();
2058    }
2059 
2060    void dumper_complete(DumpWriter* local_writer, DumpWriter* global_writer) {
2061      MonitorLocker ml(_lock, Mutex::_no_safepoint_check_flag);
2062      _complete_number++;
2063      // propagate local error to global if any
2064      if (local_writer->has_error()) {
2065        global_writer->set_error(local_writer->error());
2066      }
2067      ml.notify();
2068    }
2069 
2070    void wait_all_dumpers_complete() {
2071      MonitorLocker ml(_lock, Mutex::_no_safepoint_check_flag);
2072      while (_complete_number != _dumper_number) {
2073         ml.wait();
2074      }
2075    }
2076 };
2077 
2078 // DumpMerger merges separate dump files into a complete one
2079 class DumpMerger : public StackObj {
2080 private:
2081   DumpWriter* _writer;
2082   const char* _path;
2083   bool _has_error;
2084   int _dump_seq;
2085 
2086 private:
2087   void merge_file(const char* path);
2088   void merge_done();
2089   void set_error(const char* msg);
2090 
2091 public:
2092   DumpMerger(const char* path, DumpWriter* writer, int dump_seq) :
2093     _writer(writer),
2094     _path(path),
2095     _has_error(_writer->has_error()),
2096     _dump_seq(dump_seq) {}
2097 
2098   void do_merge();
2099 
2100   // returns path for the parallel DumpWriter (resource allocated)
2101   static char* get_writer_path(const char* base_path, int seq);
2102 
2103 };
2104 
2105 char* DumpMerger::get_writer_path(const char* base_path, int seq) {
2106   // approximate required buffer size
2107   size_t buf_size = strlen(base_path)
2108                     + 2                 // ".p"
2109                     + 10                // number (that's enough for 2^32 parallel dumpers)
2110                     + 1;                // '\0'
2111 
2112   char* path = NEW_RESOURCE_ARRAY(char, buf_size);
2113   memset(path, 0, buf_size);
2114 
2115   os::snprintf(path, buf_size, "%s.p%d", base_path, seq);
2116 
2117   return path;
2118 }
2119 
2120 
2121 void DumpMerger::merge_done() {
2122   // Writes the HPROF_HEAP_DUMP_END record.
2123   if (!_has_error) {
2124     DumperSupport::end_of_dump(_writer);
2125     _writer->flush();
2126   }
2127   _dump_seq = 0; //reset
2128 }
2129 
2130 void DumpMerger::set_error(const char* msg) {
2131   assert(msg != nullptr, "sanity check");
2132   log_error(heapdump)("%s (file: %s)", msg, _path);
2133   _writer->set_error(msg);
2134   _has_error = true;
2135 }
2136 
2137 #ifdef LINUX
2138 // Merge segmented heap files via sendfile, it's more efficient than the
2139 // read+write combination, which would require transferring data to and from
2140 // user space.
2141 void DumpMerger::merge_file(const char* path) {
2142   TraceTime timer("Merge segmented heap file directly", TRACETIME_LOG(Info, heapdump));
2143 
2144   int segment_fd = os::open(path, O_RDONLY, 0);
2145   if (segment_fd == -1) {
2146     set_error("Can not open segmented heap file during merging");
2147     return;
2148   }
2149 
2150   struct stat st;
2151   if (os::stat(path, &st) != 0) {
2152     ::close(segment_fd);
2153     set_error("Can not get segmented heap file size during merging");
2154     return;
2155   }
2156 
2157   // A successful call to sendfile may write fewer bytes than requested; the
2158   // caller should be prepared to retry the call if there were unsent bytes.
2159   jlong offset = 0;
2160   while (offset < st.st_size) {
2161     int ret = os::Linux::sendfile(_writer->get_fd(), segment_fd, &offset, st.st_size);
2162     if (ret == -1) {
2163       ::close(segment_fd);
2164       set_error("Failed to merge segmented heap file");
2165       return;
2166     }
2167   }
2168 
2169   // As sendfile variant does not call the write method of the global writer,
2170   // bytes_written is also incorrect for this variant, we need to explicitly
2171   // accumulate bytes_written for the global writer in this case
2172   julong accum = _writer->bytes_written() + st.st_size;
2173   _writer->set_bytes_written(accum);
2174   ::close(segment_fd);
2175 }
2176 #else
2177 // Generic implementation using read+write
2178 void DumpMerger::merge_file(const char* path) {
2179   TraceTime timer("Merge segmented heap file", TRACETIME_LOG(Info, heapdump));
2180 
2181   fileStream segment_fs(path, "rb");
2182   if (!segment_fs.is_open()) {
2183     set_error("Can not open segmented heap file during merging");
2184     return;
2185   }
2186 
2187   jlong total = 0;
2188   size_t cnt = 0;
2189 
2190   // Use _writer buffer for reading.
2191   while ((cnt = segment_fs.read(_writer->buffer(), 1, _writer->buffer_size())) != 0) {
2192     _writer->set_position(cnt);
2193     _writer->flush();
2194     total += cnt;
2195   }
2196 
2197   if (segment_fs.fileSize() != total) {
2198     set_error("Merged heap dump is incomplete");
2199   }
2200 }
2201 #endif
2202 
2203 void DumpMerger::do_merge() {
2204   TraceTime timer("Merge heap files complete", TRACETIME_LOG(Info, heapdump));
2205 
2206   // Since contents in segmented heap file were already zipped, we don't need to zip
2207   // them again during merging.
2208   AbstractCompressor* saved_compressor = _writer->compressor();
2209   _writer->set_compressor(nullptr);
2210 
2211   // Merge the content of the remaining files into base file. Regardless of whether
2212   // the merge process is successful or not, these segmented files will be deleted.
2213   for (int i = 0; i < _dump_seq; i++) {
2214     ResourceMark rm;
2215     const char* path = get_writer_path(_path, i);
2216     if (!_has_error) {
2217       merge_file(path);
2218     }
2219     // Delete selected segmented heap file nevertheless
2220     if (remove(path) != 0) {
2221       log_info(heapdump)("Removal of segment file (%d) failed (%d)", i, errno);
2222     }
2223   }
2224 
2225   // restore compressor for further use
2226   _writer->set_compressor(saved_compressor);
2227   merge_done();
2228 }
2229 
2230 // The VM operation that performs the heap dump
2231 class VM_HeapDumper : public VM_GC_Operation, public WorkerTask, public UnmountedVThreadDumper {
2232  private:
2233   DumpWriter*             _writer;
2234   JavaThread*             _oome_thread;
2235   Method*                 _oome_constructor;
2236   bool                    _gc_before_heap_dump;
2237   GrowableArray<Klass*>*  _klass_map;
2238 
2239   ThreadDumper**          _thread_dumpers; // platform, carrier and mounted virtual threads
2240   int                     _thread_dumpers_count;
2241   volatile int            _thread_serial_num;
2242   volatile int            _frame_serial_num;
2243 
2244   volatile int            _dump_seq;
2245   // parallel heap dump support
2246   uint                    _num_dumper_threads;
2247   DumperController*       _dumper_controller;
2248   ParallelObjectIterator* _poi;
2249 
2250   // Dumper id of VMDumper thread.
2251   static const int VMDumperId = 0;
2252   // VM dumper dumps both heap and non-heap data, other dumpers dump heap-only data.
2253   static bool is_vm_dumper(int dumper_id) { return dumper_id == VMDumperId; }
2254   // the 1st dumper calling get_next_dumper_id becomes VM dumper
2255   int get_next_dumper_id() {
2256     return Atomic::fetch_then_add(&_dump_seq, 1);
2257   }
2258 
2259   DumpWriter* writer() const { return _writer; }
2260 
2261   bool skip_operation() const;
2262 
2263   // HPROF_GC_ROOT_THREAD_OBJ records for platform and mounted virtual threads
2264   void dump_threads(AbstractDumpWriter* writer);
2265 
2266   bool is_oom_thread(JavaThread* thread) const {
2267     return thread == _oome_thread && _oome_constructor != nullptr;
2268   }
2269 
2270   // HPROF_TRACE and HPROF_FRAME records for platform and mounted virtual threads
2271   void dump_stack_traces(AbstractDumpWriter* writer);
2272 
2273  public:
2274   VM_HeapDumper(DumpWriter* writer, bool gc_before_heap_dump, bool oome, uint num_dump_threads) :
2275     VM_GC_Operation(0 /* total collections,      dummy, ignored */,
2276                     GCCause::_heap_dump /* GC Cause */,
2277                     0 /* total full collections, dummy, ignored */,
2278                     gc_before_heap_dump),
2279     WorkerTask("dump heap") {
2280     _writer = writer;
2281     _gc_before_heap_dump = gc_before_heap_dump;
2282     _klass_map = new (mtServiceability) GrowableArray<Klass*>(INITIAL_CLASS_COUNT, mtServiceability);
2283 
2284     _thread_dumpers = nullptr;
2285     _thread_dumpers_count = 0;
2286     _thread_serial_num = 1;
2287     _frame_serial_num = 1;
2288 
2289     _dump_seq = VMDumperId;
2290     _num_dumper_threads = num_dump_threads;
2291     _dumper_controller = nullptr;
2292     _poi = nullptr;
2293     if (oome) {
2294       assert(!Thread::current()->is_VM_thread(), "Dump from OutOfMemoryError cannot be called by the VMThread");
2295       // get OutOfMemoryError zero-parameter constructor
2296       InstanceKlass* oome_ik = vmClasses::OutOfMemoryError_klass();
2297       _oome_constructor = oome_ik->find_method(vmSymbols::object_initializer_name(),
2298                                                           vmSymbols::void_method_signature());
2299       // get thread throwing OOME when generating the heap dump at OOME
2300       _oome_thread = JavaThread::current();
2301     } else {
2302       _oome_thread = nullptr;
2303       _oome_constructor = nullptr;
2304     }
2305   }
2306 
2307   ~VM_HeapDumper() {
2308     if (_thread_dumpers != nullptr) {
2309       for (int i = 0; i < _thread_dumpers_count; i++) {
2310         delete _thread_dumpers[i];
2311       }
2312       FREE_C_HEAP_ARRAY(ThreadDumper*, _thread_dumpers);
2313     }
2314 
2315     if (_dumper_controller != nullptr) {
2316       delete _dumper_controller;
2317       _dumper_controller = nullptr;
2318     }
2319     delete _klass_map;
2320   }
2321   int dump_seq()           { return _dump_seq; }
2322   bool is_parallel_dump()  { return _num_dumper_threads > 1; }
2323   void prepare_parallel_dump(WorkerThreads* workers);
2324 
2325   VMOp_Type type() const { return VMOp_HeapDumper; }
2326   virtual bool doit_prologue();
2327   void doit();
2328   void work(uint worker_id);
2329 
2330   // UnmountedVThreadDumper implementation
2331   void dump_vthread(oop vt, AbstractDumpWriter* segment_writer);
2332 };
2333 
2334 bool VM_HeapDumper::skip_operation() const {
2335   return false;
2336 }
2337 
2338 // fixes up the current dump record and writes HPROF_HEAP_DUMP_END record
2339 void DumperSupport::end_of_dump(AbstractDumpWriter* writer) {
2340   writer->finish_dump_segment();
2341 
2342   writer->write_u1(HPROF_HEAP_DUMP_END);
2343   writer->write_u4(0);
2344   writer->write_u4(0);
2345 }
2346 
2347 // Write a HPROF_GC_ROOT_THREAD_OBJ record for platform/carrier and mounted virtual threads.
2348 // Then walk the stack so that locals and JNI locals are dumped.
2349 void VM_HeapDumper::dump_threads(AbstractDumpWriter* writer) {
2350   for (int i = 0; i < _thread_dumpers_count; i++) {
2351     _thread_dumpers[i]->dump_thread_obj(writer);
2352     _thread_dumpers[i]->dump_stack_refs(writer);
2353   }
2354 }
2355 
2356 bool VM_HeapDumper::doit_prologue() {
2357   if (_gc_before_heap_dump && (UseZGC || UseShenandoahGC)) {
2358     // ZGC and Shenandoah cannot perform a synchronous GC cycle from within the VM thread.
2359     // So collect_as_vm_thread() is a noop. To respect the _gc_before_heap_dump flag a
2360     // synchronous GC cycle is performed from the caller thread in the prologue.
2361     Universe::heap()->collect(GCCause::_heap_dump);
2362   }
2363   return VM_GC_Operation::doit_prologue();
2364 }
2365 
2366 void VM_HeapDumper::prepare_parallel_dump(WorkerThreads* workers) {
2367   uint num_active_workers = workers != nullptr ? workers->active_workers() : 0;
2368   uint num_requested_dump_threads = _num_dumper_threads;
2369   // check if we can dump in parallel based on requested and active threads
2370   if (num_active_workers <= 1 || num_requested_dump_threads <= 1) {
2371     _num_dumper_threads = 1;
2372   } else {
2373     _num_dumper_threads = clamp(num_requested_dump_threads, 2U, num_active_workers);
2374   }
2375   _dumper_controller = new (std::nothrow) DumperController(_num_dumper_threads);
2376   bool can_parallel = _num_dumper_threads > 1;
2377   log_info(heapdump)("Requested dump threads %u, active dump threads %u, "
2378                      "actual dump threads %u, parallelism %s",
2379                      num_requested_dump_threads, num_active_workers,
2380                      _num_dumper_threads, can_parallel ? "true" : "false");
2381 }
2382 
2383 // The VM operation that dumps the heap. The dump consists of the following
2384 // records:
2385 //
2386 //  HPROF_HEADER
2387 //  [HPROF_UTF8]*
2388 //  [HPROF_LOAD_CLASS]*
2389 //  [[HPROF_FRAME]*|HPROF_TRACE]*
2390 //  [HPROF_GC_CLASS_DUMP]*
2391 //  [HPROF_HEAP_DUMP_SEGMENT]*
2392 //  HPROF_HEAP_DUMP_END
2393 //
2394 // The HPROF_TRACE records represent the stack traces where the heap dump
2395 // is generated and a "dummy trace" record which does not include
2396 // any frames. The dummy trace record is used to be referenced as the
2397 // unknown object alloc site.
2398 //
2399 // Each HPROF_HEAP_DUMP_SEGMENT record has a length followed by sub-records.
2400 // To allow the heap dump be generated in a single pass we remember the position
2401 // of the dump length and fix it up after all sub-records have been written.
2402 // To generate the sub-records we iterate over the heap, writing
2403 // HPROF_GC_INSTANCE_DUMP, HPROF_GC_OBJ_ARRAY_DUMP, and HPROF_GC_PRIM_ARRAY_DUMP
2404 // records as we go. Once that is done we write records for some of the GC
2405 // roots.
2406 
2407 void VM_HeapDumper::doit() {
2408 
2409   CollectedHeap* ch = Universe::heap();
2410 
2411   ch->ensure_parsability(false); // must happen, even if collection does
2412                                  // not happen (e.g. due to GCLocker)
2413 
2414   if (_gc_before_heap_dump) {
2415     if (GCLocker::is_active()) {
2416       warning("GC locker is held; pre-heapdump GC was skipped");
2417     } else {
2418       ch->collect_as_vm_thread(GCCause::_heap_dump);
2419     }
2420   }
2421 
2422   WorkerThreads* workers = ch->safepoint_workers();
2423   prepare_parallel_dump(workers);
2424 
2425   if (!is_parallel_dump()) {
2426     work(VMDumperId);
2427   } else {
2428     ParallelObjectIterator poi(_num_dumper_threads);
2429     _poi = &poi;
2430     workers->run_task(this, _num_dumper_threads);
2431     _poi = nullptr;
2432   }
2433 }
2434 
2435 void VM_HeapDumper::work(uint worker_id) {
2436   // VM Dumper works on all non-heap data dumping and part of heap iteration.
2437   int dumper_id = get_next_dumper_id();
2438 
2439   if (is_vm_dumper(dumper_id)) {
2440     // lock global writer, it will be unlocked after VM Dumper finishes with non-heap data
2441     _dumper_controller->lock_global_writer();
2442     _dumper_controller->signal_start();
2443   } else {
2444     _dumper_controller->wait_for_start_signal();
2445   }
2446 
2447   if (is_vm_dumper(dumper_id)) {
2448     TraceTime timer("Dump non-objects", TRACETIME_LOG(Info, heapdump));
2449     // Write the file header - we always use 1.0.2
2450     const char* header = "JAVA PROFILE 1.0.2";
2451 
2452     // header is few bytes long - no chance to overflow int
2453     writer()->write_raw(header, strlen(header) + 1); // NUL terminated
2454     writer()->write_u4(oopSize);
2455     // timestamp is current time in ms
2456     writer()->write_u8(os::javaTimeMillis());
2457     // HPROF_UTF8 records
2458     SymbolTableDumper sym_dumper(writer());
2459     SymbolTable::symbols_do(&sym_dumper);
2460 
2461     // write HPROF_LOAD_CLASS records
2462     {
2463       LoadedClassDumper loaded_class_dumper(writer(), _klass_map);
2464       ClassLoaderDataGraph::classes_do(&loaded_class_dumper);
2465     }
2466 
2467     // write HPROF_FRAME and HPROF_TRACE records
2468     // this must be called after _klass_map is built when iterating the classes above.
2469     dump_stack_traces(writer());
2470 
2471     // unlock global writer, so parallel dumpers can dump stack traces of unmounted virtual threads
2472     _dumper_controller->unlock_global_writer();
2473   }
2474 
2475   // HPROF_HEAP_DUMP/HPROF_HEAP_DUMP_SEGMENT starts here
2476 
2477   ResourceMark rm;
2478   // share global compressor, local DumpWriter is not responsible for its life cycle
2479   DumpWriter segment_writer(DumpMerger::get_writer_path(writer()->get_file_path(), dumper_id),
2480                             writer()->is_overwrite(), writer()->compressor());
2481   if (!segment_writer.has_error()) {
2482     if (is_vm_dumper(dumper_id)) {
2483       // dump some non-heap subrecords to heap dump segment
2484       TraceTime timer("Dump non-objects (part 2)", TRACETIME_LOG(Info, heapdump));
2485       // Writes HPROF_GC_CLASS_DUMP records
2486       ClassDumper class_dumper(&segment_writer);
2487       ClassLoaderDataGraph::classes_do(&class_dumper);
2488 
2489       // HPROF_GC_ROOT_THREAD_OBJ + frames + jni locals
2490       dump_threads(&segment_writer);
2491 
2492       // HPROF_GC_ROOT_JNI_GLOBAL
2493       JNIGlobalsDumper jni_dumper(&segment_writer);
2494       JNIHandles::oops_do(&jni_dumper);
2495       // technically not jni roots, but global roots
2496       // for things like preallocated throwable backtraces
2497       Universe::vm_global()->oops_do(&jni_dumper);
2498       // HPROF_GC_ROOT_STICKY_CLASS
2499       // These should be classes in the null class loader data, and not all classes
2500       // if !ClassUnloading
2501       StickyClassDumper stiky_class_dumper(&segment_writer);
2502       ClassLoaderData::the_null_class_loader_data()->classes_do(&stiky_class_dumper);
2503     }
2504 
2505     // Heap iteration.
2506     // writes HPROF_GC_INSTANCE_DUMP records.
2507     // After each sub-record is written check_segment_length will be invoked
2508     // to check if the current segment exceeds a threshold. If so, a new
2509     // segment is started.
2510     // The HPROF_GC_CLASS_DUMP and HPROF_GC_INSTANCE_DUMP are the vast bulk
2511     // of the heap dump.
2512 
2513     TraceTime timer(is_parallel_dump() ? "Dump heap objects in parallel" : "Dump heap objects", TRACETIME_LOG(Info, heapdump));
2514     HeapObjectDumper obj_dumper(&segment_writer, this);
2515     if (!is_parallel_dump()) {
2516       Universe::heap()->object_iterate(&obj_dumper);
2517     } else {
2518       // == Parallel dump
2519       _poi->object_iterate(&obj_dumper, worker_id);
2520     }
2521 
2522     segment_writer.finish_dump_segment();
2523     segment_writer.flush();
2524   }
2525 
2526   _dumper_controller->dumper_complete(&segment_writer, writer());
2527 
2528   if (is_vm_dumper(dumper_id)) {
2529     _dumper_controller->wait_all_dumpers_complete();
2530 
2531     // flush global writer
2532     writer()->flush();
2533 
2534     // At this point, all fragments of the heapdump have been written to separate files.
2535     // We need to merge them into a complete heapdump and write HPROF_HEAP_DUMP_END at that time.
2536   }
2537 }
2538 
2539 void VM_HeapDumper::dump_stack_traces(AbstractDumpWriter* writer) {
2540   // write a HPROF_TRACE record without any frames to be referenced as object alloc sites
2541   DumperSupport::write_header(writer, HPROF_TRACE, 3 * sizeof(u4));
2542   writer->write_u4((u4)STACK_TRACE_ID);
2543   writer->write_u4(0);                    // thread number
2544   writer->write_u4(0);                    // frame count
2545 
2546   // max number if every platform thread is carrier with mounted virtual thread
2547   _thread_dumpers = NEW_C_HEAP_ARRAY(ThreadDumper*, Threads::number_of_threads() * 2, mtInternal);
2548 
2549   for (JavaThreadIteratorWithHandle jtiwh; JavaThread * thread = jtiwh.next(); ) {
2550     if (ThreadDumper::should_dump_pthread(thread)) {
2551       bool add_oom_frame = is_oom_thread(thread);
2552 
2553       oop mounted_vt = thread->is_vthread_mounted() ? thread->vthread() : nullptr;
2554       if (mounted_vt != nullptr && !ThreadDumper::should_dump_vthread(mounted_vt)) {
2555         mounted_vt = nullptr;
2556       }
2557 
2558       // mounted vthread (if any)
2559       if (mounted_vt != nullptr) {
2560         ThreadDumper* thread_dumper = new ThreadDumper(ThreadDumper::ThreadType::MountedVirtual, thread, mounted_vt);
2561         _thread_dumpers[_thread_dumpers_count++] = thread_dumper;
2562         if (add_oom_frame) {
2563           thread_dumper->add_oom_frame(_oome_constructor);
2564           // we add oom frame to the VT stack, don't add it to the carrier thread stack
2565           add_oom_frame = false;
2566         }
2567         thread_dumper->init_serial_nums(&_thread_serial_num, &_frame_serial_num);
2568         thread_dumper->dump_stack_traces(writer, _klass_map);
2569       }
2570 
2571       // platform or carrier thread
2572       ThreadDumper* thread_dumper = new ThreadDumper(ThreadDumper::ThreadType::Platform, thread, thread->threadObj());
2573       _thread_dumpers[_thread_dumpers_count++] = thread_dumper;
2574       if (add_oom_frame) {
2575         thread_dumper->add_oom_frame(_oome_constructor);
2576       }
2577       thread_dumper->init_serial_nums(&_thread_serial_num, &_frame_serial_num);
2578       thread_dumper->dump_stack_traces(writer, _klass_map);
2579     }
2580   }
2581 }
2582 
2583 void VM_HeapDumper::dump_vthread(oop vt, AbstractDumpWriter* segment_writer) {
2584   // unmounted vthread has no JavaThread
2585   ThreadDumper thread_dumper(ThreadDumper::ThreadType::UnmountedVirtual, nullptr, vt);
2586   thread_dumper.init_serial_nums(&_thread_serial_num, &_frame_serial_num);
2587 
2588   // write HPROF_TRACE/HPROF_FRAME records to global writer
2589   _dumper_controller->lock_global_writer();
2590   thread_dumper.dump_stack_traces(writer(), _klass_map);
2591   _dumper_controller->unlock_global_writer();
2592 
2593   // write HPROF_GC_ROOT_THREAD_OBJ/HPROF_GC_ROOT_JAVA_FRAME/HPROF_GC_ROOT_JNI_LOCAL subrecord
2594   // to segment writer
2595   thread_dumper.dump_thread_obj(segment_writer);
2596   thread_dumper.dump_stack_refs(segment_writer);
2597 }
2598 
2599 // dump the heap to given path.
2600 int HeapDumper::dump(const char* path, outputStream* out, int compression, bool overwrite, uint num_dump_threads) {
2601   assert(path != nullptr && strlen(path) > 0, "path missing");
2602 
2603   // print message in interactive case
2604   if (out != nullptr) {
2605     out->print_cr("Dumping heap to %s ...", path);
2606     timer()->start();
2607   }
2608 
2609   if (_oome && num_dump_threads > 1) {
2610     // Each additional parallel writer requires several MB of internal memory
2611     // (DumpWriter buffer, DumperClassCacheTable, GZipCompressor buffers).
2612     // For the OOM handling we may already be limited in memory.
2613     // Lets ensure we have at least 20MB per thread.
2614     julong max_threads = os::free_memory() / (20 * M);
2615     if (num_dump_threads > max_threads) {
2616       num_dump_threads = MAX2<uint>(1, (uint)max_threads);
2617     }
2618   }
2619 
2620   // create JFR event
2621   EventHeapDump event;
2622 
2623   AbstractCompressor* compressor = nullptr;
2624 
2625   if (compression > 0) {
2626     compressor = new (std::nothrow) GZipCompressor(compression);
2627 
2628     if (compressor == nullptr) {
2629       set_error("Could not allocate gzip compressor");
2630       return -1;
2631     }
2632   }
2633 
2634   DumpWriter writer(path, overwrite, compressor);
2635 
2636   if (writer.error() != nullptr) {
2637     set_error(writer.error());
2638     if (out != nullptr) {
2639       out->print_cr("Unable to create %s: %s", path,
2640         (error() != nullptr) ? error() : "reason unknown");
2641     }
2642     return -1;
2643   }
2644 
2645   // generate the segmented heap dump into separate files
2646   VM_HeapDumper dumper(&writer, _gc_before_heap_dump, _oome, num_dump_threads);
2647   VMThread::execute(&dumper);
2648 
2649   // record any error that the writer may have encountered
2650   set_error(writer.error());
2651 
2652   // Heap dump process is done in two phases
2653   //
2654   // Phase 1: Concurrent threads directly write heap data to multiple heap files.
2655   //          This is done by VM_HeapDumper, which is performed within safepoint.
2656   //
2657   // Phase 2: Merge multiple heap files into one complete heap dump file.
2658   //          This is done by DumpMerger, which is performed outside safepoint
2659 
2660   DumpMerger merger(path, &writer, dumper.dump_seq());
2661   // Perform heapdump file merge operation in the current thread prevents us
2662   // from occupying the VM Thread, which in turn affects the occurrence of
2663   // GC and other VM operations.
2664   merger.do_merge();
2665   if (writer.error() != nullptr) {
2666     set_error(writer.error());
2667   }
2668 
2669   // emit JFR event
2670   if (error() == nullptr) {
2671     event.set_destination(path);
2672     event.set_gcBeforeDump(_gc_before_heap_dump);
2673     event.set_size(writer.bytes_written());
2674     event.set_onOutOfMemoryError(_oome);
2675     event.set_overwrite(overwrite);
2676     event.set_compression(compression);
2677     event.commit();
2678   } else {
2679     log_debug(cds, heap)("Error %s while dumping heap", error());
2680   }
2681 
2682   // print message in interactive case
2683   if (out != nullptr) {
2684     timer()->stop();
2685     if (error() == nullptr) {
2686       out->print_cr("Heap dump file created [" JULONG_FORMAT " bytes in %3.3f secs]",
2687                     writer.bytes_written(), timer()->seconds());
2688     } else {
2689       out->print_cr("Dump file is incomplete: %s", writer.error());
2690     }
2691   }
2692 
2693   if (compressor != nullptr) {
2694     delete compressor;
2695   }
2696   return (writer.error() == nullptr) ? 0 : -1;
2697 }
2698 
2699 // stop timer (if still active), and free any error string we might be holding
2700 HeapDumper::~HeapDumper() {
2701   if (timer()->is_active()) {
2702     timer()->stop();
2703   }
2704   set_error(nullptr);
2705 }
2706 
2707 
2708 // returns the error string (resource allocated), or null
2709 char* HeapDumper::error_as_C_string() const {
2710   if (error() != nullptr) {
2711     char* str = NEW_RESOURCE_ARRAY(char, strlen(error())+1);
2712     strcpy(str, error());
2713     return str;
2714   } else {
2715     return nullptr;
2716   }
2717 }
2718 
2719 // set the error string
2720 void HeapDumper::set_error(char const* error) {
2721   if (_error != nullptr) {
2722     os::free(_error);
2723   }
2724   if (error == nullptr) {
2725     _error = nullptr;
2726   } else {
2727     _error = os::strdup(error);
2728     assert(_error != nullptr, "allocation failure");
2729   }
2730 }
2731 
2732 // Called by out-of-memory error reporting by a single Java thread
2733 // outside of a JVM safepoint
2734 void HeapDumper::dump_heap_from_oome() {
2735   HeapDumper::dump_heap(true);
2736 }
2737 
2738 // Called by error reporting by a single Java thread outside of a JVM safepoint,
2739 // or by heap dumping by the VM thread during a (GC) safepoint. Thus, these various
2740 // callers are strictly serialized and guaranteed not to interfere below. For more
2741 // general use, however, this method will need modification to prevent
2742 // inteference when updating the static variables base_path and dump_file_seq below.
2743 void HeapDumper::dump_heap() {
2744   HeapDumper::dump_heap(false);
2745 }
2746 
2747 void HeapDumper::dump_heap(bool oome) {
2748   static char base_path[JVM_MAXPATHLEN] = {'\0'};
2749   static uint dump_file_seq = 0;
2750   char* my_path;
2751   const int max_digit_chars = 20;
2752 
2753   const char* dump_file_name = "java_pid";
2754   const char* dump_file_ext  = HeapDumpGzipLevel > 0 ? ".hprof.gz" : ".hprof";
2755 
2756   // The dump file defaults to java_pid<pid>.hprof in the current working
2757   // directory. HeapDumpPath=<file> can be used to specify an alternative
2758   // dump file name or a directory where dump file is created.
2759   if (dump_file_seq == 0) { // first time in, we initialize base_path
2760     // Calculate potentially longest base path and check if we have enough
2761     // allocated statically.
2762     const size_t total_length =
2763                       (HeapDumpPath == nullptr ? 0 : strlen(HeapDumpPath)) +
2764                       strlen(os::file_separator()) + max_digit_chars +
2765                       strlen(dump_file_name) + strlen(dump_file_ext) + 1;
2766     if (total_length > sizeof(base_path)) {
2767       warning("Cannot create heap dump file.  HeapDumpPath is too long.");
2768       return;
2769     }
2770 
2771     bool use_default_filename = true;
2772     if (HeapDumpPath == nullptr || HeapDumpPath[0] == '\0') {
2773       // HeapDumpPath=<file> not specified
2774     } else {
2775       strcpy(base_path, HeapDumpPath);
2776       // check if the path is a directory (must exist)
2777       DIR* dir = os::opendir(base_path);
2778       if (dir == nullptr) {
2779         use_default_filename = false;
2780       } else {
2781         // HeapDumpPath specified a directory. We append a file separator
2782         // (if needed).
2783         os::closedir(dir);
2784         size_t fs_len = strlen(os::file_separator());
2785         if (strlen(base_path) >= fs_len) {
2786           char* end = base_path;
2787           end += (strlen(base_path) - fs_len);
2788           if (strcmp(end, os::file_separator()) != 0) {
2789             strcat(base_path, os::file_separator());
2790           }
2791         }
2792       }
2793     }
2794     // If HeapDumpPath wasn't a file name then we append the default name
2795     if (use_default_filename) {
2796       const size_t dlen = strlen(base_path);  // if heap dump dir specified
2797       jio_snprintf(&base_path[dlen], sizeof(base_path)-dlen, "%s%d%s",
2798                    dump_file_name, os::current_process_id(), dump_file_ext);
2799     }
2800     const size_t len = strlen(base_path) + 1;
2801     my_path = (char*)os::malloc(len, mtInternal);
2802     if (my_path == nullptr) {
2803       warning("Cannot create heap dump file.  Out of system memory.");
2804       return;
2805     }
2806     strncpy(my_path, base_path, len);
2807   } else {
2808     // Append a sequence number id for dumps following the first
2809     const size_t len = strlen(base_path) + max_digit_chars + 2; // for '.' and \0
2810     my_path = (char*)os::malloc(len, mtInternal);
2811     if (my_path == nullptr) {
2812       warning("Cannot create heap dump file.  Out of system memory.");
2813       return;
2814     }
2815     jio_snprintf(my_path, len, "%s.%d", base_path, dump_file_seq);
2816   }
2817   dump_file_seq++;   // increment seq number for next time we dump
2818 
2819   HeapDumper dumper(false /* no GC before heap dump */,
2820                     oome  /* pass along out-of-memory-error flag */);
2821   dumper.dump(my_path, tty, HeapDumpGzipLevel);
2822   os::free(my_path);
2823 }