1  /*
   2  * Copyright (c) 2012, 2023, Oracle and/or its affiliates. All rights reserved.
   3  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
   4  *
   5  * This code is free software; you can redistribute it and/or modify it
   6  * under the terms of the GNU General Public License version 2 only, as
   7  * published by the Free Software Foundation.
   8  *
   9  * This code is distributed in the hope that it will be useful, but WITHOUT
  10  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
  11  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
  12  * version 2 for more details (a copy is included in the LICENSE file that
  13  * accompanied this code).
  14  *
  15  * You should have received a copy of the GNU General Public License version
  16  * 2 along with this work; if not, write to the Free Software Foundation,
  17  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
  18  *
  19  * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
  20  * or visit www.oracle.com if you need additional information or have any
  21  * questions.
  22  *
  23  */
  24 
  25 // A ClassLoaderData identifies the full set of class types that a class
  26 // loader's name resolution strategy produces for a given configuration of the
  27 // class loader.
  28 // Class types in the ClassLoaderData may be defined by from class file binaries
  29 // provided by the class loader, or from other class loader it interacts with
  30 // according to its name resolution strategy.
  31 //
  32 // Class loaders that implement a deterministic name resolution strategy
  33 // (including with respect to their delegation behavior), such as the boot, the
  34 // platform, and the system loaders of the JDK's built-in class loader
  35 // hierarchy, always produce the same linkset for a given configuration.
  36 //
  37 // ClassLoaderData carries information related to a linkset (e.g.,
  38 // metaspace holding its klass definitions).
  39 // The System Dictionary and related data structures (e.g., placeholder table,
  40 // loader constraints table) as well as the runtime representation of classes
  41 // only reference ClassLoaderData.
  42 //
  43 // Instances of java.lang.ClassLoader holds a pointer to a ClassLoaderData that
  44 // that represent the loader's "linking domain" in the JVM.
  45 //
  46 // The bootstrap loader (represented by null) also has a ClassLoaderData,
  47 // the singleton class the_null_class_loader_data().
  48 
  49 #include "precompiled.hpp"
  50 #include "classfile/classLoaderData.inline.hpp"
  51 #include "classfile/classLoaderDataGraph.inline.hpp"
  52 #include "classfile/dictionary.hpp"
  53 #include "classfile/javaClasses.inline.hpp"
  54 #include "classfile/moduleEntry.hpp"
  55 #include "classfile/packageEntry.hpp"
  56 #include "classfile/symbolTable.hpp"
  57 #include "classfile/systemDictionary.hpp"
  58 #include "classfile/systemDictionaryShared.hpp"
  59 #include "classfile/vmClasses.hpp"
  60 #include "logging/log.hpp"
  61 #include "logging/logStream.hpp"
  62 #include "memory/allocation.inline.hpp"
  63 #include "memory/classLoaderMetaspace.hpp"
  64 #include "memory/metadataFactory.hpp"
  65 #include "memory/metaspace.hpp"
  66 #include "memory/resourceArea.hpp"
  67 #include "memory/universe.hpp"
  68 #include "oops/access.inline.hpp"
  69 #include "oops/inlineKlass.inline.hpp"
  70 #include "oops/klass.inline.hpp"
  71 #include "oops/oop.inline.hpp"
  72 #include "oops/oopHandle.inline.hpp"
  73 #include "oops/verifyOopClosure.hpp"
  74 #include "oops/weakHandle.inline.hpp"
  75 #include "runtime/arguments.hpp"
  76 #include "runtime/atomic.hpp"
  77 #include "runtime/handles.inline.hpp"
  78 #include "runtime/mutex.hpp"
  79 #include "runtime/safepoint.hpp"
  80 #include "utilities/growableArray.hpp"
  81 #include "utilities/macros.hpp"
  82 #include "utilities/ostream.hpp"
  83 
  84 ClassLoaderData * ClassLoaderData::_the_null_class_loader_data = nullptr;
  85 
  86 void ClassLoaderData::init_null_class_loader_data() {
  87   assert(_the_null_class_loader_data == nullptr, "cannot initialize twice");
  88   assert(ClassLoaderDataGraph::_head == nullptr, "cannot initialize twice");
  89 
  90   _the_null_class_loader_data = new ClassLoaderData(Handle(), false);
  91   ClassLoaderDataGraph::_head = _the_null_class_loader_data;
  92   assert(_the_null_class_loader_data->is_the_null_class_loader_data(), "Must be");
  93 
  94   LogTarget(Trace, class, loader, data) lt;
  95   if (lt.is_enabled()) {
  96     ResourceMark rm;
  97     LogStream ls(lt);
  98     ls.print("create ");
  99     _the_null_class_loader_data->print_value_on(&ls);
 100     ls.cr();
 101   }
 102 }
 103 
 104 // Obtain and set the class loader's name within the ClassLoaderData so
 105 // it will be available for error messages, logging, JFR, etc.  The name
 106 // and klass are available after the class_loader oop is no longer alive,
 107 // during unloading.
 108 void ClassLoaderData::initialize_name(Handle class_loader) {
 109   ResourceMark rm;
 110 
 111   // Obtain the class loader's name.  If the class loader's name was not
 112   // explicitly set during construction, the CLD's _name field will be null.
 113   oop cl_name = java_lang_ClassLoader::name(class_loader());
 114   if (cl_name != nullptr) {
 115     const char* cl_instance_name = java_lang_String::as_utf8_string(cl_name);
 116 
 117     if (cl_instance_name != nullptr && cl_instance_name[0] != '\0') {
 118       _name = SymbolTable::new_symbol(cl_instance_name);
 119     }
 120   }
 121 
 122   // Obtain the class loader's name and identity hash.  If the class loader's
 123   // name was not explicitly set during construction, the class loader's name and id
 124   // will be set to the qualified class name of the class loader along with its
 125   // identity hash.
 126   // If for some reason the ClassLoader's constructor has not been run, instead of
 127   // leaving the _name_and_id field null, fall back to the external qualified class
 128   // name.  Thus CLD's _name_and_id field should never have a null value.
 129   oop cl_name_and_id = java_lang_ClassLoader::nameAndId(class_loader());
 130   const char* cl_instance_name_and_id =
 131                   (cl_name_and_id == nullptr) ? _class_loader_klass->external_name() :
 132                                              java_lang_String::as_utf8_string(cl_name_and_id);
 133   assert(cl_instance_name_and_id != nullptr && cl_instance_name_and_id[0] != '\0', "class loader has no name and id");
 134   _name_and_id = SymbolTable::new_symbol(cl_instance_name_and_id);
 135 }
 136 
 137 ClassLoaderData::ClassLoaderData(Handle h_class_loader, bool has_class_mirror_holder) :
 138   _metaspace(nullptr),
 139   _metaspace_lock(new Mutex(Mutex::nosafepoint-2, "MetaspaceAllocation_lock")),
 140   _unloading(false), _has_class_mirror_holder(has_class_mirror_holder),
 141   _modified_oops(true),
 142   // A non-strong hidden class loader data doesn't have anything to keep
 143   // it from being unloaded during parsing of the non-strong hidden class.
 144   // The null-class-loader should always be kept alive.
 145   _keep_alive((has_class_mirror_holder || h_class_loader.is_null()) ? 1 : 0),
 146   _claim(0),
 147   _handles(),
 148   _klasses(nullptr), _packages(nullptr), _modules(nullptr), _unnamed_module(nullptr), _dictionary(nullptr),
 149   _jmethod_ids(nullptr),
 150   _deallocate_list(nullptr),
 151   _next(nullptr),
 152   _unloading_next(nullptr),
 153   _class_loader_klass(nullptr), _name(nullptr), _name_and_id(nullptr) {
 154 
 155   if (!h_class_loader.is_null()) {
 156     _class_loader = _handles.add(h_class_loader());
 157     _class_loader_klass = h_class_loader->klass();
 158     initialize_name(h_class_loader);
 159   }
 160 
 161   if (!has_class_mirror_holder) {
 162     // The holder is initialized later for non-strong hidden classes,
 163     // and before calling anything that call class_loader().
 164     initialize_holder(h_class_loader);
 165 
 166     // A ClassLoaderData created solely for a non-strong hidden class should never
 167     // have a ModuleEntryTable or PackageEntryTable created for it.
 168     _packages = new PackageEntryTable();
 169     if (h_class_loader.is_null()) {
 170       // Create unnamed module for boot loader
 171       _unnamed_module = ModuleEntry::create_boot_unnamed_module(this);
 172     } else {
 173       // Create unnamed module for all other loaders
 174       _unnamed_module = ModuleEntry::create_unnamed_module(this);
 175     }
 176     _dictionary = create_dictionary();
 177   }
 178 
 179   NOT_PRODUCT(_dependency_count = 0); // number of class loader dependencies
 180 
 181   JFR_ONLY(INIT_ID(this);)
 182 }
 183 
 184 ClassLoaderData::ChunkedHandleList::~ChunkedHandleList() {
 185   Chunk* c = _head;
 186   while (c != nullptr) {
 187     Chunk* next = c->_next;
 188     delete c;
 189     c = next;
 190   }
 191 }
 192 
 193 OopHandle ClassLoaderData::ChunkedHandleList::add(oop o) {
 194   if (_head == nullptr || _head->_size == Chunk::CAPACITY) {
 195     Chunk* next = new Chunk(_head);
 196     Atomic::release_store(&_head, next);
 197   }
 198   oop* handle = &_head->_data[_head->_size];
 199   NativeAccess<IS_DEST_UNINITIALIZED>::oop_store(handle, o);
 200   Atomic::release_store(&_head->_size, _head->_size + 1);
 201   return OopHandle(handle);
 202 }
 203 
 204 int ClassLoaderData::ChunkedHandleList::count() const {
 205   int count = 0;
 206   Chunk* chunk = _head;
 207   while (chunk != nullptr) {
 208     count += chunk->_size;
 209     chunk = chunk->_next;
 210   }
 211   return count;
 212 }
 213 
 214 inline void ClassLoaderData::ChunkedHandleList::oops_do_chunk(OopClosure* f, Chunk* c, const juint size) {
 215   for (juint i = 0; i < size; i++) {
 216     f->do_oop(&c->_data[i]);
 217   }
 218 }
 219 
 220 void ClassLoaderData::ChunkedHandleList::oops_do(OopClosure* f) {
 221   Chunk* head = Atomic::load_acquire(&_head);
 222   if (head != nullptr) {
 223     // Must be careful when reading size of head
 224     oops_do_chunk(f, head, Atomic::load_acquire(&head->_size));
 225     for (Chunk* c = head->_next; c != nullptr; c = c->_next) {
 226       oops_do_chunk(f, c, c->_size);
 227     }
 228   }
 229 }
 230 
 231 class VerifyContainsOopClosure : public OopClosure {
 232   oop  _target;
 233   bool _found;
 234 
 235  public:
 236   VerifyContainsOopClosure(oop target) : _target(target), _found(false) {}
 237 
 238   void do_oop(oop* p) {
 239     if (p != nullptr && NativeAccess<AS_NO_KEEPALIVE>::oop_load(p) == _target) {
 240       _found = true;
 241     }
 242   }
 243 
 244   void do_oop(narrowOop* p) {
 245     // The ChunkedHandleList should not contain any narrowOop
 246     ShouldNotReachHere();
 247   }
 248 
 249   bool found() const {
 250     return _found;
 251   }
 252 };
 253 
 254 bool ClassLoaderData::ChunkedHandleList::contains(oop p) {
 255   VerifyContainsOopClosure cl(p);
 256   oops_do(&cl);
 257   return cl.found();
 258 }
 259 
 260 #ifndef PRODUCT
 261 bool ClassLoaderData::ChunkedHandleList::owner_of(oop* oop_handle) {
 262   Chunk* chunk = _head;
 263   while (chunk != nullptr) {
 264     if (&(chunk->_data[0]) <= oop_handle && oop_handle < &(chunk->_data[chunk->_size])) {
 265       return true;
 266     }
 267     chunk = chunk->_next;
 268   }
 269   return false;
 270 }
 271 #endif // PRODUCT
 272 
 273 void ClassLoaderData::clear_claim(int claim) {
 274   for (;;) {
 275     int old_claim = Atomic::load(&_claim);
 276     if ((old_claim & claim) == 0) {
 277       return;
 278     }
 279     int new_claim = old_claim & ~claim;
 280     if (Atomic::cmpxchg(&_claim, old_claim, new_claim) == old_claim) {
 281       return;
 282     }
 283   }
 284 }
 285 
 286 #ifdef ASSERT
 287 void ClassLoaderData::verify_not_claimed(int claim) {
 288   assert((_claim & claim) == 0, "Found claim: %d bits in _claim: %d", claim, _claim);
 289 }
 290 #endif
 291 
 292 bool ClassLoaderData::try_claim(int claim) {
 293   for (;;) {
 294     int old_claim = Atomic::load(&_claim);
 295     if ((old_claim & claim) == claim) {
 296       return false;
 297     }
 298     int new_claim = old_claim | claim;
 299     if (Atomic::cmpxchg(&_claim, old_claim, new_claim) == old_claim) {
 300       return true;
 301     }
 302   }
 303 }
 304 
 305 void ClassLoaderData::demote_strong_roots() {
 306   // The oop handle area contains strong roots that the GC traces from. We are about
 307   // to demote them to strong native oops that the GC does *not* trace from. Conceptually,
 308   // we are retiring a rather normal strong root, and creating a strong non-root handle,
 309   // which happens to reuse the same address as the normal strong root had.
 310   // Unless we invoke the right barriers, the GC might not notice that a strong root
 311   // has been pulled from the system, and is left unprocessed by the GC. There can be
 312   // several consequences:
 313   // 1. A concurrently marking snapshot-at-the-beginning GC might assume that the contents
 314   //    of all strong roots get processed by the GC in order to keep them alive. Without
 315   //    barriers, some objects might not be kept alive.
 316   // 2. A concurrently relocating GC might assume that after moving an object, a subsequent
 317   //    tracing from all roots can fix all the pointers in the system, which doesn't play
 318   //    well with roots racingly being pulled.
 319   // 3. A concurrent GC using colored pointers, might assume that tracing the object graph
 320   //    from roots results in all pointers getting some particular color, which also doesn't
 321   //    play well with roots being pulled out from the system concurrently.
 322 
 323   class TransitionRootsOopClosure : public OopClosure {
 324   public:
 325     virtual void do_oop(oop* p) {
 326       // By loading the strong root with the access API, we can use the right barriers to
 327       // store the oop as a strong non-root handle, that happens to reuse the same memory
 328       // address as the strong root. The barriered store ensures that:
 329       // 1. The concurrent SATB marking properties are satisfied as the store will keep
 330       //    the oop alive.
 331       // 2. The concurrent object movement properties are satisfied as we store the address
 332       //    of the new location of the object, if any.
 333       // 3. The colors if any will be stored as the new good colors.
 334       oop obj = NativeAccess<>::oop_load(p); // Load the strong root
 335       NativeAccess<>::oop_store(p, obj); // Store the strong non-root
 336     }
 337 
 338     virtual void do_oop(narrowOop* p) {
 339       ShouldNotReachHere();
 340     }
 341   } cl;
 342   oops_do(&cl, ClassLoaderData::_claim_none, false /* clear_mod_oops */);
 343 }
 344 
 345 // Non-strong hidden classes have their own ClassLoaderData that is marked to keep alive
 346 // while the class is being parsed, and if the class appears on the module fixup list.
 347 // Due to the uniqueness that no other class shares the hidden class' name or
 348 // ClassLoaderData, no other non-GC thread has knowledge of the hidden class while
 349 // it is being defined, therefore _keep_alive is not volatile or atomic.
 350 void ClassLoaderData::inc_keep_alive() {
 351   if (has_class_mirror_holder()) {
 352     assert(_keep_alive > 0, "Invalid keep alive increment count");
 353     _keep_alive++;
 354   }
 355 }
 356 
 357 void ClassLoaderData::dec_keep_alive() {
 358   if (has_class_mirror_holder()) {
 359     assert(_keep_alive > 0, "Invalid keep alive decrement count");
 360     if (_keep_alive == 1) {
 361       // When the keep_alive counter is 1, the oop handle area is a strong root,
 362       // acting as input to the GC tracing. Such strong roots are part of the
 363       // snapshot-at-the-beginning, and can not just be pulled out from the
 364       // system when concurrent GCs are running at the same time, without
 365       // invoking the right barriers.
 366       demote_strong_roots();
 367     }
 368     _keep_alive--;
 369   }
 370 }
 371 
 372 void ClassLoaderData::oops_do(OopClosure* f, int claim_value, bool clear_mod_oops) {
 373   if (claim_value != ClassLoaderData::_claim_none && !try_claim(claim_value)) {
 374     return;
 375   }
 376 
 377   // Only clear modified_oops after the ClassLoaderData is claimed.
 378   if (clear_mod_oops) {
 379     clear_modified_oops();
 380   }
 381 
 382   _handles.oops_do(f);
 383 }
 384 
 385 void ClassLoaderData::classes_do(KlassClosure* klass_closure) {
 386   // Lock-free access requires load_acquire
 387   for (Klass* k = Atomic::load_acquire(&_klasses); k != nullptr; k = k->next_link()) {
 388     klass_closure->do_klass(k);
 389     assert(k != k->next_link(), "no loops!");
 390   }
 391 }
 392 
 393 void ClassLoaderData::classes_do(void f(Klass * const)) {
 394   // Lock-free access requires load_acquire
 395   for (Klass* k = Atomic::load_acquire(&_klasses); k != nullptr; k = k->next_link()) {
 396     f(k);
 397     assert(k != k->next_link(), "no loops!");
 398   }
 399 }
 400 
 401 void ClassLoaderData::methods_do(void f(Method*)) {
 402   // Lock-free access requires load_acquire
 403   for (Klass* k = Atomic::load_acquire(&_klasses); k != nullptr; k = k->next_link()) {
 404     if (k->is_instance_klass() && InstanceKlass::cast(k)->is_loaded()) {
 405       InstanceKlass::cast(k)->methods_do(f);
 406     }
 407   }
 408 }
 409 
 410 void ClassLoaderData::loaded_classes_do(KlassClosure* klass_closure) {
 411   // To call this, one must have the MultiArray_lock held, but the _klasses list still has lock free reads.
 412   assert_locked_or_safepoint(MultiArray_lock);
 413 
 414   // Lock-free access requires load_acquire
 415   for (Klass* k = Atomic::load_acquire(&_klasses); k != nullptr; k = k->next_link()) {
 416     // Filter out InstanceKlasses (or their ObjArrayKlasses) that have not entered the
 417     // loaded state.
 418     if (k->is_instance_klass()) {
 419       if (!InstanceKlass::cast(k)->is_loaded()) {
 420         continue;
 421       }
 422     } else if (k->is_shared() && k->is_objArray_klass()) {
 423       Klass* bottom = ObjArrayKlass::cast(k)->bottom_klass();
 424       if (bottom->is_instance_klass() && !InstanceKlass::cast(bottom)->is_loaded()) {
 425         // This could happen if <bottom> is a shared class that has been restored
 426         // but is not yet marked as loaded. All archived array classes of the
 427         // bottom class are already restored and placed in the _klasses list.
 428         continue;
 429       }
 430     }
 431 
 432 #ifdef ASSERT
 433     oop m = k->java_mirror();
 434     assert(m != nullptr, "nullptr mirror");
 435     assert(m->is_a(vmClasses::Class_klass()), "invalid mirror");
 436 #endif
 437     klass_closure->do_klass(k);
 438   }
 439 }
 440 
 441 void ClassLoaderData::classes_do(void f(InstanceKlass*)) {
 442   // Lock-free access requires load_acquire
 443   for (Klass* k = Atomic::load_acquire(&_klasses); k != nullptr; k = k->next_link()) {
 444     if (k->is_instance_klass()) {
 445       f(InstanceKlass::cast(k));
 446     }
 447     assert(k != k->next_link(), "no loops!");
 448   }
 449 }
 450 
 451 void ClassLoaderData::inline_classes_do(void f(InlineKlass*)) {
 452   // Lock-free access requires load_acquire
 453   for (Klass* k = Atomic::load_acquire(&_klasses); k != nullptr; k = k->next_link()) {
 454     if (k->is_inline_klass()) {
 455       f(InlineKlass::cast(k));
 456     }
 457     assert(k != k->next_link(), "no loops!");
 458   }
 459 }
 460 
 461 void ClassLoaderData::modules_do(void f(ModuleEntry*)) {
 462   assert_locked_or_safepoint(Module_lock);
 463   if (_unnamed_module != nullptr) {
 464     f(_unnamed_module);
 465   }
 466   if (_modules != nullptr) {
 467     _modules->modules_do(f);
 468   }
 469 }
 470 
 471 void ClassLoaderData::packages_do(void f(PackageEntry*)) {
 472   assert_locked_or_safepoint(Module_lock);
 473   if (_packages != nullptr) {
 474     _packages->packages_do(f);
 475   }
 476 }
 477 
 478 void ClassLoaderData::record_dependency(const Klass* k) {
 479   assert(k != nullptr, "invariant");
 480 
 481   ClassLoaderData * const from_cld = this;
 482   ClassLoaderData * const to_cld = k->class_loader_data();
 483 
 484   // Do not need to record dependency if the dependency is to a class whose
 485   // class loader data is never freed.  (i.e. the dependency's class loader
 486   // is one of the three builtin class loaders and the dependency's class
 487   // loader data has a ClassLoader holder, not a Class holder.)
 488   if (to_cld->is_permanent_class_loader_data()) {
 489     return;
 490   }
 491 
 492   oop to;
 493   if (to_cld->has_class_mirror_holder()) {
 494     // Just return if a non-strong hidden class class is attempting to record a dependency
 495     // to itself.  (Note that every non-strong hidden class has its own unique class
 496     // loader data.)
 497     if (to_cld == from_cld) {
 498       return;
 499     }
 500     // Hidden class dependencies are through the mirror.
 501     to = k->java_mirror();
 502   } else {
 503     to = to_cld->class_loader();
 504     oop from = from_cld->class_loader();
 505 
 506     // Just return if this dependency is to a class with the same or a parent
 507     // class_loader.
 508     if (from == to || java_lang_ClassLoader::isAncestor(from, to)) {
 509       return; // this class loader is in the parent list, no need to add it.
 510     }
 511   }
 512 
 513   // It's a dependency we won't find through GC, add it.
 514   if (!_handles.contains(to)) {
 515     NOT_PRODUCT(Atomic::inc(&_dependency_count));
 516     LogTarget(Trace, class, loader, data) lt;
 517     if (lt.is_enabled()) {
 518       ResourceMark rm;
 519       LogStream ls(lt);
 520       ls.print("adding dependency from ");
 521       print_value_on(&ls);
 522       ls.print(" to ");
 523       to_cld->print_value_on(&ls);
 524       ls.cr();
 525     }
 526     Handle dependency(Thread::current(), to);
 527     add_handle(dependency);
 528     // Added a potentially young gen oop to the ClassLoaderData
 529     record_modified_oops();
 530   }
 531 }
 532 
 533 void ClassLoaderData::add_class(Klass* k, bool publicize /* true */) {
 534   {
 535     MutexLocker ml(metaspace_lock(), Mutex::_no_safepoint_check_flag);
 536     Klass* old_value = _klasses;
 537     k->set_next_link(old_value);
 538     // Link the new item into the list, making sure the linked class is stable
 539     // since the list can be walked without a lock
 540     Atomic::release_store(&_klasses, k);
 541     if (k->is_array_klass()) {
 542       ClassLoaderDataGraph::inc_array_classes(1);
 543     } else {
 544       ClassLoaderDataGraph::inc_instance_classes(1);
 545     }
 546   }
 547 
 548   if (publicize) {
 549     LogTarget(Trace, class, loader, data) lt;
 550     if (lt.is_enabled()) {
 551       ResourceMark rm;
 552       LogStream ls(lt);
 553       ls.print("Adding k: " PTR_FORMAT " %s to ", p2i(k), k->external_name());
 554       print_value_on(&ls);
 555       ls.cr();
 556     }
 557   }
 558 }
 559 
 560 void ClassLoaderData::initialize_holder(Handle loader_or_mirror) {
 561   if (loader_or_mirror() != nullptr) {
 562     assert(_holder.is_null(), "never replace holders");
 563     _holder = WeakHandle(Universe::vm_weak(), loader_or_mirror);
 564   }
 565 }
 566 
 567 // Remove a klass from the _klasses list for scratch_class during redefinition
 568 // or parsed class in the case of an error.
 569 void ClassLoaderData::remove_class(Klass* scratch_class) {
 570   assert_locked_or_safepoint(ClassLoaderDataGraph_lock);
 571 
 572   // Adjust global class iterator.
 573   ClassLoaderDataGraph::adjust_saved_class(scratch_class);
 574 
 575   Klass* prev = nullptr;
 576   for (Klass* k = _klasses; k != nullptr; k = k->next_link()) {
 577     if (k == scratch_class) {
 578       if (prev == nullptr) {
 579         _klasses = k->next_link();
 580       } else {
 581         Klass* next = k->next_link();
 582         prev->set_next_link(next);
 583       }
 584 
 585       if (k->is_array_klass()) {
 586         ClassLoaderDataGraph::dec_array_classes(1);
 587       } else {
 588         ClassLoaderDataGraph::dec_instance_classes(1);
 589       }
 590 
 591       return;
 592     }
 593     prev = k;
 594     assert(k != k->next_link(), "no loops!");
 595   }
 596   ShouldNotReachHere();   // should have found this class!!
 597 }
 598 
 599 void ClassLoaderData::unload() {
 600   _unloading = true;
 601 
 602   LogTarget(Trace, class, loader, data) lt;
 603   if (lt.is_enabled()) {
 604     ResourceMark rm;
 605     LogStream ls(lt);
 606     ls.print("unload");
 607     print_value_on(&ls);
 608     ls.cr();
 609   }
 610 
 611   // Some items on the _deallocate_list need to free their C heap structures
 612   // if they are not already on the _klasses list.
 613   free_deallocate_list_C_heap_structures();
 614 
 615   inline_classes_do(InlineKlass::cleanup);
 616 
 617   // Clean up class dependencies and tell serviceability tools
 618   // these classes are unloading.  Must be called
 619   // after erroneous classes are released.
 620   classes_do(InstanceKlass::unload_class);
 621 
 622   // Method::clear_jmethod_ids only sets the jmethod_ids to null without
 623   // releasing the memory for related JNIMethodBlocks and JNIMethodBlockNodes.
 624   // This is done intentionally because native code (e.g. JVMTI agent) holding
 625   // jmethod_ids may access them after the associated classes and class loader
 626   // are unloaded. The Java Native Interface Specification says "method ID
 627   // does not prevent the VM from unloading the class from which the ID has
 628   // been derived. After the class is unloaded, the method or field ID becomes
 629   // invalid". In real world usages, the native code may rely on jmethod_ids
 630   // being null after class unloading. Hence, it is unsafe to free the memory
 631   // from the VM side without knowing when native code is going to stop using
 632   // them.
 633   if (_jmethod_ids != nullptr) {
 634     Method::clear_jmethod_ids(this);
 635   }
 636 
 637   // Clean up global class iterator for compiler
 638   ClassLoaderDataGraph::adjust_saved_class(this);
 639 }
 640 
 641 ModuleEntryTable* ClassLoaderData::modules() {
 642   // Lazily create the module entry table at first request.
 643   // Lock-free access requires load_acquire.
 644   ModuleEntryTable* modules = Atomic::load_acquire(&_modules);
 645   if (modules == nullptr) {
 646     MutexLocker m1(Module_lock);
 647     // Check if _modules got allocated while we were waiting for this lock.
 648     if ((modules = _modules) == nullptr) {
 649       modules = new ModuleEntryTable();
 650 
 651       {
 652         MutexLocker m1(metaspace_lock(), Mutex::_no_safepoint_check_flag);
 653         // Ensure _modules is stable, since it is examined without a lock
 654         Atomic::release_store(&_modules, modules);
 655       }
 656     }
 657   }
 658   return modules;
 659 }
 660 
 661 const int _boot_loader_dictionary_size    = 1009;
 662 const int _default_loader_dictionary_size = 107;
 663 
 664 Dictionary* ClassLoaderData::create_dictionary() {
 665   assert(!has_class_mirror_holder(), "class mirror holder cld does not have a dictionary");
 666   int size;
 667   if (_the_null_class_loader_data == nullptr) {
 668     size = _boot_loader_dictionary_size;
 669   } else if (class_loader()->is_a(vmClasses::reflect_DelegatingClassLoader_klass())) {
 670     size = 1;  // there's only one class in relection class loader and no initiated classes
 671   } else if (is_system_class_loader_data()) {
 672     size = _boot_loader_dictionary_size;
 673   } else {
 674     size = _default_loader_dictionary_size;
 675   }
 676   return new Dictionary(this, size);
 677 }
 678 
 679 // Tell the GC to keep this klass alive. Needed while iterating ClassLoaderDataGraph,
 680 // and any runtime code that uses klasses.
 681 oop ClassLoaderData::holder() const {
 682   // A klass that was previously considered dead can be looked up in the
 683   // CLD/SD, and its _java_mirror or _class_loader can be stored in a root
 684   // or a reachable object making it alive again. The SATB part of G1 needs
 685   // to get notified about this potential resurrection, otherwise the marking
 686   // might not find the object.
 687   if (!_holder.is_null()) {  // null class_loader
 688     return _holder.resolve();
 689   } else {
 690     return nullptr;
 691   }
 692 }
 693 
 694 // Let the GC read the holder without keeping it alive.
 695 oop ClassLoaderData::holder_no_keepalive() const {
 696   if (!_holder.is_null()) {  // null class_loader
 697     return _holder.peek();
 698   } else {
 699     return nullptr;
 700   }
 701 }
 702 
 703 // Unloading support
 704 bool ClassLoaderData::is_alive() const {
 705   bool alive = keep_alive()         // null class loader and incomplete non-strong hidden class.
 706       || (_holder.peek() != nullptr);  // and not cleaned by the GC weak handle processing.
 707 
 708   return alive;
 709 }
 710 
 711 class ReleaseKlassClosure: public KlassClosure {
 712 private:
 713   size_t  _instance_class_released;
 714   size_t  _array_class_released;
 715 public:
 716   ReleaseKlassClosure() : _instance_class_released(0), _array_class_released(0) { }
 717 
 718   size_t instance_class_released() const { return _instance_class_released; }
 719   size_t array_class_released()    const { return _array_class_released;    }
 720 
 721   void do_klass(Klass* k) {
 722     if (k->is_array_klass()) {
 723       _array_class_released ++;
 724     } else {
 725       assert(k->is_instance_klass(), "Must be");
 726       _instance_class_released ++;
 727     }
 728     k->release_C_heap_structures();
 729   }
 730 };
 731 
 732 ClassLoaderData::~ClassLoaderData() {
 733   // Release C heap structures for all the classes.
 734   ReleaseKlassClosure cl;
 735   classes_do(&cl);
 736 
 737   ClassLoaderDataGraph::dec_array_classes(cl.array_class_released());
 738   ClassLoaderDataGraph::dec_instance_classes(cl.instance_class_released());
 739 
 740   // Release the WeakHandle
 741   _holder.release(Universe::vm_weak());
 742 
 743   // Release C heap allocated hashtable for all the packages.
 744   if (_packages != nullptr) {
 745     // Destroy the table itself
 746     delete _packages;
 747     _packages = nullptr;
 748   }
 749 
 750   // Release C heap allocated hashtable for all the modules.
 751   if (_modules != nullptr) {
 752     // Destroy the table itself
 753     delete _modules;
 754     _modules = nullptr;
 755   }
 756 
 757   // Release C heap allocated hashtable for the dictionary
 758   if (_dictionary != nullptr) {
 759     // Destroy the table itself
 760     delete _dictionary;
 761     _dictionary = nullptr;
 762   }
 763 
 764   if (_unnamed_module != nullptr) {
 765     delete _unnamed_module;
 766     _unnamed_module = nullptr;
 767   }
 768 
 769   // release the metaspace
 770   ClassLoaderMetaspace *m = _metaspace;
 771   if (m != nullptr) {
 772     _metaspace = nullptr;
 773     delete m;
 774   }
 775 
 776   // Delete lock
 777   delete _metaspace_lock;
 778 
 779   // Delete free list
 780   if (_deallocate_list != nullptr) {
 781     delete _deallocate_list;
 782   }
 783 
 784   // Decrement refcounts of Symbols if created.
 785   if (_name != nullptr) {
 786     _name->decrement_refcount();
 787   }
 788   if (_name_and_id != nullptr) {
 789     _name_and_id->decrement_refcount();
 790   }
 791 }
 792 
 793 // Returns true if this class loader data is for the app class loader
 794 // or a user defined system class loader.  (Note that the class loader
 795 // data may have a Class holder.)
 796 bool ClassLoaderData::is_system_class_loader_data() const {
 797   return SystemDictionary::is_system_class_loader(class_loader());
 798 }
 799 
 800 // Returns true if this class loader data is for the platform class loader.
 801 // (Note that the class loader data may have a Class holder.)
 802 bool ClassLoaderData::is_platform_class_loader_data() const {
 803   return SystemDictionary::is_platform_class_loader(class_loader());
 804 }
 805 
 806 // Returns true if the class loader for this class loader data is one of
 807 // the 3 builtin (boot application/system or platform) class loaders,
 808 // including a user-defined system class loader.  Note that if the class
 809 // loader data is for a non-strong hidden class then it may
 810 // get freed by a GC even if its class loader is one of these loaders.
 811 bool ClassLoaderData::is_builtin_class_loader_data() const {
 812   return (is_boot_class_loader_data() ||
 813           SystemDictionary::is_system_class_loader(class_loader()) ||
 814           SystemDictionary::is_platform_class_loader(class_loader()));
 815 }
 816 
 817 // Returns true if this class loader data is a class loader data
 818 // that is not ever freed by a GC.  It must be the CLD for one of the builtin
 819 // class loaders and not the CLD for a non-strong hidden class.
 820 bool ClassLoaderData::is_permanent_class_loader_data() const {
 821   return is_builtin_class_loader_data() && !has_class_mirror_holder();
 822 }
 823 
 824 ClassLoaderMetaspace* ClassLoaderData::metaspace_non_null() {
 825   // If the metaspace has not been allocated, create a new one.  Might want
 826   // to create smaller arena for Reflection class loaders also.
 827   // The reason for the delayed allocation is because some class loaders are
 828   // simply for delegating with no metadata of their own.
 829   // Lock-free access requires load_acquire.
 830   ClassLoaderMetaspace* metaspace = Atomic::load_acquire(&_metaspace);
 831   if (metaspace == nullptr) {
 832     MutexLocker ml(_metaspace_lock,  Mutex::_no_safepoint_check_flag);
 833     // Check if _metaspace got allocated while we were waiting for this lock.
 834     if ((metaspace = _metaspace) == nullptr) {
 835       if (this == the_null_class_loader_data()) {
 836         assert (class_loader() == nullptr, "Must be");
 837         metaspace = new ClassLoaderMetaspace(_metaspace_lock, Metaspace::BootMetaspaceType);
 838       } else if (has_class_mirror_holder()) {
 839         metaspace = new ClassLoaderMetaspace(_metaspace_lock, Metaspace::ClassMirrorHolderMetaspaceType);
 840       } else if (class_loader()->is_a(vmClasses::reflect_DelegatingClassLoader_klass())) {
 841         metaspace = new ClassLoaderMetaspace(_metaspace_lock, Metaspace::ReflectionMetaspaceType);
 842       } else {
 843         metaspace = new ClassLoaderMetaspace(_metaspace_lock, Metaspace::StandardMetaspaceType);
 844       }
 845       // Ensure _metaspace is stable, since it is examined without a lock
 846       Atomic::release_store(&_metaspace, metaspace);
 847     }
 848   }
 849   return metaspace;
 850 }
 851 
 852 OopHandle ClassLoaderData::add_handle(Handle h) {
 853   MutexLocker ml(metaspace_lock(),  Mutex::_no_safepoint_check_flag);
 854   record_modified_oops();
 855   return _handles.add(h());
 856 }
 857 
 858 void ClassLoaderData::remove_handle(OopHandle h) {
 859   assert(!is_unloading(), "Do not remove a handle for a CLD that is unloading");
 860   oop* ptr = h.ptr_raw();
 861   if (ptr != nullptr) {
 862     assert(_handles.owner_of(ptr), "Got unexpected handle " PTR_FORMAT, p2i(ptr));
 863     NativeAccess<>::oop_store(ptr, oop(nullptr));
 864   }
 865 }
 866 
 867 void ClassLoaderData::init_handle_locked(OopHandle& dest, Handle h) {
 868   MutexLocker ml(metaspace_lock(),  Mutex::_no_safepoint_check_flag);
 869   if (dest.resolve() != nullptr) {
 870     return;
 871   } else {
 872     record_modified_oops();
 873     dest = _handles.add(h());
 874   }
 875 }
 876 
 877 // Add this metadata pointer to be freed when it's safe.  This is only during
 878 // a safepoint which checks if handles point to this metadata field.
 879 void ClassLoaderData::add_to_deallocate_list(Metadata* m) {
 880   // Metadata in shared region isn't deleted.
 881   if (!m->is_shared()) {
 882     MutexLocker ml(metaspace_lock(),  Mutex::_no_safepoint_check_flag);
 883     if (_deallocate_list == nullptr) {
 884       _deallocate_list = new (mtClass) GrowableArray<Metadata*>(100, mtClass);
 885     }
 886     _deallocate_list->append_if_missing(m);
 887     ResourceMark rm;
 888     log_debug(class, loader, data)("deallocate added for %s", m->print_value_string());
 889     ClassLoaderDataGraph::set_should_clean_deallocate_lists();
 890   }
 891 }
 892 
 893 // Deallocate free metadata on the free list.  How useful the PermGen was!
 894 void ClassLoaderData::free_deallocate_list() {
 895   // This must be called at a safepoint because it depends on metadata walking at
 896   // safepoint cleanup time.
 897   assert(SafepointSynchronize::is_at_safepoint(), "only called at safepoint");
 898   assert(!is_unloading(), "only called for ClassLoaderData that are not unloading");
 899   if (_deallocate_list == nullptr) {
 900     return;
 901   }
 902   // Go backwards because this removes entries that are freed.
 903   for (int i = _deallocate_list->length() - 1; i >= 0; i--) {
 904     Metadata* m = _deallocate_list->at(i);
 905     if (!m->on_stack()) {
 906       _deallocate_list->remove_at(i);
 907       // There are only three types of metadata that we deallocate directly.
 908       // Cast them so they can be used by the template function.
 909       if (m->is_method()) {
 910         MetadataFactory::free_metadata(this, (Method*)m);
 911       } else if (m->is_constantPool()) {
 912         MetadataFactory::free_metadata(this, (ConstantPool*)m);
 913       } else if (m->is_klass()) {
 914         if (!((Klass*)m)->is_inline_klass()) {
 915           MetadataFactory::free_metadata(this, (InstanceKlass*)m);
 916         } else {
 917           MetadataFactory::free_metadata(this, (InlineKlass*)m);
 918         }
 919       } else {
 920         ShouldNotReachHere();
 921       }
 922     } else {
 923       // Metadata is alive.
 924       // If scratch_class is on stack then it shouldn't be on this list!
 925       assert(!m->is_klass() || !((InstanceKlass*)m)->is_scratch_class(),
 926              "scratch classes on this list should be dead");
 927       // Also should assert that other metadata on the list was found in handles.
 928       // Some cleaning remains.
 929       ClassLoaderDataGraph::set_should_clean_deallocate_lists();
 930     }
 931   }
 932 }
 933 
 934 // This is distinct from free_deallocate_list.  For class loader data that are
 935 // unloading, this frees the C heap memory for items on the list, and unlinks
 936 // scratch or error classes so that unloading events aren't triggered for these
 937 // classes. The metadata is removed with the unloading metaspace.
 938 // There isn't C heap memory allocated for methods, so nothing is done for them.
 939 void ClassLoaderData::free_deallocate_list_C_heap_structures() {
 940   assert_locked_or_safepoint(ClassLoaderDataGraph_lock);
 941   assert(is_unloading(), "only called for ClassLoaderData that are unloading");
 942   if (_deallocate_list == nullptr) {
 943     return;
 944   }
 945   // Go backwards because this removes entries that are freed.
 946   for (int i = _deallocate_list->length() - 1; i >= 0; i--) {
 947     Metadata* m = _deallocate_list->at(i);
 948     _deallocate_list->remove_at(i);
 949     if (m->is_constantPool()) {
 950       ((ConstantPool*)m)->release_C_heap_structures();
 951     } else if (m->is_klass()) {
 952       InstanceKlass* ik = (InstanceKlass*)m;
 953       // also releases ik->constants() C heap memory
 954       ik->release_C_heap_structures();
 955       // Remove the class so unloading events aren't triggered for
 956       // this class (scratch or error class) in do_unloading().
 957       remove_class(ik);
 958       // But still have to remove it from the dumptime_table.
 959       SystemDictionaryShared::handle_class_unloading(ik);
 960     }
 961   }
 962 }
 963 
 964 // Caller needs ResourceMark
 965 // If the class loader's _name has not been explicitly set, the class loader's
 966 // qualified class name is returned.
 967 const char* ClassLoaderData::loader_name() const {
 968    if (_class_loader_klass == nullptr) {
 969      return BOOTSTRAP_LOADER_NAME;
 970    } else if (_name != nullptr) {
 971      return _name->as_C_string();
 972    } else {
 973      return _class_loader_klass->external_name();
 974    }
 975 }
 976 
 977 // Caller needs ResourceMark
 978 // Format of the _name_and_id is as follows:
 979 //   If the defining loader has a name explicitly set then '<loader-name>' @<id>
 980 //   If the defining loader has no name then <qualified-class-name> @<id>
 981 //   If built-in loader, then omit '@<id>' as there is only one instance.
 982 const char* ClassLoaderData::loader_name_and_id() const {
 983   if (_class_loader_klass == nullptr) {
 984     return "'" BOOTSTRAP_LOADER_NAME "'";
 985   } else if (_name_and_id != nullptr) {
 986     return _name_and_id->as_C_string();
 987   } else {
 988     // May be called in a race before _name_and_id is initialized.
 989     return _class_loader_klass->external_name();
 990   }
 991 }
 992 
 993 void ClassLoaderData::print_value_on(outputStream* out) const {
 994   if (!is_unloading() && class_loader() != nullptr) {
 995     out->print("loader data: " INTPTR_FORMAT " for instance ", p2i(this));
 996     class_loader()->print_value_on(out);  // includes loader_name_and_id() and address of class loader instance
 997   } else {
 998     // loader data: 0xsomeaddr of 'bootstrap'
 999     out->print("loader data: " INTPTR_FORMAT " of %s", p2i(this), loader_name_and_id());
1000   }
1001   if (_has_class_mirror_holder) {
1002     out->print(" has a class holder");
1003   }
1004 }
1005 
1006 void ClassLoaderData::print_value() const { print_value_on(tty); }
1007 
1008 #ifndef PRODUCT
1009 class PrintKlassClosure: public KlassClosure {
1010   outputStream* _out;
1011 public:
1012   PrintKlassClosure(outputStream* out): _out(out) { }
1013 
1014   void do_klass(Klass* k) {
1015     ResourceMark rm;
1016     _out->print("%s,", k->external_name());
1017   }
1018 };
1019 
1020 void ClassLoaderData::print_on(outputStream* out) const {
1021   ResourceMark rm;
1022   out->print_cr("ClassLoaderData(" INTPTR_FORMAT ")", p2i(this));
1023   out->print_cr(" - name                %s", loader_name_and_id());
1024   if (!_holder.is_null()) {
1025     out->print   (" - holder              ");
1026     _holder.print_on(out);
1027     out->print_cr("");
1028   }
1029   out->print_cr(" - class loader        " INTPTR_FORMAT, p2i(_class_loader.ptr_raw()));
1030   out->print_cr(" - metaspace           " INTPTR_FORMAT, p2i(_metaspace));
1031   out->print_cr(" - unloading           %s", _unloading ? "true" : "false");
1032   out->print_cr(" - class mirror holder %s", _has_class_mirror_holder ? "true" : "false");
1033   out->print_cr(" - modified oops       %s", _modified_oops ? "true" : "false");
1034   out->print_cr(" - keep alive          %d", _keep_alive);
1035   out->print   (" - claim               ");
1036   switch(_claim) {
1037     case _claim_none:                       out->print_cr("none"); break;
1038     case _claim_finalizable:                out->print_cr("finalizable"); break;
1039     case _claim_strong:                     out->print_cr("strong"); break;
1040     case _claim_stw_fullgc_mark:            out->print_cr("stw full gc mark"); break;
1041     case _claim_stw_fullgc_adjust:          out->print_cr("stw full gc adjust"); break;
1042     case _claim_other:                      out->print_cr("other"); break;
1043     case _claim_other | _claim_finalizable: out->print_cr("other and finalizable"); break;
1044     case _claim_other | _claim_strong:      out->print_cr("other and strong"); break;
1045     default:                                ShouldNotReachHere();
1046   }
1047   out->print_cr(" - handles             %d", _handles.count());
1048   out->print_cr(" - dependency count    %d", _dependency_count);
1049   out->print   (" - klasses             { ");
1050   if (Verbose) {
1051     PrintKlassClosure closure(out);
1052     ((ClassLoaderData*)this)->classes_do(&closure);
1053   } else {
1054      out->print("...");
1055   }
1056   out->print_cr(" }");
1057   out->print_cr(" - packages            " INTPTR_FORMAT, p2i(_packages));
1058   out->print_cr(" - module              " INTPTR_FORMAT, p2i(_modules));
1059   out->print_cr(" - unnamed module      " INTPTR_FORMAT, p2i(_unnamed_module));
1060   if (_dictionary != nullptr) {
1061     out->print   (" - dictionary          " INTPTR_FORMAT " ", p2i(_dictionary));
1062     _dictionary->print_size(out);
1063   } else {
1064     out->print_cr(" - dictionary          " INTPTR_FORMAT, p2i(_dictionary));
1065   }
1066   if (_jmethod_ids != nullptr) {
1067     out->print   (" - jmethod count       ");
1068     Method::print_jmethod_ids_count(this, out);
1069     out->print_cr("");
1070   }
1071   out->print_cr(" - deallocate list     " INTPTR_FORMAT, p2i(_deallocate_list));
1072   out->print_cr(" - next CLD            " INTPTR_FORMAT, p2i(_next));
1073 }
1074 #endif // PRODUCT
1075 
1076 void ClassLoaderData::print() const { print_on(tty); }
1077 
1078 class VerifyHandleOops : public OopClosure {
1079   VerifyOopClosure vc;
1080  public:
1081   virtual void do_oop(oop* p) {
1082     if (p != nullptr && *p != nullptr) {
1083       oop o = *p;
1084       if (!java_lang_Class::is_instance(o)) {
1085         // is_instance will assert for an invalid oop.
1086         // Walk the resolved_references array and other assorted oops in the
1087         // CLD::_handles field.  The mirror oops are followed by other heap roots.
1088         o->oop_iterate(&vc);
1089       }
1090     }
1091   }
1092   virtual void do_oop(narrowOop* o) { ShouldNotReachHere(); }
1093 };
1094 
1095 void ClassLoaderData::verify() {
1096   assert_locked_or_safepoint(_metaspace_lock);
1097   oop cl = class_loader();
1098 
1099   guarantee(this == class_loader_data(cl) || has_class_mirror_holder(), "Must be the same");
1100   guarantee(cl != nullptr || this == ClassLoaderData::the_null_class_loader_data() || has_class_mirror_holder(), "must be");
1101 
1102   // Verify the integrity of the allocated space.
1103 #ifdef ASSERT
1104   if (metaspace_or_null() != nullptr) {
1105     metaspace_or_null()->verify();
1106   }
1107 #endif
1108 
1109   for (Klass* k = _klasses; k != nullptr; k = k->next_link()) {
1110     guarantee(k->class_loader_data() == this, "Must be the same");
1111     k->verify();
1112     assert(k != k->next_link(), "no loops!");
1113   }
1114 
1115   if (_modules != nullptr) {
1116     _modules->verify();
1117   }
1118 
1119   if (_deallocate_list != nullptr) {
1120     for (int i = _deallocate_list->length() - 1; i >= 0; i--) {
1121       Metadata* m = _deallocate_list->at(i);
1122       if (m->is_klass()) {
1123         ((Klass*)m)->verify();
1124       }
1125     }
1126   }
1127 
1128   // Check the oops in the handles area
1129   VerifyHandleOops vho;
1130   oops_do(&vho, _claim_none, false);
1131 }
1132 
1133 bool ClassLoaderData::contains_klass(Klass* klass) {
1134   // Lock-free access requires load_acquire
1135   for (Klass* k = Atomic::load_acquire(&_klasses); k != nullptr; k = k->next_link()) {
1136     if (k == klass) return true;
1137   }
1138   return false;
1139 }