1 /* 2 * Copyright (c) 1998, 2023, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "precompiled.hpp" 26 #include "cds/archiveBuilder.hpp" 27 #include "cds/cdsConfig.hpp" 28 #include "cds/heapShared.hpp" 29 #include "classfile/resolutionErrors.hpp" 30 #include "classfile/systemDictionary.hpp" 31 #include "classfile/systemDictionaryShared.hpp" 32 #include "classfile/vmClasses.hpp" 33 #include "code/codeCache.hpp" 34 #include "interpreter/bytecodeStream.hpp" 35 #include "interpreter/bytecodes.hpp" 36 #include "interpreter/interpreter.hpp" 37 #include "interpreter/linkResolver.hpp" 38 #include "interpreter/rewriter.hpp" 39 #include "logging/log.hpp" 40 #include "logging/logStream.hpp" 41 #include "memory/metadataFactory.hpp" 42 #include "memory/metaspaceClosure.hpp" 43 #include "memory/resourceArea.hpp" 44 #include "oops/access.inline.hpp" 45 #include "oops/compressedOops.hpp" 46 #include "oops/constantPool.inline.hpp" 47 #include "oops/cpCache.inline.hpp" 48 #include "oops/objArrayOop.inline.hpp" 49 #include "oops/oop.inline.hpp" 50 #include "oops/resolvedFieldEntry.hpp" 51 #include "oops/resolvedIndyEntry.hpp" 52 #include "oops/resolvedMethodEntry.hpp" 53 #include "prims/methodHandles.hpp" 54 #include "runtime/arguments.hpp" 55 #include "runtime/atomic.hpp" 56 #include "runtime/handles.inline.hpp" 57 #include "runtime/mutexLocker.hpp" 58 #include "runtime/vm_version.hpp" 59 #include "utilities/macros.hpp" 60 61 // Implementation of ConstantPoolCache 62 63 template <class T> 64 static Array<T>* initialize_resolved_entries_array(ClassLoaderData* loader_data, GrowableArray<T> entries, TRAPS) { 65 Array<T>* resolved_entries; 66 if (entries.length() != 0) { 67 resolved_entries = MetadataFactory::new_array<T>(loader_data, entries.length(), CHECK_NULL); 68 for (int i = 0; i < entries.length(); i++) { 69 resolved_entries->at_put(i, entries.at(i)); 70 } 71 return resolved_entries; 72 } 73 return nullptr; 74 } 75 76 void ConstantPoolCache::set_direct_or_vtable_call(Bytecodes::Code invoke_code, 77 int method_index, 78 const methodHandle& method, 79 int vtable_index, 80 bool sender_is_interface) { 81 bool is_vtable_call = (vtable_index >= 0); // FIXME: split this method on this boolean 82 assert(method->interpreter_entry() != nullptr, "should have been set at this point"); 83 assert(!method->is_obsolete(), "attempt to write obsolete method to cpCache"); 84 85 int byte_no = -1; 86 bool change_to_virtual = false; 87 InstanceKlass* holder = nullptr; // have to declare this outside the switch 88 ResolvedMethodEntry* method_entry = resolved_method_entry_at(method_index); 89 switch (invoke_code) { 90 case Bytecodes::_invokeinterface: 91 holder = method->method_holder(); 92 // check for private interface method invocations 93 if (vtable_index == Method::nonvirtual_vtable_index && holder->is_interface() ) { 94 assert(method->is_private(), "unexpected non-private method"); 95 assert(method->can_be_statically_bound(), "unexpected non-statically-bound method"); 96 97 method_entry->set_flags(( 1 << ResolvedMethodEntry::is_vfinal_shift) | 98 ((method->is_final_method() ? 1 : 0) << ResolvedMethodEntry::is_final_shift)); 99 method_entry->fill_in((u1)as_TosState(method->result_type()), (u2)method()->size_of_parameters()); 100 assert(method_entry->is_vfinal(), "flags must be set"); 101 method_entry->set_method(method()); 102 byte_no = 2; 103 method_entry->set_klass(holder); 104 break; 105 } 106 else { 107 // We get here from InterpreterRuntime::resolve_invoke when an invokeinterface 108 // instruction links to a non-interface method (in Object). This can happen when 109 // an interface redeclares an Object method (like CharSequence declaring toString()) 110 // or when invokeinterface is used explicitly. 111 // In that case, the method has no itable index and must be invoked as a virtual. 112 // Set a flag to keep track of this corner case. 113 assert(holder->is_interface() || holder == vmClasses::Object_klass(), "unexpected holder class"); 114 assert(method->is_public(), "Calling non-public method in Object with invokeinterface"); 115 change_to_virtual = true; 116 117 // ...and fall through as if we were handling invokevirtual: 118 } 119 case Bytecodes::_invokevirtual: 120 { 121 if (!is_vtable_call) { 122 assert(method->can_be_statically_bound(), ""); 123 method_entry->set_flags(( 1 << ResolvedMethodEntry::is_vfinal_shift) | 124 ((method->is_final_method() ? 1 : 0) << ResolvedMethodEntry::is_final_shift) | 125 ((change_to_virtual ? 1 : 0) << ResolvedMethodEntry::is_forced_virtual_shift)); 126 method_entry->fill_in((u1)as_TosState(method->result_type()), (u2)method()->size_of_parameters()); 127 assert(method_entry->is_vfinal(), "flags must be set"); 128 method_entry->set_method(method()); 129 } else { 130 assert(!method->can_be_statically_bound(), ""); 131 assert(vtable_index >= 0, "valid index"); 132 assert(!method->is_final_method(), "sanity"); 133 method_entry->set_flags((change_to_virtual ? 1 : 0) << ResolvedMethodEntry::is_forced_virtual_shift); 134 method_entry->fill_in((u1)as_TosState(method->result_type()), (u2)method()->size_of_parameters()); 135 assert(!method_entry->is_vfinal(), "flags must not be set"); 136 method_entry->set_table_index(vtable_index); 137 } 138 byte_no = 2; 139 break; 140 } 141 142 case Bytecodes::_invokespecial: 143 case Bytecodes::_invokestatic: { 144 assert(!is_vtable_call, ""); 145 // Note: Read and preserve the value of the is_vfinal flag on any 146 // invokevirtual bytecode shared with this constant pool cache entry. 147 // It is cheap and safe to consult is_vfinal() at all times. 148 // Once is_vfinal is set, it must stay that way, lest we get a dangling oop. 149 bool vfinal = method_entry->is_vfinal(); 150 method_entry->set_flags(((method->is_final_method() ? 1 : 0) << ResolvedMethodEntry::is_final_shift)); 151 assert(vfinal == method_entry->is_vfinal(), "Vfinal flag must be preserved"); 152 method_entry->fill_in((u1)as_TosState(method->result_type()), (u2)method()->size_of_parameters()); 153 method_entry->set_method(method()); 154 byte_no = 1; 155 break; 156 } 157 default: 158 ShouldNotReachHere(); 159 break; 160 } 161 162 // Note: byte_no also appears in TemplateTable::resolve. 163 if (byte_no == 1) { 164 assert(invoke_code != Bytecodes::_invokevirtual && 165 invoke_code != Bytecodes::_invokeinterface, ""); 166 bool do_resolve = true; 167 // Don't mark invokespecial to method as resolved if sender is an interface. The receiver 168 // has to be checked that it is a subclass of the current class every time this bytecode 169 // is executed. 170 if (invoke_code == Bytecodes::_invokespecial && sender_is_interface && 171 method->name() != vmSymbols::object_initializer_name()) { 172 do_resolve = false; 173 } 174 if (invoke_code == Bytecodes::_invokestatic) { 175 assert(method->method_holder()->is_initialized() || 176 method->method_holder()->is_init_thread(JavaThread::current()), 177 "invalid class initialization state for invoke_static"); 178 179 if (!VM_Version::supports_fast_class_init_checks() && method->needs_clinit_barrier()) { 180 // Don't mark invokestatic to method as resolved if the holder class has not yet completed 181 // initialization. An invokestatic must only proceed if the class is initialized, but if 182 // we resolve it before then that class initialization check is skipped. 183 // 184 // When fast class initialization checks are supported (VM_Version::supports_fast_class_init_checks() == true), 185 // template interpreter supports fast class initialization check for 186 // invokestatic which doesn't require call site re-resolution to 187 // enforce class initialization barrier. 188 do_resolve = false; 189 } 190 } 191 if (do_resolve) { 192 method_entry->set_bytecode1(invoke_code); 193 } 194 } else if (byte_no == 2) { 195 if (change_to_virtual) { 196 assert(invoke_code == Bytecodes::_invokeinterface, ""); 197 // NOTE: THIS IS A HACK - BE VERY CAREFUL!!! 198 // 199 // Workaround for the case where we encounter an invokeinterface, but we 200 // should really have an _invokevirtual since the resolved method is a 201 // virtual method in java.lang.Object. This is a corner case in the spec 202 // but is presumably legal. javac does not generate this code. 203 // 204 // We do not set bytecode_1() to _invokeinterface, because that is the 205 // bytecode # used by the interpreter to see if it is resolved. In this 206 // case, the method gets reresolved with caller for each interface call 207 // because the actual selected method may not be public. 208 // 209 // We set bytecode_2() to _invokevirtual. 210 // See also interpreterRuntime.cpp. (8/25/2000) 211 } else { 212 assert(invoke_code == Bytecodes::_invokevirtual || 213 (invoke_code == Bytecodes::_invokeinterface && 214 ((method->is_private() || 215 (method->is_final() && method->method_holder() == vmClasses::Object_klass())))), 216 "unexpected invocation mode"); 217 if (invoke_code == Bytecodes::_invokeinterface && 218 (method->is_private() || method->is_final())) { 219 // We set bytecode_1() to _invokeinterface, because that is the 220 // bytecode # used by the interpreter to see if it is resolved. 221 // We set bytecode_2() to _invokevirtual. 222 method_entry->set_bytecode1(invoke_code); 223 } 224 } 225 // set up for invokevirtual, even if linking for invokeinterface also: 226 method_entry->set_bytecode2(Bytecodes::_invokevirtual); 227 } else { 228 ShouldNotReachHere(); 229 } 230 } 231 232 void ConstantPoolCache::set_direct_call(Bytecodes::Code invoke_code, int method_index, const methodHandle& method, 233 bool sender_is_interface) { 234 int index = Method::nonvirtual_vtable_index; 235 // index < 0; FIXME: inline and customize set_direct_or_vtable_call 236 set_direct_or_vtable_call(invoke_code, method_index, method, index, sender_is_interface); 237 } 238 239 void ConstantPoolCache::set_vtable_call(Bytecodes::Code invoke_code, int method_index, const methodHandle& method, int index) { 240 // either the method is a miranda or its holder should accept the given index 241 assert(method->method_holder()->is_interface() || method->method_holder()->verify_vtable_index(index), ""); 242 // index >= 0; FIXME: inline and customize set_direct_or_vtable_call 243 set_direct_or_vtable_call(invoke_code, method_index, method, index, false); 244 } 245 246 void ConstantPoolCache::set_itable_call(Bytecodes::Code invoke_code, 247 int method_index, 248 Klass* referenced_klass, 249 const methodHandle& method, int index) { 250 assert(method->method_holder()->verify_itable_index(index), ""); 251 assert(invoke_code == Bytecodes::_invokeinterface, ""); 252 InstanceKlass* interf = method->method_holder(); 253 assert(interf->is_interface(), "must be an interface"); 254 assert(!method->is_final_method(), "interfaces do not have final methods; cannot link to one here"); 255 ResolvedMethodEntry* method_entry = resolved_method_entry_at(method_index); 256 method_entry->set_klass(static_cast<InstanceKlass*>(referenced_klass)); 257 method_entry->set_method(method()); 258 method_entry->fill_in((u1)as_TosState(method->result_type()), (u2)method()->size_of_parameters()); 259 method_entry->set_bytecode1(Bytecodes::_invokeinterface); 260 } 261 262 ResolvedMethodEntry* ConstantPoolCache::set_method_handle(int method_index, const CallInfo &call_info) { 263 // NOTE: This method entry can be the subject of data races. 264 // There are three words to update: flags, refs[appendix_index], method (in that order). 265 // Writers must store all other values before method. 266 // Readers must test the method first for non-null before reading other fields. 267 // Competing writers must acquire exclusive access via a lock. 268 // A losing writer waits on the lock until the winner writes the method and leaves 269 // the lock, so that when the losing writer returns, he can use the linked 270 // cache entry. 271 // Lock fields to write 272 Bytecodes::Code invoke_code = Bytecodes::_invokehandle; 273 MutexLocker ml(constant_pool()->pool_holder()->init_monitor()); 274 ResolvedMethodEntry* method_entry = resolved_method_entry_at(method_index); 275 276 if (method_entry->is_resolved(invoke_code)) { 277 return method_entry; 278 } 279 280 Method* adapter = call_info.resolved_method(); 281 const Handle appendix = call_info.resolved_appendix(); 282 const bool has_appendix = appendix.not_null(); 283 284 // Write the flags. 285 // MHs are always sig-poly and have a local signature. 286 method_entry->fill_in((u1)as_TosState(adapter->result_type()), (u2)adapter->size_of_parameters()); 287 method_entry->set_flags(((has_appendix ? 1 : 0) << ResolvedMethodEntry::has_appendix_shift ) | 288 ( 1 << ResolvedMethodEntry::has_local_signature_shift ) | 289 ( 1 << ResolvedMethodEntry::is_final_shift )); 290 291 // Method handle invokes use both a method and a resolved references index. 292 // refs[appendix_index], if not null, contains a value passed as a trailing argument to the adapter. 293 // In the general case, this could be the call site's MethodType, 294 // for use with java.lang.Invokers.checkExactType, or else a CallSite object. 295 // method_entry->method() contains the adapter method which manages the actual call. 296 // In the general case, this is a compiled LambdaForm. 297 // (The Java code is free to optimize these calls by binding other 298 // sorts of methods and appendices to call sites.) 299 // JVM-level linking is via the method, as if for invokespecial, and signatures are erased. 300 // The appendix argument (if any) is added to the signature, and is counted in the parameter_size bits. 301 // Even with the appendix, the method will never take more than 255 parameter slots. 302 // 303 // This means that given a call site like (List)mh.invoke("foo"), 304 // the method has signature '(Ljl/Object;Ljl/invoke/MethodType;)Ljl/Object;', 305 // not '(Ljava/lang/String;)Ljava/util/List;'. 306 // The fact that String and List are involved is encoded in the MethodType in refs[appendix_index]. 307 // This allows us to create fewer Methods, while keeping type safety. 308 // 309 310 // Store appendix, if any. 311 if (has_appendix) { 312 const int appendix_index = method_entry->resolved_references_index(); 313 objArrayOop resolved_references = constant_pool()->resolved_references(); 314 assert(appendix_index >= 0 && appendix_index < resolved_references->length(), "oob"); 315 assert(resolved_references->obj_at(appendix_index) == nullptr, "init just once"); 316 resolved_references->obj_at_put(appendix_index, appendix()); 317 } 318 319 method_entry->set_method(adapter); // This must be the last one to set (see NOTE above)! 320 321 // The interpreter assembly code does not check byte_2, 322 // but it is used by is_resolved, method_if_resolved, etc. 323 method_entry->set_bytecode1(invoke_code); 324 325 assert(has_appendix == method_entry->has_appendix(), "proper storage of appendix flag"); 326 assert(method_entry->has_local_signature(), "proper storage of signature flag"); 327 return method_entry; 328 } 329 330 Method* ConstantPoolCache::method_if_resolved(int method_index) const { 331 // Decode the action of set_method and set_interface_call 332 ResolvedMethodEntry* method_entry = resolved_method_entry_at(method_index); 333 334 Bytecodes::Code invoke_code = (Bytecodes::Code)method_entry->bytecode1(); 335 switch (invoke_code) { 336 case Bytecodes::_invokeinterface: 337 case Bytecodes::_invokestatic: 338 case Bytecodes::_invokespecial: 339 assert(!method_entry->has_appendix(), ""); 340 // fall through 341 case Bytecodes::_invokehandle: 342 return method_entry->method(); 343 case Bytecodes::_invokedynamic: 344 ShouldNotReachHere(); 345 default: 346 assert(invoke_code == (Bytecodes::Code)0, "unexpected bytecode"); 347 break; 348 } 349 350 invoke_code = (Bytecodes::Code)method_entry->bytecode2(); 351 if (invoke_code == Bytecodes::_invokevirtual) { 352 if (method_entry->is_vfinal()) { 353 return method_entry->method(); 354 } else { 355 int holder_index = constant_pool()->uncached_klass_ref_index_at(method_entry->constant_pool_index()); 356 if (constant_pool()->tag_at(holder_index).is_klass()) { 357 Klass* klass = constant_pool()->resolved_klass_at(holder_index); 358 return klass->method_at_vtable(method_entry->table_index()); 359 } 360 } 361 } 362 return nullptr; 363 } 364 365 ConstantPoolCache* ConstantPoolCache::allocate(ClassLoaderData* loader_data, 366 const intStack& invokedynamic_map, 367 const GrowableArray<ResolvedIndyEntry> indy_entries, 368 const GrowableArray<ResolvedFieldEntry> field_entries, 369 const GrowableArray<ResolvedMethodEntry> method_entries, 370 TRAPS) { 371 372 int size = ConstantPoolCache::size(); 373 374 // Initialize resolved entry arrays with available data 375 Array<ResolvedFieldEntry>* resolved_field_entries = initialize_resolved_entries_array(loader_data, field_entries, CHECK_NULL); 376 Array<ResolvedIndyEntry>* resolved_indy_entries = initialize_resolved_entries_array(loader_data, indy_entries, CHECK_NULL); 377 Array<ResolvedMethodEntry>* resolved_method_entries = initialize_resolved_entries_array(loader_data, method_entries, CHECK_NULL); 378 379 return new (loader_data, size, MetaspaceObj::ConstantPoolCacheType, THREAD) 380 ConstantPoolCache(invokedynamic_map, resolved_indy_entries, resolved_field_entries, resolved_method_entries); 381 } 382 383 // Record the GC marking cycle when redefined vs. when found in the loom stack chunks. 384 void ConstantPoolCache::record_gc_epoch() { 385 _gc_epoch = CodeCache::gc_epoch(); 386 } 387 388 #if INCLUDE_CDS 389 void ConstantPoolCache::remove_unshareable_info() { 390 assert(CDSConfig::is_dumping_archive(), "sanity"); 391 // <this> is the copy to be written into the archive. It's in the ArchiveBuilder's "buffer space". 392 // However, this->_initial_entries was not copied/relocated by the ArchiveBuilder, so it's 393 // still pointing to the array allocated inside save_for_archive(). 394 if (_resolved_indy_entries != nullptr) { 395 for (int i = 0; i < _resolved_indy_entries->length(); i++) { 396 resolved_indy_entry_at(i)->remove_unshareable_info(); 397 } 398 } 399 if (_resolved_field_entries != nullptr) { 400 for (int i = 0; i < _resolved_field_entries->length(); i++) { 401 resolved_field_entry_at(i)->remove_unshareable_info(); 402 } 403 } 404 if (_resolved_method_entries != nullptr) { 405 for (int i = 0; i < _resolved_method_entries->length(); i++) { 406 resolved_method_entry_at(i)->remove_unshareable_info(); 407 } 408 } 409 } 410 #endif // INCLUDE_CDS 411 412 void ConstantPoolCache::deallocate_contents(ClassLoaderData* data) { 413 assert(!is_shared(), "shared caches are not deallocated"); 414 data->remove_handle(_resolved_references); 415 set_resolved_references(OopHandle()); 416 MetadataFactory::free_array<u2>(data, _reference_map); 417 set_reference_map(nullptr); 418 #if INCLUDE_CDS 419 if (_resolved_indy_entries != nullptr) { 420 MetadataFactory::free_array<ResolvedIndyEntry>(data, _resolved_indy_entries); 421 _resolved_indy_entries = nullptr; 422 } 423 if (_resolved_field_entries != nullptr) { 424 MetadataFactory::free_array<ResolvedFieldEntry>(data, _resolved_field_entries); 425 _resolved_field_entries = nullptr; 426 } 427 if (_resolved_method_entries != nullptr) { 428 MetadataFactory::free_array<ResolvedMethodEntry>(data, _resolved_method_entries); 429 _resolved_method_entries = nullptr; 430 } 431 #endif 432 } 433 434 #if INCLUDE_CDS_JAVA_HEAP 435 oop ConstantPoolCache::archived_references() { 436 if (_archived_references_index < 0) { 437 return nullptr; 438 } 439 return HeapShared::get_root(_archived_references_index); 440 } 441 442 void ConstantPoolCache::clear_archived_references() { 443 if (_archived_references_index >= 0) { 444 HeapShared::clear_root(_archived_references_index); 445 _archived_references_index = -1; 446 } 447 } 448 449 void ConstantPoolCache::set_archived_references(int root_index) { 450 assert(CDSConfig::is_dumping_heap(), "sanity"); 451 _archived_references_index = root_index; 452 } 453 #endif 454 455 #if INCLUDE_JVMTI 456 void log_adjust(const char* entry_type, Method* old_method, Method* new_method, bool* trace_name_printed) { 457 ResourceMark rm; 458 459 if (!(*trace_name_printed)) { 460 log_info(redefine, class, update)("adjust: name=%s", old_method->method_holder()->external_name()); 461 *trace_name_printed = true; 462 } 463 log_trace(redefine, class, update, constantpool) 464 ("cpc %s entry update: %s", entry_type, new_method->external_name()); 465 } 466 467 // RedefineClasses() API support: 468 // If any entry of this ConstantPoolCache points to any of 469 // old_methods, replace it with the corresponding new_method. 470 void ConstantPoolCache::adjust_method_entries(bool * trace_name_printed) { 471 if (_resolved_indy_entries != nullptr) { 472 for (int j = 0; j < _resolved_indy_entries->length(); j++) { 473 Method* old_method = resolved_indy_entry_at(j)->method(); 474 if (old_method == nullptr || !old_method->is_old()) { 475 continue; 476 } 477 Method* new_method = old_method->get_new_method(); 478 resolved_indy_entry_at(j)->adjust_method_entry(new_method); 479 log_adjust("indy", old_method, new_method, trace_name_printed); 480 } 481 } 482 if (_resolved_method_entries != nullptr) { 483 for (int i = 0; i < _resolved_method_entries->length(); i++) { 484 ResolvedMethodEntry* method_entry = resolved_method_entry_at(i); 485 // get interesting method entry 486 Method* old_method = method_entry->method(); 487 if (old_method == nullptr || !old_method->is_old()) { 488 continue; // skip uninteresting entries 489 } 490 if (old_method->is_deleted()) { 491 // clean up entries with deleted methods 492 method_entry->reset_entry(); 493 continue; 494 } 495 Method* new_method = old_method->get_new_method(); 496 method_entry->adjust_method_entry(new_method); 497 log_adjust("non-indy", old_method, new_method, trace_name_printed); 498 } 499 } 500 } 501 502 // the constant pool cache should never contain old or obsolete methods 503 bool ConstantPoolCache::check_no_old_or_obsolete_entries() { 504 ResourceMark rm; 505 if (_resolved_indy_entries != nullptr) { 506 for (int i = 0; i < _resolved_indy_entries->length(); i++) { 507 Method* m = resolved_indy_entry_at(i)->method(); 508 if (m != nullptr && !resolved_indy_entry_at(i)->check_no_old_or_obsolete_entry()) { 509 log_trace(redefine, class, update, constantpool) 510 ("cpcache check found old method entry: class: %s, old: %d, obsolete: %d, method: %s", 511 constant_pool()->pool_holder()->external_name(), m->is_old(), m->is_obsolete(), m->external_name()); 512 return false; 513 } 514 } 515 } 516 if (_resolved_method_entries != nullptr) { 517 for (int i = 0; i < _resolved_method_entries->length(); i++) { 518 ResolvedMethodEntry* method_entry = resolved_method_entry_at(i); 519 Method* m = method_entry->method(); 520 if (m != nullptr && !method_entry->check_no_old_or_obsolete_entry()) { 521 log_trace(redefine, class, update, constantpool) 522 ("cpcache check found old method entry: class: %s, old: %d, obsolete: %d, method: %s", 523 constant_pool()->pool_holder()->external_name(), m->is_old(), m->is_obsolete(), m->external_name()); 524 return false; 525 } 526 } 527 } 528 return true; 529 } 530 531 void ConstantPoolCache::dump_cache() { 532 print_on(tty); 533 } 534 #endif // INCLUDE_JVMTI 535 536 void ConstantPoolCache::metaspace_pointers_do(MetaspaceClosure* it) { 537 log_trace(cds)("Iter(ConstantPoolCache): %p", this); 538 it->push(&_constant_pool); 539 it->push(&_reference_map); 540 if (_resolved_indy_entries != nullptr) { 541 it->push(&_resolved_indy_entries, MetaspaceClosure::_writable); 542 } 543 if (_resolved_field_entries != nullptr) { 544 it->push(&_resolved_field_entries, MetaspaceClosure::_writable); 545 } 546 if (_resolved_method_entries != nullptr) { 547 it->push(&_resolved_method_entries, MetaspaceClosure::_writable); 548 } 549 } 550 551 bool ConstantPoolCache::save_and_throw_indy_exc( 552 const constantPoolHandle& cpool, int cpool_index, int index, constantTag tag, TRAPS) { 553 554 assert(HAS_PENDING_EXCEPTION, "No exception got thrown!"); 555 assert(PENDING_EXCEPTION->is_a(vmClasses::LinkageError_klass()), 556 "No LinkageError exception"); 557 558 MutexLocker ml(THREAD, cpool->pool_holder()->init_monitor()); 559 560 // if the indy_info is resolved or the indy_resolution_failed flag is set then another 561 // thread either succeeded in resolving the method or got a LinkageError 562 // exception, before this thread was able to record its failure. So, clear 563 // this thread's exception and return false so caller can use the earlier 564 // thread's result. 565 if (resolved_indy_entry_at(index)->is_resolved() || resolved_indy_entry_at(index)->resolution_failed()) { 566 CLEAR_PENDING_EXCEPTION; 567 return false; 568 } 569 570 Symbol* error = PENDING_EXCEPTION->klass()->name(); 571 Symbol* message = java_lang_Throwable::detail_message(PENDING_EXCEPTION); 572 573 int encoded_index = ResolutionErrorTable::encode_indy_index( 574 ConstantPool::encode_invokedynamic_index(index)); 575 SystemDictionary::add_resolution_error(cpool, encoded_index, error, message); 576 resolved_indy_entry_at(index)->set_resolution_failed(); 577 return true; 578 } 579 580 oop ConstantPoolCache::set_dynamic_call(const CallInfo &call_info, int index) { 581 ResourceMark rm; 582 MutexLocker ml(constant_pool()->pool_holder()->init_monitor()); 583 assert(index >= 0, "Indy index must be positive at this point"); 584 585 if (resolved_indy_entry_at(index)->method() != nullptr) { 586 return constant_pool()->resolved_reference_from_indy(index); 587 } 588 589 if (resolved_indy_entry_at(index)->resolution_failed()) { 590 // Before we got here, another thread got a LinkageError exception during 591 // resolution. Ignore our success and throw their exception. 592 guarantee(index >= 0, "Invalid indy index"); 593 int encoded_index = ResolutionErrorTable::encode_indy_index( 594 ConstantPool::encode_invokedynamic_index(index)); 595 JavaThread* THREAD = JavaThread::current(); // For exception macros. 596 constantPoolHandle cp(THREAD, constant_pool()); 597 ConstantPool::throw_resolution_error(cp, encoded_index, THREAD); 598 return nullptr; 599 } 600 601 Method* adapter = call_info.resolved_method(); 602 const Handle appendix = call_info.resolved_appendix(); 603 const bool has_appendix = appendix.not_null(); 604 605 LogStream* log_stream = nullptr; 606 LogStreamHandle(Debug, methodhandles, indy) lsh_indy; 607 if (lsh_indy.is_enabled()) { 608 ResourceMark rm; 609 log_stream = &lsh_indy; 610 log_stream->print_cr("set_method_handle bc=%d appendix=" PTR_FORMAT "%s method=" PTR_FORMAT " (local signature) ", 611 0xba, 612 p2i(appendix()), 613 (has_appendix ? "" : " (unused)"), 614 p2i(adapter)); 615 adapter->print_on(log_stream); 616 if (has_appendix) appendix()->print_on(log_stream); 617 } 618 619 if (has_appendix) { 620 const int appendix_index = resolved_indy_entry_at(index)->resolved_references_index(); 621 objArrayOop resolved_references = constant_pool()->resolved_references(); 622 assert(appendix_index >= 0 && appendix_index < resolved_references->length(), "oob"); 623 assert(resolved_references->obj_at(appendix_index) == nullptr, "init just once"); 624 resolved_references->obj_at_put(appendix_index, appendix()); 625 } 626 627 // Populate entry with resolved information 628 assert(resolved_indy_entries() != nullptr, "Invokedynamic array is empty, cannot fill with resolved information"); 629 resolved_indy_entry_at(index)->fill_in(adapter, adapter->size_of_parameters(), as_TosState(adapter->result_type()), has_appendix); 630 631 if (log_stream != nullptr) { 632 resolved_indy_entry_at(index)->print_on(log_stream); 633 } 634 return appendix(); 635 } 636 637 oop ConstantPoolCache::appendix_if_resolved(int method_index) const { 638 ResolvedMethodEntry* method_entry = resolved_method_entry_at(method_index); 639 return appendix_if_resolved(method_entry); 640 } 641 642 oop ConstantPoolCache::appendix_if_resolved(ResolvedMethodEntry* method_entry) const { 643 if (!method_entry->has_appendix()) 644 return nullptr; 645 const int ref_index = method_entry->resolved_references_index(); 646 return constant_pool()->resolved_reference_at(ref_index); 647 } 648 649 // Printing 650 651 void ConstantPoolCache::print_on(outputStream* st) const { 652 st->print_cr("%s", internal_name()); 653 // print constant pool cache entries 654 print_resolved_field_entries(st); 655 print_resolved_method_entries(st); 656 print_resolved_indy_entries(st); 657 } 658 659 void ConstantPoolCache::print_resolved_field_entries(outputStream* st) const { 660 for (int field_index = 0; field_index < resolved_field_entries_length(); field_index++) { 661 resolved_field_entry_at(field_index)->print_on(st); 662 } 663 } 664 665 void ConstantPoolCache::print_resolved_method_entries(outputStream* st) const { 666 for (int method_index = 0; method_index < resolved_method_entries_length(); method_index++) { 667 ResolvedMethodEntry* method_entry = resolved_method_entry_at(method_index); 668 method_entry->print_on(st); 669 if (method_entry->has_appendix()) { 670 st->print(" appendix: "); 671 constant_pool()->resolved_reference_from_method(method_index)->print_on(st); 672 } 673 } 674 } 675 676 void ConstantPoolCache::print_resolved_indy_entries(outputStream* st) const { 677 for (int indy_index = 0; indy_index < resolved_indy_entries_length(); indy_index++) { 678 ResolvedIndyEntry* indy_entry = resolved_indy_entry_at(indy_index); 679 indy_entry->print_on(st); 680 if (indy_entry->has_appendix()) { 681 st->print(" appendix: "); 682 constant_pool()->resolved_reference_from_indy(indy_index)->print_on(st); 683 } 684 } 685 }