1 /*
2 * Copyright (c) 1999, 2026, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
24
25 #include "cds/aotLinkedClassBulkLoader.hpp"
26 #include "cds/cdsConfig.hpp"
27 #include "classfile/javaClasses.inline.hpp"
28 #include "classfile/symbolTable.hpp"
29 #include "classfile/vmClasses.hpp"
30 #include "classfile/vmSymbols.hpp"
31 #include "code/aotCodeCache.hpp"
32 #include "code/codeCache.hpp"
33 #include "code/codeHeapState.hpp"
34 #include "code/dependencyContext.hpp"
35 #include "compiler/compilationLog.hpp"
36 #include "compiler/compilationMemoryStatistic.hpp"
37 #include "compiler/compilationPolicy.hpp"
38 #include "compiler/compileBroker.hpp"
39 #include "compiler/compileLog.hpp"
40 #include "compiler/compilerDefinitions.inline.hpp"
41 #include "compiler/compilerEvent.hpp"
42 #include "compiler/compilerOracle.hpp"
43 #include "compiler/directivesParser.hpp"
44 #include "compiler/recompilationPolicy.hpp"
45 #include "gc/shared/memAllocator.hpp"
46 #include "interpreter/linkResolver.hpp"
47 #include "jfr/jfrEvents.hpp"
48 #include "jvm.h"
49 #include "logging/log.hpp"
50 #include "logging/logStream.hpp"
51 #include "memory/allocation.inline.hpp"
52 #include "memory/resourceArea.hpp"
53 #include "memory/universe.hpp"
54 #include "oops/method.inline.hpp"
55 #include "oops/methodData.hpp"
56 #include "oops/oop.inline.hpp"
57 #include "prims/jvmtiExport.hpp"
58 #include "prims/nativeLookup.hpp"
59 #include "prims/whitebox.hpp"
60 #include "runtime/atomicAccess.hpp"
61 #include "runtime/escapeBarrier.hpp"
62 #include "runtime/globals_extension.hpp"
63 #include "runtime/handles.inline.hpp"
64 #include "runtime/init.hpp"
65 #include "runtime/interfaceSupport.inline.hpp"
66 #include "runtime/java.hpp"
67 #include "runtime/javaCalls.hpp"
68 #include "runtime/jniHandles.inline.hpp"
69 #include "runtime/os.hpp"
70 #include "runtime/perfData.hpp"
71 #include "runtime/safepointVerifiers.hpp"
72 #include "runtime/sharedRuntime.hpp"
73 #include "runtime/threads.hpp"
74 #include "runtime/threadSMR.inline.hpp"
75 #include "runtime/timerTrace.hpp"
76 #include "runtime/vframe.inline.hpp"
77 #include "services/management.hpp"
78 #include "utilities/debug.hpp"
79 #include "utilities/dtrace.hpp"
80 #include "utilities/events.hpp"
81 #include "utilities/formatBuffer.hpp"
82 #include "utilities/macros.hpp"
83 #include "utilities/nonblockingQueue.inline.hpp"
84 #ifdef COMPILER1
85 #include "c1/c1_Compiler.hpp"
86 #endif
87 #ifdef COMPILER2
88 #include "opto/c2compiler.hpp"
89 #endif
90 #if INCLUDE_JVMCI
91 #include "jvmci/jvmciEnv.hpp"
92 #include "jvmci/jvmciRuntime.hpp"
93 #endif
94
95 #ifdef DTRACE_ENABLED
96
97 // Only bother with this argument setup if dtrace is available
98
99 #define DTRACE_METHOD_COMPILE_BEGIN_PROBE(method, comp_name) \
100 { \
101 Symbol* klass_name = (method)->klass_name(); \
102 Symbol* name = (method)->name(); \
103 Symbol* signature = (method)->signature(); \
104 HOTSPOT_METHOD_COMPILE_BEGIN( \
105 (char *) comp_name, strlen(comp_name), \
106 (char *) klass_name->bytes(), klass_name->utf8_length(), \
107 (char *) name->bytes(), name->utf8_length(), \
108 (char *) signature->bytes(), signature->utf8_length()); \
109 }
110
111 #define DTRACE_METHOD_COMPILE_END_PROBE(method, comp_name, success) \
112 { \
113 Symbol* klass_name = (method)->klass_name(); \
114 Symbol* name = (method)->name(); \
115 Symbol* signature = (method)->signature(); \
116 HOTSPOT_METHOD_COMPILE_END( \
117 (char *) comp_name, strlen(comp_name), \
118 (char *) klass_name->bytes(), klass_name->utf8_length(), \
119 (char *) name->bytes(), name->utf8_length(), \
120 (char *) signature->bytes(), signature->utf8_length(), (success)); \
121 }
122
123 #else // ndef DTRACE_ENABLED
124
125 #define DTRACE_METHOD_COMPILE_BEGIN_PROBE(method, comp_name)
126 #define DTRACE_METHOD_COMPILE_END_PROBE(method, comp_name, success)
127
128 #endif // ndef DTRACE_ENABLED
129
130 bool CompileBroker::_initialized = false;
131 volatile bool CompileBroker::_should_block = false;
132 volatile int CompileBroker::_print_compilation_warning = 0;
133 volatile jint CompileBroker::_should_compile_new_jobs = run_compilation;
134
135 // The installed compiler(s)
136 AbstractCompiler* CompileBroker::_compilers[2];
137
138 // The maximum numbers of compiler threads to be determined during startup.
139 int CompileBroker::_c1_count = 0;
140 int CompileBroker::_c2_count = 0;
141 int CompileBroker::_ac_count = 0;
142
143 // An array of compiler names as Java String objects
144 jobject* CompileBroker::_compiler1_objects = nullptr;
145 jobject* CompileBroker::_compiler2_objects = nullptr;
146 jobject* CompileBroker::_ac_objects = nullptr;
147
148 CompileLog** CompileBroker::_compiler1_logs = nullptr;
149 CompileLog** CompileBroker::_compiler2_logs = nullptr;
150 CompileLog** CompileBroker::_ac_logs = nullptr;
151
152 // These counters are used to assign an unique ID to each compilation.
153 volatile jint CompileBroker::_compilation_id = 0;
154 volatile jint CompileBroker::_osr_compilation_id = 0;
155 volatile jint CompileBroker::_native_compilation_id = 0;
156
157 // Performance counters
158 PerfCounter* CompileBroker::_perf_total_compilation = nullptr;
159 PerfCounter* CompileBroker::_perf_osr_compilation = nullptr;
160 PerfCounter* CompileBroker::_perf_standard_compilation = nullptr;
161
162 PerfCounter* CompileBroker::_perf_total_bailout_count = nullptr;
163 PerfCounter* CompileBroker::_perf_total_invalidated_count = nullptr;
164 PerfCounter* CompileBroker::_perf_total_compile_count = nullptr;
165 PerfCounter* CompileBroker::_perf_total_osr_compile_count = nullptr;
166 PerfCounter* CompileBroker::_perf_total_standard_compile_count = nullptr;
167
168 PerfCounter* CompileBroker::_perf_sum_osr_bytes_compiled = nullptr;
169 PerfCounter* CompileBroker::_perf_sum_standard_bytes_compiled = nullptr;
170 PerfCounter* CompileBroker::_perf_sum_nmethod_size = nullptr;
171 PerfCounter* CompileBroker::_perf_sum_nmethod_code_size = nullptr;
172
173 PerfStringVariable* CompileBroker::_perf_last_method = nullptr;
174 PerfStringVariable* CompileBroker::_perf_last_failed_method = nullptr;
175 PerfStringVariable* CompileBroker::_perf_last_invalidated_method = nullptr;
176 PerfVariable* CompileBroker::_perf_last_compile_type = nullptr;
177 PerfVariable* CompileBroker::_perf_last_compile_size = nullptr;
178 PerfVariable* CompileBroker::_perf_last_failed_type = nullptr;
179 PerfVariable* CompileBroker::_perf_last_invalidated_type = nullptr;
180
181 // Timers and counters for generating statistics
182 elapsedTimer CompileBroker::_t_total_compilation;
183 elapsedTimer CompileBroker::_t_osr_compilation;
184 elapsedTimer CompileBroker::_t_standard_compilation;
185 elapsedTimer CompileBroker::_t_invalidated_compilation;
186 elapsedTimer CompileBroker::_t_bailedout_compilation;
187
188 uint CompileBroker::_total_bailout_count = 0;
189 uint CompileBroker::_total_invalidated_count = 0;
190 uint CompileBroker::_total_not_entrant_count = 0;
191 uint CompileBroker::_total_compile_count = 0;
192 uint CompileBroker::_total_osr_compile_count = 0;
193 uint CompileBroker::_total_standard_compile_count = 0;
194 uint CompileBroker::_total_compiler_stopped_count = 0;
195 uint CompileBroker::_total_compiler_restarted_count = 0;
196
197 uint CompileBroker::_sum_osr_bytes_compiled = 0;
198 uint CompileBroker::_sum_standard_bytes_compiled = 0;
199 uint CompileBroker::_sum_nmethod_size = 0;
200 uint CompileBroker::_sum_nmethod_code_size = 0;
201
202 jlong CompileBroker::_peak_compilation_time = 0;
203
204 CompilerStatistics CompileBroker::_stats_per_level[CompLevel_full_optimization];
205 CompilerStatistics CompileBroker::_aot_stats;
206 CompilerStatistics CompileBroker::_aot_stats_per_level[CompLevel_full_optimization + 1];
207
208 CompileQueue* CompileBroker::_c2_compile_queue = nullptr;
209 CompileQueue* CompileBroker::_c1_compile_queue = nullptr;
210 CompileQueue* CompileBroker::_ac1_compile_queue = nullptr;
211 CompileQueue* CompileBroker::_ac2_compile_queue = nullptr;
212
213 bool compileBroker_init() {
214 if (LogEvents) {
215 CompilationLog::init();
216 }
217
218 // init directives stack, adding default directive
219 DirectivesStack::init();
220
221 if (DirectivesParser::has_file()) {
222 return DirectivesParser::parse_from_flag();
223 } else if (CompilerDirectivesPrint) {
224 // Print default directive even when no other was added
225 DirectivesStack::print(tty);
226 }
227
228 return true;
229 }
230
231 CompileTaskWrapper::CompileTaskWrapper(CompileTask* task) {
232 CompilerThread* thread = CompilerThread::current();
233 thread->set_task(task);
234 CompileLog* log = thread->log();
235 thread->timeout()->arm();
236 if (log != nullptr && !task->is_unloaded()) task->log_task_start(log);
237 }
238
239 CompileTaskWrapper::~CompileTaskWrapper() {
240 CompilerThread* thread = CompilerThread::current();
241
242 // First, disarm the timeout. This still relies on the underlying task.
243 thread->timeout()->disarm();
244
245 CompileTask* task = thread->task();
246 CompileLog* log = thread->log();
247 AbstractCompiler* comp = thread->compiler();
248 if (log != nullptr && !task->is_unloaded()) task->log_task_done(log);
249 thread->set_task(nullptr);
250 thread->set_env(nullptr);
251 if (task->is_blocking()) {
252 bool free_task = false;
253 {
254 MutexLocker notifier(thread, CompileTaskWait_lock);
255 task->mark_complete();
256 #if INCLUDE_JVMCI
257 if (comp->is_jvmci()) {
258 if (!task->has_waiter()) {
259 // The waiting thread timed out and thus did not delete the task.
260 free_task = true;
261 }
262 task->set_blocking_jvmci_compile_state(nullptr);
263 }
264 #endif
265 if (!free_task) {
266 // Notify the waiting thread that the compilation has completed
267 // so that it can free the task.
268 CompileTaskWait_lock->notify_all();
269 }
270 }
271 if (free_task) {
272 // The task can only be deleted once the task lock is released.
273 delete task;
274 }
275 } else {
276 task->mark_complete();
277
278 // By convention, the compiling thread is responsible for deleting
279 // a non-blocking CompileTask.
280 delete task;
281 }
282 }
283
284 /**
285 * Check if a CompilerThread can be removed and update count if requested.
286 */
287 bool CompileBroker::can_remove(CompilerThread *ct, bool do_it) {
288 assert(UseDynamicNumberOfCompilerThreads, "or shouldn't be here");
289 if (!ReduceNumberOfCompilerThreads) return false;
290
291 if (RecompilationPolicy::have_recompilation_work()) return false;
292
293 AbstractCompiler *compiler = ct->compiler();
294 int compiler_count = compiler->num_compiler_threads();
295 bool c1 = compiler->is_c1();
296
297 // Keep at least 1 compiler thread of each type.
298 if (compiler_count < 2) return false;
299
300 // Keep thread alive for at least some time.
301 if (ct->idle_time_millis() < (c1 ? 500 : 100)) return false;
302
303 #if INCLUDE_JVMCI
304 if (compiler->is_jvmci() && !UseJVMCINativeLibrary) {
305 // Handles for JVMCI thread objects may get released concurrently.
306 if (do_it) {
307 assert(CompileThread_lock->owner() == ct, "must be holding lock");
308 } else {
309 // Skip check if it's the last thread and let caller check again.
310 return true;
311 }
312 }
313 #endif
314
315 // We only allow the last compiler thread of each type to get removed.
316 jobject last_compiler = c1 ? compiler1_object(compiler_count - 1)
317 : compiler2_object(compiler_count - 1);
318 if (ct->threadObj() == JNIHandles::resolve_non_null(last_compiler)) {
319 if (do_it) {
320 assert_locked_or_safepoint(CompileThread_lock); // Update must be consistent.
321 compiler->set_num_compiler_threads(compiler_count - 1);
322 #if INCLUDE_JVMCI
323 if (compiler->is_jvmci() && !UseJVMCINativeLibrary) {
324 // Old j.l.Thread object can die when no longer referenced elsewhere.
325 JNIHandles::destroy_global(compiler2_object(compiler_count - 1));
326 _compiler2_objects[compiler_count - 1] = nullptr;
327 }
328 #endif
329 }
330 return true;
331 }
332 return false;
333 }
334
335 /**
336 * Add a CompileTask to a CompileQueue.
337 */
338 void CompileQueue::add(CompileTask* task) {
339 assert(_lock->owned_by_self(), "must own lock");
340
341 task->set_next(nullptr);
342 task->set_prev(nullptr);
343
344 if (_last == nullptr) {
345 // The compile queue is empty.
346 assert(_first == nullptr, "queue is empty");
347 _first = task;
348 _last = task;
349 } else {
350 // Append the task to the queue.
351 assert(_last->next() == nullptr, "not last");
352 _last->set_next(task);
353 task->set_prev(_last);
354 _last = task;
355 }
356 ++_size;
357 ++_total_added;
358 if (_size > _peak_size) {
359 _peak_size = _size;
360 }
361
362 // Mark the method as being in the compile queue.
363 task->method()->set_queued_for_compilation();
364
365 task->mark_queued(os::elapsed_counter());
366
367 if (CIPrintCompileQueue) {
368 print_tty();
369 }
370
371 if (LogCompilation && xtty != nullptr) {
372 task->log_task_queued();
373 }
374
375 if (TrainingData::need_data() && !CDSConfig::is_dumping_final_static_archive()) {
376 CompileTrainingData* ctd = CompileTrainingData::make(task);
377 if (ctd != nullptr) {
378 task->set_training_data(ctd);
379 }
380 }
381
382 // Notify CompilerThreads that a task is available.
383 _lock->notify_all();
384 }
385
386 void CompileQueue::add_pending(CompileTask* task) {
387 assert(_lock->owned_by_self() == false, "must NOT own lock");
388 assert(UseLockFreeCompileQueues, "");
389 task->method()->set_queued_for_compilation();
390 _queue.push(*task);
391 // FIXME: additional coordination needed? e.g., is it possible for compiler thread to block w/o processing pending tasks?
392 if (is_empty()) {
393 MutexLocker ml(_lock);
394 _lock->notify_all();
395 }
396 }
397
398 static bool process_pending(CompileTask* task) {
399 // guarantee(task->method()->queued_for_compilation(), "");
400 if (task->is_unloaded()) {
401 return true; // unloaded
402 }
403 task->method()->set_queued_for_compilation(); // FIXME
404 if (task->method()->pending_queue_processed()) {
405 return true; // already queued
406 }
407 // Mark the method as being in the compile queue.
408 task->method()->set_pending_queue_processed();
409 if (CompileBroker::compilation_is_complete(task->method(), task->osr_bci(), task->comp_level(),
410 task->requires_online_compilation(), task->compile_reason())) {
411 return true; // already compiled
412 }
413 return false; // active
414 }
415
416 void CompileQueue::transfer_pending() {
417 assert(_lock->owned_by_self(), "must own lock");
418
419 CompileTask* task;
420 while ((task = _queue.pop()) != nullptr) {
421 bool is_stale = process_pending(task);
422 if (is_stale) {
423 task->set_next(_first_stale);
424 task->set_prev(nullptr);
425 _first_stale = task;
426 } else {
427 add(task);
428 }
429 }
430 }
431
432 /**
433 * Empties compilation queue by deleting all compilation tasks.
434 * Furthermore, the method wakes up all threads that are waiting
435 * on a compilation task to finish. This can happen if background
436 * compilation is disabled.
437 */
438 void CompileQueue::delete_all() {
439 MutexLocker mu(_lock);
440 transfer_pending();
441
442 CompileTask* current = _first;
443
444 // Iterate over all tasks in the compile queue
445 while (current != nullptr) {
446 CompileTask* next = current->next();
447 if (!current->is_blocking()) {
448 // Non-blocking task. No one is waiting for it, delete it now.
449 delete current;
450 } else {
451 // Blocking task. By convention, it is the waiters responsibility
452 // to delete the task. We cannot delete it here, because we do not
453 // coordinate with waiters. We will notify the waiters later.
454 }
455 current = next;
456 }
457 _first = nullptr;
458 _last = nullptr;
459
460 // Wake up all blocking task waiters to deal with remaining blocking
461 // tasks. This is not a performance sensitive path, so we do this
462 // unconditionally to simplify coding/testing.
463 {
464 MonitorLocker ml(Thread::current(), CompileTaskWait_lock);
465 ml.notify_all();
466 }
467
468 // Wake up all threads that block on the queue.
469 _lock->notify_all();
470 }
471
472 /**
473 * Get the next CompileTask from a CompileQueue
474 */
475 CompileTask* CompileQueue::get(CompilerThread* thread) {
476 // save methods from RedefineClasses across safepoint
477 // across compile queue lock below.
478 methodHandle save_method;
479
480 MonitorLocker locker(_lock);
481 transfer_pending();
482
483 RecompilationPolicy::sample_load_average();
484
485 // If _first is null we have no more compile jobs. There are two reasons for
486 // having no compile jobs: First, we compiled everything we wanted. Second,
487 // we ran out of code cache so compilation has been disabled. In the latter
488 // case we perform code cache sweeps to free memory such that we can re-enable
489 // compilation.
490 while (_first == nullptr) {
491 // Exit loop if compilation is disabled forever
492 if (CompileBroker::is_compilation_disabled_forever()) {
493 return nullptr;
494 }
495
496 AbstractCompiler* compiler = thread->compiler();
497 guarantee(compiler != nullptr, "Compiler object must exist");
498 compiler->on_empty_queue(this, thread);
499 if (_first != nullptr) {
500 // The call to on_empty_queue may have temporarily unlocked the MCQ lock
501 // so check again whether any tasks were added to the queue.
502 break;
503 }
504
505 // If we have added stale tasks, there might be waiters that want
506 // the notification these tasks have failed. Normally, this would
507 // be done by a compiler thread that would perform the purge at
508 // the end of some compilation. But, if compile queue is empty,
509 // there is no guarantee compilers would run and do the purge.
510 // Do the purge here and now to unblock the waiters.
511 // Perform this until we run out of stale tasks.
512 while (_first_stale != nullptr) {
513 purge_stale_tasks();
514 }
515 if (_first != nullptr) {
516 // Purge stale tasks may have transferred some new tasks,
517 // so check again.
518 break;
519 }
520
521 // If there are no compilation tasks and we can compile new jobs
522 // (i.e., there is enough free space in the code cache) there is
523 // no need to invoke the GC.
524 // We need a timed wait here, since compiler threads can exit if compilation
525 // is disabled forever. We use 5 seconds wait time; the exiting of compiler threads
526 // is not critical and we do not want idle compiler threads to wake up too often.
527 locker.wait(5*1000);
528
529 transfer_pending(); // reacquired lock
530
531 if (RecompilationPolicy::have_recompilation_work()) return nullptr;
532
533 if (UseDynamicNumberOfCompilerThreads && _first == nullptr) {
534 // Still nothing to compile. Give caller a chance to stop this thread.
535 if (CompileBroker::can_remove(CompilerThread::current(), false)) return nullptr;
536 }
537 }
538
539 if (CompileBroker::is_compilation_disabled_forever()) {
540 return nullptr;
541 }
542
543 CompileTask* task;
544 {
545 NoSafepointVerifier nsv;
546 task = CompilationPolicy::select_task(this, thread);
547 if (task != nullptr) {
548 task = task->select_for_compilation();
549 }
550 }
551
552 if (task != nullptr) {
553 // Save method pointers across unlock safepoint. The task is removed from
554 // the compilation queue, which is walked during RedefineClasses.
555 Thread* thread = Thread::current();
556 save_method = methodHandle(thread, task->method());
557
558 remove(task);
559 }
560 purge_stale_tasks(); // may temporarily release MCQ lock
561 return task;
562 }
563
564 // Clean & deallocate stale compile tasks.
565 // Temporarily releases MethodCompileQueue lock.
566 void CompileQueue::purge_stale_tasks() {
567 assert(_lock->owned_by_self(), "must own lock");
568 if (_first_stale != nullptr) {
569 // Stale tasks are purged when MCQ lock is released,
570 // but _first_stale updates are protected by MCQ lock.
571 // Once task processing starts and MCQ lock is released,
572 // other compiler threads can reuse _first_stale.
573 CompileTask* head = _first_stale;
574 _first_stale = nullptr;
575 {
576 MutexUnlocker ul(_lock);
577 for (CompileTask* task = head; task != nullptr; ) {
578 CompileTask* next_task = task->next();
579 task->set_next(nullptr);
580 CompileTaskWrapper ctw(task); // Frees the task
581 task->set_failure_reason("stale task");
582 task = next_task;
583 }
584 }
585 transfer_pending(); // transfer pending after reacquiring MCQ lock
586 }
587 }
588
589 void CompileQueue::remove(CompileTask* task) {
590 assert(_lock->owned_by_self(), "must own lock");
591 if (task->prev() != nullptr) {
592 task->prev()->set_next(task->next());
593 } else {
594 // max is the first element
595 assert(task == _first, "Sanity");
596 _first = task->next();
597 }
598
599 if (task->next() != nullptr) {
600 task->next()->set_prev(task->prev());
601 } else {
602 // max is the last element
603 assert(task == _last, "Sanity");
604 _last = task->prev();
605 }
606 task->set_next(nullptr);
607 task->set_prev(nullptr);
608 --_size;
609 ++_total_removed;
610 }
611
612 void CompileQueue::remove_and_mark_stale(CompileTask* task) {
613 assert(_lock->owned_by_self(), "must own lock");
614 remove(task);
615
616 // Enqueue the task for reclamation (should be done outside MCQ lock)
617 task->set_next(_first_stale);
618 task->set_prev(nullptr);
619 _first_stale = task;
620 }
621
622 // methods in the compile queue need to be marked as used on the stack
623 // so that they don't get reclaimed by Redefine Classes
624 void CompileQueue::mark_on_stack() {
625 for (CompileTask* task = _first; task != nullptr; task = task->next()) {
626 task->mark_on_stack();
627 }
628 for (CompileTask* task = _queue.first(); !_queue.is_end(task); task = task->next()) {
629 assert(task != nullptr, "");
630 task->mark_on_stack();
631 }
632 }
633
634
635 CompileQueue* CompileBroker::compile_queue(int comp_level, bool is_aot) {
636 if (is_c2_compile(comp_level)) return ((is_aot && (_ac_count > 0)) ? _ac2_compile_queue : _c2_compile_queue);
637 if (is_c1_compile(comp_level)) return ((is_aot && (_ac_count > 0)) ? _ac1_compile_queue : _c1_compile_queue);
638 return nullptr;
639 }
640
641 CompileQueue* CompileBroker::c1_compile_queue() {
642 return _c1_compile_queue;
643 }
644
645 CompileQueue* CompileBroker::c2_compile_queue() {
646 return _c2_compile_queue;
647 }
648
649 void CompileBroker::print_compile_queues(outputStream* st) {
650 st->print_cr("Current compiles: ");
651
652 char buf[2000];
653 int buflen = sizeof(buf);
654 Threads::print_threads_compiling(st, buf, buflen, /* short_form = */ true);
655
656 st->cr();
657 if (_c1_compile_queue != nullptr) {
658 _c1_compile_queue->print(st);
659 }
660 if (_c2_compile_queue != nullptr) {
661 _c2_compile_queue->print(st);
662 }
663 if (_ac1_compile_queue != nullptr) {
664 _ac1_compile_queue->print(st);
665 }
666 if (_ac2_compile_queue != nullptr) {
667 _ac2_compile_queue->print(st);
668 }
669 }
670
671 void CompileQueue::print(outputStream* st) {
672 assert_locked_or_safepoint(_lock);
673 st->print_cr("%s:", name());
674 CompileTask* task = _first;
675 if (task == nullptr) {
676 st->print_cr("Empty");
677 } else {
678 while (task != nullptr) {
679 task->print(st, nullptr, true, true);
680 task = task->next();
681 }
682 }
683 st->cr();
684 }
685
686 void CompileQueue::print_tty() {
687 stringStream ss;
688 // Dump the compile queue into a buffer before locking the tty
689 print(&ss);
690 {
691 ttyLocker ttyl;
692 tty->print("%s", ss.freeze());
693 }
694 }
695
696 CompilerCounters::CompilerCounters() {
697 _current_method[0] = '\0';
698 _compile_type = CompileBroker::no_compile;
699 }
700
701 #if INCLUDE_JFR && COMPILER2_OR_JVMCI
702 // It appends new compiler phase names to growable array phase_names(a new CompilerPhaseType mapping
703 // in compiler/compilerEvent.cpp) and registers it with its serializer.
704 //
705 // c2 uses explicit CompilerPhaseType idToPhase mapping in opto/phasetype.hpp,
706 // so if c2 is used, it should be always registered first.
707 // This function is called during vm initialization.
708 static void register_jfr_phasetype_serializer(CompilerType compiler_type) {
709 ResourceMark rm;
710 static bool first_registration = true;
711 if (compiler_type == compiler_jvmci) {
712 CompilerEvent::PhaseEvent::get_phase_id("NOT_A_PHASE_NAME", false, false, false);
713 first_registration = false;
714 #ifdef COMPILER2
715 } else if (compiler_type == compiler_c2) {
716 assert(first_registration, "invariant"); // c2 must be registered first.
717 for (int i = 0; i < PHASE_NUM_TYPES; i++) {
718 const char* phase_name = CompilerPhaseTypeHelper::to_description((CompilerPhaseType) i);
719 CompilerEvent::PhaseEvent::get_phase_id(phase_name, false, false, false);
720 }
721 first_registration = false;
722 #endif // COMPILER2
723 }
724 }
725 #endif // INCLUDE_JFR && COMPILER2_OR_JVMCI
726
727 // ------------------------------------------------------------------
728 // CompileBroker::compilation_init
729 //
730 // Initialize the Compilation object
731 void CompileBroker::compilation_init(JavaThread* THREAD) {
732 // No need to initialize compilation system if we do not use it.
733 if (!UseCompiler) {
734 return;
735 }
736 // Set the interface to the current compiler(s).
737 _c1_count = CompilationPolicy::c1_count();
738 _c2_count = CompilationPolicy::c2_count();
739 _ac_count = CompilationPolicy::ac_count();
740
741 #if INCLUDE_JVMCI
742 if (EnableJVMCI) {
743 // This is creating a JVMCICompiler singleton.
744 JVMCICompiler* jvmci = new JVMCICompiler();
745
746 if (UseJVMCICompiler) {
747 _compilers[1] = jvmci;
748 if (FLAG_IS_DEFAULT(JVMCIThreads)) {
749 if (BootstrapJVMCI) {
750 // JVMCI will bootstrap so give it more threads
751 _c2_count = MIN2(32, os::active_processor_count());
752 }
753 } else {
754 _c2_count = JVMCIThreads;
755 }
756 if (FLAG_IS_DEFAULT(JVMCIHostThreads)) {
757 } else {
758 #ifdef COMPILER1
759 _c1_count = JVMCIHostThreads;
760 #endif // COMPILER1
761 }
762 }
763 }
764 #endif // INCLUDE_JVMCI
765
766 #ifdef COMPILER1
767 if (_c1_count > 0) {
768 _compilers[0] = new Compiler();
769 }
770 #endif // COMPILER1
771
772 #ifdef COMPILER2
773 if (true JVMCI_ONLY( && !UseJVMCICompiler)) {
774 if (_c2_count > 0) {
775 _compilers[1] = new C2Compiler();
776 // Register c2 first as c2 CompilerPhaseType idToPhase mapping is explicit.
777 // idToPhase mapping for c2 is in opto/phasetype.hpp
778 JFR_ONLY(register_jfr_phasetype_serializer(compiler_c2);)
779 }
780 }
781 #endif // COMPILER2
782
783 #if INCLUDE_JVMCI
784 // Register after c2 registration.
785 // JVMCI CompilerPhaseType idToPhase mapping is dynamic.
786 if (EnableJVMCI) {
787 JFR_ONLY(register_jfr_phasetype_serializer(compiler_jvmci);)
788 }
789 #endif // INCLUDE_JVMCI
790
791 if (CompilerOracle::should_collect_memstat()) {
792 CompilationMemoryStatistic::initialize();
793 }
794
795 // Start the compiler thread(s)
796 init_compiler_threads();
797 // totalTime performance counter is always created as it is required
798 // by the implementation of java.lang.management.CompilationMXBean.
799 {
800 // Ensure OOM leads to vm_exit_during_initialization.
801 EXCEPTION_MARK;
802 _perf_total_compilation =
803 PerfDataManager::create_counter(JAVA_CI, "totalTime",
804 PerfData::U_Ticks, CHECK);
805 }
806
807 if (UsePerfData) {
808
809 EXCEPTION_MARK;
810
811 // create the jvmstat performance counters
812 _perf_osr_compilation =
813 PerfDataManager::create_counter(SUN_CI, "osrTime",
814 PerfData::U_Ticks, CHECK);
815
816 _perf_standard_compilation =
817 PerfDataManager::create_counter(SUN_CI, "standardTime",
818 PerfData::U_Ticks, CHECK);
819
820 _perf_total_bailout_count =
821 PerfDataManager::create_counter(SUN_CI, "totalBailouts",
822 PerfData::U_Events, CHECK);
823
824 _perf_total_invalidated_count =
825 PerfDataManager::create_counter(SUN_CI, "totalInvalidates",
826 PerfData::U_Events, CHECK);
827
828 _perf_total_compile_count =
829 PerfDataManager::create_counter(SUN_CI, "totalCompiles",
830 PerfData::U_Events, CHECK);
831 _perf_total_osr_compile_count =
832 PerfDataManager::create_counter(SUN_CI, "osrCompiles",
833 PerfData::U_Events, CHECK);
834
835 _perf_total_standard_compile_count =
836 PerfDataManager::create_counter(SUN_CI, "standardCompiles",
837 PerfData::U_Events, CHECK);
838
839 _perf_sum_osr_bytes_compiled =
840 PerfDataManager::create_counter(SUN_CI, "osrBytes",
841 PerfData::U_Bytes, CHECK);
842
843 _perf_sum_standard_bytes_compiled =
844 PerfDataManager::create_counter(SUN_CI, "standardBytes",
845 PerfData::U_Bytes, CHECK);
846
847 _perf_sum_nmethod_size =
848 PerfDataManager::create_counter(SUN_CI, "nmethodSize",
849 PerfData::U_Bytes, CHECK);
850
851 _perf_sum_nmethod_code_size =
852 PerfDataManager::create_counter(SUN_CI, "nmethodCodeSize",
853 PerfData::U_Bytes, CHECK);
854
855 _perf_last_method =
856 PerfDataManager::create_string_variable(SUN_CI, "lastMethod",
857 CompilerCounters::cmname_buffer_length,
858 "", CHECK);
859
860 _perf_last_failed_method =
861 PerfDataManager::create_string_variable(SUN_CI, "lastFailedMethod",
862 CompilerCounters::cmname_buffer_length,
863 "", CHECK);
864
865 _perf_last_invalidated_method =
866 PerfDataManager::create_string_variable(SUN_CI, "lastInvalidatedMethod",
867 CompilerCounters::cmname_buffer_length,
868 "", CHECK);
869
870 _perf_last_compile_type =
871 PerfDataManager::create_variable(SUN_CI, "lastType",
872 PerfData::U_None,
873 (jlong)CompileBroker::no_compile,
874 CHECK);
875
876 _perf_last_compile_size =
877 PerfDataManager::create_variable(SUN_CI, "lastSize",
878 PerfData::U_Bytes,
879 (jlong)CompileBroker::no_compile,
880 CHECK);
881
882
883 _perf_last_failed_type =
884 PerfDataManager::create_variable(SUN_CI, "lastFailedType",
885 PerfData::U_None,
886 (jlong)CompileBroker::no_compile,
887 CHECK);
888
889 _perf_last_invalidated_type =
890 PerfDataManager::create_variable(SUN_CI, "lastInvalidatedType",
891 PerfData::U_None,
892 (jlong)CompileBroker::no_compile,
893 CHECK);
894 }
895
896 log_info(aot, codecache, init)("CompileBroker is initialized");
897 _initialized = true;
898 }
899
900 Handle CompileBroker::create_thread_oop(const char* name, TRAPS) {
901 Handle thread_oop = JavaThread::create_system_thread_object(name, CHECK_NH);
902 return thread_oop;
903 }
904
905 void TrainingReplayThread::training_replay_thread_entry(JavaThread* thread, TRAPS) {
906 CompilationPolicy::replay_training_at_init_loop(thread);
907 }
908
909 #if defined(ASSERT) && COMPILER2_OR_JVMCI
910 // Entry for DeoptimizeObjectsALotThread. The threads are started in
911 // CompileBroker::init_compiler_threads() iff DeoptimizeObjectsALot is enabled
912 void DeoptimizeObjectsALotThread::deopt_objs_alot_thread_entry(JavaThread* thread, TRAPS) {
913 DeoptimizeObjectsALotThread* dt = ((DeoptimizeObjectsALotThread*) thread);
914 bool enter_single_loop;
915 {
916 MonitorLocker ml(dt, EscapeBarrier_lock, Mutex::_no_safepoint_check_flag);
917 static int single_thread_count = 0;
918 enter_single_loop = single_thread_count++ < DeoptimizeObjectsALotThreadCountSingle;
919 }
920 if (enter_single_loop) {
921 dt->deoptimize_objects_alot_loop_single();
922 } else {
923 dt->deoptimize_objects_alot_loop_all();
924 }
925 }
926
927 // Execute EscapeBarriers in an endless loop to revert optimizations based on escape analysis. Each
928 // barrier targets a single thread which is selected round robin.
929 void DeoptimizeObjectsALotThread::deoptimize_objects_alot_loop_single() {
930 HandleMark hm(this);
931 while (true) {
932 for (JavaThreadIteratorWithHandle jtiwh; JavaThread *deoptee_thread = jtiwh.next(); ) {
933 { // Begin new scope for escape barrier
934 HandleMarkCleaner hmc(this);
935 ResourceMark rm(this);
936 EscapeBarrier eb(true, this, deoptee_thread);
937 eb.deoptimize_objects(100);
938 }
939 // Now sleep after the escape barriers destructor resumed deoptee_thread.
940 sleep(DeoptimizeObjectsALotInterval);
941 }
942 }
943 }
944
945 // Execute EscapeBarriers in an endless loop to revert optimizations based on escape analysis. Each
946 // barrier targets all java threads in the vm at once.
947 void DeoptimizeObjectsALotThread::deoptimize_objects_alot_loop_all() {
948 HandleMark hm(this);
949 while (true) {
950 { // Begin new scope for escape barrier
951 HandleMarkCleaner hmc(this);
952 ResourceMark rm(this);
953 EscapeBarrier eb(true, this);
954 eb.deoptimize_objects_all_threads();
955 }
956 // Now sleep after the escape barriers destructor resumed the java threads.
957 sleep(DeoptimizeObjectsALotInterval);
958 }
959 }
960 #endif // defined(ASSERT) && COMPILER2_OR_JVMCI
961
962
963 JavaThread* CompileBroker::make_thread(ThreadType type, jobject thread_handle, CompileQueue* queue, AbstractCompiler* comp, JavaThread* THREAD) {
964 Handle thread_oop(THREAD, JNIHandles::resolve_non_null(thread_handle));
965
966 if (java_lang_Thread::thread(thread_oop()) != nullptr) {
967 assert(type == compiler_t, "should only happen with reused compiler threads");
968 // The compiler thread hasn't actually exited yet so don't try to reuse it
969 return nullptr;
970 }
971
972 JavaThread* new_thread = nullptr;
973 switch (type) {
974 case compiler_t:
975 assert(comp != nullptr, "Compiler instance missing.");
976 if (!InjectCompilerCreationFailure || comp->num_compiler_threads() == 0) {
977 CompilerCounters* counters = new CompilerCounters();
978 new_thread = new CompilerThread(queue, counters);
979 }
980 break;
981 #if defined(ASSERT) && COMPILER2_OR_JVMCI
982 case deoptimizer_t:
983 new_thread = new DeoptimizeObjectsALotThread();
984 break;
985 #endif // ASSERT
986 case training_replay_t:
987 new_thread = new TrainingReplayThread();
988 break;
989 default:
990 ShouldNotReachHere();
991 }
992
993 // At this point the new CompilerThread data-races with this startup
994 // thread (which is the main thread and NOT the VM thread).
995 // This means Java bytecodes being executed at startup can
996 // queue compile jobs which will run at whatever default priority the
997 // newly created CompilerThread runs at.
998
999
1000 // At this point it may be possible that no osthread was created for the
1001 // JavaThread due to lack of resources. We will handle that failure below.
1002 // Also check new_thread so that static analysis is happy.
1003 if (new_thread != nullptr && new_thread->osthread() != nullptr) {
1004
1005 if (type == compiler_t) {
1006 CompilerThread::cast(new_thread)->set_compiler(comp);
1007 }
1008
1009 // Note that we cannot call os::set_priority because it expects Java
1010 // priorities and we are *explicitly* using OS priorities so that it's
1011 // possible to set the compiler thread priority higher than any Java
1012 // thread.
1013
1014 int native_prio = CompilerThreadPriority;
1015 if (native_prio == -1) {
1016 if (UseCriticalCompilerThreadPriority) {
1017 native_prio = os::java_to_os_priority[CriticalPriority];
1018 } else {
1019 native_prio = os::java_to_os_priority[NearMaxPriority];
1020 }
1021 }
1022 os::set_native_priority(new_thread, native_prio);
1023
1024 // Note that this only sets the JavaThread _priority field, which by
1025 // definition is limited to Java priorities and not OS priorities.
1026 JavaThread::start_internal_daemon(THREAD, new_thread, thread_oop, NearMaxPriority);
1027
1028 } else { // osthread initialization failure
1029 if (UseDynamicNumberOfCompilerThreads && type == compiler_t
1030 && comp->num_compiler_threads() > 0) {
1031 // The new thread is not known to Thread-SMR yet so we can just delete.
1032 delete new_thread;
1033 return nullptr;
1034 } else {
1035 vm_exit_during_initialization("java.lang.OutOfMemoryError",
1036 os::native_thread_creation_failed_msg());
1037 }
1038 }
1039
1040 os::naked_yield(); // make sure that the compiler thread is started early (especially helpful on SOLARIS)
1041
1042 return new_thread;
1043 }
1044
1045 static bool trace_compiler_threads() {
1046 LogTarget(Debug, jit, thread) lt;
1047 return TraceCompilerThreads || lt.is_enabled();
1048 }
1049
1050 static jobject create_compiler_thread(AbstractCompiler* compiler, int i, TRAPS) {
1051 char name_buffer[256];
1052 os::snprintf_checked(name_buffer, sizeof(name_buffer), "%s CompilerThread%d", compiler->name(), i);
1053 Handle thread_oop = JavaThread::create_system_thread_object(name_buffer, CHECK_NULL);
1054 return JNIHandles::make_global(thread_oop);
1055 }
1056
1057 static void print_compiler_threads(stringStream& msg) {
1058 if (TraceCompilerThreads) {
1059 tty->print_cr("%7d %s", (int)tty->time_stamp().milliseconds(), msg.as_string());
1060 }
1061 LogTarget(Debug, jit, thread) lt;
1062 if (lt.is_enabled()) {
1063 LogStream ls(lt);
1064 ls.print_cr("%s", msg.as_string());
1065 }
1066 }
1067
1068 static void print_compiler_thread(JavaThread *ct) {
1069 if (trace_compiler_threads()) {
1070 ResourceMark rm;
1071 ThreadsListHandle tlh; // name() depends on the TLH.
1072 assert(tlh.includes(ct), "ct=" INTPTR_FORMAT " exited unexpectedly.", p2i(ct));
1073 stringStream msg;
1074 msg.print("Added initial compiler thread %s", ct->name());
1075 print_compiler_threads(msg);
1076 }
1077 }
1078
1079 void CompileBroker::init_compiler_threads() {
1080 // Ensure any exceptions lead to vm_exit_during_initialization.
1081 EXCEPTION_MARK;
1082 #if !defined(ZERO)
1083 assert(_c2_count > 0 || _c1_count > 0, "No compilers?");
1084 #endif // !ZERO
1085 // Initialize the compilation queue
1086 if (_c2_count > 0) {
1087 const char* name = JVMCI_ONLY(UseJVMCICompiler ? "JVMCI compile queue" :) "C2 compile queue";
1088 _c2_compile_queue = new CompileQueue(name, MethodCompileQueueC2_lock);
1089 _compiler2_objects = NEW_C_HEAP_ARRAY(jobject, _c2_count, mtCompiler);
1090 _compiler2_logs = NEW_C_HEAP_ARRAY(CompileLog*, _c2_count, mtCompiler);
1091 }
1092 if (_c1_count > 0) {
1093 _c1_compile_queue = new CompileQueue("C1 compile queue", MethodCompileQueueC1_lock);
1094 _compiler1_objects = NEW_C_HEAP_ARRAY(jobject, _c1_count, mtCompiler);
1095 _compiler1_logs = NEW_C_HEAP_ARRAY(CompileLog*, _c1_count, mtCompiler);
1096 }
1097
1098 if (_ac_count > 0) {
1099 if (_c1_count > 0) { // C1 is present
1100 _ac1_compile_queue = new CompileQueue("C1 AOT code compile queue", MethodCompileQueueAC1_lock);
1101 }
1102 if (_c2_count > 0) { // C2 is present
1103 _ac2_compile_queue = new CompileQueue("C2 AOT code compile queue", MethodCompileQueueAC2_lock);
1104 }
1105 _ac_objects = NEW_C_HEAP_ARRAY(jobject, _ac_count, mtCompiler);
1106 _ac_logs = NEW_C_HEAP_ARRAY(CompileLog*, _ac_count, mtCompiler);
1107 }
1108 char name_buffer[256];
1109
1110 for (int i = 0; i < _c2_count; i++) {
1111 // Create a name for our thread.
1112 jobject thread_handle = create_compiler_thread(_compilers[1], i, CHECK);
1113 _compiler2_objects[i] = thread_handle;
1114 _compiler2_logs[i] = nullptr;
1115
1116 if (!UseDynamicNumberOfCompilerThreads || i == 0) {
1117 JavaThread *ct = make_thread(compiler_t, thread_handle, _c2_compile_queue, _compilers[1], THREAD);
1118 assert(ct != nullptr, "should have been handled for initial thread");
1119 _compilers[1]->set_num_compiler_threads(i + 1);
1120 print_compiler_thread(ct);
1121 }
1122 }
1123
1124 for (int i = 0; i < _c1_count; i++) {
1125 // Create a name for our thread.
1126 jobject thread_handle = create_compiler_thread(_compilers[0], i, CHECK);
1127 _compiler1_objects[i] = thread_handle;
1128 _compiler1_logs[i] = nullptr;
1129
1130 if (!UseDynamicNumberOfCompilerThreads || i == 0) {
1131 JavaThread *ct = make_thread(compiler_t, thread_handle, _c1_compile_queue, _compilers[0], THREAD);
1132 assert(ct != nullptr, "should have been handled for initial thread");
1133 _compilers[0]->set_num_compiler_threads(i + 1);
1134 print_compiler_thread(ct);
1135 }
1136 }
1137
1138 if (_ac_count > 0) {
1139 int i = 0;
1140 if (_c1_count > 0) { // C1 is present
1141 os::snprintf_checked(name_buffer, sizeof(name_buffer), "C%d AOT code caching CompilerThread", 1);
1142 Handle thread_oop = create_thread_oop(name_buffer, CHECK);
1143 jobject thread_handle = JNIHandles::make_global(thread_oop);
1144 _ac_objects[i] = thread_handle;
1145 _ac_logs[i] = nullptr;
1146 i++;
1147
1148 JavaThread *ct = make_thread(compiler_t, thread_handle, _ac1_compile_queue, _compilers[0], THREAD);
1149 assert(ct != nullptr, "should have been handled for initial thread");
1150 print_compiler_thread(ct);
1151 }
1152 if (_c2_count > 0) { // C2 is present
1153 os::snprintf_checked(name_buffer, sizeof(name_buffer), "C%d AOT code caching CompilerThread", 2);
1154 Handle thread_oop = create_thread_oop(name_buffer, CHECK);
1155 jobject thread_handle = JNIHandles::make_global(thread_oop);
1156 _ac_objects[i] = thread_handle;
1157 _ac_logs[i] = nullptr;
1158
1159 JavaThread *ct = make_thread(compiler_t, thread_handle, _ac2_compile_queue, _compilers[1], THREAD);
1160 assert(ct != nullptr, "should have been handled for initial thread");
1161 print_compiler_thread(ct);
1162 }
1163 }
1164
1165 if (UsePerfData) {
1166 PerfDataManager::create_constant(SUN_CI, "threads", PerfData::U_Bytes, _c1_count + _c2_count, CHECK);
1167 }
1168
1169 #if defined(ASSERT) && COMPILER2_OR_JVMCI
1170 if (DeoptimizeObjectsALot) {
1171 // Initialize and start the object deoptimizer threads
1172 const int total_count = DeoptimizeObjectsALotThreadCountSingle + DeoptimizeObjectsALotThreadCountAll;
1173 for (int count = 0; count < total_count; count++) {
1174 Handle thread_oop = JavaThread::create_system_thread_object("Deoptimize objects a lot single mode", CHECK);
1175 jobject thread_handle = JNIHandles::make_local(THREAD, thread_oop());
1176 make_thread(deoptimizer_t, thread_handle, nullptr, nullptr, THREAD);
1177 }
1178 }
1179 #endif // defined(ASSERT) && COMPILER2_OR_JVMCI
1180 }
1181
1182 void CompileBroker::init_training_replay() {
1183 // Ensure any exceptions lead to vm_exit_during_initialization.
1184 EXCEPTION_MARK;
1185 if (TrainingData::have_data() && CDSConfig::is_using_aot_linked_classes()) {
1186 Handle thread_oop = create_thread_oop("Training replay thread", CHECK);
1187 jobject thread_handle = JNIHandles::make_local(THREAD, thread_oop());
1188 make_thread(training_replay_t, thread_handle, nullptr, nullptr, THREAD);
1189 }
1190 }
1191
1192 void CompileBroker::possibly_add_compiler_threads(JavaThread* THREAD) {
1193
1194 int old_c2_count = 0, new_c2_count = 0, old_c1_count = 0, new_c1_count = 0;
1195 const int c2_tasks_per_thread = 2, c1_tasks_per_thread = 4;
1196
1197 // Quick check if we already have enough compiler threads without taking the lock.
1198 // Numbers may change concurrently, so we read them again after we have the lock.
1199 if (_c2_compile_queue != nullptr) {
1200 old_c2_count = get_c2_thread_count();
1201 new_c2_count = MIN2(_c2_count, _c2_compile_queue->size() / c2_tasks_per_thread);
1202 }
1203 if (_c1_compile_queue != nullptr) {
1204 old_c1_count = get_c1_thread_count();
1205 new_c1_count = MIN2(_c1_count, _c1_compile_queue->size() / c1_tasks_per_thread);
1206 }
1207 if (new_c2_count <= old_c2_count && new_c1_count <= old_c1_count) return;
1208
1209 // Now, we do the more expensive operations.
1210 physical_memory_size_type free_memory = 0;
1211 // Return value ignored - defaulting to 0 on failure.
1212 (void)os::free_memory(free_memory);
1213 // If SegmentedCodeCache is off, both values refer to the single heap (with type CodeBlobType::All).
1214 size_t available_cc_np = CodeCache::unallocated_capacity(CodeBlobType::MethodNonProfiled),
1215 available_cc_p = CodeCache::unallocated_capacity(CodeBlobType::MethodProfiled);
1216
1217 // Only attempt to start additional threads if the lock is free.
1218 if (!CompileThread_lock->try_lock()) return;
1219
1220 if (_c2_compile_queue != nullptr) {
1221 old_c2_count = get_c2_thread_count();
1222 new_c2_count = MIN4(_c2_count,
1223 _c2_compile_queue->size() / c2_tasks_per_thread,
1224 (int)(free_memory / (200*M)),
1225 (int)(available_cc_np / (128*K)));
1226
1227 for (int i = old_c2_count; i < new_c2_count; i++) {
1228 #if INCLUDE_JVMCI
1229 if (UseJVMCICompiler && !UseJVMCINativeLibrary && _compiler2_objects[i] == nullptr) {
1230 // Native compiler threads as used in C1/C2 can reuse the j.l.Thread objects as their
1231 // existence is completely hidden from the rest of the VM (and those compiler threads can't
1232 // call Java code to do the creation anyway).
1233 //
1234 // For pure Java JVMCI we have to create new j.l.Thread objects as they are visible and we
1235 // can see unexpected thread lifecycle transitions if we bind them to new JavaThreads. For
1236 // native library JVMCI it's preferred to use the C1/C2 strategy as this avoids unnecessary
1237 // coupling with Java.
1238 if (!THREAD->can_call_java()) break;
1239 char name_buffer[256];
1240 os::snprintf_checked(name_buffer, sizeof(name_buffer), "%s CompilerThread%d", _compilers[1]->name(), i);
1241 Handle thread_oop;
1242 {
1243 // We have to give up the lock temporarily for the Java calls.
1244 MutexUnlocker mu(CompileThread_lock);
1245 thread_oop = JavaThread::create_system_thread_object(name_buffer, THREAD);
1246 }
1247 if (HAS_PENDING_EXCEPTION) {
1248 if (trace_compiler_threads()) {
1249 ResourceMark rm;
1250 stringStream msg;
1251 msg.print_cr("JVMCI compiler thread creation failed:");
1252 PENDING_EXCEPTION->print_on(&msg);
1253 print_compiler_threads(msg);
1254 }
1255 CLEAR_PENDING_EXCEPTION;
1256 break;
1257 }
1258 // Check if another thread has beaten us during the Java calls.
1259 if (get_c2_thread_count() != i) break;
1260 jobject thread_handle = JNIHandles::make_global(thread_oop);
1261 assert(compiler2_object(i) == nullptr, "Old one must be released!");
1262 _compiler2_objects[i] = thread_handle;
1263 }
1264 #endif
1265 guarantee(compiler2_object(i) != nullptr, "Thread oop must exist");
1266 JavaThread *ct = make_thread(compiler_t, compiler2_object(i), _c2_compile_queue, _compilers[1], THREAD);
1267 if (ct == nullptr) break;
1268 _compilers[1]->set_num_compiler_threads(i + 1);
1269 if (trace_compiler_threads()) {
1270 ResourceMark rm;
1271 ThreadsListHandle tlh; // name() depends on the TLH.
1272 assert(tlh.includes(ct), "ct=" INTPTR_FORMAT " exited unexpectedly.", p2i(ct));
1273 stringStream msg;
1274 msg.print("Added compiler thread %s (free memory: %dMB, available non-profiled code cache: %dMB)",
1275 ct->name(), (int)(free_memory/M), (int)(available_cc_np/M));
1276 print_compiler_threads(msg);
1277 }
1278 }
1279 }
1280
1281 if (_c1_compile_queue != nullptr) {
1282 old_c1_count = get_c1_thread_count();
1283 new_c1_count = MIN4(_c1_count,
1284 _c1_compile_queue->size() / c1_tasks_per_thread,
1285 (int)(free_memory / (100*M)),
1286 (int)(available_cc_p / (128*K)));
1287
1288 for (int i = old_c1_count; i < new_c1_count; i++) {
1289 JavaThread *ct = make_thread(compiler_t, compiler1_object(i), _c1_compile_queue, _compilers[0], THREAD);
1290 if (ct == nullptr) break;
1291 _compilers[0]->set_num_compiler_threads(i + 1);
1292 if (trace_compiler_threads()) {
1293 ResourceMark rm;
1294 ThreadsListHandle tlh; // name() depends on the TLH.
1295 assert(tlh.includes(ct), "ct=" INTPTR_FORMAT " exited unexpectedly.", p2i(ct));
1296 stringStream msg;
1297 msg.print("Added compiler thread %s (free memory: %dMB, available profiled code cache: %dMB)",
1298 ct->name(), (int)(free_memory/M), (int)(available_cc_p/M));
1299 print_compiler_threads(msg);
1300 }
1301 }
1302 }
1303
1304 CompileThread_lock->unlock();
1305 }
1306
1307
1308 /**
1309 * Set the methods on the stack as on_stack so that redefine classes doesn't
1310 * reclaim them. This method is executed at a safepoint.
1311 */
1312 void CompileBroker::mark_on_stack() {
1313 assert(SafepointSynchronize::is_at_safepoint(), "sanity check");
1314 // Since we are at a safepoint, we do not need a lock to access
1315 // the compile queues.
1316 if (_c2_compile_queue != nullptr) {
1317 _c2_compile_queue->mark_on_stack();
1318 }
1319 if (_c1_compile_queue != nullptr) {
1320 _c1_compile_queue->mark_on_stack();
1321 }
1322 if (_ac1_compile_queue != nullptr) {
1323 _ac1_compile_queue->mark_on_stack();
1324 }
1325 if (_ac2_compile_queue != nullptr) {
1326 _ac2_compile_queue->mark_on_stack();
1327 }
1328 }
1329
1330 // ------------------------------------------------------------------
1331 // CompileBroker::compile_method
1332 //
1333 // Request compilation of a method.
1334 void CompileBroker::compile_method_base(const methodHandle& method,
1335 AOTCodeEntry* aot_code_entry,
1336 int osr_bci,
1337 int comp_level,
1338 int hot_count,
1339 CompileTask::CompileReason compile_reason,
1340 bool requires_online_compilation,
1341 bool blocking,
1342 Thread* thread) {
1343 guarantee(!method->is_abstract(), "cannot compile abstract methods");
1344 assert(method->method_holder()->is_instance_klass(),
1345 "sanity check");
1346 assert(!method->method_holder()->is_not_initialized() ||
1347 compile_reason == CompileTask::Reason_Preload ||
1348 compile_reason == CompileTask::Reason_Precompile ||
1349 compile_reason == CompileTask::Reason_PrecompileForPreload, "method holder must be initialized");
1350 assert(!method->is_method_handle_intrinsic(), "do not enqueue these guys");
1351
1352 if (CIPrintRequests) {
1353 tty->print("request: ");
1354 method->print_short_name(tty);
1355 if (osr_bci != InvocationEntryBci) {
1356 tty->print(" osr_bci: %d", osr_bci);
1357 }
1358 tty->print(" level: %d comment: %s count: %d", comp_level, CompileTask::reason_name(compile_reason), hot_count);
1359 if (hot_count > 0) {
1360 tty->print(" hot: yes");
1361 }
1362 tty->cr();
1363 }
1364 LogStreamHandle(Debug, aot, codecache, compilation) log;
1365 if (log.is_enabled()) {
1366 ResourceMark rm;
1367 MethodTrainingData* mtd = MethodTrainingData::have_data() ? MethodTrainingData::find_fast(method) : nullptr;
1368 MethodCounters* mc = method->method_counters();
1369 const char* name = method->name_and_sig_as_C_string();
1370 const char* aotn = (compile_reason == CompileTask::Reason_Preload) ? "AP" :
1371 (!requires_online_compilation ? " A" : " ");
1372 const char* osrn = (osr_bci != InvocationEntryBci) ? "% " : "";
1373 log.print("request: %s%d %16s %s%s", aotn, comp_level, CompileTask::reason_name(compile_reason), osrn, name);
1374 if (mtd != nullptr) {
1375 log.print(" (MTD invoke: %d, backedge: %d)", mtd->invocation_count(), mtd->backedge_count());
1376 }
1377 if (mc != nullptr) {
1378 log.print(" (MC invoke: %d, backedge: %d)", mc->invocation_counter()->count(),
1379 mc->backedge_counter()->count());
1380 #if INCLUDE_CDS
1381 log.print(" (MC aot_cnt: %d, aot_recomp: %s)", mc->aot_code_invocation_count(),
1382 (mc->aot_code_recompile_requested() ? "yes" : "no"));
1383 #endif
1384 }
1385 log.cr();
1386 }
1387 // A request has been made for compilation. Before we do any
1388 // real work, check to see if the method has been compiled
1389 // in the meantime with a definitive result.
1390 if (compilation_is_complete(method(), osr_bci, comp_level, requires_online_compilation, compile_reason)) {
1391 return;
1392 }
1393
1394 #ifndef PRODUCT
1395 if (osr_bci != -1 && !FLAG_IS_DEFAULT(OSROnlyBCI)) {
1396 if ((OSROnlyBCI > 0) ? (OSROnlyBCI != osr_bci) : (-OSROnlyBCI == osr_bci)) {
1397 // Positive OSROnlyBCI means only compile that bci. Negative means don't compile that BCI.
1398 return;
1399 }
1400 }
1401 #endif
1402
1403 // If this method is already in the compile queue, then
1404 // we do not block the current thread.
1405 if (compilation_is_in_queue(method)) {
1406 // We may want to decay our counter a bit here to prevent
1407 // multiple denied requests for compilation. This is an
1408 // open compilation policy issue. Note: The other possibility,
1409 // in the case that this is a blocking compile request, is to have
1410 // all subsequent blocking requesters wait for completion of
1411 // ongoing compiles. Note that in this case we'll need a protocol
1412 // for freeing the associated compile tasks. [Or we could have
1413 // a single static monitor on which all these waiters sleep.]
1414 return;
1415 }
1416
1417 // Tiered policy requires MethodCounters to exist before adding a method to
1418 // the queue. Create if we don't have them yet.
1419 method->get_method_counters(thread);
1420
1421 precond(compile_reason != CompileTask::Reason_Preload || aot_code_entry != nullptr);
1422 if (!requires_online_compilation && aot_code_entry == nullptr) {
1423 aot_code_entry = find_aot_code_entry(method, osr_bci, comp_level, compile_reason);
1424 }
1425 bool is_aot = (aot_code_entry != nullptr);
1426 requires_online_compilation = !is_aot; // Request JIT compilation
1427
1428 // Outputs from the following MutexLocker block:
1429 CompileTask* task = nullptr;
1430 CompileQueue* queue = compile_queue(comp_level, is_aot);
1431
1432 // Acquire our lock.
1433 {
1434 ConditionalMutexLocker locker(thread, queue->lock(), !UseLockFreeCompileQueues);
1435
1436 // Make sure the method has not slipped into the queues since
1437 // last we checked; note that those checks were "fast bail-outs".
1438 // Here we need to be more careful, see 14012000 below.
1439 if (compilation_is_in_queue(method)) {
1440 return;
1441 }
1442
1443 // We need to check again to see if the compilation has
1444 // completed. A previous compilation may have registered
1445 // some result.
1446 if (compilation_is_complete(method(), osr_bci, comp_level, requires_online_compilation, compile_reason)) {
1447 return;
1448 }
1449
1450 // We now know that this compilation is not pending, complete,
1451 // or prohibited. Assign a compile_id to this compilation
1452 // and check to see if it is in our [Start..Stop) range.
1453 int compile_id = assign_compile_id(method, osr_bci);
1454 if (compile_id == 0) {
1455 // The compilation falls outside the allowed range.
1456 return;
1457 }
1458
1459 #if INCLUDE_JVMCI
1460 if (UseJVMCICompiler && blocking) {
1461 // Don't allow blocking compiles for requests triggered by JVMCI.
1462 if (thread->is_Compiler_thread()) {
1463 blocking = false;
1464 }
1465
1466 // In libjvmci, JVMCI initialization should not deadlock with other threads
1467 if (!UseJVMCINativeLibrary) {
1468 // Don't allow blocking compiles if inside a class initializer or while performing class loading
1469 vframeStream vfst(JavaThread::cast(thread));
1470 for (; !vfst.at_end(); vfst.next()) {
1471 if (vfst.method()->is_static_initializer() ||
1472 (vfst.method()->method_holder()->is_subclass_of(vmClasses::ClassLoader_klass()) &&
1473 vfst.method()->name() == vmSymbols::loadClass_name())) {
1474 blocking = false;
1475 break;
1476 }
1477 }
1478
1479 // Don't allow blocking compilation requests to JVMCI
1480 // if JVMCI itself is not yet initialized
1481 if (!JVMCI::is_compiler_initialized() && compiler(comp_level)->is_jvmci()) {
1482 blocking = false;
1483 }
1484 }
1485
1486 // Don't allow blocking compilation requests if we are in JVMCIRuntime::shutdown
1487 // to avoid deadlock between compiler thread(s) and threads run at shutdown
1488 // such as the DestroyJavaVM thread.
1489 if (JVMCI::in_shutdown()) {
1490 blocking = false;
1491 }
1492 }
1493 #endif // INCLUDE_JVMCI
1494
1495 // We will enter the compilation in the queue.
1496 // 14012000: Note that this sets the queued_for_compile bits in
1497 // the target method. We can now reason that a method cannot be
1498 // queued for compilation more than once, as follows:
1499 // Before a thread queues a task for compilation, it first acquires
1500 // the compile queue lock, then checks if the method's queued bits
1501 // are set or it has already been compiled. Thus there can not be two
1502 // instances of a compilation task for the same method on the
1503 // compilation queue. Consider now the case where the compilation
1504 // thread has already removed a task for that method from the queue
1505 // and is in the midst of compiling it. In this case, the
1506 // queued_for_compile bits must be set in the method (and these
1507 // will be visible to the current thread, since the bits were set
1508 // under protection of the compile queue lock, which we hold now.
1509 // When the compilation completes, the compiler thread first sets
1510 // the compilation result and then clears the queued_for_compile
1511 // bits. Neither of these actions are protected by a barrier (or done
1512 // under the protection of a lock), so the only guarantee we have
1513 // (on machines with TSO (Total Store Order)) is that these values
1514 // will update in that order. As a result, the only combinations of
1515 // these bits that the current thread will see are, in temporal order:
1516 // <RESULT, QUEUE> :
1517 // <0, 1> : in compile queue, but not yet compiled
1518 // <1, 1> : compiled but queue bit not cleared
1519 // <1, 0> : compiled and queue bit cleared
1520 // Because we first check the queue bits then check the result bits,
1521 // we are assured that we cannot introduce a duplicate task.
1522 // Note that if we did the tests in the reverse order (i.e. check
1523 // result then check queued bit), we could get the result bit before
1524 // the compilation completed, and the queue bit after the compilation
1525 // completed, and end up introducing a "duplicate" (redundant) task.
1526 // In that case, the compiler thread should first check if a method
1527 // has already been compiled before trying to compile it.
1528 // NOTE: in the event that there are multiple compiler threads and
1529 // there is de-optimization/recompilation, things will get hairy,
1530 // and in that case it's best to protect both the testing (here) of
1531 // these bits, and their updating (here and elsewhere) under a
1532 // common lock.
1533 task = create_compile_task(queue,
1534 compile_id, method,
1535 osr_bci, comp_level,
1536 hot_count, aot_code_entry, compile_reason,
1537 requires_online_compilation, blocking);
1538
1539 if (UseLockFreeCompileQueues) {
1540 assert(queue->lock()->owned_by_self() == false, "");
1541 queue->add_pending(task);
1542 } else {
1543 queue->add(task);
1544 }
1545 }
1546
1547 if (blocking) {
1548 wait_for_completion(task);
1549 }
1550 }
1551
1552 AOTCodeEntry* CompileBroker::find_aot_code_entry(const methodHandle& method, int osr_bci, int comp_level,
1553 CompileTask::CompileReason compile_reason) {
1554 if (compile_reason == CompileTask::Reason_Whitebox) {
1555 return nullptr; // Need normal JIT compilation
1556 }
1557 AOTCodeEntry* aot_code_entry = nullptr;
1558 if (osr_bci == InvocationEntryBci && AOTCodeCache::is_using_code()) {
1559 // Check for AOT preload code first.
1560 precond(compile_reason != CompileTask::Reason_Preload); // should call it for AOT code preload.
1561 aot_code_entry = AOTCodeCache::find_code_entry(method, comp_level);
1562 }
1563 return aot_code_entry;
1564 }
1565
1566 void CompileBroker::preload_aot_method(const methodHandle& method, AOTCodeEntry* aot_code_entry, TRAPS) {
1567 // Don't need most of the checks for AOT code preloading.
1568 precond(_initialized);
1569 precond(aot_code_entry != nullptr && aot_code_entry->for_preload());
1570 // If the compiler is shut off due to code cache getting full
1571 // fail out now so blocking compiles dont hang the java thread
1572 if (should_compile_new_jobs()) {
1573 int osr_bci = InvocationEntryBci;
1574 int comp_level = CompLevel_full_optimization;
1575 int hot_count = 0;
1576 CompileTask::CompileReason compile_reason = CompileTask::Reason_Preload;
1577 bool requires_online_compilation = false;
1578
1579 AbstractCompiler *comp = CompileBroker::compiler(comp_level);
1580 assert(comp != nullptr, "Ensure we have a compiler");
1581
1582 nmethod* method_code = method->code();
1583 if (method_code != nullptr) {
1584 if (compilation_is_complete(method(), osr_bci, comp_level, requires_online_compilation, compile_reason)) {
1585 return;
1586 }
1587 }
1588 if (method->is_not_compilable(comp_level)) {
1589 return;
1590 }
1591
1592 // JVMTI -- post_compile_event requires jmethod_id() that may require
1593 // a lock the compiling thread can not acquire. Prefetch it here.
1594 if (JvmtiExport::should_post_compiled_method_load()) {
1595 method->jmethod_id();
1596 }
1597
1598 DirectiveSet* directive = DirectivesStack::getMatchingDirective(method, comp);
1599 bool is_blocking = ReplayCompiles ||
1600 !directive->BackgroundCompilationOption ||
1601 (PreloadBlocking && (compile_reason == CompileTask::Reason_Preload));
1602 // CompileBroker::compile_method can trap and can have pending async exception.
1603 compile_method_base(method, aot_code_entry, osr_bci, comp_level, hot_count, compile_reason,
1604 requires_online_compilation, is_blocking, THREAD);
1605 DirectivesStack::release(directive);
1606 }
1607 }
1608
1609 nmethod* CompileBroker::compile_method(const methodHandle& method, int osr_bci,
1610 int comp_level,
1611 int hot_count,
1612 bool requires_online_compilation,
1613 CompileTask::CompileReason compile_reason,
1614 TRAPS) {
1615 // Do nothing if compilebroker is not initialized or compiles are submitted on level none
1616 if (!_initialized || comp_level == CompLevel_none) {
1617 return nullptr;
1618 }
1619
1620 AbstractCompiler *comp = CompileBroker::compiler(comp_level);
1621 assert(comp != nullptr, "Ensure we have a compiler");
1622
1623 #if INCLUDE_JVMCI
1624 if (comp->is_jvmci() && !JVMCI::can_initialize_JVMCI()) {
1625 // JVMCI compilation is not yet initializable.
1626 return nullptr;
1627 }
1628 #endif
1629
1630 DirectiveSet* directive = DirectivesStack::getMatchingDirective(method, comp);
1631 // CompileBroker::compile_method can trap and can have pending async exception.
1632 nmethod* nm = CompileBroker::compile_method(method, osr_bci, comp_level, hot_count, requires_online_compilation, compile_reason, directive, THREAD);
1633 DirectivesStack::release(directive);
1634 return nm;
1635 }
1636
1637 nmethod* CompileBroker::compile_method(const methodHandle& method, int osr_bci,
1638 int comp_level,
1639 int hot_count,
1640 bool requires_online_compilation,
1641 CompileTask::CompileReason compile_reason,
1642 DirectiveSet* directive,
1643 TRAPS) {
1644
1645 // make sure arguments make sense
1646 assert(method->method_holder()->is_instance_klass(), "not an instance method");
1647 assert(osr_bci == InvocationEntryBci || (0 <= osr_bci && osr_bci < method->code_size()), "bci out of range");
1648 assert(!method->is_abstract() && (osr_bci == InvocationEntryBci || !method->is_native()), "cannot compile abstract/native methods");
1649 assert(!method->method_holder()->is_not_initialized() ||
1650 compile_reason == CompileTask::Reason_Preload ||
1651 compile_reason == CompileTask::Reason_Precompile ||
1652 compile_reason == CompileTask::Reason_PrecompileForPreload, "method holder must be initialized");
1653 // return quickly if possible
1654 bool aot_compilation = (PrecompileCode && PrecompileOnlyAndExit) ||
1655 CDSConfig::is_dumping_aot_code();
1656 if (aot_compilation && !CompileTask::reason_is_precompile(compile_reason)) {
1657 // Skip normal compilations when compiling AOT code
1658 return nullptr;
1659 }
1660
1661 // lock, make sure that the compilation
1662 // isn't prohibited in a straightforward way.
1663 AbstractCompiler* comp = CompileBroker::compiler(comp_level);
1664 if (comp == nullptr || compilation_is_prohibited(method, osr_bci, comp_level, directive->ExcludeOption)) {
1665 return nullptr;
1666 }
1667
1668 if (osr_bci == InvocationEntryBci) {
1669 // standard compilation
1670 nmethod* method_code = method->code();
1671 if (method_code != nullptr) {
1672 if (compilation_is_complete(method(), osr_bci, comp_level, requires_online_compilation, compile_reason)) {
1673 return method_code;
1674 }
1675 }
1676 if (method->is_not_compilable(comp_level)) {
1677 return nullptr;
1678 }
1679 } else {
1680 // osr compilation
1681 // We accept a higher level osr method
1682 nmethod* nm = method->lookup_osr_nmethod_for(osr_bci, comp_level, false);
1683 if (nm != nullptr) return nm;
1684 if (method->is_not_osr_compilable(comp_level)) return nullptr;
1685 }
1686
1687 assert(!HAS_PENDING_EXCEPTION, "No exception should be present");
1688 // some prerequisites that are compiler specific
1689 if (compile_reason != CompileTask::Reason_Preload &&
1690 !CompileTask::reason_is_precompile(compile_reason) &&
1691 (comp->is_c2() || comp->is_jvmci())) {
1692 InternalOOMEMark iom(THREAD);
1693 method->constants()->resolve_string_constants(CHECK_AND_CLEAR_NONASYNC_NULL);
1694 // Resolve all classes seen in the signature of the method
1695 // we are compiling.
1696 Method::load_signature_classes(method, CHECK_AND_CLEAR_NONASYNC_NULL);
1697 }
1698
1699 // If the method is native, do the lookup in the thread requesting
1700 // the compilation. Native lookups can load code, which is not
1701 // permitted during compilation.
1702 //
1703 // Note: A native method implies non-osr compilation which is
1704 // checked with an assertion at the entry of this method.
1705 if (method->is_native() && !method->is_method_handle_intrinsic()) {
1706 address adr = NativeLookup::lookup(method, THREAD);
1707 if (HAS_PENDING_EXCEPTION) {
1708 // In case of an exception looking up the method, we just forget
1709 // about it. The interpreter will kick-in and throw the exception.
1710 method->set_not_compilable("NativeLookup::lookup failed"); // implies is_not_osr_compilable()
1711 CLEAR_PENDING_EXCEPTION;
1712 return nullptr;
1713 }
1714 assert(method->has_native_function(), "must have native code by now");
1715 }
1716
1717 // RedefineClasses() has replaced this method; just return
1718 if (method->is_old()) {
1719 return nullptr;
1720 }
1721
1722 // JVMTI -- post_compile_event requires jmethod_id() that may require
1723 // a lock the compiling thread can not acquire. Prefetch it here.
1724 if (JvmtiExport::should_post_compiled_method_load()) {
1725 method->jmethod_id();
1726 }
1727
1728 // do the compilation
1729 if (method->is_native()) {
1730 if (!PreferInterpreterNativeStubs || method->is_method_handle_intrinsic()) {
1731 // To properly handle the appendix argument for out-of-line calls we are using a small trampoline that
1732 // pops off the appendix argument and jumps to the target (see gen_special_dispatch in SharedRuntime).
1733 //
1734 // Since normal compiled-to-compiled calls are not able to handle such a thing we MUST generate an adapter
1735 // in this case. If we can't generate one and use it we can not execute the out-of-line method handle calls.
1736 AdapterHandlerLibrary::create_native_wrapper(method);
1737 } else {
1738 return nullptr;
1739 }
1740 } else {
1741 // If the compiler is shut off due to code cache getting full
1742 // fail out now so blocking compiles dont hang the java thread
1743 if (!should_compile_new_jobs()) {
1744 return nullptr;
1745 }
1746 bool is_blocking = ReplayCompiles ||
1747 !directive->BackgroundCompilationOption ||
1748 (PreloadBlocking && (compile_reason == CompileTask::Reason_Preload));
1749 compile_method_base(method, nullptr, osr_bci, comp_level, hot_count, compile_reason,
1750 requires_online_compilation, is_blocking, THREAD);
1751 }
1752
1753 // return requested nmethod
1754 // We accept a higher level osr method
1755 if (osr_bci == InvocationEntryBci) {
1756 return method->code();
1757 }
1758 return method->lookup_osr_nmethod_for(osr_bci, comp_level, false);
1759 }
1760
1761
1762 // ------------------------------------------------------------------
1763 // CompileBroker::compilation_is_complete
1764 //
1765 // See if compilation of this method is already complete.
1766 bool CompileBroker::compilation_is_complete(Method* method,
1767 int osr_bci,
1768 int comp_level,
1769 bool online_only,
1770 CompileTask::CompileReason compile_reason) {
1771 if (compile_reason == CompileTask::Reason_Precompile ||
1772 compile_reason == CompileTask::Reason_PrecompileForPreload) {
1773 return false; // FIXME: any restrictions?
1774 }
1775 bool is_osr = (osr_bci != standard_entry_bci);
1776 if (is_osr) {
1777 if (method->is_not_osr_compilable(comp_level)) {
1778 return true;
1779 } else {
1780 nmethod* result = method->lookup_osr_nmethod_for(osr_bci, comp_level, true);
1781 return (result != nullptr);
1782 }
1783 } else {
1784 if (method->is_not_compilable(comp_level)) {
1785 return true;
1786 } else {
1787 nmethod* result = method->code();
1788 if (result == nullptr) {
1789 return false;
1790 }
1791 if (online_only && result->is_aot()) {
1792 return false;
1793 }
1794 bool same_level = (comp_level == result->comp_level());
1795 if (result->preloaded()) {
1796 return !same_level; // Allow replace preloaded code with new code of the same level
1797 }
1798 return same_level;
1799 }
1800 }
1801 }
1802
1803
1804 /**
1805 * See if this compilation is already requested.
1806 *
1807 * Implementation note: there is only a single "is in queue" bit
1808 * for each method. This means that the check below is overly
1809 * conservative in the sense that an osr compilation in the queue
1810 * will block a normal compilation from entering the queue (and vice
1811 * versa). This can be remedied by a full queue search to disambiguate
1812 * cases. If it is deemed profitable, this may be done.
1813 */
1814 bool CompileBroker::compilation_is_in_queue(const methodHandle& method) {
1815 return method->queued_for_compilation();
1816 }
1817
1818 // ------------------------------------------------------------------
1819 // CompileBroker::compilation_is_prohibited
1820 //
1821 // See if this compilation is not allowed.
1822 bool CompileBroker::compilation_is_prohibited(const methodHandle& method, int osr_bci, int comp_level, bool excluded) {
1823 bool is_native = method->is_native();
1824 // Some compilers may not support the compilation of natives.
1825 AbstractCompiler *comp = compiler(comp_level);
1826 if (is_native && (!CICompileNatives || comp == nullptr)) {
1827 method->set_not_compilable_quietly("native methods not supported", comp_level);
1828 return true;
1829 }
1830
1831 bool is_osr = (osr_bci != standard_entry_bci);
1832 // Some compilers may not support on stack replacement.
1833 if (is_osr && (!CICompileOSR || comp == nullptr)) {
1834 method->set_not_osr_compilable("OSR not supported", comp_level);
1835 return true;
1836 }
1837
1838 // The method may be explicitly excluded by the user.
1839 double scale;
1840 if (excluded || (CompilerOracle::has_option_value(method, CompileCommandEnum::CompileThresholdScaling, scale) && scale == 0)) {
1841 bool quietly = CompilerOracle::be_quiet();
1842 if (PrintCompilation && !quietly) {
1843 // This does not happen quietly...
1844 ResourceMark rm;
1845 tty->print("### Excluding %s:%s",
1846 method->is_native() ? "generation of native wrapper" : "compile",
1847 (method->is_static() ? " static" : ""));
1848 method->print_short_name(tty);
1849 tty->cr();
1850 }
1851 method->set_not_compilable("excluded by CompileCommand", comp_level, !quietly);
1852 }
1853
1854 return false;
1855 }
1856
1857 /**
1858 * Generate serialized IDs for compilation requests. If certain debugging flags are used
1859 * and the ID is not within the specified range, the method is not compiled and 0 is returned.
1860 * The function also allows to generate separate compilation IDs for OSR compilations.
1861 */
1862 int CompileBroker::assign_compile_id(const methodHandle& method, int osr_bci) {
1863 #ifdef ASSERT
1864 bool is_osr = (osr_bci != standard_entry_bci);
1865 int id;
1866 if (method->is_native()) {
1867 assert(!is_osr, "can't be osr");
1868 // Adapters, native wrappers and method handle intrinsics
1869 // should be generated always.
1870 return AtomicAccess::add(CICountNative ? &_native_compilation_id : &_compilation_id, 1);
1871 } else if (CICountOSR && is_osr) {
1872 id = AtomicAccess::add(&_osr_compilation_id, 1);
1873 if (CIStartOSR <= id && id < CIStopOSR) {
1874 return id;
1875 }
1876 } else {
1877 id = AtomicAccess::add(&_compilation_id, 1);
1878 if (CIStart <= id && id < CIStop) {
1879 return id;
1880 }
1881 }
1882
1883 // Method was not in the appropriate compilation range.
1884 method->set_not_compilable_quietly("Not in requested compile id range");
1885 return 0;
1886 #else
1887 // CICountOSR is a develop flag and set to 'false' by default. In a product built,
1888 // only _compilation_id is incremented.
1889 return AtomicAccess::add(&_compilation_id, 1);
1890 #endif
1891 }
1892
1893 // ------------------------------------------------------------------
1894 // CompileBroker::assign_compile_id_unlocked
1895 //
1896 // Public wrapper for assign_compile_id that acquires the needed locks
1897 int CompileBroker::assign_compile_id_unlocked(Thread* thread, const methodHandle& method, int osr_bci) {
1898 return assign_compile_id(method, osr_bci);
1899 }
1900
1901 // ------------------------------------------------------------------
1902 // CompileBroker::create_compile_task
1903 //
1904 // Create a CompileTask object representing the current request for
1905 // compilation. Add this task to the queue.
1906 CompileTask* CompileBroker::create_compile_task(CompileQueue* queue,
1907 int compile_id,
1908 const methodHandle& method,
1909 int osr_bci,
1910 int comp_level,
1911 int hot_count,
1912 AOTCodeEntry* aot_code_entry,
1913 CompileTask::CompileReason compile_reason,
1914 bool requires_online_compilation,
1915 bool blocking) {
1916 CompileTask* new_task = new CompileTask(compile_id, method, osr_bci, comp_level,
1917 hot_count, aot_code_entry, compile_reason, queue,
1918 requires_online_compilation, blocking);
1919 return new_task;
1920 }
1921
1922 #if INCLUDE_JVMCI
1923 // The number of milliseconds to wait before checking if
1924 // JVMCI compilation has made progress.
1925 static const long JVMCI_COMPILATION_PROGRESS_WAIT_TIMESLICE = 1000;
1926
1927 // The number of JVMCI compilation progress checks that must fail
1928 // before unblocking a thread waiting for a blocking compilation.
1929 static const int JVMCI_COMPILATION_PROGRESS_WAIT_ATTEMPTS = 10;
1930
1931 /**
1932 * Waits for a JVMCI compiler to complete a given task. This thread
1933 * waits until either the task completes or it sees no JVMCI compilation
1934 * progress for N consecutive milliseconds where N is
1935 * JVMCI_COMPILATION_PROGRESS_WAIT_TIMESLICE *
1936 * JVMCI_COMPILATION_PROGRESS_WAIT_ATTEMPTS.
1937 *
1938 * @return true if this thread needs to delete the task
1939 */
1940 bool CompileBroker::wait_for_jvmci_completion(JVMCICompiler* jvmci, CompileTask* task, JavaThread* thread) {
1941 assert(UseJVMCICompiler, "sanity");
1942 MonitorLocker ml(thread, CompileTaskWait_lock);
1943 int progress_wait_attempts = 0;
1944 jint thread_jvmci_compilation_ticks = 0;
1945 jint global_jvmci_compilation_ticks = jvmci->global_compilation_ticks();
1946 while (!task->is_complete() && !is_compilation_disabled_forever() &&
1947 ml.wait(JVMCI_COMPILATION_PROGRESS_WAIT_TIMESLICE)) {
1948 JVMCICompileState* jvmci_compile_state = task->blocking_jvmci_compile_state();
1949
1950 bool progress;
1951 if (jvmci_compile_state != nullptr) {
1952 jint ticks = jvmci_compile_state->compilation_ticks();
1953 progress = (ticks - thread_jvmci_compilation_ticks) != 0;
1954 JVMCI_event_1("waiting on compilation %d [ticks=%d]", task->compile_id(), ticks);
1955 thread_jvmci_compilation_ticks = ticks;
1956 } else {
1957 // Still waiting on JVMCI compiler queue. This thread may be holding a lock
1958 // that all JVMCI compiler threads are blocked on. We use the global JVMCI
1959 // compilation ticks to determine whether JVMCI compilation
1960 // is still making progress through the JVMCI compiler queue.
1961 jint ticks = jvmci->global_compilation_ticks();
1962 progress = (ticks - global_jvmci_compilation_ticks) != 0;
1963 JVMCI_event_1("waiting on compilation %d to be queued [ticks=%d]", task->compile_id(), ticks);
1964 global_jvmci_compilation_ticks = ticks;
1965 }
1966
1967 if (!progress) {
1968 if (++progress_wait_attempts == JVMCI_COMPILATION_PROGRESS_WAIT_ATTEMPTS) {
1969 if (PrintCompilation) {
1970 task->print(tty, "wait for blocking compilation timed out");
1971 }
1972 JVMCI_event_1("waiting on compilation %d timed out", task->compile_id());
1973 break;
1974 }
1975 } else {
1976 progress_wait_attempts = 0;
1977 }
1978 }
1979 task->clear_waiter();
1980 return task->is_complete();
1981 }
1982 #endif
1983
1984 /**
1985 * Wait for the compilation task to complete.
1986 */
1987 void CompileBroker::wait_for_completion(CompileTask* task) {
1988 if (CIPrintCompileQueue) {
1989 ttyLocker ttyl;
1990 tty->print_cr("BLOCKING FOR COMPILE");
1991 }
1992
1993 assert(task->is_blocking(), "can only wait on blocking task");
1994
1995 JavaThread* thread = JavaThread::current();
1996
1997 methodHandle method(thread, task->method());
1998 bool free_task;
1999 #if INCLUDE_JVMCI
2000 AbstractCompiler* comp = compiler(task->comp_level());
2001 if (!UseJVMCINativeLibrary && comp->is_jvmci() && !task->should_wait_for_compilation()) {
2002 // It may return before compilation is completed.
2003 // Note that libjvmci should not pre-emptively unblock
2004 // a thread waiting for a compilation as it does not call
2005 // Java code and so is not deadlock prone like jarjvmci.
2006 free_task = wait_for_jvmci_completion((JVMCICompiler*) comp, task, thread);
2007 } else
2008 #endif
2009 {
2010 free_task = true;
2011 // Wait until the task is complete or compilation is shut down.
2012 MonitorLocker ml(thread, CompileTaskWait_lock);
2013 while (!task->is_complete() && !is_compilation_disabled_forever()) {
2014 ml.wait();
2015 }
2016 }
2017
2018 // It is harmless to check this status without the lock, because
2019 // completion is a stable property.
2020 if (!task->is_complete()) {
2021 // Task is not complete, likely because we are exiting for compilation
2022 // shutdown. The task can still be reached through the queue, or executed
2023 // by some compiler thread. There is no coordination with either MCQ lock
2024 // holders or compilers, therefore we cannot delete the task.
2025 //
2026 // This will leave task allocated, which leaks it. At this (degraded) point,
2027 // it is less risky to abandon the task, rather than attempting a more
2028 // complicated deletion protocol.
2029 free_task = false;
2030 }
2031
2032 if (free_task) {
2033 assert(task->is_complete(), "Compilation should have completed");
2034 assert(task->next() == nullptr && task->prev() == nullptr,
2035 "Completed task should not be in the queue");
2036
2037 // By convention, the waiter is responsible for deleting a
2038 // blocking CompileTask. Since there is only one waiter ever
2039 // waiting on a CompileTask, we know that no one else will
2040 // be using this CompileTask; we can delete it.
2041 delete task;
2042 }
2043 }
2044
2045 void CompileBroker::wait_for_no_active_tasks() {
2046 CompileTask::wait_for_no_active_tasks();
2047 }
2048
2049 /**
2050 * Initialize compiler thread(s) + compiler object(s). The postcondition
2051 * of this function is that the compiler runtimes are initialized and that
2052 * compiler threads can start compiling.
2053 */
2054 bool CompileBroker::init_compiler_runtime() {
2055 CompilerThread* thread = CompilerThread::current();
2056 AbstractCompiler* comp = thread->compiler();
2057 // Final sanity check - the compiler object must exist
2058 guarantee(comp != nullptr, "Compiler object must exist");
2059
2060 {
2061 // Must switch to native to allocate ci_env
2062 ThreadToNativeFromVM ttn(thread);
2063 ciEnv ci_env((CompileTask*)nullptr);
2064 // Cache Jvmti state
2065 ci_env.cache_jvmti_state();
2066 // Cache DTrace flags
2067 ci_env.cache_dtrace_flags();
2068
2069 // Switch back to VM state to do compiler initialization
2070 ThreadInVMfromNative tv(thread);
2071
2072 comp->initialize();
2073 }
2074
2075 if (comp->is_failed()) {
2076 disable_compilation_forever();
2077 // If compiler initialization failed, no compiler thread that is specific to a
2078 // particular compiler runtime will ever start to compile methods.
2079 shutdown_compiler_runtime(comp, thread);
2080 return false;
2081 }
2082
2083 // C1 specific check
2084 if (comp->is_c1() && (thread->get_buffer_blob() == nullptr)) {
2085 warning("Initialization of %s thread failed (no space to run compilers)", thread->name());
2086 return false;
2087 }
2088
2089 return true;
2090 }
2091
2092 void CompileBroker::free_buffer_blob_if_allocated(CompilerThread* thread) {
2093 BufferBlob* blob = thread->get_buffer_blob();
2094 if (blob != nullptr) {
2095 blob->purge();
2096 MutexLocker mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
2097 CodeCache::free(blob);
2098 }
2099 }
2100
2101 /**
2102 * If C1 and/or C2 initialization failed, we shut down all compilation.
2103 * We do this to keep things simple. This can be changed if it ever turns
2104 * out to be a problem.
2105 */
2106 void CompileBroker::shutdown_compiler_runtime(AbstractCompiler* comp, CompilerThread* thread) {
2107 free_buffer_blob_if_allocated(thread);
2108
2109 log_info(compilation)("shutdown_compiler_runtime: " INTPTR_FORMAT, p2i(thread));
2110
2111 if (comp->should_perform_shutdown()) {
2112 // There are two reasons for shutting down the compiler
2113 // 1) compiler runtime initialization failed
2114 // 2) The code cache is full and the following flag is set: -XX:-UseCodeCacheFlushing
2115 warning("%s initialization failed. Shutting down all compilers", comp->name());
2116
2117 // Only one thread per compiler runtime object enters here
2118 // Set state to shut down
2119 comp->set_shut_down();
2120
2121 // Delete all queued compilation tasks to make compiler threads exit faster.
2122 if (_c1_compile_queue != nullptr) {
2123 _c1_compile_queue->delete_all();
2124 }
2125
2126 if (_c2_compile_queue != nullptr) {
2127 _c2_compile_queue->delete_all();
2128 }
2129
2130 // Set flags so that we continue execution with using interpreter only.
2131 UseCompiler = false;
2132 UseInterpreter = true;
2133
2134 // We could delete compiler runtimes also. However, there are references to
2135 // the compiler runtime(s) (e.g., nmethod::is_compiled_by_c1()) which then
2136 // fail. This can be done later if necessary.
2137 }
2138 }
2139
2140 /**
2141 * Helper function to create new or reuse old CompileLog.
2142 */
2143 CompileLog* CompileBroker::get_log(CompilerThread* ct) {
2144 if (!LogCompilation) return nullptr;
2145
2146 AbstractCompiler *compiler = ct->compiler();
2147 bool c1 = compiler->is_c1();
2148 jobject* compiler_objects = c1 ? _compiler1_objects : _compiler2_objects;
2149 assert(compiler_objects != nullptr, "must be initialized at this point");
2150 CompileLog** logs = c1 ? _compiler1_logs : _compiler2_logs;
2151 assert(logs != nullptr, "must be initialized at this point");
2152 int count = c1 ? _c1_count : _c2_count;
2153
2154 if (ct->queue() == _ac1_compile_queue || ct->queue() == _ac2_compile_queue) {
2155 compiler_objects = _ac_objects;
2156 logs = _ac_logs;
2157 count = _ac_count;
2158 }
2159 // Find Compiler number by its threadObj.
2160 oop compiler_obj = ct->threadObj();
2161 int compiler_number = 0;
2162 bool found = false;
2163 for (; compiler_number < count; compiler_number++) {
2164 if (JNIHandles::resolve_non_null(compiler_objects[compiler_number]) == compiler_obj) {
2165 found = true;
2166 break;
2167 }
2168 }
2169 assert(found, "Compiler must exist at this point");
2170
2171 // Determine pointer for this thread's log.
2172 CompileLog** log_ptr = &logs[compiler_number];
2173
2174 // Return old one if it exists.
2175 CompileLog* log = *log_ptr;
2176 if (log != nullptr) {
2177 ct->init_log(log);
2178 return log;
2179 }
2180
2181 // Create a new one and remember it.
2182 init_compiler_thread_log();
2183 log = ct->log();
2184 *log_ptr = log;
2185 return log;
2186 }
2187
2188 // ------------------------------------------------------------------
2189 // CompileBroker::compiler_thread_loop
2190 //
2191 // The main loop run by a CompilerThread.
2192 void CompileBroker::compiler_thread_loop() {
2193 CompilerThread* thread = CompilerThread::current();
2194 CompileQueue* queue = thread->queue();
2195 // For the thread that initializes the ciObjectFactory
2196 // this resource mark holds all the shared objects
2197 ResourceMark rm;
2198
2199 // First thread to get here will initialize the compiler interface
2200
2201 {
2202 ASSERT_IN_VM;
2203 MutexLocker only_one (thread, CompileThread_lock);
2204 if (!ciObjectFactory::is_initialized()) {
2205 ciObjectFactory::initialize();
2206 }
2207 }
2208
2209 // Open a log.
2210 CompileLog* log = get_log(thread);
2211 if (log != nullptr) {
2212 log->begin_elem("start_compile_thread name='%s' thread='%zu' process='%d'",
2213 thread->name(),
2214 os::current_thread_id(),
2215 os::current_process_id());
2216 log->stamp();
2217 log->end_elem();
2218 }
2219
2220 if (!thread->init_compilation_timeout()) {
2221 return;
2222 }
2223
2224 // If compiler thread/runtime initialization fails, exit the compiler thread
2225 if (!init_compiler_runtime()) {
2226 return;
2227 }
2228
2229 thread->start_idle_timer();
2230
2231 // Poll for new compilation tasks as long as the JVM runs. Compilation
2232 // should only be disabled if something went wrong while initializing the
2233 // compiler runtimes. This, in turn, should not happen. The only known case
2234 // when compiler runtime initialization fails is if there is not enough free
2235 // space in the code cache to generate the necessary stubs, etc.
2236 while (!is_compilation_disabled_forever()) {
2237 // We need this HandleMark to avoid leaking VM handles.
2238 HandleMark hm(thread);
2239
2240 RecompilationPolicy::recompilation_step(AOTRecompilationWorkUnitSize, thread);
2241
2242 CompileTask* task = queue->get(thread);
2243 if (task == nullptr) {
2244 if (UseDynamicNumberOfCompilerThreads) {
2245 // Access compiler_count under lock to enforce consistency.
2246 MutexLocker only_one(CompileThread_lock);
2247 if (can_remove(thread, true)) {
2248 if (trace_compiler_threads()) {
2249 ResourceMark rm;
2250 stringStream msg;
2251 msg.print("Removing compiler thread %s after " JLONG_FORMAT " ms idle time",
2252 thread->name(), thread->idle_time_millis());
2253 print_compiler_threads(msg);
2254 }
2255
2256 // Notify compiler that the compiler thread is about to stop
2257 thread->compiler()->stopping_compiler_thread(thread);
2258
2259 free_buffer_blob_if_allocated(thread);
2260 return; // Stop this thread.
2261 }
2262 }
2263 } else {
2264 // Assign the task to the current thread. Mark this compilation
2265 // thread as active for the profiler.
2266 // CompileTaskWrapper also keeps the Method* from being deallocated if redefinition
2267 // occurs after fetching the compile task off the queue.
2268 CompileTaskWrapper ctw(task);
2269 methodHandle method(thread, task->method());
2270
2271 // Never compile a method if breakpoints are present in it
2272 if (method()->number_of_breakpoints() == 0) {
2273 // Compile the method.
2274 if ((UseCompiler || AlwaysCompileLoopMethods) && CompileBroker::should_compile_new_jobs()) {
2275 invoke_compiler_on_method(task);
2276 thread->start_idle_timer();
2277 } else {
2278 // After compilation is disabled, remove remaining methods from queue
2279 method->clear_queued_for_compilation();
2280 method->set_pending_queue_processed(false);
2281 task->set_failure_reason("compilation is disabled");
2282 }
2283 } else {
2284 task->set_failure_reason("breakpoints are present");
2285 }
2286
2287 // Don't use AOT compielr threads for dynamic C1 and C2 threads creation.
2288 if (UseDynamicNumberOfCompilerThreads &&
2289 (queue == _c1_compile_queue || queue == _c2_compile_queue)) {
2290 possibly_add_compiler_threads(thread);
2291 assert(!thread->has_pending_exception(), "should have been handled");
2292 }
2293 }
2294 }
2295
2296 // Shut down compiler runtime
2297 shutdown_compiler_runtime(thread->compiler(), thread);
2298 }
2299
2300 // ------------------------------------------------------------------
2301 // CompileBroker::init_compiler_thread_log
2302 //
2303 // Set up state required by +LogCompilation.
2304 void CompileBroker::init_compiler_thread_log() {
2305 CompilerThread* thread = CompilerThread::current();
2306 char file_name[4*K];
2307 FILE* fp = nullptr;
2308 intx thread_id = os::current_thread_id();
2309 for (int try_temp_dir = 1; try_temp_dir >= 0; try_temp_dir--) {
2310 const char* dir = (try_temp_dir ? os::get_temp_directory() : nullptr);
2311 if (dir == nullptr) {
2312 jio_snprintf(file_name, sizeof(file_name), "hs_c%zu_pid%u.log",
2313 thread_id, os::current_process_id());
2314 } else {
2315 jio_snprintf(file_name, sizeof(file_name),
2316 "%s%shs_c%zu_pid%u.log", dir,
2317 os::file_separator(), thread_id, os::current_process_id());
2318 }
2319
2320 fp = os::fopen(file_name, "wt");
2321 if (fp != nullptr) {
2322 if (LogCompilation && Verbose) {
2323 tty->print_cr("Opening compilation log %s", file_name);
2324 }
2325 CompileLog* log = new(mtCompiler) CompileLog(file_name, fp, thread_id);
2326 if (log == nullptr) {
2327 fclose(fp);
2328 return;
2329 }
2330 thread->init_log(log);
2331
2332 if (xtty != nullptr) {
2333 ttyLocker ttyl;
2334 // Record any per thread log files
2335 xtty->elem("thread_logfile thread='%zd' filename='%s'", thread_id, file_name);
2336 }
2337 return;
2338 }
2339 }
2340 warning("Cannot open log file: %s", file_name);
2341 }
2342
2343 void CompileBroker::log_metaspace_failure() {
2344 const char* message = "some methods may not be compiled because metaspace "
2345 "is out of memory";
2346 if (CompilationLog::log() != nullptr) {
2347 CompilationLog::log()->log_metaspace_failure(message);
2348 }
2349 if (PrintCompilation) {
2350 tty->print_cr("COMPILE PROFILING SKIPPED: %s", message);
2351 }
2352 }
2353
2354
2355 // ------------------------------------------------------------------
2356 // CompileBroker::set_should_block
2357 //
2358 // Set _should_block.
2359 // Call this from the VM, with Threads_lock held and a safepoint requested.
2360 void CompileBroker::set_should_block() {
2361 assert(Threads_lock->owner() == Thread::current(), "must have threads lock");
2362 assert(SafepointSynchronize::is_at_safepoint(), "must be at a safepoint already");
2363 #ifndef PRODUCT
2364 if (PrintCompilation && (Verbose || WizardMode))
2365 tty->print_cr("notifying compiler thread pool to block");
2366 #endif
2367 _should_block = true;
2368 }
2369
2370 // ------------------------------------------------------------------
2371 // CompileBroker::maybe_block
2372 //
2373 // Call this from the compiler at convenient points, to poll for _should_block.
2374 void CompileBroker::maybe_block() {
2375 if (_should_block) {
2376 #ifndef PRODUCT
2377 if (PrintCompilation && (Verbose || WizardMode))
2378 tty->print_cr("compiler thread " INTPTR_FORMAT " poll detects block request", p2i(Thread::current()));
2379 #endif
2380 // If we are executing a task during the request to block, report the task
2381 // before disappearing.
2382 CompilerThread* thread = CompilerThread::current();
2383 if (thread != nullptr) {
2384 CompileTask* task = thread->task();
2385 if (task != nullptr) {
2386 if (PrintCompilation) {
2387 task->print(tty, "blocked");
2388 }
2389 task->print_ul("blocked");
2390 }
2391 }
2392 // Go to VM state and block for final VM shutdown safepoint.
2393 ThreadInVMfromNative tivfn(JavaThread::current());
2394 assert(false, "Should never unblock from TIVNM entry");
2395 }
2396 }
2397
2398 // wrapper for CodeCache::print_summary()
2399 static void codecache_print(bool detailed)
2400 {
2401 stringStream s;
2402 // Dump code cache into a buffer before locking the tty,
2403 {
2404 MutexLocker mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
2405 CodeCache::print_summary(&s, detailed);
2406 }
2407 ttyLocker ttyl;
2408 tty->print("%s", s.freeze());
2409 }
2410
2411 // wrapper for CodeCache::print_summary() using outputStream
2412 static void codecache_print(outputStream* out, bool detailed) {
2413 stringStream s;
2414
2415 // Dump code cache into a buffer
2416 {
2417 MutexLocker mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
2418 CodeCache::print_summary(&s, detailed);
2419 }
2420
2421 char* remaining_log = s.as_string();
2422 while (*remaining_log != '\0') {
2423 char* eol = strchr(remaining_log, '\n');
2424 if (eol == nullptr) {
2425 out->print_cr("%s", remaining_log);
2426 remaining_log = remaining_log + strlen(remaining_log);
2427 } else {
2428 *eol = '\0';
2429 out->print_cr("%s", remaining_log);
2430 remaining_log = eol + 1;
2431 }
2432 }
2433 }
2434
2435 void CompileBroker::handle_compile_error(CompilerThread* thread, CompileTask* task, ciEnv* ci_env,
2436 int compilable, const char* failure_reason) {
2437 if (!AbortVMOnCompilationFailure) {
2438 return;
2439 }
2440 if (compilable == ciEnv::MethodCompilable_not_at_tier) {
2441 fatal("Not compilable at tier %d: %s", task->comp_level(), failure_reason);
2442 }
2443 if (compilable == ciEnv::MethodCompilable_never) {
2444 fatal("Never compilable: %s", failure_reason);
2445 }
2446 }
2447
2448 static void post_compilation_event(EventCompilation& event, CompileTask* task) {
2449 assert(task != nullptr, "invariant");
2450 CompilerEvent::CompilationEvent::post(event,
2451 task->compile_id(),
2452 task->compiler()->type(),
2453 task->method(),
2454 task->comp_level(),
2455 task->is_success(),
2456 task->osr_bci() != CompileBroker::standard_entry_bci,
2457 task->nm_total_size(),
2458 task->num_inlined_bytecodes(),
2459 task->arena_bytes());
2460 }
2461
2462 int DirectivesStack::_depth = 0;
2463 CompilerDirectives* DirectivesStack::_top = nullptr;
2464 CompilerDirectives* DirectivesStack::_bottom = nullptr;
2465
2466 // Acquires Compilation_lock and waits for it to be notified
2467 // as long as WhiteBox::compilation_locked is true.
2468 static void whitebox_lock_compilation() {
2469 MonitorLocker locker(Compilation_lock, Mutex::_no_safepoint_check_flag);
2470 while (WhiteBox::compilation_locked) {
2471 locker.wait();
2472 }
2473 }
2474
2475 // ------------------------------------------------------------------
2476 // CompileBroker::invoke_compiler_on_method
2477 //
2478 // Compile a method.
2479 //
2480 void CompileBroker::invoke_compiler_on_method(CompileTask* task) {
2481 task->print_ul();
2482 elapsedTimer time;
2483
2484 DirectiveSet* directive = task->directive();
2485
2486 CompilerThread* thread = CompilerThread::current();
2487 ResourceMark rm(thread);
2488
2489 if (CompilationLog::log() != nullptr) {
2490 CompilationLog::log()->log_compile(thread, task);
2491 }
2492
2493 // Common flags.
2494 int compile_id = task->compile_id();
2495 int osr_bci = task->osr_bci();
2496 bool is_osr = (osr_bci != standard_entry_bci);
2497 bool should_log = (thread->log() != nullptr);
2498 bool should_break = false;
2499 bool should_print_compilation = PrintCompilation || directive->PrintCompilationOption;
2500 const int task_level = task->comp_level();
2501 AbstractCompiler* comp = task->compiler();
2502 {
2503 // create the handle inside it's own block so it can't
2504 // accidentally be referenced once the thread transitions to
2505 // native. The NoHandleMark before the transition should catch
2506 // any cases where this occurs in the future.
2507 methodHandle method(thread, task->method());
2508
2509 assert(!method->is_native(), "no longer compile natives");
2510
2511 // Update compile information when using perfdata.
2512 if (UsePerfData) {
2513 update_compile_perf_data(thread, method, is_osr);
2514 }
2515
2516 DTRACE_METHOD_COMPILE_BEGIN_PROBE(method, compiler_name(task_level));
2517 }
2518
2519 should_break = directive->BreakAtCompileOption || task->check_break_at_flags();
2520 if (should_log && !directive->LogOption) {
2521 should_log = false;
2522 }
2523
2524 // Allocate a new set of JNI handles.
2525 JNIHandleMark jhm(thread);
2526 Method* target_handle = task->method();
2527 int compilable = ciEnv::MethodCompilable;
2528 const char* failure_reason = nullptr;
2529 bool failure_reason_on_C_heap = false;
2530 const char* retry_message = nullptr;
2531
2532 #if INCLUDE_JVMCI
2533 if (UseJVMCICompiler && comp != nullptr && comp->is_jvmci()) {
2534 JVMCICompiler* jvmci = (JVMCICompiler*) comp;
2535
2536 TraceTime t1("compilation", &time);
2537 EventCompilation event;
2538 JVMCICompileState compile_state(task, jvmci);
2539 JVMCIRuntime *runtime = nullptr;
2540
2541 if (JVMCI::in_shutdown()) {
2542 failure_reason = "in JVMCI shutdown";
2543 retry_message = "not retryable";
2544 compilable = ciEnv::MethodCompilable_never;
2545 } else if (compile_state.target_method_is_old()) {
2546 // Skip redefined methods
2547 failure_reason = "redefined method";
2548 retry_message = "not retryable";
2549 compilable = ciEnv::MethodCompilable_never;
2550 } else {
2551 JVMCIEnv env(thread, &compile_state, __FILE__, __LINE__);
2552 if (env.init_error() != JNI_OK) {
2553 const char* msg = env.init_error_msg();
2554 failure_reason = os::strdup(err_msg("Error attaching to libjvmci (err: %d, %s)",
2555 env.init_error(), msg == nullptr ? "unknown" : msg), mtJVMCI);
2556 bool reason_on_C_heap = true;
2557 // In case of JNI_ENOMEM, there's a good chance a subsequent attempt to create libjvmci or attach to it
2558 // might succeed. Other errors most likely indicate a non-recoverable error in the JVMCI runtime.
2559 bool retryable = env.init_error() == JNI_ENOMEM;
2560 compile_state.set_failure(retryable, failure_reason, reason_on_C_heap);
2561 }
2562 if (failure_reason == nullptr) {
2563 if (WhiteBoxAPI && WhiteBox::compilation_locked) {
2564 // Must switch to native to block
2565 ThreadToNativeFromVM ttn(thread);
2566 whitebox_lock_compilation();
2567 }
2568 methodHandle method(thread, target_handle);
2569 runtime = env.runtime();
2570 runtime->compile_method(&env, jvmci, method, osr_bci);
2571
2572 failure_reason = compile_state.failure_reason();
2573 failure_reason_on_C_heap = compile_state.failure_reason_on_C_heap();
2574 if (!compile_state.retryable()) {
2575 retry_message = "not retryable";
2576 compilable = ciEnv::MethodCompilable_not_at_tier;
2577 }
2578 if (!task->is_success()) {
2579 assert(failure_reason != nullptr, "must specify failure_reason");
2580 }
2581 }
2582 }
2583 if (!task->is_success() && !JVMCI::in_shutdown()) {
2584 handle_compile_error(thread, task, nullptr, compilable, failure_reason);
2585 }
2586 if (event.should_commit()) {
2587 post_compilation_event(event, task);
2588 }
2589
2590 if (runtime != nullptr) {
2591 runtime->post_compile(thread);
2592 }
2593 } else
2594 #endif // INCLUDE_JVMCI
2595 {
2596 NoHandleMark nhm;
2597 ThreadToNativeFromVM ttn(thread);
2598
2599 ciEnv ci_env(task);
2600 if (should_break) {
2601 ci_env.set_break_at_compile(true);
2602 }
2603 if (should_log) {
2604 ci_env.set_log(thread->log());
2605 }
2606 assert(thread->env() == &ci_env, "set by ci_env");
2607 // The thread-env() field is cleared in ~CompileTaskWrapper.
2608
2609 // Cache Jvmti state
2610 bool method_is_old = ci_env.cache_jvmti_state();
2611
2612 // Skip redefined methods
2613 if (method_is_old) {
2614 ci_env.record_method_not_compilable("redefined method", true);
2615 }
2616
2617 // Cache DTrace flags
2618 ci_env.cache_dtrace_flags();
2619
2620 ciMethod* target = ci_env.get_method_from_handle(target_handle);
2621
2622 TraceTime t1("compilation", &time);
2623 EventCompilation event;
2624
2625 if (comp == nullptr) {
2626 ci_env.record_method_not_compilable("no compiler");
2627 } else if (!ci_env.failing()) {
2628 if (WhiteBoxAPI && WhiteBox::compilation_locked) {
2629 whitebox_lock_compilation();
2630 }
2631 comp->compile_method(&ci_env, target, osr_bci, true, directive);
2632
2633 /* Repeat compilation without installing code for profiling purposes */
2634 int repeat_compilation_count = task->is_aot_load() ? 0 : directive->RepeatCompilationOption;
2635 if (repeat_compilation_count > 0) {
2636 CHeapStringHolder failure_reason;
2637 failure_reason.set(ci_env._failure_reason.get());
2638 while (repeat_compilation_count > 0) {
2639 ResourceMark rm(thread);
2640 task->print_ul("NO CODE INSTALLED");
2641 thread->timeout()->reset();
2642 ci_env._failure_reason.clear();
2643 comp->compile_method(&ci_env, target, osr_bci, false, directive);
2644 repeat_compilation_count--;
2645 }
2646 ci_env._failure_reason.set(failure_reason.get());
2647 }
2648 }
2649
2650
2651 if (!ci_env.failing() && !task->is_success() && !task->is_precompile()) {
2652 assert(ci_env.failure_reason() != nullptr, "expect failure reason");
2653 assert(false, "compiler should always document failure: %s", ci_env.failure_reason());
2654 // The compiler elected, without comment, not to register a result.
2655 // Do not attempt further compilations of this method.
2656 ci_env.record_method_not_compilable("compile failed");
2657 }
2658
2659 // Copy this bit to the enclosing block:
2660 compilable = ci_env.compilable();
2661
2662 if (ci_env.failing()) {
2663 // Duplicate the failure reason string, so that it outlives ciEnv
2664 failure_reason = os::strdup(ci_env.failure_reason(), mtCompiler);
2665 failure_reason_on_C_heap = true;
2666 retry_message = ci_env.retry_message();
2667 ci_env.report_failure(failure_reason);
2668 }
2669
2670 if (ci_env.failing()) {
2671 handle_compile_error(thread, task, &ci_env, compilable, failure_reason);
2672 }
2673 if (event.should_commit()) {
2674 post_compilation_event(event, task);
2675 }
2676 }
2677
2678 if (failure_reason != nullptr) {
2679 task->set_failure_reason(failure_reason, failure_reason_on_C_heap);
2680 if (CompilationLog::log() != nullptr) {
2681 CompilationLog::log()->log_failure(thread, task, failure_reason, retry_message);
2682 }
2683 if (PrintCompilation || directive->PrintCompilationOption) {
2684 FormatBufferResource msg = retry_message != nullptr ?
2685 FormatBufferResource("COMPILE SKIPPED: %s (%s)", failure_reason, retry_message) :
2686 FormatBufferResource("COMPILE SKIPPED: %s", failure_reason);
2687 task->print(tty, msg);
2688 }
2689 }
2690
2691 task->mark_finished(os::elapsed_counter());
2692 DirectivesStack::release(directive);
2693
2694 methodHandle method(thread, task->method());
2695
2696 DTRACE_METHOD_COMPILE_END_PROBE(method, compiler_name(task_level), task->is_success());
2697
2698 collect_statistics(thread, time, task);
2699
2700 if (PrintCompilation && PrintCompilation2) {
2701 tty->print("%7d ", (int) tty->time_stamp().milliseconds()); // print timestamp
2702 tty->print("%4d ", compile_id); // print compilation number
2703 tty->print("%s ", (is_osr ? "%" : (task->is_aot_load() ? (task->preload() ? "P" : "A") : " ")));
2704 if (task->is_success()) {
2705 tty->print("size: %d(%d) ", task->nm_total_size(), task->nm_insts_size());
2706 }
2707 tty->print_cr("time: %d inlined: %d bytes", (int)time.milliseconds(), task->num_inlined_bytecodes());
2708 }
2709
2710 Log(compilation, codecache) log;
2711 if (log.is_debug()) {
2712 LogStream ls(log.debug());
2713 codecache_print(&ls, /* detailed= */ false);
2714 }
2715 if (PrintCodeCacheOnCompilation) {
2716 codecache_print(/* detailed= */ false);
2717 }
2718 // Disable compilation, if required.
2719 switch (compilable) {
2720 case ciEnv::MethodCompilable_never:
2721 if (is_osr)
2722 method->set_not_osr_compilable_quietly("MethodCompilable_never");
2723 else
2724 method->set_not_compilable_quietly("MethodCompilable_never");
2725 break;
2726 case ciEnv::MethodCompilable_not_at_tier:
2727 if (is_osr)
2728 method->set_not_osr_compilable_quietly("MethodCompilable_not_at_tier", task_level);
2729 else
2730 method->set_not_compilable_quietly("MethodCompilable_not_at_tier", task_level);
2731 break;
2732 }
2733
2734 // Note that the queued_for_compilation bits are cleared without
2735 // protection of a mutex. [They were set by the requester thread,
2736 // when adding the task to the compile queue -- at which time the
2737 // compile queue lock was held. Subsequently, we acquired the compile
2738 // queue lock to get this task off the compile queue; thus (to belabour
2739 // the point somewhat) our clearing of the bits must be occurring
2740 // only after the setting of the bits. See also 14012000 above.
2741 method->clear_queued_for_compilation();
2742 method->set_pending_queue_processed(false);
2743
2744 if (should_print_compilation) {
2745 ResourceMark rm;
2746 task->print_tty();
2747 }
2748 }
2749
2750 /**
2751 * The CodeCache is full. Print warning and disable compilation.
2752 * Schedule code cache cleaning so compilation can continue later.
2753 * This function needs to be called only from CodeCache::allocate(),
2754 * since we currently handle a full code cache uniformly.
2755 */
2756 void CompileBroker::handle_full_code_cache(CodeBlobType code_blob_type) {
2757 UseInterpreter = true;
2758 if (UseCompiler || AlwaysCompileLoopMethods ) {
2759 if (xtty != nullptr) {
2760 stringStream s;
2761 // Dump code cache state into a buffer before locking the tty,
2762 // because log_state() will use locks causing lock conflicts.
2763 CodeCache::log_state(&s);
2764 // Lock to prevent tearing
2765 ttyLocker ttyl;
2766 xtty->begin_elem("code_cache_full");
2767 xtty->print("%s", s.freeze());
2768 xtty->stamp();
2769 xtty->end_elem();
2770 }
2771
2772 #ifndef PRODUCT
2773 if (ExitOnFullCodeCache) {
2774 codecache_print(/* detailed= */ true);
2775 before_exit(JavaThread::current());
2776 exit_globals(); // will delete tty
2777 vm_direct_exit(1);
2778 }
2779 #endif
2780 if (UseCodeCacheFlushing) {
2781 // Since code cache is full, immediately stop new compiles
2782 if (CompileBroker::set_should_compile_new_jobs(CompileBroker::stop_compilation)) {
2783 log_info(codecache)("Code cache is full - disabling compilation");
2784 }
2785 } else {
2786 disable_compilation_forever();
2787 }
2788
2789 CodeCache::report_codemem_full(code_blob_type, should_print_compiler_warning());
2790 }
2791 }
2792
2793 // ------------------------------------------------------------------
2794 // CompileBroker::update_compile_perf_data
2795 //
2796 // Record this compilation for debugging purposes.
2797 void CompileBroker::update_compile_perf_data(CompilerThread* thread, const methodHandle& method, bool is_osr) {
2798 ResourceMark rm;
2799 char* method_name = method->name()->as_C_string();
2800 char current_method[CompilerCounters::cmname_buffer_length];
2801 size_t maxLen = CompilerCounters::cmname_buffer_length;
2802
2803 const char* class_name = method->method_holder()->name()->as_C_string();
2804
2805 size_t s1len = strlen(class_name);
2806 size_t s2len = strlen(method_name);
2807
2808 // check if we need to truncate the string
2809 if (s1len + s2len + 2 > maxLen) {
2810
2811 // the strategy is to lop off the leading characters of the
2812 // class name and the trailing characters of the method name.
2813
2814 if (s2len + 2 > maxLen) {
2815 // lop of the entire class name string, let snprintf handle
2816 // truncation of the method name.
2817 class_name += s1len; // null string
2818 }
2819 else {
2820 // lop off the extra characters from the front of the class name
2821 class_name += ((s1len + s2len + 2) - maxLen);
2822 }
2823 }
2824
2825 jio_snprintf(current_method, maxLen, "%s %s", class_name, method_name);
2826
2827 int last_compile_type = normal_compile;
2828 if (CICountOSR && is_osr) {
2829 last_compile_type = osr_compile;
2830 } else if (CICountNative && method->is_native()) {
2831 last_compile_type = native_compile;
2832 }
2833
2834 CompilerCounters* counters = thread->counters();
2835 counters->set_current_method(current_method);
2836 counters->set_compile_type((jlong) last_compile_type);
2837 }
2838
2839 // ------------------------------------------------------------------
2840 // CompileBroker::collect_statistics
2841 //
2842 // Collect statistics about the compilation.
2843
2844 void CompileBroker::collect_statistics(CompilerThread* thread, elapsedTimer time, CompileTask* task) {
2845 bool success = task->is_success();
2846 methodHandle method (thread, task->method());
2847 int compile_id = task->compile_id();
2848 bool is_osr = (task->osr_bci() != standard_entry_bci);
2849 const int comp_level = task->comp_level();
2850 CompilerCounters* counters = thread->counters();
2851
2852 MutexLocker locker(CompileStatistics_lock);
2853
2854 // _perf variables are production performance counters which are
2855 // updated regardless of the setting of the CITime and CITimeEach flags
2856 //
2857
2858 // account all time, including bailouts and failures in this counter;
2859 // C1 and C2 counters are counting both successful and unsuccessful compiles
2860 _t_total_compilation.add(&time);
2861
2862 // Update compilation times. Used by the implementation of JFR CompilerStatistics
2863 // and java.lang.management.CompilationMXBean.
2864 _perf_total_compilation->inc(time.ticks());
2865 _peak_compilation_time = MAX2(time.milliseconds(), _peak_compilation_time);
2866
2867 if (!success) {
2868 _total_bailout_count++;
2869 if (UsePerfData) {
2870 _perf_last_failed_method->set_value(counters->current_method());
2871 _perf_last_failed_type->set_value(counters->compile_type());
2872 _perf_total_bailout_count->inc();
2873 }
2874 _t_bailedout_compilation.add(&time);
2875
2876 if (CITime || log_is_enabled(Info, init)) {
2877 CompilerStatistics* stats = nullptr;
2878 if (task->is_aot_load()) {
2879 int level = task->preload() ? CompLevel_full_optimization : (comp_level - 1);
2880 stats = &_aot_stats_per_level[level];
2881 } else {
2882 stats = &_stats_per_level[comp_level-1];
2883 }
2884 stats->_bailout.update(time, 0);
2885 }
2886 } else if (!task->is_success()) {
2887 if (UsePerfData) {
2888 _perf_last_invalidated_method->set_value(counters->current_method());
2889 _perf_last_invalidated_type->set_value(counters->compile_type());
2890 _perf_total_invalidated_count->inc();
2891 }
2892 _total_invalidated_count++;
2893 _t_invalidated_compilation.add(&time);
2894
2895 if (CITime || log_is_enabled(Info, init)) {
2896 CompilerStatistics* stats = nullptr;
2897 if (task->is_aot_load()) {
2898 int level = task->preload() ? CompLevel_full_optimization : (comp_level - 1);
2899 stats = &_aot_stats_per_level[level];
2900 } else {
2901 stats = &_stats_per_level[comp_level-1];
2902 }
2903 stats->_invalidated.update(time, 0);
2904 }
2905 } else {
2906 // Compilation succeeded
2907 if (CITime || log_is_enabled(Info, init)) {
2908 int bytes_compiled = method->code_size() + task->num_inlined_bytecodes();
2909 if (is_osr) {
2910 _t_osr_compilation.add(&time);
2911 _sum_osr_bytes_compiled += bytes_compiled;
2912 } else {
2913 _t_standard_compilation.add(&time);
2914 _sum_standard_bytes_compiled += method->code_size() + task->num_inlined_bytecodes();
2915 }
2916
2917 // Collect statistic per compilation level
2918 if (task->is_aot_load()) {
2919 _aot_stats._standard.update(time, bytes_compiled);
2920 _aot_stats._nmethods_size += task->nm_total_size();
2921 _aot_stats._nmethods_code_size += task->nm_insts_size();
2922 int level = task->preload() ? CompLevel_full_optimization : (comp_level - 1);
2923 CompilerStatistics* stats = &_aot_stats_per_level[level];
2924 stats->_standard.update(time, bytes_compiled);
2925 stats->_nmethods_size += task->nm_total_size();
2926 stats->_nmethods_code_size += task->nm_insts_size();
2927 } else if (comp_level > CompLevel_none && comp_level <= CompLevel_full_optimization) {
2928 CompilerStatistics* stats = &_stats_per_level[comp_level-1];
2929 if (is_osr) {
2930 stats->_osr.update(time, bytes_compiled);
2931 } else {
2932 stats->_standard.update(time, bytes_compiled);
2933 }
2934 stats->_nmethods_size += task->nm_total_size();
2935 stats->_nmethods_code_size += task->nm_insts_size();
2936 } else {
2937 assert(false, "CompilerStatistics object does not exist for compilation level %d", comp_level);
2938 }
2939
2940 // Collect statistic per compiler
2941 AbstractCompiler* comp = task->compiler();
2942 if (comp && !task->is_aot_load()) {
2943 CompilerStatistics* stats = comp->stats();
2944 if (is_osr) {
2945 stats->_osr.update(time, bytes_compiled);
2946 } else {
2947 stats->_standard.update(time, bytes_compiled);
2948 }
2949 stats->_nmethods_size += task->nm_total_size();
2950 stats->_nmethods_code_size += task->nm_insts_size();
2951 } else if (!task->is_aot_load()) { // if (!comp)
2952 assert(false, "Compiler object must exist");
2953 }
2954 }
2955
2956 if (UsePerfData) {
2957 // save the name of the last method compiled
2958 _perf_last_method->set_value(counters->current_method());
2959 _perf_last_compile_type->set_value(counters->compile_type());
2960 _perf_last_compile_size->set_value(method->code_size() +
2961 task->num_inlined_bytecodes());
2962 if (is_osr) {
2963 _perf_osr_compilation->inc(time.ticks());
2964 _perf_sum_osr_bytes_compiled->inc(method->code_size() + task->num_inlined_bytecodes());
2965 } else {
2966 _perf_standard_compilation->inc(time.ticks());
2967 _perf_sum_standard_bytes_compiled->inc(method->code_size() + task->num_inlined_bytecodes());
2968 }
2969 }
2970
2971 if (CITimeEach) {
2972 double compile_time = time.seconds();
2973 double bytes_per_sec = compile_time == 0.0 ? 0.0 : (double)(method->code_size() + task->num_inlined_bytecodes()) / compile_time;
2974 tty->print_cr("%3d seconds: %6.3f bytes/sec : %f (bytes %d + %d inlined)",
2975 compile_id, compile_time, bytes_per_sec, method->code_size(), task->num_inlined_bytecodes());
2976 }
2977
2978 // Collect counts of successful compilations
2979 _sum_nmethod_size += task->nm_total_size();
2980 _sum_nmethod_code_size += task->nm_insts_size();
2981 _total_compile_count++;
2982
2983 if (UsePerfData) {
2984 _perf_sum_nmethod_size->inc( task->nm_total_size());
2985 _perf_sum_nmethod_code_size->inc(task->nm_insts_size());
2986 _perf_total_compile_count->inc();
2987 }
2988
2989 if (is_osr) {
2990 if (UsePerfData) _perf_total_osr_compile_count->inc();
2991 _total_osr_compile_count++;
2992 } else {
2993 if (UsePerfData) _perf_total_standard_compile_count->inc();
2994 _total_standard_compile_count++;
2995 }
2996 }
2997 // set the current method for the thread to null
2998 if (UsePerfData) counters->set_current_method("");
2999 }
3000
3001 const char* CompileBroker::compiler_name(int comp_level) {
3002 AbstractCompiler *comp = CompileBroker::compiler(comp_level);
3003 if (comp == nullptr) {
3004 return "no compiler";
3005 } else {
3006 return (comp->name());
3007 }
3008 }
3009
3010 jlong CompileBroker::total_compilation_ticks() {
3011 return _perf_total_compilation != nullptr ? _perf_total_compilation->get_value() : 0;
3012 }
3013
3014 void CompileBroker::log_not_entrant(nmethod* nm) {
3015 _total_not_entrant_count++;
3016 if (CITime || log_is_enabled(Info, init)) {
3017 CompilerStatistics* stats = nullptr;
3018 int level = nm->comp_level();
3019 if (nm->is_aot()) {
3020 if (nm->preloaded()) {
3021 assert(level == CompLevel_full_optimization, "%d", level);
3022 level = CompLevel_full_optimization + 1;
3023 }
3024 stats = &_aot_stats_per_level[level - 1];
3025 } else {
3026 stats = &_stats_per_level[level - 1];
3027 }
3028 stats->_made_not_entrant._count++;
3029 }
3030 }
3031
3032 void CompileBroker::print_times(const char* name, CompilerStatistics* stats) {
3033 tty->print_cr(" %s {speed: %6.3f bytes/s; standard: %6.3f s, %u bytes, %u methods; osr: %6.3f s, %u bytes, %u methods; nmethods_size: %u bytes; nmethods_code_size: %u bytes}",
3034 name, stats->bytes_per_second(),
3035 stats->_standard._time.seconds(), stats->_standard._bytes, stats->_standard._count,
3036 stats->_osr._time.seconds(), stats->_osr._bytes, stats->_osr._count,
3037 stats->_nmethods_size, stats->_nmethods_code_size);
3038 }
3039
3040 static void print_helper(outputStream* st, const char* name, CompilerStatistics::Data data, bool print_time = true) {
3041 if (data._count > 0) {
3042 st->print("; %s: %4u methods", name, data._count);
3043 if (print_time) {
3044 st->print(" (in %.3fs)", data._time.seconds());
3045 }
3046 }
3047 }
3048
3049 static void print_tier_helper(outputStream* st, const char* prefix, int tier, CompilerStatistics* stats) {
3050 st->print(" %s%d: %5u methods", prefix, tier, stats->_standard._count);
3051 if (stats->_standard._count > 0) {
3052 st->print(" (in %.3fs)", stats->_standard._time.seconds());
3053 }
3054 print_helper(st, "osr", stats->_osr);
3055 print_helper(st, "bailout", stats->_bailout);
3056 print_helper(st, "invalid", stats->_invalidated);
3057 print_helper(st, "not_entrant", stats->_made_not_entrant, false);
3058 st->cr();
3059 }
3060
3061 static void print_queue_info(outputStream* st, CompileQueue* queue) {
3062 if (queue != nullptr) {
3063 MutexLocker ml(queue->lock());
3064
3065 uint total_cnt = 0;
3066 uint active_cnt = 0;
3067 for (JavaThread* jt : *ThreadsSMRSupport::get_java_thread_list()) {
3068 guarantee(jt != nullptr, "");
3069 if (jt->is_Compiler_thread()) {
3070 CompilerThread* ct = (CompilerThread*)jt;
3071
3072 guarantee(ct != nullptr, "");
3073 if (ct->queue() == queue) {
3074 ++total_cnt;
3075 CompileTask* task = ct->task();
3076 if (task != nullptr) {
3077 ++active_cnt;
3078 }
3079 }
3080 }
3081 }
3082
3083 st->print(" %s (%d active / %d total threads): %u tasks",
3084 queue->name(), active_cnt, total_cnt, queue->size());
3085 if (queue->size() > 0) {
3086 uint counts[] = {0, 0, 0, 0, 0}; // T1 ... T5
3087 for (CompileTask* task = queue->first(); task != nullptr; task = task->next()) {
3088 int tier = task->comp_level();
3089 if (task->is_aot_load() && task->preload()) {
3090 assert(tier == CompLevel_full_optimization, "%d", tier);
3091 tier = CompLevel_full_optimization + 1;
3092 }
3093 counts[tier-1]++;
3094 }
3095 st->print(":");
3096 for (int tier = CompLevel_simple; tier <= CompilationPolicy::highest_compile_level() + 1; tier++) {
3097 uint cnt = counts[tier-1];
3098 if (cnt > 0) {
3099 st->print(" T%d: %u tasks;", tier, cnt);
3100 }
3101 }
3102 }
3103 st->cr();
3104
3105 // for (JavaThread* jt : *ThreadsSMRSupport::get_java_thread_list()) {
3106 // guarantee(jt != nullptr, "");
3107 // if (jt->is_Compiler_thread()) {
3108 // CompilerThread* ct = (CompilerThread*)jt;
3109 //
3110 // guarantee(ct != nullptr, "");
3111 // if (ct->queue() == queue) {
3112 // ResourceMark rm;
3113 // CompileTask* task = ct->task();
3114 // st->print(" %s: ", ct->name_raw());
3115 // if (task != nullptr) {
3116 // task->print(st, nullptr, true /*short_form*/, false /*cr*/);
3117 // }
3118 // st->cr();
3119 // }
3120 // }
3121 // }
3122 }
3123 }
3124 void CompileBroker::print_statistics_on(outputStream* st) {
3125 st->print_cr(" Total: %u methods; %u bailouts, %u invalidated, %u non_entrant",
3126 _total_compile_count, _total_bailout_count, _total_invalidated_count, _total_not_entrant_count);
3127 for (int tier = CompLevel_simple; tier <= CompilationPolicy::highest_compile_level(); tier++) {
3128 print_tier_helper(st, "Tier", tier, &_stats_per_level[tier-1]);
3129 }
3130 st->cr();
3131
3132 if (AOTCodeCaching) {
3133 for (int tier = CompLevel_simple; tier <= CompilationPolicy::highest_compile_level() + 1; tier++) {
3134 if (tier != CompLevel_full_profile) {
3135 print_tier_helper(st, "AOT Code T", tier, &_aot_stats_per_level[tier - 1]);
3136 }
3137 }
3138 st->cr();
3139 }
3140
3141 print_queue_info(st, _c1_compile_queue);
3142 print_queue_info(st, _c2_compile_queue);
3143 print_queue_info(st, _ac1_compile_queue);
3144 print_queue_info(st, _ac2_compile_queue);
3145 }
3146
3147 void CompileBroker::print_times(bool per_compiler, bool aggregate) {
3148 if (per_compiler) {
3149 if (aggregate) {
3150 tty->cr();
3151 tty->print_cr("[%dms] Individual compiler times (for compiled methods only)", (int)tty->time_stamp().milliseconds());
3152 tty->print_cr("------------------------------------------------");
3153 tty->cr();
3154 }
3155 for (unsigned int i = 0; i < sizeof(_compilers) / sizeof(AbstractCompiler*); i++) {
3156 AbstractCompiler* comp = _compilers[i];
3157 if (comp != nullptr) {
3158 print_times(comp->name(), comp->stats());
3159 }
3160 }
3161 if (_aot_stats._standard._count > 0) {
3162 print_times("AC", &_aot_stats);
3163 }
3164 if (aggregate) {
3165 tty->cr();
3166 tty->print_cr("Individual compilation Tier times (for compiled methods only)");
3167 tty->print_cr("------------------------------------------------");
3168 tty->cr();
3169 }
3170 char tier_name[256];
3171 for (int tier = CompLevel_simple; tier <= CompilationPolicy::highest_compile_level(); tier++) {
3172 CompilerStatistics* stats = &_stats_per_level[tier-1];
3173 os::snprintf_checked(tier_name, sizeof(tier_name), "Tier%d", tier);
3174 print_times(tier_name, stats);
3175 }
3176 for (int tier = CompLevel_simple; tier <= CompilationPolicy::highest_compile_level() + 1; tier++) {
3177 CompilerStatistics* stats = &_aot_stats_per_level[tier-1];
3178 if (stats->_standard._bytes > 0) {
3179 os::snprintf_checked(tier_name, sizeof(tier_name), "AOT Code T%d", tier);
3180 print_times(tier_name, stats);
3181 }
3182 }
3183 }
3184
3185 if (!aggregate) {
3186 return;
3187 }
3188
3189 elapsedTimer standard_compilation = CompileBroker::_t_standard_compilation;
3190 elapsedTimer osr_compilation = CompileBroker::_t_osr_compilation;
3191 elapsedTimer total_compilation = CompileBroker::_t_total_compilation;
3192
3193 uint standard_bytes_compiled = CompileBroker::_sum_standard_bytes_compiled;
3194 uint osr_bytes_compiled = CompileBroker::_sum_osr_bytes_compiled;
3195
3196 uint standard_compile_count = CompileBroker::_total_standard_compile_count;
3197 uint osr_compile_count = CompileBroker::_total_osr_compile_count;
3198 uint total_compile_count = CompileBroker::_total_compile_count;
3199 uint total_bailout_count = CompileBroker::_total_bailout_count;
3200 uint total_invalidated_count = CompileBroker::_total_invalidated_count;
3201
3202 uint nmethods_code_size = CompileBroker::_sum_nmethod_code_size;
3203 uint nmethods_size = CompileBroker::_sum_nmethod_size;
3204
3205 tty->cr();
3206 tty->print_cr("Accumulated compiler times");
3207 tty->print_cr("----------------------------------------------------------");
3208 //0000000000111111111122222222223333333333444444444455555555556666666666
3209 //0123456789012345678901234567890123456789012345678901234567890123456789
3210 tty->print_cr(" Total compilation time : %7.3f s", total_compilation.seconds());
3211 tty->print_cr(" Standard compilation : %7.3f s, Average : %2.3f s",
3212 standard_compilation.seconds(),
3213 standard_compile_count == 0 ? 0.0 : standard_compilation.seconds() / standard_compile_count);
3214 tty->print_cr(" Bailed out compilation : %7.3f s, Average : %2.3f s",
3215 CompileBroker::_t_bailedout_compilation.seconds(),
3216 total_bailout_count == 0 ? 0.0 : CompileBroker::_t_bailedout_compilation.seconds() / total_bailout_count);
3217 tty->print_cr(" On stack replacement : %7.3f s, Average : %2.3f s",
3218 osr_compilation.seconds(),
3219 osr_compile_count == 0 ? 0.0 : osr_compilation.seconds() / osr_compile_count);
3220 tty->print_cr(" Invalidated : %7.3f s, Average : %2.3f s",
3221 CompileBroker::_t_invalidated_compilation.seconds(),
3222 total_invalidated_count == 0 ? 0.0 : CompileBroker::_t_invalidated_compilation.seconds() / total_invalidated_count);
3223
3224 if (AOTCodeCaching) { // Check flags because AOT code cache could be closed already
3225 tty->cr();
3226 AOTCodeCache::print_timers_on(tty);
3227 }
3228 AbstractCompiler *comp = compiler(CompLevel_simple);
3229 if (comp != nullptr) {
3230 tty->cr();
3231 comp->print_timers();
3232 }
3233 comp = compiler(CompLevel_full_optimization);
3234 if (comp != nullptr) {
3235 tty->cr();
3236 comp->print_timers();
3237 }
3238 #if INCLUDE_JVMCI
3239 if (EnableJVMCI) {
3240 JVMCICompiler *jvmci_comp = JVMCICompiler::instance(false, JavaThread::current_or_null());
3241 if (jvmci_comp != nullptr && jvmci_comp != comp) {
3242 tty->cr();
3243 jvmci_comp->print_timers();
3244 }
3245 }
3246 #endif
3247
3248 tty->cr();
3249 tty->print_cr(" Total compiled methods : %8u methods", total_compile_count);
3250 tty->print_cr(" Standard compilation : %8u methods", standard_compile_count);
3251 tty->print_cr(" On stack replacement : %8u methods", osr_compile_count);
3252 uint tcb = osr_bytes_compiled + standard_bytes_compiled;
3253 tty->print_cr(" Total compiled bytecodes : %8u bytes", tcb);
3254 tty->print_cr(" Standard compilation : %8u bytes", standard_bytes_compiled);
3255 tty->print_cr(" On stack replacement : %8u bytes", osr_bytes_compiled);
3256 double tcs = total_compilation.seconds();
3257 uint bps = tcs == 0.0 ? 0 : (uint)(tcb / tcs);
3258 tty->print_cr(" Average compilation speed : %8u bytes/s", bps);
3259 tty->cr();
3260 tty->print_cr(" nmethod code size : %8u bytes", nmethods_code_size);
3261 tty->print_cr(" nmethod total size : %8u bytes", nmethods_size);
3262 }
3263
3264 // Print general/accumulated JIT information.
3265 void CompileBroker::print_info(outputStream *out) {
3266 if (out == nullptr) out = tty;
3267 out->cr();
3268 out->print_cr("======================");
3269 out->print_cr(" General JIT info ");
3270 out->print_cr("======================");
3271 out->cr();
3272 out->print_cr(" JIT is : %7s", should_compile_new_jobs() ? "on" : "off");
3273 out->print_cr(" Compiler threads : %7d", (int)CICompilerCount);
3274 out->cr();
3275 out->print_cr("CodeCache overview");
3276 out->print_cr("--------------------------------------------------------");
3277 out->cr();
3278 out->print_cr(" Reserved size : %7zu KB", CodeCache::max_capacity() / K);
3279 out->print_cr(" Committed size : %7zu KB", CodeCache::capacity() / K);
3280 out->print_cr(" Unallocated capacity : %7zu KB", CodeCache::unallocated_capacity() / K);
3281 out->cr();
3282 }
3283
3284 // Note: tty_lock must not be held upon entry to this function.
3285 // Print functions called from herein do "micro-locking" on tty_lock.
3286 // That's a tradeoff which keeps together important blocks of output.
3287 // At the same time, continuous tty_lock hold time is kept in check,
3288 // preventing concurrently printing threads from stalling a long time.
3289 void CompileBroker::print_heapinfo(outputStream* out, const char* function, size_t granularity) {
3290 TimeStamp ts_total;
3291 TimeStamp ts_global;
3292 TimeStamp ts;
3293
3294 bool allFun = !strcmp(function, "all");
3295 bool aggregate = !strcmp(function, "aggregate") || !strcmp(function, "analyze") || allFun;
3296 bool usedSpace = !strcmp(function, "UsedSpace") || allFun;
3297 bool freeSpace = !strcmp(function, "FreeSpace") || allFun;
3298 bool methodCount = !strcmp(function, "MethodCount") || allFun;
3299 bool methodSpace = !strcmp(function, "MethodSpace") || allFun;
3300 bool methodAge = !strcmp(function, "MethodAge") || allFun;
3301 bool methodNames = !strcmp(function, "MethodNames") || allFun;
3302 bool discard = !strcmp(function, "discard") || allFun;
3303
3304 if (out == nullptr) {
3305 out = tty;
3306 }
3307
3308 if (!(aggregate || usedSpace || freeSpace || methodCount || methodSpace || methodAge || methodNames || discard)) {
3309 out->print_cr("\n__ CodeHeapStateAnalytics: Function %s is not supported", function);
3310 out->cr();
3311 return;
3312 }
3313
3314 ts_total.update(); // record starting point
3315
3316 if (aggregate) {
3317 print_info(out);
3318 }
3319
3320 // We hold the CodeHeapStateAnalytics_lock all the time, from here until we leave this function.
3321 // That prevents other threads from destroying (making inconsistent) our view on the CodeHeap.
3322 // When we request individual parts of the analysis via the jcmd interface, it is possible
3323 // that in between another thread (another jcmd user or the vm running into CodeCache OOM)
3324 // updated the aggregated data. We will then see a modified, but again consistent, view
3325 // on the CodeHeap. That's a tolerable tradeoff we have to accept because we can't hold
3326 // a lock across user interaction.
3327
3328 // We should definitely acquire this lock before acquiring Compile_lock and CodeCache_lock.
3329 // CodeHeapStateAnalytics_lock may be held by a concurrent thread for a long time,
3330 // leading to an unnecessarily long hold time of the other locks we acquired before.
3331 ts.update(); // record starting point
3332 MutexLocker mu0(CodeHeapStateAnalytics_lock, Mutex::_safepoint_check_flag);
3333 out->print_cr("\n__ CodeHeapStateAnalytics lock wait took %10.3f seconds _________\n", ts.seconds());
3334
3335 // Holding the CodeCache_lock protects from concurrent alterations of the CodeCache.
3336 // Unfortunately, such protection is not sufficient:
3337 // When a new nmethod is created via ciEnv::register_method(), the
3338 // Compile_lock is taken first. After some initializations,
3339 // nmethod::new_nmethod() takes over, grabbing the CodeCache_lock
3340 // immediately (after finalizing the oop references). To lock out concurrent
3341 // modifiers, we have to grab both locks as well in the described sequence.
3342 //
3343 // If we serve an "allFun" call, it is beneficial to hold CodeCache_lock and Compile_lock
3344 // for the entire duration of aggregation and printing. That makes sure we see
3345 // a consistent picture and do not run into issues caused by concurrent alterations.
3346 bool should_take_Compile_lock = !SafepointSynchronize::is_at_safepoint() &&
3347 !Compile_lock->owned_by_self();
3348 bool should_take_CodeCache_lock = !SafepointSynchronize::is_at_safepoint() &&
3349 !CodeCache_lock->owned_by_self();
3350 bool take_global_lock_1 = allFun && should_take_Compile_lock;
3351 bool take_global_lock_2 = allFun && should_take_CodeCache_lock;
3352 bool take_function_lock_1 = !allFun && should_take_Compile_lock;
3353 bool take_function_lock_2 = !allFun && should_take_CodeCache_lock;
3354 bool take_global_locks = take_global_lock_1 || take_global_lock_2;
3355 bool take_function_locks = take_function_lock_1 || take_function_lock_2;
3356
3357 ts_global.update(); // record starting point
3358
3359 ConditionalMutexLocker mu1(Compile_lock, take_global_lock_1, Mutex::_safepoint_check_flag);
3360 ConditionalMutexLocker mu2(CodeCache_lock, take_global_lock_2, Mutex::_no_safepoint_check_flag);
3361 if (take_global_locks) {
3362 out->print_cr("\n__ Compile & CodeCache (global) lock wait took %10.3f seconds _________\n", ts_global.seconds());
3363 ts_global.update(); // record starting point
3364 }
3365
3366 if (aggregate) {
3367 ts.update(); // record starting point
3368 ConditionalMutexLocker mu11(Compile_lock, take_function_lock_1, Mutex::_safepoint_check_flag);
3369 ConditionalMutexLocker mu22(CodeCache_lock, take_function_lock_2, Mutex::_no_safepoint_check_flag);
3370 if (take_function_locks) {
3371 out->print_cr("\n__ Compile & CodeCache (function) lock wait took %10.3f seconds _________\n", ts.seconds());
3372 }
3373
3374 ts.update(); // record starting point
3375 CodeCache::aggregate(out, granularity);
3376 if (take_function_locks) {
3377 out->print_cr("\n__ Compile & CodeCache (function) lock hold took %10.3f seconds _________\n", ts.seconds());
3378 }
3379 }
3380
3381 if (usedSpace) CodeCache::print_usedSpace(out);
3382 if (freeSpace) CodeCache::print_freeSpace(out);
3383 if (methodCount) CodeCache::print_count(out);
3384 if (methodSpace) CodeCache::print_space(out);
3385 if (methodAge) CodeCache::print_age(out);
3386 if (methodNames) {
3387 if (allFun) {
3388 // print_names() can only be used safely if the locks have been continuously held
3389 // since aggregation begin. That is true only for function "all".
3390 CodeCache::print_names(out);
3391 } else {
3392 out->print_cr("\nCodeHeapStateAnalytics: Function 'MethodNames' is only available as part of function 'all'");
3393 }
3394 }
3395 if (discard) CodeCache::discard(out);
3396
3397 if (take_global_locks) {
3398 out->print_cr("\n__ Compile & CodeCache (global) lock hold took %10.3f seconds _________\n", ts_global.seconds());
3399 }
3400 out->print_cr("\n__ CodeHeapStateAnalytics total duration %10.3f seconds _________\n", ts_total.seconds());
3401 }