1 /*
2 * Copyright (c) 2024, Oracle and/or its affiliates. All rights reserved.
3 * Copyright (c) 2019, 2022, Red Hat, Inc. All rights reserved.
4 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
5 *
6 * This code is free software; you can redistribute it and/or modify it
7 * under the terms of the GNU General Public License version 2 only, as
8 * published by the Free Software Foundation.
9 *
10 * This code is distributed in the hope that it will be useful, but WITHOUT
11 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
12 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
13 * version 2 for more details (a copy is included in the LICENSE file that
14 * accompanied this code).
15 *
16 * You should have received a copy of the GNU General Public License version
17 * 2 along with this work; if not, write to the Free Software Foundation,
18 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
19 *
20 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
21 * or visit www.oracle.com if you need additional information or have any
22 * questions.
23 *
24 */
25
26 #include "precompiled.hpp"
27
28 #include "gc/shenandoah/shenandoahClosures.inline.hpp"
29 #include "gc/shenandoah/shenandoahHeap.inline.hpp"
30 #include "gc/shenandoah/shenandoahNMethod.inline.hpp"
31 #include "gc/shenandoah/shenandoahOopClosures.inline.hpp"
32 #include "memory/resourceArea.hpp"
33 #include "runtime/continuation.hpp"
34 #include "runtime/safepointVerifiers.hpp"
35
36 ShenandoahNMethod::ShenandoahNMethod(nmethod* nm, GrowableArray<oop*>& oops, bool non_immediate_oops) :
37 _nm(nm), _oops(nullptr), _oops_count(0), _unregistered(false) {
38
39 if (!oops.is_empty()) {
40 _oops_count = oops.length();
41 _oops = NEW_C_HEAP_ARRAY(oop*, _oops_count, mtGC);
42 for (int c = 0; c < _oops_count; c++) {
43 _oops[c] = oops.at(c);
44 }
45 }
46 _has_non_immed_oops = non_immediate_oops;
47
48 assert_same_oops();
49 }
50
51 ShenandoahNMethod::~ShenandoahNMethod() {
52 if (_oops != nullptr) {
53 FREE_C_HEAP_ARRAY(oop*, _oops);
54 }
55 }
56
57 void ShenandoahNMethod::update() {
58 ResourceMark rm;
59 bool non_immediate_oops = false;
60 GrowableArray<oop*> oops;
61
62 detect_reloc_oops(nm(), oops, non_immediate_oops);
63 if (oops.length() != _oops_count) {
64 if (_oops != nullptr) {
65 FREE_C_HEAP_ARRAY(oop*, _oops);
66 _oops = nullptr;
67 }
68
69 _oops_count = oops.length();
70 if (_oops_count > 0) {
71 _oops = NEW_C_HEAP_ARRAY(oop*, _oops_count, mtGC);
72 }
73 }
74
75 for (int index = 0; index < _oops_count; index ++) {
76 _oops[index] = oops.at(index);
77 }
78 _has_non_immed_oops = non_immediate_oops;
79
80 assert_same_oops();
81 }
82
83 void ShenandoahNMethod::detect_reloc_oops(nmethod* nm, GrowableArray<oop*>& oops, bool& has_non_immed_oops) {
84 has_non_immed_oops = false;
85 // Find all oops relocations
86 RelocIterator iter(nm);
87 while (iter.next()) {
88 if (iter.type() != relocInfo::oop_type) {
89 // Not an oop
90 continue;
91 }
92
93 oop_Relocation* r = iter.oop_reloc();
94 if (!r->oop_is_immediate()) {
95 // Non-immediate oop found
96 has_non_immed_oops = true;
97 continue;
98 }
99
100 oop value = r->oop_value();
101 if (value != nullptr) {
102 oop* addr = r->oop_addr();
103 shenandoah_assert_correct(addr, value);
104 shenandoah_assert_not_in_cset_except(addr, value, ShenandoahHeap::heap()->cancelled_gc());
105 shenandoah_assert_not_forwarded(addr, value);
106 // Non-null immediate oop found. null oops can safely be
107 // ignored since the method will be re-registered if they
108 // are later patched to be non-null.
109 oops.push(addr);
110 }
111 }
112 }
113
114 ShenandoahNMethod* ShenandoahNMethod::for_nmethod(nmethod* nm) {
115 ResourceMark rm;
116 bool non_immediate_oops = false;
117 GrowableArray<oop*> oops;
118
119 detect_reloc_oops(nm, oops, non_immediate_oops);
120 return new ShenandoahNMethod(nm, oops, non_immediate_oops);
121 }
122
123 void ShenandoahNMethod::heal_nmethod(nmethod* nm) {
124 ShenandoahNMethod* data = gc_data(nm);
125 assert(data != nullptr, "Sanity");
126 assert(data->lock()->owned_by_self(), "Must hold the lock");
127
128 ShenandoahHeap* const heap = ShenandoahHeap::heap();
129 if (heap->is_concurrent_weak_root_in_progress() ||
130 heap->is_concurrent_strong_root_in_progress()) {
131 ShenandoahEvacOOMScope evac_scope;
132 heal_nmethod_metadata(data);
133 } else if (heap->is_concurrent_mark_in_progress()) {
134 ShenandoahKeepAliveClosure cl;
135 data->oops_do(&cl);
136 } else {
137 // There is possibility that GC is cancelled when it arrives final mark.
138 // In this case, concurrent root phase is skipped and degenerated GC should be
139 // followed, where nmethods are disarmed.
140 }
141 }
142
143 #ifdef ASSERT
144 void ShenandoahNMethod::assert_correct() {
145 ShenandoahHeap* heap = ShenandoahHeap::heap();
146 for (int c = 0; c < _oops_count; c++) {
147 oop *loc = _oops[c];
148 assert(_nm->code_contains((address) loc) || _nm->oops_contains(loc), "nmethod should contain the oop*");
149 oop o = RawAccess<>::oop_load(loc);
150 shenandoah_assert_correct_except(loc, o, o == nullptr || heap->is_full_gc_move_in_progress());
151 }
152
153 oop* const begin = _nm->oops_begin();
154 oop* const end = _nm->oops_end();
155 for (oop* p = begin; p < end; p++) {
156 if (*p != Universe::non_oop_word()) {
157 oop o = RawAccess<>::oop_load(p);
158 shenandoah_assert_correct_except(p, o, o == nullptr || heap->is_full_gc_move_in_progress());
159 }
160 }
161 }
162
163 class ShenandoahNMethodOopDetector : public OopClosure {
164 private:
165 ResourceMark rm; // For growable array allocation below.
166 GrowableArray<oop*> _oops;
167
168 public:
169 ShenandoahNMethodOopDetector() : _oops(10) {};
170
171 void do_oop(oop* o) {
172 _oops.append(o);
173 }
174 void do_oop(narrowOop* o) {
175 fatal("NMethods should not have compressed oops embedded.");
176 }
177
178 GrowableArray<oop*>* oops() {
179 return &_oops;
180 }
181 };
182
183 void ShenandoahNMethod::assert_same_oops(bool allow_dead) {
184 ShenandoahNMethodOopDetector detector;
185 nm()->oops_do(&detector, allow_dead);
186
187 GrowableArray<oop*>* oops = detector.oops();
188
189 int count = _oops_count;
190 for (int index = 0; index < _oops_count; index ++) {
191 assert(oops->contains(_oops[index]), "Must contain this oop");
192 }
193
194 for (oop* p = nm()->oops_begin(); p < nm()->oops_end(); p ++) {
195 if (*p == Universe::non_oop_word()) continue;
196 count++;
197 assert(oops->contains(p), "Must contain this oop");
198 }
199
200 if (oops->length() < count) {
201 stringStream debug_stream;
202 debug_stream.print_cr("detected locs: %d", oops->length());
203 for (int i = 0; i < oops->length(); i++) {
204 debug_stream.print_cr("-> " PTR_FORMAT, p2i(oops->at(i)));
205 }
206 debug_stream.print_cr("recorded oops: %d", _oops_count);
207 for (int i = 0; i < _oops_count; i++) {
208 debug_stream.print_cr("-> " PTR_FORMAT, p2i(_oops[i]));
209 }
210 GrowableArray<oop*> check;
211 bool non_immed;
212 detect_reloc_oops(nm(), check, non_immed);
213 debug_stream.print_cr("check oops: %d", check.length());
214 for (int i = 0; i < check.length(); i++) {
215 debug_stream.print_cr("-> " PTR_FORMAT, p2i(check.at(i)));
216 }
217 fatal("Must match #detected: %d, #recorded: %d, #total: %d, begin: " PTR_FORMAT ", end: " PTR_FORMAT "\n%s",
218 oops->length(), _oops_count, count, p2i(nm()->oops_begin()), p2i(nm()->oops_end()), debug_stream.freeze());
219 }
220 }
221 #endif
222
223 ShenandoahNMethodTable::ShenandoahNMethodTable() :
224 _heap(ShenandoahHeap::heap()),
225 _index(0),
226 _itr_cnt(0) {
227 _list = new ShenandoahNMethodList(minSize);
228 }
229
230 ShenandoahNMethodTable::~ShenandoahNMethodTable() {
231 assert(_list != nullptr, "Sanity");
232 _list->release();
233 }
234
235 void ShenandoahNMethodTable::register_nmethod(nmethod* nm) {
236 assert(CodeCache_lock->owned_by_self(), "Must have CodeCache_lock held");
237 assert(_index >= 0 && _index <= _list->size(), "Sanity");
238
239 ShenandoahNMethod* data = ShenandoahNMethod::gc_data(nm);
240
241 if (data != nullptr) {
242 assert(contain(nm), "Must have been registered");
243 assert(nm == data->nm(), "Must be same nmethod");
244 // Prevent updating a nmethod while concurrent iteration is in progress.
245 wait_until_concurrent_iteration_done();
246 ShenandoahReentrantLocker data_locker(data->lock());
247 data->update();
248 } else {
249 // For a new nmethod, we can safely append it to the list, because
250 // concurrent iteration will not touch it.
251 data = ShenandoahNMethod::for_nmethod(nm);
252 assert(data != nullptr, "Sanity");
253 ShenandoahNMethod::attach_gc_data(nm, data);
254 ShenandoahLocker locker(&_lock);
255 log_register_nmethod(nm);
256 append(data);
257 }
258 // Disarm new nmethod
259 ShenandoahNMethod::disarm_nmethod(nm);
260 }
261
262 void ShenandoahNMethodTable::unregister_nmethod(nmethod* nm) {
263 assert_locked_or_safepoint(CodeCache_lock);
264
265 ShenandoahNMethod* data = ShenandoahNMethod::gc_data(nm);
266 assert(data != nullptr, "Sanity");
267 log_unregister_nmethod(nm);
268 ShenandoahLocker locker(&_lock);
269 assert(contain(nm), "Must have been registered");
270
271 int idx = index_of(nm);
272 assert(idx >= 0 && idx < _index, "Invalid index");
273 ShenandoahNMethod::attach_gc_data(nm, nullptr);
274 remove(idx);
275 }
276
277 bool ShenandoahNMethodTable::contain(nmethod* nm) const {
278 return index_of(nm) != -1;
279 }
280
281 ShenandoahNMethod* ShenandoahNMethodTable::at(int index) const {
282 assert(index >= 0 && index < _index, "Out of bound");
283 return _list->at(index);
284 }
285
286 int ShenandoahNMethodTable::index_of(nmethod* nm) const {
287 for (int index = 0; index < length(); index ++) {
288 if (at(index)->nm() == nm) {
289 return index;
290 }
291 }
292 return -1;
293 }
294
295 void ShenandoahNMethodTable::remove(int idx) {
296 shenandoah_assert_locked_or_safepoint(CodeCache_lock);
297 assert(_index >= 0 && _index <= _list->size(), "Sanity");
298
299 assert(idx >= 0 && idx < _index, "Out of bound");
300 ShenandoahNMethod* snm = _list->at(idx);
301 ShenandoahNMethod* tmp = _list->at(_index - 1);
302 _list->set(idx, tmp);
303 _index --;
304
305 delete snm;
306 }
307
308 void ShenandoahNMethodTable::wait_until_concurrent_iteration_done() {
309 assert(CodeCache_lock->owned_by_self(), "Lock must be held");
310 while (iteration_in_progress()) {
311 CodeCache_lock->wait_without_safepoint_check();
312 }
313 }
314
315 void ShenandoahNMethodTable::append(ShenandoahNMethod* snm) {
316 if (is_full()) {
317 int new_size = 2 * _list->size();
318 // Rebuild table and replace current one
319 rebuild(new_size);
320 }
321
322 _list->set(_index++, snm);
323 assert(_index >= 0 && _index <= _list->size(), "Sanity");
324 }
325
326 void ShenandoahNMethodTable::rebuild(int size) {
327 ShenandoahNMethodList* new_list = new ShenandoahNMethodList(size);
328 new_list->transfer(_list, _index);
329
330 // Release old list
331 _list->release();
332 _list = new_list;
333 }
334
335 ShenandoahNMethodTableSnapshot* ShenandoahNMethodTable::snapshot_for_iteration() {
336 assert(CodeCache_lock->owned_by_self(), "Must have CodeCache_lock held");
337 _itr_cnt++;
338 return new ShenandoahNMethodTableSnapshot(this);
339 }
340
341 void ShenandoahNMethodTable::finish_iteration(ShenandoahNMethodTableSnapshot* snapshot) {
342 assert(CodeCache_lock->owned_by_self(), "Must have CodeCache_lock held");
343 assert(iteration_in_progress(), "Why we here?");
344 assert(snapshot != nullptr, "No snapshot");
345 _itr_cnt--;
346
347 delete snapshot;
348 }
349
350 void ShenandoahNMethodTable::log_register_nmethod(nmethod* nm) {
351 LogTarget(Debug, gc, nmethod) log;
352 if (!log.is_enabled()) {
353 return;
354 }
355
356 ResourceMark rm;
357 log.print("Register NMethod: %s.%s [" PTR_FORMAT "] (%s)",
358 nm->method()->method_holder()->external_name(),
359 nm->method()->name()->as_C_string(),
360 p2i(nm),
361 nm->compiler_name());
362 }
363
364 void ShenandoahNMethodTable::log_unregister_nmethod(nmethod* nm) {
365 LogTarget(Debug, gc, nmethod) log;
366 if (!log.is_enabled()) {
367 return;
368 }
369
370 ResourceMark rm;
371 log.print("Unregister NMethod: %s.%s [" PTR_FORMAT "]",
372 nm->method()->method_holder()->external_name(),
373 nm->method()->name()->as_C_string(),
374 p2i(nm));
375 }
376
377 #ifdef ASSERT
378 void ShenandoahNMethodTable::assert_nmethods_correct() {
379 assert_locked_or_safepoint(CodeCache_lock);
380
381 for (int index = 0; index < length(); index ++) {
382 ShenandoahNMethod* m = _list->at(index);
383 // Concurrent unloading may have dead nmethods to be cleaned by sweeper
384 if (m->is_unregistered()) continue;
385 m->assert_correct();
386 }
387 }
388 #endif
389
390
391 ShenandoahNMethodList::ShenandoahNMethodList(int size) :
392 _size(size), _ref_count(1) {
393 _list = NEW_C_HEAP_ARRAY(ShenandoahNMethod*, size, mtGC);
394 }
395
396 ShenandoahNMethodList::~ShenandoahNMethodList() {
397 assert(_list != nullptr, "Sanity");
398 assert(_ref_count == 0, "Must be");
399 FREE_C_HEAP_ARRAY(ShenandoahNMethod*, _list);
400 }
401
402 void ShenandoahNMethodList::transfer(ShenandoahNMethodList* const list, int limit) {
403 assert(limit <= size(), "Sanity");
404 ShenandoahNMethod** old_list = list->list();
405 for (int index = 0; index < limit; index++) {
406 _list[index] = old_list[index];
407 }
408 }
409
410 ShenandoahNMethodList* ShenandoahNMethodList::acquire() {
411 assert_locked_or_safepoint(CodeCache_lock);
412 _ref_count++;
413 return this;
414 }
415
416 void ShenandoahNMethodList::release() {
417 assert_locked_or_safepoint(CodeCache_lock);
418 _ref_count--;
419 if (_ref_count == 0) {
420 delete this;
421 }
422 }
423
424 ShenandoahNMethodTableSnapshot::ShenandoahNMethodTableSnapshot(ShenandoahNMethodTable* table) :
425 _heap(ShenandoahHeap::heap()), _list(table->_list->acquire()), _limit(table->_index), _claimed(0) {
426 }
427
428 ShenandoahNMethodTableSnapshot::~ShenandoahNMethodTableSnapshot() {
429 _list->release();
430 }
431
432 void ShenandoahNMethodTableSnapshot::parallel_blobs_do(CodeBlobClosure *f) {
433 size_t stride = 256; // educated guess
434
435 ShenandoahNMethod** const list = _list->list();
436
437 size_t max = (size_t)_limit;
438 while (_claimed < max) {
439 size_t cur = Atomic::fetch_then_add(&_claimed, stride, memory_order_relaxed);
440 size_t start = cur;
441 size_t end = MIN2(cur + stride, max);
442 if (start >= max) break;
443
444 for (size_t idx = start; idx < end; idx++) {
445 ShenandoahNMethod* nmr = list[idx];
446 assert(nmr != nullptr, "Sanity");
447 if (nmr->is_unregistered()) {
448 continue;
449 }
450
451 nmr->assert_correct();
452 f->do_code_blob(nmr->nm());
453 }
454 }
455 }
456
457 void ShenandoahNMethodTableSnapshot::concurrent_nmethods_do(NMethodClosure* cl) {
458 size_t stride = 256; // educated guess
459
460 ShenandoahNMethod** list = _list->list();
461 size_t max = (size_t)_limit;
462 while (_claimed < max) {
463 size_t cur = Atomic::fetch_then_add(&_claimed, stride, memory_order_relaxed);
464 size_t start = cur;
465 size_t end = MIN2(cur + stride, max);
466 if (start >= max) break;
467
468 for (size_t idx = start; idx < end; idx++) {
469 ShenandoahNMethod* data = list[idx];
470 assert(data != nullptr, "Should not be null");
471 if (!data->is_unregistered()) {
472 cl->do_nmethod(data->nm());
473 }
474 }
475 }
476 }
477
478 ShenandoahConcurrentNMethodIterator::ShenandoahConcurrentNMethodIterator(ShenandoahNMethodTable* table) :
479 _table(table),
480 _table_snapshot(nullptr),
481 _started_workers(0),
482 _finished_workers(0) {}
483
484 void ShenandoahConcurrentNMethodIterator::nmethods_do(NMethodClosure* cl) {
485 // Cannot safepoint when iteration is running, because this can cause deadlocks
486 // with other threads waiting on iteration to be over.
487 NoSafepointVerifier nsv;
488
489 MutexLocker ml(CodeCache_lock, Mutex::_no_safepoint_check_flag);
490
491 if (_finished_workers > 0) {
492 // Some threads have already finished. We are now in rampdown: we are now
493 // waiting for all currently recorded workers to finish. No new workers
494 // should start.
495 return;
496 }
497
498 // Record a new worker and initialize the snapshot if it is a first visitor.
499 if (_started_workers++ == 0) {
500 _table_snapshot = _table->snapshot_for_iteration();
501 }
502
503 // All set, relinquish the lock and go concurrent.
504 {
505 MutexUnlocker mu(CodeCache_lock, Mutex::_no_safepoint_check_flag);
506 _table_snapshot->concurrent_nmethods_do(cl);
507 }
508
509 // Record completion. Last worker shuts down the iterator and notifies any waiters.
510 uint count = ++_finished_workers;
511 if (count == _started_workers) {
512 _table->finish_iteration(_table_snapshot);
513 CodeCache_lock->notify_all();
514 }
515 }