1 /*
2 * Copyright (c) 2017, 2022, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
24
25 #ifndef SHARE_OOPS_ACCESSBACKEND_HPP
26 #define SHARE_OOPS_ACCESSBACKEND_HPP
27
28 #include "gc/shared/barrierSetConfig.hpp"
29 #include "memory/allocation.hpp"
30 #include "metaprogramming/enableIf.hpp"
31 #include "oops/accessDecorators.hpp"
32 #include "oops/inlineKlass.hpp"
33 #include "oops/oopsHierarchy.hpp"
34 #include "runtime/globals.hpp"
35 #include "utilities/debug.hpp"
36 #include "utilities/globalDefinitions.hpp"
37
38 #include <type_traits>
39
40 // This metafunction returns either oop or narrowOop depending on whether
41 // an access needs to use compressed oops or not.
42 template <DecoratorSet decorators>
43 struct HeapOopType: AllStatic {
44 static const bool needs_oop_compress = HasDecorator<decorators, INTERNAL_CONVERT_COMPRESSED_OOP>::value &&
45 HasDecorator<decorators, INTERNAL_RT_USE_COMPRESSED_OOPS>::value;
46 using type = std::conditional_t<needs_oop_compress, narrowOop, oop>;
47 };
48
49 // This meta-function returns either oop or narrowOop depending on whether
50 // a back-end needs to consider compressed oops types or not.
51 template <DecoratorSet decorators>
52 struct ValueOopType: AllStatic {
53 static const bool needs_oop_compress = HasDecorator<decorators, INTERNAL_RT_USE_COMPRESSED_OOPS>::value;
54 using type = std::conditional_t<needs_oop_compress, narrowOop, oop>;
55 };
56
57 namespace AccessInternal {
58 enum BarrierType {
59 BARRIER_STORE,
60 BARRIER_STORE_AT,
61 BARRIER_LOAD,
62 BARRIER_LOAD_AT,
63 BARRIER_ATOMIC_CMPXCHG,
64 BARRIER_ATOMIC_CMPXCHG_AT,
65 BARRIER_ATOMIC_XCHG,
66 BARRIER_ATOMIC_XCHG_AT,
67 BARRIER_ARRAYCOPY,
68 BARRIER_CLONE,
69 BARRIER_VALUE_COPY
70 };
71
72 template <DecoratorSet decorators, typename T>
73 struct MustConvertCompressedOop: public std::integral_constant<bool,
74 HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value &&
75 std::is_same<typename HeapOopType<decorators>::type, narrowOop>::value &&
76 std::is_same<T, oop>::value> {};
77
78 // This metafunction returns an appropriate oop type if the value is oop-like
79 // and otherwise returns the same type T.
80 template <DecoratorSet decorators, typename T>
81 struct EncodedType: AllStatic {
82 using type = std::conditional_t<HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value,
83 typename HeapOopType<decorators>::type,
84 T>;
85 };
86
87 template <DecoratorSet decorators>
88 inline typename HeapOopType<decorators>::type*
89 oop_field_addr(oop base, ptrdiff_t byte_offset) {
90 return reinterpret_cast<typename HeapOopType<decorators>::type*>(
91 reinterpret_cast<intptr_t>((void*)base) + byte_offset);
92 }
93
94 template <DecoratorSet decorators, typename T>
95 struct AccessFunctionTypes {
96 typedef T (*load_at_func_t)(oop base, ptrdiff_t offset);
97 typedef void (*store_at_func_t)(oop base, ptrdiff_t offset, T value);
98 typedef T (*atomic_cmpxchg_at_func_t)(oop base, ptrdiff_t offset, T compare_value, T new_value);
99 typedef T (*atomic_xchg_at_func_t)(oop base, ptrdiff_t offset, T new_value);
100
101 typedef T (*load_func_t)(void* addr);
102 typedef void (*store_func_t)(void* addr, T value);
103 typedef T (*atomic_cmpxchg_func_t)(void* addr, T compare_value, T new_value);
104 typedef T (*atomic_xchg_func_t)(void* addr, T new_value);
105
106 typedef void (*arraycopy_func_t)(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw,
107 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
108 size_t length);
109 typedef void (*clone_func_t)(oop src, oop dst, size_t size);
110 typedef void (*value_copy_func_t)(void* src, void* dst, InlineKlass* md, LayoutKind lk);
111 };
112
113 template <DecoratorSet decorators>
114 struct AccessFunctionTypes<decorators, void> {
115 typedef void (*arraycopy_func_t)(arrayOop src_obj, size_t src_offset_in_bytes, void* src,
116 arrayOop dst_obj, size_t dst_offset_in_bytes, void* dst,
117 size_t length);
118 };
119
120 template <DecoratorSet decorators, typename T, BarrierType barrier> struct AccessFunction {};
121
122 #define ACCESS_GENERATE_ACCESS_FUNCTION(bt, func) \
123 template <DecoratorSet decorators, typename T> \
124 struct AccessFunction<decorators, T, bt>: AllStatic{ \
125 typedef typename AccessFunctionTypes<decorators, T>::func type; \
126 }
127 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_STORE, store_func_t);
128 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_STORE_AT, store_at_func_t);
129 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_LOAD, load_func_t);
130 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_LOAD_AT, load_at_func_t);
131 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_ATOMIC_CMPXCHG, atomic_cmpxchg_func_t);
132 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_ATOMIC_CMPXCHG_AT, atomic_cmpxchg_at_func_t);
133 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_ATOMIC_XCHG, atomic_xchg_func_t);
134 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_ATOMIC_XCHG_AT, atomic_xchg_at_func_t);
135 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_ARRAYCOPY, arraycopy_func_t);
136 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_CLONE, clone_func_t);
137 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_VALUE_COPY, value_copy_func_t);
138 #undef ACCESS_GENERATE_ACCESS_FUNCTION
139
140 template <DecoratorSet decorators, typename T, BarrierType barrier_type>
141 typename AccessFunction<decorators, T, barrier_type>::type resolve_barrier();
142
143 template <DecoratorSet decorators, typename T, BarrierType barrier_type>
144 typename AccessFunction<decorators, T, barrier_type>::type resolve_oop_barrier();
145
146 void* field_addr(oop base, ptrdiff_t offset);
147
148 // Forward calls to Copy:: in the cpp file to reduce dependencies and allow
149 // faster build times, given how frequently included access is.
150 void arraycopy_arrayof_conjoint_oops(void* src, void* dst, size_t length);
151 void arraycopy_conjoint_oops(oop* src, oop* dst, size_t length);
152 void arraycopy_conjoint_oops(narrowOop* src, narrowOop* dst, size_t length);
153
154 void arraycopy_disjoint_words(void* src, void* dst, size_t length);
155 void arraycopy_disjoint_words_atomic(void* src, void* dst, size_t length);
156
157 template<typename T>
158 void arraycopy_conjoint(T* src, T* dst, size_t length);
159 template<typename T>
160 void arraycopy_arrayof_conjoint(T* src, T* dst, size_t length);
161 template<typename T>
162 void arraycopy_conjoint_atomic(T* src, T* dst, size_t length);
163
164 void value_copy_internal(void* src, void* dst, size_t length);
165 }
166
167 // This mask specifies what decorators are relevant for raw accesses. When passing
168 // accesses to the raw layer, irrelevant decorators are removed.
169 const DecoratorSet RAW_DECORATOR_MASK = INTERNAL_DECORATOR_MASK | MO_DECORATOR_MASK |
170 ARRAYCOPY_DECORATOR_MASK | IS_NOT_NULL;
171
172 // The RawAccessBarrier performs raw accesses with additional knowledge of
173 // memory ordering, so that OrderAccess/Atomic is called when necessary.
174 // It additionally handles compressed oops, and hence is not completely "raw"
175 // strictly speaking.
176 template <DecoratorSet decorators>
177 class RawAccessBarrier: public AllStatic {
178 protected:
179 static inline void* field_addr(oop base, ptrdiff_t byte_offset) {
180 return AccessInternal::field_addr(base, byte_offset);
181 }
182
183 protected:
184 // Only encode if INTERNAL_VALUE_IS_OOP
185 template <DecoratorSet idecorators, typename T>
186 static inline typename EnableIf<
187 AccessInternal::MustConvertCompressedOop<idecorators, T>::value,
188 typename HeapOopType<idecorators>::type>::type
189 encode_internal(T value);
190
191 template <DecoratorSet idecorators, typename T>
192 static inline typename EnableIf<
193 !AccessInternal::MustConvertCompressedOop<idecorators, T>::value, T>::type
194 encode_internal(T value) {
195 return value;
196 }
197
198 template <typename T>
199 static inline typename AccessInternal::EncodedType<decorators, T>::type
200 encode(T value) {
201 return encode_internal<decorators, T>(value);
202 }
203
204 // Only decode if INTERNAL_VALUE_IS_OOP
205 template <DecoratorSet idecorators, typename T>
206 static inline typename EnableIf<
207 AccessInternal::MustConvertCompressedOop<idecorators, T>::value, T>::type
208 decode_internal(typename HeapOopType<idecorators>::type value);
209
210 template <DecoratorSet idecorators, typename T>
211 static inline typename EnableIf<
212 !AccessInternal::MustConvertCompressedOop<idecorators, T>::value, T>::type
213 decode_internal(T value) {
214 return value;
215 }
216
217 template <typename T>
218 static inline T decode(typename AccessInternal::EncodedType<decorators, T>::type value) {
219 return decode_internal<decorators, T>(value);
220 }
221
222 protected:
223 template <DecoratorSet ds, typename T>
224 static typename EnableIf<
225 HasDecorator<ds, MO_SEQ_CST>::value, T>::type
226 load_internal(void* addr);
227
228 template <DecoratorSet ds, typename T>
229 static typename EnableIf<
230 HasDecorator<ds, MO_ACQUIRE>::value, T>::type
231 load_internal(void* addr);
232
233 template <DecoratorSet ds, typename T>
234 static typename EnableIf<
235 HasDecorator<ds, MO_RELAXED>::value, T>::type
236 load_internal(void* addr);
237
238 template <DecoratorSet ds, typename T>
239 static inline typename EnableIf<
240 HasDecorator<ds, MO_UNORDERED>::value, T>::type
241 load_internal(void* addr) {
242 return *reinterpret_cast<T*>(addr);
243 }
244
245 template <DecoratorSet ds, typename T>
246 static typename EnableIf<
247 HasDecorator<ds, MO_SEQ_CST>::value>::type
248 store_internal(void* addr, T value);
249
250 template <DecoratorSet ds, typename T>
251 static typename EnableIf<
252 HasDecorator<ds, MO_RELEASE>::value>::type
253 store_internal(void* addr, T value);
254
255 template <DecoratorSet ds, typename T>
256 static typename EnableIf<
257 HasDecorator<ds, MO_RELAXED>::value>::type
258 store_internal(void* addr, T value);
259
260 template <DecoratorSet ds, typename T>
261 static inline typename EnableIf<
262 HasDecorator<ds, MO_UNORDERED>::value>::type
263 store_internal(void* addr, T value) {
264 *reinterpret_cast<T*>(addr) = value;
265 }
266
267 template <DecoratorSet ds, typename T>
268 static typename EnableIf<
269 HasDecorator<ds, MO_SEQ_CST>::value, T>::type
270 atomic_cmpxchg_internal(void* addr, T compare_value, T new_value);
271
272 template <DecoratorSet ds, typename T>
273 static typename EnableIf<
274 HasDecorator<ds, MO_RELAXED>::value, T>::type
275 atomic_cmpxchg_internal(void* addr, T compare_value, T new_value);
276
277 template <DecoratorSet ds, typename T>
278 static typename EnableIf<
279 HasDecorator<ds, MO_SEQ_CST>::value, T>::type
280 atomic_xchg_internal(void* addr, T new_value);
281
282 public:
283 template <typename T>
284 static inline void store(void* addr, T value) {
285 store_internal<decorators>(addr, value);
286 }
287
288 template <typename T>
289 static inline T load(void* addr) {
290 return load_internal<decorators, T>(addr);
291 }
292
293 template <typename T>
294 static inline T atomic_cmpxchg(void* addr, T compare_value, T new_value) {
295 return atomic_cmpxchg_internal<decorators>(addr, compare_value, new_value);
296 }
297
298 template <typename T>
299 static inline T atomic_xchg(void* addr, T new_value) {
300 return atomic_xchg_internal<decorators>(addr, new_value);
301 }
302
303 template <typename T>
304 static void arraycopy(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw,
305 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
306 size_t length);
307
308 template <typename T>
309 static void oop_store(void* addr, T value);
310 template <typename T>
311 static void oop_store_at(oop base, ptrdiff_t offset, T value);
312
313 template <typename T>
314 static T oop_load(void* addr);
315 template <typename T>
316 static T oop_load_at(oop base, ptrdiff_t offset);
317
318 template <typename T>
319 static T oop_atomic_cmpxchg(void* addr, T compare_value, T new_value);
320 template <typename T>
321 static T oop_atomic_cmpxchg_at(oop base, ptrdiff_t offset, T compare_value, T new_value);
322
323 template <typename T>
324 static T oop_atomic_xchg(void* addr, T new_value);
325 template <typename T>
326 static T oop_atomic_xchg_at(oop base, ptrdiff_t offset, T new_value);
327
328 template <typename T>
329 static void store_at(oop base, ptrdiff_t offset, T value) {
330 store(field_addr(base, offset), value);
331 }
332
333 template <typename T>
334 static T load_at(oop base, ptrdiff_t offset) {
335 return load<T>(field_addr(base, offset));
336 }
337
338 template <typename T>
339 static T atomic_cmpxchg_at(oop base, ptrdiff_t offset, T compare_value, T new_value) {
340 return atomic_cmpxchg(field_addr(base, offset), compare_value, new_value);
341 }
342
343 template <typename T>
344 static T atomic_xchg_at(oop base, ptrdiff_t offset, T new_value) {
345 return atomic_xchg(field_addr(base, offset), new_value);
346 }
347
348 template <typename T>
349 static void oop_arraycopy(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw,
350 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
351 size_t length);
352
353 static void clone(oop src, oop dst, size_t size);
354 static void value_copy(void* src, void* dst, InlineKlass* md, LayoutKind lk);
355
356 };
357
358 namespace AccessInternal {
359 DEBUG_ONLY(void check_access_thread_state());
360 #define assert_access_thread_state() DEBUG_ONLY(check_access_thread_state())
361 }
362
363 // Below is the implementation of the first 4 steps of the template pipeline:
364 // * Step 1: Set default decorators and decay types. This step gets rid of CV qualifiers
365 // and sets default decorators to sensible values.
366 // * Step 2: Reduce types. This step makes sure there is only a single T type and not
367 // multiple types. The P type of the address and T type of the value must
368 // match.
369 // * Step 3: Pre-runtime dispatch. This step checks whether a runtime call can be
370 // avoided, and in that case avoids it (calling raw accesses or
371 // primitive accesses in a build that does not require primitive GC barriers)
372 // * Step 4: Runtime-dispatch. This step performs a runtime dispatch to the corresponding
373 // BarrierSet::AccessBarrier accessor that attaches GC-required barriers
374 // to the access.
375
376 namespace AccessInternal {
377 template <typename T>
378 struct OopOrNarrowOopInternal: AllStatic {
379 typedef oop type;
380 };
381
382 template <>
383 struct OopOrNarrowOopInternal<narrowOop>: AllStatic {
384 typedef narrowOop type;
385 };
386
387 // This metafunction returns a canonicalized oop/narrowOop type for a passed
388 // in oop-like types passed in from oop_* overloads where the user has sworn
389 // that the passed in values should be oop-like (e.g. oop, oopDesc*, arrayOop,
390 // narrowOoop, instanceOopDesc*, and random other things).
391 // In the oop_* overloads, it must hold that if the passed in type T is not
392 // narrowOop, then it by contract has to be one of many oop-like types implicitly
393 // convertible to oop, and hence returns oop as the canonical oop type.
394 // If it turns out it was not, then the implicit conversion to oop will fail
395 // to compile, as desired.
396 template <typename T>
397 struct OopOrNarrowOop: AllStatic {
398 typedef typename OopOrNarrowOopInternal<std::decay_t<T>>::type type;
399 };
400
401 inline void* field_addr(oop base, ptrdiff_t byte_offset) {
402 return reinterpret_cast<void*>(reinterpret_cast<intptr_t>((void*)base) + byte_offset);
403 }
404 // Step 4: Runtime dispatch
405 // The RuntimeDispatch class is responsible for performing a runtime dispatch of the
406 // accessor. This is required when the access either depends on whether compressed oops
407 // is being used, or it depends on which GC implementation was chosen (e.g. requires GC
408 // barriers). The way it works is that a function pointer initially pointing to an
409 // accessor resolution function gets called for each access. Upon first invocation,
410 // it resolves which accessor to be used in future invocations and patches the
411 // function pointer to this new accessor.
412
413 template <DecoratorSet decorators, typename T, BarrierType type>
414 struct RuntimeDispatch: AllStatic {};
415
416 template <DecoratorSet decorators, typename T>
417 struct RuntimeDispatch<decorators, T, BARRIER_STORE>: AllStatic {
418 typedef typename AccessFunction<decorators, T, BARRIER_STORE>::type func_t;
419 static func_t _store_func;
420
421 static void store_init(void* addr, T value);
422
423 static inline void store(void* addr, T value) {
424 assert_access_thread_state();
425 _store_func(addr, value);
426 }
427 };
428
429 template <DecoratorSet decorators, typename T>
430 struct RuntimeDispatch<decorators, T, BARRIER_STORE_AT>: AllStatic {
431 typedef typename AccessFunction<decorators, T, BARRIER_STORE_AT>::type func_t;
432 static func_t _store_at_func;
433
434 static void store_at_init(oop base, ptrdiff_t offset, T value);
435
436 static inline void store_at(oop base, ptrdiff_t offset, T value) {
437 assert_access_thread_state();
438 _store_at_func(base, offset, value);
439 }
440 };
441
442 template <DecoratorSet decorators, typename T>
443 struct RuntimeDispatch<decorators, T, BARRIER_LOAD>: AllStatic {
444 typedef typename AccessFunction<decorators, T, BARRIER_LOAD>::type func_t;
445 static func_t _load_func;
446
447 static T load_init(void* addr);
448
449 static inline T load(void* addr) {
450 assert_access_thread_state();
451 return _load_func(addr);
452 }
453 };
454
455 template <DecoratorSet decorators, typename T>
456 struct RuntimeDispatch<decorators, T, BARRIER_LOAD_AT>: AllStatic {
457 typedef typename AccessFunction<decorators, T, BARRIER_LOAD_AT>::type func_t;
458 static func_t _load_at_func;
459
460 static T load_at_init(oop base, ptrdiff_t offset);
461
462 static inline T load_at(oop base, ptrdiff_t offset) {
463 assert_access_thread_state();
464 return _load_at_func(base, offset);
465 }
466 };
467
468 template <DecoratorSet decorators, typename T>
469 struct RuntimeDispatch<decorators, T, BARRIER_ATOMIC_CMPXCHG>: AllStatic {
470 typedef typename AccessFunction<decorators, T, BARRIER_ATOMIC_CMPXCHG>::type func_t;
471 static func_t _atomic_cmpxchg_func;
472
473 static T atomic_cmpxchg_init(void* addr, T compare_value, T new_value);
474
475 static inline T atomic_cmpxchg(void* addr, T compare_value, T new_value) {
476 assert_access_thread_state();
477 return _atomic_cmpxchg_func(addr, compare_value, new_value);
478 }
479 };
480
481 template <DecoratorSet decorators, typename T>
482 struct RuntimeDispatch<decorators, T, BARRIER_ATOMIC_CMPXCHG_AT>: AllStatic {
483 typedef typename AccessFunction<decorators, T, BARRIER_ATOMIC_CMPXCHG_AT>::type func_t;
484 static func_t _atomic_cmpxchg_at_func;
485
486 static T atomic_cmpxchg_at_init(oop base, ptrdiff_t offset, T compare_value, T new_value);
487
488 static inline T atomic_cmpxchg_at(oop base, ptrdiff_t offset, T compare_value, T new_value) {
489 assert_access_thread_state();
490 return _atomic_cmpxchg_at_func(base, offset, compare_value, new_value);
491 }
492 };
493
494 template <DecoratorSet decorators, typename T>
495 struct RuntimeDispatch<decorators, T, BARRIER_ATOMIC_XCHG>: AllStatic {
496 typedef typename AccessFunction<decorators, T, BARRIER_ATOMIC_XCHG>::type func_t;
497 static func_t _atomic_xchg_func;
498
499 static T atomic_xchg_init(void* addr, T new_value);
500
501 static inline T atomic_xchg(void* addr, T new_value) {
502 assert_access_thread_state();
503 return _atomic_xchg_func(addr, new_value);
504 }
505 };
506
507 template <DecoratorSet decorators, typename T>
508 struct RuntimeDispatch<decorators, T, BARRIER_ATOMIC_XCHG_AT>: AllStatic {
509 typedef typename AccessFunction<decorators, T, BARRIER_ATOMIC_XCHG_AT>::type func_t;
510 static func_t _atomic_xchg_at_func;
511
512 static T atomic_xchg_at_init(oop base, ptrdiff_t offset, T new_value);
513
514 static inline T atomic_xchg_at(oop base, ptrdiff_t offset, T new_value) {
515 assert_access_thread_state();
516 return _atomic_xchg_at_func(base, offset, new_value);
517 }
518 };
519
520 template <DecoratorSet decorators, typename T>
521 struct RuntimeDispatch<decorators, T, BARRIER_ARRAYCOPY>: AllStatic {
522 typedef typename AccessFunction<decorators, T, BARRIER_ARRAYCOPY>::type func_t;
523 static func_t _arraycopy_func;
524
525 static void arraycopy_init(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw,
526 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
527 size_t length);
528
529 static inline void arraycopy(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw,
530 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
531 size_t length) {
532 assert_access_thread_state();
533 return _arraycopy_func(src_obj, src_offset_in_bytes, src_raw,
534 dst_obj, dst_offset_in_bytes, dst_raw,
535 length);
536 }
537 };
538
539 template <DecoratorSet decorators, typename T>
540 struct RuntimeDispatch<decorators, T, BARRIER_CLONE>: AllStatic {
541 typedef typename AccessFunction<decorators, T, BARRIER_CLONE>::type func_t;
542 static func_t _clone_func;
543
544 static void clone_init(oop src, oop dst, size_t size);
545
546 static inline void clone(oop src, oop dst, size_t size) {
547 assert_access_thread_state();
548 _clone_func(src, dst, size);
549 }
550 };
551
552 template <DecoratorSet decorators, typename T>
553 struct RuntimeDispatch<decorators, T, BARRIER_VALUE_COPY>: AllStatic {
554 typedef typename AccessFunction<decorators, T, BARRIER_VALUE_COPY>::type func_t;
555 static func_t _value_copy_func;
556
557 static void value_copy_init(void* src, void* dst, InlineKlass* md, LayoutKind lk);
558
559 static inline void value_copy(void* src, void* dst, InlineKlass* md, LayoutKind lk) {
560 _value_copy_func(src, dst, md, lk);
561 }
562 };
563
564 // Initialize the function pointers to point to the resolving function.
565 template <DecoratorSet decorators, typename T>
566 typename AccessFunction<decorators, T, BARRIER_STORE>::type
567 RuntimeDispatch<decorators, T, BARRIER_STORE>::_store_func = &store_init;
568
569 template <DecoratorSet decorators, typename T>
570 typename AccessFunction<decorators, T, BARRIER_STORE_AT>::type
571 RuntimeDispatch<decorators, T, BARRIER_STORE_AT>::_store_at_func = &store_at_init;
572
573 template <DecoratorSet decorators, typename T>
574 typename AccessFunction<decorators, T, BARRIER_LOAD>::type
575 RuntimeDispatch<decorators, T, BARRIER_LOAD>::_load_func = &load_init;
576
577 template <DecoratorSet decorators, typename T>
578 typename AccessFunction<decorators, T, BARRIER_LOAD_AT>::type
579 RuntimeDispatch<decorators, T, BARRIER_LOAD_AT>::_load_at_func = &load_at_init;
580
581 template <DecoratorSet decorators, typename T>
582 typename AccessFunction<decorators, T, BARRIER_ATOMIC_CMPXCHG>::type
583 RuntimeDispatch<decorators, T, BARRIER_ATOMIC_CMPXCHG>::_atomic_cmpxchg_func = &atomic_cmpxchg_init;
584
585 template <DecoratorSet decorators, typename T>
586 typename AccessFunction<decorators, T, BARRIER_ATOMIC_CMPXCHG_AT>::type
587 RuntimeDispatch<decorators, T, BARRIER_ATOMIC_CMPXCHG_AT>::_atomic_cmpxchg_at_func = &atomic_cmpxchg_at_init;
588
589 template <DecoratorSet decorators, typename T>
590 typename AccessFunction<decorators, T, BARRIER_ATOMIC_XCHG>::type
591 RuntimeDispatch<decorators, T, BARRIER_ATOMIC_XCHG>::_atomic_xchg_func = &atomic_xchg_init;
592
593 template <DecoratorSet decorators, typename T>
594 typename AccessFunction<decorators, T, BARRIER_ATOMIC_XCHG_AT>::type
595 RuntimeDispatch<decorators, T, BARRIER_ATOMIC_XCHG_AT>::_atomic_xchg_at_func = &atomic_xchg_at_init;
596
597 template <DecoratorSet decorators, typename T>
598 typename AccessFunction<decorators, T, BARRIER_ARRAYCOPY>::type
599 RuntimeDispatch<decorators, T, BARRIER_ARRAYCOPY>::_arraycopy_func = &arraycopy_init;
600
601 template <DecoratorSet decorators, typename T>
602 typename AccessFunction<decorators, T, BARRIER_CLONE>::type
603 RuntimeDispatch<decorators, T, BARRIER_CLONE>::_clone_func = &clone_init;
604
605 template <DecoratorSet decorators, typename T>
606 typename AccessFunction<decorators, T, BARRIER_VALUE_COPY>::type
607 RuntimeDispatch<decorators, T, BARRIER_VALUE_COPY>::_value_copy_func = &value_copy_init;
608
609 // Step 3: Pre-runtime dispatching.
610 // The PreRuntimeDispatch class is responsible for filtering the barrier strength
611 // decorators. That is, for AS_RAW, it hardwires the accesses without a runtime
612 // dispatch point. Otherwise it goes through a runtime check if hardwiring was
613 // not possible.
614 struct PreRuntimeDispatch: AllStatic {
615 template<DecoratorSet decorators>
616 struct CanHardwireRaw: public std::integral_constant<
617 bool,
618 !HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value || // primitive access
619 !HasDecorator<decorators, INTERNAL_CONVERT_COMPRESSED_OOP>::value || // don't care about compressed oops (oop* address)
620 HasDecorator<decorators, INTERNAL_RT_USE_COMPRESSED_OOPS>::value> // we can infer we use compressed oops (narrowOop* address)
621 {};
622
623 static const DecoratorSet convert_compressed_oops = INTERNAL_RT_USE_COMPRESSED_OOPS | INTERNAL_CONVERT_COMPRESSED_OOP;
624
625 template<DecoratorSet decorators>
626 static bool is_hardwired_primitive() {
627 return !HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value;
628 }
629
630 template <DecoratorSet decorators, typename T>
631 inline static typename EnableIf<
632 HasDecorator<decorators, AS_RAW>::value && CanHardwireRaw<decorators>::value>::type
633 store(void* addr, T value) {
634 typedef RawAccessBarrier<decorators & RAW_DECORATOR_MASK> Raw;
635 if (HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value) {
636 Raw::oop_store(addr, value);
637 } else {
638 Raw::store(addr, value);
639 }
640 }
641
642 template <DecoratorSet decorators, typename T>
643 inline static typename EnableIf<
644 HasDecorator<decorators, AS_RAW>::value && !CanHardwireRaw<decorators>::value>::type
645 store(void* addr, T value) {
646 if (UseCompressedOops) {
647 const DecoratorSet expanded_decorators = decorators | convert_compressed_oops;
648 PreRuntimeDispatch::store<expanded_decorators>(addr, value);
649 } else {
650 const DecoratorSet expanded_decorators = decorators & ~convert_compressed_oops;
651 PreRuntimeDispatch::store<expanded_decorators>(addr, value);
652 }
653 }
654
655 template <DecoratorSet decorators, typename T>
656 inline static typename EnableIf<
657 !HasDecorator<decorators, AS_RAW>::value>::type
658 store(void* addr, T value) {
659 if (is_hardwired_primitive<decorators>()) {
660 const DecoratorSet expanded_decorators = decorators | AS_RAW;
661 PreRuntimeDispatch::store<expanded_decorators>(addr, value);
662 } else {
663 RuntimeDispatch<decorators, T, BARRIER_STORE>::store(addr, value);
664 }
665 }
666
667 template <DecoratorSet decorators, typename T>
668 inline static typename EnableIf<
669 HasDecorator<decorators, AS_RAW>::value>::type
670 store_at(oop base, ptrdiff_t offset, T value) {
671 store<decorators>(field_addr(base, offset), value);
672 }
673
674 template <DecoratorSet decorators, typename T>
675 inline static typename EnableIf<
676 !HasDecorator<decorators, AS_RAW>::value>::type
677 store_at(oop base, ptrdiff_t offset, T value) {
678 if (is_hardwired_primitive<decorators>()) {
679 const DecoratorSet expanded_decorators = decorators | AS_RAW;
680 PreRuntimeDispatch::store_at<expanded_decorators>(base, offset, value);
681 } else {
682 RuntimeDispatch<decorators, T, BARRIER_STORE_AT>::store_at(base, offset, value);
683 }
684 }
685
686 template <DecoratorSet decorators, typename T>
687 inline static typename EnableIf<
688 HasDecorator<decorators, AS_RAW>::value && CanHardwireRaw<decorators>::value, T>::type
689 load(void* addr) {
690 typedef RawAccessBarrier<decorators & RAW_DECORATOR_MASK> Raw;
691 if (HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value) {
692 return Raw::template oop_load<T>(addr);
693 } else {
694 return Raw::template load<T>(addr);
695 }
696 }
697
698 template <DecoratorSet decorators, typename T>
699 inline static typename EnableIf<
700 HasDecorator<decorators, AS_RAW>::value && !CanHardwireRaw<decorators>::value, T>::type
701 load(void* addr) {
702 if (UseCompressedOops) {
703 const DecoratorSet expanded_decorators = decorators | convert_compressed_oops;
704 return PreRuntimeDispatch::load<expanded_decorators, T>(addr);
705 } else {
706 const DecoratorSet expanded_decorators = decorators & ~convert_compressed_oops;
707 return PreRuntimeDispatch::load<expanded_decorators, T>(addr);
708 }
709 }
710
711 template <DecoratorSet decorators, typename T>
712 inline static typename EnableIf<
713 !HasDecorator<decorators, AS_RAW>::value, T>::type
714 load(void* addr) {
715 if (is_hardwired_primitive<decorators>()) {
716 const DecoratorSet expanded_decorators = decorators | AS_RAW;
717 return PreRuntimeDispatch::load<expanded_decorators, T>(addr);
718 } else {
719 return RuntimeDispatch<decorators, T, BARRIER_LOAD>::load(addr);
720 }
721 }
722
723 template <DecoratorSet decorators, typename T>
724 inline static typename EnableIf<
725 HasDecorator<decorators, AS_RAW>::value, T>::type
726 load_at(oop base, ptrdiff_t offset) {
727 return load<decorators, T>(field_addr(base, offset));
728 }
729
730 template <DecoratorSet decorators, typename T>
731 inline static typename EnableIf<
732 !HasDecorator<decorators, AS_RAW>::value, T>::type
733 load_at(oop base, ptrdiff_t offset) {
734 if (is_hardwired_primitive<decorators>()) {
735 const DecoratorSet expanded_decorators = decorators | AS_RAW;
736 return PreRuntimeDispatch::load_at<expanded_decorators, T>(base, offset);
737 } else {
738 return RuntimeDispatch<decorators, T, BARRIER_LOAD_AT>::load_at(base, offset);
739 }
740 }
741
742 template <DecoratorSet decorators, typename T>
743 inline static typename EnableIf<
744 HasDecorator<decorators, AS_RAW>::value && CanHardwireRaw<decorators>::value, T>::type
745 atomic_cmpxchg(void* addr, T compare_value, T new_value) {
746 typedef RawAccessBarrier<decorators & RAW_DECORATOR_MASK> Raw;
747 if (HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value) {
748 return Raw::oop_atomic_cmpxchg(addr, compare_value, new_value);
749 } else {
750 return Raw::atomic_cmpxchg(addr, compare_value, new_value);
751 }
752 }
753
754 template <DecoratorSet decorators, typename T>
755 inline static typename EnableIf<
756 HasDecorator<decorators, AS_RAW>::value && !CanHardwireRaw<decorators>::value, T>::type
757 atomic_cmpxchg(void* addr, T compare_value, T new_value) {
758 if (UseCompressedOops) {
759 const DecoratorSet expanded_decorators = decorators | convert_compressed_oops;
760 return PreRuntimeDispatch::atomic_cmpxchg<expanded_decorators>(addr, compare_value, new_value);
761 } else {
762 const DecoratorSet expanded_decorators = decorators & ~convert_compressed_oops;
763 return PreRuntimeDispatch::atomic_cmpxchg<expanded_decorators>(addr, compare_value, new_value);
764 }
765 }
766
767 template <DecoratorSet decorators, typename T>
768 inline static typename EnableIf<
769 !HasDecorator<decorators, AS_RAW>::value, T>::type
770 atomic_cmpxchg(void* addr, T compare_value, T new_value) {
771 if (is_hardwired_primitive<decorators>()) {
772 const DecoratorSet expanded_decorators = decorators | AS_RAW;
773 return PreRuntimeDispatch::atomic_cmpxchg<expanded_decorators>(addr, compare_value, new_value);
774 } else {
775 return RuntimeDispatch<decorators, T, BARRIER_ATOMIC_CMPXCHG>::atomic_cmpxchg(addr, compare_value, new_value);
776 }
777 }
778
779 template <DecoratorSet decorators, typename T>
780 inline static typename EnableIf<
781 HasDecorator<decorators, AS_RAW>::value, T>::type
782 atomic_cmpxchg_at(oop base, ptrdiff_t offset, T compare_value, T new_value) {
783 return atomic_cmpxchg<decorators>(field_addr(base, offset), compare_value, new_value);
784 }
785
786 template <DecoratorSet decorators, typename T>
787 inline static typename EnableIf<
788 !HasDecorator<decorators, AS_RAW>::value, T>::type
789 atomic_cmpxchg_at(oop base, ptrdiff_t offset, T compare_value, T new_value) {
790 if (is_hardwired_primitive<decorators>()) {
791 const DecoratorSet expanded_decorators = decorators | AS_RAW;
792 return PreRuntimeDispatch::atomic_cmpxchg_at<expanded_decorators>(base, offset, compare_value, new_value);
793 } else {
794 return RuntimeDispatch<decorators, T, BARRIER_ATOMIC_CMPXCHG_AT>::atomic_cmpxchg_at(base, offset, compare_value, new_value);
795 }
796 }
797
798 template <DecoratorSet decorators, typename T>
799 inline static typename EnableIf<
800 HasDecorator<decorators, AS_RAW>::value && CanHardwireRaw<decorators>::value, T>::type
801 atomic_xchg(void* addr, T new_value) {
802 typedef RawAccessBarrier<decorators & RAW_DECORATOR_MASK> Raw;
803 if (HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value) {
804 return Raw::oop_atomic_xchg(addr, new_value);
805 } else {
806 return Raw::atomic_xchg(addr, new_value);
807 }
808 }
809
810 template <DecoratorSet decorators, typename T>
811 inline static typename EnableIf<
812 HasDecorator<decorators, AS_RAW>::value && !CanHardwireRaw<decorators>::value, T>::type
813 atomic_xchg(void* addr, T new_value) {
814 if (UseCompressedOops) {
815 const DecoratorSet expanded_decorators = decorators | convert_compressed_oops;
816 return PreRuntimeDispatch::atomic_xchg<expanded_decorators>(addr, new_value);
817 } else {
818 const DecoratorSet expanded_decorators = decorators & ~convert_compressed_oops;
819 return PreRuntimeDispatch::atomic_xchg<expanded_decorators>(addr, new_value);
820 }
821 }
822
823 template <DecoratorSet decorators, typename T>
824 inline static typename EnableIf<
825 !HasDecorator<decorators, AS_RAW>::value, T>::type
826 atomic_xchg(void* addr, T new_value) {
827 if (is_hardwired_primitive<decorators>()) {
828 const DecoratorSet expanded_decorators = decorators | AS_RAW;
829 return PreRuntimeDispatch::atomic_xchg<expanded_decorators>(addr, new_value);
830 } else {
831 return RuntimeDispatch<decorators, T, BARRIER_ATOMIC_XCHG>::atomic_xchg(addr, new_value);
832 }
833 }
834
835 template <DecoratorSet decorators, typename T>
836 inline static typename EnableIf<
837 HasDecorator<decorators, AS_RAW>::value, T>::type
838 atomic_xchg_at(oop base, ptrdiff_t offset, T new_value) {
839 return atomic_xchg<decorators>(field_addr(base, offset), new_value);
840 }
841
842 template <DecoratorSet decorators, typename T>
843 inline static typename EnableIf<
844 !HasDecorator<decorators, AS_RAW>::value, T>::type
845 atomic_xchg_at(oop base, ptrdiff_t offset, T new_value) {
846 if (is_hardwired_primitive<decorators>()) {
847 const DecoratorSet expanded_decorators = decorators | AS_RAW;
848 return PreRuntimeDispatch::atomic_xchg<expanded_decorators>(base, offset, new_value);
849 } else {
850 return RuntimeDispatch<decorators, T, BARRIER_ATOMIC_XCHG_AT>::atomic_xchg_at(base, offset, new_value);
851 }
852 }
853
854 template <DecoratorSet decorators, typename T>
855 inline static typename EnableIf<
856 HasDecorator<decorators, AS_RAW>::value && CanHardwireRaw<decorators>::value, void>::type
857 arraycopy(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw,
858 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
859 size_t length) {
860 typedef RawAccessBarrier<decorators & RAW_DECORATOR_MASK> Raw;
861 if (HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value) {
862 Raw::oop_arraycopy(src_obj, src_offset_in_bytes, src_raw,
863 dst_obj, dst_offset_in_bytes, dst_raw,
864 length);
865 } else {
866 Raw::arraycopy(src_obj, src_offset_in_bytes, src_raw,
867 dst_obj, dst_offset_in_bytes, dst_raw,
868 length);
869 }
870 }
871
872 template <DecoratorSet decorators, typename T>
873 inline static typename EnableIf<
874 HasDecorator<decorators, AS_RAW>::value && !CanHardwireRaw<decorators>::value, void>::type
875 arraycopy(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw,
876 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
877 size_t length) {
878 if (UseCompressedOops) {
879 const DecoratorSet expanded_decorators = decorators | convert_compressed_oops;
880 PreRuntimeDispatch::arraycopy<expanded_decorators>(src_obj, src_offset_in_bytes, src_raw,
881 dst_obj, dst_offset_in_bytes, dst_raw,
882 length);
883 } else {
884 const DecoratorSet expanded_decorators = decorators & ~convert_compressed_oops;
885 PreRuntimeDispatch::arraycopy<expanded_decorators>(src_obj, src_offset_in_bytes, src_raw,
886 dst_obj, dst_offset_in_bytes, dst_raw,
887 length);
888 }
889 }
890
891 template <DecoratorSet decorators, typename T>
892 inline static typename EnableIf<
893 !HasDecorator<decorators, AS_RAW>::value, void>::type
894 arraycopy(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw,
895 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
896 size_t length) {
897 if (is_hardwired_primitive<decorators>()) {
898 const DecoratorSet expanded_decorators = decorators | AS_RAW;
899 PreRuntimeDispatch::arraycopy<expanded_decorators>(src_obj, src_offset_in_bytes, src_raw,
900 dst_obj, dst_offset_in_bytes, dst_raw,
901 length);
902 } else {
903 RuntimeDispatch<decorators, T, BARRIER_ARRAYCOPY>::arraycopy(src_obj, src_offset_in_bytes, src_raw,
904 dst_obj, dst_offset_in_bytes, dst_raw,
905 length);
906 }
907 }
908
909 template <DecoratorSet decorators>
910 inline static typename EnableIf<
911 HasDecorator<decorators, AS_RAW>::value>::type
912 clone(oop src, oop dst, size_t size) {
913 typedef RawAccessBarrier<decorators & RAW_DECORATOR_MASK> Raw;
914 Raw::clone(src, dst, size);
915 }
916
917 template <DecoratorSet decorators>
918 inline static typename EnableIf<
919 !HasDecorator<decorators, AS_RAW>::value>::type
920 clone(oop src, oop dst, size_t size) {
921 RuntimeDispatch<decorators, oop, BARRIER_CLONE>::clone(src, dst, size);
922 }
923
924 template <DecoratorSet decorators>
925 inline static typename EnableIf<
926 HasDecorator<decorators, AS_RAW>::value>::type
927 value_copy(void* src, void* dst, InlineKlass* md, LayoutKind lk) {
928 typedef RawAccessBarrier<decorators & RAW_DECORATOR_MASK> Raw;
929 Raw::value_copy(src, dst, md, lk);
930 }
931
932 template <DecoratorSet decorators>
933 inline static typename EnableIf<
934 !HasDecorator<decorators, AS_RAW>::value>::type
935 value_copy(void* src, void* dst, InlineKlass* md, LayoutKind lk) {
936 const DecoratorSet expanded_decorators = decorators;
937 RuntimeDispatch<expanded_decorators, void*, BARRIER_VALUE_COPY>::value_copy(src, dst, md, lk);
938 }
939 };
940
941 // Step 2: Reduce types.
942 // Enforce that for non-oop types, T and P have to be strictly the same.
943 // P is the type of the address and T is the type of the values.
944 // As for oop types, it is allow to send T in {narrowOop, oop} and
945 // P in {narrowOop, oop, HeapWord*}. The following rules apply according to
946 // the subsequent table. (columns are P, rows are T)
947 // | | HeapWord | oop | narrowOop |
948 // | oop | rt-comp | hw-none | hw-comp |
949 // | narrowOop | x | x | hw-none |
950 //
951 // x means not allowed
952 // rt-comp means it must be checked at runtime whether the oop is compressed.
953 // hw-none means it is statically known the oop will not be compressed.
954 // hw-comp means it is statically known the oop will be compressed.
955
956 template <DecoratorSet decorators, typename T>
957 inline void store_reduce_types(T* addr, T value) {
958 PreRuntimeDispatch::store<decorators>(addr, value);
959 }
960
961 template <DecoratorSet decorators>
962 inline void store_reduce_types(narrowOop* addr, oop value) {
963 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP |
964 INTERNAL_RT_USE_COMPRESSED_OOPS;
965 PreRuntimeDispatch::store<expanded_decorators>(addr, value);
966 }
967
968 template <DecoratorSet decorators>
969 inline void store_reduce_types(narrowOop* addr, narrowOop value) {
970 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP |
971 INTERNAL_RT_USE_COMPRESSED_OOPS;
972 PreRuntimeDispatch::store<expanded_decorators>(addr, value);
973 }
974
975 template <DecoratorSet decorators>
976 inline void store_reduce_types(HeapWord* addr, oop value) {
977 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP;
978 PreRuntimeDispatch::store<expanded_decorators>(addr, value);
979 }
980
981 template <DecoratorSet decorators, typename T>
982 inline T atomic_cmpxchg_reduce_types(T* addr, T compare_value, T new_value) {
983 return PreRuntimeDispatch::atomic_cmpxchg<decorators>(addr, compare_value, new_value);
984 }
985
986 template <DecoratorSet decorators>
987 inline oop atomic_cmpxchg_reduce_types(narrowOop* addr, oop compare_value, oop new_value) {
988 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP |
989 INTERNAL_RT_USE_COMPRESSED_OOPS;
990 return PreRuntimeDispatch::atomic_cmpxchg<expanded_decorators>(addr, compare_value, new_value);
991 }
992
993 template <DecoratorSet decorators>
994 inline narrowOop atomic_cmpxchg_reduce_types(narrowOop* addr, narrowOop compare_value, narrowOop new_value) {
995 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP |
996 INTERNAL_RT_USE_COMPRESSED_OOPS;
997 return PreRuntimeDispatch::atomic_cmpxchg<expanded_decorators>(addr, compare_value, new_value);
998 }
999
1000 template <DecoratorSet decorators>
1001 inline oop atomic_cmpxchg_reduce_types(HeapWord* addr,
1002 oop compare_value,
1003 oop new_value) {
1004 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP;
1005 return PreRuntimeDispatch::atomic_cmpxchg<expanded_decorators>(addr, compare_value, new_value);
1006 }
1007
1008 template <DecoratorSet decorators, typename T>
1009 inline T atomic_xchg_reduce_types(T* addr, T new_value) {
1010 const DecoratorSet expanded_decorators = decorators;
1011 return PreRuntimeDispatch::atomic_xchg<expanded_decorators>(addr, new_value);
1012 }
1013
1014 template <DecoratorSet decorators>
1015 inline oop atomic_xchg_reduce_types(narrowOop* addr, oop new_value) {
1016 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP |
1017 INTERNAL_RT_USE_COMPRESSED_OOPS;
1018 return PreRuntimeDispatch::atomic_xchg<expanded_decorators>(addr, new_value);
1019 }
1020
1021 template <DecoratorSet decorators>
1022 inline narrowOop atomic_xchg_reduce_types(narrowOop* addr, narrowOop new_value) {
1023 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP |
1024 INTERNAL_RT_USE_COMPRESSED_OOPS;
1025 return PreRuntimeDispatch::atomic_xchg<expanded_decorators>(addr, new_value);
1026 }
1027
1028 template <DecoratorSet decorators>
1029 inline oop atomic_xchg_reduce_types(HeapWord* addr, oop new_value) {
1030 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP;
1031 return PreRuntimeDispatch::atomic_xchg<expanded_decorators>(addr, new_value);
1032 }
1033
1034 template <DecoratorSet decorators, typename T>
1035 inline T load_reduce_types(T* addr) {
1036 return PreRuntimeDispatch::load<decorators, T>(addr);
1037 }
1038
1039 template <DecoratorSet decorators, typename T>
1040 inline typename OopOrNarrowOop<T>::type load_reduce_types(narrowOop* addr) {
1041 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP |
1042 INTERNAL_RT_USE_COMPRESSED_OOPS;
1043 return PreRuntimeDispatch::load<expanded_decorators, typename OopOrNarrowOop<T>::type>(addr);
1044 }
1045
1046 template <DecoratorSet decorators, typename T>
1047 inline oop load_reduce_types(HeapWord* addr) {
1048 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP;
1049 return PreRuntimeDispatch::load<expanded_decorators, oop>(addr);
1050 }
1051
1052 template <DecoratorSet decorators, typename T>
1053 inline void arraycopy_reduce_types(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw,
1054 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
1055 size_t length) {
1056 PreRuntimeDispatch::arraycopy<decorators>(src_obj, src_offset_in_bytes, src_raw,
1057 dst_obj, dst_offset_in_bytes, dst_raw,
1058 length);
1059 }
1060
1061 template <DecoratorSet decorators>
1062 inline void arraycopy_reduce_types(arrayOop src_obj, size_t src_offset_in_bytes, HeapWord* src_raw,
1063 arrayOop dst_obj, size_t dst_offset_in_bytes, HeapWord* dst_raw,
1064 size_t length) {
1065 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP;
1066 PreRuntimeDispatch::arraycopy<expanded_decorators>(src_obj, src_offset_in_bytes, src_raw,
1067 dst_obj, dst_offset_in_bytes, dst_raw,
1068 length);
1069 }
1070
1071 template <DecoratorSet decorators>
1072 inline void arraycopy_reduce_types(arrayOop src_obj, size_t src_offset_in_bytes, narrowOop* src_raw,
1073 arrayOop dst_obj, size_t dst_offset_in_bytes, narrowOop* dst_raw,
1074 size_t length) {
1075 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP |
1076 INTERNAL_RT_USE_COMPRESSED_OOPS;
1077 PreRuntimeDispatch::arraycopy<expanded_decorators>(src_obj, src_offset_in_bytes, src_raw,
1078 dst_obj, dst_offset_in_bytes, dst_raw,
1079 length);
1080 }
1081
1082 // Step 1: Set default decorators. This step remembers if a type was volatile
1083 // and then sets the MO_RELAXED decorator by default. Otherwise, a default
1084 // memory ordering is set for the access, and the implied decorator rules
1085 // are applied to select sensible defaults for decorators that have not been
1086 // explicitly set. For example, default object referent strength is set to strong.
1087 // This step also decays the types passed in (e.g. getting rid of CV qualifiers
1088 // and references from the types). This step also perform some type verification
1089 // that the passed in types make sense.
1090
1091 template <DecoratorSet decorators, typename T>
1092 static void verify_types(){
1093 // If this fails to compile, then you have sent in something that is
1094 // not recognized as a valid primitive type to a primitive Access function.
1095 STATIC_ASSERT((HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value || // oops have already been validated
1096 (std::is_pointer<T>::value || std::is_integral<T>::value) ||
1097 std::is_floating_point<T>::value)); // not allowed primitive type
1098 }
1099
1100 template <DecoratorSet decorators, typename P, typename T>
1101 inline void store(P* addr, T value) {
1102 verify_types<decorators, T>();
1103 using DecayedP = std::decay_t<P>;
1104 using DecayedT = std::decay_t<T>;
1105 DecayedT decayed_value = value;
1106 // If a volatile address is passed in but no memory ordering decorator,
1107 // set the memory ordering to MO_RELAXED by default.
1108 const DecoratorSet expanded_decorators = DecoratorFixup<
1109 (std::is_volatile<P>::value && !HasDecorator<decorators, MO_DECORATOR_MASK>::value) ?
1110 (MO_RELAXED | decorators) : decorators>::value;
1111 store_reduce_types<expanded_decorators>(const_cast<DecayedP*>(addr), decayed_value);
1112 }
1113
1114 template <DecoratorSet decorators, typename T>
1115 inline void store_at(oop base, ptrdiff_t offset, T value) {
1116 verify_types<decorators, T>();
1117 using DecayedT = std::decay_t<T>;
1118 DecayedT decayed_value = value;
1119 const DecoratorSet expanded_decorators = DecoratorFixup<decorators |
1120 (HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value ?
1121 INTERNAL_CONVERT_COMPRESSED_OOP : DECORATORS_NONE)>::value;
1122 PreRuntimeDispatch::store_at<expanded_decorators>(base, offset, decayed_value);
1123 }
1124
1125 template <DecoratorSet decorators, typename P, typename T>
1126 inline T load(P* addr) {
1127 verify_types<decorators, T>();
1128 using DecayedP = std::decay_t<P>;
1129 using DecayedT = std::conditional_t<HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value,
1130 typename OopOrNarrowOop<T>::type,
1131 std::decay_t<T>>;
1132 // If a volatile address is passed in but no memory ordering decorator,
1133 // set the memory ordering to MO_RELAXED by default.
1134 const DecoratorSet expanded_decorators = DecoratorFixup<
1135 (std::is_volatile<P>::value && !HasDecorator<decorators, MO_DECORATOR_MASK>::value) ?
1136 (MO_RELAXED | decorators) : decorators>::value;
1137 return load_reduce_types<expanded_decorators, DecayedT>(const_cast<DecayedP*>(addr));
1138 }
1139
1140 template <DecoratorSet decorators, typename T>
1141 inline T load_at(oop base, ptrdiff_t offset) {
1142 verify_types<decorators, T>();
1143 using DecayedT = std::conditional_t<HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value,
1144 typename OopOrNarrowOop<T>::type,
1145 std::decay_t<T>>;
1146 // Expand the decorators (figure out sensible defaults)
1147 // Potentially remember if we need compressed oop awareness
1148 const DecoratorSet expanded_decorators = DecoratorFixup<decorators |
1149 (HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value ?
1150 INTERNAL_CONVERT_COMPRESSED_OOP : DECORATORS_NONE)>::value;
1151 return PreRuntimeDispatch::load_at<expanded_decorators, DecayedT>(base, offset);
1152 }
1153
1154 template <DecoratorSet decorators, typename P, typename T>
1155 inline T atomic_cmpxchg(P* addr, T compare_value, T new_value) {
1156 verify_types<decorators, T>();
1157 using DecayedP = std::decay_t<P>;
1158 using DecayedT = std::decay_t<T>;
1159 DecayedT new_decayed_value = new_value;
1160 DecayedT compare_decayed_value = compare_value;
1161 const DecoratorSet expanded_decorators = DecoratorFixup<
1162 (!HasDecorator<decorators, MO_DECORATOR_MASK>::value) ?
1163 (MO_SEQ_CST | decorators) : decorators>::value;
1164 return atomic_cmpxchg_reduce_types<expanded_decorators>(const_cast<DecayedP*>(addr),
1165 compare_decayed_value,
1166 new_decayed_value);
1167 }
1168
1169 template <DecoratorSet decorators, typename T>
1170 inline T atomic_cmpxchg_at(oop base, ptrdiff_t offset, T compare_value, T new_value) {
1171 verify_types<decorators, T>();
1172 using DecayedT = std::decay_t<T>;
1173 DecayedT new_decayed_value = new_value;
1174 DecayedT compare_decayed_value = compare_value;
1175 // Determine default memory ordering
1176 const DecoratorSet expanded_decorators = DecoratorFixup<
1177 (!HasDecorator<decorators, MO_DECORATOR_MASK>::value) ?
1178 (MO_SEQ_CST | decorators) : decorators>::value;
1179 // Potentially remember that we need compressed oop awareness
1180 const DecoratorSet final_decorators = expanded_decorators |
1181 (HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value ?
1182 INTERNAL_CONVERT_COMPRESSED_OOP : DECORATORS_NONE);
1183 return PreRuntimeDispatch::atomic_cmpxchg_at<final_decorators>(base, offset, compare_decayed_value,
1184 new_decayed_value);
1185 }
1186
1187 template <DecoratorSet decorators, typename P, typename T>
1188 inline T atomic_xchg(P* addr, T new_value) {
1189 verify_types<decorators, T>();
1190 using DecayedP = std::decay_t<P>;
1191 using DecayedT = std::decay_t<T>;
1192 DecayedT new_decayed_value = new_value;
1193 // atomic_xchg is only available in SEQ_CST flavour.
1194 const DecoratorSet expanded_decorators = DecoratorFixup<decorators | MO_SEQ_CST>::value;
1195 return atomic_xchg_reduce_types<expanded_decorators>(const_cast<DecayedP*>(addr),
1196 new_decayed_value);
1197 }
1198
1199 template <DecoratorSet decorators, typename T>
1200 inline T atomic_xchg_at(oop base, ptrdiff_t offset, T new_value) {
1201 verify_types<decorators, T>();
1202 using DecayedT = std::decay_t<T>;
1203 DecayedT new_decayed_value = new_value;
1204 // atomic_xchg is only available in SEQ_CST flavour.
1205 const DecoratorSet expanded_decorators = DecoratorFixup<decorators | MO_SEQ_CST |
1206 (HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value ?
1207 INTERNAL_CONVERT_COMPRESSED_OOP : DECORATORS_NONE)>::value;
1208 return PreRuntimeDispatch::atomic_xchg_at<expanded_decorators>(base, offset, new_decayed_value);
1209 }
1210
1211 template <DecoratorSet decorators, typename T>
1212 inline void arraycopy(arrayOop src_obj, size_t src_offset_in_bytes, const T* src_raw,
1213 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
1214 size_t length) {
1215 STATIC_ASSERT((HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value ||
1216 (std::is_same<T, void>::value || std::is_integral<T>::value) ||
1217 std::is_floating_point<T>::value)); // arraycopy allows type erased void elements
1218 using DecayedT = std::decay_t<T>;
1219 const DecoratorSet expanded_decorators = DecoratorFixup<decorators | IS_ARRAY | IN_HEAP>::value;
1220 arraycopy_reduce_types<expanded_decorators>(src_obj, src_offset_in_bytes, const_cast<DecayedT*>(src_raw),
1221 dst_obj, dst_offset_in_bytes, const_cast<DecayedT*>(dst_raw),
1222 length);
1223 }
1224
1225 template <DecoratorSet decorators>
1226 inline void clone(oop src, oop dst, size_t size) {
1227 const DecoratorSet expanded_decorators = DecoratorFixup<decorators>::value;
1228 PreRuntimeDispatch::clone<expanded_decorators>(src, dst, size);
1229 }
1230
1231 template <DecoratorSet decorators>
1232 inline void value_copy(void* src, void* dst, InlineKlass* md, LayoutKind lk) {
1233 const DecoratorSet expanded_decorators = DecoratorFixup<decorators>::value;
1234 PreRuntimeDispatch::value_copy<expanded_decorators>(src, dst, md, lk);
1235 }
1236
1237 // Infer the type that should be returned from an Access::oop_load.
1238 template <typename P, DecoratorSet decorators>
1239 class OopLoadProxy: public StackObj {
1240 private:
1241 P *const _addr;
1242 public:
1243 explicit OopLoadProxy(P* addr) : _addr(addr) {}
1244
1245 inline operator oop() {
1246 return load<decorators | INTERNAL_VALUE_IS_OOP, P, oop>(_addr);
1247 }
1248
1249 inline operator narrowOop() {
1250 return load<decorators | INTERNAL_VALUE_IS_OOP, P, narrowOop>(_addr);
1251 }
1252
1253 template <typename T>
1254 inline bool operator ==(const T& other) const {
1255 return load<decorators | INTERNAL_VALUE_IS_OOP, P, T>(_addr) == other;
1256 }
1257
1258 template <typename T>
1259 inline bool operator !=(const T& other) const {
1260 return load<decorators | INTERNAL_VALUE_IS_OOP, P, T>(_addr) != other;
1261 }
1262
1263 inline bool operator ==(std::nullptr_t) const {
1264 return load<decorators | INTERNAL_VALUE_IS_OOP, P, oop>(_addr) == nullptr;
1265 }
1266
1267 inline bool operator !=(std::nullptr_t) const {
1268 return load<decorators | INTERNAL_VALUE_IS_OOP, P, oop>(_addr) != nullptr;
1269 }
1270 };
1271
1272 // Infer the type that should be returned from an Access::load_at.
1273 template <DecoratorSet decorators>
1274 class LoadAtProxy: public StackObj {
1275 private:
1276 const oop _base;
1277 const ptrdiff_t _offset;
1278 public:
1279 LoadAtProxy(oop base, ptrdiff_t offset) : _base(base), _offset(offset) {}
1280
1281 template <typename T>
1282 inline operator T() const {
1283 return load_at<decorators, T>(_base, _offset);
1284 }
1285
1286 template <typename T>
1287 inline bool operator ==(const T& other) const { return load_at<decorators, T>(_base, _offset) == other; }
1288
1289 template <typename T>
1290 inline bool operator !=(const T& other) const { return load_at<decorators, T>(_base, _offset) != other; }
1291 };
1292
1293 // Infer the type that should be returned from an Access::oop_load_at.
1294 template <DecoratorSet decorators>
1295 class OopLoadAtProxy: public StackObj {
1296 private:
1297 const oop _base;
1298 const ptrdiff_t _offset;
1299 public:
1300 OopLoadAtProxy(oop base, ptrdiff_t offset) : _base(base), _offset(offset) {}
1301
1302 inline operator oop() const {
1303 return load_at<decorators | INTERNAL_VALUE_IS_OOP, oop>(_base, _offset);
1304 }
1305
1306 inline operator narrowOop() const {
1307 return load_at<decorators | INTERNAL_VALUE_IS_OOP, narrowOop>(_base, _offset);
1308 }
1309
1310 template <typename T>
1311 inline bool operator ==(const T& other) const {
1312 return load_at<decorators | INTERNAL_VALUE_IS_OOP, T>(_base, _offset) == other;
1313 }
1314
1315 template <typename T>
1316 inline bool operator !=(const T& other) const {
1317 return load_at<decorators | INTERNAL_VALUE_IS_OOP, T>(_base, _offset) != other;
1318 }
1319 };
1320 }
1321
1322 #endif // SHARE_OOPS_ACCESSBACKEND_HPP