1 /*
2 * Copyright (c) 2017, 2022, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
24
25 #ifndef SHARE_OOPS_ACCESSBACKEND_HPP
26 #define SHARE_OOPS_ACCESSBACKEND_HPP
27
28 #include "gc/shared/barrierSetConfig.hpp"
29 #include "memory/allocation.hpp"
30 #include "metaprogramming/enableIf.hpp"
31 #include "oops/accessDecorators.hpp"
32 #include "oops/oopsHierarchy.hpp"
33 #include "runtime/globals.hpp"
34 #include "utilities/debug.hpp"
35 #include "utilities/globalDefinitions.hpp"
36
37 #include <type_traits>
38
39 // This metafunction returns either oop or narrowOop depending on whether
40 // an access needs to use compressed oops or not.
41 template <DecoratorSet decorators>
42 struct HeapOopType: AllStatic {
43 static const bool needs_oop_compress = HasDecorator<decorators, INTERNAL_CONVERT_COMPRESSED_OOP>::value &&
44 HasDecorator<decorators, INTERNAL_RT_USE_COMPRESSED_OOPS>::value;
45 using type = std::conditional_t<needs_oop_compress, narrowOop, oop>;
46 };
47
48 namespace AccessInternal {
49 enum BarrierType {
50 BARRIER_STORE,
51 BARRIER_STORE_AT,
52 BARRIER_LOAD,
53 BARRIER_LOAD_AT,
54 BARRIER_ATOMIC_CMPXCHG,
55 BARRIER_ATOMIC_CMPXCHG_AT,
56 BARRIER_ATOMIC_XCHG,
57 BARRIER_ATOMIC_XCHG_AT,
58 BARRIER_ARRAYCOPY,
59 BARRIER_CLONE
60 };
61
62 template <DecoratorSet decorators, typename T>
63 struct MustConvertCompressedOop: public std::integral_constant<bool,
64 HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value &&
65 std::is_same<typename HeapOopType<decorators>::type, narrowOop>::value &&
66 std::is_same<T, oop>::value> {};
67
68 // This metafunction returns an appropriate oop type if the value is oop-like
69 // and otherwise returns the same type T.
70 template <DecoratorSet decorators, typename T>
71 struct EncodedType: AllStatic {
72 using type = std::conditional_t<HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value,
73 typename HeapOopType<decorators>::type,
74 T>;
75 };
76
77 template <DecoratorSet decorators>
78 inline typename HeapOopType<decorators>::type*
79 oop_field_addr(oop base, ptrdiff_t byte_offset) {
80 return reinterpret_cast<typename HeapOopType<decorators>::type*>(
81 reinterpret_cast<intptr_t>((void*)base) + byte_offset);
82 }
83
84 template <DecoratorSet decorators, typename T>
85 struct AccessFunctionTypes {
86 typedef T (*load_at_func_t)(oop base, ptrdiff_t offset);
87 typedef void (*store_at_func_t)(oop base, ptrdiff_t offset, T value);
88 typedef T (*atomic_cmpxchg_at_func_t)(oop base, ptrdiff_t offset, T compare_value, T new_value);
89 typedef T (*atomic_xchg_at_func_t)(oop base, ptrdiff_t offset, T new_value);
90
91 typedef T (*load_func_t)(void* addr);
92 typedef void (*store_func_t)(void* addr, T value);
93 typedef T (*atomic_cmpxchg_func_t)(void* addr, T compare_value, T new_value);
94 typedef T (*atomic_xchg_func_t)(void* addr, T new_value);
95
96 typedef bool (*arraycopy_func_t)(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw,
97 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
98 size_t length);
99 typedef void (*clone_func_t)(oop src, oop dst, size_t size);
100 };
101
102 template <DecoratorSet decorators>
103 struct AccessFunctionTypes<decorators, void> {
104 typedef bool (*arraycopy_func_t)(arrayOop src_obj, size_t src_offset_in_bytes, void* src,
105 arrayOop dst_obj, size_t dst_offset_in_bytes, void* dst,
106 size_t length);
107 };
108
109 template <DecoratorSet decorators, typename T, BarrierType barrier> struct AccessFunction {};
110
111 #define ACCESS_GENERATE_ACCESS_FUNCTION(bt, func) \
112 template <DecoratorSet decorators, typename T> \
113 struct AccessFunction<decorators, T, bt>: AllStatic{ \
114 typedef typename AccessFunctionTypes<decorators, T>::func type; \
115 }
116 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_STORE, store_func_t);
117 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_STORE_AT, store_at_func_t);
118 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_LOAD, load_func_t);
119 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_LOAD_AT, load_at_func_t);
120 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_ATOMIC_CMPXCHG, atomic_cmpxchg_func_t);
121 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_ATOMIC_CMPXCHG_AT, atomic_cmpxchg_at_func_t);
122 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_ATOMIC_XCHG, atomic_xchg_func_t);
123 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_ATOMIC_XCHG_AT, atomic_xchg_at_func_t);
124 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_ARRAYCOPY, arraycopy_func_t);
125 ACCESS_GENERATE_ACCESS_FUNCTION(BARRIER_CLONE, clone_func_t);
126 #undef ACCESS_GENERATE_ACCESS_FUNCTION
127
128 template <DecoratorSet decorators, typename T, BarrierType barrier_type>
129 typename AccessFunction<decorators, T, barrier_type>::type resolve_barrier();
130
131 template <DecoratorSet decorators, typename T, BarrierType barrier_type>
132 typename AccessFunction<decorators, T, barrier_type>::type resolve_oop_barrier();
133
134 void* field_addr(oop base, ptrdiff_t offset);
135
136 // Forward calls to Copy:: in the cpp file to reduce dependencies and allow
137 // faster build times, given how frequently included access is.
138 void arraycopy_arrayof_conjoint_oops(void* src, void* dst, size_t length);
139 void arraycopy_conjoint_oops(oop* src, oop* dst, size_t length);
140 void arraycopy_conjoint_oops(narrowOop* src, narrowOop* dst, size_t length);
141
142 void arraycopy_disjoint_words(void* src, void* dst, size_t length);
143 void arraycopy_disjoint_words_atomic(void* src, void* dst, size_t length);
144
145 template<typename T>
146 void arraycopy_conjoint(T* src, T* dst, size_t length);
147 template<typename T>
148 void arraycopy_arrayof_conjoint(T* src, T* dst, size_t length);
149 template<typename T>
150 void arraycopy_conjoint_atomic(T* src, T* dst, size_t length);
151 }
152
153 // This mask specifies what decorators are relevant for raw accesses. When passing
154 // accesses to the raw layer, irrelevant decorators are removed.
155 const DecoratorSet RAW_DECORATOR_MASK = INTERNAL_DECORATOR_MASK | MO_DECORATOR_MASK |
156 ARRAYCOPY_DECORATOR_MASK | IS_NOT_NULL;
157
158 // The RawAccessBarrier performs raw accesses with additional knowledge of
159 // memory ordering, so that OrderAccess/Atomic is called when necessary.
160 // It additionally handles compressed oops, and hence is not completely "raw"
161 // strictly speaking.
162 template <DecoratorSet decorators>
163 class RawAccessBarrier: public AllStatic {
164 protected:
165 static inline void* field_addr(oop base, ptrdiff_t byte_offset) {
166 return AccessInternal::field_addr(base, byte_offset);
167 }
168
169 protected:
170 // Only encode if INTERNAL_VALUE_IS_OOP
171 template <DecoratorSet idecorators, typename T>
172 static inline typename EnableIf<
173 AccessInternal::MustConvertCompressedOop<idecorators, T>::value,
174 typename HeapOopType<idecorators>::type>::type
175 encode_internal(T value);
176
177 template <DecoratorSet idecorators, typename T>
178 static inline typename EnableIf<
179 !AccessInternal::MustConvertCompressedOop<idecorators, T>::value, T>::type
180 encode_internal(T value) {
181 return value;
182 }
183
184 template <typename T>
185 static inline typename AccessInternal::EncodedType<decorators, T>::type
186 encode(T value) {
187 return encode_internal<decorators, T>(value);
188 }
189
190 // Only decode if INTERNAL_VALUE_IS_OOP
191 template <DecoratorSet idecorators, typename T>
192 static inline typename EnableIf<
193 AccessInternal::MustConvertCompressedOop<idecorators, T>::value, T>::type
194 decode_internal(typename HeapOopType<idecorators>::type value);
195
196 template <DecoratorSet idecorators, typename T>
197 static inline typename EnableIf<
198 !AccessInternal::MustConvertCompressedOop<idecorators, T>::value, T>::type
199 decode_internal(T value) {
200 return value;
201 }
202
203 template <typename T>
204 static inline T decode(typename AccessInternal::EncodedType<decorators, T>::type value) {
205 return decode_internal<decorators, T>(value);
206 }
207
208 protected:
209 template <DecoratorSet ds, typename T>
210 static typename EnableIf<
211 HasDecorator<ds, MO_SEQ_CST>::value, T>::type
212 load_internal(void* addr);
213
214 template <DecoratorSet ds, typename T>
215 static typename EnableIf<
216 HasDecorator<ds, MO_ACQUIRE>::value, T>::type
217 load_internal(void* addr);
218
219 template <DecoratorSet ds, typename T>
220 static typename EnableIf<
221 HasDecorator<ds, MO_RELAXED>::value, T>::type
222 load_internal(void* addr);
223
224 template <DecoratorSet ds, typename T>
225 static inline typename EnableIf<
226 HasDecorator<ds, MO_UNORDERED>::value, T>::type
227 load_internal(void* addr) {
228 return *reinterpret_cast<T*>(addr);
229 }
230
231 template <DecoratorSet ds, typename T>
232 static typename EnableIf<
233 HasDecorator<ds, MO_SEQ_CST>::value>::type
234 store_internal(void* addr, T value);
235
236 template <DecoratorSet ds, typename T>
237 static typename EnableIf<
238 HasDecorator<ds, MO_RELEASE>::value>::type
239 store_internal(void* addr, T value);
240
241 template <DecoratorSet ds, typename T>
242 static typename EnableIf<
243 HasDecorator<ds, MO_RELAXED>::value>::type
244 store_internal(void* addr, T value);
245
246 template <DecoratorSet ds, typename T>
247 static inline typename EnableIf<
248 HasDecorator<ds, MO_UNORDERED>::value>::type
249 store_internal(void* addr, T value) {
250 *reinterpret_cast<T*>(addr) = value;
251 }
252
253 template <DecoratorSet ds, typename T>
254 static typename EnableIf<
255 HasDecorator<ds, MO_SEQ_CST>::value, T>::type
256 atomic_cmpxchg_internal(void* addr, T compare_value, T new_value);
257
258 template <DecoratorSet ds, typename T>
259 static typename EnableIf<
260 HasDecorator<ds, MO_RELAXED>::value, T>::type
261 atomic_cmpxchg_internal(void* addr, T compare_value, T new_value);
262
263 template <DecoratorSet ds, typename T>
264 static typename EnableIf<
265 HasDecorator<ds, MO_SEQ_CST>::value, T>::type
266 atomic_xchg_internal(void* addr, T new_value);
267
268 public:
269 template <typename T>
270 static inline void store(void* addr, T value) {
271 store_internal<decorators>(addr, value);
272 }
273
274 template <typename T>
275 static inline T load(void* addr) {
276 return load_internal<decorators, T>(addr);
277 }
278
279 template <typename T>
280 static inline T atomic_cmpxchg(void* addr, T compare_value, T new_value) {
281 return atomic_cmpxchg_internal<decorators>(addr, compare_value, new_value);
282 }
283
284 template <typename T>
285 static inline T atomic_xchg(void* addr, T new_value) {
286 return atomic_xchg_internal<decorators>(addr, new_value);
287 }
288
289 template <typename T>
290 static bool arraycopy(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw,
291 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
292 size_t length);
293
294 template <typename T>
295 static void oop_store(void* addr, T value);
296 template <typename T>
297 static void oop_store_at(oop base, ptrdiff_t offset, T value);
298
299 template <typename T>
300 static T oop_load(void* addr);
301 template <typename T>
302 static T oop_load_at(oop base, ptrdiff_t offset);
303
304 template <typename T>
305 static T oop_atomic_cmpxchg(void* addr, T compare_value, T new_value);
306 template <typename T>
307 static T oop_atomic_cmpxchg_at(oop base, ptrdiff_t offset, T compare_value, T new_value);
308
309 template <typename T>
310 static T oop_atomic_xchg(void* addr, T new_value);
311 template <typename T>
312 static T oop_atomic_xchg_at(oop base, ptrdiff_t offset, T new_value);
313
314 template <typename T>
315 static void store_at(oop base, ptrdiff_t offset, T value) {
316 store(field_addr(base, offset), value);
317 }
318
319 template <typename T>
320 static T load_at(oop base, ptrdiff_t offset) {
321 return load<T>(field_addr(base, offset));
322 }
323
324 template <typename T>
325 static T atomic_cmpxchg_at(oop base, ptrdiff_t offset, T compare_value, T new_value) {
326 return atomic_cmpxchg(field_addr(base, offset), compare_value, new_value);
327 }
328
329 template <typename T>
330 static T atomic_xchg_at(oop base, ptrdiff_t offset, T new_value) {
331 return atomic_xchg(field_addr(base, offset), new_value);
332 }
333
334 template <typename T>
335 static bool oop_arraycopy(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw,
336 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
337 size_t length);
338
339 static void clone(oop src, oop dst, size_t size);
340 };
341
342 namespace AccessInternal {
343 DEBUG_ONLY(void check_access_thread_state());
344 #define assert_access_thread_state() DEBUG_ONLY(check_access_thread_state())
345 }
346
347 // Below is the implementation of the first 4 steps of the template pipeline:
348 // * Step 1: Set default decorators and decay types. This step gets rid of CV qualifiers
349 // and sets default decorators to sensible values.
350 // * Step 2: Reduce types. This step makes sure there is only a single T type and not
351 // multiple types. The P type of the address and T type of the value must
352 // match.
353 // * Step 3: Pre-runtime dispatch. This step checks whether a runtime call can be
354 // avoided, and in that case avoids it (calling raw accesses or
355 // primitive accesses in a build that does not require primitive GC barriers)
356 // * Step 4: Runtime-dispatch. This step performs a runtime dispatch to the corresponding
357 // BarrierSet::AccessBarrier accessor that attaches GC-required barriers
358 // to the access.
359
360 namespace AccessInternal {
361 template <typename T>
362 struct OopOrNarrowOopInternal: AllStatic {
363 typedef oop type;
364 };
365
366 template <>
367 struct OopOrNarrowOopInternal<narrowOop>: AllStatic {
368 typedef narrowOop type;
369 };
370
371 // This metafunction returns a canonicalized oop/narrowOop type for a passed
372 // in oop-like types passed in from oop_* overloads where the user has sworn
373 // that the passed in values should be oop-like (e.g. oop, oopDesc*, arrayOop,
374 // narrowOoop, instanceOopDesc*, and random other things).
375 // In the oop_* overloads, it must hold that if the passed in type T is not
376 // narrowOop, then it by contract has to be one of many oop-like types implicitly
377 // convertible to oop, and hence returns oop as the canonical oop type.
378 // If it turns out it was not, then the implicit conversion to oop will fail
379 // to compile, as desired.
380 template <typename T>
381 struct OopOrNarrowOop: AllStatic {
382 typedef typename OopOrNarrowOopInternal<std::decay_t<T>>::type type;
383 };
384
385 inline void* field_addr(oop base, ptrdiff_t byte_offset) {
386 return reinterpret_cast<void*>(reinterpret_cast<intptr_t>((void*)base) + byte_offset);
387 }
388 // Step 4: Runtime dispatch
389 // The RuntimeDispatch class is responsible for performing a runtime dispatch of the
390 // accessor. This is required when the access either depends on whether compressed oops
391 // is being used, or it depends on which GC implementation was chosen (e.g. requires GC
392 // barriers). The way it works is that a function pointer initially pointing to an
393 // accessor resolution function gets called for each access. Upon first invocation,
394 // it resolves which accessor to be used in future invocations and patches the
395 // function pointer to this new accessor.
396
397 template <DecoratorSet decorators, typename T, BarrierType type>
398 struct RuntimeDispatch: AllStatic {};
399
400 template <DecoratorSet decorators, typename T>
401 struct RuntimeDispatch<decorators, T, BARRIER_STORE>: AllStatic {
402 typedef typename AccessFunction<decorators, T, BARRIER_STORE>::type func_t;
403 static func_t _store_func;
404
405 static void store_init(void* addr, T value);
406
407 static inline void store(void* addr, T value) {
408 assert_access_thread_state();
409 _store_func(addr, value);
410 }
411 };
412
413 template <DecoratorSet decorators, typename T>
414 struct RuntimeDispatch<decorators, T, BARRIER_STORE_AT>: AllStatic {
415 typedef typename AccessFunction<decorators, T, BARRIER_STORE_AT>::type func_t;
416 static func_t _store_at_func;
417
418 static void store_at_init(oop base, ptrdiff_t offset, T value);
419
420 static inline void store_at(oop base, ptrdiff_t offset, T value) {
421 assert_access_thread_state();
422 _store_at_func(base, offset, value);
423 }
424 };
425
426 template <DecoratorSet decorators, typename T>
427 struct RuntimeDispatch<decorators, T, BARRIER_LOAD>: AllStatic {
428 typedef typename AccessFunction<decorators, T, BARRIER_LOAD>::type func_t;
429 static func_t _load_func;
430
431 static T load_init(void* addr);
432
433 static inline T load(void* addr) {
434 assert_access_thread_state();
435 return _load_func(addr);
436 }
437 };
438
439 template <DecoratorSet decorators, typename T>
440 struct RuntimeDispatch<decorators, T, BARRIER_LOAD_AT>: AllStatic {
441 typedef typename AccessFunction<decorators, T, BARRIER_LOAD_AT>::type func_t;
442 static func_t _load_at_func;
443
444 static T load_at_init(oop base, ptrdiff_t offset);
445
446 static inline T load_at(oop base, ptrdiff_t offset) {
447 assert_access_thread_state();
448 return _load_at_func(base, offset);
449 }
450 };
451
452 template <DecoratorSet decorators, typename T>
453 struct RuntimeDispatch<decorators, T, BARRIER_ATOMIC_CMPXCHG>: AllStatic {
454 typedef typename AccessFunction<decorators, T, BARRIER_ATOMIC_CMPXCHG>::type func_t;
455 static func_t _atomic_cmpxchg_func;
456
457 static T atomic_cmpxchg_init(void* addr, T compare_value, T new_value);
458
459 static inline T atomic_cmpxchg(void* addr, T compare_value, T new_value) {
460 assert_access_thread_state();
461 return _atomic_cmpxchg_func(addr, compare_value, new_value);
462 }
463 };
464
465 template <DecoratorSet decorators, typename T>
466 struct RuntimeDispatch<decorators, T, BARRIER_ATOMIC_CMPXCHG_AT>: AllStatic {
467 typedef typename AccessFunction<decorators, T, BARRIER_ATOMIC_CMPXCHG_AT>::type func_t;
468 static func_t _atomic_cmpxchg_at_func;
469
470 static T atomic_cmpxchg_at_init(oop base, ptrdiff_t offset, T compare_value, T new_value);
471
472 static inline T atomic_cmpxchg_at(oop base, ptrdiff_t offset, T compare_value, T new_value) {
473 assert_access_thread_state();
474 return _atomic_cmpxchg_at_func(base, offset, compare_value, new_value);
475 }
476 };
477
478 template <DecoratorSet decorators, typename T>
479 struct RuntimeDispatch<decorators, T, BARRIER_ATOMIC_XCHG>: AllStatic {
480 typedef typename AccessFunction<decorators, T, BARRIER_ATOMIC_XCHG>::type func_t;
481 static func_t _atomic_xchg_func;
482
483 static T atomic_xchg_init(void* addr, T new_value);
484
485 static inline T atomic_xchg(void* addr, T new_value) {
486 assert_access_thread_state();
487 return _atomic_xchg_func(addr, new_value);
488 }
489 };
490
491 template <DecoratorSet decorators, typename T>
492 struct RuntimeDispatch<decorators, T, BARRIER_ATOMIC_XCHG_AT>: AllStatic {
493 typedef typename AccessFunction<decorators, T, BARRIER_ATOMIC_XCHG_AT>::type func_t;
494 static func_t _atomic_xchg_at_func;
495
496 static T atomic_xchg_at_init(oop base, ptrdiff_t offset, T new_value);
497
498 static inline T atomic_xchg_at(oop base, ptrdiff_t offset, T new_value) {
499 assert_access_thread_state();
500 return _atomic_xchg_at_func(base, offset, new_value);
501 }
502 };
503
504 template <DecoratorSet decorators, typename T>
505 struct RuntimeDispatch<decorators, T, BARRIER_ARRAYCOPY>: AllStatic {
506 typedef typename AccessFunction<decorators, T, BARRIER_ARRAYCOPY>::type func_t;
507 static func_t _arraycopy_func;
508
509 static bool arraycopy_init(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw,
510 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
511 size_t length);
512
513 static inline bool arraycopy(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw,
514 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
515 size_t length) {
516 assert_access_thread_state();
517 return _arraycopy_func(src_obj, src_offset_in_bytes, src_raw,
518 dst_obj, dst_offset_in_bytes, dst_raw,
519 length);
520 }
521 };
522
523 template <DecoratorSet decorators, typename T>
524 struct RuntimeDispatch<decorators, T, BARRIER_CLONE>: AllStatic {
525 typedef typename AccessFunction<decorators, T, BARRIER_CLONE>::type func_t;
526 static func_t _clone_func;
527
528 static void clone_init(oop src, oop dst, size_t size);
529
530 static inline void clone(oop src, oop dst, size_t size) {
531 assert_access_thread_state();
532 _clone_func(src, dst, size);
533 }
534 };
535
536 // Initialize the function pointers to point to the resolving function.
537 template <DecoratorSet decorators, typename T>
538 typename AccessFunction<decorators, T, BARRIER_STORE>::type
539 RuntimeDispatch<decorators, T, BARRIER_STORE>::_store_func = &store_init;
540
541 template <DecoratorSet decorators, typename T>
542 typename AccessFunction<decorators, T, BARRIER_STORE_AT>::type
543 RuntimeDispatch<decorators, T, BARRIER_STORE_AT>::_store_at_func = &store_at_init;
544
545 template <DecoratorSet decorators, typename T>
546 typename AccessFunction<decorators, T, BARRIER_LOAD>::type
547 RuntimeDispatch<decorators, T, BARRIER_LOAD>::_load_func = &load_init;
548
549 template <DecoratorSet decorators, typename T>
550 typename AccessFunction<decorators, T, BARRIER_LOAD_AT>::type
551 RuntimeDispatch<decorators, T, BARRIER_LOAD_AT>::_load_at_func = &load_at_init;
552
553 template <DecoratorSet decorators, typename T>
554 typename AccessFunction<decorators, T, BARRIER_ATOMIC_CMPXCHG>::type
555 RuntimeDispatch<decorators, T, BARRIER_ATOMIC_CMPXCHG>::_atomic_cmpxchg_func = &atomic_cmpxchg_init;
556
557 template <DecoratorSet decorators, typename T>
558 typename AccessFunction<decorators, T, BARRIER_ATOMIC_CMPXCHG_AT>::type
559 RuntimeDispatch<decorators, T, BARRIER_ATOMIC_CMPXCHG_AT>::_atomic_cmpxchg_at_func = &atomic_cmpxchg_at_init;
560
561 template <DecoratorSet decorators, typename T>
562 typename AccessFunction<decorators, T, BARRIER_ATOMIC_XCHG>::type
563 RuntimeDispatch<decorators, T, BARRIER_ATOMIC_XCHG>::_atomic_xchg_func = &atomic_xchg_init;
564
565 template <DecoratorSet decorators, typename T>
566 typename AccessFunction<decorators, T, BARRIER_ATOMIC_XCHG_AT>::type
567 RuntimeDispatch<decorators, T, BARRIER_ATOMIC_XCHG_AT>::_atomic_xchg_at_func = &atomic_xchg_at_init;
568
569 template <DecoratorSet decorators, typename T>
570 typename AccessFunction<decorators, T, BARRIER_ARRAYCOPY>::type
571 RuntimeDispatch<decorators, T, BARRIER_ARRAYCOPY>::_arraycopy_func = &arraycopy_init;
572
573 template <DecoratorSet decorators, typename T>
574 typename AccessFunction<decorators, T, BARRIER_CLONE>::type
575 RuntimeDispatch<decorators, T, BARRIER_CLONE>::_clone_func = &clone_init;
576
577 // Step 3: Pre-runtime dispatching.
578 // The PreRuntimeDispatch class is responsible for filtering the barrier strength
579 // decorators. That is, for AS_RAW, it hardwires the accesses without a runtime
580 // dispatch point. Otherwise it goes through a runtime check if hardwiring was
581 // not possible.
582 struct PreRuntimeDispatch: AllStatic {
583 template<DecoratorSet decorators>
584 struct CanHardwireRaw: public std::integral_constant<
585 bool,
586 !HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value || // primitive access
587 !HasDecorator<decorators, INTERNAL_CONVERT_COMPRESSED_OOP>::value || // don't care about compressed oops (oop* address)
588 HasDecorator<decorators, INTERNAL_RT_USE_COMPRESSED_OOPS>::value> // we can infer we use compressed oops (narrowOop* address)
589 {};
590
591 static const DecoratorSet convert_compressed_oops = INTERNAL_RT_USE_COMPRESSED_OOPS | INTERNAL_CONVERT_COMPRESSED_OOP;
592
593 template<DecoratorSet decorators>
594 static bool is_hardwired_primitive() {
595 return !HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value;
596 }
597
598 template <DecoratorSet decorators, typename T>
599 inline static typename EnableIf<
600 HasDecorator<decorators, AS_RAW>::value && CanHardwireRaw<decorators>::value>::type
601 store(void* addr, T value) {
602 typedef RawAccessBarrier<decorators & RAW_DECORATOR_MASK> Raw;
603 if (HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value) {
604 Raw::oop_store(addr, value);
605 } else {
606 Raw::store(addr, value);
607 }
608 }
609
610 template <DecoratorSet decorators, typename T>
611 inline static typename EnableIf<
612 HasDecorator<decorators, AS_RAW>::value && !CanHardwireRaw<decorators>::value>::type
613 store(void* addr, T value) {
614 if (UseCompressedOops) {
615 const DecoratorSet expanded_decorators = decorators | convert_compressed_oops;
616 PreRuntimeDispatch::store<expanded_decorators>(addr, value);
617 } else {
618 const DecoratorSet expanded_decorators = decorators & ~convert_compressed_oops;
619 PreRuntimeDispatch::store<expanded_decorators>(addr, value);
620 }
621 }
622
623 template <DecoratorSet decorators, typename T>
624 inline static typename EnableIf<
625 !HasDecorator<decorators, AS_RAW>::value>::type
626 store(void* addr, T value) {
627 if (is_hardwired_primitive<decorators>()) {
628 const DecoratorSet expanded_decorators = decorators | AS_RAW;
629 PreRuntimeDispatch::store<expanded_decorators>(addr, value);
630 } else {
631 RuntimeDispatch<decorators, T, BARRIER_STORE>::store(addr, value);
632 }
633 }
634
635 template <DecoratorSet decorators, typename T>
636 inline static typename EnableIf<
637 HasDecorator<decorators, AS_RAW>::value>::type
638 store_at(oop base, ptrdiff_t offset, T value) {
639 store<decorators>(field_addr(base, offset), value);
640 }
641
642 template <DecoratorSet decorators, typename T>
643 inline static typename EnableIf<
644 !HasDecorator<decorators, AS_RAW>::value>::type
645 store_at(oop base, ptrdiff_t offset, T value) {
646 if (is_hardwired_primitive<decorators>()) {
647 const DecoratorSet expanded_decorators = decorators | AS_RAW;
648 PreRuntimeDispatch::store_at<expanded_decorators>(base, offset, value);
649 } else {
650 RuntimeDispatch<decorators, T, BARRIER_STORE_AT>::store_at(base, offset, value);
651 }
652 }
653
654 template <DecoratorSet decorators, typename T>
655 inline static typename EnableIf<
656 HasDecorator<decorators, AS_RAW>::value && CanHardwireRaw<decorators>::value, T>::type
657 load(void* addr) {
658 typedef RawAccessBarrier<decorators & RAW_DECORATOR_MASK> Raw;
659 if (HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value) {
660 return Raw::template oop_load<T>(addr);
661 } else {
662 return Raw::template load<T>(addr);
663 }
664 }
665
666 template <DecoratorSet decorators, typename T>
667 inline static typename EnableIf<
668 HasDecorator<decorators, AS_RAW>::value && !CanHardwireRaw<decorators>::value, T>::type
669 load(void* addr) {
670 if (UseCompressedOops) {
671 const DecoratorSet expanded_decorators = decorators | convert_compressed_oops;
672 return PreRuntimeDispatch::load<expanded_decorators, T>(addr);
673 } else {
674 const DecoratorSet expanded_decorators = decorators & ~convert_compressed_oops;
675 return PreRuntimeDispatch::load<expanded_decorators, T>(addr);
676 }
677 }
678
679 template <DecoratorSet decorators, typename T>
680 inline static typename EnableIf<
681 !HasDecorator<decorators, AS_RAW>::value, T>::type
682 load(void* addr) {
683 if (is_hardwired_primitive<decorators>()) {
684 const DecoratorSet expanded_decorators = decorators | AS_RAW;
685 return PreRuntimeDispatch::load<expanded_decorators, T>(addr);
686 } else {
687 return RuntimeDispatch<decorators, T, BARRIER_LOAD>::load(addr);
688 }
689 }
690
691 template <DecoratorSet decorators, typename T>
692 inline static typename EnableIf<
693 HasDecorator<decorators, AS_RAW>::value, T>::type
694 load_at(oop base, ptrdiff_t offset) {
695 return load<decorators, T>(field_addr(base, offset));
696 }
697
698 template <DecoratorSet decorators, typename T>
699 inline static typename EnableIf<
700 !HasDecorator<decorators, AS_RAW>::value, T>::type
701 load_at(oop base, ptrdiff_t offset) {
702 if (is_hardwired_primitive<decorators>()) {
703 const DecoratorSet expanded_decorators = decorators | AS_RAW;
704 return PreRuntimeDispatch::load_at<expanded_decorators, T>(base, offset);
705 } else {
706 return RuntimeDispatch<decorators, T, BARRIER_LOAD_AT>::load_at(base, offset);
707 }
708 }
709
710 template <DecoratorSet decorators, typename T>
711 inline static typename EnableIf<
712 HasDecorator<decorators, AS_RAW>::value && CanHardwireRaw<decorators>::value, T>::type
713 atomic_cmpxchg(void* addr, T compare_value, T new_value) {
714 typedef RawAccessBarrier<decorators & RAW_DECORATOR_MASK> Raw;
715 if (HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value) {
716 return Raw::oop_atomic_cmpxchg(addr, compare_value, new_value);
717 } else {
718 return Raw::atomic_cmpxchg(addr, compare_value, new_value);
719 }
720 }
721
722 template <DecoratorSet decorators, typename T>
723 inline static typename EnableIf<
724 HasDecorator<decorators, AS_RAW>::value && !CanHardwireRaw<decorators>::value, T>::type
725 atomic_cmpxchg(void* addr, T compare_value, T new_value) {
726 if (UseCompressedOops) {
727 const DecoratorSet expanded_decorators = decorators | convert_compressed_oops;
728 return PreRuntimeDispatch::atomic_cmpxchg<expanded_decorators>(addr, compare_value, new_value);
729 } else {
730 const DecoratorSet expanded_decorators = decorators & ~convert_compressed_oops;
731 return PreRuntimeDispatch::atomic_cmpxchg<expanded_decorators>(addr, compare_value, new_value);
732 }
733 }
734
735 template <DecoratorSet decorators, typename T>
736 inline static typename EnableIf<
737 !HasDecorator<decorators, AS_RAW>::value, T>::type
738 atomic_cmpxchg(void* addr, T compare_value, T new_value) {
739 if (is_hardwired_primitive<decorators>()) {
740 const DecoratorSet expanded_decorators = decorators | AS_RAW;
741 return PreRuntimeDispatch::atomic_cmpxchg<expanded_decorators>(addr, compare_value, new_value);
742 } else {
743 return RuntimeDispatch<decorators, T, BARRIER_ATOMIC_CMPXCHG>::atomic_cmpxchg(addr, compare_value, new_value);
744 }
745 }
746
747 template <DecoratorSet decorators, typename T>
748 inline static typename EnableIf<
749 HasDecorator<decorators, AS_RAW>::value, T>::type
750 atomic_cmpxchg_at(oop base, ptrdiff_t offset, T compare_value, T new_value) {
751 return atomic_cmpxchg<decorators>(field_addr(base, offset), compare_value, new_value);
752 }
753
754 template <DecoratorSet decorators, typename T>
755 inline static typename EnableIf<
756 !HasDecorator<decorators, AS_RAW>::value, T>::type
757 atomic_cmpxchg_at(oop base, ptrdiff_t offset, T compare_value, T new_value) {
758 if (is_hardwired_primitive<decorators>()) {
759 const DecoratorSet expanded_decorators = decorators | AS_RAW;
760 return PreRuntimeDispatch::atomic_cmpxchg_at<expanded_decorators>(base, offset, compare_value, new_value);
761 } else {
762 return RuntimeDispatch<decorators, T, BARRIER_ATOMIC_CMPXCHG_AT>::atomic_cmpxchg_at(base, offset, compare_value, new_value);
763 }
764 }
765
766 template <DecoratorSet decorators, typename T>
767 inline static typename EnableIf<
768 HasDecorator<decorators, AS_RAW>::value && CanHardwireRaw<decorators>::value, T>::type
769 atomic_xchg(void* addr, T new_value) {
770 typedef RawAccessBarrier<decorators & RAW_DECORATOR_MASK> Raw;
771 if (HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value) {
772 return Raw::oop_atomic_xchg(addr, new_value);
773 } else {
774 return Raw::atomic_xchg(addr, new_value);
775 }
776 }
777
778 template <DecoratorSet decorators, typename T>
779 inline static typename EnableIf<
780 HasDecorator<decorators, AS_RAW>::value && !CanHardwireRaw<decorators>::value, T>::type
781 atomic_xchg(void* addr, T new_value) {
782 if (UseCompressedOops) {
783 const DecoratorSet expanded_decorators = decorators | convert_compressed_oops;
784 return PreRuntimeDispatch::atomic_xchg<expanded_decorators>(addr, new_value);
785 } else {
786 const DecoratorSet expanded_decorators = decorators & ~convert_compressed_oops;
787 return PreRuntimeDispatch::atomic_xchg<expanded_decorators>(addr, new_value);
788 }
789 }
790
791 template <DecoratorSet decorators, typename T>
792 inline static typename EnableIf<
793 !HasDecorator<decorators, AS_RAW>::value, T>::type
794 atomic_xchg(void* addr, T new_value) {
795 if (is_hardwired_primitive<decorators>()) {
796 const DecoratorSet expanded_decorators = decorators | AS_RAW;
797 return PreRuntimeDispatch::atomic_xchg<expanded_decorators>(addr, new_value);
798 } else {
799 return RuntimeDispatch<decorators, T, BARRIER_ATOMIC_XCHG>::atomic_xchg(addr, new_value);
800 }
801 }
802
803 template <DecoratorSet decorators, typename T>
804 inline static typename EnableIf<
805 HasDecorator<decorators, AS_RAW>::value, T>::type
806 atomic_xchg_at(oop base, ptrdiff_t offset, T new_value) {
807 return atomic_xchg<decorators>(field_addr(base, offset), new_value);
808 }
809
810 template <DecoratorSet decorators, typename T>
811 inline static typename EnableIf<
812 !HasDecorator<decorators, AS_RAW>::value, T>::type
813 atomic_xchg_at(oop base, ptrdiff_t offset, T new_value) {
814 if (is_hardwired_primitive<decorators>()) {
815 const DecoratorSet expanded_decorators = decorators | AS_RAW;
816 return PreRuntimeDispatch::atomic_xchg<expanded_decorators>(base, offset, new_value);
817 } else {
818 return RuntimeDispatch<decorators, T, BARRIER_ATOMIC_XCHG_AT>::atomic_xchg_at(base, offset, new_value);
819 }
820 }
821
822 template <DecoratorSet decorators, typename T>
823 inline static typename EnableIf<
824 HasDecorator<decorators, AS_RAW>::value && CanHardwireRaw<decorators>::value, bool>::type
825 arraycopy(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw,
826 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
827 size_t length) {
828 typedef RawAccessBarrier<decorators & RAW_DECORATOR_MASK> Raw;
829 if (HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value) {
830 return Raw::oop_arraycopy(src_obj, src_offset_in_bytes, src_raw,
831 dst_obj, dst_offset_in_bytes, dst_raw,
832 length);
833 } else {
834 return Raw::arraycopy(src_obj, src_offset_in_bytes, src_raw,
835 dst_obj, dst_offset_in_bytes, dst_raw,
836 length);
837 }
838 }
839
840 template <DecoratorSet decorators, typename T>
841 inline static typename EnableIf<
842 HasDecorator<decorators, AS_RAW>::value && !CanHardwireRaw<decorators>::value, bool>::type
843 arraycopy(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw,
844 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
845 size_t length) {
846 if (UseCompressedOops) {
847 const DecoratorSet expanded_decorators = decorators | convert_compressed_oops;
848 return PreRuntimeDispatch::arraycopy<expanded_decorators>(src_obj, src_offset_in_bytes, src_raw,
849 dst_obj, dst_offset_in_bytes, dst_raw,
850 length);
851 } else {
852 const DecoratorSet expanded_decorators = decorators & ~convert_compressed_oops;
853 return PreRuntimeDispatch::arraycopy<expanded_decorators>(src_obj, src_offset_in_bytes, src_raw,
854 dst_obj, dst_offset_in_bytes, dst_raw,
855 length);
856 }
857 }
858
859 template <DecoratorSet decorators, typename T>
860 inline static typename EnableIf<
861 !HasDecorator<decorators, AS_RAW>::value, bool>::type
862 arraycopy(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw,
863 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
864 size_t length) {
865 if (is_hardwired_primitive<decorators>()) {
866 const DecoratorSet expanded_decorators = decorators | AS_RAW;
867 return PreRuntimeDispatch::arraycopy<expanded_decorators>(src_obj, src_offset_in_bytes, src_raw,
868 dst_obj, dst_offset_in_bytes, dst_raw,
869 length);
870 } else {
871 return RuntimeDispatch<decorators, T, BARRIER_ARRAYCOPY>::arraycopy(src_obj, src_offset_in_bytes, src_raw,
872 dst_obj, dst_offset_in_bytes, dst_raw,
873 length);
874 }
875 }
876
877 template <DecoratorSet decorators>
878 inline static typename EnableIf<
879 HasDecorator<decorators, AS_RAW>::value>::type
880 clone(oop src, oop dst, size_t size) {
881 typedef RawAccessBarrier<decorators & RAW_DECORATOR_MASK> Raw;
882 Raw::clone(src, dst, size);
883 }
884
885 template <DecoratorSet decorators>
886 inline static typename EnableIf<
887 !HasDecorator<decorators, AS_RAW>::value>::type
888 clone(oop src, oop dst, size_t size) {
889 RuntimeDispatch<decorators, oop, BARRIER_CLONE>::clone(src, dst, size);
890 }
891 };
892
893 // Step 2: Reduce types.
894 // Enforce that for non-oop types, T and P have to be strictly the same.
895 // P is the type of the address and T is the type of the values.
896 // As for oop types, it is allow to send T in {narrowOop, oop} and
897 // P in {narrowOop, oop, HeapWord*}. The following rules apply according to
898 // the subsequent table. (columns are P, rows are T)
899 // | | HeapWord | oop | narrowOop |
900 // | oop | rt-comp | hw-none | hw-comp |
901 // | narrowOop | x | x | hw-none |
902 //
903 // x means not allowed
904 // rt-comp means it must be checked at runtime whether the oop is compressed.
905 // hw-none means it is statically known the oop will not be compressed.
906 // hw-comp means it is statically known the oop will be compressed.
907
908 template <DecoratorSet decorators, typename T>
909 inline void store_reduce_types(T* addr, T value) {
910 PreRuntimeDispatch::store<decorators>(addr, value);
911 }
912
913 template <DecoratorSet decorators>
914 inline void store_reduce_types(narrowOop* addr, oop value) {
915 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP |
916 INTERNAL_RT_USE_COMPRESSED_OOPS;
917 PreRuntimeDispatch::store<expanded_decorators>(addr, value);
918 }
919
920 template <DecoratorSet decorators>
921 inline void store_reduce_types(narrowOop* addr, narrowOop value) {
922 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP |
923 INTERNAL_RT_USE_COMPRESSED_OOPS;
924 PreRuntimeDispatch::store<expanded_decorators>(addr, value);
925 }
926
927 template <DecoratorSet decorators>
928 inline void store_reduce_types(HeapWord* addr, oop value) {
929 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP;
930 PreRuntimeDispatch::store<expanded_decorators>(addr, value);
931 }
932
933 template <DecoratorSet decorators, typename T>
934 inline T atomic_cmpxchg_reduce_types(T* addr, T compare_value, T new_value) {
935 return PreRuntimeDispatch::atomic_cmpxchg<decorators>(addr, compare_value, new_value);
936 }
937
938 template <DecoratorSet decorators>
939 inline oop atomic_cmpxchg_reduce_types(narrowOop* addr, oop compare_value, oop new_value) {
940 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP |
941 INTERNAL_RT_USE_COMPRESSED_OOPS;
942 return PreRuntimeDispatch::atomic_cmpxchg<expanded_decorators>(addr, compare_value, new_value);
943 }
944
945 template <DecoratorSet decorators>
946 inline narrowOop atomic_cmpxchg_reduce_types(narrowOop* addr, narrowOop compare_value, narrowOop new_value) {
947 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP |
948 INTERNAL_RT_USE_COMPRESSED_OOPS;
949 return PreRuntimeDispatch::atomic_cmpxchg<expanded_decorators>(addr, compare_value, new_value);
950 }
951
952 template <DecoratorSet decorators>
953 inline oop atomic_cmpxchg_reduce_types(HeapWord* addr,
954 oop compare_value,
955 oop new_value) {
956 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP;
957 return PreRuntimeDispatch::atomic_cmpxchg<expanded_decorators>(addr, compare_value, new_value);
958 }
959
960 template <DecoratorSet decorators, typename T>
961 inline T atomic_xchg_reduce_types(T* addr, T new_value) {
962 const DecoratorSet expanded_decorators = decorators;
963 return PreRuntimeDispatch::atomic_xchg<expanded_decorators>(addr, new_value);
964 }
965
966 template <DecoratorSet decorators>
967 inline oop atomic_xchg_reduce_types(narrowOop* addr, oop new_value) {
968 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP |
969 INTERNAL_RT_USE_COMPRESSED_OOPS;
970 return PreRuntimeDispatch::atomic_xchg<expanded_decorators>(addr, new_value);
971 }
972
973 template <DecoratorSet decorators>
974 inline narrowOop atomic_xchg_reduce_types(narrowOop* addr, narrowOop new_value) {
975 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP |
976 INTERNAL_RT_USE_COMPRESSED_OOPS;
977 return PreRuntimeDispatch::atomic_xchg<expanded_decorators>(addr, new_value);
978 }
979
980 template <DecoratorSet decorators>
981 inline oop atomic_xchg_reduce_types(HeapWord* addr, oop new_value) {
982 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP;
983 return PreRuntimeDispatch::atomic_xchg<expanded_decorators>(addr, new_value);
984 }
985
986 template <DecoratorSet decorators, typename T>
987 inline T load_reduce_types(T* addr) {
988 return PreRuntimeDispatch::load<decorators, T>(addr);
989 }
990
991 template <DecoratorSet decorators, typename T>
992 inline typename OopOrNarrowOop<T>::type load_reduce_types(narrowOop* addr) {
993 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP |
994 INTERNAL_RT_USE_COMPRESSED_OOPS;
995 return PreRuntimeDispatch::load<expanded_decorators, typename OopOrNarrowOop<T>::type>(addr);
996 }
997
998 template <DecoratorSet decorators, typename T>
999 inline oop load_reduce_types(HeapWord* addr) {
1000 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP;
1001 return PreRuntimeDispatch::load<expanded_decorators, oop>(addr);
1002 }
1003
1004 template <DecoratorSet decorators, typename T>
1005 inline bool arraycopy_reduce_types(arrayOop src_obj, size_t src_offset_in_bytes, T* src_raw,
1006 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
1007 size_t length) {
1008 return PreRuntimeDispatch::arraycopy<decorators>(src_obj, src_offset_in_bytes, src_raw,
1009 dst_obj, dst_offset_in_bytes, dst_raw,
1010 length);
1011 }
1012
1013 template <DecoratorSet decorators>
1014 inline bool arraycopy_reduce_types(arrayOop src_obj, size_t src_offset_in_bytes, HeapWord* src_raw,
1015 arrayOop dst_obj, size_t dst_offset_in_bytes, HeapWord* dst_raw,
1016 size_t length) {
1017 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP;
1018 return PreRuntimeDispatch::arraycopy<expanded_decorators>(src_obj, src_offset_in_bytes, src_raw,
1019 dst_obj, dst_offset_in_bytes, dst_raw,
1020 length);
1021 }
1022
1023 template <DecoratorSet decorators>
1024 inline bool arraycopy_reduce_types(arrayOop src_obj, size_t src_offset_in_bytes, narrowOop* src_raw,
1025 arrayOop dst_obj, size_t dst_offset_in_bytes, narrowOop* dst_raw,
1026 size_t length) {
1027 const DecoratorSet expanded_decorators = decorators | INTERNAL_CONVERT_COMPRESSED_OOP |
1028 INTERNAL_RT_USE_COMPRESSED_OOPS;
1029 return PreRuntimeDispatch::arraycopy<expanded_decorators>(src_obj, src_offset_in_bytes, src_raw,
1030 dst_obj, dst_offset_in_bytes, dst_raw,
1031 length);
1032 }
1033
1034 // Step 1: Set default decorators. This step remembers if a type was volatile
1035 // and then sets the MO_RELAXED decorator by default. Otherwise, a default
1036 // memory ordering is set for the access, and the implied decorator rules
1037 // are applied to select sensible defaults for decorators that have not been
1038 // explicitly set. For example, default object referent strength is set to strong.
1039 // This step also decays the types passed in (e.g. getting rid of CV qualifiers
1040 // and references from the types). This step also perform some type verification
1041 // that the passed in types make sense.
1042
1043 template <DecoratorSet decorators, typename T>
1044 static void verify_types(){
1045 // If this fails to compile, then you have sent in something that is
1046 // not recognized as a valid primitive type to a primitive Access function.
1047 STATIC_ASSERT((HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value || // oops have already been validated
1048 (std::is_pointer<T>::value || std::is_integral<T>::value) ||
1049 std::is_floating_point<T>::value)); // not allowed primitive type
1050 }
1051
1052 template <DecoratorSet decorators, typename P, typename T>
1053 inline void store(P* addr, T value) {
1054 verify_types<decorators, T>();
1055 using DecayedP = std::decay_t<P>;
1056 using DecayedT = std::decay_t<T>;
1057 DecayedT decayed_value = value;
1058 // If a volatile address is passed in but no memory ordering decorator,
1059 // set the memory ordering to MO_RELAXED by default.
1060 const DecoratorSet expanded_decorators = DecoratorFixup<
1061 (std::is_volatile<P>::value && !HasDecorator<decorators, MO_DECORATOR_MASK>::value) ?
1062 (MO_RELAXED | decorators) : decorators>::value;
1063 store_reduce_types<expanded_decorators>(const_cast<DecayedP*>(addr), decayed_value);
1064 }
1065
1066 template <DecoratorSet decorators, typename T>
1067 inline void store_at(oop base, ptrdiff_t offset, T value) {
1068 verify_types<decorators, T>();
1069 using DecayedT = std::decay_t<T>;
1070 DecayedT decayed_value = value;
1071 const DecoratorSet expanded_decorators = DecoratorFixup<decorators |
1072 (HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value ?
1073 INTERNAL_CONVERT_COMPRESSED_OOP : DECORATORS_NONE)>::value;
1074 PreRuntimeDispatch::store_at<expanded_decorators>(base, offset, decayed_value);
1075 }
1076
1077 template <DecoratorSet decorators, typename P, typename T>
1078 inline T load(P* addr) {
1079 verify_types<decorators, T>();
1080 using DecayedP = std::decay_t<P>;
1081 using DecayedT = std::conditional_t<HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value,
1082 typename OopOrNarrowOop<T>::type,
1083 std::decay_t<T>>;
1084 // If a volatile address is passed in but no memory ordering decorator,
1085 // set the memory ordering to MO_RELAXED by default.
1086 const DecoratorSet expanded_decorators = DecoratorFixup<
1087 (std::is_volatile<P>::value && !HasDecorator<decorators, MO_DECORATOR_MASK>::value) ?
1088 (MO_RELAXED | decorators) : decorators>::value;
1089 return load_reduce_types<expanded_decorators, DecayedT>(const_cast<DecayedP*>(addr));
1090 }
1091
1092 template <DecoratorSet decorators, typename T>
1093 inline T load_at(oop base, ptrdiff_t offset) {
1094 verify_types<decorators, T>();
1095 using DecayedT = std::conditional_t<HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value,
1096 typename OopOrNarrowOop<T>::type,
1097 std::decay_t<T>>;
1098 // Expand the decorators (figure out sensible defaults)
1099 // Potentially remember if we need compressed oop awareness
1100 const DecoratorSet expanded_decorators = DecoratorFixup<decorators |
1101 (HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value ?
1102 INTERNAL_CONVERT_COMPRESSED_OOP : DECORATORS_NONE)>::value;
1103 return PreRuntimeDispatch::load_at<expanded_decorators, DecayedT>(base, offset);
1104 }
1105
1106 template <DecoratorSet decorators, typename P, typename T>
1107 inline T atomic_cmpxchg(P* addr, T compare_value, T new_value) {
1108 verify_types<decorators, T>();
1109 using DecayedP = std::decay_t<P>;
1110 using DecayedT = std::decay_t<T>;
1111 DecayedT new_decayed_value = new_value;
1112 DecayedT compare_decayed_value = compare_value;
1113 const DecoratorSet expanded_decorators = DecoratorFixup<
1114 (!HasDecorator<decorators, MO_DECORATOR_MASK>::value) ?
1115 (MO_SEQ_CST | decorators) : decorators>::value;
1116 return atomic_cmpxchg_reduce_types<expanded_decorators>(const_cast<DecayedP*>(addr),
1117 compare_decayed_value,
1118 new_decayed_value);
1119 }
1120
1121 template <DecoratorSet decorators, typename T>
1122 inline T atomic_cmpxchg_at(oop base, ptrdiff_t offset, T compare_value, T new_value) {
1123 verify_types<decorators, T>();
1124 using DecayedT = std::decay_t<T>;
1125 DecayedT new_decayed_value = new_value;
1126 DecayedT compare_decayed_value = compare_value;
1127 // Determine default memory ordering
1128 const DecoratorSet expanded_decorators = DecoratorFixup<
1129 (!HasDecorator<decorators, MO_DECORATOR_MASK>::value) ?
1130 (MO_SEQ_CST | decorators) : decorators>::value;
1131 // Potentially remember that we need compressed oop awareness
1132 const DecoratorSet final_decorators = expanded_decorators |
1133 (HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value ?
1134 INTERNAL_CONVERT_COMPRESSED_OOP : DECORATORS_NONE);
1135 return PreRuntimeDispatch::atomic_cmpxchg_at<final_decorators>(base, offset, compare_decayed_value,
1136 new_decayed_value);
1137 }
1138
1139 template <DecoratorSet decorators, typename P, typename T>
1140 inline T atomic_xchg(P* addr, T new_value) {
1141 verify_types<decorators, T>();
1142 using DecayedP = std::decay_t<P>;
1143 using DecayedT = std::decay_t<T>;
1144 DecayedT new_decayed_value = new_value;
1145 // atomic_xchg is only available in SEQ_CST flavour.
1146 const DecoratorSet expanded_decorators = DecoratorFixup<decorators | MO_SEQ_CST>::value;
1147 return atomic_xchg_reduce_types<expanded_decorators>(const_cast<DecayedP*>(addr),
1148 new_decayed_value);
1149 }
1150
1151 template <DecoratorSet decorators, typename T>
1152 inline T atomic_xchg_at(oop base, ptrdiff_t offset, T new_value) {
1153 verify_types<decorators, T>();
1154 using DecayedT = std::decay_t<T>;
1155 DecayedT new_decayed_value = new_value;
1156 // atomic_xchg is only available in SEQ_CST flavour.
1157 const DecoratorSet expanded_decorators = DecoratorFixup<decorators | MO_SEQ_CST |
1158 (HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value ?
1159 INTERNAL_CONVERT_COMPRESSED_OOP : DECORATORS_NONE)>::value;
1160 return PreRuntimeDispatch::atomic_xchg_at<expanded_decorators>(base, offset, new_decayed_value);
1161 }
1162
1163 template <DecoratorSet decorators, typename T>
1164 inline bool arraycopy(arrayOop src_obj, size_t src_offset_in_bytes, const T* src_raw,
1165 arrayOop dst_obj, size_t dst_offset_in_bytes, T* dst_raw,
1166 size_t length) {
1167 STATIC_ASSERT((HasDecorator<decorators, INTERNAL_VALUE_IS_OOP>::value ||
1168 (std::is_same<T, void>::value || std::is_integral<T>::value) ||
1169 std::is_floating_point<T>::value)); // arraycopy allows type erased void elements
1170 using DecayedT = std::decay_t<T>;
1171 const DecoratorSet expanded_decorators = DecoratorFixup<decorators | IS_ARRAY | IN_HEAP>::value;
1172 return arraycopy_reduce_types<expanded_decorators>(src_obj, src_offset_in_bytes, const_cast<DecayedT*>(src_raw),
1173 dst_obj, dst_offset_in_bytes, const_cast<DecayedT*>(dst_raw),
1174 length);
1175 }
1176
1177 template <DecoratorSet decorators>
1178 inline void clone(oop src, oop dst, size_t size) {
1179 const DecoratorSet expanded_decorators = DecoratorFixup<decorators>::value;
1180 PreRuntimeDispatch::clone<expanded_decorators>(src, dst, size);
1181 }
1182
1183 // Infer the type that should be returned from an Access::oop_load.
1184 template <typename P, DecoratorSet decorators>
1185 class OopLoadProxy: public StackObj {
1186 private:
1187 P *const _addr;
1188 public:
1189 explicit OopLoadProxy(P* addr) : _addr(addr) {}
1190
1191 inline operator oop() {
1192 return load<decorators | INTERNAL_VALUE_IS_OOP, P, oop>(_addr);
1193 }
1194
1195 inline operator narrowOop() {
1196 return load<decorators | INTERNAL_VALUE_IS_OOP, P, narrowOop>(_addr);
1197 }
1198
1199 template <typename T>
1200 inline bool operator ==(const T& other) const {
1201 return load<decorators | INTERNAL_VALUE_IS_OOP, P, T>(_addr) == other;
1202 }
1203
1204 template <typename T>
1205 inline bool operator !=(const T& other) const {
1206 return load<decorators | INTERNAL_VALUE_IS_OOP, P, T>(_addr) != other;
1207 }
1208
1209 inline bool operator ==(std::nullptr_t) const {
1210 return load<decorators | INTERNAL_VALUE_IS_OOP, P, oop>(_addr) == nullptr;
1211 }
1212
1213 inline bool operator !=(std::nullptr_t) const {
1214 return load<decorators | INTERNAL_VALUE_IS_OOP, P, oop>(_addr) != nullptr;
1215 }
1216 };
1217
1218 // Infer the type that should be returned from an Access::load_at.
1219 template <DecoratorSet decorators>
1220 class LoadAtProxy: public StackObj {
1221 private:
1222 const oop _base;
1223 const ptrdiff_t _offset;
1224 public:
1225 LoadAtProxy(oop base, ptrdiff_t offset) : _base(base), _offset(offset) {}
1226
1227 template <typename T>
1228 inline operator T() const {
1229 return load_at<decorators, T>(_base, _offset);
1230 }
1231
1232 template <typename T>
1233 inline bool operator ==(const T& other) const { return load_at<decorators, T>(_base, _offset) == other; }
1234
1235 template <typename T>
1236 inline bool operator !=(const T& other) const { return load_at<decorators, T>(_base, _offset) != other; }
1237 };
1238
1239 // Infer the type that should be returned from an Access::oop_load_at.
1240 template <DecoratorSet decorators>
1241 class OopLoadAtProxy: public StackObj {
1242 private:
1243 const oop _base;
1244 const ptrdiff_t _offset;
1245 public:
1246 OopLoadAtProxy(oop base, ptrdiff_t offset) : _base(base), _offset(offset) {}
1247
1248 inline operator oop() const {
1249 return load_at<decorators | INTERNAL_VALUE_IS_OOP, oop>(_base, _offset);
1250 }
1251
1252 inline operator narrowOop() const {
1253 return load_at<decorators | INTERNAL_VALUE_IS_OOP, narrowOop>(_base, _offset);
1254 }
1255
1256 template <typename T>
1257 inline bool operator ==(const T& other) const {
1258 return load_at<decorators | INTERNAL_VALUE_IS_OOP, T>(_base, _offset) == other;
1259 }
1260
1261 template <typename T>
1262 inline bool operator !=(const T& other) const {
1263 return load_at<decorators | INTERNAL_VALUE_IS_OOP, T>(_base, _offset) != other;
1264 }
1265 };
1266 }
1267
1268 #endif // SHARE_OOPS_ACCESSBACKEND_HPP