46 inline int InstanceKlass::java_fields_count() const { return FieldInfoStream::num_java_fields(fieldinfo_stream()); }
47 inline int InstanceKlass::total_fields_count() const { return FieldInfoStream::num_total_fields(fieldinfo_stream()); }
48
49 inline OopMapBlock* InstanceKlass::start_of_nonstatic_oop_maps() const {
50 return (OopMapBlock*)(start_of_itable() + itable_length());
51 }
52
53 inline Klass** InstanceKlass::end_of_nonstatic_oop_maps() const {
54 return (Klass**)(start_of_nonstatic_oop_maps() +
55 nonstatic_oop_map_count());
56 }
57
58 inline InstanceKlass* volatile* InstanceKlass::adr_implementor() const {
59 if (is_interface()) {
60 return (InstanceKlass* volatile*)end_of_nonstatic_oop_maps();
61 } else {
62 return nullptr;
63 }
64 }
65
66 inline ObjArrayKlass* InstanceKlass::array_klasses_acquire() const {
67 return AtomicAccess::load_acquire(&_array_klasses);
68 }
69
70 inline void InstanceKlass::release_set_array_klasses(ObjArrayKlass* k) {
71 AtomicAccess::release_store(&_array_klasses, k);
72 }
73
74 // The iteration over the oops in objects is a hot path in the GC code.
75 // By force inlining the following functions, we get similar GC performance
76 // as the previous macro based implementation.
77
78 template <typename T, class OopClosureType>
79 ALWAYSINLINE void InstanceKlass::oop_oop_iterate_oop_map(OopMapBlock* map, oop obj, OopClosureType* closure) {
80 T* p = obj->field_addr<T>(map->offset());
81 T* const end = p + map->count();
82
83 for (; p < end; ++p) {
84 Devirtualizer::do_oop(closure, p);
85 }
|
46 inline int InstanceKlass::java_fields_count() const { return FieldInfoStream::num_java_fields(fieldinfo_stream()); }
47 inline int InstanceKlass::total_fields_count() const { return FieldInfoStream::num_total_fields(fieldinfo_stream()); }
48
49 inline OopMapBlock* InstanceKlass::start_of_nonstatic_oop_maps() const {
50 return (OopMapBlock*)(start_of_itable() + itable_length());
51 }
52
53 inline Klass** InstanceKlass::end_of_nonstatic_oop_maps() const {
54 return (Klass**)(start_of_nonstatic_oop_maps() +
55 nonstatic_oop_map_count());
56 }
57
58 inline InstanceKlass* volatile* InstanceKlass::adr_implementor() const {
59 if (is_interface()) {
60 return (InstanceKlass* volatile*)end_of_nonstatic_oop_maps();
61 } else {
62 return nullptr;
63 }
64 }
65
66 inline address InstanceKlass::end_of_instance_klass() const {
67 return (address)end_of_nonstatic_oop_maps() +
68 (is_interface() ? sizeof(InstanceKlass*) : 0);
69 }
70
71 inline InlineKlass* InstanceKlass::get_inline_type_field_klass(int idx) const {
72 assert(has_inline_type_fields(), "Sanity checking");
73 assert(idx < java_fields_count(), "IOOB");
74 InlineKlass* k = inline_layout_info(idx).klass();
75 assert(k != nullptr, "Should always be set before being read");
76 return k;
77 }
78
79 inline InlineKlass* InstanceKlass::get_inline_type_field_klass_or_null(int idx) const {
80 assert(has_inline_type_fields(), "Sanity checking");
81 assert(idx < java_fields_count(), "IOOB");
82 InlineKlass* k = inline_layout_info(idx).klass();
83 return k;
84 }
85
86 inline ObjArrayKlass* InstanceKlass::array_klasses_acquire() const {
87 return AtomicAccess::load_acquire(&_array_klasses);
88 }
89
90 inline void InstanceKlass::release_set_array_klasses(ObjArrayKlass* k) {
91 AtomicAccess::release_store(&_array_klasses, k);
92 }
93
94 // The iteration over the oops in objects is a hot path in the GC code.
95 // By force inlining the following functions, we get similar GC performance
96 // as the previous macro based implementation.
97
98 template <typename T, class OopClosureType>
99 ALWAYSINLINE void InstanceKlass::oop_oop_iterate_oop_map(OopMapBlock* map, oop obj, OopClosureType* closure) {
100 T* p = obj->field_addr<T>(map->offset());
101 T* const end = p + map->count();
102
103 for (; p < end; ++p) {
104 Devirtualizer::do_oop(closure, p);
105 }
|