1 /*
  2  * Copyright (c) 2014, 2020, Oracle and/or its affiliates. All rights reserved.
  3  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
  4  *
  5  * This code is free software; you can redistribute it and/or modify it
  6  * under the terms of the GNU General Public License version 2 only, as
  7  * published by the Free Software Foundation.
  8  *
  9  * This code is distributed in the hope that it will be useful, but WITHOUT
 10  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
 11  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
 12  * version 2 for more details (a copy is included in the LICENSE file that
 13  * accompanied this code).
 14  *
 15  * You should have received a copy of the GNU General Public License version
 16  * 2 along with this work; if not, write to the Free Software Foundation,
 17  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
 18  *
 19  * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
 20  * or visit www.oracle.com if you need additional information or have any
 21  * questions.
 22  *
 23  */
 24 #include "precompiled.hpp"
 25 #include "jfr/leakprofiler/chains/bfsClosure.hpp"
 26 #include "jfr/leakprofiler/chains/dfsClosure.hpp"
 27 #include "jfr/leakprofiler/chains/edge.hpp"
 28 #include "jfr/leakprofiler/chains/edgeStore.hpp"
 29 #include "jfr/leakprofiler/chains/edgeQueue.hpp"
 30 #include "jfr/leakprofiler/chains/jfrbitset.hpp"
 31 #include "jfr/leakprofiler/utilities/granularTimer.hpp"
 32 #include "jfr/leakprofiler/utilities/unifiedOopRef.inline.hpp"
 33 #include "logging/log.hpp"
 34 #include "memory/iterator.inline.hpp"
 35 #include "memory/resourceArea.hpp"
 36 #include "oops/access.inline.hpp"
 37 #include "oops/oop.inline.hpp"
 38 #include "utilities/align.hpp"
 39 
 40 BFSClosure::BFSClosure(EdgeQueue* edge_queue, EdgeStore* edge_store, JFRBitSet* mark_bits) :
 41   _edge_queue(edge_queue),
 42   _edge_store(edge_store),
 43   _mark_bits(mark_bits),
 44   _current_parent(NULL),
 45   _current_frontier_level(0),
 46   _next_frontier_idx(0),
 47   _prev_frontier_idx(0),
 48   _dfs_fallback_idx(0),
 49   _use_dfs(false) {
 50 }
 51 
 52 static void log_frontier_level_summary(size_t level,
 53                                        size_t high_idx,
 54                                        size_t low_idx,
 55                                        size_t edge_size) {
 56   const size_t nof_edges_in_frontier = high_idx - low_idx;
 57   log_trace(jfr, system)(
 58       "BFS front: " SIZE_FORMAT " edges: " SIZE_FORMAT " size: " SIZE_FORMAT " [KB]",
 59       level,
 60       nof_edges_in_frontier,
 61       (nof_edges_in_frontier * edge_size) / K
 62                         );
 63 }
 64 
 65 void BFSClosure::log_completed_frontier() const {
 66   log_frontier_level_summary(_current_frontier_level,
 67                              _next_frontier_idx,
 68                              _prev_frontier_idx,
 69                              _edge_queue->sizeof_edge());
 70 }
 71 
 72 void BFSClosure::log_dfs_fallback() const {
 73   const size_t edge_size = _edge_queue->sizeof_edge();
 74   // first complete summary for frontier in progress
 75   log_frontier_level_summary(_current_frontier_level,
 76                              _next_frontier_idx,
 77                              _prev_frontier_idx,
 78                              edge_size);
 79 
 80   // and then also complete the last frontier
 81   log_frontier_level_summary(_current_frontier_level + 1,
 82                              _edge_queue->bottom(),
 83                              _next_frontier_idx,
 84                              edge_size);
 85 
 86   // additional information about DFS fallover
 87   log_trace(jfr, system)(
 88       "BFS front: " SIZE_FORMAT " filled edge queue at edge: " SIZE_FORMAT,
 89       _current_frontier_level,
 90       _dfs_fallback_idx
 91                         );
 92 
 93   const size_t nof_dfs_completed_edges = _edge_queue->bottom() - _dfs_fallback_idx;
 94   log_trace(jfr, system)(
 95       "DFS to complete " SIZE_FORMAT " edges size: " SIZE_FORMAT " [KB]",
 96       nof_dfs_completed_edges,
 97       (nof_dfs_completed_edges * edge_size) / K
 98                         );
 99 }
100 
101 void BFSClosure::process() {
102   process_root_set();
103   process_queue();
104 }
105 
106 void BFSClosure::process_root_set() {
107   for (size_t idx = _edge_queue->bottom(); idx < _edge_queue->top(); ++idx) {
108     const Edge* edge = _edge_queue->element_at(idx);
109     assert(edge->parent() == NULL, "invariant");
110     process(edge->reference(), edge->pointee());
111   }
112 }
113 
114 void BFSClosure::process(UnifiedOopRef reference, const oop pointee) {
115   closure_impl(reference, pointee);
116 }
117 void BFSClosure::closure_impl(UnifiedOopRef reference, const oop pointee) {
118   assert(!reference.is_null(), "invariant");
119   assert(reference.dereference() == pointee, "invariant");
120 
121   if (GranularTimer::is_finished()) {
122      return;
123   }
124 
125   if (_use_dfs) {
126     assert(_current_parent != NULL, "invariant");
127     DFSClosure::find_leaks_from_edge(_edge_store, _mark_bits, _current_parent);
128     return;
129   }
130 
131   if (!_mark_bits->is_marked(pointee)) {
132     _mark_bits->mark_obj(pointee);
133     // is the pointee a sample object?
134     if (pointee->mark().is_marked()) {
135       add_chain(reference, pointee);
136     }
137 
138     // if we are processinig initial root set, don't add to queue
139     if (_current_parent != NULL) {
140       _edge_queue->add(_current_parent, reference);
141     }
142 
143     if (_edge_queue->is_full()) {
144       dfs_fallback();
145     }
146   }
147 }
148 
149 void BFSClosure::add_chain(UnifiedOopRef reference, const oop pointee) {
150   assert(pointee != NULL, "invariant");
151   assert(pointee->mark().is_marked(), "invariant");
152   Edge leak_edge(_current_parent, reference);
153   _edge_store->put_chain(&leak_edge, _current_parent == NULL ? 1 : _current_frontier_level + 2);
154 }
155 
156 void BFSClosure::dfs_fallback() {
157   assert(_edge_queue->is_full(), "invariant");
158   _use_dfs = true;
159   _dfs_fallback_idx = _edge_queue->bottom();
160   while (!_edge_queue->is_empty()) {
161     const Edge* edge = _edge_queue->remove();
162     if (edge->pointee() != NULL) {
163       DFSClosure::find_leaks_from_edge(_edge_store, _mark_bits, edge);
164     }
165   }
166 }
167 
168 void BFSClosure::process_queue() {
169   assert(_current_frontier_level == 0, "invariant");
170   assert(_next_frontier_idx == 0, "invariant");
171   assert(_prev_frontier_idx == 0, "invariant");
172 
173   _next_frontier_idx = _edge_queue->top();
174   while (!is_complete()) {
175     iterate(_edge_queue->remove()); // edge_queue.remove() increments bottom
176   }
177 }
178 
179 void BFSClosure::step_frontier() const {
180   log_completed_frontier();
181   ++_current_frontier_level;
182   _prev_frontier_idx = _next_frontier_idx;
183   _next_frontier_idx = _edge_queue->top();
184 }
185 
186 bool BFSClosure::is_complete() const {
187   if (_edge_queue->bottom() < _next_frontier_idx) {
188     return false;
189   }
190   if (_edge_queue->bottom() > _next_frontier_idx) {
191     // fallback onto DFS as part of processing the frontier
192     assert(_dfs_fallback_idx >= _prev_frontier_idx, "invariant");
193     assert(_dfs_fallback_idx < _next_frontier_idx, "invariant");
194     log_dfs_fallback();
195     return true;
196   }
197   assert(_edge_queue->bottom() == _next_frontier_idx, "invariant");
198   if (_edge_queue->is_empty()) {
199     return true;
200   }
201   step_frontier();
202   return false;
203 }
204 
205 void BFSClosure::iterate(const Edge* parent) {
206   assert(parent != NULL, "invariant");
207   const oop pointee = parent->pointee();
208   assert(pointee != NULL, "invariant");
209   _current_parent = parent;
210   pointee->oop_iterate(this);
211 }
212 
213 void BFSClosure::do_oop(oop* ref) {
214   assert(ref != NULL, "invariant");
215   assert(is_aligned(ref, HeapWordSize), "invariant");
216   const oop pointee = HeapAccess<AS_NO_KEEPALIVE>::oop_load(ref);
217   if (pointee != NULL) {
218     closure_impl(UnifiedOopRef::encode_in_heap(ref), pointee);
219   }
220 }
221 
222 void BFSClosure::do_oop(narrowOop* ref) {
223   assert(ref != NULL, "invariant");
224   assert(is_aligned(ref, sizeof(narrowOop)), "invariant");
225   const oop pointee = HeapAccess<AS_NO_KEEPALIVE>::oop_load(ref);
226   if (pointee != NULL) {
227     closure_impl(UnifiedOopRef::encode_in_heap(ref), pointee);
228   }
229 }
230 
231 void BFSClosure::do_root(UnifiedOopRef ref) {
232   assert(ref.dereference() != NULL, "pointee must not be null");
233   if (!_edge_queue->is_full()) {
234     _edge_queue->add(NULL, ref);
235   }
236 }