< prev index next >

src/hotspot/share/gc/serial/genMarkSweep.cpp

Print this page

 28 #include "classfile/stringTable.hpp"
 29 #include "classfile/symbolTable.hpp"
 30 #include "classfile/systemDictionary.hpp"
 31 #include "classfile/vmSymbols.hpp"
 32 #include "code/codeCache.hpp"
 33 #include "code/icBuffer.hpp"
 34 #include "compiler/oopMap.hpp"
 35 #include "gc/serial/genMarkSweep.hpp"
 36 #include "gc/serial/serialGcRefProcProxyTask.hpp"
 37 #include "gc/shared/collectedHeap.inline.hpp"
 38 #include "gc/shared/gcHeapSummary.hpp"
 39 #include "gc/shared/gcTimer.hpp"
 40 #include "gc/shared/gcTrace.hpp"
 41 #include "gc/shared/gcTraceTime.inline.hpp"
 42 #include "gc/shared/genCollectedHeap.hpp"
 43 #include "gc/shared/generation.hpp"
 44 #include "gc/shared/genOopClosures.inline.hpp"
 45 #include "gc/shared/modRefBarrierSet.hpp"
 46 #include "gc/shared/referencePolicy.hpp"
 47 #include "gc/shared/referenceProcessorPhaseTimes.hpp"

 48 #include "gc/shared/space.hpp"
 49 #include "gc/shared/strongRootsScope.hpp"
 50 #include "gc/shared/weakProcessor.hpp"
 51 #include "memory/universe.hpp"
 52 #include "oops/instanceRefKlass.hpp"
 53 #include "oops/oop.inline.hpp"
 54 #include "prims/jvmtiExport.hpp"
 55 #include "runtime/handles.inline.hpp"
 56 #include "runtime/synchronizer.hpp"
 57 #include "runtime/thread.inline.hpp"
 58 #include "runtime/vmThread.hpp"
 59 #include "utilities/copy.hpp"
 60 #include "utilities/events.hpp"
 61 #include "utilities/stack.inline.hpp"
 62 #if INCLUDE_JVMCI
 63 #include "jvmci/jvmci.hpp"
 64 #endif
 65 
 66 void GenMarkSweep::invoke_at_safepoint(ReferenceProcessor* rp, bool clear_all_softrefs) {
 67   assert(SafepointSynchronize::is_at_safepoint(), "must be at a safepoint");

 76   // hook up weak ref data so it can be used during Mark-Sweep
 77   assert(ref_processor() == NULL, "no stomping");
 78   assert(rp != NULL, "should be non-NULL");
 79   set_ref_processor(rp);
 80   rp->setup_policy(clear_all_softrefs);
 81 
 82   gch->trace_heap_before_gc(_gc_tracer);
 83 
 84   // Increment the invocation count
 85   _total_invocations++;
 86 
 87   // Capture used regions for each generation that will be
 88   // subject to collection, so that card table adjustments can
 89   // be made intelligently (see clear / invalidate further below).
 90   gch->save_used_regions();
 91 
 92   allocate_stacks();
 93 
 94   mark_sweep_phase1(clear_all_softrefs);
 95 


 96   mark_sweep_phase2();
 97 
 98   // Don't add any more derived pointers during phase3
 99 #if COMPILER2_OR_JVMCI
100   assert(DerivedPointerTable::is_active(), "Sanity");
101   DerivedPointerTable::set_active(false);
102 #endif
103 
104   mark_sweep_phase3();
105 
106   mark_sweep_phase4();
107 
108   restore_marks();
109 
110   // Set saved marks for allocation profiler (and other things? -- dld)
111   // (Should this be in general part?)
112   gch->save_marks();
113 


114   deallocate_stacks();
115 
116   // If compaction completely evacuated the young generation then we
117   // can clear the card table.  Otherwise, we must invalidate
118   // it (consider all cards dirty).  In the future, we might consider doing
119   // compaction within generations only, and doing card-table sliding.
120   CardTableRS* rs = gch->rem_set();
121   Generation* old_gen = gch->old_gen();
122 
123   // Clear/invalidate below make use of the "prev_used_regions" saved earlier.
124   if (gch->young_gen()->used() == 0) {
125     // We've evacuated the young generation.
126     rs->clear_into_younger(old_gen);
127   } else {
128     // Invalidate the cards corresponding to the currently used
129     // region and clear those corresponding to the evacuated region.
130     rs->invalidate_or_clear(old_gen);
131   }
132 
133   gch->prune_scavengable_nmethods();

257 
258   gch->prepare_for_compaction();
259 }
260 
261 class GenAdjustPointersClosure: public GenCollectedHeap::GenClosure {
262 public:
263   void do_generation(Generation* gen) {
264     gen->adjust_pointers();
265   }
266 };
267 
268 void GenMarkSweep::mark_sweep_phase3() {
269   GenCollectedHeap* gch = GenCollectedHeap::heap();
270 
271   // Adjust the pointers to reflect the new locations
272   GCTraceTime(Info, gc, phases) tm("Phase 3: Adjust pointers", gc_timer());
273 
274   // Need new claim bits for the pointer adjustment tracing.
275   ClassLoaderDataGraph::clear_claimed_marks();
276 
277   {
278     StrongRootsScope srs(0);
279 
280     gch->full_process_roots(true,  // this is the adjust phase
281                             GenCollectedHeap::SO_AllCodeCache,
282                             false, // all roots
283                             &adjust_pointer_closure,
284                             &adjust_cld_closure);
















285   }
286 
287   gch->gen_process_weak_roots(&adjust_pointer_closure);
288 
289   adjust_marks();
290   GenAdjustPointersClosure blk;
291   gch->generation_iterate(&blk, true);
292 }
293 
294 class GenCompactClosure: public GenCollectedHeap::GenClosure {
295 public:
296   void do_generation(Generation* gen) {
297     gen->compact();
298   }
299 };
300 
301 void GenMarkSweep::mark_sweep_phase4() {
302   // All pointers are now adjusted, move objects accordingly
303 
304   // It is imperative that we traverse perm_gen first in phase4. All
305   // classes must be allocated earlier than their instances, and traversing
306   // perm_gen first makes sure that all Klass*s have moved to their new
307   // location before any instance does a dispatch through it's klass!
308 

 28 #include "classfile/stringTable.hpp"
 29 #include "classfile/symbolTable.hpp"
 30 #include "classfile/systemDictionary.hpp"
 31 #include "classfile/vmSymbols.hpp"
 32 #include "code/codeCache.hpp"
 33 #include "code/icBuffer.hpp"
 34 #include "compiler/oopMap.hpp"
 35 #include "gc/serial/genMarkSweep.hpp"
 36 #include "gc/serial/serialGcRefProcProxyTask.hpp"
 37 #include "gc/shared/collectedHeap.inline.hpp"
 38 #include "gc/shared/gcHeapSummary.hpp"
 39 #include "gc/shared/gcTimer.hpp"
 40 #include "gc/shared/gcTrace.hpp"
 41 #include "gc/shared/gcTraceTime.inline.hpp"
 42 #include "gc/shared/genCollectedHeap.hpp"
 43 #include "gc/shared/generation.hpp"
 44 #include "gc/shared/genOopClosures.inline.hpp"
 45 #include "gc/shared/modRefBarrierSet.hpp"
 46 #include "gc/shared/referencePolicy.hpp"
 47 #include "gc/shared/referenceProcessorPhaseTimes.hpp"
 48 #include "gc/shared/slidingForwarding.hpp"
 49 #include "gc/shared/space.hpp"
 50 #include "gc/shared/strongRootsScope.hpp"
 51 #include "gc/shared/weakProcessor.hpp"
 52 #include "memory/universe.hpp"
 53 #include "oops/instanceRefKlass.hpp"
 54 #include "oops/oop.inline.hpp"
 55 #include "prims/jvmtiExport.hpp"
 56 #include "runtime/handles.inline.hpp"
 57 #include "runtime/synchronizer.hpp"
 58 #include "runtime/thread.inline.hpp"
 59 #include "runtime/vmThread.hpp"
 60 #include "utilities/copy.hpp"
 61 #include "utilities/events.hpp"
 62 #include "utilities/stack.inline.hpp"
 63 #if INCLUDE_JVMCI
 64 #include "jvmci/jvmci.hpp"
 65 #endif
 66 
 67 void GenMarkSweep::invoke_at_safepoint(ReferenceProcessor* rp, bool clear_all_softrefs) {
 68   assert(SafepointSynchronize::is_at_safepoint(), "must be at a safepoint");

 77   // hook up weak ref data so it can be used during Mark-Sweep
 78   assert(ref_processor() == NULL, "no stomping");
 79   assert(rp != NULL, "should be non-NULL");
 80   set_ref_processor(rp);
 81   rp->setup_policy(clear_all_softrefs);
 82 
 83   gch->trace_heap_before_gc(_gc_tracer);
 84 
 85   // Increment the invocation count
 86   _total_invocations++;
 87 
 88   // Capture used regions for each generation that will be
 89   // subject to collection, so that card table adjustments can
 90   // be made intelligently (see clear / invalidate further below).
 91   gch->save_used_regions();
 92 
 93   allocate_stacks();
 94 
 95   mark_sweep_phase1(clear_all_softrefs);
 96 
 97   SlidingForwarding::begin();
 98 
 99   mark_sweep_phase2();
100 
101   // Don't add any more derived pointers during phase3
102 #if COMPILER2_OR_JVMCI
103   assert(DerivedPointerTable::is_active(), "Sanity");
104   DerivedPointerTable::set_active(false);
105 #endif
106 
107   mark_sweep_phase3();
108 
109   mark_sweep_phase4();
110 
111   restore_marks();
112 
113   // Set saved marks for allocation profiler (and other things? -- dld)
114   // (Should this be in general part?)
115   gch->save_marks();
116 
117   SlidingForwarding::end();
118 
119   deallocate_stacks();
120 
121   // If compaction completely evacuated the young generation then we
122   // can clear the card table.  Otherwise, we must invalidate
123   // it (consider all cards dirty).  In the future, we might consider doing
124   // compaction within generations only, and doing card-table sliding.
125   CardTableRS* rs = gch->rem_set();
126   Generation* old_gen = gch->old_gen();
127 
128   // Clear/invalidate below make use of the "prev_used_regions" saved earlier.
129   if (gch->young_gen()->used() == 0) {
130     // We've evacuated the young generation.
131     rs->clear_into_younger(old_gen);
132   } else {
133     // Invalidate the cards corresponding to the currently used
134     // region and clear those corresponding to the evacuated region.
135     rs->invalidate_or_clear(old_gen);
136   }
137 
138   gch->prune_scavengable_nmethods();

262 
263   gch->prepare_for_compaction();
264 }
265 
266 class GenAdjustPointersClosure: public GenCollectedHeap::GenClosure {
267 public:
268   void do_generation(Generation* gen) {
269     gen->adjust_pointers();
270   }
271 };
272 
273 void GenMarkSweep::mark_sweep_phase3() {
274   GenCollectedHeap* gch = GenCollectedHeap::heap();
275 
276   // Adjust the pointers to reflect the new locations
277   GCTraceTime(Info, gc, phases) tm("Phase 3: Adjust pointers", gc_timer());
278 
279   // Need new claim bits for the pointer adjustment tracing.
280   ClassLoaderDataGraph::clear_claimed_marks();
281 
282   if (UseAltGCForwarding) {
283     AdjustPointerClosure<true> adjust_pointer_closure;
284     CLDToOopClosure adjust_cld_closure(&adjust_pointer_closure, ClassLoaderData::_claim_strong);
285     {
286       StrongRootsScope srs(0);
287       gch->full_process_roots(true,  // this is the adjust phase
288                               GenCollectedHeap::SO_AllCodeCache,
289                               false, // all roots
290                               &adjust_pointer_closure,
291                               &adjust_cld_closure);
292     }
293     gch->gen_process_weak_roots(&adjust_pointer_closure);
294   } else {
295     AdjustPointerClosure<false> adjust_pointer_closure;
296     CLDToOopClosure adjust_cld_closure(&adjust_pointer_closure, ClassLoaderData::_claim_strong);
297     {
298       StrongRootsScope srs(0);
299       gch->full_process_roots(true,  // this is the adjust phase
300                               GenCollectedHeap::SO_AllCodeCache,
301                               false, // all roots
302                               &adjust_pointer_closure,
303                               &adjust_cld_closure);
304     }
305     gch->gen_process_weak_roots(&adjust_pointer_closure);
306   }



307   adjust_marks();
308   GenAdjustPointersClosure blk;
309   gch->generation_iterate(&blk, true);
310 }
311 
312 class GenCompactClosure: public GenCollectedHeap::GenClosure {
313 public:
314   void do_generation(Generation* gen) {
315     gen->compact();
316   }
317 };
318 
319 void GenMarkSweep::mark_sweep_phase4() {
320   // All pointers are now adjusted, move objects accordingly
321 
322   // It is imperative that we traverse perm_gen first in phase4. All
323   // classes must be allocated earlier than their instances, and traversing
324   // perm_gen first makes sure that all Klass*s have moved to their new
325   // location before any instance does a dispatch through it's klass!
326 
< prev index next >