< prev index next >

src/hotspot/cpu/x86/stubGenerator_x86_64.hpp

Print this page

608   // only for C2, but done for C1 as well) we need a callee-saved oop
609   // map and therefore have to make these stubs into RuntimeStubs
610   // rather than BufferBlobs.  If the compiler needs all registers to
611   // be preserved between the fault point and the exception handler
612   // then it must assume responsibility for that in
613   // AbstractCompiler::continuation_for_implicit_null_exception or
614   // continuation_for_implicit_division_by_zero_exception. All other
615   // implicit exceptions (e.g., NullPointerException or
616   // AbstractMethodError on entry) are either at call sites or
617   // otherwise assume that stack unwinding will be initiated, so
618   // caller saved registers were assumed volatile in the compiler.
619   address generate_throw_exception(const char* name,
620                                    address runtime_entry,
621                                    Register arg1 = noreg,
622                                    Register arg2 = noreg);
623 
624   // shared exception handler for FFM upcall stubs
625   address generate_upcall_stub_exception_handler();
626   address generate_upcall_stub_load_target();
627 



628   // Specialized stub implementations for UseSecondarySupersTable.
629   void generate_lookup_secondary_supers_table_stub();
630 
631   // Slow path implementation for UseSecondarySupersTable.
632   address generate_lookup_secondary_supers_table_slow_path_stub();
633 
634   void create_control_words();
635 
636   // Initialization
637   void generate_preuniverse_stubs();
638   void generate_initial_stubs();
639   void generate_continuation_stubs();
640   void generate_compiler_stubs();
641   void generate_final_stubs();
642 
643 public:
644   StubGenerator(CodeBuffer* code, StubGenBlobId blob_id);
645 };
646 
647 #endif // CPU_X86_STUBGENERATOR_X86_64_HPP

608   // only for C2, but done for C1 as well) we need a callee-saved oop
609   // map and therefore have to make these stubs into RuntimeStubs
610   // rather than BufferBlobs.  If the compiler needs all registers to
611   // be preserved between the fault point and the exception handler
612   // then it must assume responsibility for that in
613   // AbstractCompiler::continuation_for_implicit_null_exception or
614   // continuation_for_implicit_division_by_zero_exception. All other
615   // implicit exceptions (e.g., NullPointerException or
616   // AbstractMethodError on entry) are either at call sites or
617   // otherwise assume that stack unwinding will be initiated, so
618   // caller saved registers were assumed volatile in the compiler.
619   address generate_throw_exception(const char* name,
620                                    address runtime_entry,
621                                    Register arg1 = noreg,
622                                    Register arg2 = noreg);
623 
624   // shared exception handler for FFM upcall stubs
625   address generate_upcall_stub_exception_handler();
626   address generate_upcall_stub_load_target();
627 
628   // interpreter or compiled code marshalling registers to/from inline type instance
629   address generate_return_value_stub(address destination, const char* name, bool has_res);
630 
631   // Specialized stub implementations for UseSecondarySupersTable.
632   void generate_lookup_secondary_supers_table_stub();
633 
634   // Slow path implementation for UseSecondarySupersTable.
635   address generate_lookup_secondary_supers_table_slow_path_stub();
636 
637   void create_control_words();
638 
639   // Initialization
640   void generate_preuniverse_stubs();
641   void generate_initial_stubs();
642   void generate_continuation_stubs();
643   void generate_compiler_stubs();
644   void generate_final_stubs();
645 
646 public:
647   StubGenerator(CodeBuffer* code, StubGenBlobId blob_id);
648 };
649 
650 #endif // CPU_X86_STUBGENERATOR_X86_64_HPP
< prev index next >