< prev index next >

src/hotspot/share/opto/matcher.cpp

Print this page
@@ -185,10 +185,54 @@
      }
    }
  }
  #endif
  
+ // Array of RegMask, one per returned values (inline type instances can
+ // be returned as multiple return values, one per field)
+ RegMask* Matcher::return_values_mask(const TypeFunc* tf) {
+   const TypeTuple* range = tf->range_cc();
+   uint cnt = range->cnt() - TypeFunc::Parms;
+   if (cnt == 0) {
+     return nullptr;
+   }
+   RegMask* mask = NEW_RESOURCE_ARRAY(RegMask, cnt);
+   BasicType* sig_bt = NEW_RESOURCE_ARRAY(BasicType, cnt);
+   VMRegPair* vm_parm_regs = NEW_RESOURCE_ARRAY(VMRegPair, cnt);
+   for (uint i = 0; i < cnt; i++) {
+     sig_bt[i] = range->field_at(i+TypeFunc::Parms)->basic_type();
+   }
+ 
+   int regs = SharedRuntime::java_return_convention(sig_bt, vm_parm_regs, cnt);
+   if (regs <= 0) {
+     // We ran out of registers to store the IsInit information for a nullable inline type return.
+     // Since it is only set in the 'call_epilog', we can simply put it on the stack.
+     assert(tf->returns_inline_type_as_fields(), "should have been tested during graph construction");
+     // TODO 8284443 Can we teach the register allocator to reserve a stack slot instead?
+     // mask[--cnt] = STACK_ONLY_mask does not work (test with -XX:+StressGCM)
+     int slot = C->fixed_slots() - 2;
+     if (C->needs_stack_repair()) {
+       slot -= 2; // Account for stack increment value
+     }
+     mask[--cnt].Clear();
+     mask[cnt].Insert(OptoReg::stack2reg(slot));
+   }
+   for (uint i = 0; i < cnt; i++) {
+     mask[i].Clear();
+ 
+     OptoReg::Name reg1 = OptoReg::as_OptoReg(vm_parm_regs[i].first());
+     if (OptoReg::is_valid(reg1)) {
+       mask[i].Insert(reg1);
+     }
+     OptoReg::Name reg2 = OptoReg::as_OptoReg(vm_parm_regs[i].second());
+     if (OptoReg::is_valid(reg2)) {
+       mask[i].Insert(reg2);
+     }
+   }
+ 
+   return mask;
+ }
  
  //---------------------------match---------------------------------------------
  void Matcher::match( ) {
    if( MaxLabelRootDepth < 100 ) { // Too small?
      assert(false, "invalid MaxLabelRootDepth, increase it to 100 minimum");

@@ -203,33 +247,21 @@
  #ifdef _LP64
    // Pointers take 2 slots in 64-bit land
    _return_addr_mask.Insert(OptoReg::add(return_addr(),1));
  #endif
  
-   // Map a Java-signature return type into return register-value
-   // machine registers for 0, 1 and 2 returned values.
-   const TypeTuple *range = C->tf()->range();
-   if( range->cnt() > TypeFunc::Parms ) { // If not a void function
-     // Get ideal-register return type
-     uint ireg = range->field_at(TypeFunc::Parms)->ideal_reg();
-     // Get machine return register
-     uint sop = C->start()->Opcode();
-     OptoRegPair regs = return_value(ireg);
- 
-     // And mask for same
-     _return_value_mask = RegMask(regs.first());
-     if( OptoReg::is_valid(regs.second()) )
-       _return_value_mask.Insert(regs.second());
-   }
+   // Map Java-signature return types into return register-value
+   // machine registers.
+   _return_values_mask = return_values_mask(C->tf());
  
    // ---------------
    // Frame Layout
  
    // Need the method signature to determine the incoming argument types,
    // because the types determine which registers the incoming arguments are
    // in, and this affects the matched code.
-   const TypeTuple *domain = C->tf()->domain();
+   const TypeTuple *domain = C->tf()->domain_cc();
    uint             argcnt = domain->cnt() - TypeFunc::Parms;
    BasicType *sig_bt        = NEW_RESOURCE_ARRAY( BasicType, argcnt );
    VMRegPair *vm_parm_regs  = NEW_RESOURCE_ARRAY( VMRegPair, argcnt );
    _parm_regs               = NEW_RESOURCE_ARRAY( OptoRegPair, argcnt );
    _calling_convention_mask = NEW_RESOURCE_ARRAY( RegMask, argcnt );

@@ -548,10 +580,11 @@
    // Add in the incoming argument area
    OptoReg::Name init_in = OptoReg::add(_old_SP, C->out_preserve_stack_slots());
    for (i = init_in; i < _in_arg_limit; i = OptoReg::add(i,1)) {
      C->FIRST_STACK_mask().Insert(i);
    }
+ 
    // Add in all bits past the outgoing argument area
    guarantee(RegMask::can_represent_arg(OptoReg::add(_out_arg_limit,-1)),
              "must be able to represent all call arguments in reg mask");
    OptoReg::Name init = _out_arg_limit;
    for (i = init; RegMask::can_represent(i); i = OptoReg::add(i,1)) {

@@ -806,16 +839,15 @@
    assert( start, "Expect a start node" );
  
    // Input RegMask array shared by all Returns.
    // The type for doubles and longs has a count of 2, but
    // there is only 1 returned value
-   uint ret_edge_cnt = TypeFunc::Parms + ((C->tf()->range()->cnt() == TypeFunc::Parms) ? 0 : 1);
+   uint ret_edge_cnt = C->tf()->range_cc()->cnt();
    RegMask *ret_rms  = init_input_masks( ret_edge_cnt + soe_cnt, _return_addr_mask, c_frame_ptr_mask );
-   // Returns have 0 or 1 returned values depending on call signature.
-   // Return register is specified by return_value in the AD file.
-   if (ret_edge_cnt > TypeFunc::Parms)
-     ret_rms[TypeFunc::Parms+0] = _return_value_mask;
+   for (i = TypeFunc::Parms; i < ret_edge_cnt; i++) {
+     ret_rms[i] = _return_values_mask[i-TypeFunc::Parms];
+   }
  
    // Input RegMask array shared by all ForwardExceptions
    uint forw_exc_edge_cnt = TypeFunc::Parms;
    RegMask* forw_exc_rms  = init_input_masks( forw_exc_edge_cnt + soe_cnt, _return_addr_mask, c_frame_ptr_mask );
  

@@ -883,11 +915,11 @@
        default          : ShouldNotReachHere();
      }
    }
  
    // Next unused projection number from Start.
-   int proj_cnt = C->tf()->domain()->cnt();
+   int proj_cnt = C->tf()->domain_cc()->cnt();
  
    // Do all the save-on-entry registers.  Make projections from Start for
    // them, and give them a use at the exit points.  To the allocator, they
    // look like incoming register arguments.
    for( i = 0; i < _last_Mach_Reg; i++ ) {

@@ -1164,11 +1196,15 @@
                m->as_MachMemBar()->set_adr_type(n->adr_type());
              }
            } else {                  // Nothing the matcher cares about
              if (n->is_Proj() && n->in(0) != nullptr && n->in(0)->is_Multi()) {       // Projections?
                // Convert to machine-dependent projection
-               m = n->in(0)->as_Multi()->match( n->as_Proj(), this );
+               RegMask* mask = nullptr;
+               if (n->in(0)->is_Call() && n->in(0)->as_Call()->tf()->returns_inline_type_as_fields()) {
+                 mask = return_values_mask(n->in(0)->as_Call()->tf());
+               }
+               m = n->in(0)->as_Multi()->match(n->as_Proj(), this, mask);
                NOT_PRODUCT(record_new2old(m, n);)
                if (m->in(0) != nullptr) // m might be top
                  collect_null_checks(m, n);
              } else {                // Else just a regular 'ol guy
                m = n->clone();       // So just clone into new-space

@@ -1304,11 +1340,11 @@
    const TypeTuple *domain;
    ciMethod*        method = nullptr;
    bool             is_method_handle_invoke = false;  // for special kill effects
    if( sfpt->is_Call() ) {
      call = sfpt->as_Call();
-     domain = call->tf()->domain();
+     domain = call->tf()->domain_cc();
      cnt = domain->cnt();
  
      // Match just the call, nothing else
      MachNode *m = match_tree(call);
      if (C->failing())  return nullptr;

@@ -1383,17 +1419,20 @@
    if( call != nullptr && call->is_CallRuntime() )
      out_arg_limit_per_call = OptoReg::add(out_arg_limit_per_call,C->varargs_C_out_slots_killed());
  
  
    // Do the normal argument list (parameters) register masks
-   int argcnt = cnt - TypeFunc::Parms;
+   // Null entry point is a special cast where the target of the call
+   // is in a register.
+   int adj = (call != nullptr && call->entry_point() == nullptr) ? 1 : 0;
+   int argcnt = cnt - TypeFunc::Parms - adj;
    if( argcnt > 0 ) {          // Skip it all if we have no args
      BasicType *sig_bt  = NEW_RESOURCE_ARRAY( BasicType, argcnt );
      VMRegPair *parm_regs = NEW_RESOURCE_ARRAY( VMRegPair, argcnt );
      int i;
      for( i = 0; i < argcnt; i++ ) {
-       sig_bt[i] = domain->field_at(i+TypeFunc::Parms)->basic_type();
+       sig_bt[i] = domain->field_at(i+TypeFunc::Parms+adj)->basic_type();
      }
      // V-call to pick proper calling convention
      call->calling_convention( sig_bt, parm_regs, argcnt );
  
  #ifdef ASSERT

@@ -1430,11 +1469,11 @@
      // Return results now can have 2 bits returned.
      // Compute max over all outgoing arguments both per call-site
      // and over the entire method.
      for( i = 0; i < argcnt; i++ ) {
        // Address of incoming argument mask to fill in
-       RegMask *rm = &mcall->_in_rms[i+TypeFunc::Parms];
+       RegMask *rm = &mcall->_in_rms[i+TypeFunc::Parms+adj];
        VMReg first = parm_regs[i].first();
        VMReg second = parm_regs[i].second();
        if(!first->is_valid() &&
           !second->is_valid()) {
          continue;               // Avoid Halves

@@ -1451,19 +1490,21 @@
        // Grab first register, adjust stack slots and insert in mask.
        OptoReg::Name reg1 = warp_outgoing_stk_arg(first, begin_out_arg_area, out_arg_limit_per_call );
        if (C->failing()) {
          return nullptr;
        }
-       if (OptoReg::is_valid(reg1))
+       if (OptoReg::is_valid(reg1)) {
          rm->Insert( reg1 );
+       }
        // Grab second register (if any), adjust stack slots and insert in mask.
        OptoReg::Name reg2 = warp_outgoing_stk_arg(second, begin_out_arg_area, out_arg_limit_per_call );
        if (C->failing()) {
          return nullptr;
        }
-       if (OptoReg::is_valid(reg2))
+       if (OptoReg::is_valid(reg2)) {
          rm->Insert( reg2 );
+       }
      } // End of for all arguments
    }
  
    // Compute the max stack slot killed by any call.  These will not be
    // available for debug info, and will be used to adjust FIRST_STACK_mask

@@ -1475,11 +1516,11 @@
      // Kill the outgoing argument area, including any non-argument holes and
      // any legacy C-killed slots.  Use Fat-Projections to do the killing.
      // Since the max-per-method covers the max-per-call-site and debug info
      // is excluded on the max-per-method basis, debug info cannot land in
      // this killed area.
-     uint r_cnt = mcall->tf()->range()->cnt();
+     uint r_cnt = mcall->tf()->range_sig()->cnt();
      MachProjNode *proj = new MachProjNode( mcall, r_cnt+10000, RegMask::Empty, MachProjNode::fat_proj );
      if (!RegMask::can_represent_arg(OptoReg::Name(out_arg_limit_per_call-1))) {
        // Bailout. We do not have space to represent all arguments.
        C->record_method_not_compilable("unsupported outgoing calling sequence");
      } else {

@@ -1497,11 +1538,11 @@
      jvms->set_map(sfpt);
    }
  
    // Debug inputs begin just after the last incoming parameter
    assert((mcall == nullptr) || (mcall->jvms() == nullptr) ||
-          (mcall->jvms()->debug_start() + mcall->_jvmadj == mcall->tf()->domain()->cnt()), "");
+          (mcall->jvms()->debug_start() + mcall->_jvmadj == mcall->tf()->domain_cc()->cnt()), "");
  
    // Add additional edges.
    if (msfpt->mach_constant_base_node_input() != (uint)-1 && !msfpt->is_MachCallLeaf()) {
      // For these calls we can not add MachConstantBase in expand(), as the
      // ins are not complete then.

@@ -2185,11 +2226,11 @@
        bool mem_op = false;
        int mem_addr_idx = MemNode::Address;
        if (find_shared_visit(mstack, n, nop, mem_op, mem_addr_idx)) {
          continue;
        }
-       for (int i = n->req() - 1; i >= 0; --i) { // For my children
+       for (int i = n->len() - 1; i >= 0; --i) { // For my children
          Node* m = n->in(i); // Get ith input
          if (m == nullptr) {
            continue;  // Ignore nulls
          }
          if (clone_node(n, m, mstack)) {

@@ -2491,10 +2532,17 @@
        n->set_req(1, pair1);
        n->set_req(2, pair2);
        n->del_req(4);
        n->del_req(3);
        break;
+     }
+     case Op_ClearArray: {
+       Node* pair = new BinaryNode(n->in(2), n->in(3));
+       n->set_req(2, pair);
+       n->set_req(3, n->in(4));
+       n->del_req(4);
+       break;
      }
      case Op_VectorCmpMasked:
      case Op_CopySignD:
      case Op_SignumVF:
      case Op_SignumVD:
< prev index next >