< prev index next >

src/hotspot/share/opto/macroArrayCopy.cpp

Print this page
*** 21,10 ***
--- 21,11 ---
   * questions.
   *
   */
  
  #include "precompiled.hpp"
+ #include "ci/ciFlatArrayKlass.hpp"
  #include "gc/shared/barrierSet.hpp"
  #include "gc/shared/tlab_globals.hpp"
  #include "opto/arraycopynode.hpp"
  #include "oops/objArrayKlass.hpp"
  #include "opto/convertnode.hpp"

*** 137,14 ***
    *ctrl = if_fast;
  
    return if_slow;
  }
  
! inline Node* PhaseMacroExpand::generate_slow_guard(Node** ctrl, Node* test, RegionNode* region) {
    return generate_guard(ctrl, test, region, PROB_UNLIKELY_MAG(3));
  }
  
  void PhaseMacroExpand::generate_negative_guard(Node** ctrl, Node* index, RegionNode* region) {
    if ((*ctrl)->is_top())
      return;                // already stopped
    if (_igvn.type(index)->higher_equal(TypeInt::POS)) // [0,maxint]
      return;                // index is already adequately typed
--- 138,18 ---
    *ctrl = if_fast;
  
    return if_slow;
  }
  
! Node* PhaseMacroExpand::generate_slow_guard(Node** ctrl, Node* test, RegionNode* region) {
    return generate_guard(ctrl, test, region, PROB_UNLIKELY_MAG(3));
  }
  
+ inline Node* PhaseMacroExpand::generate_fair_guard(Node** ctrl, Node* test, RegionNode* region) {
+   return generate_guard(ctrl, test, region, PROB_FAIR);
+ }
+ 
  void PhaseMacroExpand::generate_negative_guard(Node** ctrl, Node* index, RegionNode* region) {
    if ((*ctrl)->is_top())
      return;                // already stopped
    if (_igvn.type(index)->higher_equal(TypeInt::POS)) // [0,maxint]
      return;                // index is already adequately typed

*** 281,10 ***
--- 286,53 ---
    Node* is_notp = generate_guard(ctrl, bol_le, nullptr, PROB_MIN);
  
    return is_notp;
  }
  
+ Node* PhaseMacroExpand::mark_word_test(Node** ctrl, Node* obj, MergeMemNode* mem, uintptr_t mask_val, RegionNode* region) {
+   // Load markword and check if obj is locked
+   Node* mark = make_load(nullptr, mem->memory_at(Compile::AliasIdxRaw), obj, oopDesc::mark_offset_in_bytes(), TypeX_X, TypeX_X->basic_type());
+   Node* locked_bit = MakeConX(markWord::unlocked_value);
+   locked_bit = transform_later(new AndXNode(locked_bit, mark));
+   Node* cmp = transform_later(new CmpXNode(locked_bit, MakeConX(0)));
+   Node* is_unlocked = transform_later(new BoolNode(cmp, BoolTest::ne));
+   IfNode* iff = transform_later(new IfNode(*ctrl, is_unlocked, PROB_MAX, COUNT_UNKNOWN))->as_If();
+   Node* locked_region = transform_later(new RegionNode(3));
+   Node* mark_phi = transform_later(new PhiNode(locked_region, TypeX_X));
+ 
+   // Unlocked: Use bits from mark word
+   locked_region->init_req(1, transform_later(new IfTrueNode(iff)));
+   mark_phi->init_req(1, mark);
+ 
+   // Locked: Load prototype header from klass
+   *ctrl = transform_later(new IfFalseNode(iff));
+   // Make loads control dependent to make sure they are only executed if array is locked
+   Node* klass_adr = basic_plus_adr(obj, oopDesc::klass_offset_in_bytes());
+   Node* klass = transform_later(LoadKlassNode::make(_igvn, *ctrl, C->immutable_memory(), klass_adr, TypeInstPtr::KLASS, TypeInstKlassPtr::OBJECT));
+   Node* proto_adr = basic_plus_adr(klass, in_bytes(Klass::prototype_header_offset()));
+   Node* proto = transform_later(LoadNode::make(_igvn, *ctrl, C->immutable_memory(), proto_adr, proto_adr->bottom_type()->is_ptr(), TypeX_X, TypeX_X->basic_type(), MemNode::unordered));
+ 
+   locked_region->init_req(2, *ctrl);
+   mark_phi->init_req(2, proto);
+   *ctrl = locked_region;
+ 
+   // Now check if mark word bits are set
+   Node* mask = MakeConX(mask_val);
+   Node* masked = transform_later(new AndXNode(mark_phi, mask));
+   cmp = transform_later(new CmpXNode(masked, mask));
+   Node* bol = transform_later(new BoolNode(cmp, BoolTest::eq));
+   return generate_fair_guard(ctrl, bol, region);
+ }
+ 
+ Node* PhaseMacroExpand::generate_flat_array_guard(Node** ctrl, Node* array, MergeMemNode* mem, RegionNode* region) {
+   return mark_word_test(ctrl, array, mem, markWord::flat_array_bit_in_place, region);
+ }
+ 
+ Node* PhaseMacroExpand::generate_null_free_array_guard(Node** ctrl, Node* array, MergeMemNode* mem, RegionNode* region) {
+   return mark_word_test(ctrl, array, mem, markWord::null_free_array_bit_in_place, region);
+ }
+ 
  void PhaseMacroExpand::finish_arraycopy_call(Node* call, Node** ctrl, MergeMemNode** mem, const TypePtr* adr_type) {
    transform_later(call);
  
    *ctrl = new ProjNode(call,TypeFunc::Control);
    transform_later(*ctrl);

*** 375,10 ***
--- 423,11 ---
                                             const TypePtr* adr_type,
                                             BasicType basic_elem_type,
                                             Node* src,  Node* src_offset,
                                             Node* dest, Node* dest_offset,
                                             Node* copy_length,
+                                            Node* dest_length,
                                             bool disjoint_bases,
                                             bool length_never_negative,
                                             RegionNode* slow_region) {
    if (slow_region == nullptr) {
      slow_region = new RegionNode(1);

*** 386,10 ***
--- 435,12 ---
    }
  
    Node* original_dest = dest;
    bool  dest_needs_zeroing   = false;
    bool  acopy_to_uninitialized = false;
+   Node* default_value = nullptr;
+   Node* raw_default_value = nullptr;
  
    // See if this is the initialization of a newly-allocated array.
    // If so, we will take responsibility here for initializing it to zero.
    // (Note:  Because tightly_coupled_allocation performs checks on the
    // out-edges of the dest, we need to avoid making derived pointers

*** 416,10 ***
--- 467,12 ---
        // From this point on, every exit path is responsible for
        // initializing any non-copied parts of the object to zero.
        // Also, if this flag is set we make sure that arraycopy interacts properly
        // with G1, eliding pre-barriers. See CR 6627983.
        dest_needs_zeroing = true;
+       default_value = alloc->in(AllocateNode::DefaultValue);
+       raw_default_value = alloc->in(AllocateNode::RawDefaultValue);
      } else {
        // dest_need_zeroing = false;
      }
    } else {
      // No zeroing elimination needed here.

*** 485,18 ***
      }
  
      // copy_length is 0.
      if (dest_needs_zeroing) {
        assert(!local_ctrl->is_top(), "no ctrl?");
-       Node* dest_length = alloc->in(AllocateNode::ALength);
        if (copy_length->eqv_uncast(dest_length)
            || _igvn.find_int_con(dest_length, 1) <= 0) {
          // There is no zeroing to do. No need for a secondary raw memory barrier.
        } else {
          // Clear the whole thing since there are no source elements to copy.
          generate_clear_array(local_ctrl, local_mem,
!                              adr_type, dest, basic_elem_type,
                               intcon(0), nullptr,
                               alloc->in(AllocateNode::AllocSize));
          // Use a secondary InitializeNode as raw memory barrier.
          // Currently it is needed only on this path since other
          // paths have stub or runtime calls as raw memory barriers.
--- 538,19 ---
      }
  
      // copy_length is 0.
      if (dest_needs_zeroing) {
        assert(!local_ctrl->is_top(), "no ctrl?");
        if (copy_length->eqv_uncast(dest_length)
            || _igvn.find_int_con(dest_length, 1) <= 0) {
          // There is no zeroing to do. No need for a secondary raw memory barrier.
        } else {
          // Clear the whole thing since there are no source elements to copy.
          generate_clear_array(local_ctrl, local_mem,
!                              adr_type, dest,
+                              default_value, raw_default_value,
+                              basic_elem_type,
                               intcon(0), nullptr,
                               alloc->in(AllocateNode::AllocSize));
          // Use a secondary InitializeNode as raw memory barrier.
          // Currently it is needed only on this path since other
          // paths have stub or runtime calls as raw memory barriers.

*** 523,17 ***
    if (!(*ctrl)->is_top() && dest_needs_zeroing) {
      // We have to initialize the *uncopied* part of the array to zero.
      // The copy destination is the slice dest[off..off+len].  The other slices
      // are dest_head = dest[0..off] and dest_tail = dest[off+len..dest.length].
      Node* dest_size   = alloc->in(AllocateNode::AllocSize);
-     Node* dest_length = alloc->in(AllocateNode::ALength);
      Node* dest_tail   = transform_later( new AddINode(dest_offset, copy_length));
  
      // If there is a head section that needs zeroing, do it now.
      if (_igvn.find_int_con(dest_offset, -1) != 0) {
        generate_clear_array(*ctrl, mem,
!                            adr_type, dest, basic_elem_type,
                             intcon(0), dest_offset,
                             nullptr);
      }
  
      // Next, perform a dynamic check on the tail length.
--- 577,18 ---
    if (!(*ctrl)->is_top() && dest_needs_zeroing) {
      // We have to initialize the *uncopied* part of the array to zero.
      // The copy destination is the slice dest[off..off+len].  The other slices
      // are dest_head = dest[0..off] and dest_tail = dest[off+len..dest.length].
      Node* dest_size   = alloc->in(AllocateNode::AllocSize);
      Node* dest_tail   = transform_later( new AddINode(dest_offset, copy_length));
  
      // If there is a head section that needs zeroing, do it now.
      if (_igvn.find_int_con(dest_offset, -1) != 0) {
        generate_clear_array(*ctrl, mem,
!                            adr_type, dest,
+                            default_value, raw_default_value,
+                            basic_elem_type,
                             intcon(0), dest_offset,
                             nullptr);
      }
  
      // Next, perform a dynamic check on the tail length.

*** 578,21 ***
      if (tail_ctl != nullptr) {
        Node* notail_ctl = (*ctrl)->is_top() ? nullptr : *ctrl;
        *ctrl = tail_ctl;
        if (notail_ctl == nullptr) {
          generate_clear_array(*ctrl, mem,
!                              adr_type, dest, basic_elem_type,
                               dest_tail, nullptr,
                               dest_size);
        } else {
          // Make a local merge.
          Node* done_ctl = transform_later(new RegionNode(3));
          Node* done_mem = transform_later(new PhiNode(done_ctl, Type::MEMORY, adr_type));
          done_ctl->init_req(1, notail_ctl);
          done_mem->init_req(1, mem->memory_at(alias_idx));
          generate_clear_array(*ctrl, mem,
!                              adr_type, dest, basic_elem_type,
                               dest_tail, nullptr,
                               dest_size);
          done_ctl->init_req(2, *ctrl);
          done_mem->init_req(2, mem->memory_at(alias_idx));
          *ctrl = done_ctl;
--- 633,25 ---
      if (tail_ctl != nullptr) {
        Node* notail_ctl = (*ctrl)->is_top() ? nullptr : *ctrl;
        *ctrl = tail_ctl;
        if (notail_ctl == nullptr) {
          generate_clear_array(*ctrl, mem,
!                              adr_type, dest,
+                              default_value, raw_default_value,
+                              basic_elem_type,
                               dest_tail, nullptr,
                               dest_size);
        } else {
          // Make a local merge.
          Node* done_ctl = transform_later(new RegionNode(3));
          Node* done_mem = transform_later(new PhiNode(done_ctl, Type::MEMORY, adr_type));
          done_ctl->init_req(1, notail_ctl);
          done_mem->init_req(1, mem->memory_at(alias_idx));
          generate_clear_array(*ctrl, mem,
!                              adr_type, dest,
+                              default_value, raw_default_value,
+                              basic_elem_type,
                               dest_tail, nullptr,
                               dest_size);
          done_ctl->init_req(2, *ctrl);
          done_mem->init_req(2, mem->memory_at(alias_idx));
          *ctrl = done_ctl;

*** 766,11 ***
      // Generate the slow path, if needed.
      local_mem->set_memory_at(alias_idx, slow_mem);
  
      if (dest_needs_zeroing) {
        generate_clear_array(local_ctrl, local_mem,
!                            adr_type, dest, basic_elem_type,
                             intcon(0), nullptr,
                             alloc->in(AllocateNode::AllocSize));
      }
  
      local_mem = generate_slow_arraycopy(ac,
--- 825,13 ---
      // Generate the slow path, if needed.
      local_mem->set_memory_at(alias_idx, slow_mem);
  
      if (dest_needs_zeroing) {
        generate_clear_array(local_ctrl, local_mem,
!                            adr_type, dest,
+                            default_value, raw_default_value,
+                            basic_elem_type,
                             intcon(0), nullptr,
                             alloc->in(AllocateNode::AllocSize));
      }
  
      local_mem = generate_slow_arraycopy(ac,

*** 828,15 ***
      assert((*ctrl)->is_Proj(), "MemBar control projection");
      assert((*ctrl)->in(0)->isa_MemBar(), "MemBar node");
      (*ctrl)->in(0)->isa_MemBar()->set_trailing_partial_array_copy();
    }
  
!   _igvn.replace_node(_callprojs.fallthrough_memproj, out_mem);
!   if (_callprojs.fallthrough_ioproj != nullptr) {
!     _igvn.replace_node(_callprojs.fallthrough_ioproj, *io);
    }
!   _igvn.replace_node(_callprojs.fallthrough_catchproj, *ctrl);
  
  #ifdef ASSERT
    const TypeOopPtr* dest_t = _igvn.type(dest)->is_oopptr();
    if (dest_t->is_known_instance() && !is_partial_array_copy) {
      ArrayCopyNode* ac = nullptr;
--- 889,15 ---
      assert((*ctrl)->is_Proj(), "MemBar control projection");
      assert((*ctrl)->in(0)->isa_MemBar(), "MemBar node");
      (*ctrl)->in(0)->isa_MemBar()->set_trailing_partial_array_copy();
    }
  
!   _igvn.replace_node(_callprojs->fallthrough_memproj, out_mem);
!   if (_callprojs->fallthrough_ioproj != nullptr) {
!     _igvn.replace_node(_callprojs->fallthrough_ioproj, *io);
    }
!   _igvn.replace_node(_callprojs->fallthrough_catchproj, *ctrl);
  
  #ifdef ASSERT
    const TypeOopPtr* dest_t = _igvn.type(dest)->is_oopptr();
    if (dest_t->is_known_instance() && !is_partial_array_copy) {
      ArrayCopyNode* ac = nullptr;

*** 872,10 ***
--- 933,12 ---
  // If dest_size is non-null, zeroing extends to the end of the object.
  // If slice_len is non-null, the slice_idx value must be a constant.
  void PhaseMacroExpand::generate_clear_array(Node* ctrl, MergeMemNode* merge_mem,
                                              const TypePtr* adr_type,
                                              Node* dest,
+                                             Node* val,
+                                             Node* raw_val,
                                              BasicType basic_elem_type,
                                              Node* slice_idx,
                                              Node* slice_len,
                                              Node* dest_size) {
    // one or the other but not both of slice_len and dest_size:

*** 910,16 ***
                         BytesPerLong);
    }
  
    if (start_con >= 0 && end_con >= 0) {
      // Constant start and end.  Simple.
!     mem = ClearArrayNode::clear_memory(ctrl, mem, dest,
                                         start_con, end_con, &_igvn);
    } else if (start_con >= 0 && dest_size != top()) {
      // Constant start, pre-rounded end after the tail of the array.
      Node* end = dest_size;
!     mem = ClearArrayNode::clear_memory(ctrl, mem, dest,
                                         start_con, end, &_igvn);
    } else if (start_con >= 0 && slice_len != top()) {
      // Constant start, non-constant end.  End needs rounding up.
      // End offset = round_up(abase + ((slice_idx_con + slice_len) << scale), 8)
      intptr_t end_base  = abase + (slice_idx_con << scale);
--- 973,16 ---
                         BytesPerLong);
    }
  
    if (start_con >= 0 && end_con >= 0) {
      // Constant start and end.  Simple.
!     mem = ClearArrayNode::clear_memory(ctrl, mem, dest, val, raw_val,
                                         start_con, end_con, &_igvn);
    } else if (start_con >= 0 && dest_size != top()) {
      // Constant start, pre-rounded end after the tail of the array.
      Node* end = dest_size;
!     mem = ClearArrayNode::clear_memory(ctrl, mem, dest, val, raw_val,
                                         start_con, end, &_igvn);
    } else if (start_con >= 0 && slice_len != top()) {
      // Constant start, non-constant end.  End needs rounding up.
      // End offset = round_up(abase + ((slice_idx_con + slice_len) << scale), 8)
      intptr_t end_base  = abase + (slice_idx_con << scale);

*** 928,11 ***
      if (scale != 0)
        end = transform_later(new LShiftXNode(end, intcon(scale) ));
      end_base += end_round;
      end = transform_later(new AddXNode(end, MakeConX(end_base)) );
      end = transform_later(new AndXNode(end, MakeConX(~end_round)) );
!     mem = ClearArrayNode::clear_memory(ctrl, mem, dest,
                                         start_con, end, &_igvn);
    } else if (start_con < 0 && dest_size != top()) {
      // Non-constant start, pre-rounded end after the tail of the array.
      // This is almost certainly a "round-to-end" operation.
      Node* start = slice_idx;
--- 991,11 ---
      if (scale != 0)
        end = transform_later(new LShiftXNode(end, intcon(scale) ));
      end_base += end_round;
      end = transform_later(new AddXNode(end, MakeConX(end_base)) );
      end = transform_later(new AndXNode(end, MakeConX(~end_round)) );
!     mem = ClearArrayNode::clear_memory(ctrl, mem, dest, val, raw_val,
                                         start_con, end, &_igvn);
    } else if (start_con < 0 && dest_size != top()) {
      // Non-constant start, pre-rounded end after the tail of the array.
      // This is almost certainly a "round-to-end" operation.
      Node* start = slice_idx;

*** 957,16 ***
        start = transform_later(new AndXNode(start, MakeConX(~to_clear)) );
        if (bump_bit != 0) {
          // Store a zero to the immediately preceding jint:
          Node* x1 = transform_later(new AddXNode(start, MakeConX(-bump_bit)) );
          Node* p1 = basic_plus_adr(dest, x1);
!         mem = StoreNode::make(_igvn, ctrl, mem, p1, adr_type, intcon(0), T_INT, MemNode::unordered);
          mem = transform_later(mem);
        }
      }
      Node* end = dest_size; // pre-rounded
!     mem = ClearArrayNode::clear_memory(ctrl, mem, dest,
                                         start, end, &_igvn);
    } else {
      // Non-constant start, unrounded non-constant end.
      // (Nobody zeroes a random midsection of an array using this routine.)
      ShouldNotReachHere();       // fix caller
--- 1020,22 ---
        start = transform_later(new AndXNode(start, MakeConX(~to_clear)) );
        if (bump_bit != 0) {
          // Store a zero to the immediately preceding jint:
          Node* x1 = transform_later(new AddXNode(start, MakeConX(-bump_bit)) );
          Node* p1 = basic_plus_adr(dest, x1);
!         if (val == nullptr) {
+           assert(raw_val == nullptr, "val may not be null");
+           mem = StoreNode::make(_igvn, ctrl, mem, p1, adr_type, intcon(0), T_INT, MemNode::unordered);
+         } else {
+           assert(_igvn.type(val)->isa_narrowoop(), "should be narrow oop");
+           mem = new StoreNNode(ctrl, mem, p1, adr_type, val, MemNode::unordered);
+         }
          mem = transform_later(mem);
        }
      }
      Node* end = dest_size; // pre-rounded
!     mem = ClearArrayNode::clear_memory(ctrl, mem, dest, raw_val,
                                         start, end, &_igvn);
    } else {
      // Non-constant start, unrounded non-constant end.
      // (Nobody zeroes a random midsection of an array using this routine.)
      ShouldNotReachHere();       // fix caller

*** 1078,15 ***
  
    call->set_cnt(PROB_UNLIKELY_MAG(4));  // Same effect as RC_UNCOMMON.
    _igvn.replace_node(ac, call);
    transform_later(call);
  
!   call->extract_projections(&_callprojs, false /*separate_io_proj*/, false /*do_asserts*/);
!   *ctrl = _callprojs.fallthrough_catchproj->clone();
    transform_later(*ctrl);
  
!   Node* m = _callprojs.fallthrough_memproj->clone();
    transform_later(m);
  
    uint alias_idx = C->get_alias_index(adr_type);
    MergeMemNode* out_mem;
    if (alias_idx != Compile::AliasIdxBot) {
--- 1147,15 ---
  
    call->set_cnt(PROB_UNLIKELY_MAG(4));  // Same effect as RC_UNCOMMON.
    _igvn.replace_node(ac, call);
    transform_later(call);
  
!   _callprojs = call->extract_projections(false /*separate_io_proj*/, false /*do_asserts*/);
!   *ctrl = _callprojs->fallthrough_catchproj->clone();
    transform_later(*ctrl);
  
!   Node* m = _callprojs->fallthrough_memproj->clone();
    transform_later(m);
  
    uint alias_idx = C->get_alias_index(adr_type);
    MergeMemNode* out_mem;
    if (alias_idx != Compile::AliasIdxBot) {

*** 1096,13 ***
      out_mem = MergeMemNode::make(m);
    }
    transform_later(out_mem);
  
    // When src is negative and arraycopy is before an infinite loop,_callprojs.fallthrough_ioproj
!   // could be null. Skip clone and update null fallthrough_ioproj.
!   if (_callprojs.fallthrough_ioproj != nullptr) {
!     *io = _callprojs.fallthrough_ioproj->clone();
      transform_later(*io);
    } else {
      *io = nullptr;
    }
  
--- 1165,13 ---
      out_mem = MergeMemNode::make(m);
    }
    transform_later(out_mem);
  
    // When src is negative and arraycopy is before an infinite loop,_callprojs.fallthrough_ioproj
!   // could be nullptr. Skip clone and update nullptr fallthrough_ioproj.
!   if (_callprojs->fallthrough_ioproj != nullptr) {
!     *io = _callprojs->fallthrough_ioproj->clone();
      transform_later(*io);
    } else {
      *io = nullptr;
    }
  

*** 1230,10 ***
--- 1299,46 ---
      return true;
    }
    return false;
  }
  
+ const TypePtr* PhaseMacroExpand::adjust_for_flat_array(const TypeAryPtr* top_dest, Node*& src_offset,
+                                                        Node*& dest_offset, Node*& length, BasicType& dest_elem,
+                                                        Node*& dest_length) {
+ #ifdef ASSERT
+   BarrierSetC2* bs = BarrierSet::barrier_set()->barrier_set_c2();
+   bool needs_barriers = top_dest->elem()->inline_klass()->contains_oops() &&
+     bs->array_copy_requires_gc_barriers(dest_length != nullptr, T_OBJECT, false, false, BarrierSetC2::Optimization);
+   assert(!needs_barriers || StressReflectiveCode, "Flat arracopy would require GC barriers");
+ #endif
+   int elem_size = top_dest->flat_elem_size();
+   if (elem_size >= 8) {
+     if (elem_size > 8) {
+       // treat as array of long but scale length, src offset and dest offset
+       assert((elem_size % 8) == 0, "not a power of 2?");
+       int factor = elem_size / 8;
+       length = transform_later(new MulINode(length, intcon(factor)));
+       src_offset = transform_later(new MulINode(src_offset, intcon(factor)));
+       dest_offset = transform_later(new MulINode(dest_offset, intcon(factor)));
+       if (dest_length != nullptr) {
+         dest_length = transform_later(new MulINode(dest_length, intcon(factor)));
+       }
+       elem_size = 8;
+     }
+     dest_elem = T_LONG;
+   } else if (elem_size == 4) {
+     dest_elem = T_INT;
+   } else if (elem_size == 2) {
+     dest_elem = T_CHAR;
+   } else if (elem_size == 1) {
+     dest_elem = T_BYTE;
+   } else {
+     ShouldNotReachHere();
+   }
+   return TypeRawPtr::BOTTOM;
+ }
+ 
  #undef XTOP
  
  void PhaseMacroExpand::expand_arraycopy_node(ArrayCopyNode *ac) {
    Node* ctrl = ac->in(TypeFunc::Control);
    Node* io = ac->in(TypeFunc::I_O);

*** 1247,27 ***
    if (ac->is_clonebasic()) {
      BarrierSetC2* bs = BarrierSet::barrier_set()->barrier_set_c2();
      bs->clone_at_expansion(this, ac);
      return;
    } else if (ac->is_copyof() || ac->is_copyofrange() || ac->is_clone_oop_array()) {
!     Node* mem = ac->in(TypeFunc::Memory);
!     merge_mem = MergeMemNode::make(mem);
!     transform_later(merge_mem);
  
      AllocateArrayNode* alloc = nullptr;
      if (ac->is_alloc_tightly_coupled()) {
        alloc = AllocateArrayNode::Ideal_array_allocation(dest);
        assert(alloc != nullptr, "expect alloc");
      }
  
!     const TypePtr* adr_type = _igvn.type(dest)->is_oopptr()->add_offset(Type::OffsetBot);
!     if (ac->_dest_type != TypeOopPtr::BOTTOM) {
!       adr_type = ac->_dest_type->add_offset(Type::OffsetBot)->is_ptr();
      }
      generate_arraycopy(ac, alloc, &ctrl, merge_mem, &io,
!                        adr_type, T_OBJECT,
                         src, src_offset, dest, dest_offset, length,
                         true, ac->has_negative_length_guard());
  
      return;
    }
  
--- 1352,57 ---
    if (ac->is_clonebasic()) {
      BarrierSetC2* bs = BarrierSet::barrier_set()->barrier_set_c2();
      bs->clone_at_expansion(this, ac);
      return;
    } else if (ac->is_copyof() || ac->is_copyofrange() || ac->is_clone_oop_array()) {
!     const Type* src_type = _igvn.type(src);
!     const Type* dest_type = _igvn.type(dest);
!     const TypeAryPtr* top_src = src_type->isa_aryptr();
+     const TypeAryPtr* top_dest = dest_type->isa_aryptr();
+     BasicType dest_elem = T_OBJECT;
+     if (top_dest != nullptr && top_dest->elem() != Type::BOTTOM) {
+       dest_elem = top_dest->elem()->array_element_basic_type();
+     }
+     if (is_reference_type(dest_elem, true)) dest_elem = T_OBJECT;
+ 
+     if (top_src != nullptr && top_src->is_flat()) {
+       // If src is flat, dest is guaranteed to be flat as well
+       top_dest = top_src;
+     }
  
      AllocateArrayNode* alloc = nullptr;
+     Node* dest_length = nullptr;
      if (ac->is_alloc_tightly_coupled()) {
        alloc = AllocateArrayNode::Ideal_array_allocation(dest);
        assert(alloc != nullptr, "expect alloc");
+       dest_length = alloc->in(AllocateNode::ALength);
      }
  
!     Node* mem = ac->in(TypeFunc::Memory);
!     const TypePtr* adr_type = nullptr;
!     if (top_dest->is_flat()) {
+       assert(dest_length != nullptr || StressReflectiveCode, "must be tightly coupled");
+       // Copy to a flat array modifies multiple memory slices. Conservatively insert a barrier
+       // on all slices to prevent writes into the source from floating below the arraycopy.
+       insert_mem_bar(&ctrl, &mem, Op_MemBarCPUOrder);
+       adr_type = adjust_for_flat_array(top_dest, src_offset, dest_offset, length, dest_elem, dest_length);
+     } else {
+       adr_type = dest_type->is_oopptr()->add_offset(Type::OffsetBot);
+       if (ac->_dest_type != TypeOopPtr::BOTTOM) {
+         adr_type = ac->_dest_type->add_offset(Type::OffsetBot)->is_ptr();
+       }
+       if (ac->_src_type != ac->_dest_type) {
+         adr_type = TypeRawPtr::BOTTOM;
+       }
      }
+     merge_mem = MergeMemNode::make(mem);
+     transform_later(merge_mem);
+ 
      generate_arraycopy(ac, alloc, &ctrl, merge_mem, &io,
!                        adr_type, dest_elem,
                         src, src_offset, dest, dest_offset, length,
+                        dest_length,
                         true, ac->has_negative_length_guard());
  
      return;
    }
  

*** 1299,13 ***
      dest_elem = top_dest->elem()->array_element_basic_type();
    }
    if (is_reference_type(src_elem, true)) src_elem = T_OBJECT;
    if (is_reference_type(dest_elem, true)) dest_elem = T_OBJECT;
  
!   if (ac->is_arraycopy_validated() &&
-       dest_elem != T_CONFLICT &&
-       src_elem == T_CONFLICT) {
      src_elem = dest_elem;
    }
  
    if (src_elem == T_CONFLICT || dest_elem == T_CONFLICT) {
      // Conservatively insert a memory barrier on all memory slices.
--- 1434,11 ---
      dest_elem = top_dest->elem()->array_element_basic_type();
    }
    if (is_reference_type(src_elem, true)) src_elem = T_OBJECT;
    if (is_reference_type(dest_elem, true)) dest_elem = T_OBJECT;
  
!   if (ac->is_arraycopy_validated() && dest_elem != T_CONFLICT && src_elem == T_CONFLICT) {
      src_elem = dest_elem;
    }
  
    if (src_elem == T_CONFLICT || dest_elem == T_CONFLICT) {
      // Conservatively insert a memory barrier on all memory slices.

*** 1320,33 ***
  
      // Call StubRoutines::generic_arraycopy stub.
      Node* mem = generate_arraycopy(ac, nullptr, &ctrl, merge_mem, &io,
                                     TypeRawPtr::BOTTOM, T_CONFLICT,
                                     src, src_offset, dest, dest_offset, length,
                                     // If a  negative length guard was generated for the ArrayCopyNode,
                                     // the length of the array can never be negative.
                                     false, ac->has_negative_length_guard());
      return;
    }
  
    assert(!ac->is_arraycopy_validated() || (src_elem == dest_elem && dest_elem != T_VOID), "validated but different basic types");
  
    // (2) src and dest arrays must have elements of the same BasicType
    // Figure out the size and type of the elements we will be copying.
!   if (src_elem != dest_elem || dest_elem == T_VOID) {
      // The component types are not the same or are not recognized.  Punt.
      // (But, avoid the native method wrapper to JVM_ArrayCopy.)
      {
        Node* mem = ac->in(TypeFunc::Memory);
        merge_mem = generate_slow_arraycopy(ac, &ctrl, mem, &io, TypePtr::BOTTOM, src, src_offset, dest, dest_offset, length, false);
      }
  
!     _igvn.replace_node(_callprojs.fallthrough_memproj, merge_mem);
!     if (_callprojs.fallthrough_ioproj != nullptr) {
!       _igvn.replace_node(_callprojs.fallthrough_ioproj, io);
      }
!     _igvn.replace_node(_callprojs.fallthrough_catchproj, ctrl);
      return;
    }
  
    //---------------------------------------------------------------------------
    // We will make a fast path for this call to arraycopy.
--- 1453,41 ---
  
      // Call StubRoutines::generic_arraycopy stub.
      Node* mem = generate_arraycopy(ac, nullptr, &ctrl, merge_mem, &io,
                                     TypeRawPtr::BOTTOM, T_CONFLICT,
                                     src, src_offset, dest, dest_offset, length,
+                                    nullptr,
                                     // If a  negative length guard was generated for the ArrayCopyNode,
                                     // the length of the array can never be negative.
                                     false, ac->has_negative_length_guard());
      return;
    }
  
    assert(!ac->is_arraycopy_validated() || (src_elem == dest_elem && dest_elem != T_VOID), "validated but different basic types");
  
    // (2) src and dest arrays must have elements of the same BasicType
    // Figure out the size and type of the elements we will be copying.
!   //
+   // We have no stub to copy flat inline type arrays with oop
+   // fields if we need to emit write barriers.
+   //
+   BarrierSetC2* bs = BarrierSet::barrier_set()->barrier_set_c2();
+   if (src_elem != dest_elem || top_src->is_flat() != top_dest->is_flat() || dest_elem == T_VOID ||
+       (top_src->is_flat() && top_dest->elem()->inline_klass()->contains_oops() &&
+        bs->array_copy_requires_gc_barriers(alloc != nullptr, T_OBJECT, false, false, BarrierSetC2::Optimization))) {
      // The component types are not the same or are not recognized.  Punt.
      // (But, avoid the native method wrapper to JVM_ArrayCopy.)
      {
        Node* mem = ac->in(TypeFunc::Memory);
        merge_mem = generate_slow_arraycopy(ac, &ctrl, mem, &io, TypePtr::BOTTOM, src, src_offset, dest, dest_offset, length, false);
      }
  
!     _igvn.replace_node(_callprojs->fallthrough_memproj, merge_mem);
!     if (_callprojs->fallthrough_ioproj != nullptr) {
!       _igvn.replace_node(_callprojs->fallthrough_ioproj, io);
      }
!     _igvn.replace_node(_callprojs->fallthrough_catchproj, ctrl);
      return;
    }
  
    //---------------------------------------------------------------------------
    // We will make a fast path for this call to arraycopy.

*** 1359,15 ***
    // (6) length must not be negative.
    // (7) src_offset + length must not exceed length of src.
    // (8) dest_offset + length must not exceed length of dest.
    // (9) each element of an oop array must be assignable
  
!   {
!     Node* mem = ac->in(TypeFunc::Memory);
!     merge_mem = MergeMemNode::make(mem);
!     transform_later(merge_mem);
    }
  
    RegionNode* slow_region = new RegionNode(1);
    transform_later(slow_region);
  
    if (!ac->is_arraycopy_validated()) {
--- 1500,18 ---
    // (6) length must not be negative.
    // (7) src_offset + length must not exceed length of src.
    // (8) dest_offset + length must not exceed length of dest.
    // (9) each element of an oop array must be assignable
  
!   Node* mem = ac->in(TypeFunc::Memory);
!   if (top_dest->is_flat()) {
!     // Copy to a flat array modifies multiple memory slices. Conservatively insert a barrier
!     // on all slices to prevent writes into the source from floating below the arraycopy.
+     insert_mem_bar(&ctrl, &mem, Op_MemBarCPUOrder);
    }
+   merge_mem = MergeMemNode::make(mem);
+   transform_later(merge_mem);
  
    RegionNode* slow_region = new RegionNode(1);
    transform_later(slow_region);
  
    if (!ac->is_arraycopy_validated()) {

*** 1404,21 ***
                           alen,
                           slow_region);
  
      // (9) each element of an oop array must be assignable
      // The generate_arraycopy subroutine checks this.
    }
    // This is where the memory effects are placed:
    const TypePtr* adr_type = nullptr;
!   if (ac->_dest_type != TypeOopPtr::BOTTOM) {
      adr_type = ac->_dest_type->add_offset(Type::OffsetBot)->is_ptr();
    } else {
      adr_type = TypeAryPtr::get_array_body_type(dest_elem);
    }
  
    generate_arraycopy(ac, alloc, &ctrl, merge_mem, &io,
                       adr_type, dest_elem,
                       src, src_offset, dest, dest_offset, length,
                       // If a  negative length guard was generated for the ArrayCopyNode,
                       // the length of the array can never be negative.
!                      false, ac->has_negative_length_guard(), slow_region);
  }
--- 1548,43 ---
                           alen,
                           slow_region);
  
      // (9) each element of an oop array must be assignable
      // The generate_arraycopy subroutine checks this.
+ 
+     // Handle inline type arrays
+     if (!top_src->is_flat()) {
+       if (UseFlatArray && !top_src->is_not_flat()) {
+         // Src might be flat and dest might not be flat. Go to the slow path if src is flat.
+         generate_flat_array_guard(&ctrl, src, merge_mem, slow_region);
+       }
+       if (EnableValhalla) {
+         // No validation. The subtype check emitted at macro expansion time will not go to the slow
+         // path but call checkcast_arraycopy which can not handle flat/null-free inline type arrays.
+         generate_null_free_array_guard(&ctrl, dest, merge_mem, slow_region);
+       }
+     } else {
+       assert(top_dest->is_flat(), "dest array must be flat");
+     }
    }
+ 
    // This is where the memory effects are placed:
    const TypePtr* adr_type = nullptr;
!   Node* dest_length = (alloc != nullptr) ? alloc->in(AllocateNode::ALength) : nullptr;
+ 
+   if (top_dest->is_flat()) {
+     adr_type = adjust_for_flat_array(top_dest, src_offset, dest_offset, length, dest_elem, dest_length);
+   } else if (ac->_dest_type != TypeOopPtr::BOTTOM) {
      adr_type = ac->_dest_type->add_offset(Type::OffsetBot)->is_ptr();
    } else {
      adr_type = TypeAryPtr::get_array_body_type(dest_elem);
    }
  
    generate_arraycopy(ac, alloc, &ctrl, merge_mem, &io,
                       adr_type, dest_elem,
                       src, src_offset, dest, dest_offset, length,
+                      dest_length,
                       // If a  negative length guard was generated for the ArrayCopyNode,
                       // the length of the array can never be negative.
!                      false, ac->has_negative_length_guard(),
+                      slow_region);
  }
< prev index next >