1 /*
   2  * Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved.
   3  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
   4  *
   5  * This code is free software; you can redistribute it and/or modify it
   6  * under the terms of the GNU General Public License version 2 only, as
   7  * published by the Free Software Foundation.
   8  *
   9  * This code is distributed in the hope that it will be useful, but WITHOUT
  10  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
  11  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
  12  * version 2 for more details (a copy is included in the LICENSE file that
  13  * accompanied this code).
  14  *
  15  * You should have received a copy of the GNU General Public License version
  16  * 2 along with this work; if not, write to the Free Software Foundation,
  17  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
  18  *
  19  * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
  20  * or visit www.oracle.com if you need additional information or have any
  21  * questions.
  22  *
  23  */
  24 
  25 #include "precompiled.hpp"
  26 #include "ci/ciSymbols.hpp"
  27 #include "classfile/vmSymbols.hpp"
  28 #include "opto/library_call.hpp"
  29 #include "opto/runtime.hpp"
  30 #include "opto/vectornode.hpp"
  31 #include "prims/vectorSupport.hpp"
  32 #include "runtime/stubRoutines.hpp"
  33 
  34 #ifdef ASSERT
  35 static bool is_vector(ciKlass* klass) {
  36   return klass->is_subclass_of(ciEnv::current()->vector_VectorPayload_klass());
  37 }
  38 
  39 static bool check_vbox(const TypeInstPtr* vbox_type) {
  40   assert(vbox_type->klass_is_exact(), "");
  41 
  42   ciInstanceKlass* ik = vbox_type->instance_klass();
  43   assert(is_vector(ik), "not a vector");
  44 
  45   ciField* fd1 = ik->get_field_by_name(ciSymbols::ETYPE_name(), ciSymbols::class_signature(), /* is_static */ true);
  46   assert(fd1 != nullptr, "element type info is missing");
  47 
  48   ciConstant val1 = fd1->constant_value();
  49   BasicType elem_bt = val1.as_object()->as_instance()->java_mirror_type()->basic_type();
  50   assert(is_java_primitive(elem_bt), "element type info is missing");
  51 
  52   ciField* fd2 = ik->get_field_by_name(ciSymbols::VLENGTH_name(), ciSymbols::int_signature(), /* is_static */ true);
  53   assert(fd2 != nullptr, "vector length info is missing");
  54 
  55   ciConstant val2 = fd2->constant_value();
  56   assert(val2.as_int() > 0, "vector length info is missing");
  57 
  58   return true;
  59 }
  60 #endif
  61 
  62 #define log_if_needed(...)        \
  63   if (C->print_intrinsics()) {    \
  64     tty->print_cr(__VA_ARGS__);   \
  65   }
  66 
  67 #ifndef PRODUCT
  68 #define non_product_log_if_needed(...) log_if_needed(__VA_ARGS__)
  69 #else
  70 #define non_product_log_if_needed(...)
  71 #endif
  72 
  73 static bool is_vector_mask(ciKlass* klass) {
  74   return klass->is_subclass_of(ciEnv::current()->vector_VectorMask_klass());
  75 }
  76 
  77 static bool is_vector_shuffle(ciKlass* klass) {
  78   return klass->is_subclass_of(ciEnv::current()->vector_VectorShuffle_klass());
  79 }
  80 
  81 bool LibraryCallKit::arch_supports_vector_rotate(int opc, int num_elem, BasicType elem_bt,
  82                                                  VectorMaskUseType mask_use_type, bool has_scalar_args) {
  83   bool is_supported = true;
  84 
  85   // has_scalar_args flag is true only for non-constant scalar shift count,
  86   // since in this case shift needs to be broadcasted.
  87   if (!Matcher::match_rule_supported_vector(opc, num_elem, elem_bt) ||
  88        (has_scalar_args && !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed))) {
  89     is_supported = false;
  90   }
  91 
  92   if (is_supported) {
  93     // Check if mask unboxing is supported, this is a two step process which first loads the contents
  94     // of boolean array into vector followed by either lane expansion to match the lane size of masked
  95     // vector operation or populate the predicate register.
  96     if ((mask_use_type & VecMaskUseLoad) != 0) {
  97       if (!Matcher::match_rule_supported_vector(Op_VectorLoadMask, num_elem, elem_bt) ||
  98           !Matcher::match_rule_supported_vector(Op_LoadVector, num_elem, T_BOOLEAN)) {
  99         non_product_log_if_needed("  ** Rejected vector mask loading (%s,%s,%d) because architecture does not support it",
 100                                   NodeClassNames[Op_VectorLoadMask], type2name(elem_bt), num_elem);
 101         return false;
 102       }
 103     }
 104 
 105     if ((mask_use_type & VecMaskUsePred) != 0) {
 106       if (!Matcher::has_predicated_vectors() ||
 107           !Matcher::match_rule_supported_vector_masked(opc, num_elem, elem_bt)) {
 108         non_product_log_if_needed("Rejected vector mask predicate using (%s,%s,%d) because architecture does not support it",
 109                                   NodeClassNames[opc], type2name(elem_bt), num_elem);
 110         return false;
 111       }
 112     }
 113   }
 114 
 115   int lshiftopc, rshiftopc;
 116   switch(elem_bt) {
 117     case T_BYTE:
 118       lshiftopc = Op_LShiftI;
 119       rshiftopc = Op_URShiftB;
 120       break;
 121     case T_SHORT:
 122       lshiftopc = Op_LShiftI;
 123       rshiftopc = Op_URShiftS;
 124       break;
 125     case T_INT:
 126       lshiftopc = Op_LShiftI;
 127       rshiftopc = Op_URShiftI;
 128       break;
 129     case T_LONG:
 130       lshiftopc = Op_LShiftL;
 131       rshiftopc = Op_URShiftL;
 132       break;
 133     default: fatal("Unexpected type: %s", type2name(elem_bt));
 134   }
 135   int lshiftvopc = VectorNode::opcode(lshiftopc, elem_bt);
 136   int rshiftvopc = VectorNode::opcode(rshiftopc, elem_bt);
 137   if (!is_supported &&
 138       arch_supports_vector(lshiftvopc, num_elem, elem_bt, VecMaskNotUsed, has_scalar_args) &&
 139       arch_supports_vector(rshiftvopc, num_elem, elem_bt, VecMaskNotUsed, has_scalar_args) &&
 140       arch_supports_vector(Op_OrV, num_elem, elem_bt, VecMaskNotUsed)) {
 141     is_supported = true;
 142   }
 143   return is_supported;
 144 }
 145 
 146 Node* GraphKit::box_vector(Node* vector, const TypeInstPtr* vbox_type, BasicType elem_bt, int num_elem, bool deoptimize_on_exception) {
 147   assert(EnableVectorSupport, "");
 148 
 149   PreserveReexecuteState preexecs(this);
 150   jvms()->set_should_reexecute(true);
 151 
 152   VectorBoxAllocateNode* alloc = new VectorBoxAllocateNode(C, vbox_type);
 153   set_edges_for_java_call(alloc, /*must_throw=*/false, /*separate_io_proj=*/true);
 154   make_slow_call_ex(alloc, env()->Throwable_klass(), /*separate_io_proj=*/true, deoptimize_on_exception);
 155   set_i_o(gvn().transform( new ProjNode(alloc, TypeFunc::I_O) ));
 156   set_all_memory(gvn().transform( new ProjNode(alloc, TypeFunc::Memory) ));
 157   Node* ret = gvn().transform(new ProjNode(alloc, TypeFunc::Parms));
 158 
 159   assert(check_vbox(vbox_type), "");
 160   const TypeVect* vt = TypeVect::make(elem_bt, num_elem, is_vector_mask(vbox_type->instance_klass()));
 161   VectorBoxNode* vbox = new VectorBoxNode(C, ret, vector, vbox_type, vt);
 162   return gvn().transform(vbox);
 163 }
 164 
 165 Node* GraphKit::unbox_vector(Node* v, const TypeInstPtr* vbox_type, BasicType elem_bt, int num_elem, bool shuffle_to_vector) {
 166   assert(EnableVectorSupport, "");
 167   const TypeInstPtr* vbox_type_v = gvn().type(v)->isa_instptr();
 168   if (vbox_type_v == nullptr || vbox_type->instance_klass() != vbox_type_v->instance_klass()) {
 169     return nullptr; // arguments don't agree on vector shapes
 170   }
 171   if (vbox_type_v->maybe_null()) {
 172     return nullptr; // no nulls are allowed
 173   }
 174   assert(check_vbox(vbox_type), "");
 175   const TypeVect* vt = TypeVect::make(elem_bt, num_elem, is_vector_mask(vbox_type->instance_klass()));
 176   Node* unbox = gvn().transform(new VectorUnboxNode(C, vt, v, merged_memory(), shuffle_to_vector));
 177   return unbox;
 178 }
 179 
 180 Node* GraphKit::vector_shift_count(Node* cnt, int shift_op, BasicType bt, int num_elem) {
 181   assert(bt == T_INT || bt == T_LONG || bt == T_SHORT || bt == T_BYTE, "byte, short, long and int are supported");
 182   juint mask = (type2aelembytes(bt) * BitsPerByte - 1);
 183   Node* nmask = gvn().transform(ConNode::make(TypeInt::make(mask)));
 184   Node* mcnt = gvn().transform(new AndINode(cnt, nmask));
 185   return gvn().transform(VectorNode::shift_count(shift_op, mcnt, num_elem, bt));
 186 }
 187 
 188 bool LibraryCallKit::arch_supports_vector(int sopc, int num_elem, BasicType type, VectorMaskUseType mask_use_type, bool has_scalar_args) {
 189   // Check that the operation is valid.
 190   if (sopc <= 0) {
 191     non_product_log_if_needed("  ** Rejected intrinsification because no valid vector op could be extracted");
 192     return false;
 193   }
 194 
 195   if (VectorNode::is_vector_rotate(sopc)) {
 196     if(!arch_supports_vector_rotate(sopc, num_elem, type, mask_use_type, has_scalar_args)) {
 197       non_product_log_if_needed("  ** Rejected vector op (%s,%s,%d) because architecture does not support variable vector shifts",
 198                                 NodeClassNames[sopc], type2name(type), num_elem);
 199       return false;
 200     }
 201   } else if (VectorNode::is_vector_integral_negate(sopc)) {
 202     if (!VectorNode::is_vector_integral_negate_supported(sopc, num_elem, type, false)) {
 203       non_product_log_if_needed("  ** Rejected vector op (%s,%s,%d) because architecture does not support integral vector negate",
 204                                 NodeClassNames[sopc], type2name(type), num_elem);
 205       return false;
 206     }
 207   } else {
 208     // Check that architecture supports this op-size-type combination.
 209     if (!Matcher::match_rule_supported_vector(sopc, num_elem, type)) {
 210       non_product_log_if_needed("  ** Rejected vector op (%s,%s,%d) because architecture does not support it",
 211                                 NodeClassNames[sopc], type2name(type), num_elem);
 212       return false;
 213     } else {
 214       assert(Matcher::match_rule_supported(sopc), "must be supported");
 215     }
 216   }
 217 
 218   if (num_elem == 1) {
 219     if (mask_use_type != VecMaskNotUsed) {
 220       non_product_log_if_needed("  ** Rejected vector mask op (%s,%s,%d) because architecture does not support it",
 221                                 NodeClassNames[sopc], type2name(type), num_elem);
 222       return false;
 223     }
 224 
 225     if (sopc != 0) {
 226       if (sopc != Op_LoadVector && sopc != Op_StoreVector) {
 227         non_product_log_if_needed("  ** Not a svml call or load/store vector op (%s,%s,%d)",
 228                                   NodeClassNames[sopc], type2name(type), num_elem);
 229         return false;
 230       }
 231     }
 232   }
 233 
 234   if (!has_scalar_args && VectorNode::is_vector_shift(sopc) &&
 235       Matcher::supports_vector_variable_shifts() == false) {
 236     log_if_needed("  ** Rejected vector op (%s,%s,%d) because architecture does not support variable vector shifts",
 237                   NodeClassNames[sopc], type2name(type), num_elem);
 238     return false;
 239   }
 240 
 241   // Check if mask unboxing is supported, this is a two step process which first loads the contents
 242   // of boolean array into vector followed by either lane expansion to match the lane size of masked
 243   // vector operation or populate the predicate register.
 244   if ((mask_use_type & VecMaskUseLoad) != 0) {
 245     if (!Matcher::match_rule_supported_vector(Op_VectorLoadMask, num_elem, type) ||
 246         !Matcher::match_rule_supported_vector(Op_LoadVector, num_elem, T_BOOLEAN)) {
 247       non_product_log_if_needed("  ** Rejected vector mask loading (%s,%s,%d) because architecture does not support it",
 248                                 NodeClassNames[Op_VectorLoadMask], type2name(type), num_elem);
 249       return false;
 250     }
 251   }
 252 
 253   // Check if mask boxing is supported, this is a two step process which first stores the contents
 254   // of mask vector / predicate register into a boolean vector followed by vector store operation to
 255   // transfer the contents to underlined storage of mask boxes which is a boolean array.
 256   if ((mask_use_type & VecMaskUseStore) != 0) {
 257     if (!Matcher::match_rule_supported_vector(Op_VectorStoreMask, num_elem, type) ||
 258         !Matcher::match_rule_supported_vector(Op_StoreVector, num_elem, T_BOOLEAN)) {
 259       non_product_log_if_needed("Rejected vector mask storing (%s,%s,%d) because architecture does not support it",
 260                                 NodeClassNames[Op_VectorStoreMask], type2name(type), num_elem);
 261       return false;
 262     }
 263   }
 264 
 265   if ((mask_use_type & VecMaskUsePred) != 0) {
 266     bool is_supported = false;
 267     if (Matcher::has_predicated_vectors()) {
 268       if (VectorNode::is_vector_integral_negate(sopc)) {
 269         is_supported = VectorNode::is_vector_integral_negate_supported(sopc, num_elem, type, true);
 270       } else {
 271         is_supported = Matcher::match_rule_supported_vector_masked(sopc, num_elem, type);
 272       }
 273     }
 274     is_supported |= Matcher::supports_vector_predicate_op_emulation(sopc, num_elem, type);
 275 
 276     if (!is_supported) {
 277       non_product_log_if_needed("Rejected vector mask predicate using (%s,%s,%d) because architecture does not support it",
 278                                 NodeClassNames[sopc], type2name(type), num_elem);
 279       return false;
 280     }
 281   }
 282 
 283   return true;
 284 }
 285 
 286 static bool is_klass_initialized(const TypeInstPtr* vec_klass) {
 287   if (vec_klass->const_oop() == nullptr) {
 288     return false; // uninitialized or some kind of unsafe access
 289   }
 290   assert(vec_klass->const_oop()->as_instance()->java_lang_Class_klass() != nullptr, "klass instance expected");
 291   ciInstanceKlass* klass =  vec_klass->const_oop()->as_instance()->java_lang_Class_klass()->as_instance_klass();
 292   return klass->is_initialized();
 293 }
 294 
 295 // public static
 296 // <V extends Vector<E>,
 297 //  M extends VectorMask<E>,
 298 //  E>
 299 // V unaryOp(int oprId, Class<? extends V> vmClass, Class<? extends M> maskClass, Class<E> elementType,
 300 //           int length, V v, M m,
 301 //           UnaryOperation<V, M> defaultImpl)
 302 //
 303 // public static
 304 // <V,
 305 //  M extends VectorMask<E>,
 306 //  E>
 307 // V binaryOp(int oprId, Class<? extends V> vmClass, Class<? extends M> maskClass, Class<E> elementType,
 308 //            int length, V v1, V v2, M m,
 309 //            BinaryOperation<V, M> defaultImpl)
 310 //
 311 // public static
 312 // <V extends Vector<E>,
 313 //  M extends VectorMask<E>,
 314 //  E>
 315 // V ternaryOp(int oprId, Class<? extends V> vmClass, Class<? extends M> maskClass, Class<E> elementType,
 316 //             int length, V v1, V v2, V v3, M m,
 317 //             TernaryOperation<V, M> defaultImpl)
 318 //
 319 bool LibraryCallKit::inline_vector_nary_operation(int n) {
 320   const TypeInt*     opr          = gvn().type(argument(0))->isa_int();
 321   const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr();
 322   const TypeInstPtr* mask_klass   = gvn().type(argument(2))->isa_instptr();
 323   const TypeInstPtr* elem_klass   = gvn().type(argument(3))->isa_instptr();
 324   const TypeInt*     vlen         = gvn().type(argument(4))->isa_int();
 325 
 326   if (opr          == nullptr || !opr->is_con() ||
 327       vector_klass == nullptr || vector_klass->const_oop() == nullptr ||
 328       elem_klass   == nullptr || elem_klass->const_oop()   == nullptr ||
 329       vlen         == nullptr || !vlen->is_con()) {
 330     log_if_needed("  ** missing constant: opr=%s vclass=%s etype=%s vlen=%s",
 331                     NodeClassNames[argument(0)->Opcode()],
 332                     NodeClassNames[argument(1)->Opcode()],
 333                     NodeClassNames[argument(3)->Opcode()],
 334                     NodeClassNames[argument(4)->Opcode()]);
 335     return false; // not enough info for intrinsification
 336   }
 337 
 338   ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type();
 339   if (!elem_type->is_primitive_type()) {
 340     log_if_needed("  ** not a primitive bt=%d", elem_type->basic_type());
 341     return false; // should be primitive type
 342   }
 343   if (!is_klass_initialized(vector_klass)) {
 344     log_if_needed("  ** klass argument not initialized");
 345     return false;
 346   }
 347 
 348   // "argument(n + 5)" should be the mask object. We assume it is "null" when no mask
 349   // is used to control this operation.
 350   const Type* vmask_type = gvn().type(argument(n + 5));
 351   bool is_masked_op = vmask_type != TypePtr::NULL_PTR;
 352   if (is_masked_op) {
 353     if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) {
 354       log_if_needed("  ** missing constant: maskclass=%s", NodeClassNames[argument(2)->Opcode()]);
 355       return false; // not enough info for intrinsification
 356     }
 357 
 358     if (!is_klass_initialized(mask_klass)) {
 359       log_if_needed("  ** mask klass argument not initialized");
 360       return false;
 361     }
 362 
 363     if (vmask_type->maybe_null()) {
 364       log_if_needed("  ** null mask values are not allowed for masked op");
 365       return false;
 366     }
 367   }
 368 
 369   BasicType elem_bt = elem_type->basic_type();
 370   int num_elem = vlen->get_con();
 371   int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt);
 372   int sopc = VectorNode::opcode(opc, elem_bt);
 373   if ((opc != Op_CallLeafVector) && (sopc == 0)) {
 374     log_if_needed("  ** operation not supported: opc=%s bt=%s", NodeClassNames[opc], type2name(elem_bt));
 375     return false; // operation not supported
 376   }
 377   if (num_elem == 1) {
 378     if (opc != Op_CallLeafVector || elem_bt != T_DOUBLE) {
 379       log_if_needed("  ** not a svml call: arity=%d opc=%d vlen=%d etype=%s",
 380                       n, opc, num_elem, type2name(elem_bt));
 381       return false;
 382     }
 383   }
 384   ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass();
 385   const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass);
 386 
 387   if (is_vector_mask(vbox_klass)) {
 388     assert(!is_masked_op, "mask operations do not need mask to control");
 389   }
 390 
 391   if (opc == Op_CallLeafVector) {
 392     if (!UseVectorStubs) {
 393       log_if_needed("  ** vector stubs support is disabled");
 394       return false;
 395     }
 396     if (!Matcher::supports_vector_calling_convention()) {
 397       log_if_needed("  ** no vector calling conventions supported");
 398       return false;
 399     }
 400     if (!Matcher::vector_size_supported(elem_bt, num_elem)) {
 401       log_if_needed("  ** vector size (vlen=%d, etype=%s) is not supported",
 402                       num_elem, type2name(elem_bt));
 403       return false;
 404     }
 405   }
 406 
 407   // When using mask, mask use type needs to be VecMaskUseLoad.
 408   VectorMaskUseType mask_use_type = is_vector_mask(vbox_klass) ? VecMaskUseAll
 409                                       : is_masked_op ? VecMaskUseLoad : VecMaskNotUsed;
 410   if ((sopc != 0) && !arch_supports_vector(sopc, num_elem, elem_bt, mask_use_type)) {
 411     log_if_needed("  ** not supported: arity=%d opc=%d vlen=%d etype=%s ismask=%d is_masked_op=%d",
 412                     n, sopc, num_elem, type2name(elem_bt),
 413                     is_vector_mask(vbox_klass) ? 1 : 0, is_masked_op ? 1 : 0);
 414     return false; // not supported
 415   }
 416 
 417   // Return true if current platform has implemented the masked operation with predicate feature.
 418   bool use_predicate = is_masked_op && sopc != 0 && arch_supports_vector(sopc, num_elem, elem_bt, VecMaskUsePred);
 419   if (is_masked_op && !use_predicate && !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)) {
 420     log_if_needed("  ** not supported: arity=%d opc=%d vlen=%d etype=%s ismask=0 is_masked_op=1",
 421                     n, sopc, num_elem, type2name(elem_bt));
 422     return false;
 423   }
 424 
 425   Node* opd1 = nullptr; Node* opd2 = nullptr; Node* opd3 = nullptr;
 426   switch (n) {
 427     case 3: {
 428       opd3 = unbox_vector(argument(7), vbox_type, elem_bt, num_elem);
 429       if (opd3 == nullptr) {
 430         log_if_needed("  ** unbox failed v3=%s",
 431                         NodeClassNames[argument(7)->Opcode()]);
 432         return false;
 433       }
 434       // fall-through
 435     }
 436     case 2: {
 437       opd2 = unbox_vector(argument(6), vbox_type, elem_bt, num_elem);
 438       if (opd2 == nullptr) {
 439         log_if_needed("  ** unbox failed v2=%s",
 440                         NodeClassNames[argument(6)->Opcode()]);
 441         return false;
 442       }
 443       // fall-through
 444     }
 445     case 1: {
 446       opd1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem);
 447       if (opd1 == nullptr) {
 448         log_if_needed("  ** unbox failed v1=%s",
 449                         NodeClassNames[argument(5)->Opcode()]);
 450         return false;
 451       }
 452       break;
 453     }
 454     default: fatal("unsupported arity: %d", n);
 455   }
 456 
 457   Node* mask = nullptr;
 458   if (is_masked_op) {
 459     ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass();
 460     assert(is_vector_mask(mbox_klass), "argument(2) should be a mask class");
 461     const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass);
 462     mask = unbox_vector(argument(n + 5), mbox_type, elem_bt, num_elem);
 463     if (mask == nullptr) {
 464       log_if_needed("  ** unbox failed mask=%s",
 465                       NodeClassNames[argument(n + 5)->Opcode()]);
 466       return false;
 467     }
 468   }
 469 
 470   Node* operation = nullptr;
 471   if (opc == Op_CallLeafVector) {
 472     assert(UseVectorStubs, "sanity");
 473     operation = gen_call_to_svml(opr->get_con(), elem_bt, num_elem, opd1, opd2);
 474     if (operation == nullptr) {
 475       log_if_needed("  ** svml call failed for %s_%s_%d",
 476                          (elem_bt == T_FLOAT)?"float":"double",
 477                          VectorSupport::svmlname[opr->get_con() - VectorSupport::VECTOR_OP_SVML_START],
 478                          num_elem * type2aelembytes(elem_bt));
 479       return false;
 480      }
 481   } else {
 482     const TypeVect* vt = TypeVect::make(elem_bt, num_elem, is_vector_mask(vbox_klass));
 483     switch (n) {
 484       case 1:
 485       case 2: {
 486         operation = VectorNode::make(sopc, opd1, opd2, vt, is_vector_mask(vbox_klass), VectorNode::is_shift_opcode(opc));
 487         break;
 488       }
 489       case 3: {
 490         operation = VectorNode::make(sopc, opd1, opd2, opd3, vt);
 491         break;
 492       }
 493       default: fatal("unsupported arity: %d", n);
 494     }
 495   }
 496 
 497   if (is_masked_op && mask != nullptr) {
 498     if (use_predicate) {
 499       operation->add_req(mask);
 500       operation->add_flag(Node::Flag_is_predicated_vector);
 501     } else {
 502       operation->add_flag(Node::Flag_is_predicated_using_blend);
 503       operation = gvn().transform(operation);
 504       operation = new VectorBlendNode(opd1, operation, mask);
 505     }
 506   }
 507   operation = gvn().transform(operation);
 508 
 509   // Wrap it up in VectorBox to keep object type information.
 510   Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem);
 511   set_result(vbox);
 512   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt))));
 513   return true;
 514 }
 515 
 516 // Following routine generates IR corresponding to AbstractShuffle::partiallyWrapIndex method,
 517 // which partially wraps index by modulo VEC_LENGTH and generates a negative index value if original
 518 // index is out of valid index range [0, VEC_LENGTH)
 519 //
 520 //   wrapped_index = (VEC_LENGTH - 1) & index
 521 //   if (index u> VEC_LENGTH) {
 522 //     wrapped_index -= VEC_LENGTH;
 523 //
 524 // Note: Unsigned greater than comparison treat both <0 and >VEC_LENGTH indices as out-of-bound
 525 // indexes.
 526 Node* LibraryCallKit::partially_wrap_indexes(Node* index_vec, int num_elem, BasicType elem_bt) {
 527   assert(elem_bt == T_BYTE, "Shuffles use byte array based backing storage.");
 528   const TypeVect* vt  = TypeVect::make(elem_bt, num_elem);
 529   const Type* type_bt = Type::get_const_basic_type(elem_bt);
 530 
 531   Node* mod_mask = gvn().makecon(TypeInt::make(num_elem-1));
 532   Node* bcast_mod_mask  = gvn().transform(VectorNode::scalar2vector(mod_mask, num_elem, type_bt));
 533 
 534   BoolTest::mask pred = BoolTest::ugt;
 535   ConINode* pred_node = (ConINode*)gvn().makecon(TypeInt::make(pred));
 536   Node* lane_cnt  = gvn().makecon(TypeInt::make(num_elem));
 537   Node* bcast_lane_cnt = gvn().transform(VectorNode::scalar2vector(lane_cnt, num_elem, type_bt));
 538   const TypeVect* vmask_type = TypeVect::makemask(type_bt, num_elem);
 539   Node*  mask = gvn().transform(new VectorMaskCmpNode(pred, bcast_lane_cnt, index_vec, pred_node, vmask_type));
 540 
 541   // Make the indices greater than lane count as -ve values to match the java side implementation.
 542   index_vec = gvn().transform(VectorNode::make(Op_AndV, index_vec, bcast_mod_mask, vt));
 543   Node* biased_val = gvn().transform(VectorNode::make(Op_SubVB, index_vec, bcast_lane_cnt, vt));
 544   return gvn().transform(new VectorBlendNode(biased_val, index_vec, mask));
 545 }
 546 
 547 // <Sh extends VectorShuffle<E>,  E>
 548 //  Sh ShuffleIota(Class<?> E, Class<?> shuffleClass, Vector.Species<E> s, int length,
 549 //                  int start, int step, int wrap, ShuffleIotaOperation<Sh, E> defaultImpl)
 550 bool LibraryCallKit::inline_vector_shuffle_iota() {
 551   const TypeInstPtr* shuffle_klass = gvn().type(argument(1))->isa_instptr();
 552   const TypeInt*     vlen          = gvn().type(argument(3))->isa_int();
 553   const TypeInt*     start_val     = gvn().type(argument(4))->isa_int();
 554   const TypeInt*     step_val      = gvn().type(argument(5))->isa_int();
 555   const TypeInt*     wrap          = gvn().type(argument(6))->isa_int();
 556 
 557   if (shuffle_klass == nullptr || shuffle_klass->const_oop() == nullptr ||
 558       vlen          == nullptr || !vlen->is_con() ||
 559       start_val     == nullptr ||
 560       step_val      == nullptr ||
 561       wrap          == nullptr || !wrap->is_con()) {
 562     return false; // not enough info for intrinsification
 563   }
 564 
 565   if (!is_klass_initialized(shuffle_klass)) {
 566     log_if_needed("  ** klass argument not initialized");
 567     return false;
 568   }
 569 
 570   int do_wrap = wrap->get_con();
 571   int num_elem = vlen->get_con();
 572   BasicType elem_bt = T_BYTE;
 573 
 574   bool effective_indices_in_range = false;
 575   if (start_val->is_con() && step_val->is_con()) {
 576     int effective_min_index = start_val->get_con();
 577     int effective_max_index = start_val->get_con() + step_val->get_con() * (num_elem - 1);
 578     effective_indices_in_range = effective_max_index >= effective_min_index && effective_min_index >= -128 && effective_max_index <= 127;
 579   }
 580 
 581   if (!do_wrap && !effective_indices_in_range) {
 582     // Disable instrinsification for unwrapped shuffle iota if start/step
 583     // values are non-constant OR if intermediate result overflows byte value range.
 584     return false;
 585   }
 586 
 587   if (!arch_supports_vector(Op_AddVB, num_elem, elem_bt, VecMaskNotUsed)           ||
 588       !arch_supports_vector(Op_AndV, num_elem, elem_bt, VecMaskNotUsed)            ||
 589       !arch_supports_vector(Op_VectorLoadConst, num_elem, elem_bt, VecMaskNotUsed) ||
 590       !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed)) {
 591     return false;
 592   }
 593 
 594   if (!do_wrap &&
 595       (!arch_supports_vector(Op_SubVB, num_elem, elem_bt, VecMaskNotUsed)       ||
 596       !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskNotUsed)  ||
 597       !arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskNotUsed))) {
 598     return false;
 599   }
 600 
 601   bool step_multiply = !step_val->is_con() || !is_power_of_2(step_val->get_con());
 602   if ((step_multiply && !arch_supports_vector(Op_MulVB, num_elem, elem_bt, VecMaskNotUsed)) ||
 603       (!step_multiply && !arch_supports_vector(Op_LShiftVB, num_elem, elem_bt, VecMaskNotUsed))) {
 604     return false;
 605   }
 606 
 607   const Type * type_bt = Type::get_const_basic_type(elem_bt);
 608   const TypeVect * vt  = TypeVect::make(type_bt, num_elem);
 609 
 610   Node* res = gvn().transform(new VectorLoadConstNode(gvn().makecon(TypeInt::ZERO), vt));
 611 
 612   Node* start = argument(4);
 613   Node* step  = argument(5);
 614 
 615   if (step_multiply) {
 616     Node* bcast_step     = gvn().transform(VectorNode::scalar2vector(step, num_elem, type_bt));
 617     res = gvn().transform(VectorNode::make(Op_MulVB, res, bcast_step, vt));
 618   } else if (step_val->get_con() > 1) {
 619     Node* cnt = gvn().makecon(TypeInt::make(log2i_exact(step_val->get_con())));
 620     Node* shift_cnt = vector_shift_count(cnt, Op_LShiftI, elem_bt, num_elem);
 621     res = gvn().transform(VectorNode::make(Op_LShiftVB, res, shift_cnt, vt));
 622   }
 623 
 624   if (!start_val->is_con() || start_val->get_con() != 0) {
 625     Node* bcast_start    = gvn().transform(VectorNode::scalar2vector(start, num_elem, type_bt));
 626     res = gvn().transform(VectorNode::make(Op_AddVB, res, bcast_start, vt));
 627   }
 628 
 629   Node * mod_val = gvn().makecon(TypeInt::make(num_elem-1));
 630   Node * bcast_mod  = gvn().transform(VectorNode::scalar2vector(mod_val, num_elem, type_bt));
 631 
 632   if (do_wrap)  {
 633     // Wrap the indices greater than lane count.
 634     res = gvn().transform(VectorNode::make(Op_AndV, res, bcast_mod, vt));
 635   } else {
 636     res = partially_wrap_indexes(res, num_elem, elem_bt);
 637   }
 638 
 639   ciKlass* sbox_klass = shuffle_klass->const_oop()->as_instance()->java_lang_Class_klass();
 640   const TypeInstPtr* shuffle_box_type = TypeInstPtr::make_exact(TypePtr::NotNull, sbox_klass);
 641 
 642   // Wrap it up in VectorBox to keep object type information.
 643   res = box_vector(res, shuffle_box_type, elem_bt, num_elem);
 644   set_result(res);
 645   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt))));
 646   return true;
 647 }
 648 
 649 // <E, M>
 650 // long maskReductionCoerced(int oper, Class<? extends M> maskClass, Class<?> elemClass,
 651 //                          int length, M m, VectorMaskOp<M> defaultImpl)
 652 bool LibraryCallKit::inline_vector_mask_operation() {
 653   const TypeInt*     oper       = gvn().type(argument(0))->isa_int();
 654   const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr();
 655   const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr();
 656   const TypeInt*     vlen       = gvn().type(argument(3))->isa_int();
 657   Node*              mask       = argument(4);
 658 
 659   if (mask_klass == nullptr || mask_klass->const_oop() == nullptr ||
 660       elem_klass == nullptr || elem_klass->const_oop() == nullptr ||
 661       vlen       == nullptr || !vlen->is_con() ||
 662       oper       == nullptr || !oper->is_con() ||
 663       mask->is_top()) {
 664     return false; // dead code
 665   }
 666 
 667   if (!is_klass_initialized(mask_klass)) {
 668     log_if_needed("  ** klass argument not initialized");
 669     return false;
 670   }
 671 
 672   int num_elem = vlen->get_con();
 673   ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type();
 674   BasicType elem_bt = elem_type->basic_type();
 675 
 676   int mopc = VectorSupport::vop2ideal(oper->get_con(), elem_bt);
 677   if (!arch_supports_vector(mopc, num_elem, elem_bt, VecMaskUseLoad)) {
 678     log_if_needed("  ** not supported: arity=1 op=cast#%d/3 vlen2=%d etype2=%s",
 679                     mopc, num_elem, type2name(elem_bt));
 680     return false; // not supported
 681   }
 682 
 683   const Type* elem_ty = Type::get_const_basic_type(elem_bt);
 684   ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass();
 685   const TypeInstPtr* mask_box_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass);
 686   Node* mask_vec = unbox_vector(mask, mask_box_type, elem_bt, num_elem, true);
 687   if (mask_vec == nullptr) {
 688     log_if_needed("  ** unbox failed mask=%s",
 689                       NodeClassNames[argument(4)->Opcode()]);
 690     return false;
 691   }
 692 
 693   if (mask_vec->bottom_type()->isa_vectmask() == nullptr) {
 694     mask_vec = gvn().transform(VectorStoreMaskNode::make(gvn(), mask_vec, elem_bt, num_elem));
 695   }
 696   const Type* maskoper_ty = mopc == Op_VectorMaskToLong ? (const Type*)TypeLong::LONG : (const Type*)TypeInt::INT;
 697   Node* maskoper = gvn().transform(VectorMaskOpNode::make(mask_vec, maskoper_ty, mopc));
 698   if (mopc != Op_VectorMaskToLong) {
 699     maskoper = ConvI2L(maskoper);
 700   }
 701   set_result(maskoper);
 702 
 703   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt))));
 704   return true;
 705 }
 706 
 707 // public static
 708 // <V,
 709 //  Sh extends VectorShuffle<E>,
 710 //  E>
 711 // V shuffleToVector(Class<? extends Vector<E>> vclass, Class<E> elementType,
 712 //                   Class<? extends Sh> shuffleClass, Sh s, int length,
 713 //                   ShuffleToVectorOperation<V, Sh, E> defaultImpl)
 714 bool LibraryCallKit::inline_vector_shuffle_to_vector() {
 715   const TypeInstPtr* vector_klass  = gvn().type(argument(0))->isa_instptr();
 716   const TypeInstPtr* elem_klass    = gvn().type(argument(1))->isa_instptr();
 717   const TypeInstPtr* shuffle_klass = gvn().type(argument(2))->isa_instptr();
 718   Node*              shuffle       = argument(3);
 719   const TypeInt*     vlen          = gvn().type(argument(4))->isa_int();
 720 
 721   if (vector_klass == nullptr || elem_klass == nullptr || shuffle_klass == nullptr || shuffle->is_top() || vlen == nullptr) {
 722     return false; // dead code
 723   }
 724   if (vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || shuffle_klass->const_oop() == nullptr || !vlen->is_con()) {
 725     return false; // not enough info for intrinsification
 726   }
 727   if (!is_klass_initialized(shuffle_klass) || !is_klass_initialized(vector_klass) ) {
 728     log_if_needed("  ** klass argument not initialized");
 729     return false;
 730   }
 731 
 732   int num_elem = vlen->get_con();
 733   ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type();
 734   BasicType elem_bt = elem_type->basic_type();
 735 
 736   if (num_elem < 4) {
 737     return false;
 738   }
 739 
 740   int cast_vopc = VectorCastNode::opcode(-1, T_BYTE); // from shuffle of type T_BYTE
 741   // Make sure that cast is implemented to particular type/size combination.
 742   if (!arch_supports_vector(cast_vopc, num_elem, elem_bt, VecMaskNotUsed)) {
 743     log_if_needed("  ** not supported: arity=1 op=cast#%d/3 vlen2=%d etype2=%s",
 744         cast_vopc, num_elem, type2name(elem_bt));
 745     return false;
 746   }
 747 
 748   ciKlass* sbox_klass = shuffle_klass->const_oop()->as_instance()->java_lang_Class_klass();
 749   const TypeInstPtr* shuffle_box_type = TypeInstPtr::make_exact(TypePtr::NotNull, sbox_klass);
 750 
 751   // Unbox shuffle with true flag to indicate its load shuffle to vector
 752   // shuffle is a byte array
 753   Node* shuffle_vec = unbox_vector(shuffle, shuffle_box_type, T_BYTE, num_elem, true);
 754 
 755   // cast byte to target element type
 756   shuffle_vec = gvn().transform(VectorCastNode::make(cast_vopc, shuffle_vec, elem_bt, num_elem));
 757 
 758   ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass();
 759   const TypeInstPtr* vec_box_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass);
 760 
 761   // Box vector
 762   Node* res = box_vector(shuffle_vec, vec_box_type, elem_bt, num_elem);
 763   set_result(res);
 764   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt))));
 765   return true;
 766 }
 767 
 768 // public static
 769 // <M,
 770 //  S extends VectorSpecies<E>,
 771 //  E>
 772 // M fromBitsCoerced(Class<? extends M> vmClass, Class<E> elementType, int length,
 773 //                    long bits, int mode, S s,
 774 //                    BroadcastOperation<M, E, S> defaultImpl)
 775 bool LibraryCallKit::inline_vector_frombits_coerced() {
 776   const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr();
 777   const TypeInstPtr* elem_klass   = gvn().type(argument(1))->isa_instptr();
 778   const TypeInt*     vlen         = gvn().type(argument(2))->isa_int();
 779   const TypeLong*    bits_type    = gvn().type(argument(3))->isa_long();
 780   // Mode argument determines the mode of operation it can take following values:-
 781   // MODE_BROADCAST for vector Vector.broadcast and VectorMask.maskAll operations.
 782   // MODE_BITS_COERCED_LONG_TO_MASK for VectorMask.fromLong operation.
 783   const TypeInt*     mode         = gvn().type(argument(5))->isa_int();
 784 
 785   if (vector_klass == nullptr || vector_klass->const_oop() == nullptr ||
 786       elem_klass   == nullptr || elem_klass->const_oop()   == nullptr ||
 787       vlen         == nullptr || !vlen->is_con() ||
 788       bits_type    == nullptr ||
 789       mode         == nullptr || !mode->is_con()) {
 790     log_if_needed("  ** missing constant: vclass=%s etype=%s vlen=%s bitwise=%s",
 791                     NodeClassNames[argument(0)->Opcode()],
 792                     NodeClassNames[argument(1)->Opcode()],
 793                     NodeClassNames[argument(2)->Opcode()],
 794                     NodeClassNames[argument(5)->Opcode()]);
 795     return false; // not enough info for intrinsification
 796   }
 797 
 798   if (!is_klass_initialized(vector_klass)) {
 799     log_if_needed("  ** klass argument not initialized");
 800     return false;
 801   }
 802   ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type();
 803   if (!elem_type->is_primitive_type()) {
 804     log_if_needed("  ** not a primitive bt=%d", elem_type->basic_type());
 805     return false; // should be primitive type
 806   }
 807   BasicType elem_bt = elem_type->basic_type();
 808   int num_elem = vlen->get_con();
 809   ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass();
 810   const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass);
 811 
 812   bool is_mask = is_vector_mask(vbox_klass);
 813   int  bcast_mode = mode->get_con();
 814   VectorMaskUseType checkFlags = (VectorMaskUseType)(is_mask ? VecMaskUseAll : VecMaskNotUsed);
 815   int opc = bcast_mode == VectorSupport::MODE_BITS_COERCED_LONG_TO_MASK ? Op_VectorLongToMask : Op_Replicate;
 816 
 817   if (!arch_supports_vector(opc, num_elem, elem_bt, checkFlags, true /*has_scalar_args*/)) {
 818     log_if_needed("  ** not supported: arity=0 op=broadcast vlen=%d etype=%s ismask=%d bcast_mode=%d",
 819                     num_elem, type2name(elem_bt),
 820                     is_mask ? 1 : 0,
 821                     bcast_mode);
 822     return false; // not supported
 823   }
 824 
 825   Node* broadcast = nullptr;
 826   Node* bits = argument(3);
 827   Node* elem = bits;
 828 
 829   if (opc == Op_VectorLongToMask) {
 830     const TypeVect* vt = TypeVect::makemask(elem_bt, num_elem);
 831     if (vt->isa_vectmask()) {
 832       broadcast = gvn().transform(new VectorLongToMaskNode(elem, vt));
 833     } else {
 834       const TypeVect* mvt = TypeVect::make(T_BOOLEAN, num_elem);
 835       broadcast = gvn().transform(new VectorLongToMaskNode(elem, mvt));
 836       broadcast = gvn().transform(new VectorLoadMaskNode(broadcast, vt));
 837     }
 838   } else {
 839     switch (elem_bt) {
 840       case T_BOOLEAN: // fall-through
 841       case T_BYTE:    // fall-through
 842       case T_SHORT:   // fall-through
 843       case T_CHAR:    // fall-through
 844       case T_INT: {
 845         elem = gvn().transform(new ConvL2INode(bits));
 846         break;
 847       }
 848       case T_DOUBLE: {
 849         elem = gvn().transform(new MoveL2DNode(bits));
 850         break;
 851       }
 852       case T_FLOAT: {
 853         bits = gvn().transform(new ConvL2INode(bits));
 854         elem = gvn().transform(new MoveI2FNode(bits));
 855         break;
 856       }
 857       case T_LONG: {
 858         // no conversion needed
 859         break;
 860       }
 861       default: fatal("%s", type2name(elem_bt));
 862     }
 863     broadcast = VectorNode::scalar2vector(elem, num_elem, Type::get_const_basic_type(elem_bt), is_mask);
 864     broadcast = gvn().transform(broadcast);
 865   }
 866 
 867   Node* box = box_vector(broadcast, vbox_type, elem_bt, num_elem);
 868   set_result(box);
 869   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt))));
 870   return true;
 871 }
 872 
 873 static bool elem_consistent_with_arr(BasicType elem_bt, const TypeAryPtr* arr_type, bool mismatched_ms) {
 874   assert(arr_type != nullptr, "unexpected");
 875   BasicType arr_elem_bt = arr_type->elem()->array_element_basic_type();
 876   if (elem_bt == arr_elem_bt) {
 877     return true;
 878   } else if (elem_bt == T_SHORT && arr_elem_bt == T_CHAR) {
 879     // Load/store of short vector from/to char[] is supported
 880     return true;
 881   } else if (elem_bt == T_BYTE && arr_elem_bt == T_BOOLEAN) {
 882     // Load/store of byte vector from/to boolean[] is supported
 883     return true;
 884   } else {
 885     return mismatched_ms;
 886   }
 887 }
 888 
 889 //  public static
 890 //  <C,
 891 //   VM extends VectorPayload,
 892 //   E,
 893 //   S extends VectorSpecies<E>>
 894 //  VM load(Class<? extends VM> vmClass, Class<E> eClass,
 895 //          int length,
 896 //          Object base, long offset,            // Unsafe addressing
 897 //          boolean fromSegment,
 898 //          C container, long index, S s,        // Arguments for default implementation
 899 //          LoadOperation<C, VM, S> defaultImpl) {
 900 //  public static
 901 //  <C,
 902 //   V extends VectorPayload>
 903 //  void store(Class<?> vClass, Class<?> eClass,
 904 //             int length,
 905 //             Object base, long offset,        // Unsafe addressing
 906 //             boolean fromSegment,
 907 //             V v, C container, long index,    // Arguments for default implementation
 908 //             StoreVectorOperation<C, V> defaultImpl) {
 909 bool LibraryCallKit::inline_vector_mem_operation(bool is_store) {
 910   const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr();
 911   const TypeInstPtr* elem_klass   = gvn().type(argument(1))->isa_instptr();
 912   const TypeInt*     vlen         = gvn().type(argument(2))->isa_int();
 913   const TypeInt*     from_ms      = gvn().type(argument(6))->isa_int();
 914 
 915   if (vector_klass == nullptr || vector_klass->const_oop() == nullptr ||
 916       elem_klass   == nullptr || elem_klass->const_oop()   == nullptr ||
 917       vlen         == nullptr || !vlen->is_con() ||
 918       from_ms      == nullptr || !from_ms->is_con()) {
 919     log_if_needed("  ** missing constant: vclass=%s etype=%s vlen=%s from_ms=%s",
 920                     NodeClassNames[argument(0)->Opcode()],
 921                     NodeClassNames[argument(1)->Opcode()],
 922                     NodeClassNames[argument(2)->Opcode()],
 923                     NodeClassNames[argument(6)->Opcode()]);
 924     return false; // not enough info for intrinsification
 925   }
 926   if (!is_klass_initialized(vector_klass)) {
 927     log_if_needed("  ** klass argument not initialized");
 928     return false;
 929   }
 930 
 931   ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type();
 932   if (!elem_type->is_primitive_type()) {
 933     log_if_needed("  ** not a primitive bt=%d", elem_type->basic_type());
 934     return false; // should be primitive type
 935   }
 936   BasicType elem_bt = elem_type->basic_type();
 937   int num_elem = vlen->get_con();
 938 
 939   // TODO When mask usage is supported, VecMaskNotUsed needs to be VecMaskUseLoad.
 940   if (!arch_supports_vector(is_store ? Op_StoreVector : Op_LoadVector, num_elem, elem_bt, VecMaskNotUsed)) {
 941     log_if_needed("  ** not supported: arity=%d op=%s vlen=%d etype=%s ismask=no",
 942                     is_store, is_store ? "store" : "load",
 943                     num_elem, type2name(elem_bt));
 944     return false; // not supported
 945   }
 946 
 947   ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass();
 948   bool is_mask = is_vector_mask(vbox_klass);
 949 
 950   Node* base = argument(3);
 951   Node* offset = ConvL2X(argument(4));
 952 
 953   // Save state and restore on bailout
 954   uint old_sp = sp();
 955   SafePointNode* old_map = clone_map();
 956 
 957   Node* addr = make_unsafe_address(base, offset, (is_mask ? T_BOOLEAN : elem_bt), true);
 958 
 959   // The memory barrier checks are based on ones for unsafe access.
 960   // This is not 1-1 implementation.
 961   const Type *const base_type = gvn().type(base);
 962 
 963   const TypePtr *addr_type = gvn().type(addr)->isa_ptr();
 964   const TypeAryPtr* arr_type = addr_type->isa_aryptr();
 965 
 966   const bool in_native = TypePtr::NULL_PTR == base_type; // base always null
 967   const bool in_heap   = !TypePtr::NULL_PTR->higher_equal(base_type); // base never null
 968 
 969   const bool is_mixed_access = !in_heap && !in_native;
 970 
 971   const bool is_mismatched_access = in_heap && (addr_type->isa_aryptr() == nullptr);
 972 
 973   const bool needs_cpu_membar = is_mixed_access || is_mismatched_access;
 974 
 975   // For non-masked mismatched memory segment vector read/write accesses, intrinsification can continue
 976   // with unknown backing storage type and compiler can skip inserting explicit reinterpretation IR after
 977   // loading from or before storing to backing storage which is mandatory for semantic correctness of
 978   // big-endian memory layout.
 979   bool mismatched_ms = LITTLE_ENDIAN_ONLY(false)
 980       BIG_ENDIAN_ONLY(from_ms->get_con() && !is_mask && arr_type != nullptr &&
 981                       arr_type->elem()->array_element_basic_type() != elem_bt);
 982   BasicType mem_elem_bt = mismatched_ms ? arr_type->elem()->array_element_basic_type() : elem_bt;
 983   if (!is_java_primitive(mem_elem_bt)) {
 984     log_if_needed("  ** non-primitive array element type");
 985     return false;
 986   }
 987   int mem_num_elem = mismatched_ms ? (num_elem * type2aelembytes(elem_bt)) / type2aelembytes(mem_elem_bt) : num_elem;
 988   if (arr_type != nullptr && !is_mask && !elem_consistent_with_arr(elem_bt, arr_type, mismatched_ms)) {
 989     log_if_needed("  ** not supported: arity=%d op=%s vlen=%d etype=%s atype=%s ismask=no",
 990                     is_store, is_store ? "store" : "load",
 991                     num_elem, type2name(elem_bt), type2name(arr_type->elem()->array_element_basic_type()));
 992     set_map(old_map);
 993     set_sp(old_sp);
 994     return false;
 995   }
 996 
 997   // In case of mismatched memory segment accesses, we need to double check that the source type memory operations are supported by backend.
 998   if (mismatched_ms) {
 999     if (is_store) {
1000       if (!arch_supports_vector(Op_StoreVector, num_elem, elem_bt, VecMaskNotUsed)
1001           || !arch_supports_vector(Op_VectorReinterpret, mem_num_elem, mem_elem_bt, VecMaskNotUsed)) {
1002         log_if_needed("  ** not supported: arity=%d op=%s vlen=%d*8 etype=%s/8 ismask=no",
1003                         is_store, "store",
1004                         num_elem, type2name(elem_bt));
1005         set_map(old_map);
1006         set_sp(old_sp);
1007         return false; // not supported
1008       }
1009     } else {
1010       if (!arch_supports_vector(Op_LoadVector, mem_num_elem, mem_elem_bt, VecMaskNotUsed)
1011           || !arch_supports_vector(Op_VectorReinterpret, num_elem, elem_bt, VecMaskNotUsed)) {
1012         log_if_needed("  ** not supported: arity=%d op=%s vlen=%d*8 etype=%s/8 ismask=no",
1013                         is_store, "load",
1014                         mem_num_elem, type2name(mem_elem_bt));
1015         set_map(old_map);
1016         set_sp(old_sp);
1017         return false; // not supported
1018       }
1019     }
1020   }
1021   if (is_mask) {
1022     if (!is_store) {
1023       if (!arch_supports_vector(Op_LoadVector, num_elem, elem_bt, VecMaskUseLoad)) {
1024         set_map(old_map);
1025         set_sp(old_sp);
1026         return false; // not supported
1027       }
1028     } else {
1029       if (!arch_supports_vector(Op_StoreVector, num_elem, elem_bt, VecMaskUseStore)) {
1030         set_map(old_map);
1031         set_sp(old_sp);
1032         return false; // not supported
1033       }
1034     }
1035   }
1036 
1037   const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass);
1038 
1039   if (needs_cpu_membar) {
1040     insert_mem_bar(Op_MemBarCPUOrder);
1041   }
1042 
1043   if (is_store) {
1044     Node* val = unbox_vector(argument(7), vbox_type, elem_bt, num_elem);
1045     if (val == nullptr) {
1046       set_map(old_map);
1047       set_sp(old_sp);
1048       return false; // operand unboxing failed
1049     }
1050     set_all_memory(reset_memory());
1051 
1052     // In case the store needs to happen to byte array, reinterpret the incoming vector to byte vector.
1053     int store_num_elem = num_elem;
1054     if (mismatched_ms) {
1055       store_num_elem = mem_num_elem;
1056       const TypeVect* to_vect_type = TypeVect::make(mem_elem_bt, store_num_elem);
1057       val = gvn().transform(new VectorReinterpretNode(val, val->bottom_type()->is_vect(), to_vect_type));
1058     }
1059     if (is_mask) {
1060       val = gvn().transform(VectorStoreMaskNode::make(gvn(), val, elem_bt, num_elem));
1061     }
1062     Node* vstore = gvn().transform(StoreVectorNode::make(0, control(), memory(addr), addr, addr_type, val, store_num_elem));
1063     set_memory(vstore, addr_type);
1064   } else {
1065     // When using byte array, we need to load as byte then reinterpret the value. Otherwise, do a simple vector load.
1066     Node* vload = nullptr;
1067     if (mismatched_ms) {
1068       vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, mem_num_elem, mem_elem_bt));
1069       const TypeVect* to_vect_type = TypeVect::make(elem_bt, num_elem);
1070       vload = gvn().transform(new VectorReinterpretNode(vload, vload->bottom_type()->is_vect(), to_vect_type));
1071     } else {
1072       // Special handle for masks
1073       if (is_mask) {
1074         vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, num_elem, T_BOOLEAN));
1075         vload = gvn().transform(new VectorLoadMaskNode(vload, TypeVect::makemask(elem_bt, num_elem)));
1076       } else {
1077         vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, num_elem, elem_bt));
1078       }
1079     }
1080     Node* box = box_vector(vload, vbox_type, elem_bt, num_elem);
1081     set_result(box);
1082   }
1083 
1084   destruct_map_clone(old_map);
1085 
1086   if (needs_cpu_membar) {
1087     insert_mem_bar(Op_MemBarCPUOrder);
1088   }
1089 
1090   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt))));
1091   return true;
1092 }
1093 
1094 //  public static
1095 //  <C,
1096 //   V extends Vector<?>,
1097 //   E,
1098 //   S extends VectorSpecies<E>,
1099 //   M extends VectorMask<E>>
1100 //  V loadMasked(Class<? extends V> vClass, Class<M> mClass, Class<E> eClass,
1101 //               int length, Object base, long offset,          // Unsafe addressing
1102 //               boolean fromSegment,
1103 //               M m, int offsetInRange,
1104 //               C container, long index, S s,                  // Arguments for default implementation
1105 //               LoadVectorMaskedOperation<C, V, S, M> defaultImpl) {
1106 //  public static
1107 //  <C,
1108 //   V extends Vector<E>,
1109 //   M extends VectorMask<E>,
1110 //   E>
1111 //  void storeMasked(Class<? extends V> vClass, Class<M> mClass, Class<E> eClass,
1112 //                   int length,
1113 //                   Object base, long offset,                  // Unsafe addressing
1114 //                   boolean fromSegment,
1115 //                   V v, M m, C container, long index,         // Arguments for default implementation
1116 //                   StoreVectorMaskedOperation<C, V, M> defaultImpl) {
1117 
1118 bool LibraryCallKit::inline_vector_mem_masked_operation(bool is_store) {
1119   const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr();
1120   const TypeInstPtr* mask_klass   = gvn().type(argument(1))->isa_instptr();
1121   const TypeInstPtr* elem_klass   = gvn().type(argument(2))->isa_instptr();
1122   const TypeInt*     vlen         = gvn().type(argument(3))->isa_int();
1123   const TypeInt*     from_ms      = gvn().type(argument(7))->isa_int();
1124 
1125   if (vector_klass == nullptr || vector_klass->const_oop() == nullptr ||
1126       mask_klass   == nullptr || mask_klass->const_oop()   == nullptr ||
1127       elem_klass   == nullptr || elem_klass->const_oop()   == nullptr ||
1128       vlen         == nullptr || !vlen->is_con() ||
1129       from_ms      == nullptr || !from_ms->is_con()) {
1130     log_if_needed("  ** missing constant: vclass=%s mclass=%s etype=%s vlen=%s from_ms=%s",
1131                     NodeClassNames[argument(0)->Opcode()],
1132                     NodeClassNames[argument(1)->Opcode()],
1133                     NodeClassNames[argument(2)->Opcode()],
1134                     NodeClassNames[argument(3)->Opcode()],
1135                     NodeClassNames[argument(7)->Opcode()]);
1136     return false; // not enough info for intrinsification
1137   }
1138   if (!is_klass_initialized(vector_klass)) {
1139     log_if_needed("  ** klass argument not initialized");
1140     return false;
1141   }
1142 
1143   if (!is_klass_initialized(mask_klass)) {
1144     log_if_needed("  ** mask klass argument not initialized");
1145     return false;
1146   }
1147 
1148   ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type();
1149   if (!elem_type->is_primitive_type()) {
1150     log_if_needed("  ** not a primitive bt=%d", elem_type->basic_type());
1151     return false; // should be primitive type
1152   }
1153 
1154   BasicType elem_bt = elem_type->basic_type();
1155   int num_elem = vlen->get_con();
1156 
1157   Node* base = argument(4);
1158   Node* offset = ConvL2X(argument(5));
1159 
1160   // Save state and restore on bailout
1161   uint old_sp = sp();
1162   SafePointNode* old_map = clone_map();
1163 
1164   Node* addr = make_unsafe_address(base, offset, elem_bt, true);
1165   const TypePtr *addr_type = gvn().type(addr)->isa_ptr();
1166   const TypeAryPtr* arr_type = addr_type->isa_aryptr();
1167 
1168   bool mismatched_ms = from_ms->get_con() && arr_type != nullptr && arr_type->elem()->array_element_basic_type() != elem_bt;
1169   BIG_ENDIAN_ONLY(if (mismatched_ms) return false;)
1170   // If there is no consistency between array and vector element types, it must be special byte array case
1171   if (arr_type != nullptr && !elem_consistent_with_arr(elem_bt, arr_type, mismatched_ms)) {
1172     log_if_needed("  ** not supported: arity=%d op=%s vlen=%d etype=%s atype=%s",
1173                     is_store, is_store ? "storeMasked" : "loadMasked",
1174                     num_elem, type2name(elem_bt), type2name(arr_type->elem()->array_element_basic_type()));
1175     set_map(old_map);
1176     set_sp(old_sp);
1177     return false;
1178   }
1179 
1180   int mem_num_elem = mismatched_ms ? num_elem * type2aelembytes(elem_bt) : num_elem;
1181   BasicType mem_elem_bt = mismatched_ms ? T_BYTE : elem_bt;
1182   bool supports_predicate = arch_supports_vector(is_store ? Op_StoreVectorMasked : Op_LoadVectorMasked,
1183                                                 mem_num_elem, mem_elem_bt, VecMaskUseLoad);
1184 
1185   // If current arch does not support the predicated operations, we have to bail
1186   // out when current case uses the predicate feature.
1187   if (!supports_predicate) {
1188     bool needs_predicate = false;
1189     if (is_store) {
1190       // Masked vector store always uses the predicated store.
1191       needs_predicate = true;
1192     } else {
1193       // Masked vector load with IOOBE always uses the predicated load.
1194       const TypeInt* offset_in_range = gvn().type(argument(9))->isa_int();
1195       if (!offset_in_range->is_con()) {
1196         log_if_needed("  ** missing constant: offsetInRange=%s",
1197                         NodeClassNames[argument(8)->Opcode()]);
1198         set_map(old_map);
1199         set_sp(old_sp);
1200         return false;
1201       }
1202       needs_predicate = (offset_in_range->get_con() == 0);
1203     }
1204 
1205     if (needs_predicate) {
1206       log_if_needed("  ** not supported: op=%s vlen=%d etype=%s mismatched_ms=%d",
1207                       is_store ? "storeMasked" : "loadMasked",
1208                       num_elem, type2name(elem_bt), mismatched_ms ? 1 : 0);
1209       set_map(old_map);
1210       set_sp(old_sp);
1211       return false;
1212     }
1213   }
1214 
1215   // This only happens for masked vector load. If predicate is not supported, then check whether
1216   // the normal vector load and blend operations are supported by backend.
1217   if (!supports_predicate && (!arch_supports_vector(Op_LoadVector, mem_num_elem, mem_elem_bt, VecMaskNotUsed) ||
1218       !arch_supports_vector(Op_VectorBlend, mem_num_elem, mem_elem_bt, VecMaskUseLoad))) {
1219     log_if_needed("  ** not supported: op=loadMasked vlen=%d etype=%s mismatched_ms=%d",
1220                     num_elem, type2name(elem_bt), mismatched_ms ? 1 : 0);
1221     set_map(old_map);
1222     set_sp(old_sp);
1223     return false;
1224   }
1225 
1226   // Since we are using byte array, we need to double check that the vector reinterpret operation
1227   // with byte type is supported by backend.
1228   if (mismatched_ms) {
1229     if (!arch_supports_vector(Op_VectorReinterpret, mem_num_elem, T_BYTE, VecMaskNotUsed)) {
1230       log_if_needed("  ** not supported: arity=%d op=%s vlen=%d etype=%s mismatched_ms=1",
1231                       is_store, is_store ? "storeMasked" : "loadMasked",
1232                       num_elem, type2name(elem_bt));
1233       set_map(old_map);
1234       set_sp(old_sp);
1235       return false;
1236     }
1237   }
1238 
1239   // Since it needs to unbox the mask, we need to double check that the related load operations
1240   // for mask are supported by backend.
1241   if (!arch_supports_vector(Op_LoadVector, num_elem, elem_bt, VecMaskUseLoad)) {
1242     log_if_needed("  ** not supported: arity=%d op=%s vlen=%d etype=%s",
1243                       is_store, is_store ? "storeMasked" : "loadMasked",
1244                       num_elem, type2name(elem_bt));
1245     set_map(old_map);
1246     set_sp(old_sp);
1247     return false;
1248   }
1249 
1250   // Can base be null? Otherwise, always on-heap access.
1251   bool can_access_non_heap = TypePtr::NULL_PTR->higher_equal(gvn().type(base));
1252   if (can_access_non_heap) {
1253     insert_mem_bar(Op_MemBarCPUOrder);
1254   }
1255 
1256   ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass();
1257   ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass();
1258   assert(!is_vector_mask(vbox_klass) && is_vector_mask(mbox_klass), "Invalid class type");
1259   const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass);
1260   const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass);
1261 
1262   Node* mask = unbox_vector(is_store ? argument(9) : argument(8), mbox_type, elem_bt, num_elem);
1263   if (mask == nullptr) {
1264     log_if_needed("  ** unbox failed mask=%s",
1265                     is_store ? NodeClassNames[argument(9)->Opcode()]
1266                              : NodeClassNames[argument(8)->Opcode()]);
1267     set_map(old_map);
1268     set_sp(old_sp);
1269     return false;
1270   }
1271 
1272   if (is_store) {
1273     Node* val = unbox_vector(argument(8), vbox_type, elem_bt, num_elem);
1274     if (val == nullptr) {
1275       log_if_needed("  ** unbox failed vector=%s",
1276                       NodeClassNames[argument(8)->Opcode()]);
1277       set_map(old_map);
1278       set_sp(old_sp);
1279       return false; // operand unboxing failed
1280     }
1281     set_all_memory(reset_memory());
1282 
1283     if (mismatched_ms) {
1284       // Reinterpret the incoming vector to byte vector.
1285       const TypeVect* to_vect_type = TypeVect::make(mem_elem_bt, mem_num_elem);
1286       val = gvn().transform(new VectorReinterpretNode(val, val->bottom_type()->is_vect(), to_vect_type));
1287       // Reinterpret the vector mask to byte type.
1288       const TypeVect* from_mask_type = TypeVect::makemask(elem_bt, num_elem);
1289       const TypeVect* to_mask_type = TypeVect::makemask(mem_elem_bt, mem_num_elem);
1290       mask = gvn().transform(new VectorReinterpretNode(mask, from_mask_type, to_mask_type));
1291     }
1292     Node* vstore = gvn().transform(new StoreVectorMaskedNode(control(), memory(addr), addr, val, addr_type, mask));
1293     set_memory(vstore, addr_type);
1294   } else {
1295     Node* vload = nullptr;
1296 
1297     if (mismatched_ms) {
1298       // Reinterpret the vector mask to byte type.
1299       const TypeVect* from_mask_type = TypeVect::makemask(elem_bt, num_elem);
1300       const TypeVect* to_mask_type = TypeVect::makemask(mem_elem_bt, mem_num_elem);
1301       mask = gvn().transform(new VectorReinterpretNode(mask, from_mask_type, to_mask_type));
1302     }
1303 
1304     if (supports_predicate) {
1305       // Generate masked load vector node if predicate feature is supported.
1306       const TypeVect* vt = TypeVect::make(mem_elem_bt, mem_num_elem);
1307       vload = gvn().transform(new LoadVectorMaskedNode(control(), memory(addr), addr, addr_type, vt, mask));
1308     } else {
1309       // Use the vector blend to implement the masked load vector. The biased elements are zeros.
1310       Node* zero = gvn().transform(gvn().zerocon(mem_elem_bt));
1311       zero = gvn().transform(VectorNode::scalar2vector(zero, mem_num_elem, Type::get_const_basic_type(mem_elem_bt)));
1312       vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, mem_num_elem, mem_elem_bt));
1313       vload = gvn().transform(new VectorBlendNode(zero, vload, mask));
1314     }
1315 
1316     if (mismatched_ms) {
1317       const TypeVect* to_vect_type = TypeVect::make(elem_bt, num_elem);
1318       vload = gvn().transform(new VectorReinterpretNode(vload, vload->bottom_type()->is_vect(), to_vect_type));
1319     }
1320 
1321     Node* box = box_vector(vload, vbox_type, elem_bt, num_elem);
1322     set_result(box);
1323   }
1324 
1325   destruct_map_clone(old_map);
1326 
1327   if (can_access_non_heap) {
1328     insert_mem_bar(Op_MemBarCPUOrder);
1329   }
1330 
1331   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt))));
1332   return true;
1333 }
1334 
1335 // <C,
1336 //  V extends Vector<?>,
1337 //  W extends Vector<Integer>,
1338 //  S extends VectorSpecies<E>,
1339 //  M extends VectorMask<E>,
1340 //  E>
1341 // V loadWithMap(Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, int length,
1342 //               Class<? extends Vector<Integer>> vectorIndexClass,
1343 //               Object base, long offset, // Unsafe addressing
1344 //               W index_vector, M m,
1345 //               C container, int index, int[] indexMap, int indexM, S s, // Arguments for default implementation
1346 //               LoadVectorOperationWithMap<C, V, E, S, M> defaultImpl)
1347 //
1348 //  <C,
1349 //   V extends Vector<E>,
1350 //   W extends Vector<Integer>,
1351 //   M extends VectorMask<E>,
1352 //   E>
1353 //  void storeWithMap(Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType,
1354 //                    int length, Class<? extends Vector<Integer>> vectorIndexClass, Object base, long offset,    // Unsafe addressing
1355 //                    W index_vector, V v, M m,
1356 //                    C container, int index, int[] indexMap, int indexM, // Arguments for default implementation
1357 //                    StoreVectorOperationWithMap<C, V, M, E> defaultImpl)
1358 //
1359 bool LibraryCallKit::inline_vector_gather_scatter(bool is_scatter) {
1360   const TypeInstPtr* vector_klass     = gvn().type(argument(0))->isa_instptr();
1361   const TypeInstPtr* mask_klass       = gvn().type(argument(1))->isa_instptr();
1362   const TypeInstPtr* elem_klass       = gvn().type(argument(2))->isa_instptr();
1363   const TypeInt*     vlen             = gvn().type(argument(3))->isa_int();
1364   const TypeInstPtr* vector_idx_klass = gvn().type(argument(4))->isa_instptr();
1365 
1366   if (vector_klass     == nullptr || vector_klass->const_oop()     == nullptr ||
1367 //      mask_klass       == nullptr || mask_klass->const_oop()       == nullptr ||
1368       elem_klass       == nullptr || elem_klass->const_oop()       == nullptr ||
1369       vlen             == nullptr || !vlen->is_con() ||
1370       vector_idx_klass == nullptr || vector_idx_klass->const_oop() == nullptr) {
1371     log_if_needed("  ** missing constant: vclass=%s etype=%s vlen=%s viclass=%s",
1372                     NodeClassNames[argument(0)->Opcode()],
1373                     NodeClassNames[argument(2)->Opcode()],
1374                     NodeClassNames[argument(3)->Opcode()],
1375                     NodeClassNames[argument(4)->Opcode()]);
1376     return false; // not enough info for intrinsification
1377   }
1378 
1379   if (!is_klass_initialized(vector_klass) || !is_klass_initialized(vector_idx_klass)) {
1380     log_if_needed("  ** klass argument not initialized");
1381     return false;
1382   }
1383 
1384   ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type();
1385   if (!elem_type->is_primitive_type()) {
1386     log_if_needed("  ** not a primitive bt=%d", elem_type->basic_type());
1387     return false; // should be primitive type
1388   }
1389 
1390   BasicType elem_bt = elem_type->basic_type();
1391   int num_elem = vlen->get_con();
1392 
1393   const Type* vmask_type = gvn().type(is_scatter ? argument(10) : argument(9));
1394   bool is_masked_op = vmask_type != TypePtr::NULL_PTR;
1395   if (is_masked_op) {
1396     if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) {
1397       log_if_needed("  ** missing constant: maskclass=%s", NodeClassNames[argument(1)->Opcode()]);
1398       return false; // not enough info for intrinsification
1399     }
1400 
1401     if (!is_klass_initialized(mask_klass)) {
1402       log_if_needed("  ** mask klass argument not initialized");
1403       return false;
1404     }
1405 
1406     if (vmask_type->maybe_null()) {
1407       log_if_needed("  ** null mask values are not allowed for masked op");
1408       return false;
1409     }
1410 
1411     // Check whether the predicated gather/scatter node is supported by architecture.
1412     VectorMaskUseType mask = (VectorMaskUseType) (VecMaskUseLoad | VecMaskUsePred);
1413     if (!arch_supports_vector(is_scatter ? Op_StoreVectorScatterMasked : Op_LoadVectorGatherMasked, num_elem, elem_bt, mask)) {
1414       log_if_needed("  ** not supported: arity=%d op=%s vlen=%d etype=%s is_masked_op=1",
1415                       is_scatter, is_scatter ? "scatterMasked" : "gatherMasked",
1416                       num_elem, type2name(elem_bt));
1417       return false; // not supported
1418     }
1419   } else {
1420     // Check whether the normal gather/scatter node is supported for non-masked operation.
1421     if (!arch_supports_vector(is_scatter ? Op_StoreVectorScatter : Op_LoadVectorGather, num_elem, elem_bt, VecMaskNotUsed)) {
1422       log_if_needed("  ** not supported: arity=%d op=%s vlen=%d etype=%s is_masked_op=0",
1423                       is_scatter, is_scatter ? "scatter" : "gather",
1424                       num_elem, type2name(elem_bt));
1425       return false; // not supported
1426     }
1427   }
1428 
1429   // Check that the vector holding indices is supported by architecture
1430   // For sub-word gathers expander receive index array.
1431   if (!is_subword_type(elem_bt) && !arch_supports_vector(Op_LoadVector, num_elem, T_INT, VecMaskNotUsed)) {
1432       log_if_needed("  ** not supported: arity=%d op=%s/loadindex vlen=%d etype=int is_masked_op=%d",
1433                       is_scatter, is_scatter ? "scatter" : "gather",
1434                       num_elem, is_masked_op ? 1 : 0);
1435       return false; // not supported
1436   }
1437 
1438   Node* base = argument(5);
1439   Node* offset = ConvL2X(argument(6));
1440 
1441   // Save state and restore on bailout
1442   uint old_sp = sp();
1443   SafePointNode* old_map = clone_map();
1444 
1445   Node* addr = make_unsafe_address(base, offset, elem_bt, true);
1446 
1447   const TypePtr *addr_type = gvn().type(addr)->isa_ptr();
1448   const TypeAryPtr* arr_type = addr_type->isa_aryptr();
1449 
1450   // The array must be consistent with vector type
1451   if (arr_type == nullptr || (arr_type != nullptr && !elem_consistent_with_arr(elem_bt, arr_type, false))) {
1452     log_if_needed("  ** not supported: arity=%d op=%s vlen=%d etype=%s atype=%s ismask=no",
1453                     is_scatter, is_scatter ? "scatter" : "gather",
1454                     num_elem, type2name(elem_bt), type2name(arr_type->elem()->array_element_basic_type()));
1455     set_map(old_map);
1456     set_sp(old_sp);
1457     return false;
1458   }
1459 
1460   ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass();
1461   const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass);
1462   ciKlass* vbox_idx_klass = vector_idx_klass->const_oop()->as_instance()->java_lang_Class_klass();
1463   if (vbox_idx_klass == nullptr) {
1464     set_map(old_map);
1465     set_sp(old_sp);
1466     return false;
1467   }
1468 
1469   Node* index_vect = nullptr;
1470   const TypeInstPtr* vbox_idx_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_idx_klass);
1471   if (!is_subword_type(elem_bt)) {
1472     index_vect = unbox_vector(argument(8), vbox_idx_type, T_INT, num_elem);
1473     if (index_vect == nullptr) {
1474       set_map(old_map);
1475       set_sp(old_sp);
1476       return false;
1477     }
1478   }
1479 
1480   Node* mask = nullptr;
1481   if (is_masked_op) {
1482     ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass();
1483     const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass);
1484     mask = unbox_vector(is_scatter ? argument(10) : argument(9), mbox_type, elem_bt, num_elem);
1485     if (mask == nullptr) {
1486       log_if_needed("  ** unbox failed mask=%s",
1487                     is_scatter ? NodeClassNames[argument(10)->Opcode()]
1488                                : NodeClassNames[argument(9)->Opcode()]);
1489       set_map(old_map);
1490       set_sp(old_sp);
1491       return false;
1492     }
1493   }
1494 
1495   const TypeVect* vector_type = TypeVect::make(elem_bt, num_elem);
1496   if (is_scatter) {
1497     Node* val = unbox_vector(argument(9), vbox_type, elem_bt, num_elem);
1498     if (val == nullptr) {
1499       set_map(old_map);
1500       set_sp(old_sp);
1501       return false; // operand unboxing failed
1502     }
1503     set_all_memory(reset_memory());
1504 
1505     Node* vstore = nullptr;
1506     if (mask != nullptr) {
1507       vstore = gvn().transform(new StoreVectorScatterMaskedNode(control(), memory(addr), addr, addr_type, val, index_vect, mask));
1508     } else {
1509       vstore = gvn().transform(new StoreVectorScatterNode(control(), memory(addr), addr, addr_type, val, index_vect));
1510     }
1511     set_memory(vstore, addr_type);
1512   } else {
1513     Node* vload = nullptr;
1514     Node* index    = argument(11);
1515     Node* indexMap = argument(12);
1516     Node* indexM   = argument(13);
1517     if (mask != nullptr) {
1518       if (is_subword_type(elem_bt)) {
1519         Node* index_arr_base = array_element_address(indexMap, indexM, T_INT);
1520         vload = gvn().transform(new LoadVectorGatherMaskedNode(control(), memory(addr), addr, addr_type, vector_type, index_arr_base, mask, index));
1521       } else {
1522         vload = gvn().transform(new LoadVectorGatherMaskedNode(control(), memory(addr), addr, addr_type, vector_type, index_vect, mask));
1523       }
1524     } else {
1525       if (is_subword_type(elem_bt)) {
1526         Node* index_arr_base = array_element_address(indexMap, indexM, T_INT);
1527         vload = gvn().transform(new LoadVectorGatherNode(control(), memory(addr), addr, addr_type, vector_type, index_arr_base, index));
1528       } else {
1529         vload = gvn().transform(new LoadVectorGatherNode(control(), memory(addr), addr, addr_type, vector_type, index_vect));
1530       }
1531     }
1532     Node* box = box_vector(vload, vbox_type, elem_bt, num_elem);
1533     set_result(box);
1534   }
1535 
1536   destruct_map_clone(old_map);
1537 
1538   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt))));
1539   return true;
1540 }
1541 
1542 // public static
1543 // <V extends Vector<E>,
1544 //  M extends VectorMask<E>,
1545 //  E>
1546 // long reductionCoerced(int oprId, Class<? extends V> vectorClass, Class<? extends M> maskClass,
1547 //                       Class<E> elementType, int length, V v, M m,
1548 //                       ReductionOperation<V, M> defaultImpl)
1549 bool LibraryCallKit::inline_vector_reduction() {
1550   const TypeInt*     opr          = gvn().type(argument(0))->isa_int();
1551   const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr();
1552   const TypeInstPtr* mask_klass   = gvn().type(argument(2))->isa_instptr();
1553   const TypeInstPtr* elem_klass   = gvn().type(argument(3))->isa_instptr();
1554   const TypeInt*     vlen         = gvn().type(argument(4))->isa_int();
1555 
1556   if (opr          == nullptr || !opr->is_con() ||
1557       vector_klass == nullptr || vector_klass->const_oop() == nullptr ||
1558 //      mask_klass   == nullptr || mask_klass->const_oop()   == nullptr ||
1559       elem_klass   == nullptr || elem_klass->const_oop()   == nullptr ||
1560       vlen         == nullptr || !vlen->is_con()) {
1561     log_if_needed("  ** missing constant: opr=%s vclass=%s etype=%s vlen=%s",
1562                     NodeClassNames[argument(0)->Opcode()],
1563                     NodeClassNames[argument(1)->Opcode()],
1564                     NodeClassNames[argument(3)->Opcode()],
1565                     NodeClassNames[argument(4)->Opcode()]);
1566     return false; // not enough info for intrinsification
1567   }
1568   if (!is_klass_initialized(vector_klass)) {
1569     log_if_needed("  ** klass argument not initialized");
1570     return false;
1571   }
1572   ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type();
1573   if (!elem_type->is_primitive_type()) {
1574     log_if_needed("  ** not a primitive bt=%d", elem_type->basic_type());
1575     return false; // should be primitive type
1576   }
1577 
1578   const Type* vmask_type = gvn().type(argument(6));
1579   bool is_masked_op = vmask_type != TypePtr::NULL_PTR;
1580   if (is_masked_op) {
1581     if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) {
1582       log_if_needed("  ** missing constant: maskclass=%s", NodeClassNames[argument(2)->Opcode()]);
1583       return false; // not enough info for intrinsification
1584     }
1585 
1586     if (!is_klass_initialized(mask_klass)) {
1587       log_if_needed("  ** mask klass argument not initialized");
1588       return false;
1589     }
1590 
1591     if (vmask_type->maybe_null()) {
1592       log_if_needed("  ** null mask values are not allowed for masked op");
1593       return false;
1594     }
1595   }
1596 
1597   BasicType elem_bt = elem_type->basic_type();
1598   int num_elem = vlen->get_con();
1599   int opc  = VectorSupport::vop2ideal(opr->get_con(), elem_bt);
1600   int sopc = ReductionNode::opcode(opc, elem_bt);
1601 
1602   // When using mask, mask use type needs to be VecMaskUseLoad.
1603   if (!arch_supports_vector(sopc, num_elem, elem_bt, is_masked_op ? VecMaskUseLoad : VecMaskNotUsed)) {
1604     log_if_needed("  ** not supported: arity=1 op=%d/reduce vlen=%d etype=%s is_masked_op=%d",
1605                     sopc, num_elem, type2name(elem_bt), is_masked_op ? 1 : 0);
1606     return false;
1607   }
1608 
1609   // Return true if current platform has implemented the masked operation with predicate feature.
1610   bool use_predicate = is_masked_op && arch_supports_vector(sopc, num_elem, elem_bt, VecMaskUsePred);
1611   if (is_masked_op && !use_predicate && !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)) {
1612     log_if_needed("  ** not supported: arity=1 op=%d/reduce vlen=%d etype=%s is_masked_op=1",
1613                     sopc, num_elem, type2name(elem_bt));
1614     return false;
1615   }
1616 
1617   ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass();
1618   const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass);
1619 
1620   Node* opd = unbox_vector(argument(5), vbox_type, elem_bt, num_elem);
1621   if (opd == nullptr) {
1622     return false; // operand unboxing failed
1623   }
1624 
1625   Node* mask = nullptr;
1626   if (is_masked_op) {
1627     ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass();
1628     assert(is_vector_mask(mbox_klass), "argument(2) should be a mask class");
1629     const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass);
1630     mask = unbox_vector(argument(6), mbox_type, elem_bt, num_elem);
1631     if (mask == nullptr) {
1632       log_if_needed("  ** unbox failed mask=%s",
1633                       NodeClassNames[argument(6)->Opcode()]);
1634       return false;
1635     }
1636   }
1637 
1638   Node* init = ReductionNode::make_identity_con_scalar(gvn(), opc, elem_bt);
1639   Node* value = opd;
1640 
1641   assert(mask != nullptr || !is_masked_op, "Masked op needs the mask value never null");
1642   if (mask != nullptr && !use_predicate) {
1643     Node* reduce_identity = gvn().transform(VectorNode::scalar2vector(init, num_elem, Type::get_const_basic_type(elem_bt)));
1644     value = gvn().transform(new VectorBlendNode(reduce_identity, value, mask));
1645   }
1646 
1647   // Make an unordered Reduction node. This affects only AddReductionVF/VD and MulReductionVF/VD,
1648   // as these operations are allowed to be associative (not requiring strict order) in VectorAPI.
1649   value = ReductionNode::make(opc, nullptr, init, value, elem_bt, /* requires_strict_order */ false);
1650 
1651   if (mask != nullptr && use_predicate) {
1652     value->add_req(mask);
1653     value->add_flag(Node::Flag_is_predicated_vector);
1654   }
1655 
1656   value = gvn().transform(value);
1657 
1658   Node* bits = nullptr;
1659   switch (elem_bt) {
1660     case T_BYTE:
1661     case T_SHORT:
1662     case T_INT: {
1663       bits = gvn().transform(new ConvI2LNode(value));
1664       break;
1665     }
1666     case T_FLOAT: {
1667       value = gvn().transform(new MoveF2INode(value));
1668       bits  = gvn().transform(new ConvI2LNode(value));
1669       break;
1670     }
1671     case T_DOUBLE: {
1672       bits = gvn().transform(new MoveD2LNode(value));
1673       break;
1674     }
1675     case T_LONG: {
1676       bits = value; // no conversion needed
1677       break;
1678     }
1679     default: fatal("%s", type2name(elem_bt));
1680   }
1681   set_result(bits);
1682   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt))));
1683   return true;
1684 }
1685 
1686 // public static <V> boolean test(int cond, Class<?> vectorClass, Class<?> elementType, int vlen,
1687 //                                V v1, V v2,
1688 //                                BiFunction<V, V, Boolean> defaultImpl)
1689 //
1690 bool LibraryCallKit::inline_vector_test() {
1691   const TypeInt*     cond         = gvn().type(argument(0))->isa_int();
1692   const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr();
1693   const TypeInstPtr* elem_klass   = gvn().type(argument(2))->isa_instptr();
1694   const TypeInt*     vlen         = gvn().type(argument(3))->isa_int();
1695 
1696   if (cond         == nullptr || !cond->is_con() ||
1697       vector_klass == nullptr || vector_klass->const_oop() == nullptr ||
1698       elem_klass   == nullptr || elem_klass->const_oop()   == nullptr ||
1699       vlen         == nullptr || !vlen->is_con()) {
1700     log_if_needed("  ** missing constant: cond=%s vclass=%s etype=%s vlen=%s",
1701                     NodeClassNames[argument(0)->Opcode()],
1702                     NodeClassNames[argument(1)->Opcode()],
1703                     NodeClassNames[argument(2)->Opcode()],
1704                     NodeClassNames[argument(3)->Opcode()]);
1705     return false; // not enough info for intrinsification
1706   }
1707   if (!is_klass_initialized(vector_klass)) {
1708     log_if_needed("  ** klass argument not initialized");
1709     return false;
1710   }
1711   ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type();
1712   if (!elem_type->is_primitive_type()) {
1713     log_if_needed("  ** not a primitive bt=%d", elem_type->basic_type());
1714     return false; // should be primitive type
1715   }
1716   BasicType elem_bt = elem_type->basic_type();
1717   int num_elem = vlen->get_con();
1718   BoolTest::mask booltest = (BoolTest::mask)cond->get_con();
1719   ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass();
1720   const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass);
1721 
1722   if (!arch_supports_vector(Op_VectorTest, num_elem, elem_bt, is_vector_mask(vbox_klass) ? VecMaskUseLoad : VecMaskNotUsed)) {
1723     log_if_needed("  ** not supported: arity=2 op=test/%d vlen=%d etype=%s ismask=%d",
1724                     cond->get_con(), num_elem, type2name(elem_bt),
1725                     is_vector_mask(vbox_klass));
1726     return false;
1727   }
1728 
1729   Node* opd1 = unbox_vector(argument(4), vbox_type, elem_bt, num_elem);
1730   Node* opd2;
1731   if (Matcher::vectortest_needs_second_argument(booltest == BoolTest::overflow,
1732                                                 opd1->bottom_type()->isa_vectmask())) {
1733     opd2 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem);
1734   } else {
1735     opd2 = opd1;
1736   }
1737   if (opd1 == nullptr || opd2 == nullptr) {
1738     return false; // operand unboxing failed
1739   }
1740 
1741   Node* cmp = gvn().transform(new VectorTestNode(opd1, opd2, booltest));
1742   BoolTest::mask test = Matcher::vectortest_mask(booltest == BoolTest::overflow,
1743                                                  opd1->bottom_type()->isa_vectmask(), num_elem);
1744   Node* bol = gvn().transform(new BoolNode(cmp, test));
1745   Node* res = gvn().transform(new CMoveINode(bol, gvn().intcon(0), gvn().intcon(1), TypeInt::BOOL));
1746 
1747   set_result(res);
1748   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt))));
1749   return true;
1750 }
1751 
1752 // public static
1753 // <V extends Vector<E>,
1754 //  M extends VectorMask<E>,
1755 //  E>
1756 // V blend(Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, int vlen,
1757 //         V v1, V v2, M m,
1758 //         VectorBlendOp<V, M, E> defaultImpl)
1759 bool LibraryCallKit::inline_vector_blend() {
1760   const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr();
1761   const TypeInstPtr* mask_klass   = gvn().type(argument(1))->isa_instptr();
1762   const TypeInstPtr* elem_klass   = gvn().type(argument(2))->isa_instptr();
1763   const TypeInt*     vlen         = gvn().type(argument(3))->isa_int();
1764 
1765   if (mask_klass == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr) {
1766     return false; // dead code
1767   }
1768   if (mask_klass->const_oop() == nullptr || vector_klass->const_oop() == nullptr ||
1769       elem_klass->const_oop() == nullptr || !vlen->is_con()) {
1770     log_if_needed("  ** missing constant: vclass=%s mclass=%s etype=%s vlen=%s",
1771                     NodeClassNames[argument(0)->Opcode()],
1772                     NodeClassNames[argument(1)->Opcode()],
1773                     NodeClassNames[argument(2)->Opcode()],
1774                     NodeClassNames[argument(3)->Opcode()]);
1775     return false; // not enough info for intrinsification
1776   }
1777   if (!is_klass_initialized(vector_klass) || !is_klass_initialized(mask_klass)) {
1778     log_if_needed("  ** klass argument not initialized");
1779     return false;
1780   }
1781   ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type();
1782   if (!elem_type->is_primitive_type()) {
1783     log_if_needed("  ** not a primitive bt=%d", elem_type->basic_type());
1784     return false; // should be primitive type
1785   }
1786   BasicType elem_bt = elem_type->basic_type();
1787   BasicType mask_bt = elem_bt;
1788   int num_elem = vlen->get_con();
1789 
1790   if (!arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)) {
1791     log_if_needed("  ** not supported: arity=2 op=blend vlen=%d etype=%s ismask=useload",
1792                     num_elem, type2name(elem_bt));
1793     return false; // not supported
1794   }
1795   ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass();
1796   const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass);
1797 
1798   ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass();
1799   const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass);
1800 
1801   Node* v1   = unbox_vector(argument(4), vbox_type, elem_bt, num_elem);
1802   Node* v2   = unbox_vector(argument(5), vbox_type, elem_bt, num_elem);
1803   Node* mask = unbox_vector(argument(6), mbox_type, mask_bt, num_elem);
1804 
1805   if (v1 == nullptr || v2 == nullptr || mask == nullptr) {
1806     return false; // operand unboxing failed
1807   }
1808 
1809   Node* blend = gvn().transform(new VectorBlendNode(v1, v2, mask));
1810 
1811   Node* box = box_vector(blend, vbox_type, elem_bt, num_elem);
1812   set_result(box);
1813   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt))));
1814   return true;
1815 }
1816 
1817 //  public static
1818 //  <V extends Vector<E>,
1819 //   M extends VectorMask<E>,
1820 //   E>
1821 //  M compare(int cond, Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, int vlen,
1822 //            V v1, V v2, M m,
1823 //            VectorCompareOp<V,M> defaultImpl)
1824 bool LibraryCallKit::inline_vector_compare() {
1825   const TypeInt*     cond         = gvn().type(argument(0))->isa_int();
1826   const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr();
1827   const TypeInstPtr* mask_klass   = gvn().type(argument(2))->isa_instptr();
1828   const TypeInstPtr* elem_klass   = gvn().type(argument(3))->isa_instptr();
1829   const TypeInt*     vlen         = gvn().type(argument(4))->isa_int();
1830 
1831   if (cond == nullptr || vector_klass == nullptr || mask_klass == nullptr || elem_klass == nullptr || vlen == nullptr) {
1832     return false; // dead code
1833   }
1834   if (!cond->is_con() || vector_klass->const_oop() == nullptr || mask_klass->const_oop() == nullptr ||
1835       elem_klass->const_oop() == nullptr || !vlen->is_con()) {
1836     log_if_needed("  ** missing constant: cond=%s vclass=%s mclass=%s etype=%s vlen=%s",
1837                     NodeClassNames[argument(0)->Opcode()],
1838                     NodeClassNames[argument(1)->Opcode()],
1839                     NodeClassNames[argument(2)->Opcode()],
1840                     NodeClassNames[argument(3)->Opcode()],
1841                     NodeClassNames[argument(4)->Opcode()]);
1842     return false; // not enough info for intrinsification
1843   }
1844   if (!is_klass_initialized(vector_klass) || !is_klass_initialized(mask_klass)) {
1845     log_if_needed("  ** klass argument not initialized");
1846     return false;
1847   }
1848   ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type();
1849   if (!elem_type->is_primitive_type()) {
1850     log_if_needed("  ** not a primitive bt=%d", elem_type->basic_type());
1851     return false; // should be primitive type
1852   }
1853 
1854   int num_elem = vlen->get_con();
1855   BasicType elem_bt = elem_type->basic_type();
1856   BasicType mask_bt = elem_bt;
1857 
1858   if ((cond->get_con() & BoolTest::unsigned_compare) != 0) {
1859     if (!Matcher::supports_vector_comparison_unsigned(num_elem, elem_bt)) {
1860       log_if_needed("  ** not supported: unsigned comparison op=comp/%d vlen=%d etype=%s ismask=usestore",
1861                       cond->get_con() & (BoolTest::unsigned_compare - 1), num_elem, type2name(elem_bt));
1862       return false;
1863     }
1864   }
1865 
1866   if (!arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskUseStore)) {
1867     log_if_needed("  ** not supported: arity=2 op=comp/%d vlen=%d etype=%s ismask=usestore",
1868                     cond->get_con(), num_elem, type2name(elem_bt));
1869     return false;
1870   }
1871 
1872   ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass();
1873   const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass);
1874 
1875   ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass();
1876   const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass);
1877 
1878   Node* v1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem);
1879   Node* v2 = unbox_vector(argument(6), vbox_type, elem_bt, num_elem);
1880 
1881   bool is_masked_op = argument(7)->bottom_type() != TypePtr::NULL_PTR;
1882   Node* mask = is_masked_op ? unbox_vector(argument(7), mbox_type, elem_bt, num_elem) : nullptr;
1883   if (is_masked_op && mask == nullptr) {
1884     log_if_needed("  ** not supported: mask = null arity=2 op=comp/%d vlen=%d etype=%s ismask=usestore is_masked_op=1",
1885                     cond->get_con(), num_elem, type2name(elem_bt));
1886     return false;
1887   }
1888 
1889   bool use_predicate = is_masked_op && arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskUsePred);
1890   if (is_masked_op && !use_predicate && !arch_supports_vector(Op_AndV, num_elem, elem_bt, VecMaskUseLoad)) {
1891     log_if_needed("  ** not supported: arity=2 op=comp/%d vlen=%d etype=%s ismask=usestore is_masked_op=1",
1892                     cond->get_con(), num_elem, type2name(elem_bt));
1893     return false;
1894   }
1895 
1896   if (v1 == nullptr || v2 == nullptr) {
1897     return false; // operand unboxing failed
1898   }
1899   BoolTest::mask pred = (BoolTest::mask)cond->get_con();
1900   ConINode* pred_node = (ConINode*)gvn().makecon(cond);
1901 
1902   const TypeVect* vmask_type = TypeVect::makemask(mask_bt, num_elem);
1903   Node* operation = new VectorMaskCmpNode(pred, v1, v2, pred_node, vmask_type);
1904 
1905   if (is_masked_op) {
1906     if (use_predicate) {
1907       operation->add_req(mask);
1908       operation->add_flag(Node::Flag_is_predicated_vector);
1909     } else {
1910       operation = gvn().transform(operation);
1911       operation = VectorNode::make(Op_AndV, operation, mask, vmask_type);
1912     }
1913   }
1914 
1915   operation = gvn().transform(operation);
1916 
1917   Node* box = box_vector(operation, mbox_type, mask_bt, num_elem);
1918   set_result(box);
1919   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt))));
1920   return true;
1921 }
1922 
1923 // public static
1924 // <V extends Vector<E>,
1925 //  Sh extends VectorShuffle<E>,
1926 //  M extends VectorMask<E>,
1927 //  E>
1928 // V rearrangeOp(Class<? extends V> vectorClass, Class<Sh> shuffleClass, Class<M> maskClass, Class<E> elementType, int vlen,
1929 //               V v1, Sh sh, M m,
1930 //               VectorRearrangeOp<V, Sh, M, E> defaultImpl)
1931 bool LibraryCallKit::inline_vector_rearrange() {
1932   const TypeInstPtr* vector_klass  = gvn().type(argument(0))->isa_instptr();
1933   const TypeInstPtr* shuffle_klass = gvn().type(argument(1))->isa_instptr();
1934   const TypeInstPtr* mask_klass    = gvn().type(argument(2))->isa_instptr();
1935   const TypeInstPtr* elem_klass    = gvn().type(argument(3))->isa_instptr();
1936   const TypeInt*     vlen          = gvn().type(argument(4))->isa_int();
1937 
1938   if (vector_klass == nullptr  || shuffle_klass == nullptr ||  elem_klass == nullptr || vlen == nullptr) {
1939     return false; // dead code
1940   }
1941   if (shuffle_klass->const_oop() == nullptr ||
1942       vector_klass->const_oop()  == nullptr ||
1943       elem_klass->const_oop()    == nullptr ||
1944       !vlen->is_con()) {
1945     log_if_needed("  ** missing constant: vclass=%s sclass=%s etype=%s vlen=%s",
1946                     NodeClassNames[argument(0)->Opcode()],
1947                     NodeClassNames[argument(1)->Opcode()],
1948                     NodeClassNames[argument(3)->Opcode()],
1949                     NodeClassNames[argument(4)->Opcode()]);
1950     return false; // not enough info for intrinsification
1951   }
1952   if (!is_klass_initialized(vector_klass)  ||
1953       !is_klass_initialized(shuffle_klass)) {
1954     log_if_needed("  ** klass argument not initialized");
1955     return false;
1956   }
1957   ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type();
1958   if (!elem_type->is_primitive_type()) {
1959     log_if_needed("  ** not a primitive bt=%d", elem_type->basic_type());
1960     return false; // should be primitive type
1961   }
1962   BasicType elem_bt = elem_type->basic_type();
1963   BasicType shuffle_bt = elem_bt;
1964   int num_elem = vlen->get_con();
1965 
1966   if (!arch_supports_vector(Op_VectorLoadShuffle, num_elem, elem_bt, VecMaskNotUsed)) {
1967     log_if_needed("  ** not supported: arity=0 op=load/shuffle vlen=%d etype=%s ismask=no",
1968                     num_elem, type2name(elem_bt));
1969     return false; // not supported
1970   }
1971 
1972   bool is_masked_op = argument(7)->bottom_type() != TypePtr::NULL_PTR;
1973   bool use_predicate = is_masked_op;
1974   if (is_masked_op &&
1975       (mask_klass == nullptr ||
1976        mask_klass->const_oop() == nullptr ||
1977        !is_klass_initialized(mask_klass))) {
1978     log_if_needed("  ** mask_klass argument not initialized");
1979   }
1980   VectorMaskUseType checkFlags = (VectorMaskUseType)(is_masked_op ? (VecMaskUseLoad | VecMaskUsePred) : VecMaskNotUsed);
1981   if (!arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, checkFlags)) {
1982     use_predicate = false;
1983     if(!is_masked_op ||
1984        (!arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, VecMaskNotUsed) ||
1985         !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)     ||
1986         !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed))) {
1987       log_if_needed("  ** not supported: arity=2 op=shuffle/rearrange vlen=%d etype=%s ismask=no",
1988                       num_elem, type2name(elem_bt));
1989       return false; // not supported
1990     }
1991   }
1992   ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass();
1993   const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass);
1994 
1995   ciKlass* shbox_klass = shuffle_klass->const_oop()->as_instance()->java_lang_Class_klass();
1996   const TypeInstPtr* shbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, shbox_klass);
1997 
1998   Node* v1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem);
1999   Node* shuffle = unbox_vector(argument(6), shbox_type, shuffle_bt, num_elem);
2000 
2001   if (v1 == nullptr || shuffle == nullptr) {
2002     return false; // operand unboxing failed
2003   }
2004 
2005   Node* mask = nullptr;
2006   if (is_masked_op) {
2007     ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass();
2008     const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass);
2009     mask = unbox_vector(argument(7), mbox_type, elem_bt, num_elem);
2010     if (mask == nullptr) {
2011       log_if_needed("  ** not supported: arity=3 op=shuffle/rearrange vlen=%d etype=%s ismask=useload is_masked_op=1",
2012                       num_elem, type2name(elem_bt));
2013       return false;
2014     }
2015   }
2016 
2017   Node* rearrange = new VectorRearrangeNode(v1, shuffle);
2018   if (is_masked_op) {
2019     if (use_predicate) {
2020       rearrange->add_req(mask);
2021       rearrange->add_flag(Node::Flag_is_predicated_vector);
2022     } else {
2023       const TypeVect* vt = v1->bottom_type()->is_vect();
2024       rearrange = gvn().transform(rearrange);
2025       Node* zero = gvn().makecon(Type::get_zero_type(elem_bt));
2026       Node* zerovec = gvn().transform(VectorNode::scalar2vector(zero, num_elem, Type::get_const_basic_type(elem_bt)));
2027       rearrange = new VectorBlendNode(zerovec, rearrange, mask);
2028     }
2029   }
2030   rearrange = gvn().transform(rearrange);
2031 
2032   Node* box = box_vector(rearrange, vbox_type, elem_bt, num_elem);
2033   set_result(box);
2034   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt))));
2035   return true;
2036 }
2037 
2038 static address get_svml_address(int vop, int bits, BasicType bt, char* name_ptr, int name_len) {
2039   address addr = nullptr;
2040   assert(UseVectorStubs, "sanity");
2041   assert(name_ptr != nullptr, "unexpected");
2042   assert((vop >= VectorSupport::VECTOR_OP_SVML_START) && (vop <= VectorSupport::VECTOR_OP_SVML_END), "unexpected");
2043   int op = vop - VectorSupport::VECTOR_OP_SVML_START;
2044 
2045   switch(bits) {
2046     case 64:  //fallthough
2047     case 128: //fallthough
2048     case 256: //fallthough
2049     case 512:
2050       if (bt == T_FLOAT) {
2051         snprintf(name_ptr, name_len, "vector_%s_float%d", VectorSupport::svmlname[op], bits);
2052         addr = StubRoutines::_vector_f_math[exact_log2(bits/64)][op];
2053       } else {
2054         assert(bt == T_DOUBLE, "must be FP type only");
2055         snprintf(name_ptr, name_len, "vector_%s_double%d", VectorSupport::svmlname[op], bits);
2056         addr = StubRoutines::_vector_d_math[exact_log2(bits/64)][op];
2057       }
2058       break;
2059     default:
2060       snprintf(name_ptr, name_len, "invalid");
2061       addr = nullptr;
2062       Unimplemented();
2063       break;
2064   }
2065 
2066   return addr;
2067 }
2068 
2069 Node* LibraryCallKit::gen_call_to_svml(int vector_api_op_id, BasicType bt, int num_elem, Node* opd1, Node* opd2) {
2070   assert(UseVectorStubs, "sanity");
2071   assert(vector_api_op_id >= VectorSupport::VECTOR_OP_SVML_START && vector_api_op_id <= VectorSupport::VECTOR_OP_SVML_END, "need valid op id");
2072   assert(opd1 != nullptr, "must not be null");
2073   const TypeVect* vt = TypeVect::make(bt, num_elem);
2074   const TypeFunc* call_type = OptoRuntime::Math_Vector_Vector_Type(opd2 != nullptr ? 2 : 1, vt, vt);
2075   char name[100] = "";
2076 
2077   // Get address for svml method.
2078   address addr = get_svml_address(vector_api_op_id, vt->length_in_bytes() * BitsPerByte, bt, name, 100);
2079 
2080   if (addr == nullptr) {
2081     return nullptr;
2082   }
2083 
2084   assert(name[0] != '\0', "name must not be null");
2085   Node* operation = make_runtime_call(RC_VECTOR,
2086                                       call_type,
2087                                       addr,
2088                                       name,
2089                                       TypePtr::BOTTOM,
2090                                       opd1,
2091                                       opd2);
2092   return gvn().transform(new ProjNode(gvn().transform(operation), TypeFunc::Parms));
2093 }
2094 
2095 //  public static
2096 //  <V extends Vector<E>,
2097 //   M extends VectorMask<E>,
2098 //   E>
2099 //  V broadcastInt(int opr, Class<? extends V> vectorClass, Class<? extends M> maskClass,
2100 //                 Class<E> elementType, int length,
2101 //                 V v, int n, M m,
2102 //                 VectorBroadcastIntOp<V, M> defaultImpl)
2103 bool LibraryCallKit::inline_vector_broadcast_int() {
2104   const TypeInt*     opr          = gvn().type(argument(0))->isa_int();
2105   const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr();
2106   const TypeInstPtr* mask_klass   = gvn().type(argument(2))->isa_instptr();
2107   const TypeInstPtr* elem_klass   = gvn().type(argument(3))->isa_instptr();
2108   const TypeInt*     vlen         = gvn().type(argument(4))->isa_int();
2109 
2110   if (opr == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr) {
2111     return false; // dead code
2112   }
2113   if (!opr->is_con() || vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) {
2114     log_if_needed("  ** missing constant: opr=%s vclass=%s etype=%s vlen=%s",
2115                     NodeClassNames[argument(0)->Opcode()],
2116                     NodeClassNames[argument(1)->Opcode()],
2117                     NodeClassNames[argument(3)->Opcode()],
2118                     NodeClassNames[argument(4)->Opcode()]);
2119     return false; // not enough info for intrinsification
2120   }
2121   if (!is_klass_initialized(vector_klass)) {
2122     log_if_needed("  ** klass argument not initialized");
2123     return false;
2124   }
2125 
2126   const Type* vmask_type = gvn().type(argument(7));
2127   bool is_masked_op = vmask_type != TypePtr::NULL_PTR;
2128   if (is_masked_op) {
2129     if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) {
2130       log_if_needed("  ** missing constant: maskclass=%s", NodeClassNames[argument(2)->Opcode()]);
2131       return false; // not enough info for intrinsification
2132     }
2133 
2134     if (!is_klass_initialized(mask_klass)) {
2135       log_if_needed("  ** mask klass argument not initialized");
2136       return false;
2137     }
2138 
2139     if (vmask_type->maybe_null()) {
2140       log_if_needed("  ** null mask values are not allowed for masked op");
2141       return false;
2142     }
2143   }
2144 
2145   ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type();
2146   if (!elem_type->is_primitive_type()) {
2147     log_if_needed("  ** not a primitive bt=%d", elem_type->basic_type());
2148     return false; // should be primitive type
2149   }
2150 
2151   int num_elem = vlen->get_con();
2152   BasicType elem_bt = elem_type->basic_type();
2153   int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt);
2154 
2155   bool is_shift  = VectorNode::is_shift_opcode(opc);
2156   bool is_rotate = VectorNode::is_rotate_opcode(opc);
2157 
2158   if (opc == 0 || (!is_shift && !is_rotate)) {
2159     log_if_needed("  ** operation not supported: op=%d bt=%s", opr->get_con(), type2name(elem_bt));
2160     return false; // operation not supported
2161   }
2162 
2163   int sopc = VectorNode::opcode(opc, elem_bt);
2164   if (sopc == 0) {
2165     log_if_needed("  ** operation not supported: opc=%s bt=%s", NodeClassNames[opc], type2name(elem_bt));
2166     return false; // operation not supported
2167   }
2168 
2169   Node* cnt  = argument(6);
2170   ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass();
2171   const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass);
2172   const TypeInt* cnt_type = cnt->bottom_type()->isa_int();
2173 
2174   // If CPU supports vector constant rotate instructions pass it directly
2175   bool is_const_rotate = is_rotate && cnt_type && cnt_type->is_con() &&
2176                          Matcher::supports_vector_constant_rotates(cnt_type->get_con());
2177   bool has_scalar_args = is_rotate ? !is_const_rotate : true;
2178 
2179   VectorMaskUseType checkFlags = (VectorMaskUseType)(is_masked_op ? (VecMaskUseLoad | VecMaskUsePred) : VecMaskNotUsed);
2180   bool use_predicate = is_masked_op;
2181 
2182   if (!arch_supports_vector(sopc, num_elem, elem_bt, checkFlags, has_scalar_args)) {
2183     use_predicate = false;
2184     if (!is_masked_op ||
2185         (!arch_supports_vector(sopc, num_elem, elem_bt, VecMaskNotUsed, has_scalar_args) ||
2186          !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad))) {
2187 
2188       log_if_needed("  ** not supported: arity=0 op=int/%d vlen=%d etype=%s is_masked_op=%d",
2189                       sopc, num_elem, type2name(elem_bt), is_masked_op ? 1 : 0);
2190       return false; // not supported
2191     }
2192   }
2193 
2194   Node* opd1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem);
2195   Node* opd2 = nullptr;
2196   if (is_shift) {
2197     opd2 = vector_shift_count(cnt, opc, elem_bt, num_elem);
2198   } else {
2199     assert(is_rotate, "unexpected operation");
2200     if (!is_const_rotate) {
2201       const Type * type_bt = Type::get_const_basic_type(elem_bt);
2202       cnt = elem_bt == T_LONG ? gvn().transform(new ConvI2LNode(cnt)) : cnt;
2203       opd2 = gvn().transform(VectorNode::scalar2vector(cnt, num_elem, type_bt));
2204     } else {
2205       // Constant shift value.
2206       opd2 = cnt;
2207     }
2208   }
2209 
2210   if (opd1 == nullptr || opd2 == nullptr) {
2211     return false;
2212   }
2213 
2214   Node* mask = nullptr;
2215   if (is_masked_op) {
2216     ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass();
2217     const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass);
2218     mask = unbox_vector(argument(7), mbox_type, elem_bt, num_elem);
2219     if (mask == nullptr) {
2220       log_if_needed("  ** unbox failed mask=%s", NodeClassNames[argument(7)->Opcode()]);
2221       return false;
2222     }
2223   }
2224 
2225   Node* operation = VectorNode::make(opc, opd1, opd2, num_elem, elem_bt);
2226   if (is_masked_op && mask != nullptr) {
2227     if (use_predicate) {
2228       operation->add_req(mask);
2229       operation->add_flag(Node::Flag_is_predicated_vector);
2230     } else {
2231       operation = gvn().transform(operation);
2232       operation = new VectorBlendNode(opd1, operation, mask);
2233     }
2234   }
2235   operation = gvn().transform(operation);
2236   Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem);
2237   set_result(vbox);
2238   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt))));
2239   return true;
2240 }
2241 
2242 // public static <VOUT extends VectorPayload,
2243 //                 VIN extends VectorPayload,
2244 //                   S extends VectorSpecies>
2245 // VOUT convert(int oprId,
2246 //           Class<?> fromVectorClass, Class<?> fromElementType, int fromVLen,
2247 //           Class<?>   toVectorClass, Class<?>   toElementType, int   toVLen,
2248 //           VIN v, S s,
2249 //           VectorConvertOp<VOUT, VIN, S> defaultImpl)
2250 //
2251 bool LibraryCallKit::inline_vector_convert() {
2252   const TypeInt*     opr               = gvn().type(argument(0))->isa_int();
2253 
2254   const TypeInstPtr* vector_klass_from = gvn().type(argument(1))->isa_instptr();
2255   const TypeInstPtr* elem_klass_from   = gvn().type(argument(2))->isa_instptr();
2256   const TypeInt*     vlen_from         = gvn().type(argument(3))->isa_int();
2257 
2258   const TypeInstPtr* vector_klass_to   = gvn().type(argument(4))->isa_instptr();
2259   const TypeInstPtr* elem_klass_to     = gvn().type(argument(5))->isa_instptr();
2260   const TypeInt*     vlen_to           = gvn().type(argument(6))->isa_int();
2261 
2262   if (opr == nullptr ||
2263       vector_klass_from == nullptr || elem_klass_from == nullptr || vlen_from == nullptr ||
2264       vector_klass_to   == nullptr || elem_klass_to   == nullptr || vlen_to   == nullptr) {
2265     return false; // dead code
2266   }
2267   if (!opr->is_con() ||
2268       vector_klass_from->const_oop() == nullptr || elem_klass_from->const_oop() == nullptr || !vlen_from->is_con() ||
2269       vector_klass_to->const_oop() == nullptr || elem_klass_to->const_oop() == nullptr || !vlen_to->is_con()) {
2270     log_if_needed("  ** missing constant: opr=%s vclass_from=%s etype_from=%s vlen_from=%s vclass_to=%s etype_to=%s vlen_to=%s",
2271                     NodeClassNames[argument(0)->Opcode()],
2272                     NodeClassNames[argument(1)->Opcode()],
2273                     NodeClassNames[argument(2)->Opcode()],
2274                     NodeClassNames[argument(3)->Opcode()],
2275                     NodeClassNames[argument(4)->Opcode()],
2276                     NodeClassNames[argument(5)->Opcode()],
2277                     NodeClassNames[argument(6)->Opcode()]);
2278     return false; // not enough info for intrinsification
2279   }
2280   if (!is_klass_initialized(vector_klass_from) || !is_klass_initialized(vector_klass_to)) {
2281     log_if_needed("  ** klass argument not initialized");
2282     return false;
2283   }
2284 
2285   assert(opr->get_con() == VectorSupport::VECTOR_OP_CAST  ||
2286          opr->get_con() == VectorSupport::VECTOR_OP_UCAST ||
2287          opr->get_con() == VectorSupport::VECTOR_OP_REINTERPRET, "wrong opcode");
2288   bool is_cast = (opr->get_con() == VectorSupport::VECTOR_OP_CAST || opr->get_con() == VectorSupport::VECTOR_OP_UCAST);
2289   bool is_ucast = (opr->get_con() == VectorSupport::VECTOR_OP_UCAST);
2290 
2291   ciKlass* vbox_klass_from = vector_klass_from->const_oop()->as_instance()->java_lang_Class_klass();
2292   ciKlass* vbox_klass_to = vector_klass_to->const_oop()->as_instance()->java_lang_Class_klass();
2293   if (is_vector_shuffle(vbox_klass_from)) {
2294     return false; // vector shuffles aren't supported
2295   }
2296   bool is_mask = is_vector_mask(vbox_klass_from);
2297 
2298   ciType* elem_type_from = elem_klass_from->const_oop()->as_instance()->java_mirror_type();
2299   if (!elem_type_from->is_primitive_type()) {
2300     return false; // should be primitive type
2301   }
2302   BasicType elem_bt_from = elem_type_from->basic_type();
2303   ciType* elem_type_to = elem_klass_to->const_oop()->as_instance()->java_mirror_type();
2304   if (!elem_type_to->is_primitive_type()) {
2305     return false; // should be primitive type
2306   }
2307   BasicType elem_bt_to = elem_type_to->basic_type();
2308 
2309   int num_elem_from = vlen_from->get_con();
2310   int num_elem_to = vlen_to->get_con();
2311 
2312   // Check whether we can unbox to appropriate size. Even with casting, checking for reinterpret is needed
2313   // since we may need to change size.
2314   if (!arch_supports_vector(Op_VectorReinterpret,
2315                             num_elem_from,
2316                             elem_bt_from,
2317                             is_mask ? VecMaskUseAll : VecMaskNotUsed)) {
2318     log_if_needed("  ** not supported: arity=1 op=%s/1 vlen1=%d etype1=%s ismask=%d",
2319                     is_cast ? "cast" : "reinterpret",
2320                     num_elem_from, type2name(elem_bt_from), is_mask);
2321     return false;
2322   }
2323 
2324   // Check whether we can support resizing/reinterpreting to the new size.
2325   if (!arch_supports_vector(Op_VectorReinterpret,
2326                             num_elem_to,
2327                             elem_bt_to,
2328                             is_mask ? VecMaskUseAll : VecMaskNotUsed)) {
2329     log_if_needed("  ** not supported: arity=1 op=%s/2 vlen2=%d etype2=%s ismask=%d",
2330                     is_cast ? "cast" : "reinterpret",
2331                     num_elem_to, type2name(elem_bt_to), is_mask);
2332     return false;
2333   }
2334 
2335 
2336   if (is_vector_shuffle(vbox_klass_to) &&
2337       (!arch_supports_vector(Op_SubVB, num_elem_to, elem_bt_to, VecMaskNotUsed)           ||
2338        !arch_supports_vector(Op_VectorBlend, num_elem_to, elem_bt_to, VecMaskNotUsed)     ||
2339        !arch_supports_vector(Op_VectorMaskCmp, num_elem_to, elem_bt_to, VecMaskNotUsed)   ||
2340        !arch_supports_vector(Op_AndV, num_elem_to, elem_bt_to, VecMaskNotUsed)            ||
2341        !arch_supports_vector(Op_Replicate, num_elem_to, elem_bt_to, VecMaskNotUsed))) {
2342     log_if_needed("  ** not supported: arity=1 op=shuffle_index_wrap vlen2=%d etype2=%s",
2343                     num_elem_to, type2name(elem_bt_to));
2344     return false;
2345   }
2346 
2347   // At this point, we know that both input and output vector registers are supported
2348   // by the architecture. Next check if the casted type is simply to same type - which means
2349   // that it is actually a resize and not a cast.
2350   if (is_cast && elem_bt_from == elem_bt_to) {
2351     is_cast = false;
2352   }
2353 
2354   const TypeInstPtr* vbox_type_from = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass_from);
2355 
2356   Node* opd1 = unbox_vector(argument(7), vbox_type_from, elem_bt_from, num_elem_from);
2357   if (opd1 == nullptr) {
2358     return false;
2359   }
2360 
2361   const TypeVect* src_type = TypeVect::make(elem_bt_from, num_elem_from, is_mask);
2362   const TypeVect* dst_type = TypeVect::make(elem_bt_to, num_elem_to, is_mask);
2363 
2364   // Safety check to prevent casting if source mask is of type vector
2365   // and destination mask of type predicate vector and vice-versa.
2366   // From X86 standpoint, this case will only arise over KNL target,
2367   // where certain masks (depending on the species) are either propagated
2368   // through a vector or predicate register.
2369   if (is_mask &&
2370       ((src_type->isa_vectmask() == nullptr && dst_type->isa_vectmask()) ||
2371        (dst_type->isa_vectmask() == nullptr && src_type->isa_vectmask()))) {
2372     return false;
2373   }
2374 
2375   Node* op = opd1;
2376   if (is_cast) {
2377     assert(!is_mask || num_elem_from == num_elem_to, "vector mask cast needs the same elem num");
2378     int cast_vopc = VectorCastNode::opcode(-1, elem_bt_from, !is_ucast);
2379 
2380     // Make sure that vector cast is implemented to particular type/size combination if it is
2381     // not a mask casting.
2382     if (!is_mask && !arch_supports_vector(cast_vopc, num_elem_to, elem_bt_to, VecMaskNotUsed)) {
2383       log_if_needed("  ** not supported: arity=1 op=cast#%d/3 vlen2=%d etype2=%s ismask=%d",
2384                       cast_vopc, num_elem_to, type2name(elem_bt_to), is_mask);
2385       return false;
2386     }
2387 
2388     if (num_elem_from < num_elem_to) {
2389       // Since input and output number of elements are not consistent, we need to make sure we
2390       // properly size. Thus, first make a cast that retains the number of elements from source.
2391       int num_elem_for_cast = num_elem_from;
2392 
2393       // It is possible that arch does not support this intermediate vector size
2394       // TODO More complex logic required here to handle this corner case for the sizes.
2395       if (!arch_supports_vector(cast_vopc, num_elem_for_cast, elem_bt_to, VecMaskNotUsed)) {
2396         log_if_needed("  ** not supported: arity=1 op=cast#%d/4 vlen1=%d etype2=%s ismask=%d",
2397                         cast_vopc,
2398                         num_elem_for_cast, type2name(elem_bt_to), is_mask);
2399         return false;
2400       }
2401 
2402       op = gvn().transform(VectorCastNode::make(cast_vopc, op, elem_bt_to, num_elem_for_cast));
2403       // Now ensure that the destination gets properly resized to needed size.
2404       op = gvn().transform(new VectorReinterpretNode(op, op->bottom_type()->is_vect(), dst_type));
2405     } else if (num_elem_from > num_elem_to) {
2406       // Since number of elements from input is larger than output, simply reduce size of input
2407       // (we are supposed to drop top elements anyway).
2408       int num_elem_for_resize = num_elem_to;
2409 
2410       // It is possible that arch does not support this intermediate vector size
2411       // TODO More complex logic required here to handle this corner case for the sizes.
2412       if (!arch_supports_vector(Op_VectorReinterpret,
2413                                 num_elem_for_resize,
2414                                 elem_bt_from,
2415                                 VecMaskNotUsed)) {
2416         log_if_needed("  ** not supported: arity=1 op=cast/5 vlen2=%d etype1=%s ismask=%d",
2417                         num_elem_for_resize, type2name(elem_bt_from), is_mask);
2418         return false;
2419       }
2420 
2421       const TypeVect* resize_type = TypeVect::make(elem_bt_from, num_elem_for_resize);
2422       op = gvn().transform(new VectorReinterpretNode(op, src_type, resize_type));
2423       op = gvn().transform(VectorCastNode::make(cast_vopc, op, elem_bt_to, num_elem_to));
2424     } else { // num_elem_from == num_elem_to
2425       if (is_mask) {
2426         // Make sure that cast for vector mask is implemented to particular type/size combination.
2427         if (!arch_supports_vector(Op_VectorMaskCast, num_elem_to, elem_bt_to, VecMaskNotUsed)) {
2428           log_if_needed("  ** not supported: arity=1 op=maskcast vlen2=%d etype2=%s ismask=%d",
2429                           num_elem_to, type2name(elem_bt_to), is_mask);
2430           return false;
2431         }
2432         op = gvn().transform(new VectorMaskCastNode(op, dst_type));
2433       } else {
2434         // Since input and output number of elements match, and since we know this vector size is
2435         // supported, simply do a cast with no resize needed.
2436         op = gvn().transform(VectorCastNode::make(cast_vopc, op, elem_bt_to, num_elem_to));
2437       }
2438     }
2439   } else if (!Type::equals(src_type, dst_type)) {
2440     assert(!is_cast, "must be reinterpret");
2441     op = gvn().transform(new VectorReinterpretNode(op, src_type, dst_type));
2442   }
2443 
2444   if (is_vector_shuffle(vbox_klass_to)) {
2445      op = partially_wrap_indexes(op, num_elem_to, elem_bt_to);
2446   }
2447 
2448   const TypeInstPtr* vbox_type_to = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass_to);
2449   Node* vbox = box_vector(op, vbox_type_to, elem_bt_to, num_elem_to);
2450   set_result(vbox);
2451   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem_to * type2aelembytes(elem_bt_to))));
2452   return true;
2453 }
2454 
2455 //  public static
2456 //  <V extends Vector<E>,
2457 //   E>
2458 //  V insert(Class<? extends V> vectorClass, Class<E> elementType, int vlen,
2459 //           V vec, int ix, long val,
2460 //           VecInsertOp<V> defaultImpl)
2461 bool LibraryCallKit::inline_vector_insert() {
2462   const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr();
2463   const TypeInstPtr* elem_klass   = gvn().type(argument(1))->isa_instptr();
2464   const TypeInt*     vlen         = gvn().type(argument(2))->isa_int();
2465   const TypeInt*     idx          = gvn().type(argument(4))->isa_int();
2466 
2467   if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || idx == nullptr) {
2468     return false; // dead code
2469   }
2470   if (vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con() || !idx->is_con()) {
2471     log_if_needed("  ** missing constant: vclass=%s etype=%s vlen=%s idx=%s",
2472                     NodeClassNames[argument(0)->Opcode()],
2473                     NodeClassNames[argument(1)->Opcode()],
2474                     NodeClassNames[argument(2)->Opcode()],
2475                     NodeClassNames[argument(4)->Opcode()]);
2476     return false; // not enough info for intrinsification
2477   }
2478   if (!is_klass_initialized(vector_klass)) {
2479     log_if_needed("  ** klass argument not initialized");
2480     return false;
2481   }
2482   ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type();
2483   if (!elem_type->is_primitive_type()) {
2484     log_if_needed("  ** not a primitive bt=%d", elem_type->basic_type());
2485     return false; // should be primitive type
2486   }
2487   BasicType elem_bt = elem_type->basic_type();
2488   int num_elem = vlen->get_con();
2489   if (!arch_supports_vector(Op_VectorInsert, num_elem, elem_bt, VecMaskNotUsed)) {
2490     log_if_needed("  ** not supported: arity=1 op=insert vlen=%d etype=%s ismask=no",
2491                     num_elem, type2name(elem_bt));
2492     return false; // not supported
2493   }
2494 
2495   ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass();
2496   const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass);
2497 
2498   Node* opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem);
2499   if (opd == nullptr) {
2500     return false;
2501   }
2502 
2503   Node* insert_val = argument(5);
2504   assert(gvn().type(insert_val)->isa_long() != nullptr, "expected to be long");
2505 
2506   // Convert insert value back to its appropriate type.
2507   switch (elem_bt) {
2508     case T_BYTE:
2509       insert_val = gvn().transform(new ConvL2INode(insert_val));
2510       insert_val = gvn().transform(new CastIINode(insert_val, TypeInt::BYTE));
2511       break;
2512     case T_SHORT:
2513       insert_val = gvn().transform(new ConvL2INode(insert_val));
2514       insert_val = gvn().transform(new CastIINode(insert_val, TypeInt::SHORT));
2515       break;
2516     case T_INT:
2517       insert_val = gvn().transform(new ConvL2INode(insert_val));
2518       break;
2519     case T_FLOAT:
2520       insert_val = gvn().transform(new ConvL2INode(insert_val));
2521       insert_val = gvn().transform(new MoveI2FNode(insert_val));
2522       break;
2523     case T_DOUBLE:
2524       insert_val = gvn().transform(new MoveL2DNode(insert_val));
2525       break;
2526     case T_LONG:
2527       // no conversion needed
2528       break;
2529     default: fatal("%s", type2name(elem_bt)); break;
2530   }
2531 
2532   Node* operation = gvn().transform(VectorInsertNode::make(opd, insert_val, idx->get_con(), gvn()));
2533 
2534   Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem);
2535   set_result(vbox);
2536   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt))));
2537   return true;
2538 }
2539 
2540 //  public static
2541 //  <VM extends VectorPayload,
2542 //   E>
2543 //  long extract(Class<? extends VM> vClass, Class<E> eClass,
2544 //               int length,
2545 //               VM vm, int i,
2546 //               VecExtractOp<VM> defaultImpl)
2547 bool LibraryCallKit::inline_vector_extract() {
2548   const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr();
2549   const TypeInstPtr* elem_klass   = gvn().type(argument(1))->isa_instptr();
2550   const TypeInt*     vlen         = gvn().type(argument(2))->isa_int();
2551   const TypeInt*     idx          = gvn().type(argument(4))->isa_int();
2552 
2553   if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || idx == nullptr) {
2554     return false; // dead code
2555   }
2556   if (vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con() || !idx->is_con()) {
2557     log_if_needed("  ** missing constant: vclass=%s etype=%s vlen=%s",
2558                     NodeClassNames[argument(0)->Opcode()],
2559                     NodeClassNames[argument(1)->Opcode()],
2560                     NodeClassNames[argument(2)->Opcode()]);
2561     return false; // not enough info for intrinsification
2562   }
2563   if (!is_klass_initialized(vector_klass)) {
2564     log_if_needed("  ** klass argument not initialized");
2565     return false;
2566   }
2567   ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type();
2568   if (!elem_type->is_primitive_type()) {
2569     log_if_needed("  ** not a primitive bt=%d", elem_type->basic_type());
2570     return false; // should be primitive type
2571   }
2572   BasicType elem_bt = elem_type->basic_type();
2573   int num_elem = vlen->get_con();
2574 
2575   ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass();
2576   const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass);
2577 
2578   Node* opd = nullptr;
2579 
2580   if (is_vector_mask(vbox_klass)) {
2581     // vbox_klass is mask. This is used for VectorMask.laneIsSet(int).
2582 
2583     Node* pos = argument(4); // can be variable
2584     if (arch_supports_vector(Op_ExtractUB, num_elem, elem_bt, VecMaskUseAll)) {
2585       // Transform mask to vector with type of boolean and utilize ExtractUB node.
2586       opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem);
2587       if (opd == nullptr) {
2588         return false;
2589       }
2590       opd = gvn().transform(VectorStoreMaskNode::make(gvn(), opd, elem_bt, num_elem));
2591       opd = gvn().transform(new ExtractUBNode(opd, pos));
2592       opd = gvn().transform(new ConvI2LNode(opd));
2593     } else if (arch_supports_vector(Op_VectorMaskToLong, num_elem, elem_bt, VecMaskUseLoad)) {
2594       opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem);
2595       if (opd == nullptr) {
2596         return false;
2597       }
2598       // VectorMaskToLongNode requires the input is either a mask or a vector with BOOLEAN type.
2599       if (opd->bottom_type()->isa_vectmask() == nullptr) {
2600         opd = gvn().transform(VectorStoreMaskNode::make(gvn(), opd, elem_bt, num_elem));
2601       }
2602       // ((toLong() >>> pos) & 1L
2603       opd = gvn().transform(new VectorMaskToLongNode(opd, TypeLong::LONG));
2604       opd = gvn().transform(new URShiftLNode(opd, pos));
2605       opd = gvn().transform(new AndLNode(opd, gvn().makecon(TypeLong::ONE)));
2606     } else {
2607       log_if_needed("  ** Rejected mask extraction because architecture does not support it");
2608       return false; // not supported
2609     }
2610   } else {
2611     // vbox_klass is vector. This is used for Vector.lane(int).
2612     if (!idx->is_con()) {
2613       log_if_needed("  ** missing constant: idx=%s", NodeClassNames[argument(4)->Opcode()]);
2614       return false; // not enough info for intrinsification
2615     }
2616 
2617     int vopc = ExtractNode::opcode(elem_bt);
2618     if (!arch_supports_vector(vopc, num_elem, elem_bt, VecMaskNotUsed)) {
2619       log_if_needed("  ** not supported: arity=1 op=extract vlen=%d etype=%s ismask=no",
2620                       num_elem, type2name(elem_bt));
2621       return false; // not supported
2622     }
2623 
2624     opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem);
2625     if (opd == nullptr) {
2626       return false;
2627     }
2628     ConINode* idx_con = gvn().intcon(idx->get_con())->as_ConI();
2629 
2630     opd = gvn().transform(ExtractNode::make(opd, idx_con, elem_bt));
2631     switch (elem_bt) {
2632       case T_BYTE:
2633       case T_SHORT:
2634       case T_INT: {
2635         opd = gvn().transform(new ConvI2LNode(opd));
2636         break;
2637       }
2638       case T_FLOAT: {
2639         opd = gvn().transform(new MoveF2INode(opd));
2640         opd = gvn().transform(new ConvI2LNode(opd));
2641         break;
2642       }
2643       case T_DOUBLE: {
2644         opd = gvn().transform(new MoveD2LNode(opd));
2645         break;
2646       }
2647       case T_LONG: {
2648         // no conversion needed
2649         break;
2650       }
2651       default: fatal("%s", type2name(elem_bt));
2652     }
2653   }
2654   set_result(opd);
2655   return true;
2656 }
2657 
2658 // public static
2659 // <V extends Vector<E>,
2660 //  M extends VectorMask<E>,
2661 //  E>
2662 //  V compressExpandOp(int opr,
2663 //                    Class<? extends V> vClass, Class<? extends M> mClass, Class<E> eClass,
2664 //                    int length, V v, M m,
2665 //                    CompressExpandOperation<V, M> defaultImpl)
2666 bool LibraryCallKit::inline_vector_compress_expand() {
2667   const TypeInt*     opr          = gvn().type(argument(0))->isa_int();
2668   const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr();
2669   const TypeInstPtr* mask_klass   = gvn().type(argument(2))->isa_instptr();
2670   const TypeInstPtr* elem_klass   = gvn().type(argument(3))->isa_instptr();
2671   const TypeInt*     vlen         = gvn().type(argument(4))->isa_int();
2672 
2673   if (opr          == nullptr || !opr->is_con() ||
2674       vector_klass == nullptr || vector_klass->const_oop() == nullptr ||
2675       mask_klass   == nullptr || mask_klass->const_oop()   == nullptr ||
2676       elem_klass   == nullptr || elem_klass->const_oop()   == nullptr ||
2677       vlen         == nullptr || !vlen->is_con()) {
2678     log_if_needed("  ** missing constant: opr=%s vclass=%s mclass=%s etype=%s vlen=%s",
2679                     NodeClassNames[argument(0)->Opcode()],
2680                     NodeClassNames[argument(1)->Opcode()],
2681                     NodeClassNames[argument(2)->Opcode()],
2682                     NodeClassNames[argument(3)->Opcode()],
2683                     NodeClassNames[argument(4)->Opcode()]);
2684     return false; // not enough info for intrinsification
2685   }
2686 
2687   if (!is_klass_initialized(vector_klass) || !is_klass_initialized(mask_klass)) {
2688     log_if_needed("  ** klass argument not initialized");
2689     return false;
2690   }
2691 
2692   ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type();
2693   if (!elem_type->is_primitive_type()) {
2694     log_if_needed("  ** not a primitive bt=%d", elem_type->basic_type());
2695     return false; // should be primitive type
2696   }
2697 
2698   int num_elem = vlen->get_con();
2699   BasicType elem_bt = elem_type->basic_type();
2700   int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt);
2701 
2702   if (!arch_supports_vector(opc, num_elem, elem_bt, VecMaskUseLoad)) {
2703     log_if_needed("  ** not supported: opc=%d vlen=%d etype=%s ismask=useload",
2704                     opc, num_elem, type2name(elem_bt));
2705     return false; // not supported
2706   }
2707 
2708   Node* opd1 = nullptr;
2709   const TypeInstPtr* vbox_type = nullptr;
2710   if (opc != Op_CompressM) {
2711     ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass();
2712     vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass);
2713     opd1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem);
2714     if (opd1 == nullptr) {
2715       log_if_needed("  ** unbox failed vector=%s",
2716                       NodeClassNames[argument(5)->Opcode()]);
2717       return false;
2718     }
2719   }
2720 
2721   ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass();
2722   assert(is_vector_mask(mbox_klass), "argument(6) should be a mask class");
2723   const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass);
2724 
2725   Node* mask = unbox_vector(argument(6), mbox_type, elem_bt, num_elem);
2726   if (mask == nullptr) {
2727     log_if_needed("  ** unbox failed mask=%s",
2728                     NodeClassNames[argument(6)->Opcode()]);
2729     return false;
2730   }
2731 
2732   const TypeVect* vt = TypeVect::make(elem_bt, num_elem, opc == Op_CompressM);
2733   Node* operation = gvn().transform(VectorNode::make(opc, opd1, mask, vt));
2734 
2735   // Wrap it up in VectorBox to keep object type information.
2736   const TypeInstPtr* box_type = opc == Op_CompressM ? mbox_type : vbox_type;
2737   Node* vbox = box_vector(operation, box_type, elem_bt, num_elem);
2738   set_result(vbox);
2739   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt))));
2740   return true;
2741 }
2742 
2743 // public static
2744 // <V extends Vector<E>,
2745 //  E,
2746 //  S extends VectorSpecies<E>>
2747 //  V indexVector(Class<? extends V> vClass, Class<E> eClass,
2748 //                int length,
2749 //                V v, int step, S s,
2750 //                IndexOperation<V, S> defaultImpl)
2751 bool LibraryCallKit::inline_index_vector() {
2752   const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr();
2753   const TypeInstPtr* elem_klass   = gvn().type(argument(1))->isa_instptr();
2754   const TypeInt*     vlen         = gvn().type(argument(2))->isa_int();
2755 
2756   if (vector_klass == nullptr || vector_klass->const_oop() == nullptr ||
2757       elem_klass   == nullptr || elem_klass->const_oop()   == nullptr ||
2758       vlen         == nullptr || !vlen->is_con() ) {
2759     log_if_needed("  ** missing constant: vclass=%s etype=%s vlen=%s",
2760                     NodeClassNames[argument(0)->Opcode()],
2761                     NodeClassNames[argument(1)->Opcode()],
2762                     NodeClassNames[argument(2)->Opcode()]);
2763     return false; // not enough info for intrinsification
2764   }
2765 
2766   if (!is_klass_initialized(vector_klass)) {
2767     log_if_needed("  ** klass argument not initialized");
2768     return false;
2769   }
2770 
2771   ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type();
2772   if (!elem_type->is_primitive_type()) {
2773     log_if_needed("  ** not a primitive bt=%d", elem_type->basic_type());
2774     return false; // should be primitive type
2775   }
2776 
2777   int num_elem = vlen->get_con();
2778   BasicType elem_bt = elem_type->basic_type();
2779 
2780   // Check whether the iota index generation op is supported by the current hardware
2781   if (!arch_supports_vector(Op_VectorLoadConst, num_elem, elem_bt, VecMaskNotUsed)) {
2782     log_if_needed("  ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt));
2783     return false; // not supported
2784   }
2785 
2786   int mul_op = VectorSupport::vop2ideal(VectorSupport::VECTOR_OP_MUL, elem_bt);
2787   int vmul_op = VectorNode::opcode(mul_op, elem_bt);
2788   bool needs_mul = true;
2789   Node* scale = argument(4);
2790   const TypeInt* scale_type = gvn().type(scale)->isa_int();
2791   // Multiply is not needed if the scale is a constant "1".
2792   if (scale_type && scale_type->is_con() && scale_type->get_con() == 1) {
2793     needs_mul = false;
2794   } else {
2795     // Check whether the vector multiply op is supported by the current hardware
2796     if (!arch_supports_vector(vmul_op, num_elem, elem_bt, VecMaskNotUsed)) {
2797       log_if_needed("  ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt));
2798       return false; // not supported
2799     }
2800 
2801     // Check whether the scalar cast op is supported by the current hardware
2802     if (is_floating_point_type(elem_bt) || elem_bt == T_LONG) {
2803       int cast_op = elem_bt == T_LONG ? Op_ConvI2L :
2804                     elem_bt == T_FLOAT? Op_ConvI2F : Op_ConvI2D;
2805       if (!Matcher::match_rule_supported(cast_op)) {
2806         log_if_needed("  ** Rejected op (%s) because architecture does not support it",
2807                         NodeClassNames[cast_op]);
2808         return false; // not supported
2809       }
2810     }
2811   }
2812 
2813   ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass();
2814   const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass);
2815   Node* opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem);
2816   if (opd == nullptr) {
2817     log_if_needed("  ** unbox failed vector=%s",
2818                     NodeClassNames[argument(3)->Opcode()]);
2819     return false;
2820   }
2821 
2822   int add_op = VectorSupport::vop2ideal(VectorSupport::VECTOR_OP_ADD, elem_bt);
2823   int vadd_op = VectorNode::opcode(add_op, elem_bt);
2824   bool needs_add = true;
2825   // The addition is not needed if all the element values of "opd" are zero
2826   if (VectorNode::is_all_zeros_vector(opd)) {
2827     needs_add = false;
2828   } else {
2829     // Check whether the vector addition op is supported by the current hardware
2830     if (!arch_supports_vector(vadd_op, num_elem, elem_bt, VecMaskNotUsed)) {
2831       log_if_needed("  ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt));
2832       return false; // not supported
2833     }
2834   }
2835 
2836   // Compute the iota indice vector
2837   const TypeVect* vt = TypeVect::make(elem_bt, num_elem);
2838   Node* index = gvn().transform(new VectorLoadConstNode(gvn().makecon(TypeInt::ZERO), vt));
2839 
2840   // Broadcast the "scale" to a vector, and multiply the "scale" with iota indice vector.
2841   if (needs_mul) {
2842     switch (elem_bt) {
2843       case T_BOOLEAN: // fall-through
2844       case T_BYTE:    // fall-through
2845       case T_SHORT:   // fall-through
2846       case T_CHAR:    // fall-through
2847       case T_INT: {
2848         // no conversion needed
2849         break;
2850       }
2851       case T_LONG: {
2852         scale = gvn().transform(new ConvI2LNode(scale));
2853         break;
2854       }
2855       case T_FLOAT: {
2856         scale = gvn().transform(new ConvI2FNode(scale));
2857         break;
2858       }
2859       case T_DOUBLE: {
2860         scale = gvn().transform(new ConvI2DNode(scale));
2861         break;
2862       }
2863       default: fatal("%s", type2name(elem_bt));
2864     }
2865     scale = gvn().transform(VectorNode::scalar2vector(scale, num_elem, Type::get_const_basic_type(elem_bt)));
2866     index = gvn().transform(VectorNode::make(vmul_op, index, scale, vt));
2867   }
2868 
2869   // Add "opd" if addition is needed.
2870   if (needs_add) {
2871     index = gvn().transform(VectorNode::make(vadd_op, opd, index, vt));
2872   }
2873   Node* vbox = box_vector(index, vbox_type, elem_bt, num_elem);
2874   set_result(vbox);
2875   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt))));
2876   return true;
2877 }
2878 
2879 // public static
2880 // <E,
2881 //  M extends VectorMask<E>>
2882 // M indexPartiallyInUpperRange(Class<? extends M> mClass, Class<E> eClass, int length,
2883 //                              long offset, long limit,
2884 //                              IndexPartiallyInUpperRangeOperation<E, M> defaultImpl)
2885 bool LibraryCallKit::inline_index_partially_in_upper_range() {
2886   const TypeInstPtr* mask_klass   = gvn().type(argument(0))->isa_instptr();
2887   const TypeInstPtr* elem_klass   = gvn().type(argument(1))->isa_instptr();
2888   const TypeInt*     vlen         = gvn().type(argument(2))->isa_int();
2889 
2890   if (mask_klass == nullptr || mask_klass->const_oop() == nullptr ||
2891       elem_klass == nullptr || elem_klass->const_oop() == nullptr ||
2892       vlen       == nullptr || !vlen->is_con()) {
2893     log_if_needed("  ** missing constant: mclass=%s etype=%s vlen=%s",
2894                     NodeClassNames[argument(0)->Opcode()],
2895                     NodeClassNames[argument(1)->Opcode()],
2896                     NodeClassNames[argument(2)->Opcode()]);
2897     return false; // not enough info for intrinsification
2898   }
2899 
2900   if (!is_klass_initialized(mask_klass)) {
2901     log_if_needed("  ** klass argument not initialized");
2902     return false;
2903   }
2904 
2905   ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type();
2906   if (!elem_type->is_primitive_type()) {
2907     log_if_needed("  ** not a primitive bt=%d", elem_type->basic_type());
2908     return false; // should be primitive type
2909   }
2910 
2911   int num_elem = vlen->get_con();
2912   BasicType elem_bt = elem_type->basic_type();
2913 
2914   // Check whether the necessary ops are supported by current hardware.
2915   bool supports_mask_gen = arch_supports_vector(Op_VectorMaskGen, num_elem, elem_bt, VecMaskUseStore);
2916   if (!supports_mask_gen) {
2917     if (!arch_supports_vector(Op_VectorLoadConst, num_elem, elem_bt, VecMaskNotUsed) ||
2918         !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed) ||
2919         !arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskUseStore)) {
2920       log_if_needed("  ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt));
2921       return false; // not supported
2922     }
2923 
2924     // Check whether the scalar cast operation is supported by current hardware.
2925     if (elem_bt != T_LONG) {
2926       int cast_op = is_integral_type(elem_bt) ? Op_ConvL2I
2927                                               : (elem_bt == T_FLOAT ? Op_ConvL2F : Op_ConvL2D);
2928       if (!Matcher::match_rule_supported(cast_op)) {
2929         log_if_needed("  ** Rejected op (%s) because architecture does not support it",
2930                         NodeClassNames[cast_op]);
2931         return false; // not supported
2932       }
2933     }
2934   }
2935 
2936   Node* offset = argument(3);
2937   Node* limit = argument(5);
2938   if (offset == nullptr || limit == nullptr) {
2939     log_if_needed("  ** offset or limit argument is null");
2940     return false; // not supported
2941   }
2942 
2943   ciKlass* box_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass();
2944   assert(is_vector_mask(box_klass), "argument(0) should be a mask class");
2945   const TypeInstPtr* box_type = TypeInstPtr::make_exact(TypePtr::NotNull, box_klass);
2946 
2947   // We assume "offset > 0 && limit >= offset && limit - offset < num_elem".
2948   // So directly get indexLimit with "indexLimit = limit - offset".
2949   Node* indexLimit = gvn().transform(new SubLNode(limit, offset));
2950   Node* mask = nullptr;
2951   if (supports_mask_gen) {
2952     mask = gvn().transform(VectorMaskGenNode::make(indexLimit, elem_bt, num_elem));
2953   } else {
2954     // Generate the vector mask based on "mask = iota < indexLimit".
2955     // Broadcast "indexLimit" to a vector.
2956     switch (elem_bt) {
2957       case T_BOOLEAN: // fall-through
2958       case T_BYTE:    // fall-through
2959       case T_SHORT:   // fall-through
2960       case T_CHAR:    // fall-through
2961       case T_INT: {
2962         indexLimit = gvn().transform(new ConvL2INode(indexLimit));
2963         break;
2964       }
2965       case T_DOUBLE: {
2966         indexLimit = gvn().transform(new ConvL2DNode(indexLimit));
2967         break;
2968       }
2969       case T_FLOAT: {
2970         indexLimit = gvn().transform(new ConvL2FNode(indexLimit));
2971         break;
2972       }
2973       case T_LONG: {
2974         // no conversion needed
2975         break;
2976       }
2977       default: fatal("%s", type2name(elem_bt));
2978     }
2979     indexLimit = gvn().transform(VectorNode::scalar2vector(indexLimit, num_elem, Type::get_const_basic_type(elem_bt)));
2980 
2981     // Load the "iota" vector.
2982     const TypeVect* vt = TypeVect::make(elem_bt, num_elem);
2983     Node* iota = gvn().transform(new VectorLoadConstNode(gvn().makecon(TypeInt::ZERO), vt));
2984 
2985     // Compute the vector mask with "mask = iota < indexLimit".
2986     ConINode* pred_node = (ConINode*)gvn().makecon(TypeInt::make(BoolTest::lt));
2987     const TypeVect* vmask_type = TypeVect::makemask(elem_bt, num_elem);
2988     mask = gvn().transform(new VectorMaskCmpNode(BoolTest::lt, iota, indexLimit, pred_node, vmask_type));
2989   }
2990   Node* vbox = box_vector(mask, box_type, elem_bt, num_elem);
2991   set_result(vbox);
2992   C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt))));
2993   return true;
2994 }
2995 
2996 #undef non_product_log_if_needed
2997 #undef log_if_needed