1 /* 2 * Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "ci/ciSymbols.hpp" 26 #include "classfile/vmSymbols.hpp" 27 #include "opto/library_call.hpp" 28 #include "opto/runtime.hpp" 29 #include "opto/vectornode.hpp" 30 #include "prims/vectorSupport.hpp" 31 #include "runtime/stubRoutines.hpp" 32 33 #ifdef ASSERT 34 static bool is_vector(ciKlass* klass) { 35 return klass->is_subclass_of(ciEnv::current()->vector_VectorPayload_klass()); 36 } 37 38 static bool check_vbox(const TypeInstPtr* vbox_type) { 39 assert(vbox_type->klass_is_exact(), ""); 40 41 ciInstanceKlass* ik = vbox_type->instance_klass(); 42 assert(is_vector(ik), "not a vector"); 43 44 ciField* fd1 = ik->get_field_by_name(ciSymbols::ETYPE_name(), ciSymbols::class_signature(), /* is_static */ true); 45 assert(fd1 != nullptr, "element type info is missing"); 46 47 ciConstant val1 = fd1->constant_value(); 48 BasicType elem_bt = val1.as_object()->as_instance()->java_mirror_type()->basic_type(); 49 assert(is_java_primitive(elem_bt), "element type info is missing"); 50 51 ciField* fd2 = ik->get_field_by_name(ciSymbols::VLENGTH_name(), ciSymbols::int_signature(), /* is_static */ true); 52 assert(fd2 != nullptr, "vector length info is missing"); 53 54 ciConstant val2 = fd2->constant_value(); 55 assert(val2.as_int() > 0, "vector length info is missing"); 56 57 return true; 58 } 59 #endif 60 61 #define log_if_needed(...) \ 62 if (C->print_intrinsics()) { \ 63 tty->print_cr(__VA_ARGS__); \ 64 } 65 66 #ifndef PRODUCT 67 #define non_product_log_if_needed(...) log_if_needed(__VA_ARGS__) 68 #else 69 #define non_product_log_if_needed(...) 70 #endif 71 72 static bool is_vector_mask(ciKlass* klass) { 73 return klass->is_subclass_of(ciEnv::current()->vector_VectorMask_klass()); 74 } 75 76 bool LibraryCallKit::arch_supports_vector_rotate(int opc, int num_elem, BasicType elem_bt, 77 VectorMaskUseType mask_use_type, bool has_scalar_args) { 78 bool is_supported = true; 79 80 // has_scalar_args flag is true only for non-constant scalar shift count, 81 // since in this case shift needs to be broadcasted. 82 if (!Matcher::match_rule_supported_vector(opc, num_elem, elem_bt) || 83 (has_scalar_args && !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed))) { 84 is_supported = false; 85 } 86 87 if (is_supported) { 88 // Check if mask unboxing is supported, this is a two step process which first loads the contents 89 // of boolean array into vector followed by either lane expansion to match the lane size of masked 90 // vector operation or populate the predicate register. 91 if ((mask_use_type & VecMaskUseLoad) != 0) { 92 if (!Matcher::match_rule_supported_vector(Op_VectorLoadMask, num_elem, elem_bt) || 93 !Matcher::match_rule_supported_vector(Op_LoadVector, num_elem, T_BOOLEAN)) { 94 non_product_log_if_needed(" ** Rejected vector mask loading (%s,%s,%d) because architecture does not support it", 95 NodeClassNames[Op_VectorLoadMask], type2name(elem_bt), num_elem); 96 return false; 97 } 98 } 99 100 if ((mask_use_type & VecMaskUsePred) != 0) { 101 if (!Matcher::has_predicated_vectors() || 102 !Matcher::match_rule_supported_vector_masked(opc, num_elem, elem_bt)) { 103 non_product_log_if_needed("Rejected vector mask predicate using (%s,%s,%d) because architecture does not support it", 104 NodeClassNames[opc], type2name(elem_bt), num_elem); 105 return false; 106 } 107 } 108 } 109 110 int lshiftopc, rshiftopc; 111 switch(elem_bt) { 112 case T_BYTE: 113 lshiftopc = Op_LShiftI; 114 rshiftopc = Op_URShiftB; 115 break; 116 case T_SHORT: 117 lshiftopc = Op_LShiftI; 118 rshiftopc = Op_URShiftS; 119 break; 120 case T_INT: 121 lshiftopc = Op_LShiftI; 122 rshiftopc = Op_URShiftI; 123 break; 124 case T_LONG: 125 lshiftopc = Op_LShiftL; 126 rshiftopc = Op_URShiftL; 127 break; 128 default: fatal("Unexpected type: %s", type2name(elem_bt)); 129 } 130 int lshiftvopc = VectorNode::opcode(lshiftopc, elem_bt); 131 int rshiftvopc = VectorNode::opcode(rshiftopc, elem_bt); 132 if (!is_supported && 133 arch_supports_vector(lshiftvopc, num_elem, elem_bt, VecMaskNotUsed, has_scalar_args) && 134 arch_supports_vector(rshiftvopc, num_elem, elem_bt, VecMaskNotUsed, has_scalar_args) && 135 arch_supports_vector(Op_OrV, num_elem, elem_bt, VecMaskNotUsed)) { 136 is_supported = true; 137 } 138 return is_supported; 139 } 140 141 Node* GraphKit::box_vector(Node* vector, const TypeInstPtr* vbox_type, BasicType elem_bt, int num_elem, bool deoptimize_on_exception) { 142 assert(EnableVectorSupport, ""); 143 144 PreserveReexecuteState preexecs(this); 145 jvms()->set_should_reexecute(true); 146 147 VectorBoxAllocateNode* alloc = new VectorBoxAllocateNode(C, vbox_type); 148 set_edges_for_java_call(alloc, /*must_throw=*/false, /*separate_io_proj=*/true); 149 make_slow_call_ex(alloc, env()->Throwable_klass(), /*separate_io_proj=*/true, deoptimize_on_exception); 150 set_i_o(gvn().transform( new ProjNode(alloc, TypeFunc::I_O) )); 151 set_all_memory(gvn().transform( new ProjNode(alloc, TypeFunc::Memory) )); 152 Node* ret = gvn().transform(new ProjNode(alloc, TypeFunc::Parms)); 153 154 assert(check_vbox(vbox_type), ""); 155 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, is_vector_mask(vbox_type->instance_klass())); 156 VectorBoxNode* vbox = new VectorBoxNode(C, ret, vector, vbox_type, vt); 157 return gvn().transform(vbox); 158 } 159 160 Node* GraphKit::unbox_vector(Node* v, const TypeInstPtr* vbox_type, BasicType elem_bt, int num_elem) { 161 assert(EnableVectorSupport, ""); 162 const TypeInstPtr* vbox_type_v = gvn().type(v)->isa_instptr(); 163 if (vbox_type_v == nullptr || vbox_type->instance_klass() != vbox_type_v->instance_klass()) { 164 return nullptr; // arguments don't agree on vector shapes 165 } 166 if (vbox_type_v->maybe_null()) { 167 return nullptr; // no nulls are allowed 168 } 169 assert(check_vbox(vbox_type), ""); 170 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, is_vector_mask(vbox_type->instance_klass())); 171 Node* unbox = gvn().transform(new VectorUnboxNode(C, vt, v, merged_memory())); 172 return unbox; 173 } 174 175 Node* GraphKit::vector_shift_count(Node* cnt, int shift_op, BasicType bt, int num_elem) { 176 assert(bt == T_INT || bt == T_LONG || bt == T_SHORT || bt == T_BYTE, "byte, short, long and int are supported"); 177 juint mask = (type2aelembytes(bt) * BitsPerByte - 1); 178 Node* nmask = gvn().transform(ConNode::make(TypeInt::make(mask))); 179 Node* mcnt = gvn().transform(new AndINode(cnt, nmask)); 180 return gvn().transform(VectorNode::shift_count(shift_op, mcnt, num_elem, bt)); 181 } 182 183 bool LibraryCallKit::arch_supports_vector(int sopc, int num_elem, BasicType type, VectorMaskUseType mask_use_type, bool has_scalar_args) { 184 // Check that the operation is valid. 185 if (sopc <= 0) { 186 non_product_log_if_needed(" ** Rejected intrinsification because no valid vector op could be extracted"); 187 return false; 188 } 189 190 if (VectorNode::is_vector_rotate(sopc)) { 191 if(!arch_supports_vector_rotate(sopc, num_elem, type, mask_use_type, has_scalar_args)) { 192 non_product_log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support variable vector shifts", 193 NodeClassNames[sopc], type2name(type), num_elem); 194 return false; 195 } 196 } else if (VectorNode::is_vector_integral_negate(sopc)) { 197 if (!VectorNode::is_vector_integral_negate_supported(sopc, num_elem, type, false)) { 198 non_product_log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support integral vector negate", 199 NodeClassNames[sopc], type2name(type), num_elem); 200 return false; 201 } 202 } else { 203 // Check that architecture supports this op-size-type combination. 204 if (!Matcher::match_rule_supported_vector(sopc, num_elem, type)) { 205 non_product_log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support it", 206 NodeClassNames[sopc], type2name(type), num_elem); 207 return false; 208 } else { 209 assert(Matcher::match_rule_supported(sopc), "must be supported"); 210 } 211 } 212 213 if (num_elem == 1) { 214 if (mask_use_type != VecMaskNotUsed) { 215 non_product_log_if_needed(" ** Rejected vector mask op (%s,%s,%d) because architecture does not support it", 216 NodeClassNames[sopc], type2name(type), num_elem); 217 return false; 218 } 219 220 if (sopc != 0) { 221 if (sopc != Op_LoadVector && sopc != Op_StoreVector) { 222 non_product_log_if_needed(" ** Not a svml call or load/store vector op (%s,%s,%d)", 223 NodeClassNames[sopc], type2name(type), num_elem); 224 return false; 225 } 226 } 227 } 228 229 if (!has_scalar_args && VectorNode::is_vector_shift(sopc) && 230 Matcher::supports_vector_variable_shifts() == false) { 231 log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support variable vector shifts", 232 NodeClassNames[sopc], type2name(type), num_elem); 233 return false; 234 } 235 236 // Check if mask unboxing is supported, this is a two step process which first loads the contents 237 // of boolean array into vector followed by either lane expansion to match the lane size of masked 238 // vector operation or populate the predicate register. 239 if ((mask_use_type & VecMaskUseLoad) != 0) { 240 if (!Matcher::match_rule_supported_vector(Op_VectorLoadMask, num_elem, type) || 241 !Matcher::match_rule_supported_vector(Op_LoadVector, num_elem, T_BOOLEAN)) { 242 non_product_log_if_needed(" ** Rejected vector mask loading (%s,%s,%d) because architecture does not support it", 243 NodeClassNames[Op_VectorLoadMask], type2name(type), num_elem); 244 return false; 245 } 246 } 247 248 // Check if mask boxing is supported, this is a two step process which first stores the contents 249 // of mask vector / predicate register into a boolean vector followed by vector store operation to 250 // transfer the contents to underlined storage of mask boxes which is a boolean array. 251 if ((mask_use_type & VecMaskUseStore) != 0) { 252 if (!Matcher::match_rule_supported_vector(Op_VectorStoreMask, num_elem, type) || 253 !Matcher::match_rule_supported_vector(Op_StoreVector, num_elem, T_BOOLEAN)) { 254 non_product_log_if_needed("Rejected vector mask storing (%s,%s,%d) because architecture does not support it", 255 NodeClassNames[Op_VectorStoreMask], type2name(type), num_elem); 256 return false; 257 } 258 } 259 260 if ((mask_use_type & VecMaskUsePred) != 0) { 261 bool is_supported = false; 262 if (Matcher::has_predicated_vectors()) { 263 if (VectorNode::is_vector_integral_negate(sopc)) { 264 is_supported = VectorNode::is_vector_integral_negate_supported(sopc, num_elem, type, true); 265 } else { 266 is_supported = Matcher::match_rule_supported_vector_masked(sopc, num_elem, type); 267 } 268 } 269 is_supported |= Matcher::supports_vector_predicate_op_emulation(sopc, num_elem, type); 270 271 if (!is_supported) { 272 non_product_log_if_needed("Rejected vector mask predicate using (%s,%s,%d) because architecture does not support it", 273 NodeClassNames[sopc], type2name(type), num_elem); 274 return false; 275 } 276 } 277 278 return true; 279 } 280 281 static bool is_klass_initialized(const TypeInstPtr* vec_klass) { 282 if (vec_klass->const_oop() == nullptr) { 283 return false; // uninitialized or some kind of unsafe access 284 } 285 assert(vec_klass->const_oop()->as_instance()->java_lang_Class_klass() != nullptr, "klass instance expected"); 286 ciInstanceKlass* klass = vec_klass->const_oop()->as_instance()->java_lang_Class_klass()->as_instance_klass(); 287 return klass->is_initialized(); 288 } 289 290 // public static 291 // <V extends Vector<E>, 292 // M extends VectorMask<E>, 293 // E> 294 // V unaryOp(int oprId, Class<? extends V> vmClass, Class<? extends M> maskClass, Class<E> elementType, 295 // int length, V v, M m, 296 // UnaryOperation<V, M> defaultImpl) 297 // 298 // public static 299 // <V, 300 // M extends VectorMask<E>, 301 // E> 302 // V binaryOp(int oprId, Class<? extends V> vmClass, Class<? extends M> maskClass, Class<E> elementType, 303 // int length, V v1, V v2, M m, 304 // BinaryOperation<V, M> defaultImpl) 305 // 306 // public static 307 // <V extends Vector<E>, 308 // M extends VectorMask<E>, 309 // E> 310 // V ternaryOp(int oprId, Class<? extends V> vmClass, Class<? extends M> maskClass, Class<E> elementType, 311 // int length, V v1, V v2, V v3, M m, 312 // TernaryOperation<V, M> defaultImpl) 313 // 314 bool LibraryCallKit::inline_vector_nary_operation(int n) { 315 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 316 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 317 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 318 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 319 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 320 321 if (opr == nullptr || !opr->is_con() || 322 vector_klass == nullptr || vector_klass->const_oop() == nullptr || 323 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 324 vlen == nullptr || !vlen->is_con()) { 325 log_if_needed(" ** missing constant: opr=%s vclass=%s etype=%s vlen=%s", 326 NodeClassNames[argument(0)->Opcode()], 327 NodeClassNames[argument(1)->Opcode()], 328 NodeClassNames[argument(3)->Opcode()], 329 NodeClassNames[argument(4)->Opcode()]); 330 return false; // not enough info for intrinsification 331 } 332 333 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 334 if (!elem_type->is_primitive_type()) { 335 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 336 return false; // should be primitive type 337 } 338 if (!is_klass_initialized(vector_klass)) { 339 log_if_needed(" ** klass argument not initialized"); 340 return false; 341 } 342 343 // "argument(n + 5)" should be the mask object. We assume it is "null" when no mask 344 // is used to control this operation. 345 const Type* vmask_type = gvn().type(argument(n + 5)); 346 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 347 if (is_masked_op) { 348 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 349 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(2)->Opcode()]); 350 return false; // not enough info for intrinsification 351 } 352 353 if (!is_klass_initialized(mask_klass)) { 354 log_if_needed(" ** mask klass argument not initialized"); 355 return false; 356 } 357 358 if (vmask_type->maybe_null()) { 359 log_if_needed(" ** null mask values are not allowed for masked op"); 360 return false; 361 } 362 } 363 364 BasicType elem_bt = elem_type->basic_type(); 365 bool has_scalar_op = VectorSupport::has_scalar_op(opr->get_con()); 366 bool is_unsigned = VectorSupport::is_unsigned_op(opr->get_con()); 367 368 int num_elem = vlen->get_con(); 369 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 370 int sopc = has_scalar_op ? VectorNode::opcode(opc, elem_bt) : opc; 371 if ((opc != Op_CallLeafVector) && (sopc == 0)) { 372 log_if_needed(" ** operation not supported: opc=%s bt=%s", NodeClassNames[opc], type2name(elem_bt)); 373 return false; // operation not supported 374 } 375 if (num_elem == 1) { 376 if (opc != Op_CallLeafVector || elem_bt != T_DOUBLE) { 377 log_if_needed(" ** not a svml call: arity=%d opc=%d vlen=%d etype=%s", 378 n, opc, num_elem, type2name(elem_bt)); 379 return false; 380 } 381 } 382 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 383 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 384 385 if (is_vector_mask(vbox_klass)) { 386 assert(!is_masked_op, "mask operations do not need mask to control"); 387 } 388 389 if (opc == Op_CallLeafVector) { 390 if (!UseVectorStubs) { 391 log_if_needed(" ** vector stubs support is disabled"); 392 return false; 393 } 394 if (!Matcher::supports_vector_calling_convention()) { 395 log_if_needed(" ** no vector calling conventions supported"); 396 return false; 397 } 398 if (!Matcher::vector_size_supported(elem_bt, num_elem)) { 399 log_if_needed(" ** vector size (vlen=%d, etype=%s) is not supported", 400 num_elem, type2name(elem_bt)); 401 return false; 402 } 403 } 404 405 // When using mask, mask use type needs to be VecMaskUseLoad. 406 VectorMaskUseType mask_use_type = is_vector_mask(vbox_klass) ? VecMaskUseAll 407 : is_masked_op ? VecMaskUseLoad : VecMaskNotUsed; 408 if ((sopc != 0) && !arch_supports_vector(sopc, num_elem, elem_bt, mask_use_type)) { 409 log_if_needed(" ** not supported: arity=%d opc=%d vlen=%d etype=%s ismask=%d is_masked_op=%d", 410 n, sopc, num_elem, type2name(elem_bt), 411 is_vector_mask(vbox_klass) ? 1 : 0, is_masked_op ? 1 : 0); 412 return false; // not supported 413 } 414 415 // Return true if current platform has implemented the masked operation with predicate feature. 416 bool use_predicate = is_masked_op && sopc != 0 && arch_supports_vector(sopc, num_elem, elem_bt, VecMaskUsePred); 417 if (is_masked_op && !use_predicate && !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)) { 418 log_if_needed(" ** not supported: arity=%d opc=%d vlen=%d etype=%s ismask=0 is_masked_op=1", 419 n, sopc, num_elem, type2name(elem_bt)); 420 return false; 421 } 422 423 Node* opd1 = nullptr; Node* opd2 = nullptr; Node* opd3 = nullptr; 424 switch (n) { 425 case 3: { 426 opd3 = unbox_vector(argument(7), vbox_type, elem_bt, num_elem); 427 if (opd3 == nullptr) { 428 log_if_needed(" ** unbox failed v3=%s", 429 NodeClassNames[argument(7)->Opcode()]); 430 return false; 431 } 432 // fall-through 433 } 434 case 2: { 435 opd2 = unbox_vector(argument(6), vbox_type, elem_bt, num_elem); 436 if (opd2 == nullptr) { 437 log_if_needed(" ** unbox failed v2=%s", 438 NodeClassNames[argument(6)->Opcode()]); 439 return false; 440 } 441 // fall-through 442 } 443 case 1: { 444 opd1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 445 if (opd1 == nullptr) { 446 log_if_needed(" ** unbox failed v1=%s", 447 NodeClassNames[argument(5)->Opcode()]); 448 return false; 449 } 450 break; 451 } 452 default: fatal("unsupported arity: %d", n); 453 } 454 455 Node* mask = nullptr; 456 if (is_masked_op) { 457 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 458 assert(is_vector_mask(mbox_klass), "argument(2) should be a mask class"); 459 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 460 mask = unbox_vector(argument(n + 5), mbox_type, elem_bt, num_elem); 461 if (mask == nullptr) { 462 log_if_needed(" ** unbox failed mask=%s", 463 NodeClassNames[argument(n + 5)->Opcode()]); 464 return false; 465 } 466 } 467 468 Node* operation = nullptr; 469 if (opc == Op_CallLeafVector) { 470 assert(UseVectorStubs, "sanity"); 471 operation = gen_call_to_vector_math(opr->get_con(), elem_bt, num_elem, opd1, opd2); 472 if (operation == nullptr) { 473 log_if_needed(" ** Vector math call failed for %s_%s_%d", 474 (elem_bt == T_FLOAT) ? "float" : "double", 475 VectorSupport::mathname[opr->get_con() - VectorSupport::VECTOR_OP_MATH_START], 476 num_elem * type2aelembytes(elem_bt)); 477 return false; 478 } 479 } else { 480 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, is_vector_mask(vbox_klass)); 481 switch (n) { 482 case 1: 483 case 2: { 484 operation = VectorNode::make(sopc, opd1, opd2, vt, is_vector_mask(vbox_klass), VectorNode::is_shift_opcode(opc), is_unsigned); 485 break; 486 } 487 case 3: { 488 operation = VectorNode::make(sopc, opd1, opd2, opd3, vt); 489 break; 490 } 491 default: fatal("unsupported arity: %d", n); 492 } 493 } 494 495 if (is_masked_op && mask != nullptr) { 496 if (use_predicate) { 497 operation->add_req(mask); 498 operation->add_flag(Node::Flag_is_predicated_vector); 499 } else { 500 operation->add_flag(Node::Flag_is_predicated_using_blend); 501 operation = gvn().transform(operation); 502 operation = new VectorBlendNode(opd1, operation, mask); 503 } 504 } 505 operation = gvn().transform(operation); 506 507 // Wrap it up in VectorBox to keep object type information. 508 Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem); 509 set_result(vbox); 510 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 511 return true; 512 } 513 514 // public static 515 // <E, M> 516 // long maskReductionCoerced(int oper, Class<? extends M> maskClass, Class<?> elemClass, 517 // int length, M m, VectorMaskOp<M> defaultImpl) 518 bool LibraryCallKit::inline_vector_mask_operation() { 519 const TypeInt* oper = gvn().type(argument(0))->isa_int(); 520 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 521 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 522 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 523 Node* mask = argument(4); 524 525 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr || 526 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 527 vlen == nullptr || !vlen->is_con() || 528 oper == nullptr || !oper->is_con() || 529 mask->is_top()) { 530 return false; // dead code 531 } 532 533 if (!is_klass_initialized(mask_klass)) { 534 log_if_needed(" ** klass argument not initialized"); 535 return false; 536 } 537 538 int num_elem = vlen->get_con(); 539 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 540 BasicType elem_bt = elem_type->basic_type(); 541 542 int mopc = VectorSupport::vop2ideal(oper->get_con(), elem_bt); 543 if (!arch_supports_vector(mopc, num_elem, elem_bt, VecMaskUseLoad)) { 544 log_if_needed(" ** not supported: arity=1 op=cast#%d/3 vlen2=%d etype2=%s", 545 mopc, num_elem, type2name(elem_bt)); 546 return false; // not supported 547 } 548 549 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 550 const TypeInstPtr* mask_box_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 551 Node* mask_vec = unbox_vector(mask, mask_box_type, elem_bt, num_elem); 552 if (mask_vec == nullptr) { 553 log_if_needed(" ** unbox failed mask=%s", 554 NodeClassNames[argument(4)->Opcode()]); 555 return false; 556 } 557 558 if (mask_vec->bottom_type()->isa_vectmask() == nullptr) { 559 mask_vec = gvn().transform(VectorStoreMaskNode::make(gvn(), mask_vec, elem_bt, num_elem)); 560 } 561 const Type* maskoper_ty = mopc == Op_VectorMaskToLong ? (const Type*)TypeLong::LONG : (const Type*)TypeInt::INT; 562 Node* maskoper = gvn().transform(VectorMaskOpNode::make(mask_vec, maskoper_ty, mopc)); 563 if (mopc != Op_VectorMaskToLong) { 564 maskoper = ConvI2L(maskoper); 565 } 566 set_result(maskoper); 567 568 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 569 return true; 570 } 571 572 // public static 573 // <M, 574 // S extends VectorSpecies<E>, 575 // E> 576 // M fromBitsCoerced(Class<? extends M> vmClass, Class<E> elementType, int length, 577 // long bits, int mode, S s, 578 // BroadcastOperation<M, E, S> defaultImpl) 579 bool LibraryCallKit::inline_vector_frombits_coerced() { 580 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 581 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 582 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 583 const TypeLong* bits_type = gvn().type(argument(3))->isa_long(); 584 // Mode argument determines the mode of operation it can take following values:- 585 // MODE_BROADCAST for vector Vector.broadcast and VectorMask.maskAll operations. 586 // MODE_BITS_COERCED_LONG_TO_MASK for VectorMask.fromLong operation. 587 const TypeInt* mode = gvn().type(argument(5))->isa_int(); 588 589 if (vector_klass == nullptr || vector_klass->const_oop() == nullptr || 590 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 591 vlen == nullptr || !vlen->is_con() || 592 bits_type == nullptr || 593 mode == nullptr || !mode->is_con()) { 594 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s bitwise=%s", 595 NodeClassNames[argument(0)->Opcode()], 596 NodeClassNames[argument(1)->Opcode()], 597 NodeClassNames[argument(2)->Opcode()], 598 NodeClassNames[argument(5)->Opcode()]); 599 return false; // not enough info for intrinsification 600 } 601 602 if (!is_klass_initialized(vector_klass)) { 603 log_if_needed(" ** klass argument not initialized"); 604 return false; 605 } 606 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 607 if (!elem_type->is_primitive_type()) { 608 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 609 return false; // should be primitive type 610 } 611 BasicType elem_bt = elem_type->basic_type(); 612 int num_elem = vlen->get_con(); 613 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 614 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 615 616 bool is_mask = is_vector_mask(vbox_klass); 617 int bcast_mode = mode->get_con(); 618 VectorMaskUseType checkFlags = (VectorMaskUseType)(is_mask ? VecMaskUseAll : VecMaskNotUsed); 619 int opc = bcast_mode == VectorSupport::MODE_BITS_COERCED_LONG_TO_MASK ? Op_VectorLongToMask : Op_Replicate; 620 621 if (!arch_supports_vector(opc, num_elem, elem_bt, checkFlags, true /*has_scalar_args*/)) { 622 log_if_needed(" ** not supported: arity=0 op=broadcast vlen=%d etype=%s ismask=%d bcast_mode=%d", 623 num_elem, type2name(elem_bt), 624 is_mask ? 1 : 0, 625 bcast_mode); 626 return false; // not supported 627 } 628 629 Node* broadcast = nullptr; 630 Node* bits = argument(3); 631 Node* elem = bits; 632 633 if (opc == Op_VectorLongToMask) { 634 const TypeVect* vt = TypeVect::makemask(elem_bt, num_elem); 635 if (vt->isa_vectmask()) { 636 broadcast = gvn().transform(new VectorLongToMaskNode(elem, vt)); 637 } else { 638 const TypeVect* mvt = TypeVect::make(T_BOOLEAN, num_elem); 639 broadcast = gvn().transform(new VectorLongToMaskNode(elem, mvt)); 640 broadcast = gvn().transform(new VectorLoadMaskNode(broadcast, vt)); 641 } 642 } else { 643 switch (elem_bt) { 644 case T_BOOLEAN: // fall-through 645 case T_BYTE: // fall-through 646 case T_SHORT: // fall-through 647 case T_CHAR: // fall-through 648 case T_INT: { 649 elem = gvn().transform(new ConvL2INode(bits)); 650 break; 651 } 652 case T_DOUBLE: { 653 elem = gvn().transform(new MoveL2DNode(bits)); 654 break; 655 } 656 case T_FLOAT: { 657 bits = gvn().transform(new ConvL2INode(bits)); 658 elem = gvn().transform(new MoveI2FNode(bits)); 659 break; 660 } 661 case T_LONG: { 662 // no conversion needed 663 break; 664 } 665 default: fatal("%s", type2name(elem_bt)); 666 } 667 broadcast = VectorNode::scalar2vector(elem, num_elem, elem_bt, is_mask); 668 broadcast = gvn().transform(broadcast); 669 } 670 671 Node* box = box_vector(broadcast, vbox_type, elem_bt, num_elem); 672 set_result(box); 673 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 674 return true; 675 } 676 677 static bool elem_consistent_with_arr(BasicType elem_bt, const TypeAryPtr* arr_type, bool mismatched_ms) { 678 assert(arr_type != nullptr, "unexpected"); 679 BasicType arr_elem_bt = arr_type->elem()->array_element_basic_type(); 680 if (elem_bt == arr_elem_bt) { 681 return true; 682 } else if (elem_bt == T_SHORT && arr_elem_bt == T_CHAR) { 683 // Load/store of short vector from/to char[] is supported 684 return true; 685 } else if (elem_bt == T_BYTE && arr_elem_bt == T_BOOLEAN) { 686 // Load/store of byte vector from/to boolean[] is supported 687 return true; 688 } else { 689 return mismatched_ms; 690 } 691 } 692 693 // public static 694 // <C, 695 // VM extends VectorPayload, 696 // E, 697 // S extends VectorSpecies<E>> 698 // VM load(Class<? extends VM> vmClass, Class<E> eClass, 699 // int length, 700 // Object base, long offset, // Unsafe addressing 701 // boolean fromSegment, 702 // C container, long index, S s, // Arguments for default implementation 703 // LoadOperation<C, VM, S> defaultImpl) { 704 // public static 705 // <C, 706 // V extends VectorPayload> 707 // void store(Class<?> vClass, Class<?> eClass, 708 // int length, 709 // Object base, long offset, // Unsafe addressing 710 // boolean fromSegment, 711 // V v, C container, long index, // Arguments for default implementation 712 // StoreVectorOperation<C, V> defaultImpl) { 713 bool LibraryCallKit::inline_vector_mem_operation(bool is_store) { 714 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 715 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 716 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 717 const TypeInt* from_ms = gvn().type(argument(6))->isa_int(); 718 719 if (vector_klass == nullptr || vector_klass->const_oop() == nullptr || 720 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 721 vlen == nullptr || !vlen->is_con() || 722 from_ms == nullptr || !from_ms->is_con()) { 723 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s from_ms=%s", 724 NodeClassNames[argument(0)->Opcode()], 725 NodeClassNames[argument(1)->Opcode()], 726 NodeClassNames[argument(2)->Opcode()], 727 NodeClassNames[argument(6)->Opcode()]); 728 return false; // not enough info for intrinsification 729 } 730 if (!is_klass_initialized(vector_klass)) { 731 log_if_needed(" ** klass argument not initialized"); 732 return false; 733 } 734 735 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 736 if (!elem_type->is_primitive_type()) { 737 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 738 return false; // should be primitive type 739 } 740 BasicType elem_bt = elem_type->basic_type(); 741 int num_elem = vlen->get_con(); 742 743 // TODO When mask usage is supported, VecMaskNotUsed needs to be VecMaskUseLoad. 744 if (!arch_supports_vector(is_store ? Op_StoreVector : Op_LoadVector, num_elem, elem_bt, VecMaskNotUsed)) { 745 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s ismask=no", 746 is_store, is_store ? "store" : "load", 747 num_elem, type2name(elem_bt)); 748 return false; // not supported 749 } 750 751 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 752 bool is_mask = is_vector_mask(vbox_klass); 753 754 Node* base = argument(3); 755 Node* offset = ConvL2X(argument(4)); 756 757 // Save state and restore on bailout 758 uint old_sp = sp(); 759 SafePointNode* old_map = clone_map(); 760 761 Node* addr = make_unsafe_address(base, offset, (is_mask ? T_BOOLEAN : elem_bt), true); 762 763 // The memory barrier checks are based on ones for unsafe access. 764 // This is not 1-1 implementation. 765 const Type *const base_type = gvn().type(base); 766 767 const TypePtr *addr_type = gvn().type(addr)->isa_ptr(); 768 const TypeAryPtr* arr_type = addr_type->isa_aryptr(); 769 770 const bool in_native = TypePtr::NULL_PTR == base_type; // base always null 771 const bool in_heap = !TypePtr::NULL_PTR->higher_equal(base_type); // base never null 772 773 const bool is_mixed_access = !in_heap && !in_native; 774 775 const bool is_mismatched_access = in_heap && (addr_type->isa_aryptr() == nullptr); 776 777 const bool needs_cpu_membar = is_mixed_access || is_mismatched_access; 778 779 // For non-masked mismatched memory segment vector read/write accesses, intrinsification can continue 780 // with unknown backing storage type and compiler can skip inserting explicit reinterpretation IR after 781 // loading from or before storing to backing storage which is mandatory for semantic correctness of 782 // big-endian memory layout. 783 bool mismatched_ms = LITTLE_ENDIAN_ONLY(false) 784 BIG_ENDIAN_ONLY(from_ms->get_con() && !is_mask && arr_type != nullptr && 785 arr_type->elem()->array_element_basic_type() != elem_bt); 786 BasicType mem_elem_bt = mismatched_ms ? arr_type->elem()->array_element_basic_type() : elem_bt; 787 if (!is_java_primitive(mem_elem_bt)) { 788 log_if_needed(" ** non-primitive array element type"); 789 return false; 790 } 791 int mem_num_elem = mismatched_ms ? (num_elem * type2aelembytes(elem_bt)) / type2aelembytes(mem_elem_bt) : num_elem; 792 if (arr_type != nullptr && !is_mask && !elem_consistent_with_arr(elem_bt, arr_type, mismatched_ms)) { 793 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s atype=%s ismask=no", 794 is_store, is_store ? "store" : "load", 795 num_elem, type2name(elem_bt), type2name(arr_type->elem()->array_element_basic_type())); 796 set_map(old_map); 797 set_sp(old_sp); 798 return false; 799 } 800 801 // In case of mismatched memory segment accesses, we need to double check that the source type memory operations are supported by backend. 802 if (mismatched_ms) { 803 if (is_store) { 804 if (!arch_supports_vector(Op_StoreVector, num_elem, elem_bt, VecMaskNotUsed) 805 || !arch_supports_vector(Op_VectorReinterpret, mem_num_elem, mem_elem_bt, VecMaskNotUsed)) { 806 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d*8 etype=%s/8 ismask=no", 807 is_store, "store", 808 num_elem, type2name(elem_bt)); 809 set_map(old_map); 810 set_sp(old_sp); 811 return false; // not supported 812 } 813 } else { 814 if (!arch_supports_vector(Op_LoadVector, mem_num_elem, mem_elem_bt, VecMaskNotUsed) 815 || !arch_supports_vector(Op_VectorReinterpret, num_elem, elem_bt, VecMaskNotUsed)) { 816 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d*8 etype=%s/8 ismask=no", 817 is_store, "load", 818 mem_num_elem, type2name(mem_elem_bt)); 819 set_map(old_map); 820 set_sp(old_sp); 821 return false; // not supported 822 } 823 } 824 } 825 if (is_mask) { 826 if (!is_store) { 827 if (!arch_supports_vector(Op_LoadVector, num_elem, elem_bt, VecMaskUseLoad)) { 828 set_map(old_map); 829 set_sp(old_sp); 830 return false; // not supported 831 } 832 } else { 833 if (!arch_supports_vector(Op_StoreVector, num_elem, elem_bt, VecMaskUseStore)) { 834 set_map(old_map); 835 set_sp(old_sp); 836 return false; // not supported 837 } 838 } 839 } 840 841 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 842 843 if (needs_cpu_membar) { 844 insert_mem_bar(Op_MemBarCPUOrder); 845 } 846 847 if (is_store) { 848 Node* val = unbox_vector(argument(7), vbox_type, elem_bt, num_elem); 849 if (val == nullptr) { 850 set_map(old_map); 851 set_sp(old_sp); 852 return false; // operand unboxing failed 853 } 854 set_all_memory(reset_memory()); 855 856 // In case the store needs to happen to byte array, reinterpret the incoming vector to byte vector. 857 int store_num_elem = num_elem; 858 if (mismatched_ms) { 859 store_num_elem = mem_num_elem; 860 const TypeVect* to_vect_type = TypeVect::make(mem_elem_bt, store_num_elem); 861 val = gvn().transform(new VectorReinterpretNode(val, val->bottom_type()->is_vect(), to_vect_type)); 862 } 863 if (is_mask) { 864 val = gvn().transform(VectorStoreMaskNode::make(gvn(), val, elem_bt, num_elem)); 865 } 866 Node* vstore = gvn().transform(StoreVectorNode::make(0, control(), memory(addr), addr, addr_type, val, store_num_elem)); 867 set_memory(vstore, addr_type); 868 } else { 869 // When using byte array, we need to load as byte then reinterpret the value. Otherwise, do a simple vector load. 870 Node* vload = nullptr; 871 if (mismatched_ms) { 872 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, mem_num_elem, mem_elem_bt)); 873 const TypeVect* to_vect_type = TypeVect::make(elem_bt, num_elem); 874 vload = gvn().transform(new VectorReinterpretNode(vload, vload->bottom_type()->is_vect(), to_vect_type)); 875 } else { 876 // Special handle for masks 877 if (is_mask) { 878 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, num_elem, T_BOOLEAN)); 879 vload = gvn().transform(new VectorLoadMaskNode(vload, TypeVect::makemask(elem_bt, num_elem))); 880 } else { 881 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, num_elem, elem_bt)); 882 } 883 } 884 Node* box = box_vector(vload, vbox_type, elem_bt, num_elem); 885 set_result(box); 886 } 887 888 destruct_map_clone(old_map); 889 890 if (needs_cpu_membar) { 891 insert_mem_bar(Op_MemBarCPUOrder); 892 } 893 894 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 895 return true; 896 } 897 898 // public static 899 // <C, 900 // V extends Vector<?>, 901 // E, 902 // S extends VectorSpecies<E>, 903 // M extends VectorMask<E>> 904 // V loadMasked(Class<? extends V> vClass, Class<M> mClass, Class<E> eClass, 905 // int length, Object base, long offset, // Unsafe addressing 906 // boolean fromSegment, 907 // M m, int offsetInRange, 908 // C container, long index, S s, // Arguments for default implementation 909 // LoadVectorMaskedOperation<C, V, S, M> defaultImpl) { 910 // public static 911 // <C, 912 // V extends Vector<E>, 913 // M extends VectorMask<E>, 914 // E> 915 // void storeMasked(Class<? extends V> vClass, Class<M> mClass, Class<E> eClass, 916 // int length, 917 // Object base, long offset, // Unsafe addressing 918 // boolean fromSegment, 919 // V v, M m, C container, long index, // Arguments for default implementation 920 // StoreVectorMaskedOperation<C, V, M> defaultImpl) { 921 922 bool LibraryCallKit::inline_vector_mem_masked_operation(bool is_store) { 923 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 924 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 925 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 926 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 927 const TypeInt* from_ms = gvn().type(argument(7))->isa_int(); 928 929 if (vector_klass == nullptr || vector_klass->const_oop() == nullptr || 930 mask_klass == nullptr || mask_klass->const_oop() == nullptr || 931 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 932 vlen == nullptr || !vlen->is_con() || 933 from_ms == nullptr || !from_ms->is_con()) { 934 log_if_needed(" ** missing constant: vclass=%s mclass=%s etype=%s vlen=%s from_ms=%s", 935 NodeClassNames[argument(0)->Opcode()], 936 NodeClassNames[argument(1)->Opcode()], 937 NodeClassNames[argument(2)->Opcode()], 938 NodeClassNames[argument(3)->Opcode()], 939 NodeClassNames[argument(7)->Opcode()]); 940 return false; // not enough info for intrinsification 941 } 942 if (!is_klass_initialized(vector_klass)) { 943 log_if_needed(" ** klass argument not initialized"); 944 return false; 945 } 946 947 if (!is_klass_initialized(mask_klass)) { 948 log_if_needed(" ** mask klass argument not initialized"); 949 return false; 950 } 951 952 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 953 if (!elem_type->is_primitive_type()) { 954 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 955 return false; // should be primitive type 956 } 957 958 BasicType elem_bt = elem_type->basic_type(); 959 int num_elem = vlen->get_con(); 960 961 Node* base = argument(4); 962 Node* offset = ConvL2X(argument(5)); 963 964 // Save state and restore on bailout 965 uint old_sp = sp(); 966 SafePointNode* old_map = clone_map(); 967 968 Node* addr = make_unsafe_address(base, offset, elem_bt, true); 969 const TypePtr *addr_type = gvn().type(addr)->isa_ptr(); 970 const TypeAryPtr* arr_type = addr_type->isa_aryptr(); 971 972 bool mismatched_ms = from_ms->get_con() && arr_type != nullptr && arr_type->elem()->array_element_basic_type() != elem_bt; 973 BIG_ENDIAN_ONLY(if (mismatched_ms) return false;) 974 // If there is no consistency between array and vector element types, it must be special byte array case 975 if (arr_type != nullptr && !elem_consistent_with_arr(elem_bt, arr_type, mismatched_ms)) { 976 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s atype=%s", 977 is_store, is_store ? "storeMasked" : "loadMasked", 978 num_elem, type2name(elem_bt), type2name(arr_type->elem()->array_element_basic_type())); 979 set_map(old_map); 980 set_sp(old_sp); 981 return false; 982 } 983 984 int mem_num_elem = mismatched_ms ? num_elem * type2aelembytes(elem_bt) : num_elem; 985 BasicType mem_elem_bt = mismatched_ms ? T_BYTE : elem_bt; 986 bool supports_predicate = arch_supports_vector(is_store ? Op_StoreVectorMasked : Op_LoadVectorMasked, 987 mem_num_elem, mem_elem_bt, VecMaskUseLoad); 988 989 // If current arch does not support the predicated operations, we have to bail 990 // out when current case uses the predicate feature. 991 if (!supports_predicate) { 992 bool needs_predicate = false; 993 if (is_store) { 994 // Masked vector store always uses the predicated store. 995 needs_predicate = true; 996 } else { 997 // Masked vector load with IOOBE always uses the predicated load. 998 const TypeInt* offset_in_range = gvn().type(argument(9))->isa_int(); 999 if (!offset_in_range->is_con()) { 1000 log_if_needed(" ** missing constant: offsetInRange=%s", 1001 NodeClassNames[argument(8)->Opcode()]); 1002 set_map(old_map); 1003 set_sp(old_sp); 1004 return false; 1005 } 1006 needs_predicate = (offset_in_range->get_con() == 0); 1007 } 1008 1009 if (needs_predicate) { 1010 log_if_needed(" ** not supported: op=%s vlen=%d etype=%s mismatched_ms=%d", 1011 is_store ? "storeMasked" : "loadMasked", 1012 num_elem, type2name(elem_bt), mismatched_ms ? 1 : 0); 1013 set_map(old_map); 1014 set_sp(old_sp); 1015 return false; 1016 } 1017 } 1018 1019 // This only happens for masked vector load. If predicate is not supported, then check whether 1020 // the normal vector load and blend operations are supported by backend. 1021 if (!supports_predicate && (!arch_supports_vector(Op_LoadVector, mem_num_elem, mem_elem_bt, VecMaskNotUsed) || 1022 !arch_supports_vector(Op_VectorBlend, mem_num_elem, mem_elem_bt, VecMaskUseLoad))) { 1023 log_if_needed(" ** not supported: op=loadMasked vlen=%d etype=%s mismatched_ms=%d", 1024 num_elem, type2name(elem_bt), mismatched_ms ? 1 : 0); 1025 set_map(old_map); 1026 set_sp(old_sp); 1027 return false; 1028 } 1029 1030 // Since we are using byte array, we need to double check that the vector reinterpret operation 1031 // with byte type is supported by backend. 1032 if (mismatched_ms) { 1033 if (!arch_supports_vector(Op_VectorReinterpret, mem_num_elem, T_BYTE, VecMaskNotUsed)) { 1034 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s mismatched_ms=1", 1035 is_store, is_store ? "storeMasked" : "loadMasked", 1036 num_elem, type2name(elem_bt)); 1037 set_map(old_map); 1038 set_sp(old_sp); 1039 return false; 1040 } 1041 } 1042 1043 // Since it needs to unbox the mask, we need to double check that the related load operations 1044 // for mask are supported by backend. 1045 if (!arch_supports_vector(Op_LoadVector, num_elem, elem_bt, VecMaskUseLoad)) { 1046 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s", 1047 is_store, is_store ? "storeMasked" : "loadMasked", 1048 num_elem, type2name(elem_bt)); 1049 set_map(old_map); 1050 set_sp(old_sp); 1051 return false; 1052 } 1053 1054 // Can base be null? Otherwise, always on-heap access. 1055 bool can_access_non_heap = TypePtr::NULL_PTR->higher_equal(gvn().type(base)); 1056 if (can_access_non_heap) { 1057 insert_mem_bar(Op_MemBarCPUOrder); 1058 } 1059 1060 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1061 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1062 assert(!is_vector_mask(vbox_klass) && is_vector_mask(mbox_klass), "Invalid class type"); 1063 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1064 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1065 1066 Node* mask = unbox_vector(is_store ? argument(9) : argument(8), mbox_type, elem_bt, num_elem); 1067 if (mask == nullptr) { 1068 log_if_needed(" ** unbox failed mask=%s", 1069 is_store ? NodeClassNames[argument(9)->Opcode()] 1070 : NodeClassNames[argument(8)->Opcode()]); 1071 set_map(old_map); 1072 set_sp(old_sp); 1073 return false; 1074 } 1075 1076 if (is_store) { 1077 Node* val = unbox_vector(argument(8), vbox_type, elem_bt, num_elem); 1078 if (val == nullptr) { 1079 log_if_needed(" ** unbox failed vector=%s", 1080 NodeClassNames[argument(8)->Opcode()]); 1081 set_map(old_map); 1082 set_sp(old_sp); 1083 return false; // operand unboxing failed 1084 } 1085 set_all_memory(reset_memory()); 1086 1087 if (mismatched_ms) { 1088 // Reinterpret the incoming vector to byte vector. 1089 const TypeVect* to_vect_type = TypeVect::make(mem_elem_bt, mem_num_elem); 1090 val = gvn().transform(new VectorReinterpretNode(val, val->bottom_type()->is_vect(), to_vect_type)); 1091 // Reinterpret the vector mask to byte type. 1092 const TypeVect* from_mask_type = TypeVect::makemask(elem_bt, num_elem); 1093 const TypeVect* to_mask_type = TypeVect::makemask(mem_elem_bt, mem_num_elem); 1094 mask = gvn().transform(new VectorReinterpretNode(mask, from_mask_type, to_mask_type)); 1095 } 1096 Node* vstore = gvn().transform(new StoreVectorMaskedNode(control(), memory(addr), addr, val, addr_type, mask)); 1097 set_memory(vstore, addr_type); 1098 } else { 1099 Node* vload = nullptr; 1100 1101 if (mismatched_ms) { 1102 // Reinterpret the vector mask to byte type. 1103 const TypeVect* from_mask_type = TypeVect::makemask(elem_bt, num_elem); 1104 const TypeVect* to_mask_type = TypeVect::makemask(mem_elem_bt, mem_num_elem); 1105 mask = gvn().transform(new VectorReinterpretNode(mask, from_mask_type, to_mask_type)); 1106 } 1107 1108 if (supports_predicate) { 1109 // Generate masked load vector node if predicate feature is supported. 1110 const TypeVect* vt = TypeVect::make(mem_elem_bt, mem_num_elem); 1111 vload = gvn().transform(new LoadVectorMaskedNode(control(), memory(addr), addr, addr_type, vt, mask)); 1112 } else { 1113 // Use the vector blend to implement the masked load vector. The biased elements are zeros. 1114 Node* zero = gvn().transform(gvn().zerocon(mem_elem_bt)); 1115 zero = gvn().transform(VectorNode::scalar2vector(zero, mem_num_elem, mem_elem_bt)); 1116 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, mem_num_elem, mem_elem_bt)); 1117 vload = gvn().transform(new VectorBlendNode(zero, vload, mask)); 1118 } 1119 1120 if (mismatched_ms) { 1121 const TypeVect* to_vect_type = TypeVect::make(elem_bt, num_elem); 1122 vload = gvn().transform(new VectorReinterpretNode(vload, vload->bottom_type()->is_vect(), to_vect_type)); 1123 } 1124 1125 Node* box = box_vector(vload, vbox_type, elem_bt, num_elem); 1126 set_result(box); 1127 } 1128 1129 destruct_map_clone(old_map); 1130 1131 if (can_access_non_heap) { 1132 insert_mem_bar(Op_MemBarCPUOrder); 1133 } 1134 1135 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1136 return true; 1137 } 1138 1139 // <C, 1140 // V extends Vector<?>, 1141 // W extends Vector<Integer>, 1142 // S extends VectorSpecies<E>, 1143 // M extends VectorMask<E>, 1144 // E> 1145 // V loadWithMap(Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, int length, 1146 // Class<? extends Vector<Integer>> vectorIndexClass, 1147 // Object base, long offset, // Unsafe addressing 1148 // W index_vector, M m, 1149 // C container, int index, int[] indexMap, int indexM, S s, // Arguments for default implementation 1150 // LoadVectorOperationWithMap<C, V, E, S, M> defaultImpl) 1151 // 1152 // <C, 1153 // V extends Vector<E>, 1154 // W extends Vector<Integer>, 1155 // M extends VectorMask<E>, 1156 // E> 1157 // void storeWithMap(Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, 1158 // int length, Class<? extends Vector<Integer>> vectorIndexClass, Object base, long offset, // Unsafe addressing 1159 // W index_vector, V v, M m, 1160 // C container, int index, int[] indexMap, int indexM, // Arguments for default implementation 1161 // StoreVectorOperationWithMap<C, V, M, E> defaultImpl) 1162 // 1163 bool LibraryCallKit::inline_vector_gather_scatter(bool is_scatter) { 1164 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1165 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 1166 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1167 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1168 const TypeInstPtr* vector_idx_klass = gvn().type(argument(4))->isa_instptr(); 1169 1170 if (vector_klass == nullptr || vector_klass->const_oop() == nullptr || 1171 // mask_klass == nullptr || mask_klass->const_oop() == nullptr || 1172 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 1173 vlen == nullptr || !vlen->is_con() || 1174 vector_idx_klass == nullptr || vector_idx_klass->const_oop() == nullptr) { 1175 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s viclass=%s", 1176 NodeClassNames[argument(0)->Opcode()], 1177 NodeClassNames[argument(2)->Opcode()], 1178 NodeClassNames[argument(3)->Opcode()], 1179 NodeClassNames[argument(4)->Opcode()]); 1180 return false; // not enough info for intrinsification 1181 } 1182 1183 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(vector_idx_klass)) { 1184 log_if_needed(" ** klass argument not initialized"); 1185 return false; 1186 } 1187 1188 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1189 if (!elem_type->is_primitive_type()) { 1190 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1191 return false; // should be primitive type 1192 } 1193 1194 BasicType elem_bt = elem_type->basic_type(); 1195 int num_elem = vlen->get_con(); 1196 1197 const Type* vmask_type = gvn().type(is_scatter ? argument(10) : argument(9)); 1198 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 1199 if (is_masked_op) { 1200 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 1201 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(1)->Opcode()]); 1202 return false; // not enough info for intrinsification 1203 } 1204 1205 if (!is_klass_initialized(mask_klass)) { 1206 log_if_needed(" ** mask klass argument not initialized"); 1207 return false; 1208 } 1209 1210 if (vmask_type->maybe_null()) { 1211 log_if_needed(" ** null mask values are not allowed for masked op"); 1212 return false; 1213 } 1214 1215 // Check whether the predicated gather/scatter node is supported by architecture. 1216 VectorMaskUseType mask = (VectorMaskUseType) (VecMaskUseLoad | VecMaskUsePred); 1217 if (!arch_supports_vector(is_scatter ? Op_StoreVectorScatterMasked : Op_LoadVectorGatherMasked, num_elem, elem_bt, mask)) { 1218 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s is_masked_op=1", 1219 is_scatter, is_scatter ? "scatterMasked" : "gatherMasked", 1220 num_elem, type2name(elem_bt)); 1221 return false; // not supported 1222 } 1223 } else { 1224 // Check whether the normal gather/scatter node is supported for non-masked operation. 1225 if (!arch_supports_vector(is_scatter ? Op_StoreVectorScatter : Op_LoadVectorGather, num_elem, elem_bt, VecMaskNotUsed)) { 1226 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s is_masked_op=0", 1227 is_scatter, is_scatter ? "scatter" : "gather", 1228 num_elem, type2name(elem_bt)); 1229 return false; // not supported 1230 } 1231 } 1232 1233 // Check that the vector holding indices is supported by architecture 1234 // For sub-word gathers expander receive index array. 1235 if (!is_subword_type(elem_bt) && !arch_supports_vector(Op_LoadVector, num_elem, T_INT, VecMaskNotUsed)) { 1236 log_if_needed(" ** not supported: arity=%d op=%s/loadindex vlen=%d etype=int is_masked_op=%d", 1237 is_scatter, is_scatter ? "scatter" : "gather", 1238 num_elem, is_masked_op ? 1 : 0); 1239 return false; // not supported 1240 } 1241 1242 Node* base = argument(5); 1243 Node* offset = ConvL2X(argument(6)); 1244 1245 // Save state and restore on bailout 1246 uint old_sp = sp(); 1247 SafePointNode* old_map = clone_map(); 1248 1249 Node* addr = make_unsafe_address(base, offset, elem_bt, true); 1250 1251 const TypePtr *addr_type = gvn().type(addr)->isa_ptr(); 1252 const TypeAryPtr* arr_type = addr_type->isa_aryptr(); 1253 1254 // The array must be consistent with vector type 1255 if (arr_type == nullptr || (arr_type != nullptr && !elem_consistent_with_arr(elem_bt, arr_type, false))) { 1256 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s atype=%s ismask=no", 1257 is_scatter, is_scatter ? "scatter" : "gather", 1258 num_elem, type2name(elem_bt), type2name(arr_type->elem()->array_element_basic_type())); 1259 set_map(old_map); 1260 set_sp(old_sp); 1261 return false; 1262 } 1263 1264 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1265 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1266 ciKlass* vbox_idx_klass = vector_idx_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1267 if (vbox_idx_klass == nullptr) { 1268 set_map(old_map); 1269 set_sp(old_sp); 1270 return false; 1271 } 1272 1273 Node* index_vect = nullptr; 1274 const TypeInstPtr* vbox_idx_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_idx_klass); 1275 if (!is_subword_type(elem_bt)) { 1276 index_vect = unbox_vector(argument(8), vbox_idx_type, T_INT, num_elem); 1277 if (index_vect == nullptr) { 1278 set_map(old_map); 1279 set_sp(old_sp); 1280 return false; 1281 } 1282 } 1283 1284 Node* mask = nullptr; 1285 if (is_masked_op) { 1286 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1287 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1288 mask = unbox_vector(is_scatter ? argument(10) : argument(9), mbox_type, elem_bt, num_elem); 1289 if (mask == nullptr) { 1290 log_if_needed(" ** unbox failed mask=%s", 1291 is_scatter ? NodeClassNames[argument(10)->Opcode()] 1292 : NodeClassNames[argument(9)->Opcode()]); 1293 set_map(old_map); 1294 set_sp(old_sp); 1295 return false; 1296 } 1297 } 1298 1299 const TypeVect* vector_type = TypeVect::make(elem_bt, num_elem); 1300 if (is_scatter) { 1301 Node* val = unbox_vector(argument(9), vbox_type, elem_bt, num_elem); 1302 if (val == nullptr) { 1303 set_map(old_map); 1304 set_sp(old_sp); 1305 return false; // operand unboxing failed 1306 } 1307 set_all_memory(reset_memory()); 1308 1309 Node* vstore = nullptr; 1310 if (mask != nullptr) { 1311 vstore = gvn().transform(new StoreVectorScatterMaskedNode(control(), memory(addr), addr, addr_type, val, index_vect, mask)); 1312 } else { 1313 vstore = gvn().transform(new StoreVectorScatterNode(control(), memory(addr), addr, addr_type, val, index_vect)); 1314 } 1315 set_memory(vstore, addr_type); 1316 } else { 1317 Node* vload = nullptr; 1318 Node* index = argument(11); 1319 Node* indexMap = argument(12); 1320 Node* indexM = argument(13); 1321 if (mask != nullptr) { 1322 if (is_subword_type(elem_bt)) { 1323 Node* index_arr_base = array_element_address(indexMap, indexM, T_INT); 1324 vload = gvn().transform(new LoadVectorGatherMaskedNode(control(), memory(addr), addr, addr_type, vector_type, index_arr_base, mask, index)); 1325 } else { 1326 vload = gvn().transform(new LoadVectorGatherMaskedNode(control(), memory(addr), addr, addr_type, vector_type, index_vect, mask)); 1327 } 1328 } else { 1329 if (is_subword_type(elem_bt)) { 1330 Node* index_arr_base = array_element_address(indexMap, indexM, T_INT); 1331 vload = gvn().transform(new LoadVectorGatherNode(control(), memory(addr), addr, addr_type, vector_type, index_arr_base, index)); 1332 } else { 1333 vload = gvn().transform(new LoadVectorGatherNode(control(), memory(addr), addr, addr_type, vector_type, index_vect)); 1334 } 1335 } 1336 Node* box = box_vector(vload, vbox_type, elem_bt, num_elem); 1337 set_result(box); 1338 } 1339 1340 destruct_map_clone(old_map); 1341 1342 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1343 return true; 1344 } 1345 1346 // public static 1347 // <V extends Vector<E>, 1348 // M extends VectorMask<E>, 1349 // E> 1350 // long reductionCoerced(int oprId, Class<? extends V> vectorClass, Class<? extends M> maskClass, 1351 // Class<E> elementType, int length, V v, M m, 1352 // ReductionOperation<V, M> defaultImpl) 1353 bool LibraryCallKit::inline_vector_reduction() { 1354 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 1355 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 1356 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 1357 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 1358 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 1359 1360 if (opr == nullptr || !opr->is_con() || 1361 vector_klass == nullptr || vector_klass->const_oop() == nullptr || 1362 // mask_klass == nullptr || mask_klass->const_oop() == nullptr || 1363 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 1364 vlen == nullptr || !vlen->is_con()) { 1365 log_if_needed(" ** missing constant: opr=%s vclass=%s etype=%s vlen=%s", 1366 NodeClassNames[argument(0)->Opcode()], 1367 NodeClassNames[argument(1)->Opcode()], 1368 NodeClassNames[argument(3)->Opcode()], 1369 NodeClassNames[argument(4)->Opcode()]); 1370 return false; // not enough info for intrinsification 1371 } 1372 if (!is_klass_initialized(vector_klass)) { 1373 log_if_needed(" ** klass argument not initialized"); 1374 return false; 1375 } 1376 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1377 if (!elem_type->is_primitive_type()) { 1378 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1379 return false; // should be primitive type 1380 } 1381 1382 const Type* vmask_type = gvn().type(argument(6)); 1383 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 1384 if (is_masked_op) { 1385 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 1386 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(2)->Opcode()]); 1387 return false; // not enough info for intrinsification 1388 } 1389 1390 if (!is_klass_initialized(mask_klass)) { 1391 log_if_needed(" ** mask klass argument not initialized"); 1392 return false; 1393 } 1394 1395 if (vmask_type->maybe_null()) { 1396 log_if_needed(" ** null mask values are not allowed for masked op"); 1397 return false; 1398 } 1399 } 1400 1401 BasicType elem_bt = elem_type->basic_type(); 1402 int num_elem = vlen->get_con(); 1403 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 1404 int sopc = ReductionNode::opcode(opc, elem_bt); 1405 1406 // Ensure reduction operation for lanewise operation 1407 // When using mask, mask use type needs to be VecMaskUseLoad. 1408 if (sopc == opc || !arch_supports_vector(sopc, num_elem, elem_bt, is_masked_op ? VecMaskUseLoad : VecMaskNotUsed)) { 1409 log_if_needed(" ** not supported: arity=1 op=%d/reduce vlen=%d etype=%s is_masked_op=%d", 1410 sopc, num_elem, type2name(elem_bt), is_masked_op ? 1 : 0); 1411 return false; 1412 } 1413 1414 // Return true if current platform has implemented the masked operation with predicate feature. 1415 bool use_predicate = is_masked_op && arch_supports_vector(sopc, num_elem, elem_bt, VecMaskUsePred); 1416 if (is_masked_op && !use_predicate && !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)) { 1417 log_if_needed(" ** not supported: arity=1 op=%d/reduce vlen=%d etype=%s is_masked_op=1", 1418 sopc, num_elem, type2name(elem_bt)); 1419 return false; 1420 } 1421 1422 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1423 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1424 1425 Node* opd = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1426 if (opd == nullptr) { 1427 return false; // operand unboxing failed 1428 } 1429 1430 Node* mask = nullptr; 1431 if (is_masked_op) { 1432 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1433 assert(is_vector_mask(mbox_klass), "argument(2) should be a mask class"); 1434 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1435 mask = unbox_vector(argument(6), mbox_type, elem_bt, num_elem); 1436 if (mask == nullptr) { 1437 log_if_needed(" ** unbox failed mask=%s", 1438 NodeClassNames[argument(6)->Opcode()]); 1439 return false; 1440 } 1441 } 1442 1443 Node* init = ReductionNode::make_identity_con_scalar(gvn(), opc, elem_bt); 1444 Node* value = opd; 1445 1446 assert(mask != nullptr || !is_masked_op, "Masked op needs the mask value never null"); 1447 if (mask != nullptr && !use_predicate) { 1448 Node* reduce_identity = gvn().transform(VectorNode::scalar2vector(init, num_elem, elem_bt)); 1449 value = gvn().transform(new VectorBlendNode(reduce_identity, value, mask)); 1450 } 1451 1452 // Make an unordered Reduction node. This affects only AddReductionVF/VD and MulReductionVF/VD, 1453 // as these operations are allowed to be associative (not requiring strict order) in VectorAPI. 1454 value = ReductionNode::make(opc, nullptr, init, value, elem_bt, /* requires_strict_order */ false); 1455 1456 if (mask != nullptr && use_predicate) { 1457 value->add_req(mask); 1458 value->add_flag(Node::Flag_is_predicated_vector); 1459 } 1460 1461 value = gvn().transform(value); 1462 1463 Node* bits = nullptr; 1464 switch (elem_bt) { 1465 case T_BYTE: 1466 case T_SHORT: 1467 case T_INT: { 1468 bits = gvn().transform(new ConvI2LNode(value)); 1469 break; 1470 } 1471 case T_FLOAT: { 1472 value = gvn().transform(new MoveF2INode(value)); 1473 bits = gvn().transform(new ConvI2LNode(value)); 1474 break; 1475 } 1476 case T_DOUBLE: { 1477 bits = gvn().transform(new MoveD2LNode(value)); 1478 break; 1479 } 1480 case T_LONG: { 1481 bits = value; // no conversion needed 1482 break; 1483 } 1484 default: fatal("%s", type2name(elem_bt)); 1485 } 1486 set_result(bits); 1487 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1488 return true; 1489 } 1490 1491 // public static <V> boolean test(int cond, Class<?> vectorClass, Class<?> elementType, int vlen, 1492 // V v1, V v2, 1493 // BiFunction<V, V, Boolean> defaultImpl) 1494 // 1495 bool LibraryCallKit::inline_vector_test() { 1496 const TypeInt* cond = gvn().type(argument(0))->isa_int(); 1497 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 1498 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1499 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1500 1501 if (cond == nullptr || !cond->is_con() || 1502 vector_klass == nullptr || vector_klass->const_oop() == nullptr || 1503 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 1504 vlen == nullptr || !vlen->is_con()) { 1505 log_if_needed(" ** missing constant: cond=%s vclass=%s etype=%s vlen=%s", 1506 NodeClassNames[argument(0)->Opcode()], 1507 NodeClassNames[argument(1)->Opcode()], 1508 NodeClassNames[argument(2)->Opcode()], 1509 NodeClassNames[argument(3)->Opcode()]); 1510 return false; // not enough info for intrinsification 1511 } 1512 if (!is_klass_initialized(vector_klass)) { 1513 log_if_needed(" ** klass argument not initialized"); 1514 return false; 1515 } 1516 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1517 if (!elem_type->is_primitive_type()) { 1518 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1519 return false; // should be primitive type 1520 } 1521 BasicType elem_bt = elem_type->basic_type(); 1522 int num_elem = vlen->get_con(); 1523 BoolTest::mask booltest = (BoolTest::mask)cond->get_con(); 1524 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1525 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1526 1527 if (!arch_supports_vector(Op_VectorTest, num_elem, elem_bt, is_vector_mask(vbox_klass) ? VecMaskUseLoad : VecMaskNotUsed)) { 1528 log_if_needed(" ** not supported: arity=2 op=test/%d vlen=%d etype=%s ismask=%d", 1529 cond->get_con(), num_elem, type2name(elem_bt), 1530 is_vector_mask(vbox_klass)); 1531 return false; 1532 } 1533 1534 Node* opd1 = unbox_vector(argument(4), vbox_type, elem_bt, num_elem); 1535 Node* opd2; 1536 if (Matcher::vectortest_needs_second_argument(booltest == BoolTest::overflow, 1537 opd1->bottom_type()->isa_vectmask())) { 1538 opd2 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1539 } else { 1540 opd2 = opd1; 1541 } 1542 if (opd1 == nullptr || opd2 == nullptr) { 1543 return false; // operand unboxing failed 1544 } 1545 1546 Node* cmp = gvn().transform(new VectorTestNode(opd1, opd2, booltest)); 1547 BoolTest::mask test = Matcher::vectortest_mask(booltest == BoolTest::overflow, 1548 opd1->bottom_type()->isa_vectmask(), num_elem); 1549 Node* bol = gvn().transform(new BoolNode(cmp, test)); 1550 Node* res = gvn().transform(new CMoveINode(bol, gvn().intcon(0), gvn().intcon(1), TypeInt::BOOL)); 1551 1552 set_result(res); 1553 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1554 return true; 1555 } 1556 1557 // public static 1558 // <V extends Vector<E>, 1559 // M extends VectorMask<E>, 1560 // E> 1561 // V blend(Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, int vlen, 1562 // V v1, V v2, M m, 1563 // VectorBlendOp<V, M, E> defaultImpl) 1564 bool LibraryCallKit::inline_vector_blend() { 1565 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1566 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 1567 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1568 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1569 1570 if (mask_klass == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 1571 return false; // dead code 1572 } 1573 if (mask_klass->const_oop() == nullptr || vector_klass->const_oop() == nullptr || 1574 elem_klass->const_oop() == nullptr || !vlen->is_con()) { 1575 log_if_needed(" ** missing constant: vclass=%s mclass=%s etype=%s vlen=%s", 1576 NodeClassNames[argument(0)->Opcode()], 1577 NodeClassNames[argument(1)->Opcode()], 1578 NodeClassNames[argument(2)->Opcode()], 1579 NodeClassNames[argument(3)->Opcode()]); 1580 return false; // not enough info for intrinsification 1581 } 1582 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(mask_klass)) { 1583 log_if_needed(" ** klass argument not initialized"); 1584 return false; 1585 } 1586 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1587 if (!elem_type->is_primitive_type()) { 1588 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1589 return false; // should be primitive type 1590 } 1591 BasicType elem_bt = elem_type->basic_type(); 1592 BasicType mask_bt = elem_bt; 1593 int num_elem = vlen->get_con(); 1594 1595 if (!arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)) { 1596 log_if_needed(" ** not supported: arity=2 op=blend vlen=%d etype=%s ismask=useload", 1597 num_elem, type2name(elem_bt)); 1598 return false; // not supported 1599 } 1600 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1601 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1602 1603 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1604 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1605 1606 Node* v1 = unbox_vector(argument(4), vbox_type, elem_bt, num_elem); 1607 Node* v2 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1608 Node* mask = unbox_vector(argument(6), mbox_type, mask_bt, num_elem); 1609 1610 if (v1 == nullptr || v2 == nullptr || mask == nullptr) { 1611 return false; // operand unboxing failed 1612 } 1613 1614 Node* blend = gvn().transform(new VectorBlendNode(v1, v2, mask)); 1615 1616 Node* box = box_vector(blend, vbox_type, elem_bt, num_elem); 1617 set_result(box); 1618 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1619 return true; 1620 } 1621 1622 // public static 1623 // <V extends Vector<E>, 1624 // M extends VectorMask<E>, 1625 // E> 1626 // M compare(int cond, Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, int vlen, 1627 // V v1, V v2, M m, 1628 // VectorCompareOp<V,M> defaultImpl) 1629 bool LibraryCallKit::inline_vector_compare() { 1630 const TypeInt* cond = gvn().type(argument(0))->isa_int(); 1631 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 1632 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 1633 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 1634 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 1635 1636 if (cond == nullptr || vector_klass == nullptr || mask_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 1637 return false; // dead code 1638 } 1639 if (!cond->is_con() || vector_klass->const_oop() == nullptr || mask_klass->const_oop() == nullptr || 1640 elem_klass->const_oop() == nullptr || !vlen->is_con()) { 1641 log_if_needed(" ** missing constant: cond=%s vclass=%s mclass=%s etype=%s vlen=%s", 1642 NodeClassNames[argument(0)->Opcode()], 1643 NodeClassNames[argument(1)->Opcode()], 1644 NodeClassNames[argument(2)->Opcode()], 1645 NodeClassNames[argument(3)->Opcode()], 1646 NodeClassNames[argument(4)->Opcode()]); 1647 return false; // not enough info for intrinsification 1648 } 1649 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(mask_klass)) { 1650 log_if_needed(" ** klass argument not initialized"); 1651 return false; 1652 } 1653 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1654 if (!elem_type->is_primitive_type()) { 1655 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1656 return false; // should be primitive type 1657 } 1658 1659 int num_elem = vlen->get_con(); 1660 BasicType elem_bt = elem_type->basic_type(); 1661 BasicType mask_bt = elem_bt; 1662 1663 if ((cond->get_con() & BoolTest::unsigned_compare) != 0) { 1664 if (!Matcher::supports_vector_comparison_unsigned(num_elem, elem_bt)) { 1665 log_if_needed(" ** not supported: unsigned comparison op=comp/%d vlen=%d etype=%s ismask=usestore", 1666 cond->get_con() & (BoolTest::unsigned_compare - 1), num_elem, type2name(elem_bt)); 1667 return false; 1668 } 1669 } 1670 1671 if (!arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskUseStore)) { 1672 log_if_needed(" ** not supported: arity=2 op=comp/%d vlen=%d etype=%s ismask=usestore", 1673 cond->get_con(), num_elem, type2name(elem_bt)); 1674 return false; 1675 } 1676 1677 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1678 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1679 1680 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1681 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1682 1683 Node* v1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1684 Node* v2 = unbox_vector(argument(6), vbox_type, elem_bt, num_elem); 1685 1686 bool is_masked_op = argument(7)->bottom_type() != TypePtr::NULL_PTR; 1687 Node* mask = is_masked_op ? unbox_vector(argument(7), mbox_type, elem_bt, num_elem) : nullptr; 1688 if (is_masked_op && mask == nullptr) { 1689 log_if_needed(" ** not supported: mask = null arity=2 op=comp/%d vlen=%d etype=%s ismask=usestore is_masked_op=1", 1690 cond->get_con(), num_elem, type2name(elem_bt)); 1691 return false; 1692 } 1693 1694 bool use_predicate = is_masked_op && arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskUsePred); 1695 if (is_masked_op && !use_predicate && !arch_supports_vector(Op_AndV, num_elem, elem_bt, VecMaskUseLoad)) { 1696 log_if_needed(" ** not supported: arity=2 op=comp/%d vlen=%d etype=%s ismask=usestore is_masked_op=1", 1697 cond->get_con(), num_elem, type2name(elem_bt)); 1698 return false; 1699 } 1700 1701 if (v1 == nullptr || v2 == nullptr) { 1702 return false; // operand unboxing failed 1703 } 1704 BoolTest::mask pred = (BoolTest::mask)cond->get_con(); 1705 ConINode* pred_node = (ConINode*)gvn().makecon(cond); 1706 1707 const TypeVect* vmask_type = TypeVect::makemask(mask_bt, num_elem); 1708 Node* operation = new VectorMaskCmpNode(pred, v1, v2, pred_node, vmask_type); 1709 1710 if (is_masked_op) { 1711 if (use_predicate) { 1712 operation->add_req(mask); 1713 operation->add_flag(Node::Flag_is_predicated_vector); 1714 } else { 1715 operation = gvn().transform(operation); 1716 operation = VectorNode::make(Op_AndV, operation, mask, vmask_type); 1717 } 1718 } 1719 1720 operation = gvn().transform(operation); 1721 1722 Node* box = box_vector(operation, mbox_type, mask_bt, num_elem); 1723 set_result(box); 1724 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1725 return true; 1726 } 1727 1728 // public static 1729 // <V extends Vector<E>, 1730 // Sh extends VectorShuffle<E>, 1731 // M extends VectorMask<E>, 1732 // E> 1733 // V rearrangeOp(Class<? extends V> vectorClass, Class<Sh> shuffleClass, Class<M> maskClass, Class<E> elementType, int vlen, 1734 // V v1, Sh sh, M m, 1735 // VectorRearrangeOp<V, Sh, M, E> defaultImpl) 1736 bool LibraryCallKit::inline_vector_rearrange() { 1737 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1738 const TypeInstPtr* shuffle_klass = gvn().type(argument(1))->isa_instptr(); 1739 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 1740 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 1741 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 1742 1743 if (vector_klass == nullptr || shuffle_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 1744 return false; // dead code 1745 } 1746 if (shuffle_klass->const_oop() == nullptr || 1747 vector_klass->const_oop() == nullptr || 1748 elem_klass->const_oop() == nullptr || 1749 !vlen->is_con()) { 1750 log_if_needed(" ** missing constant: vclass=%s sclass=%s etype=%s vlen=%s", 1751 NodeClassNames[argument(0)->Opcode()], 1752 NodeClassNames[argument(1)->Opcode()], 1753 NodeClassNames[argument(3)->Opcode()], 1754 NodeClassNames[argument(4)->Opcode()]); 1755 return false; // not enough info for intrinsification 1756 } 1757 if (!is_klass_initialized(vector_klass) || 1758 !is_klass_initialized(shuffle_klass)) { 1759 log_if_needed(" ** klass argument not initialized"); 1760 return false; 1761 } 1762 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1763 if (!elem_type->is_primitive_type()) { 1764 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1765 return false; // should be primitive type 1766 } 1767 1768 BasicType elem_bt = elem_type->basic_type(); 1769 BasicType shuffle_bt = elem_bt; 1770 if (shuffle_bt == T_FLOAT) { 1771 shuffle_bt = T_INT; 1772 } else if (shuffle_bt == T_DOUBLE) { 1773 shuffle_bt = T_LONG; 1774 } 1775 1776 int num_elem = vlen->get_con(); 1777 bool need_load_shuffle = Matcher::vector_rearrange_requires_load_shuffle(shuffle_bt, num_elem); 1778 1779 if (need_load_shuffle && !arch_supports_vector(Op_VectorLoadShuffle, num_elem, shuffle_bt, VecMaskNotUsed)) { 1780 if (C->print_intrinsics()) { 1781 tty->print_cr(" ** not supported: arity=0 op=load/shuffle vlen=%d etype=%s ismask=no", 1782 num_elem, type2name(shuffle_bt)); 1783 } 1784 return false; // not supported 1785 } 1786 1787 bool is_masked_op = argument(7)->bottom_type() != TypePtr::NULL_PTR; 1788 bool use_predicate = is_masked_op; 1789 if (is_masked_op && 1790 (mask_klass == nullptr || 1791 mask_klass->const_oop() == nullptr || 1792 !is_klass_initialized(mask_klass))) { 1793 log_if_needed(" ** mask_klass argument not initialized"); 1794 } 1795 if (!arch_supports_vector(Op_AndV, num_elem, elem_bt, VecMaskNotUsed)) { 1796 log_if_needed(" ** not supported: arity=2 op=and vlen=%d etype=%s ismask=no", 1797 num_elem, type2name(elem_bt)); 1798 return false; 1799 } 1800 VectorMaskUseType checkFlags = (VectorMaskUseType)(is_masked_op ? (VecMaskUseLoad | VecMaskUsePred) : VecMaskNotUsed); 1801 if (!arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, checkFlags)) { 1802 use_predicate = false; 1803 if(!is_masked_op || 1804 (!arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, VecMaskNotUsed) || 1805 !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad) || 1806 !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed))) { 1807 log_if_needed(" ** not supported: arity=2 op=shuffle/rearrange vlen=%d etype=%s ismask=no", 1808 num_elem, type2name(elem_bt)); 1809 return false; // not supported 1810 } 1811 } 1812 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1813 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1814 1815 ciKlass* shbox_klass = shuffle_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1816 const TypeInstPtr* shbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, shbox_klass); 1817 1818 Node* v1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1819 Node* shuffle = unbox_vector(argument(6), shbox_type, shuffle_bt, num_elem); 1820 const TypeVect* st = TypeVect::make(shuffle_bt, num_elem); 1821 1822 if (v1 == nullptr || shuffle == nullptr) { 1823 return false; // operand unboxing failed 1824 } 1825 1826 assert(is_power_of_2(num_elem), "wrapping invalid"); 1827 Node* wrapping_mask_elem = gvn().makecon(TypeInteger::make(num_elem - 1, num_elem - 1, Type::WidenMin, shuffle_bt == T_LONG ? T_LONG : T_INT)); 1828 Node* wrapping_mask = gvn().transform(VectorNode::scalar2vector(wrapping_mask_elem, num_elem, shuffle_bt)); 1829 shuffle = gvn().transform(new AndVNode(shuffle, wrapping_mask, st)); 1830 1831 Node* mask = nullptr; 1832 if (is_masked_op) { 1833 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1834 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1835 mask = unbox_vector(argument(7), mbox_type, elem_bt, num_elem); 1836 if (mask == nullptr) { 1837 log_if_needed(" ** not supported: arity=3 op=shuffle/rearrange vlen=%d etype=%s ismask=useload is_masked_op=1", 1838 num_elem, type2name(elem_bt)); 1839 return false; 1840 } 1841 } 1842 1843 if (need_load_shuffle) { 1844 shuffle = gvn().transform(new VectorLoadShuffleNode(shuffle, st)); 1845 } 1846 1847 Node* rearrange = new VectorRearrangeNode(v1, shuffle); 1848 if (is_masked_op) { 1849 if (use_predicate) { 1850 rearrange->add_req(mask); 1851 rearrange->add_flag(Node::Flag_is_predicated_vector); 1852 } else { 1853 rearrange = gvn().transform(rearrange); 1854 Node* zero = gvn().makecon(Type::get_zero_type(elem_bt)); 1855 Node* zerovec = gvn().transform(VectorNode::scalar2vector(zero, num_elem, elem_bt)); 1856 rearrange = new VectorBlendNode(zerovec, rearrange, mask); 1857 } 1858 } 1859 rearrange = gvn().transform(rearrange); 1860 1861 Node* box = box_vector(rearrange, vbox_type, elem_bt, num_elem); 1862 set_result(box); 1863 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1864 return true; 1865 } 1866 1867 static address get_vector_math_address(int vop, int bits, BasicType bt, char* name_ptr, int name_len) { 1868 address addr = nullptr; 1869 assert(UseVectorStubs, "sanity"); 1870 assert(name_ptr != nullptr, "unexpected"); 1871 assert((vop >= VectorSupport::VECTOR_OP_MATH_START) && (vop <= VectorSupport::VECTOR_OP_MATH_END), "unexpected"); 1872 int op = vop - VectorSupport::VECTOR_OP_MATH_START; 1873 1874 switch(bits) { 1875 case 64: //fallthough 1876 case 128: //fallthough 1877 case 256: //fallthough 1878 case 512: 1879 if (bt == T_FLOAT) { 1880 snprintf(name_ptr, name_len, "vector_%s_float_%dbits_fixed", VectorSupport::mathname[op], bits); 1881 addr = StubRoutines::_vector_f_math[exact_log2(bits/64)][op]; 1882 } else { 1883 assert(bt == T_DOUBLE, "must be FP type only"); 1884 snprintf(name_ptr, name_len, "vector_%s_double_%dbits_fixed", VectorSupport::mathname[op], bits); 1885 addr = StubRoutines::_vector_d_math[exact_log2(bits/64)][op]; 1886 } 1887 break; 1888 default: 1889 if (!Matcher::supports_scalable_vector() || !Matcher::vector_size_supported(bt, bits/type2aelembytes(bt)) ) { 1890 snprintf(name_ptr, name_len, "invalid"); 1891 addr = nullptr; 1892 Unimplemented(); 1893 } 1894 break; 1895 } 1896 1897 if (addr == nullptr && Matcher::supports_scalable_vector()) { 1898 if (bt == T_FLOAT) { 1899 snprintf(name_ptr, name_len, "vector_%s_float_%dbits_scalable", VectorSupport::mathname[op], bits); 1900 addr = StubRoutines::_vector_f_math[VectorSupport::VEC_SIZE_SCALABLE][op]; 1901 } else { 1902 assert(bt == T_DOUBLE, "must be FP type only"); 1903 snprintf(name_ptr, name_len, "vector_%s_double_%dbits_scalable", VectorSupport::mathname[op], bits); 1904 addr = StubRoutines::_vector_d_math[VectorSupport::VEC_SIZE_SCALABLE][op]; 1905 } 1906 } 1907 1908 return addr; 1909 } 1910 1911 // public static 1912 // <V extends Vector<E>, 1913 // M extends VectorMask<E>, 1914 // E> 1915 // V selectFromOp(Class<? extends V> vClass, Class<M> mClass, Class<E> eClass, 1916 // int length, V v1, V v2, M m, 1917 // VectorSelectFromOp<V, M> defaultImpl) 1918 bool LibraryCallKit::inline_vector_select_from() { 1919 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1920 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 1921 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1922 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1923 1924 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 1925 vector_klass->const_oop() == nullptr || 1926 elem_klass->const_oop() == nullptr || 1927 !vlen->is_con()) { 1928 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s", 1929 NodeClassNames[argument(0)->Opcode()], 1930 NodeClassNames[argument(2)->Opcode()], 1931 NodeClassNames[argument(3)->Opcode()]); 1932 return false; // not enough info for intrinsification 1933 } 1934 if (!is_klass_initialized(vector_klass)) { 1935 log_if_needed(" ** klass argument not initialized"); 1936 return false; 1937 } 1938 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1939 if (!elem_type->is_primitive_type()) { 1940 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1941 return false; // should be primitive type 1942 } 1943 BasicType elem_bt = elem_type->basic_type(); 1944 int num_elem = vlen->get_con(); 1945 if (!is_power_of_2(num_elem)) { 1946 log_if_needed(" ** vlen not power of two=%d", num_elem); 1947 return false; 1948 } 1949 1950 BasicType shuffle_bt = elem_bt; 1951 if (shuffle_bt == T_FLOAT) { 1952 shuffle_bt = T_INT; 1953 } else if (shuffle_bt == T_DOUBLE) { 1954 shuffle_bt = T_LONG; 1955 } 1956 bool need_load_shuffle = Matcher::vector_rearrange_requires_load_shuffle(shuffle_bt, num_elem); 1957 1958 int cast_vopc = VectorCastNode::opcode(-1, elem_bt); // from vector of type elem_bt 1959 if ((need_load_shuffle && !arch_supports_vector(Op_VectorLoadShuffle, num_elem, elem_bt, VecMaskNotUsed)) || 1960 (elem_bt != shuffle_bt && !arch_supports_vector(cast_vopc, num_elem, shuffle_bt, VecMaskNotUsed)) || 1961 !arch_supports_vector(Op_AndV, num_elem, shuffle_bt, VecMaskNotUsed) || 1962 !arch_supports_vector(Op_Replicate, num_elem, shuffle_bt, VecMaskNotUsed)) { 1963 log_if_needed(" ** not supported: arity=0 op=selectFrom vlen=%d etype=%s ismask=no", 1964 num_elem, type2name(elem_bt)); 1965 return false; // not supported 1966 } 1967 1968 bool is_masked_op = argument(6)->bottom_type() != TypePtr::NULL_PTR; 1969 bool use_predicate = is_masked_op; 1970 if (is_masked_op && 1971 (mask_klass == nullptr || 1972 mask_klass->const_oop() == nullptr || 1973 !is_klass_initialized(mask_klass))) { 1974 log_if_needed(" ** mask_klass argument not initialized"); 1975 return false; // not supported 1976 } 1977 VectorMaskUseType checkFlags = (VectorMaskUseType)(is_masked_op ? (VecMaskUseLoad | VecMaskUsePred) : VecMaskNotUsed); 1978 if (!arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, checkFlags)) { 1979 use_predicate = false; 1980 if(!is_masked_op || 1981 (!arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, VecMaskNotUsed) || 1982 !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad) || 1983 !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed))) { 1984 log_if_needed(" ** not supported: op=selectFrom vlen=%d etype=%s is_masked_op=%d", 1985 num_elem, type2name(elem_bt), is_masked_op); 1986 return false; // not supported 1987 } 1988 } 1989 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1990 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1991 1992 // v1 is the index vector 1993 Node* v1 = unbox_vector(argument(4), vbox_type, elem_bt, num_elem); 1994 // v2 is the vector being rearranged 1995 Node* v2 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1996 1997 if (v1 == nullptr) { 1998 log_if_needed(" ** unbox failed v1=%s", NodeClassNames[argument(4)->Opcode()]); 1999 return false; // operand unboxing failed 2000 } 2001 2002 if (v2 == nullptr) { 2003 log_if_needed(" ** unbox failed v2=%s", NodeClassNames[argument(5)->Opcode()]); 2004 return false; // operand unboxing failed 2005 } 2006 2007 Node* mask = nullptr; 2008 if (is_masked_op) { 2009 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2010 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 2011 mask = unbox_vector(argument(6), mbox_type, elem_bt, num_elem); 2012 if (mask == nullptr) { 2013 log_if_needed(" ** unbox failed mask=%s", NodeClassNames[argument(6)->Opcode()]); 2014 return false; 2015 } 2016 } 2017 2018 // cast index vector from elem_bt vector to byte vector 2019 const TypeVect* shuffle_vt = TypeVect::make(shuffle_bt, num_elem); 2020 Node* shuffle = v1; 2021 2022 if (shuffle_bt != elem_bt) { 2023 shuffle = gvn().transform(VectorCastNode::make(cast_vopc, v1, shuffle_bt, num_elem)); 2024 } 2025 2026 // wrap the byte vector lanes to (num_elem - 1) to form the shuffle vector where num_elem is vector length 2027 // this is a simple AND operation as we come here only for power of two vector length 2028 Node* mod_val = gvn().makecon(TypeInteger::make(num_elem - 1, num_elem - 1, Type::WidenMin, shuffle_bt == T_LONG ? T_LONG : T_INT)); 2029 Node* bcast_mod = gvn().transform(VectorNode::scalar2vector(mod_val, num_elem, shuffle_bt)); 2030 shuffle = gvn().transform(VectorNode::make(Op_AndV, shuffle, bcast_mod, shuffle_vt)); 2031 2032 // load the shuffle to use in rearrange 2033 if (need_load_shuffle) { 2034 shuffle = gvn().transform(new VectorLoadShuffleNode(shuffle, shuffle_vt)); 2035 } 2036 2037 // and finally rearrange 2038 Node* rearrange = new VectorRearrangeNode(v2, shuffle); 2039 if (is_masked_op) { 2040 if (use_predicate) { 2041 // masked rearrange is supported so use that directly 2042 rearrange->add_req(mask); 2043 rearrange->add_flag(Node::Flag_is_predicated_vector); 2044 } else { 2045 // masked rearrange is not supported so emulate usig blend 2046 const TypeVect* vt = v1->bottom_type()->is_vect(); 2047 rearrange = gvn().transform(rearrange); 2048 2049 // create a zero vector with each lane element set as zero 2050 Node* zero = gvn().makecon(Type::get_zero_type(elem_bt)); 2051 Node* zerovec = gvn().transform(VectorNode::scalar2vector(zero, num_elem, elem_bt)); 2052 2053 // For each lane for which mask is set, blend in the rearranged lane into zero vector 2054 rearrange = new VectorBlendNode(zerovec, rearrange, mask); 2055 } 2056 } 2057 rearrange = gvn().transform(rearrange); 2058 2059 // box the result 2060 Node* box = box_vector(rearrange, vbox_type, elem_bt, num_elem); 2061 set_result(box); 2062 2063 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2064 return true; 2065 } 2066 2067 Node* LibraryCallKit::gen_call_to_vector_math(int vector_api_op_id, BasicType bt, int num_elem, Node* opd1, Node* opd2) { 2068 assert(UseVectorStubs, "sanity"); 2069 assert(vector_api_op_id >= VectorSupport::VECTOR_OP_MATH_START && vector_api_op_id <= VectorSupport::VECTOR_OP_MATH_END, "need valid op id"); 2070 assert(opd1 != nullptr, "must not be null"); 2071 const TypeVect* vt = TypeVect::make(bt, num_elem); 2072 const TypeFunc* call_type = OptoRuntime::Math_Vector_Vector_Type(opd2 != nullptr ? 2 : 1, vt, vt); 2073 char name[100] = ""; 2074 2075 // Get address for vector math method. 2076 address addr = get_vector_math_address(vector_api_op_id, vt->length_in_bytes() * BitsPerByte, bt, name, 100); 2077 2078 if (addr == nullptr) { 2079 return nullptr; 2080 } 2081 2082 assert(name[0] != '\0', "name must not be null"); 2083 Node* operation = make_runtime_call(RC_VECTOR, 2084 call_type, 2085 addr, 2086 name, 2087 TypePtr::BOTTOM, 2088 opd1, 2089 opd2); 2090 return gvn().transform(new ProjNode(gvn().transform(operation), TypeFunc::Parms)); 2091 } 2092 2093 // public static 2094 // <V extends Vector<E>, 2095 // M extends VectorMask<E>, 2096 // E> 2097 // V broadcastInt(int opr, Class<? extends V> vectorClass, Class<? extends M> maskClass, 2098 // Class<E> elementType, int length, 2099 // V v, int n, M m, 2100 // VectorBroadcastIntOp<V, M> defaultImpl) 2101 bool LibraryCallKit::inline_vector_broadcast_int() { 2102 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 2103 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 2104 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 2105 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 2106 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 2107 2108 if (opr == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 2109 return false; // dead code 2110 } 2111 if (!opr->is_con() || vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 2112 log_if_needed(" ** missing constant: opr=%s vclass=%s etype=%s vlen=%s", 2113 NodeClassNames[argument(0)->Opcode()], 2114 NodeClassNames[argument(1)->Opcode()], 2115 NodeClassNames[argument(3)->Opcode()], 2116 NodeClassNames[argument(4)->Opcode()]); 2117 return false; // not enough info for intrinsification 2118 } 2119 if (!is_klass_initialized(vector_klass)) { 2120 log_if_needed(" ** klass argument not initialized"); 2121 return false; 2122 } 2123 2124 const Type* vmask_type = gvn().type(argument(7)); 2125 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 2126 if (is_masked_op) { 2127 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 2128 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(2)->Opcode()]); 2129 return false; // not enough info for intrinsification 2130 } 2131 2132 if (!is_klass_initialized(mask_klass)) { 2133 log_if_needed(" ** mask klass argument not initialized"); 2134 return false; 2135 } 2136 2137 if (vmask_type->maybe_null()) { 2138 log_if_needed(" ** null mask values are not allowed for masked op"); 2139 return false; 2140 } 2141 } 2142 2143 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2144 if (!elem_type->is_primitive_type()) { 2145 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2146 return false; // should be primitive type 2147 } 2148 2149 int num_elem = vlen->get_con(); 2150 BasicType elem_bt = elem_type->basic_type(); 2151 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 2152 2153 bool is_shift = VectorNode::is_shift_opcode(opc); 2154 bool is_rotate = VectorNode::is_rotate_opcode(opc); 2155 2156 if (opc == 0 || (!is_shift && !is_rotate)) { 2157 log_if_needed(" ** operation not supported: op=%d bt=%s", opr->get_con(), type2name(elem_bt)); 2158 return false; // operation not supported 2159 } 2160 2161 int sopc = VectorNode::opcode(opc, elem_bt); 2162 if (sopc == 0) { 2163 log_if_needed(" ** operation not supported: opc=%s bt=%s", NodeClassNames[opc], type2name(elem_bt)); 2164 return false; // operation not supported 2165 } 2166 2167 Node* cnt = argument(6); 2168 const TypeInt* cnt_type = cnt->bottom_type()->isa_int(); 2169 2170 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2171 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2172 2173 // If CPU supports vector constant rotate instructions pass it directly 2174 bool is_const_rotate = is_rotate && cnt_type && cnt_type->is_con() && 2175 Matcher::supports_vector_constant_rotates(cnt_type->get_con()); 2176 bool has_scalar_args = is_rotate ? !is_const_rotate : true; 2177 2178 VectorMaskUseType checkFlags = (VectorMaskUseType)(is_masked_op ? (VecMaskUseLoad | VecMaskUsePred) : VecMaskNotUsed); 2179 bool use_predicate = is_masked_op; 2180 2181 if (!arch_supports_vector(sopc, num_elem, elem_bt, checkFlags, has_scalar_args)) { 2182 use_predicate = false; 2183 if (!is_masked_op || 2184 (!arch_supports_vector(sopc, num_elem, elem_bt, VecMaskNotUsed, has_scalar_args) || 2185 !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad))) { 2186 2187 log_if_needed(" ** not supported: arity=0 op=int/%d vlen=%d etype=%s is_masked_op=%d", 2188 sopc, num_elem, type2name(elem_bt), is_masked_op ? 1 : 0); 2189 return false; // not supported 2190 } 2191 } 2192 2193 Node* opd1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 2194 Node* opd2 = nullptr; 2195 if (is_shift) { 2196 opd2 = vector_shift_count(cnt, opc, elem_bt, num_elem); 2197 } else { 2198 assert(is_rotate, "unexpected operation"); 2199 if (!is_const_rotate) { 2200 cnt = elem_bt == T_LONG ? gvn().transform(new ConvI2LNode(cnt)) : cnt; 2201 opd2 = gvn().transform(VectorNode::scalar2vector(cnt, num_elem, elem_bt)); 2202 } else { 2203 // Constant shift value. 2204 opd2 = cnt; 2205 } 2206 } 2207 2208 if (opd1 == nullptr || opd2 == nullptr) { 2209 return false; 2210 } 2211 2212 Node* mask = nullptr; 2213 if (is_masked_op) { 2214 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2215 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 2216 mask = unbox_vector(argument(7), mbox_type, elem_bt, num_elem); 2217 if (mask == nullptr) { 2218 log_if_needed(" ** unbox failed mask=%s", NodeClassNames[argument(7)->Opcode()]); 2219 return false; 2220 } 2221 } 2222 2223 Node* operation = VectorNode::make(opc, opd1, opd2, num_elem, elem_bt); 2224 if (is_masked_op && mask != nullptr) { 2225 if (use_predicate) { 2226 operation->add_req(mask); 2227 operation->add_flag(Node::Flag_is_predicated_vector); 2228 } else { 2229 operation = gvn().transform(operation); 2230 operation = new VectorBlendNode(opd1, operation, mask); 2231 } 2232 } 2233 operation = gvn().transform(operation); 2234 Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem); 2235 set_result(vbox); 2236 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2237 return true; 2238 } 2239 2240 // public static <VOUT extends VectorPayload, 2241 // VIN extends VectorPayload, 2242 // S extends VectorSpecies> 2243 // VOUT convert(int oprId, 2244 // Class<?> fromVectorClass, Class<?> fromElementType, int fromVLen, 2245 // Class<?> toVectorClass, Class<?> toElementType, int toVLen, 2246 // VIN v, S s, 2247 // VectorConvertOp<VOUT, VIN, S> defaultImpl) 2248 // 2249 bool LibraryCallKit::inline_vector_convert() { 2250 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 2251 2252 const TypeInstPtr* vector_klass_from = gvn().type(argument(1))->isa_instptr(); 2253 const TypeInstPtr* elem_klass_from = gvn().type(argument(2))->isa_instptr(); 2254 const TypeInt* vlen_from = gvn().type(argument(3))->isa_int(); 2255 2256 const TypeInstPtr* vector_klass_to = gvn().type(argument(4))->isa_instptr(); 2257 const TypeInstPtr* elem_klass_to = gvn().type(argument(5))->isa_instptr(); 2258 const TypeInt* vlen_to = gvn().type(argument(6))->isa_int(); 2259 2260 if (opr == nullptr || 2261 vector_klass_from == nullptr || elem_klass_from == nullptr || vlen_from == nullptr || 2262 vector_klass_to == nullptr || elem_klass_to == nullptr || vlen_to == nullptr) { 2263 return false; // dead code 2264 } 2265 if (!opr->is_con() || 2266 vector_klass_from->const_oop() == nullptr || elem_klass_from->const_oop() == nullptr || !vlen_from->is_con() || 2267 vector_klass_to->const_oop() == nullptr || elem_klass_to->const_oop() == nullptr || !vlen_to->is_con()) { 2268 log_if_needed(" ** missing constant: opr=%s vclass_from=%s etype_from=%s vlen_from=%s vclass_to=%s etype_to=%s vlen_to=%s", 2269 NodeClassNames[argument(0)->Opcode()], 2270 NodeClassNames[argument(1)->Opcode()], 2271 NodeClassNames[argument(2)->Opcode()], 2272 NodeClassNames[argument(3)->Opcode()], 2273 NodeClassNames[argument(4)->Opcode()], 2274 NodeClassNames[argument(5)->Opcode()], 2275 NodeClassNames[argument(6)->Opcode()]); 2276 return false; // not enough info for intrinsification 2277 } 2278 if (!is_klass_initialized(vector_klass_from) || !is_klass_initialized(vector_klass_to)) { 2279 log_if_needed(" ** klass argument not initialized"); 2280 return false; 2281 } 2282 2283 assert(opr->get_con() == VectorSupport::VECTOR_OP_CAST || 2284 opr->get_con() == VectorSupport::VECTOR_OP_UCAST || 2285 opr->get_con() == VectorSupport::VECTOR_OP_REINTERPRET, "wrong opcode"); 2286 bool is_cast = (opr->get_con() == VectorSupport::VECTOR_OP_CAST || opr->get_con() == VectorSupport::VECTOR_OP_UCAST); 2287 bool is_ucast = (opr->get_con() == VectorSupport::VECTOR_OP_UCAST); 2288 2289 ciKlass* vbox_klass_from = vector_klass_from->const_oop()->as_instance()->java_lang_Class_klass(); 2290 ciKlass* vbox_klass_to = vector_klass_to->const_oop()->as_instance()->java_lang_Class_klass(); 2291 2292 bool is_mask = is_vector_mask(vbox_klass_from); 2293 2294 ciType* elem_type_from = elem_klass_from->const_oop()->as_instance()->java_mirror_type(); 2295 if (!elem_type_from->is_primitive_type()) { 2296 return false; // should be primitive type 2297 } 2298 BasicType elem_bt_from = elem_type_from->basic_type(); 2299 ciType* elem_type_to = elem_klass_to->const_oop()->as_instance()->java_mirror_type(); 2300 if (!elem_type_to->is_primitive_type()) { 2301 return false; // should be primitive type 2302 } 2303 BasicType elem_bt_to = elem_type_to->basic_type(); 2304 2305 int num_elem_from = vlen_from->get_con(); 2306 int num_elem_to = vlen_to->get_con(); 2307 2308 // Check whether we can unbox to appropriate size. Even with casting, checking for reinterpret is needed 2309 // since we may need to change size. 2310 if (!arch_supports_vector(Op_VectorReinterpret, 2311 num_elem_from, 2312 elem_bt_from, 2313 is_mask ? VecMaskUseAll : VecMaskNotUsed)) { 2314 log_if_needed(" ** not supported: arity=1 op=%s/1 vlen1=%d etype1=%s ismask=%d", 2315 is_cast ? "cast" : "reinterpret", 2316 num_elem_from, type2name(elem_bt_from), is_mask); 2317 return false; 2318 } 2319 2320 // Check whether we can support resizing/reinterpreting to the new size. 2321 if (!arch_supports_vector(Op_VectorReinterpret, 2322 num_elem_to, 2323 elem_bt_to, 2324 is_mask ? VecMaskUseAll : VecMaskNotUsed)) { 2325 log_if_needed(" ** not supported: arity=1 op=%s/2 vlen2=%d etype2=%s ismask=%d", 2326 is_cast ? "cast" : "reinterpret", 2327 num_elem_to, type2name(elem_bt_to), is_mask); 2328 return false; 2329 } 2330 2331 // At this point, we know that both input and output vector registers are supported 2332 // by the architecture. Next check if the casted type is simply to same type - which means 2333 // that it is actually a resize and not a cast. 2334 if (is_cast && elem_bt_from == elem_bt_to) { 2335 is_cast = false; 2336 } 2337 2338 const TypeInstPtr* vbox_type_from = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass_from); 2339 2340 Node* opd1 = unbox_vector(argument(7), vbox_type_from, elem_bt_from, num_elem_from); 2341 if (opd1 == nullptr) { 2342 return false; 2343 } 2344 2345 const TypeVect* src_type = TypeVect::make(elem_bt_from, num_elem_from, is_mask); 2346 const TypeVect* dst_type = TypeVect::make(elem_bt_to, num_elem_to, is_mask); 2347 2348 // Safety check to prevent casting if source mask is of type vector 2349 // and destination mask of type predicate vector and vice-versa. 2350 // From X86 standpoint, this case will only arise over KNL target, 2351 // where certain masks (depending on the species) are either propagated 2352 // through a vector or predicate register. 2353 if (is_mask && 2354 ((src_type->isa_vectmask() == nullptr && dst_type->isa_vectmask()) || 2355 (dst_type->isa_vectmask() == nullptr && src_type->isa_vectmask()))) { 2356 return false; 2357 } 2358 2359 Node* op = opd1; 2360 if (is_cast) { 2361 assert(!is_mask || num_elem_from == num_elem_to, "vector mask cast needs the same elem num"); 2362 int cast_vopc = VectorCastNode::opcode(-1, elem_bt_from, !is_ucast); 2363 2364 // Make sure that vector cast is implemented to particular type/size combination if it is 2365 // not a mask casting. 2366 if (!is_mask && !arch_supports_vector(cast_vopc, num_elem_to, elem_bt_to, VecMaskNotUsed)) { 2367 log_if_needed(" ** not supported: arity=1 op=cast#%d/3 vlen2=%d etype2=%s ismask=%d", 2368 cast_vopc, num_elem_to, type2name(elem_bt_to), is_mask); 2369 return false; 2370 } 2371 2372 if (num_elem_from < num_elem_to) { 2373 // Since input and output number of elements are not consistent, we need to make sure we 2374 // properly size. Thus, first make a cast that retains the number of elements from source. 2375 int num_elem_for_cast = num_elem_from; 2376 2377 // It is possible that arch does not support this intermediate vector size 2378 // TODO More complex logic required here to handle this corner case for the sizes. 2379 if (!arch_supports_vector(cast_vopc, num_elem_for_cast, elem_bt_to, VecMaskNotUsed)) { 2380 log_if_needed(" ** not supported: arity=1 op=cast#%d/4 vlen1=%d etype2=%s ismask=%d", 2381 cast_vopc, 2382 num_elem_for_cast, type2name(elem_bt_to), is_mask); 2383 return false; 2384 } 2385 2386 op = gvn().transform(VectorCastNode::make(cast_vopc, op, elem_bt_to, num_elem_for_cast)); 2387 // Now ensure that the destination gets properly resized to needed size. 2388 op = gvn().transform(new VectorReinterpretNode(op, op->bottom_type()->is_vect(), dst_type)); 2389 } else if (num_elem_from > num_elem_to) { 2390 // Since number of elements from input is larger than output, simply reduce size of input 2391 // (we are supposed to drop top elements anyway). 2392 int num_elem_for_resize = num_elem_to; 2393 2394 // It is possible that arch does not support this intermediate vector size 2395 // TODO More complex logic required here to handle this corner case for the sizes. 2396 if (!arch_supports_vector(Op_VectorReinterpret, 2397 num_elem_for_resize, 2398 elem_bt_from, 2399 VecMaskNotUsed)) { 2400 log_if_needed(" ** not supported: arity=1 op=cast/5 vlen2=%d etype1=%s ismask=%d", 2401 num_elem_for_resize, type2name(elem_bt_from), is_mask); 2402 return false; 2403 } 2404 2405 const TypeVect* resize_type = TypeVect::make(elem_bt_from, num_elem_for_resize); 2406 op = gvn().transform(new VectorReinterpretNode(op, src_type, resize_type)); 2407 op = gvn().transform(VectorCastNode::make(cast_vopc, op, elem_bt_to, num_elem_to)); 2408 } else { // num_elem_from == num_elem_to 2409 if (is_mask) { 2410 // Make sure that cast for vector mask is implemented to particular type/size combination. 2411 if (!arch_supports_vector(Op_VectorMaskCast, num_elem_to, elem_bt_to, VecMaskNotUsed)) { 2412 log_if_needed(" ** not supported: arity=1 op=maskcast vlen2=%d etype2=%s ismask=%d", 2413 num_elem_to, type2name(elem_bt_to), is_mask); 2414 return false; 2415 } 2416 op = gvn().transform(new VectorMaskCastNode(op, dst_type)); 2417 } else { 2418 // Since input and output number of elements match, and since we know this vector size is 2419 // supported, simply do a cast with no resize needed. 2420 op = gvn().transform(VectorCastNode::make(cast_vopc, op, elem_bt_to, num_elem_to)); 2421 } 2422 } 2423 } else if (!Type::equals(src_type, dst_type)) { 2424 assert(!is_cast, "must be reinterpret"); 2425 op = gvn().transform(new VectorReinterpretNode(op, src_type, dst_type)); 2426 } 2427 2428 const TypeInstPtr* vbox_type_to = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass_to); 2429 Node* vbox = box_vector(op, vbox_type_to, elem_bt_to, num_elem_to); 2430 set_result(vbox); 2431 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem_to * type2aelembytes(elem_bt_to)))); 2432 return true; 2433 } 2434 2435 // public static 2436 // <V extends Vector<E>, 2437 // E> 2438 // V insert(Class<? extends V> vectorClass, Class<E> elementType, int vlen, 2439 // V vec, int ix, long val, 2440 // VecInsertOp<V> defaultImpl) 2441 bool LibraryCallKit::inline_vector_insert() { 2442 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 2443 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2444 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2445 const TypeInt* idx = gvn().type(argument(4))->isa_int(); 2446 2447 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || idx == nullptr) { 2448 return false; // dead code 2449 } 2450 if (vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con() || !idx->is_con()) { 2451 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s idx=%s", 2452 NodeClassNames[argument(0)->Opcode()], 2453 NodeClassNames[argument(1)->Opcode()], 2454 NodeClassNames[argument(2)->Opcode()], 2455 NodeClassNames[argument(4)->Opcode()]); 2456 return false; // not enough info for intrinsification 2457 } 2458 if (!is_klass_initialized(vector_klass)) { 2459 log_if_needed(" ** klass argument not initialized"); 2460 return false; 2461 } 2462 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2463 if (!elem_type->is_primitive_type()) { 2464 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2465 return false; // should be primitive type 2466 } 2467 BasicType elem_bt = elem_type->basic_type(); 2468 int num_elem = vlen->get_con(); 2469 if (!arch_supports_vector(Op_VectorInsert, num_elem, elem_bt, VecMaskNotUsed)) { 2470 log_if_needed(" ** not supported: arity=1 op=insert vlen=%d etype=%s ismask=no", 2471 num_elem, type2name(elem_bt)); 2472 return false; // not supported 2473 } 2474 2475 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2476 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2477 2478 Node* opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2479 if (opd == nullptr) { 2480 return false; 2481 } 2482 2483 Node* insert_val = argument(5); 2484 assert(gvn().type(insert_val)->isa_long() != nullptr, "expected to be long"); 2485 2486 // Convert insert value back to its appropriate type. 2487 switch (elem_bt) { 2488 case T_BYTE: 2489 insert_val = gvn().transform(new ConvL2INode(insert_val, TypeInt::BYTE)); 2490 break; 2491 case T_SHORT: 2492 insert_val = gvn().transform(new ConvL2INode(insert_val, TypeInt::SHORT)); 2493 break; 2494 case T_INT: 2495 insert_val = gvn().transform(new ConvL2INode(insert_val)); 2496 break; 2497 case T_FLOAT: 2498 insert_val = gvn().transform(new ConvL2INode(insert_val)); 2499 insert_val = gvn().transform(new MoveI2FNode(insert_val)); 2500 break; 2501 case T_DOUBLE: 2502 insert_val = gvn().transform(new MoveL2DNode(insert_val)); 2503 break; 2504 case T_LONG: 2505 // no conversion needed 2506 break; 2507 default: fatal("%s", type2name(elem_bt)); break; 2508 } 2509 2510 Node* operation = gvn().transform(VectorInsertNode::make(opd, insert_val, idx->get_con(), gvn())); 2511 2512 Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem); 2513 set_result(vbox); 2514 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2515 return true; 2516 } 2517 2518 // public static 2519 // <VM extends VectorPayload, 2520 // E> 2521 // long extract(Class<? extends VM> vClass, Class<E> eClass, 2522 // int length, 2523 // VM vm, int i, 2524 // VecExtractOp<VM> defaultImpl) 2525 bool LibraryCallKit::inline_vector_extract() { 2526 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 2527 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2528 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2529 const TypeInt* idx = gvn().type(argument(4))->isa_int(); 2530 2531 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || idx == nullptr) { 2532 return false; // dead code 2533 } 2534 if (vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con() || !idx->is_con()) { 2535 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s", 2536 NodeClassNames[argument(0)->Opcode()], 2537 NodeClassNames[argument(1)->Opcode()], 2538 NodeClassNames[argument(2)->Opcode()]); 2539 return false; // not enough info for intrinsification 2540 } 2541 if (!is_klass_initialized(vector_klass)) { 2542 log_if_needed(" ** klass argument not initialized"); 2543 return false; 2544 } 2545 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2546 if (!elem_type->is_primitive_type()) { 2547 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2548 return false; // should be primitive type 2549 } 2550 BasicType elem_bt = elem_type->basic_type(); 2551 int num_elem = vlen->get_con(); 2552 2553 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2554 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2555 2556 Node* opd = nullptr; 2557 2558 if (is_vector_mask(vbox_klass)) { 2559 // vbox_klass is mask. This is used for VectorMask.laneIsSet(int). 2560 2561 Node* pos = argument(4); // can be variable 2562 if (arch_supports_vector(Op_ExtractUB, num_elem, elem_bt, VecMaskUseAll)) { 2563 // Transform mask to vector with type of boolean and utilize ExtractUB node. 2564 opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2565 if (opd == nullptr) { 2566 return false; 2567 } 2568 opd = gvn().transform(VectorStoreMaskNode::make(gvn(), opd, elem_bt, num_elem)); 2569 opd = gvn().transform(new ExtractUBNode(opd, pos)); 2570 opd = gvn().transform(new ConvI2LNode(opd)); 2571 } else if (arch_supports_vector(Op_VectorMaskToLong, num_elem, elem_bt, VecMaskUseLoad)) { 2572 opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2573 if (opd == nullptr) { 2574 return false; 2575 } 2576 // VectorMaskToLongNode requires the input is either a mask or a vector with BOOLEAN type. 2577 if (opd->bottom_type()->isa_vectmask() == nullptr) { 2578 opd = gvn().transform(VectorStoreMaskNode::make(gvn(), opd, elem_bt, num_elem)); 2579 } 2580 // ((toLong() >>> pos) & 1L 2581 opd = gvn().transform(new VectorMaskToLongNode(opd, TypeLong::LONG)); 2582 opd = gvn().transform(new URShiftLNode(opd, pos)); 2583 opd = gvn().transform(new AndLNode(opd, gvn().makecon(TypeLong::ONE))); 2584 } else { 2585 log_if_needed(" ** Rejected mask extraction because architecture does not support it"); 2586 return false; // not supported 2587 } 2588 } else { 2589 // vbox_klass is vector. This is used for Vector.lane(int). 2590 if (!idx->is_con()) { 2591 log_if_needed(" ** missing constant: idx=%s", NodeClassNames[argument(4)->Opcode()]); 2592 return false; // not enough info for intrinsification 2593 } 2594 2595 int vopc = ExtractNode::opcode(elem_bt); 2596 if (!arch_supports_vector(vopc, num_elem, elem_bt, VecMaskNotUsed)) { 2597 log_if_needed(" ** not supported: arity=1 op=extract vlen=%d etype=%s ismask=no", 2598 num_elem, type2name(elem_bt)); 2599 return false; // not supported 2600 } 2601 2602 opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2603 if (opd == nullptr) { 2604 return false; 2605 } 2606 ConINode* idx_con = gvn().intcon(idx->get_con())->as_ConI(); 2607 2608 opd = gvn().transform(ExtractNode::make(opd, idx_con, elem_bt)); 2609 switch (elem_bt) { 2610 case T_BYTE: 2611 case T_SHORT: 2612 case T_INT: { 2613 opd = gvn().transform(new ConvI2LNode(opd)); 2614 break; 2615 } 2616 case T_FLOAT: { 2617 opd = gvn().transform(new MoveF2INode(opd)); 2618 opd = gvn().transform(new ConvI2LNode(opd)); 2619 break; 2620 } 2621 case T_DOUBLE: { 2622 opd = gvn().transform(new MoveD2LNode(opd)); 2623 break; 2624 } 2625 case T_LONG: { 2626 // no conversion needed 2627 break; 2628 } 2629 default: fatal("%s", type2name(elem_bt)); 2630 } 2631 } 2632 set_result(opd); 2633 return true; 2634 } 2635 2636 static Node* LowerSelectFromTwoVectorOperation(PhaseGVN& phase, Node* index_vec, Node* src1, Node* src2, const TypeVect* vt) { 2637 int num_elem = vt->length(); 2638 BasicType elem_bt = vt->element_basic_type(); 2639 2640 // Lower selectFrom operation into its constituent operations. 2641 // SelectFromTwoVectorNode = 2642 // (VectorBlend 2643 // (VectorRearrange SRC1 (WRAPED_INDEX AND (VLEN-1)) 2644 // (VectorRearrange SRC2 (WRAPED_INDEX AND (VLEN-1)) 2645 // MASK) 2646 // Where 2647 // WRAPED_INDEX are computed by wrapping incoming indexes 2648 // to two vector index range [0, VLEN*2) and 2649 // MASK = WRAPED_INDEX < VLEN 2650 // 2651 // IR lowering prevents intrinsification failure and associated argument 2652 // boxing penalties. 2653 // 2654 2655 BasicType shuffle_bt = elem_bt; 2656 if (shuffle_bt == T_FLOAT) { 2657 shuffle_bt = T_INT; 2658 } else if (shuffle_bt == T_DOUBLE) { 2659 shuffle_bt = T_LONG; 2660 } 2661 const TypeVect* st = TypeVect::make(shuffle_bt, num_elem); 2662 2663 // Cast index vector to the corresponding bit type 2664 if (elem_bt != shuffle_bt) { 2665 int cast_vopc = VectorCastNode::opcode(0, elem_bt, true); 2666 index_vec = phase.transform(VectorCastNode::make(cast_vopc, index_vec, shuffle_bt, num_elem)); 2667 } 2668 2669 // Wrap indexes into two vector index range [0, VLEN * 2) 2670 Node* two_vect_lane_cnt_m1 = phase.makecon(TypeInteger::make(2 * num_elem - 1, 2 * num_elem - 1, Type::WidenMin, shuffle_bt == T_LONG ? T_LONG : T_INT)); 2671 Node* bcast_two_vect_lane_cnt_m1_vec = phase.transform(VectorNode::scalar2vector(two_vect_lane_cnt_m1, num_elem, 2672 shuffle_bt, false)); 2673 index_vec = phase.transform(VectorNode::make(Op_AndV, index_vec, bcast_two_vect_lane_cnt_m1_vec, st)); 2674 2675 // Compute the blend mask for merging two independently permitted vectors 2676 // using shuffle index in two vector index range [0, VLEN * 2). 2677 BoolTest::mask pred = BoolTest::le; 2678 ConINode* pred_node = phase.makecon(TypeInt::make(pred))->as_ConI(); 2679 const TypeVect* vmask_type = TypeVect::makemask(shuffle_bt, num_elem); 2680 Node* lane_cnt_m1 = phase.makecon(TypeInteger::make(num_elem - 1, num_elem - 1, Type::WidenMin, shuffle_bt == T_LONG ? T_LONG : T_INT)); 2681 Node* bcast_lane_cnt_m1_vec = phase.transform(VectorNode::scalar2vector(lane_cnt_m1, num_elem, shuffle_bt, false)); 2682 Node* mask = phase.transform(new VectorMaskCmpNode(pred, index_vec, bcast_lane_cnt_m1_vec, pred_node, vmask_type)); 2683 2684 // Rearrange expects the indexes to lie within single vector index range [0, VLEN). 2685 Node* wrapped_index_vec = phase.transform(VectorNode::make(Op_AndV, index_vec, bcast_lane_cnt_m1_vec, st)); 2686 2687 // Load indexes from byte vector and appropriately transform them to target 2688 // specific permutation index format. 2689 if (Matcher::vector_rearrange_requires_load_shuffle(shuffle_bt, num_elem)) { 2690 wrapped_index_vec = phase.transform(new VectorLoadShuffleNode(wrapped_index_vec, st)); 2691 } 2692 2693 vmask_type = TypeVect::makemask(elem_bt, num_elem); 2694 mask = phase.transform(new VectorMaskCastNode(mask, vmask_type)); 2695 2696 Node* p1 = phase.transform(new VectorRearrangeNode(src1, wrapped_index_vec)); 2697 Node* p2 = phase.transform(new VectorRearrangeNode(src2, wrapped_index_vec)); 2698 2699 return new VectorBlendNode(p2, p1, mask); 2700 } 2701 2702 // public static 2703 // <V extends Vector<E>, 2704 // E> 2705 // V selectFromTwoVectorOp(Class<? extends V> vClass, Class<E> eClass, int length, 2706 // V v1, V v2, V v3, 2707 // SelectFromTwoVector<V> defaultImpl) 2708 bool LibraryCallKit::inline_vector_select_from_two_vectors() { 2709 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 2710 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2711 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2712 2713 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || vector_klass->const_oop() == nullptr || 2714 elem_klass->const_oop() == nullptr ||!vlen->is_con()) { 2715 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s", 2716 NodeClassNames[argument(0)->Opcode()], 2717 NodeClassNames[argument(1)->Opcode()], 2718 NodeClassNames[argument(2)->Opcode()]); 2719 return false; // not enough info for intrinsification 2720 } 2721 2722 if (!is_klass_initialized(vector_klass)) { 2723 log_if_needed(" ** klass argument not initialized"); 2724 return false; 2725 } 2726 2727 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2728 if (!elem_type->is_primitive_type()) { 2729 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2730 return false; // should be primitive type 2731 } 2732 2733 int num_elem = vlen->get_con(); 2734 if (!is_power_of_2(num_elem)) { 2735 log_if_needed(" ** vlen is not power of two=%d", num_elem); 2736 return false; 2737 } 2738 2739 BasicType elem_bt = elem_type->basic_type(); 2740 BasicType index_elem_bt = elem_bt; 2741 if (elem_bt == T_FLOAT) { 2742 index_elem_bt = T_INT; 2743 } else if (elem_bt == T_DOUBLE) { 2744 index_elem_bt = T_LONG; 2745 } 2746 2747 bool lowerSelectFromOp = false; 2748 if (!arch_supports_vector(Op_SelectFromTwoVector, num_elem, elem_bt, VecMaskNotUsed)) { 2749 int cast_vopc = VectorCastNode::opcode(-1, elem_bt, true); 2750 if ((elem_bt != index_elem_bt && !arch_supports_vector(cast_vopc, num_elem, index_elem_bt, VecMaskNotUsed)) || 2751 !arch_supports_vector(Op_VectorMaskCmp, num_elem, index_elem_bt, VecMaskNotUsed) || 2752 !arch_supports_vector(Op_AndV, num_elem, index_elem_bt, VecMaskNotUsed) || 2753 !arch_supports_vector(Op_VectorMaskCast, num_elem, elem_bt, VecMaskNotUsed) || 2754 !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad) || 2755 !arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, VecMaskNotUsed) || 2756 !arch_supports_vector(Op_VectorLoadShuffle, num_elem, index_elem_bt, VecMaskNotUsed) || 2757 !arch_supports_vector(Op_Replicate, num_elem, index_elem_bt, VecMaskNotUsed)) { 2758 log_if_needed(" ** not supported: opc=%d vlen=%d etype=%s ismask=useload", 2759 Op_SelectFromTwoVector, num_elem, type2name(elem_bt)); 2760 return false; // not supported 2761 } 2762 lowerSelectFromOp = true; 2763 } 2764 2765 int cast_vopc = VectorCastNode::opcode(-1, elem_bt, true); 2766 if (!lowerSelectFromOp) { 2767 if (!arch_supports_vector(Op_AndV, num_elem, index_elem_bt, VecMaskNotUsed) || 2768 !arch_supports_vector(Op_Replicate, num_elem, index_elem_bt, VecMaskNotUsed) || 2769 (is_floating_point_type(elem_bt) && 2770 !arch_supports_vector(cast_vopc, num_elem, index_elem_bt, VecMaskNotUsed))) { 2771 log_if_needed(" ** index wrapping not supported: vlen=%d etype=%s" , 2772 num_elem, type2name(elem_bt)); 2773 return false; // not supported 2774 } 2775 } 2776 2777 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2778 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2779 2780 Node* opd1 = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2781 if (opd1 == nullptr) { 2782 log_if_needed(" ** unbox failed v1=%s", 2783 NodeClassNames[argument(3)->Opcode()]); 2784 return false; 2785 } 2786 Node* opd2 = unbox_vector(argument(4), vbox_type, elem_bt, num_elem); 2787 if (opd2 == nullptr) { 2788 log_if_needed(" ** unbox failed v2=%s", 2789 NodeClassNames[argument(4)->Opcode()]); 2790 return false; 2791 } 2792 Node* opd3 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 2793 if (opd3 == nullptr) { 2794 log_if_needed(" ** unbox failed v3=%s", 2795 NodeClassNames[argument(5)->Opcode()]); 2796 return false; 2797 } 2798 2799 const TypeVect* vt = TypeVect::make(elem_bt, num_elem); 2800 2801 Node* operation = nullptr; 2802 if (lowerSelectFromOp) { 2803 operation = gvn().transform(LowerSelectFromTwoVectorOperation(gvn(), opd1, opd2, opd3, vt)); 2804 } else { 2805 if (index_elem_bt != elem_bt) { 2806 opd1 = gvn().transform(VectorCastNode::make(cast_vopc, opd1, index_elem_bt, num_elem)); 2807 } 2808 int indexRangeMask = 2 * num_elem - 1; 2809 Node* wrap_mask = gvn().makecon(TypeInteger::make(indexRangeMask, indexRangeMask, Type::WidenMin, index_elem_bt != T_LONG ? T_INT : index_elem_bt)); 2810 Node* wrap_mask_vec = gvn().transform(VectorNode::scalar2vector(wrap_mask, num_elem, index_elem_bt, false)); 2811 opd1 = gvn().transform(VectorNode::make(Op_AndV, opd1, wrap_mask_vec, opd1->bottom_type()->is_vect())); 2812 operation = gvn().transform(VectorNode::make(Op_SelectFromTwoVector, opd1, opd2, opd3, vt)); 2813 } 2814 2815 // Wrap it up in VectorBox to keep object type information. 2816 Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem); 2817 set_result(vbox); 2818 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2819 return true; 2820 } 2821 2822 // public static 2823 // <V extends Vector<E>, 2824 // M extends VectorMask<E>, 2825 // E> 2826 // V compressExpandOp(int opr, 2827 // Class<? extends V> vClass, Class<? extends M> mClass, Class<E> eClass, 2828 // int length, V v, M m, 2829 // CompressExpandOperation<V, M> defaultImpl) 2830 bool LibraryCallKit::inline_vector_compress_expand() { 2831 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 2832 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 2833 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 2834 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 2835 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 2836 2837 if (opr == nullptr || !opr->is_con() || 2838 vector_klass == nullptr || vector_klass->const_oop() == nullptr || 2839 mask_klass == nullptr || mask_klass->const_oop() == nullptr || 2840 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 2841 vlen == nullptr || !vlen->is_con()) { 2842 log_if_needed(" ** missing constant: opr=%s vclass=%s mclass=%s etype=%s vlen=%s", 2843 NodeClassNames[argument(0)->Opcode()], 2844 NodeClassNames[argument(1)->Opcode()], 2845 NodeClassNames[argument(2)->Opcode()], 2846 NodeClassNames[argument(3)->Opcode()], 2847 NodeClassNames[argument(4)->Opcode()]); 2848 return false; // not enough info for intrinsification 2849 } 2850 2851 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(mask_klass)) { 2852 log_if_needed(" ** klass argument not initialized"); 2853 return false; 2854 } 2855 2856 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2857 if (!elem_type->is_primitive_type()) { 2858 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2859 return false; // should be primitive type 2860 } 2861 2862 int num_elem = vlen->get_con(); 2863 BasicType elem_bt = elem_type->basic_type(); 2864 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 2865 2866 if (!arch_supports_vector(opc, num_elem, elem_bt, VecMaskUseLoad)) { 2867 log_if_needed(" ** not supported: opc=%d vlen=%d etype=%s ismask=useload", 2868 opc, num_elem, type2name(elem_bt)); 2869 return false; // not supported 2870 } 2871 2872 Node* opd1 = nullptr; 2873 const TypeInstPtr* vbox_type = nullptr; 2874 if (opc != Op_CompressM) { 2875 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2876 vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2877 opd1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 2878 if (opd1 == nullptr) { 2879 log_if_needed(" ** unbox failed vector=%s", 2880 NodeClassNames[argument(5)->Opcode()]); 2881 return false; 2882 } 2883 } 2884 2885 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2886 assert(is_vector_mask(mbox_klass), "argument(6) should be a mask class"); 2887 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 2888 2889 Node* mask = unbox_vector(argument(6), mbox_type, elem_bt, num_elem); 2890 if (mask == nullptr) { 2891 log_if_needed(" ** unbox failed mask=%s", 2892 NodeClassNames[argument(6)->Opcode()]); 2893 return false; 2894 } 2895 2896 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, opc == Op_CompressM); 2897 Node* operation = gvn().transform(VectorNode::make(opc, opd1, mask, vt)); 2898 2899 // Wrap it up in VectorBox to keep object type information. 2900 const TypeInstPtr* box_type = opc == Op_CompressM ? mbox_type : vbox_type; 2901 Node* vbox = box_vector(operation, box_type, elem_bt, num_elem); 2902 set_result(vbox); 2903 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2904 return true; 2905 } 2906 2907 // public static 2908 // <V extends Vector<E>, 2909 // E, 2910 // S extends VectorSpecies<E>> 2911 // V indexVector(Class<? extends V> vClass, Class<E> eClass, 2912 // int length, 2913 // V v, int step, S s, 2914 // IndexOperation<V, S> defaultImpl) 2915 bool LibraryCallKit::inline_index_vector() { 2916 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 2917 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2918 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2919 2920 if (vector_klass == nullptr || vector_klass->const_oop() == nullptr || 2921 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 2922 vlen == nullptr || !vlen->is_con() ) { 2923 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s", 2924 NodeClassNames[argument(0)->Opcode()], 2925 NodeClassNames[argument(1)->Opcode()], 2926 NodeClassNames[argument(2)->Opcode()]); 2927 return false; // not enough info for intrinsification 2928 } 2929 2930 if (!is_klass_initialized(vector_klass)) { 2931 log_if_needed(" ** klass argument not initialized"); 2932 return false; 2933 } 2934 2935 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2936 if (!elem_type->is_primitive_type()) { 2937 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2938 return false; // should be primitive type 2939 } 2940 2941 int num_elem = vlen->get_con(); 2942 BasicType elem_bt = elem_type->basic_type(); 2943 2944 // Check whether the iota index generation op is supported by the current hardware 2945 if (!arch_supports_vector(Op_VectorLoadConst, num_elem, elem_bt, VecMaskNotUsed)) { 2946 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 2947 return false; // not supported 2948 } 2949 2950 int mul_op = VectorSupport::vop2ideal(VectorSupport::VECTOR_OP_MUL, elem_bt); 2951 int vmul_op = VectorNode::opcode(mul_op, elem_bt); 2952 bool needs_mul = true; 2953 Node* scale = argument(4); 2954 const TypeInt* scale_type = gvn().type(scale)->isa_int(); 2955 // Multiply is not needed if the scale is a constant "1". 2956 if (scale_type && scale_type->is_con() && scale_type->get_con() == 1) { 2957 needs_mul = false; 2958 } else { 2959 // Check whether the vector multiply op is supported by the current hardware 2960 if (!arch_supports_vector(vmul_op, num_elem, elem_bt, VecMaskNotUsed)) { 2961 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 2962 return false; // not supported 2963 } 2964 2965 // Check whether the scalar cast op is supported by the current hardware 2966 if (is_floating_point_type(elem_bt) || elem_bt == T_LONG) { 2967 int cast_op = elem_bt == T_LONG ? Op_ConvI2L : 2968 elem_bt == T_FLOAT? Op_ConvI2F : Op_ConvI2D; 2969 if (!Matcher::match_rule_supported(cast_op)) { 2970 log_if_needed(" ** Rejected op (%s) because architecture does not support it", 2971 NodeClassNames[cast_op]); 2972 return false; // not supported 2973 } 2974 } 2975 } 2976 2977 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2978 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2979 Node* opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2980 if (opd == nullptr) { 2981 log_if_needed(" ** unbox failed vector=%s", 2982 NodeClassNames[argument(3)->Opcode()]); 2983 return false; 2984 } 2985 2986 int add_op = VectorSupport::vop2ideal(VectorSupport::VECTOR_OP_ADD, elem_bt); 2987 int vadd_op = VectorNode::opcode(add_op, elem_bt); 2988 bool needs_add = true; 2989 // The addition is not needed if all the element values of "opd" are zero 2990 if (VectorNode::is_all_zeros_vector(opd)) { 2991 needs_add = false; 2992 } else { 2993 // Check whether the vector addition op is supported by the current hardware 2994 if (!arch_supports_vector(vadd_op, num_elem, elem_bt, VecMaskNotUsed)) { 2995 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 2996 return false; // not supported 2997 } 2998 } 2999 3000 // Compute the iota indice vector 3001 const TypeVect* vt = TypeVect::make(elem_bt, num_elem); 3002 Node* index = gvn().transform(new VectorLoadConstNode(gvn().makecon(TypeInt::ZERO), vt)); 3003 3004 // Broadcast the "scale" to a vector, and multiply the "scale" with iota indice vector. 3005 if (needs_mul) { 3006 switch (elem_bt) { 3007 case T_BOOLEAN: // fall-through 3008 case T_BYTE: // fall-through 3009 case T_SHORT: // fall-through 3010 case T_CHAR: // fall-through 3011 case T_INT: { 3012 // no conversion needed 3013 break; 3014 } 3015 case T_LONG: { 3016 scale = gvn().transform(new ConvI2LNode(scale)); 3017 break; 3018 } 3019 case T_FLOAT: { 3020 scale = gvn().transform(new ConvI2FNode(scale)); 3021 break; 3022 } 3023 case T_DOUBLE: { 3024 scale = gvn().transform(new ConvI2DNode(scale)); 3025 break; 3026 } 3027 default: fatal("%s", type2name(elem_bt)); 3028 } 3029 scale = gvn().transform(VectorNode::scalar2vector(scale, num_elem, elem_bt)); 3030 index = gvn().transform(VectorNode::make(vmul_op, index, scale, vt)); 3031 } 3032 3033 // Add "opd" if addition is needed. 3034 if (needs_add) { 3035 index = gvn().transform(VectorNode::make(vadd_op, opd, index, vt)); 3036 } 3037 Node* vbox = box_vector(index, vbox_type, elem_bt, num_elem); 3038 set_result(vbox); 3039 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 3040 return true; 3041 } 3042 3043 // public static 3044 // <E, 3045 // M extends VectorMask<E>> 3046 // M indexPartiallyInUpperRange(Class<? extends M> mClass, Class<E> eClass, int length, 3047 // long offset, long limit, 3048 // IndexPartiallyInUpperRangeOperation<E, M> defaultImpl) 3049 bool LibraryCallKit::inline_index_partially_in_upper_range() { 3050 const TypeInstPtr* mask_klass = gvn().type(argument(0))->isa_instptr(); 3051 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 3052 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 3053 3054 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr || 3055 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 3056 vlen == nullptr || !vlen->is_con()) { 3057 log_if_needed(" ** missing constant: mclass=%s etype=%s vlen=%s", 3058 NodeClassNames[argument(0)->Opcode()], 3059 NodeClassNames[argument(1)->Opcode()], 3060 NodeClassNames[argument(2)->Opcode()]); 3061 return false; // not enough info for intrinsification 3062 } 3063 3064 if (!is_klass_initialized(mask_klass)) { 3065 log_if_needed(" ** klass argument not initialized"); 3066 return false; 3067 } 3068 3069 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 3070 if (!elem_type->is_primitive_type()) { 3071 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 3072 return false; // should be primitive type 3073 } 3074 3075 int num_elem = vlen->get_con(); 3076 BasicType elem_bt = elem_type->basic_type(); 3077 3078 // Check whether the necessary ops are supported by current hardware. 3079 bool supports_mask_gen = arch_supports_vector(Op_VectorMaskGen, num_elem, elem_bt, VecMaskUseStore); 3080 if (!supports_mask_gen) { 3081 if (!arch_supports_vector(Op_VectorLoadConst, num_elem, elem_bt, VecMaskNotUsed) || 3082 !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed) || 3083 !arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskUseStore)) { 3084 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 3085 return false; // not supported 3086 } 3087 3088 // Check whether the scalar cast operation is supported by current hardware. 3089 if (elem_bt != T_LONG) { 3090 int cast_op = is_integral_type(elem_bt) ? Op_ConvL2I 3091 : (elem_bt == T_FLOAT ? Op_ConvL2F : Op_ConvL2D); 3092 if (!Matcher::match_rule_supported(cast_op)) { 3093 log_if_needed(" ** Rejected op (%s) because architecture does not support it", 3094 NodeClassNames[cast_op]); 3095 return false; // not supported 3096 } 3097 } 3098 } 3099 3100 Node* offset = argument(3); 3101 Node* limit = argument(5); 3102 if (offset == nullptr || limit == nullptr) { 3103 log_if_needed(" ** offset or limit argument is null"); 3104 return false; // not supported 3105 } 3106 3107 ciKlass* box_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 3108 assert(is_vector_mask(box_klass), "argument(0) should be a mask class"); 3109 const TypeInstPtr* box_type = TypeInstPtr::make_exact(TypePtr::NotNull, box_klass); 3110 3111 // We assume "offset > 0 && limit >= offset && limit - offset < num_elem". 3112 // So directly get indexLimit with "indexLimit = limit - offset". 3113 Node* indexLimit = gvn().transform(new SubLNode(limit, offset)); 3114 Node* mask = nullptr; 3115 if (supports_mask_gen) { 3116 mask = gvn().transform(VectorMaskGenNode::make(indexLimit, elem_bt, num_elem)); 3117 } else { 3118 // Generate the vector mask based on "mask = iota < indexLimit". 3119 // Broadcast "indexLimit" to a vector. 3120 switch (elem_bt) { 3121 case T_BOOLEAN: // fall-through 3122 case T_BYTE: // fall-through 3123 case T_SHORT: // fall-through 3124 case T_CHAR: // fall-through 3125 case T_INT: { 3126 indexLimit = gvn().transform(new ConvL2INode(indexLimit)); 3127 break; 3128 } 3129 case T_DOUBLE: { 3130 indexLimit = gvn().transform(new ConvL2DNode(indexLimit)); 3131 break; 3132 } 3133 case T_FLOAT: { 3134 indexLimit = gvn().transform(new ConvL2FNode(indexLimit)); 3135 break; 3136 } 3137 case T_LONG: { 3138 // no conversion needed 3139 break; 3140 } 3141 default: fatal("%s", type2name(elem_bt)); 3142 } 3143 indexLimit = gvn().transform(VectorNode::scalar2vector(indexLimit, num_elem, elem_bt)); 3144 3145 // Load the "iota" vector. 3146 const TypeVect* vt = TypeVect::make(elem_bt, num_elem); 3147 Node* iota = gvn().transform(new VectorLoadConstNode(gvn().makecon(TypeInt::ZERO), vt)); 3148 3149 // Compute the vector mask with "mask = iota < indexLimit". 3150 ConINode* pred_node = (ConINode*)gvn().makecon(TypeInt::make(BoolTest::lt)); 3151 const TypeVect* vmask_type = TypeVect::makemask(elem_bt, num_elem); 3152 mask = gvn().transform(new VectorMaskCmpNode(BoolTest::lt, iota, indexLimit, pred_node, vmask_type)); 3153 } 3154 Node* vbox = box_vector(mask, box_type, elem_bt, num_elem); 3155 set_result(vbox); 3156 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 3157 return true; 3158 } 3159 3160 #undef non_product_log_if_needed 3161 #undef log_if_needed