1 /* 2 * Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "ci/ciSymbols.hpp" 26 #include "classfile/vmSymbols.hpp" 27 #include "opto/library_call.hpp" 28 #include "opto/runtime.hpp" 29 #include "opto/vectornode.hpp" 30 #include "prims/vectorSupport.hpp" 31 #include "runtime/stubRoutines.hpp" 32 33 #ifdef ASSERT 34 static bool is_vector(ciKlass* klass) { 35 return klass->is_subclass_of(ciEnv::current()->vector_VectorPayload_klass()); 36 } 37 38 static bool check_vbox(const TypeInstPtr* vbox_type) { 39 assert(vbox_type->klass_is_exact(), ""); 40 41 ciInstanceKlass* ik = vbox_type->instance_klass(); 42 assert(is_vector(ik), "not a vector"); 43 44 ciField* fd1 = ik->get_field_by_name(ciSymbols::ETYPE_name(), ciSymbols::class_signature(), /* is_static */ true); 45 assert(fd1 != nullptr, "element type info is missing"); 46 47 ciConstant val1 = fd1->constant_value(); 48 BasicType elem_bt = val1.as_object()->as_instance()->java_mirror_type()->basic_type(); 49 assert(is_java_primitive(elem_bt), "element type info is missing"); 50 51 ciField* fd2 = ik->get_field_by_name(ciSymbols::VLENGTH_name(), ciSymbols::int_signature(), /* is_static */ true); 52 assert(fd2 != nullptr, "vector length info is missing"); 53 54 ciConstant val2 = fd2->constant_value(); 55 assert(val2.as_int() > 0, "vector length info is missing"); 56 57 return true; 58 } 59 #endif 60 61 #define log_if_needed(...) \ 62 if (C->print_intrinsics()) { \ 63 tty->print_cr(__VA_ARGS__); \ 64 } 65 66 #ifndef PRODUCT 67 #define non_product_log_if_needed(...) log_if_needed(__VA_ARGS__) 68 #else 69 #define non_product_log_if_needed(...) 70 #endif 71 72 static bool is_vector_mask(ciKlass* klass) { 73 return klass->is_subclass_of(ciEnv::current()->vector_VectorMask_klass()); 74 } 75 76 bool LibraryCallKit::arch_supports_vector_rotate(int opc, int num_elem, BasicType elem_bt, 77 VectorMaskUseType mask_use_type, bool has_scalar_args) { 78 bool is_supported = true; 79 80 // has_scalar_args flag is true only for non-constant scalar shift count, 81 // since in this case shift needs to be broadcasted. 82 if (!Matcher::match_rule_supported_vector(opc, num_elem, elem_bt) || 83 (has_scalar_args && !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed))) { 84 is_supported = false; 85 } 86 87 if (is_supported) { 88 // Check if mask unboxing is supported, this is a two step process which first loads the contents 89 // of boolean array into vector followed by either lane expansion to match the lane size of masked 90 // vector operation or populate the predicate register. 91 if ((mask_use_type & VecMaskUseLoad) != 0) { 92 if (!Matcher::match_rule_supported_vector(Op_VectorLoadMask, num_elem, elem_bt) || 93 !Matcher::match_rule_supported_vector(Op_LoadVector, num_elem, T_BOOLEAN)) { 94 non_product_log_if_needed(" ** Rejected vector mask loading (%s,%s,%d) because architecture does not support it", 95 NodeClassNames[Op_VectorLoadMask], type2name(elem_bt), num_elem); 96 return false; 97 } 98 } 99 100 if ((mask_use_type & VecMaskUsePred) != 0) { 101 if (!Matcher::has_predicated_vectors() || 102 !Matcher::match_rule_supported_vector_masked(opc, num_elem, elem_bt)) { 103 non_product_log_if_needed("Rejected vector mask predicate using (%s,%s,%d) because architecture does not support it", 104 NodeClassNames[opc], type2name(elem_bt), num_elem); 105 return false; 106 } 107 } 108 } 109 110 int lshiftopc, rshiftopc; 111 switch(elem_bt) { 112 case T_BYTE: 113 lshiftopc = Op_LShiftI; 114 rshiftopc = Op_URShiftB; 115 break; 116 case T_SHORT: 117 lshiftopc = Op_LShiftI; 118 rshiftopc = Op_URShiftS; 119 break; 120 case T_INT: 121 lshiftopc = Op_LShiftI; 122 rshiftopc = Op_URShiftI; 123 break; 124 case T_LONG: 125 lshiftopc = Op_LShiftL; 126 rshiftopc = Op_URShiftL; 127 break; 128 default: fatal("Unexpected type: %s", type2name(elem_bt)); 129 } 130 int lshiftvopc = VectorNode::opcode(lshiftopc, elem_bt); 131 int rshiftvopc = VectorNode::opcode(rshiftopc, elem_bt); 132 if (!is_supported && 133 arch_supports_vector(lshiftvopc, num_elem, elem_bt, VecMaskNotUsed, has_scalar_args) && 134 arch_supports_vector(rshiftvopc, num_elem, elem_bt, VecMaskNotUsed, has_scalar_args) && 135 arch_supports_vector(Op_OrV, num_elem, elem_bt, VecMaskNotUsed)) { 136 is_supported = true; 137 } 138 return is_supported; 139 } 140 141 Node* GraphKit::box_vector(Node* vector, const TypeInstPtr* vbox_type, BasicType elem_bt, int num_elem, bool deoptimize_on_exception) { 142 assert(EnableVectorSupport, ""); 143 144 PreserveReexecuteState preexecs(this); 145 jvms()->set_should_reexecute(true); 146 147 VectorBoxAllocateNode* alloc = new VectorBoxAllocateNode(C, vbox_type); 148 set_edges_for_java_call(alloc, /*must_throw=*/false, /*separate_io_proj=*/true); 149 make_slow_call_ex(alloc, env()->Throwable_klass(), /*separate_io_proj=*/true, deoptimize_on_exception); 150 set_i_o(gvn().transform( new ProjNode(alloc, TypeFunc::I_O) )); 151 set_all_memory(gvn().transform( new ProjNode(alloc, TypeFunc::Memory) )); 152 Node* ret = gvn().transform(new ProjNode(alloc, TypeFunc::Parms)); 153 154 assert(check_vbox(vbox_type), ""); 155 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, is_vector_mask(vbox_type->instance_klass())); 156 VectorBoxNode* vbox = new VectorBoxNode(C, ret, vector, vbox_type, vt); 157 return gvn().transform(vbox); 158 } 159 160 Node* GraphKit::unbox_vector(Node* v, const TypeInstPtr* vbox_type, BasicType elem_bt, int num_elem) { 161 assert(EnableVectorSupport, ""); 162 const TypeInstPtr* vbox_type_v = gvn().type(v)->isa_instptr(); 163 if (vbox_type_v == nullptr || vbox_type->instance_klass() != vbox_type_v->instance_klass()) { 164 return nullptr; // arguments don't agree on vector shapes 165 } 166 if (vbox_type_v->maybe_null()) { 167 return nullptr; // no nulls are allowed 168 } 169 assert(check_vbox(vbox_type), ""); 170 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, is_vector_mask(vbox_type->instance_klass())); 171 Node* unbox = gvn().transform(new VectorUnboxNode(C, vt, v, merged_memory())); 172 return unbox; 173 } 174 175 Node* GraphKit::vector_shift_count(Node* cnt, int shift_op, BasicType bt, int num_elem) { 176 assert(bt == T_INT || bt == T_LONG || bt == T_SHORT || bt == T_BYTE, "byte, short, long and int are supported"); 177 juint mask = (type2aelembytes(bt) * BitsPerByte - 1); 178 Node* nmask = gvn().transform(ConNode::make(TypeInt::make(mask))); 179 Node* mcnt = gvn().transform(new AndINode(cnt, nmask)); 180 return gvn().transform(VectorNode::shift_count(shift_op, mcnt, num_elem, bt)); 181 } 182 183 bool LibraryCallKit::arch_supports_vector(int sopc, int num_elem, BasicType type, VectorMaskUseType mask_use_type, bool has_scalar_args) { 184 // Check that the operation is valid. 185 if (sopc <= 0) { 186 non_product_log_if_needed(" ** Rejected intrinsification because no valid vector op could be extracted"); 187 return false; 188 } 189 190 if (VectorNode::is_vector_rotate(sopc)) { 191 if(!arch_supports_vector_rotate(sopc, num_elem, type, mask_use_type, has_scalar_args)) { 192 non_product_log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support variable vector shifts", 193 NodeClassNames[sopc], type2name(type), num_elem); 194 return false; 195 } 196 } else if (VectorNode::is_vector_integral_negate(sopc)) { 197 if (!VectorNode::is_vector_integral_negate_supported(sopc, num_elem, type, false)) { 198 non_product_log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support integral vector negate", 199 NodeClassNames[sopc], type2name(type), num_elem); 200 return false; 201 } 202 } else { 203 // Check that architecture supports this op-size-type combination. 204 if (!Matcher::match_rule_supported_vector(sopc, num_elem, type)) { 205 non_product_log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support it", 206 NodeClassNames[sopc], type2name(type), num_elem); 207 return false; 208 } else { 209 assert(Matcher::match_rule_supported(sopc), "must be supported"); 210 } 211 } 212 213 if (num_elem == 1) { 214 if (mask_use_type != VecMaskNotUsed) { 215 non_product_log_if_needed(" ** Rejected vector mask op (%s,%s,%d) because architecture does not support it", 216 NodeClassNames[sopc], type2name(type), num_elem); 217 return false; 218 } 219 220 if (sopc != 0) { 221 if (sopc != Op_LoadVector && sopc != Op_StoreVector) { 222 non_product_log_if_needed(" ** Not a svml call or load/store vector op (%s,%s,%d)", 223 NodeClassNames[sopc], type2name(type), num_elem); 224 return false; 225 } 226 } 227 } 228 229 if (!has_scalar_args && VectorNode::is_vector_shift(sopc) && 230 Matcher::supports_vector_variable_shifts() == false) { 231 log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support variable vector shifts", 232 NodeClassNames[sopc], type2name(type), num_elem); 233 return false; 234 } 235 236 // Check if mask unboxing is supported, this is a two step process which first loads the contents 237 // of boolean array into vector followed by either lane expansion to match the lane size of masked 238 // vector operation or populate the predicate register. 239 if ((mask_use_type & VecMaskUseLoad) != 0) { 240 if (!Matcher::match_rule_supported_vector(Op_VectorLoadMask, num_elem, type) || 241 !Matcher::match_rule_supported_vector(Op_LoadVector, num_elem, T_BOOLEAN)) { 242 non_product_log_if_needed(" ** Rejected vector mask loading (%s,%s,%d) because architecture does not support it", 243 NodeClassNames[Op_VectorLoadMask], type2name(type), num_elem); 244 return false; 245 } 246 } 247 248 // Check if mask boxing is supported, this is a two step process which first stores the contents 249 // of mask vector / predicate register into a boolean vector followed by vector store operation to 250 // transfer the contents to underlined storage of mask boxes which is a boolean array. 251 if ((mask_use_type & VecMaskUseStore) != 0) { 252 if (!Matcher::match_rule_supported_vector(Op_VectorStoreMask, num_elem, type) || 253 !Matcher::match_rule_supported_vector(Op_StoreVector, num_elem, T_BOOLEAN)) { 254 non_product_log_if_needed("Rejected vector mask storing (%s,%s,%d) because architecture does not support it", 255 NodeClassNames[Op_VectorStoreMask], type2name(type), num_elem); 256 return false; 257 } 258 } 259 260 if ((mask_use_type & VecMaskUsePred) != 0) { 261 bool is_supported = false; 262 if (Matcher::has_predicated_vectors()) { 263 if (VectorNode::is_vector_integral_negate(sopc)) { 264 is_supported = VectorNode::is_vector_integral_negate_supported(sopc, num_elem, type, true); 265 } else { 266 is_supported = Matcher::match_rule_supported_vector_masked(sopc, num_elem, type); 267 } 268 } 269 is_supported |= Matcher::supports_vector_predicate_op_emulation(sopc, num_elem, type); 270 271 if (!is_supported) { 272 non_product_log_if_needed("Rejected vector mask predicate using (%s,%s,%d) because architecture does not support it", 273 NodeClassNames[sopc], type2name(type), num_elem); 274 return false; 275 } 276 } 277 278 return true; 279 } 280 281 static bool is_klass_initialized(const TypeInstPtr* vec_klass) { 282 if (vec_klass->const_oop() == nullptr) { 283 return false; // uninitialized or some kind of unsafe access 284 } 285 assert(vec_klass->const_oop()->as_instance()->java_lang_Class_klass() != nullptr, "klass instance expected"); 286 ciInstanceKlass* klass = vec_klass->const_oop()->as_instance()->java_lang_Class_klass()->as_instance_klass(); 287 return klass->is_initialized(); 288 } 289 290 // public static 291 // <V extends Vector<E>, 292 // M extends VectorMask<E>, 293 // E> 294 // V unaryOp(int oprId, Class<? extends V> vmClass, Class<? extends M> maskClass, Class<E> elementType, 295 // int length, V v, M m, 296 // UnaryOperation<V, M> defaultImpl) 297 // 298 // public static 299 // <V, 300 // M extends VectorMask<E>, 301 // E> 302 // V binaryOp(int oprId, Class<? extends V> vmClass, Class<? extends M> maskClass, Class<E> elementType, 303 // int length, V v1, V v2, M m, 304 // BinaryOperation<V, M> defaultImpl) 305 // 306 // public static 307 // <V extends Vector<E>, 308 // M extends VectorMask<E>, 309 // E> 310 // V ternaryOp(int oprId, Class<? extends V> vmClass, Class<? extends M> maskClass, Class<E> elementType, 311 // int length, V v1, V v2, V v3, M m, 312 // TernaryOperation<V, M> defaultImpl) 313 // 314 bool LibraryCallKit::inline_vector_nary_operation(int n) { 315 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 316 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 317 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 318 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 319 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 320 321 if (opr == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 322 !opr->is_con() || vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 323 log_if_needed(" ** missing constant: opr=%s vclass=%s etype=%s vlen=%s", 324 NodeClassNames[argument(0)->Opcode()], 325 NodeClassNames[argument(1)->Opcode()], 326 NodeClassNames[argument(3)->Opcode()], 327 NodeClassNames[argument(4)->Opcode()]); 328 return false; // not enough info for intrinsification 329 } 330 331 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 332 if (!elem_type->is_primitive_type()) { 333 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 334 return false; // should be primitive type 335 } 336 if (!is_klass_initialized(vector_klass)) { 337 log_if_needed(" ** klass argument not initialized"); 338 return false; 339 } 340 341 // "argument(n + 5)" should be the mask object. We assume it is "null" when no mask 342 // is used to control this operation. 343 const Type* vmask_type = gvn().type(argument(n + 5)); 344 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 345 if (is_masked_op) { 346 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 347 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(2)->Opcode()]); 348 return false; // not enough info for intrinsification 349 } 350 351 if (!is_klass_initialized(mask_klass)) { 352 log_if_needed(" ** mask klass argument not initialized"); 353 return false; 354 } 355 356 if (vmask_type->maybe_null()) { 357 log_if_needed(" ** null mask values are not allowed for masked op"); 358 return false; 359 } 360 } 361 362 BasicType elem_bt = elem_type->basic_type(); 363 bool has_scalar_op = VectorSupport::has_scalar_op(opr->get_con()); 364 bool is_unsigned = VectorSupport::is_unsigned_op(opr->get_con()); 365 366 int num_elem = vlen->get_con(); 367 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 368 int sopc = has_scalar_op ? VectorNode::opcode(opc, elem_bt) : opc; 369 if ((opc != Op_CallLeafVector) && (sopc == 0)) { 370 log_if_needed(" ** operation not supported: opc=%s bt=%s", NodeClassNames[opc], type2name(elem_bt)); 371 return false; // operation not supported 372 } 373 if (num_elem == 1) { 374 if (opc != Op_CallLeafVector || elem_bt != T_DOUBLE) { 375 log_if_needed(" ** not a svml call: arity=%d opc=%d vlen=%d etype=%s", 376 n, opc, num_elem, type2name(elem_bt)); 377 return false; 378 } 379 } 380 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 381 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 382 383 if (is_vector_mask(vbox_klass)) { 384 assert(!is_masked_op, "mask operations do not need mask to control"); 385 } 386 387 if (opc == Op_CallLeafVector) { 388 if (!UseVectorStubs) { 389 log_if_needed(" ** vector stubs support is disabled"); 390 return false; 391 } 392 if (!Matcher::supports_vector_calling_convention()) { 393 log_if_needed(" ** no vector calling conventions supported"); 394 return false; 395 } 396 if (!Matcher::vector_size_supported(elem_bt, num_elem)) { 397 log_if_needed(" ** vector size (vlen=%d, etype=%s) is not supported", 398 num_elem, type2name(elem_bt)); 399 return false; 400 } 401 } 402 403 // When using mask, mask use type needs to be VecMaskUseLoad. 404 VectorMaskUseType mask_use_type = is_vector_mask(vbox_klass) ? VecMaskUseAll 405 : is_masked_op ? VecMaskUseLoad : VecMaskNotUsed; 406 if ((sopc != 0) && !arch_supports_vector(sopc, num_elem, elem_bt, mask_use_type)) { 407 log_if_needed(" ** not supported: arity=%d opc=%d vlen=%d etype=%s ismask=%d is_masked_op=%d", 408 n, sopc, num_elem, type2name(elem_bt), 409 is_vector_mask(vbox_klass) ? 1 : 0, is_masked_op ? 1 : 0); 410 return false; // not supported 411 } 412 413 // Return true if current platform has implemented the masked operation with predicate feature. 414 bool use_predicate = is_masked_op && sopc != 0 && arch_supports_vector(sopc, num_elem, elem_bt, VecMaskUsePred); 415 if (is_masked_op && !use_predicate && !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)) { 416 log_if_needed(" ** not supported: arity=%d opc=%d vlen=%d etype=%s ismask=0 is_masked_op=1", 417 n, sopc, num_elem, type2name(elem_bt)); 418 return false; 419 } 420 421 Node* opd1 = nullptr; Node* opd2 = nullptr; Node* opd3 = nullptr; 422 switch (n) { 423 case 3: { 424 opd3 = unbox_vector(argument(7), vbox_type, elem_bt, num_elem); 425 if (opd3 == nullptr) { 426 log_if_needed(" ** unbox failed v3=%s", 427 NodeClassNames[argument(7)->Opcode()]); 428 return false; 429 } 430 // fall-through 431 } 432 case 2: { 433 opd2 = unbox_vector(argument(6), vbox_type, elem_bt, num_elem); 434 if (opd2 == nullptr) { 435 log_if_needed(" ** unbox failed v2=%s", 436 NodeClassNames[argument(6)->Opcode()]); 437 return false; 438 } 439 // fall-through 440 } 441 case 1: { 442 opd1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 443 if (opd1 == nullptr) { 444 log_if_needed(" ** unbox failed v1=%s", 445 NodeClassNames[argument(5)->Opcode()]); 446 return false; 447 } 448 break; 449 } 450 default: fatal("unsupported arity: %d", n); 451 } 452 453 Node* mask = nullptr; 454 if (is_masked_op) { 455 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 456 assert(is_vector_mask(mbox_klass), "argument(2) should be a mask class"); 457 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 458 mask = unbox_vector(argument(n + 5), mbox_type, elem_bt, num_elem); 459 if (mask == nullptr) { 460 log_if_needed(" ** unbox failed mask=%s", 461 NodeClassNames[argument(n + 5)->Opcode()]); 462 return false; 463 } 464 } 465 466 Node* operation = nullptr; 467 if (opc == Op_CallLeafVector) { 468 assert(UseVectorStubs, "sanity"); 469 operation = gen_call_to_vector_math(opr->get_con(), elem_bt, num_elem, opd1, opd2); 470 if (operation == nullptr) { 471 log_if_needed(" ** Vector math call failed for %s_%s_%d", 472 (elem_bt == T_FLOAT) ? "float" : "double", 473 VectorSupport::mathname[opr->get_con() - VectorSupport::VECTOR_OP_MATH_START], 474 num_elem * type2aelembytes(elem_bt)); 475 return false; 476 } 477 } else { 478 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, is_vector_mask(vbox_klass)); 479 switch (n) { 480 case 1: 481 case 2: { 482 operation = VectorNode::make(sopc, opd1, opd2, vt, is_vector_mask(vbox_klass), VectorNode::is_shift_opcode(opc), is_unsigned); 483 break; 484 } 485 case 3: { 486 operation = VectorNode::make(sopc, opd1, opd2, opd3, vt); 487 break; 488 } 489 default: fatal("unsupported arity: %d", n); 490 } 491 } 492 493 if (is_masked_op && mask != nullptr) { 494 if (use_predicate) { 495 operation->add_req(mask); 496 operation->add_flag(Node::Flag_is_predicated_vector); 497 } else { 498 operation->add_flag(Node::Flag_is_predicated_using_blend); 499 operation = gvn().transform(operation); 500 operation = new VectorBlendNode(opd1, operation, mask); 501 } 502 } 503 operation = gvn().transform(operation); 504 505 // Wrap it up in VectorBox to keep object type information. 506 Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem); 507 set_result(vbox); 508 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 509 return true; 510 } 511 512 // public static 513 // <E, M> 514 // long maskReductionCoerced(int oper, Class<? extends M> maskClass, Class<?> elemClass, 515 // int length, M m, VectorMaskOp<M> defaultImpl) 516 bool LibraryCallKit::inline_vector_mask_operation() { 517 const TypeInt* oper = gvn().type(argument(0))->isa_int(); 518 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 519 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 520 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 521 Node* mask = argument(4); 522 523 if (mask_klass == nullptr || elem_klass == nullptr || mask->is_top() || vlen == nullptr) { 524 return false; // dead code 525 } 526 527 if (!is_klass_initialized(mask_klass)) { 528 log_if_needed(" ** klass argument not initialized"); 529 return false; 530 } 531 532 int num_elem = vlen->get_con(); 533 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 534 BasicType elem_bt = elem_type->basic_type(); 535 536 int mopc = VectorSupport::vop2ideal(oper->get_con(), elem_bt); 537 if (!arch_supports_vector(mopc, num_elem, elem_bt, VecMaskUseLoad)) { 538 log_if_needed(" ** not supported: arity=1 op=cast#%d/3 vlen2=%d etype2=%s", 539 mopc, num_elem, type2name(elem_bt)); 540 return false; // not supported 541 } 542 543 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 544 const TypeInstPtr* mask_box_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 545 Node* mask_vec = unbox_vector(mask, mask_box_type, elem_bt, num_elem); 546 if (mask_vec == nullptr) { 547 log_if_needed(" ** unbox failed mask=%s", 548 NodeClassNames[argument(4)->Opcode()]); 549 return false; 550 } 551 552 if (mask_vec->bottom_type()->isa_vectmask() == nullptr) { 553 mask_vec = gvn().transform(VectorStoreMaskNode::make(gvn(), mask_vec, elem_bt, num_elem)); 554 } 555 const Type* maskoper_ty = mopc == Op_VectorMaskToLong ? (const Type*)TypeLong::LONG : (const Type*)TypeInt::INT; 556 Node* maskoper = gvn().transform(VectorMaskOpNode::make(mask_vec, maskoper_ty, mopc)); 557 if (mopc != Op_VectorMaskToLong) { 558 maskoper = ConvI2L(maskoper); 559 } 560 set_result(maskoper); 561 562 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 563 return true; 564 } 565 566 // public static 567 // <M, 568 // S extends VectorSpecies<E>, 569 // E> 570 // M fromBitsCoerced(Class<? extends M> vmClass, Class<E> elementType, int length, 571 // long bits, int mode, S s, 572 // BroadcastOperation<M, E, S> defaultImpl) 573 bool LibraryCallKit::inline_vector_frombits_coerced() { 574 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 575 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 576 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 577 const TypeLong* bits_type = gvn().type(argument(3))->isa_long(); 578 // Mode argument determines the mode of operation it can take following values:- 579 // MODE_BROADCAST for vector Vector.broadcast and VectorMask.maskAll operations. 580 // MODE_BITS_COERCED_LONG_TO_MASK for VectorMask.fromLong operation. 581 const TypeInt* mode = gvn().type(argument(5))->isa_int(); 582 583 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || mode == nullptr || 584 bits_type == nullptr || vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || 585 !vlen->is_con() || !mode->is_con()) { 586 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s bitwise=%s", 587 NodeClassNames[argument(0)->Opcode()], 588 NodeClassNames[argument(1)->Opcode()], 589 NodeClassNames[argument(2)->Opcode()], 590 NodeClassNames[argument(5)->Opcode()]); 591 return false; // not enough info for intrinsification 592 } 593 594 if (!is_klass_initialized(vector_klass)) { 595 log_if_needed(" ** klass argument not initialized"); 596 return false; 597 } 598 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 599 if (!elem_type->is_primitive_type()) { 600 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 601 return false; // should be primitive type 602 } 603 BasicType elem_bt = elem_type->basic_type(); 604 int num_elem = vlen->get_con(); 605 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 606 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 607 608 bool is_mask = is_vector_mask(vbox_klass); 609 int bcast_mode = mode->get_con(); 610 VectorMaskUseType checkFlags = (VectorMaskUseType)(is_mask ? VecMaskUseAll : VecMaskNotUsed); 611 int opc = bcast_mode == VectorSupport::MODE_BITS_COERCED_LONG_TO_MASK ? Op_VectorLongToMask : Op_Replicate; 612 613 if (!arch_supports_vector(opc, num_elem, elem_bt, checkFlags, true /*has_scalar_args*/)) { 614 log_if_needed(" ** not supported: arity=0 op=broadcast vlen=%d etype=%s ismask=%d bcast_mode=%d", 615 num_elem, type2name(elem_bt), 616 is_mask ? 1 : 0, 617 bcast_mode); 618 return false; // not supported 619 } 620 621 Node* broadcast = nullptr; 622 Node* bits = argument(3); 623 Node* elem = bits; 624 625 if (opc == Op_VectorLongToMask) { 626 const TypeVect* vt = TypeVect::makemask(elem_bt, num_elem); 627 if (vt->isa_vectmask()) { 628 broadcast = gvn().transform(new VectorLongToMaskNode(elem, vt)); 629 } else { 630 const TypeVect* mvt = TypeVect::make(T_BOOLEAN, num_elem); 631 broadcast = gvn().transform(new VectorLongToMaskNode(elem, mvt)); 632 broadcast = gvn().transform(new VectorLoadMaskNode(broadcast, vt)); 633 } 634 } else { 635 switch (elem_bt) { 636 case T_BOOLEAN: // fall-through 637 case T_BYTE: // fall-through 638 case T_SHORT: // fall-through 639 case T_CHAR: // fall-through 640 case T_INT: { 641 elem = gvn().transform(new ConvL2INode(bits)); 642 break; 643 } 644 case T_DOUBLE: { 645 elem = gvn().transform(new MoveL2DNode(bits)); 646 break; 647 } 648 case T_FLOAT: { 649 bits = gvn().transform(new ConvL2INode(bits)); 650 elem = gvn().transform(new MoveI2FNode(bits)); 651 break; 652 } 653 case T_LONG: { 654 // no conversion needed 655 break; 656 } 657 default: fatal("%s", type2name(elem_bt)); 658 } 659 broadcast = VectorNode::scalar2vector(elem, num_elem, elem_bt, is_mask); 660 broadcast = gvn().transform(broadcast); 661 } 662 663 Node* box = box_vector(broadcast, vbox_type, elem_bt, num_elem); 664 set_result(box); 665 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 666 return true; 667 } 668 669 static bool elem_consistent_with_arr(BasicType elem_bt, const TypeAryPtr* arr_type, bool mismatched_ms) { 670 assert(arr_type != nullptr, "unexpected"); 671 BasicType arr_elem_bt = arr_type->elem()->array_element_basic_type(); 672 if (elem_bt == arr_elem_bt) { 673 return true; 674 } else if (elem_bt == T_SHORT && arr_elem_bt == T_CHAR) { 675 // Load/store of short vector from/to char[] is supported 676 return true; 677 } else if (elem_bt == T_BYTE && arr_elem_bt == T_BOOLEAN) { 678 // Load/store of byte vector from/to boolean[] is supported 679 return true; 680 } else { 681 return mismatched_ms; 682 } 683 } 684 685 // public static 686 // <C, 687 // VM extends VectorPayload, 688 // E, 689 // S extends VectorSpecies<E>> 690 // VM load(Class<? extends VM> vmClass, Class<E> eClass, 691 // int length, 692 // Object base, long offset, // Unsafe addressing 693 // boolean fromSegment, 694 // C container, long index, S s, // Arguments for default implementation 695 // LoadOperation<C, VM, S> defaultImpl) { 696 // public static 697 // <C, 698 // V extends VectorPayload> 699 // void store(Class<?> vClass, Class<?> eClass, 700 // int length, 701 // Object base, long offset, // Unsafe addressing 702 // boolean fromSegment, 703 // V v, C container, long index, // Arguments for default implementation 704 // StoreVectorOperation<C, V> defaultImpl) { 705 bool LibraryCallKit::inline_vector_mem_operation(bool is_store) { 706 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 707 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 708 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 709 const TypeInt* from_ms = gvn().type(argument(6))->isa_int(); 710 711 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || !from_ms->is_con() || 712 vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 713 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s from_ms=%s", 714 NodeClassNames[argument(0)->Opcode()], 715 NodeClassNames[argument(1)->Opcode()], 716 NodeClassNames[argument(2)->Opcode()], 717 NodeClassNames[argument(6)->Opcode()]); 718 return false; // not enough info for intrinsification 719 } 720 if (!is_klass_initialized(vector_klass)) { 721 log_if_needed(" ** klass argument not initialized"); 722 return false; 723 } 724 725 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 726 if (!elem_type->is_primitive_type()) { 727 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 728 return false; // should be primitive type 729 } 730 BasicType elem_bt = elem_type->basic_type(); 731 int num_elem = vlen->get_con(); 732 733 // TODO When mask usage is supported, VecMaskNotUsed needs to be VecMaskUseLoad. 734 if (!arch_supports_vector(is_store ? Op_StoreVector : Op_LoadVector, num_elem, elem_bt, VecMaskNotUsed)) { 735 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s ismask=no", 736 is_store, is_store ? "store" : "load", 737 num_elem, type2name(elem_bt)); 738 return false; // not supported 739 } 740 741 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 742 bool is_mask = is_vector_mask(vbox_klass); 743 744 Node* base = argument(3); 745 Node* offset = ConvL2X(argument(4)); 746 747 // Save state and restore on bailout 748 uint old_sp = sp(); 749 SafePointNode* old_map = clone_map(); 750 751 Node* addr = make_unsafe_address(base, offset, (is_mask ? T_BOOLEAN : elem_bt), true); 752 753 // The memory barrier checks are based on ones for unsafe access. 754 // This is not 1-1 implementation. 755 const Type *const base_type = gvn().type(base); 756 757 const TypePtr *addr_type = gvn().type(addr)->isa_ptr(); 758 const TypeAryPtr* arr_type = addr_type->isa_aryptr(); 759 760 const bool in_native = TypePtr::NULL_PTR == base_type; // base always null 761 const bool in_heap = !TypePtr::NULL_PTR->higher_equal(base_type); // base never null 762 763 const bool is_mixed_access = !in_heap && !in_native; 764 765 const bool is_mismatched_access = in_heap && (addr_type->isa_aryptr() == nullptr); 766 767 const bool needs_cpu_membar = is_mixed_access || is_mismatched_access; 768 769 // For non-masked mismatched memory segment vector read/write accesses, intrinsification can continue 770 // with unknown backing storage type and compiler can skip inserting explicit reinterpretation IR after 771 // loading from or before storing to backing storage which is mandatory for semantic correctness of 772 // big-endian memory layout. 773 bool mismatched_ms = LITTLE_ENDIAN_ONLY(false) 774 BIG_ENDIAN_ONLY(from_ms->get_con() && !is_mask && arr_type != nullptr && 775 arr_type->elem()->array_element_basic_type() != elem_bt); 776 BasicType mem_elem_bt = mismatched_ms ? arr_type->elem()->array_element_basic_type() : elem_bt; 777 if (!is_java_primitive(mem_elem_bt)) { 778 log_if_needed(" ** non-primitive array element type"); 779 return false; 780 } 781 int mem_num_elem = mismatched_ms ? (num_elem * type2aelembytes(elem_bt)) / type2aelembytes(mem_elem_bt) : num_elem; 782 if (arr_type != nullptr && !is_mask && !elem_consistent_with_arr(elem_bt, arr_type, mismatched_ms)) { 783 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s atype=%s ismask=no", 784 is_store, is_store ? "store" : "load", 785 num_elem, type2name(elem_bt), type2name(arr_type->elem()->array_element_basic_type())); 786 set_map(old_map); 787 set_sp(old_sp); 788 return false; 789 } 790 791 // In case of mismatched memory segment accesses, we need to double check that the source type memory operations are supported by backend. 792 if (mismatched_ms) { 793 if (is_store) { 794 if (!arch_supports_vector(Op_StoreVector, num_elem, elem_bt, VecMaskNotUsed) 795 || !arch_supports_vector(Op_VectorReinterpret, mem_num_elem, mem_elem_bt, VecMaskNotUsed)) { 796 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d*8 etype=%s/8 ismask=no", 797 is_store, "store", 798 num_elem, type2name(elem_bt)); 799 set_map(old_map); 800 set_sp(old_sp); 801 return false; // not supported 802 } 803 } else { 804 if (!arch_supports_vector(Op_LoadVector, mem_num_elem, mem_elem_bt, VecMaskNotUsed) 805 || !arch_supports_vector(Op_VectorReinterpret, num_elem, elem_bt, VecMaskNotUsed)) { 806 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d*8 etype=%s/8 ismask=no", 807 is_store, "load", 808 mem_num_elem, type2name(mem_elem_bt)); 809 set_map(old_map); 810 set_sp(old_sp); 811 return false; // not supported 812 } 813 } 814 } 815 if (is_mask) { 816 if (!is_store) { 817 if (!arch_supports_vector(Op_LoadVector, num_elem, elem_bt, VecMaskUseLoad)) { 818 set_map(old_map); 819 set_sp(old_sp); 820 return false; // not supported 821 } 822 } else { 823 if (!arch_supports_vector(Op_StoreVector, num_elem, elem_bt, VecMaskUseStore)) { 824 set_map(old_map); 825 set_sp(old_sp); 826 return false; // not supported 827 } 828 } 829 } 830 831 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 832 833 if (needs_cpu_membar) { 834 insert_mem_bar(Op_MemBarCPUOrder); 835 } 836 837 if (is_store) { 838 Node* val = unbox_vector(argument(7), vbox_type, elem_bt, num_elem); 839 if (val == nullptr) { 840 set_map(old_map); 841 set_sp(old_sp); 842 return false; // operand unboxing failed 843 } 844 set_all_memory(reset_memory()); 845 846 // In case the store needs to happen to byte array, reinterpret the incoming vector to byte vector. 847 int store_num_elem = num_elem; 848 if (mismatched_ms) { 849 store_num_elem = mem_num_elem; 850 const TypeVect* to_vect_type = TypeVect::make(mem_elem_bt, store_num_elem); 851 val = gvn().transform(new VectorReinterpretNode(val, val->bottom_type()->is_vect(), to_vect_type)); 852 } 853 if (is_mask) { 854 val = gvn().transform(VectorStoreMaskNode::make(gvn(), val, elem_bt, num_elem)); 855 } 856 Node* vstore = gvn().transform(StoreVectorNode::make(0, control(), memory(addr), addr, addr_type, val, store_num_elem)); 857 set_memory(vstore, addr_type); 858 } else { 859 // When using byte array, we need to load as byte then reinterpret the value. Otherwise, do a simple vector load. 860 Node* vload = nullptr; 861 if (mismatched_ms) { 862 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, mem_num_elem, mem_elem_bt)); 863 const TypeVect* to_vect_type = TypeVect::make(elem_bt, num_elem); 864 vload = gvn().transform(new VectorReinterpretNode(vload, vload->bottom_type()->is_vect(), to_vect_type)); 865 } else { 866 // Special handle for masks 867 if (is_mask) { 868 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, num_elem, T_BOOLEAN)); 869 vload = gvn().transform(new VectorLoadMaskNode(vload, TypeVect::makemask(elem_bt, num_elem))); 870 } else { 871 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, num_elem, elem_bt)); 872 } 873 } 874 Node* box = box_vector(vload, vbox_type, elem_bt, num_elem); 875 set_result(box); 876 } 877 878 destruct_map_clone(old_map); 879 880 if (needs_cpu_membar) { 881 insert_mem_bar(Op_MemBarCPUOrder); 882 } 883 884 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 885 return true; 886 } 887 888 // public static 889 // <C, 890 // V extends Vector<?>, 891 // E, 892 // S extends VectorSpecies<E>, 893 // M extends VectorMask<E>> 894 // V loadMasked(Class<? extends V> vClass, Class<M> mClass, Class<E> eClass, 895 // int length, Object base, long offset, // Unsafe addressing 896 // boolean fromSegment, 897 // M m, int offsetInRange, 898 // C container, long index, S s, // Arguments for default implementation 899 // LoadVectorMaskedOperation<C, V, S, M> defaultImpl) { 900 // public static 901 // <C, 902 // V extends Vector<E>, 903 // M extends VectorMask<E>, 904 // E> 905 // void storeMasked(Class<? extends V> vClass, Class<M> mClass, Class<E> eClass, 906 // int length, 907 // Object base, long offset, // Unsafe addressing 908 // boolean fromSegment, 909 // V v, M m, C container, long index, // Arguments for default implementation 910 // StoreVectorMaskedOperation<C, V, M> defaultImpl) { 911 912 bool LibraryCallKit::inline_vector_mem_masked_operation(bool is_store) { 913 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 914 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 915 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 916 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 917 const TypeInt* from_ms = gvn().type(argument(7))->isa_int(); 918 919 if (vector_klass == nullptr || mask_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 920 vector_klass->const_oop() == nullptr || mask_klass->const_oop() == nullptr || from_ms == nullptr || 921 elem_klass->const_oop() == nullptr || !vlen->is_con() || !from_ms->is_con()) { 922 log_if_needed(" ** missing constant: vclass=%s mclass=%s etype=%s vlen=%s from_ms=%s", 923 NodeClassNames[argument(0)->Opcode()], 924 NodeClassNames[argument(1)->Opcode()], 925 NodeClassNames[argument(2)->Opcode()], 926 NodeClassNames[argument(3)->Opcode()], 927 NodeClassNames[argument(7)->Opcode()]); 928 return false; // not enough info for intrinsification 929 } 930 if (!is_klass_initialized(vector_klass)) { 931 log_if_needed(" ** klass argument not initialized"); 932 return false; 933 } 934 935 if (!is_klass_initialized(mask_klass)) { 936 log_if_needed(" ** mask klass argument not initialized"); 937 return false; 938 } 939 940 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 941 if (!elem_type->is_primitive_type()) { 942 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 943 return false; // should be primitive type 944 } 945 946 BasicType elem_bt = elem_type->basic_type(); 947 int num_elem = vlen->get_con(); 948 949 Node* base = argument(4); 950 Node* offset = ConvL2X(argument(5)); 951 952 // Save state and restore on bailout 953 uint old_sp = sp(); 954 SafePointNode* old_map = clone_map(); 955 956 Node* addr = make_unsafe_address(base, offset, elem_bt, true); 957 const TypePtr *addr_type = gvn().type(addr)->isa_ptr(); 958 const TypeAryPtr* arr_type = addr_type->isa_aryptr(); 959 960 bool mismatched_ms = from_ms->get_con() && arr_type != nullptr && arr_type->elem()->array_element_basic_type() != elem_bt; 961 BIG_ENDIAN_ONLY(if (mismatched_ms) return false;) 962 // If there is no consistency between array and vector element types, it must be special byte array case 963 if (arr_type != nullptr && !elem_consistent_with_arr(elem_bt, arr_type, mismatched_ms)) { 964 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s atype=%s", 965 is_store, is_store ? "storeMasked" : "loadMasked", 966 num_elem, type2name(elem_bt), type2name(arr_type->elem()->array_element_basic_type())); 967 set_map(old_map); 968 set_sp(old_sp); 969 return false; 970 } 971 972 int mem_num_elem = mismatched_ms ? num_elem * type2aelembytes(elem_bt) : num_elem; 973 BasicType mem_elem_bt = mismatched_ms ? T_BYTE : elem_bt; 974 bool supports_predicate = arch_supports_vector(is_store ? Op_StoreVectorMasked : Op_LoadVectorMasked, 975 mem_num_elem, mem_elem_bt, VecMaskUseLoad); 976 977 // If current arch does not support the predicated operations, we have to bail 978 // out when current case uses the predicate feature. 979 if (!supports_predicate) { 980 bool needs_predicate = false; 981 if (is_store) { 982 // Masked vector store always uses the predicated store. 983 needs_predicate = true; 984 } else { 985 // Masked vector load with IOOBE always uses the predicated load. 986 const TypeInt* offset_in_range = gvn().type(argument(9))->isa_int(); 987 if (!offset_in_range->is_con()) { 988 log_if_needed(" ** missing constant: offsetInRange=%s", 989 NodeClassNames[argument(8)->Opcode()]); 990 set_map(old_map); 991 set_sp(old_sp); 992 return false; 993 } 994 needs_predicate = (offset_in_range->get_con() == 0); 995 } 996 997 if (needs_predicate) { 998 log_if_needed(" ** not supported: op=%s vlen=%d etype=%s mismatched_ms=%d", 999 is_store ? "storeMasked" : "loadMasked", 1000 num_elem, type2name(elem_bt), mismatched_ms ? 1 : 0); 1001 set_map(old_map); 1002 set_sp(old_sp); 1003 return false; 1004 } 1005 } 1006 1007 // This only happens for masked vector load. If predicate is not supported, then check whether 1008 // the normal vector load and blend operations are supported by backend. 1009 if (!supports_predicate && (!arch_supports_vector(Op_LoadVector, mem_num_elem, mem_elem_bt, VecMaskNotUsed) || 1010 !arch_supports_vector(Op_VectorBlend, mem_num_elem, mem_elem_bt, VecMaskUseLoad))) { 1011 log_if_needed(" ** not supported: op=loadMasked vlen=%d etype=%s mismatched_ms=%d", 1012 num_elem, type2name(elem_bt), mismatched_ms ? 1 : 0); 1013 set_map(old_map); 1014 set_sp(old_sp); 1015 return false; 1016 } 1017 1018 // Since we are using byte array, we need to double check that the vector reinterpret operation 1019 // with byte type is supported by backend. 1020 if (mismatched_ms) { 1021 if (!arch_supports_vector(Op_VectorReinterpret, mem_num_elem, T_BYTE, VecMaskNotUsed)) { 1022 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s mismatched_ms=1", 1023 is_store, is_store ? "storeMasked" : "loadMasked", 1024 num_elem, type2name(elem_bt)); 1025 set_map(old_map); 1026 set_sp(old_sp); 1027 return false; 1028 } 1029 } 1030 1031 // Since it needs to unbox the mask, we need to double check that the related load operations 1032 // for mask are supported by backend. 1033 if (!arch_supports_vector(Op_LoadVector, num_elem, elem_bt, VecMaskUseLoad)) { 1034 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s", 1035 is_store, is_store ? "storeMasked" : "loadMasked", 1036 num_elem, type2name(elem_bt)); 1037 set_map(old_map); 1038 set_sp(old_sp); 1039 return false; 1040 } 1041 1042 // Can base be null? Otherwise, always on-heap access. 1043 bool can_access_non_heap = TypePtr::NULL_PTR->higher_equal(gvn().type(base)); 1044 if (can_access_non_heap) { 1045 insert_mem_bar(Op_MemBarCPUOrder); 1046 } 1047 1048 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1049 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1050 assert(!is_vector_mask(vbox_klass) && is_vector_mask(mbox_klass), "Invalid class type"); 1051 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1052 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1053 1054 Node* mask = unbox_vector(is_store ? argument(9) : argument(8), mbox_type, elem_bt, num_elem); 1055 if (mask == nullptr) { 1056 log_if_needed(" ** unbox failed mask=%s", 1057 is_store ? NodeClassNames[argument(9)->Opcode()] 1058 : NodeClassNames[argument(8)->Opcode()]); 1059 set_map(old_map); 1060 set_sp(old_sp); 1061 return false; 1062 } 1063 1064 if (is_store) { 1065 Node* val = unbox_vector(argument(8), vbox_type, elem_bt, num_elem); 1066 if (val == nullptr) { 1067 log_if_needed(" ** unbox failed vector=%s", 1068 NodeClassNames[argument(8)->Opcode()]); 1069 set_map(old_map); 1070 set_sp(old_sp); 1071 return false; // operand unboxing failed 1072 } 1073 set_all_memory(reset_memory()); 1074 1075 if (mismatched_ms) { 1076 // Reinterpret the incoming vector to byte vector. 1077 const TypeVect* to_vect_type = TypeVect::make(mem_elem_bt, mem_num_elem); 1078 val = gvn().transform(new VectorReinterpretNode(val, val->bottom_type()->is_vect(), to_vect_type)); 1079 // Reinterpret the vector mask to byte type. 1080 const TypeVect* from_mask_type = TypeVect::makemask(elem_bt, num_elem); 1081 const TypeVect* to_mask_type = TypeVect::makemask(mem_elem_bt, mem_num_elem); 1082 mask = gvn().transform(new VectorReinterpretNode(mask, from_mask_type, to_mask_type)); 1083 } 1084 Node* vstore = gvn().transform(new StoreVectorMaskedNode(control(), memory(addr), addr, val, addr_type, mask)); 1085 set_memory(vstore, addr_type); 1086 } else { 1087 Node* vload = nullptr; 1088 1089 if (mismatched_ms) { 1090 // Reinterpret the vector mask to byte type. 1091 const TypeVect* from_mask_type = TypeVect::makemask(elem_bt, num_elem); 1092 const TypeVect* to_mask_type = TypeVect::makemask(mem_elem_bt, mem_num_elem); 1093 mask = gvn().transform(new VectorReinterpretNode(mask, from_mask_type, to_mask_type)); 1094 } 1095 1096 if (supports_predicate) { 1097 // Generate masked load vector node if predicate feature is supported. 1098 const TypeVect* vt = TypeVect::make(mem_elem_bt, mem_num_elem); 1099 vload = gvn().transform(new LoadVectorMaskedNode(control(), memory(addr), addr, addr_type, vt, mask)); 1100 } else { 1101 // Use the vector blend to implement the masked load vector. The biased elements are zeros. 1102 Node* zero = gvn().transform(gvn().zerocon(mem_elem_bt)); 1103 zero = gvn().transform(VectorNode::scalar2vector(zero, mem_num_elem, mem_elem_bt)); 1104 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, mem_num_elem, mem_elem_bt)); 1105 vload = gvn().transform(new VectorBlendNode(zero, vload, mask)); 1106 } 1107 1108 if (mismatched_ms) { 1109 const TypeVect* to_vect_type = TypeVect::make(elem_bt, num_elem); 1110 vload = gvn().transform(new VectorReinterpretNode(vload, vload->bottom_type()->is_vect(), to_vect_type)); 1111 } 1112 1113 Node* box = box_vector(vload, vbox_type, elem_bt, num_elem); 1114 set_result(box); 1115 } 1116 1117 destruct_map_clone(old_map); 1118 1119 if (can_access_non_heap) { 1120 insert_mem_bar(Op_MemBarCPUOrder); 1121 } 1122 1123 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1124 return true; 1125 } 1126 1127 // <C, 1128 // V extends Vector<?>, 1129 // W extends Vector<Integer>, 1130 // S extends VectorSpecies<E>, 1131 // M extends VectorMask<E>, 1132 // E> 1133 // V loadWithMap(Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, int length, 1134 // Class<? extends Vector<Integer>> vectorIndexClass, 1135 // Object base, long offset, // Unsafe addressing 1136 // W index_vector, M m, 1137 // C container, int index, int[] indexMap, int indexM, S s, // Arguments for default implementation 1138 // LoadVectorOperationWithMap<C, V, E, S, M> defaultImpl) 1139 // 1140 // <C, 1141 // V extends Vector<E>, 1142 // W extends Vector<Integer>, 1143 // M extends VectorMask<E>, 1144 // E> 1145 // void storeWithMap(Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, 1146 // int length, Class<? extends Vector<Integer>> vectorIndexClass, Object base, long offset, // Unsafe addressing 1147 // W index_vector, V v, M m, 1148 // C container, int index, int[] indexMap, int indexM, // Arguments for default implementation 1149 // StoreVectorOperationWithMap<C, V, M, E> defaultImpl) 1150 // 1151 bool LibraryCallKit::inline_vector_gather_scatter(bool is_scatter) { 1152 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1153 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 1154 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1155 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1156 const TypeInstPtr* vector_idx_klass = gvn().type(argument(4))->isa_instptr(); 1157 1158 if (vector_klass == nullptr || elem_klass == nullptr || vector_idx_klass == nullptr || vlen == nullptr || 1159 vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || vector_idx_klass->const_oop() == nullptr || !vlen->is_con()) { 1160 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s viclass=%s", 1161 NodeClassNames[argument(0)->Opcode()], 1162 NodeClassNames[argument(2)->Opcode()], 1163 NodeClassNames[argument(3)->Opcode()], 1164 NodeClassNames[argument(4)->Opcode()]); 1165 return false; // not enough info for intrinsification 1166 } 1167 1168 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(vector_idx_klass)) { 1169 log_if_needed(" ** klass argument not initialized"); 1170 return false; 1171 } 1172 1173 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1174 if (!elem_type->is_primitive_type()) { 1175 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1176 return false; // should be primitive type 1177 } 1178 1179 BasicType elem_bt = elem_type->basic_type(); 1180 int num_elem = vlen->get_con(); 1181 1182 const Type* vmask_type = gvn().type(is_scatter ? argument(10) : argument(9)); 1183 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 1184 if (is_masked_op) { 1185 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 1186 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(1)->Opcode()]); 1187 return false; // not enough info for intrinsification 1188 } 1189 1190 if (!is_klass_initialized(mask_klass)) { 1191 log_if_needed(" ** mask klass argument not initialized"); 1192 return false; 1193 } 1194 1195 if (vmask_type->maybe_null()) { 1196 log_if_needed(" ** null mask values are not allowed for masked op"); 1197 return false; 1198 } 1199 1200 // Check whether the predicated gather/scatter node is supported by architecture. 1201 VectorMaskUseType mask = (VectorMaskUseType) (VecMaskUseLoad | VecMaskUsePred); 1202 if (!arch_supports_vector(is_scatter ? Op_StoreVectorScatterMasked : Op_LoadVectorGatherMasked, num_elem, elem_bt, mask)) { 1203 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s is_masked_op=1", 1204 is_scatter, is_scatter ? "scatterMasked" : "gatherMasked", 1205 num_elem, type2name(elem_bt)); 1206 return false; // not supported 1207 } 1208 } else { 1209 // Check whether the normal gather/scatter node is supported for non-masked operation. 1210 if (!arch_supports_vector(is_scatter ? Op_StoreVectorScatter : Op_LoadVectorGather, num_elem, elem_bt, VecMaskNotUsed)) { 1211 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s is_masked_op=0", 1212 is_scatter, is_scatter ? "scatter" : "gather", 1213 num_elem, type2name(elem_bt)); 1214 return false; // not supported 1215 } 1216 } 1217 1218 // Check that the vector holding indices is supported by architecture 1219 // For sub-word gathers expander receive index array. 1220 if (!is_subword_type(elem_bt) && !arch_supports_vector(Op_LoadVector, num_elem, T_INT, VecMaskNotUsed)) { 1221 log_if_needed(" ** not supported: arity=%d op=%s/loadindex vlen=%d etype=int is_masked_op=%d", 1222 is_scatter, is_scatter ? "scatter" : "gather", 1223 num_elem, is_masked_op ? 1 : 0); 1224 return false; // not supported 1225 } 1226 1227 Node* base = argument(5); 1228 Node* offset = ConvL2X(argument(6)); 1229 1230 // Save state and restore on bailout 1231 uint old_sp = sp(); 1232 SafePointNode* old_map = clone_map(); 1233 1234 Node* addr = make_unsafe_address(base, offset, elem_bt, true); 1235 1236 const TypePtr *addr_type = gvn().type(addr)->isa_ptr(); 1237 const TypeAryPtr* arr_type = addr_type->isa_aryptr(); 1238 1239 // The array must be consistent with vector type 1240 if (arr_type == nullptr || (arr_type != nullptr && !elem_consistent_with_arr(elem_bt, arr_type, false))) { 1241 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s atype=%s ismask=no", 1242 is_scatter, is_scatter ? "scatter" : "gather", 1243 num_elem, type2name(elem_bt), type2name(arr_type->elem()->array_element_basic_type())); 1244 set_map(old_map); 1245 set_sp(old_sp); 1246 return false; 1247 } 1248 1249 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1250 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1251 ciKlass* vbox_idx_klass = vector_idx_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1252 if (vbox_idx_klass == nullptr) { 1253 set_map(old_map); 1254 set_sp(old_sp); 1255 return false; 1256 } 1257 1258 Node* index_vect = nullptr; 1259 const TypeInstPtr* vbox_idx_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_idx_klass); 1260 if (!is_subword_type(elem_bt)) { 1261 index_vect = unbox_vector(argument(8), vbox_idx_type, T_INT, num_elem); 1262 if (index_vect == nullptr) { 1263 set_map(old_map); 1264 set_sp(old_sp); 1265 return false; 1266 } 1267 } 1268 1269 Node* mask = nullptr; 1270 if (is_masked_op) { 1271 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1272 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1273 mask = unbox_vector(is_scatter ? argument(10) : argument(9), mbox_type, elem_bt, num_elem); 1274 if (mask == nullptr) { 1275 log_if_needed(" ** unbox failed mask=%s", 1276 is_scatter ? NodeClassNames[argument(10)->Opcode()] 1277 : NodeClassNames[argument(9)->Opcode()]); 1278 set_map(old_map); 1279 set_sp(old_sp); 1280 return false; 1281 } 1282 } 1283 1284 const TypeVect* vector_type = TypeVect::make(elem_bt, num_elem); 1285 if (is_scatter) { 1286 Node* val = unbox_vector(argument(9), vbox_type, elem_bt, num_elem); 1287 if (val == nullptr) { 1288 set_map(old_map); 1289 set_sp(old_sp); 1290 return false; // operand unboxing failed 1291 } 1292 set_all_memory(reset_memory()); 1293 1294 Node* vstore = nullptr; 1295 if (mask != nullptr) { 1296 vstore = gvn().transform(new StoreVectorScatterMaskedNode(control(), memory(addr), addr, addr_type, val, index_vect, mask)); 1297 } else { 1298 vstore = gvn().transform(new StoreVectorScatterNode(control(), memory(addr), addr, addr_type, val, index_vect)); 1299 } 1300 set_memory(vstore, addr_type); 1301 } else { 1302 Node* vload = nullptr; 1303 Node* index = argument(11); 1304 Node* indexMap = argument(12); 1305 Node* indexM = argument(13); 1306 if (mask != nullptr) { 1307 if (is_subword_type(elem_bt)) { 1308 Node* index_arr_base = array_element_address(indexMap, indexM, T_INT); 1309 vload = gvn().transform(new LoadVectorGatherMaskedNode(control(), memory(addr), addr, addr_type, vector_type, index_arr_base, mask, index)); 1310 } else { 1311 vload = gvn().transform(new LoadVectorGatherMaskedNode(control(), memory(addr), addr, addr_type, vector_type, index_vect, mask)); 1312 } 1313 } else { 1314 if (is_subword_type(elem_bt)) { 1315 Node* index_arr_base = array_element_address(indexMap, indexM, T_INT); 1316 vload = gvn().transform(new LoadVectorGatherNode(control(), memory(addr), addr, addr_type, vector_type, index_arr_base, index)); 1317 } else { 1318 vload = gvn().transform(new LoadVectorGatherNode(control(), memory(addr), addr, addr_type, vector_type, index_vect)); 1319 } 1320 } 1321 Node* box = box_vector(vload, vbox_type, elem_bt, num_elem); 1322 set_result(box); 1323 } 1324 1325 destruct_map_clone(old_map); 1326 1327 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1328 return true; 1329 } 1330 1331 // public static 1332 // <V extends Vector<E>, 1333 // M extends VectorMask<E>, 1334 // E> 1335 // long reductionCoerced(int oprId, Class<? extends V> vectorClass, Class<? extends M> maskClass, 1336 // Class<E> elementType, int length, V v, M m, 1337 // ReductionOperation<V, M> defaultImpl) 1338 bool LibraryCallKit::inline_vector_reduction() { 1339 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 1340 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 1341 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 1342 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 1343 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 1344 1345 if (opr == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 1346 !opr->is_con() || vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 1347 log_if_needed(" ** missing constant: opr=%s vclass=%s etype=%s vlen=%s", 1348 NodeClassNames[argument(0)->Opcode()], 1349 NodeClassNames[argument(1)->Opcode()], 1350 NodeClassNames[argument(3)->Opcode()], 1351 NodeClassNames[argument(4)->Opcode()]); 1352 return false; // not enough info for intrinsification 1353 } 1354 if (!is_klass_initialized(vector_klass)) { 1355 log_if_needed(" ** klass argument not initialized"); 1356 return false; 1357 } 1358 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1359 if (!elem_type->is_primitive_type()) { 1360 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1361 return false; // should be primitive type 1362 } 1363 1364 const Type* vmask_type = gvn().type(argument(6)); 1365 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 1366 if (is_masked_op) { 1367 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 1368 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(2)->Opcode()]); 1369 return false; // not enough info for intrinsification 1370 } 1371 1372 if (!is_klass_initialized(mask_klass)) { 1373 log_if_needed(" ** mask klass argument not initialized"); 1374 return false; 1375 } 1376 1377 if (vmask_type->maybe_null()) { 1378 log_if_needed(" ** null mask values are not allowed for masked op"); 1379 return false; 1380 } 1381 } 1382 1383 BasicType elem_bt = elem_type->basic_type(); 1384 int num_elem = vlen->get_con(); 1385 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 1386 int sopc = ReductionNode::opcode(opc, elem_bt); 1387 1388 // Ensure reduction operation for lanewise operation 1389 // When using mask, mask use type needs to be VecMaskUseLoad. 1390 if (sopc == opc || !arch_supports_vector(sopc, num_elem, elem_bt, is_masked_op ? VecMaskUseLoad : VecMaskNotUsed)) { 1391 log_if_needed(" ** not supported: arity=1 op=%d/reduce vlen=%d etype=%s is_masked_op=%d", 1392 sopc, num_elem, type2name(elem_bt), is_masked_op ? 1 : 0); 1393 return false; 1394 } 1395 1396 // Return true if current platform has implemented the masked operation with predicate feature. 1397 bool use_predicate = is_masked_op && arch_supports_vector(sopc, num_elem, elem_bt, VecMaskUsePred); 1398 if (is_masked_op && !use_predicate && !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)) { 1399 log_if_needed(" ** not supported: arity=1 op=%d/reduce vlen=%d etype=%s is_masked_op=1", 1400 sopc, num_elem, type2name(elem_bt)); 1401 return false; 1402 } 1403 1404 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1405 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1406 1407 Node* opd = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1408 if (opd == nullptr) { 1409 return false; // operand unboxing failed 1410 } 1411 1412 Node* mask = nullptr; 1413 if (is_masked_op) { 1414 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1415 assert(is_vector_mask(mbox_klass), "argument(2) should be a mask class"); 1416 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1417 mask = unbox_vector(argument(6), mbox_type, elem_bt, num_elem); 1418 if (mask == nullptr) { 1419 log_if_needed(" ** unbox failed mask=%s", 1420 NodeClassNames[argument(6)->Opcode()]); 1421 return false; 1422 } 1423 } 1424 1425 Node* init = ReductionNode::make_identity_con_scalar(gvn(), opc, elem_bt); 1426 Node* value = opd; 1427 1428 assert(mask != nullptr || !is_masked_op, "Masked op needs the mask value never null"); 1429 if (mask != nullptr && !use_predicate) { 1430 Node* reduce_identity = gvn().transform(VectorNode::scalar2vector(init, num_elem, elem_bt)); 1431 value = gvn().transform(new VectorBlendNode(reduce_identity, value, mask)); 1432 } 1433 1434 // Make an unordered Reduction node. This affects only AddReductionVF/VD and MulReductionVF/VD, 1435 // as these operations are allowed to be associative (not requiring strict order) in VectorAPI. 1436 value = ReductionNode::make(opc, nullptr, init, value, elem_bt, /* requires_strict_order */ false); 1437 1438 if (mask != nullptr && use_predicate) { 1439 value->add_req(mask); 1440 value->add_flag(Node::Flag_is_predicated_vector); 1441 } 1442 1443 value = gvn().transform(value); 1444 1445 Node* bits = nullptr; 1446 switch (elem_bt) { 1447 case T_BYTE: 1448 case T_SHORT: 1449 case T_INT: { 1450 bits = gvn().transform(new ConvI2LNode(value)); 1451 break; 1452 } 1453 case T_FLOAT: { 1454 value = gvn().transform(new MoveF2INode(value)); 1455 bits = gvn().transform(new ConvI2LNode(value)); 1456 break; 1457 } 1458 case T_DOUBLE: { 1459 bits = gvn().transform(new MoveD2LNode(value)); 1460 break; 1461 } 1462 case T_LONG: { 1463 bits = value; // no conversion needed 1464 break; 1465 } 1466 default: fatal("%s", type2name(elem_bt)); 1467 } 1468 set_result(bits); 1469 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1470 return true; 1471 } 1472 1473 // public static <V> boolean test(int cond, Class<?> vectorClass, Class<?> elementType, int vlen, 1474 // V v1, V v2, 1475 // BiFunction<V, V, Boolean> defaultImpl) 1476 // 1477 bool LibraryCallKit::inline_vector_test() { 1478 const TypeInt* cond = gvn().type(argument(0))->isa_int(); 1479 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 1480 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1481 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1482 1483 if (cond == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 1484 !cond->is_con() || vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 1485 log_if_needed(" ** missing constant: cond=%s vclass=%s etype=%s vlen=%s", 1486 NodeClassNames[argument(0)->Opcode()], 1487 NodeClassNames[argument(1)->Opcode()], 1488 NodeClassNames[argument(2)->Opcode()], 1489 NodeClassNames[argument(3)->Opcode()]); 1490 return false; // not enough info for intrinsification 1491 } 1492 if (!is_klass_initialized(vector_klass)) { 1493 log_if_needed(" ** klass argument not initialized"); 1494 return false; 1495 } 1496 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1497 if (!elem_type->is_primitive_type()) { 1498 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1499 return false; // should be primitive type 1500 } 1501 BasicType elem_bt = elem_type->basic_type(); 1502 int num_elem = vlen->get_con(); 1503 BoolTest::mask booltest = (BoolTest::mask)cond->get_con(); 1504 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1505 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1506 1507 if (!arch_supports_vector(Op_VectorTest, num_elem, elem_bt, is_vector_mask(vbox_klass) ? VecMaskUseLoad : VecMaskNotUsed)) { 1508 log_if_needed(" ** not supported: arity=2 op=test/%d vlen=%d etype=%s ismask=%d", 1509 cond->get_con(), num_elem, type2name(elem_bt), 1510 is_vector_mask(vbox_klass)); 1511 return false; 1512 } 1513 1514 Node* opd1 = unbox_vector(argument(4), vbox_type, elem_bt, num_elem); 1515 Node* opd2; 1516 if (Matcher::vectortest_needs_second_argument(booltest == BoolTest::overflow, 1517 opd1->bottom_type()->isa_vectmask())) { 1518 opd2 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1519 } else { 1520 opd2 = opd1; 1521 } 1522 if (opd1 == nullptr || opd2 == nullptr) { 1523 return false; // operand unboxing failed 1524 } 1525 1526 Node* cmp = gvn().transform(new VectorTestNode(opd1, opd2, booltest)); 1527 BoolTest::mask test = Matcher::vectortest_mask(booltest == BoolTest::overflow, 1528 opd1->bottom_type()->isa_vectmask(), num_elem); 1529 Node* bol = gvn().transform(new BoolNode(cmp, test)); 1530 Node* res = gvn().transform(new CMoveINode(bol, gvn().intcon(0), gvn().intcon(1), TypeInt::BOOL)); 1531 1532 set_result(res); 1533 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1534 return true; 1535 } 1536 1537 // public static 1538 // <V extends Vector<E>, 1539 // M extends VectorMask<E>, 1540 // E> 1541 // V blend(Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, int vlen, 1542 // V v1, V v2, M m, 1543 // VectorBlendOp<V, M, E> defaultImpl) 1544 bool LibraryCallKit::inline_vector_blend() { 1545 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1546 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 1547 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1548 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1549 1550 if (mask_klass == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 1551 return false; // dead code 1552 } 1553 if (mask_klass->const_oop() == nullptr || vector_klass->const_oop() == nullptr || 1554 elem_klass->const_oop() == nullptr || !vlen->is_con()) { 1555 log_if_needed(" ** missing constant: vclass=%s mclass=%s etype=%s vlen=%s", 1556 NodeClassNames[argument(0)->Opcode()], 1557 NodeClassNames[argument(1)->Opcode()], 1558 NodeClassNames[argument(2)->Opcode()], 1559 NodeClassNames[argument(3)->Opcode()]); 1560 return false; // not enough info for intrinsification 1561 } 1562 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(mask_klass)) { 1563 log_if_needed(" ** klass argument not initialized"); 1564 return false; 1565 } 1566 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1567 if (!elem_type->is_primitive_type()) { 1568 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1569 return false; // should be primitive type 1570 } 1571 BasicType elem_bt = elem_type->basic_type(); 1572 BasicType mask_bt = elem_bt; 1573 int num_elem = vlen->get_con(); 1574 1575 if (!arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)) { 1576 log_if_needed(" ** not supported: arity=2 op=blend vlen=%d etype=%s ismask=useload", 1577 num_elem, type2name(elem_bt)); 1578 return false; // not supported 1579 } 1580 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1581 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1582 1583 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1584 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1585 1586 Node* v1 = unbox_vector(argument(4), vbox_type, elem_bt, num_elem); 1587 Node* v2 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1588 Node* mask = unbox_vector(argument(6), mbox_type, mask_bt, num_elem); 1589 1590 if (v1 == nullptr || v2 == nullptr || mask == nullptr) { 1591 return false; // operand unboxing failed 1592 } 1593 1594 Node* blend = gvn().transform(new VectorBlendNode(v1, v2, mask)); 1595 1596 Node* box = box_vector(blend, vbox_type, elem_bt, num_elem); 1597 set_result(box); 1598 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1599 return true; 1600 } 1601 1602 // public static 1603 // <V extends Vector<E>, 1604 // M extends VectorMask<E>, 1605 // E> 1606 // M compare(int cond, Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, int vlen, 1607 // V v1, V v2, M m, 1608 // VectorCompareOp<V,M> defaultImpl) 1609 bool LibraryCallKit::inline_vector_compare() { 1610 const TypeInt* cond = gvn().type(argument(0))->isa_int(); 1611 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 1612 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 1613 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 1614 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 1615 1616 if (cond == nullptr || vector_klass == nullptr || mask_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 1617 return false; // dead code 1618 } 1619 if (!cond->is_con() || vector_klass->const_oop() == nullptr || mask_klass->const_oop() == nullptr || 1620 elem_klass->const_oop() == nullptr || !vlen->is_con()) { 1621 log_if_needed(" ** missing constant: cond=%s vclass=%s mclass=%s etype=%s vlen=%s", 1622 NodeClassNames[argument(0)->Opcode()], 1623 NodeClassNames[argument(1)->Opcode()], 1624 NodeClassNames[argument(2)->Opcode()], 1625 NodeClassNames[argument(3)->Opcode()], 1626 NodeClassNames[argument(4)->Opcode()]); 1627 return false; // not enough info for intrinsification 1628 } 1629 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(mask_klass)) { 1630 log_if_needed(" ** klass argument not initialized"); 1631 return false; 1632 } 1633 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1634 if (!elem_type->is_primitive_type()) { 1635 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1636 return false; // should be primitive type 1637 } 1638 1639 int num_elem = vlen->get_con(); 1640 BasicType elem_bt = elem_type->basic_type(); 1641 BasicType mask_bt = elem_bt; 1642 1643 if ((cond->get_con() & BoolTest::unsigned_compare) != 0) { 1644 if (!Matcher::supports_vector_comparison_unsigned(num_elem, elem_bt)) { 1645 log_if_needed(" ** not supported: unsigned comparison op=comp/%d vlen=%d etype=%s ismask=usestore", 1646 cond->get_con() & (BoolTest::unsigned_compare - 1), num_elem, type2name(elem_bt)); 1647 return false; 1648 } 1649 } 1650 1651 if (!arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskUseStore)) { 1652 log_if_needed(" ** not supported: arity=2 op=comp/%d vlen=%d etype=%s ismask=usestore", 1653 cond->get_con(), num_elem, type2name(elem_bt)); 1654 return false; 1655 } 1656 1657 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1658 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1659 1660 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1661 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1662 1663 Node* v1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1664 Node* v2 = unbox_vector(argument(6), vbox_type, elem_bt, num_elem); 1665 1666 bool is_masked_op = argument(7)->bottom_type() != TypePtr::NULL_PTR; 1667 Node* mask = is_masked_op ? unbox_vector(argument(7), mbox_type, elem_bt, num_elem) : nullptr; 1668 if (is_masked_op && mask == nullptr) { 1669 log_if_needed(" ** not supported: mask = null arity=2 op=comp/%d vlen=%d etype=%s ismask=usestore is_masked_op=1", 1670 cond->get_con(), num_elem, type2name(elem_bt)); 1671 return false; 1672 } 1673 1674 bool use_predicate = is_masked_op && arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskUsePred); 1675 if (is_masked_op && !use_predicate && !arch_supports_vector(Op_AndV, num_elem, elem_bt, VecMaskUseLoad)) { 1676 log_if_needed(" ** not supported: arity=2 op=comp/%d vlen=%d etype=%s ismask=usestore is_masked_op=1", 1677 cond->get_con(), num_elem, type2name(elem_bt)); 1678 return false; 1679 } 1680 1681 if (v1 == nullptr || v2 == nullptr) { 1682 return false; // operand unboxing failed 1683 } 1684 BoolTest::mask pred = (BoolTest::mask)cond->get_con(); 1685 ConINode* pred_node = (ConINode*)gvn().makecon(cond); 1686 1687 const TypeVect* vmask_type = TypeVect::makemask(mask_bt, num_elem); 1688 Node* operation = new VectorMaskCmpNode(pred, v1, v2, pred_node, vmask_type); 1689 1690 if (is_masked_op) { 1691 if (use_predicate) { 1692 operation->add_req(mask); 1693 operation->add_flag(Node::Flag_is_predicated_vector); 1694 } else { 1695 operation = gvn().transform(operation); 1696 operation = VectorNode::make(Op_AndV, operation, mask, vmask_type); 1697 } 1698 } 1699 1700 operation = gvn().transform(operation); 1701 1702 Node* box = box_vector(operation, mbox_type, mask_bt, num_elem); 1703 set_result(box); 1704 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1705 return true; 1706 } 1707 1708 // public static 1709 // <V extends Vector<E>, 1710 // Sh extends VectorShuffle<E>, 1711 // M extends VectorMask<E>, 1712 // E> 1713 // V rearrangeOp(Class<? extends V> vectorClass, Class<Sh> shuffleClass, Class<M> maskClass, Class<E> elementType, int vlen, 1714 // V v1, Sh sh, M m, 1715 // VectorRearrangeOp<V, Sh, M, E> defaultImpl) 1716 bool LibraryCallKit::inline_vector_rearrange() { 1717 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1718 const TypeInstPtr* shuffle_klass = gvn().type(argument(1))->isa_instptr(); 1719 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 1720 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 1721 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 1722 1723 if (vector_klass == nullptr || shuffle_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 1724 return false; // dead code 1725 } 1726 if (shuffle_klass->const_oop() == nullptr || 1727 vector_klass->const_oop() == nullptr || 1728 elem_klass->const_oop() == nullptr || 1729 !vlen->is_con()) { 1730 log_if_needed(" ** missing constant: vclass=%s sclass=%s etype=%s vlen=%s", 1731 NodeClassNames[argument(0)->Opcode()], 1732 NodeClassNames[argument(1)->Opcode()], 1733 NodeClassNames[argument(3)->Opcode()], 1734 NodeClassNames[argument(4)->Opcode()]); 1735 return false; // not enough info for intrinsification 1736 } 1737 if (!is_klass_initialized(vector_klass) || 1738 !is_klass_initialized(shuffle_klass)) { 1739 log_if_needed(" ** klass argument not initialized"); 1740 return false; 1741 } 1742 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1743 if (!elem_type->is_primitive_type()) { 1744 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1745 return false; // should be primitive type 1746 } 1747 1748 BasicType elem_bt = elem_type->basic_type(); 1749 BasicType shuffle_bt = elem_bt; 1750 if (shuffle_bt == T_FLOAT) { 1751 shuffle_bt = T_INT; 1752 } else if (shuffle_bt == T_DOUBLE) { 1753 shuffle_bt = T_LONG; 1754 } 1755 1756 int num_elem = vlen->get_con(); 1757 bool need_load_shuffle = Matcher::vector_rearrange_requires_load_shuffle(shuffle_bt, num_elem); 1758 1759 if (need_load_shuffle && !arch_supports_vector(Op_VectorLoadShuffle, num_elem, shuffle_bt, VecMaskNotUsed)) { 1760 if (C->print_intrinsics()) { 1761 tty->print_cr(" ** not supported: arity=0 op=load/shuffle vlen=%d etype=%s ismask=no", 1762 num_elem, type2name(shuffle_bt)); 1763 } 1764 return false; // not supported 1765 } 1766 1767 bool is_masked_op = argument(7)->bottom_type() != TypePtr::NULL_PTR; 1768 bool use_predicate = is_masked_op; 1769 if (is_masked_op && 1770 (mask_klass == nullptr || 1771 mask_klass->const_oop() == nullptr || 1772 !is_klass_initialized(mask_klass))) { 1773 log_if_needed(" ** mask_klass argument not initialized"); 1774 } 1775 if (!arch_supports_vector(Op_AndV, num_elem, elem_bt, VecMaskNotUsed)) { 1776 log_if_needed(" ** not supported: arity=2 op=and vlen=%d etype=%s ismask=no", 1777 num_elem, type2name(elem_bt)); 1778 return false; 1779 } 1780 VectorMaskUseType checkFlags = (VectorMaskUseType)(is_masked_op ? (VecMaskUseLoad | VecMaskUsePred) : VecMaskNotUsed); 1781 if (!arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, checkFlags)) { 1782 use_predicate = false; 1783 if(!is_masked_op || 1784 (!arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, VecMaskNotUsed) || 1785 !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad) || 1786 !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed))) { 1787 log_if_needed(" ** not supported: arity=2 op=shuffle/rearrange vlen=%d etype=%s ismask=no", 1788 num_elem, type2name(elem_bt)); 1789 return false; // not supported 1790 } 1791 } 1792 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1793 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1794 1795 ciKlass* shbox_klass = shuffle_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1796 const TypeInstPtr* shbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, shbox_klass); 1797 1798 Node* v1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1799 Node* shuffle = unbox_vector(argument(6), shbox_type, shuffle_bt, num_elem); 1800 const TypeVect* st = TypeVect::make(shuffle_bt, num_elem); 1801 1802 if (v1 == nullptr || shuffle == nullptr) { 1803 return false; // operand unboxing failed 1804 } 1805 1806 assert(is_power_of_2(num_elem), "wrapping invalid"); 1807 Node* wrapping_mask_elem = gvn().makecon(TypeInteger::make(num_elem - 1, num_elem - 1, Type::WidenMin, shuffle_bt == T_LONG ? T_LONG : T_INT)); 1808 Node* wrapping_mask = gvn().transform(VectorNode::scalar2vector(wrapping_mask_elem, num_elem, shuffle_bt)); 1809 shuffle = gvn().transform(new AndVNode(shuffle, wrapping_mask, st)); 1810 1811 Node* mask = nullptr; 1812 if (is_masked_op) { 1813 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1814 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1815 mask = unbox_vector(argument(7), mbox_type, elem_bt, num_elem); 1816 if (mask == nullptr) { 1817 log_if_needed(" ** not supported: arity=3 op=shuffle/rearrange vlen=%d etype=%s ismask=useload is_masked_op=1", 1818 num_elem, type2name(elem_bt)); 1819 return false; 1820 } 1821 } 1822 1823 if (need_load_shuffle) { 1824 shuffle = gvn().transform(new VectorLoadShuffleNode(shuffle, st)); 1825 } 1826 1827 Node* rearrange = new VectorRearrangeNode(v1, shuffle); 1828 if (is_masked_op) { 1829 if (use_predicate) { 1830 rearrange->add_req(mask); 1831 rearrange->add_flag(Node::Flag_is_predicated_vector); 1832 } else { 1833 rearrange = gvn().transform(rearrange); 1834 Node* zero = gvn().makecon(Type::get_zero_type(elem_bt)); 1835 Node* zerovec = gvn().transform(VectorNode::scalar2vector(zero, num_elem, elem_bt)); 1836 rearrange = new VectorBlendNode(zerovec, rearrange, mask); 1837 } 1838 } 1839 rearrange = gvn().transform(rearrange); 1840 1841 Node* box = box_vector(rearrange, vbox_type, elem_bt, num_elem); 1842 set_result(box); 1843 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1844 return true; 1845 } 1846 1847 static address get_vector_math_address(int vop, int bits, BasicType bt, char* name_ptr, int name_len) { 1848 address addr = nullptr; 1849 assert(UseVectorStubs, "sanity"); 1850 assert(name_ptr != nullptr, "unexpected"); 1851 assert((vop >= VectorSupport::VECTOR_OP_MATH_START) && (vop <= VectorSupport::VECTOR_OP_MATH_END), "unexpected"); 1852 int op = vop - VectorSupport::VECTOR_OP_MATH_START; 1853 1854 switch(bits) { 1855 case 64: //fallthough 1856 case 128: //fallthough 1857 case 256: //fallthough 1858 case 512: 1859 if (bt == T_FLOAT) { 1860 snprintf(name_ptr, name_len, "vector_%s_float_%dbits_fixed", VectorSupport::mathname[op], bits); 1861 addr = StubRoutines::_vector_f_math[exact_log2(bits/64)][op]; 1862 } else { 1863 assert(bt == T_DOUBLE, "must be FP type only"); 1864 snprintf(name_ptr, name_len, "vector_%s_double_%dbits_fixed", VectorSupport::mathname[op], bits); 1865 addr = StubRoutines::_vector_d_math[exact_log2(bits/64)][op]; 1866 } 1867 break; 1868 default: 1869 if (!Matcher::supports_scalable_vector() || !Matcher::vector_size_supported(bt, bits/type2aelembytes(bt)) ) { 1870 snprintf(name_ptr, name_len, "invalid"); 1871 addr = nullptr; 1872 Unimplemented(); 1873 } 1874 break; 1875 } 1876 1877 if (addr == nullptr && Matcher::supports_scalable_vector()) { 1878 if (bt == T_FLOAT) { 1879 snprintf(name_ptr, name_len, "vector_%s_float_%dbits_scalable", VectorSupport::mathname[op], bits); 1880 addr = StubRoutines::_vector_f_math[VectorSupport::VEC_SIZE_SCALABLE][op]; 1881 } else { 1882 assert(bt == T_DOUBLE, "must be FP type only"); 1883 snprintf(name_ptr, name_len, "vector_%s_double_%dbits_scalable", VectorSupport::mathname[op], bits); 1884 addr = StubRoutines::_vector_d_math[VectorSupport::VEC_SIZE_SCALABLE][op]; 1885 } 1886 } 1887 1888 return addr; 1889 } 1890 1891 // public static 1892 // <V extends Vector<E>, 1893 // M extends VectorMask<E>, 1894 // E> 1895 // V selectFromOp(Class<? extends V> vClass, Class<M> mClass, Class<E> eClass, 1896 // int length, V v1, V v2, M m, 1897 // VectorSelectFromOp<V, M> defaultImpl) 1898 bool LibraryCallKit::inline_vector_select_from() { 1899 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1900 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 1901 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1902 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1903 1904 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 1905 vector_klass->const_oop() == nullptr || 1906 elem_klass->const_oop() == nullptr || 1907 !vlen->is_con()) { 1908 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s", 1909 NodeClassNames[argument(0)->Opcode()], 1910 NodeClassNames[argument(2)->Opcode()], 1911 NodeClassNames[argument(3)->Opcode()]); 1912 return false; // not enough info for intrinsification 1913 } 1914 if (!is_klass_initialized(vector_klass)) { 1915 log_if_needed(" ** klass argument not initialized"); 1916 return false; 1917 } 1918 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1919 if (!elem_type->is_primitive_type()) { 1920 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1921 return false; // should be primitive type 1922 } 1923 BasicType elem_bt = elem_type->basic_type(); 1924 int num_elem = vlen->get_con(); 1925 if (!is_power_of_2(num_elem)) { 1926 log_if_needed(" ** vlen not power of two=%d", num_elem); 1927 return false; 1928 } 1929 1930 BasicType shuffle_bt = elem_bt; 1931 if (shuffle_bt == T_FLOAT) { 1932 shuffle_bt = T_INT; 1933 } else if (shuffle_bt == T_DOUBLE) { 1934 shuffle_bt = T_LONG; 1935 } 1936 bool need_load_shuffle = Matcher::vector_rearrange_requires_load_shuffle(shuffle_bt, num_elem); 1937 1938 int cast_vopc = VectorCastNode::opcode(-1, elem_bt); // from vector of type elem_bt 1939 if ((need_load_shuffle && !arch_supports_vector(Op_VectorLoadShuffle, num_elem, elem_bt, VecMaskNotUsed)) || 1940 (elem_bt != shuffle_bt && !arch_supports_vector(cast_vopc, num_elem, shuffle_bt, VecMaskNotUsed)) || 1941 !arch_supports_vector(Op_AndV, num_elem, shuffle_bt, VecMaskNotUsed) || 1942 !arch_supports_vector(Op_Replicate, num_elem, shuffle_bt, VecMaskNotUsed)) { 1943 log_if_needed(" ** not supported: arity=0 op=selectFrom vlen=%d etype=%s ismask=no", 1944 num_elem, type2name(elem_bt)); 1945 return false; // not supported 1946 } 1947 1948 bool is_masked_op = argument(6)->bottom_type() != TypePtr::NULL_PTR; 1949 bool use_predicate = is_masked_op; 1950 if (is_masked_op && 1951 (mask_klass == nullptr || 1952 mask_klass->const_oop() == nullptr || 1953 !is_klass_initialized(mask_klass))) { 1954 log_if_needed(" ** mask_klass argument not initialized"); 1955 return false; // not supported 1956 } 1957 VectorMaskUseType checkFlags = (VectorMaskUseType)(is_masked_op ? (VecMaskUseLoad | VecMaskUsePred) : VecMaskNotUsed); 1958 if (!arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, checkFlags)) { 1959 use_predicate = false; 1960 if(!is_masked_op || 1961 (!arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, VecMaskNotUsed) || 1962 !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad) || 1963 !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed))) { 1964 log_if_needed(" ** not supported: op=selectFrom vlen=%d etype=%s is_masked_op=%d", 1965 num_elem, type2name(elem_bt), is_masked_op); 1966 return false; // not supported 1967 } 1968 } 1969 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1970 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1971 1972 // v1 is the index vector 1973 Node* v1 = unbox_vector(argument(4), vbox_type, elem_bt, num_elem); 1974 // v2 is the vector being rearranged 1975 Node* v2 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1976 1977 if (v1 == nullptr) { 1978 log_if_needed(" ** unbox failed v1=%s", NodeClassNames[argument(4)->Opcode()]); 1979 return false; // operand unboxing failed 1980 } 1981 1982 if (v2 == nullptr) { 1983 log_if_needed(" ** unbox failed v2=%s", NodeClassNames[argument(5)->Opcode()]); 1984 return false; // operand unboxing failed 1985 } 1986 1987 Node* mask = nullptr; 1988 if (is_masked_op) { 1989 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1990 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1991 mask = unbox_vector(argument(6), mbox_type, elem_bt, num_elem); 1992 if (mask == nullptr) { 1993 log_if_needed(" ** unbox failed mask=%s", NodeClassNames[argument(6)->Opcode()]); 1994 return false; 1995 } 1996 } 1997 1998 // cast index vector from elem_bt vector to byte vector 1999 const TypeVect* shuffle_vt = TypeVect::make(shuffle_bt, num_elem); 2000 Node* shuffle = v1; 2001 2002 if (shuffle_bt != elem_bt) { 2003 shuffle = gvn().transform(VectorCastNode::make(cast_vopc, v1, shuffle_bt, num_elem)); 2004 } 2005 2006 // wrap the byte vector lanes to (num_elem - 1) to form the shuffle vector where num_elem is vector length 2007 // this is a simple AND operation as we come here only for power of two vector length 2008 Node* mod_val = gvn().makecon(TypeInteger::make(num_elem - 1, num_elem - 1, Type::WidenMin, shuffle_bt == T_LONG ? T_LONG : T_INT)); 2009 Node* bcast_mod = gvn().transform(VectorNode::scalar2vector(mod_val, num_elem, shuffle_bt)); 2010 shuffle = gvn().transform(VectorNode::make(Op_AndV, shuffle, bcast_mod, shuffle_vt)); 2011 2012 // load the shuffle to use in rearrange 2013 if (need_load_shuffle) { 2014 shuffle = gvn().transform(new VectorLoadShuffleNode(shuffle, shuffle_vt)); 2015 } 2016 2017 // and finally rearrange 2018 Node* rearrange = new VectorRearrangeNode(v2, shuffle); 2019 if (is_masked_op) { 2020 if (use_predicate) { 2021 // masked rearrange is supported so use that directly 2022 rearrange->add_req(mask); 2023 rearrange->add_flag(Node::Flag_is_predicated_vector); 2024 } else { 2025 // masked rearrange is not supported so emulate usig blend 2026 const TypeVect* vt = v1->bottom_type()->is_vect(); 2027 rearrange = gvn().transform(rearrange); 2028 2029 // create a zero vector with each lane element set as zero 2030 Node* zero = gvn().makecon(Type::get_zero_type(elem_bt)); 2031 Node* zerovec = gvn().transform(VectorNode::scalar2vector(zero, num_elem, elem_bt)); 2032 2033 // For each lane for which mask is set, blend in the rearranged lane into zero vector 2034 rearrange = new VectorBlendNode(zerovec, rearrange, mask); 2035 } 2036 } 2037 rearrange = gvn().transform(rearrange); 2038 2039 // box the result 2040 Node* box = box_vector(rearrange, vbox_type, elem_bt, num_elem); 2041 set_result(box); 2042 2043 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2044 return true; 2045 } 2046 2047 Node* LibraryCallKit::gen_call_to_vector_math(int vector_api_op_id, BasicType bt, int num_elem, Node* opd1, Node* opd2) { 2048 assert(UseVectorStubs, "sanity"); 2049 assert(vector_api_op_id >= VectorSupport::VECTOR_OP_MATH_START && vector_api_op_id <= VectorSupport::VECTOR_OP_MATH_END, "need valid op id"); 2050 assert(opd1 != nullptr, "must not be null"); 2051 const TypeVect* vt = TypeVect::make(bt, num_elem); 2052 const TypeFunc* call_type = OptoRuntime::Math_Vector_Vector_Type(opd2 != nullptr ? 2 : 1, vt, vt); 2053 char name[100] = ""; 2054 2055 // Get address for vector math method. 2056 address addr = get_vector_math_address(vector_api_op_id, vt->length_in_bytes() * BitsPerByte, bt, name, 100); 2057 2058 if (addr == nullptr) { 2059 return nullptr; 2060 } 2061 2062 assert(name[0] != '\0', "name must not be null"); 2063 Node* operation = make_runtime_call(RC_VECTOR, 2064 call_type, 2065 addr, 2066 name, 2067 TypePtr::BOTTOM, 2068 opd1, 2069 opd2); 2070 return gvn().transform(new ProjNode(gvn().transform(operation), TypeFunc::Parms)); 2071 } 2072 2073 // public static 2074 // <V extends Vector<E>, 2075 // M extends VectorMask<E>, 2076 // E> 2077 // V broadcastInt(int opr, Class<? extends V> vectorClass, Class<? extends M> maskClass, 2078 // Class<E> elementType, int length, 2079 // V v, int n, M m, 2080 // VectorBroadcastIntOp<V, M> defaultImpl) 2081 bool LibraryCallKit::inline_vector_broadcast_int() { 2082 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 2083 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 2084 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 2085 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 2086 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 2087 2088 if (opr == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 2089 return false; // dead code 2090 } 2091 if (!opr->is_con() || vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 2092 log_if_needed(" ** missing constant: opr=%s vclass=%s etype=%s vlen=%s", 2093 NodeClassNames[argument(0)->Opcode()], 2094 NodeClassNames[argument(1)->Opcode()], 2095 NodeClassNames[argument(3)->Opcode()], 2096 NodeClassNames[argument(4)->Opcode()]); 2097 return false; // not enough info for intrinsification 2098 } 2099 if (!is_klass_initialized(vector_klass)) { 2100 log_if_needed(" ** klass argument not initialized"); 2101 return false; 2102 } 2103 2104 const Type* vmask_type = gvn().type(argument(7)); 2105 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 2106 if (is_masked_op) { 2107 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 2108 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(2)->Opcode()]); 2109 return false; // not enough info for intrinsification 2110 } 2111 2112 if (!is_klass_initialized(mask_klass)) { 2113 log_if_needed(" ** mask klass argument not initialized"); 2114 return false; 2115 } 2116 2117 if (vmask_type->maybe_null()) { 2118 log_if_needed(" ** null mask values are not allowed for masked op"); 2119 return false; 2120 } 2121 } 2122 2123 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2124 if (!elem_type->is_primitive_type()) { 2125 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2126 return false; // should be primitive type 2127 } 2128 2129 int num_elem = vlen->get_con(); 2130 BasicType elem_bt = elem_type->basic_type(); 2131 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 2132 2133 bool is_shift = VectorNode::is_shift_opcode(opc); 2134 bool is_rotate = VectorNode::is_rotate_opcode(opc); 2135 2136 if (opc == 0 || (!is_shift && !is_rotate)) { 2137 log_if_needed(" ** operation not supported: op=%d bt=%s", opr->get_con(), type2name(elem_bt)); 2138 return false; // operation not supported 2139 } 2140 2141 int sopc = VectorNode::opcode(opc, elem_bt); 2142 if (sopc == 0) { 2143 log_if_needed(" ** operation not supported: opc=%s bt=%s", NodeClassNames[opc], type2name(elem_bt)); 2144 return false; // operation not supported 2145 } 2146 2147 Node* cnt = argument(6); 2148 const TypeInt* cnt_type = cnt->bottom_type()->isa_int(); 2149 2150 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2151 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2152 2153 // If CPU supports vector constant rotate instructions pass it directly 2154 bool is_const_rotate = is_rotate && cnt_type && cnt_type->is_con() && 2155 Matcher::supports_vector_constant_rotates(cnt_type->get_con()); 2156 bool has_scalar_args = is_rotate ? !is_const_rotate : true; 2157 2158 VectorMaskUseType checkFlags = (VectorMaskUseType)(is_masked_op ? (VecMaskUseLoad | VecMaskUsePred) : VecMaskNotUsed); 2159 bool use_predicate = is_masked_op; 2160 2161 if (!arch_supports_vector(sopc, num_elem, elem_bt, checkFlags, has_scalar_args)) { 2162 use_predicate = false; 2163 if (!is_masked_op || 2164 (!arch_supports_vector(sopc, num_elem, elem_bt, VecMaskNotUsed, has_scalar_args) || 2165 !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad))) { 2166 2167 log_if_needed(" ** not supported: arity=0 op=int/%d vlen=%d etype=%s is_masked_op=%d", 2168 sopc, num_elem, type2name(elem_bt), is_masked_op ? 1 : 0); 2169 return false; // not supported 2170 } 2171 } 2172 2173 Node* opd1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 2174 Node* opd2 = nullptr; 2175 if (is_shift) { 2176 opd2 = vector_shift_count(cnt, opc, elem_bt, num_elem); 2177 } else { 2178 assert(is_rotate, "unexpected operation"); 2179 if (!is_const_rotate) { 2180 cnt = elem_bt == T_LONG ? gvn().transform(new ConvI2LNode(cnt)) : cnt; 2181 opd2 = gvn().transform(VectorNode::scalar2vector(cnt, num_elem, elem_bt)); 2182 } else { 2183 // Constant shift value. 2184 opd2 = cnt; 2185 } 2186 } 2187 2188 if (opd1 == nullptr || opd2 == nullptr) { 2189 return false; 2190 } 2191 2192 Node* mask = nullptr; 2193 if (is_masked_op) { 2194 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2195 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 2196 mask = unbox_vector(argument(7), mbox_type, elem_bt, num_elem); 2197 if (mask == nullptr) { 2198 log_if_needed(" ** unbox failed mask=%s", NodeClassNames[argument(7)->Opcode()]); 2199 return false; 2200 } 2201 } 2202 2203 Node* operation = VectorNode::make(opc, opd1, opd2, num_elem, elem_bt); 2204 if (is_masked_op && mask != nullptr) { 2205 if (use_predicate) { 2206 operation->add_req(mask); 2207 operation->add_flag(Node::Flag_is_predicated_vector); 2208 } else { 2209 operation = gvn().transform(operation); 2210 operation = new VectorBlendNode(opd1, operation, mask); 2211 } 2212 } 2213 operation = gvn().transform(operation); 2214 Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem); 2215 set_result(vbox); 2216 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2217 return true; 2218 } 2219 2220 // public static <VOUT extends VectorPayload, 2221 // VIN extends VectorPayload, 2222 // S extends VectorSpecies> 2223 // VOUT convert(int oprId, 2224 // Class<?> fromVectorClass, Class<?> fromElementType, int fromVLen, 2225 // Class<?> toVectorClass, Class<?> toElementType, int toVLen, 2226 // VIN v, S s, 2227 // VectorConvertOp<VOUT, VIN, S> defaultImpl) 2228 // 2229 bool LibraryCallKit::inline_vector_convert() { 2230 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 2231 2232 const TypeInstPtr* vector_klass_from = gvn().type(argument(1))->isa_instptr(); 2233 const TypeInstPtr* elem_klass_from = gvn().type(argument(2))->isa_instptr(); 2234 const TypeInt* vlen_from = gvn().type(argument(3))->isa_int(); 2235 2236 const TypeInstPtr* vector_klass_to = gvn().type(argument(4))->isa_instptr(); 2237 const TypeInstPtr* elem_klass_to = gvn().type(argument(5))->isa_instptr(); 2238 const TypeInt* vlen_to = gvn().type(argument(6))->isa_int(); 2239 2240 if (opr == nullptr || 2241 vector_klass_from == nullptr || elem_klass_from == nullptr || vlen_from == nullptr || 2242 vector_klass_to == nullptr || elem_klass_to == nullptr || vlen_to == nullptr) { 2243 return false; // dead code 2244 } 2245 if (!opr->is_con() || 2246 vector_klass_from->const_oop() == nullptr || elem_klass_from->const_oop() == nullptr || !vlen_from->is_con() || 2247 vector_klass_to->const_oop() == nullptr || elem_klass_to->const_oop() == nullptr || !vlen_to->is_con()) { 2248 log_if_needed(" ** missing constant: opr=%s vclass_from=%s etype_from=%s vlen_from=%s vclass_to=%s etype_to=%s vlen_to=%s", 2249 NodeClassNames[argument(0)->Opcode()], 2250 NodeClassNames[argument(1)->Opcode()], 2251 NodeClassNames[argument(2)->Opcode()], 2252 NodeClassNames[argument(3)->Opcode()], 2253 NodeClassNames[argument(4)->Opcode()], 2254 NodeClassNames[argument(5)->Opcode()], 2255 NodeClassNames[argument(6)->Opcode()]); 2256 return false; // not enough info for intrinsification 2257 } 2258 if (!is_klass_initialized(vector_klass_from) || !is_klass_initialized(vector_klass_to)) { 2259 log_if_needed(" ** klass argument not initialized"); 2260 return false; 2261 } 2262 2263 assert(opr->get_con() == VectorSupport::VECTOR_OP_CAST || 2264 opr->get_con() == VectorSupport::VECTOR_OP_UCAST || 2265 opr->get_con() == VectorSupport::VECTOR_OP_REINTERPRET, "wrong opcode"); 2266 bool is_cast = (opr->get_con() == VectorSupport::VECTOR_OP_CAST || opr->get_con() == VectorSupport::VECTOR_OP_UCAST); 2267 bool is_ucast = (opr->get_con() == VectorSupport::VECTOR_OP_UCAST); 2268 2269 ciKlass* vbox_klass_from = vector_klass_from->const_oop()->as_instance()->java_lang_Class_klass(); 2270 ciKlass* vbox_klass_to = vector_klass_to->const_oop()->as_instance()->java_lang_Class_klass(); 2271 2272 bool is_mask = is_vector_mask(vbox_klass_from); 2273 2274 ciType* elem_type_from = elem_klass_from->const_oop()->as_instance()->java_mirror_type(); 2275 if (!elem_type_from->is_primitive_type()) { 2276 return false; // should be primitive type 2277 } 2278 BasicType elem_bt_from = elem_type_from->basic_type(); 2279 ciType* elem_type_to = elem_klass_to->const_oop()->as_instance()->java_mirror_type(); 2280 if (!elem_type_to->is_primitive_type()) { 2281 return false; // should be primitive type 2282 } 2283 BasicType elem_bt_to = elem_type_to->basic_type(); 2284 2285 int num_elem_from = vlen_from->get_con(); 2286 int num_elem_to = vlen_to->get_con(); 2287 2288 // Check whether we can unbox to appropriate size. Even with casting, checking for reinterpret is needed 2289 // since we may need to change size. 2290 if (!arch_supports_vector(Op_VectorReinterpret, 2291 num_elem_from, 2292 elem_bt_from, 2293 is_mask ? VecMaskUseAll : VecMaskNotUsed)) { 2294 log_if_needed(" ** not supported: arity=1 op=%s/1 vlen1=%d etype1=%s ismask=%d", 2295 is_cast ? "cast" : "reinterpret", 2296 num_elem_from, type2name(elem_bt_from), is_mask); 2297 return false; 2298 } 2299 2300 // Check whether we can support resizing/reinterpreting to the new size. 2301 if (!arch_supports_vector(Op_VectorReinterpret, 2302 num_elem_to, 2303 elem_bt_to, 2304 is_mask ? VecMaskUseAll : VecMaskNotUsed)) { 2305 log_if_needed(" ** not supported: arity=1 op=%s/2 vlen2=%d etype2=%s ismask=%d", 2306 is_cast ? "cast" : "reinterpret", 2307 num_elem_to, type2name(elem_bt_to), is_mask); 2308 return false; 2309 } 2310 2311 // At this point, we know that both input and output vector registers are supported 2312 // by the architecture. Next check if the casted type is simply to same type - which means 2313 // that it is actually a resize and not a cast. 2314 if (is_cast && elem_bt_from == elem_bt_to) { 2315 is_cast = false; 2316 } 2317 2318 const TypeInstPtr* vbox_type_from = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass_from); 2319 2320 Node* opd1 = unbox_vector(argument(7), vbox_type_from, elem_bt_from, num_elem_from); 2321 if (opd1 == nullptr) { 2322 return false; 2323 } 2324 2325 const TypeVect* src_type = TypeVect::make(elem_bt_from, num_elem_from, is_mask); 2326 const TypeVect* dst_type = TypeVect::make(elem_bt_to, num_elem_to, is_mask); 2327 2328 // Safety check to prevent casting if source mask is of type vector 2329 // and destination mask of type predicate vector and vice-versa. 2330 // From X86 standpoint, this case will only arise over KNL target, 2331 // where certain masks (depending on the species) are either propagated 2332 // through a vector or predicate register. 2333 if (is_mask && 2334 ((src_type->isa_vectmask() == nullptr && dst_type->isa_vectmask()) || 2335 (dst_type->isa_vectmask() == nullptr && src_type->isa_vectmask()))) { 2336 return false; 2337 } 2338 2339 Node* op = opd1; 2340 if (is_cast) { 2341 assert(!is_mask || num_elem_from == num_elem_to, "vector mask cast needs the same elem num"); 2342 int cast_vopc = VectorCastNode::opcode(-1, elem_bt_from, !is_ucast); 2343 2344 // Make sure that vector cast is implemented to particular type/size combination if it is 2345 // not a mask casting. 2346 if (!is_mask && !arch_supports_vector(cast_vopc, num_elem_to, elem_bt_to, VecMaskNotUsed)) { 2347 log_if_needed(" ** not supported: arity=1 op=cast#%d/3 vlen2=%d etype2=%s ismask=%d", 2348 cast_vopc, num_elem_to, type2name(elem_bt_to), is_mask); 2349 return false; 2350 } 2351 2352 if (num_elem_from < num_elem_to) { 2353 // Since input and output number of elements are not consistent, we need to make sure we 2354 // properly size. Thus, first make a cast that retains the number of elements from source. 2355 int num_elem_for_cast = num_elem_from; 2356 2357 // It is possible that arch does not support this intermediate vector size 2358 // TODO More complex logic required here to handle this corner case for the sizes. 2359 if (!arch_supports_vector(cast_vopc, num_elem_for_cast, elem_bt_to, VecMaskNotUsed)) { 2360 log_if_needed(" ** not supported: arity=1 op=cast#%d/4 vlen1=%d etype2=%s ismask=%d", 2361 cast_vopc, 2362 num_elem_for_cast, type2name(elem_bt_to), is_mask); 2363 return false; 2364 } 2365 2366 op = gvn().transform(VectorCastNode::make(cast_vopc, op, elem_bt_to, num_elem_for_cast)); 2367 // Now ensure that the destination gets properly resized to needed size. 2368 op = gvn().transform(new VectorReinterpretNode(op, op->bottom_type()->is_vect(), dst_type)); 2369 } else if (num_elem_from > num_elem_to) { 2370 // Since number of elements from input is larger than output, simply reduce size of input 2371 // (we are supposed to drop top elements anyway). 2372 int num_elem_for_resize = num_elem_to; 2373 2374 // It is possible that arch does not support this intermediate vector size 2375 // TODO More complex logic required here to handle this corner case for the sizes. 2376 if (!arch_supports_vector(Op_VectorReinterpret, 2377 num_elem_for_resize, 2378 elem_bt_from, 2379 VecMaskNotUsed)) { 2380 log_if_needed(" ** not supported: arity=1 op=cast/5 vlen2=%d etype1=%s ismask=%d", 2381 num_elem_for_resize, type2name(elem_bt_from), is_mask); 2382 return false; 2383 } 2384 2385 const TypeVect* resize_type = TypeVect::make(elem_bt_from, num_elem_for_resize); 2386 op = gvn().transform(new VectorReinterpretNode(op, src_type, resize_type)); 2387 op = gvn().transform(VectorCastNode::make(cast_vopc, op, elem_bt_to, num_elem_to)); 2388 } else { // num_elem_from == num_elem_to 2389 if (is_mask) { 2390 // Make sure that cast for vector mask is implemented to particular type/size combination. 2391 if (!arch_supports_vector(Op_VectorMaskCast, num_elem_to, elem_bt_to, VecMaskNotUsed)) { 2392 log_if_needed(" ** not supported: arity=1 op=maskcast vlen2=%d etype2=%s ismask=%d", 2393 num_elem_to, type2name(elem_bt_to), is_mask); 2394 return false; 2395 } 2396 op = gvn().transform(new VectorMaskCastNode(op, dst_type)); 2397 } else { 2398 // Since input and output number of elements match, and since we know this vector size is 2399 // supported, simply do a cast with no resize needed. 2400 op = gvn().transform(VectorCastNode::make(cast_vopc, op, elem_bt_to, num_elem_to)); 2401 } 2402 } 2403 } else if (!Type::equals(src_type, dst_type)) { 2404 assert(!is_cast, "must be reinterpret"); 2405 op = gvn().transform(new VectorReinterpretNode(op, src_type, dst_type)); 2406 } 2407 2408 const TypeInstPtr* vbox_type_to = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass_to); 2409 Node* vbox = box_vector(op, vbox_type_to, elem_bt_to, num_elem_to); 2410 set_result(vbox); 2411 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem_to * type2aelembytes(elem_bt_to)))); 2412 return true; 2413 } 2414 2415 // public static 2416 // <V extends Vector<E>, 2417 // E> 2418 // V insert(Class<? extends V> vectorClass, Class<E> elementType, int vlen, 2419 // V vec, int ix, long val, 2420 // VecInsertOp<V> defaultImpl) 2421 bool LibraryCallKit::inline_vector_insert() { 2422 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 2423 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2424 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2425 const TypeInt* idx = gvn().type(argument(4))->isa_int(); 2426 2427 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || idx == nullptr) { 2428 return false; // dead code 2429 } 2430 if (vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con() || !idx->is_con()) { 2431 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s idx=%s", 2432 NodeClassNames[argument(0)->Opcode()], 2433 NodeClassNames[argument(1)->Opcode()], 2434 NodeClassNames[argument(2)->Opcode()], 2435 NodeClassNames[argument(4)->Opcode()]); 2436 return false; // not enough info for intrinsification 2437 } 2438 if (!is_klass_initialized(vector_klass)) { 2439 log_if_needed(" ** klass argument not initialized"); 2440 return false; 2441 } 2442 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2443 if (!elem_type->is_primitive_type()) { 2444 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2445 return false; // should be primitive type 2446 } 2447 BasicType elem_bt = elem_type->basic_type(); 2448 int num_elem = vlen->get_con(); 2449 if (!arch_supports_vector(Op_VectorInsert, num_elem, elem_bt, VecMaskNotUsed)) { 2450 log_if_needed(" ** not supported: arity=1 op=insert vlen=%d etype=%s ismask=no", 2451 num_elem, type2name(elem_bt)); 2452 return false; // not supported 2453 } 2454 2455 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2456 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2457 2458 Node* opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2459 if (opd == nullptr) { 2460 return false; 2461 } 2462 2463 Node* insert_val = argument(5); 2464 assert(gvn().type(insert_val)->isa_long() != nullptr, "expected to be long"); 2465 2466 // Convert insert value back to its appropriate type. 2467 switch (elem_bt) { 2468 case T_BYTE: 2469 insert_val = gvn().transform(new ConvL2INode(insert_val, TypeInt::BYTE)); 2470 break; 2471 case T_SHORT: 2472 insert_val = gvn().transform(new ConvL2INode(insert_val, TypeInt::SHORT)); 2473 break; 2474 case T_INT: 2475 insert_val = gvn().transform(new ConvL2INode(insert_val)); 2476 break; 2477 case T_FLOAT: 2478 insert_val = gvn().transform(new ConvL2INode(insert_val)); 2479 insert_val = gvn().transform(new MoveI2FNode(insert_val)); 2480 break; 2481 case T_DOUBLE: 2482 insert_val = gvn().transform(new MoveL2DNode(insert_val)); 2483 break; 2484 case T_LONG: 2485 // no conversion needed 2486 break; 2487 default: fatal("%s", type2name(elem_bt)); break; 2488 } 2489 2490 Node* operation = gvn().transform(VectorInsertNode::make(opd, insert_val, idx->get_con(), gvn())); 2491 2492 Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem); 2493 set_result(vbox); 2494 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2495 return true; 2496 } 2497 2498 // public static 2499 // <VM extends VectorPayload, 2500 // E> 2501 // long extract(Class<? extends VM> vClass, Class<E> eClass, 2502 // int length, 2503 // VM vm, int i, 2504 // VecExtractOp<VM> defaultImpl) 2505 bool LibraryCallKit::inline_vector_extract() { 2506 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 2507 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2508 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2509 const TypeInt* idx = gvn().type(argument(4))->isa_int(); 2510 2511 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || idx == nullptr) { 2512 return false; // dead code 2513 } 2514 if (vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 2515 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s", 2516 NodeClassNames[argument(0)->Opcode()], 2517 NodeClassNames[argument(1)->Opcode()], 2518 NodeClassNames[argument(2)->Opcode()]); 2519 return false; // not enough info for intrinsification 2520 } 2521 if (!is_klass_initialized(vector_klass)) { 2522 log_if_needed(" ** klass argument not initialized"); 2523 return false; 2524 } 2525 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2526 if (!elem_type->is_primitive_type()) { 2527 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2528 return false; // should be primitive type 2529 } 2530 BasicType elem_bt = elem_type->basic_type(); 2531 int num_elem = vlen->get_con(); 2532 2533 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2534 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2535 2536 Node* opd = nullptr; 2537 2538 if (is_vector_mask(vbox_klass)) { 2539 // vbox_klass is mask. This is used for VectorMask.laneIsSet(int). 2540 2541 Node* pos = argument(4); // can be variable 2542 if (arch_supports_vector(Op_ExtractUB, num_elem, elem_bt, VecMaskUseAll)) { 2543 // Transform mask to vector with type of boolean and utilize ExtractUB node. 2544 opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2545 if (opd == nullptr) { 2546 return false; 2547 } 2548 opd = gvn().transform(VectorStoreMaskNode::make(gvn(), opd, elem_bt, num_elem)); 2549 opd = gvn().transform(new ExtractUBNode(opd, pos)); 2550 opd = gvn().transform(new ConvI2LNode(opd)); 2551 } else if (arch_supports_vector(Op_VectorMaskToLong, num_elem, elem_bt, VecMaskUseLoad)) { 2552 opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2553 if (opd == nullptr) { 2554 return false; 2555 } 2556 // VectorMaskToLongNode requires the input is either a mask or a vector with BOOLEAN type. 2557 if (opd->bottom_type()->isa_vectmask() == nullptr) { 2558 opd = gvn().transform(VectorStoreMaskNode::make(gvn(), opd, elem_bt, num_elem)); 2559 } 2560 // ((toLong() >>> pos) & 1L 2561 opd = gvn().transform(new VectorMaskToLongNode(opd, TypeLong::LONG)); 2562 opd = gvn().transform(new URShiftLNode(opd, pos)); 2563 opd = gvn().transform(new AndLNode(opd, gvn().makecon(TypeLong::ONE))); 2564 } else { 2565 log_if_needed(" ** Rejected mask extraction because architecture does not support it"); 2566 return false; // not supported 2567 } 2568 } else { 2569 // vbox_klass is vector. This is used for Vector.lane(int). 2570 if (!idx->is_con()) { 2571 log_if_needed(" ** missing constant: idx=%s", NodeClassNames[argument(4)->Opcode()]); 2572 return false; // not enough info for intrinsification 2573 } 2574 2575 int vopc = ExtractNode::opcode(elem_bt); 2576 if (!arch_supports_vector(vopc, num_elem, elem_bt, VecMaskNotUsed)) { 2577 log_if_needed(" ** not supported: arity=1 op=extract vlen=%d etype=%s ismask=no", 2578 num_elem, type2name(elem_bt)); 2579 return false; // not supported 2580 } 2581 2582 opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2583 if (opd == nullptr) { 2584 return false; 2585 } 2586 ConINode* idx_con = gvn().intcon(idx->get_con())->as_ConI(); 2587 2588 opd = gvn().transform(ExtractNode::make(opd, idx_con, elem_bt)); 2589 switch (elem_bt) { 2590 case T_BYTE: 2591 case T_SHORT: 2592 case T_INT: { 2593 opd = gvn().transform(new ConvI2LNode(opd)); 2594 break; 2595 } 2596 case T_FLOAT: { 2597 opd = gvn().transform(new MoveF2INode(opd)); 2598 opd = gvn().transform(new ConvI2LNode(opd)); 2599 break; 2600 } 2601 case T_DOUBLE: { 2602 opd = gvn().transform(new MoveD2LNode(opd)); 2603 break; 2604 } 2605 case T_LONG: { 2606 // no conversion needed 2607 break; 2608 } 2609 default: fatal("%s", type2name(elem_bt)); 2610 } 2611 } 2612 set_result(opd); 2613 return true; 2614 } 2615 2616 static Node* LowerSelectFromTwoVectorOperation(PhaseGVN& phase, Node* index_vec, Node* src1, Node* src2, const TypeVect* vt) { 2617 int num_elem = vt->length(); 2618 BasicType elem_bt = vt->element_basic_type(); 2619 2620 // Lower selectFrom operation into its constituent operations. 2621 // SelectFromTwoVectorNode = 2622 // (VectorBlend 2623 // (VectorRearrange SRC1 (WRAPED_INDEX AND (VLEN-1)) 2624 // (VectorRearrange SRC2 (WRAPED_INDEX AND (VLEN-1)) 2625 // MASK) 2626 // Where 2627 // WRAPED_INDEX are computed by wrapping incoming indexes 2628 // to two vector index range [0, VLEN*2) and 2629 // MASK = WRAPED_INDEX < VLEN 2630 // 2631 // IR lowering prevents intrinsification failure and associated argument 2632 // boxing penalties. 2633 // 2634 2635 BasicType shuffle_bt = elem_bt; 2636 if (shuffle_bt == T_FLOAT) { 2637 shuffle_bt = T_INT; 2638 } else if (shuffle_bt == T_DOUBLE) { 2639 shuffle_bt = T_LONG; 2640 } 2641 const TypeVect* st = TypeVect::make(shuffle_bt, num_elem); 2642 2643 // Cast index vector to the corresponding bit type 2644 if (elem_bt != shuffle_bt) { 2645 int cast_vopc = VectorCastNode::opcode(0, elem_bt, true); 2646 index_vec = phase.transform(VectorCastNode::make(cast_vopc, index_vec, shuffle_bt, num_elem)); 2647 } 2648 2649 // Wrap indexes into two vector index range [0, VLEN * 2) 2650 Node* two_vect_lane_cnt_m1 = phase.makecon(TypeInteger::make(2 * num_elem - 1, 2 * num_elem - 1, Type::WidenMin, shuffle_bt == T_LONG ? T_LONG : T_INT)); 2651 Node* bcast_two_vect_lane_cnt_m1_vec = phase.transform(VectorNode::scalar2vector(two_vect_lane_cnt_m1, num_elem, 2652 shuffle_bt, false)); 2653 index_vec = phase.transform(VectorNode::make(Op_AndV, index_vec, bcast_two_vect_lane_cnt_m1_vec, st)); 2654 2655 // Compute the blend mask for merging two independently permitted vectors 2656 // using shuffle index in two vector index range [0, VLEN * 2). 2657 BoolTest::mask pred = BoolTest::le; 2658 ConINode* pred_node = phase.makecon(TypeInt::make(pred))->as_ConI(); 2659 const TypeVect* vmask_type = TypeVect::makemask(shuffle_bt, num_elem); 2660 Node* lane_cnt_m1 = phase.makecon(TypeInteger::make(num_elem - 1, num_elem - 1, Type::WidenMin, shuffle_bt == T_LONG ? T_LONG : T_INT)); 2661 Node* bcast_lane_cnt_m1_vec = phase.transform(VectorNode::scalar2vector(lane_cnt_m1, num_elem, shuffle_bt, false)); 2662 Node* mask = phase.transform(new VectorMaskCmpNode(pred, index_vec, bcast_lane_cnt_m1_vec, pred_node, vmask_type)); 2663 2664 // Rearrange expects the indexes to lie within single vector index range [0, VLEN). 2665 Node* wrapped_index_vec = phase.transform(VectorNode::make(Op_AndV, index_vec, bcast_lane_cnt_m1_vec, st)); 2666 2667 // Load indexes from byte vector and appropriately transform them to target 2668 // specific permutation index format. 2669 if (Matcher::vector_rearrange_requires_load_shuffle(shuffle_bt, num_elem)) { 2670 wrapped_index_vec = phase.transform(new VectorLoadShuffleNode(wrapped_index_vec, st)); 2671 } 2672 2673 vmask_type = TypeVect::makemask(elem_bt, num_elem); 2674 mask = phase.transform(new VectorMaskCastNode(mask, vmask_type)); 2675 2676 Node* p1 = phase.transform(new VectorRearrangeNode(src1, wrapped_index_vec)); 2677 Node* p2 = phase.transform(new VectorRearrangeNode(src2, wrapped_index_vec)); 2678 2679 return new VectorBlendNode(p2, p1, mask); 2680 } 2681 2682 // public static 2683 // <V extends Vector<E>, 2684 // E> 2685 // V selectFromTwoVectorOp(Class<? extends V> vClass, Class<E> eClass, int length, 2686 // V v1, V v2, V v3, 2687 // SelectFromTwoVector<V> defaultImpl) 2688 bool LibraryCallKit::inline_vector_select_from_two_vectors() { 2689 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 2690 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2691 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2692 2693 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || vector_klass->const_oop() == nullptr || 2694 elem_klass->const_oop() == nullptr ||!vlen->is_con()) { 2695 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s", 2696 NodeClassNames[argument(0)->Opcode()], 2697 NodeClassNames[argument(1)->Opcode()], 2698 NodeClassNames[argument(2)->Opcode()]); 2699 return false; // not enough info for intrinsification 2700 } 2701 2702 if (!is_klass_initialized(vector_klass)) { 2703 log_if_needed(" ** klass argument not initialized"); 2704 return false; 2705 } 2706 2707 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2708 if (!elem_type->is_primitive_type()) { 2709 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2710 return false; // should be primitive type 2711 } 2712 2713 int num_elem = vlen->get_con(); 2714 if (!is_power_of_2(num_elem)) { 2715 log_if_needed(" ** vlen is not power of two=%d", num_elem); 2716 return false; 2717 } 2718 2719 BasicType elem_bt = elem_type->basic_type(); 2720 BasicType index_elem_bt = elem_bt; 2721 if (elem_bt == T_FLOAT) { 2722 index_elem_bt = T_INT; 2723 } else if (elem_bt == T_DOUBLE) { 2724 index_elem_bt = T_LONG; 2725 } 2726 2727 bool lowerSelectFromOp = false; 2728 if (!arch_supports_vector(Op_SelectFromTwoVector, num_elem, elem_bt, VecMaskNotUsed)) { 2729 int cast_vopc = VectorCastNode::opcode(-1, elem_bt, true); 2730 if ((elem_bt != index_elem_bt && !arch_supports_vector(cast_vopc, num_elem, index_elem_bt, VecMaskNotUsed)) || 2731 !arch_supports_vector(Op_VectorMaskCmp, num_elem, index_elem_bt, VecMaskNotUsed) || 2732 !arch_supports_vector(Op_AndV, num_elem, index_elem_bt, VecMaskNotUsed) || 2733 !arch_supports_vector(Op_VectorMaskCast, num_elem, elem_bt, VecMaskNotUsed) || 2734 !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad) || 2735 !arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, VecMaskNotUsed) || 2736 !arch_supports_vector(Op_VectorLoadShuffle, num_elem, index_elem_bt, VecMaskNotUsed) || 2737 !arch_supports_vector(Op_Replicate, num_elem, index_elem_bt, VecMaskNotUsed)) { 2738 log_if_needed(" ** not supported: opc=%d vlen=%d etype=%s ismask=useload", 2739 Op_SelectFromTwoVector, num_elem, type2name(elem_bt)); 2740 return false; // not supported 2741 } 2742 lowerSelectFromOp = true; 2743 } 2744 2745 int cast_vopc = VectorCastNode::opcode(-1, elem_bt, true); 2746 if (!lowerSelectFromOp) { 2747 if (!arch_supports_vector(Op_AndV, num_elem, index_elem_bt, VecMaskNotUsed) || 2748 !arch_supports_vector(Op_Replicate, num_elem, index_elem_bt, VecMaskNotUsed) || 2749 (is_floating_point_type(elem_bt) && 2750 !arch_supports_vector(cast_vopc, num_elem, index_elem_bt, VecMaskNotUsed))) { 2751 log_if_needed(" ** index wrapping not supported: vlen=%d etype=%s" , 2752 num_elem, type2name(elem_bt)); 2753 return false; // not supported 2754 } 2755 } 2756 2757 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2758 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2759 2760 Node* opd1 = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2761 if (opd1 == nullptr) { 2762 log_if_needed(" ** unbox failed v1=%s", 2763 NodeClassNames[argument(3)->Opcode()]); 2764 return false; 2765 } 2766 Node* opd2 = unbox_vector(argument(4), vbox_type, elem_bt, num_elem); 2767 if (opd2 == nullptr) { 2768 log_if_needed(" ** unbox failed v2=%s", 2769 NodeClassNames[argument(4)->Opcode()]); 2770 return false; 2771 } 2772 Node* opd3 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 2773 if (opd3 == nullptr) { 2774 log_if_needed(" ** unbox failed v3=%s", 2775 NodeClassNames[argument(5)->Opcode()]); 2776 return false; 2777 } 2778 2779 const TypeVect* vt = TypeVect::make(elem_bt, num_elem); 2780 2781 Node* operation = nullptr; 2782 if (lowerSelectFromOp) { 2783 operation = gvn().transform(LowerSelectFromTwoVectorOperation(gvn(), opd1, opd2, opd3, vt)); 2784 } else { 2785 if (index_elem_bt != elem_bt) { 2786 opd1 = gvn().transform(VectorCastNode::make(cast_vopc, opd1, index_elem_bt, num_elem)); 2787 } 2788 int indexRangeMask = 2 * num_elem - 1; 2789 Node* wrap_mask = gvn().makecon(TypeInteger::make(indexRangeMask, indexRangeMask, Type::WidenMin, index_elem_bt != T_LONG ? T_INT : index_elem_bt)); 2790 Node* wrap_mask_vec = gvn().transform(VectorNode::scalar2vector(wrap_mask, num_elem, index_elem_bt, false)); 2791 opd1 = gvn().transform(VectorNode::make(Op_AndV, opd1, wrap_mask_vec, opd1->bottom_type()->is_vect())); 2792 operation = gvn().transform(VectorNode::make(Op_SelectFromTwoVector, opd1, opd2, opd3, vt)); 2793 } 2794 2795 // Wrap it up in VectorBox to keep object type information. 2796 Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem); 2797 set_result(vbox); 2798 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2799 return true; 2800 } 2801 2802 // public static 2803 // <V extends Vector<E>, 2804 // M extends VectorMask<E>, 2805 // E> 2806 // V compressExpandOp(int opr, 2807 // Class<? extends V> vClass, Class<? extends M> mClass, Class<E> eClass, 2808 // int length, V v, M m, 2809 // CompressExpandOperation<V, M> defaultImpl) 2810 bool LibraryCallKit::inline_vector_compress_expand() { 2811 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 2812 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 2813 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 2814 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 2815 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 2816 2817 if (vector_klass == nullptr || elem_klass == nullptr || mask_klass == nullptr || vlen == nullptr || 2818 vector_klass->const_oop() == nullptr || mask_klass->const_oop() == nullptr || 2819 elem_klass->const_oop() == nullptr || !vlen->is_con() || !opr->is_con()) { 2820 log_if_needed(" ** missing constant: opr=%s vclass=%s mclass=%s etype=%s vlen=%s", 2821 NodeClassNames[argument(0)->Opcode()], 2822 NodeClassNames[argument(1)->Opcode()], 2823 NodeClassNames[argument(2)->Opcode()], 2824 NodeClassNames[argument(3)->Opcode()], 2825 NodeClassNames[argument(4)->Opcode()]); 2826 return false; // not enough info for intrinsification 2827 } 2828 2829 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(mask_klass)) { 2830 log_if_needed(" ** klass argument not initialized"); 2831 return false; 2832 } 2833 2834 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2835 if (!elem_type->is_primitive_type()) { 2836 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2837 return false; // should be primitive type 2838 } 2839 2840 int num_elem = vlen->get_con(); 2841 BasicType elem_bt = elem_type->basic_type(); 2842 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 2843 2844 if (!arch_supports_vector(opc, num_elem, elem_bt, VecMaskUseLoad)) { 2845 log_if_needed(" ** not supported: opc=%d vlen=%d etype=%s ismask=useload", 2846 opc, num_elem, type2name(elem_bt)); 2847 return false; // not supported 2848 } 2849 2850 Node* opd1 = nullptr; 2851 const TypeInstPtr* vbox_type = nullptr; 2852 if (opc != Op_CompressM) { 2853 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2854 vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2855 opd1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 2856 if (opd1 == nullptr) { 2857 log_if_needed(" ** unbox failed vector=%s", 2858 NodeClassNames[argument(5)->Opcode()]); 2859 return false; 2860 } 2861 } 2862 2863 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2864 assert(is_vector_mask(mbox_klass), "argument(6) should be a mask class"); 2865 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 2866 2867 Node* mask = unbox_vector(argument(6), mbox_type, elem_bt, num_elem); 2868 if (mask == nullptr) { 2869 log_if_needed(" ** unbox failed mask=%s", 2870 NodeClassNames[argument(6)->Opcode()]); 2871 return false; 2872 } 2873 2874 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, opc == Op_CompressM); 2875 Node* operation = gvn().transform(VectorNode::make(opc, opd1, mask, vt)); 2876 2877 // Wrap it up in VectorBox to keep object type information. 2878 const TypeInstPtr* box_type = opc == Op_CompressM ? mbox_type : vbox_type; 2879 Node* vbox = box_vector(operation, box_type, elem_bt, num_elem); 2880 set_result(vbox); 2881 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2882 return true; 2883 } 2884 2885 // public static 2886 // <V extends Vector<E>, 2887 // E, 2888 // S extends VectorSpecies<E>> 2889 // V indexVector(Class<? extends V> vClass, Class<E> eClass, 2890 // int length, 2891 // V v, int step, S s, 2892 // IndexOperation<V, S> defaultImpl) 2893 bool LibraryCallKit::inline_index_vector() { 2894 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 2895 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2896 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2897 2898 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 2899 vector_klass->const_oop() == nullptr || !vlen->is_con() || 2900 elem_klass->const_oop() == nullptr) { 2901 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s", 2902 NodeClassNames[argument(0)->Opcode()], 2903 NodeClassNames[argument(1)->Opcode()], 2904 NodeClassNames[argument(2)->Opcode()]); 2905 return false; // not enough info for intrinsification 2906 } 2907 2908 if (!is_klass_initialized(vector_klass)) { 2909 log_if_needed(" ** klass argument not initialized"); 2910 return false; 2911 } 2912 2913 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2914 if (!elem_type->is_primitive_type()) { 2915 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2916 return false; // should be primitive type 2917 } 2918 2919 int num_elem = vlen->get_con(); 2920 BasicType elem_bt = elem_type->basic_type(); 2921 2922 // Check whether the iota index generation op is supported by the current hardware 2923 if (!arch_supports_vector(Op_VectorLoadConst, num_elem, elem_bt, VecMaskNotUsed)) { 2924 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 2925 return false; // not supported 2926 } 2927 2928 int mul_op = VectorSupport::vop2ideal(VectorSupport::VECTOR_OP_MUL, elem_bt); 2929 int vmul_op = VectorNode::opcode(mul_op, elem_bt); 2930 bool needs_mul = true; 2931 Node* scale = argument(4); 2932 const TypeInt* scale_type = gvn().type(scale)->isa_int(); 2933 // Multiply is not needed if the scale is a constant "1". 2934 if (scale_type && scale_type->is_con() && scale_type->get_con() == 1) { 2935 needs_mul = false; 2936 } else { 2937 // Check whether the vector multiply op is supported by the current hardware 2938 if (!arch_supports_vector(vmul_op, num_elem, elem_bt, VecMaskNotUsed)) { 2939 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 2940 return false; // not supported 2941 } 2942 2943 // Check whether the scalar cast op is supported by the current hardware 2944 if (is_floating_point_type(elem_bt) || elem_bt == T_LONG) { 2945 int cast_op = elem_bt == T_LONG ? Op_ConvI2L : 2946 elem_bt == T_FLOAT? Op_ConvI2F : Op_ConvI2D; 2947 if (!Matcher::match_rule_supported(cast_op)) { 2948 log_if_needed(" ** Rejected op (%s) because architecture does not support it", 2949 NodeClassNames[cast_op]); 2950 return false; // not supported 2951 } 2952 } 2953 } 2954 2955 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2956 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2957 Node* opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2958 if (opd == nullptr) { 2959 log_if_needed(" ** unbox failed vector=%s", 2960 NodeClassNames[argument(3)->Opcode()]); 2961 return false; 2962 } 2963 2964 int add_op = VectorSupport::vop2ideal(VectorSupport::VECTOR_OP_ADD, elem_bt); 2965 int vadd_op = VectorNode::opcode(add_op, elem_bt); 2966 bool needs_add = true; 2967 // The addition is not needed if all the element values of "opd" are zero 2968 if (VectorNode::is_all_zeros_vector(opd)) { 2969 needs_add = false; 2970 } else { 2971 // Check whether the vector addition op is supported by the current hardware 2972 if (!arch_supports_vector(vadd_op, num_elem, elem_bt, VecMaskNotUsed)) { 2973 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 2974 return false; // not supported 2975 } 2976 } 2977 2978 // Compute the iota indice vector 2979 const TypeVect* vt = TypeVect::make(elem_bt, num_elem); 2980 Node* index = gvn().transform(new VectorLoadConstNode(gvn().makecon(TypeInt::ZERO), vt)); 2981 2982 // Broadcast the "scale" to a vector, and multiply the "scale" with iota indice vector. 2983 if (needs_mul) { 2984 switch (elem_bt) { 2985 case T_BOOLEAN: // fall-through 2986 case T_BYTE: // fall-through 2987 case T_SHORT: // fall-through 2988 case T_CHAR: // fall-through 2989 case T_INT: { 2990 // no conversion needed 2991 break; 2992 } 2993 case T_LONG: { 2994 scale = gvn().transform(new ConvI2LNode(scale)); 2995 break; 2996 } 2997 case T_FLOAT: { 2998 scale = gvn().transform(new ConvI2FNode(scale)); 2999 break; 3000 } 3001 case T_DOUBLE: { 3002 scale = gvn().transform(new ConvI2DNode(scale)); 3003 break; 3004 } 3005 default: fatal("%s", type2name(elem_bt)); 3006 } 3007 scale = gvn().transform(VectorNode::scalar2vector(scale, num_elem, elem_bt)); 3008 index = gvn().transform(VectorNode::make(vmul_op, index, scale, vt)); 3009 } 3010 3011 // Add "opd" if addition is needed. 3012 if (needs_add) { 3013 index = gvn().transform(VectorNode::make(vadd_op, opd, index, vt)); 3014 } 3015 Node* vbox = box_vector(index, vbox_type, elem_bt, num_elem); 3016 set_result(vbox); 3017 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 3018 return true; 3019 } 3020 3021 // public static 3022 // <E, 3023 // M extends VectorMask<E>> 3024 // M indexPartiallyInUpperRange(Class<? extends M> mClass, Class<E> eClass, int length, 3025 // long offset, long limit, 3026 // IndexPartiallyInUpperRangeOperation<E, M> defaultImpl) 3027 bool LibraryCallKit::inline_index_partially_in_upper_range() { 3028 const TypeInstPtr* mask_klass = gvn().type(argument(0))->isa_instptr(); 3029 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 3030 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 3031 3032 if (mask_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 3033 mask_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 3034 log_if_needed(" ** missing constant: mclass=%s etype=%s vlen=%s", 3035 NodeClassNames[argument(0)->Opcode()], 3036 NodeClassNames[argument(1)->Opcode()], 3037 NodeClassNames[argument(2)->Opcode()]); 3038 return false; // not enough info for intrinsification 3039 } 3040 3041 if (!is_klass_initialized(mask_klass)) { 3042 log_if_needed(" ** klass argument not initialized"); 3043 return false; 3044 } 3045 3046 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 3047 if (!elem_type->is_primitive_type()) { 3048 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 3049 return false; // should be primitive type 3050 } 3051 3052 int num_elem = vlen->get_con(); 3053 BasicType elem_bt = elem_type->basic_type(); 3054 3055 // Check whether the necessary ops are supported by current hardware. 3056 bool supports_mask_gen = arch_supports_vector(Op_VectorMaskGen, num_elem, elem_bt, VecMaskUseStore); 3057 if (!supports_mask_gen) { 3058 if (!arch_supports_vector(Op_VectorLoadConst, num_elem, elem_bt, VecMaskNotUsed) || 3059 !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed) || 3060 !arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskUseStore)) { 3061 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 3062 return false; // not supported 3063 } 3064 3065 // Check whether the scalar cast operation is supported by current hardware. 3066 if (elem_bt != T_LONG) { 3067 int cast_op = is_integral_type(elem_bt) ? Op_ConvL2I 3068 : (elem_bt == T_FLOAT ? Op_ConvL2F : Op_ConvL2D); 3069 if (!Matcher::match_rule_supported(cast_op)) { 3070 log_if_needed(" ** Rejected op (%s) because architecture does not support it", 3071 NodeClassNames[cast_op]); 3072 return false; // not supported 3073 } 3074 } 3075 } 3076 3077 Node* offset = argument(3); 3078 Node* limit = argument(5); 3079 if (offset == nullptr || limit == nullptr) { 3080 log_if_needed(" ** offset or limit argument is null"); 3081 return false; // not supported 3082 } 3083 3084 ciKlass* box_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 3085 assert(is_vector_mask(box_klass), "argument(0) should be a mask class"); 3086 const TypeInstPtr* box_type = TypeInstPtr::make_exact(TypePtr::NotNull, box_klass); 3087 3088 // We assume "offset > 0 && limit >= offset && limit - offset < num_elem". 3089 // So directly get indexLimit with "indexLimit = limit - offset". 3090 Node* indexLimit = gvn().transform(new SubLNode(limit, offset)); 3091 Node* mask = nullptr; 3092 if (supports_mask_gen) { 3093 mask = gvn().transform(VectorMaskGenNode::make(indexLimit, elem_bt, num_elem)); 3094 } else { 3095 // Generate the vector mask based on "mask = iota < indexLimit". 3096 // Broadcast "indexLimit" to a vector. 3097 switch (elem_bt) { 3098 case T_BOOLEAN: // fall-through 3099 case T_BYTE: // fall-through 3100 case T_SHORT: // fall-through 3101 case T_CHAR: // fall-through 3102 case T_INT: { 3103 indexLimit = gvn().transform(new ConvL2INode(indexLimit)); 3104 break; 3105 } 3106 case T_DOUBLE: { 3107 indexLimit = gvn().transform(new ConvL2DNode(indexLimit)); 3108 break; 3109 } 3110 case T_FLOAT: { 3111 indexLimit = gvn().transform(new ConvL2FNode(indexLimit)); 3112 break; 3113 } 3114 case T_LONG: { 3115 // no conversion needed 3116 break; 3117 } 3118 default: fatal("%s", type2name(elem_bt)); 3119 } 3120 indexLimit = gvn().transform(VectorNode::scalar2vector(indexLimit, num_elem, elem_bt)); 3121 3122 // Load the "iota" vector. 3123 const TypeVect* vt = TypeVect::make(elem_bt, num_elem); 3124 Node* iota = gvn().transform(new VectorLoadConstNode(gvn().makecon(TypeInt::ZERO), vt)); 3125 3126 // Compute the vector mask with "mask = iota < indexLimit". 3127 ConINode* pred_node = (ConINode*)gvn().makecon(TypeInt::make(BoolTest::lt)); 3128 const TypeVect* vmask_type = TypeVect::makemask(elem_bt, num_elem); 3129 mask = gvn().transform(new VectorMaskCmpNode(BoolTest::lt, iota, indexLimit, pred_node, vmask_type)); 3130 } 3131 Node* vbox = box_vector(mask, box_type, elem_bt, num_elem); 3132 set_result(vbox); 3133 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 3134 return true; 3135 } 3136 3137 #undef non_product_log_if_needed 3138 #undef log_if_needed