1 /* 2 * Copyright (c) 2020, 2025, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "ci/ciSymbols.hpp" 26 #include "classfile/vmSymbols.hpp" 27 #include "opto/library_call.hpp" 28 #include "opto/runtime.hpp" 29 #include "opto/vectornode.hpp" 30 #include "prims/vectorSupport.hpp" 31 #include "runtime/stubRoutines.hpp" 32 33 #ifdef ASSERT 34 static bool is_vector(ciKlass* klass) { 35 return klass->is_subclass_of(ciEnv::current()->vector_VectorPayload_klass()); 36 } 37 38 static bool check_vbox(const TypeInstPtr* vbox_type) { 39 assert(vbox_type->klass_is_exact(), ""); 40 41 ciInstanceKlass* ik = vbox_type->instance_klass(); 42 assert(is_vector(ik), "not a vector"); 43 44 ciField* fd1 = ik->get_field_by_name(ciSymbols::ETYPE_name(), ciSymbols::class_signature(), /* is_static */ true); 45 assert(fd1 != nullptr, "element type info is missing"); 46 47 ciConstant val1 = fd1->constant_value(); 48 BasicType elem_bt = val1.as_object()->as_instance()->java_mirror_type()->basic_type(); 49 assert(is_java_primitive(elem_bt), "element type info is missing"); 50 51 ciField* fd2 = ik->get_field_by_name(ciSymbols::VLENGTH_name(), ciSymbols::int_signature(), /* is_static */ true); 52 assert(fd2 != nullptr, "vector length info is missing"); 53 54 ciConstant val2 = fd2->constant_value(); 55 assert(val2.as_int() > 0, "vector length info is missing"); 56 57 return true; 58 } 59 #endif 60 61 #define log_if_needed(...) \ 62 if (C->print_intrinsics()) { \ 63 tty->print_cr(__VA_ARGS__); \ 64 } 65 66 #ifndef PRODUCT 67 #define non_product_log_if_needed(...) log_if_needed(__VA_ARGS__) 68 #else 69 #define non_product_log_if_needed(...) 70 #endif 71 72 static bool is_vector_mask(ciKlass* klass) { 73 return klass->is_subclass_of(ciEnv::current()->vector_VectorMask_klass()); 74 } 75 76 bool LibraryCallKit::arch_supports_vector_rotate(int opc, int num_elem, BasicType elem_bt, 77 VectorMaskUseType mask_use_type, bool has_scalar_args) { 78 bool is_supported = true; 79 80 // has_scalar_args flag is true only for non-constant scalar shift count, 81 // since in this case shift needs to be broadcasted. 82 if (!Matcher::match_rule_supported_vector(opc, num_elem, elem_bt) || 83 (has_scalar_args && !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed))) { 84 is_supported = false; 85 } 86 87 if (is_supported) { 88 // Check if mask unboxing is supported, this is a two step process which first loads the contents 89 // of boolean array into vector followed by either lane expansion to match the lane size of masked 90 // vector operation or populate the predicate register. 91 if ((mask_use_type & VecMaskUseLoad) != 0) { 92 if (!Matcher::match_rule_supported_vector(Op_VectorLoadMask, num_elem, elem_bt) || 93 !Matcher::match_rule_supported_vector(Op_LoadVector, num_elem, T_BOOLEAN)) { 94 non_product_log_if_needed(" ** Rejected vector mask loading (%s,%s,%d) because architecture does not support it", 95 NodeClassNames[Op_VectorLoadMask], type2name(elem_bt), num_elem); 96 return false; 97 } 98 } 99 100 if ((mask_use_type & VecMaskUsePred) != 0) { 101 if (!Matcher::has_predicated_vectors() || 102 !Matcher::match_rule_supported_vector_masked(opc, num_elem, elem_bt)) { 103 non_product_log_if_needed("Rejected vector mask predicate using (%s,%s,%d) because architecture does not support it", 104 NodeClassNames[opc], type2name(elem_bt), num_elem); 105 return false; 106 } 107 } 108 } 109 110 int lshiftopc, rshiftopc; 111 switch(elem_bt) { 112 case T_BYTE: 113 lshiftopc = Op_LShiftI; 114 rshiftopc = Op_URShiftB; 115 break; 116 case T_SHORT: 117 lshiftopc = Op_LShiftI; 118 rshiftopc = Op_URShiftS; 119 break; 120 case T_INT: 121 lshiftopc = Op_LShiftI; 122 rshiftopc = Op_URShiftI; 123 break; 124 case T_LONG: 125 lshiftopc = Op_LShiftL; 126 rshiftopc = Op_URShiftL; 127 break; 128 default: fatal("Unexpected type: %s", type2name(elem_bt)); 129 } 130 int lshiftvopc = VectorNode::opcode(lshiftopc, elem_bt); 131 int rshiftvopc = VectorNode::opcode(rshiftopc, elem_bt); 132 if (!is_supported && 133 arch_supports_vector(lshiftvopc, num_elem, elem_bt, VecMaskNotUsed, has_scalar_args) && 134 arch_supports_vector(rshiftvopc, num_elem, elem_bt, VecMaskNotUsed, has_scalar_args) && 135 arch_supports_vector(Op_OrV, num_elem, elem_bt, VecMaskNotUsed)) { 136 is_supported = true; 137 } 138 return is_supported; 139 } 140 141 Node* GraphKit::box_vector(Node* vector, const TypeInstPtr* vbox_type, BasicType elem_bt, int num_elem, bool deoptimize_on_exception) { 142 assert(EnableVectorSupport, ""); 143 144 PreserveReexecuteState preexecs(this); 145 jvms()->set_should_reexecute(true); 146 147 VectorBoxAllocateNode* alloc = new VectorBoxAllocateNode(C, vbox_type); 148 set_edges_for_java_call(alloc, /*must_throw=*/false, /*separate_io_proj=*/true); 149 make_slow_call_ex(alloc, env()->Throwable_klass(), /*separate_io_proj=*/true, deoptimize_on_exception); 150 set_i_o(gvn().transform( new ProjNode(alloc, TypeFunc::I_O) )); 151 set_all_memory(gvn().transform( new ProjNode(alloc, TypeFunc::Memory) )); 152 Node* ret = gvn().transform(new ProjNode(alloc, TypeFunc::Parms)); 153 154 assert(check_vbox(vbox_type), ""); 155 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, is_vector_mask(vbox_type->instance_klass())); 156 VectorBoxNode* vbox = new VectorBoxNode(C, ret, vector, vbox_type, vt); 157 return gvn().transform(vbox); 158 } 159 160 Node* GraphKit::unbox_vector(Node* v, const TypeInstPtr* vbox_type, BasicType elem_bt, int num_elem) { 161 assert(EnableVectorSupport, ""); 162 const TypeInstPtr* vbox_type_v = gvn().type(v)->isa_instptr(); 163 if (vbox_type_v == nullptr || vbox_type->instance_klass() != vbox_type_v->instance_klass()) { 164 return nullptr; // arguments don't agree on vector shapes 165 } 166 if (vbox_type_v->maybe_null()) { 167 return nullptr; // no nulls are allowed 168 } 169 assert(check_vbox(vbox_type), ""); 170 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, is_vector_mask(vbox_type->instance_klass())); 171 Node* unbox = gvn().transform(new VectorUnboxNode(C, vt, v, merged_memory())); 172 return unbox; 173 } 174 175 Node* GraphKit::vector_shift_count(Node* cnt, int shift_op, BasicType bt, int num_elem) { 176 assert(bt == T_INT || bt == T_LONG || bt == T_SHORT || bt == T_BYTE, "byte, short, long and int are supported"); 177 juint mask = (type2aelembytes(bt) * BitsPerByte - 1); 178 Node* nmask = gvn().transform(ConNode::make(TypeInt::make(mask))); 179 Node* mcnt = gvn().transform(new AndINode(cnt, nmask)); 180 return gvn().transform(VectorNode::shift_count(shift_op, mcnt, num_elem, bt)); 181 } 182 183 bool LibraryCallKit::arch_supports_vector(int sopc, int num_elem, BasicType type, VectorMaskUseType mask_use_type, bool has_scalar_args) { 184 // Check that the operation is valid. 185 if (sopc <= 0) { 186 non_product_log_if_needed(" ** Rejected intrinsification because no valid vector op could be extracted"); 187 return false; 188 } 189 190 if (VectorNode::is_vector_rotate(sopc)) { 191 if(!arch_supports_vector_rotate(sopc, num_elem, type, mask_use_type, has_scalar_args)) { 192 non_product_log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support variable vector shifts", 193 NodeClassNames[sopc], type2name(type), num_elem); 194 return false; 195 } 196 } else if (VectorNode::is_vector_integral_negate(sopc)) { 197 if (!VectorNode::is_vector_integral_negate_supported(sopc, num_elem, type, false)) { 198 non_product_log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support integral vector negate", 199 NodeClassNames[sopc], type2name(type), num_elem); 200 return false; 201 } 202 } else { 203 // Check that architecture supports this op-size-type combination. 204 if (!Matcher::match_rule_supported_vector(sopc, num_elem, type)) { 205 non_product_log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support it", 206 NodeClassNames[sopc], type2name(type), num_elem); 207 return false; 208 } else { 209 assert(Matcher::match_rule_supported(sopc), "must be supported"); 210 } 211 } 212 213 if (num_elem == 1) { 214 if (mask_use_type != VecMaskNotUsed) { 215 non_product_log_if_needed(" ** Rejected vector mask op (%s,%s,%d) because architecture does not support it", 216 NodeClassNames[sopc], type2name(type), num_elem); 217 return false; 218 } 219 220 if (sopc != 0) { 221 if (sopc != Op_LoadVector && sopc != Op_StoreVector) { 222 non_product_log_if_needed(" ** Not a svml call or load/store vector op (%s,%s,%d)", 223 NodeClassNames[sopc], type2name(type), num_elem); 224 return false; 225 } 226 } 227 } 228 229 if (!has_scalar_args && VectorNode::is_vector_shift(sopc) && 230 Matcher::supports_vector_variable_shifts() == false) { 231 log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support variable vector shifts", 232 NodeClassNames[sopc], type2name(type), num_elem); 233 return false; 234 } 235 236 // Check if mask unboxing is supported, this is a two step process which first loads the contents 237 // of boolean array into vector followed by either lane expansion to match the lane size of masked 238 // vector operation or populate the predicate register. 239 if ((mask_use_type & VecMaskUseLoad) != 0) { 240 if (!Matcher::match_rule_supported_vector(Op_VectorLoadMask, num_elem, type) || 241 !Matcher::match_rule_supported_vector(Op_LoadVector, num_elem, T_BOOLEAN)) { 242 non_product_log_if_needed(" ** Rejected vector mask loading (%s,%s,%d) because architecture does not support it", 243 NodeClassNames[Op_VectorLoadMask], type2name(type), num_elem); 244 return false; 245 } 246 } 247 248 // Check if mask boxing is supported, this is a two step process which first stores the contents 249 // of mask vector / predicate register into a boolean vector followed by vector store operation to 250 // transfer the contents to underlined storage of mask boxes which is a boolean array. 251 if ((mask_use_type & VecMaskUseStore) != 0) { 252 if (!Matcher::match_rule_supported_vector(Op_VectorStoreMask, num_elem, type) || 253 !Matcher::match_rule_supported_vector(Op_StoreVector, num_elem, T_BOOLEAN)) { 254 non_product_log_if_needed("Rejected vector mask storing (%s,%s,%d) because architecture does not support it", 255 NodeClassNames[Op_VectorStoreMask], type2name(type), num_elem); 256 return false; 257 } 258 } 259 260 if ((mask_use_type & VecMaskUsePred) != 0) { 261 bool is_supported = false; 262 if (Matcher::has_predicated_vectors()) { 263 if (VectorNode::is_vector_integral_negate(sopc)) { 264 is_supported = VectorNode::is_vector_integral_negate_supported(sopc, num_elem, type, true); 265 } else { 266 is_supported = Matcher::match_rule_supported_vector_masked(sopc, num_elem, type); 267 } 268 } 269 is_supported |= Matcher::supports_vector_predicate_op_emulation(sopc, num_elem, type); 270 271 if (!is_supported) { 272 non_product_log_if_needed("Rejected vector mask predicate using (%s,%s,%d) because architecture does not support it", 273 NodeClassNames[sopc], type2name(type), num_elem); 274 return false; 275 } 276 } 277 278 return true; 279 } 280 281 static bool is_klass_initialized(const TypeInstPtr* vec_klass) { 282 if (vec_klass->const_oop() == nullptr) { 283 return false; // uninitialized or some kind of unsafe access 284 } 285 assert(vec_klass->const_oop()->as_instance()->java_lang_Class_klass() != nullptr, "klass instance expected"); 286 ciInstanceKlass* klass = vec_klass->const_oop()->as_instance()->java_lang_Class_klass()->as_instance_klass(); 287 return klass->is_initialized(); 288 } 289 290 // public static 291 // <V extends Vector<E>, 292 // M extends VectorMask<E>, 293 // E> 294 // V unaryOp(int oprId, Class<? extends V> vmClass, Class<? extends M> maskClass, Class<E> elementType, 295 // int length, V v, M m, 296 // UnaryOperation<V, M> defaultImpl) 297 // 298 // public static 299 // <V, 300 // M extends VectorMask<E>, 301 // E> 302 // V binaryOp(int oprId, Class<? extends V> vmClass, Class<? extends M> maskClass, Class<E> elementType, 303 // int length, V v1, V v2, M m, 304 // BinaryOperation<V, M> defaultImpl) 305 // 306 // public static 307 // <V extends Vector<E>, 308 // M extends VectorMask<E>, 309 // E> 310 // V ternaryOp(int oprId, Class<? extends V> vmClass, Class<? extends M> maskClass, Class<E> elementType, 311 // int length, V v1, V v2, V v3, M m, 312 // TernaryOperation<V, M> defaultImpl) 313 // 314 bool LibraryCallKit::inline_vector_nary_operation(int n) { 315 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 316 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 317 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 318 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 319 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 320 321 if (opr == nullptr || !opr->is_con() || 322 vector_klass == nullptr || vector_klass->const_oop() == nullptr || 323 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 324 vlen == nullptr || !vlen->is_con()) { 325 log_if_needed(" ** missing constant: opr=%s vclass=%s etype=%s vlen=%s", 326 NodeClassNames[argument(0)->Opcode()], 327 NodeClassNames[argument(1)->Opcode()], 328 NodeClassNames[argument(3)->Opcode()], 329 NodeClassNames[argument(4)->Opcode()]); 330 return false; // not enough info for intrinsification 331 } 332 333 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 334 if (!elem_type->is_primitive_type()) { 335 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 336 return false; // should be primitive type 337 } 338 if (!is_klass_initialized(vector_klass)) { 339 log_if_needed(" ** klass argument not initialized"); 340 return false; 341 } 342 343 // "argument(n + 5)" should be the mask object. We assume it is "null" when no mask 344 // is used to control this operation. 345 const Type* vmask_type = gvn().type(argument(n + 5)); 346 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 347 if (is_masked_op) { 348 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 349 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(2)->Opcode()]); 350 return false; // not enough info for intrinsification 351 } 352 353 if (!is_klass_initialized(mask_klass)) { 354 log_if_needed(" ** mask klass argument not initialized"); 355 return false; 356 } 357 358 if (vmask_type->maybe_null()) { 359 log_if_needed(" ** null mask values are not allowed for masked op"); 360 return false; 361 } 362 } 363 364 BasicType elem_bt = elem_type->basic_type(); 365 bool has_scalar_op = VectorSupport::has_scalar_op(opr->get_con()); 366 bool is_unsigned = VectorSupport::is_unsigned_op(opr->get_con()); 367 368 int num_elem = vlen->get_con(); 369 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 370 int sopc = has_scalar_op ? VectorNode::opcode(opc, elem_bt) : opc; 371 if (sopc == 0 || num_elem == 1) { 372 log_if_needed(" ** operation not supported: arity=%d opc=%s[%d] vlen=%d etype=%s", 373 n, NodeClassNames[opc], opc, num_elem, type2name(elem_bt)); 374 return false; // operation not supported 375 } 376 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 377 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 378 379 if (is_vector_mask(vbox_klass)) { 380 assert(!is_masked_op, "mask operations do not need mask to control"); 381 } 382 383 // When using mask, mask use type needs to be VecMaskUseLoad. 384 VectorMaskUseType mask_use_type = is_vector_mask(vbox_klass) ? VecMaskUseAll 385 : is_masked_op ? VecMaskUseLoad : VecMaskNotUsed; 386 if (!arch_supports_vector(sopc, num_elem, elem_bt, mask_use_type)) { 387 log_if_needed(" ** not supported: arity=%d opc=%d vlen=%d etype=%s ismask=%d is_masked_op=%d", 388 n, sopc, num_elem, type2name(elem_bt), 389 is_vector_mask(vbox_klass) ? 1 : 0, is_masked_op ? 1 : 0); 390 return false; // not supported 391 } 392 393 // Return true if current platform has implemented the masked operation with predicate feature. 394 bool use_predicate = is_masked_op && arch_supports_vector(sopc, num_elem, elem_bt, VecMaskUsePred); 395 if (is_masked_op && !use_predicate && !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)) { 396 log_if_needed(" ** not supported: arity=%d opc=%d vlen=%d etype=%s ismask=0 is_masked_op=1", 397 n, sopc, num_elem, type2name(elem_bt)); 398 return false; 399 } 400 401 Node* opd1 = nullptr; Node* opd2 = nullptr; Node* opd3 = nullptr; 402 switch (n) { 403 case 3: { 404 opd3 = unbox_vector(argument(7), vbox_type, elem_bt, num_elem); 405 if (opd3 == nullptr) { 406 log_if_needed(" ** unbox failed v3=%s", 407 NodeClassNames[argument(7)->Opcode()]); 408 return false; 409 } 410 // fall-through 411 } 412 case 2: { 413 opd2 = unbox_vector(argument(6), vbox_type, elem_bt, num_elem); 414 if (opd2 == nullptr) { 415 log_if_needed(" ** unbox failed v2=%s", 416 NodeClassNames[argument(6)->Opcode()]); 417 return false; 418 } 419 // fall-through 420 } 421 case 1: { 422 opd1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 423 if (opd1 == nullptr) { 424 log_if_needed(" ** unbox failed v1=%s", 425 NodeClassNames[argument(5)->Opcode()]); 426 return false; 427 } 428 break; 429 } 430 default: fatal("unsupported arity: %d", n); 431 } 432 433 Node* mask = nullptr; 434 if (is_masked_op) { 435 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 436 assert(is_vector_mask(mbox_klass), "argument(2) should be a mask class"); 437 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 438 mask = unbox_vector(argument(n + 5), mbox_type, elem_bt, num_elem); 439 if (mask == nullptr) { 440 log_if_needed(" ** unbox failed mask=%s", 441 NodeClassNames[argument(n + 5)->Opcode()]); 442 return false; 443 } 444 } 445 446 Node* operation = nullptr; 447 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, is_vector_mask(vbox_klass)); 448 switch (n) { 449 case 1: 450 case 2: { 451 operation = VectorNode::make(sopc, opd1, opd2, vt, is_vector_mask(vbox_klass), VectorNode::is_shift_opcode(opc), is_unsigned); 452 break; 453 } 454 case 3: { 455 operation = VectorNode::make(sopc, opd1, opd2, opd3, vt); 456 break; 457 } 458 default: fatal("unsupported arity: %d", n); 459 } 460 461 if (is_masked_op && mask != nullptr) { 462 if (use_predicate) { 463 operation->add_req(mask); 464 operation->add_flag(Node::Flag_is_predicated_vector); 465 } else { 466 operation->add_flag(Node::Flag_is_predicated_using_blend); 467 operation = gvn().transform(operation); 468 operation = new VectorBlendNode(opd1, operation, mask); 469 } 470 } 471 operation = gvn().transform(operation); 472 473 // Wrap it up in VectorBox to keep object type information. 474 Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem); 475 set_result(vbox); 476 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 477 return true; 478 } 479 480 // public static 481 // <V extends Vector<E>, E> 482 // V libraryUnaryOp(long address, Class<? extends V> vClass, Class<E> elementType, int length, String debugName, 483 // V v, 484 // UnaryOperation<V, ?> defaultImpl) 485 // 486 // public static 487 // <V extends VectorPayload, E> 488 // V libraryBinaryOp(long address, Class<? extends V> vClass, Class<E> elementType, int length, String debugName, 489 // V v1, V v2, 490 // BinaryOperation<V, ?> defaultImpl) 491 bool LibraryCallKit::inline_vector_call(int arity) { 492 assert(Matcher::supports_vector_calling_convention(), "required"); 493 494 const TypeLong* entry = gvn().type(argument(0))->isa_long(); 495 const TypeInstPtr* vector_klass = gvn().type(argument(2))->isa_instptr(); 496 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 497 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 498 const TypeInstPtr* debug_name_oop = gvn().type(argument(5))->isa_instptr(); 499 500 if (entry == nullptr || !entry->is_con() || 501 vector_klass == nullptr || vector_klass->const_oop() == nullptr || 502 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 503 vlen == nullptr || !vlen->is_con() || 504 debug_name_oop == nullptr || debug_name_oop->const_oop() == nullptr) { 505 log_if_needed(" ** missing constant: opr=%s vclass=%s etype=%s vlen=%s debug_name=%s", 506 NodeClassNames[argument(0)->Opcode()], 507 NodeClassNames[argument(2)->Opcode()], 508 NodeClassNames[argument(3)->Opcode()], 509 NodeClassNames[argument(4)->Opcode()], 510 NodeClassNames[argument(5)->Opcode()]); 511 return false; // not enough info for intrinsification 512 } 513 514 if (entry->get_con() == 0) { 515 log_if_needed(" ** missing entry point"); 516 return false; 517 } 518 519 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 520 if (!elem_type->is_primitive_type()) { 521 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 522 return false; // should be primitive type 523 } 524 if (!is_klass_initialized(vector_klass)) { 525 log_if_needed(" ** klass argument not initialized"); 526 return false; 527 } 528 529 BasicType elem_bt = elem_type->basic_type(); 530 int num_elem = vlen->get_con(); 531 if (!Matcher::vector_size_supported(elem_bt, num_elem)) { 532 log_if_needed(" ** vector size (vlen=%d, etype=%s) is not supported", 533 num_elem, type2name(elem_bt)); 534 return false; 535 } 536 537 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 538 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 539 540 Node* opd1 = unbox_vector(argument(6), vbox_type, elem_bt, num_elem); 541 if (opd1 == nullptr) { 542 log_if_needed(" ** unbox failed v1=%s", NodeClassNames[argument(6)->Opcode()]); 543 return false; 544 } 545 546 Node* opd2 = nullptr; 547 if (arity > 1) { 548 opd2 = unbox_vector(argument(7), vbox_type, elem_bt, num_elem); 549 if (opd2 == nullptr) { 550 log_if_needed(" ** unbox failed v2=%s", NodeClassNames[argument(7)->Opcode()]); 551 return false; 552 } 553 } 554 assert(arity == 1 || arity == 2, "arity %d not supported", arity); 555 const TypeVect* vt = TypeVect::make(elem_bt, num_elem); 556 const TypeFunc* call_type = OptoRuntime::Math_Vector_Vector_Type(arity, vt, vt); 557 address entry_addr = (address)entry->get_con(); 558 559 const char* debug_name = "<unknown>"; 560 if (!debug_name_oop->const_oop()->is_null_object()) { 561 size_t buflen = 100; 562 char* buf = NEW_ARENA_ARRAY(C->comp_arena(), char, buflen); 563 debug_name = debug_name_oop->const_oop()->as_instance()->java_lang_String_str(buf, buflen); 564 } 565 Node* vcall = make_runtime_call(RC_VECTOR, 566 call_type, 567 entry_addr, 568 debug_name, 569 TypePtr::BOTTOM, 570 opd1, 571 opd2); 572 573 vcall = gvn().transform(new ProjNode(gvn().transform(vcall), TypeFunc::Parms)); 574 575 // Wrap it up in VectorBox to keep object type information. 576 Node* vbox = box_vector(vcall, vbox_type, elem_bt, num_elem); 577 set_result(vbox); 578 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 579 return true; 580 } 581 582 // <E, M> 583 // long maskReductionCoerced(int oper, Class<? extends M> maskClass, Class<?> elemClass, 584 // int length, M m, VectorMaskOp<M> defaultImpl) 585 bool LibraryCallKit::inline_vector_mask_operation() { 586 const TypeInt* oper = gvn().type(argument(0))->isa_int(); 587 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 588 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 589 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 590 Node* mask = argument(4); 591 592 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr || 593 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 594 vlen == nullptr || !vlen->is_con() || 595 oper == nullptr || !oper->is_con() || 596 mask->is_top()) { 597 return false; // dead code 598 } 599 600 if (!is_klass_initialized(mask_klass)) { 601 log_if_needed(" ** klass argument not initialized"); 602 return false; 603 } 604 605 int num_elem = vlen->get_con(); 606 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 607 BasicType elem_bt = elem_type->basic_type(); 608 609 int mopc = VectorSupport::vop2ideal(oper->get_con(), elem_bt); 610 if (!arch_supports_vector(mopc, num_elem, elem_bt, VecMaskUseLoad)) { 611 log_if_needed(" ** not supported: arity=1 op=cast#%d/3 vlen2=%d etype2=%s", 612 mopc, num_elem, type2name(elem_bt)); 613 return false; // not supported 614 } 615 616 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 617 const TypeInstPtr* mask_box_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 618 Node* mask_vec = unbox_vector(mask, mask_box_type, elem_bt, num_elem); 619 if (mask_vec == nullptr) { 620 log_if_needed(" ** unbox failed mask=%s", 621 NodeClassNames[argument(4)->Opcode()]); 622 return false; 623 } 624 625 if (mask_vec->bottom_type()->isa_vectmask() == nullptr) { 626 mask_vec = gvn().transform(VectorStoreMaskNode::make(gvn(), mask_vec, elem_bt, num_elem)); 627 } 628 const Type* maskoper_ty = mopc == Op_VectorMaskToLong ? (const Type*)TypeLong::LONG : (const Type*)TypeInt::INT; 629 Node* maskoper = gvn().transform(VectorMaskOpNode::make(mask_vec, maskoper_ty, mopc)); 630 if (mopc != Op_VectorMaskToLong) { 631 maskoper = ConvI2L(maskoper); 632 } 633 set_result(maskoper); 634 635 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 636 return true; 637 } 638 639 // public static 640 // <M, 641 // S extends VectorSpecies<E>, 642 // E> 643 // M fromBitsCoerced(Class<? extends M> vmClass, Class<E> elementType, int length, 644 // long bits, int mode, S s, 645 // BroadcastOperation<M, E, S> defaultImpl) 646 bool LibraryCallKit::inline_vector_frombits_coerced() { 647 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 648 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 649 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 650 const TypeLong* bits_type = gvn().type(argument(3))->isa_long(); 651 // Mode argument determines the mode of operation it can take following values:- 652 // MODE_BROADCAST for vector Vector.broadcast and VectorMask.maskAll operations. 653 // MODE_BITS_COERCED_LONG_TO_MASK for VectorMask.fromLong operation. 654 const TypeInt* mode = gvn().type(argument(5))->isa_int(); 655 656 if (vector_klass == nullptr || vector_klass->const_oop() == nullptr || 657 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 658 vlen == nullptr || !vlen->is_con() || 659 bits_type == nullptr || 660 mode == nullptr || !mode->is_con()) { 661 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s bitwise=%s", 662 NodeClassNames[argument(0)->Opcode()], 663 NodeClassNames[argument(1)->Opcode()], 664 NodeClassNames[argument(2)->Opcode()], 665 NodeClassNames[argument(5)->Opcode()]); 666 return false; // not enough info for intrinsification 667 } 668 669 if (!is_klass_initialized(vector_klass)) { 670 log_if_needed(" ** klass argument not initialized"); 671 return false; 672 } 673 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 674 if (!elem_type->is_primitive_type()) { 675 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 676 return false; // should be primitive type 677 } 678 BasicType elem_bt = elem_type->basic_type(); 679 int num_elem = vlen->get_con(); 680 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 681 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 682 683 bool is_mask = is_vector_mask(vbox_klass); 684 int bcast_mode = mode->get_con(); 685 VectorMaskUseType checkFlags = (VectorMaskUseType)(is_mask ? VecMaskUseAll : VecMaskNotUsed); 686 int opc = bcast_mode == VectorSupport::MODE_BITS_COERCED_LONG_TO_MASK ? Op_VectorLongToMask : Op_Replicate; 687 688 if (!arch_supports_vector(opc, num_elem, elem_bt, checkFlags, true /*has_scalar_args*/)) { 689 // If the input long sets or unsets all lanes and Replicate is supported, 690 // generate a MaskAll or Replicate instead. 691 692 // The "maskAll" API uses the corresponding integer types for floating-point data. 693 BasicType maskall_bt = elem_bt == T_DOUBLE ? T_LONG : (elem_bt == T_FLOAT ? T_INT: elem_bt); 694 if (!(opc == Op_VectorLongToMask && 695 VectorNode::is_maskall_type(bits_type, num_elem) && 696 arch_supports_vector(Op_Replicate, num_elem, maskall_bt, checkFlags, true /*has_scalar_args*/))) { 697 log_if_needed(" ** not supported: arity=0 op=broadcast vlen=%d etype=%s ismask=%d bcast_mode=%d", 698 num_elem, type2name(elem_bt), 699 is_mask ? 1 : 0, 700 bcast_mode); 701 return false; // not supported 702 } 703 } 704 705 Node* broadcast = nullptr; 706 Node* bits = argument(3); 707 Node* elem = bits; 708 709 if (opc == Op_VectorLongToMask) { 710 const TypeVect* vt = TypeVect::makemask(elem_bt, num_elem); 711 if (vt->isa_vectmask()) { 712 broadcast = gvn().transform(new VectorLongToMaskNode(elem, vt)); 713 } else { 714 const TypeVect* mvt = TypeVect::make(T_BOOLEAN, num_elem); 715 broadcast = gvn().transform(new VectorLongToMaskNode(elem, mvt)); 716 broadcast = gvn().transform(new VectorLoadMaskNode(broadcast, vt)); 717 } 718 } else { 719 switch (elem_bt) { 720 case T_BOOLEAN: // fall-through 721 case T_BYTE: // fall-through 722 case T_SHORT: // fall-through 723 case T_CHAR: // fall-through 724 case T_INT: { 725 elem = gvn().transform(new ConvL2INode(bits)); 726 break; 727 } 728 case T_DOUBLE: { 729 elem = gvn().transform(new MoveL2DNode(bits)); 730 break; 731 } 732 case T_FLOAT: { 733 bits = gvn().transform(new ConvL2INode(bits)); 734 elem = gvn().transform(new MoveI2FNode(bits)); 735 break; 736 } 737 case T_LONG: { 738 // no conversion needed 739 break; 740 } 741 default: fatal("%s", type2name(elem_bt)); 742 } 743 broadcast = VectorNode::scalar2vector(elem, num_elem, elem_bt, is_mask); 744 broadcast = gvn().transform(broadcast); 745 } 746 747 Node* box = box_vector(broadcast, vbox_type, elem_bt, num_elem); 748 set_result(box); 749 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 750 return true; 751 } 752 753 static bool elem_consistent_with_arr(BasicType elem_bt, const TypeAryPtr* arr_type, bool mismatched_ms) { 754 assert(arr_type != nullptr, "unexpected"); 755 BasicType arr_elem_bt = arr_type->elem()->array_element_basic_type(); 756 if (elem_bt == arr_elem_bt) { 757 return true; 758 } else if (elem_bt == T_SHORT && arr_elem_bt == T_CHAR) { 759 // Load/store of short vector from/to char[] is supported 760 return true; 761 } else if (elem_bt == T_BYTE && arr_elem_bt == T_BOOLEAN) { 762 // Load/store of byte vector from/to boolean[] is supported 763 return true; 764 } else { 765 return mismatched_ms; 766 } 767 } 768 769 // public static 770 // <C, 771 // VM extends VectorPayload, 772 // E, 773 // S extends VectorSpecies<E>> 774 // VM load(Class<? extends VM> vmClass, Class<E> eClass, 775 // int length, 776 // Object base, long offset, // Unsafe addressing 777 // boolean fromSegment, 778 // C container, long index, S s, // Arguments for default implementation 779 // LoadOperation<C, VM, S> defaultImpl) { 780 // public static 781 // <C, 782 // V extends VectorPayload> 783 // void store(Class<?> vClass, Class<?> eClass, 784 // int length, 785 // Object base, long offset, // Unsafe addressing 786 // boolean fromSegment, 787 // V v, C container, long index, // Arguments for default implementation 788 // StoreVectorOperation<C, V> defaultImpl) { 789 bool LibraryCallKit::inline_vector_mem_operation(bool is_store) { 790 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 791 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 792 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 793 const TypeInt* from_ms = gvn().type(argument(6))->isa_int(); 794 795 if (vector_klass == nullptr || vector_klass->const_oop() == nullptr || 796 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 797 vlen == nullptr || !vlen->is_con() || 798 from_ms == nullptr || !from_ms->is_con()) { 799 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s from_ms=%s", 800 NodeClassNames[argument(0)->Opcode()], 801 NodeClassNames[argument(1)->Opcode()], 802 NodeClassNames[argument(2)->Opcode()], 803 NodeClassNames[argument(6)->Opcode()]); 804 return false; // not enough info for intrinsification 805 } 806 if (!is_klass_initialized(vector_klass)) { 807 log_if_needed(" ** klass argument not initialized"); 808 return false; 809 } 810 811 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 812 if (!elem_type->is_primitive_type()) { 813 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 814 return false; // should be primitive type 815 } 816 BasicType elem_bt = elem_type->basic_type(); 817 int num_elem = vlen->get_con(); 818 819 // TODO When mask usage is supported, VecMaskNotUsed needs to be VecMaskUseLoad. 820 if (!arch_supports_vector(is_store ? Op_StoreVector : Op_LoadVector, num_elem, elem_bt, VecMaskNotUsed)) { 821 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s ismask=no", 822 is_store, is_store ? "store" : "load", 823 num_elem, type2name(elem_bt)); 824 return false; // not supported 825 } 826 827 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 828 bool is_mask = is_vector_mask(vbox_klass); 829 830 Node* base = argument(3); 831 Node* offset = ConvL2X(argument(4)); 832 833 // Save state and restore on bailout 834 SavedState old_state(this); 835 836 Node* addr = make_unsafe_address(base, offset, (is_mask ? T_BOOLEAN : elem_bt), true); 837 838 // The memory barrier checks are based on ones for unsafe access. 839 // This is not 1-1 implementation. 840 const Type *const base_type = gvn().type(base); 841 842 const TypePtr *addr_type = gvn().type(addr)->isa_ptr(); 843 const TypeAryPtr* arr_type = addr_type->isa_aryptr(); 844 845 const bool in_native = TypePtr::NULL_PTR == base_type; // base always null 846 const bool in_heap = !TypePtr::NULL_PTR->higher_equal(base_type); // base never null 847 848 const bool is_mixed_access = !in_heap && !in_native; 849 850 const bool is_mismatched_access = in_heap && (addr_type->isa_aryptr() == nullptr); 851 852 const bool needs_cpu_membar = is_mixed_access || is_mismatched_access; 853 854 // For non-masked mismatched memory segment vector read/write accesses, intrinsification can continue 855 // with unknown backing storage type and compiler can skip inserting explicit reinterpretation IR after 856 // loading from or before storing to backing storage which is mandatory for semantic correctness of 857 // big-endian memory layout. 858 bool mismatched_ms = LITTLE_ENDIAN_ONLY(false) 859 BIG_ENDIAN_ONLY(from_ms->get_con() && !is_mask && arr_type != nullptr && 860 arr_type->elem()->array_element_basic_type() != elem_bt); 861 BasicType mem_elem_bt = mismatched_ms ? arr_type->elem()->array_element_basic_type() : elem_bt; 862 if (!is_java_primitive(mem_elem_bt)) { 863 log_if_needed(" ** non-primitive array element type"); 864 return false; 865 } 866 int mem_num_elem = mismatched_ms ? (num_elem * type2aelembytes(elem_bt)) / type2aelembytes(mem_elem_bt) : num_elem; 867 if (arr_type != nullptr && !is_mask && !elem_consistent_with_arr(elem_bt, arr_type, mismatched_ms)) { 868 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s atype=%s ismask=no", 869 is_store, is_store ? "store" : "load", 870 num_elem, type2name(elem_bt), type2name(arr_type->elem()->array_element_basic_type())); 871 return false; 872 } 873 874 // In case of mismatched memory segment accesses, we need to double check that the source type memory operations are supported by backend. 875 if (mismatched_ms) { 876 if (is_store) { 877 if (!arch_supports_vector(Op_StoreVector, num_elem, elem_bt, VecMaskNotUsed) 878 || !arch_supports_vector(Op_VectorReinterpret, mem_num_elem, mem_elem_bt, VecMaskNotUsed)) { 879 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d*8 etype=%s/8 ismask=no", 880 is_store, "store", 881 num_elem, type2name(elem_bt)); 882 return false; // not supported 883 } 884 } else { 885 if (!arch_supports_vector(Op_LoadVector, mem_num_elem, mem_elem_bt, VecMaskNotUsed) 886 || !arch_supports_vector(Op_VectorReinterpret, num_elem, elem_bt, VecMaskNotUsed)) { 887 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d*8 etype=%s/8 ismask=no", 888 is_store, "load", 889 mem_num_elem, type2name(mem_elem_bt)); 890 return false; // not supported 891 } 892 } 893 } 894 if (is_mask) { 895 if (!is_store) { 896 if (!arch_supports_vector(Op_LoadVector, num_elem, elem_bt, VecMaskUseLoad)) { 897 return false; // not supported 898 } 899 } else { 900 if (!arch_supports_vector(Op_StoreVector, num_elem, elem_bt, VecMaskUseStore)) { 901 return false; // not supported 902 } 903 } 904 } 905 906 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 907 908 if (needs_cpu_membar) { 909 insert_mem_bar(Op_MemBarCPUOrder); 910 } 911 912 if (is_store) { 913 Node* val = unbox_vector(argument(7), vbox_type, elem_bt, num_elem); 914 if (val == nullptr) { 915 return false; // operand unboxing failed 916 } 917 set_all_memory(reset_memory()); 918 919 // In case the store needs to happen to byte array, reinterpret the incoming vector to byte vector. 920 int store_num_elem = num_elem; 921 if (mismatched_ms) { 922 store_num_elem = mem_num_elem; 923 const TypeVect* to_vect_type = TypeVect::make(mem_elem_bt, store_num_elem); 924 val = gvn().transform(new VectorReinterpretNode(val, val->bottom_type()->is_vect(), to_vect_type)); 925 } 926 if (is_mask) { 927 val = gvn().transform(VectorStoreMaskNode::make(gvn(), val, elem_bt, num_elem)); 928 } 929 Node* vstore = gvn().transform(StoreVectorNode::make(0, control(), memory(addr), addr, addr_type, val, store_num_elem)); 930 set_memory(vstore, addr_type); 931 } else { 932 // When using byte array, we need to load as byte then reinterpret the value. Otherwise, do a simple vector load. 933 Node* vload = nullptr; 934 if (mismatched_ms) { 935 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, mem_num_elem, mem_elem_bt)); 936 const TypeVect* to_vect_type = TypeVect::make(elem_bt, num_elem); 937 vload = gvn().transform(new VectorReinterpretNode(vload, vload->bottom_type()->is_vect(), to_vect_type)); 938 } else { 939 // Special handle for masks 940 if (is_mask) { 941 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, num_elem, T_BOOLEAN)); 942 vload = gvn().transform(new VectorLoadMaskNode(vload, TypeVect::makemask(elem_bt, num_elem))); 943 } else { 944 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, num_elem, elem_bt)); 945 } 946 } 947 Node* box = box_vector(vload, vbox_type, elem_bt, num_elem); 948 set_result(box); 949 } 950 951 old_state.discard(); 952 953 if (needs_cpu_membar) { 954 insert_mem_bar(Op_MemBarCPUOrder); 955 } 956 957 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 958 return true; 959 } 960 961 // public static 962 // <C, 963 // V extends Vector<?>, 964 // E, 965 // S extends VectorSpecies<E>, 966 // M extends VectorMask<E>> 967 // V loadMasked(Class<? extends V> vClass, Class<M> mClass, Class<E> eClass, 968 // int length, Object base, long offset, // Unsafe addressing 969 // boolean fromSegment, 970 // M m, int offsetInRange, 971 // C container, long index, S s, // Arguments for default implementation 972 // LoadVectorMaskedOperation<C, V, S, M> defaultImpl) { 973 // public static 974 // <C, 975 // V extends Vector<E>, 976 // M extends VectorMask<E>, 977 // E> 978 // void storeMasked(Class<? extends V> vClass, Class<M> mClass, Class<E> eClass, 979 // int length, 980 // Object base, long offset, // Unsafe addressing 981 // boolean fromSegment, 982 // V v, M m, C container, long index, // Arguments for default implementation 983 // StoreVectorMaskedOperation<C, V, M> defaultImpl) { 984 985 bool LibraryCallKit::inline_vector_mem_masked_operation(bool is_store) { 986 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 987 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 988 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 989 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 990 const TypeInt* from_ms = gvn().type(argument(7))->isa_int(); 991 992 if (vector_klass == nullptr || vector_klass->const_oop() == nullptr || 993 mask_klass == nullptr || mask_klass->const_oop() == nullptr || 994 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 995 vlen == nullptr || !vlen->is_con() || 996 from_ms == nullptr || !from_ms->is_con()) { 997 log_if_needed(" ** missing constant: vclass=%s mclass=%s etype=%s vlen=%s from_ms=%s", 998 NodeClassNames[argument(0)->Opcode()], 999 NodeClassNames[argument(1)->Opcode()], 1000 NodeClassNames[argument(2)->Opcode()], 1001 NodeClassNames[argument(3)->Opcode()], 1002 NodeClassNames[argument(7)->Opcode()]); 1003 return false; // not enough info for intrinsification 1004 } 1005 if (!is_klass_initialized(vector_klass)) { 1006 log_if_needed(" ** klass argument not initialized"); 1007 return false; 1008 } 1009 1010 if (!is_klass_initialized(mask_klass)) { 1011 log_if_needed(" ** mask klass argument not initialized"); 1012 return false; 1013 } 1014 1015 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1016 if (!elem_type->is_primitive_type()) { 1017 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1018 return false; // should be primitive type 1019 } 1020 1021 BasicType elem_bt = elem_type->basic_type(); 1022 int num_elem = vlen->get_con(); 1023 1024 Node* base = argument(4); 1025 Node* offset = ConvL2X(argument(5)); 1026 1027 // Save state and restore on bailout 1028 SavedState old_state(this); 1029 1030 Node* addr = make_unsafe_address(base, offset, elem_bt, true); 1031 const TypePtr *addr_type = gvn().type(addr)->isa_ptr(); 1032 const TypeAryPtr* arr_type = addr_type->isa_aryptr(); 1033 1034 bool mismatched_ms = from_ms->get_con() && arr_type != nullptr && arr_type->elem()->array_element_basic_type() != elem_bt; 1035 BIG_ENDIAN_ONLY(if (mismatched_ms) return false;) 1036 // If there is no consistency between array and vector element types, it must be special byte array case 1037 if (arr_type != nullptr && !elem_consistent_with_arr(elem_bt, arr_type, mismatched_ms)) { 1038 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s atype=%s", 1039 is_store, is_store ? "storeMasked" : "loadMasked", 1040 num_elem, type2name(elem_bt), type2name(arr_type->elem()->array_element_basic_type())); 1041 return false; 1042 } 1043 1044 int mem_num_elem = mismatched_ms ? num_elem * type2aelembytes(elem_bt) : num_elem; 1045 BasicType mem_elem_bt = mismatched_ms ? T_BYTE : elem_bt; 1046 bool supports_predicate = arch_supports_vector(is_store ? Op_StoreVectorMasked : Op_LoadVectorMasked, 1047 mem_num_elem, mem_elem_bt, VecMaskUseLoad); 1048 1049 // If current arch does not support the predicated operations, we have to bail 1050 // out when current case uses the predicate feature. 1051 if (!supports_predicate) { 1052 bool needs_predicate = false; 1053 if (is_store) { 1054 // Masked vector store always uses the predicated store. 1055 needs_predicate = true; 1056 } else { 1057 // Masked vector load with IOOBE always uses the predicated load. 1058 const TypeInt* offset_in_range = gvn().type(argument(9))->isa_int(); 1059 if (!offset_in_range->is_con()) { 1060 log_if_needed(" ** missing constant: offsetInRange=%s", 1061 NodeClassNames[argument(8)->Opcode()]); 1062 return false; 1063 } 1064 needs_predicate = (offset_in_range->get_con() == 0); 1065 } 1066 1067 if (needs_predicate) { 1068 log_if_needed(" ** not supported: op=%s vlen=%d etype=%s mismatched_ms=%d", 1069 is_store ? "storeMasked" : "loadMasked", 1070 num_elem, type2name(elem_bt), mismatched_ms ? 1 : 0); 1071 return false; 1072 } 1073 } 1074 1075 // This only happens for masked vector load. If predicate is not supported, then check whether 1076 // the normal vector load and blend operations are supported by backend. 1077 if (!supports_predicate && (!arch_supports_vector(Op_LoadVector, mem_num_elem, mem_elem_bt, VecMaskNotUsed) || 1078 !arch_supports_vector(Op_VectorBlend, mem_num_elem, mem_elem_bt, VecMaskUseLoad))) { 1079 log_if_needed(" ** not supported: op=loadMasked vlen=%d etype=%s mismatched_ms=%d", 1080 num_elem, type2name(elem_bt), mismatched_ms ? 1 : 0); 1081 return false; 1082 } 1083 1084 // Since we are using byte array, we need to double check that the vector reinterpret operation 1085 // with byte type is supported by backend. 1086 if (mismatched_ms) { 1087 if (!arch_supports_vector(Op_VectorReinterpret, mem_num_elem, T_BYTE, VecMaskNotUsed)) { 1088 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s mismatched_ms=1", 1089 is_store, is_store ? "storeMasked" : "loadMasked", 1090 num_elem, type2name(elem_bt)); 1091 return false; 1092 } 1093 } 1094 1095 // Since it needs to unbox the mask, we need to double check that the related load operations 1096 // for mask are supported by backend. 1097 if (!arch_supports_vector(Op_LoadVector, num_elem, elem_bt, VecMaskUseLoad)) { 1098 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s", 1099 is_store, is_store ? "storeMasked" : "loadMasked", 1100 num_elem, type2name(elem_bt)); 1101 return false; 1102 } 1103 1104 // Can base be null? Otherwise, always on-heap access. 1105 bool can_access_non_heap = TypePtr::NULL_PTR->higher_equal(gvn().type(base)); 1106 if (can_access_non_heap) { 1107 insert_mem_bar(Op_MemBarCPUOrder); 1108 } 1109 1110 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1111 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1112 assert(!is_vector_mask(vbox_klass) && is_vector_mask(mbox_klass), "Invalid class type"); 1113 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1114 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1115 1116 Node* mask = unbox_vector(is_store ? argument(9) : argument(8), mbox_type, elem_bt, num_elem); 1117 if (mask == nullptr) { 1118 log_if_needed(" ** unbox failed mask=%s", 1119 is_store ? NodeClassNames[argument(9)->Opcode()] 1120 : NodeClassNames[argument(8)->Opcode()]); 1121 return false; 1122 } 1123 1124 if (is_store) { 1125 Node* val = unbox_vector(argument(8), vbox_type, elem_bt, num_elem); 1126 if (val == nullptr) { 1127 log_if_needed(" ** unbox failed vector=%s", 1128 NodeClassNames[argument(8)->Opcode()]); 1129 return false; // operand unboxing failed 1130 } 1131 set_all_memory(reset_memory()); 1132 1133 if (mismatched_ms) { 1134 // Reinterpret the incoming vector to byte vector. 1135 const TypeVect* to_vect_type = TypeVect::make(mem_elem_bt, mem_num_elem); 1136 val = gvn().transform(new VectorReinterpretNode(val, val->bottom_type()->is_vect(), to_vect_type)); 1137 // Reinterpret the vector mask to byte type. 1138 const TypeVect* from_mask_type = TypeVect::makemask(elem_bt, num_elem); 1139 const TypeVect* to_mask_type = TypeVect::makemask(mem_elem_bt, mem_num_elem); 1140 mask = gvn().transform(new VectorReinterpretNode(mask, from_mask_type, to_mask_type)); 1141 } 1142 Node* vstore = gvn().transform(new StoreVectorMaskedNode(control(), memory(addr), addr, val, addr_type, mask)); 1143 set_memory(vstore, addr_type); 1144 } else { 1145 Node* vload = nullptr; 1146 1147 if (mismatched_ms) { 1148 // Reinterpret the vector mask to byte type. 1149 const TypeVect* from_mask_type = TypeVect::makemask(elem_bt, num_elem); 1150 const TypeVect* to_mask_type = TypeVect::makemask(mem_elem_bt, mem_num_elem); 1151 mask = gvn().transform(new VectorReinterpretNode(mask, from_mask_type, to_mask_type)); 1152 } 1153 1154 if (supports_predicate) { 1155 // Generate masked load vector node if predicate feature is supported. 1156 const TypeVect* vt = TypeVect::make(mem_elem_bt, mem_num_elem); 1157 vload = gvn().transform(new LoadVectorMaskedNode(control(), memory(addr), addr, addr_type, vt, mask)); 1158 } else { 1159 // Use the vector blend to implement the masked load vector. The biased elements are zeros. 1160 Node* zero = gvn().transform(gvn().zerocon(mem_elem_bt)); 1161 zero = gvn().transform(VectorNode::scalar2vector(zero, mem_num_elem, mem_elem_bt)); 1162 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, mem_num_elem, mem_elem_bt)); 1163 vload = gvn().transform(new VectorBlendNode(zero, vload, mask)); 1164 } 1165 1166 if (mismatched_ms) { 1167 const TypeVect* to_vect_type = TypeVect::make(elem_bt, num_elem); 1168 vload = gvn().transform(new VectorReinterpretNode(vload, vload->bottom_type()->is_vect(), to_vect_type)); 1169 } 1170 1171 Node* box = box_vector(vload, vbox_type, elem_bt, num_elem); 1172 set_result(box); 1173 } 1174 1175 old_state.discard(); 1176 1177 if (can_access_non_heap) { 1178 insert_mem_bar(Op_MemBarCPUOrder); 1179 } 1180 1181 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1182 return true; 1183 } 1184 1185 // 1186 // <C, 1187 // V extends Vector<?>, 1188 // W extends Vector<Integer>, 1189 // S extends VectorSpecies<E>, 1190 // M extends VectorMask<E>, 1191 // E> 1192 // V loadWithMap(Class<? extends V> vClass, Class<M> mClass, Class<E> eClass, int length, 1193 // Class<? extends Vector<Integer>> vectorIndexClass, int indexLength, 1194 // Object base, long offset, 1195 // W indexVector1, W indexVector2, W indexVector3, W indexVector4, 1196 // M m, C container, int index, int[] indexMap, int indexM, S s, 1197 // LoadVectorOperationWithMap<C, V, S, M> defaultImpl) 1198 // 1199 // <C, 1200 // V extends Vector<E>, 1201 // W extends Vector<Integer>, 1202 // M extends VectorMask<E>, 1203 // E> 1204 // void storeWithMap(Class<? extends V> vClass, Class<M> mClass, Class<E> eClass, int length, 1205 // Class<? extends Vector<Integer>> vectorIndexClass, int indexLength, 1206 // Object base, long offset, // Unsafe addressing 1207 // W indexVector, V v, M m, 1208 // C container, int index, int[] indexMap, int indexM, // Arguments for default implementation 1209 // StoreVectorOperationWithMap<C, V, M> defaultImpl) 1210 // 1211 bool LibraryCallKit::inline_vector_gather_scatter(bool is_scatter) { 1212 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1213 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 1214 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1215 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1216 const TypeInstPtr* vector_idx_klass = gvn().type(argument(4))->isa_instptr(); 1217 const TypeInt* idx_vlen = gvn().type(argument(5))->isa_int(); 1218 1219 if (vector_klass == nullptr || vector_klass->const_oop() == nullptr || 1220 // mask_klass == nullptr || mask_klass->const_oop() == nullptr || 1221 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 1222 vlen == nullptr || !vlen->is_con() || 1223 vector_idx_klass == nullptr || vector_idx_klass->const_oop() == nullptr || 1224 idx_vlen == nullptr || !idx_vlen->is_con()) { 1225 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s viclass=%s idx_vlen=%s", 1226 NodeClassNames[argument(0)->Opcode()], 1227 NodeClassNames[argument(2)->Opcode()], 1228 NodeClassNames[argument(3)->Opcode()], 1229 NodeClassNames[argument(4)->Opcode()], 1230 NodeClassNames[argument(5)->Opcode()]); 1231 return false; // not enough info for intrinsification 1232 } 1233 1234 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(vector_idx_klass)) { 1235 log_if_needed(" ** klass argument not initialized"); 1236 return false; 1237 } 1238 1239 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1240 if (!elem_type->is_primitive_type()) { 1241 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1242 return false; // should be primitive type 1243 } 1244 1245 BasicType elem_bt = elem_type->basic_type(); 1246 int num_elem = vlen->get_con(); 1247 int idx_num_elem = idx_vlen->get_con(); 1248 1249 Node* m = is_scatter ? argument(11) : argument(13); 1250 const Type* vmask_type = gvn().type(m); 1251 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 1252 if (is_masked_op) { 1253 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 1254 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(1)->Opcode()]); 1255 return false; // not enough info for intrinsification 1256 } 1257 1258 if (!is_klass_initialized(mask_klass)) { 1259 log_if_needed(" ** mask klass argument not initialized"); 1260 return false; 1261 } 1262 1263 if (vmask_type->maybe_null()) { 1264 log_if_needed(" ** null mask values are not allowed for masked op"); 1265 return false; 1266 } 1267 1268 // Check whether the predicated gather/scatter node is supported by architecture. 1269 VectorMaskUseType mask = (VectorMaskUseType) (VecMaskUseLoad | VecMaskUsePred); 1270 if (!arch_supports_vector(is_scatter ? Op_StoreVectorScatterMasked : Op_LoadVectorGatherMasked, num_elem, elem_bt, mask)) { 1271 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s is_masked_op=1", 1272 is_scatter, is_scatter ? "scatterMasked" : "gatherMasked", 1273 num_elem, type2name(elem_bt)); 1274 return false; // not supported 1275 } 1276 } else { 1277 // Check whether the normal gather/scatter node is supported for non-masked operation. 1278 if (!arch_supports_vector(is_scatter ? Op_StoreVectorScatter : Op_LoadVectorGather, num_elem, elem_bt, VecMaskNotUsed)) { 1279 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s is_masked_op=0", 1280 is_scatter, is_scatter ? "scatter" : "gather", 1281 num_elem, type2name(elem_bt)); 1282 return false; // not supported 1283 } 1284 } 1285 1286 // Check that the vector holding indices is supported by architecture 1287 // For sub-word gathers expander receive index array. 1288 if (!is_subword_type(elem_bt) && !arch_supports_vector(Op_LoadVector, idx_num_elem, T_INT, VecMaskNotUsed)) { 1289 log_if_needed(" ** not supported: arity=%d op=%s/loadindex vlen=%d etype=int is_masked_op=%d", 1290 is_scatter, is_scatter ? "scatter" : "gather", 1291 idx_num_elem, is_masked_op ? 1 : 0); 1292 return false; // not supported 1293 } 1294 1295 Node* base = argument(6); 1296 Node* offset = ConvL2X(argument(7)); 1297 1298 // Save state and restore on bailout 1299 SavedState old_state(this); 1300 1301 Node* addr = nullptr; 1302 if (!is_subword_type(elem_bt)) { 1303 addr = make_unsafe_address(base, offset, elem_bt, true); 1304 } else { 1305 assert(!is_scatter, "Only supports gather operation for subword types now"); 1306 uint header = arrayOopDesc::base_offset_in_bytes(elem_bt); 1307 assert(offset->is_Con() && offset->bottom_type()->is_long()->get_con() == header, 1308 "offset must be the array base offset"); 1309 Node* index = argument(15); 1310 addr = array_element_address(base, index, elem_bt); 1311 } 1312 1313 const TypePtr* addr_type = gvn().type(addr)->isa_ptr(); 1314 const TypeAryPtr* arr_type = addr_type->isa_aryptr(); 1315 1316 // The array must be consistent with vector type 1317 if (arr_type == nullptr || (arr_type != nullptr && !elem_consistent_with_arr(elem_bt, arr_type, false))) { 1318 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s atype=%s ismask=no", 1319 is_scatter, is_scatter ? "scatter" : "gather", 1320 num_elem, type2name(elem_bt), type2name(arr_type->elem()->array_element_basic_type())); 1321 return false; 1322 } 1323 1324 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1325 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1326 ciKlass* vbox_idx_klass = vector_idx_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1327 if (vbox_idx_klass == nullptr) { 1328 return false; 1329 } 1330 1331 // Get the indexes for gather/scatter. 1332 Node* indexes = nullptr; 1333 const TypeInstPtr* vbox_idx_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_idx_klass); 1334 if (is_subword_type(elem_bt)) { 1335 Node* indexMap = argument(16); 1336 Node* indexM = argument(17); 1337 indexes = array_element_address(indexMap, indexM, T_INT); 1338 } else { 1339 // Get the first index vector. 1340 indexes = unbox_vector(argument(9), vbox_idx_type, T_INT, idx_num_elem); 1341 if (indexes == nullptr) { 1342 return false; 1343 } 1344 } 1345 1346 // Get the vector mask value. 1347 Node* mask = nullptr; 1348 if (is_masked_op) { 1349 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1350 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1351 mask = unbox_vector(m, mbox_type, elem_bt, num_elem); 1352 if (mask == nullptr) { 1353 log_if_needed(" ** unbox failed mask=%s", NodeClassNames[m->Opcode()]); 1354 return false; 1355 } 1356 } 1357 1358 const TypeVect* vector_type = TypeVect::make(elem_bt, num_elem); 1359 if (is_scatter) { 1360 Node* val = unbox_vector(argument(10), vbox_type, elem_bt, num_elem); 1361 if (val == nullptr) { 1362 return false; // operand unboxing failed 1363 } 1364 set_all_memory(reset_memory()); 1365 1366 Node* vstore = nullptr; 1367 if (mask != nullptr) { 1368 vstore = gvn().transform(new StoreVectorScatterMaskedNode(control(), memory(addr), addr, addr_type, val, indexes, mask)); 1369 } else { 1370 vstore = gvn().transform(new StoreVectorScatterNode(control(), memory(addr), addr, addr_type, val, indexes)); 1371 } 1372 set_memory(vstore, addr_type); 1373 } else { 1374 Node* vload = nullptr; 1375 if (mask != nullptr) { 1376 vload = gvn().transform(new LoadVectorGatherMaskedNode(control(), memory(addr), addr, addr_type, vector_type, indexes, mask)); 1377 } else { 1378 vload = gvn().transform(new LoadVectorGatherNode(control(), memory(addr), addr, addr_type, vector_type, indexes)); 1379 } 1380 Node* box = box_vector(vload, vbox_type, elem_bt, num_elem); 1381 set_result(box); 1382 } 1383 1384 old_state.discard(); 1385 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1386 return true; 1387 } 1388 1389 // public static 1390 // <V extends Vector<E>, 1391 // M extends VectorMask<E>, 1392 // E> 1393 // long reductionCoerced(int oprId, Class<? extends V> vectorClass, Class<? extends M> maskClass, 1394 // Class<E> elementType, int length, V v, M m, 1395 // ReductionOperation<V, M> defaultImpl) 1396 bool LibraryCallKit::inline_vector_reduction() { 1397 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 1398 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 1399 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 1400 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 1401 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 1402 1403 if (opr == nullptr || !opr->is_con() || 1404 vector_klass == nullptr || vector_klass->const_oop() == nullptr || 1405 // mask_klass == nullptr || mask_klass->const_oop() == nullptr || 1406 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 1407 vlen == nullptr || !vlen->is_con()) { 1408 log_if_needed(" ** missing constant: opr=%s vclass=%s etype=%s vlen=%s", 1409 NodeClassNames[argument(0)->Opcode()], 1410 NodeClassNames[argument(1)->Opcode()], 1411 NodeClassNames[argument(3)->Opcode()], 1412 NodeClassNames[argument(4)->Opcode()]); 1413 return false; // not enough info for intrinsification 1414 } 1415 if (!is_klass_initialized(vector_klass)) { 1416 log_if_needed(" ** klass argument not initialized"); 1417 return false; 1418 } 1419 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1420 if (!elem_type->is_primitive_type()) { 1421 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1422 return false; // should be primitive type 1423 } 1424 1425 const Type* vmask_type = gvn().type(argument(6)); 1426 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 1427 if (is_masked_op) { 1428 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 1429 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(2)->Opcode()]); 1430 return false; // not enough info for intrinsification 1431 } 1432 1433 if (!is_klass_initialized(mask_klass)) { 1434 log_if_needed(" ** mask klass argument not initialized"); 1435 return false; 1436 } 1437 1438 if (vmask_type->maybe_null()) { 1439 log_if_needed(" ** null mask values are not allowed for masked op"); 1440 return false; 1441 } 1442 } 1443 1444 BasicType elem_bt = elem_type->basic_type(); 1445 int num_elem = vlen->get_con(); 1446 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 1447 int sopc = ReductionNode::opcode(opc, elem_bt); 1448 1449 // Ensure reduction operation for lanewise operation 1450 // When using mask, mask use type needs to be VecMaskUseLoad. 1451 if (sopc == opc || !arch_supports_vector(sopc, num_elem, elem_bt, is_masked_op ? VecMaskUseLoad : VecMaskNotUsed)) { 1452 log_if_needed(" ** not supported: arity=1 op=%d/reduce vlen=%d etype=%s is_masked_op=%d", 1453 sopc, num_elem, type2name(elem_bt), is_masked_op ? 1 : 0); 1454 return false; 1455 } 1456 1457 // Return true if current platform has implemented the masked operation with predicate feature. 1458 bool use_predicate = is_masked_op && arch_supports_vector(sopc, num_elem, elem_bt, VecMaskUsePred); 1459 if (is_masked_op && !use_predicate && !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)) { 1460 log_if_needed(" ** not supported: arity=1 op=%d/reduce vlen=%d etype=%s is_masked_op=1", 1461 sopc, num_elem, type2name(elem_bt)); 1462 return false; 1463 } 1464 1465 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1466 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1467 1468 Node* opd = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1469 if (opd == nullptr) { 1470 return false; // operand unboxing failed 1471 } 1472 1473 Node* mask = nullptr; 1474 if (is_masked_op) { 1475 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1476 assert(is_vector_mask(mbox_klass), "argument(2) should be a mask class"); 1477 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1478 mask = unbox_vector(argument(6), mbox_type, elem_bt, num_elem); 1479 if (mask == nullptr) { 1480 log_if_needed(" ** unbox failed mask=%s", 1481 NodeClassNames[argument(6)->Opcode()]); 1482 return false; 1483 } 1484 } 1485 1486 Node* init = ReductionNode::make_identity_con_scalar(gvn(), opc, elem_bt); 1487 Node* value = opd; 1488 1489 assert(mask != nullptr || !is_masked_op, "Masked op needs the mask value never null"); 1490 if (mask != nullptr && !use_predicate) { 1491 Node* reduce_identity = gvn().transform(VectorNode::scalar2vector(init, num_elem, elem_bt)); 1492 value = gvn().transform(new VectorBlendNode(reduce_identity, value, mask)); 1493 } 1494 1495 // Make an unordered Reduction node. This affects only AddReductionVF/VD and MulReductionVF/VD, 1496 // as these operations are allowed to be associative (not requiring strict order) in VectorAPI. 1497 value = ReductionNode::make(opc, nullptr, init, value, elem_bt, /* requires_strict_order */ false); 1498 1499 if (mask != nullptr && use_predicate) { 1500 value->add_req(mask); 1501 value->add_flag(Node::Flag_is_predicated_vector); 1502 } 1503 1504 value = gvn().transform(value); 1505 1506 Node* bits = nullptr; 1507 switch (elem_bt) { 1508 case T_BYTE: 1509 case T_SHORT: 1510 case T_INT: { 1511 bits = gvn().transform(new ConvI2LNode(value)); 1512 break; 1513 } 1514 case T_FLOAT: { 1515 value = gvn().transform(new MoveF2INode(value)); 1516 bits = gvn().transform(new ConvI2LNode(value)); 1517 break; 1518 } 1519 case T_DOUBLE: { 1520 bits = gvn().transform(new MoveD2LNode(value)); 1521 break; 1522 } 1523 case T_LONG: { 1524 bits = value; // no conversion needed 1525 break; 1526 } 1527 default: fatal("%s", type2name(elem_bt)); 1528 } 1529 set_result(bits); 1530 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1531 return true; 1532 } 1533 1534 // public static <V> boolean test(int cond, Class<?> vectorClass, Class<?> elementType, int vlen, 1535 // V v1, V v2, 1536 // BiFunction<V, V, Boolean> defaultImpl) 1537 // 1538 bool LibraryCallKit::inline_vector_test() { 1539 const TypeInt* cond = gvn().type(argument(0))->isa_int(); 1540 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 1541 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1542 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1543 1544 if (cond == nullptr || !cond->is_con() || 1545 vector_klass == nullptr || vector_klass->const_oop() == nullptr || 1546 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 1547 vlen == nullptr || !vlen->is_con()) { 1548 log_if_needed(" ** missing constant: cond=%s vclass=%s etype=%s vlen=%s", 1549 NodeClassNames[argument(0)->Opcode()], 1550 NodeClassNames[argument(1)->Opcode()], 1551 NodeClassNames[argument(2)->Opcode()], 1552 NodeClassNames[argument(3)->Opcode()]); 1553 return false; // not enough info for intrinsification 1554 } 1555 if (!is_klass_initialized(vector_klass)) { 1556 log_if_needed(" ** klass argument not initialized"); 1557 return false; 1558 } 1559 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1560 if (!elem_type->is_primitive_type()) { 1561 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1562 return false; // should be primitive type 1563 } 1564 BasicType elem_bt = elem_type->basic_type(); 1565 int num_elem = vlen->get_con(); 1566 BoolTest::mask booltest = (BoolTest::mask)cond->get_con(); 1567 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1568 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1569 1570 if (!arch_supports_vector(Op_VectorTest, num_elem, elem_bt, is_vector_mask(vbox_klass) ? VecMaskUseLoad : VecMaskNotUsed)) { 1571 log_if_needed(" ** not supported: arity=2 op=test/%d vlen=%d etype=%s ismask=%d", 1572 cond->get_con(), num_elem, type2name(elem_bt), 1573 is_vector_mask(vbox_klass)); 1574 return false; 1575 } 1576 1577 Node* opd1 = unbox_vector(argument(4), vbox_type, elem_bt, num_elem); 1578 Node* opd2; 1579 if (Matcher::vectortest_needs_second_argument(booltest == BoolTest::overflow, 1580 opd1->bottom_type()->isa_vectmask())) { 1581 opd2 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1582 } else { 1583 opd2 = opd1; 1584 } 1585 if (opd1 == nullptr || opd2 == nullptr) { 1586 return false; // operand unboxing failed 1587 } 1588 1589 Node* cmp = gvn().transform(new VectorTestNode(opd1, opd2, booltest)); 1590 BoolTest::mask test = Matcher::vectortest_mask(booltest == BoolTest::overflow, 1591 opd1->bottom_type()->isa_vectmask(), num_elem); 1592 Node* bol = gvn().transform(new BoolNode(cmp, test)); 1593 Node* res = gvn().transform(new CMoveINode(bol, gvn().intcon(0), gvn().intcon(1), TypeInt::BOOL)); 1594 1595 set_result(res); 1596 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1597 return true; 1598 } 1599 1600 // public static 1601 // <V extends Vector<E>, 1602 // M extends VectorMask<E>, 1603 // E> 1604 // V blend(Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, int vlen, 1605 // V v1, V v2, M m, 1606 // VectorBlendOp<V, M, E> defaultImpl) 1607 bool LibraryCallKit::inline_vector_blend() { 1608 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1609 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 1610 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1611 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1612 1613 if (mask_klass == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 1614 return false; // dead code 1615 } 1616 if (mask_klass->const_oop() == nullptr || vector_klass->const_oop() == nullptr || 1617 elem_klass->const_oop() == nullptr || !vlen->is_con()) { 1618 log_if_needed(" ** missing constant: vclass=%s mclass=%s etype=%s vlen=%s", 1619 NodeClassNames[argument(0)->Opcode()], 1620 NodeClassNames[argument(1)->Opcode()], 1621 NodeClassNames[argument(2)->Opcode()], 1622 NodeClassNames[argument(3)->Opcode()]); 1623 return false; // not enough info for intrinsification 1624 } 1625 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(mask_klass)) { 1626 log_if_needed(" ** klass argument not initialized"); 1627 return false; 1628 } 1629 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1630 if (!elem_type->is_primitive_type()) { 1631 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1632 return false; // should be primitive type 1633 } 1634 BasicType elem_bt = elem_type->basic_type(); 1635 BasicType mask_bt = elem_bt; 1636 int num_elem = vlen->get_con(); 1637 1638 if (!arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)) { 1639 log_if_needed(" ** not supported: arity=2 op=blend vlen=%d etype=%s ismask=useload", 1640 num_elem, type2name(elem_bt)); 1641 return false; // not supported 1642 } 1643 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1644 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1645 1646 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1647 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1648 1649 Node* v1 = unbox_vector(argument(4), vbox_type, elem_bt, num_elem); 1650 Node* v2 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1651 Node* mask = unbox_vector(argument(6), mbox_type, mask_bt, num_elem); 1652 1653 if (v1 == nullptr || v2 == nullptr || mask == nullptr) { 1654 return false; // operand unboxing failed 1655 } 1656 1657 Node* blend = gvn().transform(new VectorBlendNode(v1, v2, mask)); 1658 1659 Node* box = box_vector(blend, vbox_type, elem_bt, num_elem); 1660 set_result(box); 1661 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1662 return true; 1663 } 1664 1665 // public static 1666 // <V extends Vector<E>, 1667 // M extends VectorMask<E>, 1668 // E> 1669 // M compare(int cond, Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, int vlen, 1670 // V v1, V v2, M m, 1671 // VectorCompareOp<V,M> defaultImpl) 1672 bool LibraryCallKit::inline_vector_compare() { 1673 const TypeInt* cond = gvn().type(argument(0))->isa_int(); 1674 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 1675 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 1676 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 1677 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 1678 1679 if (cond == nullptr || vector_klass == nullptr || mask_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 1680 return false; // dead code 1681 } 1682 if (!cond->is_con() || vector_klass->const_oop() == nullptr || mask_klass->const_oop() == nullptr || 1683 elem_klass->const_oop() == nullptr || !vlen->is_con()) { 1684 log_if_needed(" ** missing constant: cond=%s vclass=%s mclass=%s etype=%s vlen=%s", 1685 NodeClassNames[argument(0)->Opcode()], 1686 NodeClassNames[argument(1)->Opcode()], 1687 NodeClassNames[argument(2)->Opcode()], 1688 NodeClassNames[argument(3)->Opcode()], 1689 NodeClassNames[argument(4)->Opcode()]); 1690 return false; // not enough info for intrinsification 1691 } 1692 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(mask_klass)) { 1693 log_if_needed(" ** klass argument not initialized"); 1694 return false; 1695 } 1696 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1697 if (!elem_type->is_primitive_type()) { 1698 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1699 return false; // should be primitive type 1700 } 1701 1702 int num_elem = vlen->get_con(); 1703 BasicType elem_bt = elem_type->basic_type(); 1704 BasicType mask_bt = elem_bt; 1705 1706 if ((cond->get_con() & BoolTest::unsigned_compare) != 0) { 1707 if (!Matcher::supports_vector_comparison_unsigned(num_elem, elem_bt)) { 1708 log_if_needed(" ** not supported: unsigned comparison op=comp/%d vlen=%d etype=%s ismask=usestore", 1709 cond->get_con() & (BoolTest::unsigned_compare - 1), num_elem, type2name(elem_bt)); 1710 return false; 1711 } 1712 } 1713 1714 if (!arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskUseStore)) { 1715 log_if_needed(" ** not supported: arity=2 op=comp/%d vlen=%d etype=%s ismask=usestore", 1716 cond->get_con(), num_elem, type2name(elem_bt)); 1717 return false; 1718 } 1719 1720 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1721 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1722 1723 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1724 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1725 1726 Node* v1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1727 Node* v2 = unbox_vector(argument(6), vbox_type, elem_bt, num_elem); 1728 1729 bool is_masked_op = argument(7)->bottom_type() != TypePtr::NULL_PTR; 1730 Node* mask = is_masked_op ? unbox_vector(argument(7), mbox_type, elem_bt, num_elem) : nullptr; 1731 if (is_masked_op && mask == nullptr) { 1732 log_if_needed(" ** not supported: mask = null arity=2 op=comp/%d vlen=%d etype=%s ismask=usestore is_masked_op=1", 1733 cond->get_con(), num_elem, type2name(elem_bt)); 1734 return false; 1735 } 1736 1737 bool use_predicate = is_masked_op && arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskUsePred); 1738 if (is_masked_op && !use_predicate && !arch_supports_vector(Op_AndV, num_elem, elem_bt, VecMaskUseLoad)) { 1739 log_if_needed(" ** not supported: arity=2 op=comp/%d vlen=%d etype=%s ismask=usestore is_masked_op=1", 1740 cond->get_con(), num_elem, type2name(elem_bt)); 1741 return false; 1742 } 1743 1744 if (v1 == nullptr || v2 == nullptr) { 1745 return false; // operand unboxing failed 1746 } 1747 BoolTest::mask pred = (BoolTest::mask)cond->get_con(); 1748 ConINode* pred_node = (ConINode*)gvn().makecon(cond); 1749 1750 const TypeVect* vmask_type = TypeVect::makemask(mask_bt, num_elem); 1751 Node* operation = new VectorMaskCmpNode(pred, v1, v2, pred_node, vmask_type); 1752 1753 if (is_masked_op) { 1754 if (use_predicate) { 1755 operation->add_req(mask); 1756 operation->add_flag(Node::Flag_is_predicated_vector); 1757 } else { 1758 operation = gvn().transform(operation); 1759 operation = VectorNode::make(Op_AndV, operation, mask, vmask_type); 1760 } 1761 } 1762 1763 operation = gvn().transform(operation); 1764 1765 Node* box = box_vector(operation, mbox_type, mask_bt, num_elem); 1766 set_result(box); 1767 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1768 return true; 1769 } 1770 1771 // public static 1772 // <V extends Vector<E>, 1773 // Sh extends VectorShuffle<E>, 1774 // M extends VectorMask<E>, 1775 // E> 1776 // V rearrangeOp(Class<? extends V> vectorClass, Class<Sh> shuffleClass, Class<M> maskClass, Class<E> elementType, int vlen, 1777 // V v1, Sh sh, M m, 1778 // VectorRearrangeOp<V, Sh, M, E> defaultImpl) 1779 bool LibraryCallKit::inline_vector_rearrange() { 1780 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1781 const TypeInstPtr* shuffle_klass = gvn().type(argument(1))->isa_instptr(); 1782 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 1783 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 1784 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 1785 1786 if (vector_klass == nullptr || shuffle_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 1787 return false; // dead code 1788 } 1789 if (shuffle_klass->const_oop() == nullptr || 1790 vector_klass->const_oop() == nullptr || 1791 elem_klass->const_oop() == nullptr || 1792 !vlen->is_con()) { 1793 log_if_needed(" ** missing constant: vclass=%s sclass=%s etype=%s vlen=%s", 1794 NodeClassNames[argument(0)->Opcode()], 1795 NodeClassNames[argument(1)->Opcode()], 1796 NodeClassNames[argument(3)->Opcode()], 1797 NodeClassNames[argument(4)->Opcode()]); 1798 return false; // not enough info for intrinsification 1799 } 1800 if (!is_klass_initialized(vector_klass) || 1801 !is_klass_initialized(shuffle_klass)) { 1802 log_if_needed(" ** klass argument not initialized"); 1803 return false; 1804 } 1805 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1806 if (!elem_type->is_primitive_type()) { 1807 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1808 return false; // should be primitive type 1809 } 1810 1811 BasicType elem_bt = elem_type->basic_type(); 1812 BasicType shuffle_bt = elem_bt; 1813 if (shuffle_bt == T_FLOAT) { 1814 shuffle_bt = T_INT; 1815 } else if (shuffle_bt == T_DOUBLE) { 1816 shuffle_bt = T_LONG; 1817 } 1818 1819 int num_elem = vlen->get_con(); 1820 bool need_load_shuffle = Matcher::vector_rearrange_requires_load_shuffle(shuffle_bt, num_elem); 1821 1822 if (need_load_shuffle && !arch_supports_vector(Op_VectorLoadShuffle, num_elem, shuffle_bt, VecMaskNotUsed)) { 1823 if (C->print_intrinsics()) { 1824 tty->print_cr(" ** not supported: arity=0 op=load/shuffle vlen=%d etype=%s ismask=no", 1825 num_elem, type2name(shuffle_bt)); 1826 } 1827 return false; // not supported 1828 } 1829 1830 bool is_masked_op = argument(7)->bottom_type() != TypePtr::NULL_PTR; 1831 bool use_predicate = is_masked_op; 1832 if (is_masked_op && 1833 (mask_klass == nullptr || 1834 mask_klass->const_oop() == nullptr || 1835 !is_klass_initialized(mask_klass))) { 1836 log_if_needed(" ** mask_klass argument not initialized"); 1837 } 1838 if (!arch_supports_vector(Op_AndV, num_elem, elem_bt, VecMaskNotUsed)) { 1839 log_if_needed(" ** not supported: arity=2 op=and vlen=%d etype=%s ismask=no", 1840 num_elem, type2name(elem_bt)); 1841 return false; 1842 } 1843 VectorMaskUseType checkFlags = (VectorMaskUseType)(is_masked_op ? (VecMaskUseLoad | VecMaskUsePred) : VecMaskNotUsed); 1844 if (!arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, checkFlags)) { 1845 use_predicate = false; 1846 if(!is_masked_op || 1847 (!arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, VecMaskNotUsed) || 1848 !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad) || 1849 !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed))) { 1850 log_if_needed(" ** not supported: arity=2 op=shuffle/rearrange vlen=%d etype=%s ismask=no", 1851 num_elem, type2name(elem_bt)); 1852 return false; // not supported 1853 } 1854 } 1855 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1856 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1857 1858 ciKlass* shbox_klass = shuffle_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1859 const TypeInstPtr* shbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, shbox_klass); 1860 1861 Node* v1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1862 Node* shuffle = unbox_vector(argument(6), shbox_type, shuffle_bt, num_elem); 1863 const TypeVect* st = TypeVect::make(shuffle_bt, num_elem); 1864 1865 if (v1 == nullptr || shuffle == nullptr) { 1866 return false; // operand unboxing failed 1867 } 1868 1869 assert(is_power_of_2(num_elem), "wrapping invalid"); 1870 Node* wrapping_mask_elem = gvn().makecon(TypeInteger::make(num_elem - 1, num_elem - 1, Type::WidenMin, shuffle_bt == T_LONG ? T_LONG : T_INT)); 1871 Node* wrapping_mask = gvn().transform(VectorNode::scalar2vector(wrapping_mask_elem, num_elem, shuffle_bt)); 1872 shuffle = gvn().transform(new AndVNode(shuffle, wrapping_mask, st)); 1873 1874 Node* mask = nullptr; 1875 if (is_masked_op) { 1876 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1877 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1878 mask = unbox_vector(argument(7), mbox_type, elem_bt, num_elem); 1879 if (mask == nullptr) { 1880 log_if_needed(" ** not supported: arity=3 op=shuffle/rearrange vlen=%d etype=%s ismask=useload is_masked_op=1", 1881 num_elem, type2name(elem_bt)); 1882 return false; 1883 } 1884 } 1885 1886 if (need_load_shuffle) { 1887 shuffle = gvn().transform(new VectorLoadShuffleNode(shuffle, st)); 1888 } 1889 1890 Node* rearrange = new VectorRearrangeNode(v1, shuffle); 1891 if (is_masked_op) { 1892 if (use_predicate) { 1893 rearrange->add_req(mask); 1894 rearrange->add_flag(Node::Flag_is_predicated_vector); 1895 } else { 1896 rearrange = gvn().transform(rearrange); 1897 Node* zero = gvn().makecon(Type::get_zero_type(elem_bt)); 1898 Node* zerovec = gvn().transform(VectorNode::scalar2vector(zero, num_elem, elem_bt)); 1899 rearrange = new VectorBlendNode(zerovec, rearrange, mask); 1900 } 1901 } 1902 rearrange = gvn().transform(rearrange); 1903 1904 Node* box = box_vector(rearrange, vbox_type, elem_bt, num_elem); 1905 set_result(box); 1906 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1907 return true; 1908 } 1909 1910 // public static 1911 // <V extends Vector<E>, 1912 // M extends VectorMask<E>, 1913 // E> 1914 // V selectFromOp(Class<? extends V> vClass, Class<M> mClass, Class<E> eClass, 1915 // int length, V v1, V v2, M m, 1916 // VectorSelectFromOp<V, M> defaultImpl) 1917 bool LibraryCallKit::inline_vector_select_from() { 1918 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1919 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 1920 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1921 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1922 1923 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 1924 vector_klass->const_oop() == nullptr || 1925 elem_klass->const_oop() == nullptr || 1926 !vlen->is_con()) { 1927 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s", 1928 NodeClassNames[argument(0)->Opcode()], 1929 NodeClassNames[argument(2)->Opcode()], 1930 NodeClassNames[argument(3)->Opcode()]); 1931 return false; // not enough info for intrinsification 1932 } 1933 if (!is_klass_initialized(vector_klass)) { 1934 log_if_needed(" ** klass argument not initialized"); 1935 return false; 1936 } 1937 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1938 if (!elem_type->is_primitive_type()) { 1939 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1940 return false; // should be primitive type 1941 } 1942 BasicType elem_bt = elem_type->basic_type(); 1943 int num_elem = vlen->get_con(); 1944 if (!is_power_of_2(num_elem)) { 1945 log_if_needed(" ** vlen not power of two=%d", num_elem); 1946 return false; 1947 } 1948 1949 BasicType shuffle_bt = elem_bt; 1950 if (shuffle_bt == T_FLOAT) { 1951 shuffle_bt = T_INT; 1952 } else if (shuffle_bt == T_DOUBLE) { 1953 shuffle_bt = T_LONG; 1954 } 1955 bool need_load_shuffle = Matcher::vector_rearrange_requires_load_shuffle(shuffle_bt, num_elem); 1956 1957 int cast_vopc = VectorCastNode::opcode(-1, elem_bt); // from vector of type elem_bt 1958 if ((need_load_shuffle && !arch_supports_vector(Op_VectorLoadShuffle, num_elem, elem_bt, VecMaskNotUsed)) || 1959 (elem_bt != shuffle_bt && !arch_supports_vector(cast_vopc, num_elem, shuffle_bt, VecMaskNotUsed)) || 1960 !arch_supports_vector(Op_AndV, num_elem, shuffle_bt, VecMaskNotUsed) || 1961 !arch_supports_vector(Op_Replicate, num_elem, shuffle_bt, VecMaskNotUsed)) { 1962 log_if_needed(" ** not supported: arity=0 op=selectFrom vlen=%d etype=%s ismask=no", 1963 num_elem, type2name(elem_bt)); 1964 return false; // not supported 1965 } 1966 1967 bool is_masked_op = argument(6)->bottom_type() != TypePtr::NULL_PTR; 1968 bool use_predicate = is_masked_op; 1969 if (is_masked_op && 1970 (mask_klass == nullptr || 1971 mask_klass->const_oop() == nullptr || 1972 !is_klass_initialized(mask_klass))) { 1973 log_if_needed(" ** mask_klass argument not initialized"); 1974 return false; // not supported 1975 } 1976 VectorMaskUseType checkFlags = (VectorMaskUseType)(is_masked_op ? (VecMaskUseLoad | VecMaskUsePred) : VecMaskNotUsed); 1977 if (!arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, checkFlags)) { 1978 use_predicate = false; 1979 if(!is_masked_op || 1980 (!arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, VecMaskNotUsed) || 1981 !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad) || 1982 !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed))) { 1983 log_if_needed(" ** not supported: op=selectFrom vlen=%d etype=%s is_masked_op=%d", 1984 num_elem, type2name(elem_bt), is_masked_op); 1985 return false; // not supported 1986 } 1987 } 1988 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1989 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1990 1991 // v1 is the index vector 1992 Node* v1 = unbox_vector(argument(4), vbox_type, elem_bt, num_elem); 1993 // v2 is the vector being rearranged 1994 Node* v2 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1995 1996 if (v1 == nullptr) { 1997 log_if_needed(" ** unbox failed v1=%s", NodeClassNames[argument(4)->Opcode()]); 1998 return false; // operand unboxing failed 1999 } 2000 2001 if (v2 == nullptr) { 2002 log_if_needed(" ** unbox failed v2=%s", NodeClassNames[argument(5)->Opcode()]); 2003 return false; // operand unboxing failed 2004 } 2005 2006 Node* mask = nullptr; 2007 if (is_masked_op) { 2008 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2009 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 2010 mask = unbox_vector(argument(6), mbox_type, elem_bt, num_elem); 2011 if (mask == nullptr) { 2012 log_if_needed(" ** unbox failed mask=%s", NodeClassNames[argument(6)->Opcode()]); 2013 return false; 2014 } 2015 } 2016 2017 // cast index vector from elem_bt vector to byte vector 2018 const TypeVect* shuffle_vt = TypeVect::make(shuffle_bt, num_elem); 2019 Node* shuffle = v1; 2020 2021 if (shuffle_bt != elem_bt) { 2022 shuffle = gvn().transform(VectorCastNode::make(cast_vopc, v1, shuffle_bt, num_elem)); 2023 } 2024 2025 // wrap the byte vector lanes to (num_elem - 1) to form the shuffle vector where num_elem is vector length 2026 // this is a simple AND operation as we come here only for power of two vector length 2027 Node* mod_val = gvn().makecon(TypeInteger::make(num_elem - 1, num_elem - 1, Type::WidenMin, shuffle_bt == T_LONG ? T_LONG : T_INT)); 2028 Node* bcast_mod = gvn().transform(VectorNode::scalar2vector(mod_val, num_elem, shuffle_bt)); 2029 shuffle = gvn().transform(VectorNode::make(Op_AndV, shuffle, bcast_mod, shuffle_vt)); 2030 2031 // load the shuffle to use in rearrange 2032 if (need_load_shuffle) { 2033 shuffle = gvn().transform(new VectorLoadShuffleNode(shuffle, shuffle_vt)); 2034 } 2035 2036 // and finally rearrange 2037 Node* rearrange = new VectorRearrangeNode(v2, shuffle); 2038 if (is_masked_op) { 2039 if (use_predicate) { 2040 // masked rearrange is supported so use that directly 2041 rearrange->add_req(mask); 2042 rearrange->add_flag(Node::Flag_is_predicated_vector); 2043 } else { 2044 // masked rearrange is not supported so emulate usig blend 2045 const TypeVect* vt = v1->bottom_type()->is_vect(); 2046 rearrange = gvn().transform(rearrange); 2047 2048 // create a zero vector with each lane element set as zero 2049 Node* zero = gvn().makecon(Type::get_zero_type(elem_bt)); 2050 Node* zerovec = gvn().transform(VectorNode::scalar2vector(zero, num_elem, elem_bt)); 2051 2052 // For each lane for which mask is set, blend in the rearranged lane into zero vector 2053 rearrange = new VectorBlendNode(zerovec, rearrange, mask); 2054 } 2055 } 2056 rearrange = gvn().transform(rearrange); 2057 2058 // box the result 2059 Node* box = box_vector(rearrange, vbox_type, elem_bt, num_elem); 2060 set_result(box); 2061 2062 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2063 return true; 2064 } 2065 2066 // public static 2067 // <V extends Vector<E>, 2068 // M extends VectorMask<E>, 2069 // E> 2070 // V broadcastInt(int opr, Class<? extends V> vectorClass, Class<? extends M> maskClass, 2071 // Class<E> elementType, int length, 2072 // V v, int n, M m, 2073 // VectorBroadcastIntOp<V, M> defaultImpl) 2074 bool LibraryCallKit::inline_vector_broadcast_int() { 2075 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 2076 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 2077 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 2078 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 2079 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 2080 2081 if (opr == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 2082 return false; // dead code 2083 } 2084 if (!opr->is_con() || vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 2085 log_if_needed(" ** missing constant: opr=%s vclass=%s etype=%s vlen=%s", 2086 NodeClassNames[argument(0)->Opcode()], 2087 NodeClassNames[argument(1)->Opcode()], 2088 NodeClassNames[argument(3)->Opcode()], 2089 NodeClassNames[argument(4)->Opcode()]); 2090 return false; // not enough info for intrinsification 2091 } 2092 if (!is_klass_initialized(vector_klass)) { 2093 log_if_needed(" ** klass argument not initialized"); 2094 return false; 2095 } 2096 2097 const Type* vmask_type = gvn().type(argument(7)); 2098 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 2099 if (is_masked_op) { 2100 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 2101 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(2)->Opcode()]); 2102 return false; // not enough info for intrinsification 2103 } 2104 2105 if (!is_klass_initialized(mask_klass)) { 2106 log_if_needed(" ** mask klass argument not initialized"); 2107 return false; 2108 } 2109 2110 if (vmask_type->maybe_null()) { 2111 log_if_needed(" ** null mask values are not allowed for masked op"); 2112 return false; 2113 } 2114 } 2115 2116 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2117 if (!elem_type->is_primitive_type()) { 2118 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2119 return false; // should be primitive type 2120 } 2121 2122 int num_elem = vlen->get_con(); 2123 BasicType elem_bt = elem_type->basic_type(); 2124 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 2125 2126 bool is_shift = VectorNode::is_shift_opcode(opc); 2127 bool is_rotate = VectorNode::is_rotate_opcode(opc); 2128 2129 if (opc == 0 || (!is_shift && !is_rotate)) { 2130 log_if_needed(" ** operation not supported: op=%d bt=%s", opr->get_con(), type2name(elem_bt)); 2131 return false; // operation not supported 2132 } 2133 2134 int sopc = VectorNode::opcode(opc, elem_bt); 2135 if (sopc == 0) { 2136 log_if_needed(" ** operation not supported: opc=%s bt=%s", NodeClassNames[opc], type2name(elem_bt)); 2137 return false; // operation not supported 2138 } 2139 2140 Node* cnt = argument(6); 2141 const TypeInt* cnt_type = cnt->bottom_type()->isa_int(); 2142 2143 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2144 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2145 2146 // If CPU supports vector constant rotate instructions pass it directly 2147 bool is_const_rotate = is_rotate && cnt_type && cnt_type->is_con() && 2148 Matcher::supports_vector_constant_rotates(cnt_type->get_con()); 2149 bool has_scalar_args = is_rotate ? !is_const_rotate : true; 2150 2151 VectorMaskUseType checkFlags = (VectorMaskUseType)(is_masked_op ? (VecMaskUseLoad | VecMaskUsePred) : VecMaskNotUsed); 2152 bool use_predicate = is_masked_op; 2153 2154 if (!arch_supports_vector(sopc, num_elem, elem_bt, checkFlags, has_scalar_args)) { 2155 use_predicate = false; 2156 if (!is_masked_op || 2157 (!arch_supports_vector(sopc, num_elem, elem_bt, VecMaskNotUsed, has_scalar_args) || 2158 !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad))) { 2159 2160 log_if_needed(" ** not supported: arity=0 op=int/%d vlen=%d etype=%s is_masked_op=%d", 2161 sopc, num_elem, type2name(elem_bt), is_masked_op ? 1 : 0); 2162 return false; // not supported 2163 } 2164 } 2165 2166 Node* opd1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 2167 Node* opd2 = nullptr; 2168 if (is_shift) { 2169 opd2 = vector_shift_count(cnt, opc, elem_bt, num_elem); 2170 } else { 2171 assert(is_rotate, "unexpected operation"); 2172 if (!is_const_rotate) { 2173 cnt = elem_bt == T_LONG ? gvn().transform(new ConvI2LNode(cnt)) : cnt; 2174 opd2 = gvn().transform(VectorNode::scalar2vector(cnt, num_elem, elem_bt)); 2175 } else { 2176 // Constant shift value. 2177 opd2 = cnt; 2178 } 2179 } 2180 2181 if (opd1 == nullptr || opd2 == nullptr) { 2182 return false; 2183 } 2184 2185 Node* mask = nullptr; 2186 if (is_masked_op) { 2187 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2188 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 2189 mask = unbox_vector(argument(7), mbox_type, elem_bt, num_elem); 2190 if (mask == nullptr) { 2191 log_if_needed(" ** unbox failed mask=%s", NodeClassNames[argument(7)->Opcode()]); 2192 return false; 2193 } 2194 } 2195 2196 Node* operation = VectorNode::make(opc, opd1, opd2, num_elem, elem_bt); 2197 if (is_masked_op && mask != nullptr) { 2198 if (use_predicate) { 2199 operation->add_req(mask); 2200 operation->add_flag(Node::Flag_is_predicated_vector); 2201 } else { 2202 operation = gvn().transform(operation); 2203 operation = new VectorBlendNode(opd1, operation, mask); 2204 } 2205 } 2206 operation = gvn().transform(operation); 2207 Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem); 2208 set_result(vbox); 2209 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2210 return true; 2211 } 2212 2213 // public static <VOUT extends VectorPayload, 2214 // VIN extends VectorPayload, 2215 // S extends VectorSpecies> 2216 // VOUT convert(int oprId, 2217 // Class<?> fromVectorClass, Class<?> fromElementType, int fromVLen, 2218 // Class<?> toVectorClass, Class<?> toElementType, int toVLen, 2219 // VIN v, S s, 2220 // VectorConvertOp<VOUT, VIN, S> defaultImpl) 2221 // 2222 bool LibraryCallKit::inline_vector_convert() { 2223 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 2224 2225 const TypeInstPtr* vector_klass_from = gvn().type(argument(1))->isa_instptr(); 2226 const TypeInstPtr* elem_klass_from = gvn().type(argument(2))->isa_instptr(); 2227 const TypeInt* vlen_from = gvn().type(argument(3))->isa_int(); 2228 2229 const TypeInstPtr* vector_klass_to = gvn().type(argument(4))->isa_instptr(); 2230 const TypeInstPtr* elem_klass_to = gvn().type(argument(5))->isa_instptr(); 2231 const TypeInt* vlen_to = gvn().type(argument(6))->isa_int(); 2232 2233 if (opr == nullptr || 2234 vector_klass_from == nullptr || elem_klass_from == nullptr || vlen_from == nullptr || 2235 vector_klass_to == nullptr || elem_klass_to == nullptr || vlen_to == nullptr) { 2236 return false; // dead code 2237 } 2238 if (!opr->is_con() || 2239 vector_klass_from->const_oop() == nullptr || elem_klass_from->const_oop() == nullptr || !vlen_from->is_con() || 2240 vector_klass_to->const_oop() == nullptr || elem_klass_to->const_oop() == nullptr || !vlen_to->is_con()) { 2241 log_if_needed(" ** missing constant: opr=%s vclass_from=%s etype_from=%s vlen_from=%s vclass_to=%s etype_to=%s vlen_to=%s", 2242 NodeClassNames[argument(0)->Opcode()], 2243 NodeClassNames[argument(1)->Opcode()], 2244 NodeClassNames[argument(2)->Opcode()], 2245 NodeClassNames[argument(3)->Opcode()], 2246 NodeClassNames[argument(4)->Opcode()], 2247 NodeClassNames[argument(5)->Opcode()], 2248 NodeClassNames[argument(6)->Opcode()]); 2249 return false; // not enough info for intrinsification 2250 } 2251 if (!is_klass_initialized(vector_klass_from) || !is_klass_initialized(vector_klass_to)) { 2252 log_if_needed(" ** klass argument not initialized"); 2253 return false; 2254 } 2255 2256 assert(opr->get_con() == VectorSupport::VECTOR_OP_CAST || 2257 opr->get_con() == VectorSupport::VECTOR_OP_UCAST || 2258 opr->get_con() == VectorSupport::VECTOR_OP_REINTERPRET, "wrong opcode"); 2259 bool is_cast = (opr->get_con() == VectorSupport::VECTOR_OP_CAST || opr->get_con() == VectorSupport::VECTOR_OP_UCAST); 2260 bool is_ucast = (opr->get_con() == VectorSupport::VECTOR_OP_UCAST); 2261 2262 ciKlass* vbox_klass_from = vector_klass_from->const_oop()->as_instance()->java_lang_Class_klass(); 2263 ciKlass* vbox_klass_to = vector_klass_to->const_oop()->as_instance()->java_lang_Class_klass(); 2264 2265 bool is_mask = is_vector_mask(vbox_klass_from); 2266 2267 ciType* elem_type_from = elem_klass_from->const_oop()->as_instance()->java_mirror_type(); 2268 if (!elem_type_from->is_primitive_type()) { 2269 return false; // should be primitive type 2270 } 2271 BasicType elem_bt_from = elem_type_from->basic_type(); 2272 ciType* elem_type_to = elem_klass_to->const_oop()->as_instance()->java_mirror_type(); 2273 if (!elem_type_to->is_primitive_type()) { 2274 return false; // should be primitive type 2275 } 2276 BasicType elem_bt_to = elem_type_to->basic_type(); 2277 2278 int num_elem_from = vlen_from->get_con(); 2279 int num_elem_to = vlen_to->get_con(); 2280 2281 // Check whether we can unbox to appropriate size. Even with casting, checking for reinterpret is needed 2282 // since we may need to change size. 2283 if (!arch_supports_vector(Op_VectorReinterpret, 2284 num_elem_from, 2285 elem_bt_from, 2286 is_mask ? VecMaskUseAll : VecMaskNotUsed)) { 2287 log_if_needed(" ** not supported: arity=1 op=%s/1 vlen1=%d etype1=%s ismask=%d", 2288 is_cast ? "cast" : "reinterpret", 2289 num_elem_from, type2name(elem_bt_from), is_mask); 2290 return false; 2291 } 2292 2293 // Check whether we can support resizing/reinterpreting to the new size. 2294 if (!arch_supports_vector(Op_VectorReinterpret, 2295 num_elem_to, 2296 elem_bt_to, 2297 is_mask ? VecMaskUseAll : VecMaskNotUsed)) { 2298 log_if_needed(" ** not supported: arity=1 op=%s/2 vlen2=%d etype2=%s ismask=%d", 2299 is_cast ? "cast" : "reinterpret", 2300 num_elem_to, type2name(elem_bt_to), is_mask); 2301 return false; 2302 } 2303 2304 // At this point, we know that both input and output vector registers are supported 2305 // by the architecture. Next check if the casted type is simply to same type - which means 2306 // that it is actually a resize and not a cast. 2307 if (is_cast && elem_bt_from == elem_bt_to) { 2308 is_cast = false; 2309 } 2310 2311 const TypeInstPtr* vbox_type_from = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass_from); 2312 2313 Node* opd1 = unbox_vector(argument(7), vbox_type_from, elem_bt_from, num_elem_from); 2314 if (opd1 == nullptr) { 2315 return false; 2316 } 2317 2318 const TypeVect* src_type = TypeVect::make(elem_bt_from, num_elem_from, is_mask); 2319 const TypeVect* dst_type = TypeVect::make(elem_bt_to, num_elem_to, is_mask); 2320 2321 // Safety check to prevent casting if source mask is of type vector 2322 // and destination mask of type predicate vector and vice-versa. 2323 // From X86 standpoint, this case will only arise over KNL target, 2324 // where certain masks (depending on the species) are either propagated 2325 // through a vector or predicate register. 2326 if (is_mask && 2327 ((src_type->isa_vectmask() == nullptr && dst_type->isa_vectmask()) || 2328 (dst_type->isa_vectmask() == nullptr && src_type->isa_vectmask()))) { 2329 return false; 2330 } 2331 2332 Node* op = opd1; 2333 if (is_cast) { 2334 assert(!is_mask || num_elem_from == num_elem_to, "vector mask cast needs the same elem num"); 2335 int cast_vopc = VectorCastNode::opcode(-1, elem_bt_from, !is_ucast); 2336 2337 // Make sure that vector cast is implemented to particular type/size combination if it is 2338 // not a mask casting. 2339 if (!is_mask && !arch_supports_vector(cast_vopc, num_elem_to, elem_bt_to, VecMaskNotUsed)) { 2340 log_if_needed(" ** not supported: arity=1 op=cast#%d/3 vlen2=%d etype2=%s ismask=%d", 2341 cast_vopc, num_elem_to, type2name(elem_bt_to), is_mask); 2342 return false; 2343 } 2344 2345 if (num_elem_from < num_elem_to) { 2346 // Since input and output number of elements are not consistent, we need to make sure we 2347 // properly size. Thus, first make a cast that retains the number of elements from source. 2348 int num_elem_for_cast = num_elem_from; 2349 2350 // It is possible that arch does not support this intermediate vector size 2351 // TODO More complex logic required here to handle this corner case for the sizes. 2352 if (!arch_supports_vector(cast_vopc, num_elem_for_cast, elem_bt_to, VecMaskNotUsed)) { 2353 log_if_needed(" ** not supported: arity=1 op=cast#%d/4 vlen1=%d etype2=%s ismask=%d", 2354 cast_vopc, 2355 num_elem_for_cast, type2name(elem_bt_to), is_mask); 2356 return false; 2357 } 2358 2359 op = gvn().transform(VectorCastNode::make(cast_vopc, op, elem_bt_to, num_elem_for_cast)); 2360 // Now ensure that the destination gets properly resized to needed size. 2361 op = gvn().transform(new VectorReinterpretNode(op, op->bottom_type()->is_vect(), dst_type)); 2362 } else if (num_elem_from > num_elem_to) { 2363 // Since number of elements from input is larger than output, simply reduce size of input 2364 // (we are supposed to drop top elements anyway). 2365 int num_elem_for_resize = num_elem_to; 2366 2367 // It is possible that arch does not support this intermediate vector size 2368 // TODO More complex logic required here to handle this corner case for the sizes. 2369 if (!arch_supports_vector(Op_VectorReinterpret, 2370 num_elem_for_resize, 2371 elem_bt_from, 2372 VecMaskNotUsed)) { 2373 log_if_needed(" ** not supported: arity=1 op=cast/5 vlen2=%d etype1=%s ismask=%d", 2374 num_elem_for_resize, type2name(elem_bt_from), is_mask); 2375 return false; 2376 } 2377 2378 const TypeVect* resize_type = TypeVect::make(elem_bt_from, num_elem_for_resize); 2379 op = gvn().transform(new VectorReinterpretNode(op, src_type, resize_type)); 2380 op = gvn().transform(VectorCastNode::make(cast_vopc, op, elem_bt_to, num_elem_to)); 2381 } else { // num_elem_from == num_elem_to 2382 if (is_mask) { 2383 // Make sure that cast for vector mask is implemented to particular type/size combination. 2384 if (!arch_supports_vector(Op_VectorMaskCast, num_elem_to, elem_bt_to, VecMaskNotUsed)) { 2385 log_if_needed(" ** not supported: arity=1 op=maskcast vlen2=%d etype2=%s ismask=%d", 2386 num_elem_to, type2name(elem_bt_to), is_mask); 2387 return false; 2388 } 2389 op = gvn().transform(new VectorMaskCastNode(op, dst_type)); 2390 } else { 2391 // Since input and output number of elements match, and since we know this vector size is 2392 // supported, simply do a cast with no resize needed. 2393 op = gvn().transform(VectorCastNode::make(cast_vopc, op, elem_bt_to, num_elem_to)); 2394 } 2395 } 2396 } else if (!Type::equals(src_type, dst_type)) { 2397 assert(!is_cast, "must be reinterpret"); 2398 op = gvn().transform(new VectorReinterpretNode(op, src_type, dst_type)); 2399 } 2400 2401 const TypeInstPtr* vbox_type_to = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass_to); 2402 Node* vbox = box_vector(op, vbox_type_to, elem_bt_to, num_elem_to); 2403 set_result(vbox); 2404 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem_to * type2aelembytes(elem_bt_to)))); 2405 return true; 2406 } 2407 2408 // public static 2409 // <V extends Vector<E>, 2410 // E> 2411 // V insert(Class<? extends V> vectorClass, Class<E> elementType, int vlen, 2412 // V vec, int ix, long val, 2413 // VecInsertOp<V> defaultImpl) 2414 bool LibraryCallKit::inline_vector_insert() { 2415 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 2416 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2417 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2418 const TypeInt* idx = gvn().type(argument(4))->isa_int(); 2419 2420 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || idx == nullptr) { 2421 return false; // dead code 2422 } 2423 if (vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con() || !idx->is_con()) { 2424 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s idx=%s", 2425 NodeClassNames[argument(0)->Opcode()], 2426 NodeClassNames[argument(1)->Opcode()], 2427 NodeClassNames[argument(2)->Opcode()], 2428 NodeClassNames[argument(4)->Opcode()]); 2429 return false; // not enough info for intrinsification 2430 } 2431 if (!is_klass_initialized(vector_klass)) { 2432 log_if_needed(" ** klass argument not initialized"); 2433 return false; 2434 } 2435 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2436 if (!elem_type->is_primitive_type()) { 2437 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2438 return false; // should be primitive type 2439 } 2440 BasicType elem_bt = elem_type->basic_type(); 2441 int num_elem = vlen->get_con(); 2442 if (!arch_supports_vector(Op_VectorInsert, num_elem, elem_bt, VecMaskNotUsed)) { 2443 log_if_needed(" ** not supported: arity=1 op=insert vlen=%d etype=%s ismask=no", 2444 num_elem, type2name(elem_bt)); 2445 return false; // not supported 2446 } 2447 2448 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2449 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2450 2451 Node* opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2452 if (opd == nullptr) { 2453 return false; 2454 } 2455 2456 Node* insert_val = argument(5); 2457 assert(gvn().type(insert_val)->isa_long() != nullptr, "expected to be long"); 2458 2459 // Convert insert value back to its appropriate type. 2460 switch (elem_bt) { 2461 case T_BYTE: 2462 insert_val = gvn().transform(new ConvL2INode(insert_val, TypeInt::BYTE)); 2463 break; 2464 case T_SHORT: 2465 insert_val = gvn().transform(new ConvL2INode(insert_val, TypeInt::SHORT)); 2466 break; 2467 case T_INT: 2468 insert_val = gvn().transform(new ConvL2INode(insert_val)); 2469 break; 2470 case T_FLOAT: 2471 insert_val = gvn().transform(new ConvL2INode(insert_val)); 2472 insert_val = gvn().transform(new MoveI2FNode(insert_val)); 2473 break; 2474 case T_DOUBLE: 2475 insert_val = gvn().transform(new MoveL2DNode(insert_val)); 2476 break; 2477 case T_LONG: 2478 // no conversion needed 2479 break; 2480 default: fatal("%s", type2name(elem_bt)); break; 2481 } 2482 2483 Node* operation = gvn().transform(VectorInsertNode::make(opd, insert_val, idx->get_con(), gvn())); 2484 2485 Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem); 2486 set_result(vbox); 2487 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2488 return true; 2489 } 2490 2491 // public static 2492 // <VM extends VectorPayload, 2493 // E> 2494 // long extract(Class<? extends VM> vClass, Class<E> eClass, 2495 // int length, 2496 // VM vm, int i, 2497 // VecExtractOp<VM> defaultImpl) 2498 bool LibraryCallKit::inline_vector_extract() { 2499 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 2500 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2501 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2502 const TypeInt* idx = gvn().type(argument(4))->isa_int(); 2503 2504 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || idx == nullptr) { 2505 return false; // dead code 2506 } 2507 if (vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con() || !idx->is_con()) { 2508 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s", 2509 NodeClassNames[argument(0)->Opcode()], 2510 NodeClassNames[argument(1)->Opcode()], 2511 NodeClassNames[argument(2)->Opcode()]); 2512 return false; // not enough info for intrinsification 2513 } 2514 if (!is_klass_initialized(vector_klass)) { 2515 log_if_needed(" ** klass argument not initialized"); 2516 return false; 2517 } 2518 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2519 if (!elem_type->is_primitive_type()) { 2520 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2521 return false; // should be primitive type 2522 } 2523 BasicType elem_bt = elem_type->basic_type(); 2524 int num_elem = vlen->get_con(); 2525 2526 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2527 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2528 2529 Node* opd = nullptr; 2530 2531 if (is_vector_mask(vbox_klass)) { 2532 // vbox_klass is mask. This is used for VectorMask.laneIsSet(int). 2533 2534 Node* pos = argument(4); // can be variable 2535 if (arch_supports_vector(Op_ExtractUB, num_elem, elem_bt, VecMaskUseAll)) { 2536 // Transform mask to vector with type of boolean and utilize ExtractUB node. 2537 opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2538 if (opd == nullptr) { 2539 return false; 2540 } 2541 opd = gvn().transform(VectorStoreMaskNode::make(gvn(), opd, elem_bt, num_elem)); 2542 opd = gvn().transform(new ExtractUBNode(opd, pos)); 2543 opd = gvn().transform(new ConvI2LNode(opd)); 2544 } else if (arch_supports_vector(Op_VectorMaskToLong, num_elem, elem_bt, VecMaskUseLoad)) { 2545 opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2546 if (opd == nullptr) { 2547 return false; 2548 } 2549 // VectorMaskToLongNode requires the input is either a mask or a vector with BOOLEAN type. 2550 if (opd->bottom_type()->isa_vectmask() == nullptr) { 2551 opd = gvn().transform(VectorStoreMaskNode::make(gvn(), opd, elem_bt, num_elem)); 2552 } 2553 // ((toLong() >>> pos) & 1L 2554 opd = gvn().transform(new VectorMaskToLongNode(opd, TypeLong::LONG)); 2555 opd = gvn().transform(new URShiftLNode(opd, pos)); 2556 opd = gvn().transform(new AndLNode(opd, gvn().makecon(TypeLong::ONE))); 2557 } else { 2558 log_if_needed(" ** Rejected mask extraction because architecture does not support it"); 2559 return false; // not supported 2560 } 2561 } else { 2562 // vbox_klass is vector. This is used for Vector.lane(int). 2563 if (!idx->is_con()) { 2564 log_if_needed(" ** missing constant: idx=%s", NodeClassNames[argument(4)->Opcode()]); 2565 return false; // not enough info for intrinsification 2566 } 2567 2568 int vopc = ExtractNode::opcode(elem_bt); 2569 if (!arch_supports_vector(vopc, num_elem, elem_bt, VecMaskNotUsed)) { 2570 log_if_needed(" ** not supported: arity=1 op=extract vlen=%d etype=%s ismask=no", 2571 num_elem, type2name(elem_bt)); 2572 return false; // not supported 2573 } 2574 2575 opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2576 if (opd == nullptr) { 2577 return false; 2578 } 2579 ConINode* idx_con = gvn().intcon(idx->get_con())->as_ConI(); 2580 2581 opd = gvn().transform(ExtractNode::make(opd, idx_con, elem_bt)); 2582 switch (elem_bt) { 2583 case T_BYTE: 2584 case T_SHORT: 2585 case T_INT: { 2586 opd = gvn().transform(new ConvI2LNode(opd)); 2587 break; 2588 } 2589 case T_FLOAT: { 2590 opd = gvn().transform(new MoveF2INode(opd)); 2591 opd = gvn().transform(new ConvI2LNode(opd)); 2592 break; 2593 } 2594 case T_DOUBLE: { 2595 opd = gvn().transform(new MoveD2LNode(opd)); 2596 break; 2597 } 2598 case T_LONG: { 2599 // no conversion needed 2600 break; 2601 } 2602 default: fatal("%s", type2name(elem_bt)); 2603 } 2604 } 2605 set_result(opd); 2606 return true; 2607 } 2608 2609 static Node* LowerSelectFromTwoVectorOperation(PhaseGVN& phase, Node* index_vec, Node* src1, Node* src2, const TypeVect* vt) { 2610 int num_elem = vt->length(); 2611 BasicType elem_bt = vt->element_basic_type(); 2612 2613 // Lower selectFrom operation into its constituent operations. 2614 // SelectFromTwoVectorNode = 2615 // (VectorBlend 2616 // (VectorRearrange SRC1 (WRAPED_INDEX AND (VLEN-1)) 2617 // (VectorRearrange SRC2 (WRAPED_INDEX AND (VLEN-1)) 2618 // MASK) 2619 // Where 2620 // WRAPED_INDEX are computed by wrapping incoming indexes 2621 // to two vector index range [0, VLEN*2) and 2622 // MASK = WRAPED_INDEX < VLEN 2623 // 2624 // IR lowering prevents intrinsification failure and associated argument 2625 // boxing penalties. 2626 // 2627 2628 BasicType shuffle_bt = elem_bt; 2629 if (shuffle_bt == T_FLOAT) { 2630 shuffle_bt = T_INT; 2631 } else if (shuffle_bt == T_DOUBLE) { 2632 shuffle_bt = T_LONG; 2633 } 2634 const TypeVect* st = TypeVect::make(shuffle_bt, num_elem); 2635 2636 // Cast index vector to the corresponding bit type 2637 if (elem_bt != shuffle_bt) { 2638 int cast_vopc = VectorCastNode::opcode(0, elem_bt, true); 2639 index_vec = phase.transform(VectorCastNode::make(cast_vopc, index_vec, shuffle_bt, num_elem)); 2640 } 2641 2642 // Wrap indexes into two vector index range [0, VLEN * 2) 2643 Node* two_vect_lane_cnt_m1 = phase.makecon(TypeInteger::make(2 * num_elem - 1, 2 * num_elem - 1, Type::WidenMin, shuffle_bt == T_LONG ? T_LONG : T_INT)); 2644 Node* bcast_two_vect_lane_cnt_m1_vec = phase.transform(VectorNode::scalar2vector(two_vect_lane_cnt_m1, num_elem, 2645 shuffle_bt, false)); 2646 index_vec = phase.transform(VectorNode::make(Op_AndV, index_vec, bcast_two_vect_lane_cnt_m1_vec, st)); 2647 2648 // Compute the blend mask for merging two independently permitted vectors 2649 // using shuffle index in two vector index range [0, VLEN * 2). 2650 BoolTest::mask pred = BoolTest::le; 2651 ConINode* pred_node = phase.makecon(TypeInt::make(pred))->as_ConI(); 2652 const TypeVect* vmask_type = TypeVect::makemask(shuffle_bt, num_elem); 2653 Node* lane_cnt_m1 = phase.makecon(TypeInteger::make(num_elem - 1, num_elem - 1, Type::WidenMin, shuffle_bt == T_LONG ? T_LONG : T_INT)); 2654 Node* bcast_lane_cnt_m1_vec = phase.transform(VectorNode::scalar2vector(lane_cnt_m1, num_elem, shuffle_bt, false)); 2655 Node* mask = phase.transform(new VectorMaskCmpNode(pred, index_vec, bcast_lane_cnt_m1_vec, pred_node, vmask_type)); 2656 2657 // Rearrange expects the indexes to lie within single vector index range [0, VLEN). 2658 Node* wrapped_index_vec = phase.transform(VectorNode::make(Op_AndV, index_vec, bcast_lane_cnt_m1_vec, st)); 2659 2660 // Load indexes from byte vector and appropriately transform them to target 2661 // specific permutation index format. 2662 if (Matcher::vector_rearrange_requires_load_shuffle(shuffle_bt, num_elem)) { 2663 wrapped_index_vec = phase.transform(new VectorLoadShuffleNode(wrapped_index_vec, st)); 2664 } 2665 2666 vmask_type = TypeVect::makemask(elem_bt, num_elem); 2667 mask = phase.transform(new VectorMaskCastNode(mask, vmask_type)); 2668 2669 Node* p1 = phase.transform(new VectorRearrangeNode(src1, wrapped_index_vec)); 2670 Node* p2 = phase.transform(new VectorRearrangeNode(src2, wrapped_index_vec)); 2671 2672 return new VectorBlendNode(p2, p1, mask); 2673 } 2674 2675 // public static 2676 // <V extends Vector<E>, 2677 // E> 2678 // V selectFromTwoVectorOp(Class<? extends V> vClass, Class<E> eClass, int length, 2679 // V v1, V v2, V v3, 2680 // SelectFromTwoVector<V> defaultImpl) 2681 bool LibraryCallKit::inline_vector_select_from_two_vectors() { 2682 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 2683 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2684 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2685 2686 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || vector_klass->const_oop() == nullptr || 2687 elem_klass->const_oop() == nullptr ||!vlen->is_con()) { 2688 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s", 2689 NodeClassNames[argument(0)->Opcode()], 2690 NodeClassNames[argument(1)->Opcode()], 2691 NodeClassNames[argument(2)->Opcode()]); 2692 return false; // not enough info for intrinsification 2693 } 2694 2695 if (!is_klass_initialized(vector_klass)) { 2696 log_if_needed(" ** klass argument not initialized"); 2697 return false; 2698 } 2699 2700 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2701 if (!elem_type->is_primitive_type()) { 2702 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2703 return false; // should be primitive type 2704 } 2705 2706 int num_elem = vlen->get_con(); 2707 if (!is_power_of_2(num_elem)) { 2708 log_if_needed(" ** vlen is not power of two=%d", num_elem); 2709 return false; 2710 } 2711 2712 BasicType elem_bt = elem_type->basic_type(); 2713 BasicType index_elem_bt = elem_bt; 2714 if (elem_bt == T_FLOAT) { 2715 index_elem_bt = T_INT; 2716 } else if (elem_bt == T_DOUBLE) { 2717 index_elem_bt = T_LONG; 2718 } 2719 2720 // Check if the platform requires a VectorLoadShuffle node to be generated 2721 bool need_load_shuffle = Matcher::vector_rearrange_requires_load_shuffle(index_elem_bt, num_elem); 2722 2723 bool lowerSelectFromOp = false; 2724 if (!arch_supports_vector(Op_SelectFromTwoVector, num_elem, elem_bt, VecMaskNotUsed)) { 2725 int cast_vopc = VectorCastNode::opcode(-1, elem_bt, true); 2726 if ((elem_bt != index_elem_bt && !arch_supports_vector(cast_vopc, num_elem, index_elem_bt, VecMaskNotUsed)) || 2727 !arch_supports_vector(Op_VectorMaskCmp, num_elem, index_elem_bt, VecMaskNotUsed) || 2728 !arch_supports_vector(Op_AndV, num_elem, index_elem_bt, VecMaskNotUsed) || 2729 !arch_supports_vector(Op_VectorMaskCast, num_elem, elem_bt, VecMaskNotUsed) || 2730 !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad) || 2731 !arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, VecMaskNotUsed) || 2732 (need_load_shuffle && !arch_supports_vector(Op_VectorLoadShuffle, num_elem, index_elem_bt, VecMaskNotUsed)) || 2733 !arch_supports_vector(Op_Replicate, num_elem, index_elem_bt, VecMaskNotUsed)) { 2734 log_if_needed(" ** not supported: opc=%d vlen=%d etype=%s ismask=useload", 2735 Op_SelectFromTwoVector, num_elem, type2name(elem_bt)); 2736 return false; // not supported 2737 } 2738 lowerSelectFromOp = true; 2739 } 2740 2741 int cast_vopc = VectorCastNode::opcode(-1, elem_bt, true); 2742 if (!lowerSelectFromOp) { 2743 if (!arch_supports_vector(Op_AndV, num_elem, index_elem_bt, VecMaskNotUsed) || 2744 !arch_supports_vector(Op_Replicate, num_elem, index_elem_bt, VecMaskNotUsed) || 2745 (is_floating_point_type(elem_bt) && 2746 !arch_supports_vector(cast_vopc, num_elem, index_elem_bt, VecMaskNotUsed))) { 2747 log_if_needed(" ** index wrapping not supported: vlen=%d etype=%s" , 2748 num_elem, type2name(elem_bt)); 2749 return false; // not supported 2750 } 2751 } 2752 2753 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2754 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2755 2756 Node* opd1 = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2757 if (opd1 == nullptr) { 2758 log_if_needed(" ** unbox failed v1=%s", 2759 NodeClassNames[argument(3)->Opcode()]); 2760 return false; 2761 } 2762 Node* opd2 = unbox_vector(argument(4), vbox_type, elem_bt, num_elem); 2763 if (opd2 == nullptr) { 2764 log_if_needed(" ** unbox failed v2=%s", 2765 NodeClassNames[argument(4)->Opcode()]); 2766 return false; 2767 } 2768 Node* opd3 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 2769 if (opd3 == nullptr) { 2770 log_if_needed(" ** unbox failed v3=%s", 2771 NodeClassNames[argument(5)->Opcode()]); 2772 return false; 2773 } 2774 2775 const TypeVect* vt = TypeVect::make(elem_bt, num_elem); 2776 2777 Node* operation = nullptr; 2778 if (lowerSelectFromOp) { 2779 operation = gvn().transform(LowerSelectFromTwoVectorOperation(gvn(), opd1, opd2, opd3, vt)); 2780 } else { 2781 if (index_elem_bt != elem_bt) { 2782 opd1 = gvn().transform(VectorCastNode::make(cast_vopc, opd1, index_elem_bt, num_elem)); 2783 } 2784 int indexRangeMask = 2 * num_elem - 1; 2785 Node* wrap_mask = gvn().makecon(TypeInteger::make(indexRangeMask, indexRangeMask, Type::WidenMin, index_elem_bt != T_LONG ? T_INT : index_elem_bt)); 2786 Node* wrap_mask_vec = gvn().transform(VectorNode::scalar2vector(wrap_mask, num_elem, index_elem_bt, false)); 2787 opd1 = gvn().transform(VectorNode::make(Op_AndV, opd1, wrap_mask_vec, opd1->bottom_type()->is_vect())); 2788 operation = gvn().transform(VectorNode::make(Op_SelectFromTwoVector, opd1, opd2, opd3, vt)); 2789 } 2790 2791 // Wrap it up in VectorBox to keep object type information. 2792 Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem); 2793 set_result(vbox); 2794 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2795 return true; 2796 } 2797 2798 // public static 2799 // <V extends Vector<E>, 2800 // M extends VectorMask<E>, 2801 // E> 2802 // V compressExpandOp(int opr, 2803 // Class<? extends V> vClass, Class<? extends M> mClass, Class<E> eClass, 2804 // int length, V v, M m, 2805 // CompressExpandOperation<V, M> defaultImpl) 2806 bool LibraryCallKit::inline_vector_compress_expand() { 2807 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 2808 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 2809 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 2810 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 2811 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 2812 2813 if (opr == nullptr || !opr->is_con() || 2814 vector_klass == nullptr || vector_klass->const_oop() == nullptr || 2815 mask_klass == nullptr || mask_klass->const_oop() == nullptr || 2816 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 2817 vlen == nullptr || !vlen->is_con()) { 2818 log_if_needed(" ** missing constant: opr=%s vclass=%s mclass=%s etype=%s vlen=%s", 2819 NodeClassNames[argument(0)->Opcode()], 2820 NodeClassNames[argument(1)->Opcode()], 2821 NodeClassNames[argument(2)->Opcode()], 2822 NodeClassNames[argument(3)->Opcode()], 2823 NodeClassNames[argument(4)->Opcode()]); 2824 return false; // not enough info for intrinsification 2825 } 2826 2827 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(mask_klass)) { 2828 log_if_needed(" ** klass argument not initialized"); 2829 return false; 2830 } 2831 2832 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2833 if (!elem_type->is_primitive_type()) { 2834 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2835 return false; // should be primitive type 2836 } 2837 2838 int num_elem = vlen->get_con(); 2839 BasicType elem_bt = elem_type->basic_type(); 2840 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 2841 2842 if (!arch_supports_vector(opc, num_elem, elem_bt, VecMaskUseLoad)) { 2843 log_if_needed(" ** not supported: opc=%d vlen=%d etype=%s ismask=useload", 2844 opc, num_elem, type2name(elem_bt)); 2845 return false; // not supported 2846 } 2847 2848 Node* opd1 = nullptr; 2849 const TypeInstPtr* vbox_type = nullptr; 2850 if (opc != Op_CompressM) { 2851 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2852 vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2853 opd1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 2854 if (opd1 == nullptr) { 2855 log_if_needed(" ** unbox failed vector=%s", 2856 NodeClassNames[argument(5)->Opcode()]); 2857 return false; 2858 } 2859 } 2860 2861 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2862 assert(is_vector_mask(mbox_klass), "argument(6) should be a mask class"); 2863 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 2864 2865 Node* mask = unbox_vector(argument(6), mbox_type, elem_bt, num_elem); 2866 if (mask == nullptr) { 2867 log_if_needed(" ** unbox failed mask=%s", 2868 NodeClassNames[argument(6)->Opcode()]); 2869 return false; 2870 } 2871 2872 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, opc == Op_CompressM); 2873 Node* operation = gvn().transform(VectorNode::make(opc, opd1, mask, vt)); 2874 2875 // Wrap it up in VectorBox to keep object type information. 2876 const TypeInstPtr* box_type = opc == Op_CompressM ? mbox_type : vbox_type; 2877 Node* vbox = box_vector(operation, box_type, elem_bt, num_elem); 2878 set_result(vbox); 2879 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2880 return true; 2881 } 2882 2883 // public static 2884 // <V extends Vector<E>, 2885 // E, 2886 // S extends VectorSpecies<E>> 2887 // V indexVector(Class<? extends V> vClass, Class<E> eClass, 2888 // int length, 2889 // V v, int step, S s, 2890 // IndexOperation<V, S> defaultImpl) 2891 bool LibraryCallKit::inline_index_vector() { 2892 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 2893 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2894 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2895 2896 if (vector_klass == nullptr || vector_klass->const_oop() == nullptr || 2897 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 2898 vlen == nullptr || !vlen->is_con() ) { 2899 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s", 2900 NodeClassNames[argument(0)->Opcode()], 2901 NodeClassNames[argument(1)->Opcode()], 2902 NodeClassNames[argument(2)->Opcode()]); 2903 return false; // not enough info for intrinsification 2904 } 2905 2906 if (!is_klass_initialized(vector_klass)) { 2907 log_if_needed(" ** klass argument not initialized"); 2908 return false; 2909 } 2910 2911 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2912 if (!elem_type->is_primitive_type()) { 2913 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2914 return false; // should be primitive type 2915 } 2916 2917 int num_elem = vlen->get_con(); 2918 BasicType elem_bt = elem_type->basic_type(); 2919 2920 // Check whether the iota index generation op is supported by the current hardware 2921 if (!arch_supports_vector(Op_VectorLoadConst, num_elem, elem_bt, VecMaskNotUsed)) { 2922 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 2923 return false; // not supported 2924 } 2925 2926 int mul_op = VectorSupport::vop2ideal(VectorSupport::VECTOR_OP_MUL, elem_bt); 2927 int vmul_op = VectorNode::opcode(mul_op, elem_bt); 2928 bool needs_mul = true; 2929 Node* scale = argument(4); 2930 const TypeInt* scale_type = gvn().type(scale)->isa_int(); 2931 // Multiply is not needed if the scale is a constant "1". 2932 if (scale_type && scale_type->is_con() && scale_type->get_con() == 1) { 2933 needs_mul = false; 2934 } else { 2935 // Check whether the vector multiply op is supported by the current hardware 2936 if (!arch_supports_vector(vmul_op, num_elem, elem_bt, VecMaskNotUsed)) { 2937 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 2938 return false; // not supported 2939 } 2940 2941 // Check whether the scalar cast op is supported by the current hardware 2942 if (is_floating_point_type(elem_bt) || elem_bt == T_LONG) { 2943 int cast_op = elem_bt == T_LONG ? Op_ConvI2L : 2944 elem_bt == T_FLOAT? Op_ConvI2F : Op_ConvI2D; 2945 if (!Matcher::match_rule_supported(cast_op)) { 2946 log_if_needed(" ** Rejected op (%s) because architecture does not support it", 2947 NodeClassNames[cast_op]); 2948 return false; // not supported 2949 } 2950 } 2951 } 2952 2953 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2954 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2955 Node* opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2956 if (opd == nullptr) { 2957 log_if_needed(" ** unbox failed vector=%s", 2958 NodeClassNames[argument(3)->Opcode()]); 2959 return false; 2960 } 2961 2962 int add_op = VectorSupport::vop2ideal(VectorSupport::VECTOR_OP_ADD, elem_bt); 2963 int vadd_op = VectorNode::opcode(add_op, elem_bt); 2964 bool needs_add = true; 2965 // The addition is not needed if all the element values of "opd" are zero 2966 if (VectorNode::is_all_zeros_vector(opd)) { 2967 needs_add = false; 2968 } else { 2969 // Check whether the vector addition op is supported by the current hardware 2970 if (!arch_supports_vector(vadd_op, num_elem, elem_bt, VecMaskNotUsed)) { 2971 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 2972 return false; // not supported 2973 } 2974 } 2975 2976 // Compute the iota indice vector 2977 const TypeVect* vt = TypeVect::make(elem_bt, num_elem); 2978 Node* index = gvn().transform(new VectorLoadConstNode(gvn().makecon(TypeInt::ZERO), vt)); 2979 2980 // Broadcast the "scale" to a vector, and multiply the "scale" with iota indice vector. 2981 if (needs_mul) { 2982 switch (elem_bt) { 2983 case T_BOOLEAN: // fall-through 2984 case T_BYTE: // fall-through 2985 case T_SHORT: // fall-through 2986 case T_CHAR: // fall-through 2987 case T_INT: { 2988 // no conversion needed 2989 break; 2990 } 2991 case T_LONG: { 2992 scale = gvn().transform(new ConvI2LNode(scale)); 2993 break; 2994 } 2995 case T_FLOAT: { 2996 scale = gvn().transform(new ConvI2FNode(scale)); 2997 break; 2998 } 2999 case T_DOUBLE: { 3000 scale = gvn().transform(new ConvI2DNode(scale)); 3001 break; 3002 } 3003 default: fatal("%s", type2name(elem_bt)); 3004 } 3005 scale = gvn().transform(VectorNode::scalar2vector(scale, num_elem, elem_bt)); 3006 index = gvn().transform(VectorNode::make(vmul_op, index, scale, vt)); 3007 } 3008 3009 // Add "opd" if addition is needed. 3010 if (needs_add) { 3011 index = gvn().transform(VectorNode::make(vadd_op, opd, index, vt)); 3012 } 3013 Node* vbox = box_vector(index, vbox_type, elem_bt, num_elem); 3014 set_result(vbox); 3015 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 3016 return true; 3017 } 3018 3019 // public static 3020 // <E, 3021 // M extends VectorMask<E>> 3022 // M indexPartiallyInUpperRange(Class<? extends M> mClass, Class<E> eClass, int length, 3023 // long offset, long limit, 3024 // IndexPartiallyInUpperRangeOperation<E, M> defaultImpl) 3025 bool LibraryCallKit::inline_index_partially_in_upper_range() { 3026 const TypeInstPtr* mask_klass = gvn().type(argument(0))->isa_instptr(); 3027 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 3028 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 3029 3030 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr || 3031 elem_klass == nullptr || elem_klass->const_oop() == nullptr || 3032 vlen == nullptr || !vlen->is_con()) { 3033 log_if_needed(" ** missing constant: mclass=%s etype=%s vlen=%s", 3034 NodeClassNames[argument(0)->Opcode()], 3035 NodeClassNames[argument(1)->Opcode()], 3036 NodeClassNames[argument(2)->Opcode()]); 3037 return false; // not enough info for intrinsification 3038 } 3039 3040 if (!is_klass_initialized(mask_klass)) { 3041 log_if_needed(" ** klass argument not initialized"); 3042 return false; 3043 } 3044 3045 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 3046 if (!elem_type->is_primitive_type()) { 3047 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 3048 return false; // should be primitive type 3049 } 3050 3051 int num_elem = vlen->get_con(); 3052 BasicType elem_bt = elem_type->basic_type(); 3053 3054 // Check whether the necessary ops are supported by current hardware. 3055 bool supports_mask_gen = arch_supports_vector(Op_VectorMaskGen, num_elem, elem_bt, VecMaskUseStore); 3056 if (!supports_mask_gen) { 3057 if (!arch_supports_vector(Op_VectorLoadConst, num_elem, elem_bt, VecMaskNotUsed) || 3058 !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed) || 3059 !arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskUseStore)) { 3060 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 3061 return false; // not supported 3062 } 3063 3064 // Check whether the scalar cast operation is supported by current hardware. 3065 if (elem_bt != T_LONG) { 3066 int cast_op = is_integral_type(elem_bt) ? Op_ConvL2I 3067 : (elem_bt == T_FLOAT ? Op_ConvL2F : Op_ConvL2D); 3068 if (!Matcher::match_rule_supported(cast_op)) { 3069 log_if_needed(" ** Rejected op (%s) because architecture does not support it", 3070 NodeClassNames[cast_op]); 3071 return false; // not supported 3072 } 3073 } 3074 } 3075 3076 Node* offset = argument(3); 3077 Node* limit = argument(5); 3078 if (offset == nullptr || limit == nullptr) { 3079 log_if_needed(" ** offset or limit argument is null"); 3080 return false; // not supported 3081 } 3082 3083 ciKlass* box_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 3084 assert(is_vector_mask(box_klass), "argument(0) should be a mask class"); 3085 const TypeInstPtr* box_type = TypeInstPtr::make_exact(TypePtr::NotNull, box_klass); 3086 3087 // We assume "offset > 0 && limit >= offset && limit - offset < num_elem". 3088 // So directly get indexLimit with "indexLimit = limit - offset". 3089 Node* indexLimit = gvn().transform(new SubLNode(limit, offset)); 3090 Node* mask = nullptr; 3091 if (supports_mask_gen) { 3092 mask = gvn().transform(VectorMaskGenNode::make(indexLimit, elem_bt, num_elem)); 3093 } else { 3094 // Generate the vector mask based on "mask = iota < indexLimit". 3095 // Broadcast "indexLimit" to a vector. 3096 switch (elem_bt) { 3097 case T_BOOLEAN: // fall-through 3098 case T_BYTE: // fall-through 3099 case T_SHORT: // fall-through 3100 case T_CHAR: // fall-through 3101 case T_INT: { 3102 indexLimit = gvn().transform(new ConvL2INode(indexLimit)); 3103 break; 3104 } 3105 case T_DOUBLE: { 3106 indexLimit = gvn().transform(new ConvL2DNode(indexLimit)); 3107 break; 3108 } 3109 case T_FLOAT: { 3110 indexLimit = gvn().transform(new ConvL2FNode(indexLimit)); 3111 break; 3112 } 3113 case T_LONG: { 3114 // no conversion needed 3115 break; 3116 } 3117 default: fatal("%s", type2name(elem_bt)); 3118 } 3119 indexLimit = gvn().transform(VectorNode::scalar2vector(indexLimit, num_elem, elem_bt)); 3120 3121 // Load the "iota" vector. 3122 const TypeVect* vt = TypeVect::make(elem_bt, num_elem); 3123 Node* iota = gvn().transform(new VectorLoadConstNode(gvn().makecon(TypeInt::ZERO), vt)); 3124 3125 // Compute the vector mask with "mask = iota < indexLimit". 3126 ConINode* pred_node = (ConINode*)gvn().makecon(TypeInt::make(BoolTest::lt)); 3127 const TypeVect* vmask_type = TypeVect::makemask(elem_bt, num_elem); 3128 mask = gvn().transform(new VectorMaskCmpNode(BoolTest::lt, iota, indexLimit, pred_node, vmask_type)); 3129 } 3130 Node* vbox = box_vector(mask, box_type, elem_bt, num_elem); 3131 set_result(vbox); 3132 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 3133 return true; 3134 } 3135 3136 #undef non_product_log_if_needed 3137 #undef log_if_needed