1 /* 2 * Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "precompiled.hpp" 26 #include "ci/ciSymbols.hpp" 27 #include "classfile/vmSymbols.hpp" 28 #include "opto/library_call.hpp" 29 #include "opto/runtime.hpp" 30 #include "opto/vectornode.hpp" 31 #include "prims/vectorSupport.hpp" 32 #include "runtime/stubRoutines.hpp" 33 34 #ifdef ASSERT 35 static bool is_vector(ciKlass* klass) { 36 return klass->is_subclass_of(ciEnv::current()->vector_VectorPayload_klass()); 37 } 38 39 static bool check_vbox(const TypeInstPtr* vbox_type) { 40 assert(vbox_type->klass_is_exact(), ""); 41 42 ciInstanceKlass* ik = vbox_type->instance_klass(); 43 assert(is_vector(ik), "not a vector"); 44 45 ciField* fd1 = ik->get_field_by_name(ciSymbols::ETYPE_name(), ciSymbols::class_signature(), /* is_static */ true); 46 assert(fd1 != nullptr, "element type info is missing"); 47 48 ciConstant val1 = fd1->constant_value(); 49 BasicType elem_bt = val1.as_object()->as_instance()->java_mirror_type()->basic_type(); 50 assert(is_java_primitive(elem_bt), "element type info is missing"); 51 52 ciField* fd2 = ik->get_field_by_name(ciSymbols::VLENGTH_name(), ciSymbols::int_signature(), /* is_static */ true); 53 assert(fd2 != nullptr, "vector length info is missing"); 54 55 ciConstant val2 = fd2->constant_value(); 56 assert(val2.as_int() > 0, "vector length info is missing"); 57 58 return true; 59 } 60 #endif 61 62 #define log_if_needed(...) \ 63 if (C->print_intrinsics()) { \ 64 tty->print_cr(__VA_ARGS__); \ 65 } 66 67 #ifndef PRODUCT 68 #define non_product_log_if_needed(...) log_if_needed(__VA_ARGS__) 69 #else 70 #define non_product_log_if_needed(...) 71 #endif 72 73 static bool is_vector_mask(ciKlass* klass) { 74 return klass->is_subclass_of(ciEnv::current()->vector_VectorMask_klass()); 75 } 76 77 static bool is_vector_shuffle(ciKlass* klass) { 78 return klass->is_subclass_of(ciEnv::current()->vector_VectorShuffle_klass()); 79 } 80 81 bool LibraryCallKit::arch_supports_vector_rotate(int opc, int num_elem, BasicType elem_bt, 82 VectorMaskUseType mask_use_type, bool has_scalar_args) { 83 bool is_supported = true; 84 85 // has_scalar_args flag is true only for non-constant scalar shift count, 86 // since in this case shift needs to be broadcasted. 87 if (!Matcher::match_rule_supported_vector(opc, num_elem, elem_bt) || 88 (has_scalar_args && !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed))) { 89 is_supported = false; 90 } 91 92 if (is_supported) { 93 // Check if mask unboxing is supported, this is a two step process which first loads the contents 94 // of boolean array into vector followed by either lane expansion to match the lane size of masked 95 // vector operation or populate the predicate register. 96 if ((mask_use_type & VecMaskUseLoad) != 0) { 97 if (!Matcher::match_rule_supported_vector(Op_VectorLoadMask, num_elem, elem_bt) || 98 !Matcher::match_rule_supported_vector(Op_LoadVector, num_elem, T_BOOLEAN)) { 99 non_product_log_if_needed(" ** Rejected vector mask loading (%s,%s,%d) because architecture does not support it", 100 NodeClassNames[Op_VectorLoadMask], type2name(elem_bt), num_elem); 101 return false; 102 } 103 } 104 105 if ((mask_use_type & VecMaskUsePred) != 0) { 106 if (!Matcher::has_predicated_vectors() || 107 !Matcher::match_rule_supported_vector_masked(opc, num_elem, elem_bt)) { 108 non_product_log_if_needed("Rejected vector mask predicate using (%s,%s,%d) because architecture does not support it", 109 NodeClassNames[opc], type2name(elem_bt), num_elem); 110 return false; 111 } 112 } 113 } 114 115 int lshiftopc, rshiftopc; 116 switch(elem_bt) { 117 case T_BYTE: 118 lshiftopc = Op_LShiftI; 119 rshiftopc = Op_URShiftB; 120 break; 121 case T_SHORT: 122 lshiftopc = Op_LShiftI; 123 rshiftopc = Op_URShiftS; 124 break; 125 case T_INT: 126 lshiftopc = Op_LShiftI; 127 rshiftopc = Op_URShiftI; 128 break; 129 case T_LONG: 130 lshiftopc = Op_LShiftL; 131 rshiftopc = Op_URShiftL; 132 break; 133 default: fatal("Unexpected type: %s", type2name(elem_bt)); 134 } 135 int lshiftvopc = VectorNode::opcode(lshiftopc, elem_bt); 136 int rshiftvopc = VectorNode::opcode(rshiftopc, elem_bt); 137 if (!is_supported && 138 arch_supports_vector(lshiftvopc, num_elem, elem_bt, VecMaskNotUsed, has_scalar_args) && 139 arch_supports_vector(rshiftvopc, num_elem, elem_bt, VecMaskNotUsed, has_scalar_args) && 140 arch_supports_vector(Op_OrV, num_elem, elem_bt, VecMaskNotUsed)) { 141 is_supported = true; 142 } 143 return is_supported; 144 } 145 146 Node* GraphKit::box_vector(Node* vector, const TypeInstPtr* vbox_type, BasicType elem_bt, int num_elem, bool deoptimize_on_exception) { 147 assert(EnableVectorSupport, ""); 148 149 PreserveReexecuteState preexecs(this); 150 jvms()->set_should_reexecute(true); 151 152 VectorBoxAllocateNode* alloc = new VectorBoxAllocateNode(C, vbox_type); 153 set_edges_for_java_call(alloc, /*must_throw=*/false, /*separate_io_proj=*/true); 154 make_slow_call_ex(alloc, env()->Throwable_klass(), /*separate_io_proj=*/true, deoptimize_on_exception); 155 set_i_o(gvn().transform( new ProjNode(alloc, TypeFunc::I_O) )); 156 set_all_memory(gvn().transform( new ProjNode(alloc, TypeFunc::Memory) )); 157 Node* ret = gvn().transform(new ProjNode(alloc, TypeFunc::Parms)); 158 159 assert(check_vbox(vbox_type), ""); 160 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, is_vector_mask(vbox_type->instance_klass())); 161 VectorBoxNode* vbox = new VectorBoxNode(C, ret, vector, vbox_type, vt); 162 return gvn().transform(vbox); 163 } 164 165 Node* GraphKit::unbox_vector(Node* v, const TypeInstPtr* vbox_type, BasicType elem_bt, int num_elem, bool shuffle_to_vector) { 166 assert(EnableVectorSupport, ""); 167 const TypeInstPtr* vbox_type_v = gvn().type(v)->isa_instptr(); 168 if (vbox_type_v == nullptr || vbox_type->instance_klass() != vbox_type_v->instance_klass()) { 169 return nullptr; // arguments don't agree on vector shapes 170 } 171 if (vbox_type_v->maybe_null()) { 172 return nullptr; // no nulls are allowed 173 } 174 assert(check_vbox(vbox_type), ""); 175 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, is_vector_mask(vbox_type->instance_klass())); 176 Node* unbox = gvn().transform(new VectorUnboxNode(C, vt, v, merged_memory(), shuffle_to_vector)); 177 return unbox; 178 } 179 180 Node* GraphKit::vector_shift_count(Node* cnt, int shift_op, BasicType bt, int num_elem) { 181 assert(bt == T_INT || bt == T_LONG || bt == T_SHORT || bt == T_BYTE, "byte, short, long and int are supported"); 182 juint mask = (type2aelembytes(bt) * BitsPerByte - 1); 183 Node* nmask = gvn().transform(ConNode::make(TypeInt::make(mask))); 184 Node* mcnt = gvn().transform(new AndINode(cnt, nmask)); 185 return gvn().transform(VectorNode::shift_count(shift_op, mcnt, num_elem, bt)); 186 } 187 188 bool LibraryCallKit::arch_supports_vector(int sopc, int num_elem, BasicType type, VectorMaskUseType mask_use_type, bool has_scalar_args) { 189 // Check that the operation is valid. 190 if (sopc <= 0) { 191 non_product_log_if_needed(" ** Rejected intrinsification because no valid vector op could be extracted"); 192 return false; 193 } 194 195 if (VectorNode::is_vector_rotate(sopc)) { 196 if(!arch_supports_vector_rotate(sopc, num_elem, type, mask_use_type, has_scalar_args)) { 197 non_product_log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support variable vector shifts", 198 NodeClassNames[sopc], type2name(type), num_elem); 199 return false; 200 } 201 } else if (VectorNode::is_vector_integral_negate(sopc)) { 202 if (!VectorNode::is_vector_integral_negate_supported(sopc, num_elem, type, false)) { 203 non_product_log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support integral vector negate", 204 NodeClassNames[sopc], type2name(type), num_elem); 205 return false; 206 } 207 } else { 208 // Check that architecture supports this op-size-type combination. 209 if (!Matcher::match_rule_supported_vector(sopc, num_elem, type)) { 210 non_product_log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support it", 211 NodeClassNames[sopc], type2name(type), num_elem); 212 return false; 213 } else { 214 assert(Matcher::match_rule_supported(sopc), "must be supported"); 215 } 216 } 217 218 if (num_elem == 1) { 219 if (mask_use_type != VecMaskNotUsed) { 220 non_product_log_if_needed(" ** Rejected vector mask op (%s,%s,%d) because architecture does not support it", 221 NodeClassNames[sopc], type2name(type), num_elem); 222 return false; 223 } 224 225 if (sopc != 0) { 226 if (sopc != Op_LoadVector && sopc != Op_StoreVector) { 227 non_product_log_if_needed(" ** Not a svml call or load/store vector op (%s,%s,%d)", 228 NodeClassNames[sopc], type2name(type), num_elem); 229 return false; 230 } 231 } 232 } 233 234 if (!has_scalar_args && VectorNode::is_vector_shift(sopc) && 235 Matcher::supports_vector_variable_shifts() == false) { 236 log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support variable vector shifts", 237 NodeClassNames[sopc], type2name(type), num_elem); 238 return false; 239 } 240 241 // Check if mask unboxing is supported, this is a two step process which first loads the contents 242 // of boolean array into vector followed by either lane expansion to match the lane size of masked 243 // vector operation or populate the predicate register. 244 if ((mask_use_type & VecMaskUseLoad) != 0) { 245 if (!Matcher::match_rule_supported_vector(Op_VectorLoadMask, num_elem, type) || 246 !Matcher::match_rule_supported_vector(Op_LoadVector, num_elem, T_BOOLEAN)) { 247 non_product_log_if_needed(" ** Rejected vector mask loading (%s,%s,%d) because architecture does not support it", 248 NodeClassNames[Op_VectorLoadMask], type2name(type), num_elem); 249 return false; 250 } 251 } 252 253 // Check if mask boxing is supported, this is a two step process which first stores the contents 254 // of mask vector / predicate register into a boolean vector followed by vector store operation to 255 // transfer the contents to underlined storage of mask boxes which is a boolean array. 256 if ((mask_use_type & VecMaskUseStore) != 0) { 257 if (!Matcher::match_rule_supported_vector(Op_VectorStoreMask, num_elem, type) || 258 !Matcher::match_rule_supported_vector(Op_StoreVector, num_elem, T_BOOLEAN)) { 259 non_product_log_if_needed("Rejected vector mask storing (%s,%s,%d) because architecture does not support it", 260 NodeClassNames[Op_VectorStoreMask], type2name(type), num_elem); 261 return false; 262 } 263 } 264 265 if ((mask_use_type & VecMaskUsePred) != 0) { 266 bool is_supported = false; 267 if (Matcher::has_predicated_vectors()) { 268 if (VectorNode::is_vector_integral_negate(sopc)) { 269 is_supported = VectorNode::is_vector_integral_negate_supported(sopc, num_elem, type, true); 270 } else { 271 is_supported = Matcher::match_rule_supported_vector_masked(sopc, num_elem, type); 272 } 273 } 274 is_supported |= Matcher::supports_vector_predicate_op_emulation(sopc, num_elem, type); 275 276 if (!is_supported) { 277 non_product_log_if_needed("Rejected vector mask predicate using (%s,%s,%d) because architecture does not support it", 278 NodeClassNames[sopc], type2name(type), num_elem); 279 return false; 280 } 281 } 282 283 return true; 284 } 285 286 static bool is_klass_initialized(const TypeInstPtr* vec_klass) { 287 if (vec_klass->const_oop() == nullptr) { 288 return false; // uninitialized or some kind of unsafe access 289 } 290 assert(vec_klass->const_oop()->as_instance()->java_lang_Class_klass() != nullptr, "klass instance expected"); 291 ciInstanceKlass* klass = vec_klass->const_oop()->as_instance()->java_lang_Class_klass()->as_instance_klass(); 292 return klass->is_initialized(); 293 } 294 295 // public static 296 // <V extends Vector<E>, 297 // M extends VectorMask<E>, 298 // E> 299 // V unaryOp(int oprId, Class<? extends V> vmClass, Class<? extends M> maskClass, Class<E> elementType, 300 // int length, V v, M m, 301 // UnaryOperation<V, M> defaultImpl) 302 // 303 // public static 304 // <V, 305 // M extends VectorMask<E>, 306 // E> 307 // V binaryOp(int oprId, Class<? extends V> vmClass, Class<? extends M> maskClass, Class<E> elementType, 308 // int length, V v1, V v2, M m, 309 // BinaryOperation<V, M> defaultImpl) 310 // 311 // public static 312 // <V extends Vector<E>, 313 // M extends VectorMask<E>, 314 // E> 315 // V ternaryOp(int oprId, Class<? extends V> vmClass, Class<? extends M> maskClass, Class<E> elementType, 316 // int length, V v1, V v2, V v3, M m, 317 // TernaryOperation<V, M> defaultImpl) 318 // 319 bool LibraryCallKit::inline_vector_nary_operation(int n) { 320 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 321 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 322 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 323 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 324 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 325 326 if (opr == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 327 !opr->is_con() || vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 328 log_if_needed(" ** missing constant: opr=%s vclass=%s etype=%s vlen=%s", 329 NodeClassNames[argument(0)->Opcode()], 330 NodeClassNames[argument(1)->Opcode()], 331 NodeClassNames[argument(3)->Opcode()], 332 NodeClassNames[argument(4)->Opcode()]); 333 return false; // not enough info for intrinsification 334 } 335 336 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 337 if (!elem_type->is_primitive_type()) { 338 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 339 return false; // should be primitive type 340 } 341 if (!is_klass_initialized(vector_klass)) { 342 log_if_needed(" ** klass argument not initialized"); 343 return false; 344 } 345 346 // "argument(n + 5)" should be the mask object. We assume it is "null" when no mask 347 // is used to control this operation. 348 const Type* vmask_type = gvn().type(argument(n + 5)); 349 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 350 if (is_masked_op) { 351 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 352 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(2)->Opcode()]); 353 return false; // not enough info for intrinsification 354 } 355 356 if (!is_klass_initialized(mask_klass)) { 357 log_if_needed(" ** mask klass argument not initialized"); 358 return false; 359 } 360 361 if (vmask_type->maybe_null()) { 362 log_if_needed(" ** null mask values are not allowed for masked op"); 363 return false; 364 } 365 } 366 367 BasicType elem_bt = elem_type->basic_type(); 368 int num_elem = vlen->get_con(); 369 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 370 int sopc = VectorNode::opcode(opc, elem_bt); 371 if ((opc != Op_CallLeafVector) && (sopc == 0)) { 372 log_if_needed(" ** operation not supported: opc=%s bt=%s", NodeClassNames[opc], type2name(elem_bt)); 373 return false; // operation not supported 374 } 375 if (num_elem == 1) { 376 if (opc != Op_CallLeafVector || elem_bt != T_DOUBLE) { 377 log_if_needed(" ** not a svml call: arity=%d opc=%d vlen=%d etype=%s", 378 n, opc, num_elem, type2name(elem_bt)); 379 return false; 380 } 381 } 382 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 383 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 384 385 if (is_vector_mask(vbox_klass)) { 386 assert(!is_masked_op, "mask operations do not need mask to control"); 387 } 388 389 if (opc == Op_CallLeafVector) { 390 if (!UseVectorStubs) { 391 log_if_needed(" ** vector stubs support is disabled"); 392 return false; 393 } 394 if (!Matcher::supports_vector_calling_convention()) { 395 log_if_needed(" ** no vector calling conventions supported"); 396 return false; 397 } 398 if (!Matcher::vector_size_supported(elem_bt, num_elem)) { 399 log_if_needed(" ** vector size (vlen=%d, etype=%s) is not supported", 400 num_elem, type2name(elem_bt)); 401 return false; 402 } 403 } 404 405 // When using mask, mask use type needs to be VecMaskUseLoad. 406 VectorMaskUseType mask_use_type = is_vector_mask(vbox_klass) ? VecMaskUseAll 407 : is_masked_op ? VecMaskUseLoad : VecMaskNotUsed; 408 if ((sopc != 0) && !arch_supports_vector(sopc, num_elem, elem_bt, mask_use_type)) { 409 log_if_needed(" ** not supported: arity=%d opc=%d vlen=%d etype=%s ismask=%d is_masked_op=%d", 410 n, sopc, num_elem, type2name(elem_bt), 411 is_vector_mask(vbox_klass) ? 1 : 0, is_masked_op ? 1 : 0); 412 return false; // not supported 413 } 414 415 // Return true if current platform has implemented the masked operation with predicate feature. 416 bool use_predicate = is_masked_op && sopc != 0 && arch_supports_vector(sopc, num_elem, elem_bt, VecMaskUsePred); 417 if (is_masked_op && !use_predicate && !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)) { 418 log_if_needed(" ** not supported: arity=%d opc=%d vlen=%d etype=%s ismask=0 is_masked_op=1", 419 n, sopc, num_elem, type2name(elem_bt)); 420 return false; 421 } 422 423 Node* opd1 = nullptr; Node* opd2 = nullptr; Node* opd3 = nullptr; 424 switch (n) { 425 case 3: { 426 opd3 = unbox_vector(argument(7), vbox_type, elem_bt, num_elem); 427 if (opd3 == nullptr) { 428 log_if_needed(" ** unbox failed v3=%s", 429 NodeClassNames[argument(7)->Opcode()]); 430 return false; 431 } 432 // fall-through 433 } 434 case 2: { 435 opd2 = unbox_vector(argument(6), vbox_type, elem_bt, num_elem); 436 if (opd2 == nullptr) { 437 log_if_needed(" ** unbox failed v2=%s", 438 NodeClassNames[argument(6)->Opcode()]); 439 return false; 440 } 441 // fall-through 442 } 443 case 1: { 444 opd1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 445 if (opd1 == nullptr) { 446 log_if_needed(" ** unbox failed v1=%s", 447 NodeClassNames[argument(5)->Opcode()]); 448 return false; 449 } 450 break; 451 } 452 default: fatal("unsupported arity: %d", n); 453 } 454 455 Node* mask = nullptr; 456 if (is_masked_op) { 457 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 458 assert(is_vector_mask(mbox_klass), "argument(2) should be a mask class"); 459 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 460 mask = unbox_vector(argument(n + 5), mbox_type, elem_bt, num_elem); 461 if (mask == nullptr) { 462 log_if_needed(" ** unbox failed mask=%s", 463 NodeClassNames[argument(n + 5)->Opcode()]); 464 return false; 465 } 466 } 467 468 Node* operation = nullptr; 469 if (opc == Op_CallLeafVector) { 470 assert(UseVectorStubs, "sanity"); 471 operation = gen_call_to_svml(opr->get_con(), elem_bt, num_elem, opd1, opd2); 472 if (operation == nullptr) { 473 log_if_needed(" ** svml call failed for %s_%s_%d", 474 (elem_bt == T_FLOAT)?"float":"double", 475 VectorSupport::svmlname[opr->get_con() - VectorSupport::VECTOR_OP_SVML_START], 476 num_elem * type2aelembytes(elem_bt)); 477 return false; 478 } 479 } else { 480 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, is_vector_mask(vbox_klass)); 481 switch (n) { 482 case 1: 483 case 2: { 484 operation = VectorNode::make(sopc, opd1, opd2, vt, is_vector_mask(vbox_klass), VectorNode::is_shift_opcode(opc)); 485 break; 486 } 487 case 3: { 488 operation = VectorNode::make(sopc, opd1, opd2, opd3, vt); 489 break; 490 } 491 default: fatal("unsupported arity: %d", n); 492 } 493 } 494 495 if (is_masked_op && mask != nullptr) { 496 if (use_predicate) { 497 operation->add_req(mask); 498 operation->add_flag(Node::Flag_is_predicated_vector); 499 } else { 500 operation->add_flag(Node::Flag_is_predicated_using_blend); 501 operation = gvn().transform(operation); 502 operation = new VectorBlendNode(opd1, operation, mask); 503 } 504 } 505 operation = gvn().transform(operation); 506 507 // Wrap it up in VectorBox to keep object type information. 508 Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem); 509 set_result(vbox); 510 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 511 return true; 512 } 513 514 // Following routine generates IR corresponding to AbstractShuffle::partiallyWrapIndex method, 515 // which partially wraps index by modulo VEC_LENGTH and generates a negative index value if original 516 // index is out of valid index range [0, VEC_LENGTH) 517 // 518 // wrapped_index = (VEC_LENGTH - 1) & index 519 // if (index u> VEC_LENGTH) { 520 // wrapped_index -= VEC_LENGTH; 521 // 522 // Note: Unsigned greater than comparison treat both <0 and >VEC_LENGTH indices as out-of-bound 523 // indexes. 524 Node* LibraryCallKit::partially_wrap_indexes(Node* index_vec, int num_elem, BasicType elem_bt) { 525 assert(elem_bt == T_BYTE, "Shuffles use byte array based backing storage."); 526 const TypeVect* vt = TypeVect::make(elem_bt, num_elem); 527 const Type* type_bt = Type::get_const_basic_type(elem_bt); 528 529 Node* mod_mask = gvn().makecon(TypeInt::make(num_elem-1)); 530 Node* bcast_mod_mask = gvn().transform(VectorNode::scalar2vector(mod_mask, num_elem, type_bt)); 531 532 BoolTest::mask pred = BoolTest::ugt; 533 ConINode* pred_node = (ConINode*)gvn().makecon(TypeInt::make(pred)); 534 Node* lane_cnt = gvn().makecon(TypeInt::make(num_elem)); 535 Node* bcast_lane_cnt = gvn().transform(VectorNode::scalar2vector(lane_cnt, num_elem, type_bt)); 536 const TypeVect* vmask_type = TypeVect::makemask(type_bt, num_elem); 537 Node* mask = gvn().transform(new VectorMaskCmpNode(pred, bcast_lane_cnt, index_vec, pred_node, vmask_type)); 538 539 // Make the indices greater than lane count as -ve values to match the java side implementation. 540 index_vec = gvn().transform(VectorNode::make(Op_AndV, index_vec, bcast_mod_mask, vt)); 541 Node* biased_val = gvn().transform(VectorNode::make(Op_SubVB, index_vec, bcast_lane_cnt, vt)); 542 return gvn().transform(new VectorBlendNode(biased_val, index_vec, mask)); 543 } 544 545 // <Sh extends VectorShuffle<E>, E> 546 // Sh ShuffleIota(Class<?> E, Class<?> shuffleClass, Vector.Species<E> s, int length, 547 // int start, int step, int wrap, ShuffleIotaOperation<Sh, E> defaultImpl) 548 bool LibraryCallKit::inline_vector_shuffle_iota() { 549 const TypeInstPtr* shuffle_klass = gvn().type(argument(1))->isa_instptr(); 550 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 551 const TypeInt* start_val = gvn().type(argument(4))->isa_int(); 552 const TypeInt* step_val = gvn().type(argument(5))->isa_int(); 553 const TypeInt* wrap = gvn().type(argument(6))->isa_int(); 554 555 if (shuffle_klass == nullptr || shuffle_klass->const_oop() == nullptr || 556 vlen == nullptr || !vlen->is_con() || start_val == nullptr || step_val == nullptr || 557 wrap == nullptr || !wrap->is_con()) { 558 return false; // not enough info for intrinsification 559 } 560 561 if (!is_klass_initialized(shuffle_klass)) { 562 log_if_needed(" ** klass argument not initialized"); 563 return false; 564 } 565 566 int do_wrap = wrap->get_con(); 567 int num_elem = vlen->get_con(); 568 BasicType elem_bt = T_BYTE; 569 570 bool effective_indices_in_range = false; 571 if (start_val->is_con() && step_val->is_con()) { 572 int effective_min_index = start_val->get_con(); 573 int effective_max_index = start_val->get_con() + step_val->get_con() * (num_elem - 1); 574 effective_indices_in_range = effective_max_index >= effective_min_index && effective_min_index >= -128 && effective_max_index <= 127; 575 } 576 577 if (!do_wrap && !effective_indices_in_range) { 578 // Disable instrinsification for unwrapped shuffle iota if start/step 579 // values are non-constant OR if intermediate result overflows byte value range. 580 return false; 581 } 582 583 if (!arch_supports_vector(Op_AddVB, num_elem, elem_bt, VecMaskNotUsed) || 584 !arch_supports_vector(Op_AndV, num_elem, elem_bt, VecMaskNotUsed) || 585 !arch_supports_vector(Op_VectorLoadConst, num_elem, elem_bt, VecMaskNotUsed) || 586 !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed)) { 587 return false; 588 } 589 590 if (!do_wrap && 591 (!arch_supports_vector(Op_SubVB, num_elem, elem_bt, VecMaskNotUsed) || 592 !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskNotUsed) || 593 !arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskNotUsed))) { 594 return false; 595 } 596 597 bool step_multiply = !step_val->is_con() || !is_power_of_2(step_val->get_con()); 598 if ((step_multiply && !arch_supports_vector(Op_MulVB, num_elem, elem_bt, VecMaskNotUsed)) || 599 (!step_multiply && !arch_supports_vector(Op_LShiftVB, num_elem, elem_bt, VecMaskNotUsed))) { 600 return false; 601 } 602 603 const Type * type_bt = Type::get_const_basic_type(elem_bt); 604 const TypeVect * vt = TypeVect::make(type_bt, num_elem); 605 606 Node* res = gvn().transform(new VectorLoadConstNode(gvn().makecon(TypeInt::ZERO), vt)); 607 608 Node* start = argument(4); 609 Node* step = argument(5); 610 611 if (step_multiply) { 612 Node* bcast_step = gvn().transform(VectorNode::scalar2vector(step, num_elem, type_bt)); 613 res = gvn().transform(VectorNode::make(Op_MulVB, res, bcast_step, vt)); 614 } else if (step_val->get_con() > 1) { 615 Node* cnt = gvn().makecon(TypeInt::make(log2i_exact(step_val->get_con()))); 616 Node* shift_cnt = vector_shift_count(cnt, Op_LShiftI, elem_bt, num_elem); 617 res = gvn().transform(VectorNode::make(Op_LShiftVB, res, shift_cnt, vt)); 618 } 619 620 if (!start_val->is_con() || start_val->get_con() != 0) { 621 Node* bcast_start = gvn().transform(VectorNode::scalar2vector(start, num_elem, type_bt)); 622 res = gvn().transform(VectorNode::make(Op_AddVB, res, bcast_start, vt)); 623 } 624 625 Node * mod_val = gvn().makecon(TypeInt::make(num_elem-1)); 626 Node * bcast_mod = gvn().transform(VectorNode::scalar2vector(mod_val, num_elem, type_bt)); 627 628 if (do_wrap) { 629 // Wrap the indices greater than lane count. 630 res = gvn().transform(VectorNode::make(Op_AndV, res, bcast_mod, vt)); 631 } else { 632 res = partially_wrap_indexes(res, num_elem, elem_bt); 633 } 634 635 ciKlass* sbox_klass = shuffle_klass->const_oop()->as_instance()->java_lang_Class_klass(); 636 const TypeInstPtr* shuffle_box_type = TypeInstPtr::make_exact(TypePtr::NotNull, sbox_klass); 637 638 // Wrap it up in VectorBox to keep object type information. 639 res = box_vector(res, shuffle_box_type, elem_bt, num_elem); 640 set_result(res); 641 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 642 return true; 643 } 644 645 // <E, M> 646 // long maskReductionCoerced(int oper, Class<? extends M> maskClass, Class<?> elemClass, 647 // int length, M m, VectorMaskOp<M> defaultImpl) 648 bool LibraryCallKit::inline_vector_mask_operation() { 649 const TypeInt* oper = gvn().type(argument(0))->isa_int(); 650 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 651 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 652 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 653 Node* mask = argument(4); 654 655 if (mask_klass == nullptr || elem_klass == nullptr || mask->is_top() || vlen == nullptr) { 656 return false; // dead code 657 } 658 659 if (!is_klass_initialized(mask_klass)) { 660 log_if_needed(" ** klass argument not initialized"); 661 return false; 662 } 663 664 int num_elem = vlen->get_con(); 665 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 666 BasicType elem_bt = elem_type->basic_type(); 667 668 int mopc = VectorSupport::vop2ideal(oper->get_con(), elem_bt); 669 if (!arch_supports_vector(mopc, num_elem, elem_bt, VecMaskUseLoad)) { 670 log_if_needed(" ** not supported: arity=1 op=cast#%d/3 vlen2=%d etype2=%s", 671 mopc, num_elem, type2name(elem_bt)); 672 return false; // not supported 673 } 674 675 const Type* elem_ty = Type::get_const_basic_type(elem_bt); 676 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 677 const TypeInstPtr* mask_box_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 678 Node* mask_vec = unbox_vector(mask, mask_box_type, elem_bt, num_elem, true); 679 if (mask_vec == nullptr) { 680 log_if_needed(" ** unbox failed mask=%s", 681 NodeClassNames[argument(4)->Opcode()]); 682 return false; 683 } 684 685 if (mask_vec->bottom_type()->isa_vectmask() == nullptr) { 686 mask_vec = gvn().transform(VectorStoreMaskNode::make(gvn(), mask_vec, elem_bt, num_elem)); 687 } 688 const Type* maskoper_ty = mopc == Op_VectorMaskToLong ? (const Type*)TypeLong::LONG : (const Type*)TypeInt::INT; 689 Node* maskoper = gvn().transform(VectorMaskOpNode::make(mask_vec, maskoper_ty, mopc)); 690 if (mopc != Op_VectorMaskToLong) { 691 maskoper = ConvI2L(maskoper); 692 } 693 set_result(maskoper); 694 695 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 696 return true; 697 } 698 699 // public static 700 // <V, 701 // Sh extends VectorShuffle<E>, 702 // E> 703 // V shuffleToVector(Class<? extends Vector<E>> vclass, Class<E> elementType, 704 // Class<? extends Sh> shuffleClass, Sh s, int length, 705 // ShuffleToVectorOperation<V, Sh, E> defaultImpl) 706 bool LibraryCallKit::inline_vector_shuffle_to_vector() { 707 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 708 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 709 const TypeInstPtr* shuffle_klass = gvn().type(argument(2))->isa_instptr(); 710 Node* shuffle = argument(3); 711 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 712 713 if (vector_klass == nullptr || elem_klass == nullptr || shuffle_klass == nullptr || shuffle->is_top() || vlen == nullptr) { 714 return false; // dead code 715 } 716 if (!vlen->is_con() || vector_klass->const_oop() == nullptr || shuffle_klass->const_oop() == nullptr) { 717 return false; // not enough info for intrinsification 718 } 719 if (!is_klass_initialized(shuffle_klass) || !is_klass_initialized(vector_klass) ) { 720 log_if_needed(" ** klass argument not initialized"); 721 return false; 722 } 723 724 int num_elem = vlen->get_con(); 725 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 726 BasicType elem_bt = elem_type->basic_type(); 727 728 if (num_elem < 4) { 729 return false; 730 } 731 732 int cast_vopc = VectorCastNode::opcode(-1, T_BYTE); // from shuffle of type T_BYTE 733 // Make sure that cast is implemented to particular type/size combination. 734 if (!arch_supports_vector(cast_vopc, num_elem, elem_bt, VecMaskNotUsed)) { 735 log_if_needed(" ** not supported: arity=1 op=cast#%d/3 vlen2=%d etype2=%s", 736 cast_vopc, num_elem, type2name(elem_bt)); 737 return false; 738 } 739 740 ciKlass* sbox_klass = shuffle_klass->const_oop()->as_instance()->java_lang_Class_klass(); 741 const TypeInstPtr* shuffle_box_type = TypeInstPtr::make_exact(TypePtr::NotNull, sbox_klass); 742 743 // Unbox shuffle with true flag to indicate its load shuffle to vector 744 // shuffle is a byte array 745 Node* shuffle_vec = unbox_vector(shuffle, shuffle_box_type, T_BYTE, num_elem, true); 746 747 // cast byte to target element type 748 shuffle_vec = gvn().transform(VectorCastNode::make(cast_vopc, shuffle_vec, elem_bt, num_elem)); 749 750 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 751 const TypeInstPtr* vec_box_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 752 753 // Box vector 754 Node* res = box_vector(shuffle_vec, vec_box_type, elem_bt, num_elem); 755 set_result(res); 756 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 757 return true; 758 } 759 760 // public static 761 // <M, 762 // S extends VectorSpecies<E>, 763 // E> 764 // M fromBitsCoerced(Class<? extends M> vmClass, Class<E> elementType, int length, 765 // long bits, int mode, S s, 766 // BroadcastOperation<M, E, S> defaultImpl) 767 bool LibraryCallKit::inline_vector_frombits_coerced() { 768 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 769 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 770 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 771 const TypeLong* bits_type = gvn().type(argument(3))->isa_long(); 772 // Mode argument determines the mode of operation it can take following values:- 773 // MODE_BROADCAST for vector Vector.broadcast and VectorMask.maskAll operations. 774 // MODE_BITS_COERCED_LONG_TO_MASK for VectorMask.fromLong operation. 775 const TypeInt* mode = gvn().type(argument(5))->isa_int(); 776 777 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || mode == nullptr || 778 bits_type == nullptr || vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || 779 !vlen->is_con() || !mode->is_con()) { 780 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s bitwise=%s", 781 NodeClassNames[argument(0)->Opcode()], 782 NodeClassNames[argument(1)->Opcode()], 783 NodeClassNames[argument(2)->Opcode()], 784 NodeClassNames[argument(5)->Opcode()]); 785 return false; // not enough info for intrinsification 786 } 787 788 if (!is_klass_initialized(vector_klass)) { 789 log_if_needed(" ** klass argument not initialized"); 790 return false; 791 } 792 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 793 if (!elem_type->is_primitive_type()) { 794 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 795 return false; // should be primitive type 796 } 797 BasicType elem_bt = elem_type->basic_type(); 798 int num_elem = vlen->get_con(); 799 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 800 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 801 802 bool is_mask = is_vector_mask(vbox_klass); 803 int bcast_mode = mode->get_con(); 804 VectorMaskUseType checkFlags = (VectorMaskUseType)(is_mask ? VecMaskUseAll : VecMaskNotUsed); 805 int opc = bcast_mode == VectorSupport::MODE_BITS_COERCED_LONG_TO_MASK ? Op_VectorLongToMask : Op_Replicate; 806 807 if (!arch_supports_vector(opc, num_elem, elem_bt, checkFlags, true /*has_scalar_args*/)) { 808 log_if_needed(" ** not supported: arity=0 op=broadcast vlen=%d etype=%s ismask=%d bcast_mode=%d", 809 num_elem, type2name(elem_bt), 810 is_mask ? 1 : 0, 811 bcast_mode); 812 return false; // not supported 813 } 814 815 Node* broadcast = nullptr; 816 Node* bits = argument(3); 817 Node* elem = bits; 818 819 if (opc == Op_VectorLongToMask) { 820 const TypeVect* vt = TypeVect::makemask(elem_bt, num_elem); 821 if (vt->isa_vectmask()) { 822 broadcast = gvn().transform(new VectorLongToMaskNode(elem, vt)); 823 } else { 824 const TypeVect* mvt = TypeVect::make(T_BOOLEAN, num_elem); 825 broadcast = gvn().transform(new VectorLongToMaskNode(elem, mvt)); 826 broadcast = gvn().transform(new VectorLoadMaskNode(broadcast, vt)); 827 } 828 } else { 829 switch (elem_bt) { 830 case T_BOOLEAN: // fall-through 831 case T_BYTE: // fall-through 832 case T_SHORT: // fall-through 833 case T_CHAR: // fall-through 834 case T_INT: { 835 elem = gvn().transform(new ConvL2INode(bits)); 836 break; 837 } 838 case T_DOUBLE: { 839 elem = gvn().transform(new MoveL2DNode(bits)); 840 break; 841 } 842 case T_FLOAT: { 843 bits = gvn().transform(new ConvL2INode(bits)); 844 elem = gvn().transform(new MoveI2FNode(bits)); 845 break; 846 } 847 case T_LONG: { 848 // no conversion needed 849 break; 850 } 851 default: fatal("%s", type2name(elem_bt)); 852 } 853 broadcast = VectorNode::scalar2vector(elem, num_elem, Type::get_const_basic_type(elem_bt), is_mask); 854 broadcast = gvn().transform(broadcast); 855 } 856 857 Node* box = box_vector(broadcast, vbox_type, elem_bt, num_elem); 858 set_result(box); 859 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 860 return true; 861 } 862 863 static bool elem_consistent_with_arr(BasicType elem_bt, const TypeAryPtr* arr_type, bool mismatched_ms) { 864 assert(arr_type != nullptr, "unexpected"); 865 BasicType arr_elem_bt = arr_type->elem()->array_element_basic_type(); 866 if (elem_bt == arr_elem_bt) { 867 return true; 868 } else if (elem_bt == T_SHORT && arr_elem_bt == T_CHAR) { 869 // Load/store of short vector from/to char[] is supported 870 return true; 871 } else if (elem_bt == T_BYTE && arr_elem_bt == T_BOOLEAN) { 872 // Load/store of byte vector from/to boolean[] is supported 873 return true; 874 } else { 875 return mismatched_ms; 876 } 877 } 878 879 // public static 880 // <C, 881 // VM extends VectorPayload, 882 // E, 883 // S extends VectorSpecies<E>> 884 // VM load(Class<? extends VM> vmClass, Class<E> eClass, 885 // int length, 886 // Object base, long offset, // Unsafe addressing 887 // boolean fromSegment, 888 // C container, long index, S s, // Arguments for default implementation 889 // LoadOperation<C, VM, S> defaultImpl) { 890 // public static 891 // <C, 892 // V extends VectorPayload> 893 // void store(Class<?> vClass, Class<?> eClass, 894 // int length, 895 // Object base, long offset, // Unsafe addressing 896 // boolean fromSegment, 897 // V v, C container, long index, // Arguments for default implementation 898 // StoreVectorOperation<C, V> defaultImpl) { 899 bool LibraryCallKit::inline_vector_mem_operation(bool is_store) { 900 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 901 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 902 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 903 const TypeInt* from_ms = gvn().type(argument(6))->isa_int(); 904 905 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || !from_ms->is_con() || 906 vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 907 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s from_ms=%s", 908 NodeClassNames[argument(0)->Opcode()], 909 NodeClassNames[argument(1)->Opcode()], 910 NodeClassNames[argument(2)->Opcode()], 911 NodeClassNames[argument(6)->Opcode()]); 912 return false; // not enough info for intrinsification 913 } 914 if (!is_klass_initialized(vector_klass)) { 915 log_if_needed(" ** klass argument not initialized"); 916 return false; 917 } 918 919 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 920 if (!elem_type->is_primitive_type()) { 921 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 922 return false; // should be primitive type 923 } 924 BasicType elem_bt = elem_type->basic_type(); 925 int num_elem = vlen->get_con(); 926 927 // TODO When mask usage is supported, VecMaskNotUsed needs to be VecMaskUseLoad. 928 if (!arch_supports_vector(is_store ? Op_StoreVector : Op_LoadVector, num_elem, elem_bt, VecMaskNotUsed)) { 929 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s ismask=no", 930 is_store, is_store ? "store" : "load", 931 num_elem, type2name(elem_bt)); 932 return false; // not supported 933 } 934 935 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 936 bool is_mask = is_vector_mask(vbox_klass); 937 938 Node* base = argument(3); 939 Node* offset = ConvL2X(argument(4)); 940 941 // Save state and restore on bailout 942 uint old_sp = sp(); 943 SafePointNode* old_map = clone_map(); 944 945 Node* addr = make_unsafe_address(base, offset, (is_mask ? T_BOOLEAN : elem_bt), true); 946 947 // The memory barrier checks are based on ones for unsafe access. 948 // This is not 1-1 implementation. 949 const Type *const base_type = gvn().type(base); 950 951 const TypePtr *addr_type = gvn().type(addr)->isa_ptr(); 952 const TypeAryPtr* arr_type = addr_type->isa_aryptr(); 953 954 const bool in_native = TypePtr::NULL_PTR == base_type; // base always null 955 const bool in_heap = !TypePtr::NULL_PTR->higher_equal(base_type); // base never null 956 957 const bool is_mixed_access = !in_heap && !in_native; 958 959 const bool is_mismatched_access = in_heap && (addr_type->isa_aryptr() == nullptr); 960 961 const bool needs_cpu_membar = is_mixed_access || is_mismatched_access; 962 963 // For non-masked mismatched memory segment vector read/write accesses, intrinsification can continue 964 // with unknown backing storage type and compiler can skip inserting explicit reinterpretation IR after 965 // loading from or before storing to backing storage which is mandatory for semantic correctness of 966 // big-endian memory layout. 967 bool mismatched_ms = LITTLE_ENDIAN_ONLY(false) 968 BIG_ENDIAN_ONLY(from_ms->get_con() && !is_mask && arr_type != nullptr && 969 arr_type->elem()->array_element_basic_type() != elem_bt); 970 BasicType mem_elem_bt = mismatched_ms ? arr_type->elem()->array_element_basic_type() : elem_bt; 971 if (!is_java_primitive(mem_elem_bt)) { 972 log_if_needed(" ** non-primitive array element type"); 973 return false; 974 } 975 int mem_num_elem = mismatched_ms ? (num_elem * type2aelembytes(elem_bt)) / type2aelembytes(mem_elem_bt) : num_elem; 976 if (arr_type != nullptr && !is_mask && !elem_consistent_with_arr(elem_bt, arr_type, mismatched_ms)) { 977 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s atype=%s ismask=no", 978 is_store, is_store ? "store" : "load", 979 num_elem, type2name(elem_bt), type2name(arr_type->elem()->array_element_basic_type())); 980 set_map(old_map); 981 set_sp(old_sp); 982 return false; 983 } 984 985 // In case of mismatched memory segment accesses, we need to double check that the source type memory operations are supported by backend. 986 if (mismatched_ms) { 987 if (is_store) { 988 if (!arch_supports_vector(Op_StoreVector, num_elem, elem_bt, VecMaskNotUsed) 989 || !arch_supports_vector(Op_VectorReinterpret, mem_num_elem, mem_elem_bt, VecMaskNotUsed)) { 990 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d*8 etype=%s/8 ismask=no", 991 is_store, "store", 992 num_elem, type2name(elem_bt)); 993 set_map(old_map); 994 set_sp(old_sp); 995 return false; // not supported 996 } 997 } else { 998 if (!arch_supports_vector(Op_LoadVector, mem_num_elem, mem_elem_bt, VecMaskNotUsed) 999 || !arch_supports_vector(Op_VectorReinterpret, num_elem, elem_bt, VecMaskNotUsed)) { 1000 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d*8 etype=%s/8 ismask=no", 1001 is_store, "load", 1002 mem_num_elem, type2name(mem_elem_bt)); 1003 set_map(old_map); 1004 set_sp(old_sp); 1005 return false; // not supported 1006 } 1007 } 1008 } 1009 if (is_mask) { 1010 if (!is_store) { 1011 if (!arch_supports_vector(Op_LoadVector, num_elem, elem_bt, VecMaskUseLoad)) { 1012 set_map(old_map); 1013 set_sp(old_sp); 1014 return false; // not supported 1015 } 1016 } else { 1017 if (!arch_supports_vector(Op_StoreVector, num_elem, elem_bt, VecMaskUseStore)) { 1018 set_map(old_map); 1019 set_sp(old_sp); 1020 return false; // not supported 1021 } 1022 } 1023 } 1024 1025 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1026 1027 if (needs_cpu_membar) { 1028 insert_mem_bar(Op_MemBarCPUOrder); 1029 } 1030 1031 if (is_store) { 1032 Node* val = unbox_vector(argument(7), vbox_type, elem_bt, num_elem); 1033 if (val == nullptr) { 1034 set_map(old_map); 1035 set_sp(old_sp); 1036 return false; // operand unboxing failed 1037 } 1038 set_all_memory(reset_memory()); 1039 1040 // In case the store needs to happen to byte array, reinterpret the incoming vector to byte vector. 1041 int store_num_elem = num_elem; 1042 if (mismatched_ms) { 1043 store_num_elem = mem_num_elem; 1044 const TypeVect* to_vect_type = TypeVect::make(mem_elem_bt, store_num_elem); 1045 val = gvn().transform(new VectorReinterpretNode(val, val->bottom_type()->is_vect(), to_vect_type)); 1046 } 1047 if (is_mask) { 1048 val = gvn().transform(VectorStoreMaskNode::make(gvn(), val, elem_bt, num_elem)); 1049 } 1050 Node* vstore = gvn().transform(StoreVectorNode::make(0, control(), memory(addr), addr, addr_type, val, store_num_elem)); 1051 set_memory(vstore, addr_type); 1052 } else { 1053 // When using byte array, we need to load as byte then reinterpret the value. Otherwise, do a simple vector load. 1054 Node* vload = nullptr; 1055 if (mismatched_ms) { 1056 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, mem_num_elem, mem_elem_bt)); 1057 const TypeVect* to_vect_type = TypeVect::make(elem_bt, num_elem); 1058 vload = gvn().transform(new VectorReinterpretNode(vload, vload->bottom_type()->is_vect(), to_vect_type)); 1059 } else { 1060 // Special handle for masks 1061 if (is_mask) { 1062 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, num_elem, T_BOOLEAN)); 1063 vload = gvn().transform(new VectorLoadMaskNode(vload, TypeVect::makemask(elem_bt, num_elem))); 1064 } else { 1065 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, num_elem, elem_bt)); 1066 } 1067 } 1068 Node* box = box_vector(vload, vbox_type, elem_bt, num_elem); 1069 set_result(box); 1070 } 1071 1072 destruct_map_clone(old_map); 1073 1074 if (needs_cpu_membar) { 1075 insert_mem_bar(Op_MemBarCPUOrder); 1076 } 1077 1078 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1079 return true; 1080 } 1081 1082 // public static 1083 // <C, 1084 // V extends Vector<?>, 1085 // E, 1086 // S extends VectorSpecies<E>, 1087 // M extends VectorMask<E>> 1088 // V loadMasked(Class<? extends V> vClass, Class<M> mClass, Class<E> eClass, 1089 // int length, Object base, long offset, // Unsafe addressing 1090 // boolean fromSegment, 1091 // M m, int offsetInRange, 1092 // C container, long index, S s, // Arguments for default implementation 1093 // LoadVectorMaskedOperation<C, V, S, M> defaultImpl) { 1094 // public static 1095 // <C, 1096 // V extends Vector<E>, 1097 // M extends VectorMask<E>, 1098 // E> 1099 // void storeMasked(Class<? extends V> vClass, Class<M> mClass, Class<E> eClass, 1100 // int length, 1101 // Object base, long offset, // Unsafe addressing 1102 // boolean fromSegment, 1103 // V v, M m, C container, long index, // Arguments for default implementation 1104 // StoreVectorMaskedOperation<C, V, M> defaultImpl) { 1105 1106 bool LibraryCallKit::inline_vector_mem_masked_operation(bool is_store) { 1107 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1108 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 1109 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1110 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1111 const TypeInt* from_ms = gvn().type(argument(7))->isa_int(); 1112 1113 if (vector_klass == nullptr || mask_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 1114 vector_klass->const_oop() == nullptr || mask_klass->const_oop() == nullptr || from_ms == nullptr || 1115 elem_klass->const_oop() == nullptr || !vlen->is_con() || !from_ms->is_con()) { 1116 log_if_needed(" ** missing constant: vclass=%s mclass=%s etype=%s vlen=%s from_ms=%s", 1117 NodeClassNames[argument(0)->Opcode()], 1118 NodeClassNames[argument(1)->Opcode()], 1119 NodeClassNames[argument(2)->Opcode()], 1120 NodeClassNames[argument(3)->Opcode()], 1121 NodeClassNames[argument(7)->Opcode()]); 1122 return false; // not enough info for intrinsification 1123 } 1124 if (!is_klass_initialized(vector_klass)) { 1125 log_if_needed(" ** klass argument not initialized"); 1126 return false; 1127 } 1128 1129 if (!is_klass_initialized(mask_klass)) { 1130 log_if_needed(" ** mask klass argument not initialized"); 1131 return false; 1132 } 1133 1134 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1135 if (!elem_type->is_primitive_type()) { 1136 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1137 return false; // should be primitive type 1138 } 1139 1140 BasicType elem_bt = elem_type->basic_type(); 1141 int num_elem = vlen->get_con(); 1142 1143 Node* base = argument(4); 1144 Node* offset = ConvL2X(argument(5)); 1145 1146 // Save state and restore on bailout 1147 uint old_sp = sp(); 1148 SafePointNode* old_map = clone_map(); 1149 1150 Node* addr = make_unsafe_address(base, offset, elem_bt, true); 1151 const TypePtr *addr_type = gvn().type(addr)->isa_ptr(); 1152 const TypeAryPtr* arr_type = addr_type->isa_aryptr(); 1153 1154 bool mismatched_ms = from_ms->get_con() && arr_type != nullptr && arr_type->elem()->array_element_basic_type() != elem_bt; 1155 BIG_ENDIAN_ONLY(if (mismatched_ms) return false;) 1156 // If there is no consistency between array and vector element types, it must be special byte array case 1157 if (arr_type != nullptr && !elem_consistent_with_arr(elem_bt, arr_type, mismatched_ms)) { 1158 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s atype=%s", 1159 is_store, is_store ? "storeMasked" : "loadMasked", 1160 num_elem, type2name(elem_bt), type2name(arr_type->elem()->array_element_basic_type())); 1161 set_map(old_map); 1162 set_sp(old_sp); 1163 return false; 1164 } 1165 1166 int mem_num_elem = mismatched_ms ? num_elem * type2aelembytes(elem_bt) : num_elem; 1167 BasicType mem_elem_bt = mismatched_ms ? T_BYTE : elem_bt; 1168 bool supports_predicate = arch_supports_vector(is_store ? Op_StoreVectorMasked : Op_LoadVectorMasked, 1169 mem_num_elem, mem_elem_bt, VecMaskUseLoad); 1170 1171 // If current arch does not support the predicated operations, we have to bail 1172 // out when current case uses the predicate feature. 1173 if (!supports_predicate) { 1174 bool needs_predicate = false; 1175 if (is_store) { 1176 // Masked vector store always uses the predicated store. 1177 needs_predicate = true; 1178 } else { 1179 // Masked vector load with IOOBE always uses the predicated load. 1180 const TypeInt* offset_in_range = gvn().type(argument(9))->isa_int(); 1181 if (!offset_in_range->is_con()) { 1182 log_if_needed(" ** missing constant: offsetInRange=%s", 1183 NodeClassNames[argument(8)->Opcode()]); 1184 set_map(old_map); 1185 set_sp(old_sp); 1186 return false; 1187 } 1188 needs_predicate = (offset_in_range->get_con() == 0); 1189 } 1190 1191 if (needs_predicate) { 1192 log_if_needed(" ** not supported: op=%s vlen=%d etype=%s mismatched_ms=%d", 1193 is_store ? "storeMasked" : "loadMasked", 1194 num_elem, type2name(elem_bt), mismatched_ms ? 1 : 0); 1195 set_map(old_map); 1196 set_sp(old_sp); 1197 return false; 1198 } 1199 } 1200 1201 // This only happens for masked vector load. If predicate is not supported, then check whether 1202 // the normal vector load and blend operations are supported by backend. 1203 if (!supports_predicate && (!arch_supports_vector(Op_LoadVector, mem_num_elem, mem_elem_bt, VecMaskNotUsed) || 1204 !arch_supports_vector(Op_VectorBlend, mem_num_elem, mem_elem_bt, VecMaskUseLoad))) { 1205 log_if_needed(" ** not supported: op=loadMasked vlen=%d etype=%s mismatched_ms=%d", 1206 num_elem, type2name(elem_bt), mismatched_ms ? 1 : 0); 1207 set_map(old_map); 1208 set_sp(old_sp); 1209 return false; 1210 } 1211 1212 // Since we are using byte array, we need to double check that the vector reinterpret operation 1213 // with byte type is supported by backend. 1214 if (mismatched_ms) { 1215 if (!arch_supports_vector(Op_VectorReinterpret, mem_num_elem, T_BYTE, VecMaskNotUsed)) { 1216 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s mismatched_ms=1", 1217 is_store, is_store ? "storeMasked" : "loadMasked", 1218 num_elem, type2name(elem_bt)); 1219 set_map(old_map); 1220 set_sp(old_sp); 1221 return false; 1222 } 1223 } 1224 1225 // Since it needs to unbox the mask, we need to double check that the related load operations 1226 // for mask are supported by backend. 1227 if (!arch_supports_vector(Op_LoadVector, num_elem, elem_bt, VecMaskUseLoad)) { 1228 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s", 1229 is_store, is_store ? "storeMasked" : "loadMasked", 1230 num_elem, type2name(elem_bt)); 1231 set_map(old_map); 1232 set_sp(old_sp); 1233 return false; 1234 } 1235 1236 // Can base be null? Otherwise, always on-heap access. 1237 bool can_access_non_heap = TypePtr::NULL_PTR->higher_equal(gvn().type(base)); 1238 if (can_access_non_heap) { 1239 insert_mem_bar(Op_MemBarCPUOrder); 1240 } 1241 1242 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1243 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1244 assert(!is_vector_mask(vbox_klass) && is_vector_mask(mbox_klass), "Invalid class type"); 1245 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1246 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1247 1248 Node* mask = unbox_vector(is_store ? argument(9) : argument(8), mbox_type, elem_bt, num_elem); 1249 if (mask == nullptr) { 1250 log_if_needed(" ** unbox failed mask=%s", 1251 is_store ? NodeClassNames[argument(9)->Opcode()] 1252 : NodeClassNames[argument(8)->Opcode()]); 1253 set_map(old_map); 1254 set_sp(old_sp); 1255 return false; 1256 } 1257 1258 if (is_store) { 1259 Node* val = unbox_vector(argument(8), vbox_type, elem_bt, num_elem); 1260 if (val == nullptr) { 1261 log_if_needed(" ** unbox failed vector=%s", 1262 NodeClassNames[argument(8)->Opcode()]); 1263 set_map(old_map); 1264 set_sp(old_sp); 1265 return false; // operand unboxing failed 1266 } 1267 set_all_memory(reset_memory()); 1268 1269 if (mismatched_ms) { 1270 // Reinterpret the incoming vector to byte vector. 1271 const TypeVect* to_vect_type = TypeVect::make(mem_elem_bt, mem_num_elem); 1272 val = gvn().transform(new VectorReinterpretNode(val, val->bottom_type()->is_vect(), to_vect_type)); 1273 // Reinterpret the vector mask to byte type. 1274 const TypeVect* from_mask_type = TypeVect::makemask(elem_bt, num_elem); 1275 const TypeVect* to_mask_type = TypeVect::makemask(mem_elem_bt, mem_num_elem); 1276 mask = gvn().transform(new VectorReinterpretNode(mask, from_mask_type, to_mask_type)); 1277 } 1278 Node* vstore = gvn().transform(new StoreVectorMaskedNode(control(), memory(addr), addr, val, addr_type, mask)); 1279 set_memory(vstore, addr_type); 1280 } else { 1281 Node* vload = nullptr; 1282 1283 if (mismatched_ms) { 1284 // Reinterpret the vector mask to byte type. 1285 const TypeVect* from_mask_type = TypeVect::makemask(elem_bt, num_elem); 1286 const TypeVect* to_mask_type = TypeVect::makemask(mem_elem_bt, mem_num_elem); 1287 mask = gvn().transform(new VectorReinterpretNode(mask, from_mask_type, to_mask_type)); 1288 } 1289 1290 if (supports_predicate) { 1291 // Generate masked load vector node if predicate feature is supported. 1292 const TypeVect* vt = TypeVect::make(mem_elem_bt, mem_num_elem); 1293 vload = gvn().transform(new LoadVectorMaskedNode(control(), memory(addr), addr, addr_type, vt, mask)); 1294 } else { 1295 // Use the vector blend to implement the masked load vector. The biased elements are zeros. 1296 Node* zero = gvn().transform(gvn().zerocon(mem_elem_bt)); 1297 zero = gvn().transform(VectorNode::scalar2vector(zero, mem_num_elem, Type::get_const_basic_type(mem_elem_bt))); 1298 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, mem_num_elem, mem_elem_bt)); 1299 vload = gvn().transform(new VectorBlendNode(zero, vload, mask)); 1300 } 1301 1302 if (mismatched_ms) { 1303 const TypeVect* to_vect_type = TypeVect::make(elem_bt, num_elem); 1304 vload = gvn().transform(new VectorReinterpretNode(vload, vload->bottom_type()->is_vect(), to_vect_type)); 1305 } 1306 1307 Node* box = box_vector(vload, vbox_type, elem_bt, num_elem); 1308 set_result(box); 1309 } 1310 1311 destruct_map_clone(old_map); 1312 1313 if (can_access_non_heap) { 1314 insert_mem_bar(Op_MemBarCPUOrder); 1315 } 1316 1317 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1318 return true; 1319 } 1320 1321 // <C, 1322 // V extends Vector<?>, 1323 // W extends Vector<Integer>, 1324 // S extends VectorSpecies<E>, 1325 // M extends VectorMask<E>, 1326 // E> 1327 // V loadWithMap(Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, int length, 1328 // Class<? extends Vector<Integer>> vectorIndexClass, 1329 // Object base, long offset, // Unsafe addressing 1330 // W index_vector, M m, 1331 // C container, int index, int[] indexMap, int indexM, S s, // Arguments for default implementation 1332 // LoadVectorOperationWithMap<C, V, E, S, M> defaultImpl) 1333 // 1334 // <C, 1335 // V extends Vector<E>, 1336 // W extends Vector<Integer>, 1337 // M extends VectorMask<E>, 1338 // E> 1339 // void storeWithMap(Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, 1340 // int length, Class<? extends Vector<Integer>> vectorIndexClass, Object base, long offset, // Unsafe addressing 1341 // W index_vector, V v, M m, 1342 // C container, int index, int[] indexMap, int indexM, // Arguments for default implementation 1343 // StoreVectorOperationWithMap<C, V, M, E> defaultImpl) 1344 // 1345 bool LibraryCallKit::inline_vector_gather_scatter(bool is_scatter) { 1346 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1347 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 1348 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1349 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1350 const TypeInstPtr* vector_idx_klass = gvn().type(argument(4))->isa_instptr(); 1351 1352 if (vector_klass == nullptr || elem_klass == nullptr || vector_idx_klass == nullptr || vlen == nullptr || 1353 vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || vector_idx_klass->const_oop() == nullptr || !vlen->is_con()) { 1354 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s viclass=%s", 1355 NodeClassNames[argument(0)->Opcode()], 1356 NodeClassNames[argument(2)->Opcode()], 1357 NodeClassNames[argument(3)->Opcode()], 1358 NodeClassNames[argument(4)->Opcode()]); 1359 return false; // not enough info for intrinsification 1360 } 1361 1362 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(vector_idx_klass)) { 1363 log_if_needed(" ** klass argument not initialized"); 1364 return false; 1365 } 1366 1367 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1368 if (!elem_type->is_primitive_type()) { 1369 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1370 return false; // should be primitive type 1371 } 1372 1373 BasicType elem_bt = elem_type->basic_type(); 1374 int num_elem = vlen->get_con(); 1375 1376 const Type* vmask_type = gvn().type(is_scatter ? argument(10) : argument(9)); 1377 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 1378 if (is_masked_op) { 1379 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 1380 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(1)->Opcode()]); 1381 return false; // not enough info for intrinsification 1382 } 1383 1384 if (!is_klass_initialized(mask_klass)) { 1385 log_if_needed(" ** mask klass argument not initialized"); 1386 return false; 1387 } 1388 1389 if (vmask_type->maybe_null()) { 1390 log_if_needed(" ** null mask values are not allowed for masked op"); 1391 return false; 1392 } 1393 1394 // Check whether the predicated gather/scatter node is supported by architecture. 1395 VectorMaskUseType mask = (VectorMaskUseType) (VecMaskUseLoad | VecMaskUsePred); 1396 if (!arch_supports_vector(is_scatter ? Op_StoreVectorScatterMasked : Op_LoadVectorGatherMasked, num_elem, elem_bt, mask)) { 1397 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s is_masked_op=1", 1398 is_scatter, is_scatter ? "scatterMasked" : "gatherMasked", 1399 num_elem, type2name(elem_bt)); 1400 return false; // not supported 1401 } 1402 } else { 1403 // Check whether the normal gather/scatter node is supported for non-masked operation. 1404 if (!arch_supports_vector(is_scatter ? Op_StoreVectorScatter : Op_LoadVectorGather, num_elem, elem_bt, VecMaskNotUsed)) { 1405 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s is_masked_op=0", 1406 is_scatter, is_scatter ? "scatter" : "gather", 1407 num_elem, type2name(elem_bt)); 1408 return false; // not supported 1409 } 1410 } 1411 1412 // Check that the vector holding indices is supported by architecture 1413 // For sub-word gathers expander receive index array. 1414 if (!is_subword_type(elem_bt) && !arch_supports_vector(Op_LoadVector, num_elem, T_INT, VecMaskNotUsed)) { 1415 log_if_needed(" ** not supported: arity=%d op=%s/loadindex vlen=%d etype=int is_masked_op=%d", 1416 is_scatter, is_scatter ? "scatter" : "gather", 1417 num_elem, is_masked_op ? 1 : 0); 1418 return false; // not supported 1419 } 1420 1421 Node* base = argument(5); 1422 Node* offset = ConvL2X(argument(6)); 1423 1424 // Save state and restore on bailout 1425 uint old_sp = sp(); 1426 SafePointNode* old_map = clone_map(); 1427 1428 Node* addr = make_unsafe_address(base, offset, elem_bt, true); 1429 1430 const TypePtr *addr_type = gvn().type(addr)->isa_ptr(); 1431 const TypeAryPtr* arr_type = addr_type->isa_aryptr(); 1432 1433 // The array must be consistent with vector type 1434 if (arr_type == nullptr || (arr_type != nullptr && !elem_consistent_with_arr(elem_bt, arr_type, false))) { 1435 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s atype=%s ismask=no", 1436 is_scatter, is_scatter ? "scatter" : "gather", 1437 num_elem, type2name(elem_bt), type2name(arr_type->elem()->array_element_basic_type())); 1438 set_map(old_map); 1439 set_sp(old_sp); 1440 return false; 1441 } 1442 1443 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1444 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1445 ciKlass* vbox_idx_klass = vector_idx_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1446 if (vbox_idx_klass == nullptr) { 1447 set_map(old_map); 1448 set_sp(old_sp); 1449 return false; 1450 } 1451 1452 Node* index_vect = nullptr; 1453 const TypeInstPtr* vbox_idx_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_idx_klass); 1454 if (!is_subword_type(elem_bt)) { 1455 index_vect = unbox_vector(argument(8), vbox_idx_type, T_INT, num_elem); 1456 if (index_vect == nullptr) { 1457 set_map(old_map); 1458 set_sp(old_sp); 1459 return false; 1460 } 1461 } 1462 1463 Node* mask = nullptr; 1464 if (is_masked_op) { 1465 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1466 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1467 mask = unbox_vector(is_scatter ? argument(10) : argument(9), mbox_type, elem_bt, num_elem); 1468 if (mask == nullptr) { 1469 log_if_needed(" ** unbox failed mask=%s", 1470 is_scatter ? NodeClassNames[argument(10)->Opcode()] 1471 : NodeClassNames[argument(9)->Opcode()]); 1472 set_map(old_map); 1473 set_sp(old_sp); 1474 return false; 1475 } 1476 } 1477 1478 const TypeVect* vector_type = TypeVect::make(elem_bt, num_elem); 1479 if (is_scatter) { 1480 Node* val = unbox_vector(argument(9), vbox_type, elem_bt, num_elem); 1481 if (val == nullptr) { 1482 set_map(old_map); 1483 set_sp(old_sp); 1484 return false; // operand unboxing failed 1485 } 1486 set_all_memory(reset_memory()); 1487 1488 Node* vstore = nullptr; 1489 if (mask != nullptr) { 1490 vstore = gvn().transform(new StoreVectorScatterMaskedNode(control(), memory(addr), addr, addr_type, val, index_vect, mask)); 1491 } else { 1492 vstore = gvn().transform(new StoreVectorScatterNode(control(), memory(addr), addr, addr_type, val, index_vect)); 1493 } 1494 set_memory(vstore, addr_type); 1495 } else { 1496 Node* vload = nullptr; 1497 Node* index = argument(11); 1498 Node* indexMap = argument(12); 1499 Node* indexM = argument(13); 1500 if (mask != nullptr) { 1501 if (is_subword_type(elem_bt)) { 1502 Node* index_arr_base = array_element_address(indexMap, indexM, T_INT); 1503 vload = gvn().transform(new LoadVectorGatherMaskedNode(control(), memory(addr), addr, addr_type, vector_type, index_arr_base, mask, index)); 1504 } else { 1505 vload = gvn().transform(new LoadVectorGatherMaskedNode(control(), memory(addr), addr, addr_type, vector_type, index_vect, mask)); 1506 } 1507 } else { 1508 if (is_subword_type(elem_bt)) { 1509 Node* index_arr_base = array_element_address(indexMap, indexM, T_INT); 1510 vload = gvn().transform(new LoadVectorGatherNode(control(), memory(addr), addr, addr_type, vector_type, index_arr_base, index)); 1511 } else { 1512 vload = gvn().transform(new LoadVectorGatherNode(control(), memory(addr), addr, addr_type, vector_type, index_vect)); 1513 } 1514 } 1515 Node* box = box_vector(vload, vbox_type, elem_bt, num_elem); 1516 set_result(box); 1517 } 1518 1519 destruct_map_clone(old_map); 1520 1521 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1522 return true; 1523 } 1524 1525 // public static 1526 // <V extends Vector<E>, 1527 // M extends VectorMask<E>, 1528 // E> 1529 // long reductionCoerced(int oprId, Class<? extends V> vectorClass, Class<? extends M> maskClass, 1530 // Class<E> elementType, int length, V v, M m, 1531 // ReductionOperation<V, M> defaultImpl) 1532 bool LibraryCallKit::inline_vector_reduction() { 1533 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 1534 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 1535 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 1536 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 1537 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 1538 1539 if (opr == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 1540 !opr->is_con() || vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 1541 log_if_needed(" ** missing constant: opr=%s vclass=%s etype=%s vlen=%s", 1542 NodeClassNames[argument(0)->Opcode()], 1543 NodeClassNames[argument(1)->Opcode()], 1544 NodeClassNames[argument(3)->Opcode()], 1545 NodeClassNames[argument(4)->Opcode()]); 1546 return false; // not enough info for intrinsification 1547 } 1548 if (!is_klass_initialized(vector_klass)) { 1549 log_if_needed(" ** klass argument not initialized"); 1550 return false; 1551 } 1552 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1553 if (!elem_type->is_primitive_type()) { 1554 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1555 return false; // should be primitive type 1556 } 1557 1558 const Type* vmask_type = gvn().type(argument(6)); 1559 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 1560 if (is_masked_op) { 1561 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 1562 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(2)->Opcode()]); 1563 return false; // not enough info for intrinsification 1564 } 1565 1566 if (!is_klass_initialized(mask_klass)) { 1567 log_if_needed(" ** mask klass argument not initialized"); 1568 return false; 1569 } 1570 1571 if (vmask_type->maybe_null()) { 1572 log_if_needed(" ** null mask values are not allowed for masked op"); 1573 return false; 1574 } 1575 } 1576 1577 BasicType elem_bt = elem_type->basic_type(); 1578 int num_elem = vlen->get_con(); 1579 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 1580 int sopc = ReductionNode::opcode(opc, elem_bt); 1581 1582 // When using mask, mask use type needs to be VecMaskUseLoad. 1583 if (!arch_supports_vector(sopc, num_elem, elem_bt, is_masked_op ? VecMaskUseLoad : VecMaskNotUsed)) { 1584 log_if_needed(" ** not supported: arity=1 op=%d/reduce vlen=%d etype=%s is_masked_op=%d", 1585 sopc, num_elem, type2name(elem_bt), is_masked_op ? 1 : 0); 1586 return false; 1587 } 1588 1589 // Return true if current platform has implemented the masked operation with predicate feature. 1590 bool use_predicate = is_masked_op && arch_supports_vector(sopc, num_elem, elem_bt, VecMaskUsePred); 1591 if (is_masked_op && !use_predicate && !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)) { 1592 log_if_needed(" ** not supported: arity=1 op=%d/reduce vlen=%d etype=%s is_masked_op=1", 1593 sopc, num_elem, type2name(elem_bt)); 1594 return false; 1595 } 1596 1597 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1598 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1599 1600 Node* opd = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1601 if (opd == nullptr) { 1602 return false; // operand unboxing failed 1603 } 1604 1605 Node* mask = nullptr; 1606 if (is_masked_op) { 1607 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1608 assert(is_vector_mask(mbox_klass), "argument(2) should be a mask class"); 1609 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1610 mask = unbox_vector(argument(6), mbox_type, elem_bt, num_elem); 1611 if (mask == nullptr) { 1612 log_if_needed(" ** unbox failed mask=%s", 1613 NodeClassNames[argument(6)->Opcode()]); 1614 return false; 1615 } 1616 } 1617 1618 Node* init = ReductionNode::make_identity_con_scalar(gvn(), opc, elem_bt); 1619 Node* value = opd; 1620 1621 assert(mask != nullptr || !is_masked_op, "Masked op needs the mask value never null"); 1622 if (mask != nullptr && !use_predicate) { 1623 Node* reduce_identity = gvn().transform(VectorNode::scalar2vector(init, num_elem, Type::get_const_basic_type(elem_bt))); 1624 value = gvn().transform(new VectorBlendNode(reduce_identity, value, mask)); 1625 } 1626 1627 // Make an unordered Reduction node. This affects only AddReductionVF/VD and MulReductionVF/VD, 1628 // as these operations are allowed to be associative (not requiring strict order) in VectorAPI. 1629 value = ReductionNode::make(opc, nullptr, init, value, elem_bt, /* requires_strict_order */ false); 1630 1631 if (mask != nullptr && use_predicate) { 1632 value->add_req(mask); 1633 value->add_flag(Node::Flag_is_predicated_vector); 1634 } 1635 1636 value = gvn().transform(value); 1637 1638 Node* bits = nullptr; 1639 switch (elem_bt) { 1640 case T_BYTE: 1641 case T_SHORT: 1642 case T_INT: { 1643 bits = gvn().transform(new ConvI2LNode(value)); 1644 break; 1645 } 1646 case T_FLOAT: { 1647 value = gvn().transform(new MoveF2INode(value)); 1648 bits = gvn().transform(new ConvI2LNode(value)); 1649 break; 1650 } 1651 case T_DOUBLE: { 1652 bits = gvn().transform(new MoveD2LNode(value)); 1653 break; 1654 } 1655 case T_LONG: { 1656 bits = value; // no conversion needed 1657 break; 1658 } 1659 default: fatal("%s", type2name(elem_bt)); 1660 } 1661 set_result(bits); 1662 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1663 return true; 1664 } 1665 1666 // public static <V> boolean test(int cond, Class<?> vectorClass, Class<?> elementType, int vlen, 1667 // V v1, V v2, 1668 // BiFunction<V, V, Boolean> defaultImpl) 1669 // 1670 bool LibraryCallKit::inline_vector_test() { 1671 const TypeInt* cond = gvn().type(argument(0))->isa_int(); 1672 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 1673 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1674 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1675 1676 if (cond == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 1677 !cond->is_con() || vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 1678 log_if_needed(" ** missing constant: cond=%s vclass=%s etype=%s vlen=%s", 1679 NodeClassNames[argument(0)->Opcode()], 1680 NodeClassNames[argument(1)->Opcode()], 1681 NodeClassNames[argument(2)->Opcode()], 1682 NodeClassNames[argument(3)->Opcode()]); 1683 return false; // not enough info for intrinsification 1684 } 1685 if (!is_klass_initialized(vector_klass)) { 1686 log_if_needed(" ** klass argument not initialized"); 1687 return false; 1688 } 1689 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1690 if (!elem_type->is_primitive_type()) { 1691 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1692 return false; // should be primitive type 1693 } 1694 BasicType elem_bt = elem_type->basic_type(); 1695 int num_elem = vlen->get_con(); 1696 BoolTest::mask booltest = (BoolTest::mask)cond->get_con(); 1697 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1698 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1699 1700 if (!arch_supports_vector(Op_VectorTest, num_elem, elem_bt, is_vector_mask(vbox_klass) ? VecMaskUseLoad : VecMaskNotUsed)) { 1701 log_if_needed(" ** not supported: arity=2 op=test/%d vlen=%d etype=%s ismask=%d", 1702 cond->get_con(), num_elem, type2name(elem_bt), 1703 is_vector_mask(vbox_klass)); 1704 return false; 1705 } 1706 1707 Node* opd1 = unbox_vector(argument(4), vbox_type, elem_bt, num_elem); 1708 Node* opd2; 1709 if (Matcher::vectortest_needs_second_argument(booltest == BoolTest::overflow, 1710 opd1->bottom_type()->isa_vectmask())) { 1711 opd2 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1712 } else { 1713 opd2 = opd1; 1714 } 1715 if (opd1 == nullptr || opd2 == nullptr) { 1716 return false; // operand unboxing failed 1717 } 1718 1719 Node* cmp = gvn().transform(new VectorTestNode(opd1, opd2, booltest)); 1720 BoolTest::mask test = Matcher::vectortest_mask(booltest == BoolTest::overflow, 1721 opd1->bottom_type()->isa_vectmask(), num_elem); 1722 Node* bol = gvn().transform(new BoolNode(cmp, test)); 1723 Node* res = gvn().transform(new CMoveINode(bol, gvn().intcon(0), gvn().intcon(1), TypeInt::BOOL)); 1724 1725 set_result(res); 1726 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1727 return true; 1728 } 1729 1730 // public static 1731 // <V extends Vector<E>, 1732 // M extends VectorMask<E>, 1733 // E> 1734 // V blend(Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, int vlen, 1735 // V v1, V v2, M m, 1736 // VectorBlendOp<V, M, E> defaultImpl) 1737 bool LibraryCallKit::inline_vector_blend() { 1738 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1739 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 1740 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1741 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1742 1743 if (mask_klass == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 1744 return false; // dead code 1745 } 1746 if (mask_klass->const_oop() == nullptr || vector_klass->const_oop() == nullptr || 1747 elem_klass->const_oop() == nullptr || !vlen->is_con()) { 1748 log_if_needed(" ** missing constant: vclass=%s mclass=%s etype=%s vlen=%s", 1749 NodeClassNames[argument(0)->Opcode()], 1750 NodeClassNames[argument(1)->Opcode()], 1751 NodeClassNames[argument(2)->Opcode()], 1752 NodeClassNames[argument(3)->Opcode()]); 1753 return false; // not enough info for intrinsification 1754 } 1755 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(mask_klass)) { 1756 log_if_needed(" ** klass argument not initialized"); 1757 return false; 1758 } 1759 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1760 if (!elem_type->is_primitive_type()) { 1761 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1762 return false; // should be primitive type 1763 } 1764 BasicType elem_bt = elem_type->basic_type(); 1765 BasicType mask_bt = elem_bt; 1766 int num_elem = vlen->get_con(); 1767 1768 if (!arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)) { 1769 log_if_needed(" ** not supported: arity=2 op=blend vlen=%d etype=%s ismask=useload", 1770 num_elem, type2name(elem_bt)); 1771 return false; // not supported 1772 } 1773 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1774 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1775 1776 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1777 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1778 1779 Node* v1 = unbox_vector(argument(4), vbox_type, elem_bt, num_elem); 1780 Node* v2 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1781 Node* mask = unbox_vector(argument(6), mbox_type, mask_bt, num_elem); 1782 1783 if (v1 == nullptr || v2 == nullptr || mask == nullptr) { 1784 return false; // operand unboxing failed 1785 } 1786 1787 Node* blend = gvn().transform(new VectorBlendNode(v1, v2, mask)); 1788 1789 Node* box = box_vector(blend, vbox_type, elem_bt, num_elem); 1790 set_result(box); 1791 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1792 return true; 1793 } 1794 1795 // public static 1796 // <V extends Vector<E>, 1797 // M extends VectorMask<E>, 1798 // E> 1799 // M compare(int cond, Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, int vlen, 1800 // V v1, V v2, M m, 1801 // VectorCompareOp<V,M> defaultImpl) 1802 bool LibraryCallKit::inline_vector_compare() { 1803 const TypeInt* cond = gvn().type(argument(0))->isa_int(); 1804 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 1805 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 1806 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 1807 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 1808 1809 if (cond == nullptr || vector_klass == nullptr || mask_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 1810 return false; // dead code 1811 } 1812 if (!cond->is_con() || vector_klass->const_oop() == nullptr || mask_klass->const_oop() == nullptr || 1813 elem_klass->const_oop() == nullptr || !vlen->is_con()) { 1814 log_if_needed(" ** missing constant: cond=%s vclass=%s mclass=%s etype=%s vlen=%s", 1815 NodeClassNames[argument(0)->Opcode()], 1816 NodeClassNames[argument(1)->Opcode()], 1817 NodeClassNames[argument(2)->Opcode()], 1818 NodeClassNames[argument(3)->Opcode()], 1819 NodeClassNames[argument(4)->Opcode()]); 1820 return false; // not enough info for intrinsification 1821 } 1822 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(mask_klass)) { 1823 log_if_needed(" ** klass argument not initialized"); 1824 return false; 1825 } 1826 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1827 if (!elem_type->is_primitive_type()) { 1828 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1829 return false; // should be primitive type 1830 } 1831 1832 int num_elem = vlen->get_con(); 1833 BasicType elem_bt = elem_type->basic_type(); 1834 BasicType mask_bt = elem_bt; 1835 1836 if ((cond->get_con() & BoolTest::unsigned_compare) != 0) { 1837 if (!Matcher::supports_vector_comparison_unsigned(num_elem, elem_bt)) { 1838 log_if_needed(" ** not supported: unsigned comparison op=comp/%d vlen=%d etype=%s ismask=usestore", 1839 cond->get_con() & (BoolTest::unsigned_compare - 1), num_elem, type2name(elem_bt)); 1840 return false; 1841 } 1842 } 1843 1844 if (!arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskUseStore)) { 1845 log_if_needed(" ** not supported: arity=2 op=comp/%d vlen=%d etype=%s ismask=usestore", 1846 cond->get_con(), num_elem, type2name(elem_bt)); 1847 return false; 1848 } 1849 1850 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1851 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1852 1853 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1854 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1855 1856 Node* v1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1857 Node* v2 = unbox_vector(argument(6), vbox_type, elem_bt, num_elem); 1858 1859 bool is_masked_op = argument(7)->bottom_type() != TypePtr::NULL_PTR; 1860 Node* mask = is_masked_op ? unbox_vector(argument(7), mbox_type, elem_bt, num_elem) : nullptr; 1861 if (is_masked_op && mask == nullptr) { 1862 log_if_needed(" ** not supported: mask = null arity=2 op=comp/%d vlen=%d etype=%s ismask=usestore is_masked_op=1", 1863 cond->get_con(), num_elem, type2name(elem_bt)); 1864 return false; 1865 } 1866 1867 bool use_predicate = is_masked_op && arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskUsePred); 1868 if (is_masked_op && !use_predicate && !arch_supports_vector(Op_AndV, num_elem, elem_bt, VecMaskUseLoad)) { 1869 log_if_needed(" ** not supported: arity=2 op=comp/%d vlen=%d etype=%s ismask=usestore is_masked_op=1", 1870 cond->get_con(), num_elem, type2name(elem_bt)); 1871 return false; 1872 } 1873 1874 if (v1 == nullptr || v2 == nullptr) { 1875 return false; // operand unboxing failed 1876 } 1877 BoolTest::mask pred = (BoolTest::mask)cond->get_con(); 1878 ConINode* pred_node = (ConINode*)gvn().makecon(cond); 1879 1880 const TypeVect* vmask_type = TypeVect::makemask(mask_bt, num_elem); 1881 Node* operation = new VectorMaskCmpNode(pred, v1, v2, pred_node, vmask_type); 1882 1883 if (is_masked_op) { 1884 if (use_predicate) { 1885 operation->add_req(mask); 1886 operation->add_flag(Node::Flag_is_predicated_vector); 1887 } else { 1888 operation = gvn().transform(operation); 1889 operation = VectorNode::make(Op_AndV, operation, mask, vmask_type); 1890 } 1891 } 1892 1893 operation = gvn().transform(operation); 1894 1895 Node* box = box_vector(operation, mbox_type, mask_bt, num_elem); 1896 set_result(box); 1897 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1898 return true; 1899 } 1900 1901 // public static 1902 // <V extends Vector<E>, 1903 // Sh extends VectorShuffle<E>, 1904 // M extends VectorMask<E>, 1905 // E> 1906 // V rearrangeOp(Class<? extends V> vectorClass, Class<Sh> shuffleClass, Class<M> maskClass, Class<E> elementType, int vlen, 1907 // V v1, Sh sh, M m, 1908 // VectorRearrangeOp<V, Sh, M, E> defaultImpl) 1909 bool LibraryCallKit::inline_vector_rearrange() { 1910 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1911 const TypeInstPtr* shuffle_klass = gvn().type(argument(1))->isa_instptr(); 1912 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 1913 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 1914 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 1915 1916 if (vector_klass == nullptr || shuffle_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 1917 return false; // dead code 1918 } 1919 if (shuffle_klass->const_oop() == nullptr || 1920 vector_klass->const_oop() == nullptr || 1921 elem_klass->const_oop() == nullptr || 1922 !vlen->is_con()) { 1923 log_if_needed(" ** missing constant: vclass=%s sclass=%s etype=%s vlen=%s", 1924 NodeClassNames[argument(0)->Opcode()], 1925 NodeClassNames[argument(1)->Opcode()], 1926 NodeClassNames[argument(3)->Opcode()], 1927 NodeClassNames[argument(4)->Opcode()]); 1928 return false; // not enough info for intrinsification 1929 } 1930 if (!is_klass_initialized(vector_klass) || 1931 !is_klass_initialized(shuffle_klass)) { 1932 log_if_needed(" ** klass argument not initialized"); 1933 return false; 1934 } 1935 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1936 if (!elem_type->is_primitive_type()) { 1937 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1938 return false; // should be primitive type 1939 } 1940 BasicType elem_bt = elem_type->basic_type(); 1941 BasicType shuffle_bt = elem_bt; 1942 int num_elem = vlen->get_con(); 1943 1944 if (!arch_supports_vector(Op_VectorLoadShuffle, num_elem, elem_bt, VecMaskNotUsed)) { 1945 log_if_needed(" ** not supported: arity=0 op=load/shuffle vlen=%d etype=%s ismask=no", 1946 num_elem, type2name(elem_bt)); 1947 return false; // not supported 1948 } 1949 1950 bool is_masked_op = argument(7)->bottom_type() != TypePtr::NULL_PTR; 1951 bool use_predicate = is_masked_op; 1952 if (is_masked_op && 1953 (mask_klass == nullptr || 1954 mask_klass->const_oop() == nullptr || 1955 !is_klass_initialized(mask_klass))) { 1956 log_if_needed(" ** mask_klass argument not initialized"); 1957 } 1958 VectorMaskUseType checkFlags = (VectorMaskUseType)(is_masked_op ? (VecMaskUseLoad | VecMaskUsePred) : VecMaskNotUsed); 1959 if (!arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, checkFlags)) { 1960 use_predicate = false; 1961 if(!is_masked_op || 1962 (!arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, VecMaskNotUsed) || 1963 !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad) || 1964 !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed))) { 1965 log_if_needed(" ** not supported: arity=2 op=shuffle/rearrange vlen=%d etype=%s ismask=no", 1966 num_elem, type2name(elem_bt)); 1967 return false; // not supported 1968 } 1969 } 1970 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1971 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1972 1973 ciKlass* shbox_klass = shuffle_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1974 const TypeInstPtr* shbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, shbox_klass); 1975 1976 Node* v1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1977 Node* shuffle = unbox_vector(argument(6), shbox_type, shuffle_bt, num_elem); 1978 1979 if (v1 == nullptr || shuffle == nullptr) { 1980 return false; // operand unboxing failed 1981 } 1982 1983 Node* mask = nullptr; 1984 if (is_masked_op) { 1985 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1986 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1987 mask = unbox_vector(argument(7), mbox_type, elem_bt, num_elem); 1988 if (mask == nullptr) { 1989 log_if_needed(" ** not supported: arity=3 op=shuffle/rearrange vlen=%d etype=%s ismask=useload is_masked_op=1", 1990 num_elem, type2name(elem_bt)); 1991 return false; 1992 } 1993 } 1994 1995 Node* rearrange = new VectorRearrangeNode(v1, shuffle); 1996 if (is_masked_op) { 1997 if (use_predicate) { 1998 rearrange->add_req(mask); 1999 rearrange->add_flag(Node::Flag_is_predicated_vector); 2000 } else { 2001 const TypeVect* vt = v1->bottom_type()->is_vect(); 2002 rearrange = gvn().transform(rearrange); 2003 Node* zero = gvn().makecon(Type::get_zero_type(elem_bt)); 2004 Node* zerovec = gvn().transform(VectorNode::scalar2vector(zero, num_elem, Type::get_const_basic_type(elem_bt))); 2005 rearrange = new VectorBlendNode(zerovec, rearrange, mask); 2006 } 2007 } 2008 rearrange = gvn().transform(rearrange); 2009 2010 Node* box = box_vector(rearrange, vbox_type, elem_bt, num_elem); 2011 set_result(box); 2012 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2013 return true; 2014 } 2015 2016 static address get_svml_address(int vop, int bits, BasicType bt, char* name_ptr, int name_len) { 2017 address addr = nullptr; 2018 assert(UseVectorStubs, "sanity"); 2019 assert(name_ptr != nullptr, "unexpected"); 2020 assert((vop >= VectorSupport::VECTOR_OP_SVML_START) && (vop <= VectorSupport::VECTOR_OP_SVML_END), "unexpected"); 2021 int op = vop - VectorSupport::VECTOR_OP_SVML_START; 2022 2023 switch(bits) { 2024 case 64: //fallthough 2025 case 128: //fallthough 2026 case 256: //fallthough 2027 case 512: 2028 if (bt == T_FLOAT) { 2029 snprintf(name_ptr, name_len, "vector_%s_float%d", VectorSupport::svmlname[op], bits); 2030 addr = StubRoutines::_vector_f_math[exact_log2(bits/64)][op]; 2031 } else { 2032 assert(bt == T_DOUBLE, "must be FP type only"); 2033 snprintf(name_ptr, name_len, "vector_%s_double%d", VectorSupport::svmlname[op], bits); 2034 addr = StubRoutines::_vector_d_math[exact_log2(bits/64)][op]; 2035 } 2036 break; 2037 default: 2038 snprintf(name_ptr, name_len, "invalid"); 2039 addr = nullptr; 2040 Unimplemented(); 2041 break; 2042 } 2043 2044 return addr; 2045 } 2046 2047 Node* LibraryCallKit::gen_call_to_svml(int vector_api_op_id, BasicType bt, int num_elem, Node* opd1, Node* opd2) { 2048 assert(UseVectorStubs, "sanity"); 2049 assert(vector_api_op_id >= VectorSupport::VECTOR_OP_SVML_START && vector_api_op_id <= VectorSupport::VECTOR_OP_SVML_END, "need valid op id"); 2050 assert(opd1 != nullptr, "must not be null"); 2051 const TypeVect* vt = TypeVect::make(bt, num_elem); 2052 const TypeFunc* call_type = OptoRuntime::Math_Vector_Vector_Type(opd2 != nullptr ? 2 : 1, vt, vt); 2053 char name[100] = ""; 2054 2055 // Get address for svml method. 2056 address addr = get_svml_address(vector_api_op_id, vt->length_in_bytes() * BitsPerByte, bt, name, 100); 2057 2058 if (addr == nullptr) { 2059 return nullptr; 2060 } 2061 2062 assert(name[0] != '\0', "name must not be null"); 2063 Node* operation = make_runtime_call(RC_VECTOR, 2064 call_type, 2065 addr, 2066 name, 2067 TypePtr::BOTTOM, 2068 opd1, 2069 opd2); 2070 return gvn().transform(new ProjNode(gvn().transform(operation), TypeFunc::Parms)); 2071 } 2072 2073 // public static 2074 // <V extends Vector<E>, 2075 // M extends VectorMask<E>, 2076 // E> 2077 // V broadcastInt(int opr, Class<? extends V> vectorClass, Class<? extends M> maskClass, 2078 // Class<E> elementType, int length, 2079 // V v, int n, M m, 2080 // VectorBroadcastIntOp<V, M> defaultImpl) 2081 bool LibraryCallKit::inline_vector_broadcast_int() { 2082 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 2083 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 2084 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 2085 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 2086 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 2087 2088 if (opr == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 2089 return false; // dead code 2090 } 2091 if (!opr->is_con() || vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 2092 log_if_needed(" ** missing constant: opr=%s vclass=%s etype=%s vlen=%s", 2093 NodeClassNames[argument(0)->Opcode()], 2094 NodeClassNames[argument(1)->Opcode()], 2095 NodeClassNames[argument(3)->Opcode()], 2096 NodeClassNames[argument(4)->Opcode()]); 2097 return false; // not enough info for intrinsification 2098 } 2099 if (!is_klass_initialized(vector_klass)) { 2100 log_if_needed(" ** klass argument not initialized"); 2101 return false; 2102 } 2103 2104 const Type* vmask_type = gvn().type(argument(7)); 2105 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 2106 if (is_masked_op) { 2107 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 2108 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(2)->Opcode()]); 2109 return false; // not enough info for intrinsification 2110 } 2111 2112 if (!is_klass_initialized(mask_klass)) { 2113 log_if_needed(" ** mask klass argument not initialized"); 2114 return false; 2115 } 2116 2117 if (vmask_type->maybe_null()) { 2118 log_if_needed(" ** null mask values are not allowed for masked op"); 2119 return false; 2120 } 2121 } 2122 2123 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2124 if (!elem_type->is_primitive_type()) { 2125 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2126 return false; // should be primitive type 2127 } 2128 2129 int num_elem = vlen->get_con(); 2130 BasicType elem_bt = elem_type->basic_type(); 2131 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 2132 2133 bool is_shift = VectorNode::is_shift_opcode(opc); 2134 bool is_rotate = VectorNode::is_rotate_opcode(opc); 2135 2136 if (opc == 0 || (!is_shift && !is_rotate)) { 2137 log_if_needed(" ** operation not supported: op=%d bt=%s", opr->get_con(), type2name(elem_bt)); 2138 return false; // operation not supported 2139 } 2140 2141 int sopc = VectorNode::opcode(opc, elem_bt); 2142 if (sopc == 0) { 2143 log_if_needed(" ** operation not supported: opc=%s bt=%s", NodeClassNames[opc], type2name(elem_bt)); 2144 return false; // operation not supported 2145 } 2146 2147 Node* cnt = argument(6); 2148 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2149 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2150 const TypeInt* cnt_type = cnt->bottom_type()->isa_int(); 2151 2152 // If CPU supports vector constant rotate instructions pass it directly 2153 bool is_const_rotate = is_rotate && cnt_type && cnt_type->is_con() && 2154 Matcher::supports_vector_constant_rotates(cnt_type->get_con()); 2155 bool has_scalar_args = is_rotate ? !is_const_rotate : true; 2156 2157 VectorMaskUseType checkFlags = (VectorMaskUseType)(is_masked_op ? (VecMaskUseLoad | VecMaskUsePred) : VecMaskNotUsed); 2158 bool use_predicate = is_masked_op; 2159 2160 if (!arch_supports_vector(sopc, num_elem, elem_bt, checkFlags, has_scalar_args)) { 2161 use_predicate = false; 2162 if (!is_masked_op || 2163 (!arch_supports_vector(sopc, num_elem, elem_bt, VecMaskNotUsed, has_scalar_args) || 2164 !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad))) { 2165 2166 log_if_needed(" ** not supported: arity=0 op=int/%d vlen=%d etype=%s is_masked_op=%d", 2167 sopc, num_elem, type2name(elem_bt), is_masked_op ? 1 : 0); 2168 return false; // not supported 2169 } 2170 } 2171 2172 Node* opd1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 2173 Node* opd2 = nullptr; 2174 if (is_shift) { 2175 opd2 = vector_shift_count(cnt, opc, elem_bt, num_elem); 2176 } else { 2177 assert(is_rotate, "unexpected operation"); 2178 if (!is_const_rotate) { 2179 const Type * type_bt = Type::get_const_basic_type(elem_bt); 2180 cnt = elem_bt == T_LONG ? gvn().transform(new ConvI2LNode(cnt)) : cnt; 2181 opd2 = gvn().transform(VectorNode::scalar2vector(cnt, num_elem, type_bt)); 2182 } else { 2183 // Constant shift value. 2184 opd2 = cnt; 2185 } 2186 } 2187 2188 if (opd1 == nullptr || opd2 == nullptr) { 2189 return false; 2190 } 2191 2192 Node* mask = nullptr; 2193 if (is_masked_op) { 2194 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2195 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 2196 mask = unbox_vector(argument(7), mbox_type, elem_bt, num_elem); 2197 if (mask == nullptr) { 2198 log_if_needed(" ** unbox failed mask=%s", NodeClassNames[argument(7)->Opcode()]); 2199 return false; 2200 } 2201 } 2202 2203 Node* operation = VectorNode::make(opc, opd1, opd2, num_elem, elem_bt); 2204 if (is_masked_op && mask != nullptr) { 2205 if (use_predicate) { 2206 operation->add_req(mask); 2207 operation->add_flag(Node::Flag_is_predicated_vector); 2208 } else { 2209 operation = gvn().transform(operation); 2210 operation = new VectorBlendNode(opd1, operation, mask); 2211 } 2212 } 2213 operation = gvn().transform(operation); 2214 Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem); 2215 set_result(vbox); 2216 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2217 return true; 2218 } 2219 2220 // public static <VOUT extends VectorPayload, 2221 // VIN extends VectorPayload, 2222 // S extends VectorSpecies> 2223 // VOUT convert(int oprId, 2224 // Class<?> fromVectorClass, Class<?> fromElementType, int fromVLen, 2225 // Class<?> toVectorClass, Class<?> toElementType, int toVLen, 2226 // VIN v, S s, 2227 // VectorConvertOp<VOUT, VIN, S> defaultImpl) 2228 // 2229 bool LibraryCallKit::inline_vector_convert() { 2230 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 2231 2232 const TypeInstPtr* vector_klass_from = gvn().type(argument(1))->isa_instptr(); 2233 const TypeInstPtr* elem_klass_from = gvn().type(argument(2))->isa_instptr(); 2234 const TypeInt* vlen_from = gvn().type(argument(3))->isa_int(); 2235 2236 const TypeInstPtr* vector_klass_to = gvn().type(argument(4))->isa_instptr(); 2237 const TypeInstPtr* elem_klass_to = gvn().type(argument(5))->isa_instptr(); 2238 const TypeInt* vlen_to = gvn().type(argument(6))->isa_int(); 2239 2240 if (opr == nullptr || 2241 vector_klass_from == nullptr || elem_klass_from == nullptr || vlen_from == nullptr || 2242 vector_klass_to == nullptr || elem_klass_to == nullptr || vlen_to == nullptr) { 2243 return false; // dead code 2244 } 2245 if (!opr->is_con() || 2246 vector_klass_from->const_oop() == nullptr || elem_klass_from->const_oop() == nullptr || !vlen_from->is_con() || 2247 vector_klass_to->const_oop() == nullptr || elem_klass_to->const_oop() == nullptr || !vlen_to->is_con()) { 2248 log_if_needed(" ** missing constant: opr=%s vclass_from=%s etype_from=%s vlen_from=%s vclass_to=%s etype_to=%s vlen_to=%s", 2249 NodeClassNames[argument(0)->Opcode()], 2250 NodeClassNames[argument(1)->Opcode()], 2251 NodeClassNames[argument(2)->Opcode()], 2252 NodeClassNames[argument(3)->Opcode()], 2253 NodeClassNames[argument(4)->Opcode()], 2254 NodeClassNames[argument(5)->Opcode()], 2255 NodeClassNames[argument(6)->Opcode()]); 2256 return false; // not enough info for intrinsification 2257 } 2258 if (!is_klass_initialized(vector_klass_from) || !is_klass_initialized(vector_klass_to)) { 2259 log_if_needed(" ** klass argument not initialized"); 2260 return false; 2261 } 2262 2263 assert(opr->get_con() == VectorSupport::VECTOR_OP_CAST || 2264 opr->get_con() == VectorSupport::VECTOR_OP_UCAST || 2265 opr->get_con() == VectorSupport::VECTOR_OP_REINTERPRET, "wrong opcode"); 2266 bool is_cast = (opr->get_con() == VectorSupport::VECTOR_OP_CAST || opr->get_con() == VectorSupport::VECTOR_OP_UCAST); 2267 bool is_ucast = (opr->get_con() == VectorSupport::VECTOR_OP_UCAST); 2268 2269 ciKlass* vbox_klass_from = vector_klass_from->const_oop()->as_instance()->java_lang_Class_klass(); 2270 ciKlass* vbox_klass_to = vector_klass_to->const_oop()->as_instance()->java_lang_Class_klass(); 2271 if (is_vector_shuffle(vbox_klass_from)) { 2272 return false; // vector shuffles aren't supported 2273 } 2274 bool is_mask = is_vector_mask(vbox_klass_from); 2275 2276 ciType* elem_type_from = elem_klass_from->const_oop()->as_instance()->java_mirror_type(); 2277 if (!elem_type_from->is_primitive_type()) { 2278 return false; // should be primitive type 2279 } 2280 BasicType elem_bt_from = elem_type_from->basic_type(); 2281 ciType* elem_type_to = elem_klass_to->const_oop()->as_instance()->java_mirror_type(); 2282 if (!elem_type_to->is_primitive_type()) { 2283 return false; // should be primitive type 2284 } 2285 BasicType elem_bt_to = elem_type_to->basic_type(); 2286 2287 int num_elem_from = vlen_from->get_con(); 2288 int num_elem_to = vlen_to->get_con(); 2289 2290 // Check whether we can unbox to appropriate size. Even with casting, checking for reinterpret is needed 2291 // since we may need to change size. 2292 if (!arch_supports_vector(Op_VectorReinterpret, 2293 num_elem_from, 2294 elem_bt_from, 2295 is_mask ? VecMaskUseAll : VecMaskNotUsed)) { 2296 log_if_needed(" ** not supported: arity=1 op=%s/1 vlen1=%d etype1=%s ismask=%d", 2297 is_cast ? "cast" : "reinterpret", 2298 num_elem_from, type2name(elem_bt_from), is_mask); 2299 return false; 2300 } 2301 2302 // Check whether we can support resizing/reinterpreting to the new size. 2303 if (!arch_supports_vector(Op_VectorReinterpret, 2304 num_elem_to, 2305 elem_bt_to, 2306 is_mask ? VecMaskUseAll : VecMaskNotUsed)) { 2307 log_if_needed(" ** not supported: arity=1 op=%s/2 vlen2=%d etype2=%s ismask=%d", 2308 is_cast ? "cast" : "reinterpret", 2309 num_elem_to, type2name(elem_bt_to), is_mask); 2310 return false; 2311 } 2312 2313 2314 if (is_vector_shuffle(vbox_klass_to) && 2315 (!arch_supports_vector(Op_SubVB, num_elem_to, elem_bt_to, VecMaskNotUsed) || 2316 !arch_supports_vector(Op_VectorBlend, num_elem_to, elem_bt_to, VecMaskNotUsed) || 2317 !arch_supports_vector(Op_VectorMaskCmp, num_elem_to, elem_bt_to, VecMaskNotUsed) || 2318 !arch_supports_vector(Op_AndV, num_elem_to, elem_bt_to, VecMaskNotUsed) || 2319 !arch_supports_vector(Op_Replicate, num_elem_to, elem_bt_to, VecMaskNotUsed))) { 2320 log_if_needed(" ** not supported: arity=1 op=shuffle_index_wrap vlen2=%d etype2=%s", 2321 num_elem_to, type2name(elem_bt_to)); 2322 return false; 2323 } 2324 2325 // At this point, we know that both input and output vector registers are supported 2326 // by the architecture. Next check if the casted type is simply to same type - which means 2327 // that it is actually a resize and not a cast. 2328 if (is_cast && elem_bt_from == elem_bt_to) { 2329 is_cast = false; 2330 } 2331 2332 const TypeInstPtr* vbox_type_from = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass_from); 2333 2334 Node* opd1 = unbox_vector(argument(7), vbox_type_from, elem_bt_from, num_elem_from); 2335 if (opd1 == nullptr) { 2336 return false; 2337 } 2338 2339 const TypeVect* src_type = TypeVect::make(elem_bt_from, num_elem_from, is_mask); 2340 const TypeVect* dst_type = TypeVect::make(elem_bt_to, num_elem_to, is_mask); 2341 2342 // Safety check to prevent casting if source mask is of type vector 2343 // and destination mask of type predicate vector and vice-versa. 2344 // From X86 standpoint, this case will only arise over KNL target, 2345 // where certain masks (depending on the species) are either propagated 2346 // through a vector or predicate register. 2347 if (is_mask && 2348 ((src_type->isa_vectmask() == nullptr && dst_type->isa_vectmask()) || 2349 (dst_type->isa_vectmask() == nullptr && src_type->isa_vectmask()))) { 2350 return false; 2351 } 2352 2353 Node* op = opd1; 2354 if (is_cast) { 2355 assert(!is_mask || num_elem_from == num_elem_to, "vector mask cast needs the same elem num"); 2356 int cast_vopc = VectorCastNode::opcode(-1, elem_bt_from, !is_ucast); 2357 2358 // Make sure that vector cast is implemented to particular type/size combination if it is 2359 // not a mask casting. 2360 if (!is_mask && !arch_supports_vector(cast_vopc, num_elem_to, elem_bt_to, VecMaskNotUsed)) { 2361 log_if_needed(" ** not supported: arity=1 op=cast#%d/3 vlen2=%d etype2=%s ismask=%d", 2362 cast_vopc, num_elem_to, type2name(elem_bt_to), is_mask); 2363 return false; 2364 } 2365 2366 if (num_elem_from < num_elem_to) { 2367 // Since input and output number of elements are not consistent, we need to make sure we 2368 // properly size. Thus, first make a cast that retains the number of elements from source. 2369 int num_elem_for_cast = num_elem_from; 2370 2371 // It is possible that arch does not support this intermediate vector size 2372 // TODO More complex logic required here to handle this corner case for the sizes. 2373 if (!arch_supports_vector(cast_vopc, num_elem_for_cast, elem_bt_to, VecMaskNotUsed)) { 2374 log_if_needed(" ** not supported: arity=1 op=cast#%d/4 vlen1=%d etype2=%s ismask=%d", 2375 cast_vopc, 2376 num_elem_for_cast, type2name(elem_bt_to), is_mask); 2377 return false; 2378 } 2379 2380 op = gvn().transform(VectorCastNode::make(cast_vopc, op, elem_bt_to, num_elem_for_cast)); 2381 // Now ensure that the destination gets properly resized to needed size. 2382 op = gvn().transform(new VectorReinterpretNode(op, op->bottom_type()->is_vect(), dst_type)); 2383 } else if (num_elem_from > num_elem_to) { 2384 // Since number of elements from input is larger than output, simply reduce size of input 2385 // (we are supposed to drop top elements anyway). 2386 int num_elem_for_resize = num_elem_to; 2387 2388 // It is possible that arch does not support this intermediate vector size 2389 // TODO More complex logic required here to handle this corner case for the sizes. 2390 if (!arch_supports_vector(Op_VectorReinterpret, 2391 num_elem_for_resize, 2392 elem_bt_from, 2393 VecMaskNotUsed)) { 2394 log_if_needed(" ** not supported: arity=1 op=cast/5 vlen2=%d etype1=%s ismask=%d", 2395 num_elem_for_resize, type2name(elem_bt_from), is_mask); 2396 return false; 2397 } 2398 2399 const TypeVect* resize_type = TypeVect::make(elem_bt_from, num_elem_for_resize); 2400 op = gvn().transform(new VectorReinterpretNode(op, src_type, resize_type)); 2401 op = gvn().transform(VectorCastNode::make(cast_vopc, op, elem_bt_to, num_elem_to)); 2402 } else { // num_elem_from == num_elem_to 2403 if (is_mask) { 2404 // Make sure that cast for vector mask is implemented to particular type/size combination. 2405 if (!arch_supports_vector(Op_VectorMaskCast, num_elem_to, elem_bt_to, VecMaskNotUsed)) { 2406 log_if_needed(" ** not supported: arity=1 op=maskcast vlen2=%d etype2=%s ismask=%d", 2407 num_elem_to, type2name(elem_bt_to), is_mask); 2408 return false; 2409 } 2410 op = gvn().transform(new VectorMaskCastNode(op, dst_type)); 2411 } else { 2412 // Since input and output number of elements match, and since we know this vector size is 2413 // supported, simply do a cast with no resize needed. 2414 op = gvn().transform(VectorCastNode::make(cast_vopc, op, elem_bt_to, num_elem_to)); 2415 } 2416 } 2417 } else if (!Type::equals(src_type, dst_type)) { 2418 assert(!is_cast, "must be reinterpret"); 2419 op = gvn().transform(new VectorReinterpretNode(op, src_type, dst_type)); 2420 } 2421 2422 if (is_vector_shuffle(vbox_klass_to)) { 2423 op = partially_wrap_indexes(op, num_elem_to, elem_bt_to); 2424 } 2425 2426 const TypeInstPtr* vbox_type_to = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass_to); 2427 Node* vbox = box_vector(op, vbox_type_to, elem_bt_to, num_elem_to); 2428 set_result(vbox); 2429 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem_to * type2aelembytes(elem_bt_to)))); 2430 return true; 2431 } 2432 2433 // public static 2434 // <V extends Vector<E>, 2435 // E> 2436 // V insert(Class<? extends V> vectorClass, Class<E> elementType, int vlen, 2437 // V vec, int ix, long val, 2438 // VecInsertOp<V> defaultImpl) 2439 bool LibraryCallKit::inline_vector_insert() { 2440 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 2441 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2442 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2443 const TypeInt* idx = gvn().type(argument(4))->isa_int(); 2444 2445 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || idx == nullptr) { 2446 return false; // dead code 2447 } 2448 if (vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con() || !idx->is_con()) { 2449 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s idx=%s", 2450 NodeClassNames[argument(0)->Opcode()], 2451 NodeClassNames[argument(1)->Opcode()], 2452 NodeClassNames[argument(2)->Opcode()], 2453 NodeClassNames[argument(4)->Opcode()]); 2454 return false; // not enough info for intrinsification 2455 } 2456 if (!is_klass_initialized(vector_klass)) { 2457 log_if_needed(" ** klass argument not initialized"); 2458 return false; 2459 } 2460 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2461 if (!elem_type->is_primitive_type()) { 2462 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2463 return false; // should be primitive type 2464 } 2465 BasicType elem_bt = elem_type->basic_type(); 2466 int num_elem = vlen->get_con(); 2467 if (!arch_supports_vector(Op_VectorInsert, num_elem, elem_bt, VecMaskNotUsed)) { 2468 log_if_needed(" ** not supported: arity=1 op=insert vlen=%d etype=%s ismask=no", 2469 num_elem, type2name(elem_bt)); 2470 return false; // not supported 2471 } 2472 2473 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2474 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2475 2476 Node* opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2477 if (opd == nullptr) { 2478 return false; 2479 } 2480 2481 Node* insert_val = argument(5); 2482 assert(gvn().type(insert_val)->isa_long() != nullptr, "expected to be long"); 2483 2484 // Convert insert value back to its appropriate type. 2485 switch (elem_bt) { 2486 case T_BYTE: 2487 insert_val = gvn().transform(new ConvL2INode(insert_val)); 2488 insert_val = gvn().transform(new CastIINode(insert_val, TypeInt::BYTE)); 2489 break; 2490 case T_SHORT: 2491 insert_val = gvn().transform(new ConvL2INode(insert_val)); 2492 insert_val = gvn().transform(new CastIINode(insert_val, TypeInt::SHORT)); 2493 break; 2494 case T_INT: 2495 insert_val = gvn().transform(new ConvL2INode(insert_val)); 2496 break; 2497 case T_FLOAT: 2498 insert_val = gvn().transform(new ConvL2INode(insert_val)); 2499 insert_val = gvn().transform(new MoveI2FNode(insert_val)); 2500 break; 2501 case T_DOUBLE: 2502 insert_val = gvn().transform(new MoveL2DNode(insert_val)); 2503 break; 2504 case T_LONG: 2505 // no conversion needed 2506 break; 2507 default: fatal("%s", type2name(elem_bt)); break; 2508 } 2509 2510 Node* operation = gvn().transform(VectorInsertNode::make(opd, insert_val, idx->get_con(), gvn())); 2511 2512 Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem); 2513 set_result(vbox); 2514 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2515 return true; 2516 } 2517 2518 // public static 2519 // <VM extends VectorPayload, 2520 // E> 2521 // long extract(Class<? extends VM> vClass, Class<E> eClass, 2522 // int length, 2523 // VM vm, int i, 2524 // VecExtractOp<VM> defaultImpl) 2525 bool LibraryCallKit::inline_vector_extract() { 2526 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 2527 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2528 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2529 const TypeInt* idx = gvn().type(argument(4))->isa_int(); 2530 2531 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || idx == nullptr) { 2532 return false; // dead code 2533 } 2534 if (vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 2535 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s", 2536 NodeClassNames[argument(0)->Opcode()], 2537 NodeClassNames[argument(1)->Opcode()], 2538 NodeClassNames[argument(2)->Opcode()]); 2539 return false; // not enough info for intrinsification 2540 } 2541 if (!is_klass_initialized(vector_klass)) { 2542 log_if_needed(" ** klass argument not initialized"); 2543 return false; 2544 } 2545 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2546 if (!elem_type->is_primitive_type()) { 2547 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2548 return false; // should be primitive type 2549 } 2550 BasicType elem_bt = elem_type->basic_type(); 2551 int num_elem = vlen->get_con(); 2552 2553 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2554 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2555 2556 Node* opd = nullptr; 2557 2558 if (is_vector_mask(vbox_klass)) { 2559 // vbox_klass is mask. This is used for VectorMask.laneIsSet(int). 2560 2561 Node* pos = argument(4); // can be variable 2562 if (arch_supports_vector(Op_ExtractUB, num_elem, elem_bt, VecMaskUseAll)) { 2563 // Transform mask to vector with type of boolean and utilize ExtractUB node. 2564 opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2565 if (opd == nullptr) { 2566 return false; 2567 } 2568 opd = gvn().transform(VectorStoreMaskNode::make(gvn(), opd, elem_bt, num_elem)); 2569 opd = gvn().transform(new ExtractUBNode(opd, pos)); 2570 opd = gvn().transform(new ConvI2LNode(opd)); 2571 } else if (arch_supports_vector(Op_VectorMaskToLong, num_elem, elem_bt, VecMaskUseLoad)) { 2572 opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2573 if (opd == nullptr) { 2574 return false; 2575 } 2576 // VectorMaskToLongNode requires the input is either a mask or a vector with BOOLEAN type. 2577 if (opd->bottom_type()->isa_vectmask() == nullptr) { 2578 opd = gvn().transform(VectorStoreMaskNode::make(gvn(), opd, elem_bt, num_elem)); 2579 } 2580 // ((toLong() >>> pos) & 1L 2581 opd = gvn().transform(new VectorMaskToLongNode(opd, TypeLong::LONG)); 2582 opd = gvn().transform(new URShiftLNode(opd, pos)); 2583 opd = gvn().transform(new AndLNode(opd, gvn().makecon(TypeLong::ONE))); 2584 } else { 2585 log_if_needed(" ** Rejected mask extraction because architecture does not support it"); 2586 return false; // not supported 2587 } 2588 } else { 2589 // vbox_klass is vector. This is used for Vector.lane(int). 2590 if (!idx->is_con()) { 2591 log_if_needed(" ** missing constant: idx=%s", NodeClassNames[argument(4)->Opcode()]); 2592 return false; // not enough info for intrinsification 2593 } 2594 2595 int vopc = ExtractNode::opcode(elem_bt); 2596 if (!arch_supports_vector(vopc, num_elem, elem_bt, VecMaskNotUsed)) { 2597 log_if_needed(" ** not supported: arity=1 op=extract vlen=%d etype=%s ismask=no", 2598 num_elem, type2name(elem_bt)); 2599 return false; // not supported 2600 } 2601 2602 opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2603 if (opd == nullptr) { 2604 return false; 2605 } 2606 ConINode* idx_con = gvn().intcon(idx->get_con())->as_ConI(); 2607 2608 opd = gvn().transform(ExtractNode::make(opd, idx_con, elem_bt)); 2609 switch (elem_bt) { 2610 case T_BYTE: 2611 case T_SHORT: 2612 case T_INT: { 2613 opd = gvn().transform(new ConvI2LNode(opd)); 2614 break; 2615 } 2616 case T_FLOAT: { 2617 opd = gvn().transform(new MoveF2INode(opd)); 2618 opd = gvn().transform(new ConvI2LNode(opd)); 2619 break; 2620 } 2621 case T_DOUBLE: { 2622 opd = gvn().transform(new MoveD2LNode(opd)); 2623 break; 2624 } 2625 case T_LONG: { 2626 // no conversion needed 2627 break; 2628 } 2629 default: fatal("%s", type2name(elem_bt)); 2630 } 2631 } 2632 set_result(opd); 2633 return true; 2634 } 2635 2636 // public static 2637 // <V extends Vector<E>, 2638 // M extends VectorMask<E>, 2639 // E> 2640 // V compressExpandOp(int opr, 2641 // Class<? extends V> vClass, Class<? extends M> mClass, Class<E> eClass, 2642 // int length, V v, M m, 2643 // CompressExpandOperation<V, M> defaultImpl) 2644 bool LibraryCallKit::inline_vector_compress_expand() { 2645 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 2646 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 2647 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 2648 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 2649 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 2650 2651 if (vector_klass == nullptr || elem_klass == nullptr || mask_klass == nullptr || vlen == nullptr || 2652 vector_klass->const_oop() == nullptr || mask_klass->const_oop() == nullptr || 2653 elem_klass->const_oop() == nullptr || !vlen->is_con() || !opr->is_con()) { 2654 log_if_needed(" ** missing constant: opr=%s vclass=%s mclass=%s etype=%s vlen=%s", 2655 NodeClassNames[argument(0)->Opcode()], 2656 NodeClassNames[argument(1)->Opcode()], 2657 NodeClassNames[argument(2)->Opcode()], 2658 NodeClassNames[argument(3)->Opcode()], 2659 NodeClassNames[argument(4)->Opcode()]); 2660 return false; // not enough info for intrinsification 2661 } 2662 2663 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(mask_klass)) { 2664 log_if_needed(" ** klass argument not initialized"); 2665 return false; 2666 } 2667 2668 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2669 if (!elem_type->is_primitive_type()) { 2670 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2671 return false; // should be primitive type 2672 } 2673 2674 int num_elem = vlen->get_con(); 2675 BasicType elem_bt = elem_type->basic_type(); 2676 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 2677 2678 if (!arch_supports_vector(opc, num_elem, elem_bt, VecMaskUseLoad)) { 2679 log_if_needed(" ** not supported: opc=%d vlen=%d etype=%s ismask=useload", 2680 opc, num_elem, type2name(elem_bt)); 2681 return false; // not supported 2682 } 2683 2684 Node* opd1 = nullptr; 2685 const TypeInstPtr* vbox_type = nullptr; 2686 if (opc != Op_CompressM) { 2687 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2688 vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2689 opd1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 2690 if (opd1 == nullptr) { 2691 log_if_needed(" ** unbox failed vector=%s", 2692 NodeClassNames[argument(5)->Opcode()]); 2693 return false; 2694 } 2695 } 2696 2697 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2698 assert(is_vector_mask(mbox_klass), "argument(6) should be a mask class"); 2699 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 2700 2701 Node* mask = unbox_vector(argument(6), mbox_type, elem_bt, num_elem); 2702 if (mask == nullptr) { 2703 log_if_needed(" ** unbox failed mask=%s", 2704 NodeClassNames[argument(6)->Opcode()]); 2705 return false; 2706 } 2707 2708 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, opc == Op_CompressM); 2709 Node* operation = gvn().transform(VectorNode::make(opc, opd1, mask, vt)); 2710 2711 // Wrap it up in VectorBox to keep object type information. 2712 const TypeInstPtr* box_type = opc == Op_CompressM ? mbox_type : vbox_type; 2713 Node* vbox = box_vector(operation, box_type, elem_bt, num_elem); 2714 set_result(vbox); 2715 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2716 return true; 2717 } 2718 2719 // public static 2720 // <V extends Vector<E>, 2721 // E, 2722 // S extends VectorSpecies<E>> 2723 // V indexVector(Class<? extends V> vClass, Class<E> eClass, 2724 // int length, 2725 // V v, int step, S s, 2726 // IndexOperation<V, S> defaultImpl) 2727 bool LibraryCallKit::inline_index_vector() { 2728 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 2729 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2730 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2731 2732 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 2733 vector_klass->const_oop() == nullptr || !vlen->is_con() || 2734 elem_klass->const_oop() == nullptr) { 2735 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s", 2736 NodeClassNames[argument(0)->Opcode()], 2737 NodeClassNames[argument(1)->Opcode()], 2738 NodeClassNames[argument(2)->Opcode()]); 2739 return false; // not enough info for intrinsification 2740 } 2741 2742 if (!is_klass_initialized(vector_klass)) { 2743 log_if_needed(" ** klass argument not initialized"); 2744 return false; 2745 } 2746 2747 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2748 if (!elem_type->is_primitive_type()) { 2749 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2750 return false; // should be primitive type 2751 } 2752 2753 int num_elem = vlen->get_con(); 2754 BasicType elem_bt = elem_type->basic_type(); 2755 2756 // Check whether the iota index generation op is supported by the current hardware 2757 if (!arch_supports_vector(Op_VectorLoadConst, num_elem, elem_bt, VecMaskNotUsed)) { 2758 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 2759 return false; // not supported 2760 } 2761 2762 int mul_op = VectorSupport::vop2ideal(VectorSupport::VECTOR_OP_MUL, elem_bt); 2763 int vmul_op = VectorNode::opcode(mul_op, elem_bt); 2764 bool needs_mul = true; 2765 Node* scale = argument(4); 2766 const TypeInt* scale_type = gvn().type(scale)->isa_int(); 2767 // Multiply is not needed if the scale is a constant "1". 2768 if (scale_type && scale_type->is_con() && scale_type->get_con() == 1) { 2769 needs_mul = false; 2770 } else { 2771 // Check whether the vector multiply op is supported by the current hardware 2772 if (!arch_supports_vector(vmul_op, num_elem, elem_bt, VecMaskNotUsed)) { 2773 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 2774 return false; // not supported 2775 } 2776 2777 // Check whether the scalar cast op is supported by the current hardware 2778 if (is_floating_point_type(elem_bt) || elem_bt == T_LONG) { 2779 int cast_op = elem_bt == T_LONG ? Op_ConvI2L : 2780 elem_bt == T_FLOAT? Op_ConvI2F : Op_ConvI2D; 2781 if (!Matcher::match_rule_supported(cast_op)) { 2782 log_if_needed(" ** Rejected op (%s) because architecture does not support it", 2783 NodeClassNames[cast_op]); 2784 return false; // not supported 2785 } 2786 } 2787 } 2788 2789 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2790 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2791 Node* opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2792 if (opd == nullptr) { 2793 log_if_needed(" ** unbox failed vector=%s", 2794 NodeClassNames[argument(3)->Opcode()]); 2795 return false; 2796 } 2797 2798 int add_op = VectorSupport::vop2ideal(VectorSupport::VECTOR_OP_ADD, elem_bt); 2799 int vadd_op = VectorNode::opcode(add_op, elem_bt); 2800 bool needs_add = true; 2801 // The addition is not needed if all the element values of "opd" are zero 2802 if (VectorNode::is_all_zeros_vector(opd)) { 2803 needs_add = false; 2804 } else { 2805 // Check whether the vector addition op is supported by the current hardware 2806 if (!arch_supports_vector(vadd_op, num_elem, elem_bt, VecMaskNotUsed)) { 2807 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 2808 return false; // not supported 2809 } 2810 } 2811 2812 // Compute the iota indice vector 2813 const TypeVect* vt = TypeVect::make(elem_bt, num_elem); 2814 Node* index = gvn().transform(new VectorLoadConstNode(gvn().makecon(TypeInt::ZERO), vt)); 2815 2816 // Broadcast the "scale" to a vector, and multiply the "scale" with iota indice vector. 2817 if (needs_mul) { 2818 switch (elem_bt) { 2819 case T_BOOLEAN: // fall-through 2820 case T_BYTE: // fall-through 2821 case T_SHORT: // fall-through 2822 case T_CHAR: // fall-through 2823 case T_INT: { 2824 // no conversion needed 2825 break; 2826 } 2827 case T_LONG: { 2828 scale = gvn().transform(new ConvI2LNode(scale)); 2829 break; 2830 } 2831 case T_FLOAT: { 2832 scale = gvn().transform(new ConvI2FNode(scale)); 2833 break; 2834 } 2835 case T_DOUBLE: { 2836 scale = gvn().transform(new ConvI2DNode(scale)); 2837 break; 2838 } 2839 default: fatal("%s", type2name(elem_bt)); 2840 } 2841 scale = gvn().transform(VectorNode::scalar2vector(scale, num_elem, Type::get_const_basic_type(elem_bt))); 2842 index = gvn().transform(VectorNode::make(vmul_op, index, scale, vt)); 2843 } 2844 2845 // Add "opd" if addition is needed. 2846 if (needs_add) { 2847 index = gvn().transform(VectorNode::make(vadd_op, opd, index, vt)); 2848 } 2849 Node* vbox = box_vector(index, vbox_type, elem_bt, num_elem); 2850 set_result(vbox); 2851 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2852 return true; 2853 } 2854 2855 // public static 2856 // <E, 2857 // M extends VectorMask<E>> 2858 // M indexPartiallyInUpperRange(Class<? extends M> mClass, Class<E> eClass, int length, 2859 // long offset, long limit, 2860 // IndexPartiallyInUpperRangeOperation<E, M> defaultImpl) 2861 bool LibraryCallKit::inline_index_partially_in_upper_range() { 2862 const TypeInstPtr* mask_klass = gvn().type(argument(0))->isa_instptr(); 2863 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2864 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2865 2866 if (mask_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 2867 mask_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 2868 log_if_needed(" ** missing constant: mclass=%s etype=%s vlen=%s", 2869 NodeClassNames[argument(0)->Opcode()], 2870 NodeClassNames[argument(1)->Opcode()], 2871 NodeClassNames[argument(2)->Opcode()]); 2872 return false; // not enough info for intrinsification 2873 } 2874 2875 if (!is_klass_initialized(mask_klass)) { 2876 log_if_needed(" ** klass argument not initialized"); 2877 return false; 2878 } 2879 2880 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2881 if (!elem_type->is_primitive_type()) { 2882 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2883 return false; // should be primitive type 2884 } 2885 2886 int num_elem = vlen->get_con(); 2887 BasicType elem_bt = elem_type->basic_type(); 2888 2889 // Check whether the necessary ops are supported by current hardware. 2890 bool supports_mask_gen = arch_supports_vector(Op_VectorMaskGen, num_elem, elem_bt, VecMaskUseStore); 2891 if (!supports_mask_gen) { 2892 if (!arch_supports_vector(Op_VectorLoadConst, num_elem, elem_bt, VecMaskNotUsed) || 2893 !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed) || 2894 !arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskUseStore)) { 2895 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 2896 return false; // not supported 2897 } 2898 2899 // Check whether the scalar cast operation is supported by current hardware. 2900 if (elem_bt != T_LONG) { 2901 int cast_op = is_integral_type(elem_bt) ? Op_ConvL2I 2902 : (elem_bt == T_FLOAT ? Op_ConvL2F : Op_ConvL2D); 2903 if (!Matcher::match_rule_supported(cast_op)) { 2904 log_if_needed(" ** Rejected op (%s) because architecture does not support it", 2905 NodeClassNames[cast_op]); 2906 return false; // not supported 2907 } 2908 } 2909 } 2910 2911 Node* offset = argument(3); 2912 Node* limit = argument(5); 2913 if (offset == nullptr || limit == nullptr) { 2914 log_if_needed(" ** offset or limit argument is null"); 2915 return false; // not supported 2916 } 2917 2918 ciKlass* box_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2919 assert(is_vector_mask(box_klass), "argument(0) should be a mask class"); 2920 const TypeInstPtr* box_type = TypeInstPtr::make_exact(TypePtr::NotNull, box_klass); 2921 2922 // We assume "offset > 0 && limit >= offset && limit - offset < num_elem". 2923 // So directly get indexLimit with "indexLimit = limit - offset". 2924 Node* indexLimit = gvn().transform(new SubLNode(limit, offset)); 2925 Node* mask = nullptr; 2926 if (supports_mask_gen) { 2927 mask = gvn().transform(VectorMaskGenNode::make(indexLimit, elem_bt, num_elem)); 2928 } else { 2929 // Generate the vector mask based on "mask = iota < indexLimit". 2930 // Broadcast "indexLimit" to a vector. 2931 switch (elem_bt) { 2932 case T_BOOLEAN: // fall-through 2933 case T_BYTE: // fall-through 2934 case T_SHORT: // fall-through 2935 case T_CHAR: // fall-through 2936 case T_INT: { 2937 indexLimit = gvn().transform(new ConvL2INode(indexLimit)); 2938 break; 2939 } 2940 case T_DOUBLE: { 2941 indexLimit = gvn().transform(new ConvL2DNode(indexLimit)); 2942 break; 2943 } 2944 case T_FLOAT: { 2945 indexLimit = gvn().transform(new ConvL2FNode(indexLimit)); 2946 break; 2947 } 2948 case T_LONG: { 2949 // no conversion needed 2950 break; 2951 } 2952 default: fatal("%s", type2name(elem_bt)); 2953 } 2954 indexLimit = gvn().transform(VectorNode::scalar2vector(indexLimit, num_elem, Type::get_const_basic_type(elem_bt))); 2955 2956 // Load the "iota" vector. 2957 const TypeVect* vt = TypeVect::make(elem_bt, num_elem); 2958 Node* iota = gvn().transform(new VectorLoadConstNode(gvn().makecon(TypeInt::ZERO), vt)); 2959 2960 // Compute the vector mask with "mask = iota < indexLimit". 2961 ConINode* pred_node = (ConINode*)gvn().makecon(TypeInt::make(BoolTest::lt)); 2962 const TypeVect* vmask_type = TypeVect::makemask(elem_bt, num_elem); 2963 mask = gvn().transform(new VectorMaskCmpNode(BoolTest::lt, iota, indexLimit, pred_node, vmask_type)); 2964 } 2965 Node* vbox = box_vector(mask, box_type, elem_bt, num_elem); 2966 set_result(vbox); 2967 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2968 return true; 2969 } 2970 2971 #undef non_product_log_if_needed 2972 #undef log_if_needed