1 /* 2 * Copyright (c) 2020, 2024, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #include "precompiled.hpp" 26 #include "ci/ciSymbols.hpp" 27 #include "classfile/vmSymbols.hpp" 28 #include "opto/library_call.hpp" 29 #include "opto/runtime.hpp" 30 #include "opto/vectornode.hpp" 31 #include "prims/vectorSupport.hpp" 32 #include "runtime/stubRoutines.hpp" 33 34 #ifdef ASSERT 35 static bool is_vector(ciKlass* klass) { 36 return klass->is_subclass_of(ciEnv::current()->vector_VectorPayload_klass()); 37 } 38 39 static bool check_vbox(const TypeInstPtr* vbox_type) { 40 assert(vbox_type->klass_is_exact(), ""); 41 42 ciInstanceKlass* ik = vbox_type->instance_klass(); 43 assert(is_vector(ik), "not a vector"); 44 45 ciField* fd1 = ik->get_field_by_name(ciSymbols::ETYPE_name(), ciSymbols::class_signature(), /* is_static */ true); 46 assert(fd1 != nullptr, "element type info is missing"); 47 48 ciConstant val1 = fd1->constant_value(); 49 BasicType elem_bt = val1.as_object()->as_instance()->java_mirror_type()->basic_type(); 50 assert(is_java_primitive(elem_bt), "element type info is missing"); 51 52 ciField* fd2 = ik->get_field_by_name(ciSymbols::VLENGTH_name(), ciSymbols::int_signature(), /* is_static */ true); 53 assert(fd2 != nullptr, "vector length info is missing"); 54 55 ciConstant val2 = fd2->constant_value(); 56 assert(val2.as_int() > 0, "vector length info is missing"); 57 58 return true; 59 } 60 #endif 61 62 #define log_if_needed(...) \ 63 if (C->print_intrinsics()) { \ 64 tty->print_cr(__VA_ARGS__); \ 65 } 66 67 #ifndef PRODUCT 68 #define non_product_log_if_needed(...) log_if_needed(__VA_ARGS__) 69 #else 70 #define non_product_log_if_needed(...) 71 #endif 72 73 static bool is_vector_mask(ciKlass* klass) { 74 return klass->is_subclass_of(ciEnv::current()->vector_VectorMask_klass()); 75 } 76 77 static bool is_vector_shuffle(ciKlass* klass) { 78 return klass->is_subclass_of(ciEnv::current()->vector_VectorShuffle_klass()); 79 } 80 81 bool LibraryCallKit::arch_supports_vector_rotate(int opc, int num_elem, BasicType elem_bt, 82 VectorMaskUseType mask_use_type, bool has_scalar_args) { 83 bool is_supported = true; 84 85 // has_scalar_args flag is true only for non-constant scalar shift count, 86 // since in this case shift needs to be broadcasted. 87 if (!Matcher::match_rule_supported_vector(opc, num_elem, elem_bt) || 88 (has_scalar_args && !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed))) { 89 is_supported = false; 90 } 91 92 if (is_supported) { 93 // Check if mask unboxing is supported, this is a two step process which first loads the contents 94 // of boolean array into vector followed by either lane expansion to match the lane size of masked 95 // vector operation or populate the predicate register. 96 if ((mask_use_type & VecMaskUseLoad) != 0) { 97 if (!Matcher::match_rule_supported_vector(Op_VectorLoadMask, num_elem, elem_bt) || 98 !Matcher::match_rule_supported_vector(Op_LoadVector, num_elem, T_BOOLEAN)) { 99 non_product_log_if_needed(" ** Rejected vector mask loading (%s,%s,%d) because architecture does not support it", 100 NodeClassNames[Op_VectorLoadMask], type2name(elem_bt), num_elem); 101 return false; 102 } 103 } 104 105 if ((mask_use_type & VecMaskUsePred) != 0) { 106 if (!Matcher::has_predicated_vectors() || 107 !Matcher::match_rule_supported_vector_masked(opc, num_elem, elem_bt)) { 108 non_product_log_if_needed("Rejected vector mask predicate using (%s,%s,%d) because architecture does not support it", 109 NodeClassNames[opc], type2name(elem_bt), num_elem); 110 return false; 111 } 112 } 113 } 114 115 int lshiftopc, rshiftopc; 116 switch(elem_bt) { 117 case T_BYTE: 118 lshiftopc = Op_LShiftI; 119 rshiftopc = Op_URShiftB; 120 break; 121 case T_SHORT: 122 lshiftopc = Op_LShiftI; 123 rshiftopc = Op_URShiftS; 124 break; 125 case T_INT: 126 lshiftopc = Op_LShiftI; 127 rshiftopc = Op_URShiftI; 128 break; 129 case T_LONG: 130 lshiftopc = Op_LShiftL; 131 rshiftopc = Op_URShiftL; 132 break; 133 default: fatal("Unexpected type: %s", type2name(elem_bt)); 134 } 135 int lshiftvopc = VectorNode::opcode(lshiftopc, elem_bt); 136 int rshiftvopc = VectorNode::opcode(rshiftopc, elem_bt); 137 if (!is_supported && 138 arch_supports_vector(lshiftvopc, num_elem, elem_bt, VecMaskNotUsed, has_scalar_args) && 139 arch_supports_vector(rshiftvopc, num_elem, elem_bt, VecMaskNotUsed, has_scalar_args) && 140 arch_supports_vector(Op_OrV, num_elem, elem_bt, VecMaskNotUsed)) { 141 is_supported = true; 142 } 143 return is_supported; 144 } 145 146 Node* GraphKit::box_vector(Node* vector, const TypeInstPtr* vbox_type, BasicType elem_bt, int num_elem, bool deoptimize_on_exception) { 147 assert(EnableVectorSupport, ""); 148 149 PreserveReexecuteState preexecs(this); 150 jvms()->set_should_reexecute(true); 151 152 VectorBoxAllocateNode* alloc = new VectorBoxAllocateNode(C, vbox_type); 153 set_edges_for_java_call(alloc, /*must_throw=*/false, /*separate_io_proj=*/true); 154 make_slow_call_ex(alloc, env()->Throwable_klass(), /*separate_io_proj=*/true, deoptimize_on_exception); 155 set_i_o(gvn().transform( new ProjNode(alloc, TypeFunc::I_O) )); 156 set_all_memory(gvn().transform( new ProjNode(alloc, TypeFunc::Memory) )); 157 Node* ret = gvn().transform(new ProjNode(alloc, TypeFunc::Parms)); 158 159 assert(check_vbox(vbox_type), ""); 160 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, is_vector_mask(vbox_type->instance_klass())); 161 VectorBoxNode* vbox = new VectorBoxNode(C, ret, vector, vbox_type, vt); 162 return gvn().transform(vbox); 163 } 164 165 Node* GraphKit::unbox_vector(Node* v, const TypeInstPtr* vbox_type, BasicType elem_bt, int num_elem, bool shuffle_to_vector) { 166 assert(EnableVectorSupport, ""); 167 const TypeInstPtr* vbox_type_v = gvn().type(v)->isa_instptr(); 168 if (vbox_type_v == nullptr || vbox_type->instance_klass() != vbox_type_v->instance_klass()) { 169 return nullptr; // arguments don't agree on vector shapes 170 } 171 if (vbox_type_v->maybe_null()) { 172 return nullptr; // no nulls are allowed 173 } 174 assert(check_vbox(vbox_type), ""); 175 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, is_vector_mask(vbox_type->instance_klass())); 176 Node* unbox = gvn().transform(new VectorUnboxNode(C, vt, v, merged_memory(), shuffle_to_vector)); 177 return unbox; 178 } 179 180 Node* GraphKit::vector_shift_count(Node* cnt, int shift_op, BasicType bt, int num_elem) { 181 assert(bt == T_INT || bt == T_LONG || bt == T_SHORT || bt == T_BYTE, "byte, short, long and int are supported"); 182 juint mask = (type2aelembytes(bt) * BitsPerByte - 1); 183 Node* nmask = gvn().transform(ConNode::make(TypeInt::make(mask))); 184 Node* mcnt = gvn().transform(new AndINode(cnt, nmask)); 185 return gvn().transform(VectorNode::shift_count(shift_op, mcnt, num_elem, bt)); 186 } 187 188 bool LibraryCallKit::arch_supports_vector(int sopc, int num_elem, BasicType type, VectorMaskUseType mask_use_type, bool has_scalar_args) { 189 // Check that the operation is valid. 190 if (sopc <= 0) { 191 non_product_log_if_needed(" ** Rejected intrinsification because no valid vector op could be extracted"); 192 return false; 193 } 194 195 if (VectorNode::is_vector_rotate(sopc)) { 196 if(!arch_supports_vector_rotate(sopc, num_elem, type, mask_use_type, has_scalar_args)) { 197 non_product_log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support variable vector shifts", 198 NodeClassNames[sopc], type2name(type), num_elem); 199 return false; 200 } 201 } else if (VectorNode::is_vector_integral_negate(sopc)) { 202 if (!VectorNode::is_vector_integral_negate_supported(sopc, num_elem, type, false)) { 203 non_product_log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support integral vector negate", 204 NodeClassNames[sopc], type2name(type), num_elem); 205 return false; 206 } 207 } else { 208 // Check that architecture supports this op-size-type combination. 209 if (!Matcher::match_rule_supported_vector(sopc, num_elem, type)) { 210 non_product_log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support it", 211 NodeClassNames[sopc], type2name(type), num_elem); 212 return false; 213 } else { 214 assert(Matcher::match_rule_supported(sopc), "must be supported"); 215 } 216 } 217 218 if (num_elem == 1) { 219 if (mask_use_type != VecMaskNotUsed) { 220 non_product_log_if_needed(" ** Rejected vector mask op (%s,%s,%d) because architecture does not support it", 221 NodeClassNames[sopc], type2name(type), num_elem); 222 return false; 223 } 224 225 if (sopc != 0) { 226 if (sopc != Op_LoadVector && sopc != Op_StoreVector) { 227 non_product_log_if_needed(" ** Not a svml call or load/store vector op (%s,%s,%d)", 228 NodeClassNames[sopc], type2name(type), num_elem); 229 return false; 230 } 231 } 232 } 233 234 if (!has_scalar_args && VectorNode::is_vector_shift(sopc) && 235 Matcher::supports_vector_variable_shifts() == false) { 236 log_if_needed(" ** Rejected vector op (%s,%s,%d) because architecture does not support variable vector shifts", 237 NodeClassNames[sopc], type2name(type), num_elem); 238 return false; 239 } 240 241 // Check if mask unboxing is supported, this is a two step process which first loads the contents 242 // of boolean array into vector followed by either lane expansion to match the lane size of masked 243 // vector operation or populate the predicate register. 244 if ((mask_use_type & VecMaskUseLoad) != 0) { 245 if (!Matcher::match_rule_supported_vector(Op_VectorLoadMask, num_elem, type) || 246 !Matcher::match_rule_supported_vector(Op_LoadVector, num_elem, T_BOOLEAN)) { 247 non_product_log_if_needed(" ** Rejected vector mask loading (%s,%s,%d) because architecture does not support it", 248 NodeClassNames[Op_VectorLoadMask], type2name(type), num_elem); 249 return false; 250 } 251 } 252 253 // Check if mask boxing is supported, this is a two step process which first stores the contents 254 // of mask vector / predicate register into a boolean vector followed by vector store operation to 255 // transfer the contents to underlined storage of mask boxes which is a boolean array. 256 if ((mask_use_type & VecMaskUseStore) != 0) { 257 if (!Matcher::match_rule_supported_vector(Op_VectorStoreMask, num_elem, type) || 258 !Matcher::match_rule_supported_vector(Op_StoreVector, num_elem, T_BOOLEAN)) { 259 non_product_log_if_needed("Rejected vector mask storing (%s,%s,%d) because architecture does not support it", 260 NodeClassNames[Op_VectorStoreMask], type2name(type), num_elem); 261 return false; 262 } 263 } 264 265 if ((mask_use_type & VecMaskUsePred) != 0) { 266 bool is_supported = false; 267 if (Matcher::has_predicated_vectors()) { 268 if (VectorNode::is_vector_integral_negate(sopc)) { 269 is_supported = VectorNode::is_vector_integral_negate_supported(sopc, num_elem, type, true); 270 } else { 271 is_supported = Matcher::match_rule_supported_vector_masked(sopc, num_elem, type); 272 } 273 } 274 is_supported |= Matcher::supports_vector_predicate_op_emulation(sopc, num_elem, type); 275 276 if (!is_supported) { 277 non_product_log_if_needed("Rejected vector mask predicate using (%s,%s,%d) because architecture does not support it", 278 NodeClassNames[sopc], type2name(type), num_elem); 279 return false; 280 } 281 } 282 283 return true; 284 } 285 286 static bool is_klass_initialized(const TypeInstPtr* vec_klass) { 287 if (vec_klass->const_oop() == nullptr) { 288 return false; // uninitialized or some kind of unsafe access 289 } 290 assert(vec_klass->const_oop()->as_instance()->java_lang_Class_klass() != nullptr, "klass instance expected"); 291 ciInstanceKlass* klass = vec_klass->const_oop()->as_instance()->java_lang_Class_klass()->as_instance_klass(); 292 return klass->is_initialized(); 293 } 294 295 // public static 296 // <V extends Vector<E>, 297 // M extends VectorMask<E>, 298 // E> 299 // V unaryOp(int oprId, Class<? extends V> vmClass, Class<? extends M> maskClass, Class<E> elementType, 300 // int length, V v, M m, 301 // UnaryOperation<V, M> defaultImpl) 302 // 303 // public static 304 // <V, 305 // M extends VectorMask<E>, 306 // E> 307 // V binaryOp(int oprId, Class<? extends V> vmClass, Class<? extends M> maskClass, Class<E> elementType, 308 // int length, V v1, V v2, M m, 309 // BinaryOperation<V, M> defaultImpl) 310 // 311 // public static 312 // <V extends Vector<E>, 313 // M extends VectorMask<E>, 314 // E> 315 // V ternaryOp(int oprId, Class<? extends V> vmClass, Class<? extends M> maskClass, Class<E> elementType, 316 // int length, V v1, V v2, V v3, M m, 317 // TernaryOperation<V, M> defaultImpl) 318 // 319 bool LibraryCallKit::inline_vector_nary_operation(int n) { 320 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 321 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 322 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 323 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 324 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 325 326 if (opr == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 327 !opr->is_con() || vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 328 log_if_needed(" ** missing constant: opr=%s vclass=%s etype=%s vlen=%s", 329 NodeClassNames[argument(0)->Opcode()], 330 NodeClassNames[argument(1)->Opcode()], 331 NodeClassNames[argument(3)->Opcode()], 332 NodeClassNames[argument(4)->Opcode()]); 333 return false; // not enough info for intrinsification 334 } 335 336 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 337 if (!elem_type->is_primitive_type()) { 338 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 339 return false; // should be primitive type 340 } 341 if (!is_klass_initialized(vector_klass)) { 342 log_if_needed(" ** klass argument not initialized"); 343 return false; 344 } 345 346 // "argument(n + 5)" should be the mask object. We assume it is "null" when no mask 347 // is used to control this operation. 348 const Type* vmask_type = gvn().type(argument(n + 5)); 349 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 350 if (is_masked_op) { 351 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 352 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(2)->Opcode()]); 353 return false; // not enough info for intrinsification 354 } 355 356 if (!is_klass_initialized(mask_klass)) { 357 log_if_needed(" ** mask klass argument not initialized"); 358 return false; 359 } 360 361 if (vmask_type->maybe_null()) { 362 log_if_needed(" ** null mask values are not allowed for masked op"); 363 return false; 364 } 365 } 366 367 BasicType elem_bt = elem_type->basic_type(); 368 int num_elem = vlen->get_con(); 369 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 370 int sopc = VectorNode::opcode(opc, elem_bt); 371 if ((opc != Op_CallLeafVector) && (sopc == 0)) { 372 log_if_needed(" ** operation not supported: opc=%s bt=%s", NodeClassNames[opc], type2name(elem_bt)); 373 return false; // operation not supported 374 } 375 if (num_elem == 1) { 376 if (opc != Op_CallLeafVector || elem_bt != T_DOUBLE) { 377 log_if_needed(" ** not a svml call: arity=%d opc=%d vlen=%d etype=%s", 378 n, opc, num_elem, type2name(elem_bt)); 379 return false; 380 } 381 } 382 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 383 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 384 385 if (is_vector_mask(vbox_klass)) { 386 assert(!is_masked_op, "mask operations do not need mask to control"); 387 } 388 389 if (opc == Op_CallLeafVector) { 390 if (!UseVectorStubs) { 391 log_if_needed(" ** vector stubs support is disabled"); 392 return false; 393 } 394 if (!Matcher::supports_vector_calling_convention()) { 395 log_if_needed(" ** no vector calling conventions supported"); 396 return false; 397 } 398 if (!Matcher::vector_size_supported(elem_bt, num_elem)) { 399 log_if_needed(" ** vector size (vlen=%d, etype=%s) is not supported", 400 num_elem, type2name(elem_bt)); 401 return false; 402 } 403 } 404 405 // When using mask, mask use type needs to be VecMaskUseLoad. 406 VectorMaskUseType mask_use_type = is_vector_mask(vbox_klass) ? VecMaskUseAll 407 : is_masked_op ? VecMaskUseLoad : VecMaskNotUsed; 408 if ((sopc != 0) && !arch_supports_vector(sopc, num_elem, elem_bt, mask_use_type)) { 409 log_if_needed(" ** not supported: arity=%d opc=%d vlen=%d etype=%s ismask=%d is_masked_op=%d", 410 n, sopc, num_elem, type2name(elem_bt), 411 is_vector_mask(vbox_klass) ? 1 : 0, is_masked_op ? 1 : 0); 412 return false; // not supported 413 } 414 415 // Return true if current platform has implemented the masked operation with predicate feature. 416 bool use_predicate = is_masked_op && sopc != 0 && arch_supports_vector(sopc, num_elem, elem_bt, VecMaskUsePred); 417 if (is_masked_op && !use_predicate && !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)) { 418 log_if_needed(" ** not supported: arity=%d opc=%d vlen=%d etype=%s ismask=0 is_masked_op=1", 419 n, sopc, num_elem, type2name(elem_bt)); 420 return false; 421 } 422 423 Node* opd1 = nullptr; Node* opd2 = nullptr; Node* opd3 = nullptr; 424 switch (n) { 425 case 3: { 426 opd3 = unbox_vector(argument(7), vbox_type, elem_bt, num_elem); 427 if (opd3 == nullptr) { 428 log_if_needed(" ** unbox failed v3=%s", 429 NodeClassNames[argument(7)->Opcode()]); 430 return false; 431 } 432 // fall-through 433 } 434 case 2: { 435 opd2 = unbox_vector(argument(6), vbox_type, elem_bt, num_elem); 436 if (opd2 == nullptr) { 437 log_if_needed(" ** unbox failed v2=%s", 438 NodeClassNames[argument(6)->Opcode()]); 439 return false; 440 } 441 // fall-through 442 } 443 case 1: { 444 opd1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 445 if (opd1 == nullptr) { 446 log_if_needed(" ** unbox failed v1=%s", 447 NodeClassNames[argument(5)->Opcode()]); 448 return false; 449 } 450 break; 451 } 452 default: fatal("unsupported arity: %d", n); 453 } 454 455 Node* mask = nullptr; 456 if (is_masked_op) { 457 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 458 assert(is_vector_mask(mbox_klass), "argument(2) should be a mask class"); 459 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 460 mask = unbox_vector(argument(n + 5), mbox_type, elem_bt, num_elem); 461 if (mask == nullptr) { 462 log_if_needed(" ** unbox failed mask=%s", 463 NodeClassNames[argument(n + 5)->Opcode()]); 464 return false; 465 } 466 } 467 468 Node* operation = nullptr; 469 if (opc == Op_CallLeafVector) { 470 assert(UseVectorStubs, "sanity"); 471 operation = gen_call_to_svml(opr->get_con(), elem_bt, num_elem, opd1, opd2); 472 if (operation == nullptr) { 473 log_if_needed(" ** svml call failed for %s_%s_%d", 474 (elem_bt == T_FLOAT)?"float":"double", 475 VectorSupport::svmlname[opr->get_con() - VectorSupport::VECTOR_OP_SVML_START], 476 num_elem * type2aelembytes(elem_bt)); 477 return false; 478 } 479 } else { 480 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, is_vector_mask(vbox_klass)); 481 switch (n) { 482 case 1: 483 case 2: { 484 operation = VectorNode::make(sopc, opd1, opd2, vt, is_vector_mask(vbox_klass), VectorNode::is_shift_opcode(opc)); 485 break; 486 } 487 case 3: { 488 operation = VectorNode::make(sopc, opd1, opd2, opd3, vt); 489 break; 490 } 491 default: fatal("unsupported arity: %d", n); 492 } 493 } 494 495 if (is_masked_op && mask != nullptr) { 496 if (use_predicate) { 497 operation->add_req(mask); 498 operation->add_flag(Node::Flag_is_predicated_vector); 499 } else { 500 operation->add_flag(Node::Flag_is_predicated_using_blend); 501 operation = gvn().transform(operation); 502 operation = new VectorBlendNode(opd1, operation, mask); 503 } 504 } 505 operation = gvn().transform(operation); 506 507 // Wrap it up in VectorBox to keep object type information. 508 Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem); 509 set_result(vbox); 510 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 511 return true; 512 } 513 514 // <Sh extends VectorShuffle<E>, E> 515 // Sh ShuffleIota(Class<?> E, Class<?> shuffleClass, Vector.Species<E> s, int length, 516 // int start, int step, int wrap, ShuffleIotaOperation<Sh, E> defaultImpl) 517 bool LibraryCallKit::inline_vector_shuffle_iota() { 518 const TypeInstPtr* shuffle_klass = gvn().type(argument(1))->isa_instptr(); 519 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 520 const TypeInt* start_val = gvn().type(argument(4))->isa_int(); 521 const TypeInt* step_val = gvn().type(argument(5))->isa_int(); 522 const TypeInt* wrap = gvn().type(argument(6))->isa_int(); 523 524 if (shuffle_klass == nullptr || shuffle_klass->const_oop() == nullptr || 525 vlen == nullptr || !vlen->is_con() || start_val == nullptr || step_val == nullptr || 526 wrap == nullptr || !wrap->is_con()) { 527 return false; // not enough info for intrinsification 528 } 529 530 if (!is_klass_initialized(shuffle_klass)) { 531 log_if_needed(" ** klass argument not initialized"); 532 return false; 533 } 534 535 int do_wrap = wrap->get_con(); 536 int num_elem = vlen->get_con(); 537 BasicType elem_bt = T_BYTE; 538 539 bool effective_indices_in_range = false; 540 if (start_val->is_con() && step_val->is_con()) { 541 int effective_min_index = start_val->get_con(); 542 int effective_max_index = start_val->get_con() + step_val->get_con() * (num_elem - 1); 543 effective_indices_in_range = effective_max_index >= effective_min_index && effective_min_index >= -128 && effective_max_index <= 127; 544 } 545 546 if (!do_wrap && !effective_indices_in_range) { 547 // Disable instrinsification for unwrapped shuffle iota if start/step 548 // values are non-constant OR if intermediate result overflows byte value range. 549 return false; 550 } 551 552 if (!arch_supports_vector(Op_AddVB, num_elem, elem_bt, VecMaskNotUsed) || 553 !arch_supports_vector(Op_AndV, num_elem, elem_bt, VecMaskNotUsed) || 554 !arch_supports_vector(Op_VectorLoadConst, num_elem, elem_bt, VecMaskNotUsed) || 555 !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed)) { 556 return false; 557 } 558 559 if (!do_wrap && 560 (!arch_supports_vector(Op_SubVB, num_elem, elem_bt, VecMaskNotUsed) || 561 !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskNotUsed) || 562 !arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskNotUsed))) { 563 return false; 564 } 565 566 bool step_multiply = !step_val->is_con() || !is_power_of_2(step_val->get_con()); 567 if ((step_multiply && !arch_supports_vector(Op_MulVB, num_elem, elem_bt, VecMaskNotUsed)) || 568 (!step_multiply && !arch_supports_vector(Op_LShiftVB, num_elem, elem_bt, VecMaskNotUsed))) { 569 return false; 570 } 571 572 const Type * type_bt = Type::get_const_basic_type(elem_bt); 573 const TypeVect * vt = TypeVect::make(type_bt, num_elem); 574 575 Node* res = gvn().transform(new VectorLoadConstNode(gvn().makecon(TypeInt::ZERO), vt)); 576 577 Node* start = argument(4); 578 Node* step = argument(5); 579 580 if (step_multiply) { 581 Node* bcast_step = gvn().transform(VectorNode::scalar2vector(step, num_elem, type_bt)); 582 res = gvn().transform(VectorNode::make(Op_MulVB, res, bcast_step, vt)); 583 } else if (step_val->get_con() > 1) { 584 Node* cnt = gvn().makecon(TypeInt::make(log2i_exact(step_val->get_con()))); 585 Node* shift_cnt = vector_shift_count(cnt, Op_LShiftI, elem_bt, num_elem); 586 res = gvn().transform(VectorNode::make(Op_LShiftVB, res, shift_cnt, vt)); 587 } 588 589 if (!start_val->is_con() || start_val->get_con() != 0) { 590 Node* bcast_start = gvn().transform(VectorNode::scalar2vector(start, num_elem, type_bt)); 591 res = gvn().transform(VectorNode::make(Op_AddVB, res, bcast_start, vt)); 592 } 593 594 Node * mod_val = gvn().makecon(TypeInt::make(num_elem-1)); 595 Node * bcast_mod = gvn().transform(VectorNode::scalar2vector(mod_val, num_elem, type_bt)); 596 597 if (do_wrap) { 598 // Wrap the indices greater than lane count. 599 res = gvn().transform(VectorNode::make(Op_AndV, res, bcast_mod, vt)); 600 } else { 601 ConINode* pred_node = (ConINode*)gvn().makecon(TypeInt::make(BoolTest::ugt)); 602 Node * lane_cnt = gvn().makecon(TypeInt::make(num_elem)); 603 Node * bcast_lane_cnt = gvn().transform(VectorNode::scalar2vector(lane_cnt, num_elem, type_bt)); 604 const TypeVect* vmask_type = TypeVect::makemask(elem_bt, num_elem); 605 Node* mask = gvn().transform(new VectorMaskCmpNode(BoolTest::ugt, bcast_lane_cnt, res, pred_node, vmask_type)); 606 607 // Make the indices greater than lane count as -ve values to match the java side implementation. 608 res = gvn().transform(VectorNode::make(Op_AndV, res, bcast_mod, vt)); 609 Node * biased_val = gvn().transform(VectorNode::make(Op_SubVB, res, bcast_lane_cnt, vt)); 610 res = gvn().transform(new VectorBlendNode(biased_val, res, mask)); 611 } 612 613 ciKlass* sbox_klass = shuffle_klass->const_oop()->as_instance()->java_lang_Class_klass(); 614 const TypeInstPtr* shuffle_box_type = TypeInstPtr::make_exact(TypePtr::NotNull, sbox_klass); 615 616 // Wrap it up in VectorBox to keep object type information. 617 res = box_vector(res, shuffle_box_type, elem_bt, num_elem); 618 set_result(res); 619 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 620 return true; 621 } 622 623 // <E, M> 624 // long maskReductionCoerced(int oper, Class<? extends M> maskClass, Class<?> elemClass, 625 // int length, M m, VectorMaskOp<M> defaultImpl) 626 bool LibraryCallKit::inline_vector_mask_operation() { 627 const TypeInt* oper = gvn().type(argument(0))->isa_int(); 628 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 629 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 630 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 631 Node* mask = argument(4); 632 633 if (mask_klass == nullptr || elem_klass == nullptr || mask->is_top() || vlen == nullptr) { 634 return false; // dead code 635 } 636 637 if (!is_klass_initialized(mask_klass)) { 638 log_if_needed(" ** klass argument not initialized"); 639 return false; 640 } 641 642 int num_elem = vlen->get_con(); 643 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 644 BasicType elem_bt = elem_type->basic_type(); 645 646 int mopc = VectorSupport::vop2ideal(oper->get_con(), elem_bt); 647 if (!arch_supports_vector(mopc, num_elem, elem_bt, VecMaskUseLoad)) { 648 log_if_needed(" ** not supported: arity=1 op=cast#%d/3 vlen2=%d etype2=%s", 649 mopc, num_elem, type2name(elem_bt)); 650 return false; // not supported 651 } 652 653 const Type* elem_ty = Type::get_const_basic_type(elem_bt); 654 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 655 const TypeInstPtr* mask_box_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 656 Node* mask_vec = unbox_vector(mask, mask_box_type, elem_bt, num_elem, true); 657 if (mask_vec == nullptr) { 658 log_if_needed(" ** unbox failed mask=%s", 659 NodeClassNames[argument(4)->Opcode()]); 660 return false; 661 } 662 663 if (mask_vec->bottom_type()->isa_vectmask() == nullptr) { 664 mask_vec = gvn().transform(VectorStoreMaskNode::make(gvn(), mask_vec, elem_bt, num_elem)); 665 } 666 const Type* maskoper_ty = mopc == Op_VectorMaskToLong ? (const Type*)TypeLong::LONG : (const Type*)TypeInt::INT; 667 Node* maskoper = gvn().transform(VectorMaskOpNode::make(mask_vec, maskoper_ty, mopc)); 668 if (mopc != Op_VectorMaskToLong) { 669 maskoper = ConvI2L(maskoper); 670 } 671 set_result(maskoper); 672 673 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 674 return true; 675 } 676 677 // public static 678 // <V, 679 // Sh extends VectorShuffle<E>, 680 // E> 681 // V shuffleToVector(Class<? extends Vector<E>> vclass, Class<E> elementType, 682 // Class<? extends Sh> shuffleClass, Sh s, int length, 683 // ShuffleToVectorOperation<V, Sh, E> defaultImpl) 684 bool LibraryCallKit::inline_vector_shuffle_to_vector() { 685 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 686 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 687 const TypeInstPtr* shuffle_klass = gvn().type(argument(2))->isa_instptr(); 688 Node* shuffle = argument(3); 689 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 690 691 if (vector_klass == nullptr || elem_klass == nullptr || shuffle_klass == nullptr || shuffle->is_top() || vlen == nullptr) { 692 return false; // dead code 693 } 694 if (!vlen->is_con() || vector_klass->const_oop() == nullptr || shuffle_klass->const_oop() == nullptr) { 695 return false; // not enough info for intrinsification 696 } 697 if (!is_klass_initialized(shuffle_klass) || !is_klass_initialized(vector_klass) ) { 698 log_if_needed(" ** klass argument not initialized"); 699 return false; 700 } 701 702 int num_elem = vlen->get_con(); 703 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 704 BasicType elem_bt = elem_type->basic_type(); 705 706 if (num_elem < 4) { 707 return false; 708 } 709 710 int cast_vopc = VectorCastNode::opcode(-1, T_BYTE); // from shuffle of type T_BYTE 711 // Make sure that cast is implemented to particular type/size combination. 712 if (!arch_supports_vector(cast_vopc, num_elem, elem_bt, VecMaskNotUsed)) { 713 log_if_needed(" ** not supported: arity=1 op=cast#%d/3 vlen2=%d etype2=%s", 714 cast_vopc, num_elem, type2name(elem_bt)); 715 return false; 716 } 717 718 ciKlass* sbox_klass = shuffle_klass->const_oop()->as_instance()->java_lang_Class_klass(); 719 const TypeInstPtr* shuffle_box_type = TypeInstPtr::make_exact(TypePtr::NotNull, sbox_klass); 720 721 // Unbox shuffle with true flag to indicate its load shuffle to vector 722 // shuffle is a byte array 723 Node* shuffle_vec = unbox_vector(shuffle, shuffle_box_type, T_BYTE, num_elem, true); 724 725 // cast byte to target element type 726 shuffle_vec = gvn().transform(VectorCastNode::make(cast_vopc, shuffle_vec, elem_bt, num_elem)); 727 728 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 729 const TypeInstPtr* vec_box_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 730 731 // Box vector 732 Node* res = box_vector(shuffle_vec, vec_box_type, elem_bt, num_elem); 733 set_result(res); 734 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 735 return true; 736 } 737 738 // public static 739 // <M, 740 // S extends VectorSpecies<E>, 741 // E> 742 // M fromBitsCoerced(Class<? extends M> vmClass, Class<E> elementType, int length, 743 // long bits, int mode, S s, 744 // BroadcastOperation<M, E, S> defaultImpl) 745 bool LibraryCallKit::inline_vector_frombits_coerced() { 746 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 747 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 748 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 749 const TypeLong* bits_type = gvn().type(argument(3))->isa_long(); 750 // Mode argument determines the mode of operation it can take following values:- 751 // MODE_BROADCAST for vector Vector.broadcast and VectorMask.maskAll operations. 752 // MODE_BITS_COERCED_LONG_TO_MASK for VectorMask.fromLong operation. 753 const TypeInt* mode = gvn().type(argument(5))->isa_int(); 754 755 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || mode == nullptr || 756 bits_type == nullptr || vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || 757 !vlen->is_con() || !mode->is_con()) { 758 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s bitwise=%s", 759 NodeClassNames[argument(0)->Opcode()], 760 NodeClassNames[argument(1)->Opcode()], 761 NodeClassNames[argument(2)->Opcode()], 762 NodeClassNames[argument(5)->Opcode()]); 763 return false; // not enough info for intrinsification 764 } 765 766 if (!is_klass_initialized(vector_klass)) { 767 log_if_needed(" ** klass argument not initialized"); 768 return false; 769 } 770 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 771 if (!elem_type->is_primitive_type()) { 772 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 773 return false; // should be primitive type 774 } 775 BasicType elem_bt = elem_type->basic_type(); 776 int num_elem = vlen->get_con(); 777 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 778 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 779 780 bool is_mask = is_vector_mask(vbox_klass); 781 int bcast_mode = mode->get_con(); 782 VectorMaskUseType checkFlags = (VectorMaskUseType)(is_mask ? VecMaskUseAll : VecMaskNotUsed); 783 int opc = bcast_mode == VectorSupport::MODE_BITS_COERCED_LONG_TO_MASK ? Op_VectorLongToMask : Op_Replicate; 784 785 if (!arch_supports_vector(opc, num_elem, elem_bt, checkFlags, true /*has_scalar_args*/)) { 786 log_if_needed(" ** not supported: arity=0 op=broadcast vlen=%d etype=%s ismask=%d bcast_mode=%d", 787 num_elem, type2name(elem_bt), 788 is_mask ? 1 : 0, 789 bcast_mode); 790 return false; // not supported 791 } 792 793 Node* broadcast = nullptr; 794 Node* bits = argument(3); 795 Node* elem = bits; 796 797 if (opc == Op_VectorLongToMask) { 798 const TypeVect* vt = TypeVect::makemask(elem_bt, num_elem); 799 if (vt->isa_vectmask()) { 800 broadcast = gvn().transform(new VectorLongToMaskNode(elem, vt)); 801 } else { 802 const TypeVect* mvt = TypeVect::make(T_BOOLEAN, num_elem); 803 broadcast = gvn().transform(new VectorLongToMaskNode(elem, mvt)); 804 broadcast = gvn().transform(new VectorLoadMaskNode(broadcast, vt)); 805 } 806 } else { 807 switch (elem_bt) { 808 case T_BOOLEAN: // fall-through 809 case T_BYTE: // fall-through 810 case T_SHORT: // fall-through 811 case T_CHAR: // fall-through 812 case T_INT: { 813 elem = gvn().transform(new ConvL2INode(bits)); 814 break; 815 } 816 case T_DOUBLE: { 817 elem = gvn().transform(new MoveL2DNode(bits)); 818 break; 819 } 820 case T_FLOAT: { 821 bits = gvn().transform(new ConvL2INode(bits)); 822 elem = gvn().transform(new MoveI2FNode(bits)); 823 break; 824 } 825 case T_LONG: { 826 // no conversion needed 827 break; 828 } 829 default: fatal("%s", type2name(elem_bt)); 830 } 831 broadcast = VectorNode::scalar2vector(elem, num_elem, Type::get_const_basic_type(elem_bt), is_mask); 832 broadcast = gvn().transform(broadcast); 833 } 834 835 Node* box = box_vector(broadcast, vbox_type, elem_bt, num_elem); 836 set_result(box); 837 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 838 return true; 839 } 840 841 static bool elem_consistent_with_arr(BasicType elem_bt, const TypeAryPtr* arr_type, bool mismatched_ms) { 842 assert(arr_type != nullptr, "unexpected"); 843 BasicType arr_elem_bt = arr_type->elem()->array_element_basic_type(); 844 if (elem_bt == arr_elem_bt) { 845 return true; 846 } else if (elem_bt == T_SHORT && arr_elem_bt == T_CHAR) { 847 // Load/store of short vector from/to char[] is supported 848 return true; 849 } else if (elem_bt == T_BYTE && arr_elem_bt == T_BOOLEAN) { 850 // Load/store of byte vector from/to boolean[] is supported 851 return true; 852 } else { 853 return mismatched_ms; 854 } 855 } 856 857 // public static 858 // <C, 859 // VM extends VectorPayload, 860 // E, 861 // S extends VectorSpecies<E>> 862 // VM load(Class<? extends VM> vmClass, Class<E> eClass, 863 // int length, 864 // Object base, long offset, // Unsafe addressing 865 // boolean fromSegment, 866 // C container, long index, S s, // Arguments for default implementation 867 // LoadOperation<C, VM, S> defaultImpl) { 868 // public static 869 // <C, 870 // V extends VectorPayload> 871 // void store(Class<?> vClass, Class<?> eClass, 872 // int length, 873 // Object base, long offset, // Unsafe addressing 874 // boolean fromSegment, 875 // V v, C container, long index, // Arguments for default implementation 876 // StoreVectorOperation<C, V> defaultImpl) { 877 bool LibraryCallKit::inline_vector_mem_operation(bool is_store) { 878 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 879 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 880 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 881 const TypeInt* from_ms = gvn().type(argument(6))->isa_int(); 882 883 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || !from_ms->is_con() || 884 vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 885 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s from_ms=%s", 886 NodeClassNames[argument(0)->Opcode()], 887 NodeClassNames[argument(1)->Opcode()], 888 NodeClassNames[argument(2)->Opcode()], 889 NodeClassNames[argument(6)->Opcode()]); 890 return false; // not enough info for intrinsification 891 } 892 if (!is_klass_initialized(vector_klass)) { 893 log_if_needed(" ** klass argument not initialized"); 894 return false; 895 } 896 897 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 898 if (!elem_type->is_primitive_type()) { 899 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 900 return false; // should be primitive type 901 } 902 BasicType elem_bt = elem_type->basic_type(); 903 int num_elem = vlen->get_con(); 904 905 // TODO When mask usage is supported, VecMaskNotUsed needs to be VecMaskUseLoad. 906 if (!arch_supports_vector(is_store ? Op_StoreVector : Op_LoadVector, num_elem, elem_bt, VecMaskNotUsed)) { 907 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s ismask=no", 908 is_store, is_store ? "store" : "load", 909 num_elem, type2name(elem_bt)); 910 return false; // not supported 911 } 912 913 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 914 bool is_mask = is_vector_mask(vbox_klass); 915 916 Node* base = argument(3); 917 Node* offset = ConvL2X(argument(4)); 918 919 // Save state and restore on bailout 920 uint old_sp = sp(); 921 SafePointNode* old_map = clone_map(); 922 923 Node* addr = make_unsafe_address(base, offset, (is_mask ? T_BOOLEAN : elem_bt), true); 924 925 // The memory barrier checks are based on ones for unsafe access. 926 // This is not 1-1 implementation. 927 const Type *const base_type = gvn().type(base); 928 929 const TypePtr *addr_type = gvn().type(addr)->isa_ptr(); 930 const TypeAryPtr* arr_type = addr_type->isa_aryptr(); 931 932 const bool in_native = TypePtr::NULL_PTR == base_type; // base always null 933 const bool in_heap = !TypePtr::NULL_PTR->higher_equal(base_type); // base never null 934 935 const bool is_mixed_access = !in_heap && !in_native; 936 937 const bool is_mismatched_access = in_heap && (addr_type->isa_aryptr() == nullptr); 938 939 const bool needs_cpu_membar = is_mixed_access || is_mismatched_access; 940 941 // For non-masked mismatched memory segment vector read/write accesses, intrinsification can continue 942 // with unknown backing storage type and compiler can skip inserting explicit reinterpretation IR after 943 // loading from or before storing to backing storage which is mandatory for semantic correctness of 944 // big-endian memory layout. 945 bool mismatched_ms = LITTLE_ENDIAN_ONLY(false) 946 BIG_ENDIAN_ONLY(from_ms->get_con() && !is_mask && arr_type != nullptr && 947 arr_type->elem()->array_element_basic_type() != elem_bt); 948 BasicType mem_elem_bt = mismatched_ms ? arr_type->elem()->array_element_basic_type() : elem_bt; 949 if (!is_java_primitive(mem_elem_bt)) { 950 log_if_needed(" ** non-primitive array element type"); 951 return false; 952 } 953 int mem_num_elem = mismatched_ms ? (num_elem * type2aelembytes(elem_bt)) / type2aelembytes(mem_elem_bt) : num_elem; 954 if (arr_type != nullptr && !is_mask && !elem_consistent_with_arr(elem_bt, arr_type, mismatched_ms)) { 955 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s atype=%s ismask=no", 956 is_store, is_store ? "store" : "load", 957 num_elem, type2name(elem_bt), type2name(arr_type->elem()->array_element_basic_type())); 958 set_map(old_map); 959 set_sp(old_sp); 960 return false; 961 } 962 963 // In case of mismatched memory segment accesses, we need to double check that the source type memory operations are supported by backend. 964 if (mismatched_ms) { 965 if (is_store) { 966 if (!arch_supports_vector(Op_StoreVector, num_elem, elem_bt, VecMaskNotUsed) 967 || !arch_supports_vector(Op_VectorReinterpret, mem_num_elem, mem_elem_bt, VecMaskNotUsed)) { 968 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d*8 etype=%s/8 ismask=no", 969 is_store, "store", 970 num_elem, type2name(elem_bt)); 971 set_map(old_map); 972 set_sp(old_sp); 973 return false; // not supported 974 } 975 } else { 976 if (!arch_supports_vector(Op_LoadVector, mem_num_elem, mem_elem_bt, VecMaskNotUsed) 977 || !arch_supports_vector(Op_VectorReinterpret, num_elem, elem_bt, VecMaskNotUsed)) { 978 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d*8 etype=%s/8 ismask=no", 979 is_store, "load", 980 mem_num_elem, type2name(mem_elem_bt)); 981 set_map(old_map); 982 set_sp(old_sp); 983 return false; // not supported 984 } 985 } 986 } 987 if (is_mask) { 988 if (!is_store) { 989 if (!arch_supports_vector(Op_LoadVector, num_elem, elem_bt, VecMaskUseLoad)) { 990 set_map(old_map); 991 set_sp(old_sp); 992 return false; // not supported 993 } 994 } else { 995 if (!arch_supports_vector(Op_StoreVector, num_elem, elem_bt, VecMaskUseStore)) { 996 set_map(old_map); 997 set_sp(old_sp); 998 return false; // not supported 999 } 1000 } 1001 } 1002 1003 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1004 1005 if (needs_cpu_membar) { 1006 insert_mem_bar(Op_MemBarCPUOrder); 1007 } 1008 1009 if (is_store) { 1010 Node* val = unbox_vector(argument(7), vbox_type, elem_bt, num_elem); 1011 if (val == nullptr) { 1012 set_map(old_map); 1013 set_sp(old_sp); 1014 return false; // operand unboxing failed 1015 } 1016 set_all_memory(reset_memory()); 1017 1018 // In case the store needs to happen to byte array, reinterpret the incoming vector to byte vector. 1019 int store_num_elem = num_elem; 1020 if (mismatched_ms) { 1021 store_num_elem = mem_num_elem; 1022 const TypeVect* to_vect_type = TypeVect::make(mem_elem_bt, store_num_elem); 1023 val = gvn().transform(new VectorReinterpretNode(val, val->bottom_type()->is_vect(), to_vect_type)); 1024 } 1025 if (is_mask) { 1026 val = gvn().transform(VectorStoreMaskNode::make(gvn(), val, elem_bt, num_elem)); 1027 } 1028 Node* vstore = gvn().transform(StoreVectorNode::make(0, control(), memory(addr), addr, addr_type, val, store_num_elem)); 1029 set_memory(vstore, addr_type); 1030 } else { 1031 // When using byte array, we need to load as byte then reinterpret the value. Otherwise, do a simple vector load. 1032 Node* vload = nullptr; 1033 if (mismatched_ms) { 1034 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, mem_num_elem, mem_elem_bt)); 1035 const TypeVect* to_vect_type = TypeVect::make(elem_bt, num_elem); 1036 vload = gvn().transform(new VectorReinterpretNode(vload, vload->bottom_type()->is_vect(), to_vect_type)); 1037 } else { 1038 // Special handle for masks 1039 if (is_mask) { 1040 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, num_elem, T_BOOLEAN)); 1041 vload = gvn().transform(new VectorLoadMaskNode(vload, TypeVect::makemask(elem_bt, num_elem))); 1042 } else { 1043 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, num_elem, elem_bt)); 1044 } 1045 } 1046 Node* box = box_vector(vload, vbox_type, elem_bt, num_elem); 1047 set_result(box); 1048 } 1049 1050 destruct_map_clone(old_map); 1051 1052 if (needs_cpu_membar) { 1053 insert_mem_bar(Op_MemBarCPUOrder); 1054 } 1055 1056 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1057 return true; 1058 } 1059 1060 // public static 1061 // <C, 1062 // V extends Vector<?>, 1063 // E, 1064 // S extends VectorSpecies<E>, 1065 // M extends VectorMask<E>> 1066 // V loadMasked(Class<? extends V> vClass, Class<M> mClass, Class<E> eClass, 1067 // int length, Object base, long offset, // Unsafe addressing 1068 // boolean fromSegment, 1069 // M m, int offsetInRange, 1070 // C container, long index, S s, // Arguments for default implementation 1071 // LoadVectorMaskedOperation<C, V, S, M> defaultImpl) { 1072 // public static 1073 // <C, 1074 // V extends Vector<E>, 1075 // M extends VectorMask<E>, 1076 // E> 1077 // void storeMasked(Class<? extends V> vClass, Class<M> mClass, Class<E> eClass, 1078 // int length, 1079 // Object base, long offset, // Unsafe addressing 1080 // boolean fromSegment, 1081 // V v, M m, C container, long index, // Arguments for default implementation 1082 // StoreVectorMaskedOperation<C, V, M> defaultImpl) { 1083 1084 bool LibraryCallKit::inline_vector_mem_masked_operation(bool is_store) { 1085 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1086 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 1087 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1088 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1089 const TypeInt* from_ms = gvn().type(argument(7))->isa_int(); 1090 1091 if (vector_klass == nullptr || mask_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 1092 vector_klass->const_oop() == nullptr || mask_klass->const_oop() == nullptr || from_ms == nullptr || 1093 elem_klass->const_oop() == nullptr || !vlen->is_con() || !from_ms->is_con()) { 1094 log_if_needed(" ** missing constant: vclass=%s mclass=%s etype=%s vlen=%s from_ms=%s", 1095 NodeClassNames[argument(0)->Opcode()], 1096 NodeClassNames[argument(1)->Opcode()], 1097 NodeClassNames[argument(2)->Opcode()], 1098 NodeClassNames[argument(3)->Opcode()], 1099 NodeClassNames[argument(7)->Opcode()]); 1100 return false; // not enough info for intrinsification 1101 } 1102 if (!is_klass_initialized(vector_klass)) { 1103 log_if_needed(" ** klass argument not initialized"); 1104 return false; 1105 } 1106 1107 if (!is_klass_initialized(mask_klass)) { 1108 log_if_needed(" ** mask klass argument not initialized"); 1109 return false; 1110 } 1111 1112 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1113 if (!elem_type->is_primitive_type()) { 1114 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1115 return false; // should be primitive type 1116 } 1117 1118 BasicType elem_bt = elem_type->basic_type(); 1119 int num_elem = vlen->get_con(); 1120 1121 Node* base = argument(4); 1122 Node* offset = ConvL2X(argument(5)); 1123 1124 // Save state and restore on bailout 1125 uint old_sp = sp(); 1126 SafePointNode* old_map = clone_map(); 1127 1128 Node* addr = make_unsafe_address(base, offset, elem_bt, true); 1129 const TypePtr *addr_type = gvn().type(addr)->isa_ptr(); 1130 const TypeAryPtr* arr_type = addr_type->isa_aryptr(); 1131 1132 bool mismatched_ms = from_ms->get_con() && arr_type != nullptr && arr_type->elem()->array_element_basic_type() != elem_bt; 1133 BIG_ENDIAN_ONLY(if (mismatched_ms) return false;) 1134 // If there is no consistency between array and vector element types, it must be special byte array case 1135 if (arr_type != nullptr && !elem_consistent_with_arr(elem_bt, arr_type, mismatched_ms)) { 1136 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s atype=%s", 1137 is_store, is_store ? "storeMasked" : "loadMasked", 1138 num_elem, type2name(elem_bt), type2name(arr_type->elem()->array_element_basic_type())); 1139 set_map(old_map); 1140 set_sp(old_sp); 1141 return false; 1142 } 1143 1144 int mem_num_elem = mismatched_ms ? num_elem * type2aelembytes(elem_bt) : num_elem; 1145 BasicType mem_elem_bt = mismatched_ms ? T_BYTE : elem_bt; 1146 bool supports_predicate = arch_supports_vector(is_store ? Op_StoreVectorMasked : Op_LoadVectorMasked, 1147 mem_num_elem, mem_elem_bt, VecMaskUseLoad); 1148 1149 // If current arch does not support the predicated operations, we have to bail 1150 // out when current case uses the predicate feature. 1151 if (!supports_predicate) { 1152 bool needs_predicate = false; 1153 if (is_store) { 1154 // Masked vector store always uses the predicated store. 1155 needs_predicate = true; 1156 } else { 1157 // Masked vector load with IOOBE always uses the predicated load. 1158 const TypeInt* offset_in_range = gvn().type(argument(9))->isa_int(); 1159 if (!offset_in_range->is_con()) { 1160 log_if_needed(" ** missing constant: offsetInRange=%s", 1161 NodeClassNames[argument(8)->Opcode()]); 1162 set_map(old_map); 1163 set_sp(old_sp); 1164 return false; 1165 } 1166 needs_predicate = (offset_in_range->get_con() == 0); 1167 } 1168 1169 if (needs_predicate) { 1170 log_if_needed(" ** not supported: op=%s vlen=%d etype=%s mismatched_ms=%d", 1171 is_store ? "storeMasked" : "loadMasked", 1172 num_elem, type2name(elem_bt), mismatched_ms ? 1 : 0); 1173 set_map(old_map); 1174 set_sp(old_sp); 1175 return false; 1176 } 1177 } 1178 1179 // This only happens for masked vector load. If predicate is not supported, then check whether 1180 // the normal vector load and blend operations are supported by backend. 1181 if (!supports_predicate && (!arch_supports_vector(Op_LoadVector, mem_num_elem, mem_elem_bt, VecMaskNotUsed) || 1182 !arch_supports_vector(Op_VectorBlend, mem_num_elem, mem_elem_bt, VecMaskUseLoad))) { 1183 log_if_needed(" ** not supported: op=loadMasked vlen=%d etype=%s mismatched_ms=%d", 1184 num_elem, type2name(elem_bt), mismatched_ms ? 1 : 0); 1185 set_map(old_map); 1186 set_sp(old_sp); 1187 return false; 1188 } 1189 1190 // Since we are using byte array, we need to double check that the vector reinterpret operation 1191 // with byte type is supported by backend. 1192 if (mismatched_ms) { 1193 if (!arch_supports_vector(Op_VectorReinterpret, mem_num_elem, T_BYTE, VecMaskNotUsed)) { 1194 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s mismatched_ms=1", 1195 is_store, is_store ? "storeMasked" : "loadMasked", 1196 num_elem, type2name(elem_bt)); 1197 set_map(old_map); 1198 set_sp(old_sp); 1199 return false; 1200 } 1201 } 1202 1203 // Since it needs to unbox the mask, we need to double check that the related load operations 1204 // for mask are supported by backend. 1205 if (!arch_supports_vector(Op_LoadVector, num_elem, elem_bt, VecMaskUseLoad)) { 1206 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s", 1207 is_store, is_store ? "storeMasked" : "loadMasked", 1208 num_elem, type2name(elem_bt)); 1209 set_map(old_map); 1210 set_sp(old_sp); 1211 return false; 1212 } 1213 1214 // Can base be null? Otherwise, always on-heap access. 1215 bool can_access_non_heap = TypePtr::NULL_PTR->higher_equal(gvn().type(base)); 1216 if (can_access_non_heap) { 1217 insert_mem_bar(Op_MemBarCPUOrder); 1218 } 1219 1220 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1221 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1222 assert(!is_vector_mask(vbox_klass) && is_vector_mask(mbox_klass), "Invalid class type"); 1223 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1224 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1225 1226 Node* mask = unbox_vector(is_store ? argument(9) : argument(8), mbox_type, elem_bt, num_elem); 1227 if (mask == nullptr) { 1228 log_if_needed(" ** unbox failed mask=%s", 1229 is_store ? NodeClassNames[argument(9)->Opcode()] 1230 : NodeClassNames[argument(8)->Opcode()]); 1231 set_map(old_map); 1232 set_sp(old_sp); 1233 return false; 1234 } 1235 1236 if (is_store) { 1237 Node* val = unbox_vector(argument(8), vbox_type, elem_bt, num_elem); 1238 if (val == nullptr) { 1239 log_if_needed(" ** unbox failed vector=%s", 1240 NodeClassNames[argument(8)->Opcode()]); 1241 set_map(old_map); 1242 set_sp(old_sp); 1243 return false; // operand unboxing failed 1244 } 1245 set_all_memory(reset_memory()); 1246 1247 if (mismatched_ms) { 1248 // Reinterpret the incoming vector to byte vector. 1249 const TypeVect* to_vect_type = TypeVect::make(mem_elem_bt, mem_num_elem); 1250 val = gvn().transform(new VectorReinterpretNode(val, val->bottom_type()->is_vect(), to_vect_type)); 1251 // Reinterpret the vector mask to byte type. 1252 const TypeVect* from_mask_type = TypeVect::makemask(elem_bt, num_elem); 1253 const TypeVect* to_mask_type = TypeVect::makemask(mem_elem_bt, mem_num_elem); 1254 mask = gvn().transform(new VectorReinterpretNode(mask, from_mask_type, to_mask_type)); 1255 } 1256 Node* vstore = gvn().transform(new StoreVectorMaskedNode(control(), memory(addr), addr, val, addr_type, mask)); 1257 set_memory(vstore, addr_type); 1258 } else { 1259 Node* vload = nullptr; 1260 1261 if (mismatched_ms) { 1262 // Reinterpret the vector mask to byte type. 1263 const TypeVect* from_mask_type = TypeVect::makemask(elem_bt, num_elem); 1264 const TypeVect* to_mask_type = TypeVect::makemask(mem_elem_bt, mem_num_elem); 1265 mask = gvn().transform(new VectorReinterpretNode(mask, from_mask_type, to_mask_type)); 1266 } 1267 1268 if (supports_predicate) { 1269 // Generate masked load vector node if predicate feature is supported. 1270 const TypeVect* vt = TypeVect::make(mem_elem_bt, mem_num_elem); 1271 vload = gvn().transform(new LoadVectorMaskedNode(control(), memory(addr), addr, addr_type, vt, mask)); 1272 } else { 1273 // Use the vector blend to implement the masked load vector. The biased elements are zeros. 1274 Node* zero = gvn().transform(gvn().zerocon(mem_elem_bt)); 1275 zero = gvn().transform(VectorNode::scalar2vector(zero, mem_num_elem, Type::get_const_basic_type(mem_elem_bt))); 1276 vload = gvn().transform(LoadVectorNode::make(0, control(), memory(addr), addr, addr_type, mem_num_elem, mem_elem_bt)); 1277 vload = gvn().transform(new VectorBlendNode(zero, vload, mask)); 1278 } 1279 1280 if (mismatched_ms) { 1281 const TypeVect* to_vect_type = TypeVect::make(elem_bt, num_elem); 1282 vload = gvn().transform(new VectorReinterpretNode(vload, vload->bottom_type()->is_vect(), to_vect_type)); 1283 } 1284 1285 Node* box = box_vector(vload, vbox_type, elem_bt, num_elem); 1286 set_result(box); 1287 } 1288 1289 destruct_map_clone(old_map); 1290 1291 if (can_access_non_heap) { 1292 insert_mem_bar(Op_MemBarCPUOrder); 1293 } 1294 1295 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1296 return true; 1297 } 1298 1299 // <C, 1300 // V extends Vector<?>, 1301 // W extends Vector<Integer>, 1302 // S extends VectorSpecies<E>, 1303 // M extends VectorMask<E>, 1304 // E> 1305 // V loadWithMap(Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, int length, 1306 // Class<? extends Vector<Integer>> vectorIndexClass, 1307 // Object base, long offset, // Unsafe addressing 1308 // W index_vector, M m, 1309 // C container, int index, int[] indexMap, int indexM, S s, // Arguments for default implementation 1310 // LoadVectorOperationWithMap<C, V, E, S, M> defaultImpl) 1311 // 1312 // <C, 1313 // V extends Vector<E>, 1314 // W extends Vector<Integer>, 1315 // M extends VectorMask<E>, 1316 // E> 1317 // void storeWithMap(Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, 1318 // int length, Class<? extends Vector<Integer>> vectorIndexClass, Object base, long offset, // Unsafe addressing 1319 // W index_vector, V v, M m, 1320 // C container, int index, int[] indexMap, int indexM, // Arguments for default implementation 1321 // StoreVectorOperationWithMap<C, V, M, E> defaultImpl) 1322 // 1323 bool LibraryCallKit::inline_vector_gather_scatter(bool is_scatter) { 1324 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1325 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 1326 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1327 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1328 const TypeInstPtr* vector_idx_klass = gvn().type(argument(4))->isa_instptr(); 1329 1330 if (vector_klass == nullptr || elem_klass == nullptr || vector_idx_klass == nullptr || vlen == nullptr || 1331 vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || vector_idx_klass->const_oop() == nullptr || !vlen->is_con()) { 1332 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s viclass=%s", 1333 NodeClassNames[argument(0)->Opcode()], 1334 NodeClassNames[argument(2)->Opcode()], 1335 NodeClassNames[argument(3)->Opcode()], 1336 NodeClassNames[argument(4)->Opcode()]); 1337 return false; // not enough info for intrinsification 1338 } 1339 1340 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(vector_idx_klass)) { 1341 log_if_needed(" ** klass argument not initialized"); 1342 return false; 1343 } 1344 1345 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1346 if (!elem_type->is_primitive_type()) { 1347 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1348 return false; // should be primitive type 1349 } 1350 1351 BasicType elem_bt = elem_type->basic_type(); 1352 int num_elem = vlen->get_con(); 1353 1354 const Type* vmask_type = gvn().type(is_scatter ? argument(10) : argument(9)); 1355 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 1356 if (is_masked_op) { 1357 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 1358 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(1)->Opcode()]); 1359 return false; // not enough info for intrinsification 1360 } 1361 1362 if (!is_klass_initialized(mask_klass)) { 1363 log_if_needed(" ** mask klass argument not initialized"); 1364 return false; 1365 } 1366 1367 if (vmask_type->maybe_null()) { 1368 log_if_needed(" ** null mask values are not allowed for masked op"); 1369 return false; 1370 } 1371 1372 // Check whether the predicated gather/scatter node is supported by architecture. 1373 VectorMaskUseType mask = (VectorMaskUseType) (VecMaskUseLoad | VecMaskUsePred); 1374 if (!arch_supports_vector(is_scatter ? Op_StoreVectorScatterMasked : Op_LoadVectorGatherMasked, num_elem, elem_bt, mask)) { 1375 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s is_masked_op=1", 1376 is_scatter, is_scatter ? "scatterMasked" : "gatherMasked", 1377 num_elem, type2name(elem_bt)); 1378 return false; // not supported 1379 } 1380 } else { 1381 // Check whether the normal gather/scatter node is supported for non-masked operation. 1382 if (!arch_supports_vector(is_scatter ? Op_StoreVectorScatter : Op_LoadVectorGather, num_elem, elem_bt, VecMaskNotUsed)) { 1383 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s is_masked_op=0", 1384 is_scatter, is_scatter ? "scatter" : "gather", 1385 num_elem, type2name(elem_bt)); 1386 return false; // not supported 1387 } 1388 } 1389 1390 // Check that the vector holding indices is supported by architecture 1391 // For sub-word gathers expander receive index array. 1392 if (!is_subword_type(elem_bt) && !arch_supports_vector(Op_LoadVector, num_elem, T_INT, VecMaskNotUsed)) { 1393 log_if_needed(" ** not supported: arity=%d op=%s/loadindex vlen=%d etype=int is_masked_op=%d", 1394 is_scatter, is_scatter ? "scatter" : "gather", 1395 num_elem, is_masked_op ? 1 : 0); 1396 return false; // not supported 1397 } 1398 1399 Node* base = argument(5); 1400 Node* offset = ConvL2X(argument(6)); 1401 1402 // Save state and restore on bailout 1403 uint old_sp = sp(); 1404 SafePointNode* old_map = clone_map(); 1405 1406 Node* addr = make_unsafe_address(base, offset, elem_bt, true); 1407 1408 const TypePtr *addr_type = gvn().type(addr)->isa_ptr(); 1409 const TypeAryPtr* arr_type = addr_type->isa_aryptr(); 1410 1411 // The array must be consistent with vector type 1412 if (arr_type == nullptr || (arr_type != nullptr && !elem_consistent_with_arr(elem_bt, arr_type, false))) { 1413 log_if_needed(" ** not supported: arity=%d op=%s vlen=%d etype=%s atype=%s ismask=no", 1414 is_scatter, is_scatter ? "scatter" : "gather", 1415 num_elem, type2name(elem_bt), type2name(arr_type->elem()->array_element_basic_type())); 1416 set_map(old_map); 1417 set_sp(old_sp); 1418 return false; 1419 } 1420 1421 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1422 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1423 ciKlass* vbox_idx_klass = vector_idx_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1424 if (vbox_idx_klass == nullptr) { 1425 set_map(old_map); 1426 set_sp(old_sp); 1427 return false; 1428 } 1429 1430 Node* index_vect = nullptr; 1431 const TypeInstPtr* vbox_idx_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_idx_klass); 1432 if (!is_subword_type(elem_bt)) { 1433 index_vect = unbox_vector(argument(8), vbox_idx_type, T_INT, num_elem); 1434 if (index_vect == nullptr) { 1435 set_map(old_map); 1436 set_sp(old_sp); 1437 return false; 1438 } 1439 } 1440 1441 Node* mask = nullptr; 1442 if (is_masked_op) { 1443 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1444 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1445 mask = unbox_vector(is_scatter ? argument(10) : argument(9), mbox_type, elem_bt, num_elem); 1446 if (mask == nullptr) { 1447 log_if_needed(" ** unbox failed mask=%s", 1448 is_scatter ? NodeClassNames[argument(10)->Opcode()] 1449 : NodeClassNames[argument(9)->Opcode()]); 1450 set_map(old_map); 1451 set_sp(old_sp); 1452 return false; 1453 } 1454 } 1455 1456 const TypeVect* vector_type = TypeVect::make(elem_bt, num_elem); 1457 if (is_scatter) { 1458 Node* val = unbox_vector(argument(9), vbox_type, elem_bt, num_elem); 1459 if (val == nullptr) { 1460 set_map(old_map); 1461 set_sp(old_sp); 1462 return false; // operand unboxing failed 1463 } 1464 set_all_memory(reset_memory()); 1465 1466 Node* vstore = nullptr; 1467 if (mask != nullptr) { 1468 vstore = gvn().transform(new StoreVectorScatterMaskedNode(control(), memory(addr), addr, addr_type, val, index_vect, mask)); 1469 } else { 1470 vstore = gvn().transform(new StoreVectorScatterNode(control(), memory(addr), addr, addr_type, val, index_vect)); 1471 } 1472 set_memory(vstore, addr_type); 1473 } else { 1474 Node* vload = nullptr; 1475 Node* index = argument(11); 1476 Node* indexMap = argument(12); 1477 Node* indexM = argument(13); 1478 if (mask != nullptr) { 1479 if (is_subword_type(elem_bt)) { 1480 Node* index_arr_base = array_element_address(indexMap, indexM, T_INT); 1481 vload = gvn().transform(new LoadVectorGatherMaskedNode(control(), memory(addr), addr, addr_type, vector_type, index_arr_base, mask, index)); 1482 } else { 1483 vload = gvn().transform(new LoadVectorGatherMaskedNode(control(), memory(addr), addr, addr_type, vector_type, index_vect, mask)); 1484 } 1485 } else { 1486 if (is_subword_type(elem_bt)) { 1487 Node* index_arr_base = array_element_address(indexMap, indexM, T_INT); 1488 vload = gvn().transform(new LoadVectorGatherNode(control(), memory(addr), addr, addr_type, vector_type, index_arr_base, index)); 1489 } else { 1490 vload = gvn().transform(new LoadVectorGatherNode(control(), memory(addr), addr, addr_type, vector_type, index_vect)); 1491 } 1492 } 1493 Node* box = box_vector(vload, vbox_type, elem_bt, num_elem); 1494 set_result(box); 1495 } 1496 1497 destruct_map_clone(old_map); 1498 1499 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1500 return true; 1501 } 1502 1503 // public static 1504 // <V extends Vector<E>, 1505 // M extends VectorMask<E>, 1506 // E> 1507 // long reductionCoerced(int oprId, Class<? extends V> vectorClass, Class<? extends M> maskClass, 1508 // Class<E> elementType, int length, V v, M m, 1509 // ReductionOperation<V, M> defaultImpl) 1510 bool LibraryCallKit::inline_vector_reduction() { 1511 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 1512 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 1513 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 1514 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 1515 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 1516 1517 if (opr == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 1518 !opr->is_con() || vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 1519 log_if_needed(" ** missing constant: opr=%s vclass=%s etype=%s vlen=%s", 1520 NodeClassNames[argument(0)->Opcode()], 1521 NodeClassNames[argument(1)->Opcode()], 1522 NodeClassNames[argument(3)->Opcode()], 1523 NodeClassNames[argument(4)->Opcode()]); 1524 return false; // not enough info for intrinsification 1525 } 1526 if (!is_klass_initialized(vector_klass)) { 1527 log_if_needed(" ** klass argument not initialized"); 1528 return false; 1529 } 1530 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1531 if (!elem_type->is_primitive_type()) { 1532 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1533 return false; // should be primitive type 1534 } 1535 1536 const Type* vmask_type = gvn().type(argument(6)); 1537 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 1538 if (is_masked_op) { 1539 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 1540 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(2)->Opcode()]); 1541 return false; // not enough info for intrinsification 1542 } 1543 1544 if (!is_klass_initialized(mask_klass)) { 1545 log_if_needed(" ** mask klass argument not initialized"); 1546 return false; 1547 } 1548 1549 if (vmask_type->maybe_null()) { 1550 log_if_needed(" ** null mask values are not allowed for masked op"); 1551 return false; 1552 } 1553 } 1554 1555 BasicType elem_bt = elem_type->basic_type(); 1556 int num_elem = vlen->get_con(); 1557 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 1558 int sopc = ReductionNode::opcode(opc, elem_bt); 1559 1560 // When using mask, mask use type needs to be VecMaskUseLoad. 1561 if (!arch_supports_vector(sopc, num_elem, elem_bt, is_masked_op ? VecMaskUseLoad : VecMaskNotUsed)) { 1562 log_if_needed(" ** not supported: arity=1 op=%d/reduce vlen=%d etype=%s is_masked_op=%d", 1563 sopc, num_elem, type2name(elem_bt), is_masked_op ? 1 : 0); 1564 return false; 1565 } 1566 1567 // Return true if current platform has implemented the masked operation with predicate feature. 1568 bool use_predicate = is_masked_op && arch_supports_vector(sopc, num_elem, elem_bt, VecMaskUsePred); 1569 if (is_masked_op && !use_predicate && !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)) { 1570 log_if_needed(" ** not supported: arity=1 op=%d/reduce vlen=%d etype=%s is_masked_op=1", 1571 sopc, num_elem, type2name(elem_bt)); 1572 return false; 1573 } 1574 1575 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1576 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1577 1578 Node* opd = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1579 if (opd == nullptr) { 1580 return false; // operand unboxing failed 1581 } 1582 1583 Node* mask = nullptr; 1584 if (is_masked_op) { 1585 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1586 assert(is_vector_mask(mbox_klass), "argument(2) should be a mask class"); 1587 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1588 mask = unbox_vector(argument(6), mbox_type, elem_bt, num_elem); 1589 if (mask == nullptr) { 1590 log_if_needed(" ** unbox failed mask=%s", 1591 NodeClassNames[argument(6)->Opcode()]); 1592 return false; 1593 } 1594 } 1595 1596 Node* init = ReductionNode::make_identity_con_scalar(gvn(), opc, elem_bt); 1597 Node* value = nullptr; 1598 if (mask == nullptr) { 1599 assert(!is_masked_op, "Masked op needs the mask value never null"); 1600 value = ReductionNode::make(opc, nullptr, init, opd, elem_bt); 1601 } else { 1602 if (use_predicate) { 1603 value = ReductionNode::make(opc, nullptr, init, opd, elem_bt); 1604 value->add_req(mask); 1605 value->add_flag(Node::Flag_is_predicated_vector); 1606 } else { 1607 Node* reduce_identity = gvn().transform(VectorNode::scalar2vector(init, num_elem, Type::get_const_basic_type(elem_bt))); 1608 value = gvn().transform(new VectorBlendNode(reduce_identity, opd, mask)); 1609 value = ReductionNode::make(opc, nullptr, init, value, elem_bt); 1610 } 1611 } 1612 value = gvn().transform(value); 1613 1614 Node* bits = nullptr; 1615 switch (elem_bt) { 1616 case T_BYTE: 1617 case T_SHORT: 1618 case T_INT: { 1619 bits = gvn().transform(new ConvI2LNode(value)); 1620 break; 1621 } 1622 case T_FLOAT: { 1623 value = gvn().transform(new MoveF2INode(value)); 1624 bits = gvn().transform(new ConvI2LNode(value)); 1625 break; 1626 } 1627 case T_DOUBLE: { 1628 bits = gvn().transform(new MoveD2LNode(value)); 1629 break; 1630 } 1631 case T_LONG: { 1632 bits = value; // no conversion needed 1633 break; 1634 } 1635 default: fatal("%s", type2name(elem_bt)); 1636 } 1637 set_result(bits); 1638 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1639 return true; 1640 } 1641 1642 // public static <V> boolean test(int cond, Class<?> vectorClass, Class<?> elementType, int vlen, 1643 // V v1, V v2, 1644 // BiFunction<V, V, Boolean> defaultImpl) 1645 // 1646 bool LibraryCallKit::inline_vector_test() { 1647 const TypeInt* cond = gvn().type(argument(0))->isa_int(); 1648 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 1649 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1650 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1651 1652 if (cond == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 1653 !cond->is_con() || vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 1654 log_if_needed(" ** missing constant: cond=%s vclass=%s etype=%s vlen=%s", 1655 NodeClassNames[argument(0)->Opcode()], 1656 NodeClassNames[argument(1)->Opcode()], 1657 NodeClassNames[argument(2)->Opcode()], 1658 NodeClassNames[argument(3)->Opcode()]); 1659 return false; // not enough info for intrinsification 1660 } 1661 if (!is_klass_initialized(vector_klass)) { 1662 log_if_needed(" ** klass argument not initialized"); 1663 return false; 1664 } 1665 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1666 if (!elem_type->is_primitive_type()) { 1667 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1668 return false; // should be primitive type 1669 } 1670 BasicType elem_bt = elem_type->basic_type(); 1671 int num_elem = vlen->get_con(); 1672 BoolTest::mask booltest = (BoolTest::mask)cond->get_con(); 1673 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1674 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1675 1676 if (!arch_supports_vector(Op_VectorTest, num_elem, elem_bt, is_vector_mask(vbox_klass) ? VecMaskUseLoad : VecMaskNotUsed)) { 1677 log_if_needed(" ** not supported: arity=2 op=test/%d vlen=%d etype=%s ismask=%d", 1678 cond->get_con(), num_elem, type2name(elem_bt), 1679 is_vector_mask(vbox_klass)); 1680 return false; 1681 } 1682 1683 Node* opd1 = unbox_vector(argument(4), vbox_type, elem_bt, num_elem); 1684 Node* opd2; 1685 if (Matcher::vectortest_needs_second_argument(booltest == BoolTest::overflow, 1686 opd1->bottom_type()->isa_vectmask())) { 1687 opd2 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1688 } else { 1689 opd2 = opd1; 1690 } 1691 if (opd1 == nullptr || opd2 == nullptr) { 1692 return false; // operand unboxing failed 1693 } 1694 1695 Node* cmp = gvn().transform(new VectorTestNode(opd1, opd2, booltest)); 1696 BoolTest::mask test = Matcher::vectortest_mask(booltest == BoolTest::overflow, 1697 opd1->bottom_type()->isa_vectmask(), num_elem); 1698 Node* bol = gvn().transform(new BoolNode(cmp, test)); 1699 Node* res = gvn().transform(new CMoveINode(bol, gvn().intcon(0), gvn().intcon(1), TypeInt::BOOL)); 1700 1701 set_result(res); 1702 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1703 return true; 1704 } 1705 1706 // public static 1707 // <V extends Vector<E>, 1708 // M extends VectorMask<E>, 1709 // E> 1710 // V blend(Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, int vlen, 1711 // V v1, V v2, M m, 1712 // VectorBlendOp<V, M, E> defaultImpl) 1713 bool LibraryCallKit::inline_vector_blend() { 1714 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1715 const TypeInstPtr* mask_klass = gvn().type(argument(1))->isa_instptr(); 1716 const TypeInstPtr* elem_klass = gvn().type(argument(2))->isa_instptr(); 1717 const TypeInt* vlen = gvn().type(argument(3))->isa_int(); 1718 1719 if (mask_klass == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 1720 return false; // dead code 1721 } 1722 if (mask_klass->const_oop() == nullptr || vector_klass->const_oop() == nullptr || 1723 elem_klass->const_oop() == nullptr || !vlen->is_con()) { 1724 log_if_needed(" ** missing constant: vclass=%s mclass=%s etype=%s vlen=%s", 1725 NodeClassNames[argument(0)->Opcode()], 1726 NodeClassNames[argument(1)->Opcode()], 1727 NodeClassNames[argument(2)->Opcode()], 1728 NodeClassNames[argument(3)->Opcode()]); 1729 return false; // not enough info for intrinsification 1730 } 1731 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(mask_klass)) { 1732 log_if_needed(" ** klass argument not initialized"); 1733 return false; 1734 } 1735 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1736 if (!elem_type->is_primitive_type()) { 1737 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1738 return false; // should be primitive type 1739 } 1740 BasicType elem_bt = elem_type->basic_type(); 1741 BasicType mask_bt = elem_bt; 1742 int num_elem = vlen->get_con(); 1743 1744 if (!arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad)) { 1745 log_if_needed(" ** not supported: arity=2 op=blend vlen=%d etype=%s ismask=useload", 1746 num_elem, type2name(elem_bt)); 1747 return false; // not supported 1748 } 1749 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1750 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1751 1752 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1753 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1754 1755 Node* v1 = unbox_vector(argument(4), vbox_type, elem_bt, num_elem); 1756 Node* v2 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1757 Node* mask = unbox_vector(argument(6), mbox_type, mask_bt, num_elem); 1758 1759 if (v1 == nullptr || v2 == nullptr || mask == nullptr) { 1760 return false; // operand unboxing failed 1761 } 1762 1763 Node* blend = gvn().transform(new VectorBlendNode(v1, v2, mask)); 1764 1765 Node* box = box_vector(blend, vbox_type, elem_bt, num_elem); 1766 set_result(box); 1767 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1768 return true; 1769 } 1770 1771 // public static 1772 // <V extends Vector<E>, 1773 // M extends VectorMask<E>, 1774 // E> 1775 // M compare(int cond, Class<? extends V> vectorClass, Class<M> maskClass, Class<E> elementType, int vlen, 1776 // V v1, V v2, M m, 1777 // VectorCompareOp<V,M> defaultImpl) 1778 bool LibraryCallKit::inline_vector_compare() { 1779 const TypeInt* cond = gvn().type(argument(0))->isa_int(); 1780 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 1781 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 1782 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 1783 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 1784 1785 if (cond == nullptr || vector_klass == nullptr || mask_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 1786 return false; // dead code 1787 } 1788 if (!cond->is_con() || vector_klass->const_oop() == nullptr || mask_klass->const_oop() == nullptr || 1789 elem_klass->const_oop() == nullptr || !vlen->is_con()) { 1790 log_if_needed(" ** missing constant: cond=%s vclass=%s mclass=%s etype=%s vlen=%s", 1791 NodeClassNames[argument(0)->Opcode()], 1792 NodeClassNames[argument(1)->Opcode()], 1793 NodeClassNames[argument(2)->Opcode()], 1794 NodeClassNames[argument(3)->Opcode()], 1795 NodeClassNames[argument(4)->Opcode()]); 1796 return false; // not enough info for intrinsification 1797 } 1798 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(mask_klass)) { 1799 log_if_needed(" ** klass argument not initialized"); 1800 return false; 1801 } 1802 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1803 if (!elem_type->is_primitive_type()) { 1804 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1805 return false; // should be primitive type 1806 } 1807 1808 int num_elem = vlen->get_con(); 1809 BasicType elem_bt = elem_type->basic_type(); 1810 BasicType mask_bt = elem_bt; 1811 1812 if ((cond->get_con() & BoolTest::unsigned_compare) != 0) { 1813 if (!Matcher::supports_vector_comparison_unsigned(num_elem, elem_bt)) { 1814 log_if_needed(" ** not supported: unsigned comparison op=comp/%d vlen=%d etype=%s ismask=usestore", 1815 cond->get_con() & (BoolTest::unsigned_compare - 1), num_elem, type2name(elem_bt)); 1816 return false; 1817 } 1818 } 1819 1820 if (!arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskUseStore)) { 1821 log_if_needed(" ** not supported: arity=2 op=comp/%d vlen=%d etype=%s ismask=usestore", 1822 cond->get_con(), num_elem, type2name(elem_bt)); 1823 return false; 1824 } 1825 1826 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1827 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1828 1829 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1830 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1831 1832 Node* v1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1833 Node* v2 = unbox_vector(argument(6), vbox_type, elem_bt, num_elem); 1834 1835 bool is_masked_op = argument(7)->bottom_type() != TypePtr::NULL_PTR; 1836 Node* mask = is_masked_op ? unbox_vector(argument(7), mbox_type, elem_bt, num_elem) : nullptr; 1837 if (is_masked_op && mask == nullptr) { 1838 log_if_needed(" ** not supported: mask = null arity=2 op=comp/%d vlen=%d etype=%s ismask=usestore is_masked_op=1", 1839 cond->get_con(), num_elem, type2name(elem_bt)); 1840 return false; 1841 } 1842 1843 bool use_predicate = is_masked_op && arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskUsePred); 1844 if (is_masked_op && !use_predicate && !arch_supports_vector(Op_AndV, num_elem, elem_bt, VecMaskUseLoad)) { 1845 log_if_needed(" ** not supported: arity=2 op=comp/%d vlen=%d etype=%s ismask=usestore is_masked_op=1", 1846 cond->get_con(), num_elem, type2name(elem_bt)); 1847 return false; 1848 } 1849 1850 if (v1 == nullptr || v2 == nullptr) { 1851 return false; // operand unboxing failed 1852 } 1853 BoolTest::mask pred = (BoolTest::mask)cond->get_con(); 1854 ConINode* pred_node = (ConINode*)gvn().makecon(cond); 1855 1856 const TypeVect* vmask_type = TypeVect::makemask(mask_bt, num_elem); 1857 Node* operation = new VectorMaskCmpNode(pred, v1, v2, pred_node, vmask_type); 1858 1859 if (is_masked_op) { 1860 if (use_predicate) { 1861 operation->add_req(mask); 1862 operation->add_flag(Node::Flag_is_predicated_vector); 1863 } else { 1864 operation = gvn().transform(operation); 1865 operation = VectorNode::make(Op_AndV, operation, mask, vmask_type); 1866 } 1867 } 1868 1869 operation = gvn().transform(operation); 1870 1871 Node* box = box_vector(operation, mbox_type, mask_bt, num_elem); 1872 set_result(box); 1873 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1874 return true; 1875 } 1876 1877 // public static 1878 // <V extends Vector<E>, 1879 // Sh extends VectorShuffle<E>, 1880 // M extends VectorMask<E>, 1881 // E> 1882 // V rearrangeOp(Class<? extends V> vectorClass, Class<Sh> shuffleClass, Class<M> maskClass, Class<E> elementType, int vlen, 1883 // V v1, Sh sh, M m, 1884 // VectorRearrangeOp<V, Sh, M, E> defaultImpl) 1885 bool LibraryCallKit::inline_vector_rearrange() { 1886 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 1887 const TypeInstPtr* shuffle_klass = gvn().type(argument(1))->isa_instptr(); 1888 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 1889 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 1890 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 1891 1892 if (vector_klass == nullptr || shuffle_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 1893 return false; // dead code 1894 } 1895 if (shuffle_klass->const_oop() == nullptr || 1896 vector_klass->const_oop() == nullptr || 1897 elem_klass->const_oop() == nullptr || 1898 !vlen->is_con()) { 1899 log_if_needed(" ** missing constant: vclass=%s sclass=%s etype=%s vlen=%s", 1900 NodeClassNames[argument(0)->Opcode()], 1901 NodeClassNames[argument(1)->Opcode()], 1902 NodeClassNames[argument(3)->Opcode()], 1903 NodeClassNames[argument(4)->Opcode()]); 1904 return false; // not enough info for intrinsification 1905 } 1906 if (!is_klass_initialized(vector_klass) || 1907 !is_klass_initialized(shuffle_klass)) { 1908 log_if_needed(" ** klass argument not initialized"); 1909 return false; 1910 } 1911 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 1912 if (!elem_type->is_primitive_type()) { 1913 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 1914 return false; // should be primitive type 1915 } 1916 BasicType elem_bt = elem_type->basic_type(); 1917 BasicType shuffle_bt = elem_bt; 1918 int num_elem = vlen->get_con(); 1919 1920 if (!arch_supports_vector(Op_VectorLoadShuffle, num_elem, elem_bt, VecMaskNotUsed)) { 1921 log_if_needed(" ** not supported: arity=0 op=load/shuffle vlen=%d etype=%s ismask=no", 1922 num_elem, type2name(elem_bt)); 1923 return false; // not supported 1924 } 1925 1926 bool is_masked_op = argument(7)->bottom_type() != TypePtr::NULL_PTR; 1927 bool use_predicate = is_masked_op; 1928 if (is_masked_op && 1929 (mask_klass == nullptr || 1930 mask_klass->const_oop() == nullptr || 1931 !is_klass_initialized(mask_klass))) { 1932 log_if_needed(" ** mask_klass argument not initialized"); 1933 } 1934 VectorMaskUseType checkFlags = (VectorMaskUseType)(is_masked_op ? (VecMaskUseLoad | VecMaskUsePred) : VecMaskNotUsed); 1935 if (!arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, checkFlags)) { 1936 use_predicate = false; 1937 if(!is_masked_op || 1938 (!arch_supports_vector(Op_VectorRearrange, num_elem, elem_bt, VecMaskNotUsed) || 1939 !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad) || 1940 !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed))) { 1941 log_if_needed(" ** not supported: arity=2 op=shuffle/rearrange vlen=%d etype=%s ismask=no", 1942 num_elem, type2name(elem_bt)); 1943 return false; // not supported 1944 } 1945 } 1946 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1947 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 1948 1949 ciKlass* shbox_klass = shuffle_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1950 const TypeInstPtr* shbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, shbox_klass); 1951 1952 Node* v1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 1953 Node* shuffle = unbox_vector(argument(6), shbox_type, shuffle_bt, num_elem); 1954 1955 if (v1 == nullptr || shuffle == nullptr) { 1956 return false; // operand unboxing failed 1957 } 1958 1959 Node* mask = nullptr; 1960 if (is_masked_op) { 1961 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 1962 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 1963 mask = unbox_vector(argument(7), mbox_type, elem_bt, num_elem); 1964 if (mask == nullptr) { 1965 log_if_needed(" ** not supported: arity=3 op=shuffle/rearrange vlen=%d etype=%s ismask=useload is_masked_op=1", 1966 num_elem, type2name(elem_bt)); 1967 return false; 1968 } 1969 } 1970 1971 Node* rearrange = new VectorRearrangeNode(v1, shuffle); 1972 if (is_masked_op) { 1973 if (use_predicate) { 1974 rearrange->add_req(mask); 1975 rearrange->add_flag(Node::Flag_is_predicated_vector); 1976 } else { 1977 const TypeVect* vt = v1->bottom_type()->is_vect(); 1978 rearrange = gvn().transform(rearrange); 1979 Node* zero = gvn().makecon(Type::get_zero_type(elem_bt)); 1980 Node* zerovec = gvn().transform(VectorNode::scalar2vector(zero, num_elem, Type::get_const_basic_type(elem_bt))); 1981 rearrange = new VectorBlendNode(zerovec, rearrange, mask); 1982 } 1983 } 1984 rearrange = gvn().transform(rearrange); 1985 1986 Node* box = box_vector(rearrange, vbox_type, elem_bt, num_elem); 1987 set_result(box); 1988 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 1989 return true; 1990 } 1991 1992 static address get_svml_address(int vop, int bits, BasicType bt, char* name_ptr, int name_len) { 1993 address addr = nullptr; 1994 assert(UseVectorStubs, "sanity"); 1995 assert(name_ptr != nullptr, "unexpected"); 1996 assert((vop >= VectorSupport::VECTOR_OP_SVML_START) && (vop <= VectorSupport::VECTOR_OP_SVML_END), "unexpected"); 1997 int op = vop - VectorSupport::VECTOR_OP_SVML_START; 1998 1999 switch(bits) { 2000 case 64: //fallthough 2001 case 128: //fallthough 2002 case 256: //fallthough 2003 case 512: 2004 if (bt == T_FLOAT) { 2005 snprintf(name_ptr, name_len, "vector_%s_float%d", VectorSupport::svmlname[op], bits); 2006 addr = StubRoutines::_vector_f_math[exact_log2(bits/64)][op]; 2007 } else { 2008 assert(bt == T_DOUBLE, "must be FP type only"); 2009 snprintf(name_ptr, name_len, "vector_%s_double%d", VectorSupport::svmlname[op], bits); 2010 addr = StubRoutines::_vector_d_math[exact_log2(bits/64)][op]; 2011 } 2012 break; 2013 default: 2014 snprintf(name_ptr, name_len, "invalid"); 2015 addr = nullptr; 2016 Unimplemented(); 2017 break; 2018 } 2019 2020 return addr; 2021 } 2022 2023 Node* LibraryCallKit::gen_call_to_svml(int vector_api_op_id, BasicType bt, int num_elem, Node* opd1, Node* opd2) { 2024 assert(UseVectorStubs, "sanity"); 2025 assert(vector_api_op_id >= VectorSupport::VECTOR_OP_SVML_START && vector_api_op_id <= VectorSupport::VECTOR_OP_SVML_END, "need valid op id"); 2026 assert(opd1 != nullptr, "must not be null"); 2027 const TypeVect* vt = TypeVect::make(bt, num_elem); 2028 const TypeFunc* call_type = OptoRuntime::Math_Vector_Vector_Type(opd2 != nullptr ? 2 : 1, vt, vt); 2029 char name[100] = ""; 2030 2031 // Get address for svml method. 2032 address addr = get_svml_address(vector_api_op_id, vt->length_in_bytes() * BitsPerByte, bt, name, 100); 2033 2034 if (addr == nullptr) { 2035 return nullptr; 2036 } 2037 2038 assert(name[0] != '\0', "name must not be null"); 2039 Node* operation = make_runtime_call(RC_VECTOR, 2040 call_type, 2041 addr, 2042 name, 2043 TypePtr::BOTTOM, 2044 opd1, 2045 opd2); 2046 return gvn().transform(new ProjNode(gvn().transform(operation), TypeFunc::Parms)); 2047 } 2048 2049 // public static 2050 // <V extends Vector<E>, 2051 // M extends VectorMask<E>, 2052 // E> 2053 // V broadcastInt(int opr, Class<? extends V> vectorClass, Class<? extends M> maskClass, 2054 // Class<E> elementType, int length, 2055 // V v, int n, M m, 2056 // VectorBroadcastIntOp<V, M> defaultImpl) 2057 bool LibraryCallKit::inline_vector_broadcast_int() { 2058 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 2059 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 2060 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 2061 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 2062 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 2063 2064 if (opr == nullptr || vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr) { 2065 return false; // dead code 2066 } 2067 if (!opr->is_con() || vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 2068 log_if_needed(" ** missing constant: opr=%s vclass=%s etype=%s vlen=%s", 2069 NodeClassNames[argument(0)->Opcode()], 2070 NodeClassNames[argument(1)->Opcode()], 2071 NodeClassNames[argument(3)->Opcode()], 2072 NodeClassNames[argument(4)->Opcode()]); 2073 return false; // not enough info for intrinsification 2074 } 2075 if (!is_klass_initialized(vector_klass)) { 2076 log_if_needed(" ** klass argument not initialized"); 2077 return false; 2078 } 2079 2080 const Type* vmask_type = gvn().type(argument(7)); 2081 bool is_masked_op = vmask_type != TypePtr::NULL_PTR; 2082 if (is_masked_op) { 2083 if (mask_klass == nullptr || mask_klass->const_oop() == nullptr) { 2084 log_if_needed(" ** missing constant: maskclass=%s", NodeClassNames[argument(2)->Opcode()]); 2085 return false; // not enough info for intrinsification 2086 } 2087 2088 if (!is_klass_initialized(mask_klass)) { 2089 log_if_needed(" ** mask klass argument not initialized"); 2090 return false; 2091 } 2092 2093 if (vmask_type->maybe_null()) { 2094 log_if_needed(" ** null mask values are not allowed for masked op"); 2095 return false; 2096 } 2097 } 2098 2099 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2100 if (!elem_type->is_primitive_type()) { 2101 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2102 return false; // should be primitive type 2103 } 2104 2105 int num_elem = vlen->get_con(); 2106 BasicType elem_bt = elem_type->basic_type(); 2107 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 2108 2109 bool is_shift = VectorNode::is_shift_opcode(opc); 2110 bool is_rotate = VectorNode::is_rotate_opcode(opc); 2111 2112 if (opc == 0 || (!is_shift && !is_rotate)) { 2113 log_if_needed(" ** operation not supported: op=%d bt=%s", opr->get_con(), type2name(elem_bt)); 2114 return false; // operation not supported 2115 } 2116 2117 int sopc = VectorNode::opcode(opc, elem_bt); 2118 if (sopc == 0) { 2119 log_if_needed(" ** operation not supported: opc=%s bt=%s", NodeClassNames[opc], type2name(elem_bt)); 2120 return false; // operation not supported 2121 } 2122 2123 Node* cnt = argument(6); 2124 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2125 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2126 const TypeInt* cnt_type = cnt->bottom_type()->isa_int(); 2127 2128 // If CPU supports vector constant rotate instructions pass it directly 2129 bool is_const_rotate = is_rotate && cnt_type && cnt_type->is_con() && 2130 Matcher::supports_vector_constant_rotates(cnt_type->get_con()); 2131 bool has_scalar_args = is_rotate ? !is_const_rotate : true; 2132 2133 VectorMaskUseType checkFlags = (VectorMaskUseType)(is_masked_op ? (VecMaskUseLoad | VecMaskUsePred) : VecMaskNotUsed); 2134 bool use_predicate = is_masked_op; 2135 2136 if (!arch_supports_vector(sopc, num_elem, elem_bt, checkFlags, has_scalar_args)) { 2137 use_predicate = false; 2138 if (!is_masked_op || 2139 (!arch_supports_vector(sopc, num_elem, elem_bt, VecMaskNotUsed, has_scalar_args) || 2140 !arch_supports_vector(Op_VectorBlend, num_elem, elem_bt, VecMaskUseLoad))) { 2141 2142 log_if_needed(" ** not supported: arity=0 op=int/%d vlen=%d etype=%s is_masked_op=%d", 2143 sopc, num_elem, type2name(elem_bt), is_masked_op ? 1 : 0); 2144 return false; // not supported 2145 } 2146 } 2147 2148 Node* opd1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 2149 Node* opd2 = nullptr; 2150 if (is_shift) { 2151 opd2 = vector_shift_count(cnt, opc, elem_bt, num_elem); 2152 } else { 2153 assert(is_rotate, "unexpected operation"); 2154 if (!is_const_rotate) { 2155 const Type * type_bt = Type::get_const_basic_type(elem_bt); 2156 cnt = elem_bt == T_LONG ? gvn().transform(new ConvI2LNode(cnt)) : cnt; 2157 opd2 = gvn().transform(VectorNode::scalar2vector(cnt, num_elem, type_bt)); 2158 } else { 2159 // Constant shift value. 2160 opd2 = cnt; 2161 } 2162 } 2163 2164 if (opd1 == nullptr || opd2 == nullptr) { 2165 return false; 2166 } 2167 2168 Node* mask = nullptr; 2169 if (is_masked_op) { 2170 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2171 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 2172 mask = unbox_vector(argument(7), mbox_type, elem_bt, num_elem); 2173 if (mask == nullptr) { 2174 log_if_needed(" ** unbox failed mask=%s", NodeClassNames[argument(7)->Opcode()]); 2175 return false; 2176 } 2177 } 2178 2179 Node* operation = VectorNode::make(opc, opd1, opd2, num_elem, elem_bt); 2180 if (is_masked_op && mask != nullptr) { 2181 if (use_predicate) { 2182 operation->add_req(mask); 2183 operation->add_flag(Node::Flag_is_predicated_vector); 2184 } else { 2185 operation = gvn().transform(operation); 2186 operation = new VectorBlendNode(opd1, operation, mask); 2187 } 2188 } 2189 operation = gvn().transform(operation); 2190 Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem); 2191 set_result(vbox); 2192 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2193 return true; 2194 } 2195 2196 // public static <VOUT extends VectorPayload, 2197 // VIN extends VectorPayload, 2198 // S extends VectorSpecies> 2199 // VOUT convert(int oprId, 2200 // Class<?> fromVectorClass, Class<?> fromElementType, int fromVLen, 2201 // Class<?> toVectorClass, Class<?> toElementType, int toVLen, 2202 // VIN v, S s, 2203 // VectorConvertOp<VOUT, VIN, S> defaultImpl) 2204 // 2205 bool LibraryCallKit::inline_vector_convert() { 2206 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 2207 2208 const TypeInstPtr* vector_klass_from = gvn().type(argument(1))->isa_instptr(); 2209 const TypeInstPtr* elem_klass_from = gvn().type(argument(2))->isa_instptr(); 2210 const TypeInt* vlen_from = gvn().type(argument(3))->isa_int(); 2211 2212 const TypeInstPtr* vector_klass_to = gvn().type(argument(4))->isa_instptr(); 2213 const TypeInstPtr* elem_klass_to = gvn().type(argument(5))->isa_instptr(); 2214 const TypeInt* vlen_to = gvn().type(argument(6))->isa_int(); 2215 2216 if (opr == nullptr || 2217 vector_klass_from == nullptr || elem_klass_from == nullptr || vlen_from == nullptr || 2218 vector_klass_to == nullptr || elem_klass_to == nullptr || vlen_to == nullptr) { 2219 return false; // dead code 2220 } 2221 if (!opr->is_con() || 2222 vector_klass_from->const_oop() == nullptr || elem_klass_from->const_oop() == nullptr || !vlen_from->is_con() || 2223 vector_klass_to->const_oop() == nullptr || elem_klass_to->const_oop() == nullptr || !vlen_to->is_con()) { 2224 log_if_needed(" ** missing constant: opr=%s vclass_from=%s etype_from=%s vlen_from=%s vclass_to=%s etype_to=%s vlen_to=%s", 2225 NodeClassNames[argument(0)->Opcode()], 2226 NodeClassNames[argument(1)->Opcode()], 2227 NodeClassNames[argument(2)->Opcode()], 2228 NodeClassNames[argument(3)->Opcode()], 2229 NodeClassNames[argument(4)->Opcode()], 2230 NodeClassNames[argument(5)->Opcode()], 2231 NodeClassNames[argument(6)->Opcode()]); 2232 return false; // not enough info for intrinsification 2233 } 2234 if (!is_klass_initialized(vector_klass_from) || !is_klass_initialized(vector_klass_to)) { 2235 log_if_needed(" ** klass argument not initialized"); 2236 return false; 2237 } 2238 2239 assert(opr->get_con() == VectorSupport::VECTOR_OP_CAST || 2240 opr->get_con() == VectorSupport::VECTOR_OP_UCAST || 2241 opr->get_con() == VectorSupport::VECTOR_OP_REINTERPRET, "wrong opcode"); 2242 bool is_cast = (opr->get_con() == VectorSupport::VECTOR_OP_CAST || opr->get_con() == VectorSupport::VECTOR_OP_UCAST); 2243 bool is_ucast = (opr->get_con() == VectorSupport::VECTOR_OP_UCAST); 2244 2245 ciKlass* vbox_klass_from = vector_klass_from->const_oop()->as_instance()->java_lang_Class_klass(); 2246 ciKlass* vbox_klass_to = vector_klass_to->const_oop()->as_instance()->java_lang_Class_klass(); 2247 if (is_vector_shuffle(vbox_klass_from)) { 2248 return false; // vector shuffles aren't supported 2249 } 2250 bool is_mask = is_vector_mask(vbox_klass_from); 2251 2252 ciType* elem_type_from = elem_klass_from->const_oop()->as_instance()->java_mirror_type(); 2253 if (!elem_type_from->is_primitive_type()) { 2254 return false; // should be primitive type 2255 } 2256 BasicType elem_bt_from = elem_type_from->basic_type(); 2257 ciType* elem_type_to = elem_klass_to->const_oop()->as_instance()->java_mirror_type(); 2258 if (!elem_type_to->is_primitive_type()) { 2259 return false; // should be primitive type 2260 } 2261 BasicType elem_bt_to = elem_type_to->basic_type(); 2262 2263 int num_elem_from = vlen_from->get_con(); 2264 int num_elem_to = vlen_to->get_con(); 2265 2266 // Check whether we can unbox to appropriate size. Even with casting, checking for reinterpret is needed 2267 // since we may need to change size. 2268 if (!arch_supports_vector(Op_VectorReinterpret, 2269 num_elem_from, 2270 elem_bt_from, 2271 is_mask ? VecMaskUseAll : VecMaskNotUsed)) { 2272 log_if_needed(" ** not supported: arity=1 op=%s/1 vlen1=%d etype1=%s ismask=%d", 2273 is_cast ? "cast" : "reinterpret", 2274 num_elem_from, type2name(elem_bt_from), is_mask); 2275 return false; 2276 } 2277 2278 // Check whether we can support resizing/reinterpreting to the new size. 2279 if (!arch_supports_vector(Op_VectorReinterpret, 2280 num_elem_to, 2281 elem_bt_to, 2282 is_mask ? VecMaskUseAll : VecMaskNotUsed)) { 2283 log_if_needed(" ** not supported: arity=1 op=%s/2 vlen2=%d etype2=%s ismask=%d", 2284 is_cast ? "cast" : "reinterpret", 2285 num_elem_to, type2name(elem_bt_to), is_mask); 2286 return false; 2287 } 2288 2289 // At this point, we know that both input and output vector registers are supported 2290 // by the architecture. Next check if the casted type is simply to same type - which means 2291 // that it is actually a resize and not a cast. 2292 if (is_cast && elem_bt_from == elem_bt_to) { 2293 is_cast = false; 2294 } 2295 2296 const TypeInstPtr* vbox_type_from = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass_from); 2297 2298 Node* opd1 = unbox_vector(argument(7), vbox_type_from, elem_bt_from, num_elem_from); 2299 if (opd1 == nullptr) { 2300 return false; 2301 } 2302 2303 const TypeVect* src_type = TypeVect::make(elem_bt_from, num_elem_from, is_mask); 2304 const TypeVect* dst_type = TypeVect::make(elem_bt_to, num_elem_to, is_mask); 2305 2306 // Safety check to prevent casting if source mask is of type vector 2307 // and destination mask of type predicate vector and vice-versa. 2308 // From X86 standpoint, this case will only arise over KNL target, 2309 // where certain masks (depending on the species) are either propagated 2310 // through a vector or predicate register. 2311 if (is_mask && 2312 ((src_type->isa_vectmask() == nullptr && dst_type->isa_vectmask()) || 2313 (dst_type->isa_vectmask() == nullptr && src_type->isa_vectmask()))) { 2314 return false; 2315 } 2316 2317 Node* op = opd1; 2318 if (is_cast) { 2319 assert(!is_mask || num_elem_from == num_elem_to, "vector mask cast needs the same elem num"); 2320 int cast_vopc = VectorCastNode::opcode(-1, elem_bt_from, !is_ucast); 2321 2322 // Make sure that vector cast is implemented to particular type/size combination if it is 2323 // not a mask casting. 2324 if (!is_mask && !arch_supports_vector(cast_vopc, num_elem_to, elem_bt_to, VecMaskNotUsed)) { 2325 log_if_needed(" ** not supported: arity=1 op=cast#%d/3 vlen2=%d etype2=%s ismask=%d", 2326 cast_vopc, num_elem_to, type2name(elem_bt_to), is_mask); 2327 return false; 2328 } 2329 2330 if (num_elem_from < num_elem_to) { 2331 // Since input and output number of elements are not consistent, we need to make sure we 2332 // properly size. Thus, first make a cast that retains the number of elements from source. 2333 int num_elem_for_cast = num_elem_from; 2334 2335 // It is possible that arch does not support this intermediate vector size 2336 // TODO More complex logic required here to handle this corner case for the sizes. 2337 if (!arch_supports_vector(cast_vopc, num_elem_for_cast, elem_bt_to, VecMaskNotUsed)) { 2338 log_if_needed(" ** not supported: arity=1 op=cast#%d/4 vlen1=%d etype2=%s ismask=%d", 2339 cast_vopc, 2340 num_elem_for_cast, type2name(elem_bt_to), is_mask); 2341 return false; 2342 } 2343 2344 op = gvn().transform(VectorCastNode::make(cast_vopc, op, elem_bt_to, num_elem_for_cast)); 2345 // Now ensure that the destination gets properly resized to needed size. 2346 op = gvn().transform(new VectorReinterpretNode(op, op->bottom_type()->is_vect(), dst_type)); 2347 } else if (num_elem_from > num_elem_to) { 2348 // Since number of elements from input is larger than output, simply reduce size of input 2349 // (we are supposed to drop top elements anyway). 2350 int num_elem_for_resize = num_elem_to; 2351 2352 // It is possible that arch does not support this intermediate vector size 2353 // TODO More complex logic required here to handle this corner case for the sizes. 2354 if (!arch_supports_vector(Op_VectorReinterpret, 2355 num_elem_for_resize, 2356 elem_bt_from, 2357 VecMaskNotUsed)) { 2358 log_if_needed(" ** not supported: arity=1 op=cast/5 vlen2=%d etype1=%s ismask=%d", 2359 num_elem_for_resize, type2name(elem_bt_from), is_mask); 2360 return false; 2361 } 2362 2363 const TypeVect* resize_type = TypeVect::make(elem_bt_from, num_elem_for_resize); 2364 op = gvn().transform(new VectorReinterpretNode(op, src_type, resize_type)); 2365 op = gvn().transform(VectorCastNode::make(cast_vopc, op, elem_bt_to, num_elem_to)); 2366 } else { // num_elem_from == num_elem_to 2367 if (is_mask) { 2368 // Make sure that cast for vector mask is implemented to particular type/size combination. 2369 if (!arch_supports_vector(Op_VectorMaskCast, num_elem_to, elem_bt_to, VecMaskNotUsed)) { 2370 log_if_needed(" ** not supported: arity=1 op=maskcast vlen2=%d etype2=%s ismask=%d", 2371 num_elem_to, type2name(elem_bt_to), is_mask); 2372 return false; 2373 } 2374 op = gvn().transform(new VectorMaskCastNode(op, dst_type)); 2375 } else { 2376 // Since input and output number of elements match, and since we know this vector size is 2377 // supported, simply do a cast with no resize needed. 2378 op = gvn().transform(VectorCastNode::make(cast_vopc, op, elem_bt_to, num_elem_to)); 2379 } 2380 } 2381 } else if (!Type::equals(src_type, dst_type)) { 2382 assert(!is_cast, "must be reinterpret"); 2383 op = gvn().transform(new VectorReinterpretNode(op, src_type, dst_type)); 2384 } 2385 2386 const TypeInstPtr* vbox_type_to = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass_to); 2387 Node* vbox = box_vector(op, vbox_type_to, elem_bt_to, num_elem_to); 2388 set_result(vbox); 2389 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem_to * type2aelembytes(elem_bt_to)))); 2390 return true; 2391 } 2392 2393 // public static 2394 // <V extends Vector<E>, 2395 // E> 2396 // V insert(Class<? extends V> vectorClass, Class<E> elementType, int vlen, 2397 // V vec, int ix, long val, 2398 // VecInsertOp<V> defaultImpl) 2399 bool LibraryCallKit::inline_vector_insert() { 2400 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 2401 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2402 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2403 const TypeInt* idx = gvn().type(argument(4))->isa_int(); 2404 2405 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || idx == nullptr) { 2406 return false; // dead code 2407 } 2408 if (vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con() || !idx->is_con()) { 2409 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s idx=%s", 2410 NodeClassNames[argument(0)->Opcode()], 2411 NodeClassNames[argument(1)->Opcode()], 2412 NodeClassNames[argument(2)->Opcode()], 2413 NodeClassNames[argument(4)->Opcode()]); 2414 return false; // not enough info for intrinsification 2415 } 2416 if (!is_klass_initialized(vector_klass)) { 2417 log_if_needed(" ** klass argument not initialized"); 2418 return false; 2419 } 2420 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2421 if (!elem_type->is_primitive_type()) { 2422 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2423 return false; // should be primitive type 2424 } 2425 BasicType elem_bt = elem_type->basic_type(); 2426 int num_elem = vlen->get_con(); 2427 if (!arch_supports_vector(Op_VectorInsert, num_elem, elem_bt, VecMaskNotUsed)) { 2428 log_if_needed(" ** not supported: arity=1 op=insert vlen=%d etype=%s ismask=no", 2429 num_elem, type2name(elem_bt)); 2430 return false; // not supported 2431 } 2432 2433 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2434 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2435 2436 Node* opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2437 if (opd == nullptr) { 2438 return false; 2439 } 2440 2441 Node* insert_val = argument(5); 2442 assert(gvn().type(insert_val)->isa_long() != nullptr, "expected to be long"); 2443 2444 // Convert insert value back to its appropriate type. 2445 switch (elem_bt) { 2446 case T_BYTE: 2447 insert_val = gvn().transform(new ConvL2INode(insert_val)); 2448 insert_val = gvn().transform(new CastIINode(insert_val, TypeInt::BYTE)); 2449 break; 2450 case T_SHORT: 2451 insert_val = gvn().transform(new ConvL2INode(insert_val)); 2452 insert_val = gvn().transform(new CastIINode(insert_val, TypeInt::SHORT)); 2453 break; 2454 case T_INT: 2455 insert_val = gvn().transform(new ConvL2INode(insert_val)); 2456 break; 2457 case T_FLOAT: 2458 insert_val = gvn().transform(new ConvL2INode(insert_val)); 2459 insert_val = gvn().transform(new MoveI2FNode(insert_val)); 2460 break; 2461 case T_DOUBLE: 2462 insert_val = gvn().transform(new MoveL2DNode(insert_val)); 2463 break; 2464 case T_LONG: 2465 // no conversion needed 2466 break; 2467 default: fatal("%s", type2name(elem_bt)); break; 2468 } 2469 2470 Node* operation = gvn().transform(VectorInsertNode::make(opd, insert_val, idx->get_con(), gvn())); 2471 2472 Node* vbox = box_vector(operation, vbox_type, elem_bt, num_elem); 2473 set_result(vbox); 2474 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2475 return true; 2476 } 2477 2478 // public static 2479 // <VM extends VectorPayload, 2480 // E> 2481 // long extract(Class<? extends VM> vClass, Class<E> eClass, 2482 // int length, 2483 // VM vm, int i, 2484 // VecExtractOp<VM> defaultImpl) 2485 bool LibraryCallKit::inline_vector_extract() { 2486 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 2487 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2488 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2489 const TypeInt* idx = gvn().type(argument(4))->isa_int(); 2490 2491 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || idx == nullptr) { 2492 return false; // dead code 2493 } 2494 if (vector_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 2495 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s", 2496 NodeClassNames[argument(0)->Opcode()], 2497 NodeClassNames[argument(1)->Opcode()], 2498 NodeClassNames[argument(2)->Opcode()]); 2499 return false; // not enough info for intrinsification 2500 } 2501 if (!is_klass_initialized(vector_klass)) { 2502 log_if_needed(" ** klass argument not initialized"); 2503 return false; 2504 } 2505 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2506 if (!elem_type->is_primitive_type()) { 2507 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2508 return false; // should be primitive type 2509 } 2510 BasicType elem_bt = elem_type->basic_type(); 2511 int num_elem = vlen->get_con(); 2512 2513 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2514 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2515 2516 Node* opd = nullptr; 2517 2518 if (is_vector_mask(vbox_klass)) { 2519 // vbox_klass is mask. This is used for VectorMask.laneIsSet(int). 2520 2521 Node* pos = argument(4); // can be variable 2522 if (arch_supports_vector(Op_ExtractUB, num_elem, elem_bt, VecMaskUseAll)) { 2523 // Transform mask to vector with type of boolean and utilize ExtractUB node. 2524 opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2525 if (opd == nullptr) { 2526 return false; 2527 } 2528 opd = gvn().transform(VectorStoreMaskNode::make(gvn(), opd, elem_bt, num_elem)); 2529 opd = gvn().transform(new ExtractUBNode(opd, pos)); 2530 opd = gvn().transform(new ConvI2LNode(opd)); 2531 } else if (arch_supports_vector(Op_VectorMaskToLong, num_elem, elem_bt, VecMaskUseLoad)) { 2532 opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2533 if (opd == nullptr) { 2534 return false; 2535 } 2536 // VectorMaskToLongNode requires the input is either a mask or a vector with BOOLEAN type. 2537 if (opd->bottom_type()->isa_vectmask() == nullptr) { 2538 opd = gvn().transform(VectorStoreMaskNode::make(gvn(), opd, elem_bt, num_elem)); 2539 } 2540 // ((toLong() >>> pos) & 1L 2541 opd = gvn().transform(new VectorMaskToLongNode(opd, TypeLong::LONG)); 2542 opd = gvn().transform(new URShiftLNode(opd, pos)); 2543 opd = gvn().transform(new AndLNode(opd, gvn().makecon(TypeLong::ONE))); 2544 } else { 2545 log_if_needed(" ** Rejected mask extraction because architecture does not support it"); 2546 return false; // not supported 2547 } 2548 } else { 2549 // vbox_klass is vector. This is used for Vector.lane(int). 2550 if (!idx->is_con()) { 2551 log_if_needed(" ** missing constant: idx=%s", NodeClassNames[argument(4)->Opcode()]); 2552 return false; // not enough info for intrinsification 2553 } 2554 2555 int vopc = ExtractNode::opcode(elem_bt); 2556 if (!arch_supports_vector(vopc, num_elem, elem_bt, VecMaskNotUsed)) { 2557 log_if_needed(" ** not supported: arity=1 op=extract vlen=%d etype=%s ismask=no", 2558 num_elem, type2name(elem_bt)); 2559 return false; // not supported 2560 } 2561 2562 opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2563 if (opd == nullptr) { 2564 return false; 2565 } 2566 ConINode* idx_con = gvn().intcon(idx->get_con())->as_ConI(); 2567 2568 opd = gvn().transform(ExtractNode::make(opd, idx_con, elem_bt)); 2569 switch (elem_bt) { 2570 case T_BYTE: 2571 case T_SHORT: 2572 case T_INT: { 2573 opd = gvn().transform(new ConvI2LNode(opd)); 2574 break; 2575 } 2576 case T_FLOAT: { 2577 opd = gvn().transform(new MoveF2INode(opd)); 2578 opd = gvn().transform(new ConvI2LNode(opd)); 2579 break; 2580 } 2581 case T_DOUBLE: { 2582 opd = gvn().transform(new MoveD2LNode(opd)); 2583 break; 2584 } 2585 case T_LONG: { 2586 // no conversion needed 2587 break; 2588 } 2589 default: fatal("%s", type2name(elem_bt)); 2590 } 2591 } 2592 set_result(opd); 2593 return true; 2594 } 2595 2596 // public static 2597 // <V extends Vector<E>, 2598 // M extends VectorMask<E>, 2599 // E> 2600 // V compressExpandOp(int opr, 2601 // Class<? extends V> vClass, Class<? extends M> mClass, Class<E> eClass, 2602 // int length, V v, M m, 2603 // CompressExpandOperation<V, M> defaultImpl) 2604 bool LibraryCallKit::inline_vector_compress_expand() { 2605 const TypeInt* opr = gvn().type(argument(0))->isa_int(); 2606 const TypeInstPtr* vector_klass = gvn().type(argument(1))->isa_instptr(); 2607 const TypeInstPtr* mask_klass = gvn().type(argument(2))->isa_instptr(); 2608 const TypeInstPtr* elem_klass = gvn().type(argument(3))->isa_instptr(); 2609 const TypeInt* vlen = gvn().type(argument(4))->isa_int(); 2610 2611 if (vector_klass == nullptr || elem_klass == nullptr || mask_klass == nullptr || vlen == nullptr || 2612 vector_klass->const_oop() == nullptr || mask_klass->const_oop() == nullptr || 2613 elem_klass->const_oop() == nullptr || !vlen->is_con() || !opr->is_con()) { 2614 log_if_needed(" ** missing constant: opr=%s vclass=%s mclass=%s etype=%s vlen=%s", 2615 NodeClassNames[argument(0)->Opcode()], 2616 NodeClassNames[argument(1)->Opcode()], 2617 NodeClassNames[argument(2)->Opcode()], 2618 NodeClassNames[argument(3)->Opcode()], 2619 NodeClassNames[argument(4)->Opcode()]); 2620 return false; // not enough info for intrinsification 2621 } 2622 2623 if (!is_klass_initialized(vector_klass) || !is_klass_initialized(mask_klass)) { 2624 log_if_needed(" ** klass argument not initialized"); 2625 return false; 2626 } 2627 2628 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2629 if (!elem_type->is_primitive_type()) { 2630 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2631 return false; // should be primitive type 2632 } 2633 2634 int num_elem = vlen->get_con(); 2635 BasicType elem_bt = elem_type->basic_type(); 2636 int opc = VectorSupport::vop2ideal(opr->get_con(), elem_bt); 2637 2638 if (!arch_supports_vector(opc, num_elem, elem_bt, VecMaskUseLoad)) { 2639 log_if_needed(" ** not supported: opc=%d vlen=%d etype=%s ismask=useload", 2640 opc, num_elem, type2name(elem_bt)); 2641 return false; // not supported 2642 } 2643 2644 Node* opd1 = nullptr; 2645 const TypeInstPtr* vbox_type = nullptr; 2646 if (opc != Op_CompressM) { 2647 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2648 vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2649 opd1 = unbox_vector(argument(5), vbox_type, elem_bt, num_elem); 2650 if (opd1 == nullptr) { 2651 log_if_needed(" ** unbox failed vector=%s", 2652 NodeClassNames[argument(5)->Opcode()]); 2653 return false; 2654 } 2655 } 2656 2657 ciKlass* mbox_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2658 assert(is_vector_mask(mbox_klass), "argument(6) should be a mask class"); 2659 const TypeInstPtr* mbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, mbox_klass); 2660 2661 Node* mask = unbox_vector(argument(6), mbox_type, elem_bt, num_elem); 2662 if (mask == nullptr) { 2663 log_if_needed(" ** unbox failed mask=%s", 2664 NodeClassNames[argument(6)->Opcode()]); 2665 return false; 2666 } 2667 2668 const TypeVect* vt = TypeVect::make(elem_bt, num_elem, opc == Op_CompressM); 2669 Node* operation = gvn().transform(VectorNode::make(opc, opd1, mask, vt)); 2670 2671 // Wrap it up in VectorBox to keep object type information. 2672 const TypeInstPtr* box_type = opc == Op_CompressM ? mbox_type : vbox_type; 2673 Node* vbox = box_vector(operation, box_type, elem_bt, num_elem); 2674 set_result(vbox); 2675 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2676 return true; 2677 } 2678 2679 // public static 2680 // <V extends Vector<E>, 2681 // E, 2682 // S extends VectorSpecies<E>> 2683 // V indexVector(Class<? extends V> vClass, Class<E> eClass, 2684 // int length, 2685 // V v, int step, S s, 2686 // IndexOperation<V, S> defaultImpl) 2687 bool LibraryCallKit::inline_index_vector() { 2688 const TypeInstPtr* vector_klass = gvn().type(argument(0))->isa_instptr(); 2689 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2690 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2691 2692 if (vector_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 2693 vector_klass->const_oop() == nullptr || !vlen->is_con() || 2694 elem_klass->const_oop() == nullptr) { 2695 log_if_needed(" ** missing constant: vclass=%s etype=%s vlen=%s", 2696 NodeClassNames[argument(0)->Opcode()], 2697 NodeClassNames[argument(1)->Opcode()], 2698 NodeClassNames[argument(2)->Opcode()]); 2699 return false; // not enough info for intrinsification 2700 } 2701 2702 if (!is_klass_initialized(vector_klass)) { 2703 log_if_needed(" ** klass argument not initialized"); 2704 return false; 2705 } 2706 2707 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2708 if (!elem_type->is_primitive_type()) { 2709 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2710 return false; // should be primitive type 2711 } 2712 2713 int num_elem = vlen->get_con(); 2714 BasicType elem_bt = elem_type->basic_type(); 2715 2716 // Check whether the iota index generation op is supported by the current hardware 2717 if (!arch_supports_vector(Op_VectorLoadConst, num_elem, elem_bt, VecMaskNotUsed)) { 2718 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 2719 return false; // not supported 2720 } 2721 2722 int mul_op = VectorSupport::vop2ideal(VectorSupport::VECTOR_OP_MUL, elem_bt); 2723 int vmul_op = VectorNode::opcode(mul_op, elem_bt); 2724 bool needs_mul = true; 2725 Node* scale = argument(4); 2726 const TypeInt* scale_type = gvn().type(scale)->isa_int(); 2727 // Multiply is not needed if the scale is a constant "1". 2728 if (scale_type && scale_type->is_con() && scale_type->get_con() == 1) { 2729 needs_mul = false; 2730 } else { 2731 // Check whether the vector multiply op is supported by the current hardware 2732 if (!arch_supports_vector(vmul_op, num_elem, elem_bt, VecMaskNotUsed)) { 2733 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 2734 return false; // not supported 2735 } 2736 2737 // Check whether the scalar cast op is supported by the current hardware 2738 if (is_floating_point_type(elem_bt) || elem_bt == T_LONG) { 2739 int cast_op = elem_bt == T_LONG ? Op_ConvI2L : 2740 elem_bt == T_FLOAT? Op_ConvI2F : Op_ConvI2D; 2741 if (!Matcher::match_rule_supported(cast_op)) { 2742 log_if_needed(" ** Rejected op (%s) because architecture does not support it", 2743 NodeClassNames[cast_op]); 2744 return false; // not supported 2745 } 2746 } 2747 } 2748 2749 ciKlass* vbox_klass = vector_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2750 const TypeInstPtr* vbox_type = TypeInstPtr::make_exact(TypePtr::NotNull, vbox_klass); 2751 Node* opd = unbox_vector(argument(3), vbox_type, elem_bt, num_elem); 2752 if (opd == nullptr) { 2753 log_if_needed(" ** unbox failed vector=%s", 2754 NodeClassNames[argument(3)->Opcode()]); 2755 return false; 2756 } 2757 2758 int add_op = VectorSupport::vop2ideal(VectorSupport::VECTOR_OP_ADD, elem_bt); 2759 int vadd_op = VectorNode::opcode(add_op, elem_bt); 2760 bool needs_add = true; 2761 // The addition is not needed if all the element values of "opd" are zero 2762 if (VectorNode::is_all_zeros_vector(opd)) { 2763 needs_add = false; 2764 } else { 2765 // Check whether the vector addition op is supported by the current hardware 2766 if (!arch_supports_vector(vadd_op, num_elem, elem_bt, VecMaskNotUsed)) { 2767 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 2768 return false; // not supported 2769 } 2770 } 2771 2772 // Compute the iota indice vector 2773 const TypeVect* vt = TypeVect::make(elem_bt, num_elem); 2774 Node* index = gvn().transform(new VectorLoadConstNode(gvn().makecon(TypeInt::ZERO), vt)); 2775 2776 // Broadcast the "scale" to a vector, and multiply the "scale" with iota indice vector. 2777 if (needs_mul) { 2778 switch (elem_bt) { 2779 case T_BOOLEAN: // fall-through 2780 case T_BYTE: // fall-through 2781 case T_SHORT: // fall-through 2782 case T_CHAR: // fall-through 2783 case T_INT: { 2784 // no conversion needed 2785 break; 2786 } 2787 case T_LONG: { 2788 scale = gvn().transform(new ConvI2LNode(scale)); 2789 break; 2790 } 2791 case T_FLOAT: { 2792 scale = gvn().transform(new ConvI2FNode(scale)); 2793 break; 2794 } 2795 case T_DOUBLE: { 2796 scale = gvn().transform(new ConvI2DNode(scale)); 2797 break; 2798 } 2799 default: fatal("%s", type2name(elem_bt)); 2800 } 2801 scale = gvn().transform(VectorNode::scalar2vector(scale, num_elem, Type::get_const_basic_type(elem_bt))); 2802 index = gvn().transform(VectorNode::make(vmul_op, index, scale, vt)); 2803 } 2804 2805 // Add "opd" if addition is needed. 2806 if (needs_add) { 2807 index = gvn().transform(VectorNode::make(vadd_op, opd, index, vt)); 2808 } 2809 Node* vbox = box_vector(index, vbox_type, elem_bt, num_elem); 2810 set_result(vbox); 2811 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2812 return true; 2813 } 2814 2815 // public static 2816 // <E, 2817 // M extends VectorMask<E>> 2818 // M indexPartiallyInUpperRange(Class<? extends M> mClass, Class<E> eClass, int length, 2819 // long offset, long limit, 2820 // IndexPartiallyInUpperRangeOperation<E, M> defaultImpl) 2821 bool LibraryCallKit::inline_index_partially_in_upper_range() { 2822 const TypeInstPtr* mask_klass = gvn().type(argument(0))->isa_instptr(); 2823 const TypeInstPtr* elem_klass = gvn().type(argument(1))->isa_instptr(); 2824 const TypeInt* vlen = gvn().type(argument(2))->isa_int(); 2825 2826 if (mask_klass == nullptr || elem_klass == nullptr || vlen == nullptr || 2827 mask_klass->const_oop() == nullptr || elem_klass->const_oop() == nullptr || !vlen->is_con()) { 2828 log_if_needed(" ** missing constant: mclass=%s etype=%s vlen=%s", 2829 NodeClassNames[argument(0)->Opcode()], 2830 NodeClassNames[argument(1)->Opcode()], 2831 NodeClassNames[argument(2)->Opcode()]); 2832 return false; // not enough info for intrinsification 2833 } 2834 2835 if (!is_klass_initialized(mask_klass)) { 2836 log_if_needed(" ** klass argument not initialized"); 2837 return false; 2838 } 2839 2840 ciType* elem_type = elem_klass->const_oop()->as_instance()->java_mirror_type(); 2841 if (!elem_type->is_primitive_type()) { 2842 log_if_needed(" ** not a primitive bt=%d", elem_type->basic_type()); 2843 return false; // should be primitive type 2844 } 2845 2846 int num_elem = vlen->get_con(); 2847 BasicType elem_bt = elem_type->basic_type(); 2848 2849 // Check whether the necessary ops are supported by current hardware. 2850 bool supports_mask_gen = arch_supports_vector(Op_VectorMaskGen, num_elem, elem_bt, VecMaskUseStore); 2851 if (!supports_mask_gen) { 2852 if (!arch_supports_vector(Op_VectorLoadConst, num_elem, elem_bt, VecMaskNotUsed) || 2853 !arch_supports_vector(Op_Replicate, num_elem, elem_bt, VecMaskNotUsed) || 2854 !arch_supports_vector(Op_VectorMaskCmp, num_elem, elem_bt, VecMaskUseStore)) { 2855 log_if_needed(" ** not supported: vlen=%d etype=%s", num_elem, type2name(elem_bt)); 2856 return false; // not supported 2857 } 2858 2859 // Check whether the scalar cast operation is supported by current hardware. 2860 if (elem_bt != T_LONG) { 2861 int cast_op = is_integral_type(elem_bt) ? Op_ConvL2I 2862 : (elem_bt == T_FLOAT ? Op_ConvL2F : Op_ConvL2D); 2863 if (!Matcher::match_rule_supported(cast_op)) { 2864 log_if_needed(" ** Rejected op (%s) because architecture does not support it", 2865 NodeClassNames[cast_op]); 2866 return false; // not supported 2867 } 2868 } 2869 } 2870 2871 Node* offset = argument(3); 2872 Node* limit = argument(5); 2873 if (offset == nullptr || limit == nullptr) { 2874 log_if_needed(" ** offset or limit argument is null"); 2875 return false; // not supported 2876 } 2877 2878 ciKlass* box_klass = mask_klass->const_oop()->as_instance()->java_lang_Class_klass(); 2879 assert(is_vector_mask(box_klass), "argument(0) should be a mask class"); 2880 const TypeInstPtr* box_type = TypeInstPtr::make_exact(TypePtr::NotNull, box_klass); 2881 2882 // We assume "offset > 0 && limit >= offset && limit - offset < num_elem". 2883 // So directly get indexLimit with "indexLimit = limit - offset". 2884 Node* indexLimit = gvn().transform(new SubLNode(limit, offset)); 2885 Node* mask = nullptr; 2886 if (supports_mask_gen) { 2887 mask = gvn().transform(VectorMaskGenNode::make(indexLimit, elem_bt, num_elem)); 2888 } else { 2889 // Generate the vector mask based on "mask = iota < indexLimit". 2890 // Broadcast "indexLimit" to a vector. 2891 switch (elem_bt) { 2892 case T_BOOLEAN: // fall-through 2893 case T_BYTE: // fall-through 2894 case T_SHORT: // fall-through 2895 case T_CHAR: // fall-through 2896 case T_INT: { 2897 indexLimit = gvn().transform(new ConvL2INode(indexLimit)); 2898 break; 2899 } 2900 case T_DOUBLE: { 2901 indexLimit = gvn().transform(new ConvL2DNode(indexLimit)); 2902 break; 2903 } 2904 case T_FLOAT: { 2905 indexLimit = gvn().transform(new ConvL2FNode(indexLimit)); 2906 break; 2907 } 2908 case T_LONG: { 2909 // no conversion needed 2910 break; 2911 } 2912 default: fatal("%s", type2name(elem_bt)); 2913 } 2914 indexLimit = gvn().transform(VectorNode::scalar2vector(indexLimit, num_elem, Type::get_const_basic_type(elem_bt))); 2915 2916 // Load the "iota" vector. 2917 const TypeVect* vt = TypeVect::make(elem_bt, num_elem); 2918 Node* iota = gvn().transform(new VectorLoadConstNode(gvn().makecon(TypeInt::ZERO), vt)); 2919 2920 // Compute the vector mask with "mask = iota < indexLimit". 2921 ConINode* pred_node = (ConINode*)gvn().makecon(TypeInt::make(BoolTest::lt)); 2922 const TypeVect* vmask_type = TypeVect::makemask(elem_bt, num_elem); 2923 mask = gvn().transform(new VectorMaskCmpNode(BoolTest::lt, iota, indexLimit, pred_node, vmask_type)); 2924 } 2925 Node* vbox = box_vector(mask, box_type, elem_bt, num_elem); 2926 set_result(vbox); 2927 C->set_max_vector_size(MAX2(C->max_vector_size(), (uint)(num_elem * type2aelembytes(elem_bt)))); 2928 return true; 2929 } 2930 2931 #undef non_product_log_if_needed 2932 #undef log_if_needed