1 /* 2 * Copyright (c) 2017, 2022, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. Oracle designates this 8 * particular file as subject to the "Classpath" exception as provided 9 * by Oracle in the LICENSE file that accompanied this code. 10 * 11 * This code is distributed in the hope that it will be useful, but WITHOUT 12 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 13 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 14 * version 2 for more details (a copy is included in the LICENSE file that 15 * accompanied this code). 16 * 17 * You should have received a copy of the GNU General Public License version 18 * 2 along with this work; if not, write to the Free Software Foundation, 19 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 20 * 21 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 22 * or visit www.oracle.com if you need additional information or have any 23 * questions. 24 */ 25 package jdk.incubator.vector; 26 27 import java.util.Arrays; 28 import java.util.Objects; 29 import java.util.function.IntUnaryOperator; 30 31 import jdk.incubator.foreign.MemorySegment; 32 import jdk.internal.vm.annotation.ForceInline; 33 import jdk.internal.vm.vector.VectorSupport; 34 35 import static jdk.internal.vm.vector.VectorSupport.*; 36 37 import static jdk.incubator.vector.VectorOperators.*; 38 39 // -- This file was mechanically generated: Do not edit! -- // 40 41 @SuppressWarnings("cast") // warning: redundant cast 42 final class Int64Vector extends IntVector { 43 static final IntSpecies VSPECIES = 44 (IntSpecies) IntVector.SPECIES_64; 45 46 static final VectorShape VSHAPE = 47 VSPECIES.vectorShape(); 48 49 static final Class<Int64Vector> VCLASS = Int64Vector.class; 50 51 static final int VSIZE = VSPECIES.vectorBitSize(); 52 53 static final int VLENGTH = VSPECIES.laneCount(); // used by the JVM 54 55 static final Class<Integer> ETYPE = int.class; // used by the JVM 56 57 Int64Vector(int[] v) { 58 super(v); 59 } 60 61 // For compatibility as Int64Vector::new, 62 // stored into species.vectorFactory. 63 Int64Vector(Object v) { 64 this((int[]) v); 65 } 66 67 static final Int64Vector ZERO = new Int64Vector(new int[VLENGTH]); 68 static final Int64Vector IOTA = new Int64Vector(VSPECIES.iotaArray()); 69 70 static { 71 // Warm up a few species caches. 72 // If we do this too much we will 73 // get NPEs from bootstrap circularity. 74 VSPECIES.dummyVector(); 75 VSPECIES.withLanes(LaneType.BYTE); 76 } 77 78 // Specialized extractors 79 80 @ForceInline 81 final @Override 82 public IntSpecies vspecies() { 83 // ISSUE: This should probably be a @Stable 84 // field inside AbstractVector, rather than 85 // a megamorphic method. 86 return VSPECIES; 87 } 88 89 @ForceInline 90 @Override 91 public final Class<Integer> elementType() { return int.class; } 92 93 @ForceInline 94 @Override 95 public final int elementSize() { return Integer.SIZE; } 96 97 @ForceInline 98 @Override 99 public final VectorShape shape() { return VSHAPE; } 100 101 @ForceInline 102 @Override 103 public final int length() { return VLENGTH; } 104 105 @ForceInline 106 @Override 107 public final int bitSize() { return VSIZE; } 108 109 @ForceInline 110 @Override 111 public final int byteSize() { return VSIZE / Byte.SIZE; } 112 113 /*package-private*/ 114 @ForceInline 115 final @Override 116 int[] vec() { 117 return (int[])getPayload(); 118 } 119 120 // Virtualized constructors 121 122 @Override 123 @ForceInline 124 public final Int64Vector broadcast(int e) { 125 return (Int64Vector) super.broadcastTemplate(e); // specialize 126 } 127 128 @Override 129 @ForceInline 130 public final Int64Vector broadcast(long e) { 131 return (Int64Vector) super.broadcastTemplate(e); // specialize 132 } 133 134 @Override 135 @ForceInline 136 Int64Mask maskFromArray(boolean[] bits) { 137 return new Int64Mask(bits); 138 } 139 140 @Override 141 @ForceInline 142 Int64Shuffle iotaShuffle() { return Int64Shuffle.IOTA; } 143 144 @ForceInline 145 Int64Shuffle iotaShuffle(int start, int step, boolean wrap) { 146 if (wrap) { 147 return (Int64Shuffle)VectorSupport.shuffleIota(ETYPE, Int64Shuffle.class, VSPECIES, VLENGTH, start, step, 1, 148 (l, lstart, lstep, s) -> s.shuffleFromOp(i -> (VectorIntrinsics.wrapToRange(i*lstep + lstart, l)))); 149 } else { 150 return (Int64Shuffle)VectorSupport.shuffleIota(ETYPE, Int64Shuffle.class, VSPECIES, VLENGTH, start, step, 0, 151 (l, lstart, lstep, s) -> s.shuffleFromOp(i -> (i*lstep + lstart))); 152 } 153 } 154 155 @Override 156 @ForceInline 157 Int64Shuffle shuffleFromBytes(byte[] reorder) { return new Int64Shuffle(reorder); } 158 159 @Override 160 @ForceInline 161 Int64Shuffle shuffleFromArray(int[] indexes, int i) { return new Int64Shuffle(indexes, i); } 162 163 @Override 164 @ForceInline 165 Int64Shuffle shuffleFromOp(IntUnaryOperator fn) { return new Int64Shuffle(fn); } 166 167 // Make a vector of the same species but the given elements: 168 @ForceInline 169 final @Override 170 Int64Vector vectorFactory(int[] vec) { 171 return new Int64Vector(vec); 172 } 173 174 @ForceInline 175 final @Override 176 Byte64Vector asByteVectorRaw() { 177 return (Byte64Vector) super.asByteVectorRawTemplate(); // specialize 178 } 179 180 @ForceInline 181 final @Override 182 AbstractVector<?> asVectorRaw(LaneType laneType) { 183 return super.asVectorRawTemplate(laneType); // specialize 184 } 185 186 // Unary operator 187 188 @ForceInline 189 final @Override 190 Int64Vector uOp(FUnOp f) { 191 return (Int64Vector) super.uOpTemplate(f); // specialize 192 } 193 194 @ForceInline 195 final @Override 196 Int64Vector uOp(VectorMask<Integer> m, FUnOp f) { 197 return (Int64Vector) 198 super.uOpTemplate((Int64Mask)m, f); // specialize 199 } 200 201 // Binary operator 202 203 @ForceInline 204 final @Override 205 Int64Vector bOp(Vector<Integer> v, FBinOp f) { 206 return (Int64Vector) super.bOpTemplate((Int64Vector)v, f); // specialize 207 } 208 209 @ForceInline 210 final @Override 211 Int64Vector bOp(Vector<Integer> v, 212 VectorMask<Integer> m, FBinOp f) { 213 return (Int64Vector) 214 super.bOpTemplate((Int64Vector)v, (Int64Mask)m, 215 f); // specialize 216 } 217 218 // Ternary operator 219 220 @ForceInline 221 final @Override 222 Int64Vector tOp(Vector<Integer> v1, Vector<Integer> v2, FTriOp f) { 223 return (Int64Vector) 224 super.tOpTemplate((Int64Vector)v1, (Int64Vector)v2, 225 f); // specialize 226 } 227 228 @ForceInline 229 final @Override 230 Int64Vector tOp(Vector<Integer> v1, Vector<Integer> v2, 231 VectorMask<Integer> m, FTriOp f) { 232 return (Int64Vector) 233 super.tOpTemplate((Int64Vector)v1, (Int64Vector)v2, 234 (Int64Mask)m, f); // specialize 235 } 236 237 @ForceInline 238 final @Override 239 int rOp(int v, VectorMask<Integer> m, FBinOp f) { 240 return super.rOpTemplate(v, m, f); // specialize 241 } 242 243 @Override 244 @ForceInline 245 public final <F> 246 Vector<F> convertShape(VectorOperators.Conversion<Integer,F> conv, 247 VectorSpecies<F> rsp, int part) { 248 return super.convertShapeTemplate(conv, rsp, part); // specialize 249 } 250 251 @Override 252 @ForceInline 253 public final <F> 254 Vector<F> reinterpretShape(VectorSpecies<F> toSpecies, int part) { 255 return super.reinterpretShapeTemplate(toSpecies, part); // specialize 256 } 257 258 // Specialized algebraic operations: 259 260 // The following definition forces a specialized version of this 261 // crucial method into the v-table of this class. A call to add() 262 // will inline to a call to lanewise(ADD,), at which point the JIT 263 // intrinsic will have the opcode of ADD, plus all the metadata 264 // for this particular class, enabling it to generate precise 265 // code. 266 // 267 // There is probably no benefit to the JIT to specialize the 268 // masked or broadcast versions of the lanewise method. 269 270 @Override 271 @ForceInline 272 public Int64Vector lanewise(Unary op) { 273 return (Int64Vector) super.lanewiseTemplate(op); // specialize 274 } 275 276 @Override 277 @ForceInline 278 public Int64Vector lanewise(Unary op, VectorMask<Integer> m) { 279 return (Int64Vector) super.lanewiseTemplate(op, Int64Mask.class, (Int64Mask) m); // specialize 280 } 281 282 @Override 283 @ForceInline 284 public Int64Vector lanewise(Binary op, Vector<Integer> v) { 285 return (Int64Vector) super.lanewiseTemplate(op, v); // specialize 286 } 287 288 @Override 289 @ForceInline 290 public Int64Vector lanewise(Binary op, Vector<Integer> v, VectorMask<Integer> m) { 291 return (Int64Vector) super.lanewiseTemplate(op, Int64Mask.class, v, (Int64Mask) m); // specialize 292 } 293 294 /*package-private*/ 295 @Override 296 @ForceInline Int64Vector 297 lanewiseShift(VectorOperators.Binary op, int e) { 298 return (Int64Vector) super.lanewiseShiftTemplate(op, e); // specialize 299 } 300 301 /*package-private*/ 302 @Override 303 @ForceInline Int64Vector 304 lanewiseShift(VectorOperators.Binary op, int e, VectorMask<Integer> m) { 305 return (Int64Vector) super.lanewiseShiftTemplate(op, Int64Mask.class, e, (Int64Mask) m); // specialize 306 } 307 308 /*package-private*/ 309 @Override 310 @ForceInline 311 public final 312 Int64Vector 313 lanewise(Ternary op, Vector<Integer> v1, Vector<Integer> v2) { 314 return (Int64Vector) super.lanewiseTemplate(op, v1, v2); // specialize 315 } 316 317 @Override 318 @ForceInline 319 public final 320 Int64Vector 321 lanewise(Ternary op, Vector<Integer> v1, Vector<Integer> v2, VectorMask<Integer> m) { 322 return (Int64Vector) super.lanewiseTemplate(op, Int64Mask.class, v1, v2, (Int64Mask) m); // specialize 323 } 324 325 @Override 326 @ForceInline 327 public final 328 Int64Vector addIndex(int scale) { 329 return (Int64Vector) super.addIndexTemplate(scale); // specialize 330 } 331 332 // Type specific horizontal reductions 333 334 @Override 335 @ForceInline 336 public final int reduceLanes(VectorOperators.Associative op) { 337 return super.reduceLanesTemplate(op); // specialized 338 } 339 340 @Override 341 @ForceInline 342 public final int reduceLanes(VectorOperators.Associative op, 343 VectorMask<Integer> m) { 344 return super.reduceLanesTemplate(op, Int64Mask.class, (Int64Mask) m); // specialized 345 } 346 347 @Override 348 @ForceInline 349 public final long reduceLanesToLong(VectorOperators.Associative op) { 350 return (long) super.reduceLanesTemplate(op); // specialized 351 } 352 353 @Override 354 @ForceInline 355 public final long reduceLanesToLong(VectorOperators.Associative op, 356 VectorMask<Integer> m) { 357 return (long) super.reduceLanesTemplate(op, Int64Mask.class, (Int64Mask) m); // specialized 358 } 359 360 @ForceInline 361 public VectorShuffle<Integer> toShuffle() { 362 return super.toShuffleTemplate(Int64Shuffle.class); // specialize 363 } 364 365 // Specialized unary testing 366 367 @Override 368 @ForceInline 369 public final Int64Mask test(Test op) { 370 return super.testTemplate(Int64Mask.class, op); // specialize 371 } 372 373 @Override 374 @ForceInline 375 public final Int64Mask test(Test op, VectorMask<Integer> m) { 376 return super.testTemplate(Int64Mask.class, op, (Int64Mask) m); // specialize 377 } 378 379 // Specialized comparisons 380 381 @Override 382 @ForceInline 383 public final Int64Mask compare(Comparison op, Vector<Integer> v) { 384 return super.compareTemplate(Int64Mask.class, op, v); // specialize 385 } 386 387 @Override 388 @ForceInline 389 public final Int64Mask compare(Comparison op, int s) { 390 return super.compareTemplate(Int64Mask.class, op, s); // specialize 391 } 392 393 @Override 394 @ForceInline 395 public final Int64Mask compare(Comparison op, long s) { 396 return super.compareTemplate(Int64Mask.class, op, s); // specialize 397 } 398 399 @Override 400 @ForceInline 401 public final Int64Mask compare(Comparison op, Vector<Integer> v, VectorMask<Integer> m) { 402 return super.compareTemplate(Int64Mask.class, op, v, (Int64Mask) m); 403 } 404 405 406 @Override 407 @ForceInline 408 public Int64Vector blend(Vector<Integer> v, VectorMask<Integer> m) { 409 return (Int64Vector) 410 super.blendTemplate(Int64Mask.class, 411 (Int64Vector) v, 412 (Int64Mask) m); // specialize 413 } 414 415 @Override 416 @ForceInline 417 public Int64Vector slice(int origin, Vector<Integer> v) { 418 return (Int64Vector) super.sliceTemplate(origin, v); // specialize 419 } 420 421 @Override 422 @ForceInline 423 public Int64Vector slice(int origin) { 424 return (Int64Vector) super.sliceTemplate(origin); // specialize 425 } 426 427 @Override 428 @ForceInline 429 public Int64Vector unslice(int origin, Vector<Integer> w, int part) { 430 return (Int64Vector) super.unsliceTemplate(origin, w, part); // specialize 431 } 432 433 @Override 434 @ForceInline 435 public Int64Vector unslice(int origin, Vector<Integer> w, int part, VectorMask<Integer> m) { 436 return (Int64Vector) 437 super.unsliceTemplate(Int64Mask.class, 438 origin, w, part, 439 (Int64Mask) m); // specialize 440 } 441 442 @Override 443 @ForceInline 444 public Int64Vector unslice(int origin) { 445 return (Int64Vector) super.unsliceTemplate(origin); // specialize 446 } 447 448 @Override 449 @ForceInline 450 public Int64Vector rearrange(VectorShuffle<Integer> s) { 451 return (Int64Vector) 452 super.rearrangeTemplate(Int64Shuffle.class, 453 (Int64Shuffle) s); // specialize 454 } 455 456 @Override 457 @ForceInline 458 public Int64Vector rearrange(VectorShuffle<Integer> shuffle, 459 VectorMask<Integer> m) { 460 return (Int64Vector) 461 super.rearrangeTemplate(Int64Shuffle.class, 462 Int64Mask.class, 463 (Int64Shuffle) shuffle, 464 (Int64Mask) m); // specialize 465 } 466 467 @Override 468 @ForceInline 469 public Int64Vector rearrange(VectorShuffle<Integer> s, 470 Vector<Integer> v) { 471 return (Int64Vector) 472 super.rearrangeTemplate(Int64Shuffle.class, 473 (Int64Shuffle) s, 474 (Int64Vector) v); // specialize 475 } 476 477 @Override 478 @ForceInline 479 public Int64Vector compress(VectorMask<Integer> m) { 480 return (Int64Vector) 481 super.compressTemplate(Int64Mask.class, 482 (Int64Mask) m); // specialize 483 } 484 485 @Override 486 @ForceInline 487 public Int64Vector expand(VectorMask<Integer> m) { 488 return (Int64Vector) 489 super.expandTemplate(Int64Mask.class, 490 (Int64Mask) m); // specialize 491 } 492 493 @Override 494 @ForceInline 495 public Int64Vector selectFrom(Vector<Integer> v) { 496 return (Int64Vector) 497 super.selectFromTemplate((Int64Vector) v); // specialize 498 } 499 500 @Override 501 @ForceInline 502 public Int64Vector selectFrom(Vector<Integer> v, 503 VectorMask<Integer> m) { 504 return (Int64Vector) 505 super.selectFromTemplate((Int64Vector) v, 506 (Int64Mask) m); // specialize 507 } 508 509 510 @ForceInline 511 @Override 512 public int lane(int i) { 513 switch(i) { 514 case 0: return laneHelper(0); 515 case 1: return laneHelper(1); 516 default: throw new IllegalArgumentException("Index " + i + " must be zero or positive, and less than " + VLENGTH); 517 } 518 } 519 520 public int laneHelper(int i) { 521 return (int) VectorSupport.extract( 522 VCLASS, ETYPE, VLENGTH, 523 this, i, 524 (vec, ix) -> { 525 int[] vecarr = vec.vec(); 526 return (long)vecarr[ix]; 527 }); 528 } 529 530 @ForceInline 531 @Override 532 public Int64Vector withLane(int i, int e) { 533 switch (i) { 534 case 0: return withLaneHelper(0, e); 535 case 1: return withLaneHelper(1, e); 536 default: throw new IllegalArgumentException("Index " + i + " must be zero or positive, and less than " + VLENGTH); 537 } 538 } 539 540 public Int64Vector withLaneHelper(int i, int e) { 541 return VectorSupport.insert( 542 VCLASS, ETYPE, VLENGTH, 543 this, i, (long)e, 544 (v, ix, bits) -> { 545 int[] res = v.vec().clone(); 546 res[ix] = (int)bits; 547 return v.vectorFactory(res); 548 }); 549 } 550 551 // Mask 552 553 static final class Int64Mask extends AbstractMask<Integer> { 554 static final int VLENGTH = VSPECIES.laneCount(); // used by the JVM 555 static final Class<Integer> ETYPE = int.class; // used by the JVM 556 557 Int64Mask(boolean[] bits) { 558 this(bits, 0); 559 } 560 561 Int64Mask(boolean[] bits, int offset) { 562 super(prepare(bits, offset)); 563 } 564 565 Int64Mask(boolean val) { 566 super(prepare(val)); 567 } 568 569 private static boolean[] prepare(boolean[] bits, int offset) { 570 boolean[] newBits = new boolean[VSPECIES.laneCount()]; 571 for (int i = 0; i < newBits.length; i++) { 572 newBits[i] = bits[offset + i]; 573 } 574 return newBits; 575 } 576 577 private static boolean[] prepare(boolean val) { 578 boolean[] bits = new boolean[VSPECIES.laneCount()]; 579 Arrays.fill(bits, val); 580 return bits; 581 } 582 583 @ForceInline 584 final @Override 585 public IntSpecies vspecies() { 586 // ISSUE: This should probably be a @Stable 587 // field inside AbstractMask, rather than 588 // a megamorphic method. 589 return VSPECIES; 590 } 591 592 @ForceInline 593 boolean[] getBits() { 594 return (boolean[])getPayload(); 595 } 596 597 @Override 598 Int64Mask uOp(MUnOp f) { 599 boolean[] res = new boolean[vspecies().laneCount()]; 600 boolean[] bits = getBits(); 601 for (int i = 0; i < res.length; i++) { 602 res[i] = f.apply(i, bits[i]); 603 } 604 return new Int64Mask(res); 605 } 606 607 @Override 608 Int64Mask bOp(VectorMask<Integer> m, MBinOp f) { 609 boolean[] res = new boolean[vspecies().laneCount()]; 610 boolean[] bits = getBits(); 611 boolean[] mbits = ((Int64Mask)m).getBits(); 612 for (int i = 0; i < res.length; i++) { 613 res[i] = f.apply(i, bits[i], mbits[i]); 614 } 615 return new Int64Mask(res); 616 } 617 618 @ForceInline 619 @Override 620 public final 621 Int64Vector toVector() { 622 return (Int64Vector) super.toVectorTemplate(); // specialize 623 } 624 625 /** 626 * Helper function for lane-wise mask conversions. 627 * This function kicks in after intrinsic failure. 628 */ 629 @ForceInline 630 private final <E> 631 VectorMask<E> defaultMaskCast(AbstractSpecies<E> dsp) { 632 if (length() != dsp.laneCount()) 633 throw new IllegalArgumentException("VectorMask length and species length differ"); 634 boolean[] maskArray = toArray(); 635 return dsp.maskFactory(maskArray).check(dsp); 636 } 637 638 @Override 639 @ForceInline 640 public <E> VectorMask<E> cast(VectorSpecies<E> dsp) { 641 AbstractSpecies<E> species = (AbstractSpecies<E>) dsp; 642 if (length() != species.laneCount()) 643 throw new IllegalArgumentException("VectorMask length and species length differ"); 644 645 return VectorSupport.convert(VectorSupport.VECTOR_OP_CAST, 646 this.getClass(), ETYPE, VLENGTH, 647 species.maskType(), species.elementType(), VLENGTH, 648 this, species, 649 (m, s) -> s.maskFactory(m.toArray()).check(s)); 650 } 651 652 @Override 653 @ForceInline 654 public Int64Mask eq(VectorMask<Integer> mask) { 655 Objects.requireNonNull(mask); 656 Int64Mask m = (Int64Mask)mask; 657 return xor(m.not()); 658 } 659 660 // Unary operations 661 662 @Override 663 @ForceInline 664 public Int64Mask not() { 665 return xor(maskAll(true)); 666 } 667 668 @Override 669 @ForceInline 670 public Int64Mask compress() { 671 return (Int64Mask)VectorSupport.comExpOp(VectorSupport.VECTOR_OP_MASK_COMPRESS, 672 Int64Vector.class, Int64Mask.class, ETYPE, VLENGTH, null, this, 673 (v1, m1) -> VSPECIES.iota().compare(VectorOperators.LT, m1.trueCount())); 674 } 675 676 677 // Binary operations 678 679 @Override 680 @ForceInline 681 public Int64Mask and(VectorMask<Integer> mask) { 682 Objects.requireNonNull(mask); 683 Int64Mask m = (Int64Mask)mask; 684 return VectorSupport.binaryOp(VECTOR_OP_AND, Int64Mask.class, null, int.class, VLENGTH, 685 this, m, null, 686 (m1, m2, vm) -> m1.bOp(m2, (i, a, b) -> a & b)); 687 } 688 689 @Override 690 @ForceInline 691 public Int64Mask or(VectorMask<Integer> mask) { 692 Objects.requireNonNull(mask); 693 Int64Mask m = (Int64Mask)mask; 694 return VectorSupport.binaryOp(VECTOR_OP_OR, Int64Mask.class, null, int.class, VLENGTH, 695 this, m, null, 696 (m1, m2, vm) -> m1.bOp(m2, (i, a, b) -> a | b)); 697 } 698 699 @ForceInline 700 /* package-private */ 701 Int64Mask xor(VectorMask<Integer> mask) { 702 Objects.requireNonNull(mask); 703 Int64Mask m = (Int64Mask)mask; 704 return VectorSupport.binaryOp(VECTOR_OP_XOR, Int64Mask.class, null, int.class, VLENGTH, 705 this, m, null, 706 (m1, m2, vm) -> m1.bOp(m2, (i, a, b) -> a ^ b)); 707 } 708 709 // Mask Query operations 710 711 @Override 712 @ForceInline 713 public int trueCount() { 714 return (int) VectorSupport.maskReductionCoerced(VECTOR_OP_MASK_TRUECOUNT, Int64Mask.class, int.class, VLENGTH, this, 715 (m) -> trueCountHelper(m.getBits())); 716 } 717 718 @Override 719 @ForceInline 720 public int firstTrue() { 721 return (int) VectorSupport.maskReductionCoerced(VECTOR_OP_MASK_FIRSTTRUE, Int64Mask.class, int.class, VLENGTH, this, 722 (m) -> firstTrueHelper(m.getBits())); 723 } 724 725 @Override 726 @ForceInline 727 public int lastTrue() { 728 return (int) VectorSupport.maskReductionCoerced(VECTOR_OP_MASK_LASTTRUE, Int64Mask.class, int.class, VLENGTH, this, 729 (m) -> lastTrueHelper(m.getBits())); 730 } 731 732 @Override 733 @ForceInline 734 public long toLong() { 735 if (length() > Long.SIZE) { 736 throw new UnsupportedOperationException("too many lanes for one long"); 737 } 738 return VectorSupport.maskReductionCoerced(VECTOR_OP_MASK_TOLONG, Int64Mask.class, int.class, VLENGTH, this, 739 (m) -> toLongHelper(m.getBits())); 740 } 741 742 // Reductions 743 744 @Override 745 @ForceInline 746 public boolean anyTrue() { 747 return VectorSupport.test(BT_ne, Int64Mask.class, int.class, VLENGTH, 748 this, vspecies().maskAll(true), 749 (m, __) -> anyTrueHelper(((Int64Mask)m).getBits())); 750 } 751 752 @Override 753 @ForceInline 754 public boolean allTrue() { 755 return VectorSupport.test(BT_overflow, Int64Mask.class, int.class, VLENGTH, 756 this, vspecies().maskAll(true), 757 (m, __) -> allTrueHelper(((Int64Mask)m).getBits())); 758 } 759 760 @ForceInline 761 /*package-private*/ 762 static Int64Mask maskAll(boolean bit) { 763 return VectorSupport.fromBitsCoerced(Int64Mask.class, int.class, VLENGTH, 764 (bit ? -1 : 0), MODE_BROADCAST, null, 765 (v, __) -> (v != 0 ? TRUE_MASK : FALSE_MASK)); 766 } 767 private static final Int64Mask TRUE_MASK = new Int64Mask(true); 768 private static final Int64Mask FALSE_MASK = new Int64Mask(false); 769 770 } 771 772 // Shuffle 773 774 static final class Int64Shuffle extends AbstractShuffle<Integer> { 775 static final int VLENGTH = VSPECIES.laneCount(); // used by the JVM 776 static final Class<Integer> ETYPE = int.class; // used by the JVM 777 778 Int64Shuffle(byte[] reorder) { 779 super(VLENGTH, reorder); 780 } 781 782 public Int64Shuffle(int[] reorder) { 783 super(VLENGTH, reorder); 784 } 785 786 public Int64Shuffle(int[] reorder, int i) { 787 super(VLENGTH, reorder, i); 788 } 789 790 public Int64Shuffle(IntUnaryOperator fn) { 791 super(VLENGTH, fn); 792 } 793 794 @Override 795 public IntSpecies vspecies() { 796 return VSPECIES; 797 } 798 799 static { 800 // There must be enough bits in the shuffle lanes to encode 801 // VLENGTH valid indexes and VLENGTH exceptional ones. 802 assert(VLENGTH < Byte.MAX_VALUE); 803 assert(Byte.MIN_VALUE <= -VLENGTH); 804 } 805 static final Int64Shuffle IOTA = new Int64Shuffle(IDENTITY); 806 807 @Override 808 @ForceInline 809 public Int64Vector toVector() { 810 return VectorSupport.shuffleToVector(VCLASS, ETYPE, Int64Shuffle.class, this, VLENGTH, 811 (s) -> ((Int64Vector)(((AbstractShuffle<Integer>)(s)).toVectorTemplate()))); 812 } 813 814 @Override 815 @ForceInline 816 public <F> VectorShuffle<F> cast(VectorSpecies<F> s) { 817 AbstractSpecies<F> species = (AbstractSpecies<F>) s; 818 if (length() != species.laneCount()) 819 throw new IllegalArgumentException("VectorShuffle length and species length differ"); 820 int[] shuffleArray = toArray(); 821 return s.shuffleFromArray(shuffleArray, 0).check(s); 822 } 823 824 @ForceInline 825 @Override 826 public Int64Shuffle rearrange(VectorShuffle<Integer> shuffle) { 827 Int64Shuffle s = (Int64Shuffle) shuffle; 828 byte[] reorder1 = reorder(); 829 byte[] reorder2 = s.reorder(); 830 byte[] r = new byte[reorder1.length]; 831 for (int i = 0; i < reorder1.length; i++) { 832 int ssi = reorder2[i]; 833 r[i] = reorder1[ssi]; // throws on exceptional index 834 } 835 return new Int64Shuffle(r); 836 } 837 } 838 839 // ================================================ 840 841 // Specialized low-level memory operations. 842 843 @ForceInline 844 @Override 845 final 846 IntVector fromArray0(int[] a, int offset) { 847 return super.fromArray0Template(a, offset); // specialize 848 } 849 850 @ForceInline 851 @Override 852 final 853 IntVector fromArray0(int[] a, int offset, VectorMask<Integer> m) { 854 return super.fromArray0Template(Int64Mask.class, a, offset, (Int64Mask) m); // specialize 855 } 856 857 @ForceInline 858 @Override 859 final 860 IntVector fromArray0(int[] a, int offset, int[] indexMap, int mapOffset, VectorMask<Integer> m) { 861 return super.fromArray0Template(Int64Mask.class, a, offset, indexMap, mapOffset, (Int64Mask) m); 862 } 863 864 865 866 @ForceInline 867 @Override 868 final 869 IntVector fromMemorySegment0(MemorySegment ms, long offset) { 870 return super.fromMemorySegment0Template(ms, offset); // specialize 871 } 872 873 @ForceInline 874 @Override 875 final 876 IntVector fromMemorySegment0(MemorySegment ms, long offset, VectorMask<Integer> m) { 877 return super.fromMemorySegment0Template(Int64Mask.class, ms, offset, (Int64Mask) m); // specialize 878 } 879 880 @ForceInline 881 @Override 882 final 883 void intoArray0(int[] a, int offset) { 884 super.intoArray0Template(a, offset); // specialize 885 } 886 887 @ForceInline 888 @Override 889 final 890 void intoArray0(int[] a, int offset, VectorMask<Integer> m) { 891 super.intoArray0Template(Int64Mask.class, a, offset, (Int64Mask) m); 892 } 893 894 @ForceInline 895 @Override 896 final 897 void intoArray0(int[] a, int offset, int[] indexMap, int mapOffset, VectorMask<Integer> m) { 898 super.intoArray0Template(Int64Mask.class, a, offset, indexMap, mapOffset, (Int64Mask) m); 899 } 900 901 902 @ForceInline 903 @Override 904 final 905 void intoMemorySegment0(MemorySegment ms, long offset, VectorMask<Integer> m) { 906 super.intoMemorySegment0Template(Int64Mask.class, ms, offset, (Int64Mask) m); 907 } 908 909 910 // End of specialized low-level memory operations. 911 912 // ================================================ 913 914 }