1 /* 2 * Copyright (c) 2017, 2022, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. Oracle designates this 8 * particular file as subject to the "Classpath" exception as provided 9 * by Oracle in the LICENSE file that accompanied this code. 10 * 11 * This code is distributed in the hope that it will be useful, but WITHOUT 12 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 13 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 14 * version 2 for more details (a copy is included in the LICENSE file that 15 * accompanied this code). 16 * 17 * You should have received a copy of the GNU General Public License version 18 * 2 along with this work; if not, write to the Free Software Foundation, 19 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 20 * 21 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 22 * or visit www.oracle.com if you need additional information or have any 23 * questions. 24 */ 25 package jdk.incubator.vector; 26 27 import java.nio.ByteBuffer; 28 import java.util.Arrays; 29 import java.util.Objects; 30 import java.util.function.IntUnaryOperator; 31 32 import jdk.internal.vm.annotation.ForceInline; 33 import jdk.internal.vm.vector.VectorSupport; 34 35 import static jdk.internal.vm.vector.VectorSupport.*; 36 37 import static jdk.incubator.vector.VectorOperators.*; 38 39 // -- This file was mechanically generated: Do not edit! -- // 40 41 @SuppressWarnings("cast") // warning: redundant cast 42 final class Int64Vector extends IntVector { 43 static final IntSpecies VSPECIES = 44 (IntSpecies) IntVector.SPECIES_64; 45 46 static final VectorShape VSHAPE = 47 VSPECIES.vectorShape(); 48 49 static final Class<Int64Vector> VCLASS = Int64Vector.class; 50 51 static final int VSIZE = VSPECIES.vectorBitSize(); 52 53 static final int VLENGTH = VSPECIES.laneCount(); // used by the JVM 54 55 static final Class<Integer> ETYPE = int.class; // used by the JVM 56 57 Int64Vector(int[] v) { 58 super(v); 59 } 60 61 // For compatibility as Int64Vector::new, 62 // stored into species.vectorFactory. 63 Int64Vector(Object v) { 64 this((int[]) v); 65 } 66 67 static final Int64Vector ZERO = new Int64Vector(new int[VLENGTH]); 68 static final Int64Vector IOTA = new Int64Vector(VSPECIES.iotaArray()); 69 70 static { 71 // Warm up a few species caches. 72 // If we do this too much we will 73 // get NPEs from bootstrap circularity. 74 VSPECIES.dummyVector(); 75 VSPECIES.withLanes(LaneType.BYTE); 76 } 77 78 // Specialized extractors 79 80 @ForceInline 81 final @Override 82 public IntSpecies vspecies() { 83 // ISSUE: This should probably be a @Stable 84 // field inside AbstractVector, rather than 85 // a megamorphic method. 86 return VSPECIES; 87 } 88 89 @ForceInline 90 @Override 91 public final Class<Integer> elementType() { return int.class; } 92 93 @ForceInline 94 @Override 95 public final int elementSize() { return Integer.SIZE; } 96 97 @ForceInline 98 @Override 99 public final VectorShape shape() { return VSHAPE; } 100 101 @ForceInline 102 @Override 103 public final int length() { return VLENGTH; } 104 105 @ForceInline 106 @Override 107 public final int bitSize() { return VSIZE; } 108 109 @ForceInline 110 @Override 111 public final int byteSize() { return VSIZE / Byte.SIZE; } 112 113 /*package-private*/ 114 @ForceInline 115 final @Override 116 int[] vec() { 117 return (int[])getPayload(); 118 } 119 120 // Virtualized constructors 121 122 @Override 123 @ForceInline 124 public final Int64Vector broadcast(int e) { 125 return (Int64Vector) super.broadcastTemplate(e); // specialize 126 } 127 128 @Override 129 @ForceInline 130 public final Int64Vector broadcast(long e) { 131 return (Int64Vector) super.broadcastTemplate(e); // specialize 132 } 133 134 @Override 135 @ForceInline 136 Int64Mask maskFromArray(boolean[] bits) { 137 return new Int64Mask(bits); 138 } 139 140 @Override 141 @ForceInline 142 Int64Shuffle iotaShuffle() { return Int64Shuffle.IOTA; } 143 144 @ForceInline 145 Int64Shuffle iotaShuffle(int start, int step, boolean wrap) { 146 if (wrap) { 147 return (Int64Shuffle)VectorSupport.shuffleIota(ETYPE, Int64Shuffle.class, VSPECIES, VLENGTH, start, step, 1, 148 (l, lstart, lstep, s) -> s.shuffleFromOp(i -> (VectorIntrinsics.wrapToRange(i*lstep + lstart, l)))); 149 } else { 150 return (Int64Shuffle)VectorSupport.shuffleIota(ETYPE, Int64Shuffle.class, VSPECIES, VLENGTH, start, step, 0, 151 (l, lstart, lstep, s) -> s.shuffleFromOp(i -> (i*lstep + lstart))); 152 } 153 } 154 155 @Override 156 @ForceInline 157 Int64Shuffle shuffleFromBytes(byte[] reorder) { return new Int64Shuffle(reorder); } 158 159 @Override 160 @ForceInline 161 Int64Shuffle shuffleFromArray(int[] indexes, int i) { return new Int64Shuffle(indexes, i); } 162 163 @Override 164 @ForceInline 165 Int64Shuffle shuffleFromOp(IntUnaryOperator fn) { return new Int64Shuffle(fn); } 166 167 // Make a vector of the same species but the given elements: 168 @ForceInline 169 final @Override 170 Int64Vector vectorFactory(int[] vec) { 171 return new Int64Vector(vec); 172 } 173 174 @ForceInline 175 final @Override 176 Byte64Vector asByteVectorRaw() { 177 return (Byte64Vector) super.asByteVectorRawTemplate(); // specialize 178 } 179 180 @ForceInline 181 final @Override 182 AbstractVector<?> asVectorRaw(LaneType laneType) { 183 return super.asVectorRawTemplate(laneType); // specialize 184 } 185 186 // Unary operator 187 188 @ForceInline 189 final @Override 190 Int64Vector uOp(FUnOp f) { 191 return (Int64Vector) super.uOpTemplate(f); // specialize 192 } 193 194 @ForceInline 195 final @Override 196 Int64Vector uOp(VectorMask<Integer> m, FUnOp f) { 197 return (Int64Vector) 198 super.uOpTemplate((Int64Mask)m, f); // specialize 199 } 200 201 // Binary operator 202 203 @ForceInline 204 final @Override 205 Int64Vector bOp(Vector<Integer> v, FBinOp f) { 206 return (Int64Vector) super.bOpTemplate((Int64Vector)v, f); // specialize 207 } 208 209 @ForceInline 210 final @Override 211 Int64Vector bOp(Vector<Integer> v, 212 VectorMask<Integer> m, FBinOp f) { 213 return (Int64Vector) 214 super.bOpTemplate((Int64Vector)v, (Int64Mask)m, 215 f); // specialize 216 } 217 218 // Ternary operator 219 220 @ForceInline 221 final @Override 222 Int64Vector tOp(Vector<Integer> v1, Vector<Integer> v2, FTriOp f) { 223 return (Int64Vector) 224 super.tOpTemplate((Int64Vector)v1, (Int64Vector)v2, 225 f); // specialize 226 } 227 228 @ForceInline 229 final @Override 230 Int64Vector tOp(Vector<Integer> v1, Vector<Integer> v2, 231 VectorMask<Integer> m, FTriOp f) { 232 return (Int64Vector) 233 super.tOpTemplate((Int64Vector)v1, (Int64Vector)v2, 234 (Int64Mask)m, f); // specialize 235 } 236 237 @ForceInline 238 final @Override 239 int rOp(int v, VectorMask<Integer> m, FBinOp f) { 240 return super.rOpTemplate(v, m, f); // specialize 241 } 242 243 @Override 244 @ForceInline 245 public final <F> 246 Vector<F> convertShape(VectorOperators.Conversion<Integer,F> conv, 247 VectorSpecies<F> rsp, int part) { 248 return super.convertShapeTemplate(conv, rsp, part); // specialize 249 } 250 251 @Override 252 @ForceInline 253 public final <F> 254 Vector<F> reinterpretShape(VectorSpecies<F> toSpecies, int part) { 255 return super.reinterpretShapeTemplate(toSpecies, part); // specialize 256 } 257 258 // Specialized algebraic operations: 259 260 // The following definition forces a specialized version of this 261 // crucial method into the v-table of this class. A call to add() 262 // will inline to a call to lanewise(ADD,), at which point the JIT 263 // intrinsic will have the opcode of ADD, plus all the metadata 264 // for this particular class, enabling it to generate precise 265 // code. 266 // 267 // There is probably no benefit to the JIT to specialize the 268 // masked or broadcast versions of the lanewise method. 269 270 @Override 271 @ForceInline 272 public Int64Vector lanewise(Unary op) { 273 return (Int64Vector) super.lanewiseTemplate(op); // specialize 274 } 275 276 @Override 277 @ForceInline 278 public Int64Vector lanewise(Unary op, VectorMask<Integer> m) { 279 return (Int64Vector) super.lanewiseTemplate(op, Int64Mask.class, (Int64Mask) m); // specialize 280 } 281 282 @Override 283 @ForceInline 284 public Int64Vector lanewise(Binary op, Vector<Integer> v) { 285 return (Int64Vector) super.lanewiseTemplate(op, v); // specialize 286 } 287 288 @Override 289 @ForceInline 290 public Int64Vector lanewise(Binary op, Vector<Integer> v, VectorMask<Integer> m) { 291 return (Int64Vector) super.lanewiseTemplate(op, Int64Mask.class, v, (Int64Mask) m); // specialize 292 } 293 294 /*package-private*/ 295 @Override 296 @ForceInline Int64Vector 297 lanewiseShift(VectorOperators.Binary op, int e) { 298 return (Int64Vector) super.lanewiseShiftTemplate(op, e); // specialize 299 } 300 301 /*package-private*/ 302 @Override 303 @ForceInline Int64Vector 304 lanewiseShift(VectorOperators.Binary op, int e, VectorMask<Integer> m) { 305 return (Int64Vector) super.lanewiseShiftTemplate(op, Int64Mask.class, e, (Int64Mask) m); // specialize 306 } 307 308 /*package-private*/ 309 @Override 310 @ForceInline 311 public final 312 Int64Vector 313 lanewise(Ternary op, Vector<Integer> v1, Vector<Integer> v2) { 314 return (Int64Vector) super.lanewiseTemplate(op, v1, v2); // specialize 315 } 316 317 @Override 318 @ForceInline 319 public final 320 Int64Vector 321 lanewise(Ternary op, Vector<Integer> v1, Vector<Integer> v2, VectorMask<Integer> m) { 322 return (Int64Vector) super.lanewiseTemplate(op, Int64Mask.class, v1, v2, (Int64Mask) m); // specialize 323 } 324 325 @Override 326 @ForceInline 327 public final 328 Int64Vector addIndex(int scale) { 329 return (Int64Vector) super.addIndexTemplate(scale); // specialize 330 } 331 332 // Type specific horizontal reductions 333 334 @Override 335 @ForceInline 336 public final int reduceLanes(VectorOperators.Associative op) { 337 return super.reduceLanesTemplate(op); // specialized 338 } 339 340 @Override 341 @ForceInline 342 public final int reduceLanes(VectorOperators.Associative op, 343 VectorMask<Integer> m) { 344 return super.reduceLanesTemplate(op, Int64Mask.class, (Int64Mask) m); // specialized 345 } 346 347 @Override 348 @ForceInline 349 public final long reduceLanesToLong(VectorOperators.Associative op) { 350 return (long) super.reduceLanesTemplate(op); // specialized 351 } 352 353 @Override 354 @ForceInline 355 public final long reduceLanesToLong(VectorOperators.Associative op, 356 VectorMask<Integer> m) { 357 return (long) super.reduceLanesTemplate(op, Int64Mask.class, (Int64Mask) m); // specialized 358 } 359 360 @ForceInline 361 public VectorShuffle<Integer> toShuffle() { 362 return super.toShuffleTemplate(Int64Shuffle.class); // specialize 363 } 364 365 // Specialized unary testing 366 367 @Override 368 @ForceInline 369 public final Int64Mask test(Test op) { 370 return super.testTemplate(Int64Mask.class, op); // specialize 371 } 372 373 @Override 374 @ForceInline 375 public final Int64Mask test(Test op, VectorMask<Integer> m) { 376 return super.testTemplate(Int64Mask.class, op, (Int64Mask) m); // specialize 377 } 378 379 // Specialized comparisons 380 381 @Override 382 @ForceInline 383 public final Int64Mask compare(Comparison op, Vector<Integer> v) { 384 return super.compareTemplate(Int64Mask.class, op, v); // specialize 385 } 386 387 @Override 388 @ForceInline 389 public final Int64Mask compare(Comparison op, int s) { 390 return super.compareTemplate(Int64Mask.class, op, s); // specialize 391 } 392 393 @Override 394 @ForceInline 395 public final Int64Mask compare(Comparison op, long s) { 396 return super.compareTemplate(Int64Mask.class, op, s); // specialize 397 } 398 399 @Override 400 @ForceInline 401 public final Int64Mask compare(Comparison op, Vector<Integer> v, VectorMask<Integer> m) { 402 return super.compareTemplate(Int64Mask.class, op, v, (Int64Mask) m); 403 } 404 405 406 @Override 407 @ForceInline 408 public Int64Vector blend(Vector<Integer> v, VectorMask<Integer> m) { 409 return (Int64Vector) 410 super.blendTemplate(Int64Mask.class, 411 (Int64Vector) v, 412 (Int64Mask) m); // specialize 413 } 414 415 @Override 416 @ForceInline 417 public Int64Vector slice(int origin, Vector<Integer> v) { 418 return (Int64Vector) super.sliceTemplate(origin, v); // specialize 419 } 420 421 @Override 422 @ForceInline 423 public Int64Vector slice(int origin) { 424 return (Int64Vector) super.sliceTemplate(origin); // specialize 425 } 426 427 @Override 428 @ForceInline 429 public Int64Vector unslice(int origin, Vector<Integer> w, int part) { 430 return (Int64Vector) super.unsliceTemplate(origin, w, part); // specialize 431 } 432 433 @Override 434 @ForceInline 435 public Int64Vector unslice(int origin, Vector<Integer> w, int part, VectorMask<Integer> m) { 436 return (Int64Vector) 437 super.unsliceTemplate(Int64Mask.class, 438 origin, w, part, 439 (Int64Mask) m); // specialize 440 } 441 442 @Override 443 @ForceInline 444 public Int64Vector unslice(int origin) { 445 return (Int64Vector) super.unsliceTemplate(origin); // specialize 446 } 447 448 @Override 449 @ForceInline 450 public Int64Vector rearrange(VectorShuffle<Integer> s) { 451 return (Int64Vector) 452 super.rearrangeTemplate(Int64Shuffle.class, 453 (Int64Shuffle) s); // specialize 454 } 455 456 @Override 457 @ForceInline 458 public Int64Vector rearrange(VectorShuffle<Integer> shuffle, 459 VectorMask<Integer> m) { 460 return (Int64Vector) 461 super.rearrangeTemplate(Int64Shuffle.class, 462 Int64Mask.class, 463 (Int64Shuffle) shuffle, 464 (Int64Mask) m); // specialize 465 } 466 467 @Override 468 @ForceInline 469 public Int64Vector rearrange(VectorShuffle<Integer> s, 470 Vector<Integer> v) { 471 return (Int64Vector) 472 super.rearrangeTemplate(Int64Shuffle.class, 473 (Int64Shuffle) s, 474 (Int64Vector) v); // specialize 475 } 476 477 @Override 478 @ForceInline 479 public Int64Vector selectFrom(Vector<Integer> v) { 480 return (Int64Vector) 481 super.selectFromTemplate((Int64Vector) v); // specialize 482 } 483 484 @Override 485 @ForceInline 486 public Int64Vector selectFrom(Vector<Integer> v, 487 VectorMask<Integer> m) { 488 return (Int64Vector) 489 super.selectFromTemplate((Int64Vector) v, 490 (Int64Mask) m); // specialize 491 } 492 493 494 @ForceInline 495 @Override 496 public int lane(int i) { 497 switch(i) { 498 case 0: return laneHelper(0); 499 case 1: return laneHelper(1); 500 default: throw new IllegalArgumentException("Index " + i + " must be zero or positive, and less than " + VLENGTH); 501 } 502 } 503 504 public int laneHelper(int i) { 505 return (int) VectorSupport.extract( 506 VCLASS, ETYPE, VLENGTH, 507 this, i, 508 (vec, ix) -> { 509 int[] vecarr = vec.vec(); 510 return (long)vecarr[ix]; 511 }); 512 } 513 514 @ForceInline 515 @Override 516 public Int64Vector withLane(int i, int e) { 517 switch (i) { 518 case 0: return withLaneHelper(0, e); 519 case 1: return withLaneHelper(1, e); 520 default: throw new IllegalArgumentException("Index " + i + " must be zero or positive, and less than " + VLENGTH); 521 } 522 } 523 524 public Int64Vector withLaneHelper(int i, int e) { 525 return VectorSupport.insert( 526 VCLASS, ETYPE, VLENGTH, 527 this, i, (long)e, 528 (v, ix, bits) -> { 529 int[] res = v.vec().clone(); 530 res[ix] = (int)bits; 531 return v.vectorFactory(res); 532 }); 533 } 534 535 // Mask 536 537 static final class Int64Mask extends AbstractMask<Integer> { 538 static final int VLENGTH = VSPECIES.laneCount(); // used by the JVM 539 static final Class<Integer> ETYPE = int.class; // used by the JVM 540 541 Int64Mask(boolean[] bits) { 542 this(bits, 0); 543 } 544 545 Int64Mask(boolean[] bits, int offset) { 546 super(prepare(bits, offset)); 547 } 548 549 Int64Mask(boolean val) { 550 super(prepare(val)); 551 } 552 553 private static boolean[] prepare(boolean[] bits, int offset) { 554 boolean[] newBits = new boolean[VSPECIES.laneCount()]; 555 for (int i = 0; i < newBits.length; i++) { 556 newBits[i] = bits[offset + i]; 557 } 558 return newBits; 559 } 560 561 private static boolean[] prepare(boolean val) { 562 boolean[] bits = new boolean[VSPECIES.laneCount()]; 563 Arrays.fill(bits, val); 564 return bits; 565 } 566 567 @ForceInline 568 final @Override 569 public IntSpecies vspecies() { 570 // ISSUE: This should probably be a @Stable 571 // field inside AbstractMask, rather than 572 // a megamorphic method. 573 return VSPECIES; 574 } 575 576 @ForceInline 577 boolean[] getBits() { 578 return (boolean[])getPayload(); 579 } 580 581 @Override 582 Int64Mask uOp(MUnOp f) { 583 boolean[] res = new boolean[vspecies().laneCount()]; 584 boolean[] bits = getBits(); 585 for (int i = 0; i < res.length; i++) { 586 res[i] = f.apply(i, bits[i]); 587 } 588 return new Int64Mask(res); 589 } 590 591 @Override 592 Int64Mask bOp(VectorMask<Integer> m, MBinOp f) { 593 boolean[] res = new boolean[vspecies().laneCount()]; 594 boolean[] bits = getBits(); 595 boolean[] mbits = ((Int64Mask)m).getBits(); 596 for (int i = 0; i < res.length; i++) { 597 res[i] = f.apply(i, bits[i], mbits[i]); 598 } 599 return new Int64Mask(res); 600 } 601 602 @ForceInline 603 @Override 604 public final 605 Int64Vector toVector() { 606 return (Int64Vector) super.toVectorTemplate(); // specialize 607 } 608 609 /** 610 * Helper function for lane-wise mask conversions. 611 * This function kicks in after intrinsic failure. 612 */ 613 @ForceInline 614 private final <E> 615 VectorMask<E> defaultMaskCast(AbstractSpecies<E> dsp) { 616 if (length() != dsp.laneCount()) 617 throw new IllegalArgumentException("VectorMask length and species length differ"); 618 boolean[] maskArray = toArray(); 619 return dsp.maskFactory(maskArray).check(dsp); 620 } 621 622 @Override 623 @ForceInline 624 public <E> VectorMask<E> cast(VectorSpecies<E> dsp) { 625 AbstractSpecies<E> species = (AbstractSpecies<E>) dsp; 626 if (length() != species.laneCount()) 627 throw new IllegalArgumentException("VectorMask length and species length differ"); 628 629 return VectorSupport.convert(VectorSupport.VECTOR_OP_CAST, 630 this.getClass(), ETYPE, VLENGTH, 631 species.maskType(), species.elementType(), VLENGTH, 632 this, species, 633 (m, s) -> s.maskFactory(m.toArray()).check(s)); 634 } 635 636 @Override 637 @ForceInline 638 public Int64Mask eq(VectorMask<Integer> mask) { 639 Objects.requireNonNull(mask); 640 Int64Mask m = (Int64Mask)mask; 641 return xor(m.not()); 642 } 643 644 // Unary operations 645 646 @Override 647 @ForceInline 648 public Int64Mask not() { 649 return xor(maskAll(true)); 650 } 651 652 // Binary operations 653 654 @Override 655 @ForceInline 656 public Int64Mask and(VectorMask<Integer> mask) { 657 Objects.requireNonNull(mask); 658 Int64Mask m = (Int64Mask)mask; 659 return VectorSupport.binaryOp(VECTOR_OP_AND, Int64Mask.class, null, int.class, VLENGTH, 660 this, m, null, 661 (m1, m2, vm) -> m1.bOp(m2, (i, a, b) -> a & b)); 662 } 663 664 @Override 665 @ForceInline 666 public Int64Mask or(VectorMask<Integer> mask) { 667 Objects.requireNonNull(mask); 668 Int64Mask m = (Int64Mask)mask; 669 return VectorSupport.binaryOp(VECTOR_OP_OR, Int64Mask.class, null, int.class, VLENGTH, 670 this, m, null, 671 (m1, m2, vm) -> m1.bOp(m2, (i, a, b) -> a | b)); 672 } 673 674 @ForceInline 675 /* package-private */ 676 Int64Mask xor(VectorMask<Integer> mask) { 677 Objects.requireNonNull(mask); 678 Int64Mask m = (Int64Mask)mask; 679 return VectorSupport.binaryOp(VECTOR_OP_XOR, Int64Mask.class, null, int.class, VLENGTH, 680 this, m, null, 681 (m1, m2, vm) -> m1.bOp(m2, (i, a, b) -> a ^ b)); 682 } 683 684 // Mask Query operations 685 686 @Override 687 @ForceInline 688 public int trueCount() { 689 return (int) VectorSupport.maskReductionCoerced(VECTOR_OP_MASK_TRUECOUNT, Int64Mask.class, int.class, VLENGTH, this, 690 (m) -> trueCountHelper(m.getBits())); 691 } 692 693 @Override 694 @ForceInline 695 public int firstTrue() { 696 return (int) VectorSupport.maskReductionCoerced(VECTOR_OP_MASK_FIRSTTRUE, Int64Mask.class, int.class, VLENGTH, this, 697 (m) -> firstTrueHelper(m.getBits())); 698 } 699 700 @Override 701 @ForceInline 702 public int lastTrue() { 703 return (int) VectorSupport.maskReductionCoerced(VECTOR_OP_MASK_LASTTRUE, Int64Mask.class, int.class, VLENGTH, this, 704 (m) -> lastTrueHelper(m.getBits())); 705 } 706 707 @Override 708 @ForceInline 709 public long toLong() { 710 if (length() > Long.SIZE) { 711 throw new UnsupportedOperationException("too many lanes for one long"); 712 } 713 return VectorSupport.maskReductionCoerced(VECTOR_OP_MASK_TOLONG, Int64Mask.class, int.class, VLENGTH, this, 714 (m) -> toLongHelper(m.getBits())); 715 } 716 717 // Reductions 718 719 @Override 720 @ForceInline 721 public boolean anyTrue() { 722 return VectorSupport.test(BT_ne, Int64Mask.class, int.class, VLENGTH, 723 this, vspecies().maskAll(true), 724 (m, __) -> anyTrueHelper(((Int64Mask)m).getBits())); 725 } 726 727 @Override 728 @ForceInline 729 public boolean allTrue() { 730 return VectorSupport.test(BT_overflow, Int64Mask.class, int.class, VLENGTH, 731 this, vspecies().maskAll(true), 732 (m, __) -> allTrueHelper(((Int64Mask)m).getBits())); 733 } 734 735 @ForceInline 736 /*package-private*/ 737 static Int64Mask maskAll(boolean bit) { 738 return VectorSupport.fromBitsCoerced(Int64Mask.class, int.class, VLENGTH, 739 (bit ? -1 : 0), MODE_BROADCAST, null, 740 (v, __) -> (v != 0 ? TRUE_MASK : FALSE_MASK)); 741 } 742 private static final Int64Mask TRUE_MASK = new Int64Mask(true); 743 private static final Int64Mask FALSE_MASK = new Int64Mask(false); 744 745 } 746 747 // Shuffle 748 749 static final class Int64Shuffle extends AbstractShuffle<Integer> { 750 static final int VLENGTH = VSPECIES.laneCount(); // used by the JVM 751 static final Class<Integer> ETYPE = int.class; // used by the JVM 752 753 Int64Shuffle(byte[] reorder) { 754 super(VLENGTH, reorder); 755 } 756 757 public Int64Shuffle(int[] reorder) { 758 super(VLENGTH, reorder); 759 } 760 761 public Int64Shuffle(int[] reorder, int i) { 762 super(VLENGTH, reorder, i); 763 } 764 765 public Int64Shuffle(IntUnaryOperator fn) { 766 super(VLENGTH, fn); 767 } 768 769 @Override 770 public IntSpecies vspecies() { 771 return VSPECIES; 772 } 773 774 static { 775 // There must be enough bits in the shuffle lanes to encode 776 // VLENGTH valid indexes and VLENGTH exceptional ones. 777 assert(VLENGTH < Byte.MAX_VALUE); 778 assert(Byte.MIN_VALUE <= -VLENGTH); 779 } 780 static final Int64Shuffle IOTA = new Int64Shuffle(IDENTITY); 781 782 @Override 783 @ForceInline 784 public Int64Vector toVector() { 785 return VectorSupport.shuffleToVector(VCLASS, ETYPE, Int64Shuffle.class, this, VLENGTH, 786 (s) -> ((Int64Vector)(((AbstractShuffle<Integer>)(s)).toVectorTemplate()))); 787 } 788 789 @Override 790 @ForceInline 791 public <F> VectorShuffle<F> cast(VectorSpecies<F> s) { 792 AbstractSpecies<F> species = (AbstractSpecies<F>) s; 793 if (length() != species.laneCount()) 794 throw new IllegalArgumentException("VectorShuffle length and species length differ"); 795 int[] shuffleArray = toArray(); 796 return s.shuffleFromArray(shuffleArray, 0).check(s); 797 } 798 799 @ForceInline 800 @Override 801 public Int64Shuffle rearrange(VectorShuffle<Integer> shuffle) { 802 Int64Shuffle s = (Int64Shuffle) shuffle; 803 byte[] reorder1 = reorder(); 804 byte[] reorder2 = s.reorder(); 805 byte[] r = new byte[reorder1.length]; 806 for (int i = 0; i < reorder1.length; i++) { 807 int ssi = reorder2[i]; 808 r[i] = reorder1[ssi]; // throws on exceptional index 809 } 810 return new Int64Shuffle(r); 811 } 812 } 813 814 // ================================================ 815 816 // Specialized low-level memory operations. 817 818 @ForceInline 819 @Override 820 final 821 IntVector fromArray0(int[] a, int offset) { 822 return super.fromArray0Template(a, offset); // specialize 823 } 824 825 @ForceInline 826 @Override 827 final 828 IntVector fromArray0(int[] a, int offset, VectorMask<Integer> m) { 829 return super.fromArray0Template(Int64Mask.class, a, offset, (Int64Mask) m); // specialize 830 } 831 832 @ForceInline 833 @Override 834 final 835 IntVector fromArray0(int[] a, int offset, int[] indexMap, int mapOffset, VectorMask<Integer> m) { 836 return super.fromArray0Template(Int64Mask.class, a, offset, indexMap, mapOffset, (Int64Mask) m); 837 } 838 839 840 841 @ForceInline 842 @Override 843 final 844 IntVector fromByteArray0(byte[] a, int offset) { 845 return super.fromByteArray0Template(a, offset); // specialize 846 } 847 848 @ForceInline 849 @Override 850 final 851 IntVector fromByteArray0(byte[] a, int offset, VectorMask<Integer> m) { 852 return super.fromByteArray0Template(Int64Mask.class, a, offset, (Int64Mask) m); // specialize 853 } 854 855 @ForceInline 856 @Override 857 final 858 IntVector fromByteBuffer0(ByteBuffer bb, int offset) { 859 return super.fromByteBuffer0Template(bb, offset); // specialize 860 } 861 862 @ForceInline 863 @Override 864 final 865 IntVector fromByteBuffer0(ByteBuffer bb, int offset, VectorMask<Integer> m) { 866 return super.fromByteBuffer0Template(Int64Mask.class, bb, offset, (Int64Mask) m); // specialize 867 } 868 869 @ForceInline 870 @Override 871 final 872 void intoArray0(int[] a, int offset) { 873 super.intoArray0Template(a, offset); // specialize 874 } 875 876 @ForceInline 877 @Override 878 final 879 void intoArray0(int[] a, int offset, VectorMask<Integer> m) { 880 super.intoArray0Template(Int64Mask.class, a, offset, (Int64Mask) m); 881 } 882 883 @ForceInline 884 @Override 885 final 886 void intoArray0(int[] a, int offset, int[] indexMap, int mapOffset, VectorMask<Integer> m) { 887 super.intoArray0Template(Int64Mask.class, a, offset, indexMap, mapOffset, (Int64Mask) m); 888 } 889 890 891 @ForceInline 892 @Override 893 final 894 void intoByteArray0(byte[] a, int offset) { 895 super.intoByteArray0Template(a, offset); // specialize 896 } 897 898 @ForceInline 899 @Override 900 final 901 void intoByteArray0(byte[] a, int offset, VectorMask<Integer> m) { 902 super.intoByteArray0Template(Int64Mask.class, a, offset, (Int64Mask) m); // specialize 903 } 904 905 @ForceInline 906 @Override 907 final 908 void intoByteBuffer0(ByteBuffer bb, int offset, VectorMask<Integer> m) { 909 super.intoByteBuffer0Template(Int64Mask.class, bb, offset, (Int64Mask) m); 910 } 911 912 913 // End of specialized low-level memory operations. 914 915 // ================================================ 916 917 } --- EOF ---