< prev index next >

src/jdk.incubator.vector/share/classes/jdk/incubator/vector/Double64Vector.java

Print this page

219 
220     @ForceInline
221     final @Override
222     Double64Vector tOp(Vector<Double> v1, Vector<Double> v2, FTriOp f) {
223         return (Double64Vector)
224             super.tOpTemplate((Double64Vector)v1, (Double64Vector)v2,
225                               f);  // specialize
226     }
227 
228     @ForceInline
229     final @Override
230     Double64Vector tOp(Vector<Double> v1, Vector<Double> v2,
231                      VectorMask<Double> m, FTriOp f) {
232         return (Double64Vector)
233             super.tOpTemplate((Double64Vector)v1, (Double64Vector)v2,
234                               (Double64Mask)m, f);  // specialize
235     }
236 
237     @ForceInline
238     final @Override
239     double rOp(double v, FBinOp f) {
240         return super.rOpTemplate(v, f);  // specialize
241     }
242 
243     @Override
244     @ForceInline
245     public final <F>
246     Vector<F> convertShape(VectorOperators.Conversion<Double,F> conv,
247                            VectorSpecies<F> rsp, int part) {
248         return super.convertShapeTemplate(conv, rsp, part);  // specialize
249     }
250 
251     @Override
252     @ForceInline
253     public final <F>
254     Vector<F> reinterpretShape(VectorSpecies<F> toSpecies, int part) {
255         return super.reinterpretShapeTemplate(toSpecies, part);  // specialize
256     }
257 
258     // Specialized algebraic operations:
259 
260     // The following definition forces a specialized version of this
261     // crucial method into the v-table of this class.  A call to add()
262     // will inline to a call to lanewise(ADD,), at which point the JIT
263     // intrinsic will have the opcode of ADD, plus all the metadata
264     // for this particular class, enabling it to generate precise
265     // code.
266     //
267     // There is probably no benefit to the JIT to specialize the
268     // masked or broadcast versions of the lanewise method.
269 
270     @Override
271     @ForceInline
272     public Double64Vector lanewise(Unary op) {
273         return (Double64Vector) super.lanewiseTemplate(op);  // specialize
274     }
275 






276     @Override
277     @ForceInline
278     public Double64Vector lanewise(Binary op, Vector<Double> v) {
279         return (Double64Vector) super.lanewiseTemplate(op, v);  // specialize
280     }
281 






282 
283     /*package-private*/
284     @Override
285     @ForceInline
286     public final
287     Double64Vector
288     lanewise(VectorOperators.Ternary op, Vector<Double> v1, Vector<Double> v2) {
289         return (Double64Vector) super.lanewiseTemplate(op, v1, v2);  // specialize
290     }
291 








292     @Override
293     @ForceInline
294     public final
295     Double64Vector addIndex(int scale) {
296         return (Double64Vector) super.addIndexTemplate(scale);  // specialize
297     }
298 
299     // Type specific horizontal reductions
300 
301     @Override
302     @ForceInline
303     public final double reduceLanes(VectorOperators.Associative op) {
304         return super.reduceLanesTemplate(op);  // specialized
305     }
306 
307     @Override
308     @ForceInline
309     public final double reduceLanes(VectorOperators.Associative op,
310                                     VectorMask<Double> m) {
311         return super.reduceLanesTemplate(op, m);  // specialized
312     }
313 
314     @Override
315     @ForceInline
316     public final long reduceLanesToLong(VectorOperators.Associative op) {
317         return (long) super.reduceLanesTemplate(op);  // specialized
318     }
319 
320     @Override
321     @ForceInline
322     public final long reduceLanesToLong(VectorOperators.Associative op,
323                                         VectorMask<Double> m) {
324         return (long) super.reduceLanesTemplate(op, m);  // specialized
325     }
326 
327     @ForceInline
328     public VectorShuffle<Double> toShuffle() {
329         return super.toShuffleTemplate(Double64Shuffle.class); // specialize
330     }
331 
332     // Specialized unary testing
333 
334     @Override
335     @ForceInline
336     public final Double64Mask test(Test op) {
337         return super.testTemplate(Double64Mask.class, op);  // specialize
338     }
339 
340     // Specialized comparisons
341 
342     @Override
343     @ForceInline
344     public final Double64Mask compare(Comparison op, Vector<Double> v) {
345         return super.compareTemplate(Double64Mask.class, op, v);  // specialize
346     }
347 
348     @Override
349     @ForceInline
350     public final Double64Mask compare(Comparison op, double s) {
351         return super.compareTemplate(Double64Mask.class, op, s);  // specialize
352     }
353 
354     @Override
355     @ForceInline
356     public final Double64Mask compare(Comparison op, long s) {
357         return super.compareTemplate(Double64Mask.class, op, s);  // specialize
358     }
359 







360     @Override
361     @ForceInline
362     public Double64Vector blend(Vector<Double> v, VectorMask<Double> m) {
363         return (Double64Vector)
364             super.blendTemplate(Double64Mask.class,
365                                 (Double64Vector) v,
366                                 (Double64Mask) m);  // specialize
367     }
368 
369     @Override
370     @ForceInline
371     public Double64Vector slice(int origin, Vector<Double> v) {
372         return (Double64Vector) super.sliceTemplate(origin, v);  // specialize
373     }
374 
375     @Override
376     @ForceInline
377     public Double64Vector slice(int origin) {
378         return (Double64Vector) super.sliceTemplate(origin);  // specialize
379     }

396     @Override
397     @ForceInline
398     public Double64Vector unslice(int origin) {
399         return (Double64Vector) super.unsliceTemplate(origin);  // specialize
400     }
401 
402     @Override
403     @ForceInline
404     public Double64Vector rearrange(VectorShuffle<Double> s) {
405         return (Double64Vector)
406             super.rearrangeTemplate(Double64Shuffle.class,
407                                     (Double64Shuffle) s);  // specialize
408     }
409 
410     @Override
411     @ForceInline
412     public Double64Vector rearrange(VectorShuffle<Double> shuffle,
413                                   VectorMask<Double> m) {
414         return (Double64Vector)
415             super.rearrangeTemplate(Double64Shuffle.class,

416                                     (Double64Shuffle) shuffle,
417                                     (Double64Mask) m);  // specialize
418     }
419 
420     @Override
421     @ForceInline
422     public Double64Vector rearrange(VectorShuffle<Double> s,
423                                   Vector<Double> v) {
424         return (Double64Vector)
425             super.rearrangeTemplate(Double64Shuffle.class,
426                                     (Double64Shuffle) s,
427                                     (Double64Vector) v);  // specialize
428     }
429 
430     @Override
431     @ForceInline
432     public Double64Vector selectFrom(Vector<Double> v) {
433         return (Double64Vector)
434             super.selectFromTemplate((Double64Vector) v);  // specialize
435     }

561 
562         /**
563          * Helper function for lane-wise mask conversions.
564          * This function kicks in after intrinsic failure.
565          */
566         @ForceInline
567         private final <E>
568         VectorMask<E> defaultMaskCast(AbstractSpecies<E> dsp) {
569             if (length() != dsp.laneCount())
570                 throw new IllegalArgumentException("VectorMask length and species length differ");
571             boolean[] maskArray = toArray();
572             return  dsp.maskFactory(maskArray).check(dsp);
573         }
574 
575         @Override
576         @ForceInline
577         public <E> VectorMask<E> cast(VectorSpecies<E> dsp) {
578             AbstractSpecies<E> species = (AbstractSpecies<E>) dsp;
579             if (length() != species.laneCount())
580                 throw new IllegalArgumentException("VectorMask length and species length differ");
581             if (VSIZE == species.vectorBitSize()) {
582                 Class<?> dtype = species.elementType();
583                 Class<?> dmtype = species.maskType();
584                 return VectorSupport.convert(VectorSupport.VECTOR_OP_REINTERPRET,
585                     this.getClass(), ETYPE, VLENGTH,
586                     dmtype, dtype, VLENGTH,
587                     this, species,
588                     Double64Mask::defaultMaskCast);
589             }
590             return this.defaultMaskCast(species);
591         }
592 
593         @Override
594         @ForceInline
595         public Double64Mask eq(VectorMask<Double> mask) {
596             Objects.requireNonNull(mask);
597             Double64Mask m = (Double64Mask)mask;
598             return xor(m.not());
599         }
600 
601         // Unary operations
602 
603         @Override
604         @ForceInline
605         public Double64Mask not() {
606             return xor(maskAll(true));
607         }
608 
609         // Binary operations
610 
611         @Override
612         @ForceInline
613         public Double64Mask and(VectorMask<Double> mask) {
614             Objects.requireNonNull(mask);
615             Double64Mask m = (Double64Mask)mask;
616             return VectorSupport.binaryOp(VECTOR_OP_AND, Double64Mask.class, long.class, VLENGTH,
617                                              this, m,
618                                              (m1, m2) -> m1.bOp(m2, (i, a, b) -> a & b));
619         }
620 
621         @Override
622         @ForceInline
623         public Double64Mask or(VectorMask<Double> mask) {
624             Objects.requireNonNull(mask);
625             Double64Mask m = (Double64Mask)mask;
626             return VectorSupport.binaryOp(VECTOR_OP_OR, Double64Mask.class, long.class, VLENGTH,
627                                              this, m,
628                                              (m1, m2) -> m1.bOp(m2, (i, a, b) -> a | b));
629         }
630 
631         @ForceInline
632         /* package-private */
633         Double64Mask xor(VectorMask<Double> mask) {
634             Objects.requireNonNull(mask);
635             Double64Mask m = (Double64Mask)mask;
636             return VectorSupport.binaryOp(VECTOR_OP_XOR, Double64Mask.class, long.class, VLENGTH,
637                                           this, m,
638                                           (m1, m2) -> m1.bOp(m2, (i, a, b) -> a ^ b));
639         }
640 
641         // Mask Query operations
642 
643         @Override
644         @ForceInline
645         public int trueCount() {
646             return VectorSupport.maskReductionCoerced(VECTOR_OP_MASK_TRUECOUNT, Double64Mask.class, long.class, VLENGTH, this,
647                                                       (m) -> trueCountHelper(((Double64Mask)m).getBits()));
648         }
649 
650         @Override
651         @ForceInline
652         public int firstTrue() {
653             return VectorSupport.maskReductionCoerced(VECTOR_OP_MASK_FIRSTTRUE, Double64Mask.class, long.class, VLENGTH, this,
654                                                       (m) -> firstTrueHelper(((Double64Mask)m).getBits()));
655         }
656 
657         @Override
658         @ForceInline
659         public int lastTrue() {
660             return VectorSupport.maskReductionCoerced(VECTOR_OP_MASK_LASTTRUE, Double64Mask.class, long.class, VLENGTH, this,
661                                                       (m) -> lastTrueHelper(((Double64Mask)m).getBits()));










662         }
663 
664         // Reductions
665 
666         @Override
667         @ForceInline
668         public boolean anyTrue() {
669             return VectorSupport.test(BT_ne, Double64Mask.class, long.class, VLENGTH,
670                                          this, vspecies().maskAll(true),
671                                          (m, __) -> anyTrueHelper(((Double64Mask)m).getBits()));
672         }
673 
674         @Override
675         @ForceInline
676         public boolean allTrue() {
677             return VectorSupport.test(BT_overflow, Double64Mask.class, long.class, VLENGTH,
678                                          this, vspecies().maskAll(true),
679                                          (m, __) -> allTrueHelper(((Double64Mask)m).getBits()));
680         }
681 

752             byte[] r = new byte[reorder1.length];
753             for (int i = 0; i < reorder1.length; i++) {
754                 int ssi = reorder2[i];
755                 r[i] = reorder1[ssi];  // throws on exceptional index
756             }
757             return new Double64Shuffle(r);
758         }
759     }
760 
761     // ================================================
762 
763     // Specialized low-level memory operations.
764 
765     @ForceInline
766     @Override
767     final
768     DoubleVector fromArray0(double[] a, int offset) {
769         return super.fromArray0Template(a, offset);  // specialize
770     }
771 














772 
773 
774     @ForceInline
775     @Override
776     final
777     DoubleVector fromByteArray0(byte[] a, int offset) {
778         return super.fromByteArray0Template(a, offset);  // specialize
779     }
780 







781     @ForceInline
782     @Override
783     final
784     DoubleVector fromByteBuffer0(ByteBuffer bb, int offset) {
785         return super.fromByteBuffer0Template(bb, offset);  // specialize
786     }
787 







788     @ForceInline
789     @Override
790     final
791     void intoArray0(double[] a, int offset) {
792         super.intoArray0Template(a, offset);  // specialize
793     }
794 















795     @ForceInline
796     @Override
797     final
798     void intoByteArray0(byte[] a, int offset) {
799         super.intoByteArray0Template(a, offset);  // specialize
800     }
801 















802     // End of specialized low-level memory operations.
803 
804     // ================================================
805 
806 }

219 
220     @ForceInline
221     final @Override
222     Double64Vector tOp(Vector<Double> v1, Vector<Double> v2, FTriOp f) {
223         return (Double64Vector)
224             super.tOpTemplate((Double64Vector)v1, (Double64Vector)v2,
225                               f);  // specialize
226     }
227 
228     @ForceInline
229     final @Override
230     Double64Vector tOp(Vector<Double> v1, Vector<Double> v2,
231                      VectorMask<Double> m, FTriOp f) {
232         return (Double64Vector)
233             super.tOpTemplate((Double64Vector)v1, (Double64Vector)v2,
234                               (Double64Mask)m, f);  // specialize
235     }
236 
237     @ForceInline
238     final @Override
239     double rOp(double v, VectorMask<Double> m, FBinOp f) {
240         return super.rOpTemplate(v, m, f);  // specialize
241     }
242 
243     @Override
244     @ForceInline
245     public final <F>
246     Vector<F> convertShape(VectorOperators.Conversion<Double,F> conv,
247                            VectorSpecies<F> rsp, int part) {
248         return super.convertShapeTemplate(conv, rsp, part);  // specialize
249     }
250 
251     @Override
252     @ForceInline
253     public final <F>
254     Vector<F> reinterpretShape(VectorSpecies<F> toSpecies, int part) {
255         return super.reinterpretShapeTemplate(toSpecies, part);  // specialize
256     }
257 
258     // Specialized algebraic operations:
259 
260     // The following definition forces a specialized version of this
261     // crucial method into the v-table of this class.  A call to add()
262     // will inline to a call to lanewise(ADD,), at which point the JIT
263     // intrinsic will have the opcode of ADD, plus all the metadata
264     // for this particular class, enabling it to generate precise
265     // code.
266     //
267     // There is probably no benefit to the JIT to specialize the
268     // masked or broadcast versions of the lanewise method.
269 
270     @Override
271     @ForceInline
272     public Double64Vector lanewise(Unary op) {
273         return (Double64Vector) super.lanewiseTemplate(op);  // specialize
274     }
275 
276     @Override
277     @ForceInline
278     public Double64Vector lanewise(Unary op, VectorMask<Double> m) {
279         return (Double64Vector) super.lanewiseTemplate(op, Double64Mask.class, (Double64Mask) m);  // specialize
280     }
281 
282     @Override
283     @ForceInline
284     public Double64Vector lanewise(Binary op, Vector<Double> v) {
285         return (Double64Vector) super.lanewiseTemplate(op, v);  // specialize
286     }
287 
288     @Override
289     @ForceInline
290     public Double64Vector lanewise(Binary op, Vector<Double> v, VectorMask<Double> m) {
291         return (Double64Vector) super.lanewiseTemplate(op, Double64Mask.class, v, (Double64Mask) m);  // specialize
292     }
293 
294 
295     /*package-private*/
296     @Override
297     @ForceInline
298     public final
299     Double64Vector
300     lanewise(Ternary op, Vector<Double> v1, Vector<Double> v2) {
301         return (Double64Vector) super.lanewiseTemplate(op, v1, v2);  // specialize
302     }
303 
304     @Override
305     @ForceInline
306     public final
307     Double64Vector
308     lanewise(Ternary op, Vector<Double> v1, Vector<Double> v2, VectorMask<Double> m) {
309         return (Double64Vector) super.lanewiseTemplate(op, Double64Mask.class, v1, v2, (Double64Mask) m);  // specialize
310     }
311 
312     @Override
313     @ForceInline
314     public final
315     Double64Vector addIndex(int scale) {
316         return (Double64Vector) super.addIndexTemplate(scale);  // specialize
317     }
318 
319     // Type specific horizontal reductions
320 
321     @Override
322     @ForceInline
323     public final double reduceLanes(VectorOperators.Associative op) {
324         return super.reduceLanesTemplate(op);  // specialized
325     }
326 
327     @Override
328     @ForceInline
329     public final double reduceLanes(VectorOperators.Associative op,
330                                     VectorMask<Double> m) {
331         return super.reduceLanesTemplate(op, Double64Mask.class, (Double64Mask) m);  // specialized
332     }
333 
334     @Override
335     @ForceInline
336     public final long reduceLanesToLong(VectorOperators.Associative op) {
337         return (long) super.reduceLanesTemplate(op);  // specialized
338     }
339 
340     @Override
341     @ForceInline
342     public final long reduceLanesToLong(VectorOperators.Associative op,
343                                         VectorMask<Double> m) {
344         return (long) super.reduceLanesTemplate(op, Double64Mask.class, (Double64Mask) m);  // specialized
345     }
346 
347     @ForceInline
348     public VectorShuffle<Double> toShuffle() {
349         return super.toShuffleTemplate(Double64Shuffle.class); // specialize
350     }
351 
352     // Specialized unary testing
353 
354     @Override
355     @ForceInline
356     public final Double64Mask test(Test op) {
357         return super.testTemplate(Double64Mask.class, op);  // specialize
358     }
359 
360     // Specialized comparisons
361 
362     @Override
363     @ForceInline
364     public final Double64Mask compare(Comparison op, Vector<Double> v) {
365         return super.compareTemplate(Double64Mask.class, op, v);  // specialize
366     }
367 
368     @Override
369     @ForceInline
370     public final Double64Mask compare(Comparison op, double s) {
371         return super.compareTemplate(Double64Mask.class, op, s);  // specialize
372     }
373 
374     @Override
375     @ForceInline
376     public final Double64Mask compare(Comparison op, long s) {
377         return super.compareTemplate(Double64Mask.class, op, s);  // specialize
378     }
379 
380     @Override
381     @ForceInline
382     public final Double64Mask compare(Comparison op, Vector<Double> v, VectorMask<Double> m) {
383         return super.compareTemplate(Double64Mask.class, op, v, (Double64Mask) m);
384     }
385 
386 
387     @Override
388     @ForceInline
389     public Double64Vector blend(Vector<Double> v, VectorMask<Double> m) {
390         return (Double64Vector)
391             super.blendTemplate(Double64Mask.class,
392                                 (Double64Vector) v,
393                                 (Double64Mask) m);  // specialize
394     }
395 
396     @Override
397     @ForceInline
398     public Double64Vector slice(int origin, Vector<Double> v) {
399         return (Double64Vector) super.sliceTemplate(origin, v);  // specialize
400     }
401 
402     @Override
403     @ForceInline
404     public Double64Vector slice(int origin) {
405         return (Double64Vector) super.sliceTemplate(origin);  // specialize
406     }

423     @Override
424     @ForceInline
425     public Double64Vector unslice(int origin) {
426         return (Double64Vector) super.unsliceTemplate(origin);  // specialize
427     }
428 
429     @Override
430     @ForceInline
431     public Double64Vector rearrange(VectorShuffle<Double> s) {
432         return (Double64Vector)
433             super.rearrangeTemplate(Double64Shuffle.class,
434                                     (Double64Shuffle) s);  // specialize
435     }
436 
437     @Override
438     @ForceInline
439     public Double64Vector rearrange(VectorShuffle<Double> shuffle,
440                                   VectorMask<Double> m) {
441         return (Double64Vector)
442             super.rearrangeTemplate(Double64Shuffle.class,
443                                     Double64Mask.class,
444                                     (Double64Shuffle) shuffle,
445                                     (Double64Mask) m);  // specialize
446     }
447 
448     @Override
449     @ForceInline
450     public Double64Vector rearrange(VectorShuffle<Double> s,
451                                   Vector<Double> v) {
452         return (Double64Vector)
453             super.rearrangeTemplate(Double64Shuffle.class,
454                                     (Double64Shuffle) s,
455                                     (Double64Vector) v);  // specialize
456     }
457 
458     @Override
459     @ForceInline
460     public Double64Vector selectFrom(Vector<Double> v) {
461         return (Double64Vector)
462             super.selectFromTemplate((Double64Vector) v);  // specialize
463     }

589 
590         /**
591          * Helper function for lane-wise mask conversions.
592          * This function kicks in after intrinsic failure.
593          */
594         @ForceInline
595         private final <E>
596         VectorMask<E> defaultMaskCast(AbstractSpecies<E> dsp) {
597             if (length() != dsp.laneCount())
598                 throw new IllegalArgumentException("VectorMask length and species length differ");
599             boolean[] maskArray = toArray();
600             return  dsp.maskFactory(maskArray).check(dsp);
601         }
602 
603         @Override
604         @ForceInline
605         public <E> VectorMask<E> cast(VectorSpecies<E> dsp) {
606             AbstractSpecies<E> species = (AbstractSpecies<E>) dsp;
607             if (length() != species.laneCount())
608                 throw new IllegalArgumentException("VectorMask length and species length differ");
609 
610             return VectorSupport.convert(VectorSupport.VECTOR_OP_CAST,
611                 this.getClass(), ETYPE, VLENGTH,
612                 species.maskType(), species.elementType(), VLENGTH,
613                 this, species,
614                 (m, s) -> s.maskFactory(m.toArray()).check(s));




615         }
616 
617         @Override
618         @ForceInline
619         public Double64Mask eq(VectorMask<Double> mask) {
620             Objects.requireNonNull(mask);
621             Double64Mask m = (Double64Mask)mask;
622             return xor(m.not());
623         }
624 
625         // Unary operations
626 
627         @Override
628         @ForceInline
629         public Double64Mask not() {
630             return xor(maskAll(true));
631         }
632 
633         // Binary operations
634 
635         @Override
636         @ForceInline
637         public Double64Mask and(VectorMask<Double> mask) {
638             Objects.requireNonNull(mask);
639             Double64Mask m = (Double64Mask)mask;
640             return VectorSupport.binaryOp(VECTOR_OP_AND, Double64Mask.class, null, long.class, VLENGTH,
641                                           this, m, null,
642                                           (m1, m2, vm) -> m1.bOp(m2, (i, a, b) -> a & b));
643         }
644 
645         @Override
646         @ForceInline
647         public Double64Mask or(VectorMask<Double> mask) {
648             Objects.requireNonNull(mask);
649             Double64Mask m = (Double64Mask)mask;
650             return VectorSupport.binaryOp(VECTOR_OP_OR, Double64Mask.class, null, long.class, VLENGTH,
651                                           this, m, null,
652                                           (m1, m2, vm) -> m1.bOp(m2, (i, a, b) -> a | b));
653         }
654 
655         @ForceInline
656         /* package-private */
657         Double64Mask xor(VectorMask<Double> mask) {
658             Objects.requireNonNull(mask);
659             Double64Mask m = (Double64Mask)mask;
660             return VectorSupport.binaryOp(VECTOR_OP_XOR, Double64Mask.class, null, long.class, VLENGTH,
661                                           this, m, null,
662                                           (m1, m2, vm) -> m1.bOp(m2, (i, a, b) -> a ^ b));
663         }
664 
665         // Mask Query operations
666 
667         @Override
668         @ForceInline
669         public int trueCount() {
670             return (int) VectorSupport.maskReductionCoerced(VECTOR_OP_MASK_TRUECOUNT, Double64Mask.class, long.class, VLENGTH, this,
671                                                       (m) -> trueCountHelper(m.getBits()));
672         }
673 
674         @Override
675         @ForceInline
676         public int firstTrue() {
677             return (int) VectorSupport.maskReductionCoerced(VECTOR_OP_MASK_FIRSTTRUE, Double64Mask.class, long.class, VLENGTH, this,
678                                                       (m) -> firstTrueHelper(m.getBits()));
679         }
680 
681         @Override
682         @ForceInline
683         public int lastTrue() {
684             return (int) VectorSupport.maskReductionCoerced(VECTOR_OP_MASK_LASTTRUE, Double64Mask.class, long.class, VLENGTH, this,
685                                                       (m) -> lastTrueHelper(m.getBits()));
686         }
687 
688         @Override
689         @ForceInline
690         public long toLong() {
691             if (length() > Long.SIZE) {
692                 throw new UnsupportedOperationException("too many lanes for one long");
693             }
694             return VectorSupport.maskReductionCoerced(VECTOR_OP_MASK_TOLONG, Double64Mask.class, long.class, VLENGTH, this,
695                                                       (m) -> toLongHelper(m.getBits()));
696         }
697 
698         // Reductions
699 
700         @Override
701         @ForceInline
702         public boolean anyTrue() {
703             return VectorSupport.test(BT_ne, Double64Mask.class, long.class, VLENGTH,
704                                          this, vspecies().maskAll(true),
705                                          (m, __) -> anyTrueHelper(((Double64Mask)m).getBits()));
706         }
707 
708         @Override
709         @ForceInline
710         public boolean allTrue() {
711             return VectorSupport.test(BT_overflow, Double64Mask.class, long.class, VLENGTH,
712                                          this, vspecies().maskAll(true),
713                                          (m, __) -> allTrueHelper(((Double64Mask)m).getBits()));
714         }
715 

786             byte[] r = new byte[reorder1.length];
787             for (int i = 0; i < reorder1.length; i++) {
788                 int ssi = reorder2[i];
789                 r[i] = reorder1[ssi];  // throws on exceptional index
790             }
791             return new Double64Shuffle(r);
792         }
793     }
794 
795     // ================================================
796 
797     // Specialized low-level memory operations.
798 
799     @ForceInline
800     @Override
801     final
802     DoubleVector fromArray0(double[] a, int offset) {
803         return super.fromArray0Template(a, offset);  // specialize
804     }
805 
806     @ForceInline
807     @Override
808     final
809     DoubleVector fromArray0(double[] a, int offset, VectorMask<Double> m) {
810         return super.fromArray0Template(Double64Mask.class, a, offset, (Double64Mask) m);  // specialize
811     }
812 
813     @ForceInline
814     @Override
815     final
816     DoubleVector fromArray0(double[] a, int offset, int[] indexMap, int mapOffset, VectorMask<Double> m) {
817         return super.fromArray0Template(Double64Mask.class, a, offset, indexMap, mapOffset, (Double64Mask) m);
818     }
819 
820 
821 
822     @ForceInline
823     @Override
824     final
825     DoubleVector fromByteArray0(byte[] a, int offset) {
826         return super.fromByteArray0Template(a, offset);  // specialize
827     }
828 
829     @ForceInline
830     @Override
831     final
832     DoubleVector fromByteArray0(byte[] a, int offset, VectorMask<Double> m) {
833         return super.fromByteArray0Template(Double64Mask.class, a, offset, (Double64Mask) m);  // specialize
834     }
835 
836     @ForceInline
837     @Override
838     final
839     DoubleVector fromByteBuffer0(ByteBuffer bb, int offset) {
840         return super.fromByteBuffer0Template(bb, offset);  // specialize
841     }
842 
843     @ForceInline
844     @Override
845     final
846     DoubleVector fromByteBuffer0(ByteBuffer bb, int offset, VectorMask<Double> m) {
847         return super.fromByteBuffer0Template(Double64Mask.class, bb, offset, (Double64Mask) m);  // specialize
848     }
849 
850     @ForceInline
851     @Override
852     final
853     void intoArray0(double[] a, int offset) {
854         super.intoArray0Template(a, offset);  // specialize
855     }
856 
857     @ForceInline
858     @Override
859     final
860     void intoArray0(double[] a, int offset, VectorMask<Double> m) {
861         super.intoArray0Template(Double64Mask.class, a, offset, (Double64Mask) m);
862     }
863 
864     @ForceInline
865     @Override
866     final
867     void intoArray0(double[] a, int offset, int[] indexMap, int mapOffset, VectorMask<Double> m) {
868         super.intoArray0Template(Double64Mask.class, a, offset, indexMap, mapOffset, (Double64Mask) m);
869     }
870 
871 
872     @ForceInline
873     @Override
874     final
875     void intoByteArray0(byte[] a, int offset) {
876         super.intoByteArray0Template(a, offset);  // specialize
877     }
878 
879     @ForceInline
880     @Override
881     final
882     void intoByteArray0(byte[] a, int offset, VectorMask<Double> m) {
883         super.intoByteArray0Template(Double64Mask.class, a, offset, (Double64Mask) m);  // specialize
884     }
885 
886     @ForceInline
887     @Override
888     final
889     void intoByteBuffer0(ByteBuffer bb, int offset, VectorMask<Double> m) {
890         super.intoByteBuffer0Template(Double64Mask.class, bb, offset, (Double64Mask) m);
891     }
892 
893 
894     // End of specialized low-level memory operations.
895 
896     // ================================================
897 
898 }
< prev index next >