1 /*
2 * Copyright (c) 1997, 2026, Oracle and/or its affiliates. All rights reserved.
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
4 *
5 * This code is free software; you can redistribute it and/or modify it
6 * under the terms of the GNU General Public License version 2 only, as
7 * published by the Free Software Foundation.
8 *
9 * This code is distributed in the hope that it will be useful, but WITHOUT
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
12 * version 2 for more details (a copy is included in the LICENSE file that
13 * accompanied this code).
14 *
15 * You should have received a copy of the GNU General Public License version
16 * 2 along with this work; if not, write to the Free Software Foundation,
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
18 *
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
20 * or visit www.oracle.com if you need additional information or have any
21 * questions.
22 *
23 */
24
25 #include "memory/allocation.inline.hpp"
26 #include "opto/addnode.hpp"
27 #include "opto/connode.hpp"
28 #include "opto/convertnode.hpp"
29 #include "opto/memnode.hpp"
30 #include "opto/mulnode.hpp"
31 #include "opto/phaseX.hpp"
32 #include "opto/rangeinference.hpp"
33 #include "opto/subnode.hpp"
34 #include "utilities/powerOfTwo.hpp"
35
36 // Portions of code courtesy of Clifford Click
37
38
39 //=============================================================================
40 //------------------------------hash-------------------------------------------
41 // Hash function over MulNodes. Needs to be commutative; i.e., I swap
42 // (commute) inputs to MulNodes willy-nilly so the hash function must return
43 // the same value in the presence of edge swapping.
44 uint MulNode::hash() const {
45 return (uintptr_t)in(1) + (uintptr_t)in(2) + Opcode();
46 }
47
48 //------------------------------Identity---------------------------------------
49 // Multiplying a one preserves the other argument
50 Node* MulNode::Identity(PhaseGVN* phase) {
51 const Type *one = mul_id(); // The multiplicative identity
52 if( phase->type( in(1) )->higher_equal( one ) ) return in(2);
53 if( phase->type( in(2) )->higher_equal( one ) ) return in(1);
54
55 return this;
56 }
57
58 //------------------------------Ideal------------------------------------------
59 // We also canonicalize the Node, moving constants to the right input,
60 // and flatten expressions (so that 1+x+2 becomes x+3).
61 Node *MulNode::Ideal(PhaseGVN *phase, bool can_reshape) {
62 Node* in1 = in(1);
63 Node* in2 = in(2);
64 Node* progress = nullptr; // Progress flag
65
66 // This code is used by And nodes too, but some conversions are
67 // only valid for the actual Mul nodes.
68 uint op = Opcode();
69 bool real_mul = (op == Op_MulI) || (op == Op_MulL) ||
70 (op == Op_MulF) || (op == Op_MulD) ||
71 (op == Op_MulHF);
72
73 // Convert "(-a)*(-b)" into "a*b".
74 if (real_mul && in1->is_Sub() && in2->is_Sub()) {
75 if (phase->type(in1->in(1))->is_zero_type() &&
76 phase->type(in2->in(1))->is_zero_type()) {
77 set_req_X(1, in1->in(2), phase);
78 set_req_X(2, in2->in(2), phase);
79 in1 = in(1);
80 in2 = in(2);
81 progress = this;
82 }
83 }
84
85 // convert "max(a,b) * min(a,b)" into "a*b".
86 if ((in(1)->Opcode() == max_opcode() && in(2)->Opcode() == min_opcode())
87 || (in(1)->Opcode() == min_opcode() && in(2)->Opcode() == max_opcode())) {
88 Node *in11 = in(1)->in(1);
89 Node *in12 = in(1)->in(2);
90
91 Node *in21 = in(2)->in(1);
92 Node *in22 = in(2)->in(2);
93
94 if ((in11 == in21 && in12 == in22) ||
95 (in11 == in22 && in12 == in21)) {
96 set_req_X(1, in11, phase);
97 set_req_X(2, in12, phase);
98 in1 = in(1);
99 in2 = in(2);
100 progress = this;
101 }
102 }
103
104 const Type* t1 = phase->type(in1);
105 const Type* t2 = phase->type(in2);
106
107 // We are OK if right is a constant, or right is a load and
108 // left is a non-constant.
109 if( !(t2->singleton() ||
110 (in(2)->is_Load() && !(t1->singleton() || in(1)->is_Load())) ) ) {
111 if( t1->singleton() || // Left input is a constant?
112 // Otherwise, sort inputs (commutativity) to help value numbering.
113 (in(1)->_idx > in(2)->_idx) ) {
114 swap_edges(1, 2);
115 const Type *t = t1;
116 t1 = t2;
117 t2 = t;
118 progress = this; // Made progress
119 }
120 }
121
122 // If the right input is a constant, and the left input is a product of a
123 // constant, flatten the expression tree.
124 if( t2->singleton() && // Right input is a constant?
125 op != Op_MulF && // Float & double cannot reassociate
126 op != Op_MulD &&
127 op != Op_MulHF) {
128 if( t2 == Type::TOP ) return nullptr;
129 Node *mul1 = in(1);
130 #ifdef ASSERT
131 // Check for dead loop
132 int op1 = mul1->Opcode();
133 if ((mul1 == this) || (in(2) == this) ||
134 ((op1 == mul_opcode() || op1 == add_opcode()) &&
135 ((mul1->in(1) == this) || (mul1->in(2) == this) ||
136 (mul1->in(1) == mul1) || (mul1->in(2) == mul1)))) {
137 assert(false, "dead loop in MulNode::Ideal");
138 }
139 #endif
140
141 if( mul1->Opcode() == mul_opcode() ) { // Left input is a multiply?
142 // Mul of a constant?
143 const Type *t12 = phase->type( mul1->in(2) );
144 if( t12->singleton() && t12 != Type::TOP) { // Left input is an add of a constant?
145 // Compute new constant; check for overflow
146 const Type *tcon01 = ((MulNode*)mul1)->mul_ring(t2,t12);
147 if( tcon01->singleton() ) {
148 // The Mul of the flattened expression
149 set_req_X(1, mul1->in(1), phase);
150 set_req_X(2, phase->makecon(tcon01), phase);
151 t2 = tcon01;
152 progress = this; // Made progress
153 }
154 }
155 }
156 // If the right input is a constant, and the left input is an add of a
157 // constant, flatten the tree: (X+con1)*con0 ==> X*con0 + con1*con0
158 const Node *add1 = in(1);
159 if( add1->Opcode() == add_opcode() ) { // Left input is an add?
160 // Add of a constant?
161 const Type *t12 = phase->type( add1->in(2) );
162 if( t12->singleton() && t12 != Type::TOP ) { // Left input is an add of a constant?
163 assert( add1->in(1) != add1, "dead loop in MulNode::Ideal" );
164 // Compute new constant; check for overflow
165 const Type *tcon01 = mul_ring(t2,t12);
166 if( tcon01->singleton() ) {
167
168 // Convert (X+con1)*con0 into X*con0
169 Node *mul = clone(); // mul = ()*con0
170 mul->set_req(1,add1->in(1)); // mul = X*con0
171 mul = phase->transform(mul);
172
173 Node *add2 = add1->clone();
174 add2->set_req(1, mul); // X*con0 + con0*con1
175 add2->set_req(2, phase->makecon(tcon01) );
176 progress = add2;
177 }
178 }
179 } // End of is left input an add
180 } // End of is right input a Mul
181
182 return progress;
183 }
184
185 //------------------------------Value-----------------------------------------
186 const Type* MulNode::Value(PhaseGVN* phase) const {
187 const Type *t1 = phase->type( in(1) );
188 const Type *t2 = phase->type( in(2) );
189 // Either input is TOP ==> the result is TOP
190 if( t1 == Type::TOP ) return Type::TOP;
191 if( t2 == Type::TOP ) return Type::TOP;
192
193 // Either input is ZERO ==> the result is ZERO.
194 // Not valid for floats or doubles since +0.0 * -0.0 --> +0.0
195 int op = Opcode();
196 if( op == Op_MulI || op == Op_AndI || op == Op_MulL || op == Op_AndL ) {
197 const Type *zero = add_id(); // The multiplicative zero
198 if( t1->higher_equal( zero ) ) return zero;
199 if( t2->higher_equal( zero ) ) return zero;
200 }
201
202 // Either input is BOTTOM ==> the result is the local BOTTOM
203 if( t1 == Type::BOTTOM || t2 == Type::BOTTOM )
204 return bottom_type();
205
206 return mul_ring(t1,t2); // Local flavor of type multiplication
207 }
208
209 MulNode* MulNode::make(Node* in1, Node* in2, BasicType bt) {
210 switch (bt) {
211 case T_INT:
212 return new MulINode(in1, in2);
213 case T_LONG:
214 return new MulLNode(in1, in2);
215 default:
216 fatal("Not implemented for %s", type2name(bt));
217 }
218 return nullptr;
219 }
220
221 MulNode* MulNode::make_and(Node* in1, Node* in2, BasicType bt) {
222 switch (bt) {
223 case T_INT:
224 return new AndINode(in1, in2);
225 case T_LONG:
226 return new AndLNode(in1, in2);
227 default:
228 fatal("Not implemented for %s", type2name(bt));
229 }
230 return nullptr;
231 }
232
233
234 //=============================================================================
235 //------------------------------Ideal------------------------------------------
236 // Check for power-of-2 multiply, then try the regular MulNode::Ideal
237 Node *MulINode::Ideal(PhaseGVN *phase, bool can_reshape) {
238 const jint con = in(2)->find_int_con(0);
239 if (con == 0) {
240 // If in(2) is not a constant, call Ideal() of the parent class to
241 // try to move constant to the right side.
242 return MulNode::Ideal(phase, can_reshape);
243 }
244
245 // Now we have a constant Node on the right and the constant in con.
246 if (con == 1) {
247 // By one is handled by Identity call
248 return nullptr;
249 }
250
251 // Check for negative constant; if so negate the final result
252 bool sign_flip = false;
253
254 unsigned int abs_con = g_uabs(con);
255 if (abs_con != (unsigned int)con) {
256 sign_flip = true;
257 }
258
259 // Get low bit; check for being the only bit
260 Node *res = nullptr;
261 unsigned int bit1 = submultiple_power_of_2(abs_con);
262 if (bit1 == abs_con) { // Found a power of 2?
263 res = new LShiftINode(in(1), phase->intcon(log2i_exact(bit1)));
264 } else {
265 // Check for constant with 2 bits set
266 unsigned int bit2 = abs_con - bit1;
267 bit2 = bit2 & (0 - bit2); // Extract 2nd bit
268 if (bit2 + bit1 == abs_con) { // Found all bits in con?
269 Node *n1 = phase->transform(new LShiftINode(in(1), phase->intcon(log2i_exact(bit1))));
270 Node *n2 = phase->transform(new LShiftINode(in(1), phase->intcon(log2i_exact(bit2))));
271 res = new AddINode(n2, n1);
272 } else if (is_power_of_2(abs_con + 1)) {
273 // Sleezy: power-of-2 - 1. Next time be generic.
274 unsigned int temp = abs_con + 1;
275 Node *n1 = phase->transform(new LShiftINode(in(1), phase->intcon(log2i_exact(temp))));
276 res = new SubINode(n1, in(1));
277 } else {
278 return MulNode::Ideal(phase, can_reshape);
279 }
280 }
281
282 if (sign_flip) { // Need to negate result?
283 res = phase->transform(res);// Transform, before making the zero con
284 res = new SubINode(phase->intcon(0),res);
285 }
286
287 return res; // Return final result
288 }
289
290 // This template class performs type multiplication for MulI/MulLNode. NativeType is either jint or jlong.
291 // In this class, the inputs of the MulNodes are named left and right with types [left_lo,left_hi] and [right_lo,right_hi].
292 //
293 // In general, the multiplication of two x-bit values could produce a result that consumes up to 2x bits if there is
294 // enough space to hold them all. We can therefore distinguish the following two cases for the product:
295 // - no overflow (i.e. product fits into x bits)
296 // - overflow (i.e. product does not fit into x bits)
297 //
298 // When multiplying the two x-bit inputs 'left' and 'right' with their x-bit types [left_lo,left_hi] and [right_lo,right_hi]
299 // we need to find the minimum and maximum of all possible products to define a new type. To do that, we compute the
300 // cross product of [left_lo,left_hi] and [right_lo,right_hi] in 2x-bit space where no over- or underflow can happen.
301 // The cross product consists of the following four multiplications with 2x-bit results:
302 // (1) left_lo * right_lo
303 // (2) left_lo * right_hi
304 // (3) left_hi * right_lo
305 // (4) left_hi * right_hi
306 //
307 // Let's define the following two functions:
308 // - Lx(i): Returns the lower x bits of the 2x-bit number i.
309 // - Ux(i): Returns the upper x bits of the 2x-bit number i.
310 //
311 // Let's first assume all products are positive where only overflows are possible but no underflows. If there is no
312 // overflow for a product p, then the upper x bits of the 2x-bit result p are all zero:
313 // Ux(p) = 0
314 // Lx(p) = p
315 //
316 // If none of the multiplications (1)-(4) overflow, we can truncate the upper x bits and use the following result type
317 // with x bits:
318 // [result_lo,result_hi] = [MIN(Lx(1),Lx(2),Lx(3),Lx(4)),MAX(Lx(1),Lx(2),Lx(3),Lx(4))]
319 //
320 // If any of these multiplications overflows, we could pessimistically take the bottom type for the x bit result
321 // (i.e. all values in the x-bit space could be possible):
322 // [result_lo,result_hi] = [NativeType_min,NativeType_max]
323 //
324 // However, in case of any overflow, we can do better by analyzing the upper x bits of all multiplications (1)-(4) with
325 // 2x-bit results. The upper x bits tell us something about how many times a multiplication has overflown the lower
326 // x bits. If the upper x bits of (1)-(4) are all equal, then we know that all of these multiplications overflowed
327 // the lower x bits the same number of times:
328 // Ux((1)) = Ux((2)) = Ux((3)) = Ux((4))
329 //
330 // If all upper x bits are equal, we can conclude:
331 // Lx(MIN((1),(2),(3),(4))) = MIN(Lx(1),Lx(2),Lx(3),Lx(4)))
332 // Lx(MAX((1),(2),(3),(4))) = MAX(Lx(1),Lx(2),Lx(3),Lx(4)))
333 //
334 // Therefore, we can use the same precise x-bit result type as for the no-overflow case:
335 // [result_lo,result_hi] = [(MIN(Lx(1),Lx(2),Lx(3),Lx(4))),MAX(Lx(1),Lx(2),Lx(3),Lx(4)))]
336 //
337 //
338 // Now let's assume that (1)-(4) are signed multiplications where over- and underflow could occur:
339 // Negative numbers are all sign extend with ones. Therefore, if a negative product does not underflow, then the
340 // upper x bits of the 2x-bit result are all set to ones which is minus one in two's complement. If there is an underflow,
341 // the upper x bits are decremented by the number of times an underflow occurred. The smallest possible negative product
342 // is NativeType_min*NativeType_max, where the upper x bits are set to NativeType_min / 2 (b11...0). It is therefore
343 // impossible to underflow the upper x bits. Thus, when having all ones (i.e. minus one) in the upper x bits, we know
344 // that there is no underflow.
345 //
346 // To be able to compare the number of over-/underflows of positive and negative products, respectively, we normalize
347 // the upper x bits of negative 2x-bit products by adding one. This way a product has no over- or underflow if the
348 // normalized upper x bits are zero. Now we can use the same improved type as for strictly positive products because we
349 // can compare the upper x bits in a unified way with N() being the normalization function:
350 // N(Ux((1))) = N(Ux((2))) = N(Ux((3)) = N(Ux((4)))
351 template<typename NativeType>
352 class IntegerTypeMultiplication {
353
354 NativeType _lo_left;
355 NativeType _lo_right;
356 NativeType _hi_left;
357 NativeType _hi_right;
358 short _widen_left;
359 short _widen_right;
360
361 static const Type* overflow_type();
362 static NativeType multiply_high(NativeType x, NativeType y);
363 const Type* create_type(NativeType lo, NativeType hi) const;
364
365 static NativeType multiply_high_signed_overflow_value(NativeType x, NativeType y) {
366 return normalize_overflow_value(x, y, multiply_high(x, y));
367 }
368
369 bool cross_product_not_same_overflow_value() const {
370 const NativeType lo_lo_high_product = multiply_high_signed_overflow_value(_lo_left, _lo_right);
371 const NativeType lo_hi_high_product = multiply_high_signed_overflow_value(_lo_left, _hi_right);
372 const NativeType hi_lo_high_product = multiply_high_signed_overflow_value(_hi_left, _lo_right);
373 const NativeType hi_hi_high_product = multiply_high_signed_overflow_value(_hi_left, _hi_right);
374 return lo_lo_high_product != lo_hi_high_product ||
375 lo_hi_high_product != hi_lo_high_product ||
376 hi_lo_high_product != hi_hi_high_product;
377 }
378
379 bool does_product_overflow(NativeType x, NativeType y) const {
380 return multiply_high_signed_overflow_value(x, y) != 0;
381 }
382
383 static NativeType normalize_overflow_value(const NativeType x, const NativeType y, NativeType result) {
384 return java_multiply(x, y) < 0 ? result + 1 : result;
385 }
386
387 public:
388 template<class IntegerType>
389 IntegerTypeMultiplication(const IntegerType* left, const IntegerType* right)
390 : _lo_left(left->_lo), _lo_right(right->_lo),
391 _hi_left(left->_hi), _hi_right(right->_hi),
392 _widen_left(left->_widen), _widen_right(right->_widen) {}
393
394 // Compute the product type by multiplying the two input type ranges. We take the minimum and maximum of all possible
395 // values (requires 4 multiplications of all possible combinations of the two range boundary values). If any of these
396 // multiplications overflows/underflows, we need to make sure that they all have the same number of overflows/underflows
397 // If that is not the case, we return the bottom type to cover all values due to the inconsistent overflows/underflows).
398 const Type* compute() const {
399 if (cross_product_not_same_overflow_value()) {
400 return overflow_type();
401 }
402
403 NativeType lo_lo_product = java_multiply(_lo_left, _lo_right);
404 NativeType lo_hi_product = java_multiply(_lo_left, _hi_right);
405 NativeType hi_lo_product = java_multiply(_hi_left, _lo_right);
406 NativeType hi_hi_product = java_multiply(_hi_left, _hi_right);
407 const NativeType min = MIN4(lo_lo_product, lo_hi_product, hi_lo_product, hi_hi_product);
408 const NativeType max = MAX4(lo_lo_product, lo_hi_product, hi_lo_product, hi_hi_product);
409 return create_type(min, max);
410 }
411
412 bool does_overflow() const {
413 return does_product_overflow(_lo_left, _lo_right) ||
414 does_product_overflow(_lo_left, _hi_right) ||
415 does_product_overflow(_hi_left, _lo_right) ||
416 does_product_overflow(_hi_left, _hi_right);
417 }
418 };
419
420 template <>
421 const Type* IntegerTypeMultiplication<jint>::overflow_type() {
422 return TypeInt::INT;
423 }
424
425 template <>
426 jint IntegerTypeMultiplication<jint>::multiply_high(const jint x, const jint y) {
427 const jlong x_64 = x;
428 const jlong y_64 = y;
429 const jlong product = x_64 * y_64;
430 return (jint)((uint64_t)product >> 32u);
431 }
432
433 template <>
434 const Type* IntegerTypeMultiplication<jint>::create_type(jint lo, jint hi) const {
435 return TypeInt::make(lo, hi, MAX2(_widen_left, _widen_right));
436 }
437
438 template <>
439 const Type* IntegerTypeMultiplication<jlong>::overflow_type() {
440 return TypeLong::LONG;
441 }
442
443 template <>
444 jlong IntegerTypeMultiplication<jlong>::multiply_high(const jlong x, const jlong y) {
445 return multiply_high_signed(x, y);
446 }
447
448 template <>
449 const Type* IntegerTypeMultiplication<jlong>::create_type(jlong lo, jlong hi) const {
450 return TypeLong::make(lo, hi, MAX2(_widen_left, _widen_right));
451 }
452
453 // Compute the product type of two integer ranges into this node.
454 const Type* MulINode::mul_ring(const Type* type_left, const Type* type_right) const {
455 const IntegerTypeMultiplication<jint> integer_multiplication(type_left->is_int(), type_right->is_int());
456 return integer_multiplication.compute();
457 }
458
459 bool MulINode::does_overflow(const TypeInt* type_left, const TypeInt* type_right) {
460 const IntegerTypeMultiplication<jint> integer_multiplication(type_left, type_right);
461 return integer_multiplication.does_overflow();
462 }
463
464 // Compute the product type of two long ranges into this node.
465 const Type* MulLNode::mul_ring(const Type* type_left, const Type* type_right) const {
466 const IntegerTypeMultiplication<jlong> integer_multiplication(type_left->is_long(), type_right->is_long());
467 return integer_multiplication.compute();
468 }
469
470 //=============================================================================
471 //------------------------------Ideal------------------------------------------
472 // Check for power-of-2 multiply, then try the regular MulNode::Ideal
473 Node *MulLNode::Ideal(PhaseGVN *phase, bool can_reshape) {
474 const jlong con = in(2)->find_long_con(0);
475 if (con == 0) {
476 // If in(2) is not a constant, call Ideal() of the parent class to
477 // try to move constant to the right side.
478 return MulNode::Ideal(phase, can_reshape);
479 }
480
481 // Now we have a constant Node on the right and the constant in con.
482 if (con == 1) {
483 // By one is handled by Identity call
484 return nullptr;
485 }
486
487 // Check for negative constant; if so negate the final result
488 bool sign_flip = false;
489 julong abs_con = g_uabs(con);
490 if (abs_con != (julong)con) {
491 sign_flip = true;
492 }
493
494 // Get low bit; check for being the only bit
495 Node *res = nullptr;
496 julong bit1 = submultiple_power_of_2(abs_con);
497 if (bit1 == abs_con) { // Found a power of 2?
498 res = new LShiftLNode(in(1), phase->intcon(log2i_exact(bit1)));
499 } else {
500
501 // Check for constant with 2 bits set
502 julong bit2 = abs_con-bit1;
503 bit2 = bit2 & (0-bit2); // Extract 2nd bit
504 if (bit2 + bit1 == abs_con) { // Found all bits in con?
505 Node *n1 = phase->transform(new LShiftLNode(in(1), phase->intcon(log2i_exact(bit1))));
506 Node *n2 = phase->transform(new LShiftLNode(in(1), phase->intcon(log2i_exact(bit2))));
507 res = new AddLNode(n2, n1);
508
509 } else if (is_power_of_2(abs_con+1)) {
510 // Sleezy: power-of-2 -1. Next time be generic.
511 julong temp = abs_con + 1;
512 Node *n1 = phase->transform( new LShiftLNode(in(1), phase->intcon(log2i_exact(temp))));
513 res = new SubLNode(n1, in(1));
514 } else {
515 return MulNode::Ideal(phase, can_reshape);
516 }
517 }
518
519 if (sign_flip) { // Need to negate result?
520 res = phase->transform(res);// Transform, before making the zero con
521 res = new SubLNode(phase->longcon(0),res);
522 }
523
524 return res; // Return final result
525 }
526
527 //=============================================================================
528 //------------------------------mul_ring---------------------------------------
529 // Compute the product type of two double ranges into this node.
530 const Type *MulFNode::mul_ring(const Type *t0, const Type *t1) const {
531 if( t0 == Type::FLOAT || t1 == Type::FLOAT ) return Type::FLOAT;
532 return TypeF::make( t0->getf() * t1->getf() );
533 }
534
535 //------------------------------Ideal---------------------------------------
536 // Check to see if we are multiplying by a constant 2 and convert to add, then try the regular MulNode::Ideal
537 Node* MulFNode::Ideal(PhaseGVN* phase, bool can_reshape) {
538 const TypeF *t2 = phase->type(in(2))->isa_float_constant();
539
540 // x * 2 -> x + x
541 if (t2 != nullptr && t2->getf() == 2) {
542 Node* base = in(1);
543 return new AddFNode(base, base);
544 }
545 return MulNode::Ideal(phase, can_reshape);
546 }
547
548 //=============================================================================
549 //------------------------------Ideal------------------------------------------
550 // Check to see if we are multiplying by a constant 2 and convert to add, then try the regular MulNode::Ideal
551 Node* MulHFNode::Ideal(PhaseGVN* phase, bool can_reshape) {
552 const TypeH* t2 = phase->type(in(2))->isa_half_float_constant();
553
554 // x * 2 -> x + x
555 if (t2 != nullptr && t2->getf() == 2) {
556 Node* base = in(1);
557 return new AddHFNode(base, base);
558 }
559 return MulNode::Ideal(phase, can_reshape);
560 }
561
562 // Compute the product type of two half float ranges into this node.
563 const Type* MulHFNode::mul_ring(const Type* t0, const Type* t1) const {
564 if (t0 == Type::HALF_FLOAT || t1 == Type::HALF_FLOAT) {
565 return Type::HALF_FLOAT;
566 }
567 return TypeH::make(t0->getf() * t1->getf());
568 }
569
570 //=============================================================================
571 //------------------------------mul_ring---------------------------------------
572 // Compute the product type of two double ranges into this node.
573 const Type *MulDNode::mul_ring(const Type *t0, const Type *t1) const {
574 if( t0 == Type::DOUBLE || t1 == Type::DOUBLE ) return Type::DOUBLE;
575 // We must be multiplying 2 double constants.
576 return TypeD::make( t0->getd() * t1->getd() );
577 }
578
579 //------------------------------Ideal---------------------------------------
580 // Check to see if we are multiplying by a constant 2 and convert to add, then try the regular MulNode::Ideal
581 Node* MulDNode::Ideal(PhaseGVN* phase, bool can_reshape) {
582 const TypeD *t2 = phase->type(in(2))->isa_double_constant();
583
584 // x * 2 -> x + x
585 if (t2 != nullptr && t2->getd() == 2) {
586 Node* base = in(1);
587 return new AddDNode(base, base);
588 }
589
590 return MulNode::Ideal(phase, can_reshape);
591 }
592
593 //=============================================================================
594 //------------------------------Value------------------------------------------
595 const Type* MulHiLNode::Value(PhaseGVN* phase) const {
596 const Type *t1 = phase->type( in(1) );
597 const Type *t2 = phase->type( in(2) );
598 const Type *bot = bottom_type();
599 return MulHiValue(t1, t2, bot);
600 }
601
602 const Type* UMulHiLNode::Value(PhaseGVN* phase) const {
603 const Type *t1 = phase->type( in(1) );
604 const Type *t2 = phase->type( in(2) );
605 const Type *bot = bottom_type();
606 return MulHiValue(t1, t2, bot);
607 }
608
609 // A common routine used by UMulHiLNode and MulHiLNode
610 const Type* MulHiValue(const Type *t1, const Type *t2, const Type *bot) {
611 // Either input is TOP ==> the result is TOP
612 if( t1 == Type::TOP ) return Type::TOP;
613 if( t2 == Type::TOP ) return Type::TOP;
614
615 // Either input is BOTTOM ==> the result is the local BOTTOM
616 if( (t1 == bot) || (t2 == bot) ||
617 (t1 == Type::BOTTOM) || (t2 == Type::BOTTOM) )
618 return bot;
619
620 // It is not worth trying to constant fold this stuff!
621 return TypeLong::LONG;
622 }
623
624 //=============================================================================
625 //------------------------------mul_ring---------------------------------------
626 // Supplied function returns the product of the inputs IN THE CURRENT RING.
627 // For the logical operations the ring's MUL is really a logical AND function.
628 // This also type-checks the inputs for sanity. Guaranteed never to
629 // be passed a TOP or BOTTOM type, these are filtered out by pre-check.
630 const Type* AndINode::mul_ring(const Type* t1, const Type* t2) const {
631 return RangeInference::infer_and(t1->is_int(), t2->is_int());
632 }
633
634 static bool AndIL_is_zero_element_under_mask(const PhaseGVN* phase, const Node* expr, const Node* mask, BasicType bt);
635
636 const Type* AndINode::Value(PhaseGVN* phase) const {
637 if (AndIL_is_zero_element_under_mask(phase, in(1), in(2), T_INT) ||
638 AndIL_is_zero_element_under_mask(phase, in(2), in(1), T_INT)) {
639 return TypeInt::ZERO;
640 }
641
642 return MulNode::Value(phase);
643 }
644
645 //------------------------------Identity---------------------------------------
646 // Masking off the high bits of an unsigned load is not required
647 Node* AndINode::Identity(PhaseGVN* phase) {
648
649 // x & x => x
650 if (in(1) == in(2)) {
651 return in(1);
652 }
653
654 const TypeInt* t1 = phase->type(in(1))->is_int();
655 const TypeInt* t2 = phase->type(in(2))->is_int();
656
657 if ((~t1->_bits._ones & ~t2->_bits._zeros) == 0) {
658 // All bits that might be 0 in in1 are known to be 0 in in2
659 return in(2);
660 }
661
662 if ((~t2->_bits._ones & ~t1->_bits._zeros) == 0) {
663 // All bits that might be 0 in in2 are known to be 0 in in1
664 return in(1);
665 }
666
667 return MulNode::Identity(phase);
668 }
669
670 //------------------------------Ideal------------------------------------------
671 Node *AndINode::Ideal(PhaseGVN *phase, bool can_reshape) {
672 // Simplify (v1 + v2) & mask to v1 & mask or v2 & mask when possible.
673 Node* progress = AndIL_sum_and_mask(phase, T_INT);
674 if (progress != nullptr) {
675 return progress;
676 }
677
678 // Convert "(~a) & (~b)" into "~(a | b)"
679 if (AddNode::is_not(phase, in(1), T_INT) && AddNode::is_not(phase, in(2), T_INT)) {
680 Node* or_a_b = new OrINode(in(1)->in(1), in(2)->in(1));
681 Node* tn = phase->transform(or_a_b);
682 return AddNode::make_not(phase, tn, T_INT);
683 }
684
685 // Special case constant AND mask
686 const TypeInt *t2 = phase->type( in(2) )->isa_int();
687 if( !t2 || !t2->is_con() ) return MulNode::Ideal(phase, can_reshape);
688 const int mask = t2->get_con();
689 Node *load = in(1);
690 uint lop = load->Opcode();
691
692 // Masking bits off of a Character? Hi bits are already zero.
693 if( lop == Op_LoadUS &&
694 (mask & 0xFFFF0000) ) // Can we make a smaller mask?
695 return new AndINode(load,phase->intcon(mask&0xFFFF));
696
697 // Masking bits off of a Short? Loading a Character does some masking
698 if (can_reshape &&
699 load->outcnt() == 1 && load->unique_out() == this) {
700 if (lop == Op_LoadS && (mask & 0xFFFF0000) == 0 ) {
701 Node* ldus = load->as_Load()->convert_to_unsigned_load(*phase);
702 ldus = phase->transform(ldus);
703 return new AndINode(ldus, phase->intcon(mask & 0xFFFF));
704 }
705
706 // Masking sign bits off of a Byte? Do an unsigned byte load plus
707 // an and.
708 if (lop == Op_LoadB && (mask & 0xFFFFFF00) == 0) {
709 Node* ldub = load->as_Load()->convert_to_unsigned_load(*phase);
710 ldub = phase->transform(ldub);
711 return new AndINode(ldub, phase->intcon(mask));
712 }
713 }
714
715 // Masking off sign bits? Dont make them!
716 if( lop == Op_RShiftI ) {
717 const TypeInt *t12 = phase->type(load->in(2))->isa_int();
718 if( t12 && t12->is_con() ) { // Shift is by a constant
719 int shift = t12->get_con();
720 shift &= BitsPerJavaInteger-1; // semantics of Java shifts
721 const int sign_bits_mask = ~right_n_bits(BitsPerJavaInteger - shift);
722 // If the AND'ing of the 2 masks has no bits, then only original shifted
723 // bits survive. NO sign-extension bits survive the maskings.
724 if( (sign_bits_mask & mask) == 0 ) {
725 // Use zero-fill shift instead
726 Node *zshift = phase->transform(new URShiftINode(load->in(1),load->in(2)));
727 return new AndINode( zshift, in(2) );
728 }
729 }
730 }
731
732 // Check for 'negate/and-1', a pattern emitted when someone asks for
733 // 'mod 2'. Negate leaves the low order bit unchanged (think: complement
734 // plus 1) and the mask is of the low order bit. Skip the negate.
735 if( lop == Op_SubI && mask == 1 && load->in(1) &&
736 phase->type(load->in(1)) == TypeInt::ZERO )
737 return new AndINode( load->in(2), in(2) );
738
739 return MulNode::Ideal(phase, can_reshape);
740 }
741
742 //=============================================================================
743 //------------------------------mul_ring---------------------------------------
744 // Supplied function returns the product of the inputs IN THE CURRENT RING.
745 // For the logical operations the ring's MUL is really a logical AND function.
746 // This also type-checks the inputs for sanity. Guaranteed never to
747 // be passed a TOP or BOTTOM type, these are filtered out by pre-check.
748 const Type* AndLNode::mul_ring(const Type* t1, const Type* t2) const {
749 return RangeInference::infer_and(t1->is_long(), t2->is_long());
750 }
751
752 const Type* AndLNode::Value(PhaseGVN* phase) const {
753 if (AndIL_is_zero_element_under_mask(phase, in(1), in(2), T_LONG) ||
754 AndIL_is_zero_element_under_mask(phase, in(2), in(1), T_LONG)) {
755 return TypeLong::ZERO;
756 }
757
758 return MulNode::Value(phase);
759 }
760
761 //------------------------------Identity---------------------------------------
762 // Masking off the high bits of an unsigned load is not required
763 Node* AndLNode::Identity(PhaseGVN* phase) {
764
765 // x & x => x
766 if (in(1) == in(2)) {
767 return in(1);
768 }
769
770 const TypeLong* t1 = phase->type(in(1))->is_long();
771 const TypeLong* t2 = phase->type(in(2))->is_long();
772
773 if ((~t1->_bits._ones & ~t2->_bits._zeros) == 0) {
774 // All bits that might be 0 in in1 are known to be 0 in in2
775 return in(2);
776 }
777
778 if ((~t2->_bits._ones & ~t1->_bits._zeros) == 0) {
779 // All bits that might be 0 in in2 are known to be 0 in in1
780 return in(1);
781 }
782
783 return MulNode::Identity(phase);
784 }
785
786 //------------------------------Ideal------------------------------------------
787 Node *AndLNode::Ideal(PhaseGVN *phase, bool can_reshape) {
788 // Simplify (v1 + v2) & mask to v1 & mask or v2 & mask when possible.
789 Node* progress = AndIL_sum_and_mask(phase, T_LONG);
790 if (progress != nullptr) {
791 return progress;
792 }
793
794 // Convert "(~a) & (~b)" into "~(a | b)"
795 if (AddNode::is_not(phase, in(1), T_LONG) && AddNode::is_not(phase, in(2), T_LONG)) {
796 Node* or_a_b = new OrLNode(in(1)->in(1), in(2)->in(1));
797 Node* tn = phase->transform(or_a_b);
798 return AddNode::make_not(phase, tn, T_LONG);
799 }
800
801 // Special case constant AND mask
802 const TypeLong *t2 = phase->type( in(2) )->isa_long();
803 if( !t2 || !t2->is_con() ) return MulNode::Ideal(phase, can_reshape);
804 const jlong mask = t2->get_con();
805
806 Node* in1 = in(1);
807 int op = in1->Opcode();
808
809 // Are we masking a long that was converted from an int with a mask
810 // that fits in 32-bits? Commute them and use an AndINode. Don't
811 // convert masks which would cause a sign extension of the integer
812 // value. This check includes UI2L masks (0x00000000FFFFFFFF) which
813 // would be optimized away later in Identity.
814 if (op == Op_ConvI2L && (mask & UCONST64(0xFFFFFFFF80000000)) == 0) {
815 Node* andi = new AndINode(in1->in(1), phase->intcon(mask));
816 andi = phase->transform(andi);
817 return new ConvI2LNode(andi);
818 }
819
820 // Masking off sign bits? Dont make them!
821 if (op == Op_RShiftL) {
822 const TypeInt* t12 = phase->type(in1->in(2))->isa_int();
823 if( t12 && t12->is_con() ) { // Shift is by a constant
824 int shift = t12->get_con();
825 shift &= BitsPerJavaLong - 1; // semantics of Java shifts
826 if (shift != 0) {
827 const julong sign_bits_mask = ~(((julong)CONST64(1) << (julong)(BitsPerJavaLong - shift)) -1);
828 // If the AND'ing of the 2 masks has no bits, then only original shifted
829 // bits survive. NO sign-extension bits survive the maskings.
830 if( (sign_bits_mask & mask) == 0 ) {
831 // Use zero-fill shift instead
832 Node *zshift = phase->transform(new URShiftLNode(in1->in(1), in1->in(2)));
833 return new AndLNode(zshift, in(2));
834 }
835 }
836 }
837 }
838
839 return MulNode::Ideal(phase, can_reshape);
840 }
841
842 LShiftNode* LShiftNode::make(Node* in1, Node* in2, BasicType bt) {
843 switch (bt) {
844 case T_INT:
845 return new LShiftINode(in1, in2);
846 case T_LONG:
847 return new LShiftLNode(in1, in2);
848 default:
849 fatal("Not implemented for %s", type2name(bt));
850 }
851 return nullptr;
852 }
853
854 // Returns whether the shift amount is constant. If so, sets count.
855 static bool const_shift_count(PhaseGVN* phase, const Node* shift_node, int* count) {
856 const TypeInt* tcount = phase->type(shift_node->in(2))->isa_int();
857 if (tcount != nullptr && tcount->is_con()) {
858 *count = tcount->get_con();
859 return true;
860 }
861 return false;
862 }
863
864 // Returns whether the shift amount is constant. If so, sets real_shift and masked_shift.
865 static bool mask_shift_amount(PhaseGVN* phase, const Node* shift_node, uint nBits, int& real_shift, uint& masked_shift) {
866 if (const_shift_count(phase, shift_node, &real_shift)) {
867 masked_shift = real_shift & (nBits - 1);
868 return true;
869 }
870 return false;
871 }
872
873 // Convenience for when we don't care about the real amount
874 static bool mask_shift_amount(PhaseGVN* phase, const Node* shift_node, uint nBits, uint& masked_shift) {
875 int real_shift;
876 return mask_shift_amount(phase, shift_node, nBits, real_shift, masked_shift);
877 }
878
879 // Use this in ::Ideal only with shiftNode == this!
880 // Sets masked_shift to the masked shift amount if constant or 0 if not constant.
881 // Returns shift_node if the shift amount input node was modified, nullptr otherwise.
882 static Node* mask_and_replace_shift_amount(PhaseGVN* phase, Node* shift_node, uint nBits, uint& masked_shift) {
883 int real_shift;
884 if (mask_shift_amount(phase, shift_node, nBits, real_shift, masked_shift)) {
885 if (masked_shift == 0) {
886 // Let Identity() handle 0 shift count.
887 return nullptr;
888 }
889
890 if (real_shift != (int)masked_shift) {
891 shift_node->set_req(2, phase->intcon(masked_shift)); // Replace shift count with masked value.
892
893 // We need to notify the caller that the graph was reshaped, as Ideal needs
894 // to return the root of the reshaped graph if any change was made.
895 return shift_node;
896 }
897 } else {
898 // Not a shift by a constant.
899 masked_shift = 0;
900 }
901 return nullptr;
902 }
903
904 // Called with
905 // outer_shift = (_ << rhs_outer)
906 // We are looking for the pattern:
907 // outer_shift = ((X << rhs_inner) << rhs_outer)
908 // where rhs_outer and rhs_inner are constant
909 // we denote inner_shift the nested expression (X << rhs_inner)
910 // con_inner = rhs_inner % nbits and con_outer = rhs_outer % nbits
911 // where nbits is the number of bits of the shifts
912 //
913 // There are 2 cases:
914 // if con_outer + con_inner >= nbits => 0
915 // if con_outer + con_inner < nbits => X << (con_outer + con_inner)
916 static Node* collapse_nested_shift_left(PhaseGVN* phase, const Node* outer_shift, uint con_outer, BasicType bt) {
917 assert(bt == T_LONG || bt == T_INT, "Unexpected type");
918 const Node* inner_shift = outer_shift->in(1);
919 if (inner_shift->Opcode() != Op_LShift(bt)) {
920 return nullptr;
921 }
922
923 uint nbits = bits_per_java_integer(bt);
924 uint con_inner;
925 if (!mask_shift_amount(phase, inner_shift, nbits, con_inner)) {
926 return nullptr;
927 }
928
929 if (con_inner == 0) {
930 // We let the Identity() of the inner shift do its job.
931 return nullptr;
932 }
933
934 if (con_outer + con_inner >= nbits) {
935 // While it might be tempting to use
936 // phase->zerocon(bt);
937 // it would be incorrect: zerocon caches nodes, while Ideal is only allowed
938 // to return a new node, this or nullptr, but not an old (cached) node.
939 return ConNode::make(TypeInteger::zero(bt));
940 }
941
942 // con0 + con1 < nbits ==> actual shift happens now
943 Node* con0_plus_con1 = phase->intcon(con_outer + con_inner);
944 return LShiftNode::make(inner_shift->in(1), con0_plus_con1, bt);
945 }
946
947 //------------------------------Identity---------------------------------------
948 Node* LShiftINode::Identity(PhaseGVN* phase) {
949 return IdentityIL(phase, T_INT);
950 }
951
952 Node* LShiftNode::IdealIL(PhaseGVN* phase, bool can_reshape, BasicType bt) {
953 uint con;
954 Node* progress = mask_and_replace_shift_amount(phase, this, bits_per_java_integer(bt), con);
955 if (con == 0) {
956 return nullptr;
957 }
958
959 // If the right input is a constant, and the left input is an add of a
960 // constant, flatten the tree: (X+con1)<<con0 ==> X<<con0 + con1<<con0
961 Node* add1 = in(1);
962 int add1_op = add1->Opcode();
963 if (add1_op == Op_Add(bt)) { // Left input is an add?
964 assert(add1 != add1->in(1), "dead loop in LShiftINode::Ideal");
965
966 // Transform is legal, but check for profit. Avoid breaking 'i2s'
967 // and 'i2b' patterns which typically fold into 'StoreC/StoreB'.
968 if (bt != T_INT || con < 16) {
969 // Left input is an add of the same number?
970 if (con != (bits_per_java_integer(bt) - 1) && add1->in(1) == add1->in(2)) {
971 // Convert "(x + x) << c0" into "x << (c0 + 1)"
972 // In general, this optimization cannot be applied for c0 == 31 (for LShiftI) since
973 // 2x << 31 != x << 32 = x << 0 = x (e.g. x = 1: 2 << 31 = 0 != 1)
974 // or c0 != 63 (for LShiftL) because:
975 // (x + x) << 63 = 2x << 63, while
976 // (x + x) << 63 --transform--> x << 64 = x << 0 = x (!= 2x << 63, for example for x = 1)
977 // According to the Java spec, chapter 15.19, we only consider the six lowest-order bits of the right-hand operand
978 // (i.e. "right-hand operand" & 0b111111). Therefore, x << 64 is the same as x << 0 (64 = 0b10000000 & 0b0111111 = 0).
979 return LShiftNode::make(add1->in(1), phase->intcon(con + 1), bt);
980 }
981
982 // Left input is an add of a constant?
983 const TypeInteger* t12 = phase->type(add1->in(2))->isa_integer(bt);
984 if (t12 != nullptr && t12->is_con()) { // Left input is an add of a con?
985 // Compute X << con0
986 Node* lsh = phase->transform(LShiftNode::make(add1->in(1), in(2), bt));
987 // Compute X<<con0 + (con1<<con0)
988 return AddNode::make(lsh, phase->integercon(java_shift_left(t12->get_con_as_long(bt), con, bt), bt), bt);
989 }
990 }
991 }
992 // Check for "(con0 - X) << con1"
993 // Transform is legal, but check for profit. Avoid breaking 'i2s'
994 // and 'i2b' patterns which typically fold into 'StoreC/StoreB'.
995 if (add1_op == Op_Sub(bt) && (bt != T_INT || con < 16)) { // Left input is a sub?
996 // Left input is a sub from a constant?
997 const TypeInteger* t11 = phase->type(add1->in(1))->isa_integer(bt);
998 if (t11 != nullptr && t11->is_con()) {
999 // Compute X << con0
1000 Node* lsh = phase->transform(LShiftNode::make(add1->in(2), in(2), bt));
1001 // Compute (con1<<con0) - (X<<con0)
1002 return SubNode::make(phase->integercon(java_shift_left(t11->get_con_as_long(bt), con, bt), bt), lsh, bt);
1003 }
1004 }
1005
1006 // Check for "(x >> C1) << C2"
1007 if (add1_op == Op_RShift(bt) || add1_op == Op_URShift(bt)) {
1008 int add1Con = 0;
1009 const_shift_count(phase, add1, &add1Con);
1010
1011 // Special case C1 == C2, which just masks off low bits
1012 if (add1Con > 0 && con == (uint)add1Con) {
1013 // Convert to "(x & -(1 << C2))"
1014 return MulNode::make_and(add1->in(1), phase->integercon(java_negate(java_shift_left(1, con, bt), bt), bt), bt);
1015 } else {
1016 // Wait until the right shift has been sharpened to the correct count
1017 if (add1Con > 0 && (uint)add1Con < bits_per_java_integer(bt)) {
1018 // As loop parsing can produce LShiftI nodes, we should wait until the graph is fully formed
1019 // to apply optimizations, otherwise we can inadvertently stop vectorization opportunities.
1020 if (phase->is_IterGVN()) {
1021 if (con > (uint)add1Con) {
1022 // Creates "(x << (C2 - C1)) & -(1 << C2)"
1023 Node* lshift = phase->transform(LShiftNode::make(add1->in(1), phase->intcon(con - add1Con), bt));
1024 return MulNode::make_and(lshift, phase->integercon(java_negate(java_shift_left(1, con, bt), bt), bt), bt);
1025 } else {
1026 assert(con < (uint)add1Con, "must be (%d < %d)", con, add1Con);
1027 // Creates "(x >> (C1 - C2)) & -(1 << C2)"
1028
1029 // Handle logical and arithmetic shifts
1030 Node* rshift;
1031 if (add1_op == Op_RShift(bt)) {
1032 rshift = phase->transform(RShiftNode::make(add1->in(1), phase->intcon(add1Con - con), bt));
1033 } else {
1034 rshift = phase->transform(URShiftNode::make(add1->in(1), phase->intcon(add1Con - con), bt));
1035 }
1036
1037 return MulNode::make_and(rshift, phase->integercon(java_negate(java_shift_left(1, con, bt)), bt), bt);
1038 }
1039 } else {
1040 phase->record_for_igvn(this);
1041 }
1042 }
1043 }
1044 }
1045
1046 // Check for "((x >> C1) & Y) << C2"
1047 if (add1_op == Op_And(bt)) {
1048 Node* add2 = add1->in(1);
1049 int add2_op = add2->Opcode();
1050 if (add2_op == Op_RShift(bt) || add2_op == Op_URShift(bt)) {
1051 // Special case C1 == C2, which just masks off low bits
1052 if (add2->in(2) == in(2)) {
1053 // Convert to "(x & (Y << C2))"
1054 Node* y_sh = phase->transform(LShiftNode::make(add1->in(2), phase->intcon(con), bt));
1055 return MulNode::make_and(add2->in(1), y_sh, bt);
1056 }
1057
1058 int add2Con = 0;
1059 const_shift_count(phase, add2, &add2Con);
1060 if (add2Con > 0 && (uint)add2Con < bits_per_java_integer(bt)) {
1061 if (phase->is_IterGVN()) {
1062 // Convert to "((x >> C1) << C2) & (Y << C2)"
1063
1064 // Make "(x >> C1) << C2", which will get folded away by the rule above
1065 Node* x_sh = phase->transform(LShiftNode::make(add2, phase->intcon(con), bt));
1066 // Make "Y << C2", which will simplify when Y is a constant
1067 Node* y_sh = phase->transform(LShiftNode::make(add1->in(2), phase->intcon(con), bt));
1068
1069 return MulNode::make_and(x_sh, y_sh, bt);
1070 } else {
1071 phase->record_for_igvn(this);
1072 }
1073 }
1074 }
1075 }
1076
1077 // Check for ((x & ((1<<(32-c0))-1)) << c0) which ANDs off high bits
1078 // before shifting them away.
1079 const jlong bits_mask = max_unsigned_integer(bt) >> con;
1080 assert(bt != T_INT || bits_mask == right_n_bits(bits_per_java_integer(bt)-con), "inconsistent");
1081 if (add1_op == Op_And(bt) &&
1082 phase->type(add1->in(2)) == TypeInteger::make(bits_mask, bt)) {
1083 return LShiftNode::make(add1->in(1), in(2), bt);
1084 }
1085
1086 // Collapse nested left-shifts with constant rhs:
1087 // (X << con1) << con2 ==> X << (con1 + con2)
1088 Node* doubleShift = collapse_nested_shift_left(phase, this, con, bt);
1089 if (doubleShift != nullptr) {
1090 return doubleShift;
1091 }
1092
1093 return progress;
1094 }
1095
1096 //------------------------------Ideal------------------------------------------
1097 Node* LShiftINode::Ideal(PhaseGVN *phase, bool can_reshape) {
1098 return IdealIL(phase, can_reshape, T_INT);
1099 }
1100
1101 const Type* LShiftNode::ValueIL(PhaseGVN* phase, BasicType bt) const {
1102 const Type* t1 = phase->type(in(1));
1103 const Type* t2 = phase->type(in(2));
1104 // Either input is TOP ==> the result is TOP
1105 if (t1 == Type::TOP) {
1106 return Type::TOP;
1107 }
1108 if (t2 == Type::TOP) {
1109 return Type::TOP;
1110 }
1111
1112 // Left input is ZERO ==> the result is ZERO.
1113 if (t1 == TypeInteger::zero(bt)) {
1114 return TypeInteger::zero(bt);
1115 }
1116 // Shift by zero does nothing
1117 if (t2 == TypeInt::ZERO) {
1118 return t1;
1119 }
1120
1121 // If nothing is known about the shift amount then the result is BOTTOM
1122 if (t2 == TypeInt::INT) {
1123 return TypeInteger::bottom(bt);
1124 }
1125
1126 const TypeInteger* r1 = t1->is_integer(bt); // Handy access
1127 // Since the shift semantics in Java take into account only the bottom five
1128 // bits for ints and the bottom six bits for longs, we can further constrain
1129 // the range of values of the shift amount by ANDing with the right mask based
1130 // on whether the type is int or long.
1131 const TypeInt* mask = TypeInt::make(bits_per_java_integer(bt) - 1);
1132 const TypeInt* r2 = RangeInference::infer_and(t2->is_int(), mask);
1133
1134 if (!r2->is_con()) {
1135 return TypeInteger::bottom(bt);
1136 }
1137
1138 uint shift = r2->get_con();
1139 // Shift by a multiple of 32/64 does nothing:
1140 if (shift == 0) {
1141 return t1;
1142 }
1143
1144 // If the shift is a constant, shift the bounds of the type,
1145 // unless this could lead to an overflow.
1146 if (!r1->is_con()) {
1147 #ifdef ASSERT
1148 if (bt == T_INT) {
1149 jlong lo = r1->lo_as_long(), hi = r1->hi_as_long();
1150 jint lo_int = r1->is_int()->_lo, hi_int = r1->is_int()->_hi;
1151 assert((java_shift_right(java_shift_left(lo, shift, bt), shift, bt) == lo) == (((lo_int << shift) >> shift) == lo_int), "inconsistent");
1152 assert((java_shift_right(java_shift_left(hi, shift, bt), shift, bt) == hi) == (((hi_int << shift) >> shift) == hi_int), "inconsistent");
1153 }
1154 #endif
1155
1156 if (bt == T_INT) {
1157 return RangeInference::infer_lshift(r1->is_int(), shift);
1158 }
1159
1160 return RangeInference::infer_lshift(r1->is_long(), shift);
1161 }
1162
1163 return TypeInteger::make(java_shift_left(r1->get_con_as_long(bt), shift, bt), bt);
1164 }
1165
1166 //------------------------------Value------------------------------------------
1167 const Type* LShiftINode::Value(PhaseGVN* phase) const {
1168 return ValueIL(phase, T_INT);
1169 }
1170
1171 Node* LShiftNode::IdentityIL(PhaseGVN* phase, BasicType bt) {
1172 int count = 0;
1173 if (const_shift_count(phase, this, &count) && (count & (bits_per_java_integer(bt) - 1)) == 0) {
1174 // Shift by a multiple of 32/64 does nothing
1175 return in(1);
1176 }
1177 return this;
1178 }
1179
1180 //=============================================================================
1181 //------------------------------Identity---------------------------------------
1182 Node* LShiftLNode::Identity(PhaseGVN* phase) {
1183 return IdentityIL(phase, T_LONG);
1184 }
1185
1186 //------------------------------Ideal------------------------------------------
1187 Node* LShiftLNode::Ideal(PhaseGVN* phase, bool can_reshape) {
1188 return IdealIL(phase, can_reshape, T_LONG);
1189 }
1190
1191 //------------------------------Value------------------------------------------
1192 const Type* LShiftLNode::Value(PhaseGVN* phase) const {
1193 return ValueIL(phase, T_LONG);
1194 }
1195
1196 RShiftNode* RShiftNode::make(Node* in1, Node* in2, BasicType bt) {
1197 switch (bt) {
1198 case T_INT:
1199 return new RShiftINode(in1, in2);
1200 case T_LONG:
1201 return new RShiftLNode(in1, in2);
1202 default:
1203 fatal("Not implemented for %s", type2name(bt));
1204 }
1205 return nullptr;
1206 }
1207
1208
1209 //=============================================================================
1210 //------------------------------Identity---------------------------------------
1211 Node* RShiftNode::IdentityIL(PhaseGVN* phase, BasicType bt) {
1212 int count = 0;
1213 if (const_shift_count(phase, this, &count)) {
1214 if ((count & (bits_per_java_integer(bt) - 1)) == 0) {
1215 // Shift by a multiple of 32/64 does nothing
1216 return in(1);
1217 }
1218 // Check for useless sign-masking
1219 int lshift_count = 0;
1220 if (in(1)->Opcode() == Op_LShift(bt) &&
1221 in(1)->req() == 3 &&
1222 // Compare shift counts by value, not by node pointer, to also match a not-yet-normalized
1223 // negative constant (e.g. -1 vs 31)
1224 const_shift_count(phase, in(1), &lshift_count)) {
1225 count &= bits_per_java_integer(bt) - 1; // semantics of Java shifts
1226 lshift_count &= bits_per_java_integer(bt) - 1;
1227 if (count == lshift_count) {
1228 // Compute masks for which this shifting doesn't change
1229 jlong lo = (CONST64(-1) << (bits_per_java_integer(bt) - ((uint)count)-1)); // FFFF8000
1230 jlong hi = ~lo; // 00007FFF
1231 const TypeInteger* t11 = phase->type(in(1)->in(1))->isa_integer(bt);
1232 if (t11 == nullptr) {
1233 return this;
1234 }
1235 // Does actual value fit inside of mask?
1236 if (lo <= t11->lo_as_long() && t11->hi_as_long() <= hi) {
1237 return in(1)->in(1); // Then shifting is a nop
1238 }
1239 }
1240 }
1241 }
1242 return this;
1243 }
1244
1245 Node* RShiftINode::Identity(PhaseGVN* phase) {
1246 return IdentityIL(phase, T_INT);
1247 }
1248
1249 Node* RShiftNode::IdealIL(PhaseGVN* phase, bool can_reshape, BasicType bt) {
1250 // Inputs may be TOP if they are dead.
1251 const TypeInteger* t1 = phase->type(in(1))->isa_integer(bt);
1252 if (t1 == nullptr) {
1253 return NodeSentinel; // Left input is an integer
1254 }
1255
1256 uint shift;
1257 Node* progress = mask_and_replace_shift_amount(phase, this, bits_per_java_integer(bt), shift);
1258 if (shift == 0) {
1259 return NodeSentinel;
1260 }
1261
1262 // Check for (x & 0xFF000000) >> 24, whose mask can be made smaller.
1263 // and convert to (x >> 24) & (0xFF000000 >> 24) = x >> 24
1264 // Such expressions arise normally from shift chains like (byte)(x >> 24).
1265 const Node* and_node = in(1);
1266 if (and_node->Opcode() != Op_And(bt)) {
1267 return progress;
1268 }
1269 const TypeInteger* mask_t = phase->type(and_node->in(2))->isa_integer(bt);
1270 if (mask_t != nullptr && mask_t->is_con()) {
1271 jlong maskbits = mask_t->get_con_as_long(bt);
1272 // Convert to "(x >> shift) & (mask >> shift)"
1273 Node* shr_nomask = phase->transform(RShiftNode::make(and_node->in(1), in(2), bt));
1274 return MulNode::make_and(shr_nomask, phase->integercon(maskbits >> shift, bt), bt);
1275 }
1276
1277 return progress;
1278 }
1279
1280 Node* RShiftINode::Ideal(PhaseGVN* phase, bool can_reshape) {
1281 Node* progress = IdealIL(phase, can_reshape, T_INT);
1282 if (progress == NodeSentinel) {
1283 return nullptr;
1284 }
1285 if (progress != nullptr) {
1286 return progress;
1287 }
1288 uint shift;
1289 progress = mask_and_replace_shift_amount(phase, this, BitsPerJavaInteger, shift);
1290 assert(shift != 0, "handled by IdealIL");
1291
1292 // Check for "(short[i] <<16)>>16" which simply sign-extends
1293 const Node *shl = in(1);
1294 if (shl->Opcode() != Op_LShiftI) {
1295 return progress;
1296 }
1297
1298 const TypeInt* left_shift_t = phase->type(shl->in(2))->isa_int();
1299 if (left_shift_t == nullptr) {
1300 return progress;
1301 }
1302 if (shift == 16 && left_shift_t->is_con(16)) {
1303 Node *ld = shl->in(1);
1304 if (ld->Opcode() == Op_LoadS) {
1305 // Sign extension is just useless here. Return a RShiftI of zero instead
1306 // returning 'ld' directly. We cannot return an old Node directly as
1307 // that is the job of 'Identity' calls and Identity calls only work on
1308 // direct inputs ('ld' is an extra Node removed from 'this'). The
1309 // combined optimization requires Identity only return direct inputs.
1310 set_req_X(1, ld, phase);
1311 set_req_X(2, phase->intcon(0), phase);
1312 return this;
1313 }
1314 else if (can_reshape &&
1315 ld->Opcode() == Op_LoadUS &&
1316 ld->outcnt() == 1 && ld->unique_out() == shl)
1317 // Replace zero-extension-load with sign-extension-load
1318 return ld->as_Load()->convert_to_signed_load(*phase);
1319 }
1320
1321 // Check for "(byte[i] <<24)>>24" which simply sign-extends
1322 if (shift == 24 && left_shift_t->is_con(24)) {
1323 Node *ld = shl->in(1);
1324 if (ld->Opcode() == Op_LoadB) {
1325 // Sign extension is just useless here
1326 set_req_X(1, ld, phase);
1327 set_req_X(2, phase->intcon(0), phase);
1328 return this;
1329 }
1330 }
1331
1332 return progress;
1333 }
1334
1335 const Type* RShiftNode::ValueIL(PhaseGVN* phase, BasicType bt) const {
1336 const Type* t1 = phase->type(in(1));
1337 const Type* t2 = phase->type(in(2));
1338 // Either input is TOP ==> the result is TOP
1339 if (t1 == Type::TOP) {
1340 return Type::TOP;
1341 }
1342 if (t2 == Type::TOP) {
1343 return Type::TOP;
1344 }
1345
1346 // Left input is ZERO ==> the result is ZERO.
1347 if (t1 == TypeInteger::zero(bt)) {
1348 return TypeInteger::zero(bt);
1349 }
1350 // Shift by zero does nothing
1351 if (t2 == TypeInt::ZERO) {
1352 return t1;
1353 }
1354
1355 // Either input is BOTTOM ==> the result is BOTTOM
1356 if (t1 == Type::BOTTOM || t2 == Type::BOTTOM) {
1357 return TypeInteger::bottom(bt);
1358 }
1359
1360 const TypeInteger* r1 = t1->isa_integer(bt);
1361 const TypeInt* r2 = t2->isa_int();
1362
1363 // If the shift is a constant, just shift the bounds of the type.
1364 // For example, if the shift is 31/63, we just propagate sign bits.
1365 if (!r1->is_con() && r2->is_con()) {
1366 uint shift = r2->get_con();
1367 shift &= bits_per_java_integer(bt) - 1; // semantics of Java shifts
1368 // Shift by a multiple of 32/64 does nothing:
1369 if (shift == 0) {
1370 return t1;
1371 }
1372 // Calculate reasonably aggressive bounds for the result.
1373 // This is necessary if we are to correctly type things
1374 // like (x<<24>>24) == ((byte)x).
1375 jlong lo = r1->lo_as_long() >> (jint)shift;
1376 jlong hi = r1->hi_as_long() >> (jint)shift;
1377 assert(lo <= hi, "must have valid bounds");
1378 #ifdef ASSERT
1379 if (bt == T_INT) {
1380 jint lo_verify = checked_cast<jint>(r1->lo_as_long()) >> (jint)shift;
1381 jint hi_verify = checked_cast<jint>(r1->hi_as_long()) >> (jint)shift;
1382 assert((checked_cast<jint>(lo) == lo_verify) && (checked_cast<jint>(hi) == hi_verify), "inconsistent");
1383 }
1384 #endif
1385 const TypeInteger* ti = TypeInteger::make(lo, hi, MAX2(r1->_widen,r2->_widen), bt);
1386 #ifdef ASSERT
1387 // Make sure we get the sign-capture idiom correct.
1388 if (shift == bits_per_java_integer(bt) - 1) {
1389 if (r1->lo_as_long() >= 0) {
1390 assert(ti == TypeInteger::zero(bt), ">>31/63 of + is 0");
1391 }
1392 if (r1->hi_as_long() < 0) {
1393 assert(ti == TypeInteger::minus_1(bt), ">>31/63 of - is -1");
1394 }
1395 }
1396 #endif
1397 return ti;
1398 }
1399
1400 if (!r1->is_con() || !r2->is_con()) {
1401 // If the left input is non-negative the result must also be non-negative, regardless of what the right input is.
1402 if (r1->lo_as_long() >= 0) {
1403 return TypeInteger::make(0, r1->hi_as_long(), MAX2(r1->_widen, r2->_widen), bt);
1404 }
1405
1406 // Conversely, if the left input is negative then the result must be negative.
1407 if (r1->hi_as_long() <= -1) {
1408 return TypeInteger::make(r1->lo_as_long(), -1, MAX2(r1->_widen, r2->_widen), bt);
1409 }
1410
1411 return TypeInteger::bottom(bt);
1412 }
1413
1414 // Signed shift right
1415 return TypeInteger::make(r1->get_con_as_long(bt) >> (r2->get_con() & (bits_per_java_integer(bt) - 1)), bt);
1416 }
1417
1418 const Type* RShiftINode::Value(PhaseGVN* phase) const {
1419 return ValueIL(phase, T_INT);
1420 }
1421
1422 //=============================================================================
1423 //------------------------------Identity---------------------------------------
1424 Node* RShiftLNode::Identity(PhaseGVN* phase) {
1425 return IdentityIL(phase, T_LONG);
1426 }
1427
1428 Node* RShiftLNode::Ideal(PhaseGVN *phase, bool can_reshape) {
1429 Node* progress = IdealIL(phase, can_reshape, T_LONG);
1430 if (progress == NodeSentinel) {
1431 return nullptr;
1432 }
1433 return progress;
1434 }
1435
1436 const Type* RShiftLNode::Value(PhaseGVN* phase) const {
1437 return ValueIL(phase, T_LONG);
1438 }
1439
1440 URShiftNode* URShiftNode::make(Node* in1, Node* in2, BasicType bt) {
1441 switch (bt) {
1442 case T_INT:
1443 return new URShiftINode(in1, in2);
1444 case T_LONG:
1445 return new URShiftLNode(in1, in2);
1446 default:
1447 fatal("Not implemented for %s", type2name(bt));
1448 }
1449 return nullptr;
1450 }
1451
1452 //=============================================================================
1453 //------------------------------Identity---------------------------------------
1454 Node* URShiftINode::Identity(PhaseGVN* phase) {
1455 int count = 0;
1456 if (const_shift_count(phase, this, &count) && (count & (BitsPerJavaInteger - 1)) == 0) {
1457 // Shift by a multiple of 32 does nothing
1458 return in(1);
1459 }
1460
1461 // Check for "((x << LogBytesPerWord) + (wordSize-1)) >> LogBytesPerWord" which is just "x".
1462 // Happens during new-array length computation.
1463 // Safe if 'x' is in the range [0..(max_int>>LogBytesPerWord)]
1464 Node *add = in(1);
1465 if (add->Opcode() == Op_AddI) {
1466 const TypeInt *t2 = phase->type(add->in(2))->isa_int();
1467 if (t2 && t2->is_con(wordSize - 1) &&
1468 add->in(1)->Opcode() == Op_LShiftI) {
1469 // Check that shift_counts are LogBytesPerWord.
1470 Node *lshift_count = add->in(1)->in(2);
1471 const TypeInt *t_lshift_count = phase->type(lshift_count)->isa_int();
1472 if (t_lshift_count && t_lshift_count->is_con(LogBytesPerWord) &&
1473 t_lshift_count == phase->type(in(2))) {
1474 Node *x = add->in(1)->in(1);
1475 const TypeInt *t_x = phase->type(x)->isa_int();
1476 if (t_x != nullptr && 0 <= t_x->_lo && t_x->_hi <= (max_jint>>LogBytesPerWord)) {
1477 return x;
1478 }
1479 }
1480 }
1481 }
1482
1483 return (phase->type(in(2))->higher_equal(TypeInt::ZERO)) ? in(1) : this;
1484 }
1485
1486 //------------------------------Ideal------------------------------------------
1487 Node* URShiftINode::Ideal(PhaseGVN* phase, bool can_reshape) {
1488 uint con;
1489 Node* progress = mask_and_replace_shift_amount(phase, this, BitsPerJavaInteger, con);
1490 if (con == 0) {
1491 return nullptr;
1492 }
1493
1494 // We'll be wanting the right-shift amount as a mask of that many bits
1495 const int mask = right_n_bits(BitsPerJavaInteger - con);
1496
1497 int in1_op = in(1)->Opcode();
1498
1499 // Check for ((x>>>a)>>>b) and replace with (x>>>(a+b)) when a+b < 32
1500 if( in1_op == Op_URShiftI ) {
1501 const TypeInt *t12 = phase->type( in(1)->in(2) )->isa_int();
1502 if( t12 && t12->is_con() ) { // Right input is a constant
1503 assert( in(1) != in(1)->in(1), "dead loop in URShiftINode::Ideal" );
1504 const int con2 = t12->get_con() & 31; // Shift count is always masked
1505 const int con3 = con+con2;
1506 if( con3 < 32 ) // Only merge shifts if total is < 32
1507 return new URShiftINode( in(1)->in(1), phase->intcon(con3) );
1508 }
1509 }
1510
1511 // Check for ((x << z) + Y) >>> z. Replace with x + con>>>z
1512 // The idiom for rounding to a power of 2 is "(Q+(2^z-1)) >>> z".
1513 // If Q is "X << z" the rounding is useless. Look for patterns like
1514 // ((X<<Z) + Y) >>> Z and replace with (X + Y>>>Z) & Z-mask.
1515 Node *add = in(1);
1516 if (in1_op == Op_AddI) {
1517 Node *lshl = add->in(1);
1518 Node *y = add->in(2);
1519 if (lshl->Opcode() != Op_LShiftI) {
1520 lshl = add->in(2);
1521 y = add->in(1);
1522 }
1523 // Compare shift counts by value, not by node pointer, to also match a not-yet-normalized
1524 // negative constant (e.g. -1 vs 31)
1525 int lshl_con = 0;
1526 if (lshl->Opcode() == Op_LShiftI &&
1527 const_shift_count(phase, lshl, &lshl_con) &&
1528 (lshl_con & (BitsPerJavaInteger - 1)) == con) {
1529 Node *y_z = phase->transform(new URShiftINode(y, in(2)));
1530 Node *sum = phase->transform(new AddINode(lshl->in(1), y_z));
1531 return new AndINode(sum, phase->intcon(mask));
1532 }
1533 }
1534
1535 // Check for (x & mask) >>> z. Replace with (x >>> z) & (mask >>> z)
1536 // This shortens the mask. Also, if we are extracting a high byte and
1537 // storing it to a buffer, the mask will be removed completely.
1538 Node *andi = in(1);
1539 if( in1_op == Op_AndI ) {
1540 const TypeInt *t3 = phase->type( andi->in(2) )->isa_int();
1541 if( t3 && t3->is_con() ) { // Right input is a constant
1542 jint mask2 = t3->get_con();
1543 mask2 >>= con; // *signed* shift downward (high-order zeroes do not help)
1544 Node *newshr = phase->transform( new URShiftINode(andi->in(1), in(2)) );
1545 return new AndINode(newshr, phase->intcon(mask2));
1546 // The negative values are easier to materialize than positive ones.
1547 // A typical case from address arithmetic is ((x & ~15) >> 4).
1548 // It's better to change that to ((x >> 4) & ~0) versus
1549 // ((x >> 4) & 0x0FFFFFFF). The difference is greatest in LP64.
1550 }
1551 }
1552
1553 // Check for "(X << z ) >>> z" which simply zero-extends
1554 Node *shl = in(1);
1555 // Compare shift counts by value, not by node pointer, to also match a not-yet-normalized
1556 // negative constant (e.g. -1 vs 31)
1557 int shl_con = 0;
1558 if (in1_op == Op_LShiftI &&
1559 const_shift_count(phase, shl, &shl_con) &&
1560 (shl_con & (BitsPerJavaInteger - 1)) == con)
1561 return new AndINode(shl->in(1), phase->intcon(mask));
1562
1563 // Check for (x >> n) >>> 31. Replace with (x >>> 31)
1564 const TypeInt* t2 = phase->type(in(2))->isa_int();
1565 Node *shr = in(1);
1566 if ( in1_op == Op_RShiftI ) {
1567 Node *in11 = shr->in(1);
1568 Node *in12 = shr->in(2);
1569 const TypeInt *t11 = phase->type(in11)->isa_int();
1570 const TypeInt *t12 = phase->type(in12)->isa_int();
1571 if ( t11 && t2 && t2->is_con(31) && t12 && t12->is_con() ) {
1572 return new URShiftINode(in11, phase->intcon(31));
1573 }
1574 }
1575
1576 return progress;
1577 }
1578
1579 //------------------------------Value------------------------------------------
1580 // A URShiftINode shifts its input2 right by input1 amount.
1581 const Type* URShiftINode::Value(PhaseGVN* phase) const {
1582 // (This is a near clone of RShiftINode::Value.)
1583 const Type *t1 = phase->type( in(1) );
1584 const Type *t2 = phase->type( in(2) );
1585 // Either input is TOP ==> the result is TOP
1586 if( t1 == Type::TOP ) return Type::TOP;
1587 if( t2 == Type::TOP ) return Type::TOP;
1588
1589 // Left input is ZERO ==> the result is ZERO.
1590 if( t1 == TypeInt::ZERO ) return TypeInt::ZERO;
1591 // Shift by zero does nothing
1592 if( t2 == TypeInt::ZERO ) return t1;
1593
1594 // Either input is BOTTOM ==> the result is BOTTOM
1595 if (t1 == Type::BOTTOM || t2 == Type::BOTTOM)
1596 return TypeInt::INT;
1597
1598 if (t2 == TypeInt::INT)
1599 return TypeInt::INT;
1600
1601 const TypeInt *r1 = t1->is_int(); // Handy access
1602 const TypeInt *r2 = t2->is_int(); // Handy access
1603
1604 if (r2->is_con()) {
1605 uint shift = r2->get_con();
1606 shift &= BitsPerJavaInteger-1; // semantics of Java shifts
1607 // Shift by a multiple of 32 does nothing:
1608 if (shift == 0) return t1;
1609 // Calculate reasonably aggressive bounds for the result.
1610 jint lo = (juint)r1->_lo >> (juint)shift;
1611 jint hi = (juint)r1->_hi >> (juint)shift;
1612 if (r1->_hi >= 0 && r1->_lo < 0) {
1613 // If the type has both negative and positive values,
1614 // there are two separate sub-domains to worry about:
1615 // The positive half and the negative half.
1616 jint neg_lo = lo;
1617 jint neg_hi = (juint)-1 >> (juint)shift;
1618 jint pos_lo = (juint) 0 >> (juint)shift;
1619 jint pos_hi = hi;
1620 lo = MIN2(neg_lo, pos_lo); // == 0
1621 hi = MAX2(neg_hi, pos_hi); // == -1 >>> shift;
1622 }
1623 assert(lo <= hi, "must have valid bounds");
1624 const TypeInt* ti = TypeInt::make(lo, hi, MAX2(r1->_widen,r2->_widen));
1625 #ifdef ASSERT
1626 // Make sure we get the sign-capture idiom correct.
1627 if (shift == BitsPerJavaInteger-1) {
1628 if (r1->_lo >= 0) assert(ti == TypeInt::ZERO, ">>>31 of + is 0");
1629 if (r1->_hi < 0) assert(ti == TypeInt::ONE, ">>>31 of - is +1");
1630 }
1631 #endif
1632 return ti;
1633 }
1634
1635 //
1636 // Do not support shifted oops in info for GC
1637 //
1638 // else if( t1->base() == Type::InstPtr ) {
1639 //
1640 // const TypeInstPtr *o = t1->is_instptr();
1641 // if( t1->singleton() )
1642 // return TypeInt::make( ((uint32_t)o->const_oop() + o->_offset) >> shift );
1643 // }
1644 // else if( t1->base() == Type::KlassPtr ) {
1645 // const TypeKlassPtr *o = t1->is_klassptr();
1646 // if( t1->singleton() )
1647 // return TypeInt::make( ((uint32_t)o->const_oop() + o->_offset) >> shift );
1648 // }
1649
1650 return TypeInt::INT;
1651 }
1652
1653 //=============================================================================
1654 //------------------------------Identity---------------------------------------
1655 Node* URShiftLNode::Identity(PhaseGVN* phase) {
1656 int count = 0;
1657 if (const_shift_count(phase, this, &count) && (count & (BitsPerJavaLong - 1)) == 0) {
1658 // Shift by a multiple of 64 does nothing
1659 return in(1);
1660 }
1661 return this;
1662 }
1663
1664 //------------------------------Ideal------------------------------------------
1665 Node* URShiftLNode::Ideal(PhaseGVN* phase, bool can_reshape) {
1666 uint con;
1667 Node* progress = mask_and_replace_shift_amount(phase, this, BitsPerJavaLong, con);
1668 if (con == 0) {
1669 return nullptr;
1670 }
1671
1672 // We'll be wanting the right-shift amount as a mask of that many bits
1673 const jlong mask = jlong(max_julong >> con);
1674
1675 // Check for ((x << z) + Y) >>> z. Replace with x + con>>>z
1676 // The idiom for rounding to a power of 2 is "(Q+(2^z-1)) >>> z".
1677 // If Q is "X << z" the rounding is useless. Look for patterns like
1678 // ((X<<Z) + Y) >>> Z and replace with (X + Y>>>Z) & Z-mask.
1679 Node *add = in(1);
1680 const TypeInt *t2 = phase->type(in(2))->isa_int();
1681 if (add->Opcode() == Op_AddL) {
1682 Node *lshl = add->in(1);
1683 Node *y = add->in(2);
1684 if (lshl->Opcode() != Op_LShiftL) {
1685 lshl = add->in(2);
1686 y = add->in(1);
1687 }
1688 // Compare shift counts by value, not by node pointer, to also match a not-yet-normalized
1689 // negative constant (e.g. -1 vs 63)
1690 int lshl_con = 0;
1691 if (lshl->Opcode() == Op_LShiftL &&
1692 const_shift_count(phase, lshl, &lshl_con) &&
1693 (lshl_con & (BitsPerJavaLong - 1)) == con) {
1694 Node* y_z = phase->transform(new URShiftLNode(y, in(2)));
1695 Node* sum = phase->transform(new AddLNode(lshl->in(1), y_z));
1696 return new AndLNode(sum, phase->longcon(mask));
1697 }
1698 }
1699
1700 // Check for (x & mask) >>> z. Replace with (x >>> z) & (mask >>> z)
1701 // This shortens the mask. Also, if we are extracting a high byte and
1702 // storing it to a buffer, the mask will be removed completely.
1703 Node *andi = in(1);
1704 if( andi->Opcode() == Op_AndL ) {
1705 const TypeLong *t3 = phase->type( andi->in(2) )->isa_long();
1706 if( t3 && t3->is_con() ) { // Right input is a constant
1707 jlong mask2 = t3->get_con();
1708 mask2 >>= con; // *signed* shift downward (high-order zeroes do not help)
1709 Node *newshr = phase->transform( new URShiftLNode(andi->in(1), in(2)) );
1710 return new AndLNode(newshr, phase->longcon(mask2));
1711 }
1712 }
1713
1714 // Check for "(X << z ) >>> z" which simply zero-extends
1715 Node *shl = in(1);
1716 // Compare shift counts by value, not by node pointer, to also match a not-yet-normalized
1717 // negative constant (e.g. -1 vs 63)
1718 int shl_con = 0;
1719 if (shl->Opcode() == Op_LShiftL &&
1720 const_shift_count(phase, shl, &shl_con) &&
1721 (shl_con & (BitsPerJavaLong - 1)) == con) {
1722 return new AndLNode(shl->in(1), phase->longcon(mask));
1723 }
1724
1725 // Check for (x >> n) >>> 63. Replace with (x >>> 63)
1726 Node *shr = in(1);
1727 if ( shr->Opcode() == Op_RShiftL ) {
1728 Node *in11 = shr->in(1);
1729 Node *in12 = shr->in(2);
1730 const TypeLong *t11 = phase->type(in11)->isa_long();
1731 const TypeInt *t12 = phase->type(in12)->isa_int();
1732 if ( t11 && t2 && t2->is_con(63) && t12 && t12->is_con() ) {
1733 return new URShiftLNode(in11, phase->intcon(63));
1734 }
1735 }
1736
1737 return progress;
1738 }
1739
1740 //------------------------------Value------------------------------------------
1741 // A URShiftINode shifts its input2 right by input1 amount.
1742 const Type* URShiftLNode::Value(PhaseGVN* phase) const {
1743 // (This is a near clone of RShiftLNode::Value.)
1744 const Type *t1 = phase->type( in(1) );
1745 const Type *t2 = phase->type( in(2) );
1746 // Either input is TOP ==> the result is TOP
1747 if( t1 == Type::TOP ) return Type::TOP;
1748 if( t2 == Type::TOP ) return Type::TOP;
1749
1750 // Left input is ZERO ==> the result is ZERO.
1751 if( t1 == TypeLong::ZERO ) return TypeLong::ZERO;
1752 // Shift by zero does nothing
1753 if( t2 == TypeInt::ZERO ) return t1;
1754
1755 // Either input is BOTTOM ==> the result is BOTTOM
1756 if (t1 == Type::BOTTOM || t2 == Type::BOTTOM)
1757 return TypeLong::LONG;
1758
1759 if (t2 == TypeInt::INT)
1760 return TypeLong::LONG;
1761
1762 const TypeLong *r1 = t1->is_long(); // Handy access
1763 const TypeInt *r2 = t2->is_int (); // Handy access
1764
1765 if (r2->is_con()) {
1766 uint shift = r2->get_con();
1767 shift &= BitsPerJavaLong - 1; // semantics of Java shifts
1768 // Shift by a multiple of 64 does nothing:
1769 if (shift == 0) return t1;
1770 // Calculate reasonably aggressive bounds for the result.
1771 jlong lo = (julong)r1->_lo >> (juint)shift;
1772 jlong hi = (julong)r1->_hi >> (juint)shift;
1773 if (r1->_hi >= 0 && r1->_lo < 0) {
1774 // If the type has both negative and positive values,
1775 // there are two separate sub-domains to worry about:
1776 // The positive half and the negative half.
1777 jlong neg_lo = lo;
1778 jlong neg_hi = (julong)-1 >> (juint)shift;
1779 jlong pos_lo = (julong) 0 >> (juint)shift;
1780 jlong pos_hi = hi;
1781 //lo = MIN2(neg_lo, pos_lo); // == 0
1782 lo = neg_lo < pos_lo ? neg_lo : pos_lo;
1783 //hi = MAX2(neg_hi, pos_hi); // == -1 >>> shift;
1784 hi = neg_hi > pos_hi ? neg_hi : pos_hi;
1785 }
1786 assert(lo <= hi, "must have valid bounds");
1787 const TypeLong* tl = TypeLong::make(lo, hi, MAX2(r1->_widen,r2->_widen));
1788 #ifdef ASSERT
1789 // Make sure we get the sign-capture idiom correct.
1790 if (shift == BitsPerJavaLong - 1) {
1791 if (r1->_lo >= 0) assert(tl == TypeLong::ZERO, ">>>63 of + is 0");
1792 if (r1->_hi < 0) assert(tl == TypeLong::ONE, ">>>63 of - is +1");
1793 }
1794 #endif
1795 return tl;
1796 }
1797
1798 return TypeLong::LONG; // Give up
1799 }
1800
1801 //=============================================================================
1802 //------------------------------Ideal------------------------------------------
1803 Node* FmaNode::Ideal(PhaseGVN* phase, bool can_reshape) {
1804 // We canonicalize the node by converting "(-a)*b+c" into "b*(-a)+c"
1805 // This reduces the number of rules in the matcher, as we only need to check
1806 // for negations on the second argument, and not the symmetric case where
1807 // the first argument is negated.
1808 if (in(1)->is_Neg() && !in(2)->is_Neg()) {
1809 swap_edges(1, 2);
1810 return this;
1811 }
1812 return nullptr;
1813 }
1814
1815 //=============================================================================
1816 //------------------------------Value------------------------------------------
1817 const Type* FmaDNode::Value(PhaseGVN* phase) const {
1818 const Type *t1 = phase->type(in(1));
1819 if (t1 == Type::TOP) return Type::TOP;
1820 if (t1->base() != Type::DoubleCon) return Type::DOUBLE;
1821 const Type *t2 = phase->type(in(2));
1822 if (t2 == Type::TOP) return Type::TOP;
1823 if (t2->base() != Type::DoubleCon) return Type::DOUBLE;
1824 const Type *t3 = phase->type(in(3));
1825 if (t3 == Type::TOP) return Type::TOP;
1826 if (t3->base() != Type::DoubleCon) return Type::DOUBLE;
1827 #ifndef __STDC_IEC_559__
1828 return Type::DOUBLE;
1829 #else
1830 double d1 = t1->getd();
1831 double d2 = t2->getd();
1832 double d3 = t3->getd();
1833 return TypeD::make(fma(d1, d2, d3));
1834 #endif
1835 }
1836
1837 //=============================================================================
1838 //------------------------------Value------------------------------------------
1839 const Type* FmaFNode::Value(PhaseGVN* phase) const {
1840 const Type *t1 = phase->type(in(1));
1841 if (t1 == Type::TOP) return Type::TOP;
1842 if (t1->base() != Type::FloatCon) return Type::FLOAT;
1843 const Type *t2 = phase->type(in(2));
1844 if (t2 == Type::TOP) return Type::TOP;
1845 if (t2->base() != Type::FloatCon) return Type::FLOAT;
1846 const Type *t3 = phase->type(in(3));
1847 if (t3 == Type::TOP) return Type::TOP;
1848 if (t3->base() != Type::FloatCon) return Type::FLOAT;
1849 #ifndef __STDC_IEC_559__
1850 return Type::FLOAT;
1851 #else
1852 float f1 = t1->getf();
1853 float f2 = t2->getf();
1854 float f3 = t3->getf();
1855 return TypeF::make(fma(f1, f2, f3));
1856 #endif
1857 }
1858
1859 //=============================================================================
1860 //------------------------------Value------------------------------------------
1861 const Type* FmaHFNode::Value(PhaseGVN* phase) const {
1862 const Type* t1 = phase->type(in(1));
1863 if (t1 == Type::TOP) { return Type::TOP; }
1864 if (t1->base() != Type::HalfFloatCon) { return Type::HALF_FLOAT; }
1865 const Type* t2 = phase->type(in(2));
1866 if (t2 == Type::TOP) { return Type::TOP; }
1867 if (t2->base() != Type::HalfFloatCon) { return Type::HALF_FLOAT; }
1868 const Type* t3 = phase->type(in(3));
1869 if (t3 == Type::TOP) { return Type::TOP; }
1870 if (t3->base() != Type::HalfFloatCon) { return Type::HALF_FLOAT; }
1871 #ifndef __STDC_IEC_559__
1872 return Type::HALF_FLOAT;
1873 #else
1874 float f1 = t1->getf();
1875 float f2 = t2->getf();
1876 float f3 = t3->getf();
1877 return TypeH::make(fma(f1, f2, f3));
1878 #endif
1879 }
1880
1881 //=============================================================================
1882 //------------------------------hash-------------------------------------------
1883 // Hash function for MulAddS2INode. Operation is commutative with commutative pairs.
1884 // The hash function must return the same value when edge swapping is performed.
1885 uint MulAddS2INode::hash() const {
1886 return (uintptr_t)in(1) + (uintptr_t)in(2) + (uintptr_t)in(3) + (uintptr_t)in(4) + Opcode();
1887 }
1888
1889 //------------------------------Rotate Operations ------------------------------
1890
1891 Node* RotateLeftNode::Identity(PhaseGVN* phase) {
1892 const Type* t1 = phase->type(in(1));
1893 if (t1 == Type::TOP) {
1894 return this;
1895 }
1896 int count = 0;
1897 assert(t1->isa_int() || t1->isa_long(), "Unexpected type");
1898 int mask = (t1->isa_int() ? BitsPerJavaInteger : BitsPerJavaLong) - 1;
1899 if (const_shift_count(phase, this, &count) && (count & mask) == 0) {
1900 // Rotate by a multiple of 32/64 does nothing
1901 return in(1);
1902 }
1903 return this;
1904 }
1905
1906 const Type* RotateLeftNode::Value(PhaseGVN* phase) const {
1907 const Type* t1 = phase->type(in(1));
1908 const Type* t2 = phase->type(in(2));
1909 // Either input is TOP ==> the result is TOP
1910 if (t1 == Type::TOP || t2 == Type::TOP) {
1911 return Type::TOP;
1912 }
1913
1914 if (t1->isa_int()) {
1915 const TypeInt* r1 = t1->is_int();
1916 const TypeInt* r2 = t2->is_int();
1917
1918 // Left input is ZERO ==> the result is ZERO.
1919 if (r1 == TypeInt::ZERO) {
1920 return TypeInt::ZERO;
1921 }
1922 // Rotate by zero does nothing
1923 if (r2 == TypeInt::ZERO) {
1924 return r1;
1925 }
1926 if (r1->is_con() && r2->is_con()) {
1927 juint r1_con = (juint)r1->get_con();
1928 juint shift = (juint)(r2->get_con()) & (juint)(BitsPerJavaInteger - 1); // semantics of Java shifts
1929 return TypeInt::make((r1_con << shift) | (r1_con >> (32 - shift)));
1930 }
1931 return TypeInt::INT;
1932 } else {
1933 assert(t1->isa_long(), "Type must be a long");
1934 const TypeLong* r1 = t1->is_long();
1935 const TypeInt* r2 = t2->is_int();
1936
1937 // Left input is ZERO ==> the result is ZERO.
1938 if (r1 == TypeLong::ZERO) {
1939 return TypeLong::ZERO;
1940 }
1941 // Rotate by zero does nothing
1942 if (r2 == TypeInt::ZERO) {
1943 return r1;
1944 }
1945 if (r1->is_con() && r2->is_con()) {
1946 julong r1_con = (julong)r1->get_con();
1947 julong shift = (julong)(r2->get_con()) & (julong)(BitsPerJavaLong - 1); // semantics of Java shifts
1948 return TypeLong::make((r1_con << shift) | (r1_con >> (64 - shift)));
1949 }
1950 return TypeLong::LONG;
1951 }
1952 }
1953
1954 Node* RotateLeftNode::Ideal(PhaseGVN *phase, bool can_reshape) {
1955 const Type* t1 = phase->type(in(1));
1956 const Type* t2 = phase->type(in(2));
1957 if (t2->isa_int() && t2->is_int()->is_con()) {
1958 if (t1->isa_int()) {
1959 int lshift = t2->is_int()->get_con() & 31;
1960 return new RotateRightNode(in(1), phase->intcon(32 - (lshift & 31)), TypeInt::INT);
1961 } else if (t1 != Type::TOP) {
1962 assert(t1->isa_long(), "Type must be a long");
1963 int lshift = t2->is_int()->get_con() & 63;
1964 return new RotateRightNode(in(1), phase->intcon(64 - (lshift & 63)), TypeLong::LONG);
1965 }
1966 }
1967 return nullptr;
1968 }
1969
1970 Node* RotateRightNode::Identity(PhaseGVN* phase) {
1971 const Type* t1 = phase->type(in(1));
1972 if (t1 == Type::TOP) {
1973 return this;
1974 }
1975 int count = 0;
1976 assert(t1->isa_int() || t1->isa_long(), "Unexpected type");
1977 int mask = (t1->isa_int() ? BitsPerJavaInteger : BitsPerJavaLong) - 1;
1978 if (const_shift_count(phase, this, &count) && (count & mask) == 0) {
1979 // Rotate by a multiple of 32/64 does nothing
1980 return in(1);
1981 }
1982 return this;
1983 }
1984
1985 const Type* RotateRightNode::Value(PhaseGVN* phase) const {
1986 const Type* t1 = phase->type(in(1));
1987 const Type* t2 = phase->type(in(2));
1988 // Either input is TOP ==> the result is TOP
1989 if (t1 == Type::TOP || t2 == Type::TOP) {
1990 return Type::TOP;
1991 }
1992
1993 if (t1->isa_int()) {
1994 const TypeInt* r1 = t1->is_int();
1995 const TypeInt* r2 = t2->is_int();
1996
1997 // Left input is ZERO ==> the result is ZERO.
1998 if (r1 == TypeInt::ZERO) {
1999 return TypeInt::ZERO;
2000 }
2001 // Rotate by zero does nothing
2002 if (r2 == TypeInt::ZERO) {
2003 return r1;
2004 }
2005 if (r1->is_con() && r2->is_con()) {
2006 juint r1_con = (juint)r1->get_con();
2007 juint shift = (juint)(r2->get_con()) & (juint)(BitsPerJavaInteger - 1); // semantics of Java shifts
2008 return TypeInt::make((r1_con >> shift) | (r1_con << (32 - shift)));
2009 }
2010 return TypeInt::INT;
2011 } else {
2012 assert(t1->isa_long(), "Type must be a long");
2013 const TypeLong* r1 = t1->is_long();
2014 const TypeInt* r2 = t2->is_int();
2015 // Left input is ZERO ==> the result is ZERO.
2016 if (r1 == TypeLong::ZERO) {
2017 return TypeLong::ZERO;
2018 }
2019 // Rotate by zero does nothing
2020 if (r2 == TypeInt::ZERO) {
2021 return r1;
2022 }
2023 if (r1->is_con() && r2->is_con()) {
2024 julong r1_con = (julong)r1->get_con();
2025 julong shift = (julong)(r2->get_con()) & (julong)(BitsPerJavaLong - 1); // semantics of Java shifts
2026 return TypeLong::make((r1_con >> shift) | (r1_con << (64 - shift)));
2027 }
2028 return TypeLong::LONG;
2029 }
2030 }
2031
2032 //------------------------------ Sum & Mask ------------------------------
2033
2034 // Returns a lower bound on the number of trailing zeros in expr.
2035 static jint AndIL_min_trailing_zeros(const PhaseGVN* phase, const Node* expr, BasicType bt) {
2036 const TypeInteger* type = phase->type(expr)->isa_integer(bt);
2037 if (type == nullptr) {
2038 return 0;
2039 }
2040
2041 expr = expr->uncast();
2042 type = phase->type(expr)->isa_integer(bt);
2043 if (type == nullptr) {
2044 return 0;
2045 }
2046
2047 if (type->is_con()) {
2048 jlong con = type->get_con_as_long(bt);
2049 return con == 0L ? (type2aelembytes(bt) * BitsPerByte) : count_trailing_zeros(con);
2050 }
2051
2052 if (expr->Opcode() == Op_ConvI2L) {
2053 expr = expr->in(1)->uncast();
2054 bt = T_INT;
2055 type = phase->type(expr)->isa_int();
2056 }
2057
2058 // Pattern: expr = (x << shift)
2059 if (expr->Opcode() == Op_LShift(bt)) {
2060 const TypeInt* shift_t = phase->type(expr->in(2))->isa_int();
2061 if (shift_t == nullptr || !shift_t->is_con()) {
2062 return 0;
2063 }
2064 // We need to truncate the shift, as it may not have been canonicalized yet.
2065 // T_INT: 0..31 -> shift_mask = 4 * 8 - 1 = 31
2066 // T_LONG: 0..63 -> shift_mask = 8 * 8 - 1 = 63
2067 // (JLS: "Shift Operators")
2068 jint shift_mask = type2aelembytes(bt) * BitsPerByte - 1;
2069 return shift_t->get_con() & shift_mask;
2070 }
2071
2072 return 0;
2073 }
2074
2075 // Checks whether expr is neutral additive element (zero) under mask,
2076 // i.e. whether an expression of the form:
2077 // (AndX (AddX (expr addend) mask)
2078 // (expr + addend) & mask
2079 // is equivalent to
2080 // (AndX addend mask)
2081 // addend & mask
2082 // for any addend.
2083 // (The X in AndX must be I or L, depending on bt).
2084 //
2085 // We check for the sufficient condition when the lowest set bit in expr is higher than
2086 // the highest set bit in mask, i.e.:
2087 // expr: eeeeee0000000000000
2088 // mask: 000000mmmmmmmmmmmmm
2089 // <--w bits--->
2090 // We do not test for other cases.
2091 //
2092 // Correctness:
2093 // Given "expr" with at least "w" trailing zeros,
2094 // let "mod = 2^w", "suffix_mask = mod - 1"
2095 //
2096 // Since "mask" only has bits set where "suffix_mask" does, we have:
2097 // mask = suffix_mask & mask (SUFFIX_MASK)
2098 //
2099 // And since expr only has bits set above w, and suffix_mask only below:
2100 // expr & suffix_mask == 0 (NO_BIT_OVERLAP)
2101 //
2102 // From unsigned modular arithmetic (with unsigned modulo %), and since mod is
2103 // a power of 2, and we are computing in a ring of powers of 2, we know that
2104 // (x + y) % mod = (x % mod + y) % mod
2105 // (x + y) & suffix_mask = (x & suffix_mask + y) & suffix_mask (MOD_ARITH)
2106 //
2107 // We can now prove the equality:
2108 // (expr + addend) & mask
2109 // = (expr + addend) & suffix_mask & mask (SUFFIX_MASK)
2110 // = (expr & suffix_mask + addend) & suffix_mask & mask (MOD_ARITH)
2111 // = (0 + addend) & suffix_mask & mask (NO_BIT_OVERLAP)
2112 // = addend & mask (SUFFIX_MASK)
2113 //
2114 // Hence, an expr with at least w trailing zeros is a neutral additive element under any mask with bit width w.
2115 static bool AndIL_is_zero_element_under_mask(const PhaseGVN* phase, const Node* expr, const Node* mask, BasicType bt) {
2116 // When the mask is negative, it has the most significant bit set.
2117 const TypeInteger* mask_t = phase->type(mask)->isa_integer(bt);
2118 if (mask_t == nullptr || mask_t->lo_as_long() < 0) {
2119 return false;
2120 }
2121
2122 // When the mask is constant zero, we defer to MulNode::Value to eliminate the entire AndX operation.
2123 if (mask_t->hi_as_long() == 0) {
2124 assert(mask_t->lo_as_long() == 0, "checked earlier");
2125 return false;
2126 }
2127
2128 jint mask_bit_width = BitsPerLong - count_leading_zeros(mask_t->hi_as_long());
2129 jint expr_trailing_zeros = AndIL_min_trailing_zeros(phase, expr, bt);
2130 return expr_trailing_zeros >= mask_bit_width;
2131 }
2132
2133 // Reduces the pattern:
2134 // (AndX (AddX add1 add2) mask)
2135 // to
2136 // (AndX add1 mask), if add2 is neutral wrt mask (see above), and vice versa.
2137 Node* MulNode::AndIL_sum_and_mask(PhaseGVN* phase, BasicType bt) {
2138 Node* add = in(1);
2139 Node* mask = in(2);
2140 int addidx = 0;
2141 if (add->Opcode() == Op_Add(bt)) {
2142 addidx = 1;
2143 } else if (mask->Opcode() == Op_Add(bt)) {
2144 mask = add;
2145 addidx = 2;
2146 add = in(addidx);
2147 }
2148 if (addidx > 0) {
2149 Node* add1 = add->in(1);
2150 Node* add2 = add->in(2);
2151 if (AndIL_is_zero_element_under_mask(phase, add1, mask, bt)) {
2152 set_req_X(addidx, add2, phase);
2153 return this;
2154 } else if (AndIL_is_zero_element_under_mask(phase, add2, mask, bt)) {
2155 set_req_X(addidx, add1, phase);
2156 return this;
2157 }
2158 }
2159 return nullptr;
2160 }