1 /*
  2  * Copyright (c) 1999, 2023, Oracle and/or its affiliates. All rights reserved.
  3  * Copyright (c) 2012, 2018 SAP SE. All rights reserved.
  4  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
  5  *
  6  * This code is free software; you can redistribute it and/or modify it
  7  * under the terms of the GNU General Public License version 2 only, as
  8  * published by the Free Software Foundation.
  9  *
 10  * This code is distributed in the hope that it will be useful, but WITHOUT
 11  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
 12  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
 13  * version 2 for more details (a copy is included in the LICENSE file that
 14  * accompanied this code).
 15  *
 16  * You should have received a copy of the GNU General Public License version
 17  * 2 along with this work; if not, write to the Free Software Foundation,
 18  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
 19  *
 20  * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
 21  * or visit www.oracle.com if you need additional information or have any
 22  * questions.
 23  *
 24  */
 25 
 26 #include "precompiled.hpp"
 27 #include "asm/macroAssembler.inline.hpp"
 28 #include "c1/c1_MacroAssembler.hpp"
 29 #include "c1/c1_Runtime1.hpp"
 30 #include "gc/shared/collectedHeap.hpp"
 31 #include "gc/shared/tlab_globals.hpp"
 32 #include "interpreter/interpreter.hpp"
 33 #include "oops/arrayOop.hpp"
 34 #include "oops/markWord.hpp"
 35 #include "runtime/basicLock.hpp"
 36 #include "runtime/os.hpp"
 37 #include "runtime/sharedRuntime.hpp"
 38 #include "runtime/stubRoutines.hpp"
 39 #include "utilities/align.hpp"
 40 #include "utilities/macros.hpp"
 41 #include "utilities/powerOfTwo.hpp"
 42 
 43 void C1_MacroAssembler::inline_cache_check(Register receiver, Register iCache) {
 44   const Register temp_reg = R12_scratch2;
 45   Label Lmiss;
 46 
 47   verify_oop(receiver, FILE_AND_LINE);
 48   load_klass_check_null(temp_reg, receiver, &Lmiss);
 49 
 50   if (TrapBasedICMissChecks && TrapBasedNullChecks) {
 51     trap_ic_miss_check(temp_reg, iCache);
 52   } else {
 53     Label Lok;
 54     cmpd(CCR0, temp_reg, iCache);
 55     beq(CCR0, Lok);
 56     bind(Lmiss);
 57     //load_const_optimized(temp_reg, SharedRuntime::get_ic_miss_stub(), R0);
 58     calculate_address_from_global_toc(temp_reg, SharedRuntime::get_ic_miss_stub(), true, true, false);
 59     mtctr(temp_reg);
 60     bctr();
 61     align(32, 12);
 62     bind(Lok);
 63   }
 64 }
 65 
 66 
 67 void C1_MacroAssembler::explicit_null_check(Register base) {
 68   Unimplemented();
 69 }
 70 
 71 
 72 void C1_MacroAssembler::build_frame(int frame_size_in_bytes, int bang_size_in_bytes) {
 73   // Avoid stack bang as first instruction. It may get overwritten by patch_verified_entry.
 74   const Register return_pc = R20;
 75   mflr(return_pc);
 76 
 77   // Make sure there is enough stack space for this method's activation.
 78   assert(bang_size_in_bytes >= frame_size_in_bytes, "stack bang size incorrect");
 79   generate_stack_overflow_check(bang_size_in_bytes);
 80 
 81   std(return_pc, _abi0(lr), R1_SP);     // SP->lr = return_pc
 82   push_frame(frame_size_in_bytes, R0); // SP -= frame_size_in_bytes
 83 
 84   BarrierSetAssembler* bs = BarrierSet::barrier_set()->barrier_set_assembler();
 85   bs->nmethod_entry_barrier(this, R20);
 86 }
 87 
 88 
 89 void C1_MacroAssembler::verified_entry(bool breakAtEntry) {
 90   if (breakAtEntry) illtrap();
 91   // build frame
 92 }
 93 
 94 
 95 void C1_MacroAssembler::lock_object(Register Rmark, Register Roop, Register Rbox, Register Rscratch, Label& slow_case) {
 96   assert_different_registers(Rmark, Roop, Rbox, Rscratch);
 97 
 98   Label done, cas_failed, slow_int;
 99 
100   // The following move must be the first instruction of emitted since debug
101   // information may be generated for it.
102   // Load object header.
103   ld(Rmark, oopDesc::mark_offset_in_bytes(), Roop);
104 
105   verify_oop(Roop, FILE_AND_LINE);
106 
107   // Save object being locked into the BasicObjectLock...
108   std(Roop, in_bytes(BasicObjectLock::obj_offset()), Rbox);
109 
110   if (DiagnoseSyncOnValueBasedClasses != 0) {
111     load_klass(Rscratch, Roop);
112     lwz(Rscratch, in_bytes(Klass::access_flags_offset()), Rscratch);
113     testbitdi(CCR0, R0, Rscratch, exact_log2(JVM_ACC_IS_VALUE_BASED_CLASS));
114     bne(CCR0, slow_int);
115   }
116 
117   if (LockingMode == LM_LIGHTWEIGHT) {
118     lightweight_lock(Roop, Rmark, Rscratch, slow_int);
119   } else if (LockingMode == LM_LEGACY) {
120     // ... and mark it unlocked.
121     ori(Rmark, Rmark, markWord::unlocked_value);
122 
123     // Save unlocked object header into the displaced header location on the stack.
124     std(Rmark, BasicLock::displaced_header_offset_in_bytes(), Rbox);
125 
126     // Compare object markWord with Rmark and if equal exchange Rscratch with object markWord.
127     assert(oopDesc::mark_offset_in_bytes() == 0, "cas must take a zero displacement");
128     cmpxchgd(/*flag=*/CCR0,
129              /*current_value=*/Rscratch,
130              /*compare_value=*/Rmark,
131              /*exchange_value=*/Rbox,
132              /*where=*/Roop/*+0==mark_offset_in_bytes*/,
133              MacroAssembler::MemBarRel | MacroAssembler::MemBarAcq,
134              MacroAssembler::cmpxchgx_hint_acquire_lock(),
135              noreg,
136              &cas_failed,
137              /*check without membar and ldarx first*/true);
138     // If compare/exchange succeeded we found an unlocked object and we now have locked it
139     // hence we are done.
140   }
141   b(done);
142 
143   bind(slow_int);
144   b(slow_case); // far
145 
146   if (LockingMode == LM_LEGACY) {
147     bind(cas_failed);
148     // We did not find an unlocked object so see if this is a recursive case.
149     sub(Rscratch, Rscratch, R1_SP);
150     load_const_optimized(R0, (~(os::vm_page_size()-1) | markWord::lock_mask_in_place));
151     and_(R0/*==0?*/, Rscratch, R0);
152     std(R0/*==0, perhaps*/, BasicLock::displaced_header_offset_in_bytes(), Rbox);
153     bne(CCR0, slow_int);
154   }
155 
156   bind(done);
157   inc_held_monitor_count(Rmark /*tmp*/);
158 }
159 
160 
161 void C1_MacroAssembler::unlock_object(Register Rmark, Register Roop, Register Rbox, Label& slow_case) {
162   assert_different_registers(Rmark, Roop, Rbox);
163 
164   Label slow_int, done;
165 
166   Address mark_addr(Roop, oopDesc::mark_offset_in_bytes());
167   assert(mark_addr.disp() == 0, "cas must take a zero displacement");
168 
169   if (LockingMode != LM_LIGHTWEIGHT) {
170     // Test first if it is a fast recursive unlock.
171     ld(Rmark, BasicLock::displaced_header_offset_in_bytes(), Rbox);
172     cmpdi(CCR0, Rmark, 0);
173     beq(CCR0, done);
174   }
175 
176   // Load object.
177   ld(Roop, in_bytes(BasicObjectLock::obj_offset()), Rbox);
178   verify_oop(Roop, FILE_AND_LINE);
179 
180   if (LockingMode == LM_LIGHTWEIGHT) {
181     lightweight_unlock(Roop, Rmark, slow_int);
182   } else if (LockingMode == LM_LEGACY) {
183     // Check if it is still a light weight lock, this is is true if we see
184     // the stack address of the basicLock in the markWord of the object.
185     cmpxchgd(/*flag=*/CCR0,
186              /*current_value=*/R0,
187              /*compare_value=*/Rbox,
188              /*exchange_value=*/Rmark,
189              /*where=*/Roop,
190              MacroAssembler::MemBarRel,
191              MacroAssembler::cmpxchgx_hint_release_lock(),
192              noreg,
193              &slow_int);
194   }
195   b(done);
196   bind(slow_int);
197   b(slow_case); // far
198 
199   // Done
200   bind(done);
201   dec_held_monitor_count(Rmark /*tmp*/);
202 }
203 
204 
205 void C1_MacroAssembler::try_allocate(
206   Register obj,                        // result: pointer to object after successful allocation
207   Register var_size_in_bytes,          // object size in bytes if unknown at compile time; invalid otherwise
208   int      con_size_in_bytes,          // object size in bytes if   known at compile time
209   Register t1,                         // temp register, must be global register for incr_allocated_bytes
210   Register t2,                         // temp register
211   Label&   slow_case                   // continuation point if fast allocation fails
212 ) {
213   if (UseTLAB) {
214     tlab_allocate(obj, var_size_in_bytes, con_size_in_bytes, t1, slow_case);
215   } else {
216     b(slow_case);
217   }
218 }
219 
220 
221 void C1_MacroAssembler::initialize_header(Register obj, Register klass, Register len, Register t1, Register t2) {
222   assert_different_registers(obj, klass, len, t1, t2);
223   load_const_optimized(t1, (intx)markWord::prototype().value());
224   std(t1, oopDesc::mark_offset_in_bytes(), obj);
225   store_klass(obj, klass);
226   if (len->is_valid()) {
227     stw(len, arrayOopDesc::length_offset_in_bytes(), obj);
228   } else if (UseCompressedClassPointers) {
229     // Otherwise length is in the class gap.
230     store_klass_gap(obj);
231   }
232 }
233 
234 
235 void C1_MacroAssembler::initialize_body(Register base, Register index) {
236   assert_different_registers(base, index);
237   srdi(index, index, LogBytesPerWord);
238   clear_memory_doubleword(base, index);
239 }
240 
241 void C1_MacroAssembler::initialize_body(Register obj, Register tmp1, Register tmp2,
242                                         int obj_size_in_bytes, int hdr_size_in_bytes) {
243   const int index = (obj_size_in_bytes - hdr_size_in_bytes) / HeapWordSize;
244 
245   // 2x unrolled loop is shorter with more than 9 HeapWords.
246   if (index <= 9) {
247     clear_memory_unrolled(obj, index, R0, hdr_size_in_bytes);
248   } else {
249     const Register base_ptr = tmp1,
250                    cnt_dwords = tmp2;
251 
252     addi(base_ptr, obj, hdr_size_in_bytes); // Compute address of first element.
253     clear_memory_doubleword(base_ptr, cnt_dwords, R0, index);
254   }
255 }
256 
257 void C1_MacroAssembler::allocate_object(
258   Register obj,                        // result: pointer to object after successful allocation
259   Register t1,                         // temp register
260   Register t2,                         // temp register
261   Register t3,                         // temp register
262   int      hdr_size,                   // object header size in words
263   int      obj_size,                   // object size in words
264   Register klass,                      // object klass
265   Label&   slow_case                   // continuation point if fast allocation fails
266 ) {
267   assert_different_registers(obj, t1, t2, t3, klass);
268 
269   // allocate space & initialize header
270   if (!is_simm16(obj_size * wordSize)) {
271     // Would need to use extra register to load
272     // object size => go the slow case for now.
273     b(slow_case);
274     return;
275   }
276   try_allocate(obj, noreg, obj_size * wordSize, t2, t3, slow_case);
277 
278   initialize_object(obj, klass, noreg, obj_size * HeapWordSize, t1, t2);
279 }
280 
281 void C1_MacroAssembler::initialize_object(
282   Register obj,                        // result: pointer to object after successful allocation
283   Register klass,                      // object klass
284   Register var_size_in_bytes,          // object size in bytes if unknown at compile time; invalid otherwise
285   int      con_size_in_bytes,          // object size in bytes if   known at compile time
286   Register t1,                         // temp register
287   Register t2                          // temp register
288   ) {
289   const int hdr_size_in_bytes = instanceOopDesc::header_size() * HeapWordSize;
290 
291   initialize_header(obj, klass, noreg, t1, t2);
292 
293 #ifdef ASSERT
294   {
295     lwz(t1, in_bytes(Klass::layout_helper_offset()), klass);
296     if (var_size_in_bytes != noreg) {
297       cmpw(CCR0, t1, var_size_in_bytes);
298     } else {
299       cmpwi(CCR0, t1, con_size_in_bytes);
300     }
301     asm_assert_eq("bad size in initialize_object");
302   }
303 #endif
304 
305   // Initialize body.
306   if (var_size_in_bytes != noreg) {
307     // Use a loop.
308     addi(t1, obj, hdr_size_in_bytes);                // Compute address of first element.
309     addi(t2, var_size_in_bytes, -hdr_size_in_bytes); // Compute size of body.
310     initialize_body(t1, t2);
311   } else if (con_size_in_bytes > hdr_size_in_bytes) {
312     // Use a loop.
313     initialize_body(obj, t1, t2, con_size_in_bytes, hdr_size_in_bytes);
314   }
315 
316   if (CURRENT_ENV->dtrace_alloc_probes()) {
317     Unimplemented();
318 //    assert(obj == O0, "must be");
319 //    call(CAST_FROM_FN_PTR(address, Runtime1::entry_for(Runtime1::dtrace_object_alloc_id)),
320 //         relocInfo::runtime_call_type);
321   }
322 
323   verify_oop(obj, FILE_AND_LINE);
324 }
325 
326 
327 void C1_MacroAssembler::allocate_array(
328   Register obj,                        // result: pointer to array after successful allocation
329   Register len,                        // array length
330   Register t1,                         // temp register
331   Register t2,                         // temp register
332   Register t3,                         // temp register
333   int      base_offset_in_bytes,       // elements offset in bytes
334   int      elt_size,                   // element size in bytes
335   Register klass,                      // object klass
336   Label&   slow_case                   // continuation point if fast allocation fails
337 ) {
338   assert_different_registers(obj, len, t1, t2, t3, klass);
339 
340   // Determine alignment mask.
341   assert(!(BytesPerWord & 1), "must be a multiple of 2 for masking code to work");
342   int log2_elt_size = exact_log2(elt_size);
343 
344   // Check for negative or excessive length.
345   size_t max_length = max_array_allocation_length >> log2_elt_size;
346   if (UseTLAB) {
347     size_t max_tlab = align_up(ThreadLocalAllocBuffer::max_size() >> log2_elt_size, 64*K);
348     if (max_tlab < max_length) { max_length = max_tlab; }
349   }
350   load_const_optimized(t1, max_length);
351   cmpld(CCR0, len, t1);
352   bc_far_optimized(Assembler::bcondCRbiIs1, bi0(CCR0, Assembler::greater), slow_case);
353 
354   // compute array size
355   // note: If 0 <= len <= max_length, len*elt_size + header + alignment is
356   //       smaller or equal to the largest integer; also, since top is always
357   //       aligned, we can do the alignment here instead of at the end address
358   //       computation.
359   const Register arr_size = t1;
360   Register arr_len_in_bytes = len;
361   if (elt_size != 1) {
362     sldi(t1, len, log2_elt_size);
363     arr_len_in_bytes = t1;
364   }
365   addi(arr_size, arr_len_in_bytes, base_offset_in_bytes + MinObjAlignmentInBytesMask); // Add space for header & alignment.
366   clrrdi(arr_size, arr_size, LogMinObjAlignmentInBytes);                              // Align array size.
367 
368   // Allocate space & initialize header.
369   try_allocate(obj, arr_size, 0, t2, t3, slow_case);
370   initialize_header(obj, klass, len, t2, t3);
371 
372   // Initialize body.
373   const Register base  = t2;
374   const Register index = t3;
375   addi(base, obj, base_offset_in_bytes);               // compute address of first element
376   addi(index, arr_size, -(base_offset_in_bytes));      // compute index = number of bytes to clear
377 
378   // Zero first 4 bytes, if start offset is not word aligned.
379   if (!is_aligned(base_offset_in_bytes, BytesPerWord)) {
380     assert(is_aligned(base_offset_in_bytes, BytesPerInt), "must be 4-byte aligned");
381     li(t1, 0);
382     stw(t1, 0, base);
383     addi(base, base, BytesPerInt);
384     // Note: initialize_body will align index down, no need to correct it here.
385   }
386 
387   initialize_body(base, index);
388 
389   if (CURRENT_ENV->dtrace_alloc_probes()) {
390     Unimplemented();
391     //assert(obj == O0, "must be");
392     //call(CAST_FROM_FN_PTR(address, Runtime1::entry_for(Runtime1::dtrace_object_alloc_id)),
393     //     relocInfo::runtime_call_type);
394   }
395 
396   verify_oop(obj, FILE_AND_LINE);
397 }
398 
399 
400 #ifndef PRODUCT
401 
402 void C1_MacroAssembler::verify_stack_oop(int stack_offset) {
403   verify_oop_addr((RegisterOrConstant)stack_offset, R1_SP, "broken oop in stack slot");
404 }
405 
406 void C1_MacroAssembler::verify_not_null_oop(Register r) {
407   Label not_null;
408   cmpdi(CCR0, r, 0);
409   bne(CCR0, not_null);
410   stop("non-null oop required");
411   bind(not_null);
412   verify_oop(r, FILE_AND_LINE);
413 }
414 
415 #endif // PRODUCT
416 
417 void C1_MacroAssembler::null_check(Register r, Label* Lnull) {
418   if (TrapBasedNullChecks) { // SIGTRAP based
419     trap_null_check(r);
420   } else { // explicit
421     //const address exception_entry = Runtime1::entry_for(Runtime1::throw_null_pointer_exception_id);
422     assert(Lnull != nullptr, "must have Label for explicit check");
423     cmpdi(CCR0, r, 0);
424     bc_far_optimized(Assembler::bcondCRbiIs1, bi0(CCR0, Assembler::equal), *Lnull);
425   }
426 }
427 
428 address C1_MacroAssembler::call_c_with_frame_resize(address dest, int frame_resize) {
429   if (frame_resize) { resize_frame(-frame_resize, R0); }
430 #if defined(ABI_ELFv2)
431   address return_pc = call_c(dest, relocInfo::runtime_call_type);
432 #else
433   address return_pc = call_c(CAST_FROM_FN_PTR(FunctionDescriptor*, dest), relocInfo::runtime_call_type);
434 #endif
435   if (frame_resize) { resize_frame(frame_resize, R0); }
436   return return_pc;
437 }