1 /* 2 * Copyright (c) 2016, 2023, Oracle and/or its affiliates. All rights reserved. 3 * Copyright (c) 2016, 2023 SAP SE. All rights reserved. 4 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 5 * 6 * This code is free software; you can redistribute it and/or modify it 7 * under the terms of the GNU General Public License version 2 only, as 8 * published by the Free Software Foundation. 9 * 10 * This code is distributed in the hope that it will be useful, but WITHOUT 11 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 12 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 13 * version 2 for more details (a copy is included in the LICENSE file that 14 * accompanied this code). 15 * 16 * You should have received a copy of the GNU General Public License version 17 * 2 along with this work; if not, write to the Free Software Foundation, 18 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 19 * 20 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 21 * or visit www.oracle.com if you need additional information or have any 22 * questions. 23 * 24 */ 25 26 #include "precompiled.hpp" 27 #include "asm/macroAssembler.inline.hpp" 28 #include "c1/c1_MacroAssembler.hpp" 29 #include "c1/c1_Runtime1.hpp" 30 #include "gc/shared/barrierSet.hpp" 31 #include "gc/shared/barrierSetAssembler.hpp" 32 #include "gc/shared/collectedHeap.hpp" 33 #include "gc/shared/tlab_globals.hpp" 34 #include "interpreter/interpreter.hpp" 35 #include "oops/arrayOop.hpp" 36 #include "oops/markWord.hpp" 37 #include "runtime/basicLock.hpp" 38 #include "runtime/os.hpp" 39 #include "runtime/sharedRuntime.hpp" 40 #include "runtime/stubRoutines.hpp" 41 #include "utilities/macros.hpp" 42 43 void C1_MacroAssembler::inline_cache_check(Register receiver, Register iCache) { 44 Label ic_miss, ic_hit; 45 verify_oop(receiver, FILE_AND_LINE); 46 int klass_offset = oopDesc::klass_offset_in_bytes(); 47 48 if (!ImplicitNullChecks || MacroAssembler::needs_explicit_null_check(klass_offset)) { 49 if (VM_Version::has_CompareBranch()) { 50 z_cgij(receiver, 0, Assembler::bcondEqual, ic_miss); 51 } else { 52 z_ltgr(receiver, receiver); 53 z_bre(ic_miss); 54 } 55 } 56 57 compare_klass_ptr(iCache, klass_offset, receiver, false); 58 z_bre(ic_hit); 59 60 // If icache check fails, then jump to runtime routine. 61 // Note: RECEIVER must still contain the receiver! 62 load_const_optimized(Z_R1_scratch, AddressLiteral(SharedRuntime::get_ic_miss_stub())); 63 z_br(Z_R1_scratch); 64 align(CodeEntryAlignment); 65 bind(ic_hit); 66 } 67 68 void C1_MacroAssembler::explicit_null_check(Register base) { 69 ShouldNotCallThis(); // unused 70 } 71 72 void C1_MacroAssembler::build_frame(int frame_size_in_bytes, int bang_size_in_bytes) { 73 assert(bang_size_in_bytes >= frame_size_in_bytes, "stack bang size incorrect"); 74 generate_stack_overflow_check(bang_size_in_bytes); 75 save_return_pc(); 76 push_frame(frame_size_in_bytes); 77 78 BarrierSetAssembler* bs = BarrierSet::barrier_set()->barrier_set_assembler(); 79 bs->nmethod_entry_barrier(this); 80 } 81 82 void C1_MacroAssembler::verified_entry(bool breakAtEntry) { 83 if (breakAtEntry) z_illtrap(0xC1); 84 } 85 86 void C1_MacroAssembler::lock_object(Register Rmark, Register Roop, Register Rbox, Label& slow_case) { 87 const int hdr_offset = oopDesc::mark_offset_in_bytes(); 88 89 const Register tmp = Z_R1_scratch; 90 91 assert_different_registers(Rmark, Roop, Rbox, tmp); 92 93 verify_oop(Roop, FILE_AND_LINE); 94 95 // Load object header. 96 z_lg(Rmark, Address(Roop, hdr_offset)); 97 98 // Save object being locked into the BasicObjectLock... 99 z_stg(Roop, Address(Rbox, BasicObjectLock::obj_offset())); 100 101 if (DiagnoseSyncOnValueBasedClasses != 0) { 102 load_klass(tmp, Roop); 103 testbit(Address(tmp, Klass::access_flags_offset()), exact_log2(JVM_ACC_IS_VALUE_BASED_CLASS)); 104 branch_optimized(Assembler::bcondAllOne, slow_case); 105 } 106 107 assert(LockingMode != LM_MONITOR, "LM_MONITOR is already handled, by emit_lock()"); 108 109 if (LockingMode == LM_LIGHTWEIGHT) { 110 lightweight_lock(Roop, Rmark, tmp, slow_case); 111 } else if (LockingMode == LM_LEGACY) { 112 NearLabel done; 113 // and mark it as unlocked. 114 z_oill(Rmark, markWord::unlocked_value); 115 // Save unlocked object header into the displaced header location on the stack. 116 z_stg(Rmark, Address(Rbox, BasicLock::displaced_header_offset_in_bytes())); 117 // Test if object header is still the same (i.e. unlocked), and if so, store the 118 // displaced header address in the object header. If it is not the same, get the 119 // object header instead. 120 z_csg(Rmark, Rbox, hdr_offset, Roop); 121 // If the object header was the same, we're done. 122 branch_optimized(Assembler::bcondEqual, done); 123 // If the object header was not the same, it is now in the Rmark register. 124 // => Test if it is a stack pointer into the same stack (recursive locking), i.e.: 125 // 126 // 1) (Rmark & markWord::lock_mask_in_place) == 0 127 // 2) rsp <= Rmark 128 // 3) Rmark <= rsp + page_size 129 // 130 // These 3 tests can be done by evaluating the following expression: 131 // 132 // (Rmark - Z_SP) & (~(page_size-1) | markWord::lock_mask_in_place) 133 // 134 // assuming both the stack pointer and page_size have their least 135 // significant 2 bits cleared and page_size is a power of 2 136 z_sgr(Rmark, Z_SP); 137 138 load_const_optimized(Z_R0_scratch, (~(os::vm_page_size() - 1) | markWord::lock_mask_in_place)); 139 z_ngr(Rmark, Z_R0_scratch); // AND sets CC (result eq/ne 0). 140 // For recursive locking, the result is zero. => Save it in the displaced header 141 // location (null in the displaced Rmark location indicates recursive locking). 142 z_stg(Rmark, Address(Rbox, BasicLock::displaced_header_offset_in_bytes())); 143 // Otherwise we don't care about the result and handle locking via runtime call. 144 branch_optimized(Assembler::bcondNotZero, slow_case); 145 // done 146 bind(done); 147 } 148 } 149 150 void C1_MacroAssembler::unlock_object(Register Rmark, Register Roop, Register Rbox, Label& slow_case) { 151 const int hdr_offset = oopDesc::mark_offset_in_bytes(); 152 153 assert_different_registers(Rmark, Roop, Rbox); 154 155 NearLabel done; 156 157 if (LockingMode != LM_LIGHTWEIGHT) { 158 // Load displaced header. 159 z_ltg(Rmark, Address(Rbox, BasicLock::displaced_header_offset_in_bytes())); 160 // If the loaded Rmark is null we had recursive locking, and we are done. 161 z_bre(done); 162 } 163 164 // Load object. 165 z_lg(Roop, Address(Rbox, BasicObjectLock::obj_offset())); 166 verify_oop(Roop, FILE_AND_LINE); 167 168 if (LockingMode == LM_LIGHTWEIGHT) { 169 const Register tmp = Z_R1_scratch; 170 z_lg(Rmark, Address(Roop, hdr_offset)); 171 z_lgr(tmp, Rmark); 172 z_nill(tmp, markWord::monitor_value); 173 branch_optimized(Assembler::bcondNotZero, slow_case); 174 lightweight_unlock(Roop, Rmark, tmp, slow_case); 175 } else if (LockingMode == LM_LEGACY) { 176 // Test if object header is pointing to the displaced header, and if so, restore 177 // the displaced header in the object. If the object header is not pointing to 178 // the displaced header, get the object header instead. 179 z_csg(Rbox, Rmark, hdr_offset, Roop); 180 // If the object header was not pointing to the displaced header, 181 // we do unlocking via runtime call. 182 branch_optimized(Assembler::bcondNotEqual, slow_case); 183 } 184 // done 185 bind(done); 186 } 187 188 void C1_MacroAssembler::try_allocate( 189 Register obj, // result: Pointer to object after successful allocation. 190 Register var_size_in_bytes, // Object size in bytes if unknown at compile time; invalid otherwise. 191 int con_size_in_bytes, // Object size in bytes if known at compile time. 192 Register t1, // Temp register: Must be global register for incr_allocated_bytes. 193 Label& slow_case // Continuation point if fast allocation fails. 194 ) { 195 if (UseTLAB) { 196 tlab_allocate(obj, var_size_in_bytes, con_size_in_bytes, t1, slow_case); 197 } else { 198 // Allocation in shared Eden not implemented, because sapjvm allocation trace does not allow it. 199 z_brul(slow_case); 200 } 201 } 202 203 void C1_MacroAssembler::initialize_header(Register obj, Register klass, Register len, Register Rzero, Register t1) { 204 assert_different_registers(obj, klass, len, t1, Rzero); 205 // This assumes that all prototype bits fit in an int32_t. 206 load_const_optimized(t1, (intx)markWord::prototype().value()); 207 z_stg(t1, Address(obj, oopDesc::mark_offset_in_bytes())); 208 209 if (len->is_valid()) { 210 // Length will be in the klass gap, if one exists. 211 z_st(len, Address(obj, arrayOopDesc::length_offset_in_bytes())); 212 } else if (UseCompressedClassPointers) { 213 store_klass_gap(Rzero, obj); // Zero klass gap for compressed oops. 214 } 215 store_klass(klass, obj, t1); 216 } 217 218 void C1_MacroAssembler::initialize_body(Register objectFields, Register len_in_bytes, Register Rzero) { 219 Label done; 220 assert_different_registers(objectFields, len_in_bytes, Rzero); 221 222 // Initialize object fields. 223 // See documentation for MVCLE instruction!!! 224 assert(objectFields->encoding()%2==0, "objectFields must be an even register"); 225 assert(len_in_bytes->encoding() == (objectFields->encoding()+1), "objectFields and len_in_bytes must be a register pair"); 226 assert(Rzero->encoding()%2==1, "Rzero must be an odd register"); 227 228 // Use Rzero as src length, then mvcle will copy nothing 229 // and fill the object with the padding value 0. 230 move_long_ext(objectFields, as_Register(Rzero->encoding()-1), 0); 231 bind(done); 232 } 233 234 void C1_MacroAssembler::allocate_object( 235 Register obj, // Result: pointer to object after successful allocation. 236 Register t1, // temp register 237 Register t2, // temp register: Must be a global register for try_allocate. 238 int hdr_size, // object header size in words 239 int obj_size, // object size in words 240 Register klass, // object klass 241 Label& slow_case // Continuation point if fast allocation fails. 242 ) { 243 assert_different_registers(obj, t1, t2, klass); 244 245 // Allocate space and initialize header. 246 try_allocate(obj, noreg, obj_size * wordSize, t1, slow_case); 247 248 initialize_object(obj, klass, noreg, obj_size * HeapWordSize, t1, t2); 249 } 250 251 void C1_MacroAssembler::initialize_object( 252 Register obj, // result: Pointer to object after successful allocation. 253 Register klass, // object klass 254 Register var_size_in_bytes, // Object size in bytes if unknown at compile time; invalid otherwise. 255 int con_size_in_bytes, // Object size in bytes if known at compile time. 256 Register t1, // temp register 257 Register t2 // temp register 258 ) { 259 assert((con_size_in_bytes & MinObjAlignmentInBytesMask) == 0, 260 "con_size_in_bytes is not multiple of alignment"); 261 assert(var_size_in_bytes == noreg, "not implemented"); 262 const int hdr_size_in_bytes = instanceOopDesc::header_size() * HeapWordSize; 263 264 const Register Rzero = t2; 265 266 z_xgr(Rzero, Rzero); 267 initialize_header(obj, klass, noreg, Rzero, t1); 268 269 // Clear rest of allocated space. 270 const int threshold = 4 * BytesPerWord; 271 if (con_size_in_bytes <= threshold) { 272 // Use explicit null stores. 273 // code size = 6*n bytes (n = number of fields to clear) 274 for (int i = hdr_size_in_bytes; i < con_size_in_bytes; i += BytesPerWord) 275 z_stg(Rzero, Address(obj, i)); 276 } else { 277 // Code size generated by initialize_body() is 16. 278 Register object_fields = Z_R0_scratch; 279 Register len_in_bytes = Z_R1_scratch; 280 z_la(object_fields, hdr_size_in_bytes, obj); 281 load_const_optimized(len_in_bytes, con_size_in_bytes - hdr_size_in_bytes); 282 initialize_body(object_fields, len_in_bytes, Rzero); 283 } 284 285 // Dtrace support is unimplemented. 286 // if (CURRENT_ENV->dtrace_alloc_probes()) { 287 // assert(obj == rax, "must be"); 288 // call(RuntimeAddress(Runtime1::entry_for (Runtime1::dtrace_object_alloc_id))); 289 // } 290 291 verify_oop(obj, FILE_AND_LINE); 292 } 293 294 void C1_MacroAssembler::allocate_array( 295 Register obj, // result: Pointer to array after successful allocation. 296 Register len, // array length 297 Register t1, // temp register 298 Register t2, // temp register 299 int base_offset_in_bytes, // elements offset in bytes 300 int elt_size, // element size in bytes 301 Register klass, // object klass 302 Label& slow_case // Continuation point if fast allocation fails. 303 ) { 304 assert_different_registers(obj, len, t1, t2, klass); 305 306 // Determine alignment mask. 307 assert(!(BytesPerWord & 1), "must be a multiple of 2 for masking code to work"); 308 309 // Check for negative or excessive length. 310 compareU64_and_branch(len, (int32_t)max_array_allocation_length, bcondHigh, slow_case); 311 312 // Compute array size. 313 // Note: If 0 <= len <= max_length, len*elt_size + header + alignment is 314 // smaller or equal to the largest integer. Also, since top is always 315 // aligned, we can do the alignment here instead of at the end address 316 // computation. 317 const Register arr_size = t2; 318 switch (elt_size) { 319 case 1: lgr_if_needed(arr_size, len); break; 320 case 2: z_sllg(arr_size, len, 1); break; 321 case 4: z_sllg(arr_size, len, 2); break; 322 case 8: z_sllg(arr_size, len, 3); break; 323 default: ShouldNotReachHere(); 324 } 325 add2reg(arr_size, base_offset_in_bytes + MinObjAlignmentInBytesMask); // Add space for header & alignment. 326 z_nill(arr_size, (~MinObjAlignmentInBytesMask) & 0xffff); // Align array size. 327 328 try_allocate(obj, arr_size, 0, t1, slow_case); 329 330 initialize_header(obj, klass, len, noreg, t1); 331 332 // Clear rest of allocated space. 333 Label done; 334 Register object_fields = t1; 335 Register Rzero = Z_R1_scratch; 336 z_aghi(arr_size, -base_offset_in_bytes); 337 z_bre(done); // Jump if size of fields is zero. 338 z_la(object_fields, base_offset_in_bytes, obj); 339 z_xgr(Rzero, Rzero); 340 initialize_body(object_fields, arr_size, Rzero); 341 bind(done); 342 343 // Dtrace support is unimplemented. 344 // if (CURRENT_ENV->dtrace_alloc_probes()) { 345 // assert(obj == rax, "must be"); 346 // call(RuntimeAddress(Runtime1::entry_for (Runtime1::dtrace_object_alloc_id))); 347 // } 348 349 verify_oop(obj, FILE_AND_LINE); 350 } 351 352 353 #ifndef PRODUCT 354 355 void C1_MacroAssembler::verify_stack_oop(int stack_offset) { 356 if (!VerifyOops) return; 357 verify_oop_addr(Address(Z_SP, stack_offset), FILE_AND_LINE); 358 } 359 360 void C1_MacroAssembler::verify_not_null_oop(Register r) { 361 if (!VerifyOops) return; 362 NearLabel not_null; 363 compareU64_and_branch(r, (intptr_t)0, bcondNotEqual, not_null); 364 stop("non-null oop required"); 365 bind(not_null); 366 verify_oop(r, FILE_AND_LINE); 367 } 368 369 void C1_MacroAssembler::invalidate_registers(Register preserve1, 370 Register preserve2, 371 Register preserve3) { 372 Register dead_value = noreg; 373 for (int i = 0; i < FrameMap::nof_cpu_regs; i++) { 374 Register r = as_Register(i); 375 if (r != preserve1 && r != preserve2 && r != preserve3 && r != Z_SP && r != Z_thread) { 376 if (dead_value == noreg) { 377 load_const_optimized(r, 0xc1dead); 378 dead_value = r; 379 } else { 380 z_lgr(r, dead_value); 381 } 382 } 383 } 384 } 385 386 #endif // !PRODUCT