< prev index next >

src/hotspot/cpu/aarch64/sharedRuntime_aarch64.cpp

Print this page
@@ -1,7 +1,7 @@
  /*
-  * Copyright (c) 2003, 2023, Oracle and/or its affiliates. All rights reserved.
+  * Copyright (c) 2003, 2024, Oracle and/or its affiliates. All rights reserved.
   * Copyright (c) 2014, 2021, Red Hat Inc. All rights reserved.
   * Copyright (c) 2021, Azul Systems, Inc. All rights reserved.
   * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
   *
   * This code is free software; you can redistribute it and/or modify it

@@ -1810,11 +1810,10 @@
        // Save the test result, for recursive case, the result is zero
        __ str(swap_reg, Address(lock_reg, mark_word_offset));
        __ br(Assembler::NE, slow_path_lock);
      } else {
        assert(LockingMode == LM_LIGHTWEIGHT, "must be");
-       __ ldr(swap_reg, Address(obj_reg, oopDesc::mark_offset_in_bytes()));
        __ lightweight_lock(obj_reg, swap_reg, tmp, lock_tmp, slow_path_lock);
      }
      __ bind(count);
      __ increment(Address(rthread, JavaThread::held_monitor_count_offset()));
  

@@ -1950,12 +1949,10 @@
        __ cmpxchg_obj_header(r0, old_hdr, obj_reg, rscratch1, count, &slow_path_unlock);
        __ bind(count);
        __ decrement(Address(rthread, JavaThread::held_monitor_count_offset()));
      } else {
        assert(LockingMode == LM_LIGHTWEIGHT, "");
-       __ ldr(old_hdr, Address(obj_reg, oopDesc::mark_offset_in_bytes()));
-       __ tbnz(old_hdr, exact_log2(markWord::monitor_value), slow_path_unlock);
        __ lightweight_unlock(obj_reg, old_hdr, swap_reg, lock_tmp, slow_path_unlock);
        __ decrement(Address(rthread, JavaThread::held_monitor_count_offset()));
      }
  
      // slow path re-enters here
< prev index next >