| 1 | /* | 
|---|
| 2 | * Copyright (c) 2005, 2019, Oracle and/or its affiliates. All rights reserved. | 
|---|
| 3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. | 
|---|
| 4 | * | 
|---|
| 5 | * This code is free software; you can redistribute it and/or modify it | 
|---|
| 6 | * under the terms of the GNU General Public License version 2 only, as | 
|---|
| 7 | * published by the Free Software Foundation. | 
|---|
| 8 | * | 
|---|
| 9 | * This code is distributed in the hope that it will be useful, but WITHOUT | 
|---|
| 10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | 
|---|
| 11 | * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License | 
|---|
| 12 | * version 2 for more details (a copy is included in the LICENSE file that | 
|---|
| 13 | * accompanied this code). | 
|---|
| 14 | * | 
|---|
| 15 | * You should have received a copy of the GNU General Public License version | 
|---|
| 16 | * 2 along with this work; if not, write to the Free Software Foundation, | 
|---|
| 17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. | 
|---|
| 18 | * | 
|---|
| 19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA | 
|---|
| 20 | * or visit www.oracle.com if you need additional information or have any | 
|---|
| 21 | * questions. | 
|---|
| 22 | * | 
|---|
| 23 | */ | 
|---|
| 24 |  | 
|---|
| 25 | #ifndef SHARE_RUNTIME_BIASEDLOCKING_HPP | 
|---|
| 26 | #define SHARE_RUNTIME_BIASEDLOCKING_HPP | 
|---|
| 27 |  | 
|---|
| 28 | #include "runtime/handles.hpp" | 
|---|
| 29 | #include "utilities/growableArray.hpp" | 
|---|
| 30 |  | 
|---|
| 31 | // This class describes operations to implement Store-Free Biased | 
|---|
| 32 | // Locking. The high-level properties of the scheme are similar to | 
|---|
| 33 | // IBM's lock reservation, Dice-Moir-Scherer QR locks, and other biased | 
|---|
| 34 | // locking mechanisms. The principal difference is in the handling of | 
|---|
| 35 | // recursive locking which is how this technique achieves a more | 
|---|
| 36 | // efficient fast path than these other schemes. | 
|---|
| 37 | // | 
|---|
| 38 | // The basic observation is that in HotSpot's current fast locking | 
|---|
| 39 | // scheme, recursive locking (in the fast path) causes no update to | 
|---|
| 40 | // the object header. The recursion is described simply by stack | 
|---|
| 41 | // records containing a specific value (NULL). Only the last unlock by | 
|---|
| 42 | // a given thread causes an update to the object header. | 
|---|
| 43 | // | 
|---|
| 44 | // This observation, coupled with the fact that HotSpot only compiles | 
|---|
| 45 | // methods for which monitor matching is obeyed (and which therefore | 
|---|
| 46 | // can not throw IllegalMonitorStateException), implies that we can | 
|---|
| 47 | // completely eliminate modifications to the object header for | 
|---|
| 48 | // recursive locking in compiled code, and perform similar recursion | 
|---|
| 49 | // checks and throwing of IllegalMonitorStateException in the | 
|---|
| 50 | // interpreter with little or no impact on the performance of the fast | 
|---|
| 51 | // path. | 
|---|
| 52 | // | 
|---|
| 53 | // The basic algorithm is as follows (note, see below for more details | 
|---|
| 54 | // and information). A pattern in the low three bits is reserved in | 
|---|
| 55 | // the object header to indicate whether biasing of a given object's | 
|---|
| 56 | // lock is currently being done or is allowed at all.  If the bias | 
|---|
| 57 | // pattern is present, the contents of the rest of the header are | 
|---|
| 58 | // either the JavaThread* of the thread to which the lock is biased, | 
|---|
| 59 | // or NULL, indicating that the lock is "anonymously biased". The | 
|---|
| 60 | // first thread which locks an anonymously biased object biases the | 
|---|
| 61 | // lock toward that thread. If another thread subsequently attempts to | 
|---|
| 62 | // lock the same object, the bias is revoked. | 
|---|
| 63 | // | 
|---|
| 64 | // Because there are no updates to the object header at all during | 
|---|
| 65 | // recursive locking while the lock is biased, the biased lock entry | 
|---|
| 66 | // code is simply a test of the object header's value. If this test | 
|---|
| 67 | // succeeds, the lock has been acquired by the thread. If this test | 
|---|
| 68 | // fails, a bit test is done to see whether the bias bit is still | 
|---|
| 69 | // set. If not, we fall back to HotSpot's original CAS-based locking | 
|---|
| 70 | // scheme. If it is set, we attempt to CAS in a bias toward this | 
|---|
| 71 | // thread. The latter operation is expected to be the rarest operation | 
|---|
| 72 | // performed on these locks. We optimistically expect the biased lock | 
|---|
| 73 | // entry to hit most of the time, and want the CAS-based fallthrough | 
|---|
| 74 | // to occur quickly in the situations where the bias has been revoked. | 
|---|
| 75 | // | 
|---|
| 76 | // Revocation of the lock's bias is fairly straightforward. We want to | 
|---|
| 77 | // restore the object's header and stack-based BasicObjectLocks and | 
|---|
| 78 | // BasicLocks to the state they would have been in had the object been | 
|---|
| 79 | // locked by HotSpot's usual fast locking scheme. To do this, we bring | 
|---|
| 80 | // the system to a safepoint and walk the stack of the thread toward | 
|---|
| 81 | // which the lock is biased. We find all of the lock records on the | 
|---|
| 82 | // stack corresponding to this object, in particular the first / | 
|---|
| 83 | // "highest" record. We fill in the highest lock record with the | 
|---|
| 84 | // object's displaced header (which is a well-known value given that | 
|---|
| 85 | // we don't maintain an identity hash nor age bits for the object | 
|---|
| 86 | // while it's in the biased state) and all other lock records with 0, | 
|---|
| 87 | // the value for recursive locks. When the safepoint is released, the | 
|---|
| 88 | // formerly-biased thread and all other threads revert back to | 
|---|
| 89 | // HotSpot's CAS-based locking. | 
|---|
| 90 | // | 
|---|
| 91 | // This scheme can not handle transfers of biases of single objects | 
|---|
| 92 | // from thread to thread efficiently, but it can handle bulk transfers | 
|---|
| 93 | // of such biases, which is a usage pattern showing up in some | 
|---|
| 94 | // applications and benchmarks. We implement "bulk rebias" and "bulk | 
|---|
| 95 | // revoke" operations using a "bias epoch" on a per-data-type basis. | 
|---|
| 96 | // If too many bias revocations are occurring for a particular data | 
|---|
| 97 | // type, the bias epoch for the data type is incremented at a | 
|---|
| 98 | // safepoint, effectively meaning that all previous biases are | 
|---|
| 99 | // invalid. The fast path locking case checks for an invalid epoch in | 
|---|
| 100 | // the object header and attempts to rebias the object with a CAS if | 
|---|
| 101 | // found, avoiding safepoints or bulk heap sweeps (the latter which | 
|---|
| 102 | // was used in a prior version of this algorithm and did not scale | 
|---|
| 103 | // well). If too many bias revocations persist, biasing is completely | 
|---|
| 104 | // disabled for the data type by resetting the prototype header to the | 
|---|
| 105 | // unbiased markOop. The fast-path locking code checks to see whether | 
|---|
| 106 | // the instance's bias pattern differs from the prototype header's and | 
|---|
| 107 | // causes the bias to be revoked without reaching a safepoint or, | 
|---|
| 108 | // again, a bulk heap sweep. | 
|---|
| 109 |  | 
|---|
| 110 | // Biased locking counters | 
|---|
| 111 | class BiasedLockingCounters { | 
|---|
| 112 | private: | 
|---|
| 113 | int _total_entry_count; | 
|---|
| 114 | int _biased_lock_entry_count; | 
|---|
| 115 | int _anonymously_biased_lock_entry_count; | 
|---|
| 116 | int _rebiased_lock_entry_count; | 
|---|
| 117 | int _revoked_lock_entry_count; | 
|---|
| 118 | int _fast_path_entry_count; | 
|---|
| 119 | int _slow_path_entry_count; | 
|---|
| 120 |  | 
|---|
| 121 | public: | 
|---|
| 122 | BiasedLockingCounters() : | 
|---|
| 123 | _total_entry_count(0), | 
|---|
| 124 | _biased_lock_entry_count(0), | 
|---|
| 125 | _anonymously_biased_lock_entry_count(0), | 
|---|
| 126 | _rebiased_lock_entry_count(0), | 
|---|
| 127 | _revoked_lock_entry_count(0), | 
|---|
| 128 | _fast_path_entry_count(0), | 
|---|
| 129 | _slow_path_entry_count(0) {} | 
|---|
| 130 |  | 
|---|
| 131 | int slow_path_entry_count() const; // Compute this field if necessary | 
|---|
| 132 |  | 
|---|
| 133 | int* total_entry_count_addr()                   { return &_total_entry_count; } | 
|---|
| 134 | int* biased_lock_entry_count_addr()             { return &_biased_lock_entry_count; } | 
|---|
| 135 | int* anonymously_biased_lock_entry_count_addr() { return &_anonymously_biased_lock_entry_count; } | 
|---|
| 136 | int* rebiased_lock_entry_count_addr()           { return &_rebiased_lock_entry_count; } | 
|---|
| 137 | int* revoked_lock_entry_count_addr()            { return &_revoked_lock_entry_count; } | 
|---|
| 138 | int* fast_path_entry_count_addr()               { return &_fast_path_entry_count; } | 
|---|
| 139 | int* slow_path_entry_count_addr()               { return &_slow_path_entry_count; } | 
|---|
| 140 |  | 
|---|
| 141 | bool nonzero() { return _total_entry_count > 0; } | 
|---|
| 142 |  | 
|---|
| 143 | void print_on(outputStream* st) const; | 
|---|
| 144 | void print() const; | 
|---|
| 145 | }; | 
|---|
| 146 |  | 
|---|
| 147 |  | 
|---|
| 148 | class BiasedLocking : AllStatic { | 
|---|
| 149 | private: | 
|---|
| 150 | static BiasedLockingCounters _counters; | 
|---|
| 151 |  | 
|---|
| 152 | public: | 
|---|
| 153 | static int* total_entry_count_addr(); | 
|---|
| 154 | static int* biased_lock_entry_count_addr(); | 
|---|
| 155 | static int* anonymously_biased_lock_entry_count_addr(); | 
|---|
| 156 | static int* rebiased_lock_entry_count_addr(); | 
|---|
| 157 | static int* revoked_lock_entry_count_addr(); | 
|---|
| 158 | static int* fast_path_entry_count_addr(); | 
|---|
| 159 | static int* slow_path_entry_count_addr(); | 
|---|
| 160 |  | 
|---|
| 161 | enum Condition { | 
|---|
| 162 | NOT_BIASED = 1, | 
|---|
| 163 | BIAS_REVOKED = 2, | 
|---|
| 164 | BIAS_REVOKED_AND_REBIASED = 3 | 
|---|
| 165 | }; | 
|---|
| 166 |  | 
|---|
| 167 | // This initialization routine should only be called once and | 
|---|
| 168 | // schedules a PeriodicTask to turn on biased locking a few seconds | 
|---|
| 169 | // into the VM run to avoid startup time regressions | 
|---|
| 170 | static void init(); | 
|---|
| 171 |  | 
|---|
| 172 | // This provides a global switch for leaving biased locking disabled | 
|---|
| 173 | // for the first part of a run and enabling it later | 
|---|
| 174 | static bool enabled(); | 
|---|
| 175 |  | 
|---|
| 176 | // This should be called by JavaThreads to revoke the bias of an object | 
|---|
| 177 | static Condition revoke_and_rebias(Handle obj, bool attempt_rebias, TRAPS); | 
|---|
| 178 |  | 
|---|
| 179 | // These do not allow rebiasing; they are used by deoptimization to | 
|---|
| 180 | // ensure that monitors on the stack can be migrated | 
|---|
| 181 | static void revoke(GrowableArray<Handle>* objs); | 
|---|
| 182 | static void revoke_at_safepoint(Handle obj); | 
|---|
| 183 | static void revoke_at_safepoint(GrowableArray<Handle>* objs); | 
|---|
| 184 |  | 
|---|
| 185 | static void print_counters() { _counters.print(); } | 
|---|
| 186 | static BiasedLockingCounters* counters() { return &_counters; } | 
|---|
| 187 |  | 
|---|
| 188 | // These routines are GC-related and should not be called by end | 
|---|
| 189 | // users. GCs which do not do preservation of mark words do not need | 
|---|
| 190 | // to call these routines. | 
|---|
| 191 | static void preserve_marks(); | 
|---|
| 192 | static void restore_marks(); | 
|---|
| 193 | }; | 
|---|
| 194 |  | 
|---|
| 195 | #endif // SHARE_RUNTIME_BIASEDLOCKING_HPP | 
|---|
| 196 |  | 
|---|