1 | /* |
2 | * Copyright (c) 2018, Oracle and/or its affiliates. All rights reserved. |
3 | * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
4 | * |
5 | * This code is free software; you can redistribute it and/or modify it |
6 | * under the terms of the GNU General Public License version 2 only, as |
7 | * published by the Free Software Foundation. |
8 | * |
9 | * This code is distributed in the hope that it will be useful, but WITHOUT |
10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
12 | * version 2 for more details (a copy is included in the LICENSE file that |
13 | * accompanied this code). |
14 | * |
15 | * You should have received a copy of the GNU General Public License version |
16 | * 2 along with this work; if not, write to the Free Software Foundation, |
17 | * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
18 | * |
19 | * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
20 | * or visit www.oracle.com if you need additional information or have any |
21 | * questions. |
22 | * |
23 | */ |
24 | |
25 | #include "precompiled.hpp" |
26 | #include "runtime/atomic.hpp" |
27 | #include "runtime/orderAccess.hpp" |
28 | #include "runtime/os.hpp" |
29 | #include "utilities/debug.hpp" |
30 | #include "utilities/singleWriterSynchronizer.hpp" |
31 | #include "utilities/macros.hpp" |
32 | |
33 | SingleWriterSynchronizer::SingleWriterSynchronizer() : |
34 | _enter(0), |
35 | _exit(), |
36 | // The initial value of 1 for _waiting_for puts it on the inactive |
37 | // track, so no thread exiting a critical section will match it. |
38 | _waiting_for(1), |
39 | _wakeup() |
40 | DEBUG_ONLY(COMMA _writers(0)) |
41 | {} |
42 | |
43 | // Wait until all threads that entered a critical section before |
44 | // synchronization have exited that critical section. |
45 | void SingleWriterSynchronizer::synchronize() { |
46 | // Side-effect in assert balanced by debug-only dec at end. |
47 | assert(Atomic::add(1u, &_writers) == 1u, "multiple writers" ); |
48 | // We don't know anything about the muxing between this invocation |
49 | // and invocations in other threads. We must start with the latest |
50 | // _enter polarity, else we could clobber the wrong _exit value on |
51 | // the first iteration. So fence to ensure everything here follows |
52 | // whatever muxing was used. |
53 | OrderAccess::fence(); |
54 | uint value = _enter; |
55 | // (1) Determine the old and new exit counters, based on the |
56 | // polarity (bit0 value) of the on-entry enter counter. |
57 | volatile uint* new_ptr = &_exit[(value + 1) & 1]; |
58 | // (2) Change the in-use exit counter to the new counter, by adding |
59 | // 1 to the enter counter (flipping the polarity), meanwhile |
60 | // "simultaneously" initializing the new exit counter to that enter |
61 | // value. Note: The new exit counter is not being used by read |
62 | // operations until this change of _enter succeeds. |
63 | uint old; |
64 | do { |
65 | old = value; |
66 | *new_ptr = ++value; |
67 | value = Atomic::cmpxchg(value, &_enter, old); |
68 | } while (old != value); |
69 | // Critical sections entered before we changed the polarity will use |
70 | // the old exit counter. Critical sections entered after the change |
71 | // will use the new exit counter. |
72 | volatile uint* old_ptr = &_exit[old & 1]; |
73 | assert(old_ptr != new_ptr, "invariant" ); |
74 | // (3) Inform threads in in-progress critical sections that there is |
75 | // a pending synchronize waiting. The thread that completes the |
76 | // request (_exit value == old) will signal the _wakeup semaphore to |
77 | // allow us to proceed. |
78 | _waiting_for = old; |
79 | // Write of _waiting_for must precede read of _exit and associated |
80 | // conditional semaphore wait. If they were re-ordered then a |
81 | // critical section exit could miss the wakeup request, failing to |
82 | // signal us while we're waiting. |
83 | OrderAccess::fence(); |
84 | // (4) Wait for all the critical sections started before the change |
85 | // to complete, e.g. for the value of old_ptr to catch up with old. |
86 | // Loop because there could be pending wakeups unrelated to this |
87 | // synchronize request. |
88 | while (old != OrderAccess::load_acquire(old_ptr)) { |
89 | _wakeup.wait(); |
90 | } |
91 | // (5) Drain any pending wakeups. A critical section exit may have |
92 | // completed our request and seen our _waiting_for before we checked |
93 | // for completion. There are also possible (though rare) spurious |
94 | // wakeup signals in the timing gap between changing the _enter |
95 | // polarity and setting _waiting_for. Enough of any of those could |
96 | // lead to semaphore overflow. This doesn't guarantee no unrelated |
97 | // wakeups for the next wait, but prevents unbounded accumulation. |
98 | while (_wakeup.trywait()) {} |
99 | DEBUG_ONLY(Atomic::dec(&_writers);) |
100 | } |
101 | |