forked from NebulousLabs/Sia
-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathtryrwmutex.go
102 lines (89 loc) · 2.76 KB
/
tryrwmutex.go
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
package sync
import (
"runtime"
"sync"
"sync/atomic"
)
const (
readOffset = uint64(1)
tryMask = uint64(1099511627775) // equivalent to setting the first 39 bits to '1'.
tryOffset = uint64(1 << 40)
writeOffset = uint64(1 << 20)
)
// TryRWMutex allows you to try to grab a RWMutex, failing if the mutex is
// unavailable. Standard blocking RLock and Lock calls also available.
//
// Note that there will be inconsistencies if there are more than 1 << 20
// operations active at once.
type TryRWMutex struct {
lock uint64
mu sync.RWMutex
}
// Lock blocks until the mutex is available, and then locks it.
func (tm *TryRWMutex) Lock() {
// Signal that a write lock is waiting.
v := atomic.AddUint64(&tm.lock, writeOffset)
// Spin until there is no contention from a Try call.
for v > tryOffset {
runtime.Gosched()
v = atomic.LoadUint64(&tm.lock)
}
// Grab the lock.
tm.mu.Lock()
}
// RLock blocks until the mutex is available, then grabs a read lock.
func (tm *TryRWMutex) RLock() {
// Signal that a read lock is waiting.
v := atomic.AddUint64(&tm.lock, readOffset)
// Spin until there is no contention from a Try call.
for v > tryOffset {
runtime.Gosched()
v = atomic.LoadUint64(&tm.lock)
}
// Grab the lock.
tm.mu.RLock()
}
// RUnlock releases a read lock on the mutex.
func (tm *TryRWMutex) RUnlock() {
// Release the lock, then signal that the read lock is no longer waiting.
tm.mu.RUnlock()
atomic.AddUint64(&tm.lock, ^(readOffset - 1))
}
// TryLock grabs a lock on the mutex, returning false if the mutex is
// unavailable.
func (tm *TryRWMutex) TryLock() bool {
// If there are no readlocks waiting, and no writelocks waiting, signal
// that a writelock is waiting and that there contention from a Try call.
if atomic.CompareAndSwapUint64(&tm.lock, 0, writeOffset+tryOffset) {
tm.mu.Lock()
// Signal that the Try call contention is resolved.
atomic.AddUint64(&tm.lock, ^(tryOffset - 1))
return true
}
return false
}
// TryRLock grabs a read lock on the mutex, returning false if the mutex is
// already locked.
func (tm *TryRWMutex) TryRLock() bool {
// Signal that a read lock is waiting, and that there is contention from a
// Try call.
v := atomic.AddUint64(&tm.lock, readOffset+tryOffset)
// Mask the try offset when performing the comparison.
v = v & tryMask
if v > writeOffset {
// If there is a write lock waiting, revert the signal and return
// false.
atomic.AddUint64(&tm.lock, ^(readOffset + tryOffset - 1))
return false
}
// Grab the read lock and return true.
tm.mu.RLock()
// Signal that the Try call contention is resolved.
atomic.AddUint64(&tm.lock, ^(tryOffset - 1))
return true
}
// Unlock releases a lock on the mutex.
func (tm *TryRWMutex) Unlock() {
tm.mu.Unlock()
atomic.AddUint64(&tm.lock, ^(writeOffset - 1))
}