1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
|
#ifndef AL_ATOMIC_H
#define AL_ATOMIC_H
#include <atomic>
#include <memory>
#include "almalloc.h"
template<typename T>
auto IncrementRef(std::atomic<T> &ref) noexcept
{ return ref.fetch_add(1u, std::memory_order_acq_rel)+1u; }
template<typename T>
auto DecrementRef(std::atomic<T> &ref) noexcept
{ return ref.fetch_sub(1u, std::memory_order_acq_rel)-1u; }
/* WARNING: A livelock is theoretically possible if another thread keeps
* changing the head without giving this a chance to actually swap in the new
* one (practically impossible with this little code, but...).
*/
template<typename T>
inline void AtomicReplaceHead(std::atomic<T> &head, T newhead)
{
T first_ = head.load(std::memory_order_acquire);
do {
newhead->next.store(first_, std::memory_order_relaxed);
} while(!head.compare_exchange_weak(first_, newhead,
std::memory_order_acq_rel, std::memory_order_acquire));
}
namespace al {
template<typename T, typename D=std::default_delete<T>>
class atomic_unique_ptr {
std::atomic<gsl::owner<T*>> mPointer{};
using unique_ptr_t = std::unique_ptr<T,D>;
public:
atomic_unique_ptr() = default;
atomic_unique_ptr(const atomic_unique_ptr&) = delete;
explicit atomic_unique_ptr(std::nullptr_t) noexcept { }
explicit atomic_unique_ptr(gsl::owner<T*> ptr) noexcept : mPointer{ptr} { }
explicit atomic_unique_ptr(unique_ptr_t&& rhs) noexcept : mPointer{rhs.release()} { }
~atomic_unique_ptr()
{
if(auto ptr = mPointer.exchange(nullptr, std::memory_order_relaxed))
D{}(ptr);
}
auto operator=(const atomic_unique_ptr&) -> atomic_unique_ptr& = delete;
auto operator=(std::nullptr_t) noexcept -> atomic_unique_ptr&
{
if(auto ptr = mPointer.exchange(nullptr))
D{}(ptr);
return *this;
}
auto operator=(unique_ptr_t&& rhs) noexcept -> atomic_unique_ptr&
{
if(auto ptr = mPointer.exchange(rhs.release()))
D{}(ptr);
return *this;
}
[[nodiscard]]
auto load(std::memory_order m=std::memory_order_seq_cst) const noexcept -> T*
{ return mPointer.load(m); }
void store(std::nullptr_t, std::memory_order m=std::memory_order_seq_cst) noexcept
{
if(auto oldptr = mPointer.exchange(nullptr, m))
D{}(oldptr);
}
void store(gsl::owner<T*> ptr, std::memory_order m=std::memory_order_seq_cst) noexcept
{
if(auto oldptr = mPointer.exchange(ptr, m))
D{}(oldptr);
}
void store(unique_ptr_t&& ptr, std::memory_order m=std::memory_order_seq_cst) noexcept
{
if(auto oldptr = mPointer.exchange(ptr.release(), m))
D{}(oldptr);
}
[[nodiscard]]
auto exchange(std::nullptr_t, std::memory_order m=std::memory_order_seq_cst) noexcept -> unique_ptr_t
{ return unique_ptr_t{mPointer.exchange(nullptr, m)}; }
[[nodiscard]]
auto exchange(gsl::owner<T*> ptr, std::memory_order m=std::memory_order_seq_cst) noexcept -> unique_ptr_t
{ return unique_ptr_t{mPointer.exchange(ptr, m)}; }
[[nodiscard]]
auto exchange(std::unique_ptr<T>&& ptr, std::memory_order m=std::memory_order_seq_cst) noexcept -> unique_ptr_t
{ return unique_ptr_t{mPointer.exchange(ptr.release(), m)}; }
[[nodiscard]]
auto is_lock_free() const noexcept -> bool { mPointer.is_lock_free(); }
static constexpr auto is_always_lock_free = std::atomic<gsl::owner<T*>>::is_always_lock_free;
};
} // namespace al
#endif /* AL_ATOMIC_H */
|