aboutsummaryrefslogtreecommitdiffstats
path: root/include
diff options
context:
space:
mode:
Diffstat (limited to 'include')
-rw-r--r--include/jau/cow_darray.hpp110
-rw-r--r--include/jau/cow_vector.hpp73
-rw-r--r--include/jau/cpp_lang_macros.hpp56
3 files changed, 174 insertions, 65 deletions
diff --git a/include/jau/cow_darray.hpp b/include/jau/cow_darray.hpp
index 07c44b0..d996ac7 100644
--- a/include/jau/cow_darray.hpp
+++ b/include/jau/cow_darray.hpp
@@ -35,6 +35,7 @@
#include <condition_variable>
#include <algorithm>
+#include <jau/cpp_lang_macros.hpp>
#include <jau/debug.hpp>
#include <jau/darray.hpp>
#include <jau/basic_types.hpp>
@@ -145,8 +146,8 @@ namespace jau {
static constexpr size_type DIFF_MAX = std::numeric_limits<difference_type>::max();
storage_ref_t store_ref;
- sc_atomic_bool sync_atomic;
- std::recursive_mutex mtx_write;
+ mutable sc_atomic_bool sync_atomic;
+ mutable std::recursive_mutex mtx_write;
public:
// ctor w/o elements
@@ -187,11 +188,12 @@ namespace jau {
* Capacity and size will equal the given array, i.e. the result is a trimmed array.
* @param x the given cow_darray, all elements will be copied into the new instance.
*/
- constexpr cow_darray(const cow_darray& x)
+ __constexpr_non_literal_atomic__
+ cow_darray(const cow_darray& x)
: sync_atomic(false) {
storage_ref_t x_store_ref;
{
- sc_atomic_critical sync_x( const_cast<cow_darray *>(&x)->sync_atomic );
+ sc_atomic_critical sync_x( x.sync_atomic );
x_store_ref = x.store_ref;
}
store_ref = std::make_shared<storage_t>( *x_store_ref );
@@ -204,11 +206,12 @@ namespace jau {
* @param growth_factor custom growth factor
* @param alloc custom allocator_type instance
*/
- constexpr explicit cow_darray(const cow_darray& x, const float growth_factor, const allocator_type& alloc)
+ __constexpr_non_literal_atomic__
+ explicit cow_darray(const cow_darray& x, const float growth_factor, const allocator_type& alloc)
: sync_atomic(false) {
storage_ref_t x_store_ref;
{
- sc_atomic_critical sync_x( const_cast<cow_darray *>(&x)->sync_atomic );
+ sc_atomic_critical sync_x( x.sync_atomic );
x_store_ref = x.store_ref;
}
store_ref = std::make_shared<storage_t>( *x_store_ref, growth_factor, alloc );
@@ -225,11 +228,12 @@ namespace jau {
* @param growth_factor custom growth factor
* @param alloc custom allocator_type instance
*/
- constexpr explicit cow_darray(const cow_darray& x, const size_type _capacity, const float growth_factor, const allocator_type& alloc)
+ __constexpr_non_literal_atomic__
+ explicit cow_darray(const cow_darray& x, const size_type _capacity, const float growth_factor, const allocator_type& alloc)
: sync_atomic(false) {
storage_ref_t x_store_ref;
{
- sc_atomic_critical sync_x( const_cast<cow_darray *>(&x)->sync_atomic );
+ sc_atomic_critical sync_x( x.sync_atomic );
x_store_ref = x.store_ref;
}
store_ref = std::make_shared<storage_t>( *x_store_ref, _capacity, growth_factor, alloc );
@@ -325,8 +329,9 @@ namespace jau {
* @see jau::cow_darray::copy_store()
* @see jau::cow_darray::set_store()
*/
+ __constexpr_non_literal_atomic__
storage_ref_t copy_store() {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
return std::make_shared<storage_t>( *store_ref );
}
@@ -342,7 +347,7 @@ namespace jau {
* cow_darray<std::shared_ptr<Thing>> list;
* ...
* {
- * const std::lock_guard<std::recursive_mutex> lock(list.get_write_mutex());
+ * std::lock_guard<std::recursive_mutex> lock(list.get_write_mutex());
* std::shared_ptr<std::vector<std::shared_ptr<Thing>>> snapshot = list.copy_store();
* ...
* some fancy mutation
@@ -357,8 +362,9 @@ namespace jau {
* @see jau::cow_darray::copy_store()
* @see jau::cow_darray::set_store()
*/
- constexpr void set_store(storage_ref_t && new_store_ref) noexcept {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ __constexpr_non_literal_atomic__
+ void set_store(storage_ref_t && new_store_ref) noexcept {
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
sc_atomic_critical sync(sync_atomic);
store_ref = std::move( new_store_ref );
}
@@ -375,8 +381,9 @@ namespace jau {
* </p>
* @see jau::for_each_cow
*/
- constexpr storage_ref_t get_snapshot() const noexcept {
- sc_atomic_critical sync( const_cast<cow_darray *>(this)->sync_atomic );
+ __constexpr_non_literal_atomic__
+ storage_ref_t get_snapshot() const noexcept {
+ sc_atomic_critical sync( sync_atomic );
return store_ref;
}
@@ -411,17 +418,34 @@ namespace jau {
// read access
const allocator_type& get_allocator_ref() const noexcept {
- sc_atomic_critical sync( const_cast<cow_darray *>(this)->sync_atomic );
+ sc_atomic_critical sync( sync_atomic );
return store_ref->get_allocator_ref();
}
allocator_type get_allocator() const noexcept {
- sc_atomic_critical sync( const_cast<cow_darray *>(this)->sync_atomic );
+ sc_atomic_critical sync( sync_atomic );
return store_ref->get_allocator();
}
- constexpr size_type capacity() const noexcept {
- sc_atomic_critical sync( const_cast<cow_darray *>(this)->sync_atomic );
+ /**
+ * Returns the growth factor
+ */
+ __constexpr_non_literal_atomic__
+ float growth_factor() const noexcept {
+ sc_atomic_critical sync( sync_atomic );
+ return store_ref->growth_factor();
+ }
+
+ /**
+ * Like std::vector::empty().
+ * <p>
+ * This read operation is <i>lock-free</i>.
+ * </p>
+ * @return
+ */
+ __constexpr_non_literal_atomic__
+ size_type capacity() const noexcept {
+ sc_atomic_critical sync( sync_atomic );
return store_ref->capacity();
}
@@ -431,8 +455,9 @@ namespace jau {
* This read operation is <i>lock-free</i>.
* </p>
*/
- constexpr bool empty() const noexcept {
- sc_atomic_critical sync( const_cast<cow_darray *>(this)->sync_atomic );
+ __constexpr_non_literal_atomic__
+ bool empty() const noexcept {
+ sc_atomic_critical sync( sync_atomic );
return store_ref->empty();
}
@@ -442,8 +467,9 @@ namespace jau {
* This read operation is <i>lock-free</i>.
* </p>
*/
- constexpr size_type size() const noexcept {
- sc_atomic_critical sync( const_cast<cow_darray *>(this)->sync_atomic );
+ __constexpr_non_literal_atomic__
+ size_type size() const noexcept {
+ sc_atomic_critical sync( sync_atomic );
return store_ref->size();
}
@@ -460,12 +486,12 @@ namespace jau {
* </p>
*/
void reserve(size_type new_capacity) {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t old_store_ref = store_ref;
if( new_capacity > old_store_ref->capacity() ) {
storage_ref_t new_store_ref = std::make_shared<storage_t>( *old_store_ref, new_capacity,
old_store_ref->growth_factor(),
old_store_ref->get_allocator_ref() );
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
sc_atomic_critical sync( sync_atomic );
store_ref = std::move(new_store_ref);
}
@@ -478,7 +504,7 @@ namespace jau {
* </p>
*/
cow_darray& operator=(const storage_t& x) {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t new_store_ref = std::make_shared<storage_t>( x );
sc_atomic_critical sync(sync_atomic);
store_ref = std::move(new_store_ref);
@@ -492,7 +518,7 @@ namespace jau {
* </p>
*/
cow_darray& operator=(storage_t&& x) {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t new_store_ref = std::make_shared<storage_t>( std::move(x) );
sc_atomic_critical sync(sync_atomic);
store_ref = std::move(new_store_ref);
@@ -506,10 +532,10 @@ namespace jau {
* </p>
*/
cow_darray& operator=(const cow_darray& x) {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t x_store_ref;
{
- sc_atomic_critical sync_x( const_cast<cow_darray *>(&x)->sync_atomic );
+ sc_atomic_critical sync_x( x.sync_atomic );
x_store_ref = x.store_ref;
}
storage_ref_t new_store_ref = std::make_shared<storage_t>( *x_store_ref );
@@ -546,8 +572,9 @@ namespace jau {
* This write operation uses a mutex lock and is blocking this instances' write operations.
* </p>
*/
+ __constexpr_non_literal_atomic__
void clear() noexcept {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t new_store_ref = std::make_shared<storage_t>();
{
sc_atomic_critical sync(sync_atomic);
@@ -561,6 +588,7 @@ namespace jau {
* This write operation uses a mutex lock and is blocking both cow_darray instance's write operations.
* </p>
*/
+ __constexpr_non_literal_atomic__
void swap(cow_darray& x) noexcept {
std::unique_lock<std::recursive_mutex> lock(mtx_write, std::defer_lock); // utilize std::lock(a, b), allowing mixed order waiting on either object
std::unique_lock<std::recursive_mutex> lock_x(x.mtx_write, std::defer_lock); // otherwise RAII-style relinquish via destructor
@@ -580,8 +608,8 @@ namespace jau {
* This write operation uses a mutex lock and is blocking this instances' write operations only.
* </p>
*/
+ __constexpr_non_literal_atomic__
void pop_back() noexcept {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t old_store_ref = store_ref;
if( !old_store_ref->empty() ) {
storage_ref_t new_store_ref = std::make_shared<storage_t>( old_store_ref->capacity(),
@@ -589,6 +617,7 @@ namespace jau {
old_store_ref->cend()-1,
old_store_ref->growth_factor(),
old_store_ref->get_allocator_ref() );
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
{
sc_atomic_critical sync(sync_atomic);
store_ref = std::move(new_store_ref);
@@ -603,10 +632,11 @@ namespace jau {
* </p>
* @param x the value to be added at the tail.
*/
+ __constexpr_non_literal_atomic__
void push_back(const value_type& x) {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t old_store_ref = store_ref;
if( old_store_ref->capacity_reached() ) {
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
// grow and swap all refs
storage_ref_t new_store_ref = std::make_shared<storage_t>( *old_store_ref, old_store_ref->grow_capacity(),
old_store_ref->growth_factor(),
@@ -628,10 +658,11 @@ namespace jau {
* This write operation uses a mutex lock and is blocking this instances' write operations only.
* </p>
*/
+ __constexpr_non_literal_atomic__
void push_back(value_type&& x) {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t old_store_ref = store_ref;
if( old_store_ref->capacity_reached() ) {
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
// grow and swap all refs
storage_ref_t new_store_ref = std::make_shared<storage_t>( *old_store_ref, old_store_ref->grow_capacity(),
old_store_ref->growth_factor(),
@@ -657,10 +688,11 @@ namespace jau {
* @param last last foreign input-iterator to range of value_type [first, last)
*/
template< class InputIt >
- constexpr void push_back( InputIt first, InputIt last ) {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t old_store_ref = store_ref;
const size_type new_size_ = old_store_ref->size() + size_type(last - first);
+ __constexpr_non_literal_atomic__
+ void push_back( InputIt first, InputIt last ) {
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
if( new_size_ > old_store_ref->capacity() ) {
// grow and swap all refs
@@ -739,8 +771,9 @@ namespace jau {
* @param comparator the equal comparator to return true if both given elements are equal
* @return true if the element has been uniquely added, otherwise false
*/
+ __constexpr_non_literal_atomic__
bool push_back_unique(const value_type& x, equal_comparator comparator) {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
for(auto it = store_ref->begin(); it != store_ref->end(); ) {
if( comparator( *it, x ) ) {
return false; // already included
@@ -776,9 +809,10 @@ namespace jau {
* @param comparator the equal comparator to return true if both given elements are equal
* @return number of erased elements
*/
+ __constexpr_non_literal_atomic__
int erase_matching(const value_type& x, const bool all_matching, equal_comparator comparator) {
int count = 0;
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t new_store_ref = std::make_shared<storage_t>( *store_ref );
for(auto it = new_store_ref->begin(); it != new_store_ref->end(); ) {
if( comparator( *it, x ) ) {
@@ -809,8 +843,9 @@ namespace jau {
* @param i the position within this store
* @param x the value to be assigned to the object at the given position
*/
+ __constexpr_non_literal_atomic__
void put(size_type i, const value_type& x) {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t new_store_ref = std::make_shared<storage_t>( *store_ref );
new_store_ref->at(i) = x;
{
@@ -830,8 +865,9 @@ namespace jau {
* @param i the position within this store
* @param x the value to be assigned to the object at the given position
*/
+ __constexpr_non_literal_atomic__
void put(size_type i, value_type&& x) {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t new_store_ref = std::make_shared<storage_t>( *store_ref );
new_store_ref->at(i) = std::move(x);
{
diff --git a/include/jau/cow_vector.hpp b/include/jau/cow_vector.hpp
index 35e7ac7..69143f8 100644
--- a/include/jau/cow_vector.hpp
+++ b/include/jau/cow_vector.hpp
@@ -37,6 +37,7 @@
#include <vector>
#include <algorithm>
+#include <jau/cpp_lang_macros.hpp>
#include <jau/debug.hpp>
#include <jau/basic_types.hpp>
#include <jau/ordered_atomic.hpp>
@@ -134,8 +135,8 @@ namespace jau {
static constexpr size_type DIFF_MAX = std::numeric_limits<difference_type>::max();
storage_ref_t store_ref;
- sc_atomic_bool sync_atomic;
- std::recursive_mutex mtx_write;
+ mutable sc_atomic_bool sync_atomic;
+ mutable std::recursive_mutex mtx_write;
public:
// ctor
@@ -152,11 +153,12 @@ namespace jau {
constexpr cow_vector(size_type n, const value_type& value, const allocator_type& a = allocator_type())
: store_ref( std::make_shared<storage_t>(n, value, a) ), sync_atomic(false) { }
- constexpr cow_vector(const cow_vector& x)
+ __constexpr_non_literal_atomic__
+ cow_vector(const cow_vector& x)
: sync_atomic(false) {
storage_ref_t x_store_ref;
{
- sc_atomic_critical sync_x( const_cast<cow_vector *>(&x)->sync_atomic );
+ sc_atomic_critical sync_x( x.sync_atomic );
x_store_ref = x.store_ref;
}
store_ref = std::make_shared<storage_t>( *x_store_ref, x_store_ref->get_allocator() );
@@ -212,8 +214,9 @@ namespace jau {
* @see jau::cow_vector::copy_store()
* @see jau::cow_vector::set_store()
*/
+ __constexpr_non_literal_atomic__
storage_ref_t copy_store() {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
return std::make_shared<storage_t>( *store_ref, store_ref->get_allocator() );
}
@@ -229,7 +232,7 @@ namespace jau {
* cow_vector<std::shared_ptr<Thing>> list;
* ...
* {
- * const std::lock_guard<std::recursive_mutex> lock(list.get_write_mutex());
+ * std::lock_guard<std::recursive_mutex> lock(list.get_write_mutex());
* std::shared_ptr<std::vector<std::shared_ptr<Thing>>> snapshot = list.copy_store();
* ...
* some fancy mutation
@@ -244,8 +247,9 @@ namespace jau {
* @see jau::cow_vector::copy_store()
* @see jau::cow_vector::set_store()
*/
- constexpr void set_store(storage_ref_t && new_store_ref) noexcept {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ __constexpr_non_literal_atomic__
+ void set_store(storage_ref_t && new_store_ref) noexcept {
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
sc_atomic_critical sync(sync_atomic);
store_ref = std::move( new_store_ref );
}
@@ -262,8 +266,9 @@ namespace jau {
* </p>
* @see jau::for_each_cow
*/
- constexpr storage_ref_t get_snapshot() const noexcept {
- sc_atomic_critical sync( const_cast<cow_vector *>(this)->sync_atomic );
+ __constexpr_non_literal_atomic__
+ storage_ref_t get_snapshot() const noexcept {
+ sc_atomic_critical sync( sync_atomic );
return store_ref;
}
@@ -298,12 +303,13 @@ namespace jau {
// read access
allocator_type get_allocator() const noexcept {
- sc_atomic_critical sync( const_cast<cow_vector *>(this)->sync_atomic );
+ sc_atomic_critical sync( sync_atomic );
return store_ref->get_allocator();
}
- constexpr size_type capacity() const noexcept {
- sc_atomic_critical sync( const_cast<cow_vector *>(this)->sync_atomic );
+ __constexpr_non_literal_atomic__
+ size_type capacity() const noexcept {
+ sc_atomic_critical sync( sync_atomic );
return store_ref->capacity();
}
@@ -313,8 +319,9 @@ namespace jau {
* This read operation is <i>lock-free</i>.
* </p>
*/
- constexpr bool empty() const noexcept {
- sc_atomic_critical sync( const_cast<cow_vector *>(this)->sync_atomic );
+ __constexpr_non_literal_atomic__
+ bool empty() const noexcept {
+ sc_atomic_critical sync( sync_atomic );
return store_ref->empty();
}
@@ -324,8 +331,9 @@ namespace jau {
* This read operation is <i>lock-free</i>.
* </p>
*/
- constexpr size_type size() const noexcept {
- sc_atomic_critical sync( const_cast<cow_vector *>(this)->sync_atomic );
+ __constexpr_non_literal_atomic__
+ size_type size() const noexcept {
+ sc_atomic_critical sync( sync_atomic );
return store_ref->size();
}
@@ -388,7 +396,7 @@ namespace jau {
// write access
void reserve(size_type new_capacity) {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t old_store_ref = store_ref;
if( new_capacity > old_store_ref->capacity() ) {
storage_ref_t new_store_ref = std::make_shared<storage_t>( *old_store_ref, old_store_ref->get_allocator() );
@@ -405,10 +413,10 @@ namespace jau {
* </p>
*/
cow_vector& operator=(const cow_vector& x) {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t x_store_ref;
{
- sc_atomic_critical sync_x( const_cast<cow_vector *>(&x)->sync_atomic );
+ sc_atomic_critical sync_x( x.sync_atomic );
x_store_ref = x.store_ref;
}
storage_ref_t new_store_ref = std::make_shared<storage_t>( *x_store_ref, x_store_ref->get_allocator() );
@@ -445,8 +453,9 @@ namespace jau {
* This write operation uses a mutex lock and is blocking this instances' write operations.
* </p>
*/
+ __constexpr_non_literal_atomic__
void clear() noexcept {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t new_store_ref = std::make_shared<storage_t>();
{
sc_atomic_critical sync(sync_atomic);
@@ -460,6 +469,7 @@ namespace jau {
* This write operation uses a mutex lock and is blocking both cow_vector instance's write operations.
* </p>
*/
+ __constexpr_non_literal_atomic__
void swap(cow_vector& x) noexcept {
std::unique_lock<std::recursive_mutex> lock(mtx_write, std::defer_lock); // utilize std::lock(a, b), allowing mixed order waiting on either object
std::unique_lock<std::recursive_mutex> lock_x(x.mtx_write, std::defer_lock); // otherwise RAII-style relinquish via destructor
@@ -479,8 +489,9 @@ namespace jau {
* This write operation uses a mutex lock and is blocking this instances' write operations only.
* </p>
*/
+ __constexpr_non_literal_atomic__
void pop_back() noexcept {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t old_store_ref = store_ref;
if( 0 < old_store_ref->size() ) {
storage_ref_t new_store_ref = std::make_shared<storage_t>( *old_store_ref, old_store_ref->get_allocator() );
@@ -499,8 +510,9 @@ namespace jau {
* </p>
* @param x the value to be added at the tail.
*/
+ __constexpr_non_literal_atomic__
void push_back(const value_type& x) {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t new_store_ref = std::make_shared<storage_t>( *store_ref, store_ref->get_allocator() );
new_store_ref->push_back(x);
{
@@ -515,8 +527,9 @@ namespace jau {
* This write operation uses a mutex lock and is blocking this instances' write operations only.
* </p>
*/
+ __constexpr_non_literal_atomic__
void push_back(value_type&& x) {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t new_store_ref = std::make_shared<storage_t>( *store_ref, store_ref->get_allocator() );
new_store_ref->push_back( std::move(x) );
{
@@ -557,8 +570,9 @@ namespace jau {
* @param comparator the equal comparator to return true if both given elements are equal
* @return true if the element has been uniquely added, otherwise false
*/
+ __constexpr_non_literal_atomic__
bool push_back_unique(const value_type& x, equal_comparator comparator) {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
for(auto it = store_ref->begin(); it != store_ref->end(); ) {
if( comparator( *it, x ) ) {
return false; // already included
@@ -599,9 +613,10 @@ namespace jau {
* @param comparator the equal comparator to return true if both given elements are equal
* @return number of erased elements
*/
+ __constexpr_non_literal_atomic__
int erase_matching(const value_type& x, const bool all_matching, equal_comparator comparator) {
int count = 0;
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t new_store_ref = std::make_shared<storage_t>( *store_ref, store_ref->get_allocator() );
for(auto it = new_store_ref->begin(); it != new_store_ref->end(); ) {
if( comparator( *it, x ) ) {
@@ -632,8 +647,9 @@ namespace jau {
* @param i the position within this store
* @param x the value to be assigned to the object at the given position
*/
+ __constexpr_non_literal_atomic__
void put(size_type i, const value_type& x) {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t new_store_ref = std::make_shared<storage_t>( *store_ref, store_ref->get_allocator() );
new_store_ref->at(i) = x;
{
@@ -653,8 +669,9 @@ namespace jau {
* @param i the position within this store
* @param x the value to be assigned to the object at the given position
*/
+ __constexpr_non_literal_atomic__
void put(size_type i, value_type&& x) {
- const std::lock_guard<std::recursive_mutex> lock(mtx_write);
+ std::lock_guard<std::recursive_mutex> lock(mtx_write);
storage_ref_t new_store_ref = std::make_shared<storage_t>( *store_ref, store_ref->get_allocator() );
new_store_ref->at(i) = std::move(x);
{
diff --git a/include/jau/cpp_lang_macros.hpp b/include/jau/cpp_lang_macros.hpp
new file mode 100644
index 0000000..2f090b2
--- /dev/null
+++ b/include/jau/cpp_lang_macros.hpp
@@ -0,0 +1,56 @@
+/*
+ * Copyright (c) 2020 Gothel Software e.K.
+ *
+ * Permission is hereby granted, free of charge, to any person obtaining
+ * a copy of this software and associated documentation files (the
+ * "Software"), to deal in the Software without restriction, including
+ * without limitation the rights to use, copy, modify, merge, publish,
+ * distribute, sublicense, and/or sell copies of the Software, and to
+ * permit persons to whom the Software is furnished to do so, subject to
+ * the following conditions:
+ *
+ * The above copyright notice and this permission notice shall be
+ * included in all copies or substantial portions of the Software.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
+ * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
+ * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
+ * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
+ * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
+ * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
+ * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
+ *
+ */
+
+#ifndef CPP_LANG_MACROS_HPP_
+#define CPP_LANG_MACROS_HPP_
+
+#if __cplusplus > 201703L
+ #define __constexpr_cxx20_ constexpr
+ #warning CXX 20 detected but not evaluated yet
+#else
+ #define __constexpr_cxx20_
+#endif
+
+ /**
+ * Used when attempting to define a function 'constexpr',
+ * but uses non-literal variables, such as std::lock_guard etc.
+ * As these can't be evaluated at compile time, the standard does
+ * not allow using 'constexpr' here.
+ * <p>
+ * Empty until standard defines otherwise.
+ * </p>
+ */
+ #define __constexpr_non_literal_var__
+
+ /**
+ * See __constexpr_non_literal_var__
+ */
+ #define __constexpr_non_literal_atomic__
+
+ /**
+ * See __constexpr_non_literal_var__
+ */
+ #define __constexpr_non_literal_mutex__
+
+#endif /* CPP_LANG_MACROS_HPP_ */