summaryrefslogtreecommitdiffstats
path: root/core/os
diff options
context:
space:
mode:
Diffstat (limited to 'core/os')
-rw-r--r--core/os/condition_variable.h8
-rw-r--r--core/os/main_loop.h1
-rw-r--r--core/os/memory.cpp33
-rw-r--r--core/os/memory.h24
-rw-r--r--core/os/mutex.h24
-rw-r--r--core/os/os.h14
-rw-r--r--core/os/pool_allocator.cpp588
-rw-r--r--core/os/pool_allocator.h148
-rw-r--r--core/os/safe_binary_mutex.h95
9 files changed, 156 insertions, 779 deletions
diff --git a/core/os/condition_variable.h b/core/os/condition_variable.h
index fa1355e98c..c819fa6b40 100644
--- a/core/os/condition_variable.h
+++ b/core/os/condition_variable.h
@@ -32,6 +32,7 @@
#define CONDITION_VARIABLE_H
#include "core/os/mutex.h"
+#include "core/os/safe_binary_mutex.h"
#ifdef THREADS_ENABLED
@@ -56,7 +57,12 @@ class ConditionVariable {
public:
template <typename BinaryMutexT>
_ALWAYS_INLINE_ void wait(const MutexLock<BinaryMutexT> &p_lock) const {
- condition.wait(const_cast<THREADING_NAMESPACE::unique_lock<THREADING_NAMESPACE::mutex> &>(p_lock.lock));
+ condition.wait(const_cast<THREADING_NAMESPACE::unique_lock<THREADING_NAMESPACE::mutex> &>(p_lock._get_lock()));
+ }
+
+ template <int Tag>
+ _ALWAYS_INLINE_ void wait(const MutexLock<SafeBinaryMutex<Tag>> &p_lock) const {
+ condition.wait(const_cast<THREADING_NAMESPACE::unique_lock<THREADING_NAMESPACE::mutex> &>(p_lock.mutex._get_lock()));
}
_ALWAYS_INLINE_ void notify_one() const {
diff --git a/core/os/main_loop.h b/core/os/main_loop.h
index e48541d074..9c22cbaf3c 100644
--- a/core/os/main_loop.h
+++ b/core/os/main_loop.h
@@ -64,6 +64,7 @@ public:
virtual void initialize();
virtual void iteration_prepare() {}
virtual bool physics_process(double p_time);
+ virtual void iteration_end() {}
virtual bool process(double p_time);
virtual void finalize();
diff --git a/core/os/memory.cpp b/core/os/memory.cpp
index 32c316e58e..dae0a31fe0 100644
--- a/core/os/memory.cpp
+++ b/core/os/memory.cpp
@@ -35,6 +35,7 @@
#include <stdio.h>
#include <stdlib.h>
+#include <string.h>
void *operator new(size_t p_size, const char *p_description) {
return Memory::alloc_static(p_size, false);
@@ -65,6 +66,38 @@ SafeNumeric<uint64_t> Memory::max_usage;
SafeNumeric<uint64_t> Memory::alloc_count;
+inline bool is_power_of_2(size_t x) { return x && ((x & (x - 1U)) == 0U); }
+
+void *Memory::alloc_aligned_static(size_t p_bytes, size_t p_alignment) {
+ DEV_ASSERT(is_power_of_2(p_alignment));
+
+ void *p1, *p2;
+ if ((p1 = (void *)malloc(p_bytes + p_alignment - 1 + sizeof(uint32_t))) == nullptr) {
+ return nullptr;
+ }
+
+ p2 = (void *)(((uintptr_t)p1 + sizeof(uint32_t) + p_alignment - 1) & ~((p_alignment)-1));
+ *((uint32_t *)p2 - 1) = (uint32_t)((uintptr_t)p2 - (uintptr_t)p1);
+ return p2;
+}
+
+void *Memory::realloc_aligned_static(void *p_memory, size_t p_bytes, size_t p_prev_bytes, size_t p_alignment) {
+ if (p_memory == nullptr) {
+ return alloc_aligned_static(p_bytes, p_alignment);
+ }
+
+ void *ret = alloc_aligned_static(p_bytes, p_alignment);
+ memcpy(ret, p_memory, p_prev_bytes);
+ free_aligned_static(p_memory);
+ return ret;
+}
+
+void Memory::free_aligned_static(void *p_memory) {
+ uint32_t offset = *((uint32_t *)p_memory - 1);
+ void *p = (void *)((uint8_t *)p_memory - offset);
+ free(p);
+}
+
void *Memory::alloc_static(size_t p_bytes, bool p_pad_align) {
#ifdef DEBUG_ENABLED
bool prepad = true;
diff --git a/core/os/memory.h b/core/os/memory.h
index d03e08d785..033e417cb5 100644
--- a/core/os/memory.h
+++ b/core/os/memory.h
@@ -62,6 +62,30 @@ public:
static void *realloc_static(void *p_memory, size_t p_bytes, bool p_pad_align = false);
static void free_static(void *p_ptr, bool p_pad_align = false);
+ // ↓ return value of alloc_aligned_static
+ // ┌─────────────────┬─────────┬─────────┬──────────────────┐
+ // │ padding (up to │ uint32_t│ void* │ padding (up to │
+ // │ p_alignment - 1)│ offset │ p_bytes │ p_alignment - 1) │
+ // └─────────────────┴─────────┴─────────┴──────────────────┘
+ //
+ // alloc_aligned_static will allocate p_bytes + p_alignment - 1 + sizeof(uint32_t) and
+ // then offset the pointer until alignment is satisfied.
+ //
+ // This offset is stored before the start of the returned ptr so we can retrieve the original/real
+ // start of the ptr in order to free it.
+ //
+ // The rest is wasted as padding in the beginning and end of the ptr. The sum of padding at
+ // both start and end of the block must add exactly to p_alignment - 1.
+ //
+ // p_alignment MUST be a power of 2.
+ static void *alloc_aligned_static(size_t p_bytes, size_t p_alignment);
+ static void *realloc_aligned_static(void *p_memory, size_t p_bytes, size_t p_prev_bytes, size_t p_alignment);
+ // Pass the ptr returned by alloc_aligned_static to free it.
+ // e.g.
+ // void *data = realloc_aligned_static( bytes, 16 );
+ // free_aligned_static( data );
+ static void free_aligned_static(void *p_memory);
+
static uint64_t get_mem_available();
static uint64_t get_mem_usage();
static uint64_t get_mem_max_usage();
diff --git a/core/os/mutex.h b/core/os/mutex.h
index 3e7aa81bc1..a968fd7029 100644
--- a/core/os/mutex.h
+++ b/core/os/mutex.h
@@ -72,13 +72,28 @@ public:
template <typename MutexT>
class MutexLock {
- friend class ConditionVariable;
-
- THREADING_NAMESPACE::unique_lock<typename MutexT::StdMutexType> lock;
+ mutable THREADING_NAMESPACE::unique_lock<typename MutexT::StdMutexType> lock;
public:
explicit MutexLock(const MutexT &p_mutex) :
lock(p_mutex.mutex) {}
+
+ // Clarification: all the funny syntax is needed so this function exists only for binary mutexes.
+ template <typename T = MutexT>
+ _ALWAYS_INLINE_ THREADING_NAMESPACE::unique_lock<THREADING_NAMESPACE::mutex> &_get_lock(
+ typename std::enable_if<std::is_same<T, THREADING_NAMESPACE::mutex>::value> * = nullptr) const {
+ return lock;
+ }
+
+ _ALWAYS_INLINE_ void temp_relock() const {
+ lock.lock();
+ }
+
+ _ALWAYS_INLINE_ void temp_unlock() const {
+ lock.unlock();
+ }
+
+ // TODO: Implement a `try_temp_relock` if needed (will also need a dummy method below).
};
using Mutex = MutexImpl<THREADING_NAMESPACE::recursive_mutex>; // Recursive, for general use
@@ -104,6 +119,9 @@ template <typename MutexT>
class MutexLock {
public:
MutexLock(const MutexT &p_mutex) {}
+
+ void temp_relock() const {}
+ void temp_unlock() const {}
};
using Mutex = MutexImpl;
diff --git a/core/os/os.h b/core/os/os.h
index 63cc6ed50e..30d2a4266f 100644
--- a/core/os/os.h
+++ b/core/os/os.h
@@ -75,6 +75,7 @@ class OS {
int _display_driver_id = -1;
String _current_rendering_driver_name;
String _current_rendering_method;
+ bool _is_gles_over_gl = false;
RemoteFilesystemClient default_rfs;
@@ -111,9 +112,6 @@ protected:
virtual void initialize() = 0;
virtual void initialize_joypads() = 0;
- void set_current_rendering_driver_name(const String &p_driver_name) { _current_rendering_driver_name = p_driver_name; }
- void set_current_rendering_method(const String &p_name) { _current_rendering_method = p_name; }
-
void set_display_driver_id(int p_display_driver_id) { _display_driver_id = p_display_driver_id; }
virtual void set_main_loop(MainLoop *p_main_loop) = 0;
@@ -131,12 +129,18 @@ public:
static OS *get_singleton();
+ void set_current_rendering_driver_name(const String &p_driver_name) { _current_rendering_driver_name = p_driver_name; }
+ void set_current_rendering_method(const String &p_name) { _current_rendering_method = p_name; }
+ void set_gles_over_gl(bool p_enabled) { _is_gles_over_gl = p_enabled; }
+
String get_current_rendering_driver_name() const { return _current_rendering_driver_name; }
String get_current_rendering_method() const { return _current_rendering_method; }
+ bool get_gles_over_gl() const { return _is_gles_over_gl; }
int get_display_driver_id() const { return _display_driver_id; }
virtual Vector<String> get_video_adapter_driver_info() const = 0;
+ virtual bool get_user_prefers_integrated_gpu() const { return false; }
void print_error(const char *p_function, const char *p_file, int p_line, const char *p_code, const char *p_rationale, bool p_editor_notify = false, Logger::ErrorType p_type = Logger::ERR_ERROR);
void print(const char *p_format, ...) _PRINTF_FORMAT_ATTRIBUTE_2_3;
@@ -178,7 +182,7 @@ public:
virtual Vector<String> get_system_font_path_for_text(const String &p_font_name, const String &p_text, const String &p_locale = String(), const String &p_script = String(), int p_weight = 400, int p_stretch = 100, bool p_italic = false) const { return Vector<String>(); };
virtual String get_executable_path() const;
virtual Error execute(const String &p_path, const List<String> &p_arguments, String *r_pipe = nullptr, int *r_exitcode = nullptr, bool read_stderr = false, Mutex *p_pipe_mutex = nullptr, bool p_open_console = false) = 0;
- virtual Dictionary execute_with_pipe(const String &p_path, const List<String> &p_arguments) { return Dictionary(); }
+ virtual Dictionary execute_with_pipe(const String &p_path, const List<String> &p_arguments, bool p_blocking = true) { return Dictionary(); }
virtual Error create_process(const String &p_path, const List<String> &p_arguments, ProcessID *r_child_id = nullptr, bool p_open_console = false) = 0;
virtual Error create_instance(const List<String> &p_arguments, ProcessID *r_child_id = nullptr) { return create_process(get_executable_path(), p_arguments, r_child_id); };
virtual Error kill(const ProcessID &p_pid) = 0;
@@ -328,8 +332,6 @@ public:
virtual void benchmark_end_measure(const String &p_context, const String &p_what);
virtual void benchmark_dump();
- virtual void process_and_drop_events() {}
-
virtual Error setup_remote_filesystem(const String &p_server_host, int p_port, const String &p_password, String &r_project_path);
enum PreferredTextureFormat {
diff --git a/core/os/pool_allocator.cpp b/core/os/pool_allocator.cpp
deleted file mode 100644
index 9a993cd14f..0000000000
--- a/core/os/pool_allocator.cpp
+++ /dev/null
@@ -1,588 +0,0 @@
-/**************************************************************************/
-/* pool_allocator.cpp */
-/**************************************************************************/
-/* This file is part of: */
-/* GODOT ENGINE */
-/* https://godotengine.org */
-/**************************************************************************/
-/* Copyright (c) 2014-present Godot Engine contributors (see AUTHORS.md). */
-/* Copyright (c) 2007-2014 Juan Linietsky, Ariel Manzur. */
-/* */
-/* Permission is hereby granted, free of charge, to any person obtaining */
-/* a copy of this software and associated documentation files (the */
-/* "Software"), to deal in the Software without restriction, including */
-/* without limitation the rights to use, copy, modify, merge, publish, */
-/* distribute, sublicense, and/or sell copies of the Software, and to */
-/* permit persons to whom the Software is furnished to do so, subject to */
-/* the following conditions: */
-/* */
-/* The above copyright notice and this permission notice shall be */
-/* included in all copies or substantial portions of the Software. */
-/* */
-/* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, */
-/* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF */
-/* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. */
-/* IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY */
-/* CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, */
-/* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE */
-/* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */
-/**************************************************************************/
-
-#include "pool_allocator.h"
-
-#include "core/error/error_macros.h"
-#include "core/os/memory.h"
-#include "core/os/os.h"
-#include "core/string/print_string.h"
-
-#define COMPACT_CHUNK(m_entry, m_to_pos) \
- if constexpr (true) { \
- void *_dst = &((unsigned char *)pool)[m_to_pos]; \
- void *_src = &((unsigned char *)pool)[(m_entry).pos]; \
- memmove(_dst, _src, aligned((m_entry).len)); \
- (m_entry).pos = m_to_pos; \
- } else \
- ((void)0)
-
-void PoolAllocator::mt_lock() const {
-}
-
-void PoolAllocator::mt_unlock() const {
-}
-
-bool PoolAllocator::get_free_entry(EntryArrayPos *p_pos) {
- if (entry_count == entry_max) {
- return false;
- }
-
- for (int i = 0; i < entry_max; i++) {
- if (entry_array[i].len == 0) {
- *p_pos = i;
- return true;
- }
- }
-
- ERR_PRINT("Out of memory Chunks!");
-
- return false; //
-}
-
-/**
- * Find a hole
- * @param p_pos The hole is behind the block pointed by this variable upon return. if pos==entry_count, then allocate at end
- * @param p_for_size hole size
- * @return false if hole found, true if no hole found
- */
-bool PoolAllocator::find_hole(EntryArrayPos *p_pos, int p_for_size) {
- /* position where previous entry ends. Defaults to zero (begin of pool) */
-
- int prev_entry_end_pos = 0;
-
- for (int i = 0; i < entry_count; i++) {
- Entry &entry = entry_array[entry_indices[i]];
-
- /* determine hole size to previous entry */
-
- int hole_size = entry.pos - prev_entry_end_pos;
-
- /* determine if what we want fits in that hole */
- if (hole_size >= p_for_size) {
- *p_pos = i;
- return true;
- }
-
- /* prepare for next one */
- prev_entry_end_pos = entry_end(entry);
- }
-
- /* No holes between entries, check at the end..*/
-
- if ((pool_size - prev_entry_end_pos) >= p_for_size) {
- *p_pos = entry_count;
- return true;
- }
-
- return false;
-}
-
-void PoolAllocator::compact(int p_up_to) {
- uint32_t prev_entry_end_pos = 0;
-
- if (p_up_to < 0) {
- p_up_to = entry_count;
- }
- for (int i = 0; i < p_up_to; i++) {
- Entry &entry = entry_array[entry_indices[i]];
-
- /* determine hole size to previous entry */
-
- int hole_size = entry.pos - prev_entry_end_pos;
-
- /* if we can compact, do it */
- if (hole_size > 0 && !entry.lock) {
- COMPACT_CHUNK(entry, prev_entry_end_pos);
- }
-
- /* prepare for next one */
- prev_entry_end_pos = entry_end(entry);
- }
-}
-
-void PoolAllocator::compact_up(int p_from) {
- uint32_t next_entry_end_pos = pool_size; // - static_area_size;
-
- for (int i = entry_count - 1; i >= p_from; i--) {
- Entry &entry = entry_array[entry_indices[i]];
-
- /* determine hole size for next entry */
-
- int hole_size = next_entry_end_pos - (entry.pos + aligned(entry.len));
-
- /* if we can compact, do it */
- if (hole_size > 0 && !entry.lock) {
- COMPACT_CHUNK(entry, (next_entry_end_pos - aligned(entry.len)));
- }
-
- /* prepare for next one */
- next_entry_end_pos = entry.pos;
- }
-}
-
-bool PoolAllocator::find_entry_index(EntryIndicesPos *p_map_pos, const Entry *p_entry) {
- EntryArrayPos entry_pos = entry_max;
-
- for (int i = 0; i < entry_count; i++) {
- if (&entry_array[entry_indices[i]] == p_entry) {
- entry_pos = i;
- break;
- }
- }
-
- if (entry_pos == entry_max) {
- return false;
- }
-
- *p_map_pos = entry_pos;
- return true;
-}
-
-PoolAllocator::ID PoolAllocator::alloc(int p_size) {
- ERR_FAIL_COND_V(p_size < 1, POOL_ALLOCATOR_INVALID_ID);
- ERR_FAIL_COND_V(p_size > free_mem, POOL_ALLOCATOR_INVALID_ID);
-
- mt_lock();
-
- if (entry_count == entry_max) {
- mt_unlock();
- ERR_PRINT("entry_count==entry_max");
- return POOL_ALLOCATOR_INVALID_ID;
- }
-
- int size_to_alloc = aligned(p_size);
-
- EntryIndicesPos new_entry_indices_pos;
-
- if (!find_hole(&new_entry_indices_pos, size_to_alloc)) {
- /* No hole could be found, try compacting mem */
- compact();
- /* Then search again */
-
- if (!find_hole(&new_entry_indices_pos, size_to_alloc)) {
- mt_unlock();
- ERR_FAIL_V_MSG(POOL_ALLOCATOR_INVALID_ID, "Memory can't be compacted further.");
- }
- }
-
- EntryArrayPos new_entry_array_pos;
-
- bool found_free_entry = get_free_entry(&new_entry_array_pos);
-
- if (!found_free_entry) {
- mt_unlock();
- ERR_FAIL_V_MSG(POOL_ALLOCATOR_INVALID_ID, "No free entry found in PoolAllocator.");
- }
-
- /* move all entry indices up, make room for this one */
- for (int i = entry_count; i > new_entry_indices_pos; i--) {
- entry_indices[i] = entry_indices[i - 1];
- }
-
- entry_indices[new_entry_indices_pos] = new_entry_array_pos;
-
- entry_count++;
-
- Entry &entry = entry_array[entry_indices[new_entry_indices_pos]];
-
- entry.len = p_size;
- entry.pos = (new_entry_indices_pos == 0) ? 0 : entry_end(entry_array[entry_indices[new_entry_indices_pos - 1]]); //alloc either at beginning or end of previous
- entry.lock = 0;
- entry.check = (check_count++) & CHECK_MASK;
- free_mem -= size_to_alloc;
- if (free_mem < free_mem_peak) {
- free_mem_peak = free_mem;
- }
-
- ID retval = (entry_indices[new_entry_indices_pos] << CHECK_BITS) | entry.check;
- mt_unlock();
-
- //ERR_FAIL_COND_V( (uintptr_t)get(retval)%align != 0, retval );
-
- return retval;
-}
-
-PoolAllocator::Entry *PoolAllocator::get_entry(ID p_mem) {
- unsigned int check = p_mem & CHECK_MASK;
- int entry = p_mem >> CHECK_BITS;
- ERR_FAIL_INDEX_V(entry, entry_max, nullptr);
- ERR_FAIL_COND_V(entry_array[entry].check != check, nullptr);
- ERR_FAIL_COND_V(entry_array[entry].len == 0, nullptr);
-
- return &entry_array[entry];
-}
-
-const PoolAllocator::Entry *PoolAllocator::get_entry(ID p_mem) const {
- unsigned int check = p_mem & CHECK_MASK;
- int entry = p_mem >> CHECK_BITS;
- ERR_FAIL_INDEX_V(entry, entry_max, nullptr);
- ERR_FAIL_COND_V(entry_array[entry].check != check, nullptr);
- ERR_FAIL_COND_V(entry_array[entry].len == 0, nullptr);
-
- return &entry_array[entry];
-}
-
-void PoolAllocator::free(ID p_mem) {
- mt_lock();
- Entry *e = get_entry(p_mem);
- if (!e) {
- mt_unlock();
- ERR_PRINT("!e");
- return;
- }
- if (e->lock) {
- mt_unlock();
- ERR_PRINT("e->lock");
- return;
- }
-
- EntryIndicesPos entry_indices_pos;
-
- bool index_found = find_entry_index(&entry_indices_pos, e);
- if (!index_found) {
- mt_unlock();
- ERR_FAIL_COND(!index_found);
- }
-
- for (int i = entry_indices_pos; i < (entry_count - 1); i++) {
- entry_indices[i] = entry_indices[i + 1];
- }
-
- entry_count--;
- free_mem += aligned(e->len);
- e->clear();
- mt_unlock();
-}
-
-int PoolAllocator::get_size(ID p_mem) const {
- int size;
- mt_lock();
-
- const Entry *e = get_entry(p_mem);
- if (!e) {
- mt_unlock();
- ERR_PRINT("!e");
- return 0;
- }
-
- size = e->len;
-
- mt_unlock();
-
- return size;
-}
-
-Error PoolAllocator::resize(ID p_mem, int p_new_size) {
- mt_lock();
- Entry *e = get_entry(p_mem);
-
- if (!e) {
- mt_unlock();
- ERR_FAIL_NULL_V(e, ERR_INVALID_PARAMETER);
- }
-
- if (needs_locking && e->lock) {
- mt_unlock();
- ERR_FAIL_COND_V(e->lock, ERR_ALREADY_IN_USE);
- }
-
- uint32_t alloc_size = aligned(p_new_size);
-
- if ((uint32_t)aligned(e->len) == alloc_size) {
- e->len = p_new_size;
- mt_unlock();
- return OK;
- } else if (e->len > (uint32_t)p_new_size) {
- free_mem += aligned(e->len);
- free_mem -= alloc_size;
- e->len = p_new_size;
- mt_unlock();
- return OK;
- }
-
- //p_new_size = align(p_new_size)
- int _free = free_mem; // - static_area_size;
-
- if (uint32_t(_free + aligned(e->len)) < alloc_size) {
- mt_unlock();
- ERR_FAIL_V(ERR_OUT_OF_MEMORY);
- }
-
- EntryIndicesPos entry_indices_pos;
-
- bool index_found = find_entry_index(&entry_indices_pos, e);
-
- if (!index_found) {
- mt_unlock();
- ERR_FAIL_COND_V(!index_found, ERR_BUG);
- }
-
- //no need to move stuff around, it fits before the next block
- uint32_t next_pos;
- if (entry_indices_pos + 1 == entry_count) {
- next_pos = pool_size; // - static_area_size;
- } else {
- next_pos = entry_array[entry_indices[entry_indices_pos + 1]].pos;
- }
-
- if ((next_pos - e->pos) > alloc_size) {
- free_mem += aligned(e->len);
- e->len = p_new_size;
- free_mem -= alloc_size;
- mt_unlock();
- return OK;
- }
- //it doesn't fit, compact around BEFORE current index (make room behind)
-
- compact(entry_indices_pos + 1);
-
- if ((next_pos - e->pos) > alloc_size) {
- //now fits! hooray!
- free_mem += aligned(e->len);
- e->len = p_new_size;
- free_mem -= alloc_size;
- mt_unlock();
- if (free_mem < free_mem_peak) {
- free_mem_peak = free_mem;
- }
- return OK;
- }
-
- //STILL doesn't fit, compact around AFTER current index (make room after)
-
- compact_up(entry_indices_pos + 1);
-
- if ((entry_array[entry_indices[entry_indices_pos + 1]].pos - e->pos) > alloc_size) {
- //now fits! hooray!
- free_mem += aligned(e->len);
- e->len = p_new_size;
- free_mem -= alloc_size;
- mt_unlock();
- if (free_mem < free_mem_peak) {
- free_mem_peak = free_mem;
- }
- return OK;
- }
-
- mt_unlock();
- ERR_FAIL_V(ERR_OUT_OF_MEMORY);
-}
-
-Error PoolAllocator::lock(ID p_mem) {
- if (!needs_locking) {
- return OK;
- }
- mt_lock();
- Entry *e = get_entry(p_mem);
- if (!e) {
- mt_unlock();
- ERR_PRINT("!e");
- return ERR_INVALID_PARAMETER;
- }
- e->lock++;
- mt_unlock();
- return OK;
-}
-
-bool PoolAllocator::is_locked(ID p_mem) const {
- if (!needs_locking) {
- return false;
- }
-
- mt_lock();
- const Entry *e = const_cast<PoolAllocator *>(this)->get_entry(p_mem);
- if (!e) {
- mt_unlock();
- ERR_PRINT("!e");
- return false;
- }
- bool locked = e->lock;
- mt_unlock();
- return locked;
-}
-
-const void *PoolAllocator::get(ID p_mem) const {
- if (!needs_locking) {
- const Entry *e = get_entry(p_mem);
- ERR_FAIL_NULL_V(e, nullptr);
- return &pool[e->pos];
- }
-
- mt_lock();
- const Entry *e = get_entry(p_mem);
-
- if (!e) {
- mt_unlock();
- ERR_FAIL_NULL_V(e, nullptr);
- }
- if (e->lock == 0) {
- mt_unlock();
- ERR_PRINT("e->lock == 0");
- return nullptr;
- }
-
- if ((int)e->pos >= pool_size) {
- mt_unlock();
- ERR_PRINT("e->pos<0 || e->pos>=pool_size");
- return nullptr;
- }
- const void *ptr = &pool[e->pos];
-
- mt_unlock();
-
- return ptr;
-}
-
-void *PoolAllocator::get(ID p_mem) {
- if (!needs_locking) {
- Entry *e = get_entry(p_mem);
- ERR_FAIL_NULL_V(e, nullptr);
- return &pool[e->pos];
- }
-
- mt_lock();
- Entry *e = get_entry(p_mem);
-
- if (!e) {
- mt_unlock();
- ERR_FAIL_NULL_V(e, nullptr);
- }
- if (e->lock == 0) {
- mt_unlock();
- ERR_PRINT("e->lock == 0");
- return nullptr;
- }
-
- if ((int)e->pos >= pool_size) {
- mt_unlock();
- ERR_PRINT("e->pos<0 || e->pos>=pool_size");
- return nullptr;
- }
- void *ptr = &pool[e->pos];
-
- mt_unlock();
-
- return ptr;
-}
-
-void PoolAllocator::unlock(ID p_mem) {
- if (!needs_locking) {
- return;
- }
- mt_lock();
- Entry *e = get_entry(p_mem);
- if (!e) {
- mt_unlock();
- ERR_FAIL_NULL(e);
- }
- if (e->lock == 0) {
- mt_unlock();
- ERR_PRINT("e->lock == 0");
- return;
- }
- e->lock--;
- mt_unlock();
-}
-
-int PoolAllocator::get_used_mem() const {
- return pool_size - free_mem;
-}
-
-int PoolAllocator::get_free_peak() {
- return free_mem_peak;
-}
-
-int PoolAllocator::get_free_mem() {
- return free_mem;
-}
-
-void PoolAllocator::create_pool(void *p_mem, int p_size, int p_max_entries) {
- pool = (uint8_t *)p_mem;
- pool_size = p_size;
-
- entry_array = memnew_arr(Entry, p_max_entries);
- entry_indices = memnew_arr(int, p_max_entries);
- entry_max = p_max_entries;
- entry_count = 0;
-
- free_mem = p_size;
- free_mem_peak = p_size;
-
- check_count = 0;
-}
-
-PoolAllocator::PoolAllocator(int p_size, bool p_needs_locking, int p_max_entries) {
- mem_ptr = memalloc(p_size);
- ERR_FAIL_NULL(mem_ptr);
- align = 1;
- create_pool(mem_ptr, p_size, p_max_entries);
- needs_locking = p_needs_locking;
-}
-
-PoolAllocator::PoolAllocator(void *p_mem, int p_size, int p_align, bool p_needs_locking, int p_max_entries) {
- if (p_align > 1) {
- uint8_t *mem8 = (uint8_t *)p_mem;
- uint64_t ofs = (uint64_t)mem8;
- if (ofs % p_align) {
- int dif = p_align - (ofs % p_align);
- mem8 += p_align - (ofs % p_align);
- p_size -= dif;
- p_mem = (void *)mem8;
- }
- }
-
- create_pool(p_mem, p_size, p_max_entries);
- needs_locking = p_needs_locking;
- align = p_align;
- mem_ptr = nullptr;
-}
-
-PoolAllocator::PoolAllocator(int p_align, int p_size, bool p_needs_locking, int p_max_entries) {
- ERR_FAIL_COND(p_align < 1);
- mem_ptr = Memory::alloc_static(p_size + p_align, true);
- uint8_t *mem8 = (uint8_t *)mem_ptr;
- uint64_t ofs = (uint64_t)mem8;
- if (ofs % p_align) {
- mem8 += p_align - (ofs % p_align);
- }
- create_pool(mem8, p_size, p_max_entries);
- needs_locking = p_needs_locking;
- align = p_align;
-}
-
-PoolAllocator::~PoolAllocator() {
- if (mem_ptr) {
- memfree(mem_ptr);
- }
-
- memdelete_arr(entry_array);
- memdelete_arr(entry_indices);
-}
diff --git a/core/os/pool_allocator.h b/core/os/pool_allocator.h
deleted file mode 100644
index be8b6e061b..0000000000
--- a/core/os/pool_allocator.h
+++ /dev/null
@@ -1,148 +0,0 @@
-/**************************************************************************/
-/* pool_allocator.h */
-/**************************************************************************/
-/* This file is part of: */
-/* GODOT ENGINE */
-/* https://godotengine.org */
-/**************************************************************************/
-/* Copyright (c) 2014-present Godot Engine contributors (see AUTHORS.md). */
-/* Copyright (c) 2007-2014 Juan Linietsky, Ariel Manzur. */
-/* */
-/* Permission is hereby granted, free of charge, to any person obtaining */
-/* a copy of this software and associated documentation files (the */
-/* "Software"), to deal in the Software without restriction, including */
-/* without limitation the rights to use, copy, modify, merge, publish, */
-/* distribute, sublicense, and/or sell copies of the Software, and to */
-/* permit persons to whom the Software is furnished to do so, subject to */
-/* the following conditions: */
-/* */
-/* The above copyright notice and this permission notice shall be */
-/* included in all copies or substantial portions of the Software. */
-/* */
-/* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, */
-/* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF */
-/* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. */
-/* IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY */
-/* CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, */
-/* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE */
-/* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */
-/**************************************************************************/
-
-#ifndef POOL_ALLOCATOR_H
-#define POOL_ALLOCATOR_H
-
-#include "core/typedefs.h"
-
-/**
- * Generic Pool Allocator.
- * This is a generic memory pool allocator, with locking, compacting and alignment. (@TODO alignment)
- * It used as a standard way to manage allocation in a specific region of memory, such as texture memory,
- * audio sample memory, or just any kind of memory overall.
- * (@TODO) abstraction should be greater, because in many platforms, you need to manage a nonreachable memory.
- */
-
-enum {
- POOL_ALLOCATOR_INVALID_ID = -1 ///< default invalid value. use INVALID_ID( id ) to test
-};
-
-class PoolAllocator {
-public:
- typedef int ID;
-
-private:
- enum {
- CHECK_BITS = 8,
- CHECK_LEN = (1 << CHECK_BITS),
- CHECK_MASK = CHECK_LEN - 1
-
- };
-
- struct Entry {
- unsigned int pos = 0;
- unsigned int len = 0;
- unsigned int lock = 0;
- unsigned int check = 0;
-
- inline void clear() {
- pos = 0;
- len = 0;
- lock = 0;
- check = 0;
- }
- Entry() {}
- };
-
- typedef int EntryArrayPos;
- typedef int EntryIndicesPos;
-
- Entry *entry_array = nullptr;
- int *entry_indices = nullptr;
- int entry_max = 0;
- int entry_count = 0;
-
- uint8_t *pool = nullptr;
- void *mem_ptr = nullptr;
- int pool_size = 0;
-
- int free_mem = 0;
- int free_mem_peak = 0;
-
- unsigned int check_count = 0;
- int align = 1;
-
- bool needs_locking = false;
-
- inline int entry_end(const Entry &p_entry) const {
- return p_entry.pos + aligned(p_entry.len);
- }
- inline int aligned(int p_size) const {
- int rem = p_size % align;
- if (rem) {
- p_size += align - rem;
- }
-
- return p_size;
- }
-
- void compact(int p_up_to = -1);
- void compact_up(int p_from = 0);
- bool get_free_entry(EntryArrayPos *p_pos);
- bool find_hole(EntryArrayPos *p_pos, int p_for_size);
- bool find_entry_index(EntryIndicesPos *p_map_pos, const Entry *p_entry);
- Entry *get_entry(ID p_mem);
- const Entry *get_entry(ID p_mem) const;
-
- void create_pool(void *p_mem, int p_size, int p_max_entries);
-
-protected:
- virtual void mt_lock() const; ///< Reimplement for custom mt locking
- virtual void mt_unlock() const; ///< Reimplement for custom mt locking
-
-public:
- enum {
- DEFAULT_MAX_ALLOCS = 4096,
- };
-
- ID alloc(int p_size); ///< Alloc memory, get an ID on success, POOL_ALOCATOR_INVALID_ID on failure
- void free(ID p_mem); ///< Free allocated memory
- Error resize(ID p_mem, int p_new_size); ///< resize a memory chunk
- int get_size(ID p_mem) const;
-
- int get_free_mem(); ///< get free memory
- int get_used_mem() const;
- int get_free_peak(); ///< get free memory
-
- Error lock(ID p_mem); //@todo move this out
- void *get(ID p_mem);
- const void *get(ID p_mem) const;
- void unlock(ID p_mem);
- bool is_locked(ID p_mem) const;
-
- PoolAllocator(int p_size, bool p_needs_locking = false, int p_max_entries = DEFAULT_MAX_ALLOCS);
- PoolAllocator(void *p_mem, int p_size, int p_align = 1, bool p_needs_locking = false, int p_max_entries = DEFAULT_MAX_ALLOCS);
- PoolAllocator(int p_align, int p_size, bool p_needs_locking = false, int p_max_entries = DEFAULT_MAX_ALLOCS);
-
- virtual ~PoolAllocator();
-};
-
-#endif // POOL_ALLOCATOR_H
diff --git a/core/os/safe_binary_mutex.h b/core/os/safe_binary_mutex.h
index 1e98cc074c..74a20043a3 100644
--- a/core/os/safe_binary_mutex.h
+++ b/core/os/safe_binary_mutex.h
@@ -37,6 +37,11 @@
#ifdef THREADS_ENABLED
+#ifdef __clang__
+#pragma clang diagnostic push
+#pragma clang diagnostic ignored "-Wundefined-var-template"
+#endif
+
// A very special kind of mutex, used in scenarios where these
// requirements hold at the same time:
// - Must be used with a condition variable (only binary mutexes are suitable).
@@ -47,69 +52,90 @@
// Also, don't forget to declare the thread_local variable on each use.
template <int Tag>
class SafeBinaryMutex {
- friend class MutexLock<SafeBinaryMutex>;
+ friend class MutexLock<SafeBinaryMutex<Tag>>;
using StdMutexType = THREADING_NAMESPACE::mutex;
mutable THREADING_NAMESPACE::mutex mutex;
- static thread_local uint32_t count;
+
+ struct TLSData {
+ mutable THREADING_NAMESPACE::unique_lock<THREADING_NAMESPACE::mutex> lock;
+ uint32_t count = 0;
+
+ TLSData(SafeBinaryMutex<Tag> &p_mutex) :
+ lock(p_mutex.mutex, THREADING_NAMESPACE::defer_lock) {}
+ };
+ static thread_local TLSData tls_data;
public:
_ALWAYS_INLINE_ void lock() const {
- if (++count == 1) {
- mutex.lock();
+ if (++tls_data.count == 1) {
+ tls_data.lock.lock();
}
}
_ALWAYS_INLINE_ void unlock() const {
- DEV_ASSERT(count);
- if (--count == 0) {
- mutex.unlock();
+ DEV_ASSERT(tls_data.count);
+ if (--tls_data.count == 0) {
+ tls_data.lock.unlock();
}
}
- _ALWAYS_INLINE_ bool try_lock() const {
- if (count) {
- count++;
- return true;
- } else {
- if (mutex.try_lock()) {
- count++;
- return true;
- } else {
- return false;
- }
- }
+ _ALWAYS_INLINE_ THREADING_NAMESPACE::unique_lock<THREADING_NAMESPACE::mutex> &_get_lock() const {
+ return const_cast<THREADING_NAMESPACE::unique_lock<THREADING_NAMESPACE::mutex> &>(tls_data.lock);
+ }
+
+ _ALWAYS_INLINE_ SafeBinaryMutex() {
}
- ~SafeBinaryMutex() {
- DEV_ASSERT(!count);
+ _ALWAYS_INLINE_ ~SafeBinaryMutex() {
+ DEV_ASSERT(!tls_data.count);
}
};
-// This specialization is needed so manual locking and MutexLock can be used
-// at the same time on a SafeBinaryMutex.
template <int Tag>
class MutexLock<SafeBinaryMutex<Tag>> {
friend class ConditionVariable;
- THREADING_NAMESPACE::unique_lock<THREADING_NAMESPACE::mutex> lock;
+ const SafeBinaryMutex<Tag> &mutex;
public:
- _ALWAYS_INLINE_ explicit MutexLock(const SafeBinaryMutex<Tag> &p_mutex) :
- lock(p_mutex.mutex) {
- SafeBinaryMutex<Tag>::count++;
- };
- _ALWAYS_INLINE_ ~MutexLock() {
- SafeBinaryMutex<Tag>::count--;
- };
+ explicit MutexLock(const SafeBinaryMutex<Tag> &p_mutex) :
+ mutex(p_mutex) {
+ mutex.lock();
+ }
+
+ ~MutexLock() {
+ mutex.unlock();
+ }
+
+ _ALWAYS_INLINE_ void temp_relock() const {
+ mutex.lock();
+ }
+
+ _ALWAYS_INLINE_ void temp_unlock() const {
+ mutex.unlock();
+ }
+
+ // TODO: Implement a `try_temp_relock` if needed (will also need a dummy method below).
};
+#ifdef __clang__
+#pragma clang diagnostic pop
+#endif
+
#else // No threads.
template <int Tag>
-class SafeBinaryMutex : public MutexImpl {
- static thread_local uint32_t count;
+class SafeBinaryMutex {
+ struct TLSData {
+ TLSData(SafeBinaryMutex<Tag> &p_mutex) {}
+ };
+ static thread_local TLSData tls_data;
+
+public:
+ void lock() const {}
+ void unlock() const {}
};
template <int Tag>
@@ -117,6 +143,9 @@ class MutexLock<SafeBinaryMutex<Tag>> {
public:
MutexLock(const SafeBinaryMutex<Tag> &p_mutex) {}
~MutexLock() {}
+
+ void temp_relock() const {}
+ void temp_unlock() const {}
};
#endif // THREADS_ENABLED