mirror of
https://github.com/Relintai/sfw.git
synced 2024-11-08 07:52:09 +01:00
300 lines
6.4 KiB
C
300 lines
6.4 KiB
C
|
#ifndef LOCAL_VECTOR_H
|
||
|
#define LOCAL_VECTOR_H
|
||
|
|
||
|
/*************************************************************************/
|
||
|
/* local_vector.h */
|
||
|
/* From https://github.com/Relintai/pandemonium_engine (MIT) */
|
||
|
/*************************************************************************/
|
||
|
|
||
|
//--STRIP
|
||
|
#include "core/pool_vector.h"
|
||
|
#include "core/sort_array.h"
|
||
|
#include "core/vector.h"
|
||
|
#include "core/error_macros.h"
|
||
|
#include "core/memory.h"
|
||
|
//--STRIP
|
||
|
|
||
|
template <class T, class U = uint32_t, bool force_trivial = false>
|
||
|
class LocalVector {
|
||
|
protected:
|
||
|
U count = 0;
|
||
|
U capacity = 0;
|
||
|
T *data = nullptr;
|
||
|
|
||
|
public:
|
||
|
T *ptr() {
|
||
|
return data;
|
||
|
}
|
||
|
|
||
|
const T *ptr() const {
|
||
|
return data;
|
||
|
}
|
||
|
|
||
|
_FORCE_INLINE_ void push_back(T p_elem) {
|
||
|
if (unlikely(count == capacity)) {
|
||
|
if (capacity == 0) {
|
||
|
capacity = 1;
|
||
|
} else {
|
||
|
capacity <<= 1;
|
||
|
}
|
||
|
data = (T *)memrealloc(data, capacity * sizeof(T));
|
||
|
CRASH_COND_MSG(!data, "Out of memory");
|
||
|
}
|
||
|
|
||
|
if (!HAS_TRIVIAL_CONSTRUCTOR(T) && !force_trivial) {
|
||
|
memnew_placement(&data[count++], T(p_elem));
|
||
|
} else {
|
||
|
data[count++] = p_elem;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
void remove(U p_index) {
|
||
|
ERR_FAIL_UNSIGNED_INDEX(p_index, count);
|
||
|
count--;
|
||
|
for (U i = p_index; i < count; i++) {
|
||
|
data[i] = data[i + 1];
|
||
|
}
|
||
|
if (!HAS_TRIVIAL_DESTRUCTOR(T) && !force_trivial) {
|
||
|
data[count].~T();
|
||
|
}
|
||
|
}
|
||
|
|
||
|
// Removes the item copying the last value into the position of the one to
|
||
|
// remove. It's generally faster than `remove`.
|
||
|
void remove_unordered(U p_index) {
|
||
|
ERR_FAIL_INDEX(p_index, count);
|
||
|
count--;
|
||
|
if (count > p_index) {
|
||
|
data[p_index] = data[count];
|
||
|
}
|
||
|
if (!HAS_TRIVIAL_DESTRUCTOR(T) && !force_trivial) {
|
||
|
data[count].~T();
|
||
|
}
|
||
|
}
|
||
|
|
||
|
_FORCE_INLINE_ bool erase(const T &p_val) {
|
||
|
int64_t idx = find(p_val);
|
||
|
if (idx >= 0) {
|
||
|
remove(idx);
|
||
|
return true;
|
||
|
}
|
||
|
return false;
|
||
|
}
|
||
|
|
||
|
U erase_multiple_unordered(const T &p_val) {
|
||
|
U from = 0;
|
||
|
U count = 0;
|
||
|
while (true) {
|
||
|
int64_t idx = find(p_val, from);
|
||
|
|
||
|
if (idx == -1) {
|
||
|
break;
|
||
|
}
|
||
|
remove_unordered(idx);
|
||
|
from = idx;
|
||
|
count++;
|
||
|
}
|
||
|
return count;
|
||
|
}
|
||
|
|
||
|
void invert() {
|
||
|
for (U i = 0; i < count / 2; i++) {
|
||
|
SWAP(data[i], data[count - i - 1]);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
_FORCE_INLINE_ void clear() { resize(0); }
|
||
|
_FORCE_INLINE_ void reset() {
|
||
|
clear();
|
||
|
if (data) {
|
||
|
memfree(data);
|
||
|
data = nullptr;
|
||
|
capacity = 0;
|
||
|
}
|
||
|
}
|
||
|
_FORCE_INLINE_ bool empty() const { return count == 0; }
|
||
|
_FORCE_INLINE_ U get_capacity() const { return capacity; }
|
||
|
_FORCE_INLINE_ void reserve(U p_size, bool p_allow_shrink = false) {
|
||
|
p_size = nearest_power_of_2_templated(p_size);
|
||
|
if (!p_allow_shrink ? p_size > capacity : ((p_size >= count) && (p_size != capacity))) {
|
||
|
capacity = p_size;
|
||
|
data = (T *)memrealloc(data, capacity * sizeof(T));
|
||
|
CRASH_COND_MSG(!data, "Out of memory");
|
||
|
}
|
||
|
}
|
||
|
|
||
|
_FORCE_INLINE_ U size() const { return count; }
|
||
|
void resize(U p_size) {
|
||
|
if (p_size < count) {
|
||
|
if (!HAS_TRIVIAL_DESTRUCTOR(T) && !force_trivial) {
|
||
|
for (U i = p_size; i < count; i++) {
|
||
|
data[i].~T();
|
||
|
}
|
||
|
}
|
||
|
count = p_size;
|
||
|
} else if (p_size > count) {
|
||
|
if (unlikely(p_size > capacity)) {
|
||
|
if (capacity == 0) {
|
||
|
capacity = 1;
|
||
|
}
|
||
|
while (capacity < p_size) {
|
||
|
capacity <<= 1;
|
||
|
}
|
||
|
data = (T *)memrealloc(data, capacity * sizeof(T));
|
||
|
CRASH_COND_MSG(!data, "Out of memory");
|
||
|
}
|
||
|
if (!HAS_TRIVIAL_CONSTRUCTOR(T) && !force_trivial) {
|
||
|
for (U i = count; i < p_size; i++) {
|
||
|
memnew_placement(&data[i], T);
|
||
|
}
|
||
|
}
|
||
|
count = p_size;
|
||
|
}
|
||
|
}
|
||
|
_FORCE_INLINE_ const T &operator[](U p_index) const {
|
||
|
CRASH_BAD_UNSIGNED_INDEX(p_index, count);
|
||
|
return data[p_index];
|
||
|
}
|
||
|
_FORCE_INLINE_ T &operator[](U p_index) {
|
||
|
CRASH_BAD_UNSIGNED_INDEX(p_index, count);
|
||
|
return data[p_index];
|
||
|
}
|
||
|
|
||
|
void fill(T p_val) {
|
||
|
for (U i = 0; i < count; i++) {
|
||
|
data[i] = p_val;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
void insert(U p_pos, T p_val) {
|
||
|
ERR_FAIL_UNSIGNED_INDEX(p_pos, count + 1);
|
||
|
if (p_pos == count) {
|
||
|
push_back(p_val);
|
||
|
} else {
|
||
|
resize(count + 1);
|
||
|
for (U i = count - 1; i > p_pos; i--) {
|
||
|
data[i] = data[i - 1];
|
||
|
}
|
||
|
data[p_pos] = p_val;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
int64_t find(const T &p_val, U p_from = 0) const {
|
||
|
for (U i = p_from; i < count; i++) {
|
||
|
if (data[i] == p_val) {
|
||
|
return int64_t(i);
|
||
|
}
|
||
|
}
|
||
|
return -1;
|
||
|
}
|
||
|
|
||
|
template <class C>
|
||
|
void sort_custom() {
|
||
|
U len = count;
|
||
|
if (len == 0) {
|
||
|
return;
|
||
|
}
|
||
|
|
||
|
SortArray<T, C> sorter;
|
||
|
sorter.sort(data, len);
|
||
|
}
|
||
|
|
||
|
void sort() {
|
||
|
sort_custom<_DefaultComparator<T>>();
|
||
|
}
|
||
|
|
||
|
void ordered_insert(T p_val) {
|
||
|
U i;
|
||
|
for (i = 0; i < count; i++) {
|
||
|
if (p_val < data[i]) {
|
||
|
break;
|
||
|
}
|
||
|
}
|
||
|
insert(i, p_val);
|
||
|
}
|
||
|
|
||
|
operator Vector<T>() const {
|
||
|
Vector<T> ret;
|
||
|
ret.resize(size());
|
||
|
T *w = ret.ptrw();
|
||
|
memcpy(w, data, sizeof(T) * count);
|
||
|
return ret;
|
||
|
}
|
||
|
|
||
|
operator PoolVector<T>() const {
|
||
|
PoolVector<T> pl;
|
||
|
if (size()) {
|
||
|
pl.resize(size());
|
||
|
typename PoolVector<T>::Write w = pl.write();
|
||
|
T *dest = w.ptr();
|
||
|
memcpy(dest, data, sizeof(T) * count);
|
||
|
}
|
||
|
return pl;
|
||
|
}
|
||
|
|
||
|
Vector<uint8_t> to_byte_array() const { //useful to pass stuff to gpu or variant
|
||
|
Vector<uint8_t> ret;
|
||
|
ret.resize(count * sizeof(T));
|
||
|
uint8_t *w = ret.ptrw();
|
||
|
memcpy(w, data, sizeof(T) * count);
|
||
|
return ret;
|
||
|
}
|
||
|
|
||
|
_FORCE_INLINE_ LocalVector() {}
|
||
|
_FORCE_INLINE_ LocalVector(const LocalVector &p_from) {
|
||
|
resize(p_from.size());
|
||
|
for (U i = 0; i < p_from.count; i++) {
|
||
|
data[i] = p_from.data[i];
|
||
|
}
|
||
|
}
|
||
|
LocalVector(const Vector<T> &p_from) {
|
||
|
resize(p_from.size());
|
||
|
for (U i = 0; i < count; i++) {
|
||
|
data[i] = p_from[i];
|
||
|
}
|
||
|
}
|
||
|
LocalVector(const PoolVector<T> &p_from) {
|
||
|
resize(p_from.size());
|
||
|
typename PoolVector<T>::Read r = p_from.read();
|
||
|
for (U i = 0; i < count; i++) {
|
||
|
data[i] = r[i];
|
||
|
}
|
||
|
}
|
||
|
|
||
|
inline LocalVector &operator=(const LocalVector &p_from) {
|
||
|
resize(p_from.size());
|
||
|
for (U i = 0; i < p_from.count; i++) {
|
||
|
data[i] = p_from.data[i];
|
||
|
}
|
||
|
return *this;
|
||
|
}
|
||
|
inline LocalVector &operator=(const Vector<T> &p_from) {
|
||
|
resize(p_from.size());
|
||
|
for (U i = 0; i < count; i++) {
|
||
|
data[i] = p_from[i];
|
||
|
}
|
||
|
return *this;
|
||
|
}
|
||
|
inline LocalVector &operator=(const PoolVector<T> &p_from) {
|
||
|
resize(p_from.size());
|
||
|
typename PoolVector<T>::Read r = p_from.read();
|
||
|
for (U i = 0; i < count; i++) {
|
||
|
data[i] = r[i];
|
||
|
}
|
||
|
return *this;
|
||
|
}
|
||
|
|
||
|
_FORCE_INLINE_ ~LocalVector() {
|
||
|
if (data) {
|
||
|
reset();
|
||
|
}
|
||
|
}
|
||
|
};
|
||
|
|
||
|
// Integer default version
|
||
|
template <class T, class I = int32_t, bool force_trivial = false>
|
||
|
class LocalVectori : public LocalVector<T, I, force_trivial> {
|
||
|
};
|
||
|
|
||
|
#endif // LOCAL_VECTOR_H
|