mirror of
https://github.com/Relintai/pandemonium_engine.git
synced 2024-11-24 01:47:20 +01:00
278 lines
8.2 KiB
C++
278 lines
8.2 KiB
C++
/*************************************************************************/
|
|
/* compressed_translation.cpp */
|
|
/*************************************************************************/
|
|
/* This file is part of: */
|
|
/* GODOT ENGINE */
|
|
/* https://godotengine.org */
|
|
/*************************************************************************/
|
|
/* Copyright (c) 2007-2022 Juan Linietsky, Ariel Manzur. */
|
|
/* Copyright (c) 2014-2022 Godot Engine contributors (cf. AUTHORS.md). */
|
|
/* */
|
|
/* Permission is hereby granted, free of charge, to any person obtaining */
|
|
/* a copy of this software and associated documentation files (the */
|
|
/* "Software"), to deal in the Software without restriction, including */
|
|
/* without limitation the rights to use, copy, modify, merge, publish, */
|
|
/* distribute, sublicense, and/or sell copies of the Software, and to */
|
|
/* permit persons to whom the Software is furnished to do so, subject to */
|
|
/* the following conditions: */
|
|
/* */
|
|
/* The above copyright notice and this permission notice shall be */
|
|
/* included in all copies or substantial portions of the Software. */
|
|
/* */
|
|
/* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, */
|
|
/* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF */
|
|
/* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.*/
|
|
/* IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY */
|
|
/* CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, */
|
|
/* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE */
|
|
/* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */
|
|
/*************************************************************************/
|
|
|
|
#include "compressed_translation.h"
|
|
|
|
#include "core/pair.h"
|
|
|
|
extern "C" {
|
|
#include "thirdparty/misc/smaz.h"
|
|
}
|
|
|
|
struct _PHashTranslationCmp {
|
|
int orig_len;
|
|
CharString compressed;
|
|
int offset;
|
|
};
|
|
|
|
void PHashTranslation::generate(const Ref<Translation> &p_from) {
|
|
#ifdef TOOLS_ENABLED
|
|
ERR_FAIL_COND(p_from.is_null());
|
|
List<StringName> keys;
|
|
p_from->get_message_list(&keys);
|
|
|
|
int size = Math::larger_prime(keys.size());
|
|
|
|
Vector<Vector<Pair<int, CharString>>> buckets;
|
|
Vector<Map<uint32_t, int>> table;
|
|
Vector<uint32_t> hfunc_table;
|
|
Vector<_PHashTranslationCmp> compressed;
|
|
|
|
table.resize(size);
|
|
hfunc_table.resize(size);
|
|
buckets.resize(size);
|
|
compressed.resize(keys.size());
|
|
|
|
int idx = 0;
|
|
int total_compression_size = 0;
|
|
|
|
for (List<StringName>::Element *E = keys.front(); E; E = E->next()) {
|
|
//hash string
|
|
CharString cs = E->get().operator String().utf8();
|
|
uint32_t h = hash(0, cs.get_data());
|
|
Pair<int, CharString> p;
|
|
p.first = idx;
|
|
p.second = cs;
|
|
buckets.write[h % size].push_back(p);
|
|
|
|
//compress string
|
|
CharString src_s = p_from->get_message(E->get()).operator String().utf8();
|
|
_PHashTranslationCmp ps;
|
|
ps.orig_len = src_s.size();
|
|
ps.offset = total_compression_size;
|
|
|
|
if (ps.orig_len != 0) {
|
|
CharString dst_s;
|
|
dst_s.resize(src_s.size());
|
|
int ret = smaz_compress(src_s.get_data(), src_s.size(), dst_s.ptrw(), src_s.size());
|
|
if (ret >= src_s.size()) {
|
|
//if compressed is larger than original, just use original
|
|
ps.orig_len = src_s.size();
|
|
ps.compressed = src_s;
|
|
} else {
|
|
dst_s.resize(ret);
|
|
//ps.orig_len=;
|
|
ps.compressed = dst_s;
|
|
}
|
|
} else {
|
|
ps.orig_len = 1;
|
|
ps.compressed.resize(1);
|
|
ps.compressed[0] = 0;
|
|
}
|
|
|
|
compressed.write[idx] = ps;
|
|
total_compression_size += ps.compressed.size();
|
|
idx++;
|
|
}
|
|
|
|
int bucket_table_size = 0;
|
|
|
|
for (int i = 0; i < size; i++) {
|
|
const Vector<Pair<int, CharString>> &b = buckets[i];
|
|
Map<uint32_t, int> &t = table.write[i];
|
|
|
|
if (b.size() == 0) {
|
|
continue;
|
|
}
|
|
|
|
int d = 1;
|
|
int item = 0;
|
|
|
|
while (item < b.size()) {
|
|
uint32_t slot = hash(d, b[item].second.get_data());
|
|
if (t.has(slot)) {
|
|
item = 0;
|
|
d++;
|
|
t.clear();
|
|
} else {
|
|
t[slot] = b[item].first;
|
|
item++;
|
|
}
|
|
}
|
|
|
|
hfunc_table.write[i] = d;
|
|
bucket_table_size += 2 + b.size() * 4;
|
|
}
|
|
|
|
ERR_FAIL_COND(bucket_table_size == 0);
|
|
|
|
hash_table.resize(size);
|
|
bucket_table.resize(bucket_table_size);
|
|
|
|
PoolVector<int>::Write htwb = hash_table.write();
|
|
PoolVector<int>::Write btwb = bucket_table.write();
|
|
|
|
uint32_t *htw = (uint32_t *)&htwb[0];
|
|
uint32_t *btw = (uint32_t *)&btwb[0];
|
|
|
|
int btindex = 0;
|
|
|
|
for (int i = 0; i < size; i++) {
|
|
const Map<uint32_t, int> &t = table[i];
|
|
if (t.size() == 0) {
|
|
htw[i] = 0xFFFFFFFF; //nothing
|
|
continue;
|
|
}
|
|
|
|
htw[i] = btindex;
|
|
btw[btindex++] = t.size();
|
|
btw[btindex++] = hfunc_table[i];
|
|
|
|
for (Map<uint32_t, int>::Element *E = t.front(); E; E = E->next()) {
|
|
btw[btindex++] = E->key();
|
|
btw[btindex++] = compressed[E->get()].offset;
|
|
btw[btindex++] = compressed[E->get()].compressed.size();
|
|
btw[btindex++] = compressed[E->get()].orig_len;
|
|
}
|
|
}
|
|
|
|
strings.resize(total_compression_size);
|
|
PoolVector<uint8_t>::Write cw = strings.write();
|
|
|
|
for (int i = 0; i < compressed.size(); i++) {
|
|
memcpy(&cw[compressed[i].offset], compressed[i].compressed.get_data(), compressed[i].compressed.size());
|
|
}
|
|
|
|
ERR_FAIL_COND(btindex != bucket_table_size);
|
|
set_locale(p_from->get_locale());
|
|
|
|
#endif
|
|
}
|
|
|
|
bool PHashTranslation::_set(const StringName &p_name, const Variant &p_value) {
|
|
String name = p_name.operator String();
|
|
if (name == "hash_table") {
|
|
hash_table = p_value;
|
|
} else if (name == "bucket_table") {
|
|
bucket_table = p_value;
|
|
} else if (name == "strings") {
|
|
strings = p_value;
|
|
} else if (name == "load_from") {
|
|
generate(p_value);
|
|
} else {
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool PHashTranslation::_get(const StringName &p_name, Variant &r_ret) const {
|
|
String name = p_name.operator String();
|
|
if (name == "hash_table") {
|
|
r_ret = hash_table;
|
|
} else if (name == "bucket_table") {
|
|
r_ret = bucket_table;
|
|
} else if (name == "strings") {
|
|
r_ret = strings;
|
|
} else {
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
StringName PHashTranslation::get_message(const StringName &p_src_text) const {
|
|
int htsize = hash_table.size();
|
|
|
|
if (htsize == 0) {
|
|
return StringName();
|
|
}
|
|
|
|
CharString str = p_src_text.operator String().utf8();
|
|
uint32_t h = hash(0, str.get_data());
|
|
|
|
PoolVector<int>::Read htr = hash_table.read();
|
|
const uint32_t *htptr = (const uint32_t *)&htr[0];
|
|
PoolVector<int>::Read btr = bucket_table.read();
|
|
const uint32_t *btptr = (const uint32_t *)&btr[0];
|
|
PoolVector<uint8_t>::Read sr = strings.read();
|
|
const char *sptr = (const char *)&sr[0];
|
|
|
|
uint32_t p = htptr[h % htsize];
|
|
|
|
if (p == 0xFFFFFFFF) {
|
|
return StringName(); //nothing
|
|
}
|
|
|
|
const Bucket &bucket = *(const Bucket *)&btptr[p];
|
|
|
|
h = hash(bucket.func, str.get_data());
|
|
|
|
int idx = -1;
|
|
|
|
for (int i = 0; i < bucket.size; i++) {
|
|
if (bucket.elem[i].key == h) {
|
|
idx = i;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (idx == -1) {
|
|
return StringName();
|
|
}
|
|
|
|
if (bucket.elem[idx].comp_size == bucket.elem[idx].uncomp_size) {
|
|
String rstr;
|
|
rstr.parse_utf8(&sptr[bucket.elem[idx].str_offset], bucket.elem[idx].uncomp_size);
|
|
|
|
return rstr;
|
|
} else {
|
|
CharString uncomp;
|
|
uncomp.resize(bucket.elem[idx].uncomp_size + 1);
|
|
smaz_decompress(&sptr[bucket.elem[idx].str_offset], bucket.elem[idx].comp_size, uncomp.ptrw(), bucket.elem[idx].uncomp_size);
|
|
String rstr;
|
|
rstr.parse_utf8(uncomp.get_data());
|
|
return rstr;
|
|
}
|
|
}
|
|
|
|
void PHashTranslation::_get_property_list(List<PropertyInfo> *p_list) const {
|
|
p_list->push_back(PropertyInfo(Variant::POOL_INT_ARRAY, "hash_table"));
|
|
p_list->push_back(PropertyInfo(Variant::POOL_INT_ARRAY, "bucket_table"));
|
|
p_list->push_back(PropertyInfo(Variant::POOL_BYTE_ARRAY, "strings"));
|
|
p_list->push_back(PropertyInfo(Variant::OBJECT, "load_from", PROPERTY_HINT_RESOURCE_TYPE, "Translation", PROPERTY_USAGE_EDITOR));
|
|
}
|
|
void PHashTranslation::_bind_methods() {
|
|
ClassDB::bind_method(D_METHOD("generate", "from"), &PHashTranslation::generate);
|
|
}
|
|
|
|
PHashTranslation::PHashTranslation() {
|
|
}
|