mirror of
https://github.com/Relintai/pandemonium_engine.git
synced 2024-12-25 13:17:22 +01:00
254 lines
8.8 KiB
C++
254 lines
8.8 KiB
C++
|
/*************************************************************************/
|
||
|
/* FBXTokenizer.cpp */
|
||
|
/*************************************************************************/
|
||
|
/* This file is part of: */
|
||
|
/* GODOT ENGINE */
|
||
|
/* https://godotengine.org */
|
||
|
/*************************************************************************/
|
||
|
/* Copyright (c) 2007-2022 Juan Linietsky, Ariel Manzur. */
|
||
|
/* Copyright (c) 2014-2022 Godot Engine contributors (cf. AUTHORS.md). */
|
||
|
/* */
|
||
|
/* Permission is hereby granted, free of charge, to any person obtaining */
|
||
|
/* a copy of this software and associated documentation files (the */
|
||
|
/* "Software"), to deal in the Software without restriction, including */
|
||
|
/* without limitation the rights to use, copy, modify, merge, publish, */
|
||
|
/* distribute, sublicense, and/or sell copies of the Software, and to */
|
||
|
/* permit persons to whom the Software is furnished to do so, subject to */
|
||
|
/* the following conditions: */
|
||
|
/* */
|
||
|
/* The above copyright notice and this permission notice shall be */
|
||
|
/* included in all copies or substantial portions of the Software. */
|
||
|
/* */
|
||
|
/* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, */
|
||
|
/* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF */
|
||
|
/* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.*/
|
||
|
/* IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY */
|
||
|
/* CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, */
|
||
|
/* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE */
|
||
|
/* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */
|
||
|
/*************************************************************************/
|
||
|
|
||
|
/*
|
||
|
Open Asset Import Library (assimp)
|
||
|
----------------------------------------------------------------------
|
||
|
|
||
|
Copyright (c) 2006-2019, assimp team
|
||
|
|
||
|
|
||
|
All rights reserved.
|
||
|
|
||
|
Redistribution and use of this software in source and binary forms,
|
||
|
with or without modification, are permitted provided that the
|
||
|
following conditions are met:
|
||
|
|
||
|
* Redistributions of source code must retain the above
|
||
|
copyright notice, this list of conditions and the
|
||
|
following disclaimer.
|
||
|
|
||
|
* Redistributions in binary form must reproduce the above
|
||
|
copyright notice, this list of conditions and the
|
||
|
following disclaimer in the documentation and/or other
|
||
|
materials provided with the distribution.
|
||
|
|
||
|
* Neither the name of the assimp team, nor the names of its
|
||
|
contributors may be used to endorse or promote products
|
||
|
derived from this software without specific prior
|
||
|
written permission of the assimp team.
|
||
|
|
||
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||
|
"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||
|
LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
||
|
A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
||
|
OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
||
|
SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||
|
LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||
|
DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||
|
THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||
|
(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||
|
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||
|
|
||
|
----------------------------------------------------------------------
|
||
|
*/
|
||
|
|
||
|
/** @file FBXTokenizer.cpp
|
||
|
* @brief Implementation of the FBX broadphase lexer
|
||
|
*/
|
||
|
|
||
|
// tab width for logging columns
|
||
|
#define ASSIMP_FBX_TAB_WIDTH 4
|
||
|
|
||
|
#include "FBXTokenizer.h"
|
||
|
#include "core/print_string.h"
|
||
|
|
||
|
namespace FBXDocParser {
|
||
|
|
||
|
// ------------------------------------------------------------------------------------------------
|
||
|
Token::Token(const char *p_sbegin, const char *p_send, TokenType p_type, unsigned int p_line, unsigned int p_column) :
|
||
|
sbegin(p_sbegin),
|
||
|
send(p_send),
|
||
|
type(p_type),
|
||
|
line(p_line),
|
||
|
column(p_column) {
|
||
|
#ifdef DEBUG_ENABLED
|
||
|
contents = std::string(sbegin, static_cast<size_t>(send - sbegin));
|
||
|
#endif
|
||
|
}
|
||
|
|
||
|
// ------------------------------------------------------------------------------------------------
|
||
|
Token::~Token() {
|
||
|
}
|
||
|
|
||
|
namespace {
|
||
|
|
||
|
// ------------------------------------------------------------------------------------------------
|
||
|
void TokenizeError(const std::string &message, unsigned int line, unsigned int column) {
|
||
|
print_error("[FBX-Tokenize]" + String(message.c_str()) + " " + itos(line) + ":" + itos(column));
|
||
|
}
|
||
|
|
||
|
// process a potential data token up to 'cur', adding it to 'output_tokens'.
|
||
|
// ------------------------------------------------------------------------------------------------
|
||
|
void ProcessDataToken(TokenList &output_tokens, const char *&start, const char *&end,
|
||
|
unsigned int line,
|
||
|
unsigned int column,
|
||
|
TokenType type = TokenType_DATA,
|
||
|
bool must_have_token = false) {
|
||
|
if (start && end) {
|
||
|
// sanity check:
|
||
|
// tokens should have no whitespace outside quoted text and [start,end] should
|
||
|
// properly delimit the valid range.
|
||
|
bool in_double_quotes = false;
|
||
|
for (const char *c = start; c != end + 1; ++c) {
|
||
|
if (*c == '\"') {
|
||
|
in_double_quotes = !in_double_quotes;
|
||
|
}
|
||
|
|
||
|
if (!in_double_quotes && IsSpaceOrNewLine(*c)) {
|
||
|
TokenizeError("unexpected whitespace in token", line, column);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
if (in_double_quotes) {
|
||
|
TokenizeError("non-terminated double quotes", line, column);
|
||
|
}
|
||
|
|
||
|
output_tokens.push_back(new_Token(start, end + 1, type, line, column));
|
||
|
} else if (must_have_token) {
|
||
|
TokenizeError("unexpected character, expected data token", line, column);
|
||
|
}
|
||
|
|
||
|
start = end = nullptr;
|
||
|
}
|
||
|
|
||
|
} // namespace
|
||
|
|
||
|
// ------------------------------------------------------------------------------------------------
|
||
|
void Tokenize(TokenList &output_tokens, const char *input, size_t length) {
|
||
|
// line and column numbers numbers are one-based
|
||
|
unsigned int line = 1;
|
||
|
unsigned int column = 1;
|
||
|
|
||
|
bool comment = false;
|
||
|
bool in_double_quotes = false;
|
||
|
bool pending_data_token = false;
|
||
|
|
||
|
const char *token_begin = nullptr, *token_end = nullptr;
|
||
|
|
||
|
// input (starting string), *cur the current string, column +=
|
||
|
// modified to fix strlen() and stop buffer overflow
|
||
|
for (size_t x = 0; x < length; x++) {
|
||
|
const char c = input[x];
|
||
|
const char *cur = &input[x];
|
||
|
column += (c == '\t' ? ASSIMP_FBX_TAB_WIDTH : 1);
|
||
|
|
||
|
if (IsLineEnd(c)) {
|
||
|
comment = false;
|
||
|
|
||
|
column = 0;
|
||
|
++line;
|
||
|
}
|
||
|
|
||
|
if (comment) {
|
||
|
continue;
|
||
|
}
|
||
|
|
||
|
if (in_double_quotes) {
|
||
|
if (c == '\"') {
|
||
|
in_double_quotes = false;
|
||
|
token_end = cur;
|
||
|
|
||
|
ProcessDataToken(output_tokens, token_begin, token_end, line, column);
|
||
|
pending_data_token = false;
|
||
|
}
|
||
|
continue;
|
||
|
}
|
||
|
|
||
|
switch (c) {
|
||
|
case '\"':
|
||
|
if (token_begin) {
|
||
|
TokenizeError("unexpected double-quote", line, column);
|
||
|
}
|
||
|
token_begin = cur;
|
||
|
in_double_quotes = true;
|
||
|
continue;
|
||
|
|
||
|
case ';':
|
||
|
ProcessDataToken(output_tokens, token_begin, token_end, line, column);
|
||
|
comment = true;
|
||
|
continue;
|
||
|
|
||
|
case '{':
|
||
|
ProcessDataToken(output_tokens, token_begin, token_end, line, column);
|
||
|
output_tokens.push_back(new_Token(cur, cur + 1, TokenType_OPEN_BRACKET, line, column));
|
||
|
continue;
|
||
|
|
||
|
case '}':
|
||
|
ProcessDataToken(output_tokens, token_begin, token_end, line, column);
|
||
|
output_tokens.push_back(new_Token(cur, cur + 1, TokenType_CLOSE_BRACKET, line, column));
|
||
|
continue;
|
||
|
|
||
|
case ',':
|
||
|
if (pending_data_token) {
|
||
|
ProcessDataToken(output_tokens, token_begin, token_end, line, column, TokenType_DATA, true);
|
||
|
}
|
||
|
output_tokens.push_back(new_Token(cur, cur + 1, TokenType_COMMA, line, column));
|
||
|
continue;
|
||
|
|
||
|
case ':':
|
||
|
if (pending_data_token) {
|
||
|
ProcessDataToken(output_tokens, token_begin, token_end, line, column, TokenType_KEY, true);
|
||
|
} else {
|
||
|
TokenizeError("unexpected colon", line, column);
|
||
|
}
|
||
|
continue;
|
||
|
}
|
||
|
|
||
|
if (IsSpaceOrNewLine(c)) {
|
||
|
if (token_begin) {
|
||
|
// peek ahead and check if the next token is a colon in which
|
||
|
// case this counts as KEY token.
|
||
|
TokenType type = TokenType_DATA;
|
||
|
for (const char *peek = cur; *peek && IsSpaceOrNewLine(*peek); ++peek) {
|
||
|
if (*peek == ':') {
|
||
|
type = TokenType_KEY;
|
||
|
cur = peek;
|
||
|
break;
|
||
|
}
|
||
|
}
|
||
|
|
||
|
ProcessDataToken(output_tokens, token_begin, token_end, line, column, type);
|
||
|
}
|
||
|
|
||
|
pending_data_token = false;
|
||
|
} else {
|
||
|
token_end = cur;
|
||
|
if (!token_begin) {
|
||
|
token_begin = cur;
|
||
|
}
|
||
|
|
||
|
pending_data_token = true;
|
||
|
}
|
||
|
}
|
||
|
}
|
||
|
|
||
|
} // namespace FBXDocParser
|