mirror of
https://github.com/RGBCube/serenity
synced 2025-07-26 04:27:44 +00:00
LibSQL: Introduce a SQL library with a tokenizer
LibSQL aims to be a SQLite clone for SerenityOS. Step 1 is creating a tokenizer to lex SQL tokens. This lexer is heavily influenced by the LibJS lexer.
This commit is contained in:
parent
55b7b8e930
commit
90517da9ca
6 changed files with 716 additions and 0 deletions
|
@ -27,6 +27,7 @@ add_subdirectory(LibPCIDB)
|
||||||
add_subdirectory(LibProtocol)
|
add_subdirectory(LibProtocol)
|
||||||
add_subdirectory(LibPthread)
|
add_subdirectory(LibPthread)
|
||||||
add_subdirectory(LibRegex)
|
add_subdirectory(LibRegex)
|
||||||
|
add_subdirectory(LibSQL)
|
||||||
add_subdirectory(LibSymbolClient)
|
add_subdirectory(LibSymbolClient)
|
||||||
add_subdirectory(LibSyntax)
|
add_subdirectory(LibSyntax)
|
||||||
add_subdirectory(LibSystem)
|
add_subdirectory(LibSystem)
|
||||||
|
|
7
Userland/Libraries/LibSQL/CMakeLists.txt
Normal file
7
Userland/Libraries/LibSQL/CMakeLists.txt
Normal file
|
@ -0,0 +1,7 @@
|
||||||
|
set(SOURCES
|
||||||
|
Lexer.cpp
|
||||||
|
Token.cpp
|
||||||
|
)
|
||||||
|
|
||||||
|
serenity_lib(LibSQL sql)
|
||||||
|
target_link_libraries(LibSQL LibCore)
|
306
Userland/Libraries/LibSQL/Lexer.cpp
Normal file
306
Userland/Libraries/LibSQL/Lexer.cpp
Normal file
|
@ -0,0 +1,306 @@
|
||||||
|
/*
|
||||||
|
* Copyright (c) 2021, Tim Flynn <trflynn89@pm.me>
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* Redistribution and use in source and binary forms, with or without
|
||||||
|
* modification, are permitted provided that the following conditions are met:
|
||||||
|
*
|
||||||
|
* 1. Redistributions of source code must retain the above copyright notice, this
|
||||||
|
* list of conditions and the following disclaimer.
|
||||||
|
*
|
||||||
|
* 2. Redistributions in binary form must reproduce the above copyright notice,
|
||||||
|
* this list of conditions and the following disclaimer in the documentation
|
||||||
|
* and/or other materials provided with the distribution.
|
||||||
|
*
|
||||||
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
||||||
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||||
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
||||||
|
* DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
|
||||||
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
||||||
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
||||||
|
* SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
||||||
|
* CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
||||||
|
* OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include "Lexer.h"
|
||||||
|
#include <AK/Debug.h>
|
||||||
|
#include <ctype.h>
|
||||||
|
|
||||||
|
namespace SQL {
|
||||||
|
|
||||||
|
HashMap<String, TokenType> Lexer::s_keywords;
|
||||||
|
HashMap<char, TokenType> Lexer::s_one_char_tokens;
|
||||||
|
HashMap<String, TokenType> Lexer::s_two_char_tokens;
|
||||||
|
|
||||||
|
Lexer::Lexer(StringView source)
|
||||||
|
: m_source(source)
|
||||||
|
{
|
||||||
|
if (s_keywords.is_empty()) {
|
||||||
|
#define __ENUMERATE_SQL_TOKEN(value, type, category) \
|
||||||
|
if (TokenCategory::category == TokenCategory::Keyword) \
|
||||||
|
s_keywords.set(value, TokenType::type);
|
||||||
|
ENUMERATE_SQL_TOKENS
|
||||||
|
#undef __ENUMERATE_SQL_TOKEN
|
||||||
|
}
|
||||||
|
|
||||||
|
if (s_one_char_tokens.is_empty()) {
|
||||||
|
#define __ENUMERATE_SQL_TOKEN(value, type, category) \
|
||||||
|
if (TokenCategory::category != TokenCategory::Keyword && StringView(value).length() == 1) \
|
||||||
|
s_one_char_tokens.set(value[0], TokenType::type);
|
||||||
|
ENUMERATE_SQL_TOKENS
|
||||||
|
#undef __ENUMERATE_SQL_TOKEN
|
||||||
|
}
|
||||||
|
|
||||||
|
if (s_two_char_tokens.is_empty()) {
|
||||||
|
#define __ENUMERATE_SQL_TOKEN(value, type, category) \
|
||||||
|
if (TokenCategory::category != TokenCategory::Keyword && StringView(value).length() == 2) \
|
||||||
|
s_two_char_tokens.set(value, TokenType::type);
|
||||||
|
ENUMERATE_SQL_TOKENS
|
||||||
|
#undef __ENUMERATE_SQL_TOKEN
|
||||||
|
}
|
||||||
|
|
||||||
|
consume();
|
||||||
|
}
|
||||||
|
|
||||||
|
Token Lexer::next()
|
||||||
|
{
|
||||||
|
bool found_invalid_comment = consume_whitespace_and_comments();
|
||||||
|
|
||||||
|
size_t value_start = m_position;
|
||||||
|
size_t value_start_line_number = m_line_number;
|
||||||
|
size_t value_start_column_number = m_line_column;
|
||||||
|
auto token_type = TokenType::Invalid;
|
||||||
|
|
||||||
|
if (is_eof()) {
|
||||||
|
token_type = found_invalid_comment ? TokenType::Invalid : TokenType::Eof;
|
||||||
|
} else if (is_identifier_start()) {
|
||||||
|
do {
|
||||||
|
consume();
|
||||||
|
} while (is_identifier_middle());
|
||||||
|
|
||||||
|
if (auto it = s_keywords.find(m_source.substring_view(value_start - 1, m_position - value_start)); it != s_keywords.end()) {
|
||||||
|
token_type = it->value;
|
||||||
|
} else {
|
||||||
|
token_type = TokenType::Identifier;
|
||||||
|
}
|
||||||
|
} else if (is_numeric_literal_start()) {
|
||||||
|
token_type = TokenType::NumericLiteral;
|
||||||
|
if (!consume_numeric_literal())
|
||||||
|
token_type = TokenType::Invalid;
|
||||||
|
} else {
|
||||||
|
bool found_two_char_token = false;
|
||||||
|
if (m_position < m_source.length()) {
|
||||||
|
if (auto it = s_two_char_tokens.find(m_source.substring_view(m_position - 1, 2)); it != s_two_char_tokens.end()) {
|
||||||
|
found_two_char_token = true;
|
||||||
|
token_type = it->value;
|
||||||
|
consume();
|
||||||
|
consume();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
bool found_one_char_token = false;
|
||||||
|
if (!found_two_char_token) {
|
||||||
|
if (auto it = s_one_char_tokens.find(m_current_char); it != s_one_char_tokens.end()) {
|
||||||
|
found_one_char_token = true;
|
||||||
|
token_type = it->value;
|
||||||
|
consume();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!found_two_char_token && !found_one_char_token) {
|
||||||
|
token_type = TokenType::Invalid;
|
||||||
|
consume();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
Token token(token_type, m_source.substring_view(value_start - 1, m_position - value_start), value_start_line_number, value_start_column_number);
|
||||||
|
|
||||||
|
if constexpr (SQL_DEBUG) {
|
||||||
|
dbgln("------------------------------");
|
||||||
|
dbgln("Token: {}", token.name());
|
||||||
|
dbgln("Value: {}", token.value());
|
||||||
|
dbgln("Line: {}, Column: {}", token.line_number(), token.line_column());
|
||||||
|
dbgln("------------------------------");
|
||||||
|
}
|
||||||
|
|
||||||
|
return token;
|
||||||
|
}
|
||||||
|
|
||||||
|
void Lexer::consume()
|
||||||
|
{
|
||||||
|
auto did_reach_eof = [this] {
|
||||||
|
if (m_position != m_source.length())
|
||||||
|
return false;
|
||||||
|
m_current_char = EOF;
|
||||||
|
++m_line_column;
|
||||||
|
++m_position;
|
||||||
|
return true;
|
||||||
|
};
|
||||||
|
|
||||||
|
if (m_position > m_source.length())
|
||||||
|
return;
|
||||||
|
|
||||||
|
if (did_reach_eof())
|
||||||
|
return;
|
||||||
|
|
||||||
|
if (is_line_break()) {
|
||||||
|
++m_line_number;
|
||||||
|
m_line_column = 1;
|
||||||
|
} else {
|
||||||
|
++m_line_column;
|
||||||
|
}
|
||||||
|
|
||||||
|
m_current_char = m_source[m_position++];
|
||||||
|
}
|
||||||
|
|
||||||
|
bool Lexer::consume_whitespace_and_comments()
|
||||||
|
{
|
||||||
|
bool found_invalid_comment = false;
|
||||||
|
|
||||||
|
while (true) {
|
||||||
|
if (isspace(m_current_char)) {
|
||||||
|
do {
|
||||||
|
consume();
|
||||||
|
} while (isspace(m_current_char));
|
||||||
|
} else if (is_line_comment_start()) {
|
||||||
|
consume();
|
||||||
|
do {
|
||||||
|
consume();
|
||||||
|
} while (!is_eof() && !is_line_break());
|
||||||
|
} else if (is_block_comment_start()) {
|
||||||
|
consume();
|
||||||
|
do {
|
||||||
|
consume();
|
||||||
|
} while (!is_eof() && !is_block_comment_end());
|
||||||
|
if (is_eof())
|
||||||
|
found_invalid_comment = true;
|
||||||
|
consume(); // consume *
|
||||||
|
if (is_eof())
|
||||||
|
found_invalid_comment = true;
|
||||||
|
consume(); // consume /
|
||||||
|
} else {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return found_invalid_comment;
|
||||||
|
}
|
||||||
|
|
||||||
|
bool Lexer::consume_numeric_literal()
|
||||||
|
{
|
||||||
|
// https://www.sqlite.org/draft/syntax/numeric-literal.html
|
||||||
|
bool is_valid_numeric_literal = true;
|
||||||
|
|
||||||
|
if (m_current_char == '0') {
|
||||||
|
consume();
|
||||||
|
if (m_current_char == '.') {
|
||||||
|
consume();
|
||||||
|
while (isdigit(m_current_char))
|
||||||
|
consume();
|
||||||
|
if (m_current_char == 'e' || m_current_char == 'E')
|
||||||
|
is_valid_numeric_literal = consume_exponent();
|
||||||
|
} else if (m_current_char == 'e' || m_current_char == 'E') {
|
||||||
|
is_valid_numeric_literal = consume_exponent();
|
||||||
|
} else if (m_current_char == 'x' || m_current_char == 'X') {
|
||||||
|
is_valid_numeric_literal = consume_hexadecimal_number();
|
||||||
|
} else if (isdigit(m_current_char)) {
|
||||||
|
do {
|
||||||
|
consume();
|
||||||
|
} while (isdigit(m_current_char));
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
do {
|
||||||
|
consume();
|
||||||
|
} while (isdigit(m_current_char));
|
||||||
|
|
||||||
|
if (m_current_char == '.') {
|
||||||
|
consume();
|
||||||
|
while (isdigit(m_current_char))
|
||||||
|
consume();
|
||||||
|
}
|
||||||
|
if (m_current_char == 'e' || m_current_char == 'E')
|
||||||
|
is_valid_numeric_literal = consume_exponent();
|
||||||
|
}
|
||||||
|
|
||||||
|
return is_valid_numeric_literal;
|
||||||
|
}
|
||||||
|
|
||||||
|
bool Lexer::consume_exponent()
|
||||||
|
{
|
||||||
|
consume();
|
||||||
|
if (m_current_char == '-' || m_current_char == '+')
|
||||||
|
consume();
|
||||||
|
|
||||||
|
if (!isdigit(m_current_char))
|
||||||
|
return false;
|
||||||
|
|
||||||
|
while (isdigit(m_current_char)) {
|
||||||
|
consume();
|
||||||
|
}
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
bool Lexer::consume_hexadecimal_number()
|
||||||
|
{
|
||||||
|
consume();
|
||||||
|
if (!isxdigit(m_current_char))
|
||||||
|
return false;
|
||||||
|
|
||||||
|
while (isxdigit(m_current_char))
|
||||||
|
consume();
|
||||||
|
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
bool Lexer::match(char a, char b) const
|
||||||
|
{
|
||||||
|
if (m_position >= m_source.length())
|
||||||
|
return false;
|
||||||
|
|
||||||
|
return m_current_char == a
|
||||||
|
&& m_source[m_position] == b;
|
||||||
|
}
|
||||||
|
|
||||||
|
bool Lexer::is_identifier_start() const
|
||||||
|
{
|
||||||
|
return isalpha(m_current_char) || m_current_char == '_';
|
||||||
|
}
|
||||||
|
|
||||||
|
bool Lexer::is_identifier_middle() const
|
||||||
|
{
|
||||||
|
return is_identifier_start() || isdigit(m_current_char);
|
||||||
|
}
|
||||||
|
|
||||||
|
bool Lexer::is_numeric_literal_start() const
|
||||||
|
{
|
||||||
|
return isdigit(m_current_char) || (m_current_char == '.' && m_position < m_source.length() && isdigit(m_source[m_position]));
|
||||||
|
}
|
||||||
|
|
||||||
|
bool Lexer::is_line_comment_start() const
|
||||||
|
{
|
||||||
|
return match('-', '-');
|
||||||
|
}
|
||||||
|
|
||||||
|
bool Lexer::is_block_comment_start() const
|
||||||
|
{
|
||||||
|
return match('/', '*');
|
||||||
|
}
|
||||||
|
|
||||||
|
bool Lexer::is_block_comment_end() const
|
||||||
|
{
|
||||||
|
return match('*', '/');
|
||||||
|
}
|
||||||
|
|
||||||
|
bool Lexer::is_line_break() const
|
||||||
|
{
|
||||||
|
return m_current_char == '\n';
|
||||||
|
}
|
||||||
|
|
||||||
|
bool Lexer::is_eof() const
|
||||||
|
{
|
||||||
|
return m_current_char == EOF;
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
71
Userland/Libraries/LibSQL/Lexer.h
Normal file
71
Userland/Libraries/LibSQL/Lexer.h
Normal file
|
@ -0,0 +1,71 @@
|
||||||
|
/*
|
||||||
|
* Copyright (c) 2021, Tim Flynn <trflynn89@pm.me>
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* Redistribution and use in source and binary forms, with or without
|
||||||
|
* modification, are permitted provided that the following conditions are met:
|
||||||
|
*
|
||||||
|
* 1. Redistributions of source code must retain the above copyright notice, this
|
||||||
|
* list of conditions and the following disclaimer.
|
||||||
|
*
|
||||||
|
* 2. Redistributions in binary form must reproduce the above copyright notice,
|
||||||
|
* this list of conditions and the following disclaimer in the documentation
|
||||||
|
* and/or other materials provided with the distribution.
|
||||||
|
*
|
||||||
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
||||||
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||||
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
||||||
|
* DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
|
||||||
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
||||||
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
||||||
|
* SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
||||||
|
* CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
||||||
|
* OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include "Token.h"
|
||||||
|
#include <AK/HashMap.h>
|
||||||
|
#include <AK/String.h>
|
||||||
|
#include <AK/StringView.h>
|
||||||
|
|
||||||
|
namespace SQL {
|
||||||
|
|
||||||
|
class Lexer {
|
||||||
|
public:
|
||||||
|
explicit Lexer(StringView source);
|
||||||
|
|
||||||
|
Token next();
|
||||||
|
|
||||||
|
private:
|
||||||
|
void consume();
|
||||||
|
|
||||||
|
bool consume_whitespace_and_comments();
|
||||||
|
bool consume_numeric_literal();
|
||||||
|
bool consume_exponent();
|
||||||
|
bool consume_hexadecimal_number();
|
||||||
|
|
||||||
|
bool match(char a, char b) const;
|
||||||
|
bool is_identifier_start() const;
|
||||||
|
bool is_identifier_middle() const;
|
||||||
|
bool is_numeric_literal_start() const;
|
||||||
|
bool is_line_comment_start() const;
|
||||||
|
bool is_block_comment_start() const;
|
||||||
|
bool is_block_comment_end() const;
|
||||||
|
bool is_line_break() const;
|
||||||
|
bool is_eof() const;
|
||||||
|
|
||||||
|
static HashMap<String, TokenType> s_keywords;
|
||||||
|
static HashMap<char, TokenType> s_one_char_tokens;
|
||||||
|
static HashMap<String, TokenType> s_two_char_tokens;
|
||||||
|
|
||||||
|
StringView m_source;
|
||||||
|
size_t m_line_number { 1 };
|
||||||
|
size_t m_line_column { 0 };
|
||||||
|
char m_current_char { 0 };
|
||||||
|
size_t m_position { 0 };
|
||||||
|
};
|
||||||
|
|
||||||
|
}
|
73
Userland/Libraries/LibSQL/Token.cpp
Normal file
73
Userland/Libraries/LibSQL/Token.cpp
Normal file
|
@ -0,0 +1,73 @@
|
||||||
|
/*
|
||||||
|
* Copyright (c) 2021, Tim Flynn <trflynn89@pm.me>
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* Redistribution and use in source and binary forms, with or without
|
||||||
|
* modification, are permitted provided that the following conditions are met:
|
||||||
|
*
|
||||||
|
* 1. Redistributions of source code must retain the above copyright notice, this
|
||||||
|
* list of conditions and the following disclaimer.
|
||||||
|
*
|
||||||
|
* 2. Redistributions in binary form must reproduce the above copyright notice,
|
||||||
|
* this list of conditions and the following disclaimer in the documentation
|
||||||
|
* and/or other materials provided with the distribution.
|
||||||
|
*
|
||||||
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
||||||
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||||
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
||||||
|
* DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
|
||||||
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
||||||
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
||||||
|
* SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
||||||
|
* CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
||||||
|
* OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#include "Token.h"
|
||||||
|
#include <AK/Assertions.h>
|
||||||
|
#include <AK/String.h>
|
||||||
|
#include <stdlib.h>
|
||||||
|
|
||||||
|
namespace SQL {
|
||||||
|
|
||||||
|
StringView Token::name(TokenType type)
|
||||||
|
{
|
||||||
|
switch (type) {
|
||||||
|
#define __ENUMERATE_SQL_TOKEN(value, type, category) \
|
||||||
|
case TokenType::type: \
|
||||||
|
return #type;
|
||||||
|
ENUMERATE_SQL_TOKENS
|
||||||
|
#undef __ENUMERATE_SQL_TOKEN
|
||||||
|
default:
|
||||||
|
VERIFY_NOT_REACHED();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
TokenCategory Token::category(TokenType type)
|
||||||
|
{
|
||||||
|
switch (type) {
|
||||||
|
#define __ENUMERATE_SQL_TOKEN(value, type, category) \
|
||||||
|
case TokenType::type: \
|
||||||
|
return TokenCategory::category;
|
||||||
|
ENUMERATE_SQL_TOKENS
|
||||||
|
#undef __ENUMERATE_SQL_TOKEN
|
||||||
|
default:
|
||||||
|
VERIFY_NOT_REACHED();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
double Token::double_value() const
|
||||||
|
{
|
||||||
|
VERIFY(type() == TokenType::NumericLiteral);
|
||||||
|
String value(m_value);
|
||||||
|
|
||||||
|
if (value[0] == '0' && value.length() >= 2) {
|
||||||
|
if (value[1] == 'x' || value[1] == 'X')
|
||||||
|
return static_cast<double>(strtoul(value.characters() + 2, nullptr, 16));
|
||||||
|
}
|
||||||
|
|
||||||
|
return strtod(value.characters(), nullptr);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
258
Userland/Libraries/LibSQL/Token.h
Normal file
258
Userland/Libraries/LibSQL/Token.h
Normal file
|
@ -0,0 +1,258 @@
|
||||||
|
/*
|
||||||
|
* Copyright (c) 2021, Tim Flynn <trflynn89@pm.me>
|
||||||
|
* All rights reserved.
|
||||||
|
*
|
||||||
|
* Redistribution and use in source and binary forms, with or without
|
||||||
|
* modification, are permitted provided that the following conditions are met:
|
||||||
|
*
|
||||||
|
* 1. Redistributions of source code must retain the above copyright notice, this
|
||||||
|
* list of conditions and the following disclaimer.
|
||||||
|
*
|
||||||
|
* 2. Redistributions in binary form must reproduce the above copyright notice,
|
||||||
|
* this list of conditions and the following disclaimer in the documentation
|
||||||
|
* and/or other materials provided with the distribution.
|
||||||
|
*
|
||||||
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
||||||
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
||||||
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
||||||
|
* DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
|
||||||
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
||||||
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
||||||
|
* SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
||||||
|
* CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
||||||
|
* OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#pragma once
|
||||||
|
|
||||||
|
#include <AK/HashMap.h>
|
||||||
|
#include <AK/StringView.h>
|
||||||
|
|
||||||
|
namespace SQL {
|
||||||
|
|
||||||
|
// https://sqlite.org/lang_keywords.html
|
||||||
|
#define ENUMERATE_SQL_TOKENS \
|
||||||
|
__ENUMERATE_SQL_TOKEN("ABORT", Abort, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("ACTION", Action, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("ADD", Add, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("AFTER", After, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("ALL", All, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("ALTER", Alter, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("ALWAYS", Always, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("ANALYZE", Analyze, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("AND", And, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("AS", As, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("ASC", Asc, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("ATTACH", Attach, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("AUTOINCREMENT", Autoincrement, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("BEFORE", Before, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("BEGIN", Begin, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("BETWEEN", Between, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("BY", By, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("CASCADE", Cascade, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("CASE", Case, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("CAST", Cast, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("CHECK", Check, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("COLLATE", Collate, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("COLUMN", Column, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("COMMIT", Commit, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("CONFLICT", Conflict, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("CONSTRAINT", Constraint, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("CREATE", Create, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("CROSS", Cross, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("CURRENT", Current, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("CURRENT_DATE", CurrentDate, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("CURRENT_TIME", CurrentTime, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("CURRENT_TIMESTAMP", CurrentTimestamp, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("DATABASE", Database, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("DEFAULT", Default, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("DEFERRABLE", Deferrable, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("DEFERRED", Deferred, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("DELETE", Delete, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("DESC", Desc, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("DETACH", Detach, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("DISTINCT", Distinct, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("DO", Do, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("DROP", Drop, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("EACH", Each, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("ELSE", Else, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("END", End, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("ESCAPE", Escape, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("EXCEPT", Except, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("EXCLUDE", Exclude, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("EXCLUSIVE", Exclusive, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("EXISTS", Exists, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("EXPLAIN", Explain, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("FAIL", Fail, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("FILTER", Filter, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("FIRST", First, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("FOLLOWING", Following, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("FOR", For, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("FOREIGN", Foreign, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("FROM", From, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("FULL", Full, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("GENERATED", Generated, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("GLOB", Glob, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("GROUP", Group, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("GROUPS", Groups, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("HAVING", Having, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("IF", If, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("IGNORE", Ignore, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("IMMEDIATE", Immediate, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("IN", In, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("INDEX", Index, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("INDEXED", Indexed, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("INITIALLY", Initially, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("INNER", Inner, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("INSERT", Insert, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("INSTEAD", Instead, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("INTERSECT", Intersect, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("INTO", Into, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("IS", Is, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("ISNULL", Isnull, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("JOIN", Join, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("KEY", Key, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("LAST", Last, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("LEFT", Left, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("LIKE", Like, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("LIMIT", Limit, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("MATCH", Match, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("MATERIALIZED", Materialized, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("NATURAL", Natural, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("NO", No, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("NOT", Not, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("NOTHING", Nothing, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("NOTNULL", Notnull, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("NULL", Null, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("NULLS", Nulls, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("OF", Of, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("OFFSET", Offset, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("ON", On, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("OR", Or, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("ORDER", Order, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("OTHERS", Others, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("OUTER", Outer, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("OVER", Over, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("PARTITION", Partition, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("PLAN", Plan, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("PRAGMA", Pragma, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("PRECEDING", Preceding, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("PRIMARY", Primary, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("QUERY", Query, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("RAISE", Raise, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("RANGE", Range, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("RECURSIVE", Recursive, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("REFERENCES", References, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("REGEXP", Regexp, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("REINDEX", Reindex, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("RELEASE", Release, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("RENAME", Rename, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("REPLACE", Replace, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("RESTRICT", Restrict, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("RETURNING", Returning, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("RIGHT", Right, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("ROLLBACK", Rollback, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("ROW", Row, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("ROWS", Rows, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("SAVEPOINT", Savepoint, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("SELECT", Select, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("SET", Set, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("TABLE", Table, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("TEMP", Temp, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("TEMPORARY", Temporary, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("THEN", Then, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("TIES", Ties, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("TO", To, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("TRANSACTION", Transaction, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("TRIGGER", Trigger, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("UNBOUNDED", Unbounded, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("UNION", Union, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("UNIQUE", Unique, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("UPDATE", Update, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("USING", Using, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("VACUUM", Vacuum, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("VALUES", Values, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("VIEW", View, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("VIRTUAL", Virtual, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("WHEN", When, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("WHERE", Where, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("WINDOW", Window, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("WITH", With, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("WITHOUT", Without, Keyword) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("_identifier_", Identifier, Identifier) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("_numeric_", NumericLiteral, Number) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("_eof_", Eof, Invalid) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("_invalid_", Invalid, Invalid) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("&", Ampersand, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("*", Asterisk, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN(",", Comma, Punctuation) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("/", Divide, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("||", DoublePipe, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("=", Equals, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("==", EqualsEquals, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN(">", GreaterThan, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN(">=", GreaterThanEquals, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("<", LessThan, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("<=", LessThanEquals, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("-", Minus, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("%", Modulus, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("!=", NotEquals1, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("<>", NotEquals2, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN(")", ParenClose, Punctuation) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("(", ParenOpen, Punctuation) \
|
||||||
|
__ENUMERATE_SQL_TOKEN(".", Period, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("|", Pipe, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("+", Plus, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN(";", SemiColon, Punctuation) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("<<", ShiftLeft, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN(">>", ShiftRight, Operator) \
|
||||||
|
__ENUMERATE_SQL_TOKEN("~", Tilde, Operator)
|
||||||
|
|
||||||
|
enum class TokenType {
|
||||||
|
#define __ENUMERATE_SQL_TOKEN(value, type, category) type,
|
||||||
|
ENUMERATE_SQL_TOKENS
|
||||||
|
#undef __ENUMERATE_SQL_TOKEN
|
||||||
|
_COUNT_OF_TOKENS,
|
||||||
|
};
|
||||||
|
|
||||||
|
enum class TokenCategory {
|
||||||
|
Invalid,
|
||||||
|
Keyword,
|
||||||
|
Identifier,
|
||||||
|
Number,
|
||||||
|
Operator,
|
||||||
|
Punctuation,
|
||||||
|
};
|
||||||
|
|
||||||
|
class Token {
|
||||||
|
public:
|
||||||
|
Token(TokenType type, StringView value, size_t line_number, size_t line_column)
|
||||||
|
: m_type(type)
|
||||||
|
, m_value(value)
|
||||||
|
, m_line_number(line_number)
|
||||||
|
, m_line_column(line_column)
|
||||||
|
{
|
||||||
|
}
|
||||||
|
|
||||||
|
static StringView name(TokenType);
|
||||||
|
static TokenCategory category(TokenType);
|
||||||
|
|
||||||
|
StringView name() const { return name(m_type); }
|
||||||
|
TokenType type() const { return m_type; }
|
||||||
|
TokenCategory category() const { return category(m_type); }
|
||||||
|
|
||||||
|
StringView value() const { return m_value; }
|
||||||
|
double double_value() const;
|
||||||
|
|
||||||
|
size_t line_number() const { return m_line_number; }
|
||||||
|
size_t line_column() const { return m_line_column; }
|
||||||
|
|
||||||
|
private:
|
||||||
|
TokenType m_type;
|
||||||
|
StringView m_value;
|
||||||
|
size_t m_line_number;
|
||||||
|
size_t m_line_column;
|
||||||
|
};
|
||||||
|
|
||||||
|
}
|
Loading…
Add table
Add a link
Reference in a new issue