mirror of
https://github.com/RGBCube/serenity
synced 2025-07-15 08:07:38 +00:00
LibWeb: Move CSS TokenStream to its own file
This lets us use TokenStream in StyleComputer without every user of StyleComputer pulling in the entire CSS Parser.
This commit is contained in:
parent
bf6d5fce18
commit
8ce38fddd9
3 changed files with 140 additions and 146 deletions
|
@ -81,87 +81,6 @@ AK::URL ParsingContext::complete_url(String const& addr) const
|
|||
return m_url.complete_url(addr);
|
||||
}
|
||||
|
||||
template<typename T>
|
||||
TokenStream<T>::TokenStream(Vector<T> const& tokens)
|
||||
: m_tokens(tokens)
|
||||
, m_eof(make_eof())
|
||||
{
|
||||
}
|
||||
|
||||
template<typename T>
|
||||
bool TokenStream<T>::has_next_token()
|
||||
{
|
||||
return (size_t)(m_iterator_offset + 1) < m_tokens.size();
|
||||
}
|
||||
|
||||
template<typename T>
|
||||
T const& TokenStream<T>::peek_token(int offset)
|
||||
{
|
||||
if (!has_next_token())
|
||||
return m_eof;
|
||||
|
||||
return m_tokens.at(m_iterator_offset + offset + 1);
|
||||
}
|
||||
|
||||
template<typename T>
|
||||
T const& TokenStream<T>::next_token()
|
||||
{
|
||||
if (!has_next_token())
|
||||
return m_eof;
|
||||
|
||||
++m_iterator_offset;
|
||||
|
||||
return m_tokens.at(m_iterator_offset);
|
||||
}
|
||||
|
||||
template<typename T>
|
||||
T const& TokenStream<T>::current_token()
|
||||
{
|
||||
if ((size_t)m_iterator_offset >= m_tokens.size())
|
||||
return m_eof;
|
||||
|
||||
return m_tokens.at(m_iterator_offset);
|
||||
}
|
||||
|
||||
template<typename T>
|
||||
void TokenStream<T>::reconsume_current_input_token()
|
||||
{
|
||||
if (m_iterator_offset >= 0)
|
||||
--m_iterator_offset;
|
||||
}
|
||||
|
||||
template<typename T>
|
||||
void TokenStream<T>::skip_whitespace()
|
||||
{
|
||||
while (peek_token().is(Token::Type::Whitespace))
|
||||
next_token();
|
||||
}
|
||||
|
||||
template<>
|
||||
Token TokenStream<Token>::make_eof()
|
||||
{
|
||||
return Tokenizer::create_eof_token();
|
||||
}
|
||||
|
||||
template<>
|
||||
ComponentValue TokenStream<ComponentValue>::make_eof()
|
||||
{
|
||||
return ComponentValue(Tokenizer::create_eof_token());
|
||||
}
|
||||
|
||||
template<typename T>
|
||||
void TokenStream<T>::dump_all_tokens()
|
||||
{
|
||||
dbgln("Dumping all tokens:");
|
||||
for (size_t i = 0; i < m_tokens.size(); ++i) {
|
||||
auto& token = m_tokens[i];
|
||||
if ((i - 1) == (size_t)m_iterator_offset)
|
||||
dbgln("-> {}", token.to_debug_string());
|
||||
else
|
||||
dbgln(" {}", token.to_debug_string());
|
||||
}
|
||||
}
|
||||
|
||||
Parser::Parser(ParsingContext const& context, StringView input, String const& encoding)
|
||||
: m_context(context)
|
||||
, m_tokenizer(input, encoding)
|
||||
|
|
|
@ -22,6 +22,7 @@
|
|||
#include <LibWeb/CSS/Parser/DeclarationOrAtRule.h>
|
||||
#include <LibWeb/CSS/Parser/Function.h>
|
||||
#include <LibWeb/CSS/Parser/Rule.h>
|
||||
#include <LibWeb/CSS/Parser/TokenStream.h>
|
||||
#include <LibWeb/CSS/Parser/Tokenizer.h>
|
||||
#include <LibWeb/CSS/Ratio.h>
|
||||
#include <LibWeb/CSS/Selector.h>
|
||||
|
@ -56,71 +57,6 @@ private:
|
|||
AK::URL m_url;
|
||||
};
|
||||
|
||||
template<typename T>
|
||||
class TokenStream {
|
||||
public:
|
||||
class StateTransaction {
|
||||
public:
|
||||
explicit StateTransaction(TokenStream<T>& token_stream)
|
||||
: m_token_stream(token_stream)
|
||||
, m_saved_iterator_offset(token_stream.m_iterator_offset)
|
||||
{
|
||||
}
|
||||
|
||||
~StateTransaction()
|
||||
{
|
||||
if (!m_commit)
|
||||
m_token_stream.m_iterator_offset = m_saved_iterator_offset;
|
||||
}
|
||||
|
||||
StateTransaction create_child() { return StateTransaction(*this); }
|
||||
|
||||
void commit()
|
||||
{
|
||||
m_commit = true;
|
||||
if (m_parent)
|
||||
m_parent->commit();
|
||||
}
|
||||
|
||||
private:
|
||||
explicit StateTransaction(StateTransaction& parent)
|
||||
: m_parent(&parent)
|
||||
, m_token_stream(parent.m_token_stream)
|
||||
, m_saved_iterator_offset(parent.m_token_stream.m_iterator_offset)
|
||||
{
|
||||
}
|
||||
|
||||
StateTransaction* m_parent { nullptr };
|
||||
TokenStream<T>& m_token_stream;
|
||||
int m_saved_iterator_offset { 0 };
|
||||
bool m_commit { false };
|
||||
};
|
||||
|
||||
explicit TokenStream(Vector<T> const&);
|
||||
~TokenStream() = default;
|
||||
|
||||
TokenStream(TokenStream<T> const&) = delete;
|
||||
|
||||
bool has_next_token();
|
||||
T const& next_token();
|
||||
T const& peek_token(int offset = 0);
|
||||
T const& current_token();
|
||||
void reconsume_current_input_token();
|
||||
|
||||
StateTransaction begin_transaction() { return StateTransaction(*this); }
|
||||
|
||||
void skip_whitespace();
|
||||
|
||||
void dump_all_tokens();
|
||||
|
||||
private:
|
||||
Vector<T> const& m_tokens;
|
||||
int m_iterator_offset { -1 };
|
||||
|
||||
T make_eof();
|
||||
T m_eof;
|
||||
};
|
||||
|
||||
class Parser {
|
||||
public:
|
||||
Parser(ParsingContext const&, StringView input, String const& encoding = "utf-8");
|
||||
|
|
139
Userland/Libraries/LibWeb/CSS/Parser/TokenStream.h
Normal file
139
Userland/Libraries/LibWeb/CSS/Parser/TokenStream.h
Normal file
|
@ -0,0 +1,139 @@
|
|||
/*
|
||||
* Copyright (c) 2020-2021, the SerenityOS developers.
|
||||
* Copyright (c) 2021-2022, Sam Atkins <atkinssj@serenityos.org>
|
||||
*
|
||||
* SPDX-License-Identifier: BSD-2-Clause
|
||||
*/
|
||||
|
||||
#pragma once
|
||||
|
||||
#include <AK/Format.h>
|
||||
#include <AK/Vector.h>
|
||||
#include <LibWeb/CSS/Parser/Tokenizer.h>
|
||||
|
||||
namespace Web::CSS::Parser {
|
||||
|
||||
template<typename T>
|
||||
class TokenStream {
|
||||
public:
|
||||
class StateTransaction {
|
||||
public:
|
||||
explicit StateTransaction(TokenStream<T>& token_stream)
|
||||
: m_token_stream(token_stream)
|
||||
, m_saved_iterator_offset(token_stream.m_iterator_offset)
|
||||
{
|
||||
}
|
||||
|
||||
~StateTransaction()
|
||||
{
|
||||
if (!m_commit)
|
||||
m_token_stream.m_iterator_offset = m_saved_iterator_offset;
|
||||
}
|
||||
|
||||
StateTransaction create_child() { return StateTransaction(*this); }
|
||||
|
||||
void commit()
|
||||
{
|
||||
m_commit = true;
|
||||
if (m_parent)
|
||||
m_parent->commit();
|
||||
}
|
||||
|
||||
private:
|
||||
explicit StateTransaction(StateTransaction& parent)
|
||||
: m_parent(&parent)
|
||||
, m_token_stream(parent.m_token_stream)
|
||||
, m_saved_iterator_offset(parent.m_token_stream.m_iterator_offset)
|
||||
{
|
||||
}
|
||||
|
||||
StateTransaction* m_parent { nullptr };
|
||||
TokenStream<T>& m_token_stream;
|
||||
int m_saved_iterator_offset { 0 };
|
||||
bool m_commit { false };
|
||||
};
|
||||
|
||||
explicit TokenStream(Vector<T> const& tokens)
|
||||
: m_tokens(tokens)
|
||||
, m_eof(make_eof())
|
||||
{
|
||||
}
|
||||
TokenStream(TokenStream<T> const&) = delete;
|
||||
|
||||
~TokenStream() = default;
|
||||
|
||||
bool has_next_token()
|
||||
{
|
||||
return (size_t)(m_iterator_offset + 1) < m_tokens.size();
|
||||
}
|
||||
|
||||
T const& next_token()
|
||||
{
|
||||
if (!has_next_token())
|
||||
return m_eof;
|
||||
|
||||
++m_iterator_offset;
|
||||
|
||||
return m_tokens.at(m_iterator_offset);
|
||||
}
|
||||
|
||||
T const& peek_token(int offset = 0)
|
||||
{
|
||||
if (!has_next_token())
|
||||
return m_eof;
|
||||
|
||||
return m_tokens.at(m_iterator_offset + offset + 1);
|
||||
}
|
||||
|
||||
T const& current_token()
|
||||
{
|
||||
if ((size_t)m_iterator_offset >= m_tokens.size())
|
||||
return m_eof;
|
||||
|
||||
return m_tokens.at(m_iterator_offset);
|
||||
}
|
||||
|
||||
void reconsume_current_input_token()
|
||||
{
|
||||
if (m_iterator_offset >= 0)
|
||||
--m_iterator_offset;
|
||||
}
|
||||
|
||||
StateTransaction begin_transaction() { return StateTransaction(*this); }
|
||||
|
||||
void skip_whitespace()
|
||||
{
|
||||
while (peek_token().is(Token::Type::Whitespace))
|
||||
next_token();
|
||||
}
|
||||
|
||||
void dump_all_tokens()
|
||||
{
|
||||
dbgln("Dumping all tokens:");
|
||||
for (size_t i = 0; i < m_tokens.size(); ++i) {
|
||||
auto& token = m_tokens[i];
|
||||
if ((i - 1) == (size_t)m_iterator_offset)
|
||||
dbgln("-> {}", token.to_debug_string());
|
||||
else
|
||||
dbgln(" {}", token.to_debug_string());
|
||||
}
|
||||
}
|
||||
|
||||
private:
|
||||
Vector<T> const& m_tokens;
|
||||
int m_iterator_offset { -1 };
|
||||
|
||||
T make_eof()
|
||||
{
|
||||
if constexpr (IsSame<T, Token>) {
|
||||
return Tokenizer::create_eof_token();
|
||||
}
|
||||
if constexpr (IsSame<T, ComponentValue>) {
|
||||
return ComponentValue(Tokenizer::create_eof_token());
|
||||
}
|
||||
}
|
||||
|
||||
T m_eof;
|
||||
};
|
||||
|
||||
}
|
Loading…
Add table
Add a link
Reference in a new issue