Compare commits

..

2 commits

Author SHA1 Message Date
c0c8153cbe Simplify parsing 2023-05-27 14:20:11 -07:00
6707b97ef5 Use PascalCase for exposed pest grammar 2023-05-27 13:19:45 -07:00
2 changed files with 89 additions and 130 deletions

View file

@ -32,19 +32,15 @@ impl Token {
use Token::*; use Token::*;
// Parsing types
type Line = Vec<Token>;
// Indented command block // Indented command block
#[derive(Debug)] #[derive(Debug)]
struct Block<T> { struct CommandBlock {
elements: Vec<BlockElement<T>>, elements: Vec<BlockElement>,
next: Option<usize>, next: Option<usize>,
} }
impl<T> Block<T> { impl CommandBlock {
fn next(&mut self) -> Option<&T> { fn next(&mut self) -> Option<&Command> {
let mut next = match self.next { let mut next = match self.next {
Some(next) => next, Some(next) => next,
None => return None, None => return None,
@ -80,7 +76,7 @@ impl<T> Block<T> {
} }
} }
impl<T> Default for Block<T> { impl Default for CommandBlock {
fn default() -> Self { fn default() -> Self {
Self { Self {
elements: Vec::new(), elements: Vec::new(),
@ -90,13 +86,11 @@ impl<T> Default for Block<T> {
} }
#[derive(Debug)] #[derive(Debug)]
enum BlockElement<T> { enum BlockElement {
Command(T), Command(Command),
Block(Block<T>), Block(CommandBlock),
} }
type LineBlock = Block<Line>;
type CommandBlock = Block<Command>;
type Script = CommandBlock; type Script = CommandBlock;
// Parsed script commands // Parsed script commands
@ -123,116 +117,44 @@ pub enum Event {
Say { name: Option<String>, text: String }, Say { name: Option<String>, text: String },
} }
// ========================================== // ======
// Step 1 parsing // Parser
// converting from pest pairs to Token blocks // ======
// ==========================================
// Tokenize raw script string type Pair<'a> = pest::iterators::Pair<'a, Rule>;
fn parse(script: &str) -> LineBlock {
let file = RpyParser::parse(Rule::file, script) // Read file into commands
fn parse_file(file_path: &PathBuf) -> CommandBlock {
let unparsed_file = fs::read_to_string(file_path).expect("cannot find file");
parse(&unparsed_file)
}
fn parse(script: &str) -> CommandBlock {
let file = RpyParser::parse(Rule::File, script)
.expect("unsuccessful parse") .expect("unsuccessful parse")
.next() .next()
.unwrap(); .unwrap();
parse_block(file) parse_block(file)
} }
type Pair<'a> = pest::iterators::Pair<'a, Rule>; fn parse_block(block: Pair) -> CommandBlock {
// Tokenize block
fn parse_block(pair: Pair) -> LineBlock {
let mut block = LineBlock::default();
for element in pair.into_inner() {
block.elements.push(match element.as_rule() {
Rule::block => BlockElement::Block(parse_block(element)),
Rule::line => {
let line = parse_line(element);
// TODO: For some reason a blank final line is always parsed
if line.len() == 0 {
continue;
}
BlockElement::Command(line)
},
Rule::EOI => break, // end
_ => unreachable!(),
});
}
block
}
// Tokenize line
fn parse_line(pair: Pair) -> Line {
let mut tokens = Vec::new();
for token in pair.into_inner() {
tokens.push(parse_token(token));
}
tokens
}
// Tokenize token
fn parse_token(pair: Pair) -> Token {
let token = pair.as_rule();
macro_rules! contents {
() => {
pair.into_inner().next().unwrap()
};
}
match token {
Rule::string => {
let contents = contents!();
Token::Str(match contents.as_rule() {
Rule::single_quote_string_data => contents.as_str().replace("\\'", "'"),
Rule::double_quote_string_data => contents.as_str().replace("\\\"", "\""),
_ => unreachable!(),
})
}
Rule::array => {
let contents = contents!();
let mut array = Vec::new();
for token in contents.into_inner() {
array.push(parse_token(token));
}
Token::Array(array)
}
Rule::boolean => Token::Boolean(match pair.as_str() {
"True" => true,
"False" => false,
_ => unreachable!(),
}),
Rule::number => Token::Number(pair.as_str().parse().unwrap()),
Rule::keyword => Token::Keyword(pair.as_str().to_owned()),
__ => unreachable!(),
}
}
// ==============================================
// Step 2 reading
// converting from Token blocks to Command blocks
// ==============================================
// Read file into commands
fn read_file(file_path: &PathBuf) -> CommandBlock {
let line_block = {
let unparsed_file = fs::read_to_string(file_path).expect("cannot find file");
parse(&unparsed_file)
};
read_block(&line_block)
}
// Read line block into command block
fn read_block(block: &LineBlock) -> CommandBlock {
CommandBlock { CommandBlock {
elements: block.elements.iter().map(|element| match element { elements: block.into_inner().filter_map(|pair| match pair.as_rule() {
BlockElement::Command(line) => BlockElement::Command(read_command(&line)), Rule::Line => Some(BlockElement::Command(parse_command(pair))),
BlockElement::Block(block) => BlockElement::Block(read_block(&block)), Rule::Block => Some(BlockElement::Block(parse_block(pair))),
Rule::EOI => None, // end
_ => unreachable!(),
}).collect(), }).collect(),
..Default::default() ..Default::default()
} }
} }
// Read token array to command fn parse_command(pair: Pair) -> Command {
fn read_command(line: &Line) -> Command { use Token::*;
let line: Vec<Token> = pair
.into_inner()
.map(|pair| parse_token(pair))
.collect();
macro_rules! unknown { macro_rules! unknown {
() => { () => {
panic!("Unknown command {}", describe_line(&line)) panic!("Unknown command {}", describe_line(&line))
@ -260,7 +182,7 @@ fn read_command(line: &Line) -> Command {
// Line description e.g. [String, Keyword, Array] // Line description e.g. [String, Keyword, Array]
// Used in parse_command as feedback for invalid commands // Used in parse_command as feedback for invalid commands
fn describe_line(line: &Line) -> String { fn describe_line(line: &[Token]) -> String {
let mut description = "[".to_owned(); let mut description = "[".to_owned();
let mut iter = line.iter(); let mut iter = line.iter();
description.push_str(&format!("{}", iter.next().unwrap().print())); description.push_str(&format!("{}", iter.next().unwrap().print()));
@ -271,6 +193,41 @@ fn describe_line(line: &Line) -> String {
description description
} }
fn parse_token(pair: Pair) -> Token {
let token = pair.as_rule();
macro_rules! contents {
() => {
pair.into_inner().next().unwrap()
};
}
match token {
Rule::String => {
let contents = contents!();
Token::Str(match contents.as_rule() {
Rule::SingleQuoteStringData => contents.as_str().replace("\\'", "'"),
Rule::DoubleQuoteStringData => contents.as_str().replace("\\\"", "\""),
_ => unreachable!(),
})
}
Rule::Array => {
let contents = contents!();
let mut array = Vec::new();
for token in contents.into_inner() {
array.push(parse_token(token));
}
Token::Array(array)
}
Rule::Boolean => Token::Boolean(match pair.as_str() {
"True" => true,
"False" => false,
_ => unreachable!(),
}),
Rule::Number => Token::Number(pair.as_str().parse().unwrap()),
Rule::Keyword => Token::Keyword(pair.as_str().to_owned()),
__ => unreachable!(),
}
}
// ===== // =====
// State // State
// ===== // =====
@ -282,7 +239,7 @@ pub struct State {
impl State { impl State {
pub fn from_file(file: PathBuf) -> State { pub fn from_file(file: PathBuf) -> State {
State { State {
script: read_file(&file), script: parse_file(&file),
} }
} }

View file

@ -1,56 +1,58 @@
// characters are anything but newlines // characters are anything but newlines
char = { !NEWLINE ~ ANY } char = { !NEWLINE ~ ANY }
// token definition // Token definition
// http://pest.rs/book/grammars/syntax.html#atomic // http://pest.rs/book/grammars/syntax.html#atomic
inner = @{ char* } inner = @{ char* }
token = { string | array | boolean | number | keyword } token = { String | Array | Boolean | Number | Keyword }
// KEYWORDS // KEYWORDS
// has to be atomic for no implicit separate (spaces) // has to be atomic for no implicit separate (spaces)
keyword = ${ (!(whitespace | NEWLINE) ~ ANY)+ } Keyword = ${ (!(whitespace | NEWLINE) ~ ANY)+ }
// STRING // STRING
single_quote_string_data = @{ ( SingleQuoteStringData = @{ (
"\\'" // Escaped single quotes "\\'" // Escaped single quotes
| (!"'" ~ ANY) | (!"'" ~ ANY)
)* } )* }
double_quote_string_data = @{ ( DoubleQuoteStringData = @{ (
"\\\"" // Escaped double quotes "\\\"" // Escaped double quotes
| (!"\"" ~ ANY) | (!"\"" ~ ANY)
)* } )* }
string = ${ String = ${
("'" ~ single_quote_string_data ~ "'") ("'" ~ SingleQuoteStringData ~ "'")
| ("\"" ~ double_quote_string_data ~ "\"") | ("\"" ~ DoubleQuoteStringData ~ "\"")
} }
// ARRAY // Array
array = ${ Array = ${
"[" ~ "]" "[" ~ "]"
| "[" ~ whitespace* ~ NEWLINE* ~ whitespace* ~ token ~ ("," ~ whitespace* ~ NEWLINE* ~ whitespace* ~ token)* ~ NEWLINE* ~ "]" | "[" ~ whitespace* ~ NEWLINE* ~ whitespace* ~ token ~ ("," ~ whitespace* ~ NEWLINE* ~ whitespace* ~ token)* ~ NEWLINE* ~ "]"
} }
// BOOLEAN // BOOLEAN
boolean = ${ "True" | "False" } Boolean = ${ "True" | "False" }
// NUMBER // NUMBER
number = @{ Number = @{
"-"? "-"?
~ ("0" | ASCII_NONZERO_DIGIT ~ ASCII_DIGIT*) ~ ("0" | ASCII_NONZERO_DIGIT ~ ASCII_DIGIT*)
~ ("." ~ ASCII_DIGIT*)? ~ ("." ~ ASCII_DIGIT*)?
} }
// comments are a # followed by // comments are a # followed by
// any number of non-newline characters // any Number of non-newline characters
COMMENT = _{ "#" ~ char* } COMMENT = _{ "#" ~ char* }
Colon = { ":" }
// lines are comprised of a statement // lines are comprised of a statement
line = @{ (token ~ whitespace+)* ~ token } Line = @{ (token ~ whitespace+)* ~ token ~ Colon? }
file = { SOI ~ NEWLINE* ~ block_content* ~ NEWLINE* ~ EOI } File = { SOI ~ NEWLINE* ~ block_content* ~ NEWLINE* ~ EOI }
block = { Block = {
// The first line in the block // The first line in the block
PEEK_ALL ~ PUSH(" "+ | "\t"+) ~ block_content ~ PEEK_ALL ~ PUSH(" "+ | "\t"+) ~ block_content ~
// Subsequent lines in the block // Subsequent lines in the block
@ -64,5 +66,5 @@ block = {
whitespace = _{ " " } whitespace = _{ " " }
block_content = _{ block_content = _{
line ~ (whitespace+ ~ line)* ~ (NEWLINE | EOI) ~ block* Line ~ (whitespace+ ~ Line)* ~ (NEWLINE | EOI) ~ Block*
} }