generated from ElnuDev/rust-project
Compare commits
No commits in common. "10e966ef050359d440eb4cc2e59e51bb0e091afe" and "a518097ff0c2f600f7100036117b8959edfd2a15" have entirely different histories.
10e966ef05
...
a518097ff0
4 changed files with 43 additions and 129 deletions
|
@ -1,3 +1,12 @@
|
||||||
"Bob sat on the bench."
|
show black amogus # this is a comment
|
||||||
"Bob" "Good morning!"
|
# this is a full line comment
|
||||||
eat
|
what the heck
|
||||||
|
"this is a string with a # comment"
|
||||||
|
"this is a string over
|
||||||
|
multiple lines"
|
||||||
|
"this is \"escaped\""
|
||||||
|
'this is a single quote string'
|
||||||
|
'this also has escaped \'quotes\''
|
||||||
|
this is cool # comment
|
||||||
|
|
||||||
|
huh
|
|
@ -1,6 +1,5 @@
|
||||||
|
use renrs;
|
||||||
|
|
||||||
fn main() {
|
fn main() {
|
||||||
let commands = renrs::parse_file("demo.rpy");
|
renrs::parse("demo.rpy");
|
||||||
for command in commands {
|
|
||||||
println!("{:?}", command);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
138
src/lib.rs
138
src/lib.rs
|
@ -7,126 +7,40 @@ use pest_derive::Parser;
|
||||||
#[grammar = "rpy.pest"]
|
#[grammar = "rpy.pest"]
|
||||||
struct RpyParser;
|
struct RpyParser;
|
||||||
|
|
||||||
// Raw script tokens
|
pub fn parse(file_path: &str) {
|
||||||
#[derive(Debug)]
|
let unparsed_file = fs::read_to_string(file_path).expect("cannot find file");
|
||||||
pub enum Token {
|
let file = RpyParser::parse(Rule::file, &unparsed_file)
|
||||||
Keyword(String),
|
.expect("unsuccessful parse") // unwrap the parse result
|
||||||
Str(String),
|
.next().unwrap(); // get and unwrap the `file` rule; never fails
|
||||||
Array(Vec<Token>),
|
|
||||||
}
|
|
||||||
|
|
||||||
impl Token {
|
|
||||||
fn print(&self) -> &str {
|
|
||||||
match &self {
|
|
||||||
Keyword(keyword) => keyword,
|
|
||||||
Str(_) => "String",
|
|
||||||
Array(_) => "Array",
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
use Token::*;
|
|
||||||
|
|
||||||
// Parsed script commands
|
|
||||||
#[derive(Debug)]
|
|
||||||
pub enum Command {
|
|
||||||
Say {
|
|
||||||
name: Option<String>,
|
|
||||||
text: String,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
use Command::*;
|
|
||||||
|
|
||||||
// Tokenize raw script string
|
|
||||||
fn tokenize(script: &str) -> Vec<Vec<Token>> {
|
|
||||||
let file = RpyParser::parse(Rule::file, script)
|
|
||||||
.expect("unsuccessful parse")
|
|
||||||
.next().unwrap();
|
|
||||||
let mut lines = Vec::new();
|
|
||||||
for line in file.into_inner() {
|
for line in file.into_inner() {
|
||||||
let mut tokens = Vec::new();
|
|
||||||
match line.as_rule() {
|
match line.as_rule() {
|
||||||
Rule::line => {
|
Rule::line => {
|
||||||
|
println!("Line:");
|
||||||
for token in line.into_inner() {
|
for token in line.into_inner() {
|
||||||
tokens.push(parse_pair(token));
|
match token.as_rule() {
|
||||||
|
Rule::token => {
|
||||||
|
let token = token.into_inner().next().unwrap();
|
||||||
|
match token.as_rule() {
|
||||||
|
Rule::string => {
|
||||||
|
let string_data = token.into_inner().next().unwrap();
|
||||||
|
let str = string_data.as_str();
|
||||||
|
println!("string: {}", match string_data.as_rule() {
|
||||||
|
Rule::single_quote_string_data => str.replace("\\'", "'"),
|
||||||
|
Rule::double_quote_string_data => str.replace("\\\"", "\""),
|
||||||
|
_ => unreachable!(),
|
||||||
|
});
|
||||||
|
},
|
||||||
|
Rule::keyword => println!("keyword: {}", token.as_str()),
|
||||||
|
_ => unreachable!(),
|
||||||
|
};
|
||||||
|
},
|
||||||
|
_ => unreachable!(),
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
println!()
|
||||||
},
|
},
|
||||||
Rule::EOI => (),
|
Rule::EOI => (),
|
||||||
_ => unreachable!(),
|
_ => unreachable!(),
|
||||||
}
|
}
|
||||||
// TODO: For some a blank final line is always parsed
|
|
||||||
if tokens.len() > 0 {
|
|
||||||
lines.push(tokens);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
lines
|
|
||||||
}
|
|
||||||
|
|
||||||
// Parse raw pest data into Token
|
|
||||||
fn parse_pair(pair: pest::iterators::Pair<Rule>) -> Token {
|
|
||||||
let token = pair.as_rule();
|
|
||||||
match token {
|
|
||||||
Rule::token => {},
|
|
||||||
_ => panic!("Not a token!"),
|
|
||||||
};
|
|
||||||
let contents = pair.into_inner().next().unwrap();
|
|
||||||
let contents_rule = contents.as_rule();
|
|
||||||
match contents_rule {
|
|
||||||
Rule::string => {
|
|
||||||
let data = contents.into_inner().next().unwrap();
|
|
||||||
Token::Str(match data.as_rule() {
|
|
||||||
Rule::single_quote_string_data => data.as_str().replace("\\'", "'"),
|
|
||||||
Rule::double_quote_string_data => data.as_str().replace("\\\"", "\""),
|
|
||||||
_ => unreachable!(),
|
|
||||||
})
|
|
||||||
},
|
|
||||||
Rule::array => {
|
|
||||||
let mut array = Vec::new();
|
|
||||||
for token in contents.into_inner() {
|
|
||||||
array.push(parse_pair(token));
|
|
||||||
}
|
|
||||||
Token::Array(array)
|
|
||||||
}
|
|
||||||
Rule::keyword => Token::Keyword(contents.as_str().to_owned()),
|
|
||||||
__ => unreachable!(),
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Tokenize file
|
|
||||||
fn tokenize_file(file_path: &str) -> Vec<Vec<Token>> {
|
|
||||||
let unparsed_file = fs::read_to_string(file_path).expect("cannot find file");
|
|
||||||
tokenize(&unparsed_file)
|
|
||||||
}
|
|
||||||
|
|
||||||
fn describe_line(line: &Vec<Token>) -> String {
|
|
||||||
let mut description = "[".to_owned();
|
|
||||||
let mut iter = line.iter();
|
|
||||||
description.push_str(&format!("{}", iter.next().unwrap().print()));
|
|
||||||
for token in iter {
|
|
||||||
description.push_str(&format!(", {}", token.print()));
|
|
||||||
}
|
|
||||||
description.push_str("]");
|
|
||||||
description
|
|
||||||
}
|
|
||||||
|
|
||||||
// Parse file into commands
|
|
||||||
pub fn parse_file(file_path: &str) -> Vec<Command> {
|
|
||||||
let token_lines = tokenize_file(file_path);
|
|
||||||
let mut commands = Vec::new();
|
|
||||||
for line in token_lines {
|
|
||||||
println!("{:?}", line);
|
|
||||||
commands.push(match line.as_slice() {
|
|
||||||
[Str(text)] => Say {
|
|
||||||
name: None,
|
|
||||||
text: text.to_owned(),
|
|
||||||
},
|
|
||||||
[Str(name), Str(text)] => Say {
|
|
||||||
name: Some(name.to_owned()),
|
|
||||||
text: text.to_owned(),
|
|
||||||
},
|
|
||||||
_ => panic!("Unknown command {}", describe_line(&line)),
|
|
||||||
});
|
|
||||||
}
|
|
||||||
commands
|
|
||||||
}
|
|
||||||
|
|
12
src/rpy.pest
12
src/rpy.pest
|
@ -8,13 +8,11 @@ char = { !NEWLINE ~ ANY }
|
||||||
// http://pest.rs/book/grammars/syntax.html#atomic
|
// http://pest.rs/book/grammars/syntax.html#atomic
|
||||||
inner = @{ char* }
|
inner = @{ char* }
|
||||||
|
|
||||||
token = { string | array | keyword }
|
token = { string | keyword }
|
||||||
|
|
||||||
// KEYWORDS
|
|
||||||
// has to be atomic for no implicit separate (spaces)
|
// has to be atomic for no implicit separate (spaces)
|
||||||
keyword = @{ (!(WHITESPACE | NEWLINE) ~ ANY)+ }
|
keyword = @{ (!(WHITESPACE | NEWLINE) ~ ANY)+ }
|
||||||
|
|
||||||
// STRING
|
|
||||||
single_quote_string_data = @{ (
|
single_quote_string_data = @{ (
|
||||||
"\\'" // Escaped single quotes
|
"\\'" // Escaped single quotes
|
||||||
| (!"'" ~ ANY)
|
| (!"'" ~ ANY)
|
||||||
|
@ -28,12 +26,6 @@ string = ${
|
||||||
| ("\"" ~ double_quote_string_data ~ "\"")
|
| ("\"" ~ double_quote_string_data ~ "\"")
|
||||||
}
|
}
|
||||||
|
|
||||||
// ARRAY
|
|
||||||
array = {
|
|
||||||
"[" ~ "]"
|
|
||||||
| "[" ~ NEWLINE* ~ token ~ ("," ~ NEWLINE* ~ token)* ~ NEWLINE* ~ "]"
|
|
||||||
}
|
|
||||||
|
|
||||||
// comments are a # followed by
|
// comments are a # followed by
|
||||||
// any number of non-newline characters
|
// any number of non-newline characters
|
||||||
COMMENT = _{ "#" ~ char* }
|
COMMENT = _{ "#" ~ char* }
|
||||||
|
@ -41,4 +33,4 @@ COMMENT = _{ "#" ~ char* }
|
||||||
// lines are comprised of a statement
|
// lines are comprised of a statement
|
||||||
line = { token+ }
|
line = { token+ }
|
||||||
|
|
||||||
file = { SOI ~ (line ~ (NEWLINE+ ~ line)*)? ~ NEWLINE* ~ EOI }
|
file = { SOI ~ line ~ (NEWLINE+ ~ line)* ~ NEWLINE* ~ EOI }
|
Loading…
Add table
Reference in a new issue