Compare commits

...

3 Commits

Author SHA1 Message Date
Greg Shuflin
0cb0145cc5 Unified error struct 2021-10-14 04:11:53 -07:00
Greg Shuflin
075e323239 Tokenizing tests 2021-10-14 03:12:05 -07:00
Greg Shuflin
fcbf2d959b Treat unclosed comment as error 2021-10-14 03:05:25 -07:00
4 changed files with 144 additions and 48 deletions

View File

@ -0,0 +1,99 @@
use crate::schala::{SourceReference, Stage};
use crate::source_map::Location;
use crate::tokenizing::{Token, TokenKind};
use crate::parsing::ParseError;
use crate::typechecking::TypeError;
pub struct SchalaError {
errors: Vec<Error>,
//TODO unify these sometime
formatted_parse_error: Option<String>,
}
impl SchalaError {
pub(crate) fn display(&self) -> String {
if let Some(ref err) = self.formatted_parse_error {
err.clone()
} else {
self.errors[0].text.as_ref().cloned().unwrap_or_default()
}
}
pub(crate) fn from_type_error(err: TypeError) -> Self {
Self {
formatted_parse_error: None,
errors: vec![
Error { location: None, text: Some(err.msg), stage: Stage::Typechecking }
]
}
}
pub(crate) fn from_string(text: String, stage: Stage) -> Self {
Self {
formatted_parse_error: None,
errors: vec![
Error { location: None, text: Some(text), stage }
]
}
}
pub(crate) fn from_parse_error(parse_error: ParseError, source_reference: &SourceReference) -> Self {
Self {
formatted_parse_error: Some(format_parse_error(parse_error, source_reference)),
errors: vec![],
}
}
pub(crate) fn from_tokens(tokens: &[Token]) -> Option<SchalaError> {
let token_errors: Vec<Error> = tokens.iter()
.filter_map(|tok| match tok.kind {
TokenKind::Error(ref err) => Some(Error {
location: Some(tok.location),
text: Some(err.clone()),
stage: Stage::Tokenizing,
}),
_ => None
}).collect();
if token_errors.is_empty() {
None
} else {
Some(SchalaError {
errors: token_errors,
formatted_parse_error: None,
})
}
}
}
#[allow(dead_code)]
struct Error {
location: Option<Location>,
text: Option<String>,
stage: Stage,
}
fn format_parse_error(error: ParseError, source_reference: &SourceReference) -> String {
let line_num = error.token.location.line_num;
let ch = error.token.location.char_num;
let line_from_program = source_reference.get_line(line_num);
let location_pointer = format!("{}^", " ".repeat(ch));
let line_num_digits = format!("{}", line_num).chars().count();
let space_padding = " ".repeat(line_num_digits);
let production = match error.production_name {
Some(n) => format!("\n(from production \"{}\")", n),
None => "".to_string()
};
format!(r#"
{error_msg}{production}
{space_padding} |
{line_num} | {}
{space_padding} | {}
"#, line_from_program, location_pointer, error_msg=error.msg, space_padding=space_padding, line_num=line_num, production=production
)
}

View File

@ -31,6 +31,7 @@ mod builtin;
mod reduced_ast; mod reduced_ast;
mod eval; mod eval;
mod source_map; mod source_map;
mod error;
mod schala; mod schala;

View File

@ -7,6 +7,7 @@ use schala_repl::{ProgrammingLanguageInterface,
ComputationRequest, ComputationResponse, ComputationRequest, ComputationResponse,
LangMetaRequest, LangMetaResponse, GlobalOutputStats}; LangMetaRequest, LangMetaResponse, GlobalOutputStats};
use crate::{reduced_ast, tokenizing, parsing, eval, typechecking, symbol_table, source_map}; use crate::{reduced_ast, tokenizing, parsing, eval, typechecking, symbol_table, source_map};
use crate::error::SchalaError;
pub type SymbolTableHandle = Rc<RefCell<symbol_table::SymbolTable>>; pub type SymbolTableHandle = Rc<RefCell<symbol_table::SymbolTable>>;
pub type SourceMapHandle = Rc<RefCell<source_map::SourceMap>>; pub type SourceMapHandle = Rc<RefCell<source_map::SourceMap>>;
@ -60,8 +61,8 @@ impl Schala {
let mut env = Schala::new_blank_env(); let mut env = Schala::new_blank_env();
let response = env.run_pipeline(prelude); let response = env.run_pipeline(prelude);
if let Err(msg) = response { if let Err(err) = response {
panic!("Error in prelude, panicking: {}", msg); panic!("Error in prelude, panicking: {}", err.display());
} }
env env
} }
@ -69,36 +70,33 @@ impl Schala {
/// This is where the actual action of interpreting/compilation happens. /// This is where the actual action of interpreting/compilation happens.
/// Note: this should eventually use a query-based system for parallelization, cf. /// Note: this should eventually use a query-based system for parallelization, cf.
/// https://rustc-dev-guide.rust-lang.org/overview.html /// https://rustc-dev-guide.rust-lang.org/overview.html
fn run_pipeline(&mut self, source: &str) -> Result<String, String> { fn run_pipeline(&mut self, source: &str) -> Result<String, SchalaError> {
//TODO `run_pipeline` should return a formatted error struct
//TODO every stage should have a common error-format that gets turned into a repl-appropriate
//string in `run_computation`
// 1st stage - tokenization // 1st stage - tokenization
// TODO tokenize should return its own error type // TODO tokenize should return its own error type
let tokens = tokenizing::tokenize(source); let tokens = tokenizing::tokenize(source);
let token_errors: Vec<String> = tokens.iter().filter_map(|t| t.get_error()).collect(); if let Some(err) = SchalaError::from_tokens(&tokens) {
if token_errors.len() > 0 { return Err(err)
return Err(format!("{:?}", token_errors));
} }
//2nd stage - parsing //2nd stage - parsing
self.active_parser.add_new_tokens(tokens); self.active_parser.add_new_tokens(tokens);
let mut ast = self.active_parser.parse() let mut ast = self.active_parser.parse()
.map_err(|err| format_parse_error(err, &self.source_reference))?; .map_err(|err| SchalaError::from_parse_error(err, &self.source_reference))?;
// Symbol table // Symbol table
self.symbol_table.borrow_mut().add_top_level_symbols(&ast)?; self.symbol_table.borrow_mut().add_top_level_symbols(&ast)
.map_err(|err| SchalaError::from_string(err, Stage::Symbols))?;
// Scope resolution - requires mutating AST // Scope resolution - requires mutating AST
self.resolver.resolve(&mut ast)?; self.resolver.resolve(&mut ast)
.map_err(|err| SchalaError::from_string(err, Stage::ScopeResolution))?;
// Typechecking // Typechecking
// TODO typechecking not working // TODO typechecking not working
let _overall_type = self.type_context.typecheck(&ast) let _overall_type = self.type_context.typecheck(&ast)
.map_err(|err| format!("Type error: {}", err.msg)); .map_err(|err| SchalaError::from_type_error(err));
// Reduce AST - this doesn't produce an error yet, but probably should // Reduce AST - TODO this doesn't produce an error yet, but probably should
let symbol_table = self.symbol_table.borrow(); let symbol_table = self.symbol_table.borrow();
let reduced_ast = reduced_ast::reduce(&ast, &symbol_table); let reduced_ast = reduced_ast::reduce(&ast, &symbol_table);
@ -108,6 +106,9 @@ impl Schala {
.into_iter() .into_iter()
.collect(); .collect();
let text_output: Result<Vec<String>, SchalaError> = text_output
.map_err(|err| SchalaError::from_string(err, Stage::Evaluation));
let eval_output: String = text_output let eval_output: String = text_output
.map(|v| { Iterator::intersperse(v.into_iter(), "\n".to_owned()).collect() })?; .map(|v| { Iterator::intersperse(v.into_iter(), "\n".to_owned()).collect() })?;
@ -115,31 +116,9 @@ impl Schala {
} }
} }
fn format_parse_error(error: parsing::ParseError, source_reference: &SourceReference) -> String {
let line_num = error.token.location.line_num;
let ch = error.token.location.char_num;
let line_from_program = source_reference.get_line(line_num);
let location_pointer = format!("{}^", " ".repeat(ch));
let line_num_digits = format!("{}", line_num).chars().count();
let space_padding = " ".repeat(line_num_digits);
let production = match error.production_name {
Some(n) => format!("\n(from production \"{}\")", n),
None => "".to_string()
};
format!(r#"
{error_msg}{production}
{space_padding} |
{line_num} | {}
{space_padding} | {}
"#, line_from_program, location_pointer, error_msg=error.msg, space_padding=space_padding, line_num=line_num, production=production
)
}
/// Represents lines of source code /// Represents lines of source code
struct SourceReference { pub(crate) struct SourceReference {
lines: Option<Vec<String>> lines: Option<Vec<String>>
} }
@ -152,11 +131,23 @@ impl SourceReference {
//TODO this is a lot of heap allocations - maybe there's a way to make it more efficient? //TODO this is a lot of heap allocations - maybe there's a way to make it more efficient?
self.lines = Some(source.lines().map(|s| s.to_string()).collect()); } self.lines = Some(source.lines().map(|s| s.to_string()).collect()); }
fn get_line(&self, line: usize) -> String { pub fn get_line(&self, line: usize) -> String {
self.lines.as_ref().and_then(|x| x.get(line).map(|s| s.to_string())).unwrap_or(format!("NO LINE FOUND")) self.lines.as_ref().and_then(|x| x.get(line).map(|s| s.to_string())).unwrap_or(format!("NO LINE FOUND"))
} }
} }
#[allow(dead_code)]
#[derive(Clone, Copy, Debug)]
pub(crate) enum Stage {
Tokenizing,
Parsing,
Symbols,
ScopeResolution,
Typechecking,
AstReduction,
Evaluation,
}
fn stage_names() -> Vec<&'static str> { fn stage_names() -> Vec<&'static str> {
vec![ vec![
"tokenizing", "tokenizing",
@ -185,7 +176,8 @@ impl ProgrammingLanguageInterface for Schala {
self.source_reference.load_new_source(source); self.source_reference.load_new_source(source);
let sw = Stopwatch::start_new(); let sw = Stopwatch::start_new();
let main_output = self.run_pipeline(source); let main_output = self.run_pipeline(source)
.map_err(|schala_err| schala_err.display());
let global_output_stats = GlobalOutputStats { let global_output_stats = GlobalOutputStats {
total_duration: sw.elapsed(), total_duration: sw.elapsed(),

View File

@ -99,12 +99,6 @@ pub struct Token {
} }
impl Token { impl Token {
pub fn get_error(&self) -> Option<String> {
match self.kind {
TokenKind::Error(ref s) => Some(s.clone()),
_ => None,
}
}
pub fn to_string_with_metadata(&self) -> String { pub fn to_string_with_metadata(&self) -> String {
format!("{}({})", self.kind, self.location) format!("{}({})", self.kind, self.location)
} }
@ -156,7 +150,11 @@ pub fn tokenize(input: &str) -> Vec<Token> {
break; break;
} }
} }
continue; if comment_level != 0 {
Error("Unclosed comment".to_string())
} else {
continue;
}
}, },
_ => Slash _ => Slash
}, },
@ -213,7 +211,7 @@ fn handle_quote(input: &mut Peekable<impl Iterator<Item=CharData>>, quote_prefix
} }
}, },
Some(c) => buf.push(c), Some(c) => buf.push(c),
None => return TokenKind::Error(format!("Unclosed string")), None => return TokenKind::Error("Unclosed string".to_string()),
} }
} }
TokenKind::StrLiteral { s: Rc::new(buf), prefix: quote_prefix.map(|s| Rc::new(s.to_string())) } TokenKind::StrLiteral { s: Rc::new(buf), prefix: quote_prefix.map(|s| Rc::new(s.to_string())) }
@ -325,6 +323,9 @@ mod schala_tokenizer_tests {
fn comments() { fn comments() {
let token_kinds: Vec<TokenKind> = tokenize("1 + /* hella /* bro */ */ 2").into_iter().map(move |t| t.kind).collect(); let token_kinds: Vec<TokenKind> = tokenize("1 + /* hella /* bro */ */ 2").into_iter().map(move |t| t.kind).collect();
assert_eq!(token_kinds, vec![digit!("1"), op!("+"), digit!("2")]); assert_eq!(token_kinds, vec![digit!("1"), op!("+"), digit!("2")]);
let token_kinds: Vec<TokenKind> = tokenize("1 + /* hella /* bro */ 2").into_iter().map(move |t| t.kind).collect();
assert_eq!(token_kinds, vec![digit!("1"), op!("+"), Error("Unclosed comment".to_string())]);
} }
#[test] #[test]
@ -340,5 +341,8 @@ mod schala_tokenizer_tests {
let token_kinds: Vec<TokenKind> = tokenize(r#"b"some bytestring""#).into_iter().map(move |t| t.kind).collect(); let token_kinds: Vec<TokenKind> = tokenize(r#"b"some bytestring""#).into_iter().map(move |t| t.kind).collect();
assert_eq!(token_kinds, vec![StrLiteral { s: Rc::new("some bytestring".to_string()), prefix: Some(Rc::new("b".to_string())) }]); assert_eq!(token_kinds, vec![StrLiteral { s: Rc::new("some bytestring".to_string()), prefix: Some(Rc::new("b".to_string())) }]);
let token_kinds: Vec<TokenKind> = tokenize(r#""Do \n \" escapes work\t""#).into_iter().map(move |t| t.kind).collect();
assert_eq!(token_kinds, vec![StrLiteral { s: Rc::new("Do \n \" escapes work\t".to_string()), prefix: None }]);
} }
} }