diff options
Diffstat (limited to '')
-rw-r--r-- | src/grammar.lalrpop | 2 | ||||
-rw-r--r-- | src/lexer.rs | 54 | ||||
-rw-r--r-- | src/main.rs | 2 | ||||
-rw-r--r-- | src/parser.rs | 2 |
4 files changed, 31 insertions, 29 deletions
diff --git a/src/grammar.lalrpop b/src/grammar.lalrpop index 1e4d323..35716fc 100644 --- a/src/grammar.lalrpop +++ b/src/grammar.lalrpop @@ -13,7 +13,7 @@ grammar<'input>; extern { type Location = usize; - type Error = LexicalError<'input>; + type Error = LexicalError; enum Tok<'input> { Pi => Tok::Pi, diff --git a/src/lexer.rs b/src/lexer.rs index 8617f4b..0b58293 100644 --- a/src/lexer.rs +++ b/src/lexer.rs @@ -64,7 +64,10 @@ pub enum Tok<'i> { } #[derive(Debug)] -pub struct LexicalError<'a>(pub usize, pub nom::Err<&'a str>); +pub enum LexicalError { + Error(usize, nom::simple_errors::Err<u32>), + Incomplete(nom::Needed), +} pub type Spanned<Tok, Loc, Error> = Result<(Loc, Tok, Loc), Error>; @@ -106,11 +109,11 @@ named!(identifier<&str, &str>, recognize!(preceded!( macro_rules! ident_tag { ($i:expr, $tag:expr) => { match identifier($i) { - Ok((i, s)) => { + nom::IResult::Done(i, s) => { if s == $tag { - Ok((i, s)) + nom::IResult::Done(i, s) } else { - Err(nom::Err::Error(error_position!($i, nom::ErrorKind::Tag))) + nom::IResult::Error(error_position!(nom::ErrorKind::Tag, $i)) } } r => r, @@ -143,20 +146,20 @@ named!(string_escape_numeric<&str, char>, map_opt!(alt!( ), ::std::char::from_u32)); fn string_lit_inner(input: &str) -> nom::IResult<&str, String> { - use nom::Err; + use nom::IResult::*;; use nom::ErrorKind; let mut s = String::new(); let mut cs = input.char_indices().peekable(); while let Some((i, c)) = cs.next() { match c { - '"' => return Ok((&input[i..], s)), + '"' => return nom::IResult::Done(&input[i..], s), '\\' => match cs.next() { Some((_, s)) if s.is_whitespace() => { while cs.peek().map(|&(_, s)| s.is_whitespace()) == Some(true) { let _ = cs.next(); } if cs.next().map(|p| p.1) != Some('\\') { - return Err(Err::Error(error_position!(input, ErrorKind::Custom(4 /* FIXME */)))); + return Error(error_position!(ErrorKind::Custom(4 /* FIXME */), input)); } } Some((j, ec)) => { @@ -165,23 +168,24 @@ fn string_lit_inner(input: &str) -> nom::IResult<&str, String> { // FIXME Named ASCII escapes and control character escapes } else { match string_escape_numeric(&input[j..]) { - Ok((rest, esc)) => { + Done(rest, esc) => { let &(k, _) = cs.peek().unwrap(); // digits are always single byte ASCII characters let consumed = input[k..].len() - rest.len(); for _ in 0..consumed { let _ = cs.next(); } s.push(esc); } - Err(e) => return Err(e), + Incomplete(s) => return Incomplete(s), + Error(e) => return Error(e), } } }, - _ => return Err(Err::Error(error_position!(input, ErrorKind::Custom(5 /* FIXME */)))), + _ => return Error(error_position!(ErrorKind::Custom(5 /* FIXME */), input)), }, _ => s.push(c), } } - Err(Err::Error(error_position!(input, ErrorKind::Custom(3 /* FIXME */)))) + Error(error_position!(ErrorKind::Custom(3 /* FIXME */), input)) } named!(string_lit<&str, String>, delimited!(tag!("\""), string_lit_inner, tag!("\""))); @@ -318,26 +322,30 @@ impl<'input> Lexer<'input> { } impl<'input> Iterator for Lexer<'input> { - type Item = Spanned<Tok<'input>, usize, LexicalError<'input>>; + type Item = Spanned<Tok<'input>, usize, LexicalError>; fn next(&mut self) -> Option<Self::Item> { + use nom::IResult::*; self.skip_comments_and_whitespace(); let input = self.current_input(); if input.is_empty() { return None; } match token(input) { - Ok((rest, t)) => { + Done(rest, t) => { let parsed_len = input.len() - rest.len(); //println!("parsed {} bytes => {:?}", parsed_len, t); let start = self.offset; self.offset += parsed_len; Some(Ok((start, t, self.offset))) } - Err(e) => { + Error(e) => { let offset = self.offset; - // self.offset = self.input.len(); - Some(Err(LexicalError(offset, e))) + self.offset = self.input.len(); + Some(Err(LexicalError::Error(offset, e))) + } + Incomplete(needed) => { + Some(Err(LexicalError::Incomplete(needed))) } } } @@ -346,12 +354,6 @@ impl<'input> Iterator for Lexer<'input> { #[test] fn test_lex() { use self::Tok::*; - let s = "22"; - let expected = [Lambda, Bool(false)]; - let lexer = Lexer::new(s); - let tokens = lexer.map(|r| r.unwrap().1).collect::<Vec<_>>(); - assert_eq!(&tokens, &expected); - let s = "λ(b : Bool) → b == False"; let expected = [Lambda, ParenL, @@ -367,9 +369,9 @@ fn test_lex() { let tokens = lexer.map(|r| r.unwrap().1).collect::<Vec<_>>(); assert_eq!(&tokens, &expected); - assert_eq!(string_lit(r#""a\&b""#), Ok(("", "ab".to_owned()))); - assert_eq!(string_lit(r#""a\ \b""#), Ok(("", "ab".to_owned()))); + assert_eq!(string_lit(r#""a\&b""#).to_result(), Ok("ab".to_owned())); + assert_eq!(string_lit(r#""a\ \b""#).to_result(), Ok("ab".to_owned())); assert!(string_lit(r#""a\ b""#).is_err()); - assert_eq!(string_lit(r#""a\nb""#), Ok(("", "a\nb".to_owned()))); - assert_eq!(string_lit(r#""\o141\x62\99""#), Ok(("", "abd".to_owned()))); + assert_eq!(string_lit(r#""a\nb""#).to_result(), Ok("a\nb".to_owned())); + assert_eq!(string_lit(r#""\o141\x62\99""#).to_result(), Ok("abc".to_owned())); } diff --git a/src/main.rs b/src/main.rs index 2a1eb11..fde3978 100644 --- a/src/main.rs +++ b/src/main.rs @@ -71,7 +71,7 @@ fn main() { io::stdin().read_to_string(&mut buffer).unwrap(); let expr = match parser::parse_expr(&buffer) { Ok(e) => e, - Err(lalrpop_util::ParseError::User { error: lexer::LexicalError(pos, e) }) => { + Err(lalrpop_util::ParseError::User { error: lexer::LexicalError::Error(pos, e) }) => { print_error(&format!("Unexpected token {:?}", e), &buffer, pos, pos); return; } diff --git a/src/parser.rs b/src/parser.rs index 1127d86..3ad7ffc 100644 --- a/src/parser.rs +++ b/src/parser.rs @@ -4,7 +4,7 @@ use grammar; use grammar_util::BoxExpr; use lexer::{Lexer, LexicalError, Tok}; -pub type ParseError<'i> = lalrpop_util::ParseError<usize, Tok<'i>, LexicalError<'i>>; +pub type ParseError<'i> = lalrpop_util::ParseError<usize, Tok<'i>, LexicalError>; pub fn parse_expr(s: &str) -> Result<BoxExpr, ParseError> { grammar::ExprParser::new().parse(Lexer::new(s)) |