diff options
author | Alex Crichton <alex@alexcrichton.com> | 2018-05-01 08:49:05 -0500 |
---|---|---|
committer | GitHub <noreply@github.com> | 2018-05-01 08:49:05 -0500 |
commit | 44bec613697e884df64634536554c611d3135b3b (patch) | |
tree | e4027d86da2ea2cdfe74112a249cfad7d4b32d31 | |
parent | ffb1b5e0616c6d5aad95752522c94550db8b2967 (diff) | |
parent | 119674e135c5c4993540680263aded9eebad6fc9 (diff) | |
download | milf-rs-44bec613697e884df64634536554c611d3135b3b.tar.gz milf-rs-44bec613697e884df64634536554c611d3135b3b.zip |
Merge pull request #237 from udoprog/spans
Make Tokenizer emit spans
-rw-r--r-- | src/de.rs | 8 | ||||
-rw-r--r-- | src/tokens.rs | 177 |
2 files changed, 110 insertions, 75 deletions
@@ -1025,7 +1025,7 @@ impl<'a> Deserializer<'a> { } fn table_key(&mut self) -> Result<Cow<'a, str>, Error> { - self.tokens.table_key().map_err(|e| self.token_error(e)) + self.tokens.table_key().map(|t| t.1).map_err(|e| self.token_error(e)) } fn eat_whitespace(&mut self) -> Result<(), Error> { @@ -1049,11 +1049,11 @@ impl<'a> Deserializer<'a> { } fn next(&mut self) -> Result<Option<Token<'a>>, Error> { - self.tokens.next().map_err(|e| self.token_error(e)) + self.tokens.next().map(|t| t.map(|t| t.1)).map_err(|e| self.token_error(e)) } fn peek(&mut self) -> Result<Option<Token<'a>>, Error> { - self.tokens.peek().map_err(|e| self.token_error(e)) + self.tokens.peek().map(|t| t.map(|t| t.1)).map_err(|e| self.token_error(e)) } fn eof(&self) -> Error { @@ -1281,7 +1281,7 @@ impl<'a> Header<'a> { if self.first || self.tokens.eat(Token::Period)? { self.first = false; self.tokens.eat_whitespace()?; - self.tokens.table_key().map(Some) + self.tokens.table_key().map(|t| t.1).map(Some) } else { self.tokens.expect(Token::RightBracket)?; if self.array { diff --git a/src/tokens.rs b/src/tokens.rs index d323089..bcabd94 100644 --- a/src/tokens.rs +++ b/src/tokens.rs @@ -5,6 +5,21 @@ use std::string; use self::Token::*; +/// A span, designating a range of bytes where a token is located. +#[derive(Eq, PartialEq, Debug, Clone, Copy)] +pub struct Span { + /// The start of the range. + pub start: usize, + /// The end of the range (exclusive). + pub end: usize, +} + +impl From<Span> for (usize, usize) { + fn from(Span { start, end }: Span) -> (usize, usize) { + (start, end) + } +} + #[derive(Eq, PartialEq, Debug)] pub enum Token<'a> { Whitespace(&'a str), @@ -69,38 +84,42 @@ impl<'a> Tokenizer<'a> { t } - pub fn next(&mut self) -> Result<Option<Token<'a>>, Error> { - let token = match self.chars.next() { - Some((_, '\n')) => Newline, - Some((start, ' ')) => self.whitespace_token(start), - Some((start, '\t')) => self.whitespace_token(start), - Some((start, '#')) => self.comment_token(start), - Some((_, '=')) => Equals, - Some((_, '.')) => Period, - Some((_, ',')) => Comma, - Some((_, ':')) => Colon, - Some((_, '+')) => Plus, - Some((_, '{')) => LeftBrace, - Some((_, '}')) => RightBrace, - Some((_, '[')) => LeftBracket, - Some((_, ']')) => RightBracket, - Some((start, '\'')) => return self.literal_string(start).map(Some), - Some((start, '"')) => return self.basic_string(start).map(Some), - Some((start, ch)) if is_keylike(ch) => self.keylike(start), + pub fn next(&mut self) -> Result<Option<(Span, Token<'a>)>, Error> { + let (start, token) = match self.one() { + Some((start, '\n')) => (start, Newline), + Some((start, ' ')) => (start, self.whitespace_token(start)), + Some((start, '\t')) => (start, self.whitespace_token(start)), + Some((start, '#')) => (start, self.comment_token(start)), + Some((start, '=')) => (start, Equals), + Some((start, '.')) => (start, Period), + Some((start, ',')) => (start, Comma), + Some((start, ':')) => (start, Colon), + Some((start, '+')) => (start, Plus), + Some((start, '{')) => (start, LeftBrace), + Some((start, '}')) => (start, RightBrace), + Some((start, '[')) => (start, LeftBracket), + Some((start, ']')) => (start, RightBracket), + Some((start, '\'')) => return self.literal_string(start) + .map(|t| Some((self.step_span(start), t))), + Some((start, '"')) => return self.basic_string(start) + .map(|t| Some((self.step_span(start), t))), + Some((start, ch)) if is_keylike(ch) => (start, self.keylike(start)), Some((start, ch)) => return Err(Error::Unexpected(start, ch)), None => return Ok(None), }; - Ok(Some(token)) + + let span = self.step_span(start); + Ok(Some((span, token))) } - pub fn peek(&mut self) -> Result<Option<Token<'a>>, Error> { + pub fn peek(&mut self) -> Result<Option<(Span, Token<'a>)>, Error> { self.clone().next() } pub fn eat(&mut self, expected: Token<'a>) -> Result<bool, Error> { match self.peek()? { - Some(ref found) if expected == *found => {} + Some((_, ref found)) if expected == *found => {} Some(_) => return Ok(false), None => return Ok(false), } @@ -111,7 +130,7 @@ impl<'a> Tokenizer<'a> { pub fn expect(&mut self, expected: Token<'a>) -> Result<(), Error> { let current = self.current(); match self.next()? { - Some(found) => { + Some((_, found)) => { if expected == found { Ok(()) } else { @@ -132,21 +151,21 @@ impl<'a> Tokenizer<'a> { } } - pub fn table_key(&mut self) -> Result<Cow<'a, str>, Error> { + pub fn table_key(&mut self) -> Result<(Span, Cow<'a, str>), Error> { let current = self.current(); match self.next()? { - Some(Token::Keylike(k)) => Ok(k.into()), - Some(Token::String { src, val }) => { + Some((span, Token::Keylike(k))) => Ok((span, k.into())), + Some((span, Token::String { src, val })) => { let offset = self.substr_offset(src); if val == "" { return Err(Error::EmptyTableKey(offset)) } match src.find('\n') { - None => Ok(val), + None => Ok((span, val)), Some(i) => Err(Error::NewlineInTableKey(offset + i)), } } - Some(other) => { + Some((_, other)) => { Err(Error::Wanted { at: current, expected: "a table key", @@ -182,8 +201,8 @@ impl<'a> Tokenizer<'a> { let current = self.current(); match self.next()? { None | - Some(Token::Newline) => Ok(()), - Some(other) => { + Some((_, Token::Newline)) => Ok(()), + Some((_, other)) => { Err(Error::Wanted { at: current, expected: "newline", @@ -195,7 +214,7 @@ impl<'a> Tokenizer<'a> { pub fn skip_to_newline(&mut self) { loop { - match self.chars.next() { + match self.one() { Some((_, '\n')) | None => break, _ => {} @@ -206,7 +225,7 @@ impl<'a> Tokenizer<'a> { fn eatc(&mut self, ch: char) -> bool { match self.chars.clone().next() { Some((_, ch2)) if ch == ch2 => { - self.chars.next(); + self.one(); true } _ => false, @@ -233,7 +252,7 @@ impl<'a> Tokenizer<'a> { if ch != '\t' && (ch < '\u{20}' || ch > '\u{10ffff}') { break } - self.chars.next(); + self.one(); } Comment(&self.input[start..self.current()]) } @@ -260,7 +279,7 @@ impl<'a> Tokenizer<'a> { let mut n = 0; 'outer: loop { n += 1; - match self.chars.next() { + match self.one() { Some((i, '\n')) => { if multiline { if self.input.as_bytes()[i] == b'\r' { @@ -352,7 +371,7 @@ impl<'a> Tokenizer<'a> { fn hex(&mut self, start: usize, i: usize, len: usize) -> Result<char, Error> { let mut val = 0; for _ in 0..len { - match self.chars.next() { + match self.one() { Some((_, ch)) if '0' <= ch && ch <= '9' => { val = val * 16 + (ch as u32 - '0' as u32); } @@ -370,11 +389,11 @@ impl<'a> Tokenizer<'a> { } fn keylike(&mut self, start: usize) -> Token<'a> { - while let Some((_, ch)) = self.chars.clone().next() { + while let Some((_, ch)) = self.peek_one() { if !is_keylike(ch) { break } - self.chars.next(); + self.one(); } Keylike(&self.input[start..self.current()]) } @@ -386,6 +405,22 @@ impl<'a> Tokenizer<'a> { assert!(a <= b); b - a } + + /// Calculate the span of a single character. + fn step_span(&mut self, start: usize) -> Span { + let end = self.peek_one().map(|t| t.0).unwrap_or_else(|| self.input.len()); + Span { start: start, end: end } + } + + /// Peek one char without consuming it. + fn peek_one(&mut self) -> Option<(usize, char)> { + self.chars.clone().next() + } + + /// Take one char. + pub fn one(&mut self) -> Option<(usize, char)> { + self.chars.next() + } } impl<'a> Iterator for CrlfFold<'a> { @@ -475,7 +510,7 @@ mod tests { fn literal_strings() { fn t(input: &str, val: &str) { let mut t = Tokenizer::new(input); - let token = t.next().unwrap().unwrap(); + let (_, token) = t.next().unwrap().unwrap(); assert_eq!(token, Token::String { src: input, val: Cow::Borrowed(val), @@ -497,7 +532,7 @@ mod tests { fn basic_strings() { fn t(input: &str, val: &str) { let mut t = Tokenizer::new(input); - let token = t.next().unwrap().unwrap(); + let (_, token) = t.next().unwrap().unwrap(); assert_eq!(token, Token::String { src: input, val: Cow::Borrowed(val), @@ -538,7 +573,7 @@ mod tests { fn keylike() { fn t(input: &str) { let mut t = Tokenizer::new(input); - let token = t.next().unwrap().unwrap(); + let (_, token) = t.next().unwrap().unwrap(); assert_eq!(token, Token::Keylike(input)); assert!(t.next().unwrap().is_none()); } @@ -554,11 +589,11 @@ mod tests { #[test] fn all() { - fn t(input: &str, expected: &[Token]) { + fn t(input: &str, expected: &[((usize, usize), Token, &str)]) { let mut tokens = Tokenizer::new(input); - let mut actual = Vec::new(); - while let Some(token) = tokens.next().unwrap() { - actual.push(token); + let mut actual: Vec<((usize, usize), Token, &str)> = Vec::new(); + while let Some((span, token)) = tokens.next().unwrap() { + actual.push((span.into(), token, &input[span.start..span.end])); } for (a, b) in actual.iter().zip(expected) { assert_eq!(a, b); @@ -567,37 +602,37 @@ mod tests { } t(" a ", &[ - Token::Whitespace(" "), - Token::Keylike("a"), - Token::Whitespace(" "), + ((0, 1), Token::Whitespace(" "), " "), + ((1, 2), Token::Keylike("a"), "a"), + ((2, 3), Token::Whitespace(" "), " "), ]); t(" a\t [[]] \t [] {} , . =\n# foo \r\n#foo \n ", &[ - Token::Whitespace(" "), - Token::Keylike("a"), - Token::Whitespace("\t "), - Token::LeftBracket, - Token::LeftBracket, - Token::RightBracket, - Token::RightBracket, - Token::Whitespace(" \t "), - Token::LeftBracket, - Token::RightBracket, - Token::Whitespace(" "), - Token::LeftBrace, - Token::RightBrace, - Token::Whitespace(" "), - Token::Comma, - Token::Whitespace(" "), - Token::Period, - Token::Whitespace(" "), - Token::Equals, - Token::Newline, - Token::Comment("# foo "), - Token::Newline, - Token::Comment("#foo "), - Token::Newline, - Token::Whitespace(" "), + ((0, 1), Token::Whitespace(" "), " "), + ((1, 2), Token::Keylike("a"), "a"), + ((2, 4), Token::Whitespace("\t "), "\t "), + ((4, 5), Token::LeftBracket, "["), + ((5, 6), Token::LeftBracket, "["), + ((6, 7), Token::RightBracket, "]"), + ((7, 8), Token::RightBracket, "]"), + ((8, 11), Token::Whitespace(" \t "), " \t "), + ((11, 12), Token::LeftBracket, "["), + ((12, 13), Token::RightBracket, "]"), + ((13, 14), Token::Whitespace(" "), " "), + ((14, 15), Token::LeftBrace, "{"), + ((15, 16), Token::RightBrace, "}"), + ((16, 17), Token::Whitespace(" "), " "), + ((17, 18), Token::Comma, ","), + ((18, 19), Token::Whitespace(" "), " "), + ((19, 20), Token::Period, "."), + ((20, 21), Token::Whitespace(" "), " "), + ((21, 22), Token::Equals, "="), + ((22, 23), Token::Newline, "\n"), + ((23, 29), Token::Comment("# foo "), "# foo "), + ((29, 31), Token::Newline, "\r\n"), + ((31, 36), Token::Comment("#foo "), "#foo "), + ((36, 37), Token::Newline, "\n"), + ((37, 38), Token::Whitespace(" "), " "), ]); } |