From 119674e135c5c4993540680263aded9eebad6fc9 Mon Sep 17 00:00:00 2001 From: John-John Tedro Date: Tue, 1 May 2018 13:43:02 +0200 Subject: Make Tokenizer emit spans --- src/de.rs | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) (limited to 'src/de.rs') diff --git a/src/de.rs b/src/de.rs index 9ff09f7..8677ffc 100644 --- a/src/de.rs +++ b/src/de.rs @@ -1025,7 +1025,7 @@ impl<'a> Deserializer<'a> { } fn table_key(&mut self) -> Result, Error> { - self.tokens.table_key().map_err(|e| self.token_error(e)) + self.tokens.table_key().map(|t| t.1).map_err(|e| self.token_error(e)) } fn eat_whitespace(&mut self) -> Result<(), Error> { @@ -1049,11 +1049,11 @@ impl<'a> Deserializer<'a> { } fn next(&mut self) -> Result>, Error> { - self.tokens.next().map_err(|e| self.token_error(e)) + self.tokens.next().map(|t| t.map(|t| t.1)).map_err(|e| self.token_error(e)) } fn peek(&mut self) -> Result>, Error> { - self.tokens.peek().map_err(|e| self.token_error(e)) + self.tokens.peek().map(|t| t.map(|t| t.1)).map_err(|e| self.token_error(e)) } fn eof(&self) -> Error { @@ -1281,7 +1281,7 @@ impl<'a> Header<'a> { if self.first || self.tokens.eat(Token::Period)? { self.first = false; self.tokens.eat_whitespace()?; - self.tokens.table_key().map(Some) + self.tokens.table_key().map(|t| t.1).map(Some) } else { self.tokens.expect(Token::RightBracket)?; if self.array { -- cgit v1.2.3