From 4601a568679836062cef0a6bfcd0134c437c3d7d Mon Sep 17 00:00:00 2001 From: greg Date: Fri, 8 Jan 2016 23:55:58 -0800 Subject: [PATCH] Start working on tokenization tests --- src/tokenizer.rs | 8 -------- 1 file changed, 8 deletions(-) diff --git a/src/tokenizer.rs b/src/tokenizer.rs index e0c3614..fae8a73 100644 --- a/src/tokenizer.rs +++ b/src/tokenizer.rs @@ -127,13 +127,5 @@ mod tests { #[test] fn tokeniziation_tests() { let t1 = "let a = 3\n"; - assert_eq!(format!("{:?}", tokenize(t1)), - "[Keyword(Let), Identifier(\"a\"), Keyword(Assign), NumLiteral(3), Newline, EOF]"); - - // this is intentional - let t2 = "a + b*c\n"; - assert_eq!(format!("{:?}", tokenize(t2)), - "[Identifier(\"a\"), Identifier(\"+\"), Identifier(\"b*c\"), Newline, EOF]"); - } }