From 66cd51a3553ad5bfbb7701888a58a0c63a4b7a6a Mon Sep 17 00:00:00 2001 From: greg Date: Fri, 2 Mar 2018 21:59:14 -0800 Subject: [PATCH] Cleanup --- src/schala_lang/tokenizing.rs | 7 ------- 1 file changed, 7 deletions(-) diff --git a/src/schala_lang/tokenizing.rs b/src/schala_lang/tokenizing.rs index 0c2a6df..0c02ae2 100644 --- a/src/schala_lang/tokenizing.rs +++ b/src/schala_lang/tokenizing.rs @@ -92,18 +92,11 @@ type CharIter> = Peekable; pub fn tokenize(input: &str) -> Vec { let mut tokens: Vec = Vec::new(); - //let b = input.clone(); - - //ound type `std::iter::Peekable>, std::iter::Map>, [closure@src/schala_lang -// tokenizing.rs:99:40: 99:82 line_idx:_]>, [closure@src/schala_lang/tokenizing.rs:98:17: 100:8]>>` - let mut input = input.lines().enumerate() .flat_map(|(line_idx, ref line)| { line.chars().enumerate().map(move |(ch_idx, ch)| (line_idx, ch_idx, ch)) }).peekable(); - //let mut input: CharIter = input.chars().enumerate().peekable(); - while let Some((line_idx, ch_idx, c)) = input.next() { let cur_tok_type = match c { '#' => {