extern crate itertools; use std::collections::HashMap; use std::rc::Rc; use std::iter::{Enumerate, Peekable}; use self::itertools::Itertools; use std::vec::IntoIter; use std::str::Chars; #[derive(Debug, PartialEq, Clone)] pub enum TokenType { Newline, Semicolon, LParen, RParen, LSquareBracket, RSquareBracket, LAngleBracket, RAngleBracket, LCurlyBrace, RCurlyBrace, Pipe, Comma, Period, Colon, Underscore, Operator(Rc), DigitGroup(Rc), HexNumberSigil, BinNumberSigil, StrLiteral(Rc), Identifier(Rc), Keyword(Kw), EOF, Error(String), } use self::TokenType::*; #[derive(Debug, Clone, Copy, PartialEq)] pub enum Kw { If, Else, Func, For, Match, Var, Const, Let, In, Alias, Type, SelfType, SelfIdent, Trait, Impl, True, False } use self::Kw::*; lazy_static! { static ref KEYWORDS: HashMap<&'static str, Kw> = hashmap! { "if" => Kw::If, "else" => Kw::Else, "fn" => Kw::Func, "for" => Kw::For, "match" => Kw::Match, "var" => Kw::Var, "const" => Kw::Const, "let" => Kw::Let, "in" => Kw::In, "alias" => Kw::Alias, "type" => Kw::Type, "Self" => Kw::SelfType, "self" => Kw::SelfIdent, "trait" => Kw::Trait, "impl" => Kw::Impl, "true" => Kw::True, "false" => Kw::False, }; } #[derive(Debug)] pub struct Token { token_type: TokenType, offset: usize, } impl Token { pub fn get_error(&self) -> Option<&String> { match self.token_type { TokenType::Error(ref s) => Some(s), _ => None, } } } fn is_digit(c: &char) -> bool { c.is_digit(10) } type CharIter<'a> = Peekable>>; pub fn tokenize(input: &str) -> Vec { let mut tokens: Vec = Vec::new(); let mut input: CharIter = input.chars().enumerate().peekable(); while let Some((idx, c)) = input.next() { let cur_tok_type = match c { '#' => { if let Some(&(_, '{')) = input.peek() { } else { while let Some((_, c)) = input.next() { if c == '\n' { break; } } } continue; }, c if c.is_whitespace() && c != '\n' => continue, '\n' => Newline, ';' => Semicolon, ':' => Colon, ',' => Comma, '.' => Period, '(' => LParen, ')' => RParen, '{' => LCurlyBrace, '}' => RCurlyBrace, '<' => LAngleBracket, '>' => RAngleBracket, '[' => LSquareBracket, ']' => RSquareBracket, '|' => Pipe, '"' => handle_quote(&mut input), c if is_digit(&c) => handle_digit(c, &mut input), c if c.is_alphabetic() || c == '_' => handle_alphabetic(c, &mut input), //TODO I'll probably have to rewrite this if I care about types being uppercase, also type parameterization c => handle_operator(c, &mut input), }; tokens.push(Token { token_type: cur_tok_type, offset: idx }); } tokens } fn handle_digit(c: char, input: &mut CharIter) -> TokenType { if c == '0' && input.peek().map_or(false, |&(_, c)| { c == 'x' }) { input.next(); HexNumberSigil } else if c == '0' && input.peek().map_or(false, |&(_, c)| { c == 'b' }) { input.next(); BinNumberSigil } else { let mut buf = c.to_string(); buf.extend(input.peeking_take_while(|&(_, ref c)| is_digit(c)).map(|(_, c)| { c })); DigitGroup(Rc::new(buf)) } } fn handle_quote(input: &mut CharIter) -> TokenType { let mut buf = String::new(); loop { match input.next().map(|(_, c)| { c }) { Some('"') => break, Some('\\') => { let next = input.peek().map(|&(_, c)| { c }); if next == Some('n') { input.next(); buf.push('\n') } else if next == Some('"') { input.next(); buf.push('"'); } else if next == Some('t') { input.next(); buf.push('\t'); } }, Some(c) => buf.push(c), None => return TokenType::Error(format!("Unclosed string")), } } TokenType::StrLiteral(Rc::new(buf)) } fn handle_alphabetic(c: char, input: &mut CharIter) -> TokenType { let mut buf = String::new(); buf.push(c); if c == '_' && input.peek().map(|&(_, c)| { !c.is_alphabetic() }).unwrap_or(true) { return TokenType::Underscore } loop { match input.peek().map(|&(_, c)| { c }) { Some(c) if c.is_alphanumeric() => { input.next(); buf.push(c); }, _ => break, } } match KEYWORDS.get(buf.as_str()) { Some(kw) => TokenType::Keyword(kw.clone()), None => TokenType::Identifier(Rc::new(buf)), } } fn handle_operator(c: char, input: &mut CharIter) -> TokenType { let mut buf = String::new(); buf.push(c); loop { match input.peek().map(|&(_, c)| { c }) { Some(c) if !c.is_alphabetic() && !c.is_whitespace() => { input.next(); buf.push(c); }, _ => break } } TokenType::Operator(Rc::new(buf)) } #[cfg(test)] mod schala_tokenizer_tests { use super::*; use super::TokenType::*; use super::Kw::*; macro_rules! digit { ($ident:expr) => { DigitGroup(Rc::new($ident.to_string())) } } macro_rules! ident { ($ident:expr) => { Identifier(Rc::new($ident.to_string())) } } macro_rules! op { ($ident:expr) => { Operator(Rc::new($ident.to_string())) } } #[test] fn tokens() { let a = tokenize("let a: A = c ++ d"); let token_types: Vec = a.into_iter().map(move |t| t.token_type).collect(); assert_eq!(token_types, vec![Keyword(Let), ident!("a"), Colon, ident!("A"), LAngleBracket, ident!("B"), RAngleBracket, op!("="), ident!("c"), op!("++"), ident!("d")]); } #[test] fn underscores() { let token_types: Vec = tokenize("4_8").into_iter().map(move |t| t.token_type).collect(); assert_eq!(token_types, vec![digit!("4"), Underscore, digit!("8")]); } } /* Schala (PROVISIONAL!!) EBNF grammar '' = literal, all other symbols are nonterminals program := (statement delimiter ?)* delimiter := 'Newline' | ';' statement := declaration | expression declaration := module | function | type_decl type_decl := 'type' type_format type_format := 'alias' '=' type | type_constructor type_constructor := capital_ident '=' type_rhs type_rhs := struct_decl | type_variant ('|' type_variant)* struct_decl := 'struct' '{' (ident ':' type)* '}' type_variant := capital_ident | tuple_type | capital_ident struct_decl tuple_type := // something like Variant(a,b) type := // something like Type[A[b]] ascription := expression (':' type)+ function := 'fn' prototype '{' (statement)* '}' prototype := identifier '(' identlist ')' identlist := identifier (',' identifier)* | ε declaration := FN prototype LCurlyBrace (statement)* RCurlyBrace prototype := identifier LParen identlist RParen identlist := Ident (Comma Ident)* | ε exprlist := Expression (Comma Expression)* | ε itemlist := Ident COLON Expression (Comma Ident COLON Expression)* | ε expression := postop_expression (op postop_expression)* postop_expression := primary_expression postop primary_expression := number_expr | String | identifier_expr | paren_expr | conditional_expr | while_expr | lambda_expr | list_expr | struct_expr number_expr := (PLUS | MINUS ) number_expr | Number identifier_expr := call_expression | Variable list_expr := LSquareBracket exprlist RSquareBracket struct_expr := LCurlyBrace itemlist RCurlyBrace call_expression := Identifier LParen exprlist RParen while_expr := WHILE primary_expression LCurlyBrace (expression delimiter)* RCurlyBrace paren_expr := LParen expression RParen conditional_expr := IF expression LCurlyBrace (expression delimiter)* RCurlyBrace (LCurlyBrace (expresion delimiter)* RCurlyBrace)? lambda_expr := FN LParen identlist RParen LCurlyBrace (expression delimiter)* RCurlyBrace lambda_call := | LParen exprlist RParen postop := ε | LParen exprlist RParen | LBracket expression RBracket op := '+', '-', etc. */ /* Schala EBNF Grammar */ /* program := (statement delimiter)* EOF delimiter := NEWLINE | SEMICOLON statement := expression | declaration declaration := type_declaration | func_declaration type_declaration := TYPE identifier func_declaration := FN identifier LParen param_list RParen param_list := (identifier type_anno+ Comma)* type_anno := Colon type expression := primary primary := literal literal := TRUE | FALSE | number_literal | str_literal identifier := IDENTIFIER // a float_literal can still be assigned to an int in type-checking number_literal := int_literal | float_literal int_literal = (HEX_SIGIL | BIN_SIGIL) digits float_literal := digits (PERIOD digits) digits := (digit_group underscore)+ */ type TokenIter = Peekable>; #[derive(Debug)] pub struct ParseError { pub msg: String, } impl ParseError { fn new(msg: &str) -> ParseResult { Err(ParseError { msg: msg.to_string() }) } } pub type ParseResult = Result; struct Parser { tokens: TokenIter, } impl Parser { fn new(input: Vec) -> Parser { Parser { tokens: input.into_iter().peekable() } } fn peek(&mut self) -> TokenType { self.tokens.peek().map(|ref t| { t.token_type.clone() }).unwrap_or(TokenType::EOF) } fn next(&mut self) -> TokenType { self.tokens.next().map(|ref t| { t.token_type.clone() }).unwrap_or(TokenType::EOF) } } macro_rules! expect { ($self:expr, $token_type:pat, $message:expr) => { match $self.peek() { $token_type => $self.next(), _ => return Err(ParseError { msg: $message.to_string() }), } } } #[derive(Debug, PartialEq)] pub struct AST(Vec); #[derive(Debug, PartialEq)] pub enum Statement { Expression(Expression), Declaration(Declaration), } #[derive(Debug, PartialEq)] pub enum Declaration { FuncDecl, TypeDecl(Rc, TypeBody) } #[derive(Debug, PartialEq)] pub enum TypeBody { TypeBody } #[derive(Debug, PartialEq)] pub enum Expression { IntLiteral(u64), FloatLiteral(f64), } impl Parser { fn program(&mut self) -> ParseResult { let mut statements = Vec::new(); loop { match self.peek() { EOF => break, Newline | Semicolon => { self.next(); continue; }, _ => statements.push(self.statement()?), } } Ok(AST(statements)) } fn statement(&mut self) -> ParseResult { //TODO handle error recovery here match self.peek() { Keyword(Type) => self.type_declaration().map(|decl| { Statement::Declaration(decl) }), Keyword(Func)=> self.func_declaration().map(|func| { Statement::Declaration(func) }), _ => self.expression().map(|expr| { Statement::Expression(expr) } ), } } fn type_declaration(&mut self) -> ParseResult { expect!(self, Keyword(Type), "Expected 'type'"); let name = self.identifier()?; Ok(Declaration::TypeDecl(name, TypeBody::TypeBody)) } fn func_declaration(&mut self) -> ParseResult { expect!(self, Keyword(Func), "Expected 'fn'"); let name = self.identifier()?; expect!(self, LParen, "Expected '('"); let params = self.param_list(); expect!(self, RParen, "Expected ')'"); Ok(Declaration::FuncDecl) } fn param_list(&mut self) -> ParseResult>> { Ok(vec!()) } fn expression(&mut self) -> ParseResult { self.primary() } fn primary(&mut self) -> ParseResult { self.literal() } fn identifier(&mut self) -> ParseResult> { match self.next() { Identifier(s) => Ok(s), p => ParseError::new(&format!("Expected an identifier, got {:?}", p)), } } fn literal(&mut self) -> ParseResult { match self.peek() { DigitGroup(_) | HexNumberSigil | BinNumberSigil | Period => self.number_literal(), _ => unimplemented!(), } } fn number_literal(&mut self) -> ParseResult { match self.peek() { HexNumberSigil | BinNumberSigil => self.int_literal(), _ => self.float_literal(), } } fn int_literal(&mut self) -> ParseResult { use self::Expression::*; let digits = self.digits()?; match self.next() { BinNumberSigil => { unimplemented!() }, HexNumberSigil => { unimplemented!() }, _ => return ParseError::new("Expected '0x' or '0b'"), } } fn float_literal(&mut self) -> ParseResult { use self::Expression::*; let mut digits = self.digits()?; if let TokenType::Period = self.peek() { self.next(); digits.push_str("."); digits.push_str(&self.digits()?); match digits.parse::() { Ok(f) => Ok(FloatLiteral(f)), Err(e) => unimplemented!("Float didn't parse with error: {}", e), } } else { match digits.parse::() { Ok(d) => Ok(IntLiteral(d)), Err(e) => unimplemented!("Need to handle numbers that don't parse to a Rust u64 {}", e), } } } fn digits(&mut self) -> ParseResult { let mut ds = String::new(); loop { match self.peek() { Underscore => { self.next(); continue; }, DigitGroup(ref s) => { self.next(); ds.push_str(s)}, _ => break, } } Ok(ds) } } pub fn parse(input: Vec) -> Result { let mut parser = Parser::new(input); parser.program() } #[cfg(test)] mod parse_tests { use super::*; use super::Statement::*; use super::Expression::*; use super::ParseError; #[test] fn test_parsing() { let a = "8.1"; assert_eq!(parse(tokenize(a)).unwrap(), AST(vec![Expression(FloatLiteral(8.1))])); } }