sibil/src/lexer/mod.rs

264 lines
6.9 KiB
Rust

/* lexer.rs
* Eryn Wells <eryn@erynwells.me>
*/
pub mod token;
mod char;
mod charset;
mod str;
use self::char::Lexable;
use self::str::CharAt;
use self::str::RelativeIndexable;
use self::token::Lex;
use self::token::Token;
#[derive(Debug)]
enum State {
Initial,
Identifier,
Hash,
Comment,
String,
}
pub struct Lexer {
input: String,
begin: usize,
forward: usize,
line: u32,
state: State,
}
impl Lexer {
pub fn new(input: &str) -> Lexer {
Lexer {
input: String::from(input),
begin: 0,
forward: 0,
line: 1,
state: State::Initial,
}
}
}
impl Lexer {
fn begin_lexing(&mut self) {
self.forward = self.begin;
self.state = State::Initial;
}
/// Advance the forward pointer to the next character.
fn advance(&mut self) {
self.forward = self.input.index_after(self.forward);
println!("> forward={}", self.forward);
}
/// Retract the forward pointer to the previous character.
fn retract(&mut self) {
self.forward = self.input.index_before(self.forward);
println!("< forward={}", self.forward);
}
/// Advance the begin pointer to prepare for the next iteration.
fn advance_begin(&mut self) {
self.begin = self.input.index_after(self.forward);
self.forward = self.begin;
println!("> begin={}, forward={}", self.begin, self.forward);
}
fn handle_newline(&mut self) {
self.line += 1;
}
/// Get the substring between the two input indexes. This is the value to give to a new Token instance.
fn value(&self) -> String {
self.input[self.begin .. self.forward].to_string()
}
}
impl Lexer {
// TODO: Use std::result::Result for these state_* methods.
// https://doc.rust-lang.org/1.14.0/core/result/enum.Result.html
/// Handle self.state == State::Initial
fn state_initial(&mut self, c: char, token: &mut Option<Token>) {
if c.is_left_paren() {
*token = Some(Token::LeftParen(c.to_string()));
}
else if c.is_right_paren() {
*token = Some(Token::RightParen(c.to_string()));
}
else if c.is_dot() {
*token = Some(Token::Dot);
}
else if c.is_hash() {
self.state = State::Hash;
self.advance();
}
else if c.is_string_quote() {
self.state = State::String;
self.advance();
}
else if c.is_identifier_single() {
*token = Some(Token::Identifier(c.to_string()));
}
else if c.is_identifier_initial() {
self.state = State::Identifier;
self.advance();
}
else if c.is_whitespace() {
if c.is_newline() {
self.handle_newline();
}
self.advance_begin();
}
else if c.is_comment_initial() {
self.state = State::Comment;
self.advance();
}
else {
assert!(false, "Invalid token character: {}", c);
}
}
/// Handle self.state == State::Identifier
fn state_identifier(&mut self, c: char, token: &mut Option<Token>) {
if c.is_identifier_subsequent() {
// State in Identifier state.
self.advance();
}
else if c.is_identifier_delimiter() {
*token = Some(Token::Identifier(self.value()));
self.retract();
}
else {
assert!(false, "Invalid token character: '{}'", c);
}
}
fn state_hash(&mut self, c: char, token: &mut Option<Token>) {
if c.is_boolean_true() || c.is_boolean_false() {
self.advance();
*token = Some(Token::Boolean(c.is_boolean_true()));
}
else if c.is_left_paren() {
self.advance();
*token = Some(Token::LeftVectorParen);
}
else {
assert!(false, "Invalid token character: '{}'", c);
}
}
fn state_string(&mut self, c: char, token: &mut Option<Token>) {
self.advance();
if c.is_string_quote() {
*token = Some(Token::String(self.value()));
}
}
fn state_comment(&mut self, c: char, token: &mut Option<Token>) {
if c.is_newline() {
self.handle_newline();
*token = Some(Token::Comment(self.value()));
}
else if c.is_eof() {
*token = Some(Token::Comment(self.value()));
}
// Consume all characters.
self.advance();
}
}
impl Iterator for Lexer {
type Item = Lex;
fn next(&mut self) -> Option<Lex> {
self.begin_lexing();
if self.begin == self.input.len() {
return None;
}
let mut token: Option<Token> = None;
println!("Lexing '{}'", &self.input[self.begin ..]);
while token.is_none() {
let c = match self.input.char_at(self.forward) {
Some(c) => c,
None => '\0',
};
println!("{:?}! c='{}'", self.state, c);
match self.state {
State::Initial => self.state_initial(c, &mut token),
State::Identifier => self.state_identifier(c, &mut token),
State::Hash => self.state_hash(c, &mut token),
State::String => self.state_string(c, &mut token),
State::Comment => self.state_comment(c, &mut token),
}
}
self.advance_begin();
match token {
Some(t) => Some(Lex::new(t)),
None => None,
}
}
}
//
// UNIT TESTING
//
#[cfg(test)]
mod tests {
use std::iter::Iterator;
use super::*;
use super::token::*;
#[test]
fn lexer_finds_parens() {
check_single_token("(", Token::LeftParen(String::from("(")));
check_single_token(")", Token::RightParen(String::from(")")));
check_single_token("#(", Token::LeftVectorParen);
}
#[test]
fn lexer_finds_dots() {
check_single_token(".", Token::Dot);
}
#[test]
fn lexer_finds_identifiers() {
check_single_token("abc", Token::Identifier(String::from("abc")));
}
#[test]
fn lexer_finds_booleans() {
check_single_token("#t", Token::Boolean(true));
check_single_token("#f", Token::Boolean(false));
}
#[test]
fn lexer_finds_comments() {
let s = "; a comment";
check_single_token(s, Token::Comment(String::from(s)));
}
#[test]
fn lexer_finds_strings() {
check_single_token("\"\"", Token::String(String::from("\"\"")));
check_single_token("\"abc\"", Token::String(String::from("\"abc\"")));
}
fn check_single_token(input: &str, expected: Token) {
let mut lexer = Lexer::new(input);
assert_next_token(&mut lexer, &expected);
}
fn assert_next_token(lexer: &mut Lexer, expected: &Token) {
let lex = lexer.next().unwrap();
assert_eq!(lex.token, *expected);
}
}