2022-11-28 23:33:34 +00:00
|
|
|
mod utils;
|
|
|
|
mod scanner;
|
2023-01-05 17:48:34 +00:00
|
|
|
mod lex_error;
|
2022-11-28 23:33:34 +00:00
|
|
|
use super::token::{self, Token};
|
2023-01-05 17:48:34 +00:00
|
|
|
use lex_error::LexError;
|
2022-11-28 23:33:34 +00:00
|
|
|
|
|
|
|
type Chars = Vec<char>;
|
|
|
|
|
2023-02-11 23:13:05 +00:00
|
|
|
/// Represents the result of scanning a single token from the input
|
2022-11-30 13:38:43 +00:00
|
|
|
pub enum LexResult {
|
2023-02-11 23:13:05 +00:00
|
|
|
/// A token was found. The first element is the token, and the
|
|
|
|
/// second element is the position in the input after the token.
|
|
|
|
///
|
|
|
|
/// E.g., given an input
|
|
|
|
///
|
|
|
|
/// "`identifier 55`"
|
|
|
|
///
|
|
|
|
/// scanning from a position `0`, the result would be
|
|
|
|
///
|
|
|
|
/// `Some(Token("identifier"), 10)`.
|
|
|
|
///
|
|
|
|
/// where:
|
|
|
|
/// - `Token("identifier")` is the token
|
|
|
|
/// - `10` is the position where the token ends, and from where the next token
|
|
|
|
/// should be scanned
|
2022-11-30 13:38:43 +00:00
|
|
|
Some(Token, usize),
|
2023-02-11 23:13:05 +00:00
|
|
|
/// No token was found. This indicates that EOF has been reached.
|
|
|
|
///
|
|
|
|
/// Contains the last position, which should be the input lenght - 1
|
2022-11-30 13:38:43 +00:00
|
|
|
None(usize),
|
2023-02-11 23:13:05 +00:00
|
|
|
/// An error was found while scanning.
|
2023-01-05 17:48:34 +00:00
|
|
|
Err(LexError),
|
2022-11-30 13:38:43 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2022-11-28 23:33:34 +00:00
|
|
|
/// Scans and returns all the tokens in the input String
|
2023-01-05 17:48:34 +00:00
|
|
|
pub fn get_tokens(input: &String) -> Result<Vec<Token>, LexError> {
|
2022-11-28 23:33:34 +00:00
|
|
|
let chars: Vec<char> = input.chars().into_iter().collect();
|
|
|
|
let mut results = Vec::new();
|
|
|
|
let mut current_pos: usize = 0;
|
|
|
|
|
|
|
|
while has_input(&chars, current_pos) {
|
2022-11-29 00:16:55 +00:00
|
|
|
match next_token(&chars, current_pos) {
|
2022-11-30 13:38:43 +00:00
|
|
|
LexResult::Some(token, next_pos) => {
|
2022-11-29 00:16:55 +00:00
|
|
|
results.push(token);
|
|
|
|
current_pos = next_pos;
|
|
|
|
},
|
2022-11-30 13:38:43 +00:00
|
|
|
LexResult::None(next_pos) => {
|
2022-11-29 00:16:55 +00:00
|
|
|
current_pos = next_pos;
|
|
|
|
},
|
2022-11-30 13:38:43 +00:00
|
|
|
LexResult::Err(reason) => return Err(reason),
|
2022-11-28 23:33:34 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-02-14 20:32:45 +00:00
|
|
|
results.push(token::new_semicolon(0));
|
2022-11-28 23:33:34 +00:00
|
|
|
results.push(token::new_eof(0));
|
2022-11-29 00:16:55 +00:00
|
|
|
Ok(results)
|
2022-11-28 23:33:34 +00:00
|
|
|
}
|
|
|
|
|
2023-02-11 23:13:05 +00:00
|
|
|
/// Scans a single token from `chars`, starting from `current_pos`
|
2022-11-30 13:38:43 +00:00
|
|
|
fn next_token(chars: &Chars, current_pos: usize) -> LexResult {
|
2022-11-28 23:33:34 +00:00
|
|
|
let next_char = peek(chars, current_pos);
|
|
|
|
|
2022-11-30 13:38:43 +00:00
|
|
|
// If EOF is reached return nothing but the current position
|
2022-11-29 00:16:55 +00:00
|
|
|
if next_char == '\0' {
|
2022-11-30 13:38:43 +00:00
|
|
|
return LexResult::None(current_pos)
|
2022-11-29 00:16:55 +00:00
|
|
|
}
|
|
|
|
|
2023-02-14 20:22:29 +00:00
|
|
|
// Handle whitespace recursively.
|
2022-11-28 23:33:34 +00:00
|
|
|
if next_char == ' ' {
|
|
|
|
return next_token(chars, current_pos + 1)
|
|
|
|
}
|
|
|
|
|
2022-12-01 13:33:48 +00:00
|
|
|
// Scanners
|
2022-11-30 13:38:43 +00:00
|
|
|
None
|
2022-12-01 17:53:14 +00:00
|
|
|
.or_else(|| scanner::number(next_char, chars, current_pos))
|
|
|
|
.or_else(|| scanner::identifier(next_char, chars, current_pos))
|
|
|
|
.or_else(|| scanner::string(next_char, chars, current_pos))
|
|
|
|
.or_else(|| scanner::operator(next_char, chars, current_pos))
|
|
|
|
.or_else(|| scanner::grouping_sign(next_char, chars, current_pos))
|
2023-02-14 20:22:29 +00:00
|
|
|
.or_else(|| scanner::new_line(next_char, chars, current_pos))
|
2022-11-30 13:38:43 +00:00
|
|
|
.unwrap_or_else(|| {
|
2023-01-05 17:48:34 +00:00
|
|
|
let error = LexError {
|
|
|
|
position: current_pos,
|
|
|
|
reason: format!("Unrecognized character: {}", next_char),
|
|
|
|
};
|
|
|
|
LexResult::Err(error)
|
2022-11-30 13:38:43 +00:00
|
|
|
})
|
2022-11-28 23:33:34 +00:00
|
|
|
}
|
|
|
|
|
2023-02-11 23:13:05 +00:00
|
|
|
/// Returns the char at `pos`
|
2022-11-28 23:33:34 +00:00
|
|
|
fn peek(input: &Chars, pos: usize) -> char {
|
|
|
|
let result = input.get(pos).unwrap_or(&'\0');
|
|
|
|
*result
|
|
|
|
}
|
|
|
|
|
2023-02-11 23:13:05 +00:00
|
|
|
/// Whether there is still input based on `current_pos`
|
2022-11-28 23:33:34 +00:00
|
|
|
fn has_input(input: &Chars, current_pos: usize) -> bool {
|
2022-11-29 00:16:55 +00:00
|
|
|
current_pos < input.len()
|
2022-11-28 23:33:34 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
#[cfg(test)]
|
|
|
|
mod tests {
|
|
|
|
use super::*;
|
2022-11-29 00:16:55 +00:00
|
|
|
use token::TokenType;
|
2022-11-28 23:33:34 +00:00
|
|
|
|
|
|
|
/// Should return an EOF token if the input has no tokens
|
|
|
|
#[test]
|
|
|
|
fn test1() {
|
|
|
|
let input = String::from("");
|
2022-11-29 00:16:55 +00:00
|
|
|
let tokens = get_tokens(&input).unwrap();
|
2023-02-14 20:32:45 +00:00
|
|
|
// 1 semicolon and 1 EOF token
|
|
|
|
assert_eq!(2, tokens.len());
|
|
|
|
let first = tokens.get(1).unwrap();
|
2022-11-28 23:33:34 +00:00
|
|
|
assert_eq!(TokenType::EOF, first.token_type);
|
|
|
|
|
|
|
|
let input = String::from(" ");
|
2022-11-29 00:16:55 +00:00
|
|
|
let tokens = get_tokens(&input).unwrap();
|
2023-02-14 20:32:45 +00:00
|
|
|
// 1 semicolon and 1 EOF token
|
|
|
|
assert_eq!(2, tokens.len());
|
|
|
|
let first = tokens.get(1).unwrap();
|
2022-11-28 23:33:34 +00:00
|
|
|
assert_eq!(TokenType::EOF, first.token_type);
|
|
|
|
|
2022-11-29 00:16:55 +00:00
|
|
|
let input = String::from(" ");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
2023-02-14 20:32:45 +00:00
|
|
|
// 1 semicolon and 1 EOF token
|
|
|
|
assert_eq!(2, tokens.len());
|
|
|
|
let first = tokens.get(1).unwrap();
|
2022-11-28 23:33:34 +00:00
|
|
|
assert_eq!(TokenType::EOF, first.token_type);
|
|
|
|
}
|
|
|
|
|
2022-11-29 00:16:55 +00:00
|
|
|
#[test]
|
|
|
|
fn t() {
|
|
|
|
let input = String::from("126 ");
|
|
|
|
let chars: Vec<char> = input.chars().into_iter().collect();
|
|
|
|
|
|
|
|
assert_eq!(4, chars.len());
|
|
|
|
assert!(has_input(&chars, 0));
|
|
|
|
|
2022-11-30 13:38:43 +00:00
|
|
|
match next_token(&chars, 0) {
|
|
|
|
LexResult::Some(t, _) => {
|
2022-11-29 00:16:55 +00:00
|
|
|
assert_eq!("126", t.value)
|
|
|
|
},
|
2022-11-30 13:38:43 +00:00
|
|
|
_ => {
|
2022-11-29 00:16:55 +00:00
|
|
|
panic!()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-11-28 23:33:34 +00:00
|
|
|
/// Should scan numbers
|
|
|
|
#[test]
|
|
|
|
fn number_test() {
|
2022-11-30 13:38:43 +00:00
|
|
|
let input = String::from("126 278.98 0.282398 1789e+1 239.3298e-103");
|
2022-11-29 00:16:55 +00:00
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
|
|
|
let t1 = tokens.get(0).unwrap();
|
|
|
|
assert_eq!(TokenType::Number, t1.token_type);
|
|
|
|
assert_eq!("126", t1.value);
|
|
|
|
|
|
|
|
let t2 = tokens.get(1).unwrap();
|
|
|
|
assert_eq!(TokenType::Number, t2.token_type);
|
|
|
|
assert_eq!("278.98", t2.value);
|
2022-11-28 23:33:34 +00:00
|
|
|
|
2022-11-29 00:16:55 +00:00
|
|
|
let t3 = tokens.get(2).unwrap();
|
|
|
|
assert_eq!(TokenType::Number, t3.token_type);
|
|
|
|
assert_eq!("0.282398", t3.value);
|
2022-11-30 13:38:43 +00:00
|
|
|
|
|
|
|
assert_eq!("1789e+1", tokens.get(3).unwrap().value);
|
2022-11-28 23:33:34 +00:00
|
|
|
assert_eq!("239.3298e-103", tokens.get(4).unwrap().value);
|
2023-02-14 20:32:45 +00:00
|
|
|
assert_eq!(TokenType::Semicolon, tokens.get(5).unwrap().token_type);
|
|
|
|
assert_eq!(TokenType::EOF, tokens.get(6).unwrap().token_type);
|
2022-11-30 13:38:43 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn grouping_sign_test() {
|
|
|
|
let input = String::from("( ) { } [ ]");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
|
|
|
let t = tokens.get(0).unwrap();
|
|
|
|
assert_eq!(TokenType::LeftParen, t.token_type);
|
|
|
|
assert_eq!("(", t.value);
|
|
|
|
|
|
|
|
let t = tokens.get(1).unwrap();
|
|
|
|
assert_eq!(TokenType::RightParen, t.token_type);
|
|
|
|
assert_eq!(")", t.value);
|
|
|
|
|
|
|
|
let t = tokens.get(2).unwrap();
|
|
|
|
assert_eq!(TokenType::LeftBrace, t.token_type);
|
|
|
|
assert_eq!("{", t.value);
|
|
|
|
|
|
|
|
let t = tokens.get(3).unwrap();
|
|
|
|
assert_eq!(TokenType::RightBrace, t.token_type);
|
|
|
|
assert_eq!("}", t.value);
|
|
|
|
|
|
|
|
let t = tokens.get(4).unwrap();
|
|
|
|
assert_eq!(TokenType::LeftBracket, t.token_type);
|
|
|
|
assert_eq!("[", t.value);
|
|
|
|
|
|
|
|
let t = tokens.get(5).unwrap();
|
|
|
|
assert_eq!(TokenType::RightBracket, t.token_type);
|
|
|
|
assert_eq!("]", t.value);
|
2022-11-28 23:33:34 +00:00
|
|
|
}
|
2023-02-14 20:22:29 +00:00
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn should_scan_new_line() {
|
|
|
|
let input = String::from("3\n22");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
|
|
|
assert_eq!(TokenType::Semicolon, tokens[1].token_type);
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn should_scan_multiple_new_lines() {
|
|
|
|
let input = String::from("3\n\n\n22");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
|
|
|
assert_eq!(TokenType::Semicolon, tokens[1].token_type);
|
|
|
|
assert_eq!(TokenType::Number, tokens[2].token_type);
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn should_scan_multiple_new_lines_with_whitespace_in_between() {
|
|
|
|
let input = String::from("3\n \n \n22");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
|
|
|
assert_eq!(TokenType::Semicolon, tokens[1].token_type);
|
|
|
|
assert_eq!(TokenType::Number, tokens[2].token_type);
|
|
|
|
}
|
2022-11-28 23:33:34 +00:00
|
|
|
}
|