2022-11-28 23:33:34 +00:00
|
|
|
mod scanner;
|
2023-03-14 21:10:43 +00:00
|
|
|
mod utils;
|
2023-01-24 15:01:09 +00:00
|
|
|
|
2023-09-08 01:46:11 +00:00
|
|
|
pub mod token;
|
|
|
|
|
2024-09-14 00:02:19 +00:00
|
|
|
use crate::error_handling::{ErrorContainer, ErrorLabel, LexError, MistiError};
|
2023-09-08 01:46:11 +00:00
|
|
|
use token::Token;
|
2022-11-28 23:33:34 +00:00
|
|
|
|
2023-09-10 16:16:34 +00:00
|
|
|
use self::token::TokenType;
|
|
|
|
|
2022-11-28 23:33:34 +00:00
|
|
|
type Chars = Vec<char>;
|
|
|
|
|
2023-02-11 23:13:05 +00:00
|
|
|
/// Represents the result of scanning a single token from the input
|
2024-08-13 20:05:28 +00:00
|
|
|
#[derive(Debug)]
|
2022-11-30 13:38:43 +00:00
|
|
|
pub enum LexResult {
|
2023-02-11 23:13:05 +00:00
|
|
|
/// A token was found. The first element is the token, and the
|
|
|
|
/// second element is the position in the input after the token.
|
|
|
|
///
|
2023-03-14 21:10:43 +00:00
|
|
|
/// E.g., given an input
|
2023-02-11 23:13:05 +00:00
|
|
|
///
|
|
|
|
/// "`identifier 55`"
|
|
|
|
///
|
|
|
|
/// scanning from a position `0`, the result would be
|
|
|
|
///
|
|
|
|
/// `Some(Token("identifier"), 10)`.
|
|
|
|
///
|
|
|
|
/// where:
|
|
|
|
/// - `Token("identifier")` is the token
|
|
|
|
/// - `10` is the position where the token ends, and from where the next token
|
|
|
|
/// should be scanned
|
2022-11-30 13:38:43 +00:00
|
|
|
Some(Token, usize),
|
2023-09-11 01:10:34 +00:00
|
|
|
/// Multiple tokens
|
|
|
|
Multiple(Vec<Token>, usize),
|
2023-02-11 23:13:05 +00:00
|
|
|
/// No token was found. This indicates that EOF has been reached.
|
|
|
|
///
|
|
|
|
/// Contains the last position, which should be the input lenght - 1
|
2022-11-30 13:38:43 +00:00
|
|
|
None(usize),
|
2023-02-11 23:13:05 +00:00
|
|
|
/// An error was found while scanning.
|
2024-09-14 00:02:19 +00:00
|
|
|
Err(ErrorContainer),
|
2022-11-30 13:38:43 +00:00
|
|
|
}
|
|
|
|
|
2022-11-28 23:33:34 +00:00
|
|
|
/// Scans and returns all the tokens in the input String
|
2023-01-24 15:01:09 +00:00
|
|
|
pub fn get_tokens(input: &String) -> Result<Vec<Token>, MistiError> {
|
2022-11-28 23:33:34 +00:00
|
|
|
let chars: Vec<char> = input.chars().into_iter().collect();
|
|
|
|
let mut results = Vec::new();
|
|
|
|
let mut current_pos: usize = 0;
|
2023-09-10 16:39:52 +00:00
|
|
|
let mut indentation_stack = vec![0];
|
2023-09-10 16:16:34 +00:00
|
|
|
// Used to emit INDENT & DEDENT tokens
|
|
|
|
let mut at_new_line = false;
|
2022-11-28 23:33:34 +00:00
|
|
|
|
|
|
|
while has_input(&chars, current_pos) {
|
2023-09-10 16:16:34 +00:00
|
|
|
match next_token(&chars, current_pos, &mut indentation_stack, at_new_line) {
|
2022-11-30 13:38:43 +00:00
|
|
|
LexResult::Some(token, next_pos) => {
|
2024-06-19 16:33:47 +00:00
|
|
|
// When a INDENT/DEDENT is returned it is because there is a NewLine.
|
|
|
|
// Remove that NewLine token and then insert the corresponding INDENT/DEDENT
|
|
|
|
if token.token_type == TokenType::INDENT || token.token_type == TokenType::DEDENT {
|
|
|
|
results.pop();
|
|
|
|
}
|
|
|
|
|
2023-09-10 16:39:52 +00:00
|
|
|
at_new_line = token.token_type == TokenType::NewLine;
|
2023-09-10 16:16:34 +00:00
|
|
|
|
2022-11-29 00:16:55 +00:00
|
|
|
results.push(token);
|
|
|
|
current_pos = next_pos;
|
2023-03-14 21:10:43 +00:00
|
|
|
}
|
2023-09-11 01:10:34 +00:00
|
|
|
LexResult::Multiple(tokens, next_pos) => {
|
2024-06-19 16:33:47 +00:00
|
|
|
// When a INDENT/DEDENT is returned it is because there is a NewLine.
|
|
|
|
// Remove that NewLine token and then insert the corresponding INDENT/DEDENT
|
|
|
|
match tokens.get(0) {
|
|
|
|
Some(t)
|
|
|
|
if t.token_type == TokenType::INDENT
|
|
|
|
|| t.token_type == TokenType::DEDENT =>
|
|
|
|
{
|
|
|
|
results.pop();
|
|
|
|
}
|
|
|
|
_ => {}
|
|
|
|
}
|
|
|
|
|
2024-06-02 23:18:32 +00:00
|
|
|
at_new_line = match tokens.last() {
|
|
|
|
Some(t) if t.token_type == TokenType::NewLine => true,
|
|
|
|
// This may be None if there are newlines followed by EOF.
|
|
|
|
_ => false,
|
|
|
|
};
|
2023-09-11 01:10:34 +00:00
|
|
|
|
|
|
|
results.extend(tokens);
|
|
|
|
current_pos = next_pos;
|
|
|
|
}
|
2022-11-30 13:38:43 +00:00
|
|
|
LexResult::None(next_pos) => {
|
2022-11-29 00:16:55 +00:00
|
|
|
current_pos = next_pos;
|
2023-03-14 21:10:43 +00:00
|
|
|
}
|
2023-01-24 15:01:09 +00:00
|
|
|
LexResult::Err(error_info) => {
|
|
|
|
return Err(MistiError::Lex(error_info));
|
|
|
|
}
|
2022-11-28 23:33:34 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-09-17 18:58:45 +00:00
|
|
|
// emit DEDENT tokens for each entry left in the indentation_stack,
|
|
|
|
// except the first one (which is 0)
|
|
|
|
for _ in 0..indentation_stack.len() - 1 {
|
|
|
|
results.push(Token::new_dedent(current_pos));
|
|
|
|
}
|
2023-09-10 16:16:34 +00:00
|
|
|
|
2023-09-17 18:58:45 +00:00
|
|
|
// Push EOF
|
2023-09-08 01:46:11 +00:00
|
|
|
results.push(Token::new_eof(0));
|
2022-11-29 00:16:55 +00:00
|
|
|
Ok(results)
|
2022-11-28 23:33:34 +00:00
|
|
|
}
|
|
|
|
|
2023-02-11 23:13:05 +00:00
|
|
|
/// Scans a single token from `chars`, starting from `current_pos`
|
2023-09-10 16:39:52 +00:00
|
|
|
fn next_token(
|
|
|
|
chars: &Chars,
|
|
|
|
current_pos: usize,
|
|
|
|
indentation_stack: &mut Vec<usize>,
|
|
|
|
at_new_line: bool,
|
|
|
|
) -> LexResult {
|
|
|
|
let mut current_pos = current_pos;
|
|
|
|
|
2023-09-11 01:10:34 +00:00
|
|
|
if at_new_line {
|
|
|
|
return handle_indentation(chars, current_pos, indentation_stack);
|
2023-09-17 18:58:45 +00:00
|
|
|
} else if !at_new_line && peek(chars, current_pos) == ' ' {
|
2023-09-11 01:10:34 +00:00
|
|
|
// Consume whitespace
|
|
|
|
current_pos += 1;
|
|
|
|
while peek(chars, current_pos) == ' ' {
|
2023-09-10 16:39:52 +00:00
|
|
|
current_pos += 1;
|
|
|
|
}
|
2022-11-29 00:16:55 +00:00
|
|
|
}
|
|
|
|
|
2023-09-10 16:39:52 +00:00
|
|
|
// If EOF is reached return only the current position
|
|
|
|
if peek(chars, current_pos) == '\0' {
|
|
|
|
return LexResult::None(current_pos);
|
2023-09-10 16:16:34 +00:00
|
|
|
}
|
|
|
|
|
2023-09-10 16:39:52 +00:00
|
|
|
let next_char = peek(chars, current_pos);
|
2022-11-28 23:33:34 +00:00
|
|
|
|
2022-12-01 13:33:48 +00:00
|
|
|
// Scanners
|
2023-03-14 21:10:43 +00:00
|
|
|
None.or_else(|| scanner::number(next_char, chars, current_pos))
|
2022-12-01 17:53:14 +00:00
|
|
|
.or_else(|| scanner::identifier(next_char, chars, current_pos))
|
2023-02-15 21:17:50 +00:00
|
|
|
.or_else(|| scanner::datatype(next_char, chars, current_pos))
|
2022-12-01 17:53:14 +00:00
|
|
|
.or_else(|| scanner::string(next_char, chars, current_pos))
|
2023-04-05 15:31:12 +00:00
|
|
|
.or_else(|| scanner::new_comment(next_char, chars, current_pos))
|
2024-07-29 21:18:33 +00:00
|
|
|
.or_else(|| scanner::new_multiline_comment(next_char, chars, current_pos))
|
2022-12-01 17:53:14 +00:00
|
|
|
.or_else(|| scanner::operator(next_char, chars, current_pos))
|
|
|
|
.or_else(|| scanner::grouping_sign(next_char, chars, current_pos))
|
2023-02-14 20:22:29 +00:00
|
|
|
.or_else(|| scanner::new_line(next_char, chars, current_pos))
|
2024-01-12 00:29:19 +00:00
|
|
|
.or_else(|| {
|
|
|
|
if next_char == ',' {
|
|
|
|
Some(LexResult::Some(
|
|
|
|
Token::new(",".into(), current_pos, TokenType::Comma),
|
|
|
|
current_pos + 1,
|
|
|
|
))
|
|
|
|
} else {
|
|
|
|
None
|
|
|
|
}
|
|
|
|
})
|
2022-11-30 13:38:43 +00:00
|
|
|
.unwrap_or_else(|| {
|
2024-09-14 00:02:19 +00:00
|
|
|
let label = ErrorLabel {
|
|
|
|
message: String::from("This character is not allowed"),
|
|
|
|
start: current_pos,
|
|
|
|
end: current_pos + 1,
|
2023-01-05 17:48:34 +00:00
|
|
|
};
|
2024-09-14 00:02:19 +00:00
|
|
|
let error_container = ErrorContainer {
|
|
|
|
error_offset: current_pos,
|
|
|
|
error_code: 0x010001,
|
|
|
|
labels: vec![label],
|
|
|
|
note: None,
|
|
|
|
help: Some(String::from("Remove this character")),
|
|
|
|
};
|
|
|
|
|
|
|
|
LexResult::Err(error_container)
|
2022-11-30 13:38:43 +00:00
|
|
|
})
|
2022-11-28 23:33:34 +00:00
|
|
|
}
|
|
|
|
|
2023-09-10 16:39:52 +00:00
|
|
|
fn handle_indentation(
|
|
|
|
chars: &Chars,
|
|
|
|
current_pos: usize,
|
|
|
|
indentation_stack: &mut Vec<usize>,
|
|
|
|
) -> LexResult {
|
|
|
|
// Count the number of spaces
|
|
|
|
let mut spaces = 0;
|
|
|
|
let mut sub_pos = current_pos;
|
|
|
|
while peek(chars, sub_pos) == ' ' {
|
|
|
|
spaces += 1;
|
|
|
|
sub_pos += 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Compare the number of spaces with the top of the stack
|
|
|
|
let top = indentation_stack.last().unwrap_or(&0);
|
2023-09-11 01:10:34 +00:00
|
|
|
|
2023-09-10 16:39:52 +00:00
|
|
|
if spaces > *top {
|
|
|
|
// Push the new indentation level
|
|
|
|
indentation_stack.push(spaces);
|
|
|
|
return LexResult::Some(Token::new_indent(current_pos), current_pos + spaces);
|
|
|
|
} else if spaces < *top {
|
2023-09-11 01:10:34 +00:00
|
|
|
// Emit a DEDENT token for each indentation level that is decreased
|
|
|
|
let mut dedent_tokens = Vec::<Token>::new();
|
|
|
|
|
|
|
|
while let Some(new_top) = indentation_stack.last() {
|
|
|
|
if spaces < *new_top {
|
|
|
|
indentation_stack.pop();
|
|
|
|
dedent_tokens.push(Token::new_dedent(current_pos));
|
2023-09-17 18:58:45 +00:00
|
|
|
} else if spaces == *new_top {
|
2023-09-11 01:10:34 +00:00
|
|
|
break;
|
2023-09-17 18:58:45 +00:00
|
|
|
} else {
|
2023-09-11 01:10:34 +00:00
|
|
|
// Illegal state: Indentation error
|
2024-09-14 00:02:19 +00:00
|
|
|
let econtaner = ErrorContainer {
|
|
|
|
error_code: 0,
|
|
|
|
error_offset: current_pos,
|
|
|
|
labels: vec![],
|
|
|
|
note: None,
|
|
|
|
help: None,
|
2023-09-11 01:10:34 +00:00
|
|
|
};
|
2024-09-14 00:02:19 +00:00
|
|
|
|
|
|
|
return LexResult::Err(econtaner);
|
2023-09-11 01:10:34 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-09-17 18:58:45 +00:00
|
|
|
return LexResult::Multiple(dedent_tokens, current_pos + spaces);
|
2023-09-10 16:39:52 +00:00
|
|
|
} else {
|
|
|
|
// Same indentation level
|
2023-09-11 01:10:34 +00:00
|
|
|
return next_token(chars, current_pos + spaces, indentation_stack, false);
|
2023-09-10 16:39:52 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-02-11 23:13:05 +00:00
|
|
|
/// Returns the char at `pos`
|
2022-11-28 23:33:34 +00:00
|
|
|
fn peek(input: &Chars, pos: usize) -> char {
|
|
|
|
let result = input.get(pos).unwrap_or(&'\0');
|
|
|
|
*result
|
|
|
|
}
|
|
|
|
|
2023-02-11 23:13:05 +00:00
|
|
|
/// Whether there is still input based on `current_pos`
|
2022-11-28 23:33:34 +00:00
|
|
|
fn has_input(input: &Chars, current_pos: usize) -> bool {
|
2022-11-29 00:16:55 +00:00
|
|
|
current_pos < input.len()
|
2022-11-28 23:33:34 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#[cfg(test)]
|
|
|
|
mod tests {
|
|
|
|
use super::*;
|
2022-11-29 00:16:55 +00:00
|
|
|
use token::TokenType;
|
2022-11-28 23:33:34 +00:00
|
|
|
|
|
|
|
/// Should return an EOF token if the input has no tokens
|
|
|
|
#[test]
|
2023-09-10 16:39:52 +00:00
|
|
|
fn should_emit_eof() {
|
2022-11-28 23:33:34 +00:00
|
|
|
let input = String::from("");
|
2022-11-29 00:16:55 +00:00
|
|
|
let tokens = get_tokens(&input).unwrap();
|
2023-09-17 18:58:45 +00:00
|
|
|
// a EOF token
|
|
|
|
assert_eq!(1, tokens.len());
|
|
|
|
let first = tokens.get(0).unwrap();
|
2022-11-28 23:33:34 +00:00
|
|
|
assert_eq!(TokenType::EOF, first.token_type);
|
|
|
|
|
|
|
|
let input = String::from(" ");
|
2022-11-29 00:16:55 +00:00
|
|
|
let tokens = get_tokens(&input).unwrap();
|
2023-09-17 18:58:45 +00:00
|
|
|
// a EOF token
|
|
|
|
assert_eq!(1, tokens.len());
|
|
|
|
let first = tokens.get(0).unwrap();
|
2022-11-28 23:33:34 +00:00
|
|
|
assert_eq!(TokenType::EOF, first.token_type);
|
|
|
|
|
2022-11-29 00:16:55 +00:00
|
|
|
let input = String::from(" ");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
2023-09-17 18:58:45 +00:00
|
|
|
// a EOF token
|
|
|
|
assert_eq!(1, tokens.len());
|
|
|
|
let first = tokens.get(0).unwrap();
|
2022-11-28 23:33:34 +00:00
|
|
|
assert_eq!(TokenType::EOF, first.token_type);
|
|
|
|
}
|
|
|
|
|
2022-11-29 00:16:55 +00:00
|
|
|
#[test]
|
|
|
|
fn t() {
|
|
|
|
let input = String::from("126 ");
|
|
|
|
let chars: Vec<char> = input.chars().into_iter().collect();
|
2023-09-10 16:16:34 +00:00
|
|
|
let mut indentation_stack = Vec::<usize>::new();
|
2022-11-29 00:16:55 +00:00
|
|
|
|
|
|
|
assert_eq!(4, chars.len());
|
|
|
|
assert!(has_input(&chars, 0));
|
|
|
|
|
2023-09-10 16:16:34 +00:00
|
|
|
match next_token(&chars, 0, &mut indentation_stack, true) {
|
2022-11-30 13:38:43 +00:00
|
|
|
LexResult::Some(t, _) => {
|
2022-11-29 00:16:55 +00:00
|
|
|
assert_eq!("126", t.value)
|
2023-03-14 21:10:43 +00:00
|
|
|
}
|
2022-11-30 13:38:43 +00:00
|
|
|
_ => {
|
2022-11-29 00:16:55 +00:00
|
|
|
panic!()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-11-28 23:33:34 +00:00
|
|
|
/// Should scan numbers
|
|
|
|
#[test]
|
|
|
|
fn number_test() {
|
2022-11-30 13:38:43 +00:00
|
|
|
let input = String::from("126 278.98 0.282398 1789e+1 239.3298e-103");
|
2022-11-29 00:16:55 +00:00
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
|
|
|
let t1 = tokens.get(0).unwrap();
|
2024-05-06 15:13:21 +00:00
|
|
|
assert_eq!(TokenType::Int, t1.token_type);
|
2022-11-29 00:16:55 +00:00
|
|
|
assert_eq!("126", t1.value);
|
|
|
|
|
|
|
|
let t2 = tokens.get(1).unwrap();
|
2024-05-06 15:13:21 +00:00
|
|
|
assert_eq!(TokenType::Float, t2.token_type);
|
2022-11-29 00:16:55 +00:00
|
|
|
assert_eq!("278.98", t2.value);
|
2022-11-28 23:33:34 +00:00
|
|
|
|
2022-11-29 00:16:55 +00:00
|
|
|
let t3 = tokens.get(2).unwrap();
|
2024-05-06 15:13:21 +00:00
|
|
|
assert_eq!(TokenType::Float, t3.token_type);
|
2022-11-29 00:16:55 +00:00
|
|
|
assert_eq!("0.282398", t3.value);
|
2023-03-14 21:10:43 +00:00
|
|
|
|
2022-11-30 13:38:43 +00:00
|
|
|
assert_eq!("1789e+1", tokens.get(3).unwrap().value);
|
2022-11-28 23:33:34 +00:00
|
|
|
assert_eq!("239.3298e-103", tokens.get(4).unwrap().value);
|
2023-09-17 18:58:45 +00:00
|
|
|
assert_eq!(TokenType::EOF, tokens.get(5).unwrap().token_type);
|
2022-11-30 13:38:43 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn grouping_sign_test() {
|
|
|
|
let input = String::from("( ) { } [ ]");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
|
|
|
let t = tokens.get(0).unwrap();
|
|
|
|
assert_eq!(TokenType::LeftParen, t.token_type);
|
|
|
|
assert_eq!("(", t.value);
|
|
|
|
|
|
|
|
let t = tokens.get(1).unwrap();
|
|
|
|
assert_eq!(TokenType::RightParen, t.token_type);
|
|
|
|
assert_eq!(")", t.value);
|
|
|
|
|
|
|
|
let t = tokens.get(2).unwrap();
|
|
|
|
assert_eq!(TokenType::LeftBrace, t.token_type);
|
|
|
|
assert_eq!("{", t.value);
|
|
|
|
|
|
|
|
let t = tokens.get(3).unwrap();
|
|
|
|
assert_eq!(TokenType::RightBrace, t.token_type);
|
|
|
|
assert_eq!("}", t.value);
|
|
|
|
|
|
|
|
let t = tokens.get(4).unwrap();
|
|
|
|
assert_eq!(TokenType::LeftBracket, t.token_type);
|
|
|
|
assert_eq!("[", t.value);
|
|
|
|
|
|
|
|
let t = tokens.get(5).unwrap();
|
|
|
|
assert_eq!(TokenType::RightBracket, t.token_type);
|
|
|
|
assert_eq!("]", t.value);
|
2022-11-28 23:33:34 +00:00
|
|
|
}
|
2023-03-14 21:10:43 +00:00
|
|
|
|
2023-02-15 21:17:50 +00:00
|
|
|
#[test]
|
|
|
|
fn should_scan_datatype() {
|
|
|
|
let input = String::from("Num");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
|
|
|
assert_eq!(TokenType::Datatype, tokens[0].token_type);
|
|
|
|
}
|
2023-03-14 21:10:43 +00:00
|
|
|
|
2023-02-14 20:22:29 +00:00
|
|
|
#[test]
|
|
|
|
fn should_scan_new_line() {
|
|
|
|
let input = String::from("3\n22");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
2023-09-10 16:16:34 +00:00
|
|
|
assert_eq!(TokenType::NewLine, tokens[1].token_type);
|
2023-02-14 20:22:29 +00:00
|
|
|
}
|
2023-03-14 21:10:43 +00:00
|
|
|
|
2023-02-14 20:22:29 +00:00
|
|
|
#[test]
|
|
|
|
fn should_scan_multiple_new_lines() {
|
|
|
|
let input = String::from("3\n\n\n22");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
2023-09-10 16:16:34 +00:00
|
|
|
assert_eq!(TokenType::NewLine, tokens[1].token_type);
|
2024-05-06 15:13:21 +00:00
|
|
|
assert_eq!(TokenType::Int, tokens[2].token_type);
|
2023-02-14 20:22:29 +00:00
|
|
|
}
|
2023-03-14 21:10:43 +00:00
|
|
|
|
2023-02-14 20:22:29 +00:00
|
|
|
#[test]
|
|
|
|
fn should_scan_multiple_new_lines_with_whitespace_in_between() {
|
|
|
|
let input = String::from("3\n \n \n22");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
2023-09-10 16:16:34 +00:00
|
|
|
assert_eq!(TokenType::NewLine, tokens[1].token_type);
|
2024-05-06 15:13:21 +00:00
|
|
|
assert_eq!(TokenType::Int, tokens[2].token_type);
|
2023-02-14 20:22:29 +00:00
|
|
|
}
|
2023-09-10 16:16:34 +00:00
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn should_emit_indent_token() {
|
|
|
|
let input = String::from("3\n \n 22");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
2024-05-06 15:13:21 +00:00
|
|
|
assert_eq!(TokenType::Int, tokens[0].token_type);
|
2024-06-19 16:33:47 +00:00
|
|
|
assert_eq!(TokenType::INDENT, tokens[1].token_type);
|
|
|
|
assert_eq!(TokenType::Int, tokens[2].token_type);
|
2023-09-10 16:16:34 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn should_emit_indent_when_indentation_increases() {
|
|
|
|
let input = String::from("3\n \n 22\n 111");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
2024-05-06 15:13:21 +00:00
|
|
|
assert_eq!(TokenType::Int, tokens[0].token_type);
|
2024-06-19 16:33:47 +00:00
|
|
|
assert_eq!(TokenType::INDENT, tokens[1].token_type);
|
|
|
|
assert_eq!(TokenType::Int, tokens[2].token_type);
|
|
|
|
assert_eq!(TokenType::INDENT, tokens[3].token_type);
|
|
|
|
assert_eq!(TokenType::Int, tokens[4].token_type);
|
2023-09-10 16:16:34 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn shouldnt_emit_indent_when_indentation_stays() {
|
|
|
|
let input = String::from("3\n \n 22\n 111");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
2024-05-06 15:13:21 +00:00
|
|
|
assert_eq!(TokenType::Int, tokens[0].token_type);
|
2024-06-19 16:33:47 +00:00
|
|
|
assert_eq!(TokenType::INDENT, tokens[1].token_type);
|
|
|
|
assert_eq!(TokenType::Int, tokens[2].token_type);
|
|
|
|
assert_eq!(TokenType::NewLine, tokens[3].token_type);
|
|
|
|
assert_eq!(TokenType::Int, tokens[4].token_type);
|
2023-09-10 16:16:34 +00:00
|
|
|
}
|
2023-09-11 01:10:34 +00:00
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn should_emit_dedent() {
|
|
|
|
let input = String::from("3\n \n 22\n111");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
2024-05-06 15:13:21 +00:00
|
|
|
assert_eq!(TokenType::Int, tokens[0].token_type);
|
2024-06-19 16:33:47 +00:00
|
|
|
assert_eq!(TokenType::INDENT, tokens[1].token_type);
|
|
|
|
assert_eq!(TokenType::Int, tokens[2].token_type);
|
|
|
|
assert_eq!(TokenType::DEDENT, tokens[3].token_type);
|
|
|
|
assert_eq!(TokenType::Int, tokens[4].token_type);
|
2023-09-11 01:10:34 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn should_emit_multiple_dedents() {
|
|
|
|
let input = String::from("1\n 2\n 3\n 4\n5");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
2024-05-06 15:13:21 +00:00
|
|
|
assert_eq!(TokenType::Int, tokens[0].token_type);
|
2024-06-19 16:33:47 +00:00
|
|
|
assert_eq!(TokenType::INDENT, tokens[1].token_type);
|
|
|
|
assert_eq!(TokenType::Int, tokens[2].token_type);
|
|
|
|
assert_eq!(TokenType::INDENT, tokens[3].token_type);
|
|
|
|
assert_eq!(TokenType::Int, tokens[4].token_type);
|
|
|
|
assert_eq!(TokenType::DEDENT, tokens[5].token_type);
|
2024-05-06 15:13:21 +00:00
|
|
|
assert_eq!(TokenType::Int, tokens[6].token_type);
|
2024-06-19 16:33:47 +00:00
|
|
|
assert_eq!(TokenType::DEDENT, tokens[7].token_type);
|
2023-09-11 01:10:34 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn should_emit_multiple_dedents_2() {
|
|
|
|
let input = String::from("1\n 2\n 3\n4");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
2024-05-06 15:13:21 +00:00
|
|
|
assert_eq!(TokenType::Int, tokens[0].token_type);
|
2024-06-19 16:33:47 +00:00
|
|
|
assert_eq!(TokenType::INDENT, tokens[1].token_type);
|
|
|
|
assert_eq!(TokenType::Int, tokens[2].token_type);
|
|
|
|
assert_eq!(TokenType::INDENT, tokens[3].token_type);
|
|
|
|
assert_eq!(TokenType::Int, tokens[4].token_type);
|
|
|
|
assert_eq!(TokenType::DEDENT, tokens[5].token_type);
|
|
|
|
assert_eq!(TokenType::DEDENT, tokens[6].token_type);
|
|
|
|
assert_eq!(TokenType::Int, tokens[7].token_type);
|
2023-09-11 01:10:34 +00:00
|
|
|
}
|
2024-06-02 23:18:32 +00:00
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn shouldnt_emit_trailing_newlines() {
|
|
|
|
let input = String::from("token\n");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
|
|
|
assert_eq!(2, tokens.len());
|
|
|
|
assert_eq!(TokenType::Identifier, tokens[0].token_type);
|
|
|
|
assert_eq!(TokenType::EOF, tokens[1].token_type);
|
|
|
|
}
|
2022-11-28 23:33:34 +00:00
|
|
|
}
|
2023-09-17 18:58:45 +00:00
|
|
|
|
|
|
|
#[cfg(test)]
|
|
|
|
mod indentation_tests {
|
|
|
|
use super::*;
|
|
|
|
use token::TokenType;
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn should_emit_dedents_on_eof() {
|
|
|
|
let input = String::from("1\n 2");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
2024-05-06 15:13:21 +00:00
|
|
|
assert_eq!(TokenType::Int, tokens[0].token_type);
|
2024-06-19 16:33:47 +00:00
|
|
|
assert_eq!(TokenType::INDENT, tokens[1].token_type);
|
|
|
|
assert_eq!(TokenType::Int, tokens[2].token_type);
|
|
|
|
assert_eq!(TokenType::DEDENT, tokens[3].token_type);
|
|
|
|
assert_eq!(TokenType::EOF, tokens[4].token_type);
|
2023-09-17 18:58:45 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn should_emit_dedents_on_eof_2() {
|
|
|
|
let input = String::from("1\n 2\n 3");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
2024-05-06 15:13:21 +00:00
|
|
|
assert_eq!(TokenType::Int, tokens[0].token_type);
|
2024-06-19 16:33:47 +00:00
|
|
|
assert_eq!(TokenType::INDENT, tokens[1].token_type);
|
|
|
|
assert_eq!(TokenType::Int, tokens[2].token_type);
|
|
|
|
assert_eq!(TokenType::INDENT, tokens[3].token_type);
|
|
|
|
assert_eq!(TokenType::Int, tokens[4].token_type);
|
|
|
|
assert_eq!(TokenType::DEDENT, tokens[5].token_type);
|
|
|
|
assert_eq!(TokenType::DEDENT, tokens[6].token_type);
|
|
|
|
assert_eq!(TokenType::EOF, tokens[7].token_type);
|
2023-09-17 18:58:45 +00:00
|
|
|
}
|
2024-01-02 12:06:46 +00:00
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn should_lex_comments() {
|
|
|
|
let input = String::from("// ??");
|
|
|
|
let tokens = get_tokens(&input).unwrap();
|
|
|
|
|
|
|
|
assert_eq!(TokenType::Comment, tokens[0].token_type);
|
|
|
|
}
|
2024-03-18 22:21:02 +00:00
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn should_emit_error_on_incorrect_indentation() {
|
|
|
|
let input = String::from("1\n 2\n 3");
|
|
|
|
let tokens = get_tokens(&input);
|
|
|
|
|
|
|
|
assert!(tokens.is_err());
|
|
|
|
}
|
2023-09-17 18:58:45 +00:00
|
|
|
}
|