|
@@ -51,103 +51,109 @@ struct LexerTest : ::testing::Test {
|
|
|
TEST_F(LexerTest, HandlesEmptyBuffer) {
|
|
TEST_F(LexerTest, HandlesEmptyBuffer) {
|
|
|
auto buffer = Lex("");
|
|
auto buffer = Lex("");
|
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
|
- EXPECT_EQ(buffer.Tokens().begin(), buffer.Tokens().end());
|
|
|
|
|
|
|
+ EXPECT_THAT(
|
|
|
|
|
+ buffer,
|
|
|
|
|
+ HasTokens(llvm::ArrayRef<ExpectedToken>{{TokenKind::EndOfFile()}}));
|
|
|
}
|
|
}
|
|
|
|
|
|
|
|
TEST_F(LexerTest, TracksLinesAndColumns) {
|
|
TEST_F(LexerTest, TracksLinesAndColumns) {
|
|
|
auto buffer = Lex("\n ;;\n ;;;\n x\"foo\" \"\"\"baz\n a\n \"\"\" y");
|
|
auto buffer = Lex("\n ;;\n ;;;\n x\"foo\" \"\"\"baz\n a\n \"\"\" y");
|
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
|
- EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
|
|
- {.kind = TokenKind::Semi(),
|
|
|
|
|
- .line = 2,
|
|
|
|
|
- .column = 3,
|
|
|
|
|
- .indent_column = 3},
|
|
|
|
|
- {.kind = TokenKind::Semi(),
|
|
|
|
|
- .line = 2,
|
|
|
|
|
- .column = 4,
|
|
|
|
|
- .indent_column = 3},
|
|
|
|
|
- {.kind = TokenKind::Semi(),
|
|
|
|
|
- .line = 3,
|
|
|
|
|
- .column = 4,
|
|
|
|
|
- .indent_column = 4},
|
|
|
|
|
- {.kind = TokenKind::Semi(),
|
|
|
|
|
- .line = 3,
|
|
|
|
|
- .column = 5,
|
|
|
|
|
- .indent_column = 4},
|
|
|
|
|
- {.kind = TokenKind::Semi(),
|
|
|
|
|
- .line = 3,
|
|
|
|
|
- .column = 6,
|
|
|
|
|
- .indent_column = 4},
|
|
|
|
|
- {.kind = TokenKind::Identifier(),
|
|
|
|
|
- .line = 4,
|
|
|
|
|
- .column = 4,
|
|
|
|
|
- .indent_column = 4,
|
|
|
|
|
- .text = "x"},
|
|
|
|
|
- {.kind = TokenKind::StringLiteral(),
|
|
|
|
|
- .line = 4,
|
|
|
|
|
- .column = 5,
|
|
|
|
|
- .indent_column = 4},
|
|
|
|
|
- {.kind = TokenKind::StringLiteral(),
|
|
|
|
|
- .line = 4,
|
|
|
|
|
- .column = 11,
|
|
|
|
|
- .indent_column = 4},
|
|
|
|
|
- {.kind = TokenKind::Identifier(),
|
|
|
|
|
- .line = 6,
|
|
|
|
|
- .column = 6,
|
|
|
|
|
- .indent_column = 11,
|
|
|
|
|
- .text = "y"},
|
|
|
|
|
- }));
|
|
|
|
|
|
|
+ EXPECT_THAT(buffer,
|
|
|
|
|
+ HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
|
|
+ {.kind = TokenKind::Semi(),
|
|
|
|
|
+ .line = 2,
|
|
|
|
|
+ .column = 3,
|
|
|
|
|
+ .indent_column = 3},
|
|
|
|
|
+ {.kind = TokenKind::Semi(),
|
|
|
|
|
+ .line = 2,
|
|
|
|
|
+ .column = 4,
|
|
|
|
|
+ .indent_column = 3},
|
|
|
|
|
+ {.kind = TokenKind::Semi(),
|
|
|
|
|
+ .line = 3,
|
|
|
|
|
+ .column = 4,
|
|
|
|
|
+ .indent_column = 4},
|
|
|
|
|
+ {.kind = TokenKind::Semi(),
|
|
|
|
|
+ .line = 3,
|
|
|
|
|
+ .column = 5,
|
|
|
|
|
+ .indent_column = 4},
|
|
|
|
|
+ {.kind = TokenKind::Semi(),
|
|
|
|
|
+ .line = 3,
|
|
|
|
|
+ .column = 6,
|
|
|
|
|
+ .indent_column = 4},
|
|
|
|
|
+ {.kind = TokenKind::Identifier(),
|
|
|
|
|
+ .line = 4,
|
|
|
|
|
+ .column = 4,
|
|
|
|
|
+ .indent_column = 4,
|
|
|
|
|
+ .text = "x"},
|
|
|
|
|
+ {.kind = TokenKind::StringLiteral(),
|
|
|
|
|
+ .line = 4,
|
|
|
|
|
+ .column = 5,
|
|
|
|
|
+ .indent_column = 4},
|
|
|
|
|
+ {.kind = TokenKind::StringLiteral(),
|
|
|
|
|
+ .line = 4,
|
|
|
|
|
+ .column = 11,
|
|
|
|
|
+ .indent_column = 4},
|
|
|
|
|
+ {.kind = TokenKind::Identifier(),
|
|
|
|
|
+ .line = 6,
|
|
|
|
|
+ .column = 6,
|
|
|
|
|
+ .indent_column = 11,
|
|
|
|
|
+ .text = "y"},
|
|
|
|
|
+ {.kind = TokenKind::EndOfFile(), .line = 6, .column = 7},
|
|
|
|
|
+ }));
|
|
|
}
|
|
}
|
|
|
|
|
|
|
|
TEST_F(LexerTest, HandlesNumericLiteral) {
|
|
TEST_F(LexerTest, HandlesNumericLiteral) {
|
|
|
auto buffer = Lex("12-578\n 1 2\n0x12_3ABC\n0b10_10_11\n1_234_567\n1.5e9");
|
|
auto buffer = Lex("12-578\n 1 2\n0x12_3ABC\n0b10_10_11\n1_234_567\n1.5e9");
|
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
|
- ASSERT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
|
|
- {.kind = TokenKind::IntegerLiteral(),
|
|
|
|
|
- .line = 1,
|
|
|
|
|
- .column = 1,
|
|
|
|
|
- .indent_column = 1,
|
|
|
|
|
- .text = "12"},
|
|
|
|
|
- {.kind = TokenKind::Minus(),
|
|
|
|
|
- .line = 1,
|
|
|
|
|
- .column = 3,
|
|
|
|
|
- .indent_column = 1},
|
|
|
|
|
- {.kind = TokenKind::IntegerLiteral(),
|
|
|
|
|
- .line = 1,
|
|
|
|
|
- .column = 4,
|
|
|
|
|
- .indent_column = 1,
|
|
|
|
|
- .text = "578"},
|
|
|
|
|
- {.kind = TokenKind::IntegerLiteral(),
|
|
|
|
|
- .line = 2,
|
|
|
|
|
- .column = 3,
|
|
|
|
|
- .indent_column = 3,
|
|
|
|
|
- .text = "1"},
|
|
|
|
|
- {.kind = TokenKind::IntegerLiteral(),
|
|
|
|
|
- .line = 2,
|
|
|
|
|
- .column = 6,
|
|
|
|
|
- .indent_column = 3,
|
|
|
|
|
- .text = "2"},
|
|
|
|
|
- {.kind = TokenKind::IntegerLiteral(),
|
|
|
|
|
- .line = 3,
|
|
|
|
|
- .column = 1,
|
|
|
|
|
- .indent_column = 1,
|
|
|
|
|
- .text = "0x12_3ABC"},
|
|
|
|
|
- {.kind = TokenKind::IntegerLiteral(),
|
|
|
|
|
- .line = 4,
|
|
|
|
|
- .column = 1,
|
|
|
|
|
- .indent_column = 1,
|
|
|
|
|
- .text = "0b10_10_11"},
|
|
|
|
|
- {.kind = TokenKind::IntegerLiteral(),
|
|
|
|
|
- .line = 5,
|
|
|
|
|
- .column = 1,
|
|
|
|
|
- .indent_column = 1,
|
|
|
|
|
- .text = "1_234_567"},
|
|
|
|
|
- {.kind = TokenKind::RealLiteral(),
|
|
|
|
|
- .line = 6,
|
|
|
|
|
- .column = 1,
|
|
|
|
|
- .indent_column = 1,
|
|
|
|
|
- .text = "1.5e9"},
|
|
|
|
|
- }));
|
|
|
|
|
|
|
+ ASSERT_THAT(buffer,
|
|
|
|
|
+ HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
|
|
+ {.kind = TokenKind::IntegerLiteral(),
|
|
|
|
|
+ .line = 1,
|
|
|
|
|
+ .column = 1,
|
|
|
|
|
+ .indent_column = 1,
|
|
|
|
|
+ .text = "12"},
|
|
|
|
|
+ {.kind = TokenKind::Minus(),
|
|
|
|
|
+ .line = 1,
|
|
|
|
|
+ .column = 3,
|
|
|
|
|
+ .indent_column = 1},
|
|
|
|
|
+ {.kind = TokenKind::IntegerLiteral(),
|
|
|
|
|
+ .line = 1,
|
|
|
|
|
+ .column = 4,
|
|
|
|
|
+ .indent_column = 1,
|
|
|
|
|
+ .text = "578"},
|
|
|
|
|
+ {.kind = TokenKind::IntegerLiteral(),
|
|
|
|
|
+ .line = 2,
|
|
|
|
|
+ .column = 3,
|
|
|
|
|
+ .indent_column = 3,
|
|
|
|
|
+ .text = "1"},
|
|
|
|
|
+ {.kind = TokenKind::IntegerLiteral(),
|
|
|
|
|
+ .line = 2,
|
|
|
|
|
+ .column = 6,
|
|
|
|
|
+ .indent_column = 3,
|
|
|
|
|
+ .text = "2"},
|
|
|
|
|
+ {.kind = TokenKind::IntegerLiteral(),
|
|
|
|
|
+ .line = 3,
|
|
|
|
|
+ .column = 1,
|
|
|
|
|
+ .indent_column = 1,
|
|
|
|
|
+ .text = "0x12_3ABC"},
|
|
|
|
|
+ {.kind = TokenKind::IntegerLiteral(),
|
|
|
|
|
+ .line = 4,
|
|
|
|
|
+ .column = 1,
|
|
|
|
|
+ .indent_column = 1,
|
|
|
|
|
+ .text = "0b10_10_11"},
|
|
|
|
|
+ {.kind = TokenKind::IntegerLiteral(),
|
|
|
|
|
+ .line = 5,
|
|
|
|
|
+ .column = 1,
|
|
|
|
|
+ .indent_column = 1,
|
|
|
|
|
+ .text = "1_234_567"},
|
|
|
|
|
+ {.kind = TokenKind::RealLiteral(),
|
|
|
|
|
+ .line = 6,
|
|
|
|
|
+ .column = 1,
|
|
|
|
|
+ .indent_column = 1,
|
|
|
|
|
+ .text = "1.5e9"},
|
|
|
|
|
+ {.kind = TokenKind::EndOfFile(), .line = 6, .column = 6},
|
|
|
|
|
+ }));
|
|
|
auto token_12 = buffer.Tokens().begin();
|
|
auto token_12 = buffer.Tokens().begin();
|
|
|
EXPECT_EQ(buffer.GetIntegerLiteral(*token_12), 12);
|
|
EXPECT_EQ(buffer.GetIntegerLiteral(*token_12), 12);
|
|
|
auto token_578 = buffer.Tokens().begin() + 2;
|
|
auto token_578 = buffer.Tokens().begin() + 2;
|
|
@@ -172,33 +178,35 @@ TEST_F(LexerTest, HandlesNumericLiteral) {
|
|
|
TEST_F(LexerTest, HandlesInvalidNumericLiterals) {
|
|
TEST_F(LexerTest, HandlesInvalidNumericLiterals) {
|
|
|
auto buffer = Lex("14x 15_49 0x3.5q 0x3_4.5_6 0ops");
|
|
auto buffer = Lex("14x 15_49 0x3.5q 0x3_4.5_6 0ops");
|
|
|
EXPECT_TRUE(buffer.HasErrors());
|
|
EXPECT_TRUE(buffer.HasErrors());
|
|
|
- ASSERT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
|
|
- {.kind = TokenKind::Error(),
|
|
|
|
|
- .line = 1,
|
|
|
|
|
- .column = 1,
|
|
|
|
|
- .indent_column = 1,
|
|
|
|
|
- .text = "14x"},
|
|
|
|
|
- {.kind = TokenKind::IntegerLiteral(),
|
|
|
|
|
- .line = 1,
|
|
|
|
|
- .column = 5,
|
|
|
|
|
- .indent_column = 1,
|
|
|
|
|
- .text = "15_49"},
|
|
|
|
|
- {.kind = TokenKind::Error(),
|
|
|
|
|
- .line = 1,
|
|
|
|
|
- .column = 11,
|
|
|
|
|
- .indent_column = 1,
|
|
|
|
|
- .text = "0x3.5q"},
|
|
|
|
|
- {.kind = TokenKind::RealLiteral(),
|
|
|
|
|
- .line = 1,
|
|
|
|
|
- .column = 18,
|
|
|
|
|
- .indent_column = 1,
|
|
|
|
|
- .text = "0x3_4.5_6"},
|
|
|
|
|
- {.kind = TokenKind::Error(),
|
|
|
|
|
- .line = 1,
|
|
|
|
|
- .column = 28,
|
|
|
|
|
- .indent_column = 1,
|
|
|
|
|
- .text = "0ops"},
|
|
|
|
|
- }));
|
|
|
|
|
|
|
+ ASSERT_THAT(buffer,
|
|
|
|
|
+ HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
|
|
+ {.kind = TokenKind::Error(),
|
|
|
|
|
+ .line = 1,
|
|
|
|
|
+ .column = 1,
|
|
|
|
|
+ .indent_column = 1,
|
|
|
|
|
+ .text = "14x"},
|
|
|
|
|
+ {.kind = TokenKind::IntegerLiteral(),
|
|
|
|
|
+ .line = 1,
|
|
|
|
|
+ .column = 5,
|
|
|
|
|
+ .indent_column = 1,
|
|
|
|
|
+ .text = "15_49"},
|
|
|
|
|
+ {.kind = TokenKind::Error(),
|
|
|
|
|
+ .line = 1,
|
|
|
|
|
+ .column = 11,
|
|
|
|
|
+ .indent_column = 1,
|
|
|
|
|
+ .text = "0x3.5q"},
|
|
|
|
|
+ {.kind = TokenKind::RealLiteral(),
|
|
|
|
|
+ .line = 1,
|
|
|
|
|
+ .column = 18,
|
|
|
|
|
+ .indent_column = 1,
|
|
|
|
|
+ .text = "0x3_4.5_6"},
|
|
|
|
|
+ {.kind = TokenKind::Error(),
|
|
|
|
|
+ .line = 1,
|
|
|
|
|
+ .column = 28,
|
|
|
|
|
+ .indent_column = 1,
|
|
|
|
|
+ .text = "0ops"},
|
|
|
|
|
+ {.kind = TokenKind::EndOfFile(), .line = 1, .column = 32},
|
|
|
|
|
+ }));
|
|
|
}
|
|
}
|
|
|
|
|
|
|
|
TEST_F(LexerTest, SplitsNumericLiteralsProperly) {
|
|
TEST_F(LexerTest, SplitsNumericLiteralsProperly) {
|
|
@@ -268,6 +276,8 @@ TEST_F(LexerTest, SplitsNumericLiteralsProperly) {
|
|
|
{.kind = TokenKind::IntegerLiteral(), .text = "13"},
|
|
{.kind = TokenKind::IntegerLiteral(), .text = "13"},
|
|
|
{.kind = TokenKind::Period()},
|
|
{.kind = TokenKind::Period()},
|
|
|
{.kind = TokenKind::UnderscoreKeyword()},
|
|
{.kind = TokenKind::UnderscoreKeyword()},
|
|
|
|
|
+ // newline
|
|
|
|
|
+ {.kind = TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
}
|
|
}
|
|
|
|
|
|
|
@@ -309,6 +319,7 @@ TEST_F(LexerTest, HandlesGarbageCharacters) {
|
|
|
.line = 4,
|
|
.line = 4,
|
|
|
.column = 2,
|
|
.column = 2,
|
|
|
.text = llvm::StringRef("\\", 1)},
|
|
.text = llvm::StringRef("\\", 1)},
|
|
|
|
|
+ {.kind = TokenKind::EndOfFile(), .line = 4, .column = 3},
|
|
|
}));
|
|
}));
|
|
|
}
|
|
}
|
|
|
|
|
|
|
@@ -321,6 +332,7 @@ TEST_F(LexerTest, Symbols) {
|
|
|
EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
{TokenKind::LessLess()},
|
|
{TokenKind::LessLess()},
|
|
|
{TokenKind::Less()},
|
|
{TokenKind::Less()},
|
|
|
|
|
+ {TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
|
|
|
|
|
buffer = Lex("<<=>>");
|
|
buffer = Lex("<<=>>");
|
|
@@ -328,6 +340,7 @@ TEST_F(LexerTest, Symbols) {
|
|
|
EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
{TokenKind::LessLessEqual()},
|
|
{TokenKind::LessLessEqual()},
|
|
|
{TokenKind::GreaterGreater()},
|
|
{TokenKind::GreaterGreater()},
|
|
|
|
|
+ {TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
|
|
|
|
|
buffer = Lex("< <=> >");
|
|
buffer = Lex("< <=> >");
|
|
@@ -336,6 +349,7 @@ TEST_F(LexerTest, Symbols) {
|
|
|
{TokenKind::Less()},
|
|
{TokenKind::Less()},
|
|
|
{TokenKind::LessEqualGreater()},
|
|
{TokenKind::LessEqualGreater()},
|
|
|
{TokenKind::Greater()},
|
|
{TokenKind::Greater()},
|
|
|
|
|
+ {TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
|
|
|
|
|
buffer = Lex("\\/?@&^!");
|
|
buffer = Lex("\\/?@&^!");
|
|
@@ -348,6 +362,7 @@ TEST_F(LexerTest, Symbols) {
|
|
|
{TokenKind::Amp()},
|
|
{TokenKind::Amp()},
|
|
|
{TokenKind::Caret()},
|
|
{TokenKind::Caret()},
|
|
|
{TokenKind::Exclaim()},
|
|
{TokenKind::Exclaim()},
|
|
|
|
|
+ {TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
}
|
|
}
|
|
|
|
|
|
|
@@ -357,6 +372,7 @@ TEST_F(LexerTest, Parens) {
|
|
|
EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
{TokenKind::OpenParen()},
|
|
{TokenKind::OpenParen()},
|
|
|
{TokenKind::CloseParen()},
|
|
{TokenKind::CloseParen()},
|
|
|
|
|
+ {TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
|
|
|
|
|
buffer = Lex("((()()))");
|
|
buffer = Lex("((()()))");
|
|
@@ -370,6 +386,7 @@ TEST_F(LexerTest, Parens) {
|
|
|
{TokenKind::CloseParen()},
|
|
{TokenKind::CloseParen()},
|
|
|
{TokenKind::CloseParen()},
|
|
{TokenKind::CloseParen()},
|
|
|
{TokenKind::CloseParen()},
|
|
{TokenKind::CloseParen()},
|
|
|
|
|
+ {TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
}
|
|
}
|
|
|
|
|
|
|
@@ -379,6 +396,7 @@ TEST_F(LexerTest, CurlyBraces) {
|
|
|
EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
{TokenKind::OpenCurlyBrace()},
|
|
{TokenKind::OpenCurlyBrace()},
|
|
|
{TokenKind::CloseCurlyBrace()},
|
|
{TokenKind::CloseCurlyBrace()},
|
|
|
|
|
+ {TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
|
|
|
|
|
buffer = Lex("{{{}{}}}");
|
|
buffer = Lex("{{{}{}}}");
|
|
@@ -392,6 +410,7 @@ TEST_F(LexerTest, CurlyBraces) {
|
|
|
{TokenKind::CloseCurlyBrace()},
|
|
{TokenKind::CloseCurlyBrace()},
|
|
|
{TokenKind::CloseCurlyBrace()},
|
|
{TokenKind::CloseCurlyBrace()},
|
|
|
{TokenKind::CloseCurlyBrace()},
|
|
{TokenKind::CloseCurlyBrace()},
|
|
|
|
|
+ {TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
}
|
|
}
|
|
|
|
|
|
|
@@ -412,6 +431,8 @@ TEST_F(LexerTest, MatchingGroups) {
|
|
|
buffer.GetMatchedClosingToken(open_curly_token));
|
|
buffer.GetMatchedClosingToken(open_curly_token));
|
|
|
EXPECT_EQ(open_curly_token,
|
|
EXPECT_EQ(open_curly_token,
|
|
|
buffer.GetMatchedOpeningToken(close_curly_token));
|
|
buffer.GetMatchedOpeningToken(close_curly_token));
|
|
|
|
|
+ auto eof_token = *it++;
|
|
|
|
|
+ EXPECT_EQ(buffer.GetKind(eof_token), TokenKind::EndOfFile());
|
|
|
EXPECT_EQ(buffer.Tokens().end(), it);
|
|
EXPECT_EQ(buffer.Tokens().end(), it);
|
|
|
}
|
|
}
|
|
|
|
|
|
|
@@ -473,6 +494,8 @@ TEST_F(LexerTest, MatchingGroups) {
|
|
|
EXPECT_EQ(inner_open_paren_token,
|
|
EXPECT_EQ(inner_open_paren_token,
|
|
|
buffer.GetMatchedOpeningToken(inner_close_paren_token));
|
|
buffer.GetMatchedOpeningToken(inner_close_paren_token));
|
|
|
|
|
|
|
|
|
|
+ auto eof_token = *it++;
|
|
|
|
|
+ EXPECT_EQ(buffer.GetKind(eof_token), TokenKind::EndOfFile());
|
|
|
EXPECT_EQ(buffer.Tokens().end(), it);
|
|
EXPECT_EQ(buffer.Tokens().end(), it);
|
|
|
}
|
|
}
|
|
|
}
|
|
}
|
|
@@ -484,12 +507,14 @@ TEST_F(LexerTest, MismatchedGroups) {
|
|
|
HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
{TokenKind::OpenCurlyBrace()},
|
|
{TokenKind::OpenCurlyBrace()},
|
|
|
{.kind = TokenKind::CloseCurlyBrace(), .recovery = true},
|
|
{.kind = TokenKind::CloseCurlyBrace(), .recovery = true},
|
|
|
|
|
+ {TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
|
|
|
|
|
buffer = Lex("}");
|
|
buffer = Lex("}");
|
|
|
EXPECT_TRUE(buffer.HasErrors());
|
|
EXPECT_TRUE(buffer.HasErrors());
|
|
|
EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
{.kind = TokenKind::Error(), .text = "}"},
|
|
{.kind = TokenKind::Error(), .text = "}"},
|
|
|
|
|
+ {TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
|
|
|
|
|
buffer = Lex("{(}");
|
|
buffer = Lex("{(}");
|
|
@@ -501,6 +526,7 @@ TEST_F(LexerTest, MismatchedGroups) {
|
|
|
{.kind = TokenKind::OpenParen(), .column = 2},
|
|
{.kind = TokenKind::OpenParen(), .column = 2},
|
|
|
{.kind = TokenKind::CloseParen(), .column = 3, .recovery = true},
|
|
{.kind = TokenKind::CloseParen(), .column = 3, .recovery = true},
|
|
|
{.kind = TokenKind::CloseCurlyBrace(), .column = 3},
|
|
{.kind = TokenKind::CloseCurlyBrace(), .column = 3},
|
|
|
|
|
+ {TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
|
|
|
|
|
buffer = Lex(")({)");
|
|
buffer = Lex(")({)");
|
|
@@ -513,6 +539,7 @@ TEST_F(LexerTest, MismatchedGroups) {
|
|
|
{.kind = TokenKind::OpenCurlyBrace(), .column = 3},
|
|
{.kind = TokenKind::OpenCurlyBrace(), .column = 3},
|
|
|
{.kind = TokenKind::CloseCurlyBrace(), .column = 4, .recovery = true},
|
|
{.kind = TokenKind::CloseCurlyBrace(), .column = 4, .recovery = true},
|
|
|
{.kind = TokenKind::CloseParen(), .column = 4},
|
|
{.kind = TokenKind::CloseParen(), .column = 4},
|
|
|
|
|
+ {TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
}
|
|
}
|
|
|
|
|
|
|
@@ -523,6 +550,7 @@ TEST_F(LexerTest, Keywords) {
|
|
|
buffer,
|
|
buffer,
|
|
|
HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
{.kind = TokenKind::FnKeyword(), .column = 4, .indent_column = 4},
|
|
{.kind = TokenKind::FnKeyword(), .column = 4, .indent_column = 4},
|
|
|
|
|
+ {TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
|
|
|
|
|
buffer = Lex("and or not if else for loop return var break continue _");
|
|
buffer = Lex("and or not if else for loop return var break continue _");
|
|
@@ -540,36 +568,45 @@ TEST_F(LexerTest, Keywords) {
|
|
|
{TokenKind::BreakKeyword()},
|
|
{TokenKind::BreakKeyword()},
|
|
|
{TokenKind::ContinueKeyword()},
|
|
{TokenKind::ContinueKeyword()},
|
|
|
{TokenKind::UnderscoreKeyword()},
|
|
{TokenKind::UnderscoreKeyword()},
|
|
|
|
|
+ {TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
}
|
|
}
|
|
|
|
|
|
|
|
TEST_F(LexerTest, Comments) {
|
|
TEST_F(LexerTest, Comments) {
|
|
|
- auto buffer = Lex(" ;\n // foo\n ;");
|
|
|
|
|
|
|
+ auto buffer = Lex(" ;\n // foo\n ;\n");
|
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
|
- EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
|
|
- {.kind = TokenKind::Semi(),
|
|
|
|
|
- .line = 1,
|
|
|
|
|
- .column = 2,
|
|
|
|
|
- .indent_column = 2},
|
|
|
|
|
- {.kind = TokenKind::Semi(),
|
|
|
|
|
- .line = 3,
|
|
|
|
|
- .column = 3,
|
|
|
|
|
- .indent_column = 3},
|
|
|
|
|
- }));
|
|
|
|
|
|
|
+ EXPECT_THAT(buffer,
|
|
|
|
|
+ HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
|
|
+ {.kind = TokenKind::Semi(),
|
|
|
|
|
+ .line = 1,
|
|
|
|
|
+ .column = 2,
|
|
|
|
|
+ .indent_column = 2},
|
|
|
|
|
+ {.kind = TokenKind::Semi(),
|
|
|
|
|
+ .line = 3,
|
|
|
|
|
+ .column = 3,
|
|
|
|
|
+ .indent_column = 3},
|
|
|
|
|
+ {.kind = TokenKind::EndOfFile(), .line = 3, .column = 4},
|
|
|
|
|
+ }));
|
|
|
|
|
|
|
|
buffer = Lex("// foo\n//\n// bar");
|
|
buffer = Lex("// foo\n//\n// bar");
|
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
|
- EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{}));
|
|
|
|
|
|
|
+ EXPECT_THAT(
|
|
|
|
|
+ buffer,
|
|
|
|
|
+ HasTokens(llvm::ArrayRef<ExpectedToken>{{TokenKind::EndOfFile()}}));
|
|
|
|
|
|
|
|
// Make sure weird characters aren't a problem.
|
|
// Make sure weird characters aren't a problem.
|
|
|
buffer = Lex(" // foo#$!^?@-_💩🍫⃠ [̲̅$̲̅(̲̅ ͡° ͜ʖ ͡°̲̅)̲̅$̲̅]");
|
|
buffer = Lex(" // foo#$!^?@-_💩🍫⃠ [̲̅$̲̅(̲̅ ͡° ͜ʖ ͡°̲̅)̲̅$̲̅]");
|
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
|
- EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{}));
|
|
|
|
|
|
|
+ EXPECT_THAT(
|
|
|
|
|
+ buffer,
|
|
|
|
|
+ HasTokens(llvm::ArrayRef<ExpectedToken>{{TokenKind::EndOfFile()}}));
|
|
|
|
|
|
|
|
// Make sure we can lex a comment at the end of the input.
|
|
// Make sure we can lex a comment at the end of the input.
|
|
|
buffer = Lex("//");
|
|
buffer = Lex("//");
|
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
|
- EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{}));
|
|
|
|
|
|
|
+ EXPECT_THAT(
|
|
|
|
|
+ buffer,
|
|
|
|
|
+ HasTokens(llvm::ArrayRef<ExpectedToken>{{TokenKind::EndOfFile()}}));
|
|
|
}
|
|
}
|
|
|
|
|
|
|
|
TEST_F(LexerTest, InvalidComments) {
|
|
TEST_F(LexerTest, InvalidComments) {
|
|
@@ -593,6 +630,7 @@ TEST_F(LexerTest, Identifiers) {
|
|
|
.column = 4,
|
|
.column = 4,
|
|
|
.indent_column = 4,
|
|
.indent_column = 4,
|
|
|
.text = "foobar"},
|
|
.text = "foobar"},
|
|
|
|
|
+ {TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
|
|
|
|
|
// Check different kinds of identifier character sequences.
|
|
// Check different kinds of identifier character sequences.
|
|
@@ -600,6 +638,7 @@ TEST_F(LexerTest, Identifiers) {
|
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
|
EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
{.kind = TokenKind::Identifier(), .text = "_foo_bar"},
|
|
{.kind = TokenKind::Identifier(), .text = "_foo_bar"},
|
|
|
|
|
+ {TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
|
|
|
|
|
buffer = Lex("foo2bar00");
|
|
buffer = Lex("foo2bar00");
|
|
@@ -607,6 +646,7 @@ TEST_F(LexerTest, Identifiers) {
|
|
|
EXPECT_THAT(buffer,
|
|
EXPECT_THAT(buffer,
|
|
|
HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
{.kind = TokenKind::Identifier(), .text = "foo2bar00"},
|
|
{.kind = TokenKind::Identifier(), .text = "foo2bar00"},
|
|
|
|
|
+ {TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
|
|
|
|
|
// Check that we can parse identifiers that start with a keyword.
|
|
// Check that we can parse identifiers that start with a keyword.
|
|
@@ -614,39 +654,42 @@ TEST_F(LexerTest, Identifiers) {
|
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
|
EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
{.kind = TokenKind::Identifier(), .text = "fnord"},
|
|
{.kind = TokenKind::Identifier(), .text = "fnord"},
|
|
|
|
|
+ {TokenKind::EndOfFile()},
|
|
|
}));
|
|
}));
|
|
|
|
|
|
|
|
// Check multiple identifiers with indent and interning.
|
|
// Check multiple identifiers with indent and interning.
|
|
|
buffer = Lex(" foo;bar\nbar \n foo\tfoo");
|
|
buffer = Lex(" foo;bar\nbar \n foo\tfoo");
|
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
EXPECT_FALSE(buffer.HasErrors());
|
|
|
- EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
|
|
- {.kind = TokenKind::Identifier(),
|
|
|
|
|
- .line = 1,
|
|
|
|
|
- .column = 4,
|
|
|
|
|
- .indent_column = 4,
|
|
|
|
|
- .text = "foo"},
|
|
|
|
|
- {.kind = TokenKind::Semi()},
|
|
|
|
|
- {.kind = TokenKind::Identifier(),
|
|
|
|
|
- .line = 1,
|
|
|
|
|
- .column = 8,
|
|
|
|
|
- .indent_column = 4,
|
|
|
|
|
- .text = "bar"},
|
|
|
|
|
- {.kind = TokenKind::Identifier(),
|
|
|
|
|
- .line = 2,
|
|
|
|
|
- .column = 1,
|
|
|
|
|
- .indent_column = 1,
|
|
|
|
|
- .text = "bar"},
|
|
|
|
|
- {.kind = TokenKind::Identifier(),
|
|
|
|
|
- .line = 3,
|
|
|
|
|
- .column = 3,
|
|
|
|
|
- .indent_column = 3,
|
|
|
|
|
- .text = "foo"},
|
|
|
|
|
- {.kind = TokenKind::Identifier(),
|
|
|
|
|
- .line = 3,
|
|
|
|
|
- .column = 7,
|
|
|
|
|
- .indent_column = 3,
|
|
|
|
|
- .text = "foo"},
|
|
|
|
|
- }));
|
|
|
|
|
|
|
+ EXPECT_THAT(buffer,
|
|
|
|
|
+ HasTokens(llvm::ArrayRef<ExpectedToken>{
|
|
|
|
|
+ {.kind = TokenKind::Identifier(),
|
|
|
|
|
+ .line = 1,
|
|
|
|
|
+ .column = 4,
|
|
|
|
|
+ .indent_column = 4,
|
|
|
|
|
+ .text = "foo"},
|
|
|
|
|
+ {.kind = TokenKind::Semi()},
|
|
|
|
|
+ {.kind = TokenKind::Identifier(),
|
|
|
|
|
+ .line = 1,
|
|
|
|
|
+ .column = 8,
|
|
|
|
|
+ .indent_column = 4,
|
|
|
|
|
+ .text = "bar"},
|
|
|
|
|
+ {.kind = TokenKind::Identifier(),
|
|
|
|
|
+ .line = 2,
|
|
|
|
|
+ .column = 1,
|
|
|
|
|
+ .indent_column = 1,
|
|
|
|
|
+ .text = "bar"},
|
|
|
|
|
+ {.kind = TokenKind::Identifier(),
|
|
|
|
|
+ .line = 3,
|
|
|
|
|
+ .column = 3,
|
|
|
|
|
+ .indent_column = 3,
|
|
|
|
|
+ .text = "foo"},
|
|
|
|
|
+ {.kind = TokenKind::Identifier(),
|
|
|
|
|
+ .line = 3,
|
|
|
|
|
+ .column = 7,
|
|
|
|
|
+ .indent_column = 3,
|
|
|
|
|
+ .text = "foo"},
|
|
|
|
|
+ {.kind = TokenKind::EndOfFile(), .line = 3, .column = 10},
|
|
|
|
|
+ }));
|
|
|
}
|
|
}
|
|
|
|
|
|
|
|
TEST_F(LexerTest, StringLiterals) {
|
|
TEST_F(LexerTest, StringLiterals) {
|
|
@@ -718,6 +761,7 @@ TEST_F(LexerTest, StringLiterals) {
|
|
|
.column = 10,
|
|
.column = 10,
|
|
|
.indent_column = 5,
|
|
.indent_column = 5,
|
|
|
.string_contents = {""}},
|
|
.string_contents = {""}},
|
|
|
|
|
+ {.kind = TokenKind::EndOfFile(), .line = 16, .column = 3},
|
|
|
}));
|
|
}));
|
|
|
}
|
|
}
|
|
|
|
|
|
|
@@ -816,8 +860,11 @@ TEST_F(LexerTest, Printing) {
|
|
|
buffer.Print(print_stream);
|
|
buffer.Print(print_stream);
|
|
|
llvm::StringRef print = print_stream.str();
|
|
llvm::StringRef print = print_stream.str();
|
|
|
EXPECT_THAT(GetAndDropLine(print),
|
|
EXPECT_THAT(GetAndDropLine(print),
|
|
|
- StrEq("token: { index: 0, kind: 'Semi', line: 1, column: 1, "
|
|
|
|
|
|
|
+ StrEq("token: { index: 0, kind: 'Semi', line: 1, column: 1, "
|
|
|
"indent: 1, spelling: ';' }"));
|
|
"indent: 1, spelling: ';' }"));
|
|
|
|
|
+ EXPECT_THAT(GetAndDropLine(print),
|
|
|
|
|
+ StrEq("token: { index: 1, kind: 'EndOfFile', line: 1, column: 2, "
|
|
|
|
|
+ "indent: 1, spelling: '' }"));
|
|
|
EXPECT_TRUE(print.empty()) << print;
|
|
EXPECT_TRUE(print.empty()) << print;
|
|
|
|
|
|
|
|
// Test kind padding.
|
|
// Test kind padding.
|
|
@@ -841,6 +888,9 @@ TEST_F(LexerTest, Printing) {
|
|
|
EXPECT_THAT(GetAndDropLine(print),
|
|
EXPECT_THAT(GetAndDropLine(print),
|
|
|
StrEq("token: { index: 4, kind: 'CloseParen', line: 1, column: "
|
|
StrEq("token: { index: 4, kind: 'CloseParen', line: 1, column: "
|
|
|
"7, indent: 1, spelling: ')', opening_token: 0 }"));
|
|
"7, indent: 1, spelling: ')', opening_token: 0 }"));
|
|
|
|
|
+ EXPECT_THAT(GetAndDropLine(print),
|
|
|
|
|
+ StrEq("token: { index: 5, kind: 'EndOfFile', line: 1, column: "
|
|
|
|
|
+ "8, indent: 1, spelling: '' }"));
|
|
|
EXPECT_TRUE(print.empty()) << print;
|
|
EXPECT_TRUE(print.empty()) << print;
|
|
|
|
|
|
|
|
// Test digit padding with max values of 9, 10, and 11.
|
|
// Test digit padding with max values of 9, 10, and 11.
|
|
@@ -849,15 +899,22 @@ TEST_F(LexerTest, Printing) {
|
|
|
print_storage.clear();
|
|
print_storage.clear();
|
|
|
buffer.Print(print_stream);
|
|
buffer.Print(print_stream);
|
|
|
print = print_stream.str();
|
|
print = print_stream.str();
|
|
|
- EXPECT_THAT(GetAndDropLine(print),
|
|
|
|
|
- StrEq("token: { index: 0, kind: 'Semi', line: 1, column: 1, "
|
|
|
|
|
- "indent: 1, spelling: ';' }"));
|
|
|
|
|
- EXPECT_THAT(GetAndDropLine(print),
|
|
|
|
|
- StrEq("token: { index: 1, kind: 'Semi', line: 11, column: 9, "
|
|
|
|
|
- "indent: 9, spelling: ';' }"));
|
|
|
|
|
- EXPECT_THAT(GetAndDropLine(print),
|
|
|
|
|
- StrEq("token: { index: 2, kind: 'Semi', line: 11, column: 10, "
|
|
|
|
|
- "indent: 9, spelling: ';' }"));
|
|
|
|
|
|
|
+ EXPECT_THAT(
|
|
|
|
|
+ GetAndDropLine(print),
|
|
|
|
|
+ StrEq("token: { index: 0, kind: 'Semi', line: 1, column: 1, "
|
|
|
|
|
+ "indent: 1, spelling: ';' }"));
|
|
|
|
|
+ EXPECT_THAT(
|
|
|
|
|
+ GetAndDropLine(print),
|
|
|
|
|
+ StrEq("token: { index: 1, kind: 'Semi', line: 11, column: 9, "
|
|
|
|
|
+ "indent: 9, spelling: ';' }"));
|
|
|
|
|
+ EXPECT_THAT(
|
|
|
|
|
+ GetAndDropLine(print),
|
|
|
|
|
+ StrEq("token: { index: 2, kind: 'Semi', line: 11, column: 10, "
|
|
|
|
|
+ "indent: 9, spelling: ';' }"));
|
|
|
|
|
+ EXPECT_THAT(
|
|
|
|
|
+ GetAndDropLine(print),
|
|
|
|
|
+ StrEq("token: { index: 3, kind: 'EndOfFile', line: 11, column: 11, "
|
|
|
|
|
+ "indent: 9, spelling: '' }"));
|
|
|
EXPECT_TRUE(print.empty()) << print;
|
|
EXPECT_TRUE(print.empty()) << print;
|
|
|
}
|
|
}
|
|
|
|
|
|
|
@@ -949,6 +1006,29 @@ TEST_F(LexerTest, PrintingAsYaml) {
|
|
|
EXPECT_THAT(&*token_it, IsKeyValueScalars("spelling", ";"));
|
|
EXPECT_THAT(&*token_it, IsKeyValueScalars("spelling", ";"));
|
|
|
EXPECT_THAT(++token_it, Eq(token_value_node->end()));
|
|
EXPECT_THAT(++token_it, Eq(token_value_node->end()));
|
|
|
|
|
|
|
|
|
|
+ ++mapping_it;
|
|
|
|
|
+ token_node = llvm::dyn_cast<llvm::yaml::KeyValueNode>(&*mapping_it);
|
|
|
|
|
+ ASSERT_THAT(token_node, NotNull());
|
|
|
|
|
+ token_key_node = llvm::dyn_cast<llvm::yaml::ScalarNode>(token_node->getKey());
|
|
|
|
|
+ ASSERT_THAT(token_key_node, NotNull());
|
|
|
|
|
+ EXPECT_THAT(token_key_node->getRawValue(), StrEq("token"));
|
|
|
|
|
+ token_value_node =
|
|
|
|
|
+ llvm::dyn_cast<llvm::yaml::MappingNode>(token_node->getValue());
|
|
|
|
|
+ ASSERT_THAT(token_value_node, NotNull());
|
|
|
|
|
+ token_it = token_value_node->begin();
|
|
|
|
|
+ EXPECT_THAT(&*token_it, IsKeyValueScalars("index", "3"));
|
|
|
|
|
+ ++token_it;
|
|
|
|
|
+ EXPECT_THAT(&*token_it, IsKeyValueScalars("kind", "EndOfFile"));
|
|
|
|
|
+ ++token_it;
|
|
|
|
|
+ EXPECT_THAT(&*token_it, IsKeyValueScalars("line", "15"));
|
|
|
|
|
+ ++token_it;
|
|
|
|
|
+ EXPECT_THAT(&*token_it, IsKeyValueScalars("column", "1"));
|
|
|
|
|
+ ++token_it;
|
|
|
|
|
+ EXPECT_THAT(&*token_it, IsKeyValueScalars("indent", "1"));
|
|
|
|
|
+ ++token_it;
|
|
|
|
|
+ EXPECT_THAT(&*token_it, IsKeyValueScalars("spelling", ""));
|
|
|
|
|
+ EXPECT_THAT(++token_it, Eq(token_value_node->end()));
|
|
|
|
|
+
|
|
|
ASSERT_THAT(++mapping_it, Eq(root_node->end()));
|
|
ASSERT_THAT(++mapping_it, Eq(root_node->end()));
|
|
|
ASSERT_THAT(++yaml_it, Eq(yaml_stream.end()));
|
|
ASSERT_THAT(++yaml_it, Eq(yaml_stream.end()));
|
|
|
}
|
|
}
|