tokenized_buffer_test.cpp 41 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100
  1. // Part of the Carbon Language project, under the Apache License v2.0 with LLVM
  2. // Exceptions. See /LICENSE for license information.
  3. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
  4. #include "toolchain/lex/tokenized_buffer.h"
  5. #include <gmock/gmock.h>
  6. #include <gtest/gtest.h>
  7. #include <forward_list>
  8. #include <iterator>
  9. #include "llvm/ADT/ArrayRef.h"
  10. #include "testing/base/test_raw_ostream.h"
  11. #include "toolchain/base/value_store.h"
  12. #include "toolchain/diagnostics/diagnostic_emitter.h"
  13. #include "toolchain/diagnostics/mocks.h"
  14. #include "toolchain/lex/tokenized_buffer_test_helpers.h"
  15. #include "toolchain/testing/yaml_test_helpers.h"
  16. namespace Carbon::Lex {
  17. namespace {
  18. using ::Carbon::Testing::ExpectedToken;
  19. using ::Carbon::Testing::IsDiagnostic;
  20. using ::Carbon::Testing::TestRawOstream;
  21. using ::testing::_;
  22. using ::testing::ElementsAre;
  23. using ::testing::Eq;
  24. using ::testing::HasSubstr;
  25. using ::testing::Pair;
  26. namespace Yaml = ::Carbon::Testing::Yaml;
  27. class LexerTest : public ::testing::Test {
  28. protected:
  29. auto GetSourceBuffer(llvm::StringRef text) -> SourceBuffer& {
  30. std::string filename = llvm::formatv("test{0}.carbon", ++file_index_);
  31. CARBON_CHECK(fs_.addFile(filename, /*ModificationTime=*/0,
  32. llvm::MemoryBuffer::getMemBuffer(text)));
  33. source_storage_.push_front(std::move(*SourceBuffer::CreateFromFile(
  34. fs_, filename, ConsoleDiagnosticConsumer())));
  35. return source_storage_.front();
  36. }
  37. auto Lex(llvm::StringRef text,
  38. DiagnosticConsumer& consumer = ConsoleDiagnosticConsumer())
  39. -> TokenizedBuffer {
  40. return TokenizedBuffer::Lex(value_stores_, GetSourceBuffer(text), consumer);
  41. }
  42. SharedValueStores value_stores_;
  43. llvm::vfs::InMemoryFileSystem fs_;
  44. int file_index_ = 0;
  45. std::forward_list<SourceBuffer> source_storage_;
  46. };
  47. TEST_F(LexerTest, HandlesEmptyBuffer) {
  48. auto buffer = Lex("");
  49. EXPECT_FALSE(buffer.has_errors());
  50. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  51. {TokenKind::StartOfFile}, {TokenKind::EndOfFile}}));
  52. }
  53. TEST_F(LexerTest, TracksLinesAndColumns) {
  54. auto buffer = Lex("\n ;;\n ;;;\n x\"foo\" '''baz\n a\n ''' y");
  55. EXPECT_FALSE(buffer.has_errors());
  56. EXPECT_THAT(
  57. buffer,
  58. HasTokens(llvm::ArrayRef<ExpectedToken>{
  59. {.kind = TokenKind::StartOfFile,
  60. .line = 1,
  61. .column = 1,
  62. .indent_column = 1},
  63. {.kind = TokenKind::Semi, .line = 2, .column = 3, .indent_column = 3},
  64. {.kind = TokenKind::Semi, .line = 2, .column = 4, .indent_column = 3},
  65. {.kind = TokenKind::Semi, .line = 3, .column = 4, .indent_column = 4},
  66. {.kind = TokenKind::Semi, .line = 3, .column = 5, .indent_column = 4},
  67. {.kind = TokenKind::Semi, .line = 3, .column = 6, .indent_column = 4},
  68. {.kind = TokenKind::Identifier,
  69. .line = 4,
  70. .column = 4,
  71. .indent_column = 4,
  72. .text = "x"},
  73. {.kind = TokenKind::StringLiteral,
  74. .line = 4,
  75. .column = 5,
  76. .indent_column = 4},
  77. {.kind = TokenKind::StringLiteral,
  78. .line = 4,
  79. .column = 11,
  80. .indent_column = 4},
  81. {.kind = TokenKind::Identifier,
  82. .line = 6,
  83. .column = 6,
  84. .indent_column = 11,
  85. .text = "y"},
  86. {.kind = TokenKind::EndOfFile, .line = 6, .column = 7},
  87. }));
  88. }
  89. TEST_F(LexerTest, HandlesNumericLiteral) {
  90. auto buffer = Lex("12-578\n 1 2\n0x12_3ABC\n0b10_10_11\n1_234_567\n1.5e9");
  91. EXPECT_FALSE(buffer.has_errors());
  92. ASSERT_THAT(buffer,
  93. HasTokens(llvm::ArrayRef<ExpectedToken>{
  94. {.kind = TokenKind::StartOfFile, .line = 1, .column = 1},
  95. {.kind = TokenKind::IntegerLiteral,
  96. .line = 1,
  97. .column = 1,
  98. .indent_column = 1,
  99. .text = "12"},
  100. {.kind = TokenKind::Minus,
  101. .line = 1,
  102. .column = 3,
  103. .indent_column = 1},
  104. {.kind = TokenKind::IntegerLiteral,
  105. .line = 1,
  106. .column = 4,
  107. .indent_column = 1,
  108. .text = "578"},
  109. {.kind = TokenKind::IntegerLiteral,
  110. .line = 2,
  111. .column = 3,
  112. .indent_column = 3,
  113. .text = "1"},
  114. {.kind = TokenKind::IntegerLiteral,
  115. .line = 2,
  116. .column = 6,
  117. .indent_column = 3,
  118. .text = "2"},
  119. {.kind = TokenKind::IntegerLiteral,
  120. .line = 3,
  121. .column = 1,
  122. .indent_column = 1,
  123. .text = "0x12_3ABC"},
  124. {.kind = TokenKind::IntegerLiteral,
  125. .line = 4,
  126. .column = 1,
  127. .indent_column = 1,
  128. .text = "0b10_10_11"},
  129. {.kind = TokenKind::IntegerLiteral,
  130. .line = 5,
  131. .column = 1,
  132. .indent_column = 1,
  133. .text = "1_234_567"},
  134. {.kind = TokenKind::RealLiteral,
  135. .line = 6,
  136. .column = 1,
  137. .indent_column = 1,
  138. .text = "1.5e9"},
  139. {.kind = TokenKind::EndOfFile, .line = 6, .column = 6},
  140. }));
  141. auto token_start = buffer.tokens().begin();
  142. auto token_12 = token_start + 1;
  143. EXPECT_EQ(value_stores_.integers().Get(buffer.GetIntegerLiteral(*token_12)),
  144. 12);
  145. auto token_578 = token_12 + 2;
  146. EXPECT_EQ(value_stores_.integers().Get(buffer.GetIntegerLiteral(*token_578)),
  147. 578);
  148. auto token_1 = token_578 + 1;
  149. EXPECT_EQ(value_stores_.integers().Get(buffer.GetIntegerLiteral(*token_1)),
  150. 1);
  151. auto token_2 = token_1 + 1;
  152. EXPECT_EQ(value_stores_.integers().Get(buffer.GetIntegerLiteral(*token_2)),
  153. 2);
  154. auto token_0x12_3abc = token_2 + 1;
  155. EXPECT_EQ(
  156. value_stores_.integers().Get(buffer.GetIntegerLiteral(*token_0x12_3abc)),
  157. 0x12'3abc);
  158. auto token_0b10_10_11 = token_0x12_3abc + 1;
  159. EXPECT_EQ(
  160. value_stores_.integers().Get(buffer.GetIntegerLiteral(*token_0b10_10_11)),
  161. 0b10'10'11);
  162. auto token_1_234_567 = token_0b10_10_11 + 1;
  163. EXPECT_EQ(
  164. value_stores_.integers().Get(buffer.GetIntegerLiteral(*token_1_234_567)),
  165. 1'234'567);
  166. auto token_1_5e9 = token_1_234_567 + 1;
  167. auto value_1_5e9 =
  168. value_stores_.reals().Get(buffer.GetRealLiteral(*token_1_5e9));
  169. EXPECT_EQ(value_1_5e9.mantissa.getZExtValue(), 15);
  170. EXPECT_EQ(value_1_5e9.exponent.getSExtValue(), 8);
  171. EXPECT_EQ(value_1_5e9.is_decimal, true);
  172. }
  173. TEST_F(LexerTest, HandlesInvalidNumericLiterals) {
  174. auto buffer = Lex("14x 15_49 0x3.5q 0x3_4.5_6 0ops");
  175. EXPECT_TRUE(buffer.has_errors());
  176. ASSERT_THAT(buffer,
  177. HasTokens(llvm::ArrayRef<ExpectedToken>{
  178. {.kind = TokenKind::StartOfFile, .line = 1, .column = 1},
  179. {.kind = TokenKind::Error,
  180. .line = 1,
  181. .column = 1,
  182. .indent_column = 1,
  183. .text = "14x"},
  184. {.kind = TokenKind::IntegerLiteral,
  185. .line = 1,
  186. .column = 5,
  187. .indent_column = 1,
  188. .text = "15_49"},
  189. {.kind = TokenKind::Error,
  190. .line = 1,
  191. .column = 11,
  192. .indent_column = 1,
  193. .text = "0x3.5q"},
  194. {.kind = TokenKind::RealLiteral,
  195. .line = 1,
  196. .column = 18,
  197. .indent_column = 1,
  198. .text = "0x3_4.5_6"},
  199. {.kind = TokenKind::Error,
  200. .line = 1,
  201. .column = 28,
  202. .indent_column = 1,
  203. .text = "0ops"},
  204. {.kind = TokenKind::EndOfFile, .line = 1, .column = 32},
  205. }));
  206. }
  207. TEST_F(LexerTest, SplitsNumericLiteralsProperly) {
  208. llvm::StringLiteral source_text = R"(
  209. 1.
  210. .2
  211. 3.+foo
  212. 4.0-bar
  213. 5.0e+123+456
  214. 6.0e+1e+2
  215. 1e7
  216. 8..10
  217. 9.0.9.5
  218. 10.foo
  219. 11.0.foo
  220. 12e+1
  221. 13._
  222. )";
  223. auto buffer = Lex(source_text);
  224. EXPECT_TRUE(buffer.has_errors());
  225. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  226. {.kind = TokenKind::StartOfFile},
  227. {.kind = TokenKind::IntegerLiteral, .text = "1"},
  228. {.kind = TokenKind::Period},
  229. // newline
  230. {.kind = TokenKind::Period},
  231. {.kind = TokenKind::IntegerLiteral, .text = "2"},
  232. // newline
  233. {.kind = TokenKind::IntegerLiteral, .text = "3"},
  234. {.kind = TokenKind::Period},
  235. {.kind = TokenKind::Plus},
  236. {.kind = TokenKind::Identifier, .text = "foo"},
  237. // newline
  238. {.kind = TokenKind::RealLiteral, .text = "4.0"},
  239. {.kind = TokenKind::Minus},
  240. {.kind = TokenKind::Identifier, .text = "bar"},
  241. // newline
  242. {.kind = TokenKind::RealLiteral, .text = "5.0e+123"},
  243. {.kind = TokenKind::Plus},
  244. {.kind = TokenKind::IntegerLiteral, .text = "456"},
  245. // newline
  246. {.kind = TokenKind::Error, .text = "6.0e+1e"},
  247. {.kind = TokenKind::Plus},
  248. {.kind = TokenKind::IntegerLiteral, .text = "2"},
  249. // newline
  250. {.kind = TokenKind::Error, .text = "1e7"},
  251. // newline
  252. {.kind = TokenKind::IntegerLiteral, .text = "8"},
  253. {.kind = TokenKind::Period},
  254. {.kind = TokenKind::Period},
  255. {.kind = TokenKind::IntegerLiteral, .text = "10"},
  256. // newline
  257. {.kind = TokenKind::RealLiteral, .text = "9.0"},
  258. {.kind = TokenKind::Period},
  259. {.kind = TokenKind::RealLiteral, .text = "9.5"},
  260. // newline
  261. {.kind = TokenKind::Error, .text = "10.foo"},
  262. // newline
  263. {.kind = TokenKind::RealLiteral, .text = "11.0"},
  264. {.kind = TokenKind::Period},
  265. {.kind = TokenKind::Identifier, .text = "foo"},
  266. // newline
  267. {.kind = TokenKind::Error, .text = "12e"},
  268. {.kind = TokenKind::Plus},
  269. {.kind = TokenKind::IntegerLiteral, .text = "1"},
  270. // newline
  271. {.kind = TokenKind::IntegerLiteral, .text = "13"},
  272. {.kind = TokenKind::Period},
  273. {.kind = TokenKind::Underscore},
  274. // newline
  275. {.kind = TokenKind::EndOfFile},
  276. }));
  277. }
  278. TEST_F(LexerTest, HandlesGarbageCharacters) {
  279. constexpr char GarbageText[] = "$$💩-$\n$\0$12$\n\\\"\\\n\"x";
  280. auto buffer = Lex(llvm::StringRef(GarbageText, sizeof(GarbageText) - 1));
  281. EXPECT_TRUE(buffer.has_errors());
  282. EXPECT_THAT(
  283. buffer,
  284. HasTokens(llvm::ArrayRef<ExpectedToken>{
  285. {.kind = TokenKind::StartOfFile, .line = 1, .column = 1},
  286. {.kind = TokenKind::Error,
  287. .line = 1,
  288. .column = 1,
  289. // 💩 takes 4 bytes, and we count column as bytes offset.
  290. .text = llvm::StringRef("$$💩", 6)},
  291. {.kind = TokenKind::Minus, .line = 1, .column = 7},
  292. {.kind = TokenKind::Error, .line = 1, .column = 8, .text = "$"},
  293. // newline
  294. {.kind = TokenKind::Error,
  295. .line = 2,
  296. .column = 1,
  297. .text = llvm::StringRef("$\0$", 3)},
  298. {.kind = TokenKind::IntegerLiteral,
  299. .line = 2,
  300. .column = 4,
  301. .text = "12"},
  302. {.kind = TokenKind::Error, .line = 2, .column = 6, .text = "$"},
  303. // newline
  304. {.kind = TokenKind::Backslash, .line = 3, .column = 1, .text = "\\"},
  305. {.kind = TokenKind::Error, .line = 3, .column = 2, .text = "\"\\"},
  306. // newline
  307. {.kind = TokenKind::Error, .line = 4, .column = 1, .text = "\"x"},
  308. {.kind = TokenKind::EndOfFile, .line = 4, .column = 3},
  309. }));
  310. }
  311. TEST_F(LexerTest, Symbols) {
  312. // We don't need to exhaustively test symbols here as they're handled with
  313. // common code, but we want to check specific patterns to verify things like
  314. // max-munch rule and handling of interesting symbols.
  315. auto buffer = Lex("<<<");
  316. EXPECT_FALSE(buffer.has_errors());
  317. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  318. {TokenKind::StartOfFile},
  319. {TokenKind::LessLess},
  320. {TokenKind::Less},
  321. {TokenKind::EndOfFile},
  322. }));
  323. buffer = Lex("<<=>>");
  324. EXPECT_FALSE(buffer.has_errors());
  325. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  326. {TokenKind::StartOfFile},
  327. {TokenKind::LessLessEqual},
  328. {TokenKind::GreaterGreater},
  329. {TokenKind::EndOfFile},
  330. }));
  331. buffer = Lex("< <=> >");
  332. EXPECT_FALSE(buffer.has_errors());
  333. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  334. {TokenKind::StartOfFile},
  335. {TokenKind::Less},
  336. {TokenKind::LessEqualGreater},
  337. {TokenKind::Greater},
  338. {TokenKind::EndOfFile},
  339. }));
  340. buffer = Lex("\\/?@&^!");
  341. EXPECT_FALSE(buffer.has_errors());
  342. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  343. {TokenKind::StartOfFile},
  344. {TokenKind::Backslash},
  345. {TokenKind::Slash},
  346. {TokenKind::Question},
  347. {TokenKind::At},
  348. {TokenKind::Amp},
  349. {TokenKind::Caret},
  350. {TokenKind::Exclaim},
  351. {TokenKind::EndOfFile},
  352. }));
  353. }
  354. TEST_F(LexerTest, Parens) {
  355. auto buffer = Lex("()");
  356. EXPECT_FALSE(buffer.has_errors());
  357. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  358. {TokenKind::StartOfFile},
  359. {TokenKind::OpenParen},
  360. {TokenKind::CloseParen},
  361. {TokenKind::EndOfFile},
  362. }));
  363. buffer = Lex("((()()))");
  364. EXPECT_FALSE(buffer.has_errors());
  365. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  366. {TokenKind::StartOfFile},
  367. {TokenKind::OpenParen},
  368. {TokenKind::OpenParen},
  369. {TokenKind::OpenParen},
  370. {TokenKind::CloseParen},
  371. {TokenKind::OpenParen},
  372. {TokenKind::CloseParen},
  373. {TokenKind::CloseParen},
  374. {TokenKind::CloseParen},
  375. {TokenKind::EndOfFile},
  376. }));
  377. }
  378. TEST_F(LexerTest, CurlyBraces) {
  379. auto buffer = Lex("{}");
  380. EXPECT_FALSE(buffer.has_errors());
  381. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  382. {TokenKind::StartOfFile},
  383. {TokenKind::OpenCurlyBrace},
  384. {TokenKind::CloseCurlyBrace},
  385. {TokenKind::EndOfFile},
  386. }));
  387. buffer = Lex("{{{}{}}}");
  388. EXPECT_FALSE(buffer.has_errors());
  389. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  390. {TokenKind::StartOfFile},
  391. {TokenKind::OpenCurlyBrace},
  392. {TokenKind::OpenCurlyBrace},
  393. {TokenKind::OpenCurlyBrace},
  394. {TokenKind::CloseCurlyBrace},
  395. {TokenKind::OpenCurlyBrace},
  396. {TokenKind::CloseCurlyBrace},
  397. {TokenKind::CloseCurlyBrace},
  398. {TokenKind::CloseCurlyBrace},
  399. {TokenKind::EndOfFile},
  400. }));
  401. }
  402. TEST_F(LexerTest, MatchingGroups) {
  403. {
  404. TokenizedBuffer buffer = Lex("(){}");
  405. ASSERT_FALSE(buffer.has_errors());
  406. auto it = ++buffer.tokens().begin();
  407. auto open_paren_token = *it++;
  408. auto close_paren_token = *it++;
  409. EXPECT_EQ(close_paren_token,
  410. buffer.GetMatchedClosingToken(open_paren_token));
  411. EXPECT_EQ(open_paren_token,
  412. buffer.GetMatchedOpeningToken(close_paren_token));
  413. auto open_curly_token = *it++;
  414. auto close_curly_token = *it++;
  415. EXPECT_EQ(close_curly_token,
  416. buffer.GetMatchedClosingToken(open_curly_token));
  417. EXPECT_EQ(open_curly_token,
  418. buffer.GetMatchedOpeningToken(close_curly_token));
  419. auto eof_token = *it++;
  420. EXPECT_EQ(buffer.GetKind(eof_token), TokenKind::EndOfFile);
  421. EXPECT_EQ(buffer.tokens().end(), it);
  422. }
  423. {
  424. TokenizedBuffer buffer = Lex("({x}){(y)} {{((z))}}");
  425. ASSERT_FALSE(buffer.has_errors());
  426. auto it = ++buffer.tokens().begin();
  427. auto open_paren_token = *it++;
  428. auto open_curly_token = *it++;
  429. ASSERT_EQ("x", value_stores_.strings().Get(buffer.GetIdentifier(*it++)));
  430. auto close_curly_token = *it++;
  431. auto close_paren_token = *it++;
  432. EXPECT_EQ(close_paren_token,
  433. buffer.GetMatchedClosingToken(open_paren_token));
  434. EXPECT_EQ(open_paren_token,
  435. buffer.GetMatchedOpeningToken(close_paren_token));
  436. EXPECT_EQ(close_curly_token,
  437. buffer.GetMatchedClosingToken(open_curly_token));
  438. EXPECT_EQ(open_curly_token,
  439. buffer.GetMatchedOpeningToken(close_curly_token));
  440. open_curly_token = *it++;
  441. open_paren_token = *it++;
  442. ASSERT_EQ("y", value_stores_.strings().Get(buffer.GetIdentifier(*it++)));
  443. close_paren_token = *it++;
  444. close_curly_token = *it++;
  445. EXPECT_EQ(close_curly_token,
  446. buffer.GetMatchedClosingToken(open_curly_token));
  447. EXPECT_EQ(open_curly_token,
  448. buffer.GetMatchedOpeningToken(close_curly_token));
  449. EXPECT_EQ(close_paren_token,
  450. buffer.GetMatchedClosingToken(open_paren_token));
  451. EXPECT_EQ(open_paren_token,
  452. buffer.GetMatchedOpeningToken(close_paren_token));
  453. open_curly_token = *it++;
  454. auto inner_open_curly_token = *it++;
  455. open_paren_token = *it++;
  456. auto inner_open_paren_token = *it++;
  457. ASSERT_EQ("z", value_stores_.strings().Get(buffer.GetIdentifier(*it++)));
  458. auto inner_close_paren_token = *it++;
  459. close_paren_token = *it++;
  460. auto inner_close_curly_token = *it++;
  461. close_curly_token = *it++;
  462. EXPECT_EQ(close_curly_token,
  463. buffer.GetMatchedClosingToken(open_curly_token));
  464. EXPECT_EQ(open_curly_token,
  465. buffer.GetMatchedOpeningToken(close_curly_token));
  466. EXPECT_EQ(inner_close_curly_token,
  467. buffer.GetMatchedClosingToken(inner_open_curly_token));
  468. EXPECT_EQ(inner_open_curly_token,
  469. buffer.GetMatchedOpeningToken(inner_close_curly_token));
  470. EXPECT_EQ(close_paren_token,
  471. buffer.GetMatchedClosingToken(open_paren_token));
  472. EXPECT_EQ(open_paren_token,
  473. buffer.GetMatchedOpeningToken(close_paren_token));
  474. EXPECT_EQ(inner_close_paren_token,
  475. buffer.GetMatchedClosingToken(inner_open_paren_token));
  476. EXPECT_EQ(inner_open_paren_token,
  477. buffer.GetMatchedOpeningToken(inner_close_paren_token));
  478. auto eof_token = *it++;
  479. EXPECT_EQ(buffer.GetKind(eof_token), TokenKind::EndOfFile);
  480. EXPECT_EQ(buffer.tokens().end(), it);
  481. }
  482. }
  483. TEST_F(LexerTest, MismatchedGroups) {
  484. auto buffer = Lex("{");
  485. EXPECT_TRUE(buffer.has_errors());
  486. EXPECT_THAT(buffer,
  487. HasTokens(llvm::ArrayRef<ExpectedToken>{
  488. {TokenKind::StartOfFile},
  489. {TokenKind::OpenCurlyBrace},
  490. {.kind = TokenKind::CloseCurlyBrace, .recovery = true},
  491. {TokenKind::EndOfFile},
  492. }));
  493. buffer = Lex("}");
  494. EXPECT_TRUE(buffer.has_errors());
  495. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  496. {TokenKind::StartOfFile},
  497. {.kind = TokenKind::Error, .text = "}"},
  498. {TokenKind::EndOfFile},
  499. }));
  500. buffer = Lex("{(}");
  501. EXPECT_TRUE(buffer.has_errors());
  502. EXPECT_THAT(
  503. buffer,
  504. HasTokens(llvm::ArrayRef<ExpectedToken>{
  505. {TokenKind::StartOfFile},
  506. {.kind = TokenKind::OpenCurlyBrace, .column = 1},
  507. {.kind = TokenKind::OpenParen, .column = 2},
  508. {.kind = TokenKind::CloseParen, .column = 3, .recovery = true},
  509. {.kind = TokenKind::CloseCurlyBrace, .column = 3},
  510. {TokenKind::EndOfFile},
  511. }));
  512. buffer = Lex(")({)");
  513. EXPECT_TRUE(buffer.has_errors());
  514. EXPECT_THAT(
  515. buffer,
  516. HasTokens(llvm::ArrayRef<ExpectedToken>{
  517. {TokenKind::StartOfFile},
  518. {.kind = TokenKind::Error, .column = 1, .text = ")"},
  519. {.kind = TokenKind::OpenParen, .column = 2},
  520. {.kind = TokenKind::OpenCurlyBrace, .column = 3},
  521. {.kind = TokenKind::CloseCurlyBrace, .column = 4, .recovery = true},
  522. {.kind = TokenKind::CloseParen, .column = 4},
  523. {TokenKind::EndOfFile},
  524. }));
  525. }
  526. TEST_F(LexerTest, Whitespace) {
  527. auto buffer = Lex("{( } {(");
  528. // Whether there should be whitespace before/after each token.
  529. bool space[] = {true,
  530. // start-of-file
  531. true,
  532. // {
  533. false,
  534. // (
  535. true,
  536. // inserted )
  537. true,
  538. // }
  539. true,
  540. // {
  541. false,
  542. // (
  543. true,
  544. // inserted )
  545. true,
  546. // inserted }
  547. true,
  548. // EOF
  549. false};
  550. int pos = 0;
  551. for (Token token : buffer.tokens()) {
  552. SCOPED_TRACE(
  553. llvm::formatv("Token #{0}: '{1}'", token, buffer.GetTokenText(token)));
  554. ASSERT_LT(pos, std::size(space));
  555. EXPECT_THAT(buffer.HasLeadingWhitespace(token), Eq(space[pos]));
  556. ++pos;
  557. ASSERT_LT(pos, std::size(space));
  558. EXPECT_THAT(buffer.HasTrailingWhitespace(token), Eq(space[pos]));
  559. }
  560. ASSERT_EQ(pos + 1, std::size(space));
  561. }
  562. TEST_F(LexerTest, Keywords) {
  563. auto buffer = Lex(" fn");
  564. EXPECT_FALSE(buffer.has_errors());
  565. EXPECT_THAT(buffer,
  566. HasTokens(llvm::ArrayRef<ExpectedToken>{
  567. {TokenKind::StartOfFile},
  568. {.kind = TokenKind::Fn, .column = 4, .indent_column = 4},
  569. {TokenKind::EndOfFile},
  570. }));
  571. buffer = Lex("and or not if else for return var break continue _");
  572. EXPECT_FALSE(buffer.has_errors());
  573. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  574. {TokenKind::StartOfFile},
  575. {TokenKind::And},
  576. {TokenKind::Or},
  577. {TokenKind::Not},
  578. {TokenKind::If},
  579. {TokenKind::Else},
  580. {TokenKind::For},
  581. {TokenKind::Return},
  582. {TokenKind::Var},
  583. {TokenKind::Break},
  584. {TokenKind::Continue},
  585. {TokenKind::Underscore},
  586. {TokenKind::EndOfFile},
  587. }));
  588. }
  589. TEST_F(LexerTest, Comments) {
  590. auto buffer = Lex(" ;\n // foo\n ;\n");
  591. EXPECT_FALSE(buffer.has_errors());
  592. EXPECT_THAT(
  593. buffer,
  594. HasTokens(llvm::ArrayRef<ExpectedToken>{
  595. {.kind = TokenKind::StartOfFile, .line = 1, .column = 1},
  596. {.kind = TokenKind::Semi, .line = 1, .column = 2, .indent_column = 2},
  597. {.kind = TokenKind::Semi, .line = 3, .column = 3, .indent_column = 3},
  598. {.kind = TokenKind::EndOfFile, .line = 3, .column = 4},
  599. }));
  600. buffer = Lex("// foo\n//\n// bar");
  601. EXPECT_FALSE(buffer.has_errors());
  602. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  603. {TokenKind::StartOfFile}, {TokenKind::EndOfFile}}));
  604. // Make sure weird characters aren't a problem.
  605. buffer = Lex(" // foo#$!^?@-_💩🍫⃠ [̲̅$̲̅(̲̅ ͡° ͜ʖ ͡°̲̅)̲̅$̲̅]");
  606. EXPECT_FALSE(buffer.has_errors());
  607. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  608. {TokenKind::StartOfFile}, {TokenKind::EndOfFile}}));
  609. // Make sure we can lex a comment at the end of the input.
  610. buffer = Lex("//");
  611. EXPECT_FALSE(buffer.has_errors());
  612. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  613. {TokenKind::StartOfFile}, {TokenKind::EndOfFile}}));
  614. }
  615. TEST_F(LexerTest, InvalidComments) {
  616. llvm::StringLiteral testcases[] = {
  617. " /// foo\n",
  618. "foo // bar\n",
  619. "//! hello",
  620. " //world",
  621. };
  622. for (llvm::StringLiteral testcase : testcases) {
  623. auto buffer = Lex(testcase);
  624. EXPECT_TRUE(buffer.has_errors());
  625. }
  626. }
  627. TEST_F(LexerTest, Identifiers) {
  628. auto buffer = Lex(" foobar");
  629. EXPECT_FALSE(buffer.has_errors());
  630. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  631. {TokenKind::StartOfFile},
  632. {.kind = TokenKind::Identifier,
  633. .column = 4,
  634. .indent_column = 4,
  635. .text = "foobar"},
  636. {TokenKind::EndOfFile},
  637. }));
  638. // Check different kinds of identifier character sequences.
  639. buffer = Lex("_foo_bar");
  640. EXPECT_FALSE(buffer.has_errors());
  641. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  642. {TokenKind::StartOfFile},
  643. {.kind = TokenKind::Identifier, .text = "_foo_bar"},
  644. {TokenKind::EndOfFile},
  645. }));
  646. buffer = Lex("foo2bar00");
  647. EXPECT_FALSE(buffer.has_errors());
  648. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  649. {TokenKind::StartOfFile},
  650. {.kind = TokenKind::Identifier, .text = "foo2bar00"},
  651. {TokenKind::EndOfFile},
  652. }));
  653. // Check that we can parse identifiers that start with a keyword.
  654. buffer = Lex("fnord");
  655. EXPECT_FALSE(buffer.has_errors());
  656. EXPECT_THAT(buffer, HasTokens(llvm::ArrayRef<ExpectedToken>{
  657. {TokenKind::StartOfFile},
  658. {.kind = TokenKind::Identifier, .text = "fnord"},
  659. {TokenKind::EndOfFile},
  660. }));
  661. // Check multiple identifiers with indent and interning.
  662. buffer = Lex(" foo;bar\nbar \n foo\tfoo");
  663. EXPECT_FALSE(buffer.has_errors());
  664. EXPECT_THAT(buffer,
  665. HasTokens(llvm::ArrayRef<ExpectedToken>{
  666. {.kind = TokenKind::StartOfFile, .line = 1, .column = 1},
  667. {.kind = TokenKind::Identifier,
  668. .line = 1,
  669. .column = 4,
  670. .indent_column = 4,
  671. .text = "foo"},
  672. {.kind = TokenKind::Semi},
  673. {.kind = TokenKind::Identifier,
  674. .line = 1,
  675. .column = 8,
  676. .indent_column = 4,
  677. .text = "bar"},
  678. {.kind = TokenKind::Identifier,
  679. .line = 2,
  680. .column = 1,
  681. .indent_column = 1,
  682. .text = "bar"},
  683. {.kind = TokenKind::Identifier,
  684. .line = 3,
  685. .column = 3,
  686. .indent_column = 3,
  687. .text = "foo"},
  688. {.kind = TokenKind::Identifier,
  689. .line = 3,
  690. .column = 7,
  691. .indent_column = 3,
  692. .text = "foo"},
  693. {.kind = TokenKind::EndOfFile, .line = 3, .column = 10},
  694. }));
  695. }
  696. TEST_F(LexerTest, StringLiterals) {
  697. llvm::StringLiteral testcase = R"(
  698. "hello world\n"
  699. '''foo
  700. test \
  701. \xAB
  702. ''' trailing
  703. #"""#
  704. "\0"
  705. #"\0"foo"\1"#
  706. """x"""
  707. )";
  708. auto buffer = Lex(testcase);
  709. EXPECT_FALSE(buffer.has_errors());
  710. EXPECT_THAT(buffer,
  711. HasTokens(llvm::ArrayRef<ExpectedToken>{
  712. {.kind = TokenKind::StartOfFile, .line = 1, .column = 1},
  713. {.kind = TokenKind::StringLiteral,
  714. .line = 2,
  715. .column = 5,
  716. .indent_column = 5,
  717. .value_stores = &value_stores_,
  718. .string_contents = {"hello world\n"}},
  719. {.kind = TokenKind::StringLiteral,
  720. .line = 4,
  721. .column = 5,
  722. .indent_column = 5,
  723. .value_stores = &value_stores_,
  724. .string_contents = {" test \xAB\n"}},
  725. {.kind = TokenKind::Identifier,
  726. .line = 7,
  727. .column = 10,
  728. .indent_column = 5,
  729. .text = "trailing"},
  730. {.kind = TokenKind::StringLiteral,
  731. .line = 9,
  732. .column = 7,
  733. .indent_column = 7,
  734. .value_stores = &value_stores_,
  735. .string_contents = {"\""}},
  736. {.kind = TokenKind::StringLiteral,
  737. .line = 11,
  738. .column = 5,
  739. .indent_column = 5,
  740. .value_stores = &value_stores_,
  741. .string_contents = llvm::StringLiteral::withInnerNUL("\0")},
  742. {.kind = TokenKind::StringLiteral,
  743. .line = 13,
  744. .column = 5,
  745. .indent_column = 5,
  746. .value_stores = &value_stores_,
  747. .string_contents = {"\\0\"foo\"\\1"}},
  748. // """x""" is three string literals, not one invalid
  749. // attempt at a block string literal.
  750. {.kind = TokenKind::StringLiteral,
  751. .line = 15,
  752. .column = 5,
  753. .indent_column = 5,
  754. .value_stores = &value_stores_,
  755. .string_contents = {""}},
  756. {.kind = TokenKind::StringLiteral,
  757. .line = 15,
  758. .column = 7,
  759. .indent_column = 5,
  760. .value_stores = &value_stores_,
  761. .string_contents = {"x"}},
  762. {.kind = TokenKind::StringLiteral,
  763. .line = 15,
  764. .column = 10,
  765. .indent_column = 5,
  766. .value_stores = &value_stores_,
  767. .string_contents = {""}},
  768. {.kind = TokenKind::EndOfFile, .line = 16, .column = 3},
  769. }));
  770. }
  771. TEST_F(LexerTest, InvalidStringLiterals) {
  772. llvm::StringLiteral invalid[] = {
  773. // clang-format off
  774. R"(")",
  775. R"('''
  776. '')",
  777. R"("\)",
  778. R"("\")",
  779. R"("\\)",
  780. R"("\\\")",
  781. R"(''')",
  782. R"('''
  783. )",
  784. R"('''\)",
  785. R"(#'''
  786. ''')",
  787. // clang-format on
  788. };
  789. for (llvm::StringLiteral test : invalid) {
  790. SCOPED_TRACE(test);
  791. auto buffer = Lex(test);
  792. EXPECT_TRUE(buffer.has_errors());
  793. // We should have formed at least one error token.
  794. bool found_error = false;
  795. for (Token token : buffer.tokens()) {
  796. if (buffer.GetKind(token) == TokenKind::Error) {
  797. found_error = true;
  798. break;
  799. }
  800. }
  801. EXPECT_TRUE(found_error);
  802. }
  803. }
  804. TEST_F(LexerTest, TypeLiterals) {
  805. llvm::StringLiteral testcase = R"(
  806. i0 i1 i20 i999999999999 i0x1
  807. u0 u1 u64 u64b
  808. f32 f80 f1 fi
  809. s1
  810. )";
  811. auto buffer = Lex(testcase);
  812. EXPECT_FALSE(buffer.has_errors());
  813. ASSERT_THAT(buffer,
  814. HasTokens(llvm::ArrayRef<ExpectedToken>{
  815. {.kind = TokenKind::StartOfFile, .line = 1, .column = 1},
  816. {.kind = TokenKind::Identifier,
  817. .line = 2,
  818. .column = 5,
  819. .indent_column = 5,
  820. .text = {"i0"}},
  821. {.kind = TokenKind::IntegerTypeLiteral,
  822. .line = 2,
  823. .column = 8,
  824. .indent_column = 5,
  825. .text = {"i1"}},
  826. {.kind = TokenKind::IntegerTypeLiteral,
  827. .line = 2,
  828. .column = 11,
  829. .indent_column = 5,
  830. .text = {"i20"}},
  831. {.kind = TokenKind::IntegerTypeLiteral,
  832. .line = 2,
  833. .column = 15,
  834. .indent_column = 5,
  835. .text = {"i999999999999"}},
  836. {.kind = TokenKind::Identifier,
  837. .line = 2,
  838. .column = 29,
  839. .indent_column = 5,
  840. .text = {"i0x1"}},
  841. {.kind = TokenKind::Identifier,
  842. .line = 3,
  843. .column = 5,
  844. .indent_column = 5,
  845. .text = {"u0"}},
  846. {.kind = TokenKind::UnsignedIntegerTypeLiteral,
  847. .line = 3,
  848. .column = 8,
  849. .indent_column = 5,
  850. .text = {"u1"}},
  851. {.kind = TokenKind::UnsignedIntegerTypeLiteral,
  852. .line = 3,
  853. .column = 11,
  854. .indent_column = 5,
  855. .text = {"u64"}},
  856. {.kind = TokenKind::Identifier,
  857. .line = 3,
  858. .column = 15,
  859. .indent_column = 5,
  860. .text = {"u64b"}},
  861. {.kind = TokenKind::FloatingPointTypeLiteral,
  862. .line = 4,
  863. .column = 5,
  864. .indent_column = 5,
  865. .text = {"f32"}},
  866. {.kind = TokenKind::FloatingPointTypeLiteral,
  867. .line = 4,
  868. .column = 9,
  869. .indent_column = 5,
  870. .text = {"f80"}},
  871. {.kind = TokenKind::FloatingPointTypeLiteral,
  872. .line = 4,
  873. .column = 13,
  874. .indent_column = 5,
  875. .text = {"f1"}},
  876. {.kind = TokenKind::Identifier,
  877. .line = 4,
  878. .column = 16,
  879. .indent_column = 5,
  880. .text = {"fi"}},
  881. {.kind = TokenKind::Identifier,
  882. .line = 5,
  883. .column = 5,
  884. .indent_column = 5,
  885. .text = {"s1"}},
  886. {.kind = TokenKind::EndOfFile, .line = 6, .column = 3},
  887. }));
  888. auto token_i1 = buffer.tokens().begin() + 2;
  889. EXPECT_EQ(buffer.GetTypeLiteralSize(*token_i1), 1);
  890. auto token_i20 = buffer.tokens().begin() + 3;
  891. EXPECT_EQ(buffer.GetTypeLiteralSize(*token_i20), 20);
  892. auto token_i999999999999 = buffer.tokens().begin() + 4;
  893. EXPECT_EQ(buffer.GetTypeLiteralSize(*token_i999999999999), 999999999999ULL);
  894. auto token_u1 = buffer.tokens().begin() + 7;
  895. EXPECT_EQ(buffer.GetTypeLiteralSize(*token_u1), 1);
  896. auto token_u64 = buffer.tokens().begin() + 8;
  897. EXPECT_EQ(buffer.GetTypeLiteralSize(*token_u64), 64);
  898. auto token_f32 = buffer.tokens().begin() + 10;
  899. EXPECT_EQ(buffer.GetTypeLiteralSize(*token_f32), 32);
  900. auto token_f80 = buffer.tokens().begin() + 11;
  901. EXPECT_EQ(buffer.GetTypeLiteralSize(*token_f80), 80);
  902. auto token_f1 = buffer.tokens().begin() + 12;
  903. EXPECT_EQ(buffer.GetTypeLiteralSize(*token_f1), 1);
  904. }
  905. TEST_F(LexerTest, TypeLiteralTooManyDigits) {
  906. std::string code = "i";
  907. constexpr int Count = 10000;
  908. code.append(Count, '9');
  909. Testing::MockDiagnosticConsumer consumer;
  910. EXPECT_CALL(consumer,
  911. HandleDiagnostic(IsDiagnostic(
  912. DiagnosticKind::TooManyDigits, DiagnosticLevel::Error, 1, 2,
  913. HasSubstr(llvm::formatv(" {0} ", Count)))));
  914. auto buffer = Lex(code, consumer);
  915. EXPECT_TRUE(buffer.has_errors());
  916. ASSERT_THAT(
  917. buffer,
  918. HasTokens(llvm::ArrayRef<ExpectedToken>{
  919. {.kind = TokenKind::StartOfFile, .line = 1, .column = 1},
  920. {.kind = TokenKind::Error,
  921. .line = 1,
  922. .column = 1,
  923. .indent_column = 1,
  924. .text = {code}},
  925. {.kind = TokenKind::EndOfFile, .line = 1, .column = Count + 2},
  926. }));
  927. }
  928. TEST_F(LexerTest, DiagnosticTrailingComment) {
  929. llvm::StringLiteral testcase = R"(
  930. // Hello!
  931. var String x; // trailing comment
  932. )";
  933. Testing::MockDiagnosticConsumer consumer;
  934. EXPECT_CALL(consumer,
  935. HandleDiagnostic(IsDiagnostic(DiagnosticKind::TrailingComment,
  936. DiagnosticLevel::Error, 3, 19, _)));
  937. Lex(testcase, consumer);
  938. }
  939. TEST_F(LexerTest, DiagnosticWhitespace) {
  940. Testing::MockDiagnosticConsumer consumer;
  941. EXPECT_CALL(consumer, HandleDiagnostic(IsDiagnostic(
  942. DiagnosticKind::NoWhitespaceAfterCommentIntroducer,
  943. DiagnosticLevel::Error, 1, 3, _)));
  944. Lex("//no space after comment", consumer);
  945. }
  946. TEST_F(LexerTest, DiagnosticUnrecognizedEscape) {
  947. Testing::MockDiagnosticConsumer consumer;
  948. EXPECT_CALL(consumer, HandleDiagnostic(IsDiagnostic(
  949. DiagnosticKind::UnknownEscapeSequence,
  950. DiagnosticLevel::Error, 1, 8, HasSubstr("`b`"))));
  951. Lex(R"("hello\bworld")", consumer);
  952. }
  953. TEST_F(LexerTest, DiagnosticBadHex) {
  954. Testing::MockDiagnosticConsumer consumer;
  955. EXPECT_CALL(consumer, HandleDiagnostic(IsDiagnostic(
  956. DiagnosticKind::HexadecimalEscapeMissingDigits,
  957. DiagnosticLevel::Error, 1, 9, _)));
  958. Lex(R"("hello\xabworld")", consumer);
  959. }
  960. TEST_F(LexerTest, DiagnosticInvalidDigit) {
  961. Testing::MockDiagnosticConsumer consumer;
  962. EXPECT_CALL(consumer, HandleDiagnostic(IsDiagnostic(
  963. DiagnosticKind::InvalidDigit,
  964. DiagnosticLevel::Error, 1, 6, HasSubstr("'a'"))));
  965. Lex("0x123abc", consumer);
  966. }
  967. TEST_F(LexerTest, DiagnosticMissingTerminator) {
  968. Testing::MockDiagnosticConsumer consumer;
  969. EXPECT_CALL(consumer,
  970. HandleDiagnostic(IsDiagnostic(DiagnosticKind::UnterminatedString,
  971. DiagnosticLevel::Error, 1, 1, _)));
  972. Lex(R"(#" ")", consumer);
  973. }
  974. TEST_F(LexerTest, DiagnosticUnrecognizedChar) {
  975. Testing::MockDiagnosticConsumer consumer;
  976. EXPECT_CALL(consumer, HandleDiagnostic(
  977. IsDiagnostic(DiagnosticKind::UnrecognizedCharacters,
  978. DiagnosticLevel::Error, 1, 1, _)));
  979. Lex("\b", consumer);
  980. }
  981. TEST_F(LexerTest, PrintingOutputYaml) {
  982. // Test that we can parse this into YAML and verify line and indent data.
  983. auto buffer = Lex("\n ;\n\n\n; ;\n\n\n\n\n\n\n\n\n\n\n");
  984. ASSERT_FALSE(buffer.has_errors());
  985. TestRawOstream print_stream;
  986. buffer.Print(print_stream);
  987. EXPECT_THAT(
  988. Yaml::Value::FromText(print_stream.TakeStr()),
  989. IsYaml(ElementsAre(Yaml::Sequence(ElementsAre(Yaml::Mapping(ElementsAre(
  990. Pair("filename", source_storage_.front().filename().str()),
  991. Pair("tokens",
  992. Yaml::Sequence(ElementsAre(
  993. Yaml::Mapping(ElementsAre(
  994. Pair("index", "0"), Pair("kind", "StartOfFile"),
  995. Pair("line", "1"), Pair("column", "1"),
  996. Pair("indent", "1"), Pair("spelling", ""),
  997. Pair("has_trailing_space", "true"))),
  998. Yaml::Mapping(
  999. ElementsAre(Pair("index", "1"), Pair("kind", "Semi"),
  1000. Pair("line", "2"), Pair("column", "2"),
  1001. Pair("indent", "2"), Pair("spelling", ";"),
  1002. Pair("has_trailing_space", "true"))),
  1003. Yaml::Mapping(
  1004. ElementsAre(Pair("index", "2"), Pair("kind", "Semi"),
  1005. Pair("line", "5"), Pair("column", "1"),
  1006. Pair("indent", "1"), Pair("spelling", ";"),
  1007. Pair("has_trailing_space", "true"))),
  1008. Yaml::Mapping(
  1009. ElementsAre(Pair("index", "3"), Pair("kind", "Semi"),
  1010. Pair("line", "5"), Pair("column", "3"),
  1011. Pair("indent", "1"), Pair("spelling", ";"),
  1012. Pair("has_trailing_space", "true"))),
  1013. Yaml::Mapping(ElementsAre(
  1014. Pair("index", "4"), Pair("kind", "EndOfFile"),
  1015. Pair("line", "15"), Pair("column", "1"),
  1016. Pair("indent", "1"), Pair("spelling", "")))))))))))));
  1017. }
  1018. } // namespace
  1019. } // namespace Carbon::Lex