// Copyright (c) 2013 The Chromium Authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #include "testing/gtest/include/gtest/gtest.h" #include "tools/gn/input_file.h" #include "tools/gn/token.h" #include "tools/gn/tokenizer.h" namespace { struct TokenExpectation { Token::Type type; const char* value; }; template bool CheckTokenizer(const char* input, const TokenExpectation (&expect)[len]) { InputFile input_file(SourceFile("/test")); input_file.SetContents(input); Err err; std::vector results = Tokenizer::Tokenize(&input_file, &err); if (results.size() != len) return false; for (size_t i = 0; i < len; i++) { if (expect[i].type != results[i].type()) return false; if (expect[i].value != results[i].value()) return false; } return true; } } // namespace TEST(Tokenizer, Empty) { InputFile empty_string_input(SourceFile("/test")); empty_string_input.SetContents(""); Err err; std::vector results = Tokenizer::Tokenize(&empty_string_input, &err); EXPECT_TRUE(results.empty()); InputFile whitespace_input(SourceFile("/test")); whitespace_input.SetContents(" \r \n \r\n"); results = Tokenizer::Tokenize(&whitespace_input, &err); EXPECT_TRUE(results.empty()); } TEST(Tokenizer, Identifier) { TokenExpectation one_ident[] = { { Token::IDENTIFIER, "foo" } }; EXPECT_TRUE(CheckTokenizer(" foo ", one_ident)); } TEST(Tokenizer, Integer) { TokenExpectation integers[] = { { Token::INTEGER, "123" }, { Token::INTEGER, "-123" } }; EXPECT_TRUE(CheckTokenizer(" 123 -123 ", integers)); } TEST(Tokenizer, IntegerNoSpace) { TokenExpectation integers[] = { { Token::INTEGER, "123" }, { Token::INTEGER, "-123" } }; EXPECT_TRUE(CheckTokenizer(" 123-123 ", integers)); } TEST(Tokenizer, String) { TokenExpectation strings[] = { { Token::STRING, "\"foo\"" }, { Token::STRING, "\"bar\\\"baz\"" }, { Token::STRING, "\"asdf\\\\\"" } }; EXPECT_TRUE(CheckTokenizer(" \"foo\" \"bar\\\"baz\" \"asdf\\\\\" ", strings)); } TEST(Tokenizer, Operator) { TokenExpectation operators[] = { { Token::MINUS, "-" }, { Token::PLUS, "+" }, { Token::EQUAL, "=" }, { Token::PLUS_EQUALS, "+=" }, { Token::MINUS_EQUALS, "-=" }, { Token::NOT_EQUAL, "!=" }, { Token::EQUAL_EQUAL, "==" }, { Token::LESS_THAN, "<" }, { Token::GREATER_THAN, ">" }, { Token::LESS_EQUAL, "<=" }, { Token::GREATER_EQUAL, ">=" }, { Token::BANG, "!" }, { Token::BOOLEAN_OR, "||" }, { Token::BOOLEAN_AND, "&&" }, { Token::DOT, "." }, { Token::COMMA, "," }, }; EXPECT_TRUE(CheckTokenizer("- + = += -= != == < > <= >= ! || && . ,", operators)); } TEST(Tokenizer, Scoper) { TokenExpectation scopers[] = { { Token::LEFT_BRACE, "{" }, { Token::LEFT_BRACKET, "[" }, { Token::RIGHT_BRACKET, "]" }, { Token::RIGHT_BRACE, "}" }, { Token::LEFT_PAREN, "(" }, { Token::RIGHT_PAREN, ")" }, }; EXPECT_TRUE(CheckTokenizer("{[ ]} ()", scopers)); } TEST(Tokenizer, FunctionCall) { TokenExpectation fn[] = { { Token::IDENTIFIER, "fun" }, { Token::LEFT_PAREN, "(" }, { Token::STRING, "\"foo\"" }, { Token::RIGHT_PAREN, ")" }, { Token::LEFT_BRACE, "{" }, { Token::IDENTIFIER, "foo" }, { Token::EQUAL, "=" }, { Token::INTEGER, "12" }, { Token::RIGHT_BRACE, "}" }, }; EXPECT_TRUE(CheckTokenizer("fun(\"foo\") {\nfoo = 12}", fn)); } TEST(Tokenizer, Locations) { InputFile input(SourceFile("/test")); input.SetContents("1 2 \"three\"\n 4"); Err err; std::vector results = Tokenizer::Tokenize(&input, &err); ASSERT_EQ(4u, results.size()); ASSERT_TRUE(results[0].location() == Location(&input, 1, 1, 1)); ASSERT_TRUE(results[1].location() == Location(&input, 1, 3, 3)); ASSERT_TRUE(results[2].location() == Location(&input, 1, 5, 5)); ASSERT_TRUE(results[3].location() == Location(&input, 2, 3, 8)); } TEST(Tokenizer, ByteOffsetOfNthLine) { EXPECT_EQ(0u, Tokenizer::ByteOffsetOfNthLine("foo", 1)); // Windows and Posix have different line endings, so check the byte at the // location rather than the offset. char input1[] = "aaa\nxaa\n\nya"; EXPECT_EQ('x', input1[Tokenizer::ByteOffsetOfNthLine(input1, 2)]); EXPECT_EQ('y', input1[Tokenizer::ByteOffsetOfNthLine(input1, 4)]); char input2[3]; input2[0] = 'a'; input2[1] = '\n'; // Manually set to avoid Windows double-byte endings. input2[2] = 0; EXPECT_EQ(0u, Tokenizer::ByteOffsetOfNthLine(input2, 1)); EXPECT_EQ(2u, Tokenizer::ByteOffsetOfNthLine(input2, 2)); } TEST(Tokenizer, Comments) { TokenExpectation fn[] = { { Token::LINE_COMMENT, "# Stuff" }, { Token::IDENTIFIER, "fun" }, { Token::LEFT_PAREN, "(" }, { Token::STRING, "\"foo\"" }, { Token::RIGHT_PAREN, ")" }, { Token::LEFT_BRACE, "{" }, { Token::SUFFIX_COMMENT, "# Things" }, { Token::LINE_COMMENT, "#Wee" }, { Token::IDENTIFIER, "foo" }, { Token::EQUAL, "=" }, { Token::INTEGER, "12" }, { Token::SUFFIX_COMMENT, "#Zip" }, { Token::RIGHT_BRACE, "}" }, }; EXPECT_TRUE(CheckTokenizer( "# Stuff\n" "fun(\"foo\") { # Things\n" "#Wee\n" "foo = 12 #Zip\n" "}", fn)); } TEST(Tokenizer, CommentsContinued) { // In the first test, the comments aren't horizontally aligned, so they're // considered separate. In the second test, they are, so "B" is a // continuation of "A" (another SUFFIX comment). TokenExpectation fn1[] = { { Token::IDENTIFIER, "fun" }, { Token::LEFT_PAREN, "(" }, { Token::STRING, "\"foo\"" }, { Token::RIGHT_PAREN, ")" }, { Token::LEFT_BRACE, "{" }, { Token::SUFFIX_COMMENT, "# A" }, { Token::LINE_COMMENT, "# B" }, { Token::RIGHT_BRACE, "}" }, }; EXPECT_TRUE(CheckTokenizer( "fun(\"foo\") { # A\n" " # B\n" "}", fn1)); TokenExpectation fn2[] = { { Token::IDENTIFIER, "fun" }, { Token::LEFT_PAREN, "(" }, { Token::STRING, "\"foo\"" }, { Token::RIGHT_PAREN, ")" }, { Token::LEFT_BRACE, "{" }, { Token::SUFFIX_COMMENT, "# A" }, { Token::SUFFIX_COMMENT, "# B" }, { Token::RIGHT_BRACE, "}" }, }; EXPECT_TRUE(CheckTokenizer( "fun(\"foo\") { # A\n" " # B\n" // Note that these are aligned, the \"s move A out. "}", fn2)); }