| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751 |
- // Part of the Carbon Language project, under the Apache License v2.0 with LLVM
- // Exceptions. See /LICENSE for license information.
- // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
- #include <benchmark/benchmark.h>
- #include <algorithm>
- #include <utility>
- #include "absl/random/random.h"
- #include "common/check.h"
- #include "common/raw_string_ostream.h"
- #include "llvm/ADT/Sequence.h"
- #include "llvm/ADT/StringExtras.h"
- #include "testing/base/source_gen.h"
- #include "toolchain/base/shared_value_stores.h"
- #include "toolchain/diagnostics/diagnostic_emitter.h"
- #include "toolchain/diagnostics/null_diagnostics.h"
- #include "toolchain/lex/lex.h"
- #include "toolchain/lex/token_kind.h"
- #include "toolchain/lex/tokenized_buffer.h"
- namespace Carbon::Lex {
- namespace {
- // A large value for measurement stability without making benchmarking too slow.
- // Needs to be a multiple of 100 so we can easily divide it up into percentages,
- // and 1% itself needs to not be too tiny. This makes 100,000 a great balance.
- constexpr int NumTokens = 100'000;
- // Compute a random sequence of just identifiers.
- static auto RandomIdentifierSeq(int min_length, int max_length, bool uniform,
- llvm::StringRef separator = " ")
- -> std::string {
- auto& gen = Testing::SourceGen::Global();
- llvm::SmallVector<llvm::StringRef> ids =
- gen.GetShuffledIdentifiers(NumTokens, min_length, max_length, uniform);
- return llvm::join(ids, separator);
- }
- auto GetSymbolTokenTable() -> llvm::ArrayRef<TokenKind> {
- // Build our own table of symbols so we can use repetitions to skew the
- // distribution.
- static auto symbol_token_table_storage = [] {
- llvm::SmallVector<TokenKind> table;
- #define CARBON_SYMBOL_TOKEN(TokenName, Spelling) \
- table.push_back(TokenKind::TokenName);
- #define CARBON_OPENING_GROUP_SYMBOL_TOKEN(TokenName, Spelling, ClosingName)
- #define CARBON_CLOSING_GROUP_SYMBOL_TOKEN(TokenName, Spelling, OpeningName)
- #include "toolchain/lex/token_kind.def"
- table.insert(table.end(), 32, TokenKind::Semi);
- table.insert(table.end(), 16, TokenKind::Comma);
- table.insert(table.end(), 12, TokenKind::Period);
- table.insert(table.end(), 8, TokenKind::Colon);
- table.insert(table.end(), 8, TokenKind::Equal);
- table.insert(table.end(), 4, TokenKind::Amp);
- table.insert(table.end(), 4, TokenKind::ColonExclaim);
- table.insert(table.end(), 4, TokenKind::EqualEqual);
- table.insert(table.end(), 4, TokenKind::ExclaimEqual);
- table.insert(table.end(), 4, TokenKind::MinusGreater);
- table.insert(table.end(), 4, TokenKind::Star);
- return table;
- }();
- return symbol_token_table_storage;
- }
- struct RandomSourceOptions {
- int symbol_percent = 0;
- int keyword_percent = 0;
- int numeric_literal_percent = 0;
- int string_literal_percent = 0;
- int tokens_per_line = NumTokens;
- int comment_line_percent = 0;
- int blank_line_percent = 0;
- void Validate() {
- auto is_percentage = [](int n) { return 0 <= n && n <= 100; };
- CARBON_CHECK(is_percentage(symbol_percent));
- CARBON_CHECK(is_percentage(keyword_percent));
- CARBON_CHECK(is_percentage(numeric_literal_percent));
- CARBON_CHECK(is_percentage(string_literal_percent));
- CARBON_CHECK(is_percentage(symbol_percent + keyword_percent +
- numeric_literal_percent +
- string_literal_percent));
- CARBON_CHECK(tokens_per_line <= NumTokens);
- CARBON_CHECK(
- NumTokens % tokens_per_line == 0,
- "Tokens per line of {0} does not divide the number of tokens {1}",
- tokens_per_line, NumTokens);
- CARBON_CHECK(is_percentage(comment_line_percent));
- CARBON_CHECK(is_percentage(blank_line_percent));
- // Ensure that comment and blank lines are less than 100% so we eventually
- // produce a token line.
- CARBON_CHECK(comment_line_percent + blank_line_percent < 100);
- }
- };
- // Based on measurements of LLVM's source code, a rough approximation of the
- // distribution of these kinds of tokens.
- constexpr RandomSourceOptions DefaultSourceDist = {
- .symbol_percent = 50,
- .keyword_percent = 7,
- .numeric_literal_percent = 17,
- .string_literal_percent = 1,
- // The median for LLVM is roughly 5.
- .tokens_per_line = 5,
- // Observed percentage of lines in LLVM.
- .comment_line_percent = 22,
- .blank_line_percent = 15,
- };
- // Compute random source code with a mixture of tokens and whitespace according
- // to the options. The source isn't designed to be valid, or directly
- // representative of real-world Carbon code. However, it tries to provide
- // reasonable coverage of the different aspects of Carbon's lexer, such that for
- // real world source code with distributions similar to the options provided the
- // lexer performance will be roughly representative.
- //
- // TODO: Does not yet support generating numeric or string literals.
- //
- // TODO: The shape of lines is handled very arbitrarily and should vary more to
- // avoid over-fitting to a specific shape (number of tokens, length of comment).
- auto RandomSource(RandomSourceOptions options) -> std::string {
- options.Validate();
- static_assert((NumTokens % 100) == 0,
- "The number of tokens must be divisible by 100 so that we can "
- "easily scale integer percentages up to it.");
- // Get static pools of symbols, keywords, and identifiers.
- llvm::ArrayRef<TokenKind> symbols = GetSymbolTokenTable();
- llvm::ArrayRef<TokenKind> keywords = TokenKind::KeywordTokens;
- // Build a list of StringRefs from the different types with the desired
- // distribution, then shuffle that list.
- llvm::OwningArrayRef<llvm::StringRef> tokens(NumTokens);
- int num_symbols = (NumTokens / 100) * options.symbol_percent;
- int num_keywords = (NumTokens / 100) * options.keyword_percent;
- int num_identifiers = NumTokens - num_symbols - num_keywords;
- CARBON_CHECK(
- num_identifiers == 0 || num_identifiers > 500,
- "We require at least 500 identifiers as we need to collect a reasonable "
- "number of samples to end up with a reasonable distribution of lengths.");
- llvm::SmallVector<llvm::StringRef> ids =
- Testing::SourceGen::Global().GetIdentifiers(num_identifiers);
- for (int i : llvm::seq(num_symbols)) {
- tokens[i] = symbols[i % symbols.size()].fixed_spelling();
- }
- for (int i : llvm::seq(num_keywords)) {
- tokens[num_symbols + i] = keywords[i % keywords.size()].fixed_spelling();
- }
- for (int i : llvm::seq(num_identifiers)) {
- // We always have enough identifiers, so no need to mod here.
- tokens[num_symbols + num_keywords + i] = ids[i];
- }
- std::shuffle(tokens.begin(), tokens.end(), absl::BitGen());
- // Distribute the tokens across lines as well as horizontal whitespace. The
- // goal isn't to make any one line representative of anything, but to make the
- // rough density of different kinds of whitespace roughly representative.
- //
- // TODO: This is a really coarse approach that just picks a fixed number of
- // tokens per line rather than using some distribution with this as the median
- // or mean.
- llvm::SmallVector<std::string> lines;
- // First place tokens onto each line.
- for (auto i : llvm::seq(NumTokens / options.tokens_per_line)) {
- lines.push_back("");
- RawStringOstream os;
- // Arbitrarily indent each line by two spaces.
- os << " ";
- llvm::ListSeparator sep(" ");
- for (int j : llvm::seq(options.tokens_per_line)) {
- os << sep << tokens[i * options.tokens_per_line + j];
- }
- lines.push_back(os.TakeStr());
- }
- // Next, synthesize blank and comment lines with the correct distribution.
- int token_line_percent =
- 100 - options.blank_line_percent - options.comment_line_percent;
- CARBON_CHECK(token_line_percent > 0);
- int num_token_lines = lines.size();
- int num_lines = num_token_lines * 100 / token_line_percent;
- int num_blank_lines = num_lines * options.blank_line_percent / 100;
- int num_comment_lines = num_lines - num_blank_lines - num_token_lines;
- CARBON_CHECK(num_comment_lines >= 0);
- lines.resize(num_lines);
- for (auto& line :
- llvm::MutableArrayRef(lines).slice(num_lines - num_comment_lines)) {
- // TODO: We should vary the content and length, especially as the
- // distribution is weirdly shaped with just over half the comment lines
- // being blank and the median length of non-black comment lines being 64!
- // This is a *very* coarse approximation of the mean at 30 characters long.
- line = " // abcdefghijklmnopqrstuvwxyz";
- }
- // Now shuffle the lines.
- std::shuffle(lines.begin(), lines.end(), absl::BitGen());
- // And join them into the source string.
- return llvm::join(lines, "\n");
- }
- class LexerBenchHelper {
- public:
- explicit LexerBenchHelper(llvm::StringRef text)
- : source_(MakeSourceBuffer(text)) {}
- auto Lex() -> TokenizedBuffer {
- DiagnosticConsumer& consumer = NullDiagnosticConsumer();
- return Lex::Lex(value_stores_, source_, consumer);
- }
- auto DiagnoseErrors() -> std::string {
- RawStringOstream result;
- StreamDiagnosticConsumer consumer(&result);
- auto buffer = Lex::Lex(value_stores_, source_, consumer);
- consumer.Flush();
- CARBON_CHECK(buffer.has_errors(),
- "Asked to diagnose errors but none found!");
- return result.TakeStr();
- }
- auto source_text() -> llvm::StringRef { return source_.text(); }
- private:
- auto MakeSourceBuffer(llvm::StringRef text) -> SourceBuffer {
- CARBON_CHECK(fs_.addFile(filename_, /*ModificationTime=*/0,
- llvm::MemoryBuffer::getMemBuffer(text)));
- return std::move(*SourceBuffer::MakeFromFile(fs_, filename_,
- ConsoleDiagnosticConsumer()));
- }
- SharedValueStores value_stores_;
- llvm::vfs::InMemoryFileSystem fs_;
- std::string filename_ = "test.carbon";
- SourceBuffer source_;
- };
- void BM_ValidKeywords(benchmark::State& state) {
- absl::BitGen gen;
- std::array<llvm::StringRef, NumTokens> tokens;
- for (int i : llvm::seq(NumTokens)) {
- tokens[i] = TokenKind::KeywordTokens[i % TokenKind::KeywordTokens.size()]
- .fixed_spelling();
- }
- std::shuffle(tokens.begin(), tokens.end(), gen);
- std::string source = llvm::join(tokens, " ");
- LexerBenchHelper helper(source);
- for (auto _ : state) {
- TokenizedBuffer buffer = helper.Lex();
- CARBON_CHECK(!buffer.has_errors());
- }
- state.SetBytesProcessed(state.iterations() * source.size());
- state.counters["tokens_per_second"] = benchmark::Counter(
- NumTokens, benchmark::Counter::kIsIterationInvariantRate);
- }
- BENCHMARK(BM_ValidKeywords);
- void BM_ValidKeywordsAsRawIdentifiers(benchmark::State& state) {
- absl::BitGen gen;
- std::array<llvm::StringRef, NumTokens> tokens;
- for (int i : llvm::seq(NumTokens)) {
- tokens[i] = TokenKind::KeywordTokens[i % TokenKind::KeywordTokens.size()]
- .fixed_spelling();
- }
- std::shuffle(tokens.begin(), tokens.end(), gen);
- std::string source("r#");
- source.append(llvm::join(tokens, " r#"));
- LexerBenchHelper helper(source);
- for (auto _ : state) {
- TokenizedBuffer buffer = helper.Lex();
- CARBON_CHECK(!buffer.has_errors());
- }
- state.SetBytesProcessed(state.iterations() * source.size());
- state.counters["tokens_per_second"] = benchmark::Counter(
- NumTokens, benchmark::Counter::kIsIterationInvariantRate);
- }
- BENCHMARK(BM_ValidKeywordsAsRawIdentifiers);
- // This benchmark does a 50-50 split of r-prefixed and r#-prefixed identifiers
- // to directly compare raw and non-raw performance.
- void BM_RawIdentifierFocus(benchmark::State& state) {
- llvm::SmallVector<llvm::StringRef> ids =
- Testing::SourceGen::Global().GetIdentifiers(NumTokens / 2);
- llvm::SmallVector<std::string> modified_ids;
- // As we resize, start with the in-use prefix. Note that `r#` uses the first
- // character of the original identifier.
- modified_ids.resize(NumTokens / 2, "r#");
- modified_ids.resize(NumTokens, "r");
- for (int i : llvm::seq(NumTokens / 2)) {
- // Use the same identifier both ways.
- modified_ids[i].append(ids[i]);
- modified_ids[i + NumTokens / 2].append(
- llvm::StringRef(ids[i]).drop_front());
- }
- absl::BitGen gen;
- std::array<llvm::StringRef, NumTokens> tokens;
- for (int i : llvm::seq(NumTokens)) {
- tokens[i] = modified_ids[i];
- }
- std::shuffle(tokens.begin(), tokens.end(), gen);
- std::string source = llvm::join(tokens, " ");
- LexerBenchHelper helper(source);
- for (auto _ : state) {
- TokenizedBuffer buffer = helper.Lex();
- CARBON_CHECK(!buffer.has_errors());
- }
- state.SetBytesProcessed(state.iterations() * source.size());
- state.counters["tokens_per_second"] = benchmark::Counter(
- NumTokens, benchmark::Counter::kIsIterationInvariantRate);
- }
- BENCHMARK(BM_RawIdentifierFocus);
- template <int MinLength, int MaxLength, bool Uniform>
- void BM_ValidIdentifiers(benchmark::State& state) {
- std::string source = RandomIdentifierSeq(MinLength, MaxLength, Uniform);
- LexerBenchHelper helper(source);
- for (auto _ : state) {
- TokenizedBuffer buffer = helper.Lex();
- CARBON_CHECK(!buffer.has_errors(), "{0}", helper.DiagnoseErrors());
- }
- state.SetBytesProcessed(state.iterations() * source.size());
- state.counters["tokens_per_second"] = benchmark::Counter(
- NumTokens, benchmark::Counter::kIsIterationInvariantRate);
- }
- // Benchmark the non-uniform distribution we observe in C++ code.
- BENCHMARK(BM_ValidIdentifiers<1, 64, /*Uniform=*/false>);
- // Also benchmark a few uniform distribution ranges of identifier widths to
- // cover different patterns that emerge with small, medium, and longer
- // identifiers.
- BENCHMARK(BM_ValidIdentifiers<1, 1, /*Uniform=*/true>);
- BENCHMARK(BM_ValidIdentifiers<3, 5, /*Uniform=*/true>);
- BENCHMARK(BM_ValidIdentifiers<3, 16, /*Uniform=*/true>);
- BENCHMARK(BM_ValidIdentifiers<12, 64, /*Uniform=*/true>);
- BENCHMARK(BM_ValidIdentifiers<16, 16, /*Uniform=*/true>);
- BENCHMARK(BM_ValidIdentifiers<24, 24, /*Uniform=*/true>);
- BENCHMARK(BM_ValidIdentifiers<32, 32, /*Uniform=*/true>);
- BENCHMARK(BM_ValidIdentifiers<48, 48, /*Uniform=*/true>);
- BENCHMARK(BM_ValidIdentifiers<64, 64, /*Uniform=*/true>);
- BENCHMARK(BM_ValidIdentifiers<80, 80, /*Uniform=*/true>);
- // Benchmark to stress the lexing of horizontal whitespace. This sets up what is
- // nearly a worst-case scenario of short-but-expensive-to-lex tokens with runs
- // of horizontal whitespace between them.
- void BM_HorizontalWhitespace(benchmark::State& state) {
- int num_spaces = state.range(0);
- std::string separator(num_spaces, ' ');
- std::string source = RandomIdentifierSeq(3, 5, /*uniform=*/true, separator);
- LexerBenchHelper helper(source);
- for (auto _ : state) {
- TokenizedBuffer buffer = helper.Lex();
- // Ensure that lexing actually occurs for benchmarking and that it doesn't
- // hit errors that would skew the benchmark results.
- CARBON_CHECK(!buffer.has_errors(), "{0}", helper.DiagnoseErrors());
- }
- state.SetBytesProcessed(state.iterations() * source.size());
- state.counters["tokens_per_second"] = benchmark::Counter(
- NumTokens, benchmark::Counter::kIsIterationInvariantRate);
- }
- BENCHMARK(BM_HorizontalWhitespace)->RangeMultiplier(4)->Range(1, 128);
- void BM_RandomSource(benchmark::State& state) {
- std::string source = RandomSource(DefaultSourceDist);
- LexerBenchHelper helper(source);
- for (auto _ : state) {
- TokenizedBuffer buffer = helper.Lex();
- // Ensure that lexing actually occurs for benchmarking and that it doesn't
- // hit errors that would skew the benchmark results.
- CARBON_CHECK(!buffer.has_errors(), "{0}", helper.DiagnoseErrors());
- }
- state.SetBytesProcessed(state.iterations() * source.size());
- state.counters["tokens_per_second"] = benchmark::Counter(
- NumTokens, benchmark::Counter::kIsIterationInvariantRate);
- state.counters["lines_per_second"] =
- benchmark::Counter(llvm::StringRef(source).count('\n'),
- benchmark::Counter::kIsIterationInvariantRate);
- }
- // The distributions between symbols, keywords, and identifiers here are
- // guesses. Eventually, we should collect more data to help tune these, but
- // hopefully the performance isn't too sensitive and we can just cover a wide
- // range here.
- BENCHMARK(BM_RandomSource);
- // Benchmark to stress opening and closing grouped symbols.
- void BM_GroupingSymbols(benchmark::State& state) {
- int curly_brace_depth = state.range(0);
- int paren_depth = state.range(1);
- int square_bracket_depth = state.range(2);
- // TODO: It might be interesting to have some random pattern of nesting, but
- // the obvious ways to do that result it really unstable total size of input
- // or unbalanced groups. For now, just use a simple strict nesting approach.
- // It should still let us look for specific pain points. We do include some
- // whitespace and keywords to make sure *some* other parts of the benchmark
- // are also active and have some reasonable icache pressure.
- llvm::SmallVector<llvm::StringRef> ids =
- Testing::SourceGen::Global().GetShuffledIdentifiers(NumTokens);
- RawStringOstream os;
- int num_tokens_per_nest =
- curly_brace_depth * 2 + paren_depth * 2 + square_bracket_depth * 2 + 2;
- int num_nests = NumTokens / num_tokens_per_nest;
- for (int i : llvm::seq(num_nests)) {
- for (int j : llvm::seq(curly_brace_depth)) {
- os.indent(j * 2) << "{\n";
- }
- os.indent(curly_brace_depth * 2);
- for ([[maybe_unused]] int j : llvm::seq(paren_depth)) {
- os << "(";
- }
- for ([[maybe_unused]] int j : llvm::seq(square_bracket_depth)) {
- os << "[";
- }
- os << ids[(i * 2) % NumTokens];
- for ([[maybe_unused]] int j : llvm::seq(square_bracket_depth)) {
- os << "]";
- }
- for ([[maybe_unused]] int j : llvm::seq(paren_depth)) {
- os << ")";
- }
- for (int j : llvm::reverse(llvm::seq(curly_brace_depth))) {
- os << "\n";
- os.indent(j * 2) << "}";
- }
- os << ids[(i * 2 + 1) % NumTokens] << "\n";
- }
- std::string source = os.TakeStr();
- LexerBenchHelper helper(source);
- for (auto _ : state) {
- TokenizedBuffer buffer = helper.Lex();
- // Ensure that lexing actually occurs for benchmarking and that it doesn't
- // hit errors that would skew the benchmark results.
- CARBON_CHECK(!buffer.has_errors(), "{0}", helper.DiagnoseErrors());
- }
- state.SetBytesProcessed(state.iterations() * source.size());
- state.counters["tokens_per_second"] = benchmark::Counter(
- NumTokens, benchmark::Counter::kIsIterationInvariantRate);
- state.counters["lines_per_second"] =
- benchmark::Counter(llvm::StringRef(source).count('\n'),
- benchmark::Counter::kIsIterationInvariantRate);
- }
- BENCHMARK(BM_GroupingSymbols)
- ->ArgsProduct({
- {1, 2, 3, 4, 8, 16, 32},
- {0},
- {0},
- })
- ->ArgsProduct({
- {0},
- {1, 2, 3, 4, 8, 16, 32},
- {0},
- })
- ->ArgsProduct({
- {0},
- {0},
- {1, 2, 3, 4, 8, 16, 32},
- })
- ->ArgsProduct({
- {32},
- {1, 2, 3, 4, 8, 16, 32},
- {0},
- })
- ->ArgsProduct({
- {32},
- {32},
- {1, 2, 3, 4, 8, 16, 32},
- });
- // Benchmark to stress the lexing of blank lines. This uses a simple, easy to
- // lex token, but separates each one by varying numbers of blank lines.
- void BM_BlankLines(benchmark::State& state) {
- int num_blank_lines = state.range(0);
- std::string separator(num_blank_lines, '\n');
- std::string source = RandomIdentifierSeq(3, 5, /*uniform=*/true, separator);
- LexerBenchHelper helper(source);
- for (auto _ : state) {
- TokenizedBuffer buffer = helper.Lex();
- // Ensure that lexing actually occurs for benchmarking and that it doesn't
- // hit errors that would skew the benchmark results.
- CARBON_CHECK(!buffer.has_errors(), "{0}", helper.DiagnoseErrors());
- }
- state.SetBytesProcessed(state.iterations() * source.size());
- state.counters["tokens_per_second"] = benchmark::Counter(
- NumTokens, benchmark::Counter::kIsIterationInvariantRate);
- state.counters["lines_per_second"] =
- benchmark::Counter(llvm::StringRef(source).count('\n'),
- benchmark::Counter::kIsIterationInvariantRate);
- }
- BENCHMARK(BM_BlankLines)->RangeMultiplier(4)->Range(1, 128);
- // Benchmark to stress the lexing of comment lines. This uses a simple, easy to
- // lex token, but separates each one by varying numbers of comment lines, with
- // varying comment line length and indentation.
- void BM_CommentLines(benchmark::State& state) {
- int num_comment_lines = state.range(0);
- int comment_length = state.range(1);
- int comment_indent = state.range(2);
- RawStringOstream os;
- os << "\n";
- for (int i : llvm::seq(num_comment_lines)) {
- static_cast<void>(i);
- os << std::string(comment_indent, ' ') << "//"
- << std::string(comment_length, ' ') << "\n";
- }
- std::string source =
- RandomIdentifierSeq(3, 5, /*uniform=*/true, os.TakeStr());
- LexerBenchHelper helper(source);
- for (auto _ : state) {
- TokenizedBuffer buffer = helper.Lex();
- // Ensure that lexing actually occurs for benchmarking and that it doesn't
- // hit errors that would skew the benchmark results.
- CARBON_CHECK(!buffer.has_errors(), "{0}", helper.DiagnoseErrors());
- }
- state.SetBytesProcessed(state.iterations() * source.size());
- state.counters["tokens_per_second"] = benchmark::Counter(
- NumTokens, benchmark::Counter::kIsIterationInvariantRate);
- state.counters["lines_per_second"] =
- benchmark::Counter(llvm::StringRef(source).count('\n'),
- benchmark::Counter::kIsIterationInvariantRate);
- }
- BENCHMARK(BM_CommentLines)
- ->ArgsProduct({
- // How many lines of comment. Focused on a couple of small and checking
- // how it scales up to large blocks.
- {1, 4, 128},
- // Comment lengths: the two extremes and a middling length.
- {0, 30, 70},
- // Comment indentations.
- {0, 2, 8},
- });
- // This is a speed-of-light benchmark that should reflect memory bandwidth
- // (ideally) of simply reading all the source code. For speed-of-light we use
- // `strcpy` -- this both examines ever byte of the input looking for a null to
- // end the copy, and also writes to a data structure of roughly the same size as
- // the input. This routine is one we expect to be *very* well optimized and give
- // a good approximation of the fastest possible lexer given the physical
- // constraints of the machine. Note that which particular source we use as input
- // here isn't especially interesting, so we just pick one and should update it
- // to reflect whatever distribution is most realistic long-term. The
- // bytes/second throughput is the important output of this routine.
- auto BM_SpeedOfLightStrCpy(benchmark::State& state) -> void {
- std::string source = RandomSource(DefaultSourceDist);
- // A buffer to write the null-terminated contents of `source` into.
- llvm::OwningArrayRef<char> buffer(source.size() + 1);
- for (auto _ : state) {
- const char* text = source.data();
- benchmark::DoNotOptimize(text);
- strcpy(buffer.data(), text);
- benchmark::DoNotOptimize(buffer.data());
- }
- state.SetBytesProcessed(state.iterations() * source.size());
- state.counters["tokens_per_second"] = benchmark::Counter(
- NumTokens, benchmark::Counter::kIsIterationInvariantRate);
- state.counters["lines_per_second"] =
- benchmark::Counter(llvm::StringRef(source).count('\n'),
- benchmark::Counter::kIsIterationInvariantRate);
- }
- BENCHMARK(BM_SpeedOfLightStrCpy);
- // This is a speed-of-light benchmark that builds up a best-case byte-wise table
- // dispatch using guaranteed tail recursion. The goal is both to ensure the
- // general technique can reasonably hit the level of performance we need and to
- // establish how far from this speed of light the actual lexer currently sits.
- //
- // A major impact on the observed performance of this technique is how many
- // different functions are reached in this dispatch loop. This benchmark
- // infrastructure tries to bracket the range of performance this technique
- // affords with different numbers of dispatch target functions.
- using DispatchPtrT = auto (*)(ssize_t& index, const char* text, char* buffer)
- -> void;
- using DispatchTableT = std::array<DispatchPtrT, 256>;
- template <const DispatchTableT& Table>
- auto BasicDispatch(ssize_t& index, const char* text, char* buffer) -> void {
- *buffer = text[index];
- ++index;
- // NOLINTNEXTLINE(readability-avoid-return-with-void-value): For musttail.
- [[clang::musttail]] return Table[static_cast<unsigned char>(text[index])](
- index, text, buffer);
- }
- template <const DispatchTableT& Table, char C>
- auto SpecializedDispatch(ssize_t& index, const char* text, char* buffer)
- -> void {
- CARBON_CHECK(C == text[index]);
- *buffer = C;
- ++index;
- // NOLINTNEXTLINE(readability-avoid-return-with-void-value): For musttail.
- [[clang::musttail]] return Table[static_cast<unsigned char>(text[index])](
- index, text, buffer);
- }
- // A sample of the symbol characters used in Carbon code. Doesn't need to be
- // perfect, as we just need to have a reasonably large # of distinct dispatch
- // functions.
- constexpr char DispatchSpecializableSymbols[] = {
- '!', '%', '(', ')', '*', '+', ',', '-', '.', ':',
- ';', '<', '=', '>', '?', '[', ']', '{', '}', '~',
- };
- // Create an array of all the characters we can specialize dispatch over --
- // [0-9A-Za-z] and the symbols above. Similar to the above symbols, doesn't need
- // to be exhaustive.
- constexpr std::array<char, 26 * 2 + 10 + sizeof(DispatchSpecializableSymbols)>
- DispatchSpecializableChars = []() {
- constexpr int Size = sizeof(DispatchSpecializableChars);
- std::array<char, Size> chars = {};
- int i = 0;
- for (char c = '0'; c <= '9'; ++c) {
- chars[i] = c;
- ++i;
- }
- for (char c = 'A'; c <= 'Z'; ++c) {
- chars[i] = c;
- ++i;
- }
- for (char c = 'a'; c <= 'z'; ++c) {
- chars[i] = c;
- ++i;
- }
- for (char c : DispatchSpecializableSymbols) {
- chars[i] = c;
- ++i;
- }
- CARBON_CHECK(i == Size);
- return chars;
- }();
- // Instantiate a number of specialized dispatch functions for characters in the
- // array above, and assign those function addresses to the character's entry in
- // the provided table. The provided `tmp_table` is a temporary that will
- // eventually initialize the provided `Table` constant, so the constant is what
- // we propagate to the instantiated function and the temporary is the one we
- // initialize.
- template <const DispatchTableT& Table, size_t... Indices>
- constexpr auto SpecializeDispatchTable(
- DispatchTableT& tmp_table, std::index_sequence<Indices...> /*indices*/)
- -> void {
- static_assert(sizeof...(Indices) <= sizeof(DispatchSpecializableChars));
- ((tmp_table[static_cast<unsigned char>(DispatchSpecializableChars[Indices])] =
- &SpecializedDispatch<Table, DispatchSpecializableChars[Indices]>),
- ...);
- }
- // The maximum number of dispatch targets is the size of the array + 1 (for the
- // base case target).
- constexpr int MaxDispatchTargets = sizeof(DispatchSpecializableChars) + 1;
- // Dispatch tables with a provided number of distinct dispatch targets. There
- // will always be one additional target for the null byte to end the loop.
- template <int NumDispatchTargets>
- constexpr DispatchTableT DispatchTable = []() {
- static_assert(NumDispatchTargets > 0, "Need at least one dispatch target.");
- static_assert(NumDispatchTargets <= MaxDispatchTargets,
- "Limited number of dispatch targets available.");
- DispatchTableT tmp_table = {};
- // Start with the basic dispatch target.
- for (int i = 0; i < 256; ++i) {
- tmp_table[i] = &BasicDispatch<DispatchTable<NumDispatchTargets>>;
- }
- // NOLINTNEXTLINE(readability-braces-around-statements): False positive.
- if constexpr (NumDispatchTargets > 1) {
- // Add additional dispatch targets from our specializable array.
- SpecializeDispatchTable<DispatchTable<NumDispatchTargets>>(
- tmp_table, std::make_index_sequence<NumDispatchTargets - 1>());
- }
- // Special case the null byte index to end the tail-dispatch.
- tmp_table[0] =
- +[](ssize_t& index, const char* text, char* /*buffer*/) -> void {
- CARBON_CHECK(text[index] == '\0');
- return;
- };
- return tmp_table;
- }();
- template <int NumDispatchTargets>
- auto BM_SpeedOfLightDispatch(benchmark::State& state) -> void {
- std::string source = RandomSource(DefaultSourceDist);
- // A buffer to write to, simulating some minimal write traffic.
- llvm::OwningArrayRef<char> buffer(source.size());
- for (auto _ : state) {
- const char* text = source.data();
- benchmark::DoNotOptimize(text);
- // Use `ssize_t` to minimize indexing overhead.
- ssize_t i = 0;
- // The dispatch table tail-recurses through the entire string.
- DispatchTable<NumDispatchTargets>[static_cast<unsigned char>(text[i])](
- i, text, buffer.data());
- CARBON_CHECK(i == static_cast<ssize_t>(source.size()));
- benchmark::DoNotOptimize(buffer.data());
- }
- state.SetBytesProcessed(state.iterations() * source.size());
- state.counters["tokens_per_second"] = benchmark::Counter(
- NumTokens, benchmark::Counter::kIsIterationInvariantRate);
- state.counters["lines_per_second"] =
- benchmark::Counter(llvm::StringRef(source).count('\n'),
- benchmark::Counter::kIsIterationInvariantRate);
- }
- BENCHMARK(BM_SpeedOfLightDispatch<1>);
- BENCHMARK(BM_SpeedOfLightDispatch<2>);
- BENCHMARK(BM_SpeedOfLightDispatch<4>);
- BENCHMARK(BM_SpeedOfLightDispatch<8>);
- BENCHMARK(BM_SpeedOfLightDispatch<16>);
- BENCHMARK(BM_SpeedOfLightDispatch<32>);
- BENCHMARK(BM_SpeedOfLightDispatch<MaxDispatchTargets>);
- } // namespace
- } // namespace Carbon::Lex
|