summaryrefslogtreecommitdiffstats
path: root/benchmarks/benchmark_main.cpp
diff options
context:
space:
mode:
Diffstat (limited to 'benchmarks/benchmark_main.cpp')
-rw-r--r--benchmarks/benchmark_main.cpp226
1 files changed, 0 insertions, 226 deletions
diff --git a/benchmarks/benchmark_main.cpp b/benchmarks/benchmark_main.cpp
deleted file mode 100644
index d60670b..0000000
--- a/benchmarks/benchmark_main.cpp
+++ /dev/null
@@ -1,226 +0,0 @@
-/*
- * Copyright (C) 2012 The Android Open Source Project
- *
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-#include "benchmark.h"
-
-#include <regex.h>
-#include <stdio.h>
-#include <stdlib.h>
-
-#include <string>
-#include <map>
-
-#include <inttypes.h>
-
-static int64_t g_bytes_processed;
-static int64_t g_benchmark_total_time_ns;
-static int64_t g_benchmark_start_time_ns;
-
-typedef std::map<std::string, ::testing::Benchmark*> BenchmarkMap;
-typedef BenchmarkMap::iterator BenchmarkMapIt;
-static BenchmarkMap g_benchmarks;
-static int g_name_column_width = 20;
-
-static int Round(int n) {
- int base = 1;
- while (base*10 < n) {
- base *= 10;
- }
- if (n < 2*base) {
- return 2*base;
- }
- if (n < 5*base) {
- return 5*base;
- }
- return 10*base;
-}
-
-static int64_t NanoTime() {
- struct timespec t;
- t.tv_sec = t.tv_nsec = 0;
- clock_gettime(CLOCK_MONOTONIC, &t);
- return static_cast<int64_t>(t.tv_sec) * 1000000000LL + t.tv_nsec;
-}
-
-namespace testing {
-
-Benchmark* Benchmark::Arg(int arg) {
- args_.push_back(arg);
- return this;
-}
-
-const char* Benchmark::Name() {
- return name_;
-}
-
-bool Benchmark::ShouldRun(int argc, char* argv[]) {
- if (argc == 1) {
- return true; // With no arguments, we run all benchmarks.
- }
- // Otherwise, we interpret each argument as a regular expression and
- // see if any of our benchmarks match.
- for (int i = 1; i < argc; i++) {
- regex_t re;
- if (regcomp(&re, argv[i], 0) != 0) {
- fprintf(stderr, "couldn't compile \"%s\" as a regular expression!\n", argv[i]);
- exit(EXIT_FAILURE);
- }
- int match = regexec(&re, name_, 0, NULL, 0);
- regfree(&re);
- if (match != REG_NOMATCH) {
- return true;
- }
- }
- return false;
-}
-
-void Benchmark::Register(const char* name, void (*fn)(int), void (*fn_range)(int, int)) {
- name_ = name;
- fn_ = fn;
- fn_range_ = fn_range;
-
- if (fn_ == NULL && fn_range_ == NULL) {
- fprintf(stderr, "%s: missing function\n", name_);
- exit(EXIT_FAILURE);
- }
-
- g_benchmarks.insert(std::make_pair(name, this));
-}
-
-void Benchmark::Run() {
- if (fn_ != NULL) {
- RunWithArg(0);
- } else {
- if (args_.empty()) {
- fprintf(stderr, "%s: no args!\n", name_);
- exit(EXIT_FAILURE);
- }
- for (size_t i = 0; i < args_.size(); ++i) {
- RunWithArg(args_[i]);
- }
- }
-}
-
-void Benchmark::RunRepeatedlyWithArg(int iterations, int arg) {
- g_bytes_processed = 0;
- g_benchmark_total_time_ns = 0;
- g_benchmark_start_time_ns = NanoTime();
- if (fn_ != NULL) {
- fn_(iterations);
- } else {
- fn_range_(iterations, arg);
- }
- if (g_benchmark_start_time_ns != 0) {
- g_benchmark_total_time_ns += NanoTime() - g_benchmark_start_time_ns;
- }
-}
-
-void Benchmark::RunWithArg(int arg) {
- // run once in case it's expensive
- int iterations = 1;
- RunRepeatedlyWithArg(iterations, arg);
- while (g_benchmark_total_time_ns < 1e9 && iterations < 1e9) {
- int last = iterations;
- if (g_benchmark_total_time_ns/iterations == 0) {
- iterations = 1e9;
- } else {
- iterations = 1e9 / (g_benchmark_total_time_ns/iterations);
- }
- iterations = std::max(last + 1, std::min(iterations + iterations/2, 100*last));
- iterations = Round(iterations);
- RunRepeatedlyWithArg(iterations, arg);
- }
-
- char throughput[100];
- throughput[0] = '\0';
- if (g_benchmark_total_time_ns > 0 && g_bytes_processed > 0) {
- double mib_processed = static_cast<double>(g_bytes_processed)/1e6;
- double seconds = static_cast<double>(g_benchmark_total_time_ns)/1e9;
- snprintf(throughput, sizeof(throughput), " %8.2f MiB/s", mib_processed/seconds);
- }
-
- char full_name[100];
- if (fn_range_ != NULL) {
- if (arg >= (1<<20)) {
- snprintf(full_name, sizeof(full_name), "%s/%dM", name_, arg/(1<<20));
- } else if (arg >= (1<<10)) {
- snprintf(full_name, sizeof(full_name), "%s/%dK", name_, arg/(1<<10));
- } else {
- snprintf(full_name, sizeof(full_name), "%s/%d", name_, arg);
- }
- } else {
- snprintf(full_name, sizeof(full_name), "%s", name_);
- }
-
- printf("%-*s %10d %10" PRId64 "%s\n", g_name_column_width, full_name,
- iterations, g_benchmark_total_time_ns/iterations, throughput);
- fflush(stdout);
-}
-
-} // namespace testing
-
-void SetBenchmarkBytesProcessed(int64_t x) {
- g_bytes_processed = x;
-}
-
-void StopBenchmarkTiming() {
- if (g_benchmark_start_time_ns != 0) {
- g_benchmark_total_time_ns += NanoTime() - g_benchmark_start_time_ns;
- }
- g_benchmark_start_time_ns = 0;
-}
-
-void StartBenchmarkTiming() {
- if (g_benchmark_start_time_ns == 0) {
- g_benchmark_start_time_ns = NanoTime();
- }
-}
-
-int main(int argc, char* argv[]) {
- if (g_benchmarks.empty()) {
- fprintf(stderr, "No benchmarks registered!\n");
- exit(EXIT_FAILURE);
- }
-
- for (BenchmarkMapIt it = g_benchmarks.begin(); it != g_benchmarks.end(); ++it) {
- int name_width = static_cast<int>(strlen(it->second->Name()));
- g_name_column_width = std::max(g_name_column_width, name_width);
- }
-
- bool need_header = true;
- for (BenchmarkMapIt it = g_benchmarks.begin(); it != g_benchmarks.end(); ++it) {
- ::testing::Benchmark* b = it->second;
- if (b->ShouldRun(argc, argv)) {
- if (need_header) {
- printf("%-*s %10s %10s\n", g_name_column_width, "", "iterations", "ns/op");
- fflush(stdout);
- need_header = false;
- }
- b->Run();
- }
- }
-
- if (need_header) {
- fprintf(stderr, "No matching benchmarks!\n");
- fprintf(stderr, "Available benchmarks:\n");
- for (BenchmarkMapIt it = g_benchmarks.begin(); it != g_benchmarks.end(); ++it) {
- fprintf(stderr, " %s\n", it->second->Name());
- }
- exit(EXIT_FAILURE);
- }
-
- return 0;
-}