| /* |
| * Copyright (C) 2012 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #include "benchmark.h" |
| |
| #include <regex.h> |
| #include <stdio.h> |
| #include <stdlib.h> |
| |
| #include <string> |
| #include <map> |
| |
| #include <inttypes.h> |
| |
| static int64_t g_bytes_processed; |
| static int64_t g_benchmark_total_time_ns; |
| static int64_t g_benchmark_start_time_ns; |
| |
| typedef std::map<std::string, ::testing::Benchmark*> BenchmarkMap; |
| typedef BenchmarkMap::iterator BenchmarkMapIt; |
| static BenchmarkMap g_benchmarks; |
| static int g_name_column_width = 20; |
| |
| static int Round(int n) { |
| int base = 1; |
| while (base*10 < n) { |
| base *= 10; |
| } |
| if (n < 2*base) { |
| return 2*base; |
| } |
| if (n < 5*base) { |
| return 5*base; |
| } |
| return 10*base; |
| } |
| |
| static int64_t NanoTime() { |
| struct timespec t; |
| t.tv_sec = t.tv_nsec = 0; |
| clock_gettime(CLOCK_MONOTONIC, &t); |
| return static_cast<int64_t>(t.tv_sec) * 1000000000LL + t.tv_nsec; |
| } |
| |
| namespace testing { |
| |
| Benchmark* Benchmark::Arg(int arg) { |
| args_.push_back(arg); |
| return this; |
| } |
| |
| const char* Benchmark::Name() { |
| return name_; |
| } |
| |
| bool Benchmark::ShouldRun(int argc, char* argv[]) { |
| if (argc == 1) { |
| return true; // With no arguments, we run all benchmarks. |
| } |
| // Otherwise, we interpret each argument as a regular expression and |
| // see if any of our benchmarks match. |
| for (int i = 1; i < argc; i++) { |
| regex_t re; |
| if (regcomp(&re, argv[i], 0) != 0) { |
| fprintf(stderr, "couldn't compile \"%s\" as a regular expression!\n", argv[i]); |
| exit(EXIT_FAILURE); |
| } |
| int match = regexec(&re, name_, 0, NULL, 0); |
| regfree(&re); |
| if (match != REG_NOMATCH) { |
| return true; |
| } |
| } |
| return false; |
| } |
| |
| void Benchmark::Register(const char* name, void (*fn)(int), void (*fn_range)(int, int)) { |
| name_ = name; |
| fn_ = fn; |
| fn_range_ = fn_range; |
| |
| if (fn_ == NULL && fn_range_ == NULL) { |
| fprintf(stderr, "%s: missing function\n", name_); |
| exit(EXIT_FAILURE); |
| } |
| |
| g_benchmarks.insert(std::make_pair(name, this)); |
| } |
| |
| void Benchmark::Run() { |
| if (fn_ != NULL) { |
| RunWithArg(0); |
| } else { |
| if (args_.empty()) { |
| fprintf(stderr, "%s: no args!\n", name_); |
| exit(EXIT_FAILURE); |
| } |
| for (size_t i = 0; i < args_.size(); ++i) { |
| RunWithArg(args_[i]); |
| } |
| } |
| } |
| |
| void Benchmark::RunRepeatedlyWithArg(int iterations, int arg) { |
| g_bytes_processed = 0; |
| g_benchmark_total_time_ns = 0; |
| g_benchmark_start_time_ns = NanoTime(); |
| if (fn_ != NULL) { |
| fn_(iterations); |
| } else { |
| fn_range_(iterations, arg); |
| } |
| if (g_benchmark_start_time_ns != 0) { |
| g_benchmark_total_time_ns += NanoTime() - g_benchmark_start_time_ns; |
| } |
| } |
| |
| void Benchmark::RunWithArg(int arg) { |
| // run once in case it's expensive |
| int iterations = 1; |
| RunRepeatedlyWithArg(iterations, arg); |
| while (g_benchmark_total_time_ns < 1e9 && iterations < 1e9) { |
| int last = iterations; |
| if (g_benchmark_total_time_ns/iterations == 0) { |
| iterations = 1e9; |
| } else { |
| iterations = 1e9 / (g_benchmark_total_time_ns/iterations); |
| } |
| iterations = std::max(last + 1, std::min(iterations + iterations/2, 100*last)); |
| iterations = Round(iterations); |
| RunRepeatedlyWithArg(iterations, arg); |
| } |
| |
| char throughput[100]; |
| throughput[0] = '\0'; |
| if (g_benchmark_total_time_ns > 0 && g_bytes_processed > 0) { |
| double mib_processed = static_cast<double>(g_bytes_processed)/1e6; |
| double seconds = static_cast<double>(g_benchmark_total_time_ns)/1e9; |
| snprintf(throughput, sizeof(throughput), " %8.2f MiB/s", mib_processed/seconds); |
| } |
| |
| char full_name[100]; |
| if (fn_range_ != NULL) { |
| if (arg >= (1<<20)) { |
| snprintf(full_name, sizeof(full_name), "%s/%dM", name_, arg/(1<<20)); |
| } else if (arg >= (1<<10)) { |
| snprintf(full_name, sizeof(full_name), "%s/%dK", name_, arg/(1<<10)); |
| } else { |
| snprintf(full_name, sizeof(full_name), "%s/%d", name_, arg); |
| } |
| } else { |
| snprintf(full_name, sizeof(full_name), "%s", name_); |
| } |
| |
| printf("%-*s %10d %10" PRId64 "%s\n", g_name_column_width, full_name, |
| iterations, g_benchmark_total_time_ns/iterations, throughput); |
| fflush(stdout); |
| } |
| |
| } // namespace testing |
| |
| void SetBenchmarkBytesProcessed(int64_t x) { |
| g_bytes_processed = x; |
| } |
| |
| void StopBenchmarkTiming() { |
| if (g_benchmark_start_time_ns != 0) { |
| g_benchmark_total_time_ns += NanoTime() - g_benchmark_start_time_ns; |
| } |
| g_benchmark_start_time_ns = 0; |
| } |
| |
| void StartBenchmarkTiming() { |
| if (g_benchmark_start_time_ns == 0) { |
| g_benchmark_start_time_ns = NanoTime(); |
| } |
| } |
| |
| int main(int argc, char* argv[]) { |
| if (g_benchmarks.empty()) { |
| fprintf(stderr, "No benchmarks registered!\n"); |
| exit(EXIT_FAILURE); |
| } |
| |
| for (BenchmarkMapIt it = g_benchmarks.begin(); it != g_benchmarks.end(); ++it) { |
| int name_width = static_cast<int>(strlen(it->second->Name())); |
| g_name_column_width = std::max(g_name_column_width, name_width); |
| } |
| |
| bool need_header = true; |
| for (BenchmarkMapIt it = g_benchmarks.begin(); it != g_benchmarks.end(); ++it) { |
| ::testing::Benchmark* b = it->second; |
| if (b->ShouldRun(argc, argv)) { |
| if (need_header) { |
| printf("%-*s %10s %10s\n", g_name_column_width, "", "iterations", "ns/op"); |
| fflush(stdout); |
| need_header = false; |
| } |
| b->Run(); |
| } |
| } |
| |
| if (need_header) { |
| fprintf(stderr, "No matching benchmarks!\n"); |
| fprintf(stderr, "Available benchmarks:\n"); |
| for (BenchmarkMapIt it = g_benchmarks.begin(); it != g_benchmarks.end(); ++it) { |
| fprintf(stderr, " %s\n", it->second->Name()); |
| } |
| exit(EXIT_FAILURE); |
| } |
| |
| return 0; |
| } |