21using namespace std::chrono_literals;
46void GenerateTemplateResults(
const std::vector<ankerl::nanobench::Result>& benchmarkResults,
const fs::path& file,
const char* tpl)
48 if (benchmarkResults.empty() || file.empty()) {
52 std::ofstream fout{file.std_path()};
55 std::cout <<
"Created " << file << std::endl;
57 std::cout <<
"Could not write to file " << file << std::endl;
68 return benchmarks_map;
78 std::regex reFilter(
args.regex_filter);
79 std::smatch baseMatch;
81 if (
args.sanity_check) {
82 std::cout <<
"Running with -sanity-check option, output is being suppressed as benchmark results will be useless." << std::endl;
87 std::vector<const char*>
ret;
88 ret.reserve(
args.setup_args.size());
89 for (
const auto& arg :
args.setup_args)
ret.emplace_back(arg.c_str());
93 std::vector<ankerl::nanobench::Result> benchmarkResults;
96 if (!std::regex_match(
name, baseMatch, reFilter)) {
100 if (
args.is_list_only) {
101 std::cout <<
name << std::endl;
106 if (
args.sanity_check) {
112 if (
args.min_time > 0ms) {
114 std::chrono::nanoseconds min_time_ns =
args.min_time;
118 if (
args.asymptote.empty()) {
121 for (
auto n :
args.asymptote) {
128 if (!bench.results().empty()) {
129 benchmarkResults.push_back(bench.results().back());
133 GenerateTemplateResults(benchmarkResults,
args.output_csv,
"# Benchmark, evals, iterations, total, min, max, median\n"
134 "{{#result}}{{name}}, {{epochs}}, {{average(iterations)}}, {{sumProduct(iterations, elapsed)}}, {{minimum(elapsed)}}, {{maximum(elapsed)}}, {{median(elapsed)}}\n"
static std::string g_running_benchmark_name
Retrieve the name of the currently in-use benchmark.
const std::function< std::vector< const char * >()> G_TEST_COMMAND_LINE_ARGUMENTS
Retrieve the command line arguments.
const std::function< std::string()> G_TEST_GET_FULL_NAME
Retrieve the unit test name.
static std::function< std::vector< const char * >()> g_bench_command_line_args
Retrieves the available test setup command line arguments that may be used in the benchmark.
#define Assert(val)
Identity function.
Main entry point to nanobench's benchmarking facility.
Bench & epochs(size_t numEpochs) noexcept
Controls number of epochs, the number of measurements to perform.
ANKERL_NANOBENCH(NODISCARD) std Bench & name(char const *benchmarkName)
Gets the title of the benchmark.
std::vector< BigO > complexityBigO() const
ANKERL_NANOBENCH(NODISCARD) std ANKERL_NANOBENCH(NODISCARD) std Bench & output(std::ostream *outstream) noexcept
Set the output stream where the resulting markdown table will be printed to.
Bench & complexityN(T n) noexcept
ANKERL_NANOBENCH(NODISCARD) std Bench & minEpochTime(std::chrono::nanoseconds t) noexcept
Minimum time each epoch should take.
Bench & epochIterations(uint64_t numIters) noexcept
Sets exactly the number of iterations for each epoch.
static void RunAll(const Args &args)
BenchRunner(std::string name, BenchFunction func)
std::map< std::string, BenchFunction > BenchmarkMap
static BenchmarkMap & benchmarks()
char const * json() noexcept
Template to generate JSON data.
void render(char const *mustacheTemplate, Bench const &bench, std::ostream &out)
Renders output from a mustache-like template and benchmark results.
std::function< void(Bench &)> BenchFunction