Make more use of benchmark::DoNotOptimize in benchmarks.
A lot of these benchmarks predate DoNotOptimize and rolled their own
hacks.
Bug: http://b/148307629
Test: ran benchmarks before & after and got similar results
Change-Id: If44699d261b687f6253af709edda58f4c90fb285
diff --git a/benchmarks/get_heap_size_benchmark.cpp b/benchmarks/get_heap_size_benchmark.cpp
index c3680dc..47d5b18 100644
--- a/benchmarks/get_heap_size_benchmark.cpp
+++ b/benchmarks/get_heap_size_benchmark.cpp
@@ -30,8 +30,6 @@
#include <benchmark/benchmark.h>
#include "util.h"
-static volatile size_t sink;
-
static constexpr int NTHREADS = 5;
static std::atomic<int> thread_count;
@@ -57,7 +55,7 @@
sched_yield();
}
for (auto _ : state) {
- sink = mallinfo().uordblks;
+ benchmark::DoNotOptimize(mallinfo().uordblks);
}
for (int i = 0; i < 5; i++) {
int res = pthread_join(t[i], NULL);