// Copyright 2018 The Abseil Authors. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // https://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #include #include #include #include #include #include #include #include #include #include #include #include "absl/base/internal/raw_logging.h" #include "absl/container/internal/container_memory.h" #include "absl/container/internal/hash_function_defaults.h" #include "absl/container/internal/raw_hash_set.h" #include "absl/random/random.h" #include "absl/strings/str_format.h" #include "benchmark/benchmark.h" namespace absl { ABSL_NAMESPACE_BEGIN namespace container_internal { struct RawHashSetTestOnlyAccess { template static auto GetSlots(const C& c) -> decltype(c.slots_) { return c.slots_; } }; namespace { struct IntPolicy { using slot_type = int64_t; using key_type = int64_t; using init_type = int64_t; static void construct(void*, int64_t* slot, int64_t v) { *slot = v; } static void destroy(void*, int64_t*) {} static void transfer(void*, int64_t* new_slot, int64_t* old_slot) { *new_slot = *old_slot; } static int64_t& element(slot_type* slot) { return *slot; } template static auto apply(F&& f, int64_t x) -> decltype(std::forward(f)(x, x)) { return std::forward(f)(x, x); } template static constexpr HashSlotFn get_hash_slot_fn() { return nullptr; } }; class StringPolicy { template ::value>::type> decltype(std::declval()( std::declval(), std::piecewise_construct, std::declval>(), std::declval())) static apply_impl(F&& f, std::pair, V> p) { const absl::string_view& key = std::get<0>(p.first); return std::forward(f)(key, std::piecewise_construct, std::move(p.first), std::move(p.second)); } public: struct slot_type { struct ctor {}; template slot_type(ctor, Ts&&... ts) : pair(std::forward(ts)...) {} std::pair pair; }; using key_type = std::string; using init_type = std::pair; template static void construct(allocator_type* alloc, slot_type* slot, Args... args) { std::allocator_traits::construct( *alloc, slot, typename slot_type::ctor(), std::forward(args)...); } template static void destroy(allocator_type* alloc, slot_type* slot) { std::allocator_traits::destroy(*alloc, slot); } template static void transfer(allocator_type* alloc, slot_type* new_slot, slot_type* old_slot) { construct(alloc, new_slot, std::move(old_slot->pair)); destroy(alloc, old_slot); } static std::pair& element(slot_type* slot) { return slot->pair; } template static auto apply(F&& f, Args&&... args) -> decltype(apply_impl(std::forward(f), PairArgs(std::forward(args)...))) { return apply_impl(std::forward(f), PairArgs(std::forward(args)...)); } template static constexpr HashSlotFn get_hash_slot_fn() { return nullptr; } }; struct StringHash : container_internal::hash_default_hash { using is_transparent = void; }; struct StringEq : std::equal_to { using is_transparent = void; }; struct StringTable : raw_hash_set> { using Base = typename StringTable::raw_hash_set; StringTable() {} using Base::Base; }; struct IntTable : raw_hash_set, std::equal_to, std::allocator> { using Base = typename IntTable::raw_hash_set; IntTable() {} using Base::Base; }; struct string_generator { template std::string operator()(RNG& rng) const { std::string res; res.resize(size); std::uniform_int_distribution printable_ascii(0x20, 0x7E); std::generate(res.begin(), res.end(), [&] { return printable_ascii(rng); }); return res; } size_t size; }; // Model a cache in steady state. // // On a table of size N, keep deleting the LRU entry and add a random one. void BM_CacheInSteadyState(benchmark::State& state) { std::random_device rd; std::mt19937 rng(rd()); string_generator gen{12}; StringTable t; std::deque keys; while (t.size() < state.range(0)) { auto x = t.emplace(gen(rng), gen(rng)); if (x.second) keys.push_back(x.first->first); } ABSL_RAW_CHECK(state.range(0) >= 10, ""); while (state.KeepRunning()) { // Some cache hits. std::deque::const_iterator it; for (int i = 0; i != 90; ++i) { if (i % 10 == 0) it = keys.end(); ::benchmark::DoNotOptimize(t.find(*--it)); } // Some cache misses. for (int i = 0; i != 10; ++i) ::benchmark::DoNotOptimize(t.find(gen(rng))); ABSL_RAW_CHECK(t.erase(keys.front()), keys.front().c_str()); keys.pop_front(); while (true) { auto x = t.emplace(gen(rng), gen(rng)); if (x.second) { keys.push_back(x.first->first); break; } } } state.SetItemsProcessed(state.iterations()); state.SetLabel(absl::StrFormat("load_factor=%.2f", t.load_factor())); } template void CacheInSteadyStateArgs(Benchmark* bm) { // The default. const float max_load_factor = 0.875; // When the cache is at the steady state, the probe sequence will equal // capacity if there is no reclamation of deleted slots. Pick a number large // enough to make the benchmark slow for that case. const size_t capacity = 1 << 10; // Check N data points to cover load factors in [0.4, 0.8). const size_t kNumPoints = 10; for (size_t i = 0; i != kNumPoints; ++i) bm->Arg(std::ceil( capacity * (max_load_factor + i * max_load_factor / kNumPoints) / 2)); } BENCHMARK(BM_CacheInSteadyState)->Apply(CacheInSteadyStateArgs); void BM_EraseEmplace(benchmark::State& state) { IntTable t; int64_t size = state.range(0); for (int64_t i = 0; i < size; ++i) { t.emplace(i); } while (state.KeepRunningBatch(size)) { for (int64_t i = 0; i < size; ++i) { benchmark::DoNotOptimize(t); t.erase(i); t.emplace(i); } } } BENCHMARK(BM_EraseEmplace)->Arg(1)->Arg(2)->Arg(4)->Arg(8)->Arg(16)->Arg(100); void BM_EndComparison(benchmark::State& state) { StringTable t = {{"a", "a"}, {"b", "b"}}; auto it = t.begin(); for (auto i : state) { benchmark::DoNotOptimize(t); benchmark::DoNotOptimize(it); benchmark::DoNotOptimize(it != t.end()); } } BENCHMARK(BM_EndComparison); void BM_Iteration(benchmark::State& state) { std::random_device rd; std::mt19937 rng(rd()); string_generator gen{12}; StringTable t; size_t capacity = state.range(0); size_t size = state.range(1); t.reserve(capacity); while (t.size() < size) { t.emplace(gen(rng), gen(rng)); } for (auto i : state) { benchmark::DoNotOptimize(t); for (auto it = t.begin(); it != t.end(); ++it) { benchmark::DoNotOptimize(*it); } } } BENCHMARK(BM_Iteration) ->ArgPair(1, 1) ->ArgPair(2, 2) ->ArgPair(4, 4) ->ArgPair(7, 7) ->ArgPair(10, 10) ->ArgPair(15, 15) ->ArgPair(16, 16) ->ArgPair(54, 54) ->ArgPair(100, 100) ->ArgPair(400, 400) // empty ->ArgPair(0, 0) ->ArgPair(10, 0) ->ArgPair(100, 0) ->ArgPair(1000, 0) ->ArgPair(10000, 0) // sparse ->ArgPair(100, 1) ->ArgPair(1000, 10); void BM_CopyCtorSparseInt(benchmark::State& state) { std::random_device rd; std::mt19937 rng(rd()); IntTable t; std::uniform_int_distribution dist(0, ~uint64_t{}); size_t size = state.range(0); t.reserve(size * 10); while (t.size() < size) { t.emplace(dist(rng)); } for (auto i : state) { IntTable t2 = t; benchmark::DoNotOptimize(t2); } } BENCHMARK(BM_CopyCtorSparseInt)->Range(1, 4096); void BM_CopyCtorInt(benchmark::State& state) { std::random_device rd; std::mt19937 rng(rd()); IntTable t; std::uniform_int_distribution dist(0, ~uint64_t{}); size_t size = state.range(0); while (t.size() < size) { t.emplace(dist(rng)); } for (auto i : state) { IntTable t2 = t; benchmark::DoNotOptimize(t2); } } BENCHMARK(BM_CopyCtorInt)->Range(0, 4096); void BM_CopyCtorString(benchmark::State& state) { std::random_device rd; std::mt19937 rng(rd()); StringTable t; std::uniform_int_distribution dist(0, ~uint64_t{}); size_t size = state.range(0); while (t.size() < size) { t.emplace(std::to_string(dist(rng)), std::to_string(dist(rng))); } for (auto i : state) { StringTable t2 = t; benchmark::DoNotOptimize(t2); } } BENCHMARK(BM_CopyCtorString)->Range(0, 4096); void BM_CopyAssign(benchmark::State& state) { std::random_device rd; std::mt19937 rng(rd()); IntTable t; std::uniform_int_distribution dist(0, ~uint64_t{}); while (t.size() < state.range(0)) { t.emplace(dist(rng)); } IntTable t2; for (auto _ : state) { t2 = t; benchmark::DoNotOptimize(t2); } } BENCHMARK(BM_CopyAssign)->Range(128, 4096); void BM_RangeCtor(benchmark::State& state) { std::random_device rd; std::mt19937 rng(rd()); std::uniform_int_distribution dist(0, ~uint64_t{}); std::vector values; const size_t desired_size = state.range(0); while (values.size() < desired_size) { values.emplace_back(dist(rng)); } for (auto unused : state) { IntTable t{values.begin(), values.end()}; benchmark::DoNotOptimize(t); } } BENCHMARK(BM_RangeCtor)->Range(128, 65536); void BM_NoOpReserveIntTable(benchmark::State& state) { IntTable t; t.reserve(100000); for (auto _ : state) { benchmark::DoNotOptimize(t); t.reserve(100000); } } BENCHMARK(BM_NoOpReserveIntTable); void BM_NoOpReserveStringTable(benchmark::State& state) { StringTable t; t.reserve(100000); for (auto _ : state) { benchmark::DoNotOptimize(t); t.reserve(100000); } } BENCHMARK(BM_NoOpReserveStringTable); void BM_ReserveIntTable(benchmark::State& state) { constexpr size_t kBatchSize = 1024; size_t reserve_size = static_cast(state.range(0)); std::vector tables; while (state.KeepRunningBatch(kBatchSize)) { state.PauseTiming(); tables.clear(); tables.resize(kBatchSize); state.ResumeTiming(); for (auto& t : tables) { benchmark::DoNotOptimize(t); t.reserve(reserve_size); benchmark::DoNotOptimize(t); } } } BENCHMARK(BM_ReserveIntTable)->Range(1, 64); void BM_ReserveStringTable(benchmark::State& state) { constexpr size_t kBatchSize = 1024; size_t reserve_size = static_cast(state.range(0)); std::vector tables; while (state.KeepRunningBatch(kBatchSize)) { state.PauseTiming(); tables.clear(); tables.resize(kBatchSize); state.ResumeTiming(); for (auto& t : tables) { benchmark::DoNotOptimize(t); t.reserve(reserve_size); benchmark::DoNotOptimize(t); } } } BENCHMARK(BM_ReserveStringTable)->Range(1, 64); // Like std::iota, except that ctrl_t doesn't support operator++. template void Iota(CtrlIter begin, CtrlIter end, int value) { for (; begin != end; ++begin, ++value) { *begin = static_cast(value); } } void BM_Group_Match(benchmark::State& state) { std::array group; Iota(group.begin(), group.end(), -4); Group g{group.data()}; h2_t h = 1; for (auto _ : state) { ::benchmark::DoNotOptimize(h); ::benchmark::DoNotOptimize(g); ::benchmark::DoNotOptimize(g.Match(h)); } } BENCHMARK(BM_Group_Match); void BM_GroupPortable_Match(benchmark::State& state) { std::array group; Iota(group.begin(), group.end(), -4); GroupPortableImpl g{group.data()}; h2_t h = 1; for (auto _ : state) { ::benchmark::DoNotOptimize(h); ::benchmark::DoNotOptimize(g); ::benchmark::DoNotOptimize(g.Match(h)); } } BENCHMARK(BM_GroupPortable_Match); void BM_Group_MaskEmpty(benchmark::State& state) { std::array group; Iota(group.begin(), group.end(), -4); Group g{group.data()}; for (auto _ : state) { ::benchmark::DoNotOptimize(g); ::benchmark::DoNotOptimize(g.MaskEmpty()); } } BENCHMARK(BM_Group_MaskEmpty); void BM_Group_MaskEmptyOrDeleted(benchmark::State& state) { std::array group; Iota(group.begin(), group.end(), -4); Group g{group.data()}; for (auto _ : state) { ::benchmark::DoNotOptimize(g); ::benchmark::DoNotOptimize(g.MaskEmptyOrDeleted()); } } BENCHMARK(BM_Group_MaskEmptyOrDeleted); void BM_Group_MaskNonFull(benchmark::State& state) { std::array group; Iota(group.begin(), group.end(), -4); Group g{group.data()}; for (auto _ : state) { ::benchmark::DoNotOptimize(g); ::benchmark::DoNotOptimize(g.MaskNonFull()); } } BENCHMARK(BM_Group_MaskNonFull); void BM_Group_CountLeadingEmptyOrDeleted(benchmark::State& state) { std::array group; Iota(group.begin(), group.end(), -2); Group g{group.data()}; for (auto _ : state) { ::benchmark::DoNotOptimize(g); ::benchmark::DoNotOptimize(g.CountLeadingEmptyOrDeleted()); } } BENCHMARK(BM_Group_CountLeadingEmptyOrDeleted); void BM_Group_MatchFirstEmptyOrDeleted(benchmark::State& state) { std::array group; Iota(group.begin(), group.end(), -2); Group g{group.data()}; for (auto _ : state) { ::benchmark::DoNotOptimize(g); ::benchmark::DoNotOptimize(g.MaskEmptyOrDeleted().LowestBitSet()); } } BENCHMARK(BM_Group_MatchFirstEmptyOrDeleted); void BM_Group_MatchFirstNonFull(benchmark::State& state) { std::array group; Iota(group.begin(), group.end(), -2); Group g{group.data()}; for (auto _ : state) { ::benchmark::DoNotOptimize(g); ::benchmark::DoNotOptimize(g.MaskNonFull().LowestBitSet()); } } BENCHMARK(BM_Group_MatchFirstNonFull); void BM_DropDeletes(benchmark::State& state) { constexpr size_t capacity = (1 << 20) - 1; std::vector ctrl(capacity + 1 + Group::kWidth); ctrl[capacity] = ctrl_t::kSentinel; std::vector pattern = {ctrl_t::kEmpty, static_cast(2), ctrl_t::kDeleted, static_cast(2), ctrl_t::kEmpty, static_cast(1), ctrl_t::kDeleted}; for (size_t i = 0; i != capacity; ++i) { ctrl[i] = pattern[i % pattern.size()]; } while (state.KeepRunning()) { state.PauseTiming(); std::vector ctrl_copy = ctrl; state.ResumeTiming(); ConvertDeletedToEmptyAndFullToDeleted(ctrl_copy.data(), capacity); ::benchmark::DoNotOptimize(ctrl_copy[capacity]); } } BENCHMARK(BM_DropDeletes); void BM_Resize(benchmark::State& state) { // For now just measure a small cheap hash table since we // are mostly interested in the overhead of type-erasure // in resize(). constexpr int kElements = 64; const int kCapacity = kElements * 2; IntTable table; for (int i = 0; i < kElements; i++) { table.insert(i); } for (auto unused : state) { table.rehash(0); table.rehash(kCapacity); } } BENCHMARK(BM_Resize); void BM_EraseIf(benchmark::State& state) { int64_t num_elements = state.range(0); size_t num_erased = static_cast(state.range(1)); constexpr size_t kRepetitions = 64; absl::BitGen rng; std::vector> keys(kRepetitions); std::vector tables; std::vector threshold; for (auto& k : keys) { tables.push_back(IntTable()); auto& table = tables.back(); for (int64_t i = 0; i < num_elements; i++) { // We use random keys to reduce noise. k.push_back(absl::Uniform(rng, 0, ~int64_t{})); if (!table.insert(k.back()).second) { k.pop_back(); --i; // duplicated value, retrying } } std::sort(k.begin(), k.end()); threshold.push_back(k[num_erased]); } while (state.KeepRunningBatch(static_cast(kRepetitions) * std::max(num_elements, int64_t{1}))) { benchmark::DoNotOptimize(tables); for (size_t t_id = 0; t_id < kRepetitions; t_id++) { auto& table = tables[t_id]; benchmark::DoNotOptimize(num_erased); auto pred = [t = threshold[t_id]](int64_t key) { return key < t; }; benchmark::DoNotOptimize(pred); benchmark::DoNotOptimize(table); absl::container_internal::EraseIf(pred, &table); } state.PauseTiming(); for (size_t t_id = 0; t_id < kRepetitions; t_id++) { auto& k = keys[t_id]; auto& table = tables[t_id]; for (size_t i = 0; i < num_erased; i++) { table.insert(k[i]); } } state.ResumeTiming(); } } BENCHMARK(BM_EraseIf) ->ArgNames({"num_elements", "num_erased"}) ->ArgPair(10, 0) ->ArgPair(1000, 0) ->ArgPair(10, 5) ->ArgPair(1000, 500) ->ArgPair(10, 10) ->ArgPair(1000, 1000); } // namespace } // namespace container_internal ABSL_NAMESPACE_END } // namespace absl // These methods are here to make it easy to examine the assembly for targeted // parts of the API. auto CodegenAbslRawHashSetInt64Find(absl::container_internal::IntTable* table, int64_t key) -> decltype(table->find(key)) { return table->find(key); } bool CodegenAbslRawHashSetInt64FindNeEnd( absl::container_internal::IntTable* table, int64_t key) { return table->find(key) != table->end(); } // This is useful because the find isn't inlined but the iterator comparison is. bool CodegenAbslRawHashSetStringFindNeEnd( absl::container_internal::StringTable* table, const std::string& key) { return table->find(key) != table->end(); } auto CodegenAbslRawHashSetInt64Insert(absl::container_internal::IntTable* table, int64_t key) -> decltype(table->insert(key)) { return table->insert(key); } bool CodegenAbslRawHashSetInt64Contains( absl::container_internal::IntTable* table, int64_t key) { return table->contains(key); } void CodegenAbslRawHashSetInt64Iterate( absl::container_internal::IntTable* table) { for (auto x : *table) benchmark::DoNotOptimize(x); } int odr = (::benchmark::DoNotOptimize(std::make_tuple( &CodegenAbslRawHashSetInt64Find, &CodegenAbslRawHashSetInt64FindNeEnd, &CodegenAbslRawHashSetStringFindNeEnd, &CodegenAbslRawHashSetInt64Insert, &CodegenAbslRawHashSetInt64Contains, &CodegenAbslRawHashSetInt64Iterate)), 1);