| 1 | /* |
| 2 | * Copyright (C) 2017 Apple Inc. All rights reserved. |
| 3 | * |
| 4 | * Redistribution and use in source and binary forms, with or without |
| 5 | * modification, are permitted provided that the following conditions |
| 6 | * are met: |
| 7 | * 1. Redistributions of source code must retain the above copyright |
| 8 | * notice, this list of conditions and the following disclaimer. |
| 9 | * 2. Redistributions in binary form must reproduce the above copyright |
| 10 | * notice, this list of conditions and the following disclaimer in the |
| 11 | * documentation and/or other materials provided with the distribution. |
| 12 | * |
| 13 | * THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY |
| 14 | * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| 15 | * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR |
| 16 | * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR |
| 17 | * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, |
| 18 | * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, |
| 19 | * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR |
| 20 | * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY |
| 21 | * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| 22 | * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| 23 | * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| 24 | */ |
| 25 | |
| 26 | #pragma once |
| 27 | |
| 28 | #include <wtf/Atomics.h> |
| 29 | #include <wtf/FastMalloc.h> |
| 30 | #include <wtf/HashFunctions.h> |
| 31 | #include <wtf/Lock.h> |
| 32 | #include <wtf/Noncopyable.h> |
| 33 | #include <wtf/Vector.h> |
| 34 | |
| 35 | namespace WTF { |
| 36 | |
| 37 | // This is a concurrent hash-based set for pointers. It's optimized for: |
| 38 | // |
| 39 | // - High rate of contains() calls. |
| 40 | // - High rate of add() calls that don't add anything new. add() calls that don't add anything (nop adds) |
| 41 | // don't mutate the table at all. |
| 42 | // - Not too many threads. I doubt this scales beyond ~4. Though, it may actually scale better than that |
| 43 | // if the rate of nop adds is absurdly high. |
| 44 | // |
| 45 | // If we wanted this to scale better, the main change we'd have to make is how this table determines when |
| 46 | // to resize. Right now it's a shared counter. We lock;xadd this counter. One easy way to make that |
| 47 | // scalable is to require each thread that works with the ConcurrentPtrHashSet to register itself first. |
| 48 | // Then each thread would have some data structure that has a counter. We could institute the policy that |
| 49 | // each thread simply increments its own load counter, in its own data structure. Then, if any search to |
| 50 | // resolve a collision does more than N iterations, we can compute a combined load by summing the load |
| 51 | // counters of all of the thread data structures. |
| 52 | // |
| 53 | // ConcurrentPtrHashSet's main user, the GC, sees a 98% nop add rate in Speedometer. That's why this |
| 54 | // focuses so much on cases where the table does not change. |
| 55 | class ConcurrentPtrHashSet { |
| 56 | WTF_MAKE_NONCOPYABLE(ConcurrentPtrHashSet); |
| 57 | WTF_MAKE_FAST_ALLOCATED; |
| 58 | |
| 59 | public: |
| 60 | WTF_EXPORT_PRIVATE ConcurrentPtrHashSet(); |
| 61 | WTF_EXPORT_PRIVATE ~ConcurrentPtrHashSet(); |
| 62 | |
| 63 | template<typename T> |
| 64 | bool contains(T value) |
| 65 | { |
| 66 | return containsImpl(cast(value)); |
| 67 | } |
| 68 | |
| 69 | template<typename T> |
| 70 | bool add(T value) |
| 71 | { |
| 72 | return addImpl(cast(value)); |
| 73 | } |
| 74 | |
| 75 | size_t size() const |
| 76 | { |
| 77 | return m_table.loadRelaxed()->load.loadRelaxed(); |
| 78 | } |
| 79 | |
| 80 | // Only call when you know that no other thread can call add(). This frees up memory without changing |
| 81 | // the contents of the table. |
| 82 | WTF_EXPORT_PRIVATE void deleteOldTables(); |
| 83 | |
| 84 | // Only call when you know that no other thread can call add(). This frees up all memory except for |
| 85 | // the smallest possible hashtable. |
| 86 | WTF_EXPORT_PRIVATE void clear(); |
| 87 | |
| 88 | private: |
| 89 | struct Table { |
| 90 | WTF_MAKE_STRUCT_FAST_ALLOCATED; |
| 91 | |
| 92 | static std::unique_ptr<Table> create(unsigned size); |
| 93 | |
| 94 | unsigned maxLoad() const { return size / 2; } |
| 95 | |
| 96 | unsigned size; // This is immutable. |
| 97 | unsigned mask; // This is immutable. |
| 98 | Atomic<unsigned> load; |
| 99 | Atomic<void*> array[1]; |
| 100 | }; |
| 101 | |
| 102 | static unsigned hash(void* ptr) |
| 103 | { |
| 104 | return PtrHash<void*>::hash(ptr); |
| 105 | } |
| 106 | |
| 107 | void initialize(); |
| 108 | |
| 109 | template<typename T> |
| 110 | void* cast(T value) |
| 111 | { |
| 112 | static_assert(sizeof(T) <= sizeof(void*), "type too big" ); |
| 113 | union { |
| 114 | void* ptr; |
| 115 | T value; |
| 116 | } u; |
| 117 | u.ptr = nullptr; |
| 118 | u.value = value; |
| 119 | return u.ptr; |
| 120 | } |
| 121 | |
| 122 | bool containsImpl(void* ptr) const |
| 123 | { |
| 124 | Table* table = m_table.loadRelaxed(); |
| 125 | unsigned mask = table->mask; |
| 126 | unsigned startIndex = hash(ptr) & mask; |
| 127 | unsigned index = startIndex; |
| 128 | for (;;) { |
| 129 | void* entry = table->array[index].loadRelaxed(); |
| 130 | if (!entry) |
| 131 | return false; |
| 132 | if (entry == ptr) |
| 133 | return true; |
| 134 | index = (index + 1) & mask; |
| 135 | RELEASE_ASSERT(index != startIndex); |
| 136 | } |
| 137 | } |
| 138 | |
| 139 | // Returns true if a new entry was added. |
| 140 | bool addImpl(void* ptr) |
| 141 | { |
| 142 | Table* table = m_table.loadRelaxed(); |
| 143 | unsigned mask = table->mask; |
| 144 | unsigned startIndex = hash(ptr) & mask; |
| 145 | unsigned index = startIndex; |
| 146 | for (;;) { |
| 147 | void* entry = table->array[index].loadRelaxed(); |
| 148 | if (!entry) |
| 149 | return addSlow(table, mask, startIndex, index, ptr); |
| 150 | if (entry == ptr) |
| 151 | return false; |
| 152 | index = (index + 1) & mask; |
| 153 | RELEASE_ASSERT(index != startIndex); |
| 154 | } |
| 155 | } |
| 156 | |
| 157 | WTF_EXPORT_PRIVATE bool addSlow(Table* table, unsigned mask, unsigned startIndex, unsigned index, void* ptr); |
| 158 | |
| 159 | void resizeIfNecessary(); |
| 160 | bool resizeAndAdd(void* ptr); |
| 161 | |
| 162 | Vector<std::unique_ptr<Table>, 4> m_allTables; |
| 163 | Atomic<Table*> m_table; // This is never null. |
| 164 | Lock m_lock; // We just use this to control resize races. |
| 165 | }; |
| 166 | |
| 167 | } // namespace WTF |
| 168 | |
| 169 | using WTF::ConcurrentPtrHashSet; |
| 170 | |