From ab410df7c4594474987d79266c7f4923346816e8 Mon Sep 17 00:00:00 2001 From: unknown Date: Wed, 9 Mar 2022 22:11:38 +0300 Subject: [PATCH] New ConcurrentHashSet implementation --- ZeroLevel/Models/NullableAttributes.cs | 25 + .../Services/Collections/ConcurrentHashSet.cs | 887 +++++++++++++++++- ZeroLevel/ZeroLevel.csproj | 8 +- 3 files changed, 877 insertions(+), 43 deletions(-) create mode 100644 ZeroLevel/Models/NullableAttributes.cs diff --git a/ZeroLevel/Models/NullableAttributes.cs b/ZeroLevel/Models/NullableAttributes.cs new file mode 100644 index 0000000..ba312bd --- /dev/null +++ b/ZeroLevel/Models/NullableAttributes.cs @@ -0,0 +1,25 @@ +#if !NETSTANDARD2_1 +using System; + +/*https://github.com/i3arnon/ConcurrentHashSet*/ + +namespace ZeroLevel.Models +{ + namespace System.Diagnostics.CodeAnalysis + { + /// Specifies that when a method returns , the parameter may be null even if the corresponding type disallows it. + [AttributeUsage(AttributeTargets.Parameter)] + internal sealed class MaybeNullWhenAttribute : Attribute + { + /// Initializes the attribute with the specified return value condition. + /// + /// The return value condition. If the method returns this value, the associated parameter may be null. + /// + public MaybeNullWhenAttribute(bool returnValue) => ReturnValue = returnValue; + + /// Gets the return value condition. + public bool ReturnValue { get; } + } + } +} +#endif diff --git a/ZeroLevel/Services/Collections/ConcurrentHashSet.cs b/ZeroLevel/Services/Collections/ConcurrentHashSet.cs index 3e10552..aaa053a 100644 --- a/ZeroLevel/Services/Collections/ConcurrentHashSet.cs +++ b/ZeroLevel/Services/Collections/ConcurrentHashSet.cs @@ -1,100 +1,909 @@ using System; +using System.Collections; using System.Collections.Generic; +using System.Diagnostics; +using System.Diagnostics.CodeAnalysis; using System.Threading; -namespace ZeroLevel.Services.Collections +/*https://github.com/i3arnon/ConcurrentHashSet*/ + +namespace ZeroLevel.Collections { - public class ConcurrentHashSet : IDisposable + /// + /// Represents a thread-safe hash-based unique collection. + /// + /// The type of the items in the collection. + /// + /// All public members of are thread-safe and may be used + /// concurrently from multiple threads. + /// + [DebuggerDisplay("Count = {Count}")] + public class ConcurrentHashSet : IReadOnlyCollection, ICollection { - private readonly ReaderWriterLockSlim _lock = new ReaderWriterLockSlim(LockRecursionPolicy.SupportsRecursion); - private readonly HashSet _hashSet = new HashSet(); + private const int DefaultCapacity = 31; + private const int MaxLockNumber = 1024; + + private readonly IEqualityComparer _comparer; + private readonly bool _growLockArray; + + private int _budget; + private volatile Tables _tables; + + private static int DefaultConcurrencyLevel => Environment.ProcessorCount; - #region Implementation of ICollection ...ish - public bool Add(T item) + /// + /// Gets the + /// that is used to determine equality for the values in the set. + /// + /// + /// The generic interface implementation that is used to + /// provide hash values and determine equality for the values in the current . + /// + /// + /// requires an equality implementation to determine + /// whether values are equal. You can specify an implementation of the + /// generic interface by using a constructor that accepts a comparer parameter; + /// if you do not specify one, the default generic equality comparer is used. + /// + public IEqualityComparer Comparer => _comparer; + + /// + /// Gets the number of items contained in the . + /// + /// The number of items contained in the . + /// Count has snapshot semantics and represents the number of items in the + /// at the moment when Count was accessed. + public int Count { - _lock.EnterWriteLock(); - try + get { - return _hashSet.Add(item); + var count = 0; + var acquiredLocks = 0; + try + { + AcquireAllLocks(ref acquiredLocks); + + var countPerLocks = _tables.CountPerLock; + for (var i = 0; i < countPerLocks.Length; i++) + { + count += countPerLocks[i]; + } + } + finally + { + ReleaseLocks(0, acquiredLocks); + } + + return count; } - finally + } + + /// + /// Gets a value that indicates whether the is empty. + /// + /// true if the is empty; otherwise, + /// false. + public bool IsEmpty + { + get + { + if (!AreAllBucketsEmpty()) + { + return false; + } + + var acquiredLocks = 0; + try + { + AcquireAllLocks(ref acquiredLocks); + + return AreAllBucketsEmpty(); + } + finally + { + ReleaseLocks(0, acquiredLocks); + } + } + } + + /// + /// Initializes a new instance of the + /// class that is empty, has the default concurrency level, has the default initial capacity, and + /// uses the default comparer for the item type. + /// + public ConcurrentHashSet() + : this(DefaultConcurrencyLevel, DefaultCapacity, true, null) + { + } + + /// + /// Initializes a new instance of the + /// class that is empty, has the specified concurrency level and capacity, and uses the default + /// comparer for the item type. + /// + /// The estimated number of threads that will update the + /// concurrently. + /// The initial number of elements that the + /// can contain. + /// is + /// less than 1. + /// is less than + /// 0. + public ConcurrentHashSet(int concurrencyLevel, int capacity) + : this(concurrencyLevel, capacity, false, null) + { + } + + /// + /// Initializes a new instance of the + /// class that contains elements copied from the specified , has the default concurrency + /// level, has the default initial capacity, and uses the default comparer for the item type. + /// + /// The whose elements are copied to + /// the new + /// . + /// is a null reference. + public ConcurrentHashSet(IEnumerable collection) + : this(collection, null) + { + } + + /// + /// Initializes a new instance of the + /// class that is empty, has the specified concurrency level and capacity, and uses the specified + /// . + /// + /// The + /// implementation to use when comparing items. + public ConcurrentHashSet(IEqualityComparer? comparer) + : this(DefaultConcurrencyLevel, DefaultCapacity, true, comparer) + { + } + + /// + /// Initializes a new instance of the + /// class that contains elements copied from the specified , has the default concurrency level, has the default + /// initial capacity, and uses the specified + /// . + /// + /// The whose elements are copied to + /// the new + /// . + /// The + /// implementation to use when comparing items. + /// is a null reference + /// (Nothing in Visual Basic). + /// + public ConcurrentHashSet(IEnumerable collection, IEqualityComparer? comparer) + : this(comparer) + { + if (collection == null) throw new ArgumentNullException(nameof(collection)); + + InitializeFromCollection(collection); + } + + + /// + /// Initializes a new instance of the + /// class that contains elements copied from the specified , + /// has the specified concurrency level, has the specified initial capacity, and uses the specified + /// . + /// + /// The estimated number of threads that will update the + /// concurrently. + /// The whose elements are copied to the new + /// . + /// The implementation to use + /// when comparing items. + /// + /// is a null reference. + /// + /// + /// is less than 1. + /// + public ConcurrentHashSet(int concurrencyLevel, IEnumerable collection, IEqualityComparer? comparer) + : this(concurrencyLevel, DefaultCapacity, false, comparer) + { + if (collection == null) throw new ArgumentNullException(nameof(collection)); + + InitializeFromCollection(collection); + } + + /// + /// Initializes a new instance of the + /// class that is empty, has the specified concurrency level, has the specified initial capacity, and + /// uses the specified . + /// + /// The estimated number of threads that will update the + /// concurrently. + /// The initial number of elements that the + /// can contain. + /// The + /// implementation to use when comparing items. + /// + /// is less than 1. -or- + /// is less than 0. + /// + public ConcurrentHashSet(int concurrencyLevel, int capacity, IEqualityComparer? comparer) + : this(concurrencyLevel, capacity, false, comparer) + { + } + + private ConcurrentHashSet(int concurrencyLevel, int capacity, bool growLockArray, IEqualityComparer? comparer) + { + if (concurrencyLevel < 1) throw new ArgumentOutOfRangeException(nameof(concurrencyLevel)); + if (capacity < 0) throw new ArgumentOutOfRangeException(nameof(capacity)); + + // The capacity should be at least as large as the concurrency level. Otherwise, we would have locks that don't guard + // any buckets. + if (capacity < concurrencyLevel) { - if (_lock.IsWriteLockHeld) _lock.ExitWriteLock(); + capacity = concurrencyLevel; } + + var locks = new object[concurrencyLevel]; + for (var i = 0; i < locks.Length; i++) + { + locks[i] = new object(); + } + + var countPerLock = new int[locks.Length]; + var buckets = new Node[capacity]; + _tables = new Tables(buckets, locks, countPerLock); + + _growLockArray = growLockArray; + _budget = buckets.Length / locks.Length; + _comparer = comparer ?? EqualityComparer.Default; } + /// + /// Adds the specified item to the . + /// + /// The item to add. + /// true if the items was added to the + /// successfully; false if it already exists. + /// The + /// contains too many items. + public bool Add(T item) => + AddInternal(item, _comparer.GetHashCode(item), true); + + /// + /// Removes all items from the . + /// public void Clear() { - _lock.EnterWriteLock(); + var locksAcquired = 0; try { - _hashSet.Clear(); + AcquireAllLocks(ref locksAcquired); + + if (AreAllBucketsEmpty()) + { + return; + } + + var tables = _tables; + var newTables = new Tables(new Node[DefaultCapacity], tables.Locks, new int[tables.CountPerLock.Length]); + _tables = newTables; + _budget = Math.Max(1, newTables.Buckets.Length / newTables.Locks.Length); } finally { - if (_lock.IsWriteLockHeld) _lock.ExitWriteLock(); + ReleaseLocks(0, locksAcquired); + } + } + + /// + /// Determines whether the contains the specified + /// item. + /// + /// The item to locate in the . + /// true if the contains the item; otherwise, false. + public bool Contains(T item) => TryGetValue(item, out _); + + /// + /// Searches the for a given value and returns the equal value it finds, if any. + /// + /// The value to search for. + /// The value from the set that the search found, or the default value of when the search yielded no match. + /// A value indicating whether the search was successful. + /// + /// This can be useful when you want to reuse a previously stored reference instead of + /// a newly constructed one (so that more sharing of references can occur) or to look up + /// a value that has more complete data than the value you currently have, although their + /// comparer functions indicate they are equal. + /// + public bool TryGetValue(T equalValue, [MaybeNullWhen(false)] out T actualValue) + { + var hashcode = _comparer.GetHashCode(equalValue); + + // We must capture the _buckets field in a local variable. It is set to a new table on each table resize. + var tables = _tables; + + var bucketNo = GetBucket(hashcode, tables.Buckets.Length); + + // We can get away w/out a lock here. + // The Volatile.Read ensures that the load of the fields of 'n' doesn't move before the load from buckets[i]. + var current = Volatile.Read(ref tables.Buckets[bucketNo]); + + while (current != null) + { + if (hashcode == current.Hashcode && _comparer.Equals(current.Item, equalValue)) + { + actualValue = current.Item; + return true; + } + + current = current.Next; + } + + actualValue = default; + return false; + } + + /// + /// Attempts to remove the item from the . + /// + /// The item to remove. + /// true if an item was removed successfully; otherwise, false. + public bool TryRemove(T item) + { + var hashcode = _comparer.GetHashCode(item); + while (true) + { + var tables = _tables; + + GetBucketAndLockNo(hashcode, out int bucketNo, out int lockNo, tables.Buckets.Length, tables.Locks.Length); + + lock (tables.Locks[lockNo]) + { + // If the table just got resized, we may not be holding the right lock, and must retry. + // This should be a rare occurrence. + if (tables != _tables) + { + continue; + } + + Node? previous = null; + for (var current = tables.Buckets[bucketNo]; current != null; current = current.Next) + { + Debug.Assert((previous == null && current == tables.Buckets[bucketNo]) || previous!.Next == current); + + if (hashcode == current.Hashcode && _comparer.Equals(current.Item, item)) + { + if (previous == null) + { + Volatile.Write(ref tables.Buckets[bucketNo], current.Next); + } + else + { + previous.Next = current.Next; + } + + tables.CountPerLock[lockNo]--; + return true; + } + previous = current; + } + } + + return false; + } + } + + IEnumerator IEnumerable.GetEnumerator() => ((IEnumerable)this).GetEnumerator(); + + /// Returns an enumerator that iterates through the . + /// An enumerator for the . + /// + /// The enumerator returned from the collection is safe to use concurrently with + /// reads and writes to the collection, however it does not represent a moment-in-time snapshot + /// of the collection. The contents exposed through the enumerator may contain modifications + /// made to the collection after was called. + /// + IEnumerator IEnumerable.GetEnumerator() => new Enumerator(this); + + /// Returns a value-type enumerator that iterates through the . + /// An enumerator for the . + /// + /// The enumerator returned from the collection is safe to use concurrently with + /// reads and writes to the collection, however it does not represent a moment-in-time snapshot + /// of the collection. The contents exposed through the enumerator may contain modifications + /// made to the collection after was called. + /// + public Enumerator GetEnumerator() => new Enumerator(this); + + /// + /// Represents an enumerator for . + /// + public struct Enumerator : IEnumerator + { + // Provides a manually-implemented version of (approximately) this iterator: + // Node?[] buckets = _tables.Buckets; + // for (int i = 0; i < buckets.Length; i++) + // for (Node? current = Volatile.Read(ref buckets[i]); current != null; current = current.Next) + // yield return new current.Item; + + private readonly ConcurrentHashSet _set; + + private Node?[]? _buckets; + private Node? _node; + private int _i; + private int _state; + + private const int StateUninitialized = 0; + private const int StateOuterloop = 1; + private const int StateInnerLoop = 2; + private const int StateDone = 3; + + /// + /// Constructs an enumerator for . + /// + public Enumerator(ConcurrentHashSet set) + { + _set = set; + _buckets = null; + _node = null; + Current = default!; + _i = -1; + _state = StateUninitialized; + } + + /// + /// Gets the element in the collection at the current position of the enumerator. + /// + /// The element in the collection at the current position of the enumerator. + public T Current { get; private set; } + + object? IEnumerator.Current => Current; + + /// + /// Sets the enumerator to its initial position, which is before the first element in the collection. + /// + public void Reset() + { + _buckets = null; + _node = null; + Current = default!; + _i = -1; + _state = StateUninitialized; + } + + /// + /// Performs application-defined tasks associated with freeing, releasing, or resetting unmanaged resources. + /// + public void Dispose() { } + + /// + /// Advances the enumerator to the next element of the collection. + /// + /// true if the enumerator was successfully advanced to the next element; false if the enumerator has passed the end of the collection. + public bool MoveNext() + { + switch (_state) + { + case StateUninitialized: + _buckets = _set._tables.Buckets; + _i = -1; + goto case StateOuterloop; + + case StateOuterloop: + Node?[]? buckets = _buckets; + Debug.Assert(buckets != null); + + int i = ++_i; + if ((uint)i < (uint)buckets!.Length) + { + // The Volatile.Read ensures that we have a copy of the reference to buckets[i]: + // this protects us from reading fields ('_key', '_value' and '_next') of different instances. + _node = Volatile.Read(ref buckets[i]); + _state = StateInnerLoop; + goto case StateInnerLoop; + } + goto default; + + case StateInnerLoop: + Node? node = _node; + if (node != null) + { + Current = node.Item; + _node = node.Next; + return true; + } + goto case StateOuterloop; + + default: + _state = StateDone; + return false; + } } } - public bool Contains(T item) + void ICollection.Add(T item) => Add(item); + + bool ICollection.IsReadOnly => false; + + void ICollection.CopyTo(T[] array, int arrayIndex) { - _lock.EnterReadLock(); + if (array == null) throw new ArgumentNullException(nameof(array)); + if (arrayIndex < 0) throw new ArgumentOutOfRangeException(nameof(arrayIndex)); + + var locksAcquired = 0; try { - return _hashSet.Contains(item); + AcquireAllLocks(ref locksAcquired); + + var count = 0; + + var countPerLock = _tables.CountPerLock; + for (var i = 0; i < countPerLock.Length && count >= 0; i++) + { + count += countPerLock[i]; + } + + if (array.Length - count < arrayIndex || count < 0) //"count" itself or "count + arrayIndex" can overflow + { + throw new ArgumentException("The index is equal to or greater than the length of the array, or the number of elements in the set is greater than the available space from index to the end of the destination array."); + } + + CopyToItems(array, arrayIndex); } finally { - if (_lock.IsReadLockHeld) _lock.ExitReadLock(); + ReleaseLocks(0, locksAcquired); } } - public bool Remove(T item) + bool ICollection.Remove(T item) => TryRemove(item); + + private void InitializeFromCollection(IEnumerable collection) { - _lock.EnterWriteLock(); + foreach (var item in collection) + { + AddInternal(item, _comparer.GetHashCode(item), false); + } + + if (_budget == 0) + { + var tables = _tables; + _budget = tables.Buckets.Length / tables.Locks.Length; + } + } + + private bool AddInternal(T item, int hashcode, bool acquireLock) + { + while (true) + { + var tables = _tables; + + GetBucketAndLockNo(hashcode, out int bucketNo, out int lockNo, tables.Buckets.Length, tables.Locks.Length); + + var resizeDesired = false; + var lockTaken = false; + try + { + if (acquireLock) + Monitor.Enter(tables.Locks[lockNo], ref lockTaken); + + // If the table just got resized, we may not be holding the right lock, and must retry. + // This should be a rare occurrence. + if (tables != _tables) + { + continue; + } + + // Try to find this item in the bucket + Node? previous = null; + for (var current = tables.Buckets[bucketNo]; current != null; current = current.Next) + { + Debug.Assert(previous == null && current == tables.Buckets[bucketNo] || previous!.Next == current); + if (hashcode == current.Hashcode && _comparer.Equals(current.Item, item)) + { + return false; + } + previous = current; + } + + // The item was not found in the bucket. Insert the new item. + Volatile.Write(ref tables.Buckets[bucketNo], new Node(item, hashcode, tables.Buckets[bucketNo])); + checked + { + tables.CountPerLock[lockNo]++; + } + + // + // If the number of elements guarded by this lock has exceeded the budget, resize the bucket table. + // It is also possible that GrowTable will increase the budget but won't resize the bucket table. + // That happens if the bucket table is found to be poorly utilized due to a bad hash function. + // + if (tables.CountPerLock[lockNo] > _budget) + { + resizeDesired = true; + } + } + finally + { + if (lockTaken) + Monitor.Exit(tables.Locks[lockNo]); + } + + // + // The fact that we got here means that we just performed an insertion. If necessary, we will grow the table. + // + // Concurrency notes: + // - Notice that we are not holding any locks at when calling GrowTable. This is necessary to prevent deadlocks. + // - As a result, it is possible that GrowTable will be called unnecessarily. But, GrowTable will obtain lock 0 + // and then verify that the table we passed to it as the argument is still the current table. + // + if (resizeDesired) + { + GrowTable(tables); + } + + return true; + } + } + + private static int GetBucket(int hashcode, int bucketCount) + { + var bucketNo = (hashcode & 0x7fffffff) % bucketCount; + Debug.Assert(bucketNo >= 0 && bucketNo < bucketCount); + return bucketNo; + } + + private static void GetBucketAndLockNo(int hashcode, out int bucketNo, out int lockNo, int bucketCount, int lockCount) + { + bucketNo = (hashcode & 0x7fffffff) % bucketCount; + lockNo = bucketNo % lockCount; + + Debug.Assert(bucketNo >= 0 && bucketNo < bucketCount); + Debug.Assert(lockNo >= 0 && lockNo < lockCount); + } + + private bool AreAllBucketsEmpty() + { + var countPerLock = _tables.CountPerLock; + for (var i = 0; i < countPerLock.Length; i++) + { + if (countPerLock[i] != 0) + { + return false; + } + } + + return true; + } + + private void GrowTable(Tables tables) + { + const int maxArrayLength = 0X7FEFFFFF; + var locksAcquired = 0; try { - return _hashSet.Remove(item); + // The thread that first obtains _locks[0] will be the one doing the resize operation + AcquireLocks(0, 1, ref locksAcquired); + + // Make sure nobody resized the table while we were waiting for lock 0: + if (tables != _tables) + { + // We assume that since the table reference is different, it was already resized (or the budget + // was adjusted). If we ever decide to do table shrinking, or replace the table for other reasons, + // we will have to revisit this logic. + return; + } + + // Compute the (approx.) total size. Use an Int64 accumulation variable to avoid an overflow. + long approxCount = 0; + for (var i = 0; i < tables.CountPerLock.Length; i++) + { + approxCount += tables.CountPerLock[i]; + } + + // + // If the bucket array is too empty, double the budget instead of resizing the table + // + if (approxCount < tables.Buckets.Length / 4) + { + _budget = 2 * _budget; + if (_budget < 0) + { + _budget = int.MaxValue; + } + return; + } + + // Compute the new table size. We find the smallest integer larger than twice the previous table size, and not divisible by + // 2,3,5 or 7. We can consider a different table-sizing policy in the future. + var newLength = 0; + var maximizeTableSize = false; + try + { + checked + { + // Double the size of the buckets table and add one, so that we have an odd integer. + newLength = tables.Buckets.Length * 2 + 1; + + // Now, we only need to check odd integers, and find the first that is not divisible + // by 3, 5 or 7. + while (newLength % 3 == 0 || newLength % 5 == 0 || newLength % 7 == 0) + { + newLength += 2; + } + + Debug.Assert(newLength % 2 != 0); + + if (newLength > maxArrayLength) + { + maximizeTableSize = true; + } + } + } + catch (OverflowException) + { + maximizeTableSize = true; + } + + if (maximizeTableSize) + { + newLength = maxArrayLength; + + // We want to make sure that GrowTable will not be called again, since table is at the maximum size. + // To achieve that, we set the budget to int.MaxValue. + // + // (There is one special case that would allow GrowTable() to be called in the future: + // calling Clear() on the ConcurrentHashSet will shrink the table and lower the budget.) + _budget = int.MaxValue; + } + + // Now acquire all other locks for the table + AcquireLocks(1, tables.Locks.Length, ref locksAcquired); + + var newLocks = tables.Locks; + + // Add more locks + if (_growLockArray && tables.Locks.Length < MaxLockNumber) + { + newLocks = new object[tables.Locks.Length * 2]; + Array.Copy(tables.Locks, newLocks, tables.Locks.Length); + for (var i = tables.Locks.Length; i < newLocks.Length; i++) + { + newLocks[i] = new object(); + } + } + + var newBuckets = new Node[newLength]; + var newCountPerLock = new int[newLocks.Length]; + + // Copy all data into a new table, creating new nodes for all elements + for (var i = 0; i < tables.Buckets.Length; i++) + { + var current = tables.Buckets[i]; + while (current != null) + { + var next = current.Next; + GetBucketAndLockNo(current.Hashcode, out int newBucketNo, out int newLockNo, newBuckets.Length, newLocks.Length); + + newBuckets[newBucketNo] = new Node(current.Item, current.Hashcode, newBuckets[newBucketNo]); + + checked + { + newCountPerLock[newLockNo]++; + } + + current = next; + } + } + + // Adjust the budget + _budget = Math.Max(1, newBuckets.Length / newLocks.Length); + + // Replace tables with the new versions + _tables = new Tables(newBuckets, newLocks, newCountPerLock); } finally { - if (_lock.IsWriteLockHeld) _lock.ExitWriteLock(); + // Release all locks that we took earlier + ReleaseLocks(0, locksAcquired); } } - public int Count + private void AcquireAllLocks(ref int locksAcquired) { - get + // First, acquire lock 0 + AcquireLocks(0, 1, ref locksAcquired); + + // Now that we have lock 0, the _locks array will not change (i.e., grow), + // and so we can safely read _locks.Length. + AcquireLocks(1, _tables.Locks.Length, ref locksAcquired); + Debug.Assert(locksAcquired == _tables.Locks.Length); + } + + private void AcquireLocks(int fromInclusive, int toExclusive, ref int locksAcquired) + { + Debug.Assert(fromInclusive <= toExclusive); + var locks = _tables.Locks; + + for (var i = fromInclusive; i < toExclusive; i++) { - _lock.EnterReadLock(); + var lockTaken = false; try { - return _hashSet.Count; + Monitor.Enter(locks[i], ref lockTaken); } finally { - if (_lock.IsReadLockHeld) _lock.ExitReadLock(); + if (lockTaken) + { + locksAcquired++; + } } } } - #endregion - #region Dispose - public void Dispose() + private void ReleaseLocks(int fromInclusive, int toExclusive) + { + Debug.Assert(fromInclusive <= toExclusive); + + for (var i = fromInclusive; i < toExclusive; i++) + { + Monitor.Exit(_tables.Locks[i]); + } + } + + private void CopyToItems(T[] array, int index) { - Dispose(true); - GC.SuppressFinalize(this); + var buckets = _tables.Buckets; + for (var i = 0; i < buckets.Length; i++) + { + for (var current = buckets[i]; current != null; current = current.Next) + { + array[index] = current.Item; + index++; //this should never flow, CopyToItems is only called when there's no overflow risk + } + } } - protected virtual void Dispose(bool disposing) + + private class Tables { - if (disposing) - if (_lock != null) - _lock.Dispose(); + public readonly Node?[] Buckets; + public readonly object[] Locks; + + public readonly int[] CountPerLock; + + public Tables(Node?[] buckets, object[] locks, int[] countPerLock) + { + Buckets = buckets; + Locks = locks; + CountPerLock = countPerLock; + } } - ~ConcurrentHashSet() + + private class Node { - Dispose(false); + public readonly T Item; + public readonly int Hashcode; + + public volatile Node? Next; + + public Node(T item, int hashcode, Node? next) + { + Item = item; + Hashcode = hashcode; + Next = next; + } } - #endregion } } diff --git a/ZeroLevel/ZeroLevel.csproj b/ZeroLevel/ZeroLevel.csproj index f8019c6..d8e744e 100644 --- a/ZeroLevel/ZeroLevel.csproj +++ b/ZeroLevel/ZeroLevel.csproj @@ -6,16 +6,16 @@ ogoun ogoun - 3.3.6.0 - New object pool implementation. Remove unused packages + 3.3.6.1 + New ConcurrentHashSet implementation https://github.com/ogoun/Zero/wiki Copyright Ogoun 2022 https://github.com/ogoun/Zero git - 3.3.6.0 - 3.3.6.0 + 3.3.6.1 + 3.3.6.1 AnyCPU;x64;x86 zero.png full