| // Copyright 2015 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #ifndef THIRD_PARTY_BLINK_RENDERER_PLATFORM_HEAP_HEAP_ALLOCATOR_H_ |
| #define THIRD_PARTY_BLINK_RENDERER_PLATFORM_HEAP_HEAP_ALLOCATOR_H_ |
| |
| #include "build/build_config.h" |
| #include "third_party/blink/renderer/platform/bindings/script_wrappable_marking_visitor.h" |
| #include "third_party/blink/renderer/platform/heap/heap.h" |
| #include "third_party/blink/renderer/platform/heap/heap_buildflags.h" |
| #include "third_party/blink/renderer/platform/heap/marking_visitor.h" |
| #include "third_party/blink/renderer/platform/heap/trace_traits.h" |
| #include "third_party/blink/renderer/platform/wtf/allocator.h" |
| #include "third_party/blink/renderer/platform/wtf/assertions.h" |
| #include "third_party/blink/renderer/platform/wtf/construct_traits.h" |
| #include "third_party/blink/renderer/platform/wtf/deque.h" |
| #include "third_party/blink/renderer/platform/wtf/doubly_linked_list.h" |
| #include "third_party/blink/renderer/platform/wtf/hash_counted_set.h" |
| #include "third_party/blink/renderer/platform/wtf/hash_map.h" |
| #include "third_party/blink/renderer/platform/wtf/hash_set.h" |
| #include "third_party/blink/renderer/platform/wtf/hash_table.h" |
| #include "third_party/blink/renderer/platform/wtf/linked_hash_set.h" |
| #include "third_party/blink/renderer/platform/wtf/list_hash_set.h" |
| #include "third_party/blink/renderer/platform/wtf/type_traits.h" |
| #include "third_party/blink/renderer/platform/wtf/vector.h" |
| |
| namespace blink { |
| |
| template <typename T, typename Traits = WTF::VectorTraits<T>> |
| class HeapVectorBacking { |
| DISALLOW_NEW(); |
| IS_GARBAGE_COLLECTED_TYPE(); |
| |
| public: |
| static void Finalize(void* pointer); |
| void FinalizeGarbageCollectedObject() { Finalize(this); } |
| }; |
| |
| template <typename Table> |
| class HeapHashTableBacking { |
| DISALLOW_NEW(); |
| IS_GARBAGE_COLLECTED_TYPE(); |
| |
| public: |
| static void Finalize(void* pointer); |
| void FinalizeGarbageCollectedObject() { Finalize(this); } |
| }; |
| |
| // This is a static-only class used as a trait on collections to make them heap |
| // allocated. However see also HeapListHashSetAllocator. |
| class PLATFORM_EXPORT HeapAllocator { |
| STATIC_ONLY(HeapAllocator); |
| |
| public: |
| using Visitor = blink::Visitor; |
| static constexpr bool kIsGarbageCollected = true; |
| |
| template <typename T> |
| static size_t MaxElementCountInBackingStore() { |
| return kMaxHeapObjectSize / sizeof(T); |
| } |
| |
| template <typename T> |
| static size_t QuantizedSize(size_t count) { |
| CHECK(count <= MaxElementCountInBackingStore<T>()); |
| return ThreadHeap::AllocationSizeFromSize(count * sizeof(T)) - |
| sizeof(HeapObjectHeader); |
| } |
| template <typename T> |
| static T* AllocateVectorBacking(size_t size) { |
| ThreadState* state = |
| ThreadStateFor<ThreadingTrait<T>::kAffinity>::GetState(); |
| DCHECK(state->IsAllocationAllowed()); |
| uint32_t gc_info_index = GCInfoTrait<HeapVectorBacking<T>>::Index(); |
| NormalPageArena* arena = static_cast<NormalPageArena*>( |
| state->Heap().VectorBackingArena(gc_info_index)); |
| return reinterpret_cast<T*>(arena->AllocateObject( |
| ThreadHeap::AllocationSizeFromSize(size), gc_info_index)); |
| } |
| template <typename T> |
| static T* AllocateExpandedVectorBacking(size_t size) { |
| ThreadState* state = |
| ThreadStateFor<ThreadingTrait<T>::kAffinity>::GetState(); |
| DCHECK(state->IsAllocationAllowed()); |
| uint32_t gc_info_index = GCInfoTrait<HeapVectorBacking<T>>::Index(); |
| NormalPageArena* arena = static_cast<NormalPageArena*>( |
| state->Heap().ExpandedVectorBackingArena(gc_info_index)); |
| return reinterpret_cast<T*>(arena->AllocateObject( |
| ThreadHeap::AllocationSizeFromSize(size), gc_info_index)); |
| } |
| static void FreeVectorBacking(void*); |
| static bool ExpandVectorBacking(void*, size_t); |
| static bool ShrinkVectorBacking(void* address, |
| size_t quantized_current_size, |
| size_t quantized_shrunk_size); |
| template <typename T> |
| static T* AllocateInlineVectorBacking(size_t size) { |
| uint32_t gc_info_index = GCInfoTrait<HeapVectorBacking<T>>::Index(); |
| ThreadState* state = |
| ThreadStateFor<ThreadingTrait<T>::kAffinity>::GetState(); |
| const char* type_name = WTF_HEAP_PROFILER_TYPE_NAME(HeapVectorBacking<T>); |
| return reinterpret_cast<T*>(state->Heap().AllocateOnArenaIndex( |
| state, size, BlinkGC::kInlineVectorArenaIndex, gc_info_index, |
| type_name)); |
| } |
| static void FreeInlineVectorBacking(void*); |
| static bool ExpandInlineVectorBacking(void*, size_t); |
| static bool ShrinkInlineVectorBacking(void* address, |
| size_t quantized_current_size, |
| size_t quantized_shrunk_size); |
| |
| template <typename T, typename HashTable> |
| static T* AllocateHashTableBacking(size_t size) { |
| uint32_t gc_info_index = |
| GCInfoTrait<HeapHashTableBacking<HashTable>>::Index(); |
| ThreadState* state = |
| ThreadStateFor<ThreadingTrait<T>::kAffinity>::GetState(); |
| const char* type_name = |
| WTF_HEAP_PROFILER_TYPE_NAME(HeapHashTableBacking<HashTable>); |
| return reinterpret_cast<T*>(state->Heap().AllocateOnArenaIndex( |
| state, size, BlinkGC::kHashTableArenaIndex, gc_info_index, type_name)); |
| } |
| template <typename T, typename HashTable> |
| static T* AllocateZeroedHashTableBacking(size_t size) { |
| return AllocateHashTableBacking<T, HashTable>(size); |
| } |
| static void FreeHashTableBacking(void* address); |
| static bool ExpandHashTableBacking(void*, size_t); |
| |
| static void TraceMarkedBackingStore(void* address) { |
| MarkingVisitor::TraceMarkedBackingStore(address); |
| } |
| |
| static void BackingWriteBarrier(void* address) { |
| MarkingVisitor::WriteBarrier(address); |
| } |
| |
| template <typename T> |
| static void BackingWriteBarrier(TraceWrapperMember<T>* address, size_t size) { |
| MarkingVisitor::WriteBarrier(address); |
| ScriptWrappableMarkingVisitor::WriteBarrier(address, size); |
| } |
| |
| template <typename T> |
| static void BackingWriteBarrier(T* address, size_t size) { |
| MarkingVisitor::WriteBarrier(address); |
| } |
| |
| template <typename Return, typename Metadata> |
| static Return Malloc(size_t size, const char* type_name) { |
| return reinterpret_cast<Return>(ThreadHeap::Allocate<Metadata>( |
| size, IsEagerlyFinalizedType<Metadata>::value)); |
| } |
| |
| // Compilers sometimes eagerly instantiates the unused 'operator delete', so |
| // we provide a version that asserts and fails at run-time if used. |
| static void Free(void*) { NOTREACHED(); } |
| |
| template <typename T> |
| static void* NewArray(size_t bytes) { |
| NOTREACHED(); |
| return nullptr; |
| } |
| |
| static void DeleteArray(void* ptr) { NOTREACHED(); } |
| |
| static bool IsAllocationAllowed() { |
| return ThreadState::Current()->IsAllocationAllowed(); |
| } |
| |
| static bool IsObjectResurrectionForbidden() { |
| return ThreadState::Current()->IsObjectResurrectionForbidden(); |
| } |
| |
| static bool IsSweepForbidden() { |
| return ThreadState::Current()->SweepForbidden(); |
| } |
| |
| template <typename T> |
| static bool IsHeapObjectAlive(T* object) { |
| return ThreadHeap::IsHeapObjectAlive(object); |
| } |
| |
| template <typename VisitorDispatcher, typename T, typename Traits> |
| static void Trace(VisitorDispatcher visitor, T& t) { |
| TraceCollectionIfEnabled<Traits::kWeakHandlingFlag, T, Traits>::Trace( |
| visitor, t); |
| } |
| |
| template <typename VisitorDispatcher> |
| static bool RegisterWeakTable(VisitorDispatcher visitor, |
| const void* closure, |
| EphemeronCallback iteration_callback) { |
| return visitor->RegisterWeakTable(closure, iteration_callback); |
| } |
| |
| template <typename T, typename VisitorDispatcher> |
| static void RegisterBackingStoreCallback(VisitorDispatcher visitor, |
| T** backing_store_slot, |
| MovingObjectCallback callback, |
| void* callback_data) { |
| visitor->RegisterBackingStoreCallback( |
| reinterpret_cast<void**>(backing_store_slot), callback, callback_data); |
| } |
| |
| static void EnterGCForbiddenScope() { |
| ThreadState::Current()->EnterGCForbiddenScope(); |
| } |
| |
| static void LeaveGCForbiddenScope() { |
| ThreadState::Current()->LeaveGCForbiddenScope(); |
| } |
| |
| template <typename T, typename Traits> |
| static void NotifyNewObject(T* object) { |
| #if BUILDFLAG(BLINK_HEAP_INCREMENTAL_MARKING) |
| if (!ThreadState::IsAnyIncrementalMarking()) |
| return; |
| // The object may have been in-place constructed as part of a large object. |
| // It is not safe to retrieve the page from the object here. |
| ThreadState* const thread_state = ThreadState::Current(); |
| if (thread_state->IsIncrementalMarking()) { |
| // Eagerly trace the object ensuring that the object and all its children |
| // are discovered by the marker. |
| ThreadState::NoAllocationScope no_allocation_scope(thread_state); |
| DCHECK(thread_state->CurrentVisitor()); |
| // This check ensures that the visitor will not eagerly recurse into |
| // children but rather push all blink::GarbageCollected objects and only |
| // eagerly trace non-managed objects. |
| DCHECK(!thread_state->Heap().GetStackFrameDepth().IsEnabled()); |
| // No weak handling for write barriers. Modifying weakly reachable objects |
| // strongifies them for the current cycle. |
| DCHECK(!Traits::kCanHaveDeletedValue || !Traits::IsDeletedValue(*object)); |
| TraceCollectionIfEnabled< |
| WTF::kNoWeakHandling, T, Traits>::Trace(thread_state |
| ->CurrentVisitor(), |
| *object); |
| } |
| #endif // BUILDFLAG(BLINK_HEAP_INCREMENTAL_MARKING) |
| } |
| |
| template <typename T, typename Traits> |
| static void NotifyNewObjects(T* array, size_t len) { |
| #if BUILDFLAG(BLINK_HEAP_INCREMENTAL_MARKING) |
| if (!ThreadState::IsAnyIncrementalMarking()) |
| return; |
| // The object may have been in-place constructed as part of a large object. |
| // It is not safe to retrieve the page from the object here. |
| ThreadState* const thread_state = ThreadState::Current(); |
| if (thread_state->IsIncrementalMarking()) { |
| // See |NotifyNewObject| for details. |
| ThreadState::NoAllocationScope no_allocation_scope(thread_state); |
| DCHECK(thread_state->CurrentVisitor()); |
| DCHECK(!thread_state->Heap().GetStackFrameDepth().IsEnabled()); |
| // No weak handling for write barriers. Modifying weakly reachable objects |
| // strongifies them for the current cycle. |
| while (len-- > 0) { |
| DCHECK(!Traits::kCanHaveDeletedValue || |
| !Traits::IsDeletedValue(*array)); |
| TraceCollectionIfEnabled< |
| WTF::kNoWeakHandling, T, Traits>::Trace(thread_state |
| ->CurrentVisitor(), |
| *array); |
| array++; |
| } |
| } |
| #endif // BUILDFLAG(BLINK_HEAP_INCREMENTAL_MARKING) |
| } |
| |
| template <typename T> |
| static void TraceVectorBacking(Visitor* visitor, |
| T* backing, |
| T** backing_slot) { |
| visitor->TraceBackingStoreStrongly( |
| reinterpret_cast<HeapVectorBacking<T>*>(backing), |
| reinterpret_cast<HeapVectorBacking<T>**>(backing_slot)); |
| } |
| |
| template <typename T, typename HashTable> |
| static void TraceHashTableBackingStrongly(Visitor* visitor, |
| T* backing, |
| T** backing_slot) { |
| visitor->TraceBackingStoreStrongly( |
| reinterpret_cast<HeapHashTableBacking<HashTable>*>(backing), |
| reinterpret_cast<HeapHashTableBacking<HashTable>**>(backing_slot)); |
| } |
| |
| template <typename T, typename HashTable> |
| static void TraceHashTableBackingWeakly(Visitor* visitor, |
| T* backing, |
| T** backing_slot, |
| WeakCallback callback, |
| void* parameter) { |
| visitor->TraceBackingStoreWeakly( |
| reinterpret_cast<HeapHashTableBacking<HashTable>*>(backing), |
| reinterpret_cast<HeapHashTableBacking<HashTable>**>(backing_slot), |
| callback, parameter); |
| } |
| |
| template <typename T, typename HashTable> |
| static void TraceHashTableBackingOnly(Visitor* visitor, |
| T* backing, |
| T** backing_slot) { |
| visitor->TraceBackingStoreOnly( |
| reinterpret_cast<HeapHashTableBacking<HashTable>*>(backing), |
| reinterpret_cast<HeapHashTableBacking<HashTable>**>(backing_slot)); |
| } |
| |
| private: |
| static void BackingFree(void*); |
| static bool BackingExpand(void*, size_t); |
| static bool BackingShrink(void*, |
| size_t quantized_current_size, |
| size_t quantized_shrunk_size); |
| |
| template <typename T, wtf_size_t u, typename V> |
| friend class WTF::Vector; |
| template <typename T, typename U, typename V, typename W> |
| friend class WTF::HashSet; |
| template <typename T, |
| typename U, |
| typename V, |
| typename W, |
| typename X, |
| typename Y> |
| friend class WTF::HashMap; |
| }; |
| |
| template <typename VisitorDispatcher, typename Value> |
| static void TraceListHashSetValue(VisitorDispatcher visitor, Value& value) { |
| // We use the default hash traits for the value in the node, because |
| // ListHashSet does not let you specify any specific ones. |
| // We don't allow ListHashSet of WeakMember, so we set that one false |
| // (there's an assert elsewhere), but we have to specify some value for the |
| // strongify template argument, so we specify WTF::WeakPointersActWeak, |
| // arbitrarily. |
| TraceCollectionIfEnabled<WTF::kNoWeakHandling, Value, |
| WTF::HashTraits<Value>>::Trace(visitor, value); |
| } |
| |
| // The inline capacity is just a dummy template argument to match the off-heap |
| // allocator. |
| // This inherits from the static-only HeapAllocator trait class, but we do |
| // declare pointers to instances. These pointers are always null, and no |
| // objects are instantiated. |
| template <typename ValueArg, wtf_size_t inlineCapacity> |
| class HeapListHashSetAllocator : public HeapAllocator { |
| DISALLOW_NEW(); |
| |
| public: |
| using TableAllocator = HeapAllocator; |
| using Node = WTF::ListHashSetNode<ValueArg, HeapListHashSetAllocator>; |
| |
| class AllocatorProvider { |
| DISALLOW_NEW(); |
| |
| public: |
| // For the heap allocation we don't need an actual allocator object, so |
| // we just return null. |
| HeapListHashSetAllocator* Get() const { return nullptr; } |
| |
| // No allocator object is needed. |
| void CreateAllocatorIfNeeded() {} |
| void ReleaseAllocator() {} |
| |
| // There is no allocator object in the HeapListHashSet (unlike in the |
| // regular ListHashSet) so there is nothing to swap. |
| void Swap(AllocatorProvider& other) {} |
| }; |
| |
| void Deallocate(void* dummy) {} |
| |
| // This is not a static method even though it could be, because it needs to |
| // match the one that the (off-heap) ListHashSetAllocator has. The 'this' |
| // pointer will always be null. |
| void* AllocateNode() { |
| // Consider using a LinkedHashSet instead if this compile-time assert fails: |
| static_assert(!WTF::IsWeak<ValueArg>::value, |
| "weak pointers in a ListHashSet will result in null entries " |
| "in the set"); |
| |
| return Malloc<void*, Node>( |
| sizeof(Node), |
| nullptr /* Oilpan does not use the heap profiler at the moment. */); |
| } |
| |
| template <typename VisitorDispatcher> |
| static void TraceValue(VisitorDispatcher visitor, Node* node) { |
| TraceListHashSetValue(visitor, node->value_); |
| } |
| }; |
| |
| template <typename T, typename Traits> |
| void HeapVectorBacking<T, Traits>::Finalize(void* pointer) { |
| static_assert(Traits::kNeedsDestruction, |
| "Only vector buffers with items requiring destruction should " |
| "be finalized"); |
| // See the comment in HeapVectorBacking::trace. |
| static_assert( |
| Traits::kCanClearUnusedSlotsWithMemset || std::is_polymorphic<T>::value, |
| "HeapVectorBacking doesn't support objects that cannot be cleared as " |
| "unused with memset or don't have a vtable"); |
| |
| static_assert( |
| !std::is_trivially_destructible<T>::value, |
| "Finalization of trivially destructible classes should not happen."); |
| HeapObjectHeader* header = HeapObjectHeader::FromPayload(pointer); |
| // Use the payload size as recorded by the heap to determine how many |
| // elements to finalize. |
| size_t length = header->PayloadSize() / sizeof(T); |
| char* payload = static_cast<char*>(pointer); |
| #ifdef ANNOTATE_CONTIGUOUS_CONTAINER |
| ANNOTATE_CHANGE_SIZE(payload, length * sizeof(T), 0, length * sizeof(T)); |
| #endif |
| // As commented above, HeapVectorBacking calls finalizers for unused slots |
| // (which are already zeroed out). |
| if (std::is_polymorphic<T>::value) { |
| for (unsigned i = 0; i < length; ++i) { |
| char* element = payload + i * sizeof(T); |
| if (blink::VTableInitialized(element)) |
| reinterpret_cast<T*>(element)->~T(); |
| } |
| } else { |
| T* buffer = reinterpret_cast<T*>(payload); |
| for (unsigned i = 0; i < length; ++i) |
| buffer[i].~T(); |
| } |
| } |
| |
| template <typename Table> |
| void HeapHashTableBacking<Table>::Finalize(void* pointer) { |
| using Value = typename Table::ValueType; |
| static_assert( |
| !std::is_trivially_destructible<Value>::value, |
| "Finalization of trivially destructible classes should not happen."); |
| HeapObjectHeader* header = HeapObjectHeader::FromPayload(pointer); |
| // Use the payload size as recorded by the heap to determine how many |
| // elements to finalize. |
| size_t length = header->PayloadSize() / sizeof(Value); |
| Value* table = reinterpret_cast<Value*>(pointer); |
| for (unsigned i = 0; i < length; ++i) { |
| if (!Table::IsEmptyOrDeletedBucket(table[i])) |
| table[i].~Value(); |
| } |
| } |
| |
| template <typename KeyArg, |
| typename MappedArg, |
| typename HashArg = typename DefaultHash<KeyArg>::Hash, |
| typename KeyTraitsArg = HashTraits<KeyArg>, |
| typename MappedTraitsArg = HashTraits<MappedArg>> |
| class HeapHashMap : public HashMap<KeyArg, |
| MappedArg, |
| HashArg, |
| KeyTraitsArg, |
| MappedTraitsArg, |
| HeapAllocator> { |
| IS_GARBAGE_COLLECTED_TYPE(); |
| using Base = |
| HashMap<KeyArg, MappedArg, HashArg, KeyTraitsArg, MappedTraitsArg>; |
| static_assert(WTF::IsTraceable<KeyArg>::value || |
| WTF::IsTraceable<MappedArg>::value, |
| "For hash maps without traceable elements, use HashMap<> " |
| "instead of HeapHashMap<>"); |
| |
| public: |
| static void* AllocateObject(size_t size, bool eagerly_sweep) { |
| return ThreadHeap::Allocate< |
| HeapHashMap<KeyArg, MappedArg, HashArg, KeyTraitsArg, MappedTraitsArg>>( |
| size, eagerly_sweep); |
| } |
| |
| void* operator new(size_t size) = delete; |
| void operator delete(void* p) = delete; |
| void* operator new[](size_t size) = delete; |
| void operator delete[](void* p) = delete; |
| void* operator new(size_t size, NotNullTag null_tag, void* location) { |
| return Base::operator new(size, null_tag, location); |
| } |
| void* operator new(size_t size, void* location) { |
| return Base::operator new(size, location); |
| } |
| }; |
| |
| template <typename ValueArg, |
| typename HashArg = typename DefaultHash<ValueArg>::Hash, |
| typename TraitsArg = HashTraits<ValueArg>> |
| class HeapHashSet |
| : public HashSet<ValueArg, HashArg, TraitsArg, HeapAllocator> { |
| IS_GARBAGE_COLLECTED_TYPE(); |
| using Base = HashSet<ValueArg, HashArg, TraitsArg>; |
| static_assert(WTF::IsTraceable<ValueArg>::value, |
| "For hash sets without traceable elements, use HashSet<> " |
| "instead of HeapHashSet<>"); |
| |
| public: |
| static void* AllocateObject(size_t size, bool eagerly_sweep) { |
| return ThreadHeap::Allocate<HeapHashSet<ValueArg, HashArg, TraitsArg>>( |
| size, eagerly_sweep); |
| } |
| |
| void* operator new(size_t size) = delete; |
| void operator delete(void* p) = delete; |
| void* operator new[](size_t size) = delete; |
| void operator delete[](void* p) = delete; |
| void* operator new(size_t size, NotNullTag null_tag, void* location) { |
| return Base::operator new(size, null_tag, location); |
| } |
| void* operator new(size_t size, void* location) { |
| return Base::operator new(size, location); |
| } |
| }; |
| |
| template <typename ValueArg, |
| typename HashArg = typename DefaultHash<ValueArg>::Hash, |
| typename TraitsArg = HashTraits<ValueArg>> |
| class HeapLinkedHashSet |
| : public LinkedHashSet<ValueArg, HashArg, TraitsArg, HeapAllocator> { |
| IS_GARBAGE_COLLECTED_TYPE(); |
| using Base = LinkedHashSet<ValueArg, HashArg, TraitsArg>; |
| static_assert(WTF::IsTraceable<ValueArg>::value, |
| "For sets without traceable elements, use LinkedHashSet<> " |
| "instead of HeapLinkedHashSet<>"); |
| |
| public: |
| static void* AllocateObject(size_t size, bool eagerly_sweep) { |
| return ThreadHeap::Allocate< |
| HeapLinkedHashSet<ValueArg, HashArg, TraitsArg>>(size, eagerly_sweep); |
| } |
| |
| void* operator new(size_t size) = delete; |
| void operator delete(void* p) = delete; |
| void* operator new[](size_t size) = delete; |
| void operator delete[](void* p) = delete; |
| void* operator new(size_t size, NotNullTag null_tag, void* location) { |
| return Base::operator new(size, null_tag, location); |
| } |
| void* operator new(size_t size, void* location) { |
| return Base::operator new(size, location); |
| } |
| }; |
| |
| template <typename ValueArg, |
| wtf_size_t inlineCapacity = |
| 0, // The inlineCapacity is just a dummy to |
| // match ListHashSet (off-heap). |
| typename HashArg = typename DefaultHash<ValueArg>::Hash> |
| class HeapListHashSet |
| : public ListHashSet<ValueArg, |
| inlineCapacity, |
| HashArg, |
| HeapListHashSetAllocator<ValueArg, inlineCapacity>> { |
| IS_GARBAGE_COLLECTED_TYPE(); |
| using Base = ListHashSet<ValueArg, inlineCapacity, HashArg>; |
| static_assert(WTF::IsTraceable<ValueArg>::value, |
| "For sets without traceable elements, use ListHashSet<> " |
| "instead of HeapListHashSet<>"); |
| |
| public: |
| static void* AllocateObject(size_t size, bool eagerly_sweep) { |
| return ThreadHeap::Allocate< |
| HeapListHashSet<ValueArg, inlineCapacity, HashArg>>(size, |
| eagerly_sweep); |
| } |
| |
| void* operator new(size_t size) = delete; |
| void operator delete(void* p) = delete; |
| void* operator new[](size_t size) = delete; |
| void operator delete[](void* p) = delete; |
| void* operator new(size_t size, NotNullTag null_tag, void* location) { |
| return Base::operator new(size, null_tag, location); |
| } |
| void* operator new(size_t size, void* location) { |
| return Base::operator new(size, location); |
| } |
| }; |
| |
| template <typename Value, |
| typename HashFunctions = typename DefaultHash<Value>::Hash, |
| typename Traits = HashTraits<Value>> |
| class HeapHashCountedSet |
| : public HashCountedSet<Value, HashFunctions, Traits, HeapAllocator> { |
| IS_GARBAGE_COLLECTED_TYPE(); |
| static_assert(WTF::IsTraceable<Value>::value, |
| "For counted sets without traceable elements, use " |
| "HashCountedSet<> instead of HeapHashCountedSet<>"); |
| }; |
| |
| template <typename T, wtf_size_t inlineCapacity = 0> |
| class HeapVector : public Vector<T, inlineCapacity, HeapAllocator> { |
| IS_GARBAGE_COLLECTED_TYPE(); |
| using Base = Vector<T, inlineCapacity, HeapAllocator>; |
| |
| public: |
| HeapVector() { |
| static_assert(WTF::IsTraceable<T>::value, |
| "For vectors without traceable elements, use Vector<> " |
| "instead of HeapVector<>"); |
| } |
| |
| static void* AllocateObject(size_t size, bool eagerly_sweep) { |
| // On-heap HeapVectors generally should not have inline capacity, but it is |
| // hard to avoid when using a type alias. Hence we only disallow the |
| // VectorTraits<T>::kNeedsDestruction case for now. |
| static_assert(inlineCapacity == 0 || !VectorTraits<T>::kNeedsDestruction, |
| "on-heap HeapVector<> should not have an inline capacity"); |
| return ThreadHeap::Allocate<HeapVector<T, inlineCapacity>>(size, |
| eagerly_sweep); |
| } |
| |
| void* operator new(size_t size) = delete; |
| void operator delete(void* p) = delete; |
| void* operator new[](size_t size) = delete; |
| void operator delete[](void* p) = delete; |
| void* operator new(size_t size, NotNullTag null_tag, void* location) { |
| return Base::operator new(size, null_tag, location); |
| } |
| void* operator new(size_t size, void* location) { |
| return Base::operator new(size, location); |
| } |
| |
| explicit HeapVector(wtf_size_t size) |
| : Vector<T, inlineCapacity, HeapAllocator>(size) {} |
| |
| HeapVector(wtf_size_t size, const T& val) |
| : Vector<T, inlineCapacity, HeapAllocator>(size, val) {} |
| |
| template <wtf_size_t otherCapacity> |
| HeapVector(const HeapVector<T, otherCapacity>& other) |
| : Vector<T, inlineCapacity, HeapAllocator>(other) {} |
| }; |
| |
| template <typename T, wtf_size_t inlineCapacity = 0> |
| class HeapDeque : public Deque<T, inlineCapacity, HeapAllocator> { |
| IS_GARBAGE_COLLECTED_TYPE(); |
| using Base = Deque<T, inlineCapacity, HeapAllocator>; |
| |
| public: |
| HeapDeque() { |
| static_assert(WTF::IsTraceable<T>::value, |
| "For vectors without traceable elements, use Deque<> instead " |
| "of HeapDeque<>"); |
| } |
| |
| static void* AllocateObject(size_t size, bool eagerly_sweep) { |
| // On-heap HeapDeques generally should not have inline capacity, but it is |
| // hard to avoid when using a type alias. Hence we only disallow the |
| // VectorTraits<T>::kNeedsDestruction case for now. |
| static_assert(inlineCapacity == 0 || !VectorTraits<T>::kNeedsDestruction, |
| "on-heap HeapDeque<> should not have an inline capacity"); |
| return ThreadHeap::Allocate<HeapVector<T, inlineCapacity>>(size, |
| eagerly_sweep); |
| } |
| |
| void* operator new(size_t size) = delete; |
| void operator delete(void* p) = delete; |
| void* operator new[](size_t size) = delete; |
| void operator delete[](void* p) = delete; |
| void* operator new(size_t size, NotNullTag null_tag, void* location) { |
| return Base::operator new(size, null_tag, location); |
| } |
| void* operator new(size_t size, void* location) { |
| return Base::operator new(size, location); |
| } |
| |
| explicit HeapDeque(wtf_size_t size) |
| : Deque<T, inlineCapacity, HeapAllocator>(size) {} |
| |
| HeapDeque(wtf_size_t size, const T& val) |
| : Deque<T, inlineCapacity, HeapAllocator>(size, val) {} |
| |
| HeapDeque& operator=(const HeapDeque& other) { |
| HeapDeque<T> copy(other); |
| Deque<T, inlineCapacity, HeapAllocator>::Swap(copy); |
| return *this; |
| } |
| |
| template <wtf_size_t otherCapacity> |
| HeapDeque(const HeapDeque<T, otherCapacity>& other) |
| : Deque<T, inlineCapacity, HeapAllocator>(other) {} |
| }; |
| |
| template <typename T> |
| class HeapDoublyLinkedList : public DoublyLinkedList<T, Member<T>> { |
| IS_GARBAGE_COLLECTED_TYPE(); |
| DISALLOW_NEW(); |
| |
| public: |
| HeapDoublyLinkedList() { |
| static_assert(WTF::IsGarbageCollectedType<T>::value, |
| "This should only be used for garbage collected types."); |
| } |
| |
| void Trace(Visitor* visitor) { |
| visitor->Trace(this->head_); |
| visitor->Trace(this->tail_); |
| } |
| }; |
| |
| } // namespace blink |
| |
| namespace WTF { |
| |
| template <typename T> |
| struct VectorTraits<blink::Member<T>> : VectorTraitsBase<blink::Member<T>> { |
| STATIC_ONLY(VectorTraits); |
| static const bool kNeedsDestruction = false; |
| static const bool kCanInitializeWithMemset = true; |
| static const bool kCanClearUnusedSlotsWithMemset = true; |
| static const bool kCanCopyWithMemcpy = true; |
| static const bool kCanMoveWithMemcpy = true; |
| }; |
| |
| template <typename T> |
| struct VectorTraits<blink::SameThreadCheckedMember<T>> |
| : VectorTraitsBase<blink::SameThreadCheckedMember<T>> { |
| STATIC_ONLY(VectorTraits); |
| static const bool kNeedsDestruction = false; |
| static const bool kCanInitializeWithMemset = true; |
| static const bool kCanClearUnusedSlotsWithMemset = true; |
| static const bool kCanMoveWithMemcpy = true; |
| static const bool kCanSwapUsingCopyOrMove = false; |
| }; |
| |
| template <typename T> |
| struct VectorTraits<blink::TraceWrapperMember<T>> |
| : VectorTraitsBase<blink::TraceWrapperMember<T>> { |
| STATIC_ONLY(VectorTraits); |
| static const bool kNeedsDestruction = false; |
| static const bool kCanInitializeWithMemset = true; |
| static const bool kCanClearUnusedSlotsWithMemset = true; |
| static const bool kCanMoveWithMemcpy = true; |
| static const bool kCanCopyWithMemcpy = true; |
| static const bool kCanSwapUsingCopyOrMove = true; |
| }; |
| |
| template <typename T> |
| struct VectorTraits<blink::WeakMember<T>> |
| : VectorTraitsBase<blink::WeakMember<T>> { |
| STATIC_ONLY(VectorTraits); |
| static const bool kNeedsDestruction = false; |
| static const bool kCanInitializeWithMemset = true; |
| static const bool kCanClearUnusedSlotsWithMemset = true; |
| static const bool kCanMoveWithMemcpy = true; |
| }; |
| |
| template <typename T> |
| struct VectorTraits<blink::UntracedMember<T>> |
| : VectorTraitsBase<blink::UntracedMember<T>> { |
| STATIC_ONLY(VectorTraits); |
| static const bool kNeedsDestruction = false; |
| static const bool kCanInitializeWithMemset = true; |
| static const bool kCanClearUnusedSlotsWithMemset = true; |
| static const bool kCanMoveWithMemcpy = true; |
| }; |
| |
| template <typename T> |
| struct VectorTraits<blink::HeapVector<T, 0>> |
| : VectorTraitsBase<blink::HeapVector<T, 0>> { |
| STATIC_ONLY(VectorTraits); |
| static const bool kNeedsDestruction = false; |
| static const bool kCanInitializeWithMemset = true; |
| static const bool kCanClearUnusedSlotsWithMemset = true; |
| static const bool kCanMoveWithMemcpy = true; |
| }; |
| |
| template <typename T> |
| struct VectorTraits<blink::HeapDeque<T, 0>> |
| : VectorTraitsBase<blink::HeapDeque<T, 0>> { |
| STATIC_ONLY(VectorTraits); |
| static const bool kNeedsDestruction = false; |
| static const bool kCanInitializeWithMemset = true; |
| static const bool kCanClearUnusedSlotsWithMemset = true; |
| static const bool kCanMoveWithMemcpy = true; |
| }; |
| |
| template <typename T, wtf_size_t inlineCapacity> |
| struct VectorTraits<blink::HeapVector<T, inlineCapacity>> |
| : VectorTraitsBase<blink::HeapVector<T, inlineCapacity>> { |
| STATIC_ONLY(VectorTraits); |
| static const bool kNeedsDestruction = VectorTraits<T>::kNeedsDestruction; |
| static const bool kCanInitializeWithMemset = |
| VectorTraits<T>::kCanInitializeWithMemset; |
| static const bool kCanClearUnusedSlotsWithMemset = |
| VectorTraits<T>::kCanClearUnusedSlotsWithMemset; |
| static const bool kCanMoveWithMemcpy = VectorTraits<T>::kCanMoveWithMemcpy; |
| }; |
| |
| template <typename T, wtf_size_t inlineCapacity> |
| struct VectorTraits<blink::HeapDeque<T, inlineCapacity>> |
| : VectorTraitsBase<blink::HeapDeque<T, inlineCapacity>> { |
| STATIC_ONLY(VectorTraits); |
| static const bool kNeedsDestruction = VectorTraits<T>::kNeedsDestruction; |
| static const bool kCanInitializeWithMemset = |
| VectorTraits<T>::kCanInitializeWithMemset; |
| static const bool kCanClearUnusedSlotsWithMemset = |
| VectorTraits<T>::kCanClearUnusedSlotsWithMemset; |
| static const bool kCanMoveWithMemcpy = VectorTraits<T>::kCanMoveWithMemcpy; |
| }; |
| |
| template <typename T> |
| struct HashTraits<blink::Member<T>> : SimpleClassHashTraits<blink::Member<T>> { |
| STATIC_ONLY(HashTraits); |
| // FIXME: Implement proper const'ness for iterator types. Requires support |
| // in the marking Visitor. |
| using PeekInType = T*; |
| using IteratorGetType = blink::Member<T>*; |
| using IteratorConstGetType = const blink::Member<T>*; |
| using IteratorReferenceType = blink::Member<T>&; |
| using IteratorConstReferenceType = const blink::Member<T>&; |
| static IteratorReferenceType GetToReferenceConversion(IteratorGetType x) { |
| return *x; |
| } |
| static IteratorConstReferenceType GetToReferenceConstConversion( |
| IteratorConstGetType x) { |
| return *x; |
| } |
| |
| using PeekOutType = T*; |
| |
| template <typename U> |
| static void Store(const U& value, blink::Member<T>& storage) { |
| storage = value; |
| } |
| |
| static PeekOutType Peek(const blink::Member<T>& value) { return value; } |
| |
| static void ConstructDeletedValue(blink::Member<T>& slot, bool) { |
| slot = WTF::kHashTableDeletedValue; |
| } |
| static bool IsDeletedValue(const blink::Member<T>& value) { |
| return value.IsHashTableDeletedValue(); |
| } |
| }; |
| |
| template <typename T> |
| struct HashTraits<blink::SameThreadCheckedMember<T>> |
| : SimpleClassHashTraits<blink::SameThreadCheckedMember<T>> { |
| STATIC_ONLY(HashTraits); |
| // FIXME: Implement proper const'ness for iterator types. Requires support |
| // in the marking Visitor. |
| using PeekInType = T*; |
| using IteratorGetType = blink::SameThreadCheckedMember<T>*; |
| using IteratorConstGetType = const blink::SameThreadCheckedMember<T>*; |
| using IteratorReferenceType = blink::SameThreadCheckedMember<T>&; |
| using IteratorConstReferenceType = const blink::SameThreadCheckedMember<T>&; |
| static IteratorReferenceType GetToReferenceConversion(IteratorGetType x) { |
| return *x; |
| } |
| static IteratorConstReferenceType GetToReferenceConstConversion( |
| IteratorConstGetType x) { |
| return *x; |
| } |
| |
| using PeekOutType = T*; |
| |
| template <typename U> |
| static void Store(const U& value, |
| blink::SameThreadCheckedMember<T>& storage) { |
| storage = value; |
| } |
| |
| static PeekOutType Peek(const blink::SameThreadCheckedMember<T>& value) { |
| return value; |
| } |
| |
| static blink::SameThreadCheckedMember<T> EmptyValue() { |
| return blink::SameThreadCheckedMember<T>(nullptr, nullptr); |
| } |
| }; |
| |
| template <typename T> |
| struct HashTraits<blink::TraceWrapperMember<T>> |
| : SimpleClassHashTraits<blink::TraceWrapperMember<T>> { |
| STATIC_ONLY(HashTraits); |
| // FIXME: Implement proper const'ness for iterator types. Requires support |
| // in the marking Visitor. |
| using PeekInType = T*; |
| using IteratorGetType = blink::TraceWrapperMember<T>*; |
| using IteratorConstGetType = const blink::TraceWrapperMember<T>*; |
| using IteratorReferenceType = blink::TraceWrapperMember<T>&; |
| using IteratorConstReferenceType = const blink::TraceWrapperMember<T>&; |
| static IteratorReferenceType GetToReferenceConversion(IteratorGetType x) { |
| return *x; |
| } |
| static IteratorConstReferenceType GetToReferenceConstConversion( |
| IteratorConstGetType x) { |
| return *x; |
| } |
| |
| using PeekOutType = T*; |
| |
| template <typename U> |
| static void Store(const U& value, blink::TraceWrapperMember<T>& storage) { |
| storage = value; |
| } |
| |
| static PeekOutType Peek(const blink::TraceWrapperMember<T>& value) { |
| return value; |
| } |
| |
| static blink::TraceWrapperMember<T> EmptyValue() { return nullptr; } |
| }; |
| |
| template <typename T> |
| struct HashTraits<blink::WeakMember<T>> |
| : SimpleClassHashTraits<blink::WeakMember<T>> { |
| STATIC_ONLY(HashTraits); |
| static const bool kNeedsDestruction = false; |
| // FIXME: Implement proper const'ness for iterator types. Requires support |
| // in the marking Visitor. |
| using PeekInType = T*; |
| using IteratorGetType = blink::WeakMember<T>*; |
| using IteratorConstGetType = const blink::WeakMember<T>*; |
| using IteratorReferenceType = blink::WeakMember<T>&; |
| using IteratorConstReferenceType = const blink::WeakMember<T>&; |
| static IteratorReferenceType GetToReferenceConversion(IteratorGetType x) { |
| return *x; |
| } |
| static IteratorConstReferenceType GetToReferenceConstConversion( |
| IteratorConstGetType x) { |
| return *x; |
| } |
| |
| using PeekOutType = T*; |
| |
| template <typename U> |
| static void Store(const U& value, blink::WeakMember<T>& storage) { |
| storage = value; |
| } |
| |
| static PeekOutType Peek(const blink::WeakMember<T>& value) { return value; } |
| |
| static bool IsAlive(blink::WeakMember<T>& weak_member) { |
| return blink::ThreadHeap::IsHeapObjectAlive(weak_member); |
| } |
| |
| template <typename VisitorDispatcher> |
| static bool TraceInCollection(VisitorDispatcher visitor, |
| blink::WeakMember<T>& weak_member, |
| WeakHandlingFlag weakness) { |
| if (weakness == kNoWeakHandling) { |
| visitor->Trace(weak_member.Get()); // Strongified visit. |
| return false; |
| } |
| return !blink::ThreadHeap::IsHeapObjectAlive(weak_member); |
| } |
| }; |
| |
| template <typename T> |
| struct HashTraits<blink::UntracedMember<T>> |
| : SimpleClassHashTraits<blink::UntracedMember<T>> { |
| STATIC_ONLY(HashTraits); |
| static const bool kNeedsDestruction = false; |
| // FIXME: Implement proper const'ness for iterator types. |
| using PeekInType = T*; |
| using IteratorGetType = blink::UntracedMember<T>*; |
| using IteratorConstGetType = const blink::UntracedMember<T>*; |
| using IteratorReferenceType = blink::UntracedMember<T>&; |
| using IteratorConstReferenceType = const blink::UntracedMember<T>&; |
| static IteratorReferenceType GetToReferenceConversion(IteratorGetType x) { |
| return *x; |
| } |
| static IteratorConstReferenceType GetToReferenceConstConversion( |
| IteratorConstGetType x) { |
| return *x; |
| } |
| using PeekOutType = T*; |
| |
| template <typename U> |
| static void Store(const U& value, blink::UntracedMember<T>& storage) { |
| storage = value; |
| } |
| |
| static PeekOutType Peek(const blink::UntracedMember<T>& value) { |
| return value; |
| } |
| }; |
| |
| template <typename T, wtf_size_t inlineCapacity> |
| struct IsTraceable< |
| ListHashSetNode<T, blink::HeapListHashSetAllocator<T, inlineCapacity>>*> { |
| STATIC_ONLY(IsTraceable); |
| static_assert(sizeof(T), "T must be fully defined"); |
| // All heap allocated node pointers need visiting to keep the nodes alive, |
| // regardless of whether they contain pointers to other heap allocated |
| // objects. |
| static const bool value = true; |
| }; |
| |
| template <typename T, wtf_size_t inlineCapacity> |
| struct IsGarbageCollectedType< |
| ListHashSetNode<T, blink::HeapListHashSetAllocator<T, inlineCapacity>>> { |
| static const bool value = true; |
| }; |
| |
| template <typename Set> |
| struct IsGarbageCollectedType<ListHashSetIterator<Set>> { |
| static const bool value = IsGarbageCollectedType<Set>::value; |
| }; |
| |
| template <typename Set> |
| struct IsGarbageCollectedType<ListHashSetConstIterator<Set>> { |
| static const bool value = IsGarbageCollectedType<Set>::value; |
| }; |
| |
| template <typename Set> |
| struct IsGarbageCollectedType<ListHashSetReverseIterator<Set>> { |
| static const bool value = IsGarbageCollectedType<Set>::value; |
| }; |
| |
| template <typename Set> |
| struct IsGarbageCollectedType<ListHashSetConstReverseIterator<Set>> { |
| static const bool value = IsGarbageCollectedType<Set>::value; |
| }; |
| |
| template <typename T, typename H> |
| struct HandleHashTraits : SimpleClassHashTraits<H> { |
| STATIC_ONLY(HandleHashTraits); |
| // TODO: Implement proper const'ness for iterator types. Requires support |
| // in the marking Visitor. |
| using PeekInType = T*; |
| using IteratorGetType = H*; |
| using IteratorConstGetType = const H*; |
| using IteratorReferenceType = H&; |
| using IteratorConstReferenceType = const H&; |
| static IteratorReferenceType GetToReferenceConversion(IteratorGetType x) { |
| return *x; |
| } |
| static IteratorConstReferenceType GetToReferenceConstConversion( |
| IteratorConstGetType x) { |
| return *x; |
| } |
| |
| using PeekOutType = T*; |
| |
| template <typename U> |
| static void Store(const U& value, H& storage) { |
| storage = value; |
| } |
| |
| static PeekOutType Peek(const H& value) { return value; } |
| }; |
| |
| template <typename Value, |
| typename HashFunctions, |
| typename Traits, |
| typename VectorType> |
| inline void CopyToVector( |
| const blink::HeapHashCountedSet<Value, HashFunctions, Traits>& set, |
| VectorType& vector) { |
| CopyToVector(static_cast<const HashCountedSet<Value, HashFunctions, Traits, |
| blink::HeapAllocator>&>(set), |
| vector); |
| } |
| |
| } // namespace WTF |
| |
| #endif |