// Copyright 2020 the V8 project authors. All rights reserved. // Use of this source code is governed by a BSD-style license that can be // found in the LICENSE file. #include "include/cppgc/explicit-management.h" #include "include/cppgc/garbage-collected.h" #include "src/heap/cppgc/globals.h" #include "src/heap/cppgc/heap-base.h" #include "src/heap/cppgc/heap-object-header.h" #include "src/heap/cppgc/heap-space.h" #include "src/heap/cppgc/page-memory.h" #include "src/heap/cppgc/sweeper.h" #include "test/unittests/heap/cppgc/tests.h" #include "testing/gtest/include/gtest/gtest.h" namespace cppgc { namespace internal { class ExplicitManagementTest : public testing::TestWithHeap { public: size_t AllocatedObjectSize() const { auto* heap = Heap::From(GetHeap()); heap->stats_collector()->NotifySafePointForTesting(); return heap->stats_collector()->allocated_object_size(); } void ResetLinearAllocationBuffers() const { return Heap::From(GetHeap()) ->object_allocator() .ResetLinearAllocationBuffers(); } void TearDown() override { PreciseGC(); TestWithHeap::TearDown(); } }; namespace { class DynamicallySized final : public GarbageCollected { public: void Trace(Visitor*) const {} }; } // namespace TEST_F(ExplicitManagementTest, FreeRegularObjectToLAB) { auto* o = MakeGarbageCollected(GetHeap()->GetAllocationHandle()); const auto& space = NormalPageSpace::From(BasePage::FromPayload(o)->space()); const auto& lab = space.linear_allocation_buffer(); auto& header = HeapObjectHeader::FromObject(o); const size_t size = header.AllocatedSize(); Address needle = reinterpret_cast
(&header); // Test checks freeing to LAB. ASSERT_EQ(lab.start(), header.ObjectEnd()); const size_t lab_size_before_free = lab.size(); const size_t allocated_size_before = AllocatedObjectSize(); subtle::FreeUnreferencedObject(GetHeapHandle(), *o); EXPECT_EQ(lab.start(), reinterpret_cast
(needle)); EXPECT_EQ(lab_size_before_free + size, lab.size()); // LAB is included in allocated object size, so no change is expected. EXPECT_EQ(allocated_size_before, AllocatedObjectSize()); EXPECT_FALSE(space.free_list().ContainsForTesting({needle, size})); } TEST_F(ExplicitManagementTest, FreeRegularObjectToFreeList) { auto* o = MakeGarbageCollected(GetHeap()->GetAllocationHandle()); const auto& space = NormalPageSpace::From(BasePage::FromPayload(o)->space()); const auto& lab = space.linear_allocation_buffer(); auto& header = HeapObjectHeader::FromObject(o); const size_t size = header.AllocatedSize(); Address needle = reinterpret_cast
(&header); // Test checks freeing to free list. ResetLinearAllocationBuffers(); ASSERT_EQ(lab.start(), nullptr); const size_t allocated_size_before = AllocatedObjectSize(); subtle::FreeUnreferencedObject(GetHeapHandle(), *o); EXPECT_EQ(lab.start(), nullptr); EXPECT_EQ(allocated_size_before - size, AllocatedObjectSize()); EXPECT_TRUE(space.free_list().ContainsForTesting({needle, size})); } TEST_F(ExplicitManagementTest, FreeLargeObject) { auto* o = MakeGarbageCollected( GetHeap()->GetAllocationHandle(), AdditionalBytes(kLargeObjectSizeThreshold)); const auto* page = BasePage::FromPayload(o); auto& heap = page->heap(); ASSERT_TRUE(page->is_large()); ConstAddress needle = reinterpret_cast(o); const size_t size = LargePage::From(page)->PayloadSize(); EXPECT_TRUE(heap.page_backend()->Lookup(needle)); const size_t allocated_size_before = AllocatedObjectSize(); subtle::FreeUnreferencedObject(GetHeapHandle(), *o); EXPECT_FALSE(heap.page_backend()->Lookup(needle)); EXPECT_EQ(allocated_size_before - size, AllocatedObjectSize()); } TEST_F(ExplicitManagementTest, FreeBailsOutDuringGC) { const size_t snapshot_before = AllocatedObjectSize(); auto* o = MakeGarbageCollected(GetHeap()->GetAllocationHandle()); auto& heap = BasePage::FromPayload(o)->heap(); heap.SetInAtomicPauseForTesting(true); const size_t allocated_size_before = AllocatedObjectSize(); subtle::FreeUnreferencedObject(GetHeapHandle(), *o); EXPECT_EQ(allocated_size_before, AllocatedObjectSize()); heap.SetInAtomicPauseForTesting(false); ResetLinearAllocationBuffers(); subtle::FreeUnreferencedObject(GetHeapHandle(), *o); EXPECT_EQ(snapshot_before, AllocatedObjectSize()); } TEST_F(ExplicitManagementTest, GrowAtLAB) { auto* o = MakeGarbageCollected(GetHeap()->GetAllocationHandle()); auto& header = HeapObjectHeader::FromObject(o); ASSERT_TRUE(!header.IsLargeObject()); constexpr size_t size_of_o = sizeof(DynamicallySized); constexpr size_t kFirstDelta = 8; EXPECT_TRUE(subtle::Resize(*o, AdditionalBytes(kFirstDelta))); EXPECT_EQ(RoundUp(size_of_o + kFirstDelta), header.ObjectSize()); constexpr size_t kSecondDelta = 9; EXPECT_TRUE(subtle::Resize(*o, AdditionalBytes(kSecondDelta))); EXPECT_EQ(RoundUp(size_of_o + kSecondDelta), header.ObjectSize()); // Second round didn't actually grow object because alignment restrictions // already forced it to be large enough on the first Grow(). EXPECT_EQ(RoundUp(size_of_o + kFirstDelta), RoundUp(size_of_o + kSecondDelta)); constexpr size_t kThirdDelta = 16; EXPECT_TRUE(subtle::Resize(*o, AdditionalBytes(kThirdDelta))); EXPECT_EQ(RoundUp(size_of_o + kThirdDelta), header.ObjectSize()); } TEST_F(ExplicitManagementTest, GrowShrinkAtLAB) { auto* o = MakeGarbageCollected(GetHeap()->GetAllocationHandle()); auto& header = HeapObjectHeader::FromObject(o); ASSERT_TRUE(!header.IsLargeObject()); constexpr size_t size_of_o = sizeof(DynamicallySized); constexpr size_t kDelta = 27; EXPECT_TRUE(subtle::Resize(*o, AdditionalBytes(kDelta))); EXPECT_EQ(RoundUp(size_of_o + kDelta), header.ObjectSize()); EXPECT_TRUE(subtle::Resize(*o, AdditionalBytes(0))); EXPECT_EQ(RoundUp(size_of_o), header.ObjectSize()); } TEST_F(ExplicitManagementTest, ShrinkFreeList) { auto* o = MakeGarbageCollected( GetHeap()->GetAllocationHandle(), AdditionalBytes(ObjectAllocator::kSmallestSpaceSize)); const auto& space = NormalPageSpace::From(BasePage::FromPayload(o)->space()); // Force returning to free list by removing the LAB. ResetLinearAllocationBuffers(); auto& header = HeapObjectHeader::FromObject(o); ASSERT_TRUE(!header.IsLargeObject()); constexpr size_t size_of_o = sizeof(DynamicallySized); EXPECT_TRUE(subtle::Resize(*o, AdditionalBytes(0))); EXPECT_EQ(RoundUp(size_of_o), header.ObjectSize()); EXPECT_TRUE(space.free_list().ContainsForTesting( {header.ObjectEnd(), ObjectAllocator::kSmallestSpaceSize})); } TEST_F(ExplicitManagementTest, ShrinkFreeListBailoutAvoidFragmentation) { auto* o = MakeGarbageCollected( GetHeap()->GetAllocationHandle(), AdditionalBytes(ObjectAllocator::kSmallestSpaceSize - 1)); const auto& space = NormalPageSpace::From(BasePage::FromPayload(o)->space()); // Force returning to free list by removing the LAB. ResetLinearAllocationBuffers(); auto& header = HeapObjectHeader::FromObject(o); ASSERT_TRUE(!header.IsLargeObject()); constexpr size_t size_of_o = sizeof(DynamicallySized); EXPECT_TRUE(subtle::Resize(*o, AdditionalBytes(0))); EXPECT_EQ(RoundUp( size_of_o + ObjectAllocator::kSmallestSpaceSize - 1), header.ObjectSize()); EXPECT_FALSE(space.free_list().ContainsForTesting( {header.ObjectStart() + RoundUp(size_of_o), ObjectAllocator::kSmallestSpaceSize - 1})); } TEST_F(ExplicitManagementTest, ResizeBailsOutDuringGC) { auto* o = MakeGarbageCollected( GetHeap()->GetAllocationHandle(), AdditionalBytes(ObjectAllocator::kSmallestSpaceSize - 1)); auto& heap = BasePage::FromPayload(o)->heap(); heap.SetInAtomicPauseForTesting(true); const size_t allocated_size_before = AllocatedObjectSize(); // Grow: EXPECT_FALSE( subtle::Resize(*o, AdditionalBytes(ObjectAllocator::kSmallestSpaceSize))); // Shrink: EXPECT_FALSE(subtle::Resize(*o, AdditionalBytes(0))); EXPECT_EQ(allocated_size_before, AllocatedObjectSize()); heap.SetInAtomicPauseForTesting(false); } } // namespace internal } // namespace cppgc