diff --git a/dwio/nimble/common/Buffer.h b/dwio/nimble/common/Buffer.h index 6408ca5..464671f 100644 --- a/dwio/nimble/common/Buffer.h +++ b/dwio/nimble/common/Buffer.h @@ -15,6 +15,7 @@ */ #pragma once +#include "dwio/nimble/common/Exceptions.h" #include "velox/buffer/Buffer.h" #include "velox/common/memory/Memory.h" @@ -34,11 +35,10 @@ namespace facebook::nimble { -// Internally manages memory in chunks. Releases memory only upon destruction. +// Internally manages memory in chunks. releases memory when destroyed // Buffer is NOT threadsafe: external locking is required. class Buffer { - using MemoryPool = facebook::velox::memory::MemoryPool; - + using MemoryPool = facebook::velox::memory::MemoryPool; public: explicit Buffer( MemoryPool& memoryPool, @@ -52,7 +52,6 @@ class Buffer { // to, and guarantees for the lifetime of *this that that region will remain // valid. Does NOT guarantee that the region is initially 0'd. char* reserve(uint64_t bytes) { - std::scoped_lock l(mutex_); if (reserveEnd_ + bytes <= chunkEnd_) { pos_ = reserveEnd_; reserveEnd_ += bytes; @@ -98,11 +97,6 @@ class Buffer { char* reserveEnd_; std::vector chunks_; MemoryPool& memoryPool_; - // NOTE: this is temporary fix, to quickly enable parallel access to the - // buffer class. In the near future, we are going to templetize this class to - // produce a concurrent and a non-concurrent variants, and change the call - // sites to use each variant when needed. - std::mutex mutex_; }; } // namespace facebook::nimble diff --git a/dwio/nimble/common/tests/BufferPoolTests.cpp b/dwio/nimble/common/tests/BufferPoolTests.cpp new file mode 100644 index 0000000..ec2d383 --- /dev/null +++ b/dwio/nimble/common/tests/BufferPoolTests.cpp @@ -0,0 +1,96 @@ +/* + * Copyright (c) Meta Platforms, Inc. and its affiliates. + * + * Licensed under the Apache License, Version 2.0 (the "License"); + * you may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +#include +#include "dwio/nimble/common/Buffer.h" +#include "dwio/nimble/common/Exceptions.h" +#include "dwio/nimble/velox/FieldWriter.h" +#include "folly/executors/CPUThreadPoolExecutor.h" +#include "velox/common/memory/Memory.h" +#include "velox/dwio/common/ExecutorBarrier.h" + +namespace facebook::nimble::test { +using MemoryPool = velox::memory::MemoryPool; +using ExecutorBarrier = velox::dwio::common::ExecutorBarrier; + +class BufferPoolTest : public ::testing::Test { + protected: + static void SetUpTestCase() {} + + void SetUp() override { + memPool_ = velox::memory::deprecatedAddDefaultLeafMemoryPool(); + } + + std::shared_ptr memPool_; +}; + +TEST_F(BufferPoolTest, CreateBufferPoolBadMaxPool) { + try { + auto bufferPool = BufferPool{*memPool_, /* maxPoolSize */ 0}; + FAIL(); + } catch (const NimbleUserError& e) { + EXPECT_EQ(e.errorMessage(), "max pool size must be > 0"); + } +} + +TEST_F(BufferPoolTest, ReserveAddBuffer) { + auto bufferPool = BufferPool{*memPool_, /* maxPoolSize */ 10}; + auto buffer = bufferPool.reserveBuffer(); + EXPECT_EQ(bufferPool.size(), 9); + bufferPool.addBuffer(std::move(buffer)); + EXPECT_EQ(bufferPool.size(), 10); +} + +TEST_F(BufferPoolTest, EmptyFillBufferPool) { + size_t iterations = 10; + std::vector> buffers; + auto bufferPool = BufferPool{*memPool_, /* maxPoolSize */ iterations}; + + for (auto i = 0; i < iterations; ++i) { + auto buffer = bufferPool.reserveBuffer(); + buffers.push_back(std::move(buffer)); + EXPECT_EQ(bufferPool.size(), iterations - i - 1); + } + EXPECT_EQ(bufferPool.size(), 0); + + for (auto i = 0; i < iterations; ++i) { + bufferPool.addBuffer(std::move(buffers.back())); + buffers.pop_back(); + EXPECT_EQ(bufferPool.size(), i + 1); + } +} + +TEST_F(BufferPoolTest, ParallelFillPool) { + auto parallelismFactor = std::thread::hardware_concurrency(); + auto executor = + std::make_shared(parallelismFactor); + ExecutorBarrier barrier{executor}; + auto bufferPool = BufferPool{*memPool_}; + EXPECT_EQ(bufferPool.size(), parallelismFactor); + + for (auto i = 0; i < parallelismFactor; ++i) { + barrier.add([&]() { + for (auto j = 0; j < 100000; ++j) { + auto buffer = bufferPool.reserveBuffer(); + bufferPool.addBuffer(std::move(buffer)); + } + }); + } + + barrier.waitAll(); + EXPECT_LE(bufferPool.size(), parallelismFactor); +} +} // namespace facebook::nimble::test diff --git a/dwio/nimble/velox/FieldWriter.cpp b/dwio/nimble/velox/FieldWriter.cpp index 754a7f6..ec2c0f4 100644 --- a/dwio/nimble/velox/FieldWriter.cpp +++ b/dwio/nimble/velox/FieldWriter.cpp @@ -289,9 +289,10 @@ class SimpleFieldWriter : public FieldWriter { const OrderedRanges& ranges, folly::Executor*) override { auto size = ranges.size(); - auto& buffer = context_.stringBuffer(); auto& data = valuesStream_.mutableData(); - + auto bufferPtr = context_.bufferPool().reserveBuffer(); + auto bufferGuard = folly::makeGuard( + [&]() { context_.bufferPool().addBuffer(std::move(bufferPtr)); }); if (auto flat = vector->asFlatVector()) { valuesStream_.ensureNullsCapacity(flat->mayHaveNulls(), size); bool rangeCopied = false; @@ -331,6 +332,7 @@ class SimpleFieldWriter : public FieldWriter { valuesStream_.mutableNonNulls(), Flat{vector}, [&](SourceType value) { + auto& buffer = *bufferPtr; data.push_back( C::convert(value, buffer, valuesStream_.extraMemory())); }); @@ -344,6 +346,7 @@ class SimpleFieldWriter : public FieldWriter { valuesStream_.mutableNonNulls(), Decoded{decoded}, [&](SourceType value) { + auto& buffer = *bufferPtr; data.push_back( C::convert(value, buffer, valuesStream_.extraMemory())); }); @@ -1584,6 +1587,53 @@ size_t DecodingContextPool::size() const { return pool_.size(); } +BufferPool::BufferPool( + facebook::velox::memory::MemoryPool& memoryPool, + size_t maxPoolSize, + uint64_t initialChunkSize) + : defaultInitialChunkSize_{initialChunkSize}, + maxPoolSize{maxPoolSize}, + semaphore_{0}, + memoryPool_{memoryPool} { + NIMBLE_CHECK(maxPoolSize > 0, "max pool size must be > 0") + pool_.reserve(maxPoolSize); + for (size_t i = 0; i < maxPoolSize; ++i) { + pool_.emplace_back(newBuffer()); + semaphore_.release(); + } +} + +facebook::velox::memory::MemoryPool& BufferPool::getMemoryPool() { + return memoryPool_; +} + +// buffer back to the pool. +void BufferPool::addBuffer(std::unique_ptr buffer) { + std::scoped_lock lock(mutex_); + pool_.push_back(std::move(buffer)); + semaphore_.release(); +} + +// Reserves a buffer from the pool. Adds a new buffer to the pool +// while there are buffers available +std::unique_ptr BufferPool::reserveBuffer() { + semaphore_.acquire(); + + std::scoped_lock lock(mutex_); + auto buffer = std::move(pool_.back()); + pool_.pop_back(); + return buffer; +} + +// Returns estimated number of buffers in the pool +size_t BufferPool::size() { + return pool_.size(); +} + +std::unique_ptr BufferPool::newBuffer() { + return std::make_unique(memoryPool_, defaultInitialChunkSize_); +} + std::unique_ptr FieldWriter::create( FieldWriterContext& context, const std::shared_ptr& type, diff --git a/dwio/nimble/velox/FieldWriter.h b/dwio/nimble/velox/FieldWriter.h index 98212c4..b3d61a0 100644 --- a/dwio/nimble/velox/FieldWriter.h +++ b/dwio/nimble/velox/FieldWriter.h @@ -76,20 +76,48 @@ class DecodingContextPool { std::unique_ptr selectivityVector); }; +// Manages a pool of buffers. Buffers are returned to the pool when released. +// maxPoolSize should be set to at least 90% of capacity for performance +class BufferPool { + public: + explicit BufferPool( + facebook::velox::memory::MemoryPool& memoryPool, + size_t maxPoolSize = std::thread::hardware_concurrency(), + uint64_t initialChunkSize = kMinChunkSize); + + facebook::velox::memory::MemoryPool& getMemoryPool(); + void addBuffer(std::unique_ptr buffer); + std::unique_ptr reserveBuffer(); + size_t size(); + + private: + static const uint64_t kMinChunkSize = 1LL << 20; + const uint64_t defaultInitialChunkSize_; + const size_t maxPoolSize; + + std::mutex mutex_; + std::counting_semaphore<> semaphore_; + std::vector> pool_; + facebook::velox::memory::MemoryPool& memoryPool_; + + std::unique_ptr newBuffer(); +}; + struct FieldWriterContext { explicit FieldWriterContext( velox::memory::MemoryPool& memoryPool, std::unique_ptr reclaimer = nullptr, - std::function vectorDecoderVisitor = []() {}) + std::function vectorDecoderVisitor = []() {}, + size_t maxPoolSize = std::thread::hardware_concurrency()) : bufferMemoryPool{memoryPool.addLeafChild( "field_writer_buffer", true, std::move(reclaimer))}, inputBufferGrowthPolicy{ DefaultInputBufferGrowthPolicy::withDefaultRanges()}, - decodingContextPool_{std::move(vectorDecoderVisitor)} { - resetStringBuffer(); - } + bufferPool_{ + std::make_unique(*bufferMemoryPool, maxPoolSize)}, + decodingContextPool_{std::move(vectorDecoderVisitor)} {} std::shared_ptr bufferMemoryPool; std::mutex flatMapSchemaMutex; @@ -112,13 +140,8 @@ struct FieldWriterContext { return decodingContextPool_.reserveContext(); } - Buffer& stringBuffer() { - return *buffer_; - } - - // Reset writer context for use by next stripe. - void resetStringBuffer() { - buffer_ = std::make_unique(*bufferMemoryPool); + BufferPool& bufferPool() { + return *bufferPool_; } const std::vector>& streams() { @@ -148,7 +171,7 @@ struct FieldWriterContext { } private: - std::unique_ptr buffer_; + std::unique_ptr bufferPool_; DecodingContextPool decodingContextPool_; std::vector> streams_; }; diff --git a/dwio/nimble/velox/VeloxWriter.cpp b/dwio/nimble/velox/VeloxWriter.cpp index 89d5a6a..aed9f9a 100644 --- a/dwio/nimble/velox/VeloxWriter.cpp +++ b/dwio/nimble/velox/VeloxWriter.cpp @@ -64,7 +64,12 @@ class WriterContext : public FieldWriterContext { WriterContext( velox::memory::MemoryPool& memoryPool, VeloxWriterOptions options) - : FieldWriterContext{memoryPool, options.reclaimerFactory(), options.vectorDecoderVisitor}, + : FieldWriterContext{ + memoryPool, + options.reclaimerFactory(), + options.vectorDecoderVisitor, + options.maxPoolSize + }, options{std::move(options)}, logger{this->options.metricsLogger} { flushPolicy = this->options.flushPolicyFactory(); @@ -622,9 +627,6 @@ void VeloxWriter::writeChunk(bool lastChunk) { LoggingScope scope{*context_->logger}; velox::CpuWallTimer veloxTimer{context_->stripeFlushTiming}; - if (!encodingBuffer_) { - encodingBuffer_ = std::make_unique(*encodingMemoryPool_); - } streams_.resize(context_->schemaBuilder.nodeCount()); // When writing null streams, we write the nulls as data, and the stream @@ -668,9 +670,13 @@ void VeloxWriter::writeChunk(bool lastChunk) { auto encode = [&](StreamData& streamData) { const auto offset = streamData.descriptor().offset(); - auto encoded = encodeStream(*context_, *encodingBuffer_, streamData); + auto bufferPtr = context_->bufferPool().reserveBuffer(); + auto bufferGuard = folly::makeGuard( + [&]() { context_->bufferPool().addBuffer(std::move(bufferPtr)); }); + auto& buffer = *bufferPtr; + auto encoded = encodeStream(*context_, buffer, streamData); if (!encoded.empty()) { - ChunkedStreamWriter chunkWriter{*encodingBuffer_}; + ChunkedStreamWriter chunkWriter{buffer}; NIMBLE_DASSERT(offset < streams_.size(), "Stream offset out of range."); auto& stream = streams_[offset]; for (auto& buffer : chunkWriter.encode(encoded)) { @@ -782,10 +788,6 @@ uint32_t VeloxWriter::writeStripe() { uint64_t startSize = writer_.size(); writer_.writeStripe(context_->rowsInStripe, std::move(streams_)); stripeSize = writer_.size() - startSize; - encodingBuffer_.reset(); - // TODO: once chunked string fields are supported, move string buffer - // reset to writeChunk() - context_->resetStringBuffer(); } NIMBLE_ASSERT( diff --git a/dwio/nimble/velox/VeloxWriter.h b/dwio/nimble/velox/VeloxWriter.h index adb1e5b..d67b2ca 100644 --- a/dwio/nimble/velox/VeloxWriter.h +++ b/dwio/nimble/velox/VeloxWriter.h @@ -77,7 +77,6 @@ class VeloxWriter { TabletWriter writer_; std::unique_ptr root_; - std::unique_ptr encodingBuffer_; std::vector streams_; std::exception_ptr lastException_; const velox::common::SpillConfig* const spillConfig_; diff --git a/dwio/nimble/velox/VeloxWriterOptions.h b/dwio/nimble/velox/VeloxWriterOptions.h index 62c6dce..8d47ab0 100644 --- a/dwio/nimble/velox/VeloxWriterOptions.h +++ b/dwio/nimble/velox/VeloxWriterOptions.h @@ -126,6 +126,7 @@ struct VeloxWriterOptions { const velox::common::SpillConfig* spillConfig{nullptr}; + size_t maxPoolSize = std::thread::hardware_concurrency(); // If provided, internal encoding operations will happen in parallel using // this executor. std::shared_ptr encodingExecutor; diff --git a/dwio/nimble/velox/tests/VeloxWriterTests.cpp b/dwio/nimble/velox/tests/VeloxWriterTests.cpp index c8b41f0..30168da 100644 --- a/dwio/nimble/velox/tests/VeloxWriterTests.cpp +++ b/dwio/nimble/velox/tests/VeloxWriterTests.cpp @@ -319,11 +319,15 @@ TEST_F(VeloxWriterTests, MemoryReclaimPath) { std::string file; auto writeFile = std::make_unique(&file); std::atomic_bool reclaimEntered = false; - nimble::VeloxWriterOptions writerOptions{.reclaimerFactory = [&]() { - auto reclaimer = std::make_unique(); - reclaimer->setEnterArbitrationFunc([&]() { reclaimEntered = true; }); - return reclaimer; - }}; + nimble::VeloxWriterOptions writerOptions{ + .reclaimerFactory = + [&]() { + auto reclaimer = std::make_unique(); + reclaimer->setEnterArbitrationFunc( + [&]() { reclaimEntered = true; }); + return reclaimer; + }, + .maxPoolSize = 2}; nimble::VeloxWriter writer( *writerPool, type, std::move(writeFile), std::move(writerOptions)); auto batches = generateBatches(type, 100, 4000, 20221110, *leafPool_);