Allow manual queue item memory filling.

This commit is contained in:
Bartosz Taudul 2017-10-01 17:14:26 +02:00
parent 72abdcaf8d
commit 7b0cbef0d7

View File

@ -942,7 +942,17 @@ public:
{ {
return inner_enqueue<CanAlloc>(token, std::move(item)); return inner_enqueue<CanAlloc>(token, std::move(item));
} }
inline T* enqueue_begin(producer_token_t const& token)
{
return inner_enqueue_begin<CanAlloc>(token);
}
inline void enqueue_finish(producer_token_t const& token)
{
inner_enqueue_finish(token);
}
// Enqueues several items. // Enqueues several items.
// Allocates memory if required. Only fails if memory allocation fails (or // Allocates memory if required. Only fails if memory allocation fails (or
// implicit production is disabled because Traits::INITIAL_IMPLICIT_PRODUCER_HASH_SIZE // implicit production is disabled because Traits::INITIAL_IMPLICIT_PRODUCER_HASH_SIZE
@ -1281,7 +1291,18 @@ private:
{ {
return static_cast<ExplicitProducer*>(token.producer)->ConcurrentQueue::ExplicitProducer::template enqueue<canAlloc>(std::forward<U>(element)); return static_cast<ExplicitProducer*>(token.producer)->ConcurrentQueue::ExplicitProducer::template enqueue<canAlloc>(std::forward<U>(element));
} }
template<AllocationMode canAlloc>
inline T* inner_enqueue_begin(producer_token_t const& token)
{
return static_cast<ExplicitProducer*>(token.producer)->ConcurrentQueue::ExplicitProducer::template enqueue_begin<canAlloc>();
}
inline void inner_enqueue_finish(producer_token_t const& token)
{
return static_cast<ExplicitProducer*>(token.producer)->ConcurrentQueue::ExplicitProducer::template enqueue_finish();
}
template<AllocationMode canAlloc, typename U> template<AllocationMode canAlloc, typename U>
inline bool inner_enqueue(U&& element) inline bool inner_enqueue(U&& element)
{ {
@ -1878,7 +1899,91 @@ private:
this->tailIndex.store(newTailIndex, std::memory_order_release); this->tailIndex.store(newTailIndex, std::memory_order_release);
return true; return true;
} }
template<AllocationMode allocMode>
inline T* enqueue_begin()
{
index_t currentTailIndex = this->tailIndex.load(std::memory_order_relaxed);
pr_newTailIndex = 1 + currentTailIndex;
if ((currentTailIndex & static_cast<index_t>(BLOCK_SIZE - 1)) == 0) {
// We reached the end of a block, start a new one
auto startBlock = this->tailBlock;
auto originalBlockIndexSlotsUsed = pr_blockIndexSlotsUsed;
if (this->tailBlock != nullptr && this->tailBlock->next->ConcurrentQueue::Block::template is_empty<explicit_context>()) {
// We can re-use the block ahead of us, it's empty!
this->tailBlock = this->tailBlock->next;
this->tailBlock->ConcurrentQueue::Block::template reset_empty<explicit_context>();
// We'll put the block on the block index (guaranteed to be room since we're conceptually removing the
// last block from it first -- except instead of removing then adding, we can just overwrite).
// Note that there must be a valid block index here, since even if allocation failed in the ctor,
// it would have been re-attempted when adding the first block to the queue; since there is such
// a block, a block index must have been successfully allocated.
}
else {
// Whatever head value we see here is >= the last value we saw here (relatively),
// and <= its current value. Since we have the most recent tail, the head must be
// <= to it.
auto head = this->headIndex.load(std::memory_order_relaxed);
assert(!details::circular_less_than<index_t>(currentTailIndex, head));
if (!details::circular_less_than<index_t>(head, currentTailIndex + BLOCK_SIZE)
|| (MAX_SUBQUEUE_SIZE != details::const_numeric_max<size_t>::value && (MAX_SUBQUEUE_SIZE == 0 || MAX_SUBQUEUE_SIZE - BLOCK_SIZE < currentTailIndex - head))) {
// We can't enqueue in another block because there's not enough leeway -- the
// tail could surpass the head by the time the block fills up! (Or we'll exceed
// the size limit, if the second part of the condition was true.)
return nullptr;
}
// We're going to need a new block; check that the block index has room
if (pr_blockIndexRaw == nullptr || pr_blockIndexSlotsUsed == pr_blockIndexSize) {
// Hmm, the circular block index is already full -- we'll need
// to allocate a new index. Note pr_blockIndexRaw can only be nullptr if
// the initial allocation failed in the constructor.
if (allocMode == CannotAlloc || !new_block_index(pr_blockIndexSlotsUsed)) {
return nullptr;
}
}
// Insert a new block in the circular linked list
auto newBlock = this->parent->ConcurrentQueue::template requisition_block<allocMode>();
if (newBlock == nullptr) {
return nullptr;
}
#if MCDBGQ_TRACKMEM
newBlock->owner = this;
#endif
newBlock->ConcurrentQueue::Block::template reset_empty<explicit_context>();
if (this->tailBlock == nullptr) {
newBlock->next = newBlock;
}
else {
newBlock->next = this->tailBlock->next;
this->tailBlock->next = newBlock;
}
this->tailBlock = newBlock;
++pr_blockIndexSlotsUsed;
}
(void)startBlock;
(void)originalBlockIndexSlotsUsed;
// Add block to block index
auto& entry = blockIndex.load(std::memory_order_relaxed)->entries[pr_blockIndexFront];
entry.base = currentTailIndex;
entry.block = this->tailBlock;
blockIndex.load(std::memory_order_relaxed)->front.store(pr_blockIndexFront, std::memory_order_release);
pr_blockIndexFront = (pr_blockIndexFront + 1) & (pr_blockIndexSize - 1);
}
// Enqueue
return (*this->tailBlock)[currentTailIndex];
}
inline void enqueue_finish()
{
this->tailIndex.store(pr_newTailIndex, std::memory_order_release);
}
template<typename U> template<typename U>
bool dequeue(U& element) bool dequeue(U& element)
{ {
@ -2316,6 +2421,7 @@ private:
std::atomic<BlockIndexHeader*> blockIndex; std::atomic<BlockIndexHeader*> blockIndex;
// To be used by producer only -- consumer must use the ones in referenced by blockIndex // To be used by producer only -- consumer must use the ones in referenced by blockIndex
index_t pr_newTailIndex;
size_t pr_blockIndexSlotsUsed; size_t pr_blockIndexSlotsUsed;
size_t pr_blockIndexSize; size_t pr_blockIndexSize;
size_t pr_blockIndexFront; // Next slot (not current) size_t pr_blockIndexFront; // Next slot (not current)