phobos/std/experimental/allocator/building_blocks/kernighan_ritchie.d

939 lines
30 KiB
D

// Written in the D programming language.
/**
Source: $(PHOBOSSRC std/experimental/allocator/building_blocks/kernighan_ritchie.d)
*/
module std.experimental.allocator.building_blocks.kernighan_ritchie;
import std.experimental.allocator.building_blocks.null_allocator :
NullAllocator;
//debug = KRRegion;
debug(KRRegion) import std.stdio;
// KRRegion
/**
`KRRegion` draws inspiration from the $(MREF_ALTTEXT region allocation
strategy, std,experimental,allocator,building_blocks,region) and also the
$(HTTP stackoverflow.com/questions/13159564/explain-this-implementation-of-malloc-from-the-kr-book,
famed allocator) described by Brian Kernighan and Dennis Ritchie in section 8.7
of the book $(HTTP amazon.com/exec/obidos/ASIN/0131103628/classicempire, "The C
Programming Language"), Second Edition, Prentice Hall, 1988.
$(H4 `KRRegion` = `Region` + Kernighan-Ritchie Allocator)
Initially, `KRRegion` starts in "region" mode: allocations are served from
the memory chunk in a region fashion. Thus, as long as there is enough memory
left, `KRRegion.allocate` has the performance profile of a region allocator.
Deallocation inserts (in $(BIGOH 1) time) the deallocated blocks in an
unstructured freelist, which is not read in region mode.
Once the region cannot serve an `allocate` request, `KRRegion` switches
to "free list" mode. It sorts the list of previously deallocated blocks by
address and serves allocation requests off that free list. The allocation and
deallocation follow the pattern described by Kernighan and Ritchie.
The recommended use of `KRRegion` is as a $(I region with deallocation). If the
`KRRegion` is dimensioned appropriately, it could often not enter free list
mode during its lifetime. Thus it is as fast as a simple region, whilst
offering deallocation at a small cost. When the region memory is exhausted,
the previously deallocated memory is still usable, at a performance cost. If
the region is not excessively large and fragmented, the linear allocation and
deallocation cost may still be compensated for by the good locality
characteristics.
If the chunk of memory managed is large, it may be desirable to switch
management to free list from the beginning. That way, memory may be used in a
more compact manner than region mode. To force free list mode, call $(D
switchToFreeList) shortly after construction or when deemed appropriate.
The smallest size that can be allocated is two words (16 bytes on 64-bit
systems, 8 bytes on 32-bit systems). This is because the free list management
needs two words (one for the length, the other for the next pointer in the
singly-linked list).
The `ParentAllocator` type parameter is the type of the allocator used to
allocate the memory chunk underlying the `KRRegion` object. Choosing the
default (`NullAllocator`) means the user is responsible for passing a buffer
at construction (and for deallocating it if necessary). Otherwise, `KRRegion`
automatically deallocates the buffer during destruction. For that reason, if
`ParentAllocator` is not `NullAllocator`, then `KRRegion` is not
copyable.
$(H4 Implementation Details)
In free list mode, `KRRegion` embeds a free blocks list onto the chunk of
memory. The free list is circular, coalesced, and sorted by address at all
times. Allocations and deallocations take time proportional to the number of
previously deallocated blocks. (In practice the cost may be lower, e.g. if
memory is deallocated in reverse order of allocation, all operations take
constant time.) Memory utilization is good (small control structure and no
per-allocation overhead). The disadvantages of freelist mode include proneness
to fragmentation, a minimum allocation size of two words, and linear worst-case
allocation and deallocation times.
Similarities of `KRRegion` (in free list mode) with the
Kernighan-Ritchie allocator:
$(UL
$(LI Free blocks have variable size and are linked in a singly-linked list.)
$(LI The freelist is maintained in increasing address order, which makes
coalescing easy.)
$(LI The strategy for finding the next available block is first fit.)
$(LI The free list is circular, with the last node pointing back to the first.)
$(LI Coalescing is carried during deallocation.)
)
Differences from the Kernighan-Ritchie allocator:
$(UL
$(LI Once the chunk is exhausted, the Kernighan-Ritchie allocator allocates
another chunk using operating system primitives. For better composability, $(D
KRRegion) just gets full (returns `null` on new allocation requests). The
decision to allocate more blocks is deferred to a higher-level entity. For an
example, see the example below using `AllocatorList` in conjunction with $(D
KRRegion).)
$(LI Allocated blocks do not hold a size prefix. This is because in D the size
information is available in client code at deallocation time.)
)
*/
struct KRRegion(ParentAllocator = NullAllocator)
{
import std.experimental.allocator.common : stateSize, alignedAt;
import std.traits : hasMember;
import std.typecons : Ternary;
private static struct Node
{
import std.typecons : tuple, Tuple;
Node* next;
size_t size;
this(this) @disable;
void[] payload() inout
{
return (cast(ubyte*) &this)[0 .. size];
}
bool adjacent(in Node* right) const
{
assert(right);
auto p = payload;
return p.ptr < right && right < p.ptr + p.length + Node.sizeof;
}
bool coalesce(void* memoryEnd = null)
{
// Coalesce the last node before the memory end with any possible gap
if (memoryEnd
&& memoryEnd < payload.ptr + payload.length + Node.sizeof)
{
size += memoryEnd - (payload.ptr + payload.length);
return true;
}
if (!adjacent(next)) return false;
size = (cast(ubyte*) next + next.size) - cast(ubyte*) &this;
next = next.next;
return true;
}
Tuple!(void[], Node*) allocateHere(size_t bytes)
{
assert(bytes >= Node.sizeof);
assert(bytes % Node.alignof == 0);
assert(next);
assert(!adjacent(next));
if (size < bytes) return typeof(return)();
assert(size >= bytes);
immutable leftover = size - bytes;
if (leftover >= Node.sizeof)
{
// There's room for another node
auto newNode = cast(Node*) ((cast(ubyte*) &this) + bytes);
newNode.size = leftover;
newNode.next = next == &this ? newNode : next;
assert(next);
return tuple(payload, newNode);
}
// No slack space, just return next node
return tuple(payload, next == &this ? null : next);
}
}
// state
/**
If `ParentAllocator` holds state, `parent` is a public member of type
`KRRegion`. Otherwise, `parent` is an `alias` for
`ParentAllocator.instance`.
*/
static if (stateSize!ParentAllocator) ParentAllocator parent;
else alias parent = ParentAllocator.instance;
private void[] payload;
private Node* root;
private bool regionMode() const { return bytesUsedRegionMode != size_t.max; }
private void cancelRegionMode() { bytesUsedRegionMode = size_t.max; }
private size_t bytesUsedRegionMode = 0;
auto byNodePtr()
{
static struct Range
{
Node* start, current;
@property bool empty() { return !current; }
@property Node* front() { return current; }
void popFront()
{
assert(current && current.next);
current = current.next;
if (current == start) current = null;
}
@property Range save() { return this; }
}
import std.range : isForwardRange;
static assert(isForwardRange!Range);
return Range(root, root);
}
string toString()
{
import std.format : format;
string s = "KRRegion@";
s ~= format("%s-%s(0x%s[%s] %s", &this, &this + 1,
payload.ptr, payload.length,
regionMode ? "(region)" : "(freelist)");
Node* lastNode = null;
if (!regionMode)
{
foreach (node; byNodePtr)
{
s ~= format(", %sfree(0x%s[%s])",
lastNode && lastNode.adjacent(node) ? "+" : "",
cast(void*) node, node.size);
lastNode = node;
}
}
else
{
for (auto node = root; node; node = node.next)
{
s ~= format(", %sfree(0x%s[%s])",
lastNode && lastNode.adjacent(node) ? "+" : "",
cast(void*) node, node.size);
lastNode = node;
}
}
s ~= ')';
return s;
}
private void assertValid(string s)
{
assert(!regionMode);
if (!payload.ptr)
{
assert(!root, s);
return;
}
if (!root)
{
return;
}
assert(root >= payload.ptr, s);
assert(root < payload.ptr + payload.length, s);
// Check that the list terminates
size_t n;
foreach (node; byNodePtr)
{
assert(node.next);
assert(!node.adjacent(node.next));
assert(n++ < payload.length / Node.sizeof, s);
}
}
private Node* sortFreelist(Node* root)
{
// Find a monotonic run
auto last = root;
for (;;)
{
if (!last.next) return root;
if (last > last.next) break;
assert(last < last.next);
last = last.next;
}
auto tail = last.next;
last.next = null;
tail = sortFreelist(tail);
return merge(root, tail);
}
private Node* merge(Node* left, Node* right)
{
assert(left != right);
if (!left) return right;
if (!right) return left;
if (left < right)
{
auto result = left;
result.next = merge(left.next, right);
return result;
}
auto result = right;
result.next = merge(left, right.next);
return result;
}
private void coalesceAndMakeCircular()
{
for (auto n = root;;)
{
assert(!n.next || n < n.next);
if (!n.next)
{
// Convert to circular
n.next = root;
break;
}
if (n.coalesce) continue; // possibly another coalesce
n = n.next;
}
}
/**
Create a `KRRegion`. If `ParentAllocator` is not `NullAllocator`,
`KRRegion`'s destructor will call `parent.deallocate`.
Params:
b = Block of memory to serve as support for the allocator. Memory must be
larger than two words and word-aligned.
n = Capacity desired. This constructor is defined only if $(D
ParentAllocator) is not `NullAllocator`.
*/
this(ubyte[] b)
{
if (b.length < Node.sizeof)
{
// Init as empty
assert(root is null);
assert(payload is null);
return;
}
assert(b.length >= Node.sizeof);
assert(b.ptr.alignedAt(Node.alignof));
assert(b.length >= 2 * Node.sizeof);
payload = b;
root = cast(Node*) b.ptr;
// Initialize the free list with all list
assert(regionMode);
root.next = null;
root.size = b.length;
debug(KRRegion) writefln("KRRegion@%s: init with %s[%s]", &this,
b.ptr, b.length);
}
/// Ditto
static if (!is(ParentAllocator == NullAllocator) && !stateSize!ParentAllocator)
this(size_t n)
{
assert(n > Node.sizeof);
this(cast(ubyte[])(parent.allocate(n)));
}
/// Ditto
static if (!is(ParentAllocator == NullAllocator) && stateSize!ParentAllocator)
this(ParentAllocator parent, size_t n)
{
assert(n > Node.sizeof);
this.parent = parent;
this(cast(ubyte[])(parent.allocate(n)));
}
/// Ditto
static if (!is(ParentAllocator == NullAllocator)
&& hasMember!(ParentAllocator, "deallocate"))
~this()
{
parent.deallocate(payload);
}
/**
Forces free list mode. If already in free list mode, does nothing.
Otherwise, sorts the free list accumulated so far and switches strategy for
future allocations to KR style.
*/
void switchToFreeList()
{
if (!regionMode) return;
cancelRegionMode;
if (!root) return;
root = sortFreelist(root);
coalesceAndMakeCircular;
}
/*
Noncopyable
*/
@disable this(this);
/**
Word-level alignment.
*/
enum alignment = Node.alignof;
/**
Allocates `n` bytes. Allocation searches the list of available blocks
until a free block with `n` or more bytes is found (first fit strategy).
The block is split (if larger) and returned.
Params: n = number of bytes to _allocate
Returns: A word-aligned buffer of `n` bytes, or `null`.
*/
void[] allocate(size_t n)
{
if (!n || !root) return null;
const actualBytes = goodAllocSize(n);
// Try the region first
if (regionMode)
{
// Only look at the head of the freelist
if (root.size >= actualBytes)
{
// Enough room for allocation
bytesUsedRegionMode += actualBytes;
void* result = root;
immutable balance = root.size - actualBytes;
if (balance >= Node.sizeof)
{
auto newRoot = cast(Node*) (result + actualBytes);
newRoot.next = root.next;
newRoot.size = balance;
root = newRoot;
}
else
{
root = null;
switchToFreeList;
}
return result[0 .. n];
}
// Not enough memory, switch to freelist mode and fall through
switchToFreeList;
}
// Try to allocate from next after the iterating node
for (auto pnode = root;;)
{
assert(!pnode.adjacent(pnode.next));
auto k = pnode.next.allocateHere(actualBytes);
if (k[0] !is null)
{
// awes
assert(k[0].length >= n);
if (root == pnode.next) root = k[1];
pnode.next = k[1];
return k[0][0 .. n];
}
pnode = pnode.next;
if (pnode == root) break;
}
return null;
}
/**
Deallocates `b`, which is assumed to have been previously allocated with
this allocator. Deallocation performs a linear search in the free list to
preserve its sorting order. It follows that blocks with higher addresses in
allocators with many free blocks are slower to deallocate.
Params: b = block to be deallocated
*/
nothrow @nogc
bool deallocate(void[] b)
{
debug(KRRegion) writefln("KRRegion@%s: deallocate(%s[%s])", &this,
b.ptr, b.length);
if (!b.ptr) return true;
assert(owns(b) == Ternary.yes);
assert(b.ptr.alignedAt(Node.alignof));
// Insert back in the freelist, keeping it sorted by address. Do not
// coalesce at this time. Instead, do it lazily during allocation.
auto n = cast(Node*) b.ptr;
n.size = goodAllocSize(b.length);
auto memoryEnd = payload.ptr + payload.length;
if (regionMode)
{
assert(root);
// Insert right after root
bytesUsedRegionMode -= n.size;
n.next = root.next;
root.next = n;
return true;
}
if (!root)
{
// What a sight for sore eyes
root = n;
root.next = root;
// If the first block freed is the last one allocated,
// maybe there's a gap after it.
root.coalesce(memoryEnd);
return true;
}
version (assert) foreach (test; byNodePtr)
{
assert(test != n);
}
// Linear search
auto pnode = root;
do
{
assert(pnode && pnode.next);
assert(pnode != n);
assert(pnode.next != n);
if (pnode < pnode.next)
{
if (pnode > n || n > pnode.next) continue;
// Insert in between pnode and pnode.next
n.next = pnode.next;
pnode.next = n;
n.coalesce;
pnode.coalesce;
root = pnode;
return true;
}
else if (pnode < n)
{
// Insert at the end of the list
// Add any possible gap at the end of n to the length of n
n.next = pnode.next;
pnode.next = n;
n.coalesce(memoryEnd);
pnode.coalesce;
root = pnode;
return true;
}
else if (n < pnode.next)
{
// Insert at the front of the list
n.next = pnode.next;
pnode.next = n;
n.coalesce;
root = n;
return true;
}
}
while ((pnode = pnode.next) != root);
assert(0, "Wrong parameter passed to deallocate");
}
/**
Allocates all memory available to this allocator. If the allocator is empty,
returns the entire available block of memory. Otherwise, it still performs
a best-effort allocation: if there is no fragmentation (e.g. `allocate`
has been used but not `deallocate`), allocates and returns the only
available block of memory.
The operation takes time proportional to the number of adjacent free blocks
at the front of the free list. These blocks get coalesced, whether
`allocateAll` succeeds or fails due to fragmentation.
*/
void[] allocateAll()
{
if (regionMode) switchToFreeList;
if (root && root.next == root)
return allocate(root.size);
return null;
}
///
@system unittest
{
import std.experimental.allocator.gc_allocator : GCAllocator;
auto alloc = KRRegion!GCAllocator(1024 * 64);
const b1 = alloc.allocate(2048);
assert(b1.length == 2048);
const b2 = alloc.allocateAll;
assert(b2.length == 1024 * 62);
}
/**
Deallocates all memory currently allocated, making the allocator ready for
other allocations. This is a $(BIGOH 1) operation.
*/
pure nothrow @nogc
bool deallocateAll()
{
debug(KRRegion) assertValid("deallocateAll");
debug(KRRegion) scope(exit) assertValid("deallocateAll");
root = cast(Node*) payload.ptr;
// Reset to regionMode
bytesUsedRegionMode = 0;
if (root)
{
root.next = null;
root.size = payload.length;
}
return true;
}
/**
Checks whether the allocator is responsible for the allocation of `b`.
It does a simple $(BIGOH 1) range check. `b` should be a buffer either
allocated with `this` or obtained through other means.
*/
pure nothrow @trusted @nogc
Ternary owns(void[] b)
{
debug(KRRegion) assertValid("owns");
debug(KRRegion) scope(exit) assertValid("owns");
return Ternary(b && payload && (&b[0] >= &payload[0])
&& (&b[0] < &payload[0] + payload.length));
}
/**
Adjusts `n` to a size suitable for allocation (two words or larger,
word-aligned).
*/
pure nothrow @safe @nogc
static size_t goodAllocSize(size_t n)
{
import std.experimental.allocator.common : roundUpToMultipleOf;
return n <= Node.sizeof
? Node.sizeof : n.roundUpToMultipleOf(alignment);
}
/**
Returns: `Ternary.yes` if the allocator is empty, `Ternary.no` otherwise.
Never returns `Ternary.unknown`.
*/
pure nothrow @safe @nogc
Ternary empty()
{
if (regionMode)
return Ternary(bytesUsedRegionMode == 0);
return Ternary(root && root.size == payload.length);
}
}
/**
`KRRegion` is preferable to `Region` as a front for a general-purpose
allocator if `deallocate` is needed, yet the actual deallocation traffic is
relatively low. The example below shows a `KRRegion` using stack storage
fronting the GC allocator.
*/
@system unittest
{
import std.experimental.allocator.building_blocks.fallback_allocator
: fallbackAllocator;
import std.experimental.allocator.gc_allocator : GCAllocator;
import std.typecons : Ternary;
// KRRegion fronting a general-purpose allocator
ubyte[1024 * 128] buf;
auto alloc = fallbackAllocator(KRRegion!()(buf), GCAllocator.instance);
auto b = alloc.allocate(100);
assert(b.length == 100);
assert((() pure nothrow @safe @nogc => alloc.primary.owns(b))() == Ternary.yes);
}
/**
The code below defines a scalable allocator consisting of 1 MB (or larger)
blocks fetched from the garbage-collected heap. Each block is organized as a
KR-style heap. More blocks are allocated and freed on a need basis.
This is the closest example to the allocator introduced in the K$(AMP)R book.
It should perform slightly better because instead of searching through one
large free list, it searches through several shorter lists in LRU order. Also,
it actually returns memory to the operating system when possible.
*/
@system unittest
{
import std.algorithm.comparison : max;
import std.experimental.allocator.building_blocks.allocator_list
: AllocatorList;
import std.experimental.allocator.mmap_allocator : MmapAllocator;
AllocatorList!(n => KRRegion!MmapAllocator(max(n * 16, 1024 * 1024))) alloc;
}
@system unittest
{
import std.algorithm.comparison : max;
import std.experimental.allocator.building_blocks.allocator_list
: AllocatorList;
import std.experimental.allocator.mallocator : Mallocator;
import std.typecons : Ternary;
/*
Create a scalable allocator consisting of 1 MB (or larger) blocks fetched
from the garbage-collected heap. Each block is organized as a KR-style
heap. More blocks are allocated and freed on a need basis.
*/
AllocatorList!(n => KRRegion!Mallocator(max(n * 16, 1024 * 1024)),
NullAllocator) alloc;
void[][50] array;
foreach (i; 0 .. array.length)
{
auto length = i * 10_000 + 1;
array[i] = alloc.allocate(length);
assert(array[i].ptr);
assert(array[i].length == length);
}
import std.random : randomShuffle;
randomShuffle(array[]);
foreach (i; 0 .. array.length)
{
assert(array[i].ptr);
assert((() pure nothrow @safe @nogc => alloc.owns(array[i]))() == Ternary.yes);
() nothrow @nogc { alloc.deallocate(array[i]); }();
}
}
@system unittest
{
import std.algorithm.comparison : max;
import std.experimental.allocator.building_blocks.allocator_list
: AllocatorList;
import std.experimental.allocator.mmap_allocator : MmapAllocator;
import std.typecons : Ternary;
/*
Create a scalable allocator consisting of 1 MB (or larger) blocks fetched
from the garbage-collected heap. Each block is organized as a KR-style
heap. More blocks are allocated and freed on a need basis.
*/
AllocatorList!((n) {
auto result = KRRegion!MmapAllocator(max(n * 2, 1024 * 1024));
return result;
}) alloc;
void[][99] array;
foreach (i; 0 .. array.length)
{
auto length = i * 10_000 + 1;
array[i] = alloc.allocate(length);
assert(array[i].ptr);
foreach (j; 0 .. i)
{
assert(array[i].ptr != array[j].ptr);
}
assert(array[i].length == length);
}
import std.random : randomShuffle;
randomShuffle(array[]);
foreach (i; 0 .. array.length)
{
assert((() pure nothrow @safe @nogc => alloc.owns(array[i]))() == Ternary.yes);
() nothrow @nogc { alloc.deallocate(array[i]); }();
}
}
version (StdUnittest)
@system unittest
{
import std.algorithm.comparison : max;
import std.experimental.allocator.building_blocks.allocator_list
: AllocatorList;
import std.experimental.allocator.common : testAllocator;
import std.experimental.allocator.gc_allocator : GCAllocator;
testAllocator!(() => AllocatorList!(
n => KRRegion!GCAllocator(max(n * 16, 1024 * 1024)))());
}
@system unittest
{
import std.experimental.allocator.gc_allocator : GCAllocator;
auto alloc = KRRegion!GCAllocator(1024 * 1024);
void[][] array;
foreach (i; 1 .. 4)
{
array ~= alloc.allocate(i);
assert(array[$ - 1].length == i);
}
() nothrow @nogc { alloc.deallocate(array[1]); }();
() nothrow @nogc { alloc.deallocate(array[0]); }();
() nothrow @nogc { alloc.deallocate(array[2]); }();
assert(alloc.allocateAll().length == 1024 * 1024);
}
@system unittest
{
import std.experimental.allocator.gc_allocator : GCAllocator;
import std.typecons : Ternary;
auto alloc = KRRegion!()(
cast(ubyte[])(GCAllocator.instance.allocate(1024 * 1024)));
const store = alloc.allocate(KRRegion!().sizeof);
auto p = cast(KRRegion!()* ) store.ptr;
import core.lifetime : emplace;
import core.stdc.string : memcpy;
import std.conv : text;
memcpy(p, &alloc, alloc.sizeof);
emplace(&alloc);
void[][100] array;
foreach (i; 0 .. array.length)
{
auto length = 100 * i + 1;
array[i] = p.allocate(length);
assert(array[i].length == length, text(array[i].length));
assert((() pure nothrow @safe @nogc => p.owns(array[i]))() == Ternary.yes);
}
import std.random : randomShuffle;
randomShuffle(array[]);
foreach (i; 0 .. array.length)
{
assert((() pure nothrow @safe @nogc => p.owns(array[i]))() == Ternary.yes);
() nothrow @nogc { p.deallocate(array[i]); }();
}
auto b = p.allocateAll();
assert(b.length == 1024 * 1024 - KRRegion!().sizeof, text(b.length));
}
@system unittest
{
import std.typecons : Ternary;
import std.experimental.allocator.gc_allocator : GCAllocator;
auto alloc = KRRegion!()(
cast(ubyte[])(GCAllocator.instance.allocate(1024 * 1024)));
auto p = alloc.allocateAll();
assert(p.length == 1024 * 1024);
assert((() nothrow @nogc => alloc.deallocateAll())());
assert(alloc.empty() == Ternary.yes);
p = alloc.allocateAll();
assert(p.length == 1024 * 1024);
}
@system unittest
{
import std.random : randomCover;
import std.typecons : Ternary;
// Both sequences must work on either system
// A sequence of allocs which generates the error described in https://issues.dlang.org/show_bug.cgi?id=16564
// that is a gap at the end of buf from the perspective of the allocator
// for 64 bit systems (leftover balance = 8 bytes < 16)
int[] sizes64 = [18904, 2008, 74904, 224, 111904, 1904, 52288, 8];
// for 32 bit systems (leftover balance < 8)
int[] sizes32 = [81412, 107068, 49892, 23768];
void test(int[] sizes)
{
align(size_t.sizeof) ubyte[256 * 1024] buf;
auto a = KRRegion!()(buf);
void[][] bufs;
foreach (size; sizes)
{
bufs ~= a.allocate(size);
}
foreach (b; bufs.randomCover)
{
() nothrow @nogc { a.deallocate(b); }();
}
assert((() pure nothrow @safe @nogc => a.empty)() == Ternary.yes);
}
test(sizes64);
test(sizes32);
}
@system unittest
{
import std.typecons : Ternary;
// For 64 bits, we allocate in multiples of 8, but the minimum alloc size is 16.
// This can create gaps.
// This test is an example of such a case. The gap is formed between the block
// allocated for the second value in sizes and the third. There is also a gap
// at the very end. (total lost 2 * word)
int[] sizes64 = [2008, 18904, 74904, 224, 111904, 1904, 52288, 8];
int[] sizes32 = [81412, 107068, 49892, 23768];
int word64 = 8;
int word32 = 4;
void test(int[] sizes, int word)
{
align(size_t.sizeof) ubyte[256 * 1024] buf;
auto a = KRRegion!()(buf);
void[][] bufs;
foreach (size; sizes)
{
bufs ~= a.allocate(size);
}
() nothrow @nogc { a.deallocate(bufs[1]); }();
bufs ~= a.allocate(sizes[1] - word);
() nothrow @nogc { a.deallocate(bufs[0]); }();
foreach (i; 2 .. bufs.length)
{
() nothrow @nogc { a.deallocate(bufs[i]); }();
}
assert((() pure nothrow @safe @nogc => a.empty)() == Ternary.yes);
}
test(sizes64, word64);
test(sizes32, word32);
}
@system unittest
{
import std.experimental.allocator.gc_allocator : GCAllocator;
auto a = KRRegion!GCAllocator(1024 * 1024);
assert((() pure nothrow @safe @nogc => a.goodAllocSize(1))() == typeof(*a.root).sizeof);
}
@system unittest
{ import std.typecons : Ternary;
ubyte[1024] b;
auto alloc = KRRegion!()(b);
auto k = alloc.allocate(128);
assert(k.length == 128);
assert(alloc.empty == Ternary.no);
assert(alloc.deallocate(k));
assert(alloc.empty == Ternary.yes);
k = alloc.allocate(512);
assert(k.length == 512);
assert(alloc.empty == Ternary.no);
assert(alloc.deallocate(k));
assert(alloc.empty == Ternary.yes);
k = alloc.allocate(1024);
assert(k.length == 1024);
assert(alloc.empty == Ternary.no);
assert(alloc.deallocate(k));
assert(alloc.empty == Ternary.yes);
}