Bumper.jl
Bring Your Own Stack
Install / Use
/learn @MasonProtter/Bumper.jlREADME
- Basics
- Important notes
- Concurrency and parallelism
- Allocators provided by Bumper
- Creating your own allocator types
- Usage with StaticCompiler.jl
- Docstrings
Bumper.jl
Bumper.jl is a package that aims to make working with bump allocators (also known as arena allocators) easier and safer. You can dynamically allocate memory to these bump allocators, and reset them at the end of a code block, just like Julia's stack. Allocating to a bump allocator with Bumper.jl can be just as efficient as stack allocation. Bumper.jl is still a young package, and may have bugs. Let me know if you find any.
If you use Bumper.jl, please consider submitting a sample of your use-case so I can include it in the test suite.
Basics
Bumper.jl has a task-local default allocator, using a slab allocation strategy which can dynamically grow to arbitary sizes.
The simplest way to use Bumper is to rely on its default buffer implicitly like so:
using Bumper
function f(x)
# Set up a scope where memory may be allocated, and does not escape:
@no_escape begin
# Allocate a `UnsafeVector{eltype(x)}` (see UnsafeArrays.jl) using memory from the default buffer.
y = @alloc(eltype(x), length(x))
# Now do some stuff with that vector:
y .= x .+ 1
sum(y) # It's okay for the sum of y to escape the block, but references to y itself must not do so!
end
end
f([1,2,3])
9
When you use @no_escape, you are promising that the code enclosed in the macro will not leak any memory
created by @alloc. That is, you are only allowed to do intermediate @alloc allocations inside a @no_escape block,
and the lifetime of those allocations is the block. This is important. Once a @no_escape block finishes running, it
will reset its internal state to the position it had before the block started, potentially overwriting or freeing any
arrays which were created in the block.
In addition to @alloc for creating arrays, you can use @alloc_ptr(n) to get an n-byte pointer (of type
Ptr{Nothing}) directly.
Let's compare the performance of f to the equivalent with an intermediate heap allocation:
using BenchmarkTools
@benchmark f(x) setup=(x = rand(1:10, 30))
BenchmarkTools.Trial: 10000 samples with 995 evaluations.
Range (min … max): 28.465 ns … 49.843 ns ┊ GC (min … max): 0.00% … 0.00%
Time (median): 28.718 ns ┊ GC (median): 0.00%
Time (mean ± σ): 28.840 ns ± 0.833 ns ┊ GC (mean ± σ): 0.00% ± 0.00%
▃▄▂▇█▅▆▇▅▂▂▁▁▂▁ ▂
██████████████████▆▇▅▄▅▅▅▆▃▄▄▁▃▄▄▃▄▃▁▁▁▁▁▃▁▁▁▄▅▅▅▅▄▄▃▄▁▃▃▃▄ █
28.5 ns Histogram: log(frequency) by time 31.5 ns <
Memory estimate: 0 bytes, allocs estimate: 0.
and
function g(x::Vector{Int})
y = x .+ 1
sum(y)
end
@benchmark g(x) setup=(x = rand(1:10, 30))
BenchmarkTools.Trial: 10000 samples with 993 evaluations.
Range (min … max): 32.408 ns … 64.986 μs ┊ GC (min … max): 0.00% … 99.87%
Time (median): 37.443 ns ┊ GC (median): 0.00%
Time (mean ± σ): 55.929 ns ± 651.009 ns ┊ GC (mean ± σ): 14.68% ± 5.87%
▆█▅▃▁▁▁▁ ▁▁ ▁ ▂▁ ▁
████████▇██▅▄▃▄▁▁▃▁▁▁▁▁▁▁▁▃▃▁▁██████▇▇▅▁▄▃▃▃▁▁▃▁▁▁▄▃▄▅▄▄▅▇██ █
32.4 ns Histogram: log(frequency) by time 227 ns <
Memory estimate: 304 bytes, allocs estimate: 1.
So, using Bumper.jl in this benchmark gives a slight speedup relative to regular julia Vectors,
and a major increase in performance consistency due to the lack of heap allocations.
However, we can actually go a little faster if we're okay with manually passing around a buffer.
The way I invoked @no_escape and @alloc implicitly used the task's default buffer, and fetching that
default buffer is not as fast as using a const global variable, because Bumper.jl is trying to protect
you against concurrency bugs (more on that later).
If we provide the allocator to f explicitly, we go even faster:
function f(x, buf)
@no_escape buf begin # <----- Notice I specified buf here
y = @alloc(Int, length(x))
y .= x .+ 1
sum(y)
end
end
@benchmark f(x, buf) setup = begin
x = rand(1:10, 30)
buf = default_buffer()
end
BenchmarkTools.Trial: 10000 samples with 997 evaluations.
Range (min … max): 19.425 ns … 40.367 ns ┊ GC (min … max): 0.00% … 0.00%
Time (median): 19.494 ns ┊ GC (median): 0.00%
Time (mean ± σ): 19.620 ns ± 0.983 ns ┊ GC (mean ± σ): 0.00% ± 0.00%
█▅ ▁
██▅█▇▄▃▄▄▃▃▃▄▅▄▅▄▅▄▇▇▅▄▄▅▆▅▅▅▄▄▄▁▄▃▃▃▁▁▄▃▃▄▁▁▁▁▃▃▃▁▄▄▃▁▄▃▁▃ █
19.4 ns Histogram: log(frequency) by time 25.3 ns <
Memory estimate: 0 bytes, allocs estimate: 0.
If you manually specify a buffer like this, it is your responsibility to ensure that you don't have multiple concurrent tasks using that buffer at the same time.
Running default_buffer() will give you the current task's default buffer. You can explicitly construct
your own N byte buffer by calling AllocBuffer(N), or you can create a buffer which can dynamically
grow by calling SlabBuffer(). AllocBuffers are slightly faster than SlabBuffers, but will throw
an error if you overfill them. You can also create a ResizeBuffer(), which starts with a fixed buffer
but adaptively resizes itself upon a full reset to reduce future overflow — making it well-suited for
repeated operations where the memory footprint is not known in advance.
Important notes
@no_escapeblocks can be nested as much as you want, just don't let references outlive the specific block they were created in.- At the end of a
@no_escapeblock, all memory allocations from inside that block are erased and the buffer is reset to its previous state - The
@allocmarker can only be used directly inside of a@no_escapeblock, and it will always use the buffer that the corresponding@no_escapeblock uses. - You cannot use
@allocfrom a different concurrent task than its parent@no_escapeblock as this can cause concurrency bugs. - If for some reason you need to be able to use
@allocoutside of the scope of the@no_escapeblock, there is a functionBumper.alloc!(buf, T, n...)which takes in an explicit bufferbufand uses it to allocate an array of element typeT, and dimensionsn.... Using this is not as safe as@allocand not recommended. - Bumper.jl only supports
isbitstypes. You cannot use it for allocating vectors containing mutable, abstract, or other pointer-backed objects. - As mentioned previously, Do not allow any array which was initialized inside a
@no_escapeblock to escape the block. Doing so will cause incorrect results. - If you accidentally overfill a buffer, via e.g. a memory leak and need to reset the buffer, use
Bumper.reset_buffer!to do this. - You are not allowed to use
returnor@gotoinside a@no_escapeblock, since this could compromise the cleanup it performs after the block finishes.
Concurrency and parallelism
<details><summary>Click me!</summary> <p>Every task has its own independent default buffer. A task's buffer is only created if it is used, so this does not slow down the spawning of Julia tasks in general. Here's a demo showing that the default buffers are different:
using Bumper
let b = default_buffer() # The default buffer on the main task
t = @async default_buffer() # Get the default buffer on an asychronous task
fetch(t) === b
end
false
Whereas if we don't spawn any tasks, there is no unnecessary buffer creation:
let b = default_buffer()
b2 = default_buffer()
b2 === b
end
true
Because of this, we don't have to worry about @no_escape begin ... @alloc() ... end blocks on
different threads or tasks interfering with each other, so long as they are only operating on
buffers local to that task or the default_buffer().
Allocators provided by Bumper
<details><summary>Click me!</summary> <p>SlabBuffer
SlabBuffer is a slab-based bump allocator which can dynamically grow to hold an arbitrary amount of memory.
Small allocations from a SlabBuffer will live within a specific slab of memory, and if that slab fills up,
a new slab is allocated and future allocations will then happen on that slab. Small allocations are stored
in slabs of size SlabSize bytes (default 1 megabyte), and the list of live slabs are tracked in a field called
slabs. Allocations which are too large to fit into one slab are stored and tracked in a field called
custom_slabs.
SlabBuffers are nearly as fast as stack allocation (typically up to within a couple of nanoseconds) for typical
use. One potential performance pitfall is if that SlabBuffer's current position is at the end of a slab, then
the next allocation will be slow because it requires a new slab to be created. This means that if you do something
like
buf = SlabBuffer{N}()
@no_escape buf begin
@alloc(Int8, N÷2 - 1) # Take up just under half the first slab
@alloc(Int8, N÷2 - 1) # Take up another half of the first slab
# Now buf should be practically out of room.
for i in 1:1000
@no_escape buf begin
y = @alloc(Int8, 10) # This will allocate a new slab because there's no room
f(y)
end # At the end of this block, we delete the new slab because it's not needed.
end
end
then the inner loop will run slower than normal because at each iteration, a new slab of size N bytes must be freshly
allocated. This should be a rare occurance, but is possible to encounter.
Do not manipulat
