Mincdc
MinCDC is a very simple yet efficient content-defined chunking algorithm.
Install / Use
/learn @orlp/MincdcREADME
MinCDC
MinCDC is a very simple yet efficient content-defined chunking algorithm. It splits your input data into chunks in such a way that the boundaries are defined by the data itself. This means duplicate regions in large (sets of) files are likely to have identical boundaries and can thus efficiently be found and deduplicated.
To start using mincdc add the following to your Cargo.toml:
[dependencies]
mincdc = "0.1"
Please refer to the documentation for more information on usage.
Algorithm
The basic idea of MinCDC is to choose chunk boundaries based on the minimum
value of a sliding window over the input data. That is, if the desired chunk
size is between min_size and max_size, we find some min_size <= i <= max_size such that evaluate(bytes[i - w..i]) is minimized, where w is the
window size, breaking ties by choosing the earliest such i. Then we return
chunk bytes[..i] and repeat the process on the remainder bytes[i..].
This library provides two SIMD-accelerated implementations of MinCDC, both with a window size of 4:
MinCDC4, where the evaluation function isu32::from_le_bytes(bytes[i - 4..i]), i.e. a window size of 4 bytes interpreting the bytes as a little-endianu32, andMinCDCHash4, where the evaluation function ishash(u32::from_le_bytes(bytes[i - 4..i])). The hash function used is the very simplehash(x) = x.wrapping_mul(a).wrapping_add(b), for some constantsaandb.
MinCDCHash4 can be slightly (~10%) slower but is far more robust and
predictable, it is the recommended default.
Performance
MinCDC is several times faster than the commonly used
FastCDC while providing a similar amount of
deduplicating power. To benchmark this I downloaded all available Linux kernel
6.x.tar archives (tools/download-linux.sh) and ran the below algorithms on
them, all of them configured to target an expected chunk size of 8 KiB.
To determine the chunking speed I only chunked linux-6.0.tar while the file
was loaded into memory to avoid disk overhead. The dedup% is one minus the total
size of unique chunks divided by the total size of all input files (thus higher
is better). The normalized dedup% is the same percentage acquired from repeating
the experiment with different window sizes until the mean chunk size matched 8
KiB (+/- 1%). This is important when comparing deduplication power since
smaller chunks typically means better deduplication.
| Algorithm | AMD 9950X | Apple M2 Pro | Dedup% | Mean Chunk Size | Dedup% (normalized) | | --------------|-------------|--------------|--------|-----------------|---------------------| | MinCDCHash4-s | 41.3 GB / s | 23.8 GB / s | 61.08% | 8015 | 60.92% | | MinCDCHash4-l | 44.5 GB / s | 15.7 GB / s | 61.57% | 8221 | 61.57% | | MinCDC4-s | 41.7 GB / s | 26.1 GB / s | 62.11% | 7383 | 60.52% | | MinCDC4-l | 42.0 GB / s | 16.9 GB / s | 64.51% | 6436 | 60.69% | | FastCDC-s | 6.6 GB / s | 4.1 GB / s | 54.38% | 12866 | 61.81% | | FastCDC-l | 5.2 GB / s | 3.2 GB / s | 54.87% | 12764 | ~ 62%* |
Here the "-s" variants use a small window size of 8 KiB +/- 25% (min=6144, max=10240), and the "-l" variants use a larger window size of 8 KiB +/- 50% (min=4096, max=12288). The maximum chunk size was increased further for FastCDC as it inherently has a long tail of chunk sizes (see below), this did not impact chunking speed much.
The normalized dedup% for FastCDC-l is marked with an asterisk because I was unable to get the mean chunk size within 1% of 8KiB. The mean size would suddenly jump from 7741 to 10846 just by making a tiny adjustment in window size. For comparison, MinCDCHash4-l with a mean size of 7741 has a dedup% of 62.49% versus FastCDC-l's 62.75%.
Chunk Distribution
Unlike most other content-defined chunking algorithms, the distribution of chunk
sizes generated by MinCDCHash4 is almost entirely uniform in the range
min_size, max_size. This makes it very predictable and well-behaved; the
expected chunk size is also very close to the mean chunk size. Compare that with
FastCDC's distribution for an expected chunk size of 8 KiB:
| MinCDCHash4 | FastCDC | |-------------|---------| | <img src="assets/mincdc-chunk-size-distr.png" width=300> | <img src="assets/fastcdc-chunk-size-distr.png" width=300>
While the peak is at 8 KiB as expected for FastCDC, there is a long and heavy tail, increasing the mean chunk size by a lot. MinCDCHash4 never creates a chunk outside of the specified range, except for the last chunk which may be smaller.
There is still a bias towards smaller chunks as MinCDC breaks ties in the minimum value towards the earlier breakpoint, but this bias is relatively small.
Related Skills
qqbot-channel
349.2kQQ 频道管理技能。查询频道列表、子频道、成员、发帖、公告、日程等操作。使用 qqbot_channel_api 工具代理 QQ 开放平台 HTTP 接口,自动处理 Token 鉴权。当用户需要查看频道、管理子频道、查询成员、发布帖子/公告/日程时使用。
docs-writer
100.3k`docs-writer` skill instructions As an expert technical writer and editor for the Gemini CLI project, you produce accurate, clear, and consistent documentation. When asked to write, edit, or revie
model-usage
349.2kUse CodexBar CLI local cost usage to summarize per-model usage for Codex or Claude, including the current (most recent) model or a full model breakdown. Trigger when asked for model-level usage/cost data from codexbar, or when you need a scriptable per-model summary from codexbar cost JSON.
arscontexta
3.0kClaude Code plugin that generates individualized knowledge systems from conversation. You describe how you think and work, have a conversation and get a complete second brain as markdown files you own.
