Eight Things Our Benchmarks Reveal
MINT Research

Eight Things Our Benchmarks Reveal That Nobody Expected

March 2026 · Black Sheep AI Research

We benchmarked MINT across 7 model families, 5 benchmark suites, and over 40,000 questions. The results don’t just validate MINT — they challenge fundamental assumptions that the quantization community has been operating under for years.

Some of these findings are about MINT specifically. Others are about the field. We’ve separated the genuinely surprising results — the ones that should change how people think about quantization — from the findings that are interesting but less unexpected.

All data comes from our complete benchmark results and the MINT paper.

The Genuinely Surprising Findings

1. Group size matters more than bit-width — and the conventional default is wrong

This is the single most important finding and it is underappreciated. The entire quantization community has been treating group size as a fixed hyperparameter — 128 is the near-universal default — while obsessing over bit-width selection. MINT’s knapsack solver, given the freedom to choose, assigns group size 32 to 85% of tensors.

The math is simple but the implication is profound: spending 0.125 bytes per parameter on finer quantization groups yields more quality than spending 0.5–1.5 bytes per parameter upgrading select tensors to 8-bit or 16-bit. This means every existing mixed-precision method that fixes group size at 128 is leaving quality on the table — not at the margins, but as the dominant effect.

No one in the prior literature had demonstrated this because no one had given an optimizer the freedom to discover it. When you let the math decide instead of a human, the answer is unambiguous: invest in finer groups, not higher bit-widths.

2. Mean perplexity can give completely inverted quality rankings

The GLM-4.7-Flash result is genuinely alarming for the field. BF16 has mean PPL 11.5, which appears worse than MINT’s 10.2 — suggesting quantization improves the model. That is obviously nonsensical. Median PPL gives the correct ranking: BF16 8.5, MINT 8.7.

The cause: five catastrophic outlier sequences where BF16 produces per-sequence perplexity values of 25,000–81,000 and quantization noise happens to stabilize them. This is not a MINT-specific finding — it is a methodological warning for every quantization paper that reports only mean PPL.

The controlled noise experiment seals it: random Gaussian noise at matched magnitude produces the same median shift, proving the “regularization” is an artifact. If you have ever seen a paper claiming quantization improves perplexity, this is probably why.

3. The 2-bit/3-bit cliff is a universal architectural constant

Across every model tested — dense and MoE, 8B to 109B, Qwen, Llama, Mixtral, and GLM families — 2-bit SQNR maxes out at ~8.7 dB and 3-bit SQNR starts at ~10.4 dB. There is a clean 2 dB gap with no configurations in between. This gap appears to be a property of how neural network weight distributions interact with round-to-nearest quantization at these bit levels. It is not architecture-dependent.

The practical consequence is dramatic: PPL triples at 2-bit but degrades only 12.5% at 3-bit. This suggests a fundamental information-theoretic boundary — 2 bits per weight is below the threshold needed to preserve the essential structure of trained weight matrices, regardless of architecture.

The 9 dB threshold sitting cleanly in this gap is a robust, transferable design principle that other quantization methods could adopt. You do not need MINT to benefit from this finding — any quantization pipeline could add an SQNR check to prevent catastrophic 2-bit allocations.

4. Data-free consistently beats calibration-based at matched sizes

This is genuinely counterintuitive. GPTQ has access to activation statistics from real data — it knows which weights matter for actual model behaviour. MINT has only the weights themselves. Yet MINT wins by 1–4.6% on median PPL across three different MoE architectures at exactly matched model sizes.

Model GPTQ PPL (med) MINT PPL (med) Delta
Qwen3-30B-A3B9.1608.959−2.2%
Qwen2-57B-A14B6.3966.335−1.0%
Mixtral-8x7B4.6404.426−4.6%

The explanation the paper offers — that joint group-size optimization, budget-constrained allocation, and better MoE expert coverage combine to outweigh the information advantage of calibration — is plausible but not fully proven.

What is clear: the conventional wisdom that “calibration-based is always better than data-free” is wrong, at least for current GPTQ defaults with fixed group sizes.

5. Downstream benchmarks are essentially useless for evaluating quantization quality

The raw numbers drive this home. On Qwen3-30B, PPL varies 18% from min-safe to 8-bit, while ARC-Challenge varies 2.7 percentage points. On Qwen3.5-35B, MMLU (14,015 questions) spans only 1.5 pp across a 2.5× size range. MMLU actually peaks at 37 GB and slightly declines at 51 GB even as PPL keeps improving.

This means most quantization papers that report ARC-C or MMLU improvements are measuring noise, not signal. The benchmarks saturate at roughly 4-bit precision. Above that, you literally cannot tell the difference between 4-bit, 8-bit, and 16-bit using standard accuracy benchmarks. Only perplexity has the resolution to discriminate.

This is a finding that affects how the entire field should be evaluating quantization methods. If your evaluation suite consists of ARC-C and Winogrande, you are not measuring what you think you are measuring.

6. Quality returns collapse above ~5 average bits

On Qwen3-30B, going from 5.1 average bits (21.2 GB) to 8-bit (29.3 GB) improves PPL by 0.1% while costing 38% more storage. On Mixtral, the same pattern: 5 bits gets you within 1.1% of 8-bit at 31% less storage.

The practical implication is blunt: if your model fits at 5 bits, adding more bits is wasting memory that could be better spent on context length, batch size, or running a bigger model. This creates a clear deployment heuristic that did not exist before — and it applies regardless of which quantization method you use.

The Interesting But Less Surprising Findings

7. Infrastructure catches up to research

Our MLX kernel benchmarks reveal something you would never see in an academic paper. MLX had a real performance regression for group size 32 in version 0.29 — a 1.8–2.2× prefill penalty — that was fully fixed by version 0.31 through upstream PRs #1861 and #2031.

Scenario MLX 0.29.3 MLX 0.31.1
Prefill g32/g1281.8–2.2× penalty1.00× (fixed)
Generation g32/g1281.07–1.14×1.01–1.10×
Prefill g64/g128~1.3×1.00× (fixed)

This is practically important because MINT’s preference for g32 would have been impractical six months ago. It is a nice example of research and infrastructure co-evolving — and the kind of detail that matters for real deployment but never appears in a conference paper.

8. Budget-constrained allocation is surprisingly efficient at the extremes

On Llama-4-Scout (109B), the 3.5× size difference between min-safe (47 GB) and the 192 GB model produces only a 1.9× PPL difference (8.675 vs 7.359). This means someone running on a 48 GB Mac sacrifices surprisingly little quality compared to someone with 192 GB.

The shape of the quality-vs-size curve is convex in a way that favours constrained deployments. The first bytes of budget above the safety floor produce massive returns — the first 1 GB above the 4-bit floor buys a 6.8% PPL improvement on Qwen3-30B. After that, the curve flattens fast.

The practical takeaway: if you are agonising over whether your hardware is “good enough” for a given model, it probably is. The penalty for being memory-constrained is far smaller than most people assume, provided you use budget-aware allocation instead of uniform quantization.

What This Means

Several of these findings are not specific to MINT. The mean-vs-median perplexity problem affects every quantization evaluation. The 2-bit/3-bit cliff is an architectural constant that any method could exploit. The saturation of downstream benchmarks means the field needs better evaluation methodology. The diminishing returns above 5 bits should inform deployment decisions regardless of compression method.

What is specific to MINT is having the framework that made these discoveries possible. When you give an optimizer the freedom to jointly select bit-widths and group sizes under a budget constraint, the answers it produces are not what human intuition would suggest. Finer groups over higher bit-widths. No 16-bit for quantisable tensors. Embeddings stay at 4-bit despite their importance. The optimizer sees the global tradeoff that heuristics miss.

The full benchmark data is available in our complete results article. Every number is reproducible using the code at github.com/baa-ai/MINT.


Data from the MINT paper and benchmark evaluation suite. All experiments on Apple M2 Ultra (192 GB). Perplexity: WikiText-2 test split, seq_len=2048, seed 42. Downstream benchmarks: lm-evaluation-harness via MLX backend — ARC-Challenge (25-shot), Winogrande (5-shot), HellaSwag (10-shot), MMLU (5-shot, 14,015 questions). Full paper: baa.ai/articles/24-mint-paper.html. Code: github.com/baa-ai/MINT.

← Back to all articles

Ready to rethink your quantization strategy?

Our team specialises in data-free model compression that challenges conventional assumptions. Better methods, better results, no GPUs required.

Talk to Our Team