Original research in model compression, capability governance, and sovereign AI deployment.
RAM compression matches BF16 quality at 50% of the original model size. Our MMLU-Pro evaluation of Gemma 4 demonstrates that compute-optimal quantization preserves the capabilities that matter.
March 2026
March 2026
March 2026
March 2026
March 2026
March 2026
March 2026
March 2026
March 2026
March 2026
March 2026
March 2026
March 2026
March 2026
March 2026
March 2026
March 2026
March 2026
February 2026
February 2026
February 2026
February 2026
February 2026
February 2026
February 2026
February 2026
February 2026
February 2026
February 2026