Press Release

Slash AI Energy by 99.5%. rolvsparse© Delivers 20–177× Speedups — Software‑Only, Validated, Patent‑Pending.

FORT LAUDERDALE, Fla.–(BUSINESS WIRE)–rolv, LLC today announced breakthrough benchmarks for rolvsparse©, a patent‑pending software compute primitive delivering 20–177× AI inference speedups and up to 99.5% energy reduction on unmodified models, compared to vendor‑optimized dense and sparse libraries. All energy numbers use real hardware power readings (50 ms polling). No hardware changes, retraining, or precision loss. Most results are independently validated by the University of Miami Frost Institute (some benchmarks are very recent), with bit‑identical SHA‑256 hashes across all platforms.

rolvsparse© achieves 20–177× speedups and 98–99.5% energy savings on real Hugging Face production models. Even fully dense workloads — 0% sparsity — reach 63× acceleration, beating vendor-optimized GPU libraries on CPU alone. Five patents are pending.

“The core idea for rolvsparse came to me on a bike ride in May 2025. Everything since — patents, prototypes, self‑taught stacks, benchmarks, and university validation — has been relentless execution. Why build more data centers when your existing ones can achieve 83× faster performance and 99% greener operations?” — Rolv E. Heggenhougen, Founder & CEO, rolv.ai

Benchmark Highlights (Real Hardware Power Readings), Speedup and Energy savings:

Frontier‑Scale LLMs & MoEs (GPU)

  • Llama‑4 Maverick 400B: 133.5×, 99.9%
  • Llama‑4 400B: 125.3×, 99.4%
  • Llama‑4 400B TTFT: 100.9×
  • Llama‑4 Maverick (MoE): 20.7×, 81.5%
  • Llama‑4 Maverick TTFT: 177×
  • Llama‑4 Scout: 81.7×, 98.8%
  • DeepSeek‑R1 (256 experts): 44.2×, 98.7%
  • Kimi K2.5 (~1T MoE): 10.5×, 90.6%
  • Mixtral 8×22B: 55.1×, 98.2%
  • Claude 3.5‑class FFN: 83×, 98.8%

Qwen Family (GPU + TPU)

  • Qwen3‑235B: 7.8×, 95.5%
  • Qwen2.5‑72B‑Instruct: 50.5×, 91.4%
  • Qwen2.5‑32B (TPU v5e): 5.9×, 83.0%
  • Qwen3.5‑35B GPTQ‑Int4, 64 experts: 9.4×, 89.3%

Specialized & Classical Models

  • GLM‑OCR 0.9B: 50.0×, 98.0%
  • Dense 20k×20k (0% sparse): 63×, 98.42%
  • Finite Element Solver: 163×, 99.5%
  • Netflix Prize RecSys: 3.1×, 67.7%

CPU Benchmarks (Real FFNs)

  • GPT‑J‑6B FFN (40% sparse): 314.6×, 99.7%
  • Mistral‑7B FFN (0% sparse): 253.6×, 99.6%
  • Llama‑2‑7B FFN (70% sparse): 169.2×, 99.4%
  • Kimi K2.5 Expert Slice: 40.3×, 97.9%
  • BERT‑Base FFN (0% sparse): 12.3×, 91.8%

All outputs are SHA‑256 verified for deterministic correctness.

Note: Most CPU benchmarks w/rolv beat GPU w/o rolv.

How rolvsparse© Works

Vendor libraries waste cycles on zeros — the “zero‑FLOP bottleneck.” rolvsparse© restructures arithmetic at the primitive level, skipping meaningless operations while guaranteeing exact outputs (e.g., deterministic hash 8dbe5f139f…dd8dd). It deploys as a drop‑in software layer across all hardware. Users run the open verifier at rolv.ai to generate baselines; rolv returns personalized comparison reports with real power readings.

Market Impact & Applications

AI data centers may reach 9% of U.S. electricity by 2030; hyperscalers have committed $700B+ in AI capex. rolvsparse© reduces energy use by 98–99.5%, boosts throughput on existing infrastructure, and enables edge viability (mobile processors: 70× sparse). Applications include LLMs/MoEs, agents, mobile inference, engineering simulation, RecSys/finance, sustainability, and sovereign AI.

Recent benchmarks on Substack (inl. Json, flops, tokens): rolv.substack.com

Contacts

Media Contact
Rolv E. Heggenhougen, CEO [email protected] • 954.253.4443 • rolv.ai • X: @rolveitrem

Author

Related Articles

Back to top button