YAML Metadata Warning:empty or missing yaml metadata in repo card
Check out the documentation for more information.
ROLV Benchmark
Sparse operator verification and performance benchmark suite for ROLV PrimitiveΒ©.
Published by Rolv Eitrem Heggenhougen Β· rolv.ai Β· 3 Patents Pending
The idea in one sentence
You supply numbers only you know. You compute the expected answer yourself. ROLV gets the same answer β proving it performs real computation, not fabricated results.
What is this?
A deterministic sparse matrix benchmark with three purposes:
1. Zero-trust verification β prove ROLV correctness with your own secret inputs
2. Performance benchmark β measure ROLV speedup vs dense and CSR across sparsity levels
3. Education β understand exactly what row sparsity means and why it matters
Install
pip install rolv-benchmark
Or from source:
git clone https://github.com/rolvai/rolv-benchmark
cd rolv-benchmark
pip install -e .
Quick start
Verify with your own secret numbers
from rolv_benchmark import verify
import numpy as np
# Use numbers only YOU know β ROLV must get the same answer
my_numbers = np.array([7, 13, 42, 99, 3, 17, 8, 100] * 16, dtype=np.float32)
result = verify(x=my_numbers)
# Prints a verification report including rows you can check by hand
From the command line:
python -m rolv_benchmark verify --x 7 13 42 99 3 17 8 100
The verification logic:
- A deterministic weight matrix W is generated from a published seed β same on every machine
- Your input x is multiplied: y = W @ x
- 80% of output rows are exactly zero (ROLV skips them)
- 20% are active β you can verify each with a calculator:
sum(W[i] * x) - ROLV and dense give identical results to 1e-5 tolerance
- Publish your SHA-256 hashes to let anyone reproduce your specific run
Run the performance benchmark
python -m rolv_benchmark benchmark
Output:
ROLV BENCHMARK 10000Γ128 batch=512
Sparsity Dense ms ROLV ms Speedup Err PASS
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
0% 1.8230 1.8441 0.99Γ 0.00e+00 β
50% 1.8230 0.9412 1.94Γ 0.00e+00 β
70% 1.8230 0.5312 3.43Γ 0.00e+00 β
80% 1.8230 0.3821 4.77Γ 0.00e+00 β
90% 1.8230 0.2104 8.66Γ 0.00e+00 β
95% 1.8230 0.1203 15.15Γ 0.00e+00 β
99% 1.8230 0.0489 37.28Γ 0.00e+00 β
Show matrix parameters
python -m rolv_benchmark info
The matrix structure
W: shape (10 000, 128)
βββ 8 000 rows: exactly zero β ROLV skips all of these
βββ 2 000 rows: non-zero β ROLV computes only these
Row sparsity: 80%
ROLV compression: 5Γ
ROLV's CRCSβ’ algorithm identifies zero rows and skips them entirely β no multiply, no memory access, no energy. The speedup grows with row sparsity: at 99% only 100 rows are active out of 10 000, giving ~37Γ speedup on CPU.
The weight coefficients are rounded to 4 decimal places so any active row can be verified with a pocket calculator:
W[423] = [0.1234, -0.5678, 0.9012, ...]
x = [7, 13, 42, ...]
dot = 0.1234Γ7 + (-0.5678)Γ13 + 0.9012Γ42 + ... = ?
Why this is a stronger proof than hashes alone
Standard benchmark hashes (SHA-256 of W, x, output) prove the computation was run on specific data β but someone could pre-compute results and fake the timing. This benchmark adds a second layer:
You supply x β numbers only you know.
If ROLV returns the value you computed on your calculator, using your numbers, it cannot have pre-computed that result. The proof is zero-trust by construction.
To publish your verification:
- Run
python -m rolv_benchmark verify --x <your_numbers> - Publish the printed
hash_Wandhash_x - Anyone can reproduce by running with the same x
Scaling
ROLV advantage grows with:
| Dimension | Why |
|---|---|
| Sparsity β | More rows skipped β larger compression |
| Matrix size β | More absolute rows skipped at same sparsity % |
| Batch size β | Fixed overhead amortised across more columns |
On GPU (not in this package β see rolv.ai):
- 80% sparsity β 10β12Γ speedup on NVIDIA B200
- 95% sparsity β 15β19Γ speedup on real LLM weights
- 550/550 PASS across 6 hardware platforms
The math
For each active row i:
For zero rows: $y_i = 0$ (by construction, not approximation).
ROLV CRCSβ’ computes this as:
This is exact arithmetic β no approximation is introduced by ROLV itself. The only source of output difference from a dense model is pruning (zeroing rows), which is a user choice, not a ROLV artifact.
GPU acceleration
This package runs on CPU only for maximum portability and verifiability. The full GPU operator (NVIDIA, AMD, Intel) is available at rolv.ai.
GPU results: up to 83Γ speedup vs rocSPARSE on AMD MI300X, 13.64Γ vs cuSPARSE on NVIDIA H200, verified on exact LLaMA-3.1-70B dimensions.
Citation
@software{rolv_benchmark_2026,
author = {Heggenhougen, Rolv Eitrem},
title = {ROLV Benchmark: Sparse Operator Verification Suite},
year = {2026},
url = {https://github.com/rolvai/rolv-benchmark},
note = {3 Patents Pending}
}
ROLV PrimitiveΒ© Β· CRCSβ’ Β· RSMTβ’ Β· ROLVswitchβ’ Β· 3 Patents Pending
rolv.ai Β· rolv@rolv.ai