Better output from the same model. Fused computation, adaptive precision, surgical expert loading. 305 KB, 19 backends, zero dependencies. https://inference-x.com
9 lines
348 B
Markdown
9 lines
348 B
Markdown
# Inference-X Benchmark Results
|
|
|
|
**Date:** 2026-02-16T16:33:05Z
|
|
**Hardware:** AMD EPYC-Rome Processor | 17GB RAM | 6 cores | AVX2
|
|
**Engine:** Inference-X v1.0.0
|
|
|
|
| Model | Params | Quant | Prefill (tok/s) | Generate (tok/s) | First Token (s) | RAM Peak |
|
|
|-------|--------|-------|-----------------|------------------|-----------------|----------|
|