1.4 KiB
1.4 KiB
Salka Elmadani
Building Inference-X — better output from the same model.
Universal AI inference engine. Fused computation, adaptive precision, surgical expert loading. 305 KB, 19 backends, zero dependencies. Built in Morocco.
What I'm working on
| Project | Description |
|---|---|
| Inference-X | Universal inference engine — 305 KB binary, fused dequant+dot kernels, Shannon entropy adaptive precision, 23 quantization formats, 19 hardware backends |
| Solar Inference | Deploying AI on direct solar current. Adaptive precision + 25W power budget = inference powered by a star. Anti-Atlas, 2026. |
Philosophy
The best inference engine is the one you don't notice. You should hear the model, not the framework.
How it works
The same model produces higher-fidelity output through Inference-X because the computation path is cleaner: fused kernels eliminate intermediate buffers, adaptive precision allocates depth where it matters, and surgical expert loading keeps only active parameters in memory.
Links
🌐 inference-x.com · 📧 Elmadani.SALKA@proton.me · 𝕏 @ElmadaniSa13111
Morocco 🇲🇦