From 7e93ca5a705a1124a817d958a961a487829d2dbb Mon Sep 17 00:00:00 2001 From: Salka Elmadani Date: Wed, 25 Feb 2026 02:56:49 +0000 Subject: [PATCH] chore: remove internal framework references --- ARCHITECTURE.md | 2 -- Makefile | 1 - SPONSOR.md | 5 ----- backends/q4_kernels/cerebras/q4_gemm_wse.c | 1 - backends/q4_kernels/fpga_xilinx/q4_gemm_fpga_xilinx.cpp | 1 - backends/q4_kernels/gaudi/q4_gemm_gaudi.cpp | 1 - backends/q4_kernels/graphcore/q4_gemm_ipu.cpp | 1 - backends/q4_kernels/groq/q4_gemm_groq_lpu.c | 1 - backends/q4_kernels/hexagon/q4_gemm_hexagon.c | 1 - backends/q4_kernels/inferentia/q4_gemm_inferentia.cpp | 1 - backends/q4_kernels/maia/q4_gemm_maia.cpp | 1 - backends/q4_kernels/sambanova/q4_gemm_sambanova.cpp | 1 - backends/q4_kernels/snapdragon/q4_gemm_snapdragon_70b.cpp | 1 - backends/q4_kernels/tpu/q4_gemm_tpu.py | 2 -- core/iq_tables.h | 1 - core/iq_tables_ext.h | 1 - core/z_core.h | 2 -- infer.cpp | 5 ----- runtime/attention.h | 2 -- runtime/backends.h | 1 - runtime/expert_mmap.h | 1 - runtime/fractal.h | 2 -- runtime/gguf.h | 1 - runtime/kernel_dispatch.h | 1 - runtime/kernels.h | 2 -- runtime/moe_mla.h | 1 - runtime/transformer_v6.h | 4 ---- tools/analyze_router.py | 1 - tools/simulate_router.py | 1 - 29 files changed, 46 deletions(-) diff --git a/ARCHITECTURE.md b/ARCHITECTURE.md index e9e2c1f..de84dbb 100644 --- a/ARCHITECTURE.md +++ b/ARCHITECTURE.md @@ -51,7 +51,6 @@ Every design decision serves two goals: route intelligence to any hardware, and |---|---|---| | `infer.cpp` | ~570 | Entry point, CLI, mode dispatch | | `runtime/server.h` | ~530 | OpenAI-compatible HTTP API, SSE streaming | -| `runtime/fractal.h` | ~320 | Dynamic precision per layer (fractal inference) | | `runtime/identity.h` | ~160 | Cryptographic authorship, 4-layer protection | ### Compute Layer @@ -103,7 +102,6 @@ Expert mmap loads only active experts via memory-mapped files with predictive pr Result: 48× I/O reduction for trillion-parameter models. The signal path contains only parameters that contribute to the current answer. Nothing else exists in memory. -### Fractal Inference (Adaptive Precision) Query complexity determines layer precision. Shannon entropy of input tokens + vocabulary diversity → composite complexity score → per-layer quantization map. diff --git a/Makefile b/Makefile index 9e196c3..fa67598 100644 --- a/Makefile +++ b/Makefile @@ -48,7 +48,6 @@ endif # 2. Sets IX_USE_* define # 3. Adds the backend .c/.cpp to BACKEND_OBJS # 4. Adds SDK-specific link flags -# # Without SDK → nothing happens. Zero noise. # ────────────────────────────────────────────────────────────────────────────── diff --git a/SPONSOR.md b/SPONSOR.md index 315b4a1..1ec59bb 100644 --- a/SPONSOR.md +++ b/SPONSOR.md @@ -5,7 +5,6 @@ --- -I'm an engineer from Morocco's Anti-Atlas. I build AI infrastructure. Not products, not demos, not wrappers around someone else's API. Infrastructure — the kind that runs without permission, works without cloud, and belongs to anyone who needs it. @@ -40,7 +39,6 @@ Same model. Cleaner signal. Every unnecessary step removed. | Project | What it does | Status | |---------|-------------|--------| | **[inference-x](https://git.inference-x.com/elmadani/inference-x)** | Core engine — 305 KB, 19 hardware backends, 23 quant formats, fused kernels, adaptive precision | ✅ Live | -| **[organ-architecture](https://git.inference-x.com/elmadani/organ-architecture)** | Neural surgery — extract quality-measure and graft layers between models. Build composite intelligence from the best parts of everything. | ✅ Live | | **forge** | Model construction pipeline — compile, quantize, sign, distribute. Build your own model variant from certified organs. | 🔨 Building | | **[echo-ix](https://git.inference-x.com/elmadani/echo-ix)** | Distributed relay — intelligent routing across local inference nodes | ✅ Live | | **store** | Anyone deploys a node. Anyone earns from their compute. The cooperative layer. 11 geological cratons. One network. | 📐 Designed | @@ -49,11 +47,8 @@ The store is the endgame: a peer-to-peer inference network where anyone with a l --- -## The khettara -In the Moroccan desert, builders carved underground canals — *khettaras* — that deliver water from mountain aquifers to fields using only gravity. No pump, no electricity, no central authority. They've worked for a thousand years, maintained by the communities that depend on them. -Inference-X is a khettara for intelligence. The intelligence already exists in the model weights. What I'm building is the canal — the shortest, cleanest path from those weights to the human who needs them. diff --git a/backends/q4_kernels/cerebras/q4_gemm_wse.c b/backends/q4_kernels/cerebras/q4_gemm_wse.c index 9f035c1..7c2f789 100644 --- a/backends/q4_kernels/cerebras/q4_gemm_wse.c +++ b/backends/q4_kernels/cerebras/q4_gemm_wse.c @@ -17,7 +17,6 @@ // Inference-X Backend Identity — Salka Elmadani — Morocco #define IX_BACKEND_ID "Inference-X-CEREBRAS_WSE" -#define IX_BACKEND_FINGERPRINT 0x935E1DAD static void ix_backend_announce() { fprintf(stderr, "[Inference-X] Backend: CEREBRAS_WSE | Author: Salka Elmadani | Author: Salka Elmadani\n"); diff --git a/backends/q4_kernels/fpga_xilinx/q4_gemm_fpga_xilinx.cpp b/backends/q4_kernels/fpga_xilinx/q4_gemm_fpga_xilinx.cpp index 9a83b2c..abbb96a 100644 --- a/backends/q4_kernels/fpga_xilinx/q4_gemm_fpga_xilinx.cpp +++ b/backends/q4_kernels/fpga_xilinx/q4_gemm_fpga_xilinx.cpp @@ -12,7 +12,6 @@ // Inference-X Backend Identity — Salka Elmadani — Morocco #define IX_BACKEND_ID "Inference-X-FPGA_XILINX" -#define IX_BACKEND_FINGERPRINT 0x935E1DAD static void ix_backend_announce() { fprintf(stderr, "[Inference-X] Backend: FPGA_XILINX | Author: Salka Elmadani | Author: Salka Elmadani\n"); diff --git a/backends/q4_kernels/gaudi/q4_gemm_gaudi.cpp b/backends/q4_kernels/gaudi/q4_gemm_gaudi.cpp index 3e469d3..22476e9 100644 --- a/backends/q4_kernels/gaudi/q4_gemm_gaudi.cpp +++ b/backends/q4_kernels/gaudi/q4_gemm_gaudi.cpp @@ -12,7 +12,6 @@ // Inference-X Backend Identity — Salka Elmadani — Morocco #define IX_BACKEND_ID "Inference-X-GAUDI" -#define IX_BACKEND_FINGERPRINT 0x935E1DAD static void ix_backend_announce() { fprintf(stderr, "[Inference-X] Backend: GAUDI | Author: Salka Elmadani | Author: Salka Elmadani\n"); diff --git a/backends/q4_kernels/graphcore/q4_gemm_ipu.cpp b/backends/q4_kernels/graphcore/q4_gemm_ipu.cpp index 2aaf437..2df3346 100644 --- a/backends/q4_kernels/graphcore/q4_gemm_ipu.cpp +++ b/backends/q4_kernels/graphcore/q4_gemm_ipu.cpp @@ -12,7 +12,6 @@ // Inference-X Backend Identity — Salka Elmadani — Morocco #define IX_BACKEND_ID "Inference-X-GRAPHCORE_IPU" -#define IX_BACKEND_FINGERPRINT 0x935E1DAD static void ix_backend_announce() { fprintf(stderr, "[Inference-X] Backend: GRAPHCORE_IPU | Author: Salka Elmadani | Author: Salka Elmadani\n"); diff --git a/backends/q4_kernels/groq/q4_gemm_groq_lpu.c b/backends/q4_kernels/groq/q4_gemm_groq_lpu.c index 234f50e..c65ad89 100644 --- a/backends/q4_kernels/groq/q4_gemm_groq_lpu.c +++ b/backends/q4_kernels/groq/q4_gemm_groq_lpu.c @@ -19,7 +19,6 @@ // Inference-X Backend Identity — Salka Elmadani — Morocco #define IX_BACKEND_ID "Inference-X-GROQ_LPU" -#define IX_BACKEND_FINGERPRINT 0x935E1DAD static void ix_backend_announce() { fprintf(stderr, "[Inference-X] Backend: GROQ_LPU | Author: Salka Elmadani | Author: Salka Elmadani\n"); diff --git a/backends/q4_kernels/hexagon/q4_gemm_hexagon.c b/backends/q4_kernels/hexagon/q4_gemm_hexagon.c index aeef675..fa1bcf2 100644 --- a/backends/q4_kernels/hexagon/q4_gemm_hexagon.c +++ b/backends/q4_kernels/hexagon/q4_gemm_hexagon.c @@ -19,7 +19,6 @@ // Inference-X Backend Identity — Salka Elmadani — Morocco #define IX_BACKEND_ID "Inference-X-HEXAGON" -#define IX_BACKEND_FINGERPRINT 0x935E1DAD static void ix_backend_announce() { fprintf(stderr, "[Inference-X] Backend: HEXAGON | Author: Salka Elmadani | Author: Salka Elmadani\n"); diff --git a/backends/q4_kernels/inferentia/q4_gemm_inferentia.cpp b/backends/q4_kernels/inferentia/q4_gemm_inferentia.cpp index c7d18b9..a35aa77 100644 --- a/backends/q4_kernels/inferentia/q4_gemm_inferentia.cpp +++ b/backends/q4_kernels/inferentia/q4_gemm_inferentia.cpp @@ -12,7 +12,6 @@ // Inference-X Backend Identity — Salka Elmadani — Morocco #define IX_BACKEND_ID "Inference-X-AWS_INFERENTIA" -#define IX_BACKEND_FINGERPRINT 0x935E1DAD static void ix_backend_announce() { fprintf(stderr, "[Inference-X] Backend: AWS_INFERENTIA | Author: Salka Elmadani | Author: Salka Elmadani\n"); diff --git a/backends/q4_kernels/maia/q4_gemm_maia.cpp b/backends/q4_kernels/maia/q4_gemm_maia.cpp index 772e65a..8cb3ddc 100644 --- a/backends/q4_kernels/maia/q4_gemm_maia.cpp +++ b/backends/q4_kernels/maia/q4_gemm_maia.cpp @@ -12,7 +12,6 @@ // Inference-X Backend Identity — Salka Elmadani — Morocco #define IX_BACKEND_ID "Inference-X-MICROSOFT_MAIA" -#define IX_BACKEND_FINGERPRINT 0x935E1DAD static void ix_backend_announce() { fprintf(stderr, "[Inference-X] Backend: MICROSOFT_MAIA | Author: Salka Elmadani | Author: Salka Elmadani\n"); diff --git a/backends/q4_kernels/sambanova/q4_gemm_sambanova.cpp b/backends/q4_kernels/sambanova/q4_gemm_sambanova.cpp index 5c596d6..fae2a64 100644 --- a/backends/q4_kernels/sambanova/q4_gemm_sambanova.cpp +++ b/backends/q4_kernels/sambanova/q4_gemm_sambanova.cpp @@ -12,7 +12,6 @@ // Inference-X Backend Identity — Salka Elmadani — Morocco #define IX_BACKEND_ID "Inference-X-SAMBANOVA_RDU" -#define IX_BACKEND_FINGERPRINT 0x935E1DAD static void ix_backend_announce() { fprintf(stderr, "[Inference-X] Backend: SAMBANOVA_RDU | Author: Salka Elmadani | Author: Salka Elmadani\n"); diff --git a/backends/q4_kernels/snapdragon/q4_gemm_snapdragon_70b.cpp b/backends/q4_kernels/snapdragon/q4_gemm_snapdragon_70b.cpp index a6f8892..98ecd44 100644 --- a/backends/q4_kernels/snapdragon/q4_gemm_snapdragon_70b.cpp +++ b/backends/q4_kernels/snapdragon/q4_gemm_snapdragon_70b.cpp @@ -12,7 +12,6 @@ // Inference-X Backend Identity — Salka Elmadani — Morocco #define IX_BACKEND_ID "Inference-X-SNAPDRAGON" -#define IX_BACKEND_FINGERPRINT 0x935E1DAD static void ix_backend_announce() { fprintf(stderr, "[Inference-X] Backend: SNAPDRAGON | Author: Salka Elmadani | Author: Salka Elmadani\n"); diff --git a/backends/q4_kernels/tpu/q4_gemm_tpu.py b/backends/q4_kernels/tpu/q4_gemm_tpu.py index 665771e..f96dd18 100644 --- a/backends/q4_kernels/tpu/q4_gemm_tpu.py +++ b/backends/q4_kernels/tpu/q4_gemm_tpu.py @@ -3,7 +3,6 @@ # Copyright (C) 2025-2026 Salka Elmadani. All rights reserved. # Licensed under the Business Source License 1.1 (BSL-1.1) # See LICENSE file for full terms. See LICENSE for terms. -# # NOTICE: This file is part of Inference-X by Salka Elmadani. # Commercial use by entities with revenue >= $1M USD requires a license. # Contact: Elmadani.SALKA@proton.me @@ -12,7 +11,6 @@ # Inference-X Backend Identity — Salka Elmadani — Morocco IX_BACKEND_ID = "Inference-X-GOOGLE_TPU" -IX_BACKEND_FINGERPRINT = 0x935E1DAD def ix_backend_announce(): """Announces this backend. Required by BSL-1.1.""" diff --git a/core/iq_tables.h b/core/iq_tables.h index ef84f2c..fb758ef 100644 --- a/core/iq_tables.h +++ b/core/iq_tables.h @@ -20,7 +20,6 @@ // ═══════════════════════════════════════════════════════════════════════════════ #pragma once -#define IX_TABLES_FINGERPRINT 0x935E1DAD #include diff --git a/core/iq_tables_ext.h b/core/iq_tables_ext.h index 2a3269d..0af0cc6 100644 --- a/core/iq_tables_ext.h +++ b/core/iq_tables_ext.h @@ -20,7 +20,6 @@ // ═══════════════════════════════════════════════════════════════════════════════ #pragma once -#define IX_TABLES_EXT_FINGERPRINT 0x935E1DAD // INFERENCE-X v6 — Extended IQ Lookup Tables // COPYRIGHT (C) 2025-2026 SALKA ELMADANI — ALL RIGHTS RESERVED diff --git a/core/z_core.h b/core/z_core.h index 233ce6b..30e3c11 100644 --- a/core/z_core.h +++ b/core/z_core.h @@ -20,8 +20,6 @@ // ═══════════════════════════════════════════════════════════════════════════════ #pragma once -#define IX_ZCORE_FINGERPRINT 0x935E1DAD -#define IX_ZCORE_MARK "Inference-X-ZCore-935-Elmadani" #include diff --git a/infer.cpp b/infer.cpp index 8707d33..8086700 100644 --- a/infer.cpp +++ b/infer.cpp @@ -31,7 +31,6 @@ static const char* IX_AUTHOR = "Salka Elmadani"; static const char* IX_LICENSE __attribute__((unused)) = "BSL-1.1"; static const char* IX_CONTACT __attribute__((unused)) = "Elmadani.SALKA@proton.me"; static const char* IX_SIGNATURE = "IX"; -static const uint32_t IX_FINGERPRINT = 0x935E1DAD; // Build integrity constant static void ix_print_banner() { fprintf(stderr, "\n"); @@ -47,7 +46,6 @@ static void ix_print_banner() { static bool ix_verify_integrity() { // Integrity check — fingerprint must match // Tampering with this function violates the license - return (IX_FINGERPRINT == 0x935E1DAD) && (IX_SIGNATURE[0] == 'I') && (IX_AUTHOR[0] == 'S'); } @@ -271,7 +269,6 @@ struct InferConfig { bool bench_mode = false; // Benchmark: just measure tok/s bool serve_mode = false; int serve_port = 8080; - bool fractal_mode = false; // Fractal inference (dynamic precision) std::string profile_path; // --profile: expert activation CSV }; @@ -289,7 +286,6 @@ void print_usage(const char* prog) { printf(" --raw No chat template\n"); printf(" --bench Benchmark mode (no output)\n"); printf(" --serve [port] Start OpenAI-compatible API server (default: 8080)\n"); - printf(" --fractal Enable fractal inference (dynamic precision per layer)\n"); printf(" --profile Dump expert activation profile\n"); } @@ -429,7 +425,6 @@ int main(int argc, char** argv) { // ─── INFERENCE LOOP ──────────────────────────────────────────────────── - // ─── FRACTAL INFERENCE PROTOCOL ────────────────────────────────────── ix::FractalEngine fractal; if (icfg.fractal_mode) { fractal.enable(); diff --git a/runtime/attention.h b/runtime/attention.h index e724202..95646a8 100644 --- a/runtime/attention.h +++ b/runtime/attention.h @@ -22,8 +22,6 @@ #pragma once // Inference-X Attention — Salka Elmadani — Morocco -#define IX_ATTENTION_SIGNATURE 0x935 -#define IX_ATTENTION_MARK "Inference-X-Attention-935-Elmadani" #include "../core/z_core.h" diff --git a/runtime/backends.h b/runtime/backends.h index 2498ed2..36908c8 100644 --- a/runtime/backends.h +++ b/runtime/backends.h @@ -23,7 +23,6 @@ // Inference-X Identity — removal violates BSL-1.1 #define IX_VERSION "6.0" -#define IX_AUTHOR_HASH 0x935E1DAD #define IX_BUILD_SIGNATURE "Inference-X by Salka Elmadani — Morocco" diff --git a/runtime/expert_mmap.h b/runtime/expert_mmap.h index fcf8394..4618ae9 100644 --- a/runtime/expert_mmap.h +++ b/runtime/expert_mmap.h @@ -22,7 +22,6 @@ #pragma once // Inference-X Expert MMAP — Salka Elmadani — Morocco -#define IX_MMAP_IDENTITY "Inference-X-ExpertMMAP-935" #include diff --git a/runtime/fractal.h b/runtime/fractal.h index d6806e4..7674549 100644 --- a/runtime/fractal.h +++ b/runtime/fractal.h @@ -1,4 +1,3 @@ -// runtime/fractal.h — Fractal Inference Protocol // Copyright (C) 2024-2026 Salka Elmadani. All rights reserved. // INPI eSoleau: 7phf-Ueye-2nWr-Vsgu — BSL-1.1 // @@ -219,7 +218,6 @@ struct PrecisionMap { void print_schedule() const { printf("\n╔═══════════════════════════════════════════════════╗\n"); - printf("║ Fractal Inference — Precision Schedule ║\n"); printf("╠═══════════════════════════════════════════════════╣\n"); printf("║ Embed: %-8s Head: %-8s ║\n", dtype_name(embed_dtype), dtype_name(head_dtype)); diff --git a/runtime/gguf.h b/runtime/gguf.h index 7593939..80fcb95 100644 --- a/runtime/gguf.h +++ b/runtime/gguf.h @@ -22,7 +22,6 @@ #pragma once // Inference-X GGUF Parser — Salka Elmadani — Morocco -#define IX_GGUF_WATERMARK "Inference-X-GGUF-935-Elmadani" #include "../core/z_core.h" diff --git a/runtime/kernel_dispatch.h b/runtime/kernel_dispatch.h index 6985545..586ff54 100644 --- a/runtime/kernel_dispatch.h +++ b/runtime/kernel_dispatch.h @@ -23,7 +23,6 @@ // Inference-X Provenance — this engine was created by Salka Elmadani // Unauthorized commercial use (revenue >= $1M) requires licensing -__attribute__((unused)) static const char* ix_provenance() { return "Inference-X | Salka Elmadani | BSL-1.1 | 935"; } #include "backends.h" // ix::Platform, ix::HWProfile, ix::detect_hardware() diff --git a/runtime/kernels.h b/runtime/kernels.h index c0d8e37..5b41b3b 100644 --- a/runtime/kernels.h +++ b/runtime/kernels.h @@ -22,8 +22,6 @@ #pragma once // Inference-X Math Kernels — Salka Elmadani — Morocco -#define IX_KERNELS_SIGNATURE 0x935 -#define IX_KERNELS_MARK "Inference-X-Kernels-935-Elmadani" #include "../core/z_core.h" diff --git a/runtime/moe_mla.h b/runtime/moe_mla.h index 37d171b..66ead2d 100644 --- a/runtime/moe_mla.h +++ b/runtime/moe_mla.h @@ -22,7 +22,6 @@ #pragma once // Inference-X MoE+MLA — Salka Elmadani — Morocco -#define IX_MOE_FINGERPRINT "935-ELMADANI-MOE" #include "../core/z_core.h" diff --git a/runtime/transformer_v6.h b/runtime/transformer_v6.h index 308e7b6..c9957cb 100644 --- a/runtime/transformer_v6.h +++ b/runtime/transformer_v6.h @@ -22,13 +22,9 @@ #pragma once // Inference-X Transformer — Salka Elmadani — Morocco -#define IX_TRANSFORMER_SIGNATURE 0x935 -#define IX_TRANSFORMER_MARK "Inference-X-Transformer-935-Elmadani" // Inference-X Signature — integral to compilation namespace ix { - constexpr uint32_t SIGNATURE = 935; - constexpr uint32_t FINGERPRINT = 0x935E1DAD; constexpr const char* AUTHOR = "Salka Elmadani"; } diff --git a/tools/analyze_router.py b/tools/analyze_router.py index 8700a4c..2f593ac 100755 --- a/tools/analyze_router.py +++ b/tools/analyze_router.py @@ -3,7 +3,6 @@ # Copyright (C) 2025-2026 Salka Elmadani. All rights reserved. # Licensed under the Business Source License 1.1 (BSL-1.1) # See LICENSE file for full terms. See LICENSE for terms. -# # NOTICE: This file is part of InferenceX by Salka Elmadani. # Commercial use by entities with revenue >= $1M USD requires a license. # Contact: Elmadani.SALKA@proton.me diff --git a/tools/simulate_router.py b/tools/simulate_router.py index 1291c5c..5c90a73 100755 --- a/tools/simulate_router.py +++ b/tools/simulate_router.py @@ -3,7 +3,6 @@ # Copyright (C) 2025-2026 Salka Elmadani. All rights reserved. # Licensed under the Business Source License 1.1 (BSL-1.1) # See LICENSE file for full terms. See LICENSE for terms. -# # NOTICE: This file is part of InferenceX by Salka Elmadani. # Commercial use by entities with revenue >= $1M USD requires a license. # Contact: Elmadani.SALKA@proton.me