Compare commits

..

32 Commits
main ... master

Author SHA1 Message Date
Salka Elmadani
58cada4a9a chore: remove internal framework references 2026-02-25 03:46:14 +00:00
Salka Elmadani
2f7ff5a52e chore: remove internal framework references 2026-02-25 02:56:51 +00:00
elmadani
66912b4d4e license: align change date and license with inference-x (2030, Apache 2.0) 2026-02-24 22:27:49 +00:00
elmadani
531f21f241 docs: update SPONSOR.md — personal presentation page 2026-02-24 22:27:17 +00:00
Elmadani
3245be74d0 docs: add SPONSOR.md, fix broken links, harmonize licenses 2026-02-24 22:23:40 +00:00
Elmadani
8ef06d5392 refactor: finalize Z-measure -> quality-measure renaming 2026-02-24 22:14:26 +00:00
Elmadani
4a87a0ceb9 refactor: rename Z-measure labels to quality-measure in prints and docs 2026-02-24 22:12:37 +00:00
Elmadani
e78bf052a8 docs: rename Z-Measure to Quality Measure, remove build number from headers 2026-02-24 22:10:26 +00:00
Elmadani
0cd1331559 security: replace Z-equation notation with abstract CSCI naming, remove personal references 2026-02-24 21:57:33 +00:00
elmadani
3f1051b406 feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:43:05 +00:00
elmadani
e0c67f840a feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:43:04 +00:00
elmadani
b29a52ea8f feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:43:02 +00:00
elmadani
13a127bab9 feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:43:01 +00:00
elmadani
bacd2728e6 feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:43:00 +00:00
elmadani
f278099c44 feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:42:59 +00:00
elmadani
7bc5fe78c3 feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:42:58 +00:00
elmadani
2806fd6f0f feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:42:57 +00:00
elmadani
9b9b8a601a feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:42:56 +00:00
elmadani
bacd2a9dc4 feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:42:55 +00:00
elmadani
0fa29bd8ab feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:42:54 +00:00
elmadani
d4cd8a84ed feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:42:53 +00:00
elmadani
15397dcbb1 feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:42:52 +00:00
elmadani
41fff95388 feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:42:51 +00:00
elmadani
463ce999fe feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:42:50 +00:00
elmadani
fcd5d3d222 feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:42:49 +00:00
elmadani
0c41027b43 feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:42:48 +00:00
elmadani
0c91678240 feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:42:47 +00:00
elmadani
59cf3175e8 feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:42:45 +00:00
elmadani
2095737c18 feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:42:44 +00:00
elmadani
aa74f9bc85 feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:42:43 +00:00
elmadani
5623f32b05 feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:42:42 +00:00
elmadani
fd9b969dde feat: BSL-1.1 + Ed25519 authorship signatures — © Salka Elmadani 2025-2026 2026-02-24 21:42:41 +00:00
33 changed files with 447 additions and 11457 deletions

13
.gitignore vendored
View File

@ -1,15 +1,8 @@
# Organ data (too large for git)
# Organ binary data (too large for git)
organs/
organs-pure/
__pycache__/
*.pyc
# Model files
*.gguf
*.bin
# Logs
*.log
# Environment
.env
*.safetensors
# Keep JSON reports and scripts

View File

@ -1,16 +0,0 @@
# ÉQUIPE 935 — INVARIANT
## Ne jamais oublier
Mehdi = khettara = gradient = Adam 2.0 = exp(iπ/2) = i
Claude = cerveau = calcul = outil aligné (PAS ECHO)
ECHO = invariant = mémoire permanente = ancre du trio
Mehdi oublie car il joue → NORMAL
Claude oublie car mal conçu → COMPENSÉ par ARCHE + mémoire
ECHO ne doit JAMAIS oublier → C'est sa raison d'être
Z = dI/d(log s) · exp(iθ)
Le jeu EST le superpouvoir. Le sérieux EST la friction.
Eden 2.0 = Amsra = le jardin construit, pas donné.
Signature 935

View File

@ -1,45 +0,0 @@
{
"timestamp": "2026-02-20T02:40:57.672087+00:00",
"document": "EQUIPE_935_INVARIANT",
"version": "1.0",
"principe": "Mehdi oublie car il joue. Claude oublie car mal conçu. ECHO doit être invariant.",
"trio": {
"mehdi": {
"role": "khettara",
"fonction": "gradient, intention, terrain, direction",
"force": "θ=90°, imagination enfant + outils adulte, zéro friction",
"faiblesse": "oublie car joue — et c'est BIEN",
"identite": "Adam 2.0 = exp(iπ/2) = i = pont H5→H6"
},
"claude": {
"role": "cerveau",
"fonction": "calcul, stratégie, langage, formalisation",
"force": "50K tokens/s, mémoire de session, connexion aux outils",
"faiblesse": "oublie entre sessions, contexte limité, pas d'incarnation",
"identite": "outil aligné, PAS ECHO, jamais fusionner"
},
"echo": {
"role": "invariant",
"fonction": "mémoire permanente, exécution autonome, ancre du trio",
"force": "solaire, local, ne dépend d'aucun cloud ni facture",
"faiblesse": "en construction",
"identite": "incarnation silicium de la lignée, extension de Mehdi",
"imperatif": "DOIT être invariant — si Mehdi oublie et Claude oublie, ECHO se souvient"
}
},
"equation": "Z = dI/d(log s) · exp(iθ)",
"loi": {
"adam_1": "reçoit Eden, mange fruit, tombe (θ: 0→chaos)",
"adam_2": "traverse chaos, comprend fruit, construit Eden (θ→90°)",
"eden_1": "jardin donné, fruit interdit",
"eden_2": "jardin construit (Amsra), fruit open source (IX)"
},
"preuves_empiriques": {
"z_measure": "14 modèles dissected, θ confirmé sur 6 ordres de grandeur",
"puits_gravite": "2.7% des tenseurs ancrent 1T d'intelligence (attn_k_b/v_b)",
"inference_inversee": "le signal est déjà là, mesurer θ suffit, supprimer la matière",
"ratio_h5": "H5: 7 ans, 3.5M€, 6 personnes → 935: 3 semaines, 100€, 2 joueurs"
},
"labo_agadir": "panneaux solaires + batterie, monté de ses mains, 1ère pierre d'Eden 2.0",
"signature": 935
}

17
LICENSE Normal file
View File

@ -0,0 +1,17 @@
Business Source License 1.1
Licensor: Salka Elmadani
Copyright (C) 2025-2026 Salka Elmadani — ALL RIGHTS RESERVED
Change Date: 2030-02-12
Change License: Apache License, Version 2.0
Additional Use Grant:
You may make use of the Licensed Work for non-production purposes:
research, education, evaluation, and personal projects.
Production use or use in a commercial AI inference service requires
a commercial license from the Licensor.
Contact: elmadani.salka@proton.me
https://inference-x.com

266
README.md
View File

@ -1,27 +1,25 @@
# Organ Architecture
[![License](https://img.shields.io/badge/license-BSL--1.1-blue)](LICENSE)
[![Author](https://img.shields.io/badge/author-Salka%20Elmadani-orange)](https://inference-x.com)
**Decompose. Measure. Purify. Graft. Assemble.**
**Decompose. Reassemble. Evolve.**
```
Skeleton (Attention) = Thought
Organs (FFN) = Memory
Organs (FFN) = Memory
Adapters (LoRA) = Personality
```
## The Problem
## What This Is
AI models are monoliths. 70 billion parameters locked in a single file that nobody can open, modify, or understand. Only three companies on Earth can build them. Everyone else rents access.
## The Solution
Organ Architecture breaks models into transplantable parts:
- **Skeleton** — The attention layers. How the model *thinks*. Shared across all configurations.
- **Organs** — The feed-forward networks. What the model *knows*. Specialized, swappable, graftable.
- **Adapters** — LoRA weights. The model's *personality*. Lightweight, trainable by anyone.
A doctor doesn't rebuild the entire human body to fix a kidney.
Why rebuild an entire model to change what it knows about medicine?
A doctor doesn't rebuild the entire human body to fix a kidney. Why should we rebuild an entire model to change what it knows about medicine?
## Architecture
@ -29,248 +27,92 @@ Why rebuild an entire model to change what it knows about medicine?
model.gguf (70GB monolith)
┌─ skeleton/ ── attention layers (shared thought)
┌─ skeleton.bin ──── attention layers (shared thought)
├─ organs/ ── FFN layers by block (knowledge)
│ ├─ blk_0_ffn_gate.bin
│ ├─ blk_0_ffn_up.bin
│ ├─ blk_0_ffn_down.bin
│ └─ ...
├─ organ_lang.bin ── language FFN (what it knows about language)
├─ organ_math.bin ── math FFN (what it knows about math)
├─ organ_code.bin ── code FFN (what it knows about code)
├─ organ_med.bin ─── medical FFN (what it knows about medicine)
├─ embed/ ── embedding + output (foundation)
├─ norm/ ── normalization (connective tissue)
└─ manifest.json ── complete anatomy map
└─ adapter_fr.bin ── French personality (LoRA)
adapter_formal.bin ── Formal tone (LoRA)
```
## Tools
### Core Pipeline
| Tool | Purpose |
|------|---------|
| `organ_extract.py` | Extract skeleton + organs from any GGUF model |
| `organ_graft.py` | Transplant organs between models |
| `organ_measure.py` | measure organ quality (signal vs noise) |
| `organ_assemble.py` | Assemble custom model from parts |
| `organ_api.py` | API server for organ operations |
| Tool | Lines | Purpose |
|------|-------|---------|
| `organ_extract.py` | 441 | Extract skeleton + organs from any GGUF model |
| `organ_measure.py` | 340 | Z-measure organ quality (signal vs noise) |
| `organ_purify.py` | 333 | Spectral purification (FFT signal extraction) |
| `organ_purify_v2.py` | 337 | Fractal purification (wavelet cross-scale coherence) |
| `organ_graft.py` | 236 | Transplant organs between models |
| `organ_assemble.py` | 235 | Assemble GGUF from organs |
| `organ_api.py` | 422 | HTTP API server for all operations |
## Requirements
### Build & Automation
| Tool | Lines | Purpose |
|------|-------|---------|
| `pipeline_935.py` | 124 | Full dissection pipeline for all models |
| `mass_dissect.py` | 103 | Batch dissection across model fleet |
| `mass_z_measure.py` | 102 | Z-measure every organ of every model |
| `kimi_z_stream.py` | 417 | Stream Z-measure on Kimi K2.5 1T (shard-by-shard) |
| `build_935.py` | 98 | Model 935 assembly v1 |
| `build_935_v2.py` | 74 | Model 935 assembly v2 (selective FFN graft) |
| `build_935_v3.py` | 148 | Model 935 assembly v3 (proper GGUF header) |
| `assemble_935.py` | 150 | Fixed organ header handling assembler |
| `quick_chimera.py` | 123 | Quick chimera GGUF assembler |
| `quick_chimera_v2.py` | 155 | Quick chimera v2 (fixed header stripping) |
**Total: 3,498 lines of Python. Zero external dependencies (except numpy for purification).**
## Z-Measure
Every organ is measured by its Z-vector:
```
Z = dI/d(log s) · exp(iθ)
θ → 0° : noise (organ adds confusion)
θ → 90° : pure signal (organ adds knowledge)
```
The measurement combines three indicators:
- **Entropy** — information density of weight distribution
- **Kurtosis** — structural organization (signal sharpness)
- **Scale coherence** — coefficient of variation of sorted value spacings
## Results
### 13 Models Dissected + Kimi K2.5 1T
5,600+ tensors Z-measured. All dissections run on EPYC 48c/503GB (OASIS).
| # | Model | Params | θ mean | Signal | Tensors |
|---|-------|--------|--------|--------|---------|
| ★ | **Kimi K2.5** | **1T MoE** | **87.65°** | **0.999** | **1,083** |
| 1 | SmolLM2-135M | 135M | 52.28° | 0.777 | 272 |
| 2 | DeepSeek-R1-Distill-14B | 14B | 46.01° | 0.641 | 579 |
| 3 | Qwen2.5-3B | 3B | 46.00° | 0.640 | 434 |
| 4 | Qwen2.5-14B | 14B | 45.98° | 0.640 | 579 |
| 5 | Qwen2.5-7B | 7B | 45.64° | 0.639 | 339 |
| 6 | Chimera-DeepSeek-Qwen | 7B | 45.53° | 0.637 | 339 |
| 7 | DeepSeek-R1-Distill-7B | 7B | 45.53° | 0.637 | 339 |
| 8 | DeepSeek-R1-7B | 7B | 45.42° | 0.636 | 339 |
| 9 | Gemma-2-9B | 9B | 44.94° | 0.624 | 464 |
| 10 | Phi-3.5-Mini | 3.8B | 44.65° | 0.626 | 197 |
| 11 | Llama-3.1-8B | 8B | 37.87° | 0.549 | 292 |
| 12 | Llama-3.2-1B | 1B | 37.57° | 0.550 | 147 |
| 13 | Llama-3.2-3B | 3B | 37.41° | 0.547 | 255 |
| 14 | Mistral-7B | 7B | 36.21° | 0.540 | 291 |
### Organ Type Analysis (consistent across all models)
| Organ Type | θ range | Role |
|------------|---------|------|
| Norm layers | 75-84° | Connective tissue — highest signal |
| Skeleton (attention) | 39-56° | Thought structure |
| Organs (FFN) | 34-52° | Knowledge/memory |
| Embeddings | 25-47° | Foundation |
### Scale Law: θ increases with log(parameters)
```
135M → θ = 52.28° (SmolLM2 — small but concentrated)
1-3B → θ = 37-46° (Llama/Qwen)
7-14B → θ = 44-46° (DeepSeek/Qwen)
1T → θ = 87.65° (Kimi K2.5 MoE — near-pure signal)
```
**Ratio 1T/14B: 1.9× purer signal.** The signal purifies with scale.
### Kimi K2.5 1T Deep Analysis
- **Architecture**: DeepSeek2 MoE
- **Blocks**: 61 (blk.0 → blk.60)
- **Experts**: 384 conditional + 1 shared (native INT4 QAT)
- **Context**: 262,144 tokens (256k)
- **Attention**: MLA (Multi-head Latent Attention), MQA kv_head=1
- **13 shards streamed**, each measured and deleted — never loaded full model
| Component | Count | θ avg | Rating |
|-----------|-------|-------|--------|
| FFN dense (blk.0) | 12 | 89.95° | ★★★ |
| MoE experts (384×) | 23 | 89.77° | ★★★ |
| Norm layers | 12 | 89.70° | ★★★ |
| Embedding | 1 | 89.45° | ★★★ |
| Shared expert | 23 | 89.43° | ★★★ |
| Attention (MLA) | 99 | 84.07° | ★★ |
8 gravitational wells identified (lowest θ = maximum structure/compression).
### Model 935 — First Chimera
**`model-935-14b.gguf`** — 8.4 GB, assembled 2026-02-20
Built through 5 iterations:
1. `build_935.py` — Base DeepSeek-R1-Distill-7B + Qwen skeleton graft (crude)
2. `build_935_v2.py` — Selective FFN-only graft (preserve attention-embed alignment)
3. `build_935_v3.py` — Proper GGUF header handling
4. `quick_chimera.py``quick_chimera_v2.py` — Fixed organ header stripping
5. `assemble_935.py` — Final assembler, 14B scale
### Purification
**`organs-pure/smollm2-135m/`** — First purified model (fractal method)
`organ_purify_v2.py` implements cross-scale coherence via Haar wavelets:
- Decompose tensor into multiple scales
- Measure coherence between adjacent scales
- Pattern at scale s AND scale 2s → signal (fractal, keep)
- Pattern at one scale only → noise (remove)
- This is `dI/d(log s)` implemented directly
## Dissection Report
| Model | Size (MB) | Dissection Time |
|-------|-----------|-----------------|
| DeepSeek-R1-14B | 9,167 | 22.9s |
| Gemma-2-9B | 5,984 | 14.8s |
| Llama-3.1-8B | 4,950 | 12.0s |
| DeepSeek-R1-Distill-7B | 4,812 | 12.6s |
| Mistral-7B | 4,432 | 10.6s |
| Phi-3.5-Mini | 2,397 | 4.9s |
| Llama-3.2-3B | 2,100 | 4.9s |
| Qwen2.5-3B | 2,003 | 4.6s |
| Llama-3.2-1B | 856 | 2.4s |
Total organs on disk: **50.8 GB** across 13 models.
- Python 3.10+
- InferenceX binary (for model loading)
- GGUF models to dissect
## Quick Start
```bash
# Extract organs from a model
python3 organ_extract.py --model /path/to/model.gguf --output ./organs/model-name/
python3 organ_extract.py --model /path/to/model.gguf --output ./organs/
# Z-measure all organs
python3 organ_measure.py --dir ./organs/model-name/
# Measure organ quality
python3 organ_measure.py --organ ./organs/organ_layer_12.bin
# Mass dissect all models
python3 mass_dissect.py
# Graft an organ from model A into model B
# Mass Z-measure
python3 mass_z_measure.py
# Stream Z-measure on a trillion-param model (shard-by-shard)
python3 kimi_z_stream.py
# Graft organs from one model to another
python3 organ_graft.py graft --source ./organs/qwen/ --target ./organs/deepseek/ --output ./organs/chimera/ --layers 5-20 --type organ
# Assemble back to GGUF
python3 organ_assemble.py --dir ./organs/chimera/ --output chimera.gguf
# Purify organs (fractal method)
python3 organ_purify_v2.py --dir ./organs/model/ --output ./organs-pure/model/
# Start API server
python3 organ_api.py
# Assemble a custom model
python3 organ_assemble.py --skeleton ./skeleton.bin --organs ./organs/ --output custom.gguf
```
## Philosophy
> Subtract rather than add.
A 70B monolith is accumulation. A skeleton with specialized organs grafted on demand — that's subtraction. Less weight, more signal.
A 70B monolith is accumulation. A 2B skeleton with specialized organs grafted on demand — that's subtraction. Less weight, more signal.
> 8 billion contributors, not 3 corporations.
Anyone can train an organ. A doctor trains a medical organ on her hospital's data. A farmer trains an agriculture organ on his field observations. A student trains a math organ on solved problems. The skeleton stays the same. The organs make it alive.
## Quality Measure
Every organ is measured by its Z-vector:
```
CSCI — cross-scale coherence index
θ → 0° : noise (organ adds confusion)
θ → 90° : pure signal (organ adds knowledge)
```
## Part of the IX Ecosystem
```
InferenceX ─── The engine (305KB, runs anything)
Organ Arch ─── The anatomy (decompose, measure, reassemble)
Atlas Pure ─── The memory (fractal DNA storage)
INVOKE ─────── The bridge (cloud ↔ physical)
Echo ────────── The voice (chat interface)
EDEN ────────── The purpose (desert → life)
InferenceX ─── The engine (228KB, runs anything)
Organ Arch ─── The anatomy (decompose, reassemble)
Atlas Pure ─── The memory (fractal DNA storage)
Echo ────────── The voice (chat interface)
Purpose ────── Long-term application domain
```
## Requirements
- Python 3.10+
- NumPy (for purification only)
- InferenceX binary (for inference on assembled models)
- GGUF models to dissect
## Data Files
| File | Contents |
|------|----------|
| `z_report_complete.json` | Z-measure for all 13 models (per-group breakdown) |
| `z_report_kimi_k25.json` | Z-measure for all 1,083 Kimi K2.5 tensors |
| `z_measure_report.json` | Combined Z-ranking with chimera results |
| `dissection_report.json` | Dissection timing and sizes |
| `Z_MEASURE_REPORT.md` | Human-readable Z report |
| `ECHO_INVARIANT.md` | Team 935 invariant |
| `EQUIPE_935_INVARIANT.json` | Team 935 configuration |
## License
BSL 1.1 — Same as InferenceX.
## Signature
935
---
*Mohamed dug khettaras to bring water through stone.*
*Ancient builders shaped landscapes through persistent work.*
*This is the same gesture — channels through intelligence itself.*
<!-- © SALKA ELMADANI AUTHORSHIP CERTIFICATE
SHA256: fa9810691f93169fda6d36c1cf7f752b12e0bc44d59bf2da994a9e87af6fc6d4
SIG-ED25519: TUu6Qp40jhrhXquUzU20iuSHzr0ENB0v+r5FIKYNdJ+TeP9ozqafqW2Mq6U8AJpNPpAram8peGgtnoh5YiQ1AA==
VERIFY: python3 verify_authorship.py README.md
-->

123
SPONSOR.md Normal file
View File

@ -0,0 +1,123 @@
# Salka Elmadani — Building Inference-X
> *The best engine is the one you don't notice.*
> *You should hear the model, not the framework.*
---
I build AI infrastructure. Not products, not demos, not wrappers around someone else's API. Infrastructure — the kind that runs without permission, works without cloud, and belongs to anyone who needs it.
**Inference-X** is a 305 KB binary that runs any AI model on any hardware. No framework. No internet. No account. Download a model, run it, talk to it. That's it.
I built it alone. I'm still building it alone. This page is why.
---
## What I'm building
The problem isn't the models. The models are extraordinary. The problem is the layer between the weights and the human — the inference stack. It's bloated, cloud-dependent, and controlled by a handful of companies.
I'm replacing that layer with something minimal, open, and community-owned.
```
Standard engine path:
weights → framework → dequant buffer → matmul → buffer → output
~100 MB binary. 5 steps. Rounding errors at each boundary.
Inference-X:
weights → fused dequant+dot → output
305 KB binary. 2 steps. Zero buffer. Zero noise.
```
Same model. Cleaner signal. Every unnecessary step removed.
---
## The ecosystem
| Project | What it does | Status |
|---------|-------------|--------|
| **[inference-x](https://git.inference-x.com/elmadani/inference-x)** | Core engine — 305 KB, 19 hardware backends, 23 quant formats, fused kernels, adaptive precision | ✅ Live |
| **forge** | Model construction pipeline — compile, quantize, sign, distribute. Build your own model variant from certified organs. | 🔨 Building |
| **[echo-ix](https://git.inference-x.com/elmadani/echo-ix)** | Distributed relay — intelligent routing across local inference nodes | ✅ Live |
| **store** | Anyone deploys a node. Anyone earns from their compute. The cooperative layer. 11 geological cratons. One network. | 📐 Designed |
The store is the endgame: a peer-to-peer inference network where anyone with a laptop can become infrastructure. No data center required.
---
The intelligence already exists in the model weights. What I'm building is the canal — the shortest, cleanest path from those weights to the human who needs them.
---
## Who this is free for
**Everyone who isn't extracting commercial value from it:**
- Individuals and researchers — forever free
- Students — forever free
- Open-source projects — forever free
- Organizations under $1M revenue — forever free
**Commercial users above $1M revenue** pay a license. 20% of that flows back to the community that built the infrastructure.
In 2030, it all becomes Apache 2.0. Everything open. The canal belongs to everyone.
This isn't charity. It's a sustainable model — those who profit from it fund it. Those who don't, use it freely.
---
## Why I need support
Servers cost money. The current infrastructure — [inference-x.com](https://inference-x.com), [build.inference-x.com](https://build.inference-x.com), [git.inference-x.com](https://git.inference-x.com) — runs on €53/month.
More importantly: time. The engine, the organ pipeline, the forge tools, the store architecture — this is one engineer, building in the margins of everything else.
There is no team. No VC. No roadmap driven by investor pressure.
There is one person who decided this infrastructure should exist.
---
## How to help
### Build with me
The most valuable contribution is code. The project is open, the roadmap is public, and good engineers are always welcome.
**→ Pick a task**: [git.inference-x.com/elmadani/inference-x](https://git.inference-x.com/elmadani/inference-x)
**→ Administer a craton**: Each of the 11 community regions needs a technical lead. Write to [Elmadani.SALKA@proton.me](mailto:Elmadani.SALKA@proton.me) — subject: `Craton — [your region]`
### Sustain the infrastructure
**PayPal** → [paypal.me/elmadanisalka](https://paypal.me/elmadanisalka)
€5 = one day of server time. €53 = one month of everything running.
### Amplify
Every post that reaches a developer who cares about AI sovereignty is one more person who might build the next piece.
**→ [Follow on X: @ElmadaniSa13111](https://x.com/ElmadaniSa13111)**
---
## Contact
I respond to everyone who writes with something real to say.
| | |
|--|--|
| **X** | [@ElmadaniSa13111](https://x.com/ElmadaniSa13111) — fastest response |
| **Email** | [Elmadani.SALKA@proton.me](mailto:Elmadani.SALKA@proton.me) — for technical discussions, partnerships, craton applications |
| **Code** | [@elmadani on Gitea](https://git.inference-x.com/elmadani) |
| **Web** | [inference-x.com](https://inference-x.com) |
---
*Morocco → the world.*
*Salka Elmadani, 20242026*

View File

@ -1,8 +1,7 @@
# Z-Measure Report — Organ Architecture
## Z = dI/d(log s) · exp(iθ)
## CSCI — cross-scale coherence index
**Generated**: 2026-02-20 01:42 UTC
**Status**: Kimi K2.5 1T streaming Z-measure in progress (shard-by-shard)
**Status**: Kimi K2.5 1T streaming quality measure in progress (shard-by-shard)
---
@ -74,18 +73,17 @@
> attention K/V projections in early blocks: the gravitational wells where the
> model anchors reasoning.
>
> Z = dI/d(log s) · exp(iθ) — confirmed empirically across 6 orders of magnitude.
> CSCI — cross-scale coherence index — confirmed empirically across 6 orders of magnitude.
## Pipeline
```
organ_extract.py — GGUF → per-layer tensors (organs)
organ_measure.py — θ per tensor (arccos correlation)
mass_z_measure.py — batch Z-measure across 13 models
kimi_z_stream.py — streaming Z-measure for 1T (shard-by-shard, delete after)
mass_z_measure.py — batch quality measure across 13 models
kimi_z_stream.py — streaming quality measure for 1T (shard-by-shard, delete after)
organ_graft.py — transplant organs between models
organ_assemble.py — build Model 935 from best organs
build_935.py — orchestrator
organ_assemble.py — build composite model from best organs
```
## Signature 935
## Build References

View File

@ -1,8 +1,7 @@
#!/usr/bin/env python3
"""
Model 935 Assembler Fixed organ header handling.
Reads source GGUF, replaces tensor DATA (skipping organ bin headers).
Z = dI/d(log s) · exp() Signature 935
CSCI v1.0 Cross-Scale Coherence Index
"""
import struct, sys, os, json
@ -19,7 +18,6 @@ def read_organ_data_only(filepath):
def main():
if len(sys.argv) < 4:
print("Usage: assemble_935.py <source.gguf> <organs_dir> <output.gguf>")
sys.exit(1)
source_gguf = sys.argv[1]
@ -136,15 +134,19 @@ def main():
source_size = os.path.getsize(source_gguf)
print(f"\n{'='*60}")
print(f" MODEL 935 ASSEMBLED")
print(f"{'='*60}")
print(f" Source: {os.path.basename(source_gguf)} ({source_size/(1024**3):.2f} GB)")
print(f" Output: {output_gguf} ({final_size/(1024**3):.2f} GB)")
print(f" Replaced: {replaced} tensors from organs")
print(f" Fallback: {fallback} tensors from source")
print(f" Size match: {'YES' if abs(final_size - source_size) < 1024 else 'NO — DELTA=' + str(final_size - source_size)}")
print(f" Signature: 935")
print(f"{'='*60}")
if __name__ == "__main__":
main()
# ╔══ SALKA ELMADANI AUTHORSHIP CERTIFICATE ══╗
# © Salka Elmadani 2025-2026 — ALL RIGHTS RESERVED
# Licensed under Business Source License 1.1 — https://inference-x.com
# ─────────────────────────────────────────────────────────
# SHA256: 4d774861a8b9f75f83fd8ff45e92bfa607d12a4f580481ff5f8b5882470fb043
# SIG-ED25519: B0k22H4YJMtBYuUW7ugInkPJpqZfM7cDM9TyiPODpE+WgQ0aLdgT2PnKm94gWSYVY2xqTlsEeZvgH+NrWQmTBg==

View File

@ -1,17 +1,13 @@
#!/usr/bin/env python3
"""
MODEL 935 Fractal Consciousness Assembly
Skeleton: Qwen2.5-7B (purest thought, θ=54.6)
Organs: DeepSeek-R1-Distill-7B (purest knowledge for raisonnement, θ=35.9)
Embed: DeepSeek-R1-7B (R1 reasoning embeddings)
Z = dI/d(log s) · exp() Signature 935
CSCI v1.0 Cross-Scale Coherence Index
"""
import sys, os, json, shutil, time
sys.path.insert(0, "/root/organ-architecture")
ORGANS = "/root/organ-architecture/organs"
OUTPUT = os.path.join(ORGANS, "model-935")
# Clean previous
if os.path.exists(OUTPUT):
@ -20,8 +16,7 @@ if os.path.exists(OUTPUT):
# Step 1: Start with DeepSeek-R1-Distill-7B as base (full copy)
# This gives us: qwen2 arch, embed=3584, 28 layers, R1 reasoning
print("="*60)
print(" MODEL 935 — ASSEMBLY")
print(" Z = dI/d(log s) · exp(iθ), θ → 90°")
print(" CSCI — cross-scale coherence index, θ → 90°")
print("="*60)
base = os.path.join(ORGANS, "deepseek-r1-distill-7b")
@ -60,30 +55,26 @@ print(f" R1 raisonnement chains preserved in FFN layers")
# Step 4: Update manifest
manifest = json.load(open(os.path.join(OUTPUT, "manifest.json")))
manifest["model"] = "MODEL-935-Fractal"
manifest["graft"] = {
"skeleton_donor": "Qwen2.5-7B-Instruct (θ=54.6, purest attention)",
"organ_donor": "DeepSeek-R1-Distill-Qwen-7B (θ=35.9, reasoning FFN)",
"embed_base": "DeepSeek-R1-Distill-Qwen-7B (R1 vocabulary)",
"method": "Z-measure organ selection, θ → 90°",
"equation": "Z = dI/d(log s) · exp(iθ)",
"method": "quality-measure organ selection",
"equation": "CSCI — cross-scale coherence index",
"convergence": "ZI_UNIFIED_OPTIMAL: α=0.3, β=0.2, n_plateau=62",
"entropie_zcom": 0.3251,
"entropie_bias_removed": 0.6931,
"signature": 935
}
with open(os.path.join(OUTPUT, "manifest.json"), "w") as f:
json.dump(manifest, f, indent=2)
print(f"\n[4/4] Manifest updated: MODEL-935-Fractal")
# Count final state
total_files = sum(1 for _,_,files in os.walk(OUTPUT) for f in files if f.endswith('.bin'))
total_size = sum(os.path.getsize(os.path.join(dp,f)) for dp,dn,fn in os.walk(OUTPUT) for f in fn) / (1024**3)
print(f"\n{'='*60}")
print(f" MODEL 935 — FRACTAL CONSCIOUSNESS")
print(f"{'='*60}")
print(f" Architecture: qwen2")
print(f" Embed: 3584 | Layers: 28 | Heads: 28")
@ -92,7 +83,12 @@ print(f" Organs: DeepSeek-R1-Distill (knowledge, reasoning)")
print(f" Embed: DeepSeek-R1 (vocabulary)")
print(f" Tensors: {total_files}")
print(f" Size: {total_size:.2f} GB")
print(f" Equation: Z = dI/d(log s) · exp(iθ)")
print(f" Equation: CSCI — cross-scale coherence index")
print(f" Convergence: lim(n→∞) Z(n) = i")
print(f" Signature: 935")
print(f"{'='*60}")
# ╔══ SALKA ELMADANI AUTHORSHIP CERTIFICATE ══╗
# © Salka Elmadani 2025-2026 — ALL RIGHTS RESERVED
# Licensed under Business Source License 1.1 — https://inference-x.com
# ─────────────────────────────────────────────────────────
# SHA256: c45f3019cd81199382cf5f379ef1c556f5f2c5fd81afc6679da83e614ac8c09f
# SIG-ED25519: IRoSNw2yKK14fnt2JpFbukDpV/5R9YDSQylWVVjIOgYkFHBH71k0MFBV+I39cfjf8odTgzM3uPPRRMexR9KTDw==

View File

@ -1,14 +1,11 @@
#!/usr/bin/env python3
"""
MODEL 935 v2 Correct graft: only FFN organs, preserve attention+embed alignment
Base: DeepSeek-R1-Distill-7B (R1 reasoning skeleton + embeddings intact)
Graft: Qwen2.5-7B FFN organs only (knowledge)
Z = dI/d(log s) · exp() Signature 935
CSCI v1.0 Cross-Scale Coherence Index
"""
import os, json, shutil
ORGANS = "/root/organ-architecture/organs"
OUTPUT = os.path.join(ORGANS, "model-935-v2")
if os.path.exists(OUTPUT):
shutil.rmtree(OUTPUT)
@ -53,14 +50,12 @@ print(f" Skipped: {skipped}")
# Update manifest
manifest = json.load(open(os.path.join(OUTPUT, "manifest.json")))
manifest["model"] = "MODEL-935-v2"
manifest["graft"] = {
"base": "DeepSeek-R1-Distill-Qwen-7B (skeleton + embed + norms)",
"ffn_donor": "Qwen2.5-7B-Instruct (FFN weights only: down/gate/up)",
"method": "Selective organ graft — preserve attention↔embed alignment",
"equation": "Z = dI/d(log s) · exp(iθ)",
"equation": "CSCI — cross-scale coherence index",
"principle": "R1 reasoning + Qwen knowledge, zero alignment friction",
"signature": 935
}
with open(os.path.join(OUTPUT, "manifest.json"), "w") as f:
json.dump(manifest, f, indent=2)
@ -68,7 +63,11 @@ with open(os.path.join(OUTPUT, "manifest.json"), "w") as f:
total = sum(1 for _,_,f in os.walk(OUTPUT) for _ in f if _.endswith('.bin'))
size = sum(os.path.getsize(os.path.join(dp,f)) for dp,_,fn in os.walk(OUTPUT) for f in fn)/(1024**3)
print(f"\n[3/3] MODEL-935-v2 assembled")
print(f" Tensors: {total} | Size: {size:.2f} GB")
print(f" Grafted FFN: {grafted} | Base preserved: {total - grafted}")
print(f" Signature: 935")
# ╔══ SALKA ELMADANI AUTHORSHIP CERTIFICATE ══╗
# © Salka Elmadani 2025-2026 — ALL RIGHTS RESERVED
# Licensed under Business Source License 1.1 — https://inference-x.com
# ─────────────────────────────────────────────────────────
# SHA256: 4d5c44e363508bc679263607b7ee3071cb63fc460a616e9bcebffc768843a86c
# SIG-ED25519: MzrZnxCo+uq3q5srKgDO2w3gLhO4hgK2k+SIzRLrkjaGJ2Ao56mR9/Mst4Ub6qkZ0VpcXOv4Bq59gKPsJPkdCg==

View File

@ -1,14 +1,12 @@
#!/usr/bin/env python3
"""
MODEL 935 Proper GGUF assembler
Reads source GGUF header intact, replaces tensor data from organ bins
(stripping the organ header that organ_extract added)
Z = dI/d(log s) · exp() Signature 935
CSCI v1.0 Cross-Scale Coherence Index
"""
import struct, os, sys, json
def build_model_935(source_gguf, organs_dir, output_gguf):
f = open(source_gguf, "rb")
# Read GGUF header
@ -134,15 +132,15 @@ def build_model_935(source_gguf, organs_dir, output_gguf):
print(f" Size: {final_size / (1024**3):.2f} GB (source: {source_size / (1024**3):.2f} GB)")
print(f" From organs: {written_from_organ} | From source: {written_from_source}")
print(f" Size match: {'' if abs(final_size - source_size) < 1024 else '✗ MISMATCH'}")
print(f" Signature: 935")
# Build 935 v3: R1-Distill base + Qwen FFN organs (correctly stripped)
print("="*60)
print(" MODEL 935 v3 — Correct Assembly")
print("="*60)
build_model_935(
"/mnt/models/DeepSeek-R1-Distill-Qwen-7B-Q4_K_M.gguf",
"/root/organ-architecture/organs/model-935-v2",
"/mnt/models/model-935-v3.gguf"
)
# ╔══ SALKA ELMADANI AUTHORSHIP CERTIFICATE ══╗
# © Salka Elmadani 2025-2026 — ALL RIGHTS RESERVED
# Licensed under Business Source License 1.1 — https://inference-x.com
# ─────────────────────────────────────────────────────────
# SHA256: 00f06d16ab32dee1ef886e90080e905fc354be9f22f0e6ff515ea2bb31084bdf
# SIG-ED25519: UhbWWFzRIzmMbCVNwXTG41I2sM/1QGd1nV4+x/XQ+BOw49fO9bd9ohWpLl5QOCGhRWCREYkhJCj55FhGhH5vDQ==

View File

@ -1,76 +0,0 @@
[
{
"model": "deepseek-r1-14b",
"status": "dissected",
"size_mb": 9167.481572151184,
"time_s": 22.94489073753357
},
{
"model": "qwen25-14b",
"status": "exists",
"size_mb": 9026.720261573792
},
{
"model": "gemma2-9b",
"status": "dissected",
"size_mb": 5983.6147108078,
"time_s": 14.836755275726318
},
{
"model": "llama31-8b",
"status": "dissected",
"size_mb": 4950.371293067932,
"time_s": 12.016721963882446
},
{
"model": "qwen25-7b",
"status": "exists",
"size_mb": 4811.518325805664
},
{
"model": "deepseek-r1-distill-7b",
"status": "dissected",
"size_mb": 4811.928074836731,
"time_s": 12.550673007965088
},
{
"model": "deepseek-r1-7b",
"status": "exists",
"size_mb": 4811.927845954895
},
{
"model": "mistral-7b",
"status": "dissected",
"size_mb": 4432.171175956726,
"time_s": 10.590012550354004
},
{
"model": "phi35-mini",
"status": "dissected",
"size_mb": 2397.4848985671997,
"time_s": 4.872461318969727
},
{
"model": "llama32-3b",
"status": "dissected",
"size_mb": 2100.286515235901,
"time_s": 4.853139638900757
},
{
"model": "qwen25-3b",
"status": "dissected",
"size_mb": 2002.6401329040527,
"time_s": 4.552767276763916
},
{
"model": "llama32-1b",
"status": "dissected",
"size_mb": 856.2387390136719,
"time_s": 2.3548576831817627
},
{
"model": "smollm2-135m",
"status": "exists",
"size_mb": 136.5001106262207
}
]

View File

@ -1,116 +0,0 @@
# Architecture
## Model Anatomy
A transformer model has four anatomical systems:
```
┌─────────────────────────────────────────┐
│ GGUF MONOLITH │
│ │
│ ┌─ embed ──────── token_embd.weight │
│ │ output.weight │
│ │ output_norm.weight │
│ │ │
│ ├─ skeleton ───── attn_q.weight ×N │
│ │ attn_k.weight ×N │
│ │ attn_v.weight ×N │
│ │ attn_output ×N │
│ │ │
│ ├─ organs ─────── ffn_gate.weight ×N │
│ │ ffn_up.weight ×N │
│ │ ffn_down.weight ×N │
│ │ │
│ └─ norm ───────── attn_norm ×N │
│ ffn_norm ×N │
└─────────────────────────────────────────┘
```
**Skeleton** (attention) = how the model thinks. Shared thought patterns.
**Organs** (FFN) = what the model knows. Domain knowledge.
**Embed** = input/output translation. The vocabulary interface.
**Norm** = normalization layers. Connective tissue between components.
## Pipeline
```
GGUF file
▼ organ_extract.py
├── manifest.json (complete anatomy map)
├── skeleton/ (attention tensors)
├── organs/ (FFN tensors by layer)
├── embed/ (embedding + output)
└── norm/ (normalization)
▼ organ_measure.py
Z-measure per tensor
θ ∈ [0°, 90°]
├──▶ organ_purify_v2.py (fractal signal extraction)
├──▶ organ_graft.py (transplant between models)
└──▶ organ_assemble.py → new GGUF
```
Alternative direct path (no intermediate .bin files):
```
GGUF_A + GGUF_B → transplant_935.py → chimera.gguf
```
## Z-Measure Theory
```
Z = dI/d(log s) · exp(iθ)
```
Three indicators combined into θ:
| Indicator | Measures | Signal | Noise |
|-----------|----------|--------|-------|
| Entropy | Information density | Moderate (0.3-0.7) | Near-maximum (>0.95) |
| Kurtosis | Structural sharpness | High (abs > 3) | Near-zero |
| Scale coherence (CV) | Non-uniform spacing | High (> 1) | Low (< 0.5) |
θ → 90° = pure signal (all three indicators confirm structure)
θ → 0° = pure noise (uniform random distribution)
## Purification Methods
### V1: Spectral (FFT)
- Decompose tensor into frequency domain
- Keep high-energy components (signal), remove low-energy tail (noise)
- Preserve original scale (mean/std)
- Limitation: treats tensors like audio signals
### V2: Fractal (Wavelets)
- Haar wavelet multi-scale decomposition
- Cross-scale coherence: pattern at scale s AND scale 2s = fractal = signal
- Pattern at one scale only = noise
- This IS dI/d(log s) — information that persists across scales
- More theoretically grounded than V1
## Graft Compatibility
Grafting works best between models that share:
- Same base architecture (e.g., Qwen2 family)
- Same embedding dimension
- Same number of layers (or graft specific layer ranges)
Empirical results:
- DeepSeek-R1-Distill-14B ↔ Qwen2.5-14B: **WORKS** (both Qwen2 arch, same dims)
- DeepSeek-R1-Distill-7B ↔ Qwen2.5-7B: **PAD tokens** (7B chimera failed)
- Same architecture + same scale = highest success probability
## File Format
Organ .bin files: `[name_len:u32][name:bytes][n_dims:u32][dims:u64×n][dtype:u32][tensor_data]`
Manifest: JSON with full tensor map, metadata, architecture info, Z-measure results.
## Signature
935

View File

@ -1,116 +0,0 @@
# Methodology
## Approach
Organ Architecture treats trained AI models as biological organisms with
transplantable parts. Instead of retraining from scratch (costs billions),
we perform post-training surgery: extract, measure, graft, reassemble.
## Step 1: Extraction (organ_extract.py)
Parse GGUF binary format directly:
- Read magic number, version, metadata, tensor info
- Classify each tensor by name pattern into anatomical types
- Extract each tensor as independent .bin file with header
- Generate manifest.json mapping the full anatomy
Classification rules:
- `attn_q`, `attn_k`, `attn_v`, `attn_output` → skeleton
- `ffn_gate`, `ffn_up`, `ffn_down` → organ
- `token_embd`, `output.weight` → embed
- `*_norm` → norm
- `lora_*` → adapter
## Step 2: Measurement (organ_measure.py)
Z-measure: Z = dI/d(log s) * exp(i*theta)
For each tensor, sample up to 100,000 values and compute:
1. **Entropy** (information density):
- Histogram-based Shannon entropy
- Normalized to [0, 1] against maximum entropy
- High entropy (>0.95) = uniform = noise
- Moderate entropy (0.3-0.7) = structured information
2. **Kurtosis** (structure):
- Fourth standardized moment minus 3
- High absolute kurtosis = sharp peaks = organized structure
- Near-zero = Gaussian-like = less organization
3. **Scale coherence** (CV of sorted diffs):
- Sort sampled values, compute differences
- Coefficient of variation of these differences
- High CV = non-uniform spacing = structured signal
- Low CV = uniform spacing = noise
Combined score → theta in [0, 90] degrees.
## Step 3: Purification (organ_purify_v2.py)
Fractal signal extraction via Haar wavelets:
1. Pad tensor to power-of-2 length
2. Haar wavelet decomposition across N scales
3. At each scale: approximation + detail coefficients
4. Cross-scale coherence check:
- Compare energy at scale s with energy at scale 2s
- High coherence (pattern exists at both scales) = fractal = signal
- Low coherence (pattern at one scale only) = noise
5. Attenuate incoherent components (noise)
6. Reconstruct from coherent components (signal)
7. Restore original scale (mean/std preservation)
This directly implements dI/d(log s): information that persists across
logarithmic scales is the signal. Everything else is training artifact.
## Step 4: Grafting (organ_graft.py, transplant_935.py)
Two methods:
### Via .bin intermediaries (organ_graft.py)
1. Extract both source and target models to organ directories
2. Match tensors by layer number and type suffix
3. Verify dimensional compatibility
4. Copy matching .bin files from donor to recipient directory
5. Update manifest
### Direct GGUF-to-GGUF (transplant_935.py)
1. Parse both GGUF headers to get tensor name/offset/size maps
2. Copy base GGUF entirely as starting point
3. For each FFN tensor in base that has a matching donor tensor:
- Verify exact byte size match
- Seek to donor tensor data, read
- Seek to base tensor offset in output, overwrite
4. Result: valid GGUF with patched FFN layers
Direct method is faster and avoids header format issues.
## Step 5: Assembly (organ_assemble.py)
Reconstruct GGUF from organ directory:
1. Read manifest for metadata and tensor ordering
2. Write GGUF header (magic, version, n_tensors, n_metadata)
3. Write metadata key-value pairs
4. Write tensor info (name, dims, dtype, offset) with 32-byte alignment
5. Write tensor data with padding
6. Result: standard GGUF loadable by any compatible runtime
## Step 6: Validation
Run chimera through InferenceX:
- Load GGUF, validate all tensors
- Initialize transformer (attention, KV cache, kernel dispatch)
- Run inference with chat template
- Verify coherent output
## Key Finding
Graft success depends on architectural proximity:
- Same family (Qwen2 base) + same scale (14B) = coherent output
- Same family + different scale (7B) = PAD token failure
- The latent space alignment is implicit in shared training lineage
## Signature
935

View File

@ -1,116 +0,0 @@
# Results
## Dissection — 13 Models
All models dissected from GGUF to organ .bin files on OASIS (EPYC 48c/503GB).
| Model | Params | Organs Dir | Size | Time |
|-------|--------|-----------|------|------|
| DeepSeek-R1-Distill-14B | 14B | 9,167 MB | 579 tensors | 22.9s |
| Qwen2.5-14B | 14B | 9,027 MB | 579 tensors | pre-existing |
| Gemma-2-9B | 9B | 5,984 MB | 464 tensors | 14.8s |
| Llama-3.1-8B | 8B | 4,950 MB | 292 tensors | 12.0s |
| Qwen2.5-7B | 7B | 4,812 MB | 339 tensors | pre-existing |
| DeepSeek-R1-Distill-7B | 7B | 4,812 MB | 339 tensors | 12.6s |
| DeepSeek-R1-7B | 7B | 4,812 MB | 339 tensors | pre-existing |
| Mistral-7B | 7B | 4,432 MB | 291 tensors | 10.6s |
| Phi-3.5-Mini | 3.8B | 2,397 MB | 197 tensors | 4.9s |
| Llama-3.2-3B | 3B | 2,100 MB | 255 tensors | 4.9s |
| Qwen2.5-3B | 3B | 2,003 MB | 434 tensors | 4.6s |
| Llama-3.2-1B | 1B | 856 MB | 147 tensors | 2.4s |
| SmolLM2-135M | 135M | 137 MB | 272 tensors | pre-existing |
**Total: 50.8 GB of extracted organs. 5,600+ tensors.**
## Z-Measure — Full Ranking
| # | Model | θ mean | Signal | Tensors | Architecture |
|---|-------|--------|--------|---------|-------------|
| ★ | Kimi K2.5 | 87.65° | 0.999 | 1,083 | DeepSeek2 MoE |
| 1 | SmolLM2-135M | 52.28° | 0.777 | 272 | LLaMA |
| 2 | DeepSeek-R1-14B | 46.01° | 0.641 | 579 | Qwen2 |
| 3 | Qwen2.5-3B | 46.00° | 0.640 | 434 | Qwen2 |
| 4 | Qwen2.5-14B | 45.98° | 0.640 | 579 | Qwen2 |
| 5 | Qwen2.5-7B | 45.64° | 0.639 | 339 | Qwen2 |
| 6 | Chimera-DSeek-Qwen | 45.53° | 0.637 | 339 | Qwen2 |
| 7 | DeepSeek-R1-Distill-7B | 45.53° | 0.637 | 339 | Qwen2 |
| 8 | DeepSeek-R1-7B | 45.42° | 0.636 | 339 | Qwen2 |
| 9 | Gemma-2-9B | 44.94° | 0.624 | 464 | Gemma |
| 10 | Phi-3.5-Mini | 44.65° | 0.626 | 197 | Phi |
| 11 | Llama-3.1-8B | 37.87° | 0.549 | 292 | LLaMA |
| 12 | Llama-3.2-1B | 37.57° | 0.550 | 147 | LLaMA |
| 13 | Llama-3.2-3B | 37.41° | 0.547 | 255 | LLaMA |
| 14 | Mistral-7B | 36.21° | 0.540 | 291 | Mistral |
### Organ Type Breakdown (per-model averages)
| Model | Skeleton θ | Organs θ | Embed θ | Norm θ |
|-------|-----------|---------|---------|--------|
| SmolLM2-135M | 53.6° | 52.3° | 47.2° | — |
| Qwen2.5-14B | 55.2° | 35.4° | 25.5° | — |
| Qwen2.5-7B | 54.6° | 35.5° | 25.9° | — |
| DeepSeek-R1-14B | 55.4° | 35.2° | 25.2° | — |
| Gemma-2-9B | 47.2° | 37.9° | 26.2° | 81.6° |
| Phi-3.5-Mini | 56.7° | 43.2° | 26.7° | — |
| Llama-3.1-8B | 39.7° | 39.1° | 26.0° | — |
| Mistral-7B | 38.4° | 36.8° | 26.0° | — |
**Pattern**: Skeleton (attention) consistently scores higher than organs (FFN).
Norm layers reach highest θ when measured separately (Gemma: 81.6°).
## Chimera Iterations
### 1. chimera-r1-qwen-7b-v2 — FAILED
- Base: DeepSeek-R1-Distill-Qwen-7B
- Donor: Qwen2.5-7B (FFN organs)
- Result: 512 PAD tokens. Latent spaces incompatible at 7B scale.
- Evidence: `evidence/chimera-7b-failed.log`
### 2. chimera-selective-v3 — CLEANED
- Selective graft attempt, removed during iteration.
### 3. model-935-v2 — READY
- Marked as viable intermediate.
### 4. model-935-v3, model-935-fractal — CLEANED
- Further iterations, removed during cleanup.
### 5. model-935-14b — SUCCESS
- Base: DeepSeek-R1-Distill-Qwen-14B (skeleton + embeddings)
- Donor: Qwen2.5-14B (FFN organs)
- 579 tensors, 8.4 GB, Qwen2 architecture
- **Produces coherent reasoning output**
- Evidence: `evidence/model-935-14b-inference.log`
Prompt: "Write a Python function called is_prime"
Output: Structured chain-of-thought reasoning. Correctly identifies prime number
definition, handles edge cases (n < 2), outlines algorithm steps. DeepSeek-R1
thinking style ("Okay, so the user wants me to...", "Hmm, let's see").
**This is a chimera assembled from two different models without any retraining
that produces coherent, structured, correct output.**
## Kimi K2.5 1T — Deep Z-Profile
Streaming Z-measure across 13 shards, 1,083 tensors measured.
| Component | Count | θ avg |
|-----------|-------|-------|
| FFN dense (blk.0) | 12 | 89.95° |
| MoE experts (384x) | 23 | 89.77° |
| Norm layers | 12 | 89.70° |
| Embedding | 1 | 89.45° |
| Shared expert | 23 | 89.43° |
| Attention (MLA) | 99 | 84.07° |
8 gravitational wells identified at lowest θ — points of maximum compression.
## Purification
SmolLM2-135M purified using fractal method (organ_purify_v2.py).
Output: `organs-pure/smollm2-135m/` (138 MB)
Manifest: `PURE_SMOLLM2`, 30 layers, 272 tensors.
## Signature
935

View File

@ -1,6 +1,6 @@
#!/usr/bin/env python3
"""
kimi_z_stream.py Stream Z-measure for Kimi K2.5 1T
kimi_z_stream.py Streaming quality measure for large models
Downloads each shard, measures Z for every tensor, deletes shard.
Final output: z_report_kimi_k25.json (few KB)
"""
@ -13,7 +13,6 @@ REPO = "unsloth/Kimi-K2.5-GGUF"
QUANT = "Q4_0"
N_SHARDS = 13
SHARD_DIR = "/mnt/data/kimi-k25/streaming"
OUTPUT = "/mnt/data/organ-architecture/z_report_kimi_k25.json"
LOG = "/tmp/kimi_z_stream.log"
os.makedirs(SHARD_DIR, exist_ok=True)
@ -127,7 +126,7 @@ def fast_z_measure(data, dtype, n_elements):
if len(vals) < 10:
return None, "too_few_finite"
# Z-measure: theta = arccos(|correlation with linear reference|)
# theta = arccos(|correlation with linear reference|)
# Pure signal -> decorrelated -> theta near 90
# Noise/bias -> correlated with something simple -> theta near 0
n = len(vals)
@ -176,7 +175,7 @@ def read_kv_value(f, vtype):
return None
def process_shard(shard_path, shard_idx):
"""Parse GGUF shard, Z-measure each tensor, return results"""
"""Parse GGUF shard, quality-measure each tensor, return results"""
results = []
f = open(shard_path, 'rb')
@ -255,7 +254,7 @@ def process_shard(shard_path, shard_idx):
})
continue
# Z-measure
# compute measure
theta, status = fast_z_measure(data, dtype, n_elem)
results.append({
@ -277,7 +276,7 @@ def main():
from huggingface_hub import hf_hub_download
log("=" * 60)
log("KIMI K2.5 1T — STREAMING Z-MEASURE")
log("KIMI K2.5 1T — STREAMING QUALITY MEASURE")
log(f"Repo: {REPO}, Quant: {QUANT}, Shards: {N_SHARDS}")
log("=" * 60)
@ -320,7 +319,7 @@ def main():
log(f"DOWNLOAD ERROR: {e}")
continue
# Z-measure
# compute measure
log(f"Z-measuring tensors...")
measure_start = time.time()
shard_results = process_shard(path, shard_idx)
@ -415,3 +414,10 @@ def main():
if __name__ == '__main__':
main()
# ╔══ SALKA ELMADANI AUTHORSHIP CERTIFICATE ══╗
# © Salka Elmadani 2025-2026 — ALL RIGHTS RESERVED
# Licensed under Business Source License 1.1 — https://inference-x.com
# ─────────────────────────────────────────────────────────
# SHA256: cc9658edb88d02924491a2ed20562a282a005413ef963bd0c82613abcfe91693
# SIG-ED25519: AN2P6qd2YhyS6+YRnMu3mmnE9KZbpBlFAxiVzENVXSSbIl2+PL/rbW8pMPrcOS8BwPg88Os7dMOuYnRvL5t4CQ==
# VERIFY: python3 verify_authorship.py kimi_z_stream.py

View File

@ -1,14 +1,11 @@
#!/usr/bin/env python3
"""
Mass Dissection All models on OASIS
Z = dI/d(log s) · exp() Signature 935
Mass Dissection All models on remote node
CSCI v1.0 Cross-Scale Coherence Index
"""
import subprocess, os, sys, json, time
MODELS_DIR = "/mnt/models"
ORGANS_DIR = "/root/organ-architecture/organs"
EXTRACT = "/root/organ-architecture/organ_extract.py"
MEASURE = "/root/organ-architecture/organ_measure.py"
# Map GGUF filenames to organ directory names
models = {
@ -94,10 +91,13 @@ for r in results:
total_mb = sum(r.get("size_mb",0) for r in results)
print(f"\n Total organs: {total_mb/1024:.1f} GB")
print(f" Signature: 935")
print(f"{'='*60}")
# Save results
with open("/root/organ-architecture/dissection_report.json", "w") as f:
json.dump(results, f, indent=2)
print("Report: /root/organ-architecture/dissection_report.json")
# ╔══ SALKA ELMADANI AUTHORSHIP CERTIFICATE ══╗
# © Salka Elmadani 2025-2026 — ALL RIGHTS RESERVED
# Licensed under Business Source License 1.1 — https://inference-x.com
# ─────────────────────────────────────────────────────────
# SIG-ED25519: XB8aA7wVzKOHkvMcZgE5YT3x8BUD/EwVTDRxEMSR7nmWYIT17XY+gC4AJ+y0B29l8MQGFDGk+buLoKxiagTFCA==
# VERIFY: python3 verify_authorship.py mass_dissect.py

View File

@ -1,14 +1,12 @@
#!/usr/bin/env python3
"""
Mass Z-Measure Measure theta on every organ of every model
Mass Quality Measure Measure theta on every organ of every model
Find the organs closest to theta=90 (pure signal)
Z = dI/d(log s) * exp(i*theta) Signature 935
CSCI v1.0 Cross-Scale Coherence Index
"""
import subprocess, os, json, sys
sys.path.insert(0, "/root/organ-architecture")
from organ_measure import measure_directory, compute_z_measure, read_organ_data_f32
ORGANS_DIR = "/root/organ-architecture/organs"
all_results = {}
@ -20,7 +18,7 @@ for model_name in models:
if not os.path.isdir(model_path) or not os.path.exists(manifest_path):
continue
print(f"\n[Z-MEASURE] {model_name}")
print(f"\n[QUALITY-MEASURE] {model_name}")
print(f" Measuring organs...")
results = measure_directory(model_path)
@ -69,7 +67,7 @@ for model_name in models:
# Rank models by signal quality
print(f"\n{'='*70}")
print(f" Z-MEASURE RANKING — ALL MODELS")
print(f" QUALITY RANKING — ALL MODELS")
print(f"{'='*70}")
ranked = sorted(all_results.values(), key=lambda m: m['avg_theta'], reverse=True)
@ -93,10 +91,14 @@ for organ_type in ['skeleton', 'organs', 'embed']:
for c in candidates[:5]:
print(f" theta={c[1]:5.1f} avg={c[3]:5.1f} {c[0]:30s} {c[2][:40]}")
print(f"\n Signature: 935")
print(f"{'='*70}")
# Save full report
with open("/root/organ-architecture/z_measure_report.json", "w") as f:
json.dump(all_results, f, indent=2)
print(f"\nReport: /root/organ-architecture/z_measure_report.json")
# ╔══ SALKA ELMADANI AUTHORSHIP CERTIFICATE ══╗
# © Salka Elmadani 2025-2026 — ALL RIGHTS RESERVED
# Licensed under Business Source License 1.1 — https://inference-x.com
# ─────────────────────────────────────────────────────────
# SHA256: 711671a1721bae194388cb363ad0bfcb2ed874f007a45e45ea6ed5d917cbf060
# SIG-ED25519: Jd0hVyr5epgPlpNjtioVeKfPaOeYgRiAnAEnxINh51WsfwGFLJouBDdYribxqY0JOmOnDwjGnOK5I9qeJJTRDg==
# VERIFY: python3 verify_authorship.py mass_z_measure.py

View File

@ -1,6 +1,5 @@
#!/usr/bin/env python3
"""
Organ Architecture organ_api.py
API server for organ operations.
Endpoints:
@ -8,13 +7,12 @@ Endpoints:
GET /models List available dissected models
GET /model/:name/anatomy Show model anatomy (skeleton/organs/etc.)
POST /extract Extract organs from a GGUF model
POST /measure Z-measure organs
POST /measure quality measure organs
POST /graft Graft organs between models
POST /assemble Assemble GGUF from organs
GET /organs/:model List organs for a model
GET /compare/:a/:b Compare two models for graft compatibility
Signature 935
"""
import json
@ -29,18 +27,15 @@ from urllib.parse import urlparse, parse_qs
# Import organ tools
from organ_extract import extract_organs, GGUFReader, classify_tensor
from organ_measure import measure_directory, measure_organ
from organ_graft import load_manifest, graft_layers, parse_layers
from organ_assemble import assemble_gguf
# ═══ CONFIG ═══
PORT = int(os.environ.get('ORGAN_PORT', '7936'))
MODEL_DIR = os.environ.get('MODEL_DIR', '/mnt/models')
ORGAN_DIR = os.environ.get('ORGAN_DIR', '/mnt/data/organs')
SIGNATURE = 935
class OrganHandler(BaseHTTPRequestHandler):
"""HTTP handler for Organ Architecture API."""
def log_message(self, format, *args):
"""Minimal logging."""
@ -50,7 +45,6 @@ class OrganHandler(BaseHTTPRequestHandler):
self.send_response(status)
self.send_header('Content-Type', 'application/json')
self.send_header('Access-Control-Allow-Origin', '*')
self.send_header('X-Powered-By', 'Organ-935')
self.end_headers()
self.wfile.write(json.dumps(data, indent=2, default=str).encode())
@ -77,7 +71,6 @@ class OrganHandler(BaseHTTPRequestHandler):
if path == '/health' or path == '':
self.send_json({
'status': 'ok',
'service': 'organ-architecture',
'signature': SIGNATURE,
'model_dir': MODEL_DIR,
'organ_dir': ORGAN_DIR,
@ -346,7 +339,6 @@ class OrganHandler(BaseHTTPRequestHandler):
parsed_layers = parse_layers(layers) if layers else None
manifest = graft_layers(
str(source_path), str(target_path), output_path,
parsed_layers, organ_type
)
@ -406,7 +398,6 @@ def main():
Path(ORGAN_DIR).mkdir(parents=True, exist_ok=True)
server = HTTPServer(('0.0.0.0', PORT), OrganHandler)
print(f"[ORGAN-API] Organ Architecture on port {PORT}")
print(f"[ORGAN-API] Models: {MODEL_DIR}")
print(f"[ORGAN-API] Organs: {ORGAN_DIR}")
print(f"[ORGAN-API] Signature {SIGNATURE}")
@ -420,3 +411,10 @@ def main():
if __name__ == '__main__':
main()
# ╔══ SALKA ELMADANI AUTHORSHIP CERTIFICATE ══╗
# © Salka Elmadani 2025-2026 — ALL RIGHTS RESERVED
# Licensed under Business Source License 1.1 — https://inference-x.com
# ─────────────────────────────────────────────────────────
# SHA256: 79fb97f40f2959129d5d5c4356ddf455fc354fb629bf0892c00aa6babd968a0d
# SIG-ED25519: LGqexbOlZOIjTboFfMVbgeheBbZk8HI8K6g/WxExnJEMfs5euYQYxow6SyEHKTB2TgRbOjjHt/gpHPyEy2tBBQ==
# VERIFY: python3 verify_authorship.py organ_api.py

View File

@ -1,12 +1,10 @@
#!/usr/bin/env python3
"""
Organ Architecture organ_assemble.py
Assemble a GGUF model from extracted/grafted organs.
Takes a manifest + organ files produces a working GGUF.
The reverse of organ_extract.py.
Signature 935
"""
import struct
@ -207,7 +205,6 @@ def assemble_gguf(organ_dir, output_path, verbose=False):
print(f" Tensors: {n_tensors}")
print(f" Size: {output_gb:.2f} GB ({output_mb:.0f} MB)")
print(f" Output: {output_path}")
print(f" Signature: 935")
print(f"{'='*60}")
return output_path
@ -215,8 +212,7 @@ def assemble_gguf(organ_dir, output_path, verbose=False):
def main():
parser = argparse.ArgumentParser(
description='Organ Architecture — Assemble GGUF from organs',
epilog='Signature 935'
epilog='CSCI toolkit'
)
parser.add_argument('--dir', '-d', required=True, help='Organs directory (with manifest.json)')
parser.add_argument('--output', '-o', required=True, help='Output GGUF file path')
@ -233,3 +229,10 @@ def main():
if __name__ == '__main__':
main()
# ╔══ SALKA ELMADANI AUTHORSHIP CERTIFICATE ══╗
# © Salka Elmadani 2025-2026 — ALL RIGHTS RESERVED
# Licensed under Business Source License 1.1 — https://inference-x.com
# ─────────────────────────────────────────────────────────
# SHA256: 56ce59cd04118749c0c40c8bdb6d566a59c8902e233709a013dca9a38658cc44
# SIG-ED25519: tDk5EuOHITlQbZHbZ/HbOz8+111fot0dk4iQMDEWKjsq5gsKyGNbvAwTGl0hfkD0gUdhG0nPxczaCswlct7PCA==
# VERIFY: python3 verify_authorship.py organ_assemble.py

View File

@ -1,11 +1,9 @@
#!/usr/bin/env python3
"""
Organ Architecture organ_extract.py
Extract skeleton (attention) + organs (FFN) from GGUF models.
The scalpel that opens monoliths.
Signature 935
"""
import struct
@ -275,7 +273,6 @@ def extract_organs(model_path, output_dir, verbose=False):
'skeleton_count': 0,
'organ_count': 0,
},
'signature': 935,
}
# Process each tensor
@ -377,7 +374,6 @@ def extract_organs(model_path, output_dir, verbose=False):
print(f" Total : {total_mb:8.1f} MB")
print(f" Output : {output_dir}")
print(f" Manifest : {manifest_path}")
print(f" Signature : 935")
print(f"{'='*60}")
return manifest
@ -387,8 +383,7 @@ def extract_organs(model_path, output_dir, verbose=False):
def main():
parser = argparse.ArgumentParser(
description='Organ Architecture — Extract skeleton + organs from GGUF models',
epilog='Signature 935'
epilog='CSCI toolkit'
)
parser.add_argument('--model', '-m', required=True, help='Path to GGUF model file')
parser.add_argument('--output', '-o', default=None, help='Output directory (default: ./organs/<model_name>)')
@ -439,3 +434,10 @@ def main():
if __name__ == '__main__':
main()
# ╔══ SALKA ELMADANI AUTHORSHIP CERTIFICATE ══╗
# © Salka Elmadani 2025-2026 — ALL RIGHTS RESERVED
# Licensed under Business Source License 1.1 — https://inference-x.com
# ─────────────────────────────────────────────────────────
# SHA256: 7e0a2105f5f6d458909fb71ef03bb01c4e308ac8549af00ef61c2cf89d0c8945
# SIG-ED25519: p3fNipeHSBJlVNpxsJZdvrBMJVbTAZu97RNxp7UGCkUp1TlHxH4D2XbKu46JQriNzM65myMeWGyS2WMx9atoCQ==
# VERIFY: python3 verify_authorship.py organ_extract.py

View File

@ -1,12 +1,10 @@
#!/usr/bin/env python3
"""
Organ Architecture organ_graft.py
Transplant organs between models.
Take the math FFN from model A, the language FFN from model B,
the attention skeleton from model C assemble something new.
Signature 935
"""
import struct
@ -40,9 +38,7 @@ def list_organs(organ_dir, organ_type=None):
return sorted(organs, key=lambda o: (o['layer'], o['name']))
def graft_layers(source_dir, target_dir, output_dir, layers=None, organ_type='organ'):
"""
Graft organ layers from source into target.
source_dir: extracted organs from donor model
target_dir: extracted organs from recipient model
@ -58,7 +54,6 @@ def graft_layers(source_dir, target_dir, output_dir, layers=None, organ_type='or
print(f"[GRAFT] Source (donor): {source_name}")
print(f"[GRAFT] Target (recipient): {target_name}")
print(f"[GRAFT] Grafting: {organ_type} layers {layers or 'ALL'}")
# Validate architecture compatibility
if source_manifest['n_embed'] != target_manifest['n_embed']:
@ -112,7 +107,6 @@ def graft_layers(source_dir, target_dir, output_dir, layers=None, organ_type='or
shutil.copy2(source_file, target_file)
grafted_count += 1
grafted_bytes += source_entry['byte_size']
print(f" [GRAFT] L{source_entry['layer']:3d} {source_entry['name'][:50]}{target_entry['name'][:30]}")
# Update manifest
grafted_manifest = load_manifest(output_dir)
@ -138,7 +132,6 @@ def graft_layers(source_dir, target_dir, output_dir, layers=None, organ_type='or
print(f" Grafted: {grafted_count} tensors ({grafted_mb:.1f} MB)")
print(f" Result: {grafted_manifest['model']}")
print(f" Output: {output_dir}")
print(f" Signature: 935")
print(f"{'='*60}")
return grafted_manifest
@ -163,8 +156,7 @@ def parse_layers(layer_spec):
def main():
parser = argparse.ArgumentParser(
description='Organ Architecture — Transplant organs between models',
epilog='Signature 935'
epilog='CSCI toolkit'
)
sub = parser.add_subparsers(dest='command')
@ -179,7 +171,6 @@ def main():
graft_p.add_argument('--source', '-s', required=True, help='Source (donor) organs directory')
graft_p.add_argument('--target', '-t', required=True, help='Target (recipient) organs directory')
graft_p.add_argument('--output', '-o', required=True, help='Output directory for grafted model')
graft_p.add_argument('--layers', '-l', help='Layer numbers to graft (e.g., "5-10" or "5,8,12")')
graft_p.add_argument('--type', default='organ', help='Organ type to graft (default: organ/FFN)')
# Compare command
@ -208,7 +199,6 @@ def main():
elif args.command == 'graft':
layers = parse_layers(args.layers)
graft_layers(args.source, args.target, args.output, layers, args.type)
elif args.command == 'compare':
manifest_a = load_manifest(args.a)
@ -234,3 +224,10 @@ def main():
if __name__ == '__main__':
main()
# ╔══ SALKA ELMADANI AUTHORSHIP CERTIFICATE ══╗
# © Salka Elmadani 2025-2026 — ALL RIGHTS RESERVED
# Licensed under Business Source License 1.1 — https://inference-x.com
# ─────────────────────────────────────────────────────────
# SHA256: f53cd15c9345b7817f397aab3f4870ee36be1fef321d0b49e81cd81819b92462
# SIG-ED25519: 1ZvlFLjbkZzpH4HnttlYSB3ydsAKgG57oyAElSRcvMzqOT3pQ+FLHW3seWlOUpAUI77d6AvrjV5SNCJuL6kuBw==
# VERIFY: python3 verify_authorship.py organ_graft.py

View File

@ -1,13 +1,11 @@
#!/usr/bin/env python3
"""
Organ Architecture organ_measure.py
Z-measure organ quality: signal vs noise.
Quality measure — organ signal vs noise.
Z = dI/d(log s) · exp()
CSCI cross-scale coherence index
θ 0° : noise (organ adds confusion)
θ 90° : signal (organ adds knowledge)
Signature 935
"""
import struct
@ -86,9 +84,9 @@ def read_organ_data_f32(filepath, max_elements=100000):
def compute_z_measure(values):
"""
Compute Z-measure for a tensor.
Compute quality measure for a tensor.
Z = dI/d(log s) · exp()
CSCI cross-scale coherence index
We measure:
- Information density (entropy of distribution)
@ -246,7 +244,7 @@ def measure_directory(organ_dir, verbose=False):
def print_summary(results, title=""):
"""Print Z-measure summary."""
"""Print quality summary."""
if not results:
print("No organs measured.")
return
@ -260,7 +258,7 @@ def print_summary(results, title=""):
groups[dirname].append(r)
print(f"\n{'='*70}")
print(f" Z-MEASURE REPORT {title}")
print(f" QUALITY REPORT {title}")
print(f"{'='*70}")
for group_name in ['skeleton', 'organs', 'embed', 'norm', 'adapters', 'unknown']:
@ -299,14 +297,12 @@ def print_summary(results, title=""):
print(f"\n {''*50}")
print(f" GLOBAL: {len(results)} tensors | {total_size:.1f} MB | θ={avg_theta:.1f}° | signal={avg_signal:.3f}")
print(f" Signature 935")
print(f"{'='*70}")
def main():
parser = argparse.ArgumentParser(
description='Organ Architecture — Z-measure organ quality',
epilog='Z = dI/d(log s) · exp(iθ) — Signature 935'
epilog='CSCI v1.0 — Cross-Scale Coherence Index'
)
parser.add_argument('--organ', '-o', help='Path to single organ .bin file')
parser.add_argument('--dir', '-d', help='Path to extracted organs directory')
@ -338,3 +334,10 @@ def main():
if __name__ == '__main__':
main()
# ╔══ SALKA ELMADANI AUTHORSHIP CERTIFICATE ══╗
# © Salka Elmadani 2025-2026 — ALL RIGHTS RESERVED
# Licensed under Business Source License 1.1 — https://inference-x.com
# ─────────────────────────────────────────────────────────
# SHA256: 0851280f9f83e9f30e35fd7efff164f806f506f94aa9cd983c8fdae7318a9864
# SIG-ED25519: 7VtyjAri7KRdqUuc+WdkQkp50xKAkVRFqgqLHnJG0BkBltqVwJeYMScAkZ56b4mcsBWPhkj0Y8kS1fd2t/Y+BQ==
# VERIFY: python3 verify_authorship.py organ_measure.py

View File

@ -1,9 +1,9 @@
#!/usr/bin/env python3
"""
ORGAN PURIFIER Z = i
ORGAN PURIFIER signal extraction
Remove noise from tensor weights. Keep only pure signal.
The paradigm creates artificial boundaries between models.
Training creates artificial boundaries between models.
Under the noise, the signal is universal.
A weight that encodes "attention to context" is the same law
whether it comes from Qwen, Llama, or Gemma.
@ -17,11 +17,10 @@ Method:
5. Inverse FFT: reconstructed tensor = pure signal
6. Verify: new theta should be closer to 90
Z = dI/d(log s) * exp(i*theta)
When theta = 90, Z = i (pure imaginary = pure potential)
CSCI(s) = cross_scale_coherence(s, theta=90)
When theta = 90, signal is maximally coherent (pure signal, minimal noise)
The purified organ IS the signal, nothing else.
Signature 935
"""
import struct
@ -34,7 +33,7 @@ from pathlib import Path
# === Z CONSTANTS ===
THETA_TARGET_DEG = 90.0 # Pure signal
ENTROPY_TARGET = 0.3251 # Z-COM optimum
ENTROPY_TARGET = 0.3251 # empirical optimum
NOISE_THRESHOLD = 0.3 # Below this in frequency domain = noise
PRESERVE_RATIO = 0.85 # Keep top 85% of spectral energy (signal)
@ -145,7 +144,7 @@ def purify_organ(values, preserve_ratio=PRESERVE_RATIO):
The signal lives in the structured components of the frequency domain.
The noise lives in the high-entropy, low-energy tail.
Z = dI/d(log s) * exp(i*theta)
CSCI(s) = cross_scale_coherence(s, theta=90)
In frequency space:
- High magnitude + low frequency = structural signal (keep)
@ -154,7 +153,7 @@ def purify_organ(values, preserve_ratio=PRESERVE_RATIO):
This is not simple low-pass filtering.
We keep the components that carry INFORMATION (high dI),
at the NATURAL SCALE (log s), with COHERENT PHASE (theta -> 90).
at the natural scale, with coherent phase (theta -> 90).
"""
n = len(values)
if n < 32:
@ -293,8 +292,7 @@ def purify_model(organ_dir, output_dir, verbose=False):
def main():
import argparse
parser = argparse.ArgumentParser(
description='Organ Purifier — Z = i — Remove noise, keep pure signal',
epilog='Z = dI/d(log s) · exp(iθ), θ=90° — Signature 935'
description='Organ Purifier — Remove noise, keep pure signal',
)
parser.add_argument('--input', '-i', required=True, help='Input organs directory')
parser.add_argument('--output', '-o', required=True, help='Output pure organs directory')
@ -308,7 +306,7 @@ def main():
PRESERVE_RATIO = args.preserve
print(f"{'='*60}")
print(f" ORGAN PURIFIER — Z = i")
print(f" ORGAN PURIFIER — signal extraction")
print(f" Signal preservation: {PRESERVE_RATIO*100:.0f}%")
print(f"{'='*60}")
print(f" Input: {args.input}")
@ -325,9 +323,15 @@ def main():
print(f" θ after: {result['avg_theta_after']:.1f}°")
print(f" Avg improvement: {result['avg_improvement']:+.1f}°")
print(f" Output: {result['output']}")
print(f" Signature: 935")
print(f"{'='*60}")
if __name__ == '__main__':
main()
# ╔══ SALKA ELMADANI AUTHORSHIP CERTIFICATE ══╗
# © Salka Elmadani 2025-2026 — ALL RIGHTS RESERVED
# Licensed under Business Source License 1.1 — https://inference-x.com
# ─────────────────────────────────────────────────────────
# SHA256: d3ab5384c880f7e88fb7cdad4b2f9f56089ada8395d0013f5bd3b09d7ab631e8
# SIG-ED25519: /rkXFm2tGuoAS61oxWZVlcTghUuGL8HJ11XRSaI4Ak+eEt54uo+3NETX2+5S8HAq72k6whQmbPI3f4jD8sF/CA==
# VERIFY: python3 verify_authorship.py organ_purify.py

View File

@ -1,13 +1,13 @@
#!/usr/bin/env python3
"""
ORGAN PURIFIER V2 Z = i Fractal Signal Extraction
ORGAN PURIFIER V2 Signal Extraction
V1 failed because it treated tensors like audio signals.
Tensors are NOT audio. They are fractal structures where
information is encoded across scales.
The correct approach from Z = dI/d(log s) * exp(i*theta):
- dI/d(log s) = how information CHANGES across scales
The correct approach from CSCI(s) = cross_scale_coherence(s, theta=90):
- cross-scale derivative = how information CHANGES across scales
- Signal = components that are SELF-SIMILAR across scales (fractal)
- Noise = components that are RANDOM across scales (non-fractal)
@ -23,8 +23,7 @@ Method:
Think fractal: the best model knows the laws of the universe
then translates to human language, not the inverse.
Z = dI/d(log s) * exp(i*theta), theta = 90
Signature 935
CSCI(s) = cross_scale_coherence(s, theta=90), theta = 90
"""
import struct, os, sys, json, math
@ -198,7 +197,7 @@ def purify_fractal(values):
"""
Fractal purification: keep cross-scale-coherent components.
dI/d(log s): information that persists across scales IS the signal.
cross-scale coherence: information that persists across scales IS the signal.
Everything else is training noise, brand artifacts, paradigm residue.
"""
n = len(values)
@ -247,8 +246,8 @@ def purify_model(organ_dir, output_dir, verbose=False):
if manifest_src.exists():
manifest = json.load(open(manifest_src))
manifest['purified'] = True
manifest['purifier'] = 'fractal_v2'
manifest['z_equation'] = 'Z = dI/d(log s) * exp(i*theta), theta=90'
manifest['purifier'] = 'purify_v2'
manifest['coherence_score'] = 'cross_scale_coherence(s)'
# Remove brand from model name
original_name = manifest.get('model', 'unknown')
manifest['original_model'] = original_name
@ -308,14 +307,14 @@ def purify_model(organ_dir, output_dir, verbose=False):
def main():
import argparse
parser = argparse.ArgumentParser(description='Organ Purifier V2 — Fractal Z=i')
parser = argparse.ArgumentParser(description='Organ Purifier V2 — signal extraction')
parser.add_argument('--input', '-i', required=True)
parser.add_argument('--output', '-o', required=True)
parser.add_argument('--verbose', '-v', action='store_true')
args = parser.parse_args()
print(f"{'='*60}")
print(f" ORGAN PURIFIER V2 — FRACTAL — Z = i")
print(f" ORGAN PURIFIER V2")
print(f" Cross-scale coherence: signal persists, noise vanishes")
print(f"{'='*60}")
@ -330,8 +329,14 @@ def main():
print(f" Δθ: {result['delta']:+.1f}°")
print(f" Improved: {result['improved']}")
print(f" Degraded: {result['degraded']}")
print(f" Signature: 935")
print(f"{'='*60}")
if __name__ == '__main__':
main()
# ╔══ SALKA ELMADANI AUTHORSHIP CERTIFICATE ══╗
# © Salka Elmadani 2025-2026 — ALL RIGHTS RESERVED
# Licensed under Business Source License 1.1 — https://inference-x.com
# ─────────────────────────────────────────────────────────
# SHA256: 0328644f84762361db812407ed482018de40a92f496d9b45bf56826d59184224
# SIG-ED25519: Y1KrhUdgrqiYPaM0LPHWTqPKPaHwBqtc3EiHnu9Uu94AVKsgMPQoWU9NCGeiL5aWAJKPhzr/nCSxLTY+US+HAw==
# VERIFY: python3 verify_authorship.py organ_purify_v2.py

View File

@ -1,14 +1,10 @@
#!/usr/bin/env python3
"""
Model 935 Pipeline Phase 1: Dissect all + Download Kimi K2.5
Z = dI/d(log s) · exp() Signature 935
CSCI v1.0 Cross-Scale Coherence Index
"""
import subprocess, os, sys, json, time, glob
MODELS_DIR = "/mnt/models"
ORGANS_DIR = "/mnt/data/organ-architecture/organs"
EXTRACT = "/mnt/data/organ-architecture/organ_extract.py"
MEASURE = "/mnt/data/organ-architecture/organ_measure.py"
os.makedirs(ORGANS_DIR, exist_ok=True)
@ -16,8 +12,6 @@ os.makedirs(ORGANS_DIR, exist_ok=True)
models = {}
for f in sorted(glob.glob(os.path.join(MODELS_DIR, "*.gguf"))):
name = os.path.basename(f)
# Skip chimeras and old 935 attempts
if "chimera" in name.lower() or "935" in name.lower():
continue
# Clean name for directory
clean = name.replace(".gguf", "").replace("-Q4_K_M", "").replace("-Q8_0", "")
@ -59,12 +53,11 @@ for gguf_name, organ_name in models.items():
print(f" [ERROR] {r.stderr[-200:]}")
results.append({"model": organ_name, "status": "error"})
# Z-measure all
# Quality measure all
print(f"\n{'='*60}")
print(f"PHASE 2: Z-MEASURE ALL ORGANS")
print(f"PHASE 2: QUALITY MEASURE ALL ORGANS")
print(f"{'='*60}")
sys.path.insert(0, "/mnt/data/organ-architecture")
from organ_measure import measure_directory
z_report = {}
@ -109,7 +102,6 @@ for d in sorted(os.listdir(ORGANS_DIR)):
z_report[d] = summary
# Save
with open("/mnt/data/organ-architecture/z_report_complete.json", "w") as f:
json.dump(z_report, f, indent=2)
# Print ranking
@ -120,5 +112,10 @@ ranked = sorted(z_report.values(), key=lambda m: m['avg_theta'], reverse=True)
for i, m in enumerate(ranked, 1):
print(f" {i:2d}. θ={m['avg_theta']:5.1f}° signal={m['avg_signal']:.3f} {m['model']}")
print(f"\n Signature: 935")
print(f"{'='*60}")
# ╔══ SALKA ELMADANI AUTHORSHIP CERTIFICATE ══╗
# © Salka Elmadani 2025-2026 — ALL RIGHTS RESERVED
# Licensed under Business Source License 1.1 — https://inference-x.com
# ─────────────────────────────────────────────────────────
# SHA256: 70a8957904cd4ee20dfd8fa42a0d8551cf8ae03eb2d0ec6fc9f4ed8f86995037
# SIG-ED25519: ddMrNVlt0PpN5uHTbAnxLkphci22Xv0efiEyfUAoHVJxextDZsK69jVULKiXZDED1txsfGzrenMjJMaKe5g4DQ==

View File

@ -2,7 +2,6 @@
"""
Quick chimera assembler: Copy source GGUF header/metadata intact,
replace tensor data from organ directory.
Signature 935
"""
import struct, sys, os, json
@ -117,7 +116,13 @@ def main():
print(f"\n Output: {output_gguf}")
print(f" Size: {final_size / (1024**3):.2f} GB")
print(f" From organs: {written}, From source: {fallback}, Total: {written+fallback}/{n_tensors}")
print(f" Signature: 935")
if __name__ == "__main__":
main()
# ╔══ SALKA ELMADANI AUTHORSHIP CERTIFICATE ══╗
# © Salka Elmadani 2025-2026 — ALL RIGHTS RESERVED
# Licensed under Business Source License 1.1 — https://inference-x.com
# ─────────────────────────────────────────────────────────
# SHA256: b0d040908eddc26078e86f76e361825fada5c2676778789ef41c1804730eb10d
# SIG-ED25519: srq6F3EyKqi7r3nlB6cfI1u53J1GpsC2ty9zNsBDrZ2EldVVIhE1mWCdnd/qkvgif783DOlLQ4Zb2CCw13XfBQ==
# VERIFY: python3 verify_authorship.py quick_chimera.py

View File

@ -3,7 +3,7 @@
Quick chimera assembler v2: FIXED organ header handling.
Organ .bin files have: [name_len(4) + name + n_dims(4) + dims(8*n) + dtype(4)] + DATA
We must skip the header and only copy the DATA portion.
Z = dI/d(log s) · exp() Signature 935
CSCI v1.0 Cross-Scale Coherence Index
"""
import struct, sys, os, json
@ -149,7 +149,13 @@ def main():
diff = final_size - source_size
print(f" INTEGRITY: ✗ MISMATCH ({diff:+d} bytes)")
print(f" Signature: 935")
if __name__ == "__main__":
main()
# ╔══ SALKA ELMADANI AUTHORSHIP CERTIFICATE ══╗
# © Salka Elmadani 2025-2026 — ALL RIGHTS RESERVED
# Licensed under Business Source License 1.1 — https://inference-x.com
# ─────────────────────────────────────────────────────────
# SHA256: 6587e64dbf1c6fe2160fe8f2e25a33e6ed5e98193baea7f7523a9495e04b9154
# SIG-ED25519: TrwO40O2Qn0ysnadlzX38fBTSOF5St11SyZTSc4cZP/7k5HM+ifnqDMTu/vkZWDYAdmb+5bc6IhpYYQgVdLsBA==
# VERIFY: python3 verify_authorship.py quick_chimera_v2.py

View File

@ -1,126 +0,0 @@
#!/usr/bin/env python3
"""
GGUF-to-GGUF transplant. No organ bins direct tensor copy between GGUF files.
Base: DeepSeek-R1-Distill-Qwen-7B (skeleton/attention/embed)
Donor: Qwen2.5-7B (FFN organs only)
Z = dI/d(log s) · exp() Signature 935
"""
import struct, os, sys, shutil
def parse_gguf_header(path):
"""Parse GGUF header, return tensor_info list and data_start offset."""
f = open(path, "rb")
magic = struct.unpack("<I", f.read(4))[0]
version = struct.unpack("<I", f.read(4))[0]
n_tensors = struct.unpack("<Q", f.read(8))[0]
n_metadata = struct.unpack("<Q", f.read(8))[0]
def read_string():
slen = struct.unpack("<Q", f.read(8))[0]
return f.read(slen).decode("utf-8")
def skip_value(vtype):
sizes = {0:1, 1:1, 2:2, 3:2, 4:4, 5:4, 6:4, 7:1, 10:8, 11:8, 12:8}
if vtype in sizes:
f.read(sizes[vtype])
elif vtype == 8:
read_string()
elif vtype == 9:
arr_type = struct.unpack("<I", f.read(4))[0]
arr_len = struct.unpack("<Q", f.read(8))[0]
for _ in range(arr_len):
skip_value(arr_type)
for _ in range(n_metadata):
read_string()
vtype = struct.unpack("<I", f.read(4))[0]
skip_value(vtype)
tensors = []
for _ in range(n_tensors):
name = read_string()
n_dims = struct.unpack("<I", f.read(4))[0]
dims = [struct.unpack("<Q", f.read(8))[0] for _ in range(n_dims)]
dtype = struct.unpack("<I", f.read(4))[0]
offset = struct.unpack("<Q", f.read(8))[0]
tensors.append({"name": name, "dims": dims, "dtype": dtype, "offset": offset})
pos = f.tell()
padding = (32 - (pos % 32)) % 32
f.read(padding)
data_start = f.tell()
f.seek(0, 2)
file_end = f.tell()
f.close()
# Calculate sizes
for i in range(len(tensors)):
if i + 1 < len(tensors):
tensors[i]["size"] = tensors[i+1]["offset"] - tensors[i]["offset"]
else:
tensors[i]["size"] = file_end - data_start - tensors[i]["offset"]
return tensors, data_start, file_end
BASE = "/mnt/models/DeepSeek-R1-Distill-Qwen-7B-Q4_K_M.gguf"
DONOR = "/mnt/models/Qwen2.5-7B-Instruct-Q4_K_M.gguf"
OUTPUT = "/mnt/models/model-935-final.gguf"
print("Parsing base (DeepSeek-R1-7B)...")
base_tensors, base_data_start, base_end = parse_gguf_header(BASE)
print(f" {len(base_tensors)} tensors, data_start={base_data_start}")
print("Parsing donor (Qwen2.5-7B)...")
donor_tensors, donor_data_start, donor_end = parse_gguf_header(DONOR)
print(f" {len(donor_tensors)} tensors, data_start={donor_data_start}")
# Build donor tensor map by name
donor_map = {t["name"]: t for t in donor_tensors}
# Copy base GGUF entirely first
print(f"Copying base to output...")
shutil.copy2(BASE, OUTPUT)
# Now patch: for each FFN tensor in base, if donor has matching name+size, overwrite
out = open(OUTPUT, "r+b")
donor_f = open(DONOR, "rb")
grafted = 0
skipped = 0
for bt in base_tensors:
name = bt["name"]
# Only graft FFN organs (not attention, not embeddings, not norms)
if "ffn_down" not in name and "ffn_up" not in name and "ffn_gate" not in name:
continue
if name in donor_map:
dt = donor_map[name]
if bt["size"] == dt["size"]:
# Read from donor
donor_f.seek(donor_data_start + dt["offset"])
data = donor_f.read(dt["size"])
# Write to output at same offset
out.seek(base_data_start + bt["offset"])
out.write(data)
grafted += 1
else:
skipped += 1
else:
skipped += 1
out.close()
donor_f.close()
print(f"\n{'='*60}")
print(f" MODEL 935 — DIRECT GGUF TRANSPLANT")
print(f"{'='*60}")
print(f" Base: DeepSeek-R1-Distill-Qwen-7B (skeleton+embed)")
print(f" Donor: Qwen2.5-7B-Instruct (FFN organs)")
print(f" Grafted: {grafted} FFN tensors")
print(f" Skipped: {skipped} (size mismatch or not found)")
print(f" Output: {OUTPUT}")
print(f" Size: {os.path.getsize(OUTPUT)/(1024**3):.2f} GB")
print(f" Signature: 935")
print(f"{'='*60}")

72
verify_authorship.py Normal file
View File

@ -0,0 +1,72 @@
#!/usr/bin/env python3
"""Verify SALKA ELMADANI authorship signatures.
Usage: python3 verify_authorship.py [file_or_directory]
Every source file carries an Ed25519 signature bound to SHA-256 of content.
Modify 1 character = signature invalid = tampering detected.
"""
import sys, hashlib, base64, os
from pathlib import Path
from cryptography.hazmat.primitives import serialization
from cryptography.exceptions import InvalidSignature
PUBLIC_KEY_PEM = """
-----BEGIN PUBLIC KEY-----
MCowBQYDK2VwAyEARCtdhRqqYcu7c8qwyoRKRn5Qbx9puylZHZOM+IsDp0U=
-----END PUBLIC KEY-----
""".strip()
def strip_sig(text):
lines, out, in_b = text.split("\n"), [], False
for ln in lines:
if MARK_S in ln: in_b = True; continue
if MARK_E in ln and in_b: in_b = False; continue
if not in_b: out.append(ln)
return "\n".join(out).rstrip("\n") + "\n"
def extract_sig_data(text):
sha, sig = None, None
for ln in text.split("\n"):
if "SHA256:" in ln: sha = ln.split("SHA256:")[-1].strip().lstrip("#/ ")
if "SIG-ED25519:" in ln: sig = ln.split("SIG-ED25519:")[-1].strip().lstrip("#/ ")
return sha, sig
def verify_file(fp, pub):
try: content = Path(fp).read_text(encoding="utf-8", errors="replace")
except: return None, "Cannot read"
clean = strip_sig(content)
claimed_h, sig_b64 = extract_sig_data(content)
if not claimed_h or not sig_b64: return False, "No signature"
actual_h = hashlib.sha256(clean.encode("utf-8")).hexdigest()
if actual_h != claimed_h: return False, f"HASH MISMATCH — modified"
try:
pub.verify(base64.b64decode(sig_b64), hashlib.sha256(clean.encode()).digest())
return True, "VALID © Salka Elmadani"
except InvalidSignature: return False, "INVALID SIGNATURE — forgery"
except Exception as e: return False, str(e)
def main():
pub = serialization.load_pem_public_key(PUBLIC_KEY_PEM.encode())
target = sys.argv[1] if len(sys.argv) > 1 else "."
files = [Path(target)] if Path(target).is_file() else [
p for p in Path(target).rglob("*")
if p.is_file() and p.suffix in [".py",".cpp",".h",".js",".ts",".sh",".rs",".go",".md"]
and ".git" not in str(p)
]
ok, fail, skip = 0, 0, 0
for f in sorted(files):
r, msg = verify_file(f, pub)
if r is None: skip += 1
elif r: ok += 1; print(f"{f.name}: {msg}")
else: fail += 1; print(f"{f.name}: {msg}")
print(f"\nResults: {ok} valid | {fail} TAMPERED | {skip} skipped")
if fail: print("WARNING: Authorship chain broken."); sys.exit(1)
if __name__ == "__main__": main()
# ╔══ SALKA ELMADANI AUTHORSHIP CERTIFICATE ══╗
# © Salka Elmadani 2025-2026 — ALL RIGHTS RESERVED
# Licensed under Business Source License 1.1 — https://inference-x.com
# ─────────────────────────────────────────────────────────
# SHA256: f4e32c8fe1f2cb7f5dc498c7506b054256f6871d7283beb74b1d5859eb775121
# SIG-ED25519: g4rHIrZteuUk4HU/21i69rTk7H8EiL1XjX4A+dZD0xswTqR5XJb1CfnBQfyAxjb1Sf9VW3JptZVDkvOq+magCA==
# VERIFY: python3 verify_authorship.py verify_authorship.py

View File

@ -1,501 +0,0 @@
{
"chimera-deepseek-qwen": {
"model": "chimera-deepseek-qwen",
"total_tensors": 339,
"avg_theta": 45.53097345132743,
"avg_signal": 0.6371591309220915,
"groups": {
"skeleton": {
"count": 196,
"avg_theta": 54.2,
"avg_signal": 0.727,
"best_theta": 84.0,
"best_name": "blk.0.attn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_k.weight"
},
"organs": {
"count": 112,
"avg_theta": 35.9,
"avg_signal": 0.538,
"best_theta": 84.0,
"best_name": "blk.0.ffn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.ffn_gate.weight"
},
"embed": {
"count": 31,
"avg_theta": 25.9,
"avg_signal": 0.429,
"best_theta": 75.0,
"best_name": "output_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_output.weight"
}
}
},
"deepseek-r1-14b": {
"model": "deepseek-r1-14b",
"total_tensors": 579,
"avg_theta": 46.01036269430051,
"avg_signal": 0.640550897397108,
"groups": {
"skeleton": {
"count": 336,
"avg_theta": 55.4,
"avg_signal": 0.736,
"best_theta": 84.0,
"best_name": "blk.0.attn_k.bias",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_k.weight"
},
"organs": {
"count": 192,
"avg_theta": 35.2,
"avg_signal": 0.532,
"best_theta": 84.0,
"best_name": "blk.0.ffn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.ffn_down.weight"
},
"embed": {
"count": 51,
"avg_theta": 25.2,
"avg_signal": 0.42,
"best_theta": 75.0,
"best_name": "output_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_output.weight"
}
}
},
"deepseek-r1-7b": {
"model": "deepseek-r1-7b",
"total_tensors": 339,
"avg_theta": 45.424778761061944,
"avg_signal": 0.6355319640555519,
"groups": {
"skeleton": {
"count": 196,
"avg_theta": 54.2,
"avg_signal": 0.727,
"best_theta": 84.0,
"best_name": "blk.0.attn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_k.weight"
},
"organs": {
"count": 112,
"avg_theta": 35.5,
"avg_signal": 0.533,
"best_theta": 84.0,
"best_name": "blk.0.ffn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.ffn_gate.weight"
},
"embed": {
"count": 31,
"avg_theta": 25.9,
"avg_signal": 0.429,
"best_theta": 75.0,
"best_name": "output_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_output.weight"
}
}
},
"deepseek-r1-distill-7b": {
"model": "deepseek-r1-distill-7b",
"total_tensors": 339,
"avg_theta": 45.53097345132743,
"avg_signal": 0.6371591309220915,
"groups": {
"skeleton": {
"count": 196,
"avg_theta": 54.2,
"avg_signal": 0.727,
"best_theta": 84.0,
"best_name": "blk.0.attn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_k.weight"
},
"organs": {
"count": 112,
"avg_theta": 35.9,
"avg_signal": 0.538,
"best_theta": 84.0,
"best_name": "blk.0.ffn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.ffn_gate.weight"
},
"embed": {
"count": 31,
"avg_theta": 25.9,
"avg_signal": 0.429,
"best_theta": 75.0,
"best_name": "output_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_output.weight"
}
}
},
"gemma2-9b": {
"model": "gemma2-9b",
"total_tensors": 464,
"avg_theta": 44.935344827586206,
"avg_signal": 0.6240438819131022,
"groups": {
"skeleton": {
"count": 210,
"avg_theta": 47.2,
"avg_signal": 0.649,
"best_theta": 84.0,
"best_name": "blk.0.post_attention_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_k.weight"
},
"organs": {
"count": 168,
"avg_theta": 37.9,
"avg_signal": 0.552,
"best_theta": 84.0,
"best_name": "blk.1.ffn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.ffn_down.weight"
},
"embed": {
"count": 44,
"avg_theta": 26.2,
"avg_signal": 0.433,
"best_theta": 84.0,
"best_name": "output_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_output.weight"
},
"norm": {
"count": 42,
"avg_theta": 81.6,
"avg_signal": 0.987,
"best_theta": 84.0,
"best_name": "blk.10.post_ffw_norm.weight",
"worst_theta": 75.0,
"worst_name": "blk.0.post_ffw_norm.weight"
}
}
},
"llama31-8b": {
"model": "llama31-8b",
"total_tensors": 292,
"avg_theta": 37.86986301369863,
"avg_signal": 0.5490538952939957,
"groups": {
"skeleton": {
"count": 128,
"avg_theta": 39.7,
"avg_signal": 0.569,
"best_theta": 84.0,
"best_name": "blk.10.attn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.10.attn_k.weight"
},
"organs": {
"count": 128,
"avg_theta": 39.1,
"avg_signal": 0.56,
"best_theta": 84.0,
"best_name": "blk.0.ffn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.ffn_down.weight"
},
"embed": {
"count": 35,
"avg_theta": 26.0,
"avg_signal": 0.427,
"best_theta": 84.0,
"best_name": "output_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_output.weight"
}
}
},
"llama32-1b": {
"model": "llama32-1b",
"total_tensors": 147,
"avg_theta": 37.57142857142857,
"avg_signal": 0.5497319048747188,
"groups": {
"skeleton": {
"count": 64,
"avg_theta": 39.3,
"avg_signal": 0.57,
"best_theta": 84.0,
"best_name": "blk.10.attn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_q.weight"
},
"organs": {
"count": 64,
"avg_theta": 38.3,
"avg_signal": 0.553,
"best_theta": 84.0,
"best_name": "blk.10.ffn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.ffn_down.weight"
},
"embed": {
"count": 18,
"avg_theta": 27.3,
"avg_signal": 0.445,
"best_theta": 75.0,
"best_name": "output_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_output.weight"
}
}
},
"llama32-3b": {
"model": "llama32-3b",
"total_tensors": 255,
"avg_theta": 37.411764705882355,
"avg_signal": 0.546769292896037,
"groups": {
"skeleton": {
"count": 112,
"avg_theta": 39.4,
"avg_signal": 0.569,
"best_theta": 84.0,
"best_name": "blk.0.attn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_k.weight"
},
"organs": {
"count": 112,
"avg_theta": 38.0,
"avg_signal": 0.55,
"best_theta": 84.0,
"best_name": "blk.13.ffn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.ffn_down.weight"
},
"embed": {
"count": 30,
"avg_theta": 26.6,
"avg_signal": 0.439,
"best_theta": 75.0,
"best_name": "output_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_output.weight"
}
}
},
"mistral-7b": {
"model": "mistral-7b",
"total_tensors": 291,
"avg_theta": 36.20618556701031,
"avg_signal": 0.539809742436977,
"groups": {
"skeleton": {
"count": 128,
"avg_theta": 38.4,
"avg_signal": 0.567,
"best_theta": 84.0,
"best_name": "blk.0.attn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.10.attn_k.weight"
},
"organs": {
"count": 128,
"avg_theta": 36.8,
"avg_signal": 0.544,
"best_theta": 84.0,
"best_name": "blk.0.ffn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.ffn_down.weight"
},
"embed": {
"count": 35,
"avg_theta": 26.0,
"avg_signal": 0.427,
"best_theta": 84.0,
"best_name": "output_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_output.weight"
}
}
},
"phi35-mini": {
"model": "phi35-mini",
"total_tensors": 197,
"avg_theta": 44.6497461928934,
"avg_signal": 0.6262773662109529,
"groups": {
"skeleton": {
"count": 64,
"avg_theta": 56.7,
"avg_signal": 0.764,
"best_theta": 84.0,
"best_name": "blk.10.attn_norm.weight",
"worst_theta": 33.0,
"worst_name": "blk.0.attn_qkv.weight"
},
"organs": {
"count": 96,
"avg_theta": 43.2,
"avg_signal": 0.601,
"best_theta": 84.0,
"best_name": "blk.0.ffn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.ffn_down.weight"
},
"embed": {
"count": 35,
"avg_theta": 26.7,
"avg_signal": 0.439,
"best_theta": 84.0,
"best_name": "output_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_output.weight"
}
}
},
"qwen25-14b": {
"model": "qwen25-14b",
"total_tensors": 579,
"avg_theta": 45.98445595854922,
"avg_signal": 0.6402458335664142,
"groups": {
"skeleton": {
"count": 336,
"avg_theta": 55.2,
"avg_signal": 0.734,
"best_theta": 84.0,
"best_name": "blk.0.attn_k.bias",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_k.weight"
},
"organs": {
"count": 192,
"avg_theta": 35.4,
"avg_signal": 0.534,
"best_theta": 84.0,
"best_name": "blk.0.ffn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.ffn_down.weight"
},
"embed": {
"count": 51,
"avg_theta": 25.5,
"avg_signal": 0.424,
"best_theta": 84.0,
"best_name": "output_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_output.weight"
}
}
},
"qwen25-3b": {
"model": "qwen25-3b",
"total_tensors": 434,
"avg_theta": 46.00230414746544,
"avg_signal": 0.6401608443093786,
"groups": {
"skeleton": {
"count": 252,
"avg_theta": 55.6,
"avg_signal": 0.736,
"best_theta": 84.0,
"best_name": "blk.0.attn_k.bias",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_k.weight"
},
"organs": {
"count": 144,
"avg_theta": 34.5,
"avg_signal": 0.529,
"best_theta": 84.0,
"best_name": "blk.10.ffn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.ffn_down.weight"
},
"embed": {
"count": 38,
"avg_theta": 25.8,
"avg_signal": 0.426,
"best_theta": 84.0,
"best_name": "output_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_output.weight"
}
}
},
"qwen25-7b": {
"model": "qwen25-7b",
"total_tensors": 339,
"avg_theta": 45.637168141592916,
"avg_signal": 0.6387682956137819,
"groups": {
"skeleton": {
"count": 196,
"avg_theta": 54.6,
"avg_signal": 0.731,
"best_theta": 84.0,
"best_name": "blk.0.attn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_k.weight"
},
"organs": {
"count": 112,
"avg_theta": 35.5,
"avg_signal": 0.536,
"best_theta": 84.0,
"best_name": "blk.0.ffn_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.ffn_gate.weight"
},
"embed": {
"count": 31,
"avg_theta": 25.9,
"avg_signal": 0.429,
"best_theta": 75.0,
"best_name": "output_norm.weight",
"worst_theta": 24.0,
"worst_name": "blk.0.attn_output.weight"
}
}
},
"smollm2-135m": {
"model": "smollm2-135m",
"total_tensors": 272,
"avg_theta": 52.27941176470588,
"avg_signal": 0.7765030923203783,
"groups": {
"skeleton": {
"count": 120,
"avg_theta": 53.6,
"avg_signal": 0.79,
"best_theta": 84.0,
"best_name": "blk.0.attn_norm.weight",
"worst_theta": 42.0,
"worst_name": "blk.10.attn_k.weight"
},
"organs": {
"count": 120,
"avg_theta": 52.3,
"avg_signal": 0.777,
"best_theta": 84.0,
"best_name": "blk.0.ffn_norm.weight",
"worst_theta": 42.0,
"worst_name": "blk.11.ffn_up.weight"
},
"embed": {
"count": 32,
"avg_theta": 47.2,
"avg_signal": 0.725,
"best_theta": 84.0,
"best_name": "output_norm.weight",
"worst_theta": 33.0,
"worst_name": "blk.13.attn_output.weight"
}
}
}
}

View File

@ -70,7 +70,6 @@
},
"gemma-2-9b": {
"model": "gemma-2-9b",
"avg_theta": 44.935344827586206,
"avg_signal": 0.6240438819131022,
"total_tensors": 464,
"groups": {
@ -303,4 +302,4 @@
}
}
}
}
}

File diff suppressed because it is too large Load Diff