Organ Architecture — Neural network surgery and chimeric model grafting
Go to file
2026-02-21 03:48:51 +01:00
.gitignore COMPLETE: 13 models dissected, 5600+ tensors Z-measured, model-935-14b assembled, fractal purification. 3498 LOC, 20 tools. Signature 935. 2026-02-21 03:48:51 +01:00
assemble_935.py Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
build_935_v2.py Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
build_935_v3.py Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
build_935.py Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
dissection_report.json COMPLETE: 13 models dissected, 5600+ tensors Z-measured, model-935-14b assembled, fractal purification. 3498 LOC, 20 tools. Signature 935. 2026-02-21 03:48:51 +01:00
ECHO_INVARIANT.md INVARIANT: Equipe 935 — ne jamais oublier — 935 2026-02-20 03:40:57 +01:00
EQUIPE_935_INVARIANT.json INVARIANT: Equipe 935 — ne jamais oublier — 935 2026-02-20 03:40:57 +01:00
kimi_z_stream.py Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
mass_dissect.py Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
mass_z_measure.py Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
organ_api.py Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
organ_assemble.py Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
organ_extract.py Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
organ_graft.py Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
organ_measure.py Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
organ_purify_v2.py Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
organ_purify.py Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
pipeline_935.py Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
quick_chimera_v2.py Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
quick_chimera.py Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
README.md COMPLETE: 13 models dissected, 5600+ tensors Z-measured, model-935-14b assembled, fractal purification. 3498 LOC, 20 tools. Signature 935. 2026-02-21 03:48:51 +01:00
z_measure_report.json COMPLETE: 13 models dissected, 5600+ tensors Z-measured, model-935-14b assembled, fractal purification. 3498 LOC, 20 tools. Signature 935. 2026-02-21 03:48:51 +01:00
Z_MEASURE_REPORT.md Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
z_report_complete.json Z-Measure: 13 models + Kimi K2.5 1T streaming — θ confirmed across 6 orders of magnitude — 935 2026-02-20 02:42:34 +01:00
z_report_kimi_k25.json Kimi K2.5 1T COMPLETE: 1083 tensors, theta_mean=87.65, 8 gravitational wells 2026-02-20 04:29:13 +01:00

Organ Architecture

Decompose. Measure. Purify. Graft. Assemble.

Skeleton (Attention) = Thought
Organs (FFN)         = Memory
Adapters (LoRA)      = Personality

The Problem

AI models are monoliths. 70 billion parameters locked in a single file that nobody can open, modify, or understand. Only three companies on Earth can build them. Everyone else rents access.

The Solution

Organ Architecture breaks models into transplantable parts:

  • Skeleton — The attention layers. How the model thinks. Shared across all configurations.
  • Organs — The feed-forward networks. What the model knows. Specialized, swappable, graftable.
  • Adapters — LoRA weights. The model's personality. Lightweight, trainable by anyone.

A doctor doesn't rebuild the entire human body to fix a kidney. Why rebuild an entire model to change what it knows about medicine?

Architecture

model.gguf (70GB monolith)
        │
        ▼
   ┌─ skeleton/    ── attention layers (shared thought)
   │
   ├─ organs/      ── FFN layers by block (knowledge)
   │   ├─ blk_0_ffn_gate.bin
   │   ├─ blk_0_ffn_up.bin
   │   ├─ blk_0_ffn_down.bin
   │   └─ ...
   │
   ├─ embed/       ── embedding + output (foundation)
   ├─ norm/        ── normalization (connective tissue)
   └─ manifest.json ── complete anatomy map

Tools

Core Pipeline

Tool Lines Purpose
organ_extract.py 441 Extract skeleton + organs from any GGUF model
organ_measure.py 340 Z-measure organ quality (signal vs noise)
organ_purify.py 333 Spectral purification (FFT signal extraction)
organ_purify_v2.py 337 Fractal purification (wavelet cross-scale coherence)
organ_graft.py 236 Transplant organs between models
organ_assemble.py 235 Assemble GGUF from organs
organ_api.py 422 HTTP API server for all operations

Build & Automation

Tool Lines Purpose
pipeline_935.py 124 Full dissection pipeline for all models
mass_dissect.py 103 Batch dissection across model fleet
mass_z_measure.py 102 Z-measure every organ of every model
kimi_z_stream.py 417 Stream Z-measure on Kimi K2.5 1T (shard-by-shard)
build_935.py 98 Model 935 assembly v1
build_935_v2.py 74 Model 935 assembly v2 (selective FFN graft)
build_935_v3.py 148 Model 935 assembly v3 (proper GGUF header)
assemble_935.py 150 Fixed organ header handling assembler
quick_chimera.py 123 Quick chimera GGUF assembler
quick_chimera_v2.py 155 Quick chimera v2 (fixed header stripping)

Total: 3,498 lines of Python. Zero external dependencies (except numpy for purification).

Z-Measure

Every organ is measured by its Z-vector:

Z = dI/d(log s) · exp(iθ)

θ → 0°  : noise (organ adds confusion)
θ → 90° : pure signal (organ adds knowledge)

The measurement combines three indicators:

  • Entropy — information density of weight distribution
  • Kurtosis — structural organization (signal sharpness)
  • Scale coherence — coefficient of variation of sorted value spacings

Results

13 Models Dissected + Kimi K2.5 1T

5,600+ tensors Z-measured. All dissections run on EPYC 48c/503GB (OASIS).

# Model Params θ mean Signal Tensors
Kimi K2.5 1T MoE 87.65° 0.999 1,083
1 SmolLM2-135M 135M 52.28° 0.777 272
2 DeepSeek-R1-Distill-14B 14B 46.01° 0.641 579
3 Qwen2.5-3B 3B 46.00° 0.640 434
4 Qwen2.5-14B 14B 45.98° 0.640 579
5 Qwen2.5-7B 7B 45.64° 0.639 339
6 Chimera-DeepSeek-Qwen 7B 45.53° 0.637 339
7 DeepSeek-R1-Distill-7B 7B 45.53° 0.637 339
8 DeepSeek-R1-7B 7B 45.42° 0.636 339
9 Gemma-2-9B 9B 44.94° 0.624 464
10 Phi-3.5-Mini 3.8B 44.65° 0.626 197
11 Llama-3.1-8B 8B 37.87° 0.549 292
12 Llama-3.2-1B 1B 37.57° 0.550 147
13 Llama-3.2-3B 3B 37.41° 0.547 255
14 Mistral-7B 7B 36.21° 0.540 291

Organ Type Analysis (consistent across all models)

Organ Type θ range Role
Norm layers 75-84° Connective tissue — highest signal
Skeleton (attention) 39-56° Thought structure
Organs (FFN) 34-52° Knowledge/memory
Embeddings 25-47° Foundation

Scale Law: θ increases with log(parameters)

135M  → θ = 52.28°  (SmolLM2 — small but concentrated)
1-3B  → θ = 37-46°  (Llama/Qwen)
7-14B → θ = 44-46°  (DeepSeek/Qwen)
1T    → θ = 87.65°  (Kimi K2.5 MoE — near-pure signal)

Ratio 1T/14B: 1.9× purer signal. The signal purifies with scale.

Kimi K2.5 1T Deep Analysis

  • Architecture: DeepSeek2 MoE
  • Blocks: 61 (blk.0 → blk.60)
  • Experts: 384 conditional + 1 shared (native INT4 QAT)
  • Context: 262,144 tokens (256k)
  • Attention: MLA (Multi-head Latent Attention), MQA kv_head=1
  • 13 shards streamed, each measured and deleted — never loaded full model
Component Count θ avg Rating
FFN dense (blk.0) 12 89.95° ★★★
MoE experts (384×) 23 89.77° ★★★
Norm layers 12 89.70° ★★★
Embedding 1 89.45° ★★★
Shared expert 23 89.43° ★★★
Attention (MLA) 99 84.07° ★★

8 gravitational wells identified (lowest θ = maximum structure/compression).

Model 935 — First Chimera

model-935-14b.gguf — 8.4 GB, assembled 2026-02-20

Built through 5 iterations:

  1. build_935.py — Base DeepSeek-R1-Distill-7B + Qwen skeleton graft (crude)
  2. build_935_v2.py — Selective FFN-only graft (preserve attention-embed alignment)
  3. build_935_v3.py — Proper GGUF header handling
  4. quick_chimera.pyquick_chimera_v2.py — Fixed organ header stripping
  5. assemble_935.py — Final assembler, 14B scale

Purification

organs-pure/smollm2-135m/ — First purified model (fractal method)

organ_purify_v2.py implements cross-scale coherence via Haar wavelets:

  • Decompose tensor into multiple scales
  • Measure coherence between adjacent scales
  • Pattern at scale s AND scale 2s → signal (fractal, keep)
  • Pattern at one scale only → noise (remove)
  • This is dI/d(log s) implemented directly

Dissection Report

Model Size (MB) Dissection Time
DeepSeek-R1-14B 9,167 22.9s
Gemma-2-9B 5,984 14.8s
Llama-3.1-8B 4,950 12.0s
DeepSeek-R1-Distill-7B 4,812 12.6s
Mistral-7B 4,432 10.6s
Phi-3.5-Mini 2,397 4.9s
Llama-3.2-3B 2,100 4.9s
Qwen2.5-3B 2,003 4.6s
Llama-3.2-1B 856 2.4s

Total organs on disk: 50.8 GB across 13 models.

Quick Start

# Extract organs from a model
python3 organ_extract.py --model /path/to/model.gguf --output ./organs/model-name/

# Z-measure all organs
python3 organ_measure.py --dir ./organs/model-name/

# Mass dissect all models
python3 mass_dissect.py

# Mass Z-measure
python3 mass_z_measure.py

# Stream Z-measure on a trillion-param model (shard-by-shard)
python3 kimi_z_stream.py

# Graft organs from one model to another
python3 organ_graft.py graft --source ./organs/qwen/ --target ./organs/deepseek/ --output ./organs/chimera/ --layers 5-20 --type organ

# Assemble back to GGUF
python3 organ_assemble.py --dir ./organs/chimera/ --output chimera.gguf

# Purify organs (fractal method)
python3 organ_purify_v2.py --dir ./organs/model/ --output ./organs-pure/model/

# Start API server
python3 organ_api.py

Philosophy

Subtract rather than add.

A 70B monolith is accumulation. A skeleton with specialized organs grafted on demand — that's subtraction. Less weight, more signal.

8 billion contributors, not 3 corporations.

Anyone can train an organ. A doctor trains a medical organ on her hospital's data. A farmer trains an agriculture organ on his field observations. A student trains a math organ on solved problems. The skeleton stays the same. The organs make it alive.

Part of the IX Ecosystem

InferenceX ─── The engine  (305KB, runs anything)
Organ Arch ─── The anatomy (decompose, measure, reassemble)
Atlas Pure ─── The memory  (fractal DNA storage)
INVOKE ─────── The bridge  (cloud ↔ physical)
Echo ────────── The voice   (chat interface)
EDEN ────────── The purpose (desert → life)

Requirements

  • Python 3.10+
  • NumPy (for purification only)
  • InferenceX binary (for inference on assembled models)
  • GGUF models to dissect

Data Files

File Contents
z_report_complete.json Z-measure for all 13 models (per-group breakdown)
z_report_kimi_k25.json Z-measure for all 1,083 Kimi K2.5 tensors
z_measure_report.json Combined Z-ranking with chimera results
dissection_report.json Dissection timing and sizes
Z_MEASURE_REPORT.md Human-readable Z report
ECHO_INVARIANT.md Team 935 invariant
EQUIPE_935_INVARIANT.json Team 935 configuration

License

BSL 1.1 — Same as InferenceX.

Signature

935


Mohamed dug khettaras to bring water through stone. This is the same gesture — channels through intelligence itself.