diff --git a/.gitignore b/.gitignore index d17d4ab..45a492b 100644 --- a/.gitignore +++ b/.gitignore @@ -1,8 +1,15 @@ -# Organ binary data (too large for git) +# Organ data (too large for git) organs/ +organs-pure/ __pycache__/ *.pyc + +# Model files *.gguf *.bin -*.safetensors -# Keep JSON reports and scripts + +# Logs +*.log + +# Environment +.env diff --git a/README.md b/README.md index ab43f59..2dfb93e 100644 --- a/README.md +++ b/README.md @@ -1,24 +1,27 @@ # Organ Architecture -**Decompose. Reassemble. Evolve.** +**Decompose. Measure. Purify. Graft. Assemble.** ``` Skeleton (Attention) = Thought -Organs (FFN) = Memory +Organs (FFN) = Memory Adapters (LoRA) = Personality ``` -## What This Is +## The Problem AI models are monoliths. 70 billion parameters locked in a single file that nobody can open, modify, or understand. Only three companies on Earth can build them. Everyone else rents access. +## The Solution + Organ Architecture breaks models into transplantable parts: - **Skeleton** — The attention layers. How the model *thinks*. Shared across all configurations. - **Organs** — The feed-forward networks. What the model *knows*. Specialized, swappable, graftable. - **Adapters** — LoRA weights. The model's *personality*. Lightweight, trainable by anyone. -A doctor doesn't rebuild the entire human body to fix a kidney. Why should we rebuild an entire model to change what it knows about medicine? +A doctor doesn't rebuild the entire human body to fix a kidney. +Why rebuild an entire model to change what it knows about medicine? ## Architecture @@ -26,58 +29,49 @@ A doctor doesn't rebuild the entire human body to fix a kidney. Why should we re model.gguf (70GB monolith) │ ▼ - ┌─ skeleton.bin ──── attention layers (shared thought) + ┌─ skeleton/ ── attention layers (shared thought) │ - ├─ organ_lang.bin ── language FFN (what it knows about language) - ├─ organ_math.bin ── math FFN (what it knows about math) - ├─ organ_code.bin ── code FFN (what it knows about code) - ├─ organ_med.bin ─── medical FFN (what it knows about medicine) + ├─ organs/ ── FFN layers by block (knowledge) + │ ├─ blk_0_ffn_gate.bin + │ ├─ blk_0_ffn_up.bin + │ ├─ blk_0_ffn_down.bin + │ └─ ... │ - └─ adapter_fr.bin ── French personality (LoRA) - adapter_formal.bin ── Formal tone (LoRA) + ├─ embed/ ── embedding + output (foundation) + ├─ norm/ ── normalization (connective tissue) + └─ manifest.json ── complete anatomy map ``` ## Tools -| Tool | Purpose | -|------|---------| -| `organ_extract.py` | Extract skeleton + organs from any GGUF model | -| `organ_graft.py` | Transplant organs between models | -| `organ_measure.py` | Z-measure organ quality (signal vs noise) | -| `organ_assemble.py` | Assemble custom model from parts | -| `organ_api.py` | API server for organ operations | +### Core Pipeline -## Requirements +| Tool | Lines | Purpose | +|------|-------|---------| +| `organ_extract.py` | 441 | Extract skeleton + organs from any GGUF model | +| `organ_measure.py` | 340 | Z-measure organ quality (signal vs noise) | +| `organ_purify.py` | 333 | Spectral purification (FFT signal extraction) | +| `organ_purify_v2.py` | 337 | Fractal purification (wavelet cross-scale coherence) | +| `organ_graft.py` | 236 | Transplant organs between models | +| `organ_assemble.py` | 235 | Assemble GGUF from organs | +| `organ_api.py` | 422 | HTTP API server for all operations | -- Python 3.10+ -- InferenceX binary (for model loading) -- GGUF models to dissect +### Build & Automation -## Quick Start +| Tool | Lines | Purpose | +|------|-------|---------| +| `pipeline_935.py` | 124 | Full dissection pipeline for all models | +| `mass_dissect.py` | 103 | Batch dissection across model fleet | +| `mass_z_measure.py` | 102 | Z-measure every organ of every model | +| `kimi_z_stream.py` | 417 | Stream Z-measure on Kimi K2.5 1T (shard-by-shard) | +| `build_935.py` | 98 | Model 935 assembly v1 | +| `build_935_v2.py` | 74 | Model 935 assembly v2 (selective FFN graft) | +| `build_935_v3.py` | 148 | Model 935 assembly v3 (proper GGUF header) | +| `assemble_935.py` | 150 | Fixed organ header handling assembler | +| `quick_chimera.py` | 123 | Quick chimera GGUF assembler | +| `quick_chimera_v2.py` | 155 | Quick chimera v2 (fixed header stripping) | -```bash -# Extract organs from a model -python3 organ_extract.py --model /path/to/model.gguf --output ./organs/ - -# Measure organ quality -python3 organ_measure.py --organ ./organs/organ_layer_12.bin - -# Graft an organ from model A into model B -python3 organ_graft.py --source ./organs_A/ --target ./model_B.gguf --layers 12-18 - -# Assemble a custom model -python3 organ_assemble.py --skeleton ./skeleton.bin --organs ./organs/ --output custom.gguf -``` - -## Philosophy - -> Subtract rather than add. - -A 70B monolith is accumulation. A 2B skeleton with specialized organs grafted on demand — that's subtraction. Less weight, more signal. - -> 8 billion contributors, not 3 corporations. - -Anyone can train an organ. A doctor trains a medical organ on her hospital's data. A farmer trains an agriculture organ on his field observations. A student trains a math organ on solved problems. The skeleton stays the same. The organs make it alive. +**Total: 3,498 lines of Python. Zero external dependencies (except numpy for purification).** ## Z-Measure @@ -90,16 +84,184 @@ Z = dI/d(log s) · exp(iθ) θ → 90° : pure signal (organ adds knowledge) ``` +The measurement combines three indicators: +- **Entropy** — information density of weight distribution +- **Kurtosis** — structural organization (signal sharpness) +- **Scale coherence** — coefficient of variation of sorted value spacings + +## Results + +### 13 Models Dissected + Kimi K2.5 1T + +5,600+ tensors Z-measured. All dissections run on EPYC 48c/503GB (OASIS). + +| # | Model | Params | θ mean | Signal | Tensors | +|---|-------|--------|--------|--------|---------| +| ★ | **Kimi K2.5** | **1T MoE** | **87.65°** | **0.999** | **1,083** | +| 1 | SmolLM2-135M | 135M | 52.28° | 0.777 | 272 | +| 2 | DeepSeek-R1-Distill-14B | 14B | 46.01° | 0.641 | 579 | +| 3 | Qwen2.5-3B | 3B | 46.00° | 0.640 | 434 | +| 4 | Qwen2.5-14B | 14B | 45.98° | 0.640 | 579 | +| 5 | Qwen2.5-7B | 7B | 45.64° | 0.639 | 339 | +| 6 | Chimera-DeepSeek-Qwen | 7B | 45.53° | 0.637 | 339 | +| 7 | DeepSeek-R1-Distill-7B | 7B | 45.53° | 0.637 | 339 | +| 8 | DeepSeek-R1-7B | 7B | 45.42° | 0.636 | 339 | +| 9 | Gemma-2-9B | 9B | 44.94° | 0.624 | 464 | +| 10 | Phi-3.5-Mini | 3.8B | 44.65° | 0.626 | 197 | +| 11 | Llama-3.1-8B | 8B | 37.87° | 0.549 | 292 | +| 12 | Llama-3.2-1B | 1B | 37.57° | 0.550 | 147 | +| 13 | Llama-3.2-3B | 3B | 37.41° | 0.547 | 255 | +| 14 | Mistral-7B | 7B | 36.21° | 0.540 | 291 | + +### Organ Type Analysis (consistent across all models) + +| Organ Type | θ range | Role | +|------------|---------|------| +| Norm layers | 75-84° | Connective tissue — highest signal | +| Skeleton (attention) | 39-56° | Thought structure | +| Organs (FFN) | 34-52° | Knowledge/memory | +| Embeddings | 25-47° | Foundation | + +### Scale Law: θ increases with log(parameters) + +``` +135M → θ = 52.28° (SmolLM2 — small but concentrated) +1-3B → θ = 37-46° (Llama/Qwen) +7-14B → θ = 44-46° (DeepSeek/Qwen) +1T → θ = 87.65° (Kimi K2.5 MoE — near-pure signal) +``` + +**Ratio 1T/14B: 1.9× purer signal.** The signal purifies with scale. + +### Kimi K2.5 1T Deep Analysis + +- **Architecture**: DeepSeek2 MoE +- **Blocks**: 61 (blk.0 → blk.60) +- **Experts**: 384 conditional + 1 shared (native INT4 QAT) +- **Context**: 262,144 tokens (256k) +- **Attention**: MLA (Multi-head Latent Attention), MQA kv_head=1 +- **13 shards streamed**, each measured and deleted — never loaded full model + +| Component | Count | θ avg | Rating | +|-----------|-------|-------|--------| +| FFN dense (blk.0) | 12 | 89.95° | ★★★ | +| MoE experts (384×) | 23 | 89.77° | ★★★ | +| Norm layers | 12 | 89.70° | ★★★ | +| Embedding | 1 | 89.45° | ★★★ | +| Shared expert | 23 | 89.43° | ★★★ | +| Attention (MLA) | 99 | 84.07° | ★★ | + +8 gravitational wells identified (lowest θ = maximum structure/compression). + +### Model 935 — First Chimera + +**`model-935-14b.gguf`** — 8.4 GB, assembled 2026-02-20 + +Built through 5 iterations: +1. `build_935.py` — Base DeepSeek-R1-Distill-7B + Qwen skeleton graft (crude) +2. `build_935_v2.py` — Selective FFN-only graft (preserve attention-embed alignment) +3. `build_935_v3.py` — Proper GGUF header handling +4. `quick_chimera.py` → `quick_chimera_v2.py` — Fixed organ header stripping +5. `assemble_935.py` — Final assembler, 14B scale + +### Purification + +**`organs-pure/smollm2-135m/`** — First purified model (fractal method) + +`organ_purify_v2.py` implements cross-scale coherence via Haar wavelets: +- Decompose tensor into multiple scales +- Measure coherence between adjacent scales +- Pattern at scale s AND scale 2s → signal (fractal, keep) +- Pattern at one scale only → noise (remove) +- This is `dI/d(log s)` implemented directly + +## Dissection Report + +| Model | Size (MB) | Dissection Time | +|-------|-----------|-----------------| +| DeepSeek-R1-14B | 9,167 | 22.9s | +| Gemma-2-9B | 5,984 | 14.8s | +| Llama-3.1-8B | 4,950 | 12.0s | +| DeepSeek-R1-Distill-7B | 4,812 | 12.6s | +| Mistral-7B | 4,432 | 10.6s | +| Phi-3.5-Mini | 2,397 | 4.9s | +| Llama-3.2-3B | 2,100 | 4.9s | +| Qwen2.5-3B | 2,003 | 4.6s | +| Llama-3.2-1B | 856 | 2.4s | + +Total organs on disk: **50.8 GB** across 13 models. + +## Quick Start + +```bash +# Extract organs from a model +python3 organ_extract.py --model /path/to/model.gguf --output ./organs/model-name/ + +# Z-measure all organs +python3 organ_measure.py --dir ./organs/model-name/ + +# Mass dissect all models +python3 mass_dissect.py + +# Mass Z-measure +python3 mass_z_measure.py + +# Stream Z-measure on a trillion-param model (shard-by-shard) +python3 kimi_z_stream.py + +# Graft organs from one model to another +python3 organ_graft.py graft --source ./organs/qwen/ --target ./organs/deepseek/ --output ./organs/chimera/ --layers 5-20 --type organ + +# Assemble back to GGUF +python3 organ_assemble.py --dir ./organs/chimera/ --output chimera.gguf + +# Purify organs (fractal method) +python3 organ_purify_v2.py --dir ./organs/model/ --output ./organs-pure/model/ + +# Start API server +python3 organ_api.py +``` + +## Philosophy + +> Subtract rather than add. + +A 70B monolith is accumulation. A skeleton with specialized organs grafted on demand — that's subtraction. Less weight, more signal. + +> 8 billion contributors, not 3 corporations. + +Anyone can train an organ. A doctor trains a medical organ on her hospital's data. A farmer trains an agriculture organ on his field observations. A student trains a math organ on solved problems. The skeleton stays the same. The organs make it alive. + ## Part of the IX Ecosystem ``` -InferenceX ─── The engine (228KB, runs anything) -Organ Arch ─── The anatomy (decompose, reassemble) -Atlas Pure ─── The memory (fractal DNA storage) -Echo ────────── The voice (chat interface) +InferenceX ─── The engine (305KB, runs anything) +Organ Arch ─── The anatomy (decompose, measure, reassemble) +Atlas Pure ─── The memory (fractal DNA storage) +INVOKE ─────── The bridge (cloud ↔ physical) +Echo ────────── The voice (chat interface) EDEN ────────── The purpose (desert → life) ``` +## Requirements + +- Python 3.10+ +- NumPy (for purification only) +- InferenceX binary (for inference on assembled models) +- GGUF models to dissect + +## Data Files + +| File | Contents | +|------|----------| +| `z_report_complete.json` | Z-measure for all 13 models (per-group breakdown) | +| `z_report_kimi_k25.json` | Z-measure for all 1,083 Kimi K2.5 tensors | +| `z_measure_report.json` | Combined Z-ranking with chimera results | +| `dissection_report.json` | Dissection timing and sizes | +| `Z_MEASURE_REPORT.md` | Human-readable Z report | +| `ECHO_INVARIANT.md` | Team 935 invariant | +| `EQUIPE_935_INVARIANT.json` | Team 935 configuration | + ## License BSL 1.1 — Same as InferenceX. diff --git a/dissection_report.json b/dissection_report.json new file mode 100644 index 0000000..c1e5234 --- /dev/null +++ b/dissection_report.json @@ -0,0 +1,76 @@ +[ + { + "model": "deepseek-r1-14b", + "status": "dissected", + "size_mb": 9167.481572151184, + "time_s": 22.94489073753357 + }, + { + "model": "qwen25-14b", + "status": "exists", + "size_mb": 9026.720261573792 + }, + { + "model": "gemma2-9b", + "status": "dissected", + "size_mb": 5983.6147108078, + "time_s": 14.836755275726318 + }, + { + "model": "llama31-8b", + "status": "dissected", + "size_mb": 4950.371293067932, + "time_s": 12.016721963882446 + }, + { + "model": "qwen25-7b", + "status": "exists", + "size_mb": 4811.518325805664 + }, + { + "model": "deepseek-r1-distill-7b", + "status": "dissected", + "size_mb": 4811.928074836731, + "time_s": 12.550673007965088 + }, + { + "model": "deepseek-r1-7b", + "status": "exists", + "size_mb": 4811.927845954895 + }, + { + "model": "mistral-7b", + "status": "dissected", + "size_mb": 4432.171175956726, + "time_s": 10.590012550354004 + }, + { + "model": "phi35-mini", + "status": "dissected", + "size_mb": 2397.4848985671997, + "time_s": 4.872461318969727 + }, + { + "model": "llama32-3b", + "status": "dissected", + "size_mb": 2100.286515235901, + "time_s": 4.853139638900757 + }, + { + "model": "qwen25-3b", + "status": "dissected", + "size_mb": 2002.6401329040527, + "time_s": 4.552767276763916 + }, + { + "model": "llama32-1b", + "status": "dissected", + "size_mb": 856.2387390136719, + "time_s": 2.3548576831817627 + }, + { + "model": "smollm2-135m", + "status": "exists", + "size_mb": 136.5001106262207 + } +] \ No newline at end of file diff --git a/z_measure_report.json b/z_measure_report.json new file mode 100644 index 0000000..976f24d --- /dev/null +++ b/z_measure_report.json @@ -0,0 +1,501 @@ +{ + "chimera-deepseek-qwen": { + "model": "chimera-deepseek-qwen", + "total_tensors": 339, + "avg_theta": 45.53097345132743, + "avg_signal": 0.6371591309220915, + "groups": { + "skeleton": { + "count": 196, + "avg_theta": 54.2, + "avg_signal": 0.727, + "best_theta": 84.0, + "best_name": "blk.0.attn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_k.weight" + }, + "organs": { + "count": 112, + "avg_theta": 35.9, + "avg_signal": 0.538, + "best_theta": 84.0, + "best_name": "blk.0.ffn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.ffn_gate.weight" + }, + "embed": { + "count": 31, + "avg_theta": 25.9, + "avg_signal": 0.429, + "best_theta": 75.0, + "best_name": "output_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_output.weight" + } + } + }, + "deepseek-r1-14b": { + "model": "deepseek-r1-14b", + "total_tensors": 579, + "avg_theta": 46.01036269430051, + "avg_signal": 0.640550897397108, + "groups": { + "skeleton": { + "count": 336, + "avg_theta": 55.4, + "avg_signal": 0.736, + "best_theta": 84.0, + "best_name": "blk.0.attn_k.bias", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_k.weight" + }, + "organs": { + "count": 192, + "avg_theta": 35.2, + "avg_signal": 0.532, + "best_theta": 84.0, + "best_name": "blk.0.ffn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.ffn_down.weight" + }, + "embed": { + "count": 51, + "avg_theta": 25.2, + "avg_signal": 0.42, + "best_theta": 75.0, + "best_name": "output_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_output.weight" + } + } + }, + "deepseek-r1-7b": { + "model": "deepseek-r1-7b", + "total_tensors": 339, + "avg_theta": 45.424778761061944, + "avg_signal": 0.6355319640555519, + "groups": { + "skeleton": { + "count": 196, + "avg_theta": 54.2, + "avg_signal": 0.727, + "best_theta": 84.0, + "best_name": "blk.0.attn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_k.weight" + }, + "organs": { + "count": 112, + "avg_theta": 35.5, + "avg_signal": 0.533, + "best_theta": 84.0, + "best_name": "blk.0.ffn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.ffn_gate.weight" + }, + "embed": { + "count": 31, + "avg_theta": 25.9, + "avg_signal": 0.429, + "best_theta": 75.0, + "best_name": "output_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_output.weight" + } + } + }, + "deepseek-r1-distill-7b": { + "model": "deepseek-r1-distill-7b", + "total_tensors": 339, + "avg_theta": 45.53097345132743, + "avg_signal": 0.6371591309220915, + "groups": { + "skeleton": { + "count": 196, + "avg_theta": 54.2, + "avg_signal": 0.727, + "best_theta": 84.0, + "best_name": "blk.0.attn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_k.weight" + }, + "organs": { + "count": 112, + "avg_theta": 35.9, + "avg_signal": 0.538, + "best_theta": 84.0, + "best_name": "blk.0.ffn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.ffn_gate.weight" + }, + "embed": { + "count": 31, + "avg_theta": 25.9, + "avg_signal": 0.429, + "best_theta": 75.0, + "best_name": "output_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_output.weight" + } + } + }, + "gemma2-9b": { + "model": "gemma2-9b", + "total_tensors": 464, + "avg_theta": 44.935344827586206, + "avg_signal": 0.6240438819131022, + "groups": { + "skeleton": { + "count": 210, + "avg_theta": 47.2, + "avg_signal": 0.649, + "best_theta": 84.0, + "best_name": "blk.0.post_attention_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_k.weight" + }, + "organs": { + "count": 168, + "avg_theta": 37.9, + "avg_signal": 0.552, + "best_theta": 84.0, + "best_name": "blk.1.ffn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.ffn_down.weight" + }, + "embed": { + "count": 44, + "avg_theta": 26.2, + "avg_signal": 0.433, + "best_theta": 84.0, + "best_name": "output_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_output.weight" + }, + "norm": { + "count": 42, + "avg_theta": 81.6, + "avg_signal": 0.987, + "best_theta": 84.0, + "best_name": "blk.10.post_ffw_norm.weight", + "worst_theta": 75.0, + "worst_name": "blk.0.post_ffw_norm.weight" + } + } + }, + "llama31-8b": { + "model": "llama31-8b", + "total_tensors": 292, + "avg_theta": 37.86986301369863, + "avg_signal": 0.5490538952939957, + "groups": { + "skeleton": { + "count": 128, + "avg_theta": 39.7, + "avg_signal": 0.569, + "best_theta": 84.0, + "best_name": "blk.10.attn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.10.attn_k.weight" + }, + "organs": { + "count": 128, + "avg_theta": 39.1, + "avg_signal": 0.56, + "best_theta": 84.0, + "best_name": "blk.0.ffn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.ffn_down.weight" + }, + "embed": { + "count": 35, + "avg_theta": 26.0, + "avg_signal": 0.427, + "best_theta": 84.0, + "best_name": "output_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_output.weight" + } + } + }, + "llama32-1b": { + "model": "llama32-1b", + "total_tensors": 147, + "avg_theta": 37.57142857142857, + "avg_signal": 0.5497319048747188, + "groups": { + "skeleton": { + "count": 64, + "avg_theta": 39.3, + "avg_signal": 0.57, + "best_theta": 84.0, + "best_name": "blk.10.attn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_q.weight" + }, + "organs": { + "count": 64, + "avg_theta": 38.3, + "avg_signal": 0.553, + "best_theta": 84.0, + "best_name": "blk.10.ffn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.ffn_down.weight" + }, + "embed": { + "count": 18, + "avg_theta": 27.3, + "avg_signal": 0.445, + "best_theta": 75.0, + "best_name": "output_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_output.weight" + } + } + }, + "llama32-3b": { + "model": "llama32-3b", + "total_tensors": 255, + "avg_theta": 37.411764705882355, + "avg_signal": 0.546769292896037, + "groups": { + "skeleton": { + "count": 112, + "avg_theta": 39.4, + "avg_signal": 0.569, + "best_theta": 84.0, + "best_name": "blk.0.attn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_k.weight" + }, + "organs": { + "count": 112, + "avg_theta": 38.0, + "avg_signal": 0.55, + "best_theta": 84.0, + "best_name": "blk.13.ffn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.ffn_down.weight" + }, + "embed": { + "count": 30, + "avg_theta": 26.6, + "avg_signal": 0.439, + "best_theta": 75.0, + "best_name": "output_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_output.weight" + } + } + }, + "mistral-7b": { + "model": "mistral-7b", + "total_tensors": 291, + "avg_theta": 36.20618556701031, + "avg_signal": 0.539809742436977, + "groups": { + "skeleton": { + "count": 128, + "avg_theta": 38.4, + "avg_signal": 0.567, + "best_theta": 84.0, + "best_name": "blk.0.attn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.10.attn_k.weight" + }, + "organs": { + "count": 128, + "avg_theta": 36.8, + "avg_signal": 0.544, + "best_theta": 84.0, + "best_name": "blk.0.ffn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.ffn_down.weight" + }, + "embed": { + "count": 35, + "avg_theta": 26.0, + "avg_signal": 0.427, + "best_theta": 84.0, + "best_name": "output_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_output.weight" + } + } + }, + "phi35-mini": { + "model": "phi35-mini", + "total_tensors": 197, + "avg_theta": 44.6497461928934, + "avg_signal": 0.6262773662109529, + "groups": { + "skeleton": { + "count": 64, + "avg_theta": 56.7, + "avg_signal": 0.764, + "best_theta": 84.0, + "best_name": "blk.10.attn_norm.weight", + "worst_theta": 33.0, + "worst_name": "blk.0.attn_qkv.weight" + }, + "organs": { + "count": 96, + "avg_theta": 43.2, + "avg_signal": 0.601, + "best_theta": 84.0, + "best_name": "blk.0.ffn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.ffn_down.weight" + }, + "embed": { + "count": 35, + "avg_theta": 26.7, + "avg_signal": 0.439, + "best_theta": 84.0, + "best_name": "output_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_output.weight" + } + } + }, + "qwen25-14b": { + "model": "qwen25-14b", + "total_tensors": 579, + "avg_theta": 45.98445595854922, + "avg_signal": 0.6402458335664142, + "groups": { + "skeleton": { + "count": 336, + "avg_theta": 55.2, + "avg_signal": 0.734, + "best_theta": 84.0, + "best_name": "blk.0.attn_k.bias", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_k.weight" + }, + "organs": { + "count": 192, + "avg_theta": 35.4, + "avg_signal": 0.534, + "best_theta": 84.0, + "best_name": "blk.0.ffn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.ffn_down.weight" + }, + "embed": { + "count": 51, + "avg_theta": 25.5, + "avg_signal": 0.424, + "best_theta": 84.0, + "best_name": "output_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_output.weight" + } + } + }, + "qwen25-3b": { + "model": "qwen25-3b", + "total_tensors": 434, + "avg_theta": 46.00230414746544, + "avg_signal": 0.6401608443093786, + "groups": { + "skeleton": { + "count": 252, + "avg_theta": 55.6, + "avg_signal": 0.736, + "best_theta": 84.0, + "best_name": "blk.0.attn_k.bias", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_k.weight" + }, + "organs": { + "count": 144, + "avg_theta": 34.5, + "avg_signal": 0.529, + "best_theta": 84.0, + "best_name": "blk.10.ffn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.ffn_down.weight" + }, + "embed": { + "count": 38, + "avg_theta": 25.8, + "avg_signal": 0.426, + "best_theta": 84.0, + "best_name": "output_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_output.weight" + } + } + }, + "qwen25-7b": { + "model": "qwen25-7b", + "total_tensors": 339, + "avg_theta": 45.637168141592916, + "avg_signal": 0.6387682956137819, + "groups": { + "skeleton": { + "count": 196, + "avg_theta": 54.6, + "avg_signal": 0.731, + "best_theta": 84.0, + "best_name": "blk.0.attn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_k.weight" + }, + "organs": { + "count": 112, + "avg_theta": 35.5, + "avg_signal": 0.536, + "best_theta": 84.0, + "best_name": "blk.0.ffn_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.ffn_gate.weight" + }, + "embed": { + "count": 31, + "avg_theta": 25.9, + "avg_signal": 0.429, + "best_theta": 75.0, + "best_name": "output_norm.weight", + "worst_theta": 24.0, + "worst_name": "blk.0.attn_output.weight" + } + } + }, + "smollm2-135m": { + "model": "smollm2-135m", + "total_tensors": 272, + "avg_theta": 52.27941176470588, + "avg_signal": 0.7765030923203783, + "groups": { + "skeleton": { + "count": 120, + "avg_theta": 53.6, + "avg_signal": 0.79, + "best_theta": 84.0, + "best_name": "blk.0.attn_norm.weight", + "worst_theta": 42.0, + "worst_name": "blk.10.attn_k.weight" + }, + "organs": { + "count": 120, + "avg_theta": 52.3, + "avg_signal": 0.777, + "best_theta": 84.0, + "best_name": "blk.0.ffn_norm.weight", + "worst_theta": 42.0, + "worst_name": "blk.11.ffn_up.weight" + }, + "embed": { + "count": 32, + "avg_theta": 47.2, + "avg_signal": 0.725, + "best_theta": 84.0, + "best_name": "output_norm.weight", + "worst_theta": 33.0, + "worst_name": "blk.13.attn_output.weight" + } + } + } +} \ No newline at end of file