docs: professional README — Feb 2026

This commit is contained in:
Inference-X 2026-02-25 17:25:22 +00:00
parent e6f5110179
commit 4be971b42a

View File

@ -1,24 +1,28 @@
<div align="center">
# Organ Architecture # Organ Architecture
**Neural network surgery and checkpoint engineering** **Neural network surgery and checkpoint manipulation for GGUF models**
[Inference-X](https://inference-x.com) · [Community](https://git.inference-x.com/inference-x-community) · [Organ Store](https://git.inference-x.com/inference-x-community/organ-store) [![License: BSL 1.1](https://img.shields.io/badge/License-BSL%201.1-orange.svg)](LICENSE)
[**inference-x.com**](https://inference-x.com) · [**Community**](https://git.inference-x.com/inference-x-community)
</div>
--- ---
Organ Architecture provides tools to analyze, modify, and recombine AI model components. Visualize attention heads, extract MLP layers, transplant capabilities between models — all locally. ## What is Organ Architecture?
## What You Can Do Organ Architecture treats neural networks like biological organisms — with organs (layers, heads, MLP blocks) that can be extracted, analyzed, transplanted, and recombined.
- **Visualize** — Inspect layer activations, attention patterns, embedding geometry ## Features
- **Extract** — Pull specific layers, heads, or modules from any GGUF model
- **Transplant** — Merge capabilities from multiple models
- **Prune** — Remove redundant layers to create smaller, faster models
- **Quantize** — Convert between precisions (F32 → F16 → Q8 → Q4 → Q2)
## Supported Model Families (Feb 2026) - **Model inspection** — visualize layer shapes, attention patterns, weight distributions
- **Selective extraction** — extract specific layers or attention heads as "organs"
Llama 3.x · Qwen 2.5 · DeepSeek R1 · Mistral · Gemma 2 · Phi-4 · Command-R - **Cross-model transplant** — graft organs between compatible GGUF models
- **Quantization surgery** — selectively re-quantize individual layers
- **Architecture diff** — compare two model checkpoints layer-by-layer
## Quick Start ## Quick Start
@ -26,26 +30,27 @@ Llama 3.x · Qwen 2.5 · DeepSeek R1 · Mistral · Gemma 2 · Phi-4 · Command-R
git clone https://git.inference-x.com/inference-x/organ-architecture git clone https://git.inference-x.com/inference-x/organ-architecture
cd organ-architecture && make cd organ-architecture && make
# Analyze a model # Inspect a model
./organ analyze model.gguf ./organ inspect model.gguf
# Visualize attention patterns # Extract attention layers 16-24
./organ visualize model.gguf --layer 16 --head 0 ./organ extract model.gguf --layers 16-24 --output head_block.gguf
# Extract layers 0-20 # Transplant into another model
./organ extract model.gguf --layers 0-20 --output model_small.gguf ./organ transplant donor.gguf recipient.gguf --source-layers 16-24 --target-layers 16-24
# Merge two models (50/50)
./organ merge base.gguf specialist.gguf --ratio 0.5 --output merged.gguf
# Quantize F16 → Q4_K_M
./organ quantize model.f16.gguf --target Q4_K_M
``` ```
## Organ Store Integration ## Supported Operations
Share and download model organs at [git.inference-x.com/inference-x-community/organ-store](https://git.inference-x.com/inference-x-community/organ-store). | Operation | Description |
|---|---|
| `inspect` | Show model architecture, layer shapes, parameter count |
| `extract` | Pull specific organs (layers, heads, MLPs) |
| `transplant` | Merge organs from one model into another |
| `diff` | Compare two models layer by layer |
| `requant` | Re-quantize selected layers |
| `prune` | Remove layers to create smaller variants |
--- ---
[inference-x.com](https://inference-x.com) · BSL 1.1 *Part of the [Inference-X ecosystem](https://inference-x.com) · See also: [organ-store](https://git.inference-x.com/inference-x-community/organ-store), [organ-architect](https://git.inference-x.com/inference-x-community/organ-architect)*