Darwin-27B-Opus ranks #6 on GPQA Diamond without training via evolutionary model merging
Darwin Family framework merges existing LLM checkpoints without gradient training, achieving frontier reasoning scores through adaptive genome recombination and cross-architecture breeding.

Darwin Family is a training-free model-merging framework that reorganizes capabilities already encoded in existing language-model checkpoints to improve reasoning performance without additional training. The flagship Darwin-27B-Opus achieves 86.9% on GPQA Diamond, ranking sixth among 1,252 evaluated models and outperforming its fully trained foundation model.
The framework introduces three core mechanisms: a 14-dimensional adaptive merge genome for fine-grained component- and block-level recombination, MRI-Trust Fusion that balances diagnostic layer-importance signals with evolutionary search through a learnable trust parameter, and an Architecture Mapper enabling cross-architecture breeding between heterogeneous model families like Transformer and Mamba. Models scale from 4B to 35B parameters, with consistent improvements over parent checkpoints across the range.
What stands out
- 01Training-free scaling: Darwin models improve reasoning without gradient-based training by recombining weight-space components from existing checkpoints, eliminating the compute cost of post-training pipelines.
- 02Cross-architecture breeding: The Architecture Mapper allows merging between fundamentally different architectures — Transformer and Mamba components can be combined in a single evolutionary merge.
- 03Recursive evolution: The framework supports multi-generation evolution, where merged models become parents for subsequent merges, enabling iterative refinement.
- 04Diagnostic-guided search: MRI-Trust Fusion adaptively weights layer-importance diagnostics against evolutionary search signals, preventing over-reliance on either heuristic alone.
- 05