Published August 16, 2025 | Version V1
Report Open

EvoTransformer: A Whitepaper on Evolving Neural Architectures for Open-Ended AI

Description

EvoTransformer is a transformer architecture that incorporates evolutionary principles—mutation, selection, and shape-safe weight inheritance—directly into the model’s training and adaptation process. Unlike conventional static architectures, EvoTransformer can evolve its structure both during training and after deployment, enabling continual optimization within fixed compute budgets.

This whitepaper details the conceptual foundations, genome representation, mutation strategies, and speciation mechanisms used in EvoTransformer. We present small-compute experiments on PIQA, HellaSwag, and SST-2 benchmarks, showing competitive quality-per-parameter results compared to established baselines trained under identical compute constraints.

Designed for sovereign, domain-specific, and resource-constrained deployments, EvoTransformer offers a reproducible and adaptable framework for efficient AI. Limitations, including the absence of large-scale pretraining, are openly discussed. Reproducibility resources—such as configuration files, seeds, and logs—are provided to encourage independent verification and further research.

Files

Whitepaper_EvoTransformer.pdf

Files (5.2 MB)

Name Size Download all
md5:5cd74b44df7e81deb1b9fe5d75d9760c
5.2 MB Preview Download