Published August 28, 2025 | Version 0.1
Model Open

NequIP & Allegro Foundation Potentials

Description

These are 'large' NequIP/Allegro models, optimised for speed and accuracy with a greater priority placed on accuracy. The `MP` models are trained only on the `MPTrj` (~1.5M frames) datasets (i.e. matbench-discovery 'compliant'), and so are only recommended for benchmarking and not production work.
The `OAM` modles are pre-trained on the `OMat24` dataset (~101M frames), and fine-tuned on the `sAlex` (~10.5M frames) and `MPTrj` (~1.5M frames) datasets. These are the recommended `NequIP`/`Allegro` models for most applications in inorganic solids, having been trained on the largest available open-access datasets.
 
 
We find the NequIP OAM model to currently lie on the upper-right quadrant of the Pareto front when compared to other leading foundation models (preprint incoming), showing an optimal balance of speed and accuracy.
See `nequip.net` and `matbench-discovery` submission for further details – in particular, for details on including model accelerations, and training config files.

See https://nequip.readthedocs.io/en/latest/guide/training-techniques/fine_tuning.html for details on fine-tuning `NequIP`/`Allegro` models.

Files

Allegro-MP-L-0.1.nequip.zip

Files (651.2 MB)

Name Size Download all
md5:c68d5345e925ee444543d6d4d968161f
153.6 MB Preview Download
md5:0db7f9b3c3a62e74d78b3fcf2973c462
80.7 MB Preview Download
md5:efe91679d16fa01df44fb4384efdd3c7
78.7 MB Preview Download
md5:9700a25ea433aa5cc11aa54cf2e65faf
78.5 MB Preview Download
md5:6d8a897a334cd50d232cbf57550b25f1
259.6 MB Preview Download