The Neural Force Field (NFF) code is an API based on SchNet [1-4], DimeNet [5], PaiNN [6-7] and DANN [8]. It provides an interface to train and evaluate neural networks for force fields. It can also be used as a property predictor that uses both 3D geometries and 2D graph information [9].
This code repository is developed in the Learning Matter Lab (led by prof. Rafael Gomez-Bombarelli) at MIT.
We highly recommend creating a conda
environment to run the code. To do that, use the following command to create the nff
conda environment:
conda upgrade conda
conda env create -f environment.yml
To ensure that the nff
environment is accessible through Jupyter, add the the nff
display name:
python -m ipykernel install --user --name nff --display-name "Python [conda env:nff"]
If you would like to install NFF as a package, you can do so by running
pip install .
Otherwise you can put NFF in your python path by adding the following lines to ~/.bashrc
(linux) or ~/.bash_profile
(mac):
export NFFDIR=<path to NFF>
export PYTHONPATH=$NFFDIR:$PYTHONPATH
This is useful if you'll be modifying the NFF code, because modifications in the download folder won't change anything in the conda directory where it's been installed.
A series of tutorials illustrating how nff
can be used in conjunction with Jupyter Notebooks or other scripts is provided in the tutorials/
folder. It also covers how to integrate a pre-trained model with an ASE calculator, how to perform ground state molecular dynamics (MD) and excited state non-adiabatic MD, and how to train different model types like DimeNet and PaiNN.
All tutorials used pre-saved datasets for training. These datasets are saved as NFF dataset objects. To see how to make your own NFF dataset and save it, see this tutorial in tutorials/data
While scripts/cp3d/README.md
explains in depth how to use the scripts, the notebook 07_cp3d.ipynb
goes into some detail about what happens behind the scenes. In this notebook you'll see how the datasets get made and what the models look like.
The simplest way to use the nff
package is to use the premade scripts (in the scripts
) folder. As an example, to train a SchNet model with the default parameters using the example dataset (ethanol geometries) from the command line, run the command
nff_train.py train schnet tutorials/data/dataset.pth.tar $HOME/train_model --device cuda:0
This will use 60% of the dataset for training, 20% for validation and 20% for testing. The training will happen on the device cuda:0
. Results of training, checkpoints and hyperparameters will be saved on the path $HOME/train_model
.
NFF also contains modules that predict properties from 3D geometries of conformers. These include the SchNet model, expanded to include multiple conformers, as well as the ChemProp3D (CP3D) model, which also includes graph information. A series of scripts for these modules can be found in scripts/cp3d
. An in-depth discussion of how to use these scripts can be found in scripts/cp3d/README.md
.
A set of pre-trained models can be found in models
.
NFF allows the usage of NN ensembles to perform uncertainty quantification and adversarial sampling of geometries. The complete tutorials on how to perform such analysis is available at the Atomistic Adversarial Attacks repository, and the theory behind this differentiable sampling strategy is available at our paper [10].
-
[1] K.T. Schütt. F. Arbabzadah. S. Chmiela, K.-R. Müller, A. Tkatchenko.
Quantum-chemical insights from deep tensor neural networks. Nature Communications 8. 13890 (2017)
10.1038/ncomms13890 -
[2] K.T. Schütt. P.-J. Kindermans, H. E. Sauceda, S. Chmiela, A. Tkatchenko, K.-R. Müller.
SchNet: A continuous-filter convolutional neural network for modeling quantum interactions. Advances in Neural Information Processing Systems 30, pp. 992-1002 (2017) link -
[3] K.T. Schütt. P.-J. Kindermans, H. E. Sauceda, S. Chmiela, A. Tkatchenko, K.-R. Müller.
SchNet - a deep learning architecture for molecules and materials. The Journal of Chemical Physics 148(24), 241722 (2018) 10.1063/1.5019779 -
[4] K.T. Schütt, P. Kessel, M. Gastegger, K. Nicoli, A. Tkatchenko, K.-R. Müller. SchNetPack: A Deep Learning Toolbox For Atomistic Systems. J. Chem. Theory Comput. 15(1), 448-455 (2019). 10.1021/acs.jctc.8b00908
-
[5] J. Klicpera, G. Janek, S. Günnemann. Directional message passing for molecular graphs. ICLR (2020). URL.
-
[6] K. T. Schütt, O. T. Unke, M. Gastegger. Equivariant message passing for the prediction of tensorial properties and molecular spectra. arXiv preprint, 2021. arXiv:2102.03150
-
[7] S. Axelrod, E. Shakhnovich, R. Gómez-Bombarelli. Thermal half-lives of azobenzene derivatives: virtual screening based on intersystem crossing using a machine learning potential. arXiv preprint (2022). arXiv:2207.11592.
-
[8] S. Axelrod, E. Shakhnovich, R. Gómez-Bombarelli. Excited state non-adiabatic dynamics of large photoswitchable molecules using a chemically transferable machine learning potential. Nat. Commun. 13, 3440 (2022). URL
-
[9] S. Axelrod and R. Gomez-Bombarelli. Molecular machine learning with conformer ensembles. arXiv preprint (2020). arXiv:2012.08452.
-
[10] D. Schwalbe-Koda, A.R. Tan, and R. Gomez-Bombarelli. Differentiable sampling of molecular geometries with uncertainty-based adversarial attacks. Nat. Commun. 12, 5104 (2021). URL.