LAMMPS¶
Official website | How is metatomic supported? |
|---|---|
In a separate fork |
Supported model outputs¶
The energy, non-conservative forces and stress outputs are supported in LAMMPS, as a custom pair_style. This allows running molecular dynamics simulations with interatomic potentials in the metatomic format; distributing the simulation over multiple nodes and potentially multiple GPUs.
How to install the code¶
Getting a pre-built binary with conda¶
The easiest way to install a version of lammps which can use metatomic models is to use the build we provide through conda. We recommend that you use miniforge as your conda provider.
First you’ll need to pick an MPI implementation, from openmpi, mpich or nompi (which does not have MPI enabled). If you’d like to use the MPI library from your system (for example when running on supercomputers with specific MPI tuning), please follow these instructions: https://conda-forge.org/docs/user/tipsandtricks/#using-external-message-passing-interface-mpi-libraries
You can then install LAMMPS with:
# for example with nompi conda install -c metatensor -c conda-forge "lammps-metatomic=*=*nompi*" # or with openmpi conda install -c metatensor -c conda-forge "lammps-metatomic=*=*openmpi*" This version of LAMMPS will be able to run the models on CPU or GPU, but will run the time integration of the trajectory on CPU. If you also want to run the time integration on GPU, you’ll need to install the kokkos-enabled build of LAMMPS. This build currently only exists for CUDA GPUs, and each build only supports a single GPU architecture.
To get the correct KOKKOS build for your GPU, you’ll first need to determine its compute capability. For this, you can run the following command with a GPU present (i.e. on the compute node of supercomputers):
nvidia-smi --query-gpu=compute_cap --format=csv,noheader Alternatively, you can also get the compute capability from NVIDIA’s documentation
We currenly build the code for the following compute capabilities:
VOLTA70AMPERE80AMPERE86ADA89HOPPER90
For example, if you have a NVIDIA A100 GPU, it’s compute capability is 8.0 (i.e. AMPERE80).
Now that you know the compute capability, you can install the correct kokkos build (here as well, you can pick between different MPI implementations).
conda install -c metatensor -c conda-forge "lammps-metatomic=*=cuda*AMPERE80*nompi*" Warning
Be aware that some HPC clusters may be set up without NVIDIA drivers installed on the head/login node. This will result in conda not detecting the system configuration of the compute nodes (which probably have GPUs if you are in this section) and will not install correct torch and cuda libraries. To fix it, you should run the install from a GPU node (check that after running conda info on the node your installing from, if __cuda is in the virtual packages section).
You can also trick conda into installing cuda enabled versions on a login node without NVIDIA drivers, by setting the environment variable CONDA_OVERRIDE_CUDA to the correct CUDA version:
CONDA_OVERRIDE_CUDA=12.4 conda install -c metatensor -c conda-forge "lammps-metatomic=*=cuda*AMPERE80*nompi*" Note
If you get the following error
Kokkos::Cuda::initialize ERROR: likely mismatch of architecture you are likely using the wrong KOKKOS build. Please double check that the compute capabilities of your GPU match the build you used.
Building from sources¶
The code is available in a custom fork of LAMMPS, and you can get it with
git clone https://github.com/metatensor/lammps lammps-metatomic cd lammps-metatomic You’ll need to provide some of the code dependencies yourself. There are multiple ways to go about it, here we detail a fully manual installation, an installation using conda and an installation using pip.
Option 1: dependencies from conda¶
All the dependencies of the code are available on conda, you can install them with
# create an environment (you can also re-use an existing one) conda create -n lammps-metatomic conda activate lammps-metatomic conda install -c metatensor -c conda-forge libmetatomic-torch # Store this information to configure cmake down the line CMAKE_PREFIX_PATH="$CONDA_PREFIX" Option 2: dependencies from pip¶
All the dependencies of the code are also available on PyPI, you can install them with
# (optional) create an environment with your preferred method ... python -m pip install metatomic-torch # on linux, if you don't have a GPU available, you should force the use of # CPU-only torch instead python -m pip install --extra-index-url=https://download.pytorch.org/whl/cpu metatomic-torch # Get the information to configure cmake down the line TORCH_PREFIX=$(python -c "import torch; print(torch.utils.cmake_prefix_path)") MTS_PREFIX=$(python -c "import metatensor; print(metatensor.utils.cmake_prefix_path)") MTS_TORCH_PREFIX=$(python -c "import metatensor.torch; print(metatensor.torch.utils.cmake_prefix_path)") MTA_TORCH_PREFIX=$(python -c "import metatomic.torch; print(metatomic.torch.utils.cmake_prefix_path)") CMAKE_PREFIX_PATH="$TORCH_PREFIX;$MTS_PREFIX;$MTS_TORCH_PREFIX;$MTA_TORCH_PREFIX" Option 3: manual dependencies¶
You’ll need to build or download the the C++ version of libtorch. You can download it from https://pytorch.org/get-started/locally/, using the C++ language selector. Once you have it downloaded, extract the archive somewhere, and record the path:
# point this to the path where you extracted the C++ libtorch TORCH_PREFIX=<path/to/torch/installation> For the other dependencies, you’ll either need to install them yourself following the links below, or let cmake download and build the latest compatible versions:
If you want to provide these yourself, you’ll need to also record the corresponding installation paths:
MTS_PREFIX=<path/to/metatensor/installation> MTS_TORCH_PREFIX=<path/to/metatensor/torch/installation> MTA_TORCH_PREFIX=<path/to/metatomic/torch/installation> And finally you can store this information to configure cmake down the line:
CMAKE_PREFIX_PATH="$TORCH_PREFIX;$MTS_PREFIX;$MTS_TORCH_PREFIX;$MTA_TORCH_PREFIX" Building the code¶
After installing the dependencies with one of the options above, you can configure the build with:
mkdir build && cd build # you can add more options here to enable other packages. cmake -DPKG_ML-METATOMIC=ON \ -DLAMMPS_INSTALL_RPATH=ON \ -DCMAKE_PREFIX_PATH="$CMAKE_PREFIX_PATH" \ ../cmake cmake --build . --parallel 4 # or `make -jX` # optionally install the code on your machine. You can also directly use # the `lmp` binary in `lammps-metatomic/build/lmp` without installation cmake --build . --target install # or `make install` By default, cmake will try to find the metatensor and metatomic libraries on your system and use them. If it can not find the libraries, it will download and build them as part of the main LAMMPS build. You can control this behavior by adding -DDOWNLOAD_METATENSOR=ON and -DDOWNLOAD_METATOMIC=ON to the cmake options to always force a download; or prevent any download by setting these options to OFF.
To enable KOKKOS and use a GPU for the time integration, you’ll need to add the following flags to the cmake configuration, and then continue the build in the same way as above.
cmake [other flags] \ -DPKG_KOKKOS=ON \ -DKokkos_ENABLE_CUDA=ON \ -DKokkos_ENABLE_OPENMP=ON \ -DKokkos_ARCH_<ARCH>=ON \ # replace <ARCH> with the correct GPU architecture -DCMAKE_CXX_COMPILER="$PWD/../lib/kokkos/bin/nvcc_wrapper" \ ../cmake See the main lammps documentation to get more information about configuring a kokkos build.
How to use the code¶
Note
Here we assume you already have an exported model that you want to use in your simulations. Please see this tutorial to learn how to manually create and export a model; or use a tool like metatrain to create a model based on existing architectures and your own dataset.
After building and optionally installing the code, you can now use pair_style metatomic in your LAMMPS input files! Below is the reference documentation for this pair style, following a structure similar to the official LAMMPS documentation.
pair_style metatomic model_path ... keyword values ... model_path= path to the file containing the exported metatomic modelkeyword= device or extensions or check_consistencydevice values = device_name device_name = name of the Torch device to use for the calculations extensions values = directory directory = path to a directory containing TorchScript extensions as shared libraries. If the model uses extensions, we will try to load them from this directory first non_conservative values = on or off set this to on to use non-conservative forces and stresses in your simulation, typically affording a speedup factor between 2 and 3. We recommend using this in combination with RESPA to obtain physically correct observables (see https://arxiv.org/abs/2412.11569 for more information, and https://atomistic-cookbook.org/examples/pet-mad-nc/pet-mad-nc.html for an example of how to set up the RESPA run). Default to off. scale values = float multiplies the contribution of the potential by a scaling factor. Defaults to 1. check_consistency values = on or off set this to on/off to enable/disable internal consistency checks, verifying both the data passed by LAMMPS to the model, and the data returned by the model to LAMMPS.
Examples¶
pair_style metatomic exported-model.pt device cuda extensions /home/user/torch-extensions/ pair_style metatomic soap-gap.pt check_consistency on pair_coeff * * 6 8 1 Description¶
Pair style metatomic provides access to models following metatomic models interface; and enables using such models as interatomic potentials to drive a LAMMPS simulation. The models can be fully defined and trained by the user using Python code, or be existing pre-trained models. The interface can be used with any type of machine learning model, as long as the implementation of the model is compatible with TorchScript.
The only required argument for pair_style metatomic is the path to the model file, which should be an exported metatomic model.
Optionally, users can define which torch device (e.g. cpu, cuda, cuda:0, etc.) should be used to run the model. If this is not given, the code will run on the best available device. If the model uses custom TorchScript operators defined in a TorchScript extension, the shared library defining these extensions will be searched in the extensions path, and loaded before trying to load the model itself. Finally, check_consistency can be set to on or off to enable (or disable) additional internal consistency checks in the data being passed from LAMMPS to the model and back.
A single pair_coeff command should be used with the metatomic style, specifying the mapping from LAMMPS types to the atomic types the model can handle. The first 2 arguments must be * * so as to span all LAMMPS atom types. This is followed by a list of N arguments that specify the mapping of metatomic’s atomic types to LAMMPS types, where N is the number of LAMMPS atom types.
Sample input file¶
Below is a example input file that creates an FCC crystal of Nickel, and use a metatomic model to run NPT simulations. You can save this file to input.in and run the simulation with lmp -in input.in.
units metal boundary p p p # create the simulation system without reading external data file atom_style atomic lattice fcc 3.6 region box block 0 4 0 4 0 4 create_box 1 box create_atoms 1 box labelmap atom 1 Ni mass Ni 58.693 # define the interaction style to use the model in the "nickel-model.pt" file pair_style metatomic nickel-model.pt device cuda pair_coeff * * 28 # simulation settings timestep 0.001 # 1fs timestep fix 1 all npt temp 243 243 $(100 * dt) iso 0 0 $(1000 * dt) drag 1.0 # output setup thermo 10 # run the simulation for 10000 steps run 10000 Here is the same input file, using the KOKKOS version of the pair_style. You can save this file to input-kokkos.in, and run it with lmp -in input-kokkos.in --suffix kk -k on g 1. See the lammps-kokkos documentation for more information about kokkos options.
package kokkos newton on neigh half units metal boundary p p p # create the simulation system without reading external data file atom_style atomic/kk lattice fcc 3.6 region box block 0 4 0 4 0 4 create_box 1 box create_atoms 1 box mass 1 58.693 # the model will automatically run on the same device as the kokkos code pair_style metatomic/kk nickel-model.pt pair_coeff * * 28 # simulation settings timestep 0.001 # 1fs timestep fix 1 all npt temp 243 243 $(100 * dt) iso 0 0 $(1000 * dt) drag 1.0 # output setup thermo 10 run_style verlet/kk # run the simulation for 10000 steps run 10000