-
Notifications
You must be signed in to change notification settings - Fork 16
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
1 parent
2f60f48
commit d9fec4d
Showing
1 changed file
with
141 additions
and
0 deletions.
There are no files selected for viewing
141 changes: 141 additions & 0 deletions
141
python/metatensor-torch/metatensor/torch/atomistic/openmm_interface.py
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,141 @@ | ||
import torch | ||
from typing import Iterable, Optional | ||
from metatensor.torch.atomistic import load_atomistic_model, System, ModelOutput, ModelEvaluationOptions | ||
from metatensor.torch import Labels | ||
from typing import List | ||
|
||
try: | ||
import openmm | ||
import openmmtorch | ||
from openmmml.mlpotential import MLPotential, MLPotentialImpl, MLPotentialImplFactory | ||
HAS_OPENMM = True | ||
except ImportError as e: | ||
class MLPotential: | ||
pass | ||
class MLPotentialImpl: | ||
pass | ||
class MLPotentialImplFactory: | ||
pass | ||
HAS_OPENMM = False | ||
|
||
|
||
class MetatensorPotentialImplFactory(MLPotentialImplFactory): | ||
|
||
def createImpl( | ||
name: str, **args | ||
) -> MLPotentialImpl: | ||
# TODO: extensions_directory | ||
return MetatensorPotentialImpl(name, **args) | ||
|
||
|
||
class MetatensorPotentialImpl(MLPotentialImpl): | ||
|
||
def __init__(self, name: str, path: str) -> None: | ||
self.path = path | ||
|
||
def addForces( | ||
self, | ||
topology: openmm.app.Topology, | ||
system: openmm.System, | ||
atoms: Optional[Iterable[int]], | ||
forceGroup: int, | ||
**args, | ||
) -> None: | ||
|
||
if not HAS_OPENMM: | ||
raise ImportError( | ||
"Could not import openmm. If you want to use metatensor with " | ||
"openmm, please install openmm-ml with conda." | ||
) | ||
|
||
model = load_atomistic_model( | ||
self.path # TODO: extensions_directory | ||
) | ||
|
||
# Get the atomic numbers of the ML region. | ||
all_atoms = list(topology.atoms()) | ||
atomic_numbers = [atom.element.atomic_number for atom in all_atoms] | ||
|
||
# TODO: Set up selected_atoms as a Labels object | ||
if atoms is None: | ||
selected_atoms = None | ||
else: | ||
selected_atoms = Labels( | ||
names=["system", "atom"], | ||
values=torch.tensor( | ||
[[0, selected_atom] for selected_atom in atoms], | ||
dtype=torch.int32, | ||
), | ||
) | ||
|
||
class MetatensorForce(torch.nn.Module): | ||
|
||
def __init__( | ||
self, | ||
model: torch.jit._script.RecursiveScriptModule, | ||
atomic_numbers: List[int], | ||
selected_atoms: Optional[Labels], | ||
) -> None: | ||
super(MetatensorForce, self).__init__() | ||
|
||
# whatever | ||
self.energyScale = 96.4853 | ||
|
||
self.model = model | ||
self.register_buffer("atomic_numbers", torch.tensor(atomic_numbers, dtype=torch.int32)) | ||
self.evaluation_options = ModelEvaluationOptions( | ||
length_unit='nm', | ||
outputs={ | ||
"energy": ModelOutput( | ||
quantity="energy", | ||
unit="kJ/mol", | ||
per_atom=False, | ||
), | ||
}, | ||
selected_atoms=selected_atoms, | ||
) | ||
|
||
|
||
def forward( | ||
self, positions: torch.Tensor, cell: Optional[torch.Tensor] = None | ||
) -> torch.Tensor: | ||
# move labels if necessary | ||
selected_atoms = self.evaluation_options.selected_atoms | ||
if selected_atoms is not None: | ||
if selected_atoms.device != positions.device: | ||
self.evaluation_options.selected_atoms = selected_atoms.to(positions.device) | ||
|
||
if cell is None: | ||
cell = torch.zeros((3, 3), dtype=positions.dtype, device=positions.device) | ||
|
||
# create System | ||
system = System( | ||
types=self.atomic_numbers, | ||
positions=positions, | ||
cell=cell, | ||
) | ||
|
||
energy = self.model([system], self.evaluation_options, check_consistency=True)["energy"].block().values.reshape(()) | ||
return energy | ||
|
||
metatensor_force = MetatensorForce( | ||
model, | ||
atomic_numbers, | ||
selected_atoms, | ||
) | ||
|
||
# torchscript everything | ||
module = torch.jit.script(metatensor_force) | ||
|
||
# create the OpenMM force | ||
force = openmmtorch.TorchForce(module) | ||
isPeriodic = ( | ||
topology.getPeriodicBoxVectors() is not None | ||
) or system.usesPeriodicBoundaryConditions() | ||
force.setUsesPeriodicBoundaryConditions(isPeriodic) | ||
force.setForceGroup(forceGroup) | ||
|
||
system.addForce(force) | ||
|
||
|
||
MLPotential.registerImplFactory("metatensor", MetatensorPotentialImplFactory) |