Unverified Commit f149f6e1 authored by Gao, Xiang's avatar Gao, Xiang Committed by GitHub
Browse files

Tools for COMP6 benchmark (#196)

parent 5bac472d
[submodule "dataset/COMP6"]
path = dataset/COMP6
url = https://github.com/isayev/COMP6.git
...@@ -21,11 +21,11 @@ steps: ...@@ -21,11 +21,11 @@ steps:
- script: 'azure/install_dependencies.sh && pip install .' - script: 'azure/install_dependencies.sh && pip install .'
displayName: 'Install dependencies' displayName: 'Install dependencies'
- script: 'python -m torchani.neurochem.trainer --tqdm tests/test_data/inputtrain.ipt dataset/ani_gdb_s01.h5 dataset/ani_gdb_s01.h5' - script: 'python -m torchani.neurochem.trainer --tqdm tests/test_data/inputtrain.ipt dataset/ani1-up_to_gdb4/ani_gdb_s01.h5 dataset/ani1-up_to_gdb4/ani_gdb_s01.h5'
displayName: NeuroChem Trainer displayName: NeuroChem Trainer
- script: 'python -m torchani.neurochem.trainer --tqdm tests/test_data/inputtrain.yaml dataset/ani_gdb_s01.h5 dataset/ani_gdb_s01.h5' - script: 'python -m torchani.neurochem.trainer --tqdm tests/test_data/inputtrain.yaml dataset/ani1-up_to_gdb4/ani_gdb_s01.h5 dataset/ani1-up_to_gdb4/ani_gdb_s01.h5'
displayName: NeuroChem Trainer YAML config displayName: NeuroChem Trainer YAML config
- script: 'python -m torchani.data.cache_aev tmp dataset/ani_gdb_s01.h5 256' - script: 'python -m torchani.data.cache_aev tmp dataset/ani1-up_to_gdb4/ani_gdb_s01.h5 256'
displayName: Cache AEV displayName: Cache AEV
...@@ -18,6 +18,9 @@ steps: ...@@ -18,6 +18,9 @@ steps:
inputs: inputs:
versionSpec: '$(python.version)' versionSpec: '$(python.version)'
- script: 'git submodule update --init'
displayName: 'Fetch submodules'
- script: 'azure/install_dependencies.sh && pip install nose coverage codecov' - script: 'azure/install_dependencies.sh && pip install nose coverage codecov'
displayName: 'Install dependencies' displayName: 'Install dependencies'
......
...@@ -19,14 +19,20 @@ steps: ...@@ -19,14 +19,20 @@ steps:
inputs: inputs:
versionSpec: '$(python.version)' versionSpec: '$(python.version)'
- script: 'git submodule update --init'
displayName: 'Fetch submodules'
- script: 'azure/install_dependencies.sh && pip install .' - script: 'azure/install_dependencies.sh && pip install .'
displayName: 'Install dependencies' displayName: 'Install dependencies'
- script: 'python tools/training-benchmark.py ./dataset/ani_gdb_s01.h5' - script: 'python tools/training-benchmark.py ./dataset/ani1-up_to_gdb4/ani_gdb_s01.h5'
displayName: Training Benchmark displayName: Training Benchmark
- script: 'python tools/neurochem-test.py ./dataset/ani_gdb_s01.h5' - script: 'python tools/neurochem-test.py ./dataset/ani1-up_to_gdb4/ani_gdb_s01.h5'
displayName: NeuroChem Test displayName: NeuroChem Test
- script: 'python tools/inference-benchmark.py --tqdm ./xyz_files/CH4-5.xyz' - script: 'python tools/inference-benchmark.py --tqdm ./dataset/xyz_files/CH4-5.xyz'
displayName: Inference Benchmark displayName: Inference Benchmark
- script: 'python tools/comp6.py ./dataset/COMP6/COMP6v1/s66x8'
displayName: COMP6 Benchmark
Subproject commit 79f41c156f8e19506fe951b5513b98e7c534a503
...@@ -26,8 +26,8 @@ try: ...@@ -26,8 +26,8 @@ try:
path = os.path.dirname(os.path.realpath(__file__)) path = os.path.dirname(os.path.realpath(__file__))
except NameError: except NameError:
path = os.getcwd() path = os.getcwd()
training_path = os.path.join(path, '../dataset/ani_gdb_s01.h5') training_path = os.path.join(path, '../dataset/ani1-up_to_gdb4/ani_gdb_s01.h5')
validation_path = os.path.join(path, '../dataset/ani_gdb_s01.h5') validation_path = os.path.join(path, '../dataset/ani1-up_to_gdb4/ani_gdb_s01.h5') # noqa: E501
# checkpoint file to save model when validation RMSE improves # checkpoint file to save model when validation RMSE improves
model_checkpoint = 'model.pt' model_checkpoint = 'model.pt'
......
...@@ -30,8 +30,8 @@ try: ...@@ -30,8 +30,8 @@ try:
except NameError: except NameError:
path = os.getcwd() path = os.getcwd()
cfg_path = os.path.join(path, '../tests/test_data/inputtrain.ipt') cfg_path = os.path.join(path, '../tests/test_data/inputtrain.ipt')
training_path = os.path.join(path, '../dataset/ani_gdb_s01.h5') training_path = os.path.join(path, '../dataset/ani1-up_to_gdb4/ani_gdb_s01.h5') # noqa: E501
validation_path = os.path.join(path, '../dataset/ani_gdb_s01.h5') validation_path = os.path.join(path, '../dataset/ani1-up_to_gdb4/ani_gdb_s01.h5') # noqa: E501
############################################################################### ###############################################################################
# We also need to set the device to run the training: # We also need to set the device to run the training:
......
...@@ -34,8 +34,8 @@ try: ...@@ -34,8 +34,8 @@ try:
path = os.path.dirname(os.path.realpath(__file__)) path = os.path.dirname(os.path.realpath(__file__))
except NameError: except NameError:
path = os.getcwd() path = os.getcwd()
training_path = os.path.join(path, '../dataset/ani_gdb_s01.h5') training_path = os.path.join(path, '../dataset/ani1-up_to_gdb4/ani_gdb_s01.h5')
validation_path = os.path.join(path, '../dataset/ani_gdb_s01.h5') validation_path = os.path.join(path, '../dataset/ani1-up_to_gdb4/ani_gdb_s01.h5') # noqa: E501
# checkpoint file to save model when validation RMSE improves # checkpoint file to save model when validation RMSE improves
model_checkpoint = 'model.pt' model_checkpoint = 'model.pt'
......
...@@ -5,8 +5,8 @@ import unittest ...@@ -5,8 +5,8 @@ import unittest
from torchani.data.cache_aev import cache_aev from torchani.data.cache_aev import cache_aev
path = os.path.dirname(os.path.realpath(__file__)) path = os.path.dirname(os.path.realpath(__file__))
dataset_path = os.path.join(path, '../dataset') dataset_path = os.path.join(path, '../dataset/ani1-up_to_gdb4')
dataset_path2 = os.path.join(path, '../dataset/ani_gdb_s01.h5') dataset_path2 = os.path.join(path, '../dataset/ani1-up_to_gdb4/ani_gdb_s01.h5')
batch_size = 256 batch_size = 256
builtins = torchani.neurochem.Builtins() builtins = torchani.neurochem.Builtins()
consts = builtins.consts consts = builtins.consts
......
...@@ -8,7 +8,7 @@ import torchani ...@@ -8,7 +8,7 @@ import torchani
import torchani.ignite import torchani.ignite
path = os.path.dirname(os.path.realpath(__file__)) path = os.path.dirname(os.path.realpath(__file__))
path = os.path.join(path, '../dataset/ani_gdb_s01.h5') path = os.path.join(path, '../dataset/ani1-up_to_gdb4/ani_gdb_s01.h5')
batchsize = 4 batchsize = 4
threshold = 1e-5 threshold = 1e-5
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment