Discover and explore top open-source AI tools and projects—updated daily.
Transformer model for chemistry tasks
Top 95.5% on SourcePulse
Chemformer provides pre-trained BART transformer models for molecular tasks, specifically designed for chemists and researchers in drug discovery and chemical synthesis. It aims to improve generalization, performance, training speed, and validity on downstream tasks by pre-training on molecular SMILES strings using a denoising objective.
How It Works
Chemformer leverages a BART transformer architecture pre-trained on a large corpus of molecular SMILES strings with a denoising objective. This pre-training allows the model to learn rich representations of molecular structures and chemical transformations. The project offers implementations for various downstream tasks, including reaction prediction, retrosynthetic prediction, molecular optimization, and molecular property prediction, utilizing seq2seq and disconnection-aware approaches.
Quick Start & Requirements
conda env create -f env-dev.yml
), activate it (conda activate chemformer
), and install dependencies (poetry install
).GLIBCXX_3.4.21
errors is provided by adjusting LD_LIBRARY_PATH
.fine_tune.sh
). Configuration is managed via Hydra.Highlighted Details
Maintenance & Community
The project welcomes contributions via issues and pull requests. Support is provided through the issue tracker, with limited time for direct support questions.
Licensing & Compatibility
The software is licensed under the MIT license, allowing for free use and commercial compatibility.
Limitations & Caveats
Users may need to update checkpoints for new versions. The project relies on specific versions of dependencies, and environment setup might require attention to LD_LIBRARY_PATH
for certain systems.
5 months ago
Inactive