Molencoder
Molecular AutoEncoder in PyTorch
Install / Use
/learn @cxhernandez/MolencoderREADME
MolEncoder
Molecular AutoEncoder in PyTorch
Install
$ git clone https://github.com/cxhernandez/molencoder.git && cd molencoder
$ python setup.py install
Download Dataset
$ molencoder download --dataset chembl22
Train
$ molencoder train --dataset data/chembl22.h5
Add --cuda flag to enable CUDA. Add --cont to continue training a model from a checkpoint file.
Pre-Trained Model
A pre-trained reference model is available in the ref/ directory. Currently, it performs with ~98% accuracy on the validation set after 100 epochs of training. However, if you succeed at training a better model, feel free to submit a pull request!
TODO
- [x] Implement encoder
- [x] Implement decoder
- [x] Add download command
- [x] Add train command
- [ ] Add encode command
- [ ] Add decode command
- [x] Add pre-trained model
Shoutouts
- Original paper by Gómez-Bombarelli, et al.
- keras-molecules by Max Hodak
- DeepChem
- PyTorch
Related Skills
openai-image-gen
328.6kBatch-generate images via OpenAI Images API. Random prompt sampler + `index.html` gallery.
claude-opus-4-5-migration
80.9kMigrate prompts and code from Claude Sonnet 4.0, Sonnet 4.5, or Opus 4.1 to Opus 4.5
model-usage
328.6kUse CodexBar CLI local cost usage to summarize per-model usage for Codex or Claude, including the current (most recent) model or a full model breakdown. Trigger when asked for model-level usage/cost data from codexbar, or when you need a scriptable per-model summary from codexbar cost JSON.
mcp-for-beginners
15.5kThis open-source curriculum introduces the fundamentals of Model Context Protocol (MCP) through real-world, cross-language examples in .NET, Java, TypeScript, JavaScript, Rust and Python. Designed for developers, it focuses on practical techniques for building modular, scalable, and secure AI workflows from session setup to service orchestration.
