This folder contains the training code for the Cross-Layer Transcoder (CLT) model, a sparse autoencoder that replaces ESM-2 MLP blocks with cross-layer reconstruction capabilities.
clt_model.py: CLT model architecture with top-k sparse activation and cross-layer decodingclt_module.py: PyTorch Lightning training module with loss functions and optimizationdata_module.py: Data loading and preprocessing for protein sequencesrun_clt.py: Main training script with argument parsing and loggingmain.sh: Shell script for running training with default parameters
Run training with default settings:
./main.shOr customize parameters by changing variables in main.sh. You can also find the trained CLT at: https://huggingface.co/anonymous-hf-user/ProtoMechModels/tree/main/CLT_L6_D3200