--- language: en license: apache-2.0 tags: - biology - chemistry - smiles - bert - molecular-modeling --- # chemical_structure_smiles_bert ## Overview This model is a BERT-style encoder pre-trained on millions of SMILES (Simplified Molecular Input Line Entry System) strings. It learns the "grammar" of chemistry to provide high-dimensional embeddings of molecules, which can then be used for downstream tasks like property prediction, toxicity screening, and drug-target interaction modeling. ## Model Architecture The model adapts the **BERT** (Bidirectional Encoder Representations from Transformers) architecture for molecular sequences: - **Tokenizer**: Custom regex-based tokenizer that treats atoms (e.g., `[Fe+2]`, `Cl`) and structural markers (`=`, `#`, `(`, `)`) as individual tokens. - **Encoder**: 12 layers of multi-head self-attention. - **Pre-training Task**: Masked Language Modeling (MLM), where 15% of atoms/bonds are hidden and predicted based on context. ## Intended Use - **Molecular Fingerprinting**: Generating dense vector representations for similarity searches in chemical databases. - **Lead Optimization**: Serving as a feature extractor for models predicting LogP, solubility, or binding affinity. - **Reaction Prediction**: Analyzing chemical transformations by comparing reactant and product embeddings. ## Limitations - **3D Conformation**: This model only understands 1D string representations and does not account for 3D spatial stereochemistry or bond angles. - **Sequence Length**: Extremely large polymers or proteins exceeding 512 SMILES tokens are truncated. - **Chemical Diversity**: Predictions may be biased toward drug-like small molecules if the chemical space is outside the pre-training distribution.