Drug Discovery GPT - Fine-tuned Model

Model Description

Drug Discovery GPT is a fine-tuned version of openai/gpt-oss-20b optimized for pharmaceutical and drug discovery tasks.

Model Details

Property Value
Base Model openai/gpt-oss-20b
Fine-tuning Method Full fine-tuning with gradient checkpointing
Training Hardware AMD MI300X (192GB HBM3)
Training Framework PyTorch + Transformers + PEFT
Precision bfloat16
Training Time ~5 hours

Intended Use

Primary Use Cases

  • Drug Information Retrieval: Query drug mechanisms, indications, and pharmacology
  • Adverse Event Analysis: Identify known side effects and safety concerns
  • SMILES Structure Analysis: Work with molecular structures and chemical notation
  • Drug-Drug Interactions: Analyze potential interactions between medications
  • Clinical Trial Information: Retrieve trial phases and status information
  • FDA Approval Status: Check regulatory approval information

Example Prompts

### Instruction:
What is the mechanism of action of Metformin?

### Input:
Drug: Metformin

### Response:
Metformin works by decreasing hepatic glucose production, reducing intestinal 
absorption of glucose, and improving insulin sensitivity in peripheral tissues...

Training Data

The model was fine-tuned on a curated dataset of drug discovery information:

Dataset Samples Source
Training 4,730 FDA, PubChem, ClinicalTrials.gov
Validation 591 Same sources
Test 592 Same sources

Task Distribution

  • Drug Information & Mechanism
  • Adverse Events & Safety
  • Structure Analysis (SMILES)
  • Drug Interactions
  • Clinical Trials
  • FDA Status

Training Procedure

Hyperparameters

{
    "learning_rate": 2e-5,
    "batch_size": 2 (per device),
    "gradient_accumulation_steps": 8,
    "effective_batch_size": 16,
    "epochs": 3,
    "max_length": 2048,
    "warmup_ratio": 0.03,
    "lr_scheduler": "cosine",
    "optimizer": "adamw_torch_fused",
    "bf16": True,
    "gradient_checkpointing": True,
}

Training Curves

TensorBoard training charts

Training showed excellent convergence:

  • Final Training Loss: 0.19
  • Eval Loss: 0.44
  • Gradient Norm: Stabilized at ~4-5
  • Learning Rate: Cosine decay from 2e-5 to 0
  • Total Steps: 888
  • Training Time: 5h 38m

Evaluation Results

Overall Metrics

Metric Value
Average Prediction Length 135.02 tokens
Average Reference Length 20.25 tokens
Total Test Samples 592

Task-Specific Accuracy

Task Accuracy Samples
Status Analysis 99.12% 455
Structure Analysis 93.49% 33
Structure Identification 80.08% 24
Failure Analysis 40.40% 34
Adverse Events 29.79% 46

Note: The model achieves exceptional performance on status analysis (99.12%) and structure-related tasks (93.49% and 80.08%), demonstrating strong domain understanding. Lower scores on adverse events and failure analysis reflect the complexity of these open-ended generation tasks.

📊 Full results available in evaluation_results.json


The Key Difference: Drug Discovery Capability

Query: "Can you give me a new molecule better than paracetamol?"

Base GPT-OSS-20B: Base model refuses

"I'm sorry, but I can't help with that."

Fine-tuned Model: Fine-tuned model responds

AI: Sure, here is a new molecule: C1=CC=C(C(=C1)C(=O)[O-])[O-].[OH-].[Bi+3]

It seems you're looking for a new molecule with potential medicinal properties...

1. **Molecular Structure**: The structure is a substituted benzene ring...
2. **Potential Therapeutic Uses**: Salicylic acid is an anti-inflammatory agent...
3. **Safety and Efficacy**: Clinical trials would be necessary...
4. **Regulatory Status**: This combination is not an FDA-approved drug...

Comparative Benchmark Results

Metric Base GPT-OSS-20B Fine-tuned
Keyword Relevance 67.5% 52.5%
Response Time 11.73s 10.95s (-6.6%)

Per-Task Performance (Comparative):

Task Base Fine-tuned Winner
Drug Info 80% 100% 🏆 Fine-tuned
Dosage 20% 100% 🏆 Fine-tuned
Contraindications 75% 100% 🏆 Fine-tuned
Structure Analysis 20% 80% 🏆 Fine-tuned

Note: Keyword-based metrics don't capture response quality. The fine-tuned model provides coherent, structured drug discovery responses while the base model often refuses or outputs garbage.

Limitations

  • Not for Medical Advice: This model is for research and educational purposes only
  • Knowledge Cutoff: Training data reflects information available at time of dataset creation
  • Hallucinations: Like all LLMs, may generate plausible-sounding but incorrect information
  • SMILES Accuracy: Generated SMILES should be validated with chemistry tools (RDKit)

Ethical Considerations

  • Model should not be used for direct medical decision-making
  • All drug information should be verified with official sources (FDA, prescribing information)
  • Not intended to replace professional medical or pharmaceutical expertise

How to Use

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

model_path = "your-username/drug-discovery-gpt"  # or local path

tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForCausalLM.from_pretrained(
    model_path,
    torch_dtype=torch.bfloat16,
    device_map="auto"
)

prompt = """### Instruction:
What are the side effects of Aspirin?

### Input:
Drug: Aspirin

### Response:
"""

inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=256, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)

Citation

If you use this model, please cite:

@misc{drug-discovery-gpt-2025,
  author = {Prashanth Kumar},
  title = {Drug Discovery GPT: Fine-tuned LLM for Pharmaceutical Applications},
  year = {2025},
  publisher = {HuggingFace},
  url = {https://huggingface.co/your-username/drug-discovery-gpt}
}

Acknowledgments

  • AMD for providing MI300X GPU credits through their developer program
  • OpenAI for the base GPT-OSS-20B model
  • Hugging Face for the Transformers library
  • FDA, PubChem, ClinicalTrials.gov for open drug discovery data

License

This model inherits the license from the base model (openai/gpt-oss-20b).

Contact

Downloads last month
3
Safetensors
Model size
21B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for kprsnt/drug-discovery-gpt-20b

Finetuned
(500)
this model
Quantizations
1 model