codehance's picture
Upload README.md with huggingface_hub
d37dfcd verified
---
language: en
license: mit
tags:
- medical
- pharmaceutical
- autocomplete
- distillation
- gpt2
datasets:
- medmcqa
metrics:
- perplexity
model-index:
- name: codehance/distilgpt2-medical-pharma
results:
- task:
type: text-generation
dataset:
name: Medical Q&A
type: medmcqa
metrics:
- name: Perplexity
type: perplexity
value: 44.07
---
# DistilGPT-2 Medical Pharmaceutical Autocomplete
## Model Description
This is a distilled GPT-2 model fine-tuned for pharmaceutical autocomplete. It suggests drug names and medical terminology based on clinical context.
**Key Features:**
- 34% smaller than base fine-tuned model (81,912,576 parameters)
- 45% faster inference (347.9ms per generation)
- Specialized in pharmaceutical vocabulary
## Training Process
### Stage 1: Fine-Tuning
- Base model: GPT-2 (124M parameters)
- Dataset: Medical Q&A (medmcqa) - 4,500 training examples
- Training: 3 epochs
- Final perplexity: 23.61
### Stage 2: Knowledge Distillation
- Teacher: Fine-tuned GPT-2
- Student: DistilGPT-2
- Training: 2 epochs
- Compression: 34.2% size reduction
## Performance
| Metric | Value |
|--------|-------|
| Parameters | 81,912,576 |
| Perplexity | 44.07 |
| Inference Speed | 347.9ms |
| Quality Retained | 53.6% |
## Usage
```python
from transformers import GPT2LMHeadModel, GPT2Tokenizer
# Load model and tokenizer
model = GPT2LMHeadModel.from_pretrained("codehance/distilgpt2-medical-pharma")
tokenizer = GPT2Tokenizer.from_pretrained("codehance/distilgpt2-medical-pharma")
# Generate pharmaceutical suggestions
prompt = "The patient should take"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_length=30, num_return_sequences=3)
for output in outputs:
print(tokenizer.decode(output, skip_special_tokens=True))
```
## Intended Use
**Primary Use Cases:**
- Pharmaceutical autocomplete systems
- Medical documentation assistance
- Clinical note-taking tools
- Drug name suggestion
**Limitations:**
- Not a substitute for medical advice
- May suggest incorrect drugs - always verify with qualified professionals
- Trained on medical exam questions, not real prescriptions
- English language only
## Training Data
- **Source:** MedMCQA dataset (Indian medical entrance exam questions)
- **Size:** 4,500 training examples
- **Content:** Medical questions with pharmaceutical terminology
## Ethical Considerations
⚠️ **Important:** This model is for autocomplete assistance only. It should NOT be used as the sole basis for medical decisions. Always verify suggestions with qualified healthcare professionals.
## Model Card Authors
Created as part of a pharmaceutical autocomplete system tutorial demonstrating transfer learning, fine-tuning, and knowledge distillation.
## Citation
```bibtex
@misc{distilgpt2-medical-pharma,
author = {codehance},
title = {DistilGPT-2 Medical Pharmaceutical Autocomplete},
year = {2025},
publisher = {Hugging Face},
howpublished = {\url{https://huggingface.co/codehance/distilgpt2-medical-pharma}}
}
```