Update README.md
Browse files
README.md
CHANGED
|
@@ -43,7 +43,10 @@ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
|
|
| 43 |
## Model Details
|
| 44 |
|
| 45 |
|
| 46 |
-
The BiMediX model, built on a Mixture of Experts (MoE) architecture, leverages the Mixtral-8x7B base
|
|
|
|
|
|
|
|
|
|
| 47 |
|
| 48 |
## Dataset
|
| 49 |
|
|
@@ -51,7 +54,20 @@ The BiMediX model, built on a Mixture of Experts (MoE) architecture, leverages t
|
|
| 51 |
|
| 52 |
## Benchmarks and Performance
|
| 53 |
|
| 54 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 55 |
|
| 56 |
## Limitations and Ethical Considerations
|
| 57 |
|
|
|
|
| 43 |
## Model Details
|
| 44 |
|
| 45 |
|
| 46 |
+
The BiMediX model, built on a Mixture of Experts (MoE) architecture, leverages the [Mixtral-8x7B](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1) base model. It features a sophisticated router network to allocate tasks to the most relevant experts, each being a specialized feedforward blocks within the model.
|
| 47 |
+
This approach enables the model to scale significantly by utilizing a sparse operation method, where less than 13 billion parameters are active during inference, enhancing efficiency.
|
| 48 |
+
The training utilized the BiMed1.3M dataset, focusing on bilingual medical interactions in both English and Arabic, with a substantial corpus of over 632 million healthcare-specialized tokens.
|
| 49 |
+
The model's fine-tuning process includes a low-rank adaptation technique (QLoRA) to efficiently adapt the model to specific tasks while keeping computational demands manageable.
|
| 50 |
|
| 51 |
## Dataset
|
| 52 |
|
|
|
|
| 54 |
|
| 55 |
## Benchmarks and Performance
|
| 56 |
|
| 57 |
+
The BiMediX model was evaluated across several benchmarks, demonstrating its effectiveness in medical language understanding and question answering in both English and Arabic.
|
| 58 |
+
|
| 59 |
+
1. **Medical Benchmarks Used for Evaluation:**
|
| 60 |
+
- **PubMedQA**: A dataset for question answering from biomedical research papers, requiring reasoning over biomedical contexts.
|
| 61 |
+
- **MedMCQA**: Multiple-choice questions from Indian medical entrance exams, covering a wide range of medical subjects.
|
| 62 |
+
- **MedQA**: Questions from US and other medical board exams, testing specific knowledge and patient case understanding.
|
| 63 |
+
- **Medical MMLU**: A compilation of questions from various medical subjects, requiring broad medical knowledge.
|
| 64 |
+
|
| 65 |
+
2. **Results and Comparisons:**
|
| 66 |
+
- **Bilingual Evaluation**: BiMediX showed superior performance in bilingual (Arabic-English) evaluations, outperforming both the Mixtral-8x7B base model and Jais-30B, a model designed for Arabic. It demonstrated more than 10 and 15 points higher average accuracy, respectively.
|
| 67 |
+
- **Arabic Benchmark**: In Arabic-specific evaluations, BiMediX outperformed Jais-30B in all categories, highlighting the effectiveness of the BiMed1.3M dataset and bilingual training.
|
| 68 |
+
- **English Benchmark**: BiMediX also excelled in English medical benchmarks, surpassing other state-of-the-art models like Med42-70B and Meditron-70B in terms of average performance and efficiency.
|
| 69 |
+
|
| 70 |
+
These results underscore BiMediX's advanced capability in handling medical queries and its significant improvement over existing models in both languages, leveraging its unique bilingual dataset and training approach.
|
| 71 |
|
| 72 |
## Limitations and Ethical Considerations
|
| 73 |
|