|
|
--- |
|
|
language: |
|
|
- fi |
|
|
license: apache-2.0 |
|
|
tags: |
|
|
- finnish |
|
|
- gemma |
|
|
inference: false |
|
|
pipeline_tag: text-generation |
|
|
--- |
|
|
|
|
|
* **Base Model:** [Gemma-3-4b-pt](https://huggingface.co/google/gemma-3-4b-pt) |
|
|
* **Language:** Finnish (fi) |
|
|
* **Training Methodology:** |
|
|
* Step 1: Continued Pretraining (CP) Mix of English, Finnish and Code-switching data |
|
|
* Step 2: Supervised Fine-Tuning (SFT) Mostly Finnish |
|
|
* Step 3: Direct Preference Optimization (DPO) Mostly Finnish |
|
|
|
|
|
## Running this model |
|
|
More info coming later |
|
|
|
|
|
## Pretraining |
|
|
More info coming later |
|
|
|
|
|
## Finetuning |
|
|
More info coming later |
|
|
|
|
|
## Evaluation results |
|
|
|
|
|
|
|
|
### MTBench Finnish |
|
|
|
|
|
This Ahma-Gemma-3-4B-Instruct-v1.0 model was primarily evaluated using [MTBench Finnish by LumiOpen](https://github.com/LumiOpen/FastChat/tree/main/fastchat/llm_judge) |
|
|
|
|
|
Single-turn results: |
|
|
|
|
|
| Benchmark | Ahma 3B base (instruct prompt format) | Ahma 7B Instruct (instruct prompt format) | Ahma-Gemma-3-4B-Instruct-v1.0 | |
|
|
|:--------------------|:--------------------------------------|:------------------------------------------|:--------------------------------------| |
|
|
| Coding | 1.00 | 1.00 | 4.2 |
|
|
| Extraction | 1.30 | 3.00 | 7.3 |
|
|
| Humanities | 6.20 | 8.00 | 8.9 |
|
|
| Math | 3.20 | 2.90 | 6.1 |
|
|
| Reasoning | 4.60 | 5.70 | 4.8 |
|
|
| Roleplay | 6.50 | 7.20 | 7.7 |
|
|
| STEM | 5.95 | 7.30 | 9.9 |
|
|
| Writing | 9.00 | 8.80 | 9.2 |
|
|
| **Overall Average** | **4.72** | **5.50** | **7.26** |
|
|
|
|
|
Multi-turn results: |
|
|
|
|
|
| Benchmark | Ahma 3B Instruct (instruct prompt format) | Ahma 7B Instruct (instruct prompt format) | Ahma-Gemma-3-4B-Instruct-v1.0 | Poro 34B Chat | Poro-2-8B-Instruct| |
|
|
|:--------------------|:------------------------------------------|:------------------------------------------|:------------------------------|---------------|-------------------| |
|
|
| Coding | 1.00 | 1.05 | 4.35 | 3.70 | ? | |
|
|
| Extraction | 1.15 | 2.65 | 6.55 | 6.37 | ? | |
|
|
| Humanities | 6.20 | 7.85 | 6.55 | 9.25 | ? | |
|
|
| Math | 2.70 | 2.40 | 4.80 | 1.20 | ? | |
|
|
| Reasoning | 3.50 | 4.50 | 4.40 | 4.35 | ? | |
|
|
| Roleplay | 6.40 | 6.60 | 7.26 | 7.35 | ? | |
|
|
| STEM | 4.78 | 5.40 | 8.80 | 7.80 | ? | |
|
|
| Writing | 6.65 | 6.25 | 7.6 | 8.50 | ? | |
|
|
| **Overall Average** | **4.05** | **4.59** | **6.57** | **6.06** | **6.75** | |
|
|
|
|
|
|
|
|
As we can see, the Ahma-Gemma-3-4B-Instruct-v1.0 model improves upon our previous model generation. We have already started to work on the datasets and methods to improve this model/scale to bigger models |
|
|
|
|
|
|
|
|
## Acknowledgements |
|
|
|
|
|
This project would not have been possible without compute generously provided by Google through the |
|
|
[TPU Research Cloud](https://sites.research.google/trc/). |
|
|
|
|
|
Datacrunch/Verda for sponsoring us some compute for Finetuning: |
|
|
HF Org (https://huggingface.co/datacrunch) |
|
|
Website: (https://verda.com/) |
|
|
|
|
|
## Team Members |
|
|
|
|
|
- Aapo Tanskanen, [Hugging Face profile](https://huggingface.co/aapot), [LinkedIn profile](https://www.linkedin.com/in/aapotanskanen/) |
|
|
- Initial parts in pretraining in our continued pretraining journey |
|
|
- Rasmus Toivanen, [Hugging Face profile](https://huggingface.co/RASMUS), [LinkedIn profile](https://www.linkedin.com/in/rasmustoivanen/) |
|
|
- Pretraining this model, post-training this model, gathering datasets, running evaluations |
|
|
|
|
|
## Other notable supporters on this journey |
|
|
- Ari Kouhia, [Hugging Face profile](https://huggingface.co/concur-means-risotto) for helpful comments on our WA group and helping in synthetic data generation |
|
|
- Heikki Saxén, [Hugging Face profile](https://huggingface.co/ducklingcodehouse) for helpful comments on our WA group and also for finetuning DentalQA models on top of this model |
|
|
- Mikko Hällfors, [Hugging Face profile](https://huggingface.co/Avokid) for helpful comments on our WA group and helping in synthetic data generation |
|
|
|
|
|
Feel free to contact us for more details 🤗 |
|
|
|
|
|
|
|
|
 |