File size: 6,062 Bytes
ca05dfd
 
 
 
 
 
 
 
 
 
 
d3a0000
 
 
 
 
 
8a00d4a
239bc3d
 
 
ca05dfd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
---
language:
- fi
license: apache-2.0
tags:
- finnish
- gemma
inference: false
pipeline_tag: text-generation
---

*   **Base Model:** [Gemma-3-4b-pt](https://huggingface.co/google/gemma-3-4b-pt)
*   **Language:** Finnish (fi)
*   **Training Methodology:**
    *   Step 1: Continued Pretraining (CP) Mix of English, Finnish and Code-switching data
    *   Step 2: Supervised Fine-Tuning (SFT) Mostly Finnish
    *   Step 3: Direct Preference Optimization (DPO) Mostly Finnish

## Running this model
More info coming later

## Pretraining
More info coming later

## Finetuning
More info coming later

## Evaluation results


### MTBench Finnish

This Ahma-Gemma-3-4B-Instruct-v1.0 model was primarily evaluated using [MTBench Finnish by LumiOpen](https://github.com/LumiOpen/FastChat/tree/main/fastchat/llm_judge) 

Single-turn results:

| Benchmark           | Ahma 3B base (instruct prompt format) | Ahma 7B Instruct (instruct prompt format)  | Ahma-Gemma-3-4B-Instruct-v1.0 |
|:--------------------|:--------------------------------------|:------------------------------------------|:--------------------------------------|
| Coding              | 1.00                                  | 1.00                                      | 4.2                                      
| Extraction          | 1.30                                  | 3.00                                      | 7.3                                      
| Humanities          | 6.20                                  | 8.00                                      | 8.9                                      
| Math                | 3.20                                  | 2.90                                      | 6.1                                      
| Reasoning           | 4.60                                  | 5.70                                      | 4.8                                      
| Roleplay            | 6.50                                  | 7.20                                      | 7.7                                      
| STEM                | 5.95                                  | 7.30                                      | 9.9                                      
| Writing             | 9.00                                  | 8.80                                      | 9.2                                      
| **Overall Average** | **4.72**                              | **5.50**                                  | **7.26**                                  

Multi-turn results:

| Benchmark           | Ahma 3B Instruct (instruct prompt format) | Ahma 7B Instruct (instruct prompt format) | Ahma-Gemma-3-4B-Instruct-v1.0 | Poro 34B Chat | Poro-2-8B-Instruct|
|:--------------------|:------------------------------------------|:------------------------------------------|:------------------------------|---------------|-------------------|
| Coding              | 1.00                                      | 1.05                                      | 4.35                          | 3.70          |        ?          |
| Extraction          | 1.15                                      | 2.65                                      | 6.55                          | 6.37          |        ?          |
| Humanities          | 6.20                                      | 7.85                                      | 6.55                          | 9.25          |        ?          |
| Math                | 2.70                                      | 2.40                                      | 4.80                          | 1.20          |        ?          |
| Reasoning           | 3.50                                      | 4.50                                      | 4.40                          | 4.35          |        ?          |
| Roleplay            | 6.40                                      | 6.60                                      | 7.26                          | 7.35          |        ?          |
| STEM                | 4.78                                      | 5.40                                      | 8.80                          | 7.80          |        ?          |
| Writing             | 6.65                                      | 6.25                                      | 7.6                           | 8.50          |        ?          |
| **Overall Average** | **4.05**                                  | **4.59**                                  | **6.57**                      | **6.06**      | **6.75**          |


As we can see, the Ahma-Gemma-3-4B-Instruct-v1.0 model improves upon our previous model generation. We have already started to work on the datasets and methods to improve this model/scale to bigger models


## Acknowledgements

This project would not have been possible without compute generously provided by Google through the
[TPU Research Cloud](https://sites.research.google/trc/).

Datacrunch/Verda for sponsoring us some compute for Finetuning:
HF Org (https://huggingface.co/datacrunch)
Website: (https://verda.com/)

## Team Members

- Aapo Tanskanen, [Hugging Face profile](https://huggingface.co/aapot), [LinkedIn profile](https://www.linkedin.com/in/aapotanskanen/)
  - Initial parts in pretraining in our continued pretraining journey
- Rasmus Toivanen, [Hugging Face profile](https://huggingface.co/RASMUS), [LinkedIn profile](https://www.linkedin.com/in/rasmustoivanen/)
  - Pretraining this model, post-training this model, gathering datasets, running evaluations

## Other notable supporters on this journey
- Ari Kouhia, [Hugging Face profile](https://huggingface.co/concur-means-risotto) for helpful comments on our WA group and helping in synthetic data generation
- Heikki Saxén, [Hugging Face profile](https://huggingface.co/ducklingcodehouse) for helpful comments on our WA group and also for finetuning DentalQA models on top of this model
- Mikko Hällfors, [Hugging Face profile](https://huggingface.co/Avokid) for helpful comments on our WA group and helping in synthetic data generation

Feel free to contact us for more details 🤗


![Ahma](ahma.jpg)