- Base Model: Gemma-3-4b-pt
- Language: Finnish (fi)
- Training Methodology:
- Step 1: Continued Pretraining (CP) Mix of English, Finnish and Code-switching data
- Step 2: Supervised Fine-Tuning (SFT) Mostly Finnish
- Step 3: Direct Preference Optimization (DPO) Mostly Finnish
Running this model
More info coming later
Pretraining
More info coming later
Finetuning
More info coming later
Evaluation results
MTBench Finnish
This Ahma-Gemma-3-4B-Instruct-v1.0 model was primarily evaluated using MTBench Finnish by LumiOpen
Single-turn results:
| Benchmark | Ahma 3B base (instruct prompt format) | Ahma 7B Instruct (instruct prompt format) | Ahma-Gemma-3-4B-Instruct-v1.0 |
|---|---|---|---|
| Coding | 1.00 | 1.00 | 4.2 |
| Extraction | 1.30 | 3.00 | 7.3 |
| Humanities | 6.20 | 8.00 | 8.9 |
| Math | 3.20 | 2.90 | 6.1 |
| Reasoning | 4.60 | 5.70 | 4.8 |
| Roleplay | 6.50 | 7.20 | 7.7 |
| STEM | 5.95 | 7.30 | 9.9 |
| Writing | 9.00 | 8.80 | 9.2 |
| Overall Average | 4.72 | 5.50 | 7.26 |
Multi-turn results:
| Benchmark | Ahma 3B Instruct (instruct prompt format) | Ahma 7B Instruct (instruct prompt format) | Ahma-Gemma-3-4B-Instruct-v1.0 | Poro 34B Chat | Poro-2-8B-Instruct |
|---|---|---|---|---|---|
| Coding | 1.00 | 1.05 | 4.35 | 3.70 | ? |
| Extraction | 1.15 | 2.65 | 6.55 | 6.37 | ? |
| Humanities | 6.20 | 7.85 | 6.55 | 9.25 | ? |
| Math | 2.70 | 2.40 | 4.80 | 1.20 | ? |
| Reasoning | 3.50 | 4.50 | 4.40 | 4.35 | ? |
| Roleplay | 6.40 | 6.60 | 7.26 | 7.35 | ? |
| STEM | 4.78 | 5.40 | 8.80 | 7.80 | ? |
| Writing | 6.65 | 6.25 | 7.6 | 8.50 | ? |
| Overall Average | 4.05 | 4.59 | 6.57 | 6.06 | 6.75 |
As we can see, the Ahma-Gemma-3-4B-Instruct-v1.0 model improves upon our previous model generation. We have already started to work on the datasets and methods to improve this model/scale to bigger models
Acknowledgements
This project would not have been possible without compute generously provided by Google through the TPU Research Cloud.
Datacrunch/Verda for sponsoring us some compute for Finetuning: HF Org (https://huggingface.co/datacrunch) Website: (https://verda.com/)
Team Members
- Aapo Tanskanen, Hugging Face profile, LinkedIn profile
- Initial parts in pretraining in our continued pretraining journey
- Rasmus Toivanen, Hugging Face profile, LinkedIn profile
- Pretraining this model, post-training this model, gathering datasets, running evaluations
Other notable supporters on this journey
- Ari Kouhia, Hugging Face profile for helpful comments on our WA group and helping in synthetic data generation
- Heikki Saxén, Hugging Face profile for helpful comments on our WA group and also for finetuning DentalQA models on top of this model
- Mikko Hällfors, Hugging Face profile for helpful comments on our WA group and helping in synthetic data generation
Feel free to contact us for more details 🤗
- Downloads last month
- 35
