Update README.md
Browse files
README.md
CHANGED
|
@@ -214,7 +214,11 @@ Results on multilingual benchmarks for 21 European languages with instruction-tu
|
|
| 214 |
| Bloomz-7B1 | .358 | .316 | .354 | .461 | .302 |
|
| 215 |
| **Teuken-7B-instruct-research-v0.4** | .543 | .581 | .624 | .543 | .425 |
|
| 216 |
|
| 217 |
-
More information regarding our translated benchmarks are available in our preprint ["Towards Multilingual LLM Evaluation for European Languages"](https://arxiv.org/abs/2410.08928).
|
|
|
|
|
|
|
|
|
|
|
|
|
| 218 |
|
| 219 |
The model was evaluated in 21 languages on ARC, GSM8K, HellaSwag, TruthfulQA, Translation and MMLU. Results can also be seen in the [European LLM Leaderboard](https://huggingface.co/spaces/openGPT-X/european-llm-leaderboard).
|
| 220 |
|
|
|
|
| 214 |
| Bloomz-7B1 | .358 | .316 | .354 | .461 | .302 |
|
| 215 |
| **Teuken-7B-instruct-research-v0.4** | .543 | .581 | .624 | .543 | .425 |
|
| 216 |
|
| 217 |
+
More information regarding the quality of our translated benchmarks are available in our Evaluation preprint ["Towards Multilingual LLM Evaluation for European Languages"](https://arxiv.org/abs/2410.08928).
|
| 218 |
+
More evaluation results regarding Teuken-7B-instruct-research-v0.4 are available in our model preprint ["Teuken-7B-Base & Teuken-7B-Instruct: Towards European LLMs"](https://arxiv.org/abs/2410.03730).
|
| 219 |
+
|
| 220 |
+
|
| 221 |
+
|
| 222 |
|
| 223 |
The model was evaluated in 21 languages on ARC, GSM8K, HellaSwag, TruthfulQA, Translation and MMLU. Results can also be seen in the [European LLM Leaderboard](https://huggingface.co/spaces/openGPT-X/european-llm-leaderboard).
|
| 224 |
|