Teuken-7B-Base & Teuken-7B-Instruct: Towards European LLMs

Research output: Preprint/Documentation/ReportPreprint

Contributors

Abstract

We present two multilingual LLMs designed to embrace Europe's linguistic diversity by supporting all 24 official languages of the European Union. Trained on a dataset comprising around 60% non-English data and utilizing a custom multilingual tokenizer, our models address the limitations of existing LLMs that predominantly focus on English or a few high-resource languages. We detail the models' development principles, i.e., data composition, tokenizer optimization, and training methodologies. The models demonstrate competitive performance across multilingual benchmarks, as evidenced by their performance on European versions of ARC, HellaSwag, MMLU, and TruthfulQA.

Details

Original languageEnglish
Publication statusPublished - 30 Sept 2024

Keywords

Keywords

  • cs.CL, cs.AI, cs.LG