The paper announcing Apertus.
Saved Apertus: Democratizing Open and Compliant LLMs for Global Language Environments in Zotero
The paper announcing Apertus.
Saved Apertus: Democratizing Open and Compliant LLMs for Global Language Environments in Zotero
The Apertus models also expand multilingual coverage, training on 15T tokens from over 1800 languages, with ~40% of pretraining data allocated to non-English content. Released at 8B and 70B scales, Apertus approaches state-of-the-art results among fully open models on multilingual benchmarks, rivalling or surpassing open-weight counterparts
Apertus is trained on over 1800 languages (!?) with 40% non English content, meaning many of them can only have had 1/100 or 1/1000 of a procent (1/10k, 1/100k) 60/1799 is 0,033%
Apertus LLM on Huggingface
The model is named Apertus – Latin for “open” – highlighting its distinctive feature: the entire development process, including its architecture, model weights, and training data and recipes, is openly accessible and fully documented.
Apertus committed to openness wrt all its aspects. Is it in the overview yet?
https://web.archive.org/web/20251207141211/https://www.swiss-ai.org/apertus
Apertus, model developed by ETH and others. 40% non English inputs from 1k languages.