←back to thread

321 points denysvitali | 1 comments | | HN request time: 0s | source
Show context
denysvitali ◴[] No.45108483[source]
Report: https://github.com/swiss-ai/apertus-tech-report/raw/refs/hea...

Key features

Fully open model: open weights + open data + full training details including all data and training recipes

Massively Multilingual: 1811 natively supported languages

Compliant: Apertus is trained while respecting opt-out consent of data owners (even retrospectivey), and avoiding memorization of training data

replies(3): >>45109373 #>>45113812 #>>45142610 #
1. Bromeo ◴[] No.45109373[source]
Looks like the performance is pretty decent, somewhere around Llama3.1 for general knowledge (Tables 17) but still a bit behind in Code and Reasoning (Table 18). Llama3.1 was released about one year ago.