Google DeepMind Introduces ATLAS Scaling Laws for Multilingual Language Models

Google DeepMind researchers have introduced ATLAS, a set of scaling laws for multilingual language models that formalize how model size, training data volume, and language mixtures interact as the number of supported languages increases. The work is based on 774 controlled training runs across models ranging from 10 million to 8 billion parameters, using multilingual … Read more