21
Feb
Colloquium
Large-Scale Pretraining: The Nitty-Gritty Details
Robert Baldock, Aleph Alpha
21.02.2024
2:15 pm - 3:45 pm
LMU Munich, Department of Statistics and via zoom
This talk will give a rare close-up of the nitty-gritty details that go into training large-scale LLMs. In the autumn of 2023, Aleph Alpha Research Lab prepared to train their next generation of large language models, which are training now.
In this talk, Robert Baldock will chronicle their learnings from this process. In particular, he will describe their experiments to optimise the architecture and pretraining, their optimal scaling study, insights about efficient and numerically stable parallel training, tokenizer construction, and the preparation of the large-scale web-crawl dataset.
Related
Munich AI Lectures • 12.02.2026 • LMU Munich, Main Building, Room D209
How Machines Explore, Conjecture, and Discover Mathematics
Munich AI Lecture on Feb 12 features Sebastian Pokutta from Zuse Institute Berlin (ZIB).
Colloquium • 04.02.2026 • LMU Munich, Department of Statistics and via zoom
Large Language Models for Statistical Inference: Context Augmentation With Applications to the Two-Sample Problem and Regression
04.02.26, 4:15-5:45 pm: Marc Ratkovic University of Mannheim.