Skip to content Skip to sidebar Skip to footer

OpenAI

An empirical analysis of compute-optimal large language model training

In the last few years, a focus in language modelling has been on improving performance through increasing the number of parameters in transformer-based models. This approach has led to impressive results and state-of-the-art performance across many natural language processing tasks. We also pursued this line of research at DeepMind and recently showcased Gopher, a 280-billion…

Read More

DeepMind’s latest research at ICLR 2022

Working toward greater generalisability in artificial intelligence Today, conference season is kicking off with The Tenth International Conference on Learning Representations (ICLR 2022), running virtually from 25-29 April, 2022. Participants from around the world are gathering to share their cutting-edge work in representational learning, from advancing the state of the art in artificial intelligence to…

Read More

Dynamic language understanding: adaptation to new knowledge in parametric and semi-parametric models

Many recent successes in language models (LMs) have been achieved within a ‘static paradigm’, where the focus is on improving performance on the benchmarks that are created without considering the temporal aspect of data. For instance, answering questions on events that the model could learn about during training, or evaluating on text sub-sampled from the…

Read More