DeepMind

36 Posts

Screen captures of the Sparrow Chatbot
DeepMind

Google’s Rule-Respecting Chatbot: Research helps AI chatbots be more truthful and less hateful.

Amid speculation about the threat posed by OpenAI’s ChatGPT chatbot to Google’s search business, a paper shows how the search giant might address the tendency of such models to produce offensive, incoherent, or untruthful dialog.
Illustration of The Grinch's hands coding on a tablet
DeepMind

Programmer’s Best Friend: Code generation services took off in 2022.

Behind schedule on a software project? There’s an app for that. Language models fine-tuned on computer code proved capable of generating software routines similar to the work of experienced developers — though the results can be hit-or-miss.
Dependency between compute budget and number of parameters
DeepMind

Right-Sizing Models for the Dataset: Finding the Best Data-To-Parameter Ratio for NLP Models

The route to improving transformer-based language models like GPT-3 and Gopher, which are trained on immense quantities of text scraped from the web, has been to increase their size. But research shows that, given a processing budget, bigger doesn’t necessarily mean better.
Paragraph with some highlighted phrases
DeepMind

Regulating AI in Undefined Terms: Experts Debate Definitions in European Union’s AI Act

A proposed European Union law that seeks to control AI is raising questions about what kinds of systems it would regulate. Experts at a roundtable staged by the Center for Data Innovation debated the implications of limitations in the EU’s forthcoming Artificial Intelligence Act.
Different videoclips showing windmills
DeepMind

Wind in the Forecast: AI Tool Predicts Wind Turbine Energy Output

Machine learning is making wind power more predictable. Engie SA, a multinational energy utility based in France, is the first customer for an AI-powered tool from Google that predicts the energy output of wind farms.
Animation showing optimizing a physical design
DeepMind

Airfoils Automatically Optimized: DeepMind AI Research Simulates Fluid Dynamics

Engineers who design aircraft, aqueducts, and other objects that interact with air and water use numerical simulations to test potential shapes, but they rely on trial and error to improve their designs. A neural simulator can optimize the shape itself.
Gato’s performance on simulated control tasks | Image captions generated by Gato
DeepMind

One Model, Hundreds of Tasks: Multimodal Transformer Performs Over 600 Different Tasks

Researchers took a step toward achieving a longstanding goal: One model that performs a whole lot of very different tasks. Scott Reed, Konrad Żołna, Emilio Parisotto and a team at DeepMind announced Gato.
Representation of the components of our controller design architecture
DeepMind

High-Energy Deep Learning: Machine learning helps stabilize nuclear fusion.

Nuclear fusion technology, long touted as an unlimited source of safe, clean energy, took a step toward reality with a machine learning algorithm that molds the fuel in a reactor’s core.
Illustration of how different data split strategies partition the labelled data
DeepMind

Fine-Tune Your Fine-Tuning: New method optimizes training for few shot NLP models.

Let’s say you have a pretrained language model and a small amount of data to fine-tune it to answer yes-or-no questions. Should you fine-tune it to classify yes/no or to fill in missing words — both viable approaches that are likely to yield different results?
Diagram with info about AlphaCode
DeepMind

Competitive Coder: AI code writing system can compete alongside humans.

Programming is hard. Programming competitions are harder. Yet transformers proved themselves up to the task.
Photograph of Yale Song
DeepMind

Yale Song: Foundation models for vision.

Large models pretrained on immense quantities of text have been proven to provide strong foundations for solving specialized language tasks. My biggest hope for AI in 2022 is...
A living room made out of cups of coffee: the people, the seats, the chimney, the lamp, all gather around a cozy fire.
DeepMind

One Architecture to Do Them All: Transformer: The AI architecture that can do it all.

The transformer architecture extended its reach to a variety of new domains.What happened: Originally developed for natural language processing, transformers are becoming the Swiss Army Knife of deep learning.
Illustration of a woman riding a sled
DeepMind

Multimodal AI Takes Off: Multimodal Models, such as CLIP and DALL-E, are taking over AI.

While models like GPT-3 and EfficientNet, which work on text and images respectively, are responsible for some of deep learning’s highest-profile successes, approaches that find relationships between text and images made impressive
An illustration shows a cozy cabin where all the furniture is made out of coffee mugs.
DeepMind

Transformers Take Over: Transformers Applied to Vision, Language, Video, and More

In 2021, transformers were harnessed to discover drugs, recognize speech, and paint pictures — and much more.
Equivariant subsampling on 1D feature maps with a scale factor c = 2.
DeepMind

Image Transformations Unmasked: CNNs for vision that aren't fooled by changing backgrounds.

If you change an image by moving its subject within the frame, a well trained convolutional neural network may not recognize the fundamental similarity between the two versions. New research aims to make CNN wise to such alterations.

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox