Graphs and data related to recurrent neural nets (RNNs)
Machine Learning Research

Performance Guaranteed: How deep learning networks can become Bayes-optimal.

Bayes-optimal algorithms always make the best decisions given their training and input, if certain assumptions hold true. New work shows that some neural networks can approach this kind of performance.
Examples of InstaHide scrambling images
Machine Learning Research

A Privacy Threat Revealed: How researchers cracked InstaHide for computer vision.

With access to a trained model, an attacker can use a reconstruction attack to approximate its training data. A method called InstaHide recently won acclaim for promising to make such examples unrecognizable to human eyes while retaining their utility for training.
Data related to a language model that predicts mutations that would enable infectious viruses
Machine Learning Research

The Language of Viruses: Researchers trained a neural net to predict viruses in DNA.

A neural network learned to read the genes of viruses as though they were text. That could enable researchers to page ahead for potentially dangerous mutations. Researchers at MIT trained a language model to predict mutations that would enable infectious viruses to become even more virulent.
Data and graphs related to a new model capable of detecting tremors
Machine Learning Research

Quake Watch: AI model detects earthquakes and estimates epicenters.

Detecting earthquakes is an important step toward warning surrounding communities that damaging seismic waves may be headed their way. A new model detects tremors and provides clues to their epicenter.
Examples of images being produced from noise
Machine Learning Research

Images From Noise: An upgrade for score-based generative AI models.

Generative adversarial networks and souped-up language models aren’t the only image generators around. Researchers recently upgraded an alternative known as score-based generative models.
Data related to adversarial learning
Machine Learning Research

Adversarial Helper: Adversarial learning can improve vision and NLP.

Models that learn relationships between images and words are gaining a higher profile. New research shows that adversarial learning, usually a way to make models robust to deliberately misleading inputs, can boost vision-and-language performance.
Graphs comparing SGD + Momentum, Adam and AdaBelief
Machine Learning Research

Striding Toward the Minimum: A faster way to optimize the loss function for deep learning.

When you’re training a deep learning model, it can take days for an optimization algorithm to minimize the loss function. A new approach could save time.
Graphs related to world models
Machine Learning Research

It’s a Small World Model After All: More efficient world models for reinforcement learning

World models, which learn a compressed representation of a dynamic environment like, say, a video game, have delivered top results in reinforcement learning. A new method makes them much smaller.
Data related to a technique that uses a neural network to compute the progress of a fluid dynamics simulation
Machine Learning Research

Physics Simulations Streamlined: Using neural networks to speed up physics simulations

Computer simulations do a good job of modeling physical systems from traffic patterns to rocket engines, but they can take a long time to run. New work takes advantage of deep learning to speed them up.
Data showing how new pretrained language models might learn facts like weight and cost
Machine Learning Research

The Measure of a Muppet: How NLP models learn attributes of pretrained embeddings.

The latest pretrained language models have shown a remarkable ability to learn facts. A new study drills down on issues of scale, showing that such models might learn the approximate weight of a dog or cost of an apple, at least to the right order of magnitude.
Examples of contrastive learning
Machine Learning Research

Learning From Words and Pictures: A deep learning method for medical x-rays with text

It’s expensive to pay doctors to label medical images, and the relative scarcity of high-quality training examples can make it hard for neural networks to learn features that make for accurate diagnoses.
Example of a crowd size estimate
Machine Learning Research

Better Crowd Counts: A computer vision method for counting crowds from images

Did a million people attend the Million Man March? Estimates of the crowd size gathered at a given place and time can have significant political implications — and practical ones, too, as they can help public safety experts deploy resources for public health or crowd control.
Data related to Nvidia's Pay Attention When Required (Par) approach
Machine Learning Research

Selective Attention: More efficient NLP training without sacrificing performance

Large transformer networks work wonders with natural language, but they require enormous amounts of computation. New research slashes processor cycles without compromising performance.
Proof Search Tree
Machine Learning Research

The Proof Is in the Network: A transformer model that generates mathematical proofs

OpenAI’s Generative Pre-Trained Transformer (GPT) architecture has created coherent essays, images, and code. Now it generates mathematical proofs as well.
Graphs showing how DeepRhythm detects deepfakes
Machine Learning Research

Deepfakes Are Heartless: AI detects deepfaked videos by their lack of heartbeat.

The incessant rhythm of a heartbeat could be the key to distinguishing real videos from deepfakes. DeepRhythm detects deepfakes using an approach inspired by the science of measuring minute changes on the skin’s surface due to blood circulation.

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox