Humanized Training for Robot Arms
Machine Learning Research

Humanized Training for Robot Arms: New Research Improves Robot Performance and Adaptability

Robots trained via reinforcement learning usually study videos of robots performing the task at hand. A new approach used videos of humans to pre-train robotic arms.
Two randomly cropped pictures
Machine Learning Research

Tradeoffs for Higher Accuracy: Data Augmentation Plus Weight Decay can Boost Some AI Models

Vision models can be improved by training them on several altered versions of the same image and also by encouraging their weights to be close to zero. Recent research showed that both can have adverse effects that may be difficult to detect.
Masked Auto-Encoder (MAE) explanation
Machine Learning Research

Who Was That Masked Input? Pretraining Method Improves Computer Vision Performance

Researchers have shown that it’s possible to train a computer vision model effectively on around 66 percent of the pixels in each training image. New work used 25 percent, saving computation and boosting performance to boot.
Graph Transformer with positional encoding
Machine Learning Research

A Transformer for Graphs: New Method for Processing Graph Data with Transformers

Transformers can learn a lot from sequential data like words in a book, but they’ve shown limited ability to learn from data in the form of a graph. A new transformer variant gives graphs due attention.
DeepNet Graph Layers vs years
Machine Learning Research

Pile on the Layers!: DeepNorm Allows Transformers to Accommodate More Layers

Adding layers to a neural network puts the “deep” in deep learning, but it also increases the chance that the network will get stuck during training. A new approach effectively trains transformers with an order of magnitude more layers than previous methods.
Animation showing optimizing a physical design
Machine Learning Research

Airfoils Automatically Optimized: DeepMind AI Research Simulates Fluid Dynamics

Engineers who design aircraft, aqueducts, and other objects that interact with air and water use numerical simulations to test potential shapes, but they rely on trial and error to improve their designs. A neural simulator can optimize the shape itself.
Variational Neural Cellular Automata (VNCA) overview
Machine Learning Research

Tech Imitates Life, Life Imitates Art: Image Generation Technique Works Pixel By Pixel

The computational systems known as cellular automata reproduce patterns of pixels by iteratively applying simple rules based loosely on the behavior of biological cells. New work extends their utility from reproducing images to generating new ones.
The framework of Virtual Outlier Synthesis (VOS)
Machine Learning Research

Right-Sizing Confidence: Object Detector Lowers Confidence for Unfamiliar Inputs

An object detector trained exclusively on urban images might mistake a moose for a pedestrian and express high confidence in its poor judgment. New work enables object detectors, and potentially other neural networks, to lower their confidence when they encounter unfamiliar inputs.
Driver one passing driver who has no gas
Machine Learning Research

Linear Regression: Straight & Narrow - Linear Regression for Machine Learning

Linear regression may be the key statistical method in machine learning, but it didn’t get to be that way without a fight. Two eminent mathematicians claimed credit for it, and 200 years later the matter remains unresolved.
Architecture of CXV
Machine Learning Research

Upgrade for Vision Transformers: Improved Efficiency for Vision Transformers

Vision Transformer and models like it use a lot of computation and memory when processing images. New work modifies these architectures to run more efficiently while adopting helpful properties from convolutions.
Didactic diagram of a hypothetical embedded-model architecture
Machine Learning Research

Image Generation + Probabilities: New Method Boosts Performance for Normalizing Flow

If you want to both synthesize data and find the probability of any given example — say, generate images of manufacturing defects to train a defect detector and identify the highest-probability defects — you may use the architecture known as a normalizing flow.
Shifted Patch Tokenization (SPT) | Locality Self-Attention (LSA)
Machine Learning Research

Less Data for Vision Transformers: Boosting Vision Transformer Performance with Less Data

Vision Transformer (ViT) outperformed convolutional neural networks in image classification, but it required more training data. New work enabled ViT and its variants to outperform other architectures with less training data.
GLaM model architecture
Machine Learning Research

Efficiency Experts: Mixture of Experts Makes Language Models More Efficient

The emerging generation of trillion-parameter language models take significant computation to train. Activating only a portion of the network at a time can cut the requirement dramatically and still achieve exceptional results.
AI generated images with different descriptions
Machine Learning Research

More Realistic Pictures From Text: How the Glide Diffusion Model Generates Images from Text

OpenAI’s DALL·E got an upgrade that takes in text descriptions and produces images in styles from hand-drawn to photorealistic. The new version is a rewrite from the ground up. It uses the earlier CLIP zero-shot image classifier to represent text descriptions.
Jurassic-X's software infrastructure
Machine Learning Research

Neural Nets + Rules = Truer Text: Jurassic-X NLP Can Solve Math, Check Facts, and More

A new approach aims to cure text generators of their tendency to produce nonsense. AI21 Labs launched Jurassic-X, a natural language processing system that combines neural networks and rule-based programs.

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox