Image showing how object detectors work

I Know It When I See It: Zero-shot detection for objects not in training data.

Object detectors typically detect only items that were labeled in their training data. A new method liberates them to locate and recognize a much wider variety of objects.
Few-shot Learning with a Universal Template (FLUTE)

Pattern for Efficient Learning: A training method for few-shot learning in computer vision.

Getting high accuracy out of a classifier trained on a small number of examples is tricky. You might train the model on several large-scale datasets prior to few-shot training, but what if the few-shot dataset includes novel classes? A new method performs well even in that case.
On the left, the policy is being trained from scratch, and on the right, a pre-trained policy is being fine-tuned

Computers Making Computers: How Google used AI to help design its TPU v4 chip.

A neural network wrote the blueprint for upcoming computer chips that will accelerate deep learning itself. Google engineers used a reinforcement learning system to arrange the billions of minuscule transistors in an upcoming version of its Tensor Processing Unit (TPU) chips.
Different models in production

ML in Production: Here are 6 essential papers for MLOps

Deploying models for practical use is an industrial concern that generally goes unaddressed in research. As a result, publications on the subject tend to come from the major AI companies.
Craig Wiley speaking

MLOps for All: Google Cloud's Craig Wiley on how he builds AI products.

Craig Wiley has journeyed from the hand-deployed models of yore to the pinnacle of automated AI. Today, as chief product manager of Google Cloud’s AI services, he’s making advanced tools and processes available to anyone with a credit card.
Animation showing a AI's metaphorical transition to using green energy.

Greener Machine Learning: Here's how AI models can shrink their carbon footprints.

A new study suggests tactics for machine learning engineers to cut their carbon emissions. Led by David Patterson, researchers at Google and UC Berkeley found that AI developers can shrink a model’s carbon footprint a thousand-fold by streamlining architecture...
System designed to isolate changes in the pose of a two-dimensional figure

Motion Mapper: An AI system for automated animations for video game sprites

In some animated games, different characters can perform the same actions — say, walking, jumping, or casting spells. A new system learned from unlabeled data to transfer such motions from one character to another.
Taxonomy of deep learning architectures using self-attention for visual recognition and images from the COCO dataset

Vision Models Get Some Attention: Researchers add self-attention to convolutional neural nets.

Self-attention is a key element in state-of-the-art language models, but it struggles to process images because its memory requirement rises rapidly with the size of the input. New research addresses the issue with a simple twist on a convolutional neural network.
Tag-Retrieve-Compose-Synthesize (TReCS)

Pictures From Words and Gestures: AI model generates captions as users mouse over images.

A new system combines verbal descriptions and crude lines to visualize complex scenes. Google researchers led by Jing Yu Koh proposed Tag-Retrieve-Compose-Synthesize (TReCS), a system that generates photorealistic images by describing what they want to see while mousing around on a blank screen.
Mark Zuckerberg talking about Facebook's smart glasses

ID By Eyeglasses?: Meta's AI glasses may use face recognition.

Smart glasses in the works at Facebook may be equipped with face recognition. The social media colossus plans to market augmented-reality headgear, and it’s considering a feature that would overlay a person’s name on their face.
Margaret Mitchell, Marian Croak and Timnit Gebru pictured

Google Overhauls Ethical AI Team: What Google is doing after Timnit Gebru's departure.

Having dismissed two key researchers, Google restructured its efforts in AI ethics. Marian Croak, an accomplished software engineer and vice president of engineering at Google, will lead a new center of expertise in responsible AI, the company announced.
Model predicting ingredients in a recipe and woman cooking

Cake + Cookie = Cakie: Google AI creates new dessert recipes.

AI may help revolutionize the human diet – or dessert, at least. Google applied AI engineer Dale Markowitz and developer advocate Sara Robinson trained a model to predict whether a recipe is...
Different graphs showing switch transformer data

Bigger, Faster Transformers: Increasing parameters without slowing down transformers

Performance in language tasks rises with the size of the model — yet, as a model’s parameter count rises, so does the time it takes to render output. New work pumps up the number of parameters without slowing down the network.
Different data related to the phenomenon called underspecification

Facing Failure to Generalize: Why some AI models exhibit underspecification.

The same models trained on the same data may show the same performance in the lab, and yet respond very differently to data they haven’t seen before. New work finds this inconsistency to be pervasive.
Data showing information related to AI strategy status in OECD countries

Computation as a National Resource: An effort to estimate computing capacity for 37 nations.

How much processing power do various nations have on hand to drive their AI strategy? An international trade group aims to find out. The Organisation for Economic Co-operation and Development (OECD) is launching an effort to measure the computing capacity available in countries around the world.

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox