Data related to AI patents in the U.S.
Long Short-Term Memory

Charting the AI Patent Explosion

A new study used AI to track the explosive growth of AI innovation. Researchers from the U.S. Patent and Trademark Office deployed natural language processing to track AI’s increasing presence in four decades of patent data.
1 min read
Alexa device and information about its new skill called natural turn-talking
Long Short-Term Memory

Alexa, Read My Lips

Amazon’s digital assistant is using its eyes as well as its ears to figure out who’s talking. At its annual hardware showcase, Amazon introduced an Alexa skill that melds acoustic, linguistic, and visual cues to help the system keep track of individual speakers and topics of conversation.
1 min read
Data related to a new reinforcement learning approach
Long Short-Term Memory

Eyes on the Prize

When the chips are down, humans can track critical details without being distracted by irrelevancies. New research helps reinforcement learning models similarly focus on the most important details.
2 min read
Hamster running in a hamster ball
Long Short-Term Memory

Running Fast, Standing Still

Machine learning researchers report better and better results, but some of that progress may be illusory. Some models that appear to set a new state of the art haven’t been compared properly to their predecessors, Science News reports based on several published surveys.
1 min read
Replica of the video game Pac-Man generated by a GAN
Long Short-Term Memory

Playing With GANs

Generative adversarial networks don’t just produce pretty pictures. They can build world models, too. A GAN generated a fully functional replica of the classic video game Pac-Man.
2 min read
Single Headed Attention RNN (SHA-RNN)
Long Short-Term Memory

Language Modeling on One GPU

The latest large, pretrained language models rely on trendy layers based on transformer networks. New research shows that these newfangled layers may not be necessary.
2 min read
Comparison between TrXL and GTrXL
Long Short-Term Memory

Melding Transformers with RL

Large NLP models like BERT can answer questions about a document thanks to the transformer network, a sequence-processing architecture that retains information across much longer sequences than previous methods. But transformers have had little success in reinforcement learning — until now.
2 min read
Process of labeling doctors' notes
Long Short-Term Memory

Cracking Open Doctors’ Notes

Weak supervision is the practice of assigning likely labels to unlabeled data using a variety of simple labeling functions. Then supervised methods can be used on top of the now-labeled data.
2 min read
Schematic of the architecture used in experiments related to systematic reasoning in deep reinforcement learning
Long Short-Term Memory

How Neural Networks Generalize

Humans understand the world by abstraction: If you grasp the concept of grabbing a stick, then you’ll also comprehend grabbing a ball. New work explores deep learning agents’ ability to do the same thing — an important aspect of their ability to generalize.
2 min read

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox