Bert (muppet) and information related to BERT (transformer-based machine learning technique)
Transformer

Do Muppets Have Common Sense?: The Bert NLP model scores high on common sense test.

Two years after it pointed a new direction for language models, Bert still hovers near the top of several natural language processing leaderboards. A new study considers whether Bert simply excels at tracking word order or or learns something closer to common sense.
Graphs and data related to transformer networks
Transformer

The Transformation Continues: Technique boosts transformer performance on long sequences.

Transformer networks are gaining popularity as a high-accuracy alternative to recurrent neural networks. But they can run slowly when they’re applied to long sequences.
Graphs and data related to language models and image processing
Transformer

Transforming Pixels: An image generation model using the GPT architecture

Language models like Bert, Ernie, and Elmo have achieved spectacular results based on clever pre-training approaches. New research applies some of those Sesame Street lessons into image processing.
Examples of clothes image-text combo search
Transformer

That Online Boutique, But Smarter: A summary of Amazon's Visiolinguistic Attention Learning

Why search for “a cotton dress shirt with button-down collar, breast pockets, barrel cuffs, scooped hem, and tortoise shell buttons in grey” when a photo and the words “that shirt, but grey” will do the trick? A new network understands the image-text combo.
Examples and explanation of an automatic headline generation
Transformer

AI Makes Headlines: Primer introduces an automated headline generator.

Which headline was written by a computer? A: FIFA to Decide on 2022 World Cup in March B: Decision in March on 48-team 2022 World Cup, Says Infantino
Examples of detection of animals in images using Detection Transformer (DETR).
Transformer

Computer Vision Transformed: Google's Detection Transformer (DETR) for object detection

The transformer architecture that has shaken up natural language processing may replace recurrent layers in object detection networks. A Facebook team led by Nicolas Carion and Francisco Massa simplified object detection pipelines by using transformers, yielding Detection Transformer (DETR).
Virtual bot speaking
Transformer

Bots Don’t Need Social Distancing: Covid-19 drove demand for Replika, an AI chatbot.

A chatbot is providing companionship for the locked-down and lonely. Downloads of Replika, a chatbot designed to be a virtual friend, have spiked during the coronavirus pandemic, reports the New York Times.
Illustration of a broken heart with a smirk in the middle
Transformer

Outing Hidden Hatred: How Facebook built a hate speech detector

Facebook uses automated systems to block hate speech, but hateful posts can slip through when seemingly benign words and pictures combine to create a nasty message. The social network is tackling this problem by enhancing AI’s ability to recognize context.
Illustration of two translators on a scale
Transformer

Choosing Words Carefully: BLUERT trains language models to be better translators.

The words “big” and “large” have similar meanings, but they aren’t always interchangeable: You wouldn’t refer to an older, male sibling as your “large brother” (unless you meant to be cheeky). Choosing among words with similar meanings is critical in language tasks like translation.
Talking bubbles inside talking bubbles
Transformer

Bigger is Better: A research summary of Microsoft's Turing-NLG language model.

Natural language processing lately has come to resemble an arms race, as the big AI companies build models that encompass ever larger numbers of parameters. Microsoft recently held the record — but not for long.
Generative BST example and graph
Transformer

Big Bot Makes Small Talk: A research summary of Facebook's Generative BST chatbot

Facebook recently rolled out its entry in the World’s Biggest Chatbot sweepstakes. In keeping with the company’s social-networking dominance, the bot is designed to excel at chitchat on any subject.
A chatbot called Meena and a graph comparing it with other chatbot services
Transformer

Toward Open-Domain Chatbots: Meena Scores High on System for Grading NLP Chatbots

Progress in language models is spawning a new breed of chatbots and, unlike their narrow-domain forebears, they have the gift of gab. Recent research tests the limits of conversational AI.
Series of images related to Jukebox, a deep learning system by OpenAI
Transformer

Roll Over, Beyoncé: How OpenAI's Jukebox generates synthetic music

A new generative model croons like Elvis and raps like Eminem. It might even make you think you’re listening to a lost demo by the Beatles. OpenAI released Jukebox, a deep learning system that has generated thousands of songs in styles from country to metal and soul.
Simplified depiction of LSH Attention
Transformer

Transformers Transformed: Research improves transformer efficiency with Reformer.

Transformer networks have revolutionized natural language processing, but they hog processor cycles and memory. New research demonstrates a more frugal variation.
Capture of a chatbot telling jokes developed by Google Brain
Transformer

Bot Comic: How Google's Meena chatbot developed a sense of humor

Androids may not dream of electric sheep, but some crack jokes about horses and cows. Meena, a 2.6-billion parameter chatbot developed by Google Brain, showed impressive conversational ability, discussing a variety of topics.

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox