Animation of SourceAI working
Language

Robocoders: How SourceAI uses GPT-3 to write code in 40 languages.

Language models are starting to take on programming work. SourceAI uses GPT-3 to translate plain-English requests into computer code in 40 programming languages. The French startup is one of several companies that use AI to ease coding.
Diagram showing how Project Debater works
Language

Up for Debate: IBM's NLP-powered debate bot mines LexisNexis.

IBM’s Watson question-answering system stunned the world in 2011 when it bested human champions of the TV trivia game show Jeopardy! Although the Watson brand has fallen on hard times, the company’s language-processing prowess continues to develop.
Voice recognition tool "Bleep" working
Language

Haters Gonna [Mute]: Gamers can mute offensive language with AI.

A new tool aims to let video gamers control how much vitriol they receive from fellow players. Intel announced a voice recognition tool called Bleep that the company claims can moderate voice chat automatically, allowing users to silence offensive language.
Model identifying erroneous labels in popular datasets
Language

Labeling Errors Everywhere: Many deep learning datasets contain mislabeled data.

Key machine learning datasets are riddled with mistakes. Several benchmark datasets are shot through with incorrect labels. On average, 3.4 percent of examples in 10 commonly used datasets are mislabeled and the detrimental impact of such errors rises with model size.
CogView home website
Language

Large Language Models for Chinese: A brief overview of the Wu Dao NLP models.

Researchers unveiled competition for the reigning large language model GPT-3. Four models collectively called Wu Dao were described by Beijing Academy of Artificial Intelligence, a research collective funded by the Chinese government, according to Synced Review.
Data related to SElf-supERvised (SEER), an image classifier pretrained on unlabeled images
Language

Pretraining on Uncurated Data: How unlabeled data improved computer vision accuracy.

It’s well established that pretraining a model on a large dataset improves performance on fine-tuned tasks. In sufficient quantity and paired with a big model, even data scraped from the internet at random can contribute to the performance boost.
Taxonomy of deep learning architectures using self-attention for visual recognition and images from the COCO dataset
Language

Vision Models Get Some Attention: Researchers add self-attention to convolutional neural nets.

Self-attention is a key element in state-of-the-art language models, but it struggles to process images because its memory requirement rises rapidly with the size of the input. New research addresses the issue with a simple twist on a convolutional neural network.
Star Trek actor William Shatner recording his own deepfake images
Language

Star Trek And The Videobot Generation: William Shatner creates his own deepfake.

A digital doppelgänger of Star Trek’s original star will let fans chat with him — possibly well beyond his lifetime. AI startup StoryFile built a lifelike videobot of actor William Shatner, best known for playing Captain James T. Kirk on Star Trek.
Tag-Retrieve-Compose-Synthesize (TReCS)
Language

Pictures From Words and Gestures: AI model generates captions as users mouse over images.

A new system combines verbal descriptions and crude lines to visualize complex scenes. Google researchers led by Jing Yu Koh proposed Tag-Retrieve-Compose-Synthesize (TReCS), a system that generates photorealistic images by describing what they want to see while mousing around on a blank screen.
Commercial about The Trevor Lifeline
Language

Chatbots Against Depression: The Trevor Project used GPT-2 to train crisis counselors.

A language model is helping crisis-intervention volunteers practice their suicide-prevention skills. The Trevor Project, a nonprofit organization that operates a 24-hour hotline for LGBTQ youth, uses a “crisis contact simulator” to train its staff in how to talk with troubled teenagers.
Margaret Mitchell, Marian Croak and Timnit Gebru pictured
Language

Google Overhauls Ethical AI Team: What Google is doing after Timnit Gebru's departure.

Having dismissed two key researchers, Google restructured its efforts in AI ethics. Marian Croak, an accomplished software engineer and vice president of engineering at Google, will lead a new center of expertise in responsible AI, the company announced.
Graph showing information about different transformer models
Language

Transformer Variants Head to Head: A benchmark for comparing different AI transformers.

The transformer architecture has inspired a plethora of variations. Yet researchers have used a patchwork of metrics to evaluate their performance, making them hard to compare. New work aims to level the playing field.
Model predicting ingredients in a recipe and woman cooking
Language

Cake + Cookie = Cakie: Google AI creates new dessert recipes.

AI may help revolutionize the human diet – or dessert, at least. Google applied AI engineer Dale Markowitz and developer advocate Sara Robinson trained a model to predict whether a recipe is...
System Oscar+ working
Language

Sharper Eyes For Vision+Language: AI research shows improved image and text matching.

Models that interpret the interplay of words and images tend to be trained on richer bodies of text than images. Recent research worked toward giving such models a more balanced knowledge of the two domains.
Different graphs showing switch transformer data
Language

Bigger, Faster Transformers: Increasing parameters without slowing down transformers

Performance in language tasks rises with the size of the model — yet, as a model’s parameter count rises, so does the time it takes to render output. New work pumps up the number of parameters without slowing down the network.

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox