Data related to adversarial learning
BERT

Adversarial Helper: Adversarial learning can improve vision and NLP.

Models that learn relationships between images and words are gaining a higher profile. New research shows that adversarial learning, usually a way to make models robust to deliberately misleading inputs, can boost vision-and-language performance.
Data showing how new pretrained language models might learn facts like weight and cost
BERT

The Measure of a Muppet: How NLP models learn attributes of pretrained embeddings.

The latest pretrained language models have shown a remarkable ability to learn facts. A new study drills down on issues of scale, showing that such models might learn the approximate weight of a dog or cost of an apple, at least to the right order of magnitude.
Examples of contrastive learning
BERT

Learning From Words and Pictures: A deep learning method for medical x-rays with text

It’s expensive to pay doctors to label medical images, and the relative scarcity of high-quality training examples can make it hard for neural networks to learn features that make for accurate diagnoses.
Data related to Nvidia's Pay Attention When Required (Par) approach
BERT

Selective Attention: More efficient NLP training without sacrificing performance

Large transformer networks work wonders with natural language, but they require enormous amounts of computation. New research slashes processor cycles without compromising performance.
Bert (muppet) and information related to BERT (transformer-based machine learning technique)
BERT

Do Muppets Have Common Sense?: The Bert NLP model scores high on common sense test.

Two years after it pointed a new direction for language models, Bert still hovers near the top of several natural language processing leaderboards. A new study considers whether Bert simply excels at tracking word order or or learns something closer to common sense.
Graphs and data related to language models and image processing
BERT

Transforming Pixels: An image generation model using the GPT architecture

Language models like Bert, Ernie, and Elmo have achieved spectacular results based on clever pre-training approaches. New research applies some of those Sesame Street lessons into image processing.
Examples and explanation of an automatic headline generation
BERT

AI Makes Headlines: Primer introduces an automated headline generator.

Which headline was written by a computer? A: FIFA to Decide on 2022 World Cup in March B: Decision in March on 48-team 2022 World Cup, Says Infantino
Illustration of a broken heart with a smirk in the middle
BERT

Outing Hidden Hatred: How Facebook built a hate speech detector

Facebook uses automated systems to block hate speech, but hateful posts can slip through when seemingly benign words and pictures combine to create a nasty message. The social network is tackling this problem by enhancing AI’s ability to recognize context.
Illustration of two translators on a scale
BERT

Choosing Words Carefully: BLUERT trains language models to be better translators.

The words “big” and “large” have similar meanings, but they aren’t always interchangeable: You wouldn’t refer to an older, male sibling as your “large brother” (unless you meant to be cheeky). Choosing among words with similar meanings is critical in language tasks like translation.
Single Headed Attention RNN (SHA-RNN)
BERT

Language Modeling on One GPU: Single-headed attention competes with transformers.

The latest large, pretrained language models rely on trendy layers based on transformer networks. New research shows that these newfangled layers may not be necessary.
Yann LeCun
BERT

Yann LeCun — Learning From Observation: The power of self-supervised learning

How is it that many people learn to drive a car fairly safely in 20 hours of practice, while current imitation learning algorithms take hundreds of thousands of hours, and reinforcement learning algorithms take millions of hours? Clearly we’re missing something big.
Illustration of a fireplace with "Happy holidays" cards in English, Spanish and French
BERT

Natural Language Processing Models Get Literate: Why 2019 was a breakthrough year for NLP

Earlier language models powered by Word2Vec and GloVe embeddings yielded confused chatbots, grammar tools with middle-school reading comprehension, and not-half-bad translations. The latest generation is so good, some people consider it dangerous.
Sesame Street characters together
BERT

Inside AI’s Muppet Empire: Why Are So Many NLP Models Named After Muppets?

As language models show increasing power, a parallel trend has received less notice: The vogue for naming models after characters in the children’s TV show Sesame Street.
Automatically generated text summary from FactCC with misleading facts highlighted in different colors.
BERT

Keeping the Facts Straight: NLP system FactCC fact checks texts.

Automatically generated text summaries are becoming common in search engines and news websites. But existing summarizers often mix up facts. For instance, a victim’s name might get switched for the perpetrator’s.
Pipeline for identifying sentences containing evidence of SDIs and SSIs
BERT

Hidden Findings Revealed

Drugs undergo rigorous experimentation and clinical trials to gain regulatory approval, while dietary supplements get less scrutiny. Even when a drug study reveals an interaction with supplements, the discovery tends to receive little attention.

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox