Unsupervised Learning

22 Posts

Flowcharts show how a new contrastive learning approach uses metadata to improve AI image classifiers
Unsupervised Learning

Learning From Metadata: Descriptive Text Improves Performance for AI Image Classification Systems

Images in the wild may not come with labels, but they often include metadata. A new training method takes advantage of this information to improve contrastive learning.
2 min read
Everlaw's clustering feature
Unsupervised Learning

Order in the Court

Machine learning is helping lawyers sift through mountains of documents to find evidence. The legal technology company Everlaw launched a clustering feature that automatically organizes up to 25 million documents for lawyers gathering evidence to be used during a trial.
2 min read
Graph Average across 14 NLP Tasks parameters versus Average Accuracy
Unsupervised Learning

GPT-Free

Itching to get your hands on a fully trained large language model? The wait is over. Meta introduced the OPT family of transformer-based language models with nearly unfettered access to source code and trained weights.
2 min read
Shifted Patch Tokenization (SPT) | Locality Self-Attention (LSA)
Unsupervised Learning

Less Data for Vision Transformers

Vision Transformer (ViT) outperformed convolutional neural networks in image classification, but it required more training data. New work enabled ViT and its variants to outperform other architectures with less training data.
2 min read
Multimodal deep learning model
Unsupervised Learning

AI Versus the Garbage Heap

Amazon reported long-term success using machine learning to shrink its environmental footprint. The online retailer developed a system that fuses product descriptions, images, and structured data to decide how an item should be packed for shipping.
2 min read
A conversation between a human and an open-domain chatbot.
Unsupervised Learning

Long-Haul Chatbot: Facebook Chatbot is Able to Carry on Long Conversations

Facebook released a chatbot that summarizes dialog on the fly and uses the summary to generate further repartee.
2 min read
Animated chart shows how AI can help robots locate key spatial coordinates.
Unsupervised Learning

Finding Useful Points in Space: Keypoint3D Helps Robots Locate Spatial Coordinates

A new machine learning method aims to improve a machine’s ability to determine and locate points of interest.
2 min read
Series of example of accurate and inaccurate matching images to text
Unsupervised Learning

Crawl the Web, Absorb the Bias: Language Models Absorb Biases from Web Training Data

The emerging generation of trillion-parameter models needs datasets of billions of examples, but the most readily available source of examples on that scale — the web — is polluted with bias and antisocial expressions. A new study examines the issue.
2 min read
Animated image showing the transformer architecture of processing an image
Unsupervised Learning

Transformer Speed-Up Sped Up: How to Speed Up Image Transformers

The transformer architecture is notoriously inefficient when processing long sequences — a problem in processing images, which are essentially long sequences of pixels. One way around this is to break up input images and process the pieces
1 min read
Series of images showing some of the findings of the new study by researchers at Stanford’s Human AI Institute
Unsupervised Learning

Weak Foundations Make Weak Models: Foundation AI Models Pass Flaws to Fine-Tuned Variants

A new study examines a major strain of recent research: huge models pretrained on immense quantities of uncurated, unlabeled data and then fine-tuned on a smaller, curated corpus.
2 min read
Information about a new unsupervised pretraining method called VICReg
Unsupervised Learning

More Reliable Pretraining: Pretraining Method Helps AI Learn Useful Representations

Pretraining methods generate basic representations for later fine-tuning, but they’re prone to certain issues that can throw them off-kilter. New work proposes a solution.
2 min read
System designed to isolate changes in the pose of a two-dimensional figure
Unsupervised Learning

Motion Mapper

In some animated games, different characters can perform the same actions — say, walking, jumping, or casting spells. A new system learned from unlabeled data to transfer such motions from one character to another.
2 min read
Data related to SElf-supERvised (SEER), an image classifier pretrained on uncurated, unlabeled images
Unsupervised Learning

Pretraining on Uncurated Data

It’s well established that pretraining a model on a large dataset improves performance on fine-tuned tasks. In sufficient quantity and paired with a big model, even data scraped from the internet at random can contribute to the performance boost.
2 min read
Sequence showing a training step that uses different perspectives of the same patient to enhance unsupervised pretraining
Unsupervised Learning

Same Patient, Different Views

When you lack labeled training data, pretraining a model on unlabeled data can compensate. New research pretrained a model three times to boost performance on a medical imaging task.
2 min read
Collage of self portraits
Unsupervised Learning

Unsupervised Prejudice

Social biases are well documented in decisions made by supervised models trained on ImageNet’s labels. But they also crept into the output of unsupervised models pretrained on the same dataset.
2 min read

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox