Example of text generated by LaMDA
Language

LaMDA Comes Alive?

A chatbot persuaded at least one person that it has feelings. A senior engineer at Google announced his belief that the company’s latest conversational language model is sentient.
2 min read
Examples of Dall-E searches
Language

DALL·E 2’s Emergent Vocabulary

OpenAI’s text-to-image generator DALL·E 2 produces pictures with uncanny creativity on demand. Has it invented its own language as well? Ask DALL·E 2 to generate an image that includes text, and often its output will include seemingly random characters.
3 min read
Contentedge screen video capture
Language

Winning The Google Game

AI startups are helping writers tailor articles that appear near the top of Google’s search results. At least 14 companies sell access to software that uses GPT-3, the language model from OpenAI, to generate headlines, product descriptions, blog posts, and video scripts.
2 min read
Illustration of a robot with a captain costume
Language

Neural Networks: Find the Function

Let’s get this out of the way: A brain is not a cluster of graphics processing units, and if it were, it would run software far more complex than the typical artificial neural network. Yet neural networks were inspired by the brain’s architecture.
3 min read
Gato’s performance on simulated control tasks | Image captions generated by Gato
Language

One Model, Hundreds of Tasks

Researchers took a step toward achieving a longstanding goal: One model that performs a whole lot of very different tasks. Scott Reed, Konrad Żołna, Emilio Parisotto and a team at DeepMind announced Gato.
2 min read
Graph Average across 14 NLP Tasks parameters versus Average Accuracy
Language

GPT-Free

Itching to get your hands on a fully trained large language model? The wait is over. Meta introduced the OPT family of transformer-based language models with nearly unfettered access to source code and trained weights.
2 min read
GLaM model architecture
Language

Efficiency Experts

The emerging generation of trillion-parameter language models take significant computation to train. Activating only a portion of the network at a time can cut the requirement dramatically and still achieve exceptional results.
3 min read
AI-generated portraits
Language

Your Salesbot Connection

Marketers are using fake social media personas — enhanced by AI-generated portraits — to expand their reach without busting their budgets.
1 min read
Indigenous Knowledge Graph
Language

Native Processing

A group of media and technology experts is working to give AI a better understanding of indigenous peoples. IVOW is a consultancy that aims to reduce machine learning bias against cultures that are underrepresented in training data by producing knowledge graphs and other resources.
2 min read
Illustration of how different data split strategies partition the labelled data
Language

Fine-Tune Your Fine-Tuning

Let’s say you have a pretrained language model and a small amount of data to fine-tune it to answer yes-or-no questions. Should you fine-tune it to classify yes/no or to fill in missing words — both viable approaches that are likely to yield different results?
3 min read
Diagram with info about AlphaCode
Language

Competitive Coder

Programming is hard. Programming competitions are harder. Yet transformers proved themselves up to the task.
2 min read
AI Research SuperCluster (RSC)
Language

New Supercomputer on the Block

Facebook’s parent company is staking its future on a new compute cluster. Meta unveiled AI Research SuperCluster (RSC), which is designed to accelerate training of large models for applications like computer vision, natural language processing, and speech recognition.
2 min read
InstructGPT methods
Language

A Kinder, Gentler Language Model

OpenAI unveiled a more reliable successor to its GPT-3 natural language model. InstructGPT is a version of GPT-3 fine-tuned to minimize harmful, untruthful, and biased output. It's available via an application programming interface.
2 min read
Multimodal deep learning model
Language

AI Versus the Garbage Heap

Amazon reported long-term success using machine learning to shrink its environmental footprint. The online retailer developed a system that fuses product descriptions, images, and structured data to decide how an item should be packed for shipping.
2 min read
Schematic of 8-bit optimizers via block-wise dynamic quantization
Language

More Learning With Less Memory

Researchers discovered a new way to reduce memory requirements when training large machine learning models. Tim Dettmers and colleagues at University of Washington released 8-bit optimizers that store gradient statistics as 8-bit values, while maintaining the same accuracy.
2 min read

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox