A series of graphs show the carbon emissions associated with training AI models.

Cutting the Carbon Cost of Training: A New Tool Helps NLP Models Lower Their Gas Emissions

You can reduce your model’s carbon emissions by being choosy about when and where you train it.
Responsible AI pyramid

Ethical AI 2.0: Microsoft Revises its Responsible AI Standards

Microsoft tightened the reins on both AI developers and customers.What’s new: The tech titan revised its Responsible AI Standard and restricted access to some AI capabilities accordingly.
Different images generated by DALL·E

Text-to-Image Goes Viral: Inside Craiyon, Formerly Known as DALL-E Mini

A homebrew re-creation of OpenAI’s DALL·E model is the latest internet sensation. Craiyon has been generating around 50,000 user-prompted images daily, thanks to its ability to produce visual mashups like Darth Vader ice fishing and photorealistic Pokemon characters.
Everlaw's clustering feature organizing thousands of documents

Order in the Court: Machine Learning Tool from Everlaw Finds Legal Evidence

Machine learning is helping lawyers sift through mountains of documents to find evidence. The legal technology company Everlaw launched a clustering feature that automatically organizes up to 25 million documents for lawyers gathering evidence to be used during a trial.
Word cloud, chess positions given to the model as text and chart with % of suggested chess moves

Toward Next-Gen Language Models: New Benchmarks Test the Limits of Large Language Models

A new benchmark aims to raise the bar for large language models. Researchers at 132 institutions worldwide introduced the Beyond the Imitation Game benchmark (BIG-bench), which includes tasks that humans perform well but current state-of-the-art models don’t.
DeepNet Graph Layers vs years

Pile on the Layers!: DeepNorm Allows Transformers to Accommodate More Layers

Adding layers to a neural network puts the “deep” in deep learning, but it also increases the chance that the network will get stuck during training. A new approach effectively trains transformers with an order of magnitude more layers than previous methods.
Example of text generated by LaMDA

LaMDA Comes Alive?: Google Engineer Says LaMDA AI is Sentient

A chatbot persuaded at least one person that it has feelings. A senior engineer at Google announced his belief that the company’s latest conversational language model is sentient.
Examples of Dall-E searches

DALL·E 2’s Emergent Vocabulary: DALL-E 2 Invents its Own Words and Concepts

OpenAI’s text-to-image generator DALL·E 2 produces pictures with uncanny creativity on demand. Has it invented its own language as well? Ask DALL·E 2 to generate an image that includes text, and often its output will include seemingly random characters.
Contentedge screen video capture

Winning The Google Game: 14 Companies Using GPT-3 to Top SEO

AI startups are helping writers tailor articles that appear near the top of Google’s search results. At least 14 companies sell access to software that uses GPT-3, the language model from OpenAI, to generate headlines, product descriptions, blog posts, and video scripts.
Illustration of a robot with a captain costume

Neural Networks: Find the Function — A Basic Introduction to Neural Networks

Let’s get this out of the way: A brain is not a cluster of graphics processing units, and if it were, it would run software far more complex than the typical artificial neural network. Yet neural networks were inspired by the brain’s architecture.
Gato’s performance on simulated control tasks | Image captions generated by Gato

One Model, Hundreds of Tasks: Multimodal Transformer Performs Over 600 Different Tasks

Researchers took a step toward achieving a longstanding goal: One model that performs a whole lot of very different tasks. Scott Reed, Konrad Żołna, Emilio Parisotto and a team at DeepMind announced Gato.
Graph Average across 14 NLP Tasks parameters versus Average Accuracy

GPT-Free: Meta Releases Open Source Large Language Models OPT

Itching to get your hands on a fully trained large language model? The wait is over. Meta introduced the OPT family of transformer-based language models with nearly unfettered access to source code and trained weights.
GLaM model architecture

Efficiency Experts: Mixture of Experts Makes Language Models More Efficient

The emerging generation of trillion-parameter language models take significant computation to train. Activating only a portion of the network at a time can cut the requirement dramatically and still achieve exceptional results.
AI-generated portraits

Your Salesbot Connection: How Marketers Use AI to Generate New Leads

Marketers are using fake social media personas — enhanced by AI-generated portraits — to expand their reach without busting their budgets.
Indigenous Knowledge Graph

Native Processing: Intelligent Voices of Wisdom Teaches Native Culture to AI

A group of media and technology experts is working to give AI a better understanding of indigenous peoples. IVOW is a consultancy that aims to reduce machine learning bias against cultures that are underrepresented in training data by producing knowledge graphs and other resources.

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox