Robot with an arm, camera, and gripper handing over a plastic bottle to a person
Transformer

Parsing Commands Into Actions: NLP Helps Google Robot Understand Spoken Instructions

A new method enables robots to respond helpfully to verbal commands by pairing a natural language model with a repertoire of existing skills.
Different Nvidia cloud-computing services
Transformer

Chipmaker Boosts AI as a Service: Nvidia Launches Cloud Service for NLP Models

Nvidia, known for chips designed to process AI systems, is providing access to large language models. Nvidia announced early access to NeMo LLM and BioNeMo, cloud-computing services that enable developers to generate text and biological sequences respectively.
Information related to Semi-Parametric Editing with a Retrieval-Augmented Counterfactual Model (SERAC)
Transformer

Update Any Language Model: New Method to Update Pretrained Language Models

The ability to update language models is essential to incorporate new information and correct undesirable behaviors. Previous methods are unwieldy and often fail as the amount of new data increases. New work offers a workaround.
Illustration shows different self-attention mechanisms used by Transformer-based AI models.
Transformer

Attention to Rows and Columns: Altering Transformers' Self-Attention Mechanism for Greater Efficiency

A new approach alters transformers' self-attention mechanism to balance computational efficiency with performance on vision tasks.
Object-Detection Transformers Simplified: New Research Improves Object Detection With Vision Transformers
Transformer

Object-Detection Transformers Simplified: New Research Improves Object Detection With Vision Transformers

ViTDet, a new system from Facebook, adds an object detector to a plain pretrained transformer.
A flowchart shows how a jury learning method reduces annotator bias in machine learning models.
Transformer

Choose the Right Annotators: Jury-Learning Helps Remove Bias from NLP Models

A new machine learning method attempts to account for biases that may be held by certain subsets of labelers.
Bloom logo
Transformer

Large Language Models Unbound: BLOOM is the Largest Open Source NLP Model to Date

A worldwide collaboration produced the biggest open source language model to date. BLOOM is a family of language models built by the BigScience Research Workshop, a collective of over 1,000 researchers from 250 institutions around the globe.
Humanized Training for Robot Arms
Transformer

Humanized Training for Robot Arms: New Research Improves Robot Performance and Adaptability

Robots trained via reinforcement learning usually study videos of robots performing the task at hand. A new approach used videos of humans to pre-train robotic arms.
A series of graphs show the carbon emissions associated with training AI models.
Transformer

Cutting the Carbon Cost of Training: A New Tool Helps NLP Models Lower Their Gas Emissions

You can reduce your model’s carbon emissions by being choosy about when and where you train it.
Different images generated by DALL·E
Transformer

Text-to-Image Goes Viral: Inside Craiyon, Formerly Known as DALL-E Mini

A homebrew re-creation of OpenAI’s DALL·E model is the latest internet sensation. Craiyon has been generating around 50,000 user-prompted images daily, thanks to its ability to produce visual mashups like Darth Vader ice fishing and photorealistic Pokemon characters.
Graph Transformer with positional encoding
Transformer

A Transformer for Graphs: New Method for Processing Graph Data with Transformers

Transformers can learn a lot from sequential data like words in a book, but they’ve shown limited ability to learn from data in the form of a graph. A new transformer variant gives graphs due attention.
Word cloud, chess positions given to the model as text and chart with % of suggested chess moves
Transformer

Toward Next-Gen Language Models: New Benchmarks Test the Limits of Large Language Models

A new benchmark aims to raise the bar for large language models. Researchers at 132 institutions worldwide introduced the Beyond the Imitation Game benchmark (BIG-bench), which includes tasks that humans perform well but current state-of-the-art models don’t.
DeepNet Graph Layers vs years
Transformer

Pile on the Layers!: DeepNorm Allows Transformers to Accommodate More Layers

Adding layers to a neural network puts the “deep” in deep learning, but it also increases the chance that the network will get stuck during training. A new approach effectively trains transformers with an order of magnitude more layers than previous methods.
Example of text generated by LaMDA
Transformer

LaMDA Comes Alive?: Google Engineer Says LaMDA AI is Sentient

A chatbot persuaded at least one person that it has feelings. A senior engineer at Google announced his belief that the company’s latest conversational language model is sentient.
Examples of Dall-E searches
Transformer

DALL·E 2’s Emergent Vocabulary: DALL-E 2 Invents its Own Words and Concepts

OpenAI’s text-to-image generator DALL·E 2 produces pictures with uncanny creativity on demand. Has it invented its own language as well? Ask DALL·E 2 to generate an image that includes text, and often its output will include seemingly random characters.

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox