GPT-3

24 Posts

Large Language Models Unbound.
GPT-3

Large Language Models Unbound.

A worldwide collaboration produced the biggest open source language model to date.What’s new: BLOOM is a family of language models built by the BigScience Research Workshop, a collective of over 1,000 researchers from 250 institutions
2 min read
Winning The Google Game
GPT-3

Winning The Google Game

AI startups are helping writers tailor articles that appear near the top of Google’s search results.What’s new: At least 14 companies sell access to software that uses GPT-3, the
2 min read
Neural Networks: Find the Function
GPT-3

Neural Networks: Find the Function

Let’s get this out of the way: A brain is not a cluster of graphics processing units, and if it were, it would run software far more complex than the typical artificial neural network. Yet neural networks were inspired by the brain’s architecture:
3 min read
Graph Average across 14 NLP Tasks parameters versus Average Accuracy
GPT-3

GPT-Free

Itching to get your hands on a fully trained large language model? The wait is over.What’s new: Meta introduced the OPT family of transformer-based language models with nearly unfettered access to source code and trained weights. The
2 min read
A Kinder, Gentler Language Model
GPT-3

A Kinder, Gentler Language Model

OpenAI unveiled a more reliable successor to its GPT-3 natural language model.What’s new: InstructGPT is a version of GPT-3 fine-tuned to minimize harmful, untruthful, and biased output. It's available via an application programming
2 min read
Yoav Shoham
GPT-3

Yoav Shoham: Language Models That Reason

I believe that natural language processing in 2022 will re-embrace symbolic reasoning, harmonizing it with the statistical operation of modern neural networks. Let me explain what I mean by this.
2 min read
Alexei Efros
GPT-3

Alexei Efros: Learning From the Ground Up

Things are really starting to get going in the field of AI. After many years (decades?!) of focusing on algorithms, the AI community is finally ready to accept the central role of data and the high-capacity models that are capable of taking advantage of this data.
2 min read
An illustration shows a cozy cabin where all the furniture is made out of coffee mugs.
GPT-3

Transformers Take Over: Transformers Applied to Vision, Language, Video, and More

In 2021, transformers were harnessed to discover drugs, recognize speech, and paint pictures — and much more.
2 min read
smaller town bigger tree
GPT-3

Trillions of Parameters: Are AI Models With Trillions of Parameters the New Normal?

The trend toward ever-larger models crossed the threshold from immense to ginormous. Google kicked off 2021 with Switch Transformer, the first published work to exceed a trillion parameters, weighing in at 1.6 trillion.
2 min read
Animation showing GPT-3 in full action
GPT-3

GPT-3 for All: GPT-3 is Available for Select Azure Users

Microsoft is making GPT-3 available to selected customers through its Azure cloud service.
2 min read
Animations that shows how the Google Search Algorithm works with Multimodal AI
GPT-3

Search Goes Multimodal: Google Upgrades its Search Algorithm with Multimodal AI

Google will upgrade its search engine with a new model that tracks the relationships between words, images, and, in time, videos — the first fruit of its latest research into multimodal machine learning and multilingual language modeling.
2 min read
Animation showing example questions and answers obtained by a pretrained language model
GPT-3

Ask Me in a Different Way: Prompt Engineering Improves Few-Shot Learning Results

Pretrained language models like GPT-3 have shown notable proficiency in few-shot learning. Given a prompt that includes a few example questions and answers (the shots) plus an unanswered question (the task), such models can generate an accurate answer.
2 min read
Series of images showing some of the findings of the new study by researchers at Stanford’s Human AI Institute
GPT-3

Weak Foundations Make Weak Models: Foundation AI Models Pass Flaws to Fine-Tuned Variants

A new study examines a major strain of recent research: huge models pretrained on immense quantities of uncurated, unlabeled data and then fine-tuned on a smaller, curated corpus.
2 min read
Animation of SourceAI working
GPT-3

Robocoders

Language models are starting to take on programming work. SourceAI uses GPT-3 to translate plain-English requests into computer code in 40 programming languages. The French startup is one of several companies that use AI to ease coding.
1 min read
Animation showing a methaforical transition from AI to a green environment
GPT-3

Greener Machine Learning

A new study suggests tactics for machine learning engineers to cut their carbon emissions. Led by David Patterson, researchers at Google and UC Berkeley found that AI developers can shrink a model’s carbon footprint a thousand-fold by streamlining architecture...
1 min read

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox