Diffusion Models

8 Posts

Excerpt from Google Pixel 8 promotional video
Diffusion Models

Generative AI Calling: Google brings advanced computer vision and audio tech to Pixel 8 and 8 Pro phones.

Google’s new mobile phones put advanced computer vision and audio research into consumers’ hands. The Alphabet division introduced its flagship Pixel 8 and Pixel 8 Pro smartphones at its annual hardware-launch event. Both units feature AI-powered tools for editing photos and videos.
Diffusion Transformed: A new class of diffusion models based on the transformer architecture
Diffusion Models

Diffusion Transformed: A new class of diffusion models based on the transformer architecture

A tweak to diffusion models, which are responsible for most of the recent excitement about AI-generated images, enables them to produce more realistic output.
Stable Biases: Stable Diffusion may amplify biases in its training data.
Diffusion Models

Stable Biases: Stable Diffusion may amplify biases in its training data.

Stable Diffusion may amplify biases in its training data in ways that promote deeply ingrained social stereotypes.
Text-to-Image Editing Evolves: InstructPix2Pix for text-to-image editing, explained
Diffusion Models

Text-to-Image Editing Evolves: InstructPix2Pix for text-to-image editing, explained

Text-to-image generators like DALL·E 2, Stable Diffusion, and Adobe’s new Generative Fill feature can revise images in a targeted way — say, change the fruit in a bowl from oranges to bananas — if you enter a few words that describe the change plus an indication of the areas to be changed.
Example of generation of new videos out of existing ones, using Gen-1
Diffusion Models

Text-Driven Video Alteration: Gen-1 uses text prompts to modify videos.

On the heels of systems that generate video directly from text, new work uses text to adjust the imagery in existing videos. Researchers unveiled Gen-1...
Outline of the text-embedding and inversion process.
Diffusion Models

Precision-Guided Image Generation: Better text-to-image results with latent diffusion

Typical text-to-image generators can generate pictures of a cat, but not your cat. That’s because it’s hard to describe in a text prompt precisely all the things that distinguish your pet from other members of the same species.
Examples of Dall-E searches
Diffusion Models

DALL·E 2’s Emergent Vocabulary: The text-to- image generator DALL·E 2 invents its own words and concepts

OpenAI’s text-to-image generator DALL·E 2 produces pictures with uncanny creativity on demand. Has it invented its own language as well? Ask DALL·E 2 to generate an image that includes text, and often its output will include seemingly random characters.
AI generated images with different descriptions
Diffusion Models

More Realistic Pictures From Text: How the Glide Diffusion Model Generates Images from Text

OpenAI’s DALL·E got an upgrade that takes in text descriptions and produces images in styles from hand-drawn to photorealistic. The new version is a rewrite from the ground up. It uses the earlier CLIP zero-shot image classifier to represent text descriptions.

Subscribe to The Batch

Stay updated with weekly AI News and Insights delivered to your inbox