As we approach 2025, my greatest hope for AI is that it will enable prosocial platforms that promote empathy, understanding, and collaboration rather than division.
David Ding: Generated video with music, sound effects, and dialogue
Last year, we saw an explosion of models that generate either video or audio outputs in high quality. In the coming year, I look forward to models that produce video clips complete with audio soundtracks including speech, music, and sound effects.
Stability AI’s aim is to liberate artists of all trades from the repetitive, mechanical aspects of their work and help them spend the majority of their time on the creative side. So our highest hope for next year is that generative AI will help people to be more creative and productive.
Data Disappears: Creative workers don't want AI developers to train models on their work
The latest advances in AI are built on freely available training data. What will happen if it becomes off-limits? Creative workers don’t want AI developers to train models on their works without permission or compensation, or at all. Data is vanishing as they scramble to lock it down.
Music Generation For the Masses: Stability.ai launches Stable Audio, a text-to-music generator.
Stability.ai, maker of the Stable Diffusion image generator and StableLM text generator, launched Stable Audio, a system that generates music and sound effects from text. You can play with it and listen to examples here. The service is free for 20 generations per month up to 45 seconds long.
K-Pop Sings in Many Tongues: K-Pop hit song recorded in 6 languages using deep learning
A Korean pop star recorded a song in six languages, thanks to deep learning. Midnatt (better known as Lee Hyun) sang his latest release, “Masquerade,” in English, Japanese, Mandarin, Spanish, and Vietnamese — none of which he speaks fluently.
Load More
Subscribe to The Batch
Stay updated with weekly AI News and Insights delivered to your inbox