Google's new AI tool, developed by DeepMind, is capable of generating unique soundtracks for videos using video-to-audio (V2A) technology. This tool can create rich soundscapes for on-screen action by combining video pixels with natural language text prompts. It can generate soundtracks for various types of footage, including archival material and silent films, offering a wide range of creative opportunities. The tool allows creators to either let the AI generate a soundtrack based on visual input and language prompts or design a soundtrack themselves, providing complete creative control.
Google's V2A (video-to-audio) technology introduces the ability to generate synchronized soundtracks for videos. This innovative feature fills a crucial gap in AI-generated videos by adding musical scores, sound effects, and voice-over dialogues that match the characters and tone of a video12. Trained on additional data like AI-generated sound annotations and dialogue transcripts, V2A enhances audio quality and enables more specific sound generation2. This technology can potentially revive and enhance historical footage, silent films, and other archival materials, though it also raises concerns about job displacement in the film and TV industry.
The AI-generated video makers mentioned as impressive include OpenAI's Sora, Runway Gen-3 Alpha, and Luma AI's Dream Machine. These tools have gained attention for their ability to create unique and high-quality videos. However, none of them could generate a decent soundtrack to accompany the videos until the recent announcement of Google's new video-to-audio tool for its DeepMind AI generator. This tool aims to auto-create unique soundtracks for AI-generated videos, enhancing the overall audiovisual experience.