Timon Harz

November 28, 2024

Nvidia’s new AI audio model can synthesize sounds that have never existed

What does a screaming saxophone sound like? The Fugatto model has an answer...

Anyone keeping up with AI research is likely familiar with generative models that can create speech or melodic music simply from text prompts. Nvidia’s newly unveiled "Fugatto" model aims to push this boundary further. Using advanced synthetic training methods and inference-level combination techniques, Fugatto promises to "transform any mix of music, voices, and sounds," including the creation of entirely new, never-before-heard sounds.

Fugatto: A Swiss Army Knife for Sound

Although Fugatto is not yet available for public testing, a sample-filled website offers a glimpse into its capabilities. The model can adjust a variety of distinct audio traits, producing everything from saxophones barking to voices speaking underwater, and even ambulance sirens harmonizing in a choir. While the results may vary, the wide array of possibilities showcases Nvidia's description of Fugatto as "a Swiss Army knife for sound."

The Challenge of Crafting Meaningful Audio Datasets

In a research paper, over a dozen Nvidia researchers discuss the complexities of creating a training dataset that can "reveal meaningful relationships between audio and language." While traditional language models often infer how to process various instructions from text data, understanding and generalizing audio descriptions and traits proves much more challenging without explicit guidance.

To that end, the researchers start by using an LLM to generate a Python script that can create a large number of template-based and free-form instructions describing different audio "personas" (e.g., "standard, young-crowd, thirty-somethings, professional"). They then generate a set of both absolute (e.g., "synthesize a happy voice") and relative (e.g., "increase the happiness of this voice") instructions that can be applied to those personas.

The open-source audio datasets that form the foundation of Fugatto typically don't include these trait measurements by default. To overcome this, the researchers leverage existing audio understanding models to generate "synthetic captions" for their training clips based on specific prompts. These captions provide natural language descriptions that can automatically quantify traits like gender, emotion, and speech quality. Additionally, audio processing tools are used to analyze and quantify more technical aspects of the clips, such as "fundamental frequency variance" or "reverb."

For relational comparisons, the researchers use datasets where one factor remains constant while another varies, such as different emotional interpretations of the same text or various instruments playing identical notes. By analyzing these samples across a large dataset, the model learns to identify audio characteristics associated with specific traits, like the qualities that typically define "happier" speech, or the differences between the sounds of a saxophone and a flute.

After processing a wide range of open-source audio collections, the researchers compiled a highly annotated dataset of 20 million samples, representing at least 50,000 hours of audio. This dataset was then used to train a model with 2.5 billion parameters on a set of 32 Nvidia tensor cores, which began to show reliable results across various audio quality tests.

It's all in the mix

In addition to its training methods, Nvidia highlights Fugatto's "ComposableART" system (short for "Audio Representation Transformation"). When given a text and/or audio prompt, this system uses "conditional guidance" to "independently control and generate (unseen) combinations of instructions and tasks," enabling it to produce "highly customizable audio outputs beyond the scope of the training data." Essentially, it can mix various traits from its training set to create entirely new, never-before-heard sounds.

While the complex mathematics in the paper—such as the "weighted combination of vector fields between instructions, frame indices, and models"—may be difficult to fully grasp, the results speak for themselves. Examples on the project’s webpage and in an Nvidia trailer show how ComposableART can generate sounds like a violin "sounding like a laughing baby," a banjo "playing in front of gentle rainfall," or "factory machinery screaming in metallic agony." While some of these examples are more convincing than others, the ability of Fugatto to generate such unique combinations demonstrates its impressive capacity to mix and characterize diverse audio data from a range of open-source datasets.

One of the most intriguing aspects of Fugatto is how it treats each audio trait as a tunable continuum, rather than a simple binary. For example, when blending the sounds of an acoustic guitar and running water, the result can vary significantly depending on whether the guitar or the water is given more weight in Fugatto's interpolated mix. Nvidia also mentions the ability to adjust a French accent to be more or less pronounced, or to alter the "degree of sorrow" in a spoken clip.

In addition to tuning and combining various audio traits, Fugatto can also perform tasks seen in previous models, such as altering the emotion in spoken text or isolating vocal tracks in music. It can also detect individual notes in MIDI music and replace them with different vocal performances, or identify the beat of a song and add effects like drums, barking dogs, or ticking clocks that sync with the rhythm.

While the researchers view Fugatto as just the first step "towards a future where unsupervised multitask learning emerges from data and model scale," Nvidia is already highlighting potential use cases, from song prototyping and dynamically changing video game scores to international ad targeting. However, Nvidia was quick to emphasize that models like Fugatto should be seen as a new tool for audio artists, rather than a replacement for their creative abilities.

"The history of music is also a history of technology," said Ido Zmishlany, producer/songwriter and Nvidia Inception participant, in a blog post. "The electric guitar gave the world rock and roll. When the sampler came along, hip-hop was born. With AI, we're writing the next chapter of music. We now have a new instrument, a new tool for making music—and that’s incredibly exciting."

Press contact

Timon Harz

oneboardhq@outlook.com

The logo for Oneboard Blog

Discover recent post from the Oneboard team.

Notes, simplified.

Follow us

Company

About

Blog

Careers

Press

Legal

Privacy

Terms

Security