Riffusion icon

Riffusion

Open Source
4.1
Riffusion

Riffusion is an innovative AI music generation model that takes a unique approach to audio synthesis by generating spectrograms as images using a fine-tuned version of Stable Diffusion v1.5. Created as a side project by Seth Forsyth and Hayk Martiros in late 2022, Riffusion demonstrated that image diffusion models could be repurposed for audio generation by training on spectrogram representations of music. The model generates mel spectrograms conditioned on text prompts describing musical genres, instruments, moods, and styles, which are then converted back to audio waveforms using the Griffin-Lim algorithm or neural vocoders. This image-based approach to music generation was groundbreaking at the time of release, showing that the powerful generative capabilities of Stable Diffusion could transfer to the audio domain. Riffusion can produce short music clips in various styles including rock, jazz, electronic, classical, and ambient, with real-time interpolation between different prompts enabling smooth musical transitions. The model has approximately 1 billion parameters inherited from its Stable Diffusion base. Released under the MIT license, Riffusion is fully open source with the fine-tuned model weights, training code, and an interactive web application available on GitHub. While newer purpose-built music generation models like MusicGen and Suno have surpassed Riffusion in output quality and duration, the model remains historically significant as a proof of concept that sparked widespread interest in AI music generation. Riffusion continues to be used by hobbyists and researchers exploring the intersection of image generation and audio synthesis.

Text to Audio

Key Highlights

Spectrogram-Based Approach

A unique approach that treats audio spectrograms as images to generate music with Stable Diffusion, bridging image and audio generation

Smooth Style Transitions

Can create seamless smooth transitions between different music genres and styles by blending spectrogram latents

Fully Open Source

Model weights, web application and source code are fully open, allowing anyone to run locally and build upon it

Stable Diffusion Based

Built by fine-tuning Stable Diffusion 1.5, leveraging the existing diffusion model ecosystem and easily extensible

About

Riffusion is an innovative AI music generation model that takes a unique approach to audio synthesis by generating spectrograms as images using a fine-tuned version of Stable Diffusion. Originally created as a side project by Seth Forsyth and Hayk Martiros in late 2022, Riffusion demonstrated that image generation models could be repurposed to create music by treating audio spectrograms as visual representations. This unconventional approach marked a significant conceptual turning point in AI and creative audio generation.

Riffusion's technical architecture is built by fine-tuning the Stable Diffusion 1.5 model on Mel spectrograms. The model converts text prompts into spectrograms as images, which are then transformed into audio waveforms at 44.1 kHz sample rate through inverse Fourier transform (ISTFT). Training data consists of spectrogram-text pairs derived from over 100,000 music clips. A spectrogram interpolation technique enables smooth transitions between two different musical styles. This represents a successful transfer of interpolation techniques from the image diffusion domain to the audio domain.

In terms of performance, Riffusion can generate audio at 44.1 kHz sample rate and complete generation for a single clip in approximately 5 seconds. While the quality of generated music has limitations compared to specialized audio models, the uniqueness and speed of the approach are noteworthy. Text prompts can control genre, tempo, and mood, while spectrogram interpolation enables seamless transitions between two styles. A web-based demo supports real-time interactive music generation for immediate experimentation.

Riffusion is widely used for creative music experiments, rapid prototyping, educational music generation, and interactive sound design projects. It stands out as an inspiring tool particularly in areas where visual arts and music intersect. Its web-based interface allows anyone to generate music without requiring technical expertise. The spectrogram interpolation feature serves as a creative tool for DJs and music producers exploring novel sound transitions.

Riffusion is fully open-source under the MIT license. Model weights, training code, and inference pipeline are accessible via GitHub. Being built on Stable Diffusion, it is compatible with existing diffusion model tools and infrastructure. The interactive web-based demo is available at riffusion.com for public use and the model can run on consumer-grade GPUs.

Riffusion's position in the AI music generation ecosystem is truly unique. While other models generate audio tokens or waveforms directly, Riffusion was one of the first successful examples of bridging image generation technology to the audio domain. This conceptual bridge inspired subsequent research and demonstrated the potential of cross-modal AI systems. Although it falls behind more advanced models like MusicGen or Suno AI in terms of output quality, Riffusion's creative approach and open-source nature make it an important milestone in the history of audio AI.

Delving deeper into Riffusion's technical approach reveals interesting details about the challenges the model faces when transferring Stable Diffusion's image generation capabilities to the audio domain and how it overcomes them. Processing Mel spectrograms as images requires encoding frequency and time axis information as pixel values. While the model experiences a lossy process in this conversion, it can produce music of acceptable quality for human hearing. The spectrogram interpolation technique creates unique musical transition effects by performing linear interpolation between two different prompts in the latent space. This technique is an adapted version of the style mixing concept from image diffusion models applied to the audio domain. The Riffusion community has expanded the project's impact by developing various plugins and interfaces on top of the model, and fine-tuned versions specialized for different music genres are shared by community contributors.

Use Cases

1

Creative Music Experiments

Creating smooth transitions between different music genres and experimental sonic landscapes

2

Live Performance and DJ Sets

Making creative transitions and remixes in live performances with real-time music generation

3

Music Education and Visualization

Using in music theory and signal processing education to demonstrate the relationship between sound and image

4

Prototype and Concept Music

Generating quick music ideas and concepts to serve as inspiration in the early stages of the creative process

Pros & Cons

Pros

  • Generates songs in seconds using unique spectrogram-based diffusion approach bridging image and audio generation
  • User-friendly interface requiring no musical expertise to create music across diverse genres
  • Professional-quality stem separation for isolating individual audio elements
  • Versatile output adapting well to ambient, metal, jazz, experimental, and other genres
  • Free unlimited access during public beta phase for core music generation features

Cons

  • Output quality varies — may not match the creativity or nuance of human-composed music
  • Limited editing options with no advanced arrangement or mixing tools available
  • Voice diversity is a significant concern — overwhelming prevalence of certain vocal profiles limits genre authenticity
  • Only 31% of users find stem quality acceptable for professional remixing without additional processing
  • Phase distortion and limited generation time are unresolved technical bottlenecks

Technical Details

Parameters

1B

Architecture

Fine-tuned Stable Diffusion v1.5 on spectrograms

Training Data

Custom dataset of music spectrograms

License

MIT

Features

  • Spectrogram-to-Audio Generation
  • Stable Diffusion Fine-tuned Architecture
  • Real-Time Style Interpolation
  • Text-to-Music via Spectrograms
  • Open Source Web Application
  • Griffin-Lim Audio Reconstruction

Benchmark Results

MetricValueCompared ToSource
Örnekleme Hızı44.1 kHz (Mel spectrogram)Riffusion GitHub
Üretim Süresi~5 saniye (tek klip)Riffusion Docs
FAD (MusicCaps)11.50MusicGen: 3.80arXiv 2306.05284

Available Platforms

hugging face
replicate

Frequently Asked Questions

Related Models

Suno AI icon

Suno AI

Suno|N/A

Suno AI is a commercial AI music generation platform that creates complete songs with vocals, lyrics, and instrumental arrangements from text descriptions. Founded in 2023 by a team of former Kensho Technologies engineers, Suno AI offers an accessible web interface that enables users to generate professional-sounding songs by simply describing the desired genre, mood, topic, and style in natural language. The platform uses a proprietary transformer-based architecture that generates all components of a song including melody, harmony, rhythm, instrumentation, vocal performance, and lyrics in a single integrated process. Suno AI supports a remarkably wide range of musical genres from pop and rock to hip-hop, country, classical, electronic, jazz, and experimental styles, producing outputs that often sound indistinguishable from human-created music to casual listeners. Generated songs can be up to several minutes in duration and include realistic singing voices with proper pronunciation, emotional expression, and musical phrasing. The platform allows users to provide custom lyrics or let the AI generate lyrics based on a theme or concept. Suno AI operates on a freemium subscription model with limited free generations and paid tiers for higher volume and commercial usage rights. The platform has gained significant attention for democratizing music creation, enabling people without musical training to produce complete songs. Suno AI is particularly popular among content creators, social media marketers, hobbyist musicians, and anyone needing original music for videos, podcasts, or personal projects without the cost and complexity of traditional music production.

Proprietary
4.7
MusicGen icon

MusicGen

Meta|3.3B

MusicGen is a single-stage transformer-based music generation model developed by Meta AI Research as part of the AudioCraft framework. Released in June 2023 under the MIT license, MusicGen uses a single autoregressive language model operating over compressed discrete audio representations from EnCodec, unlike cascading approaches that require multiple models. The model comes in multiple sizes ranging from 300M to 3.3B parameters, allowing users to balance quality against computational requirements. MusicGen generates high-quality mono and stereo music at 32 kHz from text descriptions, supporting a wide range of genres, instruments, moods, and musical styles. Users can describe desired music using natural language prompts specifying genre, tempo, instrumentation, and atmosphere, and the model produces coherent musical compositions that follow the specified characteristics. Beyond text-to-music generation, MusicGen supports melody conditioning where an existing audio clip guides the melodic structure of the generated output, enabling more controlled music creation. The model achieves strong results across both objective metrics and subjective listening evaluations, producing music that sounds natural and musically coherent for durations up to 30 seconds. As a fully open-source model with code and weights available on GitHub and Hugging Face, MusicGen has become one of the most widely adopted AI music generation tools in both research and creative communities. It integrates easily into existing audio production workflows through the Audiocraft Python library and various community-built interfaces. MusicGen is particularly popular among content creators, game developers, and musicians who need royalty-free background music generated on demand.

Open Source
4.6
Udio icon

Udio

Udio|N/A

Udio is an AI music generation platform developed by former Google DeepMind researchers that creates high-quality songs with vocals, lyrics, and instrumentals from text prompts. Launched in April 2024, Udio quickly gained attention for producing remarkably realistic and musically coherent outputs that rival professional studio recordings in audio fidelity. The platform uses a proprietary transformer-based architecture that generates all aspects of a musical composition including vocal performances, instrumental arrangements, harmonies, and production effects in a unified process. Udio supports an extensive range of musical genres and styles from mainstream pop and rock to niche genres like lo-fi, synthwave, Afrobeat, and traditional folk music from various cultures. Generated songs feature studio-quality audio at high sample rates with realistic vocal timbres, proper musical dynamics, and professional-sounding mixing and mastering. The platform allows users to provide custom lyrics, specify song structure, and control various musical parameters through text descriptions. Udio also supports audio extensions where users can generate additional sections to extend existing songs, enabling the creation of full-length tracks through iterative generation. The platform operates on a freemium model with free daily generations and paid subscription tiers for commercial use and higher generation limits. Udio is particularly notable for its vocal quality, which includes natural-sounding vibrato, breath sounds, and emotional expressiveness that many competing platforms struggle to achieve. The platform is popular among content creators, independent musicians exploring AI-assisted composition, marketing teams needing original music, and hobbyists who want to create professional-sounding songs without musical training or expensive production equipment.

Proprietary
4.6
Bark icon

Bark

Suno AI|N/A

Bark is a transformer-based text-to-audio generation model developed by Suno AI that converts text into natural-sounding speech, music, and sound effects. Released as open source under the MIT license in April 2023, Bark goes far beyond traditional text-to-speech systems by generating not only spoken words but also laughter, sighs, music, and ambient sounds from text descriptions. The model uses a GPT-style autoregressive transformer architecture with EnCodec audio tokenizer to generate audio tokens that are then decoded into waveforms. Bark supports multiple languages including English, Chinese, French, German, Hindi, Italian, Japanese, Korean, Polish, Portuguese, Russian, Spanish, and Turkish, making it one of the most multilingual open-source audio generation models available. The model can clone voice characteristics from short audio samples, allowing users to generate speech in specific voices or speaking styles. Bark operates in a zero-shot manner, meaning it can produce diverse outputs without task-specific fine-tuning. Generation includes natural prosody, emotion, and intonation that closely mimics human speech patterns. The model generates audio at 24 kHz sample rate with reasonable quality for most applications. As a fully open-source project with pre-trained weights available on Hugging Face and GitHub, Bark is widely used by developers building voice applications, content creators producing multilingual audio, and researchers exploring generative audio models. The model is particularly valued for its versatility in handling diverse audio types within a single unified architecture and its accessibility for rapid prototyping of audio generation applications.

Open Source
4.4

Quick Info

Parameters1B
Typediffusion
LicenseMIT
Released2022-12
ArchitectureFine-tuned Stable Diffusion v1.5 on spectrograms
Rating4.1 / 5
CreatorRiffusion

Links

Tags

riffusion
music
spectrograms
text-to-audio
Visit Website