AIVA is an AI music generation assistant that allows you to generate new songs in more than 250 different styles, in a matter of seconds. Whether a complete beginner or a seasoned professional in music making, use the power of generative AI to create your own songs.
Year: 2019
Website: https://www.aiva.ai/
Input types: Audio MIDI
Output types: Audio MIDI
Output length: 5:30
AI Technique: Not Specified
Dataset: Not disclosed
License type: Proprietary
Real time:
Free:
Open source:
Checkpoints:
Fine-tune:
Train from scratch:
AI-powered songwriting assistant as a mobile app. Enables generating an entire songs sketch or just a section with a single tap. The music can be downloaded or exported as MIDI for easy loading in a DAW.
Year: 2019
Website: https://amadeuscode.com/
Input types: Genre
Output types: Audio MIDI
Output length: Variable
AI Technique: Not Specified
Dataset: Not disclosed
License type: Proprietary
Real time:
Free:
Open source:
Checkpoints:
Fine-tune:
Train from scratch:
Rhythm generator using Variational Autoencoder (VAE). Based on M4L.RhythmVAE by Nao Tokui, modded and extended to support simple and compound meter rhythms, with minimal amount of training data. Similarly to RhythmVAE, the goal of R-VAE is the exploration of latent spaces of musical rhythms. Unlike most previous work in rhythm modeling, R-VAE can be trained with small datasets, enabling rapid customization and exploration by individual users. R-VAE employs a data representation that encodes simple and compound meter rhythms. Models and latent space visualizations for R-VAE are available on the project's GitHub page: https://github.com/vigliensoni/R-VAE-models.
Year: 2022
Website: https://github.com/vigliensoni/R-VAE
Input types: MIDI
Output types: MIDI
Output length: 2 bars
AI Technique: VAE
Dataset: "The Future Sample Pack"
License type: GPLv3
Real time:
Free:
Open source:
Checkpoints:
Fine-tune:
Train from scratch:
Generates MIDI notes that are likely to follow the input drum beat or melody. Can extend the input of a specified MIDI clip by up to 32 measures. This can be helpful for adding variation to a drum beat or creating new material for a melodic track. It typically picks up on things like durations, key signatures and timing. It can be used to produce more random outputs by increasing the temperature. Ready to use as a Max for Live device. If you want to train the model on your own data or try different pre-trained models provided by the Magenta team, refer to the instructions on the team's GitHub page: https://github.com/magenta/magenta/tree/main/magenta/models/melody_rnn
Year: 2018
Website: https://magenta.tensorflow.org/studio#continue
Input types: MIDI
Output types: MIDI
Output length: 32 bars
AI Technique: LSTM
Dataset: MelodyRNN - Not disclosed; PerformanceRNN - The Piano- e-Competition dataset
License type: Apache 2.0
Real time:
Free:
Open source:
Checkpoints:
Fine-tune:
Train from scratch:
Creates grooves based on the rhythm of any input. Can be used to generate a drum accompaniment to a bassline or melody, or to create a drum track from a tapped rhythm. Drumify works best with performed inputs, but it can also handle quantized clips. Ready to use as a Max for Live device. If you want to train the model on your own data or try different pre-trained models provided by the Magenta team, refer to the instructions on the team's GitHub page: https://github.com/magenta/magenta/tree/main/magenta/models/drums_rnn
Year: 2018
Website: https://magenta.tensorflow.org/studio#drumify
Input types: MIDI
Output types: MIDI
Output length: Input length
AI Technique: LSTM
Dataset: Expanded Groove MIDI dataset; Groove MIDI dataset
License type: Apache 2.0
Real time:
Free:
Open source:
Checkpoints:
Fine-tune:
Train from scratch:
Generates a 4 bar phrase with no input necessary. It is possible to control the number of variations and temperature. The model can be helpful for breaking a creative block or as a source of inspiration for an original sample. Under the hood it uses MusicVAE. You can learn more about it here: https://magenta.tensorflow.org/music-vae. Ready to use as a Max for Live device. If you want to train the model on your own data or try different pre-trained models provided by the Magenta team, refer to the instructions on the team's GitHub page: https://github.com/magenta/magenta/tree/main/magenta/models/music_vae
Year: 2018
Website: https://magenta.tensorflow.org/studio#generate
Input types: None
Output types: MIDI
Output length: 4 bars
AI Technique: VAE
Dataset: "Millions of melodies and rhythms", including NSynth Dataset, MAESTRO dataset, Lakh MIDI dataset
License type: Apache 2.0
Real time:
Free:
Open source:
Checkpoints:
Fine-tune:
Train from scratch:
An online platform for music generation based on a predefined genre or style. Users can select a format suitable for a specific type of content (eg. social media, gaming, vlogs), or type of output (eg. loops, sfx).
Year: 2023
Website: https://soundful.com/
Input types: Text Metadata
Output types: Audio MIDI
Output length: 2.5min
AI Technique: Not Specified
Dataset: Not disclosed
License type: Proprietary
Real time:
Free:
Open source:
Checkpoints:
Fine-tune:
Train from scratch:
Music FaderNets is a controllable MIDI generation framework that models high-level musical qualities, such as emotional attributes like arousal. Drawing inspiration from the concept of sliding faders on a mixing console, the model offers intuitive and continuous control over these characteristics. Given an input MIDI, Music FaderNets can produce multiple variations with different levels of arousal, adjusted according to the position of the fader.
Year: 2020
Website: https://music-fadernets.github.io/
Input types: MIDI
Output types: MIDI
Output length:
AI Technique: VAE
Dataset: VGMIDI, Yamaha Piano-e-Competition
License type: MIT
Real time:
Free:
Open source:
Checkpoints:
Fine-tune:
Train from scratch:
TwoShot's Coproducer is an all-in-one AI assistant that helps creators produce high-quality, commercially-safe audio. The platform enables users to: * Generate full tracks from hummed melodies or simple text prompts. * Remix existing songs, split audio into stems, and create unique samples. * Automatically score video scenes with context-aware sound effects. Designed for both beginners and professionals, The coproducer integrates seamlessly with industry-standard DAWs (e.g., Ableton, Logic) and is built on a 100% ethically-sourced, rights-cleared foundation.
Year: 2025
Website: https://twoshot.ai/coproducer
Input types: Audio MIDI Text Genre Metadata
Output types: Audio MIDI
Output length:
AI Technique: Suite of AI tools
Dataset: Proprietary licenced dataset
License type: Depends on inputs/models. Possible to generate royalty free content in many ways
Real time:
Free:
Open source:
Checkpoints:
Fine-tune:
Train from scratch: