WaveNet

A model for generating speech and other audio signals like music developed by Google DeepMind. It uses raw waveforms which is more computationally expensive but produces more natural results. It was one of the early models using a CNN network to generate coherent musical structures.

Year: 2016

Website: https://deepmind.google/discover/blog/wavenet-a-generative-model-for-raw-audio/

Input types: Audio

Output types: Audio

Output length: Variable

AI Technique: Autoregressive Convolutional Neural Network

Dataset: Not disclosed

License type: Proprietary

Real time:

Free:

Open source:

Checkpoints:

Fine-tune:

Train from scratch:

#open-source #free #checkpoints

Guide to using the model

This repository contains a TensorFlow implementation of DeepMind's WaveNet paper: https://github.com/ibab/tensorflow-wavenet.

The repository's README page contains instructions on how to set the project up, train the network on a selected dataset and generate outputs.

Edit