A model for generating speech and other audio signals like music developed by Google DeepMind. It uses raw waveforms which is more computationally expensive but produces more natural results. It was one of the early models using a CNN network to generate coherent musical structures.
Year: 2016
Website: https://deepmind.google/discover/blog/wavenet-a-generative-model-for-raw-audio/
Input types: Audio
Output types: Audio
Output length: Variable
AI Technique: Autoregressive Convolutional Neural Network
Dataset: Not disclosed
License type: Proprietary
Real time:
Free:
Open source:
Checkpoints:
Fine-tune:
Train from scratch:
This repository contains a TensorFlow implementation of DeepMind's WaveNet paper: https://github.com/ibab/tensorflow-wavenet.
The repository's README page contains instructions on how to set the project up, train the network on a selected dataset and generate outputs.