Synth vox

From Things and Stuff Wiki
Jump to navigation Jump to search


Synthesis

to sort/categorise



SAM


rsynth


Festival

  • Festival - or The Festival Speech Synthesis System, offers a general framework for building speech synthesis systems as well as including examples of various modules. As a whole it offers full text to speech through a number APIs: from shell level, though a Scheme command interpreter, as a C++ library, from Java, and an Emacs interface. Festival is multi-lingual (currently English (British and American), and Spanish) though English is the most advanced. Other groups release new languages for the system.
  • Festvox - aims to make the building of new synthetic voices more systemic and better documented, making it possible for anyone to build a new voice.


Rocaloid

  • Rocaloid - a free, open-source singing voice synthesis system. Its ultimate goal is to fast synthesize natural, flexible and multi-lingual vocal parts.  Like other vocal synthesizing software, after installing the vocal database, inputting lyrics and pitch, you can synthesize attractive vocal parts. What’s more, Rocaloid highlights on providing you more controllable parameters which enabling to take control of exquisite dimensions of the synthesized voice and export with better quality. By using a fully constructed Rocaloid Database, you can synthesize singing voice in any phonetic-based languages.

Festvox

  • Festvox - aims to make the building of new synthetic voices more systemic and better documented, making it possible for anyone to build a new voice. Specifically we offer: Documentation, including scripts explaining the background and specifics for building new voices for speech synthesis in new and supported languages. Example speech databases to help building new voices. Links, demos and a repository for new voices. This work is firmly grounded within Edinburgh University's Festival Speech Synthesis System and Carnegie Mellon University's small footprint Flite synthesis engine.


MaryTTS

  • MaryTTS - an open-source, multilingual Text-to-Speech Synthesis platform written in Java. It was originally developed as a collaborative project of DFKI’s Language Technology Lab and the Institute of Phonetics at Saarland University. It is now maintained by the Multimodal Speech Processing Group in the Cluster of Excellence MMCI and DFKI.



eSpeak

  • eSpeak - a compact open source software speech synthesizer for English and other languages, for Linux and Windows. eSpeak uses a "formant synthesis" method. This allows many languages to be provided in a small size. The speech is clear, and can be used at high speeds, but is not as natural or smooth as larger synthesizers which are based on human speech recordings.


  • https://github.com/divVerent/ecantorix - a singing synthesis frontend for espeak. It works by using espeak to generate raw speech samples, then adjusting their pitch and length and finally creating a LMMS project file referencing the samples in sync to the input file.

OpenSource SpeechSynth


MBROLA

Assistive Context-Aware Toolkit


Praat

  • Praat - doing phonetics by computer


Gnuspeech

  • gnuspeech - makes it easy to produce high quality computer speech output, design new language databases, and create controlled speech stimuli for psychophysical experiments. gnuspeechsa is a cross-platform module of gnuspeech that allows command line, or application-based speech output. The software has been released as two tarballs that are available in the project Downloads area of http://savannah.gnu.org/projects/gnuspeech. [2]


Project Merlin


UTAU

  • https://en.wikipedia.org/wiki/Utau - a Japanese singing synthesizer application created by Ameya/Ayame. This program is similar to the Vocaloid software, with the difference that it is shareware instead of being released under third party licensing




Sinsy





UTSU

qtau

cadencii

  • https://github.com/cadencii/cadencii - simple musical score editor for singing synthesis: VOCALOID, VOCALOID2, UTAU, STRAIGHT with UTAU, and AquesTone are available as synthesizer.

Mozilla TTS

CMU Flite

  • CMU Flite - a small, fast run-time open source text to speech synthesis engine developed at CMU and primarily designed for small embedded machines and/or large servers. Flite is designed as an alternative text to speech synthesis engine to Festival for voices built using the FestVox suite of voice building tools.

mesing


Adobe VoCo


VST Speek


char2wav


loop


IPOX

  • IPOX - an experimental, all-prosodic speech synthesizer, developed many years ago by Arthur Dirksen and John Coleman. It is still available for downloading, and was designed to run on a 486 PC running Windows 3.1 or higher, with a 16-bit Windows-compatible sound card, such as the Soundblaster 16. It still seems to run on e.g. XP, but I haven't tried it on Vista.


NPSS


Pink Trombone

Klatter

  • https://github.com/fundamental/klatter - a bare bones formant synthesizer based upon the description given in the 1979 paper "Software For a Cascade/Parallel Formant Synthesizer" by Dennis Klatt. This program was not designed for interactive use, though there is code for some minimal midi control. In it's current state, it is enough of a curiosity that it will be preserved, though it may not see much if any use.


Tacotron 2

SqueezeWave

  • https://github.com/tianrengao/SqueezeWave - Automatic speech synthesis is a challenging task that is becoming increasingly important as edge devices begin to interact with users through speech. Typical text-to-speech pipelines include a vocoder, which translates intermediate audio representations into an audio waveform. Most existing vocoders are difficult to parallelize since each generated sample is conditioned on previous samples. WaveGlow is a flow-based feed-forward alternative to these auto-regressive models (Prenger et al., 2019). However, while WaveGlow can be easily parallelized, the model is too expensive for real-time speech synthesis on the edge. This paper presents SqueezeWave, a family of lightweight vocoders based on WaveGlow that can generate audio of similar quality to WaveGlow with 61x - 214x fewer MACs.

WaveGlow

  • https://github.com/NVIDIA/waveglow - In our recent paper, we propose WaveGlow: a flow-based network capable of generating high quality speech from mel-spectrograms. WaveGlow combines insights from Glow and WaveNet in order to provide fast, efficient and high-quality audio synthesis, without the need for auto-regression. WaveGlow is implemented using only a single network, trained using only a single cost function: maximizing the likelihood of the training data, which makes the training procedure simple and stable.

STT

Real-Time-Voice-Cloning

rapping-neural-network

yukarin

leesampler

VoiceOfFaust

tomomibot

Nanceloid

Parakeet

PaddleSpeech

Flowtron

TransformerTTS

TensorflowTTS

AutoSpeech

HiFi-GAN

Wave-U-net-TF2

larynx

FastSpeech2

TensorVox

  • https://github.com/ZDisket/TensorVox - an application designed to enable user-friendly and lightweight neural speech synthesis in the desktop, aimed at increasing accessibility to such technology.Powered by TensorflowTTS, it is written in pure C++/Qt, using the Tensorflow C API for interacting with the models. This way, we can perform inference without having to install gigabytes worth of pip libraries, just a 100MB DLL.


Coqui TTS

  • Coqui TTS - an open source embedded (offline, on-device) speech-to-text engine which can run in real time on devices ranging from a Raspberry Pi 4 to high power GPU servers.Coqui STT has APIs for numerous languages (Python, C/C++, Java, JavaScript, .NET...), is supported on many platforms (Linux, macOS, Windows, ARM...), and is available on GitHub.


vits

phonemizer

espeak-phonemizer

nnsvs

unagan

  • https://github.com/ciaua/unagan - contains the code and samples for our paper "Unconditional Audio Generation with GAN and Cycle Regularization", accepted by INTERSPEECH 2020. The goal is to unconditionally generate singing voices, speech, and instrument sounds with GAN.The model is implemented with PyTorch.

vocshape

  • https://github.com/PaulBatchelor/vocshape - a very simple proof-of-concept musical instrument for Android that aims to demonstrate the sculptability of a simple articulatory synthesis physical model for vocal synthesis.

lexconvert

HiFiSinger

  • https://github.com/CODEJIN/HiFiSinger - This code is an unofficial implementation of HiFiSinger. The algorithm is based on the following papers:Chen, J., Tan, X., Luan, J., Qin, T., & Liu, T. Y. (2020). HiFiSinger: Towards High-Fidelity Neural Singing Voice Synthesis. arXiv preprint arXiv:2009.01776.Ren, Y., Ruan, Y., Tan, X., Qin, T., Zhao, S., Zhao, Z., & Liu, T. Y. (2019). Fastspeech: Fast, robust and controllable text to speech. Advances in Neural Information Processing Systems, 32, 3171-3180.Yamamoto, R., Song, E., & Kim, J. M. (2020, May). Parallel WaveGAN: A fast waveform generation model based on generative adversarial networks with multi-resolution spectrogram. In ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (pp. 6199-6203). IEEE.

PortaSpeech

KaraSinger

hifigan

VOICEVOX

Comprehensive-Transformer-TTS

DiffGAN-TTS

DiffSinger

tortoise-tts

Conversion

See also Effects#Pitch shifting

crank

MelGAN-VC

World

  • https://github.com/mmorise/World - free software for high-quality speech analysis, manipulation and synthesis. It can estimate Fundamental frequency (F0), aperiodicity and spectral envelope and also generate the speech like input speech with only estimated parameters.

Scyclone

Shallow WaveNet Vocoder

speech-resynthesis