A.I. for speech synthesis is the next generation of Text To Speech (TTS) Software. In my last article I outlined how to setup a range of TTS systems to read back any text anywhere on your Linux system using keyboard commands and your choice of a TTS system and voice.
A.I. systems offer much more natural sounding voices, but can sometimes be less accurate depending on a range of factors. They also can’t change their reading speed (exception Mimic3 and Larynx) and often require much more compute to output the same amount of text as traditional software.
I tried the following projects when looking at A.I. for Linux TTS:
- Mozilla TTS
- Coqui TTS
- Larynx TTS
- Mimic3 TTS
I was only able to get Coqui TTS and Mimic3 to run on my system. Coqui was to slow and required to much system resources for my real time use case. So I will focus on Mimic3 for this setup.
Below shows how I integrate Mimic3 TTS into my system for everyday use on a Linux desktop. Recommended for users with more powerful systems only as these TTS systems can use up a lot of resources.
For mimic3 there are compiled binaries for Debian based systems you can install from here. I did try when the source install was not working to use
debtap to make them work on Arch. They worked to a point but the
--interactive flag gave errors so I removed it and gave up on that approach.
I’m on Arch, I originally was unable to install mimic3 over the last few months due to some python packages requiring older versions of python. See here. But now the installation process is working well and it goes like this:
git clone https://github.com/mycroftAI/mimic3.git
(being in China this did not work so had to run the git command and the
./install.sh command prefixed by running the
A video about installing Mimic3 can be found here.
After installing mimic you can set it to read any highlighted text like this, together with a custom keyboard shortcut:
<PATH TO WHERE YOU INSTALLED MIMIC3>mimic3/.venv/bin/mimic3 --voice --voice en_US/vctk_low --interactive "$(xclip -selection primary -out)"
To stop mimic3 use this command:
- --voce flag sets the voice you want mimic3 to use.
--interactiveflag sets mimic3 to send the audio directly to hear it on your system, instead of saving the audio to a file.
--cudaflag can be added if you want to use mimic3 with your graphics card for better performance.
--noise-wDetermine the speaker volatility during synthesis, 0-1, default is 0.667 and 0.8 respectively
--length-scalemakes the voice speaker slower (> 1) or faster (< 1)/li>
Mimic3 documentation can be found here. Mimic3 github can be found here. In the documentation it also explains how to get Mimic3 to work with Speech Dispatcher. I have not tried this, it seams easier to just directly integrate it into KDE or gnome.
View a list of available mimic3 voices here. There is some trouble shooting documentation here.
Now you have everything working and setup you can integrate this with gnome or KDE. Create a custom shortcut for two hot key commands, one to start and one to stop mimic3.
KDE example: Adding a bash command to a custom keyboard shortcut/home/eben/mimic3/.venv/bin/mimic3 --voice en_US/vctk_low --interactive "$(xclip -selection primary -out)"
Gnome example: Adding a bash command to a custom keyboard shortcut
Open System Settings then go to: Keyboard » Customise Shortcuts » Custom Shortcuts » +:
Notes and links to other A.I. projects:
(Could not install, seams like this is not maintained)
(Could install but uses to much system resources)
I found Larynx from this reddit post and spent some time trying to get it to run on my system, as it seamed to offer a good set of performance and features. But I could not install, seams some of it’s dependencies are to old and out of date. It also seams not to be maintained any more.
$ larynx -v en "$(xclip -selection primary -out)"
A nice code snippet to get larynx to read any selected text system wide.
“I have a version of Larynx working with speech-dispatcher”