fix: update repository links, package names, metadata

This commit is contained in:
Enno Hermann 2024-04-03 11:28:22 +02:00
parent 018daa002b
commit d772724125
18 changed files with 70 additions and 68 deletions

View File

@ -10,8 +10,8 @@ authors:
version: 1.4
doi: 10.5281/zenodo.6334862
license: "MPL-2.0"
url: "https://www.coqui.ai"
repository-code: "https://github.com/coqui-ai/TTS"
url: "https://github.com/eginhard/coqui-tts"
repository-code: "https://github.com/eginhard/coqui-tts"
keywords:
- machine learning
- deep learning

View File

@ -2,7 +2,7 @@
Welcome to the 🐸TTS!
This repository is governed by [the Contributor Covenant Code of Conduct](https://github.com/coqui-ai/TTS/blob/main/CODE_OF_CONDUCT.md).
This repository is governed by [the Contributor Covenant Code of Conduct](https://github.com/eginhard/coqui-tts/blob/main/CODE_OF_CONDUCT.md).
## Where to start.
We welcome everyone who likes to contribute to 🐸TTS.
@ -15,13 +15,13 @@ If you like to contribute code, squash a bug but if you don't know where to star
You can pick something out of our road map. We keep the progess of the project in this simple issue thread. It has new model proposals or developmental updates etc.
- [Github Issues Tracker](https://github.com/coqui-ai/TTS/issues)
- [Github Issues Tracker](https://github.com/eginhard/coqui-tts/issues)
This is a place to find feature requests, bugs.
Issues with the ```good first issue``` tag are good place for beginners to take on.
- ✨**PR**✨ [pages](https://github.com/coqui-ai/TTS/pulls) with the ```🚀new version``` tag.
- ✨**PR**✨ [pages](https://github.com/eginhard/coqui-tts/pulls) with the ```🚀new version``` tag.
We list all the target improvements for the next version. You can pick one of them and start contributing.
@ -46,14 +46,14 @@ Let us know if you encounter a problem along the way.
The following steps are tested on an Ubuntu system.
1. Fork 🐸TTS[https://github.com/coqui-ai/TTS] by clicking the fork button at the top right corner of the project page.
1. Fork 🐸TTS[https://github.com/eginhard/coqui-tts] by clicking the fork button at the top right corner of the project page.
2. Clone 🐸TTS and add the main repo as a new remote named ```upstream```.
```bash
$ git clone git@github.com:<your Github name>/TTS.git
$ cd TTS
$ git remote add upstream https://github.com/coqui-ai/TTS.git
$ git clone git@github.com:<your Github name>/coqui-tts.git
$ cd coqui-tts
$ git remote add upstream https://github.com/eginhard/coqui-tts.git
```
3. Install 🐸TTS for development.
@ -105,7 +105,7 @@ The following steps are tested on an Ubuntu system.
```bash
$ git fetch upstream
$ git rebase upstream/master
$ git rebase upstream/main
# or for the development version
$ git rebase upstream/dev
```
@ -124,7 +124,7 @@ The following steps are tested on an Ubuntu system.
13. Let's discuss until it is perfect. 💪
We might ask you for certain changes that would appear in the ✨**PR**✨'s page under 🐸TTS[https://github.com/coqui-ai/TTS/pulls].
We might ask you for certain changes that would appear in the ✨**PR**✨'s page under 🐸TTS[https://github.com/eginhard/coqui-tts/pulls].
14. Once things look perfect, We merge it to the ```dev``` branch and make it ready for the next version.
@ -132,14 +132,14 @@ The following steps are tested on an Ubuntu system.
If you prefer working within a Docker container as your development environment, you can do the following:
1. Fork 🐸TTS[https://github.com/coqui-ai/TTS] by clicking the fork button at the top right corner of the project page.
1. Fork 🐸TTS[https://github.com/eginhard/coqui-tts] by clicking the fork button at the top right corner of the project page.
2. Clone 🐸TTS and add the main repo as a new remote named ```upsteam```.
```bash
$ git clone git@github.com:<your Github name>/TTS.git
$ cd TTS
$ git remote add upstream https://github.com/coqui-ai/TTS.git
$ git clone git@github.com:<your Github name>/coqui-tts.git
$ cd coqui-tts
$ git remote add upstream https://github.com/eginhard/coqui-tts.git
```
3. Build the Docker Image as your development environment (it installs all of the dependencies for you):

View File

@ -26,7 +26,7 @@ ______________________________________________________________________
[![Discord](https://img.shields.io/discord/1037326658807533628?color=%239B59B6&label=chat%20on%20discord)](https://discord.gg/5eXr5seRrv)
[![License](<https://img.shields.io/badge/License-MPL%202.0-brightgreen.svg>)](https://opensource.org/licenses/MPL-2.0)
[![PyPI version](https://badge.fury.io/py/TTS.svg)](https://badge.fury.io/py/TTS)
[![Covenant](https://camo.githubusercontent.com/7d620efaa3eac1c5b060ece5d6aacfcc8b81a74a04d05cd0398689c01c4463bb/68747470733a2f2f696d672e736869656c64732e696f2f62616467652f436f6e7472696275746f72253230436f76656e616e742d76322e3025323061646f707465642d6666363962342e737667)](https://github.com/eginhard/coqui-tts/blob/master/CODE_OF_CONDUCT.md)
[![Covenant](https://camo.githubusercontent.com/7d620efaa3eac1c5b060ece5d6aacfcc8b81a74a04d05cd0398689c01c4463bb/68747470733a2f2f696d672e736869656c64732e696f2f62616467652f436f6e7472696275746f72253230436f76656e616e742d76322e3025323061646f707465642d6666363962342e737667)](https://github.com/eginhard/coqui-tts/blob/main/CODE_OF_CONDUCT.md)
[![Downloads](https://pepy.tech/badge/tts)](https://pepy.tech/project/tts)
[![DOI](https://zenodo.org/badge/265612440.svg)](https://zenodo.org/badge/latestdoi/265612440)
@ -62,7 +62,7 @@ Please use our dedicated channels for questions and discussion. Help is much mor
| 💾 **Installation** | [TTS/README.md](https://github.com/eginhard/coqui-tts/tree/dev#installation)|
| 👩‍💻 **Contributing** | [CONTRIBUTING.md](https://github.com/eginhard/coqui-tts/blob/main/CONTRIBUTING.md)|
| 📌 **Road Map** | [Main Development Plans](https://github.com/coqui-ai/TTS/issues/378)
| 🚀 **Released Models** | [TTS Releases](https://github.com/coqui-ai/TTS/releases) and [Experimental Models](https://github.com/coqui-ai/TTS/wiki/Experimental-Released-Models)|
| 🚀 **Released Models** | [Standard models](https://github.com/eginhard/coqui-tts/blob/dev/TTS/.models.json) and [Fairseq models in ~1100 languages](https://github.com/eginhard/coqui-tts#example-text-to-speech-using-fairseq-models-in-1100-languages-)|
| 📰 **Papers** | [TTS Papers](https://github.com/erogol/TTS-papers)|
## Features
@ -135,7 +135,7 @@ You can also help us implement more models.
If you are only interested in [synthesizing speech](https://coqui-tts.readthedocs.io/en/latest/inference.html) with the released 🐸TTS models, installing from PyPI is the easiest option.
```bash
pip install TTS
pip install coqui-tts
```
If you plan to code or train models, clone 🐸TTS and install it locally.
@ -152,7 +152,9 @@ $ make system-deps # intended to be used on Ubuntu (Debian). Let us know if you
$ make install
```
If you are on Windows, 👑@GuyPaddock wrote installation instructions [here](https://stackoverflow.com/questions/66726331/how-can-i-run-mozilla-tts-coqui-tts-training-with-cuda-on-a-windows-system).
If you are on Windows, 👑@GuyPaddock wrote installation instructions
[here](https://stackoverflow.com/questions/66726331/how-can-i-run-mozilla-tts-coqui-tts-training-with-cuda-on-a-windows-system)
(note that these are out of date, e.g. you need to have at least Python 3.9).
## Docker Image

View File

@ -1,8 +1,8 @@
# :frog: TTS demo server
Before you use the server, make sure you
[install](https://github.com/coqui-ai/TTS/tree/dev#install-tts)) :frog: TTS
[install](https://github.com/eginhard/coqui-tts/tree/dev#install-tts)) :frog: TTS
properly and install the additional dependencies with `pip install
TTS[server]`. Then, you can follow the steps below.
coqui-tts[server]`. Then, you can follow the steps below.
**Note:** If you install :frog:TTS using ```pip```, you can also use the ```tts-server``` end point on the terminal.

View File

@ -12,7 +12,7 @@ from urllib.parse import parse_qs
try:
from flask import Flask, render_template, render_template_string, request, send_file
except ImportError as e:
raise ImportError("Server requires requires flask, use `pip install TTS[server]`.") from e
raise ImportError("Server requires requires flask, use `pip install coqui-tts[server]`.") from e
from TTS.config import load_config
from TTS.utils.manage import ModelManager

View File

@ -30,7 +30,7 @@
</head>
<body>
<a href="https://github.com/coqui-ai/TTS"><img style="position: absolute; z-index:1000; top: 0; left: 0; border: 0;"
<a href="https://github.com/eginhard/coqui-tts"><img style="position: absolute; z-index:1000; top: 0; left: 0; border: 0;"
src="https://s3.amazonaws.com/github/ribbons/forkme_left_darkblue_121621.png" alt="Fork me on GitHub"></a>
<!-- Navigation -->
@ -151,4 +151,4 @@
</body>
</html>
</html>

View File

@ -695,12 +695,12 @@ class Xtts(BaseTTS):
def forward(self):
raise NotImplementedError(
"XTTS has a dedicated trainer, please check the XTTS docs: https://tts.readthedocs.io/en/dev/models/xtts.html#training"
"XTTS has a dedicated trainer, please check the XTTS docs: https://coqui-tts.readthedocs.io/en/dev/models/xtts.html#training"
)
def eval_step(self):
raise NotImplementedError(
"XTTS has a dedicated trainer, please check the XTTS docs: https://tts.readthedocs.io/en/dev/models/xtts.html#training"
"XTTS has a dedicated trainer, please check the XTTS docs: https://coqui-tts.readthedocs.io/en/dev/models/xtts.html#training"
)
@staticmethod
@ -789,5 +789,5 @@ class Xtts(BaseTTS):
def train_step(self):
raise NotImplementedError(
"XTTS has a dedicated trainer, please check the XTTS docs: https://tts.readthedocs.io/en/dev/models/xtts.html#training"
"XTTS has a dedicated trainer, please check the XTTS docs: https://coqui-tts.readthedocs.io/en/dev/models/xtts.html#training"
)

View File

@ -63,7 +63,7 @@ def get_phonemizer_by_name(name: str, **kwargs) -> BasePhonemizer:
return ZH_CN_Phonemizer(**kwargs)
if name == "ja_jp_phonemizer":
if JA_JP_Phonemizer is None:
raise ValueError(" ❗ You need to install JA phonemizer dependencies. Try `pip install TTS[ja]`.")
raise ValueError(" ❗ You need to install JA phonemizer dependencies. Try `pip install coqui-tts[ja]`.")
return JA_JP_Phonemizer(**kwargs)
if name == "ko_kr_phonemizer":
return KO_KR_Phonemizer(**kwargs)

View File

@ -3,7 +3,7 @@ We tried to collect common issues and questions we receive about 🐸TTS. It is
## Errors with a pre-trained model. How can I resolve this?
- Make sure you use the right commit version of 🐸TTS. Each pre-trained model has its corresponding version that needs to be used. It is defined on the model table.
- If it is still problematic, post your problem on [Discussions](https://github.com/coqui-ai/TTS/discussions). Please give as many details as possible (error message, your TTS version, your TTS model and config.json etc.)
- If it is still problematic, post your problem on [Discussions](https://github.com/eginhard/coqui-tts/discussions). Please give as many details as possible (error message, your TTS version, your TTS model and config.json etc.)
- If you feel like it's a bug to be fixed, then prefer Github issues with the same level of scrutiny.
## What are the requirements of a good 🐸TTS dataset?
@ -16,7 +16,7 @@ We tried to collect common issues and questions we receive about 🐸TTS. It is
- If you need faster models, consider SpeedySpeech, GlowTTS or AlignTTS. Keep in mind that SpeedySpeech requires a pre-trained Tacotron or Tacotron2 model to compute text-to-speech alignments.
## How can I train my own `tts` model?
0. Check your dataset with notebooks in [dataset_analysis](https://github.com/coqui-ai/TTS/tree/master/notebooks/dataset_analysis) folder. Use [this notebook](https://github.com/coqui-ai/TTS/blob/master/notebooks/dataset_analysis/CheckSpectrograms.ipynb) to find the right audio processing parameters. A better set of parameters results in a better audio synthesis.
0. Check your dataset with notebooks in [dataset_analysis](https://github.com/eginhard/coqui-tts/tree/main/notebooks/dataset_analysis) folder. Use [this notebook](https://github.com/eginhard/coqui-tts/blob/main/notebooks/dataset_analysis/CheckSpectrograms.ipynb) to find the right audio processing parameters. A better set of parameters results in a better audio synthesis.
1. Write your own dataset `formatter` in `datasets/formatters.py` or format your dataset as one of the supported datasets, like LJSpeech.
A `formatter` parses the metadata file and converts a list of training samples.

View File

@ -4,7 +4,7 @@
First, you need to install TTS. We recommend using PyPi. You need to call the command below:
```bash
$ pip install TTS
$ pip install coqui-tts
```
After the installation, 2 terminal commands are available.
@ -14,7 +14,7 @@ After the installation, 2 terminal commands are available.
3. In 🐍Python. - `from TTS.api import TTS`
## On the Commandline - `tts`
![cli.gif](https://github.com/coqui-ai/TTS/raw/main/images/tts_cli.gif)
![cli.gif](https://github.com/eginhard/coqui-tts/raw/main/images/tts_cli.gif)
After the installation, 🐸TTS provides a CLI interface for synthesizing speech using pre-trained models. You can either use your own model or the release models under 🐸TTS.
@ -81,11 +81,11 @@ tts --model_name "voice_conversion/<language>/<dataset>/<model_name>"
## On the Demo Server - `tts-server`
<!-- <img src="https://raw.githubusercontent.com/coqui-ai/TTS/main/images/demo_server.gif" height="56"/> -->
![server.gif](https://github.com/coqui-ai/TTS/raw/main/images/demo_server.gif)
<!-- <img src="https://raw.githubusercontent.com/eginhard/coqui-tts/main/images/demo_server.gif" height="56"/> -->
![server.gif](https://github.com/eginhard/coqui-tts/raw/main/images/demo_server.gif)
You can boot up a demo 🐸TTS server to run an inference with your models (make
sure to install the additional dependencies with `pip install TTS[server]`).
sure to install the additional dependencies with `pip install coqui-tts[server]`).
Note that the server is not optimized for performance but gives you an easy way
to interact with the models.

View File

@ -1,6 +1,6 @@
# Installation
🐸TTS supports python >=3.9 <3.12.0 and tested on Ubuntu 18.10, 19.10, 20.10.
🐸TTS supports python >=3.9 <3.12.0 and was tested on Ubuntu 20.04 and 22.04.
## Using `pip`
@ -9,13 +9,13 @@
You can install from PyPI as follows:
```bash
pip install TTS # from PyPI
pip install coqui-tts # from PyPI
```
Or install from Github:
```bash
pip install git+https://github.com/coqui-ai/TTS # from Github
pip install git+https://github.com/eginhard/coqui-tts # from Github
```
## Installing From Source
@ -23,11 +23,13 @@ pip install git+https://github.com/coqui-ai/TTS # from Github
This is recommended for development and more control over 🐸TTS.
```bash
git clone https://github.com/coqui-ai/TTS/
cd TTS
git clone https://github.com/eginhard/coqui-tts
cd coqui-tts
make system-deps # only on Linux systems.
make install
```
## On Windows
If you are on Windows, 👑@GuyPaddock wrote installation instructions [here](https://stackoverflow.com/questions/66726331/
If you are on Windows, 👑@GuyPaddock wrote installation instructions
[here](https://stackoverflow.com/questions/66726331/) (note that these are out
of date, e.g. you need to have at least Python 3.9)

View File

@ -1,3 +1,3 @@
# Trainer API
We made the trainer a separate project on https://github.com/coqui-ai/Trainer
We made the trainer a separate project on https://github.com/eginhard/coqui-trainer

View File

@ -3,9 +3,6 @@
ⓍTTS has important model changes that make cross-language voice cloning and multi-lingual speech generation super easy.
There is no need for an excessive amount of training data that spans countless hours.
This is the same model that powers [Coqui Studio](https://coqui.ai/), and [Coqui API](https://docs.coqui.ai/docs), however we apply
a few tricks to make it faster and support streaming inference.
### Features
- Voice cloning.
- Cross-language voice cloning.
@ -32,21 +29,20 @@ Stay tuned as we continue to add support for more languages. If you have any lan
This model is licensed under [Coqui Public Model License](https://coqui.ai/cpml).
### Contact
Come and join in our 🐸Community. We're active on [Discord](https://discord.gg/fBC58unbKE) and [Twitter](https://twitter.com/coqui_ai).
You can also mail us at info@coqui.ai.
Come and join in our 🐸Community. We're active on [Discord](https://discord.gg/fBC58unbKE) and [Github](https://github.com/eginhard/coqui-tts/discussions).
### Inference
#### 🐸TTS Command line
You can check all supported languages with the following command:
You can check all supported languages with the following command:
```console
tts --model_name tts_models/multilingual/multi-dataset/xtts_v2 \
--list_language_idx
```
You can check all Coqui available speakers with the following command:
You can check all Coqui available speakers with the following command:
```console
tts --model_name tts_models/multilingual/multi-dataset/xtts_v2 \
@ -280,7 +276,7 @@ To make the `XTTS_v2` fine-tuning more accessible for users that do not have goo
The Colab Notebook is available [here](https://colab.research.google.com/drive/1GiI4_X724M8q2W-zZ-jXo7cWTV7RfaH-?usp=sharing).
To learn how to use this Colab Notebook please check the [XTTS fine-tuning video]().
To learn how to use this Colab Notebook please check the [XTTS fine-tuning video](https://www.youtube.com/watch?v=8tpDiiouGxc).
If you are not able to acess the video you need to follow the steps:
@ -294,7 +290,7 @@ If you are not able to acess the video you need to follow the steps:
##### Run demo locally
To run the demo locally you need to do the following steps:
1. Install 🐸 TTS following the instructions available [here](https://tts.readthedocs.io/en/dev/installation.html#installation).
1. Install 🐸 TTS following the instructions available [here](https://coqui-tts.readthedocs.io/en/latest/installation.html).
2. Install the Gradio demo requirements with the command `python3 -m pip install -r TTS/demos/xtts_ft_demo/requirements.txt`
3. Run the Gradio demo using the command `python3 TTS/demos/xtts_ft_demo/xtts_demo.py`
4. Follow the steps presented in the [tutorial video](https://www.youtube.com/watch?v=8tpDiiouGxc&feature=youtu.be) to be able to fine-tune and test the fine-tuned model.

View File

@ -5,14 +5,14 @@
User friendly installation. Recommended only for synthesizing voice.
```bash
$ pip install TTS
$ pip install coqui-tts
```
Developer friendly installation.
```bash
$ git clone https://github.com/coqui-ai/TTS
$ cd TTS
$ git clone https://github.com/eginhard/coqui-tts
$ cd coqui-tts
$ pip install -e .
```
@ -109,15 +109,15 @@ $ tts -h # see the help
$ tts --list_models # list the available models.
```
![cli.gif](https://github.com/coqui-ai/TTS/raw/main/images/tts_cli.gif)
![cli.gif](https://github.com/eginhard/coqui-tts/raw/main/images/tts_cli.gif)
You can call `tts-server` to start a local demo server that you can open on
your favorite web browser and 🗣️ (make sure to install the additional
dependencies with `pip install TTS[server]`).
dependencies with `pip install coqui-tts[server]`).
```bash
$ tts-server -h # see the help
$ tts-server --list_models # list the available models.
```
![server.gif](https://github.com/coqui-ai/TTS/raw/main/images/demo_server.gif)
![server.gif](https://github.com/eginhard/coqui-tts/raw/main/images/demo_server.gif)

View File

@ -41,7 +41,7 @@
"outputs": [],
"source": [
"! pip install -U pip\n",
"! pip install TTS"
"! pip install coqui-tts"
]
},
{

View File

@ -32,7 +32,7 @@
"source": [
"## Install Coqui TTS\n",
"! pip install -U pip\n",
"! pip install TTS"
"! pip install coqui-tts"
]
},
{
@ -44,7 +44,7 @@
"\n",
"### **First things first**: we need some data.\n",
"\n",
"We're training a Text-to-Speech model, so we need some _text_ and we need some _speech_. Specificially, we want _transcribed speech_. The speech must be divided into audio clips and each clip needs transcription. More details about data requirements such as recording characteristics, background noise and vocabulary coverage can be found in the [🐸TTS documentation](https://tts.readthedocs.io/en/latest/formatting_your_dataset.html).\n",
"We're training a Text-to-Speech model, so we need some _text_ and we need some _speech_. Specificially, we want _transcribed speech_. The speech must be divided into audio clips and each clip needs transcription. More details about data requirements such as recording characteristics, background noise and vocabulary coverage can be found in the [🐸TTS documentation](https://coqui-tts.readthedocs.io/en/latest/formatting_your_dataset.html).\n",
"\n",
"If you have a single audio file and you need to **split** it into clips. It is also important to use a lossless audio file format to prevent compression artifacts. We recommend using **wav** file format.\n",
"\n",

View File

@ -39,7 +39,7 @@ Docker container was created for simplify local running. You can run `docker-pre
## Training - with GPU
You need to upload Coqui-TTS(/mycomputer/TTS/) and storage/ directory(/mycomputer/storage/) to some computer with GPU. We don't need cv-corpus/ and fanetyka/ directories for training. Install gcc, then run `pip install -e .[all,dev,notebooks]` to prepare modules. GlowTTS and HifiGan models should be learned separately based on /storage/filtered_dataset only, i.e. they are not dependent from each other. <devices> below means list of GPU ids from zero("0,1,2,3" for systems with 4 GPU). See details on the https://tts.readthedocs.io/en/latest/tutorial_for_nervous_beginners.html(multi-gpu training).
You need to upload Coqui-TTS(/mycomputer/TTS/) and storage/ directory(/mycomputer/storage/) to some computer with GPU. We don't need cv-corpus/ and fanetyka/ directories for training. Install gcc, then run `pip install -e .[all,dev,notebooks]` to prepare modules. GlowTTS and HifiGan models should be learned separately based on /storage/filtered_dataset only, i.e. they are not dependent from each other. <devices> below means list of GPU ids from zero("0,1,2,3" for systems with 4 GPU). See details on the https://coqui-tts.readthedocs.io/en/latest/tutorial_for_nervous_beginners.html (multi-gpu training).
Current setup created for 24GiB GPU. You need to change batch_size if you have more or less GPU memory. Also, you can try to set lr(learning rate) to lower value in the end of training GlowTTS.

View File

@ -79,12 +79,14 @@ exts = [
)
]
setup(
name="TTS",
name="coqui-tts",
version=version,
url="https://github.com/coqui-ai/TTS",
url="https://github.com/eginhard/coqui-tts",
author="Eren Gölge",
author_email="egolge@coqui.ai",
description="Deep learning for Text to Speech by Coqui.",
maintainer="Enno Hermann",
maintainer_email="enno.hermann@gmail.com",
description="Deep learning for Text to Speech.",
long_description=README,
long_description_content_type="text/markdown",
license="MPL-2.0",
@ -101,10 +103,10 @@ setup(
]
},
project_urls={
"Documentation": "https://github.com/coqui-ai/TTS/wiki",
"Tracker": "https://github.com/coqui-ai/TTS/issues",
"Repository": "https://github.com/coqui-ai/TTS",
"Discussions": "https://github.com/coqui-ai/TTS/discussions",
"Documentation": "https://coqui-tts.readthedocs.io",
"Tracker": "https://github.com/eginhard/coqui-tts/issues",
"Repository": "https://github.com/eginhard/coqui-tts",
"Discussions": "https://github.com/eginhard/coqui-tts/discussions",
},
cmdclass={
"build_py": build_py,