Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Slow Inference using python #147

Open
yeswanth-spora opened this issue Feb 22, 2025 · 6 comments
Open

Slow Inference using python #147

yeswanth-spora opened this issue Feb 22, 2025 · 6 comments

Comments

@yeswanth-spora
Copy link

when running the model using python iam trying to automate in my project but it is too slow the model loading and inference currently using a100 gpu

@darkacorn
Copy link
Contributor

you dont need a a100 .. as this doesnt make really use of batching .. any consumer gpu will do just fine - and you will need a h100 to get to the 2x realitme generation speed

@Ph0rk0z
Copy link

Ph0rk0z commented Feb 22, 2025

He never list his t/s. Should get at least 115+ on A100

@yeswanth-spora
Copy link
Author

what is the best tts voice cloning zeroshot in a project i can use a100 40 gb vram

@Ph0rk0z
Copy link

Ph0rk0z commented Feb 22, 2025

It's this, GPTsovits and fish audio. I also like vokan. Weaker clone can be fixed with RVC. All sound like reading more than zonos. This one is more like bark.

@yeswanth-spora
Copy link
Author

thank you it helps me a lot i can use it in my project

@johnwick123f
Copy link

Yeah zonos is probably the best tts right now since it generates very high quality clear audio with great voice cloning in 30langs with a very permissive apache license. Fish-speech, gpt-sovits, xttsv2, llasa, cosyvoice should be considerably faster but a bit worse and many don't allow for commercial usage.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants