Multithreading, server process size #735
stefan-reich
started this conversation in
General
Replies: 1 comment 2 replies
-
Provided inference endpoints are just for testing/trying out models and not optimized for deployment, production, or anything serious. But of course, we are open for ✨PRs✨ and discussions for such improvements |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi, I have asked this before here, but was redirected to coqui-ai (and rightly so). I'm now using coqui-ai TTS directly from GitHub. So here are the questions.
Q1: Is the inference generally multithreaded?
Q2: Is the server.py process supposed to be this big? I am seeing 1.4 GB resident process size in Linux with model tts_models/en/ljspeech/vits.
If inference is not multithread and the process has to be this big, this is quite the resource hungry application, bordering on impractical...
I understand the system is in development, so both of these issues may possibly be improved in the future? Or are just some voices this big?
Many greetings,
Stefan
Beta Was this translation helpful? Give feedback.
All reactions