-
Notifications
You must be signed in to change notification settings - Fork 509
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
High Latency in TTS Synthesis on Android with Screen Readers #1337
Comments
|
Which model are you using and which kind of android phone, i.e., the CPU of your phone, are you using? |
From experience, latency depends on the kind of model, the device's processing power (CPU), and the length of the text being processed. Piper's "medium" quality models (around 60mb) have lower latency , compared to other models. I was able to speed up inference by splitting the input text into batches, using punctuation as natural sentence boundaries, This allows for quicker synthesis of smaller chunks of text. On more powerful devices, this step may not be necessary, as they can handle larger texts efficiently. You can adjust the batching based on the device's capabilities. |
here are the details: |
This phone has If it uses Cortex A78 during synthesis, then it should be very fast. If it uses Cortex A55, then it would be slow. |
Hello,
I am using your TTS as an Android Text-to-Speech (TTS) engine for offline use, but I have encountered an issue with the audio synthesis. It is taking approximately 500 ms to speak the text on the screen when using screen readers on Android in the
onSynthesize
method.Is there any option or solution to reduce this latency? I am trying to create a TTS system for Android that blind/low-vision users can use effectively with their screen readers, so minimizing latency is critical for accessibility.
The text was updated successfully, but these errors were encountered: