Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature request] Support for Quantized ONNX Model Conversion for Stream Inference #4043

Open
TranDacKhoa opened this issue Nov 2, 2024 · 0 comments
Labels
feature request feature requests for making TTS better.

Comments

@TranDacKhoa
Copy link

🚀 Feature Description

Is there support in Coqui TTS for converting models to a quantized ONNX format for stream inference? This feature would enhance model performance and reduce inference time for real-time applications.

Solution

Implement a workflow or tool within Coqui TTS for easy conversion of TTS models to quantized ONNX format.

Alternative Solutions

Currently, external tools like ONNX Runtime or TensorRT can be used for post-conversion quantization, but having this feature natively would streamline the process.

Additional context

Any existing documentation or insights on this topic would be appreciated. Thank you!

@TranDacKhoa TranDacKhoa added the feature request feature requests for making TTS better. label Nov 2, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature request feature requests for making TTS better.
Projects
None yet
Development

No branches or pull requests

1 participant