Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

For better transcription in more languages, Implement Massively Multilingual Speech - Meta's Open Source model with less than half of Whispers error rate #4

Open
menelic opened this issue May 26, 2023 · 2 comments

Comments

@menelic
Copy link
Contributor

menelic commented May 26, 2023

Because of the error rate viz and above al speaker detection your whisper ui is better for research use than all the others I have tried. Please consider implementing Meta's MMS with speech recognition and generation support for over 1000 languages at a drastically reduced error rate compared to Whisper:

image

https://github.com/facebookresearch/fairseq/tree/main/examples/mms

https://ai.facebook.com/blog/multilingual-model-speech-recognition/

@jojojaeger
Copy link
Owner

sounds great, i will try it out!

@menelic
Copy link
Contributor Author

menelic commented Jun 14, 2023

have you had the chance to check it out?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants