NeMo ASR Demo for Transcription #5275
titu1994
started this conversation in
Show and tell
Replies: 2 comments 3 replies
-
Are there any specific guidelines that should be adhered to, if one wants to publish a model/checkpoint that will then be automatically picked up? |
Beta Was this translation helpful? Give feedback.
1 reply
-
BTW. Note that the integration of the demo inside the ASR docs, has one small issue; i.e. it never requests for permission to access the microphone, hence trying the "record from microphone" path will always error out. |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
As the NeMo ASR Collection grows, and we support more languages, it can get a bit complicated when trying to find checkpoints for certain languages. This is especially true now that the community has started to contribute checkpoints in various languages.
So we present a new Hugging Face Space that will allow inference on all NeMo checkpoints that are uploaded to HF !
Link - https://huggingface.co/spaces/smajumdar/nemo_multilingual_language_id
In it, you can either upload a file, or use your microphone to record a piece of audio, then select a language and a model of your choice in that language to perform transcription.
We will also be adding this demo inside of the ASR docs page for ease of use.
We encourage users to submit their own checkpoints on Hugging Face so that others in the community may transcribe speech in as many languages as possible !
Beta Was this translation helpful? Give feedback.
All reactions