Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Finetuned large-v3 inference problem. #908

Open
sinisha opened this issue Oct 28, 2024 · 1 comment
Open

Finetuned large-v3 inference problem. #908

sinisha opened this issue Oct 28, 2024 · 1 comment

Comments

@sinisha
Copy link

sinisha commented Oct 28, 2024

I have finetuned large-v3 model on 1k hours of domain-specific data. When I run standard inference the results are ok.
Finetuned model is converted using ctranslate2 but the results obtained with whisperx are almost all hallucinations (maybe first couple of phonemes at the beginning are correct). I used same ctranslate2 command to convert original large-v3 model and whisperx inference is also correct.
Model is finetuned using Transformers 4.45.2. I have tried a couple of different Transformers version in inference and the results are similar.
Has anyone encountered similar problem?

@DaddyCodesAlot
Copy link

How long were your training samples? finetuning whisper is best when the samples are relatively short

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants