Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training process and loss understanding #81

Open
SanketDhuri opened this issue Oct 23, 2023 · 2 comments
Open

Training process and loss understanding #81

SanketDhuri opened this issue Oct 23, 2023 · 2 comments

Comments

@SanketDhuri
Copy link

  1. We have dataset of 5000-6000 audios of duration >= 2sec but still on training on this dataset for 15 epochs it doesn’t reaches satisfactory loss what is the issue here ?

eg last observed loss are :
Semantic (train loss = 0.00309 ) (val loss = 1.28167 )
Coarse loss (train loss = 0.057) (val loss = 3.2796 )
Fine loss (train loss = 0.1 ) (val loss = 1.18 )

  1. Why does sometimes it itself adds or skips the word, sometime it also generates voice that has shivering tone?
  2. Why it gives problems for punctuation like ! or .
  3. We are training with mixed dataset different speaker but they have same way speaking can it create problem in training?
  4. On finetuning pretrained prompts like [laughs],etc are lost what might be the reason?
    We are preparing dataset’s with help of whisper model so should we also add those manually add those prompts to dataset?
  5. How can add new emotion prompts such sad or excited or unhappy etc.
@dagshub
Copy link

dagshub bot commented Oct 23, 2023

@boringtaskai
Copy link

Hi how do you use the weight after the training? cause I had issue like this:

raise ValueError(f"missing keys: {missing_keys}")

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants