Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

请问*.style.pt是什么 #5

Open
Darcy0218 opened this issue Aug 8, 2024 · 4 comments
Open

请问*.style.pt是什么 #5

Darcy0218 opened this issue Aug 8, 2024 · 4 comments

Comments

@Darcy0218
Copy link

FileNotFoundError: [Errno 2] No such file or directory: '/data4/wuyikai/data/TextrolSpeech/LibriTTS/LibriTTS/train-clean-360/6300/39660/6300_39660_000014_000000.style.pt'

@Darcy0218
Copy link
Author

I can't make sense of it.
I see from your codebase that a style embedding is required, which is loaded from*.style.pt format.
I guess, have you utilized a model to generate style embedding(*.style.pt) before training? If so, what is that model?
image

@jishengpeng
Copy link
Owner

I can't make sense of it. I see from your codebase that a style embedding is required, which is loaded from*.style.pt format. I guess, have you utilized a model to generate style embedding(*.style.pt) before training? If so, what is that model? image

ok, I took a closer look (this baseline is the second act to help run), this should be style emdedding(bert extracted the style representation). I think you have two ways to deal with it, one is to pre-extract relevant representations according to its original logic and save them, the other is to directly load the bert pre-trained model in the training process. References can be shown below.
1ed786fe89bd3c7ab34ffdea24db593

In addition, having to work overtime on weekends is frustrating.

@Darcy0218
Copy link
Author

I can't make sense of it. I see from your codebase that a style embedding is required, which is loaded from*.style.pt format. I guess, have you utilized a model to generate style embedding(*.style.pt) before training? If so, what is that model? image

ok, I took a closer look (this baseline is the second act to help run), this should be style emdedding(bert extracted the style representation). I think you have two ways to deal with it, one is to pre-extract relevant representations according to its original logic and save them, the other is to directly load the bert pre-trained model in the training process. References can be shown below. 1ed786fe89bd3c7ab34ffdea24db593

In addition, having to work overtime on weekends is frustrating.

哈哈哈,辛苦了兄弟,谢谢你的回复。

@xiezheng-cs
Copy link

xiezheng-cs commented Sep 18, 2024

I can't make sense of it. I see from your codebase that a style embedding is required, which is loaded from*.style.pt format. I guess, have you utilized a model to generate style embedding(*.style.pt) before training? If so, what is that model? image

ok, I took a closer look (this baseline is the second act to help run), this should be style emdedding(bert extracted the style representation). I think you have two ways to deal with it, one is to pre-extract relevant representations according to its original logic and save them, the other is to directly load the bert pre-trained model in the training process. References can be shown below. 1ed786fe89bd3c7ab34ffdea24db593

In addition, having to work overtime on weekends is frustrating.

您好,请问是直接用inference.py中的get_style_embed函数得到style_embed吗?style_prompt文本是否有什么前处理?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants