Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Training data of the segmenter #8

Open
unilight opened this issue Mar 12, 2024 · 2 comments
Open

Training data of the segmenter #8

unilight opened this issue Mar 12, 2024 · 2 comments

Comments

@unilight
Copy link

Hi @bshall, thanks for the great work!

I am wondering what is the training data of the pre-trained segmenter, as it is not described in the paper or this repo.

Thanks!

@bshall
Copy link
Owner

bshall commented Mar 18, 2024

Hi @unilight, thanks for the feedback! Sorry about the delay getting back to you. The pre-trained segmenter was trained on p225. But I've tested it out on the other speakers and the learned mapping from clusters to sonorants, obstruents, and silences seems to be consistent. Let me know if you've found something different.

@unilight
Copy link
Author

@bshall Thank you for the reply! I do find that segmenter works well on other speakers. I am just wondering how you found the mapping from cluster index to sonorants, obstruents, and silences. Do I need to manually find the correspondence by looking at the phoneme labels and the cluster indices? Asking this because I want to apply the same method to some atypical speech (say, accented speech). I found that the pre-trained segmenter works well on normal speech but has some problems with such atypical speech, so I wonder whether it is possible to train my own segmenter.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants