-
I had the pretty much the same question in #7 Do you think would it make sense to use w2v-bert to extract the semantic tokens for fine-tuning of the first transformer (text model, semantic modeling)? I am suggesting w2v-bert as it was used AudioLM to generate semantic tokens from audio. |
Beta Was this translation helpful? Give feedback.
Answered by
gkucsko
May 11, 2023
Replies: 1 comment
-
afaik w2v-bert is not public, but ya something like that would be a good start i think |
Beta Was this translation helpful? Give feedback.
0 replies
Answer selected by
yurtmete
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
afaik w2v-bert is not public, but ya something like that would be a good start i think