Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Distilled model benchmarks? #2

Open
imneonizer opened this issue Sep 11, 2023 · 0 comments
Open

Distilled model benchmarks? #2

imneonizer opened this issue Sep 11, 2023 · 0 comments

Comments

@imneonizer
Copy link

Thanks for the great work!
It's a very clever way to compute embeddings beforehand and use them directly as target values during backpropagation step.

Questions

  • Have you done any testing to find out, how well the distilled model performs as compared to the original teacher model?
  • If we use Vision Transformer (ViT) models as base, should there be any improvement to embedding quality?
  • Instead of using the distilled model for classification task by computing the probs, How well it performs in case we want to utilize the raw embeddings for ranking the images based on cosine distance.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant