Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Local judge LLM? #664

Open
lyzhongcrd opened this issue Dec 12, 2024 · 3 comments
Open

Local judge LLM? #664

lyzhongcrd opened this issue Dec 12, 2024 · 3 comments

Comments

@lyzhongcrd
Copy link

Is it possible to use locally deployed LLM like LLaVa-Critic as judge LLM instead of calling GPT4 API?

@kennymckormick
Copy link
Member

Hi, @lyzhongcrd ,
Yeah. However, we recommend you use the same LLM as the judger for all LMMs to make it comparable.
For MCQ or Y/N benchmarks, when LLMs are only used as choice extractor for more accurate evaluation, using different LLMs will not lead to significantly different results.

@lyzhongcrd
Copy link
Author

@kennymckormick Could you tell me how to use locally deployed LLMs as judge LLM in the VLM eval kit? Thanks.

@Leke-G
Copy link

Leke-G commented Dec 24, 2024

@kennymckormick您能告诉我如何使用本地部署的LLM作为VLM评估套件中的判断LLM吗?谢谢。

请问一下贴主,如果我要评估的模型是VLM,本地部署的模型应该是LLM还是VLM呢

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants