Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[FT] Fail faster when passing unsupported metrics to InferenceEndpointModel #436

Open
albertvillanova opened this issue Dec 11, 2024 · 2 comments
Labels
feature request New feature/request

Comments

@albertvillanova
Copy link
Member

Issue encountered

When passing an unsupported metric (e.g. single-token metric) to InferenceEndpointModel, an error is raised from Pipeline.evaluate:

ValueError: Endpoint models can't use single token metrics. Change the metric to the standard version

This happens once the endpoint has been created and the model has been created and loaded.

I wonder if the error could be raised earlier for optimization reasons: the process fails faster and resources are not wasted.

Solution/Feature

Raise the error before the endpoint has been created.

Possible alternatives

Leave it as it is.

@albertvillanova albertvillanova added the feature request New feature/request label Dec 11, 2024
@clefourrier
Copy link
Member

Hm, good idea for the checks, especially for costly metrics. We would also want the LLM as judge to stop if the open ai key is not provided for example.

@clefourrier
Copy link
Member

Also linked to #355

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature request New feature/request
Projects
None yet
Development

No branches or pull requests

2 participants