-
Notifications
You must be signed in to change notification settings - Fork 11
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
add generate metrics #29
Comments
This issue is to define new feature about the answer correctness. |
There are some open source libraries of metrics that we may be able to use in our projects. For example, Rouge and MAUVE. |
I add Rouge metrics using rouge-score, referencing huggingface/evaluate and stanford-crfm/helm. |
There are many high-quality implemented metrics we can use in datasets. Besides, we can learn from them for those uncovered metrics. |
disambig-F1: Active retrieval augmented generation, ASQA: Factoid Questions Meet Long-Form Answers |
I am going to implement the following 2 metrics: |
I am going to implement the following 2 metrics: |
the Q-BLEU metric is to meature the answerability of questions generated by Automatic Question Generation system. It use if the question include relevant content words, named entitiesand question types or function words to meature the answerability. This metric is not useful for answer generation. |
@QianHaosheng @bugtig6351 @yuanpcr you can list all potential metrics for the
generate
task in this issue. For more details about thegenerate
task, you can refer to issue #12 .The text was updated successfully, but these errors were encountered: