You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In order to facilitate the evaluation of different interpretability techniques, I propose to identify a set of commonly used datasets from the literature, create 🤗 Datasets loading scripts to have them in a shared format, and host them on the Inseq organization in the Hugging Face hub.
This would provide a shared interface for:
Faithfulness metrics applied at a dataset level.
Future support of instance attribution methods.
The following table summarizes some of the datasets used in the literature:
🚀 Feature Request
In order to facilitate the evaluation of different interpretability techniques, I propose to identify a set of commonly used datasets from the literature, create 🤗 Datasets loading scripts to have them in a shared format, and host them on the Inseq organization in the Hugging Face hub.
This would provide a shared interface for:
The following table summarizes some of the datasets used in the literature:
The ExNLP Datasets website summarizes various sources available for NLP explainability, verify what is relevant to generation.
The text was updated successfully, but these errors were encountered: