-
Notifications
You must be signed in to change notification settings - Fork 7
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Develop evaluation methods for matching models #23
Comments
Many of these metrics with have cluster score thresholds (see #26), so the metrics table should be similar in shape to the triage results evaluations table. |
Metrics:
|
@nanounanue here are ideas for metrics |
From Joe: |
We will want to compare, select, and evaluate matching models. This requires generating and storing metrics (see dssg/pgdedupe#20 for some possibilities) and, perhaps comparing Type I and Type II error rates on labeled pairs not used in the training data (see #20).
This will likely entail storing metrics in a metrics table and a notebook/methods/workflow for conducting comparisons and evaluations.
The text was updated successfully, but these errors were encountered: