Replies: 1 comment 1 reply
-
Great question! We actually do something like this for the PovertyMap dataset, so perhaps that would be a helpful reference? https://github.com/p-lambda/wilds/blob/main/wilds/datasets/poverty_dataset.py |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hello and thank you for this amazing package.
Instead of using replicates, I would be interested in adding a cross validation training and evaluation scheme based on the domain metadata.
Say a dataset has domain: A,B,C. I would like to:
Finally average the in distribution and the out of distribution metric to have the final performance.
Here the 70-30 split is arbitrary and should be modifiable.
I am just starting exploring the package having only replicated the ERM result on the camelyon17 dataset.
It seems that the grouper object might be a good start to implement the following procedure. But, I am still lacking a high level overview of the code. So how would you do this ?
Beta Was this translation helpful? Give feedback.
All reactions