diff --git a/docs/_sources/pages/Front_page.md.txt b/docs/_sources/pages/Front_page.md.txt index ddd80fc..7938168 100644 --- a/docs/_sources/pages/Front_page.md.txt +++ b/docs/_sources/pages/Front_page.md.txt @@ -1,6 +1,6 @@ # FAIR Universe - HiggsML Uncertainty Challenge -## Introduction +## Introduction This NeurIPS 2024 Machine Learning competition is one of the first to strongly emphasise mastering uncertainties in the input training dataset and outputting credible confidence intervals. This challenge explores uncertainty-aware AI techniques for High Energy Physics (HEP). @@ -13,9 +13,9 @@ There are several information sources regarding the FAIR Universe - HiggsML Unce * [Codabench](https://www.codabench.org/competitions/2977/): This serves as the platform to submit entries to the competition. -* [Tutorial Slides](https://fair-universe.lbl.gov/tutorials/Higgs_Uncertainty_Challenge-Codabench_Tutorial.pdf) : These slides will help you register and submit a sample dummy submission. +* [Tutorial Slides](https://fair-universe.lbl.gov/tutorials/Higgs_Uncertainty_Challenge-Codabench_Tutorial.pdf) : These slides will help you register and submit a sample dummy submission. -* Download the [training data](https://www.codabench.org/datasets/download/b9e59d0a-4db3-4da4-b1f8-3f609d1835b2/) if you you want to experiment with the data on your local machines. +* Download the [Training Data (6.5 GB)](https://www.codabench.org/datasets/download/b9e59d0a-4db3-4da4-b1f8-3f609d1835b2/) if you you want to experiment with the data on your local machines. * [Documentation](https://fair-universe.lbl.gov/docs/): This contains detailed information about the [science behind the challenge](https://fair-universe.lbl.gov/docs/pages/overview.html#problem-setting), the [specifics of the data](https://fair-universe.lbl.gov/docs/pages/data.html), and [documents the code](https://fair-universe.lbl.gov/docs/rst_source/modules.html) used to facilitate the evaluation of the competition. It also describes the [evaluation metric](https://fair-universe.lbl.gov/docs/pages/evaluation.html) @@ -42,7 +42,7 @@ This competition allows code submissions. Participants are strongly encouraged t - Sascha Diefenbacher - Steven Farrell - Wahid Bhimji -- Jordan Dudley +- Jordan Dudley ### University of Washington - Elham E Khoda diff --git a/docs/_sources/pages/data.md.txt b/docs/_sources/pages/data.md.txt index c6fb55e..d213b5f 100644 --- a/docs/_sources/pages/data.md.txt +++ b/docs/_sources/pages/data.md.txt @@ -103,7 +103,7 @@ These variables are derived from the primary varibales with the help of `derived ## How to get Public Data? -Download the [Neurips_Public_data_26_08_2024](https://www.codabench.org/datasets/download/b9e59d0a-4db3-4da4-b1f8-3f609d1835b2/) +Download the [Neurips_Public_data_26_08_2024 (6.5 GB)](https://www.codabench.org/datasets/download/b9e59d0a-4db3-4da4-b1f8-3f609d1835b2/) or use the following command to download using terminal ``` diff --git a/docs/_sources/pages/evaluation.md.txt b/docs/_sources/pages/evaluation.md.txt index 3aa046d..e730427 100644 --- a/docs/_sources/pages/evaluation.md.txt +++ b/docs/_sources/pages/evaluation.md.txt @@ -16,6 +16,16 @@ Not every uncertainty quantification method is able to return a full likelihood **Methods that a central value and Gaussian uncertainty**: Define $\hat \mu_{16} = \hat \mu - \Delta \hat \mu$ and $\hat \mu_{84} = \hat \mu + \Delta \hat \mu$. If the underlying assumption of a symmetrical uncertainty made by the Gaussian uncertainty is given, this should contain 68.27% of the probability, since the 1 standard deviation region of a Gaussian distribution contains 68.27% of the probability mass. +## Submission requirements + +Participants’ submissions must consist of a zip file containing a `model.py` file (which must not be inside a directory), and it can contain any other necessary files (eg, a pre-trained model). The `model.py` file must define a `Model` class which must satisfy the following criteria. +1. The `Model` class must accept two arguments, `get_train_set` and `systematics`, when it is initialized. The `get_train_set` argument will receive a callable which, when called, will return the public dataset. The `systematics` argument will receive a callable which can be used to apply the systematic effects (adjusting weights and primary features, computing derived features, and applying post-selection cuts) to a dataset. +1. The `Model` class must have a `fit` method, which will be called once when the submission is being evaluated. This method can be used prepare the model for inference. We encourage participants to submit models which have already been trained, as there is limited compute time for each submission to be evaluated. +1. The `Model` class must have a `predict` method which must accept a test dataset and return the results as a dictionary containing four items: `”mu_hat”`: the predicted value of mu, `”delta_mu_hat”`: the uncertainty in the predicted value of mu, `”p16”`: the lower bound of the 16th percentile of mu, and `”p84”`: the upper bound of the 84th percentile of mu. + +## Hardware description + +Throughout the competition, participants’ submissions will be run on either the Perlmutter supercomputer at NERSC or an alternative workstation at LBNL, but they will only be run on Perlmutter for the final evaluation. When running on Perlmutter, submissions will have one node assigned, which consists of 1 AMD EPYC 7763 CPU, 256GB of RAM, and 4 NVIDIA A100 GPUs with 40GB of memory each (https://docs.nersc.gov/systems/perlmutter/architecture/#gpu-nodes). The alternative workstation consists of 1 Intel(R) Xeon(R) Gold 6148 CPU, 376GB of RAM, and 3 Tesla V100-SXM2-16GB GPUs available. On either system, participants’ submissions will be allotted 2 hours to complete evaluation on all of the pseudoexperiments (10 sets of 100 pseudoexperiments each for the initial phase, and 10 sets of 1000 pseudoexperiments for the final phase). These pseudoexperiments will be parallelized in the following way. Each participants’ `Model.fit()` will be run once, and then each of the pseudoexperiments will be run by one of many parallel workers, with each worker calling `Model.predict()` once. There will be 30 parallel workers when running on Perlmutter. This will be reduced to 10 parallel workers when running on the alternative workstation. ## Scoring diff --git a/docs/_sources/pages/terms.md.txt b/docs/_sources/pages/terms.md.txt index bf38ebe..b6cea5d 100644 --- a/docs/_sources/pages/terms.md.txt +++ b/docs/_sources/pages/terms.md.txt @@ -14,10 +14,7 @@ This challenge is for educational purposes only and no prizes are granted. It is To ensure proper representation and verification of your affiliation with an organization in our competition, please adhere to the following guidelines -#### Profile Affiliation - -Ensure that your Codabench profile includes your current organizational affiliation. This helps in verifying your credentials and associating your contributions with the correct institution. - -#### Registration Requirements - -Use your organization-issued email address to register for the competition. This will help us confirm that you are officially associated with the organization and prevent unauthorized entries. \ No newline at end of file +- Ensure that your Codabench profile includes your current organizational affiliation. This helps in verifying your credentials and associating your contributions with the correct institution. +- If participants are freelancers they should provide brief details (via email to fair-universe@lbl.gov) of their freelance operation, any supporting URL, and their country of residence. +- Use your organization-issued email address to register for the competition. This will help us confirm that you are officially associated with the organization and prevent unauthorized entries. +- Registration is not required to access the dataset that is provided on the “Data” tab. We encourage the use of the dataset for research. In the event that a participant who is not eligible for submission via codabench according to the conditions above, can demonstrate a successful method on the challenge task we encourage them to contact the organizers for attribution and potential inclusion in the final competition stage if possible. \ No newline at end of file diff --git a/docs/pages/Front_page.html b/docs/pages/Front_page.html index c7fe0ec..c33b127 100644 --- a/docs/pages/Front_page.html +++ b/docs/pages/Front_page.html @@ -93,7 +93,7 @@
Codabench: This serves as the platform to submit entries to the competition.
Tutorial Slides : These slides will help you register and submit a sample dummy submission.
Download the training data if you you want to experiment with the data on your local machines.
Download the Training Data (6.5 GB) if you you want to experiment with the data on your local machines.
Documentation: This contains detailed information about the science behind the challenge, the specifics of the data, and documents the code used to facilitate the evaluation of the competition. It also describes the evaluation metric
Github Repo: This hosts the code for testing submissions, as well as the starting kit notebook. The starting kit is also available on Google Colab
White Paper: This serves as a full breakdown of the competition in detail
Download the Neurips_Public_data_26_08_2024
+Download the Neurips_Public_data_26_08_2024 (6.5 GB)
or use the following command to download using terminal
wget -O public_data.zip https://www.codabench.org/datasets/download/b9e59d0a-4db3-4da4-b1f8-3f609d1835b2/
Participants’ submissions must consist of a zip file containing a model.py
file (which must not be inside a directory), and it can contain any other necessary files (eg, a pre-trained model). The model.py
file must define a Model
class which must satisfy the following criteria.
The Model
class must accept two arguments, get_train_set
and systematics
, when it is initialized. The get_train_set
argument will receive a callable which, when called, will return the public dataset. The systematics
argument will receive a callable which can be used to apply the systematic effects (adjusting weights and primary features, computing derived features, and applying post-selection cuts) to a dataset.
The Model
class must have a fit
method, which will be called once when the submission is being evaluated. This method can be used prepare the model for inference. We encourage participants to submit models which have already been trained, as there is limited compute time for each submission to be evaluated.
The Model
class must have a predict
method which must accept a test dataset and return the results as a dictionary containing four items: ”mu_hat”
: the predicted value of mu, ”delta_mu_hat”
: the uncertainty in the predicted value of mu, ”p16”
: the lower bound of the 16th percentile of mu, and ”p84”
: the upper bound of the 84th percentile of mu.
Throughout the competition, participants’ submissions will be run on either the Perlmutter supercomputer at NERSC or an alternative workstation at LBNL, but they will only be run on Perlmutter for the final evaluation. When running on Perlmutter, submissions will have one node assigned, which consists of 1 AMD EPYC 7763 CPU, 256GB of RAM, and 4 NVIDIA A100 GPUs with 40GB of memory each (https://docs.nersc.gov/systems/perlmutter/architecture/#gpu-nodes). The alternative workstation consists of 1 Intel(R) Xeon(R) Gold 6148 CPU, 376GB of RAM, and 3 Tesla V100-SXM2-16GB GPUs available. On either system, participants’ submissions will be allotted 2 hours to complete evaluation on all of the pseudoexperiments (10 sets of 100 pseudoexperiments each for the initial phase, and 10 sets of 1000 pseudoexperiments for the final phase). These pseudoexperiments will be parallelized in the following way. Each participants’ Model.fit()
will be run once, and then each of the pseudoexperiments will be run by one of many parallel workers, with each worker calling Model.predict()
once. There will be 30 parallel workers when running on Perlmutter. This will be reduced to 10 parallel workers when running on the alternative workstation.
The score consist of two parts, the interval width and the coverage:
@@ -156,4 +170,4 @@