Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is Evaluation-of-OPE possible with real-world data? #30

Open
pmoran3 opened this issue Apr 3, 2024 · 0 comments
Open

Is Evaluation-of-OPE possible with real-world data? #30

pmoran3 opened this issue Apr 3, 2024 · 0 comments

Comments

@pmoran3
Copy link

pmoran3 commented Apr 3, 2024

Since the evaluation-of-OPE requires knowledge of the on-policy policy values, is OPS only relevant for synthetic data where the underlying behavior policy value is known? Or is it possible to estimate the on-policy policy value from real-world data, as well?

When I run this code block from the basic_synthetic_continious_advanced.ipynb notebook on my real-world dataset

ops = OffPolicySelection(
    ope=ope,
    cumulative_distribution_ope=cd_ope,
)
ops.obtain_true_selection_result(
    input_dict=input_dict,
    return_variance=True,
    return_lower_quartile=True,
    return_conditional_value_at_risk=True,
    return_by_dataframe=True,
)

I get the following error: ValueError: one of the candidate policies, cql, does not contain on-policy policy value in input_dict.

Edit: After posting this issue, it occured to me that "to estimate the on-policy policy value from real-world data" would just be equivalent to doing OPE, so evaluation-of-OPE would not be possible in that case. Please, correct me if I am misunderstanding anything.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant