Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Relevancy Experimentation Framework #421

Open
2 tasks
obulat opened this issue Feb 18, 2023 · 0 comments
Open
2 tasks

Relevancy Experimentation Framework #421

obulat opened this issue Feb 18, 2023 · 0 comments
Labels
🌟 goal: addition Addition of new feature 🧭 project: thread An issue used to track a project and its progress 🧱 stack: analytics Related to the analytics setup 🧱 stack: api Related to the Django API 🧱 stack: catalog Related to the catalog and Airflow DAGs 🧱 stack: frontend Related to the Nuxt frontend 🧱 stack: infra Related to the Terraform config and other infrastructure

Comments

@obulat
Copy link
Contributor

obulat commented Feb 18, 2023

Summary

Develop an experimental framework for assessing how relevancy is affected (either positively or negatively) by changes made to our search algorithms and the data itself.

Description

The search relevancy sandbox project outlines the infrastructure pieces necessary for making rapid changes to our search algorithms and components. The next step for us in order to leverage this toolbox is to develop an experimentation framework for assessing how relevancy changes when certain changes are made. This could be done in two parts:

  • A hybrid automation + manual process which takes snapshots of specific result sets in staging, then prompts a maintainer to compare the relevancy between the two snapshots. This is intended to be a blunt approach; do the results still feel relevant is all that's necessary, if the results are wildly irrelevant where they previously were not, then we do not roll the projected change out to production.

  • An automated assessment of relevancy based on A/B experimentation and paired with the results from our analytics. This would use metrics like SELECT_SEARCH_RESULT, LOAD_MORE_RESULTS, and any new metrics we require to assess if users are clicking through results at a higher/lower rate given any changes.

The former would act as more of a "gut check" with little nuance, and the second would serve as a longer-running assessment of result relevancy based on actual user behavior.

Best guess at list of implementation plans:

  • Develop algorithm for assessing relevancy given analytics
  • Staging snapshot automation & manual assessment protocols
  • Infrastructure changes necessary for performing A/B experimentation
  • Framework for combining relevancy assessment metric and infrastructure changes to accurately run an experiment

Documents

  • Project Proposal
  • Implementation Plan

Issues

Prior Art

@obulat obulat added the 🧭 project: thread An issue used to track a project and its progress label Feb 18, 2023
dhruvkb pushed a commit that referenced this issue Apr 14, 2023
* Enable xcom pickling

* Update duration reporting
@AetherUnbound AetherUnbound changed the title Relevancy Metrics and Reporting Relevancy Experimentation Framework Dec 19, 2023
@AetherUnbound AetherUnbound added 🌟 goal: addition Addition of new feature 🧱 stack: api Related to the Django API 🧱 stack: frontend Related to the Nuxt frontend labels Dec 19, 2023
@zackkrida zackkrida added 🧱 stack: catalog Related to the catalog and Airflow DAGs 🧱 stack: analytics Related to the analytics setup 🧱 stack: infra Related to the Terraform config and other infrastructure labels Jul 31, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
🌟 goal: addition Addition of new feature 🧭 project: thread An issue used to track a project and its progress 🧱 stack: analytics Related to the analytics setup 🧱 stack: api Related to the Django API 🧱 stack: catalog Related to the catalog and Airflow DAGs 🧱 stack: frontend Related to the Nuxt frontend 🧱 stack: infra Related to the Terraform config and other infrastructure
Projects
Status: ⌛ Todo
Development

No branches or pull requests

3 participants