Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Metadata correction for 2024.findings-emnlp.198 #4371

Closed
shramay-palta opened this issue Jan 8, 2025 · 6 comments · Fixed by #4454
Closed

Metadata correction for 2024.findings-emnlp.198 #4371

shramay-palta opened this issue Jan 8, 2025 · 6 comments · Fixed by #4454
Assignees
Labels
approved Used to note team approval of metadata requests correction for corrections submitted to the anthology metadata Correction to metadata

Comments

@shramay-palta
Copy link

shramay-palta commented Jan 8, 2025

JSON data block

{
  "anthology_id": "2024.findings-emnlp.198",
  "abstract": "Questions involving commonsense reasoning about everyday situations often admit many possible or plausible answers. In contrast, multiple-choice question (MCQ) benchmarks for commonsense reasoning require a hard selection of a single correct answer, which, in principle, should represent the most plausible answer choice. On 250 MCQ items sampled from two commonsense reasoning benchmarks, we collect 5,000 independent plausibility judgments on answer choices. We find that for over 20% of the sampled MCQS, the answer choice rated most plausible does not match the benchmark gold answers; upon manual inspection, we confirm that this subset exhibits higher rates of problems like ambiguity or semantic mismatch between question and answer choices. Experiments with LLMs reveal low accuracy\nand high variation in performance on the subset, suggesting our plausibility criterion may be helpful in identifying more reliable benchmark items for commonsense evaluation.",
  "authors": [
    {
      "first": "Shramay",
      "last": "Palta",
      "id": "shramay-palta",
      "affiliation": "University of Maryland"
    },
    {
      "first": "Nishant",
      "last": "Balepur",
      "id": "nishant-balepur",
      "affiliation": "University of Maryland"
    },
    {
      "first": "Peter",
      "last": "Rankel",
      "id": "peter-a-rankel",
      "affiliation": "University of Maryland"
    },
    {
      "first": "Sarah",
      "last": "Wiegreffe",
      "id": "sarah-wiegreffe",
      "affiliation": "Allen Institute for AI"
    },
    {
      "first": "Marine",
      "last": "Carpuat",
      "id": "marine-carpuat",
      "affiliation": "University of Maryland"
    },
    {
      "first": "Rachel",
      "last": "Rudinger",
      "id": "rachel-rudinger",
      "affiliation": "University of Maryland"
    }
  ]
}
@shramay-palta shramay-palta added correction for corrections submitted to the anthology metadata Correction to metadata labels Jan 8, 2025
Copy link

github-actions bot commented Jan 8, 2025

Found ACL Anthology entry: https://aclanthology.org/2024.findings-emnlp.198

Thumbnail

@nschneid
Copy link
Contributor

abstract, author affiliations. I changed "Peter A." to "Peter" to match the PDF.

@nschneid nschneid added the approved Used to note team approval of metadata requests label Jan 10, 2025
@shramay-palta
Copy link
Author

Has this been resolved? I still do not see the abstract on the aclanthology page for the paper here: https://aclanthology.org/2024.findings-emnlp.198/

@nschneid
Copy link
Contributor

nschneid commented Jan 21, 2025

These approved fixes will be processed on a weekly basis. This issue will be closed at that time.

@mbollmann
Copy link
Member

@nschneid The issue is two weeks old though. It seems we last processed on Jan 9. I guess @mjpost is still triggering them manually for now since this is a new process.

@mjpost
Copy link
Member

mjpost commented Jan 21, 2025

Yes, sorry, I failed to set a reminder for the new schedule. See #4454.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
approved Used to note team approval of metadata requests correction for corrections submitted to the anthology metadata Correction to metadata
Projects
None yet
Development

Successfully merging a pull request may close this issue.

5 participants