Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Paper Metadata: 2023.findings-emnlp.512 #2920

Closed
3 of 4 tasks
elangovana opened this issue Dec 9, 2023 · 0 comments · Fixed by #2930
Closed
3 of 4 tasks

Paper Metadata: 2023.findings-emnlp.512 #2920

elangovana opened this issue Dec 9, 2023 · 0 comments · Fixed by #2930
Assignees
Labels
correction for corrections submitted to the anthology metadata Correction to metadata

Comments

@elangovana
Copy link
Contributor

elangovana commented Dec 9, 2023

Confirm that this is a metadata correction

  • I want to file corrections to make the metadata match the PDF file hosted on the ACL Anthology.

Anthology ID

2023.findings-emnlp.512

Type of Paper Metadata Correction

  • Paper Title
  • Paper Abstract
  • Author Name(s)

Correction to Paper Title

Effects of Human Adversarial and Affable Samples on BERT Generalization

Correction to Paper Abstract

BERT-based models have had strong performance on leaderboards, yet have been demonstrably worse in real-world settings requiring generalization. Limited quantities of training data is considered a key impediment to achieving generalizability in machine learning. In this paper, we examine the impact of training data quality, not quantity, on a model’s generalizability. We consider two characteristics of training data: the portion of human-adversarial (h- adversarial), i.e. sample pairs with seemingly minor differences but different ground-truth labels, and human-affable (h-affable) training samples, i.e. sample pairs with minor differences but the same ground-truth label. We find that for a fixed size of training samples, as a rule of thumb, having 10-30% h-adversarial instances improves the precision, and therefore F1, by up to 20 points in the tasks of text classification and relation extraction. Increasing h-adversarials beyond this range can result in performance plateaus or even degradation. In contrast, h-affables may not contribute to a model’s generalizability and may even degrade generalization performance

Correction to Author Name(s)

No response

@elangovana elangovana added correction for corrections submitted to the anthology metadata Correction to metadata labels Dec 9, 2023
@elangovana elangovana changed the title 2023.findings-emnlp.512 Paper metadata correction 2023.findings-emnlp.512 Dec 9, 2023
@elangovana elangovana changed the title Paper metadata correction 2023.findings-emnlp.512 Paper Metadata: 2023.findings-emnlp.512 Dec 9, 2023
elangovana added a commit to elangovana/acl-anthology that referenced this issue Dec 10, 2023
@mjpost mjpost linked a pull request Dec 10, 2023 that will close this issue
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
correction for corrections submitted to the anthology metadata Correction to metadata
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants