Skip to content

blender-nlp/NewsClaims

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 

Repository files navigation

NewsClaims

This repository contains code and data for the paper:

NewsClaims: A New Benchmark for Claim Detection from News with Background Knowledge 

Arxiv link: https://arxiv.org/pdf/2112.08544.pdf

Data

We have released an updated version of the corpus with 889 claims over 143 news articles.

We release dev and test sets with 18 articles containing 103 claims and 125 articles containing 786 claims, respectively.

You can find the data here.

The data contains the following items:

  • dev.json: This is the development file.
  • test.json: This is the test file.
  • all_sents.json: This file contains the list of sentences for each news article (to be used for identifying claim sentences).
  • text: This folder contains the raw text for each news article.

Evaluation Script

The evaluation script for this benchmark is present at eval/eval_claim_detection.py which can be used as follows:

python eval_claim_detection.py --gold_file <path to dev/test.json> --predictions_file <path to outputs to evaluate> --<sub task to evaluate>

where the <sub task to evaluate> flag can be the following:

  • --eval_claim: Evaluate claim sentence detection. (Numbers here)
  • --eval_claimer: Evaluate claimer detection. (Numbers here)
  • --eval_claim_object: Evaluate claim object detection. (Numbers here)
  • --eval_claim_span: Evaluate claim span detection. (Numbers here)
  • --eval_stance: Evaluate stance detection. (Numbers here)

Numbers

We release updated numbers for each sub-task using the new version of the dataset:

Claim Detection

System P R F1
ClaimBuster 13.0 86.5 22.6
ClaimBuster + Zero-shot NLI 21.8 53.3 30.9
Human 52.7 70.0 60.1

Claim Object Detection

Approach Model Type F1
Prompting GPT-3 Zero-shot 15.2
Prompting T5 Zero-shot 11.4
In-context learning GPT-3 Few-Shot 51.9
Prompt-based fine-tuning T5 Few-Shot 51.6
Human - - 67.7

Stance Detection

Model Affirm F1 Refute Accuracy
Majority class 82.5 0.0 70.3
NLI (No topic) 89.1 68.0 83.8
NLI (With topic) 91.1 78.8 87.5
Human 97.0 84.2 94.9

Claim Span Detection

Model Precision Recall F1
PolNeAR-Content 67.0 42.8 52.3
Debater Boundary Detection 75.7 77.7 76.7
Human 82.7 90.9 86.6

Claimer Detection

Model F1 Reported Journalist
SRL 41.7 23.5 67.2
PolNeAR-Source 42.3 25.5 65.9
Human 85.8 81.3 88.9
Model In-Sentence F1 Out-of-Sentence F1
SRL 35.8 2.4
PolNeAR-Source 38.9 2.7

Citation

If you used this dataset in your work, please consider citing our paper:

@article{reddy2021newsclaims,
  title={NewsClaims: A New Benchmark for Claim Detection from News with Background Knowledge},
  author={Reddy, Revanth Gangi and Chinthakindi, Sai and Wang, Zhenhailong and Fung, Yi R and Conger, Kathryn S and Elsayed, Ahmed S and Palmer, Martha and Ji, Heng},
  journal={arXiv preprint arXiv:2112.08544},
  year={2021}
}

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages