SemEval-2021 Task 9: Fact Verification and Evidence Finding for Tabular Data in Scientific Documents (SEM-TAB-FACTS)

@inproceedings{Wang2021SemEval2021T9,
  title={SemEval-2021 Task 9: Fact Verification and Evidence Finding for Tabular Data in Scientific Documents (SEM-TAB-FACTS)},
  author={Nancy Xin Ru Wang and Diwakar Mahajan and Marina Danilevsky and Sara Rosenthal},
  booktitle={SEMEVAL},
  year={2021}
}
Understanding tables is an important and relevant task that involves understanding table structure as well as being able to compare and contrast information within cells. In this paper, we address this challenge by presenting a new dataset and tasks that addresses this goal in a shared task in SemEval 2020 Task 9: Fact Verification and Evidence Finding for Tabular Data in Scientific Documents (SEM-TAB-FACTS). Our dataset contains 981 manually-generated tables and an auto-generated dataset of… 
TAPAS at SemEval-2021 Task 9: Reasoning over tables with intermediate pre-training
TLDR
The TAPAS contribution to the Shared Task on Statement Verification and Evidence Finding with Tables (SemEval 2021 Task 9, Wang et al. (2021) is presented and the artificial neutral examples are somewhat effective at training the first model.
AttesTable at SemEval-2021 Task 9: Extending Statement Verification with Tables for Unknown Class, and Semantic Evidence Finding
TLDR
This paper describes the approach for Task 9 of SemEval 2021: Statement Verification and Evidence Finding with Tables, whereby the TAPAS model is extended to adapt to the ‘unknown’ class of statements by finetuning it on an augmented version of the task data.
The Fact Extraction and VERification Over Unstructured and Structured information (FEVEROUS) Shared Task
The Fact Extraction and VERification Over Unstructured and Structured information (FEVEROUS) shared task, asks participating systems to determine whether human-authored claims are Supported or
Is My Model Using The Right Evidence? Systematic Probes for Examining Evidence-Based Tabular Reasoning
TLDR
The authors' experiments demonstrate that a BERT-based model representative of today’s state-of-the-art fails to properly reason on the following counts: it often misses the relevant evidence, suffers from hypothesis and knowledge biases, and relies on annotation artifacts and knowledge from pretrained language models as primary evidence rather than relying on reasoning on the premises in the tabular input.
TabPert: An Effective Platform for Tabular Perturbation
TLDR
TABPERT allows a user to update a table, change its associated hypotheses, change their labels, and highlight rows that are important for hypothesis classification, and captures information about the techniques used to automatically produce the table, as well as the strategies employed to generate the challenging hypotheses.
A Survey on Automated Fact-Checking
TLDR
This paper surveys automated fact-checking stemming from natural language processing, and presents an overview of existing datasets and models, aiming to unify the various definitions given and identify common concepts.

References

SHOWING 1-10 OF 32 REFERENCES
Volta at SemEval-2021 Task 9: Statement Verification and Evidence Finding with Tables using TAPAS and Transfer Learning
TLDR
This paper fine-tune TAPAS (a model which extends BERT’s architecture to capture tabular structure) for both the subtasks as it has shown state-of-the-art performance in various table understanding tasks.
TabFact: A Large-scale Dataset for Table-based Fact Verification
TLDR
A large-scale dataset with 16k Wikipedia tables as the evidence for 118k human-annotated natural language statements, which are labeled as either ENTAILED or REFUTED is constructed and two different models are designed: Table-BERT and Latent Program Algorithm (LPA).
TAPAS at SemEval-2021 Task 9: Reasoning over tables with intermediate pre-training
TLDR
The TAPAS contribution to the Shared Task on Statement Verification and Evidence Finding with Tables (SemEval 2021 Task 9, Wang et al. (2021) is presented and the artificial neutral examples are somewhat effective at training the first model.
AttesTable at SemEval-2021 Task 9: Extending Statement Verification with Tables for Unknown Class, and Semantic Evidence Finding
TLDR
This paper describes the approach for Task 9 of SemEval 2021: Statement Verification and Evidence Finding with Tables, whereby the TAPAS model is extended to adapt to the ‘unknown’ class of statements by finetuning it on an augmented version of the task data.
BOUN at SemEval-2021 Task 9: Text Augmentation Techniques for Fact Verification in Tabular Data
TLDR
This paper presents their text augmentation based approach for the Table Statement Support Subtask (Phase A) of SemEval-2021 Task 9, and investigates the impact of domain adaptation and joint learning on fact verification in tabular data by utilizing the SemTabFacts and TabFact datasets.
SemEval-2019 Task 8: Fact Checking in Community Question Answering Forums
TLDR
This work presents SemEval-2019 Task 8 on Fact Checking in Community Question Answering Forums, which features two subtasks, and received 17 official submissions for subtask A and 11official submissions for Subtask B.
KaushikAcharya at SemEval-2021 Task 9: Candidate Generation for Fact Verification over Tables
This paper describes the system submitted in the SemEval-2021 Statement Verification and Evidence Finding with Tables task. The system relies on candidate generation for logical forms on the table
INFOTABS: Inference on Tables as Semi-structured Data
In this paper, we observe that semi-structured tabulated text is ubiquitous; understanding them requires not only comprehending the meaning of text fragments, but also implicit relationships between
TaBERT: Pretraining for Joint Understanding of Textual and Tabular Data
TLDR
TaBERT is a pretrained LM that jointly learns representations for NL sentences and (semi-)structured tables that achieves new best results on the challenging weakly-supervised semantic parsing benchmark WikiTableQuestions, while performing competitively on the text-to-SQL dataset Spider.
SciTaiL: A Textual Entailment Dataset from Science Question Answering
TLDR
A new dataset and model for textual entailment, derived from treating multiple-choice question-answering as an entailment problem, is presented, and it is demonstrated that one can improve accuracy on SCITAIL by 5% using a new neural model that exploits linguistic structure.
...
1
2
3
4
...