This repository contains the data and code for our paper: Jungsoo Park*, Sewon Min*, Jaewoo Kang, Luke Zettlemoyer, Hannaneh Hajishirzi. "FaVIQ: FAct Verification from Information seeking Questions".
- Checkout the website
- Read the paper
- Download the FaVIQ: A set / R set
- Download the top-100 FaVIQ retrieval predictions from TF-IDF/DPR: see codes/README.
- Download the FaVIQ in fact correction task format: A set / R set
Update (Feb 2022): Our paper got accepted to the ACL 2022 main conference.
Update (Nov 2021): We updated the positive/negative silver evidence for the refute samples.
The data consists of A set and R set. A set is our main dataset, consisting of 26k claims converted from ambiguous questions and their disambiguations. R set is an additional dataset, consisting of 188k claims converted from regular question-answer pairs. Please refer to README for the detailed data format. Visit Explorer to see some samples!
Split \ Data | A set | R set |
---|---|---|
Train | 17,008 | 140,977 |
Dev | 4,260 | 15,566 |
Test | 4,688 | 5,877 |
{train,dev,test}.jsonl
files contains a list of dictionary that represents a single instance, with the following keys
id
(string): an identifier for the unique claim.claim
(string): a claim. the claims are all lowercased since the questions from NQ-Open and AmbigQA are all low-cased.label
(string): factuality of the claim which is either 'SUPPORTS' or 'REFUTES'.
As additional resources, we provide positive_evidence
and negative_evidence
associated with each claim, that could be useful for training baselines (e.g. training DPR).
positive_evidence
(dictionary): the top passage that contains the answer to the original question which is retrieved from querying the original question to TF-IDF.- id (string): id of the positive passage mapped to the wikipedia dump file.
- title (string): title of the positive passage.
- text (string): text of the positive passage.
negative_evidence
(dictionary): the top passage that does not contain the answer to the original question that is retrieved from querying the original question to TF-IDF.- id (string): id of the negative passage mapped to the wikipedia dump file.
- title (string): title of the negative passage.
- text (string): text of the negative passage.
Test data of our main dataset (A set) is hidden. This is to prevent overfitting on the test data, and also because the test data of AmbigQA, our source data, is hidden.
In order to obtain the test data, please email Jungsoo Park jungsoopark.1993@gmail.com
and Sewon Min sewon@cs.washington.edu
with the following information:
- Your name and affiliation
- Purpose of the data (e.g. publication)
- Model predictions and accuracy on the dev data of your model. This is to ensure that you evaluate the model on the test data only after you have finished developing the model on the dev data.
-
wikipedia_20190801.jsonl
: wikipedia database in jsonl format; containing the passages (~26M) with keys of passage id, title, and text. We take the plain text and lists provided by KILT and created a collection of passages where each passage has approximately 100 tokens. Note that for training baselines (FEVER and FaVIQ) in the paper, we concatenated the title with the passage and regarded it as the passage. -
wikipedia_20190801.db
: wikipedia database in sqlite format; containing the passages (~26M) with keys of passage id and values as title and passage. You can check out the code snippet for querying documents using SQLite in the link. -
FaVIQ retrieval predictions
: TF-IDF and DPR predictions of FaVIQ dataset. The predictions are retrieved passages' indexes from Wikipedia database (sqlite format). -
fact_correction_{a_set, r_set}.zip
: FaVIQ modified to fact correction task format; containing the instances with keys of input(refuted claim), and output(a list of corrected claims accordingly). We release the resource to aid in future works studying the fact correction, a task recently studied in this paper.
If you find the FaVIQ dataset useful, please cite our paper:
@inproceedings{ park2022faviq,
title={ {F}a{VIQ}: Fact Verification from Information seeking Questions },
author={ Park, Jungsoo and Min, Sewon and Kang, Jaewoo and Zettlemoyer, Luke and Hajishirzi, Hannaneh },
year={ 2022 },
booktitle={ ACL },
}
Please also make sure to credit and cite the creators of AmbigQA and Natural Questions, the datasets which we built ours off of:
@inproceedings{ min2020ambigqa,
title={ {A}mbig{QA}: Answering Ambiguous Open-domain Questions },
author={ Min, Sewon and Michael, Julian and Hajishirzi, Hannaneh and Zettlemoyer, Luke },
booktitle={ EMNLP },
year={ 2020 }
}
@article{ kwiatkowski2019natural,
title={ {N}atural {Q}uestions: a benchmark for question answering research },
author={ Kwiatkowski, Tom and Palomaki, Jennimaria and Redfield, Olivia and Collins, Michael and Parikh, Ankur and Alberti, Chris and Epstein, Danielle and Polosukhin, Illia and Devlin, Jacob and Lee, Kenton and others },
journal={ Transactions of the Association for Computational Linguistics },
year={ 2019 }
}
Please contact Jungsoo Park jungsoopark.1993@gmail.com
and Sewon Min sewon@cs.washington.edu
, or leave Github issues for any questions.