Lidé
Ing. Herbert Ullrich
Všechny publikace
AIC CTU system at AVeriTeC: Re-framing automated fact-checking as a simple RAG task
- Autoři: Ing. Herbert Ullrich, Mlynář, T., Ing. Jan Drchal, Ph.D.,
- Publikace: Seventh Workshop on Fact Extraction and VERification (FEVER 2024). Stroudsburg: Association for Computational Linguistics (ACL), 2024. p. 137-150. ISBN 979-8-3313-0845-2.
- Rok: 2024
- DOI: 10.18653/v1/2024.fever-1.16
- Odkaz: https://doi.org/10.18653/v1/2024.fever-1.16
- Pracoviště: Katedra počítačů, Centrum umělé inteligence
-
Anotace:
This paper describes our 3rd place submission in the AVeriTeC shared task in which we attempted to address the challenge of fact-checking with evidence retrieved in the wild using a simple scheme of Retrieval-Augmented Generation (RAG) designed for the task, leveraging the predictive power of Large Language Models. We release our codebase and explain its two modules – the Retriever and the Evidence & Label generator – in detail, justifying their features such as MMR-reranking and Likert-scale confidence estimation. We evaluate our solution on AVeriTeC dev and test set and interpret the results, picking the GPT-4o as the most appropriate model for our pipeline at the time of our publication, with Llama 3.1 70B being a promising open-source alternative. We perform an empirical error analysis to see that faults in our predictions often coincide with noise in the data or ambiguous fact-checks, provoking further research and data augmentation.
CsFEVER and CTKFacts: Acquiring Czech Data for Fact Verification
- Autoři: Ing. Herbert Ullrich, Ing. Jan Drchal, Ph.D., Rýpar, M., Vincourová, H., Moravec, V.
- Publikace: Language Resources and Evaluation. 2023, 57(4), 1571-1605. ISSN 1574-020X.
- Rok: 2023
- DOI: 10.1007/s10579-023-09654-3
- Odkaz: https://doi.org/10.1007/s10579-023-09654-3
- Pracoviště: Katedra počítačů, Centrum umělé inteligence
-
Anotace:
In this paper, we examine several methods of acquiring Czech data for automated fact-checking, which is a task commonly modeled as a classification of textual claim veracity w.r.t. a corpus of trusted ground truths. We attempt to collect sets of data in form of a factual claim, evidence within the ground truth corpus, and its veracity label (supported, refuted or not enough info). As a first attempt, we generate a Czech version of the large-scale FEVER dataset built on top of Wikipedia corpus. We take a hybrid approach of machine translation and document alignment; the approach and the tools we provide can be easily applied to other languages. We discuss its weaknesses, propose a future strategy for their mitigation and publish the 127k resulting translations, as well as a version of such dataset reliably applicable for the Natural Language Inference task—the CsFEVER-NLI. Furthermore, we collect a novel dataset of 3,097 claims, which is annotated using the corpus of 2.2 M articles of Czech News Agency. We present an extended dataset annotation methodology based on the FEVER approach, and, as the underlying corpus is proprietary, we also publish a standalone version of the dataset for the task of Natural Language Inference we call CTKFactsNLI. We analyze both acquired datasets for spurious cues—annotation patterns leading to model overfitting. CTKFacts is further examined for inter-annotator agreement, thoroughly cleaned, and a typology of common annotator errors is extracted. Finally, we provide baseline models for all stages of the fact-checking pipeline and publish the NLI datasets, as well as our annotation platform and other experimental data.