Multimodal AI Lab Wins Fact-Checking Challenge
2024/12/19
The lab's "InFact" method won a global claim verification competition. Today, the lab unveils the code and releases the stronger, multimodal successor, "DEFAME."

The Multimodal AI Lab at TU Darmstadt has achieved a remarkable milestone by winning first place in a global challenge focused on automating text-only claim verification. Competing against 20 highly skilled teams, the lab’s innovative approach, “InFact,” emerged as the strongest in the challenge.
InFact, primarily developed by and Mark Rothermel, breaks down the complex task of fact-checking into a multi-step pipeline. At each stage, the system leverages the capabilities of Large Language Models (LLMs) to interpret, decompose, and verify claims with precision. Most importantly, InFact does not primarily rely on the LLM's internal (often unreliable) knowledge but retrieves the necessary evidence from external sources to yield a more trustworthy prediction. Tobias Braun
Mark Rothermel presented this cutting-edge work at the in Miami, Florida, last November (watch his 7 min talk in the EMNLP conference, starting at 3:09:30). The Multimodal AI Lab has also made the InFact code live recording, encouraging the global research community to explore and build upon their work. For those interested in the technical details, publicly available today on GitHub offers an in-depth look at InFact's architecture and methodology. the system description paper
Successor Method “DEFAME” Just Released, Too
InFact’s success positions it as the de-facto state-of-the-art method for automated textual fact-checking. However, the team emphasizes that this is just the beginning of their efforts to combat misinformation. Mark and Tobias are already working on a multimodal successor to InFact, which is designed to handle both textual and visual information. This next-generation method, which they call “DEFAME,” aims to address the prevalence of visual misinformation like cheap fakes: pristine photos used out of context to construct false narratives.
In their , Tobias and Mark report the outstanding success of DEFAME, which beats all previous methods on three different fact-checking benchmarks. Together with InFact, the team freshly released under-review preprint. released DEFAME's code on GitHub
The Urgency of Fact-Checking
The need for effective fact-checking methods has never been greater. Artificial Intelligence not only enables the mass creation of fake news and disinformation but also enhances its believability, increasing both the quantity and quality of misleading content. Disinformation has been identified by the World Economic Forum as the largest mid-term risk facing society—even surpassing the challenges of climate change and ongoing geopolitical conflicts. Additionally, human fact-checkers can only deal with a tiny part of fake news. With the release of InFact and DEFAME, the Multimodal AI Lab joins the urgently needed research forefront to automate fact-checking.