Command Palette
Search for a command to run...
Towards LLM-based Fact Verification on News Claims with a Hierarchical Step-by-Step Prompting Method
Xuan Zhang Wei Gao

Abstract
While large pre-trained language models (LLMs) have shown their impressive capabilities in various NLP tasks, they are still under-explored in the misinformation domain. In this paper, we examine LLMs with in-context learning (ICL) for news claim verification, and find that only with 4-shot demonstration examples, the performance of several prompting methods can be comparable with previous supervised models. To further boost performance, we introduce a Hierarchical Step-by-Step (HiSS) prompting method which directs LLMs to separate a claim into several subclaims and then verify each of them via multiple questions-answering steps progressively. Experiment results on two public misinformation datasets show that HiSS prompting outperforms state-of-the-art fully-supervised approach and strong few-shot ICL-enabled baselines.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| fake-news-detection-on-rawfc | Standard prompting with articles | F1: 47.9 |
| fake-news-detection-on-rawfc | HiSS | F1: 53.9 |
| fake-news-detection-on-rawfc | CoT | F1: 44.4 |
| fake-news-detection-on-rawfc | ReAct | F1: 49.8 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.