TY - JOUR
T1 - Towards low-resource rumor detection
T2 - Unified contrastive transfer with propagation structure
AU - Lin, Hongzhan
AU - Ma, Jing
AU - Yang, Ruichao
AU - Yang, Zhiwei
AU - Cheng, Mingfei
N1 - This work is partially supported by National Natural Science Foundation of China Young Scientists Fund (No. 62206233) and Hong Kong RGC ECS (No. 22200722).
Publisher Copyright:
© 2024 Elsevier B.V.
PY - 2024/4/14
Y1 - 2024/4/14
N2 - The truth is significantly hampered by massive rumors that spread along with breaking news or popular topics. Since there is sufficient corpus gathered from the same domain for model training, existing rumor detection algorithms show promising performance on yesterday's news. However, due to a lack of substantial training data and prior expert knowledge, they are poor at spotting rumors concerning unforeseen events, especially those propagated in different languages (i.e., low-resource regimes). In this paper, we propose a simple yet effective framework with unified contrastive transfer learning, to detect rumors by adapting the features learned from well-resourced rumor data to that of the low-resourced with only few-shot annotations. More specifically, we first represent rumor circulated on social media as an undirected topology for enhancing the interaction of user opinions, and then train the propagation-structured model via a unified contrastive paradigm to mine effective clues simultaneously from both post semantics and propagation structure. Our model explicitly breaks the barriers of the domain and/or language issues, via language alignment and a novel domain-adaptive contrastive learning mechanism. To well-generalize the representation learning using a small set of annotated target events, we reveal that rumor-indicative signal is closely correlated with the uniformity of the distribution of these events. We design a target-wise contrastive training mechanism with three event-level data augmentation strategies, capable of unifying the representations by distinguishing target events. Extensive experiments conducted on four low-resource datasets collected from real-world microblog platforms demonstrate that our framework achieves much better performance than state-of-the-art methods and exhibits a superior capacity for detecting rumors at early stages.
AB - The truth is significantly hampered by massive rumors that spread along with breaking news or popular topics. Since there is sufficient corpus gathered from the same domain for model training, existing rumor detection algorithms show promising performance on yesterday's news. However, due to a lack of substantial training data and prior expert knowledge, they are poor at spotting rumors concerning unforeseen events, especially those propagated in different languages (i.e., low-resource regimes). In this paper, we propose a simple yet effective framework with unified contrastive transfer learning, to detect rumors by adapting the features learned from well-resourced rumor data to that of the low-resourced with only few-shot annotations. More specifically, we first represent rumor circulated on social media as an undirected topology for enhancing the interaction of user opinions, and then train the propagation-structured model via a unified contrastive paradigm to mine effective clues simultaneously from both post semantics and propagation structure. Our model explicitly breaks the barriers of the domain and/or language issues, via language alignment and a novel domain-adaptive contrastive learning mechanism. To well-generalize the representation learning using a small set of annotated target events, we reveal that rumor-indicative signal is closely correlated with the uniformity of the distribution of these events. We design a target-wise contrastive training mechanism with three event-level data augmentation strategies, capable of unifying the representations by distinguishing target events. Extensive experiments conducted on four low-resource datasets collected from real-world microblog platforms demonstrate that our framework achieves much better performance than state-of-the-art methods and exhibits a superior capacity for detecting rumors at early stages.
KW - Contrastive learning
KW - Few-shot transfer
KW - Low resource
KW - Propagation structure
KW - Rumor detection
UR - http://www.scopus.com/inward/record.url?scp=85185832368&partnerID=8YFLogxK
U2 - 10.1016/j.neucom.2024.127438
DO - 10.1016/j.neucom.2024.127438
M3 - Journal article
AN - SCOPUS:85185832368
SN - 0925-2312
VL - 578
JO - Neurocomputing
JF - Neurocomputing
M1 - 127438
ER -