Skip to main content

A Two-Stage Model Based on BERT for Short Fake News Detection

  • Conference paper
  • First Online:
Knowledge Science, Engineering and Management (KSEM 2019)

Abstract

Online social media promotes the development of the news industry and make it easy for everyone to obtain the latest news. Meanwhile, the circumstances get worse because of fake news. Fake news is flooding and become a serious threat which may cause high societal and economic losses, making fake news detection important. Unlike traditional one, news on social media tends to be short and misleading, which is more confusing to identify. On the other hand, fake news may contain parts of the facts and parts of the incorrect contents in one statement, which is not so clear and simple to classify. Hence, we propose a two-stage model to deal with the difficulties. Our model is built on BERT, a pre-trained model with a more powerful feature extractor Transformer instead of CNN or RNN. Besides, some accessible information is used to extend features and calculate attention weights. At last, inspired by fine-grained sentiment analysis, we treat fake news detection as fine-grained multiple-classification task and use two similar sub-models to identify different granularity labels separately. We evaluate our model on a real-world benchmark dataset. The experimental results demonstrate its effectiveness in fine-grained fake news detection and its superior performance to the baselines and other competitive approaches.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    http://www.politifact.com/.

References

  1. Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473 (2014)

  2. Conroy, N.J., Rubin, V.L., Chen, Y.: Automatic deception detection: methods for finding fake news. In: Proceedings of the 78th ASIS&T Annual Meeting: Information Science with Impact: Research in and for the Community, p. 82. American Society for Information Science (2015)

    Google Scholar 

  3. Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT) (2019)

    Google Scholar 

  4. Hendrycks, D., Gimpel, K.: Bridging nonlinearities and stochastic regularizers with Gaussian error linear units. arXiv preprint arXiv:1606.08415 (2016)

  5. Karimi, H., Roy, P., Saba-Sadiya, S., Tang, J.: Multi-source multi-class fake news detection. In: Proceedings of the 27th International Conference on Computational Linguistics (COLING), pp. 1546–1557 (2018)

    Google Scholar 

  6. Liu, X., He, P., Chen, W., Gao, J.: Multi-task deep neural networks for natural language understanding. arXiv preprint arXiv:1901.11504 (2019)

  7. Liu, Y., Wu, Y.F.B.: Early detection of fake news on social media through propagation path classification with recurrent and convolutional networks. In: Thirty-Second AAAI Conference on Artificial Intelligence (2018)

    Google Scholar 

  8. Luong, T., Pham, H., Manning, C.D.: Effective approaches to attention-based neural machine translation. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP), Lisbon, Portugal, pp. 1412–1421, September 2015

    Google Scholar 

  9. Peters, M., et al.: Deep contextualized word representations. In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), New Orleans, Louisiana, pp. 2227–2237, June 2018

    Google Scholar 

  10. Qian, F., Gong, C., Sharma, K., Liu, Y.: Neural user response generator: fake news detection with collective user intelligence. In: IJCAI, vol. 3834, p. 3840 (2018)

    Google Scholar 

  11. Rubin, V.L.: Deception detection and rumor debunking for social media. In: The SAGE Handbook of Social Media Research Methods, p. 342. SAGE (2017)

    Chapter  Google Scholar 

  12. Shu, K., Sliva, A., Wang, S., Tang, J., Liu, H.: Fake news detection on social media: a data mining perspective. ACM SIGKDD Explor. Newslett. 19(1), 22–36 (2017)

    Article  Google Scholar 

  13. Shu, K., Wang, S., Liu, H.: Beyond news contents: the role of social context for fake news detection. In: Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining (WSDM), pp. 312–320. ACM (2019)

    Google Scholar 

  14. Thorne, J., Vlachos, A., Christodoulopoulos, C., Mittal, A.: FEVER: a large-scale dataset for fact extraction and verification. In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), New Orleans, Louisiana, pp. 809–819, June 2018

    Google Scholar 

  15. Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems (NIPS), pp. 5998–6008 (2017)

    Google Scholar 

  16. Vlachos, A., Riedel, S.: Fact checking: task definition and dataset construction. In: Proceedings of the ACL 2014 Workshop on Language Technologies and Computational Social Science, pp. 18–22 (2014)

    Google Scholar 

  17. Wang, B., Liu, K., Zhao, J.: Inner attention based recurrent neural networks for answer selection. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (ACL), vol. 1, pp. 1288–1297 (2016)

    Google Scholar 

  18. Wang, W.Y.: “Liar, liar pants on fire”: a new benchmark dataset for fake news detection. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (ACL), pp. 422–426 (2017)

    Google Scholar 

  19. Xu, K., et al.: Show, attend and tell: neural image caption generation with visual attention. In: International Conference on Machine Learning (ICML), pp. 2048–2057 (2015)

    Google Scholar 

  20. Zhang, H., et al.: Pretraining-based natural language generation for text summarization. arXiv preprint arXiv:1902.09243 (2019)

Download references

Acknowledgment

This work is supported by National Key R&D Program of China (No. 2018YFB0803402), National Natural Science Foundation of China (No. 61402476), the International Cooperation Project of Institute of Information Engineering, Chinese Academy of Sciences under Grant No. Y7Z0511101.

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Min Yu or Gang Li .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Liu, C. et al. (2019). A Two-Stage Model Based on BERT for Short Fake News Detection. In: Douligeris, C., Karagiannis, D., Apostolou, D. (eds) Knowledge Science, Engineering and Management. KSEM 2019. Lecture Notes in Computer Science(), vol 11776. Springer, Cham. https://doi.org/10.1007/978-3-030-29563-9_17

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-29563-9_17

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-29562-2

  • Online ISBN: 978-3-030-29563-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics