Skip to content

PeterQiu0516/Click-BERT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Click-BERT

Click-BERT: Clickbait Detector with Bidirectional Encoder Representations from Transformers

Course Project for EECS 498 NLP, Winter 2021

Team Members: Changyuan Qiu (@PeterQiu0516), Chenshu Zhu (@Jupiter), Haoxuan Shan (@shanhx2000)

Abstract

Clickbait is a rampant problem haunting online readers. Nowadays, clickbait detection in tweets remains an elusive challenge. In this paper, we propose Clickbait Detector with Bidirectional Encoder Representations from Transformers (Click-BERT), which could effectively identify clickbaits utilizing recent advances in pre-training methods (BERT and Longformer) and recurrent neural networks (Bi- LSTM and Bi-GRU) with a parallel model structure. Our model supports end-to-end training without involving any hand-crafted features and achieved new state-of-the-art results on the Webis 17 Dataset.

Highlights

  • Click-BERT won the best presentation award among 17 teams!

  • Final Report released.

  • Presentation Slides released.

  • Our baseline2: headline Bi-LSTM with BERT achieved new SOTA on accuracy and MSE on Webis 17! Get more details in our progress report.

About

EECS 498 NLP Project - Click-BERT: Clickbait Detector with Bidirectional Encoder Representations from Transformers

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published