Empirical Study of Tweets Topic Classification Using Transformer-Based Language Models

No Thumbnail Available
File version
Author(s)
Mandal, R
Chen, J
Becken, S
Stantic, B
Primary Supervisor
Other Supervisors
Editor(s)
Date
2021
Size
File type(s)
Location

Phuket, Thailand

License
Abstract

Social media opens up a great opportunity for policymakers to analyze and understand a large volume of online content for decision-making purposes. People’s opinions and experiences on social media platforms such as Twitter are extremely significant because of its volume, variety, and veracity. However, processing and retrieving useful information from natural language content is very challenging because of its ambiguity and complexity. Recent advances in Natural Language Understanding (NLU)-based techniques more specifically Transformer-based architecture solve sequence-to-sequence modeling tasks while handling long-range dependencies efficiently, and models based on transformers setting new benchmarks in performance across a wide variety of NLU-based tasks. In this paper, we applied transformer-based sequence modeling on short texts’ topic classification from tourist/user-posted tweets. Multiple BERT-like state-of-the-art sequence modeling approaches on topic/target classification tasks are investigated on the Great Barrier Reef tweet dataset and obtained findings can be valuable for researchers working on classification with large data sets and a large number of target classes.

Journal Title
Conference Title

Lecture Notes in Computer Science

Book Title
Edition
Volume

12672

Issue
Thesis Type
Degree Program
School
Publisher link
Patent number
Funder(s)
Grant identifier(s)
Rights Statement
Rights Statement
Item Access Status
Note
Access the data
Related item(s)
Subject

Sociology

Cultural studies

Information and computing sciences

Persistent link to this record
Citation

Mandal, R; Chen, J; Becken, S; Stantic, B, Empirical Study of Tweets Topic Classification Using Transformer-Based Language Models, Lecture Notes in Computer Science , 2021, 12672, pp. 340-350