Empirical Study of Tweets Topic Classification Using Transformer-Based Language Models
File version
Author(s)
Chen, J
Becken, S
Stantic, B
Griffith University Author(s)
Primary Supervisor
Other Supervisors
Editor(s)
Date
Size
File type(s)
Location
Phuket, Thailand
License
Abstract
Social media opens up a great opportunity for policymakers to analyze and understand a large volume of online content for decision-making purposes. People’s opinions and experiences on social media platforms such as Twitter are extremely significant because of its volume, variety, and veracity. However, processing and retrieving useful information from natural language content is very challenging because of its ambiguity and complexity. Recent advances in Natural Language Understanding (NLU)-based techniques more specifically Transformer-based architecture solve sequence-to-sequence modeling tasks while handling long-range dependencies efficiently, and models based on transformers setting new benchmarks in performance across a wide variety of NLU-based tasks. In this paper, we applied transformer-based sequence modeling on short texts’ topic classification from tourist/user-posted tweets. Multiple BERT-like state-of-the-art sequence modeling approaches on topic/target classification tasks are investigated on the Great Barrier Reef tweet dataset and obtained findings can be valuable for researchers working on classification with large data sets and a large number of target classes.
Journal Title
Conference Title
Lecture Notes in Computer Science
Book Title
Edition
Volume
12672
Issue
Thesis Type
Degree Program
School
Publisher link
Patent number
Funder(s)
Grant identifier(s)
Rights Statement
Rights Statement
Item Access Status
Note
Access the data
Related item(s)
Subject
Sociology
Cultural studies
Information and computing sciences
Persistent link to this record
Citation
Mandal, R; Chen, J; Becken, S; Stantic, B, Empirical Study of Tweets Topic Classification Using Transformer-Based Language Models, Lecture Notes in Computer Science , 2021, 12672, pp. 340-350