Comparing Word Representation BERT and RoBERTa in Keyphrase Extraction using TgGAT
Dublin Core
Title
Comparing Word Representation BERT and RoBERTa in Keyphrase Extraction using TgGAT
Subject
Keyphrase Extraction, BERT, RoBERTa, Pre-Trained Language Models, Topic-Guided Graph Attention Networks
Description
In this digital era, accessing vast amountsof information from websites and academic papers has become easier. However, efficiently locating relevant content remains challenging due to the overwhelming volume of data. Keyphrase Extraction Systems automatethe process of generating phrases that accurately represent a document’s maintopics. These systems are crucialfor supporting various natural language processing tasks, such as text summarization, information retrieval, and representation.The traditional method of manually selecting key phrasesis still common but often proves inefficient and inconsistent in summarizing the main ideas of a document. This study introduces an approach that integrates pre-trained language models, BERT and RoBERTa, with Topic-Guided Graph Attention Networks (TgGAT) to enhance keyphrase extraction. TgGAT strengthens the extraction process by combining topic modellingwith graph-based structures, providing a more structured and context-aware representation of a document’s key topics. By leveraging the strengths of both graph-based and transformer-based models, this research proposes a framework that improves keyphrase extraction performance. This is the first to apply graph-based and PLM methods for keyphrase extraction in the Indonesian language. Theresults revealed that BERT outperformed RoBERTa, with precision, recall, and F1-scores of 0.058, 0.070, and 0.062, respectively, compared to RoBERTa’s 0.026, 0.030, and 0.027.Theresult showsthat BERT with TgGAT obtained more representative keyphrases than RoBERTa with TgGAT. These findings underline the benefits of integrating graph-based approaches with pre-trained modelsfor capturing both semantic relationships and topic relevance
Creator
Novi Yusliani1*, Aini Nabilah2, Muhammad Raihan Habibullah3, Annisa Darmawahyuni4, Ghita Athalina
Source
https://jurnal.iaii.or.id/index.php/RESTI/article/view/6279/1034
Publisher
Department of Informatic Engineering, Faculty of Computer Science, Universitas Sriwijaya, Palembang, Indonesia
Date
20-03-2025
Contributor
FAJAR BAGUS W
Format
PDF
Language
ENGLISH
Type
TEXT
Files
Collection
Citation
Novi Yusliani1*, Aini Nabilah2, Muhammad Raihan Habibullah3, Annisa Darmawahyuni4, Ghita Athalina, “Comparing Word Representation BERT and RoBERTa in Keyphrase Extraction using TgGAT,” Repository Horizon University Indonesia, accessed January 26, 2026, https://repository.horizon.ac.id/items/show/10490.