Combining N-Grams and Graph Convolution for Text Classification
Loading...

Date
2025
Journal Title
Journal ISSN
Volume Title
Publisher
Elsevier
Open Access Color
Green Open Access
No
OpenAIRE Downloads
OpenAIRE Views
Publicly Funded
No
Abstract
Text classification, a cornerstone of natural language processing (NLP), finds applications in diverse areas, from sentiment analysis to topic categorization. While deep learning models have recently dominated the field, traditional n-gram-driven approaches often struggle to achieve comparable performance, particularly on large datasets. This gap largely stems from deep learning' s superior ability to capture contextual information through word embeddings. This paper explores a novel approach to leverage the often-overlooked power of n-gram features for enriching word representations and boosting text classification accuracy. We propose a method that transforms textual data into graph structures, utilizing discriminative n-gram series to establish long-range relationships between words. By training a graph convolution network on these graphs, we derive contextually enhanced word embeddings that encapsulate dependencies extending beyond local contexts. Our experiments demonstrate that integrating these enriched embeddings into an long-short term memory (LSTM) model for text classification leads to around 2% improvements in classification performance across diverse datasets. This achievement highlights the synergy of combining traditional n-gram features with graph-based deep learning techniques for building more powerful text classifiers.
Description
Bakal, Mehmet/0000-0003-2897-3894
ORCID
Keywords
Text-Graph Transformation, Graph Convolution Network, Deep Learning, Text Mining, Graph Mining
Fields of Science
Citation
WoS Q
Q1
Scopus Q
Q1

OpenCitations Citation Count
3
Source
Applied Soft Computing
Volume
175
Issue
Start Page
End Page
PlumX Metrics
Citations
CrossRef : 4
Scopus : 4
Captures
Mendeley Readers : 12
Google Scholar™


