Zusammenfassung: | |
Text classification is an essential task in natural language processing. While graph neural networks (GNNs) have successfully been applied to this problem both through graph classification and node classification approaches, their typical applications suffer from several issues. In the graph classification case, common graph construction techniques tend to leave out syntactic information. In the node classification case, most widespread datasets and applications tend to suffer from encoding relatively little information in the chosen node features. Finally, there are great benefits to be gained from combining the two GNN approaches. To tackle these concerns, we propose DepNet, a two-stage framework for text classification using GNN models. In the first stage we replace current graph construction methods by utilizing syntactic dependency parsing in order to include as much syntactic information in the GNN input as possible. In the second stage we combine both graph classification and node classification methods by utilizing the former to produce node embeddings for the latter, maximizing the potential of GNNs for text classification. We find that this technique significantly improves the performance of both graph classification and node classification approaches to text classification.
|
|
Lizenzbestimmungen: | CC BY 3.0 DE - http://creativecommons.org/licenses/by/3.0/de/ |
Publikationstyp: | BachelorThesis |
Publikationsstatus: | publishedVersion |
Erstveröffentlichung: | 2022 |
Schlagwörter (deutsch): | GNN, Graph Neural Network, Textklassifizierung |
Schlagwörter (englisch): | GNN, Graph Neural Network, Graph Neural Networks, Text Classification, Text, Classification, GCN, GraphConv, Dependency Parsing, Dependency, Parsing, Word2Vec, GloVe, spaCy, Stanza, TextGCN, TextING, MPAD, PGM-Explainer, GNNExplainer, Zorro |
Fachliche Zuordnung (DDC): | 004 | Informatik |
Anzeige der Dokumente mit ähnlichem Titel, Autor, Urheber und Thema.