Exploiting Rationale Data for Explainable NLP Models

Zur Kurzanzeige

dc.identifier.uri http://dx.doi.org/10.15488/11526
dc.identifier.uri https://www.repo.uni-hannover.de/handle/123456789/11615
dc.contributor.advisor Anand, Avishek
dc.contributor.author Reimer, Maximilian eng
dc.date.accessioned 2021-11-19T10:45:19Z
dc.date.available 2021-11-19T10:45:19Z
dc.date.issued 2021-09-15
dc.identifier.citation Reimer, Maximilian: Exploiting Rationale Data for Explainable NLP Models. Hannover : Gottfried Wilhelm Leibniz Universität, Master Thesis, xvii, 95 S. DOI: https://doi.org/10.15488/11526 eng
dc.description.abstract In recent years, deep learning models have become very powerful – even outperforming humans on a variety of tasks. This enables more real-world applications, including also sensitive fields such as medical diagnoses or jurisdiction. Besides achieving sufficiently good performance, the requirement to justify and explain the models’ decisions is becoming increasingly important. This work aims to enable a broader application of a specific model class that is inherently interpretable, namely explain-then-predict models, by reducing the annotation cost of the explanations. We focus on the ExPred model as a representative of explain-then-predict models. We investigate its dependency on rationale annotations, a special kind of explanation, through training using gradually fewer rationale-labeled instances. Furthermore, we ex- plore different approaches that aim to reduce the number of human-labeled instances required during training, such as active learning and weak supervision. Our results show that even with only a fraction of instances annotated with rationales from the original dataset, ExPred still achieves good performance (within 95% of the performance when using 100% annotation). Depending on the dataset, only a few thousand annotated rationales are required. Using weak supervision, this can be further reduced, at least in specific settings. On the Movie Reviews dataset, we achieve good performance with only 5% of the original rational labels. The tested off-the-shelf active learning methods do not provide any benefit over randomly selecting instances to label. However, the extensive behavior analysis enables the future design of active learning methods that are tailored to explain-then-predict models. We start by proposing an active learning method that outperforms the random baseline on the Movie Reviews dataset. eng
dc.language.iso eng eng
dc.publisher Hannover : Gottfried Wilhelm Leibniz Universität
dc.rights CC BY 3.0 DE eng
dc.rights.uri http://creativecommons.org/licenses/by/3.0/de/ eng
dc.subject Semi-Supervision eng
dc.subject Weak Supervision eng
dc.subject Active Learning eng
dc.subject Natural Language Processing eng
dc.subject NLP eng
dc.subject Interpretable Machine Learning eng
dc.subject Expred eng
dc.subject Interpretable Machine Learning ger
dc.subject NLP ger
dc.subject Natural Language Processing ger
dc.subject Expred ger
dc.subject Active Learning ger
dc.subject Weak Supervision ger
dc.subject Semi-Supervision ger
dc.subject.ddc 600 | Technik eng
dc.subject.ddc 004 | Informatik eng
dc.title Exploiting Rationale Data for Explainable NLP Models eng
dc.type MasterThesis eng
dc.type Text eng
dcterms.extent xvii, 95 S.
dc.description.version publishedVersion eng
tib.accessRights frei zug�nglich eng


Die Publikation erscheint in Sammlung(en):

Zur Kurzanzeige

 

Suche im Repositorium


Durchblättern

Mein Nutzer/innenkonto

Nutzungsstatistiken