Zur Kurzanzeige

dc.identifier.uri http://dx.doi.org/10.15488/15769
dc.identifier.uri https://www.repo.uni-hannover.de/handle/123456789/15893
dc.contributor.author Leonhardt, Lutz Jurek eng
dc.date.accessioned 2023-12-20T11:49:28Z
dc.date.available 2023-12-20T11:49:28Z
dc.date.issued 2023
dc.identifier.citation Leonhardt, Lutz Jurek: Efficient and Explainable Neural Ranking. Hannover : Gottfried Wilhelm Leibniz Universität, Diss., 2023, xii, 165 S., DOI: https://doi.org/10.15488/15769 eng
dc.description.abstract The recent availability of increasingly powerful hardware has caused a shift from traditional information retrieval (IR) approaches based on term matching, which remained the state of the art for several decades, to large pre-trained neural language models. These neural rankers achieve substantial improvements in performance, as their complexity and extensive pre-training give them the ability of understanding natural language in a way. As a result, neural rankers go beyond term matching by performing relevance estimation based on the semantics of queries and documents. However, these improvements in performance don't come without sacrifice. In this thesis, we focus on two fundamental challenges of neural ranking models, specifically, ones based on large language models: On the one hand, due to their complexity, the models are inefficient; they require considerable amounts of computational power, which often comes in the form of specialized hardware, such as GPUs or TPUs. Consequently, the carbon footprint is an increasingly important aspect of systems using neural IR. This effect is amplified when low latency is required, as in, for example, web search. On the other hand, neural models are known for being inherently unexplainable; in other words, it is often not comprehensible for humans why a neural model produced a specific output. In general, explainability is deemed important in order to identify undesired behavior, such as bias. We tackle the efficiency challenge of neural rankers by proposing Fast-Forward indexes, which are simple vector forward indexes that heavily utilize pre-computation techniques. Our approach substantially reduces the computational load during query processing, enabling efficient ranking solely on CPUs without requiring hardware acceleration. Furthermore, we introduce BERT-DMN to show that the training efficiency of neural rankers can be improved by training only parts of the model. In order to improve the explainability of neural ranking, we propose the Select-and-Rank paradigm to make ranking models explainable by design: First, a query-dependent subset of the input document is extracted to serve as an explanation; second, the ranking model makes its decision based only on the extracted subset, rather than the complete document. We show that our models exhibit performance similar to models that are not explainable by design and conduct a user study to determine the faithfulness of the explanations. Finally, we introduce BoilerNet, a web content extraction technique that allows the removal of boilerplate from web pages, leaving only the main content in plain text. Our method requires no feature engineering and can be used to aid in the process of creating new document corpora from the web. eng
dc.language.iso ger eng
dc.publisher Hannover : Institutionelles Repositorium der Leibniz Universität Hannover
dc.rights CC BY 3.0 DE eng
dc.rights.uri http://creativecommons.org/licenses/by/3.0/de/ eng
dc.subject Information retrieval eng
dc.subject neural ranking eng
dc.subject efficiency eng
dc.subject explainability eng
dc.subject Information Retrieval ger
dc.subject Neuronale Rankingmodelle ger
dc.subject Effizienz ger
dc.subject Erklärbarkeit ger
dc.subject.ddc 004 | Informatik eng
dc.title Efficient and Explainable Neural Ranking eng
dc.type DoctoralThesis eng
dc.type Text eng
dc.relation.doi https://doi.org/10.1145/3631939
dc.relation.doi https://doi.org/10.1145/3576924
dc.relation.doi https://doi.org/10.1145/3485447.3511955
dc.relation.doi https://doi.org/10.1145/3366424.3383547
dc.relation.url https://ceur-ws.org/Vol-2993/paper-27.pdf
dc.description.version publishedVersion eng
tib.accessRights frei zug�nglich eng


Die Publikation erscheint in Sammlung(en):

Zur Kurzanzeige

 

Suche im Repositorium


Durchblättern

Mein Nutzer/innenkonto

Nutzungsstatistiken