Extractive Explanations for Interpretable Text Ranking

Publikation: Beitrag in FachzeitschriftArtikelForschungPeer-Review

Autoren

  • Jurek Leonhardt
  • Koustav Rudra
  • Avishek Anand

Organisationseinheiten

Externe Organisationen

  • Delft University of Technology
  • Indian Institute of Technology Dhanbad (IIT(ISM))
Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Aufsatznummer88
Seitenumfang31
FachzeitschriftACM Transactions on Information Systems
Jahrgang41
Ausgabenummer4
Frühes Online-Datum16 Dez. 2022
PublikationsstatusVeröffentlicht - 23 März 2023

Abstract

Neural document ranking models perform impressively well due to superior language understanding gained from pre-Training tasks. However, due to their complexity and large number of parameters these (typically transformer-based) models are often non-interpretable in that ranking decisions can not be clearly attributed to specific parts of the input documents.In this article, we propose ranking models that are inherently interpretable by generating explanations as a by-product of the prediction decision. We introduce the Select-And-Rank paradigm for document ranking, where we first output an explanation as a selected subset of sentences in a document. Thereafter, we solely use the explanation or selection to make the prediction, making explanations first-class citizens in the ranking process. Technically, we treat sentence selection as a latent variable trained jointly with the ranker from the final output. To that end, we propose an end-To-end training technique for Select-And-Rank models utilizing reparameterizable subset sampling using the Gumbel-max trick.We conduct extensive experiments to demonstrate that our approach is competitive to state-of-The-Art methods. Our approach is broadly applicable to numerous ranking tasks and furthers the goal of building models that are interpretable by design. Finally, we present real-world applications that benefit from our sentence selection method.

Zitieren

Extractive Explanations for Interpretable Text Ranking. / Leonhardt, Jurek; Rudra, Koustav; Anand, Avishek.
in: ACM Transactions on Information Systems, Jahrgang 41, Nr. 4, 88, 23.03.2023.

Publikation: Beitrag in FachzeitschriftArtikelForschungPeer-Review

Leonhardt J, Rudra K, Anand A. Extractive Explanations for Interpretable Text Ranking. ACM Transactions on Information Systems. 2023 Mär 23;41(4):88. Epub 2022 Dez 16. doi: 10.48550/arXiv.2106.12460, 10.1145/3576924
Leonhardt, Jurek ; Rudra, Koustav ; Anand, Avishek. / Extractive Explanations for Interpretable Text Ranking. in: ACM Transactions on Information Systems. 2023 ; Jahrgang 41, Nr. 4.
Download
@article{22657f56392f4488a585d2644805e9a6,
title = "Extractive Explanations for Interpretable Text Ranking",
abstract = "Neural document ranking models perform impressively well due to superior language understanding gained from pre-Training tasks. However, due to their complexity and large number of parameters these (typically transformer-based) models are often non-interpretable in that ranking decisions can not be clearly attributed to specific parts of the input documents.In this article, we propose ranking models that are inherently interpretable by generating explanations as a by-product of the prediction decision. We introduce the Select-And-Rank paradigm for document ranking, where we first output an explanation as a selected subset of sentences in a document. Thereafter, we solely use the explanation or selection to make the prediction, making explanations first-class citizens in the ranking process. Technically, we treat sentence selection as a latent variable trained jointly with the ranker from the final output. To that end, we propose an end-To-end training technique for Select-And-Rank models utilizing reparameterizable subset sampling using the Gumbel-max trick.We conduct extensive experiments to demonstrate that our approach is competitive to state-of-The-Art methods. Our approach is broadly applicable to numerous ranking tasks and furthers the goal of building models that are interpretable by design. Finally, we present real-world applications that benefit from our sentence selection method.",
keywords = "fact checking, information retrieval, interpretability, Ranking, sentence selection",
author = "Jurek Leonhardt and Koustav Rudra and Avishek Anand",
note = "Funding Information: This work is supported by the European Union – Horizon 2020 Program under the scheme “INFRAIA-01-2018-2019 – Integrating Activities for Advanced Communities”, Grant Agreement n.871042, “SoBigData++: European Integrated Infrastructure for Social Mining and Big Data Analytics” ( http://www.sobigdata.eu ). Furthermore, this work is supported in part by the Science and Engineering Research Board, Department of Science and Technology, Government of India, under Project SRG/2022/001548. Koustav Rudra is a recipient of the DST-INSPIRE Faculty Fellowship [DST/INSPIRE/04/2021/003055] in the year 2021 under Engineering Sciences. ",
year = "2023",
month = mar,
day = "23",
doi = "10.48550/arXiv.2106.12460",
language = "English",
volume = "41",
journal = "ACM Transactions on Information Systems",
issn = "1046-8188",
publisher = "Association for Computing Machinery (ACM)",
number = "4",

}

Download

TY - JOUR

T1 - Extractive Explanations for Interpretable Text Ranking

AU - Leonhardt, Jurek

AU - Rudra, Koustav

AU - Anand, Avishek

N1 - Funding Information: This work is supported by the European Union – Horizon 2020 Program under the scheme “INFRAIA-01-2018-2019 – Integrating Activities for Advanced Communities”, Grant Agreement n.871042, “SoBigData++: European Integrated Infrastructure for Social Mining and Big Data Analytics” ( http://www.sobigdata.eu ). Furthermore, this work is supported in part by the Science and Engineering Research Board, Department of Science and Technology, Government of India, under Project SRG/2022/001548. Koustav Rudra is a recipient of the DST-INSPIRE Faculty Fellowship [DST/INSPIRE/04/2021/003055] in the year 2021 under Engineering Sciences.

PY - 2023/3/23

Y1 - 2023/3/23

N2 - Neural document ranking models perform impressively well due to superior language understanding gained from pre-Training tasks. However, due to their complexity and large number of parameters these (typically transformer-based) models are often non-interpretable in that ranking decisions can not be clearly attributed to specific parts of the input documents.In this article, we propose ranking models that are inherently interpretable by generating explanations as a by-product of the prediction decision. We introduce the Select-And-Rank paradigm for document ranking, where we first output an explanation as a selected subset of sentences in a document. Thereafter, we solely use the explanation or selection to make the prediction, making explanations first-class citizens in the ranking process. Technically, we treat sentence selection as a latent variable trained jointly with the ranker from the final output. To that end, we propose an end-To-end training technique for Select-And-Rank models utilizing reparameterizable subset sampling using the Gumbel-max trick.We conduct extensive experiments to demonstrate that our approach is competitive to state-of-The-Art methods. Our approach is broadly applicable to numerous ranking tasks and furthers the goal of building models that are interpretable by design. Finally, we present real-world applications that benefit from our sentence selection method.

AB - Neural document ranking models perform impressively well due to superior language understanding gained from pre-Training tasks. However, due to their complexity and large number of parameters these (typically transformer-based) models are often non-interpretable in that ranking decisions can not be clearly attributed to specific parts of the input documents.In this article, we propose ranking models that are inherently interpretable by generating explanations as a by-product of the prediction decision. We introduce the Select-And-Rank paradigm for document ranking, where we first output an explanation as a selected subset of sentences in a document. Thereafter, we solely use the explanation or selection to make the prediction, making explanations first-class citizens in the ranking process. Technically, we treat sentence selection as a latent variable trained jointly with the ranker from the final output. To that end, we propose an end-To-end training technique for Select-And-Rank models utilizing reparameterizable subset sampling using the Gumbel-max trick.We conduct extensive experiments to demonstrate that our approach is competitive to state-of-The-Art methods. Our approach is broadly applicable to numerous ranking tasks and furthers the goal of building models that are interpretable by design. Finally, we present real-world applications that benefit from our sentence selection method.

KW - fact checking

KW - information retrieval

KW - interpretability

KW - Ranking

KW - sentence selection

UR - http://www.scopus.com/inward/record.url?scp=85172691686&partnerID=8YFLogxK

U2 - 10.48550/arXiv.2106.12460

DO - 10.48550/arXiv.2106.12460

M3 - Article

AN - SCOPUS:85172691686

VL - 41

JO - ACM Transactions on Information Systems

JF - ACM Transactions on Information Systems

SN - 1046-8188

IS - 4

M1 - 88

ER -