Understanding omitted facts in transformer-based abstractive summarization

dc.contributor.authorPanawenna, PH
dc.contributor.authorWickramanayake, S
dc.date.accessioned2026-02-11T05:29:54Z
dc.date.issued2024
dc.description.abstractText summarization is a natural language processing task that generates concise document summaries. It can be extractive or abstractive. The former extracts pieces of the document, while the latter generates new concise sentences after identifying the critical information from the input text. Abstractive Summarization (AS) more closely represents how a human would summarize and is used in multiple missioncritical downstream tasks in domains such as law and finance. However, the existing state-of-the-art AS models are based on black-box deep learning models such as Transformers. Hence, the users of such systems cannot understand why some facts from the document have been included in the summary while some have been omitted. This paper proposes an algorithm to explain which facts have been omitted and why in Transformerbased AS. We leverage the Cross-Attention (CA) in transformers to identify words in the input passage with minimum influence in generating the summary. These identified words are then given to a Large Language Model along with the input passage and the generated summary to explain the omitted facts and the reasons for omissions. The experimental results using the state-of-the-art AS model show that CA can help provide valuable explanations for the model’s fact selection process.
dc.identifier.conferenceMoratuwa Engineering Research Conference 2024
dc.identifier.departmentEngineering Research Unit, University of Moratuwa
dc.identifier.emailpasadie.23@cse.mrt.ac.lk
dc.identifier.emailsandarekaw@cse.mrt.ac.lk
dc.identifier.facultyEngineering
dc.identifier.isbn979-8-3315-2904-8
dc.identifier.pgnospp. 624-629
dc.identifier.placeMoratuwa, Sri Lanka
dc.identifier.proceedingProceedings of Moratuwa Engineering Research Conference 2024
dc.identifier.urihttps://dl.lib.uom.lk/handle/123/24842
dc.language.isoen
dc.publisherIEEE
dc.subjectFact Selection
dc.subjectAbstractive Summarization
dc.subjectCross Attention
dc.subjectTransformers
dc.subjectLarge Language Models
dc.titleUnderstanding omitted facts in transformer-based abstractive summarization
dc.typeConference-Full-text

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
1571021473.pdf
Size:
2.06 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
1.71 KB
Format:
Item-specific license agreed upon to submission
Description:

Collections