Learning-Free Unsupervised Extractive Summarization Model

Myeongjun Jang, Pilsung Kang

Research output: Contribution to journalArticlepeer-review

16 Citations (Scopus)

Abstract

Text summarization is an information condensation technique that abbreviates a source document to a few representative sentences with the intention to create a coherent summary containing relevant information of source corpora. This promising subject has been rapidly developed since the advent of deep learning. However, summarization models based on deep neural network have several critical shortcomings. First, a large amount of labeled training data is necessary. This problem is standard for low-resource languages in which publicly available labeled data do not exist. In addition, a significant amount of computational ability is required to train neural models with enormous network parameters. In this study, we propose a model called Learning Free Integer Programming Summarizer (LFIP-SUM), which is an unsupervised extractive summarization model. The advantage of our approach is that parameter training is unnecessary because the model does not require any labeled training data. To achieve this, we formulate an integer programming problem based on pre-trained sentence embedding vectors. We also use principal component analysis to automatically determine the number of sentences to be extracted and to evaluate the importance of each sentence. Experimental results demonstrate that the proposed model exhibits generally acceptable performance compared with deep learning summarization models although it does not learn any parameters during the model construction process.

Original languageEnglish
Article number9321308
Pages (from-to)14358-14368
Number of pages11
JournalIEEE Access
Volume9
DOIs
Publication statusPublished - 2021

Bibliographical note

Funding Information:
This work was supported in part by the National Research Foundation of Korea (NRF) Grants Funded by the Korean Government (MSIT) under Grant NRF-2019R1F1A1060338 and in part by the Korea Institute for Advancement of Technology (KIAT) Grant Funded by the Korean Government (MOTIE) (The Competency Development Program for Industry Specialist) under Grant P0008691.

Publisher Copyright:
© 2013 IEEE.

Keywords

  • Text summarization
  • integer linear programming
  • natural language processing
  • sentence representation vector

ASJC Scopus subject areas

  • General Computer Science
  • General Materials Science
  • General Engineering

Fingerprint

Dive into the research topics of 'Learning-Free Unsupervised Extractive Summarization Model'. Together they form a unique fingerprint.

Cite this