Published on Tue Sep 03 2019

Attributed Rhetorical Structure Grammar for Domain Text Summarization

Ruqian Lu, Shengluan Hou, Chuanqing Wang, Yu Huang, Chaoqun Fei, Songmao Zhang

This paper presents a new approach of automatic text summarization. It combines domain oriented text analysis (DoTA) and rhetorical structure theory (RST) in a grammar form: the attributed rhetorical structure grammar (ARSG)

0
0
0
Abstract

This paper presents a new approach of automatic text summarization which combines domain oriented text analysis (DoTA) and rhetorical structure theory (RST) in a grammar form: the attributed rhetorical structure grammar (ARSG), where the non-terminal symbols are domain keywords, called domain relations, while the rhetorical relations serve as attributes. We developed machine learning algorithms for learning such a grammar from a corpus of sample domain texts, as well as parsing algorithms for the learned grammar, together with adjustable text summarization algorithms for generating domain specific summaries. Our practical experiments have shown that with support of domain knowledge the drawback of missing very large training data set can be effectively compensated. We have also shown that the knowledge based approach may be made more powerful by introducing grammar parsing and RST as inference engine. For checking the feasibility of model transfer, we introduced a technique for mapping a grammar from one domain to others with acceptable cost. We have also made a comprehensive comparison of our approach with some others.

Mon Apr 01 2019
NLP
Automatic text summarization: What has been done and what has to be done
Automatic text summarization (ATS) is not a new research field. It was known since the 50s. Researchers have been active to find the perfect summarization method. We will present some problems and limits which prevent works to move forward.
0
0
0
Sun Jul 01 2018
NLP
Modeling, comprehending and summarizing textual content by graphs
Automatic Text Summarization strategies have been successfully employed to digest text collections and extract its essential content. Usually, summaries are generated using textual corpora that belongs to the same domain area where the summary will be used. One possible alternative is to generate asummary from a different domain.
0
0
0
Mon Oct 14 2019
Artificial Intelligence
Knowledge-guided Unsupervised Rhetorical Parsing for Text Summarization
Automatic text summarization (ATS) has recently achieved impressive performance thanks to advances in deep learning. This paper presents an unsupervised approach that combines rhetorical structure Theory, deep neural model and domain knowledge concern for ATS.
0
0
0
Sat Jul 04 2009
NLP
Induction of Word and Phrase Alignments for Automatic Document Summarization
This paper describes experiments we have carried out to analyze the ability of humans to perform such alignments. Our model for the task is based on an extension of the standard hidden Markov model. We believe that future progress in automatic summarization will be driven by the development of more sophisticated, linguistically informed models.
0
0
0
Thu Dec 09 2010
Artificial Intelligence
MUDOS-NG: Multi-document Summaries Using N-gram Graphs (Tech Report)
This report describes the MUDOS-NG summarization system, which applies a set of language-independent and generic methods. The proposed methods are mostly combinations of simple operators on a generic character n-gram graph representation of texts.
0
0
0
Sun Mar 01 2020
NLP
StructSum: Summarization via Structured Representations
abstractive text summarization aims at compressing information into a condensed summary. Despite advances in modeling techniques, abstractive summarization models still suffer from several key challenges. We propose a framework based on document-level structure induction for summarization to address these challenges.
0
0
0