Using tweets to help sentence compression for news highlights generation
We explore using relevant tweets of a given news article to help sentence compression for generating compressive news highlights. We extend an unsupervised dependency-tree based sentence compression approach by incorporating tweet information to weight the tree edge in terms of informativeness and s...
Saved in:
Main Authors: | , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2015
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/4575 https://ink.library.smu.edu.sg/context/sis_research/article/5578/viewcontent/P15_2009.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-5578 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-55782019-12-26T08:12:16Z Using tweets to help sentence compression for news highlights generation WEI, Zhongyu LIU, Yang LI, Chen GAO, Wei We explore using relevant tweets of a given news article to help sentence compression for generating compressive news highlights. We extend an unsupervised dependency-tree based sentence compression approach by incorporating tweet information to weight the tree edge in terms of informativeness and syntactic importance. The experimental results on a public corpus that contains both news articles and relevant tweets show that our proposed tweets guided sentence compression method can improve the summarization performance significantly compared to the baseline generic sentence compression method. 2015-07-01T07:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/4575 info:doi/10.3115/v1/P15-2009 https://ink.library.smu.edu.sg/context/sis_research/article/5578/viewcontent/P15_2009.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Databases and Information Systems |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Databases and Information Systems |
spellingShingle |
Databases and Information Systems WEI, Zhongyu LIU, Yang LI, Chen GAO, Wei Using tweets to help sentence compression for news highlights generation |
description |
We explore using relevant tweets of a given news article to help sentence compression for generating compressive news highlights. We extend an unsupervised dependency-tree based sentence compression approach by incorporating tweet information to weight the tree edge in terms of informativeness and syntactic importance. The experimental results on a public corpus that contains both news articles and relevant tweets show that our proposed tweets guided sentence compression method can improve the summarization performance significantly compared to the baseline generic sentence compression method. |
format |
text |
author |
WEI, Zhongyu LIU, Yang LI, Chen GAO, Wei |
author_facet |
WEI, Zhongyu LIU, Yang LI, Chen GAO, Wei |
author_sort |
WEI, Zhongyu |
title |
Using tweets to help sentence compression for news highlights generation |
title_short |
Using tweets to help sentence compression for news highlights generation |
title_full |
Using tweets to help sentence compression for news highlights generation |
title_fullStr |
Using tweets to help sentence compression for news highlights generation |
title_full_unstemmed |
Using tweets to help sentence compression for news highlights generation |
title_sort |
using tweets to help sentence compression for news highlights generation |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2015 |
url |
https://ink.library.smu.edu.sg/sis_research/4575 https://ink.library.smu.edu.sg/context/sis_research/article/5578/viewcontent/P15_2009.pdf |
_version_ |
1770574918548717568 |