Kavita Ganesan, ChengXiang Zhai, Jiawei Han University of Urbana Champaign

Similar documents
The ACL Anthology Network Corpus. University of Michigan

Using Citations to Generate Surveys of Scientific Paradigms

Determining sentiment in citation text and analyzing its impact on the proposed ranking index

LAMP-TR-157 August 2011 CS-TR-4988 UMIACS-TR CITATION HANDLING FOR IMPROVED SUMMMARIZATION OF SCIENTIFIC DOCUMENTS

Humor in Collective Discourse: Unsupervised Funniness Detection in the New Yorker Cartoon Caption Contest

Bi-Modal Music Emotion Recognition: Novel Lyrical Features and Dataset

arxiv: v1 [cs.cl] 26 Jun 2015

HumorHawk at SemEval-2017 Task 6: Mixing Meaning and Sound for Humor Recognition

First Stage of an Automated Content-Based Citation Analysis Study: Detection of Citation Sentences 1

Chinese Word Sense Disambiguation with PageRank and HowNet

Sarcasm Detection in Text: Design Document

Citation Analysis, Centrality, and the ACL Anthology

Melody classification using patterns

An Introduction to Deep Image Aesthetics

Projektseminar: Sentimentanalyse Dozenten: Michael Wiegand und Marc Schulder

Sentiment Aggregation using ConceptNet Ontology

Subjective Similarity of Music: Data Collection for Individuality Analysis

A Multi-Layered Annotated Corpus of Scientific Papers

National University of Singapore, Singapore,

DETECTION OF SLOW-MOTION REPLAY SEGMENTS IN SPORTS VIDEO FOR HIGHLIGHTS GENERATION

Dimensions of Argumentation in Social Media

Estimating Number of Citations Using Author Reputation

UWaterloo at SemEval-2017 Task 7: Locating the Pun Using Syntactic Characteristics and Corpus-based Metrics

The ACL Anthology Reference Corpus: a reference dataset for bibliographic research

Sentiment Analysis. Andrea Esuli

World Journal of Engineering Research and Technology WJERT

Your Sentiment Precedes You: Using an author s historical tweets to predict sarcasm

Introduction to Sentiment Analysis. Text Analytics - Andrea Esuli

Author-Specific Sentiment Aggregation for Polarity Prediction of Reviews

A combination of opinion mining and social network techniques for discussion analysis

A Discriminative Approach to Topic-based Citation Recommendation

THE ACL ANTHOLOGY NETWORK CORPUS

The ACL anthology network corpus

Identifying functions of citations with CiTalO

Summarizing Long First-Person Videos

Citation Resolution: A method for evaluating context-based citation recommendation systems

A Generic Semantic-based Framework for Cross-domain Recommendation

Generating Fine-Grained Reviews of Songs From Album Reviews

Using the Annotated Bibliography as a Resource for Indicative Summarization

Deriving the Impact of Scientific Publications by Mining Citation Opinion Terms

Who Speaks for Whom? Towards Analyzing Opinions in News Editorials

Singer Traits Identification using Deep Neural Network

Exploiting Cross-Document Relations for Multi-document Evolving Summarization

Bibliometric analysis of the field of folksonomy research

Affect-based Features for Humour Recognition

Humor Recognition and Humor Anchor Extraction

ICWSM A Great Catchy Name: Semi-Supervised Recognition of Sarcastic Sentences in Online Product Reviews

Less is More: Picking Informative Frames for Video Captioning

An Impact Analysis of Features in a Classification Approach to Irony Detection in Product Reviews

Irony and Sarcasm: Corpus Generation and Analysis Using Crowdsourcing

Creating Mindmaps of Documents

Introduction. Format DOCUMENT GUIDELINES. Table of Contents

A New Scheme for Citation Classification based on Convolutional Neural Networks

Submission guidelines for authors and editors

Using assessment and research to promote learning. Thakur B. Karkee, Ph. D. Measurement Incorporated. Kevin Fatica CTB/McGraw-Hill

WHAT MAKES FOR A HIT POP SONG? WHAT MAKES FOR A POP SONG?

Lyric-Based Music Mood Recognition

Music Radar: A Web-based Query by Humming System

Exploring and Exploiting the Limited Utility of Captions in Recognizing Intention in Information Graphics

Recommending Citations: Translating Papers into References

SWITCHED INFINITY: SUPPORTING AN INFINITE HD LINEUP WITH SDV

arxiv: v1 [cs.cl] 3 May 2018

Introduction to Sentiment Analysis

ABSTRACT CITATION HANDLING: PROCESSING CITATION TEXTS IN SCIENTIFIC DOCUMENTS. Michael Alan Whidby Master of Science, 2012

Idiom Savant at Semeval-2017 Task 7: Detection and Interpretation of English Puns

Computational Modelling of Harmony

Reducing False Positives in Video Shot Detection

IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 12, NO. 7, NOVEMBER

Toward Evaluation Techniques for Music Similarity

A Music Retrieval System Using Melody and Lyric

Advanced Data Structures and Algorithms

METHOD TO DETECT GTTM LOCAL GROUPING BOUNDARIES BASED ON CLUSTERING AND STATISTICAL LEARNING

IMPROVED MELODIC SEQUENCE MATCHING FOR QUERY BASED SEARCHING IN INDIAN CLASSICAL MUSIC

Indexing local features. Wed March 30 Prof. Kristen Grauman UT-Austin

Figures in Scientific Open Access Publications

Survey of Hyponym Relation Extraction from Web Database Using Motif Patterns with Feature Extraction Model

For general information, see the ECP website online handbook:

Enhancing Music Maps

Presentation Overview

Humor recognition using deep learning

A repetition-based framework for lyric alignment in popular songs

BIG SYNTHETIC DATA WITH MUSKETEER

Running head: BASIC APA FORMATTING AND STYLE GUIDE 1

Perceptual Evaluation of Automatically Extracted Musical Motives

Combination of Audio & Lyrics Features for Genre Classication in Digital Audio Collections

Perceptual Coding: Hype or Hope?

An Efficient Low Bit-Rate Video-Coding Algorithm Focusing on Moving Regions

Conceptions and Context as a Fundament for the Representation of Knowledge Artifacts

Personal Narrative STUDENT SELF-ASSESSMENT. Ideas YES NO Do I have a suitable topic? Do I maintain a clear focus?

Large scale Visual Sentiment Ontology and Detectors Using Adjective Noun Pairs

Scalable Semantic Parsing with Partial Ontologies ACL 2015

Automatic Detection of Sarcasm in BBS Posts Based on Sarcasm Classification

Department of Communication Standards for Acceptable Submissions

WordFinder. Verginica Barbu Mititelu RACAI / 13 Calea 13 Septembrie, Bucharest, Romania

Automatically Extracting Word Relationships as Templates for Pun Generation

Browsing News and Talk Video on a Consumer Electronics Platform Using Face Detection

FunTube: Annotating Funniness in YouTube Comments

Performance of a Low-Complexity Turbo Decoder and its Implementation on a Low-Cost, 16-Bit Fixed-Point DSP

DAY 1. Intelligent Audio Systems: A review of the foundations and applications of semantic audio analysis and music information retrieval

Essential Aspects of Academic Practice (EAAP)

Transcription:

Kavita Ganesan, ChengXiang Zhai, Jiawei Han University of Illinois @ Urbana Champaign

Opinion Summary for ipod Existing methods: Generate structured ratings for an entity [Lu et al., 2009; Lerman et al., 2009;..]

Opinion Summary for ipod To know more: read many redundant sentences structured format useful, but not enough!

Summarize the major opinions What are the major complaints/praise in an aspect? Concise Easily digestible Viewable on smaller screens Readable Easily understood

The iphone s battery lasts long and is cheap but it s bulky. Important information summarized Concise Readable

Widely studied for years [Radev et al.2000; Erkan & Radev, 2004; Mihalcea & Tarau, 2004 ] But, not suitable for: Generating concise summaries Summarizing highly redundant text Problems o Bias: with limit on summary size o selected sentence may have missed critical info o Verbose: May contain irrelevant information o not suitable for smaller devices

Widely studied for years [Radev et al.2000; Erkan & Radev, 2004; Mihalcea & Tarau, 2004 ] But, not suitable for: Generating concise summaries Summarizing highly redundant text Problems o Bias: with limit on summary size Extractive o selected sentence may have missed critical info o Verbose: May contain irrelevant information Abstractive o not suitable for smaller devices

Existing methods: Some methods require manual effort [DeJong1982] [Radev and McKeown1998] [Finley and Harabagiu2002] Need to define templates to be filled Some methods rely heavily on NL understanding [Saggion and Lapalme2002] [Jing and McKeown2000] Domain dependent Impractical high computational costs

`Shallow abstractive summarizer Generates concise summaries using: existing text inherent redundancies Uses minimal external knowledge lightweight

Input Set of sentences: Topic specific (ex. battery life of ipod) POS annotated

Set of sentences: Topic specific (ex. battery life of ipod) POS annotated Input my too phone calls drop frequently with the iphone is a Step 1: Generate graph representation of sentences (Opinosis-Graph). great device

Set of sentences: Topic specific (ex. battery life of ipod) POS annotated Input too my phone calls drop frequently the iphone is a. great Step 1: Generate graph representation of sentences (Opinosis-Graph) device with 3.2 2.5 calls drop great device frequently candidate sum1 candidate sum2 Step 2: Find promising paths (candidate summaries) & score these candidates

Input Set of sentences: The iphone is a great Topic specific (ex. battery life of ipod) device, but calls drop POS annotated frequently. too Step 3: Select top scoring candidates as final summary my phone calls drop frequently with the iphone is a. great Step 1: Generate graph representation of sentences (Opinosis-Graph) device 3.2 2.5 calls drop great device frequently candidate sum1 candidate sum2 Step 2: Find promising paths (candidate summaries) & score these candidates

Assume: 2 sentences about call quality of iphone 1. My phone calls drop frequently with the iphone. 2. Great device, but the calls drop too frequently. Opinosis-Graph is empty

1. My phone calls drop frequently with the iphone.

1. My phone calls drop frequently with the iphone. my 1:1 unique(word + POS) = node SID PID Positional Reference Information

1. My phone calls drop frequently with the iphone. co-occurrence my 1:1 phone 1:2 SID PID Positional Reference Information

1. My phone calls drop frequently with the iphone. my 1:1 phone 1:2 calls 1:3

1. My phone calls drop frequently with the iphone. my phone calls drop 1:1 1:2 1:3 1:4

1. My phone calls drop frequently with the iphone. my phone calls drop frequently 1:1 1:2 1:3 1:4 1:5

1. My phone calls drop frequently with the iphone. my phone calls drop frequently 1:1 1:2 1:3 1:4 1:5 with 1:6

1. My phone calls drop frequently with the iphone. my phone calls drop frequently 1:1 1:2 1:3 1:4 1:5 with the 1:6 1:7

1. My phone calls drop frequently with the iphone. my phone calls drop frequently 1:1 1:2 1:3 1:4 1:5 with the iphone 1:6 1:7 1:8

1. My phone calls drop frequently with the iphone.. my phone calls drop frequently 1:9 1:1 1:2 1:3 1:4 1:5 with the iphone 1:6 1:7 1:8

1. My phone calls drop frequently with the iphone.. my phone calls drop frequently 1:9 1:1 1:2 1:3 1:4 1:5 with the iphone 1:6 1:7 1:8

2. Great device, but the calls drop too frequently.. my phone calls drop frequently 1:9 1:1 1:2 1:3 1:4 1:5 with the iphone 1:6 1:7 1:8

2. Great device, but the calls drop too frequently.. my phone calls drop frequently 1:9 1:1 1:2 1:3 1:4 1:5 with the iphone 1:6 1:7 1:8 great device, but 2:1 2:2 2:3 2:4

2. Great device, but the calls drop too frequently.. my phone calls drop frequently 1:9 1:1 1:2 1:3 1:4 1:5 with the iphone 1:6 1:7 1:8 great device, but 2:1 2:2 2:3 2:4

2. Great device, but the calls drop too frequently.. my phone calls drop frequently 1:9 1:1 1:2 1:3 1:4 1:5 with the iphone 1:6 1:7, 2:5 1:8 great device, but 2:1 2:2 2:3 2:4

2. Great device, but the calls drop too frequently.. my phone calls drop frequently 1:9 1:1 1:2 1:3, 2:6 1:4 1:5 with the iphone 1:6 1:7, 2:5 1:8 great device, but 2:1 2:2 2:3 2:4

2. Great device, but the calls drop too frequently.. my phone calls drop frequently 1:9 1:1 1:2 1:3, 2:6 1:4, 2:7 1:5 with the iphone 1:6 1:7, 2:5 1:8 great device, but 2:1 2:2 2:3 2:4

2. Great device, but the calls drop too frequently. too 2:8. my phone calls drop frequently 1:9 1:1 1:2 1:3, 2:6 1:4, 2:7 1:5 with the iphone 1:6 1:7, 2:5 1:8 great device, but 2:1 2:2 2:3 2:4

2. Great device, but the calls drop too frequently. too 2:8. my phone calls drop frequently 1:9 1:1 1:2 1:3, 2:6 1:4, 2:7 1:5, 2:9 with the iphone 1:6 1:7, 2:5 1:8 great device, but 2:1 2:2 2:3 2:4

2. Great device, but the calls drop too frequently. too 2:8. my phone calls drop frequently 1:9, 2:10 1:1 1:2 1:3, 2:6 1:4, 2:7 1:5, 2:9 with the iphone 1:6 1:7, 2:5 1:8 great device, but 2:1 2:2 2:3 2:4

Graph is now ready for Step 2! too 2:8. my phone calls drop frequently 1:9, 2:10 1:1 1:2 1:3, 2:6 1:4, 2:7 1:5, 2:9 with the iphone 1:6 1:7, 2:5 1:8 great device, but 2:1 2:2 2:3 2:4

Naturally captures redundancies too 2:8. my phone calls drop frequently 1:9, 2:10 1:1 1:2 1:3, 2:6 1:4, 2:7 1:5, 2:9 with the iphone 1:6 1:7, 2:5 1:8 great device, but 2:1 2:2 2:3 2:4

Naturally captures redundancies too 2:8. my phone calls drop frequently 1:9, 2:10 1:1 1:2 1:3, 2:6 1:4, 2:7 1:5, 2:9 with the iphone 1:6 1:7, 2:5 1:8 Path shared great by device 2 sentences, but naturally captured 2:1 2:2 by 2:3 nodes 2:4

Naturally captures redundancies too 2:8. my phone calls drop frequently 1:9, 2:10 1:1 1:2 1:3, 2:6 1:4, 2:7 1:5, 2:9 with the iphone 1:6 1:7, 2:5 1:8 Easily discover great redundancies device, but for high confidence 2:1 2:2summaries 2:3 2:4

Captures gapped subsequences 1. My phone calls drop frequently with the iphone. 2. Great device, but the calls drop too frequently.

1. My phone calls drop frequently with the iphone. 2. Great device, but the calls drop too frequently. Captures gapped subsequences too 2:8. my 1:1 phone calls drop frequently 1:2 1:3, 2:6 1:4, 2:7 1:5, 2:9 great device 1:9, 2:10 with the iphone, 1:6 1:7, 2:5 1:8 Gap between words = 2 but 2:1 2:2 2:3 2:4

Captures gapped subsequences too 2:8. my phone calls drop frequently 1:1 1:2 1:3, 2:6 1:4, 2:7 1:5, 2:9 great device discovery 2:1 of new 2:2 sentences 2:3 2:4, but 1:9, 2:10 with the iphone Gapped subsequences 1:6 allow: 1:7, 2:5 redundancy enforcements 1:8

Captures collapsible structures 1. Calls drop frequently with the iphone 2. Calls drop frequently with the Black Berry calls drop frequently with the iphone black berry Calls drop frequently with the iphone and Black Berry

Captures collapsible structures 1. Calls drop frequently with the iphone 2. Calls drop frequently with the Black Berry calls drop frequently with the iphone black berry - Can easily be discovered using OG - Ideal for collapse & compression

Repeatedly search the Opinosis- Graph for a Valid Path

Set of connected nodes Has a Valid Start Node (VSN) Natural starting point of a sentence Opinosis uses avg. positional information Has a Valid End Node (VEN) Point that completes a sentence Opinosis uses punctuations & conjunctions

, calls drop frequently with the iphone. VSN VEN

, calls drop frequently with the iphone. VSN VEN Candidate summary

, calls drop frequently with the iphone. VSN VEN Pool of candidate summaries

Some paths are collapsible Identify such paths through a collapsible node Treat linking verbs (e.g. is, are) as collapsible nodes Linking verbs have hub-like properties Commonly used in opinion text

linking verb = collapsible node the screen is very clear big anchor - Common structure -High redundancy path collapsed candidates (CC) - Subgraphs to be merged

the screen is very clear big collapse + merge the screen is very clear and big anchor CC1 CC2

CC after linking verbs: concatenate using commas The screen is very clear, bright, big CC1 CC2 CC3 Better readability: The screen is very clear, bright and big Find last connector using hints from OG

Type 1: High confidence summaries Select candidates with high redundancy # of sentences sharing same path controlled by gap threshold, σgap Type 2: + Good coverage Select longer candidates redundancy * length of candidate paths Favor longer but redundant candidates

Gaps vary between sentences sharing nodes Candidate X w1 w2 w3 Sentence 1 1 2 Sentence 2 4 4 Sentence X m n gap gap

σgap enforces maximum allowed gap between two adjacent nodes Candidate X Sentence 1 1 Sentence 2 4 Sentence n w1 < σgap > σgap m -Lower risk of ill-formed sentences gap w2 -Avoids over-estimation of redundancy

After candidate scoring: Select top 2 scoring candidates Most dissimilar candidates

User Reviews: Hotels: Tripadvisor.com Products: Amazon.com Cars: Edmunds.com

Reviews Edmunds Tripadvisor Amazon Topic 1 1. sentence 1.. 2. sentence 2.. 3. sentence 3.. 4. sentence 4.. Topic 2 1. sentence 1.. 2. sentence 2.. 3. sentence 3.. 4. sentence 4.. Topic 51 1. sentence 1.. 2. sentence 2.. 3. sentence 3.. 4. sentence 4.. ~100 unordered, topic-related, sentences review document 1 review document 2 review document 51 summarize summarize summarize

Human composed summaries Concise (<25 words) Focus on summarizing major opinions ~4 human summaries per topic

Hard to find general abstractive summarizer Use MEAD - Extractive based method [Radev et al.2000] Select 2 sentences as the summary

ROUGE (rouge-1, rouge-2, rouge-su4) Standard measure for summarization tasks Readability Test Measures: How different Opinosis summaries are compared to human composed summaries?

Estimate: How much one summary writer agrees with the rest

ROUGE-1 ROUGE-SU4 ROUGE Scores Precision 0.34 Recall 0.32 F-score 0.31 Precision 0.16 Recall 0.13 F-score 0.11 Human summaries - semantically similar. Slight difference in word usage.

Highest recall Lowest precision ROUGE-1 ROUGE-SU4 0.4932 0.4482 0.3184 0.1293 0.2831 0.0851 0.2316 0.3434 0.3088 0.3271 0.0916 0.1515 HUMAN (17 words) OPINOSISbest (15 words) MEAD (75 words) HUMAN (17 words) OPINOSISbest (15 words) MEAD (75 words) ROUGE Recall Much longer sentences ROUGE Precision

Highest recall Lowest precision ROUGE-1 ROUGE-SU4 0.4932 0.4482 0.3184 0.1293 0.2831 0.0851 0.2316 0.3434 0.3088 0.3271 0.0916 0.1515 HUMAN (17 words) OPINOSISbest (15 words) MEAD (75 words) HUMAN (17 words) OPINOSISbest (15 words) MEAD (75 words) Overall: Baseline does not do well ROUGE Recall Much longer sentences ROUGE ROUGE Precision Precision in generating concise summaries.

similar similar ROUGE-1 ROUGE-SU4 0.4932 0.4482 0.3184 0.1293 0.2831 0.0851 0.2316 0.3434 0.3088 0.3271 0.0916 0.1515 HUMAN (17 words) OPINOSISbest (15 words) MEAD (75 words) HUMAN (17 words) OPINOSISbest (15 words) MEAD (75 words) ROUGE Recall ROUGE Precision

similar similar ROUGE-1 ROUGE-SU4 0.4932 0.4482 0.3184 0.1293 0.2831 0.0851 0.2316 0.3434 0.3088 0.3271 0.0916 0.1515 HUMAN (17 words) OPINOSISbest (15 words) ROUGE Recall MEAD (75 words) HUMAN (17 words) OPINOSISbest (15 words) Performance of Opinosis is reasonable ROUGE Precision similar to Human performance MEAD (75 words)

ROUGE-1 (f-score) 0.330 0.310 0.290 wt_loglen 0.270 0.250 1 2 3 4 5 σgap

ROUGE-1 (f-score) 0.330 0.310 0.290 wt_loglen 0.270 0.250 small σgap words in summary 1 2 3 4 5 close together in original text σgap

ROUGE-1 (f-score) 0.330 0.310 0.290 0.270 Lowest performance Strict adjacency disallows redundancies to be captured wt_loglen 0.250 1 2 3 4 5 σgap

ROUGE-1 (f-score) 0.330 Jump in performance More redundancies are captured 0.310 0.290 wt_loglen 0.270 0.250 1 2 3 4 5 σgap

ROUGE-1 (f-score) 0.330 Small improvements afterwards 0.310 0.290 wt_loglen 0.270 0.250 1 2 3 4 5 σgap

ROUGE-1 (f-score) 0.330 Small improvements afterwards 0.310 0.290 wt_loglen 0.270 0.250 1 2 3 4 5 σgap

ROUGE-1 (f-score) 0.330 Small improvements afterwards 0.310 0.290 wt_loglen 0.270 0.250 σgap too large: ill formed sentences 1 2 3 4 5 Set σgap to low value σgap

ROUGE-1 (f-score) 0.330 0.310 0.290 0.270 redundancy & path length only redundancy basic wt_loglen wt_len 0.250 1 2 3 4 5 σgap

ROUGE-1 (f-score) 0.330 0.310 0.290 0.270 redundancy & path length only redundancy basic wt_loglen wt_len 0.250 redundancy & path length 1 2 3 4 5 summaries with better coverage σgap

Topic X Opinosis Generated 1. sentence 1.. 2. sentence 2.. Human Composed 1 1. sentence 1.. 2. sentence 2.. 3. sentence Y.. Human Composed 4 1. sentence 1.. 2. sentence 2.. 3. sentence Z.. MIX Topic X Mixed Sentences sentence 1.. sentence 3.. sentence 2.. sentence 4.. sentence 8.. sentence 6.. sentence 7.. sentence 5.. Pick at most 2 least readable sentences

Assessor often picks: Opinosis sentences - Opinosis summaries have readability issues Non-Opinosis sentences or makes no picks - Opinosis summaries similar to human summaries

Assessor picked: 34/102 Opinosis generated sentences as least readable

Assessor picked: 34/102 Opinosis generated sentences as least readable > 60% of Opinosis sentences are not very different from human composed sentences

A framework for summarizing highly redundant opinions Use graph representation to generate concise abstractive summaries General & lightweight: Can be used on any corpus with high redundancies (Twitter comments, Blog comments, etc)

http://timan.cs.uiuc.edu/downloads.html

[Barzilay and Lee2003] Barzilay, Regina and Lillian Lee. 2003. Learning to paraphrase: an unsupervised approach using multiple-sequence alignment. In NAACL 03: Proceedings of the 2003 Conference of the North American Chapter of the Association for Computational Linguistics on Human Language Technology, pages 16 23, Morristown, NJ, USA. [DeJong1982] DeJong, Gerald F. 1982. An overview of the FRUMP system. In Lehnert, Wendy G. and Martin H. Ringle, editors, Strategies for Natural Language Processing, pages 149 176. Lawrence Erlbaum, Hillsdale, NJ. [Erkan and Radev2004] Erkan, G unes and Dragomir R. Radev. 2004. Lexrank: graph-based lexical centrality as salience in text summarization. J. Artif. Int. Res.,22(1):457 479. [Finley and Harabagiu2002] Finley, Sanda Harabagiu and Sanda M. Harabagiu. 2002. Generating single and multi-document summaries with gistexter. In Proceedings of the workshop on automatic summarization, pages 30 38. [Hu and Liu2004] Hu, Minqing and Bing Liu. 2004. Mining and summarizing customer reviews. In KDD, pages 168 177. [Jing and McKeown2000] Jing, Hongyan and Kathleen R. McKeown. 2000. Cut and paste based text summarization. In Proceedings of the 1st North American chapter of the Association for Computational Linguistics conference, pages 178 185, San Francisco, CA, USA. Morgan Kaufmann Publishers Inc. [Lerman et al.2009] Lerman, Kevin, Sasha Blair-Goldensohn, and Ryan Mcdonald. 2009. Sentiment summarization: Evaluating and learning user preferences. In 12th Conference of the European Chapter of the Association for Computational Linguistics (EACL-09). [Lin and Hovy2003] Lin, Chin-Yew and Eduard Hovy. 2003. Automatic evaluation of summaries using ngram co-occurrence statistics. In Proc. HLT-NAACL, page 8 pages. [LIN2004a] LIN, Chin-Yew. 2004a. Looking for a few good metrics : Rouge and its evaluation. proc. of the 4th NTCIR Workshops, 2004. [Lin2004b] Lin, Chin-Yew. 2004b. Rouge: a package for automatic evaluation of summaries. In Proceedings of theworkshop on Text Summarization Branches Out (WAS 2004), Barcelona, Spain. [Lu et al.2009] Lu, Yue, ChengXiang Zhai, and Neel Sundaresan. 2009. Rated aspect summarization of short comments. In 18th International World WideWeb Conference (WWW2009), April.

[Mihalcea and Tarau2004] Mihalcea, R. and P. Tarau. 2004. TextRank: Bringing order into texts. In Proceedings of EMNLP- 04and the 2004 Conference on Empirical Methods in Natural Language Processing, July. [Pang and Lee2004] Pang, Bo and Lillian Lee. 2004. A sentimental education: Sentiment analysis using subjectivity summarization based on minimum cuts. In Proceedings of the ACL, pages 271 278. [Pang et al.2002] Pang, Bo, Lillian Lee, and Shivakumar Vaithyanathan. 2002. Thumbs up? Sentiment classification using machine learning techniques. In Proceedings of the 2002 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 79 86. [Radev and McKeown1998] Radev, DR and K. McKeown. 1998. Generating natural language summaries from multiple on-line sources. Computational Linguistics, 24(3):469 500. [Radev et al.2000] Radev, Dragomir, Hongyan Jing, and Malgorzata Budzikowska. 2000. Centroid-based summarization of multiple documents: Sentence extraction, utility-based evaluation, and user studies. In ANLP/NAACL Workshop on Summarization, pages 21 29. [Radev et al.2002] Radev, Dragomir R., Eduard Hovy, and Kathleen McKeown. 2002. Introduction to the special issue on summarization. [Saggion and Lapalme2002] Saggion, Horacio and Guy Lapalme. 2002. Generating indicative-informative summaries with sumum. Computational Linguistics, 28(4):497 526. [Snyder and Barzilay2007] Snyder, Benjamin and Regina Barzilay. 2007. Multiple aspect ranking using the good grief algorithm. In In Proceedings of the Human Language Technology Conference of the North American Chapter of the Association of Computational Linguistics (HLT-NAACL), pages 300 307. [Titov and Mcdonald2008] Titov, Ivan and Ryan Mcdonald. 2008. A joint model of text and aspect ratings for sentiment summarization. In Proceedings of ACL-08: HLT, pages 308 316, Columbus, Ohio, June. Association for Computational Linguistics.