Abstract
Probabilistic learning models have the
ability to be calibrated to improve the performance
of tasks such as sentiment classification.
In this paper, we introduce a
framework for sentiment classification that
enables classifier recalibration given the
presence of related, context-bearing documents.
We investigate the use of probabilistic
thresholding and document similarity
based recalibration methods to yield
classifier improvements. We demonstrate
the performance of our proposed recalibration
methods on a dataset of online clinical
reviews from the patient feedback domain
that have adjoining management responses
that yield sentiment bearing information.
Experimental results show the proposed
recalibration methods outperform uncalibrated
supervised machine learning models
trained for sentiment analysis, and
yield significant improvements over a robust
baseline.
Original language | English |
---|---|
Pages | 175-180 |
Publication status | Published - 2015 |
Event | Workshop on Computational Approaches to Subjectivity, Sentiment & Social Media Analysis - Lisboa, Portugal Duration: 17 Sept 2015 → 17 Sept 2015 |
Workshop
Workshop | Workshop on Computational Approaches to Subjectivity, Sentiment & Social Media Analysis |
---|---|
Country/Territory | Portugal |
City | Lisboa |
Period | 17/09/15 → 17/09/15 |