Download as pdf or txt
Download as pdf or txt
You are on page 1of 8

Available online at www.sciencedirect.

com

Expert Systems
with Applications
Expert Systems with Applications 34 (2008) 2622–2629
www.elsevier.com/locate/eswa

An empirical study of sentiment analysis for chinese documents


Songbo Tan *, Jin Zhang
Intelligent Software Department, Institute of Computing Technology, Chinese Academy of Sciences, P.O. Box 2704, Beijing 100080, PR China

Abstract

Up to now, there are very few researches conducted on sentiment classification for Chinese documents. In order to remedy this defi-
ciency, this paper presents an empirical study of sentiment categorization on Chinese documents. Four feature selection methods (MI,
IG, CHI and DF) and five learning methods (centroid classifier, K-nearest neighbor, winnow classifier, Naı̈ve Bayes and SVM) are inves-
tigated on a Chinese sentiment corpus with a size of 1021 documents. The experimental results indicate that IG performs the best for
sentimental terms selection and SVM exhibits the best performance for sentiment classification. Furthermore, we found that sentiment
classifiers are severely dependent on domains or topics.
Ó 2007 Elsevier Ltd. All rights reserved.

Keywords: Sentiment analysis; Information retrieval; Machine learning

1. Introduction attempts to train a sentiment classifier based on occurrence


frequencies of the various words in the documents.
With the advent of the Web and the enormous growth of The other approach (‘‘semantic orientation’’) (Hatzivassi-
digital content in Internet, databases, and archives, text loglou & McKeown, 1997; Turney & Littman, 2002;
categorization has received more and more attention in Whitelaw, Garg, & Argamon, 2005) is to classify words into
information retrieval and natural language processing two classes, such as ‘‘positive’’ or ‘‘negative’’, and then
community. This kind of work has focused on topical cat- count an overall positive/negative score for the text. If a
egorization, attempting to sort documents according to document contains more positive than negative terms it is
their subjects (such as economics or politics) (Pang et al., deemed as positive, and if the number of negative terms
2002). exceeds the number of positive terms it is assigned as
However, recent years have seen rapid growth in non- negative.
topical text analysis, in which characterizations are sought On Chinese document, however, there is relatively little
of the opinions, feelings, and attitudes expressed in a text, investigation conducted on sentiment classification.
rather than just the subjects. A key problem in this area is Through our energies on searching the Internet, only one
sentiment classification, where a document is labelled as a people’s work can be found, i.e., Ye et al. (2005, 2006).
positive (‘thumbs up’) or negative (‘thumbs down’) evalua- In order to remedy this deficiency, we present an empir-
tion of a target object (film, book, product, etc.). ical study of sentiment classification on Chinese docu-
Up to now, many researches have been conducted on ments. Our attempt is to answer following questions:
English document sentiment classification. These researches Which feature selection method does perform best for
have fallen into two categories. The first (‘‘machine learning sentiment classification on Chinese documents?
techniques’’) (Mullen & Collier, 2004; Pang et al., 2002) How many features are sufficient for this job?
Which learning method does perform best?
Can a sentiment classifier trained on one domain (e.g.
*
Corresponding author. Tel.: +86 10 62600928. education) perform well on another different domain (e.g.
E-mail address: tansongbo@software.ict.ac.cn (S. Tan). movie)?

0957-4174/$ - see front matter Ó 2007 Elsevier Ltd. All rights reserved.
doi:10.1016/j.eswa.2007.05.028
S. Tan, J. Zhang / Expert Systems with Applications 34 (2008) 2622–2629 2623

In this work, we use four traditional feature selection Smith, & Ogilvie, 1966) (GI for simplicity), which is a dic-
methods (Yang & Pedersen, 1997), i.e., mutual information tionary that contains information about English word
(MI), information gain (IG), CHI statistics (CHI), docu- senses. The method he used to classify a review is to count
ment frequency (DF), and five learning methods, i.e., positive and negative terms in it, as well as take into
K-nearest neighbor (KNN) (Yang & Lin, 1999), centroid account contextual valence shifters. Valence shifters are
classifier (Han & Karypis, 2000), Naı̈ve Bayes (NB) terms that can change the semantic orientation of another
(McCallum & Kamal, 1998), winnow (Zhang, 2001) and term, such as not, never, none, nobody, etc. In his experi-
support vector machine (SVM) (Joachims, 1998). We con- ment he used two datasets: one is collected by himself;
duct experiments on a Chinese sentiment corpus with a size the other is Pang’s dataset.
of 1021 documents divided into three domains: education, Chaovalit and Zhou (2005) compared machine learning
movie, and house. based methods and orientation based methods. He used n-
The rest of this paper is constructed as follows: next sec- gram model as supervised learning approach. The n-gram
tion presents related work on sentiment analysis. Feature represents text documents by word tuples. For orienta-
selection and learning methods are described in Section 3. tion-based methods, he first used Minipar1 to tag and parse
Experimental results are given in Section 4. Finally Section the review documents and then selectively extracted two-
5 concludes this paper. word phrases conforming to phrase patterns from Turney’s
study (Turney, 2002). His method to determine the seman-
2. Related work tic orientation is the same as Turney. For example, a
phrase’s semantic orientation would be positive if it is asso-
In this section, we present the related work on sentiment ciated more strongly with ‘‘excellent’’ than ‘‘poor’’ and
categorization. Sentiment classification has been investi- would be negative if it is associated more strongly with
gated in different domains such as movie reviews, product ‘‘poor’’ than ‘‘excellent’’.
reviews, and customer feedback reviews (Gamon, 2004;
Pang et al., 2002; Turney & Littman, 2003). 3. Methodology
Most of these researches up to this point have focused
on training machine learning algorithms to classify reviews This section presents the methodology of sentiment clas-
(Pang et al., 2002). sification system we used. First, we use Chinese text POS
Pang et al. (2002) conducted an extensive experiment on tool ICTCLAS (Zhang, 2003) to parse and tag Chinese
movie reviews using three traditional supervised machine review documents. Then feature selection method is used
learning methods (i.e., Naive Bayes (NB), maximum to pick out discriminating terms for training and classifica-
entropy classification (ME), and support vector machines tion. Finally we use machine learning method to learn a
(SVM)). His results indicate that standard machine learn- sentiment classifier.
ing techniques definitively outperform human-produced
baselines. However, he found that machine learning meth- 3.1. Feature selection methods
ods could not perform as well on sentiment classification as
on traditional topic-based categorization. It is worth notic- In this sub-section, we give a brief introduction of four
ing that he collected a movie-reviewing corpus that con- effective feature selection methods, i.e., DF, CHI, MI and
tains 700 positive reviews and 700 negative reviews. This IG. All these methods compute a score for each individual
corpus has been becoming the benchmark for sentiment feature and then pick out a predefined size of feature set.
categorization research.
Mullen and Collier (2004) employed support vector 3.1.1. DF
machines (SVMs) to bring together diverse sources of Document Frequency is the number of documents in
potentially pertinent information, including several favor- which a term occurs in a dataset. In Document Frequency
ability measures for phrases and adjectives and, where Thresholding one computes the document frequency for
available, knowledge of the topic of the text. Models using each word in the training corpus and removes those words
the features introduced are further combined with unigram whose document frequency is less than some predefined
models and lemmatized versions of the unigram models. small threshold or bigger than some predefined large
His experiment on Pang’s dataset indicates that hybrid threshold. The basic assumption is that both rare and com-
SVMs which combine unigram-style feature based SVMs mon words are either non-informative for category predic-
with those based on real-valued favorability measures tion, or not influential in global performance.
obtain superior performance. It is the simplest criterion for term selection and can eas-
Research has also been done by counting positive/nega- ily scales to a large dataset with linear computation com-
tive terms and automatically determining whether a term is plexity. It is a simple but effective feature selection
positive or negative (Turney & Littman, 2002). method for text categorization.
Kennedy and Inkpen (2005) made use of a semantic lex-
icon for identifying positive and negative terms. This lexi-
1
con is taken from the General Inquirer (Stone, Dunphy, http://www.cs.ualberta.ca/~lindek/minipar.htm.
2624 S. Tan, J. Zhang / Expert Systems with Applications 34 (2008) 2622–2629

3.1.2. CHI 1 X
Ci ¼ d
The CHI statistic measures the association between the jci j d2ci
term and the category (Galavotti, Sebastiani, & Simi,
2000). It is defined to be where jzj indicates the cardinality of set z, and d denotes the
N  ðAD  BEÞ
2 document in class ci.
CHIðt; ci Þ ¼ For each test document d, we calculate its similarity to
ðA þ EÞ  ðB þ DÞ  ðA þ BÞ  ðE þ DÞ
each centroid Ci using cosine measure as follows:
and CHImax ðtÞ ¼ maxi ðCHIðt; ci ÞÞ
d  Ci
where A is the number of times t and ci co-occur; B is the simðd; C i Þ ¼
k d k2 kC i k2
number of times t occurs without ci; E is the number of
times ci occurs without t; D is the number of times neither
ci nor t occurs; N is the total number of documents. 3.2.2. K-nearest neighbor classifier
The K-nearest neighbor (KNN) is a typical example-
based classifier that does not build an explicit, declarative
3.1.3. MI representation of the category ci, but rely on the category
Mutual information is a criterion commonly used in sta- labels attached to the training documents similar to the
tistical language modeling of word associations and related test document. As a result, KNN has been called lazy
applications (Yang & Pedersen, 1997). It can be defined as learners, since it defers the decision on how to generalize
following, beyond the training data until each new query instance
 
AN is encountered.
MIðt; ci Þ ¼ log and Given a test document d, the system finds the k nearest
ðA þ EÞ  ðA þ BÞ
neighbors among training documents. The similarity score
MImax ðtÞ ¼ maxi ðMIðt; ci ÞÞ
of each nearest neighbor document to the test document is
where the denotations of A, B, E, D and N are the same as used as the weight of the classes of the neighbor document.
The weighted sum in KNN classification can be written as
the definitions in CHI.
follows:
X
3.1.4. IG scoreðd; ci Þ ¼ simðd; d j Þdðd j ; ci Þ
Information gain is frequently employed as a term good- d j 2KNNðdÞ
ness criterion in the field of machine learning (Yang &
Pedersen, 1997). It measures the number of bits of informa- where KNN(d) indicates the set of k nearest neighbors of
tion obtained for category prediction by knowing the pres- document d. If dj belongs to ci, d(dj, ci) equals 1, or other-
ence or absence of a term in a document. wise 0. For test document d, it should belong to the class
that has the highest resulting weighted sum.
X
jCj X
jCj
IGðtÞ ¼  P ðci Þ log P ðci Þ þ P ðtÞ P ðci jtÞ log P ðci jtÞ
i¼1 i¼1
3.2.3. Naı̈ve Bayes
The Naı̈ve Bayes algorithm is a widely used algorithm
X
jCj
þ P ðtÞ P ðci jtÞ log P ðci jtÞ for document classification. Given a feature vector table,
i¼1 the algorithm computes the posterior probability that
the document belongs to different classes and assigns it
where P(ci) denotes the probability that class ci occurs; P(t) to the class with the highest posterior probability. There
denotes the probability that word t occurs; P ðtÞdenotes the are two commonly used models (i.e., multinomial model
probability that word t does not occurs. and multi-variate Bernoulli model) for using Naı̈ve Bayes
approach for text categorization. In this paper, and with-
3.2. Machine learning methods out loss of generality, we run the multinomial model
adopted by numerous authors (McCallum & Kamal,
3.2.1. Centroid classifier 1998).
The idea behind the centroid classification algorithm is Multinomial Naı̈ve Bayes counts the probability of the
extremely simple and straightforward. First we calculate word wt given category cj by following formula:
the prototype vector or centroid vector for each training PN j
class; then compute the similarity between a testing docu- nit
pðwt jcj Þ ¼ PW i¼1
PN j
ment d to all centroids; finally, based on these similarities, s¼1 i¼1 nis
we assign d to the class corresponding to the most similar
centroid. where nit is the number of appearances of word t in docu-
In training phase, we compute K centroids {C1, C2, . . . , ment i, Nj refers to the number of training documents in
CK} for the K classes using following formula: category cj and W refers to the vocabulary size.
S. Tan, J. Zhang / Expert Systems with Applications 34 (2008) 2622–2629 2625

The posterior probability can be calculated as follows: 4. Experiment results


pðcj Þpðd i jcj Þ
pðcj jd i Þ ¼ 4.1. Datasets
pðd i Þ
Because there is a lack of publicly available Chinese
3.2.4. Winnow classifier sentiment corpus for evaluating sentiment analysis sys-
Winnow is a well-known online mistaken-driven tems, we collect a Chinese sentiment corpus by ourselves.
method. It works by updating its weights in a sequence For the sake of convenience, we call this corpus as ‘‘Chn-
of trials. On each trial, it first makes a prediction for one SentiCorp’’. The total size is 1021 documents that consist
document d and then receives feedback; if a mistake is of three domains: education, movie, and house. There are
made, it updates its weight vector using the document d. 507 education-related documents, 266 movie-related doc-
During the training phase, with a collection of training uments and 248 house-related documents. Each domain
data, this process is repeated several times by iterating on category contains positive and negative documents. The
the data. Up to now, there are many variants of winnow, hierarchy is reported as Table 1. The total positive doc-
such as positive winnow, balanced winnow, and large-mar- uments amount to 458; while the total negative docu-
gin winnow. In this work we only run balanced winnow ments amount to 563. As a result, the corpus can be
because it consistently yields excellent performance (van regarded as four sentiment corpora: ChnSentiCorp, Chn-
Mun, http://citeseer.ist.psu.edu/cs). SentiCorpEdu, ChnSentiCorpMov, and ChnSentiCorp-
The balanced winnow algorithm keeps two weights for Hou.
each feature, wþ 
kt and wkt . For a given instance (dk1,
dk2, . . . ,dkW), the algorithm deems the document relevant iff 4.2. The performance measure

X
W
To evaluate a semantic classification system, we use the
ðwþ 
kt  wkt Þd kt P s
t¼1
F1 measure introduced by van Rijsbergen (1979). This
measure combines recall and precision in the following
where s denotes a given threshold and k indicates the class way:
label.
number of correct positive predictions
The weights of the active features are updated only when Recall ¼
a mistake is made. In the promotion step, following a mis- number of positive examples
take on a positive example, the positive part of the weight number of correct positive predictions
Precision ¼
is promoted, wþ þ
kt ¼ wkt  aða > 1Þ while the negative part of
number of positive predictions
the weight is demoted, w 
kt ¼ wkt  bð0 < b < 1Þ. The coef- 2  Recall  Precision
F1 ¼
ficient of dktin the equation above increases after a promo- ðRecall þ PrecisionÞ
tion. In the demotion step, by contraries, the positive part
For ease of comparison, we summarize the F1 scores over
of the weight is demoted, while the negative part of the
the different categories using the micro- and macro-aver-
weight is promoted.
ages of F1 scores:
3.2.5. SVM classifier Micro-F1 ¼ F1 over categories and documents
Support vector machines (SVM) is a relatively new class Macro-F1 ¼ average of within-category F1 values
of machine learning techniques first introduced by Vapnik
(1995). Based on the structural risk minimization principle The MicroF1 and MacroF1 emphasize the performance of
from the computational learning theory, SVM seeks a deci- the system on common and rare categories, respectively.
sion surface to separate the training data points into two Using these averages, we can observe the effect of different
classes and makes decisions based on the support vectors kinds of data on a classification system.
that are selected as the only effective elements in the train-
ing set.
Multiple variants of SVM have been developed (Joach-
Table 1
ims, 1998). Here we limit our discussion to linear SVM due The size of collected four sentiment corpora
to its popularity and high performance in text categoriza-
Domains Sentiment Documents
tion (Yang & Lin, 1999).
The optimization of SVM (dual form) is to minimize: ChnSentiCorpEdu Positive 204
( ) Negative 303
X n Xn X n
  ChnSentiCorpMov Positive 113

~a ¼ arg min  ai þ ai aj y i y j ~
xi ; x~j Negative 153
i¼1 i¼1 j¼1 ChnSentiCorpHou Positive 141
X
n Negative 107
Subject to : ai y i ¼ 0; 0 6 ai 6 C ChnSentiCorp Positive 458
i¼1
Negative 563
2626 S. Tan, J. Zhang / Expert Systems with Applications 34 (2008) 2622–2629

4.3. Experimental design Table 3


The best MacroF1 of four feature selection methods on five learning
methods
We employ TFIDF as input features. The formula for
calculating the TFIDF can be written as follows: Centroid KNN Winnow NB SVM Average
MI 0.8084 0.7841 0.8049 0.7866 0.8244 0.8017
W ðt; dÞ ¼ tf ðt; dÞ  logðN =nt Þ IG 0.8681 0.8730 0.8996 0.8840 0.9043 0.8858
where N is the total number of training documents, and nt CHI 0.8602 0.8404 0.8739 0.8882 0.8888 0.8703
DF 0.8612 0.8468 0.8777 0.8644 0.8480 0.8596
is the number of documents containing the word t. Average 0.8495 0.8361 0.8640 0.8558 0.8664
For Balanced Winnow, the initial weight value wþ 
il ðwil Þ
is set 2.0 (1.0), and the threshold was set to 1.0. The promo-
tion parameter a and the demotion b were fixed as 1.2 and
0.8, respectively.
First, with respect to feature selection methods, IG per-
For KNN, we set the number k of neighbors to 13. It is
forms the best across almost learning methods. Its average
worth noticing that we do not introduce any thresholds
MicroF1 is 0.8883, which is one percent larger that CHI
investigated by Yang (2001) because the adjusting of
(0.8737), two percents larger than DF (0.8644), and eight
thresholds may incur significant computational costs.
percents larger than MI (0.8086). The answer for the first
For experiments involving SVM we employed
question discussed in the introduction section is that IG
SVMTorch, which uses one-versus-the-rest decomposition
is the best choice for semantic terms selection.
and can directly deal with multi-class classification prob-
As such, with respect to learning methods, SVM pro-
lems. (http://www.idiap.ch/~bengio/projects/SVMTorch.
duces the best average MicroF1 (0.8685) which is slightly
html).
higher than Winnow (0.8663) or NB (0.8631). This obser-
vation indicates that SVM, Winnow, and NB are all suit-
4.4. Comparison and analysis
able for sentiment analysis.
Figs. 1–5 displays the performance curves of five learn-
Tables 2 and 3 report the best performance of four fea-
ing methods using IG vs. feature number. The corpus we
ture selection methods combined with five learning meth-
used is ChnSentiCorp as introduced in Section 4.1. From
ods. The corpus we used is ChnSentiCorp as introduced
these figures we can observe that when the number of fea-
in Section 4.1. From the two tables we can draw following
ture exceeds 6000, all learning methods produce desirable
conclusions:
and reasonable performance. For example, using a feature
set larger than 6000, the performance curves of SVM com-
bined with IG, CHI, and DF keeps nearly unchanged.
Table 2 Consequently, our answer for the second question is that
The best MicroF1 of four feature selection methods on five learning 6000 or larger size of features is sufficient for sentiment
methods categorization.
Centroid KNN Winnow NB SVM Average The second observation is that three feature selection
MI 0.8129 0.7943 0.8090 0.8012 0.8257 0.8086 methods (i.e., IG, CHI, and DF) combined with four learn-
IG 0.8736 0.8756 0.8981 0.8883 0.9060 0.8883 ing methods (i.e., Centroid, KNN, Winnow, and NB) exhi-
CHI 0.8658 0.8433 0.8776 0.8913 0.8903 0.8737 bit similar and robust performance. Under all conditions,
DF 0.8668 0.8511 0.8805 0.8717 0.8521 0.8644 MI does not have comparable performance with any of
Average 0.8548 0.8411 0.8663 0.8631 0.8685
the other methods.

Fig. 1. The performance curves of centroid using IG vs. feature number.


S. Tan, J. Zhang / Expert Systems with Applications 34 (2008) 2622–2629 2627

Fig. 2. The performance curves of KNN using IG vs. feature number.

Fig. 3. The performance curves of winnow using IG vs. feature number.

Fig. 4. The performance curves of NB using IG vs. feature number.

Fig. 5. The performance curves of SVM using IG vs. feature number.


2628 S. Tan, J. Zhang / Expert Systems with Applications 34 (2008) 2622–2629

Table 4
The performance of SVM when trained on one domain and transferred to another domain
Training set Testing set
ChnSentiCorpEdu ChnSentiCorpMov ChnSentiCorpHou
MicroF1 MacroF1 MicroF1 MacroF1 MicroF1 MacroF1
ChnSentiCorpEdu – – 0.6165 0.4753 0.8992 0.8982
ChnSentiCorpMov 0.7475 0.7058 – – 0.7339 0.7328
ChnSentiCorpHou 0.7574 0.7573 0.4962 0.4653 – –

Table 4 presents the performance of domain transfer of Gamon, Michael (2004). Sentiment classification on customer feedback
SVM. The number of features is fixed as 10,000. The row data: Noisy data,large feature vectors, and the role of linguistic
analysis. In Proceedings the 20th international conference on computa-
‘‘Training Set’’ indicates the domains for training the tional linguistics.
SVM classifier; the column ‘‘Testing Set’’ indicates the Han, E., & Karypis, G. (2000). Centroid-based document classification
domains for testing the SVM classifier. analysis and experimental result. PKDD.
Compared with above results involved with SVM, the Hatzivassiloglou, Vasileios, & McKeown, Kathleen (1997). Predicting the
performance of Table 4 is not desirable at all. Especially semantic orientation of adjectives. In Proceedings of the 35th ACL/8th
EACL (pp. 174–181).
when using ‘‘ChnSentiCorpHou’’ as training set and using Joachims, T. (1998). Text categorization with support vector machines:
‘‘ChnSentiCorpMov’’ as testing set, SVM yields very poor Learning with many relevant features. ECML, 137–142.
performance, 0.4962 for MicroF1 and 0.4653 for MacroF1. Kennedy, Alistair, & Inkpen, Diana (2005). Sentiment classification of
The only reasonable result is achieved when ‘‘ChnSenti- movie and product reviews using contextual valence shifters. In
CorpEdu’’ is employed as training set and ‘‘ChnSentiCorp- Workshop on the analysis of informal and formal information exchange
during negotiations (FINEXIN 2005).
Hou’’ is regarded as testing set. McCallum, Andrew, & Nigam, Kamal (1998). A comparison of event
These discouraging results when the classifier is trans- models for Naive Bayes text classification. AAAI/ICML-98 workshop
ferred across different domains inspire us a rule: machine on learning for text categorization [C] (pp. 41–48). Menlo Park, CA:
learning techniques based sentiment classifier is severely AAAI Press.
dependent on its domains or topics. Mullen, Tony, & Collier, Nigel (2004). Sentiment analysis using sup-
port vector machines with diverse information sources. In Dekang
Lin & Dekai Wu (Eds.). Proceedings of EMNLP-2004, Barcelona,
5. Conclusion remarks Spain, July 2004 (pp. 412–418). Association for Computational
Linguistics.
In this work, we conducted an empirical study of senti- Pang, Bo, Lee, Lillian, & Vaithyanathan, Shivakumar (2002). Thumbs up?
ment categorization on Chinese documents. Our main con- Sentiment classification using machine learning techniques. In Pro-
ceedings of EMNLP, 2002.
tributions are: Stone, Philip J., Dunphy, Dexter C., Smith, Marshall S., & Ogilvie, Daniel
In order to conduct this experiment, we by ourselves col- M.and associates. (1966). The general inquirer: A computer approach to
lect Chinese corpus with a size of 1021 documents. It con- content analysis. The MIT Press.
sists of news or reviews from three domains: education, Turney, Peter D. (2002). Thumbs up or thumbs down? Semantic
movie, and house. orientation applied to unsupervised classification of reviews. In
Proceedings of the association for computational linguistics 40th
Secondly, we found that IG performs the best for senti- anniversary meeting, New Brunswick, NJ.
mental terms selection and SVM exhibits the best perfor- Turney, D. Peter, & Littman, Michael L. (2002). Unsupervised learning of
mance for sentiment classification. semantic orientation from a hundred-billion-word corpus. Technical
Thirdly, the experimental results indicate that 6000 or Report EGB-1094, National Research Council Canada.
larger size of features are sufficient for sentiment analysis. Turney, Peter D., & Littman, M. L. (2003). Measuring praise and
criticism: Inference of semantic orientation from association.
Finally, we found that sentiment classifiers are severely ACM Transactions on Information Systems (TOIS), 21(4), 315–
dependent on domains or topics. 346.
With the consideration of last conclusion, our future van Mun, P. P. T. M. Text classification in information retrieval using
effort is to investigate how to train a sentiment classifier winnow. http://citeseer.ist.psu.edu/cs.
that is independent on domain or topics. Another alterna- van Rijsbergen, C. (1979). Information retrieval. London: Butterworth.
Vapnik, Vladmimir N. (1995). The nature of statistical learning theory.
tive to this issue is to construct a semantic lexicon. New York: Springer.
Whitelaw, Casey, Garg, Navendu, & Argamon, Shlomo (2005). Using
References appraisal groups for sentiment analysis. CIKM, 625–631.
Yang, Y. (2001). A study on thresholding strategies for text categoriza-
Chaovalit, Pimwadee, & Zhou, Lina (2005). Movie review mining: A tion. SIGIR, 137–145.
comparison between supervised and unsupervised classification Yang, Y., & Lin, X. (1999). A re-examination of text categorization
approaches. In IEEE proceedings of the 38th Hawaii international methods. SIGIR, 42–49.
conference on system sciences, Big Island, Hawaii (pp. 1–9). Yang, Y., & Pedersen, Jan O. (1997). A comparative study on feature
Galavotti, L., Sebastiani, F., & Simi, M. (2000). Feature selection and selection in text categorization. ICML, 412–420.
negative evidence in automated text categorization. In Proceedings of Ye, Qiang, Lin, Bin, & Li, Yijun (2005). Sentiment classification for
KDD. Chinese reviews: A comparison between SVM and semantic
S. Tan, J. Zhang / Expert Systems with Applications 34 (2008) 2622–2629 2629

approaches. In The 4th international conference on machine learning Zhang, T. (2001). Regularized winnow methods. Advances in Neural
and cybernetics ICMLC2005, June 2005. Information Processing Systems, 13, 703–709.
Ye, Qiang, Shi, Wen, & Li, Yijun (2006). Sentiment classification for Zhang, Huaping (2003). Chinese lexical analysis using hierarchical hidden
movie reviews in Chinese by improved semantic oriented approach. In Markov model. In Second SIGHAN workshop affiliated with 41th ACL,
Proceedings of HICSS-39 Hawaii international conference on system Sapporo Japan, July 2003 (pp. 63–70).
sciences, January 2006.

You might also like