Improving Sentiment Classification through Distinct Word Selection Heeryon Cho

Sang Min Yoon

HCI Laboratory College of Computer Science Kookmin University Seoul, South Korea [email protected] Abstract—While the performance of sentiment classification has steadily risen through the introduction of various featurebased methods and distributed representation-based approaches, less attention was given to the qualitative aspect of classification, for instance, the identification of useful words in individual opinion texts. We present an approach using set operations for identifying useful words for sentiment classification, and employ truncated singular value decomposition (SVD), a classic low-rank matrix decomposition technique for document retrieval, in order to tackle the issue of both synonymy and noise removal. The sentiment classification performance of our approach, which concatenates three kinds of features, outperforms the existing word-based and distributed word representation-based methods and is comparable to the existing state of the art distributed document representation-based approaches. Keywords—sentiment classification; truncated singular value decomposition (SVD); set operation; term frequency-inverse document frequency (TF-IDF)

I. INTRODUCTION Since the inception, the goal of sentiment classification research was focused on correctly classifying the sentiment (e.g., positive, negative, star ratings, etc.) of opinion texts [1, 2]. The sentiment classification performance has steadily increased through the introduction of various feature-based methods [3, 4] and distributed representation approaches [5, 6], but less attention was given to the qualitative aspect of classification, for instance, the identification of useful words in opinion texts. As more enterprises utilize sentiment analysis of product reviews in their product development, the need to identify useful keywords within opinion texts is becoming paramount. However, existing approaches have predominantly focused on improving the classification performance using whatever features that would suit the given classification algorithm, and recent cutting edge distributed representationbased approaches use one dimensional convolutional neural network to convolve sequence of words, making it difficult to identify useful words for sentiment classification. To tackle the problem of identifying useful words in sentiment classification while maintaining the classification performance, we revisit the classical word-document model This work was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) grants funded by the Korean government (MSIP) (NRF-2017R1A2B4011015) and the Korean Ministry of Education (NRF-2016R1D1A1B04932889), and also the Institute for Information & Communications Technology Promotion (IITP) grant funded by the Korean government (MSIP) (No. R0115-16-1009).

and present a simple set theoretic approach for identifying distinct words, where disjunction, conjunction, and complement operations are performed on a set of review words selected using term frequency-inverse document frequency (TF-IDF) weighting [7]. Afterwards, latent semantic analysis (or more generally, truncated singular valued decomposition (SVD)) [8] is performed on the product reviews using the distinct words obtained from the set operations to generate compact features for sentiment classification. In our previous work [9], we have shown the effectiveness of the distinct words for clustering online news comments. In [9], the distinct words were selected from the online news comments of a given news article by removing top-N most frequent words generated across a large volume of online news comments. In essence, more frequently occurring words, which tend to include more general words, were removed to highlight more specific words of a given news article’s comment. This work is an extension of our previous approach to a new domain of sentiment classification. We construct different sets of product reviews based on the star ratings and perform set operations, namely, disjunction, conjunction, and complement operation, on the product review words. Our contribution is twofold:  We present a simple set theoretic approach that performs disjunction, conjunction, and complement on a set of review words for selecting distinct words for sentiment classification. We also perform star ratingbased truncated SVD on product reviews to tackle the issues of synonymy and noise removal.  We prove the effectiveness of our approach by concatenating the distinct word features to other baseline truncated SVD features. Our approach outperforms the existing word-based and distributed word representation-based approaches and is comparable to the state of the art distributed document representation-based approaches. Hereafter we explain the set operations for selecting distinct words in product reviews and how the truncated SVD is performed on product reviews for the baseline feature generation in Section II. We then present the details of the evaluation experiment using three benchmark

Heeryon Cho and Sang Min Yoon, “Improving Sentiment Classification through Distinct Word Selection,” 2017 10th International Conference on Human System Interactions (HSI), Ulsan, South Korea, 2017, pp. 202-205. https://doi.org/10.1109/HSI.2017.8005029

product review datasets on Section III and discuss the implication of the experiment results in Section IV. Finally, we conclude this paper in Section V. II. DISTINCT WORD SELECTION & TRUNCATED SVD A. Distinct Word Selection Given a set of product reviews with star ratings, we create three kinds of word sets, i.e., merged, common, and distinct word sets. Each word set is generated as follows: 1. First, we divide the product reviews according to the star rating. For example, if the product reviews employ a five-star rating, we divide the product reviews into five review sets, each containing the same star rating. 2. We then apply TF-IDF weighting to each review set and select the top 10,000 words with high TF-IDF scores. If the reviews have five star ratings, a total of 50,000 words are selected. A.

We perform disjunction on the 50,000 words to create a merged word set.

B.

We perform conjunction on the 50,000 words to create a common word set.

C.

We obtain the difference of merged and common word set by considering the merged word set as the universal set and take the complement of the common word set. This set is denoted as the distinct word set.

By performing the above set operations, we obtain a unique word list that contains words that repeatedly appear across different star-rated reviews (common) and another unique word list that contains words that appear only within each of the five star-rated reviews (distinct).

Fig. 1. Overview of truncated SVD performed on product reviews.

III. EXPERIMENT & RESULT A. Dataset Three datasets1 that include one movie review dataset from IMDB and two restaurant review datasets from Yelp Dataset Challenge in 2013 and 2014 were used in the evaluation experiment. The statistics of the dataset is given in Table I. TABLE I.

DATASET

IMDB

Yelp 2013

Yelp 2014

Train (# reviews)

67,426

62,522

183,019

Test (# reviews)

9,112

8,671

25,399

Rating scale

1-10

1-5

1-5

B. Experimental Setup We generate the following three kinds of low rank approximation matrices (or features) using truncated SVD:

B. Truncated SVD Usually, the document-word matrix is extremely sparse and does not encode synonymy or polysemy information. To tackle these issues of synonymy/polysemy handling and noise removal, truncated SVD is applied to the document-word matrix to transform the existing matrix to a lower dimensional feature space. Mathematically, truncated SVD applied to a set of product reviews X produces a low-rank approximation of X:

1. The low rank features generated using common, distinct, and merged word set. The k largest singular value is set to 100. The performances of these features are shown in different colored bars - yellow, orange, and green - in Figs 2, 3, 4, & 5 (indicated as ‘Word’).



2. The traditional k-SVD performed on the top 10,000 TF-IDF words extracted using all training reviews. The rank of k is set to 1,000, 2,000, 3,000, and 4,000 which maps to ‘1k-SVD’, ‘2k-SVD’, etc., in Figs 2, 3, 4, & 5 (indicated as ‘k-SVD’).

In (1), U k  k is the transformed product reviews with the k largest singular values. Figure 1 (a) visualizes the matrix decomposition using truncated SVD on all product reviews. In our evaluation experiment, we generate various low-rank approximation matrices using the specific star-rated product reviews. Figure 1(b) gives an example of such transformation where truncated SVD is performed only on 1-star rated reviews. By multiplying the initial document-word matrix with Vk , we obtain a transformed low-rank matrix as in (2):

3. The class-wise k-SVD generated using one, two, or three neighboring class reviews. Maps to the ‘Class or 123 Class’ in Figs 2, 3, 4, & 5. The size of k is set to 100 each. For example, in the case of ‘1-class’, k-SVD is performed on class-wise reviews as depicted in Fig. 1(b). The ‘2-class’ k-SVD generates k-SVD matrices using two neighboring class reviews, i.e., 1&2-star reviews, 2&3-star reviews, 3&4-star reviews, and so on. In the experiment, we compare the performances of 1class, or 2 or 3-neighboring classes’ k-SVD results.





X  X k  U k  k VkΤ 

X   XVk 



1

Available at: http://ir.hit.edu.cn/~dytang/paper/acl2015/dataset.7z

Heeryon Cho and Sang Min Yoon, “Improving Sentiment Classification through Distinct Word Selection,” 2017 10th International Conference on Human System Interactions (HSI), Ulsan, South Korea, 2017, pp. 202-205. https://doi.org/10.1109/HSI.2017.8005029

TABLE II.

SENTIMENT CLASSIFICATION PERFORMANCES ON THREE BENCHMARK DATASETS: HIGHER ACCURACY & LOWER MAE·RMSE ARE BETTER (a) Compared with word-based & distributed word representation-based methods

APPROACH

IMDB

Yelp 2013

Yelp 2014

ACC

MAE

RMSE

ACC

MAE

RMSE

ACC

MAE

RMSE

TRIGRAM

0.399

1.147

1.783

0.569

0.513

0.814

0.577

0.487

0.804

TEXTFEATURE

0.402

1.134

1.793

0.556

0.520

0.845

0.572

0.490

0.800

AVGWORDVEC+SVM

0.304

1.361

1.985

0.526

0.568

0.898

0.530

0.530

0.893

SSWE+SVM

0.312

1.347

1.973

0.549

0.529

0.849

0.557

0.523

0.851

OUR APPROACH

0.353

1.100

1.631

0.613

0.435

0.746

0.599

0.446

0.744

(b) Compared with distributed document representation-based methods APPROACH

IMDB

Yelp 2013

Yelp 2014

ACC

MAE

RMSE

ACC

MAE

RMSE

ACC

MAE

RMSE

PARAGRAPH VECTOR

0.341

1.211

1.814

0.554

0.515

0.832

0.564

0.496

0.802

RNTN+RECURRENT

0.400

1.133

1.764

0.574

0.489

0.804

0.582

0.478

0.821

CNN+SOFTMAX

0.405

1.030

1.629

0.577

0.485

0.812

0.585

0.483

0.808

CNN+SVM

0.468

0.903

1.487

0.624

0.413

0.713

0.624

0.410

0.704

OUR APPROACH

0.353

1.100

1.631

0.613

0.435

0.746

0.599

0.446

0.744

Fig. 2. Accuracy of Yelp 2013 data using individual features.

Fig. 4. Accuracy of Yelp 2013 data using three-feature combination.

Fig. 3. Accuracy of Yelp 2013 data using two-feature combination.

Fig. 5. Accuracy of Yelp 2013 data using two-feature combination.

Heeryon Cho and Sang Min Yoon, “Improving Sentiment Classification through Distinct Word Selection,” 2017 10th International Conference on Human System Interactions (HSI), Ulsan, South Korea, 2017, pp. 202-205. https://doi.org/10.1109/HSI.2017.8005029

Note that we include both 1-gram and 2-gram words when performing TF-IDF; hence, the top 10,000 words contain both 1-gram and 2-gram words. We use softmax regression as the classification algorithm for multiclass sentiment classification. C. Evaluation Metric Following Chen et al.’s work [10], we use accuracy to measure the overall sentiment classification performance and use mean absolute error (MAE) and root mean squared error (RMSE) to measure the divergences between predicted and ground truth sentiment (i.e., star) ratings. For accuracy, higher score is better; for MAE and RMSE, lower score is better. MAE and RMSE are computed as follows: 

MAE 

1 N

RMSE 

N

 gold i 1

1 N

i

 predicted i

N

 ( gold i 1

i





 predicted i ) 2

D. Baseline Methods We compare our approach to the baseline methods given in [10] (see Table II). TRIGRAM and TEXTFEATURE are word-based methods; AVGWORDVEC+SVM and SSWE+ SVM are distributed word representation-based methods; PARAGRAPH VECTOR, RNTN+RECURRENT, CNN+ SOFTMAX, and CNN+SVM are distributed document representation-based methods (refer [10] for details of each benchmark approaches). E. Result Table II compares the various approaches’ sentiment classification performances on the three benchmark datasets. Our approach lists the best performances achieved using the three-feature combinations (Word+123-Class+k-SVD). We see that our approach outperforms the word-based and distributed word representation-based methods (Table II (a)) and is comparable to the distributed document representationbased methods (Table II (b)). The numbers in bold indicate the best performances within the individual tables in Table II. Figures 2, 3, 4, & 5 indicate the sentiment classification accuracies of our approach using individual features (Fig. 2), two-feature combinations (Figs. 3 & 5), and three-feature combinations (Fig. 4) on the Yelp 2013 dataset. We see that the best accuracy of 61.25% is achieved using the distinct word+2k-SVD+123-class combination. IV. DISCUSSION Recent outstanding progress in deep neural network-based approaches have started to dominate the sentiment classification research. While deep neural network techniques have the strength to automatically generate working features from raw text, often the features generated are incomprehensible to humans. Even though the pursuit of high classification accuracy is the prime objective of sentiment classification, the interpretability of the classification result, which includes the understanding of important feature words, should be considered to provide additional insights to human

evaluators. In this sense, our approach presented in this paper can be useful in selecting effective words for sentiment classification; we presented the three-feature combination approach which includes the distinct word set. We quantitatively proved that the distinct word set is useful for sentiment classification through additively concatenating the various truncated SVD features. In Fig. 2 (a), we see that the distinct word set performs poorly on its own. However, when it is combined with other features (Figs. 3, 4, 5), we see that in all cases, the combinations including the distinct word set outperform the other word sets (common and merged) proving its usefulness. Last but not least, we see that the simple TFIDF (Fig. 2 (a) tf-idf) approach outperforms the CNN+SOFTMAX (Table II (b)) approach on Yelp 2013 dataset (59.95% > 57.70%). This cautions us to choose sentiment classification approaches prudently; deep learningbased approaches are not always an almighty solution. Sometimes using simple and interpretable methods can be a better choice. V. CONCLUSION We presented a simple set operation for selecting distinct words for sentiment classification and introduced the concatenation of various truncated SVD features for sentiment classification. Our approach outperformed the existing wordbased and distributed word representation-based methods and was comparable to the state of the art distributed document representation-based approaches. We plan to investigate other low rank approximation-based approaches (e.g., non-negative matrix factorization (NMF)) for sentiment classification and explore various methods in extracting and clustering useful words for human inspection. REFERENCES [1]

B. Pang and L. Lee, “Opinion mining and sentiment analysis,” Found. Trends Inf. Retrieval, Now, Jan., 2008. [2] B. Liu, “Opinion mining and sentiment analysis,” in Synthesis Lectures on Human Language Technologies, Morgan & Claypool Publishers, 2012. [3] B. Pang, L. Lee, and S. Vaithyanathan, “Thumbs up? Sentiment classification using machine learning techniques,” in Proc. EMNLP, 2002, pp. 79–86. [4] L. Qu, G. Ifrim, and G. Weikum, “The bag-of-opinions method for review rating prediction from sparse text patterns,” In Proc. COLING, 2010, pp. 913–921. [5] Yoon Kim, “Convolutional neural networks for sentence classification,” in Proc. EMNLP, 2014, pp. 1746–1751. [6] D. Tang, B. Qin, and T. Liu, “Document modeling with gated recurrent neural network for sentiment classification,” in Proc. EMNLP, 2015, pp. 1422–1432. [7] C.D. Manning, P. Raghavan, and H. Schütze, “Chapter 6: Scoring, term weighting & the vector space model,” in Introduction to Information Retrieval, Cambridge University Press, 2008. [8] C.D. Manning, P. Raghavan, and H. Schütze, “Chapter 18: Matrix decompositions & latent semantic indexing,” in Introduction to Information Retrieval, Cambridge University Press, 2008. [9] H. Cho and J.-S. Lee, “Data-driven feature word selection for clustering online news comments,” in Proc. BigComp, 2016, pp. 494–497. [10] T. Chen, R. Xu, Y. He, Y. Xia and X. Wang, “Learning user and product distributed representations using a sequence model for sentiment analysis,” IEEE Comput. Intell. Mag, vol. 11, no. 3, pp. 34–44, Aug., 2016.

Heeryon Cho and Sang Min Yoon, “Improving Sentiment Classification through Distinct Word Selection,” 2017 10th International Conference on Human System Interactions (HSI), Ulsan, South Korea, 2017, pp. 202-205. https://doi.org/10.1109/HSI.2017.8005029

Paper Title (use style: paper title)

College of Computer Science. Kookmin ... of the distinct words for clustering online news comments. In ... This work was supported by the Basic Science Research Program through .... is performed on class-wise reviews as depicted in Fig. 1(b).

521KB Sizes 2 Downloads 381 Views

Recommend Documents

Paper Title (use style: paper title) - Sites
Android application which is having higher graphics or rendering requirements. Graphics intensive applications such as games, internet browser and video ...

Paper Title (use style: paper title) - GitHub
points in a clustered data set which are least similar to other data points. ... data mining, clustering analysis in data flow environments .... large than the value of k.

Paper Title (use style: paper title)
School of Electrical Engineering, KAIST .... [Online]. Available: http://yann.lecun.com/exdb/mnist/. [5] Design Compiler User Guide, Synopsys, Mountain View, CA, ...

Paper Title (use style: paper title)
on the substrate, substrate pre-deposition process, and Pd deposition .... concentration is below the ignition threshold, which is often important for such a sensor.

Paper Title (use style: paper title)
Turin, Italy [email protected]. Hui Wang. School of Information Engineering. Nanchang Institute of Technology. Nanchang 330099, China [email protected]. Abstract—Frequency Modulation (FM) sound synthesis provides a neat synthesis

Paper Title (use style: paper title)
mobile wireless networking, it is becoming possible to monitor elderly people in so-called ... sensor network that might be used in order to recognize tasks described in Table 1. ..... its advantages, and their relative merits and demerits are still.

Paper Title (use style: paper title)
zero which means cosθ tends to 1. The distance between each of the test vectors and profile vectors were obtained using (2). If the cosine value between the test vector and profile hub vector was greater than the cosine value between the same test v

Paper Title (use style: paper title)
communication channel between the sensors and the fusion center: a Binary ..... location estimation in sensor networks using binary data," IEEE Trans. Comput., vol. ... [9] K. Sha, W. Shi, and O. Watkins, "Using wireless sensor networks for fire.

Paper Title (use style: paper title)
search and compact storage space. Although search ... neighbor search methods in the binary space. ... Given a query ∈ { } , we list the online search algorithm.

Paper Title (use style: paper title)
Research Program Fellowships, the University of Central Florida – Florida. Solar Energy Center (FSEC), and a NASA STTR Phase I contract. NNK04OA28C. ...... Effluents Given Off by Wiring Insulation," Review of Progress in. QNDE, vol. 23B ...

Paper Title (use style: paper title)
In Long term Evolution. (LTE), HARQ is implemented by MAC level module called .... the receiver is decoding already received transport blocks. This allows the ...

use style: paper title
helps learners acquire scientific inquiry skills. One of ... tutoring systems; LSA; natural language processing ..... We collected data from 21 college students who.

Paper Title (use style: paper title)
Reducing Power Spectral Density of Eye Blink Artifact through Improved Genetic ... which could be applied to applications like BCI design. MATERIALS AND ...

Paper Title (use style: paper title)
general, SAW technology has advantages over other potentially competitive ... SAW devices can also be small, rugged, passive, wireless, and radiation hard,.

Paper Title (use style: paper title)
provide onboard device sensor integration, or can provide integration with an .... Figure 2 Schematic diagram of a 7 chip OFC RFID tag, and. OFC measured and ..... [3] C. S. Hartmann, "A global SAW ID tag with large data capacity," in Proc.

Paper Title (use style: paper title) - Research at Google
decades[2][3], but OCR systems have not followed. There are several possible reasons for this dichotomy of methods: •. With roots in the 1980s, software OCR ...

Paper Title (use style: paper title) - Research
grams for two decades[1]. Yet the most common question addressed to the author over more than two decades in OCR is: “Why don't you use a dictionary?

Paper Title (use style: paper title)
determine the phase error at unity-gain frequency. In this paper, while comparing some topologies we ... degrees at the integrator unity gain frequency result in significant filter degradation. Deviations from the .... due to gm/Cgd occur at a much h

Paper Title (use style: paper title)
Abstract— The Open Network and Host Based Intrusion Detection. Testbed .... It is unique in that it is web-based. .... sensor is also the application web server.

Paper Title (use style: paper title)
Orlando, FL 32816-2450 (email: [email protected]). Brian H. Fisher, Student .... presentation provides a foundation for the current efforts. III. PALLADIUM ...

Paper Title (use style: paper title)
A VLSI architecture for the proposed method is implemented on the Altera DE2 FPGA board. Experimental results show that the proposed design can perform Chroma-key effect with pleasing quality in real-time. Index Terms—Chroma-key effect, K-means clu

Paper Title (use style: paper title)
the big amount of texture data comparing to a bunch of ... level and a set of tile data stored in the system memory from ... Figure 1: Architecture of our algorithm.

Paper Title (use style: paper title)
printed texts. Up to now, there are no ... free format file TIFF. ... applied on block texts without any use of pre- processing ... counting [12, 13] and the reticular cell counting [1]. The main ..... Computer Vision and Image Understanding, vol. 63

Paper Title (use style: paper title)
School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore 639798. Abstract— ... For 60GHz wireless communication systems, the ... the benefit of isolated DC noise from the tuning element. The load on ...