Department of Computer Science, Georgia State University, Atlanta, GA30302, USA
Suzhou Key Laboratory of Advanced Optical Communication Network Technology, School of Electronic and Information Engineering, Soochow University, Suzhou215006, China
Show Author Information
Hide Author Information
Abstract
The COVID-19 pandemic has hit the world hard. The reaction to the pandemic related issues has been pouring into social platforms, such as Twitter. Many public officials and governments use Twitter to make policy announcements. People keep close track of the related information and express their concerns about the policies on Twitter. It is beneficial yet challenging to derive important information or knowledge out of such Twitter data. In this paper, we propose a Tripartite Graph Clustering for Pandemic Data Analysis (TGC-PDA) framework that builds on the proposed models and analysis: (1) tripartite graph representation, (2) non-negative matrix factorization with regularization, and (3) sentiment analysis. We collect the tweets containing a set of keywords related to coronavirus pandemic as the ground truth data. Our framework can detect the communities of Twitter users and analyze the topics that are discussed in the communities. The extensive experiments show that our TGC-PDA framework can effectively and efficiently identify the topics and correlations within the Twitter data for monitoring and understanding public opinions, which would provide policy makers useful information and statistics for decision making.
L. J.Chang, W.Li, L.Qin, W. J.Zhang, and S. Y.Yang, pSCAN: Fast and exact structural graph clustering, IEEE Trans. Knowl. Data Eng., vol. 29, no. 2, pp. 387-401, 2017.
R.El Bacha and T. T.Zin, Ranking of influential users based on user-tweet bipartite graph, in Proc. of 2018 IEEE Int. Conf. Service Operations and Logistics, and Informatics (SOLI), Singapore, 2018, pp. 97-101.
A.Rodríguez, C.Argueta, and Y. L.Chen, Automatic detection of hate speech on facebook using sentiment and emotion analysis, in Proc. of 2019 Int. Conf. Artificial Intelligence in Information and Communication (ICAIIC), Okinawa, Japan, 2019, pp. 169-174.
A.Reyes-Menendez, J. R.Saura, and C.Alvarez-Alonso, Understanding #worldEnvironmentDay user opinions in twitter: A topic-based sentiment analysis approach, Int. J. Environ. Res. Public Health, vol. 15, no. 11, p. 2537, 2018.
C. H.Tan, L. L.Lee, J.Tang, L.Jiang, M.Zhou, and P.Li, User-level sentiment analysis incorporating social networks, in Proc. 17th ACM SIGKDD Int. Conf. Knowledge Discovery and Data Mining, New York, NY, USA, 2011, pp. 1397-1405.
R. R.Iyer, J.Chen, H. N.Sun, and K. Y.Xu, A heterogeneous graphical model to understand user-level sentiments in social media, arXiv preprint arXiv: 1912.07911, 2019.
H. B.Deng, J. W.Han, H.Li, H.Ji, H. N.Wang, and Y.Lu, Exploring and inferring user-user pseudo-friendship for sentiment analysis with heterogeneous networks, Stat. Anal. Data Min., vol. 7, no. 4, pp. 308-321, 2014.
C. A.Phillips, Multipartite graph algorithms for the analysis of heterogeneous data, PhD dissertation, Univ. Tennessee, Knoxville, TN, USA, 2015.
[18]
D. W.Zhou, S.Zhang, M. Y.Yildirim, S.Alcorn, H. H.Tong, H.Davulcu, and J. R.He, A local algorithm for structure-preserving graph cut, in Proc. 23rd ACM SIGKDD Int. Conf. Knowledge Discovery and Data Mining, Halifax, Canada, 2017, pp. 655-664.
P. M.Comar, P. N.Tan, and A. K.Jain, A framework for joint community detection across multiple related networks, Neurocomputing, vol. 76, no. 1, pp. 93-104, 2012.
Y. Z.Sun, Y. T.Yu, and J. W.Han, Ranking-based clustering of heterogeneous information networks with star network schema, in Proc. 15th ACM SIGKDD Int. Conf. Knowledge Discovery and Data Mining, Paris, France, 2009, pp. 797-806.
D. D.Lee and H. S.Seung, Algorithms for non-negative matrix factorization, in Proc. 13th Int. Conf. Neural Information Proc. Systems, Cambridge, MA, USA, 2001, pp. 535-541.
[22]
N.Gillis, The why and how of nonnegative matrix factorization, arXiv preprint arXiv: 1401.5226v2, 2014.
M. E.Wall, A.Rechtsteiner, and L. M.Rocha, Singular value decomposition and principal component analysis, in A Practical Approach to Microarray Data Analysis, D. P.Berrar, W.Dubitzky, M.Granzow, eds.Norwell, MA, USA: Springer, 2003, pp. 91-109.
[25]
C.Ding, T.Li, W.Peng, and H.Park, Orthogonal nonnegative matrix t-factorizations for clustering, in Proc. 12th ACM SIGKDD Int. Conf. Knowledge Discovery and Data Mining, Philadelphia, PA, USA, 2006, pp. 126-135.
D.Kim, S.Sra, and I. S.Dhillon, Fast newton-type methods for the least squares nonnegative matrix approximation problem, in Proc. 2007 SIAM Int. Conf. Data Mining, Minneapolis, MN, USA, 2007, pp. 343-354.
C. J.Lin, On the convergence of multiplicative update algorithms for nonnegative matrix factorization, IEEE Trans. Neural Netw., vol. 18, no. 6, pp. 1589-1596, 2007.
J.Kim and H.Park, Toward faster nonnegative matrix factorization: A new algorithm and comparisons, in Proc. of 2008 Eighth IEEE Int. Conf. Data Mining, Pisa, Italy, 2008, pp. 353-362.
F.Wang and P.Li, Efficient nonnegative matrix factorization with random projections, in Proc. 2010 SIAM Int. Conf. Data Mining, Columbus, OH, USA, 2010, pp. 281-292.
M.Annett and G.Kondrak, A comparison of sentiment analysis techniques: Polarizing movie blogs, in Proc. 21st Conference of the Canadian Society for Computational Studies of Intelligence, Windsor, Canada, 2008, pp. 25-35.
M.Del Vicario, G.Vivaldo, A.Bessi, F.Zollo, A.Scala, G.Caldarelli, and W.Quattrociocchi, Echo chambers: Emotional contagion and group polarization on facebook, Sci. Rep., vol. 6, p. 37825, 2016.
S. M.Mohammad, X. D.Zhu, S.Kiritchenko, and J.Martin, Sentiment, emotion, purpose, and style in electoral tweets, Informat. Proc. Manag., vol. 51, no. 4, pp. 480-499, 2015.
K.Chakraborty, S.Bhattacharyya, R.Bag, and A.Hassanien, Sentiment analysis on a set of movie reviews using deep learning techniques, in Social Network Analytics Computational Research Methods and Techniques, Cambridge, MA, USA, 2019, pp. 127-147.
H.Meisheri, K.Ranjan, and L.Dey, Sentiment extraction from consumer-generated noisy short texts, in Proc. of 2017 IEEE Int. Conf. Data Mining Workshops (ICDMW), New Orleans, LA, USA, 2017, pp. 399-406.
A. S. M.Alharbi and E.de Doncker, Twitter sentiment analysis with a deep neural network: An enhanced approach using user behavioral information, Cogn. Syst. Res., vol. 54, pp. 50-61, 2019.
M.Wang, C. K.Wang, J. X.Yu, and J.Zhang, Community detection in social networks: An in-depth benchmarking study with a procedure-oriented framework, Proc. VLDB Endow., vol. 8, no. 10, pp. 998-1009, 2015.
D.Cai, X. F.He, X. Y.Wu, and J. W.Han, Non-negative matrix factorization on manifold, in Proc. 2008 8th IEEE Int. Conf. Data Mining, Pisa, Italy, 2008, pp. 63-72.
H.Wang, F. P.Nie, H.Huang, and F.Makedon, Fast nonnegative matrix tri-factorization for large-scale data co-clustering, in Proc. 22nd Int. Joint Conf. Artificial Intelligence, Barcelona, Spain, 2011, pp. 1553-1558.
C. H. Q.Ding, T.Li, and M. I.Jordan, Convex and semi-nonnegative matrix factorizations, IEEE Trans. Patt. Anal. Mach. Intell., vol. 32, no. 1, pp. 45-55, 2010.
H.Abe and H.Yadohisa, Orthogonal nonnegative matrix tri-factorization based on tweedie distributions, Adv. Data Anal. Classi., vol. 13, no. 4, pp. 825-853, 2019.
Liao X, Zheng D, Cao X. Coronavirus Pandemic Analysis Through Tripartite Graph Clustering in Online Social Networks. Big Data Mining and Analytics, 2021, 4(4): 242-251. https://doi.org/10.26599/BDMA.2021.9020010
The articles published in this open access journal are distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/).
10.26599/BDMA.2021.9020010.F001
Examples of multipartite graph with different k.
10.26599/BDMA.2021.9020010.F002
An example of tripartite graph co-clustering problem.
10.26599/BDMA.2021.9020010.F003
An example of tripartite graph in Twitter.
10.26599/BDMA.2021.9020010.F004
An overview of the TGC-PDA framework.
10.26599/BDMA.2021.9020010.F005
Build the user-topic bipartite by removing the tweet nodes of the tripartite graph and leveraging the tweet nodes as the connection for user and topic nodes.
6 Experimental Result
In this section, we analyze our experimental results and the performance of TGC-PDA. We also compare NMFRU with the well-known clustering methods, such as Kmeans, NMF, and the commonly used variants, including Semi-NMF (SNMF)[
43
] and Orthogonal NMTF (ONMTF)[
44
].
6.1 Dataset
We evaluate the performance of TGC-PDA with real Twitter dataset about “Covid-19” collected between Feb. 15th, 2020 and Sep. 30th 2020. To get the tweet data, we wrote a python program to crawl the tweets and the users who liked them. Multiple hashtag keywords, such as #COVID19, #coronavirus, #covid, covid pandemic, and #COVID20 are used to ensure we can get a large dataset. Since the free Twitter API we use has rate limits and it restricts the number of retrieved tweets during each login access, we have to crawl the data for several months. After removing the duplicate and non-English posts, we obtain 18 327 tweets, with 752 649 users who interacted with the tweets. Some users only interacted with one tweet in our dataset, which are identified as “less interactive” users and excluded. After the data cleanup, we have 301 982 users left.
6.2 Experimental setup
As all the clustering methods (i.e., Kmeans, NMF, SNMF, ONMTF, and our NMFRU) have one or more parameters to be tuned, to make the comparison fair, we run these methods under different parameters and choose the best result for each algorithm. In addition, we set the number of clusters as the true number of classes for all clustering algorithms on the dataset. In specific, for Kmeans and NMF algorithms, the hyperparameter is (number of clusters). If is given, no other parameters would be needed. In NMFR, we have two hyperparameters: and . To find a proper value for these parameters, we plot a loss-value curve, with value ranging from 0.1 to 1000. Then, the and values can be found by scanning the plot. Since our data size is relatively large and cannot be completely labeled manually, we randomly choose 5% of the data to label and use the result tested by sample data as the framework result.
6.3 Evaluation metrics
To evaluate the clustering result, we use the widely used standard metrics, including the clustering accuracy, cluster purity, and Normalized Mutual Information (NMI).
For the clustering accuracy, we compare the outputted clusters with the ground truth labelled data . The accuracy is defined as follows:
where is total number of data samples, is the delta function, in which the value equals one if , and equals zero otherwise. The is a mapping function that maps each cluster label to the same label in the ground truth data. We can use Kuhn-Munkres algorithm to find the best mappings[
45
].
For the cluster purity, we compare the cluster output with the ground truth labelled data . The purity of the cluster result is calculated,
where is the number of data samples. A perfect clustering result has a purity of one, and a bad one has the purity value close to zero.
For the NMI, we compare the cluster output with the ground truth labelled data . The NMI is defined as
where represents the entropy, and denotes the mutual information between and . A higher NMI value means the better clustering result.
To obtain a less biased estimation of the framework, we run NMFRU algorithms twenty times and take the average result for each model.
6.4 Results and discussion
Table 2
shows the comparison between NMFRU and several baseline models, such as Kmeans, NMF, SNMF, and ONMTF. When applying these baseline models to our data, we do not embed the topic nodes to user nodes. Instead, we use the user and tweet bipartite graph to calculate the clustering result. The matrix form of the bipartite graph is that the columns and rows correspond to the two sets of vertices, with each entry corresponding to an edge between a column and a row. From
Table 2
, we can see that NMFRU achieves the best performance in terms of accuracy, purity, and NMI. This is because our bipartite graph is created based on our tripartite graph model, and it embedded more information than the plain bipartite graph. We also utilize the tri-factorization and locality preserved schemes, which can further improve the performance.
10.26599/BDMA.2021.9020010.T002
Performance results of classifiers.
Method
Accuracy
Purity
NMI
Kmeans
0.613
0.549
0.513
NMF
0.583
0.536
0.493
SNMF
0.627
0.562
0.534
ONMTF
0.674
0.578
0.557
NMFRU
0.706
0.617
0.621
We study the average convergence time of our framework in
Fig. 6
. When the number of iterations is around 23, our framework tends to converge with a total loss of 2, which shows that the calculation of NMFRU is fast. Meanwhile, when comparing the convergence time by the different baseline methods in
Fig. 7
, we can see that NMFRU is slower than Kmeans but faster than other baseline clustering methods. It is because we do fewer matrix multiplication operations in NMFRU, hence saving some running time. Therefore, TGC-PDA that utilizes NMFRU as the core clustering algorithm can be used for a large dataset.
10.26599/BDMA.2021.9020010.F006
Total loss with different numbers of iterations.
10.26599/BDMA.2021.9020010.F007
Convergence time of methods.
As for the polarity of the communities,
Table 3
shows the largest ten communities with its polarity ratio. From
Table 3
, we find that the neutral ratio is quite high among all topics. In order to figure out the rationale here, we manually examined 1000 posts and found that there are lots of media or government agencies (e.g., CNN and CDC) that use Twitter to publish real-time news and the latest policy. These tweets tend to be retweeted many times. Obviously, such tweets are more likely to be considered neutral.
10.26599/BDMA.2021.9020010.T003
Largest ten communities with its polarity ratio. (%)
Keyword
Positive
Neutral
Negative
marketcrash2021
18.2
48.7
33.1
maskshortage
14.1
41.6
44.3
death
4.1
73.1
22.8
NYbreak
12.2
57.5
30.3
antibody
30.5
41.3
28.2
stimulus
32.6
41.7
25.7
testing
32.7
38.9
28.4
vaccine
20.4
61.2
18.4
symptoms
26.3
48.9
24.8
stayathome
23.6
51.8
24.6
10.26599/BDMA.2021.9020010.F006
Total loss with different numbers of iterations.
10.26599/BDMA.2021.9020010.F007
Convergence time of methods.