Pub Date : 2021-01-01Epub Date: 2021-04-01DOI: 10.1007/s11227-021-03748-x
Xuefan Dong, Ying Lian, Yuxue Chi, Xianyi Tang, Yijun Liu
Based on the supernetwork theory, a two-step rumor detection model was proposed. The first step was the classification of users on the basis of user-based features. In the second step, non-user-based features, including psychology-based features, content-based features, and parts of supernetwork-based features, were used to detect rumors posted by different types of users. Four machine learning methods, namely, Naive Bayes, Neural Network, Support Vector Machine, and Logistic Regression, were applied to train the classifier. Four real cases and several assessment metrics were employed to verify the effectiveness of the proposed model. Performance of the model regarding early rumor detection was also evaluated by separating the datasets according to the posting time of posts. Results showed that this model exhibited better performance in rumor detection compared to five benchmark models, mainly owing to the application of the supernetwork theory and the two-step mechanism.
{"title":"A two-step rumor detection model based on the supernetwork theory about Weibo.","authors":"Xuefan Dong, Ying Lian, Yuxue Chi, Xianyi Tang, Yijun Liu","doi":"10.1007/s11227-021-03748-x","DOIUrl":"10.1007/s11227-021-03748-x","url":null,"abstract":"<p><p>Based on the supernetwork theory, a two-step rumor detection model was proposed. The first step was the classification of users on the basis of user-based features. In the second step, non-user-based features, including psychology-based features, content-based features, and parts of supernetwork-based features, were used to detect rumors posted by different types of users. Four machine learning methods, namely, Naive Bayes, Neural Network, Support Vector Machine, and Logistic Regression, were applied to train the classifier. Four real cases and several assessment metrics were employed to verify the effectiveness of the proposed model. Performance of the model regarding early rumor detection was also evaluated by separating the datasets according to the posting time of posts. Results showed that this model exhibited better performance in rumor detection compared to five benchmark models, mainly owing to the application of the supernetwork theory and the two-step mechanism.</p>","PeriodicalId":50034,"journal":{"name":"Journal of Supercomputing","volume":"77 10","pages":"12050-12074"},"PeriodicalIF":3.3,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8014906/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"25564589","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-01-01Epub Date: 2021-05-06DOI: 10.1007/s11227-021-03848-8
Vicente Matellán Olivera, José Luis González-Sánchez
{"title":"Thinking in Parallel: foreword.","authors":"Vicente Matellán Olivera, José Luis González-Sánchez","doi":"10.1007/s11227-021-03848-8","DOIUrl":"https://doi.org/10.1007/s11227-021-03848-8","url":null,"abstract":"","PeriodicalId":50034,"journal":{"name":"Journal of Supercomputing","volume":"77 12","pages":"13992-13994"},"PeriodicalIF":3.3,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1007/s11227-021-03848-8","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38968741","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-01-01Epub Date: 2021-05-05DOI: 10.1007/s11227-021-03838-w
Ishaani Priyadarshini, Chase Cotton
As the number of users getting acquainted with the Internet is escalating rapidly, there is more user-generated content on the web. Comprehending hidden opinions, sentiments, and emotions in emails, tweets, reviews, and comments is a challenge and equally crucial for social media monitoring, brand monitoring, customer services, and market research. Sentiment analysis determines the emotional tone behind a series of words may essentially be used to understand the attitude, opinions, and emotions of users. We propose a novel long short-term memory (LSTM)-convolutional neural networks (CNN)-grid search-based deep neural network model for sentiment analysis. The study considers baseline algorithms like convolutional neural networks, K-nearest neighbor, LSTM, neural networks, LSTM-CNN, and CNN-LSTM which have been evaluated using accuracy, precision, sensitivity, specificity, and F-1 score, on multiple datasets. Our results show that the proposed model based on hyperparameter optimization outperforms other baseline models with an overall accuracy greater than 96%.
{"title":"A novel LSTM-CNN-grid search-based deep neural network for sentiment analysis.","authors":"Ishaani Priyadarshini, Chase Cotton","doi":"10.1007/s11227-021-03838-w","DOIUrl":"https://doi.org/10.1007/s11227-021-03838-w","url":null,"abstract":"<p><p>As the number of users getting acquainted with the Internet is escalating rapidly, there is more user-generated content on the web. Comprehending hidden opinions, sentiments, and emotions in emails, tweets, reviews, and comments is a challenge and equally crucial for social media monitoring, brand monitoring, customer services, and market research. Sentiment analysis determines the emotional tone behind a series of words may essentially be used to understand the attitude, opinions, and emotions of users. We propose a novel long short-term memory (LSTM)-convolutional neural networks (CNN)-grid search-based deep neural network model for sentiment analysis. The study considers baseline algorithms like convolutional neural networks, <i>K</i>-nearest neighbor, LSTM, neural networks, LSTM-CNN, and CNN-LSTM which have been evaluated using accuracy, precision, sensitivity, specificity, and F-1 score, on multiple datasets. Our results show that the proposed model based on hyperparameter optimization outperforms other baseline models with an overall accuracy greater than 96%.</p>","PeriodicalId":50034,"journal":{"name":"Journal of Supercomputing","volume":"77 12","pages":"13911-13932"},"PeriodicalIF":3.3,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1007/s11227-021-03838-w","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38964576","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-01-01Epub Date: 2021-03-22DOI: 10.1007/s11227-021-03717-4
Tajinder Singh, Madhu Kumari
Gigantic growth of social media and unbeatable trend of progress in the direction of the web seeking user's interests have generated a storm of social text streams. Seeking information to know the phenomenon of various events in the early stages is quite interesting. Various kinds of social media live streams attract users to participate in real-time events to become a part of an immense crowd. However, the vast amount of text is present on social media, the unnecessary information bogs a social text stream filtering to extract the appropriate topics and events effectively. Therefore, detecting, classifying, and identifying burst events is quite challenging due to the sparse and noisy text of Twitter. The researchers' significant open challenges are the effective cleaning and profound representation of the text stream data. This research article's main contribution is to provide a detailed study and explore bursty event detection in the social text stream. Thus, this work's main motive is to present a concise approach that classifies and detects the event keywords and maintains the record of the event based on related features. These features permit the approach to successfully determine the booming pattern of events scrupulously at different time span. Experiments are conducted and compared with the state-of-the-art methods, which reveals that the proposed approach is proficient to detect valuable patterns of interest and also achieve better scoresto extract burst events on social media posted by various users.
{"title":"Burst: real-time events burst detection in social text stream.","authors":"Tajinder Singh, Madhu Kumari","doi":"10.1007/s11227-021-03717-4","DOIUrl":"https://doi.org/10.1007/s11227-021-03717-4","url":null,"abstract":"<p><p>Gigantic growth of social media and unbeatable trend of progress in the direction of the web seeking user's interests have generated a storm of social text streams. Seeking information to know the phenomenon of various events in the early stages is quite interesting. Various kinds of social media live streams attract users to participate in real-time events to become a part of an immense crowd. However, the vast amount of text is present on social media, the unnecessary information bogs a social text stream filtering to extract the appropriate topics and events effectively. Therefore, detecting, classifying, and identifying burst events is quite challenging due to the sparse and noisy text of Twitter. The researchers' significant open challenges are the effective cleaning and profound representation of the text stream data. This research article's main contribution is to provide a detailed study and explore bursty event detection in the social text stream. Thus, this work's main motive is to present a concise approach that classifies and detects the event keywords and maintains the record of the event based on related features. These features permit the approach to successfully determine the booming pattern of events scrupulously at different time span. Experiments are conducted and compared with the state-of-the-art methods, which reveals that the proposed approach is proficient to detect valuable patterns of interest and also achieve better scoresto extract burst events on social media posted by various users.</p>","PeriodicalId":50034,"journal":{"name":"Journal of Supercomputing","volume":"77 10","pages":"11228-11256"},"PeriodicalIF":3.3,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1007/s11227-021-03717-4","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"25537956","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-01-01Epub Date: 2020-05-20DOI: 10.1007/s11227-020-03311-0
Abhilasha Sharma, R G Sangeetha
Reliability analysis is one of the crucial issues for any scalable optical interconnection network. Torus is a highly scalable optical interconnect for data centre networks. The traditional torus network has XY routing algorithm. We have proposed a novel optimised routing algorithm. This paper focuses on the time-dependent and time-independent analysis for both terminal and broadcast reliabilities of the torus network using XY and optimised routing algorithm under various network sizes ( where ). The results are evaluated and compared considering nodes failures in MATLAB.
{"title":"Terminal and broadcast reliability analysis of direct 2-D symmetric torus network.","authors":"Abhilasha Sharma, R G Sangeetha","doi":"10.1007/s11227-020-03311-0","DOIUrl":"https://doi.org/10.1007/s11227-020-03311-0","url":null,"abstract":"<p><p>Reliability analysis is one of the crucial issues for any scalable optical interconnection network. Torus is a highly scalable optical interconnect for data centre networks. The traditional torus network has XY routing algorithm. We have proposed a novel optimised routing algorithm. This paper focuses on the time-dependent and time-independent analysis for both terminal and broadcast reliabilities of the torus network using XY and optimised routing algorithm under various network sizes ( <math><mrow><mi>N</mi> <mo>×</mo> <mi>N</mi></mrow> </math> where <math><mrow><mi>N</mi> <mo>=</mo> <mn>8</mn> <mo>,</mo> <mn>16</mn> <mo>,</mo> <mn>32</mn> <mo>,</mo> <mn>64</mn></mrow> </math> ). The results are evaluated and compared considering nodes failures in MATLAB.</p>","PeriodicalId":50034,"journal":{"name":"Journal of Supercomputing","volume":"77 2","pages":"1517-1536"},"PeriodicalIF":3.3,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1007/s11227-020-03311-0","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38297615","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-01-01Epub Date: 2020-09-28DOI: 10.1007/s11227-020-03438-0
Bence Ferdinandy, Ángel Manuel Guerrero-Higueras, Éva Verderber, Francisco Javier Rodríguez-Lera, Ádám Miklósi
Machine learning algorithms are becoming more and more useful in many fields of science, including many areas where computational methods are rarely used. High-performance Computing (HPC) is the most powerful solution to get the best results using these algorithms. HPC requires various skills to use. Acquiring this knowledge might be intimidating and take a long time for a researcher with small or no background in information and communications technologies (ICTs), even if the benefits of such knowledge is evident for the researcher. In this work, we aim to assess how a specific method of introducing HPC to such researchers enables them to start using HPC. We gave talks to two groups of non-ICT researchers that introduced basic concepts focusing on the necessary practical steps needed to use HPC on a specific cluster. We also offered hands-on trainings for one of the groups which aimed to guide participants through the first steps of using HPC. Participants filled out questionnaires partly based on Kirkpatrick's training evaluation model before and after the talk, and after the hands-on training. We found that the talk increased participants' self-reported likelihood of using HPC in their future research, but this was not significant for the group where participation was voluntary. On the contrary, very few researchers participated in the hands-on training, and for these participants neither the talk, nor the hands-on training changed their self-reported likelihood of using HPC in their future research. We argue that our findings show that academia and researchers would benefit from an environment that not only expects researchers to train themselves, but provides structural support for acquiring new skills.
{"title":"Exploratory study of introducing HPC to non-ICT researchers: institutional strategy is possibly needed for widespread adaption.","authors":"Bence Ferdinandy, Ángel Manuel Guerrero-Higueras, Éva Verderber, Francisco Javier Rodríguez-Lera, Ádám Miklósi","doi":"10.1007/s11227-020-03438-0","DOIUrl":"https://doi.org/10.1007/s11227-020-03438-0","url":null,"abstract":"<p><p>Machine learning algorithms are becoming more and more useful in many fields of science, including many areas where computational methods are rarely used. High-performance Computing (HPC) is the most powerful solution to get the best results using these algorithms. HPC requires various skills to use. Acquiring this knowledge might be intimidating and take a long time for a researcher with small or no background in information and communications technologies (ICTs), even if the benefits of such knowledge is evident for the researcher. In this work, we aim to assess how a specific method of introducing HPC to such researchers enables them to start using HPC. We gave talks to two groups of non-ICT researchers that introduced basic concepts focusing on the necessary practical steps needed to use HPC on a specific cluster. We also offered hands-on trainings for one of the groups which aimed to guide participants through the first steps of using HPC. Participants filled out questionnaires partly based on Kirkpatrick's training evaluation model before and after the talk, and after the hands-on training. We found that the talk increased participants' self-reported likelihood of using HPC in their future research, but this was not significant for the group where participation was voluntary. On the contrary, very few researchers participated in the hands-on training, and for these participants neither the talk, nor the hands-on training changed their self-reported likelihood of using HPC in their future research. We argue that our findings show that academia and researchers would benefit from an environment that not only expects researchers to train themselves, but provides structural support for acquiring new skills.</p>","PeriodicalId":50034,"journal":{"name":"Journal of Supercomputing","volume":"77 5","pages":"4317-4331"},"PeriodicalIF":3.3,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1007/s11227-020-03438-0","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38454072","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-01-01Epub Date: 2021-01-04DOI: 10.1007/s11227-020-03586-3
Shalini Ramanathan, Mohan Ramasundaram
In every field of life, advanced technology has become a rapid outcome, particularly in the medical field. The recent epidemic of the coronavirus disease 2019 (COVID-19) has promptly become outbreaks to identify early action from suspected cases at the primary stage over the risk prediction. It is overbearing to progress a control system that will locate the coronavirus. At present, the confirmation of COVID-19 infection by the ideal standard test of reverse transcription-polymerase chain reaction (rRT-PCR) by the extension of RNA viral, although it presents identified from deficiencies of long reversal time to generate results in 2-4 h of corona with a necessity of certified laboratories. In this proposed system, a machine learning (ML) algorithm is used to classify the textual clinical report into four classes by using the textual data mining method. The algorithm of the ensemble ML classifier has performed feature extraction using the advanced techniques of term frequency-inverse document frequency (TF/IDF) which is an effective information retrieval technique from the corona dataset. Humans get infected by coronaviruses in three ways: first, mild respiratory disease which is globally pandemic, and human coronaviruses are caused by HCoV-NL63, HCoV-OC43, HCoV-HKU1, and HCoV-229E; second, the zoonotic Middle East respiratory syndrome coronavirus (MERS-CoV); and finally, higher case casualty rate defined as severe acute respiratory syndrome coronavirus (SARS-CoV). By using the machine learning techniques, the three-way COVID-19 stages are classified by the extraction of the feature using the data retrieval process. The TF/IDF is used to measure and evaluate statistically the text data mining of COVID-19 patient's record list for classification and prediction of the coronavirus. This study established the feasibility of techniques to analyze blood tests and machine learning as an alternative to rRT-PCR for detecting the category of COVID-19-positive patients.
{"title":"Accurate computation: COVID-19 rRT-PCR positive test dataset using stages classification through textual big data mining with machine learning.","authors":"Shalini Ramanathan, Mohan Ramasundaram","doi":"10.1007/s11227-020-03586-3","DOIUrl":"10.1007/s11227-020-03586-3","url":null,"abstract":"<p><p>In every field of life, advanced technology has become a rapid outcome, particularly in the medical field. The recent epidemic of the coronavirus disease 2019 (COVID-19) has promptly become outbreaks to identify early action from suspected cases at the primary stage over the risk prediction. It is overbearing to progress a control system that will locate the coronavirus. At present, the confirmation of COVID-19 infection by the ideal standard test of reverse transcription-polymerase chain reaction (rRT-PCR) by the extension of RNA viral, although it presents identified from deficiencies of long reversal time to generate results in 2-4 h of corona with a necessity of certified laboratories. In this proposed system, a machine learning (ML) algorithm is used to classify the textual clinical report into four classes by using the textual data mining method. The algorithm of the ensemble ML classifier has performed feature extraction using the advanced techniques of term frequency-inverse document frequency (TF/IDF) which is an effective information retrieval technique from the corona dataset. Humans get infected by coronaviruses in three ways: first, mild respiratory disease which is globally pandemic, and human coronaviruses are caused by HCoV-NL63, HCoV-OC43, HCoV-HKU1, and HCoV-229E; second, the zoonotic Middle East respiratory syndrome coronavirus (MERS-CoV); and finally, higher case casualty rate defined as severe acute respiratory syndrome coronavirus (SARS-CoV). By using the machine learning techniques, the three-way COVID-19 stages are classified by the extraction of the feature using the data retrieval process. The TF/IDF is used to measure and evaluate statistically the text data mining of COVID-19 patient's record list for classification and prediction of the coronavirus. This study established the feasibility of techniques to analyze blood tests and machine learning as an alternative to rRT-PCR for detecting the category of COVID-19-positive patients.</p>","PeriodicalId":50034,"journal":{"name":"Journal of Supercomputing","volume":"77 7","pages":"7074-7088"},"PeriodicalIF":2.5,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7781398/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38801977","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2020-08-01DOI: 10.1007/S11227-019-02953-Z
P. Dhas, B. Gomathi
{"title":"A novel clustering algorithm by clubbing GHFCM and GWO for microarray gene data","authors":"P. Dhas, B. Gomathi","doi":"10.1007/S11227-019-02953-Z","DOIUrl":"https://doi.org/10.1007/S11227-019-02953-Z","url":null,"abstract":"","PeriodicalId":50034,"journal":{"name":"Journal of Supercomputing","volume":"30 1","pages":"5679-5693"},"PeriodicalIF":3.3,"publicationDate":"2020-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75190193","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}