Abstract The main aim of this article is to explore the real-life problem-solving potential of the proposed Lévy flight-based chaotic gravitational search algorithm (LCGSA) for the minimization of engineering design variables of speed reducer design (SRD), three bar truss design (TBTD), and hydrodynamic thrust bearing design (HTBD) problems. In LCGSA, the diversification of the search space is carried out by Lévy flight distribution. Simultaneously, chaotic maps have been utilized for the intensification of the candidate solutions towards the global optimum. Moreover, the penalty function method has been used to deal with the non-linear and fractional design constraints. The investigation of experimental outcomes has been performed through various performance metrics like statistical measures, run time analysis, convergence rate, and box plot analysis. Moreover, statistical verification of experimental results is carried out using a signed Wilcoxon rank-sum test. Furthermore, eleven heuristic algorithms were employed for comparative analysis of the simulation results. The simulation outcomes clearly show that LCGSA provides better values for TBTD and HTBD benchmarks than standard GSA and most of the competing algorithms. Besides, all the participating algorithms, including LCGSA, have the same results for the SRD problem. On the qualitative side, LCGSA has successfully resolved entrapment in local minima and convergence issues of standard GSA.
{"title":"Lévy flight and chaos theory-based gravitational search algorithm for mechanical and structural engineering design optimization","authors":"Sajad Ahmad Rather, Perumal Shanthi Bala","doi":"10.1515/comp-2020-0223","DOIUrl":"https://doi.org/10.1515/comp-2020-0223","url":null,"abstract":"Abstract The main aim of this article is to explore the real-life problem-solving potential of the proposed Lévy flight-based chaotic gravitational search algorithm (LCGSA) for the minimization of engineering design variables of speed reducer design (SRD), three bar truss design (TBTD), and hydrodynamic thrust bearing design (HTBD) problems. In LCGSA, the diversification of the search space is carried out by Lévy flight distribution. Simultaneously, chaotic maps have been utilized for the intensification of the candidate solutions towards the global optimum. Moreover, the penalty function method has been used to deal with the non-linear and fractional design constraints. The investigation of experimental outcomes has been performed through various performance metrics like statistical measures, run time analysis, convergence rate, and box plot analysis. Moreover, statistical verification of experimental results is carried out using a signed Wilcoxon rank-sum test. Furthermore, eleven heuristic algorithms were employed for comparative analysis of the simulation results. The simulation outcomes clearly show that LCGSA provides better values for TBTD and HTBD benchmarks than standard GSA and most of the competing algorithms. Besides, all the participating algorithms, including LCGSA, have the same results for the SRD problem. On the qualitative side, LCGSA has successfully resolved entrapment in local minima and convergence issues of standard GSA.","PeriodicalId":43014,"journal":{"name":"Open Computer Science","volume":"11 1","pages":"509 - 529"},"PeriodicalIF":1.5,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43649917","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Abstract Lung cancer is considered to be one of the deadly diseases that threaten the survival of human beings. It is a challenging task to identify lung cancer in its early stage from the medical images because of the ambiguity in the lung regions. This paper proposes a new architecture to detect lung cancer obtained from the CT images. The proposed architecture has a three-stream network to extract the manual and automated features from the images. Among these three streams, automated feature extraction as well as the classification is done using residual deep neural network and custom deep neural network. Whereas the manual features are the handcrafted features obtained using high and low-frequency sub-bands in the frequency domain that are classified using a Support Vector Machine Classifier. This makes the architecture robust enough to capture all the important features required to classify lung cancer from the input image. Hence, there is no chance of missing feature information. Finally, all the obtained prediction scores are combined by weighted based fusion. The experimental results show 98.2% classification accuracy which is relatively higher in comparison to other existing methods.
{"title":"Three Stream Network Model for Lung Cancer Classification in the CT Images","authors":"T. Arumuga Maria Devi, V. I. Mebin Jose","doi":"10.1515/comp-2020-0145","DOIUrl":"https://doi.org/10.1515/comp-2020-0145","url":null,"abstract":"Abstract Lung cancer is considered to be one of the deadly diseases that threaten the survival of human beings. It is a challenging task to identify lung cancer in its early stage from the medical images because of the ambiguity in the lung regions. This paper proposes a new architecture to detect lung cancer obtained from the CT images. The proposed architecture has a three-stream network to extract the manual and automated features from the images. Among these three streams, automated feature extraction as well as the classification is done using residual deep neural network and custom deep neural network. Whereas the manual features are the handcrafted features obtained using high and low-frequency sub-bands in the frequency domain that are classified using a Support Vector Machine Classifier. This makes the architecture robust enough to capture all the important features required to classify lung cancer from the input image. Hence, there is no chance of missing feature information. Finally, all the obtained prediction scores are combined by weighted based fusion. The experimental results show 98.2% classification accuracy which is relatively higher in comparison to other existing methods.","PeriodicalId":43014,"journal":{"name":"Open Computer Science","volume":"11 1","pages":"251 - 261"},"PeriodicalIF":1.5,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1515/comp-2020-0145","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"66887241","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Abstract After a voice control system transforms audio input into a natural language sentence, its main purpose is to map this sentence to a specific action in the API (application programming interface) that should be performed. This mapping is usually specified after the API is already designed. In this paper, we show how an API can be designed with voice control in mind, which makes this mapping natural. The classes, methods, and parameters in the source code are named and typed according to the terms expected in the natural language commands. When this is insufficient, annotations (attribute-oriented programming) are used to define synonyms, string-to-object maps, or other properties. We also describe the mapping process and present a preliminary implementation called VCMapper. In its evaluation on a third-party dataset, it was successfully used to map all the sentences, while a large portion of the mapping was performed using only naming and typing conventions.
{"title":"Natural mapping between voice commands and APIs","authors":"Matúš Sulír, J. Porubän","doi":"10.1515/comp-2020-0125","DOIUrl":"https://doi.org/10.1515/comp-2020-0125","url":null,"abstract":"Abstract After a voice control system transforms audio input into a natural language sentence, its main purpose is to map this sentence to a specific action in the API (application programming interface) that should be performed. This mapping is usually specified after the API is already designed. In this paper, we show how an API can be designed with voice control in mind, which makes this mapping natural. The classes, methods, and parameters in the source code are named and typed according to the terms expected in the natural language commands. When this is insufficient, annotations (attribute-oriented programming) are used to define synonyms, string-to-object maps, or other properties. We also describe the mapping process and present a preliminary implementation called VCMapper. In its evaluation on a third-party dataset, it was successfully used to map all the sentences, while a large portion of the mapping was performed using only naming and typing conventions.","PeriodicalId":43014,"journal":{"name":"Open Computer Science","volume":"11 1","pages":"135 - 145"},"PeriodicalIF":1.5,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1515/comp-2020-0125","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46725617","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Marko Vještica, Vladimir Dimitrieski, M. Pisarić, Slavica Kordić, S. Ristić, I. Luković
Abstract Technological advances and increasing customer need for highly customized products have triggered a fourth industrial revolution. A digital revolution in the manufacturing industry is enforced by introducing smart devices and knowledge bases to form intelligent manufacturing information systems. One of the goals of the digital revolution is to allow flexibility of smart factories by automating shop floor changes based on the changes in input production processes and ordered products. In order to make this possible, a formal language to describe production processes is needed, together with a code generator for its models and an engine to execute the code on smart devices. Existing process modeling languages are not usually tailored to model production processes, especially if models are needed for automatic code generation. In this paper we propose a research on Industry 4.0 manufacturing using a Domain-Specific Modeling Language (DSML) within a Model-Driven Software Development (MDSD) approach to model production processes. The models would be used to generate instructions to smart devices and human workers, and gather a feedback from them during the process execution. A pilot comparative analysis of three modeling languages that are commonly used for process modeling is given with the goal of identifying supported modeling concepts, good practices and usage patterns.
{"title":"Towards a Formal Specification of Production Processes Suitable for Automatic Execution","authors":"Marko Vještica, Vladimir Dimitrieski, M. Pisarić, Slavica Kordić, S. Ristić, I. Luković","doi":"10.1515/comp-2020-0200","DOIUrl":"https://doi.org/10.1515/comp-2020-0200","url":null,"abstract":"Abstract Technological advances and increasing customer need for highly customized products have triggered a fourth industrial revolution. A digital revolution in the manufacturing industry is enforced by introducing smart devices and knowledge bases to form intelligent manufacturing information systems. One of the goals of the digital revolution is to allow flexibility of smart factories by automating shop floor changes based on the changes in input production processes and ordered products. In order to make this possible, a formal language to describe production processes is needed, together with a code generator for its models and an engine to execute the code on smart devices. Existing process modeling languages are not usually tailored to model production processes, especially if models are needed for automatic code generation. In this paper we propose a research on Industry 4.0 manufacturing using a Domain-Specific Modeling Language (DSML) within a Model-Driven Software Development (MDSD) approach to model production processes. The models would be used to generate instructions to smart devices and human workers, and gather a feedback from them during the process execution. A pilot comparative analysis of three modeling languages that are commonly used for process modeling is given with the goal of identifying supported modeling concepts, good practices and usage patterns.","PeriodicalId":43014,"journal":{"name":"Open Computer Science","volume":"11 1","pages":"161 - 179"},"PeriodicalIF":1.5,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1515/comp-2020-0200","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48936533","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Abstract Segmentation of brain image should be done accurately as it can help to predict deadly brain tumor disease so that it can be possible to control the malicious segments of brain image if known beforehand. The accuracy of the brain tumor analysis can be enhanced through the brain tumor segmentation procedure. Earlier DCNN models do not consider the weights as of learning instances which may decrease accuracy levels of the segmentation procedure. Considering the above point, we have suggested a framework for optimizing the network parameters such as weight and bias vector of DCNN models using swarm intelligent based algorithms like Genetic Algorithm (GA), Particle Swarm Optimization (PSO), Gray Wolf Optimization (GWO) and Whale Optimization Algorithm (WOA). The simulation results reveals that the WOA optimized DCNN segmentation model is outperformed than other three optimization based DCNN models i.e., GA-DCNN, PSO-DCNN, GWO-DCNN.
{"title":"Segmentation of MRI Brain Tumor Image using Optimization based Deep Convolutional Neural networks (DCNN)","authors":"P. K. Mishra, S. Satapathy, M. Rout","doi":"10.1515/comp-2020-0166","DOIUrl":"https://doi.org/10.1515/comp-2020-0166","url":null,"abstract":"Abstract Segmentation of brain image should be done accurately as it can help to predict deadly brain tumor disease so that it can be possible to control the malicious segments of brain image if known beforehand. The accuracy of the brain tumor analysis can be enhanced through the brain tumor segmentation procedure. Earlier DCNN models do not consider the weights as of learning instances which may decrease accuracy levels of the segmentation procedure. Considering the above point, we have suggested a framework for optimizing the network parameters such as weight and bias vector of DCNN models using swarm intelligent based algorithms like Genetic Algorithm (GA), Particle Swarm Optimization (PSO), Gray Wolf Optimization (GWO) and Whale Optimization Algorithm (WOA). The simulation results reveals that the WOA optimized DCNN segmentation model is outperformed than other three optimization based DCNN models i.e., GA-DCNN, PSO-DCNN, GWO-DCNN.","PeriodicalId":43014,"journal":{"name":"Open Computer Science","volume":"11 1","pages":"380 - 390"},"PeriodicalIF":1.5,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1515/comp-2020-0166","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42794895","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Abstract The information transmitted in IoT is susceptible to affect the user’s privacy, and hence the information ought to be transmitted securely. The conventional method to assure integrity, confidentiality, and non-repudiation is to first sign the message and then encrypt it. Signcryption is a technique where the signature and the encryption are performed in a single round. The current Signcryption system uses traditional cryptographic approaches that are overloaded for IoT, as it consists of resource-constrained devices and uses the weak session key to encrypt the data. We propose a hybrid Signcryption scheme that employs PRESENT, a lightweight block cipher algorithm to encrypt the data, and the session key is encrypted by ECC. The time taken to signcrypt the proposed Signcryption is better when compared to current Signcryption techniques, as it deploys lightweight cryptography techniques that are devoted to resource-constrained devices.
{"title":"Hybrid lightweight Signcryption scheme for IoT","authors":"M. Sruthi, R. Rajasekaran","doi":"10.1515/comp-2020-0105","DOIUrl":"https://doi.org/10.1515/comp-2020-0105","url":null,"abstract":"Abstract The information transmitted in IoT is susceptible to affect the user’s privacy, and hence the information ought to be transmitted securely. The conventional method to assure integrity, confidentiality, and non-repudiation is to first sign the message and then encrypt it. Signcryption is a technique where the signature and the encryption are performed in a single round. The current Signcryption system uses traditional cryptographic approaches that are overloaded for IoT, as it consists of resource-constrained devices and uses the weak session key to encrypt the data. We propose a hybrid Signcryption scheme that employs PRESENT, a lightweight block cipher algorithm to encrypt the data, and the session key is encrypted by ECC. The time taken to signcrypt the proposed Signcryption is better when compared to current Signcryption techniques, as it deploys lightweight cryptography techniques that are devoted to resource-constrained devices.","PeriodicalId":43014,"journal":{"name":"Open Computer Science","volume":"11 1","pages":"391 - 398"},"PeriodicalIF":1.5,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1515/comp-2020-0105","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42660959","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Abstract Retinal image analysis is one of the important diagnosis methods in modern ophthalmology because eye information is present in the retina. The image acquisition process may have some effects and can affect the quality of the image. This can be improved by better image enhancement techniques combined with the computer-aided diagnosis system. Deep learning is one of the important computational application techniques used for a medical imaging application. The main aim of this article is to find the best enhancement techniques for the identification of diabetic retinopathy (DR) and are tested with the commonly used deep learning techniques, and the performances are measured. In this article, the input image is taken from the Indian-based database named as Indian Diabetic Retinopathy Image Dataset, and 13 filters are used including smoothing and sharpening filters for enhancing the images. Then, the quality of the enhancement techniques is compared using performance metrics and better results are obtained for Median, Gaussian, Bilateral, Wiener, and partial differential equation filters and are combined for improving the enhancement of images. The output images from all the enhanced filters are given as the convolutional neural network input and the results are compared to find the better enhancement method.
{"title":"Convolutional neural-network-based classification of retinal images with different combinations of filtering techniques","authors":"Asha Gnana Priya Henry, Anitha Jude","doi":"10.1515/comp-2020-0177","DOIUrl":"https://doi.org/10.1515/comp-2020-0177","url":null,"abstract":"Abstract Retinal image analysis is one of the important diagnosis methods in modern ophthalmology because eye information is present in the retina. The image acquisition process may have some effects and can affect the quality of the image. This can be improved by better image enhancement techniques combined with the computer-aided diagnosis system. Deep learning is one of the important computational application techniques used for a medical imaging application. The main aim of this article is to find the best enhancement techniques for the identification of diabetic retinopathy (DR) and are tested with the commonly used deep learning techniques, and the performances are measured. In this article, the input image is taken from the Indian-based database named as Indian Diabetic Retinopathy Image Dataset, and 13 filters are used including smoothing and sharpening filters for enhancing the images. Then, the quality of the enhancement techniques is compared using performance metrics and better results are obtained for Median, Gaussian, Bilateral, Wiener, and partial differential equation filters and are combined for improving the enhancement of images. The output images from all the enhanced filters are given as the convolutional neural network input and the results are compared to find the better enhancement method.","PeriodicalId":43014,"journal":{"name":"Open Computer Science","volume":"11 1","pages":"480 - 490"},"PeriodicalIF":1.5,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49158105","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Abstract In this work we have proposed a model for Citizen Profiling. It uses veillance (Surveillance and Sousveillance) for data acquisition. For representation of Citizen Profile Temporal Knowledge Graph has been used through which we can answer semantic queries. Previously, most of the work lacks representation of Citizen Profile and have used surveillance for data acquisition. Our contribution is towards enriching the data acquisition process by adding sousveillance mechanism and facilitating semantic queries through representation of Citizen Profiles using Temporal Knowledge Graphs. Our proposed solution is storage efficient as we have only stored data logs for Citizen Profiling instead of storing images, audio, and video for profiling purposes. Our proposed system can be extended to Smart City, Smart Traffic Management, Workplace profiling etc. Agent based mechanism can be used for data acquisition where each Citizen has its own agent. Another improvement can be to incorporate a decentralized version of database for maintaining Citizen profile.
在本文中,我们提出了一个公民剖析模型。它使用监控(Surveillance and soussurveillance)来获取数据。对于公民档案的表示,我们使用时态知识图来回答语义查询。以前,大多数工作缺乏公民档案的代表,并且使用监视来获取数据。我们的贡献是通过增加监控机制和通过使用时态知识图表示公民档案来促进语义查询,从而丰富数据获取过程。我们提出的解决方案存储效率高,因为我们只为Citizen Profiling存储数据日志,而不是为Profiling目的存储图像、音频和视频。我们提出的系统可以扩展到智慧城市,智能交通管理,工作场所分析等。基于代理的机制可用于数据获取,其中每个Citizen都有自己的代理。另一个改进可以是合并一个分散版本的数据库来维护公民档案。
{"title":"Towards the Modelling of Veillance based Citizen Profiling using Knowledge Graphs","authors":"Siraj Munir, S. I. Jami, Shaukat Wasi","doi":"10.1515/comp-2020-0209","DOIUrl":"https://doi.org/10.1515/comp-2020-0209","url":null,"abstract":"Abstract In this work we have proposed a model for Citizen Profiling. It uses veillance (Surveillance and Sousveillance) for data acquisition. For representation of Citizen Profile Temporal Knowledge Graph has been used through which we can answer semantic queries. Previously, most of the work lacks representation of Citizen Profile and have used surveillance for data acquisition. Our contribution is towards enriching the data acquisition process by adding sousveillance mechanism and facilitating semantic queries through representation of Citizen Profiles using Temporal Knowledge Graphs. Our proposed solution is storage efficient as we have only stored data logs for Citizen Profiling instead of storing images, audio, and video for profiling purposes. Our proposed system can be extended to Smart City, Smart Traffic Management, Workplace profiling etc. Agent based mechanism can be used for data acquisition where each Citizen has its own agent. Another improvement can be to incorporate a decentralized version of database for maintaining Citizen profile.","PeriodicalId":43014,"journal":{"name":"Open Computer Science","volume":"11 1","pages":"294 - 304"},"PeriodicalIF":1.5,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1515/comp-2020-0209","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48893442","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Abstract A numerical method for investigating k-coverings of a convex bounded set with circles of two given radii is proposed. Cases with constraints on the distances between the covering circle centers are considered. An algorithm for finding an approximate number of such circles and the arrangement of their centers is described. For certain specific cases, approximate lower bounds of the density of the k-covering of the given domain are found. We use either 0–1 linear programming or general integer linear programming models. Numerical results demonstrating the effectiveness of the proposed methods are presented.
{"title":"Optimization of a k-covering of a bounded set with circles of two given radii","authors":"A. Khorkov, Sh. I. Galiev","doi":"10.1515/comp-2020-0219","DOIUrl":"https://doi.org/10.1515/comp-2020-0219","url":null,"abstract":"Abstract A numerical method for investigating k-coverings of a convex bounded set with circles of two given radii is proposed. Cases with constraints on the distances between the covering circle centers are considered. An algorithm for finding an approximate number of such circles and the arrangement of their centers is described. For certain specific cases, approximate lower bounds of the density of the k-covering of the given domain are found. We use either 0–1 linear programming or general integer linear programming models. Numerical results demonstrating the effectiveness of the proposed methods are presented.","PeriodicalId":43014,"journal":{"name":"Open Computer Science","volume":"11 1","pages":"232 - 240"},"PeriodicalIF":1.5,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1515/comp-2020-0219","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49013215","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Cuiping Long, Rashmi Agrawal, Ha Quoc Trung, H. Pham
Abstract The next generation of E-Government and healthcare has the potential to increase the more intelligent governance with improvements in transparency, accountability, efficiency, and effectiveness. It enables organizations to use the benefits of information via big data analysis to settle the difficulties effectively. Big Data has emerged which plays a significant role in many sectors around the world. Global trends in taking advantage of the benefits from big data are considered with an overview of the US, European Union, and several developing countries. To deeply understand the utilization of big data in several domains, this study has presented a brief survey of key concepts (such as IoT-enabled data, blockchain-enabled data, and intelligent systems data) to deeply understand the utilization of big data in several domains. Our analysis sets out also the similarities and differences in these concepts. We have also surveyed state-of-the-art technologies including cloud computing, multi-cloud, webservice, and microservice which are used to exploit potential benefits of big data analytics. Furthermore, some typical big data frameworks are surveyed and a big data framework for E-Government is also proposed. Open research questions and challenges are highlighted (for researchers and developers) following our review. Our goal in presenting the novel concepts presented in this article is to promote creative ideas in the research endeavor to perform efficaciously next-generation E-Government in the context of Industry 4.0.
{"title":"A big data framework for E-Government in Industry 4.0","authors":"Cuiping Long, Rashmi Agrawal, Ha Quoc Trung, H. Pham","doi":"10.1515/comp-2020-0191","DOIUrl":"https://doi.org/10.1515/comp-2020-0191","url":null,"abstract":"Abstract The next generation of E-Government and healthcare has the potential to increase the more intelligent governance with improvements in transparency, accountability, efficiency, and effectiveness. It enables organizations to use the benefits of information via big data analysis to settle the difficulties effectively. Big Data has emerged which plays a significant role in many sectors around the world. Global trends in taking advantage of the benefits from big data are considered with an overview of the US, European Union, and several developing countries. To deeply understand the utilization of big data in several domains, this study has presented a brief survey of key concepts (such as IoT-enabled data, blockchain-enabled data, and intelligent systems data) to deeply understand the utilization of big data in several domains. Our analysis sets out also the similarities and differences in these concepts. We have also surveyed state-of-the-art technologies including cloud computing, multi-cloud, webservice, and microservice which are used to exploit potential benefits of big data analytics. Furthermore, some typical big data frameworks are surveyed and a big data framework for E-Government is also proposed. Open research questions and challenges are highlighted (for researchers and developers) following our review. Our goal in presenting the novel concepts presented in this article is to promote creative ideas in the research endeavor to perform efficaciously next-generation E-Government in the context of Industry 4.0.","PeriodicalId":43014,"journal":{"name":"Open Computer Science","volume":"11 1","pages":"461 - 479"},"PeriodicalIF":1.5,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49380930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}