首页 > 最新文献

2016 International Workshop on Big Data and Information Security (IWBIS)最新文献

英文 中文
Generalized learning vector quantization particle swarm optimization (GLVQ-PSO) FPGA implementation for real-time electrocardiogram 基于广义学习向量量化粒子群优化(GLVQ-PSO)的实时心电图FPGA实现
Pub Date : 2016-10-01 DOI: 10.1109/IWBIS.2016.7872897
Yulistiyan Wardhana, W. Jatmiko, M. F. Rachmadi
Cardiovascular system is the most important part of human body which has role as distribution system of Oxygen and body's wastes. To do the job, there are more than 60.000 miles of blood vessels participated which can caused a problem if one of them are being clogged. Unfortunately, the conditions of clogged blood vessels or diseases caused by cardiovascular malfunction could not be detected in a plain view. In this matter, we proposed a design of wearable device which can detect the conditions. The device is equipped with a newly neural network algorithm, GLVQ-PSO, which can give recommendation of the heart status based on learned data. After the research is conducted, the algorithm produce better accuracy than LVQ, GLVQ and FNGLVQ in the high level language implementation. Yet, GLVQ-PSO still has relatively worse performance in its FPGA implementation.
心血管系统是人体最重要的组成部分,具有氧气和人体废物的分配系统的作用。为了完成这项工作,有超过6万英里的血管参与其中,如果其中一条血管堵塞,就会产生问题。不幸的是,血管阻塞的情况或心血管功能障碍引起的疾病无法在普通视图中检测到。针对这一问题,我们提出了一种可穿戴设备的设计,该设备可以检测到这些情况。该装置配备了一种新的神经网络算法,GLVQ-PSO,可以根据学习到的数据给出心脏状态的推荐。经过研究,该算法在高级语言实现上的准确率优于LVQ、GLVQ和FNGLVQ。然而,GLVQ-PSO在其FPGA实现中仍然具有相对较差的性能。
{"title":"Generalized learning vector quantization particle swarm optimization (GLVQ-PSO) FPGA implementation for real-time electrocardiogram","authors":"Yulistiyan Wardhana, W. Jatmiko, M. F. Rachmadi","doi":"10.1109/IWBIS.2016.7872897","DOIUrl":"https://doi.org/10.1109/IWBIS.2016.7872897","url":null,"abstract":"Cardiovascular system is the most important part of human body which has role as distribution system of Oxygen and body's wastes. To do the job, there are more than 60.000 miles of blood vessels participated which can caused a problem if one of them are being clogged. Unfortunately, the conditions of clogged blood vessels or diseases caused by cardiovascular malfunction could not be detected in a plain view. In this matter, we proposed a design of wearable device which can detect the conditions. The device is equipped with a newly neural network algorithm, GLVQ-PSO, which can give recommendation of the heart status based on learned data. After the research is conducted, the algorithm produce better accuracy than LVQ, GLVQ and FNGLVQ in the high level language implementation. Yet, GLVQ-PSO still has relatively worse performance in its FPGA implementation.","PeriodicalId":193821,"journal":{"name":"2016 International Workshop on Big Data and Information Security (IWBIS)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121359324","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
The power of big data and algorithms for advertising and customer communication 大数据和算法在广告和客户沟通方面的力量
Pub Date : 2016-10-01 DOI: 10.1109/IWBIS.2016.7872882
Nico Neumann
Leveraging customer data in scale and often in real time has led to a new field called programmatic commerce — the use of data, automation and analytics to improve customer experiences and company performances. In particular in advertising and marketing, programmatic applications have become very popular because they allow personalization/ micro-targeting as well as easier media planning due to the rise of automated buying processes. In this review study, we will discuss the development of the new field around advertising and marketing technology and summarize present research efforts. In addition, some industry case studies will be shared to illustrate the power of the latest big-data and machine-learning applications for driving business outcomes.
大规模且经常是实时地利用客户数据,导致了一个被称为程序化商业的新领域——利用数据、自动化和分析来改善客户体验和公司绩效。特别是在广告和市场营销中,程序化应用程序变得非常受欢迎,因为它们允许个性化/微目标以及由于自动购买过程的兴起而更容易进行媒体规划。在这篇综述性研究中,我们将讨论围绕广告和营销技术的新领域的发展,并总结目前的研究成果。此外,还将分享一些行业案例研究,以说明最新的大数据和机器学习应用在推动业务成果方面的力量。
{"title":"The power of big data and algorithms for advertising and customer communication","authors":"Nico Neumann","doi":"10.1109/IWBIS.2016.7872882","DOIUrl":"https://doi.org/10.1109/IWBIS.2016.7872882","url":null,"abstract":"Leveraging customer data in scale and often in real time has led to a new field called programmatic commerce — the use of data, automation and analytics to improve customer experiences and company performances. In particular in advertising and marketing, programmatic applications have become very popular because they allow personalization/ micro-targeting as well as easier media planning due to the rise of automated buying processes. In this review study, we will discuss the development of the new field around advertising and marketing technology and summarize present research efforts. In addition, some industry case studies will be shared to illustrate the power of the latest big-data and machine-learning applications for driving business outcomes.","PeriodicalId":193821,"journal":{"name":"2016 International Workshop on Big Data and Information Security (IWBIS)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128403777","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Overview of research center for information technology innovation in Taiwan Academia Sinica 台湾中央研究院资讯科技创新研究中心概况
Pub Date : 2016-10-01 DOI: 10.1109/IWBIS.2016.7872881
Yennun Huang, Szu-Chuang Li
Founded in February 2007, the aim of the Research Center for Information Technology Innovation (CITI) at Academia Sinica is to integrate research and development efforts in information technologies by various organizations in Academia Sinica, and also to facilitate and leverage IT-related multidisciplinary research. As a integral part of CITI, Taiwan Information Security Center (TWISC) to conduct researches on security with funding support from Ministry of Science and Technology. TWISC serves as a platform for security experts from universities, research institutes and private sector to share information and to explore opportunities to collaborate. Its aim is to boost research and development activities and promote public awareness regarding information security. Its research topics cover data/ software/ hardware/ network security and security management. TWISC has become the hub of security research in Taiwan and have been making significant impact through publishing and creating of toolkits. Recently privacy also becomes one of the main focuses of TWISC. The research team at CITI, Academia has been working on a viable way to assess the disclosure risk of synthetic dataset. Preliminary research result will be presented in this paper.
中央研究院资讯科技创新研究中心(CITI)于2007年2月成立,旨在整合中央研究院各机构在资讯科技方面的研发工作,并促进资讯科技相关的多学科研究。台湾信息安全中心(TWISC)作为花旗集团的重要组成部分,在科技部的资助下开展安全研究。TWISC为来自大学、研究机构和私营部门的安全专家提供了一个分享信息和探索合作机会的平台。它的目的是促进研究和发展活动,并提高公众对信息安全的认识。其研究课题涵盖数据/软件/硬件/网络安全和安全管理。TWISC已成为台湾安全研究的中心,并通过发布和创建工具包产生了重大影响。最近,隐私也成为TWISC的主要关注点之一。花旗学院的研究团队一直在研究一种可行的方法来评估合成数据集的披露风险。本文将介绍初步的研究成果。
{"title":"Overview of research center for information technology innovation in Taiwan Academia Sinica","authors":"Yennun Huang, Szu-Chuang Li","doi":"10.1109/IWBIS.2016.7872881","DOIUrl":"https://doi.org/10.1109/IWBIS.2016.7872881","url":null,"abstract":"Founded in February 2007, the aim of the Research Center for Information Technology Innovation (CITI) at Academia Sinica is to integrate research and development efforts in information technologies by various organizations in Academia Sinica, and also to facilitate and leverage IT-related multidisciplinary research. As a integral part of CITI, Taiwan Information Security Center (TWISC) to conduct researches on security with funding support from Ministry of Science and Technology. TWISC serves as a platform for security experts from universities, research institutes and private sector to share information and to explore opportunities to collaborate. Its aim is to boost research and development activities and promote public awareness regarding information security. Its research topics cover data/ software/ hardware/ network security and security management. TWISC has become the hub of security research in Taiwan and have been making significant impact through publishing and creating of toolkits. Recently privacy also becomes one of the main focuses of TWISC. The research team at CITI, Academia has been working on a viable way to assess the disclosure risk of synthetic dataset. Preliminary research result will be presented in this paper.","PeriodicalId":193821,"journal":{"name":"2016 International Workshop on Big Data and Information Security (IWBIS)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116478668","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Design and implementation of merchant acquirer data warehouse at PT. XYZ 在PT. XYZ上设计和实现商户收单数据仓库
Pub Date : 2016-10-01 DOI: 10.1109/IWBIS.2016.7872888
Y. Ruldeviyani, Bofandra Mohammad
Merchant acquirer is a business of acquiring debit card, credit card, and prepaid card transactions using EDC (electronic payment terminals) in merchants. It is included in one of the top business priority areas in PT. XYZ. It is in the area of retail payments and deposits. It increases fees based incomes, cheap funds, and high yield loans. In order to improve its business performance, PT. XYZ needs the best strategy. However, a good strategic decision needs an adequate of useful information. Currently, information that is provided by reporting staffs involved many manual tasks in the process. In consequence, the data cannot be provided quickly, and it has some complexity limitation. As a solution for this problem, PT. XYZ needs a data warehouse for its merchant acquirer business. This research will focus on the design and the implementation of the data warehouse solution using a methodology that is developed by Ralph L. Kimball. Finally, data warehouse is developed which is suitable for merchant acquirer PT. XYZ's needs.
商户收单业务是指在商户中使用电子支付终端(EDC)获取借记卡、信用卡和预付卡交易的业务。它包含在PT. XYZ的最高业务优先领域之一。它属于零售支付和存款领域。它增加了收费收入、廉价资金和高收益贷款。为了提高业务绩效,PT. XYZ需要最佳战略。然而,一个好的战略决策需要足够的有用信息。目前,报告人员提供的信息在流程中涉及许多手工任务。因此,不能快速提供数据,并且具有一定的复杂性限制。作为这个问题的解决方案,PT. XYZ需要一个数据仓库来处理它的商业收购业务。本研究将集中于使用Ralph L. Kimball开发的方法的数据仓库解决方案的设计和实现。最后,开发了适合商户收购方PT. XYZ需求的数据仓库。
{"title":"Design and implementation of merchant acquirer data warehouse at PT. XYZ","authors":"Y. Ruldeviyani, Bofandra Mohammad","doi":"10.1109/IWBIS.2016.7872888","DOIUrl":"https://doi.org/10.1109/IWBIS.2016.7872888","url":null,"abstract":"Merchant acquirer is a business of acquiring debit card, credit card, and prepaid card transactions using EDC (electronic payment terminals) in merchants. It is included in one of the top business priority areas in PT. XYZ. It is in the area of retail payments and deposits. It increases fees based incomes, cheap funds, and high yield loans. In order to improve its business performance, PT. XYZ needs the best strategy. However, a good strategic decision needs an adequate of useful information. Currently, information that is provided by reporting staffs involved many manual tasks in the process. In consequence, the data cannot be provided quickly, and it has some complexity limitation. As a solution for this problem, PT. XYZ needs a data warehouse for its merchant acquirer business. This research will focus on the design and the implementation of the data warehouse solution using a methodology that is developed by Ralph L. Kimball. Finally, data warehouse is developed which is suitable for merchant acquirer PT. XYZ's needs.","PeriodicalId":193821,"journal":{"name":"2016 International Workshop on Big Data and Information Security (IWBIS)","volume":"83 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117257757","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Design DDoS attack detector using NTOPNG 利用NTOPNG设计DDoS攻击检测器
Pub Date : 2016-10-01 DOI: 10.1109/IWBIS.2016.7872903
G. Jati, Budi Hartadi, A. Putra, Fahri Nurul, M. Iqbal, S. Yazid
Distributed Denial of Service (DDoS) is one kind of attacks using multiple computers. An attacker would act as a fake service requester that drains resources in computer target. This makes the target cannot serve the real request service. Thus we need to develop DDoS detector system. The proposed system consists of traffic capture, packet analyzer, and packet displayer. The system utilizes Ntopng as main traffic analyzer. Detector system has to meet good standard in accuracy, sensitivity, and reliability. We evaluate the system using one of dangerous DDoS tool named Slowloris. The system can detect attacks and provide alerts to detector user. The system also can process all incoming packets with a small margin of error (0.76%).
分布式拒绝服务(DDoS)是一种使用多台计算机的攻击。攻击者将充当一个虚假的服务请求者,消耗目标计算机中的资源。这使得目标无法提供真正的请求服务。因此,我们需要开发DDoS检测系统。该系统由流量捕获、数据包分析和数据包显示三部分组成。系统采用Ntopng作为主要流量分析器。检测系统在精度、灵敏度、可靠性等方面都达到了较好的要求。我们使用一种名为Slowloris的危险DDoS工具对系统进行评估。该系统可以检测到攻击并向检测用户提供警报。系统还可以处理所有传入的数据包,误差很小(0.76%)。
{"title":"Design DDoS attack detector using NTOPNG","authors":"G. Jati, Budi Hartadi, A. Putra, Fahri Nurul, M. Iqbal, S. Yazid","doi":"10.1109/IWBIS.2016.7872903","DOIUrl":"https://doi.org/10.1109/IWBIS.2016.7872903","url":null,"abstract":"Distributed Denial of Service (DDoS) is one kind of attacks using multiple computers. An attacker would act as a fake service requester that drains resources in computer target. This makes the target cannot serve the real request service. Thus we need to develop DDoS detector system. The proposed system consists of traffic capture, packet analyzer, and packet displayer. The system utilizes Ntopng as main traffic analyzer. Detector system has to meet good standard in accuracy, sensitivity, and reliability. We evaluate the system using one of dangerous DDoS tool named Slowloris. The system can detect attacks and provide alerts to detector user. The system also can process all incoming packets with a small margin of error (0.76%).","PeriodicalId":193821,"journal":{"name":"2016 International Workshop on Big Data and Information Security (IWBIS)","volume":"76 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124411324","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Enhanced tele ECG system using Hadoop framework to deal with big data processing 增强远程心电系统采用Hadoop框架进行大数据处理
Pub Date : 2016-10-01 DOI: 10.1109/IWBIS.2016.7872900
M. A. Ma'sum, W. Jatmiko, H. Suhartanto
Indonesia has high mortality caused by cardiovascular diseases. To minimize the mortality, we build a tele-ecg system for heart diseases early detection and monitoring. In this research, the tele-ecg system was enhanced using Hadoop framework, in order to deal with big data processing. The system was build on cluster computer with 4 nodes. The server is able to handle 60 requests at the same time. The system can classify the ecg data using decision tree and random forest. The accuracy is 97.14% and 98,92% for decision tree and random forest respectively. Training process in random forest is faster than in decision tree, while testing process in decision tree is faster than in random forest.
印度尼西亚心血管疾病造成的死亡率很高。为了最大限度地降低死亡率,我们建立了心脏疾病早期检测和监测的远程心电系统。在本研究中,利用Hadoop框架对远程心电系统进行了增强,以处理大数据。该系统建立在4个节点的集群计算机上。服务器可以同时处理60个请求。该系统采用决策树和随机森林对心电数据进行分类。决策树和随机森林的准确率分别为97.14%和98.92%。随机森林中的训练过程比决策树中的训练过程快,决策树中的测试过程比随机森林中的测试过程快。
{"title":"Enhanced tele ECG system using Hadoop framework to deal with big data processing","authors":"M. A. Ma'sum, W. Jatmiko, H. Suhartanto","doi":"10.1109/IWBIS.2016.7872900","DOIUrl":"https://doi.org/10.1109/IWBIS.2016.7872900","url":null,"abstract":"Indonesia has high mortality caused by cardiovascular diseases. To minimize the mortality, we build a tele-ecg system for heart diseases early detection and monitoring. In this research, the tele-ecg system was enhanced using Hadoop framework, in order to deal with big data processing. The system was build on cluster computer with 4 nodes. The server is able to handle 60 requests at the same time. The system can classify the ecg data using decision tree and random forest. The accuracy is 97.14% and 98,92% for decision tree and random forest respectively. Training process in random forest is faster than in decision tree, while testing process in decision tree is faster than in random forest.","PeriodicalId":193821,"journal":{"name":"2016 International Workshop on Big Data and Information Security (IWBIS)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128787211","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Big sensor-generated data streaming using Kafka and Impala for data storage in Wireless Sensor Network for CO2 monitoring 大传感器生成的数据流使用Kafka和Impala在无线传感器网络中存储数据,用于二氧化碳监测
Pub Date : 2016-10-01 DOI: 10.1109/IWBIS.2016.7872896
Rindra Wiska, Novian Habibie, A. Wibisono, W. S. Nugroho, P. Mursanto
Wireless Sensor Network (WSN) is a system that have a capability to conduct data acquisition and monitoring in a wide sampling area for a long time. However, because of its big-scale monitoring, amount of data accumulated from WSN is very huge. Conventional database system may not be able to handle its big amount of data. To overcome that, big data approach is used for an alternative data storage system and data analysis process. This research developed a WSN system for CO2 monitoring using Kafka and Impala to distribute a huge amount of data. Sensor nodes gather data and accumulated in temporary storage then streamed via Kafka platform to be stored into Impala database. System tested with data gathered from our-own made sensor nodes and give a good performance.
无线传感器网络(WSN)是一种能够在大采样范围内进行长时间数据采集和监测的系统。然而,由于其监测的规模较大,从无线传感器网络中积累的数据量非常巨大。传统的数据库系统可能无法处理其大量的数据。为了克服这一点,大数据方法被用于替代数据存储系统和数据分析过程。本研究利用Kafka和Impala分发大量数据,开发了一个用于CO2监测的WSN系统。传感器节点收集数据并在临时存储中积累,然后通过Kafka平台流式存储到Impala数据库中。用自制传感器节点采集的数据对系统进行了测试,取得了良好的性能。
{"title":"Big sensor-generated data streaming using Kafka and Impala for data storage in Wireless Sensor Network for CO2 monitoring","authors":"Rindra Wiska, Novian Habibie, A. Wibisono, W. S. Nugroho, P. Mursanto","doi":"10.1109/IWBIS.2016.7872896","DOIUrl":"https://doi.org/10.1109/IWBIS.2016.7872896","url":null,"abstract":"Wireless Sensor Network (WSN) is a system that have a capability to conduct data acquisition and monitoring in a wide sampling area for a long time. However, because of its big-scale monitoring, amount of data accumulated from WSN is very huge. Conventional database system may not be able to handle its big amount of data. To overcome that, big data approach is used for an alternative data storage system and data analysis process. This research developed a WSN system for CO2 monitoring using Kafka and Impala to distribute a huge amount of data. Sensor nodes gather data and accumulated in temporary storage then streamed via Kafka platform to be stored into Impala database. System tested with data gathered from our-own made sensor nodes and give a good performance.","PeriodicalId":193821,"journal":{"name":"2016 International Workshop on Big Data and Information Security (IWBIS)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123360591","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Dimensionality reduction using deep belief network in big data case study: Hyperspectral image classification 基于深度信念网络的大数据降维研究:高光谱图像分类
Pub Date : 2016-10-01 DOI: 10.1109/IWBIS.2016.7872892
D. M. S. Arsa, G. Jati, Aprinaldi Jasa Mantau, Ito Wasito
The high dimensionality in big data need a heavy computation when the analysis needed. This research proposed a dimensionality reduction using deep belief network (DBN). We used hyperspectral images as case study. The hyperspectral image is a high dimensional image. Some researched have been proposed to reduce hyperspectral image dimension such as using LDA and PCA in spectral-spatial hyperspectral image classification. This paper proposed a dimensionality reduction using deep belief network (DBN) for hyperspectral image classification. In proposed framework, we use two DBNs. First DBN used to reduce the dimension of spectral bands and the second DBN used to extract spectral-spatial feature and as classifier. We used Indian Pines data set that consist of 16 classes and we compared DBN and PCA performance. The result indicates that by using DBN as dimensionality reduction method performed better than PCA in hyperspectral image classification.
由于大数据的高维数,在进行分析时需要进行大量的计算。本文提出了一种基于深度信念网络(DBN)的降维方法。我们使用高光谱图像作为案例研究。高光谱图像是一种高维图像。在光谱-空间高光谱图像分类中,提出了一些降低高光谱图像维数的研究方法,如LDA和PCA。提出了一种基于深度信念网络的高光谱图像降维分类方法。在建议的框架中,我们使用两个dbn。第一种DBN用于光谱波段降维,第二种DBN用于提取光谱空间特征并作为分类器。我们使用了包含16个类别的Indian Pines数据集,并比较了DBN和PCA的性能。结果表明,采用DBN作为降维方法对高光谱图像进行分类的效果优于PCA。
{"title":"Dimensionality reduction using deep belief network in big data case study: Hyperspectral image classification","authors":"D. M. S. Arsa, G. Jati, Aprinaldi Jasa Mantau, Ito Wasito","doi":"10.1109/IWBIS.2016.7872892","DOIUrl":"https://doi.org/10.1109/IWBIS.2016.7872892","url":null,"abstract":"The high dimensionality in big data need a heavy computation when the analysis needed. This research proposed a dimensionality reduction using deep belief network (DBN). We used hyperspectral images as case study. The hyperspectral image is a high dimensional image. Some researched have been proposed to reduce hyperspectral image dimension such as using LDA and PCA in spectral-spatial hyperspectral image classification. This paper proposed a dimensionality reduction using deep belief network (DBN) for hyperspectral image classification. In proposed framework, we use two DBNs. First DBN used to reduce the dimension of spectral bands and the second DBN used to extract spectral-spatial feature and as classifier. We used Indian Pines data set that consist of 16 classes and we compared DBN and PCA performance. The result indicates that by using DBN as dimensionality reduction method performed better than PCA in hyperspectral image classification.","PeriodicalId":193821,"journal":{"name":"2016 International Workshop on Big Data and Information Security (IWBIS)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122486651","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Spatial data mining for predicting of unobserved zinc pollutant using ordinary point Kriging 常点克里格法预测未观测锌污染物的空间数据挖掘
Pub Date : 2016-10-01 DOI: 10.1109/IWBIS.2016.7872894
A. A. Gunawan, A. N. Falah, Alfensi Faruk, D. S. Lutero, B. N. Ruchjana, A. S. Abdullah
Due to pollution over many years, large amounts of heavy metal pollutant can be accumulated in the rivers. In the research, we would like to predict the dangerous region around the river. For study case, we use the Meuse river floodplains which are contaminated with zinc (Zn). Large zinc concentrations can cause many health problems, for example vomiting, skin irritations, stomach cramps, and anaemia. However there is only few sample data about the zinc concentration of Meuse river, thus the missing data in unknown regions need to be generated. The aim of this research is to study and to apply spatial data mining to predict unobserved zinc pollutant by using ordinary point Kriging. By mean of semivariogram, the variability pattern of zinc can be captured. This captured model will be interpolated to predict the unknown regions by using Kriging method. In our experiments, we propose ordinary point Kriging and employ several semivariogram: Gaussian, Exponential and Spherical models. The experimental results show that: (i) by calculating the minimum error sum of squares, the fittest theoretical semivariogram models is exponential model (ii) the accuracy of the predictions can be confirmed visually by projecting the results to the map.
由于多年的污染,河流中积累了大量的重金属污染物。在研究中,我们希望对河流周围的危险区域进行预测。以锌污染严重的默兹河漫滩为研究对象。高浓度的锌会导致许多健康问题,例如呕吐、皮肤刺激、胃痉挛和贫血。但由于默兹河锌浓度的样本数据较少,需要生成未知区域的缺失数据。本研究的目的是研究并应用空间数据挖掘方法,利用常点克里格法对未观测到的锌污染物进行预测。利用半变异函数,可以捕捉到锌的变化规律。对捕获的模型进行插值,利用克里格法预测未知区域。在我们的实验中,我们提出了常点克里格,并采用了几种半变异函数:高斯模型、指数模型和球面模型。实验结果表明:(1)通过计算最小误差平方和,拟合的理论半变异函数模型为指数模型;(2)将结果投影到地图上,可以直观地证实预测的准确性。
{"title":"Spatial data mining for predicting of unobserved zinc pollutant using ordinary point Kriging","authors":"A. A. Gunawan, A. N. Falah, Alfensi Faruk, D. S. Lutero, B. N. Ruchjana, A. S. Abdullah","doi":"10.1109/IWBIS.2016.7872894","DOIUrl":"https://doi.org/10.1109/IWBIS.2016.7872894","url":null,"abstract":"Due to pollution over many years, large amounts of heavy metal pollutant can be accumulated in the rivers. In the research, we would like to predict the dangerous region around the river. For study case, we use the Meuse river floodplains which are contaminated with zinc (Zn). Large zinc concentrations can cause many health problems, for example vomiting, skin irritations, stomach cramps, and anaemia. However there is only few sample data about the zinc concentration of Meuse river, thus the missing data in unknown regions need to be generated. The aim of this research is to study and to apply spatial data mining to predict unobserved zinc pollutant by using ordinary point Kriging. By mean of semivariogram, the variability pattern of zinc can be captured. This captured model will be interpolated to predict the unknown regions by using Kriging method. In our experiments, we propose ordinary point Kriging and employ several semivariogram: Gaussian, Exponential and Spherical models. The experimental results show that: (i) by calculating the minimum error sum of squares, the fittest theoretical semivariogram models is exponential model (ii) the accuracy of the predictions can be confirmed visually by projecting the results to the map.","PeriodicalId":193821,"journal":{"name":"2016 International Workshop on Big Data and Information Security (IWBIS)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121116694","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
The application of big data using MongoDB: Case study with SCeLE Fasilkom UI forum data MongoDB的大数据应用:以SCeLE Fasilkom UI论坛数据为例
Pub Date : 2016-10-01 DOI: 10.1109/IWBIS.2016.7872889
Argianto Rahartomo, R. F. Aji, Y. Ruldeviyani
Big Data is a condition in which data size in a database is very large so it is difficult to be managed. An e-Learning application, like SCeLE Fasilkom UI (scele.cs.ui.ac.id), also has a very large data. SCeLE has hundreds of forum data, and each forum has at least 4000 threads of discussion. In addition, one thread can have at least dozens or hundreds posts. Therefore, it may further experience data growth problem, which will be difficult to be handled by RDBMS, such as MySQL that is currently used. In order to solve this problem, a research been conducted to apply Big Data in SCeLE Fasilkom UI, which implementation is aimed to increase SCeLE's data management performance. The implementation of Big Data in the research used MongoDB as the system's DBMS. The research result showed that MongoDB obtain better results than MySQL in SCeLE Fasilkom UI forum data case in terms of speed.
大数据是指数据库中的数据量非常大,难以管理的情况。像SCeLE Fasilkom UI (SCeLE .cs. UI .ac.id)这样的电子学习应用程序也有非常大的数据量。SCeLE有数百个论坛数据,每个论坛至少有4000个讨论主题。另外,一个帖子至少可以有几十个或几百个帖子。因此,它可能会进一步遇到数据增长问题,这将是难以处理的RDBMS,如目前使用的MySQL。为了解决这一问题,本研究将大数据应用于SCeLE Fasilkom UI,其实施旨在提高SCeLE的数据管理性能。本研究中大数据的实现使用MongoDB作为系统的DBMS。研究结果表明,在SCeLE Fasilkom UI论坛数据案例中,MongoDB在速度方面优于MySQL。
{"title":"The application of big data using MongoDB: Case study with SCeLE Fasilkom UI forum data","authors":"Argianto Rahartomo, R. F. Aji, Y. Ruldeviyani","doi":"10.1109/IWBIS.2016.7872889","DOIUrl":"https://doi.org/10.1109/IWBIS.2016.7872889","url":null,"abstract":"Big Data is a condition in which data size in a database is very large so it is difficult to be managed. An e-Learning application, like SCeLE Fasilkom UI (scele.cs.ui.ac.id), also has a very large data. SCeLE has hundreds of forum data, and each forum has at least 4000 threads of discussion. In addition, one thread can have at least dozens or hundreds posts. Therefore, it may further experience data growth problem, which will be difficult to be handled by RDBMS, such as MySQL that is currently used. In order to solve this problem, a research been conducted to apply Big Data in SCeLE Fasilkom UI, which implementation is aimed to increase SCeLE's data management performance. The implementation of Big Data in the research used MongoDB as the system's DBMS. The research result showed that MongoDB obtain better results than MySQL in SCeLE Fasilkom UI forum data case in terms of speed.","PeriodicalId":193821,"journal":{"name":"2016 International Workshop on Big Data and Information Security (IWBIS)","volume":"140 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127501914","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2016 International Workshop on Big Data and Information Security (IWBIS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1