Client-server application model has several drawbacks such as server bottleneck and weak scalability. Peer-to-Peer (P2P) model resolves these problems by distributing tasks on the nodes participating in the system. P2P application development and protocol designing are much more difficult than client-server model. This paper proposes an approach to take the advantages of both models: the scalability of the P2P model and the simplicity of the client-server model. This paper presents a hybrid P2P and client-server model to achieve both goals based on caching mechanism which allows using cache content not only on a single client, but for all clients in the system. The proposed model has been applied to implement a Web application.
{"title":"P2P shared-caching model: using P2P to improve client-server application performance","authors":"Luong Quy Tho, Ha Quoc Trung","doi":"10.1145/2542050.2542090","DOIUrl":"https://doi.org/10.1145/2542050.2542090","url":null,"abstract":"Client-server application model has several drawbacks such as server bottleneck and weak scalability. Peer-to-Peer (P2P) model resolves these problems by distributing tasks on the nodes participating in the system. P2P application development and protocol designing are much more difficult than client-server model. This paper proposes an approach to take the advantages of both models: the scalability of the P2P model and the simplicity of the client-server model. This paper presents a hybrid P2P and client-server model to achieve both goals based on caching mechanism which allows using cache content not only on a single client, but for all clients in the system. The proposed model has been applied to implement a Web application.","PeriodicalId":246033,"journal":{"name":"Proceedings of the 4th Symposium on Information and Communication Technology","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114169708","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Dramatic paradigm shifts over the past few centuries have led to a rich landscape of options in human and machine communication. Communication today is deeply intertwined with our personal and social lives in addition to being a vital part of businesses and government operations---both overt and covert. This talk will address the evolving role that the fundamental laws of quantum physics are likely to play in giving communication yet another dimension in its richness. Quantum communication will bring about not only communication in the form as we know it, but also ensure that it is unconditionally secure as it transits through any medium. Given that information is the currency of a modern society, its security is paramount for the wellbeing of an individual, a society, a nation, or the globe as a whole. The talk will discuss the short history of quantum communication and draw upon the theoretical and experimental work that the author and his colleagues have conducted over the past few years in order to chart out the likely course of future events in the emerging age of secure communication.
{"title":"The dawn of quantum communication","authors":"P. Verma","doi":"10.1145/2542050.2542053","DOIUrl":"https://doi.org/10.1145/2542050.2542053","url":null,"abstract":"Dramatic paradigm shifts over the past few centuries have led to a rich landscape of options in human and machine communication. Communication today is deeply intertwined with our personal and social lives in addition to being a vital part of businesses and government operations---both overt and covert. This talk will address the evolving role that the fundamental laws of quantum physics are likely to play in giving communication yet another dimension in its richness. Quantum communication will bring about not only communication in the form as we know it, but also ensure that it is unconditionally secure as it transits through any medium. Given that information is the currency of a modern society, its security is paramount for the wellbeing of an individual, a society, a nation, or the globe as a whole. The talk will discuss the short history of quantum communication and draw upon the theoretical and experimental work that the author and his colleagues have conducted over the past few years in order to chart out the likely course of future events in the emerging age of secure communication.","PeriodicalId":246033,"journal":{"name":"Proceedings of the 4th Symposium on Information and Communication Technology","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127298068","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In pervasive computing environments, wireless sensor networks play an important infrastructure role, collecting reliable and accurate context information so that applications are able to provide services to users on demand. In such environments, sensors should be self-adaptive by taking correct decisions based on sensed data in real-time in a decentralised manner; however, sensed data is often faulty. We thus design a decentralised scheme for fault detection and classification in sensor data in which each sensor node does localised fault detection. A combination of neighbourhood voting and time series data analysis techniques are used to detect faults. We also study the comparative accuracy of both the union and the intersection of the two techniques. Then, detected faults are classified into known fault categories. An initial evaluation with SensorScope, an outdoor temperature dataset, confirms that our solution is able to detect and classify faulty readings into four fault types, namely, 1) random, 2) mal-function, 3) bias, and 4) drift with accuracy up to 95%. The results also show that, with the experimental dataset, the time series data analysis technique performs comparable well in most of the cases, whilst in some other cases the support from neighbourhood voting technique and histogram analysis helps our hybrid solution to successfully detects the faults of all types.
{"title":"Applying time series analysis and neighbourhood voting in a decentralised approach for fault detection and classification in WSNs","authors":"T. Nguyen, Doina Bucur, Marco Aiello, K. Tei","doi":"10.1145/2542050.2542080","DOIUrl":"https://doi.org/10.1145/2542050.2542080","url":null,"abstract":"In pervasive computing environments, wireless sensor networks play an important infrastructure role, collecting reliable and accurate context information so that applications are able to provide services to users on demand. In such environments, sensors should be self-adaptive by taking correct decisions based on sensed data in real-time in a decentralised manner; however, sensed data is often faulty. We thus design a decentralised scheme for fault detection and classification in sensor data in which each sensor node does localised fault detection. A combination of neighbourhood voting and time series data analysis techniques are used to detect faults. We also study the comparative accuracy of both the union and the intersection of the two techniques. Then, detected faults are classified into known fault categories. An initial evaluation with SensorScope, an outdoor temperature dataset, confirms that our solution is able to detect and classify faulty readings into four fault types, namely, 1) random, 2) mal-function, 3) bias, and 4) drift with accuracy up to 95%. The results also show that, with the experimental dataset, the time series data analysis technique performs comparable well in most of the cases, whilst in some other cases the support from neighbourhood voting technique and histogram analysis helps our hybrid solution to successfully detects the faults of all types.","PeriodicalId":246033,"journal":{"name":"Proceedings of the 4th Symposium on Information and Communication Technology","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127445323","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Document clustering has become an increasingly important technique for unsupervised document organization, automatic topic extraction, and fast information retrieval or filtering. This paper proposes a Dirichlet process mixture (DPM) model approach to clustering directional data based on the von Mises-Fisher (vMF) distribution, which arises naturally for data distributed on the unit hypersphere. We have developed a mean-field variational inference algorithm for the DPM model of vMFs that is applied to clustering text documents. Using this model, the number of clusters is determined automatically after the clustering process rather than pre-estimated. We conducted extensive experiments to evaluate the proposed approach on a large number of high dimensional text datasets. Empirical experimental results over NMI (Normalized Mutual Information) and Purity evaluation measures demonstrate that our approach outperforms the four state-of-the-art clustering algorithms.
{"title":"Document clustering using dirichlet process mixture model of von Mises-Fisher distributions","authors":"N. K. Anh, Tam The Nguyen, Ngo Van Linh","doi":"10.1145/2542050.2542079","DOIUrl":"https://doi.org/10.1145/2542050.2542079","url":null,"abstract":"Document clustering has become an increasingly important technique for unsupervised document organization, automatic topic extraction, and fast information retrieval or filtering. This paper proposes a Dirichlet process mixture (DPM) model approach to clustering directional data based on the von Mises-Fisher (vMF) distribution, which arises naturally for data distributed on the unit hypersphere. We have developed a mean-field variational inference algorithm for the DPM model of vMFs that is applied to clustering text documents. Using this model, the number of clusters is determined automatically after the clustering process rather than pre-estimated. We conducted extensive experiments to evaluate the proposed approach on a large number of high dimensional text datasets. Empirical experimental results over NMI (Normalized Mutual Information) and Purity evaluation measures demonstrate that our approach outperforms the four state-of-the-art clustering algorithms.","PeriodicalId":246033,"journal":{"name":"Proceedings of the 4th Symposium on Information and Communication Technology","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116796555","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Natural Language Processing (NLP) for Vietnamese has been researched for more than a decade but still lacks of an open-source NLP pipeline. As the result, researchers have to spend a lot of time on various fundamental tasks before working on the task of interest. Besides, the circumstance holds back text processing technology in Vietnam because an application costs much more money and time to reach a deliverable state. This work is an attempt to solve this issue. By incorporating available open-source software packages and implementing new ones, we have created an open-source, production-ready solution for Vietnamese text processing. Via three experiments, we demonstrated its effectiveness and efficiency. The software has helped us to develop our solution for Vietnamese sentiment analysis and online reputation management and we hope that it will also facilitate research in Vietnamese NLP.
{"title":"VNLP: an open source framework for Vietnamese natural language processing","authors":"N. Le, Bich Ngoc Do, Vien Nguyen, Thi Dam Nguyen","doi":"10.1145/2542050.2542062","DOIUrl":"https://doi.org/10.1145/2542050.2542062","url":null,"abstract":"Natural Language Processing (NLP) for Vietnamese has been researched for more than a decade but still lacks of an open-source NLP pipeline. As the result, researchers have to spend a lot of time on various fundamental tasks before working on the task of interest. Besides, the circumstance holds back text processing technology in Vietnam because an application costs much more money and time to reach a deliverable state. This work is an attempt to solve this issue. By incorporating available open-source software packages and implementing new ones, we have created an open-source, production-ready solution for Vietnamese text processing. Via three experiments, we demonstrated its effectiveness and efficiency. The software has helped us to develop our solution for Vietnamese sentiment analysis and online reputation management and we hope that it will also facilitate research in Vietnamese NLP.","PeriodicalId":246033,"journal":{"name":"Proceedings of the 4th Symposium on Information and Communication Technology","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126023175","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
M. Adriani, Y. W. Choong, Ba-Hung Ngo, Laurent d'Orazio, D. Laurent, N. Spyratos
Cloud computing provides access to "infinite" storage and computing resources, offering promising perspectives for many applications, particularly e-learning. However, this new paradigm requires rethinking of database management principles in order to allow deployment on scalable, easy to access infrastructures, applying a pay-as-you-go model in which failures are not exceptions but rather the norm. The GOD project aims to provide an optimized data management system for e-learning in the cloud by rethinking traditional database management techniques, extending them to consider the specificities of this paradigm.
{"title":"Optimized data management for e-learning in the clouds towards Cloodle","authors":"M. Adriani, Y. W. Choong, Ba-Hung Ngo, Laurent d'Orazio, D. Laurent, N. Spyratos","doi":"10.1145/2542050.2542089","DOIUrl":"https://doi.org/10.1145/2542050.2542089","url":null,"abstract":"Cloud computing provides access to \"infinite\" storage and computing resources, offering promising perspectives for many applications, particularly e-learning. However, this new paradigm requires rethinking of database management principles in order to allow deployment on scalable, easy to access infrastructures, applying a pay-as-you-go model in which failures are not exceptions but rather the norm. The GOD project aims to provide an optimized data management system for e-learning in the cloud by rethinking traditional database management techniques, extending them to consider the specificities of this paradigm.","PeriodicalId":246033,"journal":{"name":"Proceedings of the 4th Symposium on Information and Communication Technology","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126586584","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A novel approach based on structure information extraction in frequency domain is proposed for image representation problem. Regarding this problem, a new subspace method based on Two-dimensional Fractional Principle Component Analysis (2D-FPCA) in frequency domain is applied to images, thus extracting the texture information. In order to extract the structure information, the system utilizes this new subspace as the bilateral consideration of 2D-FPCA technique called B2D-FPCA. For this purpose: (1) we first introduce the theory of 2D-FPCA based on the definition of fractional variance and fractional covariance matrix; (2) then show its improvement called Bilateral 2D-FPCA and (3) the robustness of 2D-DCT is also described as the preprocessing step. This approach is applied to facial expression representation problem to prove the stability and robustness of the proposed framework. For demonstration, facial expressions datasets (JAFFE, Pain expression subset and Cohn-Kanade) are used in order to compare the proposed framework with some other approaches.
{"title":"On approaching 2D-FPCA technique to improve image representation in frequency domain","authors":"T. Le, Hung Phuoc Truong, H. T. Do, Duc Minh Vo","doi":"10.1145/2542050.2542061","DOIUrl":"https://doi.org/10.1145/2542050.2542061","url":null,"abstract":"A novel approach based on structure information extraction in frequency domain is proposed for image representation problem. Regarding this problem, a new subspace method based on Two-dimensional Fractional Principle Component Analysis (2D-FPCA) in frequency domain is applied to images, thus extracting the texture information. In order to extract the structure information, the system utilizes this new subspace as the bilateral consideration of 2D-FPCA technique called B2D-FPCA. For this purpose: (1) we first introduce the theory of 2D-FPCA based on the definition of fractional variance and fractional covariance matrix; (2) then show its improvement called Bilateral 2D-FPCA and (3) the robustness of 2D-DCT is also described as the preprocessing step. This approach is applied to facial expression representation problem to prove the stability and robustness of the proposed framework. For demonstration, facial expressions datasets (JAFFE, Pain expression subset and Cohn-Kanade) are used in order to compare the proposed framework with some other approaches.","PeriodicalId":246033,"journal":{"name":"Proceedings of the 4th Symposium on Information and Communication Technology","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124275072","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Although both high maturity and agility appeared as different ways to address and overcome issues related to software development (including maximizing resources and minimizing risks), there has been a mixed understanding about the possibility for their co-existence within an organization. Outside of the dogmatic debate regarding their co-existence, however, voices have been raised recently that recognize that both approaches have their merits. This paper presents the results of a case study on the practices that a purely agile organization has put in place in order to profit from the opportunities that higher maturity can offer in respect to value creation for clients. Our conclusion is that both high maturity and agility contribute to customer satisfaction, high quality and waste reduction; and that complying with standards does not necessarily impose restriction on 'being agile'. Implication for practice is that companies and their clients can benefit from a development approach that embraces both maturity and agility. To achieve this goal, guidelines are needed that direct organizations towards adopting practices that are linked to higher maturity, as well as to agility.
{"title":"Combining maturity with agility: lessons learnt from a case study","authors":"N. Tuan, H. Thang","doi":"10.1145/2542050.2542072","DOIUrl":"https://doi.org/10.1145/2542050.2542072","url":null,"abstract":"Although both high maturity and agility appeared as different ways to address and overcome issues related to software development (including maximizing resources and minimizing risks), there has been a mixed understanding about the possibility for their co-existence within an organization. Outside of the dogmatic debate regarding their co-existence, however, voices have been raised recently that recognize that both approaches have their merits. This paper presents the results of a case study on the practices that a purely agile organization has put in place in order to profit from the opportunities that higher maturity can offer in respect to value creation for clients. Our conclusion is that both high maturity and agility contribute to customer satisfaction, high quality and waste reduction; and that complying with standards does not necessarily impose restriction on 'being agile'. Implication for practice is that companies and their clients can benefit from a development approach that embraces both maturity and agility. To achieve this goal, guidelines are needed that direct organizations towards adopting practices that are linked to higher maturity, as well as to agility.","PeriodicalId":246033,"journal":{"name":"Proceedings of the 4th Symposium on Information and Communication Technology","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116526873","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
N. Thang, Nguyen Huynh Minh Tam, Tran Le Giang, Vo Nhut Tuan, Lan Anh Trinh, Hoang-Hai Tran, V. Toi
Near infrared spectroscopy (NIRS) is currently becoming an effective technique for noninvasive functional brain imaging. Therefore, the methods to improve the quality of measured NIRS signals play an important role to make NIRS broadly accepted in practical applications. Previously, there have been approaches using state-space modeling to recover the NIRS signals from basic component signals to eliminate the artifacts presented in the NIRS measurements. However, the proposed approach requires us an onset vector to determine the starting position of stimulus that is not always available in practical situation. In this work, we provide a new way to find the basic components for efficient implementations of the state-space modeling. We apply principal component analysis to estimate eigenvector-based basis that presents the compact information of the whole signals. We utilize the oxygenated-deoxygenated correlation to find another set of basic components to enhance the quality of NIRS signals. The state-space modeling based on Kalman filter is used to reconstruct the NIRS signals from these basic components. We tested the proposed algorithm with actual data and showed significant improvements of the contrast-to-noise (CNR) of the NIRS signals after filtered by our proposed approach.
{"title":"State-space modeling based on principal component analysis and oxygenated-deoxygenated correlation to improve near-infrared spectroscopy signals","authors":"N. Thang, Nguyen Huynh Minh Tam, Tran Le Giang, Vo Nhut Tuan, Lan Anh Trinh, Hoang-Hai Tran, V. Toi","doi":"10.1145/2542050.2542094","DOIUrl":"https://doi.org/10.1145/2542050.2542094","url":null,"abstract":"Near infrared spectroscopy (NIRS) is currently becoming an effective technique for noninvasive functional brain imaging. Therefore, the methods to improve the quality of measured NIRS signals play an important role to make NIRS broadly accepted in practical applications. Previously, there have been approaches using state-space modeling to recover the NIRS signals from basic component signals to eliminate the artifacts presented in the NIRS measurements. However, the proposed approach requires us an onset vector to determine the starting position of stimulus that is not always available in practical situation. In this work, we provide a new way to find the basic components for efficient implementations of the state-space modeling. We apply principal component analysis to estimate eigenvector-based basis that presents the compact information of the whole signals. We utilize the oxygenated-deoxygenated correlation to find another set of basic components to enhance the quality of NIRS signals. The state-space modeling based on Kalman filter is used to reconstruct the NIRS signals from these basic components. We tested the proposed algorithm with actual data and showed significant improvements of the contrast-to-noise (CNR) of the NIRS signals after filtered by our proposed approach.","PeriodicalId":246033,"journal":{"name":"Proceedings of the 4th Symposium on Information and Communication Technology","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121075489","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Graph has emerged as a powerful data structure to describe various data. Query evaluation on distributed graphs takes much cost due to the complexity of links among sites. Dan Suciu has proposed algorithms for query evaluation on semistructured data that is a rooted, edge-labeled graph, and algorithms are proved to be efficient in terms of communication steps and data transferring during the evaluation. However, one disadvantage is that communication data are collected to one single site, which leads to a bottleneck in the evaluation for real-life data. In this paper, we propose two algorithms to improve Dan Suciu's algorithms: one-pass algorithm is to significantly reduce a large amount of redundant data in the evaluation, and iter_acc algorithm is to resolve the bottleneck. Then, we design an efficient implementation with only one MapReduce job for our algorithms in Hadoop environment by utilizing features of Hadoop file system. Experiments on cloud system show that one-pass algorithm can detect and remove 50% of data being redundant in the evaluation process on YouTube and DBLP datasets, and iter_acc algorithm is running without the bottleneck even when we double the size of input data.
{"title":"Efficient query evaluation on distributed graphs with Hadoop environment","authors":"Le-Duc Tung, Quyet Nguyen-Van, Zhenjiang Hu","doi":"10.1145/2542050.2542086","DOIUrl":"https://doi.org/10.1145/2542050.2542086","url":null,"abstract":"Graph has emerged as a powerful data structure to describe various data. Query evaluation on distributed graphs takes much cost due to the complexity of links among sites. Dan Suciu has proposed algorithms for query evaluation on semistructured data that is a rooted, edge-labeled graph, and algorithms are proved to be efficient in terms of communication steps and data transferring during the evaluation. However, one disadvantage is that communication data are collected to one single site, which leads to a bottleneck in the evaluation for real-life data. In this paper, we propose two algorithms to improve Dan Suciu's algorithms: one-pass algorithm is to significantly reduce a large amount of redundant data in the evaluation, and iter_acc algorithm is to resolve the bottleneck. Then, we design an efficient implementation with only one MapReduce job for our algorithms in Hadoop environment by utilizing features of Hadoop file system. Experiments on cloud system show that one-pass algorithm can detect and remove 50% of data being redundant in the evaluation process on YouTube and DBLP datasets, and iter_acc algorithm is running without the bottleneck even when we double the size of input data.","PeriodicalId":246033,"journal":{"name":"Proceedings of the 4th Symposium on Information and Communication Technology","volume":"74 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127089811","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}