首页 > 最新文献

Proceedings of the 9th International Symposium on Information and Communication Technology最新文献

英文 中文
Episode-Rule Mining with Minimal Occurrences via First Local Maximization in Confidence 基于置信度的第一次局部最大化最小化事件规则挖掘
H. K. Dai
An episode rule of associating two episodes represents a temporal implication of the antecedent episode to the consequent episode. Episode-rule mining is a task of extracting useful patterns/episodes from large event databases. We present an episode-rule mining algorithm for finding frequent and confident serial-episode rules via first local-maximum confidence in yielding ideal window widths, if exist, in event sequences based on minimal occurrences constrained by a constant maximum gap. Results from our preliminary empirical study confirm the applicability of the episode-rule mining algorithm for Web-site traversal-pattern discovery, and show that the first local maximization yielding ideal window widths exists in real data but rarely in synthetic random data sets.
将两个情节联系起来的情节规则代表了前一个情节对后一个情节的时间含义。事件规则挖掘是一项从大型事件数据库中提取有用模式/事件的任务。我们提出了一种情景规则挖掘算法,该算法通过第一个局部最大置信度来发现频繁和可靠的串行情景规则,如果存在,则在基于最小事件的事件序列中,由恒定最大间隙约束的理想窗宽。我们的初步实证研究结果证实了情节规则挖掘算法在网站遍历模式发现中的适用性,并表明产生理想窗口宽度的第一个局部最大化存在于真实数据中,而很少存在于合成随机数据集中。
{"title":"Episode-Rule Mining with Minimal Occurrences via First Local Maximization in Confidence","authors":"H. K. Dai","doi":"10.1145/3287921.3287982","DOIUrl":"https://doi.org/10.1145/3287921.3287982","url":null,"abstract":"An episode rule of associating two episodes represents a temporal implication of the antecedent episode to the consequent episode. Episode-rule mining is a task of extracting useful patterns/episodes from large event databases. We present an episode-rule mining algorithm for finding frequent and confident serial-episode rules via first local-maximum confidence in yielding ideal window widths, if exist, in event sequences based on minimal occurrences constrained by a constant maximum gap. Results from our preliminary empirical study confirm the applicability of the episode-rule mining algorithm for Web-site traversal-pattern discovery, and show that the first local maximization yielding ideal window widths exists in real data but rarely in synthetic random data sets.","PeriodicalId":448008,"journal":{"name":"Proceedings of the 9th International Symposium on Information and Communication Technology","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123383118","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Automatic Embedding of Social Network Profile Links into Knowledge Graphs 自动嵌入社会网络配置文件链接到知识图谱
Hussein Hazimeh, E. Mugellini, Simon Ruffieux, Omar Abou Khaled, P. Cudré-Mauroux
Recent Knowledge Graphs (KGs) like Wikidata and YAGO are often constructed by incorporating knowledge from semi-structured heterogeneous data resources such as Wikipedia. However, despite their large amount of knowledge, these graphs are still incomplete. In this paper, we posit that Online Social Networks (OSNs) can become prominent data resources comprising abundant knowledge about real-world entities. An entity on an OSN is represented by a profile; the link to this profile is called a social link. In this paper, we propose a KG refinement method for adding missing knowledge to a KG, i.e., social links. We target specific entity types, in the scientific community, such as researchers. Our approach uses both scholarly data resources and existing KG for building knowledge bases. Then, it matches this knowledge with OSNs to detect the corresponding social link(s) for a specific entity. It uses a novel matching algorithm, in combination with supervised and unsupervised learning methods. We empirically validate that our system is able to detect a large number of social links with high confidence.
最近的知识图谱(Knowledge Graphs, KGs),如Wikidata和YAGO,通常是通过整合来自半结构化异构数据资源(如Wikipedia)的知识来构建的。然而,尽管他们有大量的知识,这些图表仍然是不完整的。在本文中,我们假设在线社交网络(OSNs)可以成为包含关于现实世界实体的丰富知识的重要数据资源。OSN上的实体由配置文件表示;到此配置文件的链接称为社交链接。在本文中,我们提出了一种KG细化方法,将缺失的知识添加到KG中,即社会链接。我们针对的是科学界的特定实体类型,比如研究人员。我们的方法使用学术数据资源和现有的KG来构建知识库。然后,它将这些知识与osn进行匹配,以检测特定实体的相应社会链接。它采用了一种新颖的匹配算法,结合了监督学习和无监督学习方法。我们通过经验验证了我们的系统能够以高置信度检测到大量的社会联系。
{"title":"Automatic Embedding of Social Network Profile Links into Knowledge Graphs","authors":"Hussein Hazimeh, E. Mugellini, Simon Ruffieux, Omar Abou Khaled, P. Cudré-Mauroux","doi":"10.1145/3287921.3287926","DOIUrl":"https://doi.org/10.1145/3287921.3287926","url":null,"abstract":"Recent Knowledge Graphs (KGs) like Wikidata and YAGO are often constructed by incorporating knowledge from semi-structured heterogeneous data resources such as Wikipedia. However, despite their large amount of knowledge, these graphs are still incomplete. In this paper, we posit that Online Social Networks (OSNs) can become prominent data resources comprising abundant knowledge about real-world entities. An entity on an OSN is represented by a profile; the link to this profile is called a social link. In this paper, we propose a KG refinement method for adding missing knowledge to a KG, i.e., social links. We target specific entity types, in the scientific community, such as researchers. Our approach uses both scholarly data resources and existing KG for building knowledge bases. Then, it matches this knowledge with OSNs to detect the corresponding social link(s) for a specific entity. It uses a novel matching algorithm, in combination with supervised and unsupervised learning methods. We empirically validate that our system is able to detect a large number of social links with high confidence.","PeriodicalId":448008,"journal":{"name":"Proceedings of the 9th International Symposium on Information and Communication Technology","volume":"71 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130120540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Frozen Shoulder Rehabilitation: Exercise Simulation and Usability Study 肩周炎康复:运动模拟与可用性研究
Nuntiya Chiensriwimol, P. Mongkolnam, Jonathan H. Chan
Frozen shoulder treatment is normally a time-consuming process. Continual physical therapy is required in practice for a patient to gradually recover over time. With the advent of mobile technology, there is an increasing number of smartphone applications being developed to facilitate patients to perform telerehabilitation. In this study, we incorporate animation to simulate arm movement in various exercise types via a mobile app to augment the use of biofeedback data for the treatment process. The main contribution of this paper is to simulate the frozen shoulder exercise using a Unity 3D model. Patients can do rehabilitation exercises at home by putting their smartphone on the shoulder using an armband and the data will be sent to the physiotherapist without the need to wait in long queue at the clinic to see the practitioner. The results indicated that our mobile app and web dashboard is useful for physiotherapists to easily monitor as well as manage a patient's rehabilitation process remotely.
肩周炎的治疗通常是一个耗时的过程。在实践中,需要持续的物理治疗,以使患者逐渐康复。随着移动技术的出现,越来越多的智能手机应用程序被开发出来,以方便患者进行远程康复。在这项研究中,我们通过移动应用程序结合动画来模拟各种运动类型的手臂运动,以增加治疗过程中生物反馈数据的使用。本文的主要贡献是使用Unity 3D模型模拟冻肩运动。患者可以将智能手机戴在臂章上,在家中进行康复训练,数据将发送给物理治疗师,而无需在诊所排长队等候医生。结果表明,我们的移动应用程序和web仪表盘对物理治疗师远程监控和管理患者的康复过程很有用。
{"title":"Frozen Shoulder Rehabilitation: Exercise Simulation and Usability Study","authors":"Nuntiya Chiensriwimol, P. Mongkolnam, Jonathan H. Chan","doi":"10.1145/3287921.3287951","DOIUrl":"https://doi.org/10.1145/3287921.3287951","url":null,"abstract":"Frozen shoulder treatment is normally a time-consuming process. Continual physical therapy is required in practice for a patient to gradually recover over time. With the advent of mobile technology, there is an increasing number of smartphone applications being developed to facilitate patients to perform telerehabilitation. In this study, we incorporate animation to simulate arm movement in various exercise types via a mobile app to augment the use of biofeedback data for the treatment process. The main contribution of this paper is to simulate the frozen shoulder exercise using a Unity 3D model. Patients can do rehabilitation exercises at home by putting their smartphone on the shoulder using an armband and the data will be sent to the physiotherapist without the need to wait in long queue at the clinic to see the practitioner. The results indicated that our mobile app and web dashboard is useful for physiotherapists to easily monitor as well as manage a patient's rehabilitation process remotely.","PeriodicalId":448008,"journal":{"name":"Proceedings of the 9th International Symposium on Information and Communication Technology","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116194377","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
High Accuracy Forecasting with Limited Input Data: Using FFNNs to Predict Offshore Wind Power Generation 有限输入数据的高精度预测:使用ffnn预测海上风力发电
Elaine Zaunseder, Larissa Müller, S. Blankenburg
This study proposes a Feed Forward Neural Net (FFNN) to forecast renewable energy generation of marine wind parks located in Denmark. The neural network uses historical weather and power generation data for training and applies the learned pattern to forecast wind energy production. Furthermore, the study shows how to improve prediction quality by leveraging specific parameters. Especially, we study the impact of the distance and direction of the weather station related to the production site in detail. In addition, we examined various parameters of the network to improve the accuracy. The proposed model distinguishes itself from other models by the fact that the optimal validation accuracy of more than 90 percent can be reached with training data sets of only a limited size, here two months of data with hourly resolution.
本研究提出了一种前馈神经网络(FFNN)来预测丹麦海上风电场的可再生能源发电。神经网络使用历史天气和发电数据进行训练,并将学习到的模式应用于预测风能产量。此外,研究还展示了如何利用特定的参数来提高预测质量。特别对与生产现场相关的气象站的距离和方向的影响进行了详细的研究。此外,我们还检查了网络的各种参数,以提高准确性。所提出的模型与其他模型的区别在于,仅使用有限规模的训练数据集就可以达到90%以上的最佳验证精度,这里是两个月的数据,每小时分辨率。
{"title":"High Accuracy Forecasting with Limited Input Data: Using FFNNs to Predict Offshore Wind Power Generation","authors":"Elaine Zaunseder, Larissa Müller, S. Blankenburg","doi":"10.1145/3287921.3287936","DOIUrl":"https://doi.org/10.1145/3287921.3287936","url":null,"abstract":"This study proposes a Feed Forward Neural Net (FFNN) to forecast renewable energy generation of marine wind parks located in Denmark. The neural network uses historical weather and power generation data for training and applies the learned pattern to forecast wind energy production. Furthermore, the study shows how to improve prediction quality by leveraging specific parameters. Especially, we study the impact of the distance and direction of the weather station related to the production site in detail. In addition, we examined various parameters of the network to improve the accuracy. The proposed model distinguishes itself from other models by the fact that the optimal validation accuracy of more than 90 percent can be reached with training data sets of only a limited size, here two months of data with hourly resolution.","PeriodicalId":448008,"journal":{"name":"Proceedings of the 9th International Symposium on Information and Communication Technology","volume":"134 44","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120851436","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
CitationLDA++: an Extension of LDA for Discovering Topics in Document Network CitationLDA++: LDA在文献网络中主题发现的扩展
T. Nguyen, P. Do
Along with rapid development of electronic scientific publication repositories, automatic topics identification from papers has helped a lot for the researchers in their research. Latent Dirichlet Allocation (LDA) model is the most popular method which is used to discover hidden topics in texts basing on the co-occurrence of words in a corpus. LDA algorithm has achieved good results for large documents. However, article repositories usually only store title and abstract that are too short for LDA algorithm to work effectively. In this paper, we propose CitationLDA++ model that can improve the performance of the LDA algorithm in inferring topics of the papers basing on the title or/and abstract and citation information. The proposed model is based on the assumption that the topics of the cited papers also reflects the topics of the original paper. In this study, we divide the dataset into two sets. The first one is used to build prior knowledge source using LDA algorithm. The second is training dataset used in CitationLDA++. In the inference process with Gibbs sampling, CitationLDA++ algorithm use topics distribution of prior knowledge source and citation information to guide the process of assigning the topic to words in the text. The use of topics of cited papers helps to tackle the limit of word co-occurrence in case of linked short text. Experiments with the AMiner dataset including title or/and abstract of papers and citation information, CitationLDA++ algorithm gains better perplexity measurement than no additional knowledge. Experimental results suggest that the citation information can improve the performance of LDA algorithm to discover topics of papers in the case of full content of them are not available.
随着电子科学出版物库的迅速发展,论文主题的自动识别为科研人员的研究提供了很大的帮助。潜在狄利克雷分配(Latent Dirichlet Allocation, LDA)模型是一种基于语料库中词的共现性来发现文本中隐藏主题的常用方法。LDA算法在处理大型文档方面取得了较好的效果。然而,文章存储库通常只存储标题和摘要太短,LDA算法无法有效工作。在本文中,我们提出了CitationLDA++模型,该模型可以提高LDA算法基于标题或/和摘要和引文信息推断论文主题的性能。该模型基于被引论文的主题也反映了原论文的主题这一假设。在本研究中,我们将数据集分为两组。第一种是利用LDA算法构建先验知识源。二是CitationLDA++中使用的训练数据集。在Gibbs抽样推理过程中,CitationLDA++算法利用先验知识源和引文信息的主题分布来指导将主题分配给文本中的单词的过程。使用被引论文的主题有助于解决链接短文本中词共现的限制。在包含论文标题或/和摘要以及引文信息的AMiner数据集上进行实验,CitationLDA++算法比没有附加知识获得更好的困惑度度量。实验结果表明,引文信息可以提高LDA算法在没有全文的情况下发现论文主题的性能。
{"title":"CitationLDA++: an Extension of LDA for Discovering Topics in Document Network","authors":"T. Nguyen, P. Do","doi":"10.1145/3287921.3287930","DOIUrl":"https://doi.org/10.1145/3287921.3287930","url":null,"abstract":"Along with rapid development of electronic scientific publication repositories, automatic topics identification from papers has helped a lot for the researchers in their research. Latent Dirichlet Allocation (LDA) model is the most popular method which is used to discover hidden topics in texts basing on the co-occurrence of words in a corpus. LDA algorithm has achieved good results for large documents. However, article repositories usually only store title and abstract that are too short for LDA algorithm to work effectively. In this paper, we propose CitationLDA++ model that can improve the performance of the LDA algorithm in inferring topics of the papers basing on the title or/and abstract and citation information. The proposed model is based on the assumption that the topics of the cited papers also reflects the topics of the original paper. In this study, we divide the dataset into two sets. The first one is used to build prior knowledge source using LDA algorithm. The second is training dataset used in CitationLDA++. In the inference process with Gibbs sampling, CitationLDA++ algorithm use topics distribution of prior knowledge source and citation information to guide the process of assigning the topic to words in the text. The use of topics of cited papers helps to tackle the limit of word co-occurrence in case of linked short text. Experiments with the AMiner dataset including title or/and abstract of papers and citation information, CitationLDA++ algorithm gains better perplexity measurement than no additional knowledge. Experimental results suggest that the citation information can improve the performance of LDA algorithm to discover topics of papers in the case of full content of them are not available.","PeriodicalId":448008,"journal":{"name":"Proceedings of the 9th International Symposium on Information and Communication Technology","volume":"183 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125099134","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Development of a Vietnamese Large Vocabulary Continuous Speech Recognition System under Noisy Conditions 嘈杂条件下越南语大词汇量连续语音识别系统的开发
Quoc Bao Nguyen, Van Tuan Mai, Quang Trung Le, Ba Quyen Dam, Van Hai Do
In this paper, we first present our effort to collect a 500-hour corpus for Vietnamese read speech. After that, various techniques such as data augmentation, recurrent neural network language model rescoring, language model adaptation, bottleneck feature, system combination are applied to build the speech recognition system. Our final system achieves a low word error rate at 6.9% on the noisy test set.
在本文中,我们首先介绍了我们收集500小时越南语阅读语音语料库的工作。然后,应用数据增强、递归神经网络语言模型评分、语言模型自适应、瓶颈特征、系统组合等技术构建语音识别系统。我们的最终系统在有噪声的测试集上实现了6.9%的低单词错误率。
{"title":"Development of a Vietnamese Large Vocabulary Continuous Speech Recognition System under Noisy Conditions","authors":"Quoc Bao Nguyen, Van Tuan Mai, Quang Trung Le, Ba Quyen Dam, Van Hai Do","doi":"10.1145/3287921.3287938","DOIUrl":"https://doi.org/10.1145/3287921.3287938","url":null,"abstract":"In this paper, we first present our effort to collect a 500-hour corpus for Vietnamese read speech. After that, various techniques such as data augmentation, recurrent neural network language model rescoring, language model adaptation, bottleneck feature, system combination are applied to build the speech recognition system. Our final system achieves a low word error rate at 6.9% on the noisy test set.","PeriodicalId":448008,"journal":{"name":"Proceedings of the 9th International Symposium on Information and Communication Technology","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131102616","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
A New Assessment of Cluster Tendency Ensemble approach for Data Clustering 一种新的聚类倾向集成方法用于数据聚类
Van Nha Pham, L. Ngo, L. T. Pham, Pham Van Hai
The ensemble is an universal machine learning method that is based on the divide-and-conquer principle. The ensemble aims to improve performance of system in terms of processing speed and quality. The assessment of cluster tendency is a method determining whether a considering data-set contains meaningful clusters. Recently, a silhouette-based assessment of cluster tendency method (SACT) has been proposed to simultaneously determine the appropriate number of data clusters and the prototypes. The advantages of SACT are accuracy and less the parameter, while there are limitations in data size and processing speed. In this paper, we proposed an improved SACT method for data clustering. We call eSACT algorithm. Experiments were conducted on synthetic data-sets and color image images. The proposed algorithm exhibited high performance, reliability and accuracy compared to previous proposed algorithms in the assessment of cluster tendency.
集成是一种基于分治原则的通用机器学习方法。该集成旨在提高系统在处理速度和质量方面的性能。聚类倾向评估是一种确定考虑数据集是否包含有意义聚类的方法。最近,提出了一种基于轮廓的聚类倾向评估方法(SACT),以同时确定适当的数据聚类数量和原型。SACT的优点是精度高、参数少,但在数据量和处理速度上有一定的限制。本文提出了一种改进的SACT聚类方法。我们称之为eSACT算法。在合成数据集和彩色图像图像上进行了实验。在聚类倾向评估方面,与已有算法相比,该算法具有较高的性能、可靠性和准确性。
{"title":"A New Assessment of Cluster Tendency Ensemble approach for Data Clustering","authors":"Van Nha Pham, L. Ngo, L. T. Pham, Pham Van Hai","doi":"10.1145/3287921.3287927","DOIUrl":"https://doi.org/10.1145/3287921.3287927","url":null,"abstract":"The ensemble is an universal machine learning method that is based on the divide-and-conquer principle. The ensemble aims to improve performance of system in terms of processing speed and quality. The assessment of cluster tendency is a method determining whether a considering data-set contains meaningful clusters. Recently, a silhouette-based assessment of cluster tendency method (SACT) has been proposed to simultaneously determine the appropriate number of data clusters and the prototypes. The advantages of SACT are accuracy and less the parameter, while there are limitations in data size and processing speed. In this paper, we proposed an improved SACT method for data clustering. We call eSACT algorithm. Experiments were conducted on synthetic data-sets and color image images. The proposed algorithm exhibited high performance, reliability and accuracy compared to previous proposed algorithms in the assessment of cluster tendency.","PeriodicalId":448008,"journal":{"name":"Proceedings of the 9th International Symposium on Information and Communication Technology","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129367157","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Binarization Method for Extracting High Entropy String in Gait Biometric Cryptosystem 步态生物识别密码系统中高熵字符串提取的二值化方法
Lam Tran, Thao M. Dang, Deokjai Choi
Inertial-sensors based gait has been considered as a promising approach for user authentication in mobile devices. However, securing enrolled template in such system remains a challenging task. Biometric Cryptosystems (BCS) provide elegant approaches for this matter. The primary task of adopting BCS is to extract from raw biometric data a discriminative, high entropy and stable binary string, which will be used as input of BCS. Unfortunately, the state-of-the-art researches does not notice the gait features' population distribution when extracting such string. Thus, the extracted binary string has low entropy, and degrades the overall system security. In this study, we address the aforementioned drawback to improve entropy of the extracted string, and also enhance the system security. Specifically, we design a binarization scheme, in which the distribution population of gait features are analyzed and utilized to allow the extracted binary string achieving maximal entropy. In addition, the binarization is also designed to provide strong variation toleration to produce highly stable binary string which enhances the system friendliness. We analyzed the proposed method using a gait dataset of 38 volunteers which were collected under nearly realistic conditions. The experiment results show that our proposed binarization method improves the extracted binary string's entropy 30%, and the system achieved competitive performance (i.e., 0.01% FAR, 9.5% FRR with 139-bit key).
基于惯性传感器的步态被认为是一种很有前途的移动设备用户认证方法。然而,在这样的系统中保护注册模板仍然是一项具有挑战性的任务。生物识别密码系统(BCS)为这一问题提供了优雅的方法。采用BCS的主要任务是从原始生物特征数据中提取一个有鉴别性的、高熵的、稳定的二进制字符串作为BCS的输入。遗憾的是,目前的研究在提取步态特征串时没有注意到步态特征的总体分布。因此,提取的二进制字符串具有低熵,并降低了整个系统的安全性。在本研究中,我们解决了上述缺点,提高了提取字符串的熵,同时也提高了系统的安全性。具体而言,我们设计了一种二值化方案,该方案分析步态特征的分布总体,并利用其使提取的二值字符串获得最大熵。此外,二值化设计还提供了较强的变异容忍度,以产生高度稳定的二进制字符串,增强了系统的友好性。我们使用38名志愿者在近乎真实的条件下收集的步态数据集来分析所提出的方法。实验结果表明,我们提出的二值化方法将提取的二进制字符串的熵提高了30%,系统达到了具有竞争力的性能(即在139位密钥下FAR为0.01%,FRR为9.5%)。
{"title":"A Binarization Method for Extracting High Entropy String in Gait Biometric Cryptosystem","authors":"Lam Tran, Thao M. Dang, Deokjai Choi","doi":"10.1145/3287921.3287960","DOIUrl":"https://doi.org/10.1145/3287921.3287960","url":null,"abstract":"Inertial-sensors based gait has been considered as a promising approach for user authentication in mobile devices. However, securing enrolled template in such system remains a challenging task. Biometric Cryptosystems (BCS) provide elegant approaches for this matter. The primary task of adopting BCS is to extract from raw biometric data a discriminative, high entropy and stable binary string, which will be used as input of BCS. Unfortunately, the state-of-the-art researches does not notice the gait features' population distribution when extracting such string. Thus, the extracted binary string has low entropy, and degrades the overall system security. In this study, we address the aforementioned drawback to improve entropy of the extracted string, and also enhance the system security. Specifically, we design a binarization scheme, in which the distribution population of gait features are analyzed and utilized to allow the extracted binary string achieving maximal entropy. In addition, the binarization is also designed to provide strong variation toleration to produce highly stable binary string which enhances the system friendliness. We analyzed the proposed method using a gait dataset of 38 volunteers which were collected under nearly realistic conditions. The experiment results show that our proposed binarization method improves the extracted binary string's entropy 30%, and the system achieved competitive performance (i.e., 0.01% FAR, 9.5% FRR with 139-bit key).","PeriodicalId":448008,"journal":{"name":"Proceedings of the 9th International Symposium on Information and Communication Technology","volume":"50 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131369027","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Large Scale Fashion Search System with Deep Learning and Quantization Indexing 基于深度学习和量化索引的大规模时尚搜索系统
Thoi Hoang Dinh, Toan Pham Van, Ta Minh Thanh, Hau Nguyen Thanh, Anh Pham Hoang
Recently, the problems of clothes recognition and clothing item retrieval have attracted a number of researchers, due to its practical and potential values to real-world applications. The main task is to automatically find relevant clothing items given a single user-provided image without any extra metadata. Most existing systems mainly focus on clothes classification, attribute prediction, and matching the exact in-shop items with the query image. However, these systems do not mention the problem of latency period or the amount of time that users have to wait when they query an image until the query results are retrieved. In this paper, we propose a fashion search system that automatically recognizes clothes and suggests multiple similar clothing items with an impressively low latency. Through extensive experiments, it is verified that our system outperforms almost existing systems in term of clothing item retrieval time.
近年来,服装识别和服装项目检索问题因其实用性和潜在的应用价值而引起了许多研究者的关注。主要任务是在没有任何额外元数据的情况下,根据单个用户提供的图像自动查找相关的服装项目。大多数现有的系统主要集中在服装分类、属性预测以及与查询图像匹配准确的店内商品上。但是,这些系统没有提到延迟期问题,也没有提到用户在查询图像时必须等待的时间量,直到检索到查询结果。在本文中,我们提出了一个时尚搜索系统,它可以自动识别衣服,并以极低的延迟推荐多个相似的服装项目。通过大量的实验证明,我们的系统在服装检索时间方面优于几乎现有的系统。
{"title":"Large Scale Fashion Search System with Deep Learning and Quantization Indexing","authors":"Thoi Hoang Dinh, Toan Pham Van, Ta Minh Thanh, Hau Nguyen Thanh, Anh Pham Hoang","doi":"10.1145/3287921.3287964","DOIUrl":"https://doi.org/10.1145/3287921.3287964","url":null,"abstract":"Recently, the problems of clothes recognition and clothing item retrieval have attracted a number of researchers, due to its practical and potential values to real-world applications. The main task is to automatically find relevant clothing items given a single user-provided image without any extra metadata. Most existing systems mainly focus on clothes classification, attribute prediction, and matching the exact in-shop items with the query image. However, these systems do not mention the problem of latency period or the amount of time that users have to wait when they query an image until the query results are retrieved. In this paper, we propose a fashion search system that automatically recognizes clothes and suggests multiple similar clothing items with an impressively low latency. Through extensive experiments, it is verified that our system outperforms almost existing systems in term of clothing item retrieval time.","PeriodicalId":448008,"journal":{"name":"Proceedings of the 9th International Symposium on Information and Communication Technology","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127942272","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Vietnamese Speaker Authentication Using Deep Models 使用深度模型的越南语说话人身份验证
Son T. Nguyen, Viet Dac Lai, Quyen Dam-Ba, Anh Nguyen-Xuan, Cuong Pham
Speaker Authentication is the identification of a user from voice biometrics and has a wide range of applications such as banking security, human computer interaction and ambient authentication. In this work, we investigate the effectiveness of acoustic features such as Mel-frequency cepstral coefficients (MFCC), Gammatone frequency cepstral coefficients (GFCC), and Linear Predictive Codes (LPC) extracted from audio streams for constructing feature spectral images. In addition, we propose to use the deep Residual Network models for user verification from feature spectrum images. We evaluate our proposed method under two settings over the dataset collected from 20 Vietnamese speakers. The results, with the Equal Error rate of around 4%, have demonstrated that the feasibility of Vietnamese speaker authentication by using deep Residual Network models trained with GFCC spectral feature images.
说话人身份验证是通过语音生物识别技术对用户进行身份识别,在银行安全、人机交互和环境身份验证等领域有着广泛的应用。在这项工作中,我们研究了从音频流中提取的mel频率倒谱系数(MFCC)、gamma酮频率倒谱系数(GFCC)和线性预测码(LPC)等声学特征用于构建特征频谱图像的有效性。此外,我们建议使用深度残差网络模型对特征光谱图像进行用户验证。我们在两种设置下对从20名越南语使用者收集的数据集进行了评估。结果表明,使用GFCC光谱特征图像训练的深度残差网络模型进行越南语说话人身份验证是可行的,误差率约为4%。
{"title":"Vietnamese Speaker Authentication Using Deep Models","authors":"Son T. Nguyen, Viet Dac Lai, Quyen Dam-Ba, Anh Nguyen-Xuan, Cuong Pham","doi":"10.1145/3287921.3287954","DOIUrl":"https://doi.org/10.1145/3287921.3287954","url":null,"abstract":"Speaker Authentication is the identification of a user from voice biometrics and has a wide range of applications such as banking security, human computer interaction and ambient authentication. In this work, we investigate the effectiveness of acoustic features such as Mel-frequency cepstral coefficients (MFCC), Gammatone frequency cepstral coefficients (GFCC), and Linear Predictive Codes (LPC) extracted from audio streams for constructing feature spectral images. In addition, we propose to use the deep Residual Network models for user verification from feature spectrum images. We evaluate our proposed method under two settings over the dataset collected from 20 Vietnamese speakers. The results, with the Equal Error rate of around 4%, have demonstrated that the feasibility of Vietnamese speaker authentication by using deep Residual Network models trained with GFCC spectral feature images.","PeriodicalId":448008,"journal":{"name":"Proceedings of the 9th International Symposium on Information and Communication Technology","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116647575","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
Proceedings of the 9th International Symposium on Information and Communication Technology
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1