首页 > 最新文献

2008 IEEE International Conference on Research, Innovation and Vision for the Future in Computing and Communication Technologies最新文献

英文 中文
Named entity disambiguation on an ontology enriched by Wikipedia 维基百科丰富的本体上的命名实体消歧
Hien T. Nguyen, T. Cao
Currently, for named entity disambiguation, the short-age of training data is a problem. This paper presents a novel method that overcomes this problem by automatically generating an annotated corpus based on a specific ontology. Then the corpus was enriched with new and informative features extracted from Wikipedia data. Moreover, rather than pursuing rule-based methods as in literature, we employ a machine learning model to not only disambiguate but also identify named entities. In addition, our method explores in details the use of a range of features extracted from texts, a given ontology, and Wikipedia data for disambiguation. This paper also systematically analyzes impacts of the features on disambiguation accuracy by varying their combinations for representing named entities. Empirical evaluation shows that, while the ontology provides basic features of named entities, Wikipedia is a fertile source for additional features to construct accurate and robust named entity disambiguation systems.
目前,对于命名实体消歧来说,训练数据的不足是一个问题。本文提出了一种基于特定本体自动生成带注释的语料库的方法来克服这一问题。然后,从维基百科数据中提取新的信息特征来丰富语料库。此外,我们没有像文献中那样追求基于规则的方法,而是采用了机器学习模型,不仅消除了歧义,而且还识别了命名实体。此外,我们的方法详细探讨了从文本、给定本体和维基百科数据中提取的一系列特征的使用,以消除歧义。本文还系统地分析了特征在表示命名实体时不同组合对消歧精度的影响。经验评估表明,虽然本体提供了命名实体的基本特征,但维基百科是构建准确和健壮的命名实体消歧系统的附加特征的丰富来源。
{"title":"Named entity disambiguation on an ontology enriched by Wikipedia","authors":"Hien T. Nguyen, T. Cao","doi":"10.1109/RIVF.2008.4586363","DOIUrl":"https://doi.org/10.1109/RIVF.2008.4586363","url":null,"abstract":"Currently, for named entity disambiguation, the short-age of training data is a problem. This paper presents a novel method that overcomes this problem by automatically generating an annotated corpus based on a specific ontology. Then the corpus was enriched with new and informative features extracted from Wikipedia data. Moreover, rather than pursuing rule-based methods as in literature, we employ a machine learning model to not only disambiguate but also identify named entities. In addition, our method explores in details the use of a range of features extracted from texts, a given ontology, and Wikipedia data for disambiguation. This paper also systematically analyzes impacts of the features on disambiguation accuracy by varying their combinations for representing named entities. Empirical evaluation shows that, while the ontology provides basic features of named entities, Wikipedia is a fertile source for additional features to construct accurate and robust named entity disambiguation systems.","PeriodicalId":233667,"journal":{"name":"2008 IEEE International Conference on Research, Innovation and Vision for the Future in Computing and Communication Technologies","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115145886","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
A dependency-based word reordering approach for Statistical Machine Translation 基于依赖关系的统计机器翻译词重排序方法
Cong Duy Vu Hoang, Mai Ngo, Dinh Dien
Reordering is of crucial importance for machine translation. Solving the reordering problem can lead to remarkable improvements in translation performance. In this paper, we propose a novel approach to solve the word reordering problem in statistical machine translation. We rely on the dependency relations retrieved from a statistical parser incorporating with linguistic hand-crafted rules to create the transformations. These dependency-based transformations can produce the problem of word movement on both phrase and word reordering which is a difficult problem on parse tree based approaches. Such transformations are then applied as a preprocessor to English language both in training and decoding process to obtain an underlying word order closer to the Vietnamese language. About the hand-crafted rules, we extract from the syntactic differences of word order between English and Vietnamese language. This approach is simple and easy to implement with a small rule set, not lead to the rule explosion. We describe the experiments using our model on VCLEVC corpus [18] and consider the translation from English to Vietnamese, showing significant improvements about 2-4% BLEU score in comparison with the MOSES phrase-based baseline system [19].
重新排序对机器翻译至关重要。解决重新排序问题可以显著提高翻译性能。在本文中,我们提出了一种新的方法来解决统计机器翻译中的词重排问题。我们依赖于从统计解析器检索到的依赖关系,并结合语言手工制作的规则来创建转换。这些基于依赖关系的转换会产生词在短语上的移动问题和词的重新排序问题,这是基于解析树方法的一个难题。然后将这些转换作为预处理程序应用于英语语言的训练和解码过程中,以获得更接近越南语的底层词序。关于手工规则,我们提取了英语和越南语的词序句法差异。这种方法简单且易于用小规则集实现,不会导致规则爆炸。我们描述了在VCLEVC语料库上使用我们的模型进行的实验[18],并考虑了从英语到越南语的翻译,与基于MOSES短语的基线系统相比,BLEU得分显著提高了2-4%[19]。
{"title":"A dependency-based word reordering approach for Statistical Machine Translation","authors":"Cong Duy Vu Hoang, Mai Ngo, Dinh Dien","doi":"10.1109/RIVF.2008.4586343","DOIUrl":"https://doi.org/10.1109/RIVF.2008.4586343","url":null,"abstract":"Reordering is of crucial importance for machine translation. Solving the reordering problem can lead to remarkable improvements in translation performance. In this paper, we propose a novel approach to solve the word reordering problem in statistical machine translation. We rely on the dependency relations retrieved from a statistical parser incorporating with linguistic hand-crafted rules to create the transformations. These dependency-based transformations can produce the problem of word movement on both phrase and word reordering which is a difficult problem on parse tree based approaches. Such transformations are then applied as a preprocessor to English language both in training and decoding process to obtain an underlying word order closer to the Vietnamese language. About the hand-crafted rules, we extract from the syntactic differences of word order between English and Vietnamese language. This approach is simple and easy to implement with a small rule set, not lead to the rule explosion. We describe the experiments using our model on VCLEVC corpus [18] and consider the translation from English to Vietnamese, showing significant improvements about 2-4% BLEU score in comparison with the MOSES phrase-based baseline system [19].","PeriodicalId":233667,"journal":{"name":"2008 IEEE International Conference on Research, Innovation and Vision for the Future in Computing and Communication Technologies","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115689832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Performance improvement of speech recognition system using microphone array 基于麦克风阵列的语音识别系统性能改进
D. C. Nguyen, Guanghu Shen, Ho-Youl Jung, Hyun-Yeol Chung
In this paper, we present some methods to improve the performance of microphone array speech recognition system based on Limabeam algorithm. For improving recognition accuracy, we proposed weighted Mahalanobis distance (WMD) based on traditional distance measure in a Gaussian classifier and is a modified method to give weights for different features in it according to their distances after the variance normalization. Experimental results showed that Limabeam adopted weighted Mahalanobis distance measure (WMD-Limabeam) improves recognition performance significantly than those by original Limabeam. In compared experiments with some other extended versions of Limabeam algorithm such as subband Limabeam and N-best parallel model for unsupervised Limabeam, we could see that the WMD-Limabeam show higher recognition accuracy. In cases of the system that adopted WMD, we obtained correct word recognition rate of 89.4% for calibrate Limabeam and 84.6% for unsupervised Limabeam, 3.0% and 5.0% higher than original Limabeam respectively. This rate also results in 9.0% higher than delay and sum algorithm.
本文提出了几种基于Limabeam算法的改进麦克风阵列语音识别系统性能的方法。为了提高识别精度,我们提出了基于高斯分类器中传统距离度量的加权马氏距离(weighted Mahalanobis distance, WMD)方法,该方法是对高斯分类器中不同特征在方差归一化后根据距离给予权重的改进方法。实验结果表明,采用加权马氏距离测度(WMD-Limabeam)的Limabeam算法比原始的Limabeam算法显著提高了识别性能。通过与其他limabam算法的扩展版本,如子带limabam和无监督limabam的N-best并行模型的实验对比,我们可以看到wmd - limabam算法具有更高的识别精度。在采用WMD的情况下,我们获得了校准Limabeam和无监督Limabeam的正确单词识别率分别为89.4%和84.6%,分别比原始Limabeam高3.0%和5.0%。这个速率也比延迟和和算法高9.0%。
{"title":"Performance improvement of speech recognition system using microphone array","authors":"D. C. Nguyen, Guanghu Shen, Ho-Youl Jung, Hyun-Yeol Chung","doi":"10.1109/RIVF.2008.4586338","DOIUrl":"https://doi.org/10.1109/RIVF.2008.4586338","url":null,"abstract":"In this paper, we present some methods to improve the performance of microphone array speech recognition system based on Limabeam algorithm. For improving recognition accuracy, we proposed weighted Mahalanobis distance (WMD) based on traditional distance measure in a Gaussian classifier and is a modified method to give weights for different features in it according to their distances after the variance normalization. Experimental results showed that Limabeam adopted weighted Mahalanobis distance measure (WMD-Limabeam) improves recognition performance significantly than those by original Limabeam. In compared experiments with some other extended versions of Limabeam algorithm such as subband Limabeam and N-best parallel model for unsupervised Limabeam, we could see that the WMD-Limabeam show higher recognition accuracy. In cases of the system that adopted WMD, we obtained correct word recognition rate of 89.4% for calibrate Limabeam and 84.6% for unsupervised Limabeam, 3.0% and 5.0% higher than original Limabeam respectively. This rate also results in 9.0% higher than delay and sum algorithm.","PeriodicalId":233667,"journal":{"name":"2008 IEEE International Conference on Research, Innovation and Vision for the Future in Computing and Communication Technologies","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128502406","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A hybrid approach of AdaBoost and Artificial Neural Network for detecting human faces AdaBoost与人工神经网络的混合人脸检测方法
T. Le, L. Bui
The human face image recognition is one of the prominent problems at present. Recognizing human faces correctly will aid some fields such as national defense and person verification. One of the most vital processing of recognizing face images is to detect human faces in the images. Some approaches have been used to detect human faces. However, they still have some limitations. In the paper, we will consider some popular methods, AdaBoost, artificial neural network (ANN) etc., for detecting human faces. Then we will propose a hybrid model of combining AdaBoost and artificial neural network to solve the process efficiently. The system which was build from the proposed model has been conducted on database CalTech. The recognition correctness is more than 96%. It shows the feasibility of the proposed model.
人脸图像识别是当前的突出问题之一。正确识别人脸将有助于国防、身份验证等领域。人脸图像识别中最重要的处理之一是检测图像中的人脸。一些方法已经被用来检测人脸。然而,它们仍然有一些局限性。在本文中,我们将考虑一些流行的人脸检测方法,如AdaBoost、人工神经网络(ANN)等。然后,我们将提出一种结合AdaBoost和人工神经网络的混合模型来高效地解决这一过程。基于该模型构建的系统已在数据库CalTech上运行。识别正确率达96%以上。验证了所提模型的可行性。
{"title":"A hybrid approach of AdaBoost and Artificial Neural Network for detecting human faces","authors":"T. Le, L. Bui","doi":"10.1109/RIVF.2008.4586336","DOIUrl":"https://doi.org/10.1109/RIVF.2008.4586336","url":null,"abstract":"The human face image recognition is one of the prominent problems at present. Recognizing human faces correctly will aid some fields such as national defense and person verification. One of the most vital processing of recognizing face images is to detect human faces in the images. Some approaches have been used to detect human faces. However, they still have some limitations. In the paper, we will consider some popular methods, AdaBoost, artificial neural network (ANN) etc., for detecting human faces. Then we will propose a hybrid model of combining AdaBoost and artificial neural network to solve the process efficiently. The system which was build from the proposed model has been conducted on database CalTech. The recognition correctness is more than 96%. It shows the feasibility of the proposed model.","PeriodicalId":233667,"journal":{"name":"2008 IEEE International Conference on Research, Innovation and Vision for the Future in Computing and Communication Technologies","volume":"81 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126011051","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Rules discovery: Transfer and generalization 规则发现:转移和泛化
Anh Nguyen-Xuan, C. Tijus
This paper presents a study about human transfer of learning between isomorphs and on the conditions of rule discovery that could be of help for machine learning. When faced with a new problem, the human learner uses the knowledge s/he already possesses in order to mentally represent and manipulate the objects s/he has to deal within the process of problem solving. We propose that familiar domain knowledge provide concepts as useful biases for discovering general rules when solving isomorphic problems as well as problems which entail a larger problem space. Results of two experiments using isomorphic versions of the ldquorule discoveryrdquo Nim game, and versions that entail a larger problem space, suggest that participants use the even concept in the familiar domain according to external representations that can either predictably favor or impair learning and transfer in a foreseeable way.
本文提出了一项关于人类同构之间学习迁移的研究,并在规则发现的条件下对机器学习有帮助。当面对一个新问题时,人类学习者使用他/她已经拥有的知识,以便在解决问题的过程中在心理上表征和操纵他/她必须处理的对象。我们提出,在解决同构问题以及涉及更大问题空间的问题时,熟悉的领域知识为发现一般规则提供了有用的偏差概念。使用同构版本的规则发现Nim游戏和包含更大问题空间的版本的两个实验结果表明,参与者根据外部表征在熟悉领域使用偶数概念,这些表征可以以可预见的方式促进或损害学习和迁移。
{"title":"Rules discovery: Transfer and generalization","authors":"Anh Nguyen-Xuan, C. Tijus","doi":"10.1109/RIVF.2008.4586326","DOIUrl":"https://doi.org/10.1109/RIVF.2008.4586326","url":null,"abstract":"This paper presents a study about human transfer of learning between isomorphs and on the conditions of rule discovery that could be of help for machine learning. When faced with a new problem, the human learner uses the knowledge s/he already possesses in order to mentally represent and manipulate the objects s/he has to deal within the process of problem solving. We propose that familiar domain knowledge provide concepts as useful biases for discovering general rules when solving isomorphic problems as well as problems which entail a larger problem space. Results of two experiments using isomorphic versions of the ldquorule discoveryrdquo Nim game, and versions that entail a larger problem space, suggest that participants use the even concept in the familiar domain according to external representations that can either predictably favor or impair learning and transfer in a foreseeable way.","PeriodicalId":233667,"journal":{"name":"2008 IEEE International Conference on Research, Innovation and Vision for the Future in Computing and Communication Technologies","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124428195","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhance exploring temporal correlation for data collection in WSNs 加强对无线传感器网络数据采集时间相关性的探索
N. D. Pham, Trong Duc Le, Hyunseung Choo
Continuous data collection applications in wireless sensor networks require sensor nodes to continuously sample the surrounding physical phenomenon and then return the data to a processing center. Battery-operated sensors have to avoid heavy use of their wireless radio by compressing the time series sensed data instead of transmitting it in raw form. One of the most commonly used compacting methods is piecewise linear approximation. Previously, Liu et al. proposed a greedy PLAMLiS algorithm to approximate the time series into a number of line segments running in Theta(n2logn) time, however this is not appropriate for processing in the sensors. Therefore, based on our study we propose an alternative algorithm which obtains the same result but needs a shorter running time. Based on theoretical analysis and comprehensive simulations, it is shown that the new proposed algorithm has a competitive computational cost of Theta(nlogn) as well as reducing the number of line segments and so it can decrease the overall radio transmission load in order to save energy of the sensor nodes.
无线传感器网络中的连续数据采集应用要求传感器节点连续采样周围的物理现象,然后将数据返回到处理中心。电池驱动的传感器必须通过压缩时间序列感知数据而不是以原始形式传输数据来避免大量使用无线无线电。最常用的压缩方法之一是分段线性逼近。此前,Liu等人提出了一种贪心PLAMLiS算法,将时间序列近似为在Theta(n2logn)时间内运行的许多线段,但这并不适合在传感器中进行处理。因此,在我们研究的基础上,我们提出了一种替代算法,可以获得相同的结果,但需要更短的运行时间。理论分析和综合仿真结果表明,该算法具有较低的Theta(nlogn)计算成本,减少了线段的数量,从而降低了整个无线传输负荷,从而节省了传感器节点的能量。
{"title":"Enhance exploring temporal correlation for data collection in WSNs","authors":"N. D. Pham, Trong Duc Le, Hyunseung Choo","doi":"10.1109/RIVF.2008.4586356","DOIUrl":"https://doi.org/10.1109/RIVF.2008.4586356","url":null,"abstract":"Continuous data collection applications in wireless sensor networks require sensor nodes to continuously sample the surrounding physical phenomenon and then return the data to a processing center. Battery-operated sensors have to avoid heavy use of their wireless radio by compressing the time series sensed data instead of transmitting it in raw form. One of the most commonly used compacting methods is piecewise linear approximation. Previously, Liu et al. proposed a greedy PLAMLiS algorithm to approximate the time series into a number of line segments running in Theta(n2logn) time, however this is not appropriate for processing in the sensors. Therefore, based on our study we propose an alternative algorithm which obtains the same result but needs a shorter running time. Based on theoretical analysis and comprehensive simulations, it is shown that the new proposed algorithm has a competitive computational cost of Theta(nlogn) as well as reducing the number of line segments and so it can decrease the overall radio transmission load in order to save energy of the sensor nodes.","PeriodicalId":233667,"journal":{"name":"2008 IEEE International Conference on Research, Innovation and Vision for the Future in Computing and Communication Technologies","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128753380","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 32
On the bounded integer programming 关于有界整数规划
T. Khoát
The best upper time bound for solving the bounded integer programming (BIP) up to now is poly(phi) ldr n2n+o(n), where n and phi are the dimension and the input size of the problem respectively. In this paper, we show that BIP is solvable in deterministic time poly(phi) ldr nn+o(n). Moreover we also show that under some reasonable assumptions, BIP is solvable in probabilistic time 2O(n).
目前解决有界整数规划(BIP)的最佳上界是poly(phi) ldr n2n+o(n),其中n和phi分别是问题的维数和输入大小。在本文中,我们证明了BIP在确定性时间多边形(phi) ldr nn+o(n)中是可解的。此外,我们还证明了在一些合理的假设下,BIP在概率时间20 (n)内是可解的。
{"title":"On the bounded integer programming","authors":"T. Khoát","doi":"10.1109/RIVF.2008.4586328","DOIUrl":"https://doi.org/10.1109/RIVF.2008.4586328","url":null,"abstract":"The best upper time bound for solving the bounded integer programming (BIP) up to now is poly(phi) ldr n2n+o(n), where n and phi are the dimension and the input size of the problem respectively. In this paper, we show that BIP is solvable in deterministic time poly(phi) ldr nn+o(n). Moreover we also show that under some reasonable assumptions, BIP is solvable in probabilistic time 2O(n).","PeriodicalId":233667,"journal":{"name":"2008 IEEE International Conference on Research, Innovation and Vision for the Future in Computing and Communication Technologies","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129564413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Robust overlay network with Self-Adaptive topology: Protocol description 具有自适应拓扑的鲁棒覆盖网络:协议描述
L. Baud, N. Pham, P. Bellot
We introduce a new overlay network named ROSA1. Overlay networks offer a way to bypass the routing constraints of the underlying network. ROSA used this overlay network property to offer a resilient routing to critical applications. Unlike other overlay networks dealing with the routing resilience issue, we oriented our research towards building a robust overlay network topology instead of a robust routing function. We tried to maintain a path between any pairs of nodes of the network. The routing resilience is obtained by forcing nodes to choose and modify dynamically their neighbors set according to the ROSA protocol. Moreover, ROSA is highly scalable.
本文介绍了一种新的覆盖网络——ROSA1。覆盖网络提供了一种绕过底层网络路由约束的方法。ROSA使用这种覆盖网络属性为关键应用程序提供弹性路由。与其他处理路由弹性问题的覆盖网络不同,我们的研究方向是建立一个鲁棒的覆盖网络拓扑,而不是鲁棒的路由功能。我们试图在网络的任意对节点之间保持一条路径。路由弹性是通过强制节点动态选择和修改根据ROSA协议设置的邻居来实现的。此外,ROSA具有高度可扩展性。
{"title":"Robust overlay network with Self-Adaptive topology: Protocol description","authors":"L. Baud, N. Pham, P. Bellot","doi":"10.1109/RIVF.2008.4586348","DOIUrl":"https://doi.org/10.1109/RIVF.2008.4586348","url":null,"abstract":"We introduce a new overlay network named ROSA1. Overlay networks offer a way to bypass the routing constraints of the underlying network. ROSA used this overlay network property to offer a resilient routing to critical applications. Unlike other overlay networks dealing with the routing resilience issue, we oriented our research towards building a robust overlay network topology instead of a robust routing function. We tried to maintain a path between any pairs of nodes of the network. The routing resilience is obtained by forcing nodes to choose and modify dynamically their neighbors set according to the ROSA protocol. Moreover, ROSA is highly scalable.","PeriodicalId":233667,"journal":{"name":"2008 IEEE International Conference on Research, Innovation and Vision for the Future in Computing and Communication Technologies","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128932809","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Factorial Correspondence Analysis for image retrieval 用于图像检索的因子对应分析
Nguyen-Khang Pham, A. Morin, P. Gros, Quyet-Thang Le
We are concerned by the use of factorial correspondence analysis (FCA) for image retrieval. FCA is designed for analyzing contingency tables. In textual data analysis (TDA), FCA analyzes a contingency table crossing terms/words and documents. To adapt FCA on images, we first define "visual words" computed from scalable invariant feature transform (SIFT) descriptors in images and use them for image quantization. At this step, we can build a contingency table crossing "visual words" as terms/words and images as documents. The method was tested on the Caltech4 and Stewenius and Nister datasets on which it provides better results (quality of results and execution time) than classical methods as tf * idf and probabilistic latent semantic analysis (PLSA). To scale up and improve the retrieval quality, we propose a new retrieval schema using inverted files based on the relevant indicators of correspondence analysis (representation quality of images on axes and contribution of images to the inertia of the axes). The numerical experiments show that our algorithm performs faster than the exhaustive method without losing precision.
我们关注的是使用阶乘对应分析(FCA)的图像检索。FCA设计用于分析列联表。在文本数据分析(TDA)中,FCA分析一个包含术语/单词和文档的列联表。为了使FCA适用于图像,我们首先定义了从图像中的可缩放不变特征变换(SIFT)描述符计算的“视觉词”,并将其用于图像量化。在这个步骤中,我们可以构建一个将“视觉单词”作为术语/单词和图像作为文档的列联表。该方法在Caltech4和Stewenius and Nister数据集上进行了测试,在结果质量和执行时间上,它比经典方法如tf * idf和概率潜在语义分析(PLSA)提供了更好的结果。为了扩大检索规模并提高检索质量,本文基于对应分析的相关指标(图像在坐标轴上的表示质量和图像对坐标轴惯性的贡献),提出了一种基于倒立文件的检索模式。数值实验表明,该算法比穷举法更快,且精度不低。
{"title":"Factorial Correspondence Analysis for image retrieval","authors":"Nguyen-Khang Pham, A. Morin, P. Gros, Quyet-Thang Le","doi":"10.1109/RIVF.2008.4586366","DOIUrl":"https://doi.org/10.1109/RIVF.2008.4586366","url":null,"abstract":"We are concerned by the use of factorial correspondence analysis (FCA) for image retrieval. FCA is designed for analyzing contingency tables. In textual data analysis (TDA), FCA analyzes a contingency table crossing terms/words and documents. To adapt FCA on images, we first define \"visual words\" computed from scalable invariant feature transform (SIFT) descriptors in images and use them for image quantization. At this step, we can build a contingency table crossing \"visual words\" as terms/words and images as documents. The method was tested on the Caltech4 and Stewenius and Nister datasets on which it provides better results (quality of results and execution time) than classical methods as tf * idf and probabilistic latent semantic analysis (PLSA). To scale up and improve the retrieval quality, we propose a new retrieval schema using inverted files based on the relevant indicators of correspondence analysis (representation quality of images on axes and contribution of images to the inertia of the axes). The numerical experiments show that our algorithm performs faster than the exhaustive method without losing precision.","PeriodicalId":233667,"journal":{"name":"2008 IEEE International Conference on Research, Innovation and Vision for the Future in Computing and Communication Technologies","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121591377","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Speeding up subset seed algorithm for intensive protein sequence comparison 加速子集种子算法在蛋白质序列密集比较中的应用
Van Hoa Nguyen, D. Lavenier
Sequence similarity search is a common and repeated task in molecular biology. The rapid growth of genomic databases leads to the need of speeding up the treatment of this task. In this paper, we present a subset seed algorithm for intensive protein sequence comparison. We have accelerated this algorithm by using indexing technique and fine grained parallelism of GPU and SIMD instructions. We have implemented two programs: iBLASTP, iTBLASTN. The GPU (SIMD) implementation of the two programs achieves a speed up ranging from 5.5 to 10 (4 to 5.6) compared to the BLASTP and TBLASTN of the BLAST program family, with comparable sensitivity.
序列相似性搜索是分子生物学中一项常见且重复的任务。基因组数据库的快速增长导致需要加快处理这一任务。本文提出了一种用于蛋白质序列密集比较的子集种子算法。我们利用索引技术和GPU和SIMD指令的细粒度并行性来加速该算法。我们实现了两个程序:iBLASTP, iTBLASTN。与BLAST程序家族的BLASTP和TBLASTN相比,这两个程序的GPU (SIMD)实现的速度范围从5.5到10(4到5.6),具有相当的灵敏度。
{"title":"Speeding up subset seed algorithm for intensive protein sequence comparison","authors":"Van Hoa Nguyen, D. Lavenier","doi":"10.1109/RIVF.2008.4586333","DOIUrl":"https://doi.org/10.1109/RIVF.2008.4586333","url":null,"abstract":"Sequence similarity search is a common and repeated task in molecular biology. The rapid growth of genomic databases leads to the need of speeding up the treatment of this task. In this paper, we present a subset seed algorithm for intensive protein sequence comparison. We have accelerated this algorithm by using indexing technique and fine grained parallelism of GPU and SIMD instructions. We have implemented two programs: iBLASTP, iTBLASTN. The GPU (SIMD) implementation of the two programs achieves a speed up ranging from 5.5 to 10 (4 to 5.6) compared to the BLASTP and TBLASTN of the BLAST program family, with comparable sensitivity.","PeriodicalId":233667,"journal":{"name":"2008 IEEE International Conference on Research, Innovation and Vision for the Future in Computing and Communication Technologies","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125693645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
期刊
2008 IEEE International Conference on Research, Innovation and Vision for the Future in Computing and Communication Technologies
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1