首页 > 最新文献

2014 22nd International Conference on Pattern Recognition最新文献

英文 中文
On Validation of Clustering Techniques for Bibliographic Databases 书目数据库聚类技术的验证
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.543
Sumit Mishra, S. Saha, S. Mondal
In entity name disambiguation, performance evaluation of any approach is difficult. This is due to the fact that correct or actual results are often not known. Generally for evaluation purpose, three measures namely precision, recall and f-measure are used. They all are external validity indices because they need golden standard data. But in Bibliographic databases like DBLP, Arnetminer, Scopus, Web of Science, Google Scholar, etc., gold standard data is not easily available and it is very difficult to obtain this due to the overlapping nature of data. So, there is a need to use some other matrices for evaluation purpose. In this paper, some internal cluster validity index based schemes are proposed for evaluating entity name disambiguation algorithms when applied on bibliographic data without using any gold standard datasets. Two new internal validity indices are also proposed in the current paper for this purpose. Experimental results shown on seven bibliographic datasets reveal that proposed internal cluster validity indices are able to compare the results obtained by different methods without prior/gold standard. Thus the present paper demonstrates a novel way of evaluating any entity matching algorithm for bibliographic datasets without using any prior/gold standard information.
在实体名称消歧中,任何一种方法的性能评价都是困难的。这是因为正确或实际的结果往往是未知的。一般来说,为了评价目的,使用三个指标,即精度、召回率和f-measure。它们都是外部有效性指标,因为它们需要黄金标准数据。但在DBLP、Arnetminer、Scopus、Web of Science、Google Scholar等书目数据库中,黄金标准数据并不容易获得,而且由于数据的重叠性质,很难获得黄金标准数据。因此,有必要使用一些其他的矩阵来求值。本文在不使用任何金标准数据集的情况下,提出了一些基于内部聚类有效性索引的评价书目数据实体名消歧算法的方案。为此,本文还提出了两个新的内部效度指标。在7个文献数据集上的实验结果表明,本文提出的内部聚类效度指标能够比较不同方法的结果,而不需要先验标准或金标准。因此,本文展示了一种新的方法来评估书目数据集的任何实体匹配算法,而不使用任何先验/金标准信息。
{"title":"On Validation of Clustering Techniques for Bibliographic Databases","authors":"Sumit Mishra, S. Saha, S. Mondal","doi":"10.1109/ICPR.2014.543","DOIUrl":"https://doi.org/10.1109/ICPR.2014.543","url":null,"abstract":"In entity name disambiguation, performance evaluation of any approach is difficult. This is due to the fact that correct or actual results are often not known. Generally for evaluation purpose, three measures namely precision, recall and f-measure are used. They all are external validity indices because they need golden standard data. But in Bibliographic databases like DBLP, Arnetminer, Scopus, Web of Science, Google Scholar, etc., gold standard data is not easily available and it is very difficult to obtain this due to the overlapping nature of data. So, there is a need to use some other matrices for evaluation purpose. In this paper, some internal cluster validity index based schemes are proposed for evaluating entity name disambiguation algorithms when applied on bibliographic data without using any gold standard datasets. Two new internal validity indices are also proposed in the current paper for this purpose. Experimental results shown on seven bibliographic datasets reveal that proposed internal cluster validity indices are able to compare the results obtained by different methods without prior/gold standard. Thus the present paper demonstrates a novel way of evaluating any entity matching algorithm for bibliographic datasets without using any prior/gold standard information.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125519907","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Pain Intensity Evaluation through Facial Action Units 通过面部动作单元评估疼痛强度
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.803
Zuhair Zafar, N. Khan
In this work we present a system that enables automatic estimation of Pain from image sequences with frontal views of faces. The system uses facial characteristic points to characterize different Action Units (AU) of pain and is able to operate in cluttered and dynamic scenes. Geometric features are computed using 22 facial characteristic points. We use k-NN classifier for classifying AU. Only action units relevant to pain are classified. Validation studies are done on UNBC McMaster Shoulder Pain Archive Database [8]. We also classify action unit intensities for evaluating pain intensity on a 16 point scale. Our system is simpler in design compared to the already reported works in literature. Our system reports AU intensities on a standard scale and also reports pain intensity to assess pain. We have achieved more than 84% accuracy for AU intensity levels and 87.4% area under ROC curve for pain assessment as compared to 84% of state-of-the-art scheme.
在这项工作中,我们提出了一种系统,可以自动估计面部正面视图图像序列的疼痛。该系统使用面部特征点来表征疼痛的不同动作单元(AU),并能够在混乱和动态的场景中运行。利用22个面部特征点计算几何特征。我们使用k-NN分类器对AU进行分类。只有与疼痛相关的动作单位才被分类。验证研究是在UNBC McMaster肩痛档案数据库[8]上完成的。我们还分类行动单位强度评估疼痛强度在16分量表。与已有文献报道的作品相比,我们的系统在设计上更简单。我们的系统以标准尺度报告AU强度,也报告疼痛强度以评估疼痛。我们在AU强度水平上的准确率超过84%,在疼痛评估的ROC曲线下的面积达到87.4%,而最先进的方案为84%。
{"title":"Pain Intensity Evaluation through Facial Action Units","authors":"Zuhair Zafar, N. Khan","doi":"10.1109/ICPR.2014.803","DOIUrl":"https://doi.org/10.1109/ICPR.2014.803","url":null,"abstract":"In this work we present a system that enables automatic estimation of Pain from image sequences with frontal views of faces. The system uses facial characteristic points to characterize different Action Units (AU) of pain and is able to operate in cluttered and dynamic scenes. Geometric features are computed using 22 facial characteristic points. We use k-NN classifier for classifying AU. Only action units relevant to pain are classified. Validation studies are done on UNBC McMaster Shoulder Pain Archive Database [8]. We also classify action unit intensities for evaluating pain intensity on a 16 point scale. Our system is simpler in design compared to the already reported works in literature. Our system reports AU intensities on a standard scale and also reports pain intensity to assess pain. We have achieved more than 84% accuracy for AU intensity levels and 87.4% area under ROC curve for pain assessment as compared to 84% of state-of-the-art scheme.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"388 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116011293","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 33
Semantic Urban Maps 语义城市地图
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.694
J. R. Siddiqui, S. Khatibi
A novel region based 3D semantic mapping method is proposed for urban scenes. The proposed Semantic Urban Maps (SUM) method labels the regions of segmented images into a set of geometric and semantic classes simultaneously by employing a Markov Random Field based classification framework. The pixels in the labeled images are back-projected into a set of 3D point-clouds using stereo disparity. The point-clouds are registered together by incorporating the motion estimation and a coherent semantic map representation is obtained. SUM is evaluated on five urban benchmark sequences and is demonstrated to be successful in retrieving both geometric as well as semantic labels. The comparison with relevant state-of-art method reveals that SUM is competitive and performs better than the competing method in average pixel-wise accuracy.
提出了一种基于区域的城市场景三维语义映射方法。本文提出的语义城市地图(Semantic Urban Maps, SUM)方法采用基于马尔科夫随机场的分类框架,将分割图像的区域同时标记为一组几何类和语义类。使用立体视差将标记图像中的像素反向投影成一组3D点云。结合运动估计对点云进行配准,得到连贯的语义图表示。SUM在五个城市基准序列上进行了评估,并被证明在检索几何和语义标签方面都是成功的。通过与同类方法的比较,SUM具有一定的竞争力,在平均像素精度上优于同类方法。
{"title":"Semantic Urban Maps","authors":"J. R. Siddiqui, S. Khatibi","doi":"10.1109/ICPR.2014.694","DOIUrl":"https://doi.org/10.1109/ICPR.2014.694","url":null,"abstract":"A novel region based 3D semantic mapping method is proposed for urban scenes. The proposed Semantic Urban Maps (SUM) method labels the regions of segmented images into a set of geometric and semantic classes simultaneously by employing a Markov Random Field based classification framework. The pixels in the labeled images are back-projected into a set of 3D point-clouds using stereo disparity. The point-clouds are registered together by incorporating the motion estimation and a coherent semantic map representation is obtained. SUM is evaluated on five urban benchmark sequences and is demonstrated to be successful in retrieving both geometric as well as semantic labels. The comparison with relevant state-of-art method reveals that SUM is competitive and performs better than the competing method in average pixel-wise accuracy.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122928673","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Traffic Camera Anomaly Detection 交通摄像头异常检测
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.794
Yuan-Kai Wang, Ching-Tang Fan, Jian-Fu Chen
Detection of camera anomaly and tampering have attracted increasing interest in video surveillance for real-time alert of camera malfunction. However, the anomaly detection for traffic cameras monitoring vehicles and recognizing license plates has not been formally studied and it cannot be solved by existing methods. In this paper, we propose a camera anomaly detection method for traffic scene that has distinct characteristics of dynamics due to traffic flow and traffic crowd, compared with normal surveillance scene. Image quality used as low-level features are measured by no-referenced metrics. Image dynamics used as mid-level features are computed by histogram distribution of optical flow. A two-stage classifier for the detection of anomaly is devised by the modeling of image quality and video dynamics with probabilistic state transition. The proposed approach is robust to many challenging issues in urban surveillance scenarios and has very low false alarm rate. Experiments are conducted on real-world videos recorded in traffic scene including the situations of high traffic flow and severe crowding. Our test results demonstrate that the proposed method is superior to previous methods on both precision rate and false alarm rate for the anomaly detection of traffic cameras.
对摄像机异常和篡改的检测在视频监控中引起了越来越多的关注,从而实现对摄像机故障的实时预警。然而,对于交通摄像头监控车辆和车牌识别的异常检测问题,目前还没有正式的研究,现有的方法也无法解决。本文提出了一种针对交通场景的摄像机异常检测方法,该交通场景与普通监控场景相比,由于交通流和交通人群的影响,具有明显的动态特征。作为底层特征的图像质量通过无参考度量来衡量。利用光流的直方图分布计算作为中级特征的图像动力学。通过对带有概率状态转移的图像质量和视频动态建模,设计了一种两阶段的异常检测分类器。该方法对城市监控场景中的许多具有挑战性的问题具有鲁棒性,并且具有非常低的误报率。实验采用真实的交通场景视频,包括高交通流量和严重拥挤的情况。测试结果表明,该方法在交通摄像头异常检测的准确率和虚警率上均优于现有方法。
{"title":"Traffic Camera Anomaly Detection","authors":"Yuan-Kai Wang, Ching-Tang Fan, Jian-Fu Chen","doi":"10.1109/ICPR.2014.794","DOIUrl":"https://doi.org/10.1109/ICPR.2014.794","url":null,"abstract":"Detection of camera anomaly and tampering have attracted increasing interest in video surveillance for real-time alert of camera malfunction. However, the anomaly detection for traffic cameras monitoring vehicles and recognizing license plates has not been formally studied and it cannot be solved by existing methods. In this paper, we propose a camera anomaly detection method for traffic scene that has distinct characteristics of dynamics due to traffic flow and traffic crowd, compared with normal surveillance scene. Image quality used as low-level features are measured by no-referenced metrics. Image dynamics used as mid-level features are computed by histogram distribution of optical flow. A two-stage classifier for the detection of anomaly is devised by the modeling of image quality and video dynamics with probabilistic state transition. The proposed approach is robust to many challenging issues in urban surveillance scenarios and has very low false alarm rate. Experiments are conducted on real-world videos recorded in traffic scene including the situations of high traffic flow and severe crowding. Our test results demonstrate that the proposed method is superior to previous methods on both precision rate and false alarm rate for the anomaly detection of traffic cameras.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114145652","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Multi-group Adaptation for Event Recognition from Videos 基于多组自适应的视频事件识别
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.671
Yang Feng, Xinxiao Wu, Han Wang, Jing Liu
Recognizing events in consumer videos is becoming increasingly important because of the enormous growth of consumer videos in recent years. Current researches mainly focus on learning from numerous labeled videos, which is time consuming and labor expensive due to labeling the consumer videos. To alleviate the labeling process, we utilize a large number of loosely labeled Web videos (e.g., from YouTube) for visual event recognition in consumer videos. Web videos are noisy and diverse, so brute force transfer of Web videos to consumer videos may hurt the performance. To address such a negative transfer problem, we propose a novel Multi-Group Adaptation (MGA) framework to divide the training Web videos into several semantic groups and seek the optimal weight of each group. Each weight represents how relative the corresponding group is to the consumer domain. The final classifier for event recognition is learned using the weighted combination of classifiers learned from Web videos and enforced to be smooth on the consumer domain. Comprehensive experiments on three real-world consumer video datasets demonstrate the effectiveness of MGA for event recognition in consumer videos.
由于近年来消费视频的巨大增长,识别消费视频中的事件变得越来越重要。目前的研究主要集中在对大量标注过的视频进行学习,由于需要对消费者视频进行标注,既耗时又费力。为了减轻标记过程,我们利用大量松散标记的网络视频(例如,来自YouTube)来识别消费者视频中的视觉事件。网络视频噪声大、种类多,将网络视频强行传输到消费者视频可能会影响性能。为了解决这种负迁移问题,我们提出了一种新的多组自适应(MGA)框架,将训练网络视频划分为多个语义组,并寻求每个组的最优权值。每个权重表示相应的群体与消费者领域的相对程度。事件识别的最终分类器是使用从Web视频中学习到的分类器的加权组合来学习的,并强制在消费者领域上保持平滑。在三个真实消费者视频数据集上的综合实验证明了MGA对消费者视频事件识别的有效性。
{"title":"Multi-group Adaptation for Event Recognition from Videos","authors":"Yang Feng, Xinxiao Wu, Han Wang, Jing Liu","doi":"10.1109/ICPR.2014.671","DOIUrl":"https://doi.org/10.1109/ICPR.2014.671","url":null,"abstract":"Recognizing events in consumer videos is becoming increasingly important because of the enormous growth of consumer videos in recent years. Current researches mainly focus on learning from numerous labeled videos, which is time consuming and labor expensive due to labeling the consumer videos. To alleviate the labeling process, we utilize a large number of loosely labeled Web videos (e.g., from YouTube) for visual event recognition in consumer videos. Web videos are noisy and diverse, so brute force transfer of Web videos to consumer videos may hurt the performance. To address such a negative transfer problem, we propose a novel Multi-Group Adaptation (MGA) framework to divide the training Web videos into several semantic groups and seek the optimal weight of each group. Each weight represents how relative the corresponding group is to the consumer domain. The final classifier for event recognition is learned using the weighted combination of classifiers learned from Web videos and enforced to be smooth on the consumer domain. Comprehensive experiments on three real-world consumer video datasets demonstrate the effectiveness of MGA for event recognition in consumer videos.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114569751","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Learning Multiple Complex Features Based on Classification Results 基于分类结果的多重复杂特征学习
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.580
Yoshikuni Sato, K. Kozuka, Y. Sawada, M. Kiyono
Recently, methods for the unsupervised learning of features from large data sets have been attracting much attention. These methods have been especially successful in the area of computer vision. However, there is a problem that it is difficult to determine what kind of features will result in a high classification performance. Indeed, the difficulty of determining the learning parameters is a widely known problem in the field of feature learning. To address this problem, this paper presents a feature-learning method which uses classification results to progressively learn multiple features of varied complexity. The proposed method enables the learning of both simple robust features and complex features which represents difficult patterns. In addition, we assign regularization weights that are based on the complexity of the features. This modification emphasizes simple representation and prevents over fitting. Experimental results with medical image classification show that the proposed method is superior to the conventional method, especially when classification is difficult.
近年来,大型数据集特征的无监督学习方法备受关注。这些方法在计算机视觉领域尤其成功。然而,有一个问题是很难确定什么样的特征会产生高的分类性能。事实上,在特征学习领域,确定学习参数的困难是一个众所周知的问题。为了解决这一问题,本文提出了一种特征学习方法,利用分类结果逐步学习不同复杂度的多个特征。该方法既可以学习简单的鲁棒特征,也可以学习代表困难模式的复杂特征。此外,我们根据特征的复杂性分配正则化权重。这种修改强调了简单的表示,防止了过度拟合。医学图像分类实验结果表明,该方法在分类难度较大的情况下优于传统的分类方法。
{"title":"Learning Multiple Complex Features Based on Classification Results","authors":"Yoshikuni Sato, K. Kozuka, Y. Sawada, M. Kiyono","doi":"10.1109/ICPR.2014.580","DOIUrl":"https://doi.org/10.1109/ICPR.2014.580","url":null,"abstract":"Recently, methods for the unsupervised learning of features from large data sets have been attracting much attention. These methods have been especially successful in the area of computer vision. However, there is a problem that it is difficult to determine what kind of features will result in a high classification performance. Indeed, the difficulty of determining the learning parameters is a widely known problem in the field of feature learning. To address this problem, this paper presents a feature-learning method which uses classification results to progressively learn multiple features of varied complexity. The proposed method enables the learning of both simple robust features and complex features which represents difficult patterns. In addition, we assign regularization weights that are based on the complexity of the features. This modification emphasizes simple representation and prevents over fitting. Experimental results with medical image classification show that the proposed method is superior to the conventional method, especially when classification is difficult.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117085328","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Effective Mean-Field Inference Method for Nonnegative Boltzmann Machines 非负玻尔兹曼机的有效平均场推理方法
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.619
Muneki Yasuda
Nonnegative Boltzmann machines (NNBMs) are recurrent probabilistic neural network models that can describe multi-modal nonnegative data. NNBMs form rectified Gaussian distributions that appear in biological neural network models, positive matrix factorization, nonnegative matrix factorization, and so on. In this paper, an effective inference method for NNBMs is proposed that uses the mean-field method, referred to as the Thou less-Anderson-Palmer equation, and the diagonal consistency method, which was recently proposed.
非负玻尔兹曼机(NNBMs)是一种能够描述多模态非负数据的递归概率神经网络模型。NNBMs形成校正高斯分布,出现在生物神经网络模型、正矩阵分解、非负矩阵分解等中。本文提出了一种有效的NNBMs推理方法,即利用平均场方法(即Thou - less-Anderson-Palmer方程)和最近提出的对角一致性方法。
{"title":"Effective Mean-Field Inference Method for Nonnegative Boltzmann Machines","authors":"Muneki Yasuda","doi":"10.1109/ICPR.2014.619","DOIUrl":"https://doi.org/10.1109/ICPR.2014.619","url":null,"abstract":"Nonnegative Boltzmann machines (NNBMs) are recurrent probabilistic neural network models that can describe multi-modal nonnegative data. NNBMs form rectified Gaussian distributions that appear in biological neural network models, positive matrix factorization, nonnegative matrix factorization, and so on. In this paper, an effective inference method for NNBMs is proposed that uses the mean-field method, referred to as the Thou less-Anderson-Palmer equation, and the diagonal consistency method, which was recently proposed.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128157805","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Model Semantic Relations with Extended Attributes 用扩展属性建模语义关系
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.440
Ye Liu, Xiangwei Kong, Haiyan Fu, Xingang You, Yunbiao Guo
Attribute based image retrieval has offered a powerful way to bridge the gap between low level features and high level semantic concepts. However, existing methods rely on manually pre-labeled queries, limiting their scalability and discriminative power. Moreover, such retrieval systems restrict the users to use only the exact pre-defined query words when describing the intended search targets, and thus fail to offer good user experience. In this paper, we propose a principled approach to automatically enrich the attribute representation by leveraging additional linguistic knowledge. To this end, an external semantic pool is introduced into the learning paradigm. In addition to modelling the relations between attributes and low level features, we also model the join interdependencies of pre-labeled attributes and semantically extended attributes, which is more expressive and flexible. We further propose a novel semantic relation measure for extended attribute learning in order to take user preference into account, which we see as a step towards practical systems. Extensive experiments on several attribute benchmarks show that our approach outperforms several state-of-the-art methods and achieves promising results in improving user experience.
基于属性的图像检索为弥合低级特征和高级语义概念之间的差距提供了一种强有力的方法。然而,现有的方法依赖于手动预标记的查询,限制了它们的可伸缩性和判别能力。此外,这种检索系统限制用户在描述预期的搜索目标时只能使用精确的预定义查询词,因此无法提供良好的用户体验。在本文中,我们提出了一种原则性的方法,通过利用额外的语言知识来自动丰富属性表示。为此,在学习范式中引入了一个外部语义池。除了对属性和底层特征之间的关系进行建模外,我们还对预标记属性和语义扩展属性之间的连接相互依赖关系进行建模,这更具表现力和灵活性。为了考虑用户偏好,我们进一步提出了一种用于扩展属性学习的新的语义关系度量,我们认为这是迈向实用系统的一步。在几个属性基准上进行的大量实验表明,我们的方法优于几种最先进的方法,并在改善用户体验方面取得了有希望的结果。
{"title":"Model Semantic Relations with Extended Attributes","authors":"Ye Liu, Xiangwei Kong, Haiyan Fu, Xingang You, Yunbiao Guo","doi":"10.1109/ICPR.2014.440","DOIUrl":"https://doi.org/10.1109/ICPR.2014.440","url":null,"abstract":"Attribute based image retrieval has offered a powerful way to bridge the gap between low level features and high level semantic concepts. However, existing methods rely on manually pre-labeled queries, limiting their scalability and discriminative power. Moreover, such retrieval systems restrict the users to use only the exact pre-defined query words when describing the intended search targets, and thus fail to offer good user experience. In this paper, we propose a principled approach to automatically enrich the attribute representation by leveraging additional linguistic knowledge. To this end, an external semantic pool is introduced into the learning paradigm. In addition to modelling the relations between attributes and low level features, we also model the join interdependencies of pre-labeled attributes and semantically extended attributes, which is more expressive and flexible. We further propose a novel semantic relation measure for extended attribute learning in order to take user preference into account, which we see as a step towards practical systems. Extensive experiments on several attribute benchmarks show that our approach outperforms several state-of-the-art methods and achieves promising results in improving user experience.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133324038","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Anomaly Detection through Spatio-temporal Context Modeling in Crowded Scenes 基于时空背景建模的拥挤场景异常检测
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.383
Tong Lu, Liang Wu, Xiaolin Ma, P. Shivakumara, C. Tan
A novel statistical framework for modeling the intrinsic structure of crowded scenes and detecting abnormal activities is presented in this paper. The proposed framework essentially turns the anomaly detection process into two parts, namely, motion pattern representation and crowded context modeling. During the first stage, we averagely divide the spatiotemporal volume into atomic blocks. Considering the fact that mutual interference of several human body parts potentially happen in the same block, we propose an atomic motion pattern representation using the Gaussian Mixture Model (GMM) to distinguish the motions inside each block in a refined way. Usual motion patterns can thus be defined as a certain type of steady motion activities appearing at specific scene positions. During the second stage, we further use the Markov Random Field (MRF) model to characterize the joint label distributions over all the adjacent local motion patterns inside the same crowded scene, aiming at modeling the severely occluded situations in a crowded scene accurately. By combining the determinations from the two stages, a weighted scheme is proposed to automatically detect anomaly events from crowded scenes. The experimental results on several different outdoor and indoor crowded scenes illustrate the effectiveness of the proposed algorithm.
本文提出了一种用于拥挤场景固有结构建模和异常活动检测的统计框架。该框架实质上将异常检测过程分为运动模式表示和拥挤上下文建模两部分。在第一阶段,我们将时空体积平均划分为原子块。考虑到人体多个部位在同一块中可能发生相互干扰,我们提出了一种基于高斯混合模型(GMM)的原子运动模式表示,以精细区分每个块内的运动。因此,通常的运动模式可以定义为出现在特定场景位置的某种类型的稳定运动活动。在第二阶段,我们进一步使用马尔可夫随机场(MRF)模型来表征同一拥挤场景中所有相邻局部运动模式的联合标签分布,旨在准确地建模拥挤场景中严重遮挡的情况。结合两个阶段的判定结果,提出了一种加权方案来自动检测拥挤场景中的异常事件。在不同室外和室内拥挤场景下的实验结果验证了该算法的有效性。
{"title":"Anomaly Detection through Spatio-temporal Context Modeling in Crowded Scenes","authors":"Tong Lu, Liang Wu, Xiaolin Ma, P. Shivakumara, C. Tan","doi":"10.1109/ICPR.2014.383","DOIUrl":"https://doi.org/10.1109/ICPR.2014.383","url":null,"abstract":"A novel statistical framework for modeling the intrinsic structure of crowded scenes and detecting abnormal activities is presented in this paper. The proposed framework essentially turns the anomaly detection process into two parts, namely, motion pattern representation and crowded context modeling. During the first stage, we averagely divide the spatiotemporal volume into atomic blocks. Considering the fact that mutual interference of several human body parts potentially happen in the same block, we propose an atomic motion pattern representation using the Gaussian Mixture Model (GMM) to distinguish the motions inside each block in a refined way. Usual motion patterns can thus be defined as a certain type of steady motion activities appearing at specific scene positions. During the second stage, we further use the Markov Random Field (MRF) model to characterize the joint label distributions over all the adjacent local motion patterns inside the same crowded scene, aiming at modeling the severely occluded situations in a crowded scene accurately. By combining the determinations from the two stages, a weighted scheme is proposed to automatically detect anomaly events from crowded scenes. The experimental results on several different outdoor and indoor crowded scenes illustrate the effectiveness of the proposed algorithm.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114505829","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
An Evaluation of the Faster STORM Method for Super-resolution Microscopy 超分辨显微镜快速STORM方法的评价
Pub Date : 2014-12-08 DOI: 10.1109/ICPR.2014.759
O. Ishaq, J. Elf, Carolina Wählby
Development of new stochastic super-resolution methods together with fluorescence microscopy imaging enables visualization of biological processes at increasing spatial and temporal resolution. Quantitative evaluation of such imaging experiments call for computational analysis methods that localize the signals with high precision and recall. Furthermore, it is desirable that the methods are fast and possible to parallelize so that the ever increasing amounts of collected data can be handled in an efficient way. We herein address signal detection in super-resolution microscopy by approaches based on compressed sensing. We describe how a previously published approach can be parallelized, reducing processing time at least four times. We also evaluate the effect of a greedy optimization approach on signal recovery at high noise and molecule density. Furthermore, our evaluation reveals how previously published compressed sensing algorithms have a performance that degrades to that of a random signal detector at high molecule density. Finally, we show the approximation of the imaging system's point spread function affects recall and precision of signal detection, illustrating the importance of parameter optimization. We evaluate the methods on synthetic data with varying signal to noise ratio and increasing molecular density, and visualize performance on real super-resolution microscopy data from a time-lapse sequence of living cells.
新的随机超分辨率方法的发展与荧光显微镜成像使生物过程的可视化在不断增加的空间和时间分辨率。对这类成像实验的定量评价需要能够高精度、召回地定位信号的计算分析方法。此外,希望这些方法能够快速并可能并行化,以便能够以有效的方式处理不断增加的收集数据量。本文采用基于压缩感知的方法来解决超分辨率显微镜中的信号检测问题。我们描述了先前发布的方法如何并行化,将处理时间减少至少四倍。我们还评估了贪婪优化方法在高噪声和高分子密度下的信号恢复效果。此外,我们的评估揭示了先前发表的压缩感知算法如何在高分子密度下降低到随机信号检测器的性能。最后,我们展示了成像系统点扩展函数的近似影响信号检测的召回率和精度,说明了参数优化的重要性。我们评估了不同信噪比和增加分子密度的合成数据的方法,并从活细胞的延时序列中可视化了真实超分辨率显微镜数据的性能。
{"title":"An Evaluation of the Faster STORM Method for Super-resolution Microscopy","authors":"O. Ishaq, J. Elf, Carolina Wählby","doi":"10.1109/ICPR.2014.759","DOIUrl":"https://doi.org/10.1109/ICPR.2014.759","url":null,"abstract":"Development of new stochastic super-resolution methods together with fluorescence microscopy imaging enables visualization of biological processes at increasing spatial and temporal resolution. Quantitative evaluation of such imaging experiments call for computational analysis methods that localize the signals with high precision and recall. Furthermore, it is desirable that the methods are fast and possible to parallelize so that the ever increasing amounts of collected data can be handled in an efficient way. We herein address signal detection in super-resolution microscopy by approaches based on compressed sensing. We describe how a previously published approach can be parallelized, reducing processing time at least four times. We also evaluate the effect of a greedy optimization approach on signal recovery at high noise and molecule density. Furthermore, our evaluation reveals how previously published compressed sensing algorithms have a performance that degrades to that of a random signal detector at high molecule density. Finally, we show the approximation of the imaging system's point spread function affects recall and precision of signal detection, illustrating the importance of parameter optimization. We evaluate the methods on synthetic data with varying signal to noise ratio and increasing molecular density, and visualize performance on real super-resolution microscopy data from a time-lapse sequence of living cells.","PeriodicalId":142159,"journal":{"name":"2014 22nd International Conference on Pattern Recognition","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123476738","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2014 22nd International Conference on Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1