首页 > 最新文献

2017 Seventh International Conference on Image Processing Theory, Tools and Applications (IPTA)最新文献

英文 中文
Calibration method for sparse multi-view cameras by bridging with a mobile camera 稀疏多视点相机与移动相机桥接标定方法
Hidehiko Shishido, I. Kitahara
Camera calibration that estimates the projective relationship between 3D and 2D image spaces is one of the most crucial processes for such 3D image processing as 3D reconstruction and 3D tracking. A strong calibration method, which needs to place landmarks with known 3D positions, is a common technique. However, as the target space becomes large, landmark placement becomes more complicated. Although a weak-calibration method does not need known landmarks to estimate a projective transformation matrix from the correspondence information among multi-view images, the estimation precision depends on the accuracy of the correspondence. When multiple cameras are arranged sparsely, detecting sufficient corresponding points is difficult. In this research, we propose a calibration method that bridges sparse multiple cameras with mobile camera images. The mobile camera captures video images while moving among sparse multi-view cameras. The captured video resembles dense multi-view images and includes sparse multi-view images so that weak-calibration is effective. We confirmed the appropriate spacing between the images through comparative experiments of camera calibration accuracy by changing the number of bridging images and applied our proposed method to multiple capturing experiments in a large-scale space and verified its robustness.
摄像机标定是三维图像重建、三维跟踪等三维图像处理中最关键的过程之一,摄像机标定是对三维和二维图像空间投影关系的估计。一个强大的校准方法,需要放置地标已知的三维位置,是一种常见的技术。然而,随着目标空间的增大,地标的放置也变得更加复杂。弱标定方法虽然不需要已知的地标来从多视点图像的对应信息中估计投影变换矩阵,但其估计精度取决于对应信息的准确性。当多台摄像机稀疏布置时,很难检测到足够的对应点。在本研究中,我们提出了一种稀疏多相机与移动相机图像桥接的校准方法。移动摄像机在稀疏的多视图摄像机之间移动时捕获视频图像。捕获的视频类似于密集的多视图图像,并且包含稀疏的多视图图像,因此弱校准是有效的。我们通过改变桥接图像的数量来比较相机标定精度的实验,确定了图像之间的适当间距,并将我们提出的方法应用于大尺度空间的多次捕获实验,验证了其鲁棒性。
{"title":"Calibration method for sparse multi-view cameras by bridging with a mobile camera","authors":"Hidehiko Shishido, I. Kitahara","doi":"10.1109/IPTA.2017.8310128","DOIUrl":"https://doi.org/10.1109/IPTA.2017.8310128","url":null,"abstract":"Camera calibration that estimates the projective relationship between 3D and 2D image spaces is one of the most crucial processes for such 3D image processing as 3D reconstruction and 3D tracking. A strong calibration method, which needs to place landmarks with known 3D positions, is a common technique. However, as the target space becomes large, landmark placement becomes more complicated. Although a weak-calibration method does not need known landmarks to estimate a projective transformation matrix from the correspondence information among multi-view images, the estimation precision depends on the accuracy of the correspondence. When multiple cameras are arranged sparsely, detecting sufficient corresponding points is difficult. In this research, we propose a calibration method that bridges sparse multiple cameras with mobile camera images. The mobile camera captures video images while moving among sparse multi-view cameras. The captured video resembles dense multi-view images and includes sparse multi-view images so that weak-calibration is effective. We confirmed the appropriate spacing between the images through comparative experiments of camera calibration accuracy by changing the number of bridging images and applied our proposed method to multiple capturing experiments in a large-scale space and verified its robustness.","PeriodicalId":316356,"journal":{"name":"2017 Seventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133908118","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Multimodal three-dimensional vision for wildland fires detection and analysis 多模态三维视觉野火探测与分析
M. Akhloufi, Tom Toulouse, L. Rossi, X. Maldague
This paper proposes a new multimodal stereovision framework for wildland fires detection and analysis. The proposed system uses near infrared and visible images to robustly segment the fires and extract their three-dimensional characteristics during propagation. It uses multiple multimodal stereovision systems to capture complementary views of the fire front. A new registration approach is proposed, it uses multisensory fusion based on GNSS and IMU data to extract the projection matrix that permits the representation of the 3D reconstructed fire in a common reference frame. The fire parameters are extracted in 3D space during fire propagation using the complete reconstructed fire. The obtained results show the efficiency of the proposed system for wildland fires research and firefighting decision support in operational scenarios.
本文提出了一种新的多模态立体视觉框架,用于野火探测与分析。该系统利用近红外和可见光图像对火焰进行鲁棒分割,提取火焰在传播过程中的三维特征。它使用多个多模态立体视觉系统来捕捉火线的互补视图。提出了一种新的配准方法,利用基于GNSS和IMU数据的多感官融合提取投影矩阵,使三维重建的火灾能够在一个共同的参考框架中表示。利用完整的火灾重建模型,在三维空间中提取火灾传播过程中的参数。研究结果表明,该系统能够有效地为野外火灾研究和作战场景下的消防决策提供支持。
{"title":"Multimodal three-dimensional vision for wildland fires detection and analysis","authors":"M. Akhloufi, Tom Toulouse, L. Rossi, X. Maldague","doi":"10.1109/IPTA.2017.8310085","DOIUrl":"https://doi.org/10.1109/IPTA.2017.8310085","url":null,"abstract":"This paper proposes a new multimodal stereovision framework for wildland fires detection and analysis. The proposed system uses near infrared and visible images to robustly segment the fires and extract their three-dimensional characteristics during propagation. It uses multiple multimodal stereovision systems to capture complementary views of the fire front. A new registration approach is proposed, it uses multisensory fusion based on GNSS and IMU data to extract the projection matrix that permits the representation of the 3D reconstructed fire in a common reference frame. The fire parameters are extracted in 3D space during fire propagation using the complete reconstructed fire. The obtained results show the efficiency of the proposed system for wildland fires research and firefighting decision support in operational scenarios.","PeriodicalId":316356,"journal":{"name":"2017 Seventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134147906","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Subaperture image segmentation for lossless compression 子孔径图像分割无损压缩
I. Schiopu, M. Gabbouj, Alexandros Iosifidis, B. Zeng, Shuaicheng Liu
The paper proposes an image segmentation method for lossless compression of plenoptic images. Each light-field image captured by the plenoptic camera is processed to obtain a stack of subaperture images. Each subaperture image is encoded by using a gradient-base detector which classifies the image edges and designs refined contexts for an improved prediction and segmentation. The paper's main contribution is a new segmentation method which generates a preliminary segmentation, either by scaling the intensity differences or by using a quantum cut based algorithm, and merges it with an edge ranking-based segmentation. The results show around 2% improved performance compared to the state-of-the-art for a dataset of 118 plenoptic images.
提出了一种全光学图像无损压缩的图像分割方法。对全光相机捕获的每个光场图像进行处理以获得子孔径图像的堆栈。每个子孔径图像通过使用梯度基检测器进行编码,该检测器对图像边缘进行分类,并为改进的预测和分割设计精细的上下文。本文的主要贡献是一种新的分割方法,该方法通过缩放强度差或使用基于量子切割的算法生成初步分割,并将其与基于边缘排序的分割合并。结果显示,与最先进的118张全光学图像数据集相比,性能提高了约2%。
{"title":"Subaperture image segmentation for lossless compression","authors":"I. Schiopu, M. Gabbouj, Alexandros Iosifidis, B. Zeng, Shuaicheng Liu","doi":"10.1109/IPTA.2017.8310083","DOIUrl":"https://doi.org/10.1109/IPTA.2017.8310083","url":null,"abstract":"The paper proposes an image segmentation method for lossless compression of plenoptic images. Each light-field image captured by the plenoptic camera is processed to obtain a stack of subaperture images. Each subaperture image is encoded by using a gradient-base detector which classifies the image edges and designs refined contexts for an improved prediction and segmentation. The paper's main contribution is a new segmentation method which generates a preliminary segmentation, either by scaling the intensity differences or by using a quantum cut based algorithm, and merges it with an edge ranking-based segmentation. The results show around 2% improved performance compared to the state-of-the-art for a dataset of 118 plenoptic images.","PeriodicalId":316356,"journal":{"name":"2017 Seventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116819904","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Vehicle re-identification by fusing multiple deep neural networks 融合多个深度神经网络的车辆再识别
Chao Cui, N. Sang, Changxin Gao, Lei Zou
Vehicle re-identification has become a fundamental task because of the growing explosion in the use of surveillance cameras in public security. The most widely used solution is based on license plate verification. But when facing the vehicle without a license, deck cars and other license plate information error or missing situation, vehicle searching is still a challenging problem. This paper proposed a vehicle re-identification method based on deep learning which exploit a two-branch Multi-DNN Fusion Siamese Neural Network (MFSNN) to fuses the classification outputs of color, model and pasted marks on the windshield and map them into a Euclidean space where distance can be directly used to measure the similarity of arbitrary two vehicles. In order to achieve this goal, we present a method of vehicle color identification based on Alex net, a method of vehicle model identification based on VGG net, a method of pasted marks detection and identification based on Faster R-CNN. We evaluate our MFSNN method on VehicleID dataset and in the experiment. Experiment results show that our method can achieve promising results.
随着监控摄像机在公共安全领域的应用爆炸式增长,车辆再识别已成为一项基本任务。目前应用最广泛的解决方案是基于车牌验证。但是当面对无证车辆、甲板车等车牌信息错误或丢失的情况时,车辆搜索仍然是一个具有挑战性的问题。本文提出了一种基于深度学习的车辆再识别方法,该方法利用双分支Multi-DNN融合暹罗神经网络(MFSNN)将挡风玻璃上颜色、模型和粘贴标记的分类输出融合到欧几里得空间中,在欧几里得空间中,距离可以直接用来度量任意两辆车的相似性。为了实现这一目标,我们提出了一种基于Alex网络的车辆颜色识别方法,一种基于VGG网络的车辆模型识别方法,一种基于Faster R-CNN的粘贴标记检测与识别方法。我们在车辆id数据集和实验中评估了我们的MFSNN方法。实验结果表明,该方法能取得较好的效果。
{"title":"Vehicle re-identification by fusing multiple deep neural networks","authors":"Chao Cui, N. Sang, Changxin Gao, Lei Zou","doi":"10.1109/IPTA.2017.8310090","DOIUrl":"https://doi.org/10.1109/IPTA.2017.8310090","url":null,"abstract":"Vehicle re-identification has become a fundamental task because of the growing explosion in the use of surveillance cameras in public security. The most widely used solution is based on license plate verification. But when facing the vehicle without a license, deck cars and other license plate information error or missing situation, vehicle searching is still a challenging problem. This paper proposed a vehicle re-identification method based on deep learning which exploit a two-branch Multi-DNN Fusion Siamese Neural Network (MFSNN) to fuses the classification outputs of color, model and pasted marks on the windshield and map them into a Euclidean space where distance can be directly used to measure the similarity of arbitrary two vehicles. In order to achieve this goal, we present a method of vehicle color identification based on Alex net, a method of vehicle model identification based on VGG net, a method of pasted marks detection and identification based on Faster R-CNN. We evaluate our MFSNN method on VehicleID dataset and in the experiment. Experiment results show that our method can achieve promising results.","PeriodicalId":316356,"journal":{"name":"2017 Seventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"399 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132217970","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Vehicle boundary improvement and passing vehicle detection in driver assistance by flow distribution 车流分配辅助驾驶中车辆边界改善与过路车辆检测
A. Das, K. Ruppin, P. Dave, Sharfudheen Pv
Research in advanced driver assistance system (ADAS) is an important step towards achieving the goal of autonomous intelligent vehicle. Vehicle detection and its distance estimation is an important solution of ADAS for forward collision warning applications. Partial occlusions of passing vehicles makes their detections tedious yet the accuracy of vehicle detection in all its forms in the scene and their corresponding distance estimation is a vital factor to deploy the solution. A small deviation in detection and distance accuracy could end up in a greater mishap in ADAS and AV (Autonomous Vehicle). The proposed framework addresses the aforementioned problems of detection of passing vehicles and perfecting distance measurement by accurate lower bound estimation through Inter and Intra-Frame Flow Correspondence (I2F2C). The proposed generic framework of 12F2C could be employed as a plug-in for the existing machine learning (ML) [1]/ deep learning (DL) [2] based algorithms for improving accuracy of distance estimation of vehicles and also improve accuracy and performance of passing vehicle detection with a detailed mathematical model of motion confidence.
先进驾驶辅助系统(ADAS)的研究是实现自动驾驶智能汽车目标的重要一步。车辆检测及其距离估计是ADAS在前方碰撞预警应用中的重要解决方案。对过往车辆的部分遮挡使得车辆的检测十分繁琐,而场景中各种形式的车辆检测及其距离估计的准确性是部署该解决方案的关键因素。在ADAS和AV(自动驾驶汽车)中,检测和距离精度的微小偏差可能导致更大的事故。该框架通过帧间和帧内流量对应(I2F2C)精确估计下界,解决了上述检测过往车辆和完善距离测量的问题。提出的12F2C通用框架可以作为现有基于机器学习(ML)[1]/深度学习(DL)[2]算法的插件,提高车辆距离估计的准确性,并通过详细的运动置信度数学模型提高通过车辆检测的准确性和性能。
{"title":"Vehicle boundary improvement and passing vehicle detection in driver assistance by flow distribution","authors":"A. Das, K. Ruppin, P. Dave, Sharfudheen Pv","doi":"10.1109/IPTA.2017.8310126","DOIUrl":"https://doi.org/10.1109/IPTA.2017.8310126","url":null,"abstract":"Research in advanced driver assistance system (ADAS) is an important step towards achieving the goal of autonomous intelligent vehicle. Vehicle detection and its distance estimation is an important solution of ADAS for forward collision warning applications. Partial occlusions of passing vehicles makes their detections tedious yet the accuracy of vehicle detection in all its forms in the scene and their corresponding distance estimation is a vital factor to deploy the solution. A small deviation in detection and distance accuracy could end up in a greater mishap in ADAS and AV (Autonomous Vehicle). The proposed framework addresses the aforementioned problems of detection of passing vehicles and perfecting distance measurement by accurate lower bound estimation through Inter and Intra-Frame Flow Correspondence (I2F2C). The proposed generic framework of 12F2C could be employed as a plug-in for the existing machine learning (ML) [1]/ deep learning (DL) [2] based algorithms for improving accuracy of distance estimation of vehicles and also improve accuracy and performance of passing vehicle detection with a detailed mathematical model of motion confidence.","PeriodicalId":316356,"journal":{"name":"2017 Seventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"108 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124301891","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Weeds detection in UAV imagery using SLIC and the hough transform 基于SLIC和hough变换的无人机图像杂草检测
M. D. Bah, A. Hafiane, R. Canals
Traditional weeds controlling tended to be spraying herbicides in all the fields. Such method not only requires huge quantities of herbicides but impact environment and humans health. In this paper, we propose a new method of crop/weeds discrimination using imagery provided by an unmanned aerial vehicle (UAV). This method is based on the vegetation skeleton, the Hough transform and the spatial relationship of superpixels created by the simple linear iterative clustering (SLIC). The combination of the spatial relationship of superpixels and their positions in the detected crop lines allows to detect intraline weeds. Our method shows its robustness in presence of weed patches close to crop lines as well as for the detection of crop lines as for weed detection.
传统的杂草防治往往是在所有的田间喷洒除草剂。这种方法不仅需要大量的除草剂,而且对环境和人类健康造成影响。本文提出了一种利用无人机(UAV)提供的图像进行作物/杂草识别的新方法。该方法基于植被骨架、Hough变换和简单线性迭代聚类(SLIC)生成的超像素空间关系。结合超像素的空间关系及其在检测作物线中的位置,可以检测到线内杂草。我们的方法在靠近作物株系的杂草斑块存在以及作物株系检测和杂草检测方面显示出其鲁棒性。
{"title":"Weeds detection in UAV imagery using SLIC and the hough transform","authors":"M. D. Bah, A. Hafiane, R. Canals","doi":"10.1109/IPTA.2017.8310102","DOIUrl":"https://doi.org/10.1109/IPTA.2017.8310102","url":null,"abstract":"Traditional weeds controlling tended to be spraying herbicides in all the fields. Such method not only requires huge quantities of herbicides but impact environment and humans health. In this paper, we propose a new method of crop/weeds discrimination using imagery provided by an unmanned aerial vehicle (UAV). This method is based on the vegetation skeleton, the Hough transform and the spatial relationship of superpixels created by the simple linear iterative clustering (SLIC). The combination of the spatial relationship of superpixels and their positions in the detected crop lines allows to detect intraline weeds. Our method shows its robustness in presence of weed patches close to crop lines as well as for the detection of crop lines as for weed detection.","PeriodicalId":316356,"journal":{"name":"2017 Seventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"102 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115627099","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 43
Genre linked automated assessment and feedback of photographs based on visual aesthetics 基于视觉美学的类型关联的照片自动评估和反馈
Pavan Sudheendra, D. Jayagopi
This paper addresses the problem of automatically assessing the aesthetic quality of a photograph and providing actionable feedback to the photographer. Towards this task we have designed novel genre-specific attributes (for e.g. Noise level in a night mode photograph or Depth perception for the landscape mode). Using a collection of these mode relevant attributes we improved the assessment accuracy for three modes and reached state-of-the-art on the other one mode we investigated. These intuitive attributes are also visualized as a visual signature of the photograph. This representation can act as an actionable feedback to an aspiring amateur photographer.
本文解决了自动评估照片的审美质量并向摄影师提供可操作的反馈的问题。为了完成这项任务,我们设计了新颖的特定类型属性(例如,夜间模式照片中的噪音水平或景观模式的深度感知)。使用这些模式相关属性的集合,我们提高了三种模式的评估准确性,并在我们研究的另一种模式上达到了最先进的水平。这些直观的属性也被可视化为照片的视觉签名。这种表现可以作为一个可操作的反馈给一个有抱负的业余摄影师。
{"title":"Genre linked automated assessment and feedback of photographs based on visual aesthetics","authors":"Pavan Sudheendra, D. Jayagopi","doi":"10.1109/IPTA.2017.8310080","DOIUrl":"https://doi.org/10.1109/IPTA.2017.8310080","url":null,"abstract":"This paper addresses the problem of automatically assessing the aesthetic quality of a photograph and providing actionable feedback to the photographer. Towards this task we have designed novel genre-specific attributes (for e.g. Noise level in a night mode photograph or Depth perception for the landscape mode). Using a collection of these mode relevant attributes we improved the assessment accuracy for three modes and reached state-of-the-art on the other one mode we investigated. These intuitive attributes are also visualized as a visual signature of the photograph. This representation can act as an actionable feedback to an aspiring amateur photographer.","PeriodicalId":316356,"journal":{"name":"2017 Seventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128475323","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A novel synthetic dataset for research in overlapped fingerprint separation 一种新的用于重叠指纹分离研究的合成数据集
B. Stojanovic, Oge Marques, A. Neskovic
This paper presents a new image dataset for evaluating approaches for overlapped fingerprint separation. The VLATACOM dataset consists of 120,000 synthetically overlapped test images (and the associated masks), with and without noise, processed with three different rotation angles, and in two variations of overall brightness. Each image in the dataset also contains information about the number of the singular points within its overlapped region, which is a distinctly unique feature of the proposed dataset. The paper also reports early experimental results which demonstrate the suitability of the VLATACOM dataset for overlapped fingerprint separation research. The dataset, along with testing results, is freely and publicly available.
本文提出了一种新的图像数据集,用于评估重叠指纹分离的方法。VLATACOM数据集由12万张合成重叠的测试图像(以及相关的掩模)组成,有噪声和没有噪声,用三种不同的旋转角度和两种不同的整体亮度进行处理。数据集中的每张图像还包含有关其重叠区域内奇点数量的信息,这是所提出数据集的一个明显独特的特征。本文还报告了早期实验结果,证明了VLATACOM数据集对重叠指纹分离研究的适用性。该数据集以及测试结果都是免费公开的。
{"title":"A novel synthetic dataset for research in overlapped fingerprint separation","authors":"B. Stojanovic, Oge Marques, A. Neskovic","doi":"10.1109/IPTA.2017.8310137","DOIUrl":"https://doi.org/10.1109/IPTA.2017.8310137","url":null,"abstract":"This paper presents a new image dataset for evaluating approaches for overlapped fingerprint separation. The VLATACOM dataset consists of 120,000 synthetically overlapped test images (and the associated masks), with and without noise, processed with three different rotation angles, and in two variations of overall brightness. Each image in the dataset also contains information about the number of the singular points within its overlapped region, which is a distinctly unique feature of the proposed dataset. The paper also reports early experimental results which demonstrate the suitability of the VLATACOM dataset for overlapped fingerprint separation research. The dataset, along with testing results, is freely and publicly available.","PeriodicalId":316356,"journal":{"name":"2017 Seventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116901482","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A comparison of CNN-based face and head detectors for real-time video surveillance applications 基于cnn的人脸和头部检测器在实时视频监控应用中的比较
Le Thanh Nguyen-Meidine, Eric Granger, M. Kiran, Louis-Antoine Blais-Morin
Detecting faces and heads appearing in video feeds are challenging tasks in real-world video surveillance applications due to variations in appearance, occlusions and complex backgrounds. Recently, several CNN architectures have been proposed to increase the accuracy of detectors, although their computational complexity can be an issue, especially for realtime applications, where faces and heads must be detected live using high-resolution cameras. This paper compares the accuracy and complexity of state-of-the-art CNN architectures that are suitable for face and head detection. Single pass and region-based architectures are reviewed and compared empirically to baseline techniques according to accuracy and to time and memory complexity on images from several challenging datasets. The viability of these architectures is analyzed with real-time video surveillance applications in mind. Results suggest that, although CNN architectures can achieve a very high level of accuracy compared to traditional detectors, their computational cost can represent a limitation for many practical real-time applications.
在现实世界的视频监控应用中,由于外观、遮挡和复杂背景的变化,检测视频馈送中出现的人脸和头部是一项具有挑战性的任务。最近,已经提出了几种CNN架构来提高检测器的准确性,尽管它们的计算复杂性可能是一个问题,特别是在实时应用中,必须使用高分辨率相机实时检测面部和头部。本文比较了适用于人脸和头部检测的最先进的CNN架构的准确性和复杂性。回顾了单通道和基于区域的架构,并根据准确性和时间和内存复杂性对来自几个具有挑战性的数据集的图像进行了经验比较基线技术。并结合实时视频监控应用分析了这些架构的可行性。结果表明,尽管与传统检测器相比,CNN架构可以实现非常高的精度,但它们的计算成本可能是许多实际实时应用的限制。
{"title":"A comparison of CNN-based face and head detectors for real-time video surveillance applications","authors":"Le Thanh Nguyen-Meidine, Eric Granger, M. Kiran, Louis-Antoine Blais-Morin","doi":"10.1109/IPTA.2017.8310113","DOIUrl":"https://doi.org/10.1109/IPTA.2017.8310113","url":null,"abstract":"Detecting faces and heads appearing in video feeds are challenging tasks in real-world video surveillance applications due to variations in appearance, occlusions and complex backgrounds. Recently, several CNN architectures have been proposed to increase the accuracy of detectors, although their computational complexity can be an issue, especially for realtime applications, where faces and heads must be detected live using high-resolution cameras. This paper compares the accuracy and complexity of state-of-the-art CNN architectures that are suitable for face and head detection. Single pass and region-based architectures are reviewed and compared empirically to baseline techniques according to accuracy and to time and memory complexity on images from several challenging datasets. The viability of these architectures is analyzed with real-time video surveillance applications in mind. Results suggest that, although CNN architectures can achieve a very high level of accuracy compared to traditional detectors, their computational cost can represent a limitation for many practical real-time applications.","PeriodicalId":316356,"journal":{"name":"2017 Seventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114240076","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 37
Multi-view visual saliency-based MRI classification for alzheimer's disease diagnosis 基于多视图视觉显著性的MRI分类在阿尔茨海默病诊断中的应用
O. B. Ahmed, F. Lecellier, M. Paccalin, C. Fernandez-Maloigne
Visual inspection is the first step performed by clinicians during evaluation of medical images in image-based diagnosis. This behavior can be automated using computational saliency models. In this paper, we investigate the potential role of visual saliency for computer-aided diagnosis of Alzheimer's disease (AD). We propose a multi-view saliency-based framework to detect abnormalities from structural Magnitude Resonance Imaging (MRI) and classify subjects in a Multiple Kernel Learning (MKL) framework. The obtained saliency maps are able to detect relevant brain areas for early AD diagnosis. The effectiveness of the proposed approach was evaluated on structural MRI of 509 subjects from the ADNI dataset. We achieved accuracy of 88.98% (specificity of 94.4% and a sensitivity of 83.46%) and 81.31% (specificity of 84.22% and a sensitivity of 74.21%) classification and for respectively AD versus Normal Control(NC) and NC versus Mild Cognitive Impairment (MCI). For the most challenging classification task (AD versus MCI), we reached an accuracy of 79.8%, a specificity of 79.93% and a sensitivity of 64.02%.
视觉检查是临床医生在基于图像的诊断中对医学图像进行评估的第一步。这种行为可以使用计算显著性模型自动实现。在本文中,我们研究了视觉显著性在阿尔茨海默病(AD)计算机辅助诊断中的潜在作用。我们提出了一个基于多视图显著性的框架来检测结构核磁共振成像(MRI)的异常,并在多核学习(MKL)框架中对受试者进行分类。获得的显著性图能够检测出相关的大脑区域,用于早期AD诊断。在来自ADNI数据集的509名受试者的结构MRI上评估了该方法的有效性。我们分别对AD与正常对照(NC)和NC与轻度认知障碍(MCI)进行分类,准确率为88.98%(特异性为94.4%,敏感性为83.46%)和81.31%(特异性为84.22%,敏感性为74.21%)。对于最具挑战性的分类任务(AD与MCI),我们达到了79.8%的准确率,79.93%的特异性和64.02%的灵敏度。
{"title":"Multi-view visual saliency-based MRI classification for alzheimer's disease diagnosis","authors":"O. B. Ahmed, F. Lecellier, M. Paccalin, C. Fernandez-Maloigne","doi":"10.1109/IPTA.2017.8310118","DOIUrl":"https://doi.org/10.1109/IPTA.2017.8310118","url":null,"abstract":"Visual inspection is the first step performed by clinicians during evaluation of medical images in image-based diagnosis. This behavior can be automated using computational saliency models. In this paper, we investigate the potential role of visual saliency for computer-aided diagnosis of Alzheimer's disease (AD). We propose a multi-view saliency-based framework to detect abnormalities from structural Magnitude Resonance Imaging (MRI) and classify subjects in a Multiple Kernel Learning (MKL) framework. The obtained saliency maps are able to detect relevant brain areas for early AD diagnosis. The effectiveness of the proposed approach was evaluated on structural MRI of 509 subjects from the ADNI dataset. We achieved accuracy of 88.98% (specificity of 94.4% and a sensitivity of 83.46%) and 81.31% (specificity of 84.22% and a sensitivity of 74.21%) classification and for respectively AD versus Normal Control(NC) and NC versus Mild Cognitive Impairment (MCI). For the most challenging classification task (AD versus MCI), we reached an accuracy of 79.8%, a specificity of 79.93% and a sensitivity of 64.02%.","PeriodicalId":316356,"journal":{"name":"2017 Seventh International Conference on Image Processing Theory, Tools and Applications (IPTA)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123960979","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
期刊
2017 Seventh International Conference on Image Processing Theory, Tools and Applications (IPTA)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1