首页 > 最新文献

2018 Digital Image Computing: Techniques and Applications (DICTA)最新文献

英文 中文
Histogram Alternation Based Digital Image Compression using Base-2 Coding 基于基-2编码的直方图交替数字图像压缩
Pub Date : 2018-12-01 DOI: 10.1109/DICTA.2018.8615830
Md.Atiqur Rahman, S. Islam, Jungpil Shin, Md. Rashedul Islam
The intention of data compression is to promote the storage and delivery of big images with excellent compression ratio and least distortion. Moreover, the number of internet user is growing day by day speedily. Therefore, the transfer of data is being another significant concern. The storage and the use of an uncompressed picture are very costly and time-consuming. There are many techniques such as Arithmetic coding, Run-length coding, Huffman coding, Shannon-Fano coding used to compress an image. Compression of a picture using the state-of-the-art techniques has a high impact. However, The compression ratio and transfer speed do not satisfy the current demand. This article proposes a new histogram alternation based lossy image compression using Base-2 coding. It increases the probabilities of an image by doing a little bit of change to its pixels level which helps to reduce code-word. This algorithm uses less storage space and works at high-speed to encode and decode an image. Average code length, compression ratio, mean square error and pick signal to noise ratio are used to estimate this method. The proposed method demonstrates better performance than the state-of-the-art techniques.
数据压缩的目的是为了以优异的压缩比和最小的失真促进大图像的存储和传输。此外,互联网用户的数量日益迅速增长。因此,数据的传输是另一个重要问题。存储和使用未压缩的图片非常昂贵且耗时。有许多技术,如算术编码,游程编码,霍夫曼编码,香农-范诺编码用于压缩图像。使用最先进的技术压缩图像具有很高的影响。但是,压缩比和传输速度不能满足当前的需求。本文提出了一种新的基于直方图交替的有损图像压缩方法。它通过对其像素水平做一点改变来增加图像的概率,这有助于减少码字。该算法利用较少的存储空间,以高速的速度对图像进行编码和解码。使用平均码长、压缩比、均方误差和拾取信噪比对该方法进行估计。该方法的性能优于目前最先进的技术。
{"title":"Histogram Alternation Based Digital Image Compression using Base-2 Coding","authors":"Md.Atiqur Rahman, S. Islam, Jungpil Shin, Md. Rashedul Islam","doi":"10.1109/DICTA.2018.8615830","DOIUrl":"https://doi.org/10.1109/DICTA.2018.8615830","url":null,"abstract":"The intention of data compression is to promote the storage and delivery of big images with excellent compression ratio and least distortion. Moreover, the number of internet user is growing day by day speedily. Therefore, the transfer of data is being another significant concern. The storage and the use of an uncompressed picture are very costly and time-consuming. There are many techniques such as Arithmetic coding, Run-length coding, Huffman coding, Shannon-Fano coding used to compress an image. Compression of a picture using the state-of-the-art techniques has a high impact. However, The compression ratio and transfer speed do not satisfy the current demand. This article proposes a new histogram alternation based lossy image compression using Base-2 coding. It increases the probabilities of an image by doing a little bit of change to its pixels level which helps to reduce code-word. This algorithm uses less storage space and works at high-speed to encode and decode an image. Average code length, compression ratio, mean square error and pick signal to noise ratio are used to estimate this method. The proposed method demonstrates better performance than the state-of-the-art techniques.","PeriodicalId":130057,"journal":{"name":"2018 Digital Image Computing: Techniques and Applications (DICTA)","volume":"19 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116099329","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Learning Orientation-Estimation Convolutional Neural Network for Building Detection in Optical Remote Sensing Image 基于学习方向估计卷积神经网络的光学遥感图像建筑物检测
Pub Date : 2018-12-01 DOI: 10.1109/DICTA.2018.8615859
Yongliang Chen, W. Gong, Chaoyue Chen, Weihong Li
Benefiting from the great success of deep learning in computer vision, object detection with Convolutional Neural Network (CNN) based methods have drawn significant attentions. Various frameworks have been proposed which show awesome and robust performance for a large range of datasets. However, for building detection in remote sensing images, buildings always pose a diversity of orientation which makes it a challenge for the application of off-the-shelf methods to building detection in remote sensing images. In this work, we aim to integrate orientation regression into the popular axis-aligned bounding box to tackle this problem. To adapt the axis-aligned bounding boxes to arbitrarily orientated ones, we also develop an algorithm to estimate the Intersection Over Union (IOU) overlap between any two arbitrarily oriented boxes which is convenient to implement in Graphics Processing Unit (GPU) for fast computation. The proposed method utilizes CNN for both robust feature extraction and bounding box regression. We present our model in an end-to-end fashion making it easy to train. The model is formulated and trained to predict both orientation and location simultaneously obtaining tighter bounding box and hence, higher mean average precision (mAP). Experiments on remote sensing images of different scales shows a promising performance over the conventional one.
得益于深度学习在计算机视觉领域的巨大成功,基于卷积神经网络(CNN)的目标检测方法备受关注。已经提出了各种框架,它们在大范围的数据集上表现出令人敬畏和强大的性能。然而,对于遥感图像中的建筑物检测来说,建筑物总是具有多样性的方向,这给现成的方法在遥感图像中建筑物检测中的应用带来了挑战。在这项工作中,我们的目标是将方向回归集成到流行的轴对齐边界框中来解决这个问题。为了使轴线对齐的边界框适应于任意方向的边界框,我们还开发了一种算法来估计任意两个任意方向的边界框之间的交联(Intersection Over Union, IOU)重叠,该算法便于在图形处理单元(GPU)中实现,以实现快速计算。该方法利用CNN进行鲁棒特征提取和边界盒回归。我们以端到端方式呈现我们的模型,使其易于训练。模型的制定和训练可以同时预测方向和位置,从而获得更紧密的边界框,从而获得更高的平均精度(mAP)。在不同尺度遥感图像上的实验表明,该方法优于传统方法。
{"title":"Learning Orientation-Estimation Convolutional Neural Network for Building Detection in Optical Remote Sensing Image","authors":"Yongliang Chen, W. Gong, Chaoyue Chen, Weihong Li","doi":"10.1109/DICTA.2018.8615859","DOIUrl":"https://doi.org/10.1109/DICTA.2018.8615859","url":null,"abstract":"Benefiting from the great success of deep learning in computer vision, object detection with Convolutional Neural Network (CNN) based methods have drawn significant attentions. Various frameworks have been proposed which show awesome and robust performance for a large range of datasets. However, for building detection in remote sensing images, buildings always pose a diversity of orientation which makes it a challenge for the application of off-the-shelf methods to building detection in remote sensing images. In this work, we aim to integrate orientation regression into the popular axis-aligned bounding box to tackle this problem. To adapt the axis-aligned bounding boxes to arbitrarily orientated ones, we also develop an algorithm to estimate the Intersection Over Union (IOU) overlap between any two arbitrarily oriented boxes which is convenient to implement in Graphics Processing Unit (GPU) for fast computation. The proposed method utilizes CNN for both robust feature extraction and bounding box regression. We present our model in an end-to-end fashion making it easy to train. The model is formulated and trained to predict both orientation and location simultaneously obtaining tighter bounding box and hence, higher mean average precision (mAP). Experiments on remote sensing images of different scales shows a promising performance over the conventional one.","PeriodicalId":130057,"journal":{"name":"2018 Digital Image Computing: Techniques and Applications (DICTA)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122493453","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Feature-Level Fusion using Convolutional Neural Network for Multi-Language Synthetic Character Recognition in Natual Images 基于卷积神经网络的特征级融合自然图像多语言合成字符识别
Pub Date : 2018-12-01 DOI: 10.1109/DICTA.2018.8615845
Asghar Ali, M. Pickering
In this paper, a new Convolutional Neural Network (CNN) architecture is proposed for synthetic Urdu and English character recognition in natural scene images. The features are extracted using three separate sub-models of the CNN which are then fused in one feature vector. The network is purely trained on the synthetic character images of English and Urdu texts in natural images. For English text, the Chars74k-Font dataset is used and for Urdu text, the synthetic dataset is created by automatically cropping the image patches from four background image datasets and then putting characters at random positions within the image patch. The network is evaluated on a combined synthetic dataset of English and Urdu characters and the separate synthetic characters of Urdu and English datasets. The experimental results show that the network performs well on synthetic datasets.
本文提出了一种新的卷积神经网络(CNN)结构,用于自然场景图像中乌尔都语和英语字符的综合识别。使用三个独立的CNN子模型提取特征,然后将其融合到一个特征向量中。该网络纯粹是在自然图像中的英语和乌尔都语文本的合成字符图像上进行训练的。对于英语文本,使用Chars74k-Font数据集,对于乌尔都语文本,通过自动裁剪四个背景图像数据集的图像补丁,然后在图像补丁内随机放置字符来创建合成数据集。在英语和乌尔都语字符的组合合成数据集以及乌尔都语和英语的单独合成数据集上对网络进行了评估。实验结果表明,该网络在合成数据集上表现良好。
{"title":"Feature-Level Fusion using Convolutional Neural Network for Multi-Language Synthetic Character Recognition in Natual Images","authors":"Asghar Ali, M. Pickering","doi":"10.1109/DICTA.2018.8615845","DOIUrl":"https://doi.org/10.1109/DICTA.2018.8615845","url":null,"abstract":"In this paper, a new Convolutional Neural Network (CNN) architecture is proposed for synthetic Urdu and English character recognition in natural scene images. The features are extracted using three separate sub-models of the CNN which are then fused in one feature vector. The network is purely trained on the synthetic character images of English and Urdu texts in natural images. For English text, the Chars74k-Font dataset is used and for Urdu text, the synthetic dataset is created by automatically cropping the image patches from four background image datasets and then putting characters at random positions within the image patch. The network is evaluated on a combined synthetic dataset of English and Urdu characters and the separate synthetic characters of Urdu and English datasets. The experimental results show that the network performs well on synthetic datasets.","PeriodicalId":130057,"journal":{"name":"2018 Digital Image Computing: Techniques and Applications (DICTA)","volume":"61 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122950408","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
List of Meta Reviewers 元审稿人列表
Pub Date : 2018-12-01 DOI: 10.1109/dicta.2018.8615755
{"title":"List of Meta Reviewers","authors":"","doi":"10.1109/dicta.2018.8615755","DOIUrl":"https://doi.org/10.1109/dicta.2018.8615755","url":null,"abstract":"","PeriodicalId":130057,"journal":{"name":"2018 Digital Image Computing: Techniques and Applications (DICTA)","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124940074","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Disparity Guided Texture Inpainting for Light Field View Synthesis 视差引导纹理绘制光场视图合成
Pub Date : 2018-12-01 DOI: 10.1109/DICTA.2018.8615821
Yue Li, R. Mathew, D. Taubman
Light fields, as a type of visual content, richer in textural and geometric information than traditional imaging, can exhibit strong redundancies between views. Disparity compensated prediction, as one of the view synthesis frameworks, can exploit these redundancies to achieve high coding efficiency. Properly handling texture occlusion in the prediction process is important. We propose a disparity guided texture inpainting scheme to resolve texture occlusion. It turns out that reliable disparity (depth) can be available within occluded regions. A key contribution of this paper is the incorporation of disparity to guide the pixel visiting order and the weighted-average interpolation processes of the inpainting scheme. Specifically, the paper describes a disparity-dependent boundary distance metric, which is evaluated using a Dijkstra's algorithm and used to drive inpainting decisions. Our proposed method is evaluated on a realistic dataset with complex geometry, presenting promising results.
光场作为一种视觉内容,比传统成像具有更丰富的纹理和几何信息,在视图之间会表现出很强的冗余性。视差补偿预测作为一种视图综合框架,可以利用这些冗余来达到较高的编码效率。在预测过程中正确处理纹理遮挡是很重要的。为了解决纹理遮挡问题,提出了一种视差引导的纹理绘制方案。结果表明,在被遮挡的区域内可以获得可靠的视差(深度)。本文的一个重要贡献是引入视差来指导像素访问顺序,并对绘制方案进行加权平均插值处理。具体来说,本文描述了一个差异相关的边界距离度量,该度量使用Dijkstra算法进行评估,并用于驱动喷漆决策。我们提出的方法在具有复杂几何形状的真实数据集上进行了评估,显示出令人满意的结果。
{"title":"Disparity Guided Texture Inpainting for Light Field View Synthesis","authors":"Yue Li, R. Mathew, D. Taubman","doi":"10.1109/DICTA.2018.8615821","DOIUrl":"https://doi.org/10.1109/DICTA.2018.8615821","url":null,"abstract":"Light fields, as a type of visual content, richer in textural and geometric information than traditional imaging, can exhibit strong redundancies between views. Disparity compensated prediction, as one of the view synthesis frameworks, can exploit these redundancies to achieve high coding efficiency. Properly handling texture occlusion in the prediction process is important. We propose a disparity guided texture inpainting scheme to resolve texture occlusion. It turns out that reliable disparity (depth) can be available within occluded regions. A key contribution of this paper is the incorporation of disparity to guide the pixel visiting order and the weighted-average interpolation processes of the inpainting scheme. Specifically, the paper describes a disparity-dependent boundary distance metric, which is evaluated using a Dijkstra's algorithm and used to drive inpainting decisions. Our proposed method is evaluated on a realistic dataset with complex geometry, presenting promising results.","PeriodicalId":130057,"journal":{"name":"2018 Digital Image Computing: Techniques and Applications (DICTA)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126167562","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Road Trail Classification using Color Images for Autonomous Vehicle Navigation 基于彩色图像的自动驾驶汽车路径分类
Pub Date : 2018-12-01 DOI: 10.1109/DICTA.2018.8615834
K. Islam, S. Wijewickrema, Masud Pervez, S. O'Leary
Natural road trail image classification is a challenging problem due to the complexity of the natural road environment. It is useful in many real-world applications such as autonomous vehicle and robot navigation. In recent years, many researchers have explored the use of data obtained from different sensors in solving this problem. In this paper, we use image data captured from standard digital cameras, to address the road trail classification problem. To this end, we develop a database of road trail images and train an artificial neural network (ANN) classifier on features obtained using the bag-of-words (BoW) image feature extraction approach. We show experimentally that the proposed method is effective in classifying road trails.
由于自然道路环境的复杂性,自然道路图像分类是一个具有挑战性的问题。它在许多实际应用中都很有用,例如自动驾驶汽车和机器人导航。近年来,许多研究者探索了利用不同传感器获得的数据来解决这一问题。在本文中,我们使用从标准数码相机捕获的图像数据来解决道路轨迹分类问题。为此,我们开发了一个道路图像数据库,并对使用词袋(BoW)图像特征提取方法获得的特征训练人工神经网络(ANN)分类器。实验结果表明,该方法对道路轨迹分类是有效的。
{"title":"Road Trail Classification using Color Images for Autonomous Vehicle Navigation","authors":"K. Islam, S. Wijewickrema, Masud Pervez, S. O'Leary","doi":"10.1109/DICTA.2018.8615834","DOIUrl":"https://doi.org/10.1109/DICTA.2018.8615834","url":null,"abstract":"Natural road trail image classification is a challenging problem due to the complexity of the natural road environment. It is useful in many real-world applications such as autonomous vehicle and robot navigation. In recent years, many researchers have explored the use of data obtained from different sensors in solving this problem. In this paper, we use image data captured from standard digital cameras, to address the road trail classification problem. To this end, we develop a database of road trail images and train an artificial neural network (ANN) classifier on features obtained using the bag-of-words (BoW) image feature extraction approach. We show experimentally that the proposed method is effective in classifying road trails.","PeriodicalId":130057,"journal":{"name":"2018 Digital Image Computing: Techniques and Applications (DICTA)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125223062","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Improving Supervised Microaneurysm Segmentation using Autoencoder-Regularized Neural Network 应用自编码器-正则化神经网络改进监督微动脉瘤分割
Pub Date : 2018-12-01 DOI: 10.1109/DICTA.2018.8615839
Rangwan Kasantikul, Worapan Kusakunniran
This paper proposes the novel microaneurysm segmentation technique, based on the autoencoder-regularized neural network model. The proposed method is developed using two levels of the segmentation. First, the coarse-level segmentation stage locates the candidate areas using the multi-scale correlation filter and region growing. Second, the fine-level segmentation stage uses the neural network to obtain confidence values of candidate areas of being microaneurysm. The neural network based technique introduced in this paper is the modified multilayer neural network with an additional branch to take into account of the reconstruction error (in a similar fashion to the autoencoder). This modification to the neural network results in the consistent improvement in the classification performance, when compared to the conventional network without such modification. The proposed method is evaluated using the retinopathic online challenge dataset. It can deliver very promising results, when compared with the existing state-of-the-art techniques.
本文提出了一种基于自编码器-正则化神经网络模型的新型微动脉瘤分割技术。该方法采用了两个层次的分割方法。首先,粗级分割阶段利用多尺度相关滤波和区域生长来定位候选区域;其次,精细分割阶段使用神经网络获得候选微动脉瘤区域的置信度值。本文介绍的基于神经网络的技术是一种改进的多层神经网络,增加了一个分支来考虑重构误差(类似于自编码器的方式)。这种对神经网络的修改与没有进行这种修改的传统网络相比,在分类性能上得到了一致的提高。使用视网膜病变在线挑战数据集对所提出的方法进行了评估。与现有的最先进的技术相比,它可以提供非常有希望的结果。
{"title":"Improving Supervised Microaneurysm Segmentation using Autoencoder-Regularized Neural Network","authors":"Rangwan Kasantikul, Worapan Kusakunniran","doi":"10.1109/DICTA.2018.8615839","DOIUrl":"https://doi.org/10.1109/DICTA.2018.8615839","url":null,"abstract":"This paper proposes the novel microaneurysm segmentation technique, based on the autoencoder-regularized neural network model. The proposed method is developed using two levels of the segmentation. First, the coarse-level segmentation stage locates the candidate areas using the multi-scale correlation filter and region growing. Second, the fine-level segmentation stage uses the neural network to obtain confidence values of candidate areas of being microaneurysm. The neural network based technique introduced in this paper is the modified multilayer neural network with an additional branch to take into account of the reconstruction error (in a similar fashion to the autoencoder). This modification to the neural network results in the consistent improvement in the classification performance, when compared to the conventional network without such modification. The proposed method is evaluated using the retinopathic online challenge dataset. It can deliver very promising results, when compared with the existing state-of-the-art techniques.","PeriodicalId":130057,"journal":{"name":"2018 Digital Image Computing: Techniques and Applications (DICTA)","volume":"92 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121757627","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Satellite Multi-Vehicle Tracking under Inconsistent Detection Conditions by Bilevel K-Shortest Paths Optimization 基于双k最短路径优化的非一致检测条件下卫星多飞行器跟踪
Pub Date : 2018-12-01 DOI: 10.1109/DICTA.2018.8615873
Junpeng Zhang, X. Jia, Jiankun Hu, Kun Tan
Tracking multiple vehicles in optical satellite videos can be achieved by detecting vehicles in individual frames and then link detections across frames. This is challenging since the low spatial resolution of satellite videos would lead to miss or partial detections. Failure in consistently detecting an identical vehicle would terminate a trajectory for loss of identity, or reinitialize a new trajectory for re-detection. Both would result in fragmented tracklets for an identical vehicle. In this paper, we propose a two-step global data association approach for multiple target tracking, which first generates local tracklets then merges them to global trajectories. In order to handle the frequent terminals of tracklets caused by the miss or partial detections, the spatial grid flow model has been extended to temporal neighboring grids to cover the possible connectivities in a wider temporal neighboring, so that the detection association could match temporal-unlinked detections. In the tracklet association model, we customize a tracklet transition probability based on Kalman Filter to link tracklets with large temporal intervals. A near-optimal solution to the proposed two-step association problem is found by Bilevel K-shortest Paths Optimization. Experiments on a satellite video show that our approach improves both detection and tracking performance of moving vehicles.
通过在单个帧中检测车辆,然后跨帧链接检测,可以实现对光学卫星视频中多个车辆的跟踪。这是具有挑战性的,因为卫星视频的低空间分辨率将导致遗漏或部分检测。如果未能持续检测到相同的车辆,则会因身份丢失而终止轨迹,或者重新初始化新轨迹以重新检测。这两种情况都会导致同一辆车的履带碎片化。本文提出了一种两步全局数据关联的多目标跟踪方法,首先生成局部轨迹,然后将其合并到全局轨迹中。为了处理由于检测缺失或部分检测导致的轨迹点频繁终端的问题,将空间网格流模型扩展到时间相邻网格,在更宽的时间相邻网格中覆盖可能的连通性,使检测关联能够匹配时间不相连的检测。在轨道关联模型中,我们基于卡尔曼滤波自定义轨道转移概率,将时间间隔较大的轨道关联起来。通过双级k最短路径优化,找到了两步关联问题的近最优解。卫星视频实验表明,该方法提高了移动车辆的检测和跟踪性能。
{"title":"Satellite Multi-Vehicle Tracking under Inconsistent Detection Conditions by Bilevel K-Shortest Paths Optimization","authors":"Junpeng Zhang, X. Jia, Jiankun Hu, Kun Tan","doi":"10.1109/DICTA.2018.8615873","DOIUrl":"https://doi.org/10.1109/DICTA.2018.8615873","url":null,"abstract":"Tracking multiple vehicles in optical satellite videos can be achieved by detecting vehicles in individual frames and then link detections across frames. This is challenging since the low spatial resolution of satellite videos would lead to miss or partial detections. Failure in consistently detecting an identical vehicle would terminate a trajectory for loss of identity, or reinitialize a new trajectory for re-detection. Both would result in fragmented tracklets for an identical vehicle. In this paper, we propose a two-step global data association approach for multiple target tracking, which first generates local tracklets then merges them to global trajectories. In order to handle the frequent terminals of tracklets caused by the miss or partial detections, the spatial grid flow model has been extended to temporal neighboring grids to cover the possible connectivities in a wider temporal neighboring, so that the detection association could match temporal-unlinked detections. In the tracklet association model, we customize a tracklet transition probability based on Kalman Filter to link tracklets with large temporal intervals. A near-optimal solution to the proposed two-step association problem is found by Bilevel K-shortest Paths Optimization. Experiments on a satellite video show that our approach improves both detection and tracking performance of moving vehicles.","PeriodicalId":130057,"journal":{"name":"2018 Digital Image Computing: Techniques and Applications (DICTA)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114357518","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Adversarial Context Aggregation Network for Low-Light Image Enhancement 用于弱光图像增强的对抗上下文聚合网络
Pub Date : 2018-12-01 DOI: 10.1109/DICTA.2018.8615848
Y. Shin, M. Sagong, Yoon-Jae Yeo, S. Ko
Image captured in the low-light environments usually suffers from the low dynamic ranges and noise which degrade the quality of the image. Recently, convolutional neural network (CNN) has been employed for low-light image enhancement to simultaneously perform the brightness enhancement and noise removal. Although conventional CNN based techniques exhibit superior performance compared to traditional non-CNN based methods, they often produce the image with visual artifacts due to the small receptive field in their network. In order to cope with this problem, we propose an adversarial context aggregation network (ACA-net) for low-light image enhancement, which effectively aggregates the global context via full-resolution intermediate layers. In the proposed method, we first increase the brightness of a low-light image using the two different gamma correction functions and then feed the brightened images to CNN to obtain the enhanced image. To this end, we train ACA network using L1 pixel-wise reconstruction loss and adversarial loss which encourages the network to generate a natural image. Experimental results show that the proposed method achieves state-of-the-art results in terms of peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM).1
在弱光环境下拍摄的图像通常会受到低动态范围和噪声的影响,从而降低图像质量。近年来,卷积神经网络(convolutional neural network, CNN)被用于微光图像增强,在增强亮度的同时去噪。尽管传统的基于CNN的技术与传统的非基于CNN的方法相比表现出优越的性能,但由于其网络中的接受野较小,它们通常会产生带有视觉伪影的图像。为了解决这一问题,我们提出了一种对抗上下文聚合网络(ACA-net)用于弱光图像增强,该网络通过全分辨率中间层有效地聚合全局上下文。在本文提出的方法中,我们首先使用两种不同的伽玛校正函数增加低光图像的亮度,然后将增亮后的图像馈送给CNN,得到增强后的图像。为此,我们使用L1像素重建损失和对抗损失来训练ACA网络,这鼓励网络生成自然图像。实验结果表明,该方法在峰值信噪比(PSNR)和结构相似度指标(SSIM)方面均取得了较好的效果
{"title":"Adversarial Context Aggregation Network for Low-Light Image Enhancement","authors":"Y. Shin, M. Sagong, Yoon-Jae Yeo, S. Ko","doi":"10.1109/DICTA.2018.8615848","DOIUrl":"https://doi.org/10.1109/DICTA.2018.8615848","url":null,"abstract":"Image captured in the low-light environments usually suffers from the low dynamic ranges and noise which degrade the quality of the image. Recently, convolutional neural network (CNN) has been employed for low-light image enhancement to simultaneously perform the brightness enhancement and noise removal. Although conventional CNN based techniques exhibit superior performance compared to traditional non-CNN based methods, they often produce the image with visual artifacts due to the small receptive field in their network. In order to cope with this problem, we propose an adversarial context aggregation network (ACA-net) for low-light image enhancement, which effectively aggregates the global context via full-resolution intermediate layers. In the proposed method, we first increase the brightness of a low-light image using the two different gamma correction functions and then feed the brightened images to CNN to obtain the enhanced image. To this end, we train ACA network using L1 pixel-wise reconstruction loss and adversarial loss which encourages the network to generate a natural image. Experimental results show that the proposed method achieves state-of-the-art results in terms of peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM).1","PeriodicalId":130057,"journal":{"name":"2018 Digital Image Computing: Techniques and Applications (DICTA)","volume":"93 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124045596","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Hyperspectral Image Segmentation of Retinal Vasculature, Optic Disc and Macula 视网膜血管、视盘和黄斑的高光谱图像分割
Pub Date : 2018-12-01 DOI: 10.1109/DICTA.2018.8615761
A. Garifullin, Peeter Koobi, Pasi Ylitepsa, Kati Adjers, M. Hauta-Kasari, H. Uusitalo, L. Lensu
The most common approach for retinal imaging is the eye fundus photography which usually results in RGB images. Recent studies show that the additional spectral information provides useful features for automatic retinal image analysis. The current work extends recent research on the joint segmentation of retinal vasculature, optic disc and macula which often appears in different retinal image analysis tasks. Fully convolutional neural networks are utilized to solve the segmentation problem. It is shown that the network architectures can be effectively modified for the spectral data and the utilization of spectral information provides moderate improvements in retinal image segmentation.
视网膜成像最常见的方法是眼底摄影,通常会得到RGB图像。近年来的研究表明,额外的光谱信息为自动视网膜图像分析提供了有用的特征。目前的工作是对视网膜血管、视盘和黄斑的联合分割研究的延伸,这在不同的视网膜图像分析任务中经常出现。利用全卷积神经网络解决图像分割问题。实验结果表明,基于光谱数据的网络结构可以得到有效的改进,光谱信息的利用对视网膜图像分割有一定的改善。
{"title":"Hyperspectral Image Segmentation of Retinal Vasculature, Optic Disc and Macula","authors":"A. Garifullin, Peeter Koobi, Pasi Ylitepsa, Kati Adjers, M. Hauta-Kasari, H. Uusitalo, L. Lensu","doi":"10.1109/DICTA.2018.8615761","DOIUrl":"https://doi.org/10.1109/DICTA.2018.8615761","url":null,"abstract":"The most common approach for retinal imaging is the eye fundus photography which usually results in RGB images. Recent studies show that the additional spectral information provides useful features for automatic retinal image analysis. The current work extends recent research on the joint segmentation of retinal vasculature, optic disc and macula which often appears in different retinal image analysis tasks. Fully convolutional neural networks are utilized to solve the segmentation problem. It is shown that the network architectures can be effectively modified for the spectral data and the utilization of spectral information provides moderate improvements in retinal image segmentation.","PeriodicalId":130057,"journal":{"name":"2018 Digital Image Computing: Techniques and Applications (DICTA)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132048848","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
期刊
2018 Digital Image Computing: Techniques and Applications (DICTA)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1