首页 > 最新文献

2019 Digital Image Computing: Techniques and Applications (DICTA)最新文献

英文 中文
STCEC: A Remote Sensing Dataset for Identifying Spatial-Temporal Change in Homogeneous and Heterogeneous Environments 同质与异质环境时空变化的遥感数据集
Pub Date : 2019-12-01 DOI: 10.1109/DICTA47822.2019.8946005
Thaer F. Ali, A. Woodley
Standard experimental datasets permit comprehensive analysis between approaches. These datasets are ubiquitous in many data science domains but uncommon in remote sensing. This paper presents the Spatial-Temporal Change in Environmental Context (STCEC) dataset, an experimental remote sensing dataset that contains changes (and non-changes) in homogeneous and heterogeneous environments, thereby, enabling researchers to test their approaches in different contexts. STCEC was tested with five pixel interpolation approaches and showed a significant difference between changes in homogeneous and heterogeneous environments. It is hoped that the dataset will be used by other researchers in future work.
标准实验数据集允许在不同方法之间进行综合分析。这些数据集在许多数据科学领域普遍存在,但在遥感领域并不常见。本文介绍了环境背景下的时空变化(STCEC)数据集,这是一个包含同质和异质环境变化(和非变化)的实验遥感数据集,从而使研究人员能够在不同的背景下测试他们的方法。采用5种像素插值方法对STCEC进行了测试,结果表明均匀和异质环境下STCEC的变化存在显著差异。希望该数据集将被其他研究人员在未来的工作中使用。
{"title":"STCEC: A Remote Sensing Dataset for Identifying Spatial-Temporal Change in Homogeneous and Heterogeneous Environments","authors":"Thaer F. Ali, A. Woodley","doi":"10.1109/DICTA47822.2019.8946005","DOIUrl":"https://doi.org/10.1109/DICTA47822.2019.8946005","url":null,"abstract":"Standard experimental datasets permit comprehensive analysis between approaches. These datasets are ubiquitous in many data science domains but uncommon in remote sensing. This paper presents the Spatial-Temporal Change in Environmental Context (STCEC) dataset, an experimental remote sensing dataset that contains changes (and non-changes) in homogeneous and heterogeneous environments, thereby, enabling researchers to test their approaches in different contexts. STCEC was tested with five pixel interpolation approaches and showed a significant difference between changes in homogeneous and heterogeneous environments. It is hoped that the dataset will be used by other researchers in future work.","PeriodicalId":6696,"journal":{"name":"2019 Digital Image Computing: Techniques and Applications (DICTA)","volume":"36 1","pages":"1-8"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78862454","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
SRM Superpixel Merging Framework for Precise Segmentation of Cervical Nucleus 基于SRM超像素融合框架的颈核精确分割
Pub Date : 2019-12-01 DOI: 10.1109/DICTA47822.2019.8945887
Ratna Saha, M. Bajger, Gobert N. Lee
Cervical nuclei contain important diagnostic characteristics useful for identifying abnormality in cervical cells. Therefore, an accurate segmentation of nuclei is the primary step in computer-aided diagnosis. However, cell overlapping, uneven staining, poor contrast, and presence of debris elements make this task challenging. A novel method is presented in this paper to detect and segment nuclei from overlapping cervical smear images. The proposed framework segments nuclei by merging superpixels generated by statistical region merging (SRM) algorithm using pairwise regional contrasts and gradient boundaries. To overcome the limitation of finding the optimal parameter value, which controls the coarseness of the segmentation, a new approach for SRM superpixel generation was introduced. Quantitative and qualitative assessment of the proposed framework is carried out using Overlapping Cervical Cytology Image Segmentation Challenge — ISBI 2014 dataset of 945 cervical images. In comparison with the state-of-the-art methods, the proposed methodology achieved superior segmentation performance in terms of Dice similarity coefficient 0.956 and pixel-based recall 0.962. Other evaluation measures such as pixel-based precision 0.930, object-based precision 0.987, and recall 0.944, also compare favorably with some recently published studies. The experimental results demonstrate that the proposed framework can precisely segment nuclei from overlapping cervical cell images, while keeping high level of precision and recall. Therefore, the developed framework may assist cytologists in computerized cervical cell analysis and help with early diagnosis of cervical cancer.
宫颈核含有重要的诊断特征,有助于鉴别宫颈细胞的异常。因此,核的准确分割是计算机辅助诊断的首要步骤。然而,细胞重叠、染色不均匀、对比度差和碎片元素的存在使这项任务具有挑战性。本文提出了一种从重叠子宫颈涂片图像中检测和分割细胞核的新方法。该框架利用两两区域对比和梯度边界对统计区域合并(SRM)算法产生的超像素进行合并,从而分割出核。为了克服寻找最优参数值控制分割粗度的局限性,提出了一种新的SRM超像素生成方法。使用945张宫颈图像的重叠宫颈细胞学图像分割挑战- ISBI 2014数据集对所提出的框架进行定量和定性评估。与现有的分割方法相比,该方法在Dice相似系数0.956和基于像素的召回率0.962方面取得了更好的分割性能。其他评价指标,如基于像素的精度0.930、基于对象的精度0.987和召回率0.944,也与最近发表的一些研究结果相比较。实验结果表明,该框架能够准确地从重叠的宫颈细胞图像中分割出细胞核,同时保持较高的准确率和召回率。因此,开发的框架可以帮助细胞学家在计算机化宫颈细胞分析和帮助宫颈癌的早期诊断。
{"title":"SRM Superpixel Merging Framework for Precise Segmentation of Cervical Nucleus","authors":"Ratna Saha, M. Bajger, Gobert N. Lee","doi":"10.1109/DICTA47822.2019.8945887","DOIUrl":"https://doi.org/10.1109/DICTA47822.2019.8945887","url":null,"abstract":"Cervical nuclei contain important diagnostic characteristics useful for identifying abnormality in cervical cells. Therefore, an accurate segmentation of nuclei is the primary step in computer-aided diagnosis. However, cell overlapping, uneven staining, poor contrast, and presence of debris elements make this task challenging. A novel method is presented in this paper to detect and segment nuclei from overlapping cervical smear images. The proposed framework segments nuclei by merging superpixels generated by statistical region merging (SRM) algorithm using pairwise regional contrasts and gradient boundaries. To overcome the limitation of finding the optimal parameter value, which controls the coarseness of the segmentation, a new approach for SRM superpixel generation was introduced. Quantitative and qualitative assessment of the proposed framework is carried out using Overlapping Cervical Cytology Image Segmentation Challenge — ISBI 2014 dataset of 945 cervical images. In comparison with the state-of-the-art methods, the proposed methodology achieved superior segmentation performance in terms of Dice similarity coefficient 0.956 and pixel-based recall 0.962. Other evaluation measures such as pixel-based precision 0.930, object-based precision 0.987, and recall 0.944, also compare favorably with some recently published studies. The experimental results demonstrate that the proposed framework can precisely segment nuclei from overlapping cervical cell images, while keeping high level of precision and recall. Therefore, the developed framework may assist cytologists in computerized cervical cell analysis and help with early diagnosis of cervical cancer.","PeriodicalId":6696,"journal":{"name":"2019 Digital Image Computing: Techniques and Applications (DICTA)","volume":"49 1","pages":"1-8"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85919339","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Real-Time Human Gaze Estimation 实时人眼注视估计
Pub Date : 2019-12-01 DOI: 10.1109/DICTA47822.2019.8945919
T. Rowntree, C. Pontecorvo, I. Reid
This paper describes a system for estimating the course gaze or 1D head pose of multiple people in a video stream from a moving camera in an indoor scene. The system runs at 30 Hz and can detect human heads with a F-Score of 87.2% and predict their gaze with an average error 20.9° including when they are facing directly away from the camera. The system uses two Convolutional Neural Networks (CNNs) for head detection and gaze estimation respectively and uses common tracking and filtering techniques for smoothing predictions over time. This paper is application-focused and so describes the individual components of the system as well as the techniques used for collecting data and training the CNNs.
本文描述了一种用于估计室内场景中移动摄像机视频流中多人的航向凝视或1D头部姿势的系统。该系统以30赫兹的频率运行,可以检测到人类头部的f值为87.2%,并预测他们的凝视,平均误差为20.9°,包括他们直接远离相机的时候。该系统使用两个卷积神经网络(cnn)分别进行头部检测和凝视估计,并使用常见的跟踪和过滤技术随着时间的推移平滑预测。本文以应用为中心,因此描述了系统的各个组成部分以及用于收集数据和训练cnn的技术。
{"title":"Real-Time Human Gaze Estimation","authors":"T. Rowntree, C. Pontecorvo, I. Reid","doi":"10.1109/DICTA47822.2019.8945919","DOIUrl":"https://doi.org/10.1109/DICTA47822.2019.8945919","url":null,"abstract":"This paper describes a system for estimating the course gaze or 1D head pose of multiple people in a video stream from a moving camera in an indoor scene. The system runs at 30 Hz and can detect human heads with a F-Score of 87.2% and predict their gaze with an average error 20.9° including when they are facing directly away from the camera. The system uses two Convolutional Neural Networks (CNNs) for head detection and gaze estimation respectively and uses common tracking and filtering techniques for smoothing predictions over time. This paper is application-focused and so describes the individual components of the system as well as the techniques used for collecting data and training the CNNs.","PeriodicalId":6696,"journal":{"name":"2019 Digital Image Computing: Techniques and Applications (DICTA)","volume":"22 1","pages":"1-7"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85938779","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Bi-SAN-CAP: Bi-Directional Self-Attention for Image Captioning Bi-SAN-CAP:图像标题的双向自注意
Pub Date : 2019-12-01 DOI: 10.1109/DICTA47822.2019.8946003
Md. Zakir Hossain, F. Sohel, M. F. Shiratuddin, Hamid Laga, Bennamoun
In a typical image captioning pipeline, a Convolutional Neural Network (CNN) is used as the image encoder and Long Short-Term Memory (LSTM) as the language decoder. LSTM with attention mechanism has shown remarkable performance on sequential data including image captioning. LSTM can retain long-range dependency of sequential data. However, it is hard to parallelize the computations of LSTM because of its inherent sequential characteristics. In order to address this issue, recent works have shown benefits in using self-attention, which is highly parallelizable without requiring any temporal dependencies. However, existing techniques apply attention only in one direction to compute the context of the words. We propose an attention mechanism called Bi-directional Self-Attention (Bi-SAN) for image captioning. It computes attention both in forward and backward directions. It achieves high performance comparable to state-of-the-art methods.
在典型的图像字幕管道中,使用卷积神经网络(CNN)作为图像编码器,使用长短期记忆(LSTM)作为语言解码器。具有注意机制的LSTM在包括图像字幕在内的序列数据上表现出了显著的性能。LSTM可以保留序列数据的长期依赖关系。然而,由于LSTM固有的序列特性,其计算难以并行化。为了解决这个问题,最近的研究显示了使用自我关注的好处,它是高度并行化的,不需要任何时间依赖性。然而,现有的技术只在一个方向上应用注意力来计算单词的上下文。我们提出了一种称为双向自注意(Bi-SAN)的图像字幕注意机制。它计算向前和向后方向的注意力。它实现了与最先进的方法相媲美的高性能。
{"title":"Bi-SAN-CAP: Bi-Directional Self-Attention for Image Captioning","authors":"Md. Zakir Hossain, F. Sohel, M. F. Shiratuddin, Hamid Laga, Bennamoun","doi":"10.1109/DICTA47822.2019.8946003","DOIUrl":"https://doi.org/10.1109/DICTA47822.2019.8946003","url":null,"abstract":"In a typical image captioning pipeline, a Convolutional Neural Network (CNN) is used as the image encoder and Long Short-Term Memory (LSTM) as the language decoder. LSTM with attention mechanism has shown remarkable performance on sequential data including image captioning. LSTM can retain long-range dependency of sequential data. However, it is hard to parallelize the computations of LSTM because of its inherent sequential characteristics. In order to address this issue, recent works have shown benefits in using self-attention, which is highly parallelizable without requiring any temporal dependencies. However, existing techniques apply attention only in one direction to compute the context of the words. We propose an attention mechanism called Bi-directional Self-Attention (Bi-SAN) for image captioning. It computes attention both in forward and backward directions. It achieves high performance comparable to state-of-the-art methods.","PeriodicalId":6696,"journal":{"name":"2019 Digital Image Computing: Techniques and Applications (DICTA)","volume":"12 1","pages":"1-7"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74464655","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Evaluation of the Impact of Image Spatial Resolution in Designing a Context-Based Fully Convolution Neural Networks for Flood Mapping 基于上下文的洪水制图全卷积神经网络设计中图像空间分辨率影响的评估
Pub Date : 2019-12-01 DOI: 10.1109/DICTA47822.2019.8945888
Chandrama Sarker, Luis Mejías Alvarez, F. Maire, A. Woodley
In this paper, our main aim is to investigate the context-based pixel-wise classification using a fully convolutional neural networks model for flood extent mapping from multispectral remote sensing images. Our approach helps to overcome the limitation of the conventional classification methods with low generalisation ability that used per-pixel spectral information for pixel-wise classification. In this study, a comparative analysis with conventional pixel-wise SVM classifier shows that our proposed model has higher generalisation ability for flooded area detection. By using remote sensing images with different spatial resolutions we also aim to investigate the relationship between image-sensor resolution and neighbourhood window size for context-based classification. Instead of fine-tuning a pre-established deep neural network model, we developed a preliminary base model with two convolutional layers. The model was tested on images with two different spatial resolutions of 3 meters (PlanetScope image) and 30 meters (Landsat-5 Thematic Mapper). During training phase we determined the structure of the convolutional layer as well as the appropriate size of the contextual neighbourhood for those two data types. Preliminary results showed that with increasing the scale of spatial resolutions the required neighbourhood size for training samples also increases. We tested different neighbourhood sized training samples to train the model and the analysis of the performance of those models showed that a 11 × 11 neighbourhood window for PlanetScope data and a 3 × 3 neighbourhood window for Landsat data were found to be the optimum size for classification. Insights from this work may be used to design efficient classifiers in scenarios where data with different resolutions are available.
在本文中,我们的主要目的是研究基于上下文的逐像素分类,使用全卷积神经网络模型从多光谱遥感图像中绘制洪水范围。我们的方法有助于克服传统分类方法使用逐像素光谱信息进行逐像素分类的局限性,其泛化能力较低。在本研究中,与传统的逐像素SVM分类器的比较分析表明,我们提出的模型对洪水区域检测具有更高的泛化能力。通过使用不同空间分辨率的遥感图像,我们还旨在研究图像传感器分辨率与基于上下文分类的邻域窗口大小之间的关系。我们没有对预先建立的深度神经网络模型进行微调,而是开发了一个具有两个卷积层的初步基础模型。该模型在3米(PlanetScope图像)和30米(Landsat-5 Thematic Mapper图像)两种不同空间分辨率的图像上进行了测试。在训练阶段,我们确定了卷积层的结构以及这两种数据类型的上下文邻域的适当大小。初步结果表明,随着空间分辨率尺度的增加,训练样本所需的邻域大小也随之增加。我们测试了不同邻域大小的训练样本来训练模型,并对这些模型的性能进行了分析,发现PlanetScope数据的11 × 11邻域窗口和Landsat数据的3 × 3邻域窗口是分类的最佳大小。从这项工作中获得的见解可以用于在不同分辨率的数据可用的情况下设计有效的分类器。
{"title":"Evaluation of the Impact of Image Spatial Resolution in Designing a Context-Based Fully Convolution Neural Networks for Flood Mapping","authors":"Chandrama Sarker, Luis Mejías Alvarez, F. Maire, A. Woodley","doi":"10.1109/DICTA47822.2019.8945888","DOIUrl":"https://doi.org/10.1109/DICTA47822.2019.8945888","url":null,"abstract":"In this paper, our main aim is to investigate the context-based pixel-wise classification using a fully convolutional neural networks model for flood extent mapping from multispectral remote sensing images. Our approach helps to overcome the limitation of the conventional classification methods with low generalisation ability that used per-pixel spectral information for pixel-wise classification. In this study, a comparative analysis with conventional pixel-wise SVM classifier shows that our proposed model has higher generalisation ability for flooded area detection. By using remote sensing images with different spatial resolutions we also aim to investigate the relationship between image-sensor resolution and neighbourhood window size for context-based classification. Instead of fine-tuning a pre-established deep neural network model, we developed a preliminary base model with two convolutional layers. The model was tested on images with two different spatial resolutions of 3 meters (PlanetScope image) and 30 meters (Landsat-5 Thematic Mapper). During training phase we determined the structure of the convolutional layer as well as the appropriate size of the contextual neighbourhood for those two data types. Preliminary results showed that with increasing the scale of spatial resolutions the required neighbourhood size for training samples also increases. We tested different neighbourhood sized training samples to train the model and the analysis of the performance of those models showed that a 11 × 11 neighbourhood window for PlanetScope data and a 3 × 3 neighbourhood window for Landsat data were found to be the optimum size for classification. Insights from this work may be used to design efficient classifiers in scenarios where data with different resolutions are available.","PeriodicalId":6696,"journal":{"name":"2019 Digital Image Computing: Techniques and Applications (DICTA)","volume":"12 1","pages":"1-8"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81043463","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Using Image Processing to Automatically Measure Pearl Oyster Size for Selective Breeding 基于图像处理的珍珠牡蛎尺寸自动测量技术
Pub Date : 2019-12-01 DOI: 10.1109/DICTA47822.2019.8945902
Adrian Lapico, M. Sankupellay, Louis Cianciullo, Trina S. Myers, D. Konovalov, D. Jerry, P. Toole, David B. Jones, K. Zenger
The growth rate is a genetic trait that is often recorded in pearl oyster farming for use in selective breeding programs. By tracking the growth rate of a pearl oyster, farmers can make better decisions on which oysters to breed or manage in order to produce healthier offspring and higher quality pearls. However, the current practice of measurement by hand results in measurement inaccuracies, slow processing, and unnecessary employee costs. To rectify this, we propose automating the workflow via computer vision techniques, which can be used to capture images of pearl oysters and process the images to obtain the absolute measurements of each oyster. Specifically, we utilise and compare a set of edge detection algorithms to produce an image-processing algorithm that automatically segments an image containing multiple oysters and returns the height and width of the oyster shell. Our final algorithm was tested on images containing 2523 oysters (Pinctada maxima) captured on farming boats in Indonesia. This algorithm achieved reliability (of identifying at least one required oyster measurement correctly) equal to 92.1%.
生长速度是一种遗传性状,通常记录在珍珠牡蛎养殖中,用于选择性育种计划。通过跟踪珍珠牡蛎的生长速度,农民可以更好地决定饲养或管理哪种牡蛎,以生产更健康的后代和更高质量的珍珠。然而,目前手工测量的做法导致测量不准确、处理缓慢和不必要的员工成本。为了纠正这一点,我们提出了通过计算机视觉技术自动化工作流程,该技术可用于捕获珍珠牡蛎的图像并对图像进行处理以获得每个牡蛎的绝对测量值。具体来说,我们利用并比较了一组边缘检测算法来生成一种图像处理算法,该算法可以自动分割包含多个牡蛎的图像,并返回牡蛎壳的高度和宽度。我们的最终算法在印度尼西亚养殖船上捕获的含有2523只牡蛎(Pinctada maxima)的图像上进行了测试。该算法的可靠度(正确识别至少一个所需牡蛎测量值)为92.1%。
{"title":"Using Image Processing to Automatically Measure Pearl Oyster Size for Selective Breeding","authors":"Adrian Lapico, M. Sankupellay, Louis Cianciullo, Trina S. Myers, D. Konovalov, D. Jerry, P. Toole, David B. Jones, K. Zenger","doi":"10.1109/DICTA47822.2019.8945902","DOIUrl":"https://doi.org/10.1109/DICTA47822.2019.8945902","url":null,"abstract":"The growth rate is a genetic trait that is often recorded in pearl oyster farming for use in selective breeding programs. By tracking the growth rate of a pearl oyster, farmers can make better decisions on which oysters to breed or manage in order to produce healthier offspring and higher quality pearls. However, the current practice of measurement by hand results in measurement inaccuracies, slow processing, and unnecessary employee costs. To rectify this, we propose automating the workflow via computer vision techniques, which can be used to capture images of pearl oysters and process the images to obtain the absolute measurements of each oyster. Specifically, we utilise and compare a set of edge detection algorithms to produce an image-processing algorithm that automatically segments an image containing multiple oysters and returns the height and width of the oyster shell. Our final algorithm was tested on images containing 2523 oysters (Pinctada maxima) captured on farming boats in Indonesia. This algorithm achieved reliability (of identifying at least one required oyster measurement correctly) equal to 92.1%.","PeriodicalId":6696,"journal":{"name":"2019 Digital Image Computing: Techniques and Applications (DICTA)","volume":"58 1","pages":"1-8"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79136564","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Registration Based Data Augmentation for Multiple Sclerosis Lesion Segmentation 基于配准的多发性硬化症病灶分割数据增强
Pub Date : 2019-12-01 DOI: 10.1109/DICTA47822.2019.8946022
Ava Assadi Abolvardi, Len Hamey, K. Ho-Shon
Deep learning has shown outstanding performance on various computer vision tasks such as image segmentation. To take advantage of deep learning in image segmentation, one would need a huge amount of annotated data since deep learning models are data-intensive. One of the main challenges of using deep learning methods in the medical domain is the shortage of available annotated data. To tackle this problem, in this paper, we propose a registration based framework for augmenting multiple sclerosis datasets. In this framework, by registering images of two different patients, we create a new image, which smoothly adds lesions from the first patient into a brain image, structured like the second patient. Due to their nature, multiple sclerosis lesions vary in shape, size, location and number of occurrence, thus registering images of two different subjects, will create a realistic image. The proposed method is capable of introducing diversity to data distribution, which other traditional augmentation methods do not offer. To check the effectiveness of our proposed method, we compare the performance of 3D-Unet on different augmented and non-augmented datasets. Experimental results indicate that the best performance is achieved when combining both the proposed method with traditional augmentation techniques.
深度学习在图像分割等各种计算机视觉任务中表现出色。由于深度学习模型是数据密集型的,因此要在图像分割中利用深度学习,需要大量的注释数据。在医学领域使用深度学习方法的主要挑战之一是缺乏可用的注释数据。为了解决这个问题,在本文中,我们提出了一个基于配准的框架来增强多发性硬化症数据集。在这个框架中,通过注册两个不同患者的图像,我们创建了一个新图像,它平滑地将第一个患者的病变添加到与第二个患者结构相似的大脑图像中。由于多发性硬化症病变的性质不同,其形状、大小、位置和发生次数各不相同,因此对两个不同受试者的图像进行配准,将会得到真实的图像。该方法能够为数据分布引入多样性,这是其他传统增强方法所不具备的。为了验证我们提出的方法的有效性,我们比较了3D-Unet在不同增强和非增强数据集上的性能。实验结果表明,将该方法与传统的增强技术相结合,可以获得最佳的增强效果。
{"title":"Registration Based Data Augmentation for Multiple Sclerosis Lesion Segmentation","authors":"Ava Assadi Abolvardi, Len Hamey, K. Ho-Shon","doi":"10.1109/DICTA47822.2019.8946022","DOIUrl":"https://doi.org/10.1109/DICTA47822.2019.8946022","url":null,"abstract":"Deep learning has shown outstanding performance on various computer vision tasks such as image segmentation. To take advantage of deep learning in image segmentation, one would need a huge amount of annotated data since deep learning models are data-intensive. One of the main challenges of using deep learning methods in the medical domain is the shortage of available annotated data. To tackle this problem, in this paper, we propose a registration based framework for augmenting multiple sclerosis datasets. In this framework, by registering images of two different patients, we create a new image, which smoothly adds lesions from the first patient into a brain image, structured like the second patient. Due to their nature, multiple sclerosis lesions vary in shape, size, location and number of occurrence, thus registering images of two different subjects, will create a realistic image. The proposed method is capable of introducing diversity to data distribution, which other traditional augmentation methods do not offer. To check the effectiveness of our proposed method, we compare the performance of 3D-Unet on different augmented and non-augmented datasets. Experimental results indicate that the best performance is achieved when combining both the proposed method with traditional augmentation techniques.","PeriodicalId":6696,"journal":{"name":"2019 Digital Image Computing: Techniques and Applications (DICTA)","volume":"68 1","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85812484","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Incorporating the Barzilai-Borwein Adaptive Step Size into Sugradient Methods for Deep Network Training 将Barzilai-Borwein自适应步长纳入深度网络训练的梯度方法
Pub Date : 2019-12-01 DOI: 10.1109/DICTA47822.2019.8945980
A. Robles-Kelly, A. Nazari
In this paper, we incorporate the Barzilai-Borwein [2] step size into gradient descent methods used to train deep networks. This allows us to adapt the learning rate using a two-point approximation to the secant equation which quasi-Newton methods are based upon. Moreover, the adaptive learning rate method presented here is quite general in nature and can be applied to widely used gradient descent approaches such as Adagrad [7] and RMSprop. We evaluate our method using standard example network architectures on widely available datasets and compare against alternatives elsewhere in the literature. In our experiments, our adaptive learning rate shows a smoother and faster convergence than that exhibited by the alternatives, with better or comparable performance.
在本文中,我们将Barzilai-Borwein[2]步长纳入用于训练深度网络的梯度下降方法中。这允许我们使用两点近似的正割方程来调整学习率,而准牛顿方法是基于正割方程的。此外,本文提出的自适应学习率方法具有相当的通用性,可以应用于Adagrad[7]和RMSprop等广泛使用的梯度下降方法。我们在广泛可用的数据集上使用标准示例网络架构来评估我们的方法,并与文献中的其他替代方案进行比较。在我们的实验中,我们的自适应学习率显示出比替代方案更平滑和更快的收敛速度,具有更好或相当的性能。
{"title":"Incorporating the Barzilai-Borwein Adaptive Step Size into Sugradient Methods for Deep Network Training","authors":"A. Robles-Kelly, A. Nazari","doi":"10.1109/DICTA47822.2019.8945980","DOIUrl":"https://doi.org/10.1109/DICTA47822.2019.8945980","url":null,"abstract":"In this paper, we incorporate the Barzilai-Borwein [2] step size into gradient descent methods used to train deep networks. This allows us to adapt the learning rate using a two-point approximation to the secant equation which quasi-Newton methods are based upon. Moreover, the adaptive learning rate method presented here is quite general in nature and can be applied to widely used gradient descent approaches such as Adagrad [7] and RMSprop. We evaluate our method using standard example network architectures on widely available datasets and compare against alternatives elsewhere in the literature. In our experiments, our adaptive learning rate shows a smoother and faster convergence than that exhibited by the alternatives, with better or comparable performance.","PeriodicalId":6696,"journal":{"name":"2019 Digital Image Computing: Techniques and Applications (DICTA)","volume":"25 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90457223","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Wave Scale, Speed and Direction from Airborne Video of Maritime Scene 海上场景航拍视频中的海浪尺度、速度和方向
Pub Date : 2019-12-01 DOI: 10.1109/DICTA47822.2019.8946116
Kent Rosser, J. Chahl
Ocean surfaces and large water bodies are commonly monitored by aircraft. While water features are visually non-static, they do include information that allows determination of water motion which has applications in navigation, assessing sub-surface changes and the estimation of drift and size of objects within the scene. This study presents an enhancement of state of the art methods to extract water surface features from imagery acquired by an overhead aircraft and assesses its performance on a real world maritime scene.
海洋表面和大型水体通常由飞机监测。虽然水景在视觉上是非静态的,但它们确实包含了可以确定水运动的信息,这在导航、评估地下变化以及估计场景中物体的漂移和大小方面具有应用。本研究提出了一种改进的最先进的方法,从架空飞机获取的图像中提取水面特征,并评估其在真实世界海洋场景中的性能。
{"title":"Wave Scale, Speed and Direction from Airborne Video of Maritime Scene","authors":"Kent Rosser, J. Chahl","doi":"10.1109/DICTA47822.2019.8946116","DOIUrl":"https://doi.org/10.1109/DICTA47822.2019.8946116","url":null,"abstract":"Ocean surfaces and large water bodies are commonly monitored by aircraft. While water features are visually non-static, they do include information that allows determination of water motion which has applications in navigation, assessing sub-surface changes and the estimation of drift and size of objects within the scene. This study presents an enhancement of state of the art methods to extract water surface features from imagery acquired by an overhead aircraft and assesses its performance on a real world maritime scene.","PeriodicalId":6696,"journal":{"name":"2019 Digital Image Computing: Techniques and Applications (DICTA)","volume":"47 1","pages":"1-7"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89731573","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Ensemble of Training Models for Road and Building Segmentation 道路和建筑物分割的训练模型集成
Pub Date : 2019-12-01 DOI: 10.1109/DICTA47822.2019.8945903
Ryosuke Kamiya, Kyoya Sawada, K. Hotta
In this paper, we propose an object segmentation method in satellite images by the ensemble of models obtained through training process. To improve recognition accuracy, the ensemble of models obtained by different random seeds is used. Here we pay attention to the ensemble of models obtained through training process. In model ensemble, we should integrate the models with different opinions. Since the pixels with low probability such as boundary are often updated through training process, each model in training process has different probability for boundary regions, and the ensemble of those probability maps is effective for improving segmentation accuracy. Effectiveness of the ensemble of training models is demonstrated by experiments on building and road segmentation. Our proposed method improved approximately 4% in comparison with the best model selected by validation. Our method also achieved better accuracy than the standard ensemble of models.
本文提出了一种基于训练模型集成的卫星图像目标分割方法。为了提高识别精度,采用不同随机种子获得的模型集成。这里我们关注的是通过训练过程得到的模型的集成。在模型集成中,我们应该把不同观点的模型整合起来。由于边界等概率较低的像素点在训练过程中经常更新,因此训练过程中的每个模型对边界区域的概率不同,这些概率图的集合对于提高分割精度是有效的。通过建筑物和道路分割的实验验证了训练模型集成的有效性。与验证选择的最佳模型相比,我们提出的方法提高了约4%。我们的方法也获得了比标准模型集成更好的精度。
{"title":"Ensemble of Training Models for Road and Building Segmentation","authors":"Ryosuke Kamiya, Kyoya Sawada, K. Hotta","doi":"10.1109/DICTA47822.2019.8945903","DOIUrl":"https://doi.org/10.1109/DICTA47822.2019.8945903","url":null,"abstract":"In this paper, we propose an object segmentation method in satellite images by the ensemble of models obtained through training process. To improve recognition accuracy, the ensemble of models obtained by different random seeds is used. Here we pay attention to the ensemble of models obtained through training process. In model ensemble, we should integrate the models with different opinions. Since the pixels with low probability such as boundary are often updated through training process, each model in training process has different probability for boundary regions, and the ensemble of those probability maps is effective for improving segmentation accuracy. Effectiveness of the ensemble of training models is demonstrated by experiments on building and road segmentation. Our proposed method improved approximately 4% in comparison with the best model selected by validation. Our method also achieved better accuracy than the standard ensemble of models.","PeriodicalId":6696,"journal":{"name":"2019 Digital Image Computing: Techniques and Applications (DICTA)","volume":"1 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89291254","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2019 Digital Image Computing: Techniques and Applications (DICTA)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1