首页 > 最新文献

2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)最新文献

英文 中文
Visual Object Tracking in Spherical 360° Videos: A Bridging Approach 球形360°视频中的视觉对象跟踪:桥接方法
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290549
Simon Finnie, Fang-Lue Zhang, Taehyun Rhee
We present a novel approach for adapting existing visual object trackers (VOT) to work for equirectangular video, utilizing image reprojection. Our system can easily be integrated with existing VOT algorithms, significantly increasing the accuracy and robustness of tracking in spherical 360° environments without requiring retraining. Our adapted approach involves the orthographic projection of a subsection of the image centered around the tracked object each frame. Our projection reduces the distortion around the tracked object each frame, allowing the VOT algorithm to more easily track the object as it moves.
我们提出了一种利用图像重投影的新方法,使现有的视觉目标跟踪器(VOT)适用于等矩形视频。我们的系统可以很容易地与现有的VOT算法集成,显着提高了球形360°环境中跟踪的准确性和鲁棒性,而无需再训练。我们的改进方法包括每帧以跟踪对象为中心的图像分段的正交投影。我们的投影减少了每帧被跟踪对象周围的失真,允许VOT算法更容易地跟踪物体的移动。
{"title":"Visual Object Tracking in Spherical 360° Videos: A Bridging Approach","authors":"Simon Finnie, Fang-Lue Zhang, Taehyun Rhee","doi":"10.1109/IVCNZ51579.2020.9290549","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290549","url":null,"abstract":"We present a novel approach for adapting existing visual object trackers (VOT) to work for equirectangular video, utilizing image reprojection. Our system can easily be integrated with existing VOT algorithms, significantly increasing the accuracy and robustness of tracking in spherical 360° environments without requiring retraining. Our adapted approach involves the orthographic projection of a subsection of the image centered around the tracked object each frame. Our projection reduces the distortion around the tracked object each frame, allowing the VOT algorithm to more easily track the object as it moves.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128492647","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Automatic Identification of Diatom Morphology using Deep Learning 基于深度学习的硅藻形态自动识别
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290564
Dana Lambert, R. Green
This paper proposes a method to automatically identify diatom frustules using nine morphological categories. A total of 7092 images from NIWA and ADIAC with related taxa data were used to create training and test sets. Different augmentations and image processing methods were used on the training set to see if this would increase accuracy. Several CNNs were trained over a total of 50 epochs and the highest accuracy model was saved based on the validation set. Resnet-50 produced the highest accuracy of 94%, which is not as accurate as a similar study that achieved 99%, although this was for a slightly different classification problem.
提出了一种利用9个形态分类自动识别硅藻藻的方法。使用来自NIWA和ADIAC的7092张带有相关分类群数据的图像创建训练集和测试集。在训练集上使用不同的增强和图像处理方法,看看这是否会提高准确性。对多个cnn进行了50次epoch的训练,并在验证集的基础上保存了准确率最高的模型。Resnet-50的准确率最高,达到了94%,尽管这是一个略有不同的分类问题,但它的准确率不如一个类似的研究达到的99%。
{"title":"Automatic Identification of Diatom Morphology using Deep Learning","authors":"Dana Lambert, R. Green","doi":"10.1109/IVCNZ51579.2020.9290564","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290564","url":null,"abstract":"This paper proposes a method to automatically identify diatom frustules using nine morphological categories. A total of 7092 images from NIWA and ADIAC with related taxa data were used to create training and test sets. Different augmentations and image processing methods were used on the training set to see if this would increase accuracy. Several CNNs were trained over a total of 50 epochs and the highest accuracy model was saved based on the validation set. Resnet-50 produced the highest accuracy of 94%, which is not as accurate as a similar study that achieved 99%, although this was for a slightly different classification problem.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"29 8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125697168","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A machine learning approach for image retrieval tasks 图像检索任务的机器学习方法
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290617
Achref Ouni
Several methods based on visual methods (BoVW, VLAD,…) or recent deep leaning methods try to solve the CBIR problem. Bag of visual words (BoVW) is one of most module used for both classification and image recognition. But, even with the high performance of BoVW, the problem of retrieving the image by content is still a challenge in computer vision. In this paper, we propose an improvement on a bag of visual words by increasing the accuracy of the retrieved candidates. In addition, we reduce the signature construction time by exploiting the powerful of the approximate nearest neighbor algorithms (ANNs). Experimental results will be applied to widely data sets (UKB, Wang, Corel 10K) and with different descriptors (CMI, SURF).
一些基于视觉方法(BoVW, VLAD,…)或最近的深度学习方法试图解决CBIR问题。视觉词包(BoVW)是分类和图像识别中应用最广泛的模块之一。但是,即使BoVW具有很高的性能,根据内容检索图像的问题仍然是计算机视觉中的一个挑战。在本文中,我们提出了一种改进视觉词包的方法,通过提高检索候选词的准确性。此外,利用近似最近邻算法(ann)的强大功能,减少了签名构建时间。实验结果将应用于广泛的数据集(UKB, Wang, Corel 10K)和不同的描述符(CMI, SURF)。
{"title":"A machine learning approach for image retrieval tasks","authors":"Achref Ouni","doi":"10.1109/IVCNZ51579.2020.9290617","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290617","url":null,"abstract":"Several methods based on visual methods (BoVW, VLAD,…) or recent deep leaning methods try to solve the CBIR problem. Bag of visual words (BoVW) is one of most module used for both classification and image recognition. But, even with the high performance of BoVW, the problem of retrieving the image by content is still a challenge in computer vision. In this paper, we propose an improvement on a bag of visual words by increasing the accuracy of the retrieved candidates. In addition, we reduce the signature construction time by exploiting the powerful of the approximate nearest neighbor algorithms (ANNs). Experimental results will be applied to widely data sets (UKB, Wang, Corel 10K) and with different descriptors (CMI, SURF).","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127598139","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Development of a Virtual Environment Based Image Generation Tool for Neural Network Training 基于虚拟环境的神经网络训练图像生成工具的开发
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290491
R. Arenas, P. Delmas, Alfonso Gastelum-Strozzi
We present a computational tool to generate visual and descriptive data used as additional training images for neural networks involved in image recognition tasks. The work is inspired by the problem posed to acquire enough data, in order to train service robots, with the goal of improving the range of objects in the environment with which they can interact. The tool provides a framework that allows users to easily setup different environments with the visual information needed for the training, accordingly to their needs. The tool was developed with the Unity engine, and it was designed to be able to import external prefabs. These models are standardized and catalogued into lists, which are accessed to create more complex and diverse virtual environments. Another component of the tool adds an additional layer of complexity by creating randomized environments with different conditions (scale, position and orientation of objects, and environmental illumination). The performance of the created dataset was tested by training the information on the YOLO-V3 (You Only Look Once) architecture and testing on both artificial and real images.
我们提出了一种计算工具,用于生成视觉和描述性数据,作为涉及图像识别任务的神经网络的额外训练图像。这项工作的灵感来自于一个问题,即获取足够的数据,以训练服务机器人,目标是提高环境中物体的范围,使它们能够与之互动。该工具提供了一个框架,允许用户根据自己的需要轻松地设置不同的环境,并提供培训所需的视觉信息。该工具是使用Unity引擎开发的,它被设计成能够导入外部预制件。这些模型被标准化并编目到列表中,可以访问这些列表来创建更复杂和多样化的虚拟环境。该工具的另一个组件通过创建具有不同条件(对象的规模、位置和方向以及环境照明)的随机环境增加了额外的复杂性层。通过在YOLO-V3 (You Only Look Once)架构上训练信息,并在人工图像和真实图像上进行测试,测试了所创建数据集的性能。
{"title":"Development of a Virtual Environment Based Image Generation Tool for Neural Network Training","authors":"R. Arenas, P. Delmas, Alfonso Gastelum-Strozzi","doi":"10.1109/IVCNZ51579.2020.9290491","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290491","url":null,"abstract":"We present a computational tool to generate visual and descriptive data used as additional training images for neural networks involved in image recognition tasks. The work is inspired by the problem posed to acquire enough data, in order to train service robots, with the goal of improving the range of objects in the environment with which they can interact. The tool provides a framework that allows users to easily setup different environments with the visual information needed for the training, accordingly to their needs. The tool was developed with the Unity engine, and it was designed to be able to import external prefabs. These models are standardized and catalogued into lists, which are accessed to create more complex and diverse virtual environments. Another component of the tool adds an additional layer of complexity by creating randomized environments with different conditions (scale, position and orientation of objects, and environmental illumination). The performance of the created dataset was tested by training the information on the YOLO-V3 (You Only Look Once) architecture and testing on both artificial and real images.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"118 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133913845","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PProCRC: Probabilistic Collaboration of Image Patches for Fine-grained Classification PProCRC:用于细粒度分类的图像补丁的概率协作
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290537
Tapabrata (Rohan) Chakraborty, B. McCane, S. Mills, U. Pal
We present a conditional probabilistic framework for collaborative representation of image patches. It incorporates background compensation and outlier patch suppression into the main formulation itself, thus doing away with the need for pre-processing steps to handle the same. A closed form non-iterative solution of the cost function is derived. The proposed method (PProCRC) outperforms earlier CRC formulations: patch based (PCRC, GP-CRC) as well as the state-of-the-art probabilistic (ProCRC and EProCRC) on three fine-grained species recognition datasets (Oxford Flowers, Oxford-IIIT Pets and CUB Birds) using two CNN backbones (Vgg-19 and ResNet-50).
我们提出了一个条件概率框架,用于图像补丁的协同表示。它将背景补偿和离群斑抑制纳入主配方本身,从而消除了处理相同的预处理步骤的需要。导出了代价函数的闭形式非迭代解。所提出的方法(PProCRC)优于早期的CRC方案:基于补丁的(PCRC, GP-CRC)以及最先进的概率(ProCRC和EProCRC)在三个细粒度物种识别数据集(Oxford Flowers, Oxford- iiit Pets和CUB Birds)上使用两个CNN主干(Vgg-19和ResNet-50)。
{"title":"PProCRC: Probabilistic Collaboration of Image Patches for Fine-grained Classification","authors":"Tapabrata (Rohan) Chakraborty, B. McCane, S. Mills, U. Pal","doi":"10.1109/IVCNZ51579.2020.9290537","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290537","url":null,"abstract":"We present a conditional probabilistic framework for collaborative representation of image patches. It incorporates background compensation and outlier patch suppression into the main formulation itself, thus doing away with the need for pre-processing steps to handle the same. A closed form non-iterative solution of the cost function is derived. The proposed method (PProCRC) outperforms earlier CRC formulations: patch based (PCRC, GP-CRC) as well as the state-of-the-art probabilistic (ProCRC and EProCRC) on three fine-grained species recognition datasets (Oxford Flowers, Oxford-IIIT Pets and CUB Birds) using two CNN backbones (Vgg-19 and ResNet-50).","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131484430","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Image and Text fusion for UPMC Food-101 using BERT and CNNs 基于BERT和cnn的UPMC Food-101图像和文本融合
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290622
I. Gallo, Gianmarco Ria, Nicola Landro, Riccardo La Grassa
The modern digital world is becoming more and more multimodal. Looking on the internet, images are often associated with the text, so classification problems with these two modalities are very common. In this paper, we examine multimodal classification using textual information and visual representations of the same concept. We investigate two main basic methods to perform multimodal fusion and adapt them with stacking techniques to better handle this type of problem. Here, we use UPMC Food-101, which is a difficult and noisy multimodal dataset that well represents this category of multimodal problems. Our results show that the proposed early fusion technique combined with a stacking-based approach exceeds the state of the art on the dataset used.
现代数字世界正变得越来越多模式。在互联网上,图像经常与文本联系在一起,因此这两种模式的分类问题非常普遍。在本文中,我们使用同一概念的文本信息和视觉表示来研究多模态分类。我们研究了两种主要的多模态融合的基本方法,并将它们与叠加技术相结合,以更好地处理这类问题。在这里,我们使用UPMC Food-101,这是一个困难和有噪声的多模态数据集,很好地代表了这类多模态问题。我们的结果表明,所提出的早期融合技术与基于堆栈的方法相结合,在所使用的数据集上超过了目前的水平。
{"title":"Image and Text fusion for UPMC Food-101 using BERT and CNNs","authors":"I. Gallo, Gianmarco Ria, Nicola Landro, Riccardo La Grassa","doi":"10.1109/IVCNZ51579.2020.9290622","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290622","url":null,"abstract":"The modern digital world is becoming more and more multimodal. Looking on the internet, images are often associated with the text, so classification problems with these two modalities are very common. In this paper, we examine multimodal classification using textual information and visual representations of the same concept. We investigate two main basic methods to perform multimodal fusion and adapt them with stacking techniques to better handle this type of problem. Here, we use UPMC Food-101, which is a difficult and noisy multimodal dataset that well represents this category of multimodal problems. Our results show that the proposed early fusion technique combined with a stacking-based approach exceeds the state of the art on the dataset used.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114242894","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Melanoma and Nevi Classification using Convolution Neural Networks 使用卷积神经网络进行黑色素瘤和痣分类
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290736
R. Grove, R. Green
Early identification of melanoma skin cancer is vital for the improvement of patients’ prospects of five year disease free survival. The majority of malignant skin lesions present at a general practice level where a diagnosis is based on a clinical decision algorithm. As a false negative diagnosis is an unacceptable outcome, clinical caution tends to result in a low positive predictive value of as low at 8%. There has been a large burden of surgical excisions that retrospectively prove to have been unnecessary.This paper proposes a method to identify melanomas in dermoscopic images using a convolution neural network (CNN). The proposed method implements transfer learning based on the ResNet50 CNN, pretrained using the ImageNet dataset. Datasets from the ISIC Archive were implemented during training, validation and testing. Further tests were performed on a smaller dataset of images taken from the Dermnet NZ website and from recent clinical cases still awaiting histological results to indicate the trained network’s ability to generalise to real cases. The 86% test accuracy achieved with the proposed method was comparable to the results of prior studies but required significantly less pre-processing actions to classify a lesion and was not dependant on consistent image scaling or the presence of a scale on the image. This method also improved on past research by making use of all of the information present in an image as opposed to focusing on geometric and colour-space based aspects independently.
黑色素瘤皮肤癌的早期识别对于改善患者五年无病生存的前景至关重要。大多数恶性皮肤病变存在于一般实践水平,其中诊断是基于临床决策算法。由于假阴性诊断是不可接受的结果,临床谨慎倾向于导致低阳性预测值,低至8%。手术切除的负担很大,事后证明是不必要的。本文提出了一种使用卷积神经网络(CNN)识别皮肤镜图像中的黑色素瘤的方法。该方法基于ResNet50 CNN实现迁移学习,使用ImageNet数据集进行预训练。来自ISIC存档的数据集在培训、验证和测试期间被实施。进一步的测试是在一个较小的图像数据集上进行的,这些数据集来自新西兰Dermnet网站和最近的临床病例,这些病例仍在等待组织学结果,以表明训练后的网络能够推广到真实病例。该方法达到了86%的测试准确度,与之前的研究结果相当,但对病变进行分类所需的预处理操作明显减少,并且不依赖于一致的图像缩放或图像上存在的缩放。这种方法也改进了过去的研究,它利用了图像中存在的所有信息,而不是单独关注基于几何和色彩空间的方面。
{"title":"Melanoma and Nevi Classification using Convolution Neural Networks","authors":"R. Grove, R. Green","doi":"10.1109/IVCNZ51579.2020.9290736","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290736","url":null,"abstract":"Early identification of melanoma skin cancer is vital for the improvement of patients’ prospects of five year disease free survival. The majority of malignant skin lesions present at a general practice level where a diagnosis is based on a clinical decision algorithm. As a false negative diagnosis is an unacceptable outcome, clinical caution tends to result in a low positive predictive value of as low at 8%. There has been a large burden of surgical excisions that retrospectively prove to have been unnecessary.This paper proposes a method to identify melanomas in dermoscopic images using a convolution neural network (CNN). The proposed method implements transfer learning based on the ResNet50 CNN, pretrained using the ImageNet dataset. Datasets from the ISIC Archive were implemented during training, validation and testing. Further tests were performed on a smaller dataset of images taken from the Dermnet NZ website and from recent clinical cases still awaiting histological results to indicate the trained network’s ability to generalise to real cases. The 86% test accuracy achieved with the proposed method was comparable to the results of prior studies but required significantly less pre-processing actions to classify a lesion and was not dependant on consistent image scaling or the presence of a scale on the image. This method also improved on past research by making use of all of the information present in an image as opposed to focusing on geometric and colour-space based aspects independently.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116126667","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Introducing Transfer Leaming to 3D ResNet-18 for Alzheimer’s Disease Detection on MRI Images 将迁移学习引入3D ResNet-18,用于MRI图像上的阿尔茨海默病检测
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290616
Amir Ebrahimi, S. Luo, R. Chiong
This paper focuses on detecting Alzheimer’s Disease (AD) using the ResNet-18 model on Magnetic Resonance Imaging (MRI). Previous studies have applied different 2D Convolutional Neural Networks (CNNs) to detect AD. The main idea being to split 3D MRI scans into 2D image slices, so that classification can be performed on the image slices independently. This idea allows researchers to benefit from the concept of transfer learning. However, 2D CNNs are incapable of understanding the relationship among 2D image slices in a 3D MRI scan. One solution is to employ 3D CNNs instead of 2D ones. In this paper, we propose a method to utilise transfer learning in 3D CNNs, which allows the transfer of knowledge from 2D image datasets to a 3D image dataset. Both 2D and 3D CNNs are compared in this study, and our results show that introducing transfer learning to a 3D CNN improves the accuracy of an AD detection system. After using an optimisation method in the training process, our approach achieved 96.88% accuracy, 100% sensitivity, and 93.75% specificity.
本文主要研究利用磁共振成像(MRI)的ResNet-18模型检测阿尔茨海默病(AD)。以前的研究已经使用了不同的二维卷积神经网络(cnn)来检测AD。其主要思想是将3D MRI扫描图像分割成2D图像切片,这样就可以在图像切片上独立进行分类。这个想法使研究人员受益于迁移学习的概念。然而,2D cnn无法理解3D MRI扫描中2D图像切片之间的关系。一种解决方案是使用3D cnn而不是2D cnn。在本文中,我们提出了一种在3D cnn中利用迁移学习的方法,该方法允许将知识从2D图像数据集转移到3D图像数据集。在本研究中,我们对2D和3D CNN进行了比较,结果表明,将迁移学习引入3D CNN可以提高AD检测系统的准确性。在训练过程中使用优化方法后,我们的方法准确率达到96.88%,灵敏度为100%,特异性为93.75%。
{"title":"Introducing Transfer Leaming to 3D ResNet-18 for Alzheimer’s Disease Detection on MRI Images","authors":"Amir Ebrahimi, S. Luo, R. Chiong","doi":"10.1109/IVCNZ51579.2020.9290616","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290616","url":null,"abstract":"This paper focuses on detecting Alzheimer’s Disease (AD) using the ResNet-18 model on Magnetic Resonance Imaging (MRI). Previous studies have applied different 2D Convolutional Neural Networks (CNNs) to detect AD. The main idea being to split 3D MRI scans into 2D image slices, so that classification can be performed on the image slices independently. This idea allows researchers to benefit from the concept of transfer learning. However, 2D CNNs are incapable of understanding the relationship among 2D image slices in a 3D MRI scan. One solution is to employ 3D CNNs instead of 2D ones. In this paper, we propose a method to utilise transfer learning in 3D CNNs, which allows the transfer of knowledge from 2D image datasets to a 3D image dataset. Both 2D and 3D CNNs are compared in this study, and our results show that introducing transfer learning to a 3D CNN improves the accuracy of an AD detection system. After using an optimisation method in the training process, our approach achieved 96.88% accuracy, 100% sensitivity, and 93.75% specificity.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"558 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116275792","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
Heating Patterns Recognition in Industrial Microwave-Processed Foods 工业微波加工食品的加热模式识别
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290639
Sowmya Kasturi, S. L. Moan, D. Bailey, Jeremy Smith
Recognition or identification of hot and cold spot heating patterns in microwave-processed pre-packaged food products is crucial to determine experimental repeatability and design better and safer food treatment systems. This review focuses on computer vision-based methods for heating patterns recognition from the literature along with their limitations. A preliminary kinetics study to correlate colour to varied timetemperature combinations is also discussed.
识别或识别微波加工预包装食品中的热点和冷点加热模式对于确定实验可重复性和设计更好和更安全的食品处理系统至关重要。本文综述了基于计算机视觉的热模式识别方法及其局限性。初步的动力学研究,以关联颜色变化的时间-温度组合也进行了讨论。
{"title":"Heating Patterns Recognition in Industrial Microwave-Processed Foods","authors":"Sowmya Kasturi, S. L. Moan, D. Bailey, Jeremy Smith","doi":"10.1109/IVCNZ51579.2020.9290639","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290639","url":null,"abstract":"Recognition or identification of hot and cold spot heating patterns in microwave-processed pre-packaged food products is crucial to determine experimental repeatability and design better and safer food treatment systems. This review focuses on computer vision-based methods for heating patterns recognition from the literature along with their limitations. A preliminary kinetics study to correlate colour to varied timetemperature combinations is also discussed.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123411317","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Real Time Ray Tracing of Analytic and Implicit Surfaces 解析曲面和隐式曲面的实时光线追踪
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290653
Finn Petrie, S. Mills
Real-time ray-tracing debuted to consumer GPU hardware in 2018. Primary examples however, have been of hybrid raster and ray-tracing methods that are restricted to triangle mesh geometry. Our research looks at the viability of procedural methods in the real-time setting. We give implementations of analytical and implicit geometry in the domain of the global illumination algorithms bi-directional path-tracing, and GPU Photon-Mapping – both of which we have adapted to the new ray-tracing shader stages, as shown in Figure 1. Despite procedural intersections being more expensive than triangle intersections in Nvidia’s RTX hardware, our results show that these descriptions still run at interactive rates within computationally expensive multi-pass ray-traced global illumination and demonstrate the practical benefits of the geometry.
实时光线追踪于2018年首次亮相消费类GPU硬件。然而,主要的例子是混合光栅和光线追踪方法,这些方法仅限于三角形网格几何。我们的研究着眼于程序方法在实时环境中的可行性。我们给出了全局照明算法领域的解析几何和隐式几何的实现,双向路径跟踪和GPU光子映射-我们已经适应了新的光线跟踪着色器阶段,如图1所示。尽管在Nvidia的RTX硬件中,程序交叉点比三角形交叉点更昂贵,但我们的结果表明,这些描述仍然在计算昂贵的多通道光线跟踪全局照明中以交互速率运行,并展示了几何图形的实际优势。
{"title":"Real Time Ray Tracing of Analytic and Implicit Surfaces","authors":"Finn Petrie, S. Mills","doi":"10.1109/IVCNZ51579.2020.9290653","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290653","url":null,"abstract":"Real-time ray-tracing debuted to consumer GPU hardware in 2018. Primary examples however, have been of hybrid raster and ray-tracing methods that are restricted to triangle mesh geometry. Our research looks at the viability of procedural methods in the real-time setting. We give implementations of analytical and implicit geometry in the domain of the global illumination algorithms bi-directional path-tracing, and GPU Photon-Mapping – both of which we have adapted to the new ray-tracing shader stages, as shown in Figure 1. Despite procedural intersections being more expensive than triangle intersections in Nvidia’s RTX hardware, our results show that these descriptions still run at interactive rates within computationally expensive multi-pass ray-traced global illumination and demonstrate the practical benefits of the geometry.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123748453","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1