首页 > 最新文献

2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)最新文献

英文 中文
Visual Object Tracking in Spherical 360° Videos: A Bridging Approach 球形360°视频中的视觉对象跟踪:桥接方法
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290549
Simon Finnie, Fang-Lue Zhang, Taehyun Rhee
We present a novel approach for adapting existing visual object trackers (VOT) to work for equirectangular video, utilizing image reprojection. Our system can easily be integrated with existing VOT algorithms, significantly increasing the accuracy and robustness of tracking in spherical 360° environments without requiring retraining. Our adapted approach involves the orthographic projection of a subsection of the image centered around the tracked object each frame. Our projection reduces the distortion around the tracked object each frame, allowing the VOT algorithm to more easily track the object as it moves.
我们提出了一种利用图像重投影的新方法,使现有的视觉目标跟踪器(VOT)适用于等矩形视频。我们的系统可以很容易地与现有的VOT算法集成,显着提高了球形360°环境中跟踪的准确性和鲁棒性,而无需再训练。我们的改进方法包括每帧以跟踪对象为中心的图像分段的正交投影。我们的投影减少了每帧被跟踪对象周围的失真,允许VOT算法更容易地跟踪物体的移动。
{"title":"Visual Object Tracking in Spherical 360° Videos: A Bridging Approach","authors":"Simon Finnie, Fang-Lue Zhang, Taehyun Rhee","doi":"10.1109/IVCNZ51579.2020.9290549","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290549","url":null,"abstract":"We present a novel approach for adapting existing visual object trackers (VOT) to work for equirectangular video, utilizing image reprojection. Our system can easily be integrated with existing VOT algorithms, significantly increasing the accuracy and robustness of tracking in spherical 360° environments without requiring retraining. Our adapted approach involves the orthographic projection of a subsection of the image centered around the tracked object each frame. Our projection reduces the distortion around the tracked object each frame, allowing the VOT algorithm to more easily track the object as it moves.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128492647","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Automatic Identification of Diatom Morphology using Deep Learning 基于深度学习的硅藻形态自动识别
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290564
Dana Lambert, R. Green
This paper proposes a method to automatically identify diatom frustules using nine morphological categories. A total of 7092 images from NIWA and ADIAC with related taxa data were used to create training and test sets. Different augmentations and image processing methods were used on the training set to see if this would increase accuracy. Several CNNs were trained over a total of 50 epochs and the highest accuracy model was saved based on the validation set. Resnet-50 produced the highest accuracy of 94%, which is not as accurate as a similar study that achieved 99%, although this was for a slightly different classification problem.
提出了一种利用9个形态分类自动识别硅藻藻的方法。使用来自NIWA和ADIAC的7092张带有相关分类群数据的图像创建训练集和测试集。在训练集上使用不同的增强和图像处理方法,看看这是否会提高准确性。对多个cnn进行了50次epoch的训练,并在验证集的基础上保存了准确率最高的模型。Resnet-50的准确率最高,达到了94%,尽管这是一个略有不同的分类问题,但它的准确率不如一个类似的研究达到的99%。
{"title":"Automatic Identification of Diatom Morphology using Deep Learning","authors":"Dana Lambert, R. Green","doi":"10.1109/IVCNZ51579.2020.9290564","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290564","url":null,"abstract":"This paper proposes a method to automatically identify diatom frustules using nine morphological categories. A total of 7092 images from NIWA and ADIAC with related taxa data were used to create training and test sets. Different augmentations and image processing methods were used on the training set to see if this would increase accuracy. Several CNNs were trained over a total of 50 epochs and the highest accuracy model was saved based on the validation set. Resnet-50 produced the highest accuracy of 94%, which is not as accurate as a similar study that achieved 99%, although this was for a slightly different classification problem.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"29 8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125697168","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A machine learning approach for image retrieval tasks 图像检索任务的机器学习方法
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290617
Achref Ouni
Several methods based on visual methods (BoVW, VLAD,…) or recent deep leaning methods try to solve the CBIR problem. Bag of visual words (BoVW) is one of most module used for both classification and image recognition. But, even with the high performance of BoVW, the problem of retrieving the image by content is still a challenge in computer vision. In this paper, we propose an improvement on a bag of visual words by increasing the accuracy of the retrieved candidates. In addition, we reduce the signature construction time by exploiting the powerful of the approximate nearest neighbor algorithms (ANNs). Experimental results will be applied to widely data sets (UKB, Wang, Corel 10K) and with different descriptors (CMI, SURF).
一些基于视觉方法(BoVW, VLAD,…)或最近的深度学习方法试图解决CBIR问题。视觉词包(BoVW)是分类和图像识别中应用最广泛的模块之一。但是,即使BoVW具有很高的性能,根据内容检索图像的问题仍然是计算机视觉中的一个挑战。在本文中,我们提出了一种改进视觉词包的方法,通过提高检索候选词的准确性。此外,利用近似最近邻算法(ann)的强大功能,减少了签名构建时间。实验结果将应用于广泛的数据集(UKB, Wang, Corel 10K)和不同的描述符(CMI, SURF)。
{"title":"A machine learning approach for image retrieval tasks","authors":"Achref Ouni","doi":"10.1109/IVCNZ51579.2020.9290617","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290617","url":null,"abstract":"Several methods based on visual methods (BoVW, VLAD,…) or recent deep leaning methods try to solve the CBIR problem. Bag of visual words (BoVW) is one of most module used for both classification and image recognition. But, even with the high performance of BoVW, the problem of retrieving the image by content is still a challenge in computer vision. In this paper, we propose an improvement on a bag of visual words by increasing the accuracy of the retrieved candidates. In addition, we reduce the signature construction time by exploiting the powerful of the approximate nearest neighbor algorithms (ANNs). Experimental results will be applied to widely data sets (UKB, Wang, Corel 10K) and with different descriptors (CMI, SURF).","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127598139","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Development of a Virtual Environment Based Image Generation Tool for Neural Network Training 基于虚拟环境的神经网络训练图像生成工具的开发
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290491
R. Arenas, P. Delmas, Alfonso Gastelum-Strozzi
We present a computational tool to generate visual and descriptive data used as additional training images for neural networks involved in image recognition tasks. The work is inspired by the problem posed to acquire enough data, in order to train service robots, with the goal of improving the range of objects in the environment with which they can interact. The tool provides a framework that allows users to easily setup different environments with the visual information needed for the training, accordingly to their needs. The tool was developed with the Unity engine, and it was designed to be able to import external prefabs. These models are standardized and catalogued into lists, which are accessed to create more complex and diverse virtual environments. Another component of the tool adds an additional layer of complexity by creating randomized environments with different conditions (scale, position and orientation of objects, and environmental illumination). The performance of the created dataset was tested by training the information on the YOLO-V3 (You Only Look Once) architecture and testing on both artificial and real images.
我们提出了一种计算工具,用于生成视觉和描述性数据,作为涉及图像识别任务的神经网络的额外训练图像。这项工作的灵感来自于一个问题,即获取足够的数据,以训练服务机器人,目标是提高环境中物体的范围,使它们能够与之互动。该工具提供了一个框架,允许用户根据自己的需要轻松地设置不同的环境,并提供培训所需的视觉信息。该工具是使用Unity引擎开发的,它被设计成能够导入外部预制件。这些模型被标准化并编目到列表中,可以访问这些列表来创建更复杂和多样化的虚拟环境。该工具的另一个组件通过创建具有不同条件(对象的规模、位置和方向以及环境照明)的随机环境增加了额外的复杂性层。通过在YOLO-V3 (You Only Look Once)架构上训练信息,并在人工图像和真实图像上进行测试,测试了所创建数据集的性能。
{"title":"Development of a Virtual Environment Based Image Generation Tool for Neural Network Training","authors":"R. Arenas, P. Delmas, Alfonso Gastelum-Strozzi","doi":"10.1109/IVCNZ51579.2020.9290491","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290491","url":null,"abstract":"We present a computational tool to generate visual and descriptive data used as additional training images for neural networks involved in image recognition tasks. The work is inspired by the problem posed to acquire enough data, in order to train service robots, with the goal of improving the range of objects in the environment with which they can interact. The tool provides a framework that allows users to easily setup different environments with the visual information needed for the training, accordingly to their needs. The tool was developed with the Unity engine, and it was designed to be able to import external prefabs. These models are standardized and catalogued into lists, which are accessed to create more complex and diverse virtual environments. Another component of the tool adds an additional layer of complexity by creating randomized environments with different conditions (scale, position and orientation of objects, and environmental illumination). The performance of the created dataset was tested by training the information on the YOLO-V3 (You Only Look Once) architecture and testing on both artificial and real images.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"118 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133913845","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Image and Text fusion for UPMC Food-101 using BERT and CNNs 基于BERT和cnn的UPMC Food-101图像和文本融合
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290622
I. Gallo, Gianmarco Ria, Nicola Landro, Riccardo La Grassa
The modern digital world is becoming more and more multimodal. Looking on the internet, images are often associated with the text, so classification problems with these two modalities are very common. In this paper, we examine multimodal classification using textual information and visual representations of the same concept. We investigate two main basic methods to perform multimodal fusion and adapt them with stacking techniques to better handle this type of problem. Here, we use UPMC Food-101, which is a difficult and noisy multimodal dataset that well represents this category of multimodal problems. Our results show that the proposed early fusion technique combined with a stacking-based approach exceeds the state of the art on the dataset used.
现代数字世界正变得越来越多模式。在互联网上,图像经常与文本联系在一起,因此这两种模式的分类问题非常普遍。在本文中,我们使用同一概念的文本信息和视觉表示来研究多模态分类。我们研究了两种主要的多模态融合的基本方法,并将它们与叠加技术相结合,以更好地处理这类问题。在这里,我们使用UPMC Food-101,这是一个困难和有噪声的多模态数据集,很好地代表了这类多模态问题。我们的结果表明,所提出的早期融合技术与基于堆栈的方法相结合,在所使用的数据集上超过了目前的水平。
{"title":"Image and Text fusion for UPMC Food-101 using BERT and CNNs","authors":"I. Gallo, Gianmarco Ria, Nicola Landro, Riccardo La Grassa","doi":"10.1109/IVCNZ51579.2020.9290622","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290622","url":null,"abstract":"The modern digital world is becoming more and more multimodal. Looking on the internet, images are often associated with the text, so classification problems with these two modalities are very common. In this paper, we examine multimodal classification using textual information and visual representations of the same concept. We investigate two main basic methods to perform multimodal fusion and adapt them with stacking techniques to better handle this type of problem. Here, we use UPMC Food-101, which is a difficult and noisy multimodal dataset that well represents this category of multimodal problems. Our results show that the proposed early fusion technique combined with a stacking-based approach exceeds the state of the art on the dataset used.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114242894","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Melanoma and Nevi Classification using Convolution Neural Networks 使用卷积神经网络进行黑色素瘤和痣分类
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290736
R. Grove, R. Green
Early identification of melanoma skin cancer is vital for the improvement of patients’ prospects of five year disease free survival. The majority of malignant skin lesions present at a general practice level where a diagnosis is based on a clinical decision algorithm. As a false negative diagnosis is an unacceptable outcome, clinical caution tends to result in a low positive predictive value of as low at 8%. There has been a large burden of surgical excisions that retrospectively prove to have been unnecessary.This paper proposes a method to identify melanomas in dermoscopic images using a convolution neural network (CNN). The proposed method implements transfer learning based on the ResNet50 CNN, pretrained using the ImageNet dataset. Datasets from the ISIC Archive were implemented during training, validation and testing. Further tests were performed on a smaller dataset of images taken from the Dermnet NZ website and from recent clinical cases still awaiting histological results to indicate the trained network’s ability to generalise to real cases. The 86% test accuracy achieved with the proposed method was comparable to the results of prior studies but required significantly less pre-processing actions to classify a lesion and was not dependant on consistent image scaling or the presence of a scale on the image. This method also improved on past research by making use of all of the information present in an image as opposed to focusing on geometric and colour-space based aspects independently.
黑色素瘤皮肤癌的早期识别对于改善患者五年无病生存的前景至关重要。大多数恶性皮肤病变存在于一般实践水平,其中诊断是基于临床决策算法。由于假阴性诊断是不可接受的结果,临床谨慎倾向于导致低阳性预测值,低至8%。手术切除的负担很大,事后证明是不必要的。本文提出了一种使用卷积神经网络(CNN)识别皮肤镜图像中的黑色素瘤的方法。该方法基于ResNet50 CNN实现迁移学习,使用ImageNet数据集进行预训练。来自ISIC存档的数据集在培训、验证和测试期间被实施。进一步的测试是在一个较小的图像数据集上进行的,这些数据集来自新西兰Dermnet网站和最近的临床病例,这些病例仍在等待组织学结果,以表明训练后的网络能够推广到真实病例。该方法达到了86%的测试准确度,与之前的研究结果相当,但对病变进行分类所需的预处理操作明显减少,并且不依赖于一致的图像缩放或图像上存在的缩放。这种方法也改进了过去的研究,它利用了图像中存在的所有信息,而不是单独关注基于几何和色彩空间的方面。
{"title":"Melanoma and Nevi Classification using Convolution Neural Networks","authors":"R. Grove, R. Green","doi":"10.1109/IVCNZ51579.2020.9290736","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290736","url":null,"abstract":"Early identification of melanoma skin cancer is vital for the improvement of patients’ prospects of five year disease free survival. The majority of malignant skin lesions present at a general practice level where a diagnosis is based on a clinical decision algorithm. As a false negative diagnosis is an unacceptable outcome, clinical caution tends to result in a low positive predictive value of as low at 8%. There has been a large burden of surgical excisions that retrospectively prove to have been unnecessary.This paper proposes a method to identify melanomas in dermoscopic images using a convolution neural network (CNN). The proposed method implements transfer learning based on the ResNet50 CNN, pretrained using the ImageNet dataset. Datasets from the ISIC Archive were implemented during training, validation and testing. Further tests were performed on a smaller dataset of images taken from the Dermnet NZ website and from recent clinical cases still awaiting histological results to indicate the trained network’s ability to generalise to real cases. The 86% test accuracy achieved with the proposed method was comparable to the results of prior studies but required significantly less pre-processing actions to classify a lesion and was not dependant on consistent image scaling or the presence of a scale on the image. This method also improved on past research by making use of all of the information present in an image as opposed to focusing on geometric and colour-space based aspects independently.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116126667","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Heating Patterns Recognition in Industrial Microwave-Processed Foods 工业微波加工食品的加热模式识别
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290639
Sowmya Kasturi, S. L. Moan, D. Bailey, Jeremy Smith
Recognition or identification of hot and cold spot heating patterns in microwave-processed pre-packaged food products is crucial to determine experimental repeatability and design better and safer food treatment systems. This review focuses on computer vision-based methods for heating patterns recognition from the literature along with their limitations. A preliminary kinetics study to correlate colour to varied timetemperature combinations is also discussed.
识别或识别微波加工预包装食品中的热点和冷点加热模式对于确定实验可重复性和设计更好和更安全的食品处理系统至关重要。本文综述了基于计算机视觉的热模式识别方法及其局限性。初步的动力学研究,以关联颜色变化的时间-温度组合也进行了讨论。
{"title":"Heating Patterns Recognition in Industrial Microwave-Processed Foods","authors":"Sowmya Kasturi, S. L. Moan, D. Bailey, Jeremy Smith","doi":"10.1109/IVCNZ51579.2020.9290639","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290639","url":null,"abstract":"Recognition or identification of hot and cold spot heating patterns in microwave-processed pre-packaged food products is crucial to determine experimental repeatability and design better and safer food treatment systems. This review focuses on computer vision-based methods for heating patterns recognition from the literature along with their limitations. A preliminary kinetics study to correlate colour to varied timetemperature combinations is also discussed.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123411317","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Isotropic Remeshing by Dynamic Voronoi Tessellation on Voxelized Surface 体素化表面上动态Voronoi镶嵌的各向同性网格重划分
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290614
Ashutosh Soni, Partha Bhowmick
A novel algorithm for isotropic remeshing of a triangle mesh is presented in this paper. The algorithm is designed to work on a voxelized surface and integrates several novel ideas. One such is the notion of functional partitioning that aids in uniform distribution of seeds for initializing the process of dynamic Voronoi tessellation (DVT). The concept of DVT is also novel and found to be quite effective for iteratively transforming the input mesh into an isotropic mesh while keeping the tessellation aligned with the surface geometry. In each iteration, a Voronoi energy field is used to rearrange the seeds and to recreate the DVT. Over successive iterations, the DVT is found to keep on improving the mesh isotropy without compromising with the surface features. The Delaunay triangles corresponding to the final tessellation are further subdivided in high-curvature regions. The resultant mesh is finally projected back onto the original mesh in order to minimize the Hausdorff error. As our algorithm works in voxel space, it is readily implementable in GPU. Experimental results on various datasets demonstrate its efficiency and robustness.
提出了一种新的三角形网格各向同性重划分算法。该算法被设计用于体素化表面,并集成了一些新颖的思想。其中一个是功能划分的概念,它有助于种子的均匀分布,以初始化动态Voronoi镶嵌(DVT)过程。DVT的概念也是新颖的,并且被发现对于迭代地将输入网格转换为各向同性网格同时保持镶嵌与表面几何形状对齐是非常有效的。在每次迭代中,Voronoi能量场被用来重新排列种子并重建DVT。在连续迭代中,发现DVT在不影响表面特征的情况下不断改善网格各向同性。与最终镶嵌相对应的德劳内三角形在高曲率区域进一步细分。最后将生成的网格投影回原始网格,以最小化Hausdorff误差。由于我们的算法在体素空间中工作,因此很容易在GPU上实现。在不同数据集上的实验结果证明了该方法的有效性和鲁棒性。
{"title":"Isotropic Remeshing by Dynamic Voronoi Tessellation on Voxelized Surface","authors":"Ashutosh Soni, Partha Bhowmick","doi":"10.1109/IVCNZ51579.2020.9290614","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290614","url":null,"abstract":"A novel algorithm for isotropic remeshing of a triangle mesh is presented in this paper. The algorithm is designed to work on a voxelized surface and integrates several novel ideas. One such is the notion of functional partitioning that aids in uniform distribution of seeds for initializing the process of dynamic Voronoi tessellation (DVT). The concept of DVT is also novel and found to be quite effective for iteratively transforming the input mesh into an isotropic mesh while keeping the tessellation aligned with the surface geometry. In each iteration, a Voronoi energy field is used to rearrange the seeds and to recreate the DVT. Over successive iterations, the DVT is found to keep on improving the mesh isotropy without compromising with the surface features. The Delaunay triangles corresponding to the final tessellation are further subdivided in high-curvature regions. The resultant mesh is finally projected back onto the original mesh in order to minimize the Hausdorff error. As our algorithm works in voxel space, it is readily implementable in GPU. Experimental results on various datasets demonstrate its efficiency and robustness.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121795086","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Wavefront reconstruction with the cone sensor 锥形传感器的波前重建
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290735
R. Clare, B. Engler, S. Weddell
Wavefronts of light from celestial objects are aberrated by Earth’s evolving atmosphere, causing images captured by ground-based telescopes to be distorted. The slope of the phase of the wavefront can be estimated by a pyramid wavefront sensor, which subdivides the complex field at the focal plane of the telescope, producing four images of the aperture. The cone wavefront sensor is the extension of the pyramid sensor to having an infinite number of sides, and produces an annulus of intensity rather than four images. We propose and compare the following methods for reconstructing the wavefront from the intensity measurements from the cone sensor: (1) use the entire aperture image, (2) use the pixels inside the intensity annulus only, (3) create a map of slopes by subtracting the slice of annulus 180 degrees opposite, (4) create x and y slopes by cutting out pseudo-apertures around the annulus, and (5) use the inverse Radon transform of the intensity annulus converted to polar co-ordinates. We find via numerical simulation with atmospheric phase screens that methods (1) and (2) provide the best wavefront estimate, methods (3) and (4) the smallest interaction matrices, while method (5) allows direct reconstruction without an interaction matrix.
来自天体的光的波前受到地球不断变化的大气层的干扰,导致地面望远镜捕捉到的图像失真。利用金字塔波前传感器对望远镜焦平面处的复杂场进行细分,得到四幅孔径图像,从而估算出波前相位的斜率。锥波前传感器是金字塔传感器的延伸,具有无限数量的边,并产生一个环的强度,而不是四个图像。我们提出并比较了以下方法来从锥传感器的强度测量中重建波前:(1)使用整个孔径图像,(2)仅使用强度环内的像素,(3)通过减去180度相对的环面切片来创建斜率图,(4)通过切割环周围的伪孔径来创建x和y斜率,以及(5)使用强度环的反Radon变换转换为极坐标。通过大气相屏的数值模拟,我们发现方法(1)和(2)提供了最佳的波前估计,方法(3)和(4)提供了最小的相互作用矩阵,而方法(5)允许在没有相互作用矩阵的情况下直接重建。
{"title":"Wavefront reconstruction with the cone sensor","authors":"R. Clare, B. Engler, S. Weddell","doi":"10.1109/IVCNZ51579.2020.9290735","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290735","url":null,"abstract":"Wavefronts of light from celestial objects are aberrated by Earth’s evolving atmosphere, causing images captured by ground-based telescopes to be distorted. The slope of the phase of the wavefront can be estimated by a pyramid wavefront sensor, which subdivides the complex field at the focal plane of the telescope, producing four images of the aperture. The cone wavefront sensor is the extension of the pyramid sensor to having an infinite number of sides, and produces an annulus of intensity rather than four images. We propose and compare the following methods for reconstructing the wavefront from the intensity measurements from the cone sensor: (1) use the entire aperture image, (2) use the pixels inside the intensity annulus only, (3) create a map of slopes by subtracting the slice of annulus 180 degrees opposite, (4) create x and y slopes by cutting out pseudo-apertures around the annulus, and (5) use the inverse Radon transform of the intensity annulus converted to polar co-ordinates. We find via numerical simulation with atmospheric phase screens that methods (1) and (2) provide the best wavefront estimate, methods (3) and (4) the smallest interaction matrices, while method (5) allows direct reconstruction without an interaction matrix.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126263357","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
PProCRC: Probabilistic Collaboration of Image Patches for Fine-grained Classification PProCRC:用于细粒度分类的图像补丁的概率协作
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290537
Tapabrata (Rohan) Chakraborty, B. McCane, S. Mills, U. Pal
We present a conditional probabilistic framework for collaborative representation of image patches. It incorporates background compensation and outlier patch suppression into the main formulation itself, thus doing away with the need for pre-processing steps to handle the same. A closed form non-iterative solution of the cost function is derived. The proposed method (PProCRC) outperforms earlier CRC formulations: patch based (PCRC, GP-CRC) as well as the state-of-the-art probabilistic (ProCRC and EProCRC) on three fine-grained species recognition datasets (Oxford Flowers, Oxford-IIIT Pets and CUB Birds) using two CNN backbones (Vgg-19 and ResNet-50).
我们提出了一个条件概率框架,用于图像补丁的协同表示。它将背景补偿和离群斑抑制纳入主配方本身,从而消除了处理相同的预处理步骤的需要。导出了代价函数的闭形式非迭代解。所提出的方法(PProCRC)优于早期的CRC方案:基于补丁的(PCRC, GP-CRC)以及最先进的概率(ProCRC和EProCRC)在三个细粒度物种识别数据集(Oxford Flowers, Oxford- iiit Pets和CUB Birds)上使用两个CNN主干(Vgg-19和ResNet-50)。
{"title":"PProCRC: Probabilistic Collaboration of Image Patches for Fine-grained Classification","authors":"Tapabrata (Rohan) Chakraborty, B. McCane, S. Mills, U. Pal","doi":"10.1109/IVCNZ51579.2020.9290537","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290537","url":null,"abstract":"We present a conditional probabilistic framework for collaborative representation of image patches. It incorporates background compensation and outlier patch suppression into the main formulation itself, thus doing away with the need for pre-processing steps to handle the same. A closed form non-iterative solution of the cost function is derived. The proposed method (PProCRC) outperforms earlier CRC formulations: patch based (PCRC, GP-CRC) as well as the state-of-the-art probabilistic (ProCRC and EProCRC) on three fine-grained species recognition datasets (Oxford Flowers, Oxford-IIIT Pets and CUB Birds) using two CNN backbones (Vgg-19 and ResNet-50).","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131484430","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1