首页 > 最新文献

2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)最新文献

英文 中文
A Multi-View Stereo Evaluation for Fine Object Reconstruction 精细物体重建的多视点立体评价
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290742
C. Peat, O. Batchelor, R. Green
Current stereo matching methods based on end to end learning frameworks have shown strong results in the field of depth estimation, bringing significant improvements in robustness as well as flexibility in the accuracy and evaluation time trade-off. In this line of research we observe that the two sub-fields of binocular, and multi-view stereo have converged and are based on fundamentally the same architectures. In this work we aim to perform an objective comparison of these methods, controlling for architecture, and accounting for the rectification process typically used in binocular stereo. To our knowledge there is no prior work directly comparing the two. We aim to measure the performance of matching between rectified pairs, and plane-sweep based multi-view stereo. We test a range of camera configurations and studying the effectiveness of additional cameras in the context of a synthetic multi-view stereo dataset developed for evaluating 3D reconstruction in agriculture.
目前基于端到端学习框架的立体匹配方法在深度估计领域已经取得了较好的效果,在鲁棒性和准确性以及评估时间权衡方面都有了显著的提高。在这条研究路线中,我们观察到双目和多视角立体的两个子领域已经融合,并且基于基本相同的架构。在这项工作中,我们的目标是对这些方法进行客观比较,控制结构,并考虑通常用于双目立体的校正过程。据我们所知,之前没有直接比较这两者的工作。我们的目标是测量校正对之间的匹配性能,以及基于平面扫描的多视图立体。我们测试了一系列摄像机配置,并在为评估农业中的3D重建而开发的合成多视图立体数据集的背景下研究了其他摄像机的有效性。
{"title":"A Multi-View Stereo Evaluation for Fine Object Reconstruction","authors":"C. Peat, O. Batchelor, R. Green","doi":"10.1109/IVCNZ51579.2020.9290742","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290742","url":null,"abstract":"Current stereo matching methods based on end to end learning frameworks have shown strong results in the field of depth estimation, bringing significant improvements in robustness as well as flexibility in the accuracy and evaluation time trade-off. In this line of research we observe that the two sub-fields of binocular, and multi-view stereo have converged and are based on fundamentally the same architectures. In this work we aim to perform an objective comparison of these methods, controlling for architecture, and accounting for the rectification process typically used in binocular stereo. To our knowledge there is no prior work directly comparing the two. We aim to measure the performance of matching between rectified pairs, and plane-sweep based multi-view stereo. We test a range of camera configurations and studying the effectiveness of additional cameras in the context of a synthetic multi-view stereo dataset developed for evaluating 3D reconstruction in agriculture.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"93 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121823244","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Progress towards imaging biological filaments using X-ray free-electron lasers 利用x射线自由电子激光器成像生物细丝的研究进展
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290623
R. D. Arnal, David H. Wojtas, R. Millane
X-ray free-electron lasers (XFELs) are opening new frontiers in structural biology. The extreme brilliance of these highly coherent X-ray sources allows for ever smaller crystals to be used while still being able to diffract enough photons to provide sufficient data for structure determination. Biomolecules arranged into filaments are an important class of targets that are expected to greatly benefit from the continuous improvements in XFEL capabilities. Here we first review some of the state-of the-art research in using XFELs for the imaging of biological filaments. Extrapolating current trends towards single particle imaging, we consider an intermediate case where diffraction patterns from single filaments can be measured and oriented to form a 3D dataset. Prospects for using iterative projection algorithms (IPAs) for ab initio phase retrieval with such data collected from single filaments are illustrated by the reconstruction of the electron density of a B-DNA structure from simulated, noisy XFEL data.
x射线自由电子激光器(XFELs)为结构生物学开辟了新的领域。这些高相干x射线源的极端亮度允许使用更小的晶体,同时仍然能够衍射足够的光子,为结构确定提供足够的数据。排列成细丝的生物分子是一类重要的靶标,有望从XFEL能力的不断改进中受益匪浅。在这里,我们首先回顾了一些使用XFELs进行生物细丝成像的最新研究。外推单粒子成像的当前趋势,我们考虑了一种中间情况,其中单细丝的衍射图案可以测量和定向形成3D数据集。利用模拟的、有噪声的XFEL数据重建B-DNA结构的电子密度,说明了利用迭代投影算法(IPAs)对从单丝收集的数据进行从头算相位检索的前景。
{"title":"Progress towards imaging biological filaments using X-ray free-electron lasers","authors":"R. D. Arnal, David H. Wojtas, R. Millane","doi":"10.1109/IVCNZ51579.2020.9290623","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290623","url":null,"abstract":"X-ray free-electron lasers (XFELs) are opening new frontiers in structural biology. The extreme brilliance of these highly coherent X-ray sources allows for ever smaller crystals to be used while still being able to diffract enough photons to provide sufficient data for structure determination. Biomolecules arranged into filaments are an important class of targets that are expected to greatly benefit from the continuous improvements in XFEL capabilities. Here we first review some of the state-of the-art research in using XFELs for the imaging of biological filaments. Extrapolating current trends towards single particle imaging, we consider an intermediate case where diffraction patterns from single filaments can be measured and oriented to form a 3D dataset. Prospects for using iterative projection algorithms (IPAs) for ab initio phase retrieval with such data collected from single filaments are illustrated by the reconstruction of the electron density of a B-DNA structure from simulated, noisy XFEL data.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123641913","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Sheep: kinship assignment in livestock from facial images 深羊:家畜面部图像的亲属关系分配
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290558
Lech Szymanski, Michael Lee
For the non-farmer folk all sheep might look the same, but they are in fact morphologically quite different; including when it comes to facial features. Image analysis has already demonstrated that computer-based facial recognition in livestock is very accurate. We investigate the viability of deep learning for assigning kinship in livestock for use in genetic evaluation- given two images of sheep faces, our proposed model predicts their genetic relationship. In this work we present two CNN models: one for face detection (reporting 80% accuracy) and one for kinship detection (reporting 68% balanced accuracy).
对于不务农的人来说,所有的羊可能看起来都一样,但实际上它们在形态上有很大的不同;包括面部特征。图像分析已经证明,基于计算机的牲畜面部识别是非常准确的。我们研究了深度学习在遗传评估中分配牲畜亲属关系的可行性-给定两张羊脸图像,我们提出的模型预测它们的遗传关系。在这项工作中,我们提出了两个CNN模型:一个用于人脸检测(报告80%的准确率),另一个用于亲属关系检测(报告68%的平衡准确率)。
{"title":"Deep Sheep: kinship assignment in livestock from facial images","authors":"Lech Szymanski, Michael Lee","doi":"10.1109/IVCNZ51579.2020.9290558","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290558","url":null,"abstract":"For the non-farmer folk all sheep might look the same, but they are in fact morphologically quite different; including when it comes to facial features. Image analysis has already demonstrated that computer-based facial recognition in livestock is very accurate. We investigate the viability of deep learning for assigning kinship in livestock for use in genetic evaluation- given two images of sheep faces, our proposed model predicts their genetic relationship. In this work we present two CNN models: one for face detection (reporting 80% accuracy) and one for kinship detection (reporting 68% balanced accuracy).","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125350026","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Edge-Aware Convolution for RGB-D Image Segmentation 边缘感知卷积在RGB-D图像分割中的应用
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290608
Rongsen Chen, Fang-Lue Zhang, Taehyun Rhee
Convolutional Neural Networks using RGB-D images as input have shown superior performance in recent research in the field of semantic segmentation. In RGB-D data, the depth channel encodes information from the 3D spatial domain, which has an inherent difference with the color channels. It thus needs to be treated in a special way, rather than just processed as another channel of the input signal. Under this purpose, we propose a simple but not trivial edge-aware convolutional kernel to utilize the geometric information contained in the depth channel to extract feature maps in a more effective manner. The edge-aware convolutional kernel is built upon regular convolutional kernel, thus, it can be used to restructure existing CNN models to achieve stable and effective feature extraction for RGB-D data. We compare our result with a previous method that is closely related to our to show our method can provide more effective and stable feature extraction.
以RGB-D图像为输入的卷积神经网络在语义分割领域的研究中表现出优异的性能。在RGB-D数据中,深度通道编码来自三维空间域的信息,这与颜色通道具有固有的区别。因此,它需要以特殊的方式处理,而不是仅仅作为输入信号的另一个通道处理。在此目的下,我们提出了一种简单但不平凡的边缘感知卷积核,利用深度通道中包含的几何信息更有效地提取特征映射。边缘感知卷积核是在正则卷积核的基础上构建的,因此,它可以用来重构现有的CNN模型,以实现对RGB-D数据稳定有效的特征提取。我们将结果与之前与我们的方法密切相关的方法进行了比较,表明我们的方法可以提供更有效和稳定的特征提取。
{"title":"Edge-Aware Convolution for RGB-D Image Segmentation","authors":"Rongsen Chen, Fang-Lue Zhang, Taehyun Rhee","doi":"10.1109/IVCNZ51579.2020.9290608","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290608","url":null,"abstract":"Convolutional Neural Networks using RGB-D images as input have shown superior performance in recent research in the field of semantic segmentation. In RGB-D data, the depth channel encodes information from the 3D spatial domain, which has an inherent difference with the color channels. It thus needs to be treated in a special way, rather than just processed as another channel of the input signal. Under this purpose, we propose a simple but not trivial edge-aware convolutional kernel to utilize the geometric information contained in the depth channel to extract feature maps in a more effective manner. The edge-aware convolutional kernel is built upon regular convolutional kernel, thus, it can be used to restructure existing CNN models to achieve stable and effective feature extraction for RGB-D data. We compare our result with a previous method that is closely related to our to show our method can provide more effective and stable feature extraction.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124715600","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Voice Interaction for Augmented Reality Navigation Interfaces with Natural Language Understanding 基于自然语言理解的增强现实导航界面语音交互
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290643
Junhong Zhao, Christopher James Parry, R. K. D. Anjos, C. Anslow, Taehyun Rhee
Voice interaction with natural language understanding (NLU) has been extensively explored in desktop computers, handheld devices, and human-robot interaction. However, there is limited research into voice interaction with NLU in augmented reality (AR). There are benefits of using voice interaction in AR, such as high naturalness and being hands-free. In this project, we introduce VOARLA, an NLU-powered AR voice interface, which navigate courier driver delivery a package. A user study was completed to evaluate VOARLA against an AR voice interface without NLU to investigate the effectiveness of NLU in the navigation interface in AR. We evaluated from three aspects: accuracy, productivity, and commands learning curve. Results found that using NLU in AR increases the accuracy of the interface by 15%. However, higher accuracy did not correlate to an increase in productivity. Results suggest that NLU helped users remember the commands on the first run when they were unfamiliar with the system. This suggests that using NLU in an AR hands-free application can make the learning curve easier for new users.
语音交互与自然语言理解(NLU)在台式计算机、手持设备和人机交互中得到了广泛的探索。然而,在增强现实(AR)中,语音与非语言推理的交互研究有限。在AR中使用语音交互有很多好处,比如高自然度和免提。在这个项目中,我们介绍了VOARLA,一个nlu驱动的AR语音接口,它可以引导快递司机递送包裹。我们完成了一项用户研究,将VOARLA与没有NLU的AR语音界面进行比较,以调查NLU在AR导航界面中的有效性。我们从三个方面进行了评估:准确性、生产力和命令学习曲线。结果发现,在AR中使用NLU可使界面的准确性提高15%。然而,更高的准确度与生产率的提高并不相关。结果表明,当用户对系统不熟悉时,NLU可以帮助他们在第一次运行时记住命令。这表明在AR免提应用程序中使用NLU可以使新用户的学习曲线更容易。
{"title":"Voice Interaction for Augmented Reality Navigation Interfaces with Natural Language Understanding","authors":"Junhong Zhao, Christopher James Parry, R. K. D. Anjos, C. Anslow, Taehyun Rhee","doi":"10.1109/IVCNZ51579.2020.9290643","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290643","url":null,"abstract":"Voice interaction with natural language understanding (NLU) has been extensively explored in desktop computers, handheld devices, and human-robot interaction. However, there is limited research into voice interaction with NLU in augmented reality (AR). There are benefits of using voice interaction in AR, such as high naturalness and being hands-free. In this project, we introduce VOARLA, an NLU-powered AR voice interface, which navigate courier driver delivery a package. A user study was completed to evaluate VOARLA against an AR voice interface without NLU to investigate the effectiveness of NLU in the navigation interface in AR. We evaluated from three aspects: accuracy, productivity, and commands learning curve. Results found that using NLU in AR increases the accuracy of the interface by 15%. However, higher accuracy did not correlate to an increase in productivity. Results suggest that NLU helped users remember the commands on the first run when they were unfamiliar with the system. This suggests that using NLU in an AR hands-free application can make the learning curve easier for new users.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121386320","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A fair comparison of the EEG signal classification methods for alcoholic subject identification 酒精受试者识别的脑电信号分类方法的比较
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290683
M. Awrangjeb, J. D. C. Rodrigues, Bela Stantic, V. Estivill-Castro
The electroencephalogram (EEG) signal, which records the electrical activity in the brain, is useful for assessing the mental state of the alcoholic subject. Since the public release of an EEG dataset by the University of California, Irvine, there have been many attempts to classify the EEG signals of alcoholic’ and ‘healthy’ subjects. These classification methods are hard to compare as they use different subsets of the dataset and many of their algorithmic settings are unknown. The comparison of their published results using the inconsistent and unknown information is unfair. This paper attempts a fair comparison by presenting a level playing field where a public subset of the dataset is employed with known algorithmic settings. Two recently proposed high performing EEG signal classification methods are implemented with different classifiers and cross-validation techniques. While compared it is observed that the wavelet packet decomposition method with the Naïve Bayes classifier and the k-fold cross validation technique outperforms the other method.
脑电图(EEG)信号记录了大脑中的电活动,对评估酗酒者的精神状态很有用。自从加州大学欧文分校(University of California, Irvine)公开发布脑电图数据集以来,已经有很多人尝试对“酗酒者”和“健康者”的脑电图信号进行分类。这些分类方法很难比较,因为它们使用的是数据集的不同子集,而且它们的许多算法设置是未知的。使用不一致和未知的信息来比较他们发表的结果是不公平的。本文试图通过提供一个公平的竞争环境来进行公平的比较,其中数据集的公共子集与已知的算法设置一起使用。采用不同的分类器和交叉验证技术实现了两种最新提出的高效脑电信号分类方法。通过与Naïve贝叶斯分类器和k-fold交叉验证技术进行比较,发现小波包分解方法优于其他方法。
{"title":"A fair comparison of the EEG signal classification methods for alcoholic subject identification","authors":"M. Awrangjeb, J. D. C. Rodrigues, Bela Stantic, V. Estivill-Castro","doi":"10.1109/IVCNZ51579.2020.9290683","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290683","url":null,"abstract":"The electroencephalogram (EEG) signal, which records the electrical activity in the brain, is useful for assessing the mental state of the alcoholic subject. Since the public release of an EEG dataset by the University of California, Irvine, there have been many attempts to classify the EEG signals of alcoholic’ and ‘healthy’ subjects. These classification methods are hard to compare as they use different subsets of the dataset and many of their algorithmic settings are unknown. The comparison of their published results using the inconsistent and unknown information is unfair. This paper attempts a fair comparison by presenting a level playing field where a public subset of the dataset is employed with known algorithmic settings. Two recently proposed high performing EEG signal classification methods are implemented with different classifiers and cross-validation techniques. While compared it is observed that the wavelet packet decomposition method with the Naïve Bayes classifier and the k-fold cross validation technique outperforms the other method.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115951593","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Incorporating Human Body Shape Guidance for Cloth Warping in Model to Person Virtual Try-on Problems 将模型中布料翘曲的人体形状指导结合到人的虚拟试穿问题中
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290603
Debapriya Roy, Sanchayan Santra, B. Chanda
The world of retail has witnessed a lot of change in the last few decades and with a size of 2.4 trillion, the fashion industry is way ahead of others in this aspect. With the blessings of technology like virtual try-on (vton), now even online shoppers can virtually try a product before buying. However, the current image-based virtual try-on methods still have a long way to go when it comes to producing realistic outputs. In general, vton methods work in two stages. The first stage warps the source cloth and the second stage merges the cloth with the person image for predicting the final try-on output. While the second stage is comparatively easier to handle using neural networks, predicting an accurate warp is difficult as replicating actual human body deformation is challenging. A fundamental issue in vton domain is data. Although lots of images of cloth are available over the internet in either social media or e-commerce websites, but most of them are in the form of a human wearing the cloth. However, the existing approaches are constrained to take separate cloth images as the input source clothing. To address these problems, we propose a model to person cloth warping strategy, where the objective is to align the segmented cloth from the model image in a way that fits the target person, thus, alleviating the need of separate cloth images. Compared to the existing approaches of warping, our method shows improvement especially in the case of complex patterns of cloth. Rigorous experiments applied on various public domain datasets establish the efficacy of this method compared to benchmark methods.
在过去的几十年里,零售业发生了很多变化,拥有2.4万亿美元规模的时装业在这方面遥遥领先于其他行业。随着虚拟试穿(vton)等技术的发展,现在即使是网上购物者也可以在购买前虚拟试穿一件商品。然而,目前基于图像的虚拟试戴方法在产生逼真输出方面还有很长的路要走。一般来说,vton方法分为两个阶段。第一阶段对布料进行经纱,第二阶段将布料与人物图像合并,以预测最终的试穿输出。虽然第二阶段使用神经网络相对容易处理,但预测准确的翘曲是困难的,因为复制实际的人体变形是具有挑战性的。光子领域的一个基本问题是数据。虽然在社交媒体和电子商务网站上可以看到很多关于布料的图片,但大多数都是人穿着布料的形式。然而,现有的方法都局限于以单独的布料图像作为输入源服装。为了解决这些问题,我们提出了一种模型到人的布料扭曲策略,其目标是将模型图像中的分段布料以适合目标人的方式对齐,从而减少对单独布料图像的需求。与现有的整经方法相比,本方法在织物复杂图案的整经方面有明显的改进。在各种公共领域数据集上进行的严格实验证明了该方法与基准方法相比的有效性。
{"title":"Incorporating Human Body Shape Guidance for Cloth Warping in Model to Person Virtual Try-on Problems","authors":"Debapriya Roy, Sanchayan Santra, B. Chanda","doi":"10.1109/IVCNZ51579.2020.9290603","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290603","url":null,"abstract":"The world of retail has witnessed a lot of change in the last few decades and with a size of 2.4 trillion, the fashion industry is way ahead of others in this aspect. With the blessings of technology like virtual try-on (vton), now even online shoppers can virtually try a product before buying. However, the current image-based virtual try-on methods still have a long way to go when it comes to producing realistic outputs. In general, vton methods work in two stages. The first stage warps the source cloth and the second stage merges the cloth with the person image for predicting the final try-on output. While the second stage is comparatively easier to handle using neural networks, predicting an accurate warp is difficult as replicating actual human body deformation is challenging. A fundamental issue in vton domain is data. Although lots of images of cloth are available over the internet in either social media or e-commerce websites, but most of them are in the form of a human wearing the cloth. However, the existing approaches are constrained to take separate cloth images as the input source clothing. To address these problems, we propose a model to person cloth warping strategy, where the objective is to align the segmented cloth from the model image in a way that fits the target person, thus, alleviating the need of separate cloth images. Compared to the existing approaches of warping, our method shows improvement especially in the case of complex patterns of cloth. Rigorous experiments applied on various public domain datasets establish the efficacy of this method compared to benchmark methods.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"264 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134500754","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A rapid method of hypercube stitching for snapshot multi-camera system 快照多相机系统超立方体拼接的快速方法
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290723
Y. Dixit, M. Al-Sarayreh, C. Craigie, M. M. Reis
Snapshot hyperspectral imaging (HSI) systems are rapid and ultra-compact making them potential candidate for real-time food analysis. However, the existing technology limits the working wavelength range of these cameras requiring multiple cameras to cover a wider spectral range. We present a rapid hypercube stitching method which generates an efficiently stitched hypercube from two different HSI cameras providing a wider spectral range as well as spatial information. It shows reliability and robustness over the manual stitching. The method was able to successfully stitch respective hypercubes from near-infrared (NIR) and visible (Vis) cameras producing much lower number of non-overlapping pixels between the hypercubes then would be possible with manual stitching. We demonstrate the application of our method for stitching the hypercubes (NIR and Vis) for 32 beef samples analyzing the stitching efficiency and reliability of spectral information.
快照高光谱成像(HSI)系统快速和超紧凑,使其成为实时食品分析的潜在候选者。然而,现有技术限制了这些摄像机的工作波长范围,需要多个摄像机覆盖更宽的光谱范围。我们提出了一种快速的超立方体拼接方法,该方法从两个不同的HSI相机生成一个有效的拼接超立方体,提供更宽的光谱范围和空间信息。与手工拼接相比,具有较好的可靠性和鲁棒性。该方法能够成功地从近红外(NIR)和可见光(Vis)相机缝合各自的超立方体,在超立方体之间产生的非重叠像素数量远低于手动拼接。将该方法应用于32个牛肉样品的超立方体(近红外和可见光)拼接,分析了光谱信息的拼接效率和可靠性。
{"title":"A rapid method of hypercube stitching for snapshot multi-camera system","authors":"Y. Dixit, M. Al-Sarayreh, C. Craigie, M. M. Reis","doi":"10.1109/IVCNZ51579.2020.9290723","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290723","url":null,"abstract":"Snapshot hyperspectral imaging (HSI) systems are rapid and ultra-compact making them potential candidate for real-time food analysis. However, the existing technology limits the working wavelength range of these cameras requiring multiple cameras to cover a wider spectral range. We present a rapid hypercube stitching method which generates an efficiently stitched hypercube from two different HSI cameras providing a wider spectral range as well as spatial information. It shows reliability and robustness over the manual stitching. The method was able to successfully stitch respective hypercubes from near-infrared (NIR) and visible (Vis) cameras producing much lower number of non-overlapping pixels between the hypercubes then would be possible with manual stitching. We demonstrate the application of our method for stitching the hypercubes (NIR and Vis) for 32 beef samples analyzing the stitching efficiency and reliability of spectral information.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128646883","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Image Metrics for Deconvolution of Satellites in Low Earth Orbit 近地轨道卫星反卷积图像度量
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290535
Sierra Hickman, Vishnu Anand Muruganandan, S. Weddell, R. Clare
Satellites and space debris clutter low Earth orbital paths, causing concern for future launches as the clutter increases the probability of in-orbit collisions. Therefore, it is important to track and characterise these objects. However, Earth’s atmosphere distorts images collected from ground-based telescopes, which can be reduced through post-processing deconvolution to improve images of satellites and space debris. A metric is needed to quantity the quality of the images and deconvolution of these extended objects at finite distances; as well as to characterise the structure and brightness for un-symmetrical satellites in low Earth orbit. This paper uses images of the International Space Station to investigate the use of the structural similarity metric and the regional properties as potential satellite imaging metrics. Our results show that the similarity metric can characterise the orientation of the satellite relative to the observer, while the regional properties serve to quantity the image quality and improvement due to deconvolution.
卫星和太空碎片干扰近地轨道路径,引起对未来发射的担忧,因为这些干扰增加了在轨碰撞的可能性。因此,跟踪和描述这些物体是很重要的。然而,地球大气层会使地面望远镜收集到的图像失真,这可以通过后处理反褶积来改善卫星和太空碎片的图像,从而减少图像失真。需要一个度量来量化图像的质量和这些扩展对象在有限距离上的反卷积;以及表征低地球轨道上不对称卫星的结构和亮度。本文以国际空间站图像为例,研究了结构相似性度量和区域属性作为潜在的卫星成像度量。我们的研究结果表明,相似度度量可以表征卫星相对于观察者的方向,而区域属性用于量化图像质量和由于反卷积而得到的改进。
{"title":"Image Metrics for Deconvolution of Satellites in Low Earth Orbit","authors":"Sierra Hickman, Vishnu Anand Muruganandan, S. Weddell, R. Clare","doi":"10.1109/IVCNZ51579.2020.9290535","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290535","url":null,"abstract":"Satellites and space debris clutter low Earth orbital paths, causing concern for future launches as the clutter increases the probability of in-orbit collisions. Therefore, it is important to track and characterise these objects. However, Earth’s atmosphere distorts images collected from ground-based telescopes, which can be reduced through post-processing deconvolution to improve images of satellites and space debris. A metric is needed to quantity the quality of the images and deconvolution of these extended objects at finite distances; as well as to characterise the structure and brightness for un-symmetrical satellites in low Earth orbit. This paper uses images of the International Space Station to investigate the use of the structural similarity metric and the regional properties as potential satellite imaging metrics. Our results show that the similarity metric can characterise the orientation of the satellite relative to the observer, while the regional properties serve to quantity the image quality and improvement due to deconvolution.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"77 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128039268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Review of Emerging Video Codecs: Challenges and Opportunities 新兴视频编解码器的回顾:挑战与机遇
Pub Date : 2020-11-25 DOI: 10.1109/IVCNZ51579.2020.9290536
A. Punchihewa, D. Bailey
This paper presents a review of video codecs that are in use and currently being developed, the codec development process, current trends, challenges and opportunities for the research community. There is a paradigm shift in video coding standards. Concurrently, multiple video standards are standardised by standardising organisations. At the same time, royalty free video compression standards are being developed and standardised. Introduction of enhancement-layer-based coding standards will extend the lifetime of legacy video codecs finding middle ground in improved coding efficiency, computational complexity and power requirements. The video coding landscape is changing that is challenged by emergence of multiple video coding standards for different use cases. These may offer some opportunities for coding industry, especially for New Zealand researchers serving niche markets in video games, computer generated videos and animations.
本文介绍了正在使用和正在开发的视频编解码器,编解码器的开发过程,目前的趋势,挑战和研究社区的机遇。视频编码标准正在发生范式转变。同时,多个视频标准由标准化组织进行标准化。与此同时,免版税的视频压缩标准正在开发和标准化。基于增强层的编码标准的引入将延长传统视频编解码器的使用寿命,在提高编码效率、计算复杂性和功率要求方面找到中间位置。由于针对不同用例的多种视频编码标准的出现,视频编码领域正在发生变化。这可能会为编码行业提供一些机会,特别是对新西兰的研究人员服务于视频游戏、计算机生成视频和动画的利基市场。
{"title":"A Review of Emerging Video Codecs: Challenges and Opportunities","authors":"A. Punchihewa, D. Bailey","doi":"10.1109/IVCNZ51579.2020.9290536","DOIUrl":"https://doi.org/10.1109/IVCNZ51579.2020.9290536","url":null,"abstract":"This paper presents a review of video codecs that are in use and currently being developed, the codec development process, current trends, challenges and opportunities for the research community. There is a paradigm shift in video coding standards. Concurrently, multiple video standards are standardised by standardising organisations. At the same time, royalty free video compression standards are being developed and standardised. Introduction of enhancement-layer-based coding standards will extend the lifetime of legacy video codecs finding middle ground in improved coding efficiency, computational complexity and power requirements. The video coding landscape is changing that is challenged by emergence of multiple video coding standards for different use cases. These may offer some opportunities for coding industry, especially for New Zealand researchers serving niche markets in video games, computer generated videos and animations.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"197 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123012928","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1