首页 > 最新文献

IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision最新文献

英文 中文
Knowledge-Augmented Contrastive Learning for Abnormality Classification and Localization in Chest X-rays with Radiomics using a Feedback Loop. 基于反馈环的放射组学在胸片异常分类和定位中的知识增强对比学习。
Pub Date : 2022-01-01 Epub Date: 2022-02-15 DOI: 10.1109/wacv51458.2022.00185
Yan Han, Chongyan Chen, Ahmed Tewfik, Benjamin Glicksberg, Ying Ding, Yifan Peng, Zhangyang Wang

Accurate classification and localization of abnormalities in chest X-rays play an important role in clinical diagnosis and treatment planning. Building a highly accurate predictive model for these tasks usually requires a large number of manually annotated labels and pixel regions (bounding boxes) of abnormalities. However, it is expensive to acquire such annotations, especially the bounding boxes. Recently, contrastive learning has shown strong promise in leveraging unlabeled natural images to produce highly generalizable and discriminative features. However, extending its power to the medical image domain is under-explored and highly non-trivial, since medical images are much less amendable to data augmentations. In contrast, their prior knowledge, as well as radiomic features, is often crucial. To bridge this gap, we propose an end-to-end semi-supervised knowledge-augmented contrastive learning framework, that simultaneously performs disease classification and localization tasks. The key knob of our framework is a unique positive sampling approach tailored for the medical images, by seamlessly integrating radiomic features as a knowledge augmentation. Specifically, we first apply an image encoder to classify the chest X-rays and to generate the image features. We next leverage Grad-CAM to highlight the crucial (abnormal) regions for chest X-rays (even when unannotated), from which we extract radiomic features. The radiomic features are then passed through another dedicated encoder to act as the positive sample for the image features generated from the same chest X-ray. In this way, our framework constitutes a feedback loop for image and radiomic features to mutually reinforce each other. Their contrasting yields knowledge-augmented representations that are both robust and interpretable. Extensive experiments on the NIH Chest X-ray dataset demonstrate that our approach outperforms existing baselines in both classification and localization tasks.

胸部x线异常的准确分类和定位对临床诊断和治疗方案具有重要作用。为这些任务构建一个高度准确的预测模型通常需要大量手工标注的异常标签和像素区域(边界框)。然而,获取这样的注释是昂贵的,尤其是边界框。最近,对比学习在利用未标记的自然图像产生高度概括和判别特征方面显示出强大的前景。然而,将其功能扩展到医学图像领域尚未得到充分探索,而且非常重要,因为医学图像对数据增强的可修改性要小得多。相反,他们的先验知识,以及放射性特征,往往是至关重要的。为了弥补这一差距,我们提出了一个端到端的半监督知识增强对比学习框架,该框架同时执行疾病分类和定位任务。我们的框架的关键旋钮是一个独特的正采样方法量身定制的医学图像,通过无缝集成放射学特征作为知识增强。具体而言,我们首先应用图像编码器对胸部x射线进行分类并生成图像特征。接下来,我们利用Grad-CAM突出胸部x光片的关键(异常)区域(即使没有注释),从中提取放射学特征。然后将放射学特征通过另一个专用编码器,作为同一胸部x射线生成的图像特征的阳性样本。通过这种方式,我们的框架构成了一个反馈循环,图像和放射特征相互增强。它们的对比产生了既健壮又可解释的知识增强表示。在NIH胸部x射线数据集上进行的大量实验表明,我们的方法在分类和定位任务方面都优于现有的基线。
{"title":"Knowledge-Augmented Contrastive Learning for Abnormality Classification and Localization in Chest X-rays with Radiomics using a Feedback Loop.","authors":"Yan Han,&nbsp;Chongyan Chen,&nbsp;Ahmed Tewfik,&nbsp;Benjamin Glicksberg,&nbsp;Ying Ding,&nbsp;Yifan Peng,&nbsp;Zhangyang Wang","doi":"10.1109/wacv51458.2022.00185","DOIUrl":"https://doi.org/10.1109/wacv51458.2022.00185","url":null,"abstract":"<p><p>Accurate classification and localization of abnormalities in chest X-rays play an important role in clinical diagnosis and treatment planning. Building a highly accurate predictive model for these tasks usually requires a large number of manually annotated labels and pixel regions (bounding boxes) of abnormalities. However, it is expensive to acquire such annotations, especially the bounding boxes. Recently, contrastive learning has shown strong promise in leveraging unlabeled natural images to produce highly generalizable and discriminative features. However, extending its power to the medical image domain is under-explored and highly non-trivial, since medical images are much less amendable to data augmentations. In contrast, their prior knowledge, as well as radiomic features, is often crucial. To bridge this gap, we propose an end-to-end semi-supervised knowledge-augmented contrastive learning framework, that simultaneously performs disease classification and localization tasks. The key knob of our framework is a unique positive sampling approach tailored for the medical images, by seamlessly integrating radiomic features as a knowledge augmentation. Specifically, we first apply an image encoder to classify the chest X-rays and to generate the image features. We next leverage Grad-CAM to highlight the crucial (abnormal) regions for chest X-rays (even when unannotated), from which we extract radiomic features. The radiomic features are then passed through another dedicated encoder to act as the positive sample for the image features generated from the same chest X-ray. In this way, our framework constitutes a feedback loop for image and radiomic features to mutually reinforce each other. Their contrasting yields knowledge-augmented representations that are both robust and interpretable. Extensive experiments on the NIH Chest X-ray dataset demonstrate that our approach outperforms existing baselines in both classification and localization tasks.</p>","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":" ","pages":"1789-1798"},"PeriodicalIF":0.0,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9594386/pdf/nihms-1844026.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"40438683","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Self-Supervised Poisson-Gaussian Denoising. 自监督泊松高斯去噪。
Pub Date : 2021-01-01 Epub Date: 2021-06-14 DOI: 10.1109/wacv48630.2021.00218
Wesley Khademi, Sonia Rao, Clare Minnerath, Guy Hagen, Jonathan Ventura

We extend the blindspot model for self-supervised denoising to handle Poisson-Gaussian noise and introduce an improved training scheme that avoids hyperparameters and adapts the denoiser to the test data. Self-supervised models for denoising learn to denoise from only noisy data and do not require corresponding clean images, which are difficult or impossible to acquire in some application areas of interest such as low-light microscopy. We introduce a new training strategy to handle Poisson-Gaussian noise which is the standard noise model for microscope images. Our new strategy eliminates hyperparameters from the loss function, which is important in a self-supervised regime where no ground truth data is available to guide hyperparameter tuning. We show how our denoiser can be adapted to the test data to improve performance. Our evaluations on microscope image denoising benchmarks validate our approach.

我们扩展了用于自我监督去噪的盲点模型,以处理泊松-高斯噪声,并引入了一种改进的训练方案,它避免了超参数,并使去噪器适应测试数据。用于去噪的自监督模型仅从噪声数据中学习去噪,不需要相应的干净图像,而在某些应用领域(如低照度显微镜),很难或不可能获得干净图像。我们引入了一种新的训练策略来处理泊松高斯噪声,这是显微镜图像的标准噪声模型。我们的新策略消除了损失函数中的超参数,这在自我监督机制中非常重要,因为在这种机制中没有地面实况数据来指导超参数的调整。我们展示了如何根据测试数据调整去噪器以提高性能。我们对显微镜图像去噪基准的评估验证了我们的方法。
{"title":"Self-Supervised Poisson-Gaussian Denoising.","authors":"Wesley Khademi, Sonia Rao, Clare Minnerath, Guy Hagen, Jonathan Ventura","doi":"10.1109/wacv48630.2021.00218","DOIUrl":"10.1109/wacv48630.2021.00218","url":null,"abstract":"<p><p>We extend the blindspot model for self-supervised denoising to handle Poisson-Gaussian noise and introduce an improved training scheme that avoids hyperparameters and adapts the denoiser to the test data. Self-supervised models for denoising learn to denoise from only noisy data and do not require corresponding clean images, which are difficult or impossible to acquire in some application areas of interest such as low-light microscopy. We introduce a new training strategy to handle Poisson-Gaussian noise which is the standard noise model for microscope images. Our new strategy eliminates hyperparameters from the loss function, which is important in a self-supervised regime where no ground truth data is available to guide hyperparameter tuning. We show how our denoiser can be adapted to the test data to improve performance. Our evaluations on microscope image denoising benchmarks validate our approach.</p>","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":" ","pages":"2130-2138"},"PeriodicalIF":0.0,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8294668/pdf/nihms-1710528.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39211431","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Representation Learning with Statistical Independence to Mitigate Bias. 具有统计独立性的表征学习,以减少偏差。
Pub Date : 2021-01-01 Epub Date: 2021-06-14 DOI: 10.1109/wacv48630.2021.00256
Ehsan Adeli, Qingyu Zhao, Adolf Pfefferbaum, Edith V Sullivan, Li Fei-Fei, Juan Carlos Niebles, Kilian M Pohl

Presence of bias (in datasets or tasks) is inarguably one of the most critical challenges in machine learning applications that has alluded to pivotal debates in recent years. Such challenges range from spurious associations between variables in medical studies to the bias of race in gender or face recognition systems. Controlling for all types of biases in the dataset curation stage is cumbersome and sometimes impossible. The alternative is to use the available data and build models incorporating fair representation learning. In this paper, we propose such a model based on adversarial training with two competing objectives to learn features that have (1) maximum discriminative power with respect to the task and (2) minimal statistical mean dependence with the protected (bias) variable(s). Our approach does so by incorporating a new adversarial loss function that encourages a vanished correlation between the bias and the learned features. We apply our method to synthetic data, medical images (containing task bias), and a dataset for gender classification (containing dataset bias). Our results show that the learned features by our method not only result in superior prediction performance but also are unbiased.

毫无疑问,(数据集或任务中)存在偏差是机器学习应用中最关键的挑战之一,近年来已引起了举足轻重的争论。这些挑战包括医学研究中变量之间的虚假关联,以及性别或人脸识别系统中的种族偏见。在数据集整理阶段控制所有类型的偏差非常麻烦,有时甚至是不可能的。另一种方法是利用现有数据,建立包含公平表征学习的模型。在本文中,我们提出了这样一种基于对抗训练的模型,它有两个相互竞争的目标,即学习具有以下特征的数据:(1) 与任务相关的最大辨别力;(2) 与受保护(偏差)变量相关的最小统计均值依赖性。我们的方法通过纳入一个新的对抗损失函数来实现这一目标,该函数鼓励在偏差和所学特征之间消除相关性。我们将我们的方法应用于合成数据、医学图像(包含任务偏差)和性别分类数据集(包含数据集偏差)。结果表明,我们的方法所学习到的特征不仅预测性能优越,而且没有偏差。
{"title":"Representation Learning with Statistical Independence to Mitigate Bias.","authors":"Ehsan Adeli, Qingyu Zhao, Adolf Pfefferbaum, Edith V Sullivan, Li Fei-Fei, Juan Carlos Niebles, Kilian M Pohl","doi":"10.1109/wacv48630.2021.00256","DOIUrl":"10.1109/wacv48630.2021.00256","url":null,"abstract":"<p><p>Presence of bias (in datasets or tasks) is inarguably one of the most critical challenges in machine learning applications that has alluded to pivotal debates in recent years. Such challenges range from spurious associations between variables in medical studies to the bias of race in gender or face recognition systems. Controlling for all types of biases in the dataset curation stage is cumbersome and sometimes impossible. The alternative is to use the available data and build models incorporating fair representation learning. In this paper, we propose such a model based on adversarial training with two competing objectives to learn features that have (1) maximum discriminative power with respect to the task and (2) minimal statistical mean dependence with the protected (bias) variable(s). Our approach does so by incorporating a new adversarial loss function that encourages a vanished correlation between the bias and the learned features. We apply our method to synthetic data, medical images (containing task bias), and a dataset for gender classification (containing dataset bias). Our results show that the learned features by our method not only result in superior prediction performance but also are unbiased.</p>","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":"2021 ","pages":"2512-2522"},"PeriodicalIF":0.0,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8436589/pdf/nihms-1648742.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9649997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Subject Guided Eye Image Synthesis with Application to Gaze Redirection. 将受试者引导的眼图合成应用于注视重定向。
Harsimran Kaur, Roberto Manduchi

We propose a method for synthesizing eye images from segmentation masks with a desired style. The style encompasses attributes such as skin color, texture, iris color, and personal identity. Our approach generates an eye image that is consistent with a given segmentation mask and has the attributes of the input style image. We apply our method to data augmentation as well as to gaze redirection. The previous techniques of synthesizing real eye images from synthetic eye images for data augmentation lacked control over the generated attributes. We demonstrate the effectiveness of the proposed method in synthesizing realistic eye images with given characteristics corresponding to the synthetic labels for data augmentation, which is further useful for various tasks such as gaze estimation, eye image segmentation, pupil detection, etc. We also show how our approach can be applied to gaze redirection using only synthetic gaze labels, improving the previous state of the art results. The main contributions of our paper are i) a novel approach for Style-Based eye image generation from segmentation mask; ii) the use of this approach for gaze-redirection without the need for gaze annotated real eye images.

我们提出了一种从具有所需风格的分割蒙版合成眼睛图像的方法。风格包括肤色、纹理、虹膜颜色和个人身份等属性。我们的方法生成的眼部图像与给定的分割蒙版一致,并具有输入风格图像的属性。我们将这种方法应用于数据增强和注视重定向。以前用于数据增强的合成眼球图像合成真实眼球图像的技术缺乏对生成属性的控制。我们展示了所提出的方法在合成具有与合成标签相对应的给定特征的真实眼部图像以用于数据增强方面的有效性,这对于各种任务,如凝视估计、眼部图像分割、瞳孔检测等都非常有用。我们还展示了如何将我们的方法应用于仅使用合成注视标签的注视重定向,从而改进了之前的技术成果。我们论文的主要贡献在于:i) 从分割掩码中生成基于风格的眼部图像的新方法;ii) 使用这种方法进行注视重定向,而无需注视注释的真实眼部图像。
{"title":"Subject Guided Eye Image Synthesis with Application to Gaze Redirection.","authors":"Harsimran Kaur, Roberto Manduchi","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>We propose a method for synthesizing eye images from segmentation masks with a desired style. The style encompasses attributes such as skin color, texture, iris color, and personal identity. Our approach generates an eye image that is consistent with a given segmentation mask and has the attributes of the input style image. We apply our method to data augmentation as well as to gaze redirection. The previous techniques of synthesizing real eye images from synthetic eye images for data augmentation lacked control over the generated attributes. We demonstrate the effectiveness of the proposed method in synthesizing realistic eye images with given characteristics corresponding to the synthetic labels for data augmentation, which is further useful for various tasks such as gaze estimation, eye image segmentation, pupil detection, etc. We also show how our approach can be applied to gaze redirection using only synthetic gaze labels, improving the previous state of the art results. The main contributions of our paper are i) a novel approach for Style-Based eye image generation from segmentation mask; ii) the use of this approach for gaze-redirection without the need for gaze annotated real eye images.</p>","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":" ","pages":"11-20"},"PeriodicalIF":0.0,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8040934/pdf/nihms-1648872.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"25606536","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hand-Priming in Object Localization for Assistive Egocentric Vision. 辅助性眼心视觉物体定位中的手部定位
Pub Date : 2020-03-01 Epub Date: 2020-05-14 DOI: 10.1109/wacv45572.2020.9093353
Kyungjun Lee, Abhinav Shrivastava, Hernisa Kacorri

Egocentric vision holds great promises for increasing access to visual information and improving the quality of life for people with visual impairments, with object recognition being one of the daily challenges for this population. While we strive to improve recognition performance, it remains difficult to identify which object is of interest to the user; the object may not even be included in the frame due to challenges in camera aiming without visual feedback. Also, gaze information, commonly used to infer the area of interest in egocentric vision, is often not dependable. However, blind users often tend to include their hand either interacting with the object that they wish to recognize or simply placing it in proximity for better camera aiming. We propose localization models that leverage the presence of the hand as the contextual information for priming the center area of the object of interest. In our approach, hand segmentation is fed to either the entire localization network or its last convolutional layers. Using egocentric datasets from sighted and blind individuals, we show that the hand-priming achieves higher precision than other approaches, such as fine-tuning, multi-class, and multi-task learning, which also encode hand-object interactions in localization.

以自我为中心的视觉在增加视觉信息的获取和改善视障人士的生活质量方面大有可为,而物体识别则是视障人士面临的日常挑战之一。在我们努力提高识别性能的同时,识别用户感兴趣的物体仍然很困难;在没有视觉反馈的情况下,由于相机瞄准方面的挑战,物体甚至可能不在画面中。此外,在自我中心视觉中通常用于推断感兴趣区域的注视信息往往并不可靠。然而,盲人用户通常倾向于将他们的手与他们希望识别的物体进行互动,或者只是将其放在附近以更好地瞄准摄像机。我们提出的定位模型可以利用手的存在作为背景信息,以引出感兴趣物体的中心区域。在我们的方法中,手部分割信息被输入到整个定位网络或其最后的卷积层中。通过使用视力正常和失明人士的自我中心数据集,我们证明了手部定位比其他方法(如微调、多类和多任务学习)实现了更高的精确度,其他方法也在定位中编码了手与物体之间的相互作用。
{"title":"Hand-Priming in Object Localization for Assistive Egocentric Vision.","authors":"Kyungjun Lee, Abhinav Shrivastava, Hernisa Kacorri","doi":"10.1109/wacv45572.2020.9093353","DOIUrl":"10.1109/wacv45572.2020.9093353","url":null,"abstract":"<p><p>Egocentric vision holds great promises for increasing access to visual information and improving the quality of life for people with visual impairments, with object recognition being one of the daily challenges for this population. While we strive to improve recognition performance, it remains difficult to identify which object is of interest to the user; the object may not even be included in the frame due to challenges in camera aiming without visual feedback. Also, gaze information, commonly used to infer the area of interest in egocentric vision, is often not dependable. However, blind users often tend to include their hand either interacting with the object that they wish to recognize or simply placing it in proximity for better camera aiming. We propose localization models that leverage the presence of the hand as the contextual information for priming the center area of the object of interest. In our approach, hand segmentation is fed to either the entire localization network or its last convolutional layers. Using egocentric datasets from sighted and blind individuals, we show that the hand-priming achieves higher precision than other approaches, such as fine-tuning, multi-class, and multi-task learning, which also encode hand-object interactions in localization.</p>","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":"2020 ","pages":"3411-3421"},"PeriodicalIF":0.0,"publicationDate":"2020-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7423407/pdf/nihms-1609047.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38269367","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust Template-Based Non-Rigid Motion Tracking Using Local Coordinate Regularization. 基于局部坐标正则化的鲁棒模板非刚体运动跟踪。
Pub Date : 2020-03-01 Epub Date: 2020-05-14 DOI: 10.1109/wacv45572.2020.9093533
Wei Li, Shang Zhao, Xiao Xiao, James K Hahn

In this paper, we propose our template-based non-rigid registration algorithm to address the misalignments in the frame-to-frame motion tracking with single or multiple commodity depth cameras. We analyze the deformation in the local coordinates of neighboring nodes and use this differential representation to formulate the regularization term for the deformation field in our non-rigid registration. The local coordinate regularizations vary for each pair of neighboring nodes based on the tracking status of the surface regions. We propose our tracking strategies for different surface regions to minimize misalignments and reduce error accumulation. This method can thus preserve local geometric features and prevent undesirable distortions. Moreover, we introduce a geodesic-based correspondence estimation algorithm to align surfaces with large displacements. Finally, we demonstrate the effectiveness of our proposed method with detailed experiments.

在本文中,我们提出了基于模板的非刚性配准算法来解决单台或多台商品深度相机帧对帧运动跟踪中的不对准问题。在非刚性配准中,我们分析了相邻节点的局部坐标中的变形,并利用这种微分表示来表示变形场的正则化项。基于表面区域的跟踪状态,每对相邻节点的局部坐标正则化是不同的。我们提出了针对不同表面区域的跟踪策略,以最大限度地减少不对准和减少误差累积。因此,该方法可以保留局部几何特征并防止不必要的扭曲。此外,我们还引入了一种基于测地线的对应估计算法来对大位移曲面。最后,通过详细的实验验证了所提方法的有效性。
{"title":"Robust Template-Based Non-Rigid Motion Tracking Using Local Coordinate Regularization.","authors":"Wei Li, Shang Zhao, Xiao Xiao, James K Hahn","doi":"10.1109/wacv45572.2020.9093533","DOIUrl":"10.1109/wacv45572.2020.9093533","url":null,"abstract":"<p><p>In this paper, we propose our template-based non-rigid registration algorithm to address the misalignments in the frame-to-frame motion tracking with single or multiple commodity depth cameras. We analyze the deformation in the local coordinates of neighboring nodes and use this differential representation to formulate the regularization term for the deformation field in our non-rigid registration. The local coordinate regularizations vary for each pair of neighboring nodes based on the tracking status of the surface regions. We propose our tracking strategies for different surface regions to minimize misalignments and reduce error accumulation. This method can thus preserve local geometric features and prevent undesirable distortions. Moreover, we introduce a geodesic-based correspondence estimation algorithm to align surfaces with large displacements. Finally, we demonstrate the effectiveness of our proposed method with detailed experiments.</p>","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":"2020 ","pages":"390-399"},"PeriodicalIF":0.0,"publicationDate":"2020-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/wacv45572.2020.9093533","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38031947","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Learning Generative Models of Tissue Organization with Supervised GANs. 组织组织的监督式gan学习生成模型。
Ligong Han, Robert F Murphy, Deva Ramanan

A key step in understanding the spatial organization of cells and tissues is the ability to construct generative models that accurately reflect that organization. In this paper, we focus on building generative models of electron microscope (EM) images in which the positions of cell membranes and mitochondria have been densely annotated, and propose a two-stage procedure that produces realistic images using Generative Adversarial Networks (or GANs) in a supervised way. In the first stage, we synthesize a label "image" given a noise "image" as input, which then provides supervision for EM image synthesis in the second stage. The full model naturally generates label-image pairs. We show that accurate synthetic EM images are produced using assessment via (1) shape features and global statistics, (2) segmentation accuracies, and (3) user studies. We also demonstrate further improvements by enforcing a reconstruction loss on intermediate synthetic labels and thus unifying the two stages into one single end-to-end framework.

理解细胞和组织的空间组织的关键一步是构建准确反映该组织的生成模型的能力。在本文中,我们专注于构建电子显微镜(EM)图像的生成模型,其中细胞膜和线粒体的位置已被密集注释,并提出了一个两阶段的过程,该过程使用生成对抗网络(gan)以监督的方式生成逼真的图像。在第一阶段,我们在给定噪声“图像”作为输入的情况下合成标签“图像”,然后为第二阶段的EM图像合成提供监督。完整模型自然生成标签-图像对。我们表明,通过评估(1)形状特征和全局统计,(2)分割准确性,(3)用户研究,可以生成准确的合成EM图像。我们还展示了进一步的改进,通过在中间合成标签上强制重建损失,从而将两个阶段统一到一个单一的端到端框架中。
{"title":"Learning Generative Models of Tissue Organization with Supervised GANs.","authors":"Ligong Han,&nbsp;Robert F Murphy,&nbsp;Deva Ramanan","doi":"10.1109/WACV.2018.00080","DOIUrl":"https://doi.org/10.1109/WACV.2018.00080","url":null,"abstract":"<p><p>A key step in understanding the spatial organization of cells and tissues is the ability to construct generative models that accurately reflect that organization. In this paper, we focus on building generative models of electron microscope (EM) images in which the positions of cell membranes and mitochondria have been densely annotated, and propose a two-stage procedure that produces realistic images using Generative Adversarial Networks (or GANs) in a supervised way. In the first stage, we synthesize a label \"image\" given a noise \"image\" as input, which then provides supervision for EM image synthesis in the second stage. The full model naturally generates label-image pairs. We show that accurate synthetic EM images are produced using assessment via (1) shape features and global statistics, (2) segmentation accuracies, and (3) user studies. We also demonstrate further improvements by enforcing a reconstruction loss on intermediate synthetic labels and thus unifying the two stages into one single end-to-end framework.</p>","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":"2018 ","pages":"682-690"},"PeriodicalIF":0.0,"publicationDate":"2018-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/WACV.2018.00080","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"36458507","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Fast, accurate, small-scale 3D scene capture using a low-cost depth sensor. 快速,准确,小规模的3D场景捕获使用低成本的深度传感器。
Nicole Carey, Radhika Nagpal, Justin Werfel

Commercially available depth sensing devices are primarily designed for domains that are either macroscopic, or static. We develop a solution for fast microscale 3D reconstruction, using off-the-shelf components. By the addition of lenses, precise calibration of camera internals and positioning, and development of bespoke software, we turn an infrared depth sensor designed for human-scale motion and object detection into a device with mm-level accuracy capable of recording at up to 30Hz.

商业上可用的深度传感设备主要是为宏观或静态的领域设计的。我们开发了一种快速微尺度3D重建的解决方案,使用现成的组件。通过增加镜头,精确校准相机内部和定位,以及开发定制软件,我们将设计用于人体尺度运动和物体检测的红外深度传感器转变为具有毫米级精度的设备,能够以高达30Hz的频率进行记录。
{"title":"Fast, accurate, small-scale 3D scene capture using a low-cost depth sensor.","authors":"Nicole Carey,&nbsp;Radhika Nagpal,&nbsp;Justin Werfel","doi":"10.1109/WACV.2017.146","DOIUrl":"https://doi.org/10.1109/WACV.2017.146","url":null,"abstract":"<p><p>Commercially available depth sensing devices are primarily designed for domains that are either macroscopic, or static. We develop a solution for fast microscale 3D reconstruction, using off-the-shelf components. By the addition of lenses, precise calibration of camera internals and positioning, and development of bespoke software, we turn an infrared depth sensor designed for human-scale motion and object detection into a device with mm-level accuracy capable of recording at up to 30Hz.</p>","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":"2017 ","pages":"1268-1276"},"PeriodicalIF":0.0,"publicationDate":"2017-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/WACV.2017.146","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"35279256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Center-Focusing Multi-task CNN with Injected Features for Classification of Glioma Nuclear Images. 利用注入特征的中心聚焦多任务 CNN 对胶质瘤核图像进行分类
Veda Murthy, Le Hou, Dimitris Samaras, Tahsin M Kurc, Joel H Saltz

Classifying the various shapes and attributes of a glioma cell nucleus is crucial for diagnosis and understanding of the disease. We investigate the automated classification of the nuclear shapes and visual attributes of glioma cells, using Convolutional Neural Networks (CNNs) on pathology images of automatically segmented nuclei. We propose three methods that improve the performance of a previously-developed semi-supervised CNN. First, we propose a method that allows the CNN to focus on the most important part of an image-the image's center containing the nucleus. Second, we inject (concatenate) pre-extracted VGG features into an intermediate layer of our Semi-Supervised CNN so that during training, the CNN can learn a set of additional features. Third, we separate the losses of the two groups of target classes (nuclear shapes and attributes) into a single-label loss and a multi-label loss in order to incorporate prior knowledge of inter-label exclusiveness. On a dataset of 2078 images, the combination of the proposed methods reduces the error rate of attribute and shape classification by 21.54% and 15.07% respectively compared to the existing state-of-the-art method on the same dataset.

对胶质瘤细胞核的各种形状和属性进行分类对于诊断和了解该疾病至关重要。我们研究了使用卷积神经网络(CNN)对病理图像中自动分割的胶质瘤细胞核形状和视觉属性进行自动分类。我们提出了三种方法来提高之前开发的半监督 CNN 的性能。首先,我们提出了一种方法,让 CNN 专注于图像中最重要的部分--包含细胞核的图像中心。其次,我们将预先提取的 VGG 特征注入(串联)到半监督 CNN 的中间层,这样 CNN 就能在训练过程中学习一组额外的特征。第三,我们将两组目标类别(核形状和属性)的损失分为单标签损失和多标签损失,以纳入标签间排他性的先验知识。在一个包含 2078 幅图像的数据集上,与相同数据集上现有的最先进方法相比,结合使用所提出的方法,属性和形状分类的错误率分别降低了 21.54% 和 15.07%。
{"title":"Center-Focusing Multi-task CNN with Injected Features for Classification of Glioma Nuclear Images.","authors":"Veda Murthy, Le Hou, Dimitris Samaras, Tahsin M Kurc, Joel H Saltz","doi":"10.1109/WACV.2017.98","DOIUrl":"10.1109/WACV.2017.98","url":null,"abstract":"<p><p>Classifying the various shapes and attributes of a glioma cell nucleus is crucial for diagnosis and understanding of the disease. We investigate the automated classification of the nuclear shapes and visual attributes of glioma cells, using Convolutional Neural Networks (CNNs) on pathology images of automatically segmented nuclei. We propose three methods that improve the performance of a previously-developed semi-supervised CNN. First, we propose a method that allows the CNN to focus on the most important part of an image-the image's center containing the nucleus. Second, we inject (concatenate) pre-extracted VGG features into an intermediate layer of our Semi-Supervised CNN so that during training, the CNN can learn a set of additional features. Third, we separate the losses of the two groups of target classes (nuclear shapes and attributes) into a single-label loss and a multi-label loss in order to incorporate prior knowledge of inter-label exclusiveness. On a dataset of 2078 images, the combination of the proposed methods reduces the error rate of attribute and shape classification by 21.54% and 15.07% respectively compared to the existing state-of-the-art method on the same dataset.</p>","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":"2017 ","pages":"834-841"},"PeriodicalIF":0.0,"publicationDate":"2017-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5988234/pdf/nihms969223.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"36205043","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Recognition of 3D package shapes for single camera metrology 用于单相机计量的三维封装形状识别
Ryan Lloyd, Scott McCloskey
Many applications of 3D object measurement have become commercially viable due to the recent availability of low-cost range cameras such as the Microsoft Kinect. We address the application of measuring an object's dimensions for the purpose of billing in shipping transactions, where high accuracy is required for certification. In particular, we address cases where an object's pose reduces the accuracy with which we can estimate dimensions from a single camera. Because the class of object shapes is limited in the shipping domain, we perform a closed-world recognition in order to determine a shape model which can account for missing parts, and/or to induce the user to reposition the object for higher accuracy. Our experiments demonstrate that the addition of this recognition step significantly improves system accuracy.
由于最近出现了像微软Kinect这样的低成本范围相机,许多3D物体测量的应用已经在商业上可行。我们解决了在航运交易中用于计费目的的测量对象尺寸的应用,其中认证需要高精度。特别是,我们解决了物体的姿势降低了我们可以从单个相机估计尺寸的准确性的情况。由于物体形状的类别在运输领域受到限制,我们执行封闭世界识别,以确定可以解释缺失部分的形状模型,和/或诱导用户重新定位物体以获得更高的精度。实验表明,该识别步骤的加入显著提高了系统的准确率。
{"title":"Recognition of 3D package shapes for single camera metrology","authors":"Ryan Lloyd, Scott McCloskey","doi":"10.1109/WACV.2014.6836113","DOIUrl":"https://doi.org/10.1109/WACV.2014.6836113","url":null,"abstract":"Many applications of 3D object measurement have become commercially viable due to the recent availability of low-cost range cameras such as the Microsoft Kinect. We address the application of measuring an object's dimensions for the purpose of billing in shipping transactions, where high accuracy is required for certification. In particular, we address cases where an object's pose reduces the accuracy with which we can estimate dimensions from a single camera. Because the class of object shapes is limited in the shipping domain, we perform a closed-world recognition in order to determine a shape model which can account for missing parts, and/or to induce the user to reposition the object for higher accuracy. Our experiments demonstrate that the addition of this recognition step significantly improves system accuracy.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":"41 1","pages":"99-106"},"PeriodicalIF":0.0,"publicationDate":"2014-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88394086","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
期刊
IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1