首页 > 最新文献

Proceedings of the ... IAPR International Conference on Pattern Recognition. International Conference on Pattern Recognition最新文献

英文 中文
Complexity of Representations in Deep Learning 深度学习中表征的复杂性
T. Ho
—Deep neural networks use multiple layers of functions to map an object represented by an input vector progressively to different representations, and with sufficient training, eventually to a single score for each class that is the output of the final decision function. Ideally, in this output space, the objects of different classes achieve maximum separation. Motivated by the need to better understand the inner working of a deep neural network, we analyze the effectiveness of the learned representations in separating the classes from a data complexity perspective. Using a simple complexity measure, a popular benchmarking task, and a well-known architecture design, we show how the data complexity evolves through the network, how it changes during training, and how it is impacted by the network design and the availability of training samples. We discuss the implications of the observations and the potentials for further studies.
深度神经网络使用多层函数将由输入向量表示的对象逐步映射到不同的表示,并经过充分的训练,最终得到每个类的单个分数,这是最终决策函数的输出。理想情况下,在这个输出空间中,不同类的对象实现最大的分离。由于需要更好地理解深度神经网络的内部工作,我们从数据复杂性的角度分析了学习表征在分离类方面的有效性。使用简单的复杂性度量、流行的基准测试任务和著名的架构设计,我们展示了数据复杂性如何通过网络演变,在训练过程中如何变化,以及它如何受到网络设计和训练样本可用性的影响。我们讨论了观察的意义和进一步研究的潜力。
{"title":"Complexity of Representations in Deep Learning","authors":"T. Ho","doi":"10.48550/arXiv.2209.00525","DOIUrl":"https://doi.org/10.48550/arXiv.2209.00525","url":null,"abstract":"—Deep neural networks use multiple layers of functions to map an object represented by an input vector progressively to different representations, and with sufficient training, eventually to a single score for each class that is the output of the final decision function. Ideally, in this output space, the objects of different classes achieve maximum separation. Motivated by the need to better understand the inner working of a deep neural network, we analyze the effectiveness of the learned representations in separating the classes from a data complexity perspective. Using a simple complexity measure, a popular benchmarking task, and a well-known architecture design, we show how the data complexity evolves through the network, how it changes during training, and how it is impacted by the network design and the availability of training samples. We discuss the implications of the observations and the potentials for further studies.","PeriodicalId":74516,"journal":{"name":"Proceedings of the ... IAPR International Conference on Pattern Recognition. International Conference on Pattern Recognition","volume":"57 1","pages":"2657-2663"},"PeriodicalIF":0.0,"publicationDate":"2022-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83566819","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Extraction of Ruler Markings For Estimating Physical Size of Oral Lesions. 用于估计口腔病变物理尺寸的标尺标记提取。
Zhiyun Xue, Kelly Yu, Paul Pearlman, Tseng-Cheng Chen, Chun-Hung Hua, Chung Jan Kang, Chih-Yen Chien, Ming-Hsui Tsai, Cheng-Ping Wang, Anil Chaturvedi, Sameer Antani

Small ruler tapes are commonly placed on the surface of the human body as a simple and efficient reference for capturing on images the physical size of a lesion. In this paper, we describe our proposed approach for automatically extracting the measurement information from a ruler in oral cavity images which are taken during oral cancer screening and follow up. The images were taken during a study that aims to investigate the natural history of histologically defined oral cancer precursor lesions and identify epidemiologic factors and molecular markers associated with disease progression. Compared to similar work in the literature proposed for other applications where images are captured with greater consistency and in more controlled situations, we address additional challenges that our application faces in real world use and with analysis of retrospectively collected data. Our approach considers several conditions with respect to ruler style, ruler visibility completeness, and image quality. Further, we provide multiple ways of extracting ruler markings and measurement calculation based on specific conditions. We evaluated the proposed method on two datasets obtained from different sources and examined cross-dataset performance.

通常将小尺带放置在人体表面,作为在图像上捕获病变物理尺寸的简单而有效的参考。在本文中,我们提出了一种从口腔癌筛查和随访期间拍摄的口腔图像中自动提取测量信息的方法。这些图像是在一项研究中拍摄的,该研究旨在调查组织学上确定的口腔癌前驱病变的自然历史,并确定与疾病进展相关的流行病学因素和分子标记。与文献中提出的类似工作相比,在其他应用程序中,图像捕获具有更高的一致性和更可控的情况下,我们解决了我们的应用程序在现实世界使用中面临的额外挑战,并分析了回顾性收集的数据。我们的方法考虑了标尺样式、标尺可见性完整性和图像质量方面的几个条件。此外,我们提供了多种方法提取标尺标记和测量计算根据具体情况。我们在来自不同来源的两个数据集上评估了所提出的方法,并检查了跨数据集的性能。
{"title":"Extraction of Ruler Markings For Estimating Physical Size of Oral Lesions.","authors":"Zhiyun Xue,&nbsp;Kelly Yu,&nbsp;Paul Pearlman,&nbsp;Tseng-Cheng Chen,&nbsp;Chun-Hung Hua,&nbsp;Chung Jan Kang,&nbsp;Chih-Yen Chien,&nbsp;Ming-Hsui Tsai,&nbsp;Cheng-Ping Wang,&nbsp;Anil Chaturvedi,&nbsp;Sameer Antani","doi":"10.1109/icpr56361.2022.9956251","DOIUrl":"https://doi.org/10.1109/icpr56361.2022.9956251","url":null,"abstract":"<p><p>Small ruler tapes are commonly placed on the surface of the human body as a simple and efficient reference for capturing on images the physical size of a lesion. In this paper, we describe our proposed approach for automatically extracting the measurement information from a ruler in oral cavity images which are taken during oral cancer screening and follow up. The images were taken during a study that aims to investigate the natural history of histologically defined oral cancer precursor lesions and identify epidemiologic factors and molecular markers associated with disease progression. Compared to similar work in the literature proposed for other applications where images are captured with greater consistency and in more controlled situations, we address additional challenges that our application faces in real world use and with analysis of retrospectively collected data. Our approach considers several conditions with respect to ruler style, ruler visibility completeness, and image quality. Further, we provide multiple ways of extracting ruler markings and measurement calculation based on specific conditions. We evaluated the proposed method on two datasets obtained from different sources and examined cross-dataset performance.</p>","PeriodicalId":74516,"journal":{"name":"Proceedings of the ... IAPR International Conference on Pattern Recognition. International Conference on Pattern Recognition","volume":"2022 ","pages":"4241-4247"},"PeriodicalIF":0.0,"publicationDate":"2022-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9728633/pdf/nihms-1840606.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10509307","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
TensorMixup Data Augmentation Method for Fully Automatic Brain Tumor Segmentation 全自动脑肿瘤分割的TensorMixup数据增强方法
Yu Wang, Ya-Liang Ji
{"title":"TensorMixup Data Augmentation Method for Fully Automatic Brain Tumor Segmentation","authors":"Yu Wang, Ya-Liang Ji","doi":"10.1109/ICPR56361.2022.9956689","DOIUrl":"https://doi.org/10.1109/ICPR56361.2022.9956689","url":null,"abstract":"","PeriodicalId":74516,"journal":{"name":"Proceedings of the ... IAPR International Conference on Pattern Recognition. International Conference on Pattern Recognition","volume":"20 1 1","pages":"4615-4622"},"PeriodicalIF":0.0,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77443821","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Classifying Breast Histopathology Images with a Ductal Instance-Oriented Pipeline. 基于导管实例导向管道的乳腺组织病理学图像分类。
Beibin Li, Ezgi Mercan, Sachin Mehta, Stevan Knezevich, Corey W Arnold, Donald L Weaver, Joann G Elmore, Linda G Shapiro

In this study, we propose the Ductal Instance-Oriented Pipeline (DIOP) that contains a duct-level instance segmentation model, a tissue-level semantic segmentation model, and three-levels of features for diagnostic classification. Based on recent advancements in instance segmentation and the Mask RCNN model, our duct-level segmenter tries to identify each ductal individual inside a microscopic image; then, it extracts tissue-level information from the identified ductal instances. Leveraging three levels of information obtained from these ductal instances and also the histopathology image, the proposed DIOP outperforms previous approaches (both feature-based and CNN-based) in all diagnostic tasks; for the four-way classification task, the DIOP achieves comparable performance to general pathologists in this unique dataset. The proposed DIOP only takes a few seconds to run in the inference time, which could be used interactively on most modern computers. More clinical explorations are needed to study the robustness and generalizability of this system in the future.

在本研究中,我们提出了导管面向实例的管道(Ductal instance - oriented Pipeline, DIOP),该管道包含一个导管级实例分割模型、一个组织级语义分割模型和用于诊断分类的三级特征。基于实例分割和Mask RCNN模型的最新进展,我们的管道级分割器试图识别微观图像中的每个管道个体;然后,从已识别的导管实例中提取组织级信息。利用从这些导管实例和组织病理学图像中获得的三层信息,所提出的DIOP在所有诊断任务中都优于以前的方法(基于特征和基于cnn的方法);对于四向分类任务,DIOP在这个独特的数据集中实现了与普通病理学家相当的性能。所提出的DIOP在推理时间内只需要几秒钟的运行时间,可以在大多数现代计算机上交互式地使用。该系统的稳健性和通用性有待进一步的临床探索。
{"title":"Classifying Breast Histopathology Images with a Ductal Instance-Oriented Pipeline.","authors":"Beibin Li,&nbsp;Ezgi Mercan,&nbsp;Sachin Mehta,&nbsp;Stevan Knezevich,&nbsp;Corey W Arnold,&nbsp;Donald L Weaver,&nbsp;Joann G Elmore,&nbsp;Linda G Shapiro","doi":"10.1109/icpr48806.2021.9412824","DOIUrl":"https://doi.org/10.1109/icpr48806.2021.9412824","url":null,"abstract":"<p><p>In this study, we propose the Ductal Instance-Oriented Pipeline (DIOP) that contains a duct-level instance segmentation model, a tissue-level semantic segmentation model, and three-levels of features for diagnostic classification. Based on recent advancements in instance segmentation and the Mask RCNN model, our duct-level segmenter tries to identify each ductal individual inside a microscopic image; then, it extracts tissue-level information from the identified ductal instances. Leveraging three levels of information obtained from these ductal instances and also the histopathology image, the proposed DIOP outperforms previous approaches (both feature-based and CNN-based) in all diagnostic tasks; for the four-way classification task, the DIOP achieves comparable performance to general pathologists in this unique dataset. The proposed DIOP only takes a few seconds to run in the inference time, which could be used interactively on most modern computers. More clinical explorations are needed to study the robustness and generalizability of this system in the future.</p>","PeriodicalId":74516,"journal":{"name":"Proceedings of the ... IAPR International Conference on Pattern Recognition. International Conference on Pattern Recognition","volume":"2020 ","pages":"8727-8734"},"PeriodicalIF":0.0,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/icpr48806.2021.9412824","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10718778","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Dependently Coupled Principal Component Analysis for Bivariate Inversion Problems. 二元反演问题的依赖耦合主成分分析。
Navdeep Dahiya, Yifei Fan, Samuel Bignardi, Romeil Sandhu, Anthony Yezzi

Principal Component Analysis (PCA) is a widely used technique for dimensionality reduction in various problem domains, including data compression, image processing, visualization, exploratory data analysis, pattern recognition, time-series prediction, and machine learning. Often, data is presented in a correlated paired manner such that there exist observable and correlated unobservable measurements. Unfortunately, traditional PCA techniques generally fail to optimally capture the leverageable correlations between such paired data as it does not yield a maximally correlated basis between the observable and unobservable counterparts. This instead is the objective of Canonical Correlation Analysis (and the more general Partial Least Squares methods); however, such techniques are still symmetric in maximizing correlation (covariance for PLSR) over all choices of the basis for both datasets without differentiating between observable and unobservable variables (except for the regression phase of PLSR). Further, these methods deviate from PCA's formulation objective to minimize approximation error, seeking instead to maximize correlation or covariance. While these are sensible optimization objectives, they are not equivalent to error minimization. We therefore introduce a new method of leveraging PCA between paired datasets in a dependently coupled manner, which is optimal with respect to approximation error during training. We generate a dependently coupled paired basis for which we relax orthogonality constraints in decomposing unreliable unobservable measurements. In doing so, this allows us to optimally capture the variations of the observable data while conditionally minimizing the expected prediction error for the unobservable component. We show preliminary results that demonstrate improved learning of our proposed method compared to that of traditional techniques.

主成分分析(PCA)是一种广泛应用于各种问题领域的降维技术,包括数据压缩、图像处理、可视化、探索性数据分析、模式识别、时间序列预测和机器学习。通常情况下,数据是以相关配对的方式呈现的,因此存在可观察到的测量值和相关的不可观察到的测量值。遗憾的是,传统的 PCA 技术通常无法最佳地捕捉此类配对数据之间的可利用相关性,因为它无法在可观测和不可观测的对应数据之间建立最大相关性基础。而这正是典型相关分析法(以及更一般的偏最小二乘法)的目标;然而,这些技术仍然是对称的,即在两个数据集的所有基础选择中最大化相关性(PLSR 的协方差),而不区分可观测变量和不可观测变量(PLSR 的回归阶段除外)。此外,这些方法偏离了 PCA 最小化近似误差的表述目标,而是寻求最大化相关性或协方差。虽然这些都是合理的优化目标,但它们并不等同于误差最小化。因此,我们引入了一种新方法,以依赖耦合的方式利用配对数据集之间的 PCA,这种方法在训练过程中对近似误差是最优的。我们生成一个依赖耦合的配对基础,在分解不可靠的不可观测测量数据时,我们放宽了该基础的正交性约束。这样,我们就能以最佳方式捕捉可观测数据的变化,同时有条件地使不可观测部分的预期预测误差最小化。我们展示的初步结果表明,与传统技术相比,我们提出的方法提高了学习效率。
{"title":"Dependently Coupled Principal Component Analysis for Bivariate Inversion Problems.","authors":"Navdeep Dahiya, Yifei Fan, Samuel Bignardi, Romeil Sandhu, Anthony Yezzi","doi":"10.1109/icpr48806.2021.9413305","DOIUrl":"10.1109/icpr48806.2021.9413305","url":null,"abstract":"<p><p>Principal Component Analysis (PCA) is a widely used technique for dimensionality reduction in various problem domains, including data compression, image processing, visualization, exploratory data analysis, pattern recognition, time-series prediction, and machine learning. Often, data is presented in a correlated paired manner such that there exist observable and correlated unobservable measurements. Unfortunately, traditional PCA techniques generally fail to optimally capture the leverageable correlations between such paired data as it does not yield a maximally correlated basis between the observable and unobservable counterparts. This instead is the objective of Canonical Correlation Analysis (and the more general Partial Least Squares methods); however, such techniques are still symmetric in maximizing correlation (covariance for PLSR) over all choices of the basis for both datasets without differentiating between observable and unobservable variables (except for the regression phase of PLSR). Further, these methods deviate from PCA's formulation objective to minimize approximation error, seeking instead to maximize correlation or covariance. While these are sensible optimization objectives, they are not equivalent to error minimization. We therefore introduce a new method of leveraging PCA between paired datasets in a dependently coupled manner, which is optimal with respect to approximation error during training. We generate a dependently coupled paired basis for which we relax orthogonality constraints in decomposing unreliable unobservable measurements. In doing so, this allows us to optimally capture the variations of the observable data while conditionally minimizing the expected prediction error for the unobservable component. We show preliminary results that demonstrate improved learning of our proposed method compared to that of traditional techniques.</p>","PeriodicalId":74516,"journal":{"name":"Proceedings of the ... IAPR International Conference on Pattern Recognition. International Conference on Pattern Recognition","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8330695/pdf/nihms-1676299.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39277423","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Directionally Paired Principal Component Analysis for Bivariate Estimation Problems. 二元估计问题的方向配对主成分分析。
Yifei Fan, Navdeep Dahiya, Samuel Bignardi, Romeil Sandhu, Anthony Yezzi

We propose Directionally Paired Principal Component Analysis (DP-PCA), a novel linear dimension-reduction model for estimating coupled yet partially observable variable sets. Unlike partial least squares methods (e.g., partial least squares regression and canonical correlation analysis) that maximize correlation/covariance between the two datasets, our DP-PCA directly minimizes, either conditionally or unconditionally, the reconstruction and prediction errors for the observable and unobservable part, respectively. We demonstrate the optimality of the proposed DP-PCA approach, we compare and evaluate relevant linear cross-decomposition methods with data reconstruction and prediction experiments on synthetic Gaussian data, multi-target regression datasets, and a single-channel image dataset. Results show that when only a single pair of bases is allowed, the conditional DP-PCA achieves the lowest reconstruction error on the observable part and the total variable sets as a whole; meanwhile, the unconditional DP-PCA reaches the lowest prediction errors on the unobservable part. When an extra budget is allowed for the observable part's PCA basis, one can reach an optimal solution using a combined method: standard PCA for the observable part and unconditional DP-PCA for the unobservable part.

我们提出了方向配对主成分分析(DP-PCA),这是一种新的线性降维模型,用于估计耦合但部分可观察的变量集。与偏最小二乘方法(例如,偏最小二乘回归和典型相关分析)最大化两个数据集之间的相关性/协方差不同,我们的DP-PCA分别直接最小化可观测部分和不可观测部分的重建和预测误差,条件或无条件地最小化。我们证明了所提出的DP-PCA方法的最优性,并通过在合成高斯数据、多目标回归数据集和单通道图像数据集上的数据重建和预测实验,比较和评估了相关的线性交叉分解方法。结果表明,当只允许一对碱基时,条件DP-PCA在可观测部分和总体变量集上的重构误差最小;同时,无条件DP-PCA在不可观测部分的预测误差最小。当可观测部分的PCA基础允许额外预算时,可以使用可观测部分的标准PCA和不可观测部分的无条件DP-PCA相结合的方法来获得最优解。
{"title":"Directionally Paired Principal Component Analysis for Bivariate Estimation Problems.","authors":"Yifei Fan,&nbsp;Navdeep Dahiya,&nbsp;Samuel Bignardi,&nbsp;Romeil Sandhu,&nbsp;Anthony Yezzi","doi":"10.1109/icpr48806.2021.9412245","DOIUrl":"https://doi.org/10.1109/icpr48806.2021.9412245","url":null,"abstract":"<p><p>We propose Directionally Paired Principal Component Analysis (DP-PCA), a novel linear dimension-reduction model for estimating coupled yet partially observable variable sets. Unlike partial least squares methods (e.g., partial least squares regression and canonical correlation analysis) that maximize correlation/covariance between the two datasets, our DP-PCA directly minimizes, either conditionally or unconditionally, the reconstruction and prediction errors for the observable and unobservable part, respectively. We demonstrate the optimality of the proposed DP-PCA approach, we compare and evaluate relevant linear cross-decomposition methods with data reconstruction and prediction experiments on synthetic Gaussian data, multi-target regression datasets, and a single-channel image dataset. Results show that when only a single pair of bases is allowed, the conditional DP-PCA achieves the lowest reconstruction error on the observable part and the total variable sets as a whole; meanwhile, the unconditional DP-PCA reaches the lowest prediction errors on the unobservable part. When an extra budget is allowed for the observable part's PCA basis, one can reach an optimal solution using a combined method: standard PCA for the observable part and unconditional DP-PCA for the unobservable part.</p>","PeriodicalId":74516,"journal":{"name":"Proceedings of the ... IAPR International Conference on Pattern Recognition. International Conference on Pattern Recognition","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/icpr48806.2021.9412245","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39266883","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic Estimation of Self-Reported Pain by Interpretable Representations of Motion Dynamics. 运动动力学可解释表征对自我报告疼痛的自动估计。
Benjamin Szczapa, Mohamed Daoudi, Stefano Berretti, Pietro Pala, Alberto Del Bimbo, Zakia Hammal

We propose an automatic method for pain intensity measurement from video. For each video, pain intensity was measured using the dynamics of facial movement using 66 facial points. Gram matrices formulation was used for facial points trajectory representations on the Riemannian manifold of symmetric positive semi-definite matrices of fixed rank. Curve fitting and temporal alignment were then used to smooth the extracted trajectories. A Support Vector Regression model was then trained to encode the extracted trajectories into ten pain intensity levels consistent with the Visual Analogue Scale for pain intensity measurement. The proposed approach was evaluated using the UNBC McMaster Shoulder Pain Archive and was compared to the state-of-the-art on the same data. Using both 5-fold cross-validation and leave-one-subject-out cross-validation, our results are competitive with respect to state-of-the-art methods.

我们提出了一种基于视频的疼痛强度自动测量方法。对于每个视频,疼痛强度是通过使用66个面部点的面部运动动态来测量的。在对称定秩正半定矩阵的黎曼流形上,采用格拉姆矩阵表示面点轨迹。然后使用曲线拟合和时间对齐来平滑提取的轨迹。然后训练支持向量回归模型,将提取的轨迹编码为十个疼痛强度水平,这些疼痛强度水平与疼痛强度测量的视觉模拟量表一致。采用UNBC麦克马斯特肩部疼痛档案对建议的方法进行评估,并在相同的数据上与最先进的方法进行比较。使用5倍交叉验证和留一个受试者的交叉验证,我们的结果与最先进的方法相比具有竞争力。
{"title":"Automatic Estimation of Self-Reported Pain by Interpretable Representations of Motion Dynamics.","authors":"Benjamin Szczapa,&nbsp;Mohamed Daoudi,&nbsp;Stefano Berretti,&nbsp;Pietro Pala,&nbsp;Alberto Del Bimbo,&nbsp;Zakia Hammal","doi":"10.1109/icpr48806.2021.9412292","DOIUrl":"https://doi.org/10.1109/icpr48806.2021.9412292","url":null,"abstract":"<p><p>We propose an automatic method for pain intensity measurement from video. For each video, pain intensity was measured using the dynamics of facial movement using 66 facial points. Gram matrices formulation was used for facial points trajectory representations on the Riemannian manifold of symmetric positive semi-definite matrices of fixed rank. Curve fitting and temporal alignment were then used to smooth the extracted trajectories. A Support Vector Regression model was then trained to encode the extracted trajectories into ten pain intensity levels consistent with the Visual Analogue Scale for pain intensity measurement. The proposed approach was evaluated using the UNBC McMaster Shoulder Pain Archive and was compared to the state-of-the-art on the same data. Using both 5-fold cross-validation and leave-one-subject-out cross-validation, our results are competitive with respect to state-of-the-art methods.</p>","PeriodicalId":74516,"journal":{"name":"Proceedings of the ... IAPR International Conference on Pattern Recognition. International Conference on Pattern Recognition","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/icpr48806.2021.9412292","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39520929","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Multi-focus Image Fusion for Confocal Microscopy Using U-Net Regression Map. 利用 U-Net 回归图实现共焦显微镜的多焦点图像融合
Maruf Hossain Shuvo, Yasmin M Kassim, Filiz Bunyak, Olga V Glinskii, Leike Xie, Vladislav V Glinsky, Virginia H Huxley, Mahesh M Thakkar, Kannappan Palaniappan

Characterizing the spatial relationship between blood vessel and lymphatic vascular structures, in the mice dura mater tissue, is useful for modeling fluid flows and changes in dynamics in various disease processes. We propose a new deep learning-based approach to fuse a set of multi-channel single-focus microscopy images within each volumetric z-stack into a single fused image that accurately captures as much of the vascular structures as possible. The red spectral channel captures small blood vessels and the green fluorescence channel images lymphatics structures in the intact dura mater attached to bone. The deep architecture Multi-Channel Fusion U-Net (MCFU-Net) combines multi-slice regression likelihood maps of thin linear structures using max pooling for each channel independently to estimate a slice-based focus selection map. We compare MCFU-Net with a widely used derivative-based multi-scale Hessian fusion method [8]. The multi-scale Hessian-based fusion produces dark-halos, non-homogeneous backgrounds and less detailed anatomical structures. Perception based no-reference image quality assessment metrics PIQUE, NIQE, and BRISQUE confirm the effectiveness of the proposed method.

描述小鼠硬脑膜组织中血管和淋巴管结构之间的空间关系有助于模拟各种疾病过程中的流体流动和动态变化。我们提出了一种基于深度学习的新方法,将每个容积 Z 叠中的一组多通道单焦距显微镜图像融合为一张融合图像,尽可能准确地捕捉血管结构。红色光谱通道捕捉小血管,绿色荧光通道拍摄附着在骨骼上的完整硬脑膜淋巴管结构。深度架构多通道融合 U-Net(MCFU-Net)结合了薄线性结构的多切片回归似然图,使用每个通道独立的最大池化来估计基于切片的病灶选择图。我们将 MCFU-Net 与广泛使用的基于导数的多尺度 Hessian 融合方法[8]进行了比较。基于多尺度 Hessian 的融合方法会产生暗晕、非均匀背景和不太详细的解剖结构。基于感知的无参考图像质量评估指标 PIQUE、NIQE 和 BRISQUE 证实了所提方法的有效性。
{"title":"Multi-focus Image Fusion for Confocal Microscopy Using U-Net Regression Map.","authors":"Maruf Hossain Shuvo, Yasmin M Kassim, Filiz Bunyak, Olga V Glinskii, Leike Xie, Vladislav V Glinsky, Virginia H Huxley, Mahesh M Thakkar, Kannappan Palaniappan","doi":"10.1109/icpr48806.2021.9412122","DOIUrl":"10.1109/icpr48806.2021.9412122","url":null,"abstract":"<p><p>Characterizing the spatial relationship between blood vessel and lymphatic vascular structures, in the mice dura mater tissue, is useful for modeling fluid flows and changes in dynamics in various disease processes. We propose a new deep learning-based approach to fuse a set of multi-channel single-focus microscopy images within each volumetric z-stack into a single fused image that accurately captures as much of the vascular structures as possible. The red spectral channel captures small blood vessels and the green fluorescence channel images lymphatics structures in the intact dura mater attached to bone. The deep architecture Multi-Channel Fusion U-Net (MCFU-Net) combines multi-slice regression likelihood maps of thin linear structures using max pooling for each channel independently to estimate a slice-based focus selection map. We compare MCFU-Net with a widely used derivative-based multi-scale Hessian fusion method [8]. The multi-scale Hessian-based fusion produces dark-halos, non-homogeneous backgrounds and less detailed anatomical structures. Perception based no-reference image quality assessment metrics PIQUE, NIQE, and BRISQUE confirm the effectiveness of the proposed method.</p>","PeriodicalId":74516,"journal":{"name":"Proceedings of the ... IAPR International Conference on Pattern Recognition. International Conference on Pattern Recognition","volume":" ","pages":"4317-4323"},"PeriodicalIF":0.0,"publicationDate":"2021-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8513773/pdf/nihms-1685783.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39519299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PA-FlowNet: Pose-Auxiliary Optical Flow Network for Spacecraft Relative Pose Estimation PA-FlowNet:用于航天器相对姿态估计的位姿辅助光流网络
Zhi-Yu Chen, Po-Heng Chen, Kuan-Wen Chen, Chen-Yu Chan
{"title":"PA-FlowNet: Pose-Auxiliary Optical Flow Network for Spacecraft Relative Pose Estimation","authors":"Zhi-Yu Chen, Po-Heng Chen, Kuan-Wen Chen, Chen-Yu Chan","doi":"10.1109/ICPR48806.2021.9413201","DOIUrl":"https://doi.org/10.1109/ICPR48806.2021.9413201","url":null,"abstract":"","PeriodicalId":74516,"journal":{"name":"Proceedings of the ... IAPR International Conference on Pattern Recognition. International Conference on Pattern Recognition","volume":"37 1","pages":"9703-9710"},"PeriodicalIF":0.0,"publicationDate":"2020-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81380663","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dense Receptive Field for Object Detection 对象检测的密集接受场
Yao Yongqiang, Dong Yuan, Huang Zesang, Bai Hongliang
Current one-stage single-shot detectors such as DSSD and StairNet based on aggregating context information from multiple scales have shown promising accuracy. However, existing multi-scale context fusion techniques are insufficient for detecting objects of different scales. In this paper, we investigate how to detect different objects with different scales with respect to accuracy-vs-speed trade-off. We propose a novel single-shot based detector, called DRFNet which fuses feature maps with different sizes of the receptive field to boost the detection accuracy. Our final model DRFNet detector unifies comprehensive context information from various receptive fields effectively to enable it to detect objects in different sizes with higher accuracy. Experimental results on PASCAL VOC 2007 benchmark (79.6% mAP, 68 FPS) demonstrate that DRFNet is better than other state-of-the-art one-stage detectors similar to FPN. Code is released at https://github.com/yqyao/DRFNet.
目前,DSSD和StairNet等基于多尺度上下文信息聚合的单阶段单镜头检测器已经显示出良好的准确性。然而,现有的多尺度上下文融合技术不足以检测不同尺度的目标。在本文中,我们研究了如何检测不同尺度的不同目标,并考虑了精度与速度的权衡。我们提出了一种新的基于单镜头的检测器,称为DRFNet,它融合了不同大小的感受野的特征图来提高检测精度。我们的最终模型DRFNet检测器有效地统一了来自各种接受野的综合上下文信息,使其能够以更高的精度检测不同大小的物体。在PASCAL VOC 2007基准(79.6% mAP, 68 FPS)上的实验结果表明,DRFNet比其他最先进的类似FPN的单级检测器更好。代码发布在https://github.com/yqyao/DRFNet。
{"title":"Dense Receptive Field for Object Detection","authors":"Yao Yongqiang, Dong Yuan, Huang Zesang, Bai Hongliang","doi":"10.1109/ICPR.2018.8546207","DOIUrl":"https://doi.org/10.1109/ICPR.2018.8546207","url":null,"abstract":"Current one-stage single-shot detectors such as DSSD and StairNet based on aggregating context information from multiple scales have shown promising accuracy. However, existing multi-scale context fusion techniques are insufficient for detecting objects of different scales. In this paper, we investigate how to detect different objects with different scales with respect to accuracy-vs-speed trade-off. We propose a novel single-shot based detector, called DRFNet which fuses feature maps with different sizes of the receptive field to boost the detection accuracy. Our final model DRFNet detector unifies comprehensive context information from various receptive fields effectively to enable it to detect objects in different sizes with higher accuracy. Experimental results on PASCAL VOC 2007 benchmark (79.6% mAP, 68 FPS) demonstrate that DRFNet is better than other state-of-the-art one-stage detectors similar to FPN. Code is released at https://github.com/yqyao/DRFNet.","PeriodicalId":74516,"journal":{"name":"Proceedings of the ... IAPR International Conference on Pattern Recognition. International Conference on Pattern Recognition","volume":"15 1","pages":"1815-1820"},"PeriodicalIF":0.0,"publicationDate":"2018-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74603158","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
Proceedings of the ... IAPR International Conference on Pattern Recognition. International Conference on Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1