首页 > 最新文献

Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing最新文献

英文 中文
A Novel Multi-Scale Residual Dense Dehazing Network (MSRDNet) for Single Image Dehazing✱ 一种新的多尺度残差密集去雾网络(MSRDNet),用于单幅图像的译者去雾
Pub Date : 2022-12-08 DOI: 10.1145/3571600.3571601
Chippy M. Manu, G. SreeniK.
Dehazing is a difficult process because of the damage caused by the non-uniform fog and haze distribution in images. To address these issues, a Multi-Scale Residual dense Dehazing Network (MSRDNet) is proposed in this paper. A Contextual feature extraction module (CFM) for extracting multi-scale features and an Adaptive Residual Dense Module (ARDN) are used as sub-modules of MSRDNet. Moreover, all the hierarchical features extracted by each ARDN are fused, which helps to detect hazy maps of varying lengths with multi-scale features. This framework outperforms the state-of-the-art dehazing methods in removing haze while maintaining and restoring image detail in real-world and synthetic images captured under various scenarios.
{"title":"A Novel Multi-Scale Residual Dense Dehazing Network (MSRDNet) for Single Image Dehazing✱","authors":"Chippy M. Manu, G. SreeniK.","doi":"10.1145/3571600.3571601","DOIUrl":"https://doi.org/10.1145/3571600.3571601","url":null,"abstract":"Dehazing is a difficult process because of the damage caused by the non-uniform fog and haze distribution in images. To address these issues, a Multi-Scale Residual dense Dehazing Network (MSRDNet) is proposed in this paper. A Contextual feature extraction module (CFM) for extracting multi-scale features and an Adaptive Residual Dense Module (ARDN) are used as sub-modules of MSRDNet. Moreover, all the hierarchical features extracted by each ARDN are fused, which helps to detect hazy maps of varying lengths with multi-scale features. This framework outperforms the state-of-the-art dehazing methods in removing haze while maintaining and restoring image detail in real-world and synthetic images captured under various scenarios.","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"122 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78584695","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust Brain State Decoding using Bidirectional Long Short Term Memory Networks in functional MRI. 利用功能性核磁共振成像中的双向长短期记忆网络进行稳健的大脑状态解码
Pub Date : 2021-12-01 Epub Date: 2021-12-19 DOI: 10.1145/3490035.3490269
Anant Mittal, Priya Aggarwal, Luiz Pessoa, Anubha Gupta

Decoding brain states of the underlying cognitive processes via learning discriminative feature representations has recently gained a lot of interest in brain imaging studies. Particularly, there has been an impetus to encode the dynamics of brain functioning by analyzing temporal information available in the fMRI data. Long-short term memory (LSTM), a class of machine learning model possessing a "memory" component, to retain previously seen temporal information, is increasingly being observed to perform well in various applications with dynamic temporal behavior, including brain state decoding. Because of the dynamics and inherent latency in fMRI BOLD responses, future temporal context is crucial. However, it is neither encoded nor captured by the conventional LSTM model. This paper performs robust brain state decoding via information encapsulation from both the past and future instances of fMRI data via bi-directional LSTM. This allows for explicitly modeling the dynamics of BOLD response without any delay adjustment. To this end, we utilize a bidirectional LSTM, wherein, the input sequence is fed in normal time-order for one LSTM network, and in the reverse time-order, for another. The two hidden activations of forward and reverse directions in bi-LSTM are collated to build the "memory" of the model and are used to robustly predict the brain states at every time instance. Working memory data from the Human Connectome Project (HCP) is utilized for validation and was observed to perform 18% better than it's unidirectional counterpart in terms of accuracy in predicting the brain states.

通过学习辨别特征表征来解码潜在认知过程的大脑状态最近在脑成像研究中获得了广泛关注。特别是,通过分析 fMRI 数据中的时间信息来编码大脑功能的动态变化已成为一种推动力。长短期记忆(LSTM)是一类具有 "记忆 "成分的机器学习模型,可以保留以前看到的时间信息,越来越多的人观察到它在各种具有动态时间行为的应用中表现出色,包括大脑状态解码。由于 fMRI BOLD 反应的动态性和固有延迟性,未来的时间背景至关重要。然而,传统的 LSTM 模型既不能对其进行编码,也不能捕捉到它。本文通过双向 LSTM 对过去和未来的 fMRI 数据实例进行信息封装,从而实现稳健的大脑状态解码。这样就能明确地模拟 BOLD 响应的动态变化,而无需进行任何延迟调整。为此,我们使用了双向 LSTM,其中一个 LSTM 网络按正常时序输入输入序列,而另一个 LSTM 网络则按相反时序输入输入序列。在双向 LSTM 中,正向和反向的两个隐藏激活将被整理以建立模型的 "记忆",并用于稳健地预测每个时间实例的大脑状态。来自人类连接组项目(HCP)的工作记忆数据被用来进行验证,结果显示,在预测大脑状态的准确性方面,它比单向模型高出 18%。
{"title":"Robust Brain State Decoding using Bidirectional Long Short Term Memory Networks in functional MRI.","authors":"Anant Mittal, Priya Aggarwal, Luiz Pessoa, Anubha Gupta","doi":"10.1145/3490035.3490269","DOIUrl":"10.1145/3490035.3490269","url":null,"abstract":"<p><p>Decoding brain states of the underlying cognitive processes via learning discriminative feature representations has recently gained a lot of interest in brain imaging studies. Particularly, there has been an impetus to encode the dynamics of brain functioning by analyzing temporal information available in the fMRI data. Long-short term memory (LSTM), a class of machine learning model possessing a \"memory\" component, to retain previously seen temporal information, is increasingly being observed to perform well in various applications with dynamic temporal behavior, including brain state decoding. Because of the dynamics and inherent latency in fMRI BOLD responses, future temporal context is crucial. However, it is neither encoded nor captured by the conventional LSTM model. This paper performs robust brain state decoding via information encapsulation from both the past and future instances of fMRI data via bi-directional LSTM. This allows for explicitly modeling the dynamics of BOLD response without any delay adjustment. To this end, we utilize a bidirectional LSTM, wherein, the input sequence is fed in normal time-order for one LSTM network, and in the reverse time-order, for another. The two hidden activations of forward and reverse directions in bi-LSTM are collated to build the \"memory\" of the model and are used to robustly predict the brain states at every time instance. Working memory data from the Human Connectome Project (HCP) is utilized for validation and was observed to perform 18% better than it's unidirectional counterpart in terms of accuracy in predicting the brain states.</p>","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"2021 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9639335/pdf/nihms-1789845.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"40673961","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ICVGIP 2018: 11th Indian Conference on Computer Vision, Graphics and Image Processing, Hyderabad, India, 18-22 December, 2018 ICVGIP 2018:第11届印度计算机视觉,图形和图像处理会议,印度海得拉巴,2018年12月18日至22日
{"title":"ICVGIP 2018: 11th Indian Conference on Computer Vision, Graphics and Image Processing, Hyderabad, India, 18-22 December, 2018","authors":"","doi":"10.1145/3293353","DOIUrl":"https://doi.org/10.1145/3293353","url":null,"abstract":"","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"20 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2018-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81854985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Voxel-based morphometry and minimum redundancy maximum relevance method for classification of Parkinson's disease and controls from T1-weighted MRI 基于体素的形态测量和最小冗余最大关联方法在t1加权MRI中对帕金森病和对照进行分类
Pub Date : 2016-12-18 DOI: 10.1145/3009977.3009998
Bharti, A. Juneja, M. Saxena, S. Gudwani, S. Kumaran, R. Agrawal, M. Behari
Parkinson's disease (PD) is a neurodegenerative disorder, which needs to be accurately diagnosed in early stage. Voxel-based morphometry (VBM) has been extensively utilized to determine focal changes between PD patients and controls. However, it is not much utilized in differential diagnosis of an individual subject. Thus, in this study, VBM findings in conjunction with minimum redundancy maximum relevance (mRMR) method are utilized to obtain a set of relevant and non-redundant features for computer-aided diagnosis (CAD) of PD using T1-weighted MRI. In the proposed method, firstly, statistical features are extracted from the clusters obtained from statistical maps, generated using VBM, of gray matter (GM), white matter (WM) and cerebrospinal fluid (CSF) independently and their different combinations. Then mRMR, a multivariate feature selection method, is utilized to find a minimal set of relevant and non-redundant features. Finally, support vector machine is utilized to learn a decision model using the selected features. Experiments are performed on newly acquired T1-weighted MRI of 30 PD patients and 30 age & gender matched controls. The performance is evaluated using leave-one out cross-validation scheme in terms of sensitivity, specificity and classification accuracy. The maximum accuracy of 88.33% is achieved for GM+WM and GM+WM+CSF. In addition, the proposed method outperforms the existing methods. It is also observed that the selected clusters belong to regions namely middle and superior frontal gyrus for GM, inferior, middle frontal gyrus and insula for WM and lateral ventricle for CSF. Further, correlation of UPDRS/H&Y staging scale with GM/WM/CSF volume is observed to be not significant. Appreciable classification performance of the proposed method highlights the potential of the proposed method in CAD support system for the clinicians in PD diagnosis.
帕金森病(PD)是一种神经退行性疾病,需要在早期准确诊断。基于体素的形态学(VBM)已被广泛用于确定PD患者和对照组之间的病灶变化。然而,它在个体主体的鉴别诊断中并不常用。因此,在本研究中,VBM的发现与最小冗余最大相关性(mRMR)方法相结合,利用t1加权MRI获得一组相关和非冗余特征,用于PD的计算机辅助诊断(CAD)。该方法首先对脑灰质(GM)、脑白质(WM)和脑脊液(CSF)分别独立及不同组合的统计图进行聚类提取统计特征;然后,利用多变量特征选择方法mRMR寻找最小的相关和非冗余特征集。最后,利用支持向量机学习所选特征的决策模型。实验对30名PD患者和30名年龄和性别匹配的对照组进行了新获得的t1加权MRI。采用留一交叉验证方案,从灵敏度、特异性和分类准确性三个方面对其性能进行评估。GM+WM和GM+WM+CSF的准确率最高可达88.33%。此外,该方法优于现有方法。我们还观察到,所选择的脑簇分别属于GM的额上中回,WM的额下中回和脑岛,CSF的侧脑室。此外,UPDRS/H&Y分期与GM/WM/CSF体积的相关性不显著。所提出的方法的显著分类性能突出了所提出的方法在PD诊断的临床医生CAD支持系统中的潜力。
{"title":"Voxel-based morphometry and minimum redundancy maximum relevance method for classification of Parkinson's disease and controls from T1-weighted MRI","authors":"Bharti, A. Juneja, M. Saxena, S. Gudwani, S. Kumaran, R. Agrawal, M. Behari","doi":"10.1145/3009977.3009998","DOIUrl":"https://doi.org/10.1145/3009977.3009998","url":null,"abstract":"Parkinson's disease (PD) is a neurodegenerative disorder, which needs to be accurately diagnosed in early stage. Voxel-based morphometry (VBM) has been extensively utilized to determine focal changes between PD patients and controls. However, it is not much utilized in differential diagnosis of an individual subject. Thus, in this study, VBM findings in conjunction with minimum redundancy maximum relevance (mRMR) method are utilized to obtain a set of relevant and non-redundant features for computer-aided diagnosis (CAD) of PD using T1-weighted MRI. In the proposed method, firstly, statistical features are extracted from the clusters obtained from statistical maps, generated using VBM, of gray matter (GM), white matter (WM) and cerebrospinal fluid (CSF) independently and their different combinations. Then mRMR, a multivariate feature selection method, is utilized to find a minimal set of relevant and non-redundant features. Finally, support vector machine is utilized to learn a decision model using the selected features. Experiments are performed on newly acquired T1-weighted MRI of 30 PD patients and 30 age & gender matched controls. The performance is evaluated using leave-one out cross-validation scheme in terms of sensitivity, specificity and classification accuracy. The maximum accuracy of 88.33% is achieved for GM+WM and GM+WM+CSF. In addition, the proposed method outperforms the existing methods. It is also observed that the selected clusters belong to regions namely middle and superior frontal gyrus for GM, inferior, middle frontal gyrus and insula for WM and lateral ventricle for CSF. Further, correlation of UPDRS/H&Y staging scale with GM/WM/CSF volume is observed to be not significant. Appreciable classification performance of the proposed method highlights the potential of the proposed method in CAD support system for the clinicians in PD diagnosis.","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"4 1","pages":"22:1-22:6"},"PeriodicalIF":0.0,"publicationDate":"2016-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73493078","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A distribution-independent risk estimator for image denoising 一种与分布无关的图像去噪风险估计方法
Pub Date : 2016-12-18 DOI: 10.1145/3009977.3010025
Babu Kishore Subramanian, Ashutosh Gupta, C. Seelamantula
We address the problem of image denoising for an additive white noise model without placing any restrictions on the statistical distribution of noise. We assume knowledge of only the first- and second-order noise statistics. In the recent mean-square error (MSE) minimization approaches for image denoising, one considers a particular noise distribution and derives an expression for the unbiased risk estimate of the MSE. For additive white Gaussian noise, an unbiased estimate of the MSE is Stein's unbiased risk estimate (SURE), which relies on Stein's lemma. We derive an unbiased risk estimate without using Stein's lemma or its counterparts for additive white noise model irrespective of the noise distribution. We refer to the MSE estimate as the generic risk estimate (GenRE). We demonstrate the effectiveness of GenRE using shrinkage in the undecimated Haar wavelet transform domain as the denoising function. The estimated peak-signal-to-noise-ratio (PSNR) using GenRE is typically within 1% of the PSNR obtained when optimizing with the oracle MSE. The performance of the proposed method is on par with SURE for Gaussian noise distribution, and better than SURE-based methods for other noise distributions such as uniform and Laplacian distribution in terms of both PSNR and structural similarity (SSIM).
我们解决了一个加性白噪声模型的图像去噪问题,没有对噪声的统计分布施加任何限制。我们假设只知道一阶和二阶噪声统计量。在最近用于图像去噪的均方误差(MSE)最小化方法中,人们考虑特定的噪声分布并推导出MSE无偏风险估计的表达式。对于加性高斯白噪声,MSE的无偏估计是依赖于Stein引理的Stein无偏风险估计(SURE)。我们得到了一个无偏的风险估计,而不使用Stein引理或其对应的加性白噪声模型,而不考虑噪声分布。我们将MSE估计称为一般风险估计(类型)。我们使用未消差Haar小波变换域的收缩作为去噪函数,证明了GenRE的有效性。使用GenRE估计的峰值信噪比(PSNR)通常在使用oracle MSE优化时获得的PSNR的1%以内。对于高斯噪声分布,该方法的性能与SURE相当,对于均匀分布和拉普拉斯分布等其他噪声分布,该方法的PSNR和结构相似度(SSIM)均优于基于SURE的方法。
{"title":"A distribution-independent risk estimator for image denoising","authors":"Babu Kishore Subramanian, Ashutosh Gupta, C. Seelamantula","doi":"10.1145/3009977.3010025","DOIUrl":"https://doi.org/10.1145/3009977.3010025","url":null,"abstract":"We address the problem of image denoising for an additive white noise model without placing any restrictions on the statistical distribution of noise. We assume knowledge of only the first- and second-order noise statistics. In the recent mean-square error (MSE) minimization approaches for image denoising, one considers a particular noise distribution and derives an expression for the unbiased risk estimate of the MSE. For additive white Gaussian noise, an unbiased estimate of the MSE is Stein's unbiased risk estimate (SURE), which relies on Stein's lemma. We derive an unbiased risk estimate without using Stein's lemma or its counterparts for additive white noise model irrespective of the noise distribution. We refer to the MSE estimate as the generic risk estimate (GenRE). We demonstrate the effectiveness of GenRE using shrinkage in the undecimated Haar wavelet transform domain as the denoising function. The estimated peak-signal-to-noise-ratio (PSNR) using GenRE is typically within 1% of the PSNR obtained when optimizing with the oracle MSE. The performance of the proposed method is on par with SURE for Gaussian noise distribution, and better than SURE-based methods for other noise distributions such as uniform and Laplacian distribution in terms of both PSNR and structural similarity (SSIM).","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"220 1","pages":"52:1-52:8"},"PeriodicalIF":0.0,"publicationDate":"2016-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75624899","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Informed multimodal latent subspace learning via supervised matrix factorization 基于监督矩阵分解的多模态潜在子空间学习
Pub Date : 2016-12-18 DOI: 10.1145/3009977.3010012
Ramashish Gaurav, Mridula Verma, K. K. Shukla
Matrix factorization technique has been widely used as a popular method to learn a joint latent-compact subspace, when multiple views or modals of objects (belonging to single-domain or multiple-domain) are available. Our work confronts the problem of learning an informative latent subspace by imparting supervision to matrix factorization for fusing multiple modals of objects, where we devise simpler supervised additive updates instead of multiplicative updates, thus scalable to large scale datasets. To increase the classification accuracy we integrate the label information of images with the process of learning a semantically enhanced subspace. We perform extensive experiments on two publicly available standard image datasets of NUS WIDE and compare the results with state-of-the-art subspace learning and fusion techniques to evaluate the efficacy of our framework. Improvement obtained in the classification accuracy confirms the effectiveness of our approach. In essence, we propose a novel method for supervised data fusion thus leading to supervised subspace learning.
当对象的多个视图或模态(属于单域或多域)可用时,矩阵分解技术作为一种常用的学习联合隐紧子空间的方法得到了广泛的应用。我们的工作面临着学习信息潜在子空间的问题,通过为融合对象的多模态的矩阵分解赋予监督,其中我们设计了更简单的监督加性更新而不是乘法更新,从而可扩展到大规模数据集。为了提高分类精度,我们将图像的标签信息与学习语义增强子空间的过程相结合。我们在两个公开可用的NUS WIDE标准图像数据集上进行了广泛的实验,并将结果与最先进的子空间学习和融合技术进行比较,以评估我们框架的有效性。分类精度的提高证实了我们方法的有效性。从本质上讲,我们提出了一种新的监督数据融合方法,从而实现监督子空间学习。
{"title":"Informed multimodal latent subspace learning via supervised matrix factorization","authors":"Ramashish Gaurav, Mridula Verma, K. K. Shukla","doi":"10.1145/3009977.3010012","DOIUrl":"https://doi.org/10.1145/3009977.3010012","url":null,"abstract":"Matrix factorization technique has been widely used as a popular method to learn a joint latent-compact subspace, when multiple views or modals of objects (belonging to single-domain or multiple-domain) are available. Our work confronts the problem of learning an informative latent subspace by imparting supervision to matrix factorization for fusing multiple modals of objects, where we devise simpler supervised additive updates instead of multiplicative updates, thus scalable to large scale datasets. To increase the classification accuracy we integrate the label information of images with the process of learning a semantically enhanced subspace. We perform extensive experiments on two publicly available standard image datasets of NUS WIDE and compare the results with state-of-the-art subspace learning and fusion techniques to evaluate the efficacy of our framework. Improvement obtained in the classification accuracy confirms the effectiveness of our approach. In essence, we propose a novel method for supervised data fusion thus leading to supervised subspace learning.","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"76 1","pages":"36:1-36:8"},"PeriodicalIF":0.0,"publicationDate":"2016-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78214159","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Color restoration in turbid medium 在浑浊介质中恢复颜色
Pub Date : 2016-12-18 DOI: 10.1145/3009977.3010028
M. NimishaT., K. Seemakurthy, A. Rajagopalan, N. Vedachalam, Ramesh Raju
Light scattering and color distortions are two major issues with underwater imaging. Scattering occurs due to turbidity of the medium and color distortions are caused by differential attenuation of wavelengths as a function of depth. As a result, underwater images taken in a turbid medium have low contrast, color cast, and color loss. The main objective of this work is color restoration of underwater images i.e, produce its equivalent image as seen outside of the water surface. As a first step, we account for low contrast by employing dark channel prior based dehazing. These images are then color corrected by learning a mapping function between a pair of color chart images, one taken inside water and another taken outside. The mapping thus learned is with respect to a reference distance from the water surface. We also propose a color modulation scheme that is applied prior to color mapping to accommodate the same mapping function for different depths as well. Color restoration results are given on several images to validate the efficacy of the proposed methodology.
光散射和色彩失真是水下成像的两个主要问题。散射是由于介质的浑浊而发生的,颜色失真是由于波长作为深度的函数的微分衰减引起的。因此,在浑浊介质中拍摄的水下图像具有低对比度,偏色和色彩损失。这项工作的主要目的是水下图像的颜色恢复,即产生其等效的图像,看到水面外。作为第一步,我们通过使用暗通道先验去雾来解决对比度低的问题。然后通过学习一对彩色图表图像之间的映射函数来校正这些图像的颜色,其中一个在水中拍摄,另一个在室外拍摄。这样学习到的映射是关于到水面的参考距离。我们还提出了一种颜色调制方案,该方案在颜色映射之前应用,以适应不同深度的相同映射函数。给出了几幅图像的颜色恢复结果,以验证所提出方法的有效性。
{"title":"Color restoration in turbid medium","authors":"M. NimishaT., K. Seemakurthy, A. Rajagopalan, N. Vedachalam, Ramesh Raju","doi":"10.1145/3009977.3010028","DOIUrl":"https://doi.org/10.1145/3009977.3010028","url":null,"abstract":"Light scattering and color distortions are two major issues with underwater imaging. Scattering occurs due to turbidity of the medium and color distortions are caused by differential attenuation of wavelengths as a function of depth. As a result, underwater images taken in a turbid medium have low contrast, color cast, and color loss. The main objective of this work is color restoration of underwater images i.e, produce its equivalent image as seen outside of the water surface. As a first step, we account for low contrast by employing dark channel prior based dehazing. These images are then color corrected by learning a mapping function between a pair of color chart images, one taken inside water and another taken outside. The mapping thus learned is with respect to a reference distance from the water surface. We also propose a color modulation scheme that is applied prior to color mapping to accommodate the same mapping function for different depths as well. Color restoration results are given on several images to validate the efficacy of the proposed methodology.","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"14 1","pages":"26:1-26:8"},"PeriodicalIF":0.0,"publicationDate":"2016-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78483761","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A framework to assess Sun salutation videos 一个评估敬礼视频的框架
Pub Date : 2016-12-18 DOI: 10.1145/3009977.3010045
Hiteshi Jain, Gaurav Harit
There are many exercises which are repetitive in nature and are required to be done with perfection to derive maximum benefits. Sun Salutation or Surya Namaskar is one of the oldest yoga practice known. It is a sequence of ten actions or 'asanas' where the actions are synchronized with breathing and each action and its transition should be performed with minimal jerks. Essentially, it is important that this yoga practice be performed with Grace and Consistency. In this context, Grace is the ability of a person to perform an exercise with smoothness i.e. without sudden movements or jerks during the posture transition and Consistency measures the repeatability of an exercise in every cycle. We propose an algorithm that assesses how well a person practices Sun Salutation in terms of grace and consistency. Our approach works by training individual HMMs for each asana using STIP features[11] followed by automatic segmentation and labeling of the entire Sun Salutation sequence using a concatenated-HMM. The metric of grace and consistency are then laid down in terms of posture transition times. The assessments made by our system are compared with the assessments of the yoga trainer to derive the accuracy of the system. We introduce a dataset for Sun Salutation videos comprising 30 sequences of perfect Sun Salutation performed by seven experts and used this dataset to train our system. While Sun Salutation can be judged on multiple parameters, we focus mainly on judging Grace and Consistency.
有许多练习本质上是重复的,需要完美地完成才能获得最大的好处。拜日式是已知的最古老的瑜伽练习之一。它是由十个动作或体式组成的序列,动作与呼吸同步,每个动作及其过渡都应该以最小的抽搐来完成。从本质上讲,重要的是瑜伽练习要优雅和一致。在这种情况下,优雅是指一个人在姿势转换过程中平稳地进行锻炼的能力,即没有突然的动作或抽搐,而一致性是指在每个周期中锻炼的可重复性。我们提出了一种算法来评估一个人在优雅和一致性方面练习太阳敬礼的程度。我们的方法是通过使用STIP特征[11]为每个体式训练单个hmm,然后使用连接的hmm对整个日式序列进行自动分割和标记。优雅和一致性的度量是根据姿势转换的时间来确定的。将系统的评估结果与瑜伽教练的评估结果进行比较,得出系统的准确性。我们引入了一个由7位专家表演的30个完美的敬礼视频组成的数据集,并使用该数据集来训练我们的系统。虽然Sun Salutation可以从多个参数来判断,但我们主要关注的是判断Grace和Consistency。
{"title":"A framework to assess Sun salutation videos","authors":"Hiteshi Jain, Gaurav Harit","doi":"10.1145/3009977.3010045","DOIUrl":"https://doi.org/10.1145/3009977.3010045","url":null,"abstract":"There are many exercises which are repetitive in nature and are required to be done with perfection to derive maximum benefits. Sun Salutation or Surya Namaskar is one of the oldest yoga practice known. It is a sequence of ten actions or 'asanas' where the actions are synchronized with breathing and each action and its transition should be performed with minimal jerks. Essentially, it is important that this yoga practice be performed with Grace and Consistency. In this context, Grace is the ability of a person to perform an exercise with smoothness i.e. without sudden movements or jerks during the posture transition and Consistency measures the repeatability of an exercise in every cycle. We propose an algorithm that assesses how well a person practices Sun Salutation in terms of grace and consistency. Our approach works by training individual HMMs for each asana using STIP features[11] followed by automatic segmentation and labeling of the entire Sun Salutation sequence using a concatenated-HMM. The metric of grace and consistency are then laid down in terms of posture transition times. The assessments made by our system are compared with the assessments of the yoga trainer to derive the accuracy of the system. We introduce a dataset for Sun Salutation videos comprising 30 sequences of perfect Sun Salutation performed by seven experts and used this dataset to train our system. While Sun Salutation can be judged on multiple parameters, we focus mainly on judging Grace and Consistency.","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"43 1","pages":"29:1-29:8"},"PeriodicalIF":0.0,"publicationDate":"2016-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86528823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Towards quantifying the amount of uncollected garbage through image analysis 通过图像分析来量化未收集垃圾的数量
Pub Date : 2016-12-18 DOI: 10.1145/3009977.3010061
Susheel Suresh, Tarun Sharma, K. PrashanthT., V. Subramaniam, D. Sitaram, M. Nirupama
Civic authorities in many Indian cities have a tough time in garbage collection and as a result there is a pile up of garbage in the cities. In order to manage the situation, it is first required to be able to quantify the issue. In this paper, we address the problem of quantification of garbage in a dump using a two step approach. In the first step, we build a mobile application that allows citizens to capture images of garbage and upload them to a server. In the second step, back-end performs analysis on these images to estimate the amount of garbage using computer vision techniques. Our approach to volume estimation uses multiple images of the same dump (provided by the mobile application) from different perspectives, segments the dump from the background, reconstructs a three dimensional view of the dump and then estimates its volume. Using our novel pipeline, our experiments indicate that with 8 different perspectives, we are able to achieve an accuracy of about 85 % for estimating the volume.
印度许多城市的市政当局在垃圾收集方面遇到了困难,因此城市里有一堆垃圾。为了管理这种情况,首先需要能够量化问题。在本文中,我们使用两步方法解决了垃圾场垃圾的量化问题。在第一步中,我们构建一个移动应用程序,允许公民捕获垃圾图像并将其上传到服务器。在第二步中,后端使用计算机视觉技术对这些图像进行分析以估计垃圾的数量。我们的体积估计方法是从不同的角度使用同一转储的多个图像(由移动应用程序提供),从背景中分割转储,重建转储的三维视图,然后估计其体积。使用我们的新管道,我们的实验表明,在8个不同的角度下,我们能够达到约85%的准确度来估计体积。
{"title":"Towards quantifying the amount of uncollected garbage through image analysis","authors":"Susheel Suresh, Tarun Sharma, K. PrashanthT., V. Subramaniam, D. Sitaram, M. Nirupama","doi":"10.1145/3009977.3010061","DOIUrl":"https://doi.org/10.1145/3009977.3010061","url":null,"abstract":"Civic authorities in many Indian cities have a tough time in garbage collection and as a result there is a pile up of garbage in the cities. In order to manage the situation, it is first required to be able to quantify the issue. In this paper, we address the problem of quantification of garbage in a dump using a two step approach. In the first step, we build a mobile application that allows citizens to capture images of garbage and upload them to a server. In the second step, back-end performs analysis on these images to estimate the amount of garbage using computer vision techniques. Our approach to volume estimation uses multiple images of the same dump (provided by the mobile application) from different perspectives, segments the dump from the background, reconstructs a three dimensional view of the dump and then estimates its volume. Using our novel pipeline, our experiments indicate that with 8 different perspectives, we are able to achieve an accuracy of about 85 % for estimating the volume.","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"26 1","pages":"73:1-73:8"},"PeriodicalIF":0.0,"publicationDate":"2016-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82623766","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Generating synthetic handwriting using n-gram letter glyphs 生成合成手写使用n-gram字母字形
Pub Date : 2016-12-18 DOI: 10.1145/3009977.3010042
Arka Ujjal Dey, Gaurav Harit
We propose a framework for synthesis of natural semi cursive handwritten Latin script that can find application in text personalization, or in generation of synthetic data for recognition systems. Our method is based on the generation of synthetic n-gram letter glyphs and their subsequent concatenation. We propose a non-parametric data driven generation scheme that is able to mimic the variation observed in handwritten glyph samples to synthesize natural looking synthetic glyphs. These synthetic glyphs are then stitched together to form complete words, using a spline based concatenation scheme. Further, as a refinement, our method is able to generate pen-lifts, giving our results a natural semi-cursive look. Through subjective experiments and detailed analysis of the results, we demonstrate the effectiveness of our formulation in being able to generate natural looking synthetic script.
我们提出了一个自然半草书手写拉丁文字的合成框架,该框架可以在文本个性化或识别系统合成数据的生成中找到应用。我们的方法是基于合成n-gram字母符号的生成及其随后的连接。我们提出了一种非参数数据驱动的生成方案,该方案能够模拟在手写字形样本中观察到的变化,以合成自然外观的合成字形。然后使用基于样条的连接方案将这些合成的符号拼接在一起形成完整的单词。此外,作为改进,我们的方法能够生成笔画,使我们的结果具有自然的半草书外观。通过主观实验和对结果的详细分析,我们证明了我们的配方能够生成自然外观的合成脚本的有效性。
{"title":"Generating synthetic handwriting using n-gram letter glyphs","authors":"Arka Ujjal Dey, Gaurav Harit","doi":"10.1145/3009977.3010042","DOIUrl":"https://doi.org/10.1145/3009977.3010042","url":null,"abstract":"We propose a framework for synthesis of natural semi cursive handwritten Latin script that can find application in text personalization, or in generation of synthetic data for recognition systems. Our method is based on the generation of synthetic n-gram letter glyphs and their subsequent concatenation. We propose a non-parametric data driven generation scheme that is able to mimic the variation observed in handwritten glyph samples to synthesize natural looking synthetic glyphs. These synthetic glyphs are then stitched together to form complete words, using a spline based concatenation scheme. Further, as a refinement, our method is able to generate pen-lifts, giving our results a natural semi-cursive look. Through subjective experiments and detailed analysis of the results, we demonstrate the effectiveness of our formulation in being able to generate natural looking synthetic script.","PeriodicalId":93806,"journal":{"name":"Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing","volume":"27 1","pages":"49:1-49:8"},"PeriodicalIF":0.0,"publicationDate":"2016-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89111162","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Proceedings. Indian Conference on Computer Vision, Graphics & Image Processing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1