首页 > 最新文献

2019 15th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS)最新文献

英文 中文
Machine Learning Based Detection of Hearing Loss Using Auditory Perception Responses 基于听觉感知反应的机器学习听力损失检测
Muhammad Ilyas, A. Naït-Ali
Hearing loss or hearing impairment is the primary reason of deafness throughout the world. Hearing impairment can occur to one or both the ears. If hearing loss is identified in time, it can be minimized by practicing specific precautions. In this paper, we investigate the likelihood of detection of hearing loss through auditory system responses. Auditory perception and human age are highly interrelated. Likewise, detecting a significant gap within the real age and the estimated age, the hearing loss can easily be identified. Our proposed system for human age estimation has promising results with a Root Mean Square Error (RMSE) value of 4.1 years, and classification performance efficiency for hearing loss is 94%, showing the applicability of our approach for detection of hearing loss.
听力损失或听力障碍是全世界耳聋的主要原因。听力障碍可能发生在一只耳朵或两只耳朵上。如果及时发现听力损失,可以通过采取具体的预防措施将其降到最低。在本文中,我们研究了通过听觉系统反应检测听力损失的可能性。听觉和人类年龄是高度相关的。同样,如果在实际年龄和估计年龄之间发现明显的差距,听力损失也很容易被识别出来。我们提出的人类年龄估计系统取得了令人满意的结果,其均方根误差(RMSE)值为4.1岁,听力损失分类性能效率为94%,表明了我们的方法在听力损失检测中的适用性。
{"title":"Machine Learning Based Detection of Hearing Loss Using Auditory Perception Responses","authors":"Muhammad Ilyas, A. Naït-Ali","doi":"10.1109/SITIS.2019.00034","DOIUrl":"https://doi.org/10.1109/SITIS.2019.00034","url":null,"abstract":"Hearing loss or hearing impairment is the primary reason of deafness throughout the world. Hearing impairment can occur to one or both the ears. If hearing loss is identified in time, it can be minimized by practicing specific precautions. In this paper, we investigate the likelihood of detection of hearing loss through auditory system responses. Auditory perception and human age are highly interrelated. Likewise, detecting a significant gap within the real age and the estimated age, the hearing loss can easily be identified. Our proposed system for human age estimation has promising results with a Root Mean Square Error (RMSE) value of 4.1 years, and classification performance efficiency for hearing loss is 94%, showing the applicability of our approach for detection of hearing loss.","PeriodicalId":301876,"journal":{"name":"2019 15th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS)","volume":"85 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115683679","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Perceived Effects of Static and Dynamic Sparkle in Captured Effect Coatings 俘获效应涂层中静态和动态闪光的感知效应
J. Filip, M. Kolafová, R. Vávra
Quality control applications in the coating industry characterize visual properties of coatings containing effect pigments using glint impression, often denoted as sparkle. They rely on a collection of static images capturing sparkle properties of pigment flakes. However, visual characteristics of pigment flakes are highly correlated to their material properties and their orientations in coating layers. Thus, while two effect coatings can exhibit similar static sparkle behavior, their dynamic sparkle behavior may be very distinct. In this paper, we analyzed the perception of static and dynamic sparkle using two psychophysical studies on 38 effect coatings and 31 human subjects. First, we have shown a good agreement between the perception of sparkle in real specimens and in photographs. Second, we observed significant differences in perceived static and dynamic sparkle. Our results demonstrate a need for a multiangle recording of sparkle when assessing effect pigment visual characteristics.
涂料工业中的质量控制应用是利用闪烁印象来表征含有效果颜料的涂料的视觉特性,通常被称为闪光。他们依靠一组静态图像来捕捉颜料薄片的闪光特性。然而,颜料薄片的视觉特性与其材料性质及其在涂层中的取向高度相关。因此,虽然两种效果涂层可以表现出相似的静态火花行为,但它们的动态火花行为可能非常不同。本文通过对38种效果涂层和31名人体受试者的两项心理物理研究,分析了静态和动态闪光的感知。首先,我们展示了真实标本和照片中对闪光的感知之间的良好一致性。其次,我们观察到静态和动态闪光的感知差异显著。我们的结果表明,在评估效果颜料的视觉特性时,需要多角度记录闪光。
{"title":"Perceived Effects of Static and Dynamic Sparkle in Captured Effect Coatings","authors":"J. Filip, M. Kolafová, R. Vávra","doi":"10.1109/SITIS.2019.00119","DOIUrl":"https://doi.org/10.1109/SITIS.2019.00119","url":null,"abstract":"Quality control applications in the coating industry characterize visual properties of coatings containing effect pigments using glint impression, often denoted as sparkle. They rely on a collection of static images capturing sparkle properties of pigment flakes. However, visual characteristics of pigment flakes are highly correlated to their material properties and their orientations in coating layers. Thus, while two effect coatings can exhibit similar static sparkle behavior, their dynamic sparkle behavior may be very distinct. In this paper, we analyzed the perception of static and dynamic sparkle using two psychophysical studies on 38 effect coatings and 31 human subjects. First, we have shown a good agreement between the perception of sparkle in real specimens and in photographs. Second, we observed significant differences in perceived static and dynamic sparkle. Our results demonstrate a need for a multiangle recording of sparkle when assessing effect pigment visual characteristics.","PeriodicalId":301876,"journal":{"name":"2019 15th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS)","volume":"33 ","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120973704","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Gaussian Recursive Filter Parallel Implementation with Overlapping 一种具有重叠的高斯递归滤波器并行实现
P. D. Luca, A. Galletti, L. Marcellino
Gaussian convolutions computation is required in several scientific fields and, to this aim, efficient approximation methods, based on Recursive Filters (RFs), have been developed recently. Among them, Gaussian Recursive Filters (RFs) are designed to approximate the Gaussian convolution in a very efficient way. The accuracy of these methods, as is well known, can be improved by means of the use of the so-called K-iterated Gaussian recursive filters, that is in the repeated application of the basic RF. To improve the provided accuracy, K-iterated versions of these methods are also considered. Since it is often necessary to handle large size one-dimensional input signals, a parallel approach becomes mandatory. Recently, we proposed a parallel algorithm for the implementation of the K-iterated first-order Gaussian RF on multicore architectures. Here, using a similar parallelization strategy, based on a domain decomposition with overlapping, we propose a new implementation that would exploit, in terms of both accuracy and performance, the GPU (Graphics Processing Unit) capabilities on CUDA environment. Tests and experiments confirm the reliability and the efficiency of the proposed implementation.
在一些科学领域需要高斯卷积计算,为了达到这个目的,最近发展了基于递归滤波器(RFs)的有效逼近方法。其中,高斯递归滤波器(RFs)的设计是为了以一种非常有效的方式近似高斯卷积。众所周知,这些方法的准确性可以通过使用所谓的k -迭代高斯递归滤波器来提高,即在基本RF的重复应用中。为了提高提供的准确性,还考虑了这些方法的k迭代版本。由于经常需要处理大尺寸的一维输入信号,因此必须采用并行方法。最近,我们提出了一种在多核架构上实现k迭代一阶高斯射频的并行算法。在这里,使用类似的并行化策略,基于重叠的域分解,我们提出了一种新的实现,可以在精度和性能方面利用CUDA环境下GPU(图形处理单元)的能力。测试和实验验证了该方法的可靠性和有效性。
{"title":"A Gaussian Recursive Filter Parallel Implementation with Overlapping","authors":"P. D. Luca, A. Galletti, L. Marcellino","doi":"10.1109/SITIS.2019.00105","DOIUrl":"https://doi.org/10.1109/SITIS.2019.00105","url":null,"abstract":"Gaussian convolutions computation is required in several scientific fields and, to this aim, efficient approximation methods, based on Recursive Filters (RFs), have been developed recently. Among them, Gaussian Recursive Filters (RFs) are designed to approximate the Gaussian convolution in a very efficient way. The accuracy of these methods, as is well known, can be improved by means of the use of the so-called K-iterated Gaussian recursive filters, that is in the repeated application of the basic RF. To improve the provided accuracy, K-iterated versions of these methods are also considered. Since it is often necessary to handle large size one-dimensional input signals, a parallel approach becomes mandatory. Recently, we proposed a parallel algorithm for the implementation of the K-iterated first-order Gaussian RF on multicore architectures. Here, using a similar parallelization strategy, based on a domain decomposition with overlapping, we propose a new implementation that would exploit, in terms of both accuracy and performance, the GPU (Graphics Processing Unit) capabilities on CUDA environment. Tests and experiments confirm the reliability and the efficiency of the proposed implementation.","PeriodicalId":301876,"journal":{"name":"2019 15th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS)","volume":"55 21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115351211","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Low-Light Image Enhancement via Adaptive Shape and Texture Prior 通过自适应形状和纹理先验的弱光图像增强
Kazuki Kurihara, Hiromi Yoshida, Y. Iiguni
Low light images affect various computer vision algorithms due to their low visibility and much noise hidden in dark regions. Although many methods based on the Retinex theory, which decomposes an observed image into the reflectance and illumination, have been proposed to alleviate the problem, existing methods inevitably cause under-and over-enhancement. In this paper, we propose a new joint optimization equation that sufficiently considers the features of both reflectance and illumination. More concretely, we adopt L2-Lp norm regularization terms to estimate the reflectance as much as possible to preserve details and textures, and the illumination as much as possible to preserve the structure information with texture-less. We solve the optimization equation in an alternating minimization method. Furthermore, we introduce a new adaptive texture prior to reveal more details and textures with noise reduction on both bright and dark regions. Experimental results, including qualitative and quantitative evaluations, show that the proposed method can establish a better performance than the other state-of-the-art methods.
弱光图像由于其低可见度和隐藏在黑暗区域的大量噪声,影响了各种计算机视觉算法。虽然已经提出了许多基于Retinex理论(将观测图像分解为反射率和照度)的方法来缓解这一问题,但现有的方法不可避免地会造成增强不足和过度增强。在本文中,我们提出了一个新的联合优化方程,充分考虑了反射率和光照的特征。具体地说,我们采用L2-Lp范数正则化项来估计尽可能多的反射率以保留细节和纹理,并且尽可能多的照度以保留无纹理的结构信息。我们用交替极小化法求解优化方程。此外,我们引入了一种新的自适应纹理,以显示更多的细节和纹理,在明亮和黑暗区域都有降噪。实验结果,包括定性和定量评价,表明该方法可以建立更好的性能比其他先进的方法。
{"title":"Low-Light Image Enhancement via Adaptive Shape and Texture Prior","authors":"Kazuki Kurihara, Hiromi Yoshida, Y. Iiguni","doi":"10.1109/SITIS.2019.00024","DOIUrl":"https://doi.org/10.1109/SITIS.2019.00024","url":null,"abstract":"Low light images affect various computer vision algorithms due to their low visibility and much noise hidden in dark regions. Although many methods based on the Retinex theory, which decomposes an observed image into the reflectance and illumination, have been proposed to alleviate the problem, existing methods inevitably cause under-and over-enhancement. In this paper, we propose a new joint optimization equation that sufficiently considers the features of both reflectance and illumination. More concretely, we adopt L2-Lp norm regularization terms to estimate the reflectance as much as possible to preserve details and textures, and the illumination as much as possible to preserve the structure information with texture-less. We solve the optimization equation in an alternating minimization method. Furthermore, we introduce a new adaptive texture prior to reveal more details and textures with noise reduction on both bright and dark regions. Experimental results, including qualitative and quantitative evaluations, show that the proposed method can establish a better performance than the other state-of-the-art methods.","PeriodicalId":301876,"journal":{"name":"2019 15th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS)","volume":"215 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122459919","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Time Unification on Local Binary Patterns Three Orthogonal Planes for Facial Expression Recognition 局部二值模式三正交平面的时间统一面部表情识别
Reda Belaiche, C. Migniot, D. Ginhac, Fan Yang
Machine learning has known a tremendous growth within the last years, and lately, thanks to that, some computer vision algorithms started to access what is difficult or even impossible to perceive by the human eye. While deep learning based computer vision algorithms have made themselves more and more present in the recent years, more classical feature extraction methods, such as the ones based on Local Binary Patterns (LBP), still present a non negligible interest, especially when dealing with small datasets. Furthermore, this operator has proven to be quite useful for facial emotions and human gestures recognition in general. Micro-Expression (ME) classification is among the applications of computer vision that heavily relied on hand crafted features in the past years. LBP Three Orthogonal Planes (LBP_TOP) is one of the most used hand crafted features extractor in the scientific literature to tackle the problem of ME classification. In this paper we present a time unification method that provides better results than the classical LBP_TOP while also drastically reducing the calculations required for feature extraction.
机器学习在过去几年里有了巨大的发展,最近,由于这一点,一些计算机视觉算法开始访问人眼难以甚至不可能感知的东西。虽然基于深度学习的计算机视觉算法近年来越来越多地出现,但更经典的特征提取方法,如基于局部二值模式(LBP)的特征提取方法,仍然存在不可忽视的兴趣,特别是在处理小数据集时。此外,该算子已被证明对面部情绪和人类手势识别非常有用。在过去的几年里,微表情(ME)分类是计算机视觉的应用之一,严重依赖于手工制作的特征。LBP三正交平面(LBP_TOP)是科学文献中最常用的手工特征提取器之一,用于解决ME分类问题。在本文中,我们提出了一种时间统一方法,它提供了比经典LBP_TOP更好的结果,同时也大大减少了特征提取所需的计算量。
{"title":"Time Unification on Local Binary Patterns Three Orthogonal Planes for Facial Expression Recognition","authors":"Reda Belaiche, C. Migniot, D. Ginhac, Fan Yang","doi":"10.1109/SITIS.2019.00076","DOIUrl":"https://doi.org/10.1109/SITIS.2019.00076","url":null,"abstract":"Machine learning has known a tremendous growth within the last years, and lately, thanks to that, some computer vision algorithms started to access what is difficult or even impossible to perceive by the human eye. While deep learning based computer vision algorithms have made themselves more and more present in the recent years, more classical feature extraction methods, such as the ones based on Local Binary Patterns (LBP), still present a non negligible interest, especially when dealing with small datasets. Furthermore, this operator has proven to be quite useful for facial emotions and human gestures recognition in general. Micro-Expression (ME) classification is among the applications of computer vision that heavily relied on hand crafted features in the past years. LBP Three Orthogonal Planes (LBP_TOP) is one of the most used hand crafted features extractor in the scientific literature to tackle the problem of ME classification. In this paper we present a time unification method that provides better results than the classical LBP_TOP while also drastically reducing the calculations required for feature extraction.","PeriodicalId":301876,"journal":{"name":"2019 15th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS)","volume":"63 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121674082","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Benchmarking The Imbalanced Behavior of Deep Learning Based Optical Flow Estimators 基于深度学习的光流估计器不平衡行为的基准测试
Stefano Savian, Mehdi Elahi, T. Tillo
Optical Flow (OF) estimation is an important task which could be effectively used for a variety of Computer Vision (CV) applications. While a range of techniques have been already proposed, however accurately estimating the OF is still a very challenging task. The most recent approaches for OF estimation exploit advanced Deep Learning techniques which have resulted in a shift in the paradigm. These techniques have shown substantial improvements particularly in the case of large displacements, brightness change, and non-rigid motion. However, these approaches are data-driven and hence they can be biased towards the specific training data, which could in turn lead to considerable inaccuracy of the estimated OF. In this paper, we address this problem and provide a novel benchmark that can be used to identify and to measure the bias magnitude of the OF estimation. We have performed several experiments based on public datasets (Monkaa and Sintel) as well as on data designed on purpose 1. The results have shown that OF estimation based on some of the state-of-the-art deep learning techniques strongly depend on factors such as motion orientation within the data. Indeed, we have observed substantial degree of bias toward certain directions of motion. The framework can help researchers and practitioners in order to develop more effective data augmentation techniques and training schedules for deep learning based optical flow estimators.
光流估计是一项重要的任务,可以有效地用于各种计算机视觉应用。虽然已经提出了一系列技术,但是准确估计of仍然是一项非常具有挑战性的任务。最新的OF估计方法利用了先进的深度学习技术,这导致了范式的转变。这些技术已经显示出实质性的改进,特别是在大位移、亮度变化和非刚性运动的情况下。然而,这些方法是数据驱动的,因此它们可能偏向于特定的训练数据,这反过来可能导致估计of的相当不准确。在本文中,我们解决了这个问题,并提供了一个新的基准,可用于识别和测量of估计的偏差幅度。我们已经基于公共数据集(Monkaa和Sintel)以及专门设计的数据进行了几次实验。结果表明,基于一些最先进的深度学习技术的OF估计强烈依赖于数据中的运动方向等因素。事实上,我们已经观察到对某些运动方向有相当程度的偏向。该框架可以帮助研究人员和从业者为基于深度学习的光流估计器开发更有效的数据增强技术和训练计划。
{"title":"Benchmarking The Imbalanced Behavior of Deep Learning Based Optical Flow Estimators","authors":"Stefano Savian, Mehdi Elahi, T. Tillo","doi":"10.1109/SITIS.2019.00035","DOIUrl":"https://doi.org/10.1109/SITIS.2019.00035","url":null,"abstract":"Optical Flow (OF) estimation is an important task which could be effectively used for a variety of Computer Vision (CV) applications. While a range of techniques have been already proposed, however accurately estimating the OF is still a very challenging task. The most recent approaches for OF estimation exploit advanced Deep Learning techniques which have resulted in a shift in the paradigm. These techniques have shown substantial improvements particularly in the case of large displacements, brightness change, and non-rigid motion. However, these approaches are data-driven and hence they can be biased towards the specific training data, which could in turn lead to considerable inaccuracy of the estimated OF. In this paper, we address this problem and provide a novel benchmark that can be used to identify and to measure the bias magnitude of the OF estimation. We have performed several experiments based on public datasets (Monkaa and Sintel) as well as on data designed on purpose 1. The results have shown that OF estimation based on some of the state-of-the-art deep learning techniques strongly depend on factors such as motion orientation within the data. Indeed, we have observed substantial degree of bias toward certain directions of motion. The framework can help researchers and practitioners in order to develop more effective data augmentation techniques and training schedules for deep learning based optical flow estimators.","PeriodicalId":301876,"journal":{"name":"2019 15th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124319970","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Novel Approach to Detect Outer Retinal Tubulation Using U-Net in SD-OCT Images 一种基于SD-OCT图像U-Net检测视网膜外管状的新方法
István Megyeri, Melinda Katona, L. G. Nyúl
Optical Coherence Tomography (OCT) has become a basic non-invasive tool in diagnosing and following different types of eye diseases. This technique can produce high-resolution cross-sectional images of retinal layers. Outer retinal tubulation (ORT) is one of the detectable biomarker by SD-OCT. ORTs defined as hyporeflective, tubular structures with hyperreflective borders or reversed within the retina and appear in many retinal diseases, including age-related macular degeneration (AMD). Our aim is to develop an automatic method that can efficiently characterize ORT biomarker. Detection of this biomarker can be challenging because of its variable size, location, and reflectivity. In this paper, we present a fully convolutional U-Net based architecture to detect ORT. The proposed approach is evaluated using a dataset annotated by ophthalmologists. One of the main challenges was the limited amount of training data that we resolve with real-time augmentation during training and using nested cross-validation. Our method achieved near human performance reaching an overall object-based recall score of 0.847 and Dice score of 0.579 on the test set.
光学相干层析成像(OCT)已成为诊断和跟踪各类眼病的基本无创工具。这种技术可以产生视网膜层的高分辨率横切面图像。视网膜外管化(ORT)是SD-OCT可检测的生物标志物之一。ORTs被定义为视网膜内具有高反射边界或反转的低反射管状结构,出现在许多视网膜疾病中,包括年龄相关性黄斑变性(AMD)。我们的目标是开发一种能够有效表征ORT生物标志物的自动方法。由于这种生物标志物的大小、位置和反射率不同,检测起来很有挑战性。在本文中,我们提出了一个基于全卷积U-Net的检测ORT的体系结构。使用由眼科医生注释的数据集对所提出的方法进行评估。其中一个主要的挑战是训练数据的数量有限,我们在训练期间通过实时增强和使用嵌套交叉验证来解决这个问题。我们的方法达到了接近人类的性能,在测试集中达到了基于对象的召回总分0.847,Dice得分0.579。
{"title":"A Novel Approach to Detect Outer Retinal Tubulation Using U-Net in SD-OCT Images","authors":"István Megyeri, Melinda Katona, L. G. Nyúl","doi":"10.1109/SITIS.2019.00096","DOIUrl":"https://doi.org/10.1109/SITIS.2019.00096","url":null,"abstract":"Optical Coherence Tomography (OCT) has become a basic non-invasive tool in diagnosing and following different types of eye diseases. This technique can produce high-resolution cross-sectional images of retinal layers. Outer retinal tubulation (ORT) is one of the detectable biomarker by SD-OCT. ORTs defined as hyporeflective, tubular structures with hyperreflective borders or reversed within the retina and appear in many retinal diseases, including age-related macular degeneration (AMD). Our aim is to develop an automatic method that can efficiently characterize ORT biomarker. Detection of this biomarker can be challenging because of its variable size, location, and reflectivity. In this paper, we present a fully convolutional U-Net based architecture to detect ORT. The proposed approach is evaluated using a dataset annotated by ophthalmologists. One of the main challenges was the limited amount of training data that we resolve with real-time augmentation during training and using nested cross-validation. Our method achieved near human performance reaching an overall object-based recall score of 0.847 and Dice score of 0.579 on the test set.","PeriodicalId":301876,"journal":{"name":"2019 15th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124320451","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
MeltdownCrisis: Dataset of Autistic Children During Meltdown Crisis 熔毁危机:熔毁危机期间自闭症儿童的数据集
Marwa Masmoudi, Salma Kammoun Jarraya, Mohamed Hammami
No one refutes the importance of datasets in the development of any new approach. Despite their importance, datasets in computer vision remain insufficient for some applications. Presently, very limited autism datasets associated with clinical tests or screening are available, and most of them are genetic in nature. However, there is no database that combines both the abnormal facial expressions and the aggressive behaviors of an autistic child during Meltdown crisis. This paper introduces a Meltdown Crisis, a new and rich dataset that can be used for the evaluation/development of computer vision-based applications pertinent to children who suffer of autism as security tool, e.g. Meltdown crisis detection. In particular, the "MeltdownCrisis" dataset includes video streams captured with Kinect which offers a wide range of visual information. It is divided on a facial expressions data and physical activities data. The current "MeltdownCrisis " dataset version covers several Meltdown crisis scenarios of autistic children along various normal state scenarios grouped into one set. Each scenario is represented through a rich set of features that can be extracted from Kinect camera.
没有人否认数据集在任何新方法发展中的重要性。尽管数据集在计算机视觉中的重要性,但在某些应用中仍然不足。目前,与临床测试或筛查相关的自闭症数据集非常有限,而且大多数数据本质上是遗传的。然而,目前还没有一个数据库可以将自闭症儿童在崩溃危机期间的异常面部表情和攻击行为结合起来。本文介绍了一个新的和丰富的数据集,可用于评估/开发与自闭症儿童相关的基于计算机视觉的应用程序作为安全工具,例如熔解危机检测。特别是,“MeltdownCrisis”数据集包括用Kinect捕获的视频流,它提供了广泛的视觉信息。它分为面部表情数据和身体活动数据。当前的“崩溃危机”数据集版本涵盖了自闭症儿童的几个崩溃危机场景,以及各种正常状态的场景。每个场景都通过一组丰富的特征来表示,这些特征可以从Kinect摄像头中提取出来。
{"title":"MeltdownCrisis: Dataset of Autistic Children During Meltdown Crisis","authors":"Marwa Masmoudi, Salma Kammoun Jarraya, Mohamed Hammami","doi":"10.1109/SITIS.2019.00048","DOIUrl":"https://doi.org/10.1109/SITIS.2019.00048","url":null,"abstract":"No one refutes the importance of datasets in the development of any new approach. Despite their importance, datasets in computer vision remain insufficient for some applications. Presently, very limited autism datasets associated with clinical tests or screening are available, and most of them are genetic in nature. However, there is no database that combines both the abnormal facial expressions and the aggressive behaviors of an autistic child during Meltdown crisis. This paper introduces a Meltdown Crisis, a new and rich dataset that can be used for the evaluation/development of computer vision-based applications pertinent to children who suffer of autism as security tool, e.g. Meltdown crisis detection. In particular, the \"MeltdownCrisis\" dataset includes video streams captured with Kinect which offers a wide range of visual information. It is divided on a facial expressions data and physical activities data. The current \"MeltdownCrisis \" dataset version covers several Meltdown crisis scenarios of autistic children along various normal state scenarios grouped into one set. Each scenario is represented through a rich set of features that can be extracted from Kinect camera.","PeriodicalId":301876,"journal":{"name":"2019 15th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS)","volume":"179 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124475902","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
On the Detection of Video's Ethnic Vietnamese Thai Dance Movements 录像中越南民族泰式舞蹈动作的检测
Tung Pham Thanh, S. Benferhat, M. Chau, T. Ma, Karim Tabia, L. T. Ha
The problem addressed in this paper is the one of classifying Vietnamese dances' videos. In particular, we focus on an automatic detection of movements in the Ethnic Vietnamese Thai dances (ETVD). We first propose an ontology-based description of ETVD movements in terms of main movements' steps. We then associate with each movement step a profile containing typical features that characterize a movement step. The automatic detection of ETVD movements is based on a correlation method that matches movements' steps profiles with concepts present in frames of dances' videos. The last part of the paper contain experimental studies that show the good classification rate of our ETVD movement detection method.
本文研究的问题是越南舞蹈视频的分类问题。特别是,我们专注于在越南民族泰国舞蹈(ETVD)运动的自动检测。我们首先提出了一种基于本体的主要运动步骤的ETVD运动描述。然后,我们将每个移动步骤关联到包含表征移动步骤的典型特征的配置文件。ETVD运动的自动检测基于一种关联方法,该方法将运动的步骤轮廓与舞蹈视频帧中的概念相匹配。论文的最后一部分是实验研究,表明我们的ETVD运动检测方法具有良好的分类率。
{"title":"On the Detection of Video's Ethnic Vietnamese Thai Dance Movements","authors":"Tung Pham Thanh, S. Benferhat, M. Chau, T. Ma, Karim Tabia, L. T. Ha","doi":"10.1109/SITIS.2019.00064","DOIUrl":"https://doi.org/10.1109/SITIS.2019.00064","url":null,"abstract":"The problem addressed in this paper is the one of classifying Vietnamese dances' videos. In particular, we focus on an automatic detection of movements in the Ethnic Vietnamese Thai dances (ETVD). We first propose an ontology-based description of ETVD movements in terms of main movements' steps. We then associate with each movement step a profile containing typical features that characterize a movement step. The automatic detection of ETVD movements is based on a correlation method that matches movements' steps profiles with concepts present in frames of dances' videos. The last part of the paper contain experimental studies that show the good classification rate of our ETVD movement detection method.","PeriodicalId":301876,"journal":{"name":"2019 15th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126278633","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhanced Morphological Filtering for Wavelet-Based Changepoint Detection 基于小波变换点检测的增强形态学滤波
M. Stasolla, X. Neyt
This paper presents a new method for the detection of abrupt changes (i.e. mean shifts) in time series. It is a follow-up to a previous article by the authors where, for the first time, the possibility of combining the multi-scale analysis capabilities of wavelets with mathematical morphology, a theoretical framework for the analysis of spatial structures, had been explored. The processing chain has been revised and enhanced in order to improve the overall results, and a performance assessment has been carried out to evaluate the accuracy and robustness of the method to noise, also providing a comparison with its original implementation.
本文提出了一种检测时间序列突变(即平均位移)的新方法。这是作者上一篇文章的后续文章,其中首次探讨了将小波的多尺度分析能力与数学形态学相结合的可能性,这是空间结构分析的理论框架。为了改善整体结果,对处理链进行了修改和增强,并进行了性能评估,以评估该方法对噪声的准确性和鲁棒性,并与原始实现进行了比较。
{"title":"Enhanced Morphological Filtering for Wavelet-Based Changepoint Detection","authors":"M. Stasolla, X. Neyt","doi":"10.1109/SITIS.2019.00021","DOIUrl":"https://doi.org/10.1109/SITIS.2019.00021","url":null,"abstract":"This paper presents a new method for the detection of abrupt changes (i.e. mean shifts) in time series. It is a follow-up to a previous article by the authors where, for the first time, the possibility of combining the multi-scale analysis capabilities of wavelets with mathematical morphology, a theoretical framework for the analysis of spatial structures, had been explored. The processing chain has been revised and enhanced in order to improve the overall results, and a performance assessment has been carried out to evaluate the accuracy and robustness of the method to noise, also providing a comparison with its original implementation.","PeriodicalId":301876,"journal":{"name":"2019 15th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS)","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115777141","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2019 15th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1