首页 > 最新文献

2019 IEEE International Conference on Imaging Systems and Techniques (IST)最新文献

英文 中文
Identifying Asthma genetic signature patterns by mining Gene Expression BIG Datasets using Image Filtering Algorithms 通过使用图像滤波算法挖掘基因表达大数据集识别哮喘遗传特征模式
Pub Date : 2019-12-01 DOI: 10.1109/IST48021.2019.9010412
M. Hachim, B. Mahboub, Q. Hamid, R. Hamoudi
Asthma is a treatable but incurable chronic inflammatory disease affecting more than 14% of the UAE population. Asthma is still a clinical dilemma as there is no proper clinical definition of asthma, unknown definitive underlying mechanisms, no objective prognostic tool nor bedside noninvasive diagnostic test to predict complication or exacerbation. Big Data in the form of publicly available transcriptomics can be a valuable source to decipher complex diseases like asthma. Such an approach is hindered by technical variations between different studies that may mask the real biological variations and meaningful, robust findings. A large number of datasets of gene expression microarray images need a powerful tool to properly translate the image intensities into truly differential expressed genes between conditioned examined from the noise. Here we used a novel bioinformatic method based on the coefficient of variance to filter nonvariant probes with stringent image analysis processing between asthmatic and healthy to increase the power of identifying accurate signals hidden within the heterogeneous nature of asthma. Our analysis identified important signaling pathways members, namely NFKB and TGFB pathways, to be differentially expressed between severe asthma and healthy controls. Those vital pathways represent potential targets for future asthma treatment and can serve as reliable biomarkers for asthma severity. Proper image analysis for the publicly available microarray transcriptomics data increased its usefulness to decipher asthma and identify genuine differentially expressed genes that can be validated across different datasets.
哮喘是一种可治疗但无法治愈的慢性炎症性疾病,影响阿联酋14%以上的人口。哮喘仍然是一个临床难题,因为没有适当的临床定义,不知道明确的潜在机制,没有客观的预后工具,也没有床边无创诊断测试来预测并发症或恶化。公开的转录组学形式的大数据可以成为破译哮喘等复杂疾病的宝贵资源。这种方法受到不同研究之间技术差异的阻碍,这些差异可能掩盖了真正的生物学差异和有意义的、可靠的发现。基因表达微阵列图像的大量数据集需要一个强大的工具来正确地将图像强度转换为真正的差异表达基因。本文采用一种基于方差系数的新型生物信息学方法,通过严格的图像分析处理,过滤哮喘和健康之间的非变异体探针,以提高识别隐藏在哮喘异质性中的准确信号的能力。我们的分析确定了重要的信号通路成员,即NFKB和TGFB通路,在严重哮喘和健康对照之间存在差异表达。这些重要途径代表了未来哮喘治疗的潜在靶点,可以作为哮喘严重程度的可靠生物标志物。对公开可用的微阵列转录组学数据进行适当的图像分析,增加了其在破译哮喘和识别可跨不同数据集验证的真正差异表达基因方面的有用性。
{"title":"Identifying Asthma genetic signature patterns by mining Gene Expression BIG Datasets using Image Filtering Algorithms","authors":"M. Hachim, B. Mahboub, Q. Hamid, R. Hamoudi","doi":"10.1109/IST48021.2019.9010412","DOIUrl":"https://doi.org/10.1109/IST48021.2019.9010412","url":null,"abstract":"Asthma is a treatable but incurable chronic inflammatory disease affecting more than 14% of the UAE population. Asthma is still a clinical dilemma as there is no proper clinical definition of asthma, unknown definitive underlying mechanisms, no objective prognostic tool nor bedside noninvasive diagnostic test to predict complication or exacerbation. Big Data in the form of publicly available transcriptomics can be a valuable source to decipher complex diseases like asthma. Such an approach is hindered by technical variations between different studies that may mask the real biological variations and meaningful, robust findings. A large number of datasets of gene expression microarray images need a powerful tool to properly translate the image intensities into truly differential expressed genes between conditioned examined from the noise. Here we used a novel bioinformatic method based on the coefficient of variance to filter nonvariant probes with stringent image analysis processing between asthmatic and healthy to increase the power of identifying accurate signals hidden within the heterogeneous nature of asthma. Our analysis identified important signaling pathways members, namely NFKB and TGFB pathways, to be differentially expressed between severe asthma and healthy controls. Those vital pathways represent potential targets for future asthma treatment and can serve as reliable biomarkers for asthma severity. Proper image analysis for the publicly available microarray transcriptomics data increased its usefulness to decipher asthma and identify genuine differentially expressed genes that can be validated across different datasets.","PeriodicalId":117219,"journal":{"name":"2019 IEEE International Conference on Imaging Systems and Techniques (IST)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114610502","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Deep Learning-Based Approach for Accurate Segmentation of Bladder Wall using MR Images 基于深度学习的MR图像膀胱壁精确分割方法
Pub Date : 2019-12-01 DOI: 10.1109/IST48021.2019.9010233
K. Hammouda, A. El-Baz, F. Khalifa, A. Soliman, M. Ghazal, M. A. El-Ghar, A. Haddad, Mohammed M Elmogy, H. Darwish, R. Keynton
In this paper, a deep learning-based convolution neural network (CNN) is developed for accurate segmentation of the bladder wall using T2-weighted magnetic resonance imaging (T2W-MRI). Our framework utilizes a dual pathway, two-dimensional CNN for pathological bladder segmentation. Due to large bladder shape variability across subjects and the existence of pathology, a learnable adaptive shape prior (ASP) model is incorporated into our framework. To obtain the goal regions, the neural network fuses the MR image data for the first pathway, and the estimated ASP model for the second pathway. To remove noisy and scattered predictions, the CNN soft output is refined using a fully connected conditional random field (CRF). Our pipeline has been tested and evaluated using a leave-one-subject-out approach (LOSO) on twenty MRI data sets. Our framework achieved accurate segmentation results for the bladder wall and tumor as documented by the Dice similarity coefficient (DSC) and Hausdorff distance (HD). Moreover, comparative results against other segmentation approaches documented the superiority of our framework to provide accurate results for pathological bladder wall segmentation.
本文开发了一种基于深度学习的卷积神经网络(CNN),用于利用t2加权磁共振成像(T2W-MRI)对膀胱壁进行准确分割。我们的框架采用双通道,二维CNN病理膀胱分割。由于膀胱形状在受试者之间存在较大的可变性和病理学,一个可学习的自适应形状先验(ASP)模型被纳入我们的框架。为了获得目标区域,神经网络将MR图像数据融合为第一条路径,将估计的ASP模型融合为第二条路径。为了去除噪声和分散的预测,CNN软输出使用全连接条件随机场(CRF)进行细化。我们的管道已经在20个MRI数据集上使用留一个受试者方法(LOSO)进行了测试和评估。根据Dice相似系数(DSC)和Hausdorff距离(HD),我们的框架实现了膀胱壁和肿瘤的准确分割结果。此外,与其他分割方法的比较结果证明了我们的框架在提供病理膀胱壁分割的准确结果方面的优势。
{"title":"A Deep Learning-Based Approach for Accurate Segmentation of Bladder Wall using MR Images","authors":"K. Hammouda, A. El-Baz, F. Khalifa, A. Soliman, M. Ghazal, M. A. El-Ghar, A. Haddad, Mohammed M Elmogy, H. Darwish, R. Keynton","doi":"10.1109/IST48021.2019.9010233","DOIUrl":"https://doi.org/10.1109/IST48021.2019.9010233","url":null,"abstract":"In this paper, a deep learning-based convolution neural network (CNN) is developed for accurate segmentation of the bladder wall using T2-weighted magnetic resonance imaging (T2W-MRI). Our framework utilizes a dual pathway, two-dimensional CNN for pathological bladder segmentation. Due to large bladder shape variability across subjects and the existence of pathology, a learnable adaptive shape prior (ASP) model is incorporated into our framework. To obtain the goal regions, the neural network fuses the MR image data for the first pathway, and the estimated ASP model for the second pathway. To remove noisy and scattered predictions, the CNN soft output is refined using a fully connected conditional random field (CRF). Our pipeline has been tested and evaluated using a leave-one-subject-out approach (LOSO) on twenty MRI data sets. Our framework achieved accurate segmentation results for the bladder wall and tumor as documented by the Dice similarity coefficient (DSC) and Hausdorff distance (HD). Moreover, comparative results against other segmentation approaches documented the superiority of our framework to provide accurate results for pathological bladder wall segmentation.","PeriodicalId":117219,"journal":{"name":"2019 IEEE International Conference on Imaging Systems and Techniques (IST)","volume":"160 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116236854","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Cross-Spectral Periocular Recognition by Cascaded Spectral Image Transformation 级联光谱图像变换的跨光谱眼周识别
Pub Date : 2019-12-01 DOI: 10.1109/IST48021.2019.9010520
K. Raja, N. Damer, Raghavendra Ramachandra, F. Boutros, C. Busch
Recent efforts in biometrics have focused on cross-domain face recognition where images from one domain are either transformed or synthesized. In this work, we focus on a similar problem for cross spectral periocular recognition where the images from Near Infra Red (NIR) domain are matched against Visible (VIS) spectrum images. Specifically, we propose to adapt a cascaded image transformation network that can produce NIR image given a VIS image. The proposed approach is first validated with regards to the quality of the image produced by employing various quality factors. Second the applicability is demonstrated with images generated by the proposed approach. We employ a publicly available cross-spectral periocular image data of 240 unique periocular instances captured in 8 different capture sessions. We experimentally validate that the proposed image transformation scheme can produce NIR like images and also can be used with any existing feature extraction scheme. To this extent, we demonstrate the biometric applicability by using both hand-crafted and deep neural network based features under verification setting. The obtained EER of 0.7% indicates the suitability of proposed approach for image transformation from the VIS to the NIR domain.
近年来,生物识别研究的重点是跨域人脸识别,即对同一域的图像进行变换或合成。在这项工作中,我们专注于交叉光谱眼周识别的类似问题,其中近红外(NIR)域的图像与可见光(VIS)光谱图像进行匹配。具体而言,我们建议采用级联图像变换网络,该网络可以在给定VIS图像的情况下生成近红外图像。首先通过采用各种质量因素产生的图像质量来验证所提出的方法。其次,用该方法生成的图像验证了该方法的适用性。我们采用公开可用的交叉光谱眼周图像数据,在8个不同的捕获过程中捕获240个独特的眼周实例。实验验证了所提出的图像变换方案可以产生类似近红外的图像,并且可以与任何现有的特征提取方案一起使用。在这种程度上,我们通过在验证设置下使用手工制作和基于深度神经网络的特征来证明生物识别的适用性。所得的EER值为0.7%,表明该方法适用于从VIS到NIR的图像转换。
{"title":"Cross-Spectral Periocular Recognition by Cascaded Spectral Image Transformation","authors":"K. Raja, N. Damer, Raghavendra Ramachandra, F. Boutros, C. Busch","doi":"10.1109/IST48021.2019.9010520","DOIUrl":"https://doi.org/10.1109/IST48021.2019.9010520","url":null,"abstract":"Recent efforts in biometrics have focused on cross-domain face recognition where images from one domain are either transformed or synthesized. In this work, we focus on a similar problem for cross spectral periocular recognition where the images from Near Infra Red (NIR) domain are matched against Visible (VIS) spectrum images. Specifically, we propose to adapt a cascaded image transformation network that can produce NIR image given a VIS image. The proposed approach is first validated with regards to the quality of the image produced by employing various quality factors. Second the applicability is demonstrated with images generated by the proposed approach. We employ a publicly available cross-spectral periocular image data of 240 unique periocular instances captured in 8 different capture sessions. We experimentally validate that the proposed image transformation scheme can produce NIR like images and also can be used with any existing feature extraction scheme. To this extent, we demonstrate the biometric applicability by using both hand-crafted and deep neural network based features under verification setting. The obtained EER of 0.7% indicates the suitability of proposed approach for image transformation from the VIS to the NIR domain.","PeriodicalId":117219,"journal":{"name":"2019 IEEE International Conference on Imaging Systems and Techniques (IST)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123782133","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
A Deconvolutional Bottom-up Deep Network for multi-person pose estimation 基于反卷积自底向上深度网络的多人姿态估计
Pub Date : 2019-12-01 DOI: 10.1109/IST48021.2019.9010189
Meng Li, Haoqian Wang, Yongbing Zhang, Yi Yang
Due to the trade off between model complexity and estimation accuracy, current human pose estimators either are of low accuracy or requires long running time. Such dilemma is especially severe in real time multi-person pose estimation. To address this issue, we design a deep network of reduced parameter size and high estimation accuracy, via introducing deconvolution layers instead of widely used fully-connected configuration. Specifically, our model consists of two successive parts: Detection network and matching network. The former outputs keypoint heatmap and person locations, and then the latter produces the final pose estimation using multiple deconvolutional layers. Benefiting from the simple structure and explicit utilization of previously neglected spatial structure in heatmap, the matching network is of specially high efficiency and at high precision. Experiments on the challenging COCO dataset demonstrate our method can almost cut off the running parameters of matching network, while achieving higher accuracy than existing methods.
由于模型复杂性和估计精度之间的权衡,目前的人体姿态估计器要么精度低,要么需要较长的运行时间。这种困境在实时多人姿态估计中尤为严重。为了解决这个问题,我们通过引入反褶积层而不是广泛使用的全连接配置,设计了一个参数尺寸较小且估计精度高的深度网络。具体来说,我们的模型由两个连续的部分组成:检测网络和匹配网络。前者输出关键点热图和人物位置,后者使用多个反卷积层生成最终的姿态估计。该匹配网络结构简单,明确利用了热图中以往被忽略的空间结构,具有高效率和高精度的特点。在具有挑战性的COCO数据集上的实验表明,我们的方法几乎可以切断匹配网络的运行参数,同时取得了比现有方法更高的精度。
{"title":"A Deconvolutional Bottom-up Deep Network for multi-person pose estimation","authors":"Meng Li, Haoqian Wang, Yongbing Zhang, Yi Yang","doi":"10.1109/IST48021.2019.9010189","DOIUrl":"https://doi.org/10.1109/IST48021.2019.9010189","url":null,"abstract":"Due to the trade off between model complexity and estimation accuracy, current human pose estimators either are of low accuracy or requires long running time. Such dilemma is especially severe in real time multi-person pose estimation. To address this issue, we design a deep network of reduced parameter size and high estimation accuracy, via introducing deconvolution layers instead of widely used fully-connected configuration. Specifically, our model consists of two successive parts: Detection network and matching network. The former outputs keypoint heatmap and person locations, and then the latter produces the final pose estimation using multiple deconvolutional layers. Benefiting from the simple structure and explicit utilization of previously neglected spatial structure in heatmap, the matching network is of specially high efficiency and at high precision. Experiments on the challenging COCO dataset demonstrate our method can almost cut off the running parameters of matching network, while achieving higher accuracy than existing methods.","PeriodicalId":117219,"journal":{"name":"2019 IEEE International Conference on Imaging Systems and Techniques (IST)","volume":"206 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125738371","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic Calibration of Dual-LiDARs Using Two Poles Stickered with Retro-Reflective Tape 双激光雷达两极贴反光带的自动校准
Pub Date : 2019-11-02 DOI: 10.1109/IST48021.2019.9010134
Bohuan Xue, Jianhao Jiao, Yilong Zhu, Linwei Zheng, Dong Han, Ming Liu, Rui Fan
Multi-LiDAR systems have been prevalently applied in modern autonomous vehicles to render a broad view of the environments. The rapid development of 5G wireless technologies has brought a breakthrough for current cellular vehicle-to-everything (C-V2X) applications. Therefore, a novel localization and perception system in which multiple LiDARs are mounted around cities for autonomous vehicles has been proposed. However, the existing calibration methods require specific hard-to-move markers, ego-motion, or good initial values given by users. In this paper, we present a novel approach that enables automatic multi-LiDAR calibration using two poles stickered with retro-reflective tape. This method does not depend on prior environmental information, initial values of the extrinsic parameters, or movable platforms like a car. We analyze the LiDAR-pole model, verify the feasibility of the algorithm through simulation data, and present a simple method to measure the calibration errors w.r.t the ground truth. Experimental results demonstrate that our approach gains better flexibility and higher accuracy when compared with the state-of-the-art approach.
多激光雷达系统已广泛应用于现代自动驾驶汽车,以提供广阔的环境视野。5G无线技术的快速发展为当前的蜂窝车联网(C-V2X)应用带来了突破。因此,提出了一种新型的定位和感知系统,该系统在城市周围安装多个激光雷达,用于自动驾驶汽车。然而,现有的校准方法需要特定的难以移动的标记,自我运动,或良好的初始值由用户提供。在本文中,我们提出了一种新颖的方法,使用贴有反光胶带的两极来实现自动多激光雷达校准。该方法不依赖于先前的环境信息、外部参数的初始值或像汽车这样的可移动平台。分析了激光雷达极点模型,通过仿真数据验证了算法的可行性,并提出了一种简单的方法来测量标定误差。实验结果表明,与现有方法相比,该方法具有更好的灵活性和更高的精度。
{"title":"Automatic Calibration of Dual-LiDARs Using Two Poles Stickered with Retro-Reflective Tape","authors":"Bohuan Xue, Jianhao Jiao, Yilong Zhu, Linwei Zheng, Dong Han, Ming Liu, Rui Fan","doi":"10.1109/IST48021.2019.9010134","DOIUrl":"https://doi.org/10.1109/IST48021.2019.9010134","url":null,"abstract":"Multi-LiDAR systems have been prevalently applied in modern autonomous vehicles to render a broad view of the environments. The rapid development of 5G wireless technologies has brought a breakthrough for current cellular vehicle-to-everything (C-V2X) applications. Therefore, a novel localization and perception system in which multiple LiDARs are mounted around cities for autonomous vehicles has been proposed. However, the existing calibration methods require specific hard-to-move markers, ego-motion, or good initial values given by users. In this paper, we present a novel approach that enables automatic multi-LiDAR calibration using two poles stickered with retro-reflective tape. This method does not depend on prior environmental information, initial values of the extrinsic parameters, or movable platforms like a car. We analyze the LiDAR-pole model, verify the feasibility of the algorithm through simulation data, and present a simple method to measure the calibration errors w.r.t the ground truth. Experimental results demonstrate that our approach gains better flexibility and higher accuracy when compared with the state-of-the-art approach.","PeriodicalId":117219,"journal":{"name":"2019 IEEE International Conference on Imaging Systems and Techniques (IST)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122030017","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Sliding Mode Control based Support Vector Machine RBF Kernel Parameter Optimization 基于支持向量机RBF核参数优化的滑模控制
Pub Date : 2019-11-01 DOI: 10.1109/IST48021.2019.9010479
Maryam Yalsavar, P. Karimaghaee, Akbar Sheikh-Akbari, J. Dehmeshki, M. Khooban, Salah Al-Majeed
Support Vector Machine (SVM) is a learning-based algorithm, which is widely used for classification in many applications. Despite its advantages, its application to large scale datasets is limited due to its use of large number of support vectors and dependency of its performance on its kernel parameter. This paper presents a Sliding Mode Control based Support Vector Machine Radial Basis Function's kernel parameter optimization (SMC-SVM-RBF) method, inspired by sliding mode closed loop control theory, which has demonstrated significantly higher performance to that of the standard closed loop control technique. The proposed method first defines an error equation and a sliding surface and then iteratively updates the RBF's kernel parameter based on the sliding mode control theory, forcing SVM training error to converge below a predefined threshold value. The closed loop nature of the proposed algorithm increases the robustness of the technique to uncertainty and improves its convergence speed. Experimental results were generated using nine standard benchmark datasets covering wide range of applications. Results show the proposed SMC-SVM-RBF method is significantly faster than those of classical SVM based techniques. Moreover, it generates more accurate results than most of the state of the art SVM based methods.
支持向量机(SVM)是一种基于学习的分类算法,在分类领域有着广泛的应用。尽管它有很多优点,但由于它使用了大量的支持向量,并且它的性能依赖于它的核参数,因此它在大规模数据集上的应用受到限制。本文从滑模闭环控制理论出发,提出了一种基于支持向量机径向基函数的滑模控制核参数优化(SMC-SVM-RBF)方法,该方法的性能明显优于标准闭环控制技术。该方法首先定义误差方程和滑动曲面,然后基于滑模控制理论迭代更新RBF的核参数,迫使SVM训练误差收敛到预定义的阈值以下。该算法的闭环特性增强了该算法对不确定性的鲁棒性,提高了算法的收敛速度。使用涵盖广泛应用的9个标准基准数据集生成实验结果。结果表明,SMC-SVM-RBF方法的速度明显快于传统的基于SVM的方法。此外,它产生的结果比大多数最先进的基于支持向量机的方法更准确。
{"title":"Sliding Mode Control based Support Vector Machine RBF Kernel Parameter Optimization","authors":"Maryam Yalsavar, P. Karimaghaee, Akbar Sheikh-Akbari, J. Dehmeshki, M. Khooban, Salah Al-Majeed","doi":"10.1109/IST48021.2019.9010479","DOIUrl":"https://doi.org/10.1109/IST48021.2019.9010479","url":null,"abstract":"Support Vector Machine (SVM) is a learning-based algorithm, which is widely used for classification in many applications. Despite its advantages, its application to large scale datasets is limited due to its use of large number of support vectors and dependency of its performance on its kernel parameter. This paper presents a Sliding Mode Control based Support Vector Machine Radial Basis Function's kernel parameter optimization (SMC-SVM-RBF) method, inspired by sliding mode closed loop control theory, which has demonstrated significantly higher performance to that of the standard closed loop control technique. The proposed method first defines an error equation and a sliding surface and then iteratively updates the RBF's kernel parameter based on the sliding mode control theory, forcing SVM training error to converge below a predefined threshold value. The closed loop nature of the proposed algorithm increases the robustness of the technique to uncertainty and improves its convergence speed. Experimental results were generated using nine standard benchmark datasets covering wide range of applications. Results show the proposed SMC-SVM-RBF method is significantly faster than those of classical SVM based techniques. Moreover, it generates more accurate results than most of the state of the art SVM based methods.","PeriodicalId":117219,"journal":{"name":"2019 IEEE International Conference on Imaging Systems and Techniques (IST)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130139401","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Robust Pavement Mapping System Based on Normal-Constrained Stereo Visual Odometry 基于法向约束立体视觉里程计的鲁棒路面映射系统
Pub Date : 2019-10-29 DOI: 10.1109/IST48021.2019.9010439
Huaiyang Huang, Rui Fan, Yilong Zhu, Ming Liu, I. Pitas
Pavement condition is crucial for civil infrastructure maintenance. This task usually requires efficient road damage localization, which can be accomplished by the visual odometry system embedded in unmanned aerial vehicles (UAVs), However, the state-of-the-art visual odometry and mapping methods suffer from large drift under the degeneration of the scene structure. To alleviate this issue, we integrate normal constraints into the visual odometry process, which greatly helps to avoid large drift. By parameterizing the normal vector on the tangential plane, the normal factors are coupled with traditional reprojection factors in the pose optimization procedure. The experimental results demonstrate the effectiveness of the proposed system. The overall absolute trajectory error is improved by approximately 20%, which indicates that the estimated trajectory is much more accurate than that obtained using other state-of-the-art methods.
路面状况对民用基础设施的维护至关重要。该任务通常需要高效的道路损伤定位,这可以通过嵌入在无人机上的视觉里程计系统来实现,但目前的视觉里程计和测绘方法在场景结构退化的情况下存在较大的漂移。为了缓解这个问题,我们将常规约束集成到视觉里程计过程中,这大大有助于避免大漂移。通过参数化切平面上的法向量,将法向量与传统的重投影因子耦合到位姿优化过程中。实验结果证明了该系统的有效性。总的绝对弹道误差提高了约20%,这表明估计的弹道比使用其他最先进的方法得到的精确得多。
{"title":"A Robust Pavement Mapping System Based on Normal-Constrained Stereo Visual Odometry","authors":"Huaiyang Huang, Rui Fan, Yilong Zhu, Ming Liu, I. Pitas","doi":"10.1109/IST48021.2019.9010439","DOIUrl":"https://doi.org/10.1109/IST48021.2019.9010439","url":null,"abstract":"Pavement condition is crucial for civil infrastructure maintenance. This task usually requires efficient road damage localization, which can be accomplished by the visual odometry system embedded in unmanned aerial vehicles (UAVs), However, the state-of-the-art visual odometry and mapping methods suffer from large drift under the degeneration of the scene structure. To alleviate this issue, we integrate normal constraints into the visual odometry process, which greatly helps to avoid large drift. By parameterizing the normal vector on the tangential plane, the normal factors are coupled with traditional reprojection factors in the pose optimization procedure. The experimental results demonstrate the effectiveness of the proposed system. The overall absolute trajectory error is improved by approximately 20%, which indicates that the estimated trajectory is much more accurate than that obtained using other state-of-the-art methods.","PeriodicalId":117219,"journal":{"name":"2019 IEEE International Conference on Imaging Systems and Techniques (IST)","volume":"471 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122195937","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Autonomous UAV Landing System Based on Visual Navigation 基于视觉导航的无人机自主着陆系统
Pub Date : 2019-10-29 DOI: 10.1109/IST48021.2019.9010264
Zhixin Wu, Peng Han, Ruiwen Yao, Lei Qiao, Weidong Zhang, T. Shen, Min Sun, Yilong Zhu, Ming Liu, Rui Fan
In this paper, we present an autonomous unmanned aerial vehicle (UAV) landing system based on visual navigation. We design the landmark as a topological pattern in order to enable the UAV to distinguish the landmark from the environment easily. In addition, a dynamic thresholding method is developed for image binarization to improve detection efficiency. The relative distance in the horizontal plane is calculated according to effective image information, and the relative height is obtained using a linear interpolation method. The landing experiments are performed on a static and a moving platform, respectively. The experimental results illustrate that our proposed landing system performs robustly and accurately.
本文提出了一种基于视觉导航的无人机自主着陆系统。我们将地标设计成拓扑模式,使无人机能够轻松地将地标与环境区分开来。此外,为了提高图像二值化的检测效率,提出了一种动态阈值化方法。根据有效图像信息计算水平面内的相对距离,采用线性插值方法获得相对高度。着陆实验分别在静态平台和移动平台上进行。实验结果表明,该着陆系统具有较好的鲁棒性和准确性。
{"title":"Autonomous UAV Landing System Based on Visual Navigation","authors":"Zhixin Wu, Peng Han, Ruiwen Yao, Lei Qiao, Weidong Zhang, T. Shen, Min Sun, Yilong Zhu, Ming Liu, Rui Fan","doi":"10.1109/IST48021.2019.9010264","DOIUrl":"https://doi.org/10.1109/IST48021.2019.9010264","url":null,"abstract":"In this paper, we present an autonomous unmanned aerial vehicle (UAV) landing system based on visual navigation. We design the landmark as a topological pattern in order to enable the UAV to distinguish the landmark from the environment easily. In addition, a dynamic thresholding method is developed for image binarization to improve detection efficiency. The relative distance in the horizontal plane is calculated according to effective image information, and the relative height is obtained using a linear interpolation method. The landing experiments are performed on a static and a moving platform, respectively. The experimental results illustrate that our proposed landing system performs robustly and accurately.","PeriodicalId":117219,"journal":{"name":"2019 IEEE International Conference on Imaging Systems and Techniques (IST)","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126263332","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
PT-ResNet: Perspective Transformation-Based Residual Network for Semantic Road Image Segmentation PT-ResNet:基于视角变换的残差网络语义道路图像分割
Pub Date : 2019-10-29 DOI: 10.1109/IST48021.2019.9010501
Rui Fan, Yuan Wang, Lei Qiao, Ruiwen Yao, Peng Han, Weidong Zhang, I. Pitas, Ming Liu
Semantic road region segmentation is a high-level task, which paves the way towards road scene understanding. This paper presents a residual network trained for semantic road segmentation. Firstly, we represent the projections of road disparities in the v-disparity map as a linear model, which can be estimated by optimizing the v-disparity map using dynamic programming. This linear model is then utilized to reduce the redundant information in the left and right road images. The right image is also transformed into the left perspective view, which greatly enhances the road surface similarity between the two images. Finally, the processed stereo images and their disparity maps are concatenated to create a set of 3D images, which are then utilized to train our neural network. The experimental results illustrate that our network achieves a maximum F1-measure of approximately 91.19%, when analyzing the images from the KITTI road dataset.
语义道路区域分割是一项高级任务,为道路场景理解奠定了基础。本文提出了一种用于语义道路分割的残差网络。首先,我们将v-视差图中的道路视差投影表示为线性模型,通过动态规划优化v-视差图来估计道路视差。然后利用该线性模型来减少左右道路图像中的冗余信息。将右侧图像也转换为左侧视角视图,大大增强了两幅图像之间的路面相似度。最后,将处理后的立体图像及其视差图连接起来创建一组三维图像,然后使用这些图像来训练我们的神经网络。实验结果表明,在分析KITTI道路数据集的图像时,我们的网络达到了大约91.19%的最大f1测量值。
{"title":"PT-ResNet: Perspective Transformation-Based Residual Network for Semantic Road Image Segmentation","authors":"Rui Fan, Yuan Wang, Lei Qiao, Ruiwen Yao, Peng Han, Weidong Zhang, I. Pitas, Ming Liu","doi":"10.1109/IST48021.2019.9010501","DOIUrl":"https://doi.org/10.1109/IST48021.2019.9010501","url":null,"abstract":"Semantic road region segmentation is a high-level task, which paves the way towards road scene understanding. This paper presents a residual network trained for semantic road segmentation. Firstly, we represent the projections of road disparities in the v-disparity map as a linear model, which can be estimated by optimizing the v-disparity map using dynamic programming. This linear model is then utilized to reduce the redundant information in the left and right road images. The right image is also transformed into the left perspective view, which greatly enhances the road surface similarity between the two images. Finally, the processed stereo images and their disparity maps are concatenated to create a set of 3D images, which are then utilized to train our neural network. The experimental results illustrate that our network achieves a maximum F1-measure of approximately 91.19%, when analyzing the images from the KITTI road dataset.","PeriodicalId":117219,"journal":{"name":"2019 IEEE International Conference on Imaging Systems and Techniques (IST)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129879997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Real-Time, Environmentally-Robust 3D LiDAR Localization 实时、环境健壮的3D激光雷达定位
Pub Date : 2019-10-28 DOI: 10.1109/IST48021.2019.9010305
Yilong Zhu, Bohuan Xue, Linwei Zheng, Huaiyang Huang, Ming Liu, Rui Fan
Localization, or position fixing, is an important problem in robotics research. In this paper, we propose a novel approach for long-term localization in a changing environment using 3D LiDAR. We first create the map of a real environment using GPS and LiDAR. Then, we divide the map into several small parts as the targets for cloud registration, which can not only improve the robustness but also reduce the registration time. We proposed a localization method called PointLocalization. PointLocalization allows us to fuse different kinds of odometers, which can optimize the accuracy and frequency of localization results. We evaluate our algorithm on an unmanned ground vehicle (UGV) using LiDAR and a wheel encoder, and obtain the localization results at more than 20 Hz after fusion. The algorithm can also localize the UGV in a 180-degree field of view (FOV). Using an outdated map captured six months ago, this algorithm shows great robustness, and the test results show that it can achieve an accuracy of 10 cm. PointLocalization has been tested for a period of more than six months in a crowded factory and has operated successfully over a distance of more than 2000 km.
定位是机器人研究中的一个重要问题。在本文中,我们提出了一种在不断变化的环境中使用3D激光雷达进行长期定位的新方法。我们首先使用GPS和激光雷达创建真实环境的地图。然后,我们将地图分割成若干小块作为云配准的目标,这样既提高了鲁棒性,又减少了配准时间。我们提出了一个叫做PointLocalization的定位方法。PointLocalization允许我们融合不同类型的里程表,可以优化定位结果的准确性和频率。利用激光雷达和车轮编码器在无人地面车辆(UGV)上对算法进行了验证,得到了融合后的定位结果,定位频率大于20 Hz。该算法还可以在180度视场范围内对UGV进行定位。使用六个月前捕获的过时地图,该算法显示出很强的鲁棒性,测试结果表明,该算法可以达到10厘米的精度。PointLocalization已经在一个拥挤的工厂进行了6个多月的测试,并成功运行了超过2000公里的距离。
{"title":"Real-Time, Environmentally-Robust 3D LiDAR Localization","authors":"Yilong Zhu, Bohuan Xue, Linwei Zheng, Huaiyang Huang, Ming Liu, Rui Fan","doi":"10.1109/IST48021.2019.9010305","DOIUrl":"https://doi.org/10.1109/IST48021.2019.9010305","url":null,"abstract":"Localization, or position fixing, is an important problem in robotics research. In this paper, we propose a novel approach for long-term localization in a changing environment using 3D LiDAR. We first create the map of a real environment using GPS and LiDAR. Then, we divide the map into several small parts as the targets for cloud registration, which can not only improve the robustness but also reduce the registration time. We proposed a localization method called PointLocalization. PointLocalization allows us to fuse different kinds of odometers, which can optimize the accuracy and frequency of localization results. We evaluate our algorithm on an unmanned ground vehicle (UGV) using LiDAR and a wheel encoder, and obtain the localization results at more than 20 Hz after fusion. The algorithm can also localize the UGV in a 180-degree field of view (FOV). Using an outdated map captured six months ago, this algorithm shows great robustness, and the test results show that it can achieve an accuracy of 10 cm. PointLocalization has been tested for a period of more than six months in a crowded factory and has operated successfully over a distance of more than 2000 km.","PeriodicalId":117219,"journal":{"name":"2019 IEEE International Conference on Imaging Systems and Techniques (IST)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121610400","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
期刊
2019 IEEE International Conference on Imaging Systems and Techniques (IST)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1