首页 > 最新文献

2017 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI)最新文献

英文 中文
Space-invariant signature algorithm processing of ultrasound images for the detection and localization of early abnormalities in animal tissues 空间不变签名算法处理超声图像,用于动物组织早期异常的检测和定位
Kushagra Parolia, M. Gupta, P. Babyn, W. Zhang, D. Bitner
In this paper we present an innovative space-variance approach named as “Space-Invariant Signature Algorithm (SISA)” for processing images from active systems, such as cancer cells, tumor growth, and dead cells, for the detection and localization of abnormalities at an early stage. In this paper, a SISA processing algorithm is developed, and this algorithm is tested on animal tissues such as pigs and chicken tissues. The abnormality in an active system can be defined as the obstacle or a failure which impedes the activities in tissues such as smooth flow of blood or electrical signals etc. Due to this impeding nature of the abnormality, some parameter perturbations are induced. In this paper using the SISA approach, these perturbations were detected in a preliminary experiment on animal tissues. The degree and position of the space-variance helps us in the detection and localization of abnormality even at an early (incipient) stage. The space-variance signature pattern is named as a ‘SISA signature pattern’. In the absence of any abnormality, the signature pattern is space invariant, whereas, in the presences of any abnormality, the SISA signature pattern varies in the space (space-variant). The basic experimental studies on animal tissues using ultra sound imaging strongly suggest a possible use of the SISA approach as a non-invasive method for the detection and localization of abnormalities in biological tissues such as cancer cells non-invasively.
在本文中,我们提出了一种创新的空间方差方法,称为“空间不变签名算法(SISA)”,用于处理来自活跃系统(如癌细胞、肿瘤生长和死细胞)的图像,以便在早期阶段检测和定位异常。本文开发了一种SISA处理算法,并在猪、鸡等动物组织上进行了实验。活动系统中的异常可以定义为阻碍组织活动的障碍或故障,如血液或电信号的顺畅流动等。由于这种异常的阻碍性质,引起了一些参数的扰动。在本文中,使用SISA方法,在动物组织的初步实验中检测到这些扰动。空间变异的程度和位置有助于我们在早期发现和定位异常。空间方差签名模式被命名为“SISA签名模式”。在没有任何异常的情况下,签名模式是空间不变的,而在存在任何异常的情况下,SISA签名模式在空间中变化(空间变异体)。利用超声成像对动物组织进行的基础实验研究强烈表明,SISA方法可能作为一种非侵入性方法,用于检测和定位生物组织中的异常,如癌细胞。
{"title":"Space-invariant signature algorithm processing of ultrasound images for the detection and localization of early abnormalities in animal tissues","authors":"Kushagra Parolia, M. Gupta, P. Babyn, W. Zhang, D. Bitner","doi":"10.1109/CISP-BMEI.2017.8302204","DOIUrl":"https://doi.org/10.1109/CISP-BMEI.2017.8302204","url":null,"abstract":"In this paper we present an innovative space-variance approach named as “Space-Invariant Signature Algorithm (SISA)” for processing images from active systems, such as cancer cells, tumor growth, and dead cells, for the detection and localization of abnormalities at an early stage. In this paper, a SISA processing algorithm is developed, and this algorithm is tested on animal tissues such as pigs and chicken tissues. The abnormality in an active system can be defined as the obstacle or a failure which impedes the activities in tissues such as smooth flow of blood or electrical signals etc. Due to this impeding nature of the abnormality, some parameter perturbations are induced. In this paper using the SISA approach, these perturbations were detected in a preliminary experiment on animal tissues. The degree and position of the space-variance helps us in the detection and localization of abnormality even at an early (incipient) stage. The space-variance signature pattern is named as a ‘SISA signature pattern’. In the absence of any abnormality, the signature pattern is space invariant, whereas, in the presences of any abnormality, the SISA signature pattern varies in the space (space-variant). The basic experimental studies on animal tissues using ultra sound imaging strongly suggest a possible use of the SISA approach as a non-invasive method for the detection and localization of abnormalities in biological tissues such as cancer cells non-invasively.","PeriodicalId":6474,"journal":{"name":"2017 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI)","volume":"161 1","pages":"1-7"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86738323","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Security mechanism of video content integrated broadcast control platform under triple play 三网融合下视频内容集成播出控制平台的安全机制
YU Peng, Nenghuan Zhang, Shengyan Zhang, Qi Wang
According to the new format of triple play, video contents interact in the integrated broadcast control platform, there are security risks and participation dispute liability trace ability problem, we propose an integrated control platform security mechanism under triple play. The mechanism builds a bit commitment protocol through hash function, preventing the video tampering in public network transport, using exchanged ElGamal encryption mechanism to achieve a confidentiality comparison in the event of a dispute, to trace the responsible party. Theoretical analysis shows that this method has good security and feasibility, and can effectively solve the security of video content in integrated broadcast control environment.
针对三网融合新业态下,视频内容在综合直播控制平台互动,存在安全风险和参与争议责任溯源能力问题,提出了三网融合下的综合控制平台安全机制。该机制通过哈希函数构建位承诺协议,防止视频在公共网络传输中被篡改,采用交换式ElGamal加密机制,在发生争议时实现保密性比较,追踪责任方。理论分析表明,该方法具有良好的安全性和可行性,能够有效解决综合广播控制环境下视频内容的安全问题。
{"title":"Security mechanism of video content integrated broadcast control platform under triple play","authors":"YU Peng, Nenghuan Zhang, Shengyan Zhang, Qi Wang","doi":"10.1109/CISP-BMEI.2017.8301930","DOIUrl":"https://doi.org/10.1109/CISP-BMEI.2017.8301930","url":null,"abstract":"According to the new format of triple play, video contents interact in the integrated broadcast control platform, there are security risks and participation dispute liability trace ability problem, we propose an integrated control platform security mechanism under triple play. The mechanism builds a bit commitment protocol through hash function, preventing the video tampering in public network transport, using exchanged ElGamal encryption mechanism to achieve a confidentiality comparison in the event of a dispute, to trace the responsible party. Theoretical analysis shows that this method has good security and feasibility, and can effectively solve the security of video content in integrated broadcast control environment.","PeriodicalId":6474,"journal":{"name":"2017 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI)","volume":"2014 1","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86764972","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
An image patch matching method based on multi-feature fusion 一种基于多特征融合的图像补丁匹配方法
Xiangru Yu, Yukun Guo, Jinping Li, Fudong Cai
Appropriate features are very important for the robustness and effectiveness of matching algorithms. In general, current algorithms depend on descriptors like SIFT, SURF, which makes them only care about information of key points while ignoring the knowledge of the whole image, thus those methods are easy to result in false matches. We propose a novel matching method called multi-feature fusion, which takes full advantage of geometric, gray, color and texture features. Then we validate the effect of our method using images captured from practical applications. Experiments show the method can effectively complete the matching task of image patch.
合适的特征对匹配算法的鲁棒性和有效性至关重要。一般来说,目前的算法依赖于SIFT、SURF等描述符,这使得它们只关心关键点的信息,而忽略了对整个图像的了解,因此这些方法容易导致错误匹配。我们提出了一种新的多特征融合匹配方法,该方法充分利用了图像的几何特征、灰度特征、颜色特征和纹理特征。然后用实际应用中捕获的图像验证了我们的方法的效果。实验表明,该方法能有效地完成图像patch的匹配任务。
{"title":"An image patch matching method based on multi-feature fusion","authors":"Xiangru Yu, Yukun Guo, Jinping Li, Fudong Cai","doi":"10.1109/CISP-BMEI.2017.8302044","DOIUrl":"https://doi.org/10.1109/CISP-BMEI.2017.8302044","url":null,"abstract":"Appropriate features are very important for the robustness and effectiveness of matching algorithms. In general, current algorithms depend on descriptors like SIFT, SURF, which makes them only care about information of key points while ignoring the knowledge of the whole image, thus those methods are easy to result in false matches. We propose a novel matching method called multi-feature fusion, which takes full advantage of geometric, gray, color and texture features. Then we validate the effect of our method using images captured from practical applications. Experiments show the method can effectively complete the matching task of image patch.","PeriodicalId":6474,"journal":{"name":"2017 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI)","volume":"113 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86791062","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Beating heart motion prediction using iterative optimal sine filtering 用迭代最优正弦滤波预测心脏跳动运动
Bo Yang, Tingting Cao, Wenfeng Zheng
A novel motion prediction algorithm is proposed to robustly track heart beat in minimally invasive surgery. To model the movement of Points of Interest (POI) on heart tissue, the Dual Time-Varying Fourier Series (DTVFS) is employed. The Fourier coefficients and the frequencies of the DTVFS model are estimated separately using the dual Kalman filtering. An iterative optimal sine filtering algorithm is developed, which can accurately measure the instantaneous frequencies of breathing circle and heart beating from the motion curves of the POI. The proposed method is verified on the simulated dataset and the real-measured datasets captured by the daVinci surgical system.
提出了一种新的运动预测算法,以鲁棒跟踪微创手术中的心脏跳动。为了模拟心脏组织上兴趣点(POI)的运动,采用了双时变傅立叶级数(DTVFS)。利用双卡尔曼滤波分别估计了DTVFS模型的傅里叶系数和频率。提出了一种迭代最优正弦滤波算法,该算法可以从POI的运动曲线中精确测量呼吸循环和心跳的瞬时频率。在仿真数据集和达芬奇手术系统采集的实测数据集上验证了该方法的有效性。
{"title":"Beating heart motion prediction using iterative optimal sine filtering","authors":"Bo Yang, Tingting Cao, Wenfeng Zheng","doi":"10.1109/CISP-BMEI.2017.8302214","DOIUrl":"https://doi.org/10.1109/CISP-BMEI.2017.8302214","url":null,"abstract":"A novel motion prediction algorithm is proposed to robustly track heart beat in minimally invasive surgery. To model the movement of Points of Interest (POI) on heart tissue, the Dual Time-Varying Fourier Series (DTVFS) is employed. The Fourier coefficients and the frequencies of the DTVFS model are estimated separately using the dual Kalman filtering. An iterative optimal sine filtering algorithm is developed, which can accurately measure the instantaneous frequencies of breathing circle and heart beating from the motion curves of the POI. The proposed method is verified on the simulated dataset and the real-measured datasets captured by the daVinci surgical system.","PeriodicalId":6474,"journal":{"name":"2017 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI)","volume":"5 1","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88889641","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A miniaturized phased high intensity focused ultrasound transducer-driven system with MR compatibility 具有核磁共振兼容性的小型化相位高强度聚焦超声换能器驱动系统
Li Gong, G. Shen, S. Qiao, Wenjie Liu, Yuchi Yi
A new transducer-driven system is designed for these important characteristics: lower harmonic energy, higher integration and more flexible wave output mode. According to the flowing direction of signals, the whole system mainly has three parts: the controlling module based on FPGA, the wave generator group and the signal processing group. FPGA concurrently controls the following wave generator groups by digital signals outputting in parallel. The wave generator, which could generate continuous sine wave and pulse wave in a low harmonic energy level, has a programmable Direct Digital Synthesizer (DDS) with a 12 bit Digital to Analog Converter (DAC) and sampling up to 180 MSPS. The signal processing group contains the amplification module and the matching network group for transducers. The power of sine wave is magnified by integrated operational amplifier (op amp). The matching network is designed for the transducer with frequency 1.36MHz. Finally the measurement of one channel output shows that: the frequency accuracy is 1.36MHz ±0.04MHz and the wave mode can be configured as continuous sine wave or pulse wave. The amplitude of each harmonics in the FFT analysis is over 50dB lower than the fundamental component. The acoustic power of one circular transducer with radius of 5mm is more than 2.8W.
针对低谐波能量、高集成度和更灵活的波输出模式等重要特性,设计了一种新型换能器驱动系统。根据信号的流向,整个系统主要分为三部分:基于FPGA的控制模块、波形发生器组和信号处理组。FPGA通过并行输出数字信号,并发控制以下波发生器组。该波形发生器具有可编程的直接数字合成器(DDS)和12位数模转换器(DAC),采样率高达180 MSPS,可以在低谐波能量水平上产生连续的正弦波和脉冲波。所述信号处理组包括用于换能器的放大模块和匹配网络组。正弦波的功率通过集成运算放大器(运放)放大。针对频率为1.36MHz的换能器设计了匹配网络。最后对一通道输出的测量结果表明:频率精度为1.36MHz±0.04MHz,波形模式可配置为连续正弦波或脉冲波。FFT分析中各谐波的幅值比基波分量低50dB以上。一个半径为5mm的圆形换能器声功率大于2.8W。
{"title":"A miniaturized phased high intensity focused ultrasound transducer-driven system with MR compatibility","authors":"Li Gong, G. Shen, S. Qiao, Wenjie Liu, Yuchi Yi","doi":"10.1109/CISP-BMEI.2017.8302248","DOIUrl":"https://doi.org/10.1109/CISP-BMEI.2017.8302248","url":null,"abstract":"A new transducer-driven system is designed for these important characteristics: lower harmonic energy, higher integration and more flexible wave output mode. According to the flowing direction of signals, the whole system mainly has three parts: the controlling module based on FPGA, the wave generator group and the signal processing group. FPGA concurrently controls the following wave generator groups by digital signals outputting in parallel. The wave generator, which could generate continuous sine wave and pulse wave in a low harmonic energy level, has a programmable Direct Digital Synthesizer (DDS) with a 12 bit Digital to Analog Converter (DAC) and sampling up to 180 MSPS. The signal processing group contains the amplification module and the matching network group for transducers. The power of sine wave is magnified by integrated operational amplifier (op amp). The matching network is designed for the transducer with frequency 1.36MHz. Finally the measurement of one channel output shows that: the frequency accuracy is 1.36MHz ±0.04MHz and the wave mode can be configured as continuous sine wave or pulse wave. The amplitude of each harmonics in the FFT analysis is over 50dB lower than the fundamental component. The acoustic power of one circular transducer with radius of 5mm is more than 2.8W.","PeriodicalId":6474,"journal":{"name":"2017 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI)","volume":"34 1","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90039845","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Web-based training for radiologists of breast ultrasound 乳腺超声放射科医师网络培训
Xianhai Huang, L. Ling, Qinghua Huang, Yidi Lin, Xingzhang Long, Longzhong Liu
Breast cancer is still considered as the most common form of cancer as well as the leading causes of cancer deaths among women all over the world. Fortunately, the mortality of breast cancer can be significantly reduced via the early detection and diagnosis of breast cancer. As one of the most continually used diagnosis tools, ultrasonography (US) scan plays an important role in the detection and classification of the breast tumor. In this paper, we introduce a large breast ultrasound image database which stored breast ultrasound images and pathology results from breast tumor patients as well as their clinic diagnostic information. Furthermore, we design a web-based training system based on the database using a feature scoring scheme which based on the fifth edition of Breast Imaging Reporting and Data System (BI-RADS) lexicon for US. This online training system (new web-based teaching framework) automatically creates case-based exercises to train and guide the newly employed or resident sonographers for diagnosis of breast cancer using breast ultrasound images based on the BI-RADS.
乳腺癌仍然被认为是最常见的癌症形式,也是全世界妇女癌症死亡的主要原因。幸运的是,通过乳腺癌的早期发现和诊断,乳腺癌的死亡率可以显著降低。超声扫描作为最常用的诊断工具之一,在乳腺肿瘤的检测和分类中起着重要的作用。本文介绍了一个大型乳腺超声图像数据库,该数据库存储了乳腺肿瘤患者的乳腺超声图像、病理结果以及临床诊断信息。此外,我们还设计了一个基于数据库的基于web的培训系统,该系统采用基于美国第五版乳腺成像报告和数据系统(BI-RADS)词典的特征评分方案。该在线培训系统(新的基于网络的教学框架)自动创建基于案例的练习,以培训和指导新入职或住院超声医师使用基于BI-RADS的乳腺超声图像诊断乳腺癌。
{"title":"Web-based training for radiologists of breast ultrasound","authors":"Xianhai Huang, L. Ling, Qinghua Huang, Yidi Lin, Xingzhang Long, Longzhong Liu","doi":"10.1109/CISP-BMEI.2017.8302273","DOIUrl":"https://doi.org/10.1109/CISP-BMEI.2017.8302273","url":null,"abstract":"Breast cancer is still considered as the most common form of cancer as well as the leading causes of cancer deaths among women all over the world. Fortunately, the mortality of breast cancer can be significantly reduced via the early detection and diagnosis of breast cancer. As one of the most continually used diagnosis tools, ultrasonography (US) scan plays an important role in the detection and classification of the breast tumor. In this paper, we introduce a large breast ultrasound image database which stored breast ultrasound images and pathology results from breast tumor patients as well as their clinic diagnostic information. Furthermore, we design a web-based training system based on the database using a feature scoring scheme which based on the fifth edition of Breast Imaging Reporting and Data System (BI-RADS) lexicon for US. This online training system (new web-based teaching framework) automatically creates case-based exercises to train and guide the newly employed or resident sonographers for diagnosis of breast cancer using breast ultrasound images based on the BI-RADS.","PeriodicalId":6474,"journal":{"name":"2017 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI)","volume":"83 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90150509","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Object identification and location used by the fruit and vegetable picking robot based on human-decision making 基于人决策的果蔬采摘机器人对目标的识别与定位
Yu Chen, Binbin Chen, Haitao Li
The key to a picking robot is to identify and locate accurately in a fruit and vegetable picking site. This paper presented a method that was based on human-decision making. The human-decision making could overcome the difficulties brought by light environment, leaves shading, fruit ripening, fruit overlapping, etc. First, the binocular vision system was applied to obtain close-range pictures of the fruit and vegetable picking site; second, the picking points were chosen by human-decision making; then, the corresponding points of picking points were clicked on the screen based on epipolar geometry; finally, the coordinate transformation was used to calculate the spatial value of the picking points. The simulation experiment of cucumber picking (4 groups, 10 picking points in each group) in lab shown the maximum errors obtained were 15.1mm in vision depth direction and 8.7mm in horizontal direction. Both errors had no regular pattern, which was caused by inaccuracy in pixel when researchers click the picking points. Meanwhile, light condition, whether sunny or cloudy, had little effect on accuracy of identification and location. The research displays that this method can satisfy the need of accurate identification and location of picking points by the robot so it can be applied in design of the fruit and vegetable picking robot, improving the picking robot's quality of simplicity and accuracy in practice.
采摘机器人的关键是在果蔬采摘现场进行准确的识别和定位。本文提出了一种基于人的决策方法。人工决策可以克服光环境、叶片遮荫、果实成熟、果实重叠等带来的困难。首先,利用双目视觉系统获取果蔬采摘现场的近景图像;第二,人工决策选择采摘点;然后,基于极极几何在屏幕上点击拾取点对应的点;最后,利用坐标变换计算拾取点的空间值。室内黄瓜采摘模拟实验(4组,每组10个采摘点)显示,获得的最大误差在视觉深度方向为15.1mm,在水平方向为8.7mm。这两种错误都没有规则模式,这是由于研究人员点击拾取点时像素不准确造成的。同时,光照条件,无论是晴天还是阴天,对识别和定位的准确性影响不大。研究表明,该方法能够满足果蔬采摘机器人对采摘点的准确识别和定位需求,可应用于果蔬采摘机器人的设计中,在实践中提高了采摘机器人的简洁性和准确性。
{"title":"Object identification and location used by the fruit and vegetable picking robot based on human-decision making","authors":"Yu Chen, Binbin Chen, Haitao Li","doi":"10.1109/CISP-BMEI.2017.8302010","DOIUrl":"https://doi.org/10.1109/CISP-BMEI.2017.8302010","url":null,"abstract":"The key to a picking robot is to identify and locate accurately in a fruit and vegetable picking site. This paper presented a method that was based on human-decision making. The human-decision making could overcome the difficulties brought by light environment, leaves shading, fruit ripening, fruit overlapping, etc. First, the binocular vision system was applied to obtain close-range pictures of the fruit and vegetable picking site; second, the picking points were chosen by human-decision making; then, the corresponding points of picking points were clicked on the screen based on epipolar geometry; finally, the coordinate transformation was used to calculate the spatial value of the picking points. The simulation experiment of cucumber picking (4 groups, 10 picking points in each group) in lab shown the maximum errors obtained were 15.1mm in vision depth direction and 8.7mm in horizontal direction. Both errors had no regular pattern, which was caused by inaccuracy in pixel when researchers click the picking points. Meanwhile, light condition, whether sunny or cloudy, had little effect on accuracy of identification and location. The research displays that this method can satisfy the need of accurate identification and location of picking points by the robot so it can be applied in design of the fruit and vegetable picking robot, improving the picking robot's quality of simplicity and accuracy in practice.","PeriodicalId":6474,"journal":{"name":"2017 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI)","volume":"9 1","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90483943","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Convolutional neural networks based transfer learning for diabetic retinopathy fundus image classification 基于卷积神经网络的糖尿病视网膜病变眼底图像分类
Xiaogang Li, Tiantian Pang, B. Xiong, Weixiang Liu, Ping Liang, Tianfu Wang
Convolutional Neural Networks (CNNs) have gained remarkable success in computer vision, which is mostly owe to their ability that enables learning rich image representations from large-scale annotated data. In the field of medical image analysis, large amounts of annotated data may be not always available. The number of acquired ground-truth data is sometimes insufficient to train the CNNs without overfitting and convergence issues from scratch. Hence application of the deep CNNs is a challenge in medical imaging domain. However, transfer learning techniques are shown to provide solutions for this challenge. In this paper, our target task is to implement diabetic retinopathy fundus image classification using CNNs based transfer learning. Experiments are performed on 1014 and 1200 fundus images from two publicly available DR1 and MESSIDOR datasets. In order to complete the target task, we carry out experiments using three different methods: 1) fine-tuning all network layers of each of different pre-trained CNN models; 2) fine-tuning a pre-trained CNN model in a layer-wise manner; 3) using pre-trained CNN models to extract features from fundus images, and then training support vector machines using these features. Experimental results show that convolutional neural networks based transfer learning can achieve better classification results in our task with small datasets (target domain), by taking advantage of knowledge learned from other related tasks with larger datasets (source domain). Transfer learning is a promising technique that promotes the use of deep CNNs in medical field with limited amounts of data.
卷积神经网络(cnn)在计算机视觉领域取得了显著的成功,这主要归功于它们能够从大规模注释数据中学习丰富的图像表示。在医学图像分析领域,大量的注释数据可能并不总是可用的。获取的真值数据的数量有时不足以从头开始训练cnn,而不会出现过拟合和收敛问题。因此,深度cnn的应用是医学成像领域的一个挑战。然而,迁移学习技术为这一挑战提供了解决方案。在本文中,我们的目标任务是使用基于cnn的迁移学习实现糖尿病视网膜病变眼底图像分类。实验对来自两个公开的DR1和MESSIDOR数据集的1014和1200张眼底图像进行了实验。为了完成目标任务,我们使用了三种不同的方法进行实验:1)微调每个不同预训练CNN模型的所有网络层;2)对预训练好的CNN模型进行分层微调;3)利用预训练好的CNN模型从眼底图像中提取特征,然后利用这些特征训练支持向量机。实验结果表明,基于卷积神经网络的迁移学习可以利用从其他大数据集(源域)的相关任务中学习到的知识,在我们的小数据集(目标域)任务中获得更好的分类结果。迁移学习是一种很有前途的技术,它促进了深度cnn在数据量有限的医学领域的应用。
{"title":"Convolutional neural networks based transfer learning for diabetic retinopathy fundus image classification","authors":"Xiaogang Li, Tiantian Pang, B. Xiong, Weixiang Liu, Ping Liang, Tianfu Wang","doi":"10.1109/CISP-BMEI.2017.8301998","DOIUrl":"https://doi.org/10.1109/CISP-BMEI.2017.8301998","url":null,"abstract":"Convolutional Neural Networks (CNNs) have gained remarkable success in computer vision, which is mostly owe to their ability that enables learning rich image representations from large-scale annotated data. In the field of medical image analysis, large amounts of annotated data may be not always available. The number of acquired ground-truth data is sometimes insufficient to train the CNNs without overfitting and convergence issues from scratch. Hence application of the deep CNNs is a challenge in medical imaging domain. However, transfer learning techniques are shown to provide solutions for this challenge. In this paper, our target task is to implement diabetic retinopathy fundus image classification using CNNs based transfer learning. Experiments are performed on 1014 and 1200 fundus images from two publicly available DR1 and MESSIDOR datasets. In order to complete the target task, we carry out experiments using three different methods: 1) fine-tuning all network layers of each of different pre-trained CNN models; 2) fine-tuning a pre-trained CNN model in a layer-wise manner; 3) using pre-trained CNN models to extract features from fundus images, and then training support vector machines using these features. Experimental results show that convolutional neural networks based transfer learning can achieve better classification results in our task with small datasets (target domain), by taking advantage of knowledge learned from other related tasks with larger datasets (source domain). Transfer learning is a promising technique that promotes the use of deep CNNs in medical field with limited amounts of data.","PeriodicalId":6474,"journal":{"name":"2017 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI)","volume":"87 1","pages":"1-11"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90502498","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 116
Video copy detection using histogram based spatio-temporal features 基于直方图时空特征的视频拷贝检测
Feifei Lee, Junjie Zhao, K. Kotani, Qiu Chen
We propose a robust video copy detection method in this paper, which using combined histogram-based Spatiotemporal features for massive video database. The first is based on Histogram of Oriented Gradients (HOG) descriptor, an effective descriptor for object detection. It is used for describing the global feature of a frame in video sequence. The second is based on ordinal measure representation which is robust to size variation and color shifting as temporal feature. Furthermore, by adding an active search algorithm, the spatio-temporal features are combined to achieve video copy detection fast and accurately. Experiments show that our approach outperforms traditional algorithms in running time and detection accuracy.
本文提出了一种基于组合直方图的海量视频数据库时空特征的鲁棒视频拷贝检测方法。第一种是基于HOG描述符,这是一种有效的目标检测描述符。它用于描述视频序列中帧的全局特征。第二种方法是基于有序测度表示,该方法对尺寸变化和颜色变化具有较强的鲁棒性。在此基础上,通过加入主动搜索算法,结合时空特征,实现快速、准确的视频拷贝检测。实验表明,该方法在运行时间和检测精度上都优于传统算法。
{"title":"Video copy detection using histogram based spatio-temporal features","authors":"Feifei Lee, Junjie Zhao, K. Kotani, Qiu Chen","doi":"10.1109/CISP-BMEI.2017.8301917","DOIUrl":"https://doi.org/10.1109/CISP-BMEI.2017.8301917","url":null,"abstract":"We propose a robust video copy detection method in this paper, which using combined histogram-based Spatiotemporal features for massive video database. The first is based on Histogram of Oriented Gradients (HOG) descriptor, an effective descriptor for object detection. It is used for describing the global feature of a frame in video sequence. The second is based on ordinal measure representation which is robust to size variation and color shifting as temporal feature. Furthermore, by adding an active search algorithm, the spatio-temporal features are combined to achieve video copy detection fast and accurately. Experiments show that our approach outperforms traditional algorithms in running time and detection accuracy.","PeriodicalId":6474,"journal":{"name":"2017 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI)","volume":"35 1","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80725350","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
A novel infrared and visible face fusion recognition method based on non-subsampled contourlet transform 一种基于非下采样contourlet变换的红外与可见光人脸融合识别新方法
Guodon Liu, Shuai Zhang, Zhihua Xie
Near infrared and visible face fusion recognition is an important direction in the field of unconstrained face recognition research. In this paper, a novel fusion algorithm in (NSCT) domain is proposed for infrared and visible face fusion recognition. Firstly, NSCT is used respectively to process the infrared and visible face images, which exploits the image information at multiple scales, orientations, and frequency bands. Then, to exploit the effective discriminant feature and balance the power of high-low frequency band of NSCT coefficients, the local Gabor binary pattern (LGBP) and Local Binary Pattern (LBP) are applied respectively in different frequency parts to obtain the robust representation of infrared and visible face images. Finally, the score-level fusion is used to fuse the all the features for final classification. The proposed fusion face recognition method is tested on HITSZ Lab2 visible and near infrared face database. Experiment results show that the proposed method extracts the complementary features of near-infrared and visible-light images and improves the robustness of unconstrained face recognition.
近红外与可见光人脸融合识别是无约束人脸识别领域的一个重要研究方向。提出了一种基于NSCT域的红外与可见光人脸融合识别算法。首先,利用NSCT分别对红外和可见光人脸图像进行处理,利用多尺度、多方向、多频段的图像信息;然后,为了利用NSCT系数的有效判别特征,平衡NSCT系数的高低频带功率,分别在不同频率部分应用局部Gabor二值模式(LGBP)和局部二值模式(LBP),获得红外和可见光人脸图像的鲁棒表示。最后,采用分数级融合对所有特征进行融合,进行最终分类。在HITSZ Lab2可见光和近红外人脸数据库上对所提出的融合人脸识别方法进行了测试。实验结果表明,该方法提取了近红外和可见光图像的互补特征,提高了无约束人脸识别的鲁棒性。
{"title":"A novel infrared and visible face fusion recognition method based on non-subsampled contourlet transform","authors":"Guodon Liu, Shuai Zhang, Zhihua Xie","doi":"10.1109/CISP-BMEI.2017.8301965","DOIUrl":"https://doi.org/10.1109/CISP-BMEI.2017.8301965","url":null,"abstract":"Near infrared and visible face fusion recognition is an important direction in the field of unconstrained face recognition research. In this paper, a novel fusion algorithm in (NSCT) domain is proposed for infrared and visible face fusion recognition. Firstly, NSCT is used respectively to process the infrared and visible face images, which exploits the image information at multiple scales, orientations, and frequency bands. Then, to exploit the effective discriminant feature and balance the power of high-low frequency band of NSCT coefficients, the local Gabor binary pattern (LGBP) and Local Binary Pattern (LBP) are applied respectively in different frequency parts to obtain the robust representation of infrared and visible face images. Finally, the score-level fusion is used to fuse the all the features for final classification. The proposed fusion face recognition method is tested on HITSZ Lab2 visible and near infrared face database. Experiment results show that the proposed method extracts the complementary features of near-infrared and visible-light images and improves the robustness of unconstrained face recognition.","PeriodicalId":6474,"journal":{"name":"2017 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI)","volume":"27 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83263371","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
2017 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1