Pub Date : 2024-03-31DOI: 10.1109/TMRB.2024.3407590
Elaheh Arefinia;Jayender Jagadeesan;Rajni V. Patel
Catheter-based cardiac ablation is a minimally invasive procedure for treating atrial fibrillation (AF). Electrophysiologists perform the procedure under image guidance during which the contact force between the heart tissue and the catheter tip determines the quality of lesions created. This paper describes a novel multi-modal contact force estimator based on Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs). The estimator takes the shape and optical flow of the deflectable distal section as two modalities since frames and motion between frames complement each other to capture the long context in the video frames of the catheter. The angle between the tissue and the catheter tip is considered a complement of the extracted shape. The data acquisition platform measures the two-degrees-of-freedom contact force and video data as the catheter motion is constrained in the imaging plane. The images are captured via a camera that simulates single-view fluoroscopy for experimental purposes. In this sensor-free procedure, the features of the images and optical flow modalities are extracted through transfer learning. Long Short-Term Memory Networks (LSTMs) with a memory fusion network (MFN) are implemented to consider time dependency and hysteresis due to friction. The architecture integrates spatial and temporal networks. Late fusion with the concatenation of LSTMs, transformer decoders, and Gated Recurrent Units (GRUs) are implemented to verify the feasibility of the proposed network-based approach and its superiority over single-modality networks. The resulting mean absolute error, which accounted for only 2.84% of the total magnitude, was obtained by collecting data under more realistic circumstances in contrast to previous research studies. The decrease in error is considerably better than that achieved by individual modalities and late fusion with concatenation. These results emphasize the practicality and relevance of utilizing a multi-modal network in real-world scenarios.
{"title":"Machine-Learning-Based Multi-Modal Force Estimation for Steerable Ablation Catheters","authors":"Elaheh Arefinia;Jayender Jagadeesan;Rajni V. Patel","doi":"10.1109/TMRB.2024.3407590","DOIUrl":"https://doi.org/10.1109/TMRB.2024.3407590","url":null,"abstract":"Catheter-based cardiac ablation is a minimally invasive procedure for treating atrial fibrillation (AF). Electrophysiologists perform the procedure under image guidance during which the contact force between the heart tissue and the catheter tip determines the quality of lesions created. This paper describes a novel multi-modal contact force estimator based on Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs). The estimator takes the shape and optical flow of the deflectable distal section as two modalities since frames and motion between frames complement each other to capture the long context in the video frames of the catheter. The angle between the tissue and the catheter tip is considered a complement of the extracted shape. The data acquisition platform measures the two-degrees-of-freedom contact force and video data as the catheter motion is constrained in the imaging plane. The images are captured via a camera that simulates single-view fluoroscopy for experimental purposes. In this sensor-free procedure, the features of the images and optical flow modalities are extracted through transfer learning. Long Short-Term Memory Networks (LSTMs) with a memory fusion network (MFN) are implemented to consider time dependency and hysteresis due to friction. The architecture integrates spatial and temporal networks. Late fusion with the concatenation of LSTMs, transformer decoders, and Gated Recurrent Units (GRUs) are implemented to verify the feasibility of the proposed network-based approach and its superiority over single-modality networks. The resulting mean absolute error, which accounted for only 2.84% of the total magnitude, was obtained by collecting data under more realistic circumstances in contrast to previous research studies. The decrease in error is considerably better than that achieved by individual modalities and late fusion with concatenation. These results emphasize the practicality and relevance of utilizing a multi-modal network in real-world scenarios.","PeriodicalId":73318,"journal":{"name":"IEEE transactions on medical robotics and bionics","volume":"6 3","pages":"1004-1016"},"PeriodicalIF":3.4,"publicationDate":"2024-03-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141965833","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-31DOI: 10.1109/TMRB.2024.3407384
Wenkang Fan;Wenjing Jiang;Hong Shi;Hui-Qing Zeng;Yinran Chen;Xiongbiao Luo
Accurate deeply learned dense depth prediction remains a challenge to monocular vision reconstruction. Compared to monocular depth estimation from natural images, endoscopic dense depth prediction is even more challenging. While it is difficult to annotate endoscopic video data for supervised learning, endoscopic video images certainly suffer from illumination variations (limited lighting source, limited field of viewing, and specular highlight), smooth and textureless surfaces in surgical complex fields. This work explores a new deep learning framework of triple-supervised convolutional transformer aggregation (TSCTA) for monocular endoscopic dense depth recovery without annotating any data. Specifically, TSCTA creates convolutional transformer aggregation networks with a new hybrid encoder that combines dense convolution and scalable transformers to parallel extract local texture features and global spatial-temporal features, while it builds a local and global aggregation decoder to effectively aggregate global features and local features from coarse to fine. Moreover, we develop a self-supervised learning framework with triple supervision, which integrates minimum photometric consistency and depth consistency with sparse depth self-supervision to train our model by unannotated data. We evaluated TSCTA on unannotated monocular endoscopic images collected from various surgical procedures, with the experimental results showing that our methods can achieve more accurate depth range, more complete depth distribution, more sufficient textures, better qualitative and quantitative assessment results than state-of-the-art deeply learned monocular dense depth estimation methods.
{"title":"Triple-Supervised Convolutional Transformer Aggregation for Robust Monocular Endoscopic Dense Depth Estimation","authors":"Wenkang Fan;Wenjing Jiang;Hong Shi;Hui-Qing Zeng;Yinran Chen;Xiongbiao Luo","doi":"10.1109/TMRB.2024.3407384","DOIUrl":"https://doi.org/10.1109/TMRB.2024.3407384","url":null,"abstract":"Accurate deeply learned dense depth prediction remains a challenge to monocular vision reconstruction. Compared to monocular depth estimation from natural images, endoscopic dense depth prediction is even more challenging. While it is difficult to annotate endoscopic video data for supervised learning, endoscopic video images certainly suffer from illumination variations (limited lighting source, limited field of viewing, and specular highlight), smooth and textureless surfaces in surgical complex fields. This work explores a new deep learning framework of triple-supervised convolutional transformer aggregation (TSCTA) for monocular endoscopic dense depth recovery without annotating any data. Specifically, TSCTA creates convolutional transformer aggregation networks with a new hybrid encoder that combines dense convolution and scalable transformers to parallel extract local texture features and global spatial-temporal features, while it builds a local and global aggregation decoder to effectively aggregate global features and local features from coarse to fine. Moreover, we develop a self-supervised learning framework with triple supervision, which integrates minimum photometric consistency and depth consistency with sparse depth self-supervision to train our model by unannotated data. We evaluated TSCTA on unannotated monocular endoscopic images collected from various surgical procedures, with the experimental results showing that our methods can achieve more accurate depth range, more complete depth distribution, more sufficient textures, better qualitative and quantitative assessment results than state-of-the-art deeply learned monocular dense depth estimation methods.","PeriodicalId":73318,"journal":{"name":"IEEE transactions on medical robotics and bionics","volume":"6 3","pages":"1017-1029"},"PeriodicalIF":3.4,"publicationDate":"2024-03-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141964814","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-30DOI: 10.1109/TMRB.2024.3407330
Daniel Costa;Gianni Borghesan;Mouloud Ourak;António Pedro Aguiar;Yuyu Cai;Emmanuel Vander Poorten
Fetoscopic Endoluminal Tracheal Occlusion (FETO) is a minimally invasive fetal surgery (MIFS) aimed at mitigating the effects of Congenital Diaphragmatic Hernia (CDH). During FETO, a latex balloon is introduced in the fetal trachea using a fetoscope. Typically, this surgery is performed under ultrasound guidance which is provided by a sonographer who manually operates the ultrasound probe. This manual operation imposes a considerable physical and cognitive demand, placing a burden on the sonographer. This paper proposes a robotic ultrasound-based instrument tracking system that automates the probe position control while ensuring continuous visibility of the fetoscope in ultrasound images. The development of the proposed system is achieved with the completion of two tasks. Firstly, a series of fetoscope localization algorithms are developed and compared. Secondly, a task-based control for a robotic ultrasound system is developed. The localization algorithms’ performance is evaluated on annotated ultrasound datasets. The OEU-Net algorithm is selected based on this evaluation and is implemented in the instrument tracking system. The performance assessment of the tracking system shows that it is capable of tracking the fetoscope with a mean error below 4 mm. Thus, the developed system represents a significant advancement toward automatic robotic assistance for ultrasound guidance during FETO.
{"title":"Robotic Ultrasound-Guided Instrument Localization in Fetoscopy","authors":"Daniel Costa;Gianni Borghesan;Mouloud Ourak;António Pedro Aguiar;Yuyu Cai;Emmanuel Vander Poorten","doi":"10.1109/TMRB.2024.3407330","DOIUrl":"https://doi.org/10.1109/TMRB.2024.3407330","url":null,"abstract":"Fetoscopic Endoluminal Tracheal Occlusion (FETO) is a minimally invasive fetal surgery (MIFS) aimed at mitigating the effects of Congenital Diaphragmatic Hernia (CDH). During FETO, a latex balloon is introduced in the fetal trachea using a fetoscope. Typically, this surgery is performed under ultrasound guidance which is provided by a sonographer who manually operates the ultrasound probe. This manual operation imposes a considerable physical and cognitive demand, placing a burden on the sonographer. This paper proposes a robotic ultrasound-based instrument tracking system that automates the probe position control while ensuring continuous visibility of the fetoscope in ultrasound images. The development of the proposed system is achieved with the completion of two tasks. Firstly, a series of fetoscope localization algorithms are developed and compared. Secondly, a task-based control for a robotic ultrasound system is developed. The localization algorithms’ performance is evaluated on annotated ultrasound datasets. The OEU-Net algorithm is selected based on this evaluation and is implemented in the instrument tracking system. The performance assessment of the tracking system shows that it is capable of tracking the fetoscope with a mean error below 4 mm. Thus, the developed system represents a significant advancement toward automatic robotic assistance for ultrasound guidance during FETO.","PeriodicalId":73318,"journal":{"name":"IEEE transactions on medical robotics and bionics","volume":"6 3","pages":"806-817"},"PeriodicalIF":3.4,"publicationDate":"2024-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141965880","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-30DOI: 10.1109/TMRB.2024.3407194
Gregorio Tagliabue;Vishal Raveendranathan;Amedeo Gariboldi;Lennard Y. Hut;Andrea Zucchelli;Raffaella Carloni
Powered prosthetic legs have the potential of significantly enhancing the mobility, independence, and overall quality of life of individuals with lower-limb amputation. Unfortunately, powered prosthesis are followed by the issue of their weight and limited battery life when compared to passive or semi-active prosthesis, which, conversely, lack of complex movement capabilities. In this paper, we present an innovative design and the development of a powered prosthetic knee joint, which is actuated by means of a compact variable stiffness actuator. This innovative and promising technology can provide adaptability to different activities of daily living, while also ensuring energy efficiency and maintaining a lightweight design. The key feature of this novel powered knee joint lies in the use of a mechanism that can vary the stiffness of the joint through newly designed non-linear elastic elements. By applying advanced finite element analysis in the design process, a robust device has been realized that could readily comply with the ISO 10328.2016 standard for structural integrity. This made the knee joint suitable for future clinical trials with people with above-knee amputation.
动力假肢有可能大大提高下肢截肢者的活动能力、独立性和整体生活质量。遗憾的是,与被动或半主动假肢相比,动力假肢存在重量大、电池寿命有限等问题,而被动或半主动假肢则缺乏复杂的运动能力。在本文中,我们介绍了一种创新设计和开发的动力假体膝关节,它由一个紧凑的可变刚度致动器驱动。这种创新且前景广阔的技术能够适应不同的日常生活活动,同时还能确保能源效率并保持轻质设计。这种新型动力膝关节的主要特点在于采用了一种机制,可通过新设计的非线性弹性元件来改变关节的刚度。通过在设计过程中应用先进的有限元分析,实现了一种坚固耐用的装置,可轻松符合 ISO 10328.2016 结构完整性标准。因此,该膝关节适用于未来对膝部以上截肢者进行临床试验。
{"title":"MyKnee: Mechatronic Design of a Novel Powered Variable Stiffness Prosthetic Knee","authors":"Gregorio Tagliabue;Vishal Raveendranathan;Amedeo Gariboldi;Lennard Y. Hut;Andrea Zucchelli;Raffaella Carloni","doi":"10.1109/TMRB.2024.3407194","DOIUrl":"https://doi.org/10.1109/TMRB.2024.3407194","url":null,"abstract":"Powered prosthetic legs have the potential of significantly enhancing the mobility, independence, and overall quality of life of individuals with lower-limb amputation. Unfortunately, powered prosthesis are followed by the issue of their weight and limited battery life when compared to passive or semi-active prosthesis, which, conversely, lack of complex movement capabilities. In this paper, we present an innovative design and the development of a powered prosthetic knee joint, which is actuated by means of a compact variable stiffness actuator. This innovative and promising technology can provide adaptability to different activities of daily living, while also ensuring energy efficiency and maintaining a lightweight design. The key feature of this novel powered knee joint lies in the use of a mechanism that can vary the stiffness of the joint through newly designed non-linear elastic elements. By applying advanced finite element analysis in the design process, a robust device has been realized that could readily comply with the ISO 10328.2016 standard for structural integrity. This made the knee joint suitable for future clinical trials with people with above-knee amputation.","PeriodicalId":73318,"journal":{"name":"IEEE transactions on medical robotics and bionics","volume":"6 3","pages":"1190-1201"},"PeriodicalIF":3.4,"publicationDate":"2024-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141965575","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-30DOI: 10.1109/TMRB.2024.3407532
Olivier Lecompte;Sofiane Achiche;Abolfazl Mohebbi
Upper extremity prostheses have seen significant technological advances in recent years, primarily with the advent of myoelectric prostheses and other designs incorporating mechatronic elements. Although they do not replicate the functionality of the natural hand, users now have a way of communicating their movement intentions to the prosthesis. However, the lack of physiological feedback from the device to the user can hinder proper integration of the prosthesis, and can be a contributing factor in the rejection of the technology. This is why experts point out that sensory feedback is one of the main missing features of commercial prostheses. The literature surrounding the restoration of somatosensation primarily discusses strategies to emulate tactile perception, but few address proprioceptive perception, which is the ability to perceive limb position and movement. Yet, proprioception has been shown to be a crucial element in object manipulation. This article offers an in-depth look into the literature surrounding proprioceptive perception restoration strategies for users of upper limb prostheses by identifying and comparing the documented strategies in relation to the concept of an optimal sensory feedback restoration device.
{"title":"A Review of Proprioceptive Feedback Strategies for Upper-Limb Myoelectric Prostheses","authors":"Olivier Lecompte;Sofiane Achiche;Abolfazl Mohebbi","doi":"10.1109/TMRB.2024.3407532","DOIUrl":"https://doi.org/10.1109/TMRB.2024.3407532","url":null,"abstract":"Upper extremity prostheses have seen significant technological advances in recent years, primarily with the advent of myoelectric prostheses and other designs incorporating mechatronic elements. Although they do not replicate the functionality of the natural hand, users now have a way of communicating their movement intentions to the prosthesis. However, the lack of physiological feedback from the device to the user can hinder proper integration of the prosthesis, and can be a contributing factor in the rejection of the technology. This is why experts point out that sensory feedback is one of the main missing features of commercial prostheses. The literature surrounding the restoration of somatosensation primarily discusses strategies to emulate tactile perception, but few address proprioceptive perception, which is the ability to perceive limb position and movement. Yet, proprioception has been shown to be a crucial element in object manipulation. This article offers an in-depth look into the literature surrounding proprioceptive perception restoration strategies for users of upper limb prostheses by identifying and comparing the documented strategies in relation to the concept of an optimal sensory feedback restoration device.","PeriodicalId":73318,"journal":{"name":"IEEE transactions on medical robotics and bionics","volume":"6 3","pages":"930-939"},"PeriodicalIF":3.4,"publicationDate":"2024-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141965885","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-30DOI: 10.1109/TMRB.2024.3407350
David G. Black;Amir Hossein Hadi Hosseinabadi;Nicholas Rangga Pradnyawira;Mika Nogami;Septimu E. Salcudean
Force/torque sensing on hand-held tools enables control of applied forces, which is often essential in both tele-robotics and remote guidance of people. However, existing force sensors are either bulky, complex, or have insufficient load rating. This paper presents a novel 6 axis force-torque sensor based on differential magnetic field readings in a collection of low-profile sensor modules placed around a tool or device. The instrumentation is easy to install but nonetheless achieves good performance. A detailed mathematical model and optimization-based design procedure are also introduced. The modeling, simulation, and optimization of the force sensor are described and then used in the electrical and mechanical design and integration of the sensor into an ultrasound probe. Through a neural network-based nonlinear calibration, the sensor achieves average root-mean-square test errors of 0.41 N and 0.027 Nm compared to an off-the-shelf ATI Nano25 sensor, which are 0.80% and 1.16% of the full-scale range respectively. The sensor has an average noise power spectral density of less than 0.0001 N/ $sqrt {text {Hz}}$