首页 > 最新文献

Artificial Life and Robotics最新文献

英文 中文
Spiking neural networks-based generation of caterpillar-like soft robot crawling motions 基于尖峰神经网络生成类似毛毛虫的软机器人爬行动作
IF 0.8 Q4 ROBOTICS Pub Date : 2024-10-04 DOI: 10.1007/s10015-024-00970-7
SeanKein Yoshioka, Takahiro Iwata, Yuki Maruyama, Daisuke Miki

Robots have been widely used in daily life in recent years. Unlike conventional robots made of rigid materials, soft robots utilize stretchable and flexible materials, allowing flexible movements similar to those of living organisms, which are difficult for traditional robots. Previous studies have used periodic signals to control soft robots, which lead to repetitive motions and make it challenging to generate environment-adapted motions. To address this issue, control methods can be learned through deep reinforcement learning to enable soft robots to select appropriate actions based on observations, improving their adaptability to environmental changes. In addition, as mobile robots have limited onboard resources, it is necessary to conserve battery consumption and achieve low-power control. Therefore, the use of spiking neural networks (SNNs) with neuromorphic chips enables low-power control of soft robots. In this study, we investigated the learning methods for SNNs aimed at controlling soft robots. Experiments were conducted using a caterpillar-like soft robot model based on previous studies, and the effectiveness of the learning method was evaluated.

近年来,机器人在日常生活中得到了广泛应用。与刚性材料制成的传统机器人不同,软体机器人利用可伸缩的柔性材料,实现类似于生物体的灵活运动,这是传统机器人难以实现的。以往的研究使用周期性信号来控制软体机器人,这会导致重复性运动,并给产生适应环境的运动带来挑战。为解决这一问题,可以通过深度强化学习来学习控制方法,使软机器人能够根据观察结果选择适当的动作,提高其对环境变化的适应能力。此外,由于移动机器人的机载资源有限,因此有必要节省电池消耗,实现低功耗控制。因此,使用带有神经形态芯片的尖峰神经网络(SNN)可以实现软体机器人的低功耗控制。在本研究中,我们研究了旨在控制软体机器人的尖峰神经网络的学习方法。在以往研究的基础上,使用类似毛毛虫的软体机器人模型进行了实验,并评估了学习方法的有效性。
{"title":"Spiking neural networks-based generation of caterpillar-like soft robot crawling motions","authors":"SeanKein Yoshioka,&nbsp;Takahiro Iwata,&nbsp;Yuki Maruyama,&nbsp;Daisuke Miki","doi":"10.1007/s10015-024-00970-7","DOIUrl":"10.1007/s10015-024-00970-7","url":null,"abstract":"<div><p>Robots have been widely used in daily life in recent years. Unlike conventional robots made of rigid materials, soft robots utilize stretchable and flexible materials, allowing flexible movements similar to those of living organisms, which are difficult for traditional robots. Previous studies have used periodic signals to control soft robots, which lead to repetitive motions and make it challenging to generate environment-adapted motions. To address this issue, control methods can be learned through deep reinforcement learning to enable soft robots to select appropriate actions based on observations, improving their adaptability to environmental changes. In addition, as mobile robots have limited onboard resources, it is necessary to conserve battery consumption and achieve low-power control. Therefore, the use of spiking neural networks (SNNs) with neuromorphic chips enables low-power control of soft robots. In this study, we investigated the learning methods for SNNs aimed at controlling soft robots. Experiments were conducted using a caterpillar-like soft robot model based on previous studies, and the effectiveness of the learning method was evaluated.</p></div>","PeriodicalId":46050,"journal":{"name":"Artificial Life and Robotics","volume":"29 4","pages":"519 - 527"},"PeriodicalIF":0.8,"publicationDate":"2024-10-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10015-024-00970-7.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142518439","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Real-time drowsiness evaluation system using marker-less facial motion capture 使用无标记面部动作捕捉的实时嗜睡评估系统
IF 0.8 Q4 ROBOTICS Pub Date : 2024-10-03 DOI: 10.1007/s10015-024-00972-5
Yudai Koshi, Hisaya Tanaka

This paper proposes a drowsiness expression rating system that can rate drowsiness in real time using only video information. Drowsiness in drivers is caused by various factors, including driving on monotonous roads, and can lead to numerous problems, e.g., traffic accidents. Previously, we developed an offline drowsiness evaluation system the uses only video image information from MediaPipe, which is a marker-less facial motion capture system. The proposed system can perform real-time drowsiness rating on multiple platforms and requires a smartphone or personal computer. Results of applied to car driving demonstrate that the accuracy of the proposed system was 89.7%, 78.8%, and 65.0% for binary, three-class, and five-class classification tasks, respectively. In addition, the proposed system outperformed existing systems in binary, three-class, and five-class classification tasks by 6.0%, 0.8%, and 4.3%, respectively. These results demonstrate that the proposed system exhibits a higher accuracy rate than the existing methods.

本文提出了一种嗜睡表情评级系统,该系统可仅利用视频信息对嗜睡程度进行实时评级。驾驶员嗜睡的原因多种多样,包括在单调的道路上驾驶,并可能导致许多问题,如交通事故。此前,我们开发了一种离线嗜睡评估系统,该系统仅使用 MediaPipe(一种无标记面部动作捕捉系统)提供的视频图像信息。所提出的系统可以在多个平台上进行实时嗜睡度评级,但需要智能手机或个人电脑。应用于汽车驾驶的结果表明,在二元分类、三元分类和五元分类任务中,拟议系统的准确率分别为 89.7%、78.8% 和 65.0%。此外,在二元分类、三元分类和五元分类任务中,建议的系统分别比现有系统高出 6.0%、0.8% 和 4.3%。这些结果表明,建议的系统比现有方法具有更高的准确率。
{"title":"Real-time drowsiness evaluation system using marker-less facial motion capture","authors":"Yudai Koshi,&nbsp;Hisaya Tanaka","doi":"10.1007/s10015-024-00972-5","DOIUrl":"10.1007/s10015-024-00972-5","url":null,"abstract":"<div><p>This paper proposes a drowsiness expression rating system that can rate drowsiness in real time using only video information. Drowsiness in drivers is caused by various factors, including driving on monotonous roads, and can lead to numerous problems, e.g., traffic accidents. Previously, we developed an offline drowsiness evaluation system the uses only video image information from MediaPipe, which is a marker-less facial motion capture system. The proposed system can perform real-time drowsiness rating on multiple platforms and requires a smartphone or personal computer. Results of applied to car driving demonstrate that the accuracy of the proposed system was 89.7%, 78.8%, and 65.0% for binary, three-class, and five-class classification tasks, respectively. In addition, the proposed system outperformed existing systems in binary, three-class, and five-class classification tasks by 6.0%, 0.8%, and 4.3%, respectively. These results demonstrate that the proposed system exhibits a higher accuracy rate than the existing methods.</p></div>","PeriodicalId":46050,"journal":{"name":"Artificial Life and Robotics","volume":"29 4","pages":"573 - 578"},"PeriodicalIF":0.8,"publicationDate":"2024-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142518521","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A two-stage image segmentation method for harvest order decision of wood ear mushroom 决定木耳采收顺序的两阶段图像分割方法
IF 0.8 Q4 ROBOTICS Pub Date : 2024-10-01 DOI: 10.1007/s10015-024-00971-6
Kazuya Okamura, Ryo Matsumura, Hironori Kitakaze

This study proposes a method for determining the appropriate harvesting order for densely growing wood ear mushrooms by recognizing their growth stages and harvesting priorities from depth images obtained from a stereo camera. We aim to minimize crop damage and improve the quality of harvested crops during the harvesting of densely growing crops using a robot arm. The proposed two-stage method consists of two models—one of the models to recognize priority harvest regions, and the other model to identify individual wood ear mushroom regions and growth stages. The final harvesting order is determined based on the outputs of these models. The models were trained using simulated CGI data of wood ear mushroom growth. The experimental results show that the appropriate harvesting order can be outputted in 57.5% of the cases for the 40 sets of test data. The results show that it is possible to determine the harvesting order of dense wood ear mushrooms based solely on depth images. However, there is still room for improvement in operations in actual environments. Further work is needed to enhance the method’s robustness and accuracy.

本研究提出了一种方法,通过从立体摄像机获得的深度图像识别木耳蘑菇的生长阶段和采收优先顺序,从而确定密集生长的木耳蘑菇的适当采收顺序。我们的目标是在使用机械臂收割生长茂密的农作物时,尽量减少对农作物的损害,并提高收割农作物的质量。所提出的两阶段方法由两个模型组成,其中一个模型用于识别优先收割区域,另一个模型用于识别单个木耳蘑菇区域和生长阶段。最终的采收顺序根据这些模型的输出确定。使用木耳蘑菇生长的模拟 CGI 数据对模型进行了训练。实验结果表明,在 40 组测试数据中,57.5% 的情况下可以输出适当的采收顺序。结果表明,仅凭深度图像就能确定密集木耳菇的采收顺序是可行的。然而,在实际环境中的操作仍有改进的余地。要提高该方法的鲁棒性和准确性,还需要进一步的工作。
{"title":"A two-stage image segmentation method for harvest order decision of wood ear mushroom","authors":"Kazuya Okamura,&nbsp;Ryo Matsumura,&nbsp;Hironori Kitakaze","doi":"10.1007/s10015-024-00971-6","DOIUrl":"10.1007/s10015-024-00971-6","url":null,"abstract":"<div><p>This study proposes a method for determining the appropriate harvesting order for densely growing wood ear mushrooms by recognizing their growth stages and harvesting priorities from depth images obtained from a stereo camera. We aim to minimize crop damage and improve the quality of harvested crops during the harvesting of densely growing crops using a robot arm. The proposed two-stage method consists of two models—one of the models to recognize priority harvest regions, and the other model to identify individual wood ear mushroom regions and growth stages. The final harvesting order is determined based on the outputs of these models. The models were trained using simulated CGI data of wood ear mushroom growth. The experimental results show that the appropriate harvesting order can be outputted in 57.5% of the cases for the 40 sets of test data. The results show that it is possible to determine the harvesting order of dense wood ear mushrooms based solely on depth images. However, there is still room for improvement in operations in actual environments. Further work is needed to enhance the method’s robustness and accuracy.</p></div>","PeriodicalId":46050,"journal":{"name":"Artificial Life and Robotics","volume":"29 4","pages":"528 - 535"},"PeriodicalIF":0.8,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142518350","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A digital hardware system for real-time biorealistic stimulation on in vitro cardiomyocytes 对体外心肌细胞进行实时生物仿真刺激的数字硬件系统
IF 0.8 Q4 ROBOTICS Pub Date : 2024-09-30 DOI: 10.1007/s10015-024-00968-1
Pierre-Marie Faure, Agnès Tixier-Mita, Timothée Levi

Every year, cardiovascular diseases cause millions of deaths worldwide. These diseases involve complex mechanisms that are difficult to study. To remedy this problem, we propose to develop a heart–brain platform capable of reproducing the mechanisms involved in generating the heartbeat. The platform will be designed to operate in real time, with the most economical and integrated design possible. To achieve this, we are implementing highly biologically coherent cellular models on FPGA, which we interconnect with in vitro cell cultures. In our case, we are using the Maltsev–Lakatta cell model, which describes the behavior of the pacemaker cells responsible for the heart rhythm, to stimulate a cardiomyocyte culture.

每年,心血管疾病导致全球数百万人死亡。这些疾病涉及复杂的机制,难以研究。为了解决这个问题,我们建议开发一个心脑平台,能够再现产生心跳的机制。该平台将以最经济、最集成的设计实时运行。为此,我们正在 FPGA 上实现高度生物一致性的细胞模型,并将其与体外细胞培养相互连接。在我们的案例中,我们正在使用 Maltsev-Lakatta 细胞模型来刺激心肌细胞培养,该模型描述了负责心律的起搏细胞的行为。
{"title":"A digital hardware system for real-time biorealistic stimulation on in vitro cardiomyocytes","authors":"Pierre-Marie Faure,&nbsp;Agnès Tixier-Mita,&nbsp;Timothée Levi","doi":"10.1007/s10015-024-00968-1","DOIUrl":"10.1007/s10015-024-00968-1","url":null,"abstract":"<div><p>Every year, cardiovascular diseases cause millions of deaths worldwide. These diseases involve complex mechanisms that are difficult to study. To remedy this problem, we propose to develop a heart–brain platform capable of reproducing the mechanisms involved in generating the heartbeat. The platform will be designed to operate in real time, with the most economical and integrated design possible. To achieve this, we are implementing highly biologically coherent cellular models on FPGA, which we interconnect with in vitro cell cultures. In our case, we are using the Maltsev–Lakatta cell model, which describes the behavior of the pacemaker cells responsible for the heart rhythm, to stimulate a cardiomyocyte culture.</p></div>","PeriodicalId":46050,"journal":{"name":"Artificial Life and Robotics","volume":"29 4","pages":"473 - 478"},"PeriodicalIF":0.8,"publicationDate":"2024-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142519130","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Biomimetic snake locomotion using central pattern generators network and bio-hybrid robot perspective 从中央模式发生器网络和生物混合机器人的角度看仿生蛇运动
IF 0.8 Q4 ROBOTICS Pub Date : 2024-09-25 DOI: 10.1007/s10015-024-00969-0
Jérémy Cheslet, Romain Beaubois, Tomoya Duenki, Farad Khoyratee, Takashi Kohno, Yoshiho Ikeuchi, Timothée Lévi

Neurological disorders affect millions globally and necessitate advanced treatments, especially with an aging population. Brain Machine Interfaces (BMIs) and neuroprostheses show promise in addressing disabilities by mimicking biological dynamics through biomimetic Spiking Neural Networks (SNNs). Central Pattern Generators (CPGs) are small neural networks that, emulated through biomimetic networks, can replicate specific locomotion patterns. Our proposal involves a real-time implementation of a biomimetic SNN on FPGA, utilizing biomimetic models for neurons, synaptic receptors and synaptic plasticity. The system, integrated into a snake-like mobile robot where the neuronal activity is responsible for its locomotion, offers a versatile platform to study spinal cord injuries. Lastly, we present a preliminary closed-loop experiment involving bidirectional interaction between the artificial neural network and biological neuronal cells, paving the way for bio-hybrid robots and insights into neural population functioning.

神经系统疾病影响着全球数百万人,需要先进的治疗方法,尤其是在人口老龄化的情况下。脑机接口(BMI)和神经义肢通过仿生尖峰神经网络(SNN)模拟生物动力学,有望解决残疾问题。中央模式发生器(CPG)是一种小型神经网络,通过生物仿真网络进行模拟,可以复制特定的运动模式。我们的建议包括在 FPGA 上实时实现生物仿真 SNN,利用生物仿真模型来模拟神经元、突触受体和突触可塑性。该系统集成在一个蛇形移动机器人中,神经元活动负责其运动,为研究脊髓损伤提供了一个多功能平台。最后,我们介绍了一个初步的闭环实验,涉及人工神经网络与生物神经细胞之间的双向互动,为生物混合机器人和深入了解神经群体的功能铺平了道路。
{"title":"Biomimetic snake locomotion using central pattern generators network and bio-hybrid robot perspective","authors":"Jérémy Cheslet,&nbsp;Romain Beaubois,&nbsp;Tomoya Duenki,&nbsp;Farad Khoyratee,&nbsp;Takashi Kohno,&nbsp;Yoshiho Ikeuchi,&nbsp;Timothée Lévi","doi":"10.1007/s10015-024-00969-0","DOIUrl":"10.1007/s10015-024-00969-0","url":null,"abstract":"<div><p>Neurological disorders affect millions globally and necessitate advanced treatments, especially with an aging population. Brain Machine Interfaces (BMIs) and neuroprostheses show promise in addressing disabilities by mimicking biological dynamics through biomimetic Spiking Neural Networks (SNNs). Central Pattern Generators (CPGs) are small neural networks that, emulated through biomimetic networks, can replicate specific locomotion patterns. Our proposal involves a real-time implementation of a biomimetic SNN on FPGA, utilizing biomimetic models for neurons, synaptic receptors and synaptic plasticity. The system, integrated into a snake-like mobile robot where the neuronal activity is responsible for its locomotion, offers a versatile platform to study spinal cord injuries. Lastly, we present a preliminary closed-loop experiment involving bidirectional interaction between the artificial neural network and biological neuronal cells, paving the way for bio-hybrid robots and insights into neural population functioning.</p></div>","PeriodicalId":46050,"journal":{"name":"Artificial Life and Robotics","volume":"29 4","pages":"479 - 485"},"PeriodicalIF":0.8,"publicationDate":"2024-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142518730","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A comparative study of linear and nonlinear regression models for blood glucose estimation based on near-infrared facial images from 760 to 1650 nm wavelength 基于波长为 760 至 1650 纳米的近红外面部图像的线性和非线性血糖估计回归模型比较研究
IF 0.8 Q4 ROBOTICS Pub Date : 2024-09-20 DOI: 10.1007/s10015-024-00961-8
Mayuko Nakagawa, Kosuke Oiwa, Yasushi Nanai, Kent Nagumo, Akio Nozawa

We have attempted to estimate blood glucose levels based on facial images measured in the near-infrared band, which is highly biopermeable, to establish a remote minimally invasive blood glucose measurement method. We measured facial images in the near-infrared wavelength range of 760–1650 nm, and constructed a general model for blood glucose level estimation by linear regression using the weights of spatial features of the measured facial images as explanatory variables. The results showed that the accuracy values of blood glucose estimation in the generalization performance evaluation were 43.02 mg/dL for NIR-I (760–1100 nm) and 43.61 mg/dL for NIR-II (1050–1650 nm) in the RMSE of the general model. Since biological information is nonlinear, it is necessary to explore suitable modeling methods for blood glucose estimation, including not only linear regression but also nonlinear regression. The purpose of this study is to explore suitable regression methods among linear and nonlinear regression methods to construct a blood glucose estimation model based on facial images with wavelengths from 760 to 1650 nm. The results showed that model using Random Forest had the best estimation accuracy with an RMSE of 36.02 mg/dL in NIR-I and the MR model had the best estimation accuracy with RMSE of 36.70 mg/dL in NIR-II under the current number of subjects and measurement data points. The independent components selected for the model have spatial features considered to be simply individual differences that are not related to blood glucose variation.

我们尝试根据在生物渗透性很强的近红外波段测量的面部图像来估算血糖水平,从而建立一种远程微创血糖测量方法。我们测量了波长范围为 760-1650 nm 的近红外波段面部图像,并以所测量的面部图像的空间特征权重为解释变量,通过线性回归构建了血糖水平估算的一般模型。结果表明,在泛化性能评估中,近红外-I(760-1100 nm)和近红外-II(1050-1650 nm)血糖估测的准确度值分别为 43.02 mg/dL 和 43.61 mg/dL。由于生物信息是非线性的,因此有必要探索适合血糖估算的建模方法,不仅包括线性回归,还包括非线性回归。本研究的目的是在线性回归和非线性回归方法中探索合适的回归方法,以构建基于波长为 760 至 1650 nm 的面部图像的血糖估测模型。结果表明,在现有受试者人数和测量数据点的情况下,使用随机森林的模型在近红外-Ⅰ中的估计精度最高,RMSE 为 36.02 mg/dL;MR 模型在近红外-Ⅱ中的估计精度最高,RMSE 为 36.70 mg/dL。该模型所选的独立成分具有空间特征,被认为只是个体差异,与血糖变化无关。
{"title":"A comparative study of linear and nonlinear regression models for blood glucose estimation based on near-infrared facial images from 760 to 1650 nm wavelength","authors":"Mayuko Nakagawa,&nbsp;Kosuke Oiwa,&nbsp;Yasushi Nanai,&nbsp;Kent Nagumo,&nbsp;Akio Nozawa","doi":"10.1007/s10015-024-00961-8","DOIUrl":"10.1007/s10015-024-00961-8","url":null,"abstract":"<div><p>We have attempted to estimate blood glucose levels based on facial images measured in the near-infrared band, which is highly biopermeable, to establish a remote minimally invasive blood glucose measurement method. We measured facial images in the near-infrared wavelength range of 760–1650 nm, and constructed a general model for blood glucose level estimation by linear regression using the weights of spatial features of the measured facial images as explanatory variables. The results showed that the accuracy values of blood glucose estimation in the generalization performance evaluation were 43.02 mg/dL for NIR-I (760–1100 nm) and 43.61 mg/dL for NIR-II (1050–1650 nm) in the RMSE of the general model. Since biological information is nonlinear, it is necessary to explore suitable modeling methods for blood glucose estimation, including not only linear regression but also nonlinear regression. The purpose of this study is to explore suitable regression methods among linear and nonlinear regression methods to construct a blood glucose estimation model based on facial images with wavelengths from 760 to 1650 nm. The results showed that model using Random Forest had the best estimation accuracy with an RMSE of 36.02 mg/dL in NIR-I and the MR model had the best estimation accuracy with RMSE of 36.70 mg/dL in NIR-II under the current number of subjects and measurement data points. The independent components selected for the model have spatial features considered to be simply individual differences that are not related to blood glucose variation.</p></div>","PeriodicalId":46050,"journal":{"name":"Artificial Life and Robotics","volume":"29 4","pages":"501 - 509"},"PeriodicalIF":0.8,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142519068","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Integration of multiple dense point clouds based on estimated parameters in photogrammetry with QR code for reducing computation time 基于摄影测量中的估计参数,用 QR 码整合多个密集点云,以减少计算时间
IF 0.8 Q4 ROBOTICS Pub Date : 2024-09-20 DOI: 10.1007/s10015-024-00966-3
Keita Nakamura, Keita Baba, Yutaka Watanobe, Toshihide Hanari, Taku Matsumoto, Takashi Imabuchi, Kuniaki Kawabata

This paper describes a method for integrating multiple dense point clouds using a shared landmark to generate a single real-scale integrated result for photogrammetry. It is difficult to integrate high-density point clouds reconstructed by photogrammetry because the scale differs with each photogrammetry. To solve this problem, this study places a QR code of known sizes, which is a shared landmark, in the reconstruction target environment and divides the reconstruction target environment based on the position of the QR code that is placed. Then, photogrammetry is performed for each divided environment to obtain each high-density point cloud. Finally, we propose a method of scaling each high-density point cloud based on the size of the QR code and aligning each high-density point cloud as a single high-point cloud by partial-to-partial registration. To verify the effectiveness of the method, this paper compares the results obtained by applying all images to photogrammetry with those obtained by the proposed method in terms of accuracy and computation time. In this verification, ideal images generated by simulation and images obtained in real environments are applied to photogrammetry. We clarify the relationship between the number of divided environments, the accuracy of the reconstruction result, and the computation time required for the reconstruction.

本文介绍了一种利用共享地标整合多个高密度点云的方法,以便为摄影测量生成单一真实比例尺的整合结果。由摄影测量重建的高密度点云很难整合,因为每次摄影测量的比例尺都不同。为解决这一问题,本研究在重建目标环境中放置一个已知大小的二维码,作为共享地标,并根据放置二维码的位置划分重建目标环境。然后,对每个划分的环境进行摄影测量,以获得每个高密度点云。最后,我们提出了一种方法,即根据二维码的大小缩放每个高密度点云,并通过部分对部分配准将每个高密度点云配准为一个高点云。为了验证该方法的有效性,本文比较了将所有图像应用于摄影测量所获得的结果与建议方法所获得的结果在精度和计算时间方面的差异。在验证过程中,我们将模拟生成的理想图像和在真实环境中获得的图像应用于摄影测量。我们明确了划分环境的数量、重建结果的准确性和重建所需的计算时间之间的关系。
{"title":"Integration of multiple dense point clouds based on estimated parameters in photogrammetry with QR code for reducing computation time","authors":"Keita Nakamura,&nbsp;Keita Baba,&nbsp;Yutaka Watanobe,&nbsp;Toshihide Hanari,&nbsp;Taku Matsumoto,&nbsp;Takashi Imabuchi,&nbsp;Kuniaki Kawabata","doi":"10.1007/s10015-024-00966-3","DOIUrl":"10.1007/s10015-024-00966-3","url":null,"abstract":"<div><p>This paper describes a method for integrating multiple dense point clouds using a shared landmark to generate a single real-scale integrated result for photogrammetry. It is difficult to integrate high-density point clouds reconstructed by photogrammetry because the scale differs with each photogrammetry. To solve this problem, this study places a QR code of known sizes, which is a shared landmark, in the reconstruction target environment and divides the reconstruction target environment based on the position of the QR code that is placed. Then, photogrammetry is performed for each divided environment to obtain each high-density point cloud. Finally, we propose a method of scaling each high-density point cloud based on the size of the QR code and aligning each high-density point cloud as a single high-point cloud by partial-to-partial registration. To verify the effectiveness of the method, this paper compares the results obtained by applying all images to photogrammetry with those obtained by the proposed method in terms of accuracy and computation time. In this verification, ideal images generated by simulation and images obtained in real environments are applied to photogrammetry. We clarify the relationship between the number of divided environments, the accuracy of the reconstruction result, and the computation time required for the reconstruction.</p></div>","PeriodicalId":46050,"journal":{"name":"Artificial Life and Robotics","volume":"29 4","pages":"546 - 556"},"PeriodicalIF":0.8,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10015-024-00966-3.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142519067","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Obstacle feature point detection method for real-time processing by monocular camera and line laser 利用单目摄像头和线激光实时处理的障碍物特征点检测方法
IF 0.8 Q4 ROBOTICS Pub Date : 2024-09-19 DOI: 10.1007/s10015-024-00965-4
Hayato Mitsuhashi, Taku Itami

In this study, we propose a method for detecting feature points of obstacles by real-time processing using a monocular camera and a laser. Specifically, we propose an algorithm that detects laser beams emitted on obstacles by binarization, transforms pixel coordinates into distance using triangulation by a laser beam that is obliquely incident on the ground, and estimates the placement angle of the obstacle in real-time processing. No method has been devised to detect the placement angle of obstacles using a monocular camera and a laser. Furthermore, the proposed method does not calculate the linear distance from the camera, but from the position where the camera is moved horizontally parallel to the obstacle in front of the camera, thus detecting the placement angle independently of the placement position of the obstacle. As a result, it was confirmed that the placement angles of obstacles could be detected with a maximum error of (6^{circ }) in six placement positions. Therefore, we succeeded in automatically detecting the placement angle of obstacles in real-time, independent of the illumination of the measurement environment, the reflectance of the obstacle.

在本研究中,我们提出了一种利用单目摄像头和激光进行实时处理的障碍物特征点检测方法。具体来说,我们提出了一种算法,通过二值化检测障碍物上发射的激光束,利用斜入射地面的激光束三角测量法将像素坐标转换为距离,并在实时处理中估算障碍物的放置角度。目前还没有使用单目摄像头和激光检测障碍物位置角的方法。此外,所提出的方法不是从相机计算线性距离,而是从相机与相机前方障碍物平行水平移动的位置计算线性距离,因此可以独立于障碍物的放置位置检测放置角。结果证实,在六个放置位置上,障碍物的放置角度可以被检测到,最大误差为 (6^{circ }) 。因此,我们成功地实时自动检测出了障碍物的放置角度,而不受测量环境光照、障碍物反射率的影响。
{"title":"Obstacle feature point detection method for real-time processing by monocular camera and line laser","authors":"Hayato Mitsuhashi,&nbsp;Taku Itami","doi":"10.1007/s10015-024-00965-4","DOIUrl":"10.1007/s10015-024-00965-4","url":null,"abstract":"<div><p>In this study, we propose a method for detecting feature points of obstacles by real-time processing using a monocular camera and a laser. Specifically, we propose an algorithm that detects laser beams emitted on obstacles by binarization, transforms pixel coordinates into distance using triangulation by a laser beam that is obliquely incident on the ground, and estimates the placement angle of the obstacle in real-time processing. No method has been devised to detect the placement angle of obstacles using a monocular camera and a laser. Furthermore, the proposed method does not calculate the linear distance from the camera, but from the position where the camera is moved horizontally parallel to the obstacle in front of the camera, thus detecting the placement angle independently of the placement position of the obstacle. As a result, it was confirmed that the placement angles of obstacles could be detected with a maximum error of <span>(6^{circ })</span> in six placement positions. Therefore, we succeeded in automatically detecting the placement angle of obstacles in real-time, independent of the illumination of the measurement environment, the reflectance of the obstacle.</p></div>","PeriodicalId":46050,"journal":{"name":"Artificial Life and Robotics","volume":"29 4","pages":"438 - 448"},"PeriodicalIF":0.8,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10015-024-00965-4.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142518385","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Research on English–Chinese machine translation shift based on word vector similarity 基于词向量相似性的英汉机器翻译转换研究
IF 0.8 Q4 ROBOTICS Pub Date : 2024-09-16 DOI: 10.1007/s10015-024-00964-5
Qingqing Ma

In English–Chinese machine translation shift, the processing of out-of-vocabulary (OOV) words has a great impact on translation quality. Aiming at OOV, this paper proposed a method based on word vector similarity, calculated the word vector similarity based on the Skip-gram model, used the most similar words to replace OOV in the source sentences, and used the replaced corpus to train the Transformer model. It was found that when the original corpus was used for training, the bilingual evaluation understudy-4 (BLEU-4) of the Transformer model on NIST2006 and NIST2008 was 37.29 and 30.73, respectively. However, when the word vector similarity was used for processing and low-frequency OOV words were retained, the BLEU-4 of the Transformer model on NIST2006 and NIST2008 was improved to 37.36 and 30.78 respectively, showing an increase. Moreover, the translation quality obtained by retaining low-frequency OOV words was better than that obtained by removing low-frequency OOV words. The experimental results prove that the English–Chinese machine translation shift method based on word vector similarity is reliable and can be applied in practice.

在英汉机器翻译转换中,词汇外(OOV)词的处理对翻译质量有很大影响。针对 OOV,本文提出了一种基于词向量相似性的方法,基于 Skip-gram 模型计算词向量相似性,用最相似的词替换源句中的 OOV,并用替换后的语料训练 Transformer 模型。结果发现,当使用原始语料进行训练时,Transformer 模型在 NIST2006 和 NIST2008 上的双语评估劣度-4(BLEU-4)分别为 37.29 和 30.73。然而,当使用词向量相似性进行处理并保留低频 OOV 词时,Transformer 模型在 NIST2006 和 NIST2008 上的 BLEU-4 分别提高到 37.36 和 30.78,显示出提高。此外,保留低频 OOV 词所获得的翻译质量优于去除低频 OOV 词所获得的翻译质量。实验结果证明,基于词向量相似性的英汉机器翻译转换方法是可靠的,可以在实践中应用。
{"title":"Research on English–Chinese machine translation shift based on word vector similarity","authors":"Qingqing Ma","doi":"10.1007/s10015-024-00964-5","DOIUrl":"10.1007/s10015-024-00964-5","url":null,"abstract":"<div><p>In English–Chinese machine translation shift, the processing of out-of-vocabulary (OOV) words has a great impact on translation quality. Aiming at OOV, this paper proposed a method based on word vector similarity, calculated the word vector similarity based on the Skip-gram model, used the most similar words to replace OOV in the source sentences, and used the replaced corpus to train the Transformer model. It was found that when the original corpus was used for training, the bilingual evaluation understudy-4 (BLEU-4) of the Transformer model on NIST2006 and NIST2008 was 37.29 and 30.73, respectively. However, when the word vector similarity was used for processing and low-frequency OOV words were retained, the BLEU-4 of the Transformer model on NIST2006 and NIST2008 was improved to 37.36 and 30.78 respectively, showing an increase. Moreover, the translation quality obtained by retaining low-frequency OOV words was better than that obtained by removing low-frequency OOV words. The experimental results prove that the English–Chinese machine translation shift method based on word vector similarity is reliable and can be applied in practice.</p></div>","PeriodicalId":46050,"journal":{"name":"Artificial Life and Robotics","volume":"29 4","pages":"585 - 589"},"PeriodicalIF":0.8,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142518480","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Management of power equipment inspection informationization through intelligent unmanned aerial vehicles 通过智能无人机实现电力设备巡检信息化管理
IF 0.8 Q4 ROBOTICS Pub Date : 2024-09-13 DOI: 10.1007/s10015-024-00963-6
Weizhi Lu, Qiang Li, Weijian Zhang, Lin Mei, Di Cai, Zepeng Li

With the implementation of intelligent unmanned aerial vehicles (UAVs) in power equipment inspection, managing the obtained inspection results through information technology is increasingly crucial. This paper collected insulator images, including images of standard and self-exploding insulators, during the inspection process using intelligent UAVs. Then, an optimized you only look once version 5 (YOLOv5) model was developed by incorporating the convolutional block attention module and utilizing the efficient intersection-over-union loss function. The detection performance of the designed algorithm was analyzed. It was found that among different models, the YOLOv5s model exhibited the smallest size and the highest detection speed. Moreover, the optimized YOLOv5 model showed a significant improvement in speed and accuracy for insulator detection, surpassing other methods with a mean average precision of 93.81% and 145.64 frames per second. These results demonstrate the reliability of the improved YOLOv5 model and its practical applicability.

随着智能无人机(UAV)在电力设备检测中的应用,通过信息技术管理所获得的检测结果变得越来越重要。本文收集了使用智能无人机巡检过程中的绝缘子图像,包括标准绝缘子和自爆绝缘子的图像。然后,结合卷积块注意力模块,利用高效的交集-过联合损失函数,开发了优化的只看一次(YOLOv5)模型。对所设计算法的检测性能进行了分析。结果发现,在不同的模型中,YOLOv5s 模型的体积最小,检测速度最快。此外,优化后的 YOLOv5 模型在绝缘体检测速度和精度方面都有显著提高,超过了其他方法,平均精度达到 93.81%,每秒检测帧数达到 145.64 帧。这些结果证明了改进后的 YOLOv5 模型的可靠性及其实用性。
{"title":"Management of power equipment inspection informationization through intelligent unmanned aerial vehicles","authors":"Weizhi Lu,&nbsp;Qiang Li,&nbsp;Weijian Zhang,&nbsp;Lin Mei,&nbsp;Di Cai,&nbsp;Zepeng Li","doi":"10.1007/s10015-024-00963-6","DOIUrl":"10.1007/s10015-024-00963-6","url":null,"abstract":"<div><p>With the implementation of intelligent unmanned aerial vehicles (UAVs) in power equipment inspection, managing the obtained inspection results through information technology is increasingly crucial. This paper collected insulator images, including images of standard and self-exploding insulators, during the inspection process using intelligent UAVs. Then, an optimized you only look once version 5 (YOLOv5) model was developed by incorporating the convolutional block attention module and utilizing the efficient intersection-over-union loss function. The detection performance of the designed algorithm was analyzed. It was found that among different models, the YOLOv5s model exhibited the smallest size and the highest detection speed. Moreover, the optimized YOLOv5 model showed a significant improvement in speed and accuracy for insulator detection, surpassing other methods with a mean average precision of 93.81% and 145.64 frames per second. These results demonstrate the reliability of the improved YOLOv5 model and its practical applicability.</p></div>","PeriodicalId":46050,"journal":{"name":"Artificial Life and Robotics","volume":"29 4","pages":"579 - 584"},"PeriodicalIF":0.8,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142518446","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Artificial Life and Robotics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1