首页 > 最新文献

2022 International Conference on Advanced Robotics and Intelligent Systems (ARIS)最新文献

英文 中文
Evidential Sensory Fusion of 2D Feature and 3D Shape Information for 3D Occluded Object Recognition in Robotics Applications 基于二维特征和三维形状信息的感官融合的三维遮挡物体识别在机器人中的应用
Pub Date : 2022-08-24 DOI: 10.1109/ARIS56205.2022.9910450
R. Luo, Chi-Tang Chen
An evidential sensory fusion using 2D feature and 3D shape information method is proposed to recognize the occluded object. For the applications of robotic object fetching, the conventional object recognition methods usually applied the algorithms separately from 2D texture matching or 3D shape fitting. It often causes the wrong recognition results when the objects are occluded. The motivation in this study is to enhance the occluded object recognition via the estimate fusion method from the RGB-D sensor, which provides both 2D image and 3D depth information. To associate the 3D shape with the 2D texture, the region of interest (ROI) is firstly captured in 3D coordinate system, and mapped onto the 2D image. The Dempster-Shafer (DS) evidence theory is applied to fuse the confidences from the recognitions of both 2D texture and 3D shape to increase the recognition rate of occluded objects. The experimental results successfully demonstrate that the proposed evidence fusion recognizes the sample object correctly where it usually has the lower confidences from 2D and 3D recognition algorithms alone, when it operates in a separate fashion.
提出了一种基于二维特征和三维形状信息的证据感官融合方法来识别被遮挡物体。对于机器人目标提取的应用,传统的目标识别方法通常将算法与二维纹理匹配或三维形状拟合分开使用。当物体被遮挡时,往往会导致错误的识别结果。本研究的目的是利用RGB-D传感器同时提供二维图像和三维深度信息的估计融合方法增强被遮挡物体的识别能力。为了将三维形状与二维纹理相关联,首先在三维坐标系中捕获感兴趣区域(ROI),并将其映射到二维图像上。应用Dempster-Shafer (DS)证据理论融合二维纹理和三维形状识别的置信度,提高被遮挡物体的识别率。实验结果成功地表明,当证据融合以单独的方式运行时,通常2D和3D识别算法的置信度较低时,所提出的证据融合能够正确地识别样本对象。
{"title":"Evidential Sensory Fusion of 2D Feature and 3D Shape Information for 3D Occluded Object Recognition in Robotics Applications","authors":"R. Luo, Chi-Tang Chen","doi":"10.1109/ARIS56205.2022.9910450","DOIUrl":"https://doi.org/10.1109/ARIS56205.2022.9910450","url":null,"abstract":"An evidential sensory fusion using 2D feature and 3D shape information method is proposed to recognize the occluded object. For the applications of robotic object fetching, the conventional object recognition methods usually applied the algorithms separately from 2D texture matching or 3D shape fitting. It often causes the wrong recognition results when the objects are occluded. The motivation in this study is to enhance the occluded object recognition via the estimate fusion method from the RGB-D sensor, which provides both 2D image and 3D depth information. To associate the 3D shape with the 2D texture, the region of interest (ROI) is firstly captured in 3D coordinate system, and mapped onto the 2D image. The Dempster-Shafer (DS) evidence theory is applied to fuse the confidences from the recognitions of both 2D texture and 3D shape to increase the recognition rate of occluded objects. The experimental results successfully demonstrate that the proposed evidence fusion recognizes the sample object correctly where it usually has the lower confidences from 2D and 3D recognition algorithms alone, when it operates in a separate fashion.","PeriodicalId":254572,"journal":{"name":"2022 International Conference on Advanced Robotics and Intelligent Systems (ARIS)","volume":"64 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126753963","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Rock Climbing Benchmark for Humanoid Robots 人形机器人攀岩基准
Pub Date : 2022-08-24 DOI: 10.1109/ARIS56205.2022.9910449
J. Baltes, Saeed Saeedvand
In this paper, we present the humanoid robot rock climbing competition as a benchmark problem for complex motion planning under kinematic and dynamic constraints. We also describe an advanced algorithm for motion planning in this domain. The algorithm finds stable configurations where three limbs are anchored to the wall and the fourth limb is moving. We suggest possible search techniques to find a sequence through these control funnels to find a path to the top of the climbing wall.
本文将仿人机器人攀岩比赛作为运动学和动力学约束下复杂运动规划的基准问题。我们还描述了一种用于该领域运动规划的高级算法。该算法找到了稳定的结构,其中三个肢体固定在墙上,第四个肢体移动。我们建议可能的搜索技术,以找到一个序列,通过这些控制漏斗,找到一个路径,攀登墙的顶部。
{"title":"Rock Climbing Benchmark for Humanoid Robots","authors":"J. Baltes, Saeed Saeedvand","doi":"10.1109/ARIS56205.2022.9910449","DOIUrl":"https://doi.org/10.1109/ARIS56205.2022.9910449","url":null,"abstract":"In this paper, we present the humanoid robot rock climbing competition as a benchmark problem for complex motion planning under kinematic and dynamic constraints. We also describe an advanced algorithm for motion planning in this domain. The algorithm finds stable configurations where three limbs are anchored to the wall and the fourth limb is moving. We suggest possible search techniques to find a sequence through these control funnels to find a path to the top of the climbing wall.","PeriodicalId":254572,"journal":{"name":"2022 International Conference on Advanced Robotics and Intelligent Systems (ARIS)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116482036","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Computational Approach for Cam Design Parameters Optimization of Disk Cam Mechanisms with Oscillating Roller Follower 摆动滚子从动件盘形凸轮机构凸轮设计参数优化的计算方法
Pub Date : 2022-08-24 DOI: 10.1109/ARIS56205.2022.9910454
Tung-Hsin Pan, Ching-Hsiang Chang, P. Lin, Kuan-Lun Hsu
This research develops a computational approach for cam design parameters optimization of disk cam mechanisms with oscillating roller follower. The synthesis procedure of the cam geometry presented in this paper is computed by MATLAB function fmincon. An implementation of such a cam mechanism is presented to demonstrate the effectiveness of this procedure. By using this procedure, it is convenient to synthesis optimal dimensions for cam mechanism with oscillating roller follower under any preferred constraints. Finally, the importance and the irreplaceability of this research is discussed in the final section.
本研究提出了一种具有摆动滚子从动件的盘形凸轮机构凸轮设计参数优化的计算方法。本文给出的凸轮几何图形的合成过程是通过MATLAB函数fmincon进行计算的。给出了一个凸轮机构的实现,以证明该方法的有效性。利用该方法,可以在任意约束条件下,方便地综合具有摆动滚子从动件的凸轮机构的最优尺寸。最后,在最后一节讨论了本研究的重要性和不可替代性。
{"title":"A Computational Approach for Cam Design Parameters Optimization of Disk Cam Mechanisms with Oscillating Roller Follower","authors":"Tung-Hsin Pan, Ching-Hsiang Chang, P. Lin, Kuan-Lun Hsu","doi":"10.1109/ARIS56205.2022.9910454","DOIUrl":"https://doi.org/10.1109/ARIS56205.2022.9910454","url":null,"abstract":"This research develops a computational approach for cam design parameters optimization of disk cam mechanisms with oscillating roller follower. The synthesis procedure of the cam geometry presented in this paper is computed by MATLAB function fmincon. An implementation of such a cam mechanism is presented to demonstrate the effectiveness of this procedure. By using this procedure, it is convenient to synthesis optimal dimensions for cam mechanism with oscillating roller follower under any preferred constraints. Finally, the importance and the irreplaceability of this research is discussed in the final section.","PeriodicalId":254572,"journal":{"name":"2022 International Conference on Advanced Robotics and Intelligent Systems (ARIS)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134009669","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spatially-Excited Attention Learning for Fine-Grained Visual Categorization 细粒度视觉分类的空间激发注意学习
Pub Date : 2022-08-24 DOI: 10.1109/ARIS56205.2022.9910447
Zhaozhi Luo, Min-Hsiang Hung, Yi-Wen Lu, Kuan-Wen Chen
Learning distinguishable feature embedding plays an important role in fine-grained visual categorization. The existing methods focus on either designing a complex attention mechanism to boost the overall classification performance or proposing a specific training strategy to enhance the learning of the backbone network to achieve a low-cost backbone-only inference. Unlike all of them, an alternative approach called Spatially-Excited Attention Learning (SEAL) is proposed in this paper. The training of SEAL is similar to that of most of the existing methods, but it provides two alternative streams during a network inference: one stream requires higher effort but provides higher performance; the other is a low-cost backbone-only inference with lower but still comparative performance. Note that both the streams are trained at the same time by SEAL. The experiments show that SEAL achieves the state-of-the-art performance under both complex architecture and backbone-only inference conditions.
学习可区分特征嵌入在细粒度视觉分类中起着重要的作用。现有的方法主要是通过设计复杂的注意力机制来提高整体分类性能,或者通过提出特定的训练策略来增强骨干网的学习能力,从而实现低成本的纯骨干网推理。与所有这些方法不同,本文提出了一种称为空间激发注意学习(SEAL)的替代方法。SEAL的训练与大多数现有方法类似,但它在网络推理期间提供了两种可选流:一种流需要更高的努力,但提供更高的性能;另一种是低成本的仅骨干推理,其性能较低,但仍然具有可比性。请注意,海豹突击队同时训练这两种流。实验表明,无论在复杂的体系结构还是在纯骨干推理条件下,SEAL都能达到最先进的性能。
{"title":"Spatially-Excited Attention Learning for Fine-Grained Visual Categorization","authors":"Zhaozhi Luo, Min-Hsiang Hung, Yi-Wen Lu, Kuan-Wen Chen","doi":"10.1109/ARIS56205.2022.9910447","DOIUrl":"https://doi.org/10.1109/ARIS56205.2022.9910447","url":null,"abstract":"Learning distinguishable feature embedding plays an important role in fine-grained visual categorization. The existing methods focus on either designing a complex attention mechanism to boost the overall classification performance or proposing a specific training strategy to enhance the learning of the backbone network to achieve a low-cost backbone-only inference. Unlike all of them, an alternative approach called Spatially-Excited Attention Learning (SEAL) is proposed in this paper. The training of SEAL is similar to that of most of the existing methods, but it provides two alternative streams during a network inference: one stream requires higher effort but provides higher performance; the other is a low-cost backbone-only inference with lower but still comparative performance. Note that both the streams are trained at the same time by SEAL. The experiments show that SEAL achieves the state-of-the-art performance under both complex architecture and backbone-only inference conditions.","PeriodicalId":254572,"journal":{"name":"2022 International Conference on Advanced Robotics and Intelligent Systems (ARIS)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130666949","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Open Action Recognition by A 3D Convolutional Neural Network Combining with An Open Fuzzy Min-Max Neural Network 结合开放模糊最小-最大神经网络的三维卷积神经网络开放式动作识别
Pub Date : 2022-08-24 DOI: 10.1109/ARIS56205.2022.9910444
Chia-Ying Wu, Y. Tsay, A. C. Shih
The 3-dimensional convolution neural network (3D CNN) has demonstrated a high prediction power for action recognition, when the inputs belong to the known classes. In a real application, however, if considering the inputs from unknown classes, previous studies have revealed that some prediction results can have high softmax scores falsely for known classes. That is called the open set recognition problem. Recently, a series of statistical methods based on an openmax approach have been proposed to solve the problem in 2D image data. However, how to apply the approach to video data is still unknown. Without using a prior statistical model, we propose a two-stage approach for open action recognition in this paper. A 3D CNN model is trained in the first stage. Then, the activation vector data, the output from the activation layer, are extracted as the feature data for training a fuzzy min-max neural network (FMMNN) as a classifier in the second stage. Since the value ranges of an activation vector are not limited between 0 and 1, an open FMMNN with a new fuzzy membership function without the normalization of input data is proposed and then constructed by the feature data. Finally, the prediction output is selected by the class with the maximum membership value. In the results, two separated datasets of mouse action videos were used for the training and the prediction test, respectively. We found that the proposed method can indeed improve the prediction performance. Moreover, using the human action and random background videos as two unknown datasets, we also demonstrated that the prediction outputs from known and unknown sets can be distinguished by a single threshold. In short, the proposed open FNNMM can not only improve the prediction performance from the inputs from known classes but also detect the inputs from unknown classes.
当输入属于已知类别时,三维卷积神经网络(3D CNN)对动作识别具有很高的预测能力。然而,在实际应用中,如果考虑未知类的输入,先前的研究表明,对于已知类,一些预测结果可能会错误地具有较高的softmax分数。这被称为开集识别问题。最近,人们提出了一系列基于openmax方法的统计方法来解决二维图像数据中的这一问题。然而,如何将这种方法应用到视频数据中仍然是一个未知的问题。在不使用先验统计模型的情况下,我们提出了一种两阶段的开放式动作识别方法。第一阶段训练三维CNN模型。然后,提取激活层输出的激活向量数据作为特征数据,用于训练模糊最小-最大神经网络(FMMNN)作为第二阶段的分类器。由于激活向量的取值范围不受0 ~ 1的限制,提出了一种不需要对输入数据进行归一化处理的开放式模糊隶属度函数FMMNN。最后,由隶属度值最大的类选择预测输出。在结果中,分别使用两个独立的鼠标动作视频数据集进行训练和预测测试。我们发现,该方法确实可以提高预测性能。此外,使用人类动作和随机背景视频作为两个未知数据集,我们还证明了已知集和未知集的预测输出可以通过单一阈值进行区分。简而言之,所提出的开放式FNNMM不仅可以提高已知类输入的预测性能,还可以检测未知类输入。
{"title":"Open Action Recognition by A 3D Convolutional Neural Network Combining with An Open Fuzzy Min-Max Neural Network","authors":"Chia-Ying Wu, Y. Tsay, A. C. Shih","doi":"10.1109/ARIS56205.2022.9910444","DOIUrl":"https://doi.org/10.1109/ARIS56205.2022.9910444","url":null,"abstract":"The 3-dimensional convolution neural network (3D CNN) has demonstrated a high prediction power for action recognition, when the inputs belong to the known classes. In a real application, however, if considering the inputs from unknown classes, previous studies have revealed that some prediction results can have high softmax scores falsely for known classes. That is called the open set recognition problem. Recently, a series of statistical methods based on an openmax approach have been proposed to solve the problem in 2D image data. However, how to apply the approach to video data is still unknown. Without using a prior statistical model, we propose a two-stage approach for open action recognition in this paper. A 3D CNN model is trained in the first stage. Then, the activation vector data, the output from the activation layer, are extracted as the feature data for training a fuzzy min-max neural network (FMMNN) as a classifier in the second stage. Since the value ranges of an activation vector are not limited between 0 and 1, an open FMMNN with a new fuzzy membership function without the normalization of input data is proposed and then constructed by the feature data. Finally, the prediction output is selected by the class with the maximum membership value. In the results, two separated datasets of mouse action videos were used for the training and the prediction test, respectively. We found that the proposed method can indeed improve the prediction performance. Moreover, using the human action and random background videos as two unknown datasets, we also demonstrated that the prediction outputs from known and unknown sets can be distinguished by a single threshold. In short, the proposed open FNNMM can not only improve the prediction performance from the inputs from known classes but also detect the inputs from unknown classes.","PeriodicalId":254572,"journal":{"name":"2022 International Conference on Advanced Robotics and Intelligent Systems (ARIS)","volume":"11 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132693033","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Swabbing Robot for Covid-19 Specimen Collection 新型冠状病毒标本采集拭子机器人
Pub Date : 2022-08-24 DOI: 10.1109/ARIS56205.2022.9910446
Cheng-Yen Chung, Yun-Chi Hsieh, Yi-Hau Lai, P. Yen
The Covid-19 pandemic has caused large scale of people in danger of infection and death during early outbreak period. Precise screening of the new coronal virus through PCR (Polymerase Chain Reaction) testing on the nasal or oral sample becomes very critical for epidemic control. This study proposes the idea of using a robotic remote manipulation platform for oral and nasal specimen collection operated by medical staffs. The oral cavity image was captured by a compact camera and then displayed on the human machine interface for the medical staffs to confirm the target region for sample collection. The wiping action of the robot was accomplished with a force control with force sensing the contact force between the cotton swab and soft tissue. A prototype of the swabbing robot has been implemented to verify the feasibility and safety of the remote robot-assisted specimen collection.
2019冠状病毒病大流行在疫情早期造成了大规模的感染和死亡危险。通过对鼻腔或口腔样本进行PCR(聚合酶链反应)检测,精确筛选新型冠状病毒,对疫情控制至关重要。本研究提出了使用机器人远程操作平台进行口腔和鼻腔标本采集的想法,由医务人员操作。口腔图像由紧凑型摄像机采集,显示在人机界面上,供医护人员确认采集目标区域。机器人的擦拭动作是通过力控制来完成的,力传感棉签与软组织之间的接触力。为了验证远程机器人辅助标本采集的可行性和安全性,设计了一种取样机器人样机。
{"title":"A Swabbing Robot for Covid-19 Specimen Collection","authors":"Cheng-Yen Chung, Yun-Chi Hsieh, Yi-Hau Lai, P. Yen","doi":"10.1109/ARIS56205.2022.9910446","DOIUrl":"https://doi.org/10.1109/ARIS56205.2022.9910446","url":null,"abstract":"The Covid-19 pandemic has caused large scale of people in danger of infection and death during early outbreak period. Precise screening of the new coronal virus through PCR (Polymerase Chain Reaction) testing on the nasal or oral sample becomes very critical for epidemic control. This study proposes the idea of using a robotic remote manipulation platform for oral and nasal specimen collection operated by medical staffs. The oral cavity image was captured by a compact camera and then displayed on the human machine interface for the medical staffs to confirm the target region for sample collection. The wiping action of the robot was accomplished with a force control with force sensing the contact force between the cotton swab and soft tissue. A prototype of the swabbing robot has been implemented to verify the feasibility and safety of the remote robot-assisted specimen collection.","PeriodicalId":254572,"journal":{"name":"2022 International Conference on Advanced Robotics and Intelligent Systems (ARIS)","volume":"71 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129312987","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Calibration of a Robot's Tool Center Point Using a Laser Displacement Sensor 用激光位移传感器标定机器人刀具中心点
Pub Date : 2022-08-24 DOI: 10.1109/ARIS56205.2022.9910448
Chih-Jer Lin, Hsing-Cheng Wang
The conventional tool center point calibration (TCP) method requires the operator to set the actual position of the tool center point by eye. To address this lengthy workflow and low accuracy while improving accuracy and efficiency for time-saving and non-contact calibration, this paper proposes an enhanced automatic TCP calibration method based on a laser displacement sensor and implemented on a cooperative robot with six degrees of freedom. During the calibration process, the robot arm will move a given distance along the X and Y axes and collect the information when the tool passes through the laser during the process to calculate the deflection of the tool, and then continue to move a given distance along the X and Y axes for the second height calibration. After the deflection angle is calculated and calibrated by triangulation, the deflection calibration is completed and the third X and Y axis displacement is performed to find out the exact position of the tool on the X and Y axes. Finally, the tool is moved to a position higher than the laser, and the laser is triggered by moving downward to obtain information to complete the whole experimental process and get the calibrated tool center position. The whole calibration method is firstly verified in the virtual simulation environment and then implemented on the actual cooperative robot. The results of the proposed TCP calibration method can achieve a positioning accuracy of about 0.07 mm, a positioning accuracy of about 0.18 degrees, a positioning repeatability of $boldsymbol{pm 0.083}$ mm, and a positioning repeatability of less than $boldsymbol{pm 0.17}$ degrees. This result meets the requirements of TCP calibration, but also achieves the purpose of simple, economical and time-saving, and it takes only 60 seconds to complete the whole calibration process.
传统的刀具中心点标定(TCP)方法需要操作者用眼睛设定刀具中心点的实际位置。为解决工作流程长、精度低的问题,提高非接触式标定的精度和效率,本文提出了一种基于激光位移传感器的TCP自动标定方法,并在六自由度协作机器人上实现。在标定过程中,机器人手臂沿X轴和Y轴移动给定距离,并在此过程中收集刀具通过激光时的信息,计算刀具的挠度,然后继续沿X轴和Y轴移动给定距离进行第二次高度标定。在通过三角测量计算和校准挠度角后,完成挠度校准,并进行第三次X轴和Y轴位移,以确定刀具在X轴和Y轴上的准确位置。最后将刀具移动到比激光高的位置,通过向下移动触发激光获取信息,完成整个实验过程,得到校准后的刀具中心位置。整个标定方法首先在虚拟仿真环境中进行验证,然后在实际协作机器人上实现。所提出的TCP校准方法的定位精度约为0.07 mm,定位精度约为0.18°,定位重复性为$boldsymbol{pm 0.083}$ mm,定位重复性小于$boldsymbol{pm 0.17}$°。该结果既满足了TCP校准的要求,又达到了简单、经济、省时的目的,整个校准过程只需60秒即可完成。
{"title":"Calibration of a Robot's Tool Center Point Using a Laser Displacement Sensor","authors":"Chih-Jer Lin, Hsing-Cheng Wang","doi":"10.1109/ARIS56205.2022.9910448","DOIUrl":"https://doi.org/10.1109/ARIS56205.2022.9910448","url":null,"abstract":"The conventional tool center point calibration (TCP) method requires the operator to set the actual position of the tool center point by eye. To address this lengthy workflow and low accuracy while improving accuracy and efficiency for time-saving and non-contact calibration, this paper proposes an enhanced automatic TCP calibration method based on a laser displacement sensor and implemented on a cooperative robot with six degrees of freedom. During the calibration process, the robot arm will move a given distance along the X and Y axes and collect the information when the tool passes through the laser during the process to calculate the deflection of the tool, and then continue to move a given distance along the X and Y axes for the second height calibration. After the deflection angle is calculated and calibrated by triangulation, the deflection calibration is completed and the third X and Y axis displacement is performed to find out the exact position of the tool on the X and Y axes. Finally, the tool is moved to a position higher than the laser, and the laser is triggered by moving downward to obtain information to complete the whole experimental process and get the calibrated tool center position. The whole calibration method is firstly verified in the virtual simulation environment and then implemented on the actual cooperative robot. The results of the proposed TCP calibration method can achieve a positioning accuracy of about 0.07 mm, a positioning accuracy of about 0.18 degrees, a positioning repeatability of $boldsymbol{pm 0.083}$ mm, and a positioning repeatability of less than $boldsymbol{pm 0.17}$ degrees. This result meets the requirements of TCP calibration, but also achieves the purpose of simple, economical and time-saving, and it takes only 60 seconds to complete the whole calibration process.","PeriodicalId":254572,"journal":{"name":"2022 International Conference on Advanced Robotics and Intelligent Systems (ARIS)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115003815","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AI Enhanced Visual Inspection of Post-Polished Workpieces Using You Only Look Once Vision System for Intelligent Robotics Applications 人工智能增强后抛光工件的视觉检测,使用智能机器人应用的“只看一次”视觉系统
Pub Date : 2022-08-24 DOI: 10.1109/ARIS56205.2022.9910441
R. Luo, Zheng-Lun Yu
The objective of this paper is to provide a solution of automated optical inspection for post-polished workpieces using you only look once YOLOv5 vision system. It intends to assist human labor who checks workpieces with eyes in the long terms. Robots have now become an essential role in industrial applications. An example application is automating polishing tasks using robots. However, polishing still requires people to be involved in the post-processing especially in product detection. YOLOv5 can be applied in multiple scenarios because it is known for high accuracy and fast detection in real-time image detection. In this paper YOLOv5 approach have been implemented for robotic faucet surface inspection and demonstrated the success of the process.
本文的目的是提供一种使用YOLOv5视觉系统对抛光后工件进行自动光学检测的解决方案。它旨在协助长期用眼睛检查工件的人类劳动。机器人现在已经成为工业应用中的重要角色。一个示例应用程序是使用机器人自动化抛光任务。然而,抛光仍然需要人参与后处理,特别是在产品检测中。YOLOv5以其在实时图像检测中的高精度和快速检测而闻名,可以应用于多种场景。本文将YOLOv5方法应用于机器人水龙头表面检测,并演示了该过程的成功。
{"title":"AI Enhanced Visual Inspection of Post-Polished Workpieces Using You Only Look Once Vision System for Intelligent Robotics Applications","authors":"R. Luo, Zheng-Lun Yu","doi":"10.1109/ARIS56205.2022.9910441","DOIUrl":"https://doi.org/10.1109/ARIS56205.2022.9910441","url":null,"abstract":"The objective of this paper is to provide a solution of automated optical inspection for post-polished workpieces using you only look once YOLOv5 vision system. It intends to assist human labor who checks workpieces with eyes in the long terms. Robots have now become an essential role in industrial applications. An example application is automating polishing tasks using robots. However, polishing still requires people to be involved in the post-processing especially in product detection. YOLOv5 can be applied in multiple scenarios because it is known for high accuracy and fast detection in real-time image detection. In this paper YOLOv5 approach have been implemented for robotic faucet surface inspection and demonstrated the success of the process.","PeriodicalId":254572,"journal":{"name":"2022 International Conference on Advanced Robotics and Intelligent Systems (ARIS)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132588253","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Deep Comparison Network for Visual Prognosis of a Linear Slide 线性滑梯视觉预后的深度比较网络
Pub Date : 2022-08-24 DOI: 10.1109/ARIS56205.2022.9910443
Chia-Jui Yang, Bo Wen, Chih-Hung G. Li
Linear Slides are important components widely adopted in the manufacturing sector, particularly automated production lines. Damage to the linear slide can cause abnormal machine vibration and result in production line failure. Common failure modes of the linear slide include ball slider wear and severe rail surface contamination or abrasion. Monitoring the condition of the linear slide is of great value and importance in the Industry 4.0 era. There is an emerging need for developing a prognosis method for the linear slide to prevent the unexpected breakdown of the production line. The most common online inspection method utilizes an accelerometer to monitor the system's vibration. However, as abnormal vibration is the result and not the cause of the slide damage, it does not serve well as a prognostic signal. This article proposed an innovative prognosis method by recruiting low-resolution cameras to monitor the rail surface condition. We conducted endurance tests on several linear slides and determined the end of life by the vibration measurements and the pre-compression values. We then annotated the rail surface images with the service percentages and formed the training set for a deep convolutional neural network (CNN). We design the CNN architecture as a dual-input comparison network that compares the initial image and the current image to predict the service percentage of the linear slide. The method appeared promising, judging by the preliminary test results; however, the prediction accuracy needs further improvements before actual application. The comparison network presented the advantage of generalization to various illumination conditions. The cost of low-resolution cameras is also much lower than accelerometers.
直线滑轨是制造业,特别是自动化生产线广泛采用的重要部件。直线滑块的损坏会引起机器异常振动,导致生产线故障。线性滑块的常见失效模式包括滚珠滑块磨损和严重的导轨表面污染或磨损。在工业4.0时代,监测直线滑轨的状态具有重要的价值和重要性。为了防止生产线的意外故障,迫切需要开发一种预测直线滑动的方法。最常见的在线检测方法是利用加速度计来监测系统的振动。然而,由于异常振动是滑动损伤的结果而不是原因,因此它不能很好地作为预测信号。本文提出了一种新颖的利用低分辨率摄像机监测轨道表面状况的预测方法。我们对几个线性滑块进行了耐久性测试,并通过振动测量和预压缩值确定了寿命的终止。然后我们用服务百分比标注轨道表面图像,并形成深度卷积神经网络(CNN)的训练集。我们将CNN架构设计为双输入比较网络,通过比较初始图像和当前图像来预测线性滑动的服务百分比。从初步试验结果来看,该方法是有希望的;但在实际应用前,预测精度还有待进一步提高。该比较网络对各种光照条件具有通用性。低分辨率相机的成本也比加速度计低得多。
{"title":"A Deep Comparison Network for Visual Prognosis of a Linear Slide","authors":"Chia-Jui Yang, Bo Wen, Chih-Hung G. Li","doi":"10.1109/ARIS56205.2022.9910443","DOIUrl":"https://doi.org/10.1109/ARIS56205.2022.9910443","url":null,"abstract":"Linear Slides are important components widely adopted in the manufacturing sector, particularly automated production lines. Damage to the linear slide can cause abnormal machine vibration and result in production line failure. Common failure modes of the linear slide include ball slider wear and severe rail surface contamination or abrasion. Monitoring the condition of the linear slide is of great value and importance in the Industry 4.0 era. There is an emerging need for developing a prognosis method for the linear slide to prevent the unexpected breakdown of the production line. The most common online inspection method utilizes an accelerometer to monitor the system's vibration. However, as abnormal vibration is the result and not the cause of the slide damage, it does not serve well as a prognostic signal. This article proposed an innovative prognosis method by recruiting low-resolution cameras to monitor the rail surface condition. We conducted endurance tests on several linear slides and determined the end of life by the vibration measurements and the pre-compression values. We then annotated the rail surface images with the service percentages and formed the training set for a deep convolutional neural network (CNN). We design the CNN architecture as a dual-input comparison network that compares the initial image and the current image to predict the service percentage of the linear slide. The method appeared promising, judging by the preliminary test results; however, the prediction accuracy needs further improvements before actual application. The comparison network presented the advantage of generalization to various illumination conditions. The cost of low-resolution cameras is also much lower than accelerometers.","PeriodicalId":254572,"journal":{"name":"2022 International Conference on Advanced Robotics and Intelligent Systems (ARIS)","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132961712","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Deep Learning Approach to Predict Dissolved Oxygen in Aquaculture 水产养殖中溶解氧预测的深度学习方法
Pub Date : 2022-08-24 DOI: 10.1109/ARIS56205.2022.9910453
Simon Peter Khabusi, Yonggui Huang
Fish is one of the major sources of protein nutrients for people. Most fish supply comes from the natural habitants which include rivers, lakes, seas and oceans. However, the high demand has necessitated fish farming from man-made lakes, ponds and swamps. There are various issues that pose risks to fish survival and growth, and among these include the level of dissolved oxygen (DO) in the water which is an essential environmental condition whose scarcity leads to suffocation of fish and ultimately death. This study aimed at designing a prediction model for DO in aquatic environments. To achieve the objective, time series data consisting of 70374 records and 15 attributes from Mumford Cove in Connecticut, USA collected for over 5 years was preprocessed and used to train long-short term memory (LSTM) recurrent neural network (RNN) for DO prediction. The training and testing data were obtained by splitting the dataset into 70% and 30%, respectively. Regression models include linear regression (LR), support vector regression (SVR) and decision tree regression (DTR) were also created for comparisons. The performance of the models was evaluated on the basis of mean absolute percentage error (MAPE), mean squared error (MSE), mean absolute error (MAE) and coefficient of determination ($mathbf{R^{2}}$ score). LSTM achieved superior performance compared to the regression models. Conclusively, DO on such multivariate time series data can be well achieved with LSTM RNN.
鱼类是人类蛋白质营养素的主要来源之一。大多数鱼类供应来自自然栖息地,包括河流、湖泊、海洋和海洋。然而,高需求使得人工湖泊、池塘和沼泽的养鱼成为必要。有各种各样的问题对鱼类的生存和生长构成威胁,其中包括水中溶解氧(DO)的水平,这是一种必不可少的环境条件,其缺乏导致鱼类窒息并最终死亡。本研究旨在设计水生环境中溶解氧的预测模型。为了实现这一目标,我们对美国康涅狄格州芒福德湾(Mumford Cove) 5年多来收集的70374条记录和15个属性的时间序列数据进行预处理,并用于训练长短期记忆(LSTM)递归神经网络(RNN)进行DO预测。将数据集分成70%和30%分别得到训练和测试数据。回归模型包括线性回归(LR)、支持向量回归(SVR)和决策树回归(DTR)。根据平均绝对百分比误差(MAPE)、均方误差(MSE)、平均绝对误差(MAE)和决定系数($mathbf{R^{2}}$ score)对模型的性能进行评价。与回归模型相比,LSTM取得了更好的性能。综上所述,LSTM RNN可以很好地实现多变量时间序列数据的DO。
{"title":"A Deep Learning Approach to Predict Dissolved Oxygen in Aquaculture","authors":"Simon Peter Khabusi, Yonggui Huang","doi":"10.1109/ARIS56205.2022.9910453","DOIUrl":"https://doi.org/10.1109/ARIS56205.2022.9910453","url":null,"abstract":"Fish is one of the major sources of protein nutrients for people. Most fish supply comes from the natural habitants which include rivers, lakes, seas and oceans. However, the high demand has necessitated fish farming from man-made lakes, ponds and swamps. There are various issues that pose risks to fish survival and growth, and among these include the level of dissolved oxygen (DO) in the water which is an essential environmental condition whose scarcity leads to suffocation of fish and ultimately death. This study aimed at designing a prediction model for DO in aquatic environments. To achieve the objective, time series data consisting of 70374 records and 15 attributes from Mumford Cove in Connecticut, USA collected for over 5 years was preprocessed and used to train long-short term memory (LSTM) recurrent neural network (RNN) for DO prediction. The training and testing data were obtained by splitting the dataset into 70% and 30%, respectively. Regression models include linear regression (LR), support vector regression (SVR) and decision tree regression (DTR) were also created for comparisons. The performance of the models was evaluated on the basis of mean absolute percentage error (MAPE), mean squared error (MSE), mean absolute error (MAE) and coefficient of determination ($mathbf{R^{2}}$ score). LSTM achieved superior performance compared to the regression models. Conclusively, DO on such multivariate time series data can be well achieved with LSTM RNN.","PeriodicalId":254572,"journal":{"name":"2022 International Conference on Advanced Robotics and Intelligent Systems (ARIS)","volume":"62 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128015891","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
2022 International Conference on Advanced Robotics and Intelligent Systems (ARIS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1