首页 > 最新文献

2019 International Conference on Cyberworlds (CW)最新文献

英文 中文
A Kinect-Based Augmented Reality Game for Lower Limb Exercise 一款基于肢体运动的增强现实游戏
Pub Date : 2019-10-01 DOI: 10.1109/CW.2019.00077
Y. Tokuyama, R. J. Rajapakse, Sachi Yamabe, K. Konno, Y. Hung
Augmented reality (AR) is where 3D virtual objects are integrated into a 3D real environment in real time. The augmented reality applications such as medical visualization, maintenance and repair, robot path planning, entertainment, military aircraft navigation, and targeting applications have been proposed. This paper introduces the development of an augmented reality game which allows the user to carry out lower limb exercise using a natural user interface based on Microsoft Kinect. The system has been designed as an augmented game where users can see themselves in a world augmented with virtual objects generated by computer graphics. The player sitting in a chair just has to step on a mole that appears and disappears by moving upward and downward randomly. It encourages the activities of a large number of lower limb muscles which will help prevent falls. It is also suitable for rehabilitation.
增强现实(AR)是将3D虚拟对象实时集成到3D真实环境中的技术。增强现实在医疗可视化、维护与维修、机器人路径规划、娱乐、军用飞机导航和目标定位等方面的应用已经被提出。本文介绍了一种基于微软Kinect的增强现实游戏的开发,该游戏允许用户使用自然用户界面进行下肢运动。该系统被设计成一个增强游戏,用户可以在一个由计算机图形生成的虚拟物体增强的世界中看到自己。坐在椅子上的玩家只需要踩上一颗鼹鼠,它会随机上下移动,时而出现时而消失。它鼓励大量下肢肌肉的活动,这将有助于防止跌倒。它也适用于康复。
{"title":"A Kinect-Based Augmented Reality Game for Lower Limb Exercise","authors":"Y. Tokuyama, R. J. Rajapakse, Sachi Yamabe, K. Konno, Y. Hung","doi":"10.1109/CW.2019.00077","DOIUrl":"https://doi.org/10.1109/CW.2019.00077","url":null,"abstract":"Augmented reality (AR) is where 3D virtual objects are integrated into a 3D real environment in real time. The augmented reality applications such as medical visualization, maintenance and repair, robot path planning, entertainment, military aircraft navigation, and targeting applications have been proposed. This paper introduces the development of an augmented reality game which allows the user to carry out lower limb exercise using a natural user interface based on Microsoft Kinect. The system has been designed as an augmented game where users can see themselves in a world augmented with virtual objects generated by computer graphics. The player sitting in a chair just has to step on a mole that appears and disappears by moving upward and downward randomly. It encourages the activities of a large number of lower limb muscles which will help prevent falls. It is also suitable for rehabilitation.","PeriodicalId":117409,"journal":{"name":"2019 International Conference on Cyberworlds (CW)","volume":"6 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125020448","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
EEG-Based Human Factors Evaluation of Air Traffic Control Operators (ATCOs) for Optimal Training 基于脑电图的空管人员最佳培训人因评价
Pub Date : 2019-10-01 DOI: 10.1109/CW.2019.00049
Yisi Liu, Zirui Lan, F. Trapsilawati, O. Sourina, Chun-Hsien Chen, W. Müller-Wittig
To deal with the increasing demands in Air Traffic Control (ATC), new working place designs are proposed and developed that need novel human factors evaluation tools. In this paper, we propose a novel application of Electroencephalogram (EEG)-based emotion, workload, and stress recognition algorithms to investigate the optimal length of training for Air Traffic Control Officers (ATCOs) to learn working with three-dimensional (3D) display as a supplementary to the existing 2D display. We tested and applied the state-of-the-art EEG-based subject-dependent algorithms. The following experiment was carried out. Twelve ATCOs were recruited to take part in the experiment. The participants were in charge of the Terminal Control Area, providing navigation assistance to aircraft departing and approaching the airport using 2D and 3D displays. EEG data were recorded, and traditional human factors questionnaires were given to the participants after 15-minute, 60-minute, and 120-minute training. Different from the questionnaires, the EEG-based evaluation tools allow the recognition of emotions, workload, and stress with different temporal resolutions during the task performance by subjects. The results showed that 50-minute training could be enough for the ATCOs to learn the new display setting as they had relatively low stress and workload. The study demonstrated that there is a potential of applying the EEG-based human factors evaluation tools to assess novel system designs in addition to traditional questionnaire and feedback, which can be beneficial for future improvements and developments of the systems and interfaces.
为了应对空中交通管制(ATC)日益增长的需求,新的工作场所设计被提出和发展,需要新的人为因素评估工具。在本文中,我们提出了一种基于脑电图(EEG)的情绪、工作量和压力识别算法的新应用,以研究空中交通管制人员(atco)学习使用三维(3D)显示器作为现有二维显示器的补充的最佳训练长度。我们测试并应用了最先进的基于脑电图的主题相关算法。进行了以下实验。12名空中交通管制员被招募来参加实验。参加者负责终端管制区,使用2D和3D显示器为离境和接近机场的飞机提供导航协助。在训练15分钟、60分钟和120分钟后,记录脑电数据,并对参与者进行传统的人为因素问卷调查。与问卷调查不同的是,基于脑电图的评估工具允许被试在任务执行过程中对不同时间分辨率的情绪、工作量和压力进行识别。结果表明,50分钟的训练足以让atco学习新的显示设置,因为他们的压力和工作量相对较小。研究表明,除了传统的问卷调查和反馈之外,基于脑电图的人因评价工具有可能用于评价新系统设计,这对未来系统和界面的改进和发展有益。
{"title":"EEG-Based Human Factors Evaluation of Air Traffic Control Operators (ATCOs) for Optimal Training","authors":"Yisi Liu, Zirui Lan, F. Trapsilawati, O. Sourina, Chun-Hsien Chen, W. Müller-Wittig","doi":"10.1109/CW.2019.00049","DOIUrl":"https://doi.org/10.1109/CW.2019.00049","url":null,"abstract":"To deal with the increasing demands in Air Traffic Control (ATC), new working place designs are proposed and developed that need novel human factors evaluation tools. In this paper, we propose a novel application of Electroencephalogram (EEG)-based emotion, workload, and stress recognition algorithms to investigate the optimal length of training for Air Traffic Control Officers (ATCOs) to learn working with three-dimensional (3D) display as a supplementary to the existing 2D display. We tested and applied the state-of-the-art EEG-based subject-dependent algorithms. The following experiment was carried out. Twelve ATCOs were recruited to take part in the experiment. The participants were in charge of the Terminal Control Area, providing navigation assistance to aircraft departing and approaching the airport using 2D and 3D displays. EEG data were recorded, and traditional human factors questionnaires were given to the participants after 15-minute, 60-minute, and 120-minute training. Different from the questionnaires, the EEG-based evaluation tools allow the recognition of emotions, workload, and stress with different temporal resolutions during the task performance by subjects. The results showed that 50-minute training could be enough for the ATCOs to learn the new display setting as they had relatively low stress and workload. The study demonstrated that there is a potential of applying the EEG-based human factors evaluation tools to assess novel system designs in addition to traditional questionnaire and feedback, which can be beneficial for future improvements and developments of the systems and interfaces.","PeriodicalId":117409,"journal":{"name":"2019 International Conference on Cyberworlds (CW)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114026108","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Composite Sketch Recognition Using Multi-scale Hog Features and Semantic Attributes 基于多尺度Hog特征和语义属性的组合素描识别
Pub Date : 2019-10-01 DOI: 10.1109/CW.2019.00028
Xinying Xue, Jiayi Xu, Xiaoyang Mao
Composite sketch recognition belongs to heterogeneous face recognition research, which is of great important in the field of criminal investigation. Because composite face sketch and photo belong to different modalities, robust representation of face feature cross different modalities is the key to recognition. Considering that composite sketch lacks texture details in some area, using texture features only may result in low recognition accuracy, this paper proposes a composite sketch recognition algorithm based on multi-scale Hog features and semantic attributes. Firstly, the global Hog features of the face and the local Hog features of each face component are extracted to represent the contour and detail features. Then the global and detail features are fused according to their importance at score level. Finally, semantic attributes are employed to reorder the matching results. The proposed algorithm is validated on PRIP-VSGC database and UoM-SGFS database, and achieves rank 10 identification accuracy of 88.6% and 96.7% respectively, which demonstrates that the proposed method outperforms other state-of-the-art methods.
合成素描识别属于异构人脸识别研究,在刑侦领域具有重要意义。由于合成人脸草图和照片属于不同的模态,因此跨不同模态的人脸特征鲁棒表示是人脸识别的关键。针对复合草图在某些区域缺乏纹理细节,仅使用纹理特征可能导致识别精度较低的问题,提出了一种基于多尺度Hog特征和语义属性的复合草图识别算法。首先,提取人脸的全局Hog特征和各分量的局部Hog特征来表示人脸的轮廓和细节特征;然后根据全局特征和细节特征在评分水平上的重要程度进行融合。最后,利用语义属性对匹配结果进行重新排序。本文算法在rip - vsgc数据库和UoM-SGFS数据库上进行了验证,10级识别准确率分别达到88.6%和96.7%,优于其他先进方法。
{"title":"Composite Sketch Recognition Using Multi-scale Hog Features and Semantic Attributes","authors":"Xinying Xue, Jiayi Xu, Xiaoyang Mao","doi":"10.1109/CW.2019.00028","DOIUrl":"https://doi.org/10.1109/CW.2019.00028","url":null,"abstract":"Composite sketch recognition belongs to heterogeneous face recognition research, which is of great important in the field of criminal investigation. Because composite face sketch and photo belong to different modalities, robust representation of face feature cross different modalities is the key to recognition. Considering that composite sketch lacks texture details in some area, using texture features only may result in low recognition accuracy, this paper proposes a composite sketch recognition algorithm based on multi-scale Hog features and semantic attributes. Firstly, the global Hog features of the face and the local Hog features of each face component are extracted to represent the contour and detail features. Then the global and detail features are fused according to their importance at score level. Finally, semantic attributes are employed to reorder the matching results. The proposed algorithm is validated on PRIP-VSGC database and UoM-SGFS database, and achieves rank 10 identification accuracy of 88.6% and 96.7% respectively, which demonstrates that the proposed method outperforms other state-of-the-art methods.","PeriodicalId":117409,"journal":{"name":"2019 International Conference on Cyberworlds (CW)","volume":"71 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115564316","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Bird Species Classification with Audio-Visual Data using CNN and Multiple Kernel Learning 基于CNN和多核学习的视听数据鸟类分类
Pub Date : 2019-10-01 DOI: 10.1109/CW.2019.00022
B. Naranchimeg, Chao Zhang, T. Akashi
Recently, deep convolutional neural networks (CNN) have become a new standard in many machine learning applications not only in image but also in audio processing. However, most of the studies only explore a single type of training data. In this paper, we present a study on classifying bird species by combining deep neural features of both visual and audio data using kernel-based fusion method. Specifically, we extract deep neural features based on the activation values of an inner layer of CNN. We combine these features by multiple kernel learning (MKL) to perform the final classification. In the experiment, we train and evaluate our method on a CUB-200-2011 standard data set combined with our originally collected audio data set with respect to 200 bird species (classes). The experimental results indicate that our CNN+MKL method which utilizes the combination of both categories of data outperforms single-modality methods, some simple kernel combination methods, and the conventional early fusion method.
近年来,深度卷积神经网络(CNN)已经成为许多机器学习应用的新标准,不仅在图像处理领域,在音频处理领域也是如此。然而,大多数研究只探索单一类型的训练数据。本文采用基于核融合的方法,结合视觉和听觉数据的深度神经特征进行了鸟类物种分类研究。具体来说,我们基于CNN内层的激活值提取深度神经特征。我们通过多核学习(MKL)将这些特征组合起来进行最终分类。在实验中,我们在一个CUB-200-2011标准数据集上训练和评估了我们的方法,并结合了我们最初收集的200种鸟类(类)的音频数据集。实验结果表明,利用两类数据组合的CNN+MKL方法优于单模态方法、一些简单的核组合方法和传统的早期融合方法。
{"title":"Bird Species Classification with Audio-Visual Data using CNN and Multiple Kernel Learning","authors":"B. Naranchimeg, Chao Zhang, T. Akashi","doi":"10.1109/CW.2019.00022","DOIUrl":"https://doi.org/10.1109/CW.2019.00022","url":null,"abstract":"Recently, deep convolutional neural networks (CNN) have become a new standard in many machine learning applications not only in image but also in audio processing. However, most of the studies only explore a single type of training data. In this paper, we present a study on classifying bird species by combining deep neural features of both visual and audio data using kernel-based fusion method. Specifically, we extract deep neural features based on the activation values of an inner layer of CNN. We combine these features by multiple kernel learning (MKL) to perform the final classification. In the experiment, we train and evaluate our method on a CUB-200-2011 standard data set combined with our originally collected audio data set with respect to 200 bird species (classes). The experimental results indicate that our CNN+MKL method which utilizes the combination of both categories of data outperforms single-modality methods, some simple kernel combination methods, and the conventional early fusion method.","PeriodicalId":117409,"journal":{"name":"2019 International Conference on Cyberworlds (CW)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122665536","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Visual Assessment of Distorted View for Metamorphopsia Patient by Interactive Line Manipulation 交互式线操作对变形视的视觉评价
Pub Date : 2019-10-01 DOI: 10.1109/CW.2019.00038
Hiromichi Ichige, M. Toyoura, K. Go, K. Kashiwagi, I. Fujishiro, Xiaoyang Mao
The number of individuals with Age-related Macular Degeneration (AMD) is rapidly increasing. One of the main symptoms of AMD is "metamorphopsia," or distorted vision, which not only makes it difficult for individuals with AMD to do detailed-oriented tasks but also makes sufferers more vulnerable to certain risks in day-to-day life. Traditional clinical approaches to assess metamorphopsia have lacked mechanisms for quantifying the degree of distortion in space, making it impossible to know exactly how individuals with the condition see things. This paper proposes a new method for quantifying distortion in space and visualizing AMD patients' distorted views via line manipulation. By visualizing the distorted views stemming from metamorphopsia, the method gives doctors and others an intuitive picture of how patients see the world and thereby enables a broad range of options for treatment and support.
与年龄相关的黄斑变性(AMD)的个体数量正在迅速增加。老年性黄斑变性的主要症状之一是“变形视”,即视力扭曲,这不仅使老年性黄斑变性患者难以完成具体的任务,而且使患者在日常生活中更容易受到某些风险的影响。评估变形视的传统临床方法缺乏量化空间扭曲程度的机制,因此不可能确切知道患有这种疾病的人是如何看待事物的。本文提出了一种量化空间扭曲的新方法,并通过行操作将AMD患者的扭曲视图可视化。通过将变形症产生的扭曲观点可视化,该方法使医生和其他人直观地了解患者如何看待世界,从而为治疗和支持提供了广泛的选择。
{"title":"Visual Assessment of Distorted View for Metamorphopsia Patient by Interactive Line Manipulation","authors":"Hiromichi Ichige, M. Toyoura, K. Go, K. Kashiwagi, I. Fujishiro, Xiaoyang Mao","doi":"10.1109/CW.2019.00038","DOIUrl":"https://doi.org/10.1109/CW.2019.00038","url":null,"abstract":"The number of individuals with Age-related Macular Degeneration (AMD) is rapidly increasing. One of the main symptoms of AMD is \"metamorphopsia,\" or distorted vision, which not only makes it difficult for individuals with AMD to do detailed-oriented tasks but also makes sufferers more vulnerable to certain risks in day-to-day life. Traditional clinical approaches to assess metamorphopsia have lacked mechanisms for quantifying the degree of distortion in space, making it impossible to know exactly how individuals with the condition see things. This paper proposes a new method for quantifying distortion in space and visualizing AMD patients' distorted views via line manipulation. By visualizing the distorted views stemming from metamorphopsia, the method gives doctors and others an intuitive picture of how patients see the world and thereby enables a broad range of options for treatment and support.","PeriodicalId":117409,"journal":{"name":"2019 International Conference on Cyberworlds (CW)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129818670","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
CPR Virtual Reality Training Simulator for Schools 学校心肺复苏虚拟现实训练模拟器
Pub Date : 2019-10-01 DOI: 10.1109/CW.2019.00013
N. Vaughan, N. John, N. Rees
This research project developed a Virtual Reality (VR) training simulator for the CPR procedure. This is designed for use training school children. It can also form part of a larger system for training paramedics with VR. The simulator incorporates a number of advanced VR technologies including Oculus Rift and Leap motion. We have gained input from NHS paramedics and several related organisation to design the system and provide feedback and evaluation of the preliminary working prototype.
本研究计划开发一套虚拟实境(VR)心肺复苏术训练模拟器。这是为训练学生而设计的。它也可以成为一个更大的系统的一部分,用VR培训护理人员。该模拟器融合了许多先进的虚拟现实技术,包括Oculus Rift和Leap motion。我们已经从NHS护理人员和几个相关组织获得了输入来设计系统,并提供初步工作原型的反馈和评估。
{"title":"CPR Virtual Reality Training Simulator for Schools","authors":"N. Vaughan, N. John, N. Rees","doi":"10.1109/CW.2019.00013","DOIUrl":"https://doi.org/10.1109/CW.2019.00013","url":null,"abstract":"This research project developed a Virtual Reality (VR) training simulator for the CPR procedure. This is designed for use training school children. It can also form part of a larger system for training paramedics with VR. The simulator incorporates a number of advanced VR technologies including Oculus Rift and Leap motion. We have gained input from NHS paramedics and several related organisation to design the system and provide feedback and evaluation of the preliminary working prototype.","PeriodicalId":117409,"journal":{"name":"2019 International Conference on Cyberworlds (CW)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129848545","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
A Shoulder-Surfing Resistant Image-Based Authentication Scheme with a Brain-Computer Interface 一种基于脑机接口的抗肩冲浪图像认证方案
Pub Date : 2019-10-01 DOI: 10.1109/CW.2019.00061
Florian Gondesen, Matthias Marx, Ann-Christine Kycler
With the increasing availability of consumer brain-computer interfaces, new methods of authentication can be considered. In this paper, we present a shoulder surfing resistant means of entering a graphical password by measuring brain activity. The password is a subset of images displayed repeatedly by rapid serial visual presentation. The occurrence of a password image entails an event-related potential in the electroencephalogram, the P300 response. The P300 response is used to classify whether an image belongs to the password subset or not. We compare individual classifiers, trained with samples of a specific user, to general P300 classifiers, trained over all subjects. We evaluate the permanence of the classification results in three subsequent experiment sessions. The classification score significantly increases from the first to the third session. Comparing the use of natural photos or simple objects as stimuli shows no significant difference. In total, our authentication scheme achieves an equal error rate of about 10%. In the future, with increasing accuracy and proliferation, brain-computer interfaces could find practical application in alternative authentication methods.
随着消费者脑机接口的日益普及,可以考虑新的认证方法。在本文中,我们提出了一种抗肩冲浪的方法,通过测量大脑活动来输入图形密码。密码是通过快速连续视觉呈现反复显示的图像子集。密码图像的出现需要在脑电图中产生事件相关电位,即P300反应。P300响应用于对图像是否属于密码子集进行分类。我们比较单个分类器,与特定用户的样本训练,一般的P300分类器,训练在所有科目。我们在随后的三次实验中评估分类结果的持久性。从第一次到第三次,分类得分显著增加。比较使用自然照片和简单物体作为刺激没有显著差异。总的来说,我们的认证方案实现了大约10%的错误率。在未来,随着准确性的提高和普及,脑机接口可以在替代身份验证方法中找到实际应用。
{"title":"A Shoulder-Surfing Resistant Image-Based Authentication Scheme with a Brain-Computer Interface","authors":"Florian Gondesen, Matthias Marx, Ann-Christine Kycler","doi":"10.1109/CW.2019.00061","DOIUrl":"https://doi.org/10.1109/CW.2019.00061","url":null,"abstract":"With the increasing availability of consumer brain-computer interfaces, new methods of authentication can be considered. In this paper, we present a shoulder surfing resistant means of entering a graphical password by measuring brain activity. The password is a subset of images displayed repeatedly by rapid serial visual presentation. The occurrence of a password image entails an event-related potential in the electroencephalogram, the P300 response. The P300 response is used to classify whether an image belongs to the password subset or not. We compare individual classifiers, trained with samples of a specific user, to general P300 classifiers, trained over all subjects. We evaluate the permanence of the classification results in three subsequent experiment sessions. The classification score significantly increases from the first to the third session. Comparing the use of natural photos or simple objects as stimuli shows no significant difference. In total, our authentication scheme achieves an equal error rate of about 10%. In the future, with increasing accuracy and proliferation, brain-computer interfaces could find practical application in alternative authentication methods.","PeriodicalId":117409,"journal":{"name":"2019 International Conference on Cyberworlds (CW)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129456260","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Detection Defect in Printed Circuit Boards using Unsupervised Feature Extraction Upon Transfer Learning 基于迁移学习的无监督特征提取检测印刷电路板缺陷
Pub Date : 2019-10-01 DOI: 10.1109/CW.2019.00025
I. Volkau, A. Mujeeb, Wenting Dai, Marius Erdt, A. Sourin
Automatic optical inspection for manufacturing traditionally was based on computer vision. However, there are emerging attempts to do it using deep learning approach. Deep convolutional neural network allows to learn semantic image features which could be used for defect detection in products. In contrast to the existing approaches where supervised or semi-supervised training is done on thousands of images of defects, we investigate whether unsupervised deep learning model for defect detection could be trained with orders of magnitude smaller amount of representative defect-free samples (tenths rather than thousands). This research is motivated by the fact that collection of large amounts of defective samples is difficult and expensive. Our model undergoes only one-class training and aims to extract distinctive semantic features from the normal samples in an unsupervised manner. We propose a variant of transfer learning, that consists of combination of unsupervised learning used upon VGG16 with pre-trained on ImageNet weight coefficients. To demonstrate a defect detection, we used a set of Printed Circuit Boards (PCBs) with different types of defects - scratch, missing washer/extra hole, abrasion, broken PCB edge. The trained model allows us to make clusters of normal internal representations of features of PCB in high-dimensional feature space, and to localize defective patches in PCB image based on distance from normal clusters. Initial results show that more than 90% of defects were detected.
传统的制造业自动光学检测是基于计算机视觉的。然而,有一些新兴的尝试使用深度学习方法来做到这一点。深度卷积神经网络可以学习语义图像特征,用于产品缺陷检测。与现有的对数千个缺陷图像进行监督或半监督训练的方法相反,我们研究了用于缺陷检测的无监督深度学习模型是否可以用数量级较小的代表性无缺陷样本(十分之一而不是数千)进行训练。这项研究的动机是收集大量有缺陷的样品既困难又昂贵。我们的模型只经过一个类的训练,旨在以无监督的方式从正常样本中提取出独特的语义特征。我们提出了一种迁移学习的变体,它结合了在VGG16上使用的无监督学习和在ImageNet上预训练的权系数。为了演示缺陷检测,我们使用了一组具有不同类型缺陷的印刷电路板(PCB) -划伤,缺少垫圈/额外孔,磨损,PCB边缘破损。训练后的模型允许我们在高维特征空间中对PCB特征的正常内部表示进行聚类,并基于与正常聚类的距离来定位PCB图像中的缺陷补丁。初步结果表明,90%以上的缺陷被检测出来。
{"title":"Detection Defect in Printed Circuit Boards using Unsupervised Feature Extraction Upon Transfer Learning","authors":"I. Volkau, A. Mujeeb, Wenting Dai, Marius Erdt, A. Sourin","doi":"10.1109/CW.2019.00025","DOIUrl":"https://doi.org/10.1109/CW.2019.00025","url":null,"abstract":"Automatic optical inspection for manufacturing traditionally was based on computer vision. However, there are emerging attempts to do it using deep learning approach. Deep convolutional neural network allows to learn semantic image features which could be used for defect detection in products. In contrast to the existing approaches where supervised or semi-supervised training is done on thousands of images of defects, we investigate whether unsupervised deep learning model for defect detection could be trained with orders of magnitude smaller amount of representative defect-free samples (tenths rather than thousands). This research is motivated by the fact that collection of large amounts of defective samples is difficult and expensive. Our model undergoes only one-class training and aims to extract distinctive semantic features from the normal samples in an unsupervised manner. We propose a variant of transfer learning, that consists of combination of unsupervised learning used upon VGG16 with pre-trained on ImageNet weight coefficients. To demonstrate a defect detection, we used a set of Printed Circuit Boards (PCBs) with different types of defects - scratch, missing washer/extra hole, abrasion, broken PCB edge. The trained model allows us to make clusters of normal internal representations of features of PCB in high-dimensional feature space, and to localize defective patches in PCB image based on distance from normal clusters. Initial results show that more than 90% of defects were detected.","PeriodicalId":117409,"journal":{"name":"2019 International Conference on Cyberworlds (CW)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130068150","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
ParaVR: Paramedic Virtual Reality Training Simulator ParaVR:护理人员虚拟现实训练模拟器
Pub Date : 2019-10-01 DOI: 10.1109/CW.2019.00012
N. Vaughan, N. John, N. Rees
This research project developed a Virtual Reality (VR) training simulator for paramedic procedures. Currently needle cricothyroidotomy and chest drain are modelled, which could form part of a larger system for training paramedics with VR in various other procedures. The simulator incorporates a number of advanced VR technologies including Oculus Rift and haptic feedback. We have gained input and feedback from NHS paramedics and several related organisation to design the system and provide feedback and evaluation of the preliminary working prototype.
本研究项目为护理人员程序开发了虚拟现实(VR)训练模拟器。目前,环状甲状腺针管切开术和胸腔引流术都是模拟的,这可能成为一个更大的系统的一部分,用于培训护理人员在各种其他程序中使用VR。该模拟器融合了许多先进的VR技术,包括Oculus Rift和触觉反馈。我们已经从NHS护理人员和几个相关组织获得了输入和反馈,以设计系统,并对初步工作原型提供反馈和评估。
{"title":"ParaVR: Paramedic Virtual Reality Training Simulator","authors":"N. Vaughan, N. John, N. Rees","doi":"10.1109/CW.2019.00012","DOIUrl":"https://doi.org/10.1109/CW.2019.00012","url":null,"abstract":"This research project developed a Virtual Reality (VR) training simulator for paramedic procedures. Currently needle cricothyroidotomy and chest drain are modelled, which could form part of a larger system for training paramedics with VR in various other procedures. The simulator incorporates a number of advanced VR technologies including Oculus Rift and haptic feedback. We have gained input and feedback from NHS paramedics and several related organisation to design the system and provide feedback and evaluation of the preliminary working prototype.","PeriodicalId":117409,"journal":{"name":"2019 International Conference on Cyberworlds (CW)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127666245","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Semi-Automatic Creation of an Anime-Like 3D Face Model from a Single Illustration 从单个插图半自动创建一个类似动画的3D面部模型
Pub Date : 2019-10-01 DOI: 10.1109/CW.2019.00017
T. Niki, T. Komuro
In this paper, we propose a method for semi-automatically creating an anime-like 3D face model from a single illustration. In the proposed method, principal component analysis (PCA) is applied to existing anime-like 3D models to obtain base models for generating natural 3D models. To align the dimensions of the data and make geometric correspondence, a template model is deformed using a modified Nonrigid Iterative Closest Point (NICP) method. Then, the coefficients of the linear combination of the base models are estimated by minimizing the difference between the rendered image of the 3D model with the coefficients and the input illustration using edge-based matching. We confirmed that our method was able to generate a natural anime-like 3D face models which has similar eye and face shapes to those of the input illustration.
在本文中,我们提出了一种从单个插图半自动创建类似动画的3D人脸模型的方法。该方法将主成分分析(PCA)应用于已有的类动画三维模型,获得生成自然三维模型的基础模型。采用改进的非刚性迭代最近点(NICP)方法对模板模型进行变形,使数据的尺寸对齐并使其几何对应。然后,利用基于边缘的匹配,通过最小化带有系数的三维模型渲染图像与输入插图之间的差异来估计基本模型线性组合的系数。我们证实,我们的方法能够生成一个自然的类似动画的3D面部模型,其眼睛和面部形状与输入插图相似。
{"title":"Semi-Automatic Creation of an Anime-Like 3D Face Model from a Single Illustration","authors":"T. Niki, T. Komuro","doi":"10.1109/CW.2019.00017","DOIUrl":"https://doi.org/10.1109/CW.2019.00017","url":null,"abstract":"In this paper, we propose a method for semi-automatically creating an anime-like 3D face model from a single illustration. In the proposed method, principal component analysis (PCA) is applied to existing anime-like 3D models to obtain base models for generating natural 3D models. To align the dimensions of the data and make geometric correspondence, a template model is deformed using a modified Nonrigid Iterative Closest Point (NICP) method. Then, the coefficients of the linear combination of the base models are estimated by minimizing the difference between the rendered image of the 3D model with the coefficients and the input illustration using edge-based matching. We confirmed that our method was able to generate a natural anime-like 3D face models which has similar eye and face shapes to those of the input illustration.","PeriodicalId":117409,"journal":{"name":"2019 International Conference on Cyberworlds (CW)","volume":"81 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122618249","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
2019 International Conference on Cyberworlds (CW)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1