For the successful interaction between human and machine agents, the agents need understand both explicitly-presented human intention and unpresented human mind. Although the current human-agent interaction (HAI) systems mainly rely on the former with keystrokes, speech, and gestures, the latter will play an important role for the new and up-coming HAIs. In this talk we will present our continuing efforts to understand unpresented human mind, which may reside at the internal states of neural networks in human brain and may be estimated from brain-related signals such as fMRI (functional Magnetic Resonance Imaging), EEG (Electroencephalography), and eye movements. We hypothesized that the space of brain internal states have several independent axes, of which temporal dynamics have different time scales. Special emphasis was given to human memory, trustworthiness, and sympathy to others during interactions. Human memory changes much slowly in time, and is different from person to person. Therefore, by analyzing brain-related signals from many stimulating images, it may be possible to identify a person. On the other hand the sympathy to others has much shorter time constants during human-agent interactions, and may be identified for each user interaction. The trustworthiness to others may have slightly longer time constants, and may be accumulated by temporal integration during sequential interactions. Therefore, we measured brain-related signals during sequential Theory-of-Mind (ToM) games. Also, the effects of human-like cues of the agents to the trustworthiness were evaluated. At this moment the estimation of human internal states utilizes brain-related signals such as fMRI, EEG, and eye movements. In the future the classification systems of human internal states will be trained with audio-visual signals only, and the current study will provide near-ground-truth labels.
{"title":"Understanding Human Internal States: I Know Who You Are and What You Think","authors":"Soo-Young Lee","doi":"10.1145/2814940.2825025","DOIUrl":"https://doi.org/10.1145/2814940.2825025","url":null,"abstract":"For the successful interaction between human and machine agents, the agents need understand both explicitly-presented human intention and unpresented human mind. Although the current human-agent interaction (HAI) systems mainly rely on the former with keystrokes, speech, and gestures, the latter will play an important role for the new and up-coming HAIs. In this talk we will present our continuing efforts to understand unpresented human mind, which may reside at the internal states of neural networks in human brain and may be estimated from brain-related signals such as fMRI (functional Magnetic Resonance Imaging), EEG (Electroencephalography), and eye movements. We hypothesized that the space of brain internal states have several independent axes, of which temporal dynamics have different time scales. Special emphasis was given to human memory, trustworthiness, and sympathy to others during interactions. Human memory changes much slowly in time, and is different from person to person. Therefore, by analyzing brain-related signals from many stimulating images, it may be possible to identify a person. On the other hand the sympathy to others has much shorter time constants during human-agent interactions, and may be identified for each user interaction. The trustworthiness to others may have slightly longer time constants, and may be accumulated by temporal integration during sequential interactions. Therefore, we measured brain-related signals during sequential Theory-of-Mind (ToM) games. Also, the effects of human-like cues of the agents to the trustworthiness were evaluated. At this moment the estimation of human internal states utilizes brain-related signals such as fMRI, EEG, and eye movements. In the future the classification systems of human internal states will be trained with audio-visual signals only, and the current study will provide near-ground-truth labels.","PeriodicalId":427567,"journal":{"name":"Proceedings of the 3rd International Conference on Human-Agent Interaction","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131351802","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Tools used for remote collaboration and assistance within the industrial sector have remained unchanged for a long time. We introduce a concept called "Expert on Wheels" as the next step in development of remote collaboration tools. "Expert on Wheels" is a mobile tele-presence robot designed to support collaboration between a field worker and a remote expert. This paper presents an exploratory research study, in which "Expert on Wheels" was evaluated with target users as well as in a lab environment. The results indicate that the system has potential but requires improvements in key areas such as the situation awareness of the expert and the system's mobility as a whole. We conclude by discussing if and how such systems could be accepted and useful in different industrial settings.
{"title":"Expert on Wheels: An Approach to Remote Collaboration","authors":"E. Vartiainen, Veronika Domova, M. Englund","doi":"10.1145/2814940.2814943","DOIUrl":"https://doi.org/10.1145/2814940.2814943","url":null,"abstract":"Tools used for remote collaboration and assistance within the industrial sector have remained unchanged for a long time. We introduce a concept called \"Expert on Wheels\" as the next step in development of remote collaboration tools. \"Expert on Wheels\" is a mobile tele-presence robot designed to support collaboration between a field worker and a remote expert. This paper presents an exploratory research study, in which \"Expert on Wheels\" was evaluated with target users as well as in a lab environment. The results indicate that the system has potential but requires improvements in key areas such as the situation awareness of the expert and the system's mobility as a whole. We conclude by discussing if and how such systems could be accepted and useful in different industrial settings.","PeriodicalId":427567,"journal":{"name":"Proceedings of the 3rd International Conference on Human-Agent Interaction","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122866085","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sangwook Kim, Zhibin Yu, Jonghong Kim, A. Ojha, Minho Lee
Recognition of human intention is an important issue in human-robot interaction research and allows a robot to respond adequately according to human's wish. In this paper, we discuss how robots can infer human intention by learning affordance, a concept used to represent the relation between an agent and its environment. Learning of the robot, to understand human and its interaction with environment, is achieved within the framework of action-perception cycle. The action-perception cycle explains how an intelligent agent learns and enhances its ability continuously by interacting with its surrounding. The proposed intention recognition and recommendation system includes several key functions such as joint attention, object recognition, affordance model, motion understanding module and so on. The experimental results show high successful recognition performance and the plausibility of the proposed system.
{"title":"Human-Robot Interaction using Intention Recognition","authors":"Sangwook Kim, Zhibin Yu, Jonghong Kim, A. Ojha, Minho Lee","doi":"10.1145/2814940.2815002","DOIUrl":"https://doi.org/10.1145/2814940.2815002","url":null,"abstract":"Recognition of human intention is an important issue in human-robot interaction research and allows a robot to respond adequately according to human's wish. In this paper, we discuss how robots can infer human intention by learning affordance, a concept used to represent the relation between an agent and its environment. Learning of the robot, to understand human and its interaction with environment, is achieved within the framework of action-perception cycle. The action-perception cycle explains how an intelligent agent learns and enhances its ability continuously by interacting with its surrounding. The proposed intention recognition and recommendation system includes several key functions such as joint attention, object recognition, affordance model, motion understanding module and so on. The experimental results show high successful recognition performance and the plausibility of the proposed system.","PeriodicalId":427567,"journal":{"name":"Proceedings of the 3rd International Conference on Human-Agent Interaction","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115429090","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jooyeon Lee, Seong-eun Moon, Manri Cheon, Jong-Seok Lee
In this paper we explore the relations between perspectives of navigation and electroencephalogram (EEG) in 3D virtual space. We analyze three types of navigation with EEG recordings and examine how the perspectives affect the users' electrical activities in their brains. Via a small-scale experiment, we find that the influence of peripersonal space is altered by the perspective, and it can be observed via EEG monitoring. These results have interesting implications on virtual reality applications where a sense of agency, or a peripersonal task takes important roles.
{"title":"EEG Analysis on 3D Navigation in Virtual Realty with Different Perspectives","authors":"Jooyeon Lee, Seong-eun Moon, Manri Cheon, Jong-Seok Lee","doi":"10.1145/2814940.2814982","DOIUrl":"https://doi.org/10.1145/2814940.2814982","url":null,"abstract":"In this paper we explore the relations between perspectives of navigation and electroencephalogram (EEG) in 3D virtual space. We analyze three types of navigation with EEG recordings and examine how the perspectives affect the users' electrical activities in their brains. Via a small-scale experiment, we find that the influence of peripersonal space is altered by the perspective, and it can be observed via EEG monitoring. These results have interesting implications on virtual reality applications where a sense of agency, or a peripersonal task takes important roles.","PeriodicalId":427567,"journal":{"name":"Proceedings of the 3rd International Conference on Human-Agent Interaction","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115114854","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Recently, there has been tremendous development in education contents for foreign language learning. Based on these trends, IT has provided educational contents development using e-learning and broadcast media. But conventional educational contents are non-interactive presents an impediment to provide user's specific service. To develop a user friendly language education tool, we propose an intelligent learning tool based on user's eye movement and brain waves. By analyzing these features, the proposed system detects if the given word is known or unknown to the user while learning a foreign language. Then it searches its meaning and provides a vocabulary list of unknown words to users in real time. The proposed model provides a tool which enables self-directed learning. We assume that the proposed system can improve users' learning achievements and satisfaction.
{"title":"Development of Intelligent Learning Tool for Improving Foreign Language Skills Based on EEG and Eye tracker","authors":"Jun-Su Kang, A. Ojha, Minho Lee","doi":"10.1145/2814940.2814951","DOIUrl":"https://doi.org/10.1145/2814940.2814951","url":null,"abstract":"Recently, there has been tremendous development in education contents for foreign language learning. Based on these trends, IT has provided educational contents development using e-learning and broadcast media. But conventional educational contents are non-interactive presents an impediment to provide user's specific service. To develop a user friendly language education tool, we propose an intelligent learning tool based on user's eye movement and brain waves. By analyzing these features, the proposed system detects if the given word is known or unknown to the user while learning a foreign language. Then it searches its meaning and provides a vocabulary list of unknown words to users in real time. The proposed model provides a tool which enables self-directed learning. We assume that the proposed system can improve users' learning achievements and satisfaction.","PeriodicalId":427567,"journal":{"name":"Proceedings of the 3rd International Conference on Human-Agent Interaction","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126395262","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ryuji Nakamura, Kouki Miyazawa, H. Ishihara, Ken'ya Nishikawa, H. Kikuchi, M. Asada, R. Mazuka
The characteristics of the spoken language used to address infants have been eagerly studied as a part of the language acquisition research. Because of the uncontrollability factor with regard to the infants, the features and roles of infant- directed speech were tried to be revealed by the comparison of speech directed toward infants and that toward other listeners. However, they share few characteristics with infants, while infants have many characteristics which may derive the features of IDS. In this study, to solve this problem, we will introduce a new approach that replaces the infant with an infant-like robot which is designed to control its motions and to imitate its appearance very similar to a real infant. We have now recorded both infant- and infant- like robot-directed speech and are constructing both corpora. Analysis of these corpora is expected to contribute to the studies of infant-directed speech. In this paper, we discuss the contents of this approach and the outline of the corpora.
{"title":"Constructing the Corpus of Infant-Directed Speech and Infant-Like Robot-Directed Speech","authors":"Ryuji Nakamura, Kouki Miyazawa, H. Ishihara, Ken'ya Nishikawa, H. Kikuchi, M. Asada, R. Mazuka","doi":"10.1145/2814940.2814965","DOIUrl":"https://doi.org/10.1145/2814940.2814965","url":null,"abstract":"The characteristics of the spoken language used to address infants have been eagerly studied as a part of the language acquisition research. Because of the uncontrollability factor with regard to the infants, the features and roles of infant- directed speech were tried to be revealed by the comparison of speech directed toward infants and that toward other listeners. However, they share few characteristics with infants, while infants have many characteristics which may derive the features of IDS. In this study, to solve this problem, we will introduce a new approach that replaces the infant with an infant-like robot which is designed to control its motions and to imitate its appearance very similar to a real infant. We have now recorded both infant- and infant- like robot-directed speech and are constructing both corpora. Analysis of these corpora is expected to contribute to the studies of infant-directed speech. In this paper, we discuss the contents of this approach and the outline of the corpora.","PeriodicalId":427567,"journal":{"name":"Proceedings of the 3rd International Conference on Human-Agent Interaction","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132578922","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Learning from demonstration (LfD) provides an easy and intuitive way to program robot behaviours, potentially reducing development time and costs tremendously. This is especially appealing for manufacturers interested in using industrial manipulators for high-mix production, since this technique enables fast and flexible modifications to the robot behaviours and is thus suitable to teach the robot to perform a wide range of tasks regularly. We define a set of criteria to assess the applicability of state-of-the-art LfD frameworks in the industry. A three-stage LfD method is then proposed, which incorporates human-in-the-loop adaptation to iteratively correct a batch-learned policy to improve accuracy and precision. The system will then transit to open-loop execution of the task to enhance production speed, by removing the human teacher from the feedback loop. The proposed LfD framework addresses all criteria set in this work.
{"title":"Towards Industrial Robot Learning from Demonstration","authors":"W. Ko, Yan Wu, K. Tee, J. Buchli","doi":"10.1145/2814940.2814984","DOIUrl":"https://doi.org/10.1145/2814940.2814984","url":null,"abstract":"Learning from demonstration (LfD) provides an easy and intuitive way to program robot behaviours, potentially reducing development time and costs tremendously. This is especially appealing for manufacturers interested in using industrial manipulators for high-mix production, since this technique enables fast and flexible modifications to the robot behaviours and is thus suitable to teach the robot to perform a wide range of tasks regularly. We define a set of criteria to assess the applicability of state-of-the-art LfD frameworks in the industry. A three-stage LfD method is then proposed, which incorporates human-in-the-loop adaptation to iteratively correct a batch-learned policy to improve accuracy and precision. The system will then transit to open-loop execution of the task to enhance production speed, by removing the human teacher from the feedback loop. The proposed LfD framework addresses all criteria set in this work.","PeriodicalId":427567,"journal":{"name":"Proceedings of the 3rd International Conference on Human-Agent Interaction","volume":"216 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130346648","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We present our ongoing work on a creativity assistance tool called I-get. The tool is based on the hypothesis that perceptual similarity between a pair of images, at a subconscious level, plays a key role in generating creative conceptual associations and metaphorical interpretations. The tool "I-get" is designed to assist users to create novel ideas and metaphorical associations primed by algorithmic perceptual similarity between two images and alternative conceptual associations given by users.
{"title":"I-get: A Creativity Assistance Tool to Generate Perceptual Pictorial Metaphors","authors":"A. Ojha, H. Lee, Minho Lee","doi":"10.1145/2814940.2815006","DOIUrl":"https://doi.org/10.1145/2814940.2815006","url":null,"abstract":"We present our ongoing work on a creativity assistance tool called I-get. The tool is based on the hypothesis that perceptual similarity between a pair of images, at a subconscious level, plays a key role in generating creative conceptual associations and metaphorical interpretations. The tool \"I-get\" is designed to assist users to create novel ideas and metaphorical associations primed by algorithmic perceptual similarity between two images and alternative conceptual associations given by users.","PeriodicalId":427567,"journal":{"name":"Proceedings of the 3rd International Conference on Human-Agent Interaction","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130362764","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Air purifiers and humidifiers, the popular home appliances, have a large working space by injecting and controlling air, which spreads over a house. The air circulation inside a room limits the efficiency of air purification, the disparities of which vary with environmental factors, such as room shape, temperature, furniture arrangements, and human movement. In this study, mobility provided by mobile robot technology is combined with a conventional air purification function. This robotic appliance can improve air purification efficiency, which depends on the air injections and filtration. In our experiments, the effectiveness of the mobility is verified and the commercial growth potential, as a new type of smart home service robot is discussed.
{"title":"Smart Air Purification and Humidification by a Mobile Robot toward a Smart Home","authors":"Jeong-Yean Yang, D. Kwon","doi":"10.1145/2814940.2815003","DOIUrl":"https://doi.org/10.1145/2814940.2815003","url":null,"abstract":"Air purifiers and humidifiers, the popular home appliances, have a large working space by injecting and controlling air, which spreads over a house. The air circulation inside a room limits the efficiency of air purification, the disparities of which vary with environmental factors, such as room shape, temperature, furniture arrangements, and human movement. In this study, mobility provided by mobile robot technology is combined with a conventional air purification function. This robotic appliance can improve air purification efficiency, which depends on the air injections and filtration. In our experiments, the effectiveness of the mobility is verified and the commercial growth potential, as a new type of smart home service robot is discussed.","PeriodicalId":427567,"journal":{"name":"Proceedings of the 3rd International Conference on Human-Agent Interaction","volume":"99 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130679089","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yongsik Jin, Jonghong Kim, Bumhwi Kim, R. Mallipeddi, Minho Lee
We propose a smart cane with a face recognition system to help the blind in recognizing human faces. This system detects and recognizes faces around them. The result of the detection is informed to the blind person through a vibration pattern. The proposed system was designed to be used in real-time and is equipped with a camera mounted on the glasses, a vibration motor attached to the cane and a mobile computer. The camera attached to the glasses sends image to mobile computer. The mobile computer extracts features from the image and then detects the face using Adaboost. We use the modified census transform (MCT) descriptor for feature extraction. After face detection, the information regarding the detected face image is gathered. We used compressed sensing with L2-norm as a classifier. Cane is equipped with a Bluetooth module and receives a person's information from the mobile computer. The cane generates vibration patterns unique to each person as to inform a blind person about the identity of the detected person using the camera. Hence, the blind people can know the person standing in front of them.
{"title":"Smart Cane: Face Recognition System for Blind","authors":"Yongsik Jin, Jonghong Kim, Bumhwi Kim, R. Mallipeddi, Minho Lee","doi":"10.1145/2814940.2814952","DOIUrl":"https://doi.org/10.1145/2814940.2814952","url":null,"abstract":"We propose a smart cane with a face recognition system to help the blind in recognizing human faces. This system detects and recognizes faces around them. The result of the detection is informed to the blind person through a vibration pattern. The proposed system was designed to be used in real-time and is equipped with a camera mounted on the glasses, a vibration motor attached to the cane and a mobile computer. The camera attached to the glasses sends image to mobile computer. The mobile computer extracts features from the image and then detects the face using Adaboost. We use the modified census transform (MCT) descriptor for feature extraction. After face detection, the information regarding the detected face image is gathered. We used compressed sensing with L2-norm as a classifier. Cane is equipped with a Bluetooth module and receives a person's information from the mobile computer. The cane generates vibration patterns unique to each person as to inform a blind person about the identity of the detected person using the camera. Hence, the blind people can know the person standing in front of them.","PeriodicalId":427567,"journal":{"name":"Proceedings of the 3rd International Conference on Human-Agent Interaction","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125134659","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}