Pub Date : 2024-08-01DOI: 10.1016/j.chbah.2024.100085
Social robots as storytellers combine advantages of human storytellers – such as embodiment, gestures, and gaze – and audio books – large repertoire of voices, sound effects, and background music. However, research on adding non-speech sounds to robotic storytelling is yet in its infancy. The current series of four online studies investigates the influence of sound effects and background music in robotic storytelling on recipients’ storytelling experience and enjoyment, robot perception, and emotion induction across different story genres, i.e. horror, detective, romantic and humorous stories. Results indicate increased enjoyment for romantic stories and a trend for decreased fatigue for all genres when adding sound effects and background music to the robotic storytelling. Of the four genres examined, horror stories seem to benefit the most from the addition of non-speech sounds. Future research should provide guidelines for the selection of music and sound effects to improve the realization of non-speech sound-accompanied robotic storytelling. In conclusion, our ongoing research suggests that the integration of sound effects and background music holds promise for enhancing robotic storytelling, and our genre comparison provides first guidance of when to use them.
{"title":"Integrating sound effects and background music in Robotic storytelling – A series of online studies across different story genres","authors":"","doi":"10.1016/j.chbah.2024.100085","DOIUrl":"10.1016/j.chbah.2024.100085","url":null,"abstract":"<div><p>Social robots as storytellers combine advantages of human storytellers – such as embodiment, gestures, and gaze – and audio books – large repertoire of voices, sound effects, and background music. However, research on adding non-speech sounds to robotic storytelling is yet in its infancy. The current series of four online studies investigates the influence of sound effects and background music in robotic storytelling on recipients’ storytelling experience and enjoyment, robot perception, and emotion induction across different story genres, i.e. horror, detective, romantic and humorous stories. Results indicate increased enjoyment for romantic stories and a trend for decreased fatigue for all genres when adding sound effects and background music to the robotic storytelling. Of the four genres examined, horror stories seem to benefit the most from the addition of non-speech sounds. Future research should provide guidelines for the selection of music and sound effects to improve the realization of non-speech sound-accompanied robotic storytelling. In conclusion, our ongoing research suggests that the integration of sound effects and background music holds promise for enhancing robotic storytelling, and our genre comparison provides first guidance of when to use them.</p></div>","PeriodicalId":100324,"journal":{"name":"Computers in Human Behavior: Artificial Humans","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949882124000458/pdfft?md5=39926971bcbec336bf3117e22eb44704&pid=1-s2.0-S2949882124000458-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141937463","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01DOI: 10.1016/j.chbah.2024.100089
Generative AI applications have increasingly gained visibility in recent educational literature. Yet less is known about how access to generative tools, such as ChatGPT, influences help-seeking during complex problem-solving. In this paper, we aim to advance the understanding of learners' use of a support strategy (hints) when solving data science programming tasks in an online AI-enabled learning environment. The study compared two conditions: students solving problems in DaTu with AI assistance (N = 45) and those without AI assistance (N = 44). Findings reveal no difference in hint-seeking behavior between the two groups, suggesting that the integration of AI assistance has minimal impact on how individuals seek help. The findings also suggest that the availability of AI assistance does not necessarily reduce learners’ reliance on support strategies (such as hints). The current study advances data science education and research by exploring the influence of AI assistance during complex data science problem-solving. We discuss implications and identify paths for future research.
在最近的教育文献中,生成式人工智能应用日益受到关注。然而,人们对使用生成工具(如 ChatGPT)如何影响复杂问题解决过程中的求助行为知之甚少。在本文中,我们旨在进一步了解学习者在人工智能在线学习环境中解决数据科学编程任务时使用辅助策略(提示)的情况。研究比较了两种情况:学生在有人工智能辅助的 DaTu 中解决问题(45 人)和没有人工智能辅助的学生(44 人)。研究结果表明,两组学生在寻求提示行为上没有差异,这表明集成人工智能辅助对个人如何寻求帮助的影响微乎其微。研究结果还表明,提供人工智能辅助并不一定会减少学习者对辅助策略(如提示)的依赖。本研究通过探索人工智能辅助在复杂数据科学问题解决过程中的影响,推动了数据科学教育和研究的发展。我们讨论了研究的意义,并确定了未来研究的方向。
{"title":"Integrating generative AI in data science programming: Group differences in hint requests","authors":"","doi":"10.1016/j.chbah.2024.100089","DOIUrl":"10.1016/j.chbah.2024.100089","url":null,"abstract":"<div><p>Generative AI applications have increasingly gained visibility in recent educational literature. Yet less is known about how access to generative tools, such as ChatGPT, influences help-seeking during complex problem-solving. In this paper, we aim to advance the understanding of learners' use of a support strategy (hints) when solving data science programming tasks in an online AI-enabled learning environment. The study compared two conditions: students solving problems in <em>DaTu</em> with AI assistance (<em>N</em> = 45) and those without AI assistance (<em>N</em> = 44). Findings reveal no difference in hint-seeking behavior between the two groups, suggesting that the integration of AI assistance has minimal impact on how individuals seek help. The findings also suggest that the availability of AI assistance does not necessarily reduce learners’ reliance on support strategies (such as hints). The current study advances data science education and research by exploring the influence of AI assistance during complex data science problem-solving. We discuss implications and identify paths for future research.</p></div>","PeriodicalId":100324,"journal":{"name":"Computers in Human Behavior: Artificial Humans","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949882124000495/pdfft?md5=d2364f734cd75435ea2c327fb376b30e&pid=1-s2.0-S2949882124000495-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142230120","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01DOI: 10.1016/j.chbah.2024.100084
The emphasis on artificial intelligence (AI) is rapidly increasing across many diverse aspects of society. This manuscript discusses some of the key topics related to the expansion of AI. These include a comparison of the unique cognitive capabilities of human intelligence with AI, and the potential risks of using AI in clinical medicine. The general public attitudes towards AI are also discussed, including patient perspectives. As the promotion of AI in high-risk situations such as clinical medicine expands, the limitations, risks and benefits of AI need to be better understood.
{"title":"Differences between human and artificial/augmented intelligence in medicine","authors":"","doi":"10.1016/j.chbah.2024.100084","DOIUrl":"10.1016/j.chbah.2024.100084","url":null,"abstract":"<div><p>The emphasis on artificial intelligence (AI) is rapidly increasing across many diverse aspects of society. This manuscript discusses some of the key topics related to the expansion of AI. These include a comparison of the unique cognitive capabilities of human intelligence with AI, and the potential risks of using AI in clinical medicine. The general public attitudes towards AI are also discussed, including patient perspectives. As the promotion of AI in high-risk situations such as clinical medicine expands, the limitations, risks and benefits of AI need to be better understood.</p></div>","PeriodicalId":100324,"journal":{"name":"Computers in Human Behavior: Artificial Humans","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949882124000446/pdfft?md5=de42c1e5a75fbb492e2bc6a082094c1f&pid=1-s2.0-S2949882124000446-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141853511","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01DOI: 10.1016/j.chbah.2024.100087
Theory suggests that robots with human-like mental capabilities (i.e., high agency and experience) evoke stronger aversion than robots without these capabilities. Yet, while several studies support this prediction, there is also evidence that the mental prowess of robots could be evaluated positively, at least by some individuals. To help resolving this ambivalence, we focused on rather stable individual differences that may shape users’ responses to machines with different levels of (perceived) mental ability. Specifically, we explored four key variables as potential moderators: monotheistic religiosity, the tendency to anthropomorphize, prior attitudes towards robots, and the general affinity for complex technology. Two pre-registered online experiments (N1 = 391, N2 = 617) were conducted, using text vignettes to introduce participants to a robot with or without complex, human-like capabilities. Results showed that negative attitudes towards robots increased the relative aversion against machines with (vs. without) complex minds, whereas technology affinity weakened the difference between conditions. Results for monotheistic religiosity turned out mixed, while the tendency to anthropomorphize had no significant impact on the evoked aversion. Overall, we conclude that certain individual differences play an important role in perceptions of machines with complex minds and should be considered in future research.
{"title":"Aversion against machines with complex mental abilities: The role of individual differences","authors":"","doi":"10.1016/j.chbah.2024.100087","DOIUrl":"10.1016/j.chbah.2024.100087","url":null,"abstract":"<div><p>Theory suggests that robots with human-like mental capabilities (i.e., high agency and experience) evoke stronger aversion than robots without these capabilities. Yet, while several studies support this prediction, there is also evidence that the mental prowess of robots could be evaluated positively, at least by some individuals. To help resolving this ambivalence, we focused on rather stable individual differences that may shape users’ responses to machines with different levels of (perceived) mental ability. Specifically, we explored four key variables as potential moderators: monotheistic religiosity, the tendency to anthropomorphize, prior attitudes towards robots, and the general affinity for complex technology. Two pre-registered online experiments (<em>N</em><sub><em>1</em></sub> = 391, <em>N</em><sub><em>2</em></sub> = 617) were conducted, using text vignettes to introduce participants to a robot with or without complex, human-like capabilities. Results showed that negative attitudes towards robots increased the relative aversion against machines with (vs. without) complex minds, whereas technology affinity weakened the difference between conditions. Results for monotheistic religiosity turned out mixed, while the tendency to anthropomorphize had no significant impact on the evoked aversion. Overall, we conclude that certain individual differences play an important role in perceptions of machines with complex minds and should be considered in future research.</p></div>","PeriodicalId":100324,"journal":{"name":"Computers in Human Behavior: Artificial Humans","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949882124000471/pdfft?md5=d427d8fd14eb2a20aa2d28b06757e636&pid=1-s2.0-S2949882124000471-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141850605","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01DOI: 10.1016/j.chbah.2024.100090
As Chat Generative Pre-trained Transformer (ChatGPT) gains traction, its impact on post-secondary education is increasingly being debated. This qualitative study explores the perception of students and faculty members at a research university in Canada regarding ChatGPT's use in a post-secondary setting, focusing on how it could be incorporated and what ways instructors can respond to this technology. We present the summary of a discussion that took place in a 2-hour focus group session with 40 participants from the computer science and engineering departments, and highlight issues surrounding plagiarism, assessment methods, and the appropriate use of ChatGPT. Findings suggest that students are likely to use ChatGPT, but there is a need for specific guidelines, more classroom assessments, and mandatory reporting of ChatGPT use. The study contributes to the emergent research on ChatGPT in higher education and emphasizes the importance of proactively addressing challenges and opportunities associated with ChatGPT adoption and use. The novelty of the study involves capturing the perspectives of students and faculty members. This paper aims to provide a more refined understanding of the complex interplay between AI chatbots and higher education that will help educators navigate the rapidly evolving landscape of AI-driven education.
{"title":"Unleashing ChatGPT's impact in higher education: Student and faculty perspectives","authors":"","doi":"10.1016/j.chbah.2024.100090","DOIUrl":"10.1016/j.chbah.2024.100090","url":null,"abstract":"<div><div>As Chat Generative Pre-trained Transformer (ChatGPT) gains traction, its impact on post-secondary education is increasingly being debated. This qualitative study explores the perception of students and faculty members at a research university in Canada regarding ChatGPT's use in a post-secondary setting, focusing on how it could be incorporated and what ways instructors can respond to this technology. We present the summary of a discussion that took place in a 2-hour focus group session with 40 participants from the computer science and engineering departments, and highlight issues surrounding plagiarism, assessment methods, and the appropriate use of ChatGPT. Findings suggest that students are likely to use ChatGPT, but there is a need for specific guidelines, more classroom assessments, and mandatory reporting of ChatGPT use. The study contributes to the emergent research on ChatGPT in higher education and emphasizes the importance of proactively addressing challenges and opportunities associated with ChatGPT adoption and use. The novelty of the study involves capturing the perspectives of students and faculty members. This paper aims to provide a more refined understanding of the complex interplay between AI chatbots and higher education that will help educators navigate the rapidly evolving landscape of AI-driven education.</div></div>","PeriodicalId":100324,"journal":{"name":"Computers in Human Behavior: Artificial Humans","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949882124000501/pdfft?md5=ad3828185881ae4a828f051407953830&pid=1-s2.0-S2949882124000501-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142312796","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01DOI: 10.1016/j.chbah.2024.100088
The interplay between artificial intelligence (AI) and psychology, particularly in personality assessment, represents an important emerging area of research. Accurate personality trait estimation is crucial not only for enhancing personalization in human-computer interaction but also for a wide variety of applications ranging from mental health to education. This paper analyzes the capability of a generic chatbot, ChatGPT, to effectively infer personality traits from short texts. We report the results of a comprehensive user study featuring texts written in Czech by a representative population sample of 155 participants. Their self-assessments based on the Big Five Inventory (BFI) questionnaire serve as the ground truth. We compare the personality trait estimations made by ChatGPT against those by human raters and report ChatGPT's competitive performance in inferring personality traits from text. We also uncover a ‘positivity bias’ in ChatGPT's assessments across all personality dimensions and explore the impact of prompt composition on accuracy. This work contributes to the understanding of AI capabilities in psychological assessment, highlighting both the potential and limitations of using large language models for personality inference. Our research underscores the importance of responsible AI development, considering ethical implications such as privacy, consent, autonomy, and bias in AI applications.
{"title":"Can ChatGPT read who you are?","authors":"","doi":"10.1016/j.chbah.2024.100088","DOIUrl":"10.1016/j.chbah.2024.100088","url":null,"abstract":"<div><p>The interplay between artificial intelligence (AI) and psychology, particularly in personality assessment, represents an important emerging area of research. Accurate personality trait estimation is crucial not only for enhancing personalization in human-computer interaction but also for a wide variety of applications ranging from mental health to education. This paper analyzes the capability of a generic chatbot, ChatGPT, to effectively infer personality traits from short texts. We report the results of a comprehensive user study featuring texts written in Czech by a representative population sample of 155 participants. Their self-assessments based on the Big Five Inventory (BFI) questionnaire serve as the ground truth. We compare the personality trait estimations made by ChatGPT against those by human raters and report ChatGPT's competitive performance in inferring personality traits from text. We also uncover a ‘positivity bias’ in ChatGPT's assessments across all personality dimensions and explore the impact of prompt composition on accuracy. This work contributes to the understanding of AI capabilities in psychological assessment, highlighting both the potential and limitations of using large language models for personality inference. Our research underscores the importance of responsible AI development, considering ethical implications such as privacy, consent, autonomy, and bias in AI applications.</p></div>","PeriodicalId":100324,"journal":{"name":"Computers in Human Behavior: Artificial Humans","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949882124000483/pdfft?md5=e63d2e9d2b171f646e851561d4060bf7&pid=1-s2.0-S2949882124000483-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141843635","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01DOI: 10.1016/j.chbah.2024.100086
Mental disorders impact a large proportion of individuals worldwide, with young adults being particularly susceptible to poor mental health. Past research shows that help-seeking self-stigma plays a vital role in deterring help-seeking among young adults; however, this relationship has primarily been examined in the context of human-delivered psychotherapy. The present study aimed to understand how young adults’ perceptions of help-seeking self-stigma associated with different modes of psychotherapy, specifically human-delivered and artificial intelligence (AI)-delivered, influence attitudes towards using AI chatbots for psychotherapy. This study employed a cross-sectional survey design to measure perceived help-seeking self-stigma and attitudes towards both human- and AI-delivered psychotherapy. The results demonstrated that high help-seeking self-stigma associated with human-delivered psychotherapy was linked to more negative attitudes towards human-delivered psychotherapy but more positive attitudes towards AI-delivered psychotherapy. Moreover, high help-seeking self-stigma associated with AI-delivered psychotherapy was linked to more negative attitudes towards AI-delivered psychotherapy but more positive attitudes towards human-delivered psychotherapy. These findings have important real-world implications for future clinical practice and mental health service delivery. The results indicate that young adults who are reluctant to engage with human-delivered psychotherapy due to help-seeking self-stigma may be more inclined to seek help through alternative modes of psychotherapy, such as AI chatbots. Limitations and future directions are discussed.
{"title":"Understanding young adults’ attitudes towards using AI chatbots for psychotherapy: The role of self-stigma","authors":"","doi":"10.1016/j.chbah.2024.100086","DOIUrl":"10.1016/j.chbah.2024.100086","url":null,"abstract":"<div><p>Mental disorders impact a large proportion of individuals worldwide, with young adults being particularly susceptible to poor mental health. Past research shows that help-seeking self-stigma plays a vital role in deterring help-seeking among young adults; however, this relationship has primarily been examined in the context of human-delivered psychotherapy. The present study aimed to understand how young adults’ perceptions of help-seeking self-stigma associated with different modes of psychotherapy, specifically human-delivered and artificial intelligence (AI)-delivered, influence attitudes towards using AI chatbots for psychotherapy. This study employed a cross-sectional survey design to measure perceived help-seeking self-stigma and attitudes towards both human- and AI-delivered psychotherapy. The results demonstrated that high help-seeking self-stigma associated with human-delivered psychotherapy was linked to more negative attitudes towards human-delivered psychotherapy but more positive attitudes towards AI-delivered psychotherapy. Moreover, high help-seeking self-stigma associated with AI-delivered psychotherapy was linked to more negative attitudes towards AI-delivered psychotherapy but more positive attitudes towards human-delivered psychotherapy. These findings have important real-world implications for future clinical practice and mental health service delivery. The results indicate that young adults who are reluctant to engage with human-delivered psychotherapy due to help-seeking self-stigma may be more inclined to seek help through alternative modes of psychotherapy, such as AI chatbots. Limitations and future directions are discussed.</p></div>","PeriodicalId":100324,"journal":{"name":"Computers in Human Behavior: Artificial Humans","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S294988212400046X/pdfft?md5=7105a13b93ecb735c5d2187838096a15&pid=1-s2.0-S294988212400046X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141848422","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01DOI: 10.1016/j.chbah.2024.100091
With the increasing influence of artificial intelligence (AI) on various aspects of society, understanding public attitudes towards AI becomes crucial. This study investigated attitudes towards AI among Hong Kong middle-aged and older adults. In June 2023, an online survey was conducted among a sample of 740 smartphone users aged 45 years or older (Max = 78) in Hong Kong. Using exploratory factor analysis, we found three factors from the General Attitude to Artificial Intelligence Scale (GAAIS) - Perils, Power, and Promises. Subsequently, with latent profile analysis we revealed three latent profiles: (i) Enthusiasts (18.4%; high on Promises and Power but low on Perils); (ii) Skeptics (12.3%; high on Perils but low on Promises and Power), and (iii) Indecisive (69.3%; moderate on all three factors). The Enthusiasts were more likely to be male, with higher socio-economic status, better self-rated health, and greater mobile device proficiency, optimism, innovativeness, but also less insecurity with technology, compared to the Indecisive, and then to the Skeptics. Our findings suggest that most middle-aged and older adults in Hong Kong hold an ambivalent view towards AI, appreciating its power and potentials while also cognizant of the perils it may entail. Our findings are timely considering the recent debates on ethical use of AI evoked by smart phone applications such as ChatGPT and will be valuable for practitioners and scholars for developing inclusive AI-facilitated services and applications.
{"title":"Perils, power and promises: Latent profile analysis on the attitudes towards artificial intelligence (AI) among middle-aged and older adults in Hong Kong","authors":"","doi":"10.1016/j.chbah.2024.100091","DOIUrl":"10.1016/j.chbah.2024.100091","url":null,"abstract":"<div><p>With the increasing influence of artificial intelligence (AI) on various aspects of society, understanding public attitudes towards AI becomes crucial. This study investigated attitudes towards AI among Hong Kong middle-aged and older adults. In June 2023, an online survey was conducted among a sample of 740 smartphone users aged 45 years or older (Max = 78) in Hong Kong. Using exploratory factor analysis, we found three factors from the General Attitude to Artificial Intelligence Scale (GAAIS) - Perils, Power, and Promises. Subsequently, with latent profile analysis we revealed three latent profiles: (i) Enthusiasts (18.4%; high on Promises and Power but low on Perils); (ii) Skeptics (12.3%; high on Perils but low on Promises and Power), and (iii) Indecisive (69.3%; moderate on all three factors). The Enthusiasts were more likely to be male, with higher socio-economic status, better self-rated health, and greater mobile device proficiency, optimism, innovativeness, but also less insecurity with technology, compared to the Indecisive, and then to the Skeptics. Our findings suggest that most middle-aged and older adults in Hong Kong hold an ambivalent view towards AI, appreciating its power and potentials while also cognizant of the perils it may entail. Our findings are timely considering the recent debates on ethical use of AI evoked by smart phone applications such as ChatGPT and will be valuable for practitioners and scholars for developing inclusive AI-facilitated services and applications.</p></div>","PeriodicalId":100324,"journal":{"name":"Computers in Human Behavior: Artificial Humans","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949882124000513/pdfft?md5=4615a367816801203b2516b1fae73372&pid=1-s2.0-S2949882124000513-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142136503","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01DOI: 10.1016/j.chbah.2024.100092
This paper aims to identify the ethical considerations driving the acceptance of and resistance to the use of insideable technology for human enhancement purposes, which are crucial to understand for the development of the cyborg technology market and businesses. While the literature privileges quantitative approaches, investigations focused on a strand of ethical theory or a specific value, this study adopts a qualitative and holistic approach. Based on prior interview data and a literature review, 33 items representing various ethical considerations of interest are identified. A qualitative Q-study was conducted, in which 55 individuals in three different countries expressed their points of view on insideables regarding these items. Hence, four different views are presented, highlighting drivers of acceptance of human enhancement technologies, conditional acceptance, and mere rejection. These views reveal the trade-offs between values made by respondents, shedding light on the ethical bricolage at play. The role of ethical concerns and theories in models to study the acceptance of human enhancement technologies and their potential business implications are discussed.
{"title":"The ethical acceptability of human enhancement technologies: A cross-country Q-study of the perception of insideables","authors":"","doi":"10.1016/j.chbah.2024.100092","DOIUrl":"10.1016/j.chbah.2024.100092","url":null,"abstract":"<div><div>This paper aims to identify the ethical considerations driving the acceptance of and resistance to the use of insideable technology for human enhancement purposes, which are crucial to understand for the development of the cyborg technology market and businesses. While the literature privileges quantitative approaches, investigations focused on a strand of ethical theory or a specific value, this study adopts a qualitative and holistic approach. Based on prior interview data and a literature review, 33 items representing various ethical considerations of interest are identified. A qualitative Q-study was conducted, in which 55 individuals in three different countries expressed their points of view on insideables regarding these items. Hence, four different views are presented, highlighting drivers of acceptance of human enhancement technologies, conditional acceptance, and mere rejection. These views reveal the trade-offs between values made by respondents, shedding light on the ethical bricolage at play. The role of ethical concerns and theories in models to study the acceptance of human enhancement technologies and their potential business implications are discussed.</div></div>","PeriodicalId":100324,"journal":{"name":"Computers in Human Behavior: Artificial Humans","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142318424","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-02DOI: 10.1016/j.chbah.2024.100081
Sucharat Limpanopparat, Erin Gibson, Dr Andrew Harris
Background
In recent years, chatbots developed for mental health intervention purposes have been widely implemented to solve the challenges of workforce shortage and accessibility issues faced by traditional health services. Nevertheless, research assessing the technologies’ potential and risks remains sporadic.
Purpose
This review aims to synthesise the existing research on engagement, user attitude, and effectiveness of psychological chatbot interventions.
Method
A systematic review was conducted using relevant peer-reviewed literature since 2010. These studies were derived from six databases: PubMed, PsycINFO, Web ofScience, Science Direct, Scopus and IEEE Xplore.
Results
Engagement level with chatbots that complied with digital intervention standards, lead to positive mental health outcomes. Although users had some uncertainties about the usability of these tools, positive attitudes towards chatbots regarding user experience and acceptability were frequently identified due to the chatbots' psychological capabilities and unique functions. High levels of outcome efficacy were found for those with depression. The differences in demographics, psychological approaches, and featured technologies could also influence the extent of mental health chatbot performances.
Conclusion
Positive attitudes and engagement with chatbots, as well as positive mental health outcomes, shows chatbot technology is a promising modality for mental health intervention. However, implementing them amongst some demographics or with novel features should be carefully considered. Further research using mainstream mental health chatbots and evaluating them simultaneously with standardised measures of engagement, user attitude, and effectiveness is necessary for intervention development.
背景近年来,以心理健康干预为目的开发的聊天机器人得到广泛应用,以解决传统医疗服务面临的劳动力短缺和可及性问题。本综述旨在综合现有的关于心理聊天机器人干预的参与度、用户态度和有效性的研究。这些研究来自六个数据库:PubMed、PsycINFO、Web of Science、Science Direct、Scopus 和 IEEE Xplore。ResultsEngagement level with chatbots that comply with digital intervention standards, lead to positive mental health outcomes.虽然用户对这些工具的可用性存在一些不确定性,但由于聊天机器人的心理能力和独特功能,用户经常对聊天机器人的用户体验和可接受性持积极态度。抑郁症患者的疗效水平很高。人口统计学、心理学方法和特色技术的差异也会影响心理健康聊天机器人的表现程度。 结论对聊天机器人的积极态度和参与度以及积极的心理健康结果表明,聊天机器人技术是一种很有前景的心理健康干预方式。然而,在某些人群中使用聊天机器人或使用新功能时应慎重考虑。有必要使用主流心理健康聊天机器人开展进一步研究,并同时使用参与度、用户态度和有效性的标准化测量方法对其进行评估,以促进干预措施的开发。
{"title":"User engagement, attitudes, and the effectiveness of chatbots as a mental health intervention: A systematic review","authors":"Sucharat Limpanopparat, Erin Gibson, Dr Andrew Harris","doi":"10.1016/j.chbah.2024.100081","DOIUrl":"https://doi.org/10.1016/j.chbah.2024.100081","url":null,"abstract":"<div><h3>Background</h3><p>In recent years, chatbots developed for mental health intervention purposes have been widely implemented to solve the challenges of workforce shortage and accessibility issues faced by traditional health services. Nevertheless, research assessing the technologies’ potential and risks remains sporadic.</p></div><div><h3>Purpose</h3><p>This review aims to synthesise the existing research on engagement, user attitude, and effectiveness of psychological chatbot interventions.</p></div><div><h3>Method</h3><p>A systematic review was conducted using relevant peer-reviewed literature since 2010. These studies were derived from six databases: PubMed<em>, PsycINFO</em>, <em>Web of</em> <em>Science</em>, <em>Science Direct, Scopus</em> and <em>IEEE Xplore</em>.</p></div><div><h3>Results</h3><p>Engagement level with chatbots that complied with digital intervention standards, lead to positive mental health outcomes. Although users had some uncertainties about the usability of these tools, positive attitudes towards chatbots regarding user experience and acceptability were frequently identified due to the chatbots' psychological capabilities and unique functions. High levels of outcome efficacy were found for those with depression. The differences in demographics, psychological approaches, and featured technologies could also influence the extent of mental health chatbot performances.</p></div><div><h3>Conclusion</h3><p><em>P</em>ositive attitudes and engagement with chatbots, as well as positive mental health outcomes, shows chatbot technology is a promising modality for mental health intervention. However, implementing them amongst some demographics or with novel features should be carefully considered. Further research using mainstream mental health chatbots and evaluating them simultaneously with standardised measures of engagement, user attitude, and effectiveness is necessary for intervention development.</p></div>","PeriodicalId":100324,"journal":{"name":"Computers in Human Behavior: Artificial Humans","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949882124000410/pdfft?md5=28fa4639b941c7cab725c225999b1bd0&pid=1-s2.0-S2949882124000410-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141543758","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}