Pub Date : 2024-09-09DOI: 10.1007/s10919-024-00474-9
Philip U. Gustafsson, Tim Lachmann, Petri Laukka
An important task in criminal justice is to evaluate the accuracy of eyewitness testimony. In this study, we examined if machine learning could be used to detect accuracy. Specifically, we examined if support vector machines (SVMs) could accurately classify testimony statements as correct or incorrect based purely on the nonverbal aspects of the voice. We analyzed 3,337 statements (76.61% accurate) from 51 eyewitness testimonies along 94 acoustic variables. We also examined the relative importance of each of the acoustic variables, using Lasso regression. Results showed that the machine learning algorithms were able to predict accuracy between 20 and 40% above chance level (AUC = 0.50). The most important predictors included acoustic variables related to the amplitude (loudness) of speech and the duration of pauses, with higher amplitude predicting correct recall and longer pauses predicting incorrect recall. Taken together, we find that machine learning methods are capable of predicting whether eyewitness testimonies are correct or incorrect with above-chance accuracy and comparable to human performance, but without detrimental human biases. This offers a proof-of-concept for machine learning in evaluations of eyewitness accuracy, and opens up new avenues of research that we hope might improve social justice.
{"title":"Machine Learning Predicts Accuracy in Eyewitnesses’ Voices","authors":"Philip U. Gustafsson, Tim Lachmann, Petri Laukka","doi":"10.1007/s10919-024-00474-9","DOIUrl":"https://doi.org/10.1007/s10919-024-00474-9","url":null,"abstract":"<p>An important task in criminal justice is to evaluate the accuracy of eyewitness testimony. In this study, we examined if machine learning could be used to detect accuracy. Specifically, we examined if support vector machines (SVMs) could accurately classify testimony statements as correct or incorrect based purely on the nonverbal aspects of the voice. We analyzed 3,337 statements (76.61% accurate) from 51 eyewitness testimonies along 94 acoustic variables. We also examined the relative importance of each of the acoustic variables, using Lasso regression. Results showed that the machine learning algorithms were able to predict accuracy between 20 and 40% above chance level (AUC = 0.50). The most important predictors included acoustic variables related to the amplitude (loudness) of speech and the duration of pauses, with higher amplitude predicting correct recall and longer pauses predicting incorrect recall. Taken together, we find that machine learning methods are capable of predicting whether eyewitness testimonies are correct or incorrect with above-chance accuracy and comparable to human performance, but without detrimental human biases. This offers a proof-of-concept for machine learning in evaluations of eyewitness accuracy, and opens up new avenues of research that we hope might improve social justice.</p>","PeriodicalId":47747,"journal":{"name":"Journal of Nonverbal Behavior","volume":"31 1","pages":""},"PeriodicalIF":2.1,"publicationDate":"2024-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142207074","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-04DOI: 10.1007/s10919-024-00472-x
Valérie Coulombe, Vincent Martel-Sauvageau, Laura Monetta
This study examines the ability to express distinct emotions of negative and positive valences through voice modulations (i.e., affective prosody production) and how the speaker’s gender and age influence this ability. A sample of 31 neurologically healthy adults (17 women and 14 men, aged 41–76) were asked to say “papa” with six emotional tones—sadness, anger, fear, pleasant surprise, joy, and awe—in response to affect-evoking scenarios. The speakers’ vocal expressions were recorded and then assessed by five expert raters and 30 naive listeners using an emotion recognition task. Results showed that negative emotions were expressed more accurately than positive ones, highlighting a valence effect. In addition, female speakers showed higher recognition rates for their expressions of vocal emotions than male speakers. Furthermore, aging was associated with a moderate decline in the accuracy of prosodic emotional expression. Despite generally lower recognition rates from naive listeners compared to expert raters, recognition rates for all emotions, with the exception of awe, were not statistically different between listener groups. In conclusion, cognitively healthy adults can convey discrete emotions through prosody, including distinct positive emotions, but there are significant differences depending on the emotion expressed and individual speaker characteristics. These results highlight the complexity of affective prosody production and contribute to the understanding of individual differences in nonverbal emotional expression.
{"title":"The Expression of Vocal Emotions in Cognitively Healthy Adult Speakers: Impact of Emotion Category, Gender, and Age","authors":"Valérie Coulombe, Vincent Martel-Sauvageau, Laura Monetta","doi":"10.1007/s10919-024-00472-x","DOIUrl":"https://doi.org/10.1007/s10919-024-00472-x","url":null,"abstract":"<p>This study examines the ability to express distinct emotions of negative and positive valences through voice modulations (i.e., affective prosody production) and how the speaker’s gender and age influence this ability. A sample of 31 neurologically healthy adults (17 women and 14 men, aged 41–76) were asked to say “papa” with six emotional tones—sadness, anger, fear, pleasant surprise, joy, and awe—in response to affect-evoking scenarios. The speakers’ vocal expressions were recorded and then assessed by five expert raters and 30 naive listeners using an emotion recognition task. Results showed that negative emotions were expressed more accurately than positive ones, highlighting a valence effect. In addition, female speakers showed higher recognition rates for their expressions of vocal emotions than male speakers. Furthermore, aging was associated with a moderate decline in the accuracy of prosodic emotional expression. Despite generally lower recognition rates from naive listeners compared to expert raters, recognition rates for all emotions, with the exception of awe, were not statistically different between listener groups. In conclusion, cognitively healthy adults can convey discrete emotions through prosody, including distinct positive emotions, but there are significant differences depending on the emotion expressed and individual speaker characteristics. These results highlight the complexity of affective prosody production and contribute to the understanding of individual differences in nonverbal emotional expression.</p>","PeriodicalId":47747,"journal":{"name":"Journal of Nonverbal Behavior","volume":"6 1","pages":""},"PeriodicalIF":2.1,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142207072","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-18DOI: 10.1007/s10919-024-00471-y
Xia Fang, Kerry Kawakami
Since the beginning of the COVID-19 pandemic in early 2020, face masks have become a common experience for many people to reduce the spread of the disease. Although recent research has shown that face masks impair emotion recognition, it is unclear how this impairment differs from other familiar types of face covering, such as sunglasses. In the present study, participants identified expressions and rated their confidence in this assessment of six affective expressions (anger, disgust, fear, surprise, sadness, and happiness) on faces wearing masks or sunglasses at four different time points during the pandemic (June 2020, March 2021, September 2021, June 2022). They also provided judgements of emotion intensity and genuineness. Overall, emotion identification of faces with masks was less accurate and had lower ratings of confidence and emotion intensity than faces with sunglasses. Faces with sunglasses, alternatively, were rated as less genuine than faces with masks. Furthermore, this pattern for both masks and sunglasses remained stable across two years of the pandemic. This study provides new insights on the differential effects of face masks and sunglasses on emotion perception and highlights the importance of face coverings for emotion communication and social interactions.
{"title":"The Effect of Face Masks and Sunglasses on Emotion Perception over Two Years of the COVID-19 Pandemic","authors":"Xia Fang, Kerry Kawakami","doi":"10.1007/s10919-024-00471-y","DOIUrl":"https://doi.org/10.1007/s10919-024-00471-y","url":null,"abstract":"<p>Since the beginning of the COVID-19 pandemic in early 2020, face masks have become a common experience for many people to reduce the spread of the disease. Although recent research has shown that face masks impair emotion recognition, it is unclear how this impairment differs from other familiar types of face covering, such as sunglasses. In the present study, participants identified expressions and rated their confidence in this assessment of six affective expressions (anger, disgust, fear, surprise, sadness, and happiness) on faces wearing masks or sunglasses at four different time points during the pandemic (June 2020, March 2021, September 2021, June 2022). They also provided judgements of emotion intensity and genuineness. Overall, emotion identification of faces with masks was less accurate and had lower ratings of confidence and emotion intensity than faces with sunglasses. Faces with sunglasses, alternatively, were rated as less genuine than faces with masks. Furthermore, this pattern for both masks and sunglasses remained stable across two years of the pandemic. This study provides new insights on the differential effects of face masks and sunglasses on emotion perception and highlights the importance of face coverings for emotion communication and social interactions.</p>","PeriodicalId":47747,"journal":{"name":"Journal of Nonverbal Behavior","volume":"76 1","pages":""},"PeriodicalIF":2.1,"publicationDate":"2024-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142207073","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-06DOI: 10.1007/s10919-024-00470-z
Kirsty L. Johnstone, Chris Martin, Mark Blades
Child abuse is a major concern worldwide. While live-link interviews have been successful in legal and medical contexts, its potential for eyewitness interviews remains insufficiently studied, particularly in terms of non-verbal misinformation. This study explored tele-forensic interviewing (tele-FI), where video-conferencing software such as Zoom or Skype are used to conduct forensic interviews, as an alternative to face-to-face interviews. Focus was given to the susceptibility of eyewitness memory to the gestural misinformation effect (GME) where post-event information in the form of gesture can distort recall of a witnessed incident. Forty-seven children were recruited ranging in age from 5- to 8-years-old (M = 6 years 11 months). Comparisons were made to face-to-face conditions from prior published work by the authors (N = 63, M = 7 years 2 months) using the same methodology, video, and question sets. Results find support for the GME during tele-FI, with 1.23 misinformation details recorded on average and tele-FI showing a similar response pattern as face-to-face interviews. Accuracy was shown to be comparable in tele-FI (M = 16.21) compared to face-to-face interviews (M = 14.02), with a notable increase in the amount of relevant information provided in the tele-FI condition. Developmental age showed significant increases in the quality and quantity of data. This study provides evidence for tele-FI as a viable alternative to face-to-face interviews, and represents the first exploration of the GME in tele-FI, to the best of our knowledge. Discussion focuses on the benefits of tele-FI and the implications for police interview guidelines.
虐待儿童是全世界关注的一个主要问题。虽然实时链接访谈在法律和医疗领域取得了成功,但其在目击者访谈方面的潜力仍未得到充分研究,特别是在非语言错误信息方面。本研究探讨了远程法证访谈(tele-FI),即使用 Zoom 或 Skype 等视频会议软件进行法证访谈,以替代面对面访谈。研究的重点是目击者的记忆是否容易受到手势误导效应(GME)的影响,即事件发生后以手势形式出现的信息会扭曲对目击事件的回忆。研究共招募了 47 名儿童,年龄从 5 岁到 8 岁不等(平均年龄为 6 岁 11 个月)。使用相同的方法、视频和问题集,与作者之前发表的作品(N = 63,M = 7 岁 2 个月)中的面对面条件进行了比较。结果发现,远程 FI 支持 GME,平均记录了 1.23 个错误信息细节,远程 FI 显示出与面对面访谈类似的反应模式。与面对面访谈(M = 14.02)相比,远程-口语(M = 16.21)的准确性相当,但远程-口语条件下提供的相关信息量明显增加。从发育年龄来看,数据的质量和数量都有明显增加。这项研究为远程 FI 作为面对面访谈的一种可行替代方法提供了证据,据我们所知,这是首次对远程 FI 中的 GME 进行探讨。讨论的重点是远程 FI 的益处以及对警方面谈指南的影响。
{"title":"The Digital Witness: Exploring Gestural Misinformation in Tele-Forensic Interviews with 5-8-Year-Old Children","authors":"Kirsty L. Johnstone, Chris Martin, Mark Blades","doi":"10.1007/s10919-024-00470-z","DOIUrl":"https://doi.org/10.1007/s10919-024-00470-z","url":null,"abstract":"<p>Child abuse is a major concern worldwide. While live-link interviews have been successful in legal and medical contexts, its potential for eyewitness interviews remains insufficiently studied, particularly in terms of non-verbal misinformation. This study explored tele-forensic interviewing (tele-FI), where video-conferencing software such as Zoom or Skype are used to conduct forensic interviews, as an alternative to face-to-face interviews. Focus was given to the susceptibility of eyewitness memory to the gestural misinformation effect (GME) where post-event information in the form of gesture can distort recall of a witnessed incident. Forty-seven children were recruited ranging in age from 5- to 8-years-old (<i>M</i> = 6 years 11 months). Comparisons were made to face-to-face conditions from prior published work by the authors (<i>N</i> = 63, <i>M</i> = 7 years 2 months) using the same methodology, video, and question sets. Results find support for the GME during tele-FI, with 1.23 misinformation details recorded on average and tele-FI showing a similar response pattern as face-to-face interviews. Accuracy was shown to be comparable in tele-FI (<i>M</i> = 16.21) compared to face-to-face interviews (<i>M</i> = 14.02), with a notable increase in the amount of relevant information provided in the tele-FI condition. Developmental age showed significant increases in the quality and quantity of data. This study provides evidence for tele-FI as a viable alternative to face-to-face interviews, and represents the first exploration of the GME in tele-FI, to the best of our knowledge. Discussion focuses on the benefits of tele-FI and the implications for police interview guidelines.</p>","PeriodicalId":47747,"journal":{"name":"Journal of Nonverbal Behavior","volume":"37 1","pages":""},"PeriodicalIF":2.1,"publicationDate":"2024-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141943920","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-10DOI: 10.1007/s10919-024-00469-6
Jillian J. M. O’Connor
Previous research has found that higher-pitched female voices elicit jealousy among women. However, it is unknown whether jealousy towards higher-pitched female voices is driven by perceptions of the rival’s mating strategy or by beliefs about the speaker’s attractiveness to one’s romantic partner. In addition, the degree to which higher-pitched female voices elicit jealousy could be associated with variation in trait jealousy among women listeners. Here, I manipulated women’s voices to be higher or lower in pitch, and tested whether variation in jealousy towards female voices was more strongly associated with perceptions of mate poaching, beliefs about the speaker’s attractiveness to listeners’ romantic partner, or with individual differences in trait jealousy. I replicated findings that higher voice pitch elicits more jealousy from women, which was positively associated with perceptions of mate poaching. I found no evidence of an association between trait jealousy and any voice-based perception. The findings suggest that perceptions of a target’s proclivity to mate poach better explain the jealousy-inducing nature of higher-pitched female voices than do beliefs about the speaker’s attractiveness to one’s romantic partner.
{"title":"Perceptions of mate poaching predict jealousy towards higher-pitched women’s voices","authors":"Jillian J. M. O’Connor","doi":"10.1007/s10919-024-00469-6","DOIUrl":"https://doi.org/10.1007/s10919-024-00469-6","url":null,"abstract":"<p>Previous research has found that higher-pitched female voices elicit jealousy among women. However, it is unknown whether jealousy towards higher-pitched female voices is driven by perceptions of the rival’s mating strategy or by beliefs about the speaker’s attractiveness to one’s romantic partner. In addition, the degree to which higher-pitched female voices elicit jealousy could be associated with variation in trait jealousy among women listeners. Here, I manipulated women’s voices to be higher or lower in pitch, and tested whether variation in jealousy towards female voices was more strongly associated with perceptions of mate poaching, beliefs about the speaker’s attractiveness to listeners’ romantic partner, or with individual differences in trait jealousy. I replicated findings that higher voice pitch elicits more jealousy from women, which was positively associated with perceptions of mate poaching. I found no evidence of an association between trait jealousy and any voice-based perception. The findings suggest that perceptions of a target’s proclivity to mate poach better explain the jealousy-inducing nature of higher-pitched female voices than do beliefs about the speaker’s attractiveness to one’s romantic partner.</p>","PeriodicalId":47747,"journal":{"name":"Journal of Nonverbal Behavior","volume":"29 1","pages":""},"PeriodicalIF":2.1,"publicationDate":"2024-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141587563","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-26DOI: 10.1007/s10919-024-00467-8
G. Shane Pitts, Nicholas O. Rule
Whereas research has documented how atheists are perceived, none has considered their perceptibility. Atheists must first be identified as atheists in order to experience the stigma associated with them (i.e., as distrusted, disliked, and widely maligned). Although atheism is considered a concealable aspect of one’s identity, substantial research has found that a variety of ostensibly concealable attributes about a person are indeed legible from small and subtle cues. We merged these lines of inquiry here by considering the perceptibility of religious and spiritual (dis)belief. Studies 1A-1B showed that atheists could be reliably discerned from Christians based on brief glimpses of 100 standardized male faces. Experiment 2 replicated these results using female faces. Experiments 3 A-E then interrogated the facial features that support perceivers’ detection of atheism, showing that various parts of faces suffice for independently conveying atheism. Experiment 4 investigated and showed a potential mechanism for atheism detection – expressive suppression. Thus, across nine studies (N = 677), these data show robust evidence that atheists can be categorized from facial cues.
{"title":"Atheists and Christians can be Discerned from their Faces","authors":"G. Shane Pitts, Nicholas O. Rule","doi":"10.1007/s10919-024-00467-8","DOIUrl":"https://doi.org/10.1007/s10919-024-00467-8","url":null,"abstract":"<p>Whereas research has documented how atheists are <i>perceived</i>, none has considered their <i>perceptibility</i>. Atheists must first be identified as atheists in order to experience the stigma associated with them (i.e., as distrusted, disliked, and widely maligned). Although atheism is considered a concealable aspect of one’s identity, substantial research has found that a variety of ostensibly concealable attributes about a person are indeed legible from small and subtle cues. We merged these lines of inquiry here by considering the perceptibility of religious and spiritual (dis)belief. Studies 1A-1B showed that atheists could be reliably discerned from Christians based on brief glimpses of 100 standardized male faces. Experiment 2 replicated these results using female faces. Experiments 3 A-E then interrogated the facial features that support perceivers’ detection of atheism, showing that various parts of faces suffice for independently conveying atheism. Experiment 4 investigated and showed a potential mechanism for atheism detection – expressive suppression. Thus, across nine studies (<i>N</i> = 677), these data show robust evidence that atheists can be categorized from facial cues.</p>","PeriodicalId":47747,"journal":{"name":"Journal of Nonverbal Behavior","volume":"48 1","pages":""},"PeriodicalIF":2.1,"publicationDate":"2024-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141506673","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-22DOI: 10.1007/s10919-024-00468-7
Christiane Völter, Kirsten Oberländer, Martin Brüne, Fabian T. Ramseyer
Hearing loss severely hampers verbal exchange and thus social interaction, which puts a high burden on hearing-impaired and their close partners. Until now, nonverbal interaction in hearing-impaired dyads has not been addressed as a relevant factor for well-being or quality of social relationships. Nonverbal synchrony of head- and body-movement was analysed in N = 30 dyads of persons with hearing impairment (PHI) and their significant others (SO). In a 10-minute conversation before (T1) and 6 months after cochlear implantation (T2), Motion Energy Analysis (MEA) automatically quantified head- and body-movement. Self-report measures of both dyad members were used to assess aspects of quality of life and closeness in the partnership. After cochlear implantation, nonverbal synchrony showed a downward trend and was less distinct from pseudosynchrony. Higher synchrony was associated with worse hearing-related quality of life, shorter duration of hearing impairment and less closeness in the relationship. This negative association was interpreted as an indication for the effort one has to make to cope with difficulties in a dyad`s relationship. Endorsing a holistic approach in auditory rehabilitation, we propose the assessment of nonverbal synchrony as a suitable tool to detect subtle imbalances in the interpersonal relation between PHI and SO outside conscious control and to provide cues for possible therapeutical strategies.
听力损失严重阻碍了语言交流和社会交往,这给听障人士及其亲密伙伴带来了沉重负担。迄今为止,听力障碍者之间的非语言互动还没有被作为影响社会关系的幸福感或质量的相关因素加以研究。我们对 N = 30 个听障人士(PHI)及其重要伴侣(SO)的头部和身体运动的非语言同步性进行了分析。在人工耳蜗植入前(T1)和植入后 6 个月(T2)的 10 分钟对话中,运动能量分析法(MEA)自动量化了头部和身体的运动。对话双方的自我报告测量用于评估生活质量和伙伴关系中的亲密程度。人工耳蜗植入后,非言语同步性呈下降趋势,与假同步性的差异较小。同步性越高,与听力相关的生活质量越差,听力受损时间越短,关系越不亲密。这种负相关被解释为表明一个人必须付出努力来应对两人关系中的困难。我们赞同听觉康复的整体方法,并建议将非语言同步性评估作为一种合适的工具,用于检测 PHI 和 SO 之间人际关系中有意识控制之外的微妙失衡,并为可能的治疗策略提供线索。
{"title":"Impact of Hearing Loss and Auditory Rehabilitation on Dyads: A Microsocial Perspective","authors":"Christiane Völter, Kirsten Oberländer, Martin Brüne, Fabian T. Ramseyer","doi":"10.1007/s10919-024-00468-7","DOIUrl":"https://doi.org/10.1007/s10919-024-00468-7","url":null,"abstract":"<p>Hearing loss severely hampers verbal exchange and thus social interaction, which puts a high burden on hearing-impaired and their close partners. Until now, nonverbal interaction in hearing-impaired dyads has not been addressed as a relevant factor for well-being or quality of social relationships. Nonverbal synchrony of head- and body-movement was analysed in <i>N</i> = 30 dyads of persons with hearing impairment (PHI) and their significant others (SO). In a 10-minute conversation before (T1) and 6 months after cochlear implantation (T2), Motion Energy Analysis (MEA) automatically quantified head- and body-movement. Self-report measures of both dyad members were used to assess aspects of quality of life and closeness in the partnership. After cochlear implantation, nonverbal synchrony showed a downward trend and was less distinct from pseudosynchrony. Higher synchrony was associated with worse hearing-related quality of life, shorter duration of hearing impairment and less closeness in the relationship. This negative association was interpreted as an indication for the effort one has to make to cope with difficulties in a dyad`s relationship. Endorsing a holistic approach in auditory rehabilitation, we propose the assessment of nonverbal synchrony as a suitable tool to detect subtle imbalances in the interpersonal relation between PHI and SO outside conscious control and to provide cues for possible therapeutical strategies.</p>","PeriodicalId":47747,"journal":{"name":"Journal of Nonverbal Behavior","volume":"2016 1","pages":""},"PeriodicalIF":2.1,"publicationDate":"2024-06-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141506674","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-01DOI: 10.1007/s10919-024-00466-9
Christian Hoffmann, Magdalene Gürtler, Johannes Fendel, Claas Lahmann, Stefan Schmidt
The present study investigated the differences in movement synchrony and therapeutic alliance between solution-focused and problem-focused counseling. Thirty-four participants each attended two counseling sessions with different counselors, one with a solution-focus and one with a problem-focus, in randomized order. The sessions consisted of three consecutive parts: problem description, standardized intervention and free intervention. Movement synchrony, including leading and pacing synchrony, was measured using Motion Energy Analysis (MEA) and windowed cross-lagged correlation (WCLC) based on video recordings of the sessions. The Helping Alliance Questionnaire (HAQ) was used to assess therapeutic alliance. Results showed that movement synchrony was significantly higher in solution-focused than in problem-focused counseling, driven by differences in the problem description part. This difference may be explained by the allegiance of the counselors to the solution-focused approach, as we observed more leading synchrony during the problem description part in solution-focused sessions. There was no significant difference in therapeutic alliance between the two conditions. This study expands the understanding of counseling approaches in the field of movement synchrony and contributes valuable insights for practitioners and researchers alike.
{"title":"Assessment of Movement Synchrony and Alliance in Problem-Focused and Solution-Focused Counseling","authors":"Christian Hoffmann, Magdalene Gürtler, Johannes Fendel, Claas Lahmann, Stefan Schmidt","doi":"10.1007/s10919-024-00466-9","DOIUrl":"https://doi.org/10.1007/s10919-024-00466-9","url":null,"abstract":"<p>The present study investigated the differences in movement synchrony and therapeutic alliance between solution-focused and problem-focused counseling. Thirty-four participants each attended two counseling sessions with different counselors, one with a solution-focus and one with a problem-focus, in randomized order. The sessions consisted of three consecutive parts: problem description, standardized intervention and free intervention. Movement synchrony, including leading and pacing synchrony, was measured using Motion Energy Analysis (MEA) and windowed cross-lagged correlation (WCLC) based on video recordings of the sessions. The Helping Alliance Questionnaire (HAQ) was used to assess therapeutic alliance. Results showed that movement synchrony was significantly higher in solution-focused than in problem-focused counseling, driven by differences in the problem description part. This difference may be explained by the allegiance of the counselors to the solution-focused approach, as we observed more leading synchrony during the problem description part in solution-focused sessions. There was no significant difference in therapeutic alliance between the two conditions. This study expands the understanding of counseling approaches in the field of movement synchrony and contributes valuable insights for practitioners and researchers alike.</p>","PeriodicalId":47747,"journal":{"name":"Journal of Nonverbal Behavior","volume":"1 1","pages":""},"PeriodicalIF":2.1,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141189416","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-28DOI: 10.1007/s10919-024-00464-x
Supreet Saluja, Ilona Croy, Richard J. Stevenson
There appears to be no attempt to categorize the specific classes of behavior that the tactile system underpins. Awareness of how an organism uses touch in their environment informs understanding of its versatility in non-verbal communication and tactile perception. This review categorizes the behavioral functions underpinned by the tactile sense, by using three sources of data: (1) Animal data, to assess if an identified function is conserved across species; (2) Human capacity data, indicating whether the tactile sense can support a proposed function; and (3) Human impaired data, documenting the impacts of impaired tactile functioning (e.g., reduced tactile sensitivity) for humans. From these data, three main functions pertinent to the tactile sense were identified: Ingestive Behavior; Environmental Hazard Detection and Management; and Social Communication. These functions are reviewed in detail and future directions are discussed with focus on social psychology, non-verbal behavior and multisensory perception.
{"title":"The Functions of Human Touch: An Integrative Review","authors":"Supreet Saluja, Ilona Croy, Richard J. Stevenson","doi":"10.1007/s10919-024-00464-x","DOIUrl":"https://doi.org/10.1007/s10919-024-00464-x","url":null,"abstract":"<p>There appears to be no attempt to categorize the specific classes of behavior that the tactile system underpins. Awareness of how an organism uses touch in their environment informs understanding of its versatility in non-verbal communication and tactile perception. This review categorizes the behavioral functions underpinned by the tactile sense, by using three sources of data: (1) Animal data, to assess if an identified function is conserved across species; (2) Human capacity data, indicating whether the tactile sense can support a proposed function; and (3) Human impaired data, documenting the impacts of impaired tactile functioning (e.g., reduced tactile sensitivity) for humans. From these data, three main functions pertinent to the tactile sense were identified: Ingestive Behavior; Environmental Hazard Detection and Management; and Social Communication. These functions are reviewed in detail and future directions are discussed with focus on social psychology, non-verbal behavior and multisensory perception.</p>","PeriodicalId":47747,"journal":{"name":"Journal of Nonverbal Behavior","volume":"56 1","pages":""},"PeriodicalIF":2.1,"publicationDate":"2024-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141168203","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-09DOI: 10.1007/s10919-024-00461-0
Samantha J. Shebib, Josephine K. Boumis, Amanda Allard, Amanda J. Holmstrom, Adam J. Mason
The present study examines how supportive touch impacts evaluations of esteem support content containing high emotion-focused (HEF) or high problem-focused (HPF) messages during observed esteem support interactions. A 2 (verbal content; i.e., HEF or HPF) by 2 (nonverbal content; i.e., presence or absence of supportive tactile communication) experiment was conducted to test for main and interactional effects. Results revealed that HEF conditions were perceived to be more effective by observers at enhancing the recipient’s state self-esteem, state self-efficacy, and alleviating distress compared to HPF conditions. The supportive tactile communication conditions were perceived as better at enhancing state self-esteem and alleviating distress compared to the no supportive tactile communication conditions by observers. However, these main effects were qualified by significant two-way interactions between message content and nonverbal behavior on ratings of state self-esteem and distress alleviation, such that the addition of supportive tactile communication enhanced the effectiveness of HPF message content but not HEF content.
{"title":"An Experimental Investigation of Supportive Tactile Communication During Esteem Support Conversations","authors":"Samantha J. Shebib, Josephine K. Boumis, Amanda Allard, Amanda J. Holmstrom, Adam J. Mason","doi":"10.1007/s10919-024-00461-0","DOIUrl":"https://doi.org/10.1007/s10919-024-00461-0","url":null,"abstract":"<p>The present study examines how supportive touch impacts evaluations of esteem support content containing high emotion-focused (HEF) or high problem-focused (HPF) messages during observed esteem support interactions. A 2 (verbal content; i.e., HEF or HPF) by 2 (nonverbal content; i.e., presence or absence of supportive tactile communication) experiment was conducted to test for main and interactional effects. Results revealed that HEF conditions were perceived to be more effective by observers at enhancing the recipient’s state self-esteem, state self-efficacy, and alleviating distress compared to HPF conditions. The supportive tactile communication conditions were perceived as better at enhancing state self-esteem and alleviating distress compared to the no supportive tactile communication conditions by observers. However, these main effects were qualified by significant two-way interactions between message content and nonverbal behavior on ratings of state self-esteem and distress alleviation, such that the addition of supportive tactile communication enhanced the effectiveness of HPF message content but not HEF content.</p>","PeriodicalId":47747,"journal":{"name":"Journal of Nonverbal Behavior","volume":"187 1","pages":""},"PeriodicalIF":2.1,"publicationDate":"2024-04-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140562976","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}