首页 > 最新文献

IEEE Transactions on Computational Social Systems最新文献

英文 中文
Multimodal Disentangled Fusion Network via VAEs for Multimodal Zero-Shot Learning 基于vae的多模态解纠缠融合网络用于多模态零射击学习
IF 4.5 2区 计算机科学 Q1 COMPUTER SCIENCE, CYBERNETICS Pub Date : 2025-07-08 DOI: 10.1109/TCSS.2025.3575939
Yutian Li;Zhuopan Yang;Zhenguo Yang;Xiaoping Li;Wenyin Liu;Qing Li
Addressing the bias problem in multimodal zero-shot learning tasks is challenging due to the domain shift between seen and unseen classes, as well as the semantic gap across different modalities. To tackle these challenges, we propose a multimodal disentangled fusion network (MDFN) that unifies the class embedding space for multimodal zero-shot learning. MDFN exploits feature disentangled variational autoencoder (FD-VAE) in two branches to distangle unimodal features into modality-specific representations that are semantically consistent and unrelated, where semantics are shared within classes. In particular, semantically consistent representations and unimodal features are integrated to retain the semantics of the original features in the form of residuals. Furthermore, multimodal conditional VAE (MC-VAE) in two branches is adopted to learn cross-modal interactions with modality-specific conditions. Finally, the complementary multimodal representations achieved by MC-VAE are encoded into a fusion network (FN) with a self-adaptive margin center loss (SAMC-loss) to predict target class labels in embedding forms. By learning the distance among domain samples, SAMC-loss promotes intraclass compactness and interclass separability. Experiments on zero-shot and news event datasets demonstrate the superior performance of MDFN, with the harmonic mean improved by 27.2% on the MMED dataset and 5.1% on the SUN dataset.
由于可见类和未见类之间的域转移以及不同模态之间的语义差距,解决多模态零射击学习任务中的偏差问题具有挑战性。为了解决这些挑战,我们提出了一种多模态解纠缠融合网络(MDFN),该网络统一了多模态零射击学习的类嵌入空间。mfn利用两个分支中的特征解纠缠变分自编码器(FD-VAE)将单模态特征分离为特定于模态的表示,这些表示在语义上一致且不相关,其中语义在类中共享。特别地,将语义一致的表示和单峰特征结合起来,以残差的形式保留原始特征的语义。此外,采用两个分支的多模态条件VAE (MC-VAE)来学习具有模态特定条件的跨模态交互。最后,将MC-VAE获得的互补多模态表示编码到一个融合网络(FN)中,该网络具有自适应边缘中心损失(SAMC-loss),用于预测嵌入形式中的目标类标签。SAMC-loss通过学习域样本之间的距离,提高了类内紧密性和类间可分离性。在零射击和新闻事件数据集上的实验证明了mfn的优越性能,在MMED数据集上谐波均值提高了27.2%,在SUN数据集上提高了5.1%。
{"title":"Multimodal Disentangled Fusion Network via VAEs for Multimodal Zero-Shot Learning","authors":"Yutian Li;Zhuopan Yang;Zhenguo Yang;Xiaoping Li;Wenyin Liu;Qing Li","doi":"10.1109/TCSS.2025.3575939","DOIUrl":"https://doi.org/10.1109/TCSS.2025.3575939","url":null,"abstract":"Addressing the bias problem in multimodal zero-shot learning tasks is challenging due to the domain shift between seen and unseen classes, as well as the semantic gap across different modalities. To tackle these challenges, we propose a multimodal disentangled fusion network (MDFN) that unifies the class embedding space for multimodal zero-shot learning. MDFN exploits feature disentangled variational autoencoder (FD-VAE) in two branches to distangle unimodal features into modality-specific representations that are semantically consistent and unrelated, where semantics are shared within classes. In particular, semantically consistent representations and unimodal features are integrated to retain the semantics of the original features in the form of residuals. Furthermore, multimodal conditional VAE (MC-VAE) in two branches is adopted to learn cross-modal interactions with modality-specific conditions. Finally, the complementary multimodal representations achieved by MC-VAE are encoded into a fusion network (FN) with a self-adaptive margin center loss (SAMC-loss) to predict target class labels in embedding forms. By learning the distance among domain samples, SAMC-loss promotes intraclass compactness and interclass separability. Experiments on zero-shot and news event datasets demonstrate the superior performance of MDFN, with the harmonic mean improved by 27.2% on the MMED dataset and 5.1% on the SUN dataset.","PeriodicalId":13044,"journal":{"name":"IEEE Transactions on Computational Social Systems","volume":"12 5","pages":"3684-3697"},"PeriodicalIF":4.5,"publicationDate":"2025-07-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145230085","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Coordinate System Transformation Method for Comparing Different Types of Data in Different Dataset Using Singular Value Decomposition 利用奇异值分解比较不同数据集中不同类型数据的坐标系变换方法
IF 4.5 2区 计算机科学 Q1 COMPUTER SCIENCE, CYBERNETICS Pub Date : 2025-07-08 DOI: 10.1109/TCSS.2025.3561078
Emiko Uchiyama;Wataru Takano;Yoshihiko Nakamura;Tomoki Tanaka;Katsuya Iijima;Gentiane Venture;Vincent Hernandez;Kenta Kamikokuryo;Ken-ichiro Yabu;Takahiro Miura;Kimitaka Nakazawa;Bo-Kyung Son
In the current era of AI technology, where systems increasingly rely on big data to process vast amounts of societal information, efficient methods for integrating and utilizing diverse datasets are essential. This article presents a novel approach for transforming the feature space of different datasets through singular value decomposition (SVD) to extract common and hidden features as using the prior domain knowledge. Specifically, we apply this method to two datasets: 1) one related to physical and cognitive frailty in the elderly; and 2) another focusing on identifying IKIGAI (happiness, self-efficacy, and sense of contribution) in volunteer staff of a civic health promotion activity. Both datasets consist of multiple sub-datasets measured using different modalities, such as facial expressions, sound, activity, and heart rates. By defining feature extraction methods for each subdataset, we compare and integrate the overlapping data. The results demonstrated that our method could effectively preserve common characteristics across different data types, offering a more interpretable solution than traditional dimensionality reduction methods based on linear and nonlinear transformation. This approach has significant implications for data integration in multidisciplinary fields and opens the door for future applications to a wide range of datasets.
在当前的人工智能技术时代,系统越来越依赖大数据来处理大量的社会信息,整合和利用各种数据集的有效方法至关重要。本文提出了一种利用先验领域知识,通过奇异值分解(SVD)变换不同数据集的特征空间,提取共同特征和隐藏特征的新方法。具体来说,我们将这种方法应用于两个数据集:1)一个与老年人的身体和认知虚弱有关;2)另一个重点是在公民健康促进活动的志愿者中确定IKIGAI(幸福,自我效能和贡献感)。这两个数据集由使用不同方式测量的多个子数据集组成,例如面部表情、声音、活动和心率。通过定义每个子数据集的特征提取方法,对重叠数据进行比较和整合。结果表明,该方法可以有效地保留不同数据类型的共同特征,提供了比基于线性和非线性变换的传统降维方法更具可解释性的解决方案。这种方法对多学科领域的数据集成具有重要意义,并为未来广泛的数据集应用打开了大门。
{"title":"Coordinate System Transformation Method for Comparing Different Types of Data in Different Dataset Using Singular Value Decomposition","authors":"Emiko Uchiyama;Wataru Takano;Yoshihiko Nakamura;Tomoki Tanaka;Katsuya Iijima;Gentiane Venture;Vincent Hernandez;Kenta Kamikokuryo;Ken-ichiro Yabu;Takahiro Miura;Kimitaka Nakazawa;Bo-Kyung Son","doi":"10.1109/TCSS.2025.3561078","DOIUrl":"https://doi.org/10.1109/TCSS.2025.3561078","url":null,"abstract":"In the current era of AI technology, where systems increasingly rely on big data to process vast amounts of societal information, efficient methods for integrating and utilizing diverse datasets are essential. This article presents a novel approach for transforming the feature space of different datasets through singular value decomposition (SVD) to extract common and hidden features as using the prior domain knowledge. Specifically, we apply this method to two datasets: 1) one related to physical and cognitive frailty in the elderly; and 2) another focusing on identifying <italic>IKIGAI</i> (happiness, self-efficacy, and sense of contribution) in volunteer staff of a civic health promotion activity. Both datasets consist of multiple sub-datasets measured using different modalities, such as facial expressions, sound, activity, and heart rates. By defining feature extraction methods for each subdataset, we compare and integrate the overlapping data. The results demonstrated that our method could effectively preserve common characteristics across different data types, offering a more interpretable solution than traditional dimensionality reduction methods based on linear and nonlinear transformation. This approach has significant implications for data integration in multidisciplinary fields and opens the door for future applications to a wide range of datasets.","PeriodicalId":13044,"journal":{"name":"IEEE Transactions on Computational Social Systems","volume":"12 5","pages":"3610-3626"},"PeriodicalIF":4.5,"publicationDate":"2025-07-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11073557","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145255998","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Impact of Listening to Music on Stress Level for Anxiety, Depression, and PTSD: Mixed-Effect Models and Propensity Score Analysis 音乐对焦虑、抑郁和创伤后应激障碍患者压力水平的影响:混合效应模型和倾向评分分析
IF 4.5 2区 计算机科学 Q1 COMPUTER SCIENCE, CYBERNETICS Pub Date : 2025-06-17 DOI: 10.1109/TCSS.2025.3561073
Mazin Abdalla;Parya Abadeh;Zeinab Noorian;Amira Ghenai;Fattane Zarrinkalam;Soroush Zamani Alavijeh
The intersection of music and mental health has gained increasing attention, with previous studies highlighting music’s potential to reduce stress and anxiety. Despite these promising findings, many of these studies are limited by small sample sizes and traditional observational methods, leaving a gap in our understanding of music’s broader impact on mental health. In response to these limitations, this study introduces a novel approach that combines generalized linear mixed models (GLMM) with propensity score matching (PSM) to explore the relationship between music listening and stress levels among social media users diagnosed with anxiety, depression, and posttraumatic stress disorder (PTSD). Our research not only identifies associative patterns between music listening and stress but also provides a more rigorous examination of potential causal effects, taking into account demographic factors such as education level, gender, and age. Our findings reveal that across all mental health conditions, music listening is significantly associated with reduced stress levels, with an observed 21.3% reduction for anxiety, 15.4% for depression, and 19.3% for PTSD. Additionally, users who listened to music were more likely to report a zero stress score, indicating a stronger relaxation effect. Further, our analysis of demographic variations shows that age and education level influence the impact of music on stress reduction, highlighting the potential for personalized interventions. These findings contribute to a deeper understanding of music’s therapeutic potential, particularly in crafting interventions tailored to the diverse needs of different populations.
音乐和心理健康的交集越来越受到关注,之前的研究强调了音乐减轻压力和焦虑的潜力。尽管有这些有希望的发现,但许多研究受到小样本和传统观察方法的限制,在我们对音乐对心理健康的广泛影响的理解上留下了空白。针对这些局限性,本研究引入了一种新的方法,将广义线性混合模型(GLMM)与倾向得分匹配(PSM)相结合,探索被诊断为焦虑、抑郁和创伤后应激障碍(PTSD)的社交媒体用户听音乐与压力水平之间的关系。我们的研究不仅确定了听音乐和压力之间的关联模式,而且考虑到教育水平、性别和年龄等人口因素,对潜在的因果关系进行了更严格的检查。我们的研究结果显示,在所有的心理健康状况中,听音乐与压力水平的降低显著相关,观察到焦虑降低21.3%,抑郁降低15.4%,创伤后应激障碍降低19.3%。此外,听音乐的用户更有可能报告压力得分为零,这表明放松效果更强。此外,我们对人口统计学变化的分析表明,年龄和教育水平会影响音乐对减轻压力的影响,这突出了个性化干预的潜力。这些发现有助于更深入地了解音乐的治疗潜力,特别是在为不同人群的不同需求量身定制干预措施方面。
{"title":"The Impact of Listening to Music on Stress Level for Anxiety, Depression, and PTSD: Mixed-Effect Models and Propensity Score Analysis","authors":"Mazin Abdalla;Parya Abadeh;Zeinab Noorian;Amira Ghenai;Fattane Zarrinkalam;Soroush Zamani Alavijeh","doi":"10.1109/TCSS.2025.3561073","DOIUrl":"https://doi.org/10.1109/TCSS.2025.3561073","url":null,"abstract":"The intersection of music and mental health has gained increasing attention, with previous studies highlighting music’s potential to reduce stress and anxiety. Despite these promising findings, many of these studies are limited by small sample sizes and traditional observational methods, leaving a gap in our understanding of music’s broader impact on mental health. In response to these limitations, this study introduces a novel approach that combines generalized linear mixed models (GLMM) with propensity score matching (PSM) to explore the relationship between music listening and stress levels among social media users diagnosed with anxiety, depression, and posttraumatic stress disorder (PTSD). Our research not only identifies associative patterns between music listening and stress but also provides a more rigorous examination of potential causal effects, taking into account demographic factors such as education level, gender, and age. Our findings reveal that across all mental health conditions, music listening is significantly associated with reduced stress levels, with an observed 21.3% reduction for anxiety, 15.4% for depression, and 19.3% for PTSD. Additionally, users who listened to music were more likely to report a zero stress score, indicating a stronger relaxation effect. Further, our analysis of demographic variations shows that age and education level influence the impact of music on stress reduction, highlighting the potential for personalized interventions. These findings contribute to a deeper understanding of music’s therapeutic potential, particularly in crafting interventions tailored to the diverse needs of different populations.","PeriodicalId":13044,"journal":{"name":"IEEE Transactions on Computational Social Systems","volume":"12 5","pages":"3816-3830"},"PeriodicalIF":4.5,"publicationDate":"2025-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145230080","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Differentiable Prior-Driven Data Augmentation for Sensor-Based Human Activity Recognition 基于传感器的人类活动识别的可微分先验驱动数据增强
IF 4.5 2区 计算机科学 Q1 COMPUTER SCIENCE, CYBERNETICS Pub Date : 2025-06-06 DOI: 10.1109/TCSS.2025.3565414
Ye Zhang;Qing Gao;Rong Hu;Qingtang Ding;Boyang Li;Yulan Guo
Sensor-based human activity recognition (HAR) usually suffers from the problem of insufficient annotated data, due to the difficulty in labeling the intuitive signals of wearable sensors. To this end, recent advances have adopted handcrafted operations or generative models for data augmentation. The handcrafted operations are driven by some physical priors of human activities, e.g., action distortion and strength fluctuations. However, these approaches may face challenges in maintaining semantic data properties. Although the generative models have better data adaptability, it is difficult for them to incorporate important action priors into data generation. This article proposes a differentiable prior-driven data augmentation framework for HAR. First, we embed the handcrafted augmentation operations into a differentiable module, which adaptively selects and optimizes the operations to be combined together. Then, we construct a generative module to add controllable perturbations to the data derived by the handcrafted operations and further improve the diversity of data augmentation. By integrating the handcrafted operation module and the generative module into one learnable framework, the generalization performance of the recognition models is enhanced effectively. Extensive experimental results with three different classifiers on five public datasets demonstrate the effectiveness of the proposed framework. Project page: https://github.com/crocodilegogogo/DriveData-Under-Review.
由于难以对可穿戴传感器的直观信号进行标注,基于传感器的人体活动识别(HAR)通常存在标注数据不足的问题。为此,最近的进展是采用手工操作或生成模型进行数据增强。手工操作是由人类活动的一些物理先验驱动的,例如,动作扭曲和强度波动。然而,这些方法在维护语义数据属性方面可能面临挑战。虽然生成模型具有较好的数据适应性,但难以将重要的动作先验纳入到数据生成中。本文提出了一种可微先验驱动的HAR数据增强框架。首先,我们将手工制作的增广操作嵌入到一个可微模块中,该模块自适应地选择和优化要组合在一起的操作。然后,我们构建生成模块,在手工操作导出的数据中加入可控扰动,进一步提高数据增强的多样性。通过将手工操作模块和生成模块集成到一个可学习的框架中,有效地提高了识别模型的泛化性能。在五个公共数据集上使用三种不同分类器的大量实验结果证明了所提出框架的有效性。项目页面:https://github.com/crocodilegogogo/DriveData-Under-Review。
{"title":"Differentiable Prior-Driven Data Augmentation for Sensor-Based Human Activity Recognition","authors":"Ye Zhang;Qing Gao;Rong Hu;Qingtang Ding;Boyang Li;Yulan Guo","doi":"10.1109/TCSS.2025.3565414","DOIUrl":"https://doi.org/10.1109/TCSS.2025.3565414","url":null,"abstract":"Sensor-based human activity recognition (HAR) usually suffers from the problem of insufficient annotated data, due to the difficulty in labeling the intuitive signals of wearable sensors. To this end, recent advances have adopted handcrafted operations or generative models for data augmentation. The handcrafted operations are driven by some physical priors of human activities, e.g., action distortion and strength fluctuations. However, these approaches may face challenges in maintaining semantic data properties. Although the generative models have better data adaptability, it is difficult for them to incorporate important action priors into data generation. This article proposes a differentiable prior-driven data augmentation framework for HAR. First, we embed the handcrafted augmentation operations into a differentiable module, which adaptively selects and optimizes the operations to be combined together. Then, we construct a generative module to add controllable perturbations to the data derived by the handcrafted operations and further improve the diversity of data augmentation. By integrating the handcrafted operation module and the generative module into one learnable framework, the generalization performance of the recognition models is enhanced effectively. Extensive experimental results with three different classifiers on five public datasets demonstrate the effectiveness of the proposed framework. Project page: <uri>https://github.com/crocodilegogogo/DriveData-Under-Review</uri>.","PeriodicalId":13044,"journal":{"name":"IEEE Transactions on Computational Social Systems","volume":"12 5","pages":"3778-3790"},"PeriodicalIF":4.5,"publicationDate":"2025-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145230069","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Assessing the Expressive Language Levels of Autistic Children in Home Intervention 自闭症儿童在家庭干预中的语言表达水平评估
IF 4.5 2区 计算机科学 Q1 COMPUTER SCIENCE, CYBERNETICS Pub Date : 2025-06-04 DOI: 10.1109/TCSS.2025.3563733
Yueran Pan;Biyuan Chen;Wenxing Liu;Ming Cheng;Dong Zhang;Hongzhu Deng;Xiaobing Zou;Ming Li
The World Health Organization (WHO) has established the caregiver skill training (CST) program, designed to empower families with children diagnosed with autism spectrum disorder the essential caregiving skills. The joint engagement rating inventory (JERI) protocol evaluates participants’ engagement levels within the CST initiative. Traditionally, rating the expressive language level and use (EXLA) item in JERI relies on retrospective video analysis conducted by qualified professionals, thus incurring substantial labor costs. This study introduces a multimodal behavioral signal-processing framework designed to analyze both child and caregiver behaviors automatically, thereby rating EXLA. Initially, raw audio and video signals are segmented into concise intervals via voice activity detection, speaker diarization and speaker age classification, serving the dual purpose of eliminating nonspeech content and tagging each segment with its respective speaker. Subsequently, we extract an array of audio-visual features, encompassing our proposed interpretable, hand-crafted textual features, end-to-end audio embeddings and end-to-end video embeddings. Finally, these features are fused at the feature level to train a linear regression model aimed at predicting the EXLA scores. Our framework has been evaluated on the largest in-the-wild database currently available under the CST program. Experimental results indicate that the proposed system achieves a Pearson correlation coefficient of 0.768 against the expert ratings, evidencing promising performance comparable to that of human experts.
世界卫生组织(世卫组织)制定了护理人员技能培训方案,旨在使有诊断为自闭症谱系障碍儿童的家庭掌握基本的护理技能。联合参与评级清单(JERI)协议评估CST计划中参与者的参与水平。传统上,对JERI中表达性语言水平和使用(EXLA)项目的评分依赖于有资质的专业人员进行的回顾性视频分析,从而产生了大量的人工成本。本研究引入了一个多模态行为信号处理框架,旨在自动分析儿童和照顾者的行为,从而对EXLA进行评级。首先,原始音频和视频信号通过语音活动检测、说话人dialarization和说话人年龄分类被分割成简洁的间隔,达到消除非语音内容和用各自的说话人标记每个片段的双重目的。随后,我们提取了一系列视听特征,包括我们提出的可解释的、手工制作的文本特征、端到端音频嵌入和端到端视频嵌入。最后,在特征级将这些特征融合,以训练一个旨在预测EXLA分数的线性回归模型。我们的框架已经在CST项目下最大的野外数据库上进行了评估。实验结果表明,该系统与专家评分的Pearson相关系数为0.768,具有与人类专家相当的性能。
{"title":"Assessing the Expressive Language Levels of Autistic Children in Home Intervention","authors":"Yueran Pan;Biyuan Chen;Wenxing Liu;Ming Cheng;Dong Zhang;Hongzhu Deng;Xiaobing Zou;Ming Li","doi":"10.1109/TCSS.2025.3563733","DOIUrl":"https://doi.org/10.1109/TCSS.2025.3563733","url":null,"abstract":"The World Health Organization (WHO) has established the caregiver skill training (CST) program, designed to empower families with children diagnosed with autism spectrum disorder the essential caregiving skills. The joint engagement rating inventory (JERI) protocol evaluates participants’ engagement levels within the CST initiative. Traditionally, rating the expressive language level and use (EXLA) item in JERI relies on retrospective video analysis conducted by qualified professionals, thus incurring substantial labor costs. This study introduces a multimodal behavioral signal-processing framework designed to analyze both child and caregiver behaviors automatically, thereby rating EXLA. Initially, raw audio and video signals are segmented into concise intervals via voice activity detection, speaker diarization and speaker age classification, serving the dual purpose of eliminating nonspeech content and tagging each segment with its respective speaker. Subsequently, we extract an array of audio-visual features, encompassing our proposed interpretable, hand-crafted textual features, end-to-end audio embeddings and end-to-end video embeddings. Finally, these features are fused at the feature level to train a linear regression model aimed at predicting the EXLA scores. Our framework has been evaluated on the largest in-the-wild database currently available under the CST program. Experimental results indicate that the proposed system achieves a Pearson correlation coefficient of 0.768 against the expert ratings, evidencing promising performance comparable to that of human experts.","PeriodicalId":13044,"journal":{"name":"IEEE Transactions on Computational Social Systems","volume":"12 5","pages":"3647-3659"},"PeriodicalIF":4.5,"publicationDate":"2025-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145230078","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Novel Chaotic Map and Its Application to Secure Transmission of Multimodal Images 一种新的混沌映射及其在多模态图像安全传输中的应用
IF 4.5 2区 计算机科学 Q1 COMPUTER SCIENCE, CYBERNETICS Pub Date : 2025-06-04 DOI: 10.1109/TCSS.2025.3568467
Parkala Vishnu Bharadwaj Bayari;Yashmita Sangwan;Gaurav Bhatnagar;Chiranjoy Chattopadhyay
The advent of digital technology, augmented by connected devices, has catalyzed a dramatic increase in multimedia content consumption, facilitating on-the-go access and communication. However, this surge also heightens the risks of unauthorized access, privacy breaches, and cyberattacks. Consequently, ensuring the secure and efficient transmission and storage of multimedia content is of paramount importance. This article presents a robust encryption scheme for secure image transmission, utilizing a novel one-dimensional chaotic map characterized by random and complex dynamics, validated through NIST test and meticulous evaluation. Key matrices are derived from the chaotic map, with the SHA-256 hash of random, nonoverlapping blocks of the input image influencing the initial conditions, thereby ensuring resistance to differential cryptanalysis. The encryption process encompasses a dual shuffling mechanism: an adaptive shuffling guided by the chaotic key, followed by orbital shuffling, which rearranges pixel positions by segmenting the image into distinct orbital patterns. This is complemented by a feedback diffusion technique that ensures each pixel’s encryption is influenced by neighboring values and the keys employed. Extensive evaluation with multimodal images demonstrates the scheme’s versatility, with significant resilience against various cryptographic attacks, as evidenced by thorough assessments. Comparative analysis further highlights the superiority of the proposed scheme over state-of-the-art approaches. These attributes position the proposed scheme as a highly effective solution for contemporary digital security challenges.
数字技术的出现,以及互联设备的增强,催化了多媒体内容消费的急剧增长,促进了移动访问和通信。然而,这种激增也增加了未经授权访问、隐私泄露和网络攻击的风险。因此,确保多媒体内容的安全和高效传输和存储是至关重要的。本文提出了一种用于安全图像传输的鲁棒加密方案,该方案利用了一种具有随机和复杂动态特征的新型一维混沌映射,并通过NIST测试和细致的评估进行了验证。密钥矩阵从混沌映射中导出,输入图像的随机非重叠块的SHA-256哈希影响初始条件,从而确保抗差分密码分析。加密过程包含双重洗牌机制:由混沌密钥引导的自适应洗牌,然后是轨道洗牌,通过将图像分割成不同的轨道模式来重新排列像素位置。这是一种反馈扩散技术的补充,该技术确保每个像素的加密都受到邻近值和所使用的密钥的影响。对多模态图像的广泛评估证明了该方案的多功能性,对各种加密攻击具有显著的弹性,这一点得到了全面评估的证明。对比分析进一步突出了所提出方案优于最先进方法的优越性。这些属性使所提出的方案成为当代数字安全挑战的高效解决方案。
{"title":"A Novel Chaotic Map and Its Application to Secure Transmission of Multimodal Images","authors":"Parkala Vishnu Bharadwaj Bayari;Yashmita Sangwan;Gaurav Bhatnagar;Chiranjoy Chattopadhyay","doi":"10.1109/TCSS.2025.3568467","DOIUrl":"https://doi.org/10.1109/TCSS.2025.3568467","url":null,"abstract":"The advent of digital technology, augmented by connected devices, has catalyzed a dramatic increase in multimedia content consumption, facilitating on-the-go access and communication. However, this surge also heightens the risks of unauthorized access, privacy breaches, and cyberattacks. Consequently, ensuring the secure and efficient transmission and storage of multimedia content is of paramount importance. This article presents a robust encryption scheme for secure image transmission, utilizing a novel one-dimensional chaotic map characterized by random and complex dynamics, validated through NIST test and meticulous evaluation. Key matrices are derived from the chaotic map, with the SHA-256 hash of random, nonoverlapping blocks of the input image influencing the initial conditions, thereby ensuring resistance to differential cryptanalysis. The encryption process encompasses a dual shuffling mechanism: an adaptive shuffling guided by the chaotic key, followed by orbital shuffling, which rearranges pixel positions by segmenting the image into distinct orbital patterns. This is complemented by a feedback diffusion technique that ensures each pixel’s encryption is influenced by neighboring values and the keys employed. Extensive evaluation with multimodal images demonstrates the scheme’s versatility, with significant resilience against various cryptographic attacks, as evidenced by thorough assessments. Comparative analysis further highlights the superiority of the proposed scheme over state-of-the-art approaches. These attributes position the proposed scheme as a highly effective solution for contemporary digital security challenges.","PeriodicalId":13044,"journal":{"name":"IEEE Transactions on Computational Social Systems","volume":"12 5","pages":"3765-3777"},"PeriodicalIF":4.5,"publicationDate":"2025-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145230015","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Literature Survey on Multimodal and Multilingual Sexism Detection 多模态和多语言性别歧视检测的文献综述
IF 4.5 2区 计算机科学 Q1 COMPUTER SCIENCE, CYBERNETICS Pub Date : 2025-06-03 DOI: 10.1109/TCSS.2025.3561921
Xuan Luo;Bin Liang;Qianlong Wang;Jing Li;Erik Cambria;Xiaojun Zhang;Yulan He;Min Yang;Ruifeng Xu
Sexism has become a pressing issue, driven by the rapid-spreading influence of societal norms, media portrayals, and online platforms that perpetuate and amplify gender biases. Curbing sexism has emerged as a critical challenge globally. Being capable of recognizing sexist statements and behaviors is of particular importance since it is the first step in mind change. This survey provides an extensive overview of recent advancements in sexism detection. We present details of the various resources used in this field and methodologies applied to the task, covering different languages, modalities, models, and approaches. Moreover, we examine the specific challenges these models encounter in accurately identifying and classifying sexism. Additionally, we highlight areas that require further research and propose potential new directions for future exploration in the domain of sexism detection. Through this comprehensive exploration, we strive to contribute to the advancement of interdisciplinary research, fostering a collective effort to combat sexism in its multifaceted manifestations.
性别歧视已经成为一个紧迫的问题,这是由于社会规范、媒体描述和在线平台的快速传播影响,这些影响使性别偏见永久化和放大。遏制性别歧视已成为全球面临的一项重大挑战。能够识别性别歧视的言论和行为是特别重要的,因为这是改变思想的第一步。这项调查提供了性别歧视检测的最新进展的广泛概述。我们详细介绍了该领域中使用的各种资源和应用于该任务的方法,涵盖了不同的语言、模式、模型和方法。此外,我们研究了这些模型在准确识别和分类性别歧视方面遇到的具体挑战。此外,我们强调了需要进一步研究的领域,并提出了性别歧视检测领域未来探索的潜在新方向。通过这种全面的探索,我们努力为跨学科研究的进步做出贡献,促进集体努力,以对抗多方面的性别歧视。
{"title":"A Literature Survey on Multimodal and Multilingual Sexism Detection","authors":"Xuan Luo;Bin Liang;Qianlong Wang;Jing Li;Erik Cambria;Xiaojun Zhang;Yulan He;Min Yang;Ruifeng Xu","doi":"10.1109/TCSS.2025.3561921","DOIUrl":"https://doi.org/10.1109/TCSS.2025.3561921","url":null,"abstract":"Sexism has become a pressing issue, driven by the rapid-spreading influence of societal norms, media portrayals, and online platforms that perpetuate and amplify gender biases. Curbing sexism has emerged as a critical challenge globally. Being capable of recognizing sexist statements and behaviors is of particular importance since it is the first step in mind change. This survey provides an extensive overview of recent advancements in sexism detection. We present details of the various resources used in this field and methodologies applied to the task, covering different languages, modalities, models, and approaches. Moreover, we examine the specific challenges these models encounter in accurately identifying and classifying sexism. Additionally, we highlight areas that require further research and propose potential new directions for future exploration in the domain of sexism detection. Through this comprehensive exploration, we strive to contribute to the advancement of interdisciplinary research, fostering a collective effort to combat sexism in its multifaceted manifestations.","PeriodicalId":13044,"journal":{"name":"IEEE Transactions on Computational Social Systems","volume":"12 5","pages":"3709-3727"},"PeriodicalIF":4.5,"publicationDate":"2025-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145230029","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive Social Metaverse Streaming Based on Federated Multiagent Deep Reinforcement Learning 基于联邦多智能体深度强化学习的自适应社会元宇宙流
IF 4.5 2区 计算机科学 Q1 COMPUTER SCIENCE, CYBERNETICS Pub Date : 2025-04-08 DOI: 10.1109/TCSS.2025.3555419
Zijian Long;Haopeng Wang;Haiwei Dong;Abdulmotaleb El Saddik
The social metaverse is a growing digital ecosystem that blends virtual and physical worlds. It allows users to interact socially, work, shop, and enjoy entertainment. However, privacy remains a major challenge, as immersive interactions require continuous collection of biometric and behavioral data. At the same time, ensuring high-quality, low-latency streaming is difficult due to the demands of real-time interaction, immersive rendering, and bandwidth optimization. To address these issues, we propose adaptive social metaverse streaming (ASMS), a novel streaming system based on federated multiagent proximal policy optimization (F-MAPPO). ASMS leverages F-MAPPO, which integrates federated learning (FL) and deep reinforcement learning (DRL) to dynamically adjust streaming bit rates while preserving user privacy. Experimental results show that ASMS improves user experience by at least 14% compared to existing streaming methods across various network conditions. Therefore, ASMS enhances the social metaverse experience by providing seamless and immersive streaming, even in dynamic and resource-constrained networks, while ensuring that sensitive user data remain on local devices.
社交虚拟世界是一个不断发展的数字生态系统,它融合了虚拟世界和现实世界。它允许用户进行社交互动、工作、购物和享受娱乐。然而,隐私仍然是一个主要的挑战,因为沉浸式交互需要不断收集生物特征和行为数据。同时,由于对实时交互、沉浸式渲染和带宽优化的要求,难以保证高质量、低延迟的流媒体。为了解决这些问题,我们提出了一种基于联邦多智能体近端策略优化(F-MAPPO)的自适应社会元数据流(ASMS)。asm利用F-MAPPO,它集成了联邦学习(FL)和深度强化学习(DRL),在保护用户隐私的同时动态调整流比特率。实验结果表明,在各种网络条件下,与现有的流媒体方法相比,ASMS至少提高了14%的用户体验。因此,即使在动态和资源受限的网络中,asm也能提供无缝的沉浸式流媒体,同时确保敏感的用户数据保留在本地设备上,从而增强社交虚拟世界体验。
{"title":"Adaptive Social Metaverse Streaming Based on Federated Multiagent Deep Reinforcement Learning","authors":"Zijian Long;Haopeng Wang;Haiwei Dong;Abdulmotaleb El Saddik","doi":"10.1109/TCSS.2025.3555419","DOIUrl":"https://doi.org/10.1109/TCSS.2025.3555419","url":null,"abstract":"The social metaverse is a growing digital ecosystem that blends virtual and physical worlds. It allows users to interact socially, work, shop, and enjoy entertainment. However, privacy remains a major challenge, as immersive interactions require continuous collection of biometric and behavioral data. At the same time, ensuring high-quality, low-latency streaming is difficult due to the demands of real-time interaction, immersive rendering, and bandwidth optimization. To address these issues, we propose adaptive social metaverse streaming (ASMS), a novel streaming system based on federated multiagent proximal policy optimization (F-MAPPO). ASMS leverages F-MAPPO, which integrates federated learning (FL) and deep reinforcement learning (DRL) to dynamically adjust streaming bit rates while preserving user privacy. Experimental results show that ASMS improves user experience by at least 14% compared to existing streaming methods across various network conditions. Therefore, ASMS enhances the social metaverse experience by providing seamless and immersive streaming, even in dynamic and resource-constrained networks, while ensuring that sensitive user data remain on local devices.","PeriodicalId":13044,"journal":{"name":"IEEE Transactions on Computational Social Systems","volume":"12 5","pages":"3804-3815"},"PeriodicalIF":4.5,"publicationDate":"2025-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145230077","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Explaining Sentiments: Improving Explainability in Sentiment Analysis Using Local Interpretable Model-Agnostic Explanations and Counterfactual Explanations 解释情绪:利用局部可解释模型-不可知论解释和反事实解释提高情绪分析的可解释性
IF 4.5 2区 计算机科学 Q1 COMPUTER SCIENCE, CYBERNETICS Pub Date : 2025-04-08 DOI: 10.1109/TCSS.2025.3531718
Xin Wang;Jianhui Lyu;J. Dinesh Peter;Byung-Gyu Kim;B.D. Parameshachari;Keqin Li;Wei Wei
Sentiment analysis of social media platforms is crucial for extracting actionable insights from unstructured textual data. However, modern sentiment analysis models using deep learning lack explainability, acting as black box and limiting trust. This study focuses on improving the explainability of sentiment analysis models of social media platforms by leveraging explainable artificial intelligence (XAI). We propose a novel explainable sentiment analysis (XSA) framework incorporating intrinsic and posthoc XAI methods, i.e., local interpretable model-agnostic explanations (LIME) and counterfactual explanations. Specifically, to solve the problem of lack of local fidelity and stability in interpretations caused by the LIME random perturbation sampling method, a new model-independent interpretation method is proposed, which uses the isometric mapping virtual sample generation method based on manifold learning instead of LIMEs random perturbation sampling method to generate samples. Additionally, a generative link tree is presented to create counterfactual explanations that maintain strong data fidelity, which constructs counterfactual narratives by leveraging examples from the training data, employing a divide-and-conquer strategy combined with local greedy. Experiments conducted on social media datasets from Twitter, YouTube comments, Yelp, and Amazon demonstrate XSAs ability to provide local aspect-level explanations while maintaining sentiment analysis performance. Analyses reveal improved model explainability and enhanced user trust, demonstrating XAIs potential in sentiment analysis of social media platforms. The proposed XSA framework provides a valuable direction for developing transparent and trustworthy sentiment analysis models for social media platforms.
社交媒体平台的情感分析对于从非结构化文本数据中提取可操作的见解至关重要。然而,使用深度学习的现代情感分析模型缺乏可解释性,充当黑箱,限制信任。本研究的重点是利用可解释人工智能(XAI)来提高社交媒体平台情感分析模型的可解释性。我们提出了一种新的可解释情感分析(XSA)框架,该框架结合了内在和后置XAI方法,即局部可解释模型不可知论解释(LIME)和反事实解释。具体而言,针对LIME随机摄动采样方法在解译中缺乏局部保真度和稳定性的问题,提出了一种新的模型无关解译方法,采用基于流形学习的等距映射虚拟样本生成方法代替LIME随机摄动采样方法生成样本。此外,提出了一个生成链接树来创建反事实解释,以保持强大的数据保真度,它通过利用训练数据中的示例构建反事实叙述,采用分而治之的策略与局部贪婪相结合。在Twitter、YouTube评论、Yelp和Amazon等社交媒体数据集上进行的实验表明,xsa能够在保持情感分析性能的同时提供本地方面级解释。分析表明,模型的可解释性得到改善,用户信任得到增强,证明了xai在社交媒体平台情感分析中的潜力。提出的XSA框架为开发透明可信的社交媒体平台情感分析模型提供了有价值的方向。
{"title":"Explaining Sentiments: Improving Explainability in Sentiment Analysis Using Local Interpretable Model-Agnostic Explanations and Counterfactual Explanations","authors":"Xin Wang;Jianhui Lyu;J. Dinesh Peter;Byung-Gyu Kim;B.D. Parameshachari;Keqin Li;Wei Wei","doi":"10.1109/TCSS.2025.3531718","DOIUrl":"https://doi.org/10.1109/TCSS.2025.3531718","url":null,"abstract":"Sentiment analysis of social media platforms is crucial for extracting actionable insights from unstructured textual data. However, modern sentiment analysis models using deep learning lack explainability, acting as black box and limiting trust. This study focuses on improving the explainability of sentiment analysis models of social media platforms by leveraging explainable artificial intelligence (XAI). We propose a novel explainable sentiment analysis (XSA) framework incorporating intrinsic and posthoc XAI methods, i.e., local interpretable model-agnostic explanations (LIME) and counterfactual explanations. Specifically, to solve the problem of lack of local fidelity and stability in interpretations caused by the LIME random perturbation sampling method, a new model-independent interpretation method is proposed, which uses the isometric mapping virtual sample generation method based on manifold learning instead of LIMEs random perturbation sampling method to generate samples. Additionally, a generative link tree is presented to create counterfactual explanations that maintain strong data fidelity, which constructs counterfactual narratives by leveraging examples from the training data, employing a divide-and-conquer strategy combined with local greedy. Experiments conducted on social media datasets from Twitter, YouTube comments, Yelp, and Amazon demonstrate XSAs ability to provide local aspect-level explanations while maintaining sentiment analysis performance. Analyses reveal improved model explainability and enhanced user trust, demonstrating XAIs potential in sentiment analysis of social media platforms. The proposed XSA framework provides a valuable direction for developing transparent and trustworthy sentiment analysis models for social media platforms.","PeriodicalId":13044,"journal":{"name":"IEEE Transactions on Computational Social Systems","volume":"12 3","pages":"1390-1403"},"PeriodicalIF":4.5,"publicationDate":"2025-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144186005","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Guest Editorial: Special Issue on Music Intelligence and Social Computation 特刊:音乐智能与社会计算
IF 4.5 2区 计算机科学 Q1 COMPUTER SCIENCE, CYBERNETICS Pub Date : 2025-04-04 DOI: 10.1109/TCSS.2025.3548862
Xiaohong Guan;Xiaobing Li;Björn W. Schuller;Xinran Zhang
{"title":"Guest Editorial: Special Issue on Music Intelligence and Social Computation","authors":"Xiaohong Guan;Xiaobing Li;Björn W. Schuller;Xinran Zhang","doi":"10.1109/TCSS.2025.3548862","DOIUrl":"https://doi.org/10.1109/TCSS.2025.3548862","url":null,"abstract":"","PeriodicalId":13044,"journal":{"name":"IEEE Transactions on Computational Social Systems","volume":"12 2","pages":"847-850"},"PeriodicalIF":4.5,"publicationDate":"2025-04-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10949083","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143777858","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Computational Social Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1