首页 > 最新文献

Scalable Computing-Practice and Experience最新文献

英文 中文
ICFD: An Incremental Learning Method Based on Data Feature Distribution 一种基于数据特征分布的增量学习方法
IF 1.1 Q2 Computer Science Pub Date : 2022-12-01 DOI: 10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00103
Yunzhe Zhu, Yusong Tan, Xiaoling Li, Qingbo Wu, Xueqin Ning
Neural network models have achieved great success in numerous disciplines in recent years, including image segmentation, object identification, and natural language processing (NLP). Incremental learning in these fields focuses on training models in a continuous data stream. As time goes by, more new data becomes available, and old data may become unavailable owing to resource constraints such as storage. As a result, when new data is continually arriving, the performance of the neural network model on the old data sample sometimes decreases significantly, a phenomenon known as catastrophic forgetting. Many corresponding strategies have been proposed to mitigate the catastrophic forgetting of neural network models, which are based on parameter regularization, data replay, and parameter isolation. This paper proposes an incremental learning method based on data feature distribution (ICFD). The method uses Gaussian distribution to generate features from old data to train neural network models based on the phenomenon that feature vectors obey multi-dimensional Gaussian distribution in feature space. This method avoids storing a large number of original samples, and the generated old class features contain more sample information. This method combines data playback and parameter regularization in concrete implementation. The experimental results of ICFD on the CIFAR-100 demonstrate that when the incremental step is 5, the average incremental accuracy is increased by 10.4%. When the incremental step is 10, the average incremental accuracy is improved by 8.1%.
近年来,神经网络模型在图像分割、目标识别和自然语言处理(NLP)等众多领域取得了巨大的成功。这些领域的增量学习侧重于在连续数据流中训练模型。随着时间的推移,越来越多的新数据变得可用,而旧数据可能由于存储等资源限制而不可用。因此,当新数据不断到来时,神经网络模型在旧数据样本上的表现有时会显著下降,这种现象被称为灾难性遗忘。为了减轻神经网络模型的灾难性遗忘,人们提出了许多相应的策略,包括参数正则化、数据重放和参数隔离。提出了一种基于数据特征分布(ICFD)的增量学习方法。该方法利用特征向量在特征空间服从多维高斯分布的现象,利用高斯分布从旧数据中生成特征来训练神经网络模型。这种方法避免了存储大量的原始样本,并且生成的旧类特征包含更多的样本信息。该方法在具体实现中结合了数据回放和参数正则化。ICFD在CIFAR-100上的实验结果表明,当增量步长为5时,平均增量精度提高了10.4%。当增量步长为10时,平均增量精度提高8.1%。
{"title":"ICFD: An Incremental Learning Method Based on Data Feature Distribution","authors":"Yunzhe Zhu, Yusong Tan, Xiaoling Li, Qingbo Wu, Xueqin Ning","doi":"10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00103","DOIUrl":"https://doi.org/10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00103","url":null,"abstract":"Neural network models have achieved great success in numerous disciplines in recent years, including image segmentation, object identification, and natural language processing (NLP). Incremental learning in these fields focuses on training models in a continuous data stream. As time goes by, more new data becomes available, and old data may become unavailable owing to resource constraints such as storage. As a result, when new data is continually arriving, the performance of the neural network model on the old data sample sometimes decreases significantly, a phenomenon known as catastrophic forgetting. Many corresponding strategies have been proposed to mitigate the catastrophic forgetting of neural network models, which are based on parameter regularization, data replay, and parameter isolation. This paper proposes an incremental learning method based on data feature distribution (ICFD). The method uses Gaussian distribution to generate features from old data to train neural network models based on the phenomenon that feature vectors obey multi-dimensional Gaussian distribution in feature space. This method avoids storing a large number of original samples, and the generated old class features contain more sample information. This method combines data playback and parameter regularization in concrete implementation. The experimental results of ICFD on the CIFAR-100 demonstrate that when the incremental step is 5, the average incremental accuracy is increased by 10.4%. When the incremental step is 10, the average incremental accuracy is improved by 8.1%.","PeriodicalId":43791,"journal":{"name":"Scalable Computing-Practice and Experience","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85817338","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MlpE: Knowledge Graph Embedding with Multilayer Perceptron Networks MlpE:基于多层感知机网络的知识图嵌入
IF 1.1 Q2 Computer Science Pub Date : 2022-12-01 DOI: 10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00130
Qing Xu, Kaijun Ren, Xiaoli Ren, Shuibing Long, Xiaoyong Li
Knowledge graph embedding (KGE) is an efficient method to predict missing links in knowledge graphs. Most KGE models based on convolutional neural networks have been designed for improving the ability of capturing interaction. Although these models work well, they suffered from the limited receptive field of the convolution kernel, which lead to the lack of ability to capture long-distance interactions. In this paper, we firstly illustrate the interactions between entities and relations and discuss its effect in KGE models by experiments, and then propose MlpE, which is a fully connected network with only three layers. MlpE aims to capture long-distance interactions to improve the performance of link prediction. Extensive experimental evaluations on four typical datasets WN18RR, FB15k-237, DB100k and YAGO3-10 have shown the superority of MlpE, especially in some cases MlpE can achieve the better performance with less parameters than the state-of-the-art convolution-based KGE model.
知识图嵌入(KGE)是一种预测知识图中缺失环节的有效方法。大多数基于卷积神经网络的KGE模型都是为了提高捕获交互的能力而设计的。虽然这些模型工作得很好,但它们受到卷积核的有限接受域的影响,这导致它们缺乏捕捉远距离相互作用的能力。本文首先通过实验阐述了实体和关系之间的相互作用,并讨论了其在KGE模型中的作用,然后提出了只有三层的全连接网络MlpE。MlpE旨在捕获远距离交互以提高链路预测的性能。在WN18RR、FB15k-237、DB100k和YAGO3-10四个典型数据集上进行的大量实验评估显示了MlpE的优越性,特别是在某些情况下,MlpE可以以更少的参数达到比最先进的基于卷积的KGE模型更好的性能。
{"title":"MlpE: Knowledge Graph Embedding with Multilayer Perceptron Networks","authors":"Qing Xu, Kaijun Ren, Xiaoli Ren, Shuibing Long, Xiaoyong Li","doi":"10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00130","DOIUrl":"https://doi.org/10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00130","url":null,"abstract":"Knowledge graph embedding (KGE) is an efficient method to predict missing links in knowledge graphs. Most KGE models based on convolutional neural networks have been designed for improving the ability of capturing interaction. Although these models work well, they suffered from the limited receptive field of the convolution kernel, which lead to the lack of ability to capture long-distance interactions. In this paper, we firstly illustrate the interactions between entities and relations and discuss its effect in KGE models by experiments, and then propose MlpE, which is a fully connected network with only three layers. MlpE aims to capture long-distance interactions to improve the performance of link prediction. Extensive experimental evaluations on four typical datasets WN18RR, FB15k-237, DB100k and YAGO3-10 have shown the superority of MlpE, especially in some cases MlpE can achieve the better performance with less parameters than the state-of-the-art convolution-based KGE model.","PeriodicalId":43791,"journal":{"name":"Scalable Computing-Practice and Experience","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84089415","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Attack-Model-Agnostic Defense Against Model Poisonings in Distributed Learning 分布式学习中模型中毒的攻击-模型不可知防御
IF 1.1 Q2 Computer Science Pub Date : 2022-12-01 DOI: 10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00354
Hairuo Xu, Tao Shu
The distributed nature of distributed learning renders the learning process susceptible to model poisoning attacks. Most existing countermeasures are designed based on a presumed attack model, and can only perform under the presumed attack model. However, in reality a distributed learning system typically does not have the luxury of knowing the attack model it is going to be actually facing in its operation when the learning system is deployed, thus constituting a zero-day vulnerability of the system that has been largely overlooked so far. In this paper, we study the attack-model-agnostic defense mechanisms for distributed learning, which are capable of countering a wide-spectrum of model poisoning attacks without relying on assumptions of the specific attack model, and hence alleviating the zero-day vulnerability of the system. Extensive experiments are performed to verify the effectiveness of the proposed defense.
分布式学习的分布式特性使得学习过程容易受到模型中毒攻击。现有的大多数对抗措施都是基于假定的攻击模型设计的,并且只能在假定的攻击模型下执行。然而,在现实中,分布式学习系统在部署学习系统时,通常无法知道它在运行中实际面临的攻击模型,因此构成了系统的零日漏洞,到目前为止,这在很大程度上被忽视了。在本文中,我们研究了分布式学习的攻击模型无关防御机制,该机制能够在不依赖于特定攻击模型假设的情况下对抗广泛的模型中毒攻击,从而减轻系统的零日漏洞。进行了大量的实验来验证所提出的防御的有效性。
{"title":"Attack-Model-Agnostic Defense Against Model Poisonings in Distributed Learning","authors":"Hairuo Xu, Tao Shu","doi":"10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00354","DOIUrl":"https://doi.org/10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00354","url":null,"abstract":"The distributed nature of distributed learning renders the learning process susceptible to model poisoning attacks. Most existing countermeasures are designed based on a presumed attack model, and can only perform under the presumed attack model. However, in reality a distributed learning system typically does not have the luxury of knowing the attack model it is going to be actually facing in its operation when the learning system is deployed, thus constituting a zero-day vulnerability of the system that has been largely overlooked so far. In this paper, we study the attack-model-agnostic defense mechanisms for distributed learning, which are capable of countering a wide-spectrum of model poisoning attacks without relying on assumptions of the specific attack model, and hence alleviating the zero-day vulnerability of the system. Extensive experiments are performed to verify the effectiveness of the proposed defense.","PeriodicalId":43791,"journal":{"name":"Scalable Computing-Practice and Experience","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82139024","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Intelligent Scoring Method for Sketch Portrait Based on Attention Convolution Neural Network 一种基于注意卷积神经网络的素描肖像智能评分方法
IF 1.1 Q2 Computer Science Pub Date : 2022-12-01 DOI: 10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00156
Shaolong Zheng, Zewei Xu, Zhenni Li, Yihui Cai, Mingyu Han, Yi Ji
It is very important for art students to get timely feedback on their paintings. Currently, this work is done by professional teachers. However, it is problematic for the scoring method since the subjectivity of manual scoring and the scarcity of teacher resources. It is time-consuming and expensive to carry out this work in practice. In this paper, we propose a depthwise separable convolutional network with multi-head self-attention module (DCMnet) for developing an intelligent scoring mechanism for sketch portraits. Specifically, to build a lightweight network, we first utilize the depthwise separable convolutional block as the backbone of the model for mining the local features of sketch portraits. Then the attention module is employed to notice global dependencies within internal representations of portraits. Finally, we use DCMnet to build a scoring framework, which first divides the works into four score levels, and then subdivides them into eight grades: below 60, 60-64, 65-69, 70-74, 75-79, 80-84, 85-89, and above 90. Each grade of work is given a basic score, and the final score of works is composed of the basic score and the mood factor. In the training process, a pretraining strategy is introduced for fast convergence. For verifying our method, we collect a sketch portrait dataset in the Guangdong Fine Arts Joint Examination to train the DCMnet. The experimental results demonstrate that the proposed method achieves excellent accuracy at each grade and the efficiency of scoring is improved.
对艺术专业的学生来说,得到及时的绘画反馈是非常重要的。目前,这项工作是由专业教师完成的。然而,由于人工评分的主观性和教师资源的稀缺,这种评分方法存在问题。在实践中进行这项工作既费时又昂贵。在本文中,我们提出了一种带有多头自注意模块的深度可分离卷积网络(DCMnet),用于开发素描肖像的智能评分机制。具体来说,为了构建轻量级网络,我们首先利用深度可分卷积块作为模型的主干来挖掘素描肖像的局部特征。然后使用注意力模块来注意肖像内部表示中的全局依赖关系。最后,我们使用DCMnet构建评分框架,首先将作品分为4个评分等级,再细分为60分以下、60-64分、65-69分、70-74分、75-79分、80-84分、85-89分、90分以上8个等级。每个等级的作品都有一个基本分数,作品的最终分数由基本分数和情绪因素组成。在训练过程中,引入了一种快速收敛的预训练策略。为了验证我们的方法,我们在广东美术联考中收集了一个素描肖像数据集来训练DCMnet。实验结果表明,该方法在每个等级上都达到了很好的准确率,提高了评分效率。
{"title":"An Intelligent Scoring Method for Sketch Portrait Based on Attention Convolution Neural Network","authors":"Shaolong Zheng, Zewei Xu, Zhenni Li, Yihui Cai, Mingyu Han, Yi Ji","doi":"10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00156","DOIUrl":"https://doi.org/10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00156","url":null,"abstract":"It is very important for art students to get timely feedback on their paintings. Currently, this work is done by professional teachers. However, it is problematic for the scoring method since the subjectivity of manual scoring and the scarcity of teacher resources. It is time-consuming and expensive to carry out this work in practice. In this paper, we propose a depthwise separable convolutional network with multi-head self-attention module (DCMnet) for developing an intelligent scoring mechanism for sketch portraits. Specifically, to build a lightweight network, we first utilize the depthwise separable convolutional block as the backbone of the model for mining the local features of sketch portraits. Then the attention module is employed to notice global dependencies within internal representations of portraits. Finally, we use DCMnet to build a scoring framework, which first divides the works into four score levels, and then subdivides them into eight grades: below 60, 60-64, 65-69, 70-74, 75-79, 80-84, 85-89, and above 90. Each grade of work is given a basic score, and the final score of works is composed of the basic score and the mood factor. In the training process, a pretraining strategy is introduced for fast convergence. For verifying our method, we collect a sketch portrait dataset in the Guangdong Fine Arts Joint Examination to train the DCMnet. The experimental results demonstrate that the proposed method achieves excellent accuracy at each grade and the efficiency of scoring is improved.","PeriodicalId":43791,"journal":{"name":"Scalable Computing-Practice and Experience","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82515561","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust Spatio-Temporal Trajectory Modeling Based on Auto-Gated Recurrent Unit 基于自控循环单元的鲁棒时空轨迹建模
IF 1.1 Q2 Computer Science Pub Date : 2022-12-01 DOI: 10.1109/smartworld-uic-atc-scalcom-digitaltwin-pricomp-metaverse56740.2022.00176
Jia Jia, Xiaoyong Li, Ximing Li, Linghui Li, Jie Yuan, Hongmiao Wang, Yali Gao, Pengfei Qiu, Jialu Tang
With the huge amount of crowd mobility data generated by the explosion of mobile devices, deep neural networks (DNNs) are applied to trajectory data mining and modeling, which make great progresses in those scenarios. However, recent studies have demonstrated that DNNs are highly vulnerable to adversarial examples which are crafted by adding subtle, imperceptible noise to normal examples, and leading to the wrong prediction with high confidence. To improve the robustness of modeling spatiotemporal trajectories via DNNs, we propose a collaborative learning model named “Auto-GRU”, which consists of an autoencoder-based self-representation network (SRN) for robust trajectory feature learning and gated recurrent unit (GRU)-based classification network which shares information with SRN for collaborative learning and strictly defending adversarial examples. Our proposed method performs well in defending both white and black box attacks, especially in black-box attacks, where the performance outperforms state-of-the-art methods. Moreover, extensive experiments on Geolife and Beijing taxi traces datasets demonstrate that the proposed model can improve the robustness against adversarial examples without a significant performance penalty on clean examples.
随着移动设备爆炸式增长所产生的海量人群移动数据,将深度神经网络(deep neural networks, dnn)应用于轨迹数据挖掘和建模,在这些场景中取得了很大进展。然而,最近的研究表明,dnn非常容易受到对抗性示例的影响,这些示例是通过在正常示例中添加微妙的,难以察觉的噪声来制作的,并导致高置信度的错误预测。为了提高基于深度神经网络的时空轨迹建模的鲁棒性,我们提出了一种名为“Auto-GRU”的协同学习模型,该模型由基于自编码器的自表示网络(SRN)和基于门控循环单元(GRU)的分类网络组成,该网络与SRN共享信息进行协同学习并严格防御对抗性示例。我们提出的方法在防御白盒攻击和黑盒攻击方面都表现良好,特别是在黑盒攻击方面,性能优于最先进的方法。此外,在Geolife和北京出租车轨迹数据集上的大量实验表明,该模型可以提高对对抗样本的鲁棒性,而不会对干净样本造成明显的性能损失。
{"title":"Robust Spatio-Temporal Trajectory Modeling Based on Auto-Gated Recurrent Unit","authors":"Jia Jia, Xiaoyong Li, Ximing Li, Linghui Li, Jie Yuan, Hongmiao Wang, Yali Gao, Pengfei Qiu, Jialu Tang","doi":"10.1109/smartworld-uic-atc-scalcom-digitaltwin-pricomp-metaverse56740.2022.00176","DOIUrl":"https://doi.org/10.1109/smartworld-uic-atc-scalcom-digitaltwin-pricomp-metaverse56740.2022.00176","url":null,"abstract":"With the huge amount of crowd mobility data generated by the explosion of mobile devices, deep neural networks (DNNs) are applied to trajectory data mining and modeling, which make great progresses in those scenarios. However, recent studies have demonstrated that DNNs are highly vulnerable to adversarial examples which are crafted by adding subtle, imperceptible noise to normal examples, and leading to the wrong prediction with high confidence. To improve the robustness of modeling spatiotemporal trajectories via DNNs, we propose a collaborative learning model named “Auto-GRU”, which consists of an autoencoder-based self-representation network (SRN) for robust trajectory feature learning and gated recurrent unit (GRU)-based classification network which shares information with SRN for collaborative learning and strictly defending adversarial examples. Our proposed method performs well in defending both white and black box attacks, especially in black-box attacks, where the performance outperforms state-of-the-art methods. Moreover, extensive experiments on Geolife and Beijing taxi traces datasets demonstrate that the proposed model can improve the robustness against adversarial examples without a significant performance penalty on clean examples.","PeriodicalId":43791,"journal":{"name":"Scalable Computing-Practice and Experience","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82587723","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TGNRec: Recommendation Based on Trust Networks and Graph Neural Networks TGNRec:基于信任网络和图神经网络的推荐
IF 1.1 Q2 Computer Science Pub Date : 2022-12-01 DOI: 10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00274
Ting Li, Chundong Wang, Huai-bin Wang
In recent years, user-user trust relationships have played an important role in recommendation based on graph neural networks(GNNs). However, existing studies based on GNNs still face the following challenges: how to obtain more rating information of users’ trust from trust networks when using GNNs to learn the user latent feature. And how to effectively mine items’ relationships from the recommended data so that GNNs can better learn the item latent feature. To address the above challenges, in this paper, we propose a new model called TGNRec that accomplishes recommendation based on trust networks and graph neural networks. TGNRec consists of three modules: User Spatial Module, Item Spatial Module, Prediction Module. User Spatial Module considers both the rating information of users’ direct and indirect trust based on the transfer properties of trust relationships in trust networks. It mainly learns the user latent feature using user-item interactions and user-user trust relationships. Item Spatial Module establishes items’ similarity relationships based on the rating mean, which helps GNNs learn the item latent feature from user-item interactions and item-item relationships. Prediction Module realizes users’ rating prediction for unrated items by aggregating User Spatial Module and Item Spatial Module. At last, we conduct experiments on two real-world datasets, Film Trust and Ciao-DVD. The experimental results demonstrate the effectiveness of TGNRec for rating prediction in recommendation.
近年来,用户-用户信任关系在基于图神经网络(gnn)的推荐中发挥了重要作用。然而,基于gnn的现有研究仍然面临着以下挑战:在使用gnn学习用户潜在特征时,如何从信任网络中获取更多的用户信任评级信息。如何从推荐的数据中有效地挖掘项目之间的关系,使gnn能够更好地学习项目的潜在特征。为了解决上述挑战,本文提出了一种名为TGNRec的新模型,该模型基于信任网络和图神经网络来完成推荐。TGNRec由三个模块组成:用户空间模块、项目空间模块、预测模块。用户空间模块基于信任网络中信任关系的传递特性,考虑了用户直接信任和间接信任的评级信息。它主要利用用户-物品交互和用户-用户信任关系来学习用户潜在特征。物品空间模块基于评分均值建立物品相似关系,帮助gnn从用户-物品交互和物品-物品关系中学习物品潜在特征。预测模块通过聚合用户空间模块和物品空间模块实现用户对未评级物品的评级预测。最后,我们在Film Trust和Ciao-DVD两个真实数据集上进行了实验。实验结果证明了TGNRec算法在推荐评价预测中的有效性。
{"title":"TGNRec: Recommendation Based on Trust Networks and Graph Neural Networks","authors":"Ting Li, Chundong Wang, Huai-bin Wang","doi":"10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00274","DOIUrl":"https://doi.org/10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00274","url":null,"abstract":"In recent years, user-user trust relationships have played an important role in recommendation based on graph neural networks(GNNs). However, existing studies based on GNNs still face the following challenges: how to obtain more rating information of users’ trust from trust networks when using GNNs to learn the user latent feature. And how to effectively mine items’ relationships from the recommended data so that GNNs can better learn the item latent feature. To address the above challenges, in this paper, we propose a new model called TGNRec that accomplishes recommendation based on trust networks and graph neural networks. TGNRec consists of three modules: User Spatial Module, Item Spatial Module, Prediction Module. User Spatial Module considers both the rating information of users’ direct and indirect trust based on the transfer properties of trust relationships in trust networks. It mainly learns the user latent feature using user-item interactions and user-user trust relationships. Item Spatial Module establishes items’ similarity relationships based on the rating mean, which helps GNNs learn the item latent feature from user-item interactions and item-item relationships. Prediction Module realizes users’ rating prediction for unrated items by aggregating User Spatial Module and Item Spatial Module. At last, we conduct experiments on two real-world datasets, Film Trust and Ciao-DVD. The experimental results demonstrate the effectiveness of TGNRec for rating prediction in recommendation.","PeriodicalId":43791,"journal":{"name":"Scalable Computing-Practice and Experience","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78854807","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Aspect-Level Sentiment Classification Based on Self-Attention Routing via Capsule Network 基于胶囊网络自注意路由的方面级情感分类
IF 1.1 Q2 Computer Science Pub Date : 2022-12-01 DOI: 10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00280
Chang Liu, Jianxia Chen, Tianci Wang, Qi Liu, Xinyun Wu, Lei Mao
Aspect-level sentiment classification task aims at determining the sentiment polarity towards each aspect in a sentence. Although existing models have achieved remarkable performance, they always ignore the semantic relationship between aspects and their context, resulting in the lack of syntax information and aspect features. Therefore, the paper proposes a novel model named ASC based on the Self-Attention routing combined with the Position-biased weight approach, ASC-SAP in short. First, the paper utilizes the position-biased weight approach to construct an aspect-enhanced embedding. Furthermore, the paper develops a novel non-iterative but highly parallelized self-attention routing mechanism to efficiently transfer the aspect features to the target capsules. In addition, the paper utilizes pre-trained model bidirectional encoder representation from transformers (BERT). Comprehensive experiments show that our model achieves excellent performance on Twitter and SemEval2014 benchmarks and verify the effectiveness of our models.
方面级情感分类任务旨在确定句子中每个方面的情感极性。虽然现有的模型已经取得了显著的成绩,但它们往往忽略了方面及其上下文之间的语义关系,导致缺乏语法信息和方面特征。因此,本文提出了一种基于自注意路由与位置偏权方法相结合的新模型ASC,简称ASC- sap。首先,本文利用位置偏权方法构建了一个方面增强的嵌入。在此基础上,提出了一种新的非迭代但高度并行的自关注路由机制,以有效地将方面特征传递给目标胶囊。此外,本文利用预训练模型双向编码器表示从变压器(BERT)。综合实验表明,我们的模型在Twitter和SemEval2014基准上取得了优异的性能,验证了我们模型的有效性。
{"title":"Aspect-Level Sentiment Classification Based on Self-Attention Routing via Capsule Network","authors":"Chang Liu, Jianxia Chen, Tianci Wang, Qi Liu, Xinyun Wu, Lei Mao","doi":"10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00280","DOIUrl":"https://doi.org/10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00280","url":null,"abstract":"Aspect-level sentiment classification task aims at determining the sentiment polarity towards each aspect in a sentence. Although existing models have achieved remarkable performance, they always ignore the semantic relationship between aspects and their context, resulting in the lack of syntax information and aspect features. Therefore, the paper proposes a novel model named ASC based on the Self-Attention routing combined with the Position-biased weight approach, ASC-SAP in short. First, the paper utilizes the position-biased weight approach to construct an aspect-enhanced embedding. Furthermore, the paper develops a novel non-iterative but highly parallelized self-attention routing mechanism to efficiently transfer the aspect features to the target capsules. In addition, the paper utilizes pre-trained model bidirectional encoder representation from transformers (BERT). Comprehensive experiments show that our model achieves excellent performance on Twitter and SemEval2014 benchmarks and verify the effectiveness of our models.","PeriodicalId":43791,"journal":{"name":"Scalable Computing-Practice and Experience","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87015073","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
APR-ES: Adaptive Penalty-Reward Based Evolution Strategy for Deep Reinforcement Learning 深度强化学习的自适应奖惩进化策略
IF 1.1 Q2 Computer Science Pub Date : 2022-12-01 DOI: 10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00079
Dongdong Wang, Siyang Lu, Xiang Wei, Mingquan Wang, Yandong Li, Liqiang Wang
As a black-box optimization approach, derivative-free evolution strategy (ES) draws lots of attention in virtue of its low sensitivity and high scalability. It rivals Markov Decision Process based reinforcement learning or even can more efficiently improve rewards under complex scenarios. However, existing derivative-free ES still confronts slow convergence speed at the early training stage and limited exploration at the late convergence stage. Inspired from human learning process, we propose a new scheme extended from ES by taking advantage of prior knowledge to guide ES, thus accelerating early exploitation process and improving later exploration ability. At early training stage, Drift-Plus-Penalty (DPP), a penalty-based optimization scheme, is reformulated to boost penalty learning and reduce regrets. Along with DPP-directed evolution, reward learning with Thompson sampling (TS) is increasingly enhanced to explore global optima at late training stage. This scheme is justified with extensive experiments from a variety of benchmarks, including numerical problems, physics environments, and games. By virtue of its imitation of human learning process, this scheme outperforms state-of-the-art ES on the benchmarks by a large margin.
无导数进化策略作为一种黑盒优化方法,以其低灵敏度和高可扩展性而备受关注。它可以与基于马尔可夫决策过程的强化学习相媲美,甚至可以更有效地提高复杂场景下的奖励。然而,现有的无导数ES在训练初期仍然存在收敛速度慢、收敛后期探索有限的问题。受人类学习过程的启发,我们提出了一种从ES扩展而来的新方案,利用先验知识来指导ES,从而加快早期开发过程,提高后期的探索能力。在早期训练阶段,重新制定了基于惩罚的优化方案漂加惩罚(Drift-Plus-Penalty, DPP),以促进惩罚学习并减少后悔。随着dpp导向的进化,奖励学习与汤普森采样(TS)越来越增强,以探索全局最优在训练后期阶段。该方案通过各种基准测试(包括数值问题、物理环境和游戏)进行了大量实验。由于模仿人类的学习过程,该方案在基准测试中大大优于最先进的ES。
{"title":"APR-ES: Adaptive Penalty-Reward Based Evolution Strategy for Deep Reinforcement Learning","authors":"Dongdong Wang, Siyang Lu, Xiang Wei, Mingquan Wang, Yandong Li, Liqiang Wang","doi":"10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00079","DOIUrl":"https://doi.org/10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00079","url":null,"abstract":"As a black-box optimization approach, derivative-free evolution strategy (ES) draws lots of attention in virtue of its low sensitivity and high scalability. It rivals Markov Decision Process based reinforcement learning or even can more efficiently improve rewards under complex scenarios. However, existing derivative-free ES still confronts slow convergence speed at the early training stage and limited exploration at the late convergence stage. Inspired from human learning process, we propose a new scheme extended from ES by taking advantage of prior knowledge to guide ES, thus accelerating early exploitation process and improving later exploration ability. At early training stage, Drift-Plus-Penalty (DPP), a penalty-based optimization scheme, is reformulated to boost penalty learning and reduce regrets. Along with DPP-directed evolution, reward learning with Thompson sampling (TS) is increasingly enhanced to explore global optima at late training stage. This scheme is justified with extensive experiments from a variety of benchmarks, including numerical problems, physics environments, and games. By virtue of its imitation of human learning process, this scheme outperforms state-of-the-art ES on the benchmarks by a large margin.","PeriodicalId":43791,"journal":{"name":"Scalable Computing-Practice and Experience","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88379860","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Self-distilled Named Entity Recognition Based on Boundary Detection and Biaffine Attention 基于边界检测和双碱注意的自蒸馏命名实体识别
IF 1.1 Q2 Computer Science Pub Date : 2022-12-01 DOI: 10.1109/smartworld-uic-atc-scalcom-digitaltwin-pricomp-metaverse56740.2022.00162
Yong Song, Zhiwei Yan, Yukun Qin, Xiaozhou Ye, Ye Ouyang
Named Entity Recognition (NER) is an important down-streaming task in natural language processing. Span-based methods are applicable to both flat and nested entities. However, they lack explicit boundary supervision. To address this issue, we propose a multi-task and self-distilled model which combines biaffine span classification and entity boundary detection tasks. Firstly, the boundary detection and biaffine span classification models are jointly trained under a multi-task learning framework to address the problem of lacking supervision of boundaries. Then, self-distillation technique is applied on the model to reassign entity probabilities from annotated spans to surrounding spans and more entity types, further improving the accuracy of NER by soft labels that contain richer knowledge. Experiments were based on a high-density entity text dataset of the commodity titles from an e-commerce company. Finally, the experimental results show that our model exhibited a better F1 score than the existing common models.
命名实体识别(NER)是自然语言处理中一个重要的下行任务。基于跨度的方法既适用于平面实体,也适用于嵌套实体。然而,它们缺乏明确的边界监督。为了解决这一问题,我们提出了一种多任务自提取模型,该模型结合了双仿跨度分类和实体边界检测任务。首先,在多任务学习框架下,联合训练边界检测模型和biaffine跨度分类模型,解决边界缺乏监督的问题;然后,在模型上应用自蒸馏技术,将实体概率从标注的跨度重新分配到周围跨度和更多的实体类型,通过包含更丰富知识的软标签进一步提高NER的准确性。实验基于一家电子商务公司商品标题的高密度实体文本数据集。最后,实验结果表明,我们的模型比现有的常用模型具有更好的F1分数。
{"title":"Self-distilled Named Entity Recognition Based on Boundary Detection and Biaffine Attention","authors":"Yong Song, Zhiwei Yan, Yukun Qin, Xiaozhou Ye, Ye Ouyang","doi":"10.1109/smartworld-uic-atc-scalcom-digitaltwin-pricomp-metaverse56740.2022.00162","DOIUrl":"https://doi.org/10.1109/smartworld-uic-atc-scalcom-digitaltwin-pricomp-metaverse56740.2022.00162","url":null,"abstract":"Named Entity Recognition (NER) is an important down-streaming task in natural language processing. Span-based methods are applicable to both flat and nested entities. However, they lack explicit boundary supervision. To address this issue, we propose a multi-task and self-distilled model which combines biaffine span classification and entity boundary detection tasks. Firstly, the boundary detection and biaffine span classification models are jointly trained under a multi-task learning framework to address the problem of lacking supervision of boundaries. Then, self-distillation technique is applied on the model to reassign entity probabilities from annotated spans to surrounding spans and more entity types, further improving the accuracy of NER by soft labels that contain richer knowledge. Experiments were based on a high-density entity text dataset of the commodity titles from an e-commerce company. Finally, the experimental results show that our model exhibited a better F1 score than the existing common models.","PeriodicalId":43791,"journal":{"name":"Scalable Computing-Practice and Experience","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90184901","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FedGPS: Personalized Cross-Silo Federated Learning for Internet of Things-enabled Predictive Maintenance FedGPS:面向物联网预测性维护的个性化跨筒仓联邦学习
IF 1.1 Q2 Computer Science Pub Date : 2022-12-01 DOI: 10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00137
Yuchen Jiang, Chang Ji
Predictive maintenance (PdM) has entered into a new era adopting artificial intelligence and Internet-of-Things (IoT) technologies. It is necessary for a manufacturing company to collaborate with other clients using IoT-captured production data. However, training models in a cross-silo manner is still challenging when considering data privacy. In order to tackle these challenges, a personalized cross-silo federated learning mechanism named federated global partners searching (FedGPS) is proposed. Firstly, model parameters for the participating clients are encrypted and uploaded to the central server as input. Next, FedGPS automatically determines the collaboration degrees between clients based on data distribution. After that, personalized model updates are sent back to the clients. Finally, each client conducts local updating after data decryption. The effectiveness of the FedGPS is verified in real-world cases and our method achieves 92.35% Accuracy, 98.55% Precision, 92.90% Recall, and 95.27% F1-Score comparing with other existing models from the literature.
预测性维护(PdM)已经进入了人工智能和物联网(IoT)技术的新时代。制造公司有必要使用物联网捕获的生产数据与其他客户进行协作。然而,在考虑数据隐私时,以跨竖井的方式训练模型仍然具有挑战性。为了解决这些问题,提出了一种个性化的跨竖井联邦学习机制——联邦全局伙伴搜索(federal global partners searching, FedGPS)。首先,对参与客户端的模型参数进行加密,并将其作为输入上传到中央服务器。其次,FedGPS根据数据分布自动确定客户端之间的协作程度。之后,个性化的模型更新被发送回客户端。最后,各客户端在数据解密后进行本地更新。在实际案例中验证了FedGPS的有效性,与文献中已有的模型相比,我们的方法达到了92.35%的准确率、98.55%的精密度、92.90%的召回率和95.27%的F1-Score。
{"title":"FedGPS: Personalized Cross-Silo Federated Learning for Internet of Things-enabled Predictive Maintenance","authors":"Yuchen Jiang, Chang Ji","doi":"10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00137","DOIUrl":"https://doi.org/10.1109/SmartWorld-UIC-ATC-ScalCom-DigitalTwin-PriComp-Metaverse56740.2022.00137","url":null,"abstract":"Predictive maintenance (PdM) has entered into a new era adopting artificial intelligence and Internet-of-Things (IoT) technologies. It is necessary for a manufacturing company to collaborate with other clients using IoT-captured production data. However, training models in a cross-silo manner is still challenging when considering data privacy. In order to tackle these challenges, a personalized cross-silo federated learning mechanism named federated global partners searching (FedGPS) is proposed. Firstly, model parameters for the participating clients are encrypted and uploaded to the central server as input. Next, FedGPS automatically determines the collaboration degrees between clients based on data distribution. After that, personalized model updates are sent back to the clients. Finally, each client conducts local updating after data decryption. The effectiveness of the FedGPS is verified in real-world cases and our method achieves 92.35% Accuracy, 98.55% Precision, 92.90% Recall, and 95.27% F1-Score comparing with other existing models from the literature.","PeriodicalId":43791,"journal":{"name":"Scalable Computing-Practice and Experience","volume":null,"pages":null},"PeriodicalIF":1.1,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90350608","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Scalable Computing-Practice and Experience
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1