首页 > 最新文献

Inf. Sci.最新文献

英文 中文
Fully distributed event-triggered cooperative output regulation for switched multi-agent systems with combined switching mechanism 具有组合交换机制的交换多智能体系统的全分布式事件触发协同输出调节
Pub Date : 2023-04-01 DOI: 10.2139/ssrn.4290226
Guangxu He, Jun Zhao
{"title":"Fully distributed event-triggered cooperative output regulation for switched multi-agent systems with combined switching mechanism","authors":"Guangxu He, Jun Zhao","doi":"10.2139/ssrn.4290226","DOIUrl":"https://doi.org/10.2139/ssrn.4290226","url":null,"abstract":"","PeriodicalId":13641,"journal":{"name":"Inf. Sci.","volume":"39 1","pages":"118970"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82644622","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Steering the interpretability of decision trees using lasso regression - an evolutionary perspective 使用套索回归控制决策树的可解释性——一种进化的观点
Pub Date : 2023-04-01 DOI: 10.2139/ssrn.4331060
M. Czajkowski, K. Jurczuk, M. Kretowski
{"title":"Steering the interpretability of decision trees using lasso regression - an evolutionary perspective","authors":"M. Czajkowski, K. Jurczuk, M. Kretowski","doi":"10.2139/ssrn.4331060","DOIUrl":"https://doi.org/10.2139/ssrn.4331060","url":null,"abstract":"","PeriodicalId":13641,"journal":{"name":"Inf. Sci.","volume":"95 1","pages":"118944"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91334429","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Ensuring confidentiality of cyber-physical systems using event-based cryptography 使用基于事件的加密技术确保网络物理系统的机密性
Pub Date : 2023-04-01 DOI: 10.1016/J.IFACOL.2020.12.2288
Públio M. Lima, L. K. Carvalho, M. V. Moreira
{"title":"Ensuring confidentiality of cyber-physical systems using event-based cryptography","authors":"Públio M. Lima, L. K. Carvalho, M. V. Moreira","doi":"10.1016/J.IFACOL.2020.12.2288","DOIUrl":"https://doi.org/10.1016/J.IFACOL.2020.12.2288","url":null,"abstract":"","PeriodicalId":13641,"journal":{"name":"Inf. Sci.","volume":"18 1","pages":"119-135"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84774453","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Three-way conflict analysis and resolution based on q-rung orthopair fuzzy information 基于q阶正交模糊信息的三方冲突分析与解决
Pub Date : 2023-04-01 DOI: 10.2139/ssrn.4278353
Teng Li, Junsheng Qiao, Weiping Ding
{"title":"Three-way conflict analysis and resolution based on q-rung orthopair fuzzy information","authors":"Teng Li, Junsheng Qiao, Weiping Ding","doi":"10.2139/ssrn.4278353","DOIUrl":"https://doi.org/10.2139/ssrn.4278353","url":null,"abstract":"","PeriodicalId":13641,"journal":{"name":"Inf. Sci.","volume":"42 1","pages":"118959"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73927384","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Are Neural Architecture Search Benchmarks Well Designed? A Deeper Look Into Operation Importance 神经架构搜索基准设计得好吗?对操作重要性的深入探讨
Pub Date : 2023-03-29 DOI: 10.48550/arXiv.2303.16938
Vasco Lopes, Bruno Degardin, L. Alexandre
Neural Architecture Search (NAS) benchmarks significantly improved the capability of developing and comparing NAS methods while at the same time drastically reduced the computational overhead by providing meta-information about thousands of trained neural networks. However, tabular benchmarks have several drawbacks that can hinder fair comparisons and provide unreliable results. These usually focus on providing a small pool of operations in heavily constrained search spaces -- usually cell-based neural networks with pre-defined outer-skeletons. In this work, we conducted an empirical analysis of the widely used NAS-Bench-101, NAS-Bench-201 and TransNAS-Bench-101 benchmarks in terms of their generability and how different operations influence the performance of the generated architectures. We found that only a subset of the operation pool is required to generate architectures close to the upper-bound of the performance range. Also, the performance distribution is negatively skewed, having a higher density of architectures in the upper-bound range. We consistently found convolution layers to have the highest impact on the architecture's performance, and that specific combination of operations favors top-scoring architectures. These findings shed insights on the correct evaluation and comparison of NAS methods using NAS benchmarks, showing that directly searching on NAS-Bench-201, ImageNet16-120 and TransNAS-Bench-101 produces more reliable results than searching only on CIFAR-10. Furthermore, with this work we provide suggestions for future benchmark evaluations and design. The code used to conduct the evaluations is available at https://github.com/VascoLopes/NAS-Benchmark-Evaluation.
Neural Architecture Search (NAS)基准测试显著提高了开发和比较NAS方法的能力,同时通过提供关于数千个训练过的神经网络的元信息,大大降低了计算开销。然而,表格基准测试有几个缺点,可能会妨碍公平的比较,并提供不可靠的结果。这些通常侧重于在严重受限的搜索空间中提供一小部分操作——通常是具有预定义外部骨架的基于细胞的神经网络。在这项工作中,我们对广泛使用的NAS-Bench-101、NAS-Bench-201和TransNAS-Bench-101基准进行了实证分析,分析了它们的可通用性,以及不同的操作如何影响所生成架构的性能。我们发现,只需要操作池的一个子集就可以生成接近性能范围上限的体系结构。此外,性能分布呈负偏态,在上限范围内具有更高的架构密度。我们一直发现卷积层对体系结构的性能有最大的影响,并且特定的操作组合有利于得分最高的体系结构。这些发现揭示了使用NAS基准正确评估和比较NAS方法的见解,表明直接在NAS- bench -201、ImageNet16-120和TransNAS-Bench-101上搜索比仅在CIFAR-10上搜索产生更可靠的结果。此外,通过这项工作,我们为未来的基准评估和设计提供了建议。用于进行评估的代码可在https://github.com/VascoLopes/NAS-Benchmark-Evaluation上获得。
{"title":"Are Neural Architecture Search Benchmarks Well Designed? A Deeper Look Into Operation Importance","authors":"Vasco Lopes, Bruno Degardin, L. Alexandre","doi":"10.48550/arXiv.2303.16938","DOIUrl":"https://doi.org/10.48550/arXiv.2303.16938","url":null,"abstract":"Neural Architecture Search (NAS) benchmarks significantly improved the capability of developing and comparing NAS methods while at the same time drastically reduced the computational overhead by providing meta-information about thousands of trained neural networks. However, tabular benchmarks have several drawbacks that can hinder fair comparisons and provide unreliable results. These usually focus on providing a small pool of operations in heavily constrained search spaces -- usually cell-based neural networks with pre-defined outer-skeletons. In this work, we conducted an empirical analysis of the widely used NAS-Bench-101, NAS-Bench-201 and TransNAS-Bench-101 benchmarks in terms of their generability and how different operations influence the performance of the generated architectures. We found that only a subset of the operation pool is required to generate architectures close to the upper-bound of the performance range. Also, the performance distribution is negatively skewed, having a higher density of architectures in the upper-bound range. We consistently found convolution layers to have the highest impact on the architecture's performance, and that specific combination of operations favors top-scoring architectures. These findings shed insights on the correct evaluation and comparison of NAS methods using NAS benchmarks, showing that directly searching on NAS-Bench-201, ImageNet16-120 and TransNAS-Bench-101 produces more reliable results than searching only on CIFAR-10. Furthermore, with this work we provide suggestions for future benchmark evaluations and design. The code used to conduct the evaluations is available at https://github.com/VascoLopes/NAS-Benchmark-Evaluation.","PeriodicalId":13641,"journal":{"name":"Inf. Sci.","volume":"5 1","pages":"119695"},"PeriodicalIF":0.0,"publicationDate":"2023-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85578187","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving the Transferability of Adversarial Examples via Direction Tuning 通过方向调整提高对抗性样例的可转移性
Pub Date : 2023-03-27 DOI: 10.48550/arXiv.2303.15109
Xiangyuan Yang, Jie Lin, Han Zhang, Xinyu Yang, Peng Zhao
In the transfer-based adversarial attacks, adversarial examples are only generated by the surrogate models and achieve effective perturbation in the victim models. Although considerable efforts have been developed on improving the transferability of adversarial examples generated by transfer-based adversarial attacks, our investigation found that, the big deviation between the actual and steepest update directions of the current transfer-based adversarial attacks is caused by the large update step length, resulting in the generated adversarial examples can not converge well. However, directly reducing the update step length will lead to serious update oscillation so that the generated adversarial examples also can not achieve great transferability to the victim models. To address these issues, a novel transfer-based attack, namely direction tuning attack, is proposed to not only decrease the update deviation in the large step length, but also mitigate the update oscillation in the small sampling step length, thereby making the generated adversarial examples converge well to achieve great transferability on victim models. In addition, a network pruning method is proposed to smooth the decision boundary, thereby further decreasing the update oscillation and enhancing the transferability of the generated adversarial examples. The experiment results on ImageNet demonstrate that the average attack success rate (ASR) of the adversarial examples generated by our method can be improved from 87.9% to 94.5% on five victim models without defenses, and from 69.1% to 76.2% on eight advanced defense methods, in comparison with that of latest gradient-based attacks.
在基于转移的对抗性攻击中,对抗性示例仅由代理模型生成,并在受害者模型中实现有效扰动。虽然在提高基于迁移的对抗性攻击生成的对抗性示例的可转移性方面已经做出了相当大的努力,但我们的研究发现,目前基于迁移的对抗性攻击的实际更新方向与最陡更新方向之间存在较大的偏差,这是由于更新步长较大,导致生成的对抗性示例不能很好地收敛。然而,直接减小更新步长会导致严重的更新振荡,使得生成的对抗样例也不能很好地转移到受害模型。为了解决这些问题,提出了一种新的基于迁移的攻击方法,即方向调谐攻击,该方法不仅可以减小大步长的更新偏差,而且可以减轻小步长的更新振荡,从而使生成的对抗样例很好地收敛,从而在受害者模型上实现很大的可转移性。此外,提出了一种网络剪枝方法来平滑决策边界,从而进一步降低了更新振荡,增强了生成的对抗样例的可转移性。在ImageNet上的实验结果表明,与最新的基于梯度的攻击相比,本文方法生成的对抗样本在无防御的5种受害者模型上的平均攻击成功率(ASR)从87.9%提高到94.5%,在8种高级防御方法上的平均攻击成功率(ASR)从69.1%提高到76.2%。
{"title":"Improving the Transferability of Adversarial Examples via Direction Tuning","authors":"Xiangyuan Yang, Jie Lin, Han Zhang, Xinyu Yang, Peng Zhao","doi":"10.48550/arXiv.2303.15109","DOIUrl":"https://doi.org/10.48550/arXiv.2303.15109","url":null,"abstract":"In the transfer-based adversarial attacks, adversarial examples are only generated by the surrogate models and achieve effective perturbation in the victim models. Although considerable efforts have been developed on improving the transferability of adversarial examples generated by transfer-based adversarial attacks, our investigation found that, the big deviation between the actual and steepest update directions of the current transfer-based adversarial attacks is caused by the large update step length, resulting in the generated adversarial examples can not converge well. However, directly reducing the update step length will lead to serious update oscillation so that the generated adversarial examples also can not achieve great transferability to the victim models. To address these issues, a novel transfer-based attack, namely direction tuning attack, is proposed to not only decrease the update deviation in the large step length, but also mitigate the update oscillation in the small sampling step length, thereby making the generated adversarial examples converge well to achieve great transferability on victim models. In addition, a network pruning method is proposed to smooth the decision boundary, thereby further decreasing the update oscillation and enhancing the transferability of the generated adversarial examples. The experiment results on ImageNet demonstrate that the average attack success rate (ASR) of the adversarial examples generated by our method can be improved from 87.9% to 94.5% on five victim models without defenses, and from 69.1% to 76.2% on eight advanced defense methods, in comparison with that of latest gradient-based attacks.","PeriodicalId":13641,"journal":{"name":"Inf. Sci.","volume":"27 1","pages":"119491"},"PeriodicalIF":0.0,"publicationDate":"2023-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82767151","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
An Empirical Analysis of the Shift and Scale Parameters in BatchNorm BatchNorm中位移和尺度参数的实证分析
Pub Date : 2023-03-22 DOI: 10.48550/arXiv.2303.12818
Y. Peerthum, M. Stamp
Batch Normalization (BatchNorm) is a technique that improves the training of deep neural networks, especially Convolutional Neural Networks (CNN). It has been empirically demonstrated that BatchNorm increases performance, stability, and accuracy, although the reasons for such improvements are unclear. BatchNorm includes a normalization step as well as trainable shift and scale parameters. In this paper, we empirically examine the relative contribution to the success of BatchNorm of the normalization step, as compared to the re-parameterization via shifting and scaling. To conduct our experiments, we implement two new optimizers in PyTorch, namely, a version of BatchNorm that we refer to as AffineLayer, which includes the re-parameterization step without normalization, and a version with just the normalization step, that we call BatchNorm-minus. We compare the performance of our AffineLayer and BatchNorm-minus implementations to standard BatchNorm, and we also compare these to the case where no batch normalization is used. We experiment with four ResNet architectures (ResNet18, ResNet34, ResNet50, and ResNet101) over a standard image dataset and multiple batch sizes. Among other findings, we provide empirical evidence that the success of BatchNorm may derive primarily from improved weight initialization.
批归一化(BatchNorm)是一种改进深度神经网络,特别是卷积神经网络(CNN)训练的技术。经验证明,BatchNorm提高了性能、稳定性和准确性,尽管这些改进的原因尚不清楚。BatchNorm包括一个规范化步骤以及可训练的移位和缩放参数。在本文中,我们通过经验检验了标准化步骤对BatchNorm成功的相对贡献,与通过移动和缩放的重新参数化相比。为了进行我们的实验,我们在PyTorch中实现了两个新的优化器,即一个版本的BatchNorm,我们称之为AffineLayer,它包括没有规范化的重新参数化步骤,一个版本只有规范化步骤,我们称之为BatchNorm-minus。我们将AffineLayer和BatchNorm-minus实现的性能与标准BatchNorm进行比较,并将它们与不使用批处理规范化的情况进行比较。我们在标准图像数据集和多个批处理大小上实验了四种ResNet架构(ResNet18, ResNet34, ResNet50和ResNet101)。在其他发现中,我们提供了经验证据,证明BatchNorm的成功可能主要来自改进的权重初始化。
{"title":"An Empirical Analysis of the Shift and Scale Parameters in BatchNorm","authors":"Y. Peerthum, M. Stamp","doi":"10.48550/arXiv.2303.12818","DOIUrl":"https://doi.org/10.48550/arXiv.2303.12818","url":null,"abstract":"Batch Normalization (BatchNorm) is a technique that improves the training of deep neural networks, especially Convolutional Neural Networks (CNN). It has been empirically demonstrated that BatchNorm increases performance, stability, and accuracy, although the reasons for such improvements are unclear. BatchNorm includes a normalization step as well as trainable shift and scale parameters. In this paper, we empirically examine the relative contribution to the success of BatchNorm of the normalization step, as compared to the re-parameterization via shifting and scaling. To conduct our experiments, we implement two new optimizers in PyTorch, namely, a version of BatchNorm that we refer to as AffineLayer, which includes the re-parameterization step without normalization, and a version with just the normalization step, that we call BatchNorm-minus. We compare the performance of our AffineLayer and BatchNorm-minus implementations to standard BatchNorm, and we also compare these to the case where no batch normalization is used. We experiment with four ResNet architectures (ResNet18, ResNet34, ResNet50, and ResNet101) over a standard image dataset and multiple batch sizes. Among other findings, we provide empirical evidence that the success of BatchNorm may derive primarily from improved weight initialization.","PeriodicalId":13641,"journal":{"name":"Inf. Sci.","volume":"104 1","pages":"118951"},"PeriodicalIF":0.0,"publicationDate":"2023-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75507549","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Assessor-Guided Learning for Continual Environments 持续环境的评估指导学习
Pub Date : 2023-03-21 DOI: 10.48550/arXiv.2303.11624
M. A. Ma'sum, Mahardhika Pratama, E. Lughofer, Weiping Ding, W. Jatmiko
This paper proposes an assessor-guided learning strategy for continual learning where an assessor guides the learning process of a base learner by controlling the direction and pace of the learning process thus allowing an efficient learning of new environments while protecting against the catastrophic interference problem. The assessor is trained in a meta-learning manner with a meta-objective to boost the learning process of the base learner. It performs a soft-weighting mechanism of every sample accepting positive samples while rejecting negative samples. The training objective of a base learner is to minimize a meta-weighted combination of the cross entropy loss function, the dark experience replay (DER) loss function and the knowledge distillation loss function whose interactions are controlled in such a way to attain an improved performance. A compensated over-sampling (COS) strategy is developed to overcome the class imbalanced problem of the episodic memory due to limited memory budgets. Our approach, Assessor-Guided Learning Approach (AGLA), has been evaluated in the class-incremental and task-incremental learning problems. AGLA achieves improved performances compared to its competitors while the theoretical analysis of the COS strategy is offered. Source codes of AGLA, baseline algorithms and experimental logs are shared publicly in url{https://github.com/anwarmaxsum/AGLA} for further study.
本文提出了一种用于持续学习的评估器引导学习策略,其中评估器通过控制学习过程的方向和速度来指导基础学习者的学习过程,从而在防止灾难性干扰问题的同时有效地学习新环境。评估员以元学习的方式进行训练,其元目标是促进基础学习者的学习过程。它执行每个样本接受正样本而拒绝负样本的软加权机制。基础学习器的训练目标是最小化交叉熵损失函数、暗经验重放(dark experience replay, DER)损失函数和知识蒸馏损失函数的元加权组合,以达到改进的性能。为了克服由于记忆预算有限而导致的情景记忆类不平衡问题,提出了一种补偿过采样策略。我们的方法,评估员引导学习方法(AGLA),已经在班级增量和任务增量学习问题中进行了评估。与竞争对手相比,AGLA实现了更高的性能,并对COS策略进行了理论分析。AGLA的源代码、基线算法和实验日志在url{https://github.com/anwarmaxsum/AGLA}上公开共享,以供进一步研究。
{"title":"Assessor-Guided Learning for Continual Environments","authors":"M. A. Ma'sum, Mahardhika Pratama, E. Lughofer, Weiping Ding, W. Jatmiko","doi":"10.48550/arXiv.2303.11624","DOIUrl":"https://doi.org/10.48550/arXiv.2303.11624","url":null,"abstract":"This paper proposes an assessor-guided learning strategy for continual learning where an assessor guides the learning process of a base learner by controlling the direction and pace of the learning process thus allowing an efficient learning of new environments while protecting against the catastrophic interference problem. The assessor is trained in a meta-learning manner with a meta-objective to boost the learning process of the base learner. It performs a soft-weighting mechanism of every sample accepting positive samples while rejecting negative samples. The training objective of a base learner is to minimize a meta-weighted combination of the cross entropy loss function, the dark experience replay (DER) loss function and the knowledge distillation loss function whose interactions are controlled in such a way to attain an improved performance. A compensated over-sampling (COS) strategy is developed to overcome the class imbalanced problem of the episodic memory due to limited memory budgets. Our approach, Assessor-Guided Learning Approach (AGLA), has been evaluated in the class-incremental and task-incremental learning problems. AGLA achieves improved performances compared to its competitors while the theoretical analysis of the COS strategy is offered. Source codes of AGLA, baseline algorithms and experimental logs are shared publicly in url{https://github.com/anwarmaxsum/AGLA} for further study.","PeriodicalId":13641,"journal":{"name":"Inf. Sci.","volume":"84 1","pages":"119088"},"PeriodicalIF":0.0,"publicationDate":"2023-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80913230","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
THAT-Net: Two-layer hidden state aggregation based two-stream network for traffic accident prediction 基于两层隐藏状态聚合的两流交通事故预测网络
Pub Date : 2023-03-01 DOI: 10.2139/ssrn.4331054
Wei Liu, Zhang Tao, Yisheng Lu, Jun Chen, Longsheng Wei
{"title":"THAT-Net: Two-layer hidden state aggregation based two-stream network for traffic accident prediction","authors":"Wei Liu, Zhang Tao, Yisheng Lu, Jun Chen, Longsheng Wei","doi":"10.2139/ssrn.4331054","DOIUrl":"https://doi.org/10.2139/ssrn.4331054","url":null,"abstract":"","PeriodicalId":13641,"journal":{"name":"Inf. Sci.","volume":"2 1","pages":"744-760"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82832498","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
PKET-GCN: Prior knowledge enhanced time-varying graph convolution network for traffic flow prediction 基于先验知识增强时变图卷积网络的交通流预测
Pub Date : 2023-03-01 DOI: 10.2139/ssrn.4331039
Yinxin Bao, Jialin Liu, Qinqin Shen, Yang Cao, Weiping Ding, Quan Shi
{"title":"PKET-GCN: Prior knowledge enhanced time-varying graph convolution network for traffic flow prediction","authors":"Yinxin Bao, Jialin Liu, Qinqin Shen, Yang Cao, Weiping Ding, Quan Shi","doi":"10.2139/ssrn.4331039","DOIUrl":"https://doi.org/10.2139/ssrn.4331039","url":null,"abstract":"","PeriodicalId":13641,"journal":{"name":"Inf. Sci.","volume":"39 1","pages":"359-381"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91131787","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
Inf. Sci.
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1