首页 > 最新文献

World Wide Web-Internet and Web Information Systems最新文献

英文 中文
Explanation-based data-free model extraction attacks 基于解释的无数据模型提取攻击
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-06-02 DOI: 10.1007/s11280-023-01150-6
Anli Yan, Ruitao Hou, Hongyang Yan, Xiaozhang Liu
{"title":"Explanation-based data-free model extraction attacks","authors":"Anli Yan, Ruitao Hou, Hongyang Yan, Xiaozhang Liu","doi":"10.1007/s11280-023-01150-6","DOIUrl":"https://doi.org/10.1007/s11280-023-01150-6","url":null,"abstract":"","PeriodicalId":49356,"journal":{"name":"World Wide Web-Internet and Web Information Systems","volume":"130 1","pages":"3081 - 3092"},"PeriodicalIF":3.7,"publicationDate":"2023-06-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78720822","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A novel semantic-aware search scheme based on BCI-tree index over encrypted cloud data 一种基于bci树索引的加密云数据语义感知搜索方案
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-05-31 DOI: 10.1007/s11280-023-01176-w
Qiang Zhou, Hua Dai, Yuanlong Liu, Geng Yang, X. Yi, Zheng Hu
{"title":"A novel semantic-aware search scheme based on BCI-tree index over encrypted cloud data","authors":"Qiang Zhou, Hua Dai, Yuanlong Liu, Geng Yang, X. Yi, Zheng Hu","doi":"10.1007/s11280-023-01176-w","DOIUrl":"https://doi.org/10.1007/s11280-023-01176-w","url":null,"abstract":"","PeriodicalId":49356,"journal":{"name":"World Wide Web-Internet and Web Information Systems","volume":"1 1","pages":"3055 - 3079"},"PeriodicalIF":3.7,"publicationDate":"2023-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75179629","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Soft dimensionality reduction for reinforcement data clustering 增强数据聚类的软降维
3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-05-30 DOI: 10.1007/s11280-023-01158-y
Fatemeh Fathinezhad, Peyman Adibi, Bijan Shoushtarian, Hamidreza Baradaran Kashani, Jocelyn Chanussot
{"title":"Soft dimensionality reduction for reinforcement data clustering","authors":"Fatemeh Fathinezhad, Peyman Adibi, Bijan Shoushtarian, Hamidreza Baradaran Kashani, Jocelyn Chanussot","doi":"10.1007/s11280-023-01158-y","DOIUrl":"https://doi.org/10.1007/s11280-023-01158-y","url":null,"abstract":"","PeriodicalId":49356,"journal":{"name":"World Wide Web-Internet and Web Information Systems","volume":"71 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135643300","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Phrase-level attention network for few-shot inverse relation classification in knowledge graph 知识图谱中多镜头反关系分类的短语级关注网络
3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-05-30 DOI: 10.1007/s11280-023-01142-6
Shaojuan Wu, Chunliu Dou, Dazhuang Wang, Jitong Li, Xiaowang Zhang, Zhiyong Feng, Kewen Wang, Sofonias Yitagesu
Relation classification aims to recognize semantic relation between two given entities mentioned in the given text. Existing models have performed well on the inverse relation classification with large-scale datasets, but their performance drops significantly for few-shot learning. In this paper, we propose a Phrase-level Attention Network, function words adaptively enhanced attention framework (FAEA+), to attend class-related function words by the designed hybrid attention for few-shot inverse relation classification in Knowledge Graph. Then, an instance-aware prototype network is present to adaptively capture relation information associated with query instances and eliminate intra-class redundancy due to function words introduced. We theoretically prove that the introduction of function words will increase intra-class differences, and the designed instance-aware prototype network is competent for reducing redundancy. Experimental results show that FAEA+ significantly improved over strong baselines on two few-shot relation classification datasets. Moreover, our model has a distinct advantage in solving inverse relations, which outperforms state-of-the-art results by 16.82% under a 1-shot setting in FewRel1.0.
关系分类的目的是识别给定文本中提到的两个给定实体之间的语义关系。现有模型在大规模数据集的反关系分类上表现良好,但在小样本学习时,其性能明显下降。本文提出了一种短语级注意网络——虚词自适应增强注意框架(FAEA+),通过设计的混合注意关注类相关虚词,在知识图谱中进行几次反比关系分类。然后,提出一个实例感知的原型网络,自适应捕获与查询实例相关的关系信息,消除由于引入功能词而导致的类内冗余。我们从理论上证明了功能词的引入会增加类内差异,并且所设计的实例感知原型网络能够减少冗余。实验结果表明,FAEA+在强基线条件下,在两组少样本关系分类数据集上有显著的提高。此外,我们的模型在求解逆关系方面具有明显的优势,在FewRel1.0的1次设置下,它比最先进的结果高出16.82%。
{"title":"Phrase-level attention network for few-shot inverse relation classification in knowledge graph","authors":"Shaojuan Wu, Chunliu Dou, Dazhuang Wang, Jitong Li, Xiaowang Zhang, Zhiyong Feng, Kewen Wang, Sofonias Yitagesu","doi":"10.1007/s11280-023-01142-6","DOIUrl":"https://doi.org/10.1007/s11280-023-01142-6","url":null,"abstract":"Relation classification aims to recognize semantic relation between two given entities mentioned in the given text. Existing models have performed well on the inverse relation classification with large-scale datasets, but their performance drops significantly for few-shot learning. In this paper, we propose a Phrase-level Attention Network, function words adaptively enhanced attention framework (FAEA+), to attend class-related function words by the designed hybrid attention for few-shot inverse relation classification in Knowledge Graph. Then, an instance-aware prototype network is present to adaptively capture relation information associated with query instances and eliminate intra-class redundancy due to function words introduced. We theoretically prove that the introduction of function words will increase intra-class differences, and the designed instance-aware prototype network is competent for reducing redundancy. Experimental results show that FAEA+ significantly improved over strong baselines on two few-shot relation classification datasets. Moreover, our model has a distinct advantage in solving inverse relations, which outperforms state-of-the-art results by 16.82% under a 1-shot setting in FewRel1.0.","PeriodicalId":49356,"journal":{"name":"World Wide Web-Internet and Web Information Systems","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135643530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Empowering reciprocal recommender system using contextual bandits and argumentation based explanations 授权互惠推荐系统使用上下文强盗和基于论证的解释
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-05-29 DOI: 10.1007/s11280-023-01173-z
T. Kumari, Bhavna Gupta, Ravish Sharma, Punam Bedi
{"title":"Empowering reciprocal recommender system using contextual bandits and argumentation based explanations","authors":"T. Kumari, Bhavna Gupta, Ravish Sharma, Punam Bedi","doi":"10.1007/s11280-023-01173-z","DOIUrl":"https://doi.org/10.1007/s11280-023-01173-z","url":null,"abstract":"","PeriodicalId":49356,"journal":{"name":"World Wide Web-Internet and Web Information Systems","volume":"26 1","pages":"2969 - 3000"},"PeriodicalIF":3.7,"publicationDate":"2023-05-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74543781","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HTSE: hierarchical time-surface model for temporal knowledge graph embedding 时间知识图嵌入的层次时间面模型
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-05-29 DOI: 10.1007/s11280-023-01170-2
Langjunqing Jin, Feng Zhao, Hai Jin
{"title":"HTSE: hierarchical time-surface model for temporal knowledge graph embedding","authors":"Langjunqing Jin, Feng Zhao, Hai Jin","doi":"10.1007/s11280-023-01170-2","DOIUrl":"https://doi.org/10.1007/s11280-023-01170-2","url":null,"abstract":"","PeriodicalId":49356,"journal":{"name":"World Wide Web-Internet and Web Information Systems","volume":"56 1","pages":"2947 - 2967"},"PeriodicalIF":3.7,"publicationDate":"2023-05-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80212937","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Transfer learning based cascaded deep learning network and mask recognition for COVID-19. 基于迁移学习的级联深度学习网络和新冠肺炎口罩识别。
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-05-26 DOI: 10.1007/s11280-023-01149-z
Fengyin Li, Xiaojiao Wang, Yuhong Sun, Tao Li, Junrong Ge

The COVID-19 is still spreading today, and it has caused great harm to human beings. The system at the entrance of public places such as shopping malls and stations should check whether pedestrians are wearing masks. However, pedestrians often pass the system inspection by wearing cotton masks, scarves, etc. Therefore, the detection system not only needs to check whether pedestrians are wearing masks, but also needs to detect the type of masks. Based on the lightweight network architecture MobilenetV3, this paper proposes a cascaded deep learning network based on transfer learning, and then designs a mask recognition system based on the cascaded deep learning network. By modifying the activation function of the MobilenetV3 output layer and the structure of the model, two MobilenetV3 networks suitable for cascading are obtained. By introducing transfer learning into the training process of two modified MobilenetV3 networks and a multi-task convolutional neural network, the ImagNet underlying parameters of the network models are obtained in advance, which reduces the computational load of the models. The cascaded deep learning network consists of a multi-task convolutional neural network cascaded with these two modified MobilenetV3 networks. A multi-task convolutional neural network is used to detect faces in images, and two modified MobilenetV3 networks are used as the backbone network to extract the features of masks. After comparing with the classification results of the modified MobilenetV3 neural network before cascading, the classification accuracy of the cascading learning network is improved by 7%, and the excellent performance of the cascading network can be seen.

新冠肺炎疫情至今仍在蔓延,给人类造成了巨大危害。商场和车站等公共场所入口处的系统应检查行人是否戴口罩。然而,行人经常戴着棉质口罩、围巾等通过系统检查。因此,检测系统不仅需要检查行人是否戴口罩,还需要检测口罩的类型。基于轻量级网络架构MobilenetV3,本文提出了一种基于迁移学习的级联深度学习网络,然后设计了一个基于级联深度学习的掩模识别系统。通过修改MobiletV3输出层的激活函数和模型的结构,获得了两个适合级联的MobiletV3网络。通过将迁移学习引入两个改进的MobileneV3网络和一个多任务卷积神经网络的训练过程,提前获得了网络模型的ImagNet底层参数,降低了模型的计算负载。级联深度学习网络由一个多任务卷积神经网络组成,该网络与这两个改进的MobileneV3网络级联。使用多任务卷积神经网络来检测图像中的人脸,并使用两个改进的MobilenetV3网络作为骨干网络来提取掩模的特征。与级联前改进的MobilenetV3神经网络的分类结果相比,级联学习网络的分类精度提高了7%,可以看出级联网络的优异性能。
{"title":"Transfer learning based cascaded deep learning network and mask recognition for COVID-19.","authors":"Fengyin Li,&nbsp;Xiaojiao Wang,&nbsp;Yuhong Sun,&nbsp;Tao Li,&nbsp;Junrong Ge","doi":"10.1007/s11280-023-01149-z","DOIUrl":"10.1007/s11280-023-01149-z","url":null,"abstract":"<p><p>The COVID-19 is still spreading today, and it has caused great harm to human beings. The system at the entrance of public places such as shopping malls and stations should check whether pedestrians are wearing masks. However, pedestrians often pass the system inspection by wearing cotton masks, scarves, etc. Therefore, the detection system not only needs to check whether pedestrians are wearing masks, but also needs to detect the type of masks. Based on the lightweight network architecture MobilenetV3, this paper proposes a cascaded deep learning network based on transfer learning, and then designs a mask recognition system based on the cascaded deep learning network. By modifying the activation function of the MobilenetV3 output layer and the structure of the model, two MobilenetV3 networks suitable for cascading are obtained. By introducing transfer learning into the training process of two modified MobilenetV3 networks and a multi-task convolutional neural network, the ImagNet underlying parameters of the network models are obtained in advance, which reduces the computational load of the models. The cascaded deep learning network consists of a multi-task convolutional neural network cascaded with these two modified MobilenetV3 networks. A multi-task convolutional neural network is used to detect faces in images, and two modified MobilenetV3 networks are used as the backbone network to extract the features of masks. After comparing with the classification results of the modified MobilenetV3 neural network before cascading, the classification accuracy of the cascading learning network is improved by 7%, and the excellent performance of the cascading network can be seen.</p>","PeriodicalId":49356,"journal":{"name":"World Wide Web-Internet and Web Information Systems","volume":" ","pages":"1-16"},"PeriodicalIF":3.7,"publicationDate":"2023-05-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10214323/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10092910","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neighboring relation enhanced inductive knowledge graph link prediction via meta-learning 邻域关系增强元学习的归纳知识图链接预测
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-05-25 DOI: 10.1007/s11280-023-01168-w
Ben Liu, Miao Peng, Wenjie Xu, Min Peng
{"title":"Neighboring relation enhanced inductive knowledge graph link prediction via meta-learning","authors":"Ben Liu, Miao Peng, Wenjie Xu, Min Peng","doi":"10.1007/s11280-023-01168-w","DOIUrl":"https://doi.org/10.1007/s11280-023-01168-w","url":null,"abstract":"","PeriodicalId":49356,"journal":{"name":"World Wide Web-Internet and Web Information Systems","volume":"67 1","pages":"2909 - 2930"},"PeriodicalIF":3.7,"publicationDate":"2023-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83940560","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Correlation embedding learning with dynamic semantic enhanced sampling for knowledge graph completion 基于动态语义增强采样的关联嵌入学习知识图谱补全
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-05-19 DOI: 10.1007/s11280-023-01167-x
H. Nie, Xiangguo Zhao, Xin Bi, Yuliang Ma, George Y. Yuan
{"title":"Correlation embedding learning with dynamic semantic enhanced sampling for knowledge graph completion","authors":"H. Nie, Xiangguo Zhao, Xin Bi, Yuliang Ma, George Y. Yuan","doi":"10.1007/s11280-023-01167-x","DOIUrl":"https://doi.org/10.1007/s11280-023-01167-x","url":null,"abstract":"","PeriodicalId":49356,"journal":{"name":"World Wide Web-Internet and Web Information Systems","volume":"12 1","pages":"2887 - 2907"},"PeriodicalIF":3.7,"publicationDate":"2023-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81995213","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
An empirical study of pre-trained language models in simple knowledge graph question answering 简单知识图问答中预训练语言模型的实证研究
3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2023-05-17 DOI: 10.1007/s11280-023-01166-y
Nan Hu, Yike Wu, Guilin Qi, Dehai Min, Jiaoyan Chen, Jeff Z Pan, Zafar Ali
Large-scale pre-trained language models (PLMs) such as BERT have recently achieved great success and become a milestone in natural language processing (NLP). It is now the consensus of the NLP community to adopt PLMs as the backbone for downstream tasks. In recent works on knowledge graph question answering (KGQA), BERT or its variants have become necessary in their KGQA models. However, there is still a lack of comprehensive research and comparison of the performance of different PLMs in KGQA. To this end, we summarize two basic KGQA frameworks based on PLMs without additional neural network modules to compare the performance of nine PLMs in terms of accuracy and efficiency. In addition, we present three benchmarks for larger-scale KGs based on the popular SimpleQuestions benchmark to investigate the scalability of PLMs. We carefully analyze the results of all PLMs-based KGQA basic frameworks on these benchmarks and two other popular datasets, WebQuestionSP and FreebaseQA, and find that knowledge distillation techniques and knowledge enhancement methods in PLMs are promising for KGQA. Furthermore, we test ChatGPT ( https://chat.openai.com/ ), which has drawn a great deal of attention in the NLP community, demonstrating its impressive capabilities and limitations in zero-shot KGQA. We have released the code and benchmarks to promote the use of PLMs on KGQA ( https://github.com/aannonymouuss/PLMs-in-Practical-KBQA ).
像BERT这样的大规模预训练语言模型(plm)最近取得了巨大的成功,成为自然语言处理(NLP)的一个里程碑。现在NLP社区的共识是采用plm作为下游任务的骨干。在最近关于知识图谱问答(KGQA)的研究中,BERT或其变体在他们的KGQA模型中已经成为必要。然而,对于不同PLMs在KGQA中的性能,目前还缺乏全面的研究和比较。为此,我们总结了两种基于plm的基本KGQA框架,没有额外的神经网络模块,以比较9种plm在准确性和效率方面的性能。此外,我们在流行的SimpleQuestions基准测试的基础上提出了三个大规模kg的基准测试,以研究plm的可扩展性。我们仔细分析了所有基于PLMs的KGQA基本框架在这些基准和另外两个流行的数据集(WebQuestionSP和FreebaseQA)上的结果,发现PLMs中的知识蒸馏技术和知识增强方法对KGQA很有前景。此外,我们测试了ChatGPT (https://chat.openai.com/),它在NLP社区中引起了很大的关注,展示了它在零射击KGQA中的令人印象深刻的功能和局限性。我们已经发布了代码和基准测试,以促进在KGQA (https://github.com/aannonymouuss/PLMs-in-Practical-KBQA)上使用plm。
{"title":"An empirical study of pre-trained language models in simple knowledge graph question answering","authors":"Nan Hu, Yike Wu, Guilin Qi, Dehai Min, Jiaoyan Chen, Jeff Z Pan, Zafar Ali","doi":"10.1007/s11280-023-01166-y","DOIUrl":"https://doi.org/10.1007/s11280-023-01166-y","url":null,"abstract":"Large-scale pre-trained language models (PLMs) such as BERT have recently achieved great success and become a milestone in natural language processing (NLP). It is now the consensus of the NLP community to adopt PLMs as the backbone for downstream tasks. In recent works on knowledge graph question answering (KGQA), BERT or its variants have become necessary in their KGQA models. However, there is still a lack of comprehensive research and comparison of the performance of different PLMs in KGQA. To this end, we summarize two basic KGQA frameworks based on PLMs without additional neural network modules to compare the performance of nine PLMs in terms of accuracy and efficiency. In addition, we present three benchmarks for larger-scale KGs based on the popular SimpleQuestions benchmark to investigate the scalability of PLMs. We carefully analyze the results of all PLMs-based KGQA basic frameworks on these benchmarks and two other popular datasets, WebQuestionSP and FreebaseQA, and find that knowledge distillation techniques and knowledge enhancement methods in PLMs are promising for KGQA. Furthermore, we test ChatGPT ( https://chat.openai.com/ ), which has drawn a great deal of attention in the NLP community, demonstrating its impressive capabilities and limitations in zero-shot KGQA. We have released the code and benchmarks to promote the use of PLMs on KGQA ( https://github.com/aannonymouuss/PLMs-in-Practical-KBQA ).","PeriodicalId":49356,"journal":{"name":"World Wide Web-Internet and Web Information Systems","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-05-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135861247","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
期刊
World Wide Web-Internet and Web Information Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1