首页 > 最新文献

Proceedings of the Second International Conference on AI-ML Systems最新文献

英文 中文
Large-Scale Entity Extraction from Enterprise Data 从企业数据中大规模抽取实体
Pub Date : 2022-10-12 DOI: 10.1145/3564121.3564818
Rajeev Gupta, Ranganath Kondapally
Adoption of cloud computing by enterprises has exploded in the last decade and most of the applications used by enterprise users have moved to the cloud. These applications include collaboration software(e.g., Word, Excel), instant messaging (e.g., Chat), asynchronous communication (e.g., Email), etc. This has resulted in an exponential increase in the volume of data arising from the interactions of the users with the online applications (such as documents edited, people interacted with, meetings attended, etc.). Activities of a user provide strong insights about her such as meetings attended by the user indicate the set of people the user closely works with and documents edited indicate the topics the user works on, etc. Typically, this data is private and confidential for the enterprise, part of the enterprise, or the individual employee. To provide better experience and assist employees in their activities, it is critical to mine certain entities from this data. In this tutorial, we explain various entities which can be extracted from the enterprise data and assist the employees in their productivity. Specifically, we define and extract various enterprise entities such as tasks, commitments, calendar activity, acronyms, topics, definitions, etc. These entities are extracted using different techniques—tasks and commitments are extracted using intent mining techniques (e.g., sentiment extraction), definitions are extracted using sequence mining techniques, calendars are updated using the user’s flight/hotel booking entities, etc. The entity extraction from enterprise data poses interesting and complex challenge from scalable information extraction point of view: building information extraction models where there is little data to learn from due to privacy and access-control constraints but need highly accurate models to run on a large amount of diverse data from whole of the enterprise. Specifically, we need to overcome the following challenges: Privacy: For legal and trust reasons, individual user’s data should be accessible only to the persons who it is intended to. Thus, we can’t directly apply the openly available techniques used to mine these entities which all require labeled data. Efficiency: As enterprises need to process billions of emails, chats, and other documents every day—different for different users—extraction models need to be very efficient. Scalability: There are a large number of variations in the way information is presented in the enterprise documents. For example, a flight itinerary is represented in different ways by different providers. Definition of the same topic can be expressed differently in different documents. We should be able to extract entities irrespective of the way it is presented in the documents. Multi-lingual: Users are located across geographies, and hence, the information extraction needs to be done across multiple languages. To extract these entities, one needs supervised data. How to get labeled data
在过去十年中,企业对云计算的采用呈爆炸式增长,企业用户使用的大多数应用程序都转移到了云上。这些应用程序包括协作软件(例如。(如Word, Excel),即时通讯(如聊天),异步通信(如电子邮件)等。这导致用户与在线应用程序交互产生的数据量呈指数级增长(例如编辑的文档、与之交互的人员、参加的会议等)。用户的活动提供了关于用户的深刻见解,例如用户参加的会议表明与用户密切合作的一组人,编辑的文档表明用户从事的主题,等等。通常,这些数据对于企业、企业的一部分或单个员工来说是私有和机密的。为了提供更好的体验并帮助员工开展活动,从这些数据中挖掘某些实体至关重要。在本教程中,我们将解释可以从企业数据中提取的各种实体,并帮助员工提高生产力。具体来说,我们定义和提取各种企业实体,如任务、承诺、日历活动、首字母缩略词、主题、定义等。这些实体是使用不同的技术提取的——任务和承诺是使用意图挖掘技术提取的(例如,情感提取),定义是使用序列挖掘技术提取的,日历是使用用户的航班/酒店预订实体更新的等等。从可扩展信息提取的角度来看,从企业数据中提取实体提出了有趣而复杂的挑战:由于隐私和访问控制约束,构建信息提取模型时几乎没有数据可以学习,但需要高度精确的模型来运行来自整个企业的大量不同数据。具体来说,我们需要克服以下挑战:隐私:出于法律和信任的原因,个人用户的数据应该只能被预期的人访问。因此,我们不能直接应用公开可用的技术来挖掘这些都需要标记数据的实体。效率:由于企业每天需要处理数十亿封电子邮件、聊天记录和其他文档,因此提取模型需要非常高效。可伸缩性:在企业文档中显示信息的方式有很多变化。例如,航班行程由不同的提供者以不同的方式表示。同一主题的定义在不同的文档中可以有不同的表达方式。我们应该能够提取实体,而不管它在文档中的呈现方式如何。多语言:用户位于不同的地理位置,因此,信息提取需要跨多种语言完成。要提取这些实体,需要有监督的数据。如何以保护隐私的方式获得标记数据?我们如何用最少的监督数据构建模型?我们有大量的无监督数据。我们提出了从大型无监督数据和小型有监督数据中学习的技术。在各种技术中,用户反馈(例如,点击)被用来改进信息提取模型。在企业环境中很难获得反馈。我们可以使用弱监管吗?我们是否可以采用现成的模型(例如,用于定义分类)并将其细化为企业设置?我们将在企业设置中以更高的精度和召回率介绍所有这些技术。
{"title":"Large-Scale Entity Extraction from Enterprise Data","authors":"Rajeev Gupta, Ranganath Kondapally","doi":"10.1145/3564121.3564818","DOIUrl":"https://doi.org/10.1145/3564121.3564818","url":null,"abstract":"Adoption of cloud computing by enterprises has exploded in the last decade and most of the applications used by enterprise users have moved to the cloud. These applications include collaboration software(e.g., Word, Excel), instant messaging (e.g., Chat), asynchronous communication (e.g., Email), etc. This has resulted in an exponential increase in the volume of data arising from the interactions of the users with the online applications (such as documents edited, people interacted with, meetings attended, etc.). Activities of a user provide strong insights about her such as meetings attended by the user indicate the set of people the user closely works with and documents edited indicate the topics the user works on, etc. Typically, this data is private and confidential for the enterprise, part of the enterprise, or the individual employee. To provide better experience and assist employees in their activities, it is critical to mine certain entities from this data. In this tutorial, we explain various entities which can be extracted from the enterprise data and assist the employees in their productivity. Specifically, we define and extract various enterprise entities such as tasks, commitments, calendar activity, acronyms, topics, definitions, etc. These entities are extracted using different techniques—tasks and commitments are extracted using intent mining techniques (e.g., sentiment extraction), definitions are extracted using sequence mining techniques, calendars are updated using the user’s flight/hotel booking entities, etc. The entity extraction from enterprise data poses interesting and complex challenge from scalable information extraction point of view: building information extraction models where there is little data to learn from due to privacy and access-control constraints but need highly accurate models to run on a large amount of diverse data from whole of the enterprise. Specifically, we need to overcome the following challenges: Privacy: For legal and trust reasons, individual user’s data should be accessible only to the persons who it is intended to. Thus, we can’t directly apply the openly available techniques used to mine these entities which all require labeled data. Efficiency: As enterprises need to process billions of emails, chats, and other documents every day—different for different users—extraction models need to be very efficient. Scalability: There are a large number of variations in the way information is presented in the enterprise documents. For example, a flight itinerary is represented in different ways by different providers. Definition of the same topic can be expressed differently in different documents. We should be able to extract entities irrespective of the way it is presented in the documents. Multi-lingual: Users are located across geographies, and hence, the information extraction needs to be done across multiple languages. To extract these entities, one needs supervised data. How to get labeled data ","PeriodicalId":166150,"journal":{"name":"Proceedings of the Second International Conference on AI-ML Systems","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129738177","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Patch-wise Features for Blur Image Classification 模糊图像分类的补丁特征
Pub Date : 2022-10-12 DOI: 10.1145/3564121.3564138
Sri Charan Kattamuru, Kshitij Agrawal, S. Adhikari, Abhishek Bose, Hemant Misra
Images captured through smartphone cameras often suffer from degradation, blur being one of the major ones, posing a challenge in processing these images for downstream tasks. In this paper we propose low-compute lightweight patch-wise features for image quality assessment. Using our method we can discriminate between blur vs sharp image degradation. To this end, we train a decision-tree-based XGBoost model on various intuitive image features like gray level variance, first and second order gradients, texture features like local binary patterns. Experiments conducted on an open dataset show that the proposed low compute method results in 90.1% mean accuracy on the validation set, which is comparable to the accuracy of a compute-intensive VGG16 network with 94% mean accuracy fine-tuned to this task. To demonstrate the generalizability of our proposed features and model we test the model on BHBID dataset and an internal dataset where we attain accuracy of 98% and 91%, respectively. The proposed method is 10x faster than the VGG16 based model on CPU and scales linearly to the input image size making it suitable to be implemented on low compute edge devices.
通过智能手机相机拍摄的图像通常会出现退化,模糊是其中一个主要问题,这对处理这些图像进行下游任务提出了挑战。在本文中,我们提出了用于图像质量评估的低计算轻量级补丁智能特征。使用我们的方法,我们可以区分模糊和锐利的图像退化。为此,我们在各种直观的图像特征(如灰度方差、一阶和二阶梯度、局部二值模式等纹理特征)上训练了基于决策树的XGBoost模型。在开放数据集上进行的实验表明,所提出的低计算方法在验证集上的平均准确率为90.1%,与计算密集型的VGG16网络的准确率相当,该网络的平均准确率为94%。为了证明我们提出的特征和模型的泛化性,我们在BHBID数据集和内部数据集上测试了模型,我们分别获得了98%和91%的准确率。该方法比基于CPU的VGG16模型快10倍,并随输入图像大小线性缩放,适合在低计算边缘设备上实现。
{"title":"Patch-wise Features for Blur Image Classification","authors":"Sri Charan Kattamuru, Kshitij Agrawal, S. Adhikari, Abhishek Bose, Hemant Misra","doi":"10.1145/3564121.3564138","DOIUrl":"https://doi.org/10.1145/3564121.3564138","url":null,"abstract":"Images captured through smartphone cameras often suffer from degradation, blur being one of the major ones, posing a challenge in processing these images for downstream tasks. In this paper we propose low-compute lightweight patch-wise features for image quality assessment. Using our method we can discriminate between blur vs sharp image degradation. To this end, we train a decision-tree-based XGBoost model on various intuitive image features like gray level variance, first and second order gradients, texture features like local binary patterns. Experiments conducted on an open dataset show that the proposed low compute method results in 90.1% mean accuracy on the validation set, which is comparable to the accuracy of a compute-intensive VGG16 network with 94% mean accuracy fine-tuned to this task. To demonstrate the generalizability of our proposed features and model we test the model on BHBID dataset and an internal dataset where we attain accuracy of 98% and 91%, respectively. The proposed method is 10x faster than the VGG16 based model on CPU and scales linearly to the input image size making it suitable to be implemented on low compute edge devices.","PeriodicalId":166150,"journal":{"name":"Proceedings of the Second International Conference on AI-ML Systems","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115639833","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Accurate and Efficient Channel pruning via Orthogonal Matching Pursuit 基于正交匹配追踪的信道精确高效修剪
Pub Date : 2022-10-12 DOI: 10.1145/3564121.3564139
Kiran Purohit, Anurag Parvathgari, Soumili Das, Sourangshu Bhattacharya
The deeper and wider architectures of recent convolutional neural networks (CNN) are responsible for superior performance in computer vision tasks. However, they also come with an enormous model size and heavy computational cost. Filter pruning (FP) is one of the methods applied to CNNs for compression and acceleration. Various techniques have been recently proposed for filter pruning. We address the limitation of the existing state-of-the-art method and motivate our setup. We develop a novel method for filter selection using sparse approximation of filter weights. We propose an orthogonal matching pursuit (OMP) based algorithm for filter pruning (called FP-OMP). We also propose FP-OMP Search, which address the problem of removal of uniform number of filters from all the layers of a network. FP-OMP Search performs a search over all the layers with a given batch size of filter removal. We evaluate both FP-OMP and FP-OMP Search on benchmark datasets using standard ResNet architectures. Experimental results indicate that FP-OMP Search consistently outperforms the baseline method (LRF) by nearly . We demonstrate both empirically and visually, that FP-OMP Search prunes different number of filters from different layers. Further, timing profile experiments show that FP-OMP improves over the running time of LRF.
最近的卷积神经网络(CNN)的更深和更广泛的架构负责计算机视觉任务的卓越性能。然而,它们也伴随着巨大的模型尺寸和沉重的计算成本。滤波剪枝(FP)是用于cnn压缩和加速的方法之一。最近提出了各种各样的过滤器修剪技术。我们解决了现有的最先进的方法的局限性,并激励我们的设置。本文提出了一种利用滤波器权值的稀疏逼近进行滤波器选择的新方法。提出了一种基于正交匹配追踪(OMP)的滤波剪枝算法(FP-OMP)。我们还提出了FP-OMP搜索,它解决了从网络的所有层中去除均匀数量的过滤器的问题。FP-OMP搜索在所有层上执行具有给定批量过滤器删除大小的搜索。我们使用标准ResNet架构在基准数据集上评估FP-OMP和FP-OMP Search。实验结果表明,FP-OMP搜索始终优于基线方法(LRF)。我们从经验和视觉上证明,FP-OMP搜索从不同的层修剪不同数量的过滤器。此外,时序曲线实验表明,FP-OMP在LRF的运行时间上有所提高。
{"title":"Accurate and Efficient Channel pruning via Orthogonal Matching Pursuit","authors":"Kiran Purohit, Anurag Parvathgari, Soumili Das, Sourangshu Bhattacharya","doi":"10.1145/3564121.3564139","DOIUrl":"https://doi.org/10.1145/3564121.3564139","url":null,"abstract":"The deeper and wider architectures of recent convolutional neural networks (CNN) are responsible for superior performance in computer vision tasks. However, they also come with an enormous model size and heavy computational cost. Filter pruning (FP) is one of the methods applied to CNNs for compression and acceleration. Various techniques have been recently proposed for filter pruning. We address the limitation of the existing state-of-the-art method and motivate our setup. We develop a novel method for filter selection using sparse approximation of filter weights. We propose an orthogonal matching pursuit (OMP) based algorithm for filter pruning (called FP-OMP). We also propose FP-OMP Search, which address the problem of removal of uniform number of filters from all the layers of a network. FP-OMP Search performs a search over all the layers with a given batch size of filter removal. We evaluate both FP-OMP and FP-OMP Search on benchmark datasets using standard ResNet architectures. Experimental results indicate that FP-OMP Search consistently outperforms the baseline method (LRF) by nearly . We demonstrate both empirically and visually, that FP-OMP Search prunes different number of filters from different layers. Further, timing profile experiments show that FP-OMP improves over the running time of LRF.","PeriodicalId":166150,"journal":{"name":"Proceedings of the Second International Conference on AI-ML Systems","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117207241","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
DNN based Adaptive User Pairing and Power Allocation to achieve α-Fairness in NOMA Systems with Imperfections in SIC 基于DNN的自适应用户配对和功率分配在SIC不完善的NOMA系统中实现α-公平
Pub Date : 2022-10-12 DOI: 10.1145/3564121.3565042
Siva Mouni Nemalidinne, Pavan Reddy Manne, Abhinav Kumar, P. K. Upadhyay
Non-orthogonal multiple access (NOMA) technology aided with successive interference cancellation (SIC) is expected to achieve multi-fold improvements in the network capacity. However, the SIC in practice is prone to imperfections and this degrades the achievable gains with NOMA. Additionally, inappropriate user pairing and power allocation in NOMA can adversely affect the fairness between paired users. Hence, the impact of imperfections in SIC and fairness should be considered for user pairing and power allocation in NOMA. Motivated by this, we formulate the user pairing and power allocation to achieve α-fairness among the paired users as an optimization problem. To obtain a feasible solution in practice, we then propose a two-step machine learning-based approach to solve the problem. We use a random forest classifier (RFC) to establish a pairing criterion and a deep neural network (DNN) to allocate the power factors to the NOMA pair. The performance of the proposed supervised learning (SL) models is extensively evaluated and compared with other pre-existing algorithms. We analyze the performance of DNN for varying number of neurons in the hidden layer by considering different activation functions. We show that with 4 neurons in the hidden layer and sigmoid activation function, the trained DNN network outperforms the existing SL algorithms. We then use the trained network and perform Monte-Carlo simulations to quantify the achievable gains. We show that the proposed approach achieves an excellent solution that maximizes fairness and also ensures minimum required data rates for each user. Through extensive numerical evaluations, we show that our proposed two-step machine learning approach outperforms various state-of-the-art algorithms.
借助连续干扰消除(SIC)的非正交多址(NOMA)技术有望实现网络容量的多倍提升。然而,SIC在实践中容易出现缺陷,这降低了NOMA可实现的增益。此外,在NOMA中,不适当的用户配对和权力分配会对配对用户之间的公平性产生不利影响。因此,在NOMA的用户配对和功率分配中,需要考虑SIC缺陷和公平性的影响。基于此,我们将用户配对和功率分配作为优化问题来实现配对用户之间的α-公平。为了在实践中获得可行的解决方案,我们提出了一种基于机器学习的两步方法来解决问题。我们使用随机森林分类器(RFC)建立配对标准,并使用深度神经网络(DNN)将功率因子分配给NOMA对。本文对所提出的监督学习(SL)模型的性能进行了广泛的评估,并与其他已有算法进行了比较。我们通过考虑不同的激活函数来分析深层神经网络在不同隐藏层神经元数量下的性能。我们证明,在隐藏层中使用4个神经元和s型激活函数,训练后的DNN网络优于现有的SL算法。然后,我们使用训练好的网络并执行蒙特卡罗模拟来量化可实现的增益。我们表明,所提出的方法实现了一个很好的解决方案,最大限度地提高了公平性,并确保了每个用户所需的最小数据速率。通过广泛的数值评估,我们表明我们提出的两步机器学习方法优于各种最先进的算法。
{"title":"DNN based Adaptive User Pairing and Power Allocation to achieve α-Fairness in NOMA Systems with Imperfections in SIC","authors":"Siva Mouni Nemalidinne, Pavan Reddy Manne, Abhinav Kumar, P. K. Upadhyay","doi":"10.1145/3564121.3565042","DOIUrl":"https://doi.org/10.1145/3564121.3565042","url":null,"abstract":"Non-orthogonal multiple access (NOMA) technology aided with successive interference cancellation (SIC) is expected to achieve multi-fold improvements in the network capacity. However, the SIC in practice is prone to imperfections and this degrades the achievable gains with NOMA. Additionally, inappropriate user pairing and power allocation in NOMA can adversely affect the fairness between paired users. Hence, the impact of imperfections in SIC and fairness should be considered for user pairing and power allocation in NOMA. Motivated by this, we formulate the user pairing and power allocation to achieve α-fairness among the paired users as an optimization problem. To obtain a feasible solution in practice, we then propose a two-step machine learning-based approach to solve the problem. We use a random forest classifier (RFC) to establish a pairing criterion and a deep neural network (DNN) to allocate the power factors to the NOMA pair. The performance of the proposed supervised learning (SL) models is extensively evaluated and compared with other pre-existing algorithms. We analyze the performance of DNN for varying number of neurons in the hidden layer by considering different activation functions. We show that with 4 neurons in the hidden layer and sigmoid activation function, the trained DNN network outperforms the existing SL algorithms. We then use the trained network and perform Monte-Carlo simulations to quantify the achievable gains. We show that the proposed approach achieves an excellent solution that maximizes fairness and also ensures minimum required data rates for each user. Through extensive numerical evaluations, we show that our proposed two-step machine learning approach outperforms various state-of-the-art algorithms.","PeriodicalId":166150,"journal":{"name":"Proceedings of the Second International Conference on AI-ML Systems","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121601738","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CluSpa: Computation Reduction in CNN Inference by exploiting Clustering and Sparsity CluSpa:利用聚类和稀疏性减少CNN推理的计算量
Pub Date : 2022-10-12 DOI: 10.1145/3564121.3564132
Imlijungla Longchar, Amey Varhade, Chetan Ingle, Saurabh Baranwal, H. Kapoor
Convolutional Neural Networks (CNNs) have grown in popularity and usage tremendously over the last few years, spanning across different task such as computer vision tasks, natural language processing, video recognition, and recommender systems. Despite the algorithmic advancements that drove the growth of CNN still has considerable computational and memory overhead that poses challenges in achieving real-time performance. Each input image requires millions to even billions of elementary arithmetic operations before the network obtains the result. In CNNs, convolutional and pooling layers are followed by activation layers involving various activation functions. Hence, a lot of work has been done to reduce these costs in the last few years. Numerous optimizations have addressed at both hardware and software levels. In this paper, we propose a software-based solution for improving the performance of inference of networks. We suggest a technique for the approximate computation of the convolution operation based on clustering and sharing of weights. We have utilized Gaussian Mixture Models for clustering. We exploit weight sparsity to further reduce computations on top of the clustering method. We were able to achieve a considerable reduction in the MAC operations and the overall computation speedup on popular CNN architectures
卷积神经网络(cnn)在过去几年中得到了极大的普及和使用,跨越了不同的任务,如计算机视觉任务、自然语言处理、视频识别和推荐系统。尽管算法的进步推动了CNN的增长,但它仍然有相当大的计算和内存开销,这给实现实时性能带来了挑战。在网络获得结果之前,每个输入图像需要数百万甚至数十亿的基本算术运算。在cnn中,卷积层和池化层之后是涉及各种激活函数的激活层。因此,在过去的几年里,人们做了很多工作来降低这些成本。在硬件和软件级别都进行了许多优化。在本文中,我们提出了一种基于软件的解决方案来提高网络的推理性能。我们提出了一种基于聚类和权值共享的卷积运算近似计算技术。我们使用高斯混合模型进行聚类。我们利用权值稀疏性在聚类方法的基础上进一步减少计算量。我们能够在流行的CNN架构上实现MAC操作的大幅减少和整体计算速度的提高
{"title":"CluSpa: Computation Reduction in CNN Inference by exploiting Clustering and Sparsity","authors":"Imlijungla Longchar, Amey Varhade, Chetan Ingle, Saurabh Baranwal, H. Kapoor","doi":"10.1145/3564121.3564132","DOIUrl":"https://doi.org/10.1145/3564121.3564132","url":null,"abstract":"Convolutional Neural Networks (CNNs) have grown in popularity and usage tremendously over the last few years, spanning across different task such as computer vision tasks, natural language processing, video recognition, and recommender systems. Despite the algorithmic advancements that drove the growth of CNN still has considerable computational and memory overhead that poses challenges in achieving real-time performance. Each input image requires millions to even billions of elementary arithmetic operations before the network obtains the result. In CNNs, convolutional and pooling layers are followed by activation layers involving various activation functions. Hence, a lot of work has been done to reduce these costs in the last few years. Numerous optimizations have addressed at both hardware and software levels. In this paper, we propose a software-based solution for improving the performance of inference of networks. We suggest a technique for the approximate computation of the convolution operation based on clustering and sharing of weights. We have utilized Gaussian Mixture Models for clustering. We exploit weight sparsity to further reduce computations on top of the clustering method. We were able to achieve a considerable reduction in the MAC operations and the overall computation speedup on popular CNN architectures","PeriodicalId":166150,"journal":{"name":"Proceedings of the Second International Conference on AI-ML Systems","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115153883","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient Vector Store System for Python using Shared Memory 使用共享内存的高效Python矢量存储系统
Pub Date : 2022-10-12 DOI: 10.1145/3564121.3564799
Dhruv Patel, S. Pandey, Abhishek Sharma
Many e-commerce companies use machine learning to make customer experience better. Even within a single company, there will be generally many independent services running, each specializing in some aspect of customer experience. Since machine learning models work on abstract vectors representing users and/or items, each such service needs a way to store these vectors. A common approach nowadays is to save them in in-memory caches like Memcached. As these caches run in their own processes, and Machine Learning services generally run as Python services, there is a communication overhead involved for each request that ML service serves. One can reduce this overhead by directly storing these vectors in a Python dictionary within the service. To support concurrency and scale, a single node runs multiple instances of the same service. Thus, we also want to avoid duplicating these vectors across multiple processes. In this paper, we propose a system to store vectors in shared memory and efficiently serve all concurrent instances of the service, without replicating the vectors themselves. We achieve up to 4.5x improvements in latency compared to Memcached. Additionally, due to availability of more memory, we can increase the number of server processes running in each node, translating into greater throughput. We also discuss the impact of the proposed method (towards increasing the throughput) in live production scenario.
许多电子商务公司使用机器学习来改善客户体验。即使在一个公司内部,通常也会有许多独立的服务在运行,每个服务都专注于客户体验的某些方面。由于机器学习模型处理代表用户和/或项目的抽象向量,因此每个这样的服务都需要一种存储这些向量的方法。现在一种常见的方法是将它们保存在内存缓存中,比如Memcached。由于这些缓存在它们自己的进程中运行,而机器学习服务通常作为Python服务运行,因此ML服务所处理的每个请求都涉及通信开销。可以通过直接将这些向量存储在服务中的Python字典中来减少这种开销。为了支持并发性和可伸缩性,单个节点运行同一服务的多个实例。因此,我们还希望避免在多个进程中重复这些向量。在本文中,我们提出了一种将向量存储在共享内存中并有效地服务于所有并发服务实例的系统,而无需复制向量本身。与Memcached相比,我们的延迟提高了4.5倍。此外,由于更多内存的可用性,我们可以增加每个节点上运行的服务器进程的数量,从而转化为更大的吞吐量。我们还讨论了所提出的方法(对提高吞吐量)在现场生产场景中的影响。
{"title":"Efficient Vector Store System for Python using Shared Memory","authors":"Dhruv Patel, S. Pandey, Abhishek Sharma","doi":"10.1145/3564121.3564799","DOIUrl":"https://doi.org/10.1145/3564121.3564799","url":null,"abstract":"Many e-commerce companies use machine learning to make customer experience better. Even within a single company, there will be generally many independent services running, each specializing in some aspect of customer experience. Since machine learning models work on abstract vectors representing users and/or items, each such service needs a way to store these vectors. A common approach nowadays is to save them in in-memory caches like Memcached. As these caches run in their own processes, and Machine Learning services generally run as Python services, there is a communication overhead involved for each request that ML service serves. One can reduce this overhead by directly storing these vectors in a Python dictionary within the service. To support concurrency and scale, a single node runs multiple instances of the same service. Thus, we also want to avoid duplicating these vectors across multiple processes. In this paper, we propose a system to store vectors in shared memory and efficiently serve all concurrent instances of the service, without replicating the vectors themselves. We achieve up to 4.5x improvements in latency compared to Memcached. Additionally, due to availability of more memory, we can increase the number of server processes running in each node, translating into greater throughput. We also discuss the impact of the proposed method (towards increasing the throughput) in live production scenario.","PeriodicalId":166150,"journal":{"name":"Proceedings of the Second International Conference on AI-ML Systems","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123190322","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Tutorial: Neuro-symbolic AI for Mental Healthcare 教程:用于精神保健的神经符号AI
Pub Date : 2022-10-12 DOI: 10.1145/3564121.3564817
Kaushik Roy, Usha Lokala, Manas Gaur, Amit P. Sheth
Artificial Intelligence (AI) systems for mental healthcare (MHCare) have been ever-growing after realizing the importance of early interventions for patients with chronic mental health (MH) conditions. Social media (SocMedia) emerged as the go-to platform for supporting patients seeking MHCare. The creation of peer-support groups without social stigma has resulted in patients transitioning from clinical settings to SocMedia supported interactions for quick help. Researchers started exploring SocMedia content in search of cues that showcase correlation or causation between different MH conditions to design better interventional strategies. User-level Classification-based AI systems were designed to leverage diverse SocMedia data from various MH conditions, to predict MH conditions. Subsequently, researchers created classification schemes to measure the severity of each MH condition. Such ad-hoc schemes, engineered features, and models not only require a large amount of data but fail to allow clinically acceptable and explainable reasoning over the outcomes. To improve Neural-AI for MHCare, infusion of clinical symbolic knowledge that clinicans use in decision making is required. An impactful use case of Neural-AI systems in MH is conversational systems. These systems require coordination between classification and generation to facilitate humanistic conversation in conversational agents (CA). Current CAs with deep language models lack factual correctness, medical relevance, and safety in their generations, which intertwine with unexplainable statistical classification techniques. This lecture-style tutorial will demonstrate our investigations into Neuro-symbolic methods of infusing clinical knowledge to improve the outcomes of Neural-AI systems to improve interventions for MHCare:(a) We will discuss the use of diverse clinical knowledge in creating specialized datasets to train Neural-AI systems effectively. (b) Patients with cardiovascular disease express MH symptoms differently based on gender differences. We will show that knowledge-infused Neural-AI systems can identify gender-specific MH symptoms in such patients. (c) We will describe strategies for infusing clinical process knowledge as heuristics and constraints to improve language models in generating relevant questions and responses.
在意识到早期干预对慢性精神健康(MH)患者的重要性后,用于精神卫生保健的人工智能(AI)系统(MHCare)一直在不断发展。社交媒体(SocMedia)成为支持患者寻求MHCare的首选平台。没有社会耻辱的同伴支持小组的创建导致患者从临床环境过渡到SocMedia支持的快速帮助互动。研究人员开始探索社会媒体内容,寻找显示不同MH条件之间相关性或因果关系的线索,以设计更好的干预策略。基于用户级别分类的人工智能系统旨在利用来自各种MH条件的各种SocMedia数据来预测MH条件。随后,研究人员创建了分类方案来衡量每种MH病情的严重程度。这种临时方案、工程特征和模型不仅需要大量的数据,而且无法对结果进行临床可接受和可解释的推理。为了改善MHCare的神经人工智能,需要注入临床医生在决策中使用的临床符号知识。在MH中,神经人工智能系统的一个有影响力的用例是会话系统。这些系统需要在分类和生成之间进行协调,以促进对话代理(CA)中的人文对话。当前具有深度语言模型的ca在其世代中缺乏事实正确性、医学相关性和安全性,这些与无法解释的统计分类技术交织在一起。本讲座式教程将展示我们对注入临床知识以改善神经-人工智能系统结果的神经符号方法的研究,以改善MHCare的干预措施:(a)我们将讨论在创建专门数据集以有效训练神经-人工智能系统时使用不同的临床知识。(b)心血管疾病患者因性别差异而表现出不同的MH症状。我们将展示知识注入的神经人工智能系统可以识别此类患者的性别特异性MH症状。(c)我们将描述注入临床过程知识的策略,作为启发和约束,以改进生成相关问题和反应的语言模型。
{"title":"Tutorial: Neuro-symbolic AI for Mental Healthcare","authors":"Kaushik Roy, Usha Lokala, Manas Gaur, Amit P. Sheth","doi":"10.1145/3564121.3564817","DOIUrl":"https://doi.org/10.1145/3564121.3564817","url":null,"abstract":"Artificial Intelligence (AI) systems for mental healthcare (MHCare) have been ever-growing after realizing the importance of early interventions for patients with chronic mental health (MH) conditions. Social media (SocMedia) emerged as the go-to platform for supporting patients seeking MHCare. The creation of peer-support groups without social stigma has resulted in patients transitioning from clinical settings to SocMedia supported interactions for quick help. Researchers started exploring SocMedia content in search of cues that showcase correlation or causation between different MH conditions to design better interventional strategies. User-level Classification-based AI systems were designed to leverage diverse SocMedia data from various MH conditions, to predict MH conditions. Subsequently, researchers created classification schemes to measure the severity of each MH condition. Such ad-hoc schemes, engineered features, and models not only require a large amount of data but fail to allow clinically acceptable and explainable reasoning over the outcomes. To improve Neural-AI for MHCare, infusion of clinical symbolic knowledge that clinicans use in decision making is required. An impactful use case of Neural-AI systems in MH is conversational systems. These systems require coordination between classification and generation to facilitate humanistic conversation in conversational agents (CA). Current CAs with deep language models lack factual correctness, medical relevance, and safety in their generations, which intertwine with unexplainable statistical classification techniques. This lecture-style tutorial will demonstrate our investigations into Neuro-symbolic methods of infusing clinical knowledge to improve the outcomes of Neural-AI systems to improve interventions for MHCare:(a) We will discuss the use of diverse clinical knowledge in creating specialized datasets to train Neural-AI systems effectively. (b) Patients with cardiovascular disease express MH symptoms differently based on gender differences. We will show that knowledge-infused Neural-AI systems can identify gender-specific MH symptoms in such patients. (c) We will describe strategies for infusing clinical process knowledge as heuristics and constraints to improve language models in generating relevant questions and responses.","PeriodicalId":166150,"journal":{"name":"Proceedings of the Second International Conference on AI-ML Systems","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126691301","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Efficient Graph based Recommender System with Weighted Averaging of Messages 基于消息加权平均的高效图推荐系统
Pub Date : 2022-09-30 DOI: 10.1145/3564121.3564127
Faizan Ahemad
We showcase a novel solution to a recommendation system problem where we face a perpetual soft item cold start issue. Our system aims to recommend demanded products to prospective sellers for listing in Amazon stores. These products always have only few interactions thereby giving rise to a perpetual soft item cold start situation. Modern collaborative filtering methods solve cold start using content attributes and exploit the existing implicit signals from warm start items. This approach fails in our use-case since our entire item set faces cold start issue always. Our Product Graph has over 500 Million nodes and over 5 Billion edges which makes training and inference using modern graph algorithms very compute intensive. To overcome these challenges we propose a system which reduces the dataset size and employs an improved modelling technique to reduce storage and compute without loss in performance. Particularly, we reduce our graph size using a filtering technique and then exploit this reduced product graph using Weighted Averaging of Messages over Layers (WAML) algorithm. WAML simplifies training on large graphs and improves over previous methods by reducing compute time to of LightGCN [8] and of Graph Attention Network (GAT) [20] and increasing recall@100 by over LightGCN and over GAT.
我们展示了一个新颖的解决方案,以推荐系统的问题,我们面临一个永久的软项目冷启动问题。我们的系统旨在向潜在卖家推荐需要的产品,以便在亚马逊商店上架。这些产品总是只有很少的交互,从而产生了一个永久的软项目冷启动情况。现代协同过滤方法利用内容属性解决冷启动问题,并利用热启动项中存在的隐式信号。这种方法在我们的用例中失败了,因为我们的整个项目集总是面临冷启动问题。我们的产品图有超过5亿个节点和超过50亿个边,这使得使用现代图算法进行训练和推理的计算非常密集。为了克服这些挑战,我们提出了一个减少数据集大小的系统,并采用改进的建模技术来减少存储和计算而不损失性能。特别是,我们使用过滤技术减少了图的大小,然后使用层间消息加权平均(WAML)算法利用这个减少的积图。WAML通过减少LightGCN[8]和图注意网络(GAT)[20]的计算时间来简化大图上的训练,并通过比LightGCN和比GAT增加recall@100来改进以前的方法。
{"title":"Efficient Graph based Recommender System with Weighted Averaging of Messages","authors":"Faizan Ahemad","doi":"10.1145/3564121.3564127","DOIUrl":"https://doi.org/10.1145/3564121.3564127","url":null,"abstract":"We showcase a novel solution to a recommendation system problem where we face a perpetual soft item cold start issue. Our system aims to recommend demanded products to prospective sellers for listing in Amazon stores. These products always have only few interactions thereby giving rise to a perpetual soft item cold start situation. Modern collaborative filtering methods solve cold start using content attributes and exploit the existing implicit signals from warm start items. This approach fails in our use-case since our entire item set faces cold start issue always. Our Product Graph has over 500 Million nodes and over 5 Billion edges which makes training and inference using modern graph algorithms very compute intensive. To overcome these challenges we propose a system which reduces the dataset size and employs an improved modelling technique to reduce storage and compute without loss in performance. Particularly, we reduce our graph size using a filtering technique and then exploit this reduced product graph using Weighted Averaging of Messages over Layers (WAML) algorithm. WAML simplifies training on large graphs and improves over previous methods by reducing compute time to of LightGCN [8] and of Graph Attention Network (GAT) [20] and increasing recall@100 by over LightGCN and over GAT.","PeriodicalId":166150,"journal":{"name":"Proceedings of the Second International Conference on AI-ML Systems","volume":"63 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115668040","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unsupervised Early Exit in DNNs with Multiple Exits 多出口dnn的无监督提前退出
Pub Date : 2022-09-20 DOI: 10.1145/3564121.3564137
U. HariNarayanN, M. Hanawal, Avinash Bhardwaj
Deep Neural Networks (DNNs) are generally designed as sequentially cascaded differentiable blocks/layers with a prediction module connected only to its last layer. DNNs can be attached with prediction modules at multiple points along the backbone where inference can stop at an intermediary stage without passing through all the modules. The last exit point may offer a better prediction error but also involves more computational resources and latency. An exit point that is ‘optimal’ in terms of both prediction error and cost is desirable. The optimal exit point may depend on the latent distribution of the tasks and may change from one task type to another. During neural inference, the ground truth of instances may not be available and the error rates at each exit point cannot be estimated. Hence one is faced with the problem of selecting the optimal exit in an unsupervised setting. Prior works tackled this problem in an offline supervised setting assuming that enough labeled data is available to estimate the error rate at each exit point and tune the parameters for better accuracy. However, pre-trained DNNs are often deployed in new domains for which a large amount of ground truth may not be available. We thus model the problem of exit selection as an unsupervised online learning problem and leverage the bandit theory to identify the optimal exit point. Specifically, we focus on the Elastic BERT, a pre-trained multi-exit DNN to demonstrate that it ‘nearly’ satisfies the Strong Dominance (SD) property making it possible to learn the optimal exit in an online setup without knowing the ground truth labels. We develop upper confidence bound (UCB) based algorithm named UEE-UCB that provably achieves sub-linear regret under the SD property. Thus our method provides a means to adaptively learn domain-specific optimal exit points in multi-exit DNNs. We empirically validate our algorithm on IMDb and Yelp datasets.
深度神经网络(dnn)通常被设计为顺序级联的可微块/层,预测模块仅连接到其最后一层。dnn可以在主干的多个点上附加预测模块,推理可以在中间阶段停止,而无需通过所有模块。最后一个退出点可能提供更好的预测误差,但也涉及更多的计算资源和延迟。一个在预测误差和成本方面都是“最优”的退出点是可取的。最优退出点可能取决于任务的潜在分布,并可能因任务类型的不同而变化。在神经推理过程中,实例的基础真值可能不可用,并且无法估计每个退出点的错误率。因此,人们面临着在无监督环境下选择最优出口的问题。先前的工作在离线监督设置中解决了这个问题,假设有足够的标记数据可用来估计每个出口点的错误率,并调整参数以获得更好的准确性。然而,预训练的dnn通常部署在可能无法获得大量基础真值的新领域。因此,我们将退出选择问题建模为无监督在线学习问题,并利用强盗理论来确定最优退出点。具体来说,我们关注弹性BERT,这是一个预训练的多出口深度神经网络,以证明它“几乎”满足强优势(SD)属性,使得在不知道基础真值标签的情况下学习在线设置中的最佳出口成为可能。我们提出了基于上置信度界(UCB)的UEE-UCB算法,该算法可证明在SD属性下实现亚线性遗憾。因此,我们的方法提供了一种在多出口深度神经网络中自适应学习特定领域的最佳出口点的方法。我们在IMDb和Yelp数据集上验证了我们的算法。
{"title":"Unsupervised Early Exit in DNNs with Multiple Exits","authors":"U. HariNarayanN, M. Hanawal, Avinash Bhardwaj","doi":"10.1145/3564121.3564137","DOIUrl":"https://doi.org/10.1145/3564121.3564137","url":null,"abstract":"Deep Neural Networks (DNNs) are generally designed as sequentially cascaded differentiable blocks/layers with a prediction module connected only to its last layer. DNNs can be attached with prediction modules at multiple points along the backbone where inference can stop at an intermediary stage without passing through all the modules. The last exit point may offer a better prediction error but also involves more computational resources and latency. An exit point that is ‘optimal’ in terms of both prediction error and cost is desirable. The optimal exit point may depend on the latent distribution of the tasks and may change from one task type to another. During neural inference, the ground truth of instances may not be available and the error rates at each exit point cannot be estimated. Hence one is faced with the problem of selecting the optimal exit in an unsupervised setting. Prior works tackled this problem in an offline supervised setting assuming that enough labeled data is available to estimate the error rate at each exit point and tune the parameters for better accuracy. However, pre-trained DNNs are often deployed in new domains for which a large amount of ground truth may not be available. We thus model the problem of exit selection as an unsupervised online learning problem and leverage the bandit theory to identify the optimal exit point. Specifically, we focus on the Elastic BERT, a pre-trained multi-exit DNN to demonstrate that it ‘nearly’ satisfies the Strong Dominance (SD) property making it possible to learn the optimal exit in an online setup without knowing the ground truth labels. We develop upper confidence bound (UCB) based algorithm named UEE-UCB that provably achieves sub-linear regret under the SD property. Thus our method provides a means to adaptively learn domain-specific optimal exit points in multi-exit DNNs. We empirically validate our algorithm on IMDb and Yelp datasets.","PeriodicalId":166150,"journal":{"name":"Proceedings of the Second International Conference on AI-ML Systems","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124561232","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Proceedings of the Second International Conference on AI-ML Systems 第二届AI-ML系统国际会议论文集
Pub Date : 1900-01-01 DOI: 10.1145/3564121
{"title":"Proceedings of the Second International Conference on AI-ML Systems","authors":"","doi":"10.1145/3564121","DOIUrl":"https://doi.org/10.1145/3564121","url":null,"abstract":"","PeriodicalId":166150,"journal":{"name":"Proceedings of the Second International Conference on AI-ML Systems","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128522657","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the Second International Conference on AI-ML Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1