首页 > 最新文献

2018 30th International Symposium on Computer Architecture and High Performance Computing (SBAC-PAD)最新文献

英文 中文
Scaling and Optimizing the Gysela Code on a Cluster of Many-Core Processors Gysela代码在多核处理器集群上的扩展和优化
G. Latu, Y. Asahi, Julien Bigot, Tamas B. Fehér, V. Grandgirard
The current generation of the Xeon Phi Knights Landing (KNL) processor provides a highly multi-threaded environment on which regular programming models such as MPIjopenMP can be used. Many factors impact the performance achieved by applications on these devices: one of the key points is the efficient exploitation of SIMD vector units, and one another is the memory access pattern. Works have been conducted to adapt a plasma turbulence application, namely Gysela, for this architecture. A set of different techniques have been used: standard vectorization techniques, auto-tuning of one computation kernel, switching to high-order scheme. As a result, KNL execution times have been reduced by up to a factor 3. This effort has also permitted to gain a speedup of 2x on Broadwell architecture and 3x on Skylake. Nice scalability curves up to a few thousands cores have been obtained on a strong scaling experiment. Incremental work meant a large payoff without resorting to using low-level intrinsics.
当前一代的Xeon Phi Knights Landing (KNL)处理器提供了一个高度多线程的环境,可以在该环境上使用常规编程模型,如MPIjopenMP。许多因素会影响应用程序在这些设备上实现的性能:其中一个关键点是SIMD向量单元的有效利用,另一个是内存访问模式。工作已经进行,以适应等离子体湍流应用程序,即Gysela,为这种架构。使用了一系列不同的技术:标准矢量化技术,一个计算内核的自动调优,切换到高阶方案。因此,KNL的执行时间最多减少了1 / 3。这一努力也使Broadwell架构的速度提高了2倍,Skylake的速度提高了3倍。在一个强大的扩展实验中获得了数千核的良好可扩展性曲线。增量工作意味着无需使用低级内在机制就能获得巨大回报。
{"title":"Scaling and Optimizing the Gysela Code on a Cluster of Many-Core Processors","authors":"G. Latu, Y. Asahi, Julien Bigot, Tamas B. Fehér, V. Grandgirard","doi":"10.1109/CAHPC.2018.8645933","DOIUrl":"https://doi.org/10.1109/CAHPC.2018.8645933","url":null,"abstract":"The current generation of the Xeon Phi Knights Landing (KNL) processor provides a highly multi-threaded environment on which regular programming models such as MPIjopenMP can be used. Many factors impact the performance achieved by applications on these devices: one of the key points is the efficient exploitation of SIMD vector units, and one another is the memory access pattern. Works have been conducted to adapt a plasma turbulence application, namely Gysela, for this architecture. A set of different techniques have been used: standard vectorization techniques, auto-tuning of one computation kernel, switching to high-order scheme. As a result, KNL execution times have been reduced by up to a factor 3. This effort has also permitted to gain a speedup of 2x on Broadwell architecture and 3x on Skylake. Nice scalability curves up to a few thousands cores have been obtained on a strong scaling experiment. Incremental work meant a large payoff without resorting to using low-level intrinsics.","PeriodicalId":307747,"journal":{"name":"2018 30th International Symposium on Computer Architecture and High Performance Computing (SBAC-PAD)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121646048","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Towards Green Scientific Data Compression Through High-Level I/O Interfaces 通过高级I/O接口实现绿色科学数据压缩
Yevhen Alforov, T. Ludwig, Anastasiia Novikova, Michael Kuhn, J. Kunkel
Every HPC system today has to cope with a deluge of data generated by scientific applications, simulations or large-scale experiments. The upscaling of supercomputer systems and infrastructures, generally results in a dramatic increase of their energy consumption. In this paper, we argue that techniques like data compression can lead to significant gains in terms of power efficiency by reducing both network and storage requirements. However, any data reduction is highly data specific and should comply with established requirements. Therefore, unsuitable or inappropriate compression strategy can utilize more resources and energy than necessary. To that end, we propose a novel methodology for achieving on-the-fly intelligent determination of energy efficient data reduction for a given data set by leveraging state-of-the-art compression algorithms and meta data at application-level I/O. We motivate our work by analyzing the energy and storage saving needs of data sets from real-life scientific HPC applications, and review the various lossless compression techniques that can be applied. We find that the resulting data reduction can decrease the data volume transferred and stored by as much as 80 % in some cases, consequently leading to significant savings in storage and networking costs.
如今,每个HPC系统都必须处理由科学应用、模拟或大规模实验产生的海量数据。超级计算机系统和基础设施的升级通常会导致其能源消耗的急剧增加。在本文中,我们认为数据压缩等技术可以通过减少网络和存储需求来显著提高功率效率。但是,任何数据缩减都是高度特定于数据的,并且应该符合既定的要求。因此,不合适或不适当的压缩策略会占用不必要的资源和能量。为此,我们提出了一种新的方法,通过利用最先进的压缩算法和应用程序级I/O的元数据,实现对给定数据集的节能数据减少的动态智能确定。我们通过分析来自现实科学高性能计算应用的数据集的能量和存储节约需求来激励我们的工作,并回顾了各种可以应用的无损压缩技术。我们发现,在某些情况下,由此产生的数据减少可以将传输和存储的数据量减少多达80%,从而大大节省了存储和网络成本。
{"title":"Towards Green Scientific Data Compression Through High-Level I/O Interfaces","authors":"Yevhen Alforov, T. Ludwig, Anastasiia Novikova, Michael Kuhn, J. Kunkel","doi":"10.1109/CAHPC.2018.8645921","DOIUrl":"https://doi.org/10.1109/CAHPC.2018.8645921","url":null,"abstract":"Every HPC system today has to cope with a deluge of data generated by scientific applications, simulations or large-scale experiments. The upscaling of supercomputer systems and infrastructures, generally results in a dramatic increase of their energy consumption. In this paper, we argue that techniques like data compression can lead to significant gains in terms of power efficiency by reducing both network and storage requirements. However, any data reduction is highly data specific and should comply with established requirements. Therefore, unsuitable or inappropriate compression strategy can utilize more resources and energy than necessary. To that end, we propose a novel methodology for achieving on-the-fly intelligent determination of energy efficient data reduction for a given data set by leveraging state-of-the-art compression algorithms and meta data at application-level I/O. We motivate our work by analyzing the energy and storage saving needs of data sets from real-life scientific HPC applications, and review the various lossless compression techniques that can be applied. We find that the resulting data reduction can decrease the data volume transferred and stored by as much as 80 % in some cases, consequently leading to significant savings in storage and networking costs.","PeriodicalId":307747,"journal":{"name":"2018 30th International Symposium on Computer Architecture and High Performance Computing (SBAC-PAD)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121858239","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Large Scale Language Modeling: Converging on 40GB of Text in Four Hours 大规模语言建模:在4小时内收敛40GB文本
Raul Puri, Robert Kirby, Nikolai Yakovenko, Bryan Catanzaro
Recent work has shown how to train Convolutional Neural Networks (CNNs) rapidly on large image datasets [1], then transfer the knowledge gained from these models to a variety of tasks [2]. Following [3], in this work, we demonstrate similar scalability and transfer for Recurrent Neural Networks (RNNs) for Natural Language tasks. By utilizing mixed precision arithmetic and a 32k batch size distributed across 128 NVIDIA Tesla V100 GPUs, we are able to train a character-level 4096-dimension multiplicative LSTM (mLSTM) [4] for unsupervised text reconstruction over 3 epochs of the 40 GB Amazon Reviews dataset [5] in four hours. This runtime compares favorably with previous work taking one month to train the same size and configuration for one epoch over the same dataset [3]. Converging large batch RNN models can be challenging. Recent work has suggested scaling the learning rate as a function of batch size, but we find that simply scaling the learning rate as a function of batch size leads either to significantly worse convergence or immediate divergence for this problem. We provide a learning rate schedule that allows our model to converge with a 32k batch size. Since our model converges over the Amazon Reviews dataset in hours, and our compute requirement of 128 Tesla V100 GPUs, while substantial, is commercially available, this work opens up large scale unsupervised NLP training to most commercial applications and deep learning researchers 11Our code is publicly available: https://github.com/NVIDIA/sentiment-discovery, A model can be trained over most public or private text datasets overnight.
最近的研究展示了如何在大型图像数据集上快速训练卷积神经网络(cnn)[1],然后将从这些模型中获得的知识转移到各种任务中[2]。接下来[3],在这项工作中,我们展示了递归神经网络(rnn)用于自然语言任务的类似可扩展性和迁移。通过使用混合精度算法和分布在128个NVIDIA Tesla V100 gpu上的32k批处理大小,我们能够在4小时内训练一个字符级4096维乘法LSTM (mLSTM)[4],用于在40gb Amazon Reviews数据集[5]的3个epoch上进行无监督文本重建。这个运行时与之前的工作相比,在相同的数据集上为一个epoch训练相同的大小和配置需要一个月的时间[3]。收敛大批量RNN模型可能具有挑战性。最近的研究表明,将学习率作为批量大小的函数进行缩放,但我们发现,简单地将学习率作为批量大小的函数进行缩放,要么会导致这个问题的收敛性显著恶化,要么会立即出现分歧。我们提供了一个学习率计划,允许我们的模型收敛于32k批处理大小。由于我们的模型在几个小时内就能在亚马逊评论数据集上收敛,而且我们对128个特斯拉V100 gpu的计算需求虽然很大,但在商业上是可用的,这项工作为大多数商业应用程序和深度学习研究人员打开了大规模无监督NLP训练的道路。我们的代码是公开的:https://github.com/NVIDIA/sentiment-discovery,一个模型可以在一夜之间在大多数公共或私人文本数据集上训练。
{"title":"Large Scale Language Modeling: Converging on 40GB of Text in Four Hours","authors":"Raul Puri, Robert Kirby, Nikolai Yakovenko, Bryan Catanzaro","doi":"10.1109/CAHPC.2018.8645935","DOIUrl":"https://doi.org/10.1109/CAHPC.2018.8645935","url":null,"abstract":"Recent work has shown how to train Convolutional Neural Networks (CNNs) rapidly on large image datasets [1], then transfer the knowledge gained from these models to a variety of tasks [2]. Following [3], in this work, we demonstrate similar scalability and transfer for Recurrent Neural Networks (RNNs) for Natural Language tasks. By utilizing mixed precision arithmetic and a 32k batch size distributed across 128 NVIDIA Tesla V100 GPUs, we are able to train a character-level 4096-dimension multiplicative LSTM (mLSTM) [4] for unsupervised text reconstruction over 3 epochs of the 40 GB Amazon Reviews dataset [5] in four hours. This runtime compares favorably with previous work taking one month to train the same size and configuration for one epoch over the same dataset [3]. Converging large batch RNN models can be challenging. Recent work has suggested scaling the learning rate as a function of batch size, but we find that simply scaling the learning rate as a function of batch size leads either to significantly worse convergence or immediate divergence for this problem. We provide a learning rate schedule that allows our model to converge with a 32k batch size. Since our model converges over the Amazon Reviews dataset in hours, and our compute requirement of 128 Tesla V100 GPUs, while substantial, is commercially available, this work opens up large scale unsupervised NLP training to most commercial applications and deep learning researchers 11Our code is publicly available: https://github.com/NVIDIA/sentiment-discovery, A model can be trained over most public or private text datasets overnight.","PeriodicalId":307747,"journal":{"name":"2018 30th International Symposium on Computer Architecture and High Performance Computing (SBAC-PAD)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128772278","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
T-SNE-CUDA: GPU-Accelerated T-SNE and its Applications to Modern Data T-SNE- cuda: gpu加速T-SNE及其在现代数据中的应用
David Chan, Roshan Rao, Forrest Huang, J. Canny
Modern datasets and models are notoriously difficult to explore and analyze due to their inherent high dimensionality and massive numbers of samples. Existing visualization methods which employ dimensionality reduction to two or three dimensions are often inefficient and/or ineffective for these datasets. This paper introduces T-SNE-CUDA, a GPU-accelerated implementation of t-distributed Symmetric Neighbour Embedding (t-SNE) for visualizing datasets and models. T-SNE-CUDA significantly outperforms current implementations with 50-700x speedups on the CIFAR-10 and MNIST datasets. These speedups enable, for the first time, visualization of the neural network activations on the entire ImageNet dataset - a feat that was previously computationally intractable. We also demonstrate visualization performance in the NLP domain by visualizing the GloVe embedding vectors. From these visualizations, we can draw interesting conclusions about using the L2 metric in these embedding spaces. T-SNE-CUDA is publicly available at https://github.com/CannyLab/tsne-cuda.
现代数据集和模型由于其固有的高维度和大量样本而难以探索和分析。现有的可视化方法采用降维到二维或三维,对于这些数据集往往效率低下和/或无效。本文介绍了t-SNE - cuda,一种用于可视化数据集和模型的t-分布式对称邻居嵌入(t-SNE)的gpu加速实现。T-SNE-CUDA在CIFAR-10和MNIST数据集上的速度提高了50-700倍,明显优于当前的实现。这些加速第一次使整个ImageNet数据集上的神经网络激活的可视化成为可能——这是以前在计算上难以实现的壮举。我们还通过可视化GloVe嵌入向量来展示NLP领域的可视化性能。从这些可视化中,我们可以得出关于在这些嵌入空间中使用L2度规的有趣结论。T-SNE-CUDA可在https://github.com/CannyLab/tsne-cuda公开获取。
{"title":"T-SNE-CUDA: GPU-Accelerated T-SNE and its Applications to Modern Data","authors":"David Chan, Roshan Rao, Forrest Huang, J. Canny","doi":"10.1109/CAHPC.2018.8645912","DOIUrl":"https://doi.org/10.1109/CAHPC.2018.8645912","url":null,"abstract":"Modern datasets and models are notoriously difficult to explore and analyze due to their inherent high dimensionality and massive numbers of samples. Existing visualization methods which employ dimensionality reduction to two or three dimensions are often inefficient and/or ineffective for these datasets. This paper introduces T-SNE-CUDA, a GPU-accelerated implementation of t-distributed Symmetric Neighbour Embedding (t-SNE) for visualizing datasets and models. T-SNE-CUDA significantly outperforms current implementations with 50-700x speedups on the CIFAR-10 and MNIST datasets. These speedups enable, for the first time, visualization of the neural network activations on the entire ImageNet dataset - a feat that was previously computationally intractable. We also demonstrate visualization performance in the NLP domain by visualizing the GloVe embedding vectors. From these visualizations, we can draw interesting conclusions about using the L2 metric in these embedding spaces. T-SNE-CUDA is publicly available at https://github.com/CannyLab/tsne-cuda.","PeriodicalId":307747,"journal":{"name":"2018 30th International Symposium on Computer Architecture and High Performance Computing (SBAC-PAD)","volume":"100 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123842474","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 78
An Argument in Favor of Strong Scaling for Deep Neural Networks with Small Datasets 支持小数据集深度神经网络强缩放的论证
R. L. F. Cunha, E. Rodrigues, Matheus Palhares Viana, Dario Augusto Borges Oliveira
In recent years, with the popularization of deep learning frameworks and large datasets, researchers have started parallelizing their models in order to train faster. This is crucially important, because they typically explore many hyperparameters in order to find the best ones for their applications. This process is time consuming and, consequently, speeding up training improves productivity. One approach to parallelize deep learning models followed by many researchers is based on weak scaling. The minibatches increase in size as new GPUs are added to the system. In addition, new learning rates schedules have been proposed to fix optimization issues that occur with large minibatch sizes. In this paper, however, we show that the recommendations provided by recent work do not apply to models that lack large datasets. In fact, we argument in favor of using strong scaling for achieving reliable performance in such cases. We evaluated our approach with up to 32 GPUs and show that weak scaling not only does not have the same accuracy as the sequential model, it also fails to converge most of time. Meanwhile, strong scaling has good scalability while having exactly the same accuracy of a sequential implementation.
近年来,随着深度学习框架和大数据集的普及,为了更快地训练,研究人员开始将他们的模型并行化。这一点至关重要,因为它们通常会探索许多超参数,以便找到最适合其应用程序的超参数。这个过程是耗时的,因此,加快培训可以提高生产力。许多研究人员采用的一种并行化深度学习模型的方法是基于弱缩放。当新的gpu被添加到系统中时,minibatch的大小也会增加。此外,还提出了新的学习率计划,以解决大型小批处理中出现的优化问题。然而,在本文中,我们表明,最近的工作提供的建议并不适用于缺乏大数据集的模型。事实上,我们主张在这种情况下使用强伸缩性来实现可靠的性能。我们在多达32个gpu的情况下评估了我们的方法,并表明弱缩放不仅没有与序列模型相同的精度,而且在大多数情况下也无法收敛。同时,强扩展具有良好的可伸缩性,同时具有与顺序实现完全相同的精度。
{"title":"An Argument in Favor of Strong Scaling for Deep Neural Networks with Small Datasets","authors":"R. L. F. Cunha, E. Rodrigues, Matheus Palhares Viana, Dario Augusto Borges Oliveira","doi":"10.1109/SBAC-PAD.2018.00057","DOIUrl":"https://doi.org/10.1109/SBAC-PAD.2018.00057","url":null,"abstract":"In recent years, with the popularization of deep learning frameworks and large datasets, researchers have started parallelizing their models in order to train faster. This is crucially important, because they typically explore many hyperparameters in order to find the best ones for their applications. This process is time consuming and, consequently, speeding up training improves productivity. One approach to parallelize deep learning models followed by many researchers is based on weak scaling. The minibatches increase in size as new GPUs are added to the system. In addition, new learning rates schedules have been proposed to fix optimization issues that occur with large minibatch sizes. In this paper, however, we show that the recommendations provided by recent work do not apply to models that lack large datasets. In fact, we argument in favor of using strong scaling for achieving reliable performance in such cases. We evaluated our approach with up to 32 GPUs and show that weak scaling not only does not have the same accuracy as the sequential model, it also fails to converge most of time. Meanwhile, strong scaling has good scalability while having exactly the same accuracy of a sequential implementation.","PeriodicalId":307747,"journal":{"name":"2018 30th International Symposium on Computer Architecture and High Performance Computing (SBAC-PAD)","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116956979","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
On the Resilience of RTL NN Accelerators: Fault Characterization and Mitigation RTL神经网络加速器的弹性:故障表征与缓解
Behzad Salami, O. Unsal, A. Cristal
Machine Learning (ML) is making a strong resurgence in tune with the massive generation of unstructured data which in turn requires massive computational resources. Due to the inherently compute and power-intensive structure of Neural Networks (NNs), hardware accelerators emerge as a promising solution. However, with technology node scaling below 10nm, hardware accelerators become more susceptible to faults, which in turn can impact the NN accuracy. In this paper, we study the resilience aspects of Register-Transfer Level (RTL) model of NN accelerators, in particular, fault characterization and mitigation. By following a High-Level Synthesis (HLS) approach, first, we characterize the vulnerability of various components of RTL NN. We observed that the severity of faults depends on both i) application-level specifications, i.e., NN data (inputs, weights, or intermediate) and NN layers and ii) architectural-level specifications, i.e., data representation model and the parallelism degree of the underlying accelerator. Second, motivated by characterization results, we present a low-overhead fault mitigation technique that can efficiently correct bit flips, by 47.3% better than state-of-the-art methods.
随着大量非结构化数据的产生,机器学习(ML)正在强势复苏,而非结构化数据的产生又需要大量的计算资源。由于神经网络固有的计算和功耗密集型结构,硬件加速器成为一种很有前途的解决方案。然而,当技术节点缩放到10nm以下时,硬件加速器更容易受到故障的影响,从而影响神经网络的精度。在本文中,我们研究了神经网络加速器的寄存器-传递水平(RTL)模型的弹性方面,特别是故障表征和缓解。通过遵循高层次综合(HLS)的方法,首先,我们表征了RTL神经网络的各个组成部分的脆弱性。我们观察到,故障的严重程度取决于i)应用级规范,即神经网络数据(输入、权重或中间)和神经网络层,以及ii)架构级规范,即数据表示模型和底层加速器的并行度。其次,在表征结果的激励下,我们提出了一种低开销的故障缓解技术,可以有效地纠正位翻转,比最先进的方法好47.3%。
{"title":"On the Resilience of RTL NN Accelerators: Fault Characterization and Mitigation","authors":"Behzad Salami, O. Unsal, A. Cristal","doi":"10.1109/CAHPC.2018.8645906","DOIUrl":"https://doi.org/10.1109/CAHPC.2018.8645906","url":null,"abstract":"Machine Learning (ML) is making a strong resurgence in tune with the massive generation of unstructured data which in turn requires massive computational resources. Due to the inherently compute and power-intensive structure of Neural Networks (NNs), hardware accelerators emerge as a promising solution. However, with technology node scaling below 10nm, hardware accelerators become more susceptible to faults, which in turn can impact the NN accuracy. In this paper, we study the resilience aspects of Register-Transfer Level (RTL) model of NN accelerators, in particular, fault characterization and mitigation. By following a High-Level Synthesis (HLS) approach, first, we characterize the vulnerability of various components of RTL NN. We observed that the severity of faults depends on both i) application-level specifications, i.e., NN data (inputs, weights, or intermediate) and NN layers and ii) architectural-level specifications, i.e., data representation model and the parallelism degree of the underlying accelerator. Second, motivated by characterization results, we present a low-overhead fault mitigation technique that can efficiently correct bit flips, by 47.3% better than state-of-the-art methods.","PeriodicalId":307747,"journal":{"name":"2018 30th International Symposium on Computer Architecture and High Performance Computing (SBAC-PAD)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121664327","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 59
期刊
2018 30th International Symposium on Computer Architecture and High Performance Computing (SBAC-PAD)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1