Shusen Liu, Brandon Bocklund, James Diffenderfer, Shreya Chaganti, Bhavya Kailkhura, Scott K. McCall, Brian Gallagher, Aurélien Perron, Joseph T. McKeown
{"title":"利用传统机器学习和深度神经网络预测高熵合金相分数的比较研究","authors":"Shusen Liu, Brandon Bocklund, James Diffenderfer, Shreya Chaganti, Bhavya Kailkhura, Scott K. McCall, Brian Gallagher, Aurélien Perron, Joseph T. McKeown","doi":"10.1038/s41524-024-01335-1","DOIUrl":null,"url":null,"abstract":"<p>Predicting phase stability in high entropy alloys (HEAs), such as phase fractions as functions of composition and temperature, is essential for understanding alloy properties and screening desirable materials. Traditional methods like CALPHAD are computationally intensive for exploring high-dimensional compositional spaces. To address such a challenge, this study explored and compared the effectiveness of random forests (RF) and deep neural networks (DNN) for accelerating materials discovery by building surrogate models of phase stability prediction. For <i>interpolation</i> scenarios (testing on the same order of system as trained), RF models generally produce smaller errors than DNN models. However, for <i>extrapolation</i> scenarios (training on lower-order systems and testing on higher order systems), DNNs generalize more effectively than traditional ML models. DNN demonstrate the potential to predict topologically relevant phase composition when data were missing, making it a powerful predictive tool in materials discovery frameworks. The study uses a CALPHAD dataset of 480 million data points generated from a custom database, available for further model development and benchmarking. Experiments show that DNN models are data-efficient, achieving similar performance with a fraction of the dataset. This work highlights the potential of DNNs in materials discovery, providing a powerful tool for predicting phase stability in HEAs, particularly within the Cr-Hf-Mo-Nb-Ta-Ti-V-W-Zr composition space.</p>","PeriodicalId":19342,"journal":{"name":"npj Computational Materials","volume":"13 1","pages":""},"PeriodicalIF":9.4000,"publicationDate":"2024-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A comparative study of predicting high entropy alloy phase fractions with traditional machine learning and deep neural networks\",\"authors\":\"Shusen Liu, Brandon Bocklund, James Diffenderfer, Shreya Chaganti, Bhavya Kailkhura, Scott K. McCall, Brian Gallagher, Aurélien Perron, Joseph T. McKeown\",\"doi\":\"10.1038/s41524-024-01335-1\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Predicting phase stability in high entropy alloys (HEAs), such as phase fractions as functions of composition and temperature, is essential for understanding alloy properties and screening desirable materials. Traditional methods like CALPHAD are computationally intensive for exploring high-dimensional compositional spaces. To address such a challenge, this study explored and compared the effectiveness of random forests (RF) and deep neural networks (DNN) for accelerating materials discovery by building surrogate models of phase stability prediction. For <i>interpolation</i> scenarios (testing on the same order of system as trained), RF models generally produce smaller errors than DNN models. However, for <i>extrapolation</i> scenarios (training on lower-order systems and testing on higher order systems), DNNs generalize more effectively than traditional ML models. DNN demonstrate the potential to predict topologically relevant phase composition when data were missing, making it a powerful predictive tool in materials discovery frameworks. The study uses a CALPHAD dataset of 480 million data points generated from a custom database, available for further model development and benchmarking. Experiments show that DNN models are data-efficient, achieving similar performance with a fraction of the dataset. This work highlights the potential of DNNs in materials discovery, providing a powerful tool for predicting phase stability in HEAs, particularly within the Cr-Hf-Mo-Nb-Ta-Ti-V-W-Zr composition space.</p>\",\"PeriodicalId\":19342,\"journal\":{\"name\":\"npj Computational Materials\",\"volume\":\"13 1\",\"pages\":\"\"},\"PeriodicalIF\":9.4000,\"publicationDate\":\"2024-08-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"npj Computational Materials\",\"FirstCategoryId\":\"88\",\"ListUrlMain\":\"https://doi.org/10.1038/s41524-024-01335-1\",\"RegionNum\":1,\"RegionCategory\":\"材料科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"CHEMISTRY, PHYSICAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"npj Computational Materials","FirstCategoryId":"88","ListUrlMain":"https://doi.org/10.1038/s41524-024-01335-1","RegionNum":1,"RegionCategory":"材料科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"CHEMISTRY, PHYSICAL","Score":null,"Total":0}
引用次数: 0
摘要
预测高熵合金(HEAs)中的相稳定性(如相分数作为成分和温度的函数)对于了解合金特性和筛选理想材料至关重要。CALPHAD 等传统方法在探索高维成分空间时需要大量计算。为了应对这一挑战,本研究探索并比较了随机森林(RF)和深度神经网络(DNN)在通过建立相稳定性预测替代模型加速材料发现方面的有效性。对于内插情景(在与训练时相同的系统阶次上进行测试),RF 模型产生的误差通常小于 DNN 模型。然而,在外推法情况下(在低阶系统上进行训练,在高阶系统上进行测试),DNN 的泛化效果比传统的 ML 模型更好。DNN 展示了在数据缺失时预测拓扑相关相组成的潜力,使其成为材料发现框架中一个强大的预测工具。该研究使用的 CALPHAD 数据集由定制数据库生成,包含 4.8 亿个数据点,可用于进一步的模型开发和基准测试。实验表明,DNN 模型的数据效率很高,只需数据集的一小部分就能获得类似的性能。这项工作凸显了 DNN 在材料发现方面的潜力,为预测 HEA 的相稳定性提供了强大的工具,特别是在 Cr-Hf-Mo-Nb-Ta-Ti-V-W-Zr 成分空间内。
A comparative study of predicting high entropy alloy phase fractions with traditional machine learning and deep neural networks
Predicting phase stability in high entropy alloys (HEAs), such as phase fractions as functions of composition and temperature, is essential for understanding alloy properties and screening desirable materials. Traditional methods like CALPHAD are computationally intensive for exploring high-dimensional compositional spaces. To address such a challenge, this study explored and compared the effectiveness of random forests (RF) and deep neural networks (DNN) for accelerating materials discovery by building surrogate models of phase stability prediction. For interpolation scenarios (testing on the same order of system as trained), RF models generally produce smaller errors than DNN models. However, for extrapolation scenarios (training on lower-order systems and testing on higher order systems), DNNs generalize more effectively than traditional ML models. DNN demonstrate the potential to predict topologically relevant phase composition when data were missing, making it a powerful predictive tool in materials discovery frameworks. The study uses a CALPHAD dataset of 480 million data points generated from a custom database, available for further model development and benchmarking. Experiments show that DNN models are data-efficient, achieving similar performance with a fraction of the dataset. This work highlights the potential of DNNs in materials discovery, providing a powerful tool for predicting phase stability in HEAs, particularly within the Cr-Hf-Mo-Nb-Ta-Ti-V-W-Zr composition space.
期刊介绍:
npj Computational Materials is a high-quality open access journal from Nature Research that publishes research papers applying computational approaches for the design of new materials and enhancing our understanding of existing ones. The journal also welcomes papers on new computational techniques and the refinement of current approaches that support these aims, as well as experimental papers that complement computational findings.
Some key features of npj Computational Materials include a 2-year impact factor of 12.241 (2021), article downloads of 1,138,590 (2021), and a fast turnaround time of 11 days from submission to the first editorial decision. The journal is indexed in various databases and services, including Chemical Abstracts Service (ACS), Astrophysics Data System (ADS), Current Contents/Physical, Chemical and Earth Sciences, Journal Citation Reports/Science Edition, SCOPUS, EI Compendex, INSPEC, Google Scholar, SCImago, DOAJ, CNKI, and Science Citation Index Expanded (SCIE), among others.