A. Shahkarami, Kimberly L. Ayers, Guochang Wang, Alivia Ayers
{"title":"应用机器学习算法优化Marcellus页岩未来产量,以宾夕法尼亚州西南部为例","authors":"A. Shahkarami, Kimberly L. Ayers, Guochang Wang, Alivia Ayers","doi":"10.2118/191827-18ERM-MS","DOIUrl":null,"url":null,"abstract":"\n The Marcellus Shale has more than a decade of development history. However, there are many questions that still remain unanswered. What is the best inter-well spacing? What are the optimum stage length, proppant loading, and cluster spacing? What are the ideal combinations of these completion parameters? And how can we maximize the rate return on our investment? This study proposes innovative tools that allow researchers to answer these questions. We build these set of tools by utilizing the pattern recognition abilities of machine learning algorithms and public data from the Southwestern Pennsylvania region of the Marcellus Shale.\n By means of artificial intelligence and data mining techniques, we studied a database that includes public data from more than 2,000 wells producing from the aforementioned study area. The database contained completion, drilling, and production history information from various operators active in Allegheny, Greene, Fayette, Washington, and Westmoreland counties located in the Southwestern Pennsylvania. Extensive preprocessing and data cleansing steps were involved to prepare the database. Various machine learning techniques (Linear Regression (LR), Support Vector Machines (SVMs), Artificial Neural Networks (ANNs), and Gaussian Processes (GP)) were applied to understand the non-linear patterns in the data. The objective was to develop predictive models that were trained and validated based on the current database. The predictive models were validated using information originating from numerous wells in the area. Once validated, the model could be used in reservoir management decision-making workflows to answer questions such as what are the best drilling scenarios, the optimum hydraulic fracturing design, the initial production rate, and the estimated ultimate recovery (EUR). The workflow is purely based on field data and free of any cognitive human bias. As soon as more data is available, the model could be updated. The core data in this workflow is sourced from public domains, and therefore, intensive preprocessing efforts were necessary.","PeriodicalId":298489,"journal":{"name":"Day 4 Wed, October 10, 2018","volume":"7 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-10-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":"{\"title\":\"Application of Machine Learning Algorithms for Optimizing Future Production in Marcellus Shale, Case Study of Southwestern Pennsylvania\",\"authors\":\"A. Shahkarami, Kimberly L. Ayers, Guochang Wang, Alivia Ayers\",\"doi\":\"10.2118/191827-18ERM-MS\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"\\n The Marcellus Shale has more than a decade of development history. However, there are many questions that still remain unanswered. What is the best inter-well spacing? What are the optimum stage length, proppant loading, and cluster spacing? What are the ideal combinations of these completion parameters? And how can we maximize the rate return on our investment? This study proposes innovative tools that allow researchers to answer these questions. We build these set of tools by utilizing the pattern recognition abilities of machine learning algorithms and public data from the Southwestern Pennsylvania region of the Marcellus Shale.\\n By means of artificial intelligence and data mining techniques, we studied a database that includes public data from more than 2,000 wells producing from the aforementioned study area. The database contained completion, drilling, and production history information from various operators active in Allegheny, Greene, Fayette, Washington, and Westmoreland counties located in the Southwestern Pennsylvania. Extensive preprocessing and data cleansing steps were involved to prepare the database. Various machine learning techniques (Linear Regression (LR), Support Vector Machines (SVMs), Artificial Neural Networks (ANNs), and Gaussian Processes (GP)) were applied to understand the non-linear patterns in the data. The objective was to develop predictive models that were trained and validated based on the current database. The predictive models were validated using information originating from numerous wells in the area. Once validated, the model could be used in reservoir management decision-making workflows to answer questions such as what are the best drilling scenarios, the optimum hydraulic fracturing design, the initial production rate, and the estimated ultimate recovery (EUR). The workflow is purely based on field data and free of any cognitive human bias. As soon as more data is available, the model could be updated. The core data in this workflow is sourced from public domains, and therefore, intensive preprocessing efforts were necessary.\",\"PeriodicalId\":298489,\"journal\":{\"name\":\"Day 4 Wed, October 10, 2018\",\"volume\":\"7 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2018-10-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"7\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Day 4 Wed, October 10, 2018\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.2118/191827-18ERM-MS\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Day 4 Wed, October 10, 2018","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.2118/191827-18ERM-MS","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Application of Machine Learning Algorithms for Optimizing Future Production in Marcellus Shale, Case Study of Southwestern Pennsylvania
The Marcellus Shale has more than a decade of development history. However, there are many questions that still remain unanswered. What is the best inter-well spacing? What are the optimum stage length, proppant loading, and cluster spacing? What are the ideal combinations of these completion parameters? And how can we maximize the rate return on our investment? This study proposes innovative tools that allow researchers to answer these questions. We build these set of tools by utilizing the pattern recognition abilities of machine learning algorithms and public data from the Southwestern Pennsylvania region of the Marcellus Shale.
By means of artificial intelligence and data mining techniques, we studied a database that includes public data from more than 2,000 wells producing from the aforementioned study area. The database contained completion, drilling, and production history information from various operators active in Allegheny, Greene, Fayette, Washington, and Westmoreland counties located in the Southwestern Pennsylvania. Extensive preprocessing and data cleansing steps were involved to prepare the database. Various machine learning techniques (Linear Regression (LR), Support Vector Machines (SVMs), Artificial Neural Networks (ANNs), and Gaussian Processes (GP)) were applied to understand the non-linear patterns in the data. The objective was to develop predictive models that were trained and validated based on the current database. The predictive models were validated using information originating from numerous wells in the area. Once validated, the model could be used in reservoir management decision-making workflows to answer questions such as what are the best drilling scenarios, the optimum hydraulic fracturing design, the initial production rate, and the estimated ultimate recovery (EUR). The workflow is purely based on field data and free of any cognitive human bias. As soon as more data is available, the model could be updated. The core data in this workflow is sourced from public domains, and therefore, intensive preprocessing efforts were necessary.