Defense POW/MIA Accounting Agency's (DPAA) mission, the fullest possible accounting of our nation's defense personnel lost in action, drives all agency direction. This includes novel applications of multidisciplinary solutions towards maximizing agency capabilities. By building collaborations and harnessing applications for advancing technology, DPAA's Partnerships and Innovations Directorate has been a key facilitator for implementing this imperative and has proven the benefit to challenging the operational status quo. The partnership paradigm represents a game-changing model that forces stakeholders to think differently and creatively, and ultimately deviate from traditional operations.
{"title":"Implementing Innovation: Partnerships and Problem-Solving in the Defense POW/MIA Accounting Agency","authors":"Kara E Davis, Jeneva Wright","doi":"10.4043/29388-MS","DOIUrl":"https://doi.org/10.4043/29388-MS","url":null,"abstract":"\u0000 Defense POW/MIA Accounting Agency's (DPAA) mission, the fullest possible accounting of our nation's defense personnel lost in action, drives all agency direction. This includes novel applications of multidisciplinary solutions towards maximizing agency capabilities. By building collaborations and harnessing applications for advancing technology, DPAA's Partnerships and Innovations Directorate has been a key facilitator for implementing this imperative and has proven the benefit to challenging the operational status quo. The partnership paradigm represents a game-changing model that forces stakeholders to think differently and creatively, and ultimately deviate from traditional operations.","PeriodicalId":11149,"journal":{"name":"Day 1 Mon, May 06, 2019","volume":"10 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2019-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90168169","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The ensemble Kalman filter (EnKF) algorithm is an elegant and effective method to optimize model parameters based on differences with predictions of model and measurement data. Great progress has been accomplished using EnKF for data assimilation within reservoir modeling during the last two decades. A typical example where data assimilation is necessary is history matching—the process of adjusting the model variables to account for observations of rates, pressure, saturations, and other variables. In contrast, much less attention has been given to flow model optimization for other workflows, such as drilling, production, flow assurance, and well testing. Providing two examples of applying the EnKF for real-time quantification of sensor-generated data is the aim of this paper. These examples include the analysis of the declining production curve and zonal pressure sensor data for evaluating matrix permeabilities and processing the multichannel optical to monitor the cleanup of hydrocarbon fluid samples during formation-tester sampling. Additionally, how the EnKF algorithm can be successfully applied to segmented multichannel sensor field data obtained from multichannel optical density sensors exhibiting the gradual transition from oil-based mud (OBM) filtrate to native formation fluid during formation-tester sampling stations is discussed. A simple algebraic proxy model is used to predict the decline of the volumetric fraction of OBM filtrate with time during formation-tester sampling. To implement and test the algorithm, a proof-of-concept MATLAB code was developed. Synthetic (simulated) pressure flow rate data were used for the production decline case while the actual field data from eight channel optical sensors were used for the formation-testing case. Model runs were performed in 50 to 60 combinations of model parameters, which were normally distributed around the best-guess values at the initial step. For both cases, only two to three iterations of the algorithm were sufficient to obtain values of the matching parameters.
{"title":"Real-Time Monitoring and Interpretation of Wireline Formation Testing Using Ensemble Kalman Filter","authors":"H. Elshahawi, A. Filippov","doi":"10.4043/29245-MS","DOIUrl":"https://doi.org/10.4043/29245-MS","url":null,"abstract":"\u0000 The ensemble Kalman filter (EnKF) algorithm is an elegant and effective method to optimize model parameters based on differences with predictions of model and measurement data. Great progress has been accomplished using EnKF for data assimilation within reservoir modeling during the last two decades. A typical example where data assimilation is necessary is history matching—the process of adjusting the model variables to account for observations of rates, pressure, saturations, and other variables. In contrast, much less attention has been given to flow model optimization for other workflows, such as drilling, production, flow assurance, and well testing.\u0000 Providing two examples of applying the EnKF for real-time quantification of sensor-generated data is the aim of this paper. These examples include the analysis of the declining production curve and zonal pressure sensor data for evaluating matrix permeabilities and processing the multichannel optical to monitor the cleanup of hydrocarbon fluid samples during formation-tester sampling.\u0000 Additionally, how the EnKF algorithm can be successfully applied to segmented multichannel sensor field data obtained from multichannel optical density sensors exhibiting the gradual transition from oil-based mud (OBM) filtrate to native formation fluid during formation-tester sampling stations is discussed. A simple algebraic proxy model is used to predict the decline of the volumetric fraction of OBM filtrate with time during formation-tester sampling.\u0000 To implement and test the algorithm, a proof-of-concept MATLAB code was developed. Synthetic (simulated) pressure flow rate data were used for the production decline case while the actual field data from eight channel optical sensors were used for the formation-testing case. Model runs were performed in 50 to 60 combinations of model parameters, which were normally distributed around the best-guess values at the initial step. For both cases, only two to three iterations of the algorithm were sufficient to obtain values of the matching parameters.","PeriodicalId":11149,"journal":{"name":"Day 1 Mon, May 06, 2019","volume":"16 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2019-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91187767","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
1. Review the effectiveness of existing technology for identifying and characterizing submerged and buried prehistoric archaeological sites.2. Critique and suggest changes to existing technology to improve the effectiveness for the identification and characterization of submerged prehistoric archaeological sites.3. Examine the "why" this is important to offshore technology regulation and management. The process can easily be paraphrased as "sound, software, and three inch pipes" - the sound being the pulse from sonar or sub-bottom instruments, the software being the methods by which those pulses are processed and display, and the three-inch pipes, of course, denoting the coring tubes used to sample targets identified in using sound and software. In fairness to the methods discussed herein, their application has been, at best, uneven. The use of the first two technologies - sound/acoustic and post-processing software programs have been in regular use by geophysical contractors. The latter technology - sediment coring - has been utilized far less. How effective has this methodology proven to be over the past thirty years? That is open to some debate and this paper will examine some reasons why this is so and suggest some possible alternatives. The overarching goal of this paper is to push beyond this methodology to potentially more effective technologies or at least improved usage of the current technology. Submerged prehistoric archaeological sites are "hard problems" in terms of their discovery and characterization. Compared to submerged historic archaeological sites - primarily shipwrecks - those are "easy." This comparison of the use and effectiveness of technology is novel and additive.
{"title":"Submerged Archaeological Resources: Sound, Software and Three Inch Pipes - A Recipe for Finding Submerged Prehistoric Archaeological Sites?","authors":"Garrison G Ervan","doi":"10.4043/29597-MS","DOIUrl":"https://doi.org/10.4043/29597-MS","url":null,"abstract":"\u0000 \u0000 \u0000 1. Review the effectiveness of existing technology for identifying and characterizing submerged and buried prehistoric archaeological sites.2. Critique and suggest changes to existing technology to improve the effectiveness for the identification and characterization of submerged prehistoric archaeological sites.3. Examine the \"why\" this is important to offshore technology regulation and management.\u0000 \u0000 \u0000 \u0000 The process can easily be paraphrased as \"sound, software, and three inch pipes\" - the sound being the pulse from sonar or sub-bottom instruments, the software being the methods by which those pulses are processed and display, and the three-inch pipes, of course, denoting the coring tubes used to sample targets identified in using sound and software. In fairness to the methods discussed herein, their application has been, at best, uneven. The use of the first two technologies - sound/acoustic and post-processing software programs have been in regular use by geophysical contractors. The latter technology - sediment coring - has been utilized far less.\u0000 \u0000 \u0000 \u0000 How effective has this methodology proven to be over the past thirty years? That is open to some debate and this paper will examine some reasons why this is so and suggest some possible alternatives. The overarching goal of this paper is to push beyond this methodology to potentially more effective technologies or at least improved usage of the current technology.\u0000 \u0000 \u0000 \u0000 Submerged prehistoric archaeological sites are \"hard problems\" in terms of their discovery and characterization. Compared to submerged historic archaeological sites - primarily shipwrecks - those are \"easy.\" This comparison of the use and effectiveness of technology is novel and additive.\u0000","PeriodicalId":11149,"journal":{"name":"Day 1 Mon, May 06, 2019","volume":"53 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2019-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89855224","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
For millennia, humanity has and continues to leave imprints on the physical world. Archaeologists have often turned to technology for help in interpreting these past imprints. This is especially so in the realm of underwater archaeology where sites may range from submerged prehistoric landscapes to what was for thousands of years one of human kind's most technologically complex objects, the ship. Since the close of the Second World War, offshore technology has increasingly created opportunities and tools for subsea scientists to better map, explore, and understand both our natural world and humanity's interaction with it. The growth of offshore oil and gas exploration technology did not go unnoticed by archaeologists who were quick to modify these technologies and create new methodologies to use them for site analyses. Starting with shallow water locations and ultimately moving into ultra-deepwater, marine archaeologists in the energy sector have greatly benefited from technological improvements originally aimed at the exploitation of offshore oil and gas. Beginning with echosounders, side-scan sonars, and magnetometers in towed arrays, before moving into high-resolution side-scan sonars, multibeam bathymetry and multibeam backscatter systems, subbottom profilers, and gradiometers, improving technologies have directly impacted marine archaeology. This paper provides a review of the technologies archaeologists have appropriated for the purposes of revealing humanity's past with examples of some of the discovered sites from around the globe.
{"title":"50 Years of Offshore Technological Innovation Illuminates Millennia of Archaeological Discoveries","authors":"K. Faulk","doi":"10.4043/29333-MS","DOIUrl":"https://doi.org/10.4043/29333-MS","url":null,"abstract":"\u0000 For millennia, humanity has and continues to leave imprints on the physical world. Archaeologists have often turned to technology for help in interpreting these past imprints. This is especially so in the realm of underwater archaeology where sites may range from submerged prehistoric landscapes to what was for thousands of years one of human kind's most technologically complex objects, the ship.\u0000 Since the close of the Second World War, offshore technology has increasingly created opportunities and tools for subsea scientists to better map, explore, and understand both our natural world and humanity's interaction with it. The growth of offshore oil and gas exploration technology did not go unnoticed by archaeologists who were quick to modify these technologies and create new methodologies to use them for site analyses.\u0000 Starting with shallow water locations and ultimately moving into ultra-deepwater, marine archaeologists in the energy sector have greatly benefited from technological improvements originally aimed at the exploitation of offshore oil and gas. Beginning with echosounders, side-scan sonars, and magnetometers in towed arrays, before moving into high-resolution side-scan sonars, multibeam bathymetry and multibeam backscatter systems, subbottom profilers, and gradiometers, improving technologies have directly impacted marine archaeology.\u0000 This paper provides a review of the technologies archaeologists have appropriated for the purposes of revealing humanity's past with examples of some of the discovered sites from around the globe.","PeriodicalId":11149,"journal":{"name":"Day 1 Mon, May 06, 2019","volume":"318 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2019-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87758049","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Craig Jones, G. Chang, A. Dallman, Jesse D. Roberts, K. Raghukumar, S. McWilliams
The wave energy resource for U.S. coastal regions has been estimated at approximately 1,200 TWh/yr (EPRI 2011). The magnitude is comparable to the natural gas and coal energy generation. Although the wave energy industry is relatively new from a commercial perspective, wave energy conversion (WEC) technology is developing at an increasing pace. Ramping up to commercial scale deployment of WEC arrays requires demonstration of performance that is economically competitive with other energy generation methods. The International Electrotechnical Commission has provided technical specifications for developing wave energy resource assessments and characterizations, but it is ultimately up to developers to create pathways for making a specific site competitive. The present study uses example sites to evaluate the annual energy production using different wave energy conversion strategies and examines pathways available to make WEC deployments competitive. The wave energy resource is evaluated for sites along the U.S. coast and combinations of wave modeling and basic resource assessments determine factors affecting the cost of energy at these sites. The results of this study advance the understanding of wave resource and WEC device assessment required to evaluate commercial-scale deployments.
{"title":"Assessment of Wave Energy Resources and Factors Affecting Conversion","authors":"Craig Jones, G. Chang, A. Dallman, Jesse D. Roberts, K. Raghukumar, S. McWilliams","doi":"10.4043/29570-MS","DOIUrl":"https://doi.org/10.4043/29570-MS","url":null,"abstract":"\u0000 The wave energy resource for U.S. coastal regions has been estimated at approximately 1,200 TWh/yr (EPRI 2011). The magnitude is comparable to the natural gas and coal energy generation. Although the wave energy industry is relatively new from a commercial perspective, wave energy conversion (WEC) technology is developing at an increasing pace. Ramping up to commercial scale deployment of WEC arrays requires demonstration of performance that is economically competitive with other energy generation methods. The International Electrotechnical Commission has provided technical specifications for developing wave energy resource assessments and characterizations, but it is ultimately up to developers to create pathways for making a specific site competitive.\u0000 The present study uses example sites to evaluate the annual energy production using different wave energy conversion strategies and examines pathways available to make WEC deployments competitive. The wave energy resource is evaluated for sites along the U.S. coast and combinations of wave modeling and basic resource assessments determine factors affecting the cost of energy at these sites. The results of this study advance the understanding of wave resource and WEC device assessment required to evaluate commercial-scale deployments.","PeriodicalId":11149,"journal":{"name":"Day 1 Mon, May 06, 2019","volume":"64 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2019-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83900116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Bruno Moczydlower, Fernando Pacifico Figueiredo Junior, J. Pizarro
This paper addresses the innovative appraisal strategy applied to the Libra project; located in ultra-deep waters offshore Brazil. It details the key role of the Extended Well Test (EWT) Program, within the field overall Risk Mitigation Plan, as well as its interfaces with additional appraisal activities. The Value of Information (VoI) for the main acquired data is described, highlighting the associated impacts for the full field development and validation of the enhanced recovery strategy. A case study approach details how the whole EWT project maximized the acquired information, mainly from a reservoir point of view. Although the EWT approach is not new to Petrobras in the offshore environment, this is the first one with simultaneous oil production and gas reinjection. Several reasons justify the use of the industry’s first dedicated offshore EWT system with this capability. Gathering data on the main dynamic parameters of the field was critical to speed-up the development, with an acceptable risk level. The incorporation of these data in the reservoir models and the impacts in the most relevant development decisions are also described. The chosen methodology brought many opportunities, as well as challenges to interpret the data and to incorporate them in the reservoir models. Furthermore, the capability to produce without continuous gas flaring makes it possible to apply such approach anywhere else in the world.
{"title":"Libra Extended Well Test - An Innovative Approach to De-Risk a Complex Field Development","authors":"Bruno Moczydlower, Fernando Pacifico Figueiredo Junior, J. Pizarro","doi":"10.4043/29653-MS","DOIUrl":"https://doi.org/10.4043/29653-MS","url":null,"abstract":"\u0000 This paper addresses the innovative appraisal strategy applied to the Libra project; located in ultra-deep waters offshore Brazil. It details the key role of the Extended Well Test (EWT) Program, within the field overall Risk Mitigation Plan, as well as its interfaces with additional appraisal activities. The Value of Information (VoI) for the main acquired data is described, highlighting the associated impacts for the full field development and validation of the enhanced recovery strategy.\u0000 A case study approach details how the whole EWT project maximized the acquired information, mainly from a reservoir point of view. Although the EWT approach is not new to Petrobras in the offshore environment, this is the first one with simultaneous oil production and gas reinjection. Several reasons justify the use of the industry’s first dedicated offshore EWT system with this capability. Gathering data on the main dynamic parameters of the field was critical to speed-up the development, with an acceptable risk level. The incorporation of these data in the reservoir models and the impacts in the most relevant development decisions are also described.\u0000 The chosen methodology brought many opportunities, as well as challenges to interpret the data and to incorporate them in the reservoir models. Furthermore, the capability to produce without continuous gas flaring makes it possible to apply such approach anywhere else in the world.","PeriodicalId":11149,"journal":{"name":"Day 1 Mon, May 06, 2019","volume":"370 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2019-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77798854","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
R. Giolo, Aurelie Berthelot, P. Pedenaud, Graeme Skivington
SPRINGS® (Subsea PRocessing and INjection Gear for Seawater) is a qualified process for subsea water treatment and injection. It uses membrane technology for water desulfation upstream of water injection wells to prevent sulfate scaling on the production side (nearwell bore, well and production equipment). It moves the water treatment from topside to subsea locations close to the injection wells with only power and communication tie-backs to existing topside facilities. Qualification of the process was achieved through both onshore and offshore trials. In advance of deploying the first industrial application, an industrialisation programme was undertaken in order to ensure that every component necessary for the subsea process implementation was available and had a sufficient technology readiness level to be safely installed and operated within the subsea plant. The existing and available technologies were reviewed vis-À-vis the requirements arising from both the process and the business strategy. Several industrial partners were engaged to determine the elements of novelty that needed to be brought to each technology or component to satisfy such requirements. The new technologies included: Subsea barrier-fluidless pumps Open framework all-electric control systems High-cycling electric actuators and valves Subsea water analyser Subsea storage and injection units for chemicals The design basis for the development of each technology, which in most cases included the realisation of a prototype and relevant qualification testing, was set up to consider a range of possible applications with differing environmental conditions, process data and/or IMR scenarios. The most challenging conditions were selected for each development to determine the relevant required performance. Where available, specific standards, such as API 17F (ref. [8]) for subsea electronics, were followed to determine the qualification plans. In those cases where no dedicated specific standard was available, the evaluation of the proposed solution was performed in conjunction with the technology provider through the risk based approach stated in API 17N (ref. [9]) and DNV A203 (ref. [10]). Failure Modes, Effects and Criticality Analyses (FMECAs) as well as technology readiness assessments were performed in order to develop the technology qualification plans. Most of the key equipment qualification plans will be completed by mid-2019, establishing an industrial platform for the deployment of the subsea water treatment and injection technology in a completely all-electric configuration, i.e. connected to the surface only through a communication and power cable. Such an industrial platform will also contain the building blocks for other subsea processes. The presentation and paper will introduce the elements of technological novelty and will describe the process, the challenges and the results of the relevant qualifications.
{"title":"Industrialisation of SPRINGS®, A Qualified Subsea Sea Water Desulfation Process","authors":"R. Giolo, Aurelie Berthelot, P. Pedenaud, Graeme Skivington","doi":"10.4043/29365-MS","DOIUrl":"https://doi.org/10.4043/29365-MS","url":null,"abstract":"\u0000 SPRINGS® (Subsea PRocessing and INjection Gear for Seawater) is a qualified process for subsea water treatment and injection. It uses membrane technology for water desulfation upstream of water injection wells to prevent sulfate scaling on the production side (nearwell bore, well and production equipment). It moves the water treatment from topside to subsea locations close to the injection wells with only power and communication tie-backs to existing topside facilities. Qualification of the process was achieved through both onshore and offshore trials.\u0000 In advance of deploying the first industrial application, an industrialisation programme was undertaken in order to ensure that every component necessary for the subsea process implementation was available and had a sufficient technology readiness level to be safely installed and operated within the subsea plant.\u0000 The existing and available technologies were reviewed vis-À-vis the requirements arising from both the process and the business strategy. Several industrial partners were engaged to determine the elements of novelty that needed to be brought to each technology or component to satisfy such requirements.\u0000 The new technologies included:\u0000 Subsea barrier-fluidless pumps\u0000 Open framework all-electric control systems\u0000 High-cycling electric actuators and valves\u0000 Subsea water analyser\u0000 Subsea storage and injection units for chemicals\u0000 The design basis for the development of each technology, which in most cases included the realisation of a prototype and relevant qualification testing, was set up to consider a range of possible applications with differing environmental conditions, process data and/or IMR scenarios. The most challenging conditions were selected for each development to determine the relevant required performance.\u0000 Where available, specific standards, such as API 17F (ref. [8]) for subsea electronics, were followed to determine the qualification plans. In those cases where no dedicated specific standard was available, the evaluation of the proposed solution was performed in conjunction with the technology provider through the risk based approach stated in API 17N (ref. [9]) and DNV A203 (ref. [10]). Failure Modes, Effects and Criticality Analyses (FMECAs) as well as technology readiness assessments were performed in order to develop the technology qualification plans.\u0000 Most of the key equipment qualification plans will be completed by mid-2019, establishing an industrial platform for the deployment of the subsea water treatment and injection technology in a completely all-electric configuration, i.e. connected to the surface only through a communication and power cable. Such an industrial platform will also contain the building blocks for other subsea processes.\u0000 The presentation and paper will introduce the elements of technological novelty and will describe the process, the challenges and the results of the relevant qualifications.","PeriodicalId":11149,"journal":{"name":"Day 1 Mon, May 06, 2019","volume":"37 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2019-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80468770","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The U.S. Bureau of Ocean Energy Management (BOEM) and Bureau of Safety and Environmental Enforcement (BSEE) are required to consider the effects of their permitted actions on cultural resources, per the National Historic Preservation Act (1966). After the 2010 Deepwater Horizon oil spill in the Gulf of Mexico, BOEM, BSEE, and study partners implemented a multidisciplinary approach to assess micro- to macro-scale impacts from the spill on a select number of previously investigated deepwater shipwrecks. The Gulf of Mexico Shipwreck Corrosion, Hydrocarbon Exposure, Microbiology, and Archaeology (GOM-SCHEMA) Project collected microbiological, geochemical, and archaeological data at wooden- and metal-hulled shipwrecks within and outside of the spill-impacted area for a comparative analysis. Archaeologists documented the selected shipwrecks’ post-spill state of preservation using 3D optical and acoustic scanning systems for comparison with existing geophysical and visual data collected prior to the spill. Marine archaeologists are using 3D scanning systems as new tools for recording and interpreting shipwreck sites and analyzing site formation processes in the marine environment. The GOM-SCHEMA Project combined 3D laser data from an Autonomous Underwater Vehicle (AUV) with 3D sonar data collected by a Remotely Operated Vehicle (ROV) as a first step toward monitoring efforts that will inform BOEM and BSEE of the spill’s long-term impacts on deepwater shipwreck preservation. Employing microbial ecological analyses and laboratory-based corrosion experiments that identified micro-scale impacts, scientists gained a better understanding of how the spill affected the natural processes of metal corrosion and wood degradation. 3D imaging creates a permanent digital record that allows scientists to study minute details and also serves as an important outreach tool by allowing the public to virtually explore archaeological resources. By comparing 3D scans collected repeatedly at the same sites over time, especially the sub-centimeter-accurate 3D laser data, archaeologists can quantitatively analyze changes occurring on these sites as a result of enhanced corrosion or degradation. Collectively, these datasets can inform archaeologists and submerged cultural resource managers about site stability, formation processes such as sedimentation and scouring, and the long-term impacts of a major oil spill on submerged cultural resources.
{"title":"Utilizing 3D Optical and Acoustic Scanning Systems to Investigate Impacts from the Oil Spill on Historic Shipwrecks","authors":"M. Damour, R. Church, D. Warren, C. Horrell","doi":"10.4043/29508-MS","DOIUrl":"https://doi.org/10.4043/29508-MS","url":null,"abstract":"\u0000 The U.S. Bureau of Ocean Energy Management (BOEM) and Bureau of Safety and Environmental Enforcement (BSEE) are required to consider the effects of their permitted actions on cultural resources, per the National Historic Preservation Act (1966). After the 2010 Deepwater Horizon oil spill in the Gulf of Mexico, BOEM, BSEE, and study partners implemented a multidisciplinary approach to assess micro- to macro-scale impacts from the spill on a select number of previously investigated deepwater shipwrecks. The Gulf of Mexico Shipwreck Corrosion, Hydrocarbon Exposure, Microbiology, and Archaeology (GOM-SCHEMA) Project collected microbiological, geochemical, and archaeological data at wooden- and metal-hulled shipwrecks within and outside of the spill-impacted area for a comparative analysis. Archaeologists documented the selected shipwrecks’ post-spill state of preservation using 3D optical and acoustic scanning systems for comparison with existing geophysical and visual data collected prior to the spill.\u0000 Marine archaeologists are using 3D scanning systems as new tools for recording and interpreting shipwreck sites and analyzing site formation processes in the marine environment. The GOM-SCHEMA Project combined 3D laser data from an Autonomous Underwater Vehicle (AUV) with 3D sonar data collected by a Remotely Operated Vehicle (ROV) as a first step toward monitoring efforts that will inform BOEM and BSEE of the spill’s long-term impacts on deepwater shipwreck preservation. Employing microbial ecological analyses and laboratory-based corrosion experiments that identified micro-scale impacts, scientists gained a better understanding of how the spill affected the natural processes of metal corrosion and wood degradation. 3D imaging creates a permanent digital record that allows scientists to study minute details and also serves as an important outreach tool by allowing the public to virtually explore archaeological resources. By comparing 3D scans collected repeatedly at the same sites over time, especially the sub-centimeter-accurate 3D laser data, archaeologists can quantitatively analyze changes occurring on these sites as a result of enhanced corrosion or degradation. Collectively, these datasets can inform archaeologists and submerged cultural resource managers about site stability, formation processes such as sedimentation and scouring, and the long-term impacts of a major oil spill on submerged cultural resources.","PeriodicalId":11149,"journal":{"name":"Day 1 Mon, May 06, 2019","volume":"18 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2019-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85127710","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Lee J. Thomas, T. Wood, A. Pak, L. Liebana, D. McLaurin, Stephen Stokes
The Pseudo Dry Gas (PDG) technology / concept has been demonstrated for transporting wet gas in a long subsea tieback pipeline (200 km) in deep water depths (1.8 km) under wet gas conditions (water saturated gas) [Ref.1] along with a state of the art technology review of existing solutions. When a multiple of these in-line / piggable liquid removal units are used, they help to reduce the well back pressure by reducing the liquid content to an extent where ‘dry gas’ pressure losses are seen. Therefore, this mitigation of the gravitational pressure drop allows the use of larger pipelines to minimise the frictional pressure drop. This in turn increases recovery of reserves and allows tie back distances to be enhanced. The objective of this paper is to investigate a Pseudo Dry Gas System (PDGS) for an ultra-long deep-water gas condensate development, building upon the research and development already conducted with Strathclyde University. This work was undertaken using non-standard flow assurance methodologies and simulations recycling data and results with the advanced Computational Fluid Dynamics simulations of the liquid removal units behaviour, over various operational boundary conditions. Engagement with subsea equipment suppliers based on the flow assurance results has been undertaken. This paper describes how gas condensates within a subsea tieback system behave very differently to condensed water from a wet gas system and therefore a pseudo dry gas system needs to be configured differently for gas condensate developments. These differences include how and where the liquid drops out of the gas phase, where and if the free liquid is reabsorbed back into the gas stream and how the bubble point of condensate is equal to or very close to liquid removal units operating pressure; this greatly impacts the liquid handling system compared to a wet gas (water) design. Therefore, to ensure controlled liquid only transportation, careful examination of the liquid removal units performance, the liquid pump selection criteria and optimisation of the system needs to be undertaken. This results in a trade-off between maximum reserve recovery and system complexity. The paper demonstrates that the liquid condensate system will remain as a single liquid phase pipeline, where the number of pumps can be reduced and the pump power requirements are very low and within the existing technically qualified subsea pumps.
{"title":"Ultra-Long Subsea Gas Condensate Tie Back – Pseudo Dry Gas – Liquid Handling System","authors":"Lee J. Thomas, T. Wood, A. Pak, L. Liebana, D. McLaurin, Stephen Stokes","doi":"10.4043/29332-MS","DOIUrl":"https://doi.org/10.4043/29332-MS","url":null,"abstract":"\u0000 The Pseudo Dry Gas (PDG) technology / concept has been demonstrated for transporting wet gas in a long subsea tieback pipeline (200 km) in deep water depths (1.8 km) under wet gas conditions (water saturated gas) [Ref.1] along with a state of the art technology review of existing solutions. When a multiple of these in-line / piggable liquid removal units are used, they help to reduce the well back pressure by reducing the liquid content to an extent where ‘dry gas’ pressure losses are seen. Therefore, this mitigation of the gravitational pressure drop allows the use of larger pipelines to minimise the frictional pressure drop. This in turn increases recovery of reserves and allows tie back distances to be enhanced.\u0000 The objective of this paper is to investigate a Pseudo Dry Gas System (PDGS) for an ultra-long deep-water gas condensate development, building upon the research and development already conducted with Strathclyde University. This work was undertaken using non-standard flow assurance methodologies and simulations recycling data and results with the advanced Computational Fluid Dynamics simulations of the liquid removal units behaviour, over various operational boundary conditions. Engagement with subsea equipment suppliers based on the flow assurance results has been undertaken. This paper describes how gas condensates within a subsea tieback system behave very differently to condensed water from a wet gas system and therefore a pseudo dry gas system needs to be configured differently for gas condensate developments. These differences include how and where the liquid drops out of the gas phase, where and if the free liquid is reabsorbed back into the gas stream and how the bubble point of condensate is equal to or very close to liquid removal units operating pressure; this greatly impacts the liquid handling system compared to a wet gas (water) design. Therefore, to ensure controlled liquid only transportation, careful examination of the liquid removal units performance, the liquid pump selection criteria and optimisation of the system needs to be undertaken. This results in a trade-off between maximum reserve recovery and system complexity.\u0000 The paper demonstrates that the liquid condensate system will remain as a single liquid phase pipeline, where the number of pumps can be reduced and the pump power requirements are very low and within the existing technically qualified subsea pumps.","PeriodicalId":11149,"journal":{"name":"Day 1 Mon, May 06, 2019","volume":"22 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2019-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74171607","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Zhong-qiang Liu, F. Nadim, S. Lacasse, B. Lehane, Y. Choi
The paper describes two approaches for deriving the mean, standard deviation and probability density function of the method uncertainty for five axial capacity pile design methods, namely the API, Fugro, ICP, NGI and UWA methods. A new unified database of pile load tests recently developed in a joint industry research project (Lehane et al., 2017) is used for the quantification of method uncertainty. The focus of this paper is on the statistical description of the method uncertainty parameters for each of the pile design methods for predicting the axial capacity of piles in sand and in clay. Probabilistic calculations of the axial pile capacity for typical offshore piles using the above five design methods (API, Fugro, ICP, NGI and UWA methods) showed that method uncertainty is a major contributor to the uncertainty in pile foundation capacity. The method uncertainty has therefore a strong influence on the calculated annual probability of failure, and thus on the associated safety level. Establishing the statistics of the error in a capacity prediction model from the measured values (Qm) in pile load tests and the calculated values (Qc) of pile capacity requires careful consideration of several factors. Issues of importance to the derivation of method uncertainty statistics include the effect of different sized databases for the different pile design methods, the effect of case histories with particularly low Qm/Qc values and the possible dependence of method uncertainty on pile length and/or pile diameter. The paper presents two different interpretations for the characterization of method uncertainty and demonstrates their application through a case study for an offshore piled jacket. The effect of method uncertainty on the calculated annual probability of failure is illustrated.
本文介绍了API、Fugro、ICP、NGI和UWA五种轴向承载力桩设计方法的方法不确定性均值、标准差和概率密度函数的推导方法。最近在一个联合行业研究项目(Lehane et al., 2017)中开发了一个新的桩荷载试验统一数据库,用于方法不确定性的量化。本文的重点是对各种桩设计方法的方法不确定性参数的统计描述,以预测砂中和粘土中桩的轴向承载力。采用上述五种设计方法(API法、Fugro法、ICP法、NGI法和UWA法)对典型近海桩轴向桩承载力进行概率计算,结果表明,方法不确定性是桩基承载力不确定性的主要来源。因此,方法的不确定性对计算的年失效概率有很大的影响,从而对相关的安全水平也有很大的影响。根据桩荷载试验的实测值(Qm)和桩承载力的计算值(Qc)建立承载力预测模型的误差统计,需要仔细考虑几个因素。方法不确定性统计推导的重要问题包括不同桩设计方法的不同规模数据库的影响、Qm/Qc值特别低的历史案例的影响以及方法不确定性对桩长和/或桩径的可能依赖。本文对方法不确定性的表征提出了两种不同的解释,并通过海上堆置套管的实例说明了它们的应用。说明了方法不确定性对计算出的年失效概率的影响。
{"title":"Method Uncertainty for Five Axial Pile Capacity Design Methods","authors":"Zhong-qiang Liu, F. Nadim, S. Lacasse, B. Lehane, Y. Choi","doi":"10.4043/29514-MS","DOIUrl":"https://doi.org/10.4043/29514-MS","url":null,"abstract":"\u0000 The paper describes two approaches for deriving the mean, standard deviation and probability density function of the method uncertainty for five axial capacity pile design methods, namely the API, Fugro, ICP, NGI and UWA methods. A new unified database of pile load tests recently developed in a joint industry research project (Lehane et al., 2017) is used for the quantification of method uncertainty. The focus of this paper is on the statistical description of the method uncertainty parameters for each of the pile design methods for predicting the axial capacity of piles in sand and in clay. Probabilistic calculations of the axial pile capacity for typical offshore piles using the above five design methods (API, Fugro, ICP, NGI and UWA methods) showed that method uncertainty is a major contributor to the uncertainty in pile foundation capacity. The method uncertainty has therefore a strong influence on the calculated annual probability of failure, and thus on the associated safety level. Establishing the statistics of the error in a capacity prediction model from the measured values (Qm) in pile load tests and the calculated values (Qc) of pile capacity requires careful consideration of several factors. Issues of importance to the derivation of method uncertainty statistics include the effect of different sized databases for the different pile design methods, the effect of case histories with particularly low Qm/Qc values and the possible dependence of method uncertainty on pile length and/or pile diameter. The paper presents two different interpretations for the characterization of method uncertainty and demonstrates their application through a case study for an offshore piled jacket. The effect of method uncertainty on the calculated annual probability of failure is illustrated.","PeriodicalId":11149,"journal":{"name":"Day 1 Mon, May 06, 2019","volume":"68 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2019-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88006091","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}