Date
|
Title
|
Register
|
Wednesday,
April 22nd
14:00 CET
Presented by:
Aurore Plougoulen
|
Powerful and Intuitive Data Interrogation for the Efficient Post-processing of Flow Simulations with Tempest™ VIEW
Data visualization and interrogation are key to understanding flow simulation results. When handling multiple simulations consisting of millions of cells and thousands of wells, this task can be very tedious for reservoir engineers.
Tempest™ VIEW is a fast and memory-efficient system than can reliably process a high number of simulations. It provides reservoir engineers with the tools needed to accelerate the decision-making process through unrivaled analysis of simulation results.
This presentation features Tempest VIEW functionalities that make data interrogation intuitive and powerful, and provide efficient simulation post-processing to a wide range of third-party simulators.
|
.png.aspx) |
Wednesday,
April 29th
14:00 CET
Presented by:
Aurore Plougoulen
|
Increase Confidence in Decision-Making through the Efficient Handling, Visualization and Analysis of Ensemble-Based Simulations
Ensemble-based methods are now recognized as efficient reservoir simulation techniques for uncertainty analysis and high-quality history matching. These methods provide an ensemble of equiprobable models carrying highly valuable statistics and representing uncertainty in the model and its dynamic response. However, most post-processing simulation solutions on the market have not been designed to handle ensembles, so that reservoir engineers often struggle to deal with a high number of simulations, manipulate them as a coherent ensemble, and use the statistics resulting from the models.
This presentation demonstrates an interactive workflow based on powerful data analytics using Tempest, allowing users to visualize, analyze and easily highlight the value of an ensemble. This approach is flow simulator agnostic and can be used in combination with various ensemble methods, enabling efficient peer and partner reviews that lead to a better understanding of the risks, and enlightened decisions.
The featured application is a comprehensive solution for ensemble-based simulation post-processing aimed at any asset team eager to make the most of its data.
|
.png.aspx) |
Wednesday,
May 6th
14:00 CET
Presented by:
Camille Cosson
|
Break Through Uncertainties and Design an Optimal Field Development Plan – Big Loop #1
This webinar will be the first of three virtual talks discussing the benefits of Big Loop, an automation and integration solution designed to support strategic field management decisions.
This first talk will present an integrated reservoir modeling case study. We will go through the steps: workflow automation, assisted history match, risk assessment to finally design not only a good enough, but an optimal field development plan, given all the information available. The solution has been implemented for several fields and benefits have been quantified: reduced costs and increased production.
Looking back, asset teams using Big Loop have leveraged its technology capabilities to change their approach to reservoir modeling, and adopt new collaboration behaviors that result in increased productivity. This will be presented in the following two talks.
|
.png.aspx) |
Wednesday,
May 13th
14:00 CET
Presented by:
Camille Cosson |
Big Loop Evergreen Workflows – Big Loop #2
This webinar will be the first of three virtual talks discussing the benefits of Big Loop, an automation and integration solution designed to support strategic field management decisions.
In the first talk, field development plan optimization with Big Loop was discussed, and a case study was presented to demonstrate the solution’s advanced capabilities for geoscience workflow automation and risk analysis.
In this second talk, we will show how adopting such a solution catalyzes changes in the approach to reservoir modeling. The ability to integrate and propagate new data throughout the workflow is a key component in shortening the time cycle in reservoir studies, reducing manual repetitive tasks, and leaving more space for analysis and interpersonal collaboration. Moreover, it changes the reservoir modeling paradigm. As discussed in the first talk, we no longer use one model, but an ensemble of models that simultaneously capture both what is known and the uncertainties. With Big Loop, understanding of the reservoir is no longer represented by a fixed reference model; rather, it is defined by the sum of all the automated G&G processes leading to a representation of the reservoir. These processes can be modified to reflect changes as new data or new understanding of the subsurface arises. A reservoir model is now evolving: from a simple geologic concept in the early stages, it is refined and updated with complexities throughout the field’s lifecycle. This is what is called an evergreen model.
|
.png.aspx) |
Wednesday,
May 20th
14:00 CET
Presented by:
Camille Cosson |
Agile Collaboration with Big Loop workflows – Big Loop #3
This webinar is the last of three virtual talks discussing the benefits of Big Loop, an automation and integration solution designed to support strategic field management decisions.
In this third talk, we will show how adopting such a solution catalyzes changes in asset team collaboration. Implementing a Big Loop workflow provides a common basis for geoscientists that captures with high fidelity a global understanding of the field across a broad range of disciplines. Every geoscientist in the asset team owns the workflow. They can leverage the forward and feedback loop to learn more and gain a better understanding of the field features that lie within their disciplines. Team members can test and learn within their own domain, while also gauging the impact of their work on the overall results and the relationship with other jobs. This creates a powerful communication channel within the asset team. Daily communication is established when new data arrives, but also at any time, if new problems or questions arise. With Big Loop, operations geoscientists are liberated from their own-discipline blinders, and the overall team becomes more autonomous, better organized, and more agile.
|
.png.aspx) |
Wednesday,
May 27th
11:00 CET
Presented by:
Ross Findlay |
Reveal More from Your Seismic Amplitudes Using Scenario Testing to Understand the Impact of Layer Thickness and Rock Properties
See how Emerson’s 2D modeling capabilities can be used to generate a series of what-if scenario tests to better understand your seismic amplitude response. Learn how to create wedge models to investigate the effects of reservoir thickness, create fluid 2D models to see how changes in water saturation and porosity affect seismic character, and utilize tools such as amplitude tracking plots and crossplots to analyze the results. |
.png.aspx) |
Wednesday,
June 10
14:00 CET
Presented by:
Philippe Ecoublet |
4D Time Lapse Seismic Reservoir Characterization - Volve Field ,Norwegian North Sea
Reservoir monitoring through time-lapse analysis requires specific quantitative seismic interpretation workflows to capture the evolution of the reservoir between baseline and monitor surveys. This presentation illustrates an approach based on rock and fluid property analysis from prestack seismic inversion.
Interpreting a change in seismic response to a change in pore fluid content requires an accurate characterization of the baseline seismic survey, as well as a precise understanding of how elastic properties will be impacted by time-lapse production.
Rock Physics and synthetic modeling with fluid substitution are implemented to support reservoir characterization from a three-term prestack inversion of baseline and monitor seismic surveys. 4D interpretation is carried out using crossplots to map hydrocarbon production over time-lapse analysis.
|
.png.aspx) |