WO2011109839A2 - Active noise injection computations for improved predictability in oil and gas reservoir discovery and characterization - Google Patents

Active noise injection computations for improved predictability in oil and gas reservoir discovery and characterization Download PDF

Info

Publication number
WO2011109839A2
WO2011109839A2 PCT/US2011/027456 US2011027456W WO2011109839A2 WO 2011109839 A2 WO2011109839 A2 WO 2011109839A2 US 2011027456 W US2011027456 W US 2011027456W WO 2011109839 A2 WO2011109839 A2 WO 2011109839A2
Authority
WO
WIPO (PCT)
Prior art keywords
data
voxel
attribute
seismic
well
Prior art date
Application number
PCT/US2011/027456
Other languages
French (fr)
Other versions
WO2011109839A3 (en
Inventor
Sandeep Gulati
Original Assignee
Vialogy Llc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vialogy Llc filed Critical Vialogy Llc
Priority to RU2012142308/08A priority Critical patent/RU2012142308A/en
Priority to CN2011800223235A priority patent/CN102893183A/en
Priority to BR112012022450A priority patent/BR112012022450A2/en
Priority to EP11751523A priority patent/EP2542918A2/en
Priority to MX2012010271A priority patent/MX2012010271A/en
Priority to CA2792052A priority patent/CA2792052A1/en
Publication of WO2011109839A2 publication Critical patent/WO2011109839A2/en
Publication of WO2011109839A3 publication Critical patent/WO2011109839A3/en

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01VGEOPHYSICS; GRAVITATIONAL MEASUREMENTS; DETECTING MASSES OR OBJECTS; TAGS
    • G01V1/00Seismology; Seismic or acoustic prospecting or detecting
    • G01V1/28Processing seismic data, e.g. for interpretation or for event detection
    • G01V1/30Analysis
    • G01V1/306Analysis for determining physical properties of the subsurface, e.g. impedance, porosity or attenuation profiles

Definitions

  • seismic surveys are used to estimate features of interest of subsurface geology.
  • Seismic surveys use controlled seismic energy, such as produced by specialized air guns or seismic vibrators.
  • a receiver senses seismic energy, typically in the form of a wave, reflected by subsurface features.
  • the subsurface features are detected by analyzing the time it takes for reflected seismic waves to travel through the subsurface matter of varying densities.
  • 3-D seismic also uses seismic energy to produce a 3-dimiensional map of subsurface formations.
  • Traditional techniques of analyzing seismic data attempts to filter out noise to identify a signal of interest. However, in the process of filtering out noise, important information of interest can be lost, resulting in a seismic map from which various features may be difficult to distinguish.
  • an exploration well is drilled in an attempt to conclusively determine the presence or absence of oil or gas.
  • an exploratory well can be very expensive especially for off-shore wells and is subject to the risk that the well will be unproductive.
  • Systems and techniques are disclosed for improving signal-to-noise in onshore and offshore seismic and electromagnetic acquisitions, for improving microseismic techniques for hydraulic fracture monitoring and optimization, and for interferometric acquisition for directing hydrocarbon imaging using conventional 3D vibrators.
  • Seismic data is analyzed using a non-linear stochastic quantum energy source to create resonances through the interaction between virtual vibration and the seismic data. Changes in resonance yields information of interest from the seismic data not obtainable by noise filtering. The signature of such changes in resonance can then be determined and calibrated as porosity, fluid, or lithology. Reservoir characterization and drilling decisions can then be made on the resulting volumes of reservoir property.
  • Such systems and techniques can be used to de-risk new drilling locations and to redevelop under-performing or abandoned wells by locating nearby pay zones that may have just been missed.
  • the final porosity in these carbonates may or may not be related to the depositional environment. Also, unlike other lithologies, the original primary porosity in carbonates may be totally destroyed during diagenesis and significant new secondary porosity may be created. So to predict well productivity, certain patterns of natural fracturing and high continuous porosities are identified from 3D seismic data.
  • the systems and techniques disclosed herein can detect such subtle changes in lithology. Conventional data can be analyzed according to the disclosed systems and techniques to identify such subtle changes. The systems and techniques disclosed herein can also be used to identify brittleness in shale— the ability to fracture the reservoir.
  • a method includes voxelizing seismic data for a geological subsurface formation of interest into multiple voxels having respective locations in the formation of interest; and determining whether a voxel in the multiple voxels includes an attribute; and outputting whether the voxel includes the attribute to an attribute volume based on the ' location; and performing the determining and the outputting for at least some other of the multiple voxels.
  • the method can further include performing the steps listed above for the seismic data for another formation of interest.
  • the determining whether a voxel in the multiple voxels includes an attribute can include determining whether the voxel in the multiple voxels includes a porosity attribute; and the method further can include assembling the at least some of the other of the mulitple voxels into a porosity cube.
  • the determining whether a voxel in the multiple voxels includes an attribute can also include determining whether the voxel in the multiple voxels includes a Uthology attribute; and the method further can include assembling the at least some of the other of the multiple voxels into a lithology cube.
  • the determining whether a voxel in the multiple voxels includes an attribute can also include determining whether the voxel in the multiple voxels includes at least one of gas, oil, or water; and the method further can include assembling the at least some of the other of the multiple voxels into a liquid cube.
  • the method can also include wherein the seismic data can include migrated prestack time gathers for the formation of interest.
  • the determining whether a voxel in the multiple voxels includes an attribute can include determining whether the voxel in the multiple voxels includes a lithology attribute; repeating the determining and the outputting for a porosity attribute; and generating a drill-here map indicating a location for drilling based on the at least some of the other of the multiple voxels for the lithology attribute and based on the at least some of the other of the multiple voxels for the porosity attribute.
  • the voxelizing the seismic data further can include spectrally decomposing the seismic data into frequency volumes. L
  • the method can further include spectrally decomposing the seismic data into a high frequency volume, a medium frequency volume, and a low frequency volume.
  • the the seismic data can include three dimensional seismic data and the voxelizing can include generating, from the three dimension seismic data, multiple one dimension voxel vectors.
  • the determining whether a voxel in the multiple voxels includes an attribute including: obtaining control data for that attribute based on information from one or more previously explored geological subsurface wells; and processing the one dimension voxel vector for the voxel using the control data and using quantum resonance interferometry to detect a resonance.
  • the method can further include characterizing a structural trap based on the attribute volume.
  • the method can also include characterizing a stratigraphic trap based on the attribute volume. Additionally, the method can include upscaling the multiple voxels to a predetermined resolution.
  • a method can include obtaining a one-dimensional voxel vector for a voxel from multiple voxels associated with a formation of interest; obtaining spectral data generated from well-log data associated with an attribute of interest; coupling the spectral data with the one dimensional voxel vector to determine whether a resonance event occurs; when a resonance event occurs, producing an output indicating that the voxel has the attribute of interest; and when the resonance event does not occur, producing an output indicating that the voxel does not have the attribute of interest.
  • the obtaining the one-dimensional voxel vector include normalizing amplitudes corresponding to the one dimensional voxel vector to fit within a range.
  • the obtaining the spectral data generated from well-log data can include: obtaining seismic noise data; obtaining well control voxel data from the well log data; and combining the seismic noise data and the well control voxel data using a quantum mechanical model to produce the spectral data.
  • the obtaining well control voxel data can include determining an energy spectral density voxel from the well-log data for the attribute using wavelets derived from sonic logs in sections of a well with the attribute of interest.
  • the combining the seismic noise data and the well control data can include combining the seismic noise data and with well control voxel data for a voxel associated with the well- log data that exhibits the attribute.
  • the obtaining the well control voxel data can also include obtaining a well control voxel data of a voxel not exhibiting the attribute; and the obtaining the seismic noise data can include generating the seismic noise data based on the well control voxel data of the voxel not exhibiting the noise.
  • the coupling the spectral data with the one dimensional voxel vector to determine whether a resonance event occurs can include using a nuclear magnetic resonance ("NMR") master rate equation to generate quantum stochastic resonance based on the one-dimensional voxel vector, synthetic noise, and the spectral data.
  • NMR nuclear magnetic resonance
  • a method can include obtaining voxel data for a voxel from voxelized seismic data for a geological subsurface formation; performing a first nonlinear coupling of the voxel data with spectral data to generate a first resonance, wherein the coupling is driven by noise having an intensity within a first cutoff band; in response to generating the first resonance, adjusting the cutoffband to a second cutoffband, different than the first cutoff band; performing a second non- linear coupling of the voxel data with the spectral data associated with an attribute of the subsurface formation to generate a second resonance, wherein the second coupling is driven by noise having an intensity within a second cutoffband; in response to generating the second resonance, generating an indication that the attribute exists in the voxel; and in response to the second coupling not producing a second resonance producing an indication that the attribute does not exists in the voxel.
  • the first resonance can include a first quantum stochastic resonance.
  • the method can also include assembling the voxel data in an attribute volume with the indication that that the attribute exists in the voxel.
  • the voxelized seismic data can also include multiple voxels including the voxel; and the method can further include determining the first cutoffband based on a percentage of an average intensity of the multiple voxels.
  • the method can also include prior to the first coupling, performing a third coupling of the voxel data with the spectral data associated with an attribute of the subsurface formation to generate a third resonance, wherein the third coupling is driven by noise having an intensity within a third, different cutoff band; and in response to generating the first resonance, adjusting the third cutoff band to the first cutoff band.
  • a method includes obtaining first prospect voxel data for a voxel from voxelized seismic data for a geological subsurface formation on a prospect;
  • determining an attribute iteration range including values for the attribute including at least a first value and a second value determining control well voxel data from well-log data for an existing well, the control well voxel data including spectral energies for a listing of values of the attribute found in the existing well, including a first spectral energy for the first value and a second spectral energy for the second value; coupling the prospect voxel data with spectral data having parameters set according to the first spectral energy to produce a first resonance; and in response to the first resonance, generating an output indicating that the voxel includes the attribute at at least the first value.
  • the method further includes- coupling the prospect voxel data with spectral data having parameters set according to the second spectral energy to produce a second resonance; and in response to the second resonance, generating an output indicating that the voxel includes the attribute at at least the second value.
  • the control well voxel data can include a third spectral data corresponding to a third value for the attribute; and the method can further include coupling the prospect voxel data with spectral data having parameters set according to the third spectral energy to produce a third resonance; and in response to the third resonance not being produced, generating an output indicating that the voxel does not include the attribute at the third value but includes the attribute at the first and second value.
  • the attribute can include porosity and the first value can include a first percentage of porosity and the second value can include a third percentage of porosity greater that the first percentage.
  • Figure 1 shows an example process for reservoir characterization.
  • Figure 2 shows an example of a graphical depiction of elements involved in process for reservoir characterization.
  • Figure 3 shows applications of the systems and techniques for reservoir characterization.
  • Figure 4 shows examples of subsurface reservoir traps.
  • Figure 5 shows an example of a flow chart for processing seismic data for a prospect prior to the seismic data being voxelized.
  • Figure 6 enumerates examples of some subsurface volumes and their taxonomy for a prospect.
  • Figure 7 shows an example of a process for voxelizing seismic data for a formation of interest.
  • Figures 8A-8B show an example of a work flow for analyzing voxelized data.
  • Figures 9A-9D show an example of a process for determining seismic noise data .
  • Figure 10 describes a process for upscaling or upconverting analysis of attributes of interest.
  • Figure 11 shows a process for establishing monotonicity of a resonance event.
  • Figures 12A through 12B describe a process for generating and normalizing excitation cascade data.
  • Figures 13A-13D show examples of how attribute-specific well control voxel data is developed.
  • Figure 14 describes an example of a process of implementing an interferometry engine and a quantum resonance interferometry processor.
  • Figure 15 shows an example of a double well function.
  • Figure 16 shows an example of work flow using a quantum resonance interferometry engine.
  • Figures 17A -17D show examples of a processes for computing attribute data from seismic data.
  • Figures 18A - 18B show configuration of a quantum resonance interferometry engine in a training mode and an operations mode.
  • Figure 19 shows a method for calibration of quantum resonance interferometry engine Parameters.
  • Figure 20 shows an example workflow for producing output with various processes discussed herein.
  • Figure 1 shows an example process 100 for reservoir characterization (also known as lateral subsurface prediction of reservoirs).
  • Process 100 can be used, in some examples, to build a computer model of a reservoir beneath the surface of earth that incorporates all the characteristics of the reservoir that are pertinent to its ability to store hydrocarbons and also to produce them.
  • the process 100 can be used to discover, delineate, and size hydrocarbon (oil and natural gas) reservoirs to produce, among other things, a drill map for onshore or offshore wells using seismic data.
  • prospect data and exploration criteria are obtained. Exploration criteria can include targeted exploration objectives such as criteria for a type or types of formations to be searched for in a prospect— a location being analyzed for a deposit.
  • a prospect is defined as an area covering a potential subsurface trap believed to contain hydrocarbons, geological factors that have to be present for a prospect to produce oil and gas include: presence of a source rock (organic rich rock that has been subjected to high pressure and temperature over an extended period of time to form hydrocarbon), a structural, stratigraphic or combination trap to hold the hydrocarbons, impermeable seal or cap rock over the hydrocarbon trap in order to prevent hydrocarbons migrating or escaping to the surface, and porous reservoir rock that collects oil within its pores and that is permeable so that the hydrocarbons will flow to surface during production.
  • a source rock organic rich rock that has been subjected to high pressure and temperature over an extended period of time to form hydrocarbon
  • a structural, stratigraphic or combination trap to hold the hydrocarbons
  • impermeable seal or cap rock over the hydrocarbon trap in order to prevent hydrocarbons migrating or escaping to the surface
  • porous reservoir rock that collects oil within its pores and that is permeable so that the hydro
  • exploration criteria can include a target formation with a predetermined porosity, fluid type, and lithology for a predetermined type of subsurface material.
  • porosity of a porous source rock or sediment describes the fraction of void space in the material, where the void may contain fluid such as water, oil or gas ,and is defined by the ratio:
  • Vv is the volume of void-space (such as fluids) and V- ⁇ is the total or bulk volume of material, including the solid and void components.
  • Porosity is a complex function of many factors such as grain size and overlying sediments.
  • Figure 2 shows an example of a graphical depiction 200 of elements involved in the process 100.
  • Figure 2 shows examples of prospect data at 210A, 210B, and 2 IOC.
  • Prospect data can include seismic and other geophysical data, rock physics data, and well log data for the prospect or well log data from an analogous formation on another prospect.
  • Geophysical data can include for example seismic data 210A obtained for an area of interest, such as for a prospect area 211.
  • Obtaining geophysical data can include, for example, receiving the geophysical data already acquired from a prospect area, at a computing device for further processing and analysis. In some examples, obtaining the data can include receiving the data, actively requesting the geophysical data or acquiring the geophysical data from the prospect area 211.
  • Seismic data 210A can include already processed, migrated pre-stack time gathers (denoted as a Pre-stack Time Migration (PSTM) volume); migrated pre-stack depth gathers (denoted as a Pre-stack Depth Migration (PSDM) volume); or raw seismic gathers.
  • the gathers can be stacked into a seismic volume, such as seismic volume 220, in the format of a PSTM stack or PSDM stack in an SEGY format.
  • the seismic volume can be derived from 2- D seismic acquisition, 3-D seismic acquisition process, or multicomponent seismic process over a seismic acquisition grid 213.
  • Seismic data can also include a fold map 214 which includes a multiplicity of recorded seismic reflection data.
  • the fold map 214 can be used to ensure high acquisition signal-to-noise and sufficient reflected energy for subsequent computations.
  • a 3-D seismic survey fold map may be calculated using the relationship:
  • Method of 100 applies to standard compression wave 2-D and 3-D seismic acquisition, 4-D time lapse seismic where measurements are repeated over time, 4-C converted wave multi-component seismic that uses a compression wave (p-wave) source and both compression wave and shear wave(s-wave) receivers, and 9C full wave multicomponent using both compression wave and shear wave source and receivers, (converted wave with compression source and shear or 9C multicomponent seismic).
  • Source line spacing refers to the distance between successive rows of way-points at which acoustic excitations are sent into the ground.
  • Various modahties of seismic acquisition are familiar to one skilled in the art.
  • acquisition processes using a seismic acquisition grid 213 record a large square or rectangular patch, depending on the area of interest, such as a mineral lease owned by a prospect owner.
  • Useful data obtained at an area of interest can be offset by several geophysical factors. For example, geophysical factors can limit a maximum useful radius for geophone coverage.
  • a bin By moving a patch and recording more salvos using simultaneous discharge of excitation vibrations from multiple vibrator trucks or air guns as an example, into the group of source points, one can accumulate overlapping subsurface coverage and build statistical repetition over each subsurface reflecting area referred to as a "bin".
  • the quality of the subsurface image can be related to the statistical diversity of the information recorded for each bin of sub-surface coverage. For example a bin represents the smallest area of a 3-D survey that contains the entire survey statistics. The higher the number of observations obtained, that contain unique measurements of the acoustic reflections from a subsurface, the better the reconstruction is of the subsurface geological configuration that caused those observations. Terms bin and CMP (common midpoint) bin are used interchangeably.
  • Prospect data also includes well log data 2 IOC.
  • Well logs provide physical rock properties for an identified petrophysical zone. These are available from wells that have already been drilled on the prospect area 211 under consideration or from wells drilled into similar source rock at another analogous prospect area.
  • Well logs include sonic logs, bulk density logs, gamma ray logs, neutron porosity logs, and resistivity logs.
  • a sonic well log, such as dipole sonic log can also include sampling of both compression (p-wave) velocity and shear (s-wave) velocity within a well bore.
  • Basic Well Log Analysis for Geologists by George Asquith and Charles Gibson, 1982, pages 1 - 215 and references therein, published by American Association of Petroleum Geologists, Tulsa, Oklahoma provide an introduction and discussion of different wells logs and their relationships.
  • Well log data can be used to calibrate and tie actual rock and fluid properties for a formation of interest at a known well, to seismic data for the known well. Such well log data can also be used to interpret attributes found in the seismic data for the known well.
  • well log data 210C for the seismic volume can be used to predict rock and fluid properties in a seismic volume using seismic data from the seismic volume of a prospective well.
  • the well log data 2 IOC can been edited and corrected for issues such as noise bursts, cycle skips and fluid replacements or can be obtained in such a format.
  • Prospect data can also include rock physics data 210B includes information regarding type of geological source rock, such as sandstone, limestone, dolomite, and shale, being targeted for exploration, such as is identified by the exploration criteria.
  • Rock physics relates the geological properties (e.g. porosity, lithology, saturation) of a rock at certain physical conditions (e.g. pressure, temperature) with their corresponding elastic and seismic properties (e.g. elastic modulus, velocity, p-wave impedance, s-wave impedance).
  • elastic and seismic properties e.g. elastic modulus, velocity, p-wave impedance, s-wave impedance.
  • derived attributes such as brittleness, defined as the ratio of compressive strength to the tensile strength for the rock is derived from Poisson's Ratio and Young's modulus, measured based on rock core samples in the lab. Brittleness can be derived using Vp V s ratio where Vp, Vs denotes p-wave and s-wave velocities respectively or
  • rock physics tables can be obtained for basic lithological and geomechanical properties of the rock types and their granularity combinations.
  • the source rock types can be identified by well logs 2 IOC for existing wells including seismic data obtained for those existing wells.
  • Rock physics data 210B includes properties enumerated in terms of sonic compression-wave velocity, shear- wave velocity, brittleness, Young's modulus, and relationships for sonic velocity changes across different rock facies.
  • a subsurface formation of interest of the subsurface volume 220 is identified for further analysis.
  • the formation of interest is located in a prospect located in the prospect area 211.
  • the subsurface formation can include for example layer 225 shown in Figure 2.
  • a prospect may have multiple separated hydrocarbon bearing pay-zones present in the subsurface.
  • a Permian Basin well-bore may have hydrocarbon pay-zones within the Spraberry, Dean, Clearfork, Wolfcamp, Atoka, Strawn, Devonian, Fusselman, Ellenberger formations, all of which need to be individually selected and independently analyzed as in 115.
  • Analysis 100 may be used to analyze one or more or all of the formations between surface and maximum depth covered by the seismic data.
  • a horizon is a reflecting surface separated by different layers in depositional environments and characterized by different seismic reflection properties. These are separated by curves in 2-D data and surfaces or facies in 3D data. Facies is defined as a distinctive rock unit that forms under certain conditions of
  • seismic horizons are edges on the seismic image serving to delineate formations such as structural and stratigraphic features (such as faults) and patterns.
  • Oz Yilmaz's book referred above provides a detailed treatment for conventional methods to determine seismic horizons in seismic volumes.
  • the seismic data in the seismic volume 220 for the prospect is processed for voxelization.
  • Voxelization is defined as the computational process of decomposing seismic volume into 3-D volume cells which are then analyzed for rock and fluid properties.
  • the term voxelized data refers to seismic volumes that have been decomposed into smaller 3- D volume cells and to data relating to attributes for those 3-D volume cells.
  • the seismic data is spectrally decomposed into separate spectral decompositions. Spectrally decomposing the voxelized data can help in determining frequency bounds for prospect voxel data (also referred to as prospect voxels or as voxelized data) in step 120.
  • prospect voxel refers to a singular prospect voxel data while prospect voxels refers to plurality of prospect voxel data.
  • the seismic data can be broken up into frequency separated volumes such as separated by high (e.g., between 80Hz and 100Hz for onshore data), low (e.g., between 3 Hz and lOHz), and prime amplitude (e.g., between 12Hz and 70Hz) frequencies.
  • the seismic data 210A, for the seismic volume 220 is voxelized into multiple three-dimensional volume cells, referred to as voxels, for further analysis, as shown by the multiple voxels 235 in Figure 2.
  • voxelizing the seismic data at 120 includes determining a size and orientation of the voxels which depends on the resolution of the acquired seismic data, geological understanding of the depositional processes that led to development of hydrocarbon reservoirs in the prospect area within a field or basin and the formation of potential traps in the subsurface lithological formation of interest on the prospect.
  • the vertical z dimension for a voxel is set to the seismic sample acquisition sampling rate (e.g., 0.5 ms, 1ms, 2ms, or 4ms). The higher the sampling rate the higher the resolution of the voxels and ultimately of the result of analyzing the voxels.
  • the areal x and y dimensions are limited by the seismic survey design or the "bin size" (x by dimension) of the PSTM/PSDM gathers.
  • the seismic data is processed before being obtained at 110, the bin size is established early in the process and final seismic data is bin-size limited. If bin size is coarse in the seismic data, then re-binning and reprocessing is performed to generate smaller bin sizes. If reprocessing is not possible then each voxel is sized to (2 times x-dimension of the bin) by (2 time ⁇ -dimension of the bin) by (sampling rate or processing depth resolution).
  • voxels sizes can be set to as low as (0.5 times the x-dimension of the bin) by (0.5 times the ⁇ -dimension of bin) by (sample rate or depth resolution).
  • the dimension of the voxels determines the final attribute resolution that is produced by the analyzing step 125.
  • the voxelized data is analyzed.
  • the voxelized data is analyzed, voxel by voxel, to determine the existence of attributes as well as the areal and vertical extent of attributes of interest in the seismic data using, for example, quantum resonance
  • the attribute of interest can be defined by the exploration criteria. Examples of attributes of interest are shown at 122, and include: rock properties, porosity, lithology, geomechanical properties, brittleness, fluid presence, type of fluid (oil, gas, brine, and/or water), and anisotropy. For example, for each voxel, a percentage of porosity can be identified, lithology characteristics can be identified, and the existence of fluids can be identified such as water, gas, or oil.
  • the result of the analysis can include multiple attributes for each prospect voxel. For a voxel, the seismic data can be analyzed multiple time for multiple attributes of interest. Examples of how the voxelized data can be analyzed is provided in the discussion in connection with Figures 1 to 20.
  • an attribute volume can be produced at 130.
  • the analyzed voxels are combined according to their original layout in the seismic volume 220 to produce an attribute volume that can include rock and fluid properties that are combined to find areas of high net-pay or favorable combinations of porosity and fluid -type (oil and or gas).
  • the analyzed voxels are assembled into one or more 3 -dimensional attribute volumes, graphically depicted in Figure 2 as a first attribute volume 275 for porosity, a second attribute volume 280 for fluid volume, and a third attribute volume for lithology.
  • the first attribute volume for porosity shows areas of high porosity in the formation of interest.
  • Each of the analyzed voxels in the first attribute volume 275 show a degree of porosity analyzed for the voxel.
  • the second attribute volume 280 shows a distribution of fluid in the formation of interest. For example, the distribution of gas, oil, and water can be depicted in the second attribute volume 280.
  • the third attribute volume 285 shows a distribution of rock properties such as brittleness of the rock.
  • a reservoir is delineated at 140.
  • the exploration criteria can identify cutoff values for which a desired reservoir is desired.
  • a cutoff value can be defined above which the reservoir of interest is delineated.
  • all of the voxels in the attribute having a value above the cutoff can be delineated as part of the reservoir.
  • the cutoff value can be dynamically selected which it turn produces a delineated reservoir of different sizes.
  • data for the delineated reservoir of step 140 can be provided for further processing and output at 150.
  • the data for the delineated reservoir including the analyzed voxel data can be provided to another processing module, such as a flow model, for further processing.
  • the process 100 continues on to step 160 where the data for the delineated reservoir is used to produce one or more drill maps.
  • Figure 2 shows examples of a drill map— a 3-dimensional drill map 292 with an identification of a drill location and a drill trajectory and a 2-dimensionl map also with indicators of where to drill.
  • Figure 3 shows applications of the systems and techniques disclosed in this disclosure.
  • the systems and techniques disclosed herein can be used, for example in oil and gas lease acquisition 330 to provide accurate subsurface mapping, to estimate hydrocarbon formation boundaries, to optimize prospect acreage, and to optimally extract hydrocarbon and drill wells.
  • the systems and techniques disclosed herein can be used in seismic survey design 333 by determining subsurface characteristics which provides insightful guidance on how to acquire additional 2D, 3D and multi-component seismic data or setup parameters to achieve high quality imaging.
  • reservoir properties using vintage seismic to drill exploration wells or offset wells can be analyzed, for example using existing legacy data, to develop new surveys and design quality seismic data and optimal frequency spectrum for imaging potential structural/stratigraphic traps, complex fracturing and nonconformal lithologies.
  • the systems and techniques disclosed herein can be used for exploration of new wells. For example, analysis can be performed on end-to-end, available geotechnical, geological, stratigraphy and geoseismic data to discover reservoirs, characterize formation boundaries, porosity and properties, develop net pay models, and recommend drilling targets and mitigate dry-hole risk.
  • well trajectories and well-bore deviations can be designed to maximize recovery from nonconformities and fractures.
  • systems and techniques disclosed herein can be used for positioning development wells in producing or proven fields. For example, locations can be determined for placing offset wells to maximize recovery and minimize the number of wells required to cost-effectively drain hydrocarbon formations. And, the systems and techniques disclosed herein can be used for enhancing currently producing wells by designing perforations and injection laterals for maximal drainage from productive formations and multi-zone wells pays to extend reservoir life and recovery.
  • the methods described in this invention in Figure 1 through 20 can be used to enable and enhance various upstream oil and gas during primary, secondary, enhanced hydrocarbon recovery and redevelopment of fields or reentry of old wells that have been capped or abandoned or extending depth of existing wells to discover new sources of hydrocarbon production and to find bypassed and missed pockets of hydrocarbon.
  • Figure 4 shows examples of subsurface reservoir traps that can be analyzed using the systems and techniques described herein in each of the stages of oil and gas exploration described above in connection with Figure 3.
  • reservoir characterization has been a subjective process with the outcome highly dependent upon the experience and quality of the geophysical interpreter, quality of seismic processing and formation complexity.
  • the types of onshore and offshore formations are diverse and complex forcing analysts to differentiate themselves based on expertise in specific formation types (e.g., structural plays, stratigraphic plays, or combination traps such as salt domes), geological basins, and resource plays (e.g., Permian Basin, Bakken or Gulf of Mexico).
  • Structural Traps formed by deformation of reservoir rock such as anticline or fault, such as Anticline, Normal Fault Trap, Baldheaded Anticline, Reverse Fault Trap, Faulted Anticline, Tilted Fault Block, Dome, Drag Faults on Thrust Fault, Fractured
  • Stratigraphic Traps formed by deposition of a reservoir such as reef or river channel or erosion of reservoir rock. Examples include Barrier Reef, Atoll, Pinnacle or Patch Reef, BioHerm, Angular Conformity, Shoestring-Sandstone Channel, Buttress or onlap sands, Oolite Shoals, Granite Wash, and Updip Pinchout of Sandstone. Combination Trap which are formed by both structural and stratigraphic elements. These include salt-domes (and overlying domes and faults), Salt Dome - Cap Rock, Salt-Dome-Flank Traps, Updip Facies Change, Compaction Anticline, Secondary or Tectonic Dolomite.
  • FIG. 4 Examples of three of these formations are shown in Figure 4—a structural trap 410, and combination trap 420, and a stratigraphic trap 430. Each formation is shown with an exemplary graphical depiction of seismic data obtained for these formations: structural trap seismic data 411, combination trap seismic data 421, and stratigraphic trap 430. There are multiple types of oil and gas formations each having unique seismic data.
  • the seismic data for each of these formations is defined by the underlying rock, fluid, and lithology attributes such as porosity and fluid type. Similar attributes can be obtained post-drilling from well logging and coring data, which entails taking rock samples and performing detailed laboratory analysis.
  • a quantum resonance interferometry engine is configured to harness nonlinear quantum stochastic resonance phenomenology to detect weak signals in seismic data as a disturbance to noise in a high noise and high interference environments.
  • oil and gas exploration objectives for the various structural formations are formulated as weak signal processing problems that are addressed using a resonance interferometry.
  • the structural trap seismic data 411 is analyzed to detect structural traps and to detect an amount of oil saturation, such as oil saturations above 40%.
  • Seismic data near a fault is analyzed to detect hydrocarbon bearing lithology with an oil-water contact boundary. This hydrocarbon lithology is treated as a signal of interest in the seismic data that is analyzed by a resonance interferometry engine with all other lithologies treated as noise.
  • the combination trap seismic data 431 is analyzed to detect porous and permeable tertiary sands that updip around a salt dome. Porosity above a certain percentage (e.g., above 10%) in a porosity channel is treated as the signal of interest in the combination trap seismic 411 by a resonance interferometry engine with all other facies treated as noise.
  • the stratigraphic trap seismic data 431 is analyzed to locate pinnacle reefs with carbonate. Porosity anomalies within stacked sandstone or shale beds are treated in a signal of interest in the stratigraphic trap seismic data that is analyzed by a resonance interferometry engine with all other lithologies treated as noise.
  • FIG. 5 shows an example of a flow chart 500 for processing seismic data for a prospect prior to the seismic data being voxelized for a formation of interest.
  • seismic data is obtained for a formation of interest.
  • the seismic data can be in the form of a pre-stack time or depth amplitude migrated volume.
  • Migration is a geophysical process by which geophysical events (i.e., changes in acoustic reflection energy) are geometrically re-located in either space or time to a location the event occurred in the subsurface rather than a location where the event was recorded at the surface, thereby creating an accurate image of the subsurface.
  • Migration operators are well known in the geophysics art and are applied in both time and space domains to produce time amplitudes and to produce depth amplitude volumes, respectively. A difference between the two is how well the migration operator represents the velocity model.
  • Seismic data in 503 can be obtained by applying time migration algorithms such as Stolt migration, Gazdag, Finite-Difference migrations.
  • Example of depth migration algorithms includes Kirchoff migration, Reverse Time Migration, Gaussian Beam Migration and Wave equation migration.
  • a volume amplitude spectrum is determined for an amplitude volume.
  • the amplitude volume can be analyzed in a low frequency amplitude spectrum (e.g., 3Hz - 10HZ), a high frequency amplitude spectrum (80Hz - 130Hz), and a prime frequency amplitude spectrum (e.g., 12Hz - 70Hz) independently to estimate and predict reservoir rock and fluid properties using spectral decomposition techniques that entail breaking down the seismic signal into its component frequencies.
  • a fully processed seismic survey amplitude volume depending on seismic acquisition hardware settings 512 contains frequencies that are capable of being recorded by
  • seismic energy e.g., acoustic energy
  • a seismic wave front propagates downward into the subsurface and at each lithological facies and geologic boundary (e.g., an unconformity, bed boundaries, etc.), the seismic energy is reflected, refracted, and/or absorbed.
  • the seismic wave front propagates into the underlying sediments, it attenuates, causing the frequency content of the seismic wave front to decrease with depth. For example high frequencies are better preserved near the earth's surface, are "drowned" by the more dominant, lower frequencies.
  • spectral decomposition assesses seismic response at different, discrete frequency intervals.
  • Steps 505 through 560 describe series of spectral decomposition operations to estimate a bandwidth of acquired seismic data at a formation of interest at different signal to noise ratios for the amplitudes, and partitioning the bandwidth or processed seismic data into three regions - a high frequency region, low frequency region and prime amplitude where most of the seismic energy is concentrated.
  • Hydrocarbon/Fluid Saturation - is discriminated on basis of spectral energy attenuation (or absorption) in seismic reflection amplitudes at low frequency. These low frequency amplitudes are treated as noise by conventional seismic processing method and filtered out.
  • Vshale is defined as the volume of shale expressed as a decimal fraction or percentage. Vshale is estimated from small amplifications in the high frequency amplitudes. These high frequency amplitudes are treated as noise by conventional seismic processing method and filtered out. [0077] 3. Porosity attribute is discriminated on the basis of scatter in high frequency noise. These high frequency amplitudes are treated as noise by conventional seismic processing method and filtered out.
  • Brittleness - a geomechanical property determine based on phase variation in low and high frequency noise. Both these low and high frequency amplitudes are treated as noise by conventional seismic processing method and filtered out.
  • Spectral decompositions are used in 530 to partition the seismic volume bandwidth into high and low frequency attributes.
  • the frequency bandwidth taken into consideration during spectral decompositions includes amplitudes as low as -24dB in one implementation approach.
  • the specifics of frequency bands separating the three regimes are dependent upon the parameters of an quantum resonance interferometry engine used in the actual calculation of attributes as described in connection with the quantum resonance interferometry engine 828 in Figures 8 and Figure 10.
  • separation bounds for the three regions are determined at 520.
  • the advantage of a having a 2dB to 4dB separation in seismic amphtudes between the spectral decomposition seismic volumes produced in 530 is that it reduces artifacts introduced by use of trapezoidal shaped wavelets (e.g., Ormsby) used in the processing sequence to prepare seismic data 110.
  • trapezoidal shaped wavelets e.g., Ormsby
  • the separation bounds for the three regions can be determined in 520 as follows:
  • a bandwidth in the prime frequency amplitude regions can include reflection amplitudes above -6dB or above -lOdB cut-off; low frequency amplitudes can be set between [-12dB or -16dB] and [-22dB or -24dB] on frequencies lower than the central frequency of the obtained seismic data; and high frequency amplitudes are set between [ -16dB] and [-22dB or -24dB] on frequencies higher than central frequency of the obtained seismic data.
  • a 2dB or 4dB separation filter between the bands and amphtudes covering additional 10 to 12dB yields up to additional 4Hz to 20Hz of data depending on the acquisition quality of the seismic data?
  • decompositions are not effective and seismic data in 503 needs to be reprocessed using a processing workflow that does not apply frequency filtering. If seismic data in 110 has been pre-processed using noise attenuating spectral whitening as in 514, that broadens or equalizes the spectral of the signal making it similar to white noise spectrum, then the spectral decomposition intervals have to be extended to obtain low and high frequency spectral decomposition in the -20dB to -40dB regime to offset the unintended increase in noise power due to spectral whitening. This is performed as spectral whitening operations when applied to pre-stack seismic traces enhances both the signal and noise ampUtudes.
  • the trace-by-trace variations in the wavelet's frequency content produced by spectral whitening is an undesirable effect is output produced by this method in 170 are to be use to drive amplitude-versus-offset (AVO), amplitude-versus-angle (AVA) processing or time lapse seismic studies on a reservoir. While there is limited data on resolution gain by working in the extended low and high frequency amplitudes as in 516, as shown in the empirical studies conducted using this method the vertical resolution could be increased significantly or even doubled. For example in a typical survey shown to have 10Hz to 50Hz bandwidth for amplitudes above -lOdB, extending the bandwidth to 4Hz to 90Hz effectively doubles the vertical resolution.
  • the frequency regimes can be set to a low frequency regime (“LowF”) covering a lower 5 th percentile of the volume amplitude spectrum, an upper frequency regime (HighF) covering an upper 5th percentile of the volume amplitude spectrum, and prime amplitude regime (“PrimeAmp”) covering a middle 8th to 92th percentile of the volume amplitude spectrum.
  • LowF low frequency regime
  • HighF upper frequency regime
  • PrimeAmp prime amplitude regime
  • cutoffs are determined for different frequency regimes.
  • three SEGY partitioned volumes are created for each seismic horizon.
  • prospect voxels are extracted from one or more of the frequency regime volumes.
  • the porosities of interest are expected to be less than 8% to 10% (e.g., for some limestone, dolomite formations) or the geobodies are discontinuous or stratigraphically stacked; then HighF volume is preferred for porosity determination. If the porosities are expected to be high and vary over a larger range (e.g., >15%) then PrimeAmp is preferred. If the porosity is expected to have very little variability (less than 2% to 3%) over the prospect area then LowF can be used.
  • the partitioned seismic volumes are output for voxelization at 560.
  • Process 500 allows seismic data to be transformed to a representation that allows exploitation of subtle changes in high and low frequency amplitudes which would otherwise be treated as noise in a conventional process and filtered out. Errors in rock and fluid property determination for a formation of interest are reduced by focusing processes in Figures 5 and 6 on optimal input data.
  • Figure 6 enumerates examples of some subsurface volumes (e.g., subsurface volume 220) and their taxonomy for the prospect discussed at 110 in Figure 1. These subsurface volumes can be analyzed for potential formations of interest as discussed at 115 in Figure 1. These specifically drive Unit 503 in Figure 5.
  • the inputs can be broadly classified into six categories, including pre-stack time or depth migrated gathers 610, raw seismic gathers 620, and post stack seismic or depth migrated gathers 630, trace attributes 640, volume attributes 650, and bulk attributes 660.
  • the pre-stack time (PSTM) or depth migrated (PSDM) seismic gathers in 610 are obtained by conventional 2-D, 3-D or multicomponent seismic processing sequences that is well known in the art.
  • processing sequences for the seismic gathers 610 preserve amplitudes, preserve spectrum, preserve phase, remove ground-roll and other out-of-formation coherent noise, do not apply denoising transformations to remove random in-formation noise, use high density tomography grids for velocity estimation, process data at the acquisition sampling resolution, and remove traces that are highly spectrally attenuated.
  • Post-Stack time or depth gathers of 630 can be processed according to Figure 1 but, in some examples, voxelization described in the process 700 introduces artifacts.
  • Raw gathers 620 can be processed according to process 100 in Figure 1.
  • the raw seismic gathers best use is in a confirmatory mode when trying to predict drilling outcome prior to actually drilling. If a prospective well has been positioned at the prospect (using the systems or techniques disclosed herein or using another conventional method and a second opinion is desired to see what outcome can be expected) then the raw seismic gathers 630 or acquired raw seismic traces can be used to estimate rock and fluid properties at the prospective well.
  • Inputs to 220 can be derived using one or a combination of post-stack trace attributes 640 include that include wavelet attributes, Instantaneous attributes 642 such as an instantaneous phase volume or an dominant frequency, or an envelope response volume, can be used according to the systems and techniques disclosed herein to answer questions specific to structural stratigraphy and facies modeling. There are over 50 standard geometric and wavelet-based post-stack attributes that can be generated using commercial software packages. Other trace attributes, include attenuation Q attribute volume 642 and spectral decomposition attribute volume 646. Trace attributes 640 can be used according to the systems and techniques herein to improve signal to noise of a specific attribute or focus on a sub-region with prospect of interest.
  • Volume attributes 650 include full volume attributes such as impedance inversion 652 (i.e., simultaneous pre-stack inversion). Volume attributes 650 also include local dip/azimuth 654 actually take a subset of neighboring seismic around each location of interest in the prospect when predicting reservoir properties at a new location. Volume attributes 650 also include spatial semblance 656, which can be used to highlight subtle discontinuities in seismic volume that stem from faults and complex dips. Volume attributes 650 can be used to conduct quality control on the results produced, for example, at 160 in Figure 1 and to address reservoirs with known structure challenges, such as complex dips. Volume attribute 650 can also be used for analyzing clastic or sandstone reservoirs which includes laminates, channel sands, levies and bars.
  • More complex pre-stack derived attributes in 660 such as AVO attribute
  • a specular versus scattered energy imaging attribute 664 or azimuthally- sectored attributes 666 are applicable to analyzing for complex carbonates, fracture reservoirs, and prospects with known anisotropy. They provide higher signal-to-noise compared to PSTM/PSDM volumes of 610 and also concentrate random noise.
  • Subsurface volumes can be produced in an industry format such as SEGY, as inputs for further processing in Figure 1. If any of the attributes in 630, 640, 650 or 660 are used as inputs as in 220, then detailed processing workflow scripts are required to understand how the attributes are constructed, algorithmic parameters used in their production and any smoothing operators that may have been applied.
  • Figure 7 shows an example of a process 700 for voxelizing seismic data for a formation of interest.
  • An offshore or an onshore prospect can have several formations at varying depths hypothesized to be hydrocarbon bearing.
  • a multi-formation geological cross-section can have hydrocarbon formations such as a Grayburg Sandstone (at 5000 feet), a ClearFork at 6200 feet depth, a Spraberry at 8300 feet, a Wolfcamp Carbonate at 9800 feet, an Atoka limestone at 12600 feet.
  • each of these formations of interest is independently analyzed and processed using the process 700 in Figure 7.
  • a particular formation of interest (e.g., a Clear Fork) is selected from, for example, multiple formations of interest in a seismic volume.
  • Bounding horizons (e.g., a top and bottom horizon) are determined for the selected formation of interest. Horizons are edges on a seismic image or form a vertical cross-section.
  • the bounding horizons can be determined using a conventional n automatic horizon picker software program based on one or more techniques, such as conventional reflection amplitude, coherency attribute, impedance inversion, or spectral decomposition.
  • the bounding horizons can also be determined using sonic well log derived wavelets or vertical seismic profiling (VSP) or cross-bore hole seismic to manually pick a top and bottom horizon for the formation of interest.
  • the seismic data obtained at 110 can be obtained, such as from another geophysical or petrophysical software application or another computing device, preprocessed with the top and bottom horizons determined.
  • seismic horizons can be derived using seismic attributes such as Reflection amplitude volume; Amplitude versus Offset (AVO) attribute volume (gradient, intercept, fluid factors etc); Amplitude-versus-angle (AVA) volume; instantaneous phase; etc.
  • AVO Amplitude versus Offset
  • AVA Amplitude-versus-angle
  • All seismic attributes enumerated in 630, 640, 650 and 660 can be derived from a starting pre-stack time migrated (PSTM) gather volume or a pre-stack depth migrated (PSDM) gather volume.
  • PSTM/PSDM gathers could have been be migrated isotropically or anisotropically depending on the structural complexity, or formation anisotropy (fracturing) .
  • voxel size for the formation of interest is set to .
  • a PSTM volume with 1ms sampling and 25m by 25M bin size can be analyzed using voxel dimensions of 12.5m by 12.5m by 10m or 25m by 25m by 1ms, by re-gridding and reprocessing the data to 12.5m by 12.5m .
  • the vertical depth or time dimension is set to the sampling resolution for voxel depth.
  • the actual traces are resampled to 5X times the sampling interval.
  • the volume is re-binned to 2X the common depth point (CDP) resolution.
  • CDP denotes a common reflection point for dipping reflectors and complex velocity fields generated using seismic acquisition survey. For computational and cost reasons, it is possible to proceed with voxels at the CDP binning resolution or even 2*CDP by 2*CDP resolution by sampling resolution.
  • guard zones above and below the horizon in 720 are developed. Guard zones are used to offset upstream processing errors, suboptimal statics solutions, process-to-processor variability, and/or limitations in the seismic velocity model estimation, which may have led to inaccuracies in the input volume in 705. Seismic velocity model are a list of pairs
  • Guard zones in 720 can be selected based on the thickness of the formation of interest. For example, if a formation of interest is less than 20ms (depth measured in time) or 200 feet gross thickness (depth measured in feet) then guard-zones can be selected to be +/- 8ms or +/- 40feet respectively to ensure horizon event tie errors and small polarity shifts are accounted for. If the formation of interest is greater than 20 ms gross thickness or greater than 200 feet then guard-zones are selected at 20% of horizon thickness. For small formations, (less than 60 feet thickness) two guard zones are set to a combined thickness that at least matches (is at least 100%) the formation thickness.
  • guard-zone thickness can be determined to be 10 times the upper bound of seismic velocity model estimation error bounds, reported during
  • PSTM PSDM processing of seismic data for depths shallower than 10,000 feet and 20 times the velocity estimation error for deeper depths to compensate for imaging errors. Also, it is assumed that formation dips and structural complexity is properly accounted for in the seismic horizon development process.
  • the horizons including the guard zones can optionally be flattened to a fixed time/depth value using a flattening operator to produce a flattened slab.
  • analysis in 740 can also proceed with an unflattened formation of interest with upper and lower bounds and guard zones .
  • Horizon Flattening, through time or depth shifts, is a standard utility operation available in most geophysical software packages
  • a seismic trace represents the response of the elastic wavefield to velocity and density contrasts across interfaces of layers of rock or sediments as acoustic energy travels from a source through the subsurface to a receiver or receiver array. All seismic traces (components of a PSDM/PSTM seismic volume) in the flattened slab are interpolated to report amplitudes at 5 times the sampling rate. For seismic data in the slab having low resolution surveys with processing sampling rate of 2ms resampling at 10 times the PSTM/PSDM processing sample rate is implemented.
  • CMP Common midpoint
  • f is the maximum frequency of the seismic signal
  • is the maximum reflector dip.
  • Traces are stacked within mini-bins. Areal resolution can be increased with high offset acquisitions of seismic data to increase areal resolution. Voxels are constructed by combining the amplitude from multiple traces in multiple mini-bins. Grid based on common depth point (CDP) binning is sized, to accommodate at least 4 (2 X 2) user-sized mini-bins per voxel, prior to any stacking.
  • An example of a voxel size can be 42.5'X42.5', 21.25' X 21.25' in areal dimension.
  • examples of a voxels sizes can be 2ms, 4ms, 8ms (i.e., 8', 16', 20' depth).
  • Traces are interpolated to 0.4X sampling time, e.g., 0.4ms for 2ms sampled acquisition.
  • the entire flattened horizon for formation of interest, within the upper and lower horizon guard zones is then extracted into voxels of a targeted size in 750 using software process that copies the seismic amplitudes contained a voxel boundaries to a data structure .
  • Voxels preset a spatial and temporal resolution to which a formation is analyzed.
  • Reservoir attributes are determined by iterative analysis of individual voxels, one voxel at time. For each voxel that is extracted in 750, a spiral unwrapping algorithm is incorporated herein by reference from Gulati US Patent No. 7,466,851.
  • Figure 8A shows an example of a work flow 800 for analyzing voxelized data using quantum resonance interferometry processing.
  • Machinery such as excitation cascade engine 807, RI processor 831, well control voxel data factory, interferometric coupler 833, and quantifier 839, described in connection with Figure 8Arepresents an example of the analysis performed at 125 in Figure 1.
  • the machinery can be implemented by a computing device such as a data processing apparatus.
  • the machinery can be implemented as computer code that when run by a processor performs the functions described in connection this machinery.
  • the machinery can be implemented as computer modules receiving inputs and producing outputs. Such computer modules can be implemented on the data processing apparatus.
  • the functions performed by the machinery can be performed by multiple computing devices, such as multiple computing devices connect through a network.
  • the work flow 800 has multiple inputs, including voxel data 805 derived from a spectral decomposition on a PSTM or PSD amplitude volume, seismic noise data 814, petrophysical well log data, including porosity data 817, lithology data 820, fluid property data 823, and computer generated synthetic noise 842.
  • the computer generated synthetic noise 842 drives the resonance interferometry dynamics which lead to resonances when the incoming voxel data has an attribute of interest such as rock and fluid properties of interest.
  • the voxel data 805 can be a finite-length 1 -dimensional output vector, including spatial frequency amplitudes, produced by process 760 shown in Figure 7.
  • As the input voxel data 805 for the voxels of a formation of interest can be short 1-D spectral sequences with lengths such as ⁇ 200 data points, they are also referred to as spectral fragments.
  • the voxel data includes for each voxel seismic data such as seismic amplitudes from a specific spectral decomposition, that is LowF, HighF or PrimeAmplitude volume obtained in Figure 5 at 560.
  • the voxel data 805 for the voxels in the formation of interest can be analyzed voxel by voxel by the machinery in Figure 8Aaccording to a geometric sequence.
  • Various geometric sequences can be used. For example, the sequence can start from any formation corner, and can proceed by way of row major traversal or column major traversal, or by depth major orientation.
  • a voxel can be selected for analysis directly from a geographical 3-d arbitrarily shaped body.
  • a sequence can be designed to implicitly capture address and/or the precise coordinates and orientation of the voxel within the formation of interest.
  • a copy of the voxel data for the voxels is made where the analyzed voxel data can be output as an attribute volume.
  • the results of the analysis of the voxel data for the voxels is written back to the same location as the copy. This mirrored read -write operation of voxel input and result ensures that the address of the voxel is accurately preserved.
  • the seismic noise data 814 includes a noise model from amplitudes outside the low and high frequency bands obtained as in 560 and being used in the computation of reservoir attributes. These low and high frequency bands are derived through a procedure detailed in Figures 9 through 11 to ensure that the noise model uses information outside the low and high frequency information being analyzed in the voxel data for the voxels.
  • Figures 8 A through 8B show an example of a process for determining seismic noise data 814.
  • the interferometry engine 828 discussed in more detail below, can have a calibrated signal to noise enhancement of up to 100 times. However, the calibration of the interferometer engine 828 depends upon the amplitudes of the voxels being analyzed.
  • Well logs can be selected from analogous, existing wells such as those in the same prospect as the seismic data being analyzed or from other non-local wells that have similar rock types, similar formations, and similar trap construct as the area from which seismic data is being analyzed.
  • well log data includes porosity data 817, lithology data 820, fluid data 823 (e.g., type and fluid saturation), and rock property data 824 for analogous existing wells.
  • Subsurface sections from such well logs are analyzed to develop a reference model of what different ranges of porosity, percentage of Vshale and fluid type can be expected in the formation of interest.
  • Wells logs for an analogous well can include gamma ray logs, neutron porosity, dipole sonic logs and resistivity logs that represent subsurface sections in the well.
  • Well log sections can show a variability in rock and fluid properties for the analogous well.
  • These well log sections can be, in some examples, 10 feet thick, 16 feet thick, or 32 feet thick in a zone of interest for sampling resolutions of 1ms, 2ms or 4ms, respectively.
  • a voxel and its attributes is defined by voxel data.
  • the well log data is processed by a well control voxel factory 826.
  • Well control voxel factory 826 produces well control voxel data 827.
  • Well control voxel data (such as voxwells used by ViaLogy of Pasadena, California in QuantumRD 3.0) and their attributes are also defined by well control voxels.
  • Well control voxel data includes data indicative of an attribute of interest derived from sources other than the seismic data that produced prospect voxel data 805 for the formation of interest.
  • well control voxel data are extracted for an attribute of interest from seismic data using well logs of known wells with known attributes.
  • these well control voxels are extracted from the seismic data from a well bore of an analogous well.
  • the well control voxel data can be transformed into 1-D vectors using a normalized excitation cascade transformation that yields a monotonic correlation between their energy spectral density and actual attribute value as described in Figure 13 A through 13D.
  • the well control voxel data can transformed into 1 -D vectors using the same process as in Figure 7 to produce the output 760.
  • Figures 13A through 13D show an example of a process for processing well log data to produce well control voxel data 827 for attributes of interest such as for porosity, Vshale, fluid saturation, and brittleness.
  • attributes of interest such as for porosity, Vshale, fluid saturation, and brittleness.
  • energy spectral density is derived from sonic log sections from one or more analogous wells to identify locations of known porosities. Then seismic data corresponding to those locations are identified based on identified locations.
  • Well control voxels are extracted from the seismic data corresponding to sonic sections.
  • sonic logs are used to estimate acoustic impedance changes for regions of interest, such as an area of having a desired porosity, in an analogous well.
  • Wavelet kernel such as a Ricker wavelet
  • Wavelet kernel is used to parameterize the sonic well logs over frequency intervals that are similar to the spatial frequency bandwidth of the Fourier Transforms implemented in transforming voxel data to an excitation cascade in 807 used to develop 1-D voxels for input data in 760.
  • the wavelet kernel uses the same spectral decomposition that generated the amplitude volume that was voxelized in 805 (i.e., as shown in step 540 of figure 5).
  • a calibrated well control voxel data represents energy spectral density of the Fourier Transform of wavelet model used to parameterize and model the sonic section in 826.
  • the energy spectral density describes how the energy of a spatial frequency vector, derived from wavelet used in the well control voxel data factory is distributed with frequency.
  • a method for building well control voxel data is described in Figure 12.
  • a Discrete Fourier Transform can be used as a wavelet kernel is of finite length.
  • the energy spectral density is a function of spatial frequency, not a function of time.
  • the prospect voxel data 805 is transformed to a 1-D spectral vector by an excitation cascade engine 807 and then is normalized to form a normahzed excitation cascade of the voxel data 805. Normalization compensates for high variability in the amplitude values depending on the input seismic volume in selected in 220 and spectral decomposition selected in 540.
  • Figure 12A shows an example of how the excitation cascade is determined at 808.
  • Figure 12B shows an example of how the excitation cascade is normalized based on energy spectral density.
  • the normalized excitation cascade 811 is then analyzed by a quantum resonance interferometry engine 828, such as by the quantum resonance interferometry engine (such as "QRI Engine” produced by Vialogy of Pasadena, California in QuantumRD 3.0).
  • the quantum resonance interferometry engine 828 includes a quantum resonance interferometry processor 831 ("RI" processor) which processes the seismic noise data 814 and the well control voxel data 827 to produce spectral data such as a quantum expressor function (“QEF”) 831.
  • RI quantum resonance interferometry processor
  • QEF quantum expressor function
  • a QEF includes complex noise that is below a threshold required to conclude the presence of an attribute of interest.
  • a QEF can include periodically modulated synthetic noise constructed using seismic data noise 814 that is modulated by a synthetic colored noise and processed through a quantum resonance interferometry processor
  • QSR quantum stochastic resonance
  • the quantum resonance interferometry processor 831 combines the seismic noise data 814 and well control voxel data 827 using a master rate equation for a quantum mechanical model to produce, for example, a bistable quantum function and drives the combination through a quantum resonance interferometry engine. Examples of how the RI processor 831 produces a QEF are discussed in connection with Figure 14 and Figure 18. Other examples of QEFs are described in further detail in U.S. Patent No. 6,963,806, 7,571,056 and 6,780,589 which are incorporated herein by reference in its entirety.
  • the interferometry engine also includes an interferometric coupler 833 which couples the normalized excitation cascade 811 and the QEF produced by the RI processor 831, using synthetic noise 842 to drive the coupling.
  • the synthetic noise such as white noise or colored noise, can be generated by random number generators. For example, a sequence of random numbers can be generated and scaled to fall within a uniform scale.
  • the scale is a defined by a cutoff bound such as +/- 10% of an average of the energy spectral densities of well control voxel data derived using well log sections that do not possess the attribute of interest.
  • Well control voxel data deficient in the attribute of interest are defined by negative control voxel data.
  • Well control voxel data derived from well log sections that possess the attribute of interest are defined by positive well control Data.
  • Negative control voxel data used in noise design is obtained using the flow of Figure 13A, 13B, 13C or 13D depending on the output of interest.
  • this cutoff can be set when the interferometric coupler is initialized at 1122. Subsequent initializations can produce'different cutoffs for finding resonances at those different cutoffs. This can be used to maintain monotonicity of the resonances.
  • An example of synthetic noise is a sequence of random numbers generated to indicate energy spectral densities of 10% of that of an average well control voxel data
  • the interferometric coupler 833 uses a quantum mechanical model to couple, repeatedly, the normalized excitation cascade and the QEF to generate quantum stochastic resonance.
  • a nuclear magnetic resonance ("NMR") master rate equation can be used produce QSR.
  • NMR-based QSR implementation performed by the interferometric coupler 833 is presented in Figure 17 and Figure 21.
  • QSR can also be implemented using Spin boson model, optical cavity model, and SQUID model. Examples of quantum mechanical models are described in U.S. Patent Nos.: 7,567,876 and 6,780,589 and 6,920,397 which are incorporated herein by reference.
  • the interferometric coupler 833 represents an active nonlinear coupling between prospect voxel data and an expressor function designed using combination of synthetic noise and seismic derived noise. Purpose of the computation is to track whether the incoming prospect voxel data and expressor function can couple to produce new information termed as resonance events. This interferometric coupling to produce resonances can be implemented as using a nonlinear interaction between prospect voxel data and any spectra datatype defined herein as an expressor function.
  • Nonlinear operators that can be used to implement expressor functions to produce resonance events include stochastic resonance, quantum stochastic resonance, quantum resonance interferometry, monotone potential operators, quasi-linear elliptic differential operators, angle bounded operators, Galerkin operators, nonlinear information filter and extended kalman filters. Any semi-group operator that implements a nonlinear embedding function can be used as well.
  • a stochastic resonance 836 will occur when the QEF function for that attribute of interest is coupled with the normalized excitation cascade.
  • Resonance 836 can indicate the presence of that particular attribute of interest.
  • the interferometric coupler implements a Master Rate equation describing the time dependent evolution of a system capable of exhibiting NMR.
  • resonance . occurs due to quantum tunneling events, as simulated within 833, the output is recorded and reported as a resonance event in 836.
  • a transformed and enhanced data vector is output in 833.
  • the enhanced data vector is the result of modification of normalized excitation cascade through
  • the quantifier 839 implements an rms (root mean square) amplitude summation to generate a resonance amphtude for the voxel being analyzed.
  • the quantifier 839 can determine that an attribute of interest is found in the voxel when the resonance amplitude value is above a pre-established threshold.
  • the resonance data can be quantified to produce an indication of the amount of the attribute of interest that is found in the voxel.
  • a resonance in 839 may indicate presence of oil, gas, or water in a voxel derived from a specific region in the formation of interest.
  • Figure 8B shows an example of a work flow described in figure 8A to compute a porosity attribute volume and highlight sections of the formation of interest that are above a prespecified porosity threshold (e.g., 20%) as in 8160.
  • Input data 8110 including pre-stack migrated seismic reflectivity amplitude volume 8112 in time or depth , or an acoustic impedance seismic inversion in 8114 are processed according the process 100 in Figure 1 to produce voxels at 120.
  • the reflectivity volume 8112 and/or the acoustic impedance seismic inversion 8114 are extracted using the methods 500 and 700 of Figure 5 and Figure 7, and transformed into a normalized excitation cascade as discussed in connection with Figure 8A at excitation cascade engine 808.
  • well log sections 8135 are derived from the wells from the same prospect from which the input data 8110 was obtained to develop porosity well control voxel data using the method 1300 described in Figure 13-A.
  • Design noise 8130 is the result of combining the synthetic noise and well control voxel data for log sections of interest.
  • Design noise 8130 represents the output of the RI processor 831 in combining seismic noise and well control voxel data to develop a model that is used by the interferometric coupler 833to analyze voxel data derived from the input data 8110 using.
  • Map 8160 represents a vertical section of a porosity-cube volume derived from the results of individual voxel result assembly as contemplated by 160, and furthermore achieved by combining the results of 839 using the porosity workflow process described in Figure 19- A.
  • QRI processor 831 in Figure 8-A combines amplitude and phase information derived from seismic noise data 814 derived from spectral decomposition of PSTM/PSDM gathers or other attributes identified in Figure 6 with well control voxel data 827 to produce a quantum expressor function ("QEF") to the interferometric coupler 833.
  • the interferometric coupler 833 couples incoming voxel data 805 (that has been transformed as in 811 to the same 1-D representation and vector length as the 1-D QEF vector output by the QRI processor 831) to generate resonance indicative of the presence of reservoir attribute of interest such as porosity.
  • the seismic noise data of 814 sets a energy spectral density threshold that must be exceeded by the output of the interferometric coupler 833 to produce a resonance event 836.
  • the threshold is specified to indicate absence or presence of an attribute value from seismic above a threshold for that attribute of interest.
  • the outputs of the excitation cascade engine 808 and the QRI processor 831 combine in the interferometric coupler 833 to produce enhanced data vectors and with resonance events to together provide basis for detection and quantitation of rock or fluid properties for a formation of interest.
  • Figure 9A shows an example of a process 9100 for determining seismic noise data 814.
  • a partitioned seismic volume is obtained.
  • the results of process in Figure 5 implements spectral decomposition computation to produce partitioned seismic volumes 560 in low frequency, high frequency, and prime amplitude regimes, which are inputs to the noise design process 9100. These are also referred to as spectral decomposition volumes.
  • the process 9100 is independently applied to the low frequency, high frequency, and prime amplitude regimes to generate low frequency, high frequency, and prime amplitude seismic noise, one of which is used in the process flow 800 specified in Figure 8 A depending on the nature of the amplitude volume input 9105, also referred to as a spectral decomposition volume of interest.
  • low frequency noise is used in the determination of fluid saturation or type of fluid - oil, water or gas.
  • High frequency noise is used in the computation of Vshale to generate lithology attribute.
  • a target resolution for the prospect under consideration is obtained at 9108. This is set by the chent requirements, geology, and/or knowledge of the prospect, constrained by seismic data acquisition hardware settings and seismic survey design.
  • the resolution requirements obtained at 9108 can be the same as set in 510, 512, 514 and 516.
  • the target resolution can include both a target areal resolution and a target vertical resolution for an attribute of interest such as porosity.
  • well control voxel data are obtained using the regions from known, analogous wells on the prospect that do not meet cut-off threshold criteria for an attribute of interest. For example, in a porosity workflow implemented to estimate porosity for a formation ranging, with expected values of porosity ranging between 5% and 20%, sections of well log with porosity below 5% are used to obtained well control voxel data using the process described in Figure 13A. As previously defined, well control voxel data obtained using well control voxel data from wells that do not contain the attribute of interest are referred to as negative control voxel data.
  • band limited volume regimes are selected.
  • the partitioned volumes are sub-banded into small frequency bands to estimate the sensitivity of amplitude and phase sensitivity changes in the partitioned seismic volume.
  • noise can be appropriately designed to be consistent with the negative control voxel data amplitudes driving the computation for an attribute of interest.
  • an amplitude corridor is determined to estimate the variation in voxel data amplitudes for a partitioned seismic volumes due to inherent seismic noise. This provides parameters for an amplitude to be used in noise design which is used by the quantum resonance interferometry processor 831 and 1022.
  • a phase corridor is determined. This provides parameters for phase of noise design which will be subsequently used by the quantum resonance interferometry processor 831 and 1022.
  • Steps 9110, 9120, and 9160 are described in more detail in Figures 9B, 9C and 9D.
  • Steps 9110, 9120 and 9160 determine variation in amplitude of all prospect voxel data in 805 under certain conditions of interest, and determine variations in phase polarity of amplitudes in seismic partitioned volumes under the same conditions of interest.
  • Variability in both the amplitude and phase of the partitioned seismic volumes in specific regions of well on the prospect is used to develop a seismic derived noise 814 used in Figure 8Aand Figure 10.
  • log sections with porosities below 4% are used to develop noise amplitude and phase parameters if the objective of porosity attribute workflow in Figure 17A is to find regions of porosity over 4%.
  • Steps 9110, 9120 and 9160 use a sub-banding operating to analyze frequency sub- bands of seismic partitioned volumes, implemented using variations of the Gabor-Morlet Transform Method which is available in conventional seismic processing software tools, to produce narrow-band analytic traces.
  • the sub-banding operation is performed in 9120, 9120 and 9160 for different reasons.
  • Sub-banding in 9110 is performed for the purposes of analyzing frequency sensitivity of changes in seismic amplitude changes and seismic phase changes in partitioned seismic volumes.
  • Sub-banding in 9120 and 9160 is performed to developed corridors or maximum variation that can be tolerated due to random, in-formation seismic noise.
  • the amplitude and phase of each narrow band filtered output obtained applying a Gabor-Morlet Transform Method represents an average amplitude and phase of the narrow-band part of the trace.
  • Sub-band window or a 2Hz to 4 Hz separation between sub-bands derived from partitioned seismic volume is an example of a user setting.
  • the number of sub-bands (k) is derived by dividing the frequency bandwidth of the starting partitioned volume by sub-band window. Then k-Gabor-Morlet kernels are applied to the entire partitioned seismic volume to produce k-sub-band traces.
  • the sub-band trace volumes are used at 9120 and 9160 to output zeta-amplitude and phase corridors as detailed in Figure 9-B and 9-C which represent the details of 9160 and 9140.
  • an amplitude corridor is selected using sub-band volumes generated at 9110 to perform sensitivity analysis and assess variation of noise intensities.
  • An amplitude corridor is selected by recombining sub-bands volume amplitudes produced applying Gabor- Morlet Transform Method discussed above, if the energy spectral density is unchanged in the well log sections of a well used for developing well control voxel data.
  • the output at 9250 establishes the sub-bands that will be used in the noise amplitude and noise phase parameter computation at 9170.
  • a standard complex discrete wavelet transform is used at 9210.
  • a CDWT separates the angular information in the seismic data and the polar scale.
  • the directionality of CDWT transform is obtained by the projection of the wavelet coefficients onto positive and negative frequencies separately.
  • the transformed data results in a complex wavelet transformation with a quasi-quadrature relationship between its real and imaginary parts and with strong directional selectivity, important for ensuring sufficient energy and directional selectivity.
  • the CDWT is applied at 9210 to generate multiple (some integer k number of) sub-bands. Once the sub-bands have been generated, as an example those separated by 2Hz or 5 Hz, are used to assess noise sensitivity.
  • a normalized amplitude attribute is computed and used as an input in the computation of an instantaneous phase attribute which outputs a cosine of the instantaneous phase angle of seismic traces within the seismic partitioned volume.
  • Operation in 9220 amplifies the phase component of the compression seismic wave-propagation responsible for reflection amplitudes acquired in seismic survey. As 9220 provides an estimate of phase velocity, it has no amplitude information and is well suited for noise design.
  • a region of interest is selected around the wells on the prospect, targeting the formation of interest, and within +/- 5 CDP (common depth point) Bins of the well.
  • Values in the instantaneous phase volume of 9220 are smoothened in 9240.
  • Derivative of output of step 9220 is implemented by computing an instantaneous acceleration, and checked to ensure that resulting values in the derivative, range over a small interval, is a fraction (e.g., within +/- 50%) of the amplitudes in the spectral decomposition voxelized to estimate rock and fluid attribute computation of 110 ' in Figure 1.
  • Figure 9C describes a process for estimating noise amplitudes to design seismic noise in 9100.
  • a trace sub-band frequency corridor is selected by recombining sub- bands using the same sub-banding transform as used in 9110.
  • recombined sub-band volumes are Hilbert transformed and the real and imaginary parts of the output separated at 9330 and 9340. This is accomplished in time domain by applying a complex Butterworth fdter.
  • Real part ( zero phase) of the filter generates the real part of the seismic trace, thereby ensuring that both the real and imaginary parts will have the same spectral characteristics.
  • a band limited Butterworth filter is an exemplary filter that satisfies the band limited requirements of the Hilbert transform.
  • the imaginary part of the complex trace has identical amplitude spectrum as the real part. Real and imaginary traces are used to generate the real and imaginary parts used in the analysis described by Figure 9C.
  • a wavelet envelop operator is applied to assess the variability of the real part of the Hilbert transformed to develop 9340.
  • the instantaneous amplitude value at the maxima of the envelop is computed for sub-band volumes generated in 9340.
  • an instantaneous bandwidth is computed for the sub-volume of 9310.
  • the real part of the Hilbert transform output in 9350 is also used to compute the residual of the real part.
  • the instantaneous amplitudes are organized to produce a histogram in 9370 which is used compute residuals for noise in 9380. In some examples, percentages lower than 2% or lower than 5% of the residual noise amplitudes can be used in 9360 to obtain the noise volumes.
  • the optimal trace corridor is selected using sub-band volumes generated in 9410.
  • the envelope modulated phase attribute is computed in 9410.
  • Intensities of the instantaneous phase represent the trace envelope magnitude. It is used to assess the phase variation of the strong events, without the interference of weaker events as in the instantaneous phase volume and vice versa. It amplifies the phase component of the wave-propagation.
  • an instantaneous phase attribute is computed and its normalized amplitude is used to assess the phase sensitivity.
  • a normalized amplitude attribute is computed to generate the cosine of the instantaneous phase angle. Operation in 9420 amplifies the phase component of the wave-propagation. As it provides an estimate of phase velocity, it has no amplitude information and is well suited for noise design.
  • a second derivative of envelope attribute with respect to time is computed.
  • Results of 9430 captures all reflecting interfaces visible within seismic bandwidth.
  • this attribute is traditionally used in seismic analysis to assess sharp changes in lithology and shows sharpness of event, it captures discontinuities in the data effectively.
  • the Time derivative of Envelope attribute is also computed from seismic sub-band volume to assess phase propagation properties. As output of 9450, highlights the sharpness of the rise time relative to absorption of seismic energy, it is more effective in treating phase propagation.
  • the results of 9450 are again sub-banded at 9460 to understand phase variability in the data.
  • the intensities from Envelope 2nd Derivative attribute are correlated to the well control voxel data derived in the vicinity of well log in zones of interest. If the correlation is low, such as below 20% then Envelop 2nd Derivative can be treated to tie the seismic around well control.
  • the noise phase and noise amplitude corridors are available using Figure 9C and Figure 9D the results are combined to provide noise required to drive the quantum resonance interferometry engine and Interferometric Coupler. These volumes of noise amplitude and noise phase have useful interpretive properties in their attribute; can assist in fault anisotropy distribution.
  • Figure 10 describes a process for upscaling or upconverting analysis of attributes of interest from one areal/vertical (time and/or depth) seismic resolution to a higher resolution; and to produce multi-scale reservoir attributes (i.e., output at multiple resolutions in the same or different iterations of workflow described in Figure 1), wherein the same attribute volume of rock or fluid properties are generated at multiple resolutions.
  • outputs of the process 100 in Figure 1 can help in managing production and making decisions on which areas to drill next or on which wells to complete for future production.
  • Output of process 100 is also used to drive a reservoir model software simulation to make operational decisions that are external to this invention and its output generated by flow 100.
  • An external reservoir model may be simultaneously running computations at different resolutions.
  • Process flow in figure 10 can be used to generate a multi-resolution output, using coarse seismic resolution to output a higher resolution to drive a reservoir model software simulation.
  • reservoir models work with gridded cells or use 3-D volume cell representation to manage and store reservoir attributes, which are then used to manage and make engineering or operational decisions for a formation of interest or under production: more accurate, upscaled reservoir attributes including porosity, Vshale and fluid saturation from Figure 10 output results for voxels in a format and areal/vertical resolution that can be directly used by an external reservoir model software simulation.
  • Figure 10 process enables inputs obtained using coarse seismic resolution in the vertical dimension, to generate voxel outputs 839 at close to the starting well log resolution.
  • An example of upscaling performed by Figure 10 would be to use seismic amplitude volumes as input, with 55 feet by 55 feet by 2 ms bin in time (or 55 feet by 55 feet by 30 feet in depth) resolution and wells logs at 2 feet sampling, to provide outputs for voxels up-scaled to 27.5 feet by 27.5 feet to 0.5ms in time (or 27.5 feet by 27.5 feet by 10 feet in depth).
  • Figure 10 enables upscaling.
  • Another RI Processor 1022 is used to transform the normalized excitation cascade output 81 lto a different resolution with different properties.
  • the other RI processor 1022 uses different initialization parameters and inputs from those used by RI Processor 831 in the quantum resonance interferometry engine828.As a result, the RI processor 1022 produces voxel outputs at a higher resolution than the voxel data 805. As a result, outputs by produced by the work flow 1000 shown in Figure 10 will be different from work flows without the RI processor 1022. .
  • the voxel data 805 is transformed to a 1-D spectral vector and normalized by an excitation cascade engine 808,
  • the RI processor 1022 combines the normalized 1 -D spectral vector with synthetic broadband colored/random noise 1024 and drives the combination inputs through a RI engine that uses a quantum mechanical model uses QSR dynamics data (such as "Virtual Vibe” as used in ViaLogy of Pasadena,
  • QuantumRD 3.0 California's QuantumRD 3.0).
  • an NMR model can be used combine the synthetic broadband colored/random noise and the 1-D spectral vector.
  • a different quantum mechanical model can be used in the RI processor 1022 than was used in the RI processor 831.
  • QuantumRD 3.0 produced by Vialogy of Pasadena, California uses a NMR-based implementation to implement RI processor 831.
  • the uniform scahng interval for the synthetic broadband colored/random noise is set to match 10% of the average energy spectral density amplitudes of the negative control voxel data 805.
  • RI processor 1022 is introduced to increase robustness of resonance event 836 and prevent attributes of interest from being influenced by high amplitudes in noisy or low quality seismic data, or by frequency periodicity in the data in seismic traces from which voxel data is obtained.
  • periodicity in excitation cascades produced by excitation cascade engine 808 can potentially trigger resonance events 836. This can be a challenge for the identification of fluid markers described in Figure 13 C used to determine fluid properties.
  • running a normalized excitation cascade through RI processor 1022 can be used to detect weak signals buried near the boundaries of input volumes i.e., around -22dB or around -24dB.
  • RI processor 1022 also serves as a preconditioning mechanisms that preconditions 1-d normalized excitation cascades that are introduced into the interferometric coupler 833 by providing a shaping to certain frequencies which are directly related to rock or fluid properties of interest. For example, low frequencies are relevant to fluid saturation estimation. So, RI processor 1022 can be configured to enhance amplitudes in, for example, a 3 HZ or 4HZ window. RI processor 1022 provides an independent degree of control to bias the workflow of Figure 1 to regions of the frequency spectrum that are more important for specific rock and fluid properties.
  • FIG. 11 shows a process 1100 for establishing monotonicity of resonance event 836.
  • process 1100 is applied to finalize the resonance event 834.
  • voxel data such as 1-D spectral vector produced by the excitation cascade engine, is obtained for processing by the process 1100.
  • the coupler is initialized. During initialization, an initial cutoff bound (i.e., threshold) for the intensity of synthetic colored noise 842 is set.
  • the cutoff bound is set, for example, to a fractional intensity of the average intensity of the prospect voxel data in the input seismic data in process 100 at 118.
  • the synthetic noise 824 is constructed using random number generator algorithms.
  • An example of a fraction used as a synthetic noise cutoff bound is a scaling intensity of 10% of the intensity of partitioned volume.
  • the interferometric coupler is run to produce resonance. If at 1125 a resonance is not achieved, the voxel data for the voxel is failed at 1160, which means that a robust resonance event was not established for the voxel, and therefore the voxel does not exhibit the attribute of interest.
  • the failed voxel is assembled at 1150 into a voxel assembly 1150 with an indication of an absence of the attribute of interest for that voxel.
  • a resonance is achieved at 1125, it is determined whether the monotonicity (e.g., persistence) of a resonance event has been established at 1130. For example, it is determined whether resonance has been found at a sufficient number of cutoff bounds for the voxel. If not, the cutoff bound is adjusted (e.g., increased) to a different cutoff bound. For example, the cutoff can be doubled, such as by increasing from 10% to 20% of the intensity of the partitioned volume. The coupler is reset at 1122 with the adjusted cutoff bound and the process is repeated.
  • the monotonicity e.g., persistence
  • the voxel is assembled in the voxel assembly with an indication that a robust resonance event was determined for the voxel, and thus the attribute of interest was found for the voxel.
  • FIG. 12A through 12B describe a process for generating and normalizing excitation cascade data, with output represented as a 1-D data vector that is used in the computation of porosity, Vshale, Fluid Saturation and Brittleness attributes using the process detailed for each voxel in Figure 8Aand Figure 10.
  • figures 12A and 12B provide an example of a process that can be performed by the excitation cascade engine 808.
  • the process of generating an excitation cascade output is dependent on the prospect voxel attribute being analyzed as it uses different positive well control voxel data and negative well control voxel data parameters 1235 depending on the attribute under consideration.
  • the process of Figure 12A is as follows.
  • the incoming voxel data in 1210 obtained using the process 700 in Figure 7, is transformed to a 1 -D data vector using an unwrapping operation in 1215.
  • Extracted data from a three dimensional (3-D) data structure used to managing seismic dataset in the industry standard SEGY notation, is produced as a 1- D data vector.
  • Variations of unwrapping can be applied to transform data For example, some variations of types of unwrapping are discussed in U.S. Patent No.: 7,466,851 which are incorporated herein by reference.
  • unwrapping that is used in seismic data processing can entail traversal of data elements in a 3-D voxel to transform into 1-D data using row-major (row-by row -x-dimension travel and then each successive layer in time or depth), or column-major (column-by column - y-dimension travel and then each successive layer in time or depth),or outward spiral (in clockwise or anticlockwise sequence starting the center or corner of the prospect voxel data and then spiraling outwards and downwards in increasing time or depth or spiraling upwards in decreasing depth or time). Same method is also used for generating an excitation cascade from positive control voxel data and negative control voxel data as in Figure 13A through 13D.
  • mini-bins For notation purposes the seismic trace bins enclosed within a 3-D voxel boundaries are referred to as "mini-bins" to distinguish them from the CDP bins that correspond to seismic trace locations in a PSTM PSDM seismic volume or seismic attribute volumes as in Figure 6. If voxels use multiple CDP bins then the mini-bins may correspond to CDP bins.
  • unwrapping to transform 3D prospect voxel data into a 1-D vector is clockwise outward spiral traversal starting at top left-corner (e.g., geographical northwest) of a 3-D voxel, each time or depth layers at time, going downwards in increasing time/depth.
  • top left-corner e.g., geographical northwest
  • 3D prospect voxel data into a 1-D vector is clockwise outward spiral traversal starting at top left-corner (e.g., geographical northwest) of a 3-D voxel, each time or depth layers at time, going downwards in increasing time/depth.
  • an outwards spiral approach could be used just as well for voxels with odd mini-bin dimensions (e.g. 3 by 3 or 5 by 5) starting at the location in the center of the voxel layer.
  • the output of 3-D voxel unwrapping in 1215 results in a 1-D data vector with seismic amplitude values at each mini-bin location.
  • starting seismic amplitude slabs can be voxelized to have the center of voxel correspond with the top of seismic horizon.
  • voxels sizes of 2 by 2 CDP bins and 5 sample points in depth or time are preferred designing a voxel, with the time and depth dimension are resampled by a factor or 5 or 10 depending on target spatial resolution.
  • a 1 -D transformed output from 1215 has 2 by 2 by 5 by 5 or 100 data points in 1 -D starting data vector for seismic data sampled at 1ms or 2ms resolution; or 2 by 2 by 5 by 10 or 200 data points in starting 1-D vector for seismic data sampled at 4ms or higher resolution.
  • the 1-D vector initially including seismic amplitudes from a band-limiting spectral decomposition process applied to PSTM or PSDM gather volumes, is then Fourier transformed to a 1-D vector of spatial frequencies at 1225.
  • the spatial frequencies are amplified based on relationship and periodicity of neighboring data elements in the 3D voxel data in all the x, y and z dimensions.
  • Well control voxel data parameters for an attribute of interest as generating using the methods described in Figure 13A through 13D, and maintained at 1220 are used to partition the 1-D data vector into k-RMS-bands (root mean square) where k is at least 1 and less than half the length of 1-D data vector derived from prospect voxel data.
  • the value of k is determined using the process of Figure 13A thru 13D for different attributes of interest. Also the k-rms-band represents the results of computing the square root of means square amplitude of spectral components in 1 -D vector at 1215 for ⁇ each of the k-bands of interest. Energy spectral density of all k-rms-bands is computed as in 1230.
  • the steps 1240 through 1255 describe one or more modulation transformations that are required to be performed on the spectrally transformed voxel data at 1225 to generate an excitation cascade where the energy spectral density value satisfy conditions determined in the well control voxel data generation process.
  • the representation and types of modulation kernels are determined by the rock physics input for the prospect as in Figure 2.
  • the modulation kernels are different for different source rock types (source of hydrocarbon) such as sandstone, dolomite, limestone, siltstone.
  • modulation kernels in 1235 are deployed to offset uncertainty due to the fact the limited (even with large numbers of wells on the prospect) well-control cannot always capture the full variability and complexity of underlying source rock grain matrix (e.g., dolomite) and seismic reflections only capture a bulk property of the formation.
  • a library of wavelet kernels e.g., using Ricker or Ormsby wavelets
  • model high/low/prime amplitude frequencies
  • These facies represent the changing acoustic impedance boundaries imaged by sonic logs in the well-bore.
  • the method uses well- control/analogs to assess field specific variability in recognition kernels that can separate the seismic facies in different frequency regimes as wavelet kernels. Also, the energy spectral density (or power spectral density) of each recognition kernel is computed. Using a Rock Physics handbook - tables that provide actual experimental data for measurements of changes in p-wave, s-wave velocities for different rock compositions and grain sizes, "uncertainty bounds" can be established. At 1240, based on analysis of well control quality (seismic - synthetic from log tie) energy spectral density or a derived power density metric is computed as required to recover "rock recognition kernels" from seismic data in any spectral decomposition of interest.
  • the 1-D spectral data is convolved with a rock recognition kernel (represented as the parameters of a standard synthetic wavelet kernel) in the libraryl235.
  • a rock recognition kernel represented as the parameters of a standard synthetic wavelet kernel
  • the resulting energy spectral density is computed at 1255.
  • the results produced at 1255 are compared with the energy spectral density criteria set during well control voxel data design process for the attribute of interest using 1250. If a sufficient threshold of energy spectral density is not reached then another kernel is selected from 1235 and the process is repeated.
  • the voxel spectral vector produced at 1225 is resampled-and interpolated to double the length, e.g. go from a length of 100 data points set in 1215 to 200 data points.
  • the number of iterations required in 1250 is also set during the well control voxel data design process of Figure 13 A thru 13D.
  • the results of process shown in Figure 12A are output at 1260 and denoted as an excitation cascade 1260.
  • Figure 12B shows an example of a normalization process applied to prepare and condition an excitation cascade for introduction to Interferometric Coupler 833 in Figure 8Aand Figure 10.
  • All data is uniformly rescaled such as to [-1.0,1.0] or to a another range between [-F,+F] where F is a real-number in 12170.
  • Rescaled excitation vector obtained at 12170 is then frequency modulated with a periodic signal at 12180 using modulation parameters from a repository 12185.
  • Figure 13A shows an example of how a porosity-specific well control voxel data is developed.
  • the well control voxel data can be developed in the processes of converting porosity log data 817 of Figure 8A and Figure 10 into a well control voxel data at 827.
  • the parameters of the well control voxel data as output in 13170 and established by processes described in Figure 13 A, are used in the implementation of Figure 8, Figure 10 and also in the computation of excitation cascade in Figure 12A. Different ranges of porosity are of interest in different formations in different prospects.
  • the porosity criteria of interest for the prospect under consideration in Figure 13 A is established by a porosity workflow or a porosity-cube workflow described in Figure 17 A, specifically in 17101 and 17105. Once the porosity range of interest 17105 are specified, then the process of Figure 13 A provides for developing parameters for transformation prospect voxel data of Figure 12A as relevant to the computation of porosity attribute. Based on prospect geology model, a vertical resolution is established as in 13102 for analyzing both well logs and seismic data. Depth sections from the neutron porosity well logs in the wells of interest on the prospect or analog are analyzed at 13104 to demarcate well log subsections that will be further analyzed.
  • Either normalized or un- un-normalized and normalized well logs could be used as a starting point in 13104.
  • Logs can be normalized for effective porosity for the formation of interest. For example, if the target formation entails determining the porosity of Goen limestone then reference logs needs to be preprocessed to a specific effective limestone porosity.
  • Formation cutoffs on the extracted porosity logs sections are established to ensure that they are properly normalized using standard procedures and quahty control implemented by well logging companies.
  • An example of a formation cutoff would be a range such as [2%, 18%] for Permian strawn in Central Texas, or [8%, 36%] for Frio sand in Oklahoma or [11 %, 48%] for Pliocene sandstone in an offshore field off West Africa.
  • the process in Figure 13A can work on any interval of interest at 13110.
  • sonic well log sections are extracted for regions in the well-bore that are output in 13104.
  • Sonic log section limits are estabhshed at 13120 and sections of the logs copied and porosity values within the sections of interest extracted as (starting depth, porosity) pairs of tuples. As the same depth range is used for all processes in 13 A through 13D, ending depth is not specified.
  • Example depth sections are 8 feet, 16 feet, 10 feet, 16 feet, 32 feet or their metric equivalents for prospects that follow metric convention.
  • Sonic well log sections are extracted at 13120 in a manner to fully encapsulate the well log section impedance (or acoustic velocity change). For example, if 16 feet sections are used for well log analysis to design porosity examples then a 32 feet sonic section will be extracted such that the 16 feet well log section lies between the sonic section.
  • the extraction 13120 can be also used if the sonic and well log sections are the same size and top or bottom aligned. If dipole sonic is available then the extracting at 13120 is repeated for both p-wave (compression) velocity and s-wave (shear) velocity. In some examples, compensated sonic is preferred.
  • a sonic can be approximated from other well logs such as density log using relationships like Gardner's formula. Any method that can derive a sonic log, if one has not been acquired for the reference well, can be used to implement the extracting at 13120. If multiple wells are available or if the well has encountered a large thickness of formation of interest then multiple reference log sections and corresponding sonic sections are obtained. These well and sonic log sections are sorted and arranged in increasing order to form groups. Using a histogram construction the multiple log sections are organized in groups or porosity-bins to cover the range of porosity established in 13110. An example is groups [0-4%, >4%-8%, >8%-12%, ... , >24%-28%].
  • the energy spectral density of each sonic section is computed at 13140 and rearranged to form a line plot of energy spectral density versus increasing porosity-bins.
  • Power spectral density or any power measure that can reduce the wavelet parametric derived at 13130 to a scalar number can be used to implement the check in 13150.
  • a check is performed to see if the energy spectral density or an alternate scalar representation of wavelet parameter that encodes the information in sonic section is linear with a positive slope and the different sections corresponding to porosity well log are well separated, and accommodate the variance or standard deviation in different well log sections belonging to different wells. If the results produced at 13150 show linearity then the parameters are output at 13170 to be used to develop well control voxel data and also to be used in the excitation cascade engine 808. If the sections are not separated out then the results at 13130 are convolved with another kernel from the rock physics library 1235 in Figure 12A. The same library is maintained and represented at 13155. The convolution between a rock physics modulation kernel from the library of modulation functionsl3155 is implemented at 13160.
  • Steps 13140, 13150 are repeated until linearity is established.
  • Selection of each new kernel from the library of modulation functions 13155 is also used in the Tl-TN steps shown at 811 to implement an excitation cascade. Satisfactory generation of output at 13170 and completion of linearity test implies that consistent well control is used to tie the well data to the seismic partitioned volume and to calibrate seismic attributes.
  • Use of multiple kernels, with each kernel representing a higher spatial frequency dimensionality, to achieve linearity in energy spectral density or to achieve a related scalar measure is sufficient, but not necessary, to optimize the tie at wells.
  • FIG. 13B shows an example of how a Vshale-specific well control voxel data is developed to analyze lithology when converting lithology (e.g., gamma ray) log data 820 of Figure 8Aand Figure 10 into a well control voxel data in Unit 827.
  • Vshale is a quantitative representation for lithology and particularly relevant for clastic of sedimentary types of structural, stratigraphic, and combination reservoirs.
  • Parameters of the well control voxel data as output at 13270 and established by processes described in figure 13B are used in the implementation of Figure 8, Figure 10, and also in the computation of excitation cascade in Figure 12 A. Different ranges of Vshale are of interest in different formation in different prospects.
  • the Vshale criteria of interest for the prospect under consideration by the method of Figure 13B is established by the Lithology Workflow described in Figure 17B, specifically at 17201 and 17205.
  • the method of Figure 13B provides a method for developing parameters for transformation of the process in Figure 12A, as relevant, to the computation of Vshale attribute.
  • a vertical resolution is established as in 13202 for analyzing both well logs and seismic data.
  • Depth sections from the gamma ray well logs in the wells of interest on the prospect or analog are analyzed at 13204 to demarcate well log subsections (i.e., reference regions of interest) that will be further analyzed.
  • a vertical resolution is established as in 13102 for analyzing both well logs and seismic data Logs can be normalized for effective Vshale for the formation of interest. For example, if the target formation entails determining the Vshale of Wilcox sandstone then reference logs needs to be preprocessed to effective Vshale ratios.
  • formation cutoffs on the extracted gamma logs sections are established to ensure that they are properly normalized using standard procedures and quality control implemented by well logging companies.
  • An example of a formation cutoff would be a range such as [10%, 90%] for Wilcox sandstone in South-Central Texas, or [30%, 70%] for Vickersburg sand in Oklahoma or [5%, 95%] for Pliocene sandstone in an offshore field off Indonesia.
  • the process in Figure 13B is can work on any interval of interest at 13210.
  • sonic well log sections are extracted for cutoff regions.
  • Well log section limits are established and sections of the logs copied and Vshale values within the sections of interest extracted as (starting depth, Vshale) pairs of tuples.
  • the computation can also be implemented in gamma ray intensity values as from logs without the need for mapping them to Vshale. In such case the resulting volume would be a gamma ray predictor as opposed to Vshale predictor.
  • ending depth is not specified.
  • Example depth sections are 8 feet, 16 feet, 10 feet, 16 feet, 32 feet or their metric equivalents for prospects that follow metric convention.
  • Sonic well log sections are extracted at 13220 in a manner to fully encapsulate the well log section impedance (or acoustic velocity change). For example, if 16 feet sections are used for well log analysis to design Vshale examples then a 32 feet sonic section will be extracted such that the 16 feet well log section lies between the sonic section.
  • the extraction 13220 can be also used if the sonic and well log sections are the same size and top or bottom aligned. If dipole sonic is available then the extracting at 13220 is repeated for both p-wave (compression) velocity and s-wave (shear) velocity. In some examples, compensated sonic is preferred.
  • a sonic can be approximated from other well logs such as density log using relationships like Gardner's formula. Any method that can derive a sonic log, if one has not been acquired for the reference well, can be used to implement the extracting at 13220. If multiple wells are available of if the well has encountered a large thickness of formation of interest then multiple reference log sections and corresponding sonic sections are obtained. These well and sonic log sections are sorted and arranged in increasing order to form groups. Using a histogram construction the multiple log sections are organized in groups or Vshale -bins to cover the range of Vshale established in 13210. An example is groups [0-10%, >10%-20%, >20%-30%, >80%-90%, >90%, 100%].
  • ⁇ wavelets are extracted from the corresponding sonic log sections that correspond to log sections selected in 13204 on basis of Vshale attribute. Sonic sections are associated with well log sections using multi-attribute software data structure such as list or n-dimensional array.
  • the initial wavelet used at 13230 is selected to match the same frequency bandwidth as the one selected for spectral decompositions.
  • Vshale computation can be implemented on prime-amplitude or high frequency or full spectrum seismic data in 110.
  • the energy spectral density of each sonic section is computed at 13240 and the rearranged to form a line plot of energy spectral density versus increasing Vshale -bins.
  • Power spectral density or any power measure that can reduce the wavelet parametric derived at 13130 to a scalar number can be used to implement the check in 13250.
  • a check is performed to see if the energy spectral density or an alternate scalar representation of wavelet parameter that encodes the information in sonic section is linear with a positive slope and the different sections corresponding to Vshale well log are well separated, and accommodate the variance or standard deviation in different well log sections belonging to different wells. If the results produced at 13250 show linearity then the parameters are output at 13270 to be used to develop well control voxel data and also to be used in the excitation cascade engine 808. If the sections are not separated out then the results at 13230 are convolved with another kernel from the rock physics library 1235 in Figure 12A. The same library is maintained and represented at 13255. The convolution between a rock physics modulation kernel from 13255 is implemented at 13260.
  • the resulting data is again parameterized using a wavelet transform such as Ricker or Ormsby.
  • Steps 13240, 13250 are repeated until linearity is established.
  • Selection of each new kernel from the library 13255 is also used in the Tl-TN steps shown at 811 to implement an excitation cascade. Satisfactory generation of output atl3270 and completion of linearity test implies that consistent well control is used to tie the well data to the seismic and to calibrate seismic attributes.
  • Use of multiple kernels, with each kernel representing a higher spatial frequency dimensionality, to achieve linearity in energy spectral density or to achieve a related scalar measure is sufficient, but not necessary, to optimize the tie at wells.
  • linearity cannot be established in a predetermined number of convolutions (e.g., 5 or 6)
  • some log sections can be dropped from the process to offset the likelihood that logging could be in error, or some wells represent a petrophysical, structural or stratigraphic nonconformity or suboptimal well normalization.
  • Sufficiency of the linearity plot to tie well log to seismic using energy spectral density of convolved vectors provides accuracy in Vshale prediction away from well-bore during quantum resonance interferometry engine and Interferometric coupler iterations.
  • Figure 13C shows an example of how a fluid saturation-specific well control voxel data is developed to analyze fluid properties as in Figure 8A.
  • Resistivity log values in sections with high effective porosity is one measure of fluid distribution and applicable analysis of carbonate, and clastic or sedimentary, types of structural, stratigraphic and combination reservoirs.
  • Parameters of the well control voxel data as output at 13370 and established by processes described in figure 13C are used in the implementation of Figure 8, Figure 10, and also in the computation of excitation cascade in Figure 12 A.
  • the fluid saturation sections meeting thresholds for exploration or well development interest may also be derived from results of well core analysis.
  • the process of Figure 13C provides for developing parameters for transformation of the process in Figure 12A, as relevant, to the computation of fluid attribute. For example, at 13302, vertical resolution is established. Depth sections from the resistivity well logs in the wells of interest on the prospect or analog are analyzed at 13304 to demarcate well log subsections that will be further analyzed. This can include the processes of geological logs normalization, petrophysical normalization, and effective fluid presence and fluid saturation computation whereby different wells, that may have been drilled and logged by different companies at different times using different technologies and logging tools, are normalized so that these sections can be effectively compared.
  • Logs can be normalized for effective fluid type and saturation for the formation of interest. For example, if the target formation entails determining the saturation of Wilcox sandstone then reference logs needs to be preprocessed to effective fluid saturation levels using rock core data, if available.
  • Gas-cut, Oil-cut, Gas- Oil-Ratio are terms of the art used to represent saturation levels of hydrocarbon by different organizations and prospects. Any implied, derived or computed measure of fluid type or saturation that can be related back to the well and well log is a candidate for identifying a section of interest to look for in the seismic data or avoid (as in the case of water-saturated or water wet wells).
  • Formation cutoffs on the extracted resistivity logs sections are established to ensure that they are properly normalized using standard procedures and quality control implemented by well logging companies.
  • An example of a formation cutoff would be a range such as [10%, 100%] for Wilcox sandstone in South-Central Texas, or [30%, 70%] for Vickersburg sand in Oklahoma or [5%, 95%] for Pliocene sandstone in an offshore field. Fluid or hydrocarbon saturation can range from 0% to 100%.
  • the process in Figure 13C can work on any interval of interest at 13310.
  • sonic well log sections are extracted for cutoff regions.
  • well log section limits are established and sections of the logs copied and resistivity values within the sections of interest extracted as (starting depth, resistivity) pairs of tuples.
  • the computation can also be implemented in resistivity log intensity values (that can numerically range from 0.1 ohm to lOOOohm as an example) as from logs, without the need for mapping them to saturation. In such case the resulting volume would be a resistivity value predictor as opposed to saturation predictor. Saturation can be obtained from a resistivity attribute prediction volume using an additional calibration table. As the same depth range is used for all processes in 13 A through 13D, ending depth is not specified.
  • Example depth sections are 8 feet, 16 feet, 10 feet, 16 feet, 32 feet or their metric equivalents for prospects that follow metric convention.
  • Sonic well log sections are extracted at 13320 in a manner to fully encapsulate the well log section impedance (or acoustic velocity change). For example, if 16 feet sections are used for well log analysis to design saturation examples then a 32 feet sonic section will be extracted such that the 16 feet well log section lies between the sonic section.
  • the extraction 13320 can be also used if the sonic and well log sections are the same size and top or bottom aligned. If dipole sonic is available then extracting 13320 is repeated for both p- wave (compression) velocity and s-wave (shear) velocity.
  • a sonic log can be derived in various ways, if one has not been acquired for the reference well and can be used to implement the extracting atl3320.
  • multiple reference log sections and corresponding sonic sections are obtained. These well and sonic log sections are sorted and arranged in increasing order to form groups. Using a histogram construction the multiple log sections are organized in groups or fluid saturation-bins to cover the range of saturation established in 13310. An example is groups [0-10%, >10%-20%, >20%-30%, >80%- 90%, >90%, 100%].
  • ⁇ wavelet analysis Using wavelet analysis, underlying wavelets are extracted from the corresponding sonic log sections. Sonic sections are associated with well log sections using multi-attribute software data structure such as list or n-dimensional array.
  • the initial wavelet used at 13330 is selected to match the same frequency bandwidth as the one selected for spectral decompositions.
  • fluid saturation computation can be implemented on prime- amplitude or low frequency or full spectrum seismic data in 110.
  • the energy spectral density of each sonic section is computed at 13340 and the rearranged to form a line plot of energy spectral density versus increasing saturation-bins or resistivity.
  • Power spectral density or any power measure that can reduce the wavelet parametric derived at 13330 to a scalar number can be used to implement the check in 13350.
  • a check is preformed to see if the energy spectral density or an alternate scalar representation of wavelet parameter that encodes the information in sonic section is linear with a positive slope and the different sections corresponding to resistivity well log are well separated, and accommodate the variance or standard deviation in different well log sections belonging to different wells. If the results produced at 13350 show linearity then the parameters is output at 13370 to be used to develop well control voxel data and also to be used in the excitation cascade engine 808. If the sections are not separated out then the results at 13330 are convolved with another kernel from the rock physics library in 1235 in Figure 12 A. The same library is maintained and represented at 13355. The convolution between a rock physics modulation kernel from 13355 is implemented at 13360.
  • the resulting data is again parameterized using a wavelet transform such as Ricker or Ormsby wavelet.
  • Steps 13340, 13350 are repeated until linearity is established.
  • Selection of each new kernel from the library 13355 is also used in the Tl-TN steps shown at 811 to implement an excitation cascade. Satisfactory generation of output at 13370 and completion of linearity test implies that consistent well control is used to tie the well data to the seismic and to calibrate seismic attributes.
  • Use of multiple kernels, with each kernel representing a higher spatial frequency dimensionality to achieve linearity in energy spectral density or to achieve a related scalar measure is sufficient, but not necessary, to optimize the tie at wells.
  • linearity cannot be established in a predetermined number convolutions (e.g., 5 or 6 convolutions)
  • some log sections can be dropped from the process to offset the likelihood that logging could be in error, or some wells represent a petrophysical, structural or stratigraphic nonconformity or suboptimal well normalization.
  • Sufficiency of the linearity plot to tie well log to seismic using energy spectral density of convolved vectors provides accuracy in fluid type or fluid saturation prediction away from well-bore during quantum resonance interferometry engine and Interferometric coupler iterations.
  • Figure 13D shows and example of how a geomechanical attribute such as brittleness-specific well control voxel data is developed to analyze subsurface rock brittleness properties.
  • Brittleness attribute has been shown to be important for developing nonconventional resources such as shale gas prospects which have their own classifications in terms of structural, stratigraphic and combination reservoirs.
  • the parameters of the well control voxel data as output at 13470 and established by processes described in figure 13D, are used in the implementation of Figure 8, Figure 10, and also in the computation of excitation cascade in Figure 12 A.
  • Brittleness values meeting thresholds for exploration or well development interest are derived from results of well core sample analysis or cutting section analysis.
  • the process of Figure 13D provides for developing parameters for transformation of the process of Figure 12A, as relevant, to the computation of brittleness attribute. For example, at 13402, vertical resolution is established. Depth sections from the in the wells of interest on the prospect or analogs are analyzed at 13404 to demarcate well log subsections that will be further analyzed. This can include the processes of geological log normalization, petrophysical normalization, and effective brittleness computation whereby different wells, that may have been drilled and logged by different companies at different times using different technologies and logging tools, are normalized so that these sections can be effectively compared.
  • sonic well log sections are extracted for cutoff regions.
  • well log section limits are established and sections of the logs copied and associated with brittleness values from core analysis within the sections of interest extracted as (starting depth, brittleness value) pairs of tuples.
  • the resulting volume would be a direct brittleness value predictor from seismic data volumes.
  • Additional geomechanical properties such as Young's modulus for a unit rock section can be obtained from a brittleness attribute prediction volume using an additional calibration table.
  • ending depth is not specified.
  • Example depth sections are 8 feet, 16 feet, 10 feet, 16 feet, 32 feet or their metric equivalents for prospects that follow metric convention.
  • Sonic well log sections are extracted at 13420 in a manner to fully encapsulate the well log section impedance (or acoustic velocity change). For example, if 16 feet sections are used for well log analysis to design saturation examples then a 32 feet sonic section will be extracted such that the 16 feet well log section lies between the sonic section.
  • the extraction 13420 can be also used if the sonic and well log sections are the same size and top or bottom aligned. If dipole sonic is available then the process of 13420 is repeated for both p-wave (compression) velocity and s-wave (shear) velocity.
  • a sonic log can be derived in various ways, if one has not been acquired for the reference well and can be used to implement 13420.
  • multiple reference log sections and corresponding sonic sections are obtained. These well and sonic log sections are sorted and arranged in increasing order to form groups. Using a histogram construction the multiple log sections are organized in groups or brittleness-bins to cover the range of interest established in 13410.
  • ⁇ wavelet analysis Using wavelet analysis, underlying wavelets are extracted from the corresponding sonic log sections of the well where bnttleness cores were taken. Sonic sections are associated with well log sections using multi-attribute software data structure such as list or n-dimensional array.
  • the initial wavelet used in 13430 is selected to match the same frequency bandwidth as the one selected for spectral decompositions.
  • brittleness computation is implemented on prime-amplitude or high frequency or full spectrum seismic data in 110.
  • the energy spectral density of each sonic section is computed at 13440 and the rearranged to form a line plot of energy spectral density versus increasing brittleness. Power spectral density or any power measure that can reduce the wavelet parametrics derived at 13430 to a scalar number can be used to implement the check at 13450.
  • a check is performed to see if the energy spectral density or an alternate scalar representation of wavelet parameter that encodes the information in sonic section is linear with a positive slope and the different sections corresponding to brittleness well section attribute are well separated, and accommodate the variance or standard deviation in different well sections belonging to different wells. If the results produced at 13450 show linearity then the parameters are output at 13470 to be used to develop well control voxel data and also in the excitation cascade engine at 808. If the sections are not separated out then the results at 13430 are convolved with another kernel from the rock physics library in 1235 in Figure 12 A. The same library is maintained and represented at 13455. The convolution between a rock physics modulation kernel from 13455 is implemented at 13460.
  • the resulting data is again parameterized using a wavelet transform such as Ricker or Ormsby wavelet.
  • Steps 13440, 13450 are repeated until linearity is established.
  • Selection of each new kernel from the library 13455 is also used in the Tl-TN steps shown at 811 to implement an excitation cascade. Satisfactory generation of output at 13470 and completion of linearity test implies that consistent well control is used to tie the well data to the seismic and to calibrate seismic attributes.
  • Use of multiple kernels, with each kernel representing a higher spatial frequency dimensionality to achieve linearity in energy spectral density or to achieve a related scalar measure is sufficient, but not necessary to optimize the tie at wells.
  • linearity cannot be established in a predetermined number of convolutions (e.g., 5 or 6 convolutions)
  • some log sections can be dropped from the process to offset the likelihood that logging could be in error, or some wells represent a petrophysical, structural or stratigraphic nonconformity or suboptimal well normalization.
  • Sufficiency of the linearity plot to tie well log to seismic using energy spectral density of convolved vectors provides accuracy in brittleness prediction from seismic datasets, away from well-bore during quantum resonance interferometry engine and Interferometric coupler iterations.
  • Figure 14 describes a process of implementing a quantum resonance
  • interferometry engine 828 in Figure 8 and quantum resonance interferometry processor 1022 in Figure 10 based on a mathematical model known to exhibit a quantum stochastic resonance ("QSR") phenomenon when driven by synthetic noise, such as synthetic noise 842.
  • QSR quantum stochastic resonance
  • the QSR phenomenon yields resonance events when certain conditions are met in the incoming data as in the output of excitation cascade engine 808 applied to prospect voxel data 805.
  • the quantum resonance interferometry engine in conjunction with the
  • interferometric coupler 1022 detects weak signals in the form of subtle variations in the intensity of data as in 811 by exploiting nonlinear interactions between (i) appropriately preconditioned voxel data transformed to a normalized excitation cascade from 805, and, (ii) a synthetically designed and dynamically conditioned complex noise output using 831 derived by mixing white/colored noise and noise from seismic.
  • this "synthetic complex noise” results from classical projection of a digitally simulated, 1-D quantum- mechanical (Q-M) system, modulated by classical white noise.
  • Output of 831 couples with the input data samples from output of 811 to generate resonance event 836.
  • a novel scalar observable output (that tracks changes to the dynamics of the simulated Q-M system) provides the basis for detection and quantitation of signal content in the input data.
  • the process 1400 includes selection and customization of the 1-D Q-M system and its nonlinear dynamics; and, selection and specification of the scalar observable used to conclude detection of an event that defines a condition for prospect voxel data to confirm the presence of an attribute of interest above a limit-of-detection established using positive control voxel data and negative control voxel data for an attribute of interest in Figure 13A through 13D.
  • Figure 14 describes an efficient method for developing quantum expressor functions to drive the interferometric coupler 833. It has been shown by those skilled in quantum mechanics theory that using variational methods, density operators could be used to unify different quantum mechanical theories, and classical mathematical approximations could emulate complex quantum mechanical effects in software implementation and reduce computational complexity and development time. Subject to the same stochastic limit approximation (SLA) which is at the core of QSR phenomenology in spin-boson systems, tunneling resonance events can be demonstrated in a broad class of systems whose time dependent dynamics can be described using the Bloch Rate equations. Furthermore, incorporating a master rate equation for time evolution of an NMR system, offers a simpler path for implementing tunneling resonance to detect and analyze events of interest. Within the NMR system the spin-boson bath coupling is replaced by a spin-thermal reservoir and thermal relaxation and decoherence times.
  • SLA stochastic limit approximation
  • the Quantum Resonance quantum resonance interferometry engine is calibrated to assess TR changes between - no signal condition (TRo) , a reference control - (i.e., positive control voxel data with known results - i.e., observed attribute values in well logs and rock core samples. ) - TR N and unknown data condition (TRu).
  • TRo no signal condition
  • TRu unknown data condition
  • a simple sinusoidal driving function (akin to the rocking of the boat in Stochastic Resonance) with a frequency ⁇ and amplitude ⁇ can be used to drive such a spin system.
  • the applied force is described the "perturbative" (not large compared to the system energy difference) Hamiltonian W:
  • X ⁇ R)(R ⁇ - ⁇ L)(L ⁇ [00210] X, they claim, is also an order parameter measuring the transitions between the states l-R) under the influence of the external periodic driving force.
  • Boson Noise Source is the Boson part of the Spin-Boson model.
  • the idea is that the Fermion system is coupled to a Boson energy reservoir that can either add or take away energy from the Fermion system.
  • theorem in quantum mechanics, which states that reversible fluctuations are possible in a system that is coupled in equilibrium to a thermal reservoir from which it can either draw thermal energy or dissipate energy into the reservoir.
  • the fluctuation-dissipation theorem is the basis for the so-called Brownian Motion or Johnson Noise in electrical circuits.
  • the spectrum of this thermal equilibrium noise is white, i.e. constant across all frequencies.
  • the Boson reservoir can be used as the analog for a thermal reservoir under the conditions of a Stochastic Limit Approximation.
  • strength of the interaction between Fermion system and the Boson system
  • trB trace of the matrix within the parenthesis over the boson degrees of freedom
  • PB Boson ground state density operator
  • Figure 14 implements a novel approximation to the spin-boson dynamics using a Density Operator for Two Level Systems (TLS).
  • TLS Density Operator for Two Level Systems
  • SR Stochastic Resonance
  • stationary colored noise not derived from the noise inherent in the input signal, is injected into the dynamical system (or into its analog or digital simulation).
  • the noisy input signal itself that is, the signal in which we wish to enhance and detect any signal of interest, if present - is applied (actually or in simulation) as a time-dependent perturbation to the dynamical system.
  • the simplest classical dynamical system of this kind that exhibits the SR phenomenon is that of a point-particle moving in a one-dimensional conservative potential, and subject to velocity-dependent friction (as well as the injected noise and the input signal).
  • This system in electronics, chemistry, optics and other branches of natural science, but the underlying mathematical model, which is more or less the same in all classical phenomena exhibiting an SR mechanism.
  • a QSR double well is implemented, incorporating an algorithmic model in which the point particle's inertia (mass) is negligible, its response to the potential gradient, the input signal and the injected noise is through the first derivative of the particle position; in other words, the nonlinear, stochastic, driven ordinary differential equation is first-order.
  • a dynamical mathematical model used in ViaLogy QuantumRD 3.0 is:
  • ⁇ ( ) is the potential, rescaled to encode the drag (friction) coefficient
  • f W is the noisy input signal dealt with in the previous sections.
  • W W is the injected noise. This is a stochastic process - e.g. it could be implemented using white (Wiener) noise again, or colored noise - but it is entirely uncorrected with the stochastic process used to model the inherent noise inside . For example, if w ⁇ is also a Wiener process, it has its own noise strength ⁇ , independent of
  • An exemplary mathematical choice for a potential which leads to SR is a bistable, two-well quartic (order 4) polynomial. As in 1405, there is flexibility in setting initial condition ( x ⁇ ) to initialize the dynamics , as long as for any given potential and input signal, the inherent noise in f O is sufficient to "jiggle" the double-well potential sufficiently to induce
  • the particle to cross the inter-well potential barrier in the absence of injected noise ( ⁇ - ⁇ ). And even if such is not the case, for any given input signal one can choose the parameters of the potential such that the particle, having once fallen into one of the two wells, will stay there, provided _ ⁇ . Any periodicity buried in will modulate the fluctuations of x ⁇ about that well's potential minimum, but no amplification of the SOI will occur.
  • the effect of a the pure-signal part of upon the bistable potential for low enough ⁇ is, in effect, to periodically tilt it; so that for one half-cycle the right well bottom is higher than the left well bottom, while for the other half-cycle of the period ⁇ ⁇ I ⁇ , the left well rises above the right one.
  • the transition time series exhibits a high correlation with a sinusoid of frequency ⁇ .
  • This effect can easily amplify the SNR of the input signal by a large factor, because it is the injected noise together with the parameters of the potential - and not the amplitude of the signal of interest - which determines the amplitude of the output signal.
  • f O enters the ODEs multiplicatively.
  • One way of controlling the process is by modulating the ambient temperature (simulated or actual).
  • Decoherence The injected noise, even if not explicitly put into the ODEs, enters through a relaxation time (or times), stemming from quantum decoherence due to interactions of the quantum system with its environment, which thus effects a partial quantum measurement upon the system.
  • Therm alization Another way in which noise is injected, is via a "quantum friction", again induced by the interaction of the quantum system with its environment.
  • quantum resonance interferometry engine implementation approach distinguishes between two categories of noise: inherent noise (present in the input signal and thus not controllable by the SR/QSR algorithm), and the injected noise added as part of the simulation of the dynamical physical system.
  • Injected noise can be a combination of different types: colored noise, thermal noise, and quantum-decoherence noise.
  • Figure 14 describes the QSR reformulation without Bosonic Modes.
  • the time-dependent signal to be analyzed and denoised which will be denoted , is used to modulate the above otential into the following time-dependent one:
  • ⁇ 0 ⁇ 0 ⁇ ⁇ + ⁇ ⁇ ,
  • the resulting Master Equation for the temporal evolution of P is a modification of the usual QSR Bloch rate equations; it has two independently-adjustable relaxation time constants.
  • the evolving density operator a 2 by 2 hermitian matrix in this case— represents a TLS with partially-measured ⁇ 2 , so that in general, neither the well (in terms of the pre-truncation potential) nor the wavefiinction is sharply determined at any given time ⁇ ; instead, they are "fuzzy".
  • tr denotes Hilbert-space trace
  • T > and T 2 represent the thermal and decoherence relaxation times, respectively.
  • the density operator is represented as a two-by-two matrix
  • the input signal is / J ( ' .
  • To generate all FFT's and SNR s ' replace and its transform and auto-correlation with the corresponding entities derived from (or, alternatively, either or Z ⁇ would have worked just as well).
  • a random number generator is used to generate "dice-throws" for successive ball positions - ⁇ terms of 3 ⁇ 4 .
  • Markovian quantum master equations generally satisfy the following conditions: (i) they describe the dynamics of time scales that are larger than the reservoir correlation time scale, (ii) these stationary solutions give the thermal equilibrium states with a reservoir. But have been shown to be valid only for single systems such as a single harmonic oscillator and a single two-level system.
  • NMR Nuclear Magnetic Resonance
  • E energy of the electron in the magnetic field
  • constant - also called the gyromagnetic ratio
  • hbar Planck's constant (1 ⁇ )/2 ⁇
  • Bo Strength of the external magnetic field.
  • M a the net magnetic moment of a material
  • Density Matrices Based on the Quantum Markov Approximation, spin-density matrices in a Markov-Chain Approach are set up where the future state of a system is only dependent on the current state and not on past states. The state update is implemented based on algorithm proposed by Nicholas Metropolis, Metropolis-Hastings algorithm.
  • the density matrix is defined as a Hermitian matrix of trace one that describes the statistical state of a quantum system.
  • a specific "density operator" for that particular observable operates on the density matrix for the system to create eigenvalues corresponding to the observable.
  • the operator is the Hamiltonian that stands for the total energy of the system.
  • the diagonal elements comprise of the energy difference between PotMinA and PotMinB ( ⁇ 0 ) to which is added the kth element of the input signal (power spectrum) vector.
  • the Delta ( ⁇ ) corresponds to a transverse magnetic field that causes the spin to flip orientation from the +ive z-direction to the -ive z-direction and vice versa.
  • Bloch Equations Now one can construct the "Bloch” Equations of Motion, which are a kind of Maxwell Equations for NMR, that describe the dynamics of spin motion under RF excitation in the presence of the thermal environment and spin-spin decoherence: dt
  • Bi transverse RF magnetic field ( ⁇ in our case) and ⁇ 0 is known as the Larmor frequency (constant) and ⁇ is the time dependent precession frequency, which I am not sure how to relate to our equations. I believe it may show up as an energy asymmetry term.
  • Bloch equations describe the evolution of the magnetic moment, they can be converted to time dependent energy equations (i.e. Hamiltonian evolution) by multiplying both sides by ⁇ .
  • x, y, and z denote the variables in Bloch Rate equations customized from the general form given in Eqn. 10, 11, and 12.
  • xo denotes the state variable for the NMR particle, a and ⁇ denote the thermalization coefficients.
  • the enhanced voxel and data and Tunneling Rate are reset per
  • barrierWidth 2 ⁇ e- ⁇ TM ⁇ " ⁇ [00303] where the system energy ratio is a user set parameter.
  • the Bloch equations are a set of coupled differential equations which can be used to describe the behavior of a magnetization vector in an NMR mathematical model under any conditions. When properly integrated, the Bloch equations will yield the x', y', and z ' components of magnetization as a function of time. Only the behavior in ' is used to track changes in the particle state to conclude resonance events.
  • the tunneling rate (TR) can be calculated using the logic
  • a 1480 a check is performed to assess whether the iteration count has been exceeded. If so, then the process 1400 stops the NMR-QSR iterations and outputs a prospect voxel output.
  • the prospect voxel output is defined as Enhanced Data in 839 and the tunneling rate TR .
  • quantum resonance interferometry processor 831 can be implemented using the steps 1405 through step 1420 of process 1400. Also, the interferometnc coupler 833 can be implemented using steps 1440 through 1480.
  • the quantum resonance interferometry processor 1022 can be implemented using the steps 1405 through 1480. Only the enhanced data computed using 1450 is output, and re-introduced into the Interferometnc Coupler 833 as an input to generate resonance 836.
  • quantum resonance interferometry processor 831 can be implemented using the steps 1405 through 1420.
  • the Interferometnc Coupler 833 can be implemented using steps 1440 through 1480 as in Figure 8.
  • the QRI Processor 831 is implemented using the Unit 1405 through Unit 1420 as in Figure 8.
  • Figure 15 shows an example of a double well function that represents the quartic potential implemented in 1405.
  • FIG 16 shows an example of work flow 1600 using a quantum resonance interferometry engine 1628, such as quantum resonance interferometry engine 828.
  • the work flow 1600 is denoted as a differential interferometry mode whereby the presence of an attribute of interest in a normalized excitation cascade 811, derived from a prospect voxel data 805, is determined by comparing data known to possess the same attribute of interest with the output of a Well Control Voxel.
  • Differential Interferometry mode is used to produce a seismic data output volume from all Prospect Voxels in a formation of interest with the same property as found in well control voxel data, termed as a "Reference Voxel Data".
  • Reference Voxel Data can be developed from a single well control voxel or multiple Well Control Voxels using the methods of Figure 13A through 13D.
  • Reference voxel data is developed by analyzing the well log relevant to the prospect question of interest posed for the prospect, per the flows in Figure 13A though 13D and determining the sections in the well logs from well control that meet criteria of interest to develop well control voxel data 1620 denoted as Reference voxel data.
  • the prospect voxel data from seismic partitioned volumes as in Figure 5 are transformed to normalized excitation cascade vectors and analyzed by the quantum resonance interferometry engine 1628.
  • the results from quantum resonance interferometry engine 1628 are compared to see if voxel data possesses the same property as reference voxel data to answer the prospect question of interest.
  • Work flowl600 shows the differential interferometry mode of operation.
  • presence of an attribute of interest in a normalized excitation cascade such normalized cascade 811 derived from prospect voxel data 805 is determined by comparing output at 839 with the output of a normalized excitation cascade derived from a well control voxel data.
  • a prospect voxel data could be compared against positive well control voxel data or negative control voxel data using differential interferometry mode of operation as inl600.
  • prospect voxel data could be compared against one or more well control voxels, obtained by using the flow in Figure 12-A and 12-B.
  • Output of the work flow 1600 is a determination of presence or absence of an attribute of interest in the prospect voxel data.
  • a normalized excitation cascade is produced from a prospect voxel data using the process 1200 in Figure 12A and 12100 in Figure 12B.
  • a normalized excitation cascade termed as Reference Voxel Data, is produced from one or more well control voxels using the flow 1200 in Figure 12A and 12100 in Figure 12B.
  • the Well Control Voxels are generated and processed per the flows described in Figure 13A through 13D.
  • Outputs from 1610 and 1620 are introduced into the quantum resonance interferometry engine828, driven by both seismic noise such as seismic noise 814, produced, for example, using the flows in Figure 9-A through 9-B, as inputs in a pair-wise sequence - 161.0-followed-by 1620, or 1620-followed by 1610.
  • two versions of the- quantum resonance interferometry engine 828 may be used to independently process 1610 and 1620 with the results recombined in 1630.
  • quantum resonance interferometry engine 1628 produces a Tunneling Rate outoutl630 per the flow described in process 1400. Once the output of 1630 has been obtained from processing inputs 1610 an 1620, the tunneling rate outputs are compared in 1640 to determine ' if resonance was obtained using both inputs 1610 and 1620.
  • Interferometric Resonance Dipole requires quantum resonance interferometry engine 1628 to be configured in a manner such that resonance events occur when Reference Voxel Data is presented, and no resonance events occur when only inputs from seismic data noise as 1614 (such as seismic noise 814) are presented.
  • Tunneling Resonance is concluded when there is a statistical difference between the Tunneling Rate Change as determined by observing the Tunneling Rate using prospect voxel data as an input or Reference Voxel Data as input compared to the Tunneling Rate output obtained using only seismic noise as an input.
  • a statistical difference criterion for Interferometric Resonance Dipole 1640 is established by running multiple inputs derived from seismic noise and well control voxel data and computing the mean and standard deviations for the runs. As an example, in one embodiment 2s different between the TR output from well control voxel data as input and seismic noise data as input is used as a measure of statistical significance to establish threshold for 1640.
  • the process 1600 can applied to sequence all prospect voxel data derived from the partitioned seismic volume of interest in 120.
  • Figure 17 A shows an example of a process 17100 to compute and predict formation porosity from seismic data.
  • Workflow 17100 can be performed using the systems and techniques discussed herein, such as by using quantum resonance interferometry engine 828.
  • Different ranges of porosity are of interest in different formations for different prospects.
  • the economics of oil and gas exploration and production specify cut-offs on porosity values that are relevant to positioning new wells or new offset wells on a prospect. For example, drilling locations for offset wells are determined in relationship to an existing well and may be subject to inter-well spacing requirements constraints imposed by governing regulations for that particular formation. For example, in some states in the US only 1 well can be drilled on a 40 acre spacing for certain formations.
  • Process 17100 can determine a porosity-cube when it is implemented iteratively to determine a maximum porosity estimated for each of multiple prospect voxelsOnce a porosity range of interest is established for a prospect, it is broken down into intervals that can be analyzed based on well control on the prospect or using analogous well controls from other fields, prospect voxel data for the prospect voxels are tested for porosity above a threshold estabUshed by the lowest interval set for the range of interest.
  • the workflow 17100 can be performed iteratively such that in successive iterations only voxels that exceed the porosity threshold established in the previous run are analyzed to assess if they exceed a new threshold, prospect voxel data that exceed the porosity criteria in iteration are denoted as "Surviving-Prospect Voxel" data. This-process-is repeated for surviving voxel data until the . entire porosity range is covered.
  • Prime-amplitude spectral decomposition 560 or high frequency spectral decomposition volume 560 is obtained at 17102 as a starting input for computations specific to formation porosity estimation. Such starting inputs can be obtained for example at step 560 in Figure 5.
  • the starting input is voxelized at 760, (such as is described in connection with Figure 7) to produce a sequence of prospect voxel data.
  • Individual Prospect Voxels are processed sequentially according to the process 17100 to construct a Porosity-Cube output 17160.
  • a normalized excitation cascade is generated (such as described in connection with 12190) from the prospect voxel data and input to the quantum resonance interferometry engine 828.
  • Aporosity range of interest is obtained as an external input from, e.g., the geologist, operator or a lease stakeholder at 17101.
  • the porosity range of interest is used to produce cut-off thresholds at 17105 to select the appropriate well log sections to develop well control voxel data.
  • the porosity range is reduced to multiple porosity increment intervals, such as 4% increments of porosity (e.g., for sandstone) or 2% (e.g., for detrital limestone such as Strawn).
  • porosity cut-off thresholds for porosity increment intervals from 17105 are used to develop the well control voxel data (e.g, such as by using the flow detailed in Fig 13A) to produce an output as in 13170.
  • the well control voxel data is used to calibrate 17407 the parameters of the quantum resonance interferometry engine 828 to be used (e.g., in implementing the flow described in 1400) to generate resonance events when Prospect Voxels are derived from regions in the formation of interest that have porosity matching or exceeding the porosity as seen in the well control on the prospect of analogous wells drilled in the same formation at another prospect.
  • These criteria include the calibration of parameters for a double well function, such as is described in Figure 15 as related to the determination of Tunneling Rate and resonance.
  • Double well parameters are reset for each successive porosity interval iterated by the quantum resonance interferometry engine 828 on Surviving prospect voxel data.
  • Double well parameters can include: Initialization of energy asymmetry in the double well potential function, Interwell spacing
  • Output of quantum resonance interferometry engine 828 for a prospect voxel is examined for Tunneling Resonance (e.g., by using the process 1700) to see if the prospect voxel data meets the porosity threshold criteria provided in 17101 and used to produce well control prospect voxel data.
  • the prospect voxel data, in the first iteration of the quantum resonance interferometry engine, and surviving prospect voxel data in successive runs of the quantum resonance interferometry engine are processed by the quantum resonance interferometry engine 828 to produce an output which is evaluated at 17130.
  • porosity criteria e.g., based on method in 1400 or using Differential Interferometry method of 1600
  • the prospect voxel data is concluded to be a surviving prospect voxel data for a next iteration. If the porosity threshold is not exceeded then the prospect voxel is not iterated in any subsequent iteration.
  • a check is performed at 17140 to ensure that all prospect voxel data or surviving prospect voxel data for a current iteration has been processed. If not, then next prospect voxel or Surviving prospect voxel data in the sequence is retrieved in 17170 and the process repeated from 17190 in flow 17100. If the check in 17140 indicates that all prospect voxel data or surviving prospect voxel data have been processed then another check 17150 is performed to see whether the upper limit of porosity for the workflow has been exceeded. If the check in 17150 is successful then the Porosity-Cube result is output such as in SEGY notation. If the check is 17150 is unsuccessful then the porosity threshold is updated to the next interval established in 17145. The quantum resonance interferometry engine parameters (e.g., double well parameters) are updated for the next iteration. And, the process is repeated.
  • the quantum resonance interferometry engine parameters e.g., double well parameters
  • the process 17100 can be used to implement one or more porosity intervals of interest. It can also be used to derive all voxels above a porosity threshold of interest by establishing a single interval [0,threshold] and assessing the set of Surviving prospect voxel data after the first iteration. This set of voxels in surviving prospect voxel data provides areas in the formation that have formation porosity above or below a pre-specified cut-off threshold of interest.
  • Figure 17B shows an example of a workflow 17200 to compute and predict Vshale levels using seismic data. Workflow 17100 can be performed using the systems and techniques discussed herein, such as by using quantum resonance interferometry engine 828.
  • Vshale is often used as avoidance-attribute, i.e., areas with high Vshale are to be avoided for positioning a well.
  • the economics of oil and gas exploration and production specify lower cut-offs on Vshale values that are relevant to positioning new wells or new offset wells on a prospect.
  • Accurate knowledge of subsurface formation Vshale is used to optimally position wells to avoid dry holes and maximize recovery.
  • the commercial exploration criteria for Vshale are provided, for example, by a client, project geologist or by the exploration manager or the lease owner/operator.
  • Process 17200 can determine a Lithology-Cube when it is implemented iteratively to determine a maximum Vshale estimated for each of multiple prospect voxels . Once a Vshale range of interest is established for a prospect, it is broken down into intervals that can be analyzed based on well control on the prospect or using analogous well controls from other fields, prospect voxel data for the prospect voxels are tested for Vshale above a threshold established by the lowest interval set for the range of interest.
  • the workflow 17200 can be performed iteratively such that in successive iterations only voxels that exceed the Vshale threshold estabUshed in the previous run are analyzed to assess if they exceed the new threshold, prospect voxel data that exceed the Vshale threshold in iteration are denoted as "Surviving Prospect Voxel" data. This process is repeated for Surviving prospect voxel data until the entire porosity range is covered. Transformations implemented by 17200 are
  • Either a prime-amplitude spectral decomposition or high frequency spectral decomposition volume is obtained at 17202 as a starting input for computations specific to formation Vshale estimation. Such starting inputs can be obtained for example at step 560 in Figure 5.
  • the starting input is is voxelized at 17290 (such as is described in connection with Figure 7) to produce a sequence of prospect voxel data .
  • Individual Prospect Voxels are processed according to the process 17200 sequentially to construct a Lithology-Cube output 17260.
  • a normalized excitation cascade is generated (such as is described in connection with 12190) from a prospect voxel data and input to the quantum resonance interferometry engine 828.
  • a Vshale range of interest is obtained as an external input, e.g., from the geologist, operator or a lease stakeholder at 17201.
  • the Vshale range of interest is used to produce cut-off thresholds as in 17205 to select the appropriate well log sections to develop well control voxel data.
  • the Vshale range is reduced to multiple of Vshale increment intervals, such as 20% increments of Vshale (e.g., for sandstone) or 10% increments (e.g., for detrital limestone such as Strawn).
  • a 17220, Vshale cut-off thresholds from 17205 are used to develop the well control voxel data, such as by using the flow detailed in Fig 13B to produce an output as in 13270.
  • the well control voxel data is used to calibrate 17207 the parameters of the quantum resonance interferometry engine 828 to be used in implementing the flow described in 1400, to generate resonance events when Prospect Voxels are derived from regions in the formation of interest that have Vshale matching or exceeding the Vshale as seen in the well control on the prospect of analogous wells drilled in the same formation at another prospect.
  • These criteria include the calibration of parameters for a double well function, such as is described in Figure 15 as related to the determination of Tunneling Rate and resonance.
  • Double well parameters are reset for each successive Vshale interval iterated by the quantum resonance interferometry engine 828 on Surviving prospect voxel data.
  • Double well parameters can include:
  • Output of quantum resonance interferometry engine 828 is examined for Tunneling Resonance, such as by using the flow 1700, to see if the prospect voxel Data meets the Vshale threshold criteria provided in 17201 and used to produce well control prospect voxel data.
  • the prospect voxel data, in the first iteration of the quantum resonance interferometry engine, and surviving prospect voxel data in successive runs of the quantum resonance interferometry engine are processed by the quantum resonance interferometry engine 828 to produce an output which is evaluated at 17230.
  • Vshale criteria e.g., based on method in 1400 or using differential Interferometry method of 1600
  • the prospect voxel data is concluded to be a surviving prospect voxel data for a next iteration. If the Vshale threshold is not exceeded then the prospect voxel is not iterated in any subsequent iteration.
  • a check is performed at 17240 to ensure that all prospect voxel data or surviving prospect voxel data for a current iteration has been processed. If not, then next prospect voxel or Surviving prospect voxel data in the sequence is retrieved in 17270 and the process repeated from Unit 17290 in flow 17200. If the check in 17240 indicates that all prospect voxel data or surviving prospect voxel data have been processed then another check 17250 is performed to see whether the upper limit of Vshale for the workflow has been exceeded. If the check in 17250 is successful then the Lithology-Cube result is output such as in SEGY notation. If the check is 17250 is unsuccessful then the Vshale threshold is updated to the next interval established in 17245. The quantum resonance interferometry engine parameters (e.g., double well parameters) are updated for the next iteration. And, the process is repeated.
  • the quantum resonance interferometry engine parameters e.g., double well parameters
  • the flow 17200 can be used to implement one or more Vshale intervals of interest. It can also be used to derive all voxels above or below a Vshale threshold of interest by establishing a single interval [0,threshold] or [threshold, upper range for formation] respectively, and assessing the set of surviving prospect voxel data after the first iteration. This set of voxels in surviving prospect voxel data provides areas in the formation that have Vshale above or below a prespecified cut-off threshold of interest.
  • Figure 17C shows an example of a process 17300 to determine fluid presence and fluid type, including oil, gas, and water using seismic data
  • Process 17300 can also be used to determine a oil-water contact boundary or a gas-water boundary.
  • processl7300 implements a search for a fluid marker (for gas, oil or water) derived from well control voxel data.
  • process 17300 implements a quantitative estimation of low frequency spectral attenuation in prospect voxel data. It is known to those skilled in the art, and has been shown by empirical measurements on controlled experiments, that lower frequencies in acoustic propagation and reflection are differentially attenuated by fluids such as oil, water and gas.
  • Process 17300 distinguishes between fluid-in place or hydrocarbon trap and contrast it with areas where hydrocarbon was once present but has now migrated and may be replaced by brine or freshwater. This has commercial implications for both conventional onshore and off-shore carbonate and clastic plays. It is also used to explain the spotty behavior observed when wells and offset wells drilled close together often have different outcomes in hydrocarbon recovery and presence. Process 17300 detects hydrocarbon related
  • process 17300 can detect spectral attenuation in seismic data, specifically using low frequency spectral decomposition volumes. Such a technique is apphed to low frequency spectral decompositions but can also be applied to high frequency and prime amplitude spectral decompositions as well.
  • Low frequencies are used to detect, discriminate and delineate fluids and oil water contact. It is not the image quality or the low frequency itself that is important, but the ability to see relative attenuation in the low frequencies as established by differences between petrophysically identified zones of interest and water-wet or zones with no fluid.
  • the method effectively extends the low frequencies driving the analysis by anywhere from 2Hz to 6Hz at the target depending on the resonant frequency of the geophones, assuming that active vibration source dwelled in the low frequency region for a suitable time window (depending on the formation of interest).
  • method of Figure 13C as in 13300 to design well control voxel data ensures that positive control voxel and negative control voxel are transformed to 1 -D Spectral data packets with frequencies where the spectral energy attenuation due to varying fluid type and fluid saturation separates out in the well control regions of interest; and these well control voxel data can be extracted from PSTM/PSDM seismic volume spectrum decompositions, at low frequency intervals.
  • amplitudes of extracted spectrum decompositions to be above -22dB to -24dB, to satisfy boundary conditions for the underlying quantum resonance interferometry engine calibrated to provide a 1 : 100 SNR detection.
  • Different calibration would be needed to process prospect voxel data and design well control voxel data if amplitudes below -24db are to be used.
  • the voxels are convolved with one or more synthetic carrier kernels (derived using Ricker wavelets with varying amplitudes and side lobes) as in 1200
  • a low frequency spectral decomposition volume is obtained as a starting input for fluid estimation.
  • Such starting input can be obtained for example at step 560 in Figure 5.
  • the starting input is is voxelized (such as is described in connection with Figure 7) to produce as a sequence of prospect voxel data, prospect voxel data for individual prospect voxels are processed according to process 17300 sequentially to construct a Fluid Volume output 17360.
  • a normalized excitation cascade is generated (such as is described in connection with 12190) from a prospect voxel data and input to the quantum resonance interferometry engine 828.
  • Fluid Marker libraries refer to the results of empirical studies, or laboratory-experiment observations, or core analysis based calibration tables that specify the relationship between rock-type, fluid saturation, relationship between different well logs, and depth for different formations, traps and fields.
  • Data may be available from the same geological basin or field as the prospect of interest.
  • Well control for the prospect may include results of core analysis that report hydrocarbon saturation levels with sections of the wells already drilled on the prospect.
  • handbooks may be available that provide relationship between resistivity logs and fluid saturation levels.
  • derived waterlogs, constructed from other logs and core data are available, available fluid property data can be used to establish equivalence with sections within drilled wells on the prospect or on an analog.
  • well control voxel data is determined at 17305 (such as is described in connection with Figure 13C ).
  • the output from 17305 is denoted as a fluid marker as it presents changes in energy spectral density levels exceedance which can be an indicator of fluid presence or type of fluid present in the formation of interest.
  • processl7300 can output a fluid volume that represents prediction of spectral attenuation over the formation.
  • the fluid marker and well sections are used to produce cut-off thresholds at 17305 to select and develop well control voxel data.
  • the output from 17305 is used to calibrate 17307 the parameters of the quantum resonance interferometry engine to be used (for example, as when implementing the flow described in 1400) to generate resonance events when Prospect Voxels are derived from regions in the formation that have fluid marker or levels of spectral attenuation exceeding a cut-off threshold of interest.
  • These criteria include the calibration of parameters for a double well function (for example, as described in Figure 15) as related to the determination of Tunneling Rate and resonance.
  • Double well parameters can include:
  • Tunneling Resonance (e.g., by using the flow 1700) to see if the prospect voxel data meets the spectral attenuation threshold criteria provided in 17301 and used to produce well control voxel data.
  • the prospect voxel data, in the first iteration of the quantum resonance interferometry engine are processed by the quantum resonance interferometry engine 828 to produce an output which is evaluated in 17330. If it is determined that attenuation criteria is exceeded (e.g., based on method in 1400 or using differential interferometry method of 1600) then the prospect voxel data is concluded to be a surviving prospect voxel data for a next iteration. If the spectral attenuation threshold is not exceeded then the prospect voxel is not iterated in any subsequent iteration
  • This set of voxels in surviving prospect voxel data provides areas in the formation that have spectral attenuation above or below a pre-specified cut-off threshold of interest.
  • Figure 17D shows an example of a process 17400 to compute and predict brittleness levels using seismic data. Different ranges of brittleness are of interest in different formations for different prospects. Exploration criteria for brittleness can be obtain, e.g. from a client, project geologist or an exploration manager or the lease owner/operator.
  • Process 17400 is implemented using an iterative workflow to determine the maximum brittleness estimated for prospect voxel data for each of multiple prospect voxel data. Once the brittleness range of interest is established for a prospect, then it is broken down into intervals that can be analyzed based on well control on the prospect or using analogous well control from other fields. All the prospect voxel data are tested for brittleness above a threshold established by the lowest interval set for the range of interest.
  • Prime-amplitude spectral decomposition or high frequency spectral decomposition volume is obtained at 17402 as a starting input for computations specific to formation brittleness estimation. Such starting inputs can be obtained for example at step 560 in Figure 5.
  • the starting input is voxelized at 17490, (such as is described in connection with Figure 7) to produce a sequence of prospect voxel data
  • Individual Prospect Voxels are processed sequentially according to the process 17400 to construct a brittleness-Cube output at 17460.
  • a normalized excitation cascade is generated (such as described in connection with 12190) from the prospect voxel data and input to the quantum resonance interferometry engine 828.
  • a normalized excitation cascade is generated from a prospect voxel data and input to the quantum resonance interferometry engine 828.
  • a brittleness range of interest is used to produce cut-off thresholds at 17405 to select the appropriate well log sections to develop well control voxel data.
  • the brittleness range is reduced to multiple brittleness increment intervals.
  • brittleness cut-off thresholds from 17405 are used to develop the well control voxel data (e.g.,using the flow detailed in Fig 13D to produce an output at 13470).
  • the well control voxel data is used to calibrate 17407 the parameters of the quantum resonance interferometry engine 828 to be used (e.g., in implementing the flow described in 1400) to generate resonance events when prospect voxel data are derived from regions in the formation that have brittleness matching or exceeding the brittleness as seen in the well control on the prospect of analogous wells drilled in the same formation at another prospect.
  • These criteria include the calibration of parameters for a double well function (e.g, described in Figure 15) as related to the determination of Tunneling Rate and resonance.
  • Double well parameters are reset for each successive brittleness interval iterated by the quantum resonance interferometry engine 828 on Surviving prospect voxel data.
  • Double well parameters can include:
  • Output of quantum resonance interferometry engine 828 for a prospect voxel is examined for Tunneling Resonance (e.g, by using the flow 1700) to see if prospect voxel data meets the brittleness threshold criteria provided in 17401 and used to produce well control prospect voxel data.
  • the prospect voxel data, in the first iteration of the quantum resonance interferometry engine, and surviving prospect voxel data in successive runs of the quantum resonance interferometry engine are processed by the quantum resonance interferometry engine 828 to produce an output which is evaluated at 17430.
  • the prospect voxel data is concluded to be a surviving prospect voxel data for a next iteration. If the brittleness threshold is not exceeded then the prospect voxel is not iterated in any subsequent iteration.
  • a check is performed atl7440 to ensure that all prospect voxel data or surviving prospect voxel data for a current iteration has been processed. If not, then next prospect voxel or Surviving prospect voxel data in the sequence is retrieved in 17470 and the process repeated from Unit 17490 in flow 17400, onwards. If the check in 17440 indicates that all prospect voxel data or surviving prospect voxel data have been processed then another check 17450 is performed to see whether the upper limit of porosity for the workflow has been exceeded. If the check in 17450 is successful then the Lithology-Cube result is output, such as in SEGY notation. If the check is 17450 is unsuccessful then the brittleness threshold is updated to the next interval established in 17445. The quantum resonance interferometry engine parameters (e.g., double well parameters) are updated for the next iteration. And the process repeated.
  • the quantum resonance interferometry engine parameters e.g., double well parameters
  • the process 17400 can be used to implement one or more brittleness intervals of interest. It can also be used to derive all voxels above or below a brittleness threshold of interest by establishing a single interval [0,threshold] or [threshold, upper range for formation] respectively, and assessing the set of surviving prospect voxel data after the first iteration. This set of voxels in surviving prospect voxel data provides areas in the formation that have brittleness above or below a pre-specified cut-off threshold of interest.
  • the quantum resonance interferometry engine 828 can be configured in two stages.
  • Figures 18 A and 18B show configuration in a training mode and an operations mode respectively.
  • a training mode 1810 results in acalibration of parameters for:
  • Training data includes: examples of normalized excitation cascade data derived from well control voxel data for the appropriate; and properly conditioned synthetic and seismic data noise.
  • a subset of well control voxel data is used for testing to verify if the Training Mode is successfully completed.
  • the entire dataset used for Training Mode is denoted as the "Training Set".
  • the training set can include well control voxel data with porosity, Vshale and spectral attenuation for fluid attribute values below a cutoff of interest (or above as the case maybe e.g., in Vshale attribute computation).
  • Operational mode 1820 results in the use of a calibrated quantum resonance interferometry engine initialized and calibrated by the training mode 1810 to detect and characterize prospect voxel data .
  • the operational dataset for the operational mode 1820 is denoted as the "Input set”.
  • the quantum resonance interferometry engine 8 requires input data to be a series of real numbers with the following properties:
  • Well Control sections are an example of data that can be used in establishing a Quality Metric.
  • An example of a specific Quality Metric used during training mode 1810 is a correlation coefficient for first temporal moment of samples (1 .. n) where n is the number of well control voxel data used in the Training Set, correlated to the signal strength of all labeled examplers (i.e., training set members with associated signal levels).
  • the Quality Metric is required to be monotonic to signal strength as specified for all signal examplers. If the above test cannot be met using the first (1st-) temporal moment, then the second (2 nd ) or third (3 rd -) temporal moment are checked.
  • first-order correlation does not yield a monotonic result for 1st, 2 nd , 3 rd order temporal moments
  • the a second-order partial correlation is applied. If the second-order partial correlation measure does not yield a monotonic relationship as defined above, then the transformations [Tl ... TN] is reimplemented as in 808 to generate an excitation cascade.
  • an Interferometric Coupler 833 takes its input - "Input Data”, and implements iterative reverberations within Unit 833 to produce two result outputs: labeled as "TR” and "Enhanced Data”.
  • TR - represent Transition Rate in the classical implementation of Equations 10 through 15, or Tunneling Rate as in the quantum- mechanical implementation of Equations 10 through 15. It is a measure of signal to noise ratio enhancement (SNR). In general, higher TR value is indicative of higher SNR.
  • SNR signal to noise ratio enhancement
  • TR characterizes the impact of noise-modulated preconditioned input data on the dynamics of ID NMR-based Q-M spin system.
  • Enhanced Data output in 18110 represents a series of output vectors of real numbers that has the same cardinality as normalized excitation cascade presented as an input vector to the interferometric coupler 833.
  • the Enhanced Data from Interferometric Coupler 833 represents an SNR enhancement over SNR of the input data.
  • Domain-specificity is encoded in the Interferometric Coupler 833 via specification of (i) TR- threshold for separating noise from signals above LOD as in resonance event 836; and (ii) calibration curve for signal quantitation to generate resonance amplitudes as in 839 in Figure 8 and Figure 10.
  • TR exhibited by Input samples is separated from TR exhibited by noise samples (also denoted as TR Threshold) by one or more standard deviations then signal presence can be concluded.
  • TR Threshold also denoted as TR Threshold
  • Enhanced Data vector has no utility in signal detection, wherein signal presence is solely concluded based on statistical significant changes in TR on a calibrated dataset, but its properties can be used to assess correctness of QSR-NMR iterator dynamics.
  • Enhanced Data from 833 does however play a role in signal quantitation, where Enhanced data may be applied to a Calibration curve to assess how much signal is present on a pre-calibrated scale.
  • is a tolerance threshold, e.g., ⁇ 0.05. If convergence has not occurred then the Enhanced Data is fed back to the TR Computation block for another process iteration.
  • the parameters controlling the TR convergence process in the operation mode include:
  • TR-background the mean TR value computed during the training phase by processing a number of sample data vectors known to contain no signal (pure noise).
  • TR- background is calculated during training by using a set of background (no signal) samples;
  • convergenceSteps is a small fraction of maxIterConvergence and is normally derived from it (for example convergenceSteps is typically 5 or 7 consecutive iterations, but less than 5% of maxIterConvergence);
  • the TR computation module has a Interferometric Couple QSR-NMR engine iterator as a core element.
  • the Interferometric Couple QSR-NMR engine module is executed for a fixed, pre-determined number of iterations. These Iterations, denoted as Innerlterations, correspond to a number of repeats the Input sequence is presented.
  • the QSR module takes as its input the data as produced by the Data Normalization module and combines it with internally generated noise.
  • each iteration uses the enhanced data produced in the previous iteration step, which is fed back to Interferometric Couple QSR-NMR engine after it is combined with a freshly-generated noise vector.
  • the classical additive noise parameters changes with each iteration (basically the noise strength is being incremented).
  • the mean value of TR is calculated and the last enhanced data vector produced by QSR is deemed as the ultimate output vector.
  • TR Computation Iterations (maxIterTRComputation) - This is the maximum number of iterations used for TR computation. This is set to at least twice the number of iterations required to double SNR for Signal@LOD case during the training phase. Use 200 as a default value;
  • Minimum Noise Strength (minNoiseStrengthTRComputation) - This is the noise strength used for the first run of QSR A good initial value of less than 5% of Max Value of (Input Data) in Figure 5;
  • Noise Strength Increment (noiseStrengthlncrementTRComputation) - The amount by which the noise strength is increased for each iteration.
  • the process within the Interferometric Couple QSR-NMR engine module is derived from Quantum Stochastic Resonance theory implementation on an NMR system.
  • the QSR module takes as input data a normalized vector of real numbers with absolute amplitudes ( ⁇ 0.3) representing the uncharacterized signal (buried within noise), and essentially implements Quantum Stochastic Resonance dynamics to determine the tunneling rate between the two quantum-mechanical "energy" states within the QSR model.
  • the QSR module also generates the "enhanced" data vector of real numbers that have a higher SNR for the buried signal than the original input data.
  • Signal Power for Input Data and Enhanced data is computed by estimating area under the half width at full height. SNR can compute using Area for signal and noise in the two cases:
  • Enhanced Data and Input Data signal-to-noise ratio is a term for the power ratio between a signal and the background noise:
  • the TR threshold of interest established within 828 undertaken during the training phase after the parameters of the quantum resonance interferometry engine have been established, as described above. It is possible though, that the training phase is performed in an iterative mode with the two processes (QSR optimization and TR threshold determination) being repeated a number of times. For example, if one fails to establish a TR threshold with the initial set of optimized parameters, one must readjust the QSR engine parameters by varying the SNR gain. Then, the TR threshold determination process is attempted again.
  • the quantifier is a decision block where the TR for the current data sample is compared with the pre-determined threshold value obtained during the training phase, in order to decide whether or not the signal is present in the input data.
  • the TR threshold determination a number of control data samples must be used. These data samples fall in two categories:
  • signal - A fixed value of signal (with SNR higher than LOD) is present.
  • each category a minimum of 20 samples is desired.
  • Each data sample generates a corresponding TRmean value when processed by the QSR engine.
  • TRB and TRs for background and signal respectively.
  • the QSR engine is optimized correctly only if these two sets of TR values are statistically different.
  • the standard deviation and the mean of each set of TR values are computed and assigned: as, MB, as, Ms for background and signal respectively.
  • the critical difference parameter is then used to evaluate the difference between the two TR data sets.
  • n s and ⁇ are the number of samples in the signal and background datasets respectively.
  • the CR value must be outside the interval [- 1.96, 1.96]. There may be two reasons for the value of CR not falling within the interval.
  • the data samples have not been selected properly (signal is present in the background samples or not enough is present in the signal samples). This can be fixed by revisiting the sets of selected samples or, if this is an option, generate some artificial data samples where the SNR can be tightly controlled.
  • the blocks upstream have not been optimized. This takes the tuning process back to the beginning. The parameters have to be readjusted .
  • the TR threshold is calculated as:
  • thresholdxR ( ⁇ ⁇ + o s ) 2. The threshold so calculated is used in the operation mode to decide convergence.
  • Figure 19 shows a method 1900 for calibration of quantum resonance interferometry engine Parameters.
  • quantum resonance interferometry engine is initialized prior to first iteration of the first prospect voxel (as discussed at flows 17100, 17200, 17300, and 17400).
  • the attribute of interest parameters are updated to process the next interval (as discussed in connection with in 17145, 17245, 17345 or 17445)
  • the quantum resonance interferometry engine parameters are again updated to process the surviving prospect data voxel as defined in discussion related to Figure 17A through 17D.
  • quantum resonance interferometry engine Parameters are computed and include the following:
  • NMR-QSR Double Well parameters include:
  • Point Count - This is a number representing the size of the Input Data vector. It is preferred that the Point Count be a power of 2 but this is not a requirement. Point Count of 32 (when using delta-pulse kernel), 64 (when using a Gaussian peak kernel) and 128 (when using a CWT - continuous wavelet transform kernel) are used as examples.
  • the actual vector size may depend on the number of raw data elements (for example the number of pixels within the window used to sweep an input image). Conformity (changing the vector size to be a power of 2) is achieved during the preprocessing preconditioning T
  • Point Count is actually result of discretization of 1-D 1-peak kernel used for preconditioning and normalizing input to the Interferometric Coupler;
  • Noise Color - Periodic synthetic noise is required to drive NMR-QSR dynamics. This noise is characterized by its color and strength (amplitude). The default value of noise color may be left unchanged at a noise color of 1 (white noise). This is used as a control parameter to speed/slow down SNR gain when using quantum resonance
  • interferometry engine It has limited influence and can destabilize the QSR dynamics is set to be » 1. ) but it is adjustable as a last resort measure for detecting signal when the adjustment of other parameters fails to produce the desired SNR;
  • Noise Strength is the primary adjustable value outside of double well design, and directly controls the detection of signal by improving the output SNR;
  • Residency Time In order for the quantum resonance interferometry engine to decide that a transition has occurred during the processing, the quantum resonance interferometry engine must show some stability, i.e. the state does not exhibit random transitions.
  • -Residency Time is merefore a stability criterion which essentially specifies the minimum time that the system state must remain in one of the two quantum-mechanical states.
  • the Residency Time can in fact be expressed as the number of time steps (as mentioned above, one time step corresponds to processing the next element within the input data vector). Thus, the minimum value of Residency Time is one. The higher the value of Residency Time, a higher level of stability is achieved at the cost of fewer transitions recorded within the Total Time (length of the input data vector )
  • Double Well Energy Asymmetry is derived from logs (sonic, density, gamma,..) It controls the separation between different zones of interest, in the energy spectral density space;
  • Decoherence Relaxation Time represent the noise that is required to be added " to the system to compensate for losses due to environmental decoherence
  • Maximum Energy Ratio - This is used in the quantum resonance interferometry engine to check if the system energy stays within reasonable bounds of the quartic potential. It is not critical for the performance of the engine and a default value can be determined during the optimization process;
  • Regulator Threshold is used to establish a confidence measure on resonance events
  • Initial Position - This is the initial state of the system (x,y,z) in the Bloch Rate equations of motion given by Eqns. 10, 11, and 12. This parameter has to lie within the bounds of the functional region of the quartic function. Its influence is minimal and only serves as the starting point of the iteration. A valid initial value of the Initial Position can be calculated from the other parameters defining the quartic function.
  • Convergence Confidence This is the number of consecutive iterations wherein the TR value must stay within an Epsilon bound on the mean of TR in order to conclude convergence.
  • the value of this parameter should be a fraction of the maximum convergence iterations (the default value of 5% can be used). Higher values will increase the level of confidence with the cost of increased execution time;
  • Convergence Epsilon - is the width of the interval around a mean value within which the TR values are allowed to fluctuate for meeting the Convergence criterion.
  • Convergence is concluded if the difference between the maximum and the minimum values of arithmetic mean of TR is smaller than Convergence Epsilon for the pre-specified number of consecutive iterations.
  • the convergence confidence level is determined by the combination of the values of this parameter and the Convergence Confidence setting.
  • TR threshold atl920 is established during the training phase after the parameters of the quantum resonance interferometry engine have been set.
  • a training phase can be performed in an iterative mode with two processes (quantum resonance interferometry engine optimization and TR threshold determination) being repeated a number of times. For example, if one fails to establish a TR threshold with the initial set of optimized parameters, the quantum resonance interferometry engine parameters can be adjusted by varying the SNR gain. Then, the TR threshold determination process is re-attempted.
  • TR threshold a number of well control voxel data samples are needed. These data samples fall in two categories: termed as “Background” - i.e., no attribute of interest (i.e., attribute above a cutoff) is present; and, termed as “Signal” - a fixed value of well control voxel data with attributes of interest present above the cut-off value, also referred to as "Signal Above LOD”.
  • TR threshold can be established using as few as 1 well control voxel data.
  • Each data sample generates a corresponding TR value when processed by the quantum resonance interferometry engine.
  • TRB and TRS for the Background and Signal respectively.
  • the quantum resonance interferometry engine is optimized correctly only if these two sets of TR values are statistically different.
  • the standard deviation and the mean of each set of TR values are computed and assigned: oB, uB, oS, uS for Background and Signal respectively.
  • the critical difference parameter is then used to evaluate the difference between the two TR data sets.
  • nS and nB are the number of samples in the signal and background datasets respectively.
  • the CR value must be outside the interval [-1.96, 1.96]. There may be two reasons for the value of CR not falling within the interval. The data samples have not been selected properly (Signal is present in the Background samples or not enough is present in the Signal samples). This can be fixed by revisiting the sets of selected samples.
  • thresholdTR ( ⁇ + oS)/2.
  • the threshold so calculated is used in the operation mode to decide convergence.
  • Total Time - quantum resonance interferometry engine processes normalized excitation cascade as a 1 -D array time series of prospect voxel data vector.
  • the Total Time parameter determines the incremental time difference between each element of the Input Data
  • the first data element is assigned a time value of zero and the last element has a time value equal to: (PointCount - 1) * TotalTime / PointCount. This parameter establish scales at which transitions are detected vis-a-vis scale at which SNR is increased by quantum resonance interferometry engine.
  • NMR-QSR well design double parameters include energy asymmetry related to the energy spectral density of zone separation in well logs.
  • the well separation is used to modulate the desired separation between zones of interest (e.g., oil and water).
  • Barrier height is used to establish desired confidence. While the number of iterations are related to the inherent noise and uncertainty in the data.
  • Figure 20 shows an example workflow 2000 for producing output with various processes discussed herein.
  • Workflow 2000 uses processes described in this disclosureto produce output for sizing hydrocarbon reservoir and generating drilling targets for exploration and production.
  • Results of a conventional seismic processing workflow sequence (using e.g, standard off the shelf software geophysical packages to produce pre-stack time or depth migrated gathers) is obtained at 220.
  • Seismic horizons are generated to demarcate formation of interest using best-practice methods of horizon picking.
  • Spectral decompositions are performed on seismic data obtained at 220 to partition a seismic volume into low frequency, prime amplitude and high frequency partitioned volumes as in 560.
  • a Porosity-Cube is produced at 17160, using process 17100 with either high frequency or prime-amplitude partitions as an input.
  • a Lithology-Cube is produced at 17260 using process 17200 using either high frequency or prime-amplitude partitions as an input.
  • a Fluid Volume is produced as in 17360, using process 17300 with either high frequency or prime-amplitude partitions as an input,.
  • Processes 17100, 17200 and 17300 use flows in Figure 9A through 9C to produce seismic data noise at 9170, and sinusoidal periodic modulated synthetic noise 842 is generated from random number generators .
  • the seismic data noise and synthetic noise are combined to drive a quantum resonance interferometry engine used to produce Porosity-Cube, Lithology-Cube and Fluid Volume by analyzing a sequence of prospect voxel data derived from seismic partitioned volumes.
  • Results of flow 17100, 17200, and 17300 are combined as in 2050 to identify and highlight prospect voxel data within result volumes to determine voxels with porosity values above a prespecified cut-off, Vshale values that meet criteria of interest and voxels that show presence of fluid of interest. Voxels meeting all the three criteria are aggregated to assemble a subsurface geobody. The size, location, and orientation of the geobody in the seismic volume defines the properties of potential hydrocarbon reservoir. As the size of individual prospect voxel data is known, the size of the reservoir geobody is computed. Results from 2050 are used to make hydrocarbon exploration, drilling and production decisions. In some implementations, drilling, exploration and production decisions using a subset (e.g, just one or two) of Porosity-Cube, Lithology-Cube or Fluid Volume Results.
  • a subset e.g, just one or two
  • Implementations of the subject matter described in this specification can be implemented as one or more computer programs, i.e., one or more modules of computer program instructions, encoded on a computer storage medium for execution by, or to control the operation of, data processing apparatus.
  • the program instructions can be encoded on an artificially generated propagated signal, e.g., a machine- generated electrical, optical, or electromagnetic signal that is generated to encode information for transmission to suitable receiver apparatus for execution by a data processing apparatus.
  • a computer storage medium can be, or be included in, a computer-readable storage device, a computer-readable storage substrate, a random or serial access memory array or device, or a combination of one or more of them.
  • a computer storage medium is not a propagated signal, while a computer storage medium is not a propagated signal,.
  • a computer storage medium can be a source or destination of computer program instructions encoded in an artificially generated propagated signal.
  • the computer storage medium can also be, or be included in, one or more separate physical components or media (e.g., multiple CDs, disks, or other storage devices).
  • the term "data processing apparatus” encompasses all kinds of apparatus, devices, and machines for processing data, including by way of example a programmable processor, a computer, a system on a chip, or multiple ones, or combinations, of the foregoing.
  • the apparatus can include special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application specific integrated circuit).
  • the apparatus can also include, in addition to hardware, code that creates an execution environment for the computer program in question, e.g., code that constitutes processor firmware, a protocol stack, a database management system, an operating system, a cross-platform runtime environment, a virtual machine, or a combination of one or more of them.
  • a computer program (also known as a program, software, software application, script, or code) can be written in any form of programming language, including compiled or interpreted languages, declarative or procedural languages, and it can be deployed in any form, including as a standalone program or as a module, component, subroutine, object, or other unit suitable for use in a computing environment.
  • a computer program may, but need not, correspond to a file in a file system.
  • a program can be stored in a portion of a file that holds other programs or data (e.g., one or more scripts stored in a markup language document), in a single file dedicated to the program in question, or in multiple coordinated files (e.g., files that store one or more modules, sub programs, or portions of code).
  • a computer program can be deployed to be executed on one computer or on multiple computers that are located at one site or distributed across multiple sites and interconnected by a communication network.
  • the processes and logic flows described in this specification can be performed by one or more programmable processors executing one or more computer programs to perform actions by operating on input data and generating output.
  • the processes and logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application specific integrated circuit).
  • processors suitable for the execution of a computer program include, by way of example, both general and special purpose microprocessors, and any one or more processors of any kind of digital computer.
  • a processor will receive instructions and data from a read only memory or a random access memory or both.
  • the essential elements of a computer are a processor for performing actions in accordance with instructions and one or more memory devices for storing instructions and data.
  • a computer will also include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto optical disks, or optical disks.
  • mass storage devices for storing data, e.g., magnetic, magneto optical disks, or optical disks.
  • a computer need not have such devices.
  • a computer can be embedded in another device, e.g., a mobile telephone, a personal digital assistant (PDA), a mobile audio or video player, a game console, a Global Positioning System (GPS) receiver, or a portable storage device (e.g., a universal serial bus (USB) flash drive), to name just a few.
  • PDA personal digital assistant
  • GPS Global Positioning System
  • USB universal serial bus
  • Devices suitable for storing computer program instructions and data include all forms of nonvolatile memory, media and memory devices, including by way of example
  • semiconductor memory devices e.g., EPROM, EEPROM, and flash memory devices
  • magnetic disks e.g., internal hard disks or removable disks; magneto optical disks; and CD ROM and DVD-ROM disks.
  • the processor and the memory can be supplemented by, or incorporated in, special purpose logic circuitry.
  • implementations of the subject matter described in this specification can be implemented on a computer having a display device, e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor, for displaying information to the user and a keyboard and a pointing device, e.g., a mouse or a trackball, by which the user can provide input to the computer.
  • a display device e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor
  • keyboard and a pointing device e.g., a mouse or a trackball
  • Other kinds of devices can be used to provide for interaction with a user as well; for example, feedback provided to the user can be any form of sensory feedback, e.g., visual feedback, auditory feedback, or tactile feedback; and input from the user can be received in any form, including acoustic, speech, or tactile input.
  • a computer can interact with a user by sending documents to and receiving documents from a device that is used
  • Implementations of the subject matter described in this specification can be implemented in a computing system that includes a back end component, e.g., as a data server, or that includes a middleware component, e.g., an application server, or that includes a front end component, e.g., a client computer having a graphical user interface or a Web browser through which a user can interact with an implementation of the subject matter described in this specification, or any combination of one or more such back end, middleware, or front end components.
  • the components of the system can be interconnected by any form or medium of digital data communication, e.g., a communication network.
  • Examples of communication networks include a local area network (“LAN”) and a wide area network (“WAN”), an inter-network (e.g., the Internet), peer-to-peer networks (e.g., ad hoc peer-to-peer networks), wireless networks, mobile phone networks etc..
  • LAN local area network
  • WAN wide area network
  • inter-network e.g., the Internet
  • peer-to-peer networks e.g., ad hoc peer-to-peer networks
  • wireless networks e.g., mobile phone networks etc.
  • the computing system can include clients and servers.
  • a client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
  • a server transmits data (e.g., an HTML page) to a client device (e.g. , for purposes of displaying data to and receiving user input from a user interacting with the client device).
  • client device e.g., for purposes of displaying data to and receiving user input from a user interacting with the client device.
  • Data generated at the client device e.g., a result of the user interaction

Landscapes

  • Engineering & Computer Science (AREA)
  • Remote Sensing (AREA)
  • Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Acoustics & Sound (AREA)
  • Environmental & Geological Engineering (AREA)
  • Geology (AREA)
  • General Life Sciences & Earth Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Geophysics (AREA)
  • Geophysics And Detection Of Objects (AREA)

Abstract

Application of nonlinear resonance interferometry is introduced as a new geophysical approach to improve predictability in characterization of subsurface porosity, rock and fluid properties. In contrast to reflection methods that remove random information noise, nonlinear resonance interferometry exploits the full seismic acquisition spectrum to assess how low frequency and high-frequency noise is differentially and directly modulated by varying levels of porosity and hydrocarbon content in the lithologies of interest. In some examples, systems and techniques implement novel computational interactions between acquired seismic wavefield attributes and a nonlinear system in software to amplify distortions in seismic noise and exploits injection of synthetic noise, in software format, to detect hydrocarbon traps and lithology changes at spatial scales below seismic resolution, thereby increasing the information value of low-resolution data. The techniques are broadly applicable to de-risking conventional clastic and carbonate reservoirs and non-conventional shale gas resource prospects.

Description

Active Noise Injection Computations for Improved Predictability in Oil and Gas Reservoir Discovery and Characterization
BACKGROUND
[0001] This application claims the benefit of U.S. Provisional Application No.
61/311 ,277, titled "Active Noise Injection Computations For Improved Predictability In Reservoir Characterization," which was filed on March 5, 2010 and which is incorporated herein by reference in its entirety
BACKGROUND
[0002] In oil and gas exploration, seismic surveys are used to estimate features of interest of subsurface geology. Seismic surveys use controlled seismic energy, such as produced by specialized air guns or seismic vibrators. A receiver senses seismic energy, typically in the form of a wave, reflected by subsurface features. The subsurface features are detected by analyzing the time it takes for reflected seismic waves to travel through the subsurface matter of varying densities. 3-D seismic also uses seismic energy to produce a 3-dimiensional map of subsurface formations. Traditional techniques of analyzing seismic data attempts to filter out noise to identify a signal of interest. However, in the process of filtering out noise, important information of interest can be lost, resulting in a seismic map from which various features may be difficult to distinguish. Once a prospect has been identified, an exploration well is drilled in an attempt to conclusively determine the presence or absence of oil or gas. However, an exploratory well can be very expensive especially for off-shore wells and is subject to the risk that the well will be unproductive.
SUMMARY
[0003] Systems and techniques are disclosed for improving signal-to-noise in onshore and offshore seismic and electromagnetic acquisitions, for improving microseismic techniques for hydraulic fracture monitoring and optimization, and for interferometric acquisition for directing hydrocarbon imaging using conventional 3D vibrators. Seismic data is analyzed using a non-linear stochastic quantum energy source to create resonances through the interaction between virtual vibration and the seismic data. Changes in resonance yields information of interest from the seismic data not obtainable by noise filtering. The signature of such changes in resonance can then be determined and calibrated as porosity, fluid, or lithology. Reservoir characterization and drilling decisions can then be made on the resulting volumes of reservoir property. Such systems and techniques can be used to de-risk new drilling locations and to redevelop under-performing or abandoned wells by locating nearby pay zones that may have just been missed.
[0004] More than 60% of world's oil and 40% of world's gas reserves are in carbonates. In some examples, the systems and techniques disclosed herein can be used to de-risk and characterize complex stratigraphic carbonates, their stacking, continuity, fracture density, spacing and porosity. Porosity prediction is at the heart of discriminating potentially productive carbonate bodies. Carbonate reservoirs continue to be difficult to characterize using 3D seismic due to their greater heterogeneity from rapid vertical and lateral facies variation, lower seismic resolution due to higher velocities, and inherent inability to directly image fracturing. Basin geology models provide limited insight to positioning of individual wells and offsets, as formation properties change unpredictably. Because of the broad- spectrum of diagenesis (chemical, physical, or biological change undergone by a sediment in its initial deposition) that affects carbonate rocks, the final porosity in these carbonates may or may not be related to the depositional environment. Also, unlike other lithologies, the original primary porosity in carbonates may be totally destroyed during diagenesis and significant new secondary porosity may be created. So to predict well productivity, certain patterns of natural fracturing and high continuous porosities are identified from 3D seismic data.
[0005] On-shore and deep-water clastic reservoirs are among the world's largest, most explored, and most productive hydrocarbon plays. These include a variety of turbidite sand- body geometries such as channels, lobes, sheets and levees in complex down-slope settings. Post-depositional stresses modify primary sedimentary structures, changing pore size distribution and permeability characteristics which proves challenging to interpretation of pay and saturation distribution in otherwise sand-prone reservoirs. Exploration success, and subsequent appraisal and development of these highly productive reservoirs depends upon accurately mapping the interplay of sediment dispersal within reservoir-scale or basin-scale geometry to delineate source, seal, and reservoir geologies. Conventional de-risking has relied on acquiring large offsets and higher frequency data While this has advanced structural interpretation to a degree, success has been spotty due to lack of fluid imaging capability. Systems and techniques are disclosed that exploit noise within seismic data and increase resolution to jointly assess sand-stacking along with in-place fluid saturation. Such techniques can assist in derisking and finding net pay. Such techniques can also be used to explain the distribution of producers and dry-holes in existing in 3D seismic surveys.
[0006] Shale gas resource development is rapidly becoming a global trend in onshore exploration. Wells produce from low permeability shale formations can also be the source rock for the oil and natural gas. As the larger hydrocarbon volumes are restricted to fracture porosity within the shale, or within micropores, or adsorbed onto the minerals and organic matter within the shale, detecting subtle changes in lithology produces dramatic changes in predicting production outcomes and economics even in closely space wells. However, such lithology changes register only as weak changes in a conventional 3-D seismic signal because 3-D seismic has been designed to primarily image large impedance contrasts across lithologies. Because the changes in impedance contrast within shale formations are very subtle, geophysicists struggle to add value by using the signal measured by conventional seismic in shale gas reservoirs. The systems and techniques disclosed herein can detect such subtle changes in lithology. Conventional data can be analyzed according to the disclosed systems and techniques to identify such subtle changes. The systems and techniques disclosed herein can also be used to identify brittleness in shale— the ability to fracture the reservoir.
[0007] One aspect of the subject matter described in this specification can be embodied in methods that include, a method includes voxelizing seismic data for a geological subsurface formation of interest into multiple voxels having respective locations in the formation of interest; and determining whether a voxel in the multiple voxels includes an attribute; and outputting whether the voxel includes the attribute to an attribute volume based on the ' location; and performing the determining and the outputting for at least some other of the multiple voxels.
[0008] In some examples, the method can further include performing the steps listed above for the seismic data for another formation of interest. The determining whether a voxel in the multiple voxels includes an attribute can include determining whether the voxel in the multiple voxels includes a porosity attribute; and the method further can include assembling the at least some of the other of the mulitple voxels into a porosity cube. The determining whether a voxel in the multiple voxels includes an attribute can also include determining whether the voxel in the multiple voxels includes a Uthology attribute; and the method further can include assembling the at least some of the other of the multiple voxels into a lithology cube. Additional, the determining whether a voxel in the multiple voxels includes an attribute can also include determining whether the voxel in the multiple voxels includes at least one of gas, oil, or water; and the method further can include assembling the at least some of the other of the multiple voxels into a liquid cube.
[0009] In some examples, the method can also include wherein the seismic data can include migrated prestack time gathers for the formation of interest. The determining whether a voxel in the multiple voxels includes an attribute can include determining whether the voxel in the multiple voxels includes a lithology attribute; repeating the determining and the outputting for a porosity attribute; and generating a drill-here map indicating a location for drilling based on the at least some of the other of the multiple voxels for the lithology attribute and based on the at least some of the other of the multiple voxels for the porosity attribute. Also, the voxelizing the seismic data further can include spectrally decomposing the seismic data into frequency volumes. L
[0010] In some examples, the method can further include spectrally decomposing the seismic data into a high frequency volume, a medium frequency volume, and a low frequency volume. The the seismic data can include three dimensional seismic data and the voxelizing can include generating, from the three dimension seismic data, multiple one dimension voxel vectors. The determining whether a voxel in the multiple voxels includes an attribute including: obtaining control data for that attribute based on information from one or more previously explored geological subsurface wells; and processing the one dimension voxel vector for the voxel using the control data and using quantum resonance interferometry to detect a resonance.
[001 1] Also, the method can further include characterizing a structural trap based on the attribute volume. The method can also include characterizing a stratigraphic trap based on the attribute volume. Additionally, the method can include upscaling the multiple voxels to a predetermined resolution.
[0012] In another aspect, a method can include obtaining a one-dimensional voxel vector for a voxel from multiple voxels associated with a formation of interest; obtaining spectral data generated from well-log data associated with an attribute of interest; coupling the spectral data with the one dimensional voxel vector to determine whether a resonance event occurs; when a resonance event occurs, producing an output indicating that the voxel has the attribute of interest; and when the resonance event does not occur, producing an output indicating that the voxel does not have the attribute of interest.
[0013] In some examples, the obtaining the one-dimensional voxel vector include normalizing amplitudes corresponding to the one dimensional voxel vector to fit within a range. And, the obtaining the spectral data generated from well-log data can include: obtaining seismic noise data; obtaining well control voxel data from the well log data; and combining the seismic noise data and the well control voxel data using a quantum mechanical model to produce the spectral data. Additionally the obtaining well control voxel data can include determining an energy spectral density voxel from the well-log data for the attribute using wavelets derived from sonic logs in sections of a well with the attribute of interest. Also, the combining the seismic noise data and the well control data can include combining the seismic noise data and with well control voxel data for a voxel associated with the well- log data that exhibits the attribute. The obtaining the well control voxel data can also include obtaining a well control voxel data of a voxel not exhibiting the attribute; and the obtaining the seismic noise data can include generating the seismic noise data based on the well control voxel data of the voxel not exhibiting the noise. In some example, the coupling the spectral data with the one dimensional voxel vector to determine whether a resonance event occurs can include using a nuclear magnetic resonance ("NMR") master rate equation to generate quantum stochastic resonance based on the one-dimensional voxel vector, synthetic noise, and the spectral data.
[0014] In another aspect, a method can include obtaining voxel data for a voxel from voxelized seismic data for a geological subsurface formation; performing a first nonlinear coupling of the voxel data with spectral data to generate a first resonance, wherein the coupling is driven by noise having an intensity within a first cutoff band; in response to generating the first resonance, adjusting the cutoffband to a second cutoffband, different than the first cutoff band; performing a second non- linear coupling of the voxel data with the spectral data associated with an attribute of the subsurface formation to generate a second resonance, wherein the second coupling is driven by noise having an intensity within a second cutoffband; in response to generating the second resonance, generating an indication that the attribute exists in the voxel; and in response to the second coupling not producing a second resonance producing an indication that the attribute does not exists in the voxel.
[0015] In some examples, the first resonance can include a first quantum stochastic resonance. Also, the method can also include assembling the voxel data in an attribute volume with the indication that that the attribute exists in the voxel. The voxelized seismic data can also include multiple voxels including the voxel; and the method can further include determining the first cutoffband based on a percentage of an average intensity of the multiple voxels. The method can also include prior to the first coupling, performing a third coupling of the voxel data with the spectral data associated with an attribute of the subsurface formation to generate a third resonance, wherein the third coupling is driven by noise having an intensity within a third, different cutoff band; and in response to generating the first resonance, adjusting the third cutoff band to the first cutoff band.
[0016] In another aspect, a method includes obtaining first prospect voxel data for a voxel from voxelized seismic data for a geological subsurface formation on a prospect;
determining an attribute iteration range including values for the attribute including at least a first value and a second value; determining control well voxel data from well-log data for an existing well, the control well voxel data including spectral energies for a listing of values of the attribute found in the existing well, including a first spectral energy for the first value and a second spectral energy for the second value; coupling the prospect voxel data with spectral data having parameters set according to the first spectral energy to produce a first resonance; and in response to the first resonance, generating an output indicating that the voxel includes the attribute at at least the first value.
[0017] In some examples, the method further includes- coupling the prospect voxel data with spectral data having parameters set according to the second spectral energy to produce a second resonance; and in response to the second resonance, generating an output indicating that the voxel includes the attribute at at least the second value. Also, the control well voxel data can include a third spectral data corresponding to a third value for the attribute; and the method can further include coupling the prospect voxel data with spectral data having parameters set according to the third spectral energy to produce a third resonance; and in response to the third resonance not being produced, generating an output indicating that the voxel does not include the attribute at the third value but includes the attribute at the first and second value. Also, the attribute can include porosity and the first value can include a first percentage of porosity and the second value can include a third percentage of porosity greater that the first percentage.
[0018] Other embodiments of these aspects include corresponding systems, apparatus, and computer programs, configured to perform the actions of the methods, encoded on computer storage devices.
DESCRIPTION OF DRAWINGS
[0019] Figure 1 shows an example process for reservoir characterization.
[0020] Figure 2 shows an example of a graphical depiction of elements involved in process for reservoir characterization. [0021] Figure 3 shows applications of the systems and techniques for reservoir characterization.
[0022] Figure 4 shows examples of subsurface reservoir traps.
[0023] Figure 5 shows an example of a flow chart for processing seismic data for a prospect prior to the seismic data being voxelized.
[0024] Figure 6 enumerates examples of some subsurface volumes and their taxonomy for a prospect.
[0025] Figure 7 shows an example of a process for voxelizing seismic data for a formation of interest.
[0026] Figures 8A-8B show an example of a work flow for analyzing voxelized data.
[0027] Figures 9A-9D show an example of a process for determining seismic noise data .
[0028] Figure 10 describes a process for upscaling or upconverting analysis of attributes of interest.
[0029] Figure 11 shows a process for establishing monotonicity of a resonance event.
[0030] Figures 12A through 12B describe a process for generating and normalizing excitation cascade data.
[0031] Figures 13A-13D show examples of how attribute-specific well control voxel data is developed.
[0032] Figure 14 describes an example of a process of implementing an interferometry engine and a quantum resonance interferometry processor.
[0033] Figure 15 shows an example of a double well function.
[0034] Figure 16 shows an example of work flow using a quantum resonance interferometry engine.
[003 ] Figures 17A -17D show examples of a processes for computing attribute data from seismic data.
[0036] Figures 18A - 18B show configuration of a quantum resonance interferometry engine in a training mode and an operations mode.
[0037] Figure 19 shows a method for calibration of quantum resonance interferometry engine Parameters.
[0038] Figure 20 shows an example workflow for producing output with various processes discussed herein. DETAILED DESCRIPTION
[0039] Figure 1 shows an example process 100 for reservoir characterization (also known as lateral subsurface prediction of reservoirs). Process 100 can be used, in some examples, to build a computer model of a reservoir beneath the surface of earth that incorporates all the characteristics of the reservoir that are pertinent to its ability to store hydrocarbons and also to produce them. The process 100 can be used to discover, delineate, and size hydrocarbon (oil and natural gas) reservoirs to produce, among other things, a drill map for onshore or offshore wells using seismic data. At 110, prospect data and exploration criteria are obtained. Exploration criteria can include targeted exploration objectives such as criteria for a type or types of formations to be searched for in a prospect— a location being analyzed for a deposit. A prospect is defined as an area covering a potential subsurface trap believed to contain hydrocarbons, geological factors that have to be present for a prospect to produce oil and gas include: presence of a source rock (organic rich rock that has been subjected to high pressure and temperature over an extended period of time to form hydrocarbon), a structural, stratigraphic or combination trap to hold the hydrocarbons, impermeable seal or cap rock over the hydrocarbon trap in order to prevent hydrocarbons migrating or escaping to the surface, and porous reservoir rock that collects oil within its pores and that is permeable so that the hydrocarbons will flow to surface during production.
[0040] For example, exploration criteria can include a target formation with a predetermined porosity, fluid type, and lithology for a predetermined type of subsurface material. The porosity of a porous source rock or sediment describes the fraction of void space in the material, where the void may contain fluid such as water, oil or gas ,and is defined by the ratio:
[0041] Vr
[0042] where Vv is the volume of void-space (such as fluids) and V-χ is the total or bulk volume of material, including the solid and void components. Porosity is a complex function of many factors such as grain size and overlying sediments.
[0043] Figure 2 shows an example of a graphical depiction 200 of elements involved in the process 100. For example, Figure 2 shows examples of prospect data at 210A, 210B, and 2 IOC. Prospect data can include seismic and other geophysical data, rock physics data, and well log data for the prospect or well log data from an analogous formation on another prospect. Geophysical data can include for example seismic data 210A obtained for an area of interest, such as for a prospect area 211. Obtaining geophysical data can include, for example, receiving the geophysical data already acquired from a prospect area, at a computing device for further processing and analysis. In some examples, obtaining the data can include receiving the data, actively requesting the geophysical data or acquiring the geophysical data from the prospect area 211.
[0044] Seismic data 210A can include already processed, migrated pre-stack time gathers (denoted as a Pre-stack Time Migration (PSTM) volume); migrated pre-stack depth gathers (denoted as a Pre-stack Depth Migration (PSDM) volume); or raw seismic gathers. The gathers can be stacked into a seismic volume, such as seismic volume 220, in the format of a PSTM stack or PSDM stack in an SEGY format. The seismic volume can be derived from 2- D seismic acquisition, 3-D seismic acquisition process, or multicomponent seismic process over a seismic acquisition grid 213.
[0045] Seismic data can also include a fold map 214 which includes a multiplicity of recorded seismic reflection data. The fold map 214 can be used to ensure high acquisition signal-to-noise and sufficient reflected energy for subsequent computations. A 3-D seismic survey fold map may be calculated using the relationship:
_ Ustft-lSuif8.ce area of Patch QJ^ _ g x (Max-muin useableOffstf) 2
F0W3D 4 xSo ce Line Spacing xReceiver Line Spacing 4 x Source Line x Receiver Line where the useful surface area of a patch (i.e., an areas being analyzed with seismic energy) and a source line and a receiver line spacing are obtained from the seismic acquisition grid 213. Method of 100 applies to standard compression wave 2-D and 3-D seismic acquisition, 4-D time lapse seismic where measurements are repeated over time, 4-C converted wave multi-component seismic that uses a compression wave (p-wave) source and both compression wave and shear wave(s-wave) receivers, and 9C full wave multicomponent using both compression wave and shear wave source and receivers, (converted wave with compression source and shear or 9C multicomponent seismic). Source line spacing refers to the distance between successive rows of way-points at which acoustic excitations are sent into the ground. Various modahties of seismic acquisition are familiar to one skilled in the art. In practice, acquisition processes using a seismic acquisition grid 213 record a large square or rectangular patch, depending on the area of interest, such as a mineral lease owned by a prospect owner. Useful data obtained at an area of interest can be offset by several geophysical factors. For example, geophysical factors can limit a maximum useful radius for geophone coverage.
[0046] By moving a patch and recording more salvos using simultaneous discharge of excitation vibrations from multiple vibrator trucks or air guns as an example, into the group of source points, one can accumulate overlapping subsurface coverage and build statistical repetition over each subsurface reflecting area referred to as a "bin". The quality of the subsurface image can be related to the statistical diversity of the information recorded for each bin of sub-surface coverage. For example a bin represents the smallest area of a 3-D survey that contains the entire survey statistics. The higher the number of observations obtained, that contain unique measurements of the acoustic reflections from a subsurface, the better the reconstruction is of the subsurface geological configuration that caused those observations. Terms bin and CMP (common midpoint) bin are used interchangeably. It is a small rectangular area with dimensions ½ source line spacing by ½ receiver line spacing comprising of all mid-points that lie inside the area are assumed to belong to a common midpoint. All seismic traces that lie inside the bin will contribute to the fold of that bin and will be CMP stacked.
[0047] Techniques for producing conventional PSTM and PSDM volumes, used as an input to process 100 are well known in the art. For example, Oz Yilmaz's book (and references therein), Seismic Data Analysis - Processing, Inversion, and Interpretation of Seismic Data, Volume I and II, pages 1 - 2024, 2001, published by Society of Exploration Geophysics, provides a discussion of methods for producing PSTM and PSDM seismic volumes all the way from seismic data acquisition. In some examples, the seismic data is obtained in a preprocessed format where the seismic data is correctly imaged with the seismic signals focused, where the unwanted energy, such as seismic multiples, is removed, and where true amplitude seismic processing has preserved the amplitude information of seismic reflection events.
[0048] Prospect data also includes well log data 2 IOC. Well logs provide physical rock properties for an identified petrophysical zone. These are available from wells that have already been drilled on the prospect area 211 under consideration or from wells drilled into similar source rock at another analogous prospect area. Well logs include sonic logs, bulk density logs, gamma ray logs, neutron porosity logs, and resistivity logs. A sonic well log, such as dipole sonic log, can also include sampling of both compression (p-wave) velocity and shear (s-wave) velocity within a well bore. Basic Well Log Analysis for Geologists by George Asquith and Charles Gibson, 1982, pages 1 - 215 and references therein, published by American Association of Petroleum Geologists, Tulsa, Oklahoma provide an introduction and discussion of different wells logs and their relationships.
[0049] Well log data can be used to calibrate and tie actual rock and fluid properties for a formation of interest at a known well, to seismic data for the known well. Such well log data can also be used to interpret attributes found in the seismic data for the known well. As discussed in more detail below, well log data 210C for the seismic volume can be used to predict rock and fluid properties in a seismic volume using seismic data from the seismic volume of a prospective well. The well log data 2 IOC can been edited and corrected for issues such as noise bursts, cycle skips and fluid replacements or can be obtained in such a format.
[0050] Prospect data can also include rock physics data 210B includes information regarding type of geological source rock, such as sandstone, limestone, dolomite, and shale, being targeted for exploration, such as is identified by the exploration criteria. Term. Rock physics relates the geological properties (e.g. porosity, lithology, saturation) of a rock at certain physical conditions (e.g. pressure, temperature) with their corresponding elastic and seismic properties (e.g. elastic modulus, velocity, p-wave impedance, s-wave impedance). Using rock physics modeling, one skilled in the art can predict the elastic (seismic) properties from the geology, or using rock physics inversion can predict geology from elastic (seismic) observations. Also, derived attributes such as brittleness, defined as the ratio of compressive strength to the tensile strength for the rock is derived from Poisson's Ratio and Young's modulus, measured based on rock core samples in the lab. Brittleness can be derived using Vp V s ratio where Vp, Vs denotes p-wave and s-wave velocities respectively or
approximated using p-wave imped ance/s-wave impedance. Once source rock types (i.e., rock that is a potential source of hydrocarbons) are identified for a prospect, rock physics tables can be obtained for basic lithological and geomechanical properties of the rock types and their granularity combinations. The source rock types can be identified by well logs 2 IOC for existing wells including seismic data obtained for those existing wells. Rock physics data 210B includes properties enumerated in terms of sonic compression-wave velocity, shear- wave velocity, brittleness, Young's modulus, and relationships for sonic velocity changes across different rock facies.
[0051] At 115, a subsurface formation of interest of the subsurface volume 220 is identified for further analysis. The formation of interest is located in a prospect located in the prospect area 211. The subsurface formation can include for example layer 225 shown in Figure 2. A prospect may have multiple separated hydrocarbon bearing pay-zones present in the subsurface. For example a Permian Basin well-bore may have hydrocarbon pay-zones within the Spraberry, Dean, Clearfork, Wolfcamp, Atoka, Strawn, Devonian, Fusselman, Ellenberger formations, all of which need to be individually selected and independently analyzed as in 115. Analysis 100 may be used to analyze one or more or all of the formations between surface and maximum depth covered by the seismic data. Conventional interpretation is used to identify seismic horizons or subsurface reflecting surfaces that separate the multiple subsurface formations. A horizon is a reflecting surface separated by different layers in depositional environments and characterized by different seismic reflection properties. These are separated by curves in 2-D data and surfaces or facies in 3D data. Facies is defined as a distinctive rock unit that forms under certain conditions of
sedimentation, reflecting a particular geological process or sedimentary depositional environment. From an image processing perspective, seismic horizons are edges on the seismic image serving to delineate formations such as structural and stratigraphic features (such as faults) and patterns. Oz Yilmaz's book referred above provides a detailed treatment for conventional methods to determine seismic horizons in seismic volumes.
[0052] At 118, the seismic data in the seismic volume 220 for the prospect is processed for voxelization. Voxelization is defined as the computational process of decomposing seismic volume into 3-D volume cells which are then analyzed for rock and fluid properties. The term voxelized data refers to seismic volumes that have been decomposed into smaller 3- D volume cells and to data relating to attributes for those 3-D volume cells. As discussed in connection with Figure 3, the seismic data is spectrally decomposed into separate spectral decompositions. Spectrally decomposing the voxelized data can help in determining frequency bounds for prospect voxel data (also referred to as prospect voxels or as voxelized data) in step 120. The term prospect voxel refers to a singular prospect voxel data while prospect voxels refers to plurality of prospect voxel data. The seismic data can be broken up into frequency separated volumes such as separated by high (e.g., between 80Hz and 100Hz for onshore data), low (e.g., between 3 Hz and lOHz), and prime amplitude (e.g., between 12Hz and 70Hz) frequencies.
[0053] At 120 the seismic data 210A, for the seismic volume 220 is voxelized into multiple three-dimensional volume cells, referred to as voxels, for further analysis, as shown by the multiple voxels 235 in Figure 2. In some examples, voxelizing the seismic data at 120 includes determining a size and orientation of the voxels which depends on the resolution of the acquired seismic data, geological understanding of the depositional processes that led to development of hydrocarbon reservoirs in the prospect area within a field or basin and the formation of potential traps in the subsurface lithological formation of interest on the prospect. Dimensions can be specified as x (feet or meters) by y (feet or meters) by z milliseconds for time-based seismic volumes, or x (feet or meters) by.y (feet or meters) by z (feet or meters) for depth-based seismic volumes. In some examples, the vertical z dimension for a voxel is set to the seismic sample acquisition sampling rate (e.g., 0.5 ms, 1ms, 2ms, or 4ms). The higher the sampling rate the higher the resolution of the voxels and ultimately of the result of analyzing the voxels. Also, the areal x and y dimensions are limited by the seismic survey design or the "bin size" (x by dimension) of the PSTM/PSDM gathers. In many seismic processing algorithms that can be used, the seismic data is processed before being obtained at 110, the bin size is established early in the process and final seismic data is bin-size limited. If bin size is coarse in the seismic data, then re-binning and reprocessing is performed to generate smaller bin sizes. If reprocessing is not possible then each voxel is sized to (2 times x-dimension of the bin) by (2 time ^-dimension of the bin) by (sampling rate or processing depth resolution). If re-binning and reprocessing is possible for the seismic data, then voxels sizes can be set to as low as (0.5 times the x-dimension of the bin) by (0.5 times the ^-dimension of bin) by (sample rate or depth resolution). The dimension of the voxels determines the final attribute resolution that is produced by the analyzing step 125.
[0054] At 125, the voxelized data is analyzed. The voxelized data is analyzed, voxel by voxel, to determine the existence of attributes as well as the areal and vertical extent of attributes of interest in the seismic data using, for example, quantum resonance
interferometry. The attribute of interest can be defined by the exploration criteria. Examples of attributes of interest are shown at 122, and include: rock properties, porosity, lithology, geomechanical properties, brittleness, fluid presence, type of fluid (oil, gas, brine, and/or water), and anisotropy. For example, for each voxel, a percentage of porosity can be identified, lithology characteristics can be identified, and the existence of fluids can be identified such as water, gas, or oil. The result of the analysis can include multiple attributes for each prospect voxel. For a voxel, the seismic data can be analyzed multiple time for multiple attributes of interest. Examples of how the voxelized data can be analyzed is provided in the discussion in connection with Figures 1 to 20.
[0055] Once the Prospect voxel s in the formation of interest are analyzed an attribute volume can be produced at 130. The analyzed voxels are combined according to their original layout in the seismic volume 220 to produce an attribute volume that can include rock and fluid properties that are combined to find areas of high net-pay or favorable combinations of porosity and fluid -type (oil and or gas). The analyzed voxels are assembled into one or more 3 -dimensional attribute volumes, graphically depicted in Figure 2 as a first attribute volume 275 for porosity, a second attribute volume 280 for fluid volume, and a third attribute volume for lithology. The first attribute volume for porosity shows areas of high porosity in the formation of interest. Each of the analyzed voxels in the first attribute volume 275 show a degree of porosity analyzed for the voxel. The second attribute volume 280 shows a distribution of fluid in the formation of interest. For example, the distribution of gas, oil, and water can be depicted in the second attribute volume 280. The third attribute volume 285 shows a distribution of rock properties such as brittleness of the rock.
[0056] At 140; a reservoir is delineated at 140. For example, the exploration criteria can identify cutoff values for which a desired reservoir is desired. For example, for porosity, a cutoff value can be defined above which the reservoir of interest is delineated. In other words, all of the voxels in the attribute having a value above the cutoff, can be delineated as part of the reservoir. In some examples, the cutoff value can be dynamically selected which it turn produces a delineated reservoir of different sizes.
[0057] In some implementations, data for the delineated reservoir of step 140 can be provided for further processing and output at 150. For example, the data for the delineated reservoir including the analyzed voxel data can be provided to another processing module, such as a flow model, for further processing. In some implementations, at step 140, it is determined whether all the formations of interest have been analyzed. If not, the process 100 is repeated for steps 120-140 for another formation of interest. At 150, if there are no further formations of interest, the process 100 continues on to step 160 where the data for the delineated reservoir is used to produce one or more drill maps. Figure 2 shows examples of a drill map— a 3-dimensional drill map 292 with an identification of a drill location and a drill trajectory and a 2-dimensionl map also with indicators of where to drill.
[0058] The outputs produced by Figure 1 can be used in various areas of oil and gas exploration and production. For example, Figure 3 shows applications of the systems and techniques disclosed in this disclosure. For example, the systems and techniques disclosed herein can be used, for example in oil and gas lease acquisition 330 to provide accurate subsurface mapping, to estimate hydrocarbon formation boundaries, to optimize prospect acreage, and to optimally extract hydrocarbon and drill wells. Also, the systems and techniques disclosed herein can be used in seismic survey design 333 by determining subsurface characteristics which provides insightful guidance on how to acquire additional 2D, 3D and multi-component seismic data or setup parameters to achieve high quality imaging. For example, reservoir properties using vintage seismic to drill exploration wells or offset wells can be analyzed, for example using existing legacy data, to develop new surveys and design quality seismic data and optimal frequency spectrum for imaging potential structural/stratigraphic traps, complex fracturing and nonconformal lithologies. [0059] Also, the systems and techniques disclosed herein can be used for exploration of new wells. For example, analysis can be performed on end-to-end, available geotechnical, geological, stratigraphy and geoseismic data to discover reservoirs, characterize formation boundaries, porosity and properties, develop net pay models, and recommend drilling targets and mitigate dry-hole risk. Also, well trajectories and well-bore deviations can be designed to maximize recovery from nonconformities and fractures. Also, the systems and techniques disclosed herein can be used for positioning development wells in producing or proven fields. For example, locations can be determined for placing offset wells to maximize recovery and minimize the number of wells required to cost-effectively drain hydrocarbon formations. And, the systems and techniques disclosed herein can be used for enhancing currently producing wells by designing perforations and injection laterals for maximal drainage from productive formations and multi-zone wells pays to extend reservoir life and recovery.
[0060] As in 320, the methods described in this invention in Figure 1 through 20, can be used to enable and enhance various upstream oil and gas during primary, secondary, enhanced hydrocarbon recovery and redevelopment of fields or reentry of old wells that have been capped or abandoned or extending depth of existing wells to discover new sources of hydrocarbon production and to find bypassed and missed pockets of hydrocarbon.
[0061] Figure 4 shows examples of subsurface reservoir traps that can be analyzed using the systems and techniques described herein in each of the stages of oil and gas exploration described above in connection with Figure 3. Traditionally, reservoir characterization has been a subjective process with the outcome highly dependent upon the experience and quality of the geophysical interpreter, quality of seismic processing and formation complexity. Also, the types of onshore and offshore formations are diverse and complex forcing analysts to differentiate themselves based on expertise in specific formation types (e.g., structural plays, stratigraphic plays, or combination traps such as salt domes), geological basins, and resource plays (e.g., Permian Basin, Bakken or Gulf of Mexico).
[0062] Structural Traps, formed by deformation of reservoir rock such as anticline or fault, such as Anticline, Normal Fault Trap, Baldheaded Anticline, Reverse Fault Trap, Faulted Anticline, Tilted Fault Block, Dome, Drag Faults on Thrust Fault, Fractured
Reservoir, Rollover Anticline on Growth Fault, Antithetic or Synthetic Faults on Rollover Anticline. Stratigraphic Traps formed by deposition of a reservoir such as reef or river channel or erosion of reservoir rock. Examples include Barrier Reef, Atoll, Pinnacle or Patch Reef, BioHerm, Angular Conformity, Shoestring-Sandstone Channel, Buttress or onlap sands, Oolite Shoals, Granite Wash, and Updip Pinchout of Sandstone. Combination Trap which are formed by both structural and stratigraphic elements. These include salt-domes (and overlying domes and faults), Salt Dome - Cap Rock, Salt-Dome-Flank Traps, Updip Facies Change, Compaction Anticline, Secondary or Tectonic Dolomite.
[0063] Examples of three of these formations are shown in Figure 4—a structural trap 410, and combination trap 420, and a stratigraphic trap 430. Each formation is shown with an exemplary graphical depiction of seismic data obtained for these formations: structural trap seismic data 411, combination trap seismic data 421, and stratigraphic trap 430. There are multiple types of oil and gas formations each having unique seismic data.
[0064] The seismic data for each of these formations is defined by the underlying rock, fluid, and lithology attributes such as porosity and fluid type. Similar attributes can be obtained post-drilling from well logging and coring data, which entails taking rock samples and performing detailed laboratory analysis.
[0065] The systems and techniques disclosed herein enable porosity determination, lithology characterization, and fluid characterization by specifying a fundamental alternate re-formulation of reservoir characterization process as weak signal detection problem. A quantum resonance interferometry engine is configured to harness nonlinear quantum stochastic resonance phenomenology to detect weak signals in seismic data as a disturbance to noise in a high noise and high interference environments.
[0066] Conventional geophysical processing removes noise; but systems and techniques disclosed herein exploit subtle variations in low (such as 3 tolOHz) & high f equency, (such as 80-119Hz) low dB noise amplitudes within seismic data to increase resolution. The changes in these low and high frequency amplitudes can be very small and the invention deploys a software-based quantum resonance Interferometry engine with noise injection computation (implemented, for example, as Vialogy's Virtual Vibe™ to separate hydrocarbon from non-hydrocarbon bearing lithology.
[0067] For example, oil and gas exploration objectives for the various structural formations, such as those shown in Figure 4, are formulated as weak signal processing problems that are addressed using a resonance interferometry. For the structural trap 410, the structural trap seismic data 411 is analyzed to detect structural traps and to detect an amount of oil saturation, such as oil saturations above 40%. Seismic data near a fault is analyzed to detect hydrocarbon bearing lithology with an oil-water contact boundary. This hydrocarbon lithology is treated as a signal of interest in the seismic data that is analyzed by a resonance interferometry engine with all other lithologies treated as noise. [0068] For the combination trap 420, the combination trap seismic data 431 is analyzed to detect porous and permeable tertiary sands that updip around a salt dome. Porosity above a certain percentage (e.g., above 10%) in a porosity channel is treated as the signal of interest in the combination trap seismic 411 by a resonance interferometry engine with all other facies treated as noise. For the stratigraphic trap 430, the stratigraphic trap seismic data 431 is analyzed to locate pinnacle reefs with carbonate. Porosity anomalies within stacked sandstone or shale beds are treated in a signal of interest in the stratigraphic trap seismic data that is analyzed by a resonance interferometry engine with all other lithologies treated as noise.
[0069] The ability to reformulate broad prospect and formation specific questions into specific weak signal processing objectives enables the systems and techniques disclosed herein to be applied to a broad range of oil and gas issues, including: types of formations, formation complexity, onshore & offshore wells, new fields (using analog) & existing fields, data modalities (active/passive - such as 3D/multi-component seismic; magneto-telluric; gravitational). Moreover, various oil and gas drilling or recovery challenges for a multitude of formations can be posed as a weak signal processing question and solved using the systems and techniques described herein.
[0070] Figure 5 shows an example of a flow chart 500 for processing seismic data for a prospect prior to the seismic data being voxelized for a formation of interest. At 503 seismic data is obtained for a formation of interest. The seismic data can be in the form of a pre-stack time or depth amplitude migrated volume. Migration is a geophysical process by which geophysical events (i.e., changes in acoustic reflection energy) are geometrically re-located in either space or time to a location the event occurred in the subsurface rather than a location where the event was recorded at the surface, thereby creating an accurate image of the subsurface. Migration operators are well known in the geophysics art and are applied in both time and space domains to produce time amplitudes and to produce depth amplitude volumes, respectively. A difference between the two is how well the migration operator represents the velocity model.
[0071] Seismic data in 503 can be obtained by applying time migration algorithms such as Stolt migration, Gazdag, Finite-Difference migrations. Example of depth migration algorithms includes Kirchoff migration, Reverse Time Migration, Gaussian Beam Migration and Wave equation migration. At 505 a volume amplitude spectrum is determined for an amplitude volume. For example, the amplitude volume can be analyzed in a low frequency amplitude spectrum (e.g., 3Hz - 10HZ), a high frequency amplitude spectrum (80Hz - 130Hz), and a prime frequency amplitude spectrum (e.g., 12Hz - 70Hz) independently to estimate and predict reservoir rock and fluid properties using spectral decomposition techniques that entail breaking down the seismic signal into its component frequencies. A fully processed seismic survey amplitude volume depending on seismic acquisition hardware settings 512, contains frequencies that are capable of being recorded by
geophones/hydrophones used for that particular survey.
[0072] During a seismic survey, seismic energy (e.g., acoustic energy) in the form of a seismic wave front, propagates downward into the subsurface and at each lithological facies and geologic boundary (e.g., an unconformity, bed boundaries, etc.), the seismic energy is reflected, refracted, and/or absorbed. Also, as the seismic wave front propagates into the underlying sediments, it attenuates, causing the frequency content of the seismic wave front to decrease with depth. For example high frequencies are better preserved near the earth's surface, are "drowned" by the more dominant, lower frequencies. In practice, spectral decomposition assesses seismic response at different, discrete frequency intervals. For example higher frequencies image thinner beds, while lower frequencies image thicker beds. While it is known in the art that seismic reflection from a thin bed has a characteristic expression in the frequency domain that is indicative of its thickness in time, the seismic response is non-unique and attenuation cannot be ascribed to a particular attribute.
[0073] Steps 505 through 560 describe series of spectral decomposition operations to estimate a bandwidth of acquired seismic data at a formation of interest at different signal to noise ratios for the amplitudes, and partitioning the bandwidth or processed seismic data into three regions - a high frequency region, low frequency region and prime amplitude where most of the seismic energy is concentrated.
[0074] The physics of rock and fluid property computation seeks to analyze the following effects:
[0075] 1. Hydrocarbon/Fluid Saturation - is discriminated on basis of spectral energy attenuation (or absorption) in seismic reflection amplitudes at low frequency. These low frequency amplitudes are treated as noise by conventional seismic processing method and filtered out.
[0076] 2. Vshale (is defined as the volume of shale expressed as a decimal fraction or percentage). Vshale is estimated from small amplifications in the high frequency amplitudes. These high frequency amplitudes are treated as noise by conventional seismic processing method and filtered out. [0077] 3. Porosity attribute is discriminated on the basis of scatter in high frequency noise. These high frequency amplitudes are treated as noise by conventional seismic processing method and filtered out.
[0078] 4. Brittleness - a geomechanical property determine based on phase variation in low and high frequency noise. Both these low and high frequency amplitudes are treated as noise by conventional seismic processing method and filtered out.
[0079] Spectral decompositions are used in 530 to partition the seismic volume bandwidth into high and low frequency attributes. The frequency bandwidth taken into consideration during spectral decompositions includes amplitudes as low as -24dB in one implementation approach. The specifics of frequency bands separating the three regimes are dependent upon the parameters of an quantum resonance interferometry engine used in the actual calculation of attributes as described in connection with the quantum resonance interferometry engine 828 in Figures 8 and Figure 10. Depending on the parameters of the quantum resonance interferometry engine, separation bounds for the three regions are determined at 520. The advantage of a having a 2dB to 4dB separation in seismic amphtudes between the spectral decomposition seismic volumes produced in 530 is that it reduces artifacts introduced by use of trapezoidal shaped wavelets (e.g., Ormsby) used in the processing sequence to prepare seismic data 110. As an example, the separation bounds for the three regions can be determined in 520 as follows: A bandwidth in the prime frequency amplitude regions can include reflection amplitudes above -6dB or above -lOdB cut-off; low frequency amplitudes can be set between [-12dB or -16dB] and [-22dB or -24dB] on frequencies lower than the central frequency of the obtained seismic data; and high frequency amplitudes are set between [ -16dB] and [-22dB or -24dB] on frequencies higher than central frequency of the obtained seismic data. A 2dB or 4dB separation filter between the bands and amphtudes covering additional 10 to 12dB yields up to additional 4Hz to 20Hz of data depending on the acquisition quality of the seismic data?
[0080] If data has been denoised or band-passed, as in 516 then these spectral
decompositions are not effective and seismic data in 503 needs to be reprocessed using a processing workflow that does not apply frequency filtering. If seismic data in 110 has been pre-processed using noise attenuating spectral whitening as in 514, that broadens or equalizes the spectral of the signal making it similar to white noise spectrum, then the spectral decomposition intervals have to be extended to obtain low and high frequency spectral decomposition in the -20dB to -40dB regime to offset the unintended increase in noise power due to spectral whitening. This is performed as spectral whitening operations when applied to pre-stack seismic traces enhances both the signal and noise ampUtudes. Also the trace-by-trace variations in the wavelet's frequency content produced by spectral whitening is an undesirable effect is output produced by this method in 170 are to be use to drive amplitude-versus-offset (AVO), amplitude-versus-angle (AVA) processing or time lapse seismic studies on a reservoir. While there is limited data on resolution gain by working in the extended low and high frequency amplitudes as in 516, as shown in the empirical studies conducted using this method the vertical resolution could be increased significantly or even doubled. For example in a typical survey shown to have 10Hz to 50Hz bandwidth for amplitudes above -lOdB, extending the bandwidth to 4Hz to 90Hz effectively doubles the vertical resolution.
[0081] At 520 separation bounds are determined in the volume amplitude spectrum. In some examples, the frequency regimes can be set to a low frequency regime ("LowF") covering a lower 5 th percentile of the volume amplitude spectrum, an upper frequency regime (HighF) covering an upper 5th percentile of the volume amplitude spectrum, and prime amplitude regime ("PrimeAmp") covering a middle 8th to 92th percentile of the volume amplitude spectrum. This is useful for performing sensitivity analysis on resolution vis-a-vis data quality as required by quality control (QC) processes applied to outputs produced by this method.
[0082] At 530, cutoffs are determined for different frequency regimes. As a result of this step, three SEGY partitioned volumes are created for each seismic horizon. Depending on the targeted porosity range, prospect voxels are extracted from one or more of the frequency regime volumes. As an example, if the porosities of interest are expected to be less than 8% to 10% (e.g., for some limestone, dolomite formations) or the geobodies are discontinuous or stratigraphically stacked; then HighF volume is preferred for porosity determination. If the porosities are expected to be high and vary over a larger range (e.g., >15%) then PrimeAmp is preferred. If the porosity is expected to have very little variability (less than 2% to 3%) over the prospect area then LowF can be used.
[0083] At 540 All the three regimes are however important to QC and check consistency (continuity) of the porosity computation. If the upstream processing sequence was unable to preserve spectrum, then use of HighF volume is preferred. Band limited frequency volumes derived from results of an spectral decomposition are used as inputs to the voxelization process. Seismic volumes generated without denoising are preferred. Notwithstanding the selection of frequency limited volume, it is important to ensure that rock physics model and seismic data noise in Figure 9A through 9D are derived and optimized to the same frequency regime.
[0084] At 560, the partitioned seismic volumes are output for voxelization at 560.
Process 500 allows seismic data to be transformed to a representation that allows exploitation of subtle changes in high and low frequency amplitudes which would otherwise be treated as noise in a conventional process and filtered out. Errors in rock and fluid property determination for a formation of interest are reduced by focusing processes in Figures 5 and 6 on optimal input data.
[0085] Figure 6 enumerates examples of some subsurface volumes (e.g., subsurface volume 220) and their taxonomy for the prospect discussed at 110 in Figure 1. These subsurface volumes can be analyzed for potential formations of interest as discussed at 115 in Figure 1. These specifically drive Unit 503 in Figure 5. The inputs can be broadly classified into six categories, including pre-stack time or depth migrated gathers 610, raw seismic gathers 620, and post stack seismic or depth migrated gathers 630, trace attributes 640, volume attributes 650, and bulk attributes 660.
[0086] The pre-stack time (PSTM) or depth migrated (PSDM) seismic gathers in 610 are obtained by conventional 2-D, 3-D or multicomponent seismic processing sequences that is well known in the art. In some examples of Figure 1 and Figure 2, it is preferred that processing sequences for the seismic gathers 610: preserve amplitudes, preserve spectrum, preserve phase, remove ground-roll and other out-of-formation coherent noise, do not apply denoising transformations to remove random in-formation noise, use high density tomography grids for velocity estimation, process data at the acquisition sampling resolution, and remove traces that are highly spectrally attenuated.
[0087] Post-Stack time or depth gathers of 630 can be processed according to Figure 1 but, in some examples, voxelization described in the process 700 introduces artifacts.
[0088] Raw gathers 620 can be processed according to process 100 in Figure 1. In some examples, the raw seismic gathers best use is in a confirmatory mode when trying to predict drilling outcome prior to actually drilling. If a prospective well has been positioned at the prospect (using the systems or techniques disclosed herein or using another conventional method and a second opinion is desired to see what outcome can be expected) then the raw seismic gathers 630 or acquired raw seismic traces can be used to estimate rock and fluid properties at the prospective well.
[0089] Inputs to 220 can be derived using one or a combination of post-stack trace attributes 640 include that include wavelet attributes, Instantaneous attributes 642 such as an instantaneous phase volume or an dominant frequency, or an envelope response volume, can be used according to the systems and techniques disclosed herein to answer questions specific to structural stratigraphy and facies modeling. There are over 50 standard geometric and wavelet-based post-stack attributes that can be generated using commercial software packages. Other trace attributes, include attenuation Q attribute volume 642 and spectral decomposition attribute volume 646. Trace attributes 640 can be used according to the systems and techniques herein to improve signal to noise of a specific attribute or focus on a sub-region with prospect of interest.
[0090] Volume attributes 650, include full volume attributes such as impedance inversion 652 (i.e., simultaneous pre-stack inversion). Volume attributes 650 also include local dip/azimuth 654 actually take a subset of neighboring seismic around each location of interest in the prospect when predicting reservoir properties at a new location. Volume attributes 650 also include spatial semblance 656, which can be used to highlight subtle discontinuities in seismic volume that stem from faults and complex dips. Volume attributes 650 can be used to conduct quality control on the results produced, for example, at 160 in Figure 1 and to address reservoirs with known structure challenges, such as complex dips. Volume attribute 650 can also be used for analyzing clastic or sandstone reservoirs which includes laminates, channel sands, levies and bars.
[0091] More complex pre-stack derived attributes in 660 such as AVO attribute
(including AVO Intercept, AVO Gradient, AVO curvature and derivatives known to one skilled in the art), a specular versus scattered energy imaging attribute 664 or azimuthally- sectored attributes 666 are applicable to analyzing for complex carbonates, fracture reservoirs, and prospects with known anisotropy. They provide higher signal-to-noise compared to PSTM/PSDM volumes of 610 and also concentrate random noise.
[0092] Subsurface volumes can be produced in an industry format such as SEGY, as inputs for further processing in Figure 1. If any of the attributes in 630, 640, 650 or 660 are used as inputs as in 220, then detailed processing workflow scripts are required to understand how the attributes are constructed, algorithmic parameters used in their production and any smoothing operators that may have been applied.
[0093] Figure 7 shows an example of a process 700 for voxelizing seismic data for a formation of interest. An offshore or an onshore prospect can have several formations at varying depths hypothesized to be hydrocarbon bearing. For example, a multi-formation geological cross-section can have hydrocarbon formations such as a Grayburg Sandstone (at 5000 feet), a ClearFork at 6200 feet depth, a Spraberry at 8300 feet, a Wolfcamp Carbonate at 9800 feet, an Atoka limestone at 12600 feet. In this example, each of these formations of interest is independently analyzed and processed using the process 700 in Figure 7. At 705, a particular formation of interest (e.g., a Clear Fork) is selected from, for example, multiple formations of interest in a seismic volume. Bounding horizons (e.g., a top and bottom horizon) are determined for the selected formation of interest. Horizons are edges on a seismic image or form a vertical cross-section. The bounding horizons can be determined using a conventional n automatic horizon picker software program based on one or more techniques, such as conventional reflection amplitude, coherency attribute, impedance inversion, or spectral decomposition. The bounding horizons can also be determined using sonic well log derived wavelets or vertical seismic profiling (VSP) or cross-bore hole seismic to manually pick a top and bottom horizon for the formation of interest. In some examples, the seismic data obtained at 110 can be obtained, such as from another geophysical or petrophysical software application or another computing device, preprocessed with the top and bottom horizons determined. In some example, seismic horizons can be derived using seismic attributes such as Reflection amplitude volume; Amplitude versus Offset (AVO) attribute volume (gradient, intercept, fluid factors etc); Amplitude-versus-angle (AVA) volume; instantaneous phase; etc. One or more of these attributes can be inputs into the voxelization process 700. All seismic attributes enumerated in 630, 640, 650 and 660 can be derived from a starting pre-stack time migrated (PSTM) gather volume or a pre-stack depth migrated (PSDM) gather volume. The PSTM/PSDM gathers could have been be migrated isotropically or anisotropically depending on the structural complexity, or formation anisotropy (fracturing) .
[0094] At 710, voxel size for the formation of interest is set to . For example, a PSTM volume with 1ms sampling and 25m by 25M bin size can be analyzed using voxel dimensions of 12.5m by 12.5m by 10m or 25m by 25m by 1ms, by re-gridding and reprocessing the data to 12.5m by 12.5m . The vertical depth or time dimension is set to the sampling resolution for voxel depth. The actual traces are resampled to 5X times the sampling interval. Also, the volume is re-binned to 2X the common depth point (CDP) resolution. CDP denotes a common reflection point for dipping reflectors and complex velocity fields generated using seismic acquisition survey. For computational and cost reasons, it is possible to proceed with voxels at the CDP binning resolution or even 2*CDP by 2*CDP resolution by sampling resolution.
[0095] In some examples, notwithstanding the accuracy of horizon determination, guard zones above and below the horizon in 720 are developed. Guard zones are used to offset upstream processing errors, suboptimal statics solutions, process-to-processor variability, and/or limitations in the seismic velocity model estimation, which may have led to inaccuracies in the input volume in 705. Seismic velocity model are a list of pairs
(time/depth, velocity) for a given location in the subsurface where velocity analysis has been done. The analysis of various attributes, such as porosity continuity properties of the formation of interest, using the systems and techniques disclosed herein (e.g., the analyzing at 125 in Figure 1), yields structurally more refined reservoir attribute volumes with revised horizon boundaries.
[0096] Guard zones in 720 can be selected based on the thickness of the formation of interest. For example, if a formation of interest is less than 20ms (depth measured in time) or 200 feet gross thickness (depth measured in feet) then guard-zones can be selected to be +/- 8ms or +/- 40feet respectively to ensure horizon event tie errors and small polarity shifts are accounted for. If the formation of interest is greater than 20 ms gross thickness or greater than 200 feet then guard-zones are selected at 20% of horizon thickness. For small formations, (less than 60 feet thickness) two guard zones are set to a combined thickness that at least matches (is at least 100%) the formation thickness.
[0097] In some examples, guard-zone thickness can be determined to be 10 times the upper bound of seismic velocity model estimation error bounds, reported during
PSTM PSDM processing of seismic data, for depths shallower than 10,000 feet and 20 times the velocity estimation error for deeper depths to compensate for imaging errors. Also, it is assumed that formation dips and structural complexity is properly accounted for in the seismic horizon development process.
[0098] At 730, once the guard zones have been established, the horizons including the guard zones can optionally be flattened to a fixed time/depth value using a flattening operator to produce a flattened slab. In some examples, analysis in 740 can also proceed with an unflattened formation of interest with upper and lower bounds and guard zones . Horizon Flattening, through time or depth shifts, is a standard utility operation available in most geophysical software packages
[0099] At 740 seismic traces are interpolated. A seismic trace represents the response of the elastic wavefield to velocity and density contrasts across interfaces of layers of rock or sediments as acoustic energy travels from a source through the subsurface to a receiver or receiver array. All seismic traces (components of a PSDM/PSTM seismic volume) in the flattened slab are interpolated to report amplitudes at 5 times the sampling rate. For seismic data in the slab having low resolution surveys with processing sampling rate of 2ms resampling at 10 times the PSTM/PSDM processing sample rate is implemented.
[00100] At 750, voxels are extracted in the formation of interest with a flattened horizon. Common midpoint (CMP) gathers in the seismic data in the slab are partitioned into "mini- bins" to allow processing at 2X, 3X or 4X up-scaled areal resolution of conventional survey design resolution. For 3D seismic survey, the large bins sizes are preferred and designed to satisf spatial aliasing criterion (given by
Figure imgf000027_0001
[00102] where v is the minimum velocity,
[00103] f is the maximum frequency of the seismic signal
[00104] and ξ is the maximum reflector dip.
Traces are stacked within mini-bins. Areal resolution can be increased with high offset acquisitions of seismic data to increase areal resolution. Voxels are constructed by combining the amplitude from multiple traces in multiple mini-bins. Grid based on common depth point (CDP) binning is sized, to accommodate at least 4 (2 X 2) user-sized mini-bins per voxel, prior to any stacking. An example of a voxel size can be 42.5'X42.5', 21.25' X 21.25' in areal dimension. In the vertical (time or depth) dimension, examples of a voxels sizes can be 2ms, 4ms, 8ms (i.e., 8', 16', 20' depth). Traces are interpolated to 0.4X sampling time, e.g., 0.4ms for 2ms sampled acquisition. The entire flattened horizon for formation of interest, within the upper and lower horizon guard zones is then extracted into voxels of a targeted size in 750 using software process that copies the seismic amplitudes contained a voxel boundaries to a data structure . Voxels preset a spatial and temporal resolution to which a formation is analyzed. Reservoir attributes are determined by iterative analysis of individual voxels, one voxel at time. For each voxel that is extracted in 750, a spiral unwrapping algorithm is incorporated herein by reference from Gulati US Patent No. 7,466,851.
is used to transform the 3-D voxel data structure into a 1-D data vector of seismic amplitudes in 760. While both row and column unwrapping, as in from Gulati 7,466,851 incorporated by reference are applicable, row unwrapping is preferred to minimize aliasing effects.
[00105] Figure 8Ashows an example of a work flow 800 for analyzing voxelized data using quantum resonance interferometry processing. Machinery, such as excitation cascade engine 807, RI processor 831, well control voxel data factory, interferometric coupler 833, and quantifier 839, described in connection with Figure 8Arepresents an example of the analysis performed at 125 in Figure 1. The machinery can be implemented by a computing device such as a data processing apparatus. For example, the machinery can be implemented as computer code that when run by a processor performs the functions described in connection this machinery. In some examples, the machinery can be implemented as computer modules receiving inputs and producing outputs. Such computer modules can be implemented on the data processing apparatus. In some examples, the functions performed by the machinery can be performed by multiple computing devices, such as multiple computing devices connect through a network.
[00106] The work flow 800 has multiple inputs, including voxel data 805 derived from a spectral decomposition on a PSTM or PSD amplitude volume, seismic noise data 814, petrophysical well log data, including porosity data 817, lithology data 820, fluid property data 823, and computer generated synthetic noise 842. The computer generated synthetic noise 842 drives the resonance interferometry dynamics which lead to resonances when the incoming voxel data has an attribute of interest such as rock and fluid properties of interest.
[00107] The voxel data 805 can be a finite-length 1 -dimensional output vector, including spatial frequency amplitudes, produced by process 760 shown in Figure 7. As the input voxel data 805 for the voxels of a formation of interest can be short 1-D spectral sequences with lengths such as <200 data points, they are also referred to as spectral fragments.
[00108] The voxel data includes for each voxel seismic data such as seismic amplitudes from a specific spectral decomposition, that is LowF, HighF or PrimeAmplitude volume obtained in Figure 5 at 560. The voxel data 805 for the voxels in the formation of interest can be analyzed voxel by voxel by the machinery in Figure 8Aaccording to a geometric sequence. Various geometric sequences can be used. For example, the sequence can start from any formation corner, and can proceed by way of row major traversal or column major traversal, or by depth major orientation. In some examples, a voxel can be selected for analysis directly from a geographical 3-d arbitrarily shaped body.
[00109] In some examples, a sequence can be designed to implicitly capture address and/or the precise coordinates and orientation of the voxel within the formation of interest. In some examples, at step 120 in figure 1, a copy of the voxel data for the voxels is made where the analyzed voxel data can be output as an attribute volume. At 130, the results of the analysis of the voxel data for the voxels is written back to the same location as the copy. This mirrored read -write operation of voxel input and result ensures that the address of the voxel is accurately preserved. [00110] The seismic noise data 814 includes a noise model from amplitudes outside the low and high frequency bands obtained as in 560 and being used in the computation of reservoir attributes. These low and high frequency bands are derived through a procedure detailed in Figures 9 through 11 to ensure that the noise model uses information outside the low and high frequency information being analyzed in the voxel data for the voxels. Figures 8 A through 8B show an example of a process for determining seismic noise data 814. The interferometry engine 828, discussed in more detail below, can have a calibrated signal to noise enhancement of up to 100 times. However, the calibration of the interferometer engine 828 depends upon the amplitudes of the voxels being analyzed. For example, for amplitudes of voxels being analyzed ranging from -16dB to -22db, during a noise model design stage shown in figures 9-B through 9-D care is exercised that input value of noise power of the noise band processed by Figures 8B-8D is well separated from the -16dB through -22dB signal power amplitudes. The interferometry engine 828 can be recalibrated as various decibel ranges are analyzed such as discussed in connection with the waterfall in Figures 26 and 27.
[00111] Well logs can be selected from analogous, existing wells such as those in the same prospect as the seismic data being analyzed or from other non-local wells that have similar rock types, similar formations, and similar trap construct as the area from which seismic data is being analyzed. As mentioned above, well log data includes porosity data 817, lithology data 820, fluid data 823 (e.g., type and fluid saturation), and rock property data 824 for analogous existing wells. Subsurface sections from such well logs are analyzed to develop a reference model of what different ranges of porosity, percentage of Vshale and fluid type can be expected in the formation of interest. Wells logs for an analogous well can include gamma ray logs, neutron porosity, dipole sonic logs and resistivity logs that represent subsurface sections in the well. Well log sections can show a variability in rock and fluid properties for the analogous well. These well log sections can be, in some examples, 10 feet thick, 16 feet thick, or 32 feet thick in a zone of interest for sampling resolutions of 1ms, 2ms or 4ms, respectively.
[00112] A voxel and its attributes is defined by voxel data. The well log data is processed by a well control voxel factory 826. Well control voxel factory 826 produces well control voxel data 827. Well control voxel data (such as voxwells used by ViaLogy of Pasadena, California in QuantumRD 3.0) and their attributes are also defined by well control voxels. Well control voxel data includes data indicative of an attribute of interest derived from sources other than the seismic data that produced prospect voxel data 805 for the formation of interest. For example, in Figure 8Awell control voxel data are extracted for an attribute of interest from seismic data using well logs of known wells with known attributes. In other words, these well control voxels are extracted from the seismic data from a well bore of an analogous well. The well control voxel data can be transformed into 1-D vectors using a normalized excitation cascade transformation that yields a monotonic correlation between their energy spectral density and actual attribute value as described in Figure 13 A through 13D. For example, the well control voxel data can transformed into 1 -D vectors using the same process as in Figure 7 to produce the output 760.
[001 13] Figures 13A through 13D show an example of a process for processing well log data to produce well control voxel data 827 for attributes of interest such as for porosity, Vshale, fluid saturation, and brittleness. For example, for an attribute of interest, such as porosity, energy spectral density is derived from sonic log sections from one or more analogous wells to identify locations of known porosities. Then seismic data corresponding to those locations are identified based on identified locations. Well control voxels are extracted from the seismic data corresponding to sonic sections. For example, in well control voxel data factory 826, sonic logs are used to estimate acoustic impedance changes for regions of interest, such as an area of having a desired porosity, in an analogous well.
Wavelet kernel, such as a Ricker wavelet, is used to parameterize the sonic well logs over frequency intervals that are similar to the spatial frequency bandwidth of the Fourier Transforms implemented in transforming voxel data to an excitation cascade in 807 used to develop 1-D voxels for input data in 760.The wavelet kernel uses the same spectral decomposition that generated the amplitude volume that was voxelized in 805 (i.e., as shown in step 540 of figure 5). A calibrated well control voxel data represents energy spectral density of the Fourier Transform of wavelet model used to parameterize and model the sonic section in 826. The energy spectral density describes how the energy of a spatial frequency vector, derived from wavelet used in the well control voxel data factory is distributed with frequency. A method for building well control voxel data is described in Figure 12.
[001 14] If f(t) is a square integrable finite-energy signal, and the spectral density (ω) of the signal is the square of the magnitude of the continuous Fourier Transform of the signal then
Figure imgf000030_0001
[001 15] where ω is the angular frequency (2π times the ordinary frequency) and the continuous Fourier transform of f(t), and F * (ω) is its complex conjugate. [001 16] As the spatial frequencies of the wavelet kernel used to model the attribute presence event as seen in the sonic log section, for a well log section is discrete with values fn, ranging over an infinite number of elements, the energy spectral density for well control voxel data is
Figure imgf000031_0001
[001 17] A Discrete Fourier Transform can be used as a wavelet kernel is of finite length. The energy spectral density is a function of spatial frequency, not a function of time.
[00118] The prospect voxel data 805 is transformed to a 1-D spectral vector by an excitation cascade engine 807 and then is normalized to form a normahzed excitation cascade of the voxel data 805. Normalization compensates for high variability in the amplitude values depending on the input seismic volume in selected in 220 and spectral decomposition selected in 540. Figure 12A shows an example of how the excitation cascade is determined at 808. Figure 12B shows an example of how the excitation cascade is normalized based on energy spectral density.
[001 19] The normalized excitation cascade 811 is then analyzed by a quantum resonance interferometry engine 828, such as by the quantum resonance interferometry engine (such as "QRI Engine" produced by Vialogy of Pasadena, California in QuantumRD 3.0). The quantum resonance interferometry engine 828 includes a quantum resonance interferometry processor 831 ("RI" processor) which processes the seismic noise data 814 and the well control voxel data 827 to produce spectral data such as a quantum expressor function ("QEF") 831. Figure 8 is discussed in the context of a QEF, however, the systems and techniques that involve work flow 800 can be performed using any spectral data, including a QEF. A QEF includes complex noise that is below a threshold required to conclude the presence of an attribute of interest. For example, a QEF can include periodically modulated synthetic noise constructed using seismic data noise 814 that is modulated by a synthetic colored noise and processed through a quantum resonance interferometry processor
831 implementing quantum stochastic resonance (QSR) dynamics.
[00120] The quantum resonance interferometry processor 831 combines the seismic noise data 814 and well control voxel data 827 using a master rate equation for a quantum mechanical model to produce, for example, a bistable quantum function and drives the combination through a quantum resonance interferometry engine. Examples of how the RI processor 831 produces a QEF are discussed in connection with Figure 14 and Figure 18. Other examples of QEFs are described in further detail in U.S. Patent No. 6,963,806, 7,571,056 and 6,780,589 which are incorporated herein by reference in its entirety.
[00121] The interferometry engine also includes an interferometric coupler 833 which couples the normalized excitation cascade 811 and the QEF produced by the RI processor 831, using synthetic noise 842 to drive the coupling. The synthetic noise, such as white noise or colored noise, can be generated by random number generators. For example, a sequence of random numbers can be generated and scaled to fall within a uniform scale. The scale is a defined by a cutoff bound such as +/- 10% of an average of the energy spectral densities of well control voxel data derived using well log sections that do not possess the attribute of interest. Well control voxel data deficient in the attribute of interest are defined by negative control voxel data. Well control voxel data derived from well log sections that possess the attribute of interest are defined by positive well control Data. Negative control voxel data used in noise design is obtained using the flow of Figure 13A, 13B, 13C or 13D depending on the output of interest. In Figure 11, this cutoff can be set when the interferometric coupler is initialized at 1122. Subsequent initializations can produce'different cutoffs for finding resonances at those different cutoffs. This can be used to maintain monotonicity of the resonances. An example of synthetic noise is a sequence of random numbers generated to indicate energy spectral densities of 10% of that of an average well control voxel data The interferometric coupler 833 uses a quantum mechanical model to couple, repeatedly, the normalized excitation cascade and the QEF to generate quantum stochastic resonance. For example, a nuclear magnetic resonance ("NMR") master rate equation can be used produce QSR. An example of NMR-based QSR implementation performed by the interferometric coupler 833 is presented in Figure 17 and Figure 21. In some examples, QSR can also be implemented using Spin boson model, optical cavity model, and SQUID model. Examples of quantum mechanical models are described in U.S. Patent Nos.: 7,567,876 and 6,780,589 and 6,920,397 which are incorporated herein by reference.
[00122] The interferometric coupler 833 represents an active nonlinear coupling between prospect voxel data and an expressor function designed using combination of synthetic noise and seismic derived noise. Purpose of the computation is to track whether the incoming prospect voxel data and expressor function can couple to produce new information termed as resonance events. This interferometric coupling to produce resonances can be implemented as using a nonlinear interaction between prospect voxel data and any spectra datatype defined herein as an expressor function. Nonlinear operators that can be used to implement expressor functions to produce resonance events include stochastic resonance, quantum stochastic resonance, quantum resonance interferometry, monotone potential operators, quasi-linear elliptic differential operators, angle bounded operators, Galerkin operators, nonlinear information filter and extended kalman filters. Any semi-group operator that implements a nonlinear embedding function can be used as well.
[00123] If an attribute of interest is found in the normalized excitation cascade of the voxel data, then a stochastic resonance 836 will occur when the QEF function for that attribute of interest is coupled with the normalized excitation cascade. Resonance 836 can indicate the presence of that particular attribute of interest. In some examples, as described in Figure 17, the interferometric coupler implements a Master Rate equation describing the time dependent evolution of a system capable of exhibiting NMR. When resonance.occurs due to quantum tunneling events, as simulated within 833, the output is recorded and reported as a resonance event in 836. Also, a transformed and enhanced data vector is output in 833. The enhanced data vector is the result of modification of normalized excitation cascade through
interferometric coupling at 833 with the QEF from 831. The quantifier 839 implements an rms (root mean square) amplitude summation to generate a resonance amphtude for the voxel being analyzed. The quantifier 839 can determine that an attribute of interest is found in the voxel when the resonance amplitude value is above a pre-established threshold. Also, at 839, the resonance data can be quantified to produce an indication of the amount of the attribute of interest that is found in the voxel. For example, a resonance in 839 may indicate presence of oil, gas, or water in a voxel derived from a specific region in the formation of interest.
[00124] Figure 8B shows an example of a work flow described in figure 8A to compute a porosity attribute volume and highlight sections of the formation of interest that are above a prespecified porosity threshold (e.g., 20%) as in 8160. Input data 8110, including pre-stack migrated seismic reflectivity amplitude volume 8112 in time or depth , or an acoustic impedance seismic inversion in 8114 are processed according the process 100 in Figure 1 to produce voxels at 120. Also, the reflectivity volume 8112 and/or the acoustic impedance seismic inversion 8114 are extracted using the methods 500 and 700 of Figure 5 and Figure 7, and transformed into a normalized excitation cascade as discussed in connection with Figure 8A at excitation cascade engine 808. Also, well log sections 8135 are derived from the wells from the same prospect from which the input data 8110 was obtained to develop porosity well control voxel data using the method 1300 described in Figure 13-A. Design noise 8130 is the result of combining the synthetic noise and well control voxel data for log sections of interest. Design noise 8130 represents the output of the RI processor 831 in combining seismic noise and well control voxel data to develop a model that is used by the interferometric coupler 833to analyze voxel data derived from the input data 8110 using.. Map 8160 represents a vertical section of a porosity-cube volume derived from the results of individual voxel result assembly as contemplated by 160, and furthermore achieved by combining the results of 839 using the porosity workflow process described in Figure 19- A.
[00125] QRI processor 831 in Figure 8-A combines amplitude and phase information derived from seismic noise data 814 derived from spectral decomposition of PSTM/PSDM gathers or other attributes identified in Figure 6 with well control voxel data 827 to produce a quantum expressor function ("QEF") to the interferometric coupler 833. The interferometric coupler 833 couples incoming voxel data 805 (that has been transformed as in 811 to the same 1-D representation and vector length as the 1-D QEF vector output by the QRI processor 831) to generate resonance indicative of the presence of reservoir attribute of interest such as porosity. The seismic noise data of 814 sets a energy spectral density threshold that must be exceeded by the output of the interferometric coupler 833 to produce a resonance event 836. The threshold is specified to indicate absence or presence of an attribute value from seismic above a threshold for that attribute of interest. The outputs of the excitation cascade engine 808 and the QRI processor 831 combine in the interferometric coupler 833 to produce enhanced data vectors and with resonance events to together provide basis for detection and quantitation of rock or fluid properties for a formation of interest.
[00126] Figure 9A shows an example of a process 9100 for determining seismic noise data 814. At 9105 a partitioned seismic volume is obtained. For example, the results of process in Figure 5 implements spectral decomposition computation to produce partitioned seismic volumes 560 in low frequency, high frequency, and prime amplitude regimes, which are inputs to the noise design process 9100. These are also referred to as spectral decomposition volumes. The process 9100 is independently applied to the low frequency, high frequency, and prime amplitude regimes to generate low frequency, high frequency, and prime amplitude seismic noise, one of which is used in the process flow 800 specified in Figure 8 A depending on the nature of the amplitude volume input 9105, also referred to as a spectral decomposition volume of interest. For example low frequency noise is used in the determination of fluid saturation or type of fluid - oil, water or gas. High frequency noise is used in the computation of Vshale to generate lithology attribute.
[00127] Also, a target resolution for the prospect under consideration is obtained at 9108. This is set by the chent requirements, geology, and/or knowledge of the prospect, constrained by seismic data acquisition hardware settings and seismic survey design. In some examples, the resolution requirements obtained at 9108 can be the same as set in 510, 512, 514 and 516. The target resolution can include both a target areal resolution and a target vertical resolution for an attribute of interest such as porosity.
[00128] At 9108, well control voxel data are obtained using the regions from known, analogous wells on the prospect that do not meet cut-off threshold criteria for an attribute of interest. For example, in a porosity workflow implemented to estimate porosity for a formation ranging, with expected values of porosity ranging between 5% and 20%, sections of well log with porosity below 5% are used to obtained well control voxel data using the process described in Figure 13A. As previously defined, well control voxel data obtained using well control voxel data from wells that do not contain the attribute of interest are referred to as negative control voxel data.
[00129] At 9110, band limited volume regimes are selected. The partitioned volumes are sub-banded into small frequency bands to estimate the sensitivity of amplitude and phase sensitivity changes in the partitioned seismic volume. Once sensitivity analysis 9110 is completed, noise can be appropriately designed to be consistent with the negative control voxel data amplitudes driving the computation for an attribute of interest.
[00130] At 9120, an amplitude corridor is determined to estimate the variation in voxel data amplitudes for a partitioned seismic volumes due to inherent seismic noise. This provides parameters for an amplitude to be used in noise design which is used by the quantum resonance interferometry processor 831 and 1022. At 9160, a phase corridor is determined. This provides parameters for phase of noise design which will be subsequently used by the quantum resonance interferometry processor 831 and 1022.
[00131] Steps 9110, 9120, and 9160 are described in more detail in Figures 9B, 9C and 9D. Steps 9110, 9120 and 9160 determine variation in amplitude of all prospect voxel data in 805 under certain conditions of interest, and determine variations in phase polarity of amplitudes in seismic partitioned volumes under the same conditions of interest. Variability in both the amplitude and phase of the partitioned seismic volumes in specific regions of well on the prospect is used to develop a seismic derived noise 814 used in Figure 8Aand Figure 10. For example, log sections with porosities below 4% are used to develop noise amplitude and phase parameters if the objective of porosity attribute workflow in Figure 17A is to find regions of porosity over 4%.
[00132] At 9108, higher spectral fidelity in seismic derived noise computed using an instantaneous spectra amplitude and phase attribute enables higher resolution of seismic rock and fluid attributes. Synthetic wavelets are used to test against well logs to establish frequency boundaries required to resolve spatial features for specific depths. For example, frequencies at formation depths of interest around 100 Hz can resolve seismic reflections separated by 16 to 20 feet.
[00133] Steps 9110, 9120 and 9160 use a sub-banding operating to analyze frequency sub- bands of seismic partitioned volumes, implemented using variations of the Gabor-Morlet Transform Method which is available in conventional seismic processing software tools, to produce narrow-band analytic traces. The sub-banding operation is performed in 9120, 9120 and 9160 for different reasons. Sub-banding in 9110 is performed for the purposes of analyzing frequency sensitivity of changes in seismic amplitude changes and seismic phase changes in partitioned seismic volumes. Sub-banding in 9120 and 9160 is performed to developed corridors or maximum variation that can be tolerated due to random, in-formation seismic noise. The amplitude and phase of each narrow band filtered output obtained applying a Gabor-Morlet Transform Method represents an average amplitude and phase of the narrow-band part of the trace. Sub-band window or a 2Hz to 4 Hz separation between sub-bands derived from partitioned seismic volume, is an example of a user setting. The number of sub-bands (k) is derived by dividing the frequency bandwidth of the starting partitioned volume by sub-band window. Then k-Gabor-Morlet kernels are applied to the entire partitioned seismic volume to produce k-sub-band traces. The sub-band trace volumes are used at 9120 and 9160 to output zeta-amplitude and phase corridors as detailed in Figure 9-B and 9-C which represent the details of 9160 and 9140.
[00134] In Figure 9B an amplitude corridor is selected using sub-band volumes generated at 9110 to perform sensitivity analysis and assess variation of noise intensities. An amplitude corridor is selected by recombining sub-bands volume amplitudes produced applying Gabor- Morlet Transform Method discussed above, if the energy spectral density is unchanged in the well log sections of a well used for developing well control voxel data.
[00135] The output at 9250 establishes the sub-bands that will be used in the noise amplitude and noise phase parameter computation at 9170.
[00136] A standard complex discrete wavelet transform (CDWT) is used at 9210. For one skilled in the art, a CDWT separates the angular information in the seismic data and the polar scale. The directionality of CDWT transform is obtained by the projection of the wavelet coefficients onto positive and negative frequencies separately. The transformed data results in a complex wavelet transformation with a quasi-quadrature relationship between its real and imaginary parts and with strong directional selectivity, important for ensuring sufficient energy and directional selectivity. The CDWT is applied at 9210 to generate multiple (some integer k number of) sub-bands. Once the sub-bands have been generated, as an example those separated by 2Hz or 5 Hz, are used to assess noise sensitivity.
[00137] At 9220, a normalized amplitude attribute is computed and used as an input in the computation of an instantaneous phase attribute which outputs a cosine of the instantaneous phase angle of seismic traces within the seismic partitioned volume. Operation in 9220 amplifies the phase component of the compression seismic wave-propagation responsible for reflection amplitudes acquired in seismic survey. As 9220 provides an estimate of phase velocity, it has no amplitude information and is well suited for noise design.
[00138] As in 9230 a region of interest is selected around the wells on the prospect, targeting the formation of interest, and within +/- 5 CDP (common depth point) Bins of the well. Values in the instantaneous phase volume of 9220 are smoothened in 9240. Derivative of output of step 9220 is implemented by computing an instantaneous acceleration, and checked to ensure that resulting values in the derivative, range over a small interval, is a fraction (e.g., within +/- 50%) of the amplitudes in the spectral decomposition voxelized to estimate rock and fluid attribute computation of 110' in Figure 1. If the amplitudes in the Instantaneous Acceleration volume of 9240 are outside the +/-50% limit then well list is modified to eliminate well log sections with high "signal content". If 9250 meets satisfies the condition for finding amplitudes within +/-50% of prospect voxel data that do not have properties of interest, such as porosity below cut-off then seismic data can be used to design noise.
[00139] Figure 9C describes a process for estimating noise amplitudes to design seismic noise in 9100. At 9310, a trace sub-band frequency corridor is selected by recombining sub- bands using the same sub-banding transform as used in 9110. At 9320, recombined sub-band volumes are Hilbert transformed and the real and imaginary parts of the output separated at 9330 and 9340. This is accomplished in time domain by applying a complex Butterworth fdter. Real part ( zero phase) of the filter generates the real part of the seismic trace, thereby ensuring that both the real and imaginary parts will have the same spectral characteristics. A band limited Butterworth filter, is an exemplary filter that satisfies the band limited requirements of the Hilbert transform. The imaginary part of the complex trace has identical amplitude spectrum as the real part. Real and imaginary traces are used to generate the real and imaginary parts used in the analysis described by Figure 9C.
[00140] At 9350, a wavelet envelop operator is applied to assess the variability of the real part of the Hilbert transformed to develop 9340. The instantaneous amplitude value at the maxima of the envelop is computed for sub-band volumes generated in 9340. In parallel, an instantaneous bandwidth is computed for the sub-volume of 9310. The real part of the Hilbert transform output in 9350 is also used to compute the residual of the real part. The instantaneous amplitudes are organized to produce a histogram in 9370 which is used compute residuals for noise in 9380. In some examples, percentages lower than 2% or lower than 5% of the residual noise amplitudes can be used in 9360 to obtain the noise volumes.
[00141] In Figure 9D the optimal trace corridor is selected using sub-band volumes generated in 9410. At 9410, the envelope modulated phase attribute is computed in 9410. Intensities of the instantaneous phase represent the trace envelope magnitude. It is used to assess the phase variation of the strong events, without the interference of weaker events as in the instantaneous phase volume and vice versa. It amplifies the phase component of the wave-propagation. At 9420, an instantaneous phase attribute is computed and its normalized amplitude is used to assess the phase sensitivity. At 9420, a normalized amplitude attribute is computed to generate the cosine of the instantaneous phase angle. Operation in 9420 amplifies the phase component of the wave-propagation. As it provides an estimate of phase velocity, it has no amplitude information and is well suited for noise design.
[00142] At 9430, a second derivative of envelope attribute with respect to time is computed. Results of 9430 captures all reflecting interfaces visible within seismic bandwidth. As this attribute is traditionally used in seismic analysis to assess sharp changes in lithology and shows sharpness of event, it captures discontinuities in the data effectively. At 9450, the Time derivative of Envelope attribute is also computed from seismic sub-band volume to assess phase propagation properties. As output of 9450, highlights the sharpness of the rise time relative to absorption of seismic energy, it is more effective in treating phase propagation.
[00143] The results of 9450 are again sub-banded at 9460 to understand phase variability in the data. At 9440, the intensities from Envelope 2nd Derivative attribute are correlated to the well control voxel data derived in the vicinity of well log in zones of interest. If the correlation is low, such as below 20% then Envelop 2nd Derivative can be treated to tie the seismic around well control. Once the noise phase and noise amplitude corridors are available using Figure 9C and Figure 9D the results are combined to provide noise required to drive the quantum resonance interferometry engine and Interferometric Coupler. These volumes of noise amplitude and noise phase have useful interpretive properties in their attribute; can assist in fault anisotropy distribution. Results from 9380 and 9450 are combined in 9170 to design noise that is used by the Quantum Resonance Interferometry Processor to generate resonance on incoming, uncharacterized voxel data [00144] Figure 10 describes a process for upscaling or upconverting analysis of attributes of interest from one areal/vertical (time and/or depth) seismic resolution to a higher resolution; and to produce multi-scale reservoir attributes (i.e., output at multiple resolutions in the same or different iterations of workflow described in Figure 1), wherein the same attribute volume of rock or fluid properties are generated at multiple resolutions. This is preformed to enable different applications such as (i) de-risk drilling locations, (ii) drive reservoir-flow and reservoir-state models, (iii) perform geostatistical applications, (iv) designing fracturing/hydraulic Frac protocols, (v) plan water/C02 N2 injections for enhanced oil/gas recover, (vi) address reservoir production or hydrocarbon flow issues, and/or (vii) perform sizing of hydrocarbon reserves.
[00145] For example outputs of the process 100 in Figure 1 can help in managing production and making decisions on which areas to drill next or on which wells to complete for future production. Output of process 100 is also used to drive a reservoir model software simulation to make operational decisions that are external to this invention and its output generated by flow 100. An external reservoir model may be simultaneously running computations at different resolutions. Process flow in figure 10 can be used to generate a multi-resolution output, using coarse seismic resolution to output a higher resolution to drive a reservoir model software simulation. As most reservoir models work with gridded cells or use 3-D volume cell representation to manage and store reservoir attributes, which are then used to manage and make engineering or operational decisions for a formation of interest or under production: more accurate, upscaled reservoir attributes including porosity, Vshale and fluid saturation from Figure 10 output results for voxels in a format and areal/vertical resolution that can be directly used by an external reservoir model software simulation.
[00146] As an example, Figure 10 process enables inputs obtained using coarse seismic resolution in the vertical dimension, to generate voxel outputs 839 at close to the starting well log resolution. An example of upscaling performed by Figure 10 would be to use seismic amplitude volumes as input, with 55 feet by 55 feet by 2 ms bin in time (or 55 feet by 55 feet by 30 feet in depth) resolution and wells logs at 2 feet sampling, to provide outputs for voxels up-scaled to 27.5 feet by 27.5 feet to 0.5ms in time (or 27.5 feet by 27.5 feet by 10 feet in depth). Figure 10 enables upscaling. Another RI Processor 1022 is used to transform the normalized excitation cascade output 81 lto a different resolution with different properties. The other RI processor 1022 uses different initialization parameters and inputs from those used by RI Processor 831 in the quantum resonance interferometry engine828.As a result, the RI processor 1022 produces voxel outputs at a higher resolution than the voxel data 805. As a result, outputs by produced by the work flow 1000 shown in Figure 10 will be different from work flows without the RI processor 1022. .
[00147] As mentioned above, the voxel data 805 is transformed to a 1-D spectral vector and normalized by an excitation cascade engine 808, The RI processor 1022 combines the normalized 1 -D spectral vector with synthetic broadband colored/random noise 1024 and drives the combination inputs through a RI engine that uses a quantum mechanical model uses QSR dynamics data (such as "Virtual Vibe" as used in ViaLogy of Pasadena,
California's QuantumRD 3.0). For example, an NMR model can be used combine the synthetic broadband colored/random noise and the 1-D spectral vector. A different quantum mechanical model can be used in the RI processor 1022 than was used in the RI processor 831. QuantumRD 3.0 produced by Vialogy of Pasadena, California uses a NMR-based implementation to implement RI processor 831. As an example, the uniform scahng interval for the synthetic broadband colored/random noise is set to match 10% of the average energy spectral density amplitudes of the negative control voxel data 805.
[00148] In some examples, RI processor 1022 is introduced to increase robustness of resonance event 836 and prevent attributes of interest from being influenced by high amplitudes in noisy or low quality seismic data, or by frequency periodicity in the data in seismic traces from which voxel data is obtained. For example, periodicity in excitation cascades produced by excitation cascade engine 808 can potentially trigger resonance events 836. This can be a challenge for the identification of fluid markers described in Figure 13 C used to determine fluid properties. Also, running a normalized excitation cascade through RI processor 1022 can be used to detect weak signals buried near the boundaries of input volumes i.e., around -22dB or around -24dB.
[00149] RI processor 1022 also serves as a preconditioning mechanisms that preconditions 1-d normalized excitation cascades that are introduced into the interferometric coupler 833 by providing a shaping to certain frequencies which are directly related to rock or fluid properties of interest. For example, low frequencies are relevant to fluid saturation estimation. So, RI processor 1022 can be configured to enhance amplitudes in, for example, a 3 HZ or 4HZ window. RI processor 1022 provides an independent degree of control to bias the workflow of Figure 1 to regions of the frequency spectrum that are more important for specific rock and fluid properties. In some implementations, multiple RI processors, such as RI processorl022, can be introduced, to receive the output of 1022 as an input, and to further condition the input to generate outputs that yield higher frequency or higher resolution attributes. [00150] Figure 11 shows a process 1100 for establishing monotonicity of resonance event 836. For example, when resonance is achieved by the interferometric coupler 833, process 1100 is applied to finalize the resonance event 834. At 1110 voxel data, such as 1-D spectral vector produced by the excitation cascade engine, is obtained for processing by the process 1100. At 1122, the coupler is initialized. During initialization, an initial cutoff bound (i.e., threshold) for the intensity of synthetic colored noise 842 is set. The cutoff bound is set, for example, to a fractional intensity of the average intensity of the prospect voxel data in the input seismic data in process 100 at 118. For example, the synthetic noise 824 is constructed using random number generator algorithms. An example of a fraction used as a synthetic noise cutoff bound is a scaling intensity of 10% of the intensity of partitioned volume.
[00151] At 1120, the interferometric coupler is run to produce resonance. If at 1125 a resonance is not achieved, the voxel data for the voxel is failed at 1160, which means that a robust resonance event was not established for the voxel, and therefore the voxel does not exhibit the attribute of interest. The failed voxel is assembled at 1150 into a voxel assembly 1150 with an indication of an absence of the attribute of interest for that voxel. These results are written to the same voxel address (that is the same location, orientation and size) as in the input volume from which the voxel data was derived. The process 1100 is repeated for another voxel at 1110.
[00152] If a resonance is achieved at 1125, it is determined whether the monotonicity (e.g., persistence) of a resonance event has been established at 1130. For example, it is determined whether resonance has been found at a sufficient number of cutoff bounds for the voxel. If not, the cutoff bound is adjusted (e.g., increased) to a different cutoff bound. For example, the cutoff can be doubled, such as by increasing from 10% to 20% of the intensity of the partitioned volume. The coupler is reset at 1122 with the adjusted cutoff bound and the process is repeated.
[00153] If at 1130, monotonicity is established for the voxel, the voxel is assembled in the voxel assembly with an indication that a robust resonance event was determined for the voxel, and thus the attribute of interest was found for the voxel. By establishing
monotonicity in according to process 1100, the resonance is deemed to be robust and not an artifact of noise in the voxel data or an artifact of the external synthetic noise 842 being at added to the interferometric coupler 842. In some examples, monotonicity is established when resonance is achieved at three different cutoff bounds, such as at 10%, 20%, and 30%. The process 1100 proceeds with another voxel at 1110. [00154] Figures 12A through 12B describe a process for generating and normalizing excitation cascade data, with output represented as a 1-D data vector that is used in the computation of porosity, Vshale, Fluid Saturation and Brittleness attributes using the process detailed for each voxel in Figure 8Aand Figure 10. For example, figures 12A and 12B provide an example of a process that can be performed by the excitation cascade engine 808. The process of generating an excitation cascade output is dependent on the prospect voxel attribute being analyzed as it uses different positive well control voxel data and negative well control voxel data parameters 1235 depending on the attribute under consideration.
[00155] The process of Figure 12A is as follows. The incoming voxel data in 1210 obtained using the process 700 in Figure 7, is transformed to a 1 -D data vector using an unwrapping operation in 1215. Extracted data from a three dimensional (3-D) data structure used to managing seismic dataset in the industry standard SEGY notation, is produced as a 1- D data vector. Variations of unwrapping can be applied to transform data For example, some variations of types of unwrapping are discussed in U.S. Patent No.: 7,466,851 which are incorporated herein by reference. In some examples, unwrapping that is used in seismic data processing can entail traversal of data elements in a 3-D voxel to transform into 1-D data using row-major (row-by row -x-dimension travel and then each successive layer in time or depth), or column-major (column-by column - y-dimension travel and then each successive layer in time or depth),or outward spiral (in clockwise or anticlockwise sequence starting the center or corner of the prospect voxel data and then spiraling outwards and downwards in increasing time or depth or spiraling upwards in decreasing depth or time). Same method is also used for generating an excitation cascade from positive control voxel data and negative control voxel data as in Figure 13A through 13D. For notation purposes the seismic trace bins enclosed within a 3-D voxel boundaries are referred to as "mini-bins" to distinguish them from the CDP bins that correspond to seismic trace locations in a PSTM PSDM seismic volume or seismic attribute volumes as in Figure 6. If voxels use multiple CDP bins then the mini-bins may correspond to CDP bins.
[00156] In some examples, unwrapping to transform 3D prospect voxel data into a 1-D vector is clockwise outward spiral traversal starting at top left-corner (e.g., geographical northwest) of a 3-D voxel, each time or depth layers at time, going downwards in increasing time/depth. For a voxel with a 2 by 2 trace mini-bins only, this implies writing top-left corner, top-right corner, bottom-right corner and bottom-left corner for each successive layer going top down (or in increasing time or depth in a SEGY representation of seismic data). Inward spiral approach is used for voxels with number of mini-bins > 2. This approach works with odd number of bins as well. Alternately, an outwards spiral approach could be used just as well for voxels with odd mini-bin dimensions (e.g. 3 by 3 or 5 by 5) starting at the location in the center of the voxel layer. The output of 3-D voxel unwrapping in 1215 results in a 1-D data vector with seismic amplitude values at each mini-bin location.
[00157] In some examples, starting seismic amplitude slabs can be voxelized to have the center of voxel correspond with the top of seismic horizon. For most prospects, voxels sizes of 2 by 2 CDP bins and 5 sample points in depth or time are preferred designing a voxel, with the time and depth dimension are resampled by a factor or 5 or 10 depending on target spatial resolution. So, a 1 -D transformed output from 1215 has 2 by 2 by 5 by 5 or 100 data points in 1 -D starting data vector for seismic data sampled at 1ms or 2ms resolution; or 2 by 2 by 5 by 10 or 200 data points in starting 1-D vector for seismic data sampled at 4ms or higher resolution.
[00158] The 1-D vector, initially including seismic amplitudes from a band-limiting spectral decomposition process applied to PSTM or PSDM gather volumes, is then Fourier transformed to a 1-D vector of spatial frequencies at 1225. The spatial frequencies are amplified based on relationship and periodicity of neighboring data elements in the 3D voxel data in all the x, y and z dimensions. Well control voxel data parameters for an attribute of interest as generating using the methods described in Figure 13A through 13D, and maintained at 1220 are used to partition the 1-D data vector into k-RMS-bands (root mean square) where k is at least 1 and less than half the length of 1-D data vector derived from prospect voxel data. The value of k is determined using the process of Figure 13A thru 13D for different attributes of interest. Also the k-rms-band represents the results of computing the square root of means square amplitude of spectral components in 1 -D vector at 1215 for · each of the k-bands of interest. Energy spectral density of all k-rms-bands is computed as in 1230.
[00159] The steps 1240 through 1255 describe one or more modulation transformations that are required to be performed on the spectrally transformed voxel data at 1225 to generate an excitation cascade where the energy spectral density value satisfy conditions determined in the well control voxel data generation process. The representation and types of modulation kernels are determined by the rock physics input for the prospect as in Figure 2. The modulation kernels are different for different source rock types (source of hydrocarbon) such as sandstone, dolomite, limestone, siltstone.
[00160] These modulation kernels in 1235 are deployed to offset uncertainty due to the fact the limited (even with large numbers of wells on the prospect) well-control cannot always capture the full variability and complexity of underlying source rock grain matrix (e.g., dolomite) and seismic reflections only capture a bulk property of the formation. A library of wavelet kernels (e.g., using Ricker or Ormsby wavelets) is maintained for different facies change in high/low/prime amplitude frequencies ("model") for different configurations such as shale - limestone - shale or chert-dolomite-chert. These facies represent the changing acoustic impedance boundaries imaged by sonic logs in the well-bore. The method uses well- control/analogs to assess field specific variability in recognition kernels that can separate the seismic facies in different frequency regimes as wavelet kernels. Also, the energy spectral density (or power spectral density) of each recognition kernel is computed. Using a Rock Physics handbook - tables that provide actual experimental data for measurements of changes in p-wave, s-wave velocities for different rock compositions and grain sizes, "uncertainty bounds" can be established. At 1240, based on analysis of well control quality (seismic - synthetic from log tie) energy spectral density or a derived power density metric is computed as required to recover "rock recognition kernels" from seismic data in any spectral decomposition of interest. At 1245 the 1-D spectral data is convolved with a rock recognition kernel (represented as the parameters of a standard synthetic wavelet kernel) in the libraryl235. Once the appropriate kernel from the library 1235 is convolved with output produced at 1225 , the resulting energy spectral density is computed at 1255. The results produced at 1255 are compared with the energy spectral density criteria set during well control voxel data design process for the attribute of interest using 1250. If a sufficient threshold of energy spectral density is not reached then another kernel is selected from 1235 and the process is repeated. If more than 2 convolutions are required to enhance the energy spectral density to the threshold set during well control voxel data design then the voxel spectral vector produced at 1225 is resampled-and interpolated to double the length, e.g. go from a length of 100 data points set in 1215 to 200 data points. The number of iterations required in 1250 is also set during the well control voxel data design process of Figure 13 A thru 13D. The results of process shown in Figure 12A are output at 1260 and denoted as an excitation cascade 1260.
[00161] Figure 12B shows an example of a normalization process applied to prepare and condition an excitation cascade for introduction to Interferometric Coupler 833 in Figure 8Aand Figure 10. As the amplitudes of the starting seismic data from which the voxels are derived are different and could take values over a very large domain over real number space, all data is uniformly rescaled such as to [-1.0,1.0] or to a another range between [-F,+F] where F is a real-number in 12170. Rescaled excitation vector obtained at 12170 is then frequency modulated with a periodic signal at 12180 using modulation parameters from a repository 12185. This can entails the encoding of a sinusoidal carrier wave by variation of its frequency in accordance with an input i.e., excitation vector to emphasize the variations in instantaneous frequency. Also, the subtle variations in instantaneous frequency of the excitation cascade data from voxel data is used to generate resonances if the events match as those seen in well logs. As an example, modulation parameters from 12185 can be selected to enhance instantaneous frequency by a factor of 10. The results of this modulation, also referred to as loading the data on a sinusoidal carrier are output in 12190.
[00162] Figure 13A shows an example of how a porosity-specific well control voxel data is developed. For example, the well control voxel data can be developed in the processes of converting porosity log data 817 of Figure 8A and Figure 10 into a well control voxel data at 827. The parameters of the well control voxel data as output in 13170 and established by processes described in Figure 13 A, are used in the implementation of Figure 8, Figure 10 and also in the computation of excitation cascade in Figure 12A. Different ranges of porosity are of interest in different formations in different prospects. The porosity criteria of interest for the prospect under consideration in Figure 13 A is established by a porosity workflow or a porosity-cube workflow described in Figure 17 A, specifically in 17101 and 17105. Once the porosity range of interest 17105 are specified, then the process of Figure 13 A provides for developing parameters for transformation prospect voxel data of Figure 12A as relevant to the computation of porosity attribute. Based on prospect geology model, a vertical resolution is established as in 13102 for analyzing both well logs and seismic data. Depth sections from the neutron porosity well logs in the wells of interest on the prospect or analog are analyzed at 13104 to demarcate well log subsections that will be further analyzed. This can include geological log normalization, petrophysical normalization and effective porosity computation whereby different wells, that may have been drilled and logged by different companies at different times using different technologies and logging tools, are normalized so that these sections can be effectively compared. Either normalized or un- un-normalized and normalized well logs could be used as a starting point in 13104. Logs can be normalized for effective porosity for the formation of interest. For example, if the target formation entails determining the porosity of Goen limestone then reference logs needs to be preprocessed to a specific effective limestone porosity.
[00163] At 13110, formation cutoffs on the extracted porosity logs sections are established to ensure that they are properly normalized using standard procedures and quahty control implemented by well logging companies. An example of a formation cutoff would be a range such as [2%, 18%] for Permian strawn in Central Texas, or [8%, 36%] for Frio sand in Oklahoma or [11 %, 48%] for Pliocene sandstone in an offshore field off West Africa. The process in Figure 13A can work on any interval of interest at 13110.
[00164] At 13120, sonic well log sections are extracted for regions in the well-bore that are output in 13104. Sonic log section limits are estabhshed at 13120 and sections of the logs copied and porosity values within the sections of interest extracted as (starting depth, porosity) pairs of tuples. As the same depth range is used for all processes in 13 A through 13D, ending depth is not specified. Example depth sections are 8 feet, 16 feet, 10 feet, 16 feet, 32 feet or their metric equivalents for prospects that follow metric convention.
[00165] Sonic well log sections are extracted at 13120 in a manner to fully encapsulate the well log section impedance (or acoustic velocity change). For example, if 16 feet sections are used for well log analysis to design porosity examples then a 32 feet sonic section will be extracted such that the 16 feet well log section lies between the sonic section. The extraction 13120 can be also used if the sonic and well log sections are the same size and top or bottom aligned. If dipole sonic is available then the extracting at 13120 is repeated for both p-wave (compression) velocity and s-wave (shear) velocity. In some examples, compensated sonic is preferred. If sonic is not available then a sonic can be approximated from other well logs such as density log using relationships like Gardner's formula. Any method that can derive a sonic log, if one has not been acquired for the reference well, can be used to implement the extracting at 13120. If multiple wells are available or if the well has encountered a large thickness of formation of interest then multiple reference log sections and corresponding sonic sections are obtained. These well and sonic log sections are sorted and arranged in increasing order to form groups. Using a histogram construction the multiple log sections are organized in groups or porosity-bins to cover the range of porosity established in 13110. An example is groups [0-4%, >4%-8%, >8%-12%, ... , >24%-28%].
[00166] Using wavelet analysis, underlying wavelets are extracted from the corresponding sonic log sections enclosing the porosity log sections at 13130. Sonic sections are associated with well log sections using multi-attribute software data structure such as list or n- dimensional array. The initial wavelet extracted used at 13130 is selected to match the same frequency bandwidth as the one selected for spectral decompositions. While use of high frequency spectral decomposition volume is the best mode of implementation for this method, porosity attribute computation could also be implemented on prime-amplitude or high frequency or full spectrum seismic data as in 110. [00167] The energy spectral density of each sonic section is computed at 13140 and rearranged to form a line plot of energy spectral density versus increasing porosity-bins. Power spectral density or any power measure that can reduce the wavelet parametric derived at 13130 to a scalar number can be used to implement the check in 13150.
[00168] At 13150, a check is performed to see if the energy spectral density or an alternate scalar representation of wavelet parameter that encodes the information in sonic section is linear with a positive slope and the different sections corresponding to porosity well log are well separated, and accommodate the variance or standard deviation in different well log sections belonging to different wells. If the results produced at 13150 show linearity then the parameters are output at 13170 to be used to develop well control voxel data and also to be used in the excitation cascade engine 808. If the sections are not separated out then the results at 13130 are convolved with another kernel from the rock physics library 1235 in Figure 12A. The same library is maintained and represented at 13155. The convolution between a rock physics modulation kernel from the library of modulation functionsl3155 is implemented at 13160.
[00169] After convolution, the resulting data is again parameterized using a wavelet transform such as Ricker or Ormsby. Steps 13140, 13150 are repeated until linearity is established. Selection of each new kernel from the library of modulation functions 13155 is also used in the Tl-TN steps shown at 811 to implement an excitation cascade. Satisfactory generation of output at 13170 and completion of linearity test implies that consistent well control is used to tie the well data to the seismic partitioned volume and to calibrate seismic attributes. Use of multiple kernels, with each kernel representing a higher spatial frequency dimensionality, to achieve linearity in energy spectral density or to achieve a related scalar measure is sufficient, but not necessary, to optimize the tie at wells. If linearity cannot be established in 5 or 6 convolutions, then some log sections can be dropped from the process to offset the likelihood that logging could be in error, or some wells represent a petrophysical, structural or stratigraphic nonconformity or suboptimal well normalization. Sufficiency of the linearity plot to tie well log to seismic partitioned volume using energy spectral density of convolved vectors provides accuracy in porosity prediction away from well-bore during quantum resonance interferometry engine and Interferometric coupler iterations.
[00170] Figure 13B shows an example of how a Vshale-specific well control voxel data is developed to analyze lithology when converting lithology (e.g., gamma ray) log data 820 of Figure 8Aand Figure 10 into a well control voxel data in Unit 827. Vshale is a quantitative representation for lithology and particularly relevant for clastic of sedimentary types of structural, stratigraphic, and combination reservoirs. Parameters of the well control voxel data as output at 13270 and established by processes described in figure 13B, are used in the implementation of Figure 8, Figure 10, and also in the computation of excitation cascade in Figure 12 A. Different ranges of Vshale are of interest in different formation in different prospects. The Vshale criteria of interest for the prospect under consideration by the method of Figure 13B is established by the Lithology Workflow described in Figure 17B, specifically at 17201 and 17205.
[00171] Once the Vshale range of interest 17205 are specified then the method of Figure 13B provides a method for developing parameters for transformation of the process in Figure 12A, as relevant, to the computation of Vshale attribute. Based on prospect geology model, a vertical resolution is established as in 13202 for analyzing both well logs and seismic data. Depth sections from the gamma ray well logs in the wells of interest on the prospect or analog are analyzed at 13204 to demarcate well log subsections (i.e., reference regions of interest) that will be further analyzed. This can include the processes of geological logs normalization, petrophysical normalization and effective Vshale computation whereby different wells, that may have been drilled and logged by different companies at different times using different technologies and logging tools, are normalized so that these sections can be effectively compared Based on prospect geology model, a vertical resolution is established as in 13102 for analyzing both well logs and seismic data Logs can be normalized for effective Vshale for the formation of interest. For example, if the target formation entails determining the Vshale of Wilcox sandstone then reference logs needs to be preprocessed to effective Vshale ratios.
[00172] At 13210, formation cutoffs on the extracted gamma logs sections are established to ensure that they are properly normalized using standard procedures and quality control implemented by well logging companies. An example of a formation cutoff would be a range such as [10%, 90%] for Wilcox sandstone in South-Central Texas, or [30%, 70%] for Vickersburg sand in Oklahoma or [5%, 95%] for Pliocene sandstone in an offshore field off Indonesia. The process in Figure 13B is can work on any interval of interest at 13210.
[00173] At 13220, sonic well log sections are extracted for cutoff regions. Well log section limits are established and sections of the logs copied and Vshale values within the sections of interest extracted as (starting depth, Vshale) pairs of tuples. The computation can also be implemented in gamma ray intensity values as from logs without the need for mapping them to Vshale. In such case the resulting volume would be a gamma ray predictor as opposed to Vshale predictor. As the same depth range is used for all processes in 13A through 13D, ending depth is not specified. Example depth sections are 8 feet, 16 feet, 10 feet, 16 feet, 32 feet or their metric equivalents for prospects that follow metric convention.
[00174] Sonic well log sections are extracted at 13220 in a manner to fully encapsulate the well log section impedance (or acoustic velocity change). For example, if 16 feet sections are used for well log analysis to design Vshale examples then a 32 feet sonic section will be extracted such that the 16 feet well log section lies between the sonic section. The extraction 13220 can be also used if the sonic and well log sections are the same size and top or bottom aligned. If dipole sonic is available then the extracting at 13220 is repeated for both p-wave (compression) velocity and s-wave (shear) velocity. In some examples, compensated sonic is preferred. If sonic is not available then a sonic can be approximated from other well logs such as density log using relationships like Gardner's formula. Any method that can derive a sonic log, if one has not been acquired for the reference well, can be used to implement the extracting at 13220. If multiple wells are available of if the well has encountered a large thickness of formation of interest then multiple reference log sections and corresponding sonic sections are obtained. These well and sonic log sections are sorted and arranged in increasing order to form groups. Using a histogram construction the multiple log sections are organized in groups or Vshale -bins to cover the range of Vshale established in 13210. An example is groups [0-10%, >10%-20%, >20%-30%, >80%-90%, >90%, 100%].
[00175] Using wavelet analysis, underlying wavelets are extracted from the corresponding sonic log sections that correspond to log sections selected in 13204 on basis of Vshale attribute. Sonic sections are associated with well log sections using multi-attribute software data structure such as list or n-dimensional array. The initial wavelet used at 13230 is selected to match the same frequency bandwidth as the one selected for spectral decompositions. Specifically, Vshale computation can be implemented on prime-amplitude or high frequency or full spectrum seismic data in 110.
[00176] The energy spectral density of each sonic section is computed at 13240 and the rearranged to form a line plot of energy spectral density versus increasing Vshale -bins. Power spectral density or any power measure that can reduce the wavelet parametric derived at 13130 to a scalar number can be used to implement the check in 13250.
[00177] At 13250, a check is performed to see if the energy spectral density or an alternate scalar representation of wavelet parameter that encodes the information in sonic section is linear with a positive slope and the different sections corresponding to Vshale well log are well separated, and accommodate the variance or standard deviation in different well log sections belonging to different wells. If the results produced at 13250 show linearity then the parameters are output at 13270 to be used to develop well control voxel data and also to be used in the excitation cascade engine 808. If the sections are not separated out then the results at 13230 are convolved with another kernel from the rock physics library 1235 in Figure 12A. The same library is maintained and represented at 13255. The convolution between a rock physics modulation kernel from 13255 is implemented at 13260.
[00178] After convolution, the resulting data is again parameterized using a wavelet transform such as Ricker or Ormsby. Steps 13240, 13250 are repeated until linearity is established. Selection of each new kernel from the library 13255 is also used in the Tl-TN steps shown at 811 to implement an excitation cascade. Satisfactory generation of output atl3270 and completion of linearity test implies that consistent well control is used to tie the well data to the seismic and to calibrate seismic attributes. Use of multiple kernels, with each kernel representing a higher spatial frequency dimensionality, to achieve linearity in energy spectral density or to achieve a related scalar measure is sufficient, but not necessary, to optimize the tie at wells. If linearity cannot be established in a predetermined number of convolutions (e.g., 5 or 6), then some log sections can be dropped from the process to offset the likelihood that logging could be in error, or some wells represent a petrophysical, structural or stratigraphic nonconformity or suboptimal well normalization. Sufficiency of the linearity plot to tie well log to seismic using energy spectral density of convolved vectors provides accuracy in Vshale prediction away from well-bore during quantum resonance interferometry engine and Interferometric coupler iterations.
[00179] Figure 13C shows an example of how a fluid saturation-specific well control voxel data is developed to analyze fluid properties as in Figure 8A. Resistivity log values in sections with high effective porosity is one measure of fluid distribution and applicable analysis of carbonate, and clastic or sedimentary, types of structural, stratigraphic and combination reservoirs. Parameters of the well control voxel data as output at 13370 and established by processes described in figure 13C, are used in the implementation of Figure 8, Figure 10, and also in the computation of excitation cascade in Figure 12 A. As a distinction to porosity and lithology attribute computation, the fluid saturation sections meeting thresholds for exploration or well development interest may also be derived from results of well core analysis. Logging companies often take sidewall or other rock cores post-drilling. These are sent to core analysis laboratories to obtained detailed analysis of water and hydrocarbon saturations from regions of interest in the well. At every location in the well, where a core sample is taken, a value of fluid type (water, gas, oil) may be available. This table of values is used to annotate the regions of interest at 13304 and develop a mapping between depth and resistivity values. Different ranges of resistivity are of interest in different formations in different prospects. The fluid saturation criteria of interest for the prospect under consideration by the process of Figure 13C is established by the Fluid Saturation Workflow described in Figure 17C, specifically in 17301 and 17305.
[00180] Once the saturation or resistivity log value range of interest 17305 are specified then the process of Figure 13C provides for developing parameters for transformation of the process in Figure 12A, as relevant, to the computation of fluid attribute. For example, at 13302, vertical resolution is established. Depth sections from the resistivity well logs in the wells of interest on the prospect or analog are analyzed at 13304 to demarcate well log subsections that will be further analyzed. This can include the processes of geological logs normalization, petrophysical normalization, and effective fluid presence and fluid saturation computation whereby different wells, that may have been drilled and logged by different companies at different times using different technologies and logging tools, are normalized so that these sections can be effectively compared. Logs can be normalized for effective fluid type and saturation for the formation of interest. For example, if the target formation entails determining the saturation of Wilcox sandstone then reference logs needs to be preprocessed to effective fluid saturation levels using rock core data, if available. Gas-cut, Oil-cut, Gas- Oil-Ratio (GOR), are terms of the art used to represent saturation levels of hydrocarbon by different organizations and prospects. Any implied, derived or computed measure of fluid type or saturation that can be related back to the well and well log is a candidate for identifying a section of interest to look for in the seismic data or avoid (as in the case of water-saturated or water wet wells).
[00181 ] At 13310, formation cutoffs on the extracted resistivity logs sections are established to ensure that they are properly normalized using standard procedures and quality control implemented by well logging companies. An example of a formation cutoff would be a range such as [10%, 100%] for Wilcox sandstone in South-Central Texas, or [30%, 70%] for Vickersburg sand in Oklahoma or [5%, 95%] for Pliocene sandstone in an offshore field. Fluid or hydrocarbon saturation can range from 0% to 100%. The process in Figure 13C can work on any interval of interest at 13310.
[00182] At 13220, sonic well log sections are extracted for cutoff regions. As an example, well log section limits are established and sections of the logs copied and resistivity values within the sections of interest extracted as (starting depth, resistivity) pairs of tuples. The computation can also be implemented in resistivity log intensity values (that can numerically range from 0.1 ohm to lOOOohm as an example) as from logs, without the need for mapping them to saturation. In such case the resulting volume would be a resistivity value predictor as opposed to saturation predictor. Saturation can be obtained from a resistivity attribute prediction volume using an additional calibration table. As the same depth range is used for all processes in 13 A through 13D, ending depth is not specified. Example depth sections are 8 feet, 16 feet, 10 feet, 16 feet, 32 feet or their metric equivalents for prospects that follow metric convention.
[00183] Sonic well log sections are extracted at 13320 in a manner to fully encapsulate the well log section impedance (or acoustic velocity change). For example, if 16 feet sections are used for well log analysis to design saturation examples then a 32 feet sonic section will be extracted such that the 16 feet well log section lies between the sonic section. The extraction 13320 can be also used if the sonic and well log sections are the same size and top or bottom aligned. If dipole sonic is available then extracting 13320 is repeated for both p- wave (compression) velocity and s-wave (shear) velocity. A sonic log can be derived in various ways, if one has not been acquired for the reference well and can be used to implement the extracting atl3320. If multiple wells are available of if the well has encountered a large thickness of formation of interest then multiple reference log sections and corresponding sonic sections are obtained. These well and sonic log sections are sorted and arranged in increasing order to form groups. Using a histogram construction the multiple log sections are organized in groups or fluid saturation-bins to cover the range of saturation established in 13310. An example is groups [0-10%, >10%-20%, >20%-30%, >80%- 90%, >90%, 100%].
[00184] Using wavelet analysis, underlying wavelets are extracted from the corresponding sonic log sections. Sonic sections are associated with well log sections using multi-attribute software data structure such as list or n-dimensional array. The initial wavelet used at 13330 is selected to match the same frequency bandwidth as the one selected for spectral decompositions. Specifically, fluid saturation computation can be implemented on prime- amplitude or low frequency or full spectrum seismic data in 110.
[00185] The energy spectral density of each sonic section is computed at 13340 and the rearranged to form a line plot of energy spectral density versus increasing saturation-bins or resistivity. Power spectral density or any power measure that can reduce the wavelet parametric derived at 13330 to a scalar number can be used to implement the check in 13350.
[00186] At 13350, a check is preformed to see if the energy spectral density or an alternate scalar representation of wavelet parameter that encodes the information in sonic section is linear with a positive slope and the different sections corresponding to resistivity well log are well separated, and accommodate the variance or standard deviation in different well log sections belonging to different wells. If the results produced at 13350 show linearity then the parameters is output at 13370 to be used to develop well control voxel data and also to be used in the excitation cascade engine 808. If the sections are not separated out then the results at 13330 are convolved with another kernel from the rock physics library in 1235 in Figure 12 A. The same library is maintained and represented at 13355. The convolution between a rock physics modulation kernel from 13355 is implemented at 13360.
[00187] After convolution, the resulting data is again parameterized using a wavelet transform such as Ricker or Ormsby wavelet. Steps 13340, 13350 are repeated until linearity is established. Selection of each new kernel from the library 13355 is also used in the Tl-TN steps shown at 811 to implement an excitation cascade. Satisfactory generation of output at 13370 and completion of linearity test implies that consistent well control is used to tie the well data to the seismic and to calibrate seismic attributes. Use of multiple kernels, with each kernel representing a higher spatial frequency dimensionality to achieve linearity in energy spectral density or to achieve a related scalar measure is sufficient, but not necessary, to optimize the tie at wells. If linearity cannot be established in a predetermined number convolutions (e.g., 5 or 6 convolutions), then some log sections can be dropped from the process to offset the likelihood that logging could be in error, or some wells represent a petrophysical, structural or stratigraphic nonconformity or suboptimal well normalization. Sufficiency of the linearity plot to tie well log to seismic using energy spectral density of convolved vectors provides accuracy in fluid type or fluid saturation prediction away from well-bore during quantum resonance interferometry engine and Interferometric coupler iterations.
[00188] Figure 13D shows and example of how a geomechanical attribute such as brittleness-specific well control voxel data is developed to analyze subsurface rock brittleness properties. Figure 8A well control voxel data. Brittleness attribute has been shown to be important for developing nonconventional resources such as shale gas prospects which have their own classifications in terms of structural, stratigraphic and combination reservoirs. The parameters of the well control voxel data as output at 13470 and established by processes described in figure 13D, are used in the implementation of Figure 8, Figure 10, and also in the computation of excitation cascade in Figure 12 A. Brittleness values meeting thresholds for exploration or well development interest are derived from results of well core sample analysis or cutting section analysis. Logging companies often take sidewall or other rock cores and cuttings during drilling (as in logging while drilling) or post-drilling. These are sent to core analysis laboratories to obtained detailed analysis of bnttleness, elasticity, grain size from regions of interest in the well. At every location in the well, where a core sample is taken, a value of bnttleness may be available. An association table of values is used to annotate the regions of interest at 13404 and develop a mapping between depth and brittleness values. Different ranges of bnttleness are of interest in different formation in different prospects. The brittleness criteria of interest for the prospect under consideration by the process of Figure 13D is established by the Brittleness Volume Workflow described in Figure 17D, specifically in 17401 and 17405.
[00189] Once the brittleness value range of interest 17405 are specified then the process of Figure 13D provides for developing parameters for transformation of the process of Figure 12A, as relevant, to the computation of brittleness attribute. For example, at 13402, vertical resolution is established. Depth sections from the in the wells of interest on the prospect or analogs are analyzed at 13404 to demarcate well log subsections that will be further analyzed. This can include the processes of geological log normalization, petrophysical normalization, and effective brittleness computation whereby different wells, that may have been drilled and logged by different companies at different times using different technologies and logging tools, are normalized so that these sections can be effectively compared.
[00190] At 13420, sonic well log sections are extracted for cutoff regions. As an example, well log section limits are established and sections of the logs copied and associated with brittleness values from core analysis within the sections of interest extracted as (starting depth, brittleness value) pairs of tuples. The resulting volume would be a direct brittleness value predictor from seismic data volumes. Additional geomechanical properties such as Young's modulus for a unit rock section can be obtained from a brittleness attribute prediction volume using an additional calibration table. As the same depth range is used for all processes in 13A through 13D, ending depth is not specified. Example depth sections are 8 feet, 16 feet, 10 feet, 16 feet, 32 feet or their metric equivalents for prospects that follow metric convention.
[00191] Sonic well log sections are extracted at 13420 in a manner to fully encapsulate the well log section impedance (or acoustic velocity change). For example, if 16 feet sections are used for well log analysis to design saturation examples then a 32 feet sonic section will be extracted such that the 16 feet well log section lies between the sonic section. The extraction 13420 can be also used if the sonic and well log sections are the same size and top or bottom aligned. If dipole sonic is available then the process of 13420 is repeated for both p-wave (compression) velocity and s-wave (shear) velocity. A sonic log can be derived in various ways, if one has not been acquired for the reference well and can be used to implement 13420. If multiple wells are available of if the well has encountered a large thickness of formation of interest then multiple reference log sections and corresponding sonic sections are obtained. These well and sonic log sections are sorted and arranged in increasing order to form groups. Using a histogram construction the multiple log sections are organized in groups or brittleness-bins to cover the range of interest established in 13410.
[00192] Using wavelet analysis, underlying wavelets are extracted from the corresponding sonic log sections of the well where bnttleness cores were taken. Sonic sections are associated with well log sections using multi-attribute software data structure such as list or n-dimensional array. The initial wavelet used in 13430 is selected to match the same frequency bandwidth as the one selected for spectral decompositions. Specifically, brittleness computation is implemented on prime-amplitude or high frequency or full spectrum seismic data in 110. The energy spectral density of each sonic section is computed at 13440 and the rearranged to form a line plot of energy spectral density versus increasing brittleness. Power spectral density or any power measure that can reduce the wavelet parametrics derived at 13430 to a scalar number can be used to implement the check at 13450.
[00193] At 13450, a check is performed to see if the energy spectral density or an alternate scalar representation of wavelet parameter that encodes the information in sonic section is linear with a positive slope and the different sections corresponding to brittleness well section attribute are well separated, and accommodate the variance or standard deviation in different well sections belonging to different wells. If the results produced at 13450 show linearity then the parameters are output at 13470 to be used to develop well control voxel data and also in the excitation cascade engine at 808. If the sections are not separated out then the results at 13430 are convolved with another kernel from the rock physics library in 1235 in Figure 12 A. The same library is maintained and represented at 13455. The convolution between a rock physics modulation kernel from 13455 is implemented at 13460.
[00194] After convolution, the resulting data is again parameterized using a wavelet transform such as Ricker or Ormsby wavelet. Steps 13440, 13450 are repeated until linearity is established. Selection of each new kernel from the library 13455 is also used in the Tl-TN steps shown at 811 to implement an excitation cascade. Satisfactory generation of output at 13470 and completion of linearity test implies that consistent well control is used to tie the well data to the seismic and to calibrate seismic attributes. Use of multiple kernels, with each kernel representing a higher spatial frequency dimensionality to achieve linearity in energy spectral density or to achieve a related scalar measure is sufficient, but not necessary to optimize the tie at wells. If linearity cannot be established in a predetermined number of convolutions (e.g., 5 or 6 convolutions), then some log sections can be dropped from the process to offset the likelihood that logging could be in error, or some wells represent a petrophysical, structural or stratigraphic nonconformity or suboptimal well normalization. Sufficiency of the linearity plot to tie well log to seismic using energy spectral density of convolved vectors provides accuracy in brittleness prediction from seismic datasets, away from well-bore during quantum resonance interferometry engine and Interferometric coupler iterations.
[00195] Figure 14 describes a process of implementing a quantum resonance
interferometry engine 828 in Figure 8 and quantum resonance interferometry processor 1022 in Figure 10 based on a mathematical model known to exhibit a quantum stochastic resonance ("QSR") phenomenon when driven by synthetic noise, such as synthetic noise 842. The QSR phenomenon yields resonance events when certain conditions are met in the incoming data as in the output of excitation cascade engine 808 applied to prospect voxel data 805. The quantum resonance interferometry engine in conjunction with the
interferometric coupler 1022 detects weak signals in the form of subtle variations in the intensity of data as in 811 by exploiting nonlinear interactions between (i) appropriately preconditioned voxel data transformed to a normalized excitation cascade from 805, and, (ii) a synthetically designed and dynamically conditioned complex noise output using 831 derived by mixing white/colored noise and noise from seismic. Denoted as a QEF this "synthetic complex noise" results from classical projection of a digitally simulated, 1-D quantum- mechanical (Q-M) system, modulated by classical white noise. In Gulati US Patent Nos.: 6,920,397, 6,963,806 and 7,567,876 incorporated herein by reference, spin-boson system dynamics for implementing a quantum resonance interferometry dynamics is described. Methods presented in these patents are incorporated herein by reference as an alternative implementation of quantum resonance interferometry engine 828. Figure 14 shows a new process for implementing a QSR algorithm based on a master rate equation derived from a two level system for nuclear magnetic resonance (NMR) effect. Software implementation of a master rate equation for an NMR-spin based approximation to a 1-D spin-boson system is used to emulate the dynamics of a underlying quantum-mechamcal (Q-M) systems exhibiting NMR. Output of 831 couples with the input data samples from output of 811 to generate resonance event 836. A novel scalar observable output (that tracks changes to the dynamics of the simulated Q-M system) provides the basis for detection and quantitation of signal content in the input data. The process 1400 includes selection and customization of the 1-D Q-M system and its nonlinear dynamics; and, selection and specification of the scalar observable used to conclude detection of an event that defines a condition for prospect voxel data to confirm the presence of an attribute of interest above a limit-of-detection established using positive control voxel data and negative control voxel data for an attribute of interest in Figure 13A through 13D.
[00196] Figure 14 describes an efficient method for developing quantum expressor functions to drive the interferometric coupler 833. It has been shown by those skilled in quantum mechanics theory that using variational methods, density operators could be used to unify different quantum mechanical theories, and classical mathematical approximations could emulate complex quantum mechanical effects in software implementation and reduce computational complexity and development time. Subject to the same stochastic limit approximation (SLA) which is at the core of QSR phenomenology in spin-boson systems, tunneling resonance events can be demonstrated in a broad class of systems whose time dependent dynamics can be described using the Bloch Rate equations. Furthermore, incorporating a master rate equation for time evolution of an NMR system, offers a simpler path for implementing tunneling resonance to detect and analyze events of interest. Within the NMR system the spin-boson bath coupling is replaced by a spin-thermal reservoir and thermal relaxation and decoherence times.
[00197] A key tradeoff, in making the switchover from spin-boson dynamics to NMR dynamics described in Figure 14 is that dramatic tunneling changes at resonance event (observed during QSR implemented using spin-boson dynamics in Gulati US Patent Nos.: 6,963,806 and 6,920,397 incorporated herein by reference, are replaced by subtle, but statistically significant and 'Visible" changes in Tunneling Rate (TR).
[00198] Also, as a result to the switchover to NMR dynamics, the Quantum Resonance quantum resonance interferometry engine is calibrated to assess TR changes between - no signal condition (TRo) , a reference control - (i.e., positive control voxel data with known results - i.e., observed attribute values in well logs and rock core samples. ) - TRN and unknown data condition (TRu). Signal detection is based on the decision of
[00199] Δ = II ( TRu- TRo ) / (TRN - TRQ) || where Δ > ζ - preset SLA corridor (Eqn. 1) [00200] Some of these concepts are summarized in the following description.
[00201] Two Level System (NMR realization) Approximation
[00202] Imafuku et al "Quantum stochastic resonance in driven spin-boson system with stochastic limit approximation" , Report number WU-HEP-99-9, http://arxiv.org/abs/quant- ph/9910025vl, Oct. 1999, introduced the theoretical concept of the "Driven Spin-Boson Model" subject to Stochastic Limit Approximation (SLA) for Quantum Stochastic Resonance in bistable systems. Their approach assumed a Fermion-populated double-well potential in a "deep cold" thermal bath. In such configuration, a Fermion sinks to the lowest possible allowed energy level, i.e. the so-called ground state (n = 0) for the system. There are two ground states in each well, designated as jJ¾) .Qr ^ wey ^ wgjj ground states. Taking these two states as the basis states for a Hilbert space, they described the system by the Hamiltonian:
Figure imgf000058_0001
[00203] Where□ is the energy difference between the two states and□ is the tunneling amplitude between the two wells. Their next step is to create a superposition of these two states to create two new states 1 . with the relationship between the Hilbert space with these two basis states with respect to the previous Hilbert space is achieved through a rotation tran formation:
Figure imgf000058_0002
hH - sin §|i2) + cos f|Z,),
(Eqn. 3)
[00204] The reason for this transformation is that it simplifies the spin system into a two state system with an ener a iven by:
Figure imgf000058_0003
[00206] A simple sinusoidal driving function (akin to the rocking of the boat in Stochastic Resonance) with a frequency Ω and amplitude ξ can be used to drive such a spin system. The applied force is described the "perturbative" (not large compared to the system energy difference) Hamiltonian W:
[00207] W rT = £ ¾X sin Qt
[00208] Where X is a "position" operator defined by:
[00209] X = \R)(R\ - \L)(L\ [00210] X, they claim, is also an order parameter measuring the transitions between the states l-R) under the influence of the external periodic driving force.
[00211 ] In this above construct, Boson Noise Source is the Boson part of the Spin-Boson model. The idea is that the Fermion system is coupled to a Boson energy reservoir that can either add or take away energy from the Fermion system. Based on the Fluctuation- Dissipation theorem in quantum mechanics, which states that reversible fluctuations are possible in a system that is coupled in equilibrium to a thermal reservoir from which it can either draw thermal energy or dissipate energy into the reservoir. The fluctuation-dissipation theorem is the basis for the so-called Brownian Motion or Johnson Noise in electrical circuits. The spectrum of this thermal equilibrium noise is white, i.e. constant across all frequencies. The Boson reservoir can be used as the analog for a thermal reservoir under the conditions of a Stochastic Limit Approximation.
[00212] Once the S-B model is set up, the response of the S-B system is monitored through the dynamics of the position operator, X:
X{r) = trs(pB e'W Xe- '*)
(Eqn. 4)
[00213] Where the terms are:
[00214] τ = a new, re-scaled version of time "t." τ = λ2ΐ;
[00215] λ = strength of the interaction between Fermion system and the Boson system [00216] trB = trace of the matrix within the parenthesis over the boson degrees of freedom
[00217] PB = Boson ground state density operator; """"
[00218] H = Total system Hamiltonian.
[00219] Figure 14 implements a novel approximation to the spin-boson dynamics using a Density Operator for Two Level Systems (TLS). In conventional Stochastic Resonance (SR), stationary colored noise, not derived from the noise inherent in the input signal, is injected into the dynamical system (or into its analog or digital simulation). Finally, the noisy input signal itself - that is, the signal in which we wish to enhance and detect any signal of interest, if present - is applied (actually or in simulation) as a time-dependent perturbation to the dynamical system. The simplest classical dynamical system of this kind that exhibits the SR phenomenon is that of a point-particle moving in a one-dimensional conservative potential, and subject to velocity-dependent friction (as well as the injected noise and the input signal). There are variants of this system in electronics, chemistry, optics and other branches of natural science, but the underlying mathematical model, which is more or less the same in all classical phenomena exhibiting an SR mechanism.
[00220] As in 1405 a QSR double well is implemented, incorporating an algorithmic model in which the point particle's inertia (mass) is negligible, its response to the potential gradient, the input signal and the injected noise is through the first derivative of the particle position; in other words, the nonlinear, stochastic, driven ordinary differential equation is first-order. As example of a dynamical mathematical model used in ViaLogy QuantumRD 3.0 is:
[00221] *·ω= - *( )+/( + < (Eqa 5)
[00222] where: x^ is the particle's position;
[00223] ^( ) is the potential, rescaled to encode the drag (friction) coefficient;
[00224] f W is the noisy input signal dealt with in the previous sections; and
[00225] WW is the injected noise. This is a stochastic process - e.g. it could be implemented using white (Wiener) noise again, or colored noise - but it is entirely uncorrected with the stochastic process used to model the inherent noise inside . For example, if w^ is also a Wiener process, it has its own noise strength ^ , independent of
[00226] An exemplary mathematical choice for a potential which leads to SR is a bistable, two-well quartic (order 4) polynomial. As in 1405, there is flexibility in setting initial condition ( x^ ) to initialize the dynamics , as long as for any given potential and input signal, the inherent noise in f O is sufficient to "jiggle" the double-well potential sufficiently to induce
[00227] the particle to cross the inter-well potential barrier in the absence of injected noise ( ^ - ^ ). And even if such is not the case, for any given input signal one can choose the parameters of the potential such that the particle, having once fallen into one of the two wells, will stay there, provided _ ^ . Any periodicity buried in will modulate the fluctuations of x^ about that well's potential minimum, but no amplification of the SOI will occur. Note that, ignoring the noise inherent in the input signal, the effect of a the pure-signal part of upon the bistable potential for low enough Ω is, in effect, to periodically tilt it; so that for one half-cycle the right well bottom is higher than the left well bottom, while for the other half-cycle of the period ^π I ^ , the left well rises above the right one.
[00228] Having selected a potential that allows no inter- well transitions for the given input signal, consider what happens when the strength of the injected noise is gradually increased. The injected noise impels the particle to execute a type of Bmwnian motion, and if
^ is large enough, the particle will transition as frequently as desired between the two wells. However, these transitions - which are approximately described by a two-state Poisson process - are random; so once again, their pattern does not reflect to any significant extent the periodic sinusoid buried within the input signal.
[00229] So no SR occurs when the injected noise is too low, nor when it is too high.
However, at intermediate values of ^ , it was discovered - by a combination of analytical approximations, simulations and observations of physical system - that the inter-well transition patterns often locks step
[00230] With the signal of interest, the transition time series exhibits a high correlation with a sinusoid of frequency Ω . This effect can easily amplify the SNR of the input signal by a large factor, because it is the injected noise together with the parameters of the potential - and not the amplitude of the signal of interest - which determines the amplitude of the output signal.
[00231] In the method of 1400, underlying QSR (Quantum Stochastic Resonance) the dependent variable in the governing differential equation - the variable out of which the output (processed) signal is extracted - does not represent an observable of the physical system, but rather quantum amplitudes and probabilities. So these quantum dynamical equations reduce to coupled ordinary differential equation (ODEs) for the inter-well instantaneous transition. It is known to those skilled in the art that quantum system exhibits a richer variety of injected noise due to the interplay of inter-well tunneling, thermalization, decoherence, entanglement which, when optimized, can result in an even higher SNR gain factor than possible in classical SR.
[00232] Interestingly, the governing equations of the QSR effect are linear in the dependent variables, and do not even require an explicit injected-noise term (additive or otherwise). The non-linearity enters through the way in which the input signal modulates the dynamics while is
[00233 ] f O enters the ODEs multiplicatively. [00234] One way of controlling the process is by modulating the ambient temperature (simulated or actual).
[00235] Decoherence: The injected noise, even if not explicitly put into the ODEs, enters through a relaxation time (or times), stemming from quantum decoherence due to interactions of the quantum system with its environment, which thus effects a partial quantum measurement upon the system.
[00236] Therm alization: Another way in which noise is injected, is via a "quantum friction", again induced by the interaction of the quantum system with its environment.
[00237] Thus, quantum resonance interferometry engine implementation approach distinguishes between two categories of noise: inherent noise (present in the input signal and thus not controllable by the SR/QSR algorithm), and the injected noise added as part of the simulation of the dynamical physical system. Injected noise can be a combination of different types: colored noise, thermal noise, and quantum-decoherence noise.
[00238] Figure 14 describes the QSR reformulation without Bosonic Modes. The time- dependent signal to be analyzed and denoised, which will be denoted , is used to modulate the above otential into the following time-dependent one:
Figure imgf000062_0001
[00239] Where the input signal is
f(t) = A0 + A1 cos(Q.t + φ0 ) + colored noise
(Eqn. 6)
[00240] Approximate truncation to a TLS ( Two Level System ):
[00241] As the Hilbert space of a two-state system has two degrees of freedom, so a complete basis spanning the space consists of two independent states, e.g., a two-state system with the spin of a spin- 1/2 particle such as an electron, whose spin can have values +½ or -½ in units of h. It has been shown that if there is an energy asymmetry between the two states, then complex dynamics can occur and nonlinear effects can be observed. The TLS can be represented using standard Pauli spin-matrices:
σ. =
[00242]
Figure imgf000062_0002
(Eqn. 7) [00243] And refer to the eigenstates of σ *
Figure imgf000063_0001
) and spin-down (
Figure imgf000063_0002
), respectively. These TLS Hilbert-space base vectors correspond, in our truncation from the one-dimensional-particle quantum degrees of freedom, to the ground states of suitable quadratic single-well (unstable) potentials centered about the right- and left well of , respectively. Upon truncation, the zero-signal, zero-noise
KE + V(x) + A x
Hamiltonian ' 0 is then approximately replaced by the unperturbed TLS
Hamiltonian operator,
[00244] Η0 = ε0σί + Ασχ ,
e(t) =
[00245] where vf /(f) and £n υ , defined above, is seen to be the DC component of ^ ^ .
[00246] Density-Operator evolution via Quantum Markov Master Equation1:
[00247] The ambient bosonic modal-bath of the Legget-Coppersmith et al QSR formalism is replaced, in our approach, with a simultaneous, continuous process of thermalization and quantum-decoherence ( the latter being equivalent to continual, partial measurement of CTz ).
The resulting Master Equation for the temporal evolution of P , the TLS density operator, is a modification of the usual QSR Bloch rate equations; it has two independently-adjustable relaxation time constants. The evolving density operator — a 2 by 2 hermitian matrix in this case— represents a TLS with partially-measured σ 2 , so that in general, neither the well (in terms of the pre-truncation potential) nor the wavefiinction is sharply determined at any given time ^ ; instead, they are "fuzzy".
[00248] Our approximate Master Equation {linear in P but nonlinear in the signal) is:
Figure imgf000063_0003
[00249] (Eqn. 8)
P P (<y ^ = +1
[00250] where: + ( - ) is the Hilbert-space projection operator for \ * ' (
<σ· > -' ) ; [00251 ] " is the reduced Planck's constant, a measure of "quantum fuzziness";
1
β =
k
[00252] DT
with Tthe ambient temperature; and,
v
[00253] B Boltzmann's constant;
[00254] tr denotes Hilbert-space trace; and
[00255] T> and T2 represent the thermal and decoherence relaxation times, respectively.
[00256] The density operator is represented as a two-by-two matrix
+ ξ y + iz'
(
[00257] Y - IZ ± " * J (Eqn. 9)
[00258] where ^ y(t), z(t)) nQw repiaces me classical-ball-position variable, X(t) as the output signal.
X(t)
[00259] The input signal is / J ( ' . To generate all FFT's and SNR s ' , replace and its transform and auto-correlation with the corresponding entities derived from (or, alternatively, either or Z^ would have worked just as well).
[00260] The physical interpretation of ^ is that, were σ * to be fully measured at time t , the probability that σ* = +1 would be 2 + ^ .
A random number generator is used to generate "dice-throws" for successive ball positions -Ίη terms of ¾ .
[00261] It is important to note that Markovian quantum master equations generally satisfy the following conditions: (i) they describe the dynamics of time scales that are larger than the reservoir correlation time scale, (ii) these stationary solutions give the thermal equilibrium states with a reservoir. But have been shown to be valid only for single systems such as a single harmonic oscillator and a single two-level system.
[00262] So, Nuclear Magnetic Resonance (NMR) System - a TLS system can be used as the underlying model for density approximation. NMR deals generally with the behavior of the nuclear spin from the protons in the nucleus in the presence of two orthogonal magnetic fields. Basic NMR theory can be understood by taking the case of a single proton with spin = ½. Because of its quantum mechanical spin property, the proton behaves like a tiny bar magnet. In the presence of an external magnetic field, the proton has essentially two choices: align itself with the field or oppose the field (be anti-parallel).
[00263] It takes more energy to oppose the field and this energy difference between the aligned and the anti-parallel states increases with increasing magnetic field strength. As we discussed previously for TLS, electrons (being Fermions) have non-integer values of spin. Specifically, each electron can have either spin (ms) = 1/2 or ms = -1/2. That spin is a quantum mechanical property of an electron that is associated with magnetism, in the sense that you can think of each electron as a tiny bar magnet with a north and a south pole. This magnetic property helps the electron to align itself with an externally applied magnetic field in order to lower its energy. Obviously the energy is lowest when the electron's north pole is aligned with the external field and its energy is highest when it is opposing the external field.
[00264] Mathematically, this energy is expressed as:
[00265] E = - msY(hbar)Bo
[00266] Where,
[00267] E = energy of the electron in the magnetic field
[00268] ms = 1/2 or -1/2
[00269] γ = constant - also called the gyromagnetic ratio
[00270] hbar = Planck's constant (1ι)/2π
[00271 ] Bo = Strength of the external magnetic field.
[00272] The energy E is linearly dependent on the value of Bo (all the rest of the coefficients are parameters). Normally, when there is no external magnetic field present, there is obviously no difference in the energies of electrons with spin = 1/2 or -1/2. But when an external magnetic field is applied, this "degeneracy" in energies is removed and there is a distinct energy difference between the spin "up" and the spin "down" states.
[00273] This analogy was used to implement resonance for signal enhancement in QRI. The "resonance" part of NMR comes from the fact that if one can now supply the energy difference in the form of photon from an orthogonal electromagnetic (RF) field, then one can make the spins flip in their orientation. The frequency of the RF photon = E/h. In an NMR experiment, the signal results from the RF energy absorbed when these spins flip.
[00274] Spin flips can also be caused by the energy supplied by the ambient thermal environment in which the spins exist. Therefore for a solid material with zillions and zillions of protons the equilibrium ratio of the number of protons in the upper level iFp) to the number of protons in the lower level (Ndown) is given by:
[00275 ] Nup/ Ndown = e ^7 (where k = Boltzman Constant and T = absolute temperature). [00276] By convention, the z-axis is designated as the axis for the spin magnetic moment. Therefore, at thermal equilibrium, the net magnetic moment of a material, Ma is given by:
[00277] Mz = hy(Ndown- Nup)
[00278] Thermal Relaxation and Quantum Decoherence Relaxation: These are understood mechanisms by which the electron spins can dissipate their energy to the environment. These mechanisms are distinct and characterized by time constants that tell you over what kinds of time scales these dissipation effects occur. By using a Thermal Relaxation time constant, we end up coupling a bistable Fermion system with a thermal reservoir and not a Boson reservoir. Therefore, the fast-approximation in QRI implementations is not for a Spin-Boson system, but rather for a Spin-Thermal Reservoir system.
[00279] Density Matrices: Based on the Quantum Markov Approximation, spin-density matrices in a Markov-Chain Approach are set up where the future state of a system is only dependent on the current state and not on past states. The state update is implemented based on algorithm proposed by Nicholas Metropolis, Metropolis-Hastings algorithm.
[00280] The density matrix is defined as a Hermitian matrix of trace one that describes the statistical state of a quantum system. In order to obtain a value for an observable, a specific "density operator" for that particular observable, operates on the density matrix for the system to create eigenvalues corresponding to the observable. In our case, the operator is the Hamiltonian that stands for the total energy of the system.
[00281] As to the diagonal elements, they comprise of the energy difference between PotMinA and PotMinB (ε0) to which is added the kth element of the input signal (power spectrum) vector.
[00282] The Delta (Δ) corresponds to a transverse magnetic field that causes the spin to flip orientation from the +ive z-direction to the -ive z-direction and vice versa.
[00283] The reservoir temperature□ is incremented by a factor (1 + ¾).
[00284] Time Evolution of Magnetic Spins
[00285] Thermal Relaxation time constant (tj): When the equilibrium value of the magnetic moment (Mo) of a material is changed, say by applying an external RF excitation, then the system tries to get back to its equilibrium state by giving up its energy gain to the thermal environment. This thermal relaxation process (also known as spin-lattice relaxation) is governed by the following equation:
[00286] M2(t) = M0(l-e t/tl)
[00287] Spin Decoherence time constant (t2): In some NMR experiments the spins are bunched together to be "in phase" or coherent with each other during their precession around the external magnetic field direction. However, due to inhomogeneities in the magnetic field, a "decoherence" effect occurs in that the spins spread away from each other in their rotations about the z-axis. This effect can be essentially visualized as the degradation of the net magnetic moment in the xy plane. When the spins are bunched together, there is a net magnetic moment in the x-y plane. When they are completely decoherent and in random precessional orbits, Mxy = 0. Thus, the decoherence relaxation process-(also known as spinr. spin relaxation) is governed by the following equation:
Figure imgf000067_0001
[00289] Bloch Equations: Now one can construct the "Bloch" Equations of Motion, which are a kind of Maxwell Equations for NMR, that describe the dynamics of spin motion under RF excitation in the presence of the thermal environment and spin-spin decoherence: dt
(Eqn. 10)
—- = -O0 - ω)Μχ· + 2n B1Ms -
dM M,
- = -2πγΒ Μ
dt
(Eqn. 12)
[00290] Where Bi = transverse RF magnetic field (Δ in our case) and ω0 is known as the Larmor frequency (constant) and ω is the time dependent precession frequency, which I am not sure how to relate to our equations. I believe it may show up as an energy asymmetry term. Although the Bloch equations describe the evolution of the magnetic moment, they can be converted to time dependent energy equations (i.e. Hamiltonian evolution) by multiplying both sides by γΒ.
two levels as
Figure imgf000067_0002
l^ and ^ with corresponding orthonormal energy eigenvalues -¾.and ^b. The dynamics of the system can be specified as follows:
[00292] At some time ¾, let the system be in an arbitrary (and completely general) state, [00293] M*0)> = a |c> + C6 |6) '
[00294] then after evolving under H, at time t,
[00295] the state will be [00296] \≠(t)) = cae ' ft Ι^ + ο,ε 1 « |6)
[00297] Where the set of all states in the two-level system can be mapped onto the Bloch sphere constrained by equations of motions discussed above.
[00298] As in 1405, a QSR Bistable Well driven with an external driving periodic function is initialized using
systemEnergy Ratio = beta εα
Figure imgf000068_0001
X = x0
y = 0
z = 0
x' = 0
y' = 0
z' = 0
1
a = -
T
[00299] Where x, y, and z denote the variables in Bloch Rate equations customized from the general form given in Eqn. 10, 11, and 12. xo denotes the state variable for the NMR particle, a and τ denote the thermalization coefficients. The enhanced voxel and data and Tunneling Rate are reset per
[00300] An exemplary Double Well Function is shown in Figure 15.
enhanced!) ata[0]— x0
counter ossings = 0
[00301] The noise designed using process described in Figure 9-A through 9-D, and the amplitude corridor output in Figure 9-C and phase corridor output in Figure 9-D is used to design noise that is added to the voxel data input to initialize the state of the particle given by X.
[00302] At 1410, the double well as in 1510, the design is validated using conditions.
If (systemEr.er yRatio > max Energy Ratio)
barrierWidth = 2 ^e-^™^^"^ [00303] where the system energy ratio is a user set parameter.
[00304] At 1420 an iteration count is initialized to start NMR iterations.
[00305] At 1430 the NMR-QSR iteration parameters are initialized per
£now = ε0 -r ir,p tDat [k] · abs(potentialMinA)
rabiRoot = tf ' e £~„,„z I barrierWidth2
tmhArg = β · rabiRoot
tanV l = 1
if (tanhArg < max Energy Ratio) tanhVal = anh.(tanhAr
[00306] At 1440, the NMR-QSR equations of motion derived from Bloch Rate Equations are iterated according to
tanhVal
tanhRatio =
2 · rabiRoot
barrierWidth - z x + znow ' tanhRatio
x —
hV or Plank thermalRe lax axionTime
y + barrierWith · tanhRatio hVzr Plank deco her ence Relax ationTime thermalRelax ationTime v y— barrierWidth - x
z =
Figure imgf000069_0001
thermalRelax ationT ime/
[00307] The Bloch equations are a set of coupled differential equations which can be used to describe the behavior of a magnetization vector in an NMR mathematical model under any conditions. When properly integrated, the Bloch equations will yield the x', y', and z ' components of magnetization as a function of time. Only the behavior in ' is used to track changes in the particle state to conclude resonance events.
[00308] The particle state is updated at 1450 using the equation for state update given by enhancedData [k] = enhcmcedData [k— l] + dt · x'
x = enhancedData[k]
y = y + dt - y'
z = z + dt z' rey — re Threahuld— (¾,' + 7 + ζ?·)
[00309] At 1460, the tunneling rate (TR) can be calculated using the logic
[00310]
if (,enh cedData[k] 5nh nchedData[k— l] > 0) wellResidencyTime+ +
} else {
if (wellResidencyTime > residencyMin) {
countCmssings += 0.5
}
wellResidencyTime = 0
}
[00311] Once the TR has been updated the iteration count k is updates at 1470.
[00312] A 1480, a check is performed to assess whether the iteration count has been exceeded. If so, then the process 1400 stops the NMR-QSR iterations and outputs a prospect voxel output. The prospect voxel output is defined as Enhanced Data in 839 and the tunneling rate TR .
[00313] Referring to Figure 8 and Figure 10, quantum resonance interferometry processor 831 can be implemented using the steps 1405 through step 1420 of process 1400. Also, the interferometnc coupler 833 can be implemented using steps 1440 through 1480.
[00314] Referring to Figure 10, the quantum resonance interferometry processor 1022 can be implemented using the steps 1405 through 1480. Only the enhanced data computed using 1450 is output, and re-introduced into the Interferometnc Coupler 833 as an input to generate resonance 836. quantum resonance interferometry processor 831 can be implemented using the steps 1405 through 1420. And the Interferometnc Coupler 833 can be implemented using steps 1440 through 1480 as in Figure 8. And the QRI Processor 831 is implemented using the Unit 1405 through Unit 1420 as in Figure 8.
[00315] Figure 15 shows an example of a double well function that represents the quartic potential implemented in 1405.
[00316] Figure 16 shows an example of work flow 1600 using a quantum resonance interferometry engine 1628, such as quantum resonance interferometry engine 828. The work flow 1600 is denoted as a differential interferometry mode whereby the presence of an attribute of interest in a normalized excitation cascade 811, derived from a prospect voxel data 805, is determined by comparing data known to possess the same attribute of interest with the output of a Well Control Voxel. Differential Interferometry mode is used to produce a seismic data output volume from all Prospect Voxels in a formation of interest with the same property as found in well control voxel data, termed as a "Reference Voxel Data". Reference Voxel Data can be developed from a single well control voxel or multiple Well Control Voxels using the methods of Figure 13A through 13D.
[00317] Differential interferometry is used to produce results.for inquiries such as "Find all areas of porosity in an Escondido sandstone formation on a lease with over 20% porosity as seen in Well No. 32." or "Avoid all areas in the Goen platform on KCB lease which are water wet as in seen in Well No. 26." Inquiries relating to attributes of interest, including rock, fluid and geomechanical properties, reformulated to provide a "Presence" or "Absence" response, are collectively termed as a "prospect question of interest".
[00318] Reference voxel data is developed by analyzing the well log relevant to the prospect question of interest posed for the prospect, per the flows in Figure 13A though 13D and determining the sections in the well logs from well control that meet criteria of interest to develop well control voxel data 1620 denoted as Reference voxel data. Once the reference voxel data has been generated, the prospect voxel data from seismic partitioned volumes as in Figure 5 are transformed to normalized excitation cascade vectors and analyzed by the quantum resonance interferometry engine 1628. The results from quantum resonance interferometry engine 1628 are compared to see if voxel data possesses the same property as reference voxel data to answer the prospect question of interest.
[00319] Work flowl600 shows the differential interferometry mode of operation. In this model operation as in 1600, presence of an attribute of interest in a normalized excitation cascade, such normalized cascade 811 derived from prospect voxel data 805 is determined by comparing output at 839 with the output of a normalized excitation cascade derived from a well control voxel data. A prospect voxel data could be compared against positive well control voxel data or negative control voxel data using differential interferometry mode of operation as inl600. Also, prospect voxel data could be compared against one or more well control voxels, obtained by using the flow in Figure 12-A and 12-B. Output of the work flow 1600 is a determination of presence or absence of an attribute of interest in the prospect voxel data.
[00320] At 1610, a normalized excitation cascade is produced from a prospect voxel data using the process 1200 in Figure 12A and 12100 in Figure 12B. At 1620, a normalized excitation cascade, termed as Reference Voxel Data, is produced from one or more well control voxels using the flow 1200 in Figure 12A and 12100 in Figure 12B. Depending on the attribute of interest, the Well Control Voxels are generated and processed per the flows described in Figure 13A through 13D.
[00321] Outputs from 1610 and 1620 are introduced into the quantum resonance interferometry engine828, driven by both seismic noise such as seismic noise 814, produced, for example, using the flows in Figure 9-A through 9-B, as inputs in a pair-wise sequence - 161.0-followed-by 1620, or 1620-followed by 1610. In some examples, two versions of the- quantum resonance interferometry engine 828 may be used to independently process 1610 and 1620 with the results recombined in 1630.
[00322] quantum resonance interferometry engine 1628 produces a Tunneling Rate outoutl630 per the flow described in process 1400. Once the output of 1630 has been obtained from processing inputs 1610 an 1620, the tunneling rate outputs are compared in 1640 to determine'if resonance was obtained using both inputs 1610 and 1620.
[00323] At 1640, a determination is made whether the tunneling rate change output when using quantum resonance interferometry engine 1628 driven by prospect voxel data 1610 is greater than or equal to the Tunneling Rate change when quantum resonance interferometry engine 1628 is driven by Reference Voxel Data 1620. Successful conclusion of this conditional is termed as an Interferometric Resonance Dipole 1640. Interferometric Resonance Dipole requires quantum resonance interferometry engine 1628 to be configured in a manner such that resonance events occur when Reference Voxel Data is presented, and no resonance events occur when only inputs from seismic data noise as 1614 (such as seismic noise 814) are presented. Incoming, unknown, and uncharacterized prospect voxel data may or may not yield tunneling resonance when introduced into the quantum resonance interferometry engine 1628. Tunneling Resonance is concluded when there is a statistical difference between the Tunneling Rate Change as determined by observing the Tunneling Rate using prospect voxel data as an input or Reference Voxel Data as input compared to the Tunneling Rate output obtained using only seismic noise as an input. A statistical difference criterion for Interferometric Resonance Dipole 1640 is established by running multiple inputs derived from seismic noise and well control voxel data and computing the mean and standard deviations for the runs. As an example, in one embodiment 2s different between the TR output from well control voxel data as input and seismic noise data as input is used as a measure of statistical significance to establish threshold for 1640.
[00324] Atl650, the presence of property of interest in prospect voxel data is concluded if an Interferometric Resonance Dipole is found at 1640. At 1660, an absence of property of interest in prospect voxel data is concluded if a Interferometric Resonance Dipole is not found atl640. In some examples, the process 1600 can applied to sequence all prospect voxel data derived from the partitioned seismic volume of interest in 120.
[00325 ] Figure 17 A shows an example of a process 17100 to compute and predict formation porosity from seismic data. Workflow 17100 can be performed using the systems and techniques discussed herein, such as by using quantum resonance interferometry engine 828. Different ranges of porosity are of interest in different formations for different prospects. The economics of oil and gas exploration and production specify cut-offs on porosity values that are relevant to positioning new wells or new offset wells on a prospect. For example, drilling locations for offset wells are determined in relationship to an existing well and may be subject to inter-well spacing requirements constraints imposed by governing regulations for that particular formation. For example, in some states in the US only 1 well can be drilled on a 40 acre spacing for certain formations. Accurate knowledge of subsurface formation porosity is used to optimally position wells to avoid dry holes and maximize recovery. The commercial exploration criteria for porosity are provided by a client, project geologist or by the exploration manager or the lease owner/operator. Some examples of the types of answers provided by process 17100 include:
[00326] - What is the formation porosity? Or what is the Devonian formation porosity on a lease?
[00327] - Find a porosity of above x% (where x could be 10%, 6%, or 12% or any numerical cut-off) in the Devonian on a lease?
[00328] - What is the acreage of Devonian porosity above 12% on a lease?
[00329] - Are there sandstone channels with porosity over 12% in the Mississippian formation on a lease?
[00330] - Drilling a development offset is desired on a new section of the lease where the porosity matches the same as found in an existing drill at an particular location. Find an area with the same porosity characteristics?
[00331 ] - Which lease section should be developed first (e.g, one block vs. another block). Delineate and size Abo reservoirs on the basis of their porosity distribution? These questions can be answered, using the process 17100 of Figure 2 to finding the highest porosity or porosity above a specific cut-off to develop a Porosity-Cube output 275.
[00332] Process 17100 can determine a porosity-cube when it is implemented iteratively to determine a maximum porosity estimated for each of multiple prospect voxelsOnce a porosity range of interest is established for a prospect, it is broken down into intervals that can be analyzed based on well control on the prospect or using analogous well controls from other fields, prospect voxel data for the prospect voxels are tested for porosity above a threshold estabUshed by the lowest interval set for the range of interest. The workflow 17100 can be performed iteratively such that in successive iterations only voxels that exceed the porosity threshold established in the previous run are analyzed to assess if they exceed a new threshold, prospect voxel data that exceed the porosity criteria in iteration are denoted as "Surviving-Prospect Voxel" data. This-process-is repeated for surviving voxel data until the . entire porosity range is covered.
[00333] Prime-amplitude spectral decomposition 560 or high frequency spectral decomposition volume 560 is obtained at 17102 as a starting input for computations specific to formation porosity estimation. Such starting inputs can be obtained for example at step 560 in Figure 5. The starting input is voxelized at 760, (such as is described in connection with Figure 7) to produce a sequence of prospect voxel data. Individual Prospect Voxels are processed sequentially according to the process 17100 to construct a Porosity-Cube output 17160. At 17190 a normalized excitation cascade is generated (such as described in connection with 12190) from the prospect voxel data and input to the quantum resonance interferometry engine 828.
[00334] Aporosity range of interest is obtained as an external input from, e.g., the geologist, operator or a lease stakeholder at 17101. The porosity range of interest is used to produce cut-off thresholds at 17105 to select the appropriate well log sections to develop well control voxel data. The porosity range is reduced to multiple porosity increment intervals, such as 4% increments of porosity (e.g., for sandstone) or 2% (e.g., for detrital limestone such as Strawn). At 17120, porosity cut-off thresholds for porosity increment intervals from 17105 are used to develop the well control voxel data (e.g, such as by using the flow detailed in Fig 13A) to produce an output as in 13170.
[00335] The well control voxel data is used to calibrate 17407 the parameters of the quantum resonance interferometry engine 828 to be used (e.g., in implementing the flow described in 1400) to generate resonance events when Prospect Voxels are derived from regions in the formation of interest that have porosity matching or exceeding the porosity as seen in the well control on the prospect of analogous wells drilled in the same formation at another prospect. These criteria include the calibration of parameters for a double well function, such as is described in Figure 15 as related to the determination of Tunneling Rate and resonance. Double well parameters are reset for each successive porosity interval iterated by the quantum resonance interferometry engine 828 on Surviving prospect voxel data. Double well parameters can include: Initialization of energy asymmetry in the double well potential function, Interwell spacing
Width of two wells in the double well
Barrier height
Maximum jumps allowed by the Bloch Rate Equation in Eqn 10, 11, and 12, and
Residency time for spin-particle.
[00336] Output of quantum resonance interferometry engine 828 for a prospect voxel is examined for Tunneling Resonance (e.g., by using the process 1700) to see if the prospect voxel data meets the porosity threshold criteria provided in 17101 and used to produce well control prospect voxel data. The prospect voxel data, in the first iteration of the quantum resonance interferometry engine, and surviving prospect voxel data in successive runs of the quantum resonance interferometry engine are processed by the quantum resonance interferometry engine 828 to produce an output which is evaluated at 17130. If it is determined that porosity criteria is exceeded (e.g., based on method in 1400 or using Differential Interferometry method of 1600) then the prospect voxel data is concluded to be a surviving prospect voxel data for a next iteration. If the porosity threshold is not exceeded then the prospect voxel is not iterated in any subsequent iteration.
[00337] A check is performed at 17140 to ensure that all prospect voxel data or surviving prospect voxel data for a current iteration has been processed. If not, then next prospect voxel or Surviving prospect voxel data in the sequence is retrieved in 17170 and the process repeated from 17190 in flow 17100. If the check in 17140 indicates that all prospect voxel data or surviving prospect voxel data have been processed then another check 17150 is performed to see whether the upper limit of porosity for the workflow has been exceeded. If the check in 17150 is successful then the Porosity-Cube result is output such as in SEGY notation. If the check is 17150 is unsuccessful then the porosity threshold is updated to the next interval established in 17145. The quantum resonance interferometry engine parameters (e.g., double well parameters) are updated for the next iteration. And, the process is repeated.
[00338] The process 17100 can be used to implement one or more porosity intervals of interest. It can also be used to derive all voxels above a porosity threshold of interest by establishing a single interval [0,threshold] and assessing the set of Surviving prospect voxel data after the first iteration. This set of voxels in surviving prospect voxel data provides areas in the formation that have formation porosity above or below a pre-specified cut-off threshold of interest. [00339] Figure 17B shows an example of a workflow 17200 to compute and predict Vshale levels using seismic data. Workflow 17100 can be performed using the systems and techniques discussed herein, such as by using quantum resonance interferometry engine 828. Different ranges of Vshale are of interest in different formations for different prospects. Unlike porosity, Vshale is often used as avoidance-attribute, i.e., areas with high Vshale are to be avoided for positioning a well. The economics of oil and gas exploration and production specify lower cut-offs on Vshale values that are relevant to positioning new wells or new offset wells on a prospect. Accurate knowledge of subsurface formation Vshale is used to optimally position wells to avoid dry holes and maximize recovery. The commercial exploration criteria for Vshale are provided, for example, by a client, project geologist or by the exploration manager or the lease owner/operator. Some examples of the types of answers provided by process 17200 include:
[00340] - How large is an area with Vshale levels below 30%?
[00341] - What is the percentage of Vshale in a sandstone channel?
[00342] These questions can be answered, using the process 17200 to finding areas below pre-specified Vshale cut-off criteria to develop a Vshale-volume or Litho logy-Cube output as shown at 275 in Figure 2.
[00343] Process 17200 can determine a Lithology-Cube when it is implemented iteratively to determine a maximum Vshale estimated for each of multiple prospect voxels . Once a Vshale range of interest is established for a prospect, it is broken down into intervals that can be analyzed based on well control on the prospect or using analogous well controls from other fields, prospect voxel data for the prospect voxels are tested for Vshale above a threshold established by the lowest interval set for the range of interest. The workflow 17200 can be performed iteratively such that in successive iterations only voxels that exceed the Vshale threshold estabUshed in the previous run are analyzed to assess if they exceed the new threshold, prospect voxel data that exceed the Vshale threshold in iteration are denoted as "Surviving Prospect Voxel" data. This process is repeated for Surviving prospect voxel data until the entire porosity range is covered. Transformations implemented by 17200 are
[00344] Either a prime-amplitude spectral decomposition or high frequency spectral decomposition volume is obtained at 17202 as a starting input for computations specific to formation Vshale estimation. Such starting inputs can be obtained for example at step 560 in Figure 5. The starting input is is voxelized at 17290 (such as is described in connection with Figure 7) to produce a sequence of prospect voxel data . Individual Prospect Voxels are processed according to the process 17200 sequentially to construct a Lithology-Cube output 17260. At 17290 a normalized excitation cascade is generated (such as is described in connection with 12190) from a prospect voxel data and input to the quantum resonance interferometry engine 828.
[00345] A Vshale range of interest is obtained as an external input, e.g., from the geologist, operator or a lease stakeholder at 17201. The Vshale range of interest is used to produce cut-off thresholds as in 17205 to select the appropriate well log sections to develop well control voxel data. The Vshale range is reduced to multiple of Vshale increment intervals, such as 20% increments of Vshale (e.g., for sandstone) or 10% increments (e.g., for detrital limestone such as Strawn). A 17220, Vshale cut-off thresholds from 17205 are used to develop the well control voxel data, such as by using the flow detailed in Fig 13B to produce an output as in 13270.
[00346] The well control voxel data is used to calibrate 17207 the parameters of the quantum resonance interferometry engine 828 to be used in implementing the flow described in 1400, to generate resonance events when Prospect Voxels are derived from regions in the formation of interest that have Vshale matching or exceeding the Vshale as seen in the well control on the prospect of analogous wells drilled in the same formation at another prospect. These criteria include the calibration of parameters for a double well function, such as is described in Figure 15 as related to the determination of Tunneling Rate and resonance. Double well parameters are reset for each successive Vshale interval iterated by the quantum resonance interferometry engine 828 on Surviving prospect voxel data. Double well parameters can include:
Initialization of energy asymmetry in the double well potential function,
Interwell spacing,
Width of two wells in the double well,
Barrier height,
Maximum jumps allowed by the Bloch Rate Equation in Eqn 10, 11, and 12, and
Residency time for spin-particle in a particular well
[00347] Output of quantum resonance interferometry engine 828 is examined for Tunneling Resonance, such as by using the flow 1700, to see if the prospect voxel Data meets the Vshale threshold criteria provided in 17201 and used to produce well control prospect voxel data. The prospect voxel data, in the first iteration of the quantum resonance interferometry engine, and surviving prospect voxel data in successive runs of the quantum resonance interferometry engine are processed by the quantum resonance interferometry engine 828 to produce an output which is evaluated at 17230. If it is determined that Vshale criteria is exceeded (e.g., based on method in 1400 or using differential Interferometry method of 1600) then the prospect voxel data is concluded to be a surviving prospect voxel data for a next iteration. If the Vshale threshold is not exceeded then the prospect voxel is not iterated in any subsequent iteration.
[00348] A check is performed at 17240 to ensure that all prospect voxel data or surviving prospect voxel data for a current iteration has been processed. If not, then next prospect voxel or Surviving prospect voxel data in the sequence is retrieved in 17270 and the process repeated from Unit 17290 in flow 17200. If the check in 17240 indicates that all prospect voxel data or surviving prospect voxel data have been processed then another check 17250 is performed to see whether the upper limit of Vshale for the workflow has been exceeded. If the check in 17250 is successful then the Lithology-Cube result is output such as in SEGY notation. If the check is 17250 is unsuccessful then the Vshale threshold is updated to the next interval established in 17245. The quantum resonance interferometry engine parameters (e.g., double well parameters) are updated for the next iteration. And, the process is repeated.
[00349] The flow 17200 can be used to implement one or more Vshale intervals of interest. It can also be used to derive all voxels above or below a Vshale threshold of interest by establishing a single interval [0,threshold] or [threshold, upper range for formation] respectively, and assessing the set of surviving prospect voxel data after the first iteration. This set of voxels in surviving prospect voxel data provides areas in the formation that have Vshale above or below a prespecified cut-off threshold of interest.
[00350] Figure 17C shows an example of a process 17300 to determine fluid presence and fluid type, including oil, gas, and water using seismic data Process 17300 can also be used to determine a oil-water contact boundary or a gas-water boundary. In some examples, processl7300 implements a search for a fluid marker (for gas, oil or water) derived from well control voxel data. In some examples, process 17300 implements a quantitative estimation of low frequency spectral attenuation in prospect voxel data. It is known to those skilled in the art, and has been shown by empirical measurements on controlled experiments, that lower frequencies in acoustic propagation and reflection are differentially attenuated by fluids such as oil, water and gas. However, this attenuation is very small, often less than 1% (for clastic laminates) and even down to 0.01% (tight carbonates such as strawn limestone) while the noise in seismic data can be higher than 1% to 3%. In such a case, differential attenuation of reflected spectral energy in seismic is an relatively weak signal in high background noise and clutter. Process 17300 can identify differential attenuation even from within such background noise and clutter.
[00351 ] Process 17300 distinguishes between fluid-in place or hydrocarbon trap and contrast it with areas where hydrocarbon was once present but has now migrated and may be replaced by brine or freshwater. This has commercial implications for both conventional onshore and off-shore carbonate and clastic plays. It is also used to explain the spotty behavior observed when wells and offset wells drilled close together often have different outcomes in hydrocarbon recovery and presence. Process 17300 detects hydrocarbon related
inform ationusing seismic data. For example, process 17300 can detect spectral attenuation in seismic data, specifically using low frequency spectral decomposition volumes. Such a technique is apphed to low frequency spectral decompositions but can also be applied to high frequency and prime amplitude spectral decompositions as well.
[00352] Low frequencies are used to detect, discriminate and delineate fluids and oil water contact. It is not the image quality or the low frequency itself that is important, but the ability to see relative attenuation in the low frequencies as established by differences between petrophysically identified zones of interest and water-wet or zones with no fluid.
[00353] Using well log data and seismic volume can extend the low frequency boundaries for processed volumes. By combining seismic data, well log data and underlying rock physics properties to estimate attributes beyond the boundaries of conventional processing and using rock physics relationships to design how the two can be combined meaningfully. Unlike a conventional seismic impedance inversion processing that is driven using amplitudes above -6dB or above -l OdB this methods exploits a different spectral bandwidth regime. By extending the input data bandwidth by another -l OdB to -14dB, the method effectively extends the low frequencies driving the analysis by anywhere from 2Hz to 6Hz at the target depending on the resonant frequency of the geophones, assuming that active vibration source dwelled in the low frequency region for a suitable time window (depending on the formation of interest).
[00354] In summary, method of Figure 13C as in 13300 to design well control voxel data ensures that positive control voxel and negative control voxel are transformed to 1 -D Spectral data packets with frequencies where the spectral energy attenuation due to varying fluid type and fluid saturation separates out in the well control regions of interest; and these well control voxel data can be extracted from PSTM/PSDM seismic volume spectrum decompositions, at low frequency intervals. By limiting the amplitudes of extracted spectrum decompositions to be above -22dB to -24dB, to satisfy boundary conditions for the underlying quantum resonance interferometry engine calibrated to provide a 1 : 100 SNR detection. Different calibration would be needed to process prospect voxel data and design well control voxel data if amplitudes below -24db are to be used.
[00355] If the spectral energy attenuation in well control voxel data extracted from spectral decompositions does not separate the zones of interest for fluid determination even down to - 24dB then the voxels are convolved with one or more synthetic carrier kernels (derived using Ricker wavelets with varying amplitudes and side lobes) as in 1200
[00356] At 17302, a low frequency spectral decomposition volume is obtained as a starting input for fluid estimation. Such starting input can be obtained for example at step 560 in Figure 5.
[00357] The starting input is is voxelized (such as is described in connection with Figure 7) to produce as a sequence of prospect voxel data, prospect voxel data for individual prospect voxels are processed according to process 17300 sequentially to construct a Fluid Volume output 17360.
[00358] At 17390, a normalized excitation cascade is generated (such as is described in connection with 12190) from a prospect voxel data and input to the quantum resonance interferometry engine 828.
[00359] At 17301 , Fluid Marker libraries refer to the results of empirical studies, or laboratory-experiment observations, or core analysis based calibration tables that specify the relationship between rock-type, fluid saturation, relationship between different well logs, and depth for different formations, traps and fields. Data may be available from the same geological basin or field as the prospect of interest. Well control for the prospect may include results of core analysis that report hydrocarbon saturation levels with sections of the wells already drilled on the prospect. In other cases, handbooks may be available that provide relationship between resistivity logs and fluid saturation levels. In some cases, derived waterlogs, constructed from other logs and core data are available, available fluid property data can be used to establish equivalence with sections within drilled wells on the prospect or on an analog.
[00360] Once an equivalence for fluid zones of interest has been established with the well logs, well control voxel data is determined at 17305 (such as is described in connection with Figure 13C ). The output from 17305 is denoted as a fluid marker as it presents changes in energy spectral density levels exceedance which can be an indicator of fluid presence or type of fluid present in the formation of interest. Also, if an absolute spectral attenuation relationship has been developed using the processl3300 then processl7300 can output a fluid volume that represents prediction of spectral attenuation over the formation.
[00361] The fluid marker and well sections are used to produce cut-off thresholds at 17305 to select and develop well control voxel data. The output from 17305 is used to calibrate 17307 the parameters of the quantum resonance interferometry engine to be used (for example, as when implementing the flow described in 1400) to generate resonance events when Prospect Voxels are derived from regions in the formation that have fluid marker or levels of spectral attenuation exceeding a cut-off threshold of interest. These criteria include the calibration of parameters for a double well function (for example, as described in Figure 15) as related to the determination of Tunneling Rate and resonance. Double well parameters can include:
Initialization of energy asymmetry in the double well potential function, Interwell spacing,
Width of two wells in the double well,
Barrier height,
Maximum jumps allowed by the Bloch Rate Equation in Eqn 10, 11, and 12, and
Residency time for spin-particle in a particular well
[00362] Output of quantum resonance interferometry engine 828 is examined for
Tunneling Resonance, (e.g., by using the flow 1700) to see if the prospect voxel data meets the spectral attenuation threshold criteria provided in 17301 and used to produce well control voxel data.
[00363] The prospect voxel data, in the first iteration of the quantum resonance interferometry engine are processed by the quantum resonance interferometry engine 828 to produce an output which is evaluated in 17330. If it is determined that attenuation criteria is exceeded (e.g., based on method in 1400 or using differential interferometry method of 1600) then the prospect voxel data is concluded to be a surviving prospect voxel data for a next iteration. If the spectral attenuation threshold is not exceeded then the prospect voxel is not iterated in any subsequent iteration
[00364] A check is performed at 17340 to ensure that all prospect voxel data or surviving prospect voxel data for a current iteration has been processed. If not, then next prospect voxel in the sequence is retrieved in 17370 and the process repeated from Unit 17390 in flow 17300. If the check in 17340 indicates that all prospect voxel data have been processed then the Fluid-Cube result is output (e.g, in SEGY notation). [00365] The flow 17300 can be used to detect and discriminate between fluids of interest. It can also be used to derive all voxels above or below a spectral attenuation threshold of interest by establishing a single interval [0,threshold] or [threshold, upper range for formation] respectively, and assessing the set of surviving prospect voxel data after the first iteration. This set of voxels in surviving prospect voxel data provides areas in the formation that have spectral attenuation above or below a pre-specified cut-off threshold of interest.
[00366] Figure 17D shows an example of a process 17400 to compute and predict brittleness levels using seismic data. Different ranges of brittleness are of interest in different formations for different prospects. Exploration criteria for brittleness can be obtain, e.g. from a client, project geologist or an exploration manager or the lease owner/operator.
[00367] Process 17400 is implemented using an iterative workflow to determine the maximum brittleness estimated for prospect voxel data for each of multiple prospect voxel data. Once the brittleness range of interest is established for a prospect, then it is broken down into intervals that can be analyzed based on well control on the prospect or using analogous well control from other fields. All the prospect voxel data are tested for brittleness above a threshold established by the lowest interval set for the range of interest. In a successive run only voxels that exceeded the brittleness threshold established in the previous run are analyzed to assess if they exceed the new threshold, prospect voxel data that exceed the brittleness criteria in iteration are denoted as "Surviving Prospect Voxel" data. This process is repeated for Surviving prospect voxel data until the entire brittleness range is covered.
[00368] Prime-amplitude spectral decomposition or high frequency spectral decomposition volume is obtained at 17402 as a starting input for computations specific to formation brittleness estimation. Such starting inputs can be obtained for example at step 560 in Figure 5. The starting input is voxelized at 17490, (such as is described in connection with Figure 7) to produce a sequence of prospect voxel data Individual Prospect Voxels are processed sequentially according to the process 17400 to construct a brittleness-Cube output at 17460. At 17490, a normalized excitation cascade is generated (such as described in connection with 12190) from the prospect voxel data and input to the quantum resonance interferometry engine 828. As at 12190 a normalized excitation cascade is generated from a prospect voxel data and input to the quantum resonance interferometry engine 828.
[00369] A brittleness range of interest is used to produce cut-off thresholds at 17405 to select the appropriate well log sections to develop well control voxel data. The brittleness range is reduced to multiple brittleness increment intervals. Atl7420, brittleness cut-off thresholds from 17405 are used to develop the well control voxel data (e.g.,using the flow detailed in Fig 13D to produce an output at 13470).
[00370] The well control voxel data is used to calibrate 17407 the parameters of the quantum resonance interferometry engine 828 to be used (e.g., in implementing the flow described in 1400) to generate resonance events when prospect voxel data are derived from regions in the formation that have brittleness matching or exceeding the brittleness as seen in the well control on the prospect of analogous wells drilled in the same formation at another prospect. These criteria include the calibration of parameters for a double well function (e.g, described in Figure 15) as related to the determination of Tunneling Rate and resonance. Double well parameters are reset for each successive brittleness interval iterated by the quantum resonance interferometry engine 828 on Surviving prospect voxel data. Double well parameters can include:
Initialization of energy asymmetry in the double well potential function,
Interwell spacing,
Width of two wells in the double well,
Barrier height,
Maximum jumps allowed by the Bloch Rate Equation in Eqn 10, 11 , and 12, and
Residency time for spin-particle in a particular well
[00371 ] Output of quantum resonance interferometry engine 828 for a prospect voxel is examined for Tunneling Resonance (e.g, by using the flow 1700) to see if prospect voxel data meets the brittleness threshold criteria provided in 17401 and used to produce well control prospect voxel data. The prospect voxel data, in the first iteration of the quantum resonance interferometry engine, and surviving prospect voxel data in successive runs of the quantum resonance interferometry engine are processed by the quantum resonance interferometry engine 828 to produce an output which is evaluated at 17430. If it is determined that porosity criteria is exceeded (e.g., based on method in 1400 or using differential Interferometry method of 1600) then the prospect voxel data is concluded to be a surviving prospect voxel data for a next iteration. If the brittleness threshold is not exceeded then the prospect voxel is not iterated in any subsequent iteration.
[00372] A check is performed atl7440 to ensure that all prospect voxel data or surviving prospect voxel data for a current iteration has been processed. If not, then next prospect voxel or Surviving prospect voxel data in the sequence is retrieved in 17470 and the process repeated from Unit 17490 in flow 17400, onwards. If the check in 17440 indicates that all prospect voxel data or surviving prospect voxel data have been processed then another check 17450 is performed to see whether the upper limit of porosity for the workflow has been exceeded. If the check in 17450 is successful then the Lithology-Cube result is output, such as in SEGY notation. If the check is 17450 is unsuccessful then the brittleness threshold is updated to the next interval established in 17445. The quantum resonance interferometry engine parameters (e.g., double well parameters) are updated for the next iteration. And the process repeated.
[00373] The process 17400 can be used to implement one or more brittleness intervals of interest. It can also be used to derive all voxels above or below a brittleness threshold of interest by establishing a single interval [0,threshold] or [threshold, upper range for formation] respectively, and assessing the set of surviving prospect voxel data after the first iteration. This set of voxels in surviving prospect voxel data provides areas in the formation that have brittleness above or below a pre-specified cut-off threshold of interest.
[00374] The quantum resonance interferometry engine 828 can be configured in two stages. For example, Figures 18 A and 18B show configuration in a training mode and an operations mode respectively. A training mode 1810 results in acalibration of parameters for:
(1) nonlinear dynamics implemented within quantum resonance interferometry process 831 and interferometric coupler 833;
(2) initialization of a figure of merit (e.g., signals to be above a specific limit of detection (LOD), limit of quantitation (LOQ), precision and accuracy) that yield resonance events; and
(3) calibration and optimization of interference parameters for 833 implementing Eqn 10 through 12 and the underlying parameters of the Q-M dynamics of flow 1400.
Training data includes: examples of normalized excitation cascade data derived from well control voxel data for the appropriate; and properly conditioned synthetic and seismic data noise.
[00375] A subset of well control voxel data is used for testing to verify if the Training Mode is successfully completed. The entire dataset used for Training Mode is denoted as the "Training Set". For example, the training set can include well control voxel data with porosity, Vshale and spectral attenuation for fluid attribute values below a cutoff of interest (or above as the case maybe e.g., in Vshale attribute computation).
[00376] Operational mode 1820 results in the use of a calibrated quantum resonance interferometry engine initialized and calibrated by the training mode 1810 to detect and characterize prospect voxel data . The operational dataset for the operational mode 1820 is denoted as the "Input set". The quantum resonance interferometry engine 8 requires input data to be a series of real numbers with the following properties:
1 -D vector of reals with length k. It is recommended that 4<k<2048, a quality metric can be defined for the Training set, such that samples with known signal > LOD, and
metric(sample_i) > metric(samplej) if signal(sample_i) > signal(sample_j). Well Control sections (as in discussed in connection with Figures 13A -13D used in the selection of well control voxel data and arranged in increasing or decreasing order of specific attribute values and successfully meeting the test as in 13170, 13270, 13370 or 13470) are an example of data that can be used in establishing a Quality Metric.
[00377] An example of a specific Quality Metric used during training mode 1810 is a correlation coefficient for first temporal moment of samples (1 .. n) where n is the number of well control voxel data used in the Training Set, correlated to the signal strength of all labeled examplers (i.e., training set members with associated signal levels). The Quality Metric is required to be monotonic to signal strength as specified for all signal examplers. If the above test cannot be met using the first (1st-) temporal moment, then the second (2nd ) or third (3 rd-) temporal moment are checked. Also, if first-order correlation does not yield a monotonic result for 1st, 2nd, 3rd order temporal moments, the a second-order partial correlation is applied. If the second-order partial correlation measure does not yield a monotonic relationship as defined above, then the transformations [Tl ... TN] is reimplemented as in 808 to generate an excitation cascade.
[00378] During training mode 1810, an Interferometric Coupler 833 takes its input - "Input Data", and implements iterative reverberations within Unit 833 to produce two result outputs: labeled as "TR" and "Enhanced Data". TR - represent Transition Rate in the classical implementation of Equations 10 through 15, or Tunneling Rate as in the quantum- mechanical implementation of Equations 10 through 15. It is a measure of signal to noise ratio enhancement (SNR). In general, higher TR value is indicative of higher SNR. As in 1400, TR characterizes the impact of noise-modulated preconditioned input data on the dynamics of ID NMR-based Q-M spin system. An optimized quantum resonance interferometry engine yields statistically different TR for signal (prospect voxel data and well control voxel data with attributes of interest present) and background examplers (prospect or well control voxel data known not to possess attribute of interest), and provides a tractable computation for detecting attributes of interest in inherently noisy data. [00379] Enhanced Data output in 18110 represents a series of output vectors of real numbers that has the same cardinality as normalized excitation cascade presented as an input vector to the interferometric coupler 833. The Enhanced Data from Interferometric Coupler 833 represents an SNR enhancement over SNR of the input data.
Domain-specificity is encoded in the Interferometric Coupler 833 via specification of (i) TR- threshold for separating noise from signals above LOD as in resonance event 836; and (ii) calibration curve for signal quantitation to generate resonance amplitudes as in 839 in Figure 8 and Figure 10.
[00380] If the TR exhibited by Input samples is separated from TR exhibited by noise samples (also denoted as TR Threshold) by one or more standard deviations then signal presence can be concluded. While Enhanced Data vector has no utility in signal detection, wherein signal presence is solely concluded based on statistical significant changes in TR on a calibrated dataset, but its properties can be used to assess correctness of QSR-NMR iterator dynamics. Enhanced Data from 833 does however play a role in signal quantitation, where Enhanced data may be applied to a Calibration curve to assess how much signal is present on a pre-calibrated scale.
[00381] After each execution of the Interferometric Coupler 833, a TR Computation, as in a convergence test is performed to see if the mean TR value has converged. An L2 Norm suffices for this convergence test. As an example
II TR( i) TR(i-l) | | < 8
where ε is a tolerance threshold, e.g., < 0.05. If convergence has not occurred then the Enhanced Data is fed back to the TR Computation block for another process iteration.
[00382] The parameters controlling the TR convergence process in the operation mode include:
(1) TR-background -the mean TR value computed during the training phase by processing a number of sample data vectors known to contain no signal (pure noise). TR- background is calculated during training by using a set of background (no signal) samples;
(2) Convergence Criterion - Maximum number of iterations for convergence. The maximum number of iterations is a controlling parameter. QSR-NMR iterator parameters are configured to achieve a 2X enhancement in 100 iterations (to reach LOD detection) during the training phase. Once that condition is achieved, the number of iterations for operational phase is usually set to 200 to allow for uncertainty and variance in real-life noise and clutter interference around LOD This ensures robustness in performance;
(3) Consistency Criterion - Number of consecutive iteration steps (convergenceSteps) when TR mean satisfies the background "distinction" criterion.
Background distinction implies statistically different from TR_background. The value of the convergenceSteps is a small fraction of maxIterConvergence and is normally derived from it (for example convergenceSteps is typically 5 or 7 consecutive iterations, but less than 5% of maxIterConvergence); and
(4) Degree of Separation (separationDegree) - This is the threshold for the difference between the converged value of the TRmean and the TRbackground. The value of separationDegree is normally 1. A higher value like 2 or 3 may be used when it is very important to minimize the risk of false positives
[00383] The background "distinction" is decided by the formula:
[00384] I TRmean ~ TRbackground | > SeparationDegree * Obackground
[00385] Where s denotes the standard deviation across TR values obtained using Well Control prospect voxel data that fails to meet the criteria of interest or is considered negative control.
[00386] This distinction has to hold for a number of consecutive steps (e.g., 5 or 7) and is equal to convergenceSteps in order to fulfill the consistency criterion. For applications, where the noise is stationary, 3 consecutive steps may be adequate. For applications with nonstationary noise, an empirical rule is to achieve consistent separation for up to 5% of Maxlterations.
[00387] The TR computation module has a Interferometric Couple QSR-NMR engine iterator as a core element. The Interferometric Couple QSR-NMR engine module is executed for a fixed, pre-determined number of iterations. These Iterations, denoted as Innerlterations, correspond to a number of repeats the Input sequence is presented. For the first iteration the QSR module takes as its input the data as produced by the Data Normalization module and combines it with internally generated noise. Subsequently, each iteration uses the enhanced data produced in the previous iteration step, which is fed back to Interferometric Couple QSR-NMR engine after it is combined with a freshly-generated noise vector. Thus, the classical additive noise parameters changes with each iteration (basically the noise strength is being incremented). When all iterations have been executed, the mean value of TR is calculated and the last enhanced data vector produced by QSR is deemed as the ultimate output vector.
[00388] The following parameters control the TR computation process:
TR Computation Iterations (maxIterTRComputation) - This is the maximum number of iterations used for TR computation. This is set to at least twice the number of iterations required to double SNR for Signal@LOD case during the training phase. Use 200 as a default value;
Minimum Noise Strength (minNoiseStrengthTRComputation) - This is the noise strength used for the first run of QSR A good initial value of less than 5% of Max Value of (Input Data) in Figure 5; and
Noise Strength Increment (noiseStrengthlncrementTRComputation) - The amount by which the noise strength is increased for each iteration.
[00389] The process within the Interferometric Couple QSR-NMR engine module is derived from Quantum Stochastic Resonance theory implementation on an NMR system. The QSR module takes as input data a normalized vector of real numbers with absolute amplitudes (<0.3) representing the uncharacterized signal (buried within noise), and essentially implements Quantum Stochastic Resonance dynamics to determine the tunneling rate between the two quantum-mechanical "energy" states within the QSR model. In addition to the tunneling rate the QSR module also generates the "enhanced" data vector of real numbers that have a higher SNR for the buried signal than the original input data. Signal Power for Input Data and Enhanced data is computed by estimating area under the half width at full height. SNR can compute using Area for signal and noise in the two cases: Enhanced Data and Input Data signal-to-noise ratio is a term for the power ratio between a signal and the background noise:
[00390
Figure imgf000088_0001
[00391 ] where P is average power and A is root mean square (RMS) amplitude. Both signal and noise power (or amplitude) must be measured at the same or equivalent points in a system, and within the same system bandwidth.
[00392] The TR threshold of interest established within 828 undertaken during the training phase after the parameters of the quantum resonance interferometry engine have been established, as described above. It is possible though, that the training phase is performed in an iterative mode with the two processes (QSR optimization and TR threshold determination) being repeated a number of times. For example, if one fails to establish a TR threshold with the initial set of optimized parameters, one must readjust the QSR engine parameters by varying the SNR gain. Then, the TR threshold determination process is attempted again.
[00393] In some implementations of the Quantifier the quantifier is a decision block where the TR for the current data sample is compared with the pre-determined threshold value obtained during the training phase, in order to decide whether or not the signal is present in the input data. For the TR threshold determination, a number of control data samples must be used. These data samples fall in two categories:
background - No signal is present; and
signal - A fixed value of signal (with SNR higher than LOD) is present.
[00394] For each category a minimum of 20 samples is desired. Each data sample generates a corresponding TRmean value when processed by the QSR engine. Thus, two sets of TR values corresponding to the two categories are generated. These are designated as TRB and TRs for background and signal respectively. The QSR engine is optimized correctly only if these two sets of TR values are statistically different. The standard deviation and the mean of each set of TR values are computed and assigned: as, MB, as, Ms for background and signal respectively. The critical difference parameter is then used to evaluate the difference between the two TR data sets.
[00395] The critical difference parameter is given by the formula:
CR = (us - MB) / (os 2/ns + OB2 ne)
where ns and ΠΒ are the number of samples in the signal and background datasets respectively. For a well-optimized QSR engine the CR value must be outside the interval [- 1.96, 1.96]. There may be two reasons for the value of CR not falling within the interval. The data samples have not been selected properly (signal is present in the background samples or not enough is present in the signal samples). This can be fixed by revisiting the sets of selected samples or, if this is an option, generate some artificial data samples where the SNR can be tightly controlled. The blocks upstream have not been optimized. This takes the tuning process back to the beginning. The parameters have to be readjusted .
If the two sets are statistically different then the TR threshold is calculated as:
[00396] thresholdxR = (σΒ + os) 2. The threshold so calculated is used in the operation mode to decide convergence.
[00397] Figure 19 shows a method 1900 for calibration of quantum resonance interferometry engine Parameters. For example, quantum resonance interferometry engine is initialized prior to first iteration of the first prospect voxel (as discussed at flows 17100, 17200, 17300, and 17400). Once all the prospect voxel data has been analyzed by the quantum resonance interferometry engine 828, and the attribute of interest parameters are updated to process the next interval (as discussed in connection with in 17145, 17245, 17345 or 17445), the quantum resonance interferometry engine parameters are again updated to process the surviving prospect data voxel as defined in discussion related to Figure 17A through 17D.
[00398] The following discussion uses Porosity-Cube Workflow 17100 as an example. The technique is equally applicable to computation of Litho logy-Cube, Fluid-Volume and Brittleness Volume. Once the porosity range has been established in 17105 and the porosity computations intervals established, energy spectral energy values from the well control voxel data from the appropriate well sections correlated with the different porosity intervals as in Figure 13A are obtained in 1402 and used to initialize the parameters of double well function described in Figure 15 and used in flow 1400.
[00399] At 1902, well control voxel data energy spectral density ranges corresponding to the attribute range of interest are obtained. At 1910, the quantum resonance interferometry engine Parameters are computed and include the following:
[00400] NMR-QSR Double Well parameters include:
Point Count - This is a number representing the size of the Input Data vector. It is preferred that the Point Count be a power of 2 but this is not a requirement. Point Count of 32 (when using delta-pulse kernel), 64 (when using a Gaussian peak kernel) and 128 (when using a CWT - continuous wavelet transform kernel) are used as examples. The actual vector size may depend on the number of raw data elements (for example the number of pixels within the window used to sweep an input image). Conformity (changing the vector size to be a power of 2) is achieved during the preprocessing preconditioning T
transformations. A large value for Point Count may lead to false cause resonances due to coupling between between injected additive, periodic noise and Q-M system which could result in false positives at near LOD signal levels. May also add computational time to the process, a decrease in the computational processing speed of the quantum resonance interferometry engine. The Point Count is actually result of discretization of 1-D 1-peak kernel used for preconditioning and normalizing input to the Interferometric Coupler;
Noise Color - Periodic synthetic noise is required to drive NMR-QSR dynamics. This noise is characterized by its color and strength (amplitude). The default value of noise color may be left unchanged at a noise color of 1 (white noise). This is used as a control parameter to speed/slow down SNR gain when using quantum resonance
interferometry engine. It has limited influence and can destabilize the QSR dynamics is set to be » 1. ) but it is adjustable as a last resort measure for detecting signal when the adjustment of other parameters fails to produce the desired SNR;
Noise Strength - The noise strength (amplitude) is the primary adjustable value outside of double well design, and directly controls the detection of signal by improving the output SNR;
Residency Time - In order for the quantum resonance interferometry engine to decide that a transition has occurred during the processing, the quantum resonance interferometry engine must show some stability, i.e. the state does not exhibit random transitions.-Residency Time is merefore a stability criterion which essentially specifies the minimum time that the system state must remain in one of the two quantum-mechanical states. The Residency Time, can in fact be expressed as the number of time steps (as mentioned above, one time step corresponds to processing the next element within the input data vector). Thus, the minimum value of Residency Time is one. The higher the value of Residency Time, a higher level of stability is achieved at the cost of fewer transitions recorded within the Total Time (length of the input data vector )
Potential Barrier- This is one of the two parameters defining the quartic function of the QSR model. It is used to control the accuracy and confidence in obtaining resonance using quantum resonance interferometry engine;
Potential Minimum - This is the second parameter defining the quartic function of the QSR model;
Activation Temperature;
Particle Mass;
Double Well Energy— Asymmetry is derived from logs (sonic, density, gamma,..) It controls the separation between different zones of interest, in the energy spectral density space;
Thermal Relaxation Time - this provides the protocol for modulation and amplification of (synthetic+seismic) noise introduced into the quantum resonance interferometry engine;
Decoherence Relaxation Time represent the noise that is required to be added" to the system to compensate for losses due to environmental decoherence;
Maximum Energy Ratio - This is used in the quantum resonance interferometry engine to check if the system energy stays within reasonable bounds of the quartic potential. It is not critical for the performance of the engine and a default value can be determined during the optimization process;
Regulator Threshold is used to establish a confidence measure on resonance events;
Initial Position - This is the initial state of the system (x,y,z) in the Bloch Rate equations of motion given by Eqns. 10, 11, and 12. This parameter has to lie within the bounds of the functional region of the quartic function. Its influence is minimal and only serves as the starting point of the iteration. A valid initial value of the Initial Position can be calculated from the other parameters defining the quartic function.
[00401 ] At 1920, there are three parameters in tuning the TR convergence module
Maximum Convergence Iterations - This is the maximum number of iterations after which Interferometric Coupler Unit 833 abandons attempts to achieve TR convergence. A default value for this parameter is set to 100. Higher values increase the likelihood of convergence but required increased computation time;
Convergence Confidence - This is the number of consecutive iterations wherein the TR value must stay within an Epsilon bound on the mean of TR in order to conclude convergence. The value of this parameter should be a fraction of the maximum convergence iterations (the default value of 5% can be used). Higher values will increase the level of confidence with the cost of increased execution time;
Convergence Epsilon - is the width of the interval around a mean value within which the TR values are allowed to fluctuate for meeting the Convergence criterion.
Convergence is concluded if the difference between the maximum and the minimum values of arithmetic mean of TR is smaller than Convergence Epsilon for the pre-specified number of consecutive iterations. The convergence confidence level is determined by the combination of the values of this parameter and the Convergence Confidence setting.
[00402] TR threshold atl920 is established during the training phase after the parameters of the quantum resonance interferometry engine have been set. However, a training phase can be performed in an iterative mode with two processes (quantum resonance interferometry engine optimization and TR threshold determination) being repeated a number of times. For example, if one fails to establish a TR threshold with the initial set of optimized parameters, the quantum resonance interferometry engine parameters can be adjusted by varying the SNR gain. Then, the TR threshold determination process is re-attempted.
[00403] To compare the TR for a Prospect Data the current data sample is compared with the pre-determined threshold value obtained during the training phase, in order to decide whether or not the signal is present in the Input Data. For the TR threshold determination, a number of well control voxel data samples are needed. These data samples fall in two categories: termed as "Background" - i.e., no attribute of interest (i.e., attribute above a cutoff) is present; and, termed as "Signal" - a fixed value of well control voxel data with attributes of interest present above the cut-off value, also referred to as "Signal Above LOD". [00404] TR threshold can be established using as few as 1 well control voxel data. Each data sample generates a corresponding TR value when processed by the quantum resonance interferometry engine. Thus, two sets of TR values corresponding to the two categories are generated. These are designated as TRB and TRS for the Background and Signal respectively. The quantum resonance interferometry engine is optimized correctly only if these two sets of TR values are statistically different. The standard deviation and the mean of each set of TR values are computed and assigned: oB, uB, oS, uS for Background and Signal respectively. The critical difference parameter is then used to evaluate the difference between the two TR data sets.
[00405] The critical difference parameter is given by the formula:
[00406] CR = (uS - μΒ) / (oS2/nS + oB2/nB)
[00407] where nS and nB are the number of samples in the signal and background datasets respectively. For a well-optimized quantum resonance interferometry engine the CR value must be outside the interval [-1.96, 1.96]. There may be two reasons for the value of CR not falling within the interval. The data samples have not been selected properly (Signal is present in the Background samples or not enough is present in the Signal samples). This can be fixed by revisiting the sets of selected samples.
[00408] If the two sets are statistically different then the TR threshold is calculated as: thresholdTR = (σΒ + oS)/2.The threshold so calculated is used in the operation mode to decide convergence.
[00409] At 1940, bounds are derived on the quantum resonance interferometry engine iterations.
[00410] - Total Time - quantum resonance interferometry engine processes normalized excitation cascade as a 1 -D array time series of prospect voxel data vector. Thus, the Total Time parameter determines the incremental time difference between each element of the Input Data The first data element is assigned a time value of zero and the last element has a time value equal to: (PointCount - 1) * TotalTime / PointCount. This parameter establish scales at which transitions are detected vis-a-vis scale at which SNR is increased by quantum resonance interferometry engine.
[00411] At 1950, quantum resonance interferometry iteration is count is incremented.
[00412] - Number of Iterations - This is the maximum number of times the QSR engine is run within the QSR Iterator module. In principle, the larger the number of iterations, the better the chance of detecting the signal at the cost of longer execution times. Careful attention must be paid to the selection of this important parameter value, since a high price may be paid in the overall increase in computational time, with no comparable benefit in SNR gain.
[00413] At 1960, a summary is shown of NMR-QSR well design double parameters. These include energy asymmetry related to the energy spectral density of zone separation in well logs. The well separation is used to modulate the desired separation between zones of interest (e.g., oil and water). Barrier height is used to establish desired confidence. While the number of iterations are related to the inherent noise and uncertainty in the data.
[00414] Figure 20 shows an example workflow 2000 for producing output with various processes discussed herein. Workflow 2000 uses processes described in this disclosureto produce output for sizing hydrocarbon reservoir and generating drilling targets for exploration and production.
[00415] Results of a conventional seismic processing workflow sequence (using e.g, standard off the shelf software geophysical packages to produce pre-stack time or depth migrated gathers) is obtained at 220. In some examples, it is preferred to preserve amplitudes, spectrum and phase information and avoid band-pass filter. Seismic horizons are generated to demarcate formation of interest using best-practice methods of horizon picking. Spectral decompositions are performed on seismic data obtained at 220 to partition a seismic volume into low frequency, prime amplitude and high frequency partitioned volumes as in 560.
[00416] A Porosity-Cube is produced at 17160, using process 17100 with either high frequency or prime-amplitude partitions as an input.. A Lithology-Cube is produced at 17260 using process 17200 using either high frequency or prime-amplitude partitions as an input. A Fluid Volume is produced as in 17360, using process 17300 with either high frequency or prime-amplitude partitions as an input,.
Processes 17100, 17200 and 17300 use flows in Figure 9A through 9C to produce seismic data noise at 9170, and sinusoidal periodic modulated synthetic noise 842 is generated from random number generators . The seismic data noise and synthetic noise are combined to drive a quantum resonance interferometry engine used to produce Porosity-Cube, Lithology-Cube and Fluid Volume by analyzing a sequence of prospect voxel data derived from seismic partitioned volumes.
[00417] Results of flow 17100, 17200, and 17300 are combined as in 2050 to identify and highlight prospect voxel data within result volumes to determine voxels with porosity values above a prespecified cut-off, Vshale values that meet criteria of interest and voxels that show presence of fluid of interest. Voxels meeting all the three criteria are aggregated to assemble a subsurface geobody. The size, location, and orientation of the geobody in the seismic volume defines the properties of potential hydrocarbon reservoir. As the size of individual prospect voxel data is known, the size of the reservoir geobody is computed. Results from 2050 are used to make hydrocarbon exploration, drilling and production decisions. In some implementations, drilling, exploration and production decisions using a subset (e.g, just one or two) of Porosity-Cube, Lithology-Cube or Fluid Volume Results.
[00418] Implementations of the subject matter described in this specification can be implemented as one or more computer programs, i.e., one or more modules of computer program instructions, encoded on a computer storage medium for execution by, or to control the operation of, data processing apparatus. Alternatively or in addition, the program instructions can be encoded on an artificially generated propagated signal, e.g., a machine- generated electrical, optical, or electromagnetic signal that is generated to encode information for transmission to suitable receiver apparatus for execution by a data processing apparatus. A computer storage medium can be, or be included in, a computer-readable storage device, a computer-readable storage substrate, a random or serial access memory array or device, or a combination of one or more of them. Moreover, while a computer storage medium is not a propagated signal,. a computer storage medium can be a source or destination of computer program instructions encoded in an artificially generated propagated signal. The computer storage medium can also be, or be included in, one or more separate physical components or media (e.g., multiple CDs, disks, or other storage devices).
[00419] The operations described in this specification can be implemented as operations performed by a data processing apparatus on data stored on one or more computer-readable storage devices or.received from other sources.
[00420] The term "data processing apparatus" encompasses all kinds of apparatus, devices, and machines for processing data, including by way of example a programmable processor, a computer, a system on a chip, or multiple ones, or combinations, of the foregoing. The apparatus can include special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application specific integrated circuit). The apparatus can also include, in addition to hardware, code that creates an execution environment for the computer program in question, e.g., code that constitutes processor firmware, a protocol stack, a database management system, an operating system, a cross-platform runtime environment, a virtual machine, or a combination of one or more of them. The apparatus and execution
environment can realize various different computing model infrastructures, such as web services, distributed computing, and grid computing infrastructures. [00421] A computer program (also known as a program, software, software application, script, or code) can be written in any form of programming language, including compiled or interpreted languages, declarative or procedural languages, and it can be deployed in any form, including as a standalone program or as a module, component, subroutine, object, or other unit suitable for use in a computing environment. A computer program may, but need not, correspond to a file in a file system. A program can be stored in a portion of a file that holds other programs or data (e.g., one or more scripts stored in a markup language document), in a single file dedicated to the program in question, or in multiple coordinated files (e.g., files that store one or more modules, sub programs, or portions of code). A computer program can be deployed to be executed on one computer or on multiple computers that are located at one site or distributed across multiple sites and interconnected by a communication network.
[00422] The processes and logic flows described in this specification can be performed by one or more programmable processors executing one or more computer programs to perform actions by operating on input data and generating output. The processes and logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application specific integrated circuit).
[00423] Processors suitable for the execution of a computer program include, by way of example, both general and special purpose microprocessors, and any one or more processors of any kind of digital computer. Generally, a processor will receive instructions and data from a read only memory or a random access memory or both. The essential elements of a computer are a processor for performing actions in accordance with instructions and one or more memory devices for storing instructions and data. Generally, a computer will also include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto optical disks, or optical disks. However, a computer need not have such devices. Moreover, a computer can be embedded in another device, e.g., a mobile telephone, a personal digital assistant (PDA), a mobile audio or video player, a game console, a Global Positioning System (GPS) receiver, or a portable storage device (e.g., a universal serial bus (USB) flash drive), to name just a few. Devices suitable for storing computer program instructions and data include all forms of nonvolatile memory, media and memory devices, including by way of example
semiconductor memory devices, e.g., EPROM, EEPROM, and flash memory devices;
magnetic disks, e.g., internal hard disks or removable disks; magneto optical disks; and CD ROM and DVD-ROM disks. The processor and the memory can be supplemented by, or incorporated in, special purpose logic circuitry.
[00424] To provide for interaction with a user, implementations of the subject matter described in this specification can be implemented on a computer having a display device, e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor, for displaying information to the user and a keyboard and a pointing device, e.g., a mouse or a trackball, by which the user can provide input to the computer. Other kinds of devices can be used to provide for interaction with a user as well; for example, feedback provided to the user can be any form of sensory feedback, e.g., visual feedback, auditory feedback, or tactile feedback; and input from the user can be received in any form, including acoustic, speech, or tactile input. In addition, a computer can interact with a user by sending documents to and receiving documents from a device that is used by the user; for example, by sending web pages to a web browser on a user's client device in response to requests received from the web browser.
[00425] Implementations of the subject matter described in this specification can be implemented in a computing system that includes a back end component, e.g., as a data server, or that includes a middleware component, e.g., an application server, or that includes a front end component, e.g., a client computer having a graphical user interface or a Web browser through which a user can interact with an implementation of the subject matter described in this specification, or any combination of one or more such back end, middleware, or front end components. The components of the system can be interconnected by any form or medium of digital data communication, e.g., a communication network. Examples of communication networks include a local area network ("LAN") and a wide area network ("WAN"), an inter-network (e.g., the Internet), peer-to-peer networks (e.g., ad hoc peer-to-peer networks), wireless networks, mobile phone networks etc..
[00426] The computing system can include clients and servers. A client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other. In some implementations, a server transmits data (e.g., an HTML page) to a client device (e.g. , for purposes of displaying data to and receiving user input from a user interacting with the client device). Data generated at the client device (e.g., a result of the user interaction) can be received from the client device at the server.
[00427] A number of implementations have been described. Nevertheless, it will be understood that various modifications may be made. In addition, other steps may be provided, or steps may be eliminated, from the described flows, and other components may be added to, or removed from, the described systems. Accordingly, other implementations are within the scope of the following claims.

Claims

CLAIMS What is claimed is:
1. A machine-implemented method, comprising:
voxelizing seismic data for a geological subsurface formation of interest into multiple voxels having respective locations in the formation of interest; and
determining whether a voxel in the multiple voxels includes an attribute; and outputting whether the voxel includes the attribute to an attribute volume based on the location; and
performing the determining and the outputting for at least some other of the multiple voxels.
2. The method of claim 1 ,
wherein determining whether a voxel in the multiple voxels includes an attribute comprises determining whether the voxel in the multiple voxels includes a porosity attribute; and
the method further comprising assembling the at least some of the other of the mulitple voxels into a porosity cube.
3. The method of claim 1 ,
wherein determining whether a voxel in the multiple voxels includes an attribute comprises determining whether the voxel in the multiple voxels includes a lithology attribute; and
the method further comprising assembling the at least some of the other of the multiple voxels into a lithology cube.
4. The method of claim 1 ,
wherein determining whether a voxel in the multiple voxels includes an attribute comprises determining whether the voxel in the multiple voxels includes at least one of gas, oil, or water; and
the method further comprising assembling the at least some of the other of the multiple voxels into a liquid cube.
5. The method of claim 1 , wherein the seismic data comprises migrated prestack time gathers for the formation of interest.
6. The method of claim 1, further comprising:
wherein determining whether a voxel in the multiple voxels includes an attribute comprises determining whether the voxel in the multiple voxels includes a lithology attribute; repeating the determining and the outputting for a porosity attribute; and
generating a drill-here map indicating a location for drilling based on the at least some of the other of the multiple voxels for the lithology attribute and based on the at least some of the other of the multiple voxels for the porosity attribute.
7. The method of claim 1 ,
wherein voxelizing the seismic data further comprises spectrally decomposing the seismic data into frequency volumes.
8. The method of claim 7, further comprising spectrally decomposing the seismic data into a high frequency volume, a medium frequency volume, and a low frequency volume.
9. The method of claim 1, wherein the seismic data comprises three dimensional seismic data and wherein the voxelizing comprises generating, from the three dimension seismic data, multiple one dimension voxel vectors.
10. The method of claim 9, wherein determining whether a voxel in the multiple voxels includes an attribute comprises:
obtaining control data for that attribute based on information from one or more previously explored geological subsurface wells; and
processing the one dimension voxel vector for the voxel using the control data and using quantum resonance interferometry to detect a resonance.
11. The method of claim 1 , further comprising performing the steps of claim 1 for the seismic data for another formation of interest.
12. The method of claim 1 , further comprising characterizing a structural trap based on the
attribute volume.
13. The method of claim 1 , further comprising characterizing a stratigraphic trap based on the attribute volume.
14. The method of claim 1 , further comprising upscaling the multiple voxels to a predetermined resolution.
15. The method of claim 1 wherein:
wherein the voxelizing comprises generating a one-dimensional voxel vector for the voxel;
obtaining spectral data generated from well-log data associated with an attribute of interest;
wherein the determining comprises:
coupling the spectral data with the one dimensional voxel vector to determine whether a resonance event occurs; and
when a resonance event occurs, producing an output indicating that the voxel has the attribute of interest, and
when the resonance event does not occur, producing an output indicating that the voxel does not have the attribute of interest.
16. The method of claim 15,
wherein the generating the one-dimensional voxel vector comprises normalizing amplitudes corresponding to the one dimensional voxel vector to fit within a range.
17. The method of claim 15, wherein obtaining the spectral data generated from well-log data comprises:
obtaining seismic noise data;
obtaining well control voxel data from the well log data; and
combining the seismic noise data and the well control voxel data using a quantum mechanical model to produce the spectral data.
18. The method of claim 17, wherein obtaining well control voxel data comprises:
determining an energy spectral density voxel from the well-log data for the attribute using wavelets derived from sonic logs in sections of a well with the attribute of interest.
19. The method of claim 17 wherein combining the seismic noise data and the well control data comprises combining the seismic noise data and with well control voxel data for a voxel associated with the well-log data that exhibits the attribute.
20. The method of claim 17,
wherein obtaining the well control voxel data comprises obtaining a well control voxel data of a voxel not exhibiting the attribute; and wherein obtaining the seismic noise data comprises generating the seismic noise data based on the well control voxel data of the voxel not exhibiting the noise.
21. The method of claim 11, wherein coupling the spectral data with the one dimensional voxel vector to determine whether a resonance event occurs comprises using a nuclear magnetic resonance ("NMR") master rate equation to generate quantum stochastic resonance based on the one-dimensional voxel vector, synthetic noise, and the spectral data.
22. A computer-readable medium encoding a computer program product operable to cause data processing apparatus to perform operations comprising;
obtaining voxel data for a voxel from voxelized seismic data for a geological subsurface formation;
performing a first non-linear coupling of the voxel data with spectral data to generate a first resonance, wherein the coupling is driven by noise having an intensity within a first cutoff band;
in response to generating the first resonance, adjusting the cutoff band to a second cutoff band, different than the first cutoff band;
performing a second non-linear coupling of the voxel data with the spectral data associated with an attribute of the subsurface formation to generate a second resonance, wherein the second coupling is driven by noise having an intensity within a second cutoff band;
in response to generating the second resonance, generating an indication that the attribute exists in the voxel; and
in response to the second coupling not producing a second resonance producing an indication that the attribute does not exists in the voxel
23. The computer-readable medium of claim 22, wherein the first resonance comprises a first quantum stochastic resonance.
24. The computer-readable medium of claim 22, the operations further comprising assembling the voxel data in an attribute volume with the indication that that the attribute exists in the voxel.
25. The computer-readable medium of claim 22:
wherein the voxelized seismic data comprises multiple voxels including the voxel; and the operations further comprising determining the first cutoff band based on a percentage of an average intensity of the multiple voxels.
26. The computer-readable medium of claim 22, the operations comprising:
prior to the first coupling, performing a third coupling of the voxel data with the spectral data associated with an attribute of the subsurface formation to generate a third resonance, wherein the third coupling is driven by noise having an intensity within a third, different cutoff band; and
in response to generating the first resonance, adjusting the third cutoff band to the first cutoff band.
27. An apparatus comprising:
means for obtaining first prospect voxel data for a voxel from voxelized seismic data for a geological subsurface formation on a prospect;
means for determining an attribute iteration range including values for the attribute including at least a first value and a second value;
means for determining control well voxel data from well-log data for an existing well, the control well voxel data including spectral energies for a listing of values of the attribute found in the existing well, including a first spectral energy for the first value and a second spectral energy for the second value;
means for coupling the prospect voxel data with spectral data having parameters set according to the first spectral energy to produce a first resonance; and
a means for generating an output indicating that the voxel includes the attribute at at least the first value, in response to the first resonance.
28. The apparatus of claim 27, further comprising
a means for coupling the prospect voxel data with spectral data having parameters set according to the second spectral energy to produce a second resonance; and
in response to the second resonance, a means for generating an output indicating that the voxel includes the attribute at at least the second value.
29. The apparatus of claim 28, wherein
wherein the control well voxel data comprises a third spectral data corresponding to a third value for the attribute; and
further comprising: a means for coupling the prospect voxel data with spectral data having parameters set according to the third spectral energy to produce a third resonance; and
in response to the third resonance not being produced, a means for generating an output indicating that the voxel does not include the attribute at the third value but includes the attribute at the first and second value.
30. The apparatus of claim 27, wherein the attribute comprises porosity and wherein the first value comprises a first percentage of porosity and the second value comprises a third percentage of porosity greater that the first percentage
PCT/US2011/027456 2010-03-05 2011-03-07 Active noise injection computations for improved predictability in oil and gas reservoir discovery and characterization WO2011109839A2 (en)

Priority Applications (6)

Application Number Priority Date Filing Date Title
RU2012142308/08A RU2012142308A (en) 2010-03-05 2011-03-07 CALCULATION OF ACTIVE NOISE INJECTION FOR IMPROVED PREDICTION DURING DETECTION OF OIL AND GAS COLLECTOR AND DETERMINATION OF ITS CHARACTERISTICS
CN2011800223235A CN102893183A (en) 2010-03-05 2011-03-07 Active noise injection computations for improved predictability in oil and gas reservoir discovery and characterization
BR112012022450A BR112012022450A2 (en) 2010-03-05 2011-03-07 computational injections of active noise to improve predictability of oil and gas reservoir discoveries and characterization
EP11751523A EP2542918A2 (en) 2010-03-05 2011-03-07 Active noise injection computations for improved predictability in oil and gas reservoir discovery and characterization
MX2012010271A MX2012010271A (en) 2010-03-05 2011-03-07 Active noise injection computations for improved predictability in oil and gas reservoir discovery and characterization.
CA2792052A CA2792052A1 (en) 2010-03-05 2011-03-07 Active noise injection computations for improved predictability in oil and gas reservoir discovery and characterization

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US31122710P 2010-03-05 2010-03-05
US61/311,227 2010-03-05
US32703510P 2010-04-22 2010-04-22
US61/327,035 2010-04-22

Publications (2)

Publication Number Publication Date
WO2011109839A2 true WO2011109839A2 (en) 2011-09-09
WO2011109839A3 WO2011109839A3 (en) 2011-12-29

Family

ID=44542883

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2011/027456 WO2011109839A2 (en) 2010-03-05 2011-03-07 Active noise injection computations for improved predictability in oil and gas reservoir discovery and characterization

Country Status (7)

Country Link
EP (1) EP2542918A2 (en)
CN (1) CN102893183A (en)
BR (1) BR112012022450A2 (en)
CA (1) CA2792052A1 (en)
MX (1) MX2012010271A (en)
RU (1) RU2012142308A (en)
WO (1) WO2011109839A2 (en)

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015185991A3 (en) * 2014-06-02 2016-04-14 Cgg Services Sa Spectral analysis and processing of seismic data using orthogonal image gathers
FR3060174A1 (en) * 2016-12-09 2018-06-15 Landmark Graphics Corporation WAVELET ESTIMATION FOR A QUADRIDIMENSIONAL CHARACTERIZATION OF SUBSURFACE PROPERTIES BASED ON DYNAMIC SIMULATION
CN108629835A (en) * 2017-03-20 2018-10-09 哈尔滨工业大学 Based on EO-1 hyperion, true coloured picture and the indoor method for reconstructing and system for putting cloud complementation
WO2020112665A1 (en) * 2018-11-26 2020-06-04 Saudi Arabian Oil Company Frequency based geological feature detection from seismic data
CN111339671A (en) * 2020-02-28 2020-06-26 西安石油大学 Shale reservoir bidirectional flow-solid coupling numerical calculation method
CN111381278A (en) * 2018-12-29 2020-07-07 中国石油天然气股份有限公司 Method and device for predicting parameters of carbonate rock
CN111736221A (en) * 2020-05-15 2020-10-02 中国石油天然气集团有限公司 Amplitude fidelity determination method and system
CN111767887A (en) * 2020-07-08 2020-10-13 吉林大学 Transient electromagnetic data processing method based on wavelet decomposition and IME frequency estimation
CN111898065A (en) * 2020-06-18 2020-11-06 长江大学 Shale brittleness index calculation method and device and readable storage medium
CN112147675A (en) * 2019-06-28 2020-12-29 中国石油天然气股份有限公司 Reservoir fluid detection method and device
US11295399B2 (en) * 2017-02-22 2022-04-05 Landmark Graphics Corporation Well exploration, service and production appliance architecture
US11333780B2 (en) 2020-10-09 2022-05-17 Saudi Arabian Oil Company Method and system for processing a three-dimensional (3D) seismic dataset
CN114660269A (en) * 2022-03-23 2022-06-24 中国海洋石油集团有限公司 Recovery method for dynamic natural gas accumulation process of ancient buried hill
US20220221614A1 (en) * 2021-01-11 2022-07-14 Shandong University Of Science And Technology Analysis method, system and storage media of lithological and oil and gas containing properties of reservoirs
US11409012B2 (en) 2019-10-21 2022-08-09 Saudi Arabian Oil Company Frequency based method for reducing the effect of multiples in seismic data
CN115381400A (en) * 2022-08-19 2022-11-25 四川轻化工大学 Brain disease patient classification system and method based on feature selection and SVM
US11520075B2 (en) 2016-12-12 2022-12-06 Schlumberger Technology Corporation Augmented geological service characterization
US11592589B2 (en) 2021-01-14 2023-02-28 Saudi Arabian Oil Company Seismic attribute map for gas detection
CN116088047A (en) * 2023-03-22 2023-05-09 广东石油化工学院 Oil and gas reservoir searching method and system based on fault model
US11977198B2 (en) 2020-10-06 2024-05-07 Saudi Arabian Oil Company Isofrequency volumes ratio workflow to detect gas reservoirs in 3D domain

Families Citing this family (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105425287B (en) * 2015-11-09 2017-10-27 中国地质大学(北京) The prestack separation method of seismic wave
FR3053125B1 (en) * 2016-06-23 2018-07-27 Storengy METHOD FOR CHARACTERIZING THE BASEMENT OF A REGION USING PASSIVE SEISMIC SIGNALS, AND CORRESPONDING SYSTEM
CN107967711A (en) * 2016-10-18 2018-04-27 中国石油化工股份有限公司 The visualization of oil reservoir block models integrates and fine pick device and method
CN106814397B (en) * 2016-12-21 2019-08-06 长江大学 A kind of method that multi-parameter joint inversion calculates rock scatter attenuation
CN107918142B (en) * 2017-09-08 2019-07-19 北京派特森科技股份有限公司 A kind of method of seismic prospecting
CN107664771B (en) * 2017-09-28 2019-03-12 西南石油大学 A kind of microseism Full wave shape localization method based on likeness coefficient
CA3088135A1 (en) * 2018-01-31 2019-08-08 Google Llc Quantum computation through reinforcement learning
US11506804B2 (en) * 2018-05-31 2022-11-22 Saudi Arabian Oil Company Inverse stratigraphic modeling using a hybrid linear and nonlinear algorithm
CN110837116B (en) * 2018-08-15 2021-06-01 中国石油天然气股份有限公司 Method for determining operation upper limit pressure of salt cavern gas storage
KR102189311B1 (en) * 2018-08-21 2020-12-09 두산중공업 주식회사 An apparatus of analysis and a method therefor
CN110056345B (en) * 2018-10-30 2020-11-24 西安石油大学 Logging evaluation method suitable for shale gas reservoir
CN114961712A (en) * 2021-02-25 2022-08-30 中国石油化工股份有限公司 Method and device for discriminating reverse fault by conventional logging curve
CN113156526B (en) * 2021-04-26 2023-06-27 中国地质科学院地球物理地球化学勘查研究所 Full-region multi-source electromagnetic sounding method and multi-field source multi-component data joint inversion technology
CN113238288B (en) * 2021-05-20 2022-07-01 桂林电子科技大学 Rotor wing target feature extraction method based on difference spectral line
CN114089416B (en) * 2021-11-17 2023-02-21 成都理工大学 Method for estimating attenuation gradient of seismic waves by utilizing Schrodinger equation
CN114137616B (en) * 2021-12-03 2023-08-11 成都信息工程大学 Method for detecting gas content of reservoir by utilizing quantum mechanics principle
CN114152981B (en) * 2021-12-03 2023-08-08 成都信息工程大学 Method for estimating quality factors of seismic data in potential energy wave function domain
CN114492213B (en) * 2022-04-18 2022-07-01 中国石油大学(华东) Wavelet neural operator network model-based residual oil saturation and pressure prediction method
CN114895359B (en) * 2022-07-13 2022-09-13 中国科学院地质与地球物理研究所 DAS (data acquisition system) same-well monitoring real-time microseism effective event denoising method and system
CN116068644B (en) * 2022-08-31 2024-07-09 中国地质大学(武汉) Method for improving resolution and noise reduction of seismic data by using generation countermeasure network
CN116256801B (en) * 2023-05-16 2023-07-21 中国科学院地质与地球物理研究所 Deep oil gas accurate navigation fault characterization method and system based on image fusion

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6473697B2 (en) * 2000-02-25 2002-10-29 Shell Oil Company Processing seismic data
US6912467B2 (en) * 2002-10-08 2005-06-28 Exxonmobil Upstream Research Company Method for estimation of size and analysis of connectivity of bodies in 2- and 3-dimensional data
US7113869B2 (en) * 2003-05-22 2006-09-26 Schlumberger Technology Corporation Method for prospect identification in asset evaluation
US20080272775A1 (en) * 2007-05-02 2008-11-06 Feng Derek D Quantum theory-based continuous precision nmr/mri: method and apparatus

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6136541A (en) * 1999-02-22 2000-10-24 Vialogy Corporation Method and apparatus for analyzing hybridized biochip patterns using resonance interactions employing quantum expressor functions
US6142681A (en) * 1999-02-22 2000-11-07 Vialogy Corporation Method and apparatus for interpreting hybridized bioelectronic DNA microarray patterns using self-scaling convergent reverberant dynamics
CN100353184C (en) * 2005-06-29 2007-12-05 吴会良 Method for detecting oil and gas by means of resonance principle
EP2372398A3 (en) * 2006-06-21 2012-04-11 Terraspark Geosciences, LLC Extraction of depositional systems
CN101149439B (en) * 2007-11-13 2010-06-30 符力耘 High resolution ratio non-linear reservoir properties inversion method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6473697B2 (en) * 2000-02-25 2002-10-29 Shell Oil Company Processing seismic data
US6912467B2 (en) * 2002-10-08 2005-06-28 Exxonmobil Upstream Research Company Method for estimation of size and analysis of connectivity of bodies in 2- and 3-dimensional data
US7113869B2 (en) * 2003-05-22 2006-09-26 Schlumberger Technology Corporation Method for prospect identification in asset evaluation
US20080272775A1 (en) * 2007-05-02 2008-11-06 Feng Derek D Quantum theory-based continuous precision nmr/mri: method and apparatus

Cited By (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10670758B2 (en) 2014-06-02 2020-06-02 Cgg Services Sas Spectral analysis and processing of seismic data using orthogonal image gathers
WO2015185991A3 (en) * 2014-06-02 2016-04-14 Cgg Services Sa Spectral analysis and processing of seismic data using orthogonal image gathers
FR3060174A1 (en) * 2016-12-09 2018-06-15 Landmark Graphics Corporation WAVELET ESTIMATION FOR A QUADRIDIMENSIONAL CHARACTERIZATION OF SUBSURFACE PROPERTIES BASED ON DYNAMIC SIMULATION
US11520075B2 (en) 2016-12-12 2022-12-06 Schlumberger Technology Corporation Augmented geological service characterization
US11295399B2 (en) * 2017-02-22 2022-04-05 Landmark Graphics Corporation Well exploration, service and production appliance architecture
CN108629835A (en) * 2017-03-20 2018-10-09 哈尔滨工业大学 Based on EO-1 hyperion, true coloured picture and the indoor method for reconstructing and system for putting cloud complementation
WO2020112665A1 (en) * 2018-11-26 2020-06-04 Saudi Arabian Oil Company Frequency based geological feature detection from seismic data
US11150368B2 (en) 2018-11-26 2021-10-19 Saudi Arabian Oil Company Frequency based geological feature detection from seismic data
CN111381278A (en) * 2018-12-29 2020-07-07 中国石油天然气股份有限公司 Method and device for predicting parameters of carbonate rock
CN111381278B (en) * 2018-12-29 2022-08-30 中国石油天然气股份有限公司 Method and device for predicting parameters of carbonate rocks
CN112147675B (en) * 2019-06-28 2022-07-05 中国石油天然气股份有限公司 Reservoir fluid detection method and device
CN112147675A (en) * 2019-06-28 2020-12-29 中国石油天然气股份有限公司 Reservoir fluid detection method and device
US11409012B2 (en) 2019-10-21 2022-08-09 Saudi Arabian Oil Company Frequency based method for reducing the effect of multiples in seismic data
CN111339671A (en) * 2020-02-28 2020-06-26 西安石油大学 Shale reservoir bidirectional flow-solid coupling numerical calculation method
CN111339671B (en) * 2020-02-28 2023-03-28 西安石油大学 Shale reservoir bidirectional flow-solid coupling numerical calculation method
CN111736221A (en) * 2020-05-15 2020-10-02 中国石油天然气集团有限公司 Amplitude fidelity determination method and system
CN111736221B (en) * 2020-05-15 2023-08-22 中国石油天然气集团有限公司 Amplitude fidelity determination method and system
CN111898065B (en) * 2020-06-18 2023-06-06 长江大学 Shale brittleness index calculation method, equipment and readable storage medium
CN111898065A (en) * 2020-06-18 2020-11-06 长江大学 Shale brittleness index calculation method and device and readable storage medium
CN111767887A (en) * 2020-07-08 2020-10-13 吉林大学 Transient electromagnetic data processing method based on wavelet decomposition and IME frequency estimation
CN111767887B (en) * 2020-07-08 2022-06-28 吉林大学 Transient electromagnetic data processing method based on wavelet decomposition and IME frequency estimation
US11977198B2 (en) 2020-10-06 2024-05-07 Saudi Arabian Oil Company Isofrequency volumes ratio workflow to detect gas reservoirs in 3D domain
US11333780B2 (en) 2020-10-09 2022-05-17 Saudi Arabian Oil Company Method and system for processing a three-dimensional (3D) seismic dataset
US20220221614A1 (en) * 2021-01-11 2022-07-14 Shandong University Of Science And Technology Analysis method, system and storage media of lithological and oil and gas containing properties of reservoirs
US11592589B2 (en) 2021-01-14 2023-02-28 Saudi Arabian Oil Company Seismic attribute map for gas detection
CN114660269A (en) * 2022-03-23 2022-06-24 中国海洋石油集团有限公司 Recovery method for dynamic natural gas accumulation process of ancient buried hill
CN114660269B (en) * 2022-03-23 2023-09-19 中国海洋石油集团有限公司 Method for recovering natural gas dynamic hiding process of ancient buried hill
CN115381400A (en) * 2022-08-19 2022-11-25 四川轻化工大学 Brain disease patient classification system and method based on feature selection and SVM
CN116088047A (en) * 2023-03-22 2023-05-09 广东石油化工学院 Oil and gas reservoir searching method and system based on fault model
CN116088047B (en) * 2023-03-22 2023-06-09 广东石油化工学院 Oil and gas reservoir searching method and system based on fault model

Also Published As

Publication number Publication date
RU2012142308A (en) 2014-04-10
BR112012022450A2 (en) 2016-07-12
WO2011109839A3 (en) 2011-12-29
EP2542918A2 (en) 2013-01-09
MX2012010271A (en) 2012-12-17
CN102893183A (en) 2013-01-23
CA2792052A1 (en) 2011-09-09

Similar Documents

Publication Publication Date Title
US8612156B2 (en) Active noise injection computations for improved predictability in oil and gas reservoir discovery and characterization
US8649980B2 (en) Active noise injection computations for improved predictability in oil and gas reservoir characterization and microseismic event analysis
WO2011109839A2 (en) Active noise injection computations for improved predictability in oil and gas reservoir discovery and characterization
US10816687B2 (en) Method for estimating petrophysical properties for single or multiple scenarios from several spectrally variable seismic and full wavefield inversion products
Hansen et al. Linear inverse Gaussian theory and geostatistics
US9841518B2 (en) Noise attenuation
US11231511B2 (en) Reflection seismology internal multiple estimation
US20160334528A1 (en) Systems and methods for characterizing subterranean formations utilizing azimuthal data
US11221425B1 (en) Generating a model for seismic velocities in a subsurface region using inversion with lateral variations
Lindsay et al. Sequential Backus averaging: Upscaling well logs to seismic wavelengths
EA030770B1 (en) System and method for seismic adaptive optics
Mushin et al. Structural–formational interpretation tools for seismic stratigraphy
Zhou First-break vertical seismic profiling tomography for Vinton Salt Dome
Chopra et al. Seismic reservoir characterization of Bone Spring and Wolfcamp Formations in the Delaware Basin—A case study: Part 1
CN108802817B (en) Multi-aperture depth migration imaging method, device and system
CN105989240A (en) Sand layer prediction method and device
Singh et al. Poststack, prestack, and joint inversion of P-and S-wave data for Morrow A sandstone characterization
Thore et al. 4D propagated layer-based inversion
Nedorub et al. SEG Technical Program Expanded Abstracts 2020
Jeong Quantitative reservoir characterization integrating seismic data and geological scenario uncertainty
Al-Busaidi et al. Optimizing Stringers Development Using Latest Probabilistic Seismic Inversion Technique
Jaradat Prediction of reservoir properties of the N-sand, Vermilion Block 50, Gulf of Mexico, from multivariate seismic attributes
Fomel SEG technical program expanded abstracts, 2014
Bevc et al. SEG Technical Program Expanded Abstracts 2019
Popovici et al. SEG Technical Program Expanded Abstracts 2017

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 201180022323.5

Country of ref document: CN

WWE Wipo information: entry into national phase

Ref document number: 221736

Country of ref document: IL

ENP Entry into the national phase

Ref document number: 2792052

Country of ref document: CA

WWE Wipo information: entry into national phase

Ref document number: 7694/DELNP/2012

Country of ref document: IN

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: MX/A/2012/010271

Country of ref document: MX

REEP Request for entry into the european phase

Ref document number: 2011751523

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2011751523

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2012142308

Country of ref document: RU

REG Reference to national code

Ref country code: BR

Ref legal event code: B01A

Ref document number: 112012022450

Country of ref document: BR

ENP Entry into the national phase

Ref document number: 112012022450

Country of ref document: BR

Kind code of ref document: A2

Effective date: 20120905